Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

powerpc: Fix typos

Fix typos, most reported by "codespell arch/powerpc". Only touches
comments, no code changes.

Signed-off-by: Bjorn Helgaas <bhelgaas@google.com>
Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
Link: https://msgid.link/20240103231605.1801364-8-helgaas@kernel.org

authored by

Bjorn Helgaas and committed by
Michael Ellerman
0ddbbb89 39434af1

+40 -40
+2 -2
arch/powerpc/boot/Makefile
··· 108 108 # these files into the build dir, fix up any includes and ensure that dependent 109 109 # files are copied in the right order. 110 110 111 - # these need to be seperate variables because they are copied out of different 112 - # directories in the kernel tree. Sure you COULd merge them, but it's a 111 + # these need to be separate variables because they are copied out of different 112 + # directories in the kernel tree. Sure you COULD merge them, but it's a 113 113 # cure-is-worse-than-disease situation. 114 114 zlib-decomp-$(CONFIG_KERNEL_GZIP) := decompress_inflate.c 115 115 zlib-$(CONFIG_KERNEL_GZIP) := inffast.c inflate.c inftrees.c
+1 -1
arch/powerpc/boot/dts/acadia.dts
··· 172 172 reg = <0xef602800 0x60>; 173 173 interrupt-parent = <&UIC0>; 174 174 interrupts = <0x4 0x4>; 175 - /* This thing is a bit weird. It has it's own UIC 175 + /* This thing is a bit weird. It has its own UIC 176 176 * that it uses to generate snapshot triggers. We 177 177 * don't really support this device yet, and it needs 178 178 * work to figure this out.
+1 -1
arch/powerpc/boot/main.c
··· 188 188 189 189 /* A buffer that may be edited by tools operating on a zImage binary so as to 190 190 * edit the command line passed to vmlinux (by setting /chosen/bootargs). 191 - * The buffer is put in it's own section so that tools may locate it easier. 191 + * The buffer is put in its own section so that tools may locate it easier. 192 192 */ 193 193 static char cmdline[BOOT_COMMAND_LINE_SIZE] 194 194 __attribute__((__section__("__builtin_cmdline")));
+1 -1
arch/powerpc/boot/ps3.c
··· 25 25 26 26 /* A buffer that may be edited by tools operating on a zImage binary so as to 27 27 * edit the command line passed to vmlinux (by setting /chosen/bootargs). 28 - * The buffer is put in it's own section so that tools may locate it easier. 28 + * The buffer is put in its own section so that tools may locate it easier. 29 29 */ 30 30 31 31 static char cmdline[BOOT_COMMAND_LINE_SIZE]
+1 -1
arch/powerpc/include/asm/io.h
··· 982 982 } 983 983 984 984 /* 985 - * 32 bits still uses virt_to_bus() for it's implementation of DMA 985 + * 32 bits still uses virt_to_bus() for its implementation of DMA 986 986 * mappings se we have to keep it defined here. We also have some old 987 987 * drivers (shame shame shame) that use bus_to_virt() and haven't been 988 988 * fixed yet so I need to define it here.
+2 -2
arch/powerpc/include/asm/opal-api.h
··· 1027 1027 * The host will pass on OPAL, a buffer of length OPAL_SYSEPOW_MAX 1028 1028 * with individual elements being 16 bits wide to fetch the system 1029 1029 * wide EPOW status. Each element in the buffer will contain the 1030 - * EPOW status in it's bit representation for a particular EPOW sub 1030 + * EPOW status in its bit representation for a particular EPOW sub 1031 1031 * class as defined here. So multiple detailed EPOW status bits 1032 1032 * specific for any sub class can be represented in a single buffer 1033 - * element as it's bit representation. 1033 + * element as its bit representation. 1034 1034 */ 1035 1035 1036 1036 /* System EPOW type */
+1 -1
arch/powerpc/include/asm/pmac_feature.h
··· 192 192 193 193 /* PMAC_FTR_BMAC_ENABLE (struct device_node* node, 0, int value) 194 194 * enable/disable the bmac (ethernet) cell of a mac-io ASIC, also drive 195 - * it's reset line 195 + * its reset line 196 196 */ 197 197 #define PMAC_FTR_BMAC_ENABLE PMAC_FTR_DEF(6) 198 198
+1 -1
arch/powerpc/include/asm/uninorth.h
··· 144 144 #define UNI_N_HWINIT_STATE_SLEEPING 0x01 145 145 #define UNI_N_HWINIT_STATE_RUNNING 0x02 146 146 /* This last bit appear to be used by the bootROM to know the second 147 - * CPU has started and will enter it's sleep loop with IP=0 147 + * CPU has started and will enter its sleep loop with IP=0 148 148 */ 149 149 #define UNI_N_HWINIT_STATE_CPU1_FLAG 0x10000000 150 150
+1 -1
arch/powerpc/include/uapi/asm/bootx.h
··· 108 108 /* ALL BELOW NEW (vers. 4) */ 109 109 110 110 /* This defines the physical memory. Valid with BOOT_ARCH_NUBUS flag 111 - (non-PCI) only. On PCI, memory is contiguous and it's size is in the 111 + (non-PCI) only. On PCI, memory is contiguous and its size is in the 112 112 device-tree. */ 113 113 boot_info_map_entry_t 114 114 physMemoryMap[MAX_MEM_MAP_SIZE]; /* Where the phys memory is */
+1 -1
arch/powerpc/kernel/eeh_pe.c
··· 527 527 * eeh_pe_mark_isolated 528 528 * @pe: EEH PE 529 529 * 530 - * Record that a PE has been isolated by marking the PE and it's children as 530 + * Record that a PE has been isolated by marking the PE and its children as 531 531 * EEH_PE_ISOLATED (and EEH_PE_CFG_BLOCKED, if required) and their PCI devices 532 532 * as pci_channel_io_frozen. 533 533 */
+1 -1
arch/powerpc/kernel/fadump.c
··· 681 681 * old_cpu == -1 means this is the first CPU which has come here, 682 682 * go ahead and trigger fadump. 683 683 * 684 - * old_cpu != -1 means some other CPU has already on it's way 684 + * old_cpu != -1 means some other CPU has already on its way 685 685 * to trigger fadump, just keep looping here. 686 686 */ 687 687 this_cpu = smp_processor_id();
+2 -2
arch/powerpc/kernel/misc_64.S
··· 192 192 xori r0,r0,MSR_EE 193 193 mtmsrd r0,1 194 194 195 - /* rotate 24 bits SCOM address 8 bits left and mask out it's low 8 bits 195 + /* rotate 24 bits SCOM address 8 bits left and mask out its low 8 bits 196 196 * (including parity). On current CPUs they must be 0'd, 197 197 * and finally or in RW bit 198 198 */ ··· 226 226 xori r0,r0,MSR_EE 227 227 mtmsrd r0,1 228 228 229 - /* rotate 24 bits SCOM address 8 bits left and mask out it's low 8 bits 229 + /* rotate 24 bits SCOM address 8 bits left and mask out its low 8 bits 230 230 * (including parity). On current CPUs they must be 0'd. 231 231 */ 232 232
+6 -6
arch/powerpc/kernel/process.c
··· 1661 1661 * cases will happen: 1662 1662 * 1663 1663 * 1. The correct thread is running, the wrong thread is not 1664 - * In this situation, the correct thread is woken and proceeds to pass it's 1664 + * In this situation, the correct thread is woken and proceeds to pass its 1665 1665 * condition check. 1666 1666 * 1667 1667 * 2. Neither threads are running ··· 1671 1671 * for the wrong thread, or they will execute the condition check immediately. 1672 1672 * 1673 1673 * 3. The wrong thread is running, the correct thread is not 1674 - * The wrong thread will be woken, but will fail it's condition check and 1674 + * The wrong thread will be woken, but will fail its condition check and 1675 1675 * re-execute wait. The correct thread, when scheduled, will execute either 1676 - * it's condition check (which will pass), or wait, which returns immediately 1677 - * when called the first time after the thread is scheduled, followed by it's 1676 + * its condition check (which will pass), or wait, which returns immediately 1677 + * when called the first time after the thread is scheduled, followed by its 1678 1678 * condition check (which will pass). 1679 1679 * 1680 1680 * 4. Both threads are running 1681 - * Both threads will be woken. The wrong thread will fail it's condition check 1682 - * and execute another wait, while the correct thread will pass it's condition 1681 + * Both threads will be woken. The wrong thread will fail its condition check 1682 + * and execute another wait, while the correct thread will pass its condition 1683 1683 * check. 1684 1684 * 1685 1685 * @t: the task to set the thread ID for
+1 -1
arch/powerpc/kernel/ptrace/ptrace-tm.c
··· 12 12 { 13 13 /* 14 14 * If task is not current, it will have been flushed already to 15 - * it's thread_struct during __switch_to(). 15 + * its thread_struct during __switch_to(). 16 16 * 17 17 * A reclaim flushes ALL the state or if not in TM save TM SPRs 18 18 * in the appropriate thread structures from live.
+1 -1
arch/powerpc/kernel/smp.c
··· 1567 1567 1568 1568 /* 1569 1569 * This CPU will not be in the online mask yet so we need to manually 1570 - * add it to it's own thread sibling mask. 1570 + * add it to its own thread sibling mask. 1571 1571 */ 1572 1572 map_cpu_to_node(cpu, cpu_to_node(cpu)); 1573 1573 cpumask_set_cpu(cpu, cpu_sibling_mask(cpu));
+2 -2
arch/powerpc/kernel/sysfs.c
··· 139 139 * @val: Returned cpu specific DSCR default value 140 140 * 141 141 * This function returns the per cpu DSCR default value 142 - * for any cpu which is contained in it's PACA structure. 142 + * for any cpu which is contained in its PACA structure. 143 143 */ 144 144 static void read_dscr(void *val) 145 145 { ··· 152 152 * @val: New cpu specific DSCR default value to update 153 153 * 154 154 * This function updates the per cpu DSCR default value 155 - * for any cpu which is contained in it's PACA structure. 155 + * for any cpu which is contained in its PACA structure. 156 156 */ 157 157 static void write_dscr(void *val) 158 158 {
+1 -1
arch/powerpc/kvm/book3s_xive.c
··· 531 531 xc->cppr = xive_prio_from_guest(new_cppr); 532 532 533 533 /* 534 - * IPIs are synthetized from MFRR and thus don't need 534 + * IPIs are synthesized from MFRR and thus don't need 535 535 * any special EOI handling. The underlying interrupt 536 536 * used to signal MFRR changes is EOId when fetched from 537 537 * the queue.
+1 -1
arch/powerpc/mm/cacheflush.c
··· 78 78 79 79 #ifdef CONFIG_HIGHMEM 80 80 /** 81 - * flush_dcache_icache_phys() - Flush a page by it's physical address 81 + * flush_dcache_icache_phys() - Flush a page by its physical address 82 82 * @physaddr: the physical address of the page 83 83 */ 84 84 static void flush_dcache_icache_phys(unsigned long physaddr)
+1 -1
arch/powerpc/mm/nohash/kaslr_booke.c
··· 376 376 create_kaslr_tlb_entry(1, tlb_virt, tlb_phys); 377 377 } 378 378 379 - /* Copy the kernel to it's new location and run */ 379 + /* Copy the kernel to its new location and run */ 380 380 memcpy((void *)kernstart_virt_addr, (void *)_stext, kernel_sz); 381 381 flush_icache_range(kernstart_virt_addr, kernstart_virt_addr + kernel_sz); 382 382
+1 -1
arch/powerpc/platforms/512x/mpc512x_shared.c
··· 279 279 * and so negatively affect boot time. Instead we reserve the 280 280 * already configured frame buffer area so that it won't be 281 281 * destroyed. The starting address of the area to reserve and 282 - * also it's length is passed to memblock_reserve(). It will be 282 + * also its length is passed to memblock_reserve(). It will be 283 283 * freed later on first open of fbdev, when splash image is not 284 284 * needed any more. 285 285 */
+1 -1
arch/powerpc/platforms/cell/spufs/sched.c
··· 868 868 } 869 869 870 870 /** 871 - * spu_deactivate - unbind a context from it's physical spu 871 + * spu_deactivate - unbind a context from its physical spu 872 872 * @ctx: spu context to unbind 873 873 * 874 874 * Unbind @ctx from the physical spu it is running on and schedule
+1 -1
arch/powerpc/platforms/maple/pci.c
··· 595 595 596 596 /* Probe root PCI hosts, that is on U3 the AGP host and the 597 597 * HyperTransport host. That one is actually "kept" around 598 - * and actually added last as it's resource management relies 598 + * and actually added last as its resource management relies 599 599 * on the AGP resources to have been setup first 600 600 */ 601 601 root = of_find_node_by_path("/");
+1 -1
arch/powerpc/platforms/powermac/pic.c
··· 2 2 /* 3 3 * Support for the interrupt controllers found on Power Macintosh, 4 4 * currently Apple's "Grand Central" interrupt controller in all 5 - * it's incarnations. OpenPIC support used on newer machines is 5 + * its incarnations. OpenPIC support used on newer machines is 6 6 * in a separate file 7 7 * 8 8 * Copyright (C) 1997 Paul Mackerras (paulus@samba.org)
+1 -1
arch/powerpc/platforms/powermac/sleep.S
··· 176 176 * memory location containing the PC to resume from 177 177 * at address 0. 178 178 * - On Core99, we must store the wakeup vector at 179 - * address 0x80 and eventually it's parameters 179 + * address 0x80 and eventually its parameters 180 180 * at address 0x84. I've have some trouble with those 181 181 * parameters however and I no longer use them. 182 182 */
+2 -2
arch/powerpc/platforms/powernv/pci-sriov.c
··· 238 238 } else if (pdev->is_physfn) { 239 239 /* 240 240 * For PFs adjust their allocated IOV resources to match what 241 - * the PHB can support using it's M64 BAR table. 241 + * the PHB can support using its M64 BAR table. 242 242 */ 243 243 pnv_pci_ioda_fixup_iov_resources(pdev); 244 244 } ··· 658 658 list_add_tail(&pe->list, &phb->ioda.pe_list); 659 659 mutex_unlock(&phb->ioda.pe_list_mutex); 660 660 661 - /* associate this pe to it's pdn */ 661 + /* associate this pe to its pdn */ 662 662 list_for_each_entry(vf_pdn, &pdn->parent->child_list, list) { 663 663 if (vf_pdn->busno == vf_bus && 664 664 vf_pdn->devfn == vf_devfn) {
+1 -1
arch/powerpc/platforms/powernv/vas-window.c
··· 1059 1059 } 1060 1060 } else { 1061 1061 /* 1062 - * Interrupt hanlder or fault window setup failed. Means 1062 + * Interrupt handler or fault window setup failed. Means 1063 1063 * NX can not generate fault for page fault. So not 1064 1064 * opening for user space tx window. 1065 1065 */
+1 -1
arch/powerpc/platforms/pseries/vas.c
··· 228 228 struct pseries_vas_window *txwin = data; 229 229 230 230 /* 231 - * The thread hanlder will process this interrupt if it is 231 + * The thread handler will process this interrupt if it is 232 232 * already running. 233 233 */ 234 234 atomic_inc(&txwin->pending_faults);
+2 -2
arch/powerpc/sysdev/xive/common.c
··· 383 383 * CPU. 384 384 * 385 385 * If we find that there is indeed more in there, we call 386 - * force_external_irq_replay() to make Linux synthetize an 386 + * force_external_irq_replay() to make Linux synthesize an 387 387 * external interrupt on the next call to local_irq_restore(). 388 388 */ 389 389 static void xive_do_queue_eoi(struct xive_cpu *xc) ··· 874 874 * 875 875 * This also tells us that it's in flight to a host queue 876 876 * or has already been fetched but hasn't been EOIed yet 877 - * by the host. This it's potentially using up a host 877 + * by the host. Thus it's potentially using up a host 878 878 * queue slot. This is important to know because as long 879 879 * as this is the case, we must not hard-unmask it when 880 880 * "returning" that interrupt to the host.
+1 -1
arch/powerpc/sysdev/xive/native.c
··· 415 415 return; 416 416 } 417 417 418 - /* Grab it's CAM value */ 418 + /* Grab its CAM value */ 419 419 rc = opal_xive_get_vp_info(vp, NULL, &vp_cam_be, NULL, NULL); 420 420 if (rc) { 421 421 pr_err("Failed to get pool VP info CPU %d\n", cpu);