Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

Merge branch 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/jikos/trivial

Pull trivial tree updates from Jiri Kosina.

* 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/jikos/trivial: (21 commits)
gitignore: fix wording
mfd: ab8500-debugfs: fix "between" in printk
memstick: trivial fix of spelling mistake on management
cpupowerutils: bench: fix "average"
treewide: Fix typos in printk
IB/mlx4: printk fix
pinctrl: sirf/atlas7: fix printk spelling
serial: mctrl_gpio: Grammar s/lines GPIOs/line GPIOs/, /sets/set/
w1: comment spelling s/minmum/minimum/
Blackfin: comment spelling s/divsor/divisor/
metag: Fix misspellings in comments.
ia64: Fix misspellings in comments.
hexagon: Fix misspellings in comments.
tools/perf: Fix misspellings in comments.
cris: Fix misspellings in comments.
c6x: Fix misspellings in comments.
blackfin: Fix misspelling of 'register' in comment.
avr32: Fix misspelling of 'definitions' in comment.
treewide: Fix typos in printk
Doc: treewide : Fix typos in DocBook/filesystem.xml
...

+86 -86
+1 -1
.gitignore
··· 62 62 /tar-install/ 63 63 64 64 # 65 - # git files that we don't want to ignore even it they are dot-files 65 + # git files that we don't want to ignore even if they are dot-files 66 66 # 67 67 !.gitignore 68 68 !.mailmap
+1 -1
arch/avr32/include/asm/addrspace.h
··· 1 1 /* 2 - * Defitions for the address spaces of the AVR32 CPUs. Heavily based on 2 + * Definitions for the address spaces of the AVR32 CPUs. Heavily based on 3 3 * include/asm-sh/addrspace.h 4 4 * 5 5 * Copyright (C) 2004-2006 Atmel Corporation
+1 -1
arch/blackfin/lib/udivsi3.S
··· 154 154 CC = R7 < 0; /* Check quotient(AQ) */ 155 155 /* If AQ==0, we'll sub divisor */ 156 156 IF CC R5 = R1; /* and if AQ==1, we'll add it. */ 157 - R3 = R3 + R5; /* Add/sub divsor to partial remainder */ 157 + R3 = R3 + R5; /* Add/sub divisor to partial remainder */ 158 158 R7 = R3 ^ R1; /* Generate next quotient bit */ 159 159 160 160 R5 = R7 >> 31; /* Get AQ */
+1 -1
arch/blackfin/mach-bf609/include/mach/defBF60x_base.h
··· 2689 2689 #define L2CTL0_STAT 0xFFCA3010 /* L2CTL0 L2 Status Register */ 2690 2690 #define L2CTL0_RPCR 0xFFCA3014 /* L2CTL0 L2 Read Priority Count Register */ 2691 2691 #define L2CTL0_WPCR 0xFFCA3018 /* L2CTL0 L2 Write Priority Count Register */ 2692 - #define L2CTL0_RFA 0xFFCA3024 /* L2CTL0 L2 Refresh Address Regsiter */ 2692 + #define L2CTL0_RFA 0xFFCA3024 /* L2CTL0 L2 Refresh Address Register */ 2693 2693 #define L2CTL0_ERRADDR0 0xFFCA3040 /* L2CTL0 L2 Bank 0 ECC Error Address Register */ 2694 2694 #define L2CTL0_ERRADDR1 0xFFCA3044 /* L2CTL0 L2 Bank 1 ECC Error Address Register */ 2695 2695 #define L2CTL0_ERRADDR2 0xFFCA3048 /* L2CTL0 L2 Bank 2 ECC Error Address Register */
+1 -1
arch/c6x/include/asm/clock.h
··· 101 101 #define CLK_PLL BIT(2) /* PLL-derived clock */ 102 102 #define PRE_PLL BIT(3) /* source is before PLL mult/div */ 103 103 #define FIXED_DIV_PLL BIT(4) /* fixed divisor from PLL */ 104 - #define FIXED_RATE_PLL BIT(5) /* fixed ouput rate PLL */ 104 + #define FIXED_RATE_PLL BIT(5) /* fixed output rate PLL */ 105 105 106 106 #define MAX_PLL_SYSCLKS 16 107 107
+1 -1
arch/c6x/platforms/cache.c
··· 145 145 spin_lock_irqsave(&cache_lock, flags); 146 146 147 147 /* 148 - * If another cache operation is occuring 148 + * If another cache operation is occurring 149 149 */ 150 150 if (unlikely(imcr_get(wc_reg))) { 151 151 spin_unlock_irqrestore(&cache_lock, flags);
+1 -1
arch/cris/arch-v10/drivers/axisflashmap.c
··· 212 212 /* 213 213 * Probe each chip select individually for flash chips. If there are chips on 214 214 * both cse0 and cse1, the mtd_info structs will be concatenated to one struct 215 - * so that MTD partitions can cross chip boundries. 215 + * so that MTD partitions can cross chip boundaries. 216 216 * 217 217 * The only known restriction to how you can mount your chips is that each 218 218 * chip select must hold similar flash chips. But you need external hardware
+1 -1
arch/cris/arch-v32/drivers/axisflashmap.c
··· 246 246 /* 247 247 * Probe each chip select individually for flash chips. If there are chips on 248 248 * both cse0 and cse1, the mtd_info structs will be concatenated to one struct 249 - * so that MTD partitions can cross chip boundries. 249 + * so that MTD partitions can cross chip boundaries. 250 250 * 251 251 * The only known restriction to how you can mount your chips is that each 252 252 * chip select must hold similar flash chips. But you need external hardware
+1 -1
arch/cris/arch-v32/drivers/cryptocop.c
··· 525 525 return 0; 526 526 } 527 527 528 - /* Map the ouput length of the transform to operation output starting on the inject index. */ 528 + /* Map the output length of the transform to operation output starting on the inject index. */ 529 529 static int create_input_descriptors(struct cryptocop_operation *operation, struct cryptocop_tfrm_ctx *tc, struct cryptocop_dma_desc **id, int alloc_flag) 530 530 { 531 531 int err = 0;
+1 -1
arch/cris/arch-v32/mach-a3/dram_init.S
··· 11 11 */ 12 12 13 13 /* Just to be certain the config file is included, we include it here 14 - * explicitely instead of depending on it being included in the file that 14 + * explicitly instead of depending on it being included in the file that 15 15 * uses this code. 16 16 */ 17 17
+1 -1
arch/cris/arch-v32/mach-fs/dram_init.S
··· 11 11 */ 12 12 13 13 /* Just to be certain the config file is included, we include it here 14 - * explicitely instead of depending on it being included in the file that 14 + * explicitly instead of depending on it being included in the file that 15 15 * uses this code. 16 16 */ 17 17
+1 -1
arch/hexagon/include/asm/hexagon_vm.h
··· 237 237 238 238 /* 239 239 * The initial program gets to find a system environment descriptor 240 - * on its stack when it begins exection. The first word is a version 240 + * on its stack when it begins execution. The first word is a version 241 241 * code to indicate what is there. Zero means nothing more. 242 242 */ 243 243
+1 -1
arch/hexagon/include/asm/vm_mmu.h
··· 78 78 #define __HEXAGON_C_WB_L2 0x7 /* Write-back, with L2 */ 79 79 80 80 /* 81 - * This can be overriden, but we're defaulting to the most aggressive 81 + * This can be overridden, but we're defaulting to the most aggressive 82 82 * cache policy, the better to find bugs sooner. 83 83 */ 84 84
+2 -2
arch/hexagon/kernel/kgdb.c
··· 236 236 }; 237 237 238 238 /** 239 - * kgdb_arch_init - Perform any architecture specific initalization. 239 + * kgdb_arch_init - Perform any architecture specific initialization. 240 240 * 241 - * This function will handle the initalization of any architecture 241 + * This function will handle the initialization of any architecture 242 242 * specific callbacks. 243 243 */ 244 244 int kgdb_arch_init(void)
+1 -1
arch/hexagon/kernel/vm_ops.S
··· 26 26 * could be, and perhaps some day will be, handled as in-line 27 27 * macros, but for tracing/debugging it's handy to have 28 28 * a single point of invocation for each of them. 29 - * Conveniently, they take paramters and return values 29 + * Conveniently, they take parameters and return values 30 30 * consistent with the ABI calling convention. 31 31 */ 32 32
+1 -1
arch/hexagon/lib/memcpy.S
··· 39 39 * DJH 10/14/09 Version 1.3 added special loop for aligned case, was 40 40 * overreading bloated codesize back up to 892 41 41 * DJH 4/20/10 Version 1.4 fixed Ldword_loop_epilog loop to prevent loads 42 - * occuring if only 1 left outstanding, fixes bug 42 + * occurring if only 1 left outstanding, fixes bug 43 43 * # 3888, corrected for all alignments. Peeled off 44 44 * 1 32byte chunk from kernel loop and extended 8byte 45 45 * loop at end to solve all combinations and prevent
+1 -1
arch/ia64/include/asm/sn/ioc3.h
··· 131 131 #define SSCR_PAUSE_STATE 0x40000000 /* set when PAUSE takes effect*/ 132 132 #define SSCR_RESET 0x80000000 /* reset DMA channels */ 133 133 134 - /* all producer/comsumer pointers are the same bitfield */ 134 + /* all producer/consumer pointers are the same bitfield */ 135 135 #define PROD_CONS_PTR_4K 0x00000ff8 /* for 4K buffers */ 136 136 #define PROD_CONS_PTR_1K 0x000003f8 /* for 1K buffers */ 137 137 #define PROD_CONS_PTR_OFF 3
+2 -2
arch/ia64/include/asm/sn/shubio.h
··· 1385 1385 * respones are captured until IXSS[VALID] is cleared by setting the * 1386 1386 * appropriate bit in IECLR. Every time a spurious read response is * 1387 1387 * detected, the SPUR_RD bit of the PRB corresponding to the incoming * 1388 - * message's SIDN field is set. This always happens, regarless of * 1388 + * message's SIDN field is set. This always happens, regardless of * 1389 1389 * whether a header is captured. The programmer should check * 1390 1390 * IXSM[SIDN] to determine which widget sent the spurious response, * 1391 1391 * because there may be more than one SPUR_RD bit set in the PRB * ··· 2997 2997 /* 2998 2998 * Values for field imsgtype 2999 2999 */ 3000 - #define IIO_ICRB_IMSGT_XTALK 0 /* Incoming Meessage from Xtalk */ 3000 + #define IIO_ICRB_IMSGT_XTALK 0 /* Incoming message from Xtalk */ 3001 3001 #define IIO_ICRB_IMSGT_BTE 1 /* Incoming message from BTE */ 3002 3002 #define IIO_ICRB_IMSGT_SN1NET 2 /* Incoming message from SN1 net */ 3003 3003 #define IIO_ICRB_IMSGT_CRB 3 /* Incoming message from CRB ??? */
+1 -1
arch/ia64/kernel/efi.c
··· 964 964 /* 965 965 * Look for the first granule aligned memory descriptor memory 966 966 * that is big enough to hold EFI memory map. Make sure this 967 - * descriptor is atleast granule sized so it does not get trimmed 967 + * descriptor is at least granule sized so it does not get trimmed 968 968 */ 969 969 struct kern_memdesc * 970 970 find_memmap_space (void)
+1 -1
arch/ia64/lib/idiv32.S
··· 11 11 * 12 12 * For more details on the theory behind these algorithms, see "IA-64 13 13 * and Elementary Functions" by Peter Markstein; HP Professional Books 14 - * (http://www.hp.com/go/retailbooks/) 14 + * (http://www.goodreads.com/book/show/2019887.Ia_64_and_Elementary_Functions) 15 15 */ 16 16 17 17 #include <asm/asmmacro.h>
+1 -1
arch/ia64/lib/idiv64.S
··· 11 11 * 12 12 * For more details on the theory behind these algorithms, see "IA-64 13 13 * and Elementary Functions" by Peter Markstein; HP Professional Books 14 - * (http://www.hp.com/go/retailbooks/) 14 + * (http://www.goodreads.com/book/show/2019887.Ia_64_and_Elementary_Functions) 15 15 */ 16 16 17 17 #include <asm/asmmacro.h>
+1 -1
arch/metag/include/asm/metag_regs.h
··· 1165 1165 #define TXSTATUS_IPTOGGLE_BIT 0x80000000 /* Prev PToggle of TXPRIVEXT */ 1166 1166 #define TXSTATUS_ISTATE_BIT 0x40000000 /* IState bit */ 1167 1167 #define TXSTATUS_IWAIT_BIT 0x20000000 /* wait indefinitely in decision step*/ 1168 - #define TXSTATUS_IEXCEPT_BIT 0x10000000 /* Indicate an exception occured */ 1168 + #define TXSTATUS_IEXCEPT_BIT 0x10000000 /* Indicate an exception occurred */ 1169 1169 #define TXSTATUS_IRPCOUNT_BITS 0x0E000000 /* Number of 'dirty' date entries*/ 1170 1170 #define TXSTATUS_IRPCOUNT_S 25 1171 1171 #define TXSTATUS_IRQSTAT_BITS 0x0000F000 /* IRQEnc bits, trigger or interrupts */
+3 -3
arch/metag/include/asm/tbx.h
··· 668 668 669 669 State.Sig.TrigMask will indicate the bits set within TXMASKI at 670 670 the time of the handler call that have all been cleared to prevent 671 - nested interrupt occuring immediately. 671 + nested interrupt occurring immediately. 672 672 673 673 State.Sig.SaveMask is a bit-mask which will be set to Zero when a trigger 674 674 occurs at background level and TBICTX_CRIT_BIT and optionally ··· 1083 1083 /* This routine causes the TBICTX structure specified in State.Sig.pCtx to 1084 1084 be restored. This implies that execution will not return to the caller. 1085 1085 The State.Sig.TrigMask field will be restored during the context switch 1086 - such that any immediately occuring interrupts occur in the context of the 1086 + such that any immediately occurring interrupts occur in the context of the 1087 1087 newly specified task. The State.Sig.SaveMask parameter is ignored. */ 1088 1088 void __TBIASyncResume( TBIRES State ); 1089 1089 ··· 1305 1305 1306 1306 /* 1307 1307 * Calculate linear PC value from real PC and Minim mode control, the LSB of 1308 - * the result returned indicates if address compression has occured. 1308 + * the result returned indicates if address compression has occurred. 1309 1309 */ 1310 1310 #ifndef __ASSEMBLY__ 1311 1311 #define METAG_LINPC( PCVal ) (\
+1 -1
arch/metag/tbx/tbipcx.S
··· 15 15 #include <asm/tbx.h> 16 16 17 17 /* BEGIN HACK */ 18 - /* define these for now while doing inital conversion to GAS 18 + /* define these for now while doing initial conversion to GAS 19 19 will fix properly later */ 20 20 21 21 /* Signal identifiers always have the TBID_SIGNAL_BIT set and contain the
+3 -3
arch/metag/tbx/tbisoft.S
··· 56 56 /* 57 57 * TBIRES __TBISwitch( TBIRES Switch, PTBICTX *rpSaveCtx ) 58 58 * 59 - * Software syncronous context switch between soft threads, save only the 59 + * Software synchronous context switch between soft threads, save only the 60 60 * registers which are actually valid on call entry. 61 61 * 62 62 * A0FrP, D0RtP, D0.5, D0.6, D0.7 - Saved on stack ··· 76 76 SETL [A0StP+#8++],D0FrT,D1RtP 77 77 /* 78 78 * Save current frame state - we save all regs because we don't want 79 - * uninitialised crap in the TBICTX structure that the asyncronous resumption 79 + * uninitialised crap in the TBICTX structure that the asynchronous resumption 80 80 * of a thread will restore. 81 81 */ 82 82 MOVT D1Re0,#HI($LSwitchExit) /* ASync resume point here */ ··· 117 117 * This routine causes the TBICTX structure specified in State.Sig.pCtx to 118 118 * be restored. This implies that execution will not return to the caller. 119 119 * The State.Sig.TrigMask field will be ored into TXMASKI during the 120 - * context switch such that any immediately occuring interrupts occur in 120 + * context switch such that any immediately occurring interrupts occur in 121 121 * the context of the newly specified task. The State.Sig.SaveMask parameter 122 122 * is ignored. 123 123 */
+1 -1
arch/powerpc/kernel/mce.c
··· 284 284 printk("%s Effective address: %016llx\n", 285 285 level, evt->u.ue_error.effective_address); 286 286 if (evt->u.ue_error.physical_address_provided) 287 - printk("%s Physial address: %016llx\n", 287 + printk("%s Physical address: %016llx\n", 288 288 level, evt->u.ue_error.physical_address); 289 289 break; 290 290 case MCE_ERROR_TYPE_SLB:
+1 -1
arch/x86/kernel/cpu/microcode/intel.c
··· 422 422 data_size = get_datasize(mc_saved_header); 423 423 date = mc_saved_header->date; 424 424 425 - pr_debug("mc_saved[%d]: sig=0x%x, pf=0x%x, rev=0x%x, toal size=0x%x, date = %04x-%02x-%02x\n", 425 + pr_debug("mc_saved[%d]: sig=0x%x, pf=0x%x, rev=0x%x, total size=0x%x, date = %04x-%02x-%02x\n", 426 426 i, sig, pf, rev, total_size, 427 427 date & 0xffff, 428 428 date >> 24,
+1 -1
arch/x86/kvm/iommu.c
··· 254 254 !iommu_capable(&pci_bus_type, IOMMU_CAP_INTR_REMAP)) { 255 255 printk(KERN_WARNING "%s: No interrupt remapping support," 256 256 " disallowing device assignment." 257 - " Re-enble with \"allow_unsafe_assigned_interrupts=1\"" 257 + " Re-enable with \"allow_unsafe_assigned_interrupts=1\"" 258 258 " module option.\n", __func__); 259 259 iommu_domain_free(kvm->arch.iommu_domain); 260 260 kvm->arch.iommu_domain = NULL;
+2 -2
block/partitions/efi.c
··· 430 430 } 431 431 /* Check that sizeof_partition_entry has the correct value */ 432 432 if (le32_to_cpu((*gpt)->sizeof_partition_entry) != sizeof(gpt_entry)) { 433 - pr_debug("GUID Partitition Entry Size check failed.\n"); 433 + pr_debug("GUID Partition Entry Size check failed.\n"); 434 434 goto fail; 435 435 } 436 436 ··· 443 443 le32_to_cpu((*gpt)->sizeof_partition_entry)); 444 444 445 445 if (crc != le32_to_cpu((*gpt)->partition_entry_array_crc32)) { 446 - pr_debug("GUID Partitition Entry Array CRC check failed.\n"); 446 + pr_debug("GUID Partition Entry Array CRC check failed.\n"); 447 447 goto fail_ptes; 448 448 } 449 449
+1 -1
drivers/clk/tegra/clk-tegra20.c
··· 623 623 case OSC_CTRL_PLL_REF_DIV_4: 624 624 return 4; 625 625 default: 626 - pr_err("Invalied pll ref divider %d\n", pll_ref_div); 626 + pr_err("Invalid pll ref divider %d\n", pll_ref_div); 627 627 BUG(); 628 628 } 629 629 return 0;
+1 -1
drivers/edac/amd64_edac.c
··· 640 640 input_addr = 641 641 dram_addr_to_input_addr(mci, sys_addr_to_dram_addr(mci, sys_addr)); 642 642 643 - edac_dbg(2, "SysAdddr 0x%lx translates to InputAddr 0x%lx\n", 643 + edac_dbg(2, "SysAddr 0x%lx translates to InputAddr 0x%lx\n", 644 644 (unsigned long)sys_addr, (unsigned long)input_addr); 645 645 646 646 return input_addr;
+2 -2
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
··· 796 796 debug.state[0] == debug.state[1]) { 797 797 seq_puts(m, "seems to be stuck\n"); 798 798 } else if (debug.address[0] == debug.address[1]) { 799 - seq_puts(m, "adress is constant\n"); 799 + seq_puts(m, "address is constant\n"); 800 800 } else { 801 - seq_puts(m, "is runing\n"); 801 + seq_puts(m, "is running\n"); 802 802 } 803 803 804 804 seq_printf(m, "\t address 0: 0x%08x\n", debug.address[0]);
+1 -1
drivers/infiniband/hw/cxgb3/iwch_cm.c
··· 367 367 */ 368 368 static void act_open_req_arp_failure(struct t3cdev *dev, struct sk_buff *skb) 369 369 { 370 - printk(KERN_ERR MOD "ARP failure duing connect\n"); 370 + printk(KERN_ERR MOD "ARP failure during connect\n"); 371 371 kfree_skb(skb); 372 372 } 373 373
+1 -1
drivers/infiniband/hw/mlx4/main.c
··· 1601 1601 else if (ret == -ENXIO) 1602 1602 pr_err("Device managed flow steering is disabled. Fail to register network rule.\n"); 1603 1603 else if (ret) 1604 - pr_err("Invalid argumant. Fail to register network rule.\n"); 1604 + pr_err("Invalid argument. Fail to register network rule.\n"); 1605 1605 1606 1606 mlx4_free_cmd_mailbox(mdev->dev, mailbox); 1607 1607 return ret;
+1 -1
drivers/input/touchscreen/cyttsp4_core.c
··· 1499 1499 1500 1500 if (IS_BOOTLOADER(mode[0], mode[1])) { 1501 1501 mutex_unlock(&cd->system_lock); 1502 - dev_err(cd->dev, "%s: Device in BOOTLADER mode.\n", __func__); 1502 + dev_err(cd->dev, "%s: Device in BOOTLOADER mode.\n", __func__); 1503 1503 rc = -EINVAL; 1504 1504 goto error; 1505 1505 }
+8 -8
drivers/memstick/core/ms_block.c
··· 1220 1220 } 1221 1221 1222 1222 if (extra.management_flag & MEMSTICK_MANAGEMENT_SYSFLG) { 1223 - dbg("managment flag doesn't indicate boot block %d", 1223 + dbg("management flag doesn't indicate boot block %d", 1224 1224 pba); 1225 1225 continue; 1226 1226 } ··· 1367 1367 static int msb_ftl_scan(struct msb_data *msb) 1368 1368 { 1369 1369 u16 pba, lba, other_block; 1370 - u8 overwrite_flag, managment_flag, other_overwrite_flag; 1370 + u8 overwrite_flag, management_flag, other_overwrite_flag; 1371 1371 int error; 1372 1372 struct ms_extra_data_register extra; 1373 1373 u8 *overwrite_flags = kzalloc(msb->block_count, GFP_KERNEL); ··· 1409 1409 } 1410 1410 1411 1411 lba = be16_to_cpu(extra.logical_address); 1412 - managment_flag = extra.management_flag; 1412 + management_flag = extra.management_flag; 1413 1413 overwrite_flag = extra.overwrite_flag; 1414 1414 overwrite_flags[pba] = overwrite_flag; 1415 1415 ··· 1421 1421 } 1422 1422 1423 1423 /* Skip system/drm blocks */ 1424 - if ((managment_flag & MEMSTICK_MANAGMENT_FLAG_NORMAL) != 1425 - MEMSTICK_MANAGMENT_FLAG_NORMAL) { 1426 - dbg("pba %05d -> [reserved managment flag %02x]", 1427 - pba, managment_flag); 1424 + if ((management_flag & MEMSTICK_MANAGEMENT_FLAG_NORMAL) != 1425 + MEMSTICK_MANAGEMENT_FLAG_NORMAL) { 1426 + dbg("pba %05d -> [reserved management flag %02x]", 1427 + pba, management_flag); 1428 1428 msb_mark_block_used(msb, pba); 1429 1429 continue; 1430 1430 } 1431 1431 1432 1432 /* Erase temporary tables */ 1433 - if (!(managment_flag & MEMSTICK_MANAGEMENT_ATFLG)) { 1433 + if (!(management_flag & MEMSTICK_MANAGEMENT_ATFLG)) { 1434 1434 dbg("pba %05d -> [temp table] - will erase", pba); 1435 1435 1436 1436 msb_mark_block_used(msb, pba);
+1 -1
drivers/memstick/core/ms_block.h
··· 47 47 #define MEMSTICK_OV_PG_NORMAL \ 48 48 (MEMSTICK_OVERWRITE_PGST1 | MEMSTICK_OVERWRITE_PGST0) 49 49 50 - #define MEMSTICK_MANAGMENT_FLAG_NORMAL \ 50 + #define MEMSTICK_MANAGEMENT_FLAG_NORMAL \ 51 51 (MEMSTICK_MANAGEMENT_SYSFLG | \ 52 52 MEMSTICK_MANAGEMENT_SCMS1 | \ 53 53 MEMSTICK_MANAGEMENT_SCMS0) \
+1 -1
drivers/mfd/ab8500-debugfs.c
··· 2563 2563 2564 2564 if (user_trig_timer & ~0xFF) { 2565 2565 dev_err(dev, 2566 - "debugfs error input: should be beetween 0 to 255\n"); 2566 + "debugfs error input: should be between 0 to 255\n"); 2567 2567 return -EINVAL; 2568 2568 } 2569 2569
+1 -1
drivers/mtd/sm_ftl.c
··· 386 386 if (test_bit(boffset / SM_SECTOR_SIZE, &invalid_bitmap)) { 387 387 388 388 sm_printk("sector %d of block at LBA %d of zone %d" 389 - " coudn't be read, marking it as invalid", 389 + " couldn't be read, marking it as invalid", 390 390 boffset / SM_SECTOR_SIZE, lba, zone); 391 391 392 392 oob.data_status = 0;
+1 -1
drivers/net/ethernet/broadcom/bgmac.c
··· 1515 1515 phy_dev = phy_connect(bgmac->net_dev, bus_id, &bgmac_adjust_link, 1516 1516 PHY_INTERFACE_MODE_MII); 1517 1517 if (IS_ERR(phy_dev)) { 1518 - bgmac_err(bgmac, "PHY connecton failed\n"); 1518 + bgmac_err(bgmac, "PHY connection failed\n"); 1519 1519 err = PTR_ERR(phy_dev); 1520 1520 goto err_unregister_bus; 1521 1521 }
+1 -1
drivers/net/ethernet/freescale/fman/fman.c
··· 2772 2772 /* Get the FM address */ 2773 2773 res = platform_get_resource(of_dev, IORESOURCE_MEM, 0); 2774 2774 if (!res) { 2775 - dev_err(&of_dev->dev, "%s: Can't get FMan memory resouce\n", 2775 + dev_err(&of_dev->dev, "%s: Can't get FMan memory resource\n", 2776 2776 __func__); 2777 2777 goto fman_node_put; 2778 2778 }
+2 -2
drivers/net/ethernet/mellanox/mlx5/core/health.c
··· 176 176 case MLX5_HEALTH_SYNDR_EQ_ERR: 177 177 return "EQ error"; 178 178 case MLX5_HEALTH_SYNDR_EQ_INV: 179 - return "Invalid EQ refrenced"; 179 + return "Invalid EQ referenced"; 180 180 case MLX5_HEALTH_SYNDR_FFSER_ERR: 181 181 return "FFSER error"; 182 182 case MLX5_HEALTH_SYNDR_HIGH_TEMP: 183 - return "High temprature"; 183 + return "High temperature"; 184 184 default: 185 185 return "unrecognized error"; 186 186 }
+1 -1
drivers/net/ethernet/mellanox/mlx5/core/sriov.c
··· 140 140 struct mlx5_core_sriov *sriov = &dev->priv.sriov; 141 141 int err; 142 142 143 - mlx5_core_dbg(dev, "requsted num_vfs %d\n", num_vfs); 143 + mlx5_core_dbg(dev, "requested num_vfs %d\n", num_vfs); 144 144 if (!mlx5_core_is_pf(dev)) 145 145 return -EPERM; 146 146
+2 -2
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
··· 2315 2315 2316 2316 pci_read_config_byte(rtlpci->pdev, 0x34, &cap_pointer); 2317 2317 RT_TRACE(rtlpriv, COMP_INIT, DBG_LOUD, 2318 - "PCI configration 0x34 = 0x%2x\n", cap_pointer); 2318 + "PCI configuration 0x34 = 0x%2x\n", cap_pointer); 2319 2319 2320 2320 do { 2321 2321 pci_read_config_word(rtlpci->pdev, cap_pointer, &cap_hdr); 2322 2322 cap_id = cap_hdr & 0xFF; 2323 2323 2324 2324 RT_TRACE(rtlpriv, COMP_INIT, DBG_LOUD, 2325 - "in pci configration, cap_pointer%x = %x\n", 2325 + "in pci configuration, cap_pointer%x = %x\n", 2326 2326 cap_pointer, cap_id); 2327 2327 2328 2328 if (cap_id == 0x01) {
+1 -1
drivers/pinctrl/sirf/pinctrl-atlas7.c
··· 5798 5798 5799 5799 status = readl(ATLAS7_GPIO_INT_STATUS(bank)); 5800 5800 if (!status) { 5801 - pr_warn("%s: gpio [%s] status %#x no interrupt is flaged\n", 5801 + pr_warn("%s: gpio [%s] status %#x no interrupt is flagged\n", 5802 5802 __func__, gc->label, status); 5803 5803 handle_bad_irq(desc); 5804 5804 return;
+3 -3
drivers/platform/mips/cpu_hwmon.c
··· 80 80 static ssize_t cpu0_temp_label(struct device *dev, 81 81 struct device_attribute *attr, char *buf) 82 82 { 83 - return sprintf(buf, "CPU 0 Temprature\n"); 83 + return sprintf(buf, "CPU 0 Temperature\n"); 84 84 } 85 85 86 86 static ssize_t cpu1_temp_label(struct device *dev, 87 87 struct device_attribute *attr, char *buf) 88 88 { 89 - return sprintf(buf, "CPU 1 Temprature\n"); 89 + return sprintf(buf, "CPU 1 Temperature\n"); 90 90 } 91 91 92 92 static ssize_t get_cpu0_temp(struct device *dev, ··· 169 169 170 170 ret = create_sysfs_cputemp_files(&cpu_hwmon_dev->kobj); 171 171 if (ret) { 172 - pr_err("fail to create cpu temprature interface!\n"); 172 + pr_err("fail to create cpu temperature interface!\n"); 173 173 goto fail_create_sysfs_cputemp_files; 174 174 } 175 175
+1 -1
drivers/scsi/aic94xx/aic94xx_hwi.c
··· 477 477 478 478 err = asd_start_seqs(asd_ha); 479 479 if (err) { 480 - asd_printk("coudln't start seqs for %s\n", 480 + asd_printk("couldn't start seqs for %s\n", 481 481 pci_name(asd_ha->pcidev)); 482 482 goto out; 483 483 }
+1 -1
drivers/scsi/aic94xx/aic94xx_seq.c
··· 1352 1352 for_each_sequencer(lseq_mask, lseq_mask, lseq) { 1353 1353 err = asd_seq_start_lseq(asd_ha, lseq); 1354 1354 if (err) { 1355 - asd_printk("coudln't start LSEQ %d for %s\n", lseq, 1355 + asd_printk("couldn't start LSEQ %d for %s\n", lseq, 1356 1356 pci_name(asd_ha->pcidev)); 1357 1357 return err; 1358 1358 }
+1 -1
drivers/scsi/isci/port.c
··· 794 794 * case stay in the stopped state. 795 795 */ 796 796 dev_err(sciport_to_dev(iport), 797 - "%s: SCIC Port 0x%p failed to stop before tiemout.\n", 797 + "%s: SCIC Port 0x%p failed to stop before timeout.\n", 798 798 __func__, 799 799 iport); 800 800 } else if (current_state == SCI_PORT_STOPPING) {
+1 -1
drivers/scsi/qla2xxx/qla_sup.c
··· 3222 3222 ha->fcp_prio_cfg = vmalloc(FCP_PRIO_CFG_SIZE); 3223 3223 if (!ha->fcp_prio_cfg) { 3224 3224 ql_log(ql_log_warn, vha, 0x00d5, 3225 - "Unable to allocate memory for fcp priorty data (%x).\n", 3225 + "Unable to allocate memory for fcp priority data (%x).\n", 3226 3226 FCP_PRIO_CFG_SIZE); 3227 3227 return QLA_FUNCTION_FAILED; 3228 3228 }
+1 -1
drivers/target/iscsi/iscsi_target_auth.c
··· 293 293 pr_debug("[server] MD5 Digests do not match!\n\n"); 294 294 goto out; 295 295 } else 296 - pr_debug("[server] MD5 Digests match, CHAP connetication" 296 + pr_debug("[server] MD5 Digests match, CHAP connection" 297 297 " successful.\n\n"); 298 298 /* 299 299 * One way authentication has succeeded, return now if mutual
+2 -2
drivers/tty/serial/serial_mctrl_gpio.h
··· 62 62 enum mctrl_gpio_idx gidx); 63 63 64 64 /* 65 - * Request and set direction of modem control lines GPIOs and sets up irq 65 + * Request and set direction of modem control line GPIOs and set up irq 66 66 * handling. 67 67 * devm_* functions are used, so there's no need to call mctrl_gpio_free(). 68 68 * Returns a pointer to the allocated mctrl structure if ok, -ENOMEM on ··· 71 71 struct mctrl_gpios *mctrl_gpio_init(struct uart_port *port, unsigned int idx); 72 72 73 73 /* 74 - * Request and set direction of modem control lines GPIOs. 74 + * Request and set direction of modem control line GPIOs. 75 75 * devm_* functions are used, so there's no need to call mctrl_gpio_free(). 76 76 * Returns a pointer to the allocated mctrl structure if ok, -ENOMEM on 77 77 * allocation error.
+1 -1
drivers/w1/w1_io.c
··· 352 352 w1_delay(70); 353 353 354 354 result = dev->bus_master->read_bit(dev->bus_master->data) & 0x1; 355 - /* minmum 70 (above) + 430 = 500 us 355 + /* minimum 70 (above) + 430 = 500 us 356 356 * There aren't any timing requirements between a reset and 357 357 * the following transactions. Sleeping is safe here. 358 358 */
+1 -1
drivers/watchdog/watchdog_core.c
··· 262 262 263 263 ret = register_restart_handler(&wdd->restart_nb); 264 264 if (ret) 265 - pr_warn("watchog%d: Cannot register restart handler (%d)\n", 265 + pr_warn("watchdog%d: Cannot register restart handler (%d)\n", 266 266 wdd->id, ret); 267 267 } 268 268
+1 -1
fs/exofs/super.c
··· 122 122 if (match_int(&args[0], &option)) 123 123 return -EINVAL; 124 124 if (option <= 0) { 125 - EXOFS_ERR("Timout must be > 0"); 125 + EXOFS_ERR("Timeout must be > 0"); 126 126 return -EINVAL; 127 127 } 128 128 opts->timeout = option * HZ;
+1 -1
fs/jbd2/recovery.c
··· 303 303 * Locate any valid recovery information from the journal and set up the 304 304 * journal structures in memory to ignore it (presumably because the 305 305 * caller has evidence that it is out of date). 306 - * This function does'nt appear to be exorted.. 306 + * This function doesn't appear to be exported.. 307 307 * 308 308 * We perform one pass over the journal to allow us to tell the user how 309 309 * much recovery information is being erased, and to let us initialise
+3 -3
fs/jbd2/transaction.c
··· 543 543 * 544 544 * Some transactions, such as large extends and truncates, can be done 545 545 * atomically all at once or in several stages. The operation requests 546 - * a credit for a number of buffer modications in advance, but can 546 + * a credit for a number of buffer modifications in advance, but can 547 547 * extend its credit if it needs more. 548 548 * 549 549 * jbd2_journal_extend tries to give the running handle more buffer credits. ··· 627 627 * If the jbd2_journal_extend() call above fails to grant new buffer credits 628 628 * to a running handle, a call to jbd2_journal_restart will commit the 629 629 * handle's transaction so far and reattach the handle to a new 630 - * transaction capabable of guaranteeing the requested number of 630 + * transaction capable of guaranteeing the requested number of 631 631 * credits. We preserve reserved handle if there's any attached to the 632 632 * passed in handle. 633 633 */ ··· 1586 1586 1587 1587 /** 1588 1588 * int jbd2_journal_stop() - complete a transaction 1589 - * @handle: tranaction to complete. 1589 + * @handle: transaction to complete. 1590 1590 * 1591 1591 * All done for a particular handle. 1592 1592 *
+1 -1
fs/super.c
··· 285 285 * deactivate_locked_super - drop an active reference to superblock 286 286 * @s: superblock to deactivate 287 287 * 288 - * Drops an active reference to superblock, converting it into a temprory 288 + * Drops an active reference to superblock, converting it into a temporary 289 289 * one if there is no other active references left. In that case we 290 290 * tell fs driver to shut it down and drop the temporary reference we 291 291 * had just acquired.
+1 -1
net/tipc/socket.c
··· 389 389 sock->state = state; 390 390 sock_init_data(sock, sk); 391 391 if (tipc_sk_insert(tsk)) { 392 - pr_warn("Socket create failed; port numbrer exhausted\n"); 392 + pr_warn("Socket create failed; port number exhausted\n"); 393 393 return -EINVAL; 394 394 } 395 395 msg_set_origport(msg, tsk->portid);
+1 -1
tools/perf/tests/openat-syscall-all-cpus.c
··· 73 73 } 74 74 75 75 /* 76 - * Here we need to explicitely preallocate the counts, as if 76 + * Here we need to explicitly preallocate the counts, as if 77 77 * we use the auto allocation it will allocate just for 1 cpu, 78 78 * as we start by cpu 0. 79 79 */
+1 -1
tools/perf/util/evsel.c
··· 827 827 perf_evsel__set_sample_bit(evsel, PERIOD); 828 828 829 829 /* 830 - * When the user explicitely disabled time don't force it here. 830 + * When the user explicitly disabled time don't force it here. 831 831 */ 832 832 if (opts->sample_time && 833 833 (!perf_missing_features.sample_id_all &&
+1 -1
tools/perf/util/machine.c
··· 362 362 } 363 363 364 364 /* 365 - * Caller must eventually drop thread->refcnt returned with a successfull 365 + * Caller must eventually drop thread->refcnt returned with a successful 366 366 * lookup/new thread inserted. 367 367 */ 368 368 static struct thread *____machine__findnew_thread(struct machine *machine,
+1 -1
tools/perf/util/parse-events.c
··· 1649 1649 1650 1650 buf = _buf; 1651 1651 1652 - /* We're cutting from the beggining. */ 1652 + /* We're cutting from the beginning. */ 1653 1653 if (err->idx > max_err_idx) 1654 1654 cut = err->idx - max_err_idx; 1655 1655
+1 -1
tools/perf/util/session.c
··· 557 557 558 558 /* 559 559 * XXX this is hack in attempt to carry flags bitfield 560 - * throught endian village. ABI says: 560 + * through endian village. ABI says: 561 561 * 562 562 * Bit-fields are allocated from right to left (least to most significant) 563 563 * on little-endian implementations and from left to right (most to least