Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

Merge branch 'x86-fixes-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip

* 'x86-fixes-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip:
x86: Flush TLB if PGD entry is changed in i386 PAE mode
x86, dumpstack: Correct stack dump info when frame pointer is available
x86: Clean up csum-copy_64.S a bit
x86: Fix common misspellings
x86: Fix misspelling and align params
x86: Use PentiumPro-optimized partial_csum() on VIA C7

+235 -238
+7 -7
Documentation/x86/x86_64/boot-options.txt
··· 189 189 190 190 PCI 191 191 192 - pci=off Don't use PCI 193 - pci=conf1 Use conf1 access. 194 - pci=conf2 Use conf2 access. 195 - pci=rom Assign ROMs. 196 - pci=assign-busses Assign busses 197 - pci=irqmask=MASK Set PCI interrupt mask to MASK 198 - pci=lastbus=NUMBER Scan upto NUMBER busses, no matter what the mptable says. 192 + pci=off Don't use PCI 193 + pci=conf1 Use conf1 access. 194 + pci=conf2 Use conf2 access. 195 + pci=rom Assign ROMs. 196 + pci=assign-busses Assign busses 197 + pci=irqmask=MASK Set PCI interrupt mask to MASK 198 + pci=lastbus=NUMBER Scan up to NUMBER busses, no matter what the mptable says. 199 199 pci=noacpi Don't use ACPI to set up PCI interrupt routing. 200 200 201 201 IOMMU (input/output memory management unit)
+2 -2
arch/x86/Kconfig.cpu
··· 326 326 Old PentiumPro multiprocessor systems had errata that could cause 327 327 memory operations to violate the x86 ordering standard in rare cases. 328 328 Enabling this option will attempt to work around some (but not all) 329 - occurances of this problem, at the cost of much heavier spinlock and 329 + occurrences of this problem, at the cost of much heavier spinlock and 330 330 memory barrier operations. 331 331 332 332 If unsure, say n here. Even distro kernels should think twice before ··· 366 366 367 367 config X86_USE_PPRO_CHECKSUM 368 368 def_bool y 369 - depends on MWINCHIP3D || MWINCHIPC6 || MCYRIXIII || MK7 || MK6 || MPENTIUM4 || MPENTIUMM || MPENTIUMIII || MPENTIUMII || M686 || MK8 || MVIAC3_2 || MEFFICEON || MGEODE_LX || MCORE2 || MATOM 369 + depends on MWINCHIP3D || MWINCHIPC6 || MCYRIXIII || MK7 || MK6 || MPENTIUM4 || MPENTIUMM || MPENTIUMIII || MPENTIUMII || M686 || MK8 || MVIAC3_2 || MVIAC7 || MEFFICEON || MGEODE_LX || MCORE2 || MATOM 370 370 371 371 config X86_USE_3DNOW 372 372 def_bool y
+3 -3
arch/x86/crypto/aesni-intel_asm.S
··· 1346 1346 and $15, %r13 # %r13 = arg4 (mod 16) 1347 1347 je _multiple_of_16_bytes_decrypt 1348 1348 1349 - # Handle the last <16 byte block seperately 1349 + # Handle the last <16 byte block separately 1350 1350 1351 1351 paddd ONE(%rip), %xmm0 # increment CNT to get Yn 1352 1352 movdqa SHUF_MASK(%rip), %xmm10 ··· 1355 1355 ENCRYPT_SINGLE_BLOCK %xmm0, %xmm1 # E(K, Yn) 1356 1356 sub $16, %r11 1357 1357 add %r13, %r11 1358 - movdqu (%arg3,%r11,1), %xmm1 # recieve the last <16 byte block 1358 + movdqu (%arg3,%r11,1), %xmm1 # receive the last <16 byte block 1359 1359 lea SHIFT_MASK+16(%rip), %r12 1360 1360 sub %r13, %r12 1361 1361 # adjust the shuffle mask pointer to be able to shift 16-%r13 bytes ··· 1607 1607 and $15, %r13 # %r13 = arg4 (mod 16) 1608 1608 je _multiple_of_16_bytes_encrypt 1609 1609 1610 - # Handle the last <16 Byte block seperately 1610 + # Handle the last <16 Byte block separately 1611 1611 paddd ONE(%rip), %xmm0 # INCR CNT to get Yn 1612 1612 movdqa SHUF_MASK(%rip), %xmm10 1613 1613 PSHUFB_XMM %xmm10, %xmm0
+1 -1
arch/x86/include/asm/cacheflush.h
··· 71 71 * Read/Write : ReadOnly, ReadWrite 72 72 * Presence : NotPresent 73 73 * 74 - * Within a catagory, the attributes are mutually exclusive. 74 + * Within a category, the attributes are mutually exclusive. 75 75 * 76 76 * The implementation of this API will take care of various aspects that 77 77 * are associated with changing such attributes, such as:
+1 -1
arch/x86/include/asm/kdebug.h
··· 26 26 extern int __must_check __die(const char *, struct pt_regs *, long); 27 27 extern void show_registers(struct pt_regs *regs); 28 28 extern void show_trace(struct task_struct *t, struct pt_regs *regs, 29 - unsigned long *sp); 29 + unsigned long *sp, unsigned long bp); 30 30 extern void __show_regs(struct pt_regs *regs, int all); 31 31 extern void show_regs(struct pt_regs *regs); 32 32 extern unsigned long oops_begin(void);
+2 -2
arch/x86/include/asm/nmi.h
··· 29 29 * external nmis, because the local ones are more frequent. 30 30 * 31 31 * Also setup some default high/normal/low settings for 32 - * subsystems to registers with. Using 4 bits to seperate 33 - * the priorities. This can go alot higher if needed be. 32 + * subsystems to registers with. Using 4 bits to separate 33 + * the priorities. This can go a lot higher if needed be. 34 34 */ 35 35 36 36 #define NMI_LOCAL_SHIFT 16 /* randomly picked */
+1 -1
arch/x86/include/asm/nops.h
··· 38 38 #define K8_NOP8 K8_NOP4 K8_NOP4 39 39 40 40 /* K7 nops 41 - uses eax dependencies (arbitary choice) 41 + uses eax dependencies (arbitrary choice) 42 42 1: nop 43 43 2: movl %eax,%eax 44 44 3: leal (,%eax,1),%eax
+1 -1
arch/x86/include/asm/olpc.h
··· 20 20 21 21 /* 22 22 * OLPC board IDs contain the major build number within the mask 0x0ff0, 23 - * and the minor build number withing 0x000f. Pre-builds have a minor 23 + * and the minor build number within 0x000f. Pre-builds have a minor 24 24 * number less than 8, and normal builds start at 8. For example, 0x0B10 25 25 * is a PreB1, and 0x0C18 is a C1. 26 26 */
+2 -2
arch/x86/include/asm/perf_event_p4.h
··· 1 1 /* 2 - * Netburst Perfomance Events (P4, old Xeon) 2 + * Netburst Performance Events (P4, old Xeon) 3 3 */ 4 4 5 5 #ifndef PERF_EVENT_P4_H ··· 9 9 #include <linux/bitops.h> 10 10 11 11 /* 12 - * NetBurst has perfomance MSRs shared between 12 + * NetBurst has performance MSRs shared between 13 13 * threads if HT is turned on, ie for both logical 14 14 * processors (mem: in turn in Atom with HT support 15 15 * perf-MSRs are not shared and every thread has its
+3 -8
arch/x86/include/asm/pgtable-3level.h
··· 69 69 70 70 static inline void pud_clear(pud_t *pudp) 71 71 { 72 - unsigned long pgd; 73 - 74 72 set_pud(pudp, __pud(0)); 75 73 76 74 /* ··· 77 79 * section 8.1: in PAE mode we explicitly have to flush the 78 80 * TLB via cr3 if the top-level pgd is changed... 79 81 * 80 - * Make sure the pud entry we're updating is within the 81 - * current pgd to avoid unnecessary TLB flushes. 82 + * Currently all places where pud_clear() is called either have 83 + * flush_tlb_mm() followed or don't need TLB flush (x86_64 code or 84 + * pud_clear_bad()), so we don't need TLB flush here. 82 85 */ 83 - pgd = read_cr3(); 84 - if (__pa(pudp) >= pgd && __pa(pudp) < 85 - (pgd + sizeof(pgd_t)*PTRS_PER_PGD)) 86 - write_cr3(pgd); 87 86 } 88 87 89 88 #ifdef CONFIG_SMP
+1 -1
arch/x86/include/asm/processor-flags.h
··· 7 7 */ 8 8 #define X86_EFLAGS_CF 0x00000001 /* Carry Flag */ 9 9 #define X86_EFLAGS_PF 0x00000004 /* Parity Flag */ 10 - #define X86_EFLAGS_AF 0x00000010 /* Auxillary carry Flag */ 10 + #define X86_EFLAGS_AF 0x00000010 /* Auxiliary carry Flag */ 11 11 #define X86_EFLAGS_ZF 0x00000040 /* Zero Flag */ 12 12 #define X86_EFLAGS_SF 0x00000080 /* Sign Flag */ 13 13 #define X86_EFLAGS_TF 0x00000100 /* Trap Flag */
+1 -1
arch/x86/include/asm/ptrace-abi.h
··· 31 31 #define R12 24 32 32 #define RBP 32 33 33 #define RBX 40 34 - /* arguments: interrupts/non tracing syscalls only save upto here*/ 34 + /* arguments: interrupts/non tracing syscalls only save up to here*/ 35 35 #define R11 48 36 36 #define R10 56 37 37 #define R9 64
+2 -2
arch/x86/include/asm/ptrace.h
··· 73 73 unsigned long r12; 74 74 unsigned long rbp; 75 75 unsigned long rbx; 76 - /* arguments: non interrupts/non tracing syscalls only save upto here*/ 76 + /* arguments: non interrupts/non tracing syscalls only save up to here*/ 77 77 unsigned long r11; 78 78 unsigned long r10; 79 79 unsigned long r9; ··· 103 103 unsigned long r12; 104 104 unsigned long bp; 105 105 unsigned long bx; 106 - /* arguments: non interrupts/non tracing syscalls only save upto here*/ 106 + /* arguments: non interrupts/non tracing syscalls only save up to here*/ 107 107 unsigned long r11; 108 108 unsigned long r10; 109 109 unsigned long r9;
+3 -3
arch/x86/include/asm/stacktrace.h
··· 47 47 }; 48 48 49 49 void dump_trace(struct task_struct *tsk, struct pt_regs *regs, 50 - unsigned long *stack, 50 + unsigned long *stack, unsigned long bp, 51 51 const struct stacktrace_ops *ops, void *data); 52 52 53 53 #ifdef CONFIG_X86_32 ··· 86 86 87 87 extern void 88 88 show_trace_log_lvl(struct task_struct *task, struct pt_regs *regs, 89 - unsigned long *stack, char *log_lvl); 89 + unsigned long *stack, unsigned long bp, char *log_lvl); 90 90 91 91 extern void 92 92 show_stack_log_lvl(struct task_struct *task, struct pt_regs *regs, 93 - unsigned long *sp, char *log_lvl); 93 + unsigned long *sp, unsigned long bp, char *log_lvl); 94 94 95 95 extern unsigned int code_bytes; 96 96
+1 -1
arch/x86/include/asm/tsc.h
··· 35 35 static __always_inline cycles_t vget_cycles(void) 36 36 { 37 37 /* 38 - * We only do VDSOs on TSC capable CPUs, so this shouldnt 38 + * We only do VDSOs on TSC capable CPUs, so this shouldn't 39 39 * access boot_cpu_data (which is not VDSO-safe): 40 40 */ 41 41 #ifndef CONFIG_X86_TSC
+1 -1
arch/x86/include/asm/xen/interface.h
··· 86 86 * The privilege level specifies which modes may enter a trap via a software 87 87 * interrupt. On x86/64, since rings 1 and 2 are unavailable, we allocate 88 88 * privilege levels as follows: 89 - * Level == 0: Noone may enter 89 + * Level == 0: No one may enter 90 90 * Level == 1: Kernel may enter 91 91 * Level == 2: Kernel may enter 92 92 * Level == 3: Everyone may enter
+1 -1
arch/x86/kernel/alternative.c
··· 199 199 200 200 /* Replace instructions with better alternatives for this CPU type. 201 201 This runs before SMP is initialized to avoid SMP problems with 202 - self modifying code. This implies that assymetric systems where 202 + self modifying code. This implies that asymmetric systems where 203 203 APs have less capabilities than the boot processor are not handled. 204 204 Tough. Make sure you disable such features by hand. */ 205 205
+1 -1
arch/x86/kernel/aperture_64.c
··· 73 73 /* 74 74 * using 512M as goal, in case kexec will load kernel_big 75 75 * that will do the on position decompress, and could overlap with 76 - * that positon with gart that is used. 76 + * that position with gart that is used. 77 77 * sequende: 78 78 * kernel_small 79 79 * ==> kexec (with kdump trigger path or previous doesn't shutdown gart)
+2 -2
arch/x86/kernel/apic/io_apic.c
··· 1886 1886 * 1887 1887 * With interrupt-remapping, for now we will use virtual wire A mode, 1888 1888 * as virtual wire B is little complex (need to configure both 1889 - * IOAPIC RTE aswell as interrupt-remapping table entry). 1889 + * IOAPIC RTE as well as interrupt-remapping table entry). 1890 1890 * As this gets called during crash dump, keep this simple for now. 1891 1891 */ 1892 1892 if (ioapic_i8259.pin != -1 && !intr_remapping_enabled) { ··· 2905 2905 } 2906 2906 2907 2907 /* 2908 - * Called after all the initialization is done. If we didnt find any 2908 + * Called after all the initialization is done. If we didn't find any 2909 2909 * APIC bugs then we can allow the modify fast path 2910 2910 */ 2911 2911
+1 -1
arch/x86/kernel/apm_32.c
··· 66 66 * 1.5: Fix segment register reloading (in case of bad segments saved 67 67 * across BIOS call). 68 68 * Stephen Rothwell 69 - * 1.6: Cope with complier/assembler differences. 69 + * 1.6: Cope with compiler/assembler differences. 70 70 * Only try to turn off the first display device. 71 71 * Fix OOPS at power off with no APM BIOS by Jan Echternach 72 72 * <echter@informatik.uni-rostock.de>
+2 -2
arch/x86/kernel/cpu/cpufreq/longhaul.c
··· 444 444 return -EINVAL; 445 445 } 446 446 /* Get max multiplier - as we always did. 447 - * Longhaul MSR is usefull only when voltage scaling is enabled. 447 + * Longhaul MSR is useful only when voltage scaling is enabled. 448 448 * C3 is booting at max anyway. */ 449 449 maxmult = mult; 450 450 /* Get min multiplier */ ··· 1011 1011 * trigger frequency transition in some cases. */ 1012 1012 module_param(disable_acpi_c3, int, 0644); 1013 1013 MODULE_PARM_DESC(disable_acpi_c3, "Don't use ACPI C3 support"); 1014 - /* Change CPU voltage with frequency. Very usefull to save 1014 + /* Change CPU voltage with frequency. Very useful to save 1015 1015 * power, but most VIA C3 processors aren't supporting it. */ 1016 1016 module_param(scale_voltage, int, 0644); 1017 1017 MODULE_PARM_DESC(scale_voltage, "Scale voltage of processor");
+1 -1
arch/x86/kernel/cpu/cpufreq/powernow-k8.c
··· 1275 1275 1276 1276 if (powernow_k8_cpu_init_acpi(data)) { 1277 1277 /* 1278 - * Use the PSB BIOS structure. This is only availabe on 1278 + * Use the PSB BIOS structure. This is only available on 1279 1279 * an UP version, and is deprecated by AMD. 1280 1280 */ 1281 1281 if (num_online_cpus() != 1) {
+2 -2
arch/x86/kernel/cpu/cpufreq/speedstep-smi.c
··· 292 292 293 293 result = speedstep_smi_ownership(); 294 294 if (result) { 295 - dprintk("fails in aquiring ownership of a SMI interface.\n"); 295 + dprintk("fails in acquiring ownership of a SMI interface.\n"); 296 296 return -EINVAL; 297 297 } 298 298 ··· 360 360 int result = speedstep_smi_ownership(); 361 361 362 362 if (result) 363 - dprintk("fails in re-aquiring ownership of a SMI interface.\n"); 363 + dprintk("fails in re-acquiring ownership of a SMI interface.\n"); 364 364 365 365 return result; 366 366 }
+1 -1
arch/x86/kernel/cpu/mcheck/mce-inject.c
··· 32 32 { 33 33 struct mce *i = &per_cpu(injectm, m->extcpu); 34 34 35 - /* Make sure noone reads partially written injectm */ 35 + /* Make sure no one reads partially written injectm */ 36 36 i->finished = 0; 37 37 mb(); 38 38 m->finished = 0;
+1 -1
arch/x86/kernel/cpu/mcheck/mce.c
··· 881 881 * Check if the address reported by the CPU is in a format we can parse. 882 882 * It would be possible to add code for most other cases, but all would 883 883 * be somewhat complicated (e.g. segment offset would require an instruction 884 - * parser). So only support physical addresses upto page granuality for now. 884 + * parser). So only support physical addresses up to page granuality for now. 885 885 */ 886 886 static int mce_usable_address(struct mce *m) 887 887 {
+1 -1
arch/x86/kernel/cpu/mtrr/generic.c
··· 1 1 /* 2 2 * This only handles 32bit MTRR on 32bit hosts. This is strictly wrong 3 - * because MTRRs can span upto 40 bits (36bits on most modern x86) 3 + * because MTRRs can span up to 40 bits (36bits on most modern x86) 4 4 */ 5 5 #define DEBUG 6 6
+2 -2
arch/x86/kernel/cpu/perf_event.c
··· 1109 1109 1110 1110 /* 1111 1111 * If group events scheduling transaction was started, 1112 - * skip the schedulability test here, it will be peformed 1112 + * skip the schedulability test here, it will be performed 1113 1113 * at commit time (->commit_txn) as a whole 1114 1114 */ 1115 1115 if (cpuc->group_flag & PERF_EVENT_TXN) ··· 1790 1790 1791 1791 perf_callchain_store(entry, regs->ip); 1792 1792 1793 - dump_trace(NULL, regs, NULL, &backtrace_ops, entry); 1793 + dump_trace(NULL, regs, NULL, 0, &backtrace_ops, entry); 1794 1794 } 1795 1795 1796 1796 #ifdef CONFIG_COMPAT
+4 -4
arch/x86/kernel/cpu/perf_event_p4.c
··· 1 1 /* 2 - * Netburst Perfomance Events (P4, old Xeon) 2 + * Netburst Performance Events (P4, old Xeon) 3 3 * 4 4 * Copyright (C) 2010 Parallels, Inc., Cyrill Gorcunov <gorcunov@openvz.org> 5 5 * Copyright (C) 2010 Intel Corporation, Lin Ming <ming.m.lin@intel.com> ··· 679 679 */ 680 680 681 681 /* 682 - * if an event is shared accross the logical threads 682 + * if an event is shared across the logical threads 683 683 * the user needs special permissions to be able to use it 684 684 */ 685 685 if (p4_ht_active() && p4_event_bind_map[v].shared) { ··· 790 790 * 791 791 * It's still allowed that two threads setup same cache 792 792 * events so we can't simply clear metrics until we knew 793 - * noone is depending on us, so we need kind of counter 793 + * no one is depending on us, so we need kind of counter 794 794 * for "ReplayEvent" users. 795 795 * 796 796 * What is more complex -- RAW events, if user (for some 797 797 * reason) will pass some cache event metric with improper 798 798 * event opcode -- it's fine from hardware point of view 799 - * but completely nonsence from "meaning" of such action. 799 + * but completely nonsense from "meaning" of such action. 800 800 * 801 801 * So at moment let leave metrics turned on forever -- it's 802 802 * ok for now but need to be revisited!
+1 -1
arch/x86/kernel/cpu/vmware.c
··· 86 86 } 87 87 88 88 /* 89 - * While checking the dmi string infomation, just checking the product 89 + * While checking the dmi string information, just checking the product 90 90 * serial key should be enough, as this will always have a VMware 91 91 * specific string when running under VMware hypervisor. 92 92 */
+8 -6
arch/x86/kernel/dumpstack.c
··· 175 175 176 176 void 177 177 show_trace_log_lvl(struct task_struct *task, struct pt_regs *regs, 178 - unsigned long *stack, char *log_lvl) 178 + unsigned long *stack, unsigned long bp, char *log_lvl) 179 179 { 180 180 printk("%sCall Trace:\n", log_lvl); 181 - dump_trace(task, regs, stack, &print_trace_ops, log_lvl); 181 + dump_trace(task, regs, stack, bp, &print_trace_ops, log_lvl); 182 182 } 183 183 184 184 void show_trace(struct task_struct *task, struct pt_regs *regs, 185 - unsigned long *stack) 185 + unsigned long *stack, unsigned long bp) 186 186 { 187 - show_trace_log_lvl(task, regs, stack, ""); 187 + show_trace_log_lvl(task, regs, stack, bp, ""); 188 188 } 189 189 190 190 void show_stack(struct task_struct *task, unsigned long *sp) 191 191 { 192 - show_stack_log_lvl(task, NULL, sp, ""); 192 + show_stack_log_lvl(task, NULL, sp, 0, ""); 193 193 } 194 194 195 195 /* ··· 197 197 */ 198 198 void dump_stack(void) 199 199 { 200 + unsigned long bp; 200 201 unsigned long stack; 201 202 203 + bp = stack_frame(current, NULL); 202 204 printk("Pid: %d, comm: %.20s %s %s %.*s\n", 203 205 current->pid, current->comm, print_tainted(), 204 206 init_utsname()->release, 205 207 (int)strcspn(init_utsname()->version, " "), 206 208 init_utsname()->version); 207 - show_trace(NULL, NULL, &stack); 209 + show_trace(NULL, NULL, &stack, bp); 208 210 } 209 211 EXPORT_SYMBOL(dump_stack); 210 212
+8 -7
arch/x86/kernel/dumpstack_32.c
··· 17 17 #include <asm/stacktrace.h> 18 18 19 19 20 - void dump_trace(struct task_struct *task, 21 - struct pt_regs *regs, unsigned long *stack, 20 + void dump_trace(struct task_struct *task, struct pt_regs *regs, 21 + unsigned long *stack, unsigned long bp, 22 22 const struct stacktrace_ops *ops, void *data) 23 23 { 24 24 int graph = 0; 25 - unsigned long bp; 26 25 27 26 if (!task) 28 27 task = current; ··· 34 35 stack = (unsigned long *)task->thread.sp; 35 36 } 36 37 37 - bp = stack_frame(task, regs); 38 + if (!bp) 39 + bp = stack_frame(task, regs); 40 + 38 41 for (;;) { 39 42 struct thread_info *context; 40 43 ··· 56 55 57 56 void 58 57 show_stack_log_lvl(struct task_struct *task, struct pt_regs *regs, 59 - unsigned long *sp, char *log_lvl) 58 + unsigned long *sp, unsigned long bp, char *log_lvl) 60 59 { 61 60 unsigned long *stack; 62 61 int i; ··· 78 77 touch_nmi_watchdog(); 79 78 } 80 79 printk(KERN_CONT "\n"); 81 - show_trace_log_lvl(task, regs, sp, log_lvl); 80 + show_trace_log_lvl(task, regs, sp, bp, log_lvl); 82 81 } 83 82 84 83 ··· 103 102 u8 *ip; 104 103 105 104 printk(KERN_EMERG "Stack:\n"); 106 - show_stack_log_lvl(NULL, regs, &regs->sp, KERN_EMERG); 105 + show_stack_log_lvl(NULL, regs, &regs->sp, 0, KERN_EMERG); 107 106 108 107 printk(KERN_EMERG "Code: "); 109 108
+7 -7
arch/x86/kernel/dumpstack_64.c
··· 139 139 * severe exception (double fault, nmi, stack fault, debug, mce) hardware stack 140 140 */ 141 141 142 - void dump_trace(struct task_struct *task, 143 - struct pt_regs *regs, unsigned long *stack, 142 + void dump_trace(struct task_struct *task, struct pt_regs *regs, 143 + unsigned long *stack, unsigned long bp, 144 144 const struct stacktrace_ops *ops, void *data) 145 145 { 146 146 const unsigned cpu = get_cpu(); ··· 150 150 struct thread_info *tinfo; 151 151 int graph = 0; 152 152 unsigned long dummy; 153 - unsigned long bp; 154 153 155 154 if (!task) 156 155 task = current; ··· 160 161 stack = (unsigned long *)task->thread.sp; 161 162 } 162 163 163 - bp = stack_frame(task, regs); 164 + if (!bp) 165 + bp = stack_frame(task, regs); 164 166 /* 165 167 * Print function call entries in all stacks, starting at the 166 168 * current stack address. If the stacks consist of nested ··· 225 225 226 226 void 227 227 show_stack_log_lvl(struct task_struct *task, struct pt_regs *regs, 228 - unsigned long *sp, char *log_lvl) 228 + unsigned long *sp, unsigned long bp, char *log_lvl) 229 229 { 230 230 unsigned long *irq_stack_end; 231 231 unsigned long *irq_stack; ··· 269 269 preempt_enable(); 270 270 271 271 printk(KERN_CONT "\n"); 272 - show_trace_log_lvl(task, regs, sp, log_lvl); 272 + show_trace_log_lvl(task, regs, sp, bp, log_lvl); 273 273 } 274 274 275 275 void show_registers(struct pt_regs *regs) ··· 298 298 299 299 printk(KERN_EMERG "Stack:\n"); 300 300 show_stack_log_lvl(NULL, regs, (unsigned long *)sp, 301 - KERN_EMERG); 301 + 0, KERN_EMERG); 302 302 303 303 printk(KERN_EMERG "Code: "); 304 304
+2 -2
arch/x86/kernel/entry_64.S
··· 18 18 * A note on terminology: 19 19 * - top of stack: Architecture defined interrupt frame from SS to RIP 20 20 * at the top of the kernel process stack. 21 - * - partial stack frame: partially saved registers upto R11. 21 + * - partial stack frame: partially saved registers up to R11. 22 22 * - full stack frame: Like partial stack frame, but all register saved. 23 23 * 24 24 * Some macro usage: ··· 422 422 END(ret_from_fork) 423 423 424 424 /* 425 - * System call entry. Upto 6 arguments in registers are supported. 425 + * System call entry. Up to 6 arguments in registers are supported. 426 426 * 427 427 * SYSCALL does not save anything on the stack and does not change the 428 428 * stack pointer.
+1 -1
arch/x86/kernel/i387.c
··· 145 145 * The _current_ task is using the FPU for the first time 146 146 * so initialize it and set the mxcsr to its default 147 147 * value at reset if we support XMM instructions and then 148 - * remeber the current task has used the FPU. 148 + * remember the current task has used the FPU. 149 149 */ 150 150 int init_fpu(struct task_struct *tsk) 151 151 {
+1 -1
arch/x86/kernel/irq_32.c
··· 172 172 173 173 call_on_stack(__do_softirq, isp); 174 174 /* 175 - * Shouldnt happen, we returned above if in_interrupt(): 175 + * Shouldn't happen, we returned above if in_interrupt(): 176 176 */ 177 177 WARN_ON_ONCE(softirq_count()); 178 178 }
+1 -1
arch/x86/kernel/kgdb.c
··· 278 278 pevent = per_cpu_ptr(breakinfo[breakno].pev, cpu); 279 279 if (dbg_release_bp_slot(*pevent)) 280 280 /* 281 - * The debugger is responisble for handing the retry on 281 + * The debugger is responsible for handing the retry on 282 282 * remove failure. 283 283 */ 284 284 return -1;
+1 -1
arch/x86/kernel/mca_32.c
··· 259 259 /* 260 260 * WARNING: Be careful when making changes here. Putting an adapter 261 261 * and the motherboard simultaneously into setup mode may result in 262 - * damage to chips (according to The Indispensible PC Hardware Book 262 + * damage to chips (according to The Indispensable PC Hardware Book 263 263 * by Hans-Peter Messmer). Also, we disable system interrupts (so 264 264 * that we are not disturbed in the middle of this). 265 265 */
+2 -2
arch/x86/kernel/mpparse.c
··· 883 883 884 884 if (!mpc_new_phys) { 885 885 unsigned char old, new; 886 - /* check if we can change the postion */ 886 + /* check if we can change the position */ 887 887 mpc->checksum = 0; 888 888 old = mpf_checksum((unsigned char *)mpc, mpc->length); 889 889 mpc->checksum = 0xff; ··· 892 892 printk(KERN_INFO "mpc is readonly, please try alloc_mptable instead\n"); 893 893 return 0; 894 894 } 895 - printk(KERN_INFO "use in-positon replacing\n"); 895 + printk(KERN_INFO "use in-position replacing\n"); 896 896 } else { 897 897 mpf->physptr = mpc_new_phys; 898 898 mpc_new = phys_to_virt(mpc_new_phys);
+2 -2
arch/x86/kernel/pci-calgary_64.c
··· 1279 1279 1280 1280 if (pci_dev == PCI_DEVICE_ID_IBM_CALIOC2) { 1281 1281 /* 1282 - * FIXME: properly scan for devices accross the 1282 + * FIXME: properly scan for devices across the 1283 1283 * PCI-to-PCI bridge on every CalIOC2 port. 1284 1284 */ 1285 1285 return 1; ··· 1295 1295 1296 1296 /* 1297 1297 * calgary_init_bitmap_from_tce_table(): 1298 - * Funtion for kdump case. In the second/kdump kernel initialize 1298 + * Function for kdump case. In the second/kdump kernel initialize 1299 1299 * the bitmap based on the tce table entries obtained from first kernel 1300 1300 */ 1301 1301 static void calgary_init_bitmap_from_tce_table(struct iommu_table *tbl)
+1 -1
arch/x86/kernel/process.c
··· 87 87 void show_regs(struct pt_regs *regs) 88 88 { 89 89 show_registers(regs); 90 - show_trace(NULL, regs, (unsigned long *)kernel_stack_pointer(regs)); 90 + show_trace(NULL, regs, (unsigned long *)kernel_stack_pointer(regs), 0); 91 91 } 92 92 93 93 void show_regs_common(void)
+3 -3
arch/x86/kernel/stacktrace.c
··· 73 73 */ 74 74 void save_stack_trace(struct stack_trace *trace) 75 75 { 76 - dump_trace(current, NULL, NULL, &save_stack_ops, trace); 76 + dump_trace(current, NULL, NULL, 0, &save_stack_ops, trace); 77 77 if (trace->nr_entries < trace->max_entries) 78 78 trace->entries[trace->nr_entries++] = ULONG_MAX; 79 79 } ··· 81 81 82 82 void save_stack_trace_regs(struct stack_trace *trace, struct pt_regs *regs) 83 83 { 84 - dump_trace(current, regs, NULL, &save_stack_ops, trace); 84 + dump_trace(current, regs, NULL, 0, &save_stack_ops, trace); 85 85 if (trace->nr_entries < trace->max_entries) 86 86 trace->entries[trace->nr_entries++] = ULONG_MAX; 87 87 } 88 88 89 89 void save_stack_trace_tsk(struct task_struct *tsk, struct stack_trace *trace) 90 90 { 91 - dump_trace(tsk, NULL, NULL, &save_stack_ops_nosched, trace); 91 + dump_trace(tsk, NULL, NULL, 0, &save_stack_ops_nosched, trace); 92 92 if (trace->nr_entries < trace->max_entries) 93 93 trace->entries[trace->nr_entries++] = ULONG_MAX; 94 94 }
+1 -1
arch/x86/kernel/step.c
··· 166 166 * Make sure block stepping (BTF) is not enabled unless it should be. 167 167 * Note that we don't try to worry about any is_setting_trap_flag() 168 168 * instructions after the first when using block stepping. 169 - * So noone should try to use debugger block stepping in a program 169 + * So no one should try to use debugger block stepping in a program 170 170 * that uses user-mode single stepping itself. 171 171 */ 172 172 if (enable_single_step(child) && block) {
+1 -1
arch/x86/kernel/topology.c
··· 39 39 /* 40 40 * CPU0 cannot be offlined due to several 41 41 * restrictions and assumptions in kernel. This basically 42 - * doesnt add a control file, one cannot attempt to offline 42 + * doesn't add a control file, one cannot attempt to offline 43 43 * BSP. 44 44 * 45 45 * Also certain PCI quirks require not to enable hotplug control
+2 -2
arch/x86/kernel/tsc.c
··· 427 427 * the delta to the previous read. We keep track of the min 428 428 * and max values of that delta. The delta is mostly defined 429 429 * by the IO time of the PIT access, so we can detect when a 430 - * SMI/SMM disturbance happend between the two reads. If the 430 + * SMI/SMM disturbance happened between the two reads. If the 431 431 * maximum time is significantly larger than the minimum time, 432 432 * then we discard the result and have another try. 433 433 * ··· 900 900 * timer based, instead of loop based, we don't block the boot 901 901 * process while this longer calibration is done. 902 902 * 903 - * If there are any calibration anomolies (too many SMIs, etc), 903 + * If there are any calibration anomalies (too many SMIs, etc), 904 904 * or the refined calibration is off by 1% of the fast early 905 905 * calibration, we throw out the new calibration and use the 906 906 * early calibration.
+1 -1
arch/x86/kernel/verify_cpu.S
··· 18 18 * This file is expected to run in 32bit code. Currently: 19 19 * 20 20 * arch/x86/boot/compressed/head_64.S: Boot cpu verification 21 - * arch/x86/kernel/trampoline_64.S: secondary processor verfication 21 + * arch/x86/kernel/trampoline_64.S: secondary processor verification 22 22 * arch/x86/kernel/head_32.S: processor startup 23 23 * 24 24 * verify_cpu, returns the status of longmode and SSE in register %eax.
+1 -1
arch/x86/kernel/xsave.c
··· 53 53 54 54 /* 55 55 * None of the feature bits are in init state. So nothing else 56 - * to do for us, as the memory layout is upto date. 56 + * to do for us, as the memory layout is up to date. 57 57 */ 58 58 if ((xstate_bv & pcntxt_mask) == pcntxt_mask) 59 59 return;
+1 -1
arch/x86/kvm/paging_tmpl.h
··· 346 346 return; 347 347 348 348 /* 349 - * we call mmu_set_spte() with host_writable = true beacuse that 349 + * we call mmu_set_spte() with host_writable = true because that 350 350 * vcpu->arch.update_pte.pfn was fetched from get_user_pages(write = 1). 351 351 */ 352 352 mmu_set_spte(vcpu, spte, sp->role.access, pte_access, 0, 0,
+1 -1
arch/x86/kvm/timer.c
··· 25 25 26 26 /* 27 27 * There is a race window between reading and incrementing, but we do 28 - * not care about potentially loosing timer events in the !reinject 28 + * not care about potentially losing timer events in the !reinject 29 29 * case anyway. Note: KVM_REQ_PENDING_TIMER is implicitly checked 30 30 * in vcpu_enter_guest. 31 31 */
+1 -1
arch/x86/kvm/x86.c
··· 1031 1031 /* 1032 1032 * Special case: close write to TSC within 5 seconds of 1033 1033 * another CPU is interpreted as an attempt to synchronize 1034 - * The 5 seconds is to accomodate host load / swapping as 1034 + * The 5 seconds is to accommodate host load / swapping as 1035 1035 * well as any reset of TSC during the boot process. 1036 1036 * 1037 1037 * In that case, for a reliable TSC, we can match TSC offsets,
+1 -1
arch/x86/lguest/boot.c
··· 397 397 * instead we just use the real "cpuid" instruction. Then I pretty much turned 398 398 * off feature bits until the Guest booted. (Don't say that: you'll damage 399 399 * lguest sales!) Shut up, inner voice! (Hey, just pointing out that this is 400 - * hardly future proof.) Noone's listening! They don't like you anyway, 400 + * hardly future proof.) No one's listening! They don't like you anyway, 401 401 * parenthetic weirdo! 402 402 * 403 403 * Replacing the cpuid so we can turn features off is great for the kernel, but
+1 -1
arch/x86/lib/copy_user_64.S
··· 117 117 * rdx count 118 118 * 119 119 * Output: 120 - * eax uncopied bytes or 0 if successfull. 120 + * eax uncopied bytes or 0 if successful. 121 121 */ 122 122 ENTRY(copy_user_generic_unrolled) 123 123 CFI_STARTPROC
+125 -125
arch/x86/lib/csum-copy_64.S
··· 1 1 /* 2 - * Copyright 2002,2003 Andi Kleen, SuSE Labs. 3 - * 2 + * Copyright 2002, 2003 Andi Kleen, SuSE Labs. 3 + * 4 4 * This file is subject to the terms and conditions of the GNU General Public 5 5 * License. See the file COPYING in the main directory of this archive 6 6 * for more details. No warranty for anything given at all. ··· 11 11 12 12 /* 13 13 * Checksum copy with exception handling. 14 - * On exceptions src_err_ptr or dst_err_ptr is set to -EFAULT and the 14 + * On exceptions src_err_ptr or dst_err_ptr is set to -EFAULT and the 15 15 * destination is zeroed. 16 - * 16 + * 17 17 * Input 18 18 * rdi source 19 19 * rsi destination 20 20 * edx len (32bit) 21 - * ecx sum (32bit) 21 + * ecx sum (32bit) 22 22 * r8 src_err_ptr (int) 23 23 * r9 dst_err_ptr (int) 24 24 * 25 25 * Output 26 26 * eax 64bit sum. undefined in case of exception. 27 - * 28 - * Wrappers need to take care of valid exception sum and zeroing. 27 + * 28 + * Wrappers need to take care of valid exception sum and zeroing. 29 29 * They also should align source or destination to 8 bytes. 30 30 */ 31 31 32 32 .macro source 33 33 10: 34 - .section __ex_table,"a" 34 + .section __ex_table, "a" 35 35 .align 8 36 - .quad 10b,.Lbad_source 36 + .quad 10b, .Lbad_source 37 37 .previous 38 38 .endm 39 - 39 + 40 40 .macro dest 41 41 20: 42 - .section __ex_table,"a" 42 + .section __ex_table, "a" 43 43 .align 8 44 - .quad 20b,.Lbad_dest 44 + .quad 20b, .Lbad_dest 45 45 .previous 46 46 .endm 47 - 47 + 48 48 .macro ignore L=.Lignore 49 49 30: 50 - .section __ex_table,"a" 50 + .section __ex_table, "a" 51 51 .align 8 52 - .quad 30b,\L 52 + .quad 30b, \L 53 53 .previous 54 54 .endm 55 - 56 - 55 + 56 + 57 57 ENTRY(csum_partial_copy_generic) 58 58 CFI_STARTPROC 59 - cmpl $3*64,%edx 60 - jle .Lignore 59 + cmpl $3*64, %edx 60 + jle .Lignore 61 61 62 - .Lignore: 63 - subq $7*8,%rsp 62 + .Lignore: 63 + subq $7*8, %rsp 64 64 CFI_ADJUST_CFA_OFFSET 7*8 65 - movq %rbx,2*8(%rsp) 65 + movq %rbx, 2*8(%rsp) 66 66 CFI_REL_OFFSET rbx, 2*8 67 - movq %r12,3*8(%rsp) 67 + movq %r12, 3*8(%rsp) 68 68 CFI_REL_OFFSET r12, 3*8 69 - movq %r14,4*8(%rsp) 69 + movq %r14, 4*8(%rsp) 70 70 CFI_REL_OFFSET r14, 4*8 71 - movq %r13,5*8(%rsp) 71 + movq %r13, 5*8(%rsp) 72 72 CFI_REL_OFFSET r13, 5*8 73 - movq %rbp,6*8(%rsp) 73 + movq %rbp, 6*8(%rsp) 74 74 CFI_REL_OFFSET rbp, 6*8 75 75 76 - movq %r8,(%rsp) 77 - movq %r9,1*8(%rsp) 78 - 79 - movl %ecx,%eax 80 - movl %edx,%ecx 76 + movq %r8, (%rsp) 77 + movq %r9, 1*8(%rsp) 81 78 82 - xorl %r9d,%r9d 83 - movq %rcx,%r12 79 + movl %ecx, %eax 80 + movl %edx, %ecx 84 81 85 - shrq $6,%r12 86 - jz .Lhandle_tail /* < 64 */ 82 + xorl %r9d, %r9d 83 + movq %rcx, %r12 84 + 85 + shrq $6, %r12 86 + jz .Lhandle_tail /* < 64 */ 87 87 88 88 clc 89 - 89 + 90 90 /* main loop. clear in 64 byte blocks */ 91 91 /* r9: zero, r8: temp2, rbx: temp1, rax: sum, rcx: saved length */ 92 92 /* r11: temp3, rdx: temp4, r12 loopcnt */ ··· 94 94 .p2align 4 95 95 .Lloop: 96 96 source 97 - movq (%rdi),%rbx 97 + movq (%rdi), %rbx 98 98 source 99 - movq 8(%rdi),%r8 99 + movq 8(%rdi), %r8 100 100 source 101 - movq 16(%rdi),%r11 101 + movq 16(%rdi), %r11 102 102 source 103 - movq 24(%rdi),%rdx 103 + movq 24(%rdi), %rdx 104 104 105 105 source 106 - movq 32(%rdi),%r10 106 + movq 32(%rdi), %r10 107 107 source 108 - movq 40(%rdi),%rbp 108 + movq 40(%rdi), %rbp 109 109 source 110 - movq 48(%rdi),%r14 110 + movq 48(%rdi), %r14 111 111 source 112 - movq 56(%rdi),%r13 113 - 112 + movq 56(%rdi), %r13 113 + 114 114 ignore 2f 115 115 prefetcht0 5*64(%rdi) 116 - 2: 117 - adcq %rbx,%rax 118 - adcq %r8,%rax 119 - adcq %r11,%rax 120 - adcq %rdx,%rax 121 - adcq %r10,%rax 122 - adcq %rbp,%rax 123 - adcq %r14,%rax 124 - adcq %r13,%rax 116 + 2: 117 + adcq %rbx, %rax 118 + adcq %r8, %rax 119 + adcq %r11, %rax 120 + adcq %rdx, %rax 121 + adcq %r10, %rax 122 + adcq %rbp, %rax 123 + adcq %r14, %rax 124 + adcq %r13, %rax 125 125 126 126 decl %r12d 127 - 128 - dest 129 - movq %rbx,(%rsi) 130 - dest 131 - movq %r8,8(%rsi) 132 - dest 133 - movq %r11,16(%rsi) 134 - dest 135 - movq %rdx,24(%rsi) 136 127 137 128 dest 138 - movq %r10,32(%rsi) 129 + movq %rbx, (%rsi) 139 130 dest 140 - movq %rbp,40(%rsi) 131 + movq %r8, 8(%rsi) 141 132 dest 142 - movq %r14,48(%rsi) 133 + movq %r11, 16(%rsi) 143 134 dest 144 - movq %r13,56(%rsi) 145 - 135 + movq %rdx, 24(%rsi) 136 + 137 + dest 138 + movq %r10, 32(%rsi) 139 + dest 140 + movq %rbp, 40(%rsi) 141 + dest 142 + movq %r14, 48(%rsi) 143 + dest 144 + movq %r13, 56(%rsi) 145 + 146 146 3: 147 - 148 - leaq 64(%rdi),%rdi 149 - leaq 64(%rsi),%rsi 150 147 151 - jnz .Lloop 148 + leaq 64(%rdi), %rdi 149 + leaq 64(%rsi), %rsi 152 150 153 - adcq %r9,%rax 151 + jnz .Lloop 154 152 155 - /* do last upto 56 bytes */ 153 + adcq %r9, %rax 154 + 155 + /* do last up to 56 bytes */ 156 156 .Lhandle_tail: 157 157 /* ecx: count */ 158 - movl %ecx,%r10d 159 - andl $63,%ecx 160 - shrl $3,%ecx 161 - jz .Lfold 158 + movl %ecx, %r10d 159 + andl $63, %ecx 160 + shrl $3, %ecx 161 + jz .Lfold 162 162 clc 163 163 .p2align 4 164 - .Lloop_8: 164 + .Lloop_8: 165 165 source 166 - movq (%rdi),%rbx 167 - adcq %rbx,%rax 166 + movq (%rdi), %rbx 167 + adcq %rbx, %rax 168 168 decl %ecx 169 169 dest 170 - movq %rbx,(%rsi) 171 - leaq 8(%rsi),%rsi /* preserve carry */ 172 - leaq 8(%rdi),%rdi 170 + movq %rbx, (%rsi) 171 + leaq 8(%rsi), %rsi /* preserve carry */ 172 + leaq 8(%rdi), %rdi 173 173 jnz .Lloop_8 174 - adcq %r9,%rax /* add in carry */ 174 + adcq %r9, %rax /* add in carry */ 175 175 176 176 .Lfold: 177 177 /* reduce checksum to 32bits */ 178 - movl %eax,%ebx 179 - shrq $32,%rax 180 - addl %ebx,%eax 181 - adcl %r9d,%eax 178 + movl %eax, %ebx 179 + shrq $32, %rax 180 + addl %ebx, %eax 181 + adcl %r9d, %eax 182 182 183 - /* do last upto 6 bytes */ 183 + /* do last up to 6 bytes */ 184 184 .Lhandle_7: 185 - movl %r10d,%ecx 186 - andl $7,%ecx 187 - shrl $1,%ecx 185 + movl %r10d, %ecx 186 + andl $7, %ecx 187 + shrl $1, %ecx 188 188 jz .Lhandle_1 189 - movl $2,%edx 190 - xorl %ebx,%ebx 191 - clc 189 + movl $2, %edx 190 + xorl %ebx, %ebx 191 + clc 192 192 .p2align 4 193 - .Lloop_1: 193 + .Lloop_1: 194 194 source 195 - movw (%rdi),%bx 196 - adcl %ebx,%eax 195 + movw (%rdi), %bx 196 + adcl %ebx, %eax 197 197 decl %ecx 198 198 dest 199 - movw %bx,(%rsi) 200 - leaq 2(%rdi),%rdi 201 - leaq 2(%rsi),%rsi 199 + movw %bx, (%rsi) 200 + leaq 2(%rdi), %rdi 201 + leaq 2(%rsi), %rsi 202 202 jnz .Lloop_1 203 - adcl %r9d,%eax /* add in carry */ 204 - 203 + adcl %r9d, %eax /* add in carry */ 204 + 205 205 /* handle last odd byte */ 206 206 .Lhandle_1: 207 - testl $1,%r10d 207 + testl $1, %r10d 208 208 jz .Lende 209 - xorl %ebx,%ebx 209 + xorl %ebx, %ebx 210 210 source 211 - movb (%rdi),%bl 211 + movb (%rdi), %bl 212 212 dest 213 - movb %bl,(%rsi) 214 - addl %ebx,%eax 215 - adcl %r9d,%eax /* carry */ 216 - 213 + movb %bl, (%rsi) 214 + addl %ebx, %eax 215 + adcl %r9d, %eax /* carry */ 216 + 217 217 CFI_REMEMBER_STATE 218 218 .Lende: 219 - movq 2*8(%rsp),%rbx 219 + movq 2*8(%rsp), %rbx 220 220 CFI_RESTORE rbx 221 - movq 3*8(%rsp),%r12 221 + movq 3*8(%rsp), %r12 222 222 CFI_RESTORE r12 223 - movq 4*8(%rsp),%r14 223 + movq 4*8(%rsp), %r14 224 224 CFI_RESTORE r14 225 - movq 5*8(%rsp),%r13 225 + movq 5*8(%rsp), %r13 226 226 CFI_RESTORE r13 227 - movq 6*8(%rsp),%rbp 227 + movq 6*8(%rsp), %rbp 228 228 CFI_RESTORE rbp 229 - addq $7*8,%rsp 229 + addq $7*8, %rsp 230 230 CFI_ADJUST_CFA_OFFSET -7*8 231 231 ret 232 232 CFI_RESTORE_STATE 233 233 234 234 /* Exception handlers. Very simple, zeroing is done in the wrappers */ 235 235 .Lbad_source: 236 - movq (%rsp),%rax 237 - testq %rax,%rax 236 + movq (%rsp), %rax 237 + testq %rax, %rax 238 238 jz .Lende 239 - movl $-EFAULT,(%rax) 239 + movl $-EFAULT, (%rax) 240 240 jmp .Lende 241 - 241 + 242 242 .Lbad_dest: 243 - movq 8(%rsp),%rax 244 - testq %rax,%rax 245 - jz .Lende 246 - movl $-EFAULT,(%rax) 243 + movq 8(%rsp), %rax 244 + testq %rax, %rax 245 + jz .Lende 246 + movl $-EFAULT, (%rax) 247 247 jmp .Lende 248 248 CFI_ENDPROC 249 249 ENDPROC(csum_partial_copy_generic)
+1 -1
arch/x86/lib/csum-partial_64.c
··· 84 84 count64--; 85 85 } 86 86 87 - /* last upto 7 8byte blocks */ 87 + /* last up to 7 8byte blocks */ 88 88 count %= 8; 89 89 while (count) { 90 90 asm("addq %1,%0\n\t"
+1 -1
arch/x86/mm/hugetlbpage.c
··· 326 326 if (mm->free_area_cache < len) 327 327 goto fail; 328 328 329 - /* either no address requested or cant fit in requested address hole */ 329 + /* either no address requested or can't fit in requested address hole */ 330 330 addr = (mm->free_area_cache - len) & huge_page_mask(h); 331 331 do { 332 332 /*
+1 -1
arch/x86/mm/init_32.c
··· 917 917 { 918 918 /* 919 919 * When this called, init has already been executed and released, 920 - * so everything past _etext sould be NX. 920 + * so everything past _etext should be NX. 921 921 */ 922 922 unsigned long start = PFN_ALIGN(_etext); 923 923 /*
+1 -1
arch/x86/mm/numa_64.c
··· 446 446 * @distance: NUMA distance 447 447 * 448 448 * Set the distance from node @from to @to to @distance. If distance table 449 - * doesn't exist, one which is large enough to accomodate all the currently 449 + * doesn't exist, one which is large enough to accommodate all the currently 450 450 * known nodes will be created. 451 451 * 452 452 * If such table cannot be allocated, a warning is printed and further
+1 -1
arch/x86/mm/pageattr.c
··· 310 310 * these shared mappings are made of small page mappings. 311 311 * Thus this don't enforce !RW mapping for small page kernel 312 312 * text mapping logic will help Linux Xen parvirt guest boot 313 - * aswell. 313 + * as well. 314 314 */ 315 315 if (lookup_address(address, &level) && (level != PG_LEVEL_4K)) 316 316 pgprot_val(forbidden) |= _PAGE_RW;
+1 -2
arch/x86/mm/pgtable.c
··· 168 168 * section 8.1: in PAE mode we explicitly have to flush the 169 169 * TLB via cr3 if the top-level pgd is changed... 170 170 */ 171 - if (mm == current->active_mm) 172 - write_cr3(read_cr3()); 171 + flush_tlb_mm(mm); 173 172 } 174 173 #else /* !CONFIG_X86_PAE */ 175 174
+1 -1
arch/x86/oprofile/backtrace.c
··· 126 126 if (!user_mode_vm(regs)) { 127 127 unsigned long stack = kernel_stack_pointer(regs); 128 128 if (depth) 129 - dump_trace(NULL, regs, (unsigned long *)stack, 129 + dump_trace(NULL, regs, (unsigned long *)stack, 0, 130 130 &backtrace_ops, &depth); 131 131 return; 132 132 }
+2 -2
arch/x86/pci/i386.c
··· 241 241 e820_reserve_resources_late(); 242 242 /* 243 243 * Insert the IO APIC resources after PCI initialization has 244 - * occured to handle IO APICS that are mapped in on a BAR in 244 + * occurred to handle IO APICS that are mapped in on a BAR in 245 245 * PCI space, but before trying to assign unassigned pci res. 246 246 */ 247 247 ioapic_insert_resources(); ··· 304 304 /* 305 305 * ioremap() and ioremap_nocache() defaults to UC MINUS for now. 306 306 * To avoid attribute conflicts, request UC MINUS here 307 - * aswell. 307 + * as well. 308 308 */ 309 309 prot |= _PAGE_CACHE_UC_MINUS; 310 310
+1 -1
arch/x86/xen/mmu.c
··· 1744 1744 } 1745 1745 1746 1746 /* 1747 - * Set up the inital kernel pagetable. 1747 + * Set up the initial kernel pagetable. 1748 1748 * 1749 1749 * We can construct this by grafting the Xen provided pagetable into 1750 1750 * head_64.S's preconstructed pagetables. We copy the Xen L2's into