Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

x86/cpu: Convert printk(KERN_<LEVEL> ...) to pr_<level>(...)

- Use the more current logging style pr_<level>(...) instead of the old
printk(KERN_<LEVEL> ...).

- Convert pr_warning() to pr_warn().

Signed-off-by: Chen Yucong <slaoub@gmail.com>
Cc: Borislav Petkov <bp@alien8.de>
Cc: H. Peter Anvin <hpa@zytor.com>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Link: http://lkml.kernel.org/r/1454384702-21707-1-git-send-email-slaoub@gmail.com
Signed-off-by: Ingo Molnar <mingo@kernel.org>

authored by

Chen Yucong and committed by
Ingo Molnar
1b74dde7 16aaa537

+146 -159
+10 -13
arch/x86/kernel/cpu/amd.c
··· 117 117 void (*f_vide)(void); 118 118 u64 d, d2; 119 119 120 - printk(KERN_INFO "AMD K6 stepping B detected - "); 120 + pr_info("AMD K6 stepping B detected - "); 121 121 122 122 /* 123 123 * It looks like AMD fixed the 2.6.2 bug and improved indirect ··· 133 133 d = d2-d; 134 134 135 135 if (d > 20*K6_BUG_LOOP) 136 - printk(KERN_CONT 137 - "system stability may be impaired when more than 32 MB are used.\n"); 136 + pr_cont("system stability may be impaired when more than 32 MB are used.\n"); 138 137 else 139 - printk(KERN_CONT "probably OK (after B9730xxxx).\n"); 138 + pr_cont("probably OK (after B9730xxxx).\n"); 140 139 } 141 140 142 141 /* K6 with old style WHCR */ ··· 153 154 wbinvd(); 154 155 wrmsr(MSR_K6_WHCR, l, h); 155 156 local_irq_restore(flags); 156 - printk(KERN_INFO "Enabling old style K6 write allocation for %d Mb\n", 157 + pr_info("Enabling old style K6 write allocation for %d Mb\n", 157 158 mbytes); 158 159 } 159 160 return; ··· 174 175 wbinvd(); 175 176 wrmsr(MSR_K6_WHCR, l, h); 176 177 local_irq_restore(flags); 177 - printk(KERN_INFO "Enabling new style K6 write allocation for %d Mb\n", 178 + pr_info("Enabling new style K6 write allocation for %d Mb\n", 178 179 mbytes); 179 180 } 180 181 ··· 201 202 */ 202 203 if (c->x86_model >= 6 && c->x86_model <= 10) { 203 204 if (!cpu_has(c, X86_FEATURE_XMM)) { 204 - printk(KERN_INFO "Enabling disabled K7/SSE Support.\n"); 205 + pr_info("Enabling disabled K7/SSE Support.\n"); 205 206 msr_clear_bit(MSR_K7_HWCR, 15); 206 207 set_cpu_cap(c, X86_FEATURE_XMM); 207 208 } ··· 215 216 if ((c->x86_model == 8 && c->x86_mask >= 1) || (c->x86_model > 8)) { 216 217 rdmsr(MSR_K7_CLK_CTL, l, h); 217 218 if ((l & 0xfff00000) != 0x20000000) { 218 - printk(KERN_INFO 219 - "CPU: CLK_CTL MSR was %x. Reprogramming to %x\n", 220 - l, ((l & 0x000fffff)|0x20000000)); 219 + pr_info("CPU: CLK_CTL MSR was %x. Reprogramming to %x\n", 220 + l, ((l & 0x000fffff)|0x20000000)); 221 221 wrmsr(MSR_K7_CLK_CTL, (l & 0x000fffff)|0x20000000, h); 222 222 } 223 223 } ··· 483 485 if (!rdmsrl_safe(MSR_K8_TSEG_ADDR, &tseg)) { 484 486 unsigned long pfn = tseg >> PAGE_SHIFT; 485 487 486 - printk(KERN_DEBUG "tseg: %010llx\n", tseg); 488 + pr_debug("tseg: %010llx\n", tseg); 487 489 if (pfn_range_is_mapped(pfn, pfn + 1)) 488 490 set_memory_4k((unsigned long)__va(tseg), 1); 489 491 } ··· 498 500 499 501 rdmsrl(MSR_K7_HWCR, val); 500 502 if (!(val & BIT(24))) 501 - printk(KERN_WARNING FW_BUG "TSC doesn't count " 502 - "with P0 frequency!\n"); 503 + pr_warn(FW_BUG "TSC doesn't count with P0 frequency!\n"); 503 504 } 504 505 } 505 506
+1 -1
arch/x86/kernel/cpu/bugs_64.c
··· 15 15 { 16 16 identify_boot_cpu(); 17 17 #if !defined(CONFIG_SMP) 18 - printk(KERN_INFO "CPU: "); 18 + pr_info("CPU: "); 19 19 print_cpu_info(&boot_cpu_data); 20 20 #endif 21 21 alternative_instructions();
+5 -5
arch/x86/kernel/cpu/centaur.c
··· 29 29 rdmsr(MSR_VIA_FCR, lo, hi); 30 30 lo |= ACE_FCR; /* enable ACE unit */ 31 31 wrmsr(MSR_VIA_FCR, lo, hi); 32 - printk(KERN_INFO "CPU: Enabled ACE h/w crypto\n"); 32 + pr_info("CPU: Enabled ACE h/w crypto\n"); 33 33 } 34 34 35 35 /* enable RNG unit, if present and disabled */ ··· 37 37 rdmsr(MSR_VIA_RNG, lo, hi); 38 38 lo |= RNG_ENABLE; /* enable RNG unit */ 39 39 wrmsr(MSR_VIA_RNG, lo, hi); 40 - printk(KERN_INFO "CPU: Enabled h/w RNG\n"); 40 + pr_info("CPU: Enabled h/w RNG\n"); 41 41 } 42 42 43 43 /* store Centaur Extended Feature Flags as ··· 130 130 name = "C6"; 131 131 fcr_set = ECX8|DSMC|EDCTLB|EMMX|ERETSTK; 132 132 fcr_clr = DPDC; 133 - printk(KERN_NOTICE "Disabling bugged TSC.\n"); 133 + pr_notice("Disabling bugged TSC.\n"); 134 134 clear_cpu_cap(c, X86_FEATURE_TSC); 135 135 break; 136 136 case 8: ··· 163 163 newlo = (lo|fcr_set) & (~fcr_clr); 164 164 165 165 if (newlo != lo) { 166 - printk(KERN_INFO "Centaur FCR was 0x%X now 0x%X\n", 166 + pr_info("Centaur FCR was 0x%X now 0x%X\n", 167 167 lo, newlo); 168 168 wrmsr(MSR_IDT_FCR1, newlo, hi); 169 169 } else { 170 - printk(KERN_INFO "Centaur FCR is 0x%X\n", lo); 170 + pr_info("Centaur FCR is 0x%X\n", lo); 171 171 } 172 172 /* Emulate MTRRs using Centaur's MCR. */ 173 173 set_cpu_cap(c, X86_FEATURE_CENTAUR_MCR);
+20 -22
arch/x86/kernel/cpu/common.c
··· 228 228 lo |= 0x200000; 229 229 wrmsr(MSR_IA32_BBL_CR_CTL, lo, hi); 230 230 231 - printk(KERN_NOTICE "CPU serial number disabled.\n"); 231 + pr_notice("CPU serial number disabled.\n"); 232 232 clear_cpu_cap(c, X86_FEATURE_PN); 233 233 234 234 /* Disabling the serial number may affect the cpuid level */ ··· 329 329 if (!warn) 330 330 continue; 331 331 332 - printk(KERN_WARNING 333 - "CPU: CPU feature " X86_CAP_FMT " disabled, no CPUID level 0x%x\n", 334 - x86_cap_flag(df->feature), df->level); 332 + pr_warn("CPU: CPU feature " X86_CAP_FMT " disabled, no CPUID level 0x%x\n", 333 + x86_cap_flag(df->feature), df->level); 335 334 } 336 335 } 337 336 ··· 509 510 smp_num_siblings = (ebx & 0xff0000) >> 16; 510 511 511 512 if (smp_num_siblings == 1) { 512 - printk_once(KERN_INFO "CPU0: Hyper-Threading is disabled\n"); 513 + pr_info_once("CPU0: Hyper-Threading is disabled\n"); 513 514 goto out; 514 515 } 515 516 ··· 530 531 531 532 out: 532 533 if (!printed && (c->x86_max_cores * smp_num_siblings) > 1) { 533 - printk(KERN_INFO "CPU: Physical Processor ID: %d\n", 534 - c->phys_proc_id); 535 - printk(KERN_INFO "CPU: Processor Core ID: %d\n", 536 - c->cpu_core_id); 534 + pr_info("CPU: Physical Processor ID: %d\n", 535 + c->phys_proc_id); 536 + pr_info("CPU: Processor Core ID: %d\n", 537 + c->cpu_core_id); 537 538 printed = 1; 538 539 } 539 540 #endif ··· 558 559 } 559 560 } 560 561 561 - printk_once(KERN_ERR 562 - "CPU: vendor_id '%s' unknown, using generic init.\n" \ 563 - "CPU: Your system may be unstable.\n", v); 562 + pr_err_once("CPU: vendor_id '%s' unknown, using generic init.\n" \ 563 + "CPU: Your system may be unstable.\n", v); 564 564 565 565 c->x86_vendor = X86_VENDOR_UNKNOWN; 566 566 this_cpu = &default_cpu; ··· 758 760 int count = 0; 759 761 760 762 #ifdef CONFIG_PROCESSOR_SELECT 761 - printk(KERN_INFO "KERNEL supported cpus:\n"); 763 + pr_info("KERNEL supported cpus:\n"); 762 764 #endif 763 765 764 766 for (cdev = __x86_cpu_dev_start; cdev < __x86_cpu_dev_end; cdev++) { ··· 776 778 for (j = 0; j < 2; j++) { 777 779 if (!cpudev->c_ident[j]) 778 780 continue; 779 - printk(KERN_INFO " %s %s\n", cpudev->c_vendor, 781 + pr_info(" %s %s\n", cpudev->c_vendor, 780 782 cpudev->c_ident[j]); 781 783 } 782 784 } ··· 1059 1061 for (index = index_min; index < index_max; index++) { 1060 1062 if (rdmsrl_safe(index, &val)) 1061 1063 continue; 1062 - printk(KERN_INFO " MSR%08x: %016llx\n", index, val); 1064 + pr_info(" MSR%08x: %016llx\n", index, val); 1063 1065 } 1064 1066 } 1065 1067 } ··· 1098 1100 } 1099 1101 1100 1102 if (vendor && !strstr(c->x86_model_id, vendor)) 1101 - printk(KERN_CONT "%s ", vendor); 1103 + pr_cont("%s ", vendor); 1102 1104 1103 1105 if (c->x86_model_id[0]) 1104 - printk(KERN_CONT "%s", c->x86_model_id); 1106 + pr_cont("%s", c->x86_model_id); 1105 1107 else 1106 - printk(KERN_CONT "%d86", c->x86); 1108 + pr_cont("%d86", c->x86); 1107 1109 1108 - printk(KERN_CONT " (family: 0x%x, model: 0x%x", c->x86, c->x86_model); 1110 + pr_cont(" (family: 0x%x, model: 0x%x", c->x86, c->x86_model); 1109 1111 1110 1112 if (c->x86_mask || c->cpuid_level >= 0) 1111 - printk(KERN_CONT ", stepping: 0x%x)\n", c->x86_mask); 1113 + pr_cont(", stepping: 0x%x)\n", c->x86_mask); 1112 1114 else 1113 - printk(KERN_CONT ")\n"); 1115 + pr_cont(")\n"); 1114 1116 1115 1117 print_cpu_msr(c); 1116 1118 } ··· 1436 1438 1437 1439 show_ucode_info_early(); 1438 1440 1439 - printk(KERN_INFO "Initializing CPU#%d\n", cpu); 1441 + pr_info("Initializing CPU#%d\n", cpu); 1440 1442 1441 1443 if (cpu_feature_enabled(X86_FEATURE_VME) || 1442 1444 cpu_has_tsc ||
+5 -5
arch/x86/kernel/cpu/cyrix.c
··· 103 103 local_irq_restore(flags); 104 104 105 105 if (ccr5 & 2) { /* possible wrong calibration done */ 106 - printk(KERN_INFO "Recalibrating delay loop with SLOP bit reset\n"); 106 + pr_info("Recalibrating delay loop with SLOP bit reset\n"); 107 107 calibrate_delay(); 108 108 c->loops_per_jiffy = loops_per_jiffy; 109 109 } ··· 115 115 { 116 116 u8 ccr3; 117 117 118 - printk(KERN_INFO "Enable Memory access reorder on Cyrix/NSC processor.\n"); 118 + pr_info("Enable Memory access reorder on Cyrix/NSC processor.\n"); 119 119 ccr3 = getCx86(CX86_CCR3); 120 120 setCx86(CX86_CCR3, (ccr3 & 0x0f) | 0x10); /* enable MAPEN */ 121 121 ··· 128 128 129 129 static void set_cx86_memwb(void) 130 130 { 131 - printk(KERN_INFO "Enable Memory-Write-back mode on Cyrix/NSC processor.\n"); 131 + pr_info("Enable Memory-Write-back mode on Cyrix/NSC processor.\n"); 132 132 133 133 /* CCR2 bit 2: unlock NW bit */ 134 134 setCx86_old(CX86_CCR2, getCx86_old(CX86_CCR2) & ~0x04); ··· 268 268 * VSA1 we work around however. 269 269 */ 270 270 271 - printk(KERN_INFO "Working around Cyrix MediaGX virtual DMA bugs.\n"); 271 + pr_info("Working around Cyrix MediaGX virtual DMA bugs.\n"); 272 272 isa_dma_bridge_buggy = 2; 273 273 274 274 /* We do this before the PCI layer is running. However we ··· 426 426 if (dir0 == 5 || dir0 == 3) { 427 427 unsigned char ccr3; 428 428 unsigned long flags; 429 - printk(KERN_INFO "Enabling CPUID on Cyrix processor.\n"); 429 + pr_info("Enabling CPUID on Cyrix processor.\n"); 430 430 local_irq_save(flags); 431 431 ccr3 = getCx86(CX86_CCR3); 432 432 /* enable MAPEN */
+1 -1
arch/x86/kernel/cpu/hypervisor.c
··· 56 56 } 57 57 58 58 if (max_pri) 59 - printk(KERN_INFO "Hypervisor detected: %s\n", x86_hyper->name); 59 + pr_info("Hypervisor detected: %s\n", x86_hyper->name); 60 60 } 61 61 62 62 void init_hypervisor(struct cpuinfo_x86 *c)
+5 -5
arch/x86/kernel/cpu/intel.c
··· 61 61 */ 62 62 if (c->x86 == 6 && c->x86_model == 0x1c && c->x86_mask <= 2 && 63 63 c->microcode < 0x20e) { 64 - printk(KERN_WARNING "Atom PSE erratum detected, BIOS microcode update recommended\n"); 64 + pr_warn("Atom PSE erratum detected, BIOS microcode update recommended\n"); 65 65 clear_cpu_cap(c, X86_FEATURE_PSE); 66 66 } 67 67 ··· 140 140 if (c->x86 > 6 || (c->x86 == 6 && c->x86_model >= 0xd)) { 141 141 rdmsrl(MSR_IA32_MISC_ENABLE, misc_enable); 142 142 if (!(misc_enable & MSR_IA32_MISC_ENABLE_FAST_STRING)) { 143 - printk(KERN_INFO "Disabled fast string operations\n"); 143 + pr_info("Disabled fast string operations\n"); 144 144 setup_clear_cpu_cap(X86_FEATURE_REP_GOOD); 145 145 setup_clear_cpu_cap(X86_FEATURE_ERMS); 146 146 } ··· 176 176 boot_cpu_data.x86 == 6 && 177 177 boot_cpu_data.x86_model == 1 && 178 178 boot_cpu_data.x86_mask < 8) { 179 - printk(KERN_INFO "Pentium Pro with Errata#50 detected. Taking evasive action.\n"); 179 + pr_info("Pentium Pro with Errata#50 detected. Taking evasive action.\n"); 180 180 return 1; 181 181 } 182 182 return 0; ··· 225 225 226 226 set_cpu_bug(c, X86_BUG_F00F); 227 227 if (!f00f_workaround_enabled) { 228 - printk(KERN_NOTICE "Intel Pentium with F0 0F bug - workaround enabled.\n"); 228 + pr_notice("Intel Pentium with F0 0F bug - workaround enabled.\n"); 229 229 f00f_workaround_enabled = 1; 230 230 } 231 231 } ··· 244 244 * Forcefully enable PAE if kernel parameter "forcepae" is present. 245 245 */ 246 246 if (forcepae) { 247 - printk(KERN_WARNING "PAE forced!\n"); 247 + pr_warn("PAE forced!\n"); 248 248 set_cpu_cap(c, X86_FEATURE_PAE); 249 249 add_taint(TAINT_CPU_OUT_OF_SPEC, LOCKDEP_NOW_UNRELIABLE); 250 250 }
+1 -1
arch/x86/kernel/cpu/intel_cacheinfo.c
··· 444 444 err = amd_set_l3_disable_slot(nb, cpu, slot, val); 445 445 if (err) { 446 446 if (err == -EEXIST) 447 - pr_warning("L3 slot %d in use/index already disabled!\n", 447 + pr_warn("L3 slot %d in use/index already disabled!\n", 448 448 slot); 449 449 return err; 450 450 }
+7 -8
arch/x86/kernel/cpu/mcheck/mce-inject.c
··· 115 115 int cpu = m->extcpu; 116 116 117 117 if (m->inject_flags & MCJ_EXCEPTION) { 118 - printk(KERN_INFO "Triggering MCE exception on CPU %d\n", cpu); 118 + pr_info("Triggering MCE exception on CPU %d\n", cpu); 119 119 switch (context) { 120 120 case MCJ_CTX_IRQ: 121 121 /* ··· 128 128 raise_exception(m, NULL); 129 129 break; 130 130 default: 131 - printk(KERN_INFO "Invalid MCE context\n"); 131 + pr_info("Invalid MCE context\n"); 132 132 ret = -EINVAL; 133 133 } 134 - printk(KERN_INFO "MCE exception done on CPU %d\n", cpu); 134 + pr_info("MCE exception done on CPU %d\n", cpu); 135 135 } else if (m->status) { 136 - printk(KERN_INFO "Starting machine check poll CPU %d\n", cpu); 136 + pr_info("Starting machine check poll CPU %d\n", cpu); 137 137 raise_poll(m); 138 138 mce_notify_irq(); 139 - printk(KERN_INFO "Machine check poll done on CPU %d\n", cpu); 139 + pr_info("Machine check poll done on CPU %d\n", cpu); 140 140 } else 141 141 m->finished = 0; 142 142 ··· 183 183 start = jiffies; 184 184 while (!cpumask_empty(mce_inject_cpumask)) { 185 185 if (!time_before(jiffies, start + 2*HZ)) { 186 - printk(KERN_ERR 187 - "Timeout waiting for mce inject %lx\n", 186 + pr_err("Timeout waiting for mce inject %lx\n", 188 187 *cpumask_bits(mce_inject_cpumask)); 189 188 break; 190 189 } ··· 240 241 { 241 242 if (!alloc_cpumask_var(&mce_inject_cpumask, GFP_KERNEL)) 242 243 return -ENOMEM; 243 - printk(KERN_INFO "Machine check injector initialized\n"); 244 + pr_info("Machine check injector initialized\n"); 244 245 register_mce_write_callback(mce_write); 245 246 register_nmi_handler(NMI_LOCAL, mce_raise_notify, 0, 246 247 "mce_notify");
+7 -11
arch/x86/kernel/cpu/mcheck/p5.c
··· 26 26 rdmsr(MSR_IA32_P5_MC_ADDR, loaddr, hi); 27 27 rdmsr(MSR_IA32_P5_MC_TYPE, lotype, hi); 28 28 29 - printk(KERN_EMERG 30 - "CPU#%d: Machine Check Exception: 0x%8X (type 0x%8X).\n", 31 - smp_processor_id(), loaddr, lotype); 29 + pr_emerg("CPU#%d: Machine Check Exception: 0x%8X (type 0x%8X).\n", 30 + smp_processor_id(), loaddr, lotype); 32 31 33 32 if (lotype & (1<<5)) { 34 - printk(KERN_EMERG 35 - "CPU#%d: Possible thermal failure (CPU on fire ?).\n", 36 - smp_processor_id()); 33 + pr_emerg("CPU#%d: Possible thermal failure (CPU on fire ?).\n", 34 + smp_processor_id()); 37 35 } 38 36 39 37 add_taint(TAINT_MACHINE_CHECK, LOCKDEP_NOW_UNRELIABLE); ··· 59 61 /* Read registers before enabling: */ 60 62 rdmsr(MSR_IA32_P5_MC_ADDR, l, h); 61 63 rdmsr(MSR_IA32_P5_MC_TYPE, l, h); 62 - printk(KERN_INFO 63 - "Intel old style machine check architecture supported.\n"); 64 + pr_info("Intel old style machine check architecture supported.\n"); 64 65 65 66 /* Enable MCE: */ 66 67 cr4_set_bits(X86_CR4_MCE); 67 - printk(KERN_INFO 68 - "Intel old style machine check reporting enabled on CPU#%d.\n", 69 - smp_processor_id()); 68 + pr_info("Intel old style machine check reporting enabled on CPU#%d.\n", 69 + smp_processor_id()); 70 70 }
+7 -8
arch/x86/kernel/cpu/mcheck/therm_throt.c
··· 190 190 /* if we just entered the thermal event */ 191 191 if (new_event) { 192 192 if (event == THERMAL_THROTTLING_EVENT) 193 - printk(KERN_CRIT "CPU%d: %s temperature above threshold, cpu clock throttled (total events = %lu)\n", 193 + pr_crit("CPU%d: %s temperature above threshold, cpu clock throttled (total events = %lu)\n", 194 194 this_cpu, 195 195 level == CORE_LEVEL ? "Core" : "Package", 196 196 state->count); ··· 198 198 } 199 199 if (old_event) { 200 200 if (event == THERMAL_THROTTLING_EVENT) 201 - printk(KERN_INFO "CPU%d: %s temperature/speed normal\n", 202 - this_cpu, 201 + pr_info("CPU%d: %s temperature/speed normal\n", this_cpu, 203 202 level == CORE_LEVEL ? "Core" : "Package"); 204 203 return 1; 205 204 } ··· 416 417 417 418 static void unexpected_thermal_interrupt(void) 418 419 { 419 - printk(KERN_ERR "CPU%d: Unexpected LVT thermal interrupt!\n", 420 - smp_processor_id()); 420 + pr_err("CPU%d: Unexpected LVT thermal interrupt!\n", 421 + smp_processor_id()); 421 422 } 422 423 423 424 static void (*smp_thermal_vector)(void) = unexpected_thermal_interrupt; ··· 498 499 499 500 if ((l & MSR_IA32_MISC_ENABLE_TM1) && (h & APIC_DM_SMI)) { 500 501 if (system_state == SYSTEM_BOOTING) 501 - printk(KERN_DEBUG "CPU%d: Thermal monitoring handled by SMI\n", cpu); 502 + pr_debug("CPU%d: Thermal monitoring handled by SMI\n", cpu); 502 503 return; 503 504 } 504 505 ··· 556 557 l = apic_read(APIC_LVTTHMR); 557 558 apic_write(APIC_LVTTHMR, l & ~APIC_LVT_MASKED); 558 559 559 - printk_once(KERN_INFO "CPU0: Thermal monitoring enabled (%s)\n", 560 - tm2 ? "TM2" : "TM1"); 560 + pr_info_once("CPU0: Thermal monitoring enabled (%s)\n", 561 + tm2 ? "TM2" : "TM1"); 561 562 562 563 /* enable thermal throttle processing */ 563 564 atomic_set(&therm_throt_en, 1);
+2 -2
arch/x86/kernel/cpu/mcheck/threshold.c
··· 12 12 13 13 static void default_threshold_interrupt(void) 14 14 { 15 - printk(KERN_ERR "Unexpected threshold interrupt at vector %x\n", 16 - THRESHOLD_APIC_VECTOR); 15 + pr_err("Unexpected threshold interrupt at vector %x\n", 16 + THRESHOLD_APIC_VECTOR); 17 17 } 18 18 19 19 void (*mce_threshold_vector)(void) = default_threshold_interrupt;
+2 -3
arch/x86/kernel/cpu/mcheck/winchip.c
··· 17 17 { 18 18 ist_enter(regs); 19 19 20 - printk(KERN_EMERG "CPU0: Machine Check Exception.\n"); 20 + pr_emerg("CPU0: Machine Check Exception.\n"); 21 21 add_taint(TAINT_MACHINE_CHECK, LOCKDEP_NOW_UNRELIABLE); 22 22 23 23 ist_exit(regs); ··· 39 39 40 40 cr4_set_bits(X86_CR4_MCE); 41 41 42 - printk(KERN_INFO 43 - "Winchip machine check reporting enabled on CPU#0.\n"); 42 + pr_info("Winchip machine check reporting enabled on CPU#0.\n"); 44 43 }
+1 -1
arch/x86/kernel/cpu/microcode/amd.c
··· 953 953 struct cpuinfo_x86 *c = &boot_cpu_data; 954 954 955 955 if (c->x86_vendor != X86_VENDOR_AMD || c->x86 < 0x10) { 956 - pr_warning("AMD CPU family 0x%x not supported\n", c->x86); 956 + pr_warn("AMD CPU family 0x%x not supported\n", c->x86); 957 957 return NULL; 958 958 } 959 959
+4 -4
arch/x86/kernel/cpu/mshyperv.c
··· 161 161 ms_hyperv.misc_features = cpuid_edx(HYPERV_CPUID_FEATURES); 162 162 ms_hyperv.hints = cpuid_eax(HYPERV_CPUID_ENLIGHTMENT_INFO); 163 163 164 - printk(KERN_INFO "HyperV: features 0x%x, hints 0x%x\n", 165 - ms_hyperv.features, ms_hyperv.hints); 164 + pr_info("HyperV: features 0x%x, hints 0x%x\n", 165 + ms_hyperv.features, ms_hyperv.hints); 166 166 167 167 #ifdef CONFIG_X86_LOCAL_APIC 168 168 if (ms_hyperv.features & HV_X64_MSR_APIC_FREQUENCY_AVAILABLE) { ··· 174 174 rdmsrl(HV_X64_MSR_APIC_FREQUENCY, hv_lapic_frequency); 175 175 hv_lapic_frequency = div_u64(hv_lapic_frequency, HZ); 176 176 lapic_timer_frequency = hv_lapic_frequency; 177 - printk(KERN_INFO "HyperV: LAPIC Timer Frequency: %#x\n", 178 - lapic_timer_frequency); 177 + pr_info("HyperV: LAPIC Timer Frequency: %#x\n", 178 + lapic_timer_frequency); 179 179 } 180 180 #endif 181 181
+1 -1
arch/x86/kernel/cpu/mtrr/centaur.c
··· 103 103 */ 104 104 if (type != MTRR_TYPE_WRCOMB && 105 105 (centaur_mcr_type == 0 || type != MTRR_TYPE_UNCACHABLE)) { 106 - pr_warning("mtrr: only write-combining%s supported\n", 106 + pr_warn("mtrr: only write-combining%s supported\n", 107 107 centaur_mcr_type ? " and uncacheable are" : " is"); 108 108 return -EINVAL; 109 109 }
+22 -22
arch/x86/kernel/cpu/mtrr/cleanup.c
··· 57 57 static struct var_mtrr_range_state __initdata range_state[RANGE_NUM]; 58 58 59 59 static int __initdata debug_print; 60 - #define Dprintk(x...) do { if (debug_print) printk(KERN_DEBUG x); } while (0) 60 + #define Dprintk(x...) do { if (debug_print) pr_debug(x); } while (0) 61 61 62 - #define BIOS_BUG_MSG KERN_WARNING \ 62 + #define BIOS_BUG_MSG \ 63 63 "WARNING: BIOS bug: VAR MTRR %d contains strange UC entry under 1M, check with your system vendor!\n" 64 64 65 65 static int __init ··· 81 81 base, base + size); 82 82 } 83 83 if (debug_print) { 84 - printk(KERN_DEBUG "After WB checking\n"); 84 + pr_debug("After WB checking\n"); 85 85 for (i = 0; i < nr_range; i++) 86 - printk(KERN_DEBUG "MTRR MAP PFN: %016llx - %016llx\n", 86 + pr_debug("MTRR MAP PFN: %016llx - %016llx\n", 87 87 range[i].start, range[i].end); 88 88 } 89 89 ··· 101 101 (mtrr_state.enabled & MTRR_STATE_MTRR_ENABLED) && 102 102 (mtrr_state.enabled & MTRR_STATE_MTRR_FIXED_ENABLED)) { 103 103 /* Var MTRR contains UC entry below 1M? Skip it: */ 104 - printk(BIOS_BUG_MSG, i); 104 + pr_warn(BIOS_BUG_MSG, i); 105 105 if (base + size <= (1<<(20-PAGE_SHIFT))) 106 106 continue; 107 107 size -= (1<<(20-PAGE_SHIFT)) - base; ··· 114 114 extra_remove_base + extra_remove_size); 115 115 116 116 if (debug_print) { 117 - printk(KERN_DEBUG "After UC checking\n"); 117 + pr_debug("After UC checking\n"); 118 118 for (i = 0; i < RANGE_NUM; i++) { 119 119 if (!range[i].end) 120 120 continue; 121 - printk(KERN_DEBUG "MTRR MAP PFN: %016llx - %016llx\n", 121 + pr_debug("MTRR MAP PFN: %016llx - %016llx\n", 122 122 range[i].start, range[i].end); 123 123 } 124 124 } ··· 126 126 /* sort the ranges */ 127 127 nr_range = clean_sort_range(range, RANGE_NUM); 128 128 if (debug_print) { 129 - printk(KERN_DEBUG "After sorting\n"); 129 + pr_debug("After sorting\n"); 130 130 for (i = 0; i < nr_range; i++) 131 - printk(KERN_DEBUG "MTRR MAP PFN: %016llx - %016llx\n", 131 + pr_debug("MTRR MAP PFN: %016llx - %016llx\n", 132 132 range[i].start, range[i].end); 133 133 } 134 134 ··· 544 544 start_base = to_size_factor(start_base, &start_factor), 545 545 type = range_state[i].type; 546 546 547 - printk(KERN_DEBUG "reg %d, base: %ld%cB, range: %ld%cB, type %s\n", 547 + pr_debug("reg %d, base: %ld%cB, range: %ld%cB, type %s\n", 548 548 i, start_base, start_factor, 549 549 size_base, size_factor, 550 550 (type == MTRR_TYPE_UNCACHABLE) ? "UC" : ··· 713 713 return 0; 714 714 715 715 /* Print original var MTRRs at first, for debugging: */ 716 - printk(KERN_DEBUG "original variable MTRRs\n"); 716 + pr_debug("original variable MTRRs\n"); 717 717 print_out_mtrr_range_state(); 718 718 719 719 memset(range, 0, sizeof(range)); ··· 733 733 x_remove_base, x_remove_size); 734 734 735 735 range_sums = sum_ranges(range, nr_range); 736 - printk(KERN_INFO "total RAM covered: %ldM\n", 736 + pr_info("total RAM covered: %ldM\n", 737 737 range_sums >> (20 - PAGE_SHIFT)); 738 738 739 739 if (mtrr_chunk_size && mtrr_gran_size) { ··· 745 745 746 746 if (!result[i].bad) { 747 747 set_var_mtrr_all(address_bits); 748 - printk(KERN_DEBUG "New variable MTRRs\n"); 748 + pr_debug("New variable MTRRs\n"); 749 749 print_out_mtrr_range_state(); 750 750 return 1; 751 751 } 752 - printk(KERN_INFO "invalid mtrr_gran_size or mtrr_chunk_size, " 753 - "will find optimal one\n"); 752 + pr_info("invalid mtrr_gran_size or mtrr_chunk_size, will find optimal one\n"); 754 753 } 755 754 756 755 i = 0; ··· 767 768 x_remove_base, x_remove_size, i); 768 769 if (debug_print) { 769 770 mtrr_print_out_one_result(i); 770 - printk(KERN_INFO "\n"); 771 + pr_info("\n"); 771 772 } 772 773 773 774 i++; ··· 778 779 index_good = mtrr_search_optimal_index(); 779 780 780 781 if (index_good != -1) { 781 - printk(KERN_INFO "Found optimal setting for mtrr clean up\n"); 782 + pr_info("Found optimal setting for mtrr clean up\n"); 782 783 i = index_good; 783 784 mtrr_print_out_one_result(i); 784 785 ··· 789 790 gran_size <<= 10; 790 791 x86_setup_var_mtrrs(range, nr_range, chunk_size, gran_size); 791 792 set_var_mtrr_all(address_bits); 792 - printk(KERN_DEBUG "New variable MTRRs\n"); 793 + pr_debug("New variable MTRRs\n"); 793 794 print_out_mtrr_range_state(); 794 795 return 1; 795 796 } else { ··· 798 799 mtrr_print_out_one_result(i); 799 800 } 800 801 801 - printk(KERN_INFO "mtrr_cleanup: can not find optimal value\n"); 802 - printk(KERN_INFO "please specify mtrr_gran_size/mtrr_chunk_size\n"); 802 + pr_info("mtrr_cleanup: can not find optimal value\n"); 803 + pr_info("please specify mtrr_gran_size/mtrr_chunk_size\n"); 803 804 804 805 return 0; 805 806 } ··· 917 918 918 919 /* kvm/qemu doesn't have mtrr set right, don't trim them all: */ 919 920 if (!highest_pfn) { 920 - printk(KERN_INFO "CPU MTRRs all blank - virtualized system.\n"); 921 + pr_info("CPU MTRRs all blank - virtualized system.\n"); 921 922 return 0; 922 923 } 923 924 ··· 972 973 end_pfn); 973 974 974 975 if (total_trim_size) { 975 - pr_warning("WARNING: BIOS bug: CPU MTRRs don't cover all of memory, losing %lluMB of RAM.\n", total_trim_size >> 20); 976 + pr_warn("WARNING: BIOS bug: CPU MTRRs don't cover all of memory, losing %lluMB of RAM.\n", 977 + total_trim_size >> 20); 976 978 977 979 if (!changed_by_mtrr_cleanup) 978 980 WARN_ON(1);
+11 -12
arch/x86/kernel/cpu/mtrr/generic.c
··· 55 55 56 56 rdmsr(MSR_K8_SYSCFG, lo, hi); 57 57 if (lo & K8_MTRRFIXRANGE_DRAM_MODIFY) { 58 - printk(KERN_ERR FW_WARN "MTRR: CPU %u: SYSCFG[MtrrFixDramModEn]" 58 + pr_err(FW_WARN "MTRR: CPU %u: SYSCFG[MtrrFixDramModEn]" 59 59 " not cleared by BIOS, clearing this bit\n", 60 60 smp_processor_id()); 61 61 lo &= ~K8_MTRRFIXRANGE_DRAM_MODIFY; ··· 501 501 if (!mask) 502 502 return; 503 503 if (mask & MTRR_CHANGE_MASK_FIXED) 504 - pr_warning("mtrr: your CPUs had inconsistent fixed MTRR settings\n"); 504 + pr_warn("mtrr: your CPUs had inconsistent fixed MTRR settings\n"); 505 505 if (mask & MTRR_CHANGE_MASK_VARIABLE) 506 - pr_warning("mtrr: your CPUs had inconsistent variable MTRR settings\n"); 506 + pr_warn("mtrr: your CPUs had inconsistent variable MTRR settings\n"); 507 507 if (mask & MTRR_CHANGE_MASK_DEFTYPE) 508 - pr_warning("mtrr: your CPUs had inconsistent MTRRdefType settings\n"); 508 + pr_warn("mtrr: your CPUs had inconsistent MTRRdefType settings\n"); 509 509 510 - printk(KERN_INFO "mtrr: probably your BIOS does not setup all CPUs.\n"); 511 - printk(KERN_INFO "mtrr: corrected configuration.\n"); 510 + pr_info("mtrr: probably your BIOS does not setup all CPUs.\n"); 511 + pr_info("mtrr: corrected configuration.\n"); 512 512 } 513 513 514 514 /* ··· 519 519 void mtrr_wrmsr(unsigned msr, unsigned a, unsigned b) 520 520 { 521 521 if (wrmsr_safe(msr, a, b) < 0) { 522 - printk(KERN_ERR 523 - "MTRR: CPU %u: Writing MSR %x to %x:%x failed\n", 522 + pr_err("MTRR: CPU %u: Writing MSR %x to %x:%x failed\n", 524 523 smp_processor_id(), msr, a, b); 525 524 } 526 525 } ··· 606 607 tmp |= ~((1ULL<<(hi - 1)) - 1); 607 608 608 609 if (tmp != mask) { 609 - printk(KERN_WARNING "mtrr: your BIOS has configured an incorrect mask, fixing it.\n"); 610 + pr_warn("mtrr: your BIOS has configured an incorrect mask, fixing it.\n"); 610 611 add_taint(TAINT_FIRMWARE_WORKAROUND, LOCKDEP_STILL_OK); 611 612 mask = tmp; 612 613 } ··· 857 858 boot_cpu_data.x86_model == 1 && 858 859 boot_cpu_data.x86_mask <= 7) { 859 860 if (base & ((1 << (22 - PAGE_SHIFT)) - 1)) { 860 - pr_warning("mtrr: base(0x%lx000) is not 4 MiB aligned\n", base); 861 + pr_warn("mtrr: base(0x%lx000) is not 4 MiB aligned\n", base); 861 862 return -EINVAL; 862 863 } 863 864 if (!(base + size < 0x70000 || base > 0x7003F) && 864 865 (type == MTRR_TYPE_WRCOMB 865 866 || type == MTRR_TYPE_WRBACK)) { 866 - pr_warning("mtrr: writable mtrr between 0x70000000 and 0x7003FFFF may hang the CPU.\n"); 867 + pr_warn("mtrr: writable mtrr between 0x70000000 and 0x7003FFFF may hang the CPU.\n"); 867 868 return -EINVAL; 868 869 } 869 870 } ··· 877 878 lbase = lbase >> 1, last = last >> 1) 878 879 ; 879 880 if (lbase != last) { 880 - pr_warning("mtrr: base(0x%lx000) is not aligned on a size(0x%lx000) boundary\n", base, size); 881 + pr_warn("mtrr: base(0x%lx000) is not aligned on a size(0x%lx000) boundary\n", base, size); 881 882 return -EINVAL; 882 883 } 883 884 return 0;
+10 -10
arch/x86/kernel/cpu/mtrr/main.c
··· 300 300 return error; 301 301 302 302 if (type >= MTRR_NUM_TYPES) { 303 - pr_warning("mtrr: type: %u invalid\n", type); 303 + pr_warn("mtrr: type: %u invalid\n", type); 304 304 return -EINVAL; 305 305 } 306 306 307 307 /* If the type is WC, check that this processor supports it */ 308 308 if ((type == MTRR_TYPE_WRCOMB) && !have_wrcomb()) { 309 - pr_warning("mtrr: your processor doesn't support write-combining\n"); 309 + pr_warn("mtrr: your processor doesn't support write-combining\n"); 310 310 return -ENOSYS; 311 311 } 312 312 313 313 if (!size) { 314 - pr_warning("mtrr: zero sized request\n"); 314 + pr_warn("mtrr: zero sized request\n"); 315 315 return -EINVAL; 316 316 } 317 317 318 318 if ((base | (base + size - 1)) >> 319 319 (boot_cpu_data.x86_phys_bits - PAGE_SHIFT)) { 320 - pr_warning("mtrr: base or size exceeds the MTRR width\n"); 320 + pr_warn("mtrr: base or size exceeds the MTRR width\n"); 321 321 return -EINVAL; 322 322 } 323 323 ··· 348 348 } else if (types_compatible(type, ltype)) 349 349 continue; 350 350 } 351 - pr_warning("mtrr: 0x%lx000,0x%lx000 overlaps existing" 351 + pr_warn("mtrr: 0x%lx000,0x%lx000 overlaps existing" 352 352 " 0x%lx000,0x%lx000\n", base, size, lbase, 353 353 lsize); 354 354 goto out; ··· 357 357 if (ltype != type) { 358 358 if (types_compatible(type, ltype)) 359 359 continue; 360 - pr_warning("mtrr: type mismatch for %lx000,%lx000 old: %s new: %s\n", 360 + pr_warn("mtrr: type mismatch for %lx000,%lx000 old: %s new: %s\n", 361 361 base, size, mtrr_attrib_to_str(ltype), 362 362 mtrr_attrib_to_str(type)); 363 363 goto out; ··· 395 395 static int mtrr_check(unsigned long base, unsigned long size) 396 396 { 397 397 if ((base & (PAGE_SIZE - 1)) || (size & (PAGE_SIZE - 1))) { 398 - pr_warning("mtrr: size and base must be multiples of 4 kiB\n"); 398 + pr_warn("mtrr: size and base must be multiples of 4 kiB\n"); 399 399 pr_debug("mtrr: size: 0x%lx base: 0x%lx\n", size, base); 400 400 dump_stack(); 401 401 return -1; ··· 493 493 } 494 494 } 495 495 if (reg >= max) { 496 - pr_warning("mtrr: register: %d too big\n", reg); 496 + pr_warn("mtrr: register: %d too big\n", reg); 497 497 goto out; 498 498 } 499 499 mtrr_if->get(reg, &lbase, &lsize, &ltype); 500 500 if (lsize < 1) { 501 - pr_warning("mtrr: MTRR %d not used\n", reg); 501 + pr_warn("mtrr: MTRR %d not used\n", reg); 502 502 goto out; 503 503 } 504 504 if (mtrr_usage_table[reg] < 1) { 505 - pr_warning("mtrr: reg: %d has count=0\n", reg); 505 + pr_warn("mtrr: reg: %d has count=0\n", reg); 506 506 goto out; 507 507 } 508 508 if (--mtrr_usage_table[reg] < 1)
+5 -4
arch/x86/kernel/cpu/perf_event.c
··· 254 254 * We still allow the PMU driver to operate: 255 255 */ 256 256 if (bios_fail) { 257 - printk(KERN_CONT "Broken BIOS detected, complain to your hardware vendor.\n"); 258 - printk(KERN_ERR FW_BUG "the BIOS has corrupted hw-PMU resources (MSR %x is %Lx)\n", reg_fail, val_fail); 257 + pr_cont("Broken BIOS detected, complain to your hardware vendor.\n"); 258 + pr_err(FW_BUG "the BIOS has corrupted hw-PMU resources (MSR %x is %Lx)\n", 259 + reg_fail, val_fail); 259 260 } 260 261 261 262 return true; 262 263 263 264 msr_fail: 264 - printk(KERN_CONT "Broken PMU hardware detected, using software events only.\n"); 265 - printk("%sFailed to access perfctr msr (MSR %x is %Lx)\n", 265 + pr_cont("Broken PMU hardware detected, using software events only.\n"); 266 + pr_info("%sFailed to access perfctr msr (MSR %x is %Lx)\n", 266 267 boot_cpu_has(X86_FEATURE_HYPERVISOR) ? KERN_INFO : KERN_ERR, 267 268 reg, val_new); 268 269
+5 -5
arch/x86/kernel/cpu/perf_event_amd_ibs.c
··· 670 670 perf_ibs_pmu_init(&perf_ibs_op, "ibs_op"); 671 671 672 672 register_nmi_handler(NMI_LOCAL, perf_ibs_nmi_handler, 0, "perf_ibs"); 673 - printk(KERN_INFO "perf: AMD IBS detected (0x%08x)\n", ibs_caps); 673 + pr_info("perf: AMD IBS detected (0x%08x)\n", ibs_caps); 674 674 675 675 return 0; 676 676 } ··· 774 774 pci_read_config_dword(cpu_cfg, IBSCTL, &value); 775 775 if (value != (ibs_eilvt_off | IBSCTL_LVT_OFFSET_VALID)) { 776 776 pci_dev_put(cpu_cfg); 777 - printk(KERN_DEBUG "Failed to setup IBS LVT offset, " 778 - "IBSCTL = 0x%08x\n", value); 777 + pr_debug("Failed to setup IBS LVT offset, IBSCTL = 0x%08x\n", 778 + value); 779 779 return -EINVAL; 780 780 } 781 781 } while (1); 782 782 783 783 if (!nodes) { 784 - printk(KERN_DEBUG "No CPU node configured for IBS\n"); 784 + pr_debug("No CPU node configured for IBS\n"); 785 785 return -ENODEV; 786 786 } 787 787 ··· 810 810 preempt_enable(); 811 811 812 812 if (offset == APIC_EILVT_NR_MAX) { 813 - printk(KERN_DEBUG "No EILVT entry available\n"); 813 + pr_debug("No EILVT entry available\n"); 814 814 return; 815 815 } 816 816
+2 -2
arch/x86/kernel/cpu/perf_event_amd_uncore.c
··· 536 536 if (ret) 537 537 goto fail_nb; 538 538 539 - printk(KERN_INFO "perf: AMD NB counters detected\n"); 539 + pr_info("perf: AMD NB counters detected\n"); 540 540 ret = 0; 541 541 } 542 542 ··· 550 550 if (ret) 551 551 goto fail_l2; 552 552 553 - printk(KERN_INFO "perf: AMD L2I counters detected\n"); 553 + pr_info("perf: AMD L2I counters detected\n"); 554 554 ret = 0; 555 555 } 556 556
+3 -3
arch/x86/kernel/cpu/perf_event_intel_ds.c
··· 1325 1325 1326 1326 switch (format) { 1327 1327 case 0: 1328 - printk(KERN_CONT "PEBS fmt0%c, ", pebs_type); 1328 + pr_cont("PEBS fmt0%c, ", pebs_type); 1329 1329 x86_pmu.pebs_record_size = sizeof(struct pebs_record_core); 1330 1330 x86_pmu.drain_pebs = intel_pmu_drain_pebs_core; 1331 1331 break; 1332 1332 1333 1333 case 1: 1334 - printk(KERN_CONT "PEBS fmt1%c, ", pebs_type); 1334 + pr_cont("PEBS fmt1%c, ", pebs_type); 1335 1335 x86_pmu.pebs_record_size = sizeof(struct pebs_record_nhm); 1336 1336 x86_pmu.drain_pebs = intel_pmu_drain_pebs_nhm; 1337 1337 break; ··· 1351 1351 break; 1352 1352 1353 1353 default: 1354 - printk(KERN_CONT "no PEBS fmt%d%c, ", format, pebs_type); 1354 + pr_cont("no PEBS fmt%d%c, ", format, pebs_type); 1355 1355 x86_pmu.pebs = 0; 1356 1356 } 1357 1357 }
+1 -1
arch/x86/kernel/cpu/rdrand.c
··· 51 51 for (i = 0; i < SANITY_CHECK_LOOPS; i++) { 52 52 if (!rdrand_long(&tmp)) { 53 53 clear_cpu_cap(c, X86_FEATURE_RDRAND); 54 - printk_once(KERN_WARNING "rdrand: disabled\n"); 54 + pr_warn_once("rdrand: disabled\n"); 55 55 return; 56 56 } 57 57 }
+2 -2
arch/x86/kernel/cpu/topology.c
··· 87 87 c->x86_max_cores = (core_level_siblings / smp_num_siblings); 88 88 89 89 if (!printed) { 90 - printk(KERN_INFO "CPU: Physical Processor ID: %d\n", 90 + pr_info("CPU: Physical Processor ID: %d\n", 91 91 c->phys_proc_id); 92 92 if (c->x86_max_cores > 1) 93 - printk(KERN_INFO "CPU: Processor Core ID: %d\n", 93 + pr_info("CPU: Processor Core ID: %d\n", 94 94 c->cpu_core_id); 95 95 printed = 1; 96 96 }
+4 -4
arch/x86/kernel/cpu/transmeta.c
··· 33 33 if (max >= 0x80860001) { 34 34 cpuid(0x80860001, &dummy, &cpu_rev, &cpu_freq, &cpu_flags); 35 35 if (cpu_rev != 0x02000000) { 36 - printk(KERN_INFO "CPU: Processor revision %u.%u.%u.%u, %u MHz\n", 36 + pr_info("CPU: Processor revision %u.%u.%u.%u, %u MHz\n", 37 37 (cpu_rev >> 24) & 0xff, 38 38 (cpu_rev >> 16) & 0xff, 39 39 (cpu_rev >> 8) & 0xff, ··· 44 44 if (max >= 0x80860002) { 45 45 cpuid(0x80860002, &new_cpu_rev, &cms_rev1, &cms_rev2, &dummy); 46 46 if (cpu_rev == 0x02000000) { 47 - printk(KERN_INFO "CPU: Processor revision %08X, %u MHz\n", 47 + pr_info("CPU: Processor revision %08X, %u MHz\n", 48 48 new_cpu_rev, cpu_freq); 49 49 } 50 - printk(KERN_INFO "CPU: Code Morphing Software revision %u.%u.%u-%u-%u\n", 50 + pr_info("CPU: Code Morphing Software revision %u.%u.%u-%u-%u\n", 51 51 (cms_rev1 >> 24) & 0xff, 52 52 (cms_rev1 >> 16) & 0xff, 53 53 (cms_rev1 >> 8) & 0xff, ··· 76 76 (void *)&cpu_info[56], 77 77 (void *)&cpu_info[60]); 78 78 cpu_info[64] = '\0'; 79 - printk(KERN_INFO "CPU: %s\n", cpu_info); 79 + pr_info("CPU: %s\n", cpu_info); 80 80 } 81 81 82 82 /* Unhide possibly hidden capability flags */
+2 -3
arch/x86/kernel/cpu/vmware.c
··· 62 62 tsc_hz = eax | (((uint64_t)ebx) << 32); 63 63 do_div(tsc_hz, 1000); 64 64 BUG_ON(tsc_hz >> 32); 65 - printk(KERN_INFO "TSC freq read from hypervisor : %lu.%03lu MHz\n", 65 + pr_info("TSC freq read from hypervisor : %lu.%03lu MHz\n", 66 66 (unsigned long) tsc_hz / 1000, 67 67 (unsigned long) tsc_hz % 1000); 68 68 ··· 84 84 if (ebx != UINT_MAX) 85 85 x86_platform.calibrate_tsc = vmware_get_tsc_khz; 86 86 else 87 - printk(KERN_WARNING 88 - "Failed to get TSC freq from the hypervisor\n"); 87 + pr_warn("Failed to get TSC freq from the hypervisor\n"); 89 88 } 90 89 91 90 /*