Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

Merge branch 'cpuinit_phase2' of git://git.kernel.org/pub/scm/linux/kernel/git/paulg/linux

Pull phase two of __cpuinit removal from Paul Gortmaker:
"With the __cpuinit infrastructure removed earlier, this group of
commits only removes the function/data tagging that was done with the
various (now no-op) __cpuinit related prefixes.

Now that the dust has settled with yesterday's v3.11-rc1, there
hopefully shouldn't be any new users leaking back in tree, but I think
we can leave the harmless no-op stubs there for a release as a
courtesy to those who still have out of tree stuff and weren't paying
attention.

Although the commits are against the recent tag to allow for minor
context refreshes for things like yesterday's v3.11-rc1~ slab content,
the patches have been largely unchanged for weeks, aside from such
trivial updates.

For detail junkies, the largely boring and mostly irrelevant history
of the patches can be viewed at:

http://git.kernel.org/cgit/linux/kernel/git/paulg/cpuinit-delete.git

If nothing else, I guess it does at least demonstrate the level of
involvement required to shepherd such a treewide change to completion.

This is the same repository of patches that has been applied to the
end of the daily linux-next branches for the past several weeks"

* 'cpuinit_phase2' of git://git.kernel.org/pub/scm/linux/kernel/git/paulg/linux: (28 commits)
block: delete __cpuinit usage from all block files
drivers: delete __cpuinit usage from all remaining drivers files
kernel: delete __cpuinit usage from all core kernel files
rcu: delete __cpuinit usage from all rcu files
net: delete __cpuinit usage from all net files
acpi: delete __cpuinit usage from all acpi files
hwmon: delete __cpuinit usage from all hwmon files
cpufreq: delete __cpuinit usage from all cpufreq files
clocksource+irqchip: delete __cpuinit usage from all related files
x86: delete __cpuinit usage from all x86 files
score: delete __cpuinit usage from all score files
xtensa: delete __cpuinit usage from all xtensa files
openrisc: delete __cpuinit usage from all openrisc files
m32r: delete __cpuinit usage from all m32r files
hexagon: delete __cpuinit usage from all hexagon files
frv: delete __cpuinit usage from all frv files
cris: delete __cpuinit usage from all cris files
metag: delete __cpuinit usage from all metag files
tile: delete __cpuinit usage from all tile files
sh: delete __cpuinit usage from all sh files
...

+1079 -1221
+3 -3
Documentation/cpu-hotplug.txt
··· 267 267 A: This is what you would need in your kernel code to receive notifications. 268 268 269 269 #include <linux/cpu.h> 270 - static int __cpuinit foobar_cpu_callback(struct notifier_block *nfb, 271 - unsigned long action, void *hcpu) 270 + static int foobar_cpu_callback(struct notifier_block *nfb, 271 + unsigned long action, void *hcpu) 272 272 { 273 273 unsigned int cpu = (unsigned long)hcpu; 274 274 ··· 285 285 return NOTIFY_OK; 286 286 } 287 287 288 - static struct notifier_block __cpuinitdata foobar_cpu_notifer = 288 + static struct notifier_block foobar_cpu_notifer = 289 289 { 290 290 .notifier_call = foobar_cpu_callback, 291 291 };
+5 -5
arch/alpha/kernel/smp.c
··· 116 116 /* 117 117 * Where secondaries begin a life of C. 118 118 */ 119 - void __cpuinit 119 + void 120 120 smp_callin(void) 121 121 { 122 122 int cpuid = hard_smp_processor_id(); ··· 194 194 * Send a message to a secondary's console. "START" is one such 195 195 * interesting message. ;-) 196 196 */ 197 - static void __cpuinit 197 + static void 198 198 send_secondary_console_msg(char *str, int cpuid) 199 199 { 200 200 struct percpu_struct *cpu; ··· 285 285 /* 286 286 * Convince the console to have a secondary cpu begin execution. 287 287 */ 288 - static int __cpuinit 288 + static int 289 289 secondary_cpu_start(int cpuid, struct task_struct *idle) 290 290 { 291 291 struct percpu_struct *cpu; ··· 356 356 /* 357 357 * Bring one cpu online. 358 358 */ 359 - static int __cpuinit 359 + static int 360 360 smp_boot_one_cpu(int cpuid, struct task_struct *idle) 361 361 { 362 362 unsigned long timeout; ··· 472 472 { 473 473 } 474 474 475 - int __cpuinit 475 + int 476 476 __cpu_up(unsigned int cpu, struct task_struct *tidle) 477 477 { 478 478 smp_boot_one_cpu(cpu, tidle);
+2 -2
arch/alpha/kernel/traps.c
··· 32 32 33 33 static int opDEC_fix; 34 34 35 - static void __cpuinit 35 + static void 36 36 opDEC_check(void) 37 37 { 38 38 __asm__ __volatile__ ( ··· 1059 1059 return; 1060 1060 } 1061 1061 1062 - void __cpuinit 1062 + void 1063 1063 trap_init(void) 1064 1064 { 1065 1065 /* Tell PAL-code what global pointer we want in the kernel. */
+2 -2
arch/arm/common/mcpm_platsmp.c
··· 19 19 #include <asm/smp.h> 20 20 #include <asm/smp_plat.h> 21 21 22 - static int __cpuinit mcpm_boot_secondary(unsigned int cpu, struct task_struct *idle) 22 + static int mcpm_boot_secondary(unsigned int cpu, struct task_struct *idle) 23 23 { 24 24 unsigned int mpidr, pcpu, pcluster, ret; 25 25 extern void secondary_startup(void); ··· 40 40 return 0; 41 41 } 42 42 43 - static void __cpuinit mcpm_secondary_init(unsigned int cpu) 43 + static void mcpm_secondary_init(unsigned int cpu) 44 44 { 45 45 mcpm_cpu_powered_up(); 46 46 }
+1 -1
arch/arm/include/asm/arch_timer.h
··· 89 89 return cval; 90 90 } 91 91 92 - static inline void __cpuinit arch_counter_set_user_access(void) 92 + static inline void arch_counter_set_user_access(void) 93 93 { 94 94 u32 cntkctl; 95 95
-1
arch/arm/kernel/head-common.S
··· 149 149 * r5 = proc_info pointer in physical address space 150 150 * r9 = cpuid (preserved) 151 151 */ 152 - __CPUINIT 153 152 __lookup_processor_type: 154 153 adr r3, __lookup_processor_type_data 155 154 ldmia r3, {r4 - r6}
-1
arch/arm/kernel/head-nommu.S
··· 87 87 ENDPROC(stext) 88 88 89 89 #ifdef CONFIG_SMP 90 - __CPUINIT 91 90 ENTRY(secondary_startup) 92 91 /* 93 92 * Common entry point for secondary CPUs.
-1
arch/arm/kernel/head.S
··· 343 343 .long __turn_mmu_on_end 344 344 345 345 #if defined(CONFIG_SMP) 346 - __CPUINIT 347 346 ENTRY(secondary_startup) 348 347 /* 349 348 * Common entry point for secondary CPUs.
+2 -2
arch/arm/kernel/hw_breakpoint.c
··· 1020 1020 cpumask_or(&debug_err_mask, &debug_err_mask, cpumask_of(cpu)); 1021 1021 } 1022 1022 1023 - static int __cpuinit dbg_reset_notify(struct notifier_block *self, 1023 + static int dbg_reset_notify(struct notifier_block *self, 1024 1024 unsigned long action, void *cpu) 1025 1025 { 1026 1026 if ((action & ~CPU_TASKS_FROZEN) == CPU_ONLINE) ··· 1029 1029 return NOTIFY_OK; 1030 1030 } 1031 1031 1032 - static struct notifier_block __cpuinitdata dbg_reset_nb = { 1032 + static struct notifier_block dbg_reset_nb = { 1033 1033 .notifier_call = dbg_reset_notify, 1034 1034 }; 1035 1035
+3 -3
arch/arm/kernel/perf_event_cpu.c
··· 157 157 * UNKNOWN at reset, the PMU must be explicitly reset to avoid reading 158 158 * junk values out of them. 159 159 */ 160 - static int __cpuinit cpu_pmu_notify(struct notifier_block *b, 161 - unsigned long action, void *hcpu) 160 + static int cpu_pmu_notify(struct notifier_block *b, unsigned long action, 161 + void *hcpu) 162 162 { 163 163 if ((action & ~CPU_TASKS_FROZEN) != CPU_STARTING) 164 164 return NOTIFY_DONE; ··· 171 171 return NOTIFY_OK; 172 172 } 173 173 174 - static struct notifier_block __cpuinitdata cpu_pmu_hotplug_notifier = { 174 + static struct notifier_block cpu_pmu_hotplug_notifier = { 175 175 .notifier_call = cpu_pmu_notify, 176 176 }; 177 177
+1 -2
arch/arm/kernel/psci_smp.c
··· 46 46 47 47 extern void secondary_startup(void); 48 48 49 - static int __cpuinit psci_boot_secondary(unsigned int cpu, 50 - struct task_struct *idle) 49 + static int psci_boot_secondary(unsigned int cpu, struct task_struct *idle) 51 50 { 52 51 if (psci_ops.cpu_on) 53 52 return psci_ops.cpu_on(cpu_logical_map(cpu),
+9 -9
arch/arm/kernel/smp.c
··· 58 58 * control for which core is the next to come out of the secondary 59 59 * boot "holding pen" 60 60 */ 61 - volatile int __cpuinitdata pen_release = -1; 61 + volatile int pen_release = -1; 62 62 63 63 enum ipi_msg_type { 64 64 IPI_WAKEUP, ··· 86 86 return pgdir >> ARCH_PGD_SHIFT; 87 87 } 88 88 89 - int __cpuinit __cpu_up(unsigned int cpu, struct task_struct *idle) 89 + int __cpu_up(unsigned int cpu, struct task_struct *idle) 90 90 { 91 91 int ret; 92 92 ··· 138 138 smp_ops.smp_init_cpus(); 139 139 } 140 140 141 - int __cpuinit boot_secondary(unsigned int cpu, struct task_struct *idle) 141 + int boot_secondary(unsigned int cpu, struct task_struct *idle) 142 142 { 143 143 if (smp_ops.smp_boot_secondary) 144 144 return smp_ops.smp_boot_secondary(cpu, idle); ··· 170 170 /* 171 171 * __cpu_disable runs on the processor to be shutdown. 172 172 */ 173 - int __cpuinit __cpu_disable(void) 173 + int __cpu_disable(void) 174 174 { 175 175 unsigned int cpu = smp_processor_id(); 176 176 int ret; ··· 216 216 * called on the thread which is asking for a CPU to be shutdown - 217 217 * waits until shutdown has completed, or it is timed out. 218 218 */ 219 - void __cpuinit __cpu_die(unsigned int cpu) 219 + void __cpu_die(unsigned int cpu) 220 220 { 221 221 if (!wait_for_completion_timeout(&cpu_died, msecs_to_jiffies(5000))) { 222 222 pr_err("CPU%u: cpu didn't die\n", cpu); ··· 306 306 * Called by both boot and secondaries to move global data into 307 307 * per-processor storage. 308 308 */ 309 - static void __cpuinit smp_store_cpu_info(unsigned int cpuid) 309 + static void smp_store_cpu_info(unsigned int cpuid) 310 310 { 311 311 struct cpuinfo_arm *cpu_info = &per_cpu(cpu_data, cpuid); 312 312 ··· 322 322 * This is the secondary CPU boot entry. We're using this CPUs 323 323 * idle thread stack, but a set of temporary page tables. 324 324 */ 325 - asmlinkage void __cpuinit secondary_start_kernel(void) 325 + asmlinkage void secondary_start_kernel(void) 326 326 { 327 327 struct mm_struct *mm = &init_mm; 328 328 unsigned int cpu; ··· 521 521 { 522 522 } 523 523 524 - static void __cpuinit broadcast_timer_setup(struct clock_event_device *evt) 524 + static void broadcast_timer_setup(struct clock_event_device *evt) 525 525 { 526 526 evt->name = "dummy_timer"; 527 527 evt->features = CLOCK_EVT_FEAT_ONESHOT | ··· 550 550 } 551 551 #endif 552 552 553 - static void __cpuinit percpu_timer_setup(void) 553 + static void percpu_timer_setup(void) 554 554 { 555 555 unsigned int cpu = smp_processor_id(); 556 556 struct clock_event_device *evt = &per_cpu(percpu_clockevent, cpu);
+3 -3
arch/arm/kernel/smp_twd.c
··· 187 187 188 188 #endif 189 189 190 - static void __cpuinit twd_calibrate_rate(void) 190 + static void twd_calibrate_rate(void) 191 191 { 192 192 unsigned long count; 193 193 u64 waitjiffies; ··· 265 265 /* 266 266 * Setup the local clock events for a CPU. 267 267 */ 268 - static int __cpuinit twd_timer_setup(struct clock_event_device *clk) 268 + static int twd_timer_setup(struct clock_event_device *clk) 269 269 { 270 270 struct clock_event_device **this_cpu_clk; 271 271 int cpu = smp_processor_id(); ··· 308 308 return 0; 309 309 } 310 310 311 - static struct local_timer_ops twd_lt_ops __cpuinitdata = { 311 + static struct local_timer_ops twd_lt_ops = { 312 312 .setup = twd_timer_setup, 313 313 .stop = twd_timer_stop, 314 314 };
+1 -1
arch/arm/lib/delay.c
··· 86 86 } 87 87 } 88 88 89 - unsigned long __cpuinit calibrate_delay_is_known(void) 89 + unsigned long calibrate_delay_is_known(void) 90 90 { 91 91 delay_calibrated = true; 92 92 return lpj_fine;
-2
arch/arm/mach-exynos/headsmp.S
··· 13 13 #include <linux/linkage.h> 14 14 #include <linux/init.h> 15 15 16 - __CPUINIT 17 - 18 16 /* 19 17 * exynos4 specific entry point for secondary CPUs. This provides 20 18 * a "holding pen" into which all secondary cores are held until we're
+2 -2
arch/arm/mach-exynos/platsmp.c
··· 75 75 76 76 static DEFINE_SPINLOCK(boot_lock); 77 77 78 - static void __cpuinit exynos_secondary_init(unsigned int cpu) 78 + static void exynos_secondary_init(unsigned int cpu) 79 79 { 80 80 /* 81 81 * let the primary processor know we're out of the ··· 90 90 spin_unlock(&boot_lock); 91 91 } 92 92 93 - static int __cpuinit exynos_boot_secondary(unsigned int cpu, struct task_struct *idle) 93 + static int exynos_boot_secondary(unsigned int cpu, struct task_struct *idle) 94 94 { 95 95 unsigned long timeout; 96 96 unsigned long phys_cpu = cpu_logical_map(cpu);
+1 -1
arch/arm/mach-highbank/platsmp.c
··· 24 24 25 25 extern void secondary_startup(void); 26 26 27 - static int __cpuinit highbank_boot_secondary(unsigned int cpu, struct task_struct *idle) 27 + static int highbank_boot_secondary(unsigned int cpu, struct task_struct *idle) 28 28 { 29 29 highbank_set_cpu_jump(cpu, secondary_startup); 30 30 arch_send_wakeup_ipi_mask(cpumask_of(cpu));
+1 -1
arch/arm/mach-imx/platsmp.c
··· 53 53 writel_relaxed(val, scu_base); 54 54 } 55 55 56 - static int __cpuinit imx_boot_secondary(unsigned int cpu, struct task_struct *idle) 56 + static int imx_boot_secondary(unsigned int cpu, struct task_struct *idle) 57 57 { 58 58 imx_set_cpu_jump(cpu, v7_secondary_startup); 59 59 imx_enable_cpu(cpu, true);
+1 -1
arch/arm/mach-keystone/platsmp.c
··· 21 21 22 22 #include "keystone.h" 23 23 24 - static int __cpuinit keystone_smp_boot_secondary(unsigned int cpu, 24 + static int keystone_smp_boot_secondary(unsigned int cpu, 25 25 struct task_struct *idle) 26 26 { 27 27 unsigned long start = virt_to_phys(&secondary_startup);
-2
arch/arm/mach-msm/headsmp.S
··· 11 11 #include <linux/linkage.h> 12 12 #include <linux/init.h> 13 13 14 - __CPUINIT 15 - 16 14 /* 17 15 * MSM specific entry point for secondary CPUs. This provides 18 16 * a "holding pen" into which all secondary cores are held until we're
+3 -3
arch/arm/mach-msm/platsmp.c
··· 38 38 return ((read_cpuid_id() >> 4) & 3) + 1; 39 39 } 40 40 41 - static void __cpuinit msm_secondary_init(unsigned int cpu) 41 + static void msm_secondary_init(unsigned int cpu) 42 42 { 43 43 /* 44 44 * let the primary processor know we're out of the ··· 54 54 spin_unlock(&boot_lock); 55 55 } 56 56 57 - static __cpuinit void prepare_cold_cpu(unsigned int cpu) 57 + static void prepare_cold_cpu(unsigned int cpu) 58 58 { 59 59 int ret; 60 60 ret = scm_set_boot_addr(virt_to_phys(msm_secondary_startup), ··· 73 73 "address\n"); 74 74 } 75 75 76 - static int __cpuinit msm_boot_secondary(unsigned int cpu, struct task_struct *idle) 76 + static int msm_boot_secondary(unsigned int cpu, struct task_struct *idle) 77 77 { 78 78 unsigned long timeout; 79 79 static int cold_boot_done;
+2 -2
arch/arm/mach-msm/timer.c
··· 139 139 }; 140 140 141 141 #ifdef CONFIG_LOCAL_TIMERS 142 - static int __cpuinit msm_local_timer_setup(struct clock_event_device *evt) 142 + static int msm_local_timer_setup(struct clock_event_device *evt) 143 143 { 144 144 /* Use existing clock_event for cpu 0 */ 145 145 if (!smp_processor_id()) ··· 164 164 disable_percpu_irq(evt->irq); 165 165 } 166 166 167 - static struct local_timer_ops msm_local_timer_ops __cpuinitdata = { 167 + static struct local_timer_ops msm_local_timer_ops = { 168 168 .setup = msm_local_timer_setup, 169 169 .stop = msm_local_timer_stop, 170 170 };
+1 -1
arch/arm/mach-mvebu/coherency.c
··· 28 28 #include <asm/cacheflush.h> 29 29 #include "armada-370-xp.h" 30 30 31 - unsigned long __cpuinitdata coherency_phys_base; 31 + unsigned long coherency_phys_base; 32 32 static void __iomem *coherency_base; 33 33 static void __iomem *coherency_cpu_base; 34 34
-2
arch/arm/mach-mvebu/headsmp.S
··· 21 21 #include <linux/linkage.h> 22 22 #include <linux/init.h> 23 23 24 - __CPUINIT 25 - 26 24 /* 27 25 * Armada XP specific entry point for secondary CPUs. 28 26 * We add the CPU to the coherency fabric and then jump to secondary
+2 -3
arch/arm/mach-mvebu/platsmp.c
··· 71 71 } 72 72 } 73 73 74 - static void __cpuinit armada_xp_secondary_init(unsigned int cpu) 74 + static void armada_xp_secondary_init(unsigned int cpu) 75 75 { 76 76 armada_xp_mpic_smp_cpu_init(); 77 77 } 78 78 79 - static int __cpuinit armada_xp_boot_secondary(unsigned int cpu, 80 - struct task_struct *idle) 79 + static int armada_xp_boot_secondary(unsigned int cpu, struct task_struct *idle) 81 80 { 82 81 pr_info("Booting CPU %d\n", cpu); 83 82
-2
arch/arm/mach-omap2/omap-headsmp.S
··· 20 20 21 21 #include "omap44xx.h" 22 22 23 - __CPUINIT 24 - 25 23 /* Physical address needed since MMU not enabled yet on secondary core */ 26 24 #define AUX_CORE_BOOT0_PA 0x48281800 27 25
+1 -1
arch/arm/mach-omap2/omap-mpuss-lowpower.c
··· 291 291 * @cpu : CPU ID 292 292 * @power_state: CPU low power state. 293 293 */ 294 - int __cpuinit omap4_hotplug_cpu(unsigned int cpu, unsigned int power_state) 294 + int omap4_hotplug_cpu(unsigned int cpu, unsigned int power_state) 295 295 { 296 296 struct omap4_cpu_pm_info *pm_info = &per_cpu(omap4_pm_info, cpu); 297 297 unsigned int cpu_state = 0;
+2 -2
arch/arm/mach-omap2/omap-smp.c
··· 51 51 return scu_base; 52 52 } 53 53 54 - static void __cpuinit omap4_secondary_init(unsigned int cpu) 54 + static void omap4_secondary_init(unsigned int cpu) 55 55 { 56 56 /* 57 57 * Configure ACTRL and enable NS SMP bit access on CPU1 on HS device. ··· 72 72 spin_unlock(&boot_lock); 73 73 } 74 74 75 - static int __cpuinit omap4_boot_secondary(unsigned int cpu, struct task_struct *idle) 75 + static int omap4_boot_secondary(unsigned int cpu, struct task_struct *idle) 76 76 { 77 77 static struct clockdomain *cpu1_clkdm; 78 78 static bool booted;
+2 -2
arch/arm/mach-omap2/omap-wakeupgen.c
··· 323 323 #endif 324 324 325 325 #ifdef CONFIG_HOTPLUG_CPU 326 - static int __cpuinit irq_cpu_hotplug_notify(struct notifier_block *self, 327 - unsigned long action, void *hcpu) 326 + static int irq_cpu_hotplug_notify(struct notifier_block *self, 327 + unsigned long action, void *hcpu) 328 328 { 329 329 unsigned int cpu = (unsigned int)hcpu; 330 330
-2
arch/arm/mach-prima2/headsmp.S
··· 9 9 #include <linux/linkage.h> 10 10 #include <linux/init.h> 11 11 12 - __CPUINIT 13 - 14 12 /* 15 13 * SIRFSOC specific entry point for secondary CPUs. This provides 16 14 * a "holding pen" into which all secondary cores are held until we're
+2 -2
arch/arm/mach-prima2/platsmp.c
··· 44 44 scu_base = (void __iomem *)SIRFSOC_VA(base); 45 45 } 46 46 47 - static void __cpuinit sirfsoc_secondary_init(unsigned int cpu) 47 + static void sirfsoc_secondary_init(unsigned int cpu) 48 48 { 49 49 /* 50 50 * let the primary processor know we're out of the ··· 65 65 {}, 66 66 }; 67 67 68 - static int __cpuinit sirfsoc_boot_secondary(unsigned int cpu, struct task_struct *idle) 68 + static int sirfsoc_boot_secondary(unsigned int cpu, struct task_struct *idle) 69 69 { 70 70 unsigned long timeout; 71 71 struct device_node *np;
-1
arch/arm/mach-shmobile/headsmp-scu.S
··· 23 23 #include <linux/init.h> 24 24 #include <asm/memory.h> 25 25 26 - __CPUINIT 27 26 /* 28 27 * Boot code for secondary CPUs. 29 28 *
-2
arch/arm/mach-shmobile/headsmp.S
··· 14 14 #include <linux/init.h> 15 15 #include <asm/memory.h> 16 16 17 - __CPUINIT 18 - 19 17 ENTRY(shmobile_invalidate_start) 20 18 bl v7_invalidate_l1 21 19 b secondary_startup
+1 -1
arch/arm/mach-shmobile/smp-emev2.c
··· 30 30 31 31 #define EMEV2_SCU_BASE 0x1e000000 32 32 33 - static int __cpuinit emev2_boot_secondary(unsigned int cpu, struct task_struct *idle) 33 + static int emev2_boot_secondary(unsigned int cpu, struct task_struct *idle) 34 34 { 35 35 arch_send_wakeup_ipi_mask(cpumask_of(cpu_logical_map(cpu))); 36 36 return 0;
+1 -1
arch/arm/mach-shmobile/smp-r8a7779.c
··· 81 81 return ret ? ret : 1; 82 82 } 83 83 84 - static int __cpuinit r8a7779_boot_secondary(unsigned int cpu, struct task_struct *idle) 84 + static int r8a7779_boot_secondary(unsigned int cpu, struct task_struct *idle) 85 85 { 86 86 struct r8a7779_pm_ch *ch = NULL; 87 87 int ret = -EIO;
+1 -1
arch/arm/mach-shmobile/smp-sh73a0.c
··· 48 48 } 49 49 #endif 50 50 51 - static int __cpuinit sh73a0_boot_secondary(unsigned int cpu, struct task_struct *idle) 51 + static int sh73a0_boot_secondary(unsigned int cpu, struct task_struct *idle) 52 52 { 53 53 cpu = cpu_logical_map(cpu); 54 54
-1
arch/arm/mach-socfpga/headsmp.S
··· 10 10 #include <linux/linkage.h> 11 11 #include <linux/init.h> 12 12 13 - __CPUINIT 14 13 .arch armv7-a 15 14 16 15 ENTRY(secondary_trampoline)
+1 -1
arch/arm/mach-socfpga/platsmp.c
··· 29 29 30 30 #include "core.h" 31 31 32 - static int __cpuinit socfpga_boot_secondary(unsigned int cpu, struct task_struct *idle) 32 + static int socfpga_boot_secondary(unsigned int cpu, struct task_struct *idle) 33 33 { 34 34 int trampoline_size = &secondary_trampoline_end - &secondary_trampoline; 35 35
+1 -1
arch/arm/mach-spear/generic.h
··· 37 37 void spear_restart(enum reboot_mode, const char *); 38 38 39 39 void spear13xx_secondary_startup(void); 40 - void __cpuinit spear13xx_cpu_die(unsigned int cpu); 40 + void spear13xx_cpu_die(unsigned int cpu); 41 41 42 42 extern struct smp_operations spear13xx_smp_ops; 43 43
+2 -2
arch/arm/mach-spear/platsmp.c
··· 24 24 25 25 static void __iomem *scu_base = IOMEM(VA_SCU_BASE); 26 26 27 - static void __cpuinit spear13xx_secondary_init(unsigned int cpu) 27 + static void spear13xx_secondary_init(unsigned int cpu) 28 28 { 29 29 /* 30 30 * let the primary processor know we're out of the ··· 40 40 spin_unlock(&boot_lock); 41 41 } 42 42 43 - static int __cpuinit spear13xx_boot_secondary(unsigned int cpu, struct task_struct *idle) 43 + static int spear13xx_boot_secondary(unsigned int cpu, struct task_struct *idle) 44 44 { 45 45 unsigned long timeout; 46 46
+3 -3
arch/arm/mach-sti/platsmp.c
··· 27 27 28 28 #include "smp.h" 29 29 30 - static void __cpuinit write_pen_release(int val) 30 + static void write_pen_release(int val) 31 31 { 32 32 pen_release = val; 33 33 smp_wmb(); ··· 37 37 38 38 static DEFINE_SPINLOCK(boot_lock); 39 39 40 - void __cpuinit sti_secondary_init(unsigned int cpu) 40 + void sti_secondary_init(unsigned int cpu) 41 41 { 42 42 trace_hardirqs_off(); 43 43 ··· 54 54 spin_unlock(&boot_lock); 55 55 } 56 56 57 - int __cpuinit sti_boot_secondary(unsigned int cpu, struct task_struct *idle) 57 + int sti_boot_secondary(unsigned int cpu, struct task_struct *idle) 58 58 { 59 59 unsigned long timeout; 60 60
+2 -2
arch/arm/mach-tegra/platsmp.c
··· 35 35 36 36 static cpumask_t tegra_cpu_init_mask; 37 37 38 - static void __cpuinit tegra_secondary_init(unsigned int cpu) 38 + static void tegra_secondary_init(unsigned int cpu) 39 39 { 40 40 cpumask_set_cpu(cpu, &tegra_cpu_init_mask); 41 41 } ··· 167 167 return ret; 168 168 } 169 169 170 - static int __cpuinit tegra_boot_secondary(unsigned int cpu, 170 + static int tegra_boot_secondary(unsigned int cpu, 171 171 struct task_struct *idle) 172 172 { 173 173 if (IS_ENABLED(CONFIG_ARCH_TEGRA_2x_SOC) && tegra_chip_id == TEGRA20)
+1 -1
arch/arm/mach-tegra/pm.c
··· 191 191 [TEGRA_SUSPEND_LP0] = "LP0", 192 192 }; 193 193 194 - static int __cpuinit tegra_suspend_enter(suspend_state_t state) 194 + static int tegra_suspend_enter(suspend_state_t state) 195 195 { 196 196 enum tegra_suspend_mode mode = tegra_pmc_get_suspend_mode(); 197 197
+2 -2
arch/arm/mach-ux500/platsmp.c
··· 54 54 55 55 static DEFINE_SPINLOCK(boot_lock); 56 56 57 - static void __cpuinit ux500_secondary_init(unsigned int cpu) 57 + static void ux500_secondary_init(unsigned int cpu) 58 58 { 59 59 /* 60 60 * let the primary processor know we're out of the ··· 69 69 spin_unlock(&boot_lock); 70 70 } 71 71 72 - static int __cpuinit ux500_boot_secondary(unsigned int cpu, struct task_struct *idle) 72 + static int ux500_boot_secondary(unsigned int cpu, struct task_struct *idle) 73 73 { 74 74 unsigned long timeout; 75 75
+1 -1
arch/arm/mach-zynq/common.h
··· 27 27 extern char zynq_secondary_trampoline; 28 28 extern char zynq_secondary_trampoline_jump; 29 29 extern char zynq_secondary_trampoline_end; 30 - extern int __cpuinit zynq_cpun_start(u32 address, int cpu); 30 + extern int zynq_cpun_start(u32 address, int cpu); 31 31 extern struct smp_operations zynq_smp_ops __initdata; 32 32 #endif 33 33
-2
arch/arm/mach-zynq/headsmp.S
··· 9 9 #include <linux/linkage.h> 10 10 #include <linux/init.h> 11 11 12 - __CPUINIT 13 - 14 12 ENTRY(zynq_secondary_trampoline) 15 13 ldr r0, [pc] 16 14 bx r0
+3 -3
arch/arm/mach-zynq/platsmp.c
··· 30 30 /* 31 31 * Store number of cores in the system 32 32 * Because of scu_get_core_count() must be in __init section and can't 33 - * be called from zynq_cpun_start() because it is in __cpuinit section. 33 + * be called from zynq_cpun_start() because it is not in __init section. 34 34 */ 35 35 static int ncores; 36 36 37 - int __cpuinit zynq_cpun_start(u32 address, int cpu) 37 + int zynq_cpun_start(u32 address, int cpu) 38 38 { 39 39 u32 trampoline_code_size = &zynq_secondary_trampoline_end - 40 40 &zynq_secondary_trampoline; ··· 92 92 } 93 93 EXPORT_SYMBOL(zynq_cpun_start); 94 94 95 - static int __cpuinit zynq_boot_secondary(unsigned int cpu, 95 + static int zynq_boot_secondary(unsigned int cpu, 96 96 struct task_struct *idle) 97 97 { 98 98 return zynq_cpun_start(virt_to_phys(secondary_startup), cpu);
-2
arch/arm/mm/proc-arm1020.S
··· 443 443 #endif /* CONFIG_MMU */ 444 444 mov pc, lr 445 445 446 - __CPUINIT 447 - 448 446 .type __arm1020_setup, #function 449 447 __arm1020_setup: 450 448 mov r0, #0
-2
arch/arm/mm/proc-arm1020e.S
··· 425 425 #endif /* CONFIG_MMU */ 426 426 mov pc, lr 427 427 428 - __CPUINIT 429 - 430 428 .type __arm1020e_setup, #function 431 429 __arm1020e_setup: 432 430 mov r0, #0
-2
arch/arm/mm/proc-arm1022.S
··· 407 407 #endif /* CONFIG_MMU */ 408 408 mov pc, lr 409 409 410 - __CPUINIT 411 - 412 410 .type __arm1022_setup, #function 413 411 __arm1022_setup: 414 412 mov r0, #0
-3
arch/arm/mm/proc-arm1026.S
··· 396 396 #endif /* CONFIG_MMU */ 397 397 mov pc, lr 398 398 399 - 400 - __CPUINIT 401 - 402 399 .type __arm1026_setup, #function 403 400 __arm1026_setup: 404 401 mov r0, #0
-2
arch/arm/mm/proc-arm720.S
··· 116 116 ENDPROC(cpu_arm720_reset) 117 117 .popsection 118 118 119 - __CPUINIT 120 - 121 119 .type __arm710_setup, #function 122 120 __arm710_setup: 123 121 mov r0, #0
-2
arch/arm/mm/proc-arm740.S
··· 60 60 ENDPROC(cpu_arm740_reset) 61 61 .popsection 62 62 63 - __CPUINIT 64 - 65 63 .type __arm740_setup, #function 66 64 __arm740_setup: 67 65 mov r0, #0
-2
arch/arm/mm/proc-arm7tdmi.S
··· 51 51 ENDPROC(cpu_arm7tdmi_reset) 52 52 .popsection 53 53 54 - __CPUINIT 55 - 56 54 .type __arm7tdmi_setup, #function 57 55 __arm7tdmi_setup: 58 56 mov pc, lr
-2
arch/arm/mm/proc-arm920.S
··· 410 410 ENDPROC(cpu_arm920_do_resume) 411 411 #endif 412 412 413 - __CPUINIT 414 - 415 413 .type __arm920_setup, #function 416 414 __arm920_setup: 417 415 mov r0, #0
-2
arch/arm/mm/proc-arm922.S
··· 388 388 #endif /* CONFIG_MMU */ 389 389 mov pc, lr 390 390 391 - __CPUINIT 392 - 393 391 .type __arm922_setup, #function 394 392 __arm922_setup: 395 393 mov r0, #0
-2
arch/arm/mm/proc-arm925.S
··· 438 438 #endif /* CONFIG_MMU */ 439 439 mov pc, lr 440 440 441 - __CPUINIT 442 - 443 441 .type __arm925_setup, #function 444 442 __arm925_setup: 445 443 mov r0, #0
-2
arch/arm/mm/proc-arm926.S
··· 425 425 ENDPROC(cpu_arm926_do_resume) 426 426 #endif 427 427 428 - __CPUINIT 429 - 430 428 .type __arm926_setup, #function 431 429 __arm926_setup: 432 430 mov r0, #0
-2
arch/arm/mm/proc-arm940.S
··· 273 273 @ define struct cpu_cache_fns (see <asm/cacheflush.h> and proc-macros.S) 274 274 define_cache_functions arm940 275 275 276 - __CPUINIT 277 - 278 276 .type __arm940_setup, #function 279 277 __arm940_setup: 280 278 mov r0, #0
-2
arch/arm/mm/proc-arm946.S
··· 326 326 mcr p15, 0, r0, c7, c10, 4 @ drain WB 327 327 mov pc, lr 328 328 329 - __CPUINIT 330 - 331 329 .type __arm946_setup, #function 332 330 __arm946_setup: 333 331 mov r0, #0
-2
arch/arm/mm/proc-arm9tdmi.S
··· 51 51 ENDPROC(cpu_arm9tdmi_reset) 52 52 .popsection 53 53 54 - __CPUINIT 55 - 56 54 .type __arm9tdmi_setup, #function 57 55 __arm9tdmi_setup: 58 56 mov pc, lr
-2
arch/arm/mm/proc-fa526.S
··· 135 135 #endif 136 136 mov pc, lr 137 137 138 - __CPUINIT 139 - 140 138 .type __fa526_setup, #function 141 139 __fa526_setup: 142 140 /* On return of this routine, r0 must carry correct flags for CFG register */
-2
arch/arm/mm/proc-feroceon.S
··· 514 514 #endif 515 515 mov pc, lr 516 516 517 - __CPUINIT 518 - 519 517 .type __feroceon_setup, #function 520 518 __feroceon_setup: 521 519 mov r0, #0
-2
arch/arm/mm/proc-mohawk.S
··· 383 383 ENDPROC(cpu_mohawk_do_resume) 384 384 #endif 385 385 386 - __CPUINIT 387 - 388 386 .type __mohawk_setup, #function 389 387 __mohawk_setup: 390 388 mov r0, #0
-2
arch/arm/mm/proc-sa110.S
··· 159 159 #endif 160 160 mov pc, lr 161 161 162 - __CPUINIT 163 - 164 162 .type __sa110_setup, #function 165 163 __sa110_setup: 166 164 mov r10, #0
-2
arch/arm/mm/proc-sa1100.S
··· 198 198 ENDPROC(cpu_sa1100_do_resume) 199 199 #endif 200 200 201 - __CPUINIT 202 - 203 201 .type __sa1100_setup, #function 204 202 __sa1100_setup: 205 203 mov r0, #0
-2
arch/arm/mm/proc-v6.S
··· 180 180 181 181 .align 182 182 183 - __CPUINIT 184 - 185 183 /* 186 184 * __v6_setup 187 185 *
-4
arch/arm/mm/proc-v7-2level.S
··· 160 160 mcr p15, 0, \ttbr1, c2, c0, 1 @ load TTB1 161 161 .endm 162 162 163 - __CPUINIT 164 - 165 163 /* AT 166 164 * TFR EV X F I D LR S 167 165 * .EEE ..EE PUI. .T.T 4RVI ZWRS BLDP WCAM ··· 170 172 .type v7_crval, #object 171 173 v7_crval: 172 174 crval clear=0x2120c302, mmuset=0x10c03c7d, ucset=0x00c01c7c 173 - 174 - .previous
-4
arch/arm/mm/proc-v7-3level.S
··· 140 140 mcrr p15, 0, \ttbr0, \zero, c2 @ load TTBR0 141 141 .endm 142 142 143 - __CPUINIT 144 - 145 143 /* 146 144 * AT 147 145 * TFR EV X F IHD LR S ··· 151 153 .type v7_crval, #object 152 154 v7_crval: 153 155 crval clear=0x0120c302, mmuset=0x30c23c7d, ucset=0x00c01c7c 154 - 155 - .previous
-2
arch/arm/mm/proc-v7.S
··· 167 167 168 168 #endif 169 169 170 - __CPUINIT 171 - 172 170 /* 173 171 * __v7_setup 174 172 *
-2
arch/arm/mm/proc-xsc3.S
··· 446 446 ENDPROC(cpu_xsc3_do_resume) 447 447 #endif 448 448 449 - __CPUINIT 450 - 451 449 .type __xsc3_setup, #function 452 450 __xsc3_setup: 453 451 mov r0, #PSR_F_BIT|PSR_I_BIT|SVC_MODE
-2
arch/arm/mm/proc-xscale.S
··· 558 558 ENDPROC(cpu_xscale_do_resume) 559 559 #endif 560 560 561 - __CPUINIT 562 - 563 561 .type __xscale_setup, #function 564 562 __xscale_setup: 565 563 mcr p15, 0, ip, c7, c7, 0 @ invalidate I, D caches & BTB
+3 -3
arch/arm/plat-versatile/platsmp.c
··· 23 23 * observers, irrespective of whether they're taking part in coherency 24 24 * or not. This is necessary for the hotplug code to work reliably. 25 25 */ 26 - static void __cpuinit write_pen_release(int val) 26 + static void write_pen_release(int val) 27 27 { 28 28 pen_release = val; 29 29 smp_wmb(); ··· 33 33 34 34 static DEFINE_SPINLOCK(boot_lock); 35 35 36 - void __cpuinit versatile_secondary_init(unsigned int cpu) 36 + void versatile_secondary_init(unsigned int cpu) 37 37 { 38 38 /* 39 39 * let the primary processor know we're out of the ··· 48 48 spin_unlock(&boot_lock); 49 49 } 50 50 51 - int __cpuinit versatile_boot_secondary(unsigned int cpu, struct task_struct *idle) 51 + int versatile_boot_secondary(unsigned int cpu, struct task_struct *idle) 52 52 { 53 53 unsigned long timeout; 54 54
+1 -1
arch/arm64/include/asm/arch_timer.h
··· 97 97 return val; 98 98 } 99 99 100 - static inline void __cpuinit arch_counter_set_user_access(void) 100 + static inline void arch_counter_set_user_access(void) 101 101 { 102 102 u32 cntkctl; 103 103
+3 -3
arch/arm64/kernel/debug-monitors.c
··· 141 141 isb(); 142 142 } 143 143 144 - static int __cpuinit os_lock_notify(struct notifier_block *self, 144 + static int os_lock_notify(struct notifier_block *self, 145 145 unsigned long action, void *data) 146 146 { 147 147 int cpu = (unsigned long)data; ··· 150 150 return NOTIFY_OK; 151 151 } 152 152 153 - static struct notifier_block __cpuinitdata os_lock_nb = { 153 + static struct notifier_block os_lock_nb = { 154 154 .notifier_call = os_lock_notify, 155 155 }; 156 156 157 - static int __cpuinit debug_monitors_init(void) 157 + static int debug_monitors_init(void) 158 158 { 159 159 /* Clear the OS lock. */ 160 160 smp_call_function(clear_os_lock, NULL, 1);
+2 -2
arch/arm64/kernel/hw_breakpoint.c
··· 821 821 } 822 822 } 823 823 824 - static int __cpuinit hw_breakpoint_reset_notify(struct notifier_block *self, 824 + static int hw_breakpoint_reset_notify(struct notifier_block *self, 825 825 unsigned long action, 826 826 void *hcpu) 827 827 { ··· 831 831 return NOTIFY_OK; 832 832 } 833 833 834 - static struct notifier_block __cpuinitdata hw_breakpoint_reset_nb = { 834 + static struct notifier_block hw_breakpoint_reset_nb = { 835 835 .notifier_call = hw_breakpoint_reset_notify, 836 836 }; 837 837
+4 -4
arch/arm64/kernel/smp.c
··· 71 71 * in coherency or not. This is necessary for the hotplug code to work 72 72 * reliably. 73 73 */ 74 - static void __cpuinit write_pen_release(u64 val) 74 + static void write_pen_release(u64 val) 75 75 { 76 76 void *start = (void *)&secondary_holding_pen_release; 77 77 unsigned long size = sizeof(secondary_holding_pen_release); ··· 84 84 * Boot a secondary CPU, and assign it the specified idle task. 85 85 * This also gives us the initial stack to use for this CPU. 86 86 */ 87 - static int __cpuinit boot_secondary(unsigned int cpu, struct task_struct *idle) 87 + static int boot_secondary(unsigned int cpu, struct task_struct *idle) 88 88 { 89 89 unsigned long timeout; 90 90 ··· 122 122 123 123 static DECLARE_COMPLETION(cpu_running); 124 124 125 - int __cpuinit __cpu_up(unsigned int cpu, struct task_struct *idle) 125 + int __cpu_up(unsigned int cpu, struct task_struct *idle) 126 126 { 127 127 int ret; 128 128 ··· 162 162 * This is the secondary CPU boot entry. We're using this CPUs 163 163 * idle thread stack, but a set of temporary page tables. 164 164 */ 165 - asmlinkage void __cpuinit secondary_start_kernel(void) 165 + asmlinkage void secondary_start_kernel(void) 166 166 { 167 167 struct mm_struct *mm = &init_mm; 168 168 unsigned int cpu = smp_processor_id();
+1 -1
arch/blackfin/kernel/perf_event.c
··· 468 468 memset(cpuhw, 0, sizeof(struct cpu_hw_events)); 469 469 } 470 470 471 - static int __cpuinit 471 + static int 472 472 bfin_pmu_notifier(struct notifier_block *self, unsigned long action, void *hcpu) 473 473 { 474 474 unsigned int cpu = (long)hcpu;
+2 -2
arch/blackfin/kernel/setup.c
··· 99 99 } 100 100 #endif 101 101 102 - void __cpuinit bfin_setup_caches(unsigned int cpu) 102 + void bfin_setup_caches(unsigned int cpu) 103 103 { 104 104 #ifdef CONFIG_BFIN_ICACHE 105 105 bfin_icache_init(icplb_tbl[cpu]); ··· 165 165 #endif 166 166 } 167 167 168 - void __cpuinit bfin_setup_cpudata(unsigned int cpu) 168 + void bfin_setup_cpudata(unsigned int cpu) 169 169 { 170 170 struct blackfin_cpudata *cpudata = &per_cpu(cpu_data, cpu); 171 171
+3 -3
arch/blackfin/mach-bf561/smp.c
··· 48 48 return -EINVAL; 49 49 } 50 50 51 - void __cpuinit platform_secondary_init(unsigned int cpu) 51 + void platform_secondary_init(unsigned int cpu) 52 52 { 53 53 /* Clone setup for peripheral interrupt sources from CoreA. */ 54 54 bfin_write_SICB_IMASK0(bfin_read_SIC_IMASK0()); ··· 73 73 spin_unlock(&boot_lock); 74 74 } 75 75 76 - int __cpuinit platform_boot_secondary(unsigned int cpu, struct task_struct *idle) 76 + int platform_boot_secondary(unsigned int cpu, struct task_struct *idle) 77 77 { 78 78 unsigned long timeout; 79 79 ··· 154 154 * Setup core B's local core timer. 155 155 * In SMP, core timer is used for clock event device. 156 156 */ 157 - void __cpuinit bfin_local_timer_setup(void) 157 + void bfin_local_timer_setup(void) 158 158 { 159 159 #if defined(CONFIG_TICKSOURCE_CORETMR) 160 160 struct irq_data *data = irq_get_irq_data(IRQ_CORETMR);
+2 -2
arch/blackfin/mach-common/cache-c.c
··· 52 52 } 53 53 54 54 #ifdef CONFIG_BFIN_ICACHE 55 - void __cpuinit bfin_icache_init(struct cplb_entry *icplb_tbl) 55 + void bfin_icache_init(struct cplb_entry *icplb_tbl) 56 56 { 57 57 bfin_cache_init(icplb_tbl, ICPLB_ADDR0, ICPLB_DATA0, IMEM_CONTROL, 58 58 (IMC | ENICPLB)); ··· 60 60 #endif 61 61 62 62 #ifdef CONFIG_BFIN_DCACHE 63 - void __cpuinit bfin_dcache_init(struct cplb_entry *dcplb_tbl) 63 + void bfin_dcache_init(struct cplb_entry *dcplb_tbl) 64 64 { 65 65 /* 66 66 * Anomaly notes:
+1 -1
arch/blackfin/mach-common/ints-priority.c
··· 1281 1281 .irq_set_wake = bfin_gpio_set_wake, 1282 1282 }; 1283 1283 1284 - void __cpuinit init_exception_vectors(void) 1284 + void init_exception_vectors(void) 1285 1285 { 1286 1286 /* cannot program in software: 1287 1287 * evt0 - emulation (jtag)
+6 -6
arch/blackfin/mach-common/smp.c
··· 46 46 unsigned long blackfin_iflush_l1_entry[NR_CPUS]; 47 47 #endif 48 48 49 - struct blackfin_initial_pda __cpuinitdata initial_pda_coreb; 49 + struct blackfin_initial_pda initial_pda_coreb; 50 50 51 51 enum ipi_message_type { 52 52 BFIN_IPI_NONE, ··· 246 246 return; 247 247 } 248 248 249 - int __cpuinit __cpu_up(unsigned int cpu, struct task_struct *idle) 249 + int __cpu_up(unsigned int cpu, struct task_struct *idle) 250 250 { 251 251 int ret; 252 252 ··· 259 259 return ret; 260 260 } 261 261 262 - static void __cpuinit setup_secondary(unsigned int cpu) 262 + static void setup_secondary(unsigned int cpu) 263 263 { 264 264 unsigned long ilat; 265 265 ··· 277 277 IMASK_IVG10 | IMASK_IVG9 | IMASK_IVG8 | IMASK_IVG7 | IMASK_IVGHW; 278 278 } 279 279 280 - void __cpuinit secondary_start_kernel(void) 280 + void secondary_start_kernel(void) 281 281 { 282 282 unsigned int cpu = smp_processor_id(); 283 283 struct mm_struct *mm = &init_mm; ··· 402 402 #endif 403 403 404 404 #ifdef CONFIG_HOTPLUG_CPU 405 - int __cpuexit __cpu_disable(void) 405 + int __cpu_disable(void) 406 406 { 407 407 unsigned int cpu = smp_processor_id(); 408 408 ··· 415 415 416 416 static DECLARE_COMPLETION(cpu_killed); 417 417 418 - int __cpuexit __cpu_die(unsigned int cpu) 418 + int __cpu_die(unsigned int cpu) 419 419 { 420 420 return wait_for_completion_timeout(&cpu_killed, 5000); 421 421 }
+1 -1
arch/cris/arch-v32/kernel/smp.c
··· 197 197 */ 198 198 unsigned long cache_decay_ticks = 1; 199 199 200 - int __cpuinit __cpu_up(unsigned int cpu, struct task_struct *tidle) 200 + int __cpu_up(unsigned int cpu, struct task_struct *tidle) 201 201 { 202 202 smp_boot_one_cpu(cpu, tidle); 203 203 return cpu_online(cpu) ? 0 : -ENOSYS;
+1 -1
arch/frv/kernel/setup.c
··· 709 709 /* 710 710 * calibrate the delay loop 711 711 */ 712 - void __cpuinit calibrate_delay(void) 712 + void calibrate_delay(void) 713 713 { 714 714 loops_per_jiffy = __delay_loops_MHz * (1000000 / HZ); 715 715
+1 -1
arch/hexagon/kernel/setup.c
··· 41 41 42 42 int on_simulator; 43 43 44 - void __cpuinit calibrate_delay(void) 44 + void calibrate_delay(void) 45 45 { 46 46 loops_per_jiffy = thread_freq_mhz * 1000000 / HZ; 47 47 }
+2 -2
arch/hexagon/kernel/smp.c
··· 146 146 * to point to current thread info 147 147 */ 148 148 149 - void __cpuinit start_secondary(void) 149 + void start_secondary(void) 150 150 { 151 151 unsigned int cpu; 152 152 unsigned long thread_ptr; ··· 194 194 * maintains control until "cpu_online(cpu)" is set. 195 195 */ 196 196 197 - int __cpuinit __cpu_up(unsigned int cpu, struct task_struct *idle) 197 + int __cpu_up(unsigned int cpu, struct task_struct *idle) 198 198 { 199 199 struct thread_info *thread = (struct thread_info *)idle->stack; 200 200 void *stack_start;
+1 -1
arch/m32r/kernel/smpboot.c
··· 343 343 } 344 344 } 345 345 346 - int __cpuinit __cpu_up(unsigned int cpu_id, struct task_struct *tidle) 346 + int __cpu_up(unsigned int cpu_id, struct task_struct *tidle) 347 347 { 348 348 int timeout; 349 349
+3 -3
arch/metag/kernel/perf/perf_event.c
··· 813 813 }; 814 814 815 815 /* PMU CPU hotplug notifier */ 816 - static int __cpuinit metag_pmu_cpu_notify(struct notifier_block *b, 817 - unsigned long action, void *hcpu) 816 + static int metag_pmu_cpu_notify(struct notifier_block *b, unsigned long action, 817 + void *hcpu) 818 818 { 819 819 unsigned int cpu = (unsigned int)hcpu; 820 820 struct cpu_hw_events *cpuc = &per_cpu(cpu_hw_events, cpu); ··· 828 828 return NOTIFY_OK; 829 829 } 830 830 831 - static struct notifier_block __cpuinitdata metag_pmu_notifier = { 831 + static struct notifier_block metag_pmu_notifier = { 832 832 .notifier_call = metag_pmu_cpu_notify, 833 833 }; 834 834
+10 -12
arch/metag/kernel/smp.c
··· 68 68 /* 69 69 * "thread" is assumed to be a valid Meta hardware thread ID. 70 70 */ 71 - int __cpuinit boot_secondary(unsigned int thread, struct task_struct *idle) 71 + int boot_secondary(unsigned int thread, struct task_struct *idle) 72 72 { 73 73 u32 val; 74 74 ··· 118 118 * If the cache partition has changed, prints a message to the log describing 119 119 * those changes. 120 120 */ 121 - static __cpuinit void describe_cachepart_change(unsigned int thread, 122 - const char *label, 123 - unsigned int sz, 124 - unsigned int old, 125 - unsigned int new) 121 + static void describe_cachepart_change(unsigned int thread, const char *label, 122 + unsigned int sz, unsigned int old, 123 + unsigned int new) 126 124 { 127 125 unsigned int lor1, land1, gor1, gand1; 128 126 unsigned int lor2, land2, gor2, gand2; ··· 168 170 * Ensures that coherency is enabled and that the threads share the same cache 169 171 * partitions. 170 172 */ 171 - static __cpuinit void setup_smp_cache(unsigned int thread) 173 + static void setup_smp_cache(unsigned int thread) 172 174 { 173 175 unsigned int this_thread, lflags; 174 176 unsigned int dcsz, dcpart_this, dcpart_old, dcpart_new; ··· 213 215 icpart_old, icpart_new); 214 216 } 215 217 216 - int __cpuinit __cpu_up(unsigned int cpu, struct task_struct *idle) 218 + int __cpu_up(unsigned int cpu, struct task_struct *idle) 217 219 { 218 220 unsigned int thread = cpu_2_hwthread_id[cpu]; 219 221 int ret; ··· 266 268 /* 267 269 * __cpu_disable runs on the processor to be shutdown. 268 270 */ 269 - int __cpuexit __cpu_disable(void) 271 + int __cpu_disable(void) 270 272 { 271 273 unsigned int cpu = smp_processor_id(); 272 274 ··· 297 299 * called on the thread which is asking for a CPU to be shutdown - 298 300 * waits until shutdown has completed, or it is timed out. 299 301 */ 300 - void __cpuexit __cpu_die(unsigned int cpu) 302 + void __cpu_die(unsigned int cpu) 301 303 { 302 304 if (!wait_for_completion_timeout(&cpu_killed, msecs_to_jiffies(1))) 303 305 pr_err("CPU%u: unable to kill\n", cpu); ··· 309 311 * Note that we do not return from this function. If this cpu is 310 312 * brought online again it will need to run secondary_startup(). 311 313 */ 312 - void __cpuexit cpu_die(void) 314 + void cpu_die(void) 313 315 { 314 316 local_irq_disable(); 315 317 idle_task_exit(); ··· 324 326 * Called by both boot and secondaries to move global data into 325 327 * per-processor storage. 326 328 */ 327 - void __cpuinit smp_store_cpu_info(unsigned int cpuid) 329 + void smp_store_cpu_info(unsigned int cpuid) 328 330 { 329 331 struct cpuinfo_metag *cpu_info = &per_cpu(cpu_data, cpuid); 330 332
+1 -1
arch/metag/kernel/traps.c
··· 812 812 } 813 813 #endif 814 814 815 - void __cpuinit per_cpu_trap_init(unsigned long cpu) 815 + void per_cpu_trap_init(unsigned long cpu) 816 816 { 817 817 TBIRES int_context; 818 818 unsigned int thread = cpu_2_hwthread_id[cpu];
+1 -1
arch/mips/ath79/setup.c
··· 182 182 return ath79_sys_type; 183 183 } 184 184 185 - unsigned int __cpuinit get_c0_compare_int(void) 185 + unsigned int get_c0_compare_int(void) 186 186 { 187 187 return CP0_LEGACY_COMPARE_IRQ; 188 188 }
+6 -6
arch/mips/cavium-octeon/octeon-irq.c
··· 1095 1095 1096 1096 static bool octeon_irq_use_ip4; 1097 1097 1098 - static void __cpuinit octeon_irq_local_enable_ip4(void *arg) 1098 + static void octeon_irq_local_enable_ip4(void *arg) 1099 1099 { 1100 1100 set_c0_status(STATUSF_IP4); 1101 1101 } ··· 1110 1110 static void (*octeon_irq_ip3)(void); 1111 1111 static void (*octeon_irq_ip4)(void); 1112 1112 1113 - void __cpuinitdata (*octeon_irq_setup_secondary)(void); 1113 + void (*octeon_irq_setup_secondary)(void); 1114 1114 1115 - void __cpuinit octeon_irq_set_ip4_handler(octeon_irq_ip4_handler_t h) 1115 + void octeon_irq_set_ip4_handler(octeon_irq_ip4_handler_t h) 1116 1116 { 1117 1117 octeon_irq_ip4 = h; 1118 1118 octeon_irq_use_ip4 = true; 1119 1119 on_each_cpu(octeon_irq_local_enable_ip4, NULL, 1); 1120 1120 } 1121 1121 1122 - static void __cpuinit octeon_irq_percpu_enable(void) 1122 + static void octeon_irq_percpu_enable(void) 1123 1123 { 1124 1124 irq_cpu_online(); 1125 1125 } 1126 1126 1127 - static void __cpuinit octeon_irq_init_ciu_percpu(void) 1127 + static void octeon_irq_init_ciu_percpu(void) 1128 1128 { 1129 1129 int coreid = cvmx_get_core_num(); 1130 1130 ··· 1167 1167 cvmx_read_csr(CVMX_CIU2_SUM_PPX_IP2(coreid)); 1168 1168 } 1169 1169 1170 - static void __cpuinit octeon_irq_setup_secondary_ciu(void) 1170 + static void octeon_irq_setup_secondary_ciu(void) 1171 1171 { 1172 1172 octeon_irq_init_ciu_percpu(); 1173 1173 octeon_irq_percpu_enable();
+3 -3
arch/mips/cavium-octeon/smp.c
··· 173 173 * After we've done initial boot, this function is called to allow the 174 174 * board code to clean up state, if needed 175 175 */ 176 - static void __cpuinit octeon_init_secondary(void) 176 + static void octeon_init_secondary(void) 177 177 { 178 178 unsigned int sr; 179 179 ··· 375 375 return 0; 376 376 } 377 377 378 - static int __cpuinit octeon_cpu_callback(struct notifier_block *nfb, 378 + static int octeon_cpu_callback(struct notifier_block *nfb, 379 379 unsigned long action, void *hcpu) 380 380 { 381 381 unsigned int cpu = (unsigned long)hcpu; ··· 394 394 return NOTIFY_OK; 395 395 } 396 396 397 - static int __cpuinit register_cavium_notifier(void) 397 + static int register_cavium_notifier(void) 398 398 { 399 399 hotcpu_notifier(octeon_cpu_callback, 0); 400 400 return 0;
+13 -24
arch/mips/include/asm/uasm.h
··· 13 13 14 14 #ifdef CONFIG_EXPORT_UASM 15 15 #include <linux/export.h> 16 - #define __uasminit 17 - #define __uasminitdata 18 16 #define UASM_EXPORT_SYMBOL(sym) EXPORT_SYMBOL(sym) 19 17 #else 20 - #define __uasminit __cpuinit 21 - #define __uasminitdata __cpuinitdata 22 18 #define UASM_EXPORT_SYMBOL(sym) 23 19 #endif 24 20 ··· 50 54 #endif 51 55 52 56 #define Ip_u1u2u3(op) \ 53 - void __uasminit \ 54 - ISAOPC(op)(u32 **buf, unsigned int a, unsigned int b, unsigned int c) 57 + void ISAOPC(op)(u32 **buf, unsigned int a, unsigned int b, unsigned int c) 55 58 56 59 #define Ip_u2u1u3(op) \ 57 - void __uasminit \ 58 - ISAOPC(op)(u32 **buf, unsigned int a, unsigned int b, unsigned int c) 60 + void ISAOPC(op)(u32 **buf, unsigned int a, unsigned int b, unsigned int c) 59 61 60 62 #define Ip_u3u1u2(op) \ 61 - void __uasminit \ 62 - ISAOPC(op)(u32 **buf, unsigned int a, unsigned int b, unsigned int c) 63 + void ISAOPC(op)(u32 **buf, unsigned int a, unsigned int b, unsigned int c) 63 64 64 65 #define Ip_u1u2s3(op) \ 65 - void __uasminit \ 66 - ISAOPC(op)(u32 **buf, unsigned int a, unsigned int b, signed int c) 66 + void ISAOPC(op)(u32 **buf, unsigned int a, unsigned int b, signed int c) 67 67 68 68 #define Ip_u2s3u1(op) \ 69 - void __uasminit \ 70 - ISAOPC(op)(u32 **buf, unsigned int a, signed int b, unsigned int c) 69 + void ISAOPC(op)(u32 **buf, unsigned int a, signed int b, unsigned int c) 71 70 72 71 #define Ip_u2u1s3(op) \ 73 - void __uasminit \ 74 - ISAOPC(op)(u32 **buf, unsigned int a, unsigned int b, signed int c) 72 + void ISAOPC(op)(u32 **buf, unsigned int a, unsigned int b, signed int c) 75 73 76 74 #define Ip_u2u1msbu3(op) \ 77 - void __uasminit \ 78 - ISAOPC(op)(u32 **buf, unsigned int a, unsigned int b, unsigned int c, \ 75 + void ISAOPC(op)(u32 **buf, unsigned int a, unsigned int b, unsigned int c, \ 79 76 unsigned int d) 80 77 81 78 #define Ip_u1u2(op) \ 82 - void __uasminit ISAOPC(op)(u32 **buf, unsigned int a, unsigned int b) 79 + void ISAOPC(op)(u32 **buf, unsigned int a, unsigned int b) 83 80 84 81 #define Ip_u1s2(op) \ 85 - void __uasminit ISAOPC(op)(u32 **buf, unsigned int a, signed int b) 82 + void ISAOPC(op)(u32 **buf, unsigned int a, signed int b) 86 83 87 - #define Ip_u1(op) void __uasminit ISAOPC(op)(u32 **buf, unsigned int a) 84 + #define Ip_u1(op) void ISAOPC(op)(u32 **buf, unsigned int a) 88 85 89 - #define Ip_0(op) void __uasminit ISAOPC(op)(u32 **buf) 86 + #define Ip_0(op) void ISAOPC(op)(u32 **buf) 90 87 91 88 Ip_u2u1s3(_addiu); 92 89 Ip_u3u1u2(_addu); ··· 152 163 int lab; 153 164 }; 154 165 155 - void __uasminit ISAFUNC(uasm_build_label)(struct uasm_label **lab, u32 *addr, 166 + void ISAFUNC(uasm_build_label)(struct uasm_label **lab, u32 *addr, 156 167 int lid); 157 168 #ifdef CONFIG_64BIT 158 169 int ISAFUNC(uasm_in_compat_space_p)(long addr); ··· 163 174 void ISAFUNC(UASM_i_LA)(u32 **buf, unsigned int rs, long addr); 164 175 165 176 #define UASM_L_LA(lb) \ 166 - static inline void __uasminit ISAFUNC(uasm_l##lb)(struct uasm_label **lab, u32 *addr) \ 177 + static inline void ISAFUNC(uasm_l##lb)(struct uasm_label **lab, u32 *addr) \ 167 178 { \ 168 179 ISAFUNC(uasm_build_label)(lab, addr, label##lb); \ 169 180 }
-4
arch/mips/kernel/bmips_vec.S
··· 28 28 .set mips0 29 29 .endm 30 30 31 - __CPUINIT 32 - 33 31 /*********************************************************************** 34 32 * Alternate CPU1 startup vector for BMIPS4350 35 33 * ··· 213 215 * XKS01 support 214 216 * Certain CPUs support extending kseg0 to 1024MB. 215 217 ***********************************************************************/ 216 - 217 - __CPUINIT 218 218 219 219 LEAF(bmips_enable_xks01) 220 220
+1 -1
arch/mips/kernel/cevt-bcm1480.c
··· 109 109 static DEFINE_PER_CPU(struct irqaction, sibyte_hpt_irqaction); 110 110 static DEFINE_PER_CPU(char [18], sibyte_hpt_name); 111 111 112 - void __cpuinit sb1480_clockevent_init(void) 112 + void sb1480_clockevent_init(void) 113 113 { 114 114 unsigned int cpu = smp_processor_id(); 115 115 unsigned int irq = K_BCM1480_INT_TIMER_0 + cpu;
+1 -1
arch/mips/kernel/cevt-gic.c
··· 59 59 { 60 60 } 61 61 62 - int __cpuinit gic_clockevent_init(void) 62 + int gic_clockevent_init(void) 63 63 { 64 64 unsigned int cpu = smp_processor_id(); 65 65 struct clock_event_device *cd;
+1 -1
arch/mips/kernel/cevt-r4k.c
··· 171 171 } 172 172 173 173 #ifndef CONFIG_MIPS_MT_SMTC 174 - int __cpuinit r4k_clockevent_init(void) 174 + int r4k_clockevent_init(void) 175 175 { 176 176 unsigned int cpu = smp_processor_id(); 177 177 struct clock_event_device *cd;
+1 -1
arch/mips/kernel/cevt-sb1250.c
··· 107 107 static DEFINE_PER_CPU(struct irqaction, sibyte_hpt_irqaction); 108 108 static DEFINE_PER_CPU(char [18], sibyte_hpt_name); 109 109 110 - void __cpuinit sb1250_clockevent_init(void) 110 + void sb1250_clockevent_init(void) 111 111 { 112 112 unsigned int cpu = smp_processor_id(); 113 113 unsigned int irq = K_INT_TIMER_0 + cpu;
+1 -1
arch/mips/kernel/cevt-smtc.c
··· 248 248 } 249 249 250 250 251 - int __cpuinit smtc_clockevent_init(void) 251 + int smtc_clockevent_init(void) 252 252 { 253 253 uint64_t mips_freq = mips_hpt_frequency; 254 254 unsigned int cpu = smp_processor_id();
+1 -1
arch/mips/kernel/cpu-bugs64.c
··· 168 168 panic(bug64hit, !R4000_WAR ? r4kwar : nowar); 169 169 } 170 170 171 - static volatile int daddi_ov __cpuinitdata; 171 + static volatile int daddi_ov; 172 172 173 173 asmlinkage void __init do_daddi_ov(struct pt_regs *regs) 174 174 {
+7 -7
arch/mips/kernel/cpu-probe.c
··· 27 27 #include <asm/spram.h> 28 28 #include <asm/uaccess.h> 29 29 30 - static int __cpuinitdata mips_fpu_disabled; 30 + static int mips_fpu_disabled; 31 31 32 32 static int __init fpu_disable(char *s) 33 33 { ··· 39 39 40 40 __setup("nofpu", fpu_disable); 41 41 42 - int __cpuinitdata mips_dsp_disabled; 42 + int mips_dsp_disabled; 43 43 44 44 static int __init dsp_disable(char *s) 45 45 { ··· 134 134 #endif 135 135 } 136 136 137 - static void __cpuinit set_isa(struct cpuinfo_mips *c, unsigned int isa) 137 + static void set_isa(struct cpuinfo_mips *c, unsigned int isa) 138 138 { 139 139 switch (isa) { 140 140 case MIPS_CPU_ISA_M64R2: ··· 159 159 } 160 160 } 161 161 162 - static char unknown_isa[] __cpuinitdata = KERN_ERR \ 162 + static char unknown_isa[] = KERN_ERR \ 163 163 "Unsupported ISA type, c0.config0: %d."; 164 164 165 165 static inline unsigned int decode_config0(struct cpuinfo_mips *c) ··· 290 290 return config4 & MIPS_CONF_M; 291 291 } 292 292 293 - static void __cpuinit decode_configs(struct cpuinfo_mips *c) 293 + static void decode_configs(struct cpuinfo_mips *c) 294 294 { 295 295 int ok; 296 296 ··· 962 962 const char *__cpu_name[NR_CPUS]; 963 963 const char *__elf_platform; 964 964 965 - __cpuinit void cpu_probe(void) 965 + void cpu_probe(void) 966 966 { 967 967 struct cpuinfo_mips *c = &current_cpu_data; 968 968 unsigned int cpu = smp_processor_id(); ··· 1047 1047 #endif 1048 1048 } 1049 1049 1050 - __cpuinit void cpu_report(void) 1050 + void cpu_report(void) 1051 1051 { 1052 1052 struct cpuinfo_mips *c = &current_cpu_data; 1053 1053
-4
arch/mips/kernel/head.S
··· 158 158 j start_kernel 159 159 END(kernel_entry) 160 160 161 - __CPUINIT 162 - 163 161 #ifdef CONFIG_SMP 164 162 /* 165 163 * SMP slave cpus entry point. Board specific code for bootstrap calls this ··· 186 188 j start_secondary 187 189 END(smp_bootstrap) 188 190 #endif /* CONFIG_SMP */ 189 - 190 - __FINIT
+3 -3
arch/mips/kernel/smp-bmips.c
··· 398 398 * UP BMIPS systems as well. 399 399 ***********************************************************************/ 400 400 401 - static void __cpuinit bmips_wr_vec(unsigned long dst, char *start, char *end) 401 + static void bmips_wr_vec(unsigned long dst, char *start, char *end) 402 402 { 403 403 memcpy((void *)dst, start, end - start); 404 404 dma_cache_wback((unsigned long)start, end - start); ··· 406 406 instruction_hazard(); 407 407 } 408 408 409 - static inline void __cpuinit bmips_nmi_handler_setup(void) 409 + static inline void bmips_nmi_handler_setup(void) 410 410 { 411 411 bmips_wr_vec(BMIPS_NMI_RESET_VEC, &bmips_reset_nmi_vec, 412 412 &bmips_reset_nmi_vec_end); ··· 414 414 &bmips_smp_int_vec_end); 415 415 } 416 416 417 - void __cpuinit bmips_ebase_setup(void) 417 + void bmips_ebase_setup(void) 418 418 { 419 419 unsigned long new_ebase = ebase; 420 420 void __iomem __maybe_unused *cbr;
+3 -3
arch/mips/kernel/smp-mt.c
··· 149 149 vsmp_send_ipi_single(i, action); 150 150 } 151 151 152 - static void __cpuinit vsmp_init_secondary(void) 152 + static void vsmp_init_secondary(void) 153 153 { 154 154 #ifdef CONFIG_IRQ_GIC 155 155 /* This is Malta specific: IPI,performance and timer interrupts */ ··· 162 162 STATUSF_IP6 | STATUSF_IP7); 163 163 } 164 164 165 - static void __cpuinit vsmp_smp_finish(void) 165 + static void vsmp_smp_finish(void) 166 166 { 167 167 /* CDFIXME: remove this? */ 168 168 write_c0_compare(read_c0_count() + (8* mips_hpt_frequency/HZ)); ··· 188 188 * (unsigned long)idle->thread_info the gp 189 189 * assumes a 1:1 mapping of TC => VPE 190 190 */ 191 - static void __cpuinit vsmp_boot_secondary(int cpu, struct task_struct *idle) 191 + static void vsmp_boot_secondary(int cpu, struct task_struct *idle) 192 192 { 193 193 struct thread_info *gp = task_thread_info(idle); 194 194 dvpe();
+3 -3
arch/mips/kernel/smp-up.c
··· 28 28 * After we've done initial boot, this function is called to allow the 29 29 * board code to clean up state, if needed 30 30 */ 31 - static void __cpuinit up_init_secondary(void) 31 + static void up_init_secondary(void) 32 32 { 33 33 } 34 34 35 - static void __cpuinit up_smp_finish(void) 35 + static void up_smp_finish(void) 36 36 { 37 37 } 38 38 ··· 44 44 /* 45 45 * Firmware CPU startup hook 46 46 */ 47 - static void __cpuinit up_boot_secondary(int cpu, struct task_struct *idle) 47 + static void up_boot_secondary(int cpu, struct task_struct *idle) 48 48 { 49 49 } 50 50
+3 -3
arch/mips/kernel/smp.c
··· 86 86 struct plat_smp_ops *mp_ops; 87 87 EXPORT_SYMBOL(mp_ops); 88 88 89 - __cpuinit void register_smp_ops(struct plat_smp_ops *ops) 89 + void register_smp_ops(struct plat_smp_ops *ops) 90 90 { 91 91 if (mp_ops) 92 92 printk(KERN_WARNING "Overriding previously set SMP ops\n"); ··· 98 98 * First C code run on the secondary CPUs after being started up by 99 99 * the master. 100 100 */ 101 - asmlinkage __cpuinit void start_secondary(void) 101 + asmlinkage void start_secondary(void) 102 102 { 103 103 unsigned int cpu; 104 104 ··· 197 197 cpu_set(0, cpu_callin_map); 198 198 } 199 199 200 - int __cpuinit __cpu_up(unsigned int cpu, struct task_struct *tidle) 200 + int __cpu_up(unsigned int cpu, struct task_struct *tidle) 201 201 { 202 202 mp_ops->boot_secondary(cpu, tidle); 203 203
+1 -1
arch/mips/kernel/smtc.c
··· 645 645 * (unsigned long)idle->thread_info the gp 646 646 * 647 647 */ 648 - void __cpuinit smtc_boot_secondary(int cpu, struct task_struct *idle) 648 + void smtc_boot_secondary(int cpu, struct task_struct *idle) 649 649 { 650 650 extern u32 kernelsp[NR_CPUS]; 651 651 unsigned long flags;
+7 -7
arch/mips/kernel/spram.c
··· 37 37 /* 38 38 * Different semantics to the set_c0_* function built by __BUILD_SET_C0 39 39 */ 40 - static __cpuinit unsigned int bis_c0_errctl(unsigned int set) 40 + static unsigned int bis_c0_errctl(unsigned int set) 41 41 { 42 42 unsigned int res; 43 43 res = read_c0_errctl(); ··· 45 45 return res; 46 46 } 47 47 48 - static __cpuinit void ispram_store_tag(unsigned int offset, unsigned int data) 48 + static void ispram_store_tag(unsigned int offset, unsigned int data) 49 49 { 50 50 unsigned int errctl; 51 51 ··· 64 64 } 65 65 66 66 67 - static __cpuinit unsigned int ispram_load_tag(unsigned int offset) 67 + static unsigned int ispram_load_tag(unsigned int offset) 68 68 { 69 69 unsigned int data; 70 70 unsigned int errctl; ··· 82 82 return data; 83 83 } 84 84 85 - static __cpuinit void dspram_store_tag(unsigned int offset, unsigned int data) 85 + static void dspram_store_tag(unsigned int offset, unsigned int data) 86 86 { 87 87 unsigned int errctl; 88 88 ··· 98 98 } 99 99 100 100 101 - static __cpuinit unsigned int dspram_load_tag(unsigned int offset) 101 + static unsigned int dspram_load_tag(unsigned int offset) 102 102 { 103 103 unsigned int data; 104 104 unsigned int errctl; ··· 115 115 return data; 116 116 } 117 117 118 - static __cpuinit void probe_spram(char *type, 118 + static void probe_spram(char *type, 119 119 unsigned int base, 120 120 unsigned int (*read)(unsigned int), 121 121 void (*write)(unsigned int, unsigned int)) ··· 196 196 offset += 2 * SPRAM_TAG_STRIDE; 197 197 } 198 198 } 199 - void __cpuinit spram_config(void) 199 + void spram_config(void) 200 200 { 201 201 struct cpuinfo_mips *c = &current_cpu_data; 202 202 unsigned int config0;
+6 -6
arch/mips/kernel/sync-r4k.c
··· 20 20 #include <asm/barrier.h> 21 21 #include <asm/mipsregs.h> 22 22 23 - static atomic_t __cpuinitdata count_start_flag = ATOMIC_INIT(0); 24 - static atomic_t __cpuinitdata count_count_start = ATOMIC_INIT(0); 25 - static atomic_t __cpuinitdata count_count_stop = ATOMIC_INIT(0); 26 - static atomic_t __cpuinitdata count_reference = ATOMIC_INIT(0); 23 + static atomic_t count_start_flag = ATOMIC_INIT(0); 24 + static atomic_t count_count_start = ATOMIC_INIT(0); 25 + static atomic_t count_count_stop = ATOMIC_INIT(0); 26 + static atomic_t count_reference = ATOMIC_INIT(0); 27 27 28 28 #define COUNTON 100 29 29 #define NR_LOOPS 5 30 30 31 - void __cpuinit synchronise_count_master(int cpu) 31 + void synchronise_count_master(int cpu) 32 32 { 33 33 int i; 34 34 unsigned long flags; ··· 106 106 printk("done.\n"); 107 107 } 108 108 109 - void __cpuinit synchronise_count_slave(int cpu) 109 + void synchronise_count_slave(int cpu) 110 110 { 111 111 int i; 112 112 unsigned int initcount;
+6 -6
arch/mips/kernel/traps.c
··· 90 90 void (*board_ejtag_handler_setup)(void); 91 91 void (*board_bind_eic_interrupt)(int irq, int regset); 92 92 void (*board_ebase_setup)(void); 93 - void __cpuinitdata(*board_cache_error_setup)(void); 93 + void(*board_cache_error_setup)(void); 94 94 95 95 static void show_raw_backtrace(unsigned long reg29) 96 96 { ··· 1682 1682 int cp0_perfcount_irq; 1683 1683 EXPORT_SYMBOL_GPL(cp0_perfcount_irq); 1684 1684 1685 - static int __cpuinitdata noulri; 1685 + static int noulri; 1686 1686 1687 1687 static int __init ulri_disable(char *s) 1688 1688 { ··· 1693 1693 } 1694 1694 __setup("noulri", ulri_disable); 1695 1695 1696 - void __cpuinit per_cpu_trap_init(bool is_boot_cpu) 1696 + void per_cpu_trap_init(bool is_boot_cpu) 1697 1697 { 1698 1698 unsigned int cpu = smp_processor_id(); 1699 1699 unsigned int status_set = ST0_CU0; ··· 1810 1810 } 1811 1811 1812 1812 /* Install CPU exception handler */ 1813 - void __cpuinit set_handler(unsigned long offset, void *addr, unsigned long size) 1813 + void set_handler(unsigned long offset, void *addr, unsigned long size) 1814 1814 { 1815 1815 #ifdef CONFIG_CPU_MICROMIPS 1816 1816 memcpy((void *)(ebase + offset), ((unsigned char *)addr - 1), size); ··· 1820 1820 local_flush_icache_range(ebase + offset, ebase + offset + size); 1821 1821 } 1822 1822 1823 - static char panic_null_cerr[] __cpuinitdata = 1823 + static char panic_null_cerr[] = 1824 1824 "Trying to set NULL cache error exception handler"; 1825 1825 1826 1826 /* ··· 1828 1828 * This is suitable only for the cache error exception which is the only 1829 1829 * exception handler that is being run uncached. 1830 1830 */ 1831 - void __cpuinit set_uncached_handler(unsigned long offset, void *addr, 1831 + void set_uncached_handler(unsigned long offset, void *addr, 1832 1832 unsigned long size) 1833 1833 { 1834 1834 unsigned long uncached_ebase = CKSEG1ADDR(ebase);
+1 -1
arch/mips/kernel/watch.c
··· 100 100 } 101 101 } 102 102 103 - __cpuinit void mips_probe_watch_registers(struct cpuinfo_mips *c) 103 + void mips_probe_watch_registers(struct cpuinfo_mips *c) 104 104 { 105 105 unsigned int t; 106 106
+1 -1
arch/mips/lantiq/irq.c
··· 461 461 return 0; 462 462 } 463 463 464 - unsigned int __cpuinit get_c0_compare_int(void) 464 + unsigned int get_c0_compare_int(void) 465 465 { 466 466 return MIPS_CPU_TIMER_IRQ; 467 467 }
+1 -1
arch/mips/lib/uncached.c
··· 36 36 * values, so we can avoid sharing the same stack area between a cached 37 37 * and the uncached mode. 38 38 */ 39 - unsigned long __cpuinit run_uncached(void *func) 39 + unsigned long run_uncached(void *func) 40 40 { 41 41 register long sp __asm__("$sp"); 42 42 register long ret __asm__("$2");
+3 -3
arch/mips/mm/c-octeon.c
··· 180 180 * Probe Octeon's caches 181 181 * 182 182 */ 183 - static void __cpuinit probe_octeon(void) 183 + static void probe_octeon(void) 184 184 { 185 185 unsigned long icache_size; 186 186 unsigned long dcache_size; ··· 251 251 } 252 252 } 253 253 254 - static void __cpuinit octeon_cache_error_setup(void) 254 + static void octeon_cache_error_setup(void) 255 255 { 256 256 extern char except_vec2_octeon; 257 257 set_handler(0x100, &except_vec2_octeon, 0x80); ··· 261 261 * Setup the Octeon cache flush routines 262 262 * 263 263 */ 264 - void __cpuinit octeon_cache_init(void) 264 + void octeon_cache_init(void) 265 265 { 266 266 probe_octeon(); 267 267
+4 -4
arch/mips/mm/c-r3k.c
··· 26 26 static unsigned long icache_size, dcache_size; /* Size in bytes */ 27 27 static unsigned long icache_lsize, dcache_lsize; /* Size in bytes */ 28 28 29 - unsigned long __cpuinit r3k_cache_size(unsigned long ca_flags) 29 + unsigned long r3k_cache_size(unsigned long ca_flags) 30 30 { 31 31 unsigned long flags, status, dummy, size; 32 32 volatile unsigned long *p; ··· 61 61 return size * sizeof(*p); 62 62 } 63 63 64 - unsigned long __cpuinit r3k_cache_lsize(unsigned long ca_flags) 64 + unsigned long r3k_cache_lsize(unsigned long ca_flags) 65 65 { 66 66 unsigned long flags, status, lsize, i; 67 67 volatile unsigned long *p; ··· 90 90 return lsize * sizeof(*p); 91 91 } 92 92 93 - static void __cpuinit r3k_probe_cache(void) 93 + static void r3k_probe_cache(void) 94 94 { 95 95 dcache_size = r3k_cache_size(ST0_ISC); 96 96 if (dcache_size) ··· 312 312 r3k_flush_dcache_range(start, start + size); 313 313 } 314 314 315 - void __cpuinit r3k_cache_init(void) 315 + void r3k_cache_init(void) 316 316 { 317 317 extern void build_clear_page(void); 318 318 extern void build_copy_page(void);
+17 -17
arch/mips/mm/c-r4k.c
··· 107 107 blast_dcache64_page(addr); 108 108 } 109 109 110 - static void __cpuinit r4k_blast_dcache_page_setup(void) 110 + static void r4k_blast_dcache_page_setup(void) 111 111 { 112 112 unsigned long dc_lsize = cpu_dcache_line_size(); 113 113 ··· 123 123 124 124 static void (* r4k_blast_dcache_page_indexed)(unsigned long addr); 125 125 126 - static void __cpuinit r4k_blast_dcache_page_indexed_setup(void) 126 + static void r4k_blast_dcache_page_indexed_setup(void) 127 127 { 128 128 unsigned long dc_lsize = cpu_dcache_line_size(); 129 129 ··· 140 140 void (* r4k_blast_dcache)(void); 141 141 EXPORT_SYMBOL(r4k_blast_dcache); 142 142 143 - static void __cpuinit r4k_blast_dcache_setup(void) 143 + static void r4k_blast_dcache_setup(void) 144 144 { 145 145 unsigned long dc_lsize = cpu_dcache_line_size(); 146 146 ··· 227 227 228 228 static void (* r4k_blast_icache_page)(unsigned long addr); 229 229 230 - static void __cpuinit r4k_blast_icache_page_setup(void) 230 + static void r4k_blast_icache_page_setup(void) 231 231 { 232 232 unsigned long ic_lsize = cpu_icache_line_size(); 233 233 ··· 244 244 245 245 static void (* r4k_blast_icache_page_indexed)(unsigned long addr); 246 246 247 - static void __cpuinit r4k_blast_icache_page_indexed_setup(void) 247 + static void r4k_blast_icache_page_indexed_setup(void) 248 248 { 249 249 unsigned long ic_lsize = cpu_icache_line_size(); 250 250 ··· 269 269 void (* r4k_blast_icache)(void); 270 270 EXPORT_SYMBOL(r4k_blast_icache); 271 271 272 - static void __cpuinit r4k_blast_icache_setup(void) 272 + static void r4k_blast_icache_setup(void) 273 273 { 274 274 unsigned long ic_lsize = cpu_icache_line_size(); 275 275 ··· 290 290 291 291 static void (* r4k_blast_scache_page)(unsigned long addr); 292 292 293 - static void __cpuinit r4k_blast_scache_page_setup(void) 293 + static void r4k_blast_scache_page_setup(void) 294 294 { 295 295 unsigned long sc_lsize = cpu_scache_line_size(); 296 296 ··· 308 308 309 309 static void (* r4k_blast_scache_page_indexed)(unsigned long addr); 310 310 311 - static void __cpuinit r4k_blast_scache_page_indexed_setup(void) 311 + static void r4k_blast_scache_page_indexed_setup(void) 312 312 { 313 313 unsigned long sc_lsize = cpu_scache_line_size(); 314 314 ··· 326 326 327 327 static void (* r4k_blast_scache)(void); 328 328 329 - static void __cpuinit r4k_blast_scache_setup(void) 329 + static void r4k_blast_scache_setup(void) 330 330 { 331 331 unsigned long sc_lsize = cpu_scache_line_size(); 332 332 ··· 797 797 } 798 798 } 799 799 800 - static char *way_string[] __cpuinitdata = { NULL, "direct mapped", "2-way", 800 + static char *way_string[] = { NULL, "direct mapped", "2-way", 801 801 "3-way", "4-way", "5-way", "6-way", "7-way", "8-way" 802 802 }; 803 803 804 - static void __cpuinit probe_pcache(void) 804 + static void probe_pcache(void) 805 805 { 806 806 struct cpuinfo_mips *c = &current_cpu_data; 807 807 unsigned int config = read_c0_config(); ··· 1119 1119 * executes in KSEG1 space or else you will crash and burn badly. You have 1120 1120 * been warned. 1121 1121 */ 1122 - static int __cpuinit probe_scache(void) 1122 + static int probe_scache(void) 1123 1123 { 1124 1124 unsigned long flags, addr, begin, end, pow2; 1125 1125 unsigned int config = read_c0_config(); ··· 1196 1196 extern int rm7k_sc_init(void); 1197 1197 extern int mips_sc_init(void); 1198 1198 1199 - static void __cpuinit setup_scache(void) 1199 + static void setup_scache(void) 1200 1200 { 1201 1201 struct cpuinfo_mips *c = &current_cpu_data; 1202 1202 unsigned int config = read_c0_config(); ··· 1329 1329 NXP_BARRIER(); 1330 1330 } 1331 1331 1332 - static int __cpuinitdata cca = -1; 1332 + static int cca = -1; 1333 1333 1334 1334 static int __init cca_setup(char *str) 1335 1335 { ··· 1340 1340 1341 1341 early_param("cca", cca_setup); 1342 1342 1343 - static void __cpuinit coherency_setup(void) 1343 + static void coherency_setup(void) 1344 1344 { 1345 1345 if (cca < 0 || cca > 7) 1346 1346 cca = read_c0_config() & CONF_CM_CMASK; ··· 1380 1380 } 1381 1381 } 1382 1382 1383 - static void __cpuinit r4k_cache_error_setup(void) 1383 + static void r4k_cache_error_setup(void) 1384 1384 { 1385 1385 extern char __weak except_vec2_generic; 1386 1386 extern char __weak except_vec2_sb1; ··· 1398 1398 } 1399 1399 } 1400 1400 1401 - void __cpuinit r4k_cache_init(void) 1401 + void r4k_cache_init(void) 1402 1402 { 1403 1403 extern void build_clear_page(void); 1404 1404 extern void build_copy_page(void);
+1 -1
arch/mips/mm/c-tx39.c
··· 344 344 } 345 345 } 346 346 347 - void __cpuinit tx39_cache_init(void) 347 + void tx39_cache_init(void) 348 348 { 349 349 extern void build_clear_page(void); 350 350 extern void build_copy_page(void);
+1 -1
arch/mips/mm/cache.c
··· 182 182 } 183 183 } 184 184 185 - void __cpuinit cpu_cache_init(void) 185 + void cpu_cache_init(void) 186 186 { 187 187 if (cpu_has_3k_cache) { 188 188 extern void __weak r3k_cache_init(void);
-4
arch/mips/mm/cex-sb1.S
··· 49 49 * (0x170-0x17f) are used to preserve k0, k1, and ra. 50 50 */ 51 51 52 - __CPUINIT 53 - 54 52 LEAF(except_vec2_sb1) 55 53 /* 56 54 * If this error is recoverable, we need to exit the handler ··· 139 141 #endif 140 142 141 143 END(except_vec2_sb1) 142 - 143 - __FINIT 144 144 145 145 LEAF(handle_vec2_sb1) 146 146 mfc0 k0,CP0_CONFIG
+20 -20
arch/mips/mm/page.c
··· 66 66 UASM_L_LA(_copy_pref_store) 67 67 68 68 /* We need one branch and therefore one relocation per target label. */ 69 - static struct uasm_label __cpuinitdata labels[5]; 70 - static struct uasm_reloc __cpuinitdata relocs[5]; 69 + static struct uasm_label labels[5]; 70 + static struct uasm_reloc relocs[5]; 71 71 72 72 #define cpu_is_r4600_v1_x() ((read_c0_prid() & 0xfffffff0) == 0x00002010) 73 73 #define cpu_is_r4600_v2_x() ((read_c0_prid() & 0xfffffff0) == 0x00002020) 74 74 75 - static int pref_bias_clear_store __cpuinitdata; 76 - static int pref_bias_copy_load __cpuinitdata; 77 - static int pref_bias_copy_store __cpuinitdata; 75 + static int pref_bias_clear_store; 76 + static int pref_bias_copy_load; 77 + static int pref_bias_copy_store; 78 78 79 - static u32 pref_src_mode __cpuinitdata; 80 - static u32 pref_dst_mode __cpuinitdata; 79 + static u32 pref_src_mode; 80 + static u32 pref_dst_mode; 81 81 82 - static int clear_word_size __cpuinitdata; 83 - static int copy_word_size __cpuinitdata; 82 + static int clear_word_size; 83 + static int copy_word_size; 84 84 85 - static int half_clear_loop_size __cpuinitdata; 86 - static int half_copy_loop_size __cpuinitdata; 85 + static int half_clear_loop_size; 86 + static int half_copy_loop_size; 87 87 88 - static int cache_line_size __cpuinitdata; 88 + static int cache_line_size; 89 89 #define cache_line_mask() (cache_line_size - 1) 90 90 91 - static inline void __cpuinit 91 + static inline void 92 92 pg_addiu(u32 **buf, unsigned int reg1, unsigned int reg2, unsigned int off) 93 93 { 94 94 if (cpu_has_64bit_gp_regs && DADDI_WAR && r4k_daddiu_bug()) { ··· 108 108 } 109 109 } 110 110 111 - static void __cpuinit set_prefetch_parameters(void) 111 + static void set_prefetch_parameters(void) 112 112 { 113 113 if (cpu_has_64bit_gp_regs || cpu_has_64bit_zero_reg) 114 114 clear_word_size = 8; ··· 199 199 4 * copy_word_size)); 200 200 } 201 201 202 - static void __cpuinit build_clear_store(u32 **buf, int off) 202 + static void build_clear_store(u32 **buf, int off) 203 203 { 204 204 if (cpu_has_64bit_gp_regs || cpu_has_64bit_zero_reg) { 205 205 uasm_i_sd(buf, ZERO, off, A0); ··· 208 208 } 209 209 } 210 210 211 - static inline void __cpuinit build_clear_pref(u32 **buf, int off) 211 + static inline void build_clear_pref(u32 **buf, int off) 212 212 { 213 213 if (off & cache_line_mask()) 214 214 return; ··· 240 240 extern u32 __copy_page_start; 241 241 extern u32 __copy_page_end; 242 242 243 - void __cpuinit build_clear_page(void) 243 + void build_clear_page(void) 244 244 { 245 245 int off; 246 246 u32 *buf = &__clear_page_start; ··· 333 333 pr_debug("\t.set pop\n"); 334 334 } 335 335 336 - static void __cpuinit build_copy_load(u32 **buf, int reg, int off) 336 + static void build_copy_load(u32 **buf, int reg, int off) 337 337 { 338 338 if (cpu_has_64bit_gp_regs) { 339 339 uasm_i_ld(buf, reg, off, A1); ··· 342 342 } 343 343 } 344 344 345 - static void __cpuinit build_copy_store(u32 **buf, int reg, int off) 345 + static void build_copy_store(u32 **buf, int reg, int off) 346 346 { 347 347 if (cpu_has_64bit_gp_regs) { 348 348 uasm_i_sd(buf, reg, off, A0); ··· 387 387 } 388 388 } 389 389 390 - void __cpuinit build_copy_page(void) 390 + void build_copy_page(void) 391 391 { 392 392 int off; 393 393 u32 *buf = &__copy_page_start;
+1 -1
arch/mips/mm/sc-ip22.c
··· 167 167 .bc_inv = indy_sc_wback_invalidate 168 168 }; 169 169 170 - void __cpuinit indy_sc_init(void) 170 + void indy_sc_init(void) 171 171 { 172 172 if (indy_sc_probe()) { 173 173 indy_sc_enable();
+1 -1
arch/mips/mm/sc-mips.c
··· 132 132 return 1; 133 133 } 134 134 135 - int __cpuinit mips_sc_init(void) 135 + int mips_sc_init(void) 136 136 { 137 137 int found = mips_sc_probe(); 138 138 if (found) {
+1 -1
arch/mips/mm/sc-r5k.c
··· 98 98 .bc_inv = r5k_dma_cache_inv_sc 99 99 }; 100 100 101 - void __cpuinit r5k_sc_init(void) 101 + void r5k_sc_init(void) 102 102 { 103 103 if (r5k_sc_probe()) { 104 104 r5k_sc_enable();
+6 -6
arch/mips/mm/sc-rm7k.c
··· 104 104 /* 105 105 * This function is executed in uncached address space. 106 106 */ 107 - static __cpuinit void __rm7k_tc_enable(void) 107 + static void __rm7k_tc_enable(void) 108 108 { 109 109 int i; 110 110 ··· 117 117 cache_op(Index_Store_Tag_T, CKSEG0ADDR(i)); 118 118 } 119 119 120 - static __cpuinit void rm7k_tc_enable(void) 120 + static void rm7k_tc_enable(void) 121 121 { 122 122 if (read_c0_config() & RM7K_CONF_TE) 123 123 return; ··· 130 130 /* 131 131 * This function is executed in uncached address space. 132 132 */ 133 - static __cpuinit void __rm7k_sc_enable(void) 133 + static void __rm7k_sc_enable(void) 134 134 { 135 135 int i; 136 136 ··· 143 143 cache_op(Index_Store_Tag_SD, CKSEG0ADDR(i)); 144 144 } 145 145 146 - static __cpuinit void rm7k_sc_enable(void) 146 + static void rm7k_sc_enable(void) 147 147 { 148 148 if (read_c0_config() & RM7K_CONF_SE) 149 149 return; ··· 184 184 * This is a probing function like the one found in c-r4k.c, we look for the 185 185 * wrap around point with different addresses. 186 186 */ 187 - static __cpuinit void __probe_tcache(void) 187 + static void __probe_tcache(void) 188 188 { 189 189 unsigned long flags, addr, begin, end, pow2; 190 190 ··· 226 226 local_irq_restore(flags); 227 227 } 228 228 229 - void __cpuinit rm7k_sc_init(void) 229 + void rm7k_sc_init(void) 230 230 { 231 231 struct cpuinfo_mips *c = &current_cpu_data; 232 232 unsigned int config = read_c0_config();
+1 -1
arch/mips/mm/tlb-r3k.c
··· 276 276 } 277 277 } 278 278 279 - void __cpuinit tlb_init(void) 279 + void tlb_init(void) 280 280 { 281 281 local_flush_tlb_all(); 282 282
+2 -2
arch/mips/mm/tlb-r4k.c
··· 389 389 390 390 #endif /* CONFIG_TRANSPARENT_HUGEPAGE */ 391 391 392 - static int __cpuinitdata ntlb; 392 + static int ntlb; 393 393 static int __init set_ntlb(char *str) 394 394 { 395 395 get_option(&str, &ntlb); ··· 398 398 399 399 __setup("ntlb=", set_ntlb); 400 400 401 - void __cpuinit tlb_init(void) 401 + void tlb_init(void) 402 402 { 403 403 /* 404 404 * You should never change this register:
+2 -2
arch/mips/mm/tlb-r8k.c
··· 213 213 local_irq_restore(flags); 214 214 } 215 215 216 - static void __cpuinit probe_tlb(unsigned long config) 216 + static void probe_tlb(unsigned long config) 217 217 { 218 218 struct cpuinfo_mips *c = &current_cpu_data; 219 219 220 220 c->tlbsize = 3 * 128; /* 3 sets each 128 entries */ 221 221 } 222 222 223 - void __cpuinit tlb_init(void) 223 + void tlb_init(void) 224 224 { 225 225 unsigned int config = read_c0_config(); 226 226 unsigned long status;
+67 -77
arch/mips/mm/tlbex.c
··· 136 136 * why; it's not an issue caused by the core RTL. 137 137 * 138 138 */ 139 - static int __cpuinit m4kc_tlbp_war(void) 139 + static int m4kc_tlbp_war(void) 140 140 { 141 141 return (current_cpu_data.processor_id & 0xffff00) == 142 142 (PRID_COMP_MIPS | PRID_IMP_4KC); ··· 181 181 UASM_L_LA(_tlb_huge_update) 182 182 #endif 183 183 184 - static int __cpuinitdata hazard_instance; 184 + static int hazard_instance; 185 185 186 - static void __cpuinit uasm_bgezl_hazard(u32 **p, 187 - struct uasm_reloc **r, 188 - int instance) 186 + static void uasm_bgezl_hazard(u32 **p, struct uasm_reloc **r, int instance) 189 187 { 190 188 switch (instance) { 191 189 case 0 ... 7: ··· 194 196 } 195 197 } 196 198 197 - static void __cpuinit uasm_bgezl_label(struct uasm_label **l, 198 - u32 **p, 199 - int instance) 199 + static void uasm_bgezl_label(struct uasm_label **l, u32 **p, int instance) 200 200 { 201 201 switch (instance) { 202 202 case 0 ... 7: ··· 291 295 * We deliberately chose a buffer size of 128, so we won't scribble 292 296 * over anything important on overflow before we panic. 293 297 */ 294 - static u32 tlb_handler[128] __cpuinitdata; 298 + static u32 tlb_handler[128]; 295 299 296 300 /* simply assume worst case size for labels and relocs */ 297 - static struct uasm_label labels[128] __cpuinitdata; 298 - static struct uasm_reloc relocs[128] __cpuinitdata; 301 + static struct uasm_label labels[128]; 302 + static struct uasm_reloc relocs[128]; 299 303 300 - static int check_for_high_segbits __cpuinitdata; 304 + static int check_for_high_segbits; 301 305 302 - static unsigned int kscratch_used_mask __cpuinitdata; 306 + static unsigned int kscratch_used_mask; 303 307 304 308 static inline int __maybe_unused c0_kscratch(void) 305 309 { ··· 312 316 } 313 317 } 314 318 315 - static int __cpuinit allocate_kscratch(void) 319 + static int allocate_kscratch(void) 316 320 { 317 321 int r; 318 322 unsigned int a = cpu_data[0].kscratch_mask & ~kscratch_used_mask; ··· 329 333 return r; 330 334 } 331 335 332 - static int scratch_reg __cpuinitdata; 333 - static int pgd_reg __cpuinitdata; 336 + static int scratch_reg; 337 + static int pgd_reg; 334 338 enum vmalloc64_mode {not_refill, refill_scratch, refill_noscratch}; 335 339 336 - static struct work_registers __cpuinit build_get_work_registers(u32 **p) 340 + static struct work_registers build_get_work_registers(u32 **p) 337 341 { 338 342 struct work_registers r; 339 343 ··· 389 393 return r; 390 394 } 391 395 392 - static void __cpuinit build_restore_work_registers(u32 **p) 396 + static void build_restore_work_registers(u32 **p) 393 397 { 394 398 if (scratch_reg >= 0) { 395 399 UASM_i_MFC0(p, 1, c0_kscratch(), scratch_reg); ··· 414 418 /* 415 419 * The R3000 TLB handler is simple. 416 420 */ 417 - static void __cpuinit build_r3000_tlb_refill_handler(void) 421 + static void build_r3000_tlb_refill_handler(void) 418 422 { 419 423 long pgdc = (long)pgd_current; 420 424 u32 *p; ··· 459 463 * other one.To keep things simple, we first assume linear space, 460 464 * then we relocate it to the final handler layout as needed. 461 465 */ 462 - static u32 final_handler[64] __cpuinitdata; 466 + static u32 final_handler[64]; 463 467 464 468 /* 465 469 * Hazards ··· 483 487 * 484 488 * As if we MIPS hackers wouldn't know how to nop pipelines happy ... 485 489 */ 486 - static void __cpuinit __maybe_unused build_tlb_probe_entry(u32 **p) 490 + static void __maybe_unused build_tlb_probe_entry(u32 **p) 487 491 { 488 492 switch (current_cpu_type()) { 489 493 /* Found by experiment: R4600 v2.0/R4700 needs this, too. */ ··· 507 511 */ 508 512 enum tlb_write_entry { tlb_random, tlb_indexed }; 509 513 510 - static void __cpuinit build_tlb_write_entry(u32 **p, struct uasm_label **l, 511 - struct uasm_reloc **r, 512 - enum tlb_write_entry wmode) 514 + static void build_tlb_write_entry(u32 **p, struct uasm_label **l, 515 + struct uasm_reloc **r, 516 + enum tlb_write_entry wmode) 513 517 { 514 518 void(*tlbw)(u32 **) = NULL; 515 519 ··· 643 647 } 644 648 } 645 649 646 - static __cpuinit __maybe_unused void build_convert_pte_to_entrylo(u32 **p, 647 - unsigned int reg) 650 + static __maybe_unused void build_convert_pte_to_entrylo(u32 **p, 651 + unsigned int reg) 648 652 { 649 653 if (cpu_has_rixi) { 650 654 UASM_i_ROTR(p, reg, reg, ilog2(_PAGE_GLOBAL)); ··· 659 663 660 664 #ifdef CONFIG_MIPS_HUGE_TLB_SUPPORT 661 665 662 - static __cpuinit void build_restore_pagemask(u32 **p, 663 - struct uasm_reloc **r, 664 - unsigned int tmp, 665 - enum label_id lid, 666 - int restore_scratch) 666 + static void build_restore_pagemask(u32 **p, struct uasm_reloc **r, 667 + unsigned int tmp, enum label_id lid, 668 + int restore_scratch) 667 669 { 668 670 if (restore_scratch) { 669 671 /* Reset default page size */ ··· 700 706 } 701 707 } 702 708 703 - static __cpuinit void build_huge_tlb_write_entry(u32 **p, 704 - struct uasm_label **l, 705 - struct uasm_reloc **r, 706 - unsigned int tmp, 707 - enum tlb_write_entry wmode, 708 - int restore_scratch) 709 + static void build_huge_tlb_write_entry(u32 **p, struct uasm_label **l, 710 + struct uasm_reloc **r, 711 + unsigned int tmp, 712 + enum tlb_write_entry wmode, 713 + int restore_scratch) 709 714 { 710 715 /* Set huge page tlb entry size */ 711 716 uasm_i_lui(p, tmp, PM_HUGE_MASK >> 16); ··· 719 726 /* 720 727 * Check if Huge PTE is present, if so then jump to LABEL. 721 728 */ 722 - static void __cpuinit 729 + static void 723 730 build_is_huge_pte(u32 **p, struct uasm_reloc **r, unsigned int tmp, 724 - unsigned int pmd, int lid) 731 + unsigned int pmd, int lid) 725 732 { 726 733 UASM_i_LW(p, tmp, 0, pmd); 727 734 if (use_bbit_insns()) { ··· 732 739 } 733 740 } 734 741 735 - static __cpuinit void build_huge_update_entries(u32 **p, 736 - unsigned int pte, 737 - unsigned int tmp) 742 + static void build_huge_update_entries(u32 **p, unsigned int pte, 743 + unsigned int tmp) 738 744 { 739 745 int small_sequence; 740 746 ··· 763 771 UASM_i_MTC0(p, pte, C0_ENTRYLO1); /* load it */ 764 772 } 765 773 766 - static __cpuinit void build_huge_handler_tail(u32 **p, 767 - struct uasm_reloc **r, 768 - struct uasm_label **l, 769 - unsigned int pte, 770 - unsigned int ptr) 774 + static void build_huge_handler_tail(u32 **p, struct uasm_reloc **r, 775 + struct uasm_label **l, 776 + unsigned int pte, 777 + unsigned int ptr) 771 778 { 772 779 #ifdef CONFIG_SMP 773 780 UASM_i_SC(p, pte, 0, ptr); ··· 785 794 * TMP and PTR are scratch. 786 795 * TMP will be clobbered, PTR will hold the pmd entry. 787 796 */ 788 - static void __cpuinit 797 + static void 789 798 build_get_pmde64(u32 **p, struct uasm_label **l, struct uasm_reloc **r, 790 799 unsigned int tmp, unsigned int ptr) 791 800 { ··· 877 886 * BVADDR is the faulting address, PTR is scratch. 878 887 * PTR will hold the pgd for vmalloc. 879 888 */ 880 - static void __cpuinit 889 + static void 881 890 build_get_pgd_vmalloc64(u32 **p, struct uasm_label **l, struct uasm_reloc **r, 882 891 unsigned int bvaddr, unsigned int ptr, 883 892 enum vmalloc64_mode mode) ··· 947 956 * TMP and PTR are scratch. 948 957 * TMP will be clobbered, PTR will hold the pgd entry. 949 958 */ 950 - static void __cpuinit __maybe_unused 959 + static void __maybe_unused 951 960 build_get_pgde32(u32 **p, unsigned int tmp, unsigned int ptr) 952 961 { 953 962 long pgdc = (long)pgd_current; ··· 982 991 983 992 #endif /* !CONFIG_64BIT */ 984 993 985 - static void __cpuinit build_adjust_context(u32 **p, unsigned int ctx) 994 + static void build_adjust_context(u32 **p, unsigned int ctx) 986 995 { 987 996 unsigned int shift = 4 - (PTE_T_LOG2 + 1) + PAGE_SHIFT - 12; 988 997 unsigned int mask = (PTRS_PER_PTE / 2 - 1) << (PTE_T_LOG2 + 1); ··· 1008 1017 uasm_i_andi(p, ctx, ctx, mask); 1009 1018 } 1010 1019 1011 - static void __cpuinit build_get_ptep(u32 **p, unsigned int tmp, unsigned int ptr) 1020 + static void build_get_ptep(u32 **p, unsigned int tmp, unsigned int ptr) 1012 1021 { 1013 1022 /* 1014 1023 * Bug workaround for the Nevada. It seems as if under certain ··· 1033 1042 UASM_i_ADDU(p, ptr, ptr, tmp); /* add in offset */ 1034 1043 } 1035 1044 1036 - static void __cpuinit build_update_entries(u32 **p, unsigned int tmp, 1037 - unsigned int ptep) 1045 + static void build_update_entries(u32 **p, unsigned int tmp, unsigned int ptep) 1038 1046 { 1039 1047 /* 1040 1048 * 64bit address support (36bit on a 32bit CPU) in a 32bit ··· 1094 1104 int restore_scratch; 1095 1105 }; 1096 1106 1097 - static struct mips_huge_tlb_info __cpuinit 1107 + static struct mips_huge_tlb_info 1098 1108 build_fast_tlb_refill_handler (u32 **p, struct uasm_label **l, 1099 1109 struct uasm_reloc **r, unsigned int tmp, 1100 1110 unsigned int ptr, int c0_scratch_reg) ··· 1272 1282 */ 1273 1283 #define MIPS64_REFILL_INSNS 32 1274 1284 1275 - static void __cpuinit build_r4000_tlb_refill_handler(void) 1285 + static void build_r4000_tlb_refill_handler(void) 1276 1286 { 1277 1287 u32 *p = tlb_handler; 1278 1288 struct uasm_label *l = labels; ··· 1452 1462 #ifdef CONFIG_MIPS_PGD_C0_CONTEXT 1453 1463 extern u32 tlbmiss_handler_setup_pgd[], tlbmiss_handler_setup_pgd_end[]; 1454 1464 1455 - static void __cpuinit build_r4000_setup_pgd(void) 1465 + static void build_r4000_setup_pgd(void) 1456 1466 { 1457 1467 const int a0 = 4; 1458 1468 const int a1 = 5; ··· 1503 1513 } 1504 1514 #endif 1505 1515 1506 - static void __cpuinit 1516 + static void 1507 1517 iPTE_LW(u32 **p, unsigned int pte, unsigned int ptr) 1508 1518 { 1509 1519 #ifdef CONFIG_SMP ··· 1523 1533 #endif 1524 1534 } 1525 1535 1526 - static void __cpuinit 1536 + static void 1527 1537 iPTE_SW(u32 **p, struct uasm_reloc **r, unsigned int pte, unsigned int ptr, 1528 1538 unsigned int mode) 1529 1539 { ··· 1583 1593 * the page table where this PTE is located, PTE will be re-loaded 1584 1594 * with it's original value. 1585 1595 */ 1586 - static void __cpuinit 1596 + static void 1587 1597 build_pte_present(u32 **p, struct uasm_reloc **r, 1588 1598 int pte, int ptr, int scratch, enum label_id lid) 1589 1599 { ··· 1611 1621 } 1612 1622 1613 1623 /* Make PTE valid, store result in PTR. */ 1614 - static void __cpuinit 1624 + static void 1615 1625 build_make_valid(u32 **p, struct uasm_reloc **r, unsigned int pte, 1616 1626 unsigned int ptr) 1617 1627 { ··· 1624 1634 * Check if PTE can be written to, if not branch to LABEL. Regardless 1625 1635 * restore PTE with value from PTR when done. 1626 1636 */ 1627 - static void __cpuinit 1637 + static void 1628 1638 build_pte_writable(u32 **p, struct uasm_reloc **r, 1629 1639 unsigned int pte, unsigned int ptr, int scratch, 1630 1640 enum label_id lid) ··· 1644 1654 /* Make PTE writable, update software status bits as well, then store 1645 1655 * at PTR. 1646 1656 */ 1647 - static void __cpuinit 1657 + static void 1648 1658 build_make_write(u32 **p, struct uasm_reloc **r, unsigned int pte, 1649 1659 unsigned int ptr) 1650 1660 { ··· 1658 1668 * Check if PTE can be modified, if not branch to LABEL. Regardless 1659 1669 * restore PTE with value from PTR when done. 1660 1670 */ 1661 - static void __cpuinit 1671 + static void 1662 1672 build_pte_modifiable(u32 **p, struct uasm_reloc **r, 1663 1673 unsigned int pte, unsigned int ptr, int scratch, 1664 1674 enum label_id lid) ··· 1687 1697 * This places the pte into ENTRYLO0 and writes it with tlbwi. 1688 1698 * Then it returns. 1689 1699 */ 1690 - static void __cpuinit 1700 + static void 1691 1701 build_r3000_pte_reload_tlbwi(u32 **p, unsigned int pte, unsigned int tmp) 1692 1702 { 1693 1703 uasm_i_mtc0(p, pte, C0_ENTRYLO0); /* cp0 delay */ ··· 1703 1713 * may have the probe fail bit set as a result of a trap on a 1704 1714 * kseg2 access, i.e. without refill. Then it returns. 1705 1715 */ 1706 - static void __cpuinit 1716 + static void 1707 1717 build_r3000_tlb_reload_write(u32 **p, struct uasm_label **l, 1708 1718 struct uasm_reloc **r, unsigned int pte, 1709 1719 unsigned int tmp) ··· 1721 1731 uasm_i_rfe(p); /* branch delay */ 1722 1732 } 1723 1733 1724 - static void __cpuinit 1734 + static void 1725 1735 build_r3000_tlbchange_handler_head(u32 **p, unsigned int pte, 1726 1736 unsigned int ptr) 1727 1737 { ··· 1741 1751 uasm_i_tlbp(p); /* load delay */ 1742 1752 } 1743 1753 1744 - static void __cpuinit build_r3000_tlb_load_handler(void) 1754 + static void build_r3000_tlb_load_handler(void) 1745 1755 { 1746 1756 u32 *p = handle_tlbl; 1747 1757 const int handle_tlbl_size = handle_tlbl_end - handle_tlbl; ··· 1772 1782 dump_handler("r3000_tlb_load", handle_tlbl, handle_tlbl_size); 1773 1783 } 1774 1784 1775 - static void __cpuinit build_r3000_tlb_store_handler(void) 1785 + static void build_r3000_tlb_store_handler(void) 1776 1786 { 1777 1787 u32 *p = handle_tlbs; 1778 1788 const int handle_tlbs_size = handle_tlbs_end - handle_tlbs; ··· 1803 1813 dump_handler("r3000_tlb_store", handle_tlbs, handle_tlbs_size); 1804 1814 } 1805 1815 1806 - static void __cpuinit build_r3000_tlb_modify_handler(void) 1816 + static void build_r3000_tlb_modify_handler(void) 1807 1817 { 1808 1818 u32 *p = handle_tlbm; 1809 1819 const int handle_tlbm_size = handle_tlbm_end - handle_tlbm; ··· 1838 1848 /* 1839 1849 * R4000 style TLB load/store/modify handlers. 1840 1850 */ 1841 - static struct work_registers __cpuinit 1851 + static struct work_registers 1842 1852 build_r4000_tlbchange_handler_head(u32 **p, struct uasm_label **l, 1843 1853 struct uasm_reloc **r) 1844 1854 { ··· 1874 1884 return wr; 1875 1885 } 1876 1886 1877 - static void __cpuinit 1887 + static void 1878 1888 build_r4000_tlbchange_handler_tail(u32 **p, struct uasm_label **l, 1879 1889 struct uasm_reloc **r, unsigned int tmp, 1880 1890 unsigned int ptr) ··· 1892 1902 #endif 1893 1903 } 1894 1904 1895 - static void __cpuinit build_r4000_tlb_load_handler(void) 1905 + static void build_r4000_tlb_load_handler(void) 1896 1906 { 1897 1907 u32 *p = handle_tlbl; 1898 1908 const int handle_tlbl_size = handle_tlbl_end - handle_tlbl; ··· 2075 2085 dump_handler("r4000_tlb_load", handle_tlbl, handle_tlbl_size); 2076 2086 } 2077 2087 2078 - static void __cpuinit build_r4000_tlb_store_handler(void) 2088 + static void build_r4000_tlb_store_handler(void) 2079 2089 { 2080 2090 u32 *p = handle_tlbs; 2081 2091 const int handle_tlbs_size = handle_tlbs_end - handle_tlbs; ··· 2130 2140 dump_handler("r4000_tlb_store", handle_tlbs, handle_tlbs_size); 2131 2141 } 2132 2142 2133 - static void __cpuinit build_r4000_tlb_modify_handler(void) 2143 + static void build_r4000_tlb_modify_handler(void) 2134 2144 { 2135 2145 u32 *p = handle_tlbm; 2136 2146 const int handle_tlbm_size = handle_tlbm_end - handle_tlbm; ··· 2186 2196 dump_handler("r4000_tlb_modify", handle_tlbm, handle_tlbm_size); 2187 2197 } 2188 2198 2189 - static void __cpuinit flush_tlb_handlers(void) 2199 + static void flush_tlb_handlers(void) 2190 2200 { 2191 2201 local_flush_icache_range((unsigned long)handle_tlbl, 2192 2202 (unsigned long)handle_tlbl_end); ··· 2200 2210 #endif 2201 2211 } 2202 2212 2203 - void __cpuinit build_tlb_refill_handler(void) 2213 + void build_tlb_refill_handler(void) 2204 2214 { 2205 2215 /* 2206 2216 * The refill handler is generated per-CPU, multi-node systems
+5 -5
arch/mips/mm/uasm-micromips.c
··· 49 49 50 50 #include "uasm.c" 51 51 52 - static struct insn insn_table_MM[] __uasminitdata = { 52 + static struct insn insn_table_MM[] = { 53 53 { insn_addu, M(mm_pool32a_op, 0, 0, 0, 0, mm_addu32_op), RT | RS | RD }, 54 54 { insn_addiu, M(mm_addiu32_op, 0, 0, 0, 0, 0), RT | RS | SIMM }, 55 55 { insn_and, M(mm_pool32a_op, 0, 0, 0, 0, mm_and_op), RT | RS | RD }, ··· 118 118 119 119 #undef M 120 120 121 - static inline __uasminit u32 build_bimm(s32 arg) 121 + static inline u32 build_bimm(s32 arg) 122 122 { 123 123 WARN(arg > 0xffff || arg < -0x10000, 124 124 KERN_WARNING "Micro-assembler field overflow\n"); ··· 128 128 return ((arg < 0) ? (1 << 15) : 0) | ((arg >> 1) & 0x7fff); 129 129 } 130 130 131 - static inline __uasminit u32 build_jimm(u32 arg) 131 + static inline u32 build_jimm(u32 arg) 132 132 { 133 133 134 134 WARN(arg & ~((JIMM_MASK << 2) | 1), ··· 141 141 * The order of opcode arguments is implicitly left to right, 142 142 * starting with RS and ending with FUNC or IMM. 143 143 */ 144 - static void __uasminit build_insn(u32 **buf, enum opcode opc, ...) 144 + static void build_insn(u32 **buf, enum opcode opc, ...) 145 145 { 146 146 struct insn *ip = NULL; 147 147 unsigned int i; ··· 199 199 (*buf)++; 200 200 } 201 201 202 - static inline void __uasminit 202 + static inline void 203 203 __resolve_relocs(struct uasm_reloc *rel, struct uasm_label *lab) 204 204 { 205 205 long laddr = (long)lab->addr;
+5 -5
arch/mips/mm/uasm-mips.c
··· 49 49 50 50 #include "uasm.c" 51 51 52 - static struct insn insn_table[] __uasminitdata = { 52 + static struct insn insn_table[] = { 53 53 { insn_addiu, M(addiu_op, 0, 0, 0, 0, 0), RS | RT | SIMM }, 54 54 { insn_addu, M(spec_op, 0, 0, 0, 0, addu_op), RS | RT | RD }, 55 55 { insn_andi, M(andi_op, 0, 0, 0, 0, 0), RS | RT | UIMM }, ··· 119 119 120 120 #undef M 121 121 122 - static inline __uasminit u32 build_bimm(s32 arg) 122 + static inline u32 build_bimm(s32 arg) 123 123 { 124 124 WARN(arg > 0x1ffff || arg < -0x20000, 125 125 KERN_WARNING "Micro-assembler field overflow\n"); ··· 129 129 return ((arg < 0) ? (1 << 15) : 0) | ((arg >> 2) & 0x7fff); 130 130 } 131 131 132 - static inline __uasminit u32 build_jimm(u32 arg) 132 + static inline u32 build_jimm(u32 arg) 133 133 { 134 134 WARN(arg & ~(JIMM_MASK << 2), 135 135 KERN_WARNING "Micro-assembler field overflow\n"); ··· 141 141 * The order of opcode arguments is implicitly left to right, 142 142 * starting with RS and ending with FUNC or IMM. 143 143 */ 144 - static void __uasminit build_insn(u32 **buf, enum opcode opc, ...) 144 + static void build_insn(u32 **buf, enum opcode opc, ...) 145 145 { 146 146 struct insn *ip = NULL; 147 147 unsigned int i; ··· 187 187 (*buf)++; 188 188 } 189 189 190 - static inline void __uasminit 190 + static inline void 191 191 __resolve_relocs(struct uasm_reloc *rel, struct uasm_label *lab) 192 192 { 193 193 long laddr = (long)lab->addr;
+50 -56
arch/mips/mm/uasm.c
··· 63 63 enum fields fields; 64 64 }; 65 65 66 - static inline __uasminit u32 build_rs(u32 arg) 66 + static inline u32 build_rs(u32 arg) 67 67 { 68 68 WARN(arg & ~RS_MASK, KERN_WARNING "Micro-assembler field overflow\n"); 69 69 70 70 return (arg & RS_MASK) << RS_SH; 71 71 } 72 72 73 - static inline __uasminit u32 build_rt(u32 arg) 73 + static inline u32 build_rt(u32 arg) 74 74 { 75 75 WARN(arg & ~RT_MASK, KERN_WARNING "Micro-assembler field overflow\n"); 76 76 77 77 return (arg & RT_MASK) << RT_SH; 78 78 } 79 79 80 - static inline __uasminit u32 build_rd(u32 arg) 80 + static inline u32 build_rd(u32 arg) 81 81 { 82 82 WARN(arg & ~RD_MASK, KERN_WARNING "Micro-assembler field overflow\n"); 83 83 84 84 return (arg & RD_MASK) << RD_SH; 85 85 } 86 86 87 - static inline __uasminit u32 build_re(u32 arg) 87 + static inline u32 build_re(u32 arg) 88 88 { 89 89 WARN(arg & ~RE_MASK, KERN_WARNING "Micro-assembler field overflow\n"); 90 90 91 91 return (arg & RE_MASK) << RE_SH; 92 92 } 93 93 94 - static inline __uasminit u32 build_simm(s32 arg) 94 + static inline u32 build_simm(s32 arg) 95 95 { 96 96 WARN(arg > 0x7fff || arg < -0x8000, 97 97 KERN_WARNING "Micro-assembler field overflow\n"); ··· 99 99 return arg & 0xffff; 100 100 } 101 101 102 - static inline __uasminit u32 build_uimm(u32 arg) 102 + static inline u32 build_uimm(u32 arg) 103 103 { 104 104 WARN(arg & ~IMM_MASK, KERN_WARNING "Micro-assembler field overflow\n"); 105 105 106 106 return arg & IMM_MASK; 107 107 } 108 108 109 - static inline __uasminit u32 build_scimm(u32 arg) 109 + static inline u32 build_scimm(u32 arg) 110 110 { 111 111 WARN(arg & ~SCIMM_MASK, 112 112 KERN_WARNING "Micro-assembler field overflow\n"); ··· 114 114 return (arg & SCIMM_MASK) << SCIMM_SH; 115 115 } 116 116 117 - static inline __uasminit u32 build_func(u32 arg) 117 + static inline u32 build_func(u32 arg) 118 118 { 119 119 WARN(arg & ~FUNC_MASK, KERN_WARNING "Micro-assembler field overflow\n"); 120 120 121 121 return arg & FUNC_MASK; 122 122 } 123 123 124 - static inline __uasminit u32 build_set(u32 arg) 124 + static inline u32 build_set(u32 arg) 125 125 { 126 126 WARN(arg & ~SET_MASK, KERN_WARNING "Micro-assembler field overflow\n"); 127 127 128 128 return arg & SET_MASK; 129 129 } 130 130 131 - static void __uasminit build_insn(u32 **buf, enum opcode opc, ...); 131 + static void build_insn(u32 **buf, enum opcode opc, ...); 132 132 133 133 #define I_u1u2u3(op) \ 134 134 Ip_u1u2u3(op) \ ··· 286 286 287 287 #ifdef CONFIG_CPU_CAVIUM_OCTEON 288 288 #include <asm/octeon/octeon.h> 289 - void __uasminit ISAFUNC(uasm_i_pref)(u32 **buf, unsigned int a, signed int b, 289 + void ISAFUNC(uasm_i_pref)(u32 **buf, unsigned int a, signed int b, 290 290 unsigned int c) 291 291 { 292 292 if (OCTEON_IS_MODEL(OCTEON_CN63XX_PASS1_X) && a <= 24 && a != 5) ··· 304 304 #endif 305 305 306 306 /* Handle labels. */ 307 - void __uasminit ISAFUNC(uasm_build_label)(struct uasm_label **lab, u32 *addr, int lid) 307 + void ISAFUNC(uasm_build_label)(struct uasm_label **lab, u32 *addr, int lid) 308 308 { 309 309 (*lab)->addr = addr; 310 310 (*lab)->lab = lid; ··· 312 312 } 313 313 UASM_EXPORT_SYMBOL(ISAFUNC(uasm_build_label)); 314 314 315 - int __uasminit ISAFUNC(uasm_in_compat_space_p)(long addr) 315 + int ISAFUNC(uasm_in_compat_space_p)(long addr) 316 316 { 317 317 /* Is this address in 32bit compat space? */ 318 318 #ifdef CONFIG_64BIT ··· 323 323 } 324 324 UASM_EXPORT_SYMBOL(ISAFUNC(uasm_in_compat_space_p)); 325 325 326 - static int __uasminit uasm_rel_highest(long val) 326 + static int uasm_rel_highest(long val) 327 327 { 328 328 #ifdef CONFIG_64BIT 329 329 return ((((val + 0x800080008000L) >> 48) & 0xffff) ^ 0x8000) - 0x8000; ··· 332 332 #endif 333 333 } 334 334 335 - static int __uasminit uasm_rel_higher(long val) 335 + static int uasm_rel_higher(long val) 336 336 { 337 337 #ifdef CONFIG_64BIT 338 338 return ((((val + 0x80008000L) >> 32) & 0xffff) ^ 0x8000) - 0x8000; ··· 341 341 #endif 342 342 } 343 343 344 - int __uasminit ISAFUNC(uasm_rel_hi)(long val) 344 + int ISAFUNC(uasm_rel_hi)(long val) 345 345 { 346 346 return ((((val + 0x8000L) >> 16) & 0xffff) ^ 0x8000) - 0x8000; 347 347 } 348 348 UASM_EXPORT_SYMBOL(ISAFUNC(uasm_rel_hi)); 349 349 350 - int __uasminit ISAFUNC(uasm_rel_lo)(long val) 350 + int ISAFUNC(uasm_rel_lo)(long val) 351 351 { 352 352 return ((val & 0xffff) ^ 0x8000) - 0x8000; 353 353 } 354 354 UASM_EXPORT_SYMBOL(ISAFUNC(uasm_rel_lo)); 355 355 356 - void __uasminit ISAFUNC(UASM_i_LA_mostly)(u32 **buf, unsigned int rs, long addr) 356 + void ISAFUNC(UASM_i_LA_mostly)(u32 **buf, unsigned int rs, long addr) 357 357 { 358 358 if (!ISAFUNC(uasm_in_compat_space_p)(addr)) { 359 359 ISAFUNC(uasm_i_lui)(buf, rs, uasm_rel_highest(addr)); ··· 371 371 } 372 372 UASM_EXPORT_SYMBOL(ISAFUNC(UASM_i_LA_mostly)); 373 373 374 - void __uasminit ISAFUNC(UASM_i_LA)(u32 **buf, unsigned int rs, long addr) 374 + void ISAFUNC(UASM_i_LA)(u32 **buf, unsigned int rs, long addr) 375 375 { 376 376 ISAFUNC(UASM_i_LA_mostly)(buf, rs, addr); 377 377 if (ISAFUNC(uasm_rel_lo(addr))) { ··· 386 386 UASM_EXPORT_SYMBOL(ISAFUNC(UASM_i_LA)); 387 387 388 388 /* Handle relocations. */ 389 - void __uasminit 390 - ISAFUNC(uasm_r_mips_pc16)(struct uasm_reloc **rel, u32 *addr, int lid) 389 + void ISAFUNC(uasm_r_mips_pc16)(struct uasm_reloc **rel, u32 *addr, int lid) 391 390 { 392 391 (*rel)->addr = addr; 393 392 (*rel)->type = R_MIPS_PC16; ··· 395 396 } 396 397 UASM_EXPORT_SYMBOL(ISAFUNC(uasm_r_mips_pc16)); 397 398 398 - static inline void __uasminit 399 - __resolve_relocs(struct uasm_reloc *rel, struct uasm_label *lab); 399 + static inline void __resolve_relocs(struct uasm_reloc *rel, 400 + struct uasm_label *lab); 400 401 401 - void __uasminit 402 - ISAFUNC(uasm_resolve_relocs)(struct uasm_reloc *rel, struct uasm_label *lab) 402 + void ISAFUNC(uasm_resolve_relocs)(struct uasm_reloc *rel, 403 + struct uasm_label *lab) 403 404 { 404 405 struct uasm_label *l; 405 406 ··· 410 411 } 411 412 UASM_EXPORT_SYMBOL(ISAFUNC(uasm_resolve_relocs)); 412 413 413 - void __uasminit 414 - ISAFUNC(uasm_move_relocs)(struct uasm_reloc *rel, u32 *first, u32 *end, long off) 414 + void ISAFUNC(uasm_move_relocs)(struct uasm_reloc *rel, u32 *first, u32 *end, 415 + long off) 415 416 { 416 417 for (; rel->lab != UASM_LABEL_INVALID; rel++) 417 418 if (rel->addr >= first && rel->addr < end) ··· 419 420 } 420 421 UASM_EXPORT_SYMBOL(ISAFUNC(uasm_move_relocs)); 421 422 422 - void __uasminit 423 - ISAFUNC(uasm_move_labels)(struct uasm_label *lab, u32 *first, u32 *end, long off) 423 + void ISAFUNC(uasm_move_labels)(struct uasm_label *lab, u32 *first, u32 *end, 424 + long off) 424 425 { 425 426 for (; lab->lab != UASM_LABEL_INVALID; lab++) 426 427 if (lab->addr >= first && lab->addr < end) ··· 428 429 } 429 430 UASM_EXPORT_SYMBOL(ISAFUNC(uasm_move_labels)); 430 431 431 - void __uasminit 432 - ISAFUNC(uasm_copy_handler)(struct uasm_reloc *rel, struct uasm_label *lab, u32 *first, 433 - u32 *end, u32 *target) 432 + void ISAFUNC(uasm_copy_handler)(struct uasm_reloc *rel, struct uasm_label *lab, 433 + u32 *first, u32 *end, u32 *target) 434 434 { 435 435 long off = (long)(target - first); 436 436 ··· 440 442 } 441 443 UASM_EXPORT_SYMBOL(ISAFUNC(uasm_copy_handler)); 442 444 443 - int __uasminit ISAFUNC(uasm_insn_has_bdelay)(struct uasm_reloc *rel, u32 *addr) 445 + int ISAFUNC(uasm_insn_has_bdelay)(struct uasm_reloc *rel, u32 *addr) 444 446 { 445 447 for (; rel->lab != UASM_LABEL_INVALID; rel++) { 446 448 if (rel->addr == addr ··· 454 456 UASM_EXPORT_SYMBOL(ISAFUNC(uasm_insn_has_bdelay)); 455 457 456 458 /* Convenience functions for labeled branches. */ 457 - void __uasminit 458 - ISAFUNC(uasm_il_bltz)(u32 **p, struct uasm_reloc **r, unsigned int reg, int lid) 459 + void ISAFUNC(uasm_il_bltz)(u32 **p, struct uasm_reloc **r, unsigned int reg, 460 + int lid) 459 461 { 460 462 uasm_r_mips_pc16(r, *p, lid); 461 463 ISAFUNC(uasm_i_bltz)(p, reg, 0); 462 464 } 463 465 UASM_EXPORT_SYMBOL(ISAFUNC(uasm_il_bltz)); 464 466 465 - void __uasminit 466 - ISAFUNC(uasm_il_b)(u32 **p, struct uasm_reloc **r, int lid) 467 + void ISAFUNC(uasm_il_b)(u32 **p, struct uasm_reloc **r, int lid) 467 468 { 468 469 uasm_r_mips_pc16(r, *p, lid); 469 470 ISAFUNC(uasm_i_b)(p, 0); 470 471 } 471 472 UASM_EXPORT_SYMBOL(ISAFUNC(uasm_il_b)); 472 473 473 - void __uasminit 474 - ISAFUNC(uasm_il_beqz)(u32 **p, struct uasm_reloc **r, unsigned int reg, int lid) 474 + void ISAFUNC(uasm_il_beqz)(u32 **p, struct uasm_reloc **r, unsigned int reg, 475 + int lid) 475 476 { 476 477 uasm_r_mips_pc16(r, *p, lid); 477 478 ISAFUNC(uasm_i_beqz)(p, reg, 0); 478 479 } 479 480 UASM_EXPORT_SYMBOL(ISAFUNC(uasm_il_beqz)); 480 481 481 - void __uasminit 482 - ISAFUNC(uasm_il_beqzl)(u32 **p, struct uasm_reloc **r, unsigned int reg, int lid) 482 + void ISAFUNC(uasm_il_beqzl)(u32 **p, struct uasm_reloc **r, unsigned int reg, 483 + int lid) 483 484 { 484 485 uasm_r_mips_pc16(r, *p, lid); 485 486 ISAFUNC(uasm_i_beqzl)(p, reg, 0); 486 487 } 487 488 UASM_EXPORT_SYMBOL(ISAFUNC(uasm_il_beqzl)); 488 489 489 - void __uasminit 490 - ISAFUNC(uasm_il_bne)(u32 **p, struct uasm_reloc **r, unsigned int reg1, 491 - unsigned int reg2, int lid) 490 + void ISAFUNC(uasm_il_bne)(u32 **p, struct uasm_reloc **r, unsigned int reg1, 491 + unsigned int reg2, int lid) 492 492 { 493 493 uasm_r_mips_pc16(r, *p, lid); 494 494 ISAFUNC(uasm_i_bne)(p, reg1, reg2, 0); 495 495 } 496 496 UASM_EXPORT_SYMBOL(ISAFUNC(uasm_il_bne)); 497 497 498 - void __uasminit 499 - ISAFUNC(uasm_il_bnez)(u32 **p, struct uasm_reloc **r, unsigned int reg, int lid) 498 + void ISAFUNC(uasm_il_bnez)(u32 **p, struct uasm_reloc **r, unsigned int reg, 499 + int lid) 500 500 { 501 501 uasm_r_mips_pc16(r, *p, lid); 502 502 ISAFUNC(uasm_i_bnez)(p, reg, 0); 503 503 } 504 504 UASM_EXPORT_SYMBOL(ISAFUNC(uasm_il_bnez)); 505 505 506 - void __uasminit 507 - ISAFUNC(uasm_il_bgezl)(u32 **p, struct uasm_reloc **r, unsigned int reg, int lid) 506 + void ISAFUNC(uasm_il_bgezl)(u32 **p, struct uasm_reloc **r, unsigned int reg, 507 + int lid) 508 508 { 509 509 uasm_r_mips_pc16(r, *p, lid); 510 510 ISAFUNC(uasm_i_bgezl)(p, reg, 0); 511 511 } 512 512 UASM_EXPORT_SYMBOL(ISAFUNC(uasm_il_bgezl)); 513 513 514 - void __uasminit 515 - ISAFUNC(uasm_il_bgez)(u32 **p, struct uasm_reloc **r, unsigned int reg, int lid) 514 + void ISAFUNC(uasm_il_bgez)(u32 **p, struct uasm_reloc **r, unsigned int reg, 515 + int lid) 516 516 { 517 517 uasm_r_mips_pc16(r, *p, lid); 518 518 ISAFUNC(uasm_i_bgez)(p, reg, 0); 519 519 } 520 520 UASM_EXPORT_SYMBOL(ISAFUNC(uasm_il_bgez)); 521 521 522 - void __uasminit 523 - ISAFUNC(uasm_il_bbit0)(u32 **p, struct uasm_reloc **r, unsigned int reg, 524 - unsigned int bit, int lid) 522 + void ISAFUNC(uasm_il_bbit0)(u32 **p, struct uasm_reloc **r, unsigned int reg, 523 + unsigned int bit, int lid) 525 524 { 526 525 uasm_r_mips_pc16(r, *p, lid); 527 526 ISAFUNC(uasm_i_bbit0)(p, reg, bit, 0); 528 527 } 529 528 UASM_EXPORT_SYMBOL(ISAFUNC(uasm_il_bbit0)); 530 529 531 - void __uasminit 532 - ISAFUNC(uasm_il_bbit1)(u32 **p, struct uasm_reloc **r, unsigned int reg, 533 - unsigned int bit, int lid) 530 + void ISAFUNC(uasm_il_bbit1)(u32 **p, struct uasm_reloc **r, unsigned int reg, 531 + unsigned int bit, int lid) 534 532 { 535 533 uasm_r_mips_pc16(r, *p, lid); 536 534 ISAFUNC(uasm_i_bbit1)(p, reg, bit, 0);
+3 -3
arch/mips/mti-malta/malta-smtc.c
··· 32 32 /* 33 33 * Post-config but pre-boot cleanup entry point 34 34 */ 35 - static void __cpuinit msmtc_init_secondary(void) 35 + static void msmtc_init_secondary(void) 36 36 { 37 37 int myvpe; 38 38 ··· 53 53 /* 54 54 * Platform "CPU" startup hook 55 55 */ 56 - static void __cpuinit msmtc_boot_secondary(int cpu, struct task_struct *idle) 56 + static void msmtc_boot_secondary(int cpu, struct task_struct *idle) 57 57 { 58 58 smtc_boot_secondary(cpu, idle); 59 59 } ··· 61 61 /* 62 62 * SMP initialization finalization entry point 63 63 */ 64 - static void __cpuinit msmtc_smp_finish(void) 64 + static void msmtc_smp_finish(void) 65 65 { 66 66 smtc_smp_finish(); 67 67 }
+1 -1
arch/mips/mti-malta/malta-time.c
··· 150 150 } 151 151 } 152 152 153 - unsigned int __cpuinit get_c0_compare_int(void) 153 + unsigned int get_c0_compare_int(void) 154 154 { 155 155 #ifdef MSC01E_INT_BASE 156 156 if (cpu_has_veic) {
+1 -1
arch/mips/mti-sead3/sead3-time.c
··· 91 91 } 92 92 } 93 93 94 - unsigned int __cpuinit get_c0_compare_int(void) 94 + unsigned int get_c0_compare_int(void) 95 95 { 96 96 if (cpu_has_vint) 97 97 set_vi_handler(cp0_compare_irq, mips_timer_dispatch);
+2 -2
arch/mips/netlogic/common/smp.c
··· 116 116 /* 117 117 * Code to run on secondary just after probing the CPU 118 118 */ 119 - static void __cpuinit nlm_init_secondary(void) 119 + static void nlm_init_secondary(void) 120 120 { 121 121 int hwtid; 122 122 ··· 252 252 return 0; 253 253 } 254 254 255 - int __cpuinit nlm_wakeup_secondary_cpus(void) 255 + int nlm_wakeup_secondary_cpus(void) 256 256 { 257 257 u32 *reset_data; 258 258 int threadmode;
-4
arch/mips/netlogic/common/smpboot.S
··· 70 70 nop 71 71 /* not reached */ 72 72 73 - __CPUINIT 74 73 NESTED(nlm_boot_secondary_cpus, 16, sp) 75 74 /* Initialize CP0 Status */ 76 75 move t1, zero ··· 93 94 jr t0 94 95 nop 95 96 END(nlm_boot_secondary_cpus) 96 - __FINIT 97 97 98 98 /* 99 99 * In case of RMIboot bootloader which is used on XLR boards, the CPUs ··· 100 102 * This will get them out of the bootloader code and into linux. Needed 101 103 * because the bootloader area will be taken and initialized by linux. 102 104 */ 103 - __CPUINIT 104 105 NESTED(nlm_rmiboot_preboot, 16, sp) 105 106 mfc0 t0, $15, 1 /* read ebase */ 106 107 andi t0, 0x1f /* t0 has the processor_id() */ ··· 137 140 b 1b 138 141 nop 139 142 END(nlm_rmiboot_preboot) 140 - __FINIT
+1 -1
arch/mips/netlogic/common/time.c
··· 54 54 #error "Unknown CPU" 55 55 #endif 56 56 57 - unsigned int __cpuinit get_c0_compare_int(void) 57 + unsigned int get_c0_compare_int(void) 58 58 { 59 59 return IRQ_TIMER; 60 60 }
+1 -1
arch/mips/netlogic/xlr/wakeup.c
··· 49 49 #include <asm/netlogic/xlr/iomap.h> 50 50 #include <asm/netlogic/xlr/pic.h> 51 51 52 - int __cpuinit xlr_wakeup_secondary_cpus(void) 52 + int xlr_wakeup_secondary_cpus(void) 53 53 { 54 54 struct nlm_soc_info *nodep; 55 55 unsigned int i, j, boot_cpu;
+1 -1
arch/mips/pci/pci-ip27.c
··· 42 42 43 43 extern struct pci_ops bridge_pci_ops; 44 44 45 - int __cpuinit bridge_probe(nasid_t nasid, int widget_id, int masterwid) 45 + int bridge_probe(nasid_t nasid, int widget_id, int masterwid) 46 46 { 47 47 unsigned long offset = NODE_OFFSET(nasid); 48 48 struct bridge_controller *bc;
+3 -4
arch/mips/pmcs-msp71xx/msp_smtc.c
··· 33 33 /* 34 34 * Post-config but pre-boot cleanup entry point 35 35 */ 36 - static void __cpuinit msp_smtc_init_secondary(void) 36 + static void msp_smtc_init_secondary(void) 37 37 { 38 38 int myvpe; 39 39 ··· 48 48 /* 49 49 * Platform "CPU" startup hook 50 50 */ 51 - static void __cpuinit msp_smtc_boot_secondary(int cpu, 52 - struct task_struct *idle) 51 + static void msp_smtc_boot_secondary(int cpu, struct task_struct *idle) 53 52 { 54 53 smtc_boot_secondary(cpu, idle); 55 54 } ··· 56 57 /* 57 58 * SMP initialization finalization entry point 58 59 */ 59 - static void __cpuinit msp_smtc_smp_finish(void) 60 + static void msp_smtc_smp_finish(void) 60 61 { 61 62 smtc_smp_finish(); 62 63 }
+1 -1
arch/mips/pmcs-msp71xx/msp_time.c
··· 88 88 mips_hpt_frequency = cpu_rate/2; 89 89 } 90 90 91 - unsigned int __cpuinit get_c0_compare_int(void) 91 + unsigned int get_c0_compare_int(void) 92 92 { 93 93 /* MIPS_MT modes may want timer for second VPE */ 94 94 if ((get_current_vpe()) && !tim_installed) {
+1 -1
arch/mips/pnx833x/common/interrupts.c
··· 281 281 write_c0_status(read_c0_status() | IE_IRQ2); 282 282 } 283 283 284 - unsigned int __cpuinit get_c0_compare_int(void) 284 + unsigned int get_c0_compare_int(void) 285 285 { 286 286 if (cpu_has_vint) 287 287 set_vi_handler(cp0_compare_irq, pnx833x_timer_dispatch);
+1 -1
arch/mips/powertv/time.c
··· 25 25 26 26 #include "powertv-clock.h" 27 27 28 - unsigned int __cpuinit get_c0_compare_int(void) 28 + unsigned int get_c0_compare_int(void) 29 29 { 30 30 return irq_mips_timer; 31 31 }
+1 -1
arch/mips/ralink/irq.c
··· 73 73 .irq_mask_ack = ralink_intc_irq_mask, 74 74 }; 75 75 76 - unsigned int __cpuinit get_c0_compare_int(void) 76 + unsigned int get_c0_compare_int(void) 77 77 { 78 78 return CP0_LEGACY_COMPARE_IRQ; 79 79 }
+2 -2
arch/mips/sgi-ip27/ip27-init.c
··· 54 54 55 55 extern void xtalk_probe_node(cnodeid_t nid); 56 56 57 - static void __cpuinit per_hub_init(cnodeid_t cnode) 57 + static void per_hub_init(cnodeid_t cnode) 58 58 { 59 59 struct hub_data *hub = hub_data(cnode); 60 60 nasid_t nasid = COMPACT_TO_NASID_NODEID(cnode); ··· 110 110 } 111 111 } 112 112 113 - void __cpuinit per_cpu_init(void) 113 + void per_cpu_init(void) 114 114 { 115 115 int cpu = smp_processor_id(); 116 116 int slice = LOCAL_HUB_L(PI_CPU_NUM);
+3 -3
arch/mips/sgi-ip27/ip27-smp.c
··· 173 173 ip27_send_ipi_single(i, action); 174 174 } 175 175 176 - static void __cpuinit ip27_init_secondary(void) 176 + static void ip27_init_secondary(void) 177 177 { 178 178 per_cpu_init(); 179 179 } 180 180 181 - static void __cpuinit ip27_smp_finish(void) 181 + static void ip27_smp_finish(void) 182 182 { 183 183 extern void hub_rt_clock_event_init(void); 184 184 ··· 195 195 * set sp to the kernel stack of the newly created idle process, gp to the proc 196 196 * struct so that current_thread_info() will work. 197 197 */ 198 - static void __cpuinit ip27_boot_secondary(int cpu, struct task_struct *idle) 198 + static void ip27_boot_secondary(int cpu, struct task_struct *idle) 199 199 { 200 200 unsigned long gp = (unsigned long)task_thread_info(idle); 201 201 unsigned long sp = __KSTK_TOS(idle);
+3 -3
arch/mips/sgi-ip27/ip27-timer.c
··· 106 106 #define NSEC_PER_CYCLE 800 107 107 #define CYCLES_PER_SEC (NSEC_PER_SEC / NSEC_PER_CYCLE) 108 108 109 - void __cpuinit hub_rt_clock_event_init(void) 109 + void hub_rt_clock_event_init(void) 110 110 { 111 111 unsigned int cpu = smp_processor_id(); 112 112 struct clock_event_device *cd = &per_cpu(hub_rt_clockevent, cpu); ··· 173 173 hub_rt_clock_event_init(); 174 174 } 175 175 176 - void __cpuinit cpu_time_init(void) 176 + void cpu_time_init(void) 177 177 { 178 178 lboard_t *board; 179 179 klcpu_t *cpu; ··· 194 194 set_c0_status(SRB_TIMOCLK); 195 195 } 196 196 197 - void __cpuinit hub_rtc_init(cnodeid_t cnode) 197 + void hub_rtc_init(cnodeid_t cnode) 198 198 { 199 199 200 200 /*
+3 -3
arch/mips/sgi-ip27/ip27-xtalk.c
··· 23 23 24 24 extern int bridge_probe(nasid_t nasid, int widget, int masterwid); 25 25 26 - static int __cpuinit probe_one_port(nasid_t nasid, int widget, int masterwid) 26 + static int probe_one_port(nasid_t nasid, int widget, int masterwid) 27 27 { 28 28 widgetreg_t widget_id; 29 29 xwidget_part_num_t partnum; ··· 47 47 return 0; 48 48 } 49 49 50 - static int __cpuinit xbow_probe(nasid_t nasid) 50 + static int xbow_probe(nasid_t nasid) 51 51 { 52 52 lboard_t *brd; 53 53 klxbow_t *xbow_p; ··· 100 100 return 0; 101 101 } 102 102 103 - void __cpuinit xtalk_probe_node(cnodeid_t nid) 103 + void xtalk_probe_node(cnodeid_t nid) 104 104 { 105 105 volatile u64 hubreg; 106 106 nasid_t nasid;
+4 -4
arch/mips/sibyte/bcm1480/smp.c
··· 60 60 /* 61 61 * SMP init and finish on secondary CPUs 62 62 */ 63 - void __cpuinit bcm1480_smp_init(void) 63 + void bcm1480_smp_init(void) 64 64 { 65 65 unsigned int imask = STATUSF_IP4 | STATUSF_IP3 | STATUSF_IP2 | 66 66 STATUSF_IP1 | STATUSF_IP0; ··· 95 95 /* 96 96 * Code to run on secondary just after probing the CPU 97 97 */ 98 - static void __cpuinit bcm1480_init_secondary(void) 98 + static void bcm1480_init_secondary(void) 99 99 { 100 100 extern void bcm1480_smp_init(void); 101 101 ··· 106 106 * Do any tidying up before marking online and running the idle 107 107 * loop 108 108 */ 109 - static void __cpuinit bcm1480_smp_finish(void) 109 + static void bcm1480_smp_finish(void) 110 110 { 111 111 extern void sb1480_clockevent_init(void); 112 112 ··· 125 125 * Setup the PC, SP, and GP of a secondary processor and start it 126 126 * running! 127 127 */ 128 - static void __cpuinit bcm1480_boot_secondary(int cpu, struct task_struct *idle) 128 + static void bcm1480_boot_secondary(int cpu, struct task_struct *idle) 129 129 { 130 130 int retval; 131 131
+4 -4
arch/mips/sibyte/sb1250/smp.c
··· 48 48 /* 49 49 * SMP init and finish on secondary CPUs 50 50 */ 51 - void __cpuinit sb1250_smp_init(void) 51 + void sb1250_smp_init(void) 52 52 { 53 53 unsigned int imask = STATUSF_IP4 | STATUSF_IP3 | STATUSF_IP2 | 54 54 STATUSF_IP1 | STATUSF_IP0; ··· 83 83 /* 84 84 * Code to run on secondary just after probing the CPU 85 85 */ 86 - static void __cpuinit sb1250_init_secondary(void) 86 + static void sb1250_init_secondary(void) 87 87 { 88 88 extern void sb1250_smp_init(void); 89 89 ··· 94 94 * Do any tidying up before marking online and running the idle 95 95 * loop 96 96 */ 97 - static void __cpuinit sb1250_smp_finish(void) 97 + static void sb1250_smp_finish(void) 98 98 { 99 99 extern void sb1250_clockevent_init(void); 100 100 ··· 113 113 * Setup the PC, SP, and GP of a secondary processor and start it 114 114 * running! 115 115 */ 116 - static void __cpuinit sb1250_boot_secondary(int cpu, struct task_struct *idle) 116 + static void sb1250_boot_secondary(int cpu, struct task_struct *idle) 117 117 { 118 118 int retval; 119 119
+1 -1
arch/openrisc/kernel/setup.c
··· 267 267 * 268 268 */ 269 269 270 - void __cpuinit calibrate_delay(void) 270 + void calibrate_delay(void) 271 271 { 272 272 const int *val; 273 273 struct device_node *cpu = NULL;
+8 -6
arch/parisc/kernel/firmware.c
··· 150 150 } 151 151 152 152 #ifdef CONFIG_64BIT 153 - void __cpuinit set_firmware_width_unlocked(void) 153 + void set_firmware_width_unlocked(void) 154 154 { 155 155 int ret; 156 156 ··· 167 167 * This function must be called before any pdc_* function that uses the 168 168 * convert_to_wide function. 169 169 */ 170 - void __cpuinit set_firmware_width(void) 170 + void set_firmware_width(void) 171 171 { 172 172 unsigned long flags; 173 173 spin_lock_irqsave(&pdc_lock, flags); ··· 175 175 spin_unlock_irqrestore(&pdc_lock, flags); 176 176 } 177 177 #else 178 - void __cpuinit set_firmware_width_unlocked(void) { 178 + void set_firmware_width_unlocked(void) 179 + { 179 180 return; 180 181 } 181 182 182 - void __cpuinit set_firmware_width(void) { 183 + void set_firmware_width(void) 184 + { 183 185 return; 184 186 } 185 187 #endif /*CONFIG_64BIT*/ ··· 303 301 return retval; 304 302 } 305 303 306 - int __cpuinit pdc_coproc_cfg_unlocked(struct pdc_coproc_cfg *pdc_coproc_info) 304 + int pdc_coproc_cfg_unlocked(struct pdc_coproc_cfg *pdc_coproc_info) 307 305 { 308 306 int ret; 309 307 ··· 324 322 * This PDC call returns the presence and status of all the coprocessors 325 323 * attached to the processor. 326 324 */ 327 - int __cpuinit pdc_coproc_cfg(struct pdc_coproc_cfg *pdc_coproc_info) 325 + int pdc_coproc_cfg(struct pdc_coproc_cfg *pdc_coproc_info) 328 326 { 329 327 int ret; 330 328 unsigned long flags;
+1 -1
arch/parisc/kernel/hardware.c
··· 1367 1367 1368 1368 1369 1369 /* Interpret hversion (ret[0]) from PDC_MODEL(4)/PDC_MODEL_INFO(0) */ 1370 - enum cpu_type __cpuinit 1370 + enum cpu_type 1371 1371 parisc_get_cpu_type(unsigned long hversion) 1372 1372 { 1373 1373 struct hp_cpu_type_mask *ptr;
+3 -3
arch/parisc/kernel/processor.c
··· 73 73 * 74 74 * FIXME: doesn't do much yet... 75 75 */ 76 - static void __cpuinit 76 + static void 77 77 init_percpu_prof(unsigned long cpunum) 78 78 { 79 79 struct cpuinfo_parisc *p; ··· 92 92 * (return 1). If so, initialize the chip and tell other partners in crime 93 93 * they have work to do. 94 94 */ 95 - static int __cpuinit processor_probe(struct parisc_device *dev) 95 + static int processor_probe(struct parisc_device *dev) 96 96 { 97 97 unsigned long txn_addr; 98 98 unsigned long cpuid; ··· 299 299 * 300 300 * o Enable CPU profiling hooks. 301 301 */ 302 - int __cpuinit init_per_cpu(int cpunum) 302 + int init_per_cpu(int cpunum) 303 303 { 304 304 int ret; 305 305 struct pdc_coproc_cfg coproc_cfg;
+4 -4
arch/parisc/kernel/smp.c
··· 62 62 volatile struct task_struct *smp_init_current_idle_task; 63 63 64 64 /* track which CPU is booting */ 65 - static volatile int cpu_now_booting __cpuinitdata; 65 + static volatile int cpu_now_booting; 66 66 67 - static int parisc_max_cpus __cpuinitdata = 1; 67 + static int parisc_max_cpus = 1; 68 68 69 69 static DEFINE_PER_CPU(spinlock_t, ipi_lock); 70 70 ··· 328 328 /* 329 329 * Bring one cpu online. 330 330 */ 331 - int __cpuinit smp_boot_one_cpu(int cpuid, struct task_struct *idle) 331 + int smp_boot_one_cpu(int cpuid, struct task_struct *idle) 332 332 { 333 333 const struct cpuinfo_parisc *p = &per_cpu(cpu_data, cpuid); 334 334 long timeout; ··· 424 424 } 425 425 426 426 427 - int __cpuinit __cpu_up(unsigned int cpu, struct task_struct *tidle) 427 + int __cpu_up(unsigned int cpu, struct task_struct *tidle) 428 428 { 429 429 if (cpu != 0 && cpu < parisc_max_cpus) 430 430 smp_boot_one_cpu(cpu, tidle);
+7 -8
arch/s390/kernel/cache.c
··· 173 173 } 174 174 } 175 175 176 - static struct cache_dir *__cpuinit cache_create_cache_dir(int cpu) 176 + static struct cache_dir *cache_create_cache_dir(int cpu) 177 177 { 178 178 struct cache_dir *cache_dir; 179 179 struct kobject *kobj = NULL; ··· 289 289 .default_attrs = cache_index_default_attrs, 290 290 }; 291 291 292 - static int __cpuinit cache_create_index_dir(struct cache_dir *cache_dir, 293 - struct cache *cache, int index, 294 - int cpu) 292 + static int cache_create_index_dir(struct cache_dir *cache_dir, 293 + struct cache *cache, int index, int cpu) 295 294 { 296 295 struct cache_index_dir *index_dir; 297 296 int rc; ··· 312 313 return rc; 313 314 } 314 315 315 - static int __cpuinit cache_add_cpu(int cpu) 316 + static int cache_add_cpu(int cpu) 316 317 { 317 318 struct cache_dir *cache_dir; 318 319 struct cache *cache; ··· 334 335 return 0; 335 336 } 336 337 337 - static void __cpuinit cache_remove_cpu(int cpu) 338 + static void cache_remove_cpu(int cpu) 338 339 { 339 340 struct cache_index_dir *index, *next; 340 341 struct cache_dir *cache_dir; ··· 353 354 cache_dir_cpu[cpu] = NULL; 354 355 } 355 356 356 - static int __cpuinit cache_hotplug(struct notifier_block *nfb, 357 - unsigned long action, void *hcpu) 357 + static int cache_hotplug(struct notifier_block *nfb, unsigned long action, 358 + void *hcpu) 358 359 { 359 360 int cpu = (long)hcpu; 360 361 int rc = 0;
+2 -2
arch/s390/kernel/perf_cpum_cf.c
··· 639 639 .cancel_txn = cpumf_pmu_cancel_txn, 640 640 }; 641 641 642 - static int __cpuinit cpumf_pmu_notifier(struct notifier_block *self, 643 - unsigned long action, void *hcpu) 642 + static int cpumf_pmu_notifier(struct notifier_block *self, unsigned long action, 643 + void *hcpu) 644 644 { 645 645 unsigned int cpu = (long) hcpu; 646 646 int flags;
+1 -1
arch/s390/kernel/processor.c
··· 21 21 /* 22 22 * cpu_init - initializes state that is per-CPU. 23 23 */ 24 - void __cpuinit cpu_init(void) 24 + void cpu_init(void) 25 25 { 26 26 struct s390_idle_data *idle = &__get_cpu_var(s390_idle); 27 27 struct cpuid *id = &__get_cpu_var(cpu_id);
+8 -9
arch/s390/kernel/smp.c
··· 165 165 pcpu_sigp_retry(pcpu, order, 0); 166 166 } 167 167 168 - static int __cpuinit pcpu_alloc_lowcore(struct pcpu *pcpu, int cpu) 168 + static int pcpu_alloc_lowcore(struct pcpu *pcpu, int cpu) 169 169 { 170 170 struct _lowcore *lc; 171 171 ··· 616 616 return info; 617 617 } 618 618 619 - static int __cpuinit smp_add_present_cpu(int cpu); 619 + static int smp_add_present_cpu(int cpu); 620 620 621 - static int __cpuinit __smp_rescan_cpus(struct sclp_cpu_info *info, 622 - int sysfs_add) 621 + static int __smp_rescan_cpus(struct sclp_cpu_info *info, int sysfs_add) 623 622 { 624 623 struct pcpu *pcpu; 625 624 cpumask_t avail; ··· 684 685 /* 685 686 * Activate a secondary processor. 686 687 */ 687 - static void __cpuinit smp_start_secondary(void *cpuvoid) 688 + static void smp_start_secondary(void *cpuvoid) 688 689 { 689 690 S390_lowcore.last_update_clock = get_tod_clock(); 690 691 S390_lowcore.restart_stack = (unsigned long) restart_stack; ··· 707 708 } 708 709 709 710 /* Upping and downing of CPUs */ 710 - int __cpuinit __cpu_up(unsigned int cpu, struct task_struct *tidle) 711 + int __cpu_up(unsigned int cpu, struct task_struct *tidle) 711 712 { 712 713 struct pcpu *pcpu; 713 714 int rc; ··· 963 964 .attrs = cpu_online_attrs, 964 965 }; 965 966 966 - static int __cpuinit smp_cpu_notify(struct notifier_block *self, 967 - unsigned long action, void *hcpu) 967 + static int smp_cpu_notify(struct notifier_block *self, unsigned long action, 968 + void *hcpu) 968 969 { 969 970 unsigned int cpu = (unsigned int)(long)hcpu; 970 971 struct cpu *c = &pcpu_devices[cpu].cpu; ··· 982 983 return notifier_from_errno(err); 983 984 } 984 985 985 - static int __cpuinit smp_add_present_cpu(int cpu) 986 + static int smp_add_present_cpu(int cpu) 986 987 { 987 988 struct cpu *c = &pcpu_devices[cpu].cpu; 988 989 struct device *s = &c->dev;
+1 -1
arch/s390/kernel/sysinfo.c
··· 418 418 /* 419 419 * calibrate the delay loop 420 420 */ 421 - void __cpuinit calibrate_delay(void) 421 + void calibrate_delay(void) 422 422 { 423 423 s390_adjust_jiffies(); 424 424 /* Print the good old Bogomips line .. */
+3 -3
arch/s390/kernel/vtime.c
··· 371 371 /* 372 372 * Start the virtual CPU timer on the current CPU. 373 373 */ 374 - void __cpuinit init_cpu_vtimer(void) 374 + void init_cpu_vtimer(void) 375 375 { 376 376 /* set initial cpu timer */ 377 377 set_vtimer(VTIMER_MAX_SLICE); 378 378 } 379 379 380 - static int __cpuinit s390_nohz_notify(struct notifier_block *self, 381 - unsigned long action, void *hcpu) 380 + static int s390_nohz_notify(struct notifier_block *self, unsigned long action, 381 + void *hcpu) 382 382 { 383 383 struct s390_idle_data *idle; 384 384 long cpu = (long) hcpu;
+2 -2
arch/s390/mm/fault.c
··· 639 639 put_task_struct(tsk); 640 640 } 641 641 642 - static int __cpuinit pfault_cpu_notify(struct notifier_block *self, 643 - unsigned long action, void *hcpu) 642 + static int pfault_cpu_notify(struct notifier_block *self, unsigned long action, 643 + void *hcpu) 644 644 { 645 645 struct thread_struct *thread, *next; 646 646 struct task_struct *tsk;
+1 -1
arch/score/mm/tlb-score.c
··· 240 240 local_irq_restore(flags); 241 241 } 242 242 243 - void __cpuinit tlb_init(void) 243 + void tlb_init(void) 244 244 { 245 245 tlblock_set(0); 246 246 local_flush_tlb_all();
+9 -9
arch/sh/kernel/cpu/init.c
··· 43 43 * peripherals (nofpu, nodsp, and so forth). 44 44 */ 45 45 #define onchip_setup(x) \ 46 - static int x##_disabled __cpuinitdata = !cpu_has_##x; \ 46 + static int x##_disabled = !cpu_has_##x; \ 47 47 \ 48 - static int __cpuinit x##_setup(char *opts) \ 48 + static int x##_setup(char *opts) \ 49 49 { \ 50 50 x##_disabled = 1; \ 51 51 return 1; \ ··· 59 59 #define CPUOPM 0xff2f0000 60 60 #define CPUOPM_RABD (1 << 5) 61 61 62 - static void __cpuinit speculative_execution_init(void) 62 + static void speculative_execution_init(void) 63 63 { 64 64 /* Clear RABD */ 65 65 __raw_writel(__raw_readl(CPUOPM) & ~CPUOPM_RABD, CPUOPM); ··· 78 78 #define EXPMASK_BRDSSLP (1 << 1) 79 79 #define EXPMASK_MMCAW (1 << 4) 80 80 81 - static void __cpuinit expmask_init(void) 81 + static void expmask_init(void) 82 82 { 83 83 unsigned long expmask = __raw_readl(EXPMASK); 84 84 ··· 217 217 l2_cache_shape = -1; /* No S-cache */ 218 218 } 219 219 220 - static void __cpuinit fpu_init(void) 220 + static void fpu_init(void) 221 221 { 222 222 /* Disable the FPU */ 223 223 if (fpu_disabled && (current_cpu_data.flags & CPU_HAS_FPU)) { ··· 230 230 } 231 231 232 232 #ifdef CONFIG_SH_DSP 233 - static void __cpuinit release_dsp(void) 233 + static void release_dsp(void) 234 234 { 235 235 unsigned long sr; 236 236 ··· 244 244 ); 245 245 } 246 246 247 - static void __cpuinit dsp_init(void) 247 + static void dsp_init(void) 248 248 { 249 249 unsigned long sr; 250 250 ··· 276 276 release_dsp(); 277 277 } 278 278 #else 279 - static inline void __cpuinit dsp_init(void) { } 279 + static inline void dsp_init(void) { } 280 280 #endif /* CONFIG_SH_DSP */ 281 281 282 282 /** ··· 295 295 * Each processor family is still responsible for doing its own probing 296 296 * and cache configuration in cpu_probe(). 297 297 */ 298 - asmlinkage void __cpuinit cpu_init(void) 298 + asmlinkage void cpu_init(void) 299 299 { 300 300 current_thread_info()->cpu = hard_smp_processor_id(); 301 301
+1 -1
arch/sh/kernel/cpu/sh2/probe.c
··· 13 13 #include <asm/processor.h> 14 14 #include <asm/cache.h> 15 15 16 - void __cpuinit cpu_probe(void) 16 + void cpu_probe(void) 17 17 { 18 18 #if defined(CONFIG_CPU_SUBTYPE_SH7619) 19 19 boot_cpu_data.type = CPU_SH7619;
+1 -1
arch/sh/kernel/cpu/sh2a/probe.c
··· 13 13 #include <asm/processor.h> 14 14 #include <asm/cache.h> 15 15 16 - void __cpuinit cpu_probe(void) 16 + void cpu_probe(void) 17 17 { 18 18 boot_cpu_data.family = CPU_FAMILY_SH2A; 19 19
+1 -1
arch/sh/kernel/cpu/sh3/probe.c
··· 16 16 #include <asm/cache.h> 17 17 #include <asm/io.h> 18 18 19 - void __cpuinit cpu_probe(void) 19 + void cpu_probe(void) 20 20 { 21 21 unsigned long addr0, addr1, data0, data1, data2, data3; 22 22
+1 -1
arch/sh/kernel/cpu/sh4/probe.c
··· 15 15 #include <asm/processor.h> 16 16 #include <asm/cache.h> 17 17 18 - void __cpuinit cpu_probe(void) 18 + void cpu_probe(void) 19 19 { 20 20 unsigned long pvr, prr, cvr; 21 21 unsigned long size;
+3 -3
arch/sh/kernel/cpu/sh4a/smp-shx3.c
··· 124 124 __raw_writel(STBCR_RESET, STBCR_REG(cpu)); 125 125 } 126 126 127 - static int __cpuinit 127 + static int 128 128 shx3_cpu_callback(struct notifier_block *nfb, unsigned long action, void *hcpu) 129 129 { 130 130 unsigned int cpu = (unsigned int)hcpu; ··· 143 143 return NOTIFY_OK; 144 144 } 145 145 146 - static struct notifier_block __cpuinitdata shx3_cpu_notifier = { 146 + static struct notifier_block shx3_cpu_notifier = { 147 147 .notifier_call = shx3_cpu_callback, 148 148 }; 149 149 150 - static int __cpuinit register_shx3_cpu_notifier(void) 150 + static int register_shx3_cpu_notifier(void) 151 151 { 152 152 register_hotcpu_notifier(&shx3_cpu_notifier); 153 153 return 0;
+1 -1
arch/sh/kernel/cpu/sh5/probe.c
··· 17 17 #include <asm/cache.h> 18 18 #include <asm/tlb.h> 19 19 20 - void __cpuinit cpu_probe(void) 20 + void cpu_probe(void) 21 21 { 22 22 unsigned long long cir; 23 23
+2 -2
arch/sh/kernel/perf_event.c
··· 367 367 memset(cpuhw, 0, sizeof(struct cpu_hw_events)); 368 368 } 369 369 370 - static int __cpuinit 370 + static int 371 371 sh_pmu_notifier(struct notifier_block *self, unsigned long action, void *hcpu) 372 372 { 373 373 unsigned int cpu = (long)hcpu; ··· 384 384 return NOTIFY_OK; 385 385 } 386 386 387 - int __cpuinit register_sh_pmu(struct sh_pmu *_pmu) 387 + int register_sh_pmu(struct sh_pmu *_pmu) 388 388 { 389 389 if (sh_pmu) 390 390 return -EBUSY;
+1 -1
arch/sh/kernel/process.c
··· 65 65 # define HAVE_SOFTFP 0 66 66 #endif 67 67 68 - void __cpuinit init_thread_xstate(void) 68 + void init_thread_xstate(void) 69 69 { 70 70 if (boot_cpu_data.flags & CPU_HAS_FPU) 71 71 xstate_size = sizeof(struct sh_fpu_hard_struct);
+1 -1
arch/sh/kernel/setup.c
··· 172 172 #endif 173 173 } 174 174 175 - void __cpuinit calibrate_delay(void) 175 + void calibrate_delay(void) 176 176 { 177 177 struct clk *clk = clk_get(NULL, "cpu_clk"); 178 178
+4 -4
arch/sh/kernel/smp.c
··· 37 37 /* State of each CPU */ 38 38 DEFINE_PER_CPU(int, cpu_state) = { 0 }; 39 39 40 - void __cpuinit register_smp_ops(struct plat_smp_ops *ops) 40 + void register_smp_ops(struct plat_smp_ops *ops) 41 41 { 42 42 if (mp_ops) 43 43 printk(KERN_WARNING "Overriding previously set SMP ops\n"); ··· 45 45 mp_ops = ops; 46 46 } 47 47 48 - static inline void __cpuinit smp_store_cpu_info(unsigned int cpu) 48 + static inline void smp_store_cpu_info(unsigned int cpu) 49 49 { 50 50 struct sh_cpuinfo *c = cpu_data + cpu; 51 51 ··· 174 174 } 175 175 #endif 176 176 177 - asmlinkage void __cpuinit start_secondary(void) 177 + asmlinkage void start_secondary(void) 178 178 { 179 179 unsigned int cpu = smp_processor_id(); 180 180 struct mm_struct *mm = &init_mm; ··· 215 215 void *thread_info; 216 216 } stack_start; 217 217 218 - int __cpuinit __cpu_up(unsigned int cpu, struct task_struct *tsk) 218 + int __cpu_up(unsigned int cpu, struct task_struct *tsk) 219 219 { 220 220 unsigned long timeout; 221 221
+1 -1
arch/sh/kernel/traps_32.c
··· 741 741 die_if_kernel("exception", regs, ex); 742 742 } 743 743 744 - void __cpuinit per_cpu_trap_init(void) 744 + void per_cpu_trap_init(void) 745 745 { 746 746 extern void *vbr_base; 747 747
+1 -1
arch/sh/kernel/traps_64.c
··· 810 810 poke_real_address_q(DM_EXP_CAUSE_PHY, 0x0); 811 811 } 812 812 813 - void __cpuinit per_cpu_trap_init(void) 813 + void per_cpu_trap_init(void) 814 814 { 815 815 /* Nothing to do for now, VBR initialization later. */ 816 816 }
+1 -1
arch/sh/mm/tlb-sh5.c
··· 17 17 /** 18 18 * sh64_tlb_init - Perform initial setup for the DTLB and ITLB. 19 19 */ 20 - int __cpuinit sh64_tlb_init(void) 20 + int sh64_tlb_init(void) 21 21 { 22 22 /* Assign some sane DTLB defaults */ 23 23 cpu_data->dtlb.entries = 64;
+4 -7
arch/sparc/kernel/ds.c
··· 528 528 } 529 529 } 530 530 531 - static int __cpuinit dr_cpu_configure(struct ds_info *dp, 532 - struct ds_cap_state *cp, 533 - u64 req_num, 534 - cpumask_t *mask) 531 + static int dr_cpu_configure(struct ds_info *dp, struct ds_cap_state *cp, 532 + u64 req_num, cpumask_t *mask) 535 533 { 536 534 struct ds_data *resp; 537 535 int resp_len, ncpus, cpu; ··· 625 627 return 0; 626 628 } 627 629 628 - static void __cpuinit dr_cpu_data(struct ds_info *dp, 629 - struct ds_cap_state *cp, 630 - void *buf, int len) 630 + static void dr_cpu_data(struct ds_info *dp, struct ds_cap_state *cp, void *buf, 631 + int len) 631 632 { 632 633 struct ds_data *data = buf; 633 634 struct dr_cpu_tag *tag = (struct dr_cpu_tag *) (data + 1);
+1 -1
arch/sparc/kernel/entry.h
··· 250 250 extern unsigned long ivector_table_pa; 251 251 252 252 extern void init_irqwork_curcpu(void); 253 - extern void __cpuinit sun4v_register_mondo_queues(int this_cpu); 253 + extern void sun4v_register_mondo_queues(int this_cpu); 254 254 255 255 #endif /* CONFIG_SPARC32 */ 256 256 #endif /* _ENTRY_H */
-1
arch/sparc/kernel/hvtramp.S
··· 16 16 #include <asm/asi.h> 17 17 #include <asm/pil.h> 18 18 19 - __CPUINIT 20 19 .align 8 21 20 .globl hv_cpu_startup, hv_cpu_startup_end 22 21
+3 -2
arch/sparc/kernel/irq_64.c
··· 835 835 * Therefore you cannot make any OBP calls, not even prom_printf, 836 836 * from these two routines. 837 837 */ 838 - static void __cpuinit notrace register_one_mondo(unsigned long paddr, unsigned long type, unsigned long qmask) 838 + static void notrace register_one_mondo(unsigned long paddr, unsigned long type, 839 + unsigned long qmask) 839 840 { 840 841 unsigned long num_entries = (qmask + 1) / 64; 841 842 unsigned long status; ··· 849 848 } 850 849 } 851 850 852 - void __cpuinit notrace sun4v_register_mondo_queues(int this_cpu) 851 + void notrace sun4v_register_mondo_queues(int this_cpu) 853 852 { 854 853 struct trap_per_cpu *tb = &trap_block[this_cpu]; 855 854
+5 -5
arch/sparc/kernel/leon_smp.c
··· 54 54 static int smp_processors_ready; 55 55 extern volatile unsigned long cpu_callin_map[NR_CPUS]; 56 56 extern cpumask_t smp_commenced_mask; 57 - void __cpuinit leon_configure_cache_smp(void); 57 + void leon_configure_cache_smp(void); 58 58 static void leon_ipi_init(void); 59 59 60 60 /* IRQ number of LEON IPIs */ ··· 69 69 return val; 70 70 } 71 71 72 - void __cpuinit leon_cpu_pre_starting(void *arg) 72 + void leon_cpu_pre_starting(void *arg) 73 73 { 74 74 leon_configure_cache_smp(); 75 75 } 76 76 77 - void __cpuinit leon_cpu_pre_online(void *arg) 77 + void leon_cpu_pre_online(void *arg) 78 78 { 79 79 int cpuid = hard_smp_processor_id(); 80 80 ··· 106 106 107 107 extern struct linux_prom_registers smp_penguin_ctable; 108 108 109 - void __cpuinit leon_configure_cache_smp(void) 109 + void leon_configure_cache_smp(void) 110 110 { 111 111 unsigned long cfg = sparc_leon3_get_dcachecfg(); 112 112 int me = smp_processor_id(); ··· 186 186 187 187 } 188 188 189 - int __cpuinit leon_boot_one_cpu(int i, struct task_struct *idle) 189 + int leon_boot_one_cpu(int i, struct task_struct *idle) 190 190 { 191 191 int timeout; 192 192
+17 -17
arch/sparc/kernel/mdesc.c
··· 571 571 mdesc_release(hp); 572 572 } 573 573 574 - static void __cpuinit fill_in_one_cache(cpuinfo_sparc *c, 575 - struct mdesc_handle *hp, 576 - u64 mp) 574 + static void fill_in_one_cache(cpuinfo_sparc *c, struct mdesc_handle *hp, u64 mp) 577 575 { 578 576 const u64 *level = mdesc_get_property(hp, mp, "level", NULL); 579 577 const u64 *size = mdesc_get_property(hp, mp, "size", NULL); ··· 614 616 } 615 617 } 616 618 617 - static void __cpuinit mark_core_ids(struct mdesc_handle *hp, u64 mp, int core_id) 619 + static void mark_core_ids(struct mdesc_handle *hp, u64 mp, int core_id) 618 620 { 619 621 u64 a; 620 622 ··· 647 649 } 648 650 } 649 651 650 - static void __cpuinit set_core_ids(struct mdesc_handle *hp) 652 + static void set_core_ids(struct mdesc_handle *hp) 651 653 { 652 654 int idx; 653 655 u64 mp; ··· 672 674 } 673 675 } 674 676 675 - static void __cpuinit mark_proc_ids(struct mdesc_handle *hp, u64 mp, int proc_id) 677 + static void mark_proc_ids(struct mdesc_handle *hp, u64 mp, int proc_id) 676 678 { 677 679 u64 a; 678 680 ··· 691 693 } 692 694 } 693 695 694 - static void __cpuinit __set_proc_ids(struct mdesc_handle *hp, const char *exec_unit_name) 696 + static void __set_proc_ids(struct mdesc_handle *hp, const char *exec_unit_name) 695 697 { 696 698 int idx; 697 699 u64 mp; ··· 712 714 } 713 715 } 714 716 715 - static void __cpuinit set_proc_ids(struct mdesc_handle *hp) 717 + static void set_proc_ids(struct mdesc_handle *hp) 716 718 { 717 719 __set_proc_ids(hp, "exec_unit"); 718 720 __set_proc_ids(hp, "exec-unit"); 719 721 } 720 722 721 - static void __cpuinit get_one_mondo_bits(const u64 *p, unsigned int *mask, 722 - unsigned long def, unsigned long max) 723 + static void get_one_mondo_bits(const u64 *p, unsigned int *mask, 724 + unsigned long def, unsigned long max) 723 725 { 724 726 u64 val; 725 727 ··· 740 742 *mask = ((1U << def) * 64U) - 1U; 741 743 } 742 744 743 - static void __cpuinit get_mondo_data(struct mdesc_handle *hp, u64 mp, 744 - struct trap_per_cpu *tb) 745 + static void get_mondo_data(struct mdesc_handle *hp, u64 mp, 746 + struct trap_per_cpu *tb) 745 747 { 746 748 static int printed; 747 749 const u64 *val; ··· 767 769 } 768 770 } 769 771 770 - static void * __cpuinit mdesc_iterate_over_cpus(void *(*func)(struct mdesc_handle *, u64, int, void *), void *arg, cpumask_t *mask) 772 + static void *mdesc_iterate_over_cpus(void *(*func)(struct mdesc_handle *, u64, int, void *), void *arg, cpumask_t *mask) 771 773 { 772 774 struct mdesc_handle *hp = mdesc_grab(); 773 775 void *ret = NULL; ··· 797 799 return ret; 798 800 } 799 801 800 - static void * __cpuinit record_one_cpu(struct mdesc_handle *hp, u64 mp, int cpuid, void *arg) 802 + static void *record_one_cpu(struct mdesc_handle *hp, u64 mp, int cpuid, 803 + void *arg) 801 804 { 802 805 ncpus_probed++; 803 806 #ifdef CONFIG_SMP ··· 807 808 return NULL; 808 809 } 809 810 810 - void __cpuinit mdesc_populate_present_mask(cpumask_t *mask) 811 + void mdesc_populate_present_mask(cpumask_t *mask) 811 812 { 812 813 if (tlb_type != hypervisor) 813 814 return; ··· 840 841 mdesc_iterate_over_cpus(check_one_pgsz, pgsz_mask, mask); 841 842 } 842 843 843 - static void * __cpuinit fill_in_one_cpu(struct mdesc_handle *hp, u64 mp, int cpuid, void *arg) 844 + static void *fill_in_one_cpu(struct mdesc_handle *hp, u64 mp, int cpuid, 845 + void *arg) 844 846 { 845 847 const u64 *cfreq = mdesc_get_property(hp, mp, "clock-frequency", NULL); 846 848 struct trap_per_cpu *tb; ··· 890 890 return NULL; 891 891 } 892 892 893 - void __cpuinit mdesc_fill_in_cpu_data(cpumask_t *mask) 893 + void mdesc_fill_in_cpu_data(cpumask_t *mask) 894 894 { 895 895 struct mdesc_handle *hp; 896 896
+10 -10
arch/sparc/kernel/smp_32.c
··· 39 39 #include "kernel.h" 40 40 #include "irq.h" 41 41 42 - volatile unsigned long cpu_callin_map[NR_CPUS] __cpuinitdata = {0,}; 42 + volatile unsigned long cpu_callin_map[NR_CPUS] = {0,}; 43 43 44 44 cpumask_t smp_commenced_mask = CPU_MASK_NONE; 45 45 ··· 53 53 * instruction which is much better... 54 54 */ 55 55 56 - void __cpuinit smp_store_cpu_info(int id) 56 + void smp_store_cpu_info(int id) 57 57 { 58 58 int cpu_node; 59 59 int mid; ··· 120 120 panic("SMP bolixed\n"); 121 121 } 122 122 123 - struct linux_prom_registers smp_penguin_ctable __cpuinitdata = { 0 }; 123 + struct linux_prom_registers smp_penguin_ctable = { 0 }; 124 124 125 125 void smp_send_reschedule(int cpu) 126 126 { ··· 259 259 set_cpu_possible(cpuid, true); 260 260 } 261 261 262 - int __cpuinit __cpu_up(unsigned int cpu, struct task_struct *tidle) 262 + int __cpu_up(unsigned int cpu, struct task_struct *tidle) 263 263 { 264 - extern int __cpuinit smp4m_boot_one_cpu(int, struct task_struct *); 265 - extern int __cpuinit smp4d_boot_one_cpu(int, struct task_struct *); 264 + extern int smp4m_boot_one_cpu(int, struct task_struct *); 265 + extern int smp4d_boot_one_cpu(int, struct task_struct *); 266 266 int ret=0; 267 267 268 268 switch(sparc_cpu_model) { ··· 297 297 return ret; 298 298 } 299 299 300 - void __cpuinit arch_cpu_pre_starting(void *arg) 300 + void arch_cpu_pre_starting(void *arg) 301 301 { 302 302 local_ops->cache_all(); 303 303 local_ops->tlb_all(); ··· 317 317 } 318 318 } 319 319 320 - void __cpuinit arch_cpu_pre_online(void *arg) 320 + void arch_cpu_pre_online(void *arg) 321 321 { 322 322 unsigned int cpuid = hard_smp_processor_id(); 323 323 ··· 344 344 } 345 345 } 346 346 347 - void __cpuinit sparc_start_secondary(void *arg) 347 + void sparc_start_secondary(void *arg) 348 348 { 349 349 unsigned int cpu; 350 350 ··· 375 375 BUG(); 376 376 } 377 377 378 - void __cpuinit smp_callin(void) 378 + void smp_callin(void) 379 379 { 380 380 sparc_start_secondary(NULL); 381 381 }
+5 -4
arch/sparc/kernel/smp_64.c
··· 87 87 88 88 static volatile unsigned long callin_flag = 0; 89 89 90 - void __cpuinit smp_callin(void) 90 + void smp_callin(void) 91 91 { 92 92 int cpuid = hard_smp_processor_id(); 93 93 ··· 281 281 return kern_base + (val - KERNBASE); 282 282 } 283 283 284 - static void __cpuinit ldom_startcpu_cpuid(unsigned int cpu, unsigned long thread_reg, void **descrp) 284 + static void ldom_startcpu_cpuid(unsigned int cpu, unsigned long thread_reg, 285 + void **descrp) 285 286 { 286 287 extern unsigned long sparc64_ttable_tl0; 287 288 extern unsigned long kern_locked_tte_data; ··· 343 342 */ 344 343 static struct thread_info *cpu_new_thread = NULL; 345 344 346 - static int __cpuinit smp_boot_one_cpu(unsigned int cpu, struct task_struct *idle) 345 + static int smp_boot_one_cpu(unsigned int cpu, struct task_struct *idle) 347 346 { 348 347 unsigned long entry = 349 348 (unsigned long)(&sparc64_cpu_startup); ··· 1267 1266 } 1268 1267 } 1269 1268 1270 - int __cpuinit __cpu_up(unsigned int cpu, struct task_struct *tidle) 1269 + int __cpu_up(unsigned int cpu, struct task_struct *tidle) 1271 1270 { 1272 1271 int ret = smp_boot_one_cpu(cpu, tidle); 1273 1272
+3 -3
arch/sparc/kernel/sun4d_smp.c
··· 50 50 "i" (ASI_M_CTL)); 51 51 } 52 52 53 - void __cpuinit sun4d_cpu_pre_starting(void *arg) 53 + void sun4d_cpu_pre_starting(void *arg) 54 54 { 55 55 int cpuid = hard_smp_processor_id(); 56 56 ··· 62 62 cc_set_imsk((cc_get_imsk() & ~0x8000) | 0x4000); 63 63 } 64 64 65 - void __cpuinit sun4d_cpu_pre_online(void *arg) 65 + void sun4d_cpu_pre_online(void *arg) 66 66 { 67 67 unsigned long flags; 68 68 int cpuid; ··· 118 118 local_ops->cache_all(); 119 119 } 120 120 121 - int __cpuinit smp4d_boot_one_cpu(int i, struct task_struct *idle) 121 + int smp4d_boot_one_cpu(int i, struct task_struct *idle) 122 122 { 123 123 unsigned long *entry = &sun4d_cpu_startup; 124 124 int timeout;
+3 -3
arch/sparc/kernel/sun4m_smp.c
··· 34 34 return val; 35 35 } 36 36 37 - void __cpuinit sun4m_cpu_pre_starting(void *arg) 37 + void sun4m_cpu_pre_starting(void *arg) 38 38 { 39 39 } 40 40 41 - void __cpuinit sun4m_cpu_pre_online(void *arg) 41 + void sun4m_cpu_pre_online(void *arg) 42 42 { 43 43 int cpuid = hard_smp_processor_id(); 44 44 ··· 75 75 local_ops->cache_all(); 76 76 } 77 77 78 - int __cpuinit smp4m_boot_one_cpu(int i, struct task_struct *idle) 78 + int smp4m_boot_one_cpu(int i, struct task_struct *idle) 79 79 { 80 80 unsigned long *entry = &sun4m_cpu_startup; 81 81 int timeout;
+2 -2
arch/sparc/kernel/sysfs.c
··· 246 246 } 247 247 #endif 248 248 249 - static int __cpuinit sysfs_cpu_notify(struct notifier_block *self, 249 + static int sysfs_cpu_notify(struct notifier_block *self, 250 250 unsigned long action, void *hcpu) 251 251 { 252 252 unsigned int cpu = (unsigned int)(long)hcpu; ··· 266 266 return NOTIFY_OK; 267 267 } 268 268 269 - static struct notifier_block __cpuinitdata sysfs_cpu_nb = { 269 + static struct notifier_block sysfs_cpu_nb = { 270 270 .notifier_call = sysfs_cpu_notify, 271 271 }; 272 272
-3
arch/sparc/kernel/trampoline_32.S
··· 18 18 .globl sun4m_cpu_startup 19 19 .globl sun4d_cpu_startup 20 20 21 - __CPUINIT 22 21 .align 4 23 22 24 23 /* When we start up a cpu for the first time it enters this routine. ··· 93 94 /* CPUID in bootbus can be found at PA 0xff0140000 */ 94 95 #define SUN4D_BOOTBUS_CPUID 0xf0140000 95 96 96 - __CPUINIT 97 97 .align 4 98 98 99 99 sun4d_cpu_startup: ··· 144 146 145 147 b,a smp_panic 146 148 147 - __CPUINIT 148 149 .align 4 149 150 .global leon_smp_cpu_startup, smp_penguin_ctable 150 151
-2
arch/sparc/kernel/trampoline_64.S
··· 32 32 dtlb_load: 33 33 .asciz "SUNW,dtlb-load" 34 34 35 - /* XXX __cpuinit this thing XXX */ 36 35 #define TRAMP_STACK_SIZE 1024 37 36 .align 16 38 37 tramp_stack: 39 38 .skip TRAMP_STACK_SIZE 40 39 41 - __CPUINIT 42 40 .align 8 43 41 .globl sparc64_cpu_startup, sparc64_cpu_startup_end 44 42 sparc64_cpu_startup:
+1 -1
arch/sparc/mm/init_64.c
··· 1694 1694 #endif 1695 1695 } 1696 1696 1697 - void __cpuinit sun4v_ktsb_register(void) 1697 + void sun4v_ktsb_register(void) 1698 1698 { 1699 1699 unsigned long pa, ret; 1700 1700
+6 -6
arch/sparc/mm/srmmu.c
··· 858 858 } 859 859 } 860 860 861 - void (*poke_srmmu)(void) __cpuinitdata = NULL; 861 + void (*poke_srmmu)(void) = NULL; 862 862 863 863 extern unsigned long bootmem_init(unsigned long *pages_avail); 864 864 ··· 1055 1055 (int)vac_cache_size, (int)vac_line_size); 1056 1056 } 1057 1057 1058 - static void __cpuinit poke_hypersparc(void) 1058 + static void poke_hypersparc(void) 1059 1059 { 1060 1060 volatile unsigned long clear; 1061 1061 unsigned long mreg = srmmu_get_mmureg(); ··· 1107 1107 hypersparc_setup_blockops(); 1108 1108 } 1109 1109 1110 - static void __cpuinit poke_swift(void) 1110 + static void poke_swift(void) 1111 1111 { 1112 1112 unsigned long mreg; 1113 1113 ··· 1287 1287 } 1288 1288 1289 1289 1290 - static void __cpuinit poke_turbosparc(void) 1290 + static void poke_turbosparc(void) 1291 1291 { 1292 1292 unsigned long mreg = srmmu_get_mmureg(); 1293 1293 unsigned long ccreg; ··· 1350 1350 poke_srmmu = poke_turbosparc; 1351 1351 } 1352 1352 1353 - static void __cpuinit poke_tsunami(void) 1353 + static void poke_tsunami(void) 1354 1354 { 1355 1355 unsigned long mreg = srmmu_get_mmureg(); 1356 1356 ··· 1391 1391 tsunami_setup_blockops(); 1392 1392 } 1393 1393 1394 - static void __cpuinit poke_viking(void) 1394 + static void poke_viking(void) 1395 1395 { 1396 1396 unsigned long mreg = srmmu_get_mmureg(); 1397 1397 static int smp_catch;
+1 -1
arch/tile/kernel/irq.c
··· 220 220 ipi_init(); 221 221 } 222 222 223 - void __cpuinit setup_irq_regs(void) 223 + void setup_irq_regs(void) 224 224 { 225 225 /* Enable interrupt delivery. */ 226 226 unmask_irqs(~0UL);
+1 -1
arch/tile/kernel/messaging.c
··· 25 25 /* All messages are stored here */ 26 26 static DEFINE_PER_CPU(HV_MsgState, msg_state); 27 27 28 - void __cpuinit init_messaging(void) 28 + void init_messaging(void) 29 29 { 30 30 /* Allocate storage for messages in kernel space */ 31 31 HV_MsgState *state = &__get_cpu_var(msg_state);
+6 -6
arch/tile/kernel/setup.c
··· 58 58 EXPORT_SYMBOL(node_data); 59 59 60 60 /* Information on the NUMA nodes that we compute early */ 61 - unsigned long __cpuinitdata node_start_pfn[MAX_NUMNODES]; 62 - unsigned long __cpuinitdata node_end_pfn[MAX_NUMNODES]; 61 + unsigned long node_start_pfn[MAX_NUMNODES]; 62 + unsigned long node_end_pfn[MAX_NUMNODES]; 63 63 unsigned long __initdata node_memmap_pfn[MAX_NUMNODES]; 64 64 unsigned long __initdata node_percpu_pfn[MAX_NUMNODES]; 65 65 unsigned long __initdata node_free_pfn[MAX_NUMNODES]; ··· 84 84 85 85 #ifdef CONFIG_HIGHMEM 86 86 /* Page frame index of end of lowmem on each controller. */ 87 - unsigned long __cpuinitdata node_lowmem_end_pfn[MAX_NUMNODES]; 87 + unsigned long node_lowmem_end_pfn[MAX_NUMNODES]; 88 88 89 89 /* Number of pages that can be mapped into lowmem. */ 90 90 static unsigned long __initdata mappable_physpages; ··· 290 290 * This is up to 4 mappings for lowmem, one mapping per memory 291 291 * controller, plus one for our text segment. 292 292 */ 293 - static void __cpuinit store_permanent_mappings(void) 293 + static void store_permanent_mappings(void) 294 294 { 295 295 int i; 296 296 ··· 935 935 * So the values we set up here in the hypervisor may be overridden on 936 936 * the boot cpu as arguments are parsed. 937 937 */ 938 - static __cpuinit void init_super_pages(void) 938 + static void init_super_pages(void) 939 939 { 940 940 #ifdef CONFIG_HUGETLB_SUPER_PAGES 941 941 int i; ··· 950 950 * 951 951 * Called from setup_arch() on the boot cpu, or online_secondary(). 952 952 */ 953 - void __cpuinit setup_cpu(int boot) 953 + void setup_cpu(int boot) 954 954 { 955 955 /* The boot cpu sets up its permanent mappings much earlier. */ 956 956 if (!boot)
+4 -4
arch/tile/kernel/smpboot.c
··· 133 133 } 134 134 late_initcall(reset_init_affinity); 135 135 136 - static struct cpumask cpu_started __cpuinitdata; 136 + static struct cpumask cpu_started; 137 137 138 138 /* 139 139 * Activate a secondary processor. Very minimal; don't add anything 140 140 * to this path without knowing what you're doing, since SMP booting 141 141 * is pretty fragile. 142 142 */ 143 - static void __cpuinit start_secondary(void) 143 + static void start_secondary(void) 144 144 { 145 145 int cpuid = smp_processor_id(); 146 146 ··· 183 183 /* 184 184 * Bring a secondary processor online. 185 185 */ 186 - void __cpuinit online_secondary(void) 186 + void online_secondary(void) 187 187 { 188 188 /* 189 189 * low-memory mappings have been cleared, flush them from ··· 210 210 cpu_startup_entry(CPUHP_ONLINE); 211 211 } 212 212 213 - int __cpuinit __cpu_up(unsigned int cpu, struct task_struct *tidle) 213 + int __cpu_up(unsigned int cpu, struct task_struct *tidle) 214 214 { 215 215 /* Wait 5s total for all CPUs for them to come online */ 216 216 static int timeout;
+1 -1
arch/tile/kernel/time.c
··· 159 159 .set_mode = tile_timer_set_mode, 160 160 }; 161 161 162 - void __cpuinit setup_tile_timer(void) 162 + void setup_tile_timer(void) 163 163 { 164 164 struct clock_event_device *evt = &__get_cpu_var(tile_timer); 165 165
+1 -1
arch/x86/include/asm/cpu.h
··· 28 28 #ifdef CONFIG_HOTPLUG_CPU 29 29 extern int arch_register_cpu(int num); 30 30 extern void arch_unregister_cpu(int); 31 - extern void __cpuinit start_cpu0(void); 31 + extern void start_cpu0(void); 32 32 #ifdef CONFIG_DEBUG_HOTPLUG_CPU0 33 33 extern int _debug_hotplug_cpu(int cpu, int action); 34 34 #endif
+2 -2
arch/x86/include/asm/microcode.h
··· 60 60 #ifdef CONFIG_MICROCODE_EARLY 61 61 #define MAX_UCODE_COUNT 128 62 62 extern void __init load_ucode_bsp(void); 63 - extern void __cpuinit load_ucode_ap(void); 63 + extern void load_ucode_ap(void); 64 64 extern int __init save_microcode_in_initrd(void); 65 65 #else 66 66 static inline void __init load_ucode_bsp(void) {} 67 - static inline void __cpuinit load_ucode_ap(void) {} 67 + static inline void load_ucode_ap(void) {} 68 68 static inline int __init save_microcode_in_initrd(void) 69 69 { 70 70 return 0;
+2 -2
arch/x86/include/asm/microcode_amd.h
··· 67 67 extern u8 amd_bsp_mpb[MPB_MAX_SIZE]; 68 68 #endif 69 69 extern void __init load_ucode_amd_bsp(void); 70 - extern void __cpuinit load_ucode_amd_ap(void); 70 + extern void load_ucode_amd_ap(void); 71 71 extern int __init save_microcode_in_initrd_amd(void); 72 72 #else 73 73 static inline void __init load_ucode_amd_bsp(void) {} 74 - static inline void __cpuinit load_ucode_amd_ap(void) {} 74 + static inline void load_ucode_amd_ap(void) {} 75 75 static inline int __init save_microcode_in_initrd_amd(void) { return -EINVAL; } 76 76 #endif 77 77
+2 -2
arch/x86/include/asm/microcode_intel.h
··· 65 65 66 66 #ifdef CONFIG_MICROCODE_INTEL_EARLY 67 67 extern void __init load_ucode_intel_bsp(void); 68 - extern void __cpuinit load_ucode_intel_ap(void); 68 + extern void load_ucode_intel_ap(void); 69 69 extern void show_ucode_info_early(void); 70 70 extern int __init save_microcode_in_initrd_intel(void); 71 71 #else 72 72 static inline __init void load_ucode_intel_bsp(void) {} 73 - static inline __cpuinit void load_ucode_intel_ap(void) {} 73 + static inline void load_ucode_intel_ap(void) {} 74 74 static inline void show_ucode_info_early(void) {} 75 75 static inline int __init save_microcode_in_initrd_intel(void) { return -EINVAL; } 76 76 #endif
+2 -2
arch/x86/include/asm/mmconfig.h
··· 2 2 #define _ASM_X86_MMCONFIG_H 3 3 4 4 #ifdef CONFIG_PCI_MMCONFIG 5 - extern void __cpuinit fam10h_check_enable_mmcfg(void); 6 - extern void __cpuinit check_enable_amd_mmconf_dmi(void); 5 + extern void fam10h_check_enable_mmcfg(void); 6 + extern void check_enable_amd_mmconf_dmi(void); 7 7 #else 8 8 static inline void fam10h_check_enable_mmcfg(void) { } 9 9 static inline void check_enable_amd_mmconf_dmi(void) { }
+1 -1
arch/x86/include/asm/mpspec.h
··· 94 94 #define default_get_smp_config x86_init_uint_noop 95 95 #endif 96 96 97 - void __cpuinit generic_processor_info(int apicid, int version); 97 + void generic_processor_info(int apicid, int version); 98 98 #ifdef CONFIG_ACPI 99 99 extern void mp_register_ioapic(int id, u32 address, u32 gsi_base); 100 100 extern void mp_override_legacy_irq(u8 bus_irq, u8 polarity, u8 trigger,
+3 -3
arch/x86/include/asm/numa.h
··· 39 39 __apicid_to_node[apicid] = node; 40 40 } 41 41 42 - extern int __cpuinit numa_cpu_node(int cpu); 42 + extern int numa_cpu_node(int cpu); 43 43 44 44 #else /* CONFIG_NUMA */ 45 45 static inline void set_apicid_to_node(int apicid, s16 node) ··· 60 60 extern void numa_set_node(int cpu, int node); 61 61 extern void numa_clear_node(int cpu); 62 62 extern void __init init_cpu_to_node(void); 63 - extern void __cpuinit numa_add_cpu(int cpu); 64 - extern void __cpuinit numa_remove_cpu(int cpu); 63 + extern void numa_add_cpu(int cpu); 64 + extern void numa_remove_cpu(int cpu); 65 65 #else /* CONFIG_NUMA */ 66 66 static inline void numa_set_node(int cpu, int node) { } 67 67 static inline void numa_clear_node(int cpu) { }
+1 -1
arch/x86/include/asm/processor.h
··· 164 164 #define cache_line_size() (boot_cpu_data.x86_cache_alignment) 165 165 166 166 extern void cpu_detect(struct cpuinfo_x86 *c); 167 - extern void __cpuinit fpu_detect(struct cpuinfo_x86 *c); 167 + extern void fpu_detect(struct cpuinfo_x86 *c); 168 168 169 169 extern void early_cpu_init(void); 170 170 extern void identify_boot_cpu(void);
+1 -1
arch/x86/include/asm/prom.h
··· 27 27 extern u64 initial_dtb; 28 28 extern void add_dtb(u64 data); 29 29 extern void x86_add_irq_domains(void); 30 - void __cpuinit x86_of_pci_init(void); 30 + void x86_of_pci_init(void); 31 31 void x86_dtb_init(void); 32 32 #else 33 33 static inline void add_dtb(u64 data) { }
+1 -1
arch/x86/include/asm/smp.h
··· 179 179 } 180 180 #endif /* CONFIG_SMP */ 181 181 182 - extern unsigned disabled_cpus __cpuinitdata; 182 + extern unsigned disabled_cpus; 183 183 184 184 #ifdef CONFIG_X86_32_SMP 185 185 /*
+3 -3
arch/x86/kernel/acpi/boot.c
··· 195 195 return 0; 196 196 } 197 197 198 - static void __cpuinit acpi_register_lapic(int id, u8 enabled) 198 + static void acpi_register_lapic(int id, u8 enabled) 199 199 { 200 200 unsigned int ver = 0; 201 201 ··· 607 607 #ifdef CONFIG_ACPI_HOTPLUG_CPU 608 608 #include <acpi/processor.h> 609 609 610 - static void __cpuinit acpi_map_cpu2node(acpi_handle handle, int cpu, int physid) 610 + static void acpi_map_cpu2node(acpi_handle handle, int cpu, int physid) 611 611 { 612 612 #ifdef CONFIG_ACPI_NUMA 613 613 int nid; ··· 620 620 #endif 621 621 } 622 622 623 - static int __cpuinit _acpi_map_lsapic(acpi_handle handle, int *pcpu) 623 + static int _acpi_map_lsapic(acpi_handle handle, int *pcpu) 624 624 { 625 625 struct acpi_buffer buffer = { ACPI_ALLOCATE_BUFFER, NULL }; 626 626 union acpi_object *obj;
+15 -15
arch/x86/kernel/apic/apic.c
··· 58 58 59 59 unsigned int num_processors; 60 60 61 - unsigned disabled_cpus __cpuinitdata; 61 + unsigned disabled_cpus; 62 62 63 63 /* Processor that is doing the boot up */ 64 64 unsigned int boot_cpu_physical_apicid = -1U; ··· 544 544 * Setup the local APIC timer for this CPU. Copy the initialized values 545 545 * of the boot CPU and register the clock event in the framework. 546 546 */ 547 - static void __cpuinit setup_APIC_timer(void) 547 + static void setup_APIC_timer(void) 548 548 { 549 549 struct clock_event_device *levt = &__get_cpu_var(lapic_events); 550 550 ··· 866 866 setup_APIC_timer(); 867 867 } 868 868 869 - void __cpuinit setup_secondary_APIC_clock(void) 869 + void setup_secondary_APIC_clock(void) 870 870 { 871 871 setup_APIC_timer(); 872 872 } ··· 1229 1229 apic_write(APIC_LVT1, value); 1230 1230 } 1231 1231 1232 - static void __cpuinit lapic_setup_esr(void) 1232 + static void lapic_setup_esr(void) 1233 1233 { 1234 1234 unsigned int oldvalue, value, maxlvt; 1235 1235 ··· 1276 1276 * Used to setup local APIC while initializing BSP or bringin up APs. 1277 1277 * Always called with preemption disabled. 1278 1278 */ 1279 - void __cpuinit setup_local_APIC(void) 1279 + void setup_local_APIC(void) 1280 1280 { 1281 1281 int cpu = smp_processor_id(); 1282 1282 unsigned int value, queued; ··· 1471 1471 #endif 1472 1472 } 1473 1473 1474 - void __cpuinit end_local_APIC_setup(void) 1474 + void end_local_APIC_setup(void) 1475 1475 { 1476 1476 lapic_setup_esr(); 1477 1477 ··· 2107 2107 apic_write(APIC_LVT1, value); 2108 2108 } 2109 2109 2110 - void __cpuinit generic_processor_info(int apicid, int version) 2110 + void generic_processor_info(int apicid, int version) 2111 2111 { 2112 2112 int cpu, max = nr_cpu_ids; 2113 2113 bool boot_cpu_detected = physid_isset(boot_cpu_physical_apicid, ··· 2377 2377 .suspend = lapic_suspend, 2378 2378 }; 2379 2379 2380 - static void __cpuinit apic_pm_activate(void) 2380 + static void apic_pm_activate(void) 2381 2381 { 2382 2382 apic_pm_state.active = 1; 2383 2383 } ··· 2402 2402 2403 2403 #ifdef CONFIG_X86_64 2404 2404 2405 - static int __cpuinit apic_cluster_num(void) 2405 + static int apic_cluster_num(void) 2406 2406 { 2407 2407 int i, clusters, zeros; 2408 2408 unsigned id; ··· 2447 2447 return clusters; 2448 2448 } 2449 2449 2450 - static int __cpuinitdata multi_checked; 2451 - static int __cpuinitdata multi; 2450 + static int multi_checked; 2451 + static int multi; 2452 2452 2453 - static int __cpuinit set_multi(const struct dmi_system_id *d) 2453 + static int set_multi(const struct dmi_system_id *d) 2454 2454 { 2455 2455 if (multi) 2456 2456 return 0; ··· 2459 2459 return 0; 2460 2460 } 2461 2461 2462 - static const __cpuinitconst struct dmi_system_id multi_dmi_table[] = { 2462 + static const struct dmi_system_id multi_dmi_table[] = { 2463 2463 { 2464 2464 .callback = set_multi, 2465 2465 .ident = "IBM System Summit2", ··· 2471 2471 {} 2472 2472 }; 2473 2473 2474 - static void __cpuinit dmi_check_multi(void) 2474 + static void dmi_check_multi(void) 2475 2475 { 2476 2476 if (multi_checked) 2477 2477 return; ··· 2488 2488 * multi-chassis. 2489 2489 * Use DMI to check them 2490 2490 */ 2491 - __cpuinit int apic_is_clustered_box(void) 2491 + int apic_is_clustered_box(void) 2492 2492 { 2493 2493 dmi_check_multi(); 2494 2494 if (multi)
+1 -1
arch/x86/kernel/apic/apic_numachip.c
··· 74 74 return initial_apic_id >> index_msb; 75 75 } 76 76 77 - static int __cpuinit numachip_wakeup_secondary(int phys_apicid, unsigned long start_rip) 77 + static int numachip_wakeup_secondary(int phys_apicid, unsigned long start_rip) 78 78 { 79 79 union numachip_csr_g3_ext_irq_gen int_gen; 80 80
+1 -1
arch/x86/kernel/apic/es7000_32.c
··· 130 130 */ 131 131 132 132 133 - static int __cpuinit wakeup_secondary_cpu_via_mip(int cpu, unsigned long eip) 133 + static int wakeup_secondary_cpu_via_mip(int cpu, unsigned long eip) 134 134 { 135 135 unsigned long vect = 0, psaival = 0; 136 136
+1 -1
arch/x86/kernel/apic/numaq_32.c
··· 105 105 } 106 106 } 107 107 108 - void __cpuinit numaq_tsc_disable(void) 108 + void numaq_tsc_disable(void) 109 109 { 110 110 if (!found_numaq) 111 111 return;
+1 -1
arch/x86/kernel/apic/x2apic_cluster.c
··· 148 148 /* 149 149 * At CPU state changes, update the x2apic cluster sibling info. 150 150 */ 151 - static int __cpuinit 151 + static int 152 152 update_clusterinfo(struct notifier_block *nfb, unsigned long action, void *hcpu) 153 153 { 154 154 unsigned int this_cpu = (unsigned long)hcpu;
+7 -7
arch/x86/kernel/apic/x2apic_uv_x.c
··· 209 209 unsigned long sn_rtc_cycles_per_second; 210 210 EXPORT_SYMBOL(sn_rtc_cycles_per_second); 211 211 212 - static int __cpuinit uv_wakeup_secondary(int phys_apicid, unsigned long start_rip) 212 + static int uv_wakeup_secondary(int phys_apicid, unsigned long start_rip) 213 213 { 214 214 #ifdef CONFIG_SMP 215 215 unsigned long val; ··· 416 416 .safe_wait_icr_idle = native_safe_x2apic_wait_icr_idle, 417 417 }; 418 418 419 - static __cpuinit void set_x2apic_extra_bits(int pnode) 419 + static void set_x2apic_extra_bits(int pnode) 420 420 { 421 421 __this_cpu_write(x2apic_extra_bits, pnode << uvh_apicid.s.pnode_shift); 422 422 } ··· 735 735 mod_timer_pinned(timer, jiffies + SCIR_CPU_HB_INTERVAL); 736 736 } 737 737 738 - static void __cpuinit uv_heartbeat_enable(int cpu) 738 + static void uv_heartbeat_enable(int cpu) 739 739 { 740 740 while (!uv_cpu_hub_info(cpu)->scir.enabled) { 741 741 struct timer_list *timer = &uv_cpu_hub_info(cpu)->scir.timer; ··· 752 752 } 753 753 754 754 #ifdef CONFIG_HOTPLUG_CPU 755 - static void __cpuinit uv_heartbeat_disable(int cpu) 755 + static void uv_heartbeat_disable(int cpu) 756 756 { 757 757 if (uv_cpu_hub_info(cpu)->scir.enabled) { 758 758 uv_cpu_hub_info(cpu)->scir.enabled = 0; ··· 764 764 /* 765 765 * cpu hotplug notifier 766 766 */ 767 - static __cpuinit int uv_scir_cpu_notify(struct notifier_block *self, 768 - unsigned long action, void *hcpu) 767 + static int uv_scir_cpu_notify(struct notifier_block *self, unsigned long action, 768 + void *hcpu) 769 769 { 770 770 long cpu = (long)hcpu; 771 771 ··· 835 835 * Called on each cpu to initialize the per_cpu UV data area. 836 836 * FIXME: hotplug not supported yet 837 837 */ 838 - void __cpuinit uv_cpu_init(void) 838 + void uv_cpu_init(void) 839 839 { 840 840 /* CPU 0 initilization will be done via uv_system_init. */ 841 841 if (!uv_blade_info)
+16 -17
arch/x86/kernel/cpu/amd.c
··· 69 69 extern void vide(void); 70 70 __asm__(".align 4\nvide: ret"); 71 71 72 - static void __cpuinit init_amd_k5(struct cpuinfo_x86 *c) 72 + static void init_amd_k5(struct cpuinfo_x86 *c) 73 73 { 74 74 /* 75 75 * General Systems BIOSen alias the cpu frequency registers ··· 87 87 } 88 88 89 89 90 - static void __cpuinit init_amd_k6(struct cpuinfo_x86 *c) 90 + static void init_amd_k6(struct cpuinfo_x86 *c) 91 91 { 92 92 u32 l, h; 93 93 int mbytes = get_num_physpages() >> (20-PAGE_SHIFT); ··· 179 179 } 180 180 } 181 181 182 - static void __cpuinit amd_k7_smp_check(struct cpuinfo_x86 *c) 182 + static void amd_k7_smp_check(struct cpuinfo_x86 *c) 183 183 { 184 184 /* calling is from identify_secondary_cpu() ? */ 185 185 if (!c->cpu_index) ··· 222 222 add_taint(TAINT_UNSAFE_SMP, LOCKDEP_NOW_UNRELIABLE); 223 223 } 224 224 225 - static void __cpuinit init_amd_k7(struct cpuinfo_x86 *c) 225 + static void init_amd_k7(struct cpuinfo_x86 *c) 226 226 { 227 227 u32 l, h; 228 228 ··· 267 267 * To workaround broken NUMA config. Read the comment in 268 268 * srat_detect_node(). 269 269 */ 270 - static int __cpuinit nearby_node(int apicid) 270 + static int nearby_node(int apicid) 271 271 { 272 272 int i, node; 273 273 ··· 292 292 * (2) AMD processors supporting compute units 293 293 */ 294 294 #ifdef CONFIG_X86_HT 295 - static void __cpuinit amd_get_topology(struct cpuinfo_x86 *c) 295 + static void amd_get_topology(struct cpuinfo_x86 *c) 296 296 { 297 297 u32 nodes, cores_per_cu = 1; 298 298 u8 node_id; ··· 342 342 * On a AMD dual core setup the lower bits of the APIC id distingush the cores. 343 343 * Assumes number of cores is a power of two. 344 344 */ 345 - static void __cpuinit amd_detect_cmp(struct cpuinfo_x86 *c) 345 + static void amd_detect_cmp(struct cpuinfo_x86 *c) 346 346 { 347 347 #ifdef CONFIG_X86_HT 348 348 unsigned bits; ··· 369 369 } 370 370 EXPORT_SYMBOL_GPL(amd_get_nb_id); 371 371 372 - static void __cpuinit srat_detect_node(struct cpuinfo_x86 *c) 372 + static void srat_detect_node(struct cpuinfo_x86 *c) 373 373 { 374 374 #ifdef CONFIG_NUMA 375 375 int cpu = smp_processor_id(); ··· 421 421 #endif 422 422 } 423 423 424 - static void __cpuinit early_init_amd_mc(struct cpuinfo_x86 *c) 424 + static void early_init_amd_mc(struct cpuinfo_x86 *c) 425 425 { 426 426 #ifdef CONFIG_X86_HT 427 427 unsigned bits, ecx; ··· 447 447 #endif 448 448 } 449 449 450 - static void __cpuinit bsp_init_amd(struct cpuinfo_x86 *c) 450 + static void bsp_init_amd(struct cpuinfo_x86 *c) 451 451 { 452 452 if (cpu_has(c, X86_FEATURE_CONSTANT_TSC)) { 453 453 ··· 475 475 } 476 476 } 477 477 478 - static void __cpuinit early_init_amd(struct cpuinfo_x86 *c) 478 + static void early_init_amd(struct cpuinfo_x86 *c) 479 479 { 480 480 early_init_amd_mc(c); 481 481 ··· 514 514 static const int amd_erratum_400[]; 515 515 static bool cpu_has_amd_erratum(const int *erratum); 516 516 517 - static void __cpuinit init_amd(struct cpuinfo_x86 *c) 517 + static void init_amd(struct cpuinfo_x86 *c) 518 518 { 519 519 u32 dummy; 520 520 unsigned long long value; ··· 740 740 } 741 741 742 742 #ifdef CONFIG_X86_32 743 - static unsigned int __cpuinit amd_size_cache(struct cpuinfo_x86 *c, 744 - unsigned int size) 743 + static unsigned int amd_size_cache(struct cpuinfo_x86 *c, unsigned int size) 745 744 { 746 745 /* AMD errata T13 (order #21922) */ 747 746 if ((c->x86 == 6)) { ··· 756 757 } 757 758 #endif 758 759 759 - static void __cpuinit cpu_set_tlb_flushall_shift(struct cpuinfo_x86 *c) 760 + static void cpu_set_tlb_flushall_shift(struct cpuinfo_x86 *c) 760 761 { 761 762 tlb_flushall_shift = 5; 762 763 ··· 764 765 tlb_flushall_shift = 4; 765 766 } 766 767 767 - static void __cpuinit cpu_detect_tlb_amd(struct cpuinfo_x86 *c) 768 + static void cpu_detect_tlb_amd(struct cpuinfo_x86 *c) 768 769 { 769 770 u32 ebx, eax, ecx, edx; 770 771 u16 mask = 0xfff; ··· 819 820 cpu_set_tlb_flushall_shift(c); 820 821 } 821 822 822 - static const struct cpu_dev __cpuinitconst amd_cpu_dev = { 823 + static const struct cpu_dev amd_cpu_dev = { 823 824 .c_vendor = "AMD", 824 825 .c_ident = { "AuthenticAMD" }, 825 826 #ifdef CONFIG_X86_32
+13 -13
arch/x86/kernel/cpu/centaur.c
··· 11 11 12 12 #ifdef CONFIG_X86_OOSTORE 13 13 14 - static u32 __cpuinit power2(u32 x) 14 + static u32 power2(u32 x) 15 15 { 16 16 u32 s = 1; 17 17 ··· 25 25 /* 26 26 * Set up an actual MCR 27 27 */ 28 - static void __cpuinit centaur_mcr_insert(int reg, u32 base, u32 size, int key) 28 + static void centaur_mcr_insert(int reg, u32 base, u32 size, int key) 29 29 { 30 30 u32 lo, hi; 31 31 ··· 42 42 * 43 43 * Shortcut: We know you can't put 4Gig of RAM on a winchip 44 44 */ 45 - static u32 __cpuinit ramtop(void) 45 + static u32 ramtop(void) 46 46 { 47 47 u32 clip = 0xFFFFFFFFUL; 48 48 u32 top = 0; ··· 91 91 /* 92 92 * Compute a set of MCR's to give maximum coverage 93 93 */ 94 - static int __cpuinit centaur_mcr_compute(int nr, int key) 94 + static int centaur_mcr_compute(int nr, int key) 95 95 { 96 96 u32 mem = ramtop(); 97 97 u32 root = power2(mem); ··· 157 157 return ct; 158 158 } 159 159 160 - static void __cpuinit centaur_create_optimal_mcr(void) 160 + static void centaur_create_optimal_mcr(void) 161 161 { 162 162 int used; 163 163 int i; ··· 181 181 wrmsr(MSR_IDT_MCR0+i, 0, 0); 182 182 } 183 183 184 - static void __cpuinit winchip2_create_optimal_mcr(void) 184 + static void winchip2_create_optimal_mcr(void) 185 185 { 186 186 u32 lo, hi; 187 187 int used; ··· 217 217 /* 218 218 * Handle the MCR key on the Winchip 2. 219 219 */ 220 - static void __cpuinit winchip2_unprotect_mcr(void) 220 + static void winchip2_unprotect_mcr(void) 221 221 { 222 222 u32 lo, hi; 223 223 u32 key; ··· 229 229 wrmsr(MSR_IDT_MCR_CTRL, lo, hi); 230 230 } 231 231 232 - static void __cpuinit winchip2_protect_mcr(void) 232 + static void winchip2_protect_mcr(void) 233 233 { 234 234 u32 lo, hi; 235 235 ··· 247 247 #define RNG_ENABLED (1 << 3) 248 248 #define RNG_ENABLE (1 << 6) /* MSR_VIA_RNG */ 249 249 250 - static void __cpuinit init_c3(struct cpuinfo_x86 *c) 250 + static void init_c3(struct cpuinfo_x86 *c) 251 251 { 252 252 u32 lo, hi; 253 253 ··· 318 318 EAMD3D = 1<<20, 319 319 }; 320 320 321 - static void __cpuinit early_init_centaur(struct cpuinfo_x86 *c) 321 + static void early_init_centaur(struct cpuinfo_x86 *c) 322 322 { 323 323 switch (c->x86) { 324 324 #ifdef CONFIG_X86_32 ··· 337 337 #endif 338 338 } 339 339 340 - static void __cpuinit init_centaur(struct cpuinfo_x86 *c) 340 + static void init_centaur(struct cpuinfo_x86 *c) 341 341 { 342 342 #ifdef CONFIG_X86_32 343 343 char *name; ··· 468 468 #endif 469 469 } 470 470 471 - static unsigned int __cpuinit 471 + static unsigned int 472 472 centaur_size_cache(struct cpuinfo_x86 *c, unsigned int size) 473 473 { 474 474 #ifdef CONFIG_X86_32 ··· 488 488 return size; 489 489 } 490 490 491 - static const struct cpu_dev __cpuinitconst centaur_cpu_dev = { 491 + static const struct cpu_dev centaur_cpu_dev = { 492 492 .c_vendor = "Centaur", 493 493 .c_ident = { "CentaurHauls" }, 494 494 .c_early_init = early_init_centaur,
+32 -32
arch/x86/kernel/cpu/common.c
··· 63 63 alloc_bootmem_cpumask_var(&cpu_sibling_setup_mask); 64 64 } 65 65 66 - static void __cpuinit default_init(struct cpuinfo_x86 *c) 66 + static void default_init(struct cpuinfo_x86 *c) 67 67 { 68 68 #ifdef CONFIG_X86_64 69 69 cpu_detect_cache_sizes(c); ··· 80 80 #endif 81 81 } 82 82 83 - static const struct cpu_dev __cpuinitconst default_cpu = { 83 + static const struct cpu_dev default_cpu = { 84 84 .c_init = default_init, 85 85 .c_vendor = "Unknown", 86 86 .c_x86_vendor = X86_VENDOR_UNKNOWN, 87 87 }; 88 88 89 - static const struct cpu_dev *this_cpu __cpuinitdata = &default_cpu; 89 + static const struct cpu_dev *this_cpu = &default_cpu; 90 90 91 91 DEFINE_PER_CPU_PAGE_ALIGNED(struct gdt_page, gdt_page) = { .gdt = { 92 92 #ifdef CONFIG_X86_64 ··· 160 160 __setup("noxsaveopt", x86_xsaveopt_setup); 161 161 162 162 #ifdef CONFIG_X86_32 163 - static int cachesize_override __cpuinitdata = -1; 164 - static int disable_x86_serial_nr __cpuinitdata = 1; 163 + static int cachesize_override = -1; 164 + static int disable_x86_serial_nr = 1; 165 165 166 166 static int __init cachesize_setup(char *str) 167 167 { ··· 215 215 } 216 216 217 217 /* Probe for the CPUID instruction */ 218 - int __cpuinit have_cpuid_p(void) 218 + int have_cpuid_p(void) 219 219 { 220 220 return flag_is_changeable_p(X86_EFLAGS_ID); 221 221 } 222 222 223 - static void __cpuinit squash_the_stupid_serial_number(struct cpuinfo_x86 *c) 223 + static void squash_the_stupid_serial_number(struct cpuinfo_x86 *c) 224 224 { 225 225 unsigned long lo, hi; 226 226 ··· 298 298 u32 level; 299 299 }; 300 300 301 - static const struct cpuid_dependent_feature __cpuinitconst 301 + static const struct cpuid_dependent_feature 302 302 cpuid_dependent_features[] = { 303 303 { X86_FEATURE_MWAIT, 0x00000005 }, 304 304 { X86_FEATURE_DCA, 0x00000009 }, ··· 306 306 { 0, 0 } 307 307 }; 308 308 309 - static void __cpuinit filter_cpuid_features(struct cpuinfo_x86 *c, bool warn) 309 + static void filter_cpuid_features(struct cpuinfo_x86 *c, bool warn) 310 310 { 311 311 const struct cpuid_dependent_feature *df; 312 312 ··· 344 344 */ 345 345 346 346 /* Look up CPU names by table lookup. */ 347 - static const char *__cpuinit table_lookup_model(struct cpuinfo_x86 *c) 347 + static const char *table_lookup_model(struct cpuinfo_x86 *c) 348 348 { 349 349 const struct cpu_model_info *info; 350 350 ··· 364 364 return NULL; /* Not found */ 365 365 } 366 366 367 - __u32 cpu_caps_cleared[NCAPINTS] __cpuinitdata; 368 - __u32 cpu_caps_set[NCAPINTS] __cpuinitdata; 367 + __u32 cpu_caps_cleared[NCAPINTS]; 368 + __u32 cpu_caps_set[NCAPINTS]; 369 369 370 370 void load_percpu_segment(int cpu) 371 371 { ··· 394 394 load_percpu_segment(cpu); 395 395 } 396 396 397 - static const struct cpu_dev *__cpuinitdata cpu_devs[X86_VENDOR_NUM] = {}; 397 + static const struct cpu_dev *cpu_devs[X86_VENDOR_NUM] = {}; 398 398 399 - static void __cpuinit get_model_name(struct cpuinfo_x86 *c) 399 + static void get_model_name(struct cpuinfo_x86 *c) 400 400 { 401 401 unsigned int *v; 402 402 char *p, *q; ··· 425 425 } 426 426 } 427 427 428 - void __cpuinit cpu_detect_cache_sizes(struct cpuinfo_x86 *c) 428 + void cpu_detect_cache_sizes(struct cpuinfo_x86 *c) 429 429 { 430 430 unsigned int n, dummy, ebx, ecx, edx, l2size; 431 431 ··· 479 479 */ 480 480 s8 __read_mostly tlb_flushall_shift = -1; 481 481 482 - void __cpuinit cpu_detect_tlb(struct cpuinfo_x86 *c) 482 + void cpu_detect_tlb(struct cpuinfo_x86 *c) 483 483 { 484 484 if (this_cpu->c_detect_tlb) 485 485 this_cpu->c_detect_tlb(c); ··· 493 493 tlb_flushall_shift); 494 494 } 495 495 496 - void __cpuinit detect_ht(struct cpuinfo_x86 *c) 496 + void detect_ht(struct cpuinfo_x86 *c) 497 497 { 498 498 #ifdef CONFIG_X86_HT 499 499 u32 eax, ebx, ecx, edx; ··· 544 544 #endif 545 545 } 546 546 547 - static void __cpuinit get_cpu_vendor(struct cpuinfo_x86 *c) 547 + static void get_cpu_vendor(struct cpuinfo_x86 *c) 548 548 { 549 549 char *v = c->x86_vendor_id; 550 550 int i; ··· 571 571 this_cpu = &default_cpu; 572 572 } 573 573 574 - void __cpuinit cpu_detect(struct cpuinfo_x86 *c) 574 + void cpu_detect(struct cpuinfo_x86 *c) 575 575 { 576 576 /* Get vendor name */ 577 577 cpuid(0x00000000, (unsigned int *)&c->cpuid_level, ··· 601 601 } 602 602 } 603 603 604 - void __cpuinit get_cpu_cap(struct cpuinfo_x86 *c) 604 + void get_cpu_cap(struct cpuinfo_x86 *c) 605 605 { 606 606 u32 tfms, xlvl; 607 607 u32 ebx; ··· 652 652 init_scattered_cpuid_features(c); 653 653 } 654 654 655 - static void __cpuinit identify_cpu_without_cpuid(struct cpuinfo_x86 *c) 655 + static void identify_cpu_without_cpuid(struct cpuinfo_x86 *c) 656 656 { 657 657 #ifdef CONFIG_X86_32 658 658 int i; ··· 769 769 * unless we can find a reliable way to detect all the broken cases. 770 770 * Enable it explicitly on 64-bit for non-constant inputs of cpu_has(). 771 771 */ 772 - static void __cpuinit detect_nopl(struct cpuinfo_x86 *c) 772 + static void detect_nopl(struct cpuinfo_x86 *c) 773 773 { 774 774 #ifdef CONFIG_X86_32 775 775 clear_cpu_cap(c, X86_FEATURE_NOPL); ··· 778 778 #endif 779 779 } 780 780 781 - static void __cpuinit generic_identify(struct cpuinfo_x86 *c) 781 + static void generic_identify(struct cpuinfo_x86 *c) 782 782 { 783 783 c->extended_cpuid_level = 0; 784 784 ··· 815 815 /* 816 816 * This does the hard work of actually picking apart the CPU stuff... 817 817 */ 818 - static void __cpuinit identify_cpu(struct cpuinfo_x86 *c) 818 + static void identify_cpu(struct cpuinfo_x86 *c) 819 819 { 820 820 int i; 821 821 ··· 960 960 cpu_detect_tlb(&boot_cpu_data); 961 961 } 962 962 963 - void __cpuinit identify_secondary_cpu(struct cpuinfo_x86 *c) 963 + void identify_secondary_cpu(struct cpuinfo_x86 *c) 964 964 { 965 965 BUG_ON(c == &boot_cpu_data); 966 966 identify_cpu(c); ··· 975 975 unsigned max; 976 976 }; 977 977 978 - static const struct msr_range msr_range_array[] __cpuinitconst = { 978 + static const struct msr_range msr_range_array[] = { 979 979 { 0x00000000, 0x00000418}, 980 980 { 0xc0000000, 0xc000040b}, 981 981 { 0xc0010000, 0xc0010142}, 982 982 { 0xc0011000, 0xc001103b}, 983 983 }; 984 984 985 - static void __cpuinit __print_cpu_msr(void) 985 + static void __print_cpu_msr(void) 986 986 { 987 987 unsigned index_min, index_max; 988 988 unsigned index; ··· 1001 1001 } 1002 1002 } 1003 1003 1004 - static int show_msr __cpuinitdata; 1004 + static int show_msr; 1005 1005 1006 1006 static __init int setup_show_msr(char *arg) 1007 1007 { ··· 1022 1022 } 1023 1023 __setup("noclflush", setup_noclflush); 1024 1024 1025 - void __cpuinit print_cpu_info(struct cpuinfo_x86 *c) 1025 + void print_cpu_info(struct cpuinfo_x86 *c) 1026 1026 { 1027 1027 const char *vendor = NULL; 1028 1028 ··· 1051 1051 print_cpu_msr(c); 1052 1052 } 1053 1053 1054 - void __cpuinit print_cpu_msr(struct cpuinfo_x86 *c) 1054 + void print_cpu_msr(struct cpuinfo_x86 *c) 1055 1055 { 1056 1056 if (c->cpu_index < show_msr) 1057 1057 __print_cpu_msr(); ··· 1216 1216 */ 1217 1217 #ifdef CONFIG_X86_64 1218 1218 1219 - void __cpuinit cpu_init(void) 1219 + void cpu_init(void) 1220 1220 { 1221 1221 struct orig_ist *oist; 1222 1222 struct task_struct *me; ··· 1315 1315 1316 1316 #else 1317 1317 1318 - void __cpuinit cpu_init(void) 1318 + void cpu_init(void) 1319 1319 { 1320 1320 int cpu = smp_processor_id(); 1321 1321 struct task_struct *curr = current;
+20 -20
arch/x86/kernel/cpu/cyrix.c
··· 15 15 /* 16 16 * Read NSC/Cyrix DEVID registers (DIR) to get more detailed info. about the CPU 17 17 */ 18 - static void __cpuinit __do_cyrix_devid(unsigned char *dir0, unsigned char *dir1) 18 + static void __do_cyrix_devid(unsigned char *dir0, unsigned char *dir1) 19 19 { 20 20 unsigned char ccr2, ccr3; 21 21 ··· 44 44 } 45 45 } 46 46 47 - static void __cpuinit do_cyrix_devid(unsigned char *dir0, unsigned char *dir1) 47 + static void do_cyrix_devid(unsigned char *dir0, unsigned char *dir1) 48 48 { 49 49 unsigned long flags; 50 50 ··· 59 59 * Actually since bugs.h doesn't even reference this perhaps someone should 60 60 * fix the documentation ??? 61 61 */ 62 - static unsigned char Cx86_dir0_msb __cpuinitdata = 0; 62 + static unsigned char Cx86_dir0_msb = 0; 63 63 64 - static const char __cpuinitconst Cx86_model[][9] = { 64 + static const char Cx86_model[][9] = { 65 65 "Cx486", "Cx486", "5x86 ", "6x86", "MediaGX ", "6x86MX ", 66 66 "M II ", "Unknown" 67 67 }; 68 - static const char __cpuinitconst Cx486_name[][5] = { 68 + static const char Cx486_name[][5] = { 69 69 "SLC", "DLC", "SLC2", "DLC2", "SRx", "DRx", 70 70 "SRx2", "DRx2" 71 71 }; 72 - static const char __cpuinitconst Cx486S_name[][4] = { 72 + static const char Cx486S_name[][4] = { 73 73 "S", "S2", "Se", "S2e" 74 74 }; 75 - static const char __cpuinitconst Cx486D_name[][4] = { 75 + static const char Cx486D_name[][4] = { 76 76 "DX", "DX2", "?", "?", "?", "DX4" 77 77 }; 78 - static char Cx86_cb[] __cpuinitdata = "?.5x Core/Bus Clock"; 79 - static const char __cpuinitconst cyrix_model_mult1[] = "12??43"; 80 - static const char __cpuinitconst cyrix_model_mult2[] = "12233445"; 78 + static char Cx86_cb[] = "?.5x Core/Bus Clock"; 79 + static const char cyrix_model_mult1[] = "12??43"; 80 + static const char cyrix_model_mult2[] = "12233445"; 81 81 82 82 /* 83 83 * Reset the slow-loop (SLOP) bit on the 686(L) which is set by some old ··· 87 87 * FIXME: our newer udelay uses the tsc. We don't need to frob with SLOP 88 88 */ 89 89 90 - static void __cpuinit check_cx686_slop(struct cpuinfo_x86 *c) 90 + static void check_cx686_slop(struct cpuinfo_x86 *c) 91 91 { 92 92 unsigned long flags; 93 93 ··· 112 112 } 113 113 114 114 115 - static void __cpuinit set_cx86_reorder(void) 115 + static void set_cx86_reorder(void) 116 116 { 117 117 u8 ccr3; 118 118 ··· 127 127 setCx86(CX86_CCR3, ccr3); 128 128 } 129 129 130 - static void __cpuinit set_cx86_memwb(void) 130 + static void set_cx86_memwb(void) 131 131 { 132 132 printk(KERN_INFO "Enable Memory-Write-back mode on Cyrix/NSC processor.\n"); 133 133 ··· 143 143 * Configure later MediaGX and/or Geode processor. 144 144 */ 145 145 146 - static void __cpuinit geode_configure(void) 146 + static void geode_configure(void) 147 147 { 148 148 unsigned long flags; 149 149 u8 ccr3; ··· 166 166 local_irq_restore(flags); 167 167 } 168 168 169 - static void __cpuinit early_init_cyrix(struct cpuinfo_x86 *c) 169 + static void early_init_cyrix(struct cpuinfo_x86 *c) 170 170 { 171 171 unsigned char dir0, dir0_msn, dir1 = 0; 172 172 ··· 185 185 } 186 186 } 187 187 188 - static void __cpuinit init_cyrix(struct cpuinfo_x86 *c) 188 + static void init_cyrix(struct cpuinfo_x86 *c) 189 189 { 190 190 unsigned char dir0, dir0_msn, dir0_lsn, dir1 = 0; 191 191 char *buf = c->x86_model_id; ··· 356 356 /* 357 357 * Handle National Semiconductor branded processors 358 358 */ 359 - static void __cpuinit init_nsc(struct cpuinfo_x86 *c) 359 + static void init_nsc(struct cpuinfo_x86 *c) 360 360 { 361 361 /* 362 362 * There may be GX1 processors in the wild that are branded ··· 405 405 return (unsigned char) (test >> 8) == 0x02; 406 406 } 407 407 408 - static void __cpuinit cyrix_identify(struct cpuinfo_x86 *c) 408 + static void cyrix_identify(struct cpuinfo_x86 *c) 409 409 { 410 410 /* Detect Cyrix with disabled CPUID */ 411 411 if (c->x86 == 4 && test_cyrix_52div()) { ··· 441 441 } 442 442 } 443 443 444 - static const struct cpu_dev __cpuinitconst cyrix_cpu_dev = { 444 + static const struct cpu_dev cyrix_cpu_dev = { 445 445 .c_vendor = "Cyrix", 446 446 .c_ident = { "CyrixInstead" }, 447 447 .c_early_init = early_init_cyrix, ··· 452 452 453 453 cpu_dev_register(cyrix_cpu_dev); 454 454 455 - static const struct cpu_dev __cpuinitconst nsc_cpu_dev = { 455 + static const struct cpu_dev nsc_cpu_dev = { 456 456 .c_vendor = "NSC", 457 457 .c_ident = { "Geode by NSC" }, 458 458 .c_init = init_nsc,
+1 -1
arch/x86/kernel/cpu/hypervisor.c
··· 60 60 } 61 61 } 62 62 63 - void __cpuinit init_hypervisor(struct cpuinfo_x86 *c) 63 + void init_hypervisor(struct cpuinfo_x86 *c) 64 64 { 65 65 if (x86_hyper && x86_hyper->set_cpu_features) 66 66 x86_hyper->set_cpu_features(c);
+15 -15
arch/x86/kernel/cpu/intel.c
··· 26 26 #include <asm/apic.h> 27 27 #endif 28 28 29 - static void __cpuinit early_init_intel(struct cpuinfo_x86 *c) 29 + static void early_init_intel(struct cpuinfo_x86 *c) 30 30 { 31 31 u64 misc_enable; 32 32 ··· 163 163 * This is called before we do cpu ident work 164 164 */ 165 165 166 - int __cpuinit ppro_with_ram_bug(void) 166 + int ppro_with_ram_bug(void) 167 167 { 168 168 /* Uses data from early_cpu_detect now */ 169 169 if (boot_cpu_data.x86_vendor == X86_VENDOR_INTEL && ··· 176 176 return 0; 177 177 } 178 178 179 - static void __cpuinit intel_smp_check(struct cpuinfo_x86 *c) 179 + static void intel_smp_check(struct cpuinfo_x86 *c) 180 180 { 181 181 /* calling is from identify_secondary_cpu() ? */ 182 182 if (!c->cpu_index) ··· 196 196 } 197 197 } 198 198 199 - static void __cpuinit intel_workarounds(struct cpuinfo_x86 *c) 199 + static void intel_workarounds(struct cpuinfo_x86 *c) 200 200 { 201 201 unsigned long lo, hi; 202 202 ··· 275 275 intel_smp_check(c); 276 276 } 277 277 #else 278 - static void __cpuinit intel_workarounds(struct cpuinfo_x86 *c) 278 + static void intel_workarounds(struct cpuinfo_x86 *c) 279 279 { 280 280 } 281 281 #endif 282 282 283 - static void __cpuinit srat_detect_node(struct cpuinfo_x86 *c) 283 + static void srat_detect_node(struct cpuinfo_x86 *c) 284 284 { 285 285 #ifdef CONFIG_NUMA 286 286 unsigned node; ··· 300 300 /* 301 301 * find out the number of processor cores on the die 302 302 */ 303 - static int __cpuinit intel_num_cpu_cores(struct cpuinfo_x86 *c) 303 + static int intel_num_cpu_cores(struct cpuinfo_x86 *c) 304 304 { 305 305 unsigned int eax, ebx, ecx, edx; 306 306 ··· 315 315 return 1; 316 316 } 317 317 318 - static void __cpuinit detect_vmx_virtcap(struct cpuinfo_x86 *c) 318 + static void detect_vmx_virtcap(struct cpuinfo_x86 *c) 319 319 { 320 320 /* Intel VMX MSR indicated features */ 321 321 #define X86_VMX_FEATURE_PROC_CTLS_TPR_SHADOW 0x00200000 ··· 353 353 } 354 354 } 355 355 356 - static void __cpuinit init_intel(struct cpuinfo_x86 *c) 356 + static void init_intel(struct cpuinfo_x86 *c) 357 357 { 358 358 unsigned int l2 = 0; 359 359 ··· 472 472 } 473 473 474 474 #ifdef CONFIG_X86_32 475 - static unsigned int __cpuinit intel_size_cache(struct cpuinfo_x86 *c, unsigned int size) 475 + static unsigned int intel_size_cache(struct cpuinfo_x86 *c, unsigned int size) 476 476 { 477 477 /* 478 478 * Intel PIII Tualatin. This comes in two flavours. ··· 506 506 507 507 #define STLB_4K 0x41 508 508 509 - static const struct _tlb_table intel_tlb_table[] __cpuinitconst = { 509 + static const struct _tlb_table intel_tlb_table[] = { 510 510 { 0x01, TLB_INST_4K, 32, " TLB_INST 4 KByte pages, 4-way set associative" }, 511 511 { 0x02, TLB_INST_4M, 2, " TLB_INST 4 MByte pages, full associative" }, 512 512 { 0x03, TLB_DATA_4K, 64, " TLB_DATA 4 KByte pages, 4-way set associative" }, ··· 536 536 { 0x00, 0, 0 } 537 537 }; 538 538 539 - static void __cpuinit intel_tlb_lookup(const unsigned char desc) 539 + static void intel_tlb_lookup(const unsigned char desc) 540 540 { 541 541 unsigned char k; 542 542 if (desc == 0) ··· 605 605 } 606 606 } 607 607 608 - static void __cpuinit intel_tlb_flushall_shift_set(struct cpuinfo_x86 *c) 608 + static void intel_tlb_flushall_shift_set(struct cpuinfo_x86 *c) 609 609 { 610 610 switch ((c->x86 << 8) + c->x86_model) { 611 611 case 0x60f: /* original 65 nm celeron/pentium/core2/xeon, "Merom"/"Conroe" */ ··· 634 634 } 635 635 } 636 636 637 - static void __cpuinit intel_detect_tlb(struct cpuinfo_x86 *c) 637 + static void intel_detect_tlb(struct cpuinfo_x86 *c) 638 638 { 639 639 int i, j, n; 640 640 unsigned int regs[4]; ··· 661 661 intel_tlb_flushall_shift_set(c); 662 662 } 663 663 664 - static const struct cpu_dev __cpuinitconst intel_cpu_dev = { 664 + static const struct cpu_dev intel_cpu_dev = { 665 665 .c_vendor = "Intel", 666 666 .c_ident = { "GenuineIntel" }, 667 667 #ifdef CONFIG_X86_32
+27 -28
arch/x86/kernel/cpu/intel_cacheinfo.c
··· 37 37 /* All the cache descriptor types we care about (no TLB or 38 38 trace cache entries) */ 39 39 40 - static const struct _cache_table __cpuinitconst cache_table[] = 40 + static const struct _cache_table cache_table[] = 41 41 { 42 42 { 0x06, LVL_1_INST, 8 }, /* 4-way set assoc, 32 byte line size */ 43 43 { 0x08, LVL_1_INST, 16 }, /* 4-way set assoc, 32 byte line size */ ··· 203 203 unsigned val; 204 204 }; 205 205 206 - static const unsigned short __cpuinitconst assocs[] = { 206 + static const unsigned short assocs[] = { 207 207 [1] = 1, 208 208 [2] = 2, 209 209 [4] = 4, ··· 217 217 [0xf] = 0xffff /* fully associative - no way to show this currently */ 218 218 }; 219 219 220 - static const unsigned char __cpuinitconst levels[] = { 1, 1, 2, 3 }; 221 - static const unsigned char __cpuinitconst types[] = { 1, 2, 3, 3 }; 220 + static const unsigned char levels[] = { 1, 1, 2, 3 }; 221 + static const unsigned char types[] = { 1, 2, 3, 3 }; 222 222 223 - static void __cpuinit 223 + static void 224 224 amd_cpuid4(int leaf, union _cpuid4_leaf_eax *eax, 225 225 union _cpuid4_leaf_ebx *ebx, 226 226 union _cpuid4_leaf_ecx *ecx) ··· 302 302 /* 303 303 * L3 cache descriptors 304 304 */ 305 - static void __cpuinit amd_calc_l3_indices(struct amd_northbridge *nb) 305 + static void amd_calc_l3_indices(struct amd_northbridge *nb) 306 306 { 307 307 struct amd_l3_cache *l3 = &nb->l3_cache; 308 308 unsigned int sc0, sc1, sc2, sc3; ··· 325 325 l3->indices = (max(max3(sc0, sc1, sc2), sc3) << 10) - 1; 326 326 } 327 327 328 - static void __cpuinit amd_init_l3_cache(struct _cpuid4_info_regs *this_leaf, int index) 328 + static void amd_init_l3_cache(struct _cpuid4_info_regs *this_leaf, int index) 329 329 { 330 330 int node; 331 331 ··· 528 528 #endif /* CONFIG_AMD_NB && CONFIG_SYSFS */ 529 529 530 530 static int 531 - __cpuinit cpuid4_cache_lookup_regs(int index, 532 - struct _cpuid4_info_regs *this_leaf) 531 + cpuid4_cache_lookup_regs(int index, struct _cpuid4_info_regs *this_leaf) 533 532 { 534 533 union _cpuid4_leaf_eax eax; 535 534 union _cpuid4_leaf_ebx ebx; ··· 559 560 return 0; 560 561 } 561 562 562 - static int __cpuinit find_num_cache_leaves(struct cpuinfo_x86 *c) 563 + static int find_num_cache_leaves(struct cpuinfo_x86 *c) 563 564 { 564 565 unsigned int eax, ebx, ecx, edx, op; 565 566 union _cpuid4_leaf_eax cache_eax; ··· 579 580 return i; 580 581 } 581 582 582 - void __cpuinit init_amd_cacheinfo(struct cpuinfo_x86 *c) 583 + void init_amd_cacheinfo(struct cpuinfo_x86 *c) 583 584 { 584 585 585 586 if (cpu_has_topoext) { ··· 592 593 } 593 594 } 594 595 595 - unsigned int __cpuinit init_intel_cacheinfo(struct cpuinfo_x86 *c) 596 + unsigned int init_intel_cacheinfo(struct cpuinfo_x86 *c) 596 597 { 597 598 /* Cache sizes */ 598 599 unsigned int trace = 0, l1i = 0, l1d = 0, l2 = 0, l3 = 0; ··· 743 744 744 745 #ifdef CONFIG_SMP 745 746 746 - static int __cpuinit cache_shared_amd_cpu_map_setup(unsigned int cpu, int index) 747 + static int cache_shared_amd_cpu_map_setup(unsigned int cpu, int index) 747 748 { 748 749 struct _cpuid4_info *this_leaf; 749 750 int i, sibling; ··· 792 793 return 1; 793 794 } 794 795 795 - static void __cpuinit cache_shared_cpu_map_setup(unsigned int cpu, int index) 796 + static void cache_shared_cpu_map_setup(unsigned int cpu, int index) 796 797 { 797 798 struct _cpuid4_info *this_leaf, *sibling_leaf; 798 799 unsigned long num_threads_sharing; ··· 827 828 } 828 829 } 829 830 } 830 - static void __cpuinit cache_remove_shared_cpu_map(unsigned int cpu, int index) 831 + static void cache_remove_shared_cpu_map(unsigned int cpu, int index) 831 832 { 832 833 struct _cpuid4_info *this_leaf, *sibling_leaf; 833 834 int sibling; ··· 840 841 } 841 842 } 842 843 #else 843 - static void __cpuinit cache_shared_cpu_map_setup(unsigned int cpu, int index) 844 + static void cache_shared_cpu_map_setup(unsigned int cpu, int index) 844 845 { 845 846 } 846 847 847 - static void __cpuinit cache_remove_shared_cpu_map(unsigned int cpu, int index) 848 + static void cache_remove_shared_cpu_map(unsigned int cpu, int index) 848 849 { 849 850 } 850 851 #endif 851 852 852 - static void __cpuinit free_cache_attributes(unsigned int cpu) 853 + static void free_cache_attributes(unsigned int cpu) 853 854 { 854 855 int i; 855 856 ··· 860 861 per_cpu(ici_cpuid4_info, cpu) = NULL; 861 862 } 862 863 863 - static void __cpuinit get_cpu_leaves(void *_retval) 864 + static void get_cpu_leaves(void *_retval) 864 865 { 865 866 int j, *retval = _retval, cpu = smp_processor_id(); 866 867 ··· 880 881 } 881 882 } 882 883 883 - static int __cpuinit detect_cache_attributes(unsigned int cpu) 884 + static int detect_cache_attributes(unsigned int cpu) 884 885 { 885 886 int retval; 886 887 ··· 1014 1015 }; 1015 1016 1016 1017 #ifdef CONFIG_AMD_NB 1017 - static struct attribute ** __cpuinit amd_l3_attrs(void) 1018 + static struct attribute **amd_l3_attrs(void) 1018 1019 { 1019 1020 static struct attribute **attrs; 1020 1021 int n; ··· 1090 1091 .sysfs_ops = &sysfs_ops, 1091 1092 }; 1092 1093 1093 - static void __cpuinit cpuid4_cache_sysfs_exit(unsigned int cpu) 1094 + static void cpuid4_cache_sysfs_exit(unsigned int cpu) 1094 1095 { 1095 1096 kfree(per_cpu(ici_cache_kobject, cpu)); 1096 1097 kfree(per_cpu(ici_index_kobject, cpu)); ··· 1099 1100 free_cache_attributes(cpu); 1100 1101 } 1101 1102 1102 - static int __cpuinit cpuid4_cache_sysfs_init(unsigned int cpu) 1103 + static int cpuid4_cache_sysfs_init(unsigned int cpu) 1103 1104 { 1104 1105 int err; 1105 1106 ··· 1131 1132 static DECLARE_BITMAP(cache_dev_map, NR_CPUS); 1132 1133 1133 1134 /* Add/Remove cache interface for CPU device */ 1134 - static int __cpuinit cache_add_dev(struct device *dev) 1135 + static int cache_add_dev(struct device *dev) 1135 1136 { 1136 1137 unsigned int cpu = dev->id; 1137 1138 unsigned long i, j; ··· 1182 1183 return 0; 1183 1184 } 1184 1185 1185 - static void __cpuinit cache_remove_dev(struct device *dev) 1186 + static void cache_remove_dev(struct device *dev) 1186 1187 { 1187 1188 unsigned int cpu = dev->id; 1188 1189 unsigned long i; ··· 1199 1200 cpuid4_cache_sysfs_exit(cpu); 1200 1201 } 1201 1202 1202 - static int __cpuinit cacheinfo_cpu_callback(struct notifier_block *nfb, 1203 - unsigned long action, void *hcpu) 1203 + static int cacheinfo_cpu_callback(struct notifier_block *nfb, 1204 + unsigned long action, void *hcpu) 1204 1205 { 1205 1206 unsigned int cpu = (unsigned long)hcpu; 1206 1207 struct device *dev; ··· 1219 1220 return NOTIFY_OK; 1220 1221 } 1221 1222 1222 - static struct notifier_block __cpuinitdata cacheinfo_cpu_notifier = { 1223 + static struct notifier_block cacheinfo_cpu_notifier = { 1223 1224 .notifier_call = cacheinfo_cpu_callback, 1224 1225 }; 1225 1226
+11 -12
arch/x86/kernel/cpu/mcheck/mce.c
··· 1363 1363 } 1364 1364 EXPORT_SYMBOL_GPL(mce_notify_irq); 1365 1365 1366 - static int __cpuinit __mcheck_cpu_mce_banks_init(void) 1366 + static int __mcheck_cpu_mce_banks_init(void) 1367 1367 { 1368 1368 int i; 1369 1369 u8 num_banks = mca_cfg.banks; ··· 1384 1384 /* 1385 1385 * Initialize Machine Checks for a CPU. 1386 1386 */ 1387 - static int __cpuinit __mcheck_cpu_cap_init(void) 1387 + static int __mcheck_cpu_cap_init(void) 1388 1388 { 1389 1389 unsigned b; 1390 1390 u64 cap; ··· 1483 1483 } 1484 1484 1485 1485 /* Add per CPU specific workarounds here */ 1486 - static int __cpuinit __mcheck_cpu_apply_quirks(struct cpuinfo_x86 *c) 1486 + static int __mcheck_cpu_apply_quirks(struct cpuinfo_x86 *c) 1487 1487 { 1488 1488 struct mca_config *cfg = &mca_cfg; 1489 1489 ··· 1593 1593 return 0; 1594 1594 } 1595 1595 1596 - static int __cpuinit __mcheck_cpu_ancient_init(struct cpuinfo_x86 *c) 1596 + static int __mcheck_cpu_ancient_init(struct cpuinfo_x86 *c) 1597 1597 { 1598 1598 if (c->x86 != 5) 1599 1599 return 0; ··· 1664 1664 * Called for each booted CPU to set up machine checks. 1665 1665 * Must be called with preempt off: 1666 1666 */ 1667 - void __cpuinit mcheck_cpu_init(struct cpuinfo_x86 *c) 1667 + void mcheck_cpu_init(struct cpuinfo_x86 *c) 1668 1668 { 1669 1669 if (mca_cfg.disabled) 1670 1670 return; ··· 2082 2082 2083 2083 DEFINE_PER_CPU(struct device *, mce_device); 2084 2084 2085 - __cpuinitdata 2086 2085 void (*threshold_cpu_callback)(unsigned long action, unsigned int cpu); 2087 2086 2088 2087 static inline struct mce_bank *attr_to_bank(struct device_attribute *attr) ··· 2227 2228 } 2228 2229 2229 2230 /* Per cpu device init. All of the cpus still share the same ctrl bank: */ 2230 - static __cpuinit int mce_device_create(unsigned int cpu) 2231 + static int mce_device_create(unsigned int cpu) 2231 2232 { 2232 2233 struct device *dev; 2233 2234 int err; ··· 2273 2274 return err; 2274 2275 } 2275 2276 2276 - static __cpuinit void mce_device_remove(unsigned int cpu) 2277 + static void mce_device_remove(unsigned int cpu) 2277 2278 { 2278 2279 struct device *dev = per_cpu(mce_device, cpu); 2279 2280 int i; ··· 2293 2294 } 2294 2295 2295 2296 /* Make sure there are no machine checks on offlined CPUs. */ 2296 - static void __cpuinit mce_disable_cpu(void *h) 2297 + static void mce_disable_cpu(void *h) 2297 2298 { 2298 2299 unsigned long action = *(unsigned long *)h; 2299 2300 int i; ··· 2311 2312 } 2312 2313 } 2313 2314 2314 - static void __cpuinit mce_reenable_cpu(void *h) 2315 + static void mce_reenable_cpu(void *h) 2315 2316 { 2316 2317 unsigned long action = *(unsigned long *)h; 2317 2318 int i; ··· 2330 2331 } 2331 2332 2332 2333 /* Get notified when a cpu comes on/off. Be hotplug friendly. */ 2333 - static int __cpuinit 2334 + static int 2334 2335 mce_cpu_callback(struct notifier_block *nfb, unsigned long action, void *hcpu) 2335 2336 { 2336 2337 unsigned int cpu = (unsigned long)hcpu; ··· 2366 2367 return NOTIFY_OK; 2367 2368 } 2368 2369 2369 - static struct notifier_block mce_cpu_notifier __cpuinitdata = { 2370 + static struct notifier_block mce_cpu_notifier = { 2370 2371 .notifier_call = mce_cpu_callback, 2371 2372 }; 2372 2373
+6 -8
arch/x86/kernel/cpu/mcheck/mce_amd.c
··· 458 458 .default_attrs = default_attrs, 459 459 }; 460 460 461 - static __cpuinit int allocate_threshold_blocks(unsigned int cpu, 462 - unsigned int bank, 463 - unsigned int block, 464 - u32 address) 461 + static int allocate_threshold_blocks(unsigned int cpu, unsigned int bank, 462 + unsigned int block, u32 address) 465 463 { 466 464 struct threshold_block *b = NULL; 467 465 u32 low, high; ··· 541 543 return err; 542 544 } 543 545 544 - static __cpuinit int __threshold_add_blocks(struct threshold_bank *b) 546 + static int __threshold_add_blocks(struct threshold_bank *b) 545 547 { 546 548 struct list_head *head = &b->blocks->miscj; 547 549 struct threshold_block *pos = NULL; ··· 565 567 return err; 566 568 } 567 569 568 - static __cpuinit int threshold_create_bank(unsigned int cpu, unsigned int bank) 570 + static int threshold_create_bank(unsigned int cpu, unsigned int bank) 569 571 { 570 572 struct device *dev = per_cpu(mce_device, cpu); 571 573 struct amd_northbridge *nb = NULL; ··· 630 632 } 631 633 632 634 /* create dir/files for all valid threshold banks */ 633 - static __cpuinit int threshold_create_device(unsigned int cpu) 635 + static int threshold_create_device(unsigned int cpu) 634 636 { 635 637 unsigned int bank; 636 638 struct threshold_bank **bp; ··· 734 736 } 735 737 736 738 /* get notified when a cpu comes on/off */ 737 - static void __cpuinit 739 + static void 738 740 amd_64_threshold_cpu_callback(unsigned long action, unsigned int cpu) 739 741 { 740 742 switch (action) {
+4 -5
arch/x86/kernel/cpu/mcheck/therm_throt.c
··· 240 240 241 241 #ifdef CONFIG_SYSFS 242 242 /* Add/Remove thermal_throttle interface for CPU device: */ 243 - static __cpuinit int thermal_throttle_add_dev(struct device *dev, 244 - unsigned int cpu) 243 + static int thermal_throttle_add_dev(struct device *dev, unsigned int cpu) 245 244 { 246 245 int err; 247 246 struct cpuinfo_x86 *c = &cpu_data(cpu); ··· 266 267 return err; 267 268 } 268 269 269 - static __cpuinit void thermal_throttle_remove_dev(struct device *dev) 270 + static void thermal_throttle_remove_dev(struct device *dev) 270 271 { 271 272 sysfs_remove_group(&dev->kobj, &thermal_attr_group); 272 273 } ··· 275 276 static DEFINE_MUTEX(therm_cpu_lock); 276 277 277 278 /* Get notified when a cpu comes on/off. Be hotplug friendly. */ 278 - static __cpuinit int 279 + static int 279 280 thermal_throttle_cpu_callback(struct notifier_block *nfb, 280 281 unsigned long action, 281 282 void *hcpu) ··· 306 307 return notifier_from_errno(err); 307 308 } 308 309 309 - static struct notifier_block thermal_throttle_cpu_notifier __cpuinitdata = 310 + static struct notifier_block thermal_throttle_cpu_notifier = 310 311 { 311 312 .notifier_call = thermal_throttle_cpu_callback, 312 313 };
+1 -1
arch/x86/kernel/cpu/perf_event.c
··· 1295 1295 struct event_constraint emptyconstraint; 1296 1296 struct event_constraint unconstrained; 1297 1297 1298 - static int __cpuinit 1298 + static int 1299 1299 x86_pmu_notifier(struct notifier_block *self, unsigned long action, void *hcpu) 1300 1300 { 1301 1301 unsigned int cpu = (long)hcpu;
+1 -1
arch/x86/kernel/cpu/perf_event_amd_ibs.c
··· 851 851 setup_APIC_eilvt(offset, 0, APIC_EILVT_MSG_FIX, 1); 852 852 } 853 853 854 - static int __cpuinit 854 + static int 855 855 perf_ibs_cpu_notifier(struct notifier_block *self, unsigned long action, void *hcpu) 856 856 { 857 857 switch (action & ~CPU_TASKS_FROZEN) {
+15 -16
arch/x86/kernel/cpu/perf_event_amd_uncore.c
··· 288 288 .read = amd_uncore_read, 289 289 }; 290 290 291 - static struct amd_uncore * __cpuinit amd_uncore_alloc(unsigned int cpu) 291 + static struct amd_uncore *amd_uncore_alloc(unsigned int cpu) 292 292 { 293 293 return kzalloc_node(sizeof(struct amd_uncore), GFP_KERNEL, 294 294 cpu_to_node(cpu)); 295 295 } 296 296 297 - static void __cpuinit amd_uncore_cpu_up_prepare(unsigned int cpu) 297 + static void amd_uncore_cpu_up_prepare(unsigned int cpu) 298 298 { 299 299 struct amd_uncore *uncore; 300 300 ··· 322 322 } 323 323 324 324 static struct amd_uncore * 325 - __cpuinit amd_uncore_find_online_sibling(struct amd_uncore *this, 326 - struct amd_uncore * __percpu *uncores) 325 + amd_uncore_find_online_sibling(struct amd_uncore *this, 326 + struct amd_uncore * __percpu *uncores) 327 327 { 328 328 unsigned int cpu; 329 329 struct amd_uncore *that; ··· 348 348 return this; 349 349 } 350 350 351 - static void __cpuinit amd_uncore_cpu_starting(unsigned int cpu) 351 + static void amd_uncore_cpu_starting(unsigned int cpu) 352 352 { 353 353 unsigned int eax, ebx, ecx, edx; 354 354 struct amd_uncore *uncore; ··· 376 376 } 377 377 } 378 378 379 - static void __cpuinit uncore_online(unsigned int cpu, 380 - struct amd_uncore * __percpu *uncores) 379 + static void uncore_online(unsigned int cpu, 380 + struct amd_uncore * __percpu *uncores) 381 381 { 382 382 struct amd_uncore *uncore = *per_cpu_ptr(uncores, cpu); 383 383 ··· 388 388 cpumask_set_cpu(cpu, uncore->active_mask); 389 389 } 390 390 391 - static void __cpuinit amd_uncore_cpu_online(unsigned int cpu) 391 + static void amd_uncore_cpu_online(unsigned int cpu) 392 392 { 393 393 if (amd_uncore_nb) 394 394 uncore_online(cpu, amd_uncore_nb); ··· 397 397 uncore_online(cpu, amd_uncore_l2); 398 398 } 399 399 400 - static void __cpuinit uncore_down_prepare(unsigned int cpu, 401 - struct amd_uncore * __percpu *uncores) 400 + static void uncore_down_prepare(unsigned int cpu, 401 + struct amd_uncore * __percpu *uncores) 402 402 { 403 403 unsigned int i; 404 404 struct amd_uncore *this = *per_cpu_ptr(uncores, cpu); ··· 423 423 } 424 424 } 425 425 426 - static void __cpuinit amd_uncore_cpu_down_prepare(unsigned int cpu) 426 + static void amd_uncore_cpu_down_prepare(unsigned int cpu) 427 427 { 428 428 if (amd_uncore_nb) 429 429 uncore_down_prepare(cpu, amd_uncore_nb); ··· 432 432 uncore_down_prepare(cpu, amd_uncore_l2); 433 433 } 434 434 435 - static void __cpuinit uncore_dead(unsigned int cpu, 436 - struct amd_uncore * __percpu *uncores) 435 + static void uncore_dead(unsigned int cpu, struct amd_uncore * __percpu *uncores) 437 436 { 438 437 struct amd_uncore *uncore = *per_cpu_ptr(uncores, cpu); 439 438 ··· 444 445 *per_cpu_ptr(amd_uncore_nb, cpu) = NULL; 445 446 } 446 447 447 - static void __cpuinit amd_uncore_cpu_dead(unsigned int cpu) 448 + static void amd_uncore_cpu_dead(unsigned int cpu) 448 449 { 449 450 if (amd_uncore_nb) 450 451 uncore_dead(cpu, amd_uncore_nb); ··· 453 454 uncore_dead(cpu, amd_uncore_l2); 454 455 } 455 456 456 - static int __cpuinit 457 + static int 457 458 amd_uncore_cpu_notifier(struct notifier_block *self, unsigned long action, 458 459 void *hcpu) 459 460 { ··· 488 489 return NOTIFY_OK; 489 490 } 490 491 491 - static struct notifier_block amd_uncore_cpu_notifier_block __cpuinitdata = { 492 + static struct notifier_block amd_uncore_cpu_notifier_block = { 492 493 .notifier_call = amd_uncore_cpu_notifier, 493 494 .priority = CPU_PRI_PERF + 1, 494 495 };
+10 -10
arch/x86/kernel/cpu/perf_event_intel_uncore.c
··· 3297 3297 /* CPU hot plug/unplug are serialized by cpu_add_remove_lock mutex */ 3298 3298 static LIST_HEAD(boxes_to_free); 3299 3299 3300 - static void __cpuinit uncore_kfree_boxes(void) 3300 + static void uncore_kfree_boxes(void) 3301 3301 { 3302 3302 struct intel_uncore_box *box; 3303 3303 ··· 3309 3309 } 3310 3310 } 3311 3311 3312 - static void __cpuinit uncore_cpu_dying(int cpu) 3312 + static void uncore_cpu_dying(int cpu) 3313 3313 { 3314 3314 struct intel_uncore_type *type; 3315 3315 struct intel_uncore_pmu *pmu; ··· 3328 3328 } 3329 3329 } 3330 3330 3331 - static int __cpuinit uncore_cpu_starting(int cpu) 3331 + static int uncore_cpu_starting(int cpu) 3332 3332 { 3333 3333 struct intel_uncore_type *type; 3334 3334 struct intel_uncore_pmu *pmu; ··· 3371 3371 return 0; 3372 3372 } 3373 3373 3374 - static int __cpuinit uncore_cpu_prepare(int cpu, int phys_id) 3374 + static int uncore_cpu_prepare(int cpu, int phys_id) 3375 3375 { 3376 3376 struct intel_uncore_type *type; 3377 3377 struct intel_uncore_pmu *pmu; ··· 3397 3397 return 0; 3398 3398 } 3399 3399 3400 - static void __cpuinit 3400 + static void 3401 3401 uncore_change_context(struct intel_uncore_type **uncores, int old_cpu, int new_cpu) 3402 3402 { 3403 3403 struct intel_uncore_type *type; ··· 3435 3435 } 3436 3436 } 3437 3437 3438 - static void __cpuinit uncore_event_exit_cpu(int cpu) 3438 + static void uncore_event_exit_cpu(int cpu) 3439 3439 { 3440 3440 int i, phys_id, target; 3441 3441 ··· 3463 3463 uncore_change_context(pci_uncores, cpu, target); 3464 3464 } 3465 3465 3466 - static void __cpuinit uncore_event_init_cpu(int cpu) 3466 + static void uncore_event_init_cpu(int cpu) 3467 3467 { 3468 3468 int i, phys_id; 3469 3469 ··· 3479 3479 uncore_change_context(pci_uncores, -1, cpu); 3480 3480 } 3481 3481 3482 - static int 3483 - __cpuinit uncore_cpu_notifier(struct notifier_block *self, unsigned long action, void *hcpu) 3482 + static int uncore_cpu_notifier(struct notifier_block *self, 3483 + unsigned long action, void *hcpu) 3484 3484 { 3485 3485 unsigned int cpu = (long)hcpu; 3486 3486 ··· 3520 3520 return NOTIFY_OK; 3521 3521 } 3522 3522 3523 - static struct notifier_block uncore_cpu_nb __cpuinitdata = { 3523 + static struct notifier_block uncore_cpu_nb = { 3524 3524 .notifier_call = uncore_cpu_notifier, 3525 3525 /* 3526 3526 * to migrate uncore events, our notifier should be executed
+1 -1
arch/x86/kernel/cpu/rdrand.c
··· 52 52 */ 53 53 #define RESEED_LOOP ((512*128)/sizeof(unsigned long)) 54 54 55 - void __cpuinit x86_init_rdrand(struct cpuinfo_x86 *c) 55 + void x86_init_rdrand(struct cpuinfo_x86 *c) 56 56 { 57 57 #ifdef CONFIG_ARCH_RANDOM 58 58 unsigned long tmp;
+2 -2
arch/x86/kernel/cpu/scattered.c
··· 24 24 CR_EBX 25 25 }; 26 26 27 - void __cpuinit init_scattered_cpuid_features(struct cpuinfo_x86 *c) 27 + void init_scattered_cpuid_features(struct cpuinfo_x86 *c) 28 28 { 29 29 u32 max_level; 30 30 u32 regs[4]; 31 31 const struct cpuid_bit *cb; 32 32 33 - static const struct cpuid_bit __cpuinitconst cpuid_bits[] = { 33 + static const struct cpuid_bit cpuid_bits[] = { 34 34 { X86_FEATURE_DTHERM, CR_EAX, 0, 0x00000006, 0 }, 35 35 { X86_FEATURE_IDA, CR_EAX, 1, 0x00000006, 0 }, 36 36 { X86_FEATURE_ARAT, CR_EAX, 2, 0x00000006, 0 },
+1 -1
arch/x86/kernel/cpu/topology.c
··· 26 26 * exists, use it for populating initial_apicid and cpu topology 27 27 * detection. 28 28 */ 29 - void __cpuinit detect_extended_topology(struct cpuinfo_x86 *c) 29 + void detect_extended_topology(struct cpuinfo_x86 *c) 30 30 { 31 31 #ifdef CONFIG_SMP 32 32 unsigned int eax, ebx, ecx, edx, sub_index;
+3 -3
arch/x86/kernel/cpu/transmeta.c
··· 5 5 #include <asm/msr.h> 6 6 #include "cpu.h" 7 7 8 - static void __cpuinit early_init_transmeta(struct cpuinfo_x86 *c) 8 + static void early_init_transmeta(struct cpuinfo_x86 *c) 9 9 { 10 10 u32 xlvl; 11 11 ··· 17 17 } 18 18 } 19 19 20 - static void __cpuinit init_transmeta(struct cpuinfo_x86 *c) 20 + static void init_transmeta(struct cpuinfo_x86 *c) 21 21 { 22 22 unsigned int cap_mask, uk, max, dummy; 23 23 unsigned int cms_rev1, cms_rev2; ··· 98 98 #endif 99 99 } 100 100 101 - static const struct cpu_dev __cpuinitconst transmeta_cpu_dev = { 101 + static const struct cpu_dev transmeta_cpu_dev = { 102 102 .c_vendor = "Transmeta", 103 103 .c_ident = { "GenuineTMx86", "TransmetaCPU" }, 104 104 .c_early_init = early_init_transmeta,
+1 -1
arch/x86/kernel/cpu/umc.c
··· 8 8 * so no special init takes place. 9 9 */ 10 10 11 - static const struct cpu_dev __cpuinitconst umc_cpu_dev = { 11 + static const struct cpu_dev umc_cpu_dev = { 12 12 .c_vendor = "UMC", 13 13 .c_ident = { "UMC UMC UMC" }, 14 14 .c_models = {
+1 -1
arch/x86/kernel/cpu/vmware.c
··· 122 122 * so that the kernel could just trust the hypervisor with providing a 123 123 * reliable virtual TSC that is suitable for timekeeping. 124 124 */ 125 - static void __cpuinit vmware_set_cpu_features(struct cpuinfo_x86 *c) 125 + static void vmware_set_cpu_features(struct cpuinfo_x86 *c) 126 126 { 127 127 set_cpu_cap(c, X86_FEATURE_CONSTANT_TSC); 128 128 set_cpu_cap(c, X86_FEATURE_TSC_RELIABLE);
+3 -4
arch/x86/kernel/cpuid.c
··· 137 137 .open = cpuid_open, 138 138 }; 139 139 140 - static __cpuinit int cpuid_device_create(int cpu) 140 + static int cpuid_device_create(int cpu) 141 141 { 142 142 struct device *dev; 143 143 ··· 151 151 device_destroy(cpuid_class, MKDEV(CPUID_MAJOR, cpu)); 152 152 } 153 153 154 - static int __cpuinit cpuid_class_cpu_callback(struct notifier_block *nfb, 155 - unsigned long action, 156 - void *hcpu) 154 + static int cpuid_class_cpu_callback(struct notifier_block *nfb, 155 + unsigned long action, void *hcpu) 157 156 { 158 157 unsigned int cpu = (unsigned long)hcpu; 159 158 int err = 0;
+1 -1
arch/x86/kernel/devicetree.c
··· 133 133 { 134 134 } 135 135 136 - void __cpuinit x86_of_pci_init(void) 136 + void x86_of_pci_init(void) 137 137 { 138 138 pcibios_enable_irq = x86_of_pci_irq_enable; 139 139 pcibios_disable_irq = x86_of_pci_irq_disable;
-1
arch/x86/kernel/head_32.S
··· 292 292 * If cpu hotplug is not supported then this code can go in init section 293 293 * which will be freed later 294 294 */ 295 - __CPUINIT 296 295 ENTRY(startup_32_smp) 297 296 cld 298 297 movl $(__BOOT_DS),%eax
+5 -5
arch/x86/kernel/i387.c
··· 108 108 unsigned int mxcsr_feature_mask __read_mostly = 0xffffffffu; 109 109 unsigned int xstate_size; 110 110 EXPORT_SYMBOL_GPL(xstate_size); 111 - static struct i387_fxsave_struct fx_scratch __cpuinitdata; 111 + static struct i387_fxsave_struct fx_scratch; 112 112 113 - static void __cpuinit mxcsr_feature_mask_init(void) 113 + static void mxcsr_feature_mask_init(void) 114 114 { 115 115 unsigned long mask = 0; 116 116 ··· 124 124 mxcsr_feature_mask &= mask; 125 125 } 126 126 127 - static void __cpuinit init_thread_xstate(void) 127 + static void init_thread_xstate(void) 128 128 { 129 129 /* 130 130 * Note that xstate_size might be overwriten later during ··· 153 153 * into all processes. 154 154 */ 155 155 156 - void __cpuinit fpu_init(void) 156 + void fpu_init(void) 157 157 { 158 158 unsigned long cr0; 159 159 unsigned long cr4_mask = 0; ··· 608 608 609 609 __setup("no387", no_387); 610 610 611 - void __cpuinit fpu_detect(struct cpuinfo_x86 *c) 611 + void fpu_detect(struct cpuinfo_x86 *c) 612 612 { 613 613 unsigned long cr0; 614 614 u16 fsw, fcw;
+1 -1
arch/x86/kernel/irq_32.c
··· 119 119 /* 120 120 * allocate per-cpu stacks for hardirq and for softirq processing 121 121 */ 122 - void __cpuinit irq_ctx_init(int cpu) 122 + void irq_ctx_init(int cpu) 123 123 { 124 124 union irq_ctx *irqctx; 125 125
+5 -5
arch/x86/kernel/kvm.c
··· 320 320 apic_write(APIC_EOI, APIC_EOI_ACK); 321 321 } 322 322 323 - void __cpuinit kvm_guest_cpu_init(void) 323 + void kvm_guest_cpu_init(void) 324 324 { 325 325 if (!kvm_para_available()) 326 326 return; ··· 421 421 native_smp_prepare_boot_cpu(); 422 422 } 423 423 424 - static void __cpuinit kvm_guest_cpu_online(void *dummy) 424 + static void kvm_guest_cpu_online(void *dummy) 425 425 { 426 426 kvm_guest_cpu_init(); 427 427 } ··· 435 435 apf_task_wake_all(); 436 436 } 437 437 438 - static int __cpuinit kvm_cpu_notify(struct notifier_block *self, 439 - unsigned long action, void *hcpu) 438 + static int kvm_cpu_notify(struct notifier_block *self, unsigned long action, 439 + void *hcpu) 440 440 { 441 441 int cpu = (unsigned long)hcpu; 442 442 switch (action) { ··· 455 455 return NOTIFY_OK; 456 456 } 457 457 458 - static struct notifier_block __cpuinitdata kvm_cpu_notifier = { 458 + static struct notifier_block kvm_cpu_notifier = { 459 459 .notifier_call = kvm_cpu_notify, 460 460 }; 461 461 #endif
+1 -1
arch/x86/kernel/kvmclock.c
··· 182 182 } 183 183 184 184 #ifdef CONFIG_X86_LOCAL_APIC 185 - static void __cpuinit kvm_setup_secondary_clock(void) 185 + static void kvm_setup_secondary_clock(void) 186 186 { 187 187 /* 188 188 * Now that the first cpu already had this clocksource initialized,
+4 -4
arch/x86/kernel/microcode_amd_early.c
··· 82 82 * load_microcode_amd() to save equivalent cpu table and microcode patches in 83 83 * kernel heap memory. 84 84 */ 85 - static void __cpuinit apply_ucode_in_initrd(void *ucode, size_t size) 85 + static void apply_ucode_in_initrd(void *ucode, size_t size) 86 86 { 87 87 struct equiv_cpu_entry *eq; 88 88 u32 *header; ··· 206 206 * save_microcode_in_initrd_amd() BSP's patch is copied to amd_bsp_mpb, which 207 207 * is used upon resume from suspend. 208 208 */ 209 - void __cpuinit load_ucode_amd_ap(void) 209 + void load_ucode_amd_ap(void) 210 210 { 211 211 struct microcode_amd *mc; 212 212 unsigned long *initrd; ··· 238 238 uci->cpu_sig.sig = cpuid_eax(0x00000001); 239 239 } 240 240 #else 241 - static void __cpuinit collect_cpu_info_amd_early(struct cpuinfo_x86 *c, 241 + static void collect_cpu_info_amd_early(struct cpuinfo_x86 *c, 242 242 struct ucode_cpu_info *uci) 243 243 { 244 244 u32 rev, eax; ··· 252 252 c->x86 = ((eax >> 8) & 0xf) + ((eax >> 20) & 0xff); 253 253 } 254 254 255 - void __cpuinit load_ucode_amd_ap(void) 255 + void load_ucode_amd_ap(void) 256 256 { 257 257 unsigned int cpu = smp_processor_id(); 258 258
+1 -1
arch/x86/kernel/microcode_core.c
··· 468 468 .resume = mc_bp_resume, 469 469 }; 470 470 471 - static __cpuinit int 471 + static int 472 472 mc_cpu_callback(struct notifier_block *nb, unsigned long action, void *hcpu) 473 473 { 474 474 unsigned int cpu = (unsigned long)hcpu;
+3 -3
arch/x86/kernel/microcode_core_early.c
··· 41 41 * 42 42 * x86_vendor() gets vendor information directly through cpuid. 43 43 */ 44 - static int __cpuinit x86_vendor(void) 44 + static int x86_vendor(void) 45 45 { 46 46 u32 eax = 0x00000000; 47 47 u32 ebx, ecx = 0, edx; ··· 57 57 return X86_VENDOR_UNKNOWN; 58 58 } 59 59 60 - static int __cpuinit x86_family(void) 60 + static int x86_family(void) 61 61 { 62 62 u32 eax = 0x00000001; 63 63 u32 ebx, ecx = 0, edx; ··· 96 96 } 97 97 } 98 98 99 - void __cpuinit load_ucode_ap(void) 99 + void load_ucode_ap(void) 100 100 { 101 101 int vendor, x86; 102 102
+13 -13
arch/x86/kernel/microcode_intel_early.c
··· 34 34 struct microcode_intel **mc_saved; 35 35 } mc_saved_data; 36 36 37 - static enum ucode_state __cpuinit 37 + static enum ucode_state 38 38 generic_load_microcode_early(struct microcode_intel **mc_saved_p, 39 39 unsigned int mc_saved_count, 40 40 struct ucode_cpu_info *uci) ··· 69 69 return state; 70 70 } 71 71 72 - static void __cpuinit 72 + static void 73 73 microcode_pointer(struct microcode_intel **mc_saved, 74 74 unsigned long *mc_saved_in_initrd, 75 75 unsigned long initrd_start, int mc_saved_count) ··· 82 82 } 83 83 84 84 #ifdef CONFIG_X86_32 85 - static void __cpuinit 85 + static void 86 86 microcode_phys(struct microcode_intel **mc_saved_tmp, 87 87 struct mc_saved_data *mc_saved_data) 88 88 { ··· 101 101 } 102 102 #endif 103 103 104 - static enum ucode_state __cpuinit 104 + static enum ucode_state 105 105 load_microcode(struct mc_saved_data *mc_saved_data, 106 106 unsigned long *mc_saved_in_initrd, 107 107 unsigned long initrd_start, ··· 375 375 #define native_wrmsr(msr, low, high) \ 376 376 native_write_msr(msr, low, high); 377 377 378 - static int __cpuinit collect_cpu_info_early(struct ucode_cpu_info *uci) 378 + static int collect_cpu_info_early(struct ucode_cpu_info *uci) 379 379 { 380 380 unsigned int val[2]; 381 381 u8 x86, x86_model; ··· 584 584 /* 585 585 * Print ucode update info. 586 586 */ 587 - static void __cpuinit 587 + static void 588 588 print_ucode_info(struct ucode_cpu_info *uci, unsigned int date) 589 589 { 590 590 int cpu = smp_processor_id(); ··· 605 605 /* 606 606 * Print early updated ucode info after printk works. This is delayed info dump. 607 607 */ 608 - void __cpuinit show_ucode_info_early(void) 608 + void show_ucode_info_early(void) 609 609 { 610 610 struct ucode_cpu_info uci; 611 611 ··· 621 621 * mc_saved_data.mc_saved and delay printing microcode info in 622 622 * show_ucode_info_early() until printk() works. 623 623 */ 624 - static void __cpuinit print_ucode(struct ucode_cpu_info *uci) 624 + static void print_ucode(struct ucode_cpu_info *uci) 625 625 { 626 626 struct microcode_intel *mc_intel; 627 627 int *delay_ucode_info_p; ··· 643 643 * Flush global tlb. We only do this in x86_64 where paging has been enabled 644 644 * already and PGE should be enabled as well. 645 645 */ 646 - static inline void __cpuinit flush_tlb_early(void) 646 + static inline void flush_tlb_early(void) 647 647 { 648 648 __native_flush_tlb_global_irq_disabled(); 649 649 } 650 650 651 - static inline void __cpuinit print_ucode(struct ucode_cpu_info *uci) 651 + static inline void print_ucode(struct ucode_cpu_info *uci) 652 652 { 653 653 struct microcode_intel *mc_intel; 654 654 ··· 660 660 } 661 661 #endif 662 662 663 - static int __cpuinit apply_microcode_early(struct mc_saved_data *mc_saved_data, 664 - struct ucode_cpu_info *uci) 663 + static int apply_microcode_early(struct mc_saved_data *mc_saved_data, 664 + struct ucode_cpu_info *uci) 665 665 { 666 666 struct microcode_intel *mc_intel; 667 667 unsigned int val[2]; ··· 763 763 #endif 764 764 } 765 765 766 - void __cpuinit load_ucode_intel_ap(void) 766 + void load_ucode_intel_ap(void) 767 767 { 768 768 struct mc_saved_data *mc_saved_data_p; 769 769 struct ucode_cpu_info uci;
+6 -6
arch/x86/kernel/mmconf-fam10h_64.c
··· 24 24 u32 device; 25 25 }; 26 26 27 - static u64 __cpuinitdata fam10h_pci_mmconf_base; 27 + static u64 fam10h_pci_mmconf_base; 28 28 29 - static struct pci_hostbridge_probe pci_probes[] __cpuinitdata = { 29 + static struct pci_hostbridge_probe pci_probes[] = { 30 30 { 0, 0x18, PCI_VENDOR_ID_AMD, 0x1200 }, 31 31 { 0xff, 0, PCI_VENDOR_ID_AMD, 0x1200 }, 32 32 }; 33 33 34 - static int __cpuinit cmp_range(const void *x1, const void *x2) 34 + static int cmp_range(const void *x1, const void *x2) 35 35 { 36 36 const struct range *r1 = x1; 37 37 const struct range *r2 = x2; ··· 49 49 /* need to avoid (0xfd<<32), (0xfe<<32), and (0xff<<32), ht used space */ 50 50 #define FAM10H_PCI_MMCONF_BASE (0xfcULL<<32) 51 51 #define BASE_VALID(b) ((b) + MMCONF_SIZE <= (0xfdULL<<32) || (b) >= (1ULL<<40)) 52 - static void __cpuinit get_fam10h_pci_mmconf_base(void) 52 + static void get_fam10h_pci_mmconf_base(void) 53 53 { 54 54 int i; 55 55 unsigned bus; ··· 166 166 fam10h_pci_mmconf_base = base; 167 167 } 168 168 169 - void __cpuinit fam10h_check_enable_mmcfg(void) 169 + void fam10h_check_enable_mmcfg(void) 170 170 { 171 171 u64 val; 172 172 u32 address; ··· 230 230 {} 231 231 }; 232 232 233 - /* Called from a __cpuinit function, but only on the BSP. */ 233 + /* Called from a non __init function, but only on the BSP. */ 234 234 void __ref check_enable_amd_mmconf_dmi(void) 235 235 { 236 236 dmi_check_system(mmconf_dmi_table);
+3 -3
arch/x86/kernel/msr.c
··· 200 200 .compat_ioctl = msr_ioctl, 201 201 }; 202 202 203 - static int __cpuinit msr_device_create(int cpu) 203 + static int msr_device_create(int cpu) 204 204 { 205 205 struct device *dev; 206 206 ··· 214 214 device_destroy(msr_class, MKDEV(MSR_MAJOR, cpu)); 215 215 } 216 216 217 - static int __cpuinit msr_class_cpu_callback(struct notifier_block *nfb, 218 - unsigned long action, void *hcpu) 217 + static int msr_class_cpu_callback(struct notifier_block *nfb, 218 + unsigned long action, void *hcpu) 219 219 { 220 220 unsigned int cpu = (unsigned long)hcpu; 221 221 int err = 0;
+1 -1
arch/x86/kernel/process.c
··· 398 398 default_idle(); 399 399 } 400 400 401 - void __cpuinit select_idle_routine(const struct cpuinfo_x86 *c) 401 + void select_idle_routine(const struct cpuinfo_x86 *c) 402 402 { 403 403 #ifdef CONFIG_SMP 404 404 if (boot_option_idle_override == IDLE_POLL && smp_num_siblings > 1)
+1 -1
arch/x86/kernel/setup.c
··· 170 170 171 171 #ifdef CONFIG_X86_32 172 172 /* cpu data as detected by the assembly code in head.S */ 173 - struct cpuinfo_x86 new_cpu_data __cpuinitdata = { 173 + struct cpuinfo_x86 new_cpu_data = { 174 174 .wp_works_ok = -1, 175 175 }; 176 176 /* common cpu data for all cpus */
+14 -14
arch/x86/kernel/smpboot.c
··· 130 130 * Report back to the Boot Processor during boot time or to the caller processor 131 131 * during CPU online. 132 132 */ 133 - static void __cpuinit smp_callin(void) 133 + static void smp_callin(void) 134 134 { 135 135 int cpuid, phys_id; 136 136 unsigned long timeout; ··· 237 237 /* 238 238 * Activate a secondary processor. 239 239 */ 240 - notrace static void __cpuinit start_secondary(void *unused) 240 + static void notrace start_secondary(void *unused) 241 241 { 242 242 /* 243 243 * Don't put *anything* before cpu_init(), SMP booting is too ··· 300 300 * The bootstrap kernel entry code has set these up. Save them for 301 301 * a given CPU 302 302 */ 303 - void __cpuinit smp_store_cpu_info(int id) 303 + void smp_store_cpu_info(int id) 304 304 { 305 305 struct cpuinfo_x86 *c = &cpu_data(id); 306 306 ··· 313 313 identify_secondary_cpu(c); 314 314 } 315 315 316 - static bool __cpuinit 316 + static bool 317 317 topology_sane(struct cpuinfo_x86 *c, struct cpuinfo_x86 *o, const char *name) 318 318 { 319 319 int cpu1 = c->cpu_index, cpu2 = o->cpu_index; ··· 330 330 cpumask_set_cpu((c2), cpu_##_m##_mask(c1)); \ 331 331 } while (0) 332 332 333 - static bool __cpuinit match_smt(struct cpuinfo_x86 *c, struct cpuinfo_x86 *o) 333 + static bool match_smt(struct cpuinfo_x86 *c, struct cpuinfo_x86 *o) 334 334 { 335 335 if (cpu_has_topoext) { 336 336 int cpu1 = c->cpu_index, cpu2 = o->cpu_index; ··· 348 348 return false; 349 349 } 350 350 351 - static bool __cpuinit match_llc(struct cpuinfo_x86 *c, struct cpuinfo_x86 *o) 351 + static bool match_llc(struct cpuinfo_x86 *c, struct cpuinfo_x86 *o) 352 352 { 353 353 int cpu1 = c->cpu_index, cpu2 = o->cpu_index; 354 354 ··· 359 359 return false; 360 360 } 361 361 362 - static bool __cpuinit match_mc(struct cpuinfo_x86 *c, struct cpuinfo_x86 *o) 362 + static bool match_mc(struct cpuinfo_x86 *c, struct cpuinfo_x86 *o) 363 363 { 364 364 if (c->phys_proc_id == o->phys_proc_id) { 365 365 if (cpu_has(c, X86_FEATURE_AMD_DCM)) ··· 370 370 return false; 371 371 } 372 372 373 - void __cpuinit set_cpu_sibling_map(int cpu) 373 + void set_cpu_sibling_map(int cpu) 374 374 { 375 375 bool has_smt = smp_num_siblings > 1; 376 376 bool has_mp = has_smt || boot_cpu_data.x86_max_cores > 1; ··· 499 499 * INIT, INIT, STARTUP sequence will reset the chip hard for us, and this 500 500 * won't ... remember to clear down the APIC, etc later. 501 501 */ 502 - int __cpuinit 502 + int 503 503 wakeup_secondary_cpu_via_nmi(int apicid, unsigned long start_eip) 504 504 { 505 505 unsigned long send_status, accept_status = 0; ··· 533 533 return (send_status | accept_status); 534 534 } 535 535 536 - static int __cpuinit 536 + static int 537 537 wakeup_secondary_cpu_via_init(int phys_apicid, unsigned long start_eip) 538 538 { 539 539 unsigned long send_status, accept_status = 0; ··· 649 649 } 650 650 651 651 /* reduce the number of lines printed when booting a large cpu count system */ 652 - static void __cpuinit announce_cpu(int cpu, int apicid) 652 + static void announce_cpu(int cpu, int apicid) 653 653 { 654 654 static int current_node = -1; 655 655 int node = early_cpu_to_node(cpu); ··· 691 691 * We'll change this code in the future to wake up hard offlined CPU0 if 692 692 * real platform and request are available. 693 693 */ 694 - static int __cpuinit 694 + static int 695 695 wakeup_cpu_via_init_nmi(int cpu, unsigned long start_ip, int apicid, 696 696 int *cpu0_nmi_registered) 697 697 { ··· 731 731 * Returns zero if CPU booted OK, else error code from 732 732 * ->wakeup_secondary_cpu. 733 733 */ 734 - static int __cpuinit do_boot_cpu(int apicid, int cpu, struct task_struct *idle) 734 + static int do_boot_cpu(int apicid, int cpu, struct task_struct *idle) 735 735 { 736 736 volatile u32 *trampoline_status = 737 737 (volatile u32 *) __va(real_mode_header->trampoline_status); ··· 872 872 return boot_error; 873 873 } 874 874 875 - int __cpuinit native_cpu_up(unsigned int cpu, struct task_struct *tidle) 875 + int native_cpu_up(unsigned int cpu, struct task_struct *tidle) 876 876 { 877 877 int apicid = apic->cpu_present_to_apicid(cpu); 878 878 unsigned long flags;
+3 -3
arch/x86/kernel/tboot.c
··· 320 320 return !(atomic_read((atomic_t *)&tboot->num_in_wfs) == num_aps); 321 321 } 322 322 323 - static int __cpuinit tboot_cpu_callback(struct notifier_block *nfb, 324 - unsigned long action, void *hcpu) 323 + static int tboot_cpu_callback(struct notifier_block *nfb, unsigned long action, 324 + void *hcpu) 325 325 { 326 326 switch (action) { 327 327 case CPU_DYING: ··· 334 334 return NOTIFY_OK; 335 335 } 336 336 337 - static struct notifier_block tboot_cpu_notifier __cpuinitdata = 337 + static struct notifier_block tboot_cpu_notifier = 338 338 { 339 339 .notifier_call = tboot_cpu_callback, 340 340 };
+2 -2
arch/x86/kernel/tsc.c
··· 824 824 * Make an educated guess if the TSC is trustworthy and synchronized 825 825 * over all CPUs. 826 826 */ 827 - __cpuinit int unsynchronized_tsc(void) 827 + int unsynchronized_tsc(void) 828 828 { 829 829 if (!cpu_has_tsc || tsc_unstable) 830 830 return 1; ··· 1020 1020 * been calibrated. This assumes that CONSTANT_TSC applies to all 1021 1021 * cpus in the socket - this should be a safe assumption. 1022 1022 */ 1023 - unsigned long __cpuinit calibrate_delay_is_known(void) 1023 + unsigned long calibrate_delay_is_known(void) 1024 1024 { 1025 1025 int i, cpu = smp_processor_id(); 1026 1026
+9 -9
arch/x86/kernel/tsc_sync.c
··· 25 25 * Entry/exit counters that make sure that both CPUs 26 26 * run the measurement code at once: 27 27 */ 28 - static __cpuinitdata atomic_t start_count; 29 - static __cpuinitdata atomic_t stop_count; 28 + static atomic_t start_count; 29 + static atomic_t stop_count; 30 30 31 31 /* 32 32 * We use a raw spinlock in this exceptional case, because 33 33 * we want to have the fastest, inlined, non-debug version 34 34 * of a critical section, to be able to prove TSC time-warps: 35 35 */ 36 - static __cpuinitdata arch_spinlock_t sync_lock = __ARCH_SPIN_LOCK_UNLOCKED; 36 + static arch_spinlock_t sync_lock = __ARCH_SPIN_LOCK_UNLOCKED; 37 37 38 - static __cpuinitdata cycles_t last_tsc; 39 - static __cpuinitdata cycles_t max_warp; 40 - static __cpuinitdata int nr_warps; 38 + static cycles_t last_tsc; 39 + static cycles_t max_warp; 40 + static int nr_warps; 41 41 42 42 /* 43 43 * TSC-warp measurement loop running on both CPUs: 44 44 */ 45 - static __cpuinit void check_tsc_warp(unsigned int timeout) 45 + static void check_tsc_warp(unsigned int timeout) 46 46 { 47 47 cycles_t start, now, prev, end; 48 48 int i; ··· 121 121 * Source CPU calls into this - it waits for the freshly booted 122 122 * target CPU to arrive and then starts the measurement: 123 123 */ 124 - void __cpuinit check_tsc_sync_source(int cpu) 124 + void check_tsc_sync_source(int cpu) 125 125 { 126 126 int cpus = 2; 127 127 ··· 187 187 /* 188 188 * Freshly booted CPUs call into this: 189 189 */ 190 - void __cpuinit check_tsc_sync_target(void) 190 + void check_tsc_sync_target(void) 191 191 { 192 192 int cpus = 2; 193 193
+3 -3
arch/x86/kernel/vsyscall_64.c
··· 331 331 * Assume __initcall executes before all user space. Hopefully kmod 332 332 * doesn't violate that. We'll find out if it does. 333 333 */ 334 - static void __cpuinit vsyscall_set_cpu(int cpu) 334 + static void vsyscall_set_cpu(int cpu) 335 335 { 336 336 unsigned long d; 337 337 unsigned long node = 0; ··· 353 353 write_gdt_entry(get_cpu_gdt_table(cpu), GDT_ENTRY_PER_CPU, &d, DESCTYPE_S); 354 354 } 355 355 356 - static void __cpuinit cpu_vsyscall_init(void *arg) 356 + static void cpu_vsyscall_init(void *arg) 357 357 { 358 358 /* preemption should be already off */ 359 359 vsyscall_set_cpu(raw_smp_processor_id()); 360 360 } 361 361 362 - static int __cpuinit 362 + static int 363 363 cpu_vsyscall_notifier(struct notifier_block *n, unsigned long action, void *arg) 364 364 { 365 365 long cpu = (long)arg;
+2 -2
arch/x86/kernel/x86_init.c
··· 25 25 #include <asm/iommu.h> 26 26 #include <asm/mach_traps.h> 27 27 28 - void __cpuinit x86_init_noop(void) { } 28 + void x86_init_noop(void) { } 29 29 void __init x86_init_uint_noop(unsigned int unused) { } 30 30 int __init iommu_init_noop(void) { return 0; } 31 31 void iommu_shutdown_noop(void) { } ··· 85 85 }, 86 86 }; 87 87 88 - struct x86_cpuinit_ops x86_cpuinit __cpuinitdata = { 88 + struct x86_cpuinit_ops x86_cpuinit = { 89 89 .early_percpu_clock_init = x86_init_noop, 90 90 .setup_percpu_clockev = setup_secondary_APIC_clock, 91 91 };
+2 -2
arch/x86/kernel/xsave.c
··· 573 573 * This is somewhat obfuscated due to the lack of powerful enough 574 574 * overrides for the section checks. 575 575 */ 576 - void __cpuinit xsave_init(void) 576 + void xsave_init(void) 577 577 { 578 578 static __refdata void (*next_func)(void) = xstate_enable_boot_cpu; 579 579 void (*this_func)(void); ··· 594 594 setup_init_fpu_buf(); 595 595 } 596 596 597 - void __cpuinit eager_fpu_init(void) 597 + void eager_fpu_init(void) 598 598 { 599 599 static __refdata void (*boot_func)(void) = eager_fpu_init_bp; 600 600
+1 -3
arch/x86/mm/mmio-mod.c
··· 410 410 pr_warning("multiple CPUs still online, may miss events.\n"); 411 411 } 412 412 413 - /* __ref because leave_uniprocessor calls cpu_up which is __cpuinit, 414 - but this whole function is ifdefed CONFIG_HOTPLUG_CPU */ 415 - static void __ref leave_uniprocessor(void) 413 + static void leave_uniprocessor(void) 416 414 { 417 415 int cpu; 418 416 int err;
+6 -6
arch/x86/mm/numa.c
··· 60 60 [0 ... MAX_LOCAL_APIC-1] = NUMA_NO_NODE 61 61 }; 62 62 63 - int __cpuinit numa_cpu_node(int cpu) 63 + int numa_cpu_node(int cpu) 64 64 { 65 65 int apicid = early_per_cpu(x86_cpu_to_apicid, cpu); 66 66 ··· 691 691 #ifndef CONFIG_DEBUG_PER_CPU_MAPS 692 692 693 693 # ifndef CONFIG_NUMA_EMU 694 - void __cpuinit numa_add_cpu(int cpu) 694 + void numa_add_cpu(int cpu) 695 695 { 696 696 cpumask_set_cpu(cpu, node_to_cpumask_map[early_cpu_to_node(cpu)]); 697 697 } 698 698 699 - void __cpuinit numa_remove_cpu(int cpu) 699 + void numa_remove_cpu(int cpu) 700 700 { 701 701 cpumask_clear_cpu(cpu, node_to_cpumask_map[early_cpu_to_node(cpu)]); 702 702 } ··· 763 763 } 764 764 765 765 # ifndef CONFIG_NUMA_EMU 766 - static void __cpuinit numa_set_cpumask(int cpu, bool enable) 766 + static void numa_set_cpumask(int cpu, bool enable) 767 767 { 768 768 debug_cpumask_set_cpu(cpu, early_cpu_to_node(cpu), enable); 769 769 } 770 770 771 - void __cpuinit numa_add_cpu(int cpu) 771 + void numa_add_cpu(int cpu) 772 772 { 773 773 numa_set_cpumask(cpu, true); 774 774 } 775 775 776 - void __cpuinit numa_remove_cpu(int cpu) 776 + void numa_remove_cpu(int cpu) 777 777 { 778 778 numa_set_cpumask(cpu, false); 779 779 }
+6 -6
arch/x86/mm/numa_emulation.c
··· 10 10 11 11 #include "numa_internal.h" 12 12 13 - static int emu_nid_to_phys[MAX_NUMNODES] __cpuinitdata; 13 + static int emu_nid_to_phys[MAX_NUMNODES]; 14 14 static char *emu_cmdline __initdata; 15 15 16 16 void __init numa_emu_cmdline(char *str) ··· 444 444 } 445 445 446 446 #ifndef CONFIG_DEBUG_PER_CPU_MAPS 447 - void __cpuinit numa_add_cpu(int cpu) 447 + void numa_add_cpu(int cpu) 448 448 { 449 449 int physnid, nid; 450 450 ··· 462 462 cpumask_set_cpu(cpu, node_to_cpumask_map[nid]); 463 463 } 464 464 465 - void __cpuinit numa_remove_cpu(int cpu) 465 + void numa_remove_cpu(int cpu) 466 466 { 467 467 int i; 468 468 ··· 470 470 cpumask_clear_cpu(cpu, node_to_cpumask_map[i]); 471 471 } 472 472 #else /* !CONFIG_DEBUG_PER_CPU_MAPS */ 473 - static void __cpuinit numa_set_cpumask(int cpu, bool enable) 473 + static void numa_set_cpumask(int cpu, bool enable) 474 474 { 475 475 int nid, physnid; 476 476 ··· 490 490 } 491 491 } 492 492 493 - void __cpuinit numa_add_cpu(int cpu) 493 + void numa_add_cpu(int cpu) 494 494 { 495 495 numa_set_cpumask(cpu, true); 496 496 } 497 497 498 - void __cpuinit numa_remove_cpu(int cpu) 498 + void numa_remove_cpu(int cpu) 499 499 { 500 500 numa_set_cpumask(cpu, false); 501 501 }
+2 -2
arch/x86/mm/setup_nx.c
··· 5 5 #include <asm/pgtable.h> 6 6 #include <asm/proto.h> 7 7 8 - static int disable_nx __cpuinitdata; 8 + static int disable_nx; 9 9 10 10 /* 11 11 * noexec = on|off ··· 29 29 } 30 30 early_param("noexec", noexec_setup); 31 31 32 - void __cpuinit x86_configure_nx(void) 32 + void x86_configure_nx(void) 33 33 { 34 34 if (cpu_has_nx && !disable_nx) 35 35 __supported_pte_mask |= _PAGE_NX;
+4 -4
arch/x86/pci/amd_bus.c
··· 312 312 313 313 #define ENABLE_CF8_EXT_CFG (1ULL << 46) 314 314 315 - static void __cpuinit enable_pci_io_ecs(void *unused) 315 + static void enable_pci_io_ecs(void *unused) 316 316 { 317 317 u64 reg; 318 318 rdmsrl(MSR_AMD64_NB_CFG, reg); ··· 322 322 } 323 323 } 324 324 325 - static int __cpuinit amd_cpu_notify(struct notifier_block *self, 326 - unsigned long action, void *hcpu) 325 + static int amd_cpu_notify(struct notifier_block *self, unsigned long action, 326 + void *hcpu) 327 327 { 328 328 int cpu = (long)hcpu; 329 329 switch (action) { ··· 337 337 return NOTIFY_OK; 338 338 } 339 339 340 - static struct notifier_block __cpuinitdata amd_cpu_notifier = { 340 + static struct notifier_block amd_cpu_notifier = { 341 341 .notifier_call = amd_cpu_notify, 342 342 }; 343 343
+1 -1
arch/x86/platform/ce4100/ce4100.c
··· 134 134 } 135 135 136 136 #ifdef CONFIG_X86_IO_APIC 137 - static void __cpuinit sdv_pci_init(void) 137 + static void sdv_pci_init(void) 138 138 { 139 139 x86_of_pci_init(); 140 140 /* We can't set this earlier, because we need to calibrate the timer */
+2 -2
arch/x86/platform/mrst/mrst.c
··· 65 65 * lapic (always-on,ARAT) ------ 150 66 66 */ 67 67 68 - __cpuinitdata enum mrst_timer_options mrst_timer_options; 68 + enum mrst_timer_options mrst_timer_options; 69 69 70 70 static u32 sfi_mtimer_usage[SFI_MTMR_MAX_NUM]; 71 71 static struct sfi_timer_table_entry sfi_mtimer_array[SFI_MTMR_MAX_NUM]; ··· 248 248 apbt_time_init(); 249 249 } 250 250 251 - static void __cpuinit mrst_arch_setup(void) 251 + static void mrst_arch_setup(void) 252 252 { 253 253 if (boot_cpu_data.x86 == 6 && boot_cpu_data.x86_model == 0x27) 254 254 __mrst_cpu_chip = MRST_CPU_CHIP_PENWELL;
+3 -3
arch/x86/xen/enlighten.c
··· 1681 1681 xen_domain_type = XEN_HVM_DOMAIN; 1682 1682 } 1683 1683 1684 - static int __cpuinit xen_hvm_cpu_notify(struct notifier_block *self, 1685 - unsigned long action, void *hcpu) 1684 + static int xen_hvm_cpu_notify(struct notifier_block *self, unsigned long action, 1685 + void *hcpu) 1686 1686 { 1687 1687 int cpu = (long)hcpu; 1688 1688 switch (action) { ··· 1700 1700 return NOTIFY_OK; 1701 1701 } 1702 1702 1703 - static struct notifier_block xen_hvm_cpu_notifier __cpuinitdata = { 1703 + static struct notifier_block xen_hvm_cpu_notifier = { 1704 1704 .notifier_call = xen_hvm_cpu_notify, 1705 1705 }; 1706 1706
+3 -3
arch/x86/xen/setup.c
··· 475 475 #endif 476 476 } 477 477 478 - static int __cpuinit register_callback(unsigned type, const void *func) 478 + static int register_callback(unsigned type, const void *func) 479 479 { 480 480 struct callback_register callback = { 481 481 .type = type, ··· 486 486 return HYPERVISOR_callback_op(CALLBACKOP_register, &callback); 487 487 } 488 488 489 - void __cpuinit xen_enable_sysenter(void) 489 + void xen_enable_sysenter(void) 490 490 { 491 491 int ret; 492 492 unsigned sysenter_feature; ··· 505 505 setup_clear_cpu_cap(sysenter_feature); 506 506 } 507 507 508 - void __cpuinit xen_enable_syscall(void) 508 + void xen_enable_syscall(void) 509 509 { 510 510 #ifdef CONFIG_X86_64 511 511 int ret;
+6 -6
arch/x86/xen/smp.c
··· 65 65 return IRQ_HANDLED; 66 66 } 67 67 68 - static void __cpuinit cpu_bringup(void) 68 + static void cpu_bringup(void) 69 69 { 70 70 int cpu; 71 71 ··· 97 97 wmb(); /* make sure everything is out */ 98 98 } 99 99 100 - static void __cpuinit cpu_bringup_and_idle(void) 100 + static void cpu_bringup_and_idle(void) 101 101 { 102 102 cpu_bringup(); 103 103 cpu_startup_entry(CPUHP_ONLINE); ··· 326 326 set_cpu_present(cpu, true); 327 327 } 328 328 329 - static int __cpuinit 329 + static int 330 330 cpu_initialize_context(unsigned int cpu, struct task_struct *idle) 331 331 { 332 332 struct vcpu_guest_context *ctxt; ··· 397 397 return 0; 398 398 } 399 399 400 - static int __cpuinit xen_cpu_up(unsigned int cpu, struct task_struct *idle) 400 + static int xen_cpu_up(unsigned int cpu, struct task_struct *idle) 401 401 { 402 402 int rc; 403 403 ··· 470 470 xen_teardown_timer(cpu); 471 471 } 472 472 473 - static void __cpuinit xen_play_dead(void) /* used only with HOTPLUG_CPU */ 473 + static void xen_play_dead(void) /* used only with HOTPLUG_CPU */ 474 474 { 475 475 play_dead_common(); 476 476 HYPERVISOR_vcpu_op(VCPUOP_down, smp_processor_id(), NULL); ··· 691 691 xen_init_lock_cpu(0); 692 692 } 693 693 694 - static int __cpuinit xen_hvm_cpu_up(unsigned int cpu, struct task_struct *tidle) 694 + static int xen_hvm_cpu_up(unsigned int cpu, struct task_struct *tidle) 695 695 { 696 696 int rc; 697 697 rc = native_cpu_up(cpu, tidle);
+1 -1
arch/x86/xen/spinlock.c
··· 361 361 return IRQ_HANDLED; 362 362 } 363 363 364 - void __cpuinit xen_init_lock_cpu(int cpu) 364 + void xen_init_lock_cpu(int cpu) 365 365 { 366 366 int irq; 367 367 char *name;
+1 -1
arch/x86/xen/xen-ops.h
··· 73 73 74 74 #ifdef CONFIG_PARAVIRT_SPINLOCKS 75 75 void __init xen_init_spinlocks(void); 76 - void __cpuinit xen_init_lock_cpu(int cpu); 76 + void xen_init_lock_cpu(int cpu); 77 77 void xen_uninit_lock_cpu(int cpu); 78 78 #else 79 79 static inline void xen_init_spinlocks(void)
+1 -1
arch/xtensa/kernel/time.c
··· 162 162 } 163 163 164 164 #ifndef CONFIG_GENERIC_CALIBRATE_DELAY 165 - void __cpuinit calibrate_delay(void) 165 + void calibrate_delay(void) 166 166 { 167 167 loops_per_jiffy = CCOUNT_PER_JIFFY; 168 168 printk("Calibrating delay loop (skipped)... "
+3 -3
block/blk-iopoll.c
··· 189 189 } 190 190 EXPORT_SYMBOL(blk_iopoll_init); 191 191 192 - static int __cpuinit blk_iopoll_cpu_notify(struct notifier_block *self, 193 - unsigned long action, void *hcpu) 192 + static int blk_iopoll_cpu_notify(struct notifier_block *self, 193 + unsigned long action, void *hcpu) 194 194 { 195 195 /* 196 196 * If a CPU goes away, splice its entries to the current CPU ··· 209 209 return NOTIFY_OK; 210 210 } 211 211 212 - static struct notifier_block __cpuinitdata blk_iopoll_cpu_notifier = { 212 + static struct notifier_block blk_iopoll_cpu_notifier = { 213 213 .notifier_call = blk_iopoll_cpu_notify, 214 214 }; 215 215
+3 -3
block/blk-softirq.c
··· 78 78 } 79 79 #endif 80 80 81 - static int __cpuinit blk_cpu_notify(struct notifier_block *self, 82 - unsigned long action, void *hcpu) 81 + static int blk_cpu_notify(struct notifier_block *self, unsigned long action, 82 + void *hcpu) 83 83 { 84 84 /* 85 85 * If a CPU goes away, splice its entries to the current CPU ··· 98 98 return NOTIFY_OK; 99 99 } 100 100 101 - static struct notifier_block __cpuinitdata blk_cpu_notifier = { 101 + static struct notifier_block blk_cpu_notifier = { 102 102 .notifier_call = blk_cpu_notify, 103 103 }; 104 104
+1 -1
drivers/acpi/acpi_processor.c
··· 340 340 */ 341 341 static DEFINE_PER_CPU(void *, processor_device_array); 342 342 343 - static int __cpuinit acpi_processor_add(struct acpi_device *device, 343 + static int acpi_processor_add(struct acpi_device *device, 344 344 const struct acpi_device_id *id) 345 345 { 346 346 struct acpi_processor *pr;
+4 -4
drivers/acpi/processor_core.c
··· 253 253 return true; 254 254 } 255 255 256 - static void __cpuinit acpi_set_pdc_bits(u32 *buf) 256 + static void acpi_set_pdc_bits(u32 *buf) 257 257 { 258 258 buf[0] = ACPI_PDC_REVISION_ID; 259 259 buf[1] = 1; ··· 265 265 arch_acpi_set_pdc_bits(buf); 266 266 } 267 267 268 - static struct acpi_object_list *__cpuinit acpi_processor_alloc_pdc(void) 268 + static struct acpi_object_list *acpi_processor_alloc_pdc(void) 269 269 { 270 270 struct acpi_object_list *obj_list; 271 271 union acpi_object *obj; ··· 308 308 * _PDC is required for a BIOS-OS handshake for most of the newer 309 309 * ACPI processor features. 310 310 */ 311 - static int __cpuinit 311 + static int 312 312 acpi_processor_eval_pdc(acpi_handle handle, struct acpi_object_list *pdc_in) 313 313 { 314 314 acpi_status status = AE_OK; ··· 336 336 return status; 337 337 } 338 338 339 - void __cpuinit acpi_processor_set_pdc(acpi_handle handle) 339 + void acpi_processor_set_pdc(acpi_handle handle) 340 340 { 341 341 struct acpi_object_list *obj_list; 342 342
+4 -4
drivers/acpi/processor_driver.c
··· 118 118 return; 119 119 } 120 120 121 - static __cpuinit int __acpi_processor_start(struct acpi_device *device); 121 + static int __acpi_processor_start(struct acpi_device *device); 122 122 123 - static int __cpuinit acpi_cpu_soft_notify(struct notifier_block *nfb, 123 + static int acpi_cpu_soft_notify(struct notifier_block *nfb, 124 124 unsigned long action, void *hcpu) 125 125 { 126 126 unsigned int cpu = (unsigned long)hcpu; ··· 162 162 .notifier_call = acpi_cpu_soft_notify, 163 163 }; 164 164 165 - static __cpuinit int __acpi_processor_start(struct acpi_device *device) 165 + static int __acpi_processor_start(struct acpi_device *device) 166 166 { 167 167 struct acpi_processor *pr = acpi_driver_data(device); 168 168 acpi_status status; ··· 226 226 return result; 227 227 } 228 228 229 - static int __cpuinit acpi_processor_start(struct device *dev) 229 + static int acpi_processor_start(struct device *dev) 230 230 { 231 231 struct acpi_device *device; 232 232
+2 -4
drivers/acpi/processor_idle.c
··· 96 96 return 0; 97 97 } 98 98 99 - /* Actually this shouldn't be __cpuinitdata, would be better to fix the 100 - callers to only run once -AK */ 101 - static struct dmi_system_id __cpuinitdata processor_power_dmi_table[] = { 99 + static struct dmi_system_id processor_power_dmi_table[] = { 102 100 { set_max_cstate, "Clevo 5600D", { 103 101 DMI_MATCH(DMI_BIOS_VENDOR,"Phoenix Technologies LTD"), 104 102 DMI_MATCH(DMI_BIOS_VERSION,"SHE845M0.86C.0013.D.0302131307")}, ··· 1163 1165 1164 1166 static int acpi_processor_registered; 1165 1167 1166 - int __cpuinit acpi_processor_power_init(struct acpi_processor *pr) 1168 + int acpi_processor_power_init(struct acpi_processor *pr) 1167 1169 { 1168 1170 acpi_status status = 0; 1169 1171 int retval;
+1 -1
drivers/base/cpu.c
··· 278 278 * 279 279 * Initialize and register the CPU device. 280 280 */ 281 - int __cpuinit register_cpu(struct cpu *cpu, int num) 281 + int register_cpu(struct cpu *cpu, int num) 282 282 { 283 283 int error; 284 284
+5 -5
drivers/base/topology.c
··· 143 143 }; 144 144 145 145 /* Add/Remove cpu_topology interface for CPU device */ 146 - static int __cpuinit topology_add_dev(unsigned int cpu) 146 + static int topology_add_dev(unsigned int cpu) 147 147 { 148 148 struct device *dev = get_cpu_device(cpu); 149 149 150 150 return sysfs_create_group(&dev->kobj, &topology_attr_group); 151 151 } 152 152 153 - static void __cpuinit topology_remove_dev(unsigned int cpu) 153 + static void topology_remove_dev(unsigned int cpu) 154 154 { 155 155 struct device *dev = get_cpu_device(cpu); 156 156 157 157 sysfs_remove_group(&dev->kobj, &topology_attr_group); 158 158 } 159 159 160 - static int __cpuinit topology_cpu_callback(struct notifier_block *nfb, 161 - unsigned long action, void *hcpu) 160 + static int topology_cpu_callback(struct notifier_block *nfb, 161 + unsigned long action, void *hcpu) 162 162 { 163 163 unsigned int cpu = (unsigned long)hcpu; 164 164 int rc = 0; ··· 178 178 return notifier_from_errno(rc); 179 179 } 180 180 181 - static int __cpuinit topology_sysfs_init(void) 181 + static int topology_sysfs_init(void) 182 182 { 183 183 int cpu; 184 184 int rc;
+4 -4
drivers/clocksource/arm_arch_timer.c
··· 123 123 return 0; 124 124 } 125 125 126 - static int __cpuinit arch_timer_setup(struct clock_event_device *clk) 126 + static int arch_timer_setup(struct clock_event_device *clk) 127 127 { 128 128 clk->features = CLOCK_EVT_FEAT_ONESHOT | CLOCK_EVT_FEAT_C3STOP; 129 129 clk->name = "arch_sys_timer"; ··· 221 221 return &timecounter; 222 222 } 223 223 224 - static void __cpuinit arch_timer_stop(struct clock_event_device *clk) 224 + static void arch_timer_stop(struct clock_event_device *clk) 225 225 { 226 226 pr_debug("arch_timer_teardown disable IRQ%d cpu #%d\n", 227 227 clk->irq, smp_processor_id()); ··· 237 237 clk->set_mode(CLOCK_EVT_MODE_UNUSED, clk); 238 238 } 239 239 240 - static int __cpuinit arch_timer_cpu_notify(struct notifier_block *self, 240 + static int arch_timer_cpu_notify(struct notifier_block *self, 241 241 unsigned long action, void *hcpu) 242 242 { 243 243 /* ··· 256 256 return NOTIFY_OK; 257 257 } 258 258 259 - static struct notifier_block arch_timer_cpu_nb __cpuinitdata = { 259 + static struct notifier_block arch_timer_cpu_nb = { 260 260 .notifier_call = arch_timer_cpu_notify, 261 261 }; 262 262
+4 -4
drivers/clocksource/arm_global_timer.c
··· 164 164 return IRQ_HANDLED; 165 165 } 166 166 167 - static int __cpuinit gt_clockevents_init(struct clock_event_device *clk) 167 + static int gt_clockevents_init(struct clock_event_device *clk) 168 168 { 169 169 int cpu = smp_processor_id(); 170 170 ··· 221 221 clocksource_register_hz(&gt_clocksource, gt_clk_rate); 222 222 } 223 223 224 - static int __cpuinit gt_cpu_notify(struct notifier_block *self, 225 - unsigned long action, void *hcpu) 224 + static int gt_cpu_notify(struct notifier_block *self, unsigned long action, 225 + void *hcpu) 226 226 { 227 227 switch (action & ~CPU_TASKS_FROZEN) { 228 228 case CPU_STARTING: ··· 235 235 236 236 return NOTIFY_OK; 237 237 } 238 - static struct notifier_block gt_cpu_nb __cpuinitdata = { 238 + static struct notifier_block gt_cpu_nb = { 239 239 .notifier_call = gt_cpu_notify, 240 240 }; 241 241
+3 -3
drivers/clocksource/dummy_timer.c
··· 25 25 */ 26 26 } 27 27 28 - static void __cpuinit dummy_timer_setup(void) 28 + static void dummy_timer_setup(void) 29 29 { 30 30 int cpu = smp_processor_id(); 31 31 struct clock_event_device *evt = __this_cpu_ptr(&dummy_timer_evt); ··· 41 41 clockevents_register_device(evt); 42 42 } 43 43 44 - static int __cpuinit dummy_timer_cpu_notify(struct notifier_block *self, 44 + static int dummy_timer_cpu_notify(struct notifier_block *self, 45 45 unsigned long action, void *hcpu) 46 46 { 47 47 if ((action & ~CPU_TASKS_FROZEN) == CPU_STARTING) ··· 50 50 return NOTIFY_OK; 51 51 } 52 52 53 - static struct notifier_block dummy_timer_cpu_nb __cpuinitdata = { 53 + static struct notifier_block dummy_timer_cpu_nb = { 54 54 .notifier_call = dummy_timer_cpu_notify, 55 55 }; 56 56
+2 -2
drivers/clocksource/exynos_mct.c
··· 400 400 return IRQ_HANDLED; 401 401 } 402 402 403 - static int __cpuinit exynos4_local_timer_setup(struct clock_event_device *evt) 403 + static int exynos4_local_timer_setup(struct clock_event_device *evt) 404 404 { 405 405 struct mct_clock_event_device *mevt; 406 406 unsigned int cpu = smp_processor_id(); ··· 448 448 disable_percpu_irq(mct_irqs[MCT_L0_IRQ]); 449 449 } 450 450 451 - static struct local_timer_ops exynos4_mct_tick_ops __cpuinitdata = { 451 + static struct local_timer_ops exynos4_mct_tick_ops = { 452 452 .setup = exynos4_local_timer_setup, 453 453 .stop = exynos4_local_timer_stop, 454 454 };
+3 -3
drivers/clocksource/metag_generic.c
··· 109 109 return ticks << HARDWARE_TO_NS_SHIFT; 110 110 } 111 111 112 - static void __cpuinit arch_timer_setup(unsigned int cpu) 112 + static void arch_timer_setup(unsigned int cpu) 113 113 { 114 114 unsigned int txdivtime; 115 115 struct clock_event_device *clk = &per_cpu(local_clockevent, cpu); ··· 154 154 } 155 155 } 156 156 157 - static int __cpuinit arch_timer_cpu_notify(struct notifier_block *self, 157 + static int arch_timer_cpu_notify(struct notifier_block *self, 158 158 unsigned long action, void *hcpu) 159 159 { 160 160 int cpu = (long)hcpu; ··· 169 169 return NOTIFY_OK; 170 170 } 171 171 172 - static struct notifier_block __cpuinitdata arch_timer_cpu_nb = { 172 + static struct notifier_block arch_timer_cpu_nb = { 173 173 .notifier_call = arch_timer_cpu_notify, 174 174 }; 175 175
+2 -2
drivers/clocksource/time-armada-370-xp.c
··· 167 167 /* 168 168 * Setup the local clock events for a CPU. 169 169 */ 170 - static int __cpuinit armada_370_xp_timer_setup(struct clock_event_device *evt) 170 + static int armada_370_xp_timer_setup(struct clock_event_device *evt) 171 171 { 172 172 u32 u; 173 173 int cpu = smp_processor_id(); ··· 205 205 disable_percpu_irq(evt->irq); 206 206 } 207 207 208 - static struct local_timer_ops armada_370_xp_local_timer_ops __cpuinitdata = { 208 + static struct local_timer_ops armada_370_xp_local_timer_ops = { 209 209 .setup = armada_370_xp_timer_setup, 210 210 .stop = armada_370_xp_timer_stop, 211 211 };
+2 -2
drivers/clocksource/timer-marco.c
··· 184 184 .handler = sirfsoc_timer_interrupt, 185 185 }; 186 186 187 - static int __cpuinit sirfsoc_local_timer_setup(struct clock_event_device *ce) 187 + static int sirfsoc_local_timer_setup(struct clock_event_device *ce) 188 188 { 189 189 /* Use existing clock_event for cpu 0 */ 190 190 if (!smp_processor_id()) ··· 216 216 remove_irq(sirfsoc_timer1_irq.irq, &sirfsoc_timer1_irq); 217 217 } 218 218 219 - static struct local_timer_ops sirfsoc_local_timer_ops __cpuinitdata = { 219 + static struct local_timer_ops sirfsoc_local_timer_ops = { 220 220 .setup = sirfsoc_local_timer_setup, 221 221 .stop = sirfsoc_local_timer_stop, 222 222 };
+1 -1
drivers/cpufreq/cpufreq.c
··· 1932 1932 } 1933 1933 EXPORT_SYMBOL(cpufreq_update_policy); 1934 1934 1935 - static int __cpuinit cpufreq_cpu_callback(struct notifier_block *nfb, 1935 + static int cpufreq_cpu_callback(struct notifier_block *nfb, 1936 1936 unsigned long action, void *hcpu) 1937 1937 { 1938 1938 unsigned int cpu = (unsigned long)hcpu;
+1 -1
drivers/cpufreq/cpufreq_stats.c
··· 341 341 return 0; 342 342 } 343 343 344 - static int __cpuinit cpufreq_stat_cpu_callback(struct notifier_block *nfb, 344 + static int cpufreq_stat_cpu_callback(struct notifier_block *nfb, 345 345 unsigned long action, 346 346 void *hcpu) 347 347 {
+1 -1
drivers/cpufreq/dbx500-cpufreq.c
··· 82 82 return freq_table[i].frequency; 83 83 } 84 84 85 - static int __cpuinit dbx500_cpufreq_init(struct cpufreq_policy *policy) 85 + static int dbx500_cpufreq_init(struct cpufreq_policy *policy) 86 86 { 87 87 int res; 88 88
+2 -2
drivers/cpufreq/intel_pstate.c
··· 617 617 return 0; 618 618 } 619 619 620 - static int __cpuinit intel_pstate_cpu_exit(struct cpufreq_policy *policy) 620 + static int intel_pstate_cpu_exit(struct cpufreq_policy *policy) 621 621 { 622 622 int cpu = policy->cpu; 623 623 ··· 627 627 return 0; 628 628 } 629 629 630 - static int __cpuinit intel_pstate_cpu_init(struct cpufreq_policy *policy) 630 + static int intel_pstate_cpu_init(struct cpufreq_policy *policy) 631 631 { 632 632 int rc, min_pstate, max_pstate; 633 633 struct cpudata *cpu;
+3 -3
drivers/cpufreq/longhaul.c
··· 422 422 } 423 423 424 424 425 - static int __cpuinit longhaul_get_ranges(void) 425 + static int longhaul_get_ranges(void) 426 426 { 427 427 unsigned int i, j, k = 0; 428 428 unsigned int ratio; ··· 526 526 } 527 527 528 528 529 - static void __cpuinit longhaul_setup_voltagescaling(void) 529 + static void longhaul_setup_voltagescaling(void) 530 530 { 531 531 union msr_longhaul longhaul; 532 532 struct mV_pos minvid, maxvid, vid; ··· 780 780 return 0; 781 781 } 782 782 783 - static int __cpuinit longhaul_cpu_init(struct cpufreq_policy *policy) 783 + static int longhaul_cpu_init(struct cpufreq_policy *policy) 784 784 { 785 785 struct cpuinfo_x86 *c = &cpu_data(0); 786 786 char *cpuname = NULL;
+13 -13
drivers/cpufreq/longhaul.h
··· 56 56 /* 57 57 * VIA C3 Samuel 1 & Samuel 2 (stepping 0) 58 58 */ 59 - static const int __cpuinitconst samuel1_mults[16] = { 59 + static const int samuel1_mults[16] = { 60 60 -1, /* 0000 -> RESERVED */ 61 61 30, /* 0001 -> 3.0x */ 62 62 40, /* 0010 -> 4.0x */ ··· 75 75 -1, /* 1111 -> RESERVED */ 76 76 }; 77 77 78 - static const int __cpuinitconst samuel1_eblcr[16] = { 78 + static const int samuel1_eblcr[16] = { 79 79 50, /* 0000 -> RESERVED */ 80 80 30, /* 0001 -> 3.0x */ 81 81 40, /* 0010 -> 4.0x */ ··· 97 97 /* 98 98 * VIA C3 Samuel2 Stepping 1->15 99 99 */ 100 - static const int __cpuinitconst samuel2_eblcr[16] = { 100 + static const int samuel2_eblcr[16] = { 101 101 50, /* 0000 -> 5.0x */ 102 102 30, /* 0001 -> 3.0x */ 103 103 40, /* 0010 -> 4.0x */ ··· 119 119 /* 120 120 * VIA C3 Ezra 121 121 */ 122 - static const int __cpuinitconst ezra_mults[16] = { 122 + static const int ezra_mults[16] = { 123 123 100, /* 0000 -> 10.0x */ 124 124 30, /* 0001 -> 3.0x */ 125 125 40, /* 0010 -> 4.0x */ ··· 138 138 120, /* 1111 -> 12.0x */ 139 139 }; 140 140 141 - static const int __cpuinitconst ezra_eblcr[16] = { 141 + static const int ezra_eblcr[16] = { 142 142 50, /* 0000 -> 5.0x */ 143 143 30, /* 0001 -> 3.0x */ 144 144 40, /* 0010 -> 4.0x */ ··· 160 160 /* 161 161 * VIA C3 (Ezra-T) [C5M]. 162 162 */ 163 - static const int __cpuinitconst ezrat_mults[32] = { 163 + static const int ezrat_mults[32] = { 164 164 100, /* 0000 -> 10.0x */ 165 165 30, /* 0001 -> 3.0x */ 166 166 40, /* 0010 -> 4.0x */ ··· 196 196 -1, /* 1111 -> RESERVED (12.0x) */ 197 197 }; 198 198 199 - static const int __cpuinitconst ezrat_eblcr[32] = { 199 + static const int ezrat_eblcr[32] = { 200 200 50, /* 0000 -> 5.0x */ 201 201 30, /* 0001 -> 3.0x */ 202 202 40, /* 0010 -> 4.0x */ ··· 235 235 /* 236 236 * VIA C3 Nehemiah */ 237 237 238 - static const int __cpuinitconst nehemiah_mults[32] = { 238 + static const int nehemiah_mults[32] = { 239 239 100, /* 0000 -> 10.0x */ 240 240 -1, /* 0001 -> 16.0x */ 241 241 40, /* 0010 -> 4.0x */ ··· 270 270 -1, /* 1111 -> 12.0x */ 271 271 }; 272 272 273 - static const int __cpuinitconst nehemiah_eblcr[32] = { 273 + static const int nehemiah_eblcr[32] = { 274 274 50, /* 0000 -> 5.0x */ 275 275 160, /* 0001 -> 16.0x */ 276 276 40, /* 0010 -> 4.0x */ ··· 315 315 unsigned short pos; 316 316 }; 317 317 318 - static const struct mV_pos __cpuinitconst vrm85_mV[32] = { 318 + static const struct mV_pos vrm85_mV[32] = { 319 319 {1250, 8}, {1200, 6}, {1150, 4}, {1100, 2}, 320 320 {1050, 0}, {1800, 30}, {1750, 28}, {1700, 26}, 321 321 {1650, 24}, {1600, 22}, {1550, 20}, {1500, 18}, ··· 326 326 {1475, 17}, {1425, 15}, {1375, 13}, {1325, 11} 327 327 }; 328 328 329 - static const unsigned char __cpuinitconst mV_vrm85[32] = { 329 + static const unsigned char mV_vrm85[32] = { 330 330 0x04, 0x14, 0x03, 0x13, 0x02, 0x12, 0x01, 0x11, 331 331 0x00, 0x10, 0x0f, 0x1f, 0x0e, 0x1e, 0x0d, 0x1d, 332 332 0x0c, 0x1c, 0x0b, 0x1b, 0x0a, 0x1a, 0x09, 0x19, 333 333 0x08, 0x18, 0x07, 0x17, 0x06, 0x16, 0x05, 0x15 334 334 }; 335 335 336 - static const struct mV_pos __cpuinitconst mobilevrm_mV[32] = { 336 + static const struct mV_pos mobilevrm_mV[32] = { 337 337 {1750, 31}, {1700, 30}, {1650, 29}, {1600, 28}, 338 338 {1550, 27}, {1500, 26}, {1450, 25}, {1400, 24}, 339 339 {1350, 23}, {1300, 22}, {1250, 21}, {1200, 20}, ··· 344 344 {675, 3}, {650, 2}, {625, 1}, {600, 0} 345 345 }; 346 346 347 - static const unsigned char __cpuinitconst mV_mobilevrm[32] = { 347 + static const unsigned char mV_mobilevrm[32] = { 348 348 0x1f, 0x1e, 0x1d, 0x1c, 0x1b, 0x1a, 0x19, 0x18, 349 349 0x17, 0x16, 0x15, 0x14, 0x13, 0x12, 0x11, 0x10, 350 350 0x0f, 0x0e, 0x0d, 0x0c, 0x0b, 0x0a, 0x09, 0x08,
+3 -3
drivers/cpufreq/longrun.c
··· 33 33 * Reads the current LongRun policy by access to MSR_TMTA_LONGRUN_FLAGS 34 34 * and MSR_TMTA_LONGRUN_CTRL 35 35 */ 36 - static void __cpuinit longrun_get_policy(struct cpufreq_policy *policy) 36 + static void longrun_get_policy(struct cpufreq_policy *policy) 37 37 { 38 38 u32 msr_lo, msr_hi; 39 39 ··· 163 163 * TMTA rules: 164 164 * performance_pctg = (target_freq - low_freq)/(high_freq - low_freq) 165 165 */ 166 - static int __cpuinit longrun_determine_freqs(unsigned int *low_freq, 166 + static int longrun_determine_freqs(unsigned int *low_freq, 167 167 unsigned int *high_freq) 168 168 { 169 169 u32 msr_lo, msr_hi; ··· 256 256 } 257 257 258 258 259 - static int __cpuinit longrun_cpu_init(struct cpufreq_policy *policy) 259 + static int longrun_cpu_init(struct cpufreq_policy *policy) 260 260 { 261 261 int result = 0; 262 262
+1 -1
drivers/cpufreq/omap-cpufreq.c
··· 165 165 opp_free_cpufreq_table(mpu_dev, &freq_table); 166 166 } 167 167 168 - static int __cpuinit omap_cpu_init(struct cpufreq_policy *policy) 168 + static int omap_cpu_init(struct cpufreq_policy *policy) 169 169 { 170 170 int result = 0; 171 171
+4 -4
drivers/cpufreq/powernow-k7.c
··· 563 563 * We will then get the same kind of behaviour already tested under 564 564 * the "well-known" other OS. 565 565 */ 566 - static int __cpuinit fixup_sgtc(void) 566 + static int fixup_sgtc(void) 567 567 { 568 568 unsigned int sgtc; 569 569 unsigned int m; ··· 597 597 } 598 598 599 599 600 - static int __cpuinit acer_cpufreq_pst(const struct dmi_system_id *d) 600 + static int acer_cpufreq_pst(const struct dmi_system_id *d) 601 601 { 602 602 printk(KERN_WARNING PFX 603 603 "%s laptop with broken PST tables in BIOS detected.\n", ··· 615 615 * A BIOS update is all that can save them. 616 616 * Mention this, and disable cpufreq. 617 617 */ 618 - static struct dmi_system_id __cpuinitdata powernow_dmi_table[] = { 618 + static struct dmi_system_id powernow_dmi_table[] = { 619 619 { 620 620 .callback = acer_cpufreq_pst, 621 621 .ident = "Acer Aspire", ··· 627 627 { } 628 628 }; 629 629 630 - static int __cpuinit powernow_cpu_init(struct cpufreq_policy *policy) 630 + static int powernow_cpu_init(struct cpufreq_policy *policy) 631 631 { 632 632 union msr_fidvidstatus fidvidstatus; 633 633 int result;
+3 -3
drivers/cpufreq/powernow-k8.c
··· 1069 1069 int rc; 1070 1070 }; 1071 1071 1072 - static void __cpuinit powernowk8_cpu_init_on_cpu(void *_init_on_cpu) 1072 + static void powernowk8_cpu_init_on_cpu(void *_init_on_cpu) 1073 1073 { 1074 1074 struct init_on_cpu *init_on_cpu = _init_on_cpu; 1075 1075 ··· 1096 1096 FW_BUG PFX "If that doesn't help, try upgrading your BIOS.\n"; 1097 1097 1098 1098 /* per CPU init entry point to the driver */ 1099 - static int __cpuinit powernowk8_cpu_init(struct cpufreq_policy *pol) 1099 + static int powernowk8_cpu_init(struct cpufreq_policy *pol) 1100 1100 { 1101 1101 struct powernow_k8_data *data; 1102 1102 struct init_on_cpu init_on_cpu; ··· 1263 1263 } 1264 1264 1265 1265 /* driver entry point for init */ 1266 - static int __cpuinit powernowk8_init(void) 1266 + static int powernowk8_init(void) 1267 1267 { 1268 1268 unsigned int i, supported_cpus = 0; 1269 1269 int ret;
+18 -21
drivers/hwmon/coretemp.c
··· 195 195 int tjmax; 196 196 }; 197 197 198 - static const struct tjmax __cpuinitconst tjmax_table[] = { 198 + static const struct tjmax tjmax_table[] = { 199 199 { "CPU 230", 100000 }, /* Model 0x1c, stepping 2 */ 200 200 { "CPU 330", 125000 }, /* Model 0x1c, stepping 2 */ 201 201 { "CPU CE4110", 110000 }, /* Model 0x1c, stepping 10 Sodaville */ ··· 211 211 212 212 #define ANY 0xff 213 213 214 - static const struct tjmax_model __cpuinitconst tjmax_model_table[] = { 214 + static const struct tjmax_model tjmax_model_table[] = { 215 215 { 0x1c, 10, 100000 }, /* D4xx, K4xx, N4xx, D5xx, K5xx, N5xx */ 216 216 { 0x1c, ANY, 90000 }, /* Z5xx, N2xx, possibly others 217 217 * Note: Also matches 230 and 330, ··· 226 226 { 0x36, ANY, 100000 }, /* Atom Cedar Trail/Cedarview (N2xxx, D2xxx) */ 227 227 }; 228 228 229 - static int __cpuinit adjust_tjmax(struct cpuinfo_x86 *c, u32 id, 230 - struct device *dev) 229 + static int adjust_tjmax(struct cpuinfo_x86 *c, u32 id, struct device *dev) 231 230 { 232 231 /* The 100C is default for both mobile and non mobile CPUs */ 233 232 ··· 316 317 return tjmax; 317 318 } 318 319 319 - static int __cpuinit get_tjmax(struct cpuinfo_x86 *c, u32 id, 320 - struct device *dev) 320 + static int get_tjmax(struct cpuinfo_x86 *c, u32 id, struct device *dev) 321 321 { 322 322 int err; 323 323 u32 eax, edx; ··· 365 367 return device_create_file(dev, &pdata->name_attr); 366 368 } 367 369 368 - static int __cpuinit create_core_attrs(struct temp_data *tdata, 369 - struct device *dev, int attr_no) 370 + static int create_core_attrs(struct temp_data *tdata, struct device *dev, 371 + int attr_no) 370 372 { 371 373 int err, i; 372 374 static ssize_t (*const rd_ptr[TOTAL_ATTRS]) (struct device *dev, ··· 399 401 } 400 402 401 403 402 - static int __cpuinit chk_ucode_version(unsigned int cpu) 404 + static int chk_ucode_version(unsigned int cpu) 403 405 { 404 406 struct cpuinfo_x86 *c = &cpu_data(cpu); 405 407 ··· 415 417 return 0; 416 418 } 417 419 418 - static struct platform_device __cpuinit *coretemp_get_pdev(unsigned int cpu) 420 + static struct platform_device *coretemp_get_pdev(unsigned int cpu) 419 421 { 420 422 u16 phys_proc_id = TO_PHYS_ID(cpu); 421 423 struct pdev_entry *p; ··· 432 434 return NULL; 433 435 } 434 436 435 - static struct temp_data __cpuinit *init_temp_data(unsigned int cpu, 436 - int pkg_flag) 437 + static struct temp_data *init_temp_data(unsigned int cpu, int pkg_flag) 437 438 { 438 439 struct temp_data *tdata; 439 440 ··· 450 453 return tdata; 451 454 } 452 455 453 - static int __cpuinit create_core_data(struct platform_device *pdev, 454 - unsigned int cpu, int pkg_flag) 456 + static int create_core_data(struct platform_device *pdev, unsigned int cpu, 457 + int pkg_flag) 455 458 { 456 459 struct temp_data *tdata; 457 460 struct platform_data *pdata = platform_get_drvdata(pdev); ··· 521 524 return err; 522 525 } 523 526 524 - static void __cpuinit coretemp_add_core(unsigned int cpu, int pkg_flag) 527 + static void coretemp_add_core(unsigned int cpu, int pkg_flag) 525 528 { 526 529 struct platform_device *pdev = coretemp_get_pdev(cpu); 527 530 int err; ··· 604 607 .remove = coretemp_remove, 605 608 }; 606 609 607 - static int __cpuinit coretemp_device_add(unsigned int cpu) 610 + static int coretemp_device_add(unsigned int cpu) 608 611 { 609 612 int err; 610 613 struct platform_device *pdev; ··· 648 651 return err; 649 652 } 650 653 651 - static void __cpuinit coretemp_device_remove(unsigned int cpu) 654 + static void coretemp_device_remove(unsigned int cpu) 652 655 { 653 656 struct pdev_entry *p, *n; 654 657 u16 phys_proc_id = TO_PHYS_ID(cpu); ··· 664 667 mutex_unlock(&pdev_list_mutex); 665 668 } 666 669 667 - static bool __cpuinit is_any_core_online(struct platform_data *pdata) 670 + static bool is_any_core_online(struct platform_data *pdata) 668 671 { 669 672 int i; 670 673 ··· 678 681 return false; 679 682 } 680 683 681 - static void __cpuinit get_core_online(unsigned int cpu) 684 + static void get_core_online(unsigned int cpu) 682 685 { 683 686 struct cpuinfo_x86 *c = &cpu_data(cpu); 684 687 struct platform_device *pdev = coretemp_get_pdev(cpu); ··· 720 723 coretemp_add_core(cpu, 0); 721 724 } 722 725 723 - static void __cpuinit put_core_offline(unsigned int cpu) 726 + static void put_core_offline(unsigned int cpu) 724 727 { 725 728 int i, indx; 726 729 struct platform_data *pdata; ··· 768 771 coretemp_device_remove(cpu); 769 772 } 770 773 771 - static int __cpuinit coretemp_cpu_callback(struct notifier_block *nfb, 774 + static int coretemp_cpu_callback(struct notifier_block *nfb, 772 775 unsigned long action, void *hcpu) 773 776 { 774 777 unsigned int cpu = (unsigned long) hcpu;
+4 -4
drivers/hwmon/via-cputemp.c
··· 221 221 static LIST_HEAD(pdev_list); 222 222 static DEFINE_MUTEX(pdev_list_mutex); 223 223 224 - static int __cpuinit via_cputemp_device_add(unsigned int cpu) 224 + static int via_cputemp_device_add(unsigned int cpu) 225 225 { 226 226 int err; 227 227 struct platform_device *pdev; ··· 262 262 return err; 263 263 } 264 264 265 - static void __cpuinit via_cputemp_device_remove(unsigned int cpu) 265 + static void via_cputemp_device_remove(unsigned int cpu) 266 266 { 267 267 struct pdev_entry *p; 268 268 ··· 279 279 mutex_unlock(&pdev_list_mutex); 280 280 } 281 281 282 - static int __cpuinit via_cputemp_cpu_callback(struct notifier_block *nfb, 283 - unsigned long action, void *hcpu) 282 + static int via_cputemp_cpu_callback(struct notifier_block *nfb, 283 + unsigned long action, void *hcpu) 284 284 { 285 285 unsigned int cpu = (unsigned long) hcpu; 286 286
+4 -4
drivers/irqchip/irq-gic.c
··· 414 414 writel_relaxed(1, base + GIC_DIST_CTRL); 415 415 } 416 416 417 - static void __cpuinit gic_cpu_init(struct gic_chip_data *gic) 417 + static void gic_cpu_init(struct gic_chip_data *gic) 418 418 { 419 419 void __iomem *dist_base = gic_data_dist_base(gic); 420 420 void __iomem *base = gic_data_cpu_base(gic); ··· 702 702 } 703 703 704 704 #ifdef CONFIG_SMP 705 - static int __cpuinit gic_secondary_init(struct notifier_block *nfb, 706 - unsigned long action, void *hcpu) 705 + static int gic_secondary_init(struct notifier_block *nfb, unsigned long action, 706 + void *hcpu) 707 707 { 708 708 if (action == CPU_STARTING || action == CPU_STARTING_FROZEN) 709 709 gic_cpu_init(&gic_data[0]); ··· 714 714 * Notifier for enabling the GIC CPU interface. Set an arbitrarily high 715 715 * priority because the GIC needs to be up before the ARM generic timers. 716 716 */ 717 - static struct notifier_block __cpuinitdata gic_cpu_notifier = { 717 + static struct notifier_block gic_cpu_notifier = { 718 718 .notifier_call = gic_secondary_init, 719 719 .priority = 100, 720 720 };
+2 -2
drivers/oprofile/timer_int.c
··· 74 74 put_online_cpus(); 75 75 } 76 76 77 - static int __cpuinit oprofile_cpu_notify(struct notifier_block *self, 78 - unsigned long action, void *hcpu) 77 + static int oprofile_cpu_notify(struct notifier_block *self, 78 + unsigned long action, void *hcpu) 79 79 { 80 80 long cpu = (long) hcpu; 81 81
+1 -1
drivers/xen/xen-acpi-cpuhotplug.c
··· 91 91 return 0; 92 92 } 93 93 94 - static int __cpuinit xen_acpi_processor_add(struct acpi_device *device) 94 + static int xen_acpi_processor_add(struct acpi_device *device) 95 95 { 96 96 int ret; 97 97 struct acpi_processor *pr;
+1 -1
include/linux/cpu.h
··· 114 114 /* Need to know about CPUs going up/down? */ 115 115 #if defined(CONFIG_HOTPLUG_CPU) || !defined(MODULE) 116 116 #define cpu_notifier(fn, pri) { \ 117 - static struct notifier_block fn##_nb __cpuinitdata = \ 117 + static struct notifier_block fn##_nb = \ 118 118 { .notifier_call = fn, .priority = pri }; \ 119 119 register_cpu_notifier(&fn##_nb); \ 120 120 }
+1 -1
include/linux/perf_event.h
··· 826 826 */ 827 827 #define perf_cpu_notifier(fn) \ 828 828 do { \ 829 - static struct notifier_block fn##_nb __cpuinitdata = \ 829 + static struct notifier_block fn##_nb = \ 830 830 { .notifier_call = fn, .priority = CPU_PRI_PERF }; \ 831 831 unsigned long cpu = smp_processor_id(); \ 832 832 unsigned long flags; \
+8 -5
init/calibrate.c
··· 31 31 #define DELAY_CALIBRATION_TICKS ((HZ < 100) ? 1 : (HZ/100)) 32 32 #define MAX_DIRECT_CALIBRATION_RETRIES 5 33 33 34 - static unsigned long __cpuinit calibrate_delay_direct(void) 34 + static unsigned long calibrate_delay_direct(void) 35 35 { 36 36 unsigned long pre_start, start, post_start; 37 37 unsigned long pre_end, end, post_end; ··· 166 166 return 0; 167 167 } 168 168 #else 169 - static unsigned long __cpuinit calibrate_delay_direct(void) {return 0;} 169 + static unsigned long calibrate_delay_direct(void) 170 + { 171 + return 0; 172 + } 170 173 #endif 171 174 172 175 /* ··· 183 180 */ 184 181 #define LPS_PREC 8 185 182 186 - static unsigned long __cpuinit calibrate_delay_converge(void) 183 + static unsigned long calibrate_delay_converge(void) 187 184 { 188 185 /* First stage - slowly accelerate to find initial bounds */ 189 186 unsigned long lpj, lpj_base, ticks, loopadd, loopadd_base, chop_limit; ··· 257 254 * Architectures should override this function if a faster calibration 258 255 * method is available. 259 256 */ 260 - unsigned long __attribute__((weak)) __cpuinit calibrate_delay_is_known(void) 257 + unsigned long __attribute__((weak)) calibrate_delay_is_known(void) 261 258 { 262 259 return 0; 263 260 } 264 261 265 - void __cpuinit calibrate_delay(void) 262 + void calibrate_delay(void) 266 263 { 267 264 unsigned long lpj; 268 265 static bool printed;
+3 -3
kernel/cpu.c
··· 366 366 #endif /*CONFIG_HOTPLUG_CPU*/ 367 367 368 368 /* Requires cpu_add_remove_lock to be held */ 369 - static int __cpuinit _cpu_up(unsigned int cpu, int tasks_frozen) 369 + static int _cpu_up(unsigned int cpu, int tasks_frozen) 370 370 { 371 371 int ret, nr_calls = 0; 372 372 void *hcpu = (void *)(long)cpu; ··· 419 419 return ret; 420 420 } 421 421 422 - int __cpuinit cpu_up(unsigned int cpu) 422 + int cpu_up(unsigned int cpu) 423 423 { 424 424 int err = 0; 425 425 ··· 618 618 * It must be called by the arch code on the new cpu, before the new cpu 619 619 * enables interrupts and before the "boot" cpu returns from __cpu_up(). 620 620 */ 621 - void __cpuinit notify_cpu_starting(unsigned int cpu) 621 + void notify_cpu_starting(unsigned int cpu) 622 622 { 623 623 unsigned long val = CPU_STARTING; 624 624
+2 -2
kernel/events/core.c
··· 7630 7630 } 7631 7631 } 7632 7632 7633 - static void __cpuinit perf_event_init_cpu(int cpu) 7633 + static void perf_event_init_cpu(int cpu) 7634 7634 { 7635 7635 struct swevent_htable *swhash = &per_cpu(swevent_htable, cpu); 7636 7636 ··· 7719 7719 .priority = INT_MIN, 7720 7720 }; 7721 7721 7722 - static int __cpuinit 7722 + static int 7723 7723 perf_cpu_notify(struct notifier_block *self, unsigned long action, void *hcpu) 7724 7724 { 7725 7725 unsigned int cpu = (long)hcpu;
+1 -1
kernel/fork.c
··· 1546 1546 } 1547 1547 } 1548 1548 1549 - struct task_struct * __cpuinit fork_idle(int cpu) 1549 + struct task_struct *fork_idle(int cpu) 1550 1550 { 1551 1551 struct task_struct *task; 1552 1552 task = copy_process(CLONE_VM, 0, 0, NULL, &init_struct_pid, 0);
+3 -3
kernel/hrtimer.c
··· 1659 1659 /* 1660 1660 * Functions related to boot-time initialization: 1661 1661 */ 1662 - static void __cpuinit init_hrtimers_cpu(int cpu) 1662 + static void init_hrtimers_cpu(int cpu) 1663 1663 { 1664 1664 struct hrtimer_cpu_base *cpu_base = &per_cpu(hrtimer_bases, cpu); 1665 1665 int i; ··· 1740 1740 1741 1741 #endif /* CONFIG_HOTPLUG_CPU */ 1742 1742 1743 - static int __cpuinit hrtimer_cpu_notify(struct notifier_block *self, 1743 + static int hrtimer_cpu_notify(struct notifier_block *self, 1744 1744 unsigned long action, void *hcpu) 1745 1745 { 1746 1746 int scpu = (long)hcpu; ··· 1773 1773 return NOTIFY_OK; 1774 1774 } 1775 1775 1776 - static struct notifier_block __cpuinitdata hrtimers_nb = { 1776 + static struct notifier_block hrtimers_nb = { 1777 1777 .notifier_call = hrtimer_cpu_notify, 1778 1778 }; 1779 1779
+1 -1
kernel/printk.c
··· 1921 1921 * called when a new CPU comes online (or fails to come up), and ensures 1922 1922 * that any such output gets printed. 1923 1923 */ 1924 - static int __cpuinit console_cpu_notify(struct notifier_block *self, 1924 + static int console_cpu_notify(struct notifier_block *self, 1925 1925 unsigned long action, void *hcpu) 1926 1926 { 1927 1927 switch (action) {
+1 -1
kernel/profile.c
··· 331 331 put_cpu(); 332 332 } 333 333 334 - static int __cpuinit profile_cpu_callback(struct notifier_block *info, 334 + static int profile_cpu_callback(struct notifier_block *info, 335 335 unsigned long action, void *__cpu) 336 336 { 337 337 int node, cpu = (unsigned long)__cpu;
+3 -3
kernel/rcutorture.c
··· 1476 1476 * Execute random CPU-hotplug operations at the interval specified 1477 1477 * by the onoff_interval. 1478 1478 */ 1479 - static int __cpuinit 1479 + static int 1480 1480 rcu_torture_onoff(void *arg) 1481 1481 { 1482 1482 int cpu; ··· 1558 1558 return 0; 1559 1559 } 1560 1560 1561 - static int __cpuinit 1561 + static int 1562 1562 rcu_torture_onoff_init(void) 1563 1563 { 1564 1564 int ret; ··· 1601 1601 * CPU-stall kthread. It waits as specified by stall_cpu_holdoff, then 1602 1602 * induces a CPU stall for the time specified by stall_cpu. 1603 1603 */ 1604 - static int __cpuinit rcu_torture_stall(void *args) 1604 + static int rcu_torture_stall(void *args) 1605 1605 { 1606 1606 unsigned long stop_at; 1607 1607
+3 -3
kernel/rcutree.c
··· 2910 2910 * can accept some slop in the rsp->completed access due to the fact 2911 2911 * that this CPU cannot possibly have any RCU callbacks in flight yet. 2912 2912 */ 2913 - static void __cpuinit 2913 + static void 2914 2914 rcu_init_percpu_data(int cpu, struct rcu_state *rsp, int preemptible) 2915 2915 { 2916 2916 unsigned long flags; ··· 2962 2962 mutex_unlock(&rsp->onoff_mutex); 2963 2963 } 2964 2964 2965 - static void __cpuinit rcu_prepare_cpu(int cpu) 2965 + static void rcu_prepare_cpu(int cpu) 2966 2966 { 2967 2967 struct rcu_state *rsp; 2968 2968 ··· 2974 2974 /* 2975 2975 * Handle CPU online/offline notification events. 2976 2976 */ 2977 - static int __cpuinit rcu_cpu_notify(struct notifier_block *self, 2977 + static int rcu_cpu_notify(struct notifier_block *self, 2978 2978 unsigned long action, void *hcpu) 2979 2979 { 2980 2980 long cpu = (long)hcpu;
+2 -2
kernel/rcutree.h
··· 521 521 static bool rcu_is_callbacks_kthread(void); 522 522 #ifdef CONFIG_RCU_BOOST 523 523 static void rcu_preempt_do_callbacks(void); 524 - static int __cpuinit rcu_spawn_one_boost_kthread(struct rcu_state *rsp, 524 + static int rcu_spawn_one_boost_kthread(struct rcu_state *rsp, 525 525 struct rcu_node *rnp); 526 526 #endif /* #ifdef CONFIG_RCU_BOOST */ 527 - static void __cpuinit rcu_prepare_kthreads(int cpu); 527 + static void rcu_prepare_kthreads(int cpu); 528 528 static void rcu_cleanup_after_idle(int cpu); 529 529 static void rcu_prepare_for_idle(int cpu); 530 530 static void rcu_idle_count_callbacks_posted(void);
+3 -3
kernel/rcutree_plugin.h
··· 1352 1352 * already exist. We only create this kthread for preemptible RCU. 1353 1353 * Returns zero if all is well, a negated errno otherwise. 1354 1354 */ 1355 - static int __cpuinit rcu_spawn_one_boost_kthread(struct rcu_state *rsp, 1355 + static int rcu_spawn_one_boost_kthread(struct rcu_state *rsp, 1356 1356 struct rcu_node *rnp) 1357 1357 { 1358 1358 int rnp_index = rnp - &rsp->node[0]; ··· 1507 1507 } 1508 1508 early_initcall(rcu_spawn_kthreads); 1509 1509 1510 - static void __cpuinit rcu_prepare_kthreads(int cpu) 1510 + static void rcu_prepare_kthreads(int cpu) 1511 1511 { 1512 1512 struct rcu_data *rdp = per_cpu_ptr(rcu_state->rda, cpu); 1513 1513 struct rcu_node *rnp = rdp->mynode; ··· 1549 1549 } 1550 1550 early_initcall(rcu_scheduler_really_started); 1551 1551 1552 - static void __cpuinit rcu_prepare_kthreads(int cpu) 1552 + static void rcu_prepare_kthreads(int cpu) 1553 1553 { 1554 1554 } 1555 1555
+1 -1
kernel/relay.c
··· 516 516 * 517 517 * Returns the success/failure of the operation. (%NOTIFY_OK, %NOTIFY_BAD) 518 518 */ 519 - static int __cpuinit relay_hotcpu_callback(struct notifier_block *nb, 519 + static int relay_hotcpu_callback(struct notifier_block *nb, 520 520 unsigned long action, 521 521 void *hcpu) 522 522 {
+6 -6
kernel/sched/core.c
··· 4133 4133 debug_show_all_locks(); 4134 4134 } 4135 4135 4136 - void __cpuinit init_idle_bootup_task(struct task_struct *idle) 4136 + void init_idle_bootup_task(struct task_struct *idle) 4137 4137 { 4138 4138 idle->sched_class = &idle_sched_class; 4139 4139 } ··· 4146 4146 * NOTE: this function does not set the idle thread's NEED_RESCHED 4147 4147 * flag, to make booting more robust. 4148 4148 */ 4149 - void __cpuinit init_idle(struct task_struct *idle, int cpu) 4149 + void init_idle(struct task_struct *idle, int cpu) 4150 4150 { 4151 4151 struct rq *rq = cpu_rq(cpu); 4152 4152 unsigned long flags; ··· 4630 4630 * migration_call - callback that gets triggered when a CPU is added. 4631 4631 * Here we can start up the necessary migration thread for the new CPU. 4632 4632 */ 4633 - static int __cpuinit 4633 + static int 4634 4634 migration_call(struct notifier_block *nfb, unsigned long action, void *hcpu) 4635 4635 { 4636 4636 int cpu = (long)hcpu; ··· 4684 4684 * happens before everything else. This has to be lower priority than 4685 4685 * the notifier in the perf_event subsystem, though. 4686 4686 */ 4687 - static struct notifier_block __cpuinitdata migration_notifier = { 4687 + static struct notifier_block migration_notifier = { 4688 4688 .notifier_call = migration_call, 4689 4689 .priority = CPU_PRI_MIGRATION, 4690 4690 }; 4691 4691 4692 - static int __cpuinit sched_cpu_active(struct notifier_block *nfb, 4692 + static int sched_cpu_active(struct notifier_block *nfb, 4693 4693 unsigned long action, void *hcpu) 4694 4694 { 4695 4695 switch (action & ~CPU_TASKS_FROZEN) { ··· 4702 4702 } 4703 4703 } 4704 4704 4705 - static int __cpuinit sched_cpu_inactive(struct notifier_block *nfb, 4705 + static int sched_cpu_inactive(struct notifier_block *nfb, 4706 4706 unsigned long action, void *hcpu) 4707 4707 { 4708 4708 switch (action & ~CPU_TASKS_FROZEN) {
+1 -1
kernel/sched/fair.c
··· 5506 5506 set_bit(NOHZ_TICK_STOPPED, nohz_flags(cpu)); 5507 5507 } 5508 5508 5509 - static int __cpuinit sched_ilb_notifier(struct notifier_block *nfb, 5509 + static int sched_ilb_notifier(struct notifier_block *nfb, 5510 5510 unsigned long action, void *hcpu) 5511 5511 { 5512 5512 switch (action & ~CPU_TASKS_FROZEN) {
+1 -1
kernel/smp.c
··· 73 73 return NOTIFY_OK; 74 74 } 75 75 76 - static struct notifier_block __cpuinitdata hotplug_cfd_notifier = { 76 + static struct notifier_block hotplug_cfd_notifier = { 77 77 .notifier_call = hotplug_cfd, 78 78 }; 79 79
+1 -1
kernel/smpboot.c
··· 24 24 */ 25 25 static DEFINE_PER_CPU(struct task_struct *, idle_threads); 26 26 27 - struct task_struct * __cpuinit idle_thread_get(unsigned int cpu) 27 + struct task_struct *idle_thread_get(unsigned int cpu) 28 28 { 29 29 struct task_struct *tsk = per_cpu(idle_threads, cpu); 30 30
+4 -4
kernel/softirq.c
··· 699 699 } 700 700 EXPORT_SYMBOL(send_remote_softirq); 701 701 702 - static int __cpuinit remote_softirq_cpu_notify(struct notifier_block *self, 702 + static int remote_softirq_cpu_notify(struct notifier_block *self, 703 703 unsigned long action, void *hcpu) 704 704 { 705 705 /* ··· 728 728 return NOTIFY_OK; 729 729 } 730 730 731 - static struct notifier_block __cpuinitdata remote_softirq_cpu_notifier = { 731 + static struct notifier_block remote_softirq_cpu_notifier = { 732 732 .notifier_call = remote_softirq_cpu_notify, 733 733 }; 734 734 ··· 830 830 } 831 831 #endif /* CONFIG_HOTPLUG_CPU */ 832 832 833 - static int __cpuinit cpu_callback(struct notifier_block *nfb, 833 + static int cpu_callback(struct notifier_block *nfb, 834 834 unsigned long action, 835 835 void *hcpu) 836 836 { ··· 845 845 return NOTIFY_OK; 846 846 } 847 847 848 - static struct notifier_block __cpuinitdata cpu_nfb = { 848 + static struct notifier_block cpu_nfb = { 849 849 .notifier_call = cpu_callback 850 850 }; 851 851
+1 -1
kernel/time/tick-sched.c
··· 298 298 } 299 299 __setup("nohz_full=", tick_nohz_full_setup); 300 300 301 - static int __cpuinit tick_nohz_cpu_down_callback(struct notifier_block *nfb, 301 + static int tick_nohz_cpu_down_callback(struct notifier_block *nfb, 302 302 unsigned long action, 303 303 void *hcpu) 304 304 {
+5 -5
kernel/timer.c
··· 1505 1505 } 1506 1506 EXPORT_SYMBOL(schedule_timeout_uninterruptible); 1507 1507 1508 - static int __cpuinit init_timers_cpu(int cpu) 1508 + static int init_timers_cpu(int cpu) 1509 1509 { 1510 1510 int j; 1511 1511 struct tvec_base *base; 1512 - static char __cpuinitdata tvec_base_done[NR_CPUS]; 1512 + static char tvec_base_done[NR_CPUS]; 1513 1513 1514 1514 if (!tvec_base_done[cpu]) { 1515 1515 static char boot_done; ··· 1577 1577 } 1578 1578 } 1579 1579 1580 - static void __cpuinit migrate_timers(int cpu) 1580 + static void migrate_timers(int cpu) 1581 1581 { 1582 1582 struct tvec_base *old_base; 1583 1583 struct tvec_base *new_base; ··· 1610 1610 } 1611 1611 #endif /* CONFIG_HOTPLUG_CPU */ 1612 1612 1613 - static int __cpuinit timer_cpu_notify(struct notifier_block *self, 1613 + static int timer_cpu_notify(struct notifier_block *self, 1614 1614 unsigned long action, void *hcpu) 1615 1615 { 1616 1616 long cpu = (long)hcpu; ··· 1635 1635 return NOTIFY_OK; 1636 1636 } 1637 1637 1638 - static struct notifier_block __cpuinitdata timers_nb = { 1638 + static struct notifier_block timers_nb = { 1639 1639 .notifier_call = timer_cpu_notify, 1640 1640 }; 1641 1641
+2 -2
kernel/workqueue.c
··· 4644 4644 * Workqueues should be brought up before normal priority CPU notifiers. 4645 4645 * This will be registered high priority CPU notifier. 4646 4646 */ 4647 - static int __cpuinit workqueue_cpu_up_callback(struct notifier_block *nfb, 4647 + static int workqueue_cpu_up_callback(struct notifier_block *nfb, 4648 4648 unsigned long action, 4649 4649 void *hcpu) 4650 4650 { ··· 4697 4697 * Workqueues should be brought down after normal priority CPU notifiers. 4698 4698 * This will be registered as low priority CPU notifier. 4699 4699 */ 4700 - static int __cpuinit workqueue_cpu_down_callback(struct notifier_block *nfb, 4700 + static int workqueue_cpu_down_callback(struct notifier_block *nfb, 4701 4701 unsigned long action, 4702 4702 void *hcpu) 4703 4703 {
+1 -1
lib/Kconfig.debug
··· 238 238 any use of code/data previously in these sections would 239 239 most likely result in an oops. 240 240 In the code, functions and variables are annotated with 241 - __init, __cpuinit, etc. (see the full list in include/linux/init.h), 241 + __init,, etc. (see the full list in include/linux/init.h), 242 242 which results in the code/data being placed in specific sections. 243 243 The section mismatch analysis is always performed after a full 244 244 kernel build, and enabling this option causes the following
+1 -1
lib/earlycpio.c
··· 63 63 * the match returned an empty filename string. 64 64 */ 65 65 66 - struct cpio_data __cpuinit find_cpio_data(const char *path, void *data, 66 + struct cpio_data find_cpio_data(const char *path, void *data, 67 67 size_t len, long *offset) 68 68 { 69 69 const size_t cpio_header_len = 8*C_NFIELDS - 2;
+1 -1
lib/percpu_counter.c
··· 158 158 percpu_counter_batch = max(32, nr*2); 159 159 } 160 160 161 - static int __cpuinit percpu_counter_hotcpu_callback(struct notifier_block *nb, 161 + static int percpu_counter_hotcpu_callback(struct notifier_block *nb, 162 162 unsigned long action, void *hcpu) 163 163 { 164 164 #ifdef CONFIG_HOTPLUG_CPU
+1 -1
mm/memcontrol.c
··· 2522 2522 spin_unlock(&memcg->pcp_counter_lock); 2523 2523 } 2524 2524 2525 - static int __cpuinit memcg_cpu_hotplug_callback(struct notifier_block *nb, 2525 + static int memcg_cpu_hotplug_callback(struct notifier_block *nb, 2526 2526 unsigned long action, 2527 2527 void *hcpu) 2528 2528 {
+2 -2
mm/page-writeback.c
··· 1619 1619 ratelimit_pages = 16; 1620 1620 } 1621 1621 1622 - static int __cpuinit 1622 + static int 1623 1623 ratelimit_handler(struct notifier_block *self, unsigned long action, 1624 1624 void *hcpu) 1625 1625 { ··· 1634 1634 } 1635 1635 } 1636 1636 1637 - static struct notifier_block __cpuinitdata ratelimit_nb = { 1637 + static struct notifier_block ratelimit_nb = { 1638 1638 .notifier_call = ratelimit_handler, 1639 1639 .next = NULL, 1640 1640 };
+5 -5
mm/slab.c
··· 787 787 * the CPUs getting into lockstep and contending for the global cache chain 788 788 * lock. 789 789 */ 790 - static void __cpuinit start_cpu_timer(int cpu) 790 + static void start_cpu_timer(int cpu) 791 791 { 792 792 struct delayed_work *reap_work = &per_cpu(slab_reap_work, cpu); 793 793 ··· 1186 1186 return (n->free_objects + cachep->num - 1) / cachep->num; 1187 1187 } 1188 1188 1189 - static void __cpuinit cpuup_canceled(long cpu) 1189 + static void cpuup_canceled(long cpu) 1190 1190 { 1191 1191 struct kmem_cache *cachep; 1192 1192 struct kmem_cache_node *n = NULL; ··· 1251 1251 } 1252 1252 } 1253 1253 1254 - static int __cpuinit cpuup_prepare(long cpu) 1254 + static int cpuup_prepare(long cpu) 1255 1255 { 1256 1256 struct kmem_cache *cachep; 1257 1257 struct kmem_cache_node *n = NULL; ··· 1334 1334 return -ENOMEM; 1335 1335 } 1336 1336 1337 - static int __cpuinit cpuup_callback(struct notifier_block *nfb, 1337 + static int cpuup_callback(struct notifier_block *nfb, 1338 1338 unsigned long action, void *hcpu) 1339 1339 { 1340 1340 long cpu = (long)hcpu; ··· 1390 1390 return notifier_from_errno(err); 1391 1391 } 1392 1392 1393 - static struct notifier_block __cpuinitdata cpucache_notifier = { 1393 + static struct notifier_block cpucache_notifier = { 1394 1394 &cpuup_callback, NULL, 0 1395 1395 }; 1396 1396
+2 -2
mm/slub.c
··· 3773 3773 * Use the cpu notifier to insure that the cpu slabs are flushed when 3774 3774 * necessary. 3775 3775 */ 3776 - static int __cpuinit slab_cpuup_callback(struct notifier_block *nfb, 3776 + static int slab_cpuup_callback(struct notifier_block *nfb, 3777 3777 unsigned long action, void *hcpu) 3778 3778 { 3779 3779 long cpu = (long)hcpu; ··· 3799 3799 return NOTIFY_OK; 3800 3800 } 3801 3801 3802 - static struct notifier_block __cpuinitdata slab_notifier = { 3802 + static struct notifier_block slab_notifier = { 3803 3803 .notifier_call = slab_cpuup_callback 3804 3804 }; 3805 3805
+3 -3
mm/vmstat.c
··· 1182 1182 round_jiffies_relative(sysctl_stat_interval)); 1183 1183 } 1184 1184 1185 - static void __cpuinit start_cpu_timer(int cpu) 1185 + static void start_cpu_timer(int cpu) 1186 1186 { 1187 1187 struct delayed_work *work = &per_cpu(vmstat_work, cpu); 1188 1188 ··· 1194 1194 * Use the cpu notifier to insure that the thresholds are recalculated 1195 1195 * when necessary. 1196 1196 */ 1197 - static int __cpuinit vmstat_cpuup_callback(struct notifier_block *nfb, 1197 + static int vmstat_cpuup_callback(struct notifier_block *nfb, 1198 1198 unsigned long action, 1199 1199 void *hcpu) 1200 1200 { ··· 1226 1226 return NOTIFY_OK; 1227 1227 } 1228 1228 1229 - static struct notifier_block __cpuinitdata vmstat_notifier = 1229 + static struct notifier_block vmstat_notifier = 1230 1230 { &vmstat_cpuup_callback, NULL, 0 }; 1231 1231 #endif 1232 1232
+2 -2
net/core/flow.c
··· 403 403 schedule_work(&flow_cache_flush_work); 404 404 } 405 405 406 - static int __cpuinit flow_cache_cpu_prepare(struct flow_cache *fc, int cpu) 406 + static int flow_cache_cpu_prepare(struct flow_cache *fc, int cpu) 407 407 { 408 408 struct flow_cache_percpu *fcp = per_cpu_ptr(fc->percpu, cpu); 409 409 size_t sz = sizeof(struct hlist_head) * flow_cache_hash_size(fc); ··· 421 421 return 0; 422 422 } 423 423 424 - static int __cpuinit flow_cache_cpu(struct notifier_block *nfb, 424 + static int flow_cache_cpu(struct notifier_block *nfb, 425 425 unsigned long action, 426 426 void *hcpu) 427 427 {
+1 -1
net/iucv/iucv.c
··· 621 621 put_online_cpus(); 622 622 } 623 623 624 - static int __cpuinit iucv_cpu_notify(struct notifier_block *self, 624 + static int iucv_cpu_notify(struct notifier_block *self, 625 625 unsigned long action, void *hcpu) 626 626 { 627 627 cpumask_t cpumask;