at v2.6.35 22 kB view raw
1/* 2 * linux/init/main.c 3 * 4 * Copyright (C) 1991, 1992 Linus Torvalds 5 * 6 * GK 2/5/95 - Changed to support mounting root fs via NFS 7 * Added initrd & change_root: Werner Almesberger & Hans Lermen, Feb '96 8 * Moan early if gcc is old, avoiding bogus kernels - Paul Gortmaker, May '96 9 * Simplified starting of init: Michael A. Griffith <grif@acm.org> 10 */ 11 12#include <linux/types.h> 13#include <linux/module.h> 14#include <linux/proc_fs.h> 15#include <linux/kernel.h> 16#include <linux/syscalls.h> 17#include <linux/stackprotector.h> 18#include <linux/string.h> 19#include <linux/ctype.h> 20#include <linux/delay.h> 21#include <linux/ioport.h> 22#include <linux/init.h> 23#include <linux/smp_lock.h> 24#include <linux/initrd.h> 25#include <linux/bootmem.h> 26#include <linux/acpi.h> 27#include <linux/tty.h> 28#include <linux/percpu.h> 29#include <linux/kmod.h> 30#include <linux/vmalloc.h> 31#include <linux/kernel_stat.h> 32#include <linux/start_kernel.h> 33#include <linux/security.h> 34#include <linux/smp.h> 35#include <linux/workqueue.h> 36#include <linux/profile.h> 37#include <linux/rcupdate.h> 38#include <linux/moduleparam.h> 39#include <linux/kallsyms.h> 40#include <linux/writeback.h> 41#include <linux/cpu.h> 42#include <linux/cpuset.h> 43#include <linux/cgroup.h> 44#include <linux/efi.h> 45#include <linux/tick.h> 46#include <linux/interrupt.h> 47#include <linux/taskstats_kern.h> 48#include <linux/delayacct.h> 49#include <linux/unistd.h> 50#include <linux/rmap.h> 51#include <linux/mempolicy.h> 52#include <linux/key.h> 53#include <linux/buffer_head.h> 54#include <linux/page_cgroup.h> 55#include <linux/debug_locks.h> 56#include <linux/debugobjects.h> 57#include <linux/lockdep.h> 58#include <linux/kmemleak.h> 59#include <linux/pid_namespace.h> 60#include <linux/device.h> 61#include <linux/kthread.h> 62#include <linux/sched.h> 63#include <linux/signal.h> 64#include <linux/idr.h> 65#include <linux/kgdb.h> 66#include <linux/ftrace.h> 67#include <linux/async.h> 68#include <linux/kmemcheck.h> 69#include <linux/kmemtrace.h> 70#include <linux/sfi.h> 71#include <linux/shmem_fs.h> 72#include <linux/slab.h> 73#include <trace/boot.h> 74 75#include <asm/io.h> 76#include <asm/bugs.h> 77#include <asm/setup.h> 78#include <asm/sections.h> 79#include <asm/cacheflush.h> 80 81#ifdef CONFIG_X86_LOCAL_APIC 82#include <asm/smp.h> 83#endif 84 85static int kernel_init(void *); 86 87extern void init_IRQ(void); 88extern void fork_init(unsigned long); 89extern void mca_init(void); 90extern void sbus_init(void); 91extern void prio_tree_init(void); 92extern void radix_tree_init(void); 93extern void free_initmem(void); 94#ifndef CONFIG_DEBUG_RODATA 95static inline void mark_rodata_ro(void) { } 96#endif 97 98#ifdef CONFIG_TC 99extern void tc_init(void); 100#endif 101 102enum system_states system_state __read_mostly; 103EXPORT_SYMBOL(system_state); 104 105/* 106 * Boot command-line arguments 107 */ 108#define MAX_INIT_ARGS CONFIG_INIT_ENV_ARG_LIMIT 109#define MAX_INIT_ENVS CONFIG_INIT_ENV_ARG_LIMIT 110 111extern void time_init(void); 112/* Default late time init is NULL. archs can override this later. */ 113void (*__initdata late_time_init)(void); 114extern void softirq_init(void); 115 116/* Untouched command line saved by arch-specific code. */ 117char __initdata boot_command_line[COMMAND_LINE_SIZE]; 118/* Untouched saved command line (eg. for /proc) */ 119char *saved_command_line; 120/* Command line for parameter parsing */ 121static char *static_command_line; 122 123static char *execute_command; 124static char *ramdisk_execute_command; 125 126#ifdef CONFIG_SMP 127/* Setup configured maximum number of CPUs to activate */ 128unsigned int setup_max_cpus = NR_CPUS; 129EXPORT_SYMBOL(setup_max_cpus); 130 131 132/* 133 * Setup routine for controlling SMP activation 134 * 135 * Command-line option of "nosmp" or "maxcpus=0" will disable SMP 136 * activation entirely (the MPS table probe still happens, though). 137 * 138 * Command-line option of "maxcpus=<NUM>", where <NUM> is an integer 139 * greater than 0, limits the maximum number of CPUs activated in 140 * SMP mode to <NUM>. 141 */ 142 143void __weak arch_disable_smp_support(void) { } 144 145static int __init nosmp(char *str) 146{ 147 setup_max_cpus = 0; 148 arch_disable_smp_support(); 149 150 return 0; 151} 152 153early_param("nosmp", nosmp); 154 155/* this is hard limit */ 156static int __init nrcpus(char *str) 157{ 158 int nr_cpus; 159 160 get_option(&str, &nr_cpus); 161 if (nr_cpus > 0 && nr_cpus < nr_cpu_ids) 162 nr_cpu_ids = nr_cpus; 163 164 return 0; 165} 166 167early_param("nr_cpus", nrcpus); 168 169static int __init maxcpus(char *str) 170{ 171 get_option(&str, &setup_max_cpus); 172 if (setup_max_cpus == 0) 173 arch_disable_smp_support(); 174 175 return 0; 176} 177 178early_param("maxcpus", maxcpus); 179#else 180static const unsigned int setup_max_cpus = NR_CPUS; 181#endif 182 183/* 184 * If set, this is an indication to the drivers that reset the underlying 185 * device before going ahead with the initialization otherwise driver might 186 * rely on the BIOS and skip the reset operation. 187 * 188 * This is useful if kernel is booting in an unreliable environment. 189 * For ex. kdump situaiton where previous kernel has crashed, BIOS has been 190 * skipped and devices will be in unknown state. 191 */ 192unsigned int reset_devices; 193EXPORT_SYMBOL(reset_devices); 194 195static int __init set_reset_devices(char *str) 196{ 197 reset_devices = 1; 198 return 1; 199} 200 201__setup("reset_devices", set_reset_devices); 202 203static char * argv_init[MAX_INIT_ARGS+2] = { "init", NULL, }; 204char * envp_init[MAX_INIT_ENVS+2] = { "HOME=/", "TERM=linux", NULL, }; 205static const char *panic_later, *panic_param; 206 207extern struct obs_kernel_param __setup_start[], __setup_end[]; 208 209static int __init obsolete_checksetup(char *line) 210{ 211 struct obs_kernel_param *p; 212 int had_early_param = 0; 213 214 p = __setup_start; 215 do { 216 int n = strlen(p->str); 217 if (!strncmp(line, p->str, n)) { 218 if (p->early) { 219 /* Already done in parse_early_param? 220 * (Needs exact match on param part). 221 * Keep iterating, as we can have early 222 * params and __setups of same names 8( */ 223 if (line[n] == '\0' || line[n] == '=') 224 had_early_param = 1; 225 } else if (!p->setup_func) { 226 printk(KERN_WARNING "Parameter %s is obsolete," 227 " ignored\n", p->str); 228 return 1; 229 } else if (p->setup_func(line + n)) 230 return 1; 231 } 232 p++; 233 } while (p < __setup_end); 234 235 return had_early_param; 236} 237 238/* 239 * This should be approx 2 Bo*oMips to start (note initial shift), and will 240 * still work even if initially too large, it will just take slightly longer 241 */ 242unsigned long loops_per_jiffy = (1<<12); 243 244EXPORT_SYMBOL(loops_per_jiffy); 245 246static int __init debug_kernel(char *str) 247{ 248 console_loglevel = 10; 249 return 0; 250} 251 252static int __init quiet_kernel(char *str) 253{ 254 console_loglevel = 4; 255 return 0; 256} 257 258early_param("debug", debug_kernel); 259early_param("quiet", quiet_kernel); 260 261static int __init loglevel(char *str) 262{ 263 get_option(&str, &console_loglevel); 264 return 0; 265} 266 267early_param("loglevel", loglevel); 268 269/* 270 * Unknown boot options get handed to init, unless they look like 271 * unused parameters (modprobe will find them in /proc/cmdline). 272 */ 273static int __init unknown_bootoption(char *param, char *val) 274{ 275 /* Change NUL term back to "=", to make "param" the whole string. */ 276 if (val) { 277 /* param=val or param="val"? */ 278 if (val == param+strlen(param)+1) 279 val[-1] = '='; 280 else if (val == param+strlen(param)+2) { 281 val[-2] = '='; 282 memmove(val-1, val, strlen(val)+1); 283 val--; 284 } else 285 BUG(); 286 } 287 288 /* Handle obsolete-style parameters */ 289 if (obsolete_checksetup(param)) 290 return 0; 291 292 /* Unused module parameter. */ 293 if (strchr(param, '.') && (!val || strchr(param, '.') < val)) 294 return 0; 295 296 if (panic_later) 297 return 0; 298 299 if (val) { 300 /* Environment option */ 301 unsigned int i; 302 for (i = 0; envp_init[i]; i++) { 303 if (i == MAX_INIT_ENVS) { 304 panic_later = "Too many boot env vars at `%s'"; 305 panic_param = param; 306 } 307 if (!strncmp(param, envp_init[i], val - param)) 308 break; 309 } 310 envp_init[i] = param; 311 } else { 312 /* Command line option */ 313 unsigned int i; 314 for (i = 0; argv_init[i]; i++) { 315 if (i == MAX_INIT_ARGS) { 316 panic_later = "Too many boot init vars at `%s'"; 317 panic_param = param; 318 } 319 } 320 argv_init[i] = param; 321 } 322 return 0; 323} 324 325#ifdef CONFIG_DEBUG_PAGEALLOC 326int __read_mostly debug_pagealloc_enabled = 0; 327#endif 328 329static int __init init_setup(char *str) 330{ 331 unsigned int i; 332 333 execute_command = str; 334 /* 335 * In case LILO is going to boot us with default command line, 336 * it prepends "auto" before the whole cmdline which makes 337 * the shell think it should execute a script with such name. 338 * So we ignore all arguments entered _before_ init=... [MJ] 339 */ 340 for (i = 1; i < MAX_INIT_ARGS; i++) 341 argv_init[i] = NULL; 342 return 1; 343} 344__setup("init=", init_setup); 345 346static int __init rdinit_setup(char *str) 347{ 348 unsigned int i; 349 350 ramdisk_execute_command = str; 351 /* See "auto" comment in init_setup */ 352 for (i = 1; i < MAX_INIT_ARGS; i++) 353 argv_init[i] = NULL; 354 return 1; 355} 356__setup("rdinit=", rdinit_setup); 357 358#ifndef CONFIG_SMP 359 360#ifdef CONFIG_X86_LOCAL_APIC 361static void __init smp_init(void) 362{ 363 APIC_init_uniprocessor(); 364} 365#else 366#define smp_init() do { } while (0) 367#endif 368 369static inline void setup_nr_cpu_ids(void) { } 370static inline void smp_prepare_cpus(unsigned int maxcpus) { } 371 372#else 373 374/* Setup number of possible processor ids */ 375int nr_cpu_ids __read_mostly = NR_CPUS; 376EXPORT_SYMBOL(nr_cpu_ids); 377 378/* An arch may set nr_cpu_ids earlier if needed, so this would be redundant */ 379static void __init setup_nr_cpu_ids(void) 380{ 381 nr_cpu_ids = find_last_bit(cpumask_bits(cpu_possible_mask),NR_CPUS) + 1; 382} 383 384/* Called by boot processor to activate the rest. */ 385static void __init smp_init(void) 386{ 387 unsigned int cpu; 388 389 /* FIXME: This should be done in userspace --RR */ 390 for_each_present_cpu(cpu) { 391 if (num_online_cpus() >= setup_max_cpus) 392 break; 393 if (!cpu_online(cpu)) 394 cpu_up(cpu); 395 } 396 397 /* Any cleanup work */ 398 printk(KERN_INFO "Brought up %ld CPUs\n", (long)num_online_cpus()); 399 smp_cpus_done(setup_max_cpus); 400} 401 402#endif 403 404/* 405 * We need to store the untouched command line for future reference. 406 * We also need to store the touched command line since the parameter 407 * parsing is performed in place, and we should allow a component to 408 * store reference of name/value for future reference. 409 */ 410static void __init setup_command_line(char *command_line) 411{ 412 saved_command_line = alloc_bootmem(strlen (boot_command_line)+1); 413 static_command_line = alloc_bootmem(strlen (command_line)+1); 414 strcpy (saved_command_line, boot_command_line); 415 strcpy (static_command_line, command_line); 416} 417 418/* 419 * We need to finalize in a non-__init function or else race conditions 420 * between the root thread and the init thread may cause start_kernel to 421 * be reaped by free_initmem before the root thread has proceeded to 422 * cpu_idle. 423 * 424 * gcc-3.4 accidentally inlines this function, so use noinline. 425 */ 426 427static __initdata DECLARE_COMPLETION(kthreadd_done); 428 429static noinline void __init_refok rest_init(void) 430 __releases(kernel_lock) 431{ 432 int pid; 433 434 rcu_scheduler_starting(); 435 /* 436 * We need to spawn init first so that it obtains pid 1, however 437 * the init task will end up wanting to create kthreads, which, if 438 * we schedule it before we create kthreadd, will OOPS. 439 */ 440 kernel_thread(kernel_init, NULL, CLONE_FS | CLONE_SIGHAND); 441 numa_default_policy(); 442 pid = kernel_thread(kthreadd, NULL, CLONE_FS | CLONE_FILES); 443 rcu_read_lock(); 444 kthreadd_task = find_task_by_pid_ns(pid, &init_pid_ns); 445 rcu_read_unlock(); 446 complete(&kthreadd_done); 447 unlock_kernel(); 448 449 /* 450 * The boot idle thread must execute schedule() 451 * at least once to get things moving: 452 */ 453 init_idle_bootup_task(current); 454 preempt_enable_no_resched(); 455 schedule(); 456 preempt_disable(); 457 458 /* Call into cpu_idle with preempt disabled */ 459 cpu_idle(); 460} 461 462/* Check for early params. */ 463static int __init do_early_param(char *param, char *val) 464{ 465 struct obs_kernel_param *p; 466 467 for (p = __setup_start; p < __setup_end; p++) { 468 if ((p->early && strcmp(param, p->str) == 0) || 469 (strcmp(param, "console") == 0 && 470 strcmp(p->str, "earlycon") == 0) 471 ) { 472 if (p->setup_func(val) != 0) 473 printk(KERN_WARNING 474 "Malformed early option '%s'\n", param); 475 } 476 } 477 /* We accept everything at this stage. */ 478 return 0; 479} 480 481void __init parse_early_options(char *cmdline) 482{ 483 parse_args("early options", cmdline, NULL, 0, do_early_param); 484} 485 486/* Arch code calls this early on, or if not, just before other parsing. */ 487void __init parse_early_param(void) 488{ 489 static __initdata int done = 0; 490 static __initdata char tmp_cmdline[COMMAND_LINE_SIZE]; 491 492 if (done) 493 return; 494 495 /* All fall through to do_early_param. */ 496 strlcpy(tmp_cmdline, boot_command_line, COMMAND_LINE_SIZE); 497 parse_early_options(tmp_cmdline); 498 done = 1; 499} 500 501/* 502 * Activate the first processor. 503 */ 504 505static void __init boot_cpu_init(void) 506{ 507 int cpu = smp_processor_id(); 508 /* Mark the boot cpu "present", "online" etc for SMP and UP case */ 509 set_cpu_online(cpu, true); 510 set_cpu_active(cpu, true); 511 set_cpu_present(cpu, true); 512 set_cpu_possible(cpu, true); 513} 514 515void __init __weak smp_setup_processor_id(void) 516{ 517} 518 519void __init __weak thread_info_cache_init(void) 520{ 521} 522 523/* 524 * Set up kernel memory allocators 525 */ 526static void __init mm_init(void) 527{ 528 /* 529 * page_cgroup requires countinous pages as memmap 530 * and it's bigger than MAX_ORDER unless SPARSEMEM. 531 */ 532 page_cgroup_init_flatmem(); 533 mem_init(); 534 kmem_cache_init(); 535 pgtable_cache_init(); 536 vmalloc_init(); 537} 538 539asmlinkage void __init start_kernel(void) 540{ 541 char * command_line; 542 extern struct kernel_param __start___param[], __stop___param[]; 543 544 smp_setup_processor_id(); 545 546 /* 547 * Need to run as early as possible, to initialize the 548 * lockdep hash: 549 */ 550 lockdep_init(); 551 debug_objects_early_init(); 552 553 /* 554 * Set up the the initial canary ASAP: 555 */ 556 boot_init_stack_canary(); 557 558 cgroup_init_early(); 559 560 local_irq_disable(); 561 early_boot_irqs_off(); 562 early_init_irq_lock_class(); 563 564/* 565 * Interrupts are still disabled. Do necessary setups, then 566 * enable them 567 */ 568 lock_kernel(); 569 tick_init(); 570 boot_cpu_init(); 571 page_address_init(); 572 printk(KERN_NOTICE "%s", linux_banner); 573 setup_arch(&command_line); 574 mm_init_owner(&init_mm, &init_task); 575 setup_command_line(command_line); 576 setup_nr_cpu_ids(); 577 setup_per_cpu_areas(); 578 smp_prepare_boot_cpu(); /* arch-specific boot-cpu hooks */ 579 580 build_all_zonelists(NULL); 581 page_alloc_init(); 582 583 printk(KERN_NOTICE "Kernel command line: %s\n", boot_command_line); 584 parse_early_param(); 585 parse_args("Booting kernel", static_command_line, __start___param, 586 __stop___param - __start___param, 587 &unknown_bootoption); 588 /* 589 * These use large bootmem allocations and must precede 590 * kmem_cache_init() 591 */ 592 pidhash_init(); 593 vfs_caches_init_early(); 594 sort_main_extable(); 595 trap_init(); 596 mm_init(); 597 /* 598 * Set up the scheduler prior starting any interrupts (such as the 599 * timer interrupt). Full topology setup happens at smp_init() 600 * time - but meanwhile we still have a functioning scheduler. 601 */ 602 sched_init(); 603 /* 604 * Disable preemption - early bootup scheduling is extremely 605 * fragile until we cpu_idle() for the first time. 606 */ 607 preempt_disable(); 608 if (!irqs_disabled()) { 609 printk(KERN_WARNING "start_kernel(): bug: interrupts were " 610 "enabled *very* early, fixing it\n"); 611 local_irq_disable(); 612 } 613 rcu_init(); 614 radix_tree_init(); 615 /* init some links before init_ISA_irqs() */ 616 early_irq_init(); 617 init_IRQ(); 618 prio_tree_init(); 619 init_timers(); 620 hrtimers_init(); 621 softirq_init(); 622 timekeeping_init(); 623 time_init(); 624 profile_init(); 625 if (!irqs_disabled()) 626 printk(KERN_CRIT "start_kernel(): bug: interrupts were " 627 "enabled early\n"); 628 early_boot_irqs_on(); 629 local_irq_enable(); 630 631 /* Interrupts are enabled now so all GFP allocations are safe. */ 632 gfp_allowed_mask = __GFP_BITS_MASK; 633 634 kmem_cache_init_late(); 635 636 /* 637 * HACK ALERT! This is early. We're enabling the console before 638 * we've done PCI setups etc, and console_init() must be aware of 639 * this. But we do want output early, in case something goes wrong. 640 */ 641 console_init(); 642 if (panic_later) 643 panic(panic_later, panic_param); 644 645 lockdep_info(); 646 647 /* 648 * Need to run this when irqs are enabled, because it wants 649 * to self-test [hard/soft]-irqs on/off lock inversion bugs 650 * too: 651 */ 652 locking_selftest(); 653 654#ifdef CONFIG_BLK_DEV_INITRD 655 if (initrd_start && !initrd_below_start_ok && 656 page_to_pfn(virt_to_page((void *)initrd_start)) < min_low_pfn) { 657 printk(KERN_CRIT "initrd overwritten (0x%08lx < 0x%08lx) - " 658 "disabling it.\n", 659 page_to_pfn(virt_to_page((void *)initrd_start)), 660 min_low_pfn); 661 initrd_start = 0; 662 } 663#endif 664 page_cgroup_init(); 665 enable_debug_pagealloc(); 666 kmemtrace_init(); 667 kmemleak_init(); 668 debug_objects_mem_init(); 669 idr_init_cache(); 670 setup_per_cpu_pageset(); 671 numa_policy_init(); 672 if (late_time_init) 673 late_time_init(); 674 sched_clock_init(); 675 calibrate_delay(); 676 pidmap_init(); 677 anon_vma_init(); 678#ifdef CONFIG_X86 679 if (efi_enabled) 680 efi_enter_virtual_mode(); 681#endif 682 thread_info_cache_init(); 683 cred_init(); 684 fork_init(totalram_pages); 685 proc_caches_init(); 686 buffer_init(); 687 key_init(); 688 security_init(); 689 dbg_late_init(); 690 vfs_caches_init(totalram_pages); 691 signals_init(); 692 /* rootfs populating might need page-writeback */ 693 page_writeback_init(); 694#ifdef CONFIG_PROC_FS 695 proc_root_init(); 696#endif 697 cgroup_init(); 698 cpuset_init(); 699 taskstats_init_early(); 700 delayacct_init(); 701 702 check_bugs(); 703 704 acpi_early_init(); /* before LAPIC and SMP init */ 705 sfi_init_late(); 706 707 ftrace_init(); 708 709 /* Do the rest non-__init'ed, we're now alive */ 710 rest_init(); 711} 712 713/* Call all constructor functions linked into the kernel. */ 714static void __init do_ctors(void) 715{ 716#ifdef CONFIG_CONSTRUCTORS 717 ctor_fn_t *fn = (ctor_fn_t *) __ctors_start; 718 719 for (; fn < (ctor_fn_t *) __ctors_end; fn++) 720 (*fn)(); 721#endif 722} 723 724int initcall_debug; 725core_param(initcall_debug, initcall_debug, bool, 0644); 726 727static char msgbuf[64]; 728static struct boot_trace_call call; 729static struct boot_trace_ret ret; 730 731int do_one_initcall(initcall_t fn) 732{ 733 int count = preempt_count(); 734 ktime_t calltime, delta, rettime; 735 736 if (initcall_debug) { 737 call.caller = task_pid_nr(current); 738 printk("calling %pF @ %i\n", fn, call.caller); 739 calltime = ktime_get(); 740 trace_boot_call(&call, fn); 741 enable_boot_trace(); 742 } 743 744 ret.result = fn(); 745 746 if (initcall_debug) { 747 disable_boot_trace(); 748 rettime = ktime_get(); 749 delta = ktime_sub(rettime, calltime); 750 ret.duration = (unsigned long long) ktime_to_ns(delta) >> 10; 751 trace_boot_ret(&ret, fn); 752 printk("initcall %pF returned %d after %Ld usecs\n", fn, 753 ret.result, ret.duration); 754 } 755 756 msgbuf[0] = 0; 757 758 if (ret.result && ret.result != -ENODEV && initcall_debug) 759 sprintf(msgbuf, "error code %d ", ret.result); 760 761 if (preempt_count() != count) { 762 strlcat(msgbuf, "preemption imbalance ", sizeof(msgbuf)); 763 preempt_count() = count; 764 } 765 if (irqs_disabled()) { 766 strlcat(msgbuf, "disabled interrupts ", sizeof(msgbuf)); 767 local_irq_enable(); 768 } 769 if (msgbuf[0]) { 770 printk("initcall %pF returned with %s\n", fn, msgbuf); 771 } 772 773 return ret.result; 774} 775 776 777extern initcall_t __initcall_start[], __initcall_end[], __early_initcall_end[]; 778 779static void __init do_initcalls(void) 780{ 781 initcall_t *fn; 782 783 for (fn = __early_initcall_end; fn < __initcall_end; fn++) 784 do_one_initcall(*fn); 785 786 /* Make sure there is no pending stuff from the initcall sequence */ 787 flush_scheduled_work(); 788} 789 790/* 791 * Ok, the machine is now initialized. None of the devices 792 * have been touched yet, but the CPU subsystem is up and 793 * running, and memory and process management works. 794 * 795 * Now we can finally start doing some real work.. 796 */ 797static void __init do_basic_setup(void) 798{ 799 init_workqueues(); 800 cpuset_init_smp(); 801 usermodehelper_init(); 802 init_tmpfs(); 803 driver_init(); 804 init_irq_proc(); 805 do_ctors(); 806 do_initcalls(); 807} 808 809static void __init do_pre_smp_initcalls(void) 810{ 811 initcall_t *fn; 812 813 for (fn = __initcall_start; fn < __early_initcall_end; fn++) 814 do_one_initcall(*fn); 815} 816 817static void run_init_process(char *init_filename) 818{ 819 argv_init[0] = init_filename; 820 kernel_execve(init_filename, argv_init, envp_init); 821} 822 823/* This is a non __init function. Force it to be noinline otherwise gcc 824 * makes it inline to init() and it becomes part of init.text section 825 */ 826static noinline int init_post(void) 827 __releases(kernel_lock) 828{ 829 /* need to finish all async __init code before freeing the memory */ 830 async_synchronize_full(); 831 free_initmem(); 832 unlock_kernel(); 833 mark_rodata_ro(); 834 system_state = SYSTEM_RUNNING; 835 numa_default_policy(); 836 837 838 current->signal->flags |= SIGNAL_UNKILLABLE; 839 840 if (ramdisk_execute_command) { 841 run_init_process(ramdisk_execute_command); 842 printk(KERN_WARNING "Failed to execute %s\n", 843 ramdisk_execute_command); 844 } 845 846 /* 847 * We try each of these until one succeeds. 848 * 849 * The Bourne shell can be used instead of init if we are 850 * trying to recover a really broken machine. 851 */ 852 if (execute_command) { 853 run_init_process(execute_command); 854 printk(KERN_WARNING "Failed to execute %s. Attempting " 855 "defaults...\n", execute_command); 856 } 857 run_init_process("/sbin/init"); 858 run_init_process("/etc/init"); 859 run_init_process("/bin/init"); 860 run_init_process("/bin/sh"); 861 862 panic("No init found. Try passing init= option to kernel. " 863 "See Linux Documentation/init.txt for guidance."); 864} 865 866static int __init kernel_init(void * unused) 867{ 868 /* 869 * Wait until kthreadd is all set-up. 870 */ 871 wait_for_completion(&kthreadd_done); 872 lock_kernel(); 873 874 /* 875 * init can allocate pages on any node 876 */ 877 set_mems_allowed(node_states[N_HIGH_MEMORY]); 878 /* 879 * init can run on any cpu. 880 */ 881 set_cpus_allowed_ptr(current, cpu_all_mask); 882 /* 883 * Tell the world that we're going to be the grim 884 * reaper of innocent orphaned children. 885 * 886 * We don't want people to have to make incorrect 887 * assumptions about where in the task array this 888 * can be found. 889 */ 890 init_pid_ns.child_reaper = current; 891 892 cad_pid = task_pid(current); 893 894 smp_prepare_cpus(setup_max_cpus); 895 896 do_pre_smp_initcalls(); 897 start_boot_trace(); 898 899 smp_init(); 900 sched_init_smp(); 901 902 do_basic_setup(); 903 904 /* Open the /dev/console on the rootfs, this should never fail */ 905 if (sys_open((const char __user *) "/dev/console", O_RDWR, 0) < 0) 906 printk(KERN_WARNING "Warning: unable to open an initial console.\n"); 907 908 (void) sys_dup(0); 909 (void) sys_dup(0); 910 /* 911 * check if there is an early userspace init. If yes, let it do all 912 * the work 913 */ 914 915 if (!ramdisk_execute_command) 916 ramdisk_execute_command = "/init"; 917 918 if (sys_access((const char __user *) ramdisk_execute_command, 0) != 0) { 919 ramdisk_execute_command = NULL; 920 prepare_namespace(); 921 } 922 923 /* 924 * Ok, we have completed the initial bootup, and 925 * we're essentially up and running. Get rid of the 926 * initmem segments and start the user-mode stuff.. 927 */ 928 929 init_post(); 930 return 0; 931}