Merge branch 'merge' of ssh://master.kernel.org/pub/scm/linux/kernel/git/jwboyer/powerpc-4xx into merge

+20 -3
+6 -1
arch/powerpc/kernel/cpu_setup_44x.S
··· 34 blr 35 _GLOBAL(__setup_cpu_460ex) 36 _GLOBAL(__setup_cpu_460gt) 37 - b __init_fpu_44x 38 _GLOBAL(__setup_cpu_440gx) 39 _GLOBAL(__setup_cpu_440spe) 40 b __fixup_440A_mcheck
··· 34 blr 35 _GLOBAL(__setup_cpu_460ex) 36 _GLOBAL(__setup_cpu_460gt) 37 + mflr r4 38 + bl __init_fpu_44x 39 + bl __fixup_440A_mcheck 40 + mtlr r4 41 + blr 42 + 43 _GLOBAL(__setup_cpu_440gx) 44 _GLOBAL(__setup_cpu_440spe) 45 b __fixup_440A_mcheck
+14 -2
arch/powerpc/mm/40x_mmu.c
··· 93 94 unsigned long __init mmu_mapin_ram(void) 95 { 96 - unsigned long v, s; 97 phys_addr_t p; 98 99 v = KERNELBASE; ··· 130 s -= LARGE_PAGE_SIZE_4M; 131 } 132 133 - return total_lowmem - s; 134 }
··· 93 94 unsigned long __init mmu_mapin_ram(void) 95 { 96 + unsigned long v, s, mapped; 97 phys_addr_t p; 98 99 v = KERNELBASE; ··· 130 s -= LARGE_PAGE_SIZE_4M; 131 } 132 133 + mapped = total_lowmem - s; 134 + 135 + /* If the size of RAM is not an exact power of two, we may not 136 + * have covered RAM in its entirety with 16 and 4 MiB 137 + * pages. Consequently, restrict the top end of RAM currently 138 + * allocable so that calls to the LMB to allocate PTEs for "tail" 139 + * coverage with normal-sized pages (or other reasons) do not 140 + * attempt to allocate outside the allowed range. 141 + */ 142 + 143 + __initial_memory_limit_addr = memstart_addr + mapped; 144 + 145 + return mapped; 146 }