Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

Merge branch 'fixes' of git://git.linaro.org/people/rmk/linux-arm

Pull arm fixes from Russell King:
"Not much here again.

The two most notable things here are the sched_clock() fix, which was
causing problems with the scheduling of threaded IRQs after a suspend
event, and the vfp fix, which afaik has only been seen on some older
OMAP boards. Nevertheless, both are fairly important fixes."

* 'fixes' of git://git.linaro.org/people/rmk/linux-arm:
ARM: 7569/1: mm: uninitialized warning corrections
ARM: 7567/1: io: avoid GCC's offsettable addressing modes for halfword accesses
ARM: 7566/1: vfp: fix save and restore when running on pre-VFPv3 and CONFIG_VFPv3 set
ARM: 7565/1: sched: stop sched_clock() during suspend

+21 -29
+2 -2
arch/arm/include/asm/io.h
··· 64 64 static inline void __raw_writew(u16 val, volatile void __iomem *addr) 65 65 { 66 66 asm volatile("strh %1, %0" 67 - : "+Qo" (*(volatile u16 __force *)addr) 67 + : "+Q" (*(volatile u16 __force *)addr) 68 68 : "r" (val)); 69 69 } 70 70 ··· 72 72 { 73 73 u16 val; 74 74 asm volatile("ldrh %1, %0" 75 - : "+Qo" (*(volatile u16 __force *)addr), 75 + : "+Q" (*(volatile u16 __force *)addr), 76 76 "=r" (val)); 77 77 return val; 78 78 }
-2
arch/arm/include/asm/sched_clock.h
··· 10 10 11 11 extern void sched_clock_postinit(void); 12 12 extern void setup_sched_clock(u32 (*read)(void), int bits, unsigned long rate); 13 - extern void setup_sched_clock_needs_suspend(u32 (*read)(void), int bits, 14 - unsigned long rate); 15 13 16 14 #endif
+6 -6
arch/arm/include/asm/vfpmacros.h
··· 27 27 #if __LINUX_ARM_ARCH__ <= 6 28 28 ldr \tmp, =elf_hwcap @ may not have MVFR regs 29 29 ldr \tmp, [\tmp, #0] 30 - tst \tmp, #HWCAP_VFPv3D16 31 - ldceql p11, cr0, [\base],#32*4 @ FLDMIAD \base!, {d16-d31} 32 - addne \base, \base, #32*4 @ step over unused register space 30 + tst \tmp, #HWCAP_VFPD32 31 + ldcnel p11, cr0, [\base],#32*4 @ FLDMIAD \base!, {d16-d31} 32 + addeq \base, \base, #32*4 @ step over unused register space 33 33 #else 34 34 VFPFMRX \tmp, MVFR0 @ Media and VFP Feature Register 0 35 35 and \tmp, \tmp, #MVFR0_A_SIMD_MASK @ A_SIMD field ··· 51 51 #if __LINUX_ARM_ARCH__ <= 6 52 52 ldr \tmp, =elf_hwcap @ may not have MVFR regs 53 53 ldr \tmp, [\tmp, #0] 54 - tst \tmp, #HWCAP_VFPv3D16 55 - stceql p11, cr0, [\base],#32*4 @ FSTMIAD \base!, {d16-d31} 56 - addne \base, \base, #32*4 @ step over unused register space 54 + tst \tmp, #HWCAP_VFPD32 55 + stcnel p11, cr0, [\base],#32*4 @ FSTMIAD \base!, {d16-d31} 56 + addeq \base, \base, #32*4 @ step over unused register space 57 57 #else 58 58 VFPFMRX \tmp, MVFR0 @ Media and VFP Feature Register 0 59 59 and \tmp, \tmp, #MVFR0_A_SIMD_MASK @ A_SIMD field
+2 -1
arch/arm/include/uapi/asm/hwcap.h
··· 18 18 #define HWCAP_THUMBEE (1 << 11) 19 19 #define HWCAP_NEON (1 << 12) 20 20 #define HWCAP_VFPv3 (1 << 13) 21 - #define HWCAP_VFPv3D16 (1 << 14) 21 + #define HWCAP_VFPv3D16 (1 << 14) /* also set for VFPv4-D16 */ 22 22 #define HWCAP_TLS (1 << 15) 23 23 #define HWCAP_VFPv4 (1 << 16) 24 24 #define HWCAP_IDIVA (1 << 17) 25 25 #define HWCAP_IDIVT (1 << 18) 26 + #define HWCAP_VFPD32 (1 << 19) /* set if VFP has 32 regs (not 16) */ 26 27 #define HWCAP_IDIV (HWCAP_IDIVA | HWCAP_IDIVT) 27 28 28 29
+4 -14
arch/arm/kernel/sched_clock.c
··· 107 107 update_sched_clock(); 108 108 } 109 109 110 - void __init setup_sched_clock_needs_suspend(u32 (*read)(void), int bits, 111 - unsigned long rate) 112 - { 113 - setup_sched_clock(read, bits, rate); 114 - cd.needs_suspend = true; 115 - } 116 - 117 110 void __init setup_sched_clock(u32 (*read)(void), int bits, unsigned long rate) 118 111 { 119 112 unsigned long r, w; ··· 182 189 static int sched_clock_suspend(void) 183 190 { 184 191 sched_clock_poll(sched_clock_timer.data); 185 - if (cd.needs_suspend) 186 - cd.suspended = true; 192 + cd.suspended = true; 187 193 return 0; 188 194 } 189 195 190 196 static void sched_clock_resume(void) 191 197 { 192 - if (cd.needs_suspend) { 193 - cd.epoch_cyc = read_sched_clock(); 194 - cd.epoch_cyc_copy = cd.epoch_cyc; 195 - cd.suspended = false; 196 - } 198 + cd.epoch_cyc = read_sched_clock(); 199 + cd.epoch_cyc_copy = cd.epoch_cyc; 200 + cd.suspended = false; 197 201 } 198 202 199 203 static struct syscore_ops sched_clock_ops = {
+1 -1
arch/arm/mm/alignment.c
··· 745 745 static int 746 746 do_alignment(unsigned long addr, unsigned int fsr, struct pt_regs *regs) 747 747 { 748 - union offset_union offset; 748 + union offset_union uninitialized_var(offset); 749 749 unsigned long instr = 0, instrptr; 750 750 int (*handler)(unsigned long addr, unsigned long instr, struct pt_regs *regs); 751 751 unsigned int type;
+6 -3
arch/arm/vfp/vfpmodule.c
··· 701 701 elf_hwcap |= HWCAP_VFPv3; 702 702 703 703 /* 704 - * Check for VFPv3 D16. CPUs in this configuration 705 - * only have 16 x 64bit registers. 704 + * Check for VFPv3 D16 and VFPv4 D16. CPUs in 705 + * this configuration only have 16 x 64bit 706 + * registers. 706 707 */ 707 708 if (((fmrx(MVFR0) & MVFR0_A_SIMD_MASK)) == 1) 708 - elf_hwcap |= HWCAP_VFPv3D16; 709 + elf_hwcap |= HWCAP_VFPv3D16; /* also v4-D16 */ 710 + else 711 + elf_hwcap |= HWCAP_VFPD32; 709 712 } 710 713 #endif 711 714 /*