Loading arch/arm/include/asm/io.h +2 −2 Original line number Diff line number Diff line Loading @@ -64,7 +64,7 @@ extern void __raw_readsl(const void __iomem *addr, void *data, int longlen); static inline void __raw_writew(u16 val, volatile void __iomem *addr) { asm volatile("strh %1, %0" : "+Qo" (*(volatile u16 __force *)addr) : "+Q" (*(volatile u16 __force *)addr) : "r" (val)); } Loading @@ -72,7 +72,7 @@ static inline u16 __raw_readw(const volatile void __iomem *addr) { u16 val; asm volatile("ldrh %1, %0" : "+Qo" (*(volatile u16 __force *)addr), : "+Q" (*(volatile u16 __force *)addr), "=r" (val)); return val; } Loading arch/arm/include/asm/sched_clock.h +0 −2 Original line number Diff line number Diff line Loading @@ -10,7 +10,5 @@ extern void sched_clock_postinit(void); extern void setup_sched_clock(u32 (*read)(void), int bits, unsigned long rate); extern void setup_sched_clock_needs_suspend(u32 (*read)(void), int bits, unsigned long rate); #endif arch/arm/include/asm/vfpmacros.h +6 −6 Original line number Diff line number Diff line Loading @@ -27,9 +27,9 @@ #if __LINUX_ARM_ARCH__ <= 6 ldr \tmp, =elf_hwcap @ may not have MVFR regs ldr \tmp, [\tmp, #0] tst \tmp, #HWCAP_VFPv3D16 ldceql p11, cr0, [\base],#32*4 @ FLDMIAD \base!, {d16-d31} addne \base, \base, #32*4 @ step over unused register space tst \tmp, #HWCAP_VFPD32 ldcnel p11, cr0, [\base],#32*4 @ FLDMIAD \base!, {d16-d31} addeq \base, \base, #32*4 @ step over unused register space #else VFPFMRX \tmp, MVFR0 @ Media and VFP Feature Register 0 and \tmp, \tmp, #MVFR0_A_SIMD_MASK @ A_SIMD field Loading @@ -51,9 +51,9 @@ #if __LINUX_ARM_ARCH__ <= 6 ldr \tmp, =elf_hwcap @ may not have MVFR regs ldr \tmp, [\tmp, #0] tst \tmp, #HWCAP_VFPv3D16 stceql p11, cr0, [\base],#32*4 @ FSTMIAD \base!, {d16-d31} addne \base, \base, #32*4 @ step over unused register space tst \tmp, #HWCAP_VFPD32 stcnel p11, cr0, [\base],#32*4 @ FSTMIAD \base!, {d16-d31} addeq \base, \base, #32*4 @ step over unused register space #else VFPFMRX \tmp, MVFR0 @ Media and VFP Feature Register 0 and \tmp, \tmp, #MVFR0_A_SIMD_MASK @ A_SIMD field Loading arch/arm/include/uapi/asm/hwcap.h +2 −1 Original line number Diff line number Diff line Loading @@ -18,11 +18,12 @@ #define HWCAP_THUMBEE (1 << 11) #define HWCAP_NEON (1 << 12) #define HWCAP_VFPv3 (1 << 13) #define HWCAP_VFPv3D16 (1 << 14) #define HWCAP_VFPv3D16 (1 << 14) /* also set for VFPv4-D16 */ #define HWCAP_TLS (1 << 15) #define HWCAP_VFPv4 (1 << 16) #define HWCAP_IDIVA (1 << 17) #define HWCAP_IDIVT (1 << 18) #define HWCAP_VFPD32 (1 << 19) /* set if VFP has 32 regs (not 16) */ #define HWCAP_IDIV (HWCAP_IDIVA | HWCAP_IDIVT) Loading arch/arm/kernel/sched_clock.c +4 −14 Original line number Diff line number Diff line Loading @@ -107,13 +107,6 @@ static void sched_clock_poll(unsigned long wrap_ticks) update_sched_clock(); } void __init setup_sched_clock_needs_suspend(u32 (*read)(void), int bits, unsigned long rate) { setup_sched_clock(read, bits, rate); cd.needs_suspend = true; } void __init setup_sched_clock(u32 (*read)(void), int bits, unsigned long rate) { unsigned long r, w; Loading Loading @@ -189,19 +182,16 @@ void __init sched_clock_postinit(void) static int sched_clock_suspend(void) { sched_clock_poll(sched_clock_timer.data); if (cd.needs_suspend) cd.suspended = true; return 0; } static void sched_clock_resume(void) { if (cd.needs_suspend) { cd.epoch_cyc = read_sched_clock(); cd.epoch_cyc_copy = cd.epoch_cyc; cd.suspended = false; } } static struct syscore_ops sched_clock_ops = { .suspend = sched_clock_suspend, Loading Loading
arch/arm/include/asm/io.h +2 −2 Original line number Diff line number Diff line Loading @@ -64,7 +64,7 @@ extern void __raw_readsl(const void __iomem *addr, void *data, int longlen); static inline void __raw_writew(u16 val, volatile void __iomem *addr) { asm volatile("strh %1, %0" : "+Qo" (*(volatile u16 __force *)addr) : "+Q" (*(volatile u16 __force *)addr) : "r" (val)); } Loading @@ -72,7 +72,7 @@ static inline u16 __raw_readw(const volatile void __iomem *addr) { u16 val; asm volatile("ldrh %1, %0" : "+Qo" (*(volatile u16 __force *)addr), : "+Q" (*(volatile u16 __force *)addr), "=r" (val)); return val; } Loading
arch/arm/include/asm/sched_clock.h +0 −2 Original line number Diff line number Diff line Loading @@ -10,7 +10,5 @@ extern void sched_clock_postinit(void); extern void setup_sched_clock(u32 (*read)(void), int bits, unsigned long rate); extern void setup_sched_clock_needs_suspend(u32 (*read)(void), int bits, unsigned long rate); #endif
arch/arm/include/asm/vfpmacros.h +6 −6 Original line number Diff line number Diff line Loading @@ -27,9 +27,9 @@ #if __LINUX_ARM_ARCH__ <= 6 ldr \tmp, =elf_hwcap @ may not have MVFR regs ldr \tmp, [\tmp, #0] tst \tmp, #HWCAP_VFPv3D16 ldceql p11, cr0, [\base],#32*4 @ FLDMIAD \base!, {d16-d31} addne \base, \base, #32*4 @ step over unused register space tst \tmp, #HWCAP_VFPD32 ldcnel p11, cr0, [\base],#32*4 @ FLDMIAD \base!, {d16-d31} addeq \base, \base, #32*4 @ step over unused register space #else VFPFMRX \tmp, MVFR0 @ Media and VFP Feature Register 0 and \tmp, \tmp, #MVFR0_A_SIMD_MASK @ A_SIMD field Loading @@ -51,9 +51,9 @@ #if __LINUX_ARM_ARCH__ <= 6 ldr \tmp, =elf_hwcap @ may not have MVFR regs ldr \tmp, [\tmp, #0] tst \tmp, #HWCAP_VFPv3D16 stceql p11, cr0, [\base],#32*4 @ FSTMIAD \base!, {d16-d31} addne \base, \base, #32*4 @ step over unused register space tst \tmp, #HWCAP_VFPD32 stcnel p11, cr0, [\base],#32*4 @ FSTMIAD \base!, {d16-d31} addeq \base, \base, #32*4 @ step over unused register space #else VFPFMRX \tmp, MVFR0 @ Media and VFP Feature Register 0 and \tmp, \tmp, #MVFR0_A_SIMD_MASK @ A_SIMD field Loading
arch/arm/include/uapi/asm/hwcap.h +2 −1 Original line number Diff line number Diff line Loading @@ -18,11 +18,12 @@ #define HWCAP_THUMBEE (1 << 11) #define HWCAP_NEON (1 << 12) #define HWCAP_VFPv3 (1 << 13) #define HWCAP_VFPv3D16 (1 << 14) #define HWCAP_VFPv3D16 (1 << 14) /* also set for VFPv4-D16 */ #define HWCAP_TLS (1 << 15) #define HWCAP_VFPv4 (1 << 16) #define HWCAP_IDIVA (1 << 17) #define HWCAP_IDIVT (1 << 18) #define HWCAP_VFPD32 (1 << 19) /* set if VFP has 32 regs (not 16) */ #define HWCAP_IDIV (HWCAP_IDIVA | HWCAP_IDIVT) Loading
arch/arm/kernel/sched_clock.c +4 −14 Original line number Diff line number Diff line Loading @@ -107,13 +107,6 @@ static void sched_clock_poll(unsigned long wrap_ticks) update_sched_clock(); } void __init setup_sched_clock_needs_suspend(u32 (*read)(void), int bits, unsigned long rate) { setup_sched_clock(read, bits, rate); cd.needs_suspend = true; } void __init setup_sched_clock(u32 (*read)(void), int bits, unsigned long rate) { unsigned long r, w; Loading Loading @@ -189,19 +182,16 @@ void __init sched_clock_postinit(void) static int sched_clock_suspend(void) { sched_clock_poll(sched_clock_timer.data); if (cd.needs_suspend) cd.suspended = true; return 0; } static void sched_clock_resume(void) { if (cd.needs_suspend) { cd.epoch_cyc = read_sched_clock(); cd.epoch_cyc_copy = cd.epoch_cyc; cd.suspended = false; } } static struct syscore_ops sched_clock_ops = { .suspend = sched_clock_suspend, Loading