Loading arch/sh/include/asm/fpu.h +1 −0 Original line number Diff line number Diff line Loading @@ -25,6 +25,7 @@ void fpu_state_restore(struct pt_regs *regs); #define save_fpu(tsk) do { } while (0) #define release_fpu(regs) do { } while (0) #define grab_fpu(regs) do { } while (0) #define fpu_state_restore(regs) do { } while (0) #endif Loading arch/sh/kernel/process_32.c +3 −7 Original line number Diff line number Diff line Loading @@ -300,13 +300,11 @@ __switch_to(struct task_struct *prev, struct task_struct *next) { struct thread_struct *next_t = &next->thread; #if defined(CONFIG_SH_FPU) unlazy_fpu(prev, task_pt_regs(prev)); /* we're going to use this soon, after a few expensive things */ if (next->fpu_counter > 5) prefetch(&next_t->fpu.hard); #endif #ifdef CONFIG_MMU /* Loading Loading @@ -337,15 +335,13 @@ __switch_to(struct task_struct *prev, struct task_struct *next) #endif } #if defined(CONFIG_SH_FPU) /* If the task has used fpu the last 5 timeslices, just do a full /* * If the task has used fpu the last 5 timeslices, just do a full * restore of the math state immediately to avoid the trap; the * chances of needing FPU soon are obviously high now */ if (next->fpu_counter > 5) { if (next->fpu_counter > 5) fpu_state_restore(task_pt_regs(next)); } #endif return prev; } Loading Loading
arch/sh/include/asm/fpu.h +1 −0 Original line number Diff line number Diff line Loading @@ -25,6 +25,7 @@ void fpu_state_restore(struct pt_regs *regs); #define save_fpu(tsk) do { } while (0) #define release_fpu(regs) do { } while (0) #define grab_fpu(regs) do { } while (0) #define fpu_state_restore(regs) do { } while (0) #endif Loading
arch/sh/kernel/process_32.c +3 −7 Original line number Diff line number Diff line Loading @@ -300,13 +300,11 @@ __switch_to(struct task_struct *prev, struct task_struct *next) { struct thread_struct *next_t = &next->thread; #if defined(CONFIG_SH_FPU) unlazy_fpu(prev, task_pt_regs(prev)); /* we're going to use this soon, after a few expensive things */ if (next->fpu_counter > 5) prefetch(&next_t->fpu.hard); #endif #ifdef CONFIG_MMU /* Loading Loading @@ -337,15 +335,13 @@ __switch_to(struct task_struct *prev, struct task_struct *next) #endif } #if defined(CONFIG_SH_FPU) /* If the task has used fpu the last 5 timeslices, just do a full /* * If the task has used fpu the last 5 timeslices, just do a full * restore of the math state immediately to avoid the trap; the * chances of needing FPU soon are obviously high now */ if (next->fpu_counter > 5) { if (next->fpu_counter > 5) fpu_state_restore(task_pt_regs(next)); } #endif return prev; } Loading