Loading include/linux/sched.h +11 −3 Original line number Diff line number Diff line Loading @@ -184,6 +184,8 @@ extern void sched_get_nr_running_avg(int *avg, int *iowait_avg, int *big_avg, unsigned int *big_max_nr); extern unsigned int sched_get_cpu_util(int cpu); extern u64 sched_get_cpu_last_busy_time(int cpu); extern u32 sched_get_wake_up_idle(struct task_struct *p); extern int sched_set_wake_up_idle(struct task_struct *p, int wake_up_idle); #else static inline void sched_update_nr_prod(int cpu, long delta, bool inc) { Loading @@ -201,6 +203,15 @@ static inline u64 sched_get_cpu_last_busy_time(int cpu) { return 0; } static inline u32 sched_get_wake_up_idle(struct task_struct *p) { return 0; } static inline int sched_set_wake_up_idle(struct task_struct *p, int wake_up_idle) { return 0; } #endif extern void calc_global_load(unsigned long ticks); Loading Loading @@ -2724,9 +2735,6 @@ struct sched_load { unsigned long predicted_load; }; extern int sched_set_wake_up_idle(struct task_struct *p, int wake_up_idle); extern u32 sched_get_wake_up_idle(struct task_struct *p); struct cpu_cycle_counter_cb { u64 (*get_cpu_cycle_counter)(int cpu); }; Loading kernel/sched/fair.c +2 −0 Original line number Diff line number Diff line Loading @@ -2769,6 +2769,7 @@ u32 sched_get_wake_up_idle(struct task_struct *p) return !!enabled; } EXPORT_SYMBOL(sched_get_wake_up_idle); int sched_set_wake_up_idle(struct task_struct *p, int wake_up_idle) { Loading @@ -2781,6 +2782,7 @@ int sched_set_wake_up_idle(struct task_struct *p, int wake_up_idle) return 0; } EXPORT_SYMBOL(sched_set_wake_up_idle); /* Precomputed fixed inverse multiplies for multiplication by y^n */ static const u32 runnable_avg_yN_inv[] = { Loading Loading
include/linux/sched.h +11 −3 Original line number Diff line number Diff line Loading @@ -184,6 +184,8 @@ extern void sched_get_nr_running_avg(int *avg, int *iowait_avg, int *big_avg, unsigned int *big_max_nr); extern unsigned int sched_get_cpu_util(int cpu); extern u64 sched_get_cpu_last_busy_time(int cpu); extern u32 sched_get_wake_up_idle(struct task_struct *p); extern int sched_set_wake_up_idle(struct task_struct *p, int wake_up_idle); #else static inline void sched_update_nr_prod(int cpu, long delta, bool inc) { Loading @@ -201,6 +203,15 @@ static inline u64 sched_get_cpu_last_busy_time(int cpu) { return 0; } static inline u32 sched_get_wake_up_idle(struct task_struct *p) { return 0; } static inline int sched_set_wake_up_idle(struct task_struct *p, int wake_up_idle) { return 0; } #endif extern void calc_global_load(unsigned long ticks); Loading Loading @@ -2724,9 +2735,6 @@ struct sched_load { unsigned long predicted_load; }; extern int sched_set_wake_up_idle(struct task_struct *p, int wake_up_idle); extern u32 sched_get_wake_up_idle(struct task_struct *p); struct cpu_cycle_counter_cb { u64 (*get_cpu_cycle_counter)(int cpu); }; Loading
kernel/sched/fair.c +2 −0 Original line number Diff line number Diff line Loading @@ -2769,6 +2769,7 @@ u32 sched_get_wake_up_idle(struct task_struct *p) return !!enabled; } EXPORT_SYMBOL(sched_get_wake_up_idle); int sched_set_wake_up_idle(struct task_struct *p, int wake_up_idle) { Loading @@ -2781,6 +2782,7 @@ int sched_set_wake_up_idle(struct task_struct *p, int wake_up_idle) return 0; } EXPORT_SYMBOL(sched_set_wake_up_idle); /* Precomputed fixed inverse multiplies for multiplication by y^n */ static const u32 runnable_avg_yN_inv[] = { Loading