Loading kernel/sched/fair.c +0 −2 Original line number Diff line number Diff line Loading @@ -5506,7 +5506,6 @@ static void dequeue_task_fair(struct rq *rq, struct task_struct *p, int flags) break; cfs_rq->h_nr_running--; cfs_rq->idle_h_nr_running -= idle_h_nr_running; walt_dec_cfs_rq_stats(cfs_rq, p); /* Don't dequeue parent if it has other entities besides us */ if (cfs_rq->load.weight) { Loading @@ -5527,7 +5526,6 @@ static void dequeue_task_fair(struct rq *rq, struct task_struct *p, int flags) cfs_rq = cfs_rq_of(se); cfs_rq->h_nr_running--; cfs_rq->idle_h_nr_running -= idle_h_nr_running; walt_dec_cfs_rq_stats(cfs_rq, p); if (cfs_rq_throttled(cfs_rq)) break; Loading kernel/sched/sched.h +0 −22 Original line number Diff line number Diff line Loading @@ -3273,28 +3273,6 @@ struct sched_avg_stats { }; extern void sched_get_nr_running_avg(struct sched_avg_stats *stats); #if defined(CONFIG_SCHED_WALT) && defined(CONFIG_CFS_BANDWIDTH) extern void walt_init_cfs_rq_stats(struct cfs_rq *cfs_rq); extern void walt_inc_cfs_rq_stats(struct cfs_rq *cfs_rq, struct task_struct *p); extern void walt_dec_cfs_rq_stats(struct cfs_rq *cfs_rq, struct task_struct *p); extern void walt_inc_throttled_cfs_rq_stats(struct walt_sched_stats *stats, struct cfs_rq *cfs_rq); extern void walt_dec_throttled_cfs_rq_stats(struct walt_sched_stats *stats, struct cfs_rq *cfs_rq); #else static inline void walt_init_cfs_rq_stats(struct cfs_rq *cfs_rq) {} static inline void walt_inc_cfs_rq_stats(struct cfs_rq *cfs_rq, struct task_struct *p) {} static inline void walt_dec_cfs_rq_stats(struct cfs_rq *cfs_rq, struct task_struct *p) {} #define walt_inc_throttled_cfs_rq_stats(...) #define walt_dec_throttled_cfs_rq_stats(...) #endif #ifdef CONFIG_SMP #ifdef CONFIG_SCHED_WALT extern int group_balance_cpu_not_isolated(struct sched_group *sg); Loading Loading
kernel/sched/fair.c +0 −2 Original line number Diff line number Diff line Loading @@ -5506,7 +5506,6 @@ static void dequeue_task_fair(struct rq *rq, struct task_struct *p, int flags) break; cfs_rq->h_nr_running--; cfs_rq->idle_h_nr_running -= idle_h_nr_running; walt_dec_cfs_rq_stats(cfs_rq, p); /* Don't dequeue parent if it has other entities besides us */ if (cfs_rq->load.weight) { Loading @@ -5527,7 +5526,6 @@ static void dequeue_task_fair(struct rq *rq, struct task_struct *p, int flags) cfs_rq = cfs_rq_of(se); cfs_rq->h_nr_running--; cfs_rq->idle_h_nr_running -= idle_h_nr_running; walt_dec_cfs_rq_stats(cfs_rq, p); if (cfs_rq_throttled(cfs_rq)) break; Loading
kernel/sched/sched.h +0 −22 Original line number Diff line number Diff line Loading @@ -3273,28 +3273,6 @@ struct sched_avg_stats { }; extern void sched_get_nr_running_avg(struct sched_avg_stats *stats); #if defined(CONFIG_SCHED_WALT) && defined(CONFIG_CFS_BANDWIDTH) extern void walt_init_cfs_rq_stats(struct cfs_rq *cfs_rq); extern void walt_inc_cfs_rq_stats(struct cfs_rq *cfs_rq, struct task_struct *p); extern void walt_dec_cfs_rq_stats(struct cfs_rq *cfs_rq, struct task_struct *p); extern void walt_inc_throttled_cfs_rq_stats(struct walt_sched_stats *stats, struct cfs_rq *cfs_rq); extern void walt_dec_throttled_cfs_rq_stats(struct walt_sched_stats *stats, struct cfs_rq *cfs_rq); #else static inline void walt_init_cfs_rq_stats(struct cfs_rq *cfs_rq) {} static inline void walt_inc_cfs_rq_stats(struct cfs_rq *cfs_rq, struct task_struct *p) {} static inline void walt_dec_cfs_rq_stats(struct cfs_rq *cfs_rq, struct task_struct *p) {} #define walt_inc_throttled_cfs_rq_stats(...) #define walt_dec_throttled_cfs_rq_stats(...) #endif #ifdef CONFIG_SMP #ifdef CONFIG_SCHED_WALT extern int group_balance_cpu_not_isolated(struct sched_group *sg); Loading