Loading kernel/sched/walt.c +3 −3 Original line number Diff line number Diff line Loading @@ -1376,7 +1376,7 @@ static void rollover_task_window(struct task_struct *p, bool full_window) p->ravg.curr_window_cpu[i] = 0; } if (p->ravg.active_time < NEW_TASK_ACTIVE_TIME) if (is_new_task(p)) p->ravg.active_time += p->ravg.last_win_size; } Loading Loading @@ -1515,8 +1515,6 @@ static void update_cpu_busy_time(struct task_struct *p, struct rq *rq, if (new_window) full_window = (window_start - mark_start) >= window_size; new_task = is_new_task(p); /* * Handle per-task window rollover. We don't care about the idle * task or exiting tasks. Loading @@ -1526,6 +1524,8 @@ static void update_cpu_busy_time(struct task_struct *p, struct rq *rq, rollover_task_window(p, full_window); } new_task = is_new_task(p); if (p_is_curr_task && new_window) { rollover_cpu_window(rq, full_window); rollover_top_tasks(rq, full_window); Loading kernel/sched/walt.h +1 −1 Original line number Diff line number Diff line Loading @@ -194,7 +194,7 @@ scale_load_to_freq(u64 load, unsigned int src_freq, unsigned int dst_freq) static inline bool is_new_task(struct task_struct *p) { return p->ravg.active_time <= NEW_TASK_ACTIVE_TIME; return p->ravg.active_time < NEW_TASK_ACTIVE_TIME; } static inline void clear_top_tasks_table(u8 *table) Loading Loading
kernel/sched/walt.c +3 −3 Original line number Diff line number Diff line Loading @@ -1376,7 +1376,7 @@ static void rollover_task_window(struct task_struct *p, bool full_window) p->ravg.curr_window_cpu[i] = 0; } if (p->ravg.active_time < NEW_TASK_ACTIVE_TIME) if (is_new_task(p)) p->ravg.active_time += p->ravg.last_win_size; } Loading Loading @@ -1515,8 +1515,6 @@ static void update_cpu_busy_time(struct task_struct *p, struct rq *rq, if (new_window) full_window = (window_start - mark_start) >= window_size; new_task = is_new_task(p); /* * Handle per-task window rollover. We don't care about the idle * task or exiting tasks. Loading @@ -1526,6 +1524,8 @@ static void update_cpu_busy_time(struct task_struct *p, struct rq *rq, rollover_task_window(p, full_window); } new_task = is_new_task(p); if (p_is_curr_task && new_window) { rollover_cpu_window(rq, full_window); rollover_top_tasks(rq, full_window); Loading
kernel/sched/walt.h +1 −1 Original line number Diff line number Diff line Loading @@ -194,7 +194,7 @@ scale_load_to_freq(u64 load, unsigned int src_freq, unsigned int dst_freq) static inline bool is_new_task(struct task_struct *p) { return p->ravg.active_time <= NEW_TASK_ACTIVE_TIME; return p->ravg.active_time < NEW_TASK_ACTIVE_TIME; } static inline void clear_top_tasks_table(u8 *table) Loading