Loading include/trace/events/sched.h +2 −1 Original line number Diff line number Diff line Loading @@ -1033,7 +1033,8 @@ TRACE_EVENT(walt_update_history, __entry->evt = evt; __entry->demand = p->ravg.demand; __entry->walt_avg = (__entry->demand << 10); do_div(__entry->walt_avg, walt_ravg_window); __entry->walt_avg = div_u64(__entry->walt_avg, walt_ravg_window); __entry->pelt_avg = p->se.avg.util_avg; memcpy(__entry->hist, p->ravg.sum_history, RAVG_HIST_SIZE_MAX * sizeof(u32)); Loading kernel/sched/sched.h +1 −1 Original line number Diff line number Diff line Loading @@ -1642,7 +1642,7 @@ static inline unsigned long __cpu_util(int cpu, int delta) #ifdef CONFIG_SCHED_WALT if (!walt_disabled && sysctl_sched_use_walt_cpu_util) { util = cpu_rq(cpu)->prev_runnable_sum << SCHED_CAPACITY_SHIFT; do_div(util, walt_ravg_window); util = div_u64(util, walt_ravg_window); } #endif delta += util; Loading Loading
include/trace/events/sched.h +2 −1 Original line number Diff line number Diff line Loading @@ -1033,7 +1033,8 @@ TRACE_EVENT(walt_update_history, __entry->evt = evt; __entry->demand = p->ravg.demand; __entry->walt_avg = (__entry->demand << 10); do_div(__entry->walt_avg, walt_ravg_window); __entry->walt_avg = div_u64(__entry->walt_avg, walt_ravg_window); __entry->pelt_avg = p->se.avg.util_avg; memcpy(__entry->hist, p->ravg.sum_history, RAVG_HIST_SIZE_MAX * sizeof(u32)); Loading
kernel/sched/sched.h +1 −1 Original line number Diff line number Diff line Loading @@ -1642,7 +1642,7 @@ static inline unsigned long __cpu_util(int cpu, int delta) #ifdef CONFIG_SCHED_WALT if (!walt_disabled && sysctl_sched_use_walt_cpu_util) { util = cpu_rq(cpu)->prev_runnable_sum << SCHED_CAPACITY_SHIFT; do_div(util, walt_ravg_window); util = div_u64(util, walt_ravg_window); } #endif delta += util; Loading