Loading include/trace/events/walt.h +2 −2 Original line number Diff line number Diff line Loading @@ -239,8 +239,8 @@ TRACE_EVENT(sched_update_task_ravg, __field(u64, nt_cs) __field(u64, nt_ps) __field(u32, active_windows) __field(u8, curr_top) __field(u8, prev_top) __field(u32, curr_top) __field(u32, prev_top) ), TP_fast_assign( Loading kernel/sched/sched.h +0 −2 Original line number Diff line number Diff line Loading @@ -983,7 +983,6 @@ struct rq { struct cpumask freq_domain_cpumask; struct walt_sched_stats walt_stats; int cstate, wakeup_latency, wakeup_energy; u64 window_start; s64 cum_window_start; unsigned long walt_flags; Loading @@ -991,7 +990,6 @@ struct rq { u64 cur_irqload; u64 avg_irqload; u64 irqload_ts; unsigned int static_cpu_pwr_cost; struct task_struct *ed_task; struct cpu_cycle cc; u64 old_busy_time, old_busy_time_group; Loading kernel/sched/walt.c +0 −15 Original line number Diff line number Diff line Loading @@ -114,23 +114,12 @@ __read_mostly unsigned int sysctl_sched_cpu_high_irqload = (10 * NSEC_PER_MSEC); unsigned int sysctl_sched_walt_rotate_big_tasks; unsigned int walt_rotation_enabled; /* * sched_window_stats_policy and sched_ravg_hist_size have a 'sysctl' copy * associated with them. This is required for atomic update of those variables * when being modifed via sysctl interface. * * IMPORTANT: Initialize both copies to same value!! */ __read_mostly unsigned int sched_ravg_hist_size = 5; __read_mostly unsigned int sysctl_sched_ravg_hist_size = 5; static __read_mostly unsigned int sched_io_is_busy = 1; __read_mostly unsigned int sched_window_stats_policy = WINDOW_STATS_MAX_RECENT_AVG; __read_mostly unsigned int sysctl_sched_window_stats_policy = WINDOW_STATS_MAX_RECENT_AVG; /* Window size (in ns) */ __read_mostly unsigned int sched_ravg_window = MIN_SCHED_RAVG_WINDOW; Loading Loading @@ -3354,12 +3343,8 @@ void walt_sched_init_rq(struct rq *rq) rq->cur_irqload = 0; rq->avg_irqload = 0; rq->irqload_ts = 0; rq->static_cpu_pwr_cost = 0; rq->cc.cycles = 1; rq->cc.time = 1; rq->cstate = 0; rq->wakeup_latency = 0; rq->wakeup_energy = 0; /* * All cpus part of same cluster by default. This avoids the Loading Loading
include/trace/events/walt.h +2 −2 Original line number Diff line number Diff line Loading @@ -239,8 +239,8 @@ TRACE_EVENT(sched_update_task_ravg, __field(u64, nt_cs) __field(u64, nt_ps) __field(u32, active_windows) __field(u8, curr_top) __field(u8, prev_top) __field(u32, curr_top) __field(u32, prev_top) ), TP_fast_assign( Loading
kernel/sched/sched.h +0 −2 Original line number Diff line number Diff line Loading @@ -983,7 +983,6 @@ struct rq { struct cpumask freq_domain_cpumask; struct walt_sched_stats walt_stats; int cstate, wakeup_latency, wakeup_energy; u64 window_start; s64 cum_window_start; unsigned long walt_flags; Loading @@ -991,7 +990,6 @@ struct rq { u64 cur_irqload; u64 avg_irqload; u64 irqload_ts; unsigned int static_cpu_pwr_cost; struct task_struct *ed_task; struct cpu_cycle cc; u64 old_busy_time, old_busy_time_group; Loading
kernel/sched/walt.c +0 −15 Original line number Diff line number Diff line Loading @@ -114,23 +114,12 @@ __read_mostly unsigned int sysctl_sched_cpu_high_irqload = (10 * NSEC_PER_MSEC); unsigned int sysctl_sched_walt_rotate_big_tasks; unsigned int walt_rotation_enabled; /* * sched_window_stats_policy and sched_ravg_hist_size have a 'sysctl' copy * associated with them. This is required for atomic update of those variables * when being modifed via sysctl interface. * * IMPORTANT: Initialize both copies to same value!! */ __read_mostly unsigned int sched_ravg_hist_size = 5; __read_mostly unsigned int sysctl_sched_ravg_hist_size = 5; static __read_mostly unsigned int sched_io_is_busy = 1; __read_mostly unsigned int sched_window_stats_policy = WINDOW_STATS_MAX_RECENT_AVG; __read_mostly unsigned int sysctl_sched_window_stats_policy = WINDOW_STATS_MAX_RECENT_AVG; /* Window size (in ns) */ __read_mostly unsigned int sched_ravg_window = MIN_SCHED_RAVG_WINDOW; Loading Loading @@ -3354,12 +3343,8 @@ void walt_sched_init_rq(struct rq *rq) rq->cur_irqload = 0; rq->avg_irqload = 0; rq->irqload_ts = 0; rq->static_cpu_pwr_cost = 0; rq->cc.cycles = 1; rq->cc.time = 1; rq->cstate = 0; rq->wakeup_latency = 0; rq->wakeup_energy = 0; /* * All cpus part of same cluster by default. This avoids the Loading