Loading kernel/sched/sched.h +6 −2 Original line number Diff line number Diff line Loading @@ -2338,7 +2338,7 @@ void note_task_waking(struct task_struct *p, u64 wallclock); #else /* CONFIG_SCHED_WALT */ static inline u64 sched_ktime_clock(void) { return 0; return sched_clock(); } static inline void note_task_waking(struct task_struct *p, u64 wallclock) { } #endif /* CONFIG_SCHED_WALT */ Loading Loading @@ -2371,16 +2371,20 @@ DECLARE_PER_CPU(struct update_util_data *, cpufreq_update_util_data); static inline void cpufreq_update_util(struct rq *rq, unsigned int flags) { struct update_util_data *data; u64 clock; #ifdef CONFIG_SCHED_WALT if (!(flags & SCHED_CPUFREQ_WALT)) return; clock = sched_ktime_clock(); #else clock = rq_clock(rq); #endif data = rcu_dereference_sched(*per_cpu_ptr(&cpufreq_update_util_data, cpu_of(rq))); if (data) data->func(data, sched_ktime_clock(), flags); data->func(data, clock, flags); } static inline void cpufreq_update_this_cpu(struct rq *rq, unsigned int flags) Loading Loading
kernel/sched/sched.h +6 −2 Original line number Diff line number Diff line Loading @@ -2338,7 +2338,7 @@ void note_task_waking(struct task_struct *p, u64 wallclock); #else /* CONFIG_SCHED_WALT */ static inline u64 sched_ktime_clock(void) { return 0; return sched_clock(); } static inline void note_task_waking(struct task_struct *p, u64 wallclock) { } #endif /* CONFIG_SCHED_WALT */ Loading Loading @@ -2371,16 +2371,20 @@ DECLARE_PER_CPU(struct update_util_data *, cpufreq_update_util_data); static inline void cpufreq_update_util(struct rq *rq, unsigned int flags) { struct update_util_data *data; u64 clock; #ifdef CONFIG_SCHED_WALT if (!(flags & SCHED_CPUFREQ_WALT)) return; clock = sched_ktime_clock(); #else clock = rq_clock(rq); #endif data = rcu_dereference_sched(*per_cpu_ptr(&cpufreq_update_util_data, cpu_of(rq))); if (data) data->func(data, sched_ktime_clock(), flags); data->func(data, clock, flags); } static inline void cpufreq_update_this_cpu(struct rq *rq, unsigned int flags) Loading