Loading include/linux/cpufreq.h +0 −9 Original line number Diff line number Diff line Loading @@ -540,15 +540,6 @@ static inline void cpufreq_policy_apply_limits(struct cpufreq_policy *policy) __cpufreq_driver_target(policy, policy->min, CPUFREQ_RELATION_L); } static inline void cpufreq_policy_apply_limits_fast(struct cpufreq_policy *policy) { if (policy->max < policy->cur) cpufreq_driver_fast_switch(policy, policy->max); else if (policy->min > policy->cur) cpufreq_driver_fast_switch(policy, policy->min); } /* Governor attribute set */ struct gov_attr_set { struct kobject kobj; Loading kernel/sched/cpufreq_schedutil.c +12 −4 Original line number Diff line number Diff line Loading @@ -1329,7 +1329,8 @@ static void sugov_stop(struct cpufreq_policy *policy) static void sugov_limits(struct cpufreq_policy *policy) { struct sugov_policy *sg_policy = policy->governor_data; unsigned long flags; unsigned long flags, now; unsigned int freq; if (!policy->fast_switch_enabled) { mutex_lock(&sg_policy->work_lock); Loading @@ -1341,9 +1342,16 @@ static void sugov_limits(struct cpufreq_policy *policy) mutex_unlock(&sg_policy->work_lock); } else { raw_spin_lock_irqsave(&sg_policy->update_lock, flags); sugov_track_cycles(sg_policy, sg_policy->policy->cur, ktime_get_ns()); cpufreq_policy_apply_limits_fast(policy); freq = policy->cur; now = ktime_get_ns(); /* * cpufreq_driver_resolve_freq() has a clamp, so we do not need * to do any sort of additional validation here. */ freq = cpufreq_driver_resolve_freq(policy, freq); sg_policy->cached_raw_freq = freq; sugov_fast_switch(sg_policy, now, freq); raw_spin_unlock_irqrestore(&sg_policy->update_lock, flags); } Loading Loading
include/linux/cpufreq.h +0 −9 Original line number Diff line number Diff line Loading @@ -540,15 +540,6 @@ static inline void cpufreq_policy_apply_limits(struct cpufreq_policy *policy) __cpufreq_driver_target(policy, policy->min, CPUFREQ_RELATION_L); } static inline void cpufreq_policy_apply_limits_fast(struct cpufreq_policy *policy) { if (policy->max < policy->cur) cpufreq_driver_fast_switch(policy, policy->max); else if (policy->min > policy->cur) cpufreq_driver_fast_switch(policy, policy->min); } /* Governor attribute set */ struct gov_attr_set { struct kobject kobj; Loading
kernel/sched/cpufreq_schedutil.c +12 −4 Original line number Diff line number Diff line Loading @@ -1329,7 +1329,8 @@ static void sugov_stop(struct cpufreq_policy *policy) static void sugov_limits(struct cpufreq_policy *policy) { struct sugov_policy *sg_policy = policy->governor_data; unsigned long flags; unsigned long flags, now; unsigned int freq; if (!policy->fast_switch_enabled) { mutex_lock(&sg_policy->work_lock); Loading @@ -1341,9 +1342,16 @@ static void sugov_limits(struct cpufreq_policy *policy) mutex_unlock(&sg_policy->work_lock); } else { raw_spin_lock_irqsave(&sg_policy->update_lock, flags); sugov_track_cycles(sg_policy, sg_policy->policy->cur, ktime_get_ns()); cpufreq_policy_apply_limits_fast(policy); freq = policy->cur; now = ktime_get_ns(); /* * cpufreq_driver_resolve_freq() has a clamp, so we do not need * to do any sort of additional validation here. */ freq = cpufreq_driver_resolve_freq(policy, freq); sg_policy->cached_raw_freq = freq; sugov_fast_switch(sg_policy, now, freq); raw_spin_unlock_irqrestore(&sg_policy->update_lock, flags); } Loading