Loading kernel/sched/fair.c +2 −6 Original line number Diff line number Diff line Loading @@ -8376,7 +8376,8 @@ int can_migrate_task(struct task_struct *p, struct lb_env *env) env->flags &= ~LBF_ALL_PINNED; if (energy_aware() && !env->dst_rq->rd->overutilized && env->idle == CPU_NEWLY_IDLE) { env->idle == CPU_NEWLY_IDLE && !task_in_related_thread_group(p)) { long util_cum_dst, util_cum_src; unsigned long demand; Loading Loading @@ -9695,7 +9696,6 @@ static struct sched_group *find_busiest_group(struct lb_env *env) if (energy_aware() && !env->dst_rq->rd->overutilized) { int cpu_local, cpu_busiest; long util_cum; unsigned long energy_local, energy_busiest; if (env->idle != CPU_NEWLY_IDLE) Loading @@ -9715,10 +9715,6 @@ static struct sched_group *find_busiest_group(struct lb_env *env) } else if (energy_local == energy_busiest) { if (cpu_rq(cpu_busiest)->nr_running < 2) goto out_balanced; util_cum = cpu_util_cum(cpu_busiest, 0); if (util_cum < cpu_util_cum(cpu_local, 0)) goto out_balanced; } } Loading Loading
kernel/sched/fair.c +2 −6 Original line number Diff line number Diff line Loading @@ -8376,7 +8376,8 @@ int can_migrate_task(struct task_struct *p, struct lb_env *env) env->flags &= ~LBF_ALL_PINNED; if (energy_aware() && !env->dst_rq->rd->overutilized && env->idle == CPU_NEWLY_IDLE) { env->idle == CPU_NEWLY_IDLE && !task_in_related_thread_group(p)) { long util_cum_dst, util_cum_src; unsigned long demand; Loading Loading @@ -9695,7 +9696,6 @@ static struct sched_group *find_busiest_group(struct lb_env *env) if (energy_aware() && !env->dst_rq->rd->overutilized) { int cpu_local, cpu_busiest; long util_cum; unsigned long energy_local, energy_busiest; if (env->idle != CPU_NEWLY_IDLE) Loading @@ -9715,10 +9715,6 @@ static struct sched_group *find_busiest_group(struct lb_env *env) } else if (energy_local == energy_busiest) { if (cpu_rq(cpu_busiest)->nr_running < 2) goto out_balanced; util_cum = cpu_util_cum(cpu_busiest, 0); if (util_cum < cpu_util_cum(cpu_local, 0)) goto out_balanced; } } Loading