Loading kernel/sched/fair.c +13 −0 Original line number Diff line number Diff line Loading @@ -10953,6 +10953,19 @@ static int load_balance(int this_cpu, struct rq *this_rq, raw_spin_lock_irqsave(&busiest->lock, flags); /* * The CPUs are marked as reserved if tasks * are pushed/pulled from other CPUs. In that case, * bail out from the load balancer. */ if (is_reserved(this_cpu) || is_reserved(cpu_of(busiest))) { raw_spin_unlock_irqrestore(&busiest->lock, flags); *continue_balancing = 0; goto out; } /* don't kick the active_load_balance_cpu_stop, * if the curr task on busiest cpu can't be * moved to this_cpu Loading Loading
kernel/sched/fair.c +13 −0 Original line number Diff line number Diff line Loading @@ -10953,6 +10953,19 @@ static int load_balance(int this_cpu, struct rq *this_rq, raw_spin_lock_irqsave(&busiest->lock, flags); /* * The CPUs are marked as reserved if tasks * are pushed/pulled from other CPUs. In that case, * bail out from the load balancer. */ if (is_reserved(this_cpu) || is_reserved(cpu_of(busiest))) { raw_spin_unlock_irqrestore(&busiest->lock, flags); *continue_balancing = 0; goto out; } /* don't kick the active_load_balance_cpu_stop, * if the curr task on busiest cpu can't be * moved to this_cpu Loading