Loading kernel/sched/core.c +0 −14 Original line number Diff line number Diff line Loading @@ -8052,20 +8052,6 @@ int sched_cpu_deactivate(unsigned int cpu) int ret; set_cpu_active(cpu, false); /* * We've cleared cpu_active_mask, wait for all preempt-disabled and RCU * users of this state to go away such that all new such users will * observe it. * * For CONFIG_PREEMPT we have preemptible RCU and its sync_rcu() might * not imply sync_sched(), so wait for both. * * Do sync before park smpboot threads to take care the rcu boost case. */ if (IS_ENABLED(CONFIG_PREEMPT)) synchronize_rcu_mult(call_rcu, call_rcu_sched); else synchronize_rcu(); if (!sched_smp_initialized) return 0; Loading Loading
kernel/sched/core.c +0 −14 Original line number Diff line number Diff line Loading @@ -8052,20 +8052,6 @@ int sched_cpu_deactivate(unsigned int cpu) int ret; set_cpu_active(cpu, false); /* * We've cleared cpu_active_mask, wait for all preempt-disabled and RCU * users of this state to go away such that all new such users will * observe it. * * For CONFIG_PREEMPT we have preemptible RCU and its sync_rcu() might * not imply sync_sched(), so wait for both. * * Do sync before park smpboot threads to take care the rcu boost case. */ if (IS_ENABLED(CONFIG_PREEMPT)) synchronize_rcu_mult(call_rcu, call_rcu_sched); else synchronize_rcu(); if (!sched_smp_initialized) return 0; Loading