Loading kernel/sched/core.c +1 −1 Original line number Diff line number Diff line Loading @@ -9622,11 +9622,11 @@ void sched_exit(struct task_struct *p) reset_task_stats(p); p->ravg.mark_start = wallclock; p->ravg.sum_history[0] = EXITING_TASK_MARKER; free_task_load_ptrs(p); enqueue_task(rq, p, 0); clear_ed_task(p, rq); task_rq_unlock(rq, p, &rf); free_task_load_ptrs(p); } #endif /* CONFIG_SCHED_WALT */ Loading kernel/sched/walt.c +5 −1 Original line number Diff line number Diff line /* * Copyright (c) 2016-2017, The Linux Foundation. All rights reserved. * Copyright (c) 2016-2018, The Linux Foundation. All rights reserved. * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License version 2 and Loading Loading @@ -1965,6 +1965,10 @@ void init_new_task_load(struct task_struct *p, bool idle_task) p->misfit = false; } /* * kfree() may wakeup kswapd. So this function should NOT be called * with any CPU's rq->lock acquired. */ void free_task_load_ptrs(struct task_struct *p) { kfree(p->ravg.curr_window_cpu); Loading Loading
kernel/sched/core.c +1 −1 Original line number Diff line number Diff line Loading @@ -9622,11 +9622,11 @@ void sched_exit(struct task_struct *p) reset_task_stats(p); p->ravg.mark_start = wallclock; p->ravg.sum_history[0] = EXITING_TASK_MARKER; free_task_load_ptrs(p); enqueue_task(rq, p, 0); clear_ed_task(p, rq); task_rq_unlock(rq, p, &rf); free_task_load_ptrs(p); } #endif /* CONFIG_SCHED_WALT */ Loading
kernel/sched/walt.c +5 −1 Original line number Diff line number Diff line /* * Copyright (c) 2016-2017, The Linux Foundation. All rights reserved. * Copyright (c) 2016-2018, The Linux Foundation. All rights reserved. * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License version 2 and Loading Loading @@ -1965,6 +1965,10 @@ void init_new_task_load(struct task_struct *p, bool idle_task) p->misfit = false; } /* * kfree() may wakeup kswapd. So this function should NOT be called * with any CPU's rq->lock acquired. */ void free_task_load_ptrs(struct task_struct *p) { kfree(p->ravg.curr_window_cpu); Loading