Loading include/linux/sched.h +0 −6 Original line number Diff line number Diff line Loading @@ -2130,17 +2130,11 @@ ftrace_ctx_switch(struct task_struct *prev, struct task_struct *next) #ifdef CONFIG_SCHED_TRACER extern void ftrace_wake_up_task(struct task_struct *wakee, struct task_struct *curr); extern void ftrace_wake_up_new_task(struct task_struct *wakee, struct task_struct *curr); #else static inline void ftrace_wake_up_task(struct task_struct *wakee, struct task_struct *curr) { } static inline void ftrace_wake_up_new_task(struct task_struct *wakee, struct task_struct *curr) { } #endif extern long sched_setaffinity(pid_t pid, const cpumask_t *new_mask); Loading kernel/sched.c +1 −1 Original line number Diff line number Diff line Loading @@ -2613,7 +2613,7 @@ void wake_up_new_task(struct task_struct *p, unsigned long clone_flags) p->sched_class->task_new(rq, p); inc_nr_running(rq); } ftrace_wake_up_new_task(p, rq->curr); ftrace_wake_up_task(p, rq->curr); check_preempt_curr(rq, p); #ifdef CONFIG_SMP if (p->sched_class->task_wake_up) Loading kernel/trace/trace_sched_wakeup.c +3 −10 Original line number Diff line number Diff line Loading @@ -201,20 +201,13 @@ wakeup_check_start(struct trace_array *tr, struct task_struct *p, atomic_dec(&tr->data[cpu]->disabled); } void wakeup_sched_wakeup(struct task_struct *wakee, struct task_struct *curr) void wakeup_sched_wakeup(struct task_struct *wakee, struct task_struct *curr) { if (likely(!tracer_enabled)) return; wakeup_check_start(wakeup_trace, wakee, curr); } void ftrace_wake_up_new_task(struct task_struct *wakee, struct task_struct *curr) { if (likely(!tracer_enabled)) return; tracing_record_cmdline(curr); tracing_record_cmdline(wakee); wakeup_check_start(wakeup_trace, wakee, curr); } Loading Loading
include/linux/sched.h +0 −6 Original line number Diff line number Diff line Loading @@ -2130,17 +2130,11 @@ ftrace_ctx_switch(struct task_struct *prev, struct task_struct *next) #ifdef CONFIG_SCHED_TRACER extern void ftrace_wake_up_task(struct task_struct *wakee, struct task_struct *curr); extern void ftrace_wake_up_new_task(struct task_struct *wakee, struct task_struct *curr); #else static inline void ftrace_wake_up_task(struct task_struct *wakee, struct task_struct *curr) { } static inline void ftrace_wake_up_new_task(struct task_struct *wakee, struct task_struct *curr) { } #endif extern long sched_setaffinity(pid_t pid, const cpumask_t *new_mask); Loading
kernel/sched.c +1 −1 Original line number Diff line number Diff line Loading @@ -2613,7 +2613,7 @@ void wake_up_new_task(struct task_struct *p, unsigned long clone_flags) p->sched_class->task_new(rq, p); inc_nr_running(rq); } ftrace_wake_up_new_task(p, rq->curr); ftrace_wake_up_task(p, rq->curr); check_preempt_curr(rq, p); #ifdef CONFIG_SMP if (p->sched_class->task_wake_up) Loading
kernel/trace/trace_sched_wakeup.c +3 −10 Original line number Diff line number Diff line Loading @@ -201,20 +201,13 @@ wakeup_check_start(struct trace_array *tr, struct task_struct *p, atomic_dec(&tr->data[cpu]->disabled); } void wakeup_sched_wakeup(struct task_struct *wakee, struct task_struct *curr) void wakeup_sched_wakeup(struct task_struct *wakee, struct task_struct *curr) { if (likely(!tracer_enabled)) return; wakeup_check_start(wakeup_trace, wakee, curr); } void ftrace_wake_up_new_task(struct task_struct *wakee, struct task_struct *curr) { if (likely(!tracer_enabled)) return; tracing_record_cmdline(curr); tracing_record_cmdline(wakee); wakeup_check_start(wakeup_trace, wakee, curr); } Loading