Loading include/trace/events/sched.h +6 −4 Original line number Diff line number Diff line Loading @@ -117,9 +117,9 @@ TRACE_EVENT(sched_enq_deq_task, TRACE_EVENT(sched_task_load, TP_PROTO(struct task_struct *p, int small_task, int boost, int reason, int sync), int sync, int prefer_idle), TP_ARGS(p, small_task, boost, reason, sync), TP_ARGS(p, small_task, boost, reason, sync, prefer_idle), TP_STRUCT__entry( __array( char, comm, TASK_COMM_LEN ) Loading @@ -132,6 +132,7 @@ TRACE_EVENT(sched_task_load, __field( int, boost ) __field( int, reason ) __field( int, sync ) __field( int, prefer_idle ) ), TP_fast_assign( Loading @@ -145,13 +146,14 @@ TRACE_EVENT(sched_task_load, __entry->boost = boost; __entry->reason = reason; __entry->sync = sync; __entry->prefer_idle = prefer_idle; ), TP_printk("%d (%s): sum=%u, sum_scaled=%u, period=%u demand=%u small=%d boost=%d reason=%d sync=%d", TP_printk("%d (%s): sum=%u, sum_scaled=%u, period=%u demand=%u small=%d boost=%d reason=%d sync=%d prefer_idle=%d", __entry->pid, __entry->comm, __entry->sum, __entry->sum_scaled, __entry->period, __entry->demand, __entry->small_task, __entry->boost, __entry->reason, __entry->sync) __entry->sync, __entry->prefer_idle) ); TRACE_EVENT(sched_cpu_load, Loading kernel/sched/fair.c +1 −1 Original line number Diff line number Diff line Loading @@ -1915,7 +1915,7 @@ static int select_best_cpu(struct task_struct *p, int target, int reason, int cstate, min_cstate = INT_MAX; int prefer_idle = reason ? 1 : sysctl_sched_prefer_idle; trace_sched_task_load(p, small_task, boost, reason, sync); trace_sched_task_load(p, small_task, boost, reason, sync, prefer_idle); if (small_task && !boost) { best_cpu = best_small_task_cpu(p, sync); Loading Loading
include/trace/events/sched.h +6 −4 Original line number Diff line number Diff line Loading @@ -117,9 +117,9 @@ TRACE_EVENT(sched_enq_deq_task, TRACE_EVENT(sched_task_load, TP_PROTO(struct task_struct *p, int small_task, int boost, int reason, int sync), int sync, int prefer_idle), TP_ARGS(p, small_task, boost, reason, sync), TP_ARGS(p, small_task, boost, reason, sync, prefer_idle), TP_STRUCT__entry( __array( char, comm, TASK_COMM_LEN ) Loading @@ -132,6 +132,7 @@ TRACE_EVENT(sched_task_load, __field( int, boost ) __field( int, reason ) __field( int, sync ) __field( int, prefer_idle ) ), TP_fast_assign( Loading @@ -145,13 +146,14 @@ TRACE_EVENT(sched_task_load, __entry->boost = boost; __entry->reason = reason; __entry->sync = sync; __entry->prefer_idle = prefer_idle; ), TP_printk("%d (%s): sum=%u, sum_scaled=%u, period=%u demand=%u small=%d boost=%d reason=%d sync=%d", TP_printk("%d (%s): sum=%u, sum_scaled=%u, period=%u demand=%u small=%d boost=%d reason=%d sync=%d prefer_idle=%d", __entry->pid, __entry->comm, __entry->sum, __entry->sum_scaled, __entry->period, __entry->demand, __entry->small_task, __entry->boost, __entry->reason, __entry->sync) __entry->sync, __entry->prefer_idle) ); TRACE_EVENT(sched_cpu_load, Loading
kernel/sched/fair.c +1 −1 Original line number Diff line number Diff line Loading @@ -1915,7 +1915,7 @@ static int select_best_cpu(struct task_struct *p, int target, int reason, int cstate, min_cstate = INT_MAX; int prefer_idle = reason ? 1 : sysctl_sched_prefer_idle; trace_sched_task_load(p, small_task, boost, reason, sync); trace_sched_task_load(p, small_task, boost, reason, sync, prefer_idle); if (small_task && !boost) { best_cpu = best_small_task_cpu(p, sync); Loading