Loading include/trace/events/sched.h +4 −13 Original line number Diff line number Diff line Loading @@ -71,10 +71,8 @@ TRACE_EVENT(sched_enq_deq_task, __field(unsigned long, cpu_load ) __field(unsigned int, rt_nr_running ) __field(unsigned int, cpus_allowed ) #ifdef CONFIG_SCHED_WALT __field(unsigned int, demand ) __field(unsigned int, pred_demand ) #endif ), TP_fast_assign( Loading @@ -87,24 +85,17 @@ TRACE_EVENT(sched_enq_deq_task, __entry->cpu_load = task_rq(p)->cpu_load[0]; __entry->rt_nr_running = task_rq(p)->rt.rt_nr_running; __entry->cpus_allowed = cpus_allowed; #ifdef CONFIG_SCHED_WALT __entry->demand = p->ravg.demand; __entry->pred_demand = p->ravg.pred_demand; #endif __entry->demand = task_load(p); __entry->pred_demand = task_pl(p); ), TP_printk("cpu=%d %s comm=%s pid=%d prio=%d nr_running=%u cpu_load=%lu rt_nr_running=%u affine=%x" #ifdef CONFIG_SCHED_WALT " demand=%u pred_demand=%u" #endif , __entry->cpu, TP_printk("cpu=%d %s comm=%s pid=%d prio=%d nr_running=%u cpu_load=%lu rt_nr_running=%u affine=%x demand=%u pred_demand=%u", __entry->cpu, __entry->enqueue ? "enqueue" : "dequeue", __entry->comm, __entry->pid, __entry->prio, __entry->nr_running, __entry->cpu_load, __entry->rt_nr_running, __entry->cpus_allowed #ifdef CONFIG_SCHED_WALT , __entry->demand, __entry->pred_demand #endif ) ); Loading kernel/sched/sched.h +6 −0 Original line number Diff line number Diff line Loading @@ -2536,6 +2536,11 @@ static inline unsigned int task_load(struct task_struct *p) return p->ravg.demand; } static inline unsigned int task_pl(struct task_struct *p) { return p->ravg.pred_demand; } #define pct_to_real(tunable) \ (div64_u64((u64)tunable * (u64)max_task_load(), 100)) Loading Loading @@ -2798,6 +2803,7 @@ struct related_thread_group *task_related_thread_group(struct task_struct *p) } static inline u32 task_load(struct task_struct *p) { return 0; } static inline u32 task_pl(struct task_struct *p) { return 0; } static inline int update_preferred_cluster(struct related_thread_group *grp, struct task_struct *p, u32 old_load) Loading Loading
include/trace/events/sched.h +4 −13 Original line number Diff line number Diff line Loading @@ -71,10 +71,8 @@ TRACE_EVENT(sched_enq_deq_task, __field(unsigned long, cpu_load ) __field(unsigned int, rt_nr_running ) __field(unsigned int, cpus_allowed ) #ifdef CONFIG_SCHED_WALT __field(unsigned int, demand ) __field(unsigned int, pred_demand ) #endif ), TP_fast_assign( Loading @@ -87,24 +85,17 @@ TRACE_EVENT(sched_enq_deq_task, __entry->cpu_load = task_rq(p)->cpu_load[0]; __entry->rt_nr_running = task_rq(p)->rt.rt_nr_running; __entry->cpus_allowed = cpus_allowed; #ifdef CONFIG_SCHED_WALT __entry->demand = p->ravg.demand; __entry->pred_demand = p->ravg.pred_demand; #endif __entry->demand = task_load(p); __entry->pred_demand = task_pl(p); ), TP_printk("cpu=%d %s comm=%s pid=%d prio=%d nr_running=%u cpu_load=%lu rt_nr_running=%u affine=%x" #ifdef CONFIG_SCHED_WALT " demand=%u pred_demand=%u" #endif , __entry->cpu, TP_printk("cpu=%d %s comm=%s pid=%d prio=%d nr_running=%u cpu_load=%lu rt_nr_running=%u affine=%x demand=%u pred_demand=%u", __entry->cpu, __entry->enqueue ? "enqueue" : "dequeue", __entry->comm, __entry->pid, __entry->prio, __entry->nr_running, __entry->cpu_load, __entry->rt_nr_running, __entry->cpus_allowed #ifdef CONFIG_SCHED_WALT , __entry->demand, __entry->pred_demand #endif ) ); Loading
kernel/sched/sched.h +6 −0 Original line number Diff line number Diff line Loading @@ -2536,6 +2536,11 @@ static inline unsigned int task_load(struct task_struct *p) return p->ravg.demand; } static inline unsigned int task_pl(struct task_struct *p) { return p->ravg.pred_demand; } #define pct_to_real(tunable) \ (div64_u64((u64)tunable * (u64)max_task_load(), 100)) Loading Loading @@ -2798,6 +2803,7 @@ struct related_thread_group *task_related_thread_group(struct task_struct *p) } static inline u32 task_load(struct task_struct *p) { return 0; } static inline u32 task_pl(struct task_struct *p) { return 0; } static inline int update_preferred_cluster(struct related_thread_group *grp, struct task_struct *p, u32 old_load) Loading