Loading include/trace/events/sched.h +8 −4 Original line number Diff line number Diff line Loading @@ -436,9 +436,9 @@ TRACE_EVENT(sched_update_task_ravg, TRACE_EVENT(sched_get_task_cpu_cycles, TP_PROTO(int cpu, int event, u64 cycles, u64 exec_time), TP_PROTO(int cpu, int event, u64 cycles, u64 exec_time, struct task_struct *p), TP_ARGS(cpu, event, cycles, exec_time), TP_ARGS(cpu, event, cycles, exec_time, p), TP_STRUCT__entry( __field(int, cpu ) Loading @@ -448,6 +448,8 @@ TRACE_EVENT(sched_get_task_cpu_cycles, __field(u32, freq ) __field(u32, legacy_freq ) __field(u32, max_freq) __field(pid_t, pid ) __array(char, comm, TASK_COMM_LEN ) ), TP_fast_assign( Loading @@ -458,12 +460,14 @@ TRACE_EVENT(sched_get_task_cpu_cycles, __entry->freq = cpu_cycles_to_freq(cycles, exec_time); __entry->legacy_freq = cpu_cur_freq(cpu); __entry->max_freq = cpu_max_freq(cpu); __entry->pid = p->pid; memcpy(__entry->comm, p->comm, TASK_COMM_LEN); ), TP_printk("cpu=%d event=%d cycles=%llu exec_time=%llu freq=%u legacy_freq=%u max_freq=%u", TP_printk("cpu=%d event=%d cycles=%llu exec_time=%llu freq=%u legacy_freq=%u max_freq=%u task=%d (%s)", __entry->cpu, __entry->event, __entry->cycles, __entry->exec_time, __entry->freq, __entry->legacy_freq, __entry->max_freq) __entry->max_freq, __entry->pid, __entry->comm) ); TRACE_EVENT(sched_update_history, Loading kernel/sched/hmp.c +2 −1 Original line number Diff line number Diff line Loading @@ -2612,7 +2612,8 @@ update_task_rq_cpu_cycles(struct task_struct *p, struct rq *rq, int event, p->cpu_cycles = cur_cycles; trace_sched_get_task_cpu_cycles(cpu, event, rq->cc.cycles, rq->cc.time); trace_sched_get_task_cpu_cycles(cpu, event, rq->cc.cycles, rq->cc.time, p); } static int Loading kernel/sched/sched.h +1 −1 Original line number Diff line number Diff line Loading @@ -1249,7 +1249,7 @@ static inline int cpu_min_power_cost(int cpu) return cpu_rq(cpu)->cluster->min_power_cost; } static inline u32 cpu_cycles_to_freq(u64 cycles, u32 period) static inline u32 cpu_cycles_to_freq(u64 cycles, u64 period) { return div64_u64(cycles, period); } Loading Loading
include/trace/events/sched.h +8 −4 Original line number Diff line number Diff line Loading @@ -436,9 +436,9 @@ TRACE_EVENT(sched_update_task_ravg, TRACE_EVENT(sched_get_task_cpu_cycles, TP_PROTO(int cpu, int event, u64 cycles, u64 exec_time), TP_PROTO(int cpu, int event, u64 cycles, u64 exec_time, struct task_struct *p), TP_ARGS(cpu, event, cycles, exec_time), TP_ARGS(cpu, event, cycles, exec_time, p), TP_STRUCT__entry( __field(int, cpu ) Loading @@ -448,6 +448,8 @@ TRACE_EVENT(sched_get_task_cpu_cycles, __field(u32, freq ) __field(u32, legacy_freq ) __field(u32, max_freq) __field(pid_t, pid ) __array(char, comm, TASK_COMM_LEN ) ), TP_fast_assign( Loading @@ -458,12 +460,14 @@ TRACE_EVENT(sched_get_task_cpu_cycles, __entry->freq = cpu_cycles_to_freq(cycles, exec_time); __entry->legacy_freq = cpu_cur_freq(cpu); __entry->max_freq = cpu_max_freq(cpu); __entry->pid = p->pid; memcpy(__entry->comm, p->comm, TASK_COMM_LEN); ), TP_printk("cpu=%d event=%d cycles=%llu exec_time=%llu freq=%u legacy_freq=%u max_freq=%u", TP_printk("cpu=%d event=%d cycles=%llu exec_time=%llu freq=%u legacy_freq=%u max_freq=%u task=%d (%s)", __entry->cpu, __entry->event, __entry->cycles, __entry->exec_time, __entry->freq, __entry->legacy_freq, __entry->max_freq) __entry->max_freq, __entry->pid, __entry->comm) ); TRACE_EVENT(sched_update_history, Loading
kernel/sched/hmp.c +2 −1 Original line number Diff line number Diff line Loading @@ -2612,7 +2612,8 @@ update_task_rq_cpu_cycles(struct task_struct *p, struct rq *rq, int event, p->cpu_cycles = cur_cycles; trace_sched_get_task_cpu_cycles(cpu, event, rq->cc.cycles, rq->cc.time); trace_sched_get_task_cpu_cycles(cpu, event, rq->cc.cycles, rq->cc.time, p); } static int Loading
kernel/sched/sched.h +1 −1 Original line number Diff line number Diff line Loading @@ -1249,7 +1249,7 @@ static inline int cpu_min_power_cost(int cpu) return cpu_rq(cpu)->cluster->min_power_cost; } static inline u32 cpu_cycles_to_freq(u64 cycles, u32 period) static inline u32 cpu_cycles_to_freq(u64 cycles, u64 period) { return div64_u64(cycles, period); } Loading