Loading kernel/sched/core.c +0 −2 Original line number Diff line number Diff line Loading @@ -7855,9 +7855,7 @@ void __init sched_init(void) BUG_ON(num_possible_cpus() > BITS_PER_LONG); #ifdef CONFIG_SCHED_HMP init_clusters(); #endif #ifdef CONFIG_FAIR_GROUP_SCHED alloc_size += 2 * nr_cpu_ids * sizeof(void **); Loading kernel/sched/sched.h +1 −0 Original line number Diff line number Diff line Loading @@ -1426,6 +1426,7 @@ static inline void clear_hmp_request(int cpu) { } static inline void mark_task_starting(struct task_struct *p) { } static inline void set_window_start(struct rq *rq) { } static inline void migrate_sync_cpu(int cpu, int new_cpu) {} static inline void init_clusters(void) {} static inline void update_cluster_topology(void) { } static inline void set_task_last_wake(struct task_struct *p, u64 wallclock) { } static inline void set_task_last_switch_out(struct task_struct *p, Loading Loading
kernel/sched/core.c +0 −2 Original line number Diff line number Diff line Loading @@ -7855,9 +7855,7 @@ void __init sched_init(void) BUG_ON(num_possible_cpus() > BITS_PER_LONG); #ifdef CONFIG_SCHED_HMP init_clusters(); #endif #ifdef CONFIG_FAIR_GROUP_SCHED alloc_size += 2 * nr_cpu_ids * sizeof(void **); Loading
kernel/sched/sched.h +1 −0 Original line number Diff line number Diff line Loading @@ -1426,6 +1426,7 @@ static inline void clear_hmp_request(int cpu) { } static inline void mark_task_starting(struct task_struct *p) { } static inline void set_window_start(struct rq *rq) { } static inline void migrate_sync_cpu(int cpu, int new_cpu) {} static inline void init_clusters(void) {} static inline void update_cluster_topology(void) { } static inline void set_task_last_wake(struct task_struct *p, u64 wallclock) { } static inline void set_task_last_switch_out(struct task_struct *p, Loading