Loading include/linux/sched/sysctl.h +3 −3 Original line number Diff line number Diff line Loading @@ -48,11 +48,11 @@ walt_proc_update_handler(struct ctl_table *table, int write, #endif #if defined(CONFIG_DEBUG_PREEMPT) || defined(CONFIG_PREEMPT_TRACER) || \ defined(CONFIG_PREEMPTIRQ_EVENTS) #if defined(CONFIG_PREEMPT_TRACER) || defined(CONFIG_DEBUG_PREEMPT) extern unsigned int sysctl_preemptoff_tracing_threshold_ns; #endif #if defined(CONFIG_PREEMPTIRQ_EVENTS) && defined(CONFIG_IRQSOFF_TRACER) extern unsigned int sysctl_irqsoff_tracing_threshold_ns; #endif enum sched_tunable_scaling { Loading kernel/sched/core.c +5 −6 Original line number Diff line number Diff line Loading @@ -89,12 +89,6 @@ int sysctl_sched_rt_runtime = 950000; /* CPUs with isolated domains */ cpumask_var_t cpu_isolated_map; /* * preemptoff stack tracing threshold in ns. * default: 1ms */ unsigned int sysctl_preemptoff_tracing_threshold_ns = 1000000UL; /* * __task_rq_lock - lock the rq @p resides on. */ Loading Loading @@ -3206,6 +3200,11 @@ u64 scheduler_tick_max_deferment(void) #if defined(CONFIG_PREEMPT) && (defined(CONFIG_DEBUG_PREEMPT) || \ defined(CONFIG_PREEMPT_TRACER)) /* * preemptoff stack tracing threshold in ns. * default: 1ms */ unsigned int sysctl_preemptoff_tracing_threshold_ns = 1000000UL; struct preempt_store { u64 ts; Loading kernel/sysctl.c +3 −2 Original line number Diff line number Diff line Loading @@ -310,8 +310,7 @@ static struct ctl_table kern_table[] = { .mode = 0644, .proc_handler = proc_dointvec, }, #if defined(CONFIG_DEBUG_PREEMPT) || defined(CONFIG_PREEMPT_TRACER) || \ defined(CONFIG_PREEMPTIRQ_EVENTS) #if defined(CONFIG_PREEMPT_TRACER) || defined(CONFIG_DEBUG_PREEMPT) { .procname = "preemptoff_tracing_threshold_ns", .data = &sysctl_preemptoff_tracing_threshold_ns, Loading @@ -319,6 +318,8 @@ static struct ctl_table kern_table[] = { .mode = 0644, .proc_handler = proc_dointvec, }, #endif #if defined(CONFIG_IRQSOFF_TRACER) && defined(CONFIG_PREEMPTIRQ_EVENTS) { .procname = "irqsoff_tracing_threshold_ns", .data = &sysctl_irqsoff_tracing_threshold_ns, Loading kernel/trace/trace_irqsoff.c +6 −6 Original line number Diff line number Diff line Loading @@ -41,12 +41,6 @@ static int save_flags; static void stop_irqsoff_tracer(struct trace_array *tr, int graph); static int start_irqsoff_tracer(struct trace_array *tr, int graph); /* * irqsoff stack tracing threshold in ns. * default: 1ms */ unsigned int sysctl_irqsoff_tracing_threshold_ns = 1000000UL; #ifdef CONFIG_PREEMPT_TRACER static inline int preempt_trace(void) Loading Loading @@ -475,6 +469,12 @@ void time_hardirqs_off(unsigned long a0, unsigned long a1) #else /* !CONFIG_PROVE_LOCKING */ #ifdef CONFIG_PREEMPTIRQ_EVENTS /* * irqsoff stack tracing threshold in ns. * default: 1ms */ unsigned int sysctl_irqsoff_tracing_threshold_ns = 1000000UL; struct irqsoff_store { u64 ts; unsigned long caddr[4]; Loading Loading
include/linux/sched/sysctl.h +3 −3 Original line number Diff line number Diff line Loading @@ -48,11 +48,11 @@ walt_proc_update_handler(struct ctl_table *table, int write, #endif #if defined(CONFIG_DEBUG_PREEMPT) || defined(CONFIG_PREEMPT_TRACER) || \ defined(CONFIG_PREEMPTIRQ_EVENTS) #if defined(CONFIG_PREEMPT_TRACER) || defined(CONFIG_DEBUG_PREEMPT) extern unsigned int sysctl_preemptoff_tracing_threshold_ns; #endif #if defined(CONFIG_PREEMPTIRQ_EVENTS) && defined(CONFIG_IRQSOFF_TRACER) extern unsigned int sysctl_irqsoff_tracing_threshold_ns; #endif enum sched_tunable_scaling { Loading
kernel/sched/core.c +5 −6 Original line number Diff line number Diff line Loading @@ -89,12 +89,6 @@ int sysctl_sched_rt_runtime = 950000; /* CPUs with isolated domains */ cpumask_var_t cpu_isolated_map; /* * preemptoff stack tracing threshold in ns. * default: 1ms */ unsigned int sysctl_preemptoff_tracing_threshold_ns = 1000000UL; /* * __task_rq_lock - lock the rq @p resides on. */ Loading Loading @@ -3206,6 +3200,11 @@ u64 scheduler_tick_max_deferment(void) #if defined(CONFIG_PREEMPT) && (defined(CONFIG_DEBUG_PREEMPT) || \ defined(CONFIG_PREEMPT_TRACER)) /* * preemptoff stack tracing threshold in ns. * default: 1ms */ unsigned int sysctl_preemptoff_tracing_threshold_ns = 1000000UL; struct preempt_store { u64 ts; Loading
kernel/sysctl.c +3 −2 Original line number Diff line number Diff line Loading @@ -310,8 +310,7 @@ static struct ctl_table kern_table[] = { .mode = 0644, .proc_handler = proc_dointvec, }, #if defined(CONFIG_DEBUG_PREEMPT) || defined(CONFIG_PREEMPT_TRACER) || \ defined(CONFIG_PREEMPTIRQ_EVENTS) #if defined(CONFIG_PREEMPT_TRACER) || defined(CONFIG_DEBUG_PREEMPT) { .procname = "preemptoff_tracing_threshold_ns", .data = &sysctl_preemptoff_tracing_threshold_ns, Loading @@ -319,6 +318,8 @@ static struct ctl_table kern_table[] = { .mode = 0644, .proc_handler = proc_dointvec, }, #endif #if defined(CONFIG_IRQSOFF_TRACER) && defined(CONFIG_PREEMPTIRQ_EVENTS) { .procname = "irqsoff_tracing_threshold_ns", .data = &sysctl_irqsoff_tracing_threshold_ns, Loading
kernel/trace/trace_irqsoff.c +6 −6 Original line number Diff line number Diff line Loading @@ -41,12 +41,6 @@ static int save_flags; static void stop_irqsoff_tracer(struct trace_array *tr, int graph); static int start_irqsoff_tracer(struct trace_array *tr, int graph); /* * irqsoff stack tracing threshold in ns. * default: 1ms */ unsigned int sysctl_irqsoff_tracing_threshold_ns = 1000000UL; #ifdef CONFIG_PREEMPT_TRACER static inline int preempt_trace(void) Loading Loading @@ -475,6 +469,12 @@ void time_hardirqs_off(unsigned long a0, unsigned long a1) #else /* !CONFIG_PROVE_LOCKING */ #ifdef CONFIG_PREEMPTIRQ_EVENTS /* * irqsoff stack tracing threshold in ns. * default: 1ms */ unsigned int sysctl_irqsoff_tracing_threshold_ns = 1000000UL; struct irqsoff_store { u64 ts; unsigned long caddr[4]; Loading