[PATCH 0/7 v3] tracing: Add usecase of synchronize_rcu_tasks() and stack_tracer_disable()
From: Steven Rostedt
Date: Mon Apr 10 2017 - 14:12:09 EST
Hopefully this is the final version before I push this to linux-next.
Paul, can I get an ack on the last patch from you?
-- Steve
Paul E. McKenney (1):
rcu: Fix dyntick-idle tracing
Steven Rostedt (VMware) (6):
ftrace: Add use of synchronize_rcu_tasks() with dynamic trampolines
tracing: Replace the per_cpu() with __this_cpu*() in trace_stack.c
tracing: Add stack_tracer_disable/enable() functions
tracing: Rename trace_active to disable_stack_tracer and inline its modification
rcu/tracing: Add rcu_disabled to denote when rcu_irq_enter() will not work
tracing: Make sure rcu_irq_enter() can work for trace_*_rcuidle() trace events
----
include/linux/ftrace.h | 38 ++++++++++++++++++++++++++++
include/linux/rcupdate.h | 5 ++++
include/linux/tracepoint.h | 19 ++++++++------
kernel/rcu/tree.c | 62 +++++++++++++++++++++++++++-------------------
kernel/trace/Kconfig | 3 ++-
kernel/trace/ftrace.c | 42 ++++++++++++++-----------------
kernel/trace/trace_stack.c | 35 +++++++++++++-------------
7 files changed, 128 insertions(+), 76 deletions(-)
Diff against v2 (not counting the last two patches at the end of this series)
diff --git a/include/linux/ftrace.h b/include/linux/ftrace.h
index 4bde7ff..06b2990 100644
--- a/include/linux/ftrace.h
+++ b/include/linux/ftrace.h
@@ -323,7 +323,7 @@ static inline void stack_tracer_enable(void)
}
#else
static inline void stack_tracer_disable(void) { }
-static inline void stack_tracer_enabe(void) { }
+static inline void stack_tracer_enable(void) { }
#endif
struct ftrace_func_command {
diff --git a/kernel/trace/trace_stack.c b/kernel/trace/trace_stack.c
index 84fafb6..f2f02ff 100644
--- a/kernel/trace/trace_stack.c
+++ b/kernel/trace/trace_stack.c
@@ -211,8 +211,8 @@ stack_trace_call(unsigned long ip, unsigned long parent_ip,
preempt_disable_notrace();
/* no atomic needed, we only modify this variable by this cpu */
- this_cpu_inc(disable_stack_tracer);
- if (this_cpu_read(disable_stack_tracer) != 1)
+ __this_cpu_inc(disable_stack_tracer);
+ if (__this_cpu_read(disable_stack_tracer) != 1)
goto out;
ip += MCOUNT_INSN_SIZE;
@@ -220,7 +220,7 @@ stack_trace_call(unsigned long ip, unsigned long parent_ip,
check_stack(ip, &stack);
out:
- this_cpu_dec(disable_stack_tracer);
+ __this_cpu_dec(disable_stack_tracer);
/* prevent recursion in schedule */
preempt_enable_notrace();
}
@@ -264,13 +264,13 @@ stack_max_size_write(struct file *filp, const char __user *ubuf,
* we will cause circular lock, so we also need to increase
* the percpu disable_stack_tracer here.
*/
- this_cpu_inc(disable_stack_tracer);
+ __this_cpu_inc(disable_stack_tracer);
arch_spin_lock(&stack_trace_max_lock);
*ptr = val;
arch_spin_unlock(&stack_trace_max_lock);
- this_cpu_dec(disable_stack_tracer);
+ __this_cpu_dec(disable_stack_tracer);
local_irq_restore(flags);
return count;
@@ -306,7 +306,7 @@ static void *t_start(struct seq_file *m, loff_t *pos)
{
local_irq_disable();
- this_cpu_inc(disable_stack_tracer);
+ __this_cpu_inc(disable_stack_tracer);
arch_spin_lock(&stack_trace_max_lock);
@@ -320,7 +320,7 @@ static void t_stop(struct seq_file *m, void *p)
{
arch_spin_unlock(&stack_trace_max_lock);
- this_cpu_dec(disable_stack_tracer);
+ __this_cpu_dec(disable_stack_tracer);
local_irq_enable();
}