ftrace: replace raw_local_irq_save with local_irq_save
authorSteven Rostedt <srostedt@redhat.com>
Tue, 2 Dec 2008 20:34:05 +0000 (15:34 -0500)
committerIngo Molnar <mingo@elte.hu>
Wed, 3 Dec 2008 07:56:19 +0000 (08:56 +0100)
Impact: fix for lockdep and ftrace

The raw_local_irq_save/restore confuses lockdep. This patch
converts them to the local_irq_save/restore variants.

Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
kernel/lockdep.c
kernel/trace/trace.c
kernel/trace/trace_branch.c
kernel/trace/trace_stack.c

index 46a4041..74b1878 100644 (file)
@@ -25,6 +25,7 @@
  * Thanks to Arjan van de Ven for coming up with the initial idea of
  * mapping lock dependencies runtime.
  */
+#define DISABLE_BRANCH_PROFILING
 #include <linux/mutex.h>
 #include <linux/sched.h>
 #include <linux/delay.h>
index 91887a2..380de63 100644 (file)
@@ -1209,7 +1209,7 @@ void trace_graph_entry(struct ftrace_graph_ent *trace)
        int cpu;
        int pc;
 
-       raw_local_irq_save(flags);
+       local_irq_save(flags);
        cpu = raw_smp_processor_id();
        data = tr->data[cpu];
        disabled = atomic_inc_return(&data->disabled);
@@ -1218,7 +1218,7 @@ void trace_graph_entry(struct ftrace_graph_ent *trace)
                __trace_graph_entry(tr, data, trace, flags, pc);
        }
        atomic_dec(&data->disabled);
-       raw_local_irq_restore(flags);
+       local_irq_restore(flags);
 }
 
 void trace_graph_return(struct ftrace_graph_ret *trace)
@@ -1230,7 +1230,7 @@ void trace_graph_return(struct ftrace_graph_ret *trace)
        int cpu;
        int pc;
 
-       raw_local_irq_save(flags);
+       local_irq_save(flags);
        cpu = raw_smp_processor_id();
        data = tr->data[cpu];
        disabled = atomic_inc_return(&data->disabled);
@@ -1239,7 +1239,7 @@ void trace_graph_return(struct ftrace_graph_ret *trace)
                __trace_graph_return(tr, data, trace, flags, pc);
        }
        atomic_dec(&data->disabled);
-       raw_local_irq_restore(flags);
+       local_irq_restore(flags);
 }
 #endif /* CONFIG_FUNCTION_GRAPH_TRACER */
 
@@ -2645,7 +2645,7 @@ tracing_cpumask_write(struct file *filp, const char __user *ubuf,
        if (err)
                goto err_unlock;
 
-       raw_local_irq_disable();
+       local_irq_disable();
        __raw_spin_lock(&ftrace_max_lock);
        for_each_tracing_cpu(cpu) {
                /*
@@ -2662,7 +2662,7 @@ tracing_cpumask_write(struct file *filp, const char __user *ubuf,
                }
        }
        __raw_spin_unlock(&ftrace_max_lock);
-       raw_local_irq_enable();
+       local_irq_enable();
 
        tracing_cpumask = tracing_cpumask_new;
 
index bc97275..6c00feb 100644 (file)
@@ -42,7 +42,7 @@ probe_likely_condition(struct ftrace_branch_data *f, int val, int expect)
        if (unlikely(!tr))
                return;
 
-       raw_local_irq_save(flags);
+       local_irq_save(flags);
        cpu = raw_smp_processor_id();
        if (atomic_inc_return(&tr->data[cpu]->disabled) != 1)
                goto out;
@@ -74,7 +74,7 @@ probe_likely_condition(struct ftrace_branch_data *f, int val, int expect)
 
  out:
        atomic_dec(&tr->data[cpu]->disabled);
-       raw_local_irq_restore(flags);
+       local_irq_restore(flags);
 }
 
 static inline
index fde3be1..06a1611 100644 (file)
@@ -48,7 +48,7 @@ static inline void check_stack(void)
        if (!object_is_on_stack(&this_size))
                return;
 
-       raw_local_irq_save(flags);
+       local_irq_save(flags);
        __raw_spin_lock(&max_stack_lock);
 
        /* a race could have already updated it */
@@ -96,7 +96,7 @@ static inline void check_stack(void)
 
  out:
        __raw_spin_unlock(&max_stack_lock);
-       raw_local_irq_restore(flags);
+       local_irq_restore(flags);
 }
 
 static void
@@ -162,11 +162,11 @@ stack_max_size_write(struct file *filp, const char __user *ubuf,
        if (ret < 0)
                return ret;
 
-       raw_local_irq_save(flags);
+       local_irq_save(flags);
        __raw_spin_lock(&max_stack_lock);
        *ptr = val;
        __raw_spin_unlock(&max_stack_lock);
-       raw_local_irq_restore(flags);
+       local_irq_restore(flags);
 
        return count;
 }