return ret;
}
-void trace_destroy_fields(struct ftrace_event_call *call)
+static void trace_destroy_fields(struct ftrace_event_call *call)
{
struct ftrace_event_field *field, *next;
struct list_head *head;
}
/*
- * Must be called under locking both of event_mutex and trace_event_mutex.
+ * Must be called under locking both of event_mutex and trace_event_sem.
*/
static void __trace_remove_event_call(struct ftrace_event_call *call)
{
void trace_remove_event_call(struct ftrace_event_call *call)
{
mutex_lock(&event_mutex);
- down_write(&trace_event_mutex);
+ down_write(&trace_event_sem);
__trace_remove_event_call(call);
- up_write(&trace_event_mutex);
+ up_write(&trace_event_sem);
mutex_unlock(&event_mutex);
}
struct ftrace_event_call *call, *p;
bool clear_trace = false;
- down_write(&trace_event_mutex);
+ down_write(&trace_event_sem);
list_for_each_entry_safe(call, p, &ftrace_events, list) {
if (call->mod == mod) {
if (call->flags & TRACE_EVENT_FL_WAS_ENABLED)
list_del(&file_ops->list);
kfree(file_ops);
}
- up_write(&trace_event_mutex);
+ up_write(&trace_event_sem);
/*
* It is safest to reset the ring buffer if the module being unloaded
if (ret)
goto out_unlock;
- down_write(&trace_event_mutex);
+ down_write(&trace_event_sem);
__trace_add_event_dirs(tr);
- up_write(&trace_event_mutex);
+ up_write(&trace_event_sem);
out_unlock:
mutex_unlock(&event_mutex);
if (ret)
goto out_unlock;
- down_write(&trace_event_mutex);
+ down_write(&trace_event_sem);
__trace_early_add_event_dirs(tr);
- up_write(&trace_event_mutex);
+ up_write(&trace_event_sem);
out_unlock:
mutex_unlock(&event_mutex);
mutex_lock(&event_mutex);
- down_write(&trace_event_mutex);
+ down_write(&trace_event_sem);
__trace_remove_event_dirs(tr);
debugfs_remove_recursive(tr->event_dir);
- up_write(&trace_event_mutex);
+ up_write(&trace_event_sem);
tr->event_dir = NULL;