From: Ingo Molnar Date: Tue, 21 Aug 2012 09:27:00 +0000 (+0200) Subject: Merge tag 'perf-core-for-mingo' of git://git.kernel.org/pub/scm/linux/kernel/git... X-Git-Tag: v3.8-rc1~60^2~73 X-Git-Url: http://review.tizen.org/git/?a=commitdiff_plain;h=bcada3d4b8c96b8792c2306f363992ca5ab9da42;p=platform%2Fupstream%2Fkernel-adaptation-pc.git Merge tag 'perf-core-for-mingo' of git://git./linux/kernel/git/acme/linux into perf/core Pull perf/core improvements and fixes from Arnaldo Carvalho de Melo: * Fix include order for bison/flex-generated C files, from Ben Hutchings * Build fixes and documentation corrections from David Ahern * Group parsing support, from Jiri Olsa * UI/gtk refactorings and improvements from Namhyung Kim * NULL deref fix for perf script, from Namhyung Kim * Assorted cleanups from Robert Richter * Let O= makes handle relative paths, from Steven Rostedt * perf script python fixes, from Feng Tang. * Improve 'perf lock' error message when the needed tracepoints are not present, from David Ahern. * Initial bash completion support, from Frederic Weisbecker * Allow building without libelf, from Namhyung Kim. * Support DWARF CFI based unwind to have callchains when %bp based unwinding is not possible, from Jiri Olsa. * Symbol resolution fixes, while fixing support PPC64 files with an .opt ELF section was the end goal, several fixes for code that handles all architectures and cleanups are included, from Cody Schafer. * Add a description for the JIT interface, from Andi Kleen. * Assorted fixes for Documentation and build in 32 bit, from Robert Richter * Add support for non-tracepoint events in perf script python, from Feng Tang * Cache the libtraceevent event_format associated to each evsel early, so that we avoid relookups, i.e. calling pevent_find_event repeatedly when processing tracepoint events. [ This is to reduce the surface contact with libtraceevents and make clear what is that the perf tools needs from that lib: so far parsing the common and per event fields. ] Signed-off-by: Arnaldo Carvalho de Melo Signed-off-by: Ingo Molnar --- bcada3d4b8c96b8792c2306f363992ca5ab9da42 diff --cc kernel/trace/trace_functions.c index fdff65d,a426f41..483162a --- a/kernel/trace/trace_functions.c +++ b/kernel/trace/trace_functions.c @@@ -75,10 -75,16 +76,18 @@@ function_trace_call_preempt_only(unsign preempt_enable_notrace(); } + /* Our two options */ + enum { + TRACE_FUNC_OPT_STACK = 0x1, + TRACE_FUNC_OPT_PSTORE = 0x2, + }; + + static struct tracer_flags func_flags; + static void -function_trace_call(unsigned long ip, unsigned long parent_ip) +function_trace_call(unsigned long ip, unsigned long parent_ip, + struct ftrace_ops *op, struct pt_regs *pt_regs) + { struct trace_array *tr = func_trace; struct trace_array_cpu *data; @@@ -159,14 -170,9 +174,9 @@@ static struct ftrace_ops trace_ops __re static struct ftrace_ops trace_stack_ops __read_mostly = { .func = function_stack_trace_call, - .flags = FTRACE_OPS_FL_GLOBAL, + .flags = FTRACE_OPS_FL_GLOBAL | FTRACE_OPS_FL_RECURSION_SAFE, }; - /* Our two options */ - enum { - TRACE_FUNC_OPT_STACK = 0x1, - }; - static struct tracer_opt func_opts[] = { #ifdef CONFIG_STACKTRACE { TRACER_OPT(func_stack_trace, TRACE_FUNC_OPT_STACK) },