4 * Builtin record command: Record the profile of a workload
5 * (or a CPU, or a PID) into the perf.data output file - for
6 * later analysis via perf report.
8 #define _FILE_OFFSET_BITS 64
14 #include "util/build-id.h"
15 #include "util/util.h"
16 #include "util/parse-options.h"
17 #include "util/parse-events.h"
19 #include "util/header.h"
20 #include "util/event.h"
21 #include "util/evlist.h"
22 #include "util/evsel.h"
23 #include "util/debug.h"
24 #include "util/session.h"
25 #include "util/tool.h"
26 #include "util/symbol.h"
27 #include "util/cpumap.h"
28 #include "util/thread_map.h"
38 static int __on_exit_count = 0;
39 typedef void (*on_exit_func_t) (int, void *);
40 static on_exit_func_t __on_exit_funcs[ATEXIT_MAX];
41 static void *__on_exit_args[ATEXIT_MAX];
42 static int __exitcode = 0;
43 static void __handle_on_exit_funcs(void);
44 static int on_exit(on_exit_func_t function, void *arg);
45 #define exit(x) (exit)(__exitcode = (x))
47 static int on_exit(on_exit_func_t function, void *arg)
49 if (__on_exit_count == ATEXIT_MAX)
51 else if (__on_exit_count == 0)
52 atexit(__handle_on_exit_funcs);
53 __on_exit_funcs[__on_exit_count] = function;
54 __on_exit_args[__on_exit_count++] = arg;
58 static void __handle_on_exit_funcs(void)
61 for (i = 0; i < __on_exit_count; i++)
62 __on_exit_funcs[i] (__exitcode, __on_exit_args[i]);
72 struct perf_tool tool;
73 struct perf_record_opts opts;
75 const char *output_name;
76 struct perf_evlist *evlist;
77 struct perf_session *session;
80 unsigned int page_size;
82 enum write_mode_t write_mode;
84 bool no_buildid_cache;
89 off_t post_processing_offset;
92 static void advance_output(struct perf_record *rec, size_t size)
94 rec->bytes_written += size;
97 static int write_output(struct perf_record *rec, void *buf, size_t size)
100 int ret = write(rec->output, buf, size);
103 pr_err("failed to write\n");
110 rec->bytes_written += ret;
116 static int process_synthesized_event(struct perf_tool *tool,
117 union perf_event *event,
118 struct perf_sample *sample __maybe_unused,
119 struct machine *machine __maybe_unused)
121 struct perf_record *rec = container_of(tool, struct perf_record, tool);
122 if (write_output(rec, event, event->header.size) < 0)
128 static int perf_record__mmap_read(struct perf_record *rec,
129 struct perf_mmap *md)
131 unsigned int head = perf_mmap__read_head(md);
132 unsigned int old = md->prev;
133 unsigned char *data = md->base + rec->page_size;
145 if ((old & md->mask) + size != (head & md->mask)) {
146 buf = &data[old & md->mask];
147 size = md->mask + 1 - (old & md->mask);
150 if (write_output(rec, buf, size) < 0) {
156 buf = &data[old & md->mask];
160 if (write_output(rec, buf, size) < 0) {
166 perf_mmap__write_tail(md, old);
172 static volatile int done = 0;
173 static volatile int signr = -1;
174 static volatile int child_finished = 0;
176 static void sig_handler(int sig)
185 static void perf_record__sig_exit(int exit_status __maybe_unused, void *arg)
187 struct perf_record *rec = arg;
190 if (rec->evlist->workload.pid > 0) {
192 kill(rec->evlist->workload.pid, SIGTERM);
195 if (WIFSIGNALED(status))
196 psignal(WTERMSIG(status), rec->progname);
199 if (signr == -1 || signr == SIGUSR1)
202 signal(signr, SIG_DFL);
203 kill(getpid(), signr);
206 static bool perf_evlist__equal(struct perf_evlist *evlist,
207 struct perf_evlist *other)
209 struct perf_evsel *pos, *pair;
211 if (evlist->nr_entries != other->nr_entries)
214 pair = perf_evlist__first(other);
216 list_for_each_entry(pos, &evlist->entries, node) {
217 if (memcmp(&pos->attr, &pair->attr, sizeof(pos->attr) != 0))
219 pair = perf_evsel__next(pair);
225 static int perf_record__open(struct perf_record *rec)
227 struct perf_evsel *pos;
228 struct perf_evlist *evlist = rec->evlist;
229 struct perf_session *session = rec->session;
230 struct perf_record_opts *opts = &rec->opts;
234 * Set the evsel leader links before we configure attributes,
235 * since some might depend on this info.
238 perf_evlist__set_leader(evlist);
240 perf_evlist__config_attrs(evlist, opts);
242 list_for_each_entry(pos, &evlist->entries, node) {
243 struct perf_event_attr *attr = &pos->attr;
245 * Check if parse_single_tracepoint_event has already asked for
248 * XXX this is kludgy but short term fix for problems introduced by
249 * eac23d1c that broke 'perf script' by having different sample_types
250 * when using multiple tracepoint events when we use a perf binary
251 * that tries to use sample_id_all on an older kernel.
253 * We need to move counter creation to perf_session, support
254 * different sample_types, etc.
256 bool time_needed = attr->sample_type & PERF_SAMPLE_TIME;
258 fallback_missing_features:
259 if (opts->exclude_guest_missing)
260 attr->exclude_guest = attr->exclude_host = 0;
262 attr->sample_id_all = opts->sample_id_all_missing ? 0 : 1;
264 if (perf_evsel__open(pos, evlist->cpus, evlist->threads) < 0) {
267 if (err == EPERM || err == EACCES) {
268 ui__error_paranoid();
271 } else if (err == ENODEV && opts->target.cpu_list) {
272 pr_err("No such device - did you specify"
273 " an out-of-range profile CPU?\n");
276 } else if (err == EINVAL) {
277 if (!opts->exclude_guest_missing &&
278 (attr->exclude_guest || attr->exclude_host)) {
279 pr_debug("Old kernel, cannot exclude "
280 "guest or host samples.\n");
281 opts->exclude_guest_missing = true;
282 goto fallback_missing_features;
283 } else if (!opts->sample_id_all_missing) {
285 * Old kernel, no attr->sample_id_type_all field
287 opts->sample_id_all_missing = true;
288 if (!opts->sample_time && !opts->raw_samples && !time_needed)
289 perf_evsel__reset_sample_bit(pos, TIME);
291 goto retry_sample_id;
296 * If it's cycles then fall back to hrtimer
297 * based cpu-clock-tick sw counter, which
298 * is always available even if no PMU support.
300 * PPC returns ENXIO until 2.6.37 (behavior changed
301 * with commit b0a873e).
303 if ((err == ENOENT || err == ENXIO)
304 && attr->type == PERF_TYPE_HARDWARE
305 && attr->config == PERF_COUNT_HW_CPU_CYCLES) {
308 ui__warning("The cycles event is not supported, "
309 "trying to fall back to cpu-clock-ticks\n");
310 attr->type = PERF_TYPE_SOFTWARE;
311 attr->config = PERF_COUNT_SW_CPU_CLOCK;
320 ui__error("The %s event is not supported.\n",
321 perf_evsel__name(pos));
324 } else if ((err == EOPNOTSUPP) && (attr->precise_ip)) {
325 ui__error("\'precise\' request may not be supported. "
326 "Try removing 'p' modifier\n");
332 error("sys_perf_event_open() syscall returned with %d "
333 "(%s) for event %s. /bin/dmesg may provide "
334 "additional information.\n",
335 err, strerror(err), perf_evsel__name(pos));
337 #if defined(__i386__) || defined(__x86_64__)
338 if (attr->type == PERF_TYPE_HARDWARE &&
340 pr_err("No hardware sampling interrupt available."
341 " No APIC? If so then you can boot the kernel"
342 " with the \"lapic\" boot parameter to"
343 " force-enable it.\n");
349 pr_err("No CONFIG_PERF_EVENTS=y kernel support configured?\n");
355 if (perf_evlist__apply_filters(evlist)) {
356 error("failed to set filter with %d (%s)\n", errno,
362 if (perf_evlist__mmap(evlist, opts->mmap_pages, false) < 0) {
363 if (errno == EPERM) {
364 pr_err("Permission error mapping pages.\n"
365 "Consider increasing "
366 "/proc/sys/kernel/perf_event_mlock_kb,\n"
367 "or try again with a smaller value of -m/--mmap_pages.\n"
368 "(current value: %d)\n", opts->mmap_pages);
370 } else if (!is_power_of_2(opts->mmap_pages) &&
371 (opts->mmap_pages != UINT_MAX)) {
372 pr_err("--mmap_pages/-m value must be a power of two.");
375 pr_err("failed to mmap with %d (%s)\n", errno, strerror(errno));
382 session->evlist = evlist;
384 if (!perf_evlist__equal(session->evlist, evlist)) {
385 fprintf(stderr, "incompatible append\n");
391 perf_session__set_id_hdr_size(session);
396 static int process_buildids(struct perf_record *rec)
398 u64 size = lseek(rec->output, 0, SEEK_CUR);
403 rec->session->fd = rec->output;
404 return __perf_session__process_events(rec->session, rec->post_processing_offset,
405 size - rec->post_processing_offset,
406 size, &build_id__mark_dso_hit_ops);
409 static void perf_record__exit(int status, void *arg)
411 struct perf_record *rec = arg;
416 if (!rec->opts.pipe_output) {
417 rec->session->header.data_size += rec->bytes_written;
419 if (!rec->no_buildid)
420 process_buildids(rec);
421 perf_session__write_header(rec->session, rec->evlist,
423 perf_session__delete(rec->session);
424 perf_evlist__delete(rec->evlist);
429 static void perf_event__synthesize_guest_os(struct machine *machine, void *data)
432 struct perf_tool *tool = data;
434 if (machine__is_host(machine))
438 *As for guest kernel when processing subcommand record&report,
439 *we arrange module mmap prior to guest kernel mmap and trigger
440 *a preload dso because default guest module symbols are loaded
441 *from guest kallsyms instead of /lib/modules/XXX/XXX. This
442 *method is used to avoid symbol missing when the first addr is
443 *in module instead of in guest kernel.
445 err = perf_event__synthesize_modules(tool, process_synthesized_event,
448 pr_err("Couldn't record guest kernel [%d]'s reference"
449 " relocation symbol.\n", machine->pid);
452 * We use _stext for guest kernel because guest kernel's /proc/kallsyms
453 * have no _text sometimes.
455 err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
458 err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
461 pr_err("Couldn't record guest kernel [%d]'s reference"
462 " relocation symbol.\n", machine->pid);
465 static struct perf_event_header finished_round_event = {
466 .size = sizeof(struct perf_event_header),
467 .type = PERF_RECORD_FINISHED_ROUND,
470 static int perf_record__mmap_read_all(struct perf_record *rec)
475 for (i = 0; i < rec->evlist->nr_mmaps; i++) {
476 if (rec->evlist->mmap[i].base) {
477 if (perf_record__mmap_read(rec, &rec->evlist->mmap[i]) != 0) {
484 if (perf_header__has_feat(&rec->session->header, HEADER_TRACING_DATA))
485 rc = write_output(rec, &finished_round_event,
486 sizeof(finished_round_event));
492 static int __cmd_record(struct perf_record *rec, int argc, const char **argv)
496 int err, output, feat;
497 unsigned long waking = 0;
498 const bool forks = argc > 0;
499 struct machine *machine;
500 struct perf_tool *tool = &rec->tool;
501 struct perf_record_opts *opts = &rec->opts;
502 struct perf_evlist *evsel_list = rec->evlist;
503 const char *output_name = rec->output_name;
504 struct perf_session *session;
505 bool disabled = false;
507 rec->progname = argv[0];
509 rec->page_size = sysconf(_SC_PAGE_SIZE);
511 on_exit(perf_record__sig_exit, rec);
512 signal(SIGCHLD, sig_handler);
513 signal(SIGINT, sig_handler);
514 signal(SIGUSR1, sig_handler);
517 if (!fstat(STDOUT_FILENO, &st) && S_ISFIFO(st.st_mode))
518 opts->pipe_output = true;
520 rec->output_name = output_name = "perf.data";
523 if (!strcmp(output_name, "-"))
524 opts->pipe_output = true;
525 else if (!stat(output_name, &st) && st.st_size) {
526 if (rec->write_mode == WRITE_FORCE) {
527 char oldname[PATH_MAX];
528 snprintf(oldname, sizeof(oldname), "%s.old",
531 rename(output_name, oldname);
533 } else if (rec->write_mode == WRITE_APPEND) {
534 rec->write_mode = WRITE_FORCE;
538 flags = O_CREAT|O_RDWR;
539 if (rec->write_mode == WRITE_APPEND)
544 if (opts->pipe_output)
545 output = STDOUT_FILENO;
547 output = open(output_name, flags, S_IRUSR | S_IWUSR);
549 perror("failed to create output file");
553 rec->output = output;
555 session = perf_session__new(output_name, O_WRONLY,
556 rec->write_mode == WRITE_FORCE, false, NULL);
557 if (session == NULL) {
558 pr_err("Not enough memory for reading perf file header\n");
562 rec->session = session;
564 for (feat = HEADER_FIRST_FEATURE; feat < HEADER_LAST_FEATURE; feat++)
565 perf_header__set_feat(&session->header, feat);
568 perf_header__clear_feat(&session->header, HEADER_BUILD_ID);
570 if (!have_tracepoints(&evsel_list->entries))
571 perf_header__clear_feat(&session->header, HEADER_TRACING_DATA);
573 if (!rec->opts.branch_stack)
574 perf_header__clear_feat(&session->header, HEADER_BRANCH_STACK);
576 if (!rec->file_new) {
577 err = perf_session__read_header(session, output);
579 goto out_delete_session;
583 err = perf_evlist__prepare_workload(evsel_list, opts, argv);
585 pr_err("Couldn't run the workload!\n");
586 goto out_delete_session;
590 if (perf_record__open(rec) != 0) {
592 goto out_delete_session;
596 * perf_session__delete(session) will be called at perf_record__exit()
598 on_exit(perf_record__exit, rec);
600 if (opts->pipe_output) {
601 err = perf_header__write_pipe(output);
603 goto out_delete_session;
604 } else if (rec->file_new) {
605 err = perf_session__write_header(session, evsel_list,
608 goto out_delete_session;
612 && !perf_header__has_feat(&session->header, HEADER_BUILD_ID)) {
613 pr_err("Couldn't generate buildids. "
614 "Use --no-buildid to profile anyway.\n");
616 goto out_delete_session;
619 rec->post_processing_offset = lseek(output, 0, SEEK_CUR);
621 machine = perf_session__find_host_machine(session);
623 pr_err("Couldn't find native kernel information.\n");
625 goto out_delete_session;
628 if (opts->pipe_output) {
629 err = perf_event__synthesize_attrs(tool, session,
630 process_synthesized_event);
632 pr_err("Couldn't synthesize attrs.\n");
633 goto out_delete_session;
636 err = perf_event__synthesize_event_types(tool, process_synthesized_event,
639 pr_err("Couldn't synthesize event_types.\n");
640 goto out_delete_session;
643 if (have_tracepoints(&evsel_list->entries)) {
645 * FIXME err <= 0 here actually means that
646 * there were no tracepoints so its not really
647 * an error, just that we don't need to
648 * synthesize anything. We really have to
649 * return this more properly and also
650 * propagate errors that now are calling die()
652 err = perf_event__synthesize_tracing_data(tool, output, evsel_list,
653 process_synthesized_event);
655 pr_err("Couldn't record tracing data.\n");
656 goto out_delete_session;
658 advance_output(rec, err);
662 err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
665 err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
668 pr_err("Couldn't record kernel reference relocation symbol\n"
669 "Symbol resolution may be skewed if relocation was used (e.g. kexec).\n"
670 "Check /proc/kallsyms permission or run as root.\n");
672 err = perf_event__synthesize_modules(tool, process_synthesized_event,
675 pr_err("Couldn't record kernel module information.\n"
676 "Symbol resolution may be skewed if relocation was used (e.g. kexec).\n"
677 "Check /proc/modules permission or run as root.\n");
680 perf_session__process_machines(session, tool,
681 perf_event__synthesize_guest_os);
683 if (!opts->target.system_wide)
684 err = perf_event__synthesize_thread_map(tool, evsel_list->threads,
685 process_synthesized_event,
688 err = perf_event__synthesize_threads(tool, process_synthesized_event,
692 goto out_delete_session;
694 if (rec->realtime_prio) {
695 struct sched_param param;
697 param.sched_priority = rec->realtime_prio;
698 if (sched_setscheduler(0, SCHED_FIFO, ¶m)) {
699 pr_err("Could not set realtime priority.\n");
701 goto out_delete_session;
706 * When perf is starting the traced process, all the events
707 * (apart from group members) have enable_on_exec=1 set,
708 * so don't spoil it by prematurely enabling them.
710 if (!perf_target__none(&opts->target))
711 perf_evlist__enable(evsel_list);
717 perf_evlist__start_workload(evsel_list);
720 int hits = rec->samples;
722 if (perf_record__mmap_read_all(rec) < 0) {
724 goto out_delete_session;
727 if (hits == rec->samples) {
730 err = poll(evsel_list->pollfd, evsel_list->nr_fds, -1);
735 * When perf is starting the traced process, at the end events
736 * die with the process and we wait for that. Thus no need to
737 * disable events in this case.
739 if (done && !disabled && !perf_target__none(&opts->target)) {
740 perf_evlist__disable(evsel_list);
745 if (quiet || signr == SIGUSR1)
748 fprintf(stderr, "[ perf record: Woken up %ld times to write data ]\n", waking);
751 * Approximate RIP event size: 24 bytes.
754 "[ perf record: Captured and wrote %.3f MB %s (~%" PRIu64 " samples) ]\n",
755 (double)rec->bytes_written / 1024.0 / 1024.0,
757 rec->bytes_written / 24);
762 perf_session__delete(session);
766 #define BRANCH_OPT(n, m) \
767 { .name = n, .mode = (m) }
769 #define BRANCH_END { .name = NULL }
776 static const struct branch_mode branch_modes[] = {
777 BRANCH_OPT("u", PERF_SAMPLE_BRANCH_USER),
778 BRANCH_OPT("k", PERF_SAMPLE_BRANCH_KERNEL),
779 BRANCH_OPT("hv", PERF_SAMPLE_BRANCH_HV),
780 BRANCH_OPT("any", PERF_SAMPLE_BRANCH_ANY),
781 BRANCH_OPT("any_call", PERF_SAMPLE_BRANCH_ANY_CALL),
782 BRANCH_OPT("any_ret", PERF_SAMPLE_BRANCH_ANY_RETURN),
783 BRANCH_OPT("ind_call", PERF_SAMPLE_BRANCH_IND_CALL),
788 parse_branch_stack(const struct option *opt, const char *str, int unset)
791 (PERF_SAMPLE_BRANCH_USER |\
792 PERF_SAMPLE_BRANCH_KERNEL |\
793 PERF_SAMPLE_BRANCH_HV)
795 uint64_t *mode = (uint64_t *)opt->value;
796 const struct branch_mode *br;
797 char *s, *os = NULL, *p;
804 * cannot set it twice, -b + --branch-filter for instance
809 /* str may be NULL in case no arg is passed to -b */
811 /* because str is read-only */
812 s = os = strdup(str);
821 for (br = branch_modes; br->name; br++) {
822 if (!strcasecmp(s, br->name))
826 ui__warning("unknown branch filter %s,"
827 " check man page\n", s);
841 /* default to any branch */
842 if ((*mode & ~ONLY_PLM) == 0) {
843 *mode = PERF_SAMPLE_BRANCH_ANY;
850 #ifdef LIBUNWIND_SUPPORT
851 static int get_stack_size(char *str, unsigned long *_size)
855 unsigned long max_size = round_down(USHRT_MAX, sizeof(u64));
857 size = strtoul(str, &endptr, 0);
863 size = round_up(size, sizeof(u64));
864 if (!size || size > max_size)
872 pr_err("callchain: Incorrect stack dump size (max %ld): %s\n",
876 #endif /* LIBUNWIND_SUPPORT */
879 parse_callchain_opt(const struct option *opt __maybe_unused, const char *arg,
882 struct perf_record *rec = (struct perf_record *)opt->value;
883 char *tok, *name, *saveptr = NULL;
887 /* --no-call-graph */
891 /* We specified default option if none is provided. */
894 /* We need buffer that we know we can write to. */
895 buf = malloc(strlen(arg) + 1);
901 tok = strtok_r((char *)buf, ",", &saveptr);
902 name = tok ? : (char *)buf;
905 /* Framepointer style */
906 if (!strncmp(name, "fp", sizeof("fp"))) {
907 if (!strtok_r(NULL, ",", &saveptr)) {
908 rec->opts.call_graph = CALLCHAIN_FP;
911 pr_err("callchain: No more arguments "
912 "needed for -g fp\n");
915 #ifdef LIBUNWIND_SUPPORT
917 } else if (!strncmp(name, "dwarf", sizeof("dwarf"))) {
918 const unsigned long default_stack_dump_size = 8192;
921 rec->opts.call_graph = CALLCHAIN_DWARF;
922 rec->opts.stack_dump_size = default_stack_dump_size;
924 tok = strtok_r(NULL, ",", &saveptr);
926 unsigned long size = 0;
928 ret = get_stack_size(tok, &size);
929 rec->opts.stack_dump_size = size;
933 pr_debug("callchain: stack dump size %d\n",
934 rec->opts.stack_dump_size);
935 #endif /* LIBUNWIND_SUPPORT */
937 pr_err("callchain: Unknown -g option "
947 pr_debug("callchain: type %d\n", rec->opts.call_graph);
952 static const char * const record_usage[] = {
953 "perf record [<options>] [<command>]",
954 "perf record [<options>] -- <command> [<options>]",
959 * XXX Ideally would be local to cmd_record() and passed to a perf_record__new
960 * because we need to have access to it in perf_record__exit, that is called
961 * after cmd_record() exits, but since record_options need to be accessible to
962 * builtin-script, leave it here.
964 * At least we don't ouch it in all the other functions here directly.
966 * Just say no to tons of global variables, sigh.
968 static struct perf_record record = {
970 .mmap_pages = UINT_MAX,
971 .user_freq = UINT_MAX,
972 .user_interval = ULLONG_MAX,
978 .write_mode = WRITE_FORCE,
982 #define CALLCHAIN_HELP "do call-graph (stack chain/backtrace) recording: "
984 #ifdef LIBUNWIND_SUPPORT
985 static const char callchain_help[] = CALLCHAIN_HELP "[fp] dwarf";
987 static const char callchain_help[] = CALLCHAIN_HELP "[fp]";
991 * XXX Will stay a global variable till we fix builtin-script.c to stop messing
992 * with it and switch to use the library functions in perf_evlist that came
993 * from builtin-record.c, i.e. use perf_record_opts,
994 * perf_evlist__prepare_workload, etc instead of fork+exec'in 'perf record',
997 const struct option record_options[] = {
998 OPT_CALLBACK('e', "event", &record.evlist, "event",
999 "event selector. use 'perf list' to list available events",
1000 parse_events_option),
1001 OPT_CALLBACK(0, "filter", &record.evlist, "filter",
1002 "event filter", parse_filter),
1003 OPT_STRING('p', "pid", &record.opts.target.pid, "pid",
1004 "record events on existing process id"),
1005 OPT_STRING('t', "tid", &record.opts.target.tid, "tid",
1006 "record events on existing thread id"),
1007 OPT_INTEGER('r', "realtime", &record.realtime_prio,
1008 "collect data with this RT SCHED_FIFO priority"),
1009 OPT_BOOLEAN('D', "no-delay", &record.opts.no_delay,
1010 "collect data without buffering"),
1011 OPT_BOOLEAN('R', "raw-samples", &record.opts.raw_samples,
1012 "collect raw sample records from all opened counters"),
1013 OPT_BOOLEAN('a', "all-cpus", &record.opts.target.system_wide,
1014 "system-wide collection from all CPUs"),
1015 OPT_BOOLEAN('A', "append", &record.append_file,
1016 "append to the output file to do incremental profiling"),
1017 OPT_STRING('C', "cpu", &record.opts.target.cpu_list, "cpu",
1018 "list of cpus to monitor"),
1019 OPT_BOOLEAN('f', "force", &record.force,
1020 "overwrite existing data file (deprecated)"),
1021 OPT_U64('c', "count", &record.opts.user_interval, "event period to sample"),
1022 OPT_STRING('o', "output", &record.output_name, "file",
1023 "output file name"),
1024 OPT_BOOLEAN('i', "no-inherit", &record.opts.no_inherit,
1025 "child tasks do not inherit counters"),
1026 OPT_UINTEGER('F', "freq", &record.opts.user_freq, "profile at this frequency"),
1027 OPT_UINTEGER('m', "mmap-pages", &record.opts.mmap_pages,
1028 "number of mmap data pages"),
1029 OPT_BOOLEAN(0, "group", &record.opts.group,
1030 "put the counters into a counter group"),
1031 OPT_CALLBACK_DEFAULT('g', "call-graph", &record, "mode[,dump_size]",
1032 callchain_help, &parse_callchain_opt,
1034 OPT_INCR('v', "verbose", &verbose,
1035 "be more verbose (show counter open errors, etc)"),
1036 OPT_BOOLEAN('q', "quiet", &quiet, "don't print any message"),
1037 OPT_BOOLEAN('s', "stat", &record.opts.inherit_stat,
1038 "per thread counts"),
1039 OPT_BOOLEAN('d', "data", &record.opts.sample_address,
1040 "Sample addresses"),
1041 OPT_BOOLEAN('T', "timestamp", &record.opts.sample_time, "Sample timestamps"),
1042 OPT_BOOLEAN('P', "period", &record.opts.period, "Sample period"),
1043 OPT_BOOLEAN('n', "no-samples", &record.opts.no_samples,
1045 OPT_BOOLEAN('N', "no-buildid-cache", &record.no_buildid_cache,
1046 "do not update the buildid cache"),
1047 OPT_BOOLEAN('B', "no-buildid", &record.no_buildid,
1048 "do not collect buildids in perf.data"),
1049 OPT_CALLBACK('G', "cgroup", &record.evlist, "name",
1050 "monitor event in cgroup name only",
1052 OPT_STRING('u', "uid", &record.opts.target.uid_str, "user",
1055 OPT_CALLBACK_NOOPT('b', "branch-any", &record.opts.branch_stack,
1056 "branch any", "sample any taken branches",
1057 parse_branch_stack),
1059 OPT_CALLBACK('j', "branch-filter", &record.opts.branch_stack,
1060 "branch filter mask", "branch stack filter modes",
1061 parse_branch_stack),
1065 int cmd_record(int argc, const char **argv, const char *prefix __maybe_unused)
1068 struct perf_evsel *pos;
1069 struct perf_evlist *evsel_list;
1070 struct perf_record *rec = &record;
1071 char errbuf[BUFSIZ];
1073 evsel_list = perf_evlist__new(NULL, NULL);
1074 if (evsel_list == NULL)
1077 rec->evlist = evsel_list;
1079 argc = parse_options(argc, argv, record_options, record_usage,
1080 PARSE_OPT_STOP_AT_NON_OPTION);
1081 if (!argc && perf_target__none(&rec->opts.target))
1082 usage_with_options(record_usage, record_options);
1084 if (rec->force && rec->append_file) {
1085 ui__error("Can't overwrite and append at the same time."
1086 " You need to choose between -f and -A");
1087 usage_with_options(record_usage, record_options);
1088 } else if (rec->append_file) {
1089 rec->write_mode = WRITE_APPEND;
1091 rec->write_mode = WRITE_FORCE;
1094 if (nr_cgroups && !rec->opts.target.system_wide) {
1095 ui__error("cgroup monitoring only available in"
1096 " system-wide mode\n");
1097 usage_with_options(record_usage, record_options);
1102 if (symbol_conf.kptr_restrict)
1104 "WARNING: Kernel address maps (/proc/{kallsyms,modules}) are restricted,\n"
1105 "check /proc/sys/kernel/kptr_restrict.\n\n"
1106 "Samples in kernel functions may not be resolved if a suitable vmlinux\n"
1107 "file is not found in the buildid cache or in the vmlinux path.\n\n"
1108 "Samples in kernel modules won't be resolved at all.\n\n"
1109 "If some relocation was applied (e.g. kexec) symbols may be misresolved\n"
1110 "even with a suitable vmlinux or kallsyms file.\n\n");
1112 if (rec->no_buildid_cache || rec->no_buildid)
1113 disable_buildid_cache();
1115 if (evsel_list->nr_entries == 0 &&
1116 perf_evlist__add_default(evsel_list) < 0) {
1117 pr_err("Not enough memory for event selector list\n");
1118 goto out_symbol_exit;
1121 err = perf_target__validate(&rec->opts.target);
1123 perf_target__strerror(&rec->opts.target, err, errbuf, BUFSIZ);
1124 ui__warning("%s", errbuf);
1127 err = perf_target__parse_uid(&rec->opts.target);
1129 int saved_errno = errno;
1131 perf_target__strerror(&rec->opts.target, err, errbuf, BUFSIZ);
1132 ui__error("%s", errbuf);
1139 if (perf_evlist__create_maps(evsel_list, &rec->opts.target) < 0)
1140 usage_with_options(record_usage, record_options);
1142 list_for_each_entry(pos, &evsel_list->entries, node) {
1143 if (perf_header__push_event(pos->attr.config, perf_evsel__name(pos)))
1147 if (rec->opts.user_interval != ULLONG_MAX)
1148 rec->opts.default_interval = rec->opts.user_interval;
1149 if (rec->opts.user_freq != UINT_MAX)
1150 rec->opts.freq = rec->opts.user_freq;
1153 * User specified count overrides default frequency.
1155 if (rec->opts.default_interval)
1157 else if (rec->opts.freq) {
1158 rec->opts.default_interval = rec->opts.freq;
1160 ui__error("frequency and count are zero, aborting\n");
1165 err = __cmd_record(&record, argc, argv);
1167 perf_evlist__delete_maps(evsel_list);