builtin-record.c 25.2 KB
Newer Older
I
Ingo Molnar 已提交
1
/*
2 3 4 5 6
 * builtin-record.c
 *
 * Builtin record command: Record the profile of a workload
 * (or a CPU, or a PID) into the perf.data output file - for
 * later analysis via perf report.
I
Ingo Molnar 已提交
7
 */
8
#include "builtin.h"
9 10 11

#include "perf.h"

12
#include "util/build-id.h"
13
#include "util/util.h"
14
#include "util/parse-options.h"
15
#include "util/parse-events.h"
16

17
#include "util/header.h"
18
#include "util/event.h"
19
#include "util/evlist.h"
20
#include "util/evsel.h"
21
#include "util/debug.h"
22
#include "util/session.h"
23
#include "util/tool.h"
24
#include "util/symbol.h"
25
#include "util/cpumap.h"
26
#include "util/thread_map.h"
27
#include "util/data.h"
28

29
#include <unistd.h>
30
#include <sched.h>
31
#include <sys/mman.h>
32

33
#ifndef HAVE_ON_EXIT_SUPPORT
34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64
#ifndef ATEXIT_MAX
#define ATEXIT_MAX 32
#endif
static int __on_exit_count = 0;
typedef void (*on_exit_func_t) (int, void *);
static on_exit_func_t __on_exit_funcs[ATEXIT_MAX];
static void *__on_exit_args[ATEXIT_MAX];
static int __exitcode = 0;
static void __handle_on_exit_funcs(void);
static int on_exit(on_exit_func_t function, void *arg);
#define exit(x) (exit)(__exitcode = (x))

static int on_exit(on_exit_func_t function, void *arg)
{
	if (__on_exit_count == ATEXIT_MAX)
		return -ENOMEM;
	else if (__on_exit_count == 0)
		atexit(__handle_on_exit_funcs);
	__on_exit_funcs[__on_exit_count] = function;
	__on_exit_args[__on_exit_count++] = arg;
	return 0;
}

static void __handle_on_exit_funcs(void)
{
	int i;
	for (i = 0; i < __on_exit_count; i++)
		__on_exit_funcs[i] (__exitcode, __on_exit_args[i]);
}
#endif

65
struct record {
66
	struct perf_tool	tool;
67
	struct record_opts	opts;
68
	u64			bytes_written;
69
	struct perf_data_file	file;
70 71 72 73 74 75 76
	struct perf_evlist	*evlist;
	struct perf_session	*session;
	const char		*progname;
	int			realtime_prio;
	bool			no_buildid;
	bool			no_buildid_cache;
	long			samples;
77
};
78

79
static int record__write(struct record *rec, void *bf, size_t size)
80
{
81
	if (perf_data_file__write(rec->session->file, bf, size) < 0) {
82 83
		pr_err("failed to write perf data, error: %m\n");
		return -1;
84
	}
85

86
	rec->bytes_written += size;
87
	return 0;
88 89
}

90
static int process_synthesized_event(struct perf_tool *tool,
91
				     union perf_event *event,
92 93
				     struct perf_sample *sample __maybe_unused,
				     struct machine *machine __maybe_unused)
94
{
95 96
	struct record *rec = container_of(tool, struct record, tool);
	return record__write(rec, event, event->header.size);
97 98
}

99
static int record__mmap_read(struct record *rec, struct perf_mmap *md)
100
{
101
	unsigned int head = perf_mmap__read_head(md);
102
	unsigned int old = md->prev;
J
Jiri Olsa 已提交
103
	unsigned char *data = md->base + page_size;
104 105
	unsigned long size;
	void *buf;
106
	int rc = 0;
107

108
	if (old == head)
109
		return 0;
110

111
	rec->samples++;
112 113 114 115 116 117 118

	size = head - old;

	if ((old & md->mask) + size != (head & md->mask)) {
		buf = &data[old & md->mask];
		size = md->mask + 1 - (old & md->mask);
		old += size;
119

120
		if (record__write(rec, buf, size) < 0) {
121 122 123
			rc = -1;
			goto out;
		}
124 125 126 127 128
	}

	buf = &data[old & md->mask];
	size = head - old;
	old += size;
129

130
	if (record__write(rec, buf, size) < 0) {
131 132 133
		rc = -1;
		goto out;
	}
134 135

	md->prev = old;
136
	perf_mmap__write_tail(md, old);
137 138 139

out:
	return rc;
140 141 142
}

static volatile int done = 0;
143
static volatile int signr = -1;
144
static volatile int child_finished = 0;
145

146
static void sig_handler(int sig)
147
{
148 149 150
	if (sig == SIGCHLD)
		child_finished = 1;

151
	done = 1;
152 153 154
	signr = sig;
}

155
static void record__sig_exit(int exit_status __maybe_unused, void *arg)
156
{
157
	struct record *rec = arg;
158 159
	int status;

160
	if (rec->evlist->workload.pid > 0) {
161
		if (!child_finished)
162
			kill(rec->evlist->workload.pid, SIGTERM);
163 164 165

		wait(&status);
		if (WIFSIGNALED(status))
166
			psignal(WTERMSIG(status), rec->progname);
167
	}
168

169
	if (signr == -1 || signr == SIGUSR1)
170 171 172
		return;

	signal(signr, SIG_DFL);
173 174
}

175
static int record__open(struct record *rec)
176
{
177
	char msg[512];
178
	struct perf_evsel *pos;
179 180
	struct perf_evlist *evlist = rec->evlist;
	struct perf_session *session = rec->session;
181
	struct record_opts *opts = &rec->opts;
182
	int rc = 0;
183

184
	perf_evlist__config(evlist, opts);
185

186
	evlist__for_each(evlist, pos) {
187
try_again:
188
		if (perf_evsel__open(pos, evlist->cpus, evlist->threads) < 0) {
189
			if (perf_evsel__fallback(pos, errno, msg, sizeof(msg))) {
190
				if (verbose)
191
					ui__warning("%s\n", msg);
192 193
				goto try_again;
			}
194

195 196 197 198
			rc = -errno;
			perf_evsel__open_strerror(pos, &opts->target,
						  errno, msg, sizeof(msg));
			ui__error("%s\n", msg);
199
			goto out;
L
Li Zefan 已提交
200 201
		}
	}
202

203
	if (perf_evlist__apply_filters(evlist)) {
204 205
		error("failed to set filter with %d (%s)\n", errno,
			strerror(errno));
206 207
		rc = -1;
		goto out;
208 209
	}

210
	if (perf_evlist__mmap(evlist, opts->mmap_pages, false) < 0) {
211 212 213 214 215
		if (errno == EPERM) {
			pr_err("Permission error mapping pages.\n"
			       "Consider increasing "
			       "/proc/sys/kernel/perf_event_mlock_kb,\n"
			       "or try again with a smaller value of -m/--mmap_pages.\n"
216
			       "(current value: %u)\n", opts->mmap_pages);
217 218 219 220 221 222
			rc = -errno;
		} else {
			pr_err("failed to mmap with %d (%s)\n", errno, strerror(errno));
			rc = -errno;
		}
		goto out;
223
	}
224

225
	session->evlist = evlist;
226
	perf_session__set_id_hdr_size(session);
227 228
out:
	return rc;
229 230
}

231
static int process_buildids(struct record *rec)
232
{
233 234
	struct perf_data_file *file  = &rec->file;
	struct perf_session *session = rec->session;
235
	u64 start = session->header.data_offset;
236

237
	u64 size = lseek(file->fd, 0, SEEK_CUR);
238 239 240
	if (size == 0)
		return 0;

241 242
	return __perf_session__process_events(session, start,
					      size - start,
243 244 245
					      size, &build_id__mark_dso_hit_ops);
}

246
static void record__exit(int status, void *arg)
247
{
248
	struct record *rec = arg;
249
	struct perf_data_file *file = &rec->file;
250

251 252 253
	if (status != 0)
		return;

254
	if (!file->is_pipe) {
255 256 257 258 259
		rec->session->header.data_size += rec->bytes_written;

		if (!rec->no_buildid)
			process_buildids(rec);
		perf_session__write_header(rec->session, rec->evlist,
260
					   file->fd, true);
261 262
		perf_session__delete(rec->session);
		perf_evlist__delete(rec->evlist);
263
		symbol__exit();
264
	}
265 266
}

267
static void perf_event__synthesize_guest_os(struct machine *machine, void *data)
268 269
{
	int err;
270
	struct perf_tool *tool = data;
271 272 273 274 275 276 277 278
	/*
	 *As for guest kernel when processing subcommand record&report,
	 *we arrange module mmap prior to guest kernel mmap and trigger
	 *a preload dso because default guest module symbols are loaded
	 *from guest kallsyms instead of /lib/modules/XXX/XXX. This
	 *method is used to avoid symbol missing when the first addr is
	 *in module instead of in guest kernel.
	 */
279
	err = perf_event__synthesize_modules(tool, process_synthesized_event,
280
					     machine);
281 282
	if (err < 0)
		pr_err("Couldn't record guest kernel [%d]'s reference"
283
		       " relocation symbol.\n", machine->pid);
284 285 286 287 288

	/*
	 * We use _stext for guest kernel because guest kernel's /proc/kallsyms
	 * have no _text sometimes.
	 */
289
	err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
290
						 machine);
291 292
	if (err < 0)
		pr_err("Couldn't record guest kernel [%d]'s reference"
293
		       " relocation symbol.\n", machine->pid);
294 295
}

296 297 298 299 300
static struct perf_event_header finished_round_event = {
	.size = sizeof(struct perf_event_header),
	.type = PERF_RECORD_FINISHED_ROUND,
};

301
static int record__mmap_read_all(struct record *rec)
302
{
303
	int i;
304
	int rc = 0;
305

306
	for (i = 0; i < rec->evlist->nr_mmaps; i++) {
307
		if (rec->evlist->mmap[i].base) {
308
			if (record__mmap_read(rec, &rec->evlist->mmap[i]) != 0) {
309 310 311 312
				rc = -1;
				goto out;
			}
		}
313 314
	}

315
	if (perf_header__has_feat(&rec->session->header, HEADER_TRACING_DATA))
316
		rc = record__write(rec, &finished_round_event, sizeof(finished_round_event));
317 318 319

out:
	return rc;
320 321
}

322
static void record__init_features(struct record *rec)
323 324 325 326 327 328 329 330 331 332
{
	struct perf_session *session = rec->session;
	int feat;

	for (feat = HEADER_FIRST_FEATURE; feat < HEADER_LAST_FEATURE; feat++)
		perf_header__set_feat(&session->header, feat);

	if (rec->no_buildid)
		perf_header__clear_feat(&session->header, HEADER_BUILD_ID);

333
	if (!have_tracepoints(&rec->evlist->entries))
334 335 336 337 338 339
		perf_header__clear_feat(&session->header, HEADER_TRACING_DATA);

	if (!rec->opts.branch_stack)
		perf_header__clear_feat(&session->header, HEADER_BRANCH_STACK);
}

340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355
static volatile int workload_exec_errno;

/*
 * perf_evlist__prepare_workload will send a SIGUSR1
 * if the fork fails, since we asked by setting its
 * want_signal to true.
 */
static void workload_exec_failed_signal(int signo, siginfo_t *info,
					void *ucontext __maybe_unused)
{
	workload_exec_errno = info->si_value.sival_int;
	done = 1;
	signr = signo;
	child_finished = 1;
}

356
static int __cmd_record(struct record *rec, int argc, const char **argv)
357
{
358
	int err;
359
	unsigned long waking = 0;
360
	const bool forks = argc > 0;
361
	struct machine *machine;
362
	struct perf_tool *tool = &rec->tool;
363
	struct record_opts *opts = &rec->opts;
364
	struct perf_data_file *file = &rec->file;
365
	struct perf_session *session;
366
	bool disabled = false;
367

368
	rec->progname = argv[0];
369

370
	on_exit(record__sig_exit, rec);
371 372
	signal(SIGCHLD, sig_handler);
	signal(SIGINT, sig_handler);
373
	signal(SIGTERM, sig_handler);
374

375
	session = perf_session__new(file, false, NULL);
376
	if (session == NULL) {
377 378 379 380
		pr_err("Not enough memory for reading perf file header\n");
		return -1;
	}

381 382
	rec->session = session;

383
	record__init_features(rec);
384

385
	if (forks) {
386
		err = perf_evlist__prepare_workload(rec->evlist, &opts->target,
387
						    argv, file->is_pipe,
388
						    workload_exec_failed_signal);
389 390 391
		if (err < 0) {
			pr_err("Couldn't run the workload!\n");
			goto out_delete_session;
392 393 394
		}
	}

395
	if (record__open(rec) != 0) {
396 397 398
		err = -1;
		goto out_delete_session;
	}
399

400
	if (!rec->evlist->nr_groups)
401 402
		perf_header__clear_feat(&session->header, HEADER_GROUP_DESC);

403
	/*
404
	 * perf_session__delete(session) will be called at record__exit()
405
	 */
406
	on_exit(record__exit, rec);
407

408 409
	if (file->is_pipe) {
		err = perf_header__write_pipe(file->fd);
410
		if (err < 0)
411
			goto out_delete_session;
412
	} else {
413
		err = perf_session__write_header(session, rec->evlist,
414
						 file->fd, false);
415
		if (err < 0)
416
			goto out_delete_session;
417 418
	}

419
	if (!rec->no_buildid
420
	    && !perf_header__has_feat(&session->header, HEADER_BUILD_ID)) {
421
		pr_err("Couldn't generate buildids. "
422
		       "Use --no-buildid to profile anyway.\n");
423 424
		err = -1;
		goto out_delete_session;
425 426
	}

427
	machine = &session->machines.host;
428

429
	if (file->is_pipe) {
430
		err = perf_event__synthesize_attrs(tool, session,
431
						   process_synthesized_event);
432 433
		if (err < 0) {
			pr_err("Couldn't synthesize attrs.\n");
434
			goto out_delete_session;
435
		}
436

437
		if (have_tracepoints(&rec->evlist->entries)) {
438 439 440 441 442 443 444 445
			/*
			 * FIXME err <= 0 here actually means that
			 * there were no tracepoints so its not really
			 * an error, just that we don't need to
			 * synthesize anything.  We really have to
			 * return this more properly and also
			 * propagate errors that now are calling die()
			 */
446
			err = perf_event__synthesize_tracing_data(tool, file->fd, rec->evlist,
447
								  process_synthesized_event);
448 449
			if (err <= 0) {
				pr_err("Couldn't record tracing data.\n");
450
				goto out_delete_session;
451
			}
452
			rec->bytes_written += err;
453
		}
454 455
	}

456
	err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
457
						 machine);
458 459 460 461
	if (err < 0)
		pr_err("Couldn't record kernel reference relocation symbol\n"
		       "Symbol resolution may be skewed if relocation was used (e.g. kexec).\n"
		       "Check /proc/kallsyms permission or run as root.\n");
462

463
	err = perf_event__synthesize_modules(tool, process_synthesized_event,
464
					     machine);
465 466 467 468 469
	if (err < 0)
		pr_err("Couldn't record kernel module information.\n"
		       "Symbol resolution may be skewed if relocation was used (e.g. kexec).\n"
		       "Check /proc/modules permission or run as root.\n");

470
	if (perf_guest) {
471 472
		machines__process_guests(&session->machines,
					 perf_event__synthesize_guest_os, tool);
473
	}
474

475
	err = __machine__synthesize_threads(machine, tool, &opts->target, rec->evlist->threads,
476
					    process_synthesized_event, opts->sample_address);
477 478 479
	if (err != 0)
		goto out_delete_session;

480
	if (rec->realtime_prio) {
481 482
		struct sched_param param;

483
		param.sched_priority = rec->realtime_prio;
484
		if (sched_setscheduler(0, SCHED_FIFO, &param)) {
485
			pr_err("Could not set realtime priority.\n");
486 487
			err = -1;
			goto out_delete_session;
488 489 490
		}
	}

491 492 493 494 495
	/*
	 * When perf is starting the traced process, all the events
	 * (apart from group members) have enable_on_exec=1 set,
	 * so don't spoil it by prematurely enabling them.
	 */
496
	if (!target__none(&opts->target) && !opts->initial_delay)
497
		perf_evlist__enable(rec->evlist);
498

499 500 501
	/*
	 * Let the child rip
	 */
502
	if (forks)
503
		perf_evlist__start_workload(rec->evlist);
504

505 506 507 508 509
	if (opts->initial_delay) {
		usleep(opts->initial_delay * 1000);
		perf_evlist__enable(rec->evlist);
	}

510
	for (;;) {
511
		int hits = rec->samples;
512

513
		if (record__mmap_read_all(rec) < 0) {
514 515 516
			err = -1;
			goto out_delete_session;
		}
517

518
		if (hits == rec->samples) {
519 520
			if (done)
				break;
521
			err = poll(rec->evlist->pollfd, rec->evlist->nr_fds, -1);
522 523 524
			waking++;
		}

525 526 527 528 529
		/*
		 * When perf is starting the traced process, at the end events
		 * die with the process and we wait for that. Thus no need to
		 * disable events in this case.
		 */
530
		if (done && !disabled && !target__none(&opts->target)) {
531
			perf_evlist__disable(rec->evlist);
532 533
			disabled = true;
		}
534 535
	}

536 537 538 539 540 541 542 543
	if (forks && workload_exec_errno) {
		char msg[512];
		const char *emsg = strerror_r(workload_exec_errno, msg, sizeof(msg));
		pr_err("Workload failed: %s\n", emsg);
		err = -1;
		goto out_delete_session;
	}

544
	if (quiet || signr == SIGUSR1)
545 546
		return 0;

547 548
	fprintf(stderr, "[ perf record: Woken up %ld times to write data ]\n", waking);

549 550 551 552
	/*
	 * Approximate RIP event size: 24 bytes.
	 */
	fprintf(stderr,
553
		"[ perf record: Captured and wrote %.3f MB %s (~%" PRIu64 " samples) ]\n",
554
		(double)rec->bytes_written / 1024.0 / 1024.0,
555
		file->path,
556
		rec->bytes_written / 24);
557

558
	return 0;
559 560 561 562

out_delete_session:
	perf_session__delete(session);
	return err;
563
}
564

565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582
#define BRANCH_OPT(n, m) \
	{ .name = n, .mode = (m) }

#define BRANCH_END { .name = NULL }

struct branch_mode {
	const char *name;
	int mode;
};

static const struct branch_mode branch_modes[] = {
	BRANCH_OPT("u", PERF_SAMPLE_BRANCH_USER),
	BRANCH_OPT("k", PERF_SAMPLE_BRANCH_KERNEL),
	BRANCH_OPT("hv", PERF_SAMPLE_BRANCH_HV),
	BRANCH_OPT("any", PERF_SAMPLE_BRANCH_ANY),
	BRANCH_OPT("any_call", PERF_SAMPLE_BRANCH_ANY_CALL),
	BRANCH_OPT("any_ret", PERF_SAMPLE_BRANCH_ANY_RETURN),
	BRANCH_OPT("ind_call", PERF_SAMPLE_BRANCH_IND_CALL),
583 584 585
	BRANCH_OPT("abort_tx", PERF_SAMPLE_BRANCH_ABORT_TX),
	BRANCH_OPT("in_tx", PERF_SAMPLE_BRANCH_IN_TX),
	BRANCH_OPT("no_tx", PERF_SAMPLE_BRANCH_NO_TX),
586 587 588 589
	BRANCH_END
};

static int
590
parse_branch_stack(const struct option *opt, const char *str, int unset)
591 592 593 594 595 596 597 598
{
#define ONLY_PLM \
	(PERF_SAMPLE_BRANCH_USER	|\
	 PERF_SAMPLE_BRANCH_KERNEL	|\
	 PERF_SAMPLE_BRANCH_HV)

	uint64_t *mode = (uint64_t *)opt->value;
	const struct branch_mode *br;
599
	char *s, *os = NULL, *p;
600 601
	int ret = -1;

602 603
	if (unset)
		return 0;
604

605 606 607 608
	/*
	 * cannot set it twice, -b + --branch-filter for instance
	 */
	if (*mode)
609 610
		return -1;

611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631
	/* str may be NULL in case no arg is passed to -b */
	if (str) {
		/* because str is read-only */
		s = os = strdup(str);
		if (!s)
			return -1;

		for (;;) {
			p = strchr(s, ',');
			if (p)
				*p = '\0';

			for (br = branch_modes; br->name; br++) {
				if (!strcasecmp(s, br->name))
					break;
			}
			if (!br->name) {
				ui__warning("unknown branch filter %s,"
					    " check man page\n", s);
				goto error;
			}
632

633
			*mode |= br->mode;
634

635 636
			if (!p)
				break;
637

638 639
			s = p + 1;
		}
640 641 642
	}
	ret = 0;

643
	/* default to any branch */
644
	if ((*mode & ~ONLY_PLM) == 0) {
645
		*mode = PERF_SAMPLE_BRANCH_ANY;
646 647 648 649 650 651
	}
error:
	free(os);
	return ret;
}

652
#ifdef HAVE_DWARF_UNWIND_SUPPORT
653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677
static int get_stack_size(char *str, unsigned long *_size)
{
	char *endptr;
	unsigned long size;
	unsigned long max_size = round_down(USHRT_MAX, sizeof(u64));

	size = strtoul(str, &endptr, 0);

	do {
		if (*endptr)
			break;

		size = round_up(size, sizeof(u64));
		if (!size || size > max_size)
			break;

		*_size = size;
		return 0;

	} while (0);

	pr_err("callchain: Incorrect stack dump size (max %ld): %s\n",
	       max_size, str);
	return -1;
}
678
#endif /* HAVE_DWARF_UNWIND_SUPPORT */
679

680
int record_parse_callchain(const char *arg, struct record_opts *opts)
681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699
{
	char *tok, *name, *saveptr = NULL;
	char *buf;
	int ret = -1;

	/* We need buffer that we know we can write to. */
	buf = malloc(strlen(arg) + 1);
	if (!buf)
		return -ENOMEM;

	strcpy(buf, arg);

	tok = strtok_r((char *)buf, ",", &saveptr);
	name = tok ? : (char *)buf;

	do {
		/* Framepointer style */
		if (!strncmp(name, "fp", sizeof("fp"))) {
			if (!strtok_r(NULL, ",", &saveptr)) {
700
				opts->call_graph = CALLCHAIN_FP;
701 702 703 704 705 706
				ret = 0;
			} else
				pr_err("callchain: No more arguments "
				       "needed for -g fp\n");
			break;

707
#ifdef HAVE_DWARF_UNWIND_SUPPORT
708 709
		/* Dwarf style */
		} else if (!strncmp(name, "dwarf", sizeof("dwarf"))) {
710 711
			const unsigned long default_stack_dump_size = 8192;

712
			ret = 0;
713 714
			opts->call_graph = CALLCHAIN_DWARF;
			opts->stack_dump_size = default_stack_dump_size;
715 716 717 718 719 720

			tok = strtok_r(NULL, ",", &saveptr);
			if (tok) {
				unsigned long size = 0;

				ret = get_stack_size(tok, &size);
721
				opts->stack_dump_size = size;
722
			}
723
#endif /* HAVE_DWARF_UNWIND_SUPPORT */
724
		} else {
J
Jiri Olsa 已提交
725
			pr_err("callchain: Unknown --call-graph option "
726 727 728 729 730 731 732
			       "value: %s\n", arg);
			break;
		}

	} while (0);

	free(buf);
J
Jiri Olsa 已提交
733 734 735
	return ret;
}

736
static void callchain_debug(struct record_opts *opts)
J
Jiri Olsa 已提交
737
{
738 739 740
	static const char *str[CALLCHAIN_MAX] = { "NONE", "FP", "DWARF" };

	pr_debug("callchain: type %s\n", str[opts->call_graph]);
741

J
Jiri Olsa 已提交
742 743 744 745 746 747 748 749 750
	if (opts->call_graph == CALLCHAIN_DWARF)
		pr_debug("callchain: stack dump size %d\n",
			 opts->stack_dump_size);
}

int record_parse_callchain_opt(const struct option *opt,
			       const char *arg,
			       int unset)
{
751
	struct record_opts *opts = opt->value;
J
Jiri Olsa 已提交
752 753
	int ret;

754 755
	opts->call_graph_enabled = !unset;

J
Jiri Olsa 已提交
756 757 758 759 760 761 762 763
	/* --no-call-graph */
	if (unset) {
		opts->call_graph = CALLCHAIN_NONE;
		pr_debug("callchain: disabled\n");
		return 0;
	}

	ret = record_parse_callchain(arg, opts);
764
	if (!ret)
J
Jiri Olsa 已提交
765
		callchain_debug(opts);
766 767 768 769

	return ret;
}

J
Jiri Olsa 已提交
770 771 772 773
int record_callchain_opt(const struct option *opt,
			 const char *arg __maybe_unused,
			 int unset __maybe_unused)
{
774
	struct record_opts *opts = opt->value;
J
Jiri Olsa 已提交
775

776 777
	opts->call_graph_enabled = !unset;

J
Jiri Olsa 已提交
778 779 780 781 782 783 784
	if (opts->call_graph == CALLCHAIN_NONE)
		opts->call_graph = CALLCHAIN_FP;

	callchain_debug(opts);
	return 0;
}

785 786 787 788 789 790 791 792 793 794
static int perf_record_config(const char *var, const char *value, void *cb)
{
	struct record *rec = cb;

	if (!strcmp(var, "record.call-graph"))
		return record_parse_callchain(value, &rec->opts);

	return perf_default_config(var, value, cb);
}

795
static const char * const record_usage[] = {
796 797
	"perf record [<options>] [<command>]",
	"perf record [<options>] -- <command> [<options>]",
798 799 800
	NULL
};

801
/*
802 803
 * XXX Ideally would be local to cmd_record() and passed to a record__new
 * because we need to have access to it in record__exit, that is called
804 805 806 807 808 809 810
 * after cmd_record() exits, but since record_options need to be accessible to
 * builtin-script, leave it here.
 *
 * At least we don't ouch it in all the other functions here directly.
 *
 * Just say no to tons of global variables, sigh.
 */
811
static struct record record = {
812 813 814 815
	.opts = {
		.mmap_pages	     = UINT_MAX,
		.user_freq	     = UINT_MAX,
		.user_interval	     = ULLONG_MAX,
816
		.freq		     = 4000,
N
Namhyung Kim 已提交
817 818
		.target		     = {
			.uses_mmap   = true,
819
			.default_per_cpu = true,
N
Namhyung Kim 已提交
820
		},
821 822
	},
};
823

J
Jiri Olsa 已提交
824
#define CALLCHAIN_HELP "setup and enables call-graph (stack chain/backtrace) recording: "
825

826
#ifdef HAVE_DWARF_UNWIND_SUPPORT
J
Jiri Olsa 已提交
827
const char record_callchain_help[] = CALLCHAIN_HELP "fp dwarf";
828
#else
J
Jiri Olsa 已提交
829
const char record_callchain_help[] = CALLCHAIN_HELP "fp";
830 831
#endif

832 833 834
/*
 * XXX Will stay a global variable till we fix builtin-script.c to stop messing
 * with it and switch to use the library functions in perf_evlist that came
835
 * from builtin-record.c, i.e. use record_opts,
836 837 838
 * perf_evlist__prepare_workload, etc instead of fork+exec'in 'perf record',
 * using pipes, etc.
 */
839
const struct option record_options[] = {
840
	OPT_CALLBACK('e', "event", &record.evlist, "event",
841
		     "event selector. use 'perf list' to list available events",
842
		     parse_events_option),
843
	OPT_CALLBACK(0, "filter", &record.evlist, "filter",
L
Li Zefan 已提交
844
		     "event filter", parse_filter),
845
	OPT_STRING('p', "pid", &record.opts.target.pid, "pid",
846
		    "record events on existing process id"),
847
	OPT_STRING('t', "tid", &record.opts.target.tid, "tid",
848
		    "record events on existing thread id"),
849
	OPT_INTEGER('r', "realtime", &record.realtime_prio,
850
		    "collect data with this RT SCHED_FIFO priority"),
851
	OPT_BOOLEAN(0, "no-buffering", &record.opts.no_buffering,
852
		    "collect data without buffering"),
853
	OPT_BOOLEAN('R', "raw-samples", &record.opts.raw_samples,
854
		    "collect raw sample records from all opened counters"),
855
	OPT_BOOLEAN('a', "all-cpus", &record.opts.target.system_wide,
856
			    "system-wide collection from all CPUs"),
857
	OPT_STRING('C', "cpu", &record.opts.target.cpu_list, "cpu",
858
		    "list of cpus to monitor"),
859
	OPT_U64('c', "count", &record.opts.user_interval, "event period to sample"),
860
	OPT_STRING('o', "output", &record.file.path, "file",
I
Ingo Molnar 已提交
861
		    "output file name"),
862 863 864
	OPT_BOOLEAN_SET('i', "no-inherit", &record.opts.no_inherit,
			&record.opts.no_inherit_set,
			"child tasks do not inherit counters"),
865
	OPT_UINTEGER('F', "freq", &record.opts.user_freq, "profile at this frequency"),
866 867 868
	OPT_CALLBACK('m', "mmap-pages", &record.opts.mmap_pages, "pages",
		     "number of mmap data pages",
		     perf_evlist__parse_mmap_pages),
869
	OPT_BOOLEAN(0, "group", &record.opts.group,
870
		    "put the counters into a counter group"),
J
Jiri Olsa 已提交
871 872 873 874 875 876
	OPT_CALLBACK_NOOPT('g', NULL, &record.opts,
			   NULL, "enables call-graph recording" ,
			   &record_callchain_opt),
	OPT_CALLBACK(0, "call-graph", &record.opts,
		     "mode[,dump_size]", record_callchain_help,
		     &record_parse_callchain_opt),
877
	OPT_INCR('v', "verbose", &verbose,
878
		    "be more verbose (show counter open errors, etc)"),
879
	OPT_BOOLEAN('q', "quiet", &quiet, "don't print any message"),
880
	OPT_BOOLEAN('s', "stat", &record.opts.inherit_stat,
881
		    "per thread counts"),
882
	OPT_BOOLEAN('d', "data", &record.opts.sample_address,
883
		    "Sample addresses"),
884
	OPT_BOOLEAN('T', "timestamp", &record.opts.sample_time, "Sample timestamps"),
885
	OPT_BOOLEAN('P', "period", &record.opts.period, "Sample period"),
886
	OPT_BOOLEAN('n', "no-samples", &record.opts.no_samples,
887
		    "don't sample"),
888
	OPT_BOOLEAN('N', "no-buildid-cache", &record.no_buildid_cache,
889
		    "do not update the buildid cache"),
890
	OPT_BOOLEAN('B', "no-buildid", &record.no_buildid,
891
		    "do not collect buildids in perf.data"),
892
	OPT_CALLBACK('G', "cgroup", &record.evlist, "name",
S
Stephane Eranian 已提交
893 894
		     "monitor event in cgroup name only",
		     parse_cgroups),
895
	OPT_UINTEGER('D', "delay", &record.opts.initial_delay,
896
		  "ms to wait before starting measurement after program start"),
897 898
	OPT_STRING('u', "uid", &record.opts.target.uid_str, "user",
		   "user to profile"),
899 900 901 902 903 904 905

	OPT_CALLBACK_NOOPT('b', "branch-any", &record.opts.branch_stack,
		     "branch any", "sample any taken branches",
		     parse_branch_stack),

	OPT_CALLBACK('j', "branch-filter", &record.opts.branch_stack,
		     "branch filter mask", "branch stack filter modes",
906
		     parse_branch_stack),
907 908
	OPT_BOOLEAN('W', "weight", &record.opts.sample_weight,
		    "sample by weight (on special events only)"),
909 910
	OPT_BOOLEAN(0, "transaction", &record.opts.sample_transaction,
		    "sample transaction flags (special events only)"),
911 912
	OPT_BOOLEAN(0, "per-thread", &record.opts.target.per_thread,
		    "use per-thread mmaps"),
913 914 915
	OPT_END()
};

916
int cmd_record(int argc, const char **argv, const char *prefix __maybe_unused)
917
{
918
	int err = -ENOMEM;
919
	struct record *rec = &record;
920
	char errbuf[BUFSIZ];
921

922 923
	rec->evlist = perf_evlist__new();
	if (rec->evlist == NULL)
924 925
		return -ENOMEM;

926 927
	perf_config(perf_record_config, rec);

928
	argc = parse_options(argc, argv, record_options, record_usage,
929
			    PARSE_OPT_STOP_AT_NON_OPTION);
930
	if (!argc && target__none(&rec->opts.target))
931
		usage_with_options(record_usage, record_options);
932

933
	if (nr_cgroups && !rec->opts.target.system_wide) {
934 935
		ui__error("cgroup monitoring only available in"
			  " system-wide mode\n");
S
Stephane Eranian 已提交
936 937 938
		usage_with_options(record_usage, record_options);
	}

939
	symbol__init();
940

941
	if (symbol_conf.kptr_restrict)
942 943 944 945 946 947 948 949
		pr_warning(
"WARNING: Kernel address maps (/proc/{kallsyms,modules}) are restricted,\n"
"check /proc/sys/kernel/kptr_restrict.\n\n"
"Samples in kernel functions may not be resolved if a suitable vmlinux\n"
"file is not found in the buildid cache or in the vmlinux path.\n\n"
"Samples in kernel modules won't be resolved at all.\n\n"
"If some relocation was applied (e.g. kexec) symbols may be misresolved\n"
"even with a suitable vmlinux or kallsyms file.\n\n");
950

951
	if (rec->no_buildid_cache || rec->no_buildid)
952
		disable_buildid_cache();
953

954 955
	if (rec->evlist->nr_entries == 0 &&
	    perf_evlist__add_default(rec->evlist) < 0) {
956 957
		pr_err("Not enough memory for event selector list\n");
		goto out_symbol_exit;
958
	}
959

960 961 962
	if (rec->opts.target.tid && !rec->opts.no_inherit_set)
		rec->opts.no_inherit = true;

963
	err = target__validate(&rec->opts.target);
964
	if (err) {
965
		target__strerror(&rec->opts.target, err, errbuf, BUFSIZ);
966 967 968
		ui__warning("%s", errbuf);
	}

969
	err = target__parse_uid(&rec->opts.target);
970 971
	if (err) {
		int saved_errno = errno;
972

973
		target__strerror(&rec->opts.target, err, errbuf, BUFSIZ);
974
		ui__error("%s", errbuf);
975 976

		err = -saved_errno;
977
		goto out_symbol_exit;
978
	}
979

980
	err = -ENOMEM;
981
	if (perf_evlist__create_maps(rec->evlist, &rec->opts.target) < 0)
982
		usage_with_options(record_usage, record_options);
983

984
	if (record_opts__config(&rec->opts)) {
985
		err = -EINVAL;
986
		goto out_symbol_exit;
987 988
	}

989
	err = __cmd_record(&record, argc, argv);
990 991
out_symbol_exit:
	symbol__exit();
992
	return err;
993
}