builtin-record.c 25.3 KB
Newer Older
I
Ingo Molnar 已提交
1
/*
2 3 4 5 6
 * builtin-record.c
 *
 * Builtin record command: Record the profile of a workload
 * (or a CPU, or a PID) into the perf.data output file - for
 * later analysis via perf report.
I
Ingo Molnar 已提交
7
 */
8
#include "builtin.h"
9 10 11

#include "perf.h"

12
#include "util/build-id.h"
13
#include "util/util.h"
14
#include "util/parse-options.h"
15
#include "util/parse-events.h"
16

17
#include "util/header.h"
18
#include "util/event.h"
19
#include "util/evlist.h"
20
#include "util/evsel.h"
21
#include "util/debug.h"
22
#include "util/session.h"
23
#include "util/tool.h"
24
#include "util/symbol.h"
25
#include "util/cpumap.h"
26
#include "util/thread_map.h"
27
#include "util/data.h"
28

29
#include <unistd.h>
30
#include <sched.h>
31
#include <sys/mman.h>
32

33
#ifndef HAVE_ON_EXIT_SUPPORT
34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64
#ifndef ATEXIT_MAX
#define ATEXIT_MAX 32
#endif
static int __on_exit_count = 0;
typedef void (*on_exit_func_t) (int, void *);
static on_exit_func_t __on_exit_funcs[ATEXIT_MAX];
static void *__on_exit_args[ATEXIT_MAX];
static int __exitcode = 0;
static void __handle_on_exit_funcs(void);
static int on_exit(on_exit_func_t function, void *arg);
#define exit(x) (exit)(__exitcode = (x))

static int on_exit(on_exit_func_t function, void *arg)
{
	if (__on_exit_count == ATEXIT_MAX)
		return -ENOMEM;
	else if (__on_exit_count == 0)
		atexit(__handle_on_exit_funcs);
	__on_exit_funcs[__on_exit_count] = function;
	__on_exit_args[__on_exit_count++] = arg;
	return 0;
}

static void __handle_on_exit_funcs(void)
{
	int i;
	for (i = 0; i < __on_exit_count; i++)
		__on_exit_funcs[i] (__exitcode, __on_exit_args[i]);
}
#endif

65
struct record {
66
	struct perf_tool	tool;
67
	struct record_opts	opts;
68
	u64			bytes_written;
69
	struct perf_data_file	file;
70 71 72 73 74 75 76
	struct perf_evlist	*evlist;
	struct perf_session	*session;
	const char		*progname;
	int			realtime_prio;
	bool			no_buildid;
	bool			no_buildid_cache;
	long			samples;
77
};
78

79
static int record__write(struct record *rec, void *bf, size_t size)
80
{
81
	if (perf_data_file__write(rec->session->file, bf, size) < 0) {
82 83
		pr_err("failed to write perf data, error: %m\n");
		return -1;
84
	}
85

86
	rec->bytes_written += size;
87
	return 0;
88 89
}

90
static int process_synthesized_event(struct perf_tool *tool,
91
				     union perf_event *event,
92 93
				     struct perf_sample *sample __maybe_unused,
				     struct machine *machine __maybe_unused)
94
{
95 96
	struct record *rec = container_of(tool, struct record, tool);
	return record__write(rec, event, event->header.size);
97 98
}

99
static int record__mmap_read(struct record *rec, struct perf_mmap *md)
100
{
101
	unsigned int head = perf_mmap__read_head(md);
102
	unsigned int old = md->prev;
J
Jiri Olsa 已提交
103
	unsigned char *data = md->base + page_size;
104 105
	unsigned long size;
	void *buf;
106
	int rc = 0;
107

108
	if (old == head)
109
		return 0;
110

111
	rec->samples++;
112 113 114 115 116 117 118

	size = head - old;

	if ((old & md->mask) + size != (head & md->mask)) {
		buf = &data[old & md->mask];
		size = md->mask + 1 - (old & md->mask);
		old += size;
119

120
		if (record__write(rec, buf, size) < 0) {
121 122 123
			rc = -1;
			goto out;
		}
124 125 126 127 128
	}

	buf = &data[old & md->mask];
	size = head - old;
	old += size;
129

130
	if (record__write(rec, buf, size) < 0) {
131 132 133
		rc = -1;
		goto out;
	}
134 135

	md->prev = old;
136
	perf_mmap__write_tail(md, old);
137 138 139

out:
	return rc;
140 141 142
}

static volatile int done = 0;
143
static volatile int signr = -1;
144
static volatile int child_finished = 0;
145

146
static void sig_handler(int sig)
147
{
148 149 150
	if (sig == SIGCHLD)
		child_finished = 1;

151
	done = 1;
152 153 154
	signr = sig;
}

155
static void record__sig_exit(int exit_status __maybe_unused, void *arg)
156
{
157
	struct record *rec = arg;
158 159
	int status;

160
	if (rec->evlist->workload.pid > 0) {
161
		if (!child_finished)
162
			kill(rec->evlist->workload.pid, SIGTERM);
163 164 165

		wait(&status);
		if (WIFSIGNALED(status))
166
			psignal(WTERMSIG(status), rec->progname);
167
	}
168

169
	if (signr == -1 || signr == SIGUSR1)
170 171 172
		return;

	signal(signr, SIG_DFL);
173 174
}

175
static int record__open(struct record *rec)
176
{
177
	char msg[512];
178
	struct perf_evsel *pos;
179 180
	struct perf_evlist *evlist = rec->evlist;
	struct perf_session *session = rec->session;
181
	struct record_opts *opts = &rec->opts;
182
	int rc = 0;
183

184
	perf_evlist__config(evlist, opts);
185

186 187
	list_for_each_entry(pos, &evlist->entries, node) {
try_again:
188
		if (perf_evsel__open(pos, evlist->cpus, evlist->threads) < 0) {
189
			if (perf_evsel__fallback(pos, errno, msg, sizeof(msg))) {
190
				if (verbose)
191
					ui__warning("%s\n", msg);
192 193
				goto try_again;
			}
194

195 196 197 198
			rc = -errno;
			perf_evsel__open_strerror(pos, &opts->target,
						  errno, msg, sizeof(msg));
			ui__error("%s\n", msg);
199
			goto out;
L
Li Zefan 已提交
200 201
		}
	}
202

203
	if (perf_evlist__apply_filters(evlist)) {
204 205
		error("failed to set filter with %d (%s)\n", errno,
			strerror(errno));
206 207
		rc = -1;
		goto out;
208 209
	}

210
	if (perf_evlist__mmap(evlist, opts->mmap_pages, false) < 0) {
211 212 213 214 215
		if (errno == EPERM) {
			pr_err("Permission error mapping pages.\n"
			       "Consider increasing "
			       "/proc/sys/kernel/perf_event_mlock_kb,\n"
			       "or try again with a smaller value of -m/--mmap_pages.\n"
216
			       "(current value: %u)\n", opts->mmap_pages);
217 218 219 220 221 222
			rc = -errno;
		} else {
			pr_err("failed to mmap with %d (%s)\n", errno, strerror(errno));
			rc = -errno;
		}
		goto out;
223
	}
224

225
	session->evlist = evlist;
226
	perf_session__set_id_hdr_size(session);
227 228
out:
	return rc;
229 230
}

231
static int process_buildids(struct record *rec)
232
{
233 234
	struct perf_data_file *file  = &rec->file;
	struct perf_session *session = rec->session;
235
	u64 start = session->header.data_offset;
236

237
	u64 size = lseek(file->fd, 0, SEEK_CUR);
238 239 240
	if (size == 0)
		return 0;

241 242
	return __perf_session__process_events(session, start,
					      size - start,
243 244 245
					      size, &build_id__mark_dso_hit_ops);
}

246
static void record__exit(int status, void *arg)
247
{
248
	struct record *rec = arg;
249
	struct perf_data_file *file = &rec->file;
250

251 252 253
	if (status != 0)
		return;

254
	if (!file->is_pipe) {
255 256 257 258 259
		rec->session->header.data_size += rec->bytes_written;

		if (!rec->no_buildid)
			process_buildids(rec);
		perf_session__write_header(rec->session, rec->evlist,
260
					   file->fd, true);
261 262
		perf_session__delete(rec->session);
		perf_evlist__delete(rec->evlist);
263
		symbol__exit();
264
	}
265 266
}

267
static void perf_event__synthesize_guest_os(struct machine *machine, void *data)
268 269
{
	int err;
270
	struct perf_tool *tool = data;
271 272 273 274 275 276 277 278
	/*
	 *As for guest kernel when processing subcommand record&report,
	 *we arrange module mmap prior to guest kernel mmap and trigger
	 *a preload dso because default guest module symbols are loaded
	 *from guest kallsyms instead of /lib/modules/XXX/XXX. This
	 *method is used to avoid symbol missing when the first addr is
	 *in module instead of in guest kernel.
	 */
279
	err = perf_event__synthesize_modules(tool, process_synthesized_event,
280
					     machine);
281 282
	if (err < 0)
		pr_err("Couldn't record guest kernel [%d]'s reference"
283
		       " relocation symbol.\n", machine->pid);
284 285 286 287 288

	/*
	 * We use _stext for guest kernel because guest kernel's /proc/kallsyms
	 * have no _text sometimes.
	 */
289
	err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
290
						 machine, "_text");
291
	if (err < 0)
292
		err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
293
							 machine, "_stext");
294 295
	if (err < 0)
		pr_err("Couldn't record guest kernel [%d]'s reference"
296
		       " relocation symbol.\n", machine->pid);
297 298
}

299 300 301 302 303
static struct perf_event_header finished_round_event = {
	.size = sizeof(struct perf_event_header),
	.type = PERF_RECORD_FINISHED_ROUND,
};

304
static int record__mmap_read_all(struct record *rec)
305
{
306
	int i;
307
	int rc = 0;
308

309
	for (i = 0; i < rec->evlist->nr_mmaps; i++) {
310
		if (rec->evlist->mmap[i].base) {
311
			if (record__mmap_read(rec, &rec->evlist->mmap[i]) != 0) {
312 313 314 315
				rc = -1;
				goto out;
			}
		}
316 317
	}

318
	if (perf_header__has_feat(&rec->session->header, HEADER_TRACING_DATA))
319
		rc = record__write(rec, &finished_round_event, sizeof(finished_round_event));
320 321 322

out:
	return rc;
323 324
}

325
static void record__init_features(struct record *rec)
326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343
{
	struct perf_evlist *evsel_list = rec->evlist;
	struct perf_session *session = rec->session;
	int feat;

	for (feat = HEADER_FIRST_FEATURE; feat < HEADER_LAST_FEATURE; feat++)
		perf_header__set_feat(&session->header, feat);

	if (rec->no_buildid)
		perf_header__clear_feat(&session->header, HEADER_BUILD_ID);

	if (!have_tracepoints(&evsel_list->entries))
		perf_header__clear_feat(&session->header, HEADER_TRACING_DATA);

	if (!rec->opts.branch_stack)
		perf_header__clear_feat(&session->header, HEADER_BRANCH_STACK);
}

344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359
static volatile int workload_exec_errno;

/*
 * perf_evlist__prepare_workload will send a SIGUSR1
 * if the fork fails, since we asked by setting its
 * want_signal to true.
 */
static void workload_exec_failed_signal(int signo, siginfo_t *info,
					void *ucontext __maybe_unused)
{
	workload_exec_errno = info->si_value.sival_int;
	done = 1;
	signr = signo;
	child_finished = 1;
}

360
static int __cmd_record(struct record *rec, int argc, const char **argv)
361
{
362
	int err;
363
	unsigned long waking = 0;
364
	const bool forks = argc > 0;
365
	struct machine *machine;
366
	struct perf_tool *tool = &rec->tool;
367
	struct record_opts *opts = &rec->opts;
368
	struct perf_evlist *evsel_list = rec->evlist;
369
	struct perf_data_file *file = &rec->file;
370
	struct perf_session *session;
371
	bool disabled = false;
372

373
	rec->progname = argv[0];
374

375
	on_exit(record__sig_exit, rec);
376 377
	signal(SIGCHLD, sig_handler);
	signal(SIGINT, sig_handler);
378
	signal(SIGTERM, sig_handler);
379

380
	session = perf_session__new(file, false, NULL);
381
	if (session == NULL) {
382 383 384 385
		pr_err("Not enough memory for reading perf file header\n");
		return -1;
	}

386 387
	rec->session = session;

388
	record__init_features(rec);
389

390
	if (forks) {
391
		err = perf_evlist__prepare_workload(evsel_list, &opts->target,
392
						    argv, file->is_pipe,
393
						    true);
394 395 396
		if (err < 0) {
			pr_err("Couldn't run the workload!\n");
			goto out_delete_session;
397 398 399
		}
	}

400
	if (record__open(rec) != 0) {
401 402 403
		err = -1;
		goto out_delete_session;
	}
404

405 406 407
	if (!evsel_list->nr_groups)
		perf_header__clear_feat(&session->header, HEADER_GROUP_DESC);

408
	/*
409
	 * perf_session__delete(session) will be called at record__exit()
410
	 */
411
	on_exit(record__exit, rec);
412

413 414
	if (file->is_pipe) {
		err = perf_header__write_pipe(file->fd);
415
		if (err < 0)
416
			goto out_delete_session;
417
	} else {
418
		err = perf_session__write_header(session, evsel_list,
419
						 file->fd, false);
420
		if (err < 0)
421
			goto out_delete_session;
422 423
	}

424
	if (!rec->no_buildid
425
	    && !perf_header__has_feat(&session->header, HEADER_BUILD_ID)) {
426
		pr_err("Couldn't generate buildids. "
427
		       "Use --no-buildid to profile anyway.\n");
428 429
		err = -1;
		goto out_delete_session;
430 431
	}

432
	machine = &session->machines.host;
433

434
	if (file->is_pipe) {
435
		err = perf_event__synthesize_attrs(tool, session,
436
						   process_synthesized_event);
437 438
		if (err < 0) {
			pr_err("Couldn't synthesize attrs.\n");
439
			goto out_delete_session;
440
		}
441

442
		if (have_tracepoints(&evsel_list->entries)) {
443 444 445 446 447 448 449 450
			/*
			 * FIXME err <= 0 here actually means that
			 * there were no tracepoints so its not really
			 * an error, just that we don't need to
			 * synthesize anything.  We really have to
			 * return this more properly and also
			 * propagate errors that now are calling die()
			 */
451
			err = perf_event__synthesize_tracing_data(tool, file->fd, evsel_list,
452
								  process_synthesized_event);
453 454
			if (err <= 0) {
				pr_err("Couldn't record tracing data.\n");
455
				goto out_delete_session;
456
			}
457
			rec->bytes_written += err;
458
		}
459 460
	}

461
	err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
462
						 machine, "_text");
463
	if (err < 0)
464
		err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
465
							 machine, "_stext");
466 467 468 469
	if (err < 0)
		pr_err("Couldn't record kernel reference relocation symbol\n"
		       "Symbol resolution may be skewed if relocation was used (e.g. kexec).\n"
		       "Check /proc/kallsyms permission or run as root.\n");
470

471
	err = perf_event__synthesize_modules(tool, process_synthesized_event,
472
					     machine);
473 474 475 476 477
	if (err < 0)
		pr_err("Couldn't record kernel module information.\n"
		       "Symbol resolution may be skewed if relocation was used (e.g. kexec).\n"
		       "Check /proc/modules permission or run as root.\n");

478
	if (perf_guest) {
479 480
		machines__process_guests(&session->machines,
					 perf_event__synthesize_guest_os, tool);
481
	}
482

483 484
	err = __machine__synthesize_threads(machine, tool, &opts->target, evsel_list->threads,
					    process_synthesized_event, opts->sample_address);
485 486 487
	if (err != 0)
		goto out_delete_session;

488
	if (rec->realtime_prio) {
489 490
		struct sched_param param;

491
		param.sched_priority = rec->realtime_prio;
492
		if (sched_setscheduler(0, SCHED_FIFO, &param)) {
493
			pr_err("Could not set realtime priority.\n");
494 495
			err = -1;
			goto out_delete_session;
496 497 498
		}
	}

499 500 501 502 503
	/*
	 * When perf is starting the traced process, all the events
	 * (apart from group members) have enable_on_exec=1 set,
	 * so don't spoil it by prematurely enabling them.
	 */
504
	if (!target__none(&opts->target))
505
		perf_evlist__enable(evsel_list);
506

507 508 509
	/*
	 * Let the child rip
	 */
510 511 512 513 514 515 516 517 518 519 520 521
	if (forks) {
		struct sigaction act = {
			.sa_flags     = SA_SIGINFO,
			.sa_sigaction = workload_exec_failed_signal,
		};
		/*
		 * perf_evlist__prepare_workload will, after we call
		 * perf_evlist__start_Workload, send a SIGUSR1 if the exec call
		 * fails, that we will catch in workload_signal to flip
		 * workload_exec_errno.
 		 */
		sigaction(SIGUSR1, &act, NULL);
522
		perf_evlist__start_workload(evsel_list);
523
	}
524

525
	for (;;) {
526
		int hits = rec->samples;
527

528
		if (record__mmap_read_all(rec) < 0) {
529 530 531
			err = -1;
			goto out_delete_session;
		}
532

533
		if (hits == rec->samples) {
534 535
			if (done)
				break;
536
			err = poll(evsel_list->pollfd, evsel_list->nr_fds, -1);
537 538 539
			waking++;
		}

540 541 542 543 544
		/*
		 * When perf is starting the traced process, at the end events
		 * die with the process and we wait for that. Thus no need to
		 * disable events in this case.
		 */
545
		if (done && !disabled && !target__none(&opts->target)) {
546
			perf_evlist__disable(evsel_list);
547 548
			disabled = true;
		}
549 550
	}

551 552 553 554 555 556 557 558
	if (forks && workload_exec_errno) {
		char msg[512];
		const char *emsg = strerror_r(workload_exec_errno, msg, sizeof(msg));
		pr_err("Workload failed: %s\n", emsg);
		err = -1;
		goto out_delete_session;
	}

559
	if (quiet || signr == SIGUSR1)
560 561
		return 0;

562 563
	fprintf(stderr, "[ perf record: Woken up %ld times to write data ]\n", waking);

564 565 566 567
	/*
	 * Approximate RIP event size: 24 bytes.
	 */
	fprintf(stderr,
568
		"[ perf record: Captured and wrote %.3f MB %s (~%" PRIu64 " samples) ]\n",
569
		(double)rec->bytes_written / 1024.0 / 1024.0,
570
		file->path,
571
		rec->bytes_written / 24);
572

573
	return 0;
574 575 576 577

out_delete_session:
	perf_session__delete(session);
	return err;
578
}
579

580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597
#define BRANCH_OPT(n, m) \
	{ .name = n, .mode = (m) }

#define BRANCH_END { .name = NULL }

struct branch_mode {
	const char *name;
	int mode;
};

static const struct branch_mode branch_modes[] = {
	BRANCH_OPT("u", PERF_SAMPLE_BRANCH_USER),
	BRANCH_OPT("k", PERF_SAMPLE_BRANCH_KERNEL),
	BRANCH_OPT("hv", PERF_SAMPLE_BRANCH_HV),
	BRANCH_OPT("any", PERF_SAMPLE_BRANCH_ANY),
	BRANCH_OPT("any_call", PERF_SAMPLE_BRANCH_ANY_CALL),
	BRANCH_OPT("any_ret", PERF_SAMPLE_BRANCH_ANY_RETURN),
	BRANCH_OPT("ind_call", PERF_SAMPLE_BRANCH_IND_CALL),
598 599 600
	BRANCH_OPT("abort_tx", PERF_SAMPLE_BRANCH_ABORT_TX),
	BRANCH_OPT("in_tx", PERF_SAMPLE_BRANCH_IN_TX),
	BRANCH_OPT("no_tx", PERF_SAMPLE_BRANCH_NO_TX),
601 602 603 604
	BRANCH_END
};

static int
605
parse_branch_stack(const struct option *opt, const char *str, int unset)
606 607 608 609 610 611 612 613
{
#define ONLY_PLM \
	(PERF_SAMPLE_BRANCH_USER	|\
	 PERF_SAMPLE_BRANCH_KERNEL	|\
	 PERF_SAMPLE_BRANCH_HV)

	uint64_t *mode = (uint64_t *)opt->value;
	const struct branch_mode *br;
614
	char *s, *os = NULL, *p;
615 616
	int ret = -1;

617 618
	if (unset)
		return 0;
619

620 621 622 623
	/*
	 * cannot set it twice, -b + --branch-filter for instance
	 */
	if (*mode)
624 625
		return -1;

626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646
	/* str may be NULL in case no arg is passed to -b */
	if (str) {
		/* because str is read-only */
		s = os = strdup(str);
		if (!s)
			return -1;

		for (;;) {
			p = strchr(s, ',');
			if (p)
				*p = '\0';

			for (br = branch_modes; br->name; br++) {
				if (!strcasecmp(s, br->name))
					break;
			}
			if (!br->name) {
				ui__warning("unknown branch filter %s,"
					    " check man page\n", s);
				goto error;
			}
647

648
			*mode |= br->mode;
649

650 651
			if (!p)
				break;
652

653 654
			s = p + 1;
		}
655 656 657
	}
	ret = 0;

658
	/* default to any branch */
659
	if ((*mode & ~ONLY_PLM) == 0) {
660
		*mode = PERF_SAMPLE_BRANCH_ANY;
661 662 663 664 665 666
	}
error:
	free(os);
	return ret;
}

667
#ifdef HAVE_LIBUNWIND_SUPPORT
668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692
static int get_stack_size(char *str, unsigned long *_size)
{
	char *endptr;
	unsigned long size;
	unsigned long max_size = round_down(USHRT_MAX, sizeof(u64));

	size = strtoul(str, &endptr, 0);

	do {
		if (*endptr)
			break;

		size = round_up(size, sizeof(u64));
		if (!size || size > max_size)
			break;

		*_size = size;
		return 0;

	} while (0);

	pr_err("callchain: Incorrect stack dump size (max %ld): %s\n",
	       max_size, str);
	return -1;
}
693
#endif /* HAVE_LIBUNWIND_SUPPORT */
694

695
int record_parse_callchain(const char *arg, struct record_opts *opts)
696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714
{
	char *tok, *name, *saveptr = NULL;
	char *buf;
	int ret = -1;

	/* We need buffer that we know we can write to. */
	buf = malloc(strlen(arg) + 1);
	if (!buf)
		return -ENOMEM;

	strcpy(buf, arg);

	tok = strtok_r((char *)buf, ",", &saveptr);
	name = tok ? : (char *)buf;

	do {
		/* Framepointer style */
		if (!strncmp(name, "fp", sizeof("fp"))) {
			if (!strtok_r(NULL, ",", &saveptr)) {
715
				opts->call_graph = CALLCHAIN_FP;
716 717 718 719 720 721
				ret = 0;
			} else
				pr_err("callchain: No more arguments "
				       "needed for -g fp\n");
			break;

722
#ifdef HAVE_LIBUNWIND_SUPPORT
723 724
		/* Dwarf style */
		} else if (!strncmp(name, "dwarf", sizeof("dwarf"))) {
725 726
			const unsigned long default_stack_dump_size = 8192;

727
			ret = 0;
728 729
			opts->call_graph = CALLCHAIN_DWARF;
			opts->stack_dump_size = default_stack_dump_size;
730 731 732 733 734 735

			tok = strtok_r(NULL, ",", &saveptr);
			if (tok) {
				unsigned long size = 0;

				ret = get_stack_size(tok, &size);
736
				opts->stack_dump_size = size;
737
			}
738
#endif /* HAVE_LIBUNWIND_SUPPORT */
739
		} else {
J
Jiri Olsa 已提交
740
			pr_err("callchain: Unknown --call-graph option "
741 742 743 744 745 746 747
			       "value: %s\n", arg);
			break;
		}

	} while (0);

	free(buf);
J
Jiri Olsa 已提交
748 749 750
	return ret;
}

751
static void callchain_debug(struct record_opts *opts)
J
Jiri Olsa 已提交
752 753
{
	pr_debug("callchain: type %d\n", opts->call_graph);
754

J
Jiri Olsa 已提交
755 756 757 758 759 760 761 762 763
	if (opts->call_graph == CALLCHAIN_DWARF)
		pr_debug("callchain: stack dump size %d\n",
			 opts->stack_dump_size);
}

int record_parse_callchain_opt(const struct option *opt,
			       const char *arg,
			       int unset)
{
764
	struct record_opts *opts = opt->value;
J
Jiri Olsa 已提交
765 766 767 768 769 770 771 772 773 774
	int ret;

	/* --no-call-graph */
	if (unset) {
		opts->call_graph = CALLCHAIN_NONE;
		pr_debug("callchain: disabled\n");
		return 0;
	}

	ret = record_parse_callchain(arg, opts);
775
	if (!ret)
J
Jiri Olsa 已提交
776
		callchain_debug(opts);
777 778 779 780

	return ret;
}

J
Jiri Olsa 已提交
781 782 783 784
int record_callchain_opt(const struct option *opt,
			 const char *arg __maybe_unused,
			 int unset __maybe_unused)
{
785
	struct record_opts *opts = opt->value;
J
Jiri Olsa 已提交
786 787 788 789 790 791 792 793

	if (opts->call_graph == CALLCHAIN_NONE)
		opts->call_graph = CALLCHAIN_FP;

	callchain_debug(opts);
	return 0;
}

794
static const char * const record_usage[] = {
795 796
	"perf record [<options>] [<command>]",
	"perf record [<options>] -- <command> [<options>]",
797 798 799
	NULL
};

800
/*
801 802
 * XXX Ideally would be local to cmd_record() and passed to a record__new
 * because we need to have access to it in record__exit, that is called
803 804 805 806 807 808 809
 * after cmd_record() exits, but since record_options need to be accessible to
 * builtin-script, leave it here.
 *
 * At least we don't ouch it in all the other functions here directly.
 *
 * Just say no to tons of global variables, sigh.
 */
810
static struct record record = {
811 812 813 814
	.opts = {
		.mmap_pages	     = UINT_MAX,
		.user_freq	     = UINT_MAX,
		.user_interval	     = ULLONG_MAX,
815
		.freq		     = 4000,
N
Namhyung Kim 已提交
816 817
		.target		     = {
			.uses_mmap   = true,
818
			.default_per_cpu = true,
N
Namhyung Kim 已提交
819
		},
820 821
	},
};
822

J
Jiri Olsa 已提交
823
#define CALLCHAIN_HELP "setup and enables call-graph (stack chain/backtrace) recording: "
824

825
#ifdef HAVE_LIBUNWIND_SUPPORT
J
Jiri Olsa 已提交
826
const char record_callchain_help[] = CALLCHAIN_HELP "fp dwarf";
827
#else
J
Jiri Olsa 已提交
828
const char record_callchain_help[] = CALLCHAIN_HELP "fp";
829 830
#endif

831 832 833
/*
 * XXX Will stay a global variable till we fix builtin-script.c to stop messing
 * with it and switch to use the library functions in perf_evlist that came
834
 * from builtin-record.c, i.e. use record_opts,
835 836 837
 * perf_evlist__prepare_workload, etc instead of fork+exec'in 'perf record',
 * using pipes, etc.
 */
838
const struct option record_options[] = {
839
	OPT_CALLBACK('e', "event", &record.evlist, "event",
840
		     "event selector. use 'perf list' to list available events",
841
		     parse_events_option),
842
	OPT_CALLBACK(0, "filter", &record.evlist, "filter",
L
Li Zefan 已提交
843
		     "event filter", parse_filter),
844
	OPT_STRING('p', "pid", &record.opts.target.pid, "pid",
845
		    "record events on existing process id"),
846
	OPT_STRING('t', "tid", &record.opts.target.tid, "tid",
847
		    "record events on existing thread id"),
848
	OPT_INTEGER('r', "realtime", &record.realtime_prio,
849
		    "collect data with this RT SCHED_FIFO priority"),
850
	OPT_BOOLEAN('D', "no-delay", &record.opts.no_delay,
851
		    "collect data without buffering"),
852
	OPT_BOOLEAN('R', "raw-samples", &record.opts.raw_samples,
853
		    "collect raw sample records from all opened counters"),
854
	OPT_BOOLEAN('a', "all-cpus", &record.opts.target.system_wide,
855
			    "system-wide collection from all CPUs"),
856
	OPT_STRING('C', "cpu", &record.opts.target.cpu_list, "cpu",
857
		    "list of cpus to monitor"),
858
	OPT_U64('c', "count", &record.opts.user_interval, "event period to sample"),
859
	OPT_STRING('o', "output", &record.file.path, "file",
I
Ingo Molnar 已提交
860
		    "output file name"),
861 862 863
	OPT_BOOLEAN_SET('i', "no-inherit", &record.opts.no_inherit,
			&record.opts.no_inherit_set,
			"child tasks do not inherit counters"),
864
	OPT_UINTEGER('F', "freq", &record.opts.user_freq, "profile at this frequency"),
865 866 867
	OPT_CALLBACK('m', "mmap-pages", &record.opts.mmap_pages, "pages",
		     "number of mmap data pages",
		     perf_evlist__parse_mmap_pages),
868
	OPT_BOOLEAN(0, "group", &record.opts.group,
869
		    "put the counters into a counter group"),
J
Jiri Olsa 已提交
870 871 872 873 874 875
	OPT_CALLBACK_NOOPT('g', NULL, &record.opts,
			   NULL, "enables call-graph recording" ,
			   &record_callchain_opt),
	OPT_CALLBACK(0, "call-graph", &record.opts,
		     "mode[,dump_size]", record_callchain_help,
		     &record_parse_callchain_opt),
876
	OPT_INCR('v', "verbose", &verbose,
877
		    "be more verbose (show counter open errors, etc)"),
878
	OPT_BOOLEAN('q', "quiet", &quiet, "don't print any message"),
879
	OPT_BOOLEAN('s', "stat", &record.opts.inherit_stat,
880
		    "per thread counts"),
881
	OPT_BOOLEAN('d', "data", &record.opts.sample_address,
882
		    "Sample addresses"),
883
	OPT_BOOLEAN('T', "timestamp", &record.opts.sample_time, "Sample timestamps"),
884
	OPT_BOOLEAN('P', "period", &record.opts.period, "Sample period"),
885
	OPT_BOOLEAN('n', "no-samples", &record.opts.no_samples,
886
		    "don't sample"),
887
	OPT_BOOLEAN('N', "no-buildid-cache", &record.no_buildid_cache,
888
		    "do not update the buildid cache"),
889
	OPT_BOOLEAN('B', "no-buildid", &record.no_buildid,
890
		    "do not collect buildids in perf.data"),
891
	OPT_CALLBACK('G', "cgroup", &record.evlist, "name",
S
Stephane Eranian 已提交
892 893
		     "monitor event in cgroup name only",
		     parse_cgroups),
894 895
	OPT_STRING('u', "uid", &record.opts.target.uid_str, "user",
		   "user to profile"),
896 897 898 899 900 901 902

	OPT_CALLBACK_NOOPT('b', "branch-any", &record.opts.branch_stack,
		     "branch any", "sample any taken branches",
		     parse_branch_stack),

	OPT_CALLBACK('j', "branch-filter", &record.opts.branch_stack,
		     "branch filter mask", "branch stack filter modes",
903
		     parse_branch_stack),
904 905
	OPT_BOOLEAN('W', "weight", &record.opts.sample_weight,
		    "sample by weight (on special events only)"),
906 907
	OPT_BOOLEAN(0, "transaction", &record.opts.sample_transaction,
		    "sample transaction flags (special events only)"),
908 909
	OPT_BOOLEAN(0, "per-thread", &record.opts.target.per_thread,
		    "use per-thread mmaps"),
910 911 912
	OPT_END()
};

913
int cmd_record(int argc, const char **argv, const char *prefix __maybe_unused)
914
{
915
	int err = -ENOMEM;
916
	struct perf_evlist *evsel_list;
917
	struct record *rec = &record;
918
	char errbuf[BUFSIZ];
919

920
	evsel_list = perf_evlist__new();
921 922 923
	if (evsel_list == NULL)
		return -ENOMEM;

924 925
	rec->evlist = evsel_list;

926
	argc = parse_options(argc, argv, record_options, record_usage,
927
			    PARSE_OPT_STOP_AT_NON_OPTION);
928
	if (!argc && target__none(&rec->opts.target))
929
		usage_with_options(record_usage, record_options);
930

931
	if (nr_cgroups && !rec->opts.target.system_wide) {
932 933
		ui__error("cgroup monitoring only available in"
			  " system-wide mode\n");
S
Stephane Eranian 已提交
934 935 936
		usage_with_options(record_usage, record_options);
	}

937
	symbol__init();
938

939
	if (symbol_conf.kptr_restrict)
940 941 942 943 944 945 946 947
		pr_warning(
"WARNING: Kernel address maps (/proc/{kallsyms,modules}) are restricted,\n"
"check /proc/sys/kernel/kptr_restrict.\n\n"
"Samples in kernel functions may not be resolved if a suitable vmlinux\n"
"file is not found in the buildid cache or in the vmlinux path.\n\n"
"Samples in kernel modules won't be resolved at all.\n\n"
"If some relocation was applied (e.g. kexec) symbols may be misresolved\n"
"even with a suitable vmlinux or kallsyms file.\n\n");
948

949
	if (rec->no_buildid_cache || rec->no_buildid)
950
		disable_buildid_cache();
951

952 953
	if (evsel_list->nr_entries == 0 &&
	    perf_evlist__add_default(evsel_list) < 0) {
954 955
		pr_err("Not enough memory for event selector list\n");
		goto out_symbol_exit;
956
	}
957

958 959 960
	if (rec->opts.target.tid && !rec->opts.no_inherit_set)
		rec->opts.no_inherit = true;

961
	err = target__validate(&rec->opts.target);
962
	if (err) {
963
		target__strerror(&rec->opts.target, err, errbuf, BUFSIZ);
964 965 966
		ui__warning("%s", errbuf);
	}

967
	err = target__parse_uid(&rec->opts.target);
968 969
	if (err) {
		int saved_errno = errno;
970

971
		target__strerror(&rec->opts.target, err, errbuf, BUFSIZ);
972
		ui__error("%s", errbuf);
973 974

		err = -saved_errno;
975
		goto out_symbol_exit;
976
	}
977

978
	err = -ENOMEM;
979
	if (perf_evlist__create_maps(evsel_list, &rec->opts.target) < 0)
980
		usage_with_options(record_usage, record_options);
981

982
	if (record_opts__config(&rec->opts)) {
983
		err = -EINVAL;
984
		goto out_free_fd;
985 986
	}

987
	err = __cmd_record(&record, argc, argv);
988 989 990

	perf_evlist__munmap(evsel_list);
	perf_evlist__close(evsel_list);
991
out_free_fd:
992
	perf_evlist__delete_maps(evsel_list);
993 994
out_symbol_exit:
	symbol__exit();
995
	return err;
996
}