builtin-record.c 25.0 KB
Newer Older
I
Ingo Molnar 已提交
1
/*
2 3 4 5 6
 * builtin-record.c
 *
 * Builtin record command: Record the profile of a workload
 * (or a CPU, or a PID) into the perf.data output file - for
 * later analysis via perf report.
I
Ingo Molnar 已提交
7
 */
8
#include "builtin.h"
9 10 11

#include "perf.h"

12
#include "util/build-id.h"
13
#include "util/util.h"
14
#include "util/parse-options.h"
15
#include "util/parse-events.h"
16

17
#include "util/header.h"
18
#include "util/event.h"
19
#include "util/evlist.h"
20
#include "util/evsel.h"
21
#include "util/debug.h"
22
#include "util/session.h"
23
#include "util/tool.h"
24
#include "util/symbol.h"
25
#include "util/cpumap.h"
26
#include "util/thread_map.h"
27
#include "util/data.h"
28

29
#include <unistd.h>
30
#include <sched.h>
31
#include <sys/mman.h>
32

33
#ifndef HAVE_ON_EXIT_SUPPORT
34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64
#ifndef ATEXIT_MAX
#define ATEXIT_MAX 32
#endif
static int __on_exit_count = 0;
typedef void (*on_exit_func_t) (int, void *);
static on_exit_func_t __on_exit_funcs[ATEXIT_MAX];
static void *__on_exit_args[ATEXIT_MAX];
static int __exitcode = 0;
static void __handle_on_exit_funcs(void);
static int on_exit(on_exit_func_t function, void *arg);
#define exit(x) (exit)(__exitcode = (x))

static int on_exit(on_exit_func_t function, void *arg)
{
	if (__on_exit_count == ATEXIT_MAX)
		return -ENOMEM;
	else if (__on_exit_count == 0)
		atexit(__handle_on_exit_funcs);
	__on_exit_funcs[__on_exit_count] = function;
	__on_exit_args[__on_exit_count++] = arg;
	return 0;
}

static void __handle_on_exit_funcs(void)
{
	int i;
	for (i = 0; i < __on_exit_count; i++)
		__on_exit_funcs[i] (__exitcode, __on_exit_args[i]);
}
#endif

65
struct record {
66
	struct perf_tool	tool;
67
	struct record_opts	opts;
68
	u64			bytes_written;
69
	struct perf_data_file	file;
70 71 72 73 74 75 76
	struct perf_evlist	*evlist;
	struct perf_session	*session;
	const char		*progname;
	int			realtime_prio;
	bool			no_buildid;
	bool			no_buildid_cache;
	long			samples;
77
};
78

79
static int record__write(struct record *rec, void *bf, size_t size)
80
{
81
	if (perf_data_file__write(rec->session->file, bf, size) < 0) {
82 83
		pr_err("failed to write perf data, error: %m\n");
		return -1;
84
	}
85

86
	rec->bytes_written += size;
87
	return 0;
88 89
}

90
static int process_synthesized_event(struct perf_tool *tool,
91
				     union perf_event *event,
92 93
				     struct perf_sample *sample __maybe_unused,
				     struct machine *machine __maybe_unused)
94
{
95 96
	struct record *rec = container_of(tool, struct record, tool);
	return record__write(rec, event, event->header.size);
97 98
}

99
static int record__mmap_read(struct record *rec, struct perf_mmap *md)
100
{
101
	unsigned int head = perf_mmap__read_head(md);
102
	unsigned int old = md->prev;
J
Jiri Olsa 已提交
103
	unsigned char *data = md->base + page_size;
104 105
	unsigned long size;
	void *buf;
106
	int rc = 0;
107

108
	if (old == head)
109
		return 0;
110

111
	rec->samples++;
112 113 114 115 116 117 118

	size = head - old;

	if ((old & md->mask) + size != (head & md->mask)) {
		buf = &data[old & md->mask];
		size = md->mask + 1 - (old & md->mask);
		old += size;
119

120
		if (record__write(rec, buf, size) < 0) {
121 122 123
			rc = -1;
			goto out;
		}
124 125 126 127 128
	}

	buf = &data[old & md->mask];
	size = head - old;
	old += size;
129

130
	if (record__write(rec, buf, size) < 0) {
131 132 133
		rc = -1;
		goto out;
	}
134 135

	md->prev = old;
136
	perf_mmap__write_tail(md, old);
137 138 139

out:
	return rc;
140 141 142
}

static volatile int done = 0;
143
static volatile int signr = -1;
144
static volatile int child_finished = 0;
145

146
static void sig_handler(int sig)
147
{
148 149 150
	if (sig == SIGCHLD)
		child_finished = 1;

151
	done = 1;
152 153 154
	signr = sig;
}

155
static void record__sig_exit(int exit_status __maybe_unused, void *arg)
156
{
157
	struct record *rec = arg;
158 159
	int status;

160
	if (rec->evlist->workload.pid > 0) {
161
		if (!child_finished)
162
			kill(rec->evlist->workload.pid, SIGTERM);
163 164 165

		wait(&status);
		if (WIFSIGNALED(status))
166
			psignal(WTERMSIG(status), rec->progname);
167
	}
168

169
	if (signr == -1 || signr == SIGUSR1)
170 171 172
		return;

	signal(signr, SIG_DFL);
173 174
}

175
static int record__open(struct record *rec)
176
{
177
	char msg[512];
178
	struct perf_evsel *pos;
179 180
	struct perf_evlist *evlist = rec->evlist;
	struct perf_session *session = rec->session;
181
	struct record_opts *opts = &rec->opts;
182
	int rc = 0;
183

184
	perf_evlist__config(evlist, opts);
185

186
	evlist__for_each(evlist, pos) {
187
try_again:
188
		if (perf_evsel__open(pos, evlist->cpus, evlist->threads) < 0) {
189
			if (perf_evsel__fallback(pos, errno, msg, sizeof(msg))) {
190
				if (verbose)
191
					ui__warning("%s\n", msg);
192 193
				goto try_again;
			}
194

195 196 197 198
			rc = -errno;
			perf_evsel__open_strerror(pos, &opts->target,
						  errno, msg, sizeof(msg));
			ui__error("%s\n", msg);
199
			goto out;
L
Li Zefan 已提交
200 201
		}
	}
202

203
	if (perf_evlist__apply_filters(evlist)) {
204 205
		error("failed to set filter with %d (%s)\n", errno,
			strerror(errno));
206 207
		rc = -1;
		goto out;
208 209
	}

210
	if (perf_evlist__mmap(evlist, opts->mmap_pages, false) < 0) {
211 212 213 214 215
		if (errno == EPERM) {
			pr_err("Permission error mapping pages.\n"
			       "Consider increasing "
			       "/proc/sys/kernel/perf_event_mlock_kb,\n"
			       "or try again with a smaller value of -m/--mmap_pages.\n"
216
			       "(current value: %u)\n", opts->mmap_pages);
217 218 219 220 221 222
			rc = -errno;
		} else {
			pr_err("failed to mmap with %d (%s)\n", errno, strerror(errno));
			rc = -errno;
		}
		goto out;
223
	}
224

225
	session->evlist = evlist;
226
	perf_session__set_id_hdr_size(session);
227 228
out:
	return rc;
229 230
}

231
static int process_buildids(struct record *rec)
232
{
233 234
	struct perf_data_file *file  = &rec->file;
	struct perf_session *session = rec->session;
235
	u64 start = session->header.data_offset;
236

237
	u64 size = lseek(file->fd, 0, SEEK_CUR);
238 239 240
	if (size == 0)
		return 0;

241 242
	return __perf_session__process_events(session, start,
					      size - start,
243 244 245
					      size, &build_id__mark_dso_hit_ops);
}

246
static void record__exit(int status, void *arg)
247
{
248
	struct record *rec = arg;
249
	struct perf_data_file *file = &rec->file;
250

251 252 253
	if (status != 0)
		return;

254
	if (!file->is_pipe) {
255 256 257 258 259
		rec->session->header.data_size += rec->bytes_written;

		if (!rec->no_buildid)
			process_buildids(rec);
		perf_session__write_header(rec->session, rec->evlist,
260
					   file->fd, true);
261 262
		perf_session__delete(rec->session);
		perf_evlist__delete(rec->evlist);
263
		symbol__exit();
264
	}
265 266
}

267
static void perf_event__synthesize_guest_os(struct machine *machine, void *data)
268 269
{
	int err;
270
	struct perf_tool *tool = data;
271 272 273 274 275 276 277 278
	/*
	 *As for guest kernel when processing subcommand record&report,
	 *we arrange module mmap prior to guest kernel mmap and trigger
	 *a preload dso because default guest module symbols are loaded
	 *from guest kallsyms instead of /lib/modules/XXX/XXX. This
	 *method is used to avoid symbol missing when the first addr is
	 *in module instead of in guest kernel.
	 */
279
	err = perf_event__synthesize_modules(tool, process_synthesized_event,
280
					     machine);
281 282
	if (err < 0)
		pr_err("Couldn't record guest kernel [%d]'s reference"
283
		       " relocation symbol.\n", machine->pid);
284 285 286 287 288

	/*
	 * We use _stext for guest kernel because guest kernel's /proc/kallsyms
	 * have no _text sometimes.
	 */
289
	err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
290
						 machine, "_text");
291
	if (err < 0)
292
		err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
293
							 machine, "_stext");
294 295
	if (err < 0)
		pr_err("Couldn't record guest kernel [%d]'s reference"
296
		       " relocation symbol.\n", machine->pid);
297 298
}

299 300 301 302 303
static struct perf_event_header finished_round_event = {
	.size = sizeof(struct perf_event_header),
	.type = PERF_RECORD_FINISHED_ROUND,
};

304
static int record__mmap_read_all(struct record *rec)
305
{
306
	int i;
307
	int rc = 0;
308

309
	for (i = 0; i < rec->evlist->nr_mmaps; i++) {
310
		if (rec->evlist->mmap[i].base) {
311
			if (record__mmap_read(rec, &rec->evlist->mmap[i]) != 0) {
312 313 314 315
				rc = -1;
				goto out;
			}
		}
316 317
	}

318
	if (perf_header__has_feat(&rec->session->header, HEADER_TRACING_DATA))
319
		rc = record__write(rec, &finished_round_event, sizeof(finished_round_event));
320 321 322

out:
	return rc;
323 324
}

325
static void record__init_features(struct record *rec)
326 327 328 329 330 331 332 333 334 335
{
	struct perf_session *session = rec->session;
	int feat;

	for (feat = HEADER_FIRST_FEATURE; feat < HEADER_LAST_FEATURE; feat++)
		perf_header__set_feat(&session->header, feat);

	if (rec->no_buildid)
		perf_header__clear_feat(&session->header, HEADER_BUILD_ID);

336
	if (!have_tracepoints(&rec->evlist->entries))
337 338 339 340 341 342
		perf_header__clear_feat(&session->header, HEADER_TRACING_DATA);

	if (!rec->opts.branch_stack)
		perf_header__clear_feat(&session->header, HEADER_BRANCH_STACK);
}

343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358
static volatile int workload_exec_errno;

/*
 * perf_evlist__prepare_workload will send a SIGUSR1
 * if the fork fails, since we asked by setting its
 * want_signal to true.
 */
static void workload_exec_failed_signal(int signo, siginfo_t *info,
					void *ucontext __maybe_unused)
{
	workload_exec_errno = info->si_value.sival_int;
	done = 1;
	signr = signo;
	child_finished = 1;
}

359
static int __cmd_record(struct record *rec, int argc, const char **argv)
360
{
361
	int err;
362
	unsigned long waking = 0;
363
	const bool forks = argc > 0;
364
	struct machine *machine;
365
	struct perf_tool *tool = &rec->tool;
366
	struct record_opts *opts = &rec->opts;
367
	struct perf_data_file *file = &rec->file;
368
	struct perf_session *session;
369
	bool disabled = false;
370

371
	rec->progname = argv[0];
372

373
	on_exit(record__sig_exit, rec);
374 375
	signal(SIGCHLD, sig_handler);
	signal(SIGINT, sig_handler);
376
	signal(SIGTERM, sig_handler);
377

378
	session = perf_session__new(file, false, NULL);
379
	if (session == NULL) {
380 381 382 383
		pr_err("Not enough memory for reading perf file header\n");
		return -1;
	}

384 385
	rec->session = session;

386
	record__init_features(rec);
387

388
	if (forks) {
389
		err = perf_evlist__prepare_workload(rec->evlist, &opts->target,
390
						    argv, file->is_pipe,
391
						    workload_exec_failed_signal);
392 393 394
		if (err < 0) {
			pr_err("Couldn't run the workload!\n");
			goto out_delete_session;
395 396 397
		}
	}

398
	if (record__open(rec) != 0) {
399 400 401
		err = -1;
		goto out_delete_session;
	}
402

403
	if (!rec->evlist->nr_groups)
404 405
		perf_header__clear_feat(&session->header, HEADER_GROUP_DESC);

406
	/*
407
	 * perf_session__delete(session) will be called at record__exit()
408
	 */
409
	on_exit(record__exit, rec);
410

411 412
	if (file->is_pipe) {
		err = perf_header__write_pipe(file->fd);
413
		if (err < 0)
414
			goto out_delete_session;
415
	} else {
416
		err = perf_session__write_header(session, rec->evlist,
417
						 file->fd, false);
418
		if (err < 0)
419
			goto out_delete_session;
420 421
	}

422
	if (!rec->no_buildid
423
	    && !perf_header__has_feat(&session->header, HEADER_BUILD_ID)) {
424
		pr_err("Couldn't generate buildids. "
425
		       "Use --no-buildid to profile anyway.\n");
426 427
		err = -1;
		goto out_delete_session;
428 429
	}

430
	machine = &session->machines.host;
431

432
	if (file->is_pipe) {
433
		err = perf_event__synthesize_attrs(tool, session,
434
						   process_synthesized_event);
435 436
		if (err < 0) {
			pr_err("Couldn't synthesize attrs.\n");
437
			goto out_delete_session;
438
		}
439

440
		if (have_tracepoints(&rec->evlist->entries)) {
441 442 443 444 445 446 447 448
			/*
			 * FIXME err <= 0 here actually means that
			 * there were no tracepoints so its not really
			 * an error, just that we don't need to
			 * synthesize anything.  We really have to
			 * return this more properly and also
			 * propagate errors that now are calling die()
			 */
449
			err = perf_event__synthesize_tracing_data(tool, file->fd, rec->evlist,
450
								  process_synthesized_event);
451 452
			if (err <= 0) {
				pr_err("Couldn't record tracing data.\n");
453
				goto out_delete_session;
454
			}
455
			rec->bytes_written += err;
456
		}
457 458
	}

459
	err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
460
						 machine, "_text");
461
	if (err < 0)
462
		err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
463
							 machine, "_stext");
464 465 466 467
	if (err < 0)
		pr_err("Couldn't record kernel reference relocation symbol\n"
		       "Symbol resolution may be skewed if relocation was used (e.g. kexec).\n"
		       "Check /proc/kallsyms permission or run as root.\n");
468

469
	err = perf_event__synthesize_modules(tool, process_synthesized_event,
470
					     machine);
471 472 473 474 475
	if (err < 0)
		pr_err("Couldn't record kernel module information.\n"
		       "Symbol resolution may be skewed if relocation was used (e.g. kexec).\n"
		       "Check /proc/modules permission or run as root.\n");

476
	if (perf_guest) {
477 478
		machines__process_guests(&session->machines,
					 perf_event__synthesize_guest_os, tool);
479
	}
480

481
	err = __machine__synthesize_threads(machine, tool, &opts->target, rec->evlist->threads,
482
					    process_synthesized_event, opts->sample_address);
483 484 485
	if (err != 0)
		goto out_delete_session;

486
	if (rec->realtime_prio) {
487 488
		struct sched_param param;

489
		param.sched_priority = rec->realtime_prio;
490
		if (sched_setscheduler(0, SCHED_FIFO, &param)) {
491
			pr_err("Could not set realtime priority.\n");
492 493
			err = -1;
			goto out_delete_session;
494 495 496
		}
	}

497 498 499 500 501
	/*
	 * When perf is starting the traced process, all the events
	 * (apart from group members) have enable_on_exec=1 set,
	 * so don't spoil it by prematurely enabling them.
	 */
502
	if (!target__none(&opts->target) && !opts->initial_delay)
503
		perf_evlist__enable(rec->evlist);
504

505 506 507
	/*
	 * Let the child rip
	 */
508
	if (forks)
509
		perf_evlist__start_workload(rec->evlist);
510

511 512 513 514 515
	if (opts->initial_delay) {
		usleep(opts->initial_delay * 1000);
		perf_evlist__enable(rec->evlist);
	}

516
	for (;;) {
517
		int hits = rec->samples;
518

519
		if (record__mmap_read_all(rec) < 0) {
520 521 522
			err = -1;
			goto out_delete_session;
		}
523

524
		if (hits == rec->samples) {
525 526
			if (done)
				break;
527
			err = poll(rec->evlist->pollfd, rec->evlist->nr_fds, -1);
528 529 530
			waking++;
		}

531 532 533 534 535
		/*
		 * When perf is starting the traced process, at the end events
		 * die with the process and we wait for that. Thus no need to
		 * disable events in this case.
		 */
536
		if (done && !disabled && !target__none(&opts->target)) {
537
			perf_evlist__disable(rec->evlist);
538 539
			disabled = true;
		}
540 541
	}

542 543 544 545 546 547 548 549
	if (forks && workload_exec_errno) {
		char msg[512];
		const char *emsg = strerror_r(workload_exec_errno, msg, sizeof(msg));
		pr_err("Workload failed: %s\n", emsg);
		err = -1;
		goto out_delete_session;
	}

550
	if (quiet || signr == SIGUSR1)
551 552
		return 0;

553 554
	fprintf(stderr, "[ perf record: Woken up %ld times to write data ]\n", waking);

555 556 557 558
	/*
	 * Approximate RIP event size: 24 bytes.
	 */
	fprintf(stderr,
559
		"[ perf record: Captured and wrote %.3f MB %s (~%" PRIu64 " samples) ]\n",
560
		(double)rec->bytes_written / 1024.0 / 1024.0,
561
		file->path,
562
		rec->bytes_written / 24);
563

564
	return 0;
565 566 567 568

out_delete_session:
	perf_session__delete(session);
	return err;
569
}
570

571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588
#define BRANCH_OPT(n, m) \
	{ .name = n, .mode = (m) }

#define BRANCH_END { .name = NULL }

struct branch_mode {
	const char *name;
	int mode;
};

static const struct branch_mode branch_modes[] = {
	BRANCH_OPT("u", PERF_SAMPLE_BRANCH_USER),
	BRANCH_OPT("k", PERF_SAMPLE_BRANCH_KERNEL),
	BRANCH_OPT("hv", PERF_SAMPLE_BRANCH_HV),
	BRANCH_OPT("any", PERF_SAMPLE_BRANCH_ANY),
	BRANCH_OPT("any_call", PERF_SAMPLE_BRANCH_ANY_CALL),
	BRANCH_OPT("any_ret", PERF_SAMPLE_BRANCH_ANY_RETURN),
	BRANCH_OPT("ind_call", PERF_SAMPLE_BRANCH_IND_CALL),
589 590 591
	BRANCH_OPT("abort_tx", PERF_SAMPLE_BRANCH_ABORT_TX),
	BRANCH_OPT("in_tx", PERF_SAMPLE_BRANCH_IN_TX),
	BRANCH_OPT("no_tx", PERF_SAMPLE_BRANCH_NO_TX),
592 593 594 595
	BRANCH_END
};

static int
596
parse_branch_stack(const struct option *opt, const char *str, int unset)
597 598 599 600 601 602 603 604
{
#define ONLY_PLM \
	(PERF_SAMPLE_BRANCH_USER	|\
	 PERF_SAMPLE_BRANCH_KERNEL	|\
	 PERF_SAMPLE_BRANCH_HV)

	uint64_t *mode = (uint64_t *)opt->value;
	const struct branch_mode *br;
605
	char *s, *os = NULL, *p;
606 607
	int ret = -1;

608 609
	if (unset)
		return 0;
610

611 612 613 614
	/*
	 * cannot set it twice, -b + --branch-filter for instance
	 */
	if (*mode)
615 616
		return -1;

617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637
	/* str may be NULL in case no arg is passed to -b */
	if (str) {
		/* because str is read-only */
		s = os = strdup(str);
		if (!s)
			return -1;

		for (;;) {
			p = strchr(s, ',');
			if (p)
				*p = '\0';

			for (br = branch_modes; br->name; br++) {
				if (!strcasecmp(s, br->name))
					break;
			}
			if (!br->name) {
				ui__warning("unknown branch filter %s,"
					    " check man page\n", s);
				goto error;
			}
638

639
			*mode |= br->mode;
640

641 642
			if (!p)
				break;
643

644 645
			s = p + 1;
		}
646 647 648
	}
	ret = 0;

649
	/* default to any branch */
650
	if ((*mode & ~ONLY_PLM) == 0) {
651
		*mode = PERF_SAMPLE_BRANCH_ANY;
652 653 654 655 656 657
	}
error:
	free(os);
	return ret;
}

658
#ifdef HAVE_LIBUNWIND_SUPPORT
659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683
static int get_stack_size(char *str, unsigned long *_size)
{
	char *endptr;
	unsigned long size;
	unsigned long max_size = round_down(USHRT_MAX, sizeof(u64));

	size = strtoul(str, &endptr, 0);

	do {
		if (*endptr)
			break;

		size = round_up(size, sizeof(u64));
		if (!size || size > max_size)
			break;

		*_size = size;
		return 0;

	} while (0);

	pr_err("callchain: Incorrect stack dump size (max %ld): %s\n",
	       max_size, str);
	return -1;
}
684
#endif /* HAVE_LIBUNWIND_SUPPORT */
685

686
int record_parse_callchain(const char *arg, struct record_opts *opts)
687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705
{
	char *tok, *name, *saveptr = NULL;
	char *buf;
	int ret = -1;

	/* We need buffer that we know we can write to. */
	buf = malloc(strlen(arg) + 1);
	if (!buf)
		return -ENOMEM;

	strcpy(buf, arg);

	tok = strtok_r((char *)buf, ",", &saveptr);
	name = tok ? : (char *)buf;

	do {
		/* Framepointer style */
		if (!strncmp(name, "fp", sizeof("fp"))) {
			if (!strtok_r(NULL, ",", &saveptr)) {
706
				opts->call_graph = CALLCHAIN_FP;
707 708 709 710 711 712
				ret = 0;
			} else
				pr_err("callchain: No more arguments "
				       "needed for -g fp\n");
			break;

713
#ifdef HAVE_LIBUNWIND_SUPPORT
714 715
		/* Dwarf style */
		} else if (!strncmp(name, "dwarf", sizeof("dwarf"))) {
716 717
			const unsigned long default_stack_dump_size = 8192;

718
			ret = 0;
719 720
			opts->call_graph = CALLCHAIN_DWARF;
			opts->stack_dump_size = default_stack_dump_size;
721 722 723 724 725 726

			tok = strtok_r(NULL, ",", &saveptr);
			if (tok) {
				unsigned long size = 0;

				ret = get_stack_size(tok, &size);
727
				opts->stack_dump_size = size;
728
			}
729
#endif /* HAVE_LIBUNWIND_SUPPORT */
730
		} else {
J
Jiri Olsa 已提交
731
			pr_err("callchain: Unknown --call-graph option "
732 733 734 735 736 737 738
			       "value: %s\n", arg);
			break;
		}

	} while (0);

	free(buf);
J
Jiri Olsa 已提交
739 740 741
	return ret;
}

742
static void callchain_debug(struct record_opts *opts)
J
Jiri Olsa 已提交
743 744
{
	pr_debug("callchain: type %d\n", opts->call_graph);
745

J
Jiri Olsa 已提交
746 747 748 749 750 751 752 753 754
	if (opts->call_graph == CALLCHAIN_DWARF)
		pr_debug("callchain: stack dump size %d\n",
			 opts->stack_dump_size);
}

int record_parse_callchain_opt(const struct option *opt,
			       const char *arg,
			       int unset)
{
755
	struct record_opts *opts = opt->value;
J
Jiri Olsa 已提交
756 757 758 759 760 761 762 763 764 765
	int ret;

	/* --no-call-graph */
	if (unset) {
		opts->call_graph = CALLCHAIN_NONE;
		pr_debug("callchain: disabled\n");
		return 0;
	}

	ret = record_parse_callchain(arg, opts);
766
	if (!ret)
J
Jiri Olsa 已提交
767
		callchain_debug(opts);
768 769 770 771

	return ret;
}

J
Jiri Olsa 已提交
772 773 774 775
int record_callchain_opt(const struct option *opt,
			 const char *arg __maybe_unused,
			 int unset __maybe_unused)
{
776
	struct record_opts *opts = opt->value;
J
Jiri Olsa 已提交
777 778 779 780 781 782 783 784

	if (opts->call_graph == CALLCHAIN_NONE)
		opts->call_graph = CALLCHAIN_FP;

	callchain_debug(opts);
	return 0;
}

785
static const char * const record_usage[] = {
786 787
	"perf record [<options>] [<command>]",
	"perf record [<options>] -- <command> [<options>]",
788 789 790
	NULL
};

791
/*
792 793
 * XXX Ideally would be local to cmd_record() and passed to a record__new
 * because we need to have access to it in record__exit, that is called
794 795 796 797 798 799 800
 * after cmd_record() exits, but since record_options need to be accessible to
 * builtin-script, leave it here.
 *
 * At least we don't ouch it in all the other functions here directly.
 *
 * Just say no to tons of global variables, sigh.
 */
801
static struct record record = {
802 803 804 805
	.opts = {
		.mmap_pages	     = UINT_MAX,
		.user_freq	     = UINT_MAX,
		.user_interval	     = ULLONG_MAX,
806
		.freq		     = 4000,
N
Namhyung Kim 已提交
807 808
		.target		     = {
			.uses_mmap   = true,
809
			.default_per_cpu = true,
N
Namhyung Kim 已提交
810
		},
811 812
	},
};
813

J
Jiri Olsa 已提交
814
#define CALLCHAIN_HELP "setup and enables call-graph (stack chain/backtrace) recording: "
815

816
#ifdef HAVE_LIBUNWIND_SUPPORT
J
Jiri Olsa 已提交
817
const char record_callchain_help[] = CALLCHAIN_HELP "fp dwarf";
818
#else
J
Jiri Olsa 已提交
819
const char record_callchain_help[] = CALLCHAIN_HELP "fp";
820 821
#endif

822 823 824
/*
 * XXX Will stay a global variable till we fix builtin-script.c to stop messing
 * with it and switch to use the library functions in perf_evlist that came
825
 * from builtin-record.c, i.e. use record_opts,
826 827 828
 * perf_evlist__prepare_workload, etc instead of fork+exec'in 'perf record',
 * using pipes, etc.
 */
829
const struct option record_options[] = {
830
	OPT_CALLBACK('e', "event", &record.evlist, "event",
831
		     "event selector. use 'perf list' to list available events",
832
		     parse_events_option),
833
	OPT_CALLBACK(0, "filter", &record.evlist, "filter",
L
Li Zefan 已提交
834
		     "event filter", parse_filter),
835
	OPT_STRING('p', "pid", &record.opts.target.pid, "pid",
836
		    "record events on existing process id"),
837
	OPT_STRING('t', "tid", &record.opts.target.tid, "tid",
838
		    "record events on existing thread id"),
839
	OPT_INTEGER('r', "realtime", &record.realtime_prio,
840
		    "collect data with this RT SCHED_FIFO priority"),
841
	OPT_BOOLEAN(0, "no-buffering", &record.opts.no_buffering,
842
		    "collect data without buffering"),
843
	OPT_BOOLEAN('R', "raw-samples", &record.opts.raw_samples,
844
		    "collect raw sample records from all opened counters"),
845
	OPT_BOOLEAN('a', "all-cpus", &record.opts.target.system_wide,
846
			    "system-wide collection from all CPUs"),
847
	OPT_STRING('C', "cpu", &record.opts.target.cpu_list, "cpu",
848
		    "list of cpus to monitor"),
849
	OPT_U64('c', "count", &record.opts.user_interval, "event period to sample"),
850
	OPT_STRING('o', "output", &record.file.path, "file",
I
Ingo Molnar 已提交
851
		    "output file name"),
852 853 854
	OPT_BOOLEAN_SET('i', "no-inherit", &record.opts.no_inherit,
			&record.opts.no_inherit_set,
			"child tasks do not inherit counters"),
855
	OPT_UINTEGER('F', "freq", &record.opts.user_freq, "profile at this frequency"),
856 857 858
	OPT_CALLBACK('m', "mmap-pages", &record.opts.mmap_pages, "pages",
		     "number of mmap data pages",
		     perf_evlist__parse_mmap_pages),
859
	OPT_BOOLEAN(0, "group", &record.opts.group,
860
		    "put the counters into a counter group"),
J
Jiri Olsa 已提交
861 862 863 864 865 866
	OPT_CALLBACK_NOOPT('g', NULL, &record.opts,
			   NULL, "enables call-graph recording" ,
			   &record_callchain_opt),
	OPT_CALLBACK(0, "call-graph", &record.opts,
		     "mode[,dump_size]", record_callchain_help,
		     &record_parse_callchain_opt),
867
	OPT_INCR('v', "verbose", &verbose,
868
		    "be more verbose (show counter open errors, etc)"),
869
	OPT_BOOLEAN('q', "quiet", &quiet, "don't print any message"),
870
	OPT_BOOLEAN('s', "stat", &record.opts.inherit_stat,
871
		    "per thread counts"),
872
	OPT_BOOLEAN('d', "data", &record.opts.sample_address,
873
		    "Sample addresses"),
874
	OPT_BOOLEAN('T', "timestamp", &record.opts.sample_time, "Sample timestamps"),
875
	OPT_BOOLEAN('P', "period", &record.opts.period, "Sample period"),
876
	OPT_BOOLEAN('n', "no-samples", &record.opts.no_samples,
877
		    "don't sample"),
878
	OPT_BOOLEAN('N', "no-buildid-cache", &record.no_buildid_cache,
879
		    "do not update the buildid cache"),
880
	OPT_BOOLEAN('B', "no-buildid", &record.no_buildid,
881
		    "do not collect buildids in perf.data"),
882
	OPT_CALLBACK('G', "cgroup", &record.evlist, "name",
S
Stephane Eranian 已提交
883 884
		     "monitor event in cgroup name only",
		     parse_cgroups),
885
	OPT_UINTEGER('D', "delay", &record.opts.initial_delay,
886
		  "ms to wait before starting measurement after program start"),
887 888
	OPT_STRING('u', "uid", &record.opts.target.uid_str, "user",
		   "user to profile"),
889 890 891 892 893 894 895

	OPT_CALLBACK_NOOPT('b', "branch-any", &record.opts.branch_stack,
		     "branch any", "sample any taken branches",
		     parse_branch_stack),

	OPT_CALLBACK('j', "branch-filter", &record.opts.branch_stack,
		     "branch filter mask", "branch stack filter modes",
896
		     parse_branch_stack),
897 898
	OPT_BOOLEAN('W', "weight", &record.opts.sample_weight,
		    "sample by weight (on special events only)"),
899 900
	OPT_BOOLEAN(0, "transaction", &record.opts.sample_transaction,
		    "sample transaction flags (special events only)"),
901 902
	OPT_BOOLEAN(0, "per-thread", &record.opts.target.per_thread,
		    "use per-thread mmaps"),
903 904 905
	OPT_END()
};

906
int cmd_record(int argc, const char **argv, const char *prefix __maybe_unused)
907
{
908
	int err = -ENOMEM;
909
	struct record *rec = &record;
910
	char errbuf[BUFSIZ];
911

912 913
	rec->evlist = perf_evlist__new();
	if (rec->evlist == NULL)
914 915
		return -ENOMEM;

916
	argc = parse_options(argc, argv, record_options, record_usage,
917
			    PARSE_OPT_STOP_AT_NON_OPTION);
918
	if (!argc && target__none(&rec->opts.target))
919
		usage_with_options(record_usage, record_options);
920

921
	if (nr_cgroups && !rec->opts.target.system_wide) {
922 923
		ui__error("cgroup monitoring only available in"
			  " system-wide mode\n");
S
Stephane Eranian 已提交
924 925 926
		usage_with_options(record_usage, record_options);
	}

927
	symbol__init();
928

929
	if (symbol_conf.kptr_restrict)
930 931 932 933 934 935 936 937
		pr_warning(
"WARNING: Kernel address maps (/proc/{kallsyms,modules}) are restricted,\n"
"check /proc/sys/kernel/kptr_restrict.\n\n"
"Samples in kernel functions may not be resolved if a suitable vmlinux\n"
"file is not found in the buildid cache or in the vmlinux path.\n\n"
"Samples in kernel modules won't be resolved at all.\n\n"
"If some relocation was applied (e.g. kexec) symbols may be misresolved\n"
"even with a suitable vmlinux or kallsyms file.\n\n");
938

939
	if (rec->no_buildid_cache || rec->no_buildid)
940
		disable_buildid_cache();
941

942 943
	if (rec->evlist->nr_entries == 0 &&
	    perf_evlist__add_default(rec->evlist) < 0) {
944 945
		pr_err("Not enough memory for event selector list\n");
		goto out_symbol_exit;
946
	}
947

948 949 950
	if (rec->opts.target.tid && !rec->opts.no_inherit_set)
		rec->opts.no_inherit = true;

951
	err = target__validate(&rec->opts.target);
952
	if (err) {
953
		target__strerror(&rec->opts.target, err, errbuf, BUFSIZ);
954 955 956
		ui__warning("%s", errbuf);
	}

957
	err = target__parse_uid(&rec->opts.target);
958 959
	if (err) {
		int saved_errno = errno;
960

961
		target__strerror(&rec->opts.target, err, errbuf, BUFSIZ);
962
		ui__error("%s", errbuf);
963 964

		err = -saved_errno;
965
		goto out_symbol_exit;
966
	}
967

968
	err = -ENOMEM;
969
	if (perf_evlist__create_maps(rec->evlist, &rec->opts.target) < 0)
970
		usage_with_options(record_usage, record_options);
971

972
	if (record_opts__config(&rec->opts)) {
973
		err = -EINVAL;
974
		goto out_symbol_exit;
975 976
	}

977
	err = __cmd_record(&record, argc, argv);
978 979
out_symbol_exit:
	symbol__exit();
980
	return err;
981
}