builtin-record.c 26.9 KB
Newer Older
I
Ingo Molnar 已提交
1
/*
2 3 4 5 6
 * builtin-record.c
 *
 * Builtin record command: Record the profile of a workload
 * (or a CPU, or a PID) into the perf.data output file - for
 * later analysis via perf report.
I
Ingo Molnar 已提交
7
 */
8 9
#define _FILE_OFFSET_BITS 64

10
#include "builtin.h"
11 12 13

#include "perf.h"

14
#include "util/build-id.h"
15
#include "util/util.h"
16
#include "util/parse-options.h"
17
#include "util/parse-events.h"
18

19
#include "util/header.h"
20
#include "util/event.h"
21
#include "util/evlist.h"
22
#include "util/evsel.h"
23
#include "util/debug.h"
24
#include "util/session.h"
25
#include "util/tool.h"
26
#include "util/symbol.h"
27
#include "util/cpumap.h"
28
#include "util/thread_map.h"
29

30
#include <unistd.h>
31
#include <sched.h>
32
#include <sys/mman.h>
33

34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65
#ifndef HAVE_ON_EXIT
#ifndef ATEXIT_MAX
#define ATEXIT_MAX 32
#endif
static int __on_exit_count = 0;
typedef void (*on_exit_func_t) (int, void *);
static on_exit_func_t __on_exit_funcs[ATEXIT_MAX];
static void *__on_exit_args[ATEXIT_MAX];
static int __exitcode = 0;
static void __handle_on_exit_funcs(void);
static int on_exit(on_exit_func_t function, void *arg);
#define exit(x) (exit)(__exitcode = (x))

static int on_exit(on_exit_func_t function, void *arg)
{
	if (__on_exit_count == ATEXIT_MAX)
		return -ENOMEM;
	else if (__on_exit_count == 0)
		atexit(__handle_on_exit_funcs);
	__on_exit_funcs[__on_exit_count] = function;
	__on_exit_args[__on_exit_count++] = arg;
	return 0;
}

static void __handle_on_exit_funcs(void)
{
	int i;
	for (i = 0; i < __on_exit_count; i++)
		__on_exit_funcs[i] (__exitcode, __on_exit_args[i]);
}
#endif

66 67 68 69 70
enum write_mode_t {
	WRITE_FORCE,
	WRITE_APPEND
};

71
struct perf_record {
72
	struct perf_tool	tool;
73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89
	struct perf_record_opts	opts;
	u64			bytes_written;
	const char		*output_name;
	struct perf_evlist	*evlist;
	struct perf_session	*session;
	const char		*progname;
	int			output;
	unsigned int		page_size;
	int			realtime_prio;
	enum write_mode_t	write_mode;
	bool			no_buildid;
	bool			no_buildid_cache;
	bool			force;
	bool			file_new;
	bool			append_file;
	long			samples;
	off_t			post_processing_offset;
90
};
91

92
static void advance_output(struct perf_record *rec, size_t size)
93
{
94
	rec->bytes_written += size;
95 96
}

97
static int write_output(struct perf_record *rec, void *buf, size_t size)
98 99
{
	while (size) {
100
		int ret = write(rec->output, buf, size);
101

102 103 104 105
		if (ret < 0) {
			pr_err("failed to write\n");
			return -1;
		}
106 107 108 109

		size -= ret;
		buf += ret;

110
		rec->bytes_written += ret;
111
	}
112 113

	return 0;
114 115
}

116
static int process_synthesized_event(struct perf_tool *tool,
117
				     union perf_event *event,
118 119
				     struct perf_sample *sample __maybe_unused,
				     struct machine *machine __maybe_unused)
120
{
121
	struct perf_record *rec = container_of(tool, struct perf_record, tool);
122 123 124
	if (write_output(rec, event, event->header.size) < 0)
		return -1;

125 126 127
	return 0;
}

128
static int perf_record__mmap_read(struct perf_record *rec,
129
				   struct perf_mmap *md)
130
{
131
	unsigned int head = perf_mmap__read_head(md);
132
	unsigned int old = md->prev;
133
	unsigned char *data = md->base + rec->page_size;
134 135
	unsigned long size;
	void *buf;
136
	int rc = 0;
137

138
	if (old == head)
139
		return 0;
140

141
	rec->samples++;
142 143 144 145 146 147 148

	size = head - old;

	if ((old & md->mask) + size != (head & md->mask)) {
		buf = &data[old & md->mask];
		size = md->mask + 1 - (old & md->mask);
		old += size;
149

150 151 152 153
		if (write_output(rec, buf, size) < 0) {
			rc = -1;
			goto out;
		}
154 155 156 157 158
	}

	buf = &data[old & md->mask];
	size = head - old;
	old += size;
159

160 161 162 163
	if (write_output(rec, buf, size) < 0) {
		rc = -1;
		goto out;
	}
164 165

	md->prev = old;
166
	perf_mmap__write_tail(md, old);
167 168 169

out:
	return rc;
170 171 172
}

static volatile int done = 0;
173
static volatile int signr = -1;
174
static volatile int child_finished = 0;
175

176
static void sig_handler(int sig)
177
{
178 179 180
	if (sig == SIGCHLD)
		child_finished = 1;

181
	done = 1;
182 183 184
	signr = sig;
}

185
static void perf_record__sig_exit(int exit_status __maybe_unused, void *arg)
186
{
187
	struct perf_record *rec = arg;
188 189
	int status;

190
	if (rec->evlist->workload.pid > 0) {
191
		if (!child_finished)
192
			kill(rec->evlist->workload.pid, SIGTERM);
193 194 195

		wait(&status);
		if (WIFSIGNALED(status))
196
			psignal(WTERMSIG(status), rec->progname);
197
	}
198

199
	if (signr == -1 || signr == SIGUSR1)
200 201 202 203
		return;

	signal(signr, SIG_DFL);
	kill(getpid(), signr);
204 205
}

206 207 208 209 210 211 212 213
static bool perf_evlist__equal(struct perf_evlist *evlist,
			       struct perf_evlist *other)
{
	struct perf_evsel *pos, *pair;

	if (evlist->nr_entries != other->nr_entries)
		return false;

214
	pair = perf_evlist__first(other);
215 216 217 218

	list_for_each_entry(pos, &evlist->entries, node) {
		if (memcmp(&pos->attr, &pair->attr, sizeof(pos->attr) != 0))
			return false;
219
		pair = perf_evsel__next(pair);
220 221 222 223 224
	}

	return true;
}

225
static int perf_record__open(struct perf_record *rec)
226
{
227
	char msg[512];
228
	struct perf_evsel *pos;
229 230 231
	struct perf_evlist *evlist = rec->evlist;
	struct perf_session *session = rec->session;
	struct perf_record_opts *opts = &rec->opts;
232
	int rc = 0;
233

234
	perf_evlist__config(evlist, opts);
235

236 237
	list_for_each_entry(pos, &evlist->entries, node) {
try_again:
238
		if (perf_evsel__open(pos, evlist->cpus, evlist->threads) < 0) {
239
			if (perf_evsel__fallback(pos, errno, msg, sizeof(msg))) {
240
				if (verbose)
241
					ui__warning("%s\n", msg);
242 243
				goto try_again;
			}
244

245 246 247 248
			rc = -errno;
			perf_evsel__open_strerror(pos, &opts->target,
						  errno, msg, sizeof(msg));
			ui__error("%s\n", msg);
249
			goto out;
L
Li Zefan 已提交
250 251
		}
	}
252

253
	if (perf_evlist__apply_filters(evlist)) {
254 255
		error("failed to set filter with %d (%s)\n", errno,
			strerror(errno));
256 257
		rc = -1;
		goto out;
258 259
	}

260
	if (perf_evlist__mmap(evlist, opts->mmap_pages, false) < 0) {
261 262 263 264 265 266 267
		if (errno == EPERM) {
			pr_err("Permission error mapping pages.\n"
			       "Consider increasing "
			       "/proc/sys/kernel/perf_event_mlock_kb,\n"
			       "or try again with a smaller value of -m/--mmap_pages.\n"
			       "(current value: %d)\n", opts->mmap_pages);
			rc = -errno;
268 269
		} else if (!is_power_of_2(opts->mmap_pages) &&
			   (opts->mmap_pages != UINT_MAX)) {
270 271 272 273 274 275 276
			pr_err("--mmap_pages/-m value must be a power of two.");
			rc = -EINVAL;
		} else {
			pr_err("failed to mmap with %d (%s)\n", errno, strerror(errno));
			rc = -errno;
		}
		goto out;
277
	}
278

279
	if (rec->file_new)
280 281 282 283
		session->evlist = evlist;
	else {
		if (!perf_evlist__equal(session->evlist, evlist)) {
			fprintf(stderr, "incompatible append\n");
284 285
			rc = -1;
			goto out;
286 287 288
		}
 	}

289
	perf_session__set_id_hdr_size(session);
290 291
out:
	return rc;
292 293
}

294
static int process_buildids(struct perf_record *rec)
295
{
296
	u64 size = lseek(rec->output, 0, SEEK_CUR);
297

298 299 300
	if (size == 0)
		return 0;

301 302 303
	rec->session->fd = rec->output;
	return __perf_session__process_events(rec->session, rec->post_processing_offset,
					      size - rec->post_processing_offset,
304 305 306
					      size, &build_id__mark_dso_hit_ops);
}

307
static void perf_record__exit(int status, void *arg)
308
{
309 310
	struct perf_record *rec = arg;

311 312 313
	if (status != 0)
		return;

314 315 316 317 318 319 320 321 322
	if (!rec->opts.pipe_output) {
		rec->session->header.data_size += rec->bytes_written;

		if (!rec->no_buildid)
			process_buildids(rec);
		perf_session__write_header(rec->session, rec->evlist,
					   rec->output, true);
		perf_session__delete(rec->session);
		perf_evlist__delete(rec->evlist);
323
		symbol__exit();
324
	}
325 326
}

327
static void perf_event__synthesize_guest_os(struct machine *machine, void *data)
328 329
{
	int err;
330
	struct perf_tool *tool = data;
331 332 333 334 335 336 337 338
	/*
	 *As for guest kernel when processing subcommand record&report,
	 *we arrange module mmap prior to guest kernel mmap and trigger
	 *a preload dso because default guest module symbols are loaded
	 *from guest kallsyms instead of /lib/modules/XXX/XXX. This
	 *method is used to avoid symbol missing when the first addr is
	 *in module instead of in guest kernel.
	 */
339
	err = perf_event__synthesize_modules(tool, process_synthesized_event,
340
					     machine);
341 342
	if (err < 0)
		pr_err("Couldn't record guest kernel [%d]'s reference"
343
		       " relocation symbol.\n", machine->pid);
344 345 346 347 348

	/*
	 * We use _stext for guest kernel because guest kernel's /proc/kallsyms
	 * have no _text sometimes.
	 */
349
	err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
350
						 machine, "_text");
351
	if (err < 0)
352
		err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
353
							 machine, "_stext");
354 355
	if (err < 0)
		pr_err("Couldn't record guest kernel [%d]'s reference"
356
		       " relocation symbol.\n", machine->pid);
357 358
}

359 360 361 362 363
static struct perf_event_header finished_round_event = {
	.size = sizeof(struct perf_event_header),
	.type = PERF_RECORD_FINISHED_ROUND,
};

364
static int perf_record__mmap_read_all(struct perf_record *rec)
365
{
366
	int i;
367
	int rc = 0;
368

369
	for (i = 0; i < rec->evlist->nr_mmaps; i++) {
370 371 372 373 374 375
		if (rec->evlist->mmap[i].base) {
			if (perf_record__mmap_read(rec, &rec->evlist->mmap[i]) != 0) {
				rc = -1;
				goto out;
			}
		}
376 377
	}

378
	if (perf_header__has_feat(&rec->session->header, HEADER_TRACING_DATA))
379 380 381 382 383
		rc = write_output(rec, &finished_round_event,
				  sizeof(finished_round_event));

out:
	return rc;
384 385
}

386
static int __cmd_record(struct perf_record *rec, int argc, const char **argv)
387
{
I
Ingo Molnar 已提交
388 389
	struct stat st;
	int flags;
390
	int err, output, feat;
391
	unsigned long waking = 0;
392
	const bool forks = argc > 0;
393
	struct machine *machine;
394
	struct perf_tool *tool = &rec->tool;
395 396 397 398
	struct perf_record_opts *opts = &rec->opts;
	struct perf_evlist *evsel_list = rec->evlist;
	const char *output_name = rec->output_name;
	struct perf_session *session;
399
	bool disabled = false;
400

401
	rec->progname = argv[0];
402

403
	rec->page_size = sysconf(_SC_PAGE_SIZE);
404

405
	on_exit(perf_record__sig_exit, rec);
406 407
	signal(SIGCHLD, sig_handler);
	signal(SIGINT, sig_handler);
408
	signal(SIGUSR1, sig_handler);
409

410 411
	if (!output_name) {
		if (!fstat(STDOUT_FILENO, &st) && S_ISFIFO(st.st_mode))
412
			opts->pipe_output = true;
413
		else
414
			rec->output_name = output_name = "perf.data";
415 416 417
	}
	if (output_name) {
		if (!strcmp(output_name, "-"))
418
			opts->pipe_output = true;
419
		else if (!stat(output_name, &st) && st.st_size) {
420
			if (rec->write_mode == WRITE_FORCE) {
421 422 423 424 425 426
				char oldname[PATH_MAX];
				snprintf(oldname, sizeof(oldname), "%s.old",
					 output_name);
				unlink(oldname);
				rename(output_name, oldname);
			}
427 428
		} else if (rec->write_mode == WRITE_APPEND) {
			rec->write_mode = WRITE_FORCE;
429
		}
430 431
	}

432
	flags = O_CREAT|O_RDWR;
433 434
	if (rec->write_mode == WRITE_APPEND)
		rec->file_new = 0;
I
Ingo Molnar 已提交
435 436 437
	else
		flags |= O_TRUNC;

438
	if (opts->pipe_output)
439 440 441
		output = STDOUT_FILENO;
	else
		output = open(output_name, flags, S_IRUSR | S_IWUSR);
442 443
	if (output < 0) {
		perror("failed to create output file");
444
		return -1;
445 446
	}

447 448
	rec->output = output;

449
	session = perf_session__new(output_name, O_WRONLY,
450
				    rec->write_mode == WRITE_FORCE, false, NULL);
451
	if (session == NULL) {
452 453 454 455
		pr_err("Not enough memory for reading perf file header\n");
		return -1;
	}

456 457
	rec->session = session;

458 459 460 461 462 463 464
	for (feat = HEADER_FIRST_FEATURE; feat < HEADER_LAST_FEATURE; feat++)
		perf_header__set_feat(&session->header, feat);

	if (rec->no_buildid)
		perf_header__clear_feat(&session->header, HEADER_BUILD_ID);

	if (!have_tracepoints(&evsel_list->entries))
465
		perf_header__clear_feat(&session->header, HEADER_TRACING_DATA);
466

467 468 469
	if (!rec->opts.branch_stack)
		perf_header__clear_feat(&session->header, HEADER_BRANCH_STACK);

470
	if (!rec->file_new) {
471
		err = perf_session__read_header(session, output);
472
		if (err < 0)
473
			goto out_delete_session;
474 475
	}

476
	if (forks) {
477
		err = perf_evlist__prepare_workload(evsel_list, &opts->target,
478 479
						    argv, opts->pipe_output,
						    true);
480 481 482
		if (err < 0) {
			pr_err("Couldn't run the workload!\n");
			goto out_delete_session;
483 484 485
		}
	}

486 487 488 489
	if (perf_record__open(rec) != 0) {
		err = -1;
		goto out_delete_session;
	}
490

491 492 493
	if (!evsel_list->nr_groups)
		perf_header__clear_feat(&session->header, HEADER_GROUP_DESC);

494
	/*
495
	 * perf_session__delete(session) will be called at perf_record__exit()
496
	 */
497
	on_exit(perf_record__exit, rec);
498

499
	if (opts->pipe_output) {
500 501
		err = perf_header__write_pipe(output);
		if (err < 0)
502
			goto out_delete_session;
503
	} else if (rec->file_new) {
504 505
		err = perf_session__write_header(session, evsel_list,
						 output, false);
506
		if (err < 0)
507
			goto out_delete_session;
508 509
	}

510
	if (!rec->no_buildid
511
	    && !perf_header__has_feat(&session->header, HEADER_BUILD_ID)) {
512
		pr_err("Couldn't generate buildids. "
513
		       "Use --no-buildid to profile anyway.\n");
514 515
		err = -1;
		goto out_delete_session;
516 517
	}

518
	rec->post_processing_offset = lseek(output, 0, SEEK_CUR);
519

520
	machine = &session->machines.host;
521

522
	if (opts->pipe_output) {
523
		err = perf_event__synthesize_attrs(tool, session,
524
						   process_synthesized_event);
525 526
		if (err < 0) {
			pr_err("Couldn't synthesize attrs.\n");
527
			goto out_delete_session;
528
		}
529

530
		err = perf_event__synthesize_event_types(tool, process_synthesized_event,
531
							 machine);
532 533
		if (err < 0) {
			pr_err("Couldn't synthesize event_types.\n");
534
			goto out_delete_session;
535
		}
536

537
		if (have_tracepoints(&evsel_list->entries)) {
538 539 540 541 542 543 544 545
			/*
			 * FIXME err <= 0 here actually means that
			 * there were no tracepoints so its not really
			 * an error, just that we don't need to
			 * synthesize anything.  We really have to
			 * return this more properly and also
			 * propagate errors that now are calling die()
			 */
546
			err = perf_event__synthesize_tracing_data(tool, output, evsel_list,
547
								  process_synthesized_event);
548 549
			if (err <= 0) {
				pr_err("Couldn't record tracing data.\n");
550
				goto out_delete_session;
551
			}
552
			advance_output(rec, err);
553
		}
554 555
	}

556
	err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
557
						 machine, "_text");
558
	if (err < 0)
559
		err = perf_event__synthesize_kernel_mmap(tool, process_synthesized_event,
560
							 machine, "_stext");
561 562 563 564
	if (err < 0)
		pr_err("Couldn't record kernel reference relocation symbol\n"
		       "Symbol resolution may be skewed if relocation was used (e.g. kexec).\n"
		       "Check /proc/kallsyms permission or run as root.\n");
565

566
	err = perf_event__synthesize_modules(tool, process_synthesized_event,
567
					     machine);
568 569 570 571 572
	if (err < 0)
		pr_err("Couldn't record kernel module information.\n"
		       "Symbol resolution may be skewed if relocation was used (e.g. kexec).\n"
		       "Check /proc/modules permission or run as root.\n");

573
	if (perf_guest) {
574 575
		machines__process_guests(&session->machines,
					 perf_event__synthesize_guest_os, tool);
576
	}
577

J
Jiri Olsa 已提交
578
	if (perf_target__has_task(&opts->target))
579
		err = perf_event__synthesize_thread_map(tool, evsel_list->threads,
580
						  process_synthesized_event,
581
						  machine);
J
Jiri Olsa 已提交
582
	else if (perf_target__has_cpu(&opts->target))
583
		err = perf_event__synthesize_threads(tool, process_synthesized_event,
584
					       machine);
J
Jiri Olsa 已提交
585 586
	else /* command specified */
		err = 0;
587

588 589 590
	if (err != 0)
		goto out_delete_session;

591
	if (rec->realtime_prio) {
592 593
		struct sched_param param;

594
		param.sched_priority = rec->realtime_prio;
595
		if (sched_setscheduler(0, SCHED_FIFO, &param)) {
596
			pr_err("Could not set realtime priority.\n");
597 598
			err = -1;
			goto out_delete_session;
599 600 601
		}
	}

602 603 604 605 606 607 608
	/*
	 * When perf is starting the traced process, all the events
	 * (apart from group members) have enable_on_exec=1 set,
	 * so don't spoil it by prematurely enabling them.
	 */
	if (!perf_target__none(&opts->target))
		perf_evlist__enable(evsel_list);
609

610 611 612
	/*
	 * Let the child rip
	 */
613
	if (forks)
614
		perf_evlist__start_workload(evsel_list);
615

616
	for (;;) {
617
		int hits = rec->samples;
618

619 620 621 622
		if (perf_record__mmap_read_all(rec) < 0) {
			err = -1;
			goto out_delete_session;
		}
623

624
		if (hits == rec->samples) {
625 626
			if (done)
				break;
627
			err = poll(evsel_list->pollfd, evsel_list->nr_fds, -1);
628 629 630
			waking++;
		}

631 632 633 634 635
		/*
		 * When perf is starting the traced process, at the end events
		 * die with the process and we wait for that. Thus no need to
		 * disable events in this case.
		 */
636
		if (done && !disabled && !perf_target__none(&opts->target)) {
637
			perf_evlist__disable(evsel_list);
638 639
			disabled = true;
		}
640 641
	}

642
	if (quiet || signr == SIGUSR1)
643 644
		return 0;

645 646
	fprintf(stderr, "[ perf record: Woken up %ld times to write data ]\n", waking);

647 648 649 650
	/*
	 * Approximate RIP event size: 24 bytes.
	 */
	fprintf(stderr,
651
		"[ perf record: Captured and wrote %.3f MB %s (~%" PRIu64 " samples) ]\n",
652
		(double)rec->bytes_written / 1024.0 / 1024.0,
653
		output_name,
654
		rec->bytes_written / 24);
655

656
	return 0;
657 658 659 660

out_delete_session:
	perf_session__delete(session);
	return err;
661
}
662

663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684
#define BRANCH_OPT(n, m) \
	{ .name = n, .mode = (m) }

#define BRANCH_END { .name = NULL }

struct branch_mode {
	const char *name;
	int mode;
};

static const struct branch_mode branch_modes[] = {
	BRANCH_OPT("u", PERF_SAMPLE_BRANCH_USER),
	BRANCH_OPT("k", PERF_SAMPLE_BRANCH_KERNEL),
	BRANCH_OPT("hv", PERF_SAMPLE_BRANCH_HV),
	BRANCH_OPT("any", PERF_SAMPLE_BRANCH_ANY),
	BRANCH_OPT("any_call", PERF_SAMPLE_BRANCH_ANY_CALL),
	BRANCH_OPT("any_ret", PERF_SAMPLE_BRANCH_ANY_RETURN),
	BRANCH_OPT("ind_call", PERF_SAMPLE_BRANCH_IND_CALL),
	BRANCH_END
};

static int
685
parse_branch_stack(const struct option *opt, const char *str, int unset)
686 687 688 689 690 691 692 693
{
#define ONLY_PLM \
	(PERF_SAMPLE_BRANCH_USER	|\
	 PERF_SAMPLE_BRANCH_KERNEL	|\
	 PERF_SAMPLE_BRANCH_HV)

	uint64_t *mode = (uint64_t *)opt->value;
	const struct branch_mode *br;
694
	char *s, *os = NULL, *p;
695 696
	int ret = -1;

697 698
	if (unset)
		return 0;
699

700 701 702 703
	/*
	 * cannot set it twice, -b + --branch-filter for instance
	 */
	if (*mode)
704 705
		return -1;

706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726
	/* str may be NULL in case no arg is passed to -b */
	if (str) {
		/* because str is read-only */
		s = os = strdup(str);
		if (!s)
			return -1;

		for (;;) {
			p = strchr(s, ',');
			if (p)
				*p = '\0';

			for (br = branch_modes; br->name; br++) {
				if (!strcasecmp(s, br->name))
					break;
			}
			if (!br->name) {
				ui__warning("unknown branch filter %s,"
					    " check man page\n", s);
				goto error;
			}
727

728
			*mode |= br->mode;
729

730 731
			if (!p)
				break;
732

733 734
			s = p + 1;
		}
735 736 737
	}
	ret = 0;

738
	/* default to any branch */
739
	if ((*mode & ~ONLY_PLM) == 0) {
740
		*mode = PERF_SAMPLE_BRANCH_ANY;
741 742 743 744 745 746
	}
error:
	free(os);
	return ret;
}

747
#ifdef LIBUNWIND_SUPPORT
748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772
static int get_stack_size(char *str, unsigned long *_size)
{
	char *endptr;
	unsigned long size;
	unsigned long max_size = round_down(USHRT_MAX, sizeof(u64));

	size = strtoul(str, &endptr, 0);

	do {
		if (*endptr)
			break;

		size = round_up(size, sizeof(u64));
		if (!size || size > max_size)
			break;

		*_size = size;
		return 0;

	} while (0);

	pr_err("callchain: Incorrect stack dump size (max %ld): %s\n",
	       max_size, str);
	return -1;
}
773
#endif /* LIBUNWIND_SUPPORT */
774

775 776
int record_parse_callchain_opt(const struct option *opt,
			       const char *arg, int unset)
777
{
778
	struct perf_record_opts *opts = opt->value;
779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803
	char *tok, *name, *saveptr = NULL;
	char *buf;
	int ret = -1;

	/* --no-call-graph */
	if (unset)
		return 0;

	/* We specified default option if none is provided. */
	BUG_ON(!arg);

	/* We need buffer that we know we can write to. */
	buf = malloc(strlen(arg) + 1);
	if (!buf)
		return -ENOMEM;

	strcpy(buf, arg);

	tok = strtok_r((char *)buf, ",", &saveptr);
	name = tok ? : (char *)buf;

	do {
		/* Framepointer style */
		if (!strncmp(name, "fp", sizeof("fp"))) {
			if (!strtok_r(NULL, ",", &saveptr)) {
804
				opts->call_graph = CALLCHAIN_FP;
805 806 807 808 809 810
				ret = 0;
			} else
				pr_err("callchain: No more arguments "
				       "needed for -g fp\n");
			break;

811
#ifdef LIBUNWIND_SUPPORT
812 813
		/* Dwarf style */
		} else if (!strncmp(name, "dwarf", sizeof("dwarf"))) {
814 815
			const unsigned long default_stack_dump_size = 8192;

816
			ret = 0;
817 818
			opts->call_graph = CALLCHAIN_DWARF;
			opts->stack_dump_size = default_stack_dump_size;
819 820 821 822 823 824

			tok = strtok_r(NULL, ",", &saveptr);
			if (tok) {
				unsigned long size = 0;

				ret = get_stack_size(tok, &size);
825
				opts->stack_dump_size = size;
826 827 828 829
			}

			if (!ret)
				pr_debug("callchain: stack dump size %d\n",
830
					 opts->stack_dump_size);
831
#endif /* LIBUNWIND_SUPPORT */
832 833 834 835 836 837 838 839 840 841 842
		} else {
			pr_err("callchain: Unknown -g option "
			       "value: %s\n", arg);
			break;
		}

	} while (0);

	free(buf);

	if (!ret)
843
		pr_debug("callchain: type %d\n", opts->call_graph);
844 845 846 847

	return ret;
}

848
static const char * const record_usage[] = {
849 850
	"perf record [<options>] [<command>]",
	"perf record [<options>] -- <command> [<options>]",
851 852 853
	NULL
};

854 855 856 857 858 859 860 861 862 863 864 865 866 867 868
/*
 * XXX Ideally would be local to cmd_record() and passed to a perf_record__new
 * because we need to have access to it in perf_record__exit, that is called
 * after cmd_record() exits, but since record_options need to be accessible to
 * builtin-script, leave it here.
 *
 * At least we don't ouch it in all the other functions here directly.
 *
 * Just say no to tons of global variables, sigh.
 */
static struct perf_record record = {
	.opts = {
		.mmap_pages	     = UINT_MAX,
		.user_freq	     = UINT_MAX,
		.user_interval	     = ULLONG_MAX,
869
		.freq		     = 4000,
N
Namhyung Kim 已提交
870 871 872
		.target		     = {
			.uses_mmap   = true,
		},
873 874 875 876
	},
	.write_mode = WRITE_FORCE,
	.file_new   = true,
};
877

878 879 880
#define CALLCHAIN_HELP "do call-graph (stack chain/backtrace) recording: "

#ifdef LIBUNWIND_SUPPORT
881
const char record_callchain_help[] = CALLCHAIN_HELP "[fp] dwarf";
882
#else
883
const char record_callchain_help[] = CALLCHAIN_HELP "[fp]";
884 885
#endif

886 887 888 889 890 891 892
/*
 * XXX Will stay a global variable till we fix builtin-script.c to stop messing
 * with it and switch to use the library functions in perf_evlist that came
 * from builtin-record.c, i.e. use perf_record_opts,
 * perf_evlist__prepare_workload, etc instead of fork+exec'in 'perf record',
 * using pipes, etc.
 */
893
const struct option record_options[] = {
894
	OPT_CALLBACK('e', "event", &record.evlist, "event",
895
		     "event selector. use 'perf list' to list available events",
896
		     parse_events_option),
897
	OPT_CALLBACK(0, "filter", &record.evlist, "filter",
L
Li Zefan 已提交
898
		     "event filter", parse_filter),
899
	OPT_STRING('p', "pid", &record.opts.target.pid, "pid",
900
		    "record events on existing process id"),
901
	OPT_STRING('t', "tid", &record.opts.target.tid, "tid",
902
		    "record events on existing thread id"),
903
	OPT_INTEGER('r', "realtime", &record.realtime_prio,
904
		    "collect data with this RT SCHED_FIFO priority"),
905
	OPT_BOOLEAN('D', "no-delay", &record.opts.no_delay,
906
		    "collect data without buffering"),
907
	OPT_BOOLEAN('R', "raw-samples", &record.opts.raw_samples,
908
		    "collect raw sample records from all opened counters"),
909
	OPT_BOOLEAN('a', "all-cpus", &record.opts.target.system_wide,
910
			    "system-wide collection from all CPUs"),
911
	OPT_BOOLEAN('A', "append", &record.append_file,
I
Ingo Molnar 已提交
912
			    "append to the output file to do incremental profiling"),
913
	OPT_STRING('C', "cpu", &record.opts.target.cpu_list, "cpu",
914
		    "list of cpus to monitor"),
915
	OPT_BOOLEAN('f', "force", &record.force,
916
			"overwrite existing data file (deprecated)"),
917 918
	OPT_U64('c', "count", &record.opts.user_interval, "event period to sample"),
	OPT_STRING('o', "output", &record.output_name, "file",
I
Ingo Molnar 已提交
919
		    "output file name"),
920
	OPT_BOOLEAN('i', "no-inherit", &record.opts.no_inherit,
921
		    "child tasks do not inherit counters"),
922 923
	OPT_UINTEGER('F', "freq", &record.opts.user_freq, "profile at this frequency"),
	OPT_UINTEGER('m', "mmap-pages", &record.opts.mmap_pages,
924
		     "number of mmap data pages"),
925
	OPT_BOOLEAN(0, "group", &record.opts.group,
926
		    "put the counters into a counter group"),
927
	OPT_CALLBACK_DEFAULT('g', "call-graph", &record.opts,
928 929
			     "mode[,dump_size]", record_callchain_help,
			     &record_parse_callchain_opt, "fp"),
930
	OPT_INCR('v', "verbose", &verbose,
931
		    "be more verbose (show counter open errors, etc)"),
932
	OPT_BOOLEAN('q', "quiet", &quiet, "don't print any message"),
933
	OPT_BOOLEAN('s', "stat", &record.opts.inherit_stat,
934
		    "per thread counts"),
935
	OPT_BOOLEAN('d', "data", &record.opts.sample_address,
936
		    "Sample addresses"),
937
	OPT_BOOLEAN('T', "timestamp", &record.opts.sample_time, "Sample timestamps"),
938
	OPT_BOOLEAN('P', "period", &record.opts.period, "Sample period"),
939
	OPT_BOOLEAN('n', "no-samples", &record.opts.no_samples,
940
		    "don't sample"),
941
	OPT_BOOLEAN('N', "no-buildid-cache", &record.no_buildid_cache,
942
		    "do not update the buildid cache"),
943
	OPT_BOOLEAN('B', "no-buildid", &record.no_buildid,
944
		    "do not collect buildids in perf.data"),
945
	OPT_CALLBACK('G', "cgroup", &record.evlist, "name",
S
Stephane Eranian 已提交
946 947
		     "monitor event in cgroup name only",
		     parse_cgroups),
948 949
	OPT_STRING('u', "uid", &record.opts.target.uid_str, "user",
		   "user to profile"),
950 951 952 953 954 955 956

	OPT_CALLBACK_NOOPT('b', "branch-any", &record.opts.branch_stack,
		     "branch any", "sample any taken branches",
		     parse_branch_stack),

	OPT_CALLBACK('j', "branch-filter", &record.opts.branch_stack,
		     "branch filter mask", "branch stack filter modes",
957
		     parse_branch_stack),
958 959 960
	OPT_END()
};

961
int cmd_record(int argc, const char **argv, const char *prefix __maybe_unused)
962
{
963 964
	int err = -ENOMEM;
	struct perf_evsel *pos;
965 966
	struct perf_evlist *evsel_list;
	struct perf_record *rec = &record;
967
	char errbuf[BUFSIZ];
968

969
	evsel_list = perf_evlist__new();
970 971 972
	if (evsel_list == NULL)
		return -ENOMEM;

973 974
	rec->evlist = evsel_list;

975
	argc = parse_options(argc, argv, record_options, record_usage,
976
			    PARSE_OPT_STOP_AT_NON_OPTION);
977
	if (!argc && perf_target__none(&rec->opts.target))
978
		usage_with_options(record_usage, record_options);
979

980
	if (rec->force && rec->append_file) {
981 982
		ui__error("Can't overwrite and append at the same time."
			  " You need to choose between -f and -A");
983
		usage_with_options(record_usage, record_options);
984 985
	} else if (rec->append_file) {
		rec->write_mode = WRITE_APPEND;
986
	} else {
987
		rec->write_mode = WRITE_FORCE;
988 989
	}

990
	if (nr_cgroups && !rec->opts.target.system_wide) {
991 992
		ui__error("cgroup monitoring only available in"
			  " system-wide mode\n");
S
Stephane Eranian 已提交
993 994 995
		usage_with_options(record_usage, record_options);
	}

996
	symbol__init();
997

998
	if (symbol_conf.kptr_restrict)
999 1000 1001 1002 1003 1004 1005 1006
		pr_warning(
"WARNING: Kernel address maps (/proc/{kallsyms,modules}) are restricted,\n"
"check /proc/sys/kernel/kptr_restrict.\n\n"
"Samples in kernel functions may not be resolved if a suitable vmlinux\n"
"file is not found in the buildid cache or in the vmlinux path.\n\n"
"Samples in kernel modules won't be resolved at all.\n\n"
"If some relocation was applied (e.g. kexec) symbols may be misresolved\n"
"even with a suitable vmlinux or kallsyms file.\n\n");
1007

1008
	if (rec->no_buildid_cache || rec->no_buildid)
1009
		disable_buildid_cache();
1010

1011 1012
	if (evsel_list->nr_entries == 0 &&
	    perf_evlist__add_default(evsel_list) < 0) {
1013 1014
		pr_err("Not enough memory for event selector list\n");
		goto out_symbol_exit;
1015
	}
1016

1017 1018 1019 1020 1021 1022 1023 1024 1025
	err = perf_target__validate(&rec->opts.target);
	if (err) {
		perf_target__strerror(&rec->opts.target, err, errbuf, BUFSIZ);
		ui__warning("%s", errbuf);
	}

	err = perf_target__parse_uid(&rec->opts.target);
	if (err) {
		int saved_errno = errno;
1026

1027
		perf_target__strerror(&rec->opts.target, err, errbuf, BUFSIZ);
1028
		ui__error("%s", errbuf);
1029 1030

		err = -saved_errno;
1031
		goto out_symbol_exit;
1032
	}
1033

1034
	err = -ENOMEM;
1035
	if (perf_evlist__create_maps(evsel_list, &rec->opts.target) < 0)
1036
		usage_with_options(record_usage, record_options);
1037

1038
	list_for_each_entry(pos, &evsel_list->entries, node) {
1039
		if (perf_header__push_event(pos->attr.config, perf_evsel__name(pos)))
1040
			goto out_free_fd;
1041
	}
1042

1043 1044 1045 1046
	if (rec->opts.user_interval != ULLONG_MAX)
		rec->opts.default_interval = rec->opts.user_interval;
	if (rec->opts.user_freq != UINT_MAX)
		rec->opts.freq = rec->opts.user_freq;
1047

1048 1049 1050
	/*
	 * User specified count overrides default frequency.
	 */
1051 1052 1053 1054
	if (rec->opts.default_interval)
		rec->opts.freq = 0;
	else if (rec->opts.freq) {
		rec->opts.default_interval = rec->opts.freq;
1055
	} else {
1056
		ui__error("frequency and count are zero, aborting\n");
1057
		err = -EINVAL;
1058
		goto out_free_fd;
1059 1060
	}

1061
	err = __cmd_record(&record, argc, argv);
1062 1063 1064

	perf_evlist__munmap(evsel_list);
	perf_evlist__close(evsel_list);
1065
out_free_fd:
1066
	perf_evlist__delete_maps(evsel_list);
1067 1068
out_symbol_exit:
	symbol__exit();
1069
	return err;
1070
}