builtin-top.c 34.8 KB
Newer Older
1
/*
2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17
 * builtin-top.c
 *
 * Builtin top command: Display a continuously updated profile of
 * any workload, CPU or specific PID.
 *
 * Copyright (C) 2008, Red Hat Inc, Ingo Molnar <mingo@redhat.com>
 *
 * Improvements and fixes by:
 *
 *   Arjan van de Ven <arjan@linux.intel.com>
 *   Yanmin Zhang <yanmin.zhang@intel.com>
 *   Wu Fengguang <fengguang.wu@intel.com>
 *   Mike Galbraith <efault@gmx.de>
 *   Paul Mackerras <paulus@samba.org>
 *
 * Released under the GPL v2. (and only v2, not any later version)
18
 */
19
#include "builtin.h"
20

21
#include "perf.h"
22

23
#include "util/color.h"
24 25
#include "util/session.h"
#include "util/symbol.h"
26
#include "util/thread.h"
27
#include "util/util.h"
28
#include <linux/rbtree.h>
29 30
#include "util/parse-options.h"
#include "util/parse-events.h"
31
#include "util/cpumap.h"
32

33 34
#include "util/debug.h"

35 36
#include <assert.h>
#include <fcntl.h>
37

38
#include <stdio.h>
39 40
#include <termios.h>
#include <unistd.h>
41

42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57
#include <errno.h>
#include <time.h>
#include <sched.h>
#include <pthread.h>

#include <sys/syscall.h>
#include <sys/ioctl.h>
#include <sys/poll.h>
#include <sys/prctl.h>
#include <sys/wait.h>
#include <sys/uio.h>
#include <sys/mman.h>

#include <linux/unistd.h>
#include <linux/types.h>

58
static int			*fd[MAX_NR_CPUS][MAX_COUNTERS];
59

60
static bool			system_wide			=  false;
61

62
static int			default_interval		=      0;
63

64
static int			count_filter			=      5;
65
static int			print_entries;
66

67
static int			target_pid			=     -1;
68 69 70
static int			target_tid			=     -1;
static pid_t			*all_tids			=      NULL;
static int			thread_num			=      0;
71
static bool			inherit				=  false;
72
static int			nr_cpus				=      0;
73
static int			realtime_prio			=      0;
74
static bool			group				=  false;
75
static unsigned int		page_size;
76 77
static unsigned int		mmap_pages			=     16;
static int			freq				=   1000; /* 1 KHz */
78

79
static int			delay_secs			=      2;
80 81
static bool			zero                            =  false;
static bool			dump_symtab                     =  false;
82

83 84
static bool			hide_kernel_symbols		=  false;
static bool			hide_user_symbols		=  false;
85
static struct winsize		winsize;
86

87 88 89 90 91 92 93 94 95 96 97
/*
 * Source
 */

struct source_line {
	u64			eip;
	unsigned long		count[MAX_COUNTERS];
	char			*line;
	struct source_line	*next;
};

98
static const char		*sym_filter			=   NULL;
99
struct sym_entry		*sym_filter_entry		=   NULL;
100
struct sym_entry		*sym_filter_entry_sched		=   NULL;
101 102 103
static int			sym_pcnt_filter			=      5;
static int			sym_counter			=      0;
static int			display_weighted		=     -1;
104
static const char		*cpu_list;
105

106 107 108 109
/*
 * Symbols
 */

110 111 112 113 114 115 116
struct sym_entry_source {
	struct source_line	*source;
	struct source_line	*lines;
	struct source_line	**lines_tail;
	pthread_mutex_t		lock;
};

117
struct sym_entry {
118 119
	struct rb_node		rb_node;
	struct list_head	node;
120 121
	unsigned long		snap_count;
	double			weight;
122
	int			skip;
123
	u16			name_len;
124
	u8			origin;
125
	struct map		*map;
126
	struct sym_entry_source	*src;
127
	unsigned long		count[0];
128 129
};

130 131 132 133
/*
 * Source functions
 */

134 135
static inline struct symbol *sym_entry__symbol(struct sym_entry *self)
{
136
       return ((void *)self) + symbol_conf.priv_size;
137 138
}

139
void get_term_dimensions(struct winsize *ws)
140
{
141 142 143 144 145 146 147 148 149 150
	char *s = getenv("LINES");

	if (s != NULL) {
		ws->ws_row = atoi(s);
		s = getenv("COLUMNS");
		if (s != NULL) {
			ws->ws_col = atoi(s);
			if (ws->ws_row && ws->ws_col)
				return;
		}
151
	}
152 153 154 155
#ifdef TIOCGWINSZ
	if (ioctl(1, TIOCGWINSZ, ws) == 0 &&
	    ws->ws_row && ws->ws_col)
		return;
156
#endif
157 158
	ws->ws_row = 25;
	ws->ws_col = 80;
159 160
}

161
static void update_print_entries(struct winsize *ws)
162
{
163 164
	print_entries = ws->ws_row;

165 166 167 168 169 170
	if (print_entries > 9)
		print_entries -= 9;
}

static void sig_winch_handler(int sig __used)
{
171 172
	get_term_dimensions(&winsize);
	update_print_entries(&winsize);
173 174
}

175
static int parse_source(struct sym_entry *syme)
176 177
{
	struct symbol *sym;
178
	struct sym_entry_source *source;
179
	struct map *map;
180
	FILE *file;
181
	char command[PATH_MAX*2];
182 183
	const char *path;
	u64 len;
184 185

	if (!syme)
186 187 188 189 190 191 192 193 194 195
		return -1;

	sym = sym_entry__symbol(syme);
	map = syme->map;

	/*
	 * We can't annotate with just /proc/kallsyms
	 */
	if (map->dso->origin == DSO__ORIG_KERNEL)
		return -1;
196

197
	if (syme->src == NULL) {
198
		syme->src = zalloc(sizeof(*source));
199
		if (syme->src == NULL)
200
			return -1;
201 202 203 204 205 206 207
		pthread_mutex_init(&syme->src->lock, NULL);
	}

	source = syme->src;

	if (source->lines) {
		pthread_mutex_lock(&source->lock);
208 209
		goto out_assign;
	}
210
	path = map->dso->long_name;
211 212 213

	len = sym->end - sym->start;

214
	sprintf(command,
215 216 217
		"objdump --start-address=%#0*Lx --stop-address=%#0*Lx -dS %s",
		BITS_PER_LONG / 4, map__rip_2objdump(map, sym->start),
		BITS_PER_LONG / 4, map__rip_2objdump(map, sym->end), path);
218 219 220

	file = popen(command, "r");
	if (!file)
221
		return -1;
222

223 224
	pthread_mutex_lock(&source->lock);
	source->lines_tail = &source->lines;
225 226 227
	while (!feof(file)) {
		struct source_line *src;
		size_t dummy = 0;
228
		char *c, *sep;
229 230 231 232 233 234 235 236 237 238 239 240 241 242 243

		src = malloc(sizeof(struct source_line));
		assert(src != NULL);
		memset(src, 0, sizeof(struct source_line));

		if (getline(&src->line, &dummy, file) < 0)
			break;
		if (!src->line)
			break;

		c = strchr(src->line, '\n');
		if (c)
			*c = 0;

		src->next = NULL;
244 245
		*source->lines_tail = src;
		source->lines_tail = &src->next;
246

247 248 249 250 251
		src->eip = strtoull(src->line, &sep, 16);
		if (*sep == ':')
			src->eip = map__objdump_2ip(map, src->eip);
		else /* this line has no ip info (e.g. source line) */
			src->eip = 0;
252 253 254 255
	}
	pclose(file);
out_assign:
	sym_filter_entry = syme;
256
	pthread_mutex_unlock(&source->lock);
257
	return 0;
258 259 260 261 262 263 264
}

static void __zero_source_counters(struct sym_entry *syme)
{
	int i;
	struct source_line *line;

265
	line = syme->src->lines;
266 267 268 269 270 271 272 273 274 275 276 277 278 279
	while (line) {
		for (i = 0; i < nr_counters; i++)
			line->count[i] = 0;
		line = line->next;
	}
}

static void record_precise_ip(struct sym_entry *syme, int counter, u64 ip)
{
	struct source_line *line;

	if (syme != sym_filter_entry)
		return;

280
	if (pthread_mutex_trylock(&syme->src->lock))
281 282
		return;

283
	if (syme->src == NULL || syme->src->source == NULL)
284 285
		goto out_unlock;

286
	for (line = syme->src->lines; line; line = line->next) {
287 288 289
		/* skip lines without IP info */
		if (line->eip == 0)
			continue;
290 291 292 293 294 295 296 297
		if (line->eip == ip) {
			line->count[counter]++;
			break;
		}
		if (line->eip > ip)
			break;
	}
out_unlock:
298
	pthread_mutex_unlock(&syme->src->lock);
299 300
}

301 302
#define PATTERN_LEN		(BITS_PER_LONG / 4 + 2)

303 304
static void lookup_sym_source(struct sym_entry *syme)
{
305
	struct symbol *symbol = sym_entry__symbol(syme);
306
	struct source_line *line;
307
	char pattern[PATTERN_LEN + 1];
308

309 310
	sprintf(pattern, "%0*Lx <", BITS_PER_LONG / 4,
		map__rip_2objdump(syme->map, symbol->start));
311

312 313
	pthread_mutex_lock(&syme->src->lock);
	for (line = syme->src->lines; line; line = line->next) {
314
		if (memcmp(line->line, pattern, PATTERN_LEN) == 0) {
315
			syme->src->source = line;
316 317 318
			break;
		}
	}
319
	pthread_mutex_unlock(&syme->src->lock);
320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348
}

static void show_lines(struct source_line *queue, int count, int total)
{
	int i;
	struct source_line *line;

	line = queue;
	for (i = 0; i < count; i++) {
		float pcnt = 100.0*(float)line->count[sym_counter]/(float)total;

		printf("%8li %4.1f%%\t%s\n", line->count[sym_counter], pcnt, line->line);
		line = line->next;
	}
}

#define TRACE_COUNT     3

static void show_details(struct sym_entry *syme)
{
	struct symbol *symbol;
	struct source_line *line;
	struct source_line *line_queue = NULL;
	int displayed = 0;
	int line_queue_count = 0, total = 0, more = 0;

	if (!syme)
		return;

349
	if (!syme->src->source)
350 351
		lookup_sym_source(syme);

352
	if (!syme->src->source)
353 354
		return;

355
	symbol = sym_entry__symbol(syme);
356 357 358
	printf("Showing %s for %s\n", event_name(sym_counter), symbol->name);
	printf("  Events  Pcnt (>=%d%%)\n", sym_pcnt_filter);

359 360
	pthread_mutex_lock(&syme->src->lock);
	line = syme->src->source;
361 362 363 364 365
	while (line) {
		total += line->count[sym_counter];
		line = line->next;
	}

366
	line = syme->src->source;
367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390
	while (line) {
		float pcnt = 0.0;

		if (!line_queue_count)
			line_queue = line;
		line_queue_count++;

		if (line->count[sym_counter])
			pcnt = 100.0 * line->count[sym_counter] / (float)total;
		if (pcnt >= (float)sym_pcnt_filter) {
			if (displayed <= print_entries)
				show_lines(line_queue, line_queue_count, total);
			else more++;
			displayed += line_queue_count;
			line_queue_count = 0;
			line_queue = NULL;
		} else if (line_queue_count > TRACE_COUNT) {
			line_queue = line_queue->next;
			line_queue_count--;
		}

		line->count[sym_counter] = zero ? 0 : line->count[sym_counter] * 7 / 8;
		line = line->next;
	}
391
	pthread_mutex_unlock(&syme->src->lock);
392 393 394
	if (more)
		printf("%d lines not displayed, maybe increase display entries [e]\n", more);
}
395

396
/*
397
 * Symbols will be added here in event__process_sample and will get out
398 399 400
 * after decayed.
 */
static LIST_HEAD(active_symbols);
401
static pthread_mutex_t active_symbols_lock = PTHREAD_MUTEX_INITIALIZER;
402 403 404 405 406 407

/*
 * Ordering weight: count-1 * count-2 * ... / count-n
 */
static double sym_weight(const struct sym_entry *sym)
{
408
	double weight = sym->snap_count;
409 410
	int counter;

411 412 413
	if (!display_weighted)
		return weight;

414 415 416 417 418 419 420 421
	for (counter = 1; counter < nr_counters-1; counter++)
		weight *= sym->count[counter];

	weight /= (sym->count[counter] + 1);

	return weight;
}

422
static long			samples;
423
static long			kernel_samples, us_samples;
424
static long			exact_samples;
425
static long			guest_us_samples, guest_kernel_samples;
426 427
static const char		CONSOLE_CLEAR[] = "";

428
static void __list_insert_active_sym(struct sym_entry *syme)
429 430 431 432
{
	list_add(&syme->node, &active_symbols);
}

433 434 435 436 437 438 439
static void list_remove_active_sym(struct sym_entry *syme)
{
	pthread_mutex_lock(&active_symbols_lock);
	list_del_init(&syme->node);
	pthread_mutex_unlock(&active_symbols_lock);
}

440 441 442 443 444 445 446 447 448 449
static void rb_insert_active_sym(struct rb_root *tree, struct sym_entry *se)
{
	struct rb_node **p = &tree->rb_node;
	struct rb_node *parent = NULL;
	struct sym_entry *iter;

	while (*p != NULL) {
		parent = *p;
		iter = rb_entry(parent, struct sym_entry, rb_node);

450
		if (se->weight > iter->weight)
451 452 453 454 455 456 457 458
			p = &(*p)->rb_left;
		else
			p = &(*p)->rb_right;
	}

	rb_link_node(&se->rb_node, parent, p);
	rb_insert_color(&se->rb_node, tree);
}
459 460 461

static void print_sym_table(void)
{
462
	int printed = 0, j;
463
	int counter, snap = !display_weighted ? sym_counter : 0;
464
	float samples_per_sec = samples/delay_secs;
465 466 467 468
	float ksamples_per_sec = kernel_samples/delay_secs;
	float us_samples_per_sec = (us_samples)/delay_secs;
	float guest_kernel_samples_per_sec = (guest_kernel_samples)/delay_secs;
	float guest_us_samples_per_sec = (guest_us_samples)/delay_secs;
469
	float esamples_percent = (100.0*exact_samples)/samples;
470
	float sum_ksamples = 0.0;
471 472 473
	struct sym_entry *syme, *n;
	struct rb_root tmp = RB_ROOT;
	struct rb_node *nd;
474
	int sym_width = 0, dso_width = 0, dso_short_width = 0;
475
	const int win_width = winsize.ws_col - 1;
476

477 478
	samples = us_samples = kernel_samples = exact_samples = 0;
	guest_kernel_samples = guest_us_samples = 0;
479

480
	/* Sort the active symbols */
481 482 483 484 485
	pthread_mutex_lock(&active_symbols_lock);
	syme = list_entry(active_symbols.next, struct sym_entry, node);
	pthread_mutex_unlock(&active_symbols_lock);

	list_for_each_entry_safe_from(syme, n, &active_symbols, node) {
486
		syme->snap_count = syme->count[snap];
487
		if (syme->snap_count != 0) {
488

489 490 491 492 493 494 495
			if ((hide_user_symbols &&
			     syme->origin == PERF_RECORD_MISC_USER) ||
			    (hide_kernel_symbols &&
			     syme->origin == PERF_RECORD_MISC_KERNEL)) {
				list_remove_active_sym(syme);
				continue;
			}
496
			syme->weight = sym_weight(syme);
497
			rb_insert_active_sym(&tmp, syme);
498
			sum_ksamples += syme->snap_count;
499 500

			for (j = 0; j < nr_counters; j++)
501 502
				syme->count[j] = zero ? 0 : syme->count[j] * 7 / 8;
		} else
503
			list_remove_active_sym(syme);
504 505
	}

506
	puts(CONSOLE_CLEAR);
507

508
	printf("%-*.*s\n", win_width, win_width, graph_dotted_line);
509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532
	if (!perf_guest) {
		printf("   PerfTop:%8.0f irqs/sec  kernel:%4.1f%%"
			"  exact: %4.1f%% [",
			samples_per_sec,
			100.0 - (100.0 * ((samples_per_sec - ksamples_per_sec) /
					 samples_per_sec)),
			esamples_percent);
	} else {
		printf("   PerfTop:%8.0f irqs/sec  kernel:%4.1f%% us:%4.1f%%"
			" guest kernel:%4.1f%% guest us:%4.1f%%"
			" exact: %4.1f%% [",
			samples_per_sec,
			100.0 - (100.0 * ((samples_per_sec-ksamples_per_sec) /
					  samples_per_sec)),
			100.0 - (100.0 * ((samples_per_sec-us_samples_per_sec) /
					  samples_per_sec)),
			100.0 - (100.0 * ((samples_per_sec -
						guest_kernel_samples_per_sec) /
					  samples_per_sec)),
			100.0 - (100.0 * ((samples_per_sec -
					   guest_us_samples_per_sec) /
					  samples_per_sec)),
			esamples_percent);
	}
533

534
	if (nr_counters == 1 || !display_weighted) {
535
		printf("%Ld", (u64)attrs[0].sample_period);
I
Ingo Molnar 已提交
536 537 538 539 540
		if (freq)
			printf("Hz ");
		else
			printf(" ");
	}
541

542 543 544
	if (!display_weighted)
		printf("%s", event_name(sym_counter));
	else for (counter = 0; counter < nr_counters; counter++) {
545 546 547 548 549 550 551 552
		if (counter)
			printf("/");

		printf("%s", event_name(counter));
	}

	printf( "], ");

553 554
	if (target_pid != -1)
		printf(" (target_pid: %d", target_pid);
555 556
	else if (target_tid != -1)
		printf(" (target_tid: %d", target_tid);
557 558 559
	else
		printf(" (all");

560 561
	if (cpu_list)
		printf(", CPU%s: %s)\n", nr_cpus > 1 ? "s" : "", cpu_list);
562
	else {
563
		if (target_tid != -1)
564 565
			printf(")\n");
		else
566
			printf(", %d CPU%s)\n", nr_cpus, nr_cpus > 1 ? "s" : "");
567 568
	}

569
	printf("%-*.*s\n", win_width, win_width, graph_dotted_line);
570

571 572 573 574 575
	if (sym_filter_entry) {
		show_details(sym_filter_entry);
		return;
	}

576 577 578 579 580 581 582 583 584
	/*
	 * Find the longest symbol name that will be displayed
	 */
	for (nd = rb_first(&tmp); nd; nd = rb_next(nd)) {
		syme = rb_entry(nd, struct sym_entry, rb_node);
		if (++printed > print_entries ||
		    (int)syme->snap_count < count_filter)
			continue;

585 586 587
		if (syme->map->dso->long_name_len > dso_width)
			dso_width = syme->map->dso->long_name_len;

588 589 590
		if (syme->map->dso->short_name_len > dso_short_width)
			dso_short_width = syme->map->dso->short_name_len;

591 592 593 594 595 596
		if (syme->name_len > sym_width)
			sym_width = syme->name_len;
	}

	printed = 0;

597 598 599 600 601
	if (sym_width + dso_width > winsize.ws_col - 29) {
		dso_width = dso_short_width;
		if (sym_width + dso_width > winsize.ws_col - 29)
			sym_width = winsize.ws_col - dso_width - 29;
	}
602
	putchar('\n');
603
	if (nr_counters == 1)
604
		printf("             samples  pcnt");
605
	else
606
		printf("   weight    samples  pcnt");
607

608 609
	if (verbose)
		printf("         RIP       ");
610
	printf(" %-*.*s DSO\n", sym_width, sym_width, "function");
611
	printf("   %s    _______ _____",
612 613
	       nr_counters == 1 ? "      " : "______");
	if (verbose)
614
		printf(" ________________");
615
	printf(" %-*.*s", sym_width, sym_width, graph_line);
616
	printf(" %-*.*s", dso_width, dso_width, graph_line);
617
	puts("\n");
618

619
	for (nd = rb_first(&tmp); nd; nd = rb_next(nd)) {
620
		struct symbol *sym;
621
		double pcnt;
622

623
		syme = rb_entry(nd, struct sym_entry, rb_node);
624
		sym = sym_entry__symbol(syme);
625
		if (++printed > print_entries || (int)syme->snap_count < count_filter)
626
			continue;
627

628 629
		pcnt = 100.0 - (100.0 * ((sum_ksamples - syme->snap_count) /
					 sum_ksamples));
630

631
		if (nr_counters == 1 || !display_weighted)
632
			printf("%20.2f ", syme->weight);
633
		else
634
			printf("%9.1f %10ld ", syme->weight, syme->snap_count);
635

636
		percent_color_fprintf(stdout, "%4.1f%%", pcnt);
637
		if (verbose)
638
			printf(" %016llx", sym->start);
639
		printf(" %-*.*s", sym_width, sym_width, sym->name);
640 641 642 643
		printf(" %-*.*s\n", dso_width, dso_width,
		       dso_width >= syme->map->dso->long_name_len ?
					syme->map->dso->long_name :
					syme->map->dso->short_name);
644 645 646
	}
}

647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689
static void prompt_integer(int *target, const char *msg)
{
	char *buf = malloc(0), *p;
	size_t dummy = 0;
	int tmp;

	fprintf(stdout, "\n%s: ", msg);
	if (getline(&buf, &dummy, stdin) < 0)
		return;

	p = strchr(buf, '\n');
	if (p)
		*p = 0;

	p = buf;
	while(*p) {
		if (!isdigit(*p))
			goto out_free;
		p++;
	}
	tmp = strtoul(buf, NULL, 10);
	*target = tmp;
out_free:
	free(buf);
}

static void prompt_percent(int *target, const char *msg)
{
	int tmp = 0;

	prompt_integer(&tmp, msg);
	if (tmp >= 0 && tmp <= 100)
		*target = tmp;
}

static void prompt_symbol(struct sym_entry **target, const char *msg)
{
	char *buf = malloc(0), *p;
	struct sym_entry *syme = *target, *n, *found = NULL;
	size_t dummy = 0;

	/* zero counters of active symbol */
	if (syme) {
690
		pthread_mutex_lock(&syme->src->lock);
691 692
		__zero_source_counters(syme);
		*target = NULL;
693
		pthread_mutex_unlock(&syme->src->lock);
694 695 696 697 698 699 700 701 702 703 704 705 706 707 708
	}

	fprintf(stdout, "\n%s: ", msg);
	if (getline(&buf, &dummy, stdin) < 0)
		goto out_free;

	p = strchr(buf, '\n');
	if (p)
		*p = 0;

	pthread_mutex_lock(&active_symbols_lock);
	syme = list_entry(active_symbols.next, struct sym_entry, node);
	pthread_mutex_unlock(&active_symbols_lock);

	list_for_each_entry_safe_from(syme, n, &active_symbols, node) {
709
		struct symbol *sym = sym_entry__symbol(syme);
710 711 712 713 714 715 716 717

		if (!strcmp(buf, sym->name)) {
			found = syme;
			break;
		}
	}

	if (!found) {
718
		fprintf(stderr, "Sorry, %s is not active.\n", buf);
719 720 721 722 723 724 725 726 727
		sleep(1);
		return;
	} else
		parse_source(found);

out_free:
	free(buf);
}

728
static void print_mapped_keys(void)
729
{
730 731 732
	char *name = NULL;

	if (sym_filter_entry) {
733
		struct symbol *sym = sym_entry__symbol(sym_filter_entry);
734 735 736 737 738 739 740 741 742 743 744 745
		name = sym->name;
	}

	fprintf(stdout, "\nMapped keys:\n");
	fprintf(stdout, "\t[d]     display refresh delay.             \t(%d)\n", delay_secs);
	fprintf(stdout, "\t[e]     display entries (lines).           \t(%d)\n", print_entries);

	if (nr_counters > 1)
		fprintf(stdout, "\t[E]     active event counter.              \t(%s)\n", event_name(sym_counter));

	fprintf(stdout, "\t[f]     profile display filter (count).    \t(%d)\n", count_filter);

746 747 748
	fprintf(stdout, "\t[F]     annotate display filter (percent). \t(%d%%)\n", sym_pcnt_filter);
	fprintf(stdout, "\t[s]     annotate symbol.                   \t(%s)\n", name?: "NULL");
	fprintf(stdout, "\t[S]     stop annotation.\n");
749 750 751 752

	if (nr_counters > 1)
		fprintf(stdout, "\t[w]     toggle display weighted/count[E]r. \t(%d)\n", display_weighted ? 1 : 0);

753
	fprintf(stdout,
754
		"\t[K]     hide kernel_symbols symbols.     \t(%s)\n",
755 756 757 758
		hide_kernel_symbols ? "yes" : "no");
	fprintf(stdout,
		"\t[U]     hide user symbols.               \t(%s)\n",
		hide_user_symbols ? "yes" : "no");
759
	fprintf(stdout, "\t[z]     toggle sample zeroing.             \t(%d)\n", zero ? 1 : 0);
760 761 762 763 764 765 766 767 768 769 770 771
	fprintf(stdout, "\t[qQ]    quit.\n");
}

static int key_mapped(int c)
{
	switch (c) {
		case 'd':
		case 'e':
		case 'f':
		case 'z':
		case 'q':
		case 'Q':
772 773
		case 'K':
		case 'U':
774 775 776
		case 'F':
		case 's':
		case 'S':
777 778 779 780
			return 1;
		case 'E':
		case 'w':
			return nr_counters > 1 ? 1 : 0;
781 782
		default:
			break;
783 784 785
	}

	return 0;
786 787
}

788
static void handle_keypress(struct perf_session *session, int c)
789
{
790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812
	if (!key_mapped(c)) {
		struct pollfd stdin_poll = { .fd = 0, .events = POLLIN };
		struct termios tc, save;

		print_mapped_keys();
		fprintf(stdout, "\nEnter selection, or unmapped key to continue: ");
		fflush(stdout);

		tcgetattr(0, &save);
		tc = save;
		tc.c_lflag &= ~(ICANON | ECHO);
		tc.c_cc[VMIN] = 0;
		tc.c_cc[VTIME] = 0;
		tcsetattr(0, TCSANOW, &tc);

		poll(&stdin_poll, 1, -1);
		c = getc(stdin);

		tcsetattr(0, TCSAFLUSH, &save);
		if (!key_mapped(c))
			return;
	}

813 814 815
	switch (c) {
		case 'd':
			prompt_integer(&delay_secs, "Enter display delay");
816 817
			if (delay_secs < 1)
				delay_secs = 1;
818 819 820
			break;
		case 'e':
			prompt_integer(&print_entries, "Enter display entries (lines)");
821
			if (print_entries == 0) {
822
				sig_winch_handler(SIGWINCH);
823 824 825
				signal(SIGWINCH, sig_winch_handler);
			} else
				signal(SIGWINCH, SIG_DFL);
826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849
			break;
		case 'E':
			if (nr_counters > 1) {
				int i;

				fprintf(stderr, "\nAvailable events:");
				for (i = 0; i < nr_counters; i++)
					fprintf(stderr, "\n\t%d %s", i, event_name(i));

				prompt_integer(&sym_counter, "Enter details event counter");

				if (sym_counter >= nr_counters) {
					fprintf(stderr, "Sorry, no such event, using %s.\n", event_name(0));
					sym_counter = 0;
					sleep(1);
				}
			} else sym_counter = 0;
			break;
		case 'f':
			prompt_integer(&count_filter, "Enter display event count filter");
			break;
		case 'F':
			prompt_percent(&sym_pcnt_filter, "Enter details display event filter (percent)");
			break;
850 851 852
		case 'K':
			hide_kernel_symbols = !hide_kernel_symbols;
			break;
853 854 855
		case 'q':
		case 'Q':
			printf("exiting.\n");
856
			if (dump_symtab)
857
				perf_session__fprintf_dsos(session, stderr);
858 859 860 861 862 863 864 865 866 867
			exit(0);
		case 's':
			prompt_symbol(&sym_filter_entry, "Enter details symbol");
			break;
		case 'S':
			if (!sym_filter_entry)
				break;
			else {
				struct sym_entry *syme = sym_filter_entry;

868
				pthread_mutex_lock(&syme->src->lock);
869 870
				sym_filter_entry = NULL;
				__zero_source_counters(syme);
871
				pthread_mutex_unlock(&syme->src->lock);
872 873
			}
			break;
874 875 876
		case 'U':
			hide_user_symbols = !hide_user_symbols;
			break;
877 878 879
		case 'w':
			display_weighted = ~display_weighted;
			break;
880
		case 'z':
881
			zero = !zero;
882
			break;
883 884
		default:
			break;
885 886 887
	}
}

888
static void *display_thread(void *arg __used)
889
{
890
	struct pollfd stdin_poll = { .fd = 0, .events = POLLIN };
891 892
	struct termios tc, save;
	int delay_msecs, c;
893
	struct perf_session *session = (struct perf_session *) arg;
894 895 896 897 898 899

	tcgetattr(0, &save);
	tc = save;
	tc.c_lflag &= ~(ICANON | ECHO);
	tc.c_cc[VMIN] = 0;
	tc.c_cc[VTIME] = 0;
900

901 902 903 904 905
repeat:
	delay_msecs = delay_secs * 1000;
	tcsetattr(0, TCSANOW, &tc);
	/* trash return*/
	getc(stdin);
906

907
	do {
908
		print_sym_table();
909 910
	} while (!poll(&stdin_poll, 1, delay_msecs) == 1);

911 912 913
	c = getc(stdin);
	tcsetattr(0, TCSAFLUSH, &save);

914
	handle_keypress(session, c);
915
	goto repeat;
916 917 918 919

	return NULL;
}

920
/* Tag samples to be skipped. */
921
static const char *skip_symbols[] = {
922 923 924 925 926
	"default_idle",
	"cpu_idle",
	"enter_idle",
	"exit_idle",
	"mwait_idle",
927
	"mwait_idle_with_hints",
928
	"poll_idle",
929 930
	"ppc64_runlatch_off",
	"pseries_dedicated_idle_sleep",
931 932 933
	NULL
};

934
static int symbol_filter(struct map *map, struct symbol *sym)
935
{
936 937
	struct sym_entry *syme;
	const char *name = sym->name;
938
	int i;
939

940 941 942 943 944 945 946
	/*
	 * ppc64 uses function descriptors and appends a '.' to the
	 * start of every instruction address. Remove it.
	 */
	if (name[0] == '.')
		name++;

947 948 949 950 951 952 953
	if (!strcmp(name, "_text") ||
	    !strcmp(name, "_etext") ||
	    !strcmp(name, "_sinittext") ||
	    !strncmp("init_module", name, 11) ||
	    !strncmp("cleanup_module", name, 14) ||
	    strstr(name, "_text_start") ||
	    strstr(name, "_text_end"))
954 955
		return 1;

956
	syme = symbol__priv(sym);
957
	syme->map = map;
958
	syme->src = NULL;
959 960 961 962 963 964

	if (!sym_filter_entry && sym_filter && !strcmp(name, sym_filter)) {
		/* schedule initial sym_filter_entry setup */
		sym_filter_entry_sched = syme;
		sym_filter = NULL;
	}
965

966 967 968 969 970 971
	for (i = 0; skip_symbols[i]; i++) {
		if (!strcmp(skip_symbols[i], name)) {
			syme->skip = 1;
			break;
		}
	}
972

973 974 975
	if (!syme->skip)
		syme->name_len = strlen(sym->name);

976 977 978
	return 0;
}

979
static void event__process_sample(const event_t *self,
980 981
				  struct sample_data *sample,
				  struct perf_session *session, int counter)
982
{
983 984
	u64 ip = self->ip.ip;
	struct sym_entry *syme;
985
	struct addr_location al;
986
	struct machine *machine;
987
	u8 origin = self->header.misc & PERF_RECORD_MISC_CPUMODE_MASK;
988

989 990
	++samples;

991
	switch (origin) {
992
	case PERF_RECORD_MISC_USER:
993
		++us_samples;
994 995
		if (hide_user_symbols)
			return;
996
		machine = perf_session__find_host_machine(session);
997
		break;
998
	case PERF_RECORD_MISC_KERNEL:
999
		++kernel_samples;
1000 1001
		if (hide_kernel_symbols)
			return;
1002
		machine = perf_session__find_host_machine(session);
1003 1004 1005
		break;
	case PERF_RECORD_MISC_GUEST_KERNEL:
		++guest_kernel_samples;
1006
		machine = perf_session__find_machine(session, self->ip.pid);
1007
		break;
1008 1009 1010 1011 1012 1013 1014
	case PERF_RECORD_MISC_GUEST_USER:
		++guest_us_samples;
		/*
		 * TODO: we don't process guest user from host side
		 * except simple counting.
		 */
		return;
1015 1016 1017 1018
	default:
		return;
	}

1019
	if (!machine && perf_guest) {
1020 1021 1022 1023 1024
		pr_err("Can't find guest [%d]'s kernel information\n",
			self->ip.pid);
		return;
	}

P
Peter Zijlstra 已提交
1025
	if (self->header.misc & PERF_RECORD_MISC_EXACT_IP)
1026 1027
		exact_samples++;

1028
	if (event__preprocess_sample(self, session, &al, sample,
1029
				     symbol_filter) < 0 ||
1030
	    al.filtered)
1031
		return;
1032

1033 1034 1035 1036 1037 1038 1039 1040 1041 1042 1043 1044
	if (al.sym == NULL) {
		/*
		 * As we do lazy loading of symtabs we only will know if the
		 * specified vmlinux file is invalid when we actually have a
		 * hit in kernel space and then try to load it. So if we get
		 * here and there are _no_ symbols in the DSO backing the
		 * kernel map, bail out.
		 *
		 * We may never get here, for instance, if we use -K/
		 * --hide-kernel-symbols, even if the user specifies an
		 * invalid --vmlinux ;-)
		 */
1045
		if (al.map == machine->vmlinux_maps[MAP__FUNCTION] &&
1046 1047 1048 1049 1050 1051 1052 1053 1054
		    RB_EMPTY_ROOT(&al.map->dso->symbols[MAP__FUNCTION])) {
			pr_err("The %s file can't be used\n",
			       symbol_conf.vmlinux_name);
			exit(1);
		}

		return;
	}

1055 1056 1057 1058
	/* let's see, whether we need to install initial sym_filter_entry */
	if (sym_filter_entry_sched) {
		sym_filter_entry = sym_filter_entry_sched;
		sym_filter_entry_sched = NULL;
1059 1060 1061 1062 1063 1064
		if (parse_source(sym_filter_entry) < 0) {
			struct symbol *sym = sym_entry__symbol(sym_filter_entry);

			pr_err("Can't annotate %s", sym->name);
			if (sym_filter_entry->map->dso->origin == DSO__ORIG_KERNEL) {
				pr_err(": No vmlinux file was found in the path:\n");
1065
				machine__fprintf_vmlinux_path(machine, stderr);
1066 1067 1068 1069
			} else
				pr_err(".\n");
			exit(1);
		}
1070 1071
	}

1072
	syme = symbol__priv(al.sym);
1073 1074
	if (!syme->skip) {
		syme->count[counter]++;
1075
		syme->origin = origin;
1076 1077 1078 1079 1080 1081
		record_precise_ip(syme, counter, ip);
		pthread_mutex_lock(&active_symbols_lock);
		if (list_empty(&syme->node) || !syme->node.next)
			__list_insert_active_sym(syme);
		pthread_mutex_unlock(&active_symbols_lock);
	}
1082 1083 1084
}

struct mmap_data {
1085 1086
	int			counter;
	void			*base;
1087
	int			mask;
1088
	unsigned int		prev;
1089 1090 1091 1092
};

static unsigned int mmap_read_head(struct mmap_data *md)
{
1093
	struct perf_event_mmap_page *pc = md->base;
1094 1095 1096 1097 1098 1099 1100 1101
	int head;

	head = pc->data_head;
	rmb();

	return head;
}

1102 1103
static void perf_session__mmap_read_counter(struct perf_session *self,
					    struct mmap_data *md)
1104 1105 1106 1107
{
	unsigned int head = mmap_read_head(md);
	unsigned int old = md->prev;
	unsigned char *data = md->base + page_size;
1108
	struct sample_data sample;
1109 1110 1111 1112
	int diff;

	/*
	 * If we're further behind than half the buffer, there's a chance
1113
	 * the writer will bite our tail and mess up the samples under us.
1114 1115 1116 1117 1118 1119 1120
	 *
	 * If we somehow ended up ahead of the head, we got messed up.
	 *
	 * In either case, truncate and restart at head.
	 */
	diff = head - old;
	if (diff > md->mask / 2 || diff < 0) {
1121
		fprintf(stderr, "WARNING: failed to keep up with mmap data.\n");
1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133

		/*
		 * head points to a known good entry, start there.
		 */
		old = head;
	}

	for (; old != head;) {
		event_t *event = (event_t *)&data[old & md->mask];

		event_t event_copy;

1134
		size_t size = event->header.size;
1135 1136 1137 1138 1139 1140 1141 1142 1143 1144 1145 1146 1147 1148 1149 1150 1151 1152 1153 1154 1155

		/*
		 * Event straddles the mmap boundary -- header should always
		 * be inside due to u64 alignment of output.
		 */
		if ((old & md->mask) + size != ((old + size) & md->mask)) {
			unsigned int offset = old;
			unsigned int len = min(sizeof(*event), size), cpy;
			void *dst = &event_copy;

			do {
				cpy = min(md->mask + 1 - (offset & md->mask), len);
				memcpy(dst, &data[offset & md->mask], cpy);
				offset += cpy;
				dst += cpy;
				len -= cpy;
			} while (len);

			event = &event_copy;
		}

1156
		event__parse_sample(event, self, &sample);
1157
		if (event->header.type == PERF_RECORD_SAMPLE)
1158
			event__process_sample(event, &sample, self, md->counter);
1159
		else
1160
			event__process(event, &sample, self);
1161 1162 1163 1164 1165 1166
		old += size;
	}

	md->prev = old;
}

1167 1168
static struct pollfd *event_array;
static struct mmap_data *mmap_array[MAX_NR_CPUS][MAX_COUNTERS];
M
Mike Galbraith 已提交
1169

1170
static void perf_session__mmap_read(struct perf_session *self)
1171
{
1172
	int i, counter, thread_index;
1173 1174 1175

	for (i = 0; i < nr_cpus; i++) {
		for (counter = 0; counter < nr_counters; counter++)
1176 1177 1178 1179 1180 1181
			for (thread_index = 0;
				thread_index < thread_num;
				thread_index++) {
				perf_session__mmap_read_counter(self,
					&mmap_array[i][counter][thread_index]);
			}
1182 1183 1184
	}
}

1185 1186 1187 1188
int nr_poll;
int group_fd;

static void start_counter(int i, int counter)
1189
{
1190
	struct perf_event_attr *attr;
1191
	int cpu = -1;
1192
	int thread_index;
1193

1194
	if (target_tid == -1)
1195
		cpu = cpumap[i];
1196 1197 1198 1199

	attr = attrs + counter;

	attr->sample_type	= PERF_SAMPLE_IP | PERF_SAMPLE_TID;
1200 1201 1202 1203 1204 1205 1206

	if (freq) {
		attr->sample_type	|= PERF_SAMPLE_PERIOD;
		attr->freq		= 1;
		attr->sample_freq	= freq;
	}

1207
	attr->inherit		= (cpu < 0) && inherit;
1208
	attr->mmap		= 1;
1209

1210
	for (thread_index = 0; thread_index < thread_num; thread_index++) {
1211
try_again:
1212 1213 1214 1215 1216 1217 1218
		fd[i][counter][thread_index] = sys_perf_event_open(attr,
				all_tids[thread_index], cpu, group_fd, 0);

		if (fd[i][counter][thread_index] < 0) {
			int err = errno;

			if (err == EPERM || err == EACCES)
1219 1220 1221
				die("Permission error - are you root?\n"
					"\t Consider tweaking"
					" /proc/sys/kernel/perf_event_paranoid.\n");
1222 1223 1224 1225 1226 1227 1228 1229 1230 1231 1232 1233 1234 1235 1236 1237
			/*
			 * If it's cycles then fall back to hrtimer
			 * based cpu-clock-tick sw counter, which
			 * is always available even if no PMU support:
			 */
			if (attr->type == PERF_TYPE_HARDWARE
					&& attr->config == PERF_COUNT_HW_CPU_CYCLES) {

				if (verbose)
					warning(" ... trying to fall back to cpu-clock-ticks\n");

				attr->type = PERF_TYPE_SOFTWARE;
				attr->config = PERF_COUNT_SW_CPU_CLOCK;
				goto try_again;
			}
			printf("\n");
1238
			error("sys_perf_event_open() syscall returned with %d (%s).  /bin/dmesg may provide additional information.\n",
1239 1240 1241 1242 1243 1244
					fd[i][counter][thread_index], strerror(err));
			die("No CONFIG_PERF_EVENTS=y kernel support configured?\n");
			exit(-1);
		}
		assert(fd[i][counter][thread_index] >= 0);
		fcntl(fd[i][counter][thread_index], F_SETFL, O_NONBLOCK);
1245 1246

		/*
1247
		 * First counter acts as the group leader:
1248
		 */
1249 1250 1251 1252 1253 1254 1255 1256 1257 1258 1259 1260 1261 1262
		if (group && group_fd == -1)
			group_fd = fd[i][counter][thread_index];

		event_array[nr_poll].fd = fd[i][counter][thread_index];
		event_array[nr_poll].events = POLLIN;
		nr_poll++;

		mmap_array[i][counter][thread_index].counter = counter;
		mmap_array[i][counter][thread_index].prev = 0;
		mmap_array[i][counter][thread_index].mask = mmap_pages*page_size - 1;
		mmap_array[i][counter][thread_index].base = mmap(NULL, (mmap_pages+1)*page_size,
				PROT_READ, MAP_SHARED, fd[i][counter][thread_index], 0);
		if (mmap_array[i][counter][thread_index].base == MAP_FAILED)
			die("failed to mmap with %d (%s)\n", errno, strerror(errno));
1263 1264 1265 1266 1267 1268 1269
	}
}

static int __cmd_top(void)
{
	pthread_t thread;
	int i, counter;
1270
	int ret;
1271
	/*
1272 1273
	 * FIXME: perf_session__new should allow passing a O_MMAP, so that all this
	 * mmap reading, etc is encapsulated in it. Use O_WRONLY for now.
1274
	 */
T
Tom Zanussi 已提交
1275
	struct perf_session *session = perf_session__new(NULL, O_WRONLY, false, false);
1276 1277
	if (session == NULL)
		return -ENOMEM;
1278

1279 1280
	if (target_tid != -1)
		event__synthesize_thread(target_tid, event__process, session);
1281
	else
1282
		event__synthesize_threads(event__process, session);
1283

1284 1285
	for (i = 0; i < nr_cpus; i++) {
		group_fd = -1;
1286 1287
		for (counter = 0; counter < nr_counters; counter++)
			start_counter(i, counter);
1288 1289
	}

1290
	/* Wait for a minimal set of events before starting the snapshot */
1291
	poll(&event_array[0], nr_poll, 100);
1292

1293
	perf_session__mmap_read(session);
1294

1295
	if (pthread_create(&thread, NULL, display_thread, session)) {
1296 1297 1298 1299 1300 1301 1302 1303 1304 1305 1306 1307 1308 1309 1310
		printf("Could not create display thread.\n");
		exit(-1);
	}

	if (realtime_prio) {
		struct sched_param param;

		param.sched_priority = realtime_prio;
		if (sched_setscheduler(0, SCHED_FIFO, &param)) {
			printf("Could not set realtime priority.\n");
			exit(-1);
		}
	}

	while (1) {
1311
		int hits = samples;
1312

1313
		perf_session__mmap_read(session);
1314

1315
		if (hits == samples)
1316 1317 1318 1319 1320
			ret = poll(event_array, nr_poll, 100);
	}

	return 0;
}
1321 1322 1323 1324 1325 1326 1327 1328

static const char * const top_usage[] = {
	"perf top [<options>]",
	NULL
};

static const struct option options[] = {
	OPT_CALLBACK('e', "event", NULL, "event",
1329 1330
		     "event selector. use 'perf list' to list available events",
		     parse_events),
1331 1332 1333
	OPT_INTEGER('c', "count", &default_interval,
		    "event period to sample"),
	OPT_INTEGER('p', "pid", &target_pid,
1334 1335 1336
		    "profile events on existing process id"),
	OPT_INTEGER('t', "tid", &target_tid,
		    "profile events on existing thread id"),
1337 1338
	OPT_BOOLEAN('a', "all-cpus", &system_wide,
			    "system-wide collection from all CPUs"),
1339 1340
	OPT_STRING('C', "cpu", &cpu_list, "cpu",
		    "list of cpus to monitor"),
1341 1342
	OPT_STRING('k', "vmlinux", &symbol_conf.vmlinux_name,
		   "file", "vmlinux pathname"),
1343 1344
	OPT_BOOLEAN('K', "hide_kernel_symbols", &hide_kernel_symbols,
		    "hide kernel symbols"),
1345
	OPT_UINTEGER('m', "mmap-pages", &mmap_pages, "number of mmap data pages"),
1346 1347
	OPT_INTEGER('r', "realtime", &realtime_prio,
		    "collect data with this RT SCHED_FIFO priority"),
M
Mike Galbraith 已提交
1348
	OPT_INTEGER('d', "delay", &delay_secs,
1349 1350 1351
		    "number of seconds to delay between refreshes"),
	OPT_BOOLEAN('D', "dump-symtab", &dump_symtab,
			    "dump the symbol table used for profiling"),
1352
	OPT_INTEGER('f', "count-filter", &count_filter,
1353 1354 1355
		    "only display functions with more events than this"),
	OPT_BOOLEAN('g', "group", &group,
			    "put the counters into a counter group"),
1356 1357
	OPT_BOOLEAN('i', "inherit", &inherit,
		    "child tasks inherit counters"),
1358
	OPT_STRING('s', "sym-annotate", &sym_filter, "symbol name",
1359
		    "symbol to annotate"),
A
Anton Blanchard 已提交
1360
	OPT_BOOLEAN('z', "zero", &zero,
1361
		    "zero history across updates"),
1362
	OPT_INTEGER('F', "freq", &freq,
1363
		    "profile at this frequency"),
1364 1365
	OPT_INTEGER('E', "entries", &print_entries,
		    "display this many functions"),
1366 1367
	OPT_BOOLEAN('U', "hide_user_symbols", &hide_user_symbols,
		    "hide user symbols"),
1368
	OPT_INCR('v', "verbose", &verbose,
1369
		    "be more verbose (show counter open errors, etc)"),
1370 1371 1372
	OPT_END()
};

1373
int cmd_top(int argc, const char **argv, const char *prefix __used)
1374
{
1375
	int counter;
1376
	int i,j;
1377 1378 1379 1380 1381 1382 1383

	page_size = sysconf(_SC_PAGE_SIZE);

	argc = parse_options(argc, argv, options, top_usage, 0);
	if (argc)
		usage_with_options(top_usage, options);

1384 1385 1386 1387 1388 1389 1390 1391 1392 1393 1394 1395 1396 1397 1398 1399 1400 1401 1402 1403
	if (target_pid != -1) {
		target_tid = target_pid;
		thread_num = find_all_tid(target_pid, &all_tids);
		if (thread_num <= 0) {
			fprintf(stderr, "Can't find all threads of pid %d\n",
				target_pid);
			usage_with_options(top_usage, options);
		}
	} else {
		all_tids=malloc(sizeof(pid_t));
		if (!all_tids)
			return -ENOMEM;

		all_tids[0] = target_tid;
		thread_num = 1;
	}

	for (i = 0; i < MAX_NR_CPUS; i++) {
		for (j = 0; j < MAX_COUNTERS; j++) {
			fd[i][j] = malloc(sizeof(int)*thread_num);
1404
			mmap_array[i][j] = zalloc(
1405 1406 1407 1408 1409 1410 1411 1412 1413 1414
				sizeof(struct mmap_data)*thread_num);
			if (!fd[i][j] || !mmap_array[i][j])
				return -ENOMEM;
		}
	}
	event_array = malloc(
		sizeof(struct pollfd)*MAX_NR_CPUS*MAX_COUNTERS*thread_num);
	if (!event_array)
		return -ENOMEM;

1415
	/* CPU and PID are mutually exclusive */
1416
	if (target_tid > 0 && cpu_list) {
1417 1418
		printf("WARNING: PID switch overriding CPU\n");
		sleep(1);
1419
		cpu_list = NULL;
1420 1421
	}

1422
	if (!nr_counters)
1423 1424
		nr_counters = 1;

1425 1426
	symbol_conf.priv_size = (sizeof(struct sym_entry) +
				 (nr_counters + 1) * sizeof(unsigned long));
1427 1428

	symbol_conf.try_vmlinux_path = (symbol_conf.vmlinux_name == NULL);
1429
	if (symbol__init() < 0)
1430
		return -1;
1431

1432 1433 1434
	if (delay_secs < 1)
		delay_secs = 1;

1435 1436 1437 1438 1439 1440 1441 1442 1443 1444 1445 1446
	/*
	 * User specified count overrides default frequency.
	 */
	if (default_interval)
		freq = 0;
	else if (freq) {
		default_interval = freq;
	} else {
		fprintf(stderr, "frequency and count are zero, aborting\n");
		exit(EXIT_FAILURE);
	}

1447 1448 1449
	/*
	 * Fill in the ones not specifically initialized via -c:
	 */
1450
	for (counter = 0; counter < nr_counters; counter++) {
1451
		if (attrs[counter].sample_period)
1452 1453
			continue;

1454
		attrs[counter].sample_period = default_interval;
1455 1456
	}

1457
	if (target_tid != -1)
1458
		nr_cpus = 1;
1459
	else
1460 1461 1462 1463
		nr_cpus = read_cpu_map(cpu_list);

	if (nr_cpus < 1)
		usage_with_options(top_usage, options);
1464

1465
	get_term_dimensions(&winsize);
1466
	if (print_entries == 0) {
1467
		update_print_entries(&winsize);
1468 1469 1470
		signal(SIGWINCH, sig_winch_handler);
	}

1471 1472
	return __cmd_top();
}