event.h 19.9 KB
Newer Older
1
/* SPDX-License-Identifier: GPL-2.0 */
2 3
#ifndef __PERF_RECORD_H
#define __PERF_RECORD_H
4

5
#include <limits.h>
6
#include <stdio.h>
7
#include <linux/kernel.h>
8
#include <linux/bpf.h>
9

10
#include "../perf.h"
11
#include "build-id.h"
12
#include "perf_regs.h"
13 14 15 16 17 18 19 20 21 22

struct mmap_event {
	struct perf_event_header header;
	u32 pid, tid;
	u64 start;
	u64 len;
	u64 pgoff;
	char filename[PATH_MAX];
};

23 24 25 26 27 28 29 30 31 32
struct mmap2_event {
	struct perf_event_header header;
	u32 pid, tid;
	u64 start;
	u64 len;
	u64 pgoff;
	u32 maj;
	u32 min;
	u64 ino;
	u64 ino_generation;
33 34
	u32 prot;
	u32 flags;
35 36 37
	char filename[PATH_MAX];
};

38 39 40 41 42 43
struct comm_event {
	struct perf_event_header header;
	u32 pid, tid;
	char comm[16];
};

44 45 46 47 48 49 50
struct namespaces_event {
	struct perf_event_header header;
	u32 pid, tid;
	u64 nr_namespaces;
	struct perf_ns_link_info link_info[];
};

51 52 53 54
struct fork_event {
	struct perf_event_header header;
	u32 pid, ppid;
	u32 tid, ptid;
55
	u64 time;
56 57 58 59 60 61 62 63
};

struct lost_event {
	struct perf_event_header header;
	u64 id;
	u64 lost;
};

64 65 66 67 68
struct lost_samples_event {
	struct perf_event_header header;
	u64 lost;
};

69 70 71
/*
 * PERF_FORMAT_ENABLED | PERF_FORMAT_RUNNING | PERF_FORMAT_ID
 */
72 73
struct read_event {
	struct perf_event_header header;
74
	u32 pid, tid;
75 76 77 78 79 80
	u64 value;
	u64 time_enabled;
	u64 time_running;
	u64 id;
};

81 82 83 84 85 86
struct throttle_event {
	struct perf_event_header header;
	u64 time;
	u64 id;
	u64 stream_id;
};
87

88 89 90 91 92 93 94 95 96 97 98 99 100
#ifndef KSYM_NAME_LEN
#define KSYM_NAME_LEN 256
#endif

struct ksymbol_event {
	struct perf_event_header header;
	u64 addr;
	u32 len;
	u16 ksym_type;
	u16 flags;
	char name[KSYM_NAME_LEN];
};

101 102 103 104 105 106 107 108 109 110
struct bpf_event {
	struct perf_event_header header;
	u16 type;
	u16 flags;
	u32 id;

	/* for bpf_prog types */
	u8 tag[BPF_TAG_SIZE];  // prog tag
};

111 112 113 114
#define PERF_SAMPLE_MASK				\
	(PERF_SAMPLE_IP | PERF_SAMPLE_TID |		\
	 PERF_SAMPLE_TIME | PERF_SAMPLE_ADDR |		\
	PERF_SAMPLE_ID | PERF_SAMPLE_STREAM_ID |	\
115 116
	 PERF_SAMPLE_CPU | PERF_SAMPLE_PERIOD |		\
	 PERF_SAMPLE_IDENTIFIER)
117

118 119 120
/* perf sample has 16 bits size limit */
#define PERF_SAMPLE_MAX_SIZE (1 << 16)

121
struct sample_event {
122 123 124 125
	struct perf_event_header        header;
	u64 array[];
};

126
struct regs_dump {
127
	u64 abi;
128
	u64 mask;
129
	u64 *regs;
130 131 132 133

	/* Cached values/mask filled by first register access. */
	u64 cache_regs[PERF_REGS_MAX];
	u64 cache_mask;
134 135 136 137 138 139 140 141
};

struct stack_dump {
	u16 offset;
	u64 size;
	char *data;
};

142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158
struct sample_read_value {
	u64 value;
	u64 id;
};

struct sample_read {
	u64 time_enabled;
	u64 time_running;
	union {
		struct {
			u64 nr;
			struct sample_read_value *values;
		} group;
		struct sample_read_value one;
	};
};

159 160 161 162 163 164 165 166 167 168
struct ip_callchain {
	u64 nr;
	u64 ips[0];
};

struct branch_flags {
	u64 mispred:1;
	u64 predicted:1;
	u64 in_tx:1;
	u64 abort:1;
169
	u64 cycles:16;
170 171
	u64 type:4;
	u64 reserved:40;
172 173 174 175 176 177 178 179 180 181 182 183 184
};

struct branch_entry {
	u64			from;
	u64			to;
	struct branch_flags	flags;
};

struct branch_stack {
	u64			nr;
	struct branch_entry	entries[0];
};

185 186 187 188 189 190 191 192 193 194 195 196 197 198
enum {
	PERF_IP_FLAG_BRANCH		= 1ULL << 0,
	PERF_IP_FLAG_CALL		= 1ULL << 1,
	PERF_IP_FLAG_RETURN		= 1ULL << 2,
	PERF_IP_FLAG_CONDITIONAL	= 1ULL << 3,
	PERF_IP_FLAG_SYSCALLRET		= 1ULL << 4,
	PERF_IP_FLAG_ASYNC		= 1ULL << 5,
	PERF_IP_FLAG_INTERRUPT		= 1ULL << 6,
	PERF_IP_FLAG_TX_ABORT		= 1ULL << 7,
	PERF_IP_FLAG_TRACE_BEGIN	= 1ULL << 8,
	PERF_IP_FLAG_TRACE_END		= 1ULL << 9,
	PERF_IP_FLAG_IN_TX		= 1ULL << 10,
};

199 200
#define PERF_IP_FLAG_CHARS "bcrosyiABEx"

201 202 203 204 205 206 207 208 209 210 211 212
#define PERF_BRANCH_MASK		(\
	PERF_IP_FLAG_BRANCH		|\
	PERF_IP_FLAG_CALL		|\
	PERF_IP_FLAG_RETURN		|\
	PERF_IP_FLAG_CONDITIONAL	|\
	PERF_IP_FLAG_SYSCALLRET		|\
	PERF_IP_FLAG_ASYNC		|\
	PERF_IP_FLAG_INTERRUPT		|\
	PERF_IP_FLAG_TX_ABORT		|\
	PERF_IP_FLAG_TRACE_BEGIN	|\
	PERF_IP_FLAG_TRACE_END)

213 214
#define MAX_INSN 16

215
struct perf_sample {
216 217 218 219 220 221 222
	u64 ip;
	u32 pid, tid;
	u64 time;
	u64 addr;
	u64 id;
	u64 stream_id;
	u64 period;
223
	u64 weight;
224
	u64 transaction;
225
	u32 cpu;
226
	u32 raw_size;
227
	u64 data_src;
228
	u64 phys_addr;
229 230
	u32 flags;
	u16 insn_len;
231
	u8  cpumode;
232
	u16 misc;
233
	char insn[MAX_INSN];
234
	void *raw_data;
235
	struct ip_callchain *callchain;
236
	struct branch_stack *branch_stack;
237
	struct regs_dump  user_regs;
238
	struct regs_dump  intr_regs;
239
	struct stack_dump user_stack;
240
	struct sample_read read;
241 242
};

243 244 245 246 247 248 249
#define PERF_MEM_DATA_SRC_NONE \
	(PERF_MEM_S(OP, NA) |\
	 PERF_MEM_S(LVL, NA) |\
	 PERF_MEM_S(SNOOP, NA) |\
	 PERF_MEM_S(LOCK, NA) |\
	 PERF_MEM_S(TLB, NA))

250 251
struct build_id_event {
	struct perf_event_header header;
252
	pid_t			 pid;
253
	u8			 build_id[PERF_ALIGN(BUILD_ID_SIZE, sizeof(u64))];
254 255
	char			 filename[];
};
256

257
enum perf_user_event_type { /* above any possible kernel type */
258
	PERF_RECORD_USER_TYPE_START		= 64,
259
	PERF_RECORD_HEADER_ATTR			= 64,
K
Kim Phillips 已提交
260
	PERF_RECORD_HEADER_EVENT_TYPE		= 65, /* deprecated */
261
	PERF_RECORD_HEADER_TRACING_DATA		= 66,
262
	PERF_RECORD_HEADER_BUILD_ID		= 67,
263
	PERF_RECORD_FINISHED_ROUND		= 68,
A
Adrian Hunter 已提交
264
	PERF_RECORD_ID_INDEX			= 69,
265 266
	PERF_RECORD_AUXTRACE_INFO		= 70,
	PERF_RECORD_AUXTRACE			= 71,
267
	PERF_RECORD_AUXTRACE_ERROR		= 72,
268
	PERF_RECORD_THREAD_MAP			= 73,
269
	PERF_RECORD_CPU_MAP			= 74,
270
	PERF_RECORD_STAT_CONFIG			= 75,
J
Jiri Olsa 已提交
271
	PERF_RECORD_STAT			= 76,
272
	PERF_RECORD_STAT_ROUND			= 77,
273
	PERF_RECORD_EVENT_UPDATE		= 78,
274
	PERF_RECORD_TIME_CONV			= 79,
275
	PERF_RECORD_HEADER_FEATURE		= 80,
276 277 278
	PERF_RECORD_HEADER_MAX
};

279 280 281 282 283
enum auxtrace_error_type {
	PERF_AUXTRACE_ERROR_ITRACE  = 1,
	PERF_AUXTRACE_ERROR_MAX
};

284 285 286
/* Attribute type for custom synthesized events */
#define PERF_TYPE_SYNTH		(INT_MAX + 1U)

287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404
/* Attribute config for custom synthesized events */
enum perf_synth_id {
	PERF_SYNTH_INTEL_PTWRITE,
	PERF_SYNTH_INTEL_MWAIT,
	PERF_SYNTH_INTEL_PWRE,
	PERF_SYNTH_INTEL_EXSTOP,
	PERF_SYNTH_INTEL_PWRX,
	PERF_SYNTH_INTEL_CBR,
};

/*
 * Raw data formats for synthesized events. Note that 4 bytes of padding are
 * present to match the 'size' member of PERF_SAMPLE_RAW data which is always
 * 8-byte aligned. That means we must dereference raw_data with an offset of 4.
 * Refer perf_sample__synth_ptr() and perf_synth__raw_data().  It also means the
 * structure sizes are 4 bytes bigger than the raw_size, refer
 * perf_synth__raw_size().
 */

struct perf_synth_intel_ptwrite {
	u32 padding;
	union {
		struct {
			u32	ip		:  1,
				reserved	: 31;
		};
		u32	flags;
	};
	u64	payload;
};

struct perf_synth_intel_mwait {
	u32 padding;
	u32 reserved;
	union {
		struct {
			u64	hints		:  8,
				reserved1	: 24,
				extensions	:  2,
				reserved2	: 30;
		};
		u64	payload;
	};
};

struct perf_synth_intel_pwre {
	u32 padding;
	u32 reserved;
	union {
		struct {
			u64	reserved1	:  7,
				hw		:  1,
				subcstate	:  4,
				cstate		:  4,
				reserved2	: 48;
		};
		u64	payload;
	};
};

struct perf_synth_intel_exstop {
	u32 padding;
	union {
		struct {
			u32	ip		:  1,
				reserved	: 31;
		};
		u32	flags;
	};
};

struct perf_synth_intel_pwrx {
	u32 padding;
	u32 reserved;
	union {
		struct {
			u64	deepest_cstate	:  4,
				last_cstate	:  4,
				wake_reason	:  4,
				reserved1	: 52;
		};
		u64	payload;
	};
};

struct perf_synth_intel_cbr {
	u32 padding;
	union {
		struct {
			u32	cbr		:  8,
				reserved1	:  8,
				max_nonturbo	:  8,
				reserved2	:  8;
		};
		u32	flags;
	};
	u32 freq;
	u32 reserved3;
};

/*
 * raw_data is always 4 bytes from an 8-byte boundary, so subtract 4 to get
 * 8-byte alignment.
 */
static inline void *perf_sample__synth_ptr(struct perf_sample *sample)
{
	return sample->raw_data - 4;
}

static inline void *perf_synth__raw_data(void *p)
{
	return p + 4;
}

#define perf_synth__raw_size(d) (sizeof(d) - 4)

#define perf_sample__bad_synth_size(s, d) ((s)->raw_size < sizeof(d) - 4)

405 406 407 408 409 410 411
/*
 * The kernel collects the number of events it couldn't send in a stretch and
 * when possible sends this number in a PERF_RECORD_LOST event. The number of
 * such "chunks" of lost events is stored in .nr_events[PERF_EVENT_LOST] while
 * total_lost tells exactly how many events the kernel in fact lost, i.e. it is
 * the sum of all struct lost_event.lost fields reported.
 *
412 413 414 415 416 417
 * The kernel discards mixed up samples and sends the number in a
 * PERF_RECORD_LOST_SAMPLES event. The number of lost-samples events is stored
 * in .nr_events[PERF_RECORD_LOST_SAMPLES] while total_lost_samples tells
 * exactly how many samples the kernel in fact dropped, i.e. it is the sum of
 * all struct lost_samples_event.lost fields reported.
 *
418 419 420 421 422 423 424 425 426
 * The total_period is needed because by default auto-freq is used, so
 * multipling nr_events[PERF_EVENT_SAMPLE] by a frequency isn't possible to get
 * the total number of low level events, it is necessary to to sum all struct
 * sample_event.period and stash the result in total_period.
 */
struct events_stats {
	u64 total_period;
	u64 total_non_filtered_period;
	u64 total_lost;
427
	u64 total_lost_samples;
428
	u64 total_aux_lost;
429
	u64 total_aux_partial;
430 431 432 433 434 435 436 437
	u64 total_invalid_chains;
	u32 nr_events[PERF_RECORD_HEADER_MAX];
	u32 nr_non_filtered_samples;
	u32 nr_lost_warned;
	u32 nr_unknown_events;
	u32 nr_invalid_chains;
	u32 nr_unknown_id;
	u32 nr_unprocessable_samples;
438
	u32 nr_auxtrace_errors[PERF_AUXTRACE_ERROR_MAX];
439
	u32 nr_proc_map_timeout;
440 441
};

442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467
enum {
	PERF_CPU_MAP__CPUS = 0,
	PERF_CPU_MAP__MASK = 1,
};

struct cpu_map_entries {
	u16	nr;
	u16	cpu[];
};

struct cpu_map_mask {
	u16	nr;
	u16	long_size;
	unsigned long mask[];
};

struct cpu_map_data {
	u16	type;
	char	data[];
};

struct cpu_map_event {
	struct perf_event_header	header;
	struct cpu_map_data		data;
};

468 469 470 471
struct attr_event {
	struct perf_event_header header;
	struct perf_event_attr attr;
	u64 id[];
472 473
};

474 475
enum {
	PERF_EVENT_UPDATE__UNIT  = 0,
476
	PERF_EVENT_UPDATE__SCALE = 1,
477
	PERF_EVENT_UPDATE__NAME  = 2,
478 479 480 481 482
	PERF_EVENT_UPDATE__CPUS  = 3,
};

struct event_update_event_cpus {
	struct cpu_map_data cpus;
483 484 485 486
};

struct event_update_event_scale {
	double scale;
487 488
};

489 490 491 492 493 494 495 496
struct event_update_event {
	struct perf_event_header header;
	u64 type;
	u64 id;

	char data[];
};

497 498 499 500 501 502 503 504 505 506 507 508
#define MAX_EVENT_NAME 64

struct perf_trace_event_type {
	u64	event_id;
	char	name[MAX_EVENT_NAME];
};

struct event_type_event {
	struct perf_event_header header;
	struct perf_trace_event_type event_type;
};

509 510 511 512 513
struct tracing_data_event {
	struct perf_event_header header;
	u32 size;
};

A
Adrian Hunter 已提交
514 515 516 517 518 519 520 521 522 523 524 525 526
struct id_index_entry {
	u64 id;
	u64 idx;
	u64 cpu;
	u64 tid;
};

struct id_index_event {
	struct perf_event_header header;
	u64 nr;
	struct id_index_entry entries[0];
};

527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544
struct auxtrace_info_event {
	struct perf_event_header header;
	u32 type;
	u32 reserved__; /* For alignment */
	u64 priv[];
};

struct auxtrace_event {
	struct perf_event_header header;
	u64 size;
	u64 offset;
	u64 reference;
	u32 idx;
	u32 tid;
	u32 cpu;
	u32 reserved__; /* For alignment */
};

545 546 547 548 549 550 551 552 553 554 555 556 557 558
#define MAX_AUXTRACE_ERROR_MSG 64

struct auxtrace_error_event {
	struct perf_event_header header;
	u32 type;
	u32 code;
	u32 cpu;
	u32 pid;
	u32 tid;
	u32 reserved__; /* For alignment */
	u64 ip;
	char msg[MAX_AUXTRACE_ERROR_MSG];
};

559 560 561 562 563 564 565
struct aux_event {
	struct perf_event_header header;
	u64	aux_offset;
	u64	aux_size;
	u64	flags;
};

566 567 568 569 570
struct itrace_start_event {
	struct perf_event_header header;
	u32 pid, tid;
};

571 572 573 574 575 576
struct context_switch_event {
	struct perf_event_header header;
	u32 next_prev_pid;
	u32 next_prev_tid;
};

577 578 579 580 581 582 583 584 585 586 587
struct thread_map_event_entry {
	u64	pid;
	char	comm[16];
};

struct thread_map_event {
	struct perf_event_header	header;
	u64				nr;
	struct thread_map_event_entry	entries[];
};

588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605
enum {
	PERF_STAT_CONFIG_TERM__AGGR_MODE	= 0,
	PERF_STAT_CONFIG_TERM__INTERVAL		= 1,
	PERF_STAT_CONFIG_TERM__SCALE		= 2,
	PERF_STAT_CONFIG_TERM__MAX		= 3,
};

struct stat_config_event_entry {
	u64	tag;
	u64	val;
};

struct stat_config_event {
	struct perf_event_header	header;
	u64				nr;
	struct stat_config_event_entry	data[];
};

J
Jiri Olsa 已提交
606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622
struct stat_event {
	struct perf_event_header	header;

	u64	id;
	u32	cpu;
	u32	thread;

	union {
		struct {
			u64 val;
			u64 ena;
			u64 run;
		};
		u64 values[3];
	};
};

623 624 625 626 627 628 629 630 631 632 633
enum {
	PERF_STAT_ROUND_TYPE__INTERVAL	= 0,
	PERF_STAT_ROUND_TYPE__FINAL	= 1,
};

struct stat_round_event {
	struct perf_event_header	header;
	u64				type;
	u64				time;
};

634 635 636 637 638 639 640
struct time_conv_event {
	struct perf_event_header header;
	u64 time_shift;
	u64 time_mult;
	u64 time_zero;
};

641 642 643 644 645 646
struct feature_event {
	struct perf_event_header 	header;
	u64				feat_id;
	char				data[];
};

647
union perf_event {
648 649
	struct perf_event_header	header;
	struct mmap_event		mmap;
650
	struct mmap2_event		mmap2;
651
	struct comm_event		comm;
652
	struct namespaces_event		namespaces;
653 654
	struct fork_event		fork;
	struct lost_event		lost;
655
	struct lost_samples_event	lost_samples;
656
	struct read_event		read;
657
	struct throttle_event		throttle;
658
	struct sample_event		sample;
659
	struct attr_event		attr;
660
	struct event_update_event	event_update;
661
	struct event_type_event		event_type;
662
	struct tracing_data_event	tracing_data;
663
	struct build_id_event		build_id;
A
Adrian Hunter 已提交
664
	struct id_index_event		id_index;
665 666
	struct auxtrace_info_event	auxtrace_info;
	struct auxtrace_event		auxtrace;
667
	struct auxtrace_error_event	auxtrace_error;
668
	struct aux_event		aux;
669
	struct itrace_start_event	itrace_start;
670
	struct context_switch_event	context_switch;
671
	struct thread_map_event		thread_map;
672
	struct cpu_map_event		cpu_map;
673
	struct stat_config_event	stat_config;
J
Jiri Olsa 已提交
674
	struct stat_event		stat;
675
	struct stat_round_event		stat_round;
676
	struct time_conv_event		time_conv;
677
	struct feature_event		feat;
678
	struct ksymbol_event		ksymbol_event;
679
	struct bpf_event		bpf_event;
680
};
681

682
void perf_event__print_totals(void);
683

684
struct perf_tool;
685
struct thread_map;
686
struct cpu_map;
687
struct perf_stat_config;
688
struct perf_counts_values;
689

690
typedef int (*perf_event__handler_t)(struct perf_tool *tool,
691
				     union perf_event *event,
692
				     struct perf_sample *sample,
693
				     struct machine *machine);
694

695
int perf_event__synthesize_thread_map(struct perf_tool *tool,
696
				      struct thread_map *threads,
697
				      perf_event__handler_t process,
698
				      struct machine *machine, bool mmap_data);
699 700 701 702
int perf_event__synthesize_thread_map2(struct perf_tool *tool,
				      struct thread_map *threads,
				      perf_event__handler_t process,
				      struct machine *machine);
703 704 705 706
int perf_event__synthesize_cpu_map(struct perf_tool *tool,
				   struct cpu_map *cpus,
				   perf_event__handler_t process,
				   struct machine *machine);
707
int perf_event__synthesize_threads(struct perf_tool *tool,
708
				   perf_event__handler_t process,
709
				   struct machine *machine, bool mmap_data,
710
				   unsigned int nr_threads_synthesize);
711
int perf_event__synthesize_kernel_mmap(struct perf_tool *tool,
712
				       perf_event__handler_t process,
713
				       struct machine *machine);
714 715 716 717
int perf_event__synthesize_stat_config(struct perf_tool *tool,
				       struct perf_stat_config *config,
				       perf_event__handler_t process,
				       struct machine *machine);
718 719
void perf_event__read_stat_config(struct perf_stat_config *config,
				  struct stat_config_event *event);
720 721 722 723 724
int perf_event__synthesize_stat(struct perf_tool *tool,
				u32 cpu, u32 thread, u64 id,
				struct perf_counts_values *count,
				perf_event__handler_t process,
				struct machine *machine);
725 726 727 728
int perf_event__synthesize_stat_round(struct perf_tool *tool,
				      u64 time, u64 type,
				      perf_event__handler_t process,
				      struct machine *machine);
729
int perf_event__synthesize_modules(struct perf_tool *tool,
730
				   perf_event__handler_t process,
731 732
				   struct machine *machine);

733
int perf_event__process_comm(struct perf_tool *tool,
734 735
			     union perf_event *event,
			     struct perf_sample *sample,
736
			     struct machine *machine);
737
int perf_event__process_lost(struct perf_tool *tool,
738 739
			     union perf_event *event,
			     struct perf_sample *sample,
740
			     struct machine *machine);
741 742 743 744
int perf_event__process_lost_samples(struct perf_tool *tool,
				     union perf_event *event,
				     struct perf_sample *sample,
				     struct machine *machine);
745 746 747 748
int perf_event__process_aux(struct perf_tool *tool,
			    union perf_event *event,
			    struct perf_sample *sample,
			    struct machine *machine);
749 750 751 752
int perf_event__process_itrace_start(struct perf_tool *tool,
				     union perf_event *event,
				     struct perf_sample *sample,
				     struct machine *machine);
753 754 755 756
int perf_event__process_switch(struct perf_tool *tool,
			       union perf_event *event,
			       struct perf_sample *sample,
			       struct machine *machine);
757 758 759 760
int perf_event__process_namespaces(struct perf_tool *tool,
				   union perf_event *event,
				   struct perf_sample *sample,
				   struct machine *machine);
761
int perf_event__process_mmap(struct perf_tool *tool,
762 763
			     union perf_event *event,
			     struct perf_sample *sample,
764
			     struct machine *machine);
765 766 767 768
int perf_event__process_mmap2(struct perf_tool *tool,
			     union perf_event *event,
			     struct perf_sample *sample,
			     struct machine *machine);
769 770 771 772 773
int perf_event__process_fork(struct perf_tool *tool,
			     union perf_event *event,
			     struct perf_sample *sample,
			     struct machine *machine);
int perf_event__process_exit(struct perf_tool *tool,
774 775
			     union perf_event *event,
			     struct perf_sample *sample,
776
			     struct machine *machine);
777 778 779 780
int perf_event__process_ksymbol(struct perf_tool *tool,
				union perf_event *event,
				struct perf_sample *sample,
				struct machine *machine);
781 782 783 784
int perf_event__process_bpf_event(struct perf_tool *tool,
				  union perf_event *event,
				  struct perf_sample *sample,
				  struct machine *machine);
785 786 787 788
int perf_tool__process_synth_event(struct perf_tool *tool,
				   union perf_event *event,
				   struct machine *machine,
				   perf_event__handler_t process);
789
int perf_event__process(struct perf_tool *tool,
790 791
			union perf_event *event,
			struct perf_sample *sample,
792
			struct machine *machine);
793

794
struct addr_location;
795

796 797
int machine__resolve(struct machine *machine, struct addr_location *al,
		     struct perf_sample *sample);
798

799 800
void addr_location__put(struct addr_location *al);

801 802 803 804
struct thread;

bool is_bts_event(struct perf_event_attr *attr);
bool sample_addr_correlates_sym(struct perf_event_attr *attr);
805 806
void thread__resolve(struct thread *thread, struct addr_location *al,
		     struct perf_sample *sample);
807

808
const char *perf_event__name(unsigned int id);
809

810
size_t perf_event__sample_event_size(const struct perf_sample *sample, u64 type,
811
				     u64 read_format);
812
int perf_event__synthesize_sample(union perf_event *event, u64 type,
813
				  u64 read_format,
814
				  const struct perf_sample *sample);
815

816 817 818 819 820
pid_t perf_event__synthesize_comm(struct perf_tool *tool,
				  union perf_event *event, pid_t pid,
				  perf_event__handler_t process,
				  struct machine *machine);

821 822 823 824 825 826
int perf_event__synthesize_namespaces(struct perf_tool *tool,
				      union perf_event *event,
				      pid_t pid, pid_t tgid,
				      perf_event__handler_t process,
				      struct machine *machine);

827 828 829 830 831
int perf_event__synthesize_mmap_events(struct perf_tool *tool,
				       union perf_event *event,
				       pid_t pid, pid_t tgid,
				       perf_event__handler_t process,
				       struct machine *machine,
832
				       bool mmap_data);
833

834 835 836 837
int perf_event__synthesize_extra_kmaps(struct perf_tool *tool,
				       perf_event__handler_t process,
				       struct machine *machine);

838 839
size_t perf_event__fprintf_comm(union perf_event *event, FILE *fp);
size_t perf_event__fprintf_mmap(union perf_event *event, FILE *fp);
840
size_t perf_event__fprintf_mmap2(union perf_event *event, FILE *fp);
841
size_t perf_event__fprintf_task(union perf_event *event, FILE *fp);
842
size_t perf_event__fprintf_aux(union perf_event *event, FILE *fp);
843
size_t perf_event__fprintf_itrace_start(union perf_event *event, FILE *fp);
844
size_t perf_event__fprintf_switch(union perf_event *event, FILE *fp);
845
size_t perf_event__fprintf_thread_map(union perf_event *event, FILE *fp);
846
size_t perf_event__fprintf_cpu_map(union perf_event *event, FILE *fp);
847
size_t perf_event__fprintf_namespaces(union perf_event *event, FILE *fp);
848
size_t perf_event__fprintf_ksymbol(union perf_event *event, FILE *fp);
849
size_t perf_event__fprintf_bpf_event(union perf_event *event, FILE *fp);
850 851
size_t perf_event__fprintf(union perf_event *event, FILE *fp);

852 853
int kallsyms__get_function_start(const char *kallsyms_filename,
				 const char *symbol_name, u64 *addr);
854

855 856 857
void *cpu_map_data__alloc(struct cpu_map *map, size_t *size, u16 *type, int *max);
void  cpu_map_data__synthesize(struct cpu_map_data *data, struct cpu_map *map,
			       u16 type, int max);
858 859 860 861 862 863 864

void event_attr_init(struct perf_event_attr *attr);

int perf_event_paranoid(void);

extern int sysctl_perf_event_max_stack;
extern int sysctl_perf_event_max_contexts_per_stack;
865
extern unsigned int proc_map_timeout;
866

867
#endif /* __PERF_RECORD_H */