event.h 9.8 KB
Newer Older
1
/* SPDX-License-Identifier: GPL-2.0 */
2 3
#ifndef __PERF_RECORD_H
#define __PERF_RECORD_H
4 5 6 7 8 9
/*
 * The linux/stddef.h isn't need here, but is needed for __always_inline used
 * in files included from uapi/linux/perf_event.h such as
 * /usr/include/linux/swab.h and /usr/include/linux/byteorder/little_endian.h,
 * detected in at least musl libc, used in Alpine Linux. -acme
 */
10
#include <stdio.h>
11
#include <linux/stddef.h>
12
#include <perf/event.h>
13
#include <linux/types.h>
14

15
#include "perf_regs.h"
16

17 18 19 20
struct dso;
struct machine;
struct perf_event_attr;

21 22 23 24 25 26 27 28 29
#ifdef __LP64__
/*
 * /usr/include/inttypes.h uses just 'lu' for PRIu64, but we end up defining
 * __u64 as long long unsigned int, and then -Werror=format= kicks in and
 * complains of the mismatched types, so use these two special extra PRI
 * macros to overcome that.
 */
#define PRI_lu64 "l" PRIu64
#define PRI_lx64 "l" PRIx64
30
#define PRI_ld64 "l" PRId64
31 32 33
#else
#define PRI_lu64 PRIu64
#define PRI_lx64 PRIx64
34
#define PRI_ld64 PRId64
35
#endif
36

37 38 39 40
#define PERF_SAMPLE_MASK				\
	(PERF_SAMPLE_IP | PERF_SAMPLE_TID |		\
	 PERF_SAMPLE_TIME | PERF_SAMPLE_ADDR |		\
	PERF_SAMPLE_ID | PERF_SAMPLE_STREAM_ID |	\
41 42
	 PERF_SAMPLE_CPU | PERF_SAMPLE_PERIOD |		\
	 PERF_SAMPLE_IDENTIFIER)
43

44 45 46
/* perf sample has 16 bits size limit */
#define PERF_SAMPLE_MAX_SIZE (1 << 16)

47
struct regs_dump {
48
	u64 abi;
49
	u64 mask;
50
	u64 *regs;
51 52 53 54

	/* Cached values/mask filled by first register access. */
	u64 cache_regs[PERF_REGS_MAX];
	u64 cache_mask;
55 56 57 58 59 60 61 62
};

struct stack_dump {
	u16 offset;
	u64 size;
	char *data;
};

63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79
struct sample_read_value {
	u64 value;
	u64 id;
};

struct sample_read {
	u64 time_enabled;
	u64 time_running;
	union {
		struct {
			u64 nr;
			struct sample_read_value *values;
		} group;
		struct sample_read_value one;
	};
};

80 81 82 83 84
struct ip_callchain {
	u64 nr;
	u64 ips[0];
};

85
struct branch_stack;
86

87 88 89 90 91 92 93 94 95 96 97 98 99 100
enum {
	PERF_IP_FLAG_BRANCH		= 1ULL << 0,
	PERF_IP_FLAG_CALL		= 1ULL << 1,
	PERF_IP_FLAG_RETURN		= 1ULL << 2,
	PERF_IP_FLAG_CONDITIONAL	= 1ULL << 3,
	PERF_IP_FLAG_SYSCALLRET		= 1ULL << 4,
	PERF_IP_FLAG_ASYNC		= 1ULL << 5,
	PERF_IP_FLAG_INTERRUPT		= 1ULL << 6,
	PERF_IP_FLAG_TX_ABORT		= 1ULL << 7,
	PERF_IP_FLAG_TRACE_BEGIN	= 1ULL << 8,
	PERF_IP_FLAG_TRACE_END		= 1ULL << 9,
	PERF_IP_FLAG_IN_TX		= 1ULL << 10,
};

101 102
#define PERF_IP_FLAG_CHARS "bcrosyiABEx"

103 104 105 106 107 108 109 110 111 112 113 114
#define PERF_BRANCH_MASK		(\
	PERF_IP_FLAG_BRANCH		|\
	PERF_IP_FLAG_CALL		|\
	PERF_IP_FLAG_RETURN		|\
	PERF_IP_FLAG_CONDITIONAL	|\
	PERF_IP_FLAG_SYSCALLRET		|\
	PERF_IP_FLAG_ASYNC		|\
	PERF_IP_FLAG_INTERRUPT		|\
	PERF_IP_FLAG_TX_ABORT		|\
	PERF_IP_FLAG_TRACE_BEGIN	|\
	PERF_IP_FLAG_TRACE_END)

115 116
#define MAX_INSN 16

117 118 119 120 121
struct aux_sample {
	u64 size;
	void *data;
};

122
struct perf_sample {
123 124 125 126 127 128 129
	u64 ip;
	u32 pid, tid;
	u64 time;
	u64 addr;
	u64 id;
	u64 stream_id;
	u64 period;
130
	u64 weight;
131
	u64 transaction;
132 133
	u64 insn_cnt;
	u64 cyc_cnt;
134
	u32 cpu;
135
	u32 raw_size;
136
	u64 data_src;
137
	u64 phys_addr;
138 139
	u32 flags;
	u16 insn_len;
140
	u8  cpumode;
141
	u16 misc;
142
	char insn[MAX_INSN];
143
	void *raw_data;
144
	struct ip_callchain *callchain;
145
	struct branch_stack *branch_stack;
146
	struct regs_dump  user_regs;
147
	struct regs_dump  intr_regs;
148
	struct stack_dump user_stack;
149
	struct sample_read read;
150
	struct aux_sample aux_sample;
151 152
};

153 154 155 156 157 158 159
#define PERF_MEM_DATA_SRC_NONE \
	(PERF_MEM_S(OP, NA) |\
	 PERF_MEM_S(LVL, NA) |\
	 PERF_MEM_S(SNOOP, NA) |\
	 PERF_MEM_S(LOCK, NA) |\
	 PERF_MEM_S(TLB, NA))

160 161 162
/* Attribute type for custom synthesized events */
#define PERF_TYPE_SYNTH		(INT_MAX + 1U)

163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280
/* Attribute config for custom synthesized events */
enum perf_synth_id {
	PERF_SYNTH_INTEL_PTWRITE,
	PERF_SYNTH_INTEL_MWAIT,
	PERF_SYNTH_INTEL_PWRE,
	PERF_SYNTH_INTEL_EXSTOP,
	PERF_SYNTH_INTEL_PWRX,
	PERF_SYNTH_INTEL_CBR,
};

/*
 * Raw data formats for synthesized events. Note that 4 bytes of padding are
 * present to match the 'size' member of PERF_SAMPLE_RAW data which is always
 * 8-byte aligned. That means we must dereference raw_data with an offset of 4.
 * Refer perf_sample__synth_ptr() and perf_synth__raw_data().  It also means the
 * structure sizes are 4 bytes bigger than the raw_size, refer
 * perf_synth__raw_size().
 */

struct perf_synth_intel_ptwrite {
	u32 padding;
	union {
		struct {
			u32	ip		:  1,
				reserved	: 31;
		};
		u32	flags;
	};
	u64	payload;
};

struct perf_synth_intel_mwait {
	u32 padding;
	u32 reserved;
	union {
		struct {
			u64	hints		:  8,
				reserved1	: 24,
				extensions	:  2,
				reserved2	: 30;
		};
		u64	payload;
	};
};

struct perf_synth_intel_pwre {
	u32 padding;
	u32 reserved;
	union {
		struct {
			u64	reserved1	:  7,
				hw		:  1,
				subcstate	:  4,
				cstate		:  4,
				reserved2	: 48;
		};
		u64	payload;
	};
};

struct perf_synth_intel_exstop {
	u32 padding;
	union {
		struct {
			u32	ip		:  1,
				reserved	: 31;
		};
		u32	flags;
	};
};

struct perf_synth_intel_pwrx {
	u32 padding;
	u32 reserved;
	union {
		struct {
			u64	deepest_cstate	:  4,
				last_cstate	:  4,
				wake_reason	:  4,
				reserved1	: 52;
		};
		u64	payload;
	};
};

struct perf_synth_intel_cbr {
	u32 padding;
	union {
		struct {
			u32	cbr		:  8,
				reserved1	:  8,
				max_nonturbo	:  8,
				reserved2	:  8;
		};
		u32	flags;
	};
	u32 freq;
	u32 reserved3;
};

/*
 * raw_data is always 4 bytes from an 8-byte boundary, so subtract 4 to get
 * 8-byte alignment.
 */
static inline void *perf_sample__synth_ptr(struct perf_sample *sample)
{
	return sample->raw_data - 4;
}

static inline void *perf_synth__raw_data(void *p)
{
	return p + 4;
}

#define perf_synth__raw_size(d) (sizeof(d) - 4)

#define perf_sample__bad_synth_size(s, d) ((s)->raw_size < sizeof(d) - 4)

281 282 283 284 285
enum {
	PERF_STAT_ROUND_TYPE__INTERVAL	= 0,
	PERF_STAT_ROUND_TYPE__FINAL	= 1,
};

286
void perf_event__print_totals(void);
287

288
struct perf_cpu_map;
289
struct perf_record_stat_config;
290
struct perf_stat_config;
291
struct perf_tool;
292

293
void perf_event__read_stat_config(struct perf_stat_config *config,
294
				  struct perf_record_stat_config *event);
295

296
int perf_event__process_comm(struct perf_tool *tool,
297 298
			     union perf_event *event,
			     struct perf_sample *sample,
299
			     struct machine *machine);
300
int perf_event__process_lost(struct perf_tool *tool,
301 302
			     union perf_event *event,
			     struct perf_sample *sample,
303
			     struct machine *machine);
304 305 306 307
int perf_event__process_lost_samples(struct perf_tool *tool,
				     union perf_event *event,
				     struct perf_sample *sample,
				     struct machine *machine);
308 309 310 311
int perf_event__process_aux(struct perf_tool *tool,
			    union perf_event *event,
			    struct perf_sample *sample,
			    struct machine *machine);
312 313 314 315
int perf_event__process_itrace_start(struct perf_tool *tool,
				     union perf_event *event,
				     struct perf_sample *sample,
				     struct machine *machine);
316 317 318 319
int perf_event__process_switch(struct perf_tool *tool,
			       union perf_event *event,
			       struct perf_sample *sample,
			       struct machine *machine);
320 321 322 323
int perf_event__process_namespaces(struct perf_tool *tool,
				   union perf_event *event,
				   struct perf_sample *sample,
				   struct machine *machine);
324
int perf_event__process_mmap(struct perf_tool *tool,
325 326
			     union perf_event *event,
			     struct perf_sample *sample,
327
			     struct machine *machine);
328 329 330 331
int perf_event__process_mmap2(struct perf_tool *tool,
			     union perf_event *event,
			     struct perf_sample *sample,
			     struct machine *machine);
332 333 334 335 336
int perf_event__process_fork(struct perf_tool *tool,
			     union perf_event *event,
			     struct perf_sample *sample,
			     struct machine *machine);
int perf_event__process_exit(struct perf_tool *tool,
337 338
			     union perf_event *event,
			     struct perf_sample *sample,
339
			     struct machine *machine);
340 341 342 343
int perf_event__process_ksymbol(struct perf_tool *tool,
				union perf_event *event,
				struct perf_sample *sample,
				struct machine *machine);
344 345 346 347
int perf_event__process_bpf(struct perf_tool *tool,
			    union perf_event *event,
			    struct perf_sample *sample,
			    struct machine *machine);
348
int perf_event__process(struct perf_tool *tool,
349 350
			union perf_event *event,
			struct perf_sample *sample,
351
			struct machine *machine);
352

353
struct addr_location;
354

355 356
int machine__resolve(struct machine *machine, struct addr_location *al,
		     struct perf_sample *sample);
357

358 359
void addr_location__put(struct addr_location *al);

360 361 362 363
struct thread;

bool is_bts_event(struct perf_event_attr *attr);
bool sample_addr_correlates_sym(struct perf_event_attr *attr);
364 365
void thread__resolve(struct thread *thread, struct addr_location *al,
		     struct perf_sample *sample);
366

367
const char *perf_event__name(unsigned int id);
368

369 370
size_t perf_event__fprintf_comm(union perf_event *event, FILE *fp);
size_t perf_event__fprintf_mmap(union perf_event *event, FILE *fp);
371
size_t perf_event__fprintf_mmap2(union perf_event *event, FILE *fp);
372
size_t perf_event__fprintf_task(union perf_event *event, FILE *fp);
373
size_t perf_event__fprintf_aux(union perf_event *event, FILE *fp);
374
size_t perf_event__fprintf_itrace_start(union perf_event *event, FILE *fp);
375
size_t perf_event__fprintf_switch(union perf_event *event, FILE *fp);
376
size_t perf_event__fprintf_thread_map(union perf_event *event, FILE *fp);
377
size_t perf_event__fprintf_cpu_map(union perf_event *event, FILE *fp);
378
size_t perf_event__fprintf_namespaces(union perf_event *event, FILE *fp);
379
size_t perf_event__fprintf_ksymbol(union perf_event *event, FILE *fp);
380
size_t perf_event__fprintf_bpf(union perf_event *event, FILE *fp);
381 382
size_t perf_event__fprintf(union perf_event *event, FILE *fp);

383 384
int kallsyms__get_function_start(const char *kallsyms_filename,
				 const char *symbol_name, u64 *addr);
385

386
void *cpu_map_data__alloc(struct perf_cpu_map *map, size_t *size, u16 *type, int *max);
387
void  cpu_map_data__synthesize(struct perf_record_cpu_map_data *data, struct perf_cpu_map *map,
388
			       u16 type, int max);
389 390 391 392

void event_attr_init(struct perf_event_attr *attr);

int perf_event_paranoid(void);
393
bool perf_event_paranoid_check(int max_level);
394 395 396

extern int sysctl_perf_event_max_stack;
extern int sysctl_perf_event_max_contexts_per_stack;
397
extern unsigned int proc_map_timeout;
398

399
#endif /* __PERF_RECORD_H */