perf_event.h 2.7 KB
Newer Older
1 2
#ifndef _ASM_X86_PERF_EVENT_H
#define _ASM_X86_PERF_EVENT_H
3

4
/*
5
 * Performance event hw details:
6 7 8 9 10
 */

#define X86_PMC_MAX_GENERIC					8
#define X86_PMC_MAX_FIXED					3

11 12 13 14
#define X86_PMC_IDX_GENERIC				        0
#define X86_PMC_IDX_FIXED				       32
#define X86_PMC_IDX_MAX					       64

I
Ingo Molnar 已提交
15 16
#define MSR_ARCH_PERFMON_PERFCTR0			      0xc1
#define MSR_ARCH_PERFMON_PERFCTR1			      0xc2
17

I
Ingo Molnar 已提交
18 19
#define MSR_ARCH_PERFMON_EVENTSEL0			     0x186
#define MSR_ARCH_PERFMON_EVENTSEL1			     0x187
20

I
Ingo Molnar 已提交
21 22 23 24
#define ARCH_PERFMON_EVENTSEL0_ENABLE			  (1 << 22)
#define ARCH_PERFMON_EVENTSEL_INT			  (1 << 20)
#define ARCH_PERFMON_EVENTSEL_OS			  (1 << 17)
#define ARCH_PERFMON_EVENTSEL_USR			  (1 << 16)
25

26 27 28 29 30
/*
 * Includes eventsel and unit mask as well:
 */
#define ARCH_PERFMON_EVENT_MASK				    0xffff

I
Ingo Molnar 已提交
31 32 33
#define ARCH_PERFMON_UNHALTED_CORE_CYCLES_SEL		      0x3c
#define ARCH_PERFMON_UNHALTED_CORE_CYCLES_UMASK		(0x00 << 8)
#define ARCH_PERFMON_UNHALTED_CORE_CYCLES_INDEX 		 0
34
#define ARCH_PERFMON_UNHALTED_CORE_CYCLES_PRESENT \
I
Ingo Molnar 已提交
35 36 37
		(1 << (ARCH_PERFMON_UNHALTED_CORE_CYCLES_INDEX))

#define ARCH_PERFMON_BRANCH_MISSES_RETIRED			 6
38

39 40 41 42
/*
 * Intel "Architectural Performance Monitoring" CPUID
 * detection/enumeration details:
 */
43 44 45
union cpuid10_eax {
	struct {
		unsigned int version_id:8;
46
		unsigned int num_events:8;
47 48 49 50 51 52
		unsigned int bit_width:8;
		unsigned int mask_length:8;
	} split;
	unsigned int full;
};

53 54
union cpuid10_edx {
	struct {
55
		unsigned int num_events_fixed:4;
56 57 58 59 60 61 62
		unsigned int reserved:28;
	} split;
	unsigned int full;
};


/*
63
 * Fixed-purpose performance events:
64 65
 */

66 67 68 69 70 71 72 73 74
/*
 * All 3 fixed-mode PMCs are configured via this single MSR:
 */
#define MSR_ARCH_PERFMON_FIXED_CTR_CTRL			0x38d

/*
 * The counts are available in three separate MSRs:
 */

75 76
/* Instr_Retired.Any: */
#define MSR_ARCH_PERFMON_FIXED_CTR0			0x309
77
#define X86_PMC_IDX_FIXED_INSTRUCTIONS			(X86_PMC_IDX_FIXED + 0)
78 79 80

/* CPU_CLK_Unhalted.Core: */
#define MSR_ARCH_PERFMON_FIXED_CTR1			0x30a
81
#define X86_PMC_IDX_FIXED_CPU_CYCLES			(X86_PMC_IDX_FIXED + 1)
82 83 84

/* CPU_CLK_Unhalted.Ref: */
#define MSR_ARCH_PERFMON_FIXED_CTR2			0x30b
85
#define X86_PMC_IDX_FIXED_BUS_CYCLES			(X86_PMC_IDX_FIXED + 2)
86

87 88 89
/*
 * We model BTS tracing as another fixed-mode PMC.
 *
90 91
 * We choose a value in the middle of the fixed event range, since lower
 * values are used by actual fixed events and higher values are used
92 93 94 95 96
 * to indicate other overflow conditions in the PERF_GLOBAL_STATUS msr.
 */
#define X86_PMC_IDX_FIXED_BTS				(X86_PMC_IDX_FIXED + 16)


97 98 99
#ifdef CONFIG_PERF_EVENTS
extern void init_hw_perf_events(void);
extern void perf_events_lapic_init(void);
100

101
#define PERF_EVENT_INDEX_OFFSET			0
102

I
Ingo Molnar 已提交
103
#else
104 105
static inline void init_hw_perf_events(void)		{ }
static inline void perf_events_lapic_init(void)	{ }
I
Ingo Molnar 已提交
106 107
#endif

108
#endif /* _ASM_X86_PERF_EVENT_H */