design.txt 17.5 KB
Newer Older
I
Ingo Molnar 已提交
1 2 3 4 5 6 7 8 9 10 11 12

Performance Counters for Linux
------------------------------

Performance counters are special hardware registers available on most modern
CPUs. These registers count the number of certain types of hw events: such
as instructions executed, cachemisses suffered, or branches mis-predicted -
without slowing down the kernel or applications. These registers can also
trigger interrupts when a threshold number of events have passed - and can
thus be used to profile the code that runs on that CPU.

The Linux Performance Counter subsystem provides an abstraction of these
I
Ingo Molnar 已提交
13
hardware capabilities. It provides per task and per CPU counters, counter
14 15 16
groups, and it provides event capabilities on top of those.  It
provides "virtual" 64-bit counters, regardless of the width of the
underlying hardware counters.
I
Ingo Molnar 已提交
17 18 19 20

Performance counters are accessed via special file descriptors.
There's one file descriptor per virtual counter used.

21
The special file descriptor is opened via the sys_perf_event_open()
I
Ingo Molnar 已提交
22 23
system call:

24
   int sys_perf_event_open(struct perf_event_attr *hw_event_uptr,
25 26
			     pid_t pid, int cpu, int group_fd,
			     unsigned long flags);
I
Ingo Molnar 已提交
27 28 29 30 31 32 33 34

The syscall returns the new fd. The fd can be used via the normal
VFS system calls: read() can be used to read the counter, fcntl()
can be used to set the blocking mode, etc.

Multiple counters can be kept open at a time, and the counters
can be poll()ed.

35
When creating a new counter fd, 'perf_event_attr' is:
I
Ingo Molnar 已提交
36

37
struct perf_event_attr {
38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68
        /*
         * The MSB of the config word signifies if the rest contains cpu
         * specific (raw) counter configuration data, if unset, the next
         * 7 bits are an event type and the rest of the bits are the event
         * identifier.
         */
        __u64                   config;

        __u64                   irq_period;
        __u32                   record_type;
        __u32                   read_format;

        __u64                   disabled       :  1, /* off by default        */
                                inherit        :  1, /* children inherit it   */
                                pinned         :  1, /* must always be on PMU */
                                exclusive      :  1, /* only group on PMU     */
                                exclude_user   :  1, /* don't count user      */
                                exclude_kernel :  1, /* ditto kernel          */
                                exclude_hv     :  1, /* ditto hypervisor      */
                                exclude_idle   :  1, /* don't count when idle */
                                mmap           :  1, /* include mmap data     */
                                munmap         :  1, /* include munmap data   */
                                comm           :  1, /* include comm data     */

                                __reserved_1   : 52;

        __u32                   extra_config_len;
        __u32                   wakeup_events;  /* wakeup every n events */

        __u64                   __reserved_2;
        __u64                   __reserved_3;
I
Ingo Molnar 已提交
69 70
};

71
The 'config' field specifies what the counter should count.  It
72 73
is divided into 3 bit-fields:

74 75 76
raw_type: 1 bit   (most significant bit)	0x8000_0000_0000_0000
type:	  7 bits  (next most significant)	0x7f00_0000_0000_0000
event_id: 56 bits (least significant)		0x00ff_ffff_ffff_ffff
77 78 79 80 81 82 83 84

If 'raw_type' is 1, then the counter will count a hardware event
specified by the remaining 63 bits of event_config.  The encoding is
machine-specific.

If 'raw_type' is 0, then the 'type' field says what kind of counter
this is, with the following encoding:

85
enum perf_type_id {
86 87 88 89 90 91 92 93
	PERF_TYPE_HARDWARE		= 0,
	PERF_TYPE_SOFTWARE		= 1,
	PERF_TYPE_TRACEPOINT		= 2,
};

A counter of PERF_TYPE_HARDWARE will count the hardware event
specified by 'event_id':

I
Ingo Molnar 已提交
94
/*
95
 * Generalized performance counter event types, used by the hw_event.event_id
96
 * parameter of the sys_perf_event_open() syscall:
I
Ingo Molnar 已提交
97
 */
98
enum perf_hw_id {
I
Ingo Molnar 已提交
99 100 101
	/*
	 * Common hardware events, generalized by the kernel:
	 */
102 103
	PERF_COUNT_HW_CPU_CYCLES		= 0,
	PERF_COUNT_HW_INSTRUCTIONS		= 1,
K
Kirill Smelkov 已提交
104
	PERF_COUNT_HW_CACHE_REFERENCES		= 2,
105 106
	PERF_COUNT_HW_CACHE_MISSES		= 3,
	PERF_COUNT_HW_BRANCH_INSTRUCTIONS	= 4,
K
Kirill Smelkov 已提交
107
	PERF_COUNT_HW_BRANCH_MISSES		= 5,
108
	PERF_COUNT_HW_BUS_CYCLES		= 6,
I
Ingo Molnar 已提交
109
};
I
Ingo Molnar 已提交
110

111 112 113 114 115 116
These are standardized types of events that work relatively uniformly
on all CPUs that implement Performance Counters support under Linux,
although there may be variations (e.g., different CPUs might count
cache references and misses at different levels of the cache hierarchy).
If a CPU is not able to count the selected event, then the system call
will return -EINVAL.
I
Ingo Molnar 已提交
117

118 119 120 121
More hw_event_types are supported as well, but they are CPU-specific
and accessed as raw events.  For example, to count "External bus
cycles while bus lock signal asserted" events on Intel Core CPUs, pass
in a 0x4064 event_id value and set hw_event.raw_type to 1.
I
Ingo Molnar 已提交
122

123 124
A counter of type PERF_TYPE_SOFTWARE will count one of the available
software events, selected by 'event_id':
I
Ingo Molnar 已提交
125

I
Ingo Molnar 已提交
126
/*
127 128 129 130
 * Special "software" counters provided by the kernel, even if the hardware
 * does not support performance counters. These counters measure various
 * physical and sw events of the kernel (and allow the profiling of them as
 * well):
I
Ingo Molnar 已提交
131
 */
132
enum perf_sw_ids {
133
	PERF_COUNT_SW_CPU_CLOCK		= 0,
K
Kirill Smelkov 已提交
134 135
	PERF_COUNT_SW_TASK_CLOCK	= 1,
	PERF_COUNT_SW_PAGE_FAULTS	= 2,
136 137 138 139
	PERF_COUNT_SW_CONTEXT_SWITCHES	= 3,
	PERF_COUNT_SW_CPU_MIGRATIONS	= 4,
	PERF_COUNT_SW_PAGE_FAULTS_MIN	= 5,
	PERF_COUNT_SW_PAGE_FAULTS_MAJ	= 6,
140 141
	PERF_COUNT_SW_ALIGNMENT_FAULTS	= 7,
	PERF_COUNT_SW_EMULATION_FAULTS	= 8,
I
Ingo Molnar 已提交
142
};
I
Ingo Molnar 已提交
143

144 145 146 147 148
Counters of the type PERF_TYPE_TRACEPOINT are available when the ftrace event
tracer is available, and event_id values can be obtained from
/debug/tracing/events/*/*/id


149 150 151
Counters come in two flavours: counting counters and sampling
counters.  A "counting" counter is one that is used for counting the
number of events that occur, and is characterised by having
152 153 154 155 156 157 158 159 160 161 162 163
irq_period = 0.


A read() on a counter returns the current value of the counter and possible
additional values as specified by 'read_format', each value is a u64 (8 bytes)
in size.

/*
 * Bits that can be set in hw_event.read_format to request that
 * reads on the counter should return the indicated quantities,
 * in increasing order of bit value, after the counter value.
 */
164
enum perf_event_read_format {
165 166 167 168 169 170 171 172
        PERF_FORMAT_TOTAL_TIME_ENABLED  =  1,
        PERF_FORMAT_TOTAL_TIME_RUNNING  =  2,
};

Using these additional values one can establish the overcommit ratio for a
particular counter allowing one to take the round-robin scheduling effect
into account.

I
Ingo Molnar 已提交
173

174 175
A "sampling" counter is one that is set up to generate an interrupt
every N events, where N is given by 'irq_period'.  A sampling counter
176 177
has irq_period > 0. The record_type controls what data is recorded on each
interrupt:
I
Ingo Molnar 已提交
178

179
/*
180 181
 * Bits that can be set in hw_event.record_type to request information
 * in the overflow packets.
182
 */
183
enum perf_event_record_format {
184 185 186 187 188 189
        PERF_RECORD_IP          = 1U << 0,
        PERF_RECORD_TID         = 1U << 1,
        PERF_RECORD_TIME        = 1U << 2,
        PERF_RECORD_ADDR        = 1U << 3,
        PERF_RECORD_GROUP       = 1U << 4,
        PERF_RECORD_CALLCHAIN   = 1U << 5,
190
};
I
Ingo Molnar 已提交
191

192 193
Such (and other) events will be recorded in a ring-buffer, which is
available to user-space using mmap() (see below).
194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224

The 'disabled' bit specifies whether the counter starts out disabled
or enabled.  If it is initially disabled, it can be enabled by ioctl
or prctl (see below).

The 'inherit' bit, if set, specifies that this counter should count
events on descendant tasks as well as the task specified.  This only
applies to new descendents, not to any existing descendents at the
time the counter is created (nor to any new descendents of existing
descendents).

The 'pinned' bit, if set, specifies that the counter should always be
on the CPU if at all possible.  It only applies to hardware counters
and only to group leaders.  If a pinned counter cannot be put onto the
CPU (e.g. because there are not enough hardware counters or because of
a conflict with some other event), then the counter goes into an
'error' state, where reads return end-of-file (i.e. read() returns 0)
until the counter is subsequently enabled or disabled.

The 'exclusive' bit, if set, specifies that when this counter's group
is on the CPU, it should be the only group using the CPU's counters.
In future, this will allow sophisticated monitoring programs to supply
extra configuration information via 'extra_config_len' to exploit
advanced features of the CPU's Performance Monitor Unit (PMU) that are
not otherwise accessible and that might disrupt other hardware
counters.

The 'exclude_user', 'exclude_kernel' and 'exclude_hv' bits provide a
way to request that counting of events be restricted to times when the
CPU is in user, kernel and/or hypervisor mode.

225 226 227 228
Furthermore the 'exclude_host' and 'exclude_guest' bits provide a way
to request counting of events restricted to guest and host contexts when
using Linux as the hypervisor.

229 230 231 232 233 234 235
The 'mmap' and 'munmap' bits allow recording of PROT_EXEC mmap/munmap
operations, these can be used to relate userspace IP addresses to actual
code, even after the mapping (or even the whole process) is gone,
these events are recorded in the ring-buffer (see below).

The 'comm' bit allows tracking of process comm data on process creation.
This too is recorded in the ring-buffer (see below).
236

237
The 'pid' parameter to the sys_perf_event_open() system call allows the
238
counter to be specific to a task:
I
Ingo Molnar 已提交
239 240 241 242 243 244 245 246 247

 pid == 0: if the pid parameter is zero, the counter is attached to the
 current task.

 pid > 0: the counter is attached to a specific task (if the current task
 has sufficient privilege to do so)

 pid < 0: all tasks are counted (per cpu counters)

248
The 'cpu' parameter allows a counter to be made specific to a CPU:
I
Ingo Molnar 已提交
249 250 251 252

 cpu >= 0: the counter is restricted to a specific CPU
 cpu == -1: the counter counts on all CPUs

I
Ingo Molnar 已提交
253
(Note: the combination of 'pid == -1' and 'cpu == -1' is not valid.)
I
Ingo Molnar 已提交
254 255 256 257 258 259 260 261 262

A 'pid > 0' and 'cpu == -1' counter is a per task counter that counts
events of that task and 'follows' that task to whatever CPU the task
gets schedule to. Per task counters can be created by any user, for
their own tasks.

A 'pid == -1' and 'cpu == x' counter is a per CPU counter that counts
all events on CPU-x. Per CPU counters need CAP_SYS_ADMIN privilege.

263 264 265 266
The 'flags' parameter is currently unused and must be zero.

The 'group_fd' parameter allows counter "groups" to be set up.  A
counter group has one counter which is the group "leader".  The leader
267
is created first, with group_fd = -1 in the sys_perf_event_open call
268 269 270 271 272 273 274 275 276 277 278 279
that creates it.  The rest of the group members are created
subsequently, with group_fd giving the fd of the group leader.
(A single counter on its own is created with group_fd = -1 and is
considered to be a group with only 1 member.)

A counter group is scheduled onto the CPU as a unit, that is, it will
only be put onto the CPU if all of the counters in the group can be
put onto the CPU.  This means that the values of the member counters
can be meaningfully compared, added, divided (to get ratios), etc.,
with each other, since they have counted events for the same set of
executed instructions.

280 281 282 283 284 285

Like stated, asynchronous events, like counter overflow or PROT_EXEC mmap
tracking are logged into a ring-buffer. This ring-buffer is created and
accessed through mmap().

The mmap size should be 1+2^n pages, where the first page is a meta-data page
286
(struct perf_event_mmap_page) that contains various bits of information such
287 288 289 290 291
as where the ring-buffer head is.

/*
 * Structure of the page that can be mapped via mmap
 */
292
struct perf_event_mmap_page {
293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325
        __u32   version;                /* version number of this structure */
        __u32   compat_version;         /* lowest version this is compat with */

        /*
         * Bits needed to read the hw counters in user-space.
         *
         *   u32 seq;
         *   s64 count;
         *
         *   do {
         *     seq = pc->lock;
         *
         *     barrier()
         *     if (pc->index) {
         *       count = pmc_read(pc->index - 1);
         *       count += pc->offset;
         *     } else
         *       goto regular_read;
         *
         *     barrier();
         *   } while (pc->lock != seq);
         *
         * NOTE: for obvious reason this only works on self-monitoring
         *       processes.
         */
        __u32   lock;                   /* seqlock for synchronization */
        __u32   index;                  /* hardware counter identifier */
        __s64   offset;                 /* add to hardware counter value */

        /*
         * Control data for the mmap() data buffer.
         *
         * User-space reading this value should issue an rmb(), on SMP capable
326
         * platforms, after reading this value -- see perf_event_wakeup().
327 328 329 330 331 332 333 334 335
         */
        __u32   data_head;              /* head in the data section */
};

NOTE: the hw-counter userspace bits are arch specific and are currently only
      implemented on powerpc.

The following 2^n pages are the ring-buffer which contains events of the form:

336 337 338
#define PERF_RECORD_MISC_KERNEL          (1 << 0)
#define PERF_RECORD_MISC_USER            (1 << 1)
#define PERF_RECORD_MISC_OVERFLOW        (1 << 2)
339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361

struct perf_event_header {
        __u32   type;
        __u16   misc;
        __u16   size;
};

enum perf_event_type {

        /*
         * The MMAP events record the PROT_EXEC mappings so that we can
         * correlate userspace IPs to code. They have the following structure:
         *
         * struct {
         *      struct perf_event_header        header;
         *
         *      u32                             pid, tid;
         *      u64                             addr;
         *      u64                             len;
         *      u64                             pgoff;
         *      char                            filename[];
         * };
         */
362 363
        PERF_RECORD_MMAP                 = 1,
        PERF_RECORD_MUNMAP               = 2,
364 365 366 367 368 369 370 371 372

        /*
         * struct {
         *      struct perf_event_header        header;
         *
         *      u32                             pid, tid;
         *      char                            comm[];
         * };
         */
373
        PERF_RECORD_COMM                 = 3,
374 375

        /*
376
         * When header.misc & PERF_RECORD_MISC_OVERFLOW the event_type field
377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405
         * will be PERF_RECORD_*
         *
         * struct {
         *      struct perf_event_header        header;
         *
         *      { u64                   ip;       } && PERF_RECORD_IP
         *      { u32                   pid, tid; } && PERF_RECORD_TID
         *      { u64                   time;     } && PERF_RECORD_TIME
         *      { u64                   addr;     } && PERF_RECORD_ADDR
         *
         *      { u64                   nr;
         *        { u64 event, val; }   cnt[nr];  } && PERF_RECORD_GROUP
         *
         *      { u16                   nr,
         *                              hv,
         *                              kernel,
         *                              user;
         *        u64                   ips[nr];  } && PERF_RECORD_CALLCHAIN
         * };
         */
};

NOTE: PERF_RECORD_CALLCHAIN is arch specific and currently only implemented
      on x86.

Notification of new events is possible through poll()/select()/epoll() and
fcntl() managing signals.

Normally a notification is generated for every page filled, however one can
406
additionally set perf_event_attr.wakeup_events to generate one every
407 408 409 410 411
so many counter overflow events.

Future work will include a splice() interface to the ring-buffer.


412 413 414 415
Counters can be enabled and disabled in two ways: via ioctl and via
prctl.  When a counter is disabled, it doesn't count or generate
events but does continue to exist and maintain its count value.

416
An individual counter can be enabled with
417

418
	ioctl(fd, PERF_EVENT_IOC_ENABLE, 0);
419 420 421

or disabled with

422
	ioctl(fd, PERF_EVENT_IOC_DISABLE, 0);
423

424
For a counter group, pass PERF_IOC_FLAG_GROUP as the third argument.
425 426 427 428 429 430 431
Enabling or disabling the leader of a group enables or disables the
whole group; that is, while the group leader is disabled, none of the
counters in the group will count.  Enabling or disabling a member of a
group other than the leader only affects that counter - disabling an
non-leader stops that counter from counting but doesn't affect any
other counter.

432 433
Additionally, non-inherited overflow counters can use

434
	ioctl(fd, PERF_EVENT_IOC_REFRESH, nr);
435 436 437

to enable a counter for 'nr' events, after which it gets disabled again.

438 439 440
A process can enable or disable all the counter groups that are
attached to it, using prctl:

441
	prctl(PR_TASK_PERF_EVENTS_ENABLE);
442

443
	prctl(PR_TASK_PERF_EVENTS_DISABLE);
444 445 446 447 448

This applies to all counters on the current process, whether created
by this process or by another, and doesn't affect any counters that
this process has created on other processes.  It only enables or
disables the group leaders, not any other members in the groups.
I
Ingo Molnar 已提交
449

450 451 452 453 454 455 456

Arch requirements
-----------------

If your architecture does not have hardware performance metrics, you can
still use the generic software counters based on hrtimers for sampling.

457
So to start with, in order to add HAVE_PERF_EVENTS to your Kconfig, you
458
will need at least this:
459
	- asm/perf_event.h - a basic stub will suffice at first
460 461 462
	- support for atomic64 types (and associated helper functions)

If your architecture does have hardware capabilities, you can override the
463
weak stub hw_perf_event_init() to register hardware counters.
464 465 466

Architectures that have d-cache aliassing issues, such as Sparc and ARM,
should select PERF_USE_VMALLOC in order to avoid these for perf mmap().