perf-record.c 8.3 KB
Newer Older
1
// SPDX-License-Identifier: GPL-2.0
2
#include <errno.h>
3
#include <inttypes.h>
4 5 6
/* For the CLR_() macros */
#include <pthread.h>

7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40
#include <sched.h>
#include "evlist.h"
#include "evsel.h"
#include "perf.h"
#include "debug.h"
#include "tests.h"

static int sched__get_first_possible_cpu(pid_t pid, cpu_set_t *maskp)
{
	int i, cpu = -1, nrcpus = 1024;
realloc:
	CPU_ZERO(maskp);

	if (sched_getaffinity(pid, sizeof(*maskp), maskp) == -1) {
		if (errno == EINVAL && nrcpus < (1024 << 8)) {
			nrcpus = nrcpus << 2;
			goto realloc;
		}
		perror("sched_getaffinity");
			return -1;
	}

	for (i = 0; i < nrcpus; i++) {
		if (CPU_ISSET(i, maskp)) {
			if (cpu == -1)
				cpu = i;
			else
				CPU_CLR(i, maskp);
		}
	}

	return cpu;
}

41
int test__PERF_RECORD(struct test *test __maybe_unused, int subtest __maybe_unused)
42
{
43
	struct record_opts opts = {
44 45 46 47
		.target = {
			.uid = UINT_MAX,
			.uses_mmap = true,
		},
48 49
		.no_buffering = true,
		.mmap_pages   = 256,
50 51 52
	};
	cpu_set_t cpu_mask;
	size_t cpu_mask_size = sizeof(cpu_mask);
53
	struct evlist *evlist = perf_evlist__new_dummy();
54
	struct evsel *evsel;
55 56 57
	struct perf_sample sample;
	const char *cmd = "sleep";
	const char *argv[] = { cmd, "1", NULL, };
58
	char *bname, *mmap_filename;
59 60
	u64 prev_time = 0;
	bool found_cmd_mmap = false,
61
	     found_coreutils_mmap = false,
62 63 64 65 66 67
	     found_libc_mmap = false,
	     found_vdso_mmap = false,
	     found_ld_mmap = false;
	int err = -1, errs = 0, i, wakeups = 0;
	u32 cpu;
	int total_events = 0, nr_events[PERF_RECORD_MAX] = { 0, };
68
	char sbuf[STRERR_BUFSIZE];
69

70 71 72
	if (evlist == NULL) /* Fallback for kernels lacking PERF_COUNT_SW_DUMMY */
		evlist = perf_evlist__new_default();

73
	if (evlist == NULL) {
74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95
		pr_debug("Not enough memory to create evlist\n");
		goto out;
	}

	/*
	 * Create maps of threads and cpus to monitor. In this case
	 * we start with all threads and cpus (-1, -1) but then in
	 * perf_evlist__prepare_workload we'll fill in the only thread
	 * we're monitoring, the one forked there.
	 */
	err = perf_evlist__create_maps(evlist, &opts.target);
	if (err < 0) {
		pr_debug("Not enough memory to create thread/cpu maps\n");
		goto out_delete_evlist;
	}

	/*
	 * Prepare the workload in argv[] to run, it'll fork it, and then wait
	 * for perf_evlist__start_workload() to exec it. This is done this way
	 * so that we have time to open the evlist (calling sys_perf_event_open
	 * on all the fds) and then mmap them.
	 */
96
	err = perf_evlist__prepare_workload(evlist, &opts.target, argv, false, NULL);
97 98
	if (err < 0) {
		pr_debug("Couldn't run the workload!\n");
99
		goto out_delete_evlist;
100 101 102 103 104 105
	}

	/*
	 * Config the evsels, setting attr->comm on the first one, etc.
	 */
	evsel = perf_evlist__first(evlist);
106 107 108
	perf_evsel__set_sample_bit(evsel, CPU);
	perf_evsel__set_sample_bit(evsel, TID);
	perf_evsel__set_sample_bit(evsel, TIME);
109
	perf_evlist__config(evlist, &opts, NULL);
110 111 112

	err = sched__get_first_possible_cpu(evlist->workload.pid, &cpu_mask);
	if (err < 0) {
113
		pr_debug("sched__get_first_possible_cpu: %s\n",
114
			 str_error_r(errno, sbuf, sizeof(sbuf)));
115
		goto out_delete_evlist;
116 117 118 119 120 121 122 123
	}

	cpu = err;

	/*
	 * So that we can check perf_sample.cpu on all the samples.
	 */
	if (sched_setaffinity(evlist->workload.pid, cpu_mask_size, &cpu_mask) < 0) {
124
		pr_debug("sched_setaffinity: %s\n",
125
			 str_error_r(errno, sbuf, sizeof(sbuf)));
126
		goto out_delete_evlist;
127 128 129 130 131 132 133 134
	}

	/*
	 * Call sys_perf_event_open on all the fds on all the evsels,
	 * grouping them if asked to.
	 */
	err = perf_evlist__open(evlist);
	if (err < 0) {
135
		pr_debug("perf_evlist__open: %s\n",
136
			 str_error_r(errno, sbuf, sizeof(sbuf)));
137
		goto out_delete_evlist;
138 139 140 141 142 143 144
	}

	/*
	 * mmap the first fd on a given CPU and ask for events for the other
	 * fds in the same CPU to be injected in the same mmap ring buffer
	 * (using ioctl(PERF_EVENT_IOC_SET_OUTPUT)).
	 */
145
	err = perf_evlist__mmap(evlist, opts.mmap_pages);
146
	if (err < 0) {
147
		pr_debug("perf_evlist__mmap: %s\n",
148
			 str_error_r(errno, sbuf, sizeof(sbuf)));
149
		goto out_delete_evlist;
150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167
	}

	/*
	 * Now that all is properly set up, enable the events, they will
	 * count just on workload.pid, which will start...
	 */
	perf_evlist__enable(evlist);

	/*
	 * Now!
	 */
	perf_evlist__start_workload(evlist);

	while (1) {
		int before = total_events;

		for (i = 0; i < evlist->nr_mmaps; i++) {
			union perf_event *event;
168
			struct perf_mmap *md;
169

170
			md = &evlist->mmap[i];
171
			if (perf_mmap__read_init(md) < 0)
172 173
				continue;

174
			while ((event = perf_mmap__read_event(md)) != NULL) {
175 176 177 178 179 180 181 182 183
				const u32 type = event->header.type;
				const char *name = perf_event__name(type);

				++total_events;
				if (type < PERF_RECORD_MAX)
					nr_events[type]++;

				err = perf_evlist__parse_sample(evlist, event, &sample);
				if (err < 0) {
184
					if (verbose > 0)
185 186
						perf_event__fprintf(event, stderr);
					pr_debug("Couldn't parse sample\n");
187
					goto out_delete_evlist;
188 189
				}

190
				if (verbose > 0) {
191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222
					pr_info("%" PRIu64" %d ", sample.time, sample.cpu);
					perf_event__fprintf(event, stderr);
				}

				if (prev_time > sample.time) {
					pr_debug("%s going backwards in time, prev=%" PRIu64 ", curr=%" PRIu64 "\n",
						 name, prev_time, sample.time);
					++errs;
				}

				prev_time = sample.time;

				if (sample.cpu != cpu) {
					pr_debug("%s with unexpected cpu, expected %d, got %d\n",
						 name, cpu, sample.cpu);
					++errs;
				}

				if ((pid_t)sample.pid != evlist->workload.pid) {
					pr_debug("%s with unexpected pid, expected %d, got %d\n",
						 name, evlist->workload.pid, sample.pid);
					++errs;
				}

				if ((pid_t)sample.tid != evlist->workload.pid) {
					pr_debug("%s with unexpected tid, expected %d, got %d\n",
						 name, evlist->workload.pid, sample.tid);
					++errs;
				}

				if ((type == PERF_RECORD_COMM ||
				     type == PERF_RECORD_MMAP ||
223
				     type == PERF_RECORD_MMAP2 ||
224 225 226 227 228 229 230 231
				     type == PERF_RECORD_FORK ||
				     type == PERF_RECORD_EXIT) &&
				     (pid_t)event->comm.pid != evlist->workload.pid) {
					pr_debug("%s with unexpected pid/tid\n", name);
					++errs;
				}

				if ((type == PERF_RECORD_COMM ||
232 233
				     type == PERF_RECORD_MMAP ||
				     type == PERF_RECORD_MMAP2) &&
234 235 236 237 238 239 240 241 242 243 244 245 246 247 248
				     event->comm.pid != event->comm.tid) {
					pr_debug("%s with different pid/tid!\n", name);
					++errs;
				}

				switch (type) {
				case PERF_RECORD_COMM:
					if (strcmp(event->comm.comm, cmd)) {
						pr_debug("%s with unexpected comm!\n", name);
						++errs;
					}
					break;
				case PERF_RECORD_EXIT:
					goto found_exit;
				case PERF_RECORD_MMAP:
249 250 251 252 253 254
					mmap_filename = event->mmap.filename;
					goto check_bname;
				case PERF_RECORD_MMAP2:
					mmap_filename = event->mmap2.filename;
				check_bname:
					bname = strrchr(mmap_filename, '/');
255 256 257
					if (bname != NULL) {
						if (!found_cmd_mmap)
							found_cmd_mmap = !strcmp(bname + 1, cmd);
258 259
						if (!found_coreutils_mmap)
							found_coreutils_mmap = !strcmp(bname + 1, "coreutils");
260 261 262 263 264
						if (!found_libc_mmap)
							found_libc_mmap = !strncmp(bname + 1, "libc", 4);
						if (!found_ld_mmap)
							found_ld_mmap = !strncmp(bname + 1, "ld", 2);
					} else if (!found_vdso_mmap)
265
						found_vdso_mmap = !strcmp(mmap_filename, "[vdso]");
266 267 268 269 270 271 272 273 274 275
					break;

				case PERF_RECORD_SAMPLE:
					/* Just ignore samples for now */
					break;
				default:
					pr_debug("Unexpected perf_event->header.type %d!\n",
						 type);
					++errs;
				}
276

277
				perf_mmap__consume(md);
278
			}
279
			perf_mmap__read_done(md);
280 281 282 283 284 285 286 287
		}

		/*
		 * We don't use poll here because at least at 3.1 times the
		 * PERF_RECORD_{!SAMPLE} events don't honour
		 * perf_event_attr.wakeup_events, just PERF_EVENT_SAMPLE does.
		 */
		if (total_events == before && false)
288
			perf_evlist__poll(evlist, -1);
289 290 291 292 293 294 295 296 297

		sleep(1);
		if (++wakeups > 5) {
			pr_debug("No PERF_RECORD_EXIT event!\n");
			break;
		}
	}

found_exit:
298
	if (nr_events[PERF_RECORD_COMM] > 1 + !!found_coreutils_mmap) {
299 300 301 302 303 304 305 306 307
		pr_debug("Excessive number of PERF_RECORD_COMM events!\n");
		++errs;
	}

	if (nr_events[PERF_RECORD_COMM] == 0) {
		pr_debug("Missing PERF_RECORD_COMM for %s!\n", cmd);
		++errs;
	}

308
	if (!found_cmd_mmap && !found_coreutils_mmap) {
309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327
		pr_debug("PERF_RECORD_MMAP for %s missing!\n", cmd);
		++errs;
	}

	if (!found_libc_mmap) {
		pr_debug("PERF_RECORD_MMAP for %s missing!\n", "libc");
		++errs;
	}

	if (!found_ld_mmap) {
		pr_debug("PERF_RECORD_MMAP for %s missing!\n", "ld");
		++errs;
	}

	if (!found_vdso_mmap) {
		pr_debug("PERF_RECORD_MMAP for %s missing!\n", "[vdso]");
		++errs;
	}
out_delete_evlist:
328
	evlist__delete(evlist);
329 330 331
out:
	return (err < 0 || errs > 0) ? -1 : 0;
}