Linux Audio

Check our new training course

Loading...
Note: File does not exist in v4.6.
  1// SPDX-License-Identifier: GPL-2.0
  2#include <errno.h>
  3#include <inttypes.h>
  4#include <limits.h>
  5#include <stdbool.h>
  6#include <stdio.h>
  7#include <unistd.h>
  8#include <linux/types.h>
  9#include <sys/prctl.h>
 10#include <perf/cpumap.h>
 11#include <perf/evlist.h>
 12#include <perf/mmap.h>
 13
 14#include "debug.h"
 15#include "parse-events.h"
 16#include "evlist.h"
 17#include "evsel.h"
 18#include "thread_map.h"
 19#include "record.h"
 20#include "tsc.h"
 21#include "mmap.h"
 22#include "tests.h"
 23#include "pmu.h"
 24#include "pmu-hybrid.h"
 25
 26#define CHECK__(x) {				\
 27	while ((x) < 0) {			\
 28		pr_debug(#x " failed!\n");	\
 29		goto out_err;			\
 30	}					\
 31}
 32
 33#define CHECK_NOT_NULL__(x) {			\
 34	while ((x) == NULL) {			\
 35		pr_debug(#x " failed!\n");	\
 36		goto out_err;			\
 37	}					\
 38}
 39
 40/**
 41 * test__perf_time_to_tsc - test converting perf time to TSC.
 42 *
 43 * This function implements a test that checks that the conversion of perf time
 44 * to and from TSC is consistent with the order of events.  If the test passes
 45 * %0 is returned, otherwise %-1 is returned.  If TSC conversion is not
 46 * supported then then the test passes but " (not supported)" is printed.
 47 */
 48int test__perf_time_to_tsc(struct test *test __maybe_unused, int subtest __maybe_unused)
 49{
 50	struct record_opts opts = {
 51		.mmap_pages	     = UINT_MAX,
 52		.user_freq	     = UINT_MAX,
 53		.user_interval	     = ULLONG_MAX,
 54		.target		     = {
 55			.uses_mmap   = true,
 56		},
 57		.sample_time	     = true,
 58	};
 59	struct perf_thread_map *threads = NULL;
 60	struct perf_cpu_map *cpus = NULL;
 61	struct evlist *evlist = NULL;
 62	struct evsel *evsel = NULL;
 63	int err = -1, ret, i;
 64	const char *comm1, *comm2;
 65	struct perf_tsc_conversion tc;
 66	struct perf_event_mmap_page *pc;
 67	union perf_event *event;
 68	u64 test_tsc, comm1_tsc, comm2_tsc;
 69	u64 test_time, comm1_time = 0, comm2_time = 0;
 70	struct mmap *md;
 71
 72	threads = thread_map__new(-1, getpid(), UINT_MAX);
 73	CHECK_NOT_NULL__(threads);
 74
 75	cpus = perf_cpu_map__new(NULL);
 76	CHECK_NOT_NULL__(cpus);
 77
 78	evlist = evlist__new();
 79	CHECK_NOT_NULL__(evlist);
 80
 81	perf_evlist__set_maps(&evlist->core, cpus, threads);
 82
 83	CHECK__(parse_events(evlist, "cycles:u", NULL));
 84
 85	evlist__config(evlist, &opts, NULL);
 86
 87	evsel = evlist__first(evlist);
 88
 89	evsel->core.attr.comm = 1;
 90	evsel->core.attr.disabled = 1;
 91	evsel->core.attr.enable_on_exec = 0;
 92
 93	/*
 94	 * For hybrid "cycles:u", it creates two events.
 95	 * Init the second evsel here.
 96	 */
 97	if (perf_pmu__has_hybrid() && perf_pmu__hybrid_mounted("cpu_atom")) {
 98		evsel = evsel__next(evsel);
 99		evsel->core.attr.comm = 1;
100		evsel->core.attr.disabled = 1;
101		evsel->core.attr.enable_on_exec = 0;
102	}
103
104	CHECK__(evlist__open(evlist));
105
106	CHECK__(evlist__mmap(evlist, UINT_MAX));
107
108	pc = evlist->mmap[0].core.base;
109	ret = perf_read_tsc_conversion(pc, &tc);
110	if (ret) {
111		if (ret == -EOPNOTSUPP) {
112			fprintf(stderr, " (not supported)");
113			return 0;
114		}
115		goto out_err;
116	}
117
118	evlist__enable(evlist);
119
120	comm1 = "Test COMM 1";
121	CHECK__(prctl(PR_SET_NAME, (unsigned long)comm1, 0, 0, 0));
122
123	test_tsc = rdtsc();
124
125	comm2 = "Test COMM 2";
126	CHECK__(prctl(PR_SET_NAME, (unsigned long)comm2, 0, 0, 0));
127
128	evlist__disable(evlist);
129
130	for (i = 0; i < evlist->core.nr_mmaps; i++) {
131		md = &evlist->mmap[i];
132		if (perf_mmap__read_init(&md->core) < 0)
133			continue;
134
135		while ((event = perf_mmap__read_event(&md->core)) != NULL) {
136			struct perf_sample sample;
137
138			if (event->header.type != PERF_RECORD_COMM ||
139			    (pid_t)event->comm.pid != getpid() ||
140			    (pid_t)event->comm.tid != getpid())
141				goto next_event;
142
143			if (strcmp(event->comm.comm, comm1) == 0) {
144				CHECK__(evsel__parse_sample(evsel, event, &sample));
145				comm1_time = sample.time;
146			}
147			if (strcmp(event->comm.comm, comm2) == 0) {
148				CHECK__(evsel__parse_sample(evsel, event, &sample));
149				comm2_time = sample.time;
150			}
151next_event:
152			perf_mmap__consume(&md->core);
153		}
154		perf_mmap__read_done(&md->core);
155	}
156
157	if (!comm1_time || !comm2_time)
158		goto out_err;
159
160	test_time = tsc_to_perf_time(test_tsc, &tc);
161	comm1_tsc = perf_time_to_tsc(comm1_time, &tc);
162	comm2_tsc = perf_time_to_tsc(comm2_time, &tc);
163
164	pr_debug("1st event perf time %"PRIu64" tsc %"PRIu64"\n",
165		 comm1_time, comm1_tsc);
166	pr_debug("rdtsc          time %"PRIu64" tsc %"PRIu64"\n",
167		 test_time, test_tsc);
168	pr_debug("2nd event perf time %"PRIu64" tsc %"PRIu64"\n",
169		 comm2_time, comm2_tsc);
170
171	if (test_time <= comm1_time ||
172	    test_time >= comm2_time)
173		goto out_err;
174
175	if (test_tsc <= comm1_tsc ||
176	    test_tsc >= comm2_tsc)
177		goto out_err;
178
179	err = 0;
180
181out_err:
182	evlist__delete(evlist);
183	perf_cpu_map__put(cpus);
184	perf_thread_map__put(threads);
185	return err;
186}
187
188bool test__tsc_is_supported(void)
189{
190	/*
191	 * Except x86_64/i386 and Arm64, other archs don't support TSC in perf.
192	 * Just enable the test for x86_64/i386 and Arm64 archs.
193	 */
194#if defined(__x86_64__) || defined(__i386__) || defined(__aarch64__)
195	return true;
196#else
197	return false;
198#endif
199}