Linux Audio

Check our new training course

Loading...
Note: File does not exist in v5.9.
  1// SPDX-License-Identifier: GPL-2.0
  2/* Copyright (c) 2021 Facebook */
  3
  4#define _GNU_SOURCE         /* See feature_test_macros(7) */
  5#include <unistd.h>
  6#include <sched.h>
  7#include <pthread.h>
  8#include <sys/syscall.h>   /* For SYS_xxx definitions */
  9#include <sys/types.h>
 10#include <test_progs.h>
 11#include "task_local_storage_helpers.h"
 12#include "task_local_storage.skel.h"
 13#include "task_local_storage_exit_creds.skel.h"
 14#include "task_ls_recursion.skel.h"
 15#include "task_storage_nodeadlock.skel.h"
 16
 17static void test_sys_enter_exit(void)
 18{
 19	struct task_local_storage *skel;
 20	int err;
 21
 22	skel = task_local_storage__open_and_load();
 23	if (!ASSERT_OK_PTR(skel, "skel_open_and_load"))
 24		return;
 25
 26	skel->bss->target_pid = syscall(SYS_gettid);
 27
 28	err = task_local_storage__attach(skel);
 29	if (!ASSERT_OK(err, "skel_attach"))
 30		goto out;
 31
 32	syscall(SYS_gettid);
 33	syscall(SYS_gettid);
 34
 35	/* 3x syscalls: 1x attach and 2x gettid */
 36	ASSERT_EQ(skel->bss->enter_cnt, 3, "enter_cnt");
 37	ASSERT_EQ(skel->bss->exit_cnt, 3, "exit_cnt");
 38	ASSERT_EQ(skel->bss->mismatch_cnt, 0, "mismatch_cnt");
 39out:
 40	task_local_storage__destroy(skel);
 41}
 42
 43static void test_exit_creds(void)
 44{
 45	struct task_local_storage_exit_creds *skel;
 46	int err, run_count, sync_rcu_calls = 0;
 47	const int MAX_SYNC_RCU_CALLS = 1000;
 48
 49	skel = task_local_storage_exit_creds__open_and_load();
 50	if (!ASSERT_OK_PTR(skel, "skel_open_and_load"))
 51		return;
 52
 53	err = task_local_storage_exit_creds__attach(skel);
 54	if (!ASSERT_OK(err, "skel_attach"))
 55		goto out;
 56
 57	/* trigger at least one exit_creds() */
 58	if (CHECK_FAIL(system("ls > /dev/null")))
 59		goto out;
 60
 61	/* kern_sync_rcu is not enough on its own as the read section we want
 62	 * to wait for may start after we enter synchronize_rcu, so our call
 63	 * won't wait for the section to finish. Loop on the run counter
 64	 * as well to ensure the program has run.
 65	 */
 66	do {
 67		kern_sync_rcu();
 68		run_count = __atomic_load_n(&skel->bss->run_count, __ATOMIC_SEQ_CST);
 69	} while (run_count == 0 && ++sync_rcu_calls < MAX_SYNC_RCU_CALLS);
 70
 71	ASSERT_NEQ(sync_rcu_calls, MAX_SYNC_RCU_CALLS,
 72		   "sync_rcu count too high");
 73	ASSERT_NEQ(run_count, 0, "run_count");
 74	ASSERT_EQ(skel->bss->valid_ptr_count, 0, "valid_ptr_count");
 75	ASSERT_NEQ(skel->bss->null_ptr_count, 0, "null_ptr_count");
 76out:
 77	task_local_storage_exit_creds__destroy(skel);
 78}
 79
 80static void test_recursion(void)
 81{
 82	int err, map_fd, prog_fd, task_fd;
 83	struct task_ls_recursion *skel;
 84	struct bpf_prog_info info;
 85	__u32 info_len = sizeof(info);
 86	long value;
 87
 88	task_fd = sys_pidfd_open(getpid(), 0);
 89	if (!ASSERT_NEQ(task_fd, -1, "sys_pidfd_open"))
 90		return;
 91
 92	skel = task_ls_recursion__open_and_load();
 93	if (!ASSERT_OK_PTR(skel, "skel_open_and_load"))
 94		goto out;
 95
 96	err = task_ls_recursion__attach(skel);
 97	if (!ASSERT_OK(err, "skel_attach"))
 98		goto out;
 99
100	/* trigger sys_enter, make sure it does not cause deadlock */
101	skel->bss->test_pid = getpid();
102	syscall(SYS_gettid);
103	skel->bss->test_pid = 0;
104	task_ls_recursion__detach(skel);
105
106	/* Refer to the comment in BPF_PROG(on_update) for
107	 * the explanation on the value 201 and 100.
108	 */
109	map_fd = bpf_map__fd(skel->maps.map_a);
110	err = bpf_map_lookup_elem(map_fd, &task_fd, &value);
111	ASSERT_OK(err, "lookup map_a");
112	ASSERT_EQ(value, 201, "map_a value");
113	ASSERT_EQ(skel->bss->nr_del_errs, 1, "bpf_task_storage_delete busy");
114
115	map_fd = bpf_map__fd(skel->maps.map_b);
116	err = bpf_map_lookup_elem(map_fd, &task_fd, &value);
117	ASSERT_OK(err, "lookup map_b");
118	ASSERT_EQ(value, 100, "map_b value");
119
120	prog_fd = bpf_program__fd(skel->progs.on_lookup);
121	memset(&info, 0, sizeof(info));
122	err = bpf_prog_get_info_by_fd(prog_fd, &info, &info_len);
123	ASSERT_OK(err, "get prog info");
124	ASSERT_GT(info.recursion_misses, 0, "on_lookup prog recursion");
125
126	prog_fd = bpf_program__fd(skel->progs.on_update);
127	memset(&info, 0, sizeof(info));
128	err = bpf_prog_get_info_by_fd(prog_fd, &info, &info_len);
129	ASSERT_OK(err, "get prog info");
130	ASSERT_EQ(info.recursion_misses, 0, "on_update prog recursion");
131
132	prog_fd = bpf_program__fd(skel->progs.on_enter);
133	memset(&info, 0, sizeof(info));
134	err = bpf_prog_get_info_by_fd(prog_fd, &info, &info_len);
135	ASSERT_OK(err, "get prog info");
136	ASSERT_EQ(info.recursion_misses, 0, "on_enter prog recursion");
137
138out:
139	close(task_fd);
140	task_ls_recursion__destroy(skel);
141}
142
143static bool stop;
144
145static void waitall(const pthread_t *tids, int nr)
146{
147	int i;
148
149	stop = true;
150	for (i = 0; i < nr; i++)
151		pthread_join(tids[i], NULL);
152}
153
154static void *sock_create_loop(void *arg)
155{
156	struct task_storage_nodeadlock *skel = arg;
157	int fd;
158
159	while (!stop) {
160		fd = socket(AF_INET, SOCK_STREAM, 0);
161		close(fd);
162		if (skel->bss->nr_get_errs || skel->bss->nr_del_errs)
163			stop = true;
164	}
165
166	return NULL;
167}
168
169static void test_nodeadlock(void)
170{
171	struct task_storage_nodeadlock *skel;
172	struct bpf_prog_info info = {};
173	__u32 info_len = sizeof(info);
174	const int nr_threads = 32;
175	pthread_t tids[nr_threads];
176	int i, prog_fd, err;
177	cpu_set_t old, new;
178
179	/* Pin all threads to one cpu to increase the chance of preemption
180	 * in a sleepable bpf prog.
181	 */
182	CPU_ZERO(&new);
183	CPU_SET(0, &new);
184	err = sched_getaffinity(getpid(), sizeof(old), &old);
185	if (!ASSERT_OK(err, "getaffinity"))
186		return;
187	err = sched_setaffinity(getpid(), sizeof(new), &new);
188	if (!ASSERT_OK(err, "setaffinity"))
189		return;
190
191	skel = task_storage_nodeadlock__open_and_load();
192	if (!ASSERT_OK_PTR(skel, "open_and_load"))
193		goto done;
194
195	/* Unnecessary recursion and deadlock detection are reproducible
196	 * in the preemptible kernel.
197	 */
198	if (!skel->kconfig->CONFIG_PREEMPT) {
199		test__skip();
200		goto done;
201	}
202
203	err = task_storage_nodeadlock__attach(skel);
204	ASSERT_OK(err, "attach prog");
205
206	for (i = 0; i < nr_threads; i++) {
207		err = pthread_create(&tids[i], NULL, sock_create_loop, skel);
208		if (err) {
209			/* Only assert once here to avoid excessive
210			 * PASS printing during test failure.
211			 */
212			ASSERT_OK(err, "pthread_create");
213			waitall(tids, i);
214			goto done;
215		}
216	}
217
218	/* With 32 threads, 1s is enough to reproduce the issue */
219	sleep(1);
220	waitall(tids, nr_threads);
221
222	info_len = sizeof(info);
223	prog_fd = bpf_program__fd(skel->progs.socket_post_create);
224	err = bpf_prog_get_info_by_fd(prog_fd, &info, &info_len);
225	ASSERT_OK(err, "get prog info");
226	ASSERT_EQ(info.recursion_misses, 0, "prog recursion");
227
228	ASSERT_EQ(skel->bss->nr_get_errs, 0, "bpf_task_storage_get busy");
229	ASSERT_EQ(skel->bss->nr_del_errs, 0, "bpf_task_storage_delete busy");
230
231done:
232	task_storage_nodeadlock__destroy(skel);
233	sched_setaffinity(getpid(), sizeof(old), &old);
234}
235
236void test_task_local_storage(void)
237{
238	if (test__start_subtest("sys_enter_exit"))
239		test_sys_enter_exit();
240	if (test__start_subtest("exit_creds"))
241		test_exit_creds();
242	if (test__start_subtest("recursion"))
243		test_recursion();
244	if (test__start_subtest("nodeadlock"))
245		test_nodeadlock();
246}