map_kptr.c 5.3 KB
Newer Older
1 2
// SPDX-License-Identifier: GPL-2.0
#include <test_progs.h>
3
#include <network_helpers.h>
4 5

#include "map_kptr.skel.h"
6
#include "map_kptr_fail.skel.h"
7

8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84
static char log_buf[1024 * 1024];

struct {
	const char *prog_name;
	const char *err_msg;
} map_kptr_fail_tests[] = {
	{ "size_not_bpf_dw", "kptr access size must be BPF_DW" },
	{ "non_const_var_off", "kptr access cannot have variable offset" },
	{ "non_const_var_off_kptr_xchg", "R1 doesn't have constant offset. kptr has to be" },
	{ "misaligned_access_write", "kptr access misaligned expected=8 off=7" },
	{ "misaligned_access_read", "kptr access misaligned expected=8 off=1" },
	{ "reject_var_off_store", "variable untrusted_ptr_ access var_off=(0x0; 0x1e0)" },
	{ "reject_bad_type_match", "invalid kptr access, R1 type=untrusted_ptr_prog_test_ref_kfunc" },
	{ "marked_as_untrusted_or_null", "R1 type=untrusted_ptr_or_null_ expected=percpu_ptr_" },
	{ "correct_btf_id_check_size", "access beyond struct prog_test_ref_kfunc at off 32 size 4" },
	{ "inherit_untrusted_on_walk", "R1 type=untrusted_ptr_ expected=percpu_ptr_" },
	{ "reject_kptr_xchg_on_unref", "off=8 kptr isn't referenced kptr" },
	{ "reject_kptr_get_no_map_val", "arg#0 expected pointer to map value" },
	{ "reject_kptr_get_no_null_map_val", "arg#0 expected pointer to map value" },
	{ "reject_kptr_get_no_kptr", "arg#0 no referenced kptr at map value offset=0" },
	{ "reject_kptr_get_on_unref", "arg#0 no referenced kptr at map value offset=8" },
	{ "reject_kptr_get_bad_type_match", "kernel function bpf_kfunc_call_test_kptr_get args#0" },
	{ "mark_ref_as_untrusted_or_null", "R1 type=untrusted_ptr_or_null_ expected=percpu_ptr_" },
	{ "reject_untrusted_store_to_ref", "store to referenced kptr disallowed" },
	{ "reject_bad_type_xchg", "invalid kptr access, R2 type=ptr_prog_test_ref_kfunc expected=ptr_prog_test_member" },
	{ "reject_untrusted_xchg", "R2 type=untrusted_ptr_ expected=ptr_" },
	{ "reject_member_of_ref_xchg", "invalid kptr access, R2 type=ptr_prog_test_ref_kfunc" },
	{ "reject_indirect_helper_access", "kptr cannot be accessed indirectly by helper" },
	{ "reject_indirect_global_func_access", "kptr cannot be accessed indirectly by helper" },
	{ "kptr_xchg_ref_state", "Unreleased reference id=5 alloc_insn=" },
	{ "kptr_get_ref_state", "Unreleased reference id=3 alloc_insn=" },
};

static void test_map_kptr_fail_prog(const char *prog_name, const char *err_msg)
{
	LIBBPF_OPTS(bpf_object_open_opts, opts, .kernel_log_buf = log_buf,
						.kernel_log_size = sizeof(log_buf),
						.kernel_log_level = 1);
	struct map_kptr_fail *skel;
	struct bpf_program *prog;
	int ret;

	skel = map_kptr_fail__open_opts(&opts);
	if (!ASSERT_OK_PTR(skel, "map_kptr_fail__open_opts"))
		return;

	prog = bpf_object__find_program_by_name(skel->obj, prog_name);
	if (!ASSERT_OK_PTR(prog, "bpf_object__find_program_by_name"))
		goto end;

	bpf_program__set_autoload(prog, true);

	ret = map_kptr_fail__load(skel);
	if (!ASSERT_ERR(ret, "map_kptr__load must fail"))
		goto end;

	if (!ASSERT_OK_PTR(strstr(log_buf, err_msg), "expected error message")) {
		fprintf(stderr, "Expected: %s\n", err_msg);
		fprintf(stderr, "Verifier: %s\n", log_buf);
	}

end:
	map_kptr_fail__destroy(skel);
}

static void test_map_kptr_fail(void)
{
	int i;

	for (i = 0; i < ARRAY_SIZE(map_kptr_fail_tests); i++) {
		if (!test__start_subtest(map_kptr_fail_tests[i].prog_name))
			continue;
		test_map_kptr_fail_prog(map_kptr_fail_tests[i].prog_name,
					map_kptr_fail_tests[i].err_msg);
	}
}

85
static void test_map_kptr_success(bool test_run)
86
{
87 88 89 90 91
	LIBBPF_OPTS(bpf_test_run_opts, opts,
		.data_in = &pkt_v4,
		.data_size_in = sizeof(pkt_v4),
		.repeat = 1,
	);
92 93 94 95 96 97 98 99
	struct map_kptr *skel;
	int key = 0, ret;
	char buf[24];

	skel = map_kptr__open_and_load();
	if (!ASSERT_OK_PTR(skel, "map_kptr__open_and_load"))
		return;

100 101 102 103 104 105 106 107 108 109
	ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_map_kptr_ref), &opts);
	ASSERT_OK(ret, "test_map_kptr_ref refcount");
	ASSERT_OK(opts.retval, "test_map_kptr_ref retval");
	ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.test_map_kptr_ref2), &opts);
	ASSERT_OK(ret, "test_map_kptr_ref2 refcount");
	ASSERT_OK(opts.retval, "test_map_kptr_ref2 retval");

	if (test_run)
		return;

110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131
	ret = bpf_map_update_elem(bpf_map__fd(skel->maps.array_map), &key, buf, 0);
	ASSERT_OK(ret, "array_map update");
	ret = bpf_map_update_elem(bpf_map__fd(skel->maps.array_map), &key, buf, 0);
	ASSERT_OK(ret, "array_map update2");

	ret = bpf_map_update_elem(bpf_map__fd(skel->maps.hash_map), &key, buf, 0);
	ASSERT_OK(ret, "hash_map update");
	ret = bpf_map_delete_elem(bpf_map__fd(skel->maps.hash_map), &key);
	ASSERT_OK(ret, "hash_map delete");

	ret = bpf_map_update_elem(bpf_map__fd(skel->maps.hash_malloc_map), &key, buf, 0);
	ASSERT_OK(ret, "hash_malloc_map update");
	ret = bpf_map_delete_elem(bpf_map__fd(skel->maps.hash_malloc_map), &key);
	ASSERT_OK(ret, "hash_malloc_map delete");

	ret = bpf_map_update_elem(bpf_map__fd(skel->maps.lru_hash_map), &key, buf, 0);
	ASSERT_OK(ret, "lru_hash_map update");
	ret = bpf_map_delete_elem(bpf_map__fd(skel->maps.lru_hash_map), &key);
	ASSERT_OK(ret, "lru_hash_map delete");

	map_kptr__destroy(skel);
}
132 133 134

void test_map_kptr(void)
{
135 136 137 138 139 140 141
	if (test__start_subtest("success")) {
		test_map_kptr_success(false);
		/* Do test_run twice, so that we see refcount going back to 1
		 * after we leave it in map from first iteration.
		 */
		test_map_kptr_success(true);
	}
142 143
	test_map_kptr_fail();
}