Contributors: 1
Author Tokens Token Proportion Commits Commit Proportion
Anton Protopopov 1824 100.00% 1 100.00%
Total 1824 1


// SPDX-License-Identifier: GPL-2.0

#include <test_progs.h>

#include <linux/if_ether.h>
#include <linux/in.h>
#include <linux/ip.h>
#include <linux/ipv6.h>
#include <linux/in6.h>
#include <linux/udp.h>
#include <linux/tcp.h>

#include <sys/syscall.h>
#include <bpf/bpf.h>

#include "bpf_gotox.skel.h"

static void __test_run(struct bpf_program *prog, void *ctx_in, size_t ctx_size_in)
{
	LIBBPF_OPTS(bpf_test_run_opts, topts,
			    .ctx_in = ctx_in,
			    .ctx_size_in = ctx_size_in,
		   );
	int err, prog_fd;

	prog_fd = bpf_program__fd(prog);
	err = bpf_prog_test_run_opts(prog_fd, &topts);
	ASSERT_OK(err, "test_run_opts err");
}

static void __subtest(struct bpf_gotox *skel, void (*check)(struct bpf_gotox *))
{
	if (skel->data->skip)
		test__skip();
	else
		check(skel);
}

static void check_simple(struct bpf_gotox *skel,
			 struct bpf_program *prog,
			 __u64 ctx_in,
			 __u64 expected)
{
	skel->bss->ret_user = 0;

	__test_run(prog, &ctx_in, sizeof(ctx_in));

	if (!ASSERT_EQ(skel->bss->ret_user, expected, "skel->bss->ret_user"))
		return;
}

static void check_simple_fentry(struct bpf_gotox *skel,
				struct bpf_program *prog,
				__u64 ctx_in,
				__u64 expected)
{
	skel->bss->in_user = ctx_in;
	skel->bss->ret_user = 0;

	/* trigger */
	usleep(1);

	if (!ASSERT_EQ(skel->bss->ret_user, expected, "skel->bss->ret_user"))
		return;
}

/* validate that for two loads of the same jump table libbpf generates only one map */
static void check_one_map_two_jumps(struct bpf_gotox *skel)
{
	struct bpf_prog_info prog_info;
	struct bpf_map_info map_info;
	__u32 len;
	__u32 map_ids[16];
	int prog_fd, map_fd;
	int ret;
	int i;
	bool seen = false;

	memset(&prog_info, 0, sizeof(prog_info));
	prog_info.map_ids = (long)map_ids;
	prog_info.nr_map_ids = ARRAY_SIZE(map_ids);
	prog_fd = bpf_program__fd(skel->progs.one_map_two_jumps);
	if (!ASSERT_GE(prog_fd, 0, "bpf_program__fd(one_map_two_jumps)"))
		return;

	len = sizeof(prog_info);
	ret = bpf_obj_get_info_by_fd(prog_fd, &prog_info, &len);
	if (!ASSERT_OK(ret, "bpf_obj_get_info_by_fd(prog_fd)"))
		return;

	for (i = 0; i < prog_info.nr_map_ids; i++) {
		map_fd  = bpf_map_get_fd_by_id(map_ids[i]);
		if (!ASSERT_GE(map_fd, 0, "bpf_map_get_fd_by_id"))
			return;

		len = sizeof(map_info);
		memset(&map_info, 0, len);
		ret = bpf_obj_get_info_by_fd(map_fd, &map_info, &len);
		if (!ASSERT_OK(ret, "bpf_obj_get_info_by_fd(map_fd)")) {
			close(map_fd);
			return;
		}

		if (map_info.type == BPF_MAP_TYPE_INSN_ARRAY) {
			if (!ASSERT_EQ(seen, false, "more than one INSN_ARRAY map")) {
				close(map_fd);
				return;
			}
			seen = true;
		}
		close(map_fd);
	}

	ASSERT_EQ(seen, true, "no INSN_ARRAY map");
}

static void check_one_switch(struct bpf_gotox *skel)
{
	__u64 in[]   = {0, 1, 2, 3, 4,  5, 77};
	__u64 out[]  = {2, 3, 4, 5, 7, 19, 19};
	int i;

	for (i = 0; i < ARRAY_SIZE(in); i++)
		check_simple(skel, skel->progs.one_switch, in[i], out[i]);
}

static void check_one_switch_non_zero_sec_off(struct bpf_gotox *skel)
{
	__u64 in[]   = {0, 1, 2, 3, 4,  5, 77};
	__u64 out[]  = {2, 3, 4, 5, 7, 19, 19};
	int i;

	for (i = 0; i < ARRAY_SIZE(in); i++)
		check_simple(skel, skel->progs.one_switch_non_zero_sec_off, in[i], out[i]);
}

static void check_two_switches(struct bpf_gotox *skel)
{
	__u64 in[]   = {0, 1, 2, 3, 4,  5, 77};
	__u64 out[] = {103, 104, 107, 205, 115, 1019, 1019};
	int i;

	for (i = 0; i < ARRAY_SIZE(in); i++)
		check_simple(skel, skel->progs.two_switches, in[i], out[i]);
}

static void check_big_jump_table(struct bpf_gotox *skel)
{
	__u64 in[]  = {0, 11, 27, 31, 22, 45, 99};
	__u64 out[] = {2,  3,  4,  5, 19, 19, 19};
	int i;

	for (i = 0; i < ARRAY_SIZE(in); i++)
		check_simple(skel, skel->progs.big_jump_table, in[i], out[i]);
}

static void check_one_jump_two_maps(struct bpf_gotox *skel)
{
	__u64 in[]  = {0, 1, 2, 3, 4,  5, 77};
	__u64 out[] = {12, 15, 7 , 15, 12, 15, 15};
	int i;

	for (i = 0; i < ARRAY_SIZE(in); i++)
		check_simple(skel, skel->progs.one_jump_two_maps, in[i], out[i]);
}

static void check_static_global(struct bpf_gotox *skel)
{
	__u64 in[]   = {0, 1, 2, 3, 4,  5, 77};
	__u64 out[]  = {2, 3, 4, 5, 7, 19, 19};
	int i;

	for (i = 0; i < ARRAY_SIZE(in); i++)
		check_simple(skel, skel->progs.use_static_global1, in[i], out[i]);
	for (i = 0; i < ARRAY_SIZE(in); i++)
		check_simple(skel, skel->progs.use_static_global2, in[i], out[i]);
}

static void check_nonstatic_global(struct bpf_gotox *skel)
{
	__u64 in[]   = {0, 1, 2, 3, 4,  5, 77};
	__u64 out[]  = {2, 3, 4, 5, 7, 19, 19};
	int i;

	for (i = 0; i < ARRAY_SIZE(in); i++)
		check_simple(skel, skel->progs.use_nonstatic_global1, in[i], out[i]);

	for (i = 0; i < ARRAY_SIZE(in); i++)
		check_simple(skel, skel->progs.use_nonstatic_global2, in[i], out[i]);
}

static void check_other_sec(struct bpf_gotox *skel)
{
	struct bpf_link *link;
	__u64 in[]   = {0, 1, 2, 3, 4,  5, 77};
	__u64 out[]  = {2, 3, 4, 5, 7, 19, 19};
	int i;

	link = bpf_program__attach(skel->progs.simple_test_other_sec);
	if (!ASSERT_OK_PTR(link, "link"))
		return;

	for (i = 0; i < ARRAY_SIZE(in); i++)
		check_simple_fentry(skel, skel->progs.simple_test_other_sec, in[i], out[i]);

	bpf_link__destroy(link);
}

static void check_static_global_other_sec(struct bpf_gotox *skel)
{
	struct bpf_link *link;
	__u64 in[]   = {0, 1, 2, 3, 4,  5, 77};
	__u64 out[]  = {2, 3, 4, 5, 7, 19, 19};
	int i;

	link = bpf_program__attach(skel->progs.use_static_global_other_sec);
	if (!ASSERT_OK_PTR(link, "link"))
		return;

	for (i = 0; i < ARRAY_SIZE(in); i++)
		check_simple_fentry(skel, skel->progs.use_static_global_other_sec, in[i], out[i]);

	bpf_link__destroy(link);
}

static void check_nonstatic_global_other_sec(struct bpf_gotox *skel)
{
	struct bpf_link *link;
	__u64 in[]   = {0, 1, 2, 3, 4,  5, 77};
	__u64 out[]  = {2, 3, 4, 5, 7, 19, 19};
	int i;

	link = bpf_program__attach(skel->progs.use_nonstatic_global_other_sec);
	if (!ASSERT_OK_PTR(link, "link"))
		return;

	for (i = 0; i < ARRAY_SIZE(in); i++)
		check_simple_fentry(skel, skel->progs.use_nonstatic_global_other_sec, in[i], out[i]);

	bpf_link__destroy(link);
}

void test_bpf_gotox(void)
{
	struct bpf_gotox *skel;
	int ret;

	skel = bpf_gotox__open();
	if (!ASSERT_NEQ(skel, NULL, "bpf_gotox__open"))
		return;

	ret = bpf_gotox__load(skel);
	if (!ASSERT_OK(ret, "bpf_gotox__load"))
		return;

	skel->bss->pid = getpid();

	if (test__start_subtest("one-switch"))
		__subtest(skel, check_one_switch);

	if (test__start_subtest("one-switch-non-zero-sec-offset"))
		__subtest(skel, check_one_switch_non_zero_sec_off);

	if (test__start_subtest("two-switches"))
		__subtest(skel, check_two_switches);

	if (test__start_subtest("big-jump-table"))
		__subtest(skel, check_big_jump_table);

	if (test__start_subtest("static-global"))
		__subtest(skel, check_static_global);

	if (test__start_subtest("nonstatic-global"))
		__subtest(skel, check_nonstatic_global);

	if (test__start_subtest("other-sec"))
		__subtest(skel, check_other_sec);

	if (test__start_subtest("static-global-other-sec"))
		__subtest(skel, check_static_global_other_sec);

	if (test__start_subtest("nonstatic-global-other-sec"))
		__subtest(skel, check_nonstatic_global_other_sec);

	if (test__start_subtest("one-jump-two-maps"))
		__subtest(skel, check_one_jump_two_maps);

	if (test__start_subtest("one-map-two-jumps"))
		__subtest(skel, check_one_map_two_jumps);

	bpf_gotox__destroy(skel);
}