Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Daniel Xu | 785 | 99.24% | 1 | 33.33% |
Andrii Nakryiko | 4 | 0.51% | 1 | 33.33% |
Mykola Lysenko | 2 | 0.25% | 1 | 33.33% |
Total | 791 | 3 |
// SPDX-License-Identifier: GPL-2.0 #define _GNU_SOURCE #include <pthread.h> #include <sched.h> #include <sys/socket.h> #include <test_progs.h> #include "bpf/libbpf_internal.h" #include "test_perf_branches.skel.h" static void check_good_sample(struct test_perf_branches *skel) { int written_global = skel->bss->written_global_out; int required_size = skel->bss->required_size_out; int written_stack = skel->bss->written_stack_out; int pbe_size = sizeof(struct perf_branch_entry); int duration = 0; if (CHECK(!skel->bss->valid, "output not valid", "no valid sample from prog")) return; /* * It's hard to validate the contents of the branch entries b/c it * would require some kind of disassembler and also encoding the * valid jump instructions for supported architectures. So just check * the easy stuff for now. */ CHECK(required_size <= 0, "read_branches_size", "err %d\n", required_size); CHECK(written_stack < 0, "read_branches_stack", "err %d\n", written_stack); CHECK(written_stack % pbe_size != 0, "read_branches_stack", "stack bytes written=%d not multiple of struct size=%d\n", written_stack, pbe_size); CHECK(written_global < 0, "read_branches_global", "err %d\n", written_global); CHECK(written_global % pbe_size != 0, "read_branches_global", "global bytes written=%d not multiple of struct size=%d\n", written_global, pbe_size); CHECK(written_global < written_stack, "read_branches_size", "written_global=%d < written_stack=%d\n", written_global, written_stack); } static void check_bad_sample(struct test_perf_branches *skel) { int written_global = skel->bss->written_global_out; int required_size = skel->bss->required_size_out; int written_stack = skel->bss->written_stack_out; int duration = 0; if (CHECK(!skel->bss->valid, "output not valid", "no valid sample from prog")) return; CHECK((required_size != -EINVAL && required_size != -ENOENT), "read_branches_size", "err %d\n", required_size); CHECK((written_stack != -EINVAL && written_stack != -ENOENT), "read_branches_stack", "written %d\n", written_stack); CHECK((written_global != -EINVAL && written_global != -ENOENT), "read_branches_global", "written %d\n", written_global); } static void test_perf_branches_common(int perf_fd, void (*cb)(struct test_perf_branches *)) { struct test_perf_branches *skel; int err, i, duration = 0; bool detached = false; struct bpf_link *link; volatile int j = 0; cpu_set_t cpu_set; skel = test_perf_branches__open_and_load(); if (CHECK(!skel, "test_perf_branches_load", "perf_branches skeleton failed\n")) return; /* attach perf_event */ link = bpf_program__attach_perf_event(skel->progs.perf_branches, perf_fd); if (!ASSERT_OK_PTR(link, "attach_perf_event")) goto out_destroy_skel; /* generate some branches on cpu 0 */ CPU_ZERO(&cpu_set); CPU_SET(0, &cpu_set); err = pthread_setaffinity_np(pthread_self(), sizeof(cpu_set), &cpu_set); if (CHECK(err, "set_affinity", "cpu #0, err %d\n", err)) goto out_destroy; /* spin the loop for a while (random high number) */ for (i = 0; i < 1000000; ++i) ++j; test_perf_branches__detach(skel); detached = true; cb(skel); out_destroy: bpf_link__destroy(link); out_destroy_skel: if (!detached) test_perf_branches__detach(skel); test_perf_branches__destroy(skel); } static void test_perf_branches_hw(void) { struct perf_event_attr attr = {0}; int duration = 0; int pfd; /* create perf event */ attr.size = sizeof(attr); attr.type = PERF_TYPE_HARDWARE; attr.config = PERF_COUNT_HW_CPU_CYCLES; attr.freq = 1; attr.sample_freq = 1000; attr.sample_type = PERF_SAMPLE_BRANCH_STACK; attr.branch_sample_type = PERF_SAMPLE_BRANCH_USER | PERF_SAMPLE_BRANCH_ANY; pfd = syscall(__NR_perf_event_open, &attr, -1, 0, -1, PERF_FLAG_FD_CLOEXEC); /* * Some setups don't support branch records (virtual machines, !x86), * so skip test in this case. */ if (pfd < 0) { if (errno == ENOENT || errno == EOPNOTSUPP) { printf("%s:SKIP:no PERF_SAMPLE_BRANCH_STACK\n", __func__); test__skip(); return; } if (CHECK(pfd < 0, "perf_event_open", "err %d errno %d\n", pfd, errno)) return; } test_perf_branches_common(pfd, check_good_sample); close(pfd); } /* * Tests negative case -- run bpf_read_branch_records() on improperly configured * perf event. */ static void test_perf_branches_no_hw(void) { struct perf_event_attr attr = {0}; int duration = 0; int pfd; /* create perf event */ attr.size = sizeof(attr); attr.type = PERF_TYPE_SOFTWARE; attr.config = PERF_COUNT_SW_CPU_CLOCK; attr.freq = 1; attr.sample_freq = 1000; pfd = syscall(__NR_perf_event_open, &attr, -1, 0, -1, PERF_FLAG_FD_CLOEXEC); if (CHECK(pfd < 0, "perf_event_open", "err %d\n", pfd)) return; test_perf_branches_common(pfd, check_bad_sample); close(pfd); } void test_perf_branches(void) { if (test__start_subtest("perf_branches_hw")) test_perf_branches_hw(); if (test__start_subtest("perf_branches_no_hw")) test_perf_branches_no_hw(); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1