Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Kumar Kartikeya Dwivedi | 2170 | 95.64% | 3 | 33.33% |
Andrii Nakryiko | 48 | 2.12% | 1 | 11.11% |
Alexei Starovoitov | 43 | 1.90% | 3 | 33.33% |
David Vernet | 5 | 0.22% | 1 | 11.11% |
Jiri Olsa | 3 | 0.13% | 1 | 11.11% |
Total | 2269 | 9 |
// SPDX-License-Identifier: GPL-2.0 #include <vmlinux.h> #include <bpf/bpf_tracing.h> #include <bpf/bpf_helpers.h> #include "../bpf_testmod/bpf_testmod_kfunc.h" struct map_value { struct prog_test_ref_kfunc __kptr_untrusted *unref_ptr; struct prog_test_ref_kfunc __kptr *ref_ptr; }; struct array_map { __uint(type, BPF_MAP_TYPE_ARRAY); __type(key, int); __type(value, struct map_value); __uint(max_entries, 1); } array_map SEC(".maps"); struct pcpu_array_map { __uint(type, BPF_MAP_TYPE_PERCPU_ARRAY); __type(key, int); __type(value, struct map_value); __uint(max_entries, 1); } pcpu_array_map SEC(".maps"); struct hash_map { __uint(type, BPF_MAP_TYPE_HASH); __type(key, int); __type(value, struct map_value); __uint(max_entries, 1); } hash_map SEC(".maps"); struct pcpu_hash_map { __uint(type, BPF_MAP_TYPE_PERCPU_HASH); __type(key, int); __type(value, struct map_value); __uint(max_entries, 1); } pcpu_hash_map SEC(".maps"); struct hash_malloc_map { __uint(type, BPF_MAP_TYPE_HASH); __type(key, int); __type(value, struct map_value); __uint(max_entries, 1); __uint(map_flags, BPF_F_NO_PREALLOC); } hash_malloc_map SEC(".maps"); struct pcpu_hash_malloc_map { __uint(type, BPF_MAP_TYPE_PERCPU_HASH); __type(key, int); __type(value, struct map_value); __uint(max_entries, 1); __uint(map_flags, BPF_F_NO_PREALLOC); } pcpu_hash_malloc_map SEC(".maps"); struct lru_hash_map { __uint(type, BPF_MAP_TYPE_LRU_HASH); __type(key, int); __type(value, struct map_value); __uint(max_entries, 1); } lru_hash_map SEC(".maps"); struct lru_pcpu_hash_map { __uint(type, BPF_MAP_TYPE_LRU_PERCPU_HASH); __type(key, int); __type(value, struct map_value); __uint(max_entries, 1); } lru_pcpu_hash_map SEC(".maps"); struct cgrp_ls_map { __uint(type, BPF_MAP_TYPE_CGRP_STORAGE); __uint(map_flags, BPF_F_NO_PREALLOC); __type(key, int); __type(value, struct map_value); } cgrp_ls_map SEC(".maps"); struct task_ls_map { __uint(type, BPF_MAP_TYPE_TASK_STORAGE); __uint(map_flags, BPF_F_NO_PREALLOC); __type(key, int); __type(value, struct map_value); } task_ls_map SEC(".maps"); struct inode_ls_map { __uint(type, BPF_MAP_TYPE_INODE_STORAGE); __uint(map_flags, BPF_F_NO_PREALLOC); __type(key, int); __type(value, struct map_value); } inode_ls_map SEC(".maps"); struct sk_ls_map { __uint(type, BPF_MAP_TYPE_SK_STORAGE); __uint(map_flags, BPF_F_NO_PREALLOC); __type(key, int); __type(value, struct map_value); } sk_ls_map SEC(".maps"); #define DEFINE_MAP_OF_MAP(map_type, inner_map_type, name) \ struct { \ __uint(type, map_type); \ __uint(max_entries, 1); \ __uint(key_size, sizeof(int)); \ __uint(value_size, sizeof(int)); \ __array(values, struct inner_map_type); \ } name SEC(".maps") = { \ .values = { [0] = &inner_map_type }, \ } DEFINE_MAP_OF_MAP(BPF_MAP_TYPE_ARRAY_OF_MAPS, array_map, array_of_array_maps); DEFINE_MAP_OF_MAP(BPF_MAP_TYPE_ARRAY_OF_MAPS, hash_map, array_of_hash_maps); DEFINE_MAP_OF_MAP(BPF_MAP_TYPE_ARRAY_OF_MAPS, hash_malloc_map, array_of_hash_malloc_maps); DEFINE_MAP_OF_MAP(BPF_MAP_TYPE_ARRAY_OF_MAPS, lru_hash_map, array_of_lru_hash_maps); DEFINE_MAP_OF_MAP(BPF_MAP_TYPE_ARRAY_OF_MAPS, pcpu_array_map, array_of_pcpu_array_maps); DEFINE_MAP_OF_MAP(BPF_MAP_TYPE_ARRAY_OF_MAPS, pcpu_hash_map, array_of_pcpu_hash_maps); DEFINE_MAP_OF_MAP(BPF_MAP_TYPE_HASH_OF_MAPS, array_map, hash_of_array_maps); DEFINE_MAP_OF_MAP(BPF_MAP_TYPE_HASH_OF_MAPS, hash_map, hash_of_hash_maps); DEFINE_MAP_OF_MAP(BPF_MAP_TYPE_HASH_OF_MAPS, hash_malloc_map, hash_of_hash_malloc_maps); DEFINE_MAP_OF_MAP(BPF_MAP_TYPE_HASH_OF_MAPS, lru_hash_map, hash_of_lru_hash_maps); DEFINE_MAP_OF_MAP(BPF_MAP_TYPE_HASH_OF_MAPS, pcpu_array_map, hash_of_pcpu_array_maps); DEFINE_MAP_OF_MAP(BPF_MAP_TYPE_HASH_OF_MAPS, pcpu_hash_map, hash_of_pcpu_hash_maps); #define WRITE_ONCE(x, val) ((*(volatile typeof(x) *) &(x)) = (val)) static void test_kptr_unref(struct map_value *v) { struct prog_test_ref_kfunc *p; p = v->unref_ptr; /* store untrusted_ptr_or_null_ */ WRITE_ONCE(v->unref_ptr, p); if (!p) return; if (p->a + p->b > 100) return; /* store untrusted_ptr_ */ WRITE_ONCE(v->unref_ptr, p); /* store NULL */ WRITE_ONCE(v->unref_ptr, NULL); } static void test_kptr_ref(struct map_value *v) { struct prog_test_ref_kfunc *p; p = v->ref_ptr; /* store ptr_or_null_ */ WRITE_ONCE(v->unref_ptr, p); if (!p) return; /* * p is rcu_ptr_prog_test_ref_kfunc, * because bpf prog is non-sleepable and runs in RCU CS. * p can be passed to kfunc that requires KF_RCU. */ bpf_kfunc_call_test_ref(p); if (p->a + p->b > 100) return; /* store NULL */ p = bpf_kptr_xchg(&v->ref_ptr, NULL); if (!p) return; /* * p is trusted_ptr_prog_test_ref_kfunc. * p can be passed to kfunc that requires KF_RCU. */ bpf_kfunc_call_test_ref(p); if (p->a + p->b > 100) { bpf_kfunc_call_test_release(p); return; } /* store ptr_ */ WRITE_ONCE(v->unref_ptr, p); bpf_kfunc_call_test_release(p); p = bpf_kfunc_call_test_acquire(&(unsigned long){0}); if (!p) return; /* store ptr_ */ p = bpf_kptr_xchg(&v->ref_ptr, p); if (!p) return; if (p->a + p->b > 100) { bpf_kfunc_call_test_release(p); return; } bpf_kfunc_call_test_release(p); } static void test_kptr(struct map_value *v) { test_kptr_unref(v); test_kptr_ref(v); } SEC("tc") int test_map_kptr(struct __sk_buff *ctx) { struct map_value *v; int key = 0; #define TEST(map) \ v = bpf_map_lookup_elem(&map, &key); \ if (!v) \ return 0; \ test_kptr(v) TEST(array_map); TEST(hash_map); TEST(hash_malloc_map); TEST(lru_hash_map); TEST(pcpu_array_map); TEST(pcpu_hash_map); #undef TEST return 0; } SEC("tp_btf/cgroup_mkdir") int BPF_PROG(test_cgrp_map_kptr, struct cgroup *cgrp, const char *path) { struct map_value *v; v = bpf_cgrp_storage_get(&cgrp_ls_map, cgrp, NULL, BPF_LOCAL_STORAGE_GET_F_CREATE); if (v) test_kptr(v); return 0; } SEC("lsm/inode_unlink") int BPF_PROG(test_task_map_kptr, struct inode *inode, struct dentry *victim) { struct task_struct *task; struct map_value *v; task = bpf_get_current_task_btf(); if (!task) return 0; v = bpf_task_storage_get(&task_ls_map, task, NULL, BPF_LOCAL_STORAGE_GET_F_CREATE); if (v) test_kptr(v); return 0; } SEC("lsm/inode_unlink") int BPF_PROG(test_inode_map_kptr, struct inode *inode, struct dentry *victim) { struct map_value *v; v = bpf_inode_storage_get(&inode_ls_map, inode, NULL, BPF_LOCAL_STORAGE_GET_F_CREATE); if (v) test_kptr(v); return 0; } SEC("tc") int test_sk_map_kptr(struct __sk_buff *ctx) { struct map_value *v; struct bpf_sock *sk; sk = ctx->sk; if (!sk) return 0; v = bpf_sk_storage_get(&sk_ls_map, sk, NULL, BPF_LOCAL_STORAGE_GET_F_CREATE); if (v) test_kptr(v); return 0; } SEC("tc") int test_map_in_map_kptr(struct __sk_buff *ctx) { struct map_value *v; int key = 0; void *map; #define TEST(map_in_map) \ map = bpf_map_lookup_elem(&map_in_map, &key); \ if (!map) \ return 0; \ v = bpf_map_lookup_elem(map, &key); \ if (!v) \ return 0; \ test_kptr(v) TEST(array_of_array_maps); TEST(array_of_hash_maps); TEST(array_of_hash_malloc_maps); TEST(array_of_lru_hash_maps); TEST(array_of_pcpu_array_maps); TEST(array_of_pcpu_hash_maps); TEST(hash_of_array_maps); TEST(hash_of_hash_maps); TEST(hash_of_hash_malloc_maps); TEST(hash_of_lru_hash_maps); TEST(hash_of_pcpu_array_maps); TEST(hash_of_pcpu_hash_maps); #undef TEST return 0; } int ref = 1; static __always_inline int test_map_kptr_ref_pre(struct map_value *v) { struct prog_test_ref_kfunc *p, *p_st; unsigned long arg = 0; int ret; p = bpf_kfunc_call_test_acquire(&arg); if (!p) return 1; ref++; p_st = p->next; if (p_st->cnt.refs.counter != ref) { ret = 2; goto end; } p = bpf_kptr_xchg(&v->ref_ptr, p); if (p) { ret = 3; goto end; } if (p_st->cnt.refs.counter != ref) return 4; p = bpf_kptr_xchg(&v->ref_ptr, NULL); if (!p) return 5; bpf_kfunc_call_test_release(p); ref--; if (p_st->cnt.refs.counter != ref) return 6; p = bpf_kfunc_call_test_acquire(&arg); if (!p) return 7; ref++; p = bpf_kptr_xchg(&v->ref_ptr, p); if (p) { ret = 8; goto end; } if (p_st->cnt.refs.counter != ref) return 9; /* Leave in map */ return 0; end: ref--; bpf_kfunc_call_test_release(p); return ret; } static __always_inline int test_map_kptr_ref_post(struct map_value *v) { struct prog_test_ref_kfunc *p, *p_st; p_st = v->ref_ptr; if (!p_st || p_st->cnt.refs.counter != ref) return 1; p = bpf_kptr_xchg(&v->ref_ptr, NULL); if (!p) return 2; if (p_st->cnt.refs.counter != ref) { bpf_kfunc_call_test_release(p); return 3; } p = bpf_kptr_xchg(&v->ref_ptr, p); if (p) { bpf_kfunc_call_test_release(p); return 4; } if (p_st->cnt.refs.counter != ref) return 5; return 0; } #define TEST(map) \ v = bpf_map_lookup_elem(&map, &key); \ if (!v) \ return -1; \ ret = test_map_kptr_ref_pre(v); \ if (ret) \ return ret; #define TEST_PCPU(map) \ v = bpf_map_lookup_percpu_elem(&map, &key, 0); \ if (!v) \ return -1; \ ret = test_map_kptr_ref_pre(v); \ if (ret) \ return ret; SEC("tc") int test_map_kptr_ref1(struct __sk_buff *ctx) { struct map_value *v, val = {}; int key = 0, ret; bpf_map_update_elem(&hash_map, &key, &val, 0); bpf_map_update_elem(&hash_malloc_map, &key, &val, 0); bpf_map_update_elem(&lru_hash_map, &key, &val, 0); bpf_map_update_elem(&pcpu_hash_map, &key, &val, 0); bpf_map_update_elem(&pcpu_hash_malloc_map, &key, &val, 0); bpf_map_update_elem(&lru_pcpu_hash_map, &key, &val, 0); TEST(array_map); TEST(hash_map); TEST(hash_malloc_map); TEST(lru_hash_map); TEST_PCPU(pcpu_array_map); TEST_PCPU(pcpu_hash_map); TEST_PCPU(pcpu_hash_malloc_map); TEST_PCPU(lru_pcpu_hash_map); return 0; } #undef TEST #undef TEST_PCPU #define TEST(map) \ v = bpf_map_lookup_elem(&map, &key); \ if (!v) \ return -1; \ ret = test_map_kptr_ref_post(v); \ if (ret) \ return ret; #define TEST_PCPU(map) \ v = bpf_map_lookup_percpu_elem(&map, &key, 0); \ if (!v) \ return -1; \ ret = test_map_kptr_ref_post(v); \ if (ret) \ return ret; SEC("tc") int test_map_kptr_ref2(struct __sk_buff *ctx) { struct map_value *v; int key = 0, ret; TEST(array_map); TEST(hash_map); TEST(hash_malloc_map); TEST(lru_hash_map); TEST_PCPU(pcpu_array_map); TEST_PCPU(pcpu_hash_map); TEST_PCPU(pcpu_hash_malloc_map); TEST_PCPU(lru_pcpu_hash_map); return 0; } #undef TEST #undef TEST_PCPU SEC("tc") int test_map_kptr_ref3(struct __sk_buff *ctx) { struct prog_test_ref_kfunc *p; unsigned long sp = 0; p = bpf_kfunc_call_test_acquire(&sp); if (!p) return 1; ref++; if (p->cnt.refs.counter != ref) { bpf_kfunc_call_test_release(p); return 2; } bpf_kfunc_call_test_release(p); ref--; return 0; } SEC("syscall") int test_ls_map_kptr_ref1(void *ctx) { struct task_struct *current; struct map_value *v; current = bpf_get_current_task_btf(); if (!current) return 100; v = bpf_task_storage_get(&task_ls_map, current, NULL, 0); if (v) return 150; v = bpf_task_storage_get(&task_ls_map, current, NULL, BPF_LOCAL_STORAGE_GET_F_CREATE); if (!v) return 200; return test_map_kptr_ref_pre(v); } SEC("syscall") int test_ls_map_kptr_ref2(void *ctx) { struct task_struct *current; struct map_value *v; current = bpf_get_current_task_btf(); if (!current) return 100; v = bpf_task_storage_get(&task_ls_map, current, NULL, 0); if (!v) return 200; return test_map_kptr_ref_post(v); } SEC("syscall") int test_ls_map_kptr_ref_del(void *ctx) { struct task_struct *current; struct map_value *v; current = bpf_get_current_task_btf(); if (!current) return 100; v = bpf_task_storage_get(&task_ls_map, current, NULL, 0); if (!v) return 200; if (!v->ref_ptr) return 300; return bpf_task_storage_delete(&task_ls_map, current); } char _license[] SEC("license") = "GPL";
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1