Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Andrii Nakryiko | 378 | 29.30% | 14 | 21.21% |
Yonghong Song | 194 | 15.04% | 5 | 7.58% |
Song Liu | 87 | 6.74% | 2 | 3.03% |
Stanislav Fomichev | 82 | 6.36% | 3 | 4.55% |
Alexei Starovoitov | 78 | 6.05% | 7 | 10.61% |
Wang Nan | 72 | 5.58% | 3 | 4.55% |
Martin KaFai Lau | 58 | 4.50% | 3 | 4.55% |
Andrey Ignatov | 46 | 3.57% | 4 | 6.06% |
Lorenz Bauer | 36 | 2.79% | 1 | 1.52% |
Joe Burton | 36 | 2.79% | 1 | 1.52% |
YiFei Zhu | 33 | 2.56% | 1 | 1.52% |
Jiri Olsa | 29 | 2.25% | 1 | 1.52% |
Joe Stringer | 26 | 2.02% | 2 | 3.03% |
Denis Salopek | 20 | 1.55% | 1 | 1.52% |
Christy Lee | 19 | 1.47% | 1 | 1.52% |
Mauricio Vasquez B | 17 | 1.32% | 1 | 1.52% |
He Kuang | 17 | 1.32% | 1 | 1.52% |
Quentin Monnet | 12 | 0.93% | 1 | 1.52% |
Mickaël Salaün | 8 | 0.62% | 5 | 7.58% |
Daniel Borkmann | 8 | 0.62% | 1 | 1.52% |
Kui-Feng Lee | 8 | 0.62% | 1 | 1.52% |
Grant Seltzer | 7 | 0.54% | 1 | 1.52% |
Toke Höiland-Jörgensen | 7 | 0.54% | 2 | 3.03% |
John Fastabend | 5 | 0.39% | 1 | 1.52% |
Björn Töpel | 3 | 0.23% | 1 | 1.52% |
Daniel T. Lee | 3 | 0.23% | 1 | 1.52% |
McCabe, Robert J | 1 | 0.08% | 1 | 1.52% |
Total | 1290 | 66 |
123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438
/* SPDX-License-Identifier: (LGPL-2.1 OR BSD-2-Clause) */ /* * common eBPF ELF operations. * * Copyright (C) 2013-2015 Alexei Starovoitov <ast@kernel.org> * Copyright (C) 2015 Wang Nan <wangnan0@huawei.com> * Copyright (C) 2015 Huawei Inc. * * This program is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; * version 2.1 of the License (not later!) * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with this program; if not, see <http://www.gnu.org/licenses> */ #ifndef __LIBBPF_BPF_H #define __LIBBPF_BPF_H #include <linux/bpf.h> #include <stdbool.h> #include <stddef.h> #include <stdint.h> #include "libbpf_common.h" #include "libbpf_legacy.h" #ifdef __cplusplus extern "C" { #endif int libbpf_set_memlock_rlim(size_t memlock_bytes); struct bpf_map_create_opts { size_t sz; /* size of this struct for forward/backward compatibility */ __u32 btf_fd; __u32 btf_key_type_id; __u32 btf_value_type_id; __u32 btf_vmlinux_value_type_id; __u32 inner_map_fd; __u32 map_flags; __u64 map_extra; __u32 numa_node; __u32 map_ifindex; }; #define bpf_map_create_opts__last_field map_ifindex LIBBPF_API int bpf_map_create(enum bpf_map_type map_type, const char *map_name, __u32 key_size, __u32 value_size, __u32 max_entries, const struct bpf_map_create_opts *opts); struct bpf_prog_load_opts { size_t sz; /* size of this struct for forward/backward compatibility */ /* libbpf can retry BPF_PROG_LOAD command if bpf() syscall returns * -EAGAIN. This field determines how many attempts libbpf has to * make. If not specified, libbpf will use default value of 5. */ int attempts; enum bpf_attach_type expected_attach_type; __u32 prog_btf_fd; __u32 prog_flags; __u32 prog_ifindex; __u32 kern_version; __u32 attach_btf_id; __u32 attach_prog_fd; __u32 attach_btf_obj_fd; const int *fd_array; /* .BTF.ext func info data */ const void *func_info; __u32 func_info_cnt; __u32 func_info_rec_size; /* .BTF.ext line info data */ const void *line_info; __u32 line_info_cnt; __u32 line_info_rec_size; /* verifier log options */ __u32 log_level; __u32 log_size; char *log_buf; }; #define bpf_prog_load_opts__last_field log_buf LIBBPF_API int bpf_prog_load(enum bpf_prog_type prog_type, const char *prog_name, const char *license, const struct bpf_insn *insns, size_t insn_cnt, const struct bpf_prog_load_opts *opts); /* Flags to direct loading requirements */ #define MAPS_RELAX_COMPAT 0x01 /* Recommended log buffer size */ #define BPF_LOG_BUF_SIZE (UINT32_MAX >> 8) /* verifier maximum in kernels <= 5.1 */ struct bpf_btf_load_opts { size_t sz; /* size of this struct for forward/backward compatibility */ /* kernel log options */ char *log_buf; __u32 log_level; __u32 log_size; }; #define bpf_btf_load_opts__last_field log_size LIBBPF_API int bpf_btf_load(const void *btf_data, size_t btf_size, const struct bpf_btf_load_opts *opts); LIBBPF_API int bpf_map_update_elem(int fd, const void *key, const void *value, __u64 flags); LIBBPF_API int bpf_map_lookup_elem(int fd, const void *key, void *value); LIBBPF_API int bpf_map_lookup_elem_flags(int fd, const void *key, void *value, __u64 flags); LIBBPF_API int bpf_map_lookup_and_delete_elem(int fd, const void *key, void *value); LIBBPF_API int bpf_map_lookup_and_delete_elem_flags(int fd, const void *key, void *value, __u64 flags); LIBBPF_API int bpf_map_delete_elem(int fd, const void *key); LIBBPF_API int bpf_map_delete_elem_flags(int fd, const void *key, __u64 flags); LIBBPF_API int bpf_map_get_next_key(int fd, const void *key, void *next_key); LIBBPF_API int bpf_map_freeze(int fd); struct bpf_map_batch_opts { size_t sz; /* size of this struct for forward/backward compatibility */ __u64 elem_flags; __u64 flags; }; #define bpf_map_batch_opts__last_field flags /** * @brief **bpf_map_delete_batch()** allows for batch deletion of multiple * elements in a BPF map. * * @param fd BPF map file descriptor * @param keys pointer to an array of *count* keys * @param count input and output parameter; on input **count** represents the * number of elements in the map to delete in batch; * on output if a non-EFAULT error is returned, **count** represents the number of deleted * elements if the output **count** value is not equal to the input **count** value * If EFAULT is returned, **count** should not be trusted to be correct. * @param opts options for configuring the way the batch deletion works * @return 0, on success; negative error code, otherwise (errno is also set to * the error code) */ LIBBPF_API int bpf_map_delete_batch(int fd, const void *keys, __u32 *count, const struct bpf_map_batch_opts *opts); /** * @brief **bpf_map_lookup_batch()** allows for batch lookup of BPF map elements. * * The parameter *in_batch* is the address of the first element in the batch to read. * *out_batch* is an output parameter that should be passed as *in_batch* to subsequent * calls to **bpf_map_lookup_batch()**. NULL can be passed for *in_batch* to indicate * that the batched lookup starts from the beginning of the map. * * The *keys* and *values* are output parameters which must point to memory large enough to * hold *count* items based on the key and value size of the map *map_fd*. The *keys* * buffer must be of *key_size* * *count*. The *values* buffer must be of * *value_size* * *count*. * * @param fd BPF map file descriptor * @param in_batch address of the first element in batch to read, can pass NULL to * indicate that the batched lookup starts from the beginning of the map. * @param out_batch output parameter that should be passed to next call as *in_batch* * @param keys pointer to an array large enough for *count* keys * @param values pointer to an array large enough for *count* values * @param count input and output parameter; on input it's the number of elements * in the map to read in batch; on output it's the number of elements that were * successfully read. * If a non-EFAULT error is returned, count will be set as the number of elements * that were read before the error occurred. * If EFAULT is returned, **count** should not be trusted to be correct. * @param opts options for configuring the way the batch lookup works * @return 0, on success; negative error code, otherwise (errno is also set to * the error code) */ LIBBPF_API int bpf_map_lookup_batch(int fd, void *in_batch, void *out_batch, void *keys, void *values, __u32 *count, const struct bpf_map_batch_opts *opts); /** * @brief **bpf_map_lookup_and_delete_batch()** allows for batch lookup and deletion * of BPF map elements where each element is deleted after being retrieved. * * @param fd BPF map file descriptor * @param in_batch address of the first element in batch to read, can pass NULL to * get address of the first element in *out_batch* * @param out_batch output parameter that should be passed to next call as *in_batch* * @param keys pointer to an array of *count* keys * @param values pointer to an array large enough for *count* values * @param count input and output parameter; on input it's the number of elements * in the map to read and delete in batch; on output it represents the number of * elements that were successfully read and deleted * If a non-**EFAULT** error code is returned and if the output **count** value * is not equal to the input **count** value, up to **count** elements may * have been deleted. * if **EFAULT** is returned up to *count* elements may have been deleted without * being returned via the *keys* and *values* output parameters. * @param opts options for configuring the way the batch lookup and delete works * @return 0, on success; negative error code, otherwise (errno is also set to * the error code) */ LIBBPF_API int bpf_map_lookup_and_delete_batch(int fd, void *in_batch, void *out_batch, void *keys, void *values, __u32 *count, const struct bpf_map_batch_opts *opts); /** * @brief **bpf_map_update_batch()** updates multiple elements in a map * by specifying keys and their corresponding values. * * The *keys* and *values* parameters must point to memory large enough * to hold *count* items based on the key and value size of the map. * * The *opts* parameter can be used to control how *bpf_map_update_batch()* * should handle keys that either do or do not already exist in the map. * In particular the *flags* parameter of *bpf_map_batch_opts* can be * one of the following: * * Note that *count* is an input and output parameter, where on output it * represents how many elements were successfully updated. Also note that if * **EFAULT** then *count* should not be trusted to be correct. * * **BPF_ANY** * Create new elements or update existing. * * **BPF_NOEXIST** * Create new elements only if they do not exist. * * **BPF_EXIST** * Update existing elements. * * **BPF_F_LOCK** * Update spin_lock-ed map elements. This must be * specified if the map value contains a spinlock. * * @param fd BPF map file descriptor * @param keys pointer to an array of *count* keys * @param values pointer to an array of *count* values * @param count input and output parameter; on input it's the number of elements * in the map to update in batch; on output if a non-EFAULT error is returned, * **count** represents the number of updated elements if the output **count** * value is not equal to the input **count** value. * If EFAULT is returned, **count** should not be trusted to be correct. * @param opts options for configuring the way the batch update works * @return 0, on success; negative error code, otherwise (errno is also set to * the error code) */ LIBBPF_API int bpf_map_update_batch(int fd, const void *keys, const void *values, __u32 *count, const struct bpf_map_batch_opts *opts); struct bpf_obj_get_opts { size_t sz; /* size of this struct for forward/backward compatibility */ __u32 file_flags; size_t :0; }; #define bpf_obj_get_opts__last_field file_flags LIBBPF_API int bpf_obj_pin(int fd, const char *pathname); LIBBPF_API int bpf_obj_get(const char *pathname); LIBBPF_API int bpf_obj_get_opts(const char *pathname, const struct bpf_obj_get_opts *opts); struct bpf_prog_attach_opts { size_t sz; /* size of this struct for forward/backward compatibility */ unsigned int flags; int replace_prog_fd; }; #define bpf_prog_attach_opts__last_field replace_prog_fd LIBBPF_API int bpf_prog_attach(int prog_fd, int attachable_fd, enum bpf_attach_type type, unsigned int flags); LIBBPF_API int bpf_prog_attach_opts(int prog_fd, int attachable_fd, enum bpf_attach_type type, const struct bpf_prog_attach_opts *opts); LIBBPF_API int bpf_prog_detach(int attachable_fd, enum bpf_attach_type type); LIBBPF_API int bpf_prog_detach2(int prog_fd, int attachable_fd, enum bpf_attach_type type); union bpf_iter_link_info; /* defined in up-to-date linux/bpf.h */ struct bpf_link_create_opts { size_t sz; /* size of this struct for forward/backward compatibility */ __u32 flags; union bpf_iter_link_info *iter_info; __u32 iter_info_len; __u32 target_btf_id; union { struct { __u64 bpf_cookie; } perf_event; struct { __u32 flags; __u32 cnt; const char **syms; const unsigned long *addrs; const __u64 *cookies; } kprobe_multi; struct { __u64 cookie; } tracing; }; size_t :0; }; #define bpf_link_create_opts__last_field kprobe_multi.cookies LIBBPF_API int bpf_link_create(int prog_fd, int target_fd, enum bpf_attach_type attach_type, const struct bpf_link_create_opts *opts); LIBBPF_API int bpf_link_detach(int link_fd); struct bpf_link_update_opts { size_t sz; /* size of this struct for forward/backward compatibility */ __u32 flags; /* extra flags */ __u32 old_prog_fd; /* expected old program FD */ }; #define bpf_link_update_opts__last_field old_prog_fd LIBBPF_API int bpf_link_update(int link_fd, int new_prog_fd, const struct bpf_link_update_opts *opts); LIBBPF_API int bpf_iter_create(int link_fd); struct bpf_prog_test_run_attr { int prog_fd; int repeat; const void *data_in; __u32 data_size_in; void *data_out; /* optional */ __u32 data_size_out; /* in: max length of data_out * out: length of data_out */ __u32 retval; /* out: return code of the BPF program */ __u32 duration; /* out: average per repetition in ns */ const void *ctx_in; /* optional */ __u32 ctx_size_in; void *ctx_out; /* optional */ __u32 ctx_size_out; /* in: max length of ctx_out * out: length of cxt_out */ }; LIBBPF_API int bpf_prog_get_next_id(__u32 start_id, __u32 *next_id); LIBBPF_API int bpf_map_get_next_id(__u32 start_id, __u32 *next_id); LIBBPF_API int bpf_btf_get_next_id(__u32 start_id, __u32 *next_id); LIBBPF_API int bpf_link_get_next_id(__u32 start_id, __u32 *next_id); LIBBPF_API int bpf_prog_get_fd_by_id(__u32 id); LIBBPF_API int bpf_map_get_fd_by_id(__u32 id); LIBBPF_API int bpf_btf_get_fd_by_id(__u32 id); LIBBPF_API int bpf_link_get_fd_by_id(__u32 id); LIBBPF_API int bpf_obj_get_info_by_fd(int bpf_fd, void *info, __u32 *info_len); struct bpf_prog_query_opts { size_t sz; /* size of this struct for forward/backward compatibility */ __u32 query_flags; __u32 attach_flags; /* output argument */ __u32 *prog_ids; __u32 prog_cnt; /* input+output argument */ __u32 *prog_attach_flags; }; #define bpf_prog_query_opts__last_field prog_attach_flags LIBBPF_API int bpf_prog_query_opts(int target_fd, enum bpf_attach_type type, struct bpf_prog_query_opts *opts); LIBBPF_API int bpf_prog_query(int target_fd, enum bpf_attach_type type, __u32 query_flags, __u32 *attach_flags, __u32 *prog_ids, __u32 *prog_cnt); LIBBPF_API int bpf_raw_tracepoint_open(const char *name, int prog_fd); LIBBPF_API int bpf_task_fd_query(int pid, int fd, __u32 flags, char *buf, __u32 *buf_len, __u32 *prog_id, __u32 *fd_type, __u64 *probe_offset, __u64 *probe_addr); enum bpf_stats_type; /* defined in up-to-date linux/bpf.h */ LIBBPF_API int bpf_enable_stats(enum bpf_stats_type type); struct bpf_prog_bind_opts { size_t sz; /* size of this struct for forward/backward compatibility */ __u32 flags; }; #define bpf_prog_bind_opts__last_field flags LIBBPF_API int bpf_prog_bind_map(int prog_fd, int map_fd, const struct bpf_prog_bind_opts *opts); struct bpf_test_run_opts { size_t sz; /* size of this struct for forward/backward compatibility */ const void *data_in; /* optional */ void *data_out; /* optional */ __u32 data_size_in; __u32 data_size_out; /* in: max length of data_out * out: length of data_out */ const void *ctx_in; /* optional */ void *ctx_out; /* optional */ __u32 ctx_size_in; __u32 ctx_size_out; /* in: max length of ctx_out * out: length of cxt_out */ __u32 retval; /* out: return code of the BPF program */ int repeat; __u32 duration; /* out: average per repetition in ns */ __u32 flags; __u32 cpu; __u32 batch_size; }; #define bpf_test_run_opts__last_field batch_size LIBBPF_API int bpf_prog_test_run_opts(int prog_fd, struct bpf_test_run_opts *opts); #ifdef __cplusplus } /* extern "C" */ #endif #endif /* __LIBBPF_BPF_H */
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1