Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Eduard Zingerman | 433 | 98.86% | 1 | 33.33% |
Andrei Matei | 5 | 1.14% | 2 | 66.67% |
Total | 438 | 3 |
// SPDX-License-Identifier: GPL-2.0 /* Converted from tools/testing/selftests/bpf/verifier/raw_stack.c */ #include <linux/bpf.h> #include <bpf/bpf_helpers.h> #include "bpf_misc.h" SEC("socket") __description("raw_stack: no skb_load_bytes") __success __failure_unpriv __msg_unpriv("invalid read from stack R6 off=-8 size=8") __naked void stack_no_skb_load_bytes(void) { asm volatile (" \ r2 = 4; \ r6 = r10; \ r6 += -8; \ r3 = r6; \ r4 = 8; \ /* Call to skb_load_bytes() omitted. */ \ r0 = *(u64*)(r6 + 0); \ exit; \ " ::: __clobber_all); } SEC("tc") __description("raw_stack: skb_load_bytes, negative len") __failure __msg("R4 min value is negative") __naked void skb_load_bytes_negative_len(void) { asm volatile (" \ r2 = 4; \ r6 = r10; \ r6 += -8; \ r3 = r6; \ r4 = -8; \ call %[bpf_skb_load_bytes]; \ r0 = *(u64*)(r6 + 0); \ exit; \ " : : __imm(bpf_skb_load_bytes) : __clobber_all); } SEC("tc") __description("raw_stack: skb_load_bytes, negative len 2") __failure __msg("R4 min value is negative") __naked void load_bytes_negative_len_2(void) { asm volatile (" \ r2 = 4; \ r6 = r10; \ r6 += -8; \ r3 = r6; \ r4 = %[__imm_0]; \ call %[bpf_skb_load_bytes]; \ r0 = *(u64*)(r6 + 0); \ exit; \ " : : __imm(bpf_skb_load_bytes), __imm_const(__imm_0, ~0) : __clobber_all); } SEC("tc") __description("raw_stack: skb_load_bytes, zero len") __failure __msg("R4 invalid zero-sized read: u64=[0,0]") __naked void skb_load_bytes_zero_len(void) { asm volatile (" \ r2 = 4; \ r6 = r10; \ r6 += -8; \ r3 = r6; \ r4 = 0; \ call %[bpf_skb_load_bytes]; \ r0 = *(u64*)(r6 + 0); \ exit; \ " : : __imm(bpf_skb_load_bytes) : __clobber_all); } SEC("tc") __description("raw_stack: skb_load_bytes, no init") __success __retval(0) __naked void skb_load_bytes_no_init(void) { asm volatile (" \ r2 = 4; \ r6 = r10; \ r6 += -8; \ r3 = r6; \ r4 = 8; \ call %[bpf_skb_load_bytes]; \ r0 = *(u64*)(r6 + 0); \ exit; \ " : : __imm(bpf_skb_load_bytes) : __clobber_all); } SEC("tc") __description("raw_stack: skb_load_bytes, init") __success __retval(0) __naked void stack_skb_load_bytes_init(void) { asm volatile (" \ r2 = 4; \ r6 = r10; \ r6 += -8; \ r3 = 0xcafe; \ *(u64*)(r6 + 0) = r3; \ r3 = r6; \ r4 = 8; \ call %[bpf_skb_load_bytes]; \ r0 = *(u64*)(r6 + 0); \ exit; \ " : : __imm(bpf_skb_load_bytes) : __clobber_all); } SEC("tc") __description("raw_stack: skb_load_bytes, spilled regs around bounds") __success __retval(0) __naked void bytes_spilled_regs_around_bounds(void) { asm volatile (" \ r2 = 4; \ r6 = r10; \ r6 += -16; \ *(u64*)(r6 - 8) = r1; \ *(u64*)(r6 + 8) = r1; \ r3 = r6; \ r4 = 8; \ call %[bpf_skb_load_bytes]; \ r0 = *(u64*)(r6 - 8); \ r2 = *(u64*)(r6 + 8); \ r0 = *(u32*)(r0 + %[__sk_buff_mark]); \ r2 = *(u32*)(r2 + %[__sk_buff_priority]); \ r0 += r2; \ exit; \ " : : __imm(bpf_skb_load_bytes), __imm_const(__sk_buff_mark, offsetof(struct __sk_buff, mark)), __imm_const(__sk_buff_priority, offsetof(struct __sk_buff, priority)) : __clobber_all); } SEC("tc") __description("raw_stack: skb_load_bytes, spilled regs corruption") __failure __msg("R0 invalid mem access 'scalar'") __flag(BPF_F_ANY_ALIGNMENT) __naked void load_bytes_spilled_regs_corruption(void) { asm volatile (" \ r2 = 4; \ r6 = r10; \ r6 += -8; \ *(u64*)(r6 + 0) = r1; \ r3 = r6; \ r4 = 8; \ call %[bpf_skb_load_bytes]; \ r0 = *(u64*)(r6 + 0); \ r0 = *(u32*)(r0 + %[__sk_buff_mark]); \ exit; \ " : : __imm(bpf_skb_load_bytes), __imm_const(__sk_buff_mark, offsetof(struct __sk_buff, mark)) : __clobber_all); } SEC("tc") __description("raw_stack: skb_load_bytes, spilled regs corruption 2") __failure __msg("R3 invalid mem access 'scalar'") __flag(BPF_F_ANY_ALIGNMENT) __naked void bytes_spilled_regs_corruption_2(void) { asm volatile (" \ r2 = 4; \ r6 = r10; \ r6 += -16; \ *(u64*)(r6 - 8) = r1; \ *(u64*)(r6 + 0) = r1; \ *(u64*)(r6 + 8) = r1; \ r3 = r6; \ r4 = 8; \ call %[bpf_skb_load_bytes]; \ r0 = *(u64*)(r6 - 8); \ r2 = *(u64*)(r6 + 8); \ r3 = *(u64*)(r6 + 0); \ r0 = *(u32*)(r0 + %[__sk_buff_mark]); \ r2 = *(u32*)(r2 + %[__sk_buff_priority]); \ r0 += r2; \ r3 = *(u32*)(r3 + %[__sk_buff_pkt_type]); \ r0 += r3; \ exit; \ " : : __imm(bpf_skb_load_bytes), __imm_const(__sk_buff_mark, offsetof(struct __sk_buff, mark)), __imm_const(__sk_buff_pkt_type, offsetof(struct __sk_buff, pkt_type)), __imm_const(__sk_buff_priority, offsetof(struct __sk_buff, priority)) : __clobber_all); } SEC("tc") __description("raw_stack: skb_load_bytes, spilled regs + data") __success __retval(0) __naked void load_bytes_spilled_regs_data(void) { asm volatile (" \ r2 = 4; \ r6 = r10; \ r6 += -16; \ *(u64*)(r6 - 8) = r1; \ *(u64*)(r6 + 0) = r1; \ *(u64*)(r6 + 8) = r1; \ r3 = r6; \ r4 = 8; \ call %[bpf_skb_load_bytes]; \ r0 = *(u64*)(r6 - 8); \ r2 = *(u64*)(r6 + 8); \ r3 = *(u64*)(r6 + 0); \ r0 = *(u32*)(r0 + %[__sk_buff_mark]); \ r2 = *(u32*)(r2 + %[__sk_buff_priority]); \ r0 += r2; \ r0 += r3; \ exit; \ " : : __imm(bpf_skb_load_bytes), __imm_const(__sk_buff_mark, offsetof(struct __sk_buff, mark)), __imm_const(__sk_buff_priority, offsetof(struct __sk_buff, priority)) : __clobber_all); } SEC("tc") __description("raw_stack: skb_load_bytes, invalid access 1") __failure __msg("invalid indirect access to stack R3 off=-513 size=8") __naked void load_bytes_invalid_access_1(void) { asm volatile (" \ r2 = 4; \ r6 = r10; \ r6 += -513; \ r3 = r6; \ r4 = 8; \ call %[bpf_skb_load_bytes]; \ r0 = *(u64*)(r6 + 0); \ exit; \ " : : __imm(bpf_skb_load_bytes) : __clobber_all); } SEC("tc") __description("raw_stack: skb_load_bytes, invalid access 2") __failure __msg("invalid indirect access to stack R3 off=-1 size=8") __naked void load_bytes_invalid_access_2(void) { asm volatile (" \ r2 = 4; \ r6 = r10; \ r6 += -1; \ r3 = r6; \ r4 = 8; \ call %[bpf_skb_load_bytes]; \ r0 = *(u64*)(r6 + 0); \ exit; \ " : : __imm(bpf_skb_load_bytes) : __clobber_all); } SEC("tc") __description("raw_stack: skb_load_bytes, invalid access 3") __failure __msg("R4 min value is negative") __naked void load_bytes_invalid_access_3(void) { asm volatile (" \ r2 = 4; \ r6 = r10; \ r6 += 0xffffffff; \ r3 = r6; \ r4 = 0xffffffff; \ call %[bpf_skb_load_bytes]; \ r0 = *(u64*)(r6 + 0); \ exit; \ " : : __imm(bpf_skb_load_bytes) : __clobber_all); } SEC("tc") __description("raw_stack: skb_load_bytes, invalid access 4") __failure __msg("R4 unbounded memory access, use 'var &= const' or 'if (var < const)'") __naked void load_bytes_invalid_access_4(void) { asm volatile (" \ r2 = 4; \ r6 = r10; \ r6 += -1; \ r3 = r6; \ r4 = 0x7fffffff; \ call %[bpf_skb_load_bytes]; \ r0 = *(u64*)(r6 + 0); \ exit; \ " : : __imm(bpf_skb_load_bytes) : __clobber_all); } SEC("tc") __description("raw_stack: skb_load_bytes, invalid access 5") __failure __msg("R4 unbounded memory access, use 'var &= const' or 'if (var < const)'") __naked void load_bytes_invalid_access_5(void) { asm volatile (" \ r2 = 4; \ r6 = r10; \ r6 += -512; \ r3 = r6; \ r4 = 0x7fffffff; \ call %[bpf_skb_load_bytes]; \ r0 = *(u64*)(r6 + 0); \ exit; \ " : : __imm(bpf_skb_load_bytes) : __clobber_all); } SEC("tc") __description("raw_stack: skb_load_bytes, invalid access 6") __failure __msg("invalid zero-sized read") __naked void load_bytes_invalid_access_6(void) { asm volatile (" \ r2 = 4; \ r6 = r10; \ r6 += -512; \ r3 = r6; \ r4 = 0; \ call %[bpf_skb_load_bytes]; \ r0 = *(u64*)(r6 + 0); \ exit; \ " : : __imm(bpf_skb_load_bytes) : __clobber_all); } SEC("tc") __description("raw_stack: skb_load_bytes, large access") __success __retval(0) __naked void skb_load_bytes_large_access(void) { asm volatile (" \ r2 = 4; \ r6 = r10; \ r6 += -512; \ r3 = r6; \ r4 = 512; \ call %[bpf_skb_load_bytes]; \ r0 = *(u64*)(r6 + 0); \ exit; \ " : : __imm(bpf_skb_load_bytes) : __clobber_all); } char _license[] SEC("license") = "GPL";
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1