mirror of
https://github.com/torvalds/linux.git
synced 2024-11-15 00:21:59 +00:00
selftests/bpf: verifier/cgroup_skb.c converted to inline assembly
Test verifier/cgroup_skb.c automatically converted to use inline assembly. Signed-off-by: Eduard Zingerman <eddyz87@gmail.com> Link: https://lore.kernel.org/r/20230325025524.144043-14-eddyz87@gmail.com Signed-off-by: Alexei Starovoitov <ast@kernel.org>
This commit is contained in:
parent
047687a7f4
commit
b1b6372535
@ -10,6 +10,7 @@
|
||||
#include "verifier_bounds_mix_sign_unsign.skel.h"
|
||||
#include "verifier_cfg.skel.h"
|
||||
#include "verifier_cgroup_inv_retcode.skel.h"
|
||||
#include "verifier_cgroup_skb.skel.h"
|
||||
|
||||
__maybe_unused
|
||||
static void run_tests_aux(const char *skel_name, skel_elf_bytes_fn elf_bytes_factory)
|
||||
@ -42,3 +43,4 @@ void test_verifier_bounds_deduction(void) { RUN(verifier_bounds_deduction);
|
||||
void test_verifier_bounds_mix_sign_unsign(void) { RUN(verifier_bounds_mix_sign_unsign); }
|
||||
void test_verifier_cfg(void) { RUN(verifier_cfg); }
|
||||
void test_verifier_cgroup_inv_retcode(void) { RUN(verifier_cgroup_inv_retcode); }
|
||||
void test_verifier_cgroup_skb(void) { RUN(verifier_cgroup_skb); }
|
||||
|
227
tools/testing/selftests/bpf/progs/verifier_cgroup_skb.c
Normal file
227
tools/testing/selftests/bpf/progs/verifier_cgroup_skb.c
Normal file
@ -0,0 +1,227 @@
|
||||
// SPDX-License-Identifier: GPL-2.0
|
||||
/* Converted from tools/testing/selftests/bpf/verifier/cgroup_skb.c */
|
||||
|
||||
#include <linux/bpf.h>
|
||||
#include <bpf/bpf_helpers.h>
|
||||
#include "bpf_misc.h"
|
||||
|
||||
SEC("cgroup/skb")
|
||||
__description("direct packet read test#1 for CGROUP_SKB")
|
||||
__success __failure_unpriv
|
||||
__msg_unpriv("invalid bpf_context access off=76 size=4")
|
||||
__retval(0)
|
||||
__naked void test_1_for_cgroup_skb(void)
|
||||
{
|
||||
asm volatile (" \
|
||||
r2 = *(u32*)(r1 + %[__sk_buff_data]); \
|
||||
r3 = *(u32*)(r1 + %[__sk_buff_data_end]); \
|
||||
r4 = *(u32*)(r1 + %[__sk_buff_len]); \
|
||||
r5 = *(u32*)(r1 + %[__sk_buff_pkt_type]); \
|
||||
r6 = *(u32*)(r1 + %[__sk_buff_mark]); \
|
||||
*(u32*)(r1 + %[__sk_buff_mark]) = r6; \
|
||||
r7 = *(u32*)(r1 + %[__sk_buff_queue_mapping]); \
|
||||
r8 = *(u32*)(r1 + %[__sk_buff_protocol]); \
|
||||
r9 = *(u32*)(r1 + %[__sk_buff_vlan_present]); \
|
||||
r0 = r2; \
|
||||
r0 += 8; \
|
||||
if r0 > r3 goto l0_%=; \
|
||||
r0 = *(u8*)(r2 + 0); \
|
||||
l0_%=: r0 = 0; \
|
||||
exit; \
|
||||
" :
|
||||
: __imm_const(__sk_buff_data, offsetof(struct __sk_buff, data)),
|
||||
__imm_const(__sk_buff_data_end, offsetof(struct __sk_buff, data_end)),
|
||||
__imm_const(__sk_buff_len, offsetof(struct __sk_buff, len)),
|
||||
__imm_const(__sk_buff_mark, offsetof(struct __sk_buff, mark)),
|
||||
__imm_const(__sk_buff_pkt_type, offsetof(struct __sk_buff, pkt_type)),
|
||||
__imm_const(__sk_buff_protocol, offsetof(struct __sk_buff, protocol)),
|
||||
__imm_const(__sk_buff_queue_mapping, offsetof(struct __sk_buff, queue_mapping)),
|
||||
__imm_const(__sk_buff_vlan_present, offsetof(struct __sk_buff, vlan_present))
|
||||
: __clobber_all);
|
||||
}
|
||||
|
||||
SEC("cgroup/skb")
|
||||
__description("direct packet read test#2 for CGROUP_SKB")
|
||||
__success __success_unpriv __retval(0)
|
||||
__naked void test_2_for_cgroup_skb(void)
|
||||
{
|
||||
asm volatile (" \
|
||||
r4 = *(u32*)(r1 + %[__sk_buff_vlan_tci]); \
|
||||
r5 = *(u32*)(r1 + %[__sk_buff_vlan_proto]); \
|
||||
r6 = *(u32*)(r1 + %[__sk_buff_priority]); \
|
||||
*(u32*)(r1 + %[__sk_buff_priority]) = r6; \
|
||||
r7 = *(u32*)(r1 + %[__sk_buff_ingress_ifindex]);\
|
||||
r8 = *(u32*)(r1 + %[__sk_buff_tc_index]); \
|
||||
r9 = *(u32*)(r1 + %[__sk_buff_hash]); \
|
||||
r0 = 0; \
|
||||
exit; \
|
||||
" :
|
||||
: __imm_const(__sk_buff_hash, offsetof(struct __sk_buff, hash)),
|
||||
__imm_const(__sk_buff_ingress_ifindex, offsetof(struct __sk_buff, ingress_ifindex)),
|
||||
__imm_const(__sk_buff_priority, offsetof(struct __sk_buff, priority)),
|
||||
__imm_const(__sk_buff_tc_index, offsetof(struct __sk_buff, tc_index)),
|
||||
__imm_const(__sk_buff_vlan_proto, offsetof(struct __sk_buff, vlan_proto)),
|
||||
__imm_const(__sk_buff_vlan_tci, offsetof(struct __sk_buff, vlan_tci))
|
||||
: __clobber_all);
|
||||
}
|
||||
|
||||
SEC("cgroup/skb")
|
||||
__description("direct packet read test#3 for CGROUP_SKB")
|
||||
__success __success_unpriv __retval(0)
|
||||
__naked void test_3_for_cgroup_skb(void)
|
||||
{
|
||||
asm volatile (" \
|
||||
r4 = *(u32*)(r1 + %[__sk_buff_cb_0]); \
|
||||
r5 = *(u32*)(r1 + %[__sk_buff_cb_1]); \
|
||||
r6 = *(u32*)(r1 + %[__sk_buff_cb_2]); \
|
||||
r7 = *(u32*)(r1 + %[__sk_buff_cb_3]); \
|
||||
r8 = *(u32*)(r1 + %[__sk_buff_cb_4]); \
|
||||
r9 = *(u32*)(r1 + %[__sk_buff_napi_id]); \
|
||||
*(u32*)(r1 + %[__sk_buff_cb_0]) = r4; \
|
||||
*(u32*)(r1 + %[__sk_buff_cb_1]) = r5; \
|
||||
*(u32*)(r1 + %[__sk_buff_cb_2]) = r6; \
|
||||
*(u32*)(r1 + %[__sk_buff_cb_3]) = r7; \
|
||||
*(u32*)(r1 + %[__sk_buff_cb_4]) = r8; \
|
||||
r0 = 0; \
|
||||
exit; \
|
||||
" :
|
||||
: __imm_const(__sk_buff_cb_0, offsetof(struct __sk_buff, cb[0])),
|
||||
__imm_const(__sk_buff_cb_1, offsetof(struct __sk_buff, cb[1])),
|
||||
__imm_const(__sk_buff_cb_2, offsetof(struct __sk_buff, cb[2])),
|
||||
__imm_const(__sk_buff_cb_3, offsetof(struct __sk_buff, cb[3])),
|
||||
__imm_const(__sk_buff_cb_4, offsetof(struct __sk_buff, cb[4])),
|
||||
__imm_const(__sk_buff_napi_id, offsetof(struct __sk_buff, napi_id))
|
||||
: __clobber_all);
|
||||
}
|
||||
|
||||
SEC("cgroup/skb")
|
||||
__description("direct packet read test#4 for CGROUP_SKB")
|
||||
__success __success_unpriv __retval(0)
|
||||
__naked void test_4_for_cgroup_skb(void)
|
||||
{
|
||||
asm volatile (" \
|
||||
r2 = *(u32*)(r1 + %[__sk_buff_family]); \
|
||||
r3 = *(u32*)(r1 + %[__sk_buff_remote_ip4]); \
|
||||
r4 = *(u32*)(r1 + %[__sk_buff_local_ip4]); \
|
||||
r5 = *(u32*)(r1 + %[__sk_buff_remote_ip6_0]); \
|
||||
r5 = *(u32*)(r1 + %[__sk_buff_remote_ip6_1]); \
|
||||
r5 = *(u32*)(r1 + %[__sk_buff_remote_ip6_2]); \
|
||||
r5 = *(u32*)(r1 + %[__sk_buff_remote_ip6_3]); \
|
||||
r6 = *(u32*)(r1 + %[__sk_buff_local_ip6_0]); \
|
||||
r6 = *(u32*)(r1 + %[__sk_buff_local_ip6_1]); \
|
||||
r6 = *(u32*)(r1 + %[__sk_buff_local_ip6_2]); \
|
||||
r6 = *(u32*)(r1 + %[__sk_buff_local_ip6_3]); \
|
||||
r7 = *(u32*)(r1 + %[__sk_buff_remote_port]); \
|
||||
r8 = *(u32*)(r1 + %[__sk_buff_local_port]); \
|
||||
r0 = 0; \
|
||||
exit; \
|
||||
" :
|
||||
: __imm_const(__sk_buff_family, offsetof(struct __sk_buff, family)),
|
||||
__imm_const(__sk_buff_local_ip4, offsetof(struct __sk_buff, local_ip4)),
|
||||
__imm_const(__sk_buff_local_ip6_0, offsetof(struct __sk_buff, local_ip6[0])),
|
||||
__imm_const(__sk_buff_local_ip6_1, offsetof(struct __sk_buff, local_ip6[1])),
|
||||
__imm_const(__sk_buff_local_ip6_2, offsetof(struct __sk_buff, local_ip6[2])),
|
||||
__imm_const(__sk_buff_local_ip6_3, offsetof(struct __sk_buff, local_ip6[3])),
|
||||
__imm_const(__sk_buff_local_port, offsetof(struct __sk_buff, local_port)),
|
||||
__imm_const(__sk_buff_remote_ip4, offsetof(struct __sk_buff, remote_ip4)),
|
||||
__imm_const(__sk_buff_remote_ip6_0, offsetof(struct __sk_buff, remote_ip6[0])),
|
||||
__imm_const(__sk_buff_remote_ip6_1, offsetof(struct __sk_buff, remote_ip6[1])),
|
||||
__imm_const(__sk_buff_remote_ip6_2, offsetof(struct __sk_buff, remote_ip6[2])),
|
||||
__imm_const(__sk_buff_remote_ip6_3, offsetof(struct __sk_buff, remote_ip6[3])),
|
||||
__imm_const(__sk_buff_remote_port, offsetof(struct __sk_buff, remote_port))
|
||||
: __clobber_all);
|
||||
}
|
||||
|
||||
SEC("cgroup/skb")
|
||||
__description("invalid access of tc_classid for CGROUP_SKB")
|
||||
__failure __msg("invalid bpf_context access")
|
||||
__failure_unpriv
|
||||
__naked void tc_classid_for_cgroup_skb(void)
|
||||
{
|
||||
asm volatile (" \
|
||||
r0 = *(u32*)(r1 + %[__sk_buff_tc_classid]); \
|
||||
r0 = 0; \
|
||||
exit; \
|
||||
" :
|
||||
: __imm_const(__sk_buff_tc_classid, offsetof(struct __sk_buff, tc_classid))
|
||||
: __clobber_all);
|
||||
}
|
||||
|
||||
SEC("cgroup/skb")
|
||||
__description("invalid access of data_meta for CGROUP_SKB")
|
||||
__failure __msg("invalid bpf_context access")
|
||||
__failure_unpriv
|
||||
__naked void data_meta_for_cgroup_skb(void)
|
||||
{
|
||||
asm volatile (" \
|
||||
r0 = *(u32*)(r1 + %[__sk_buff_data_meta]); \
|
||||
r0 = 0; \
|
||||
exit; \
|
||||
" :
|
||||
: __imm_const(__sk_buff_data_meta, offsetof(struct __sk_buff, data_meta))
|
||||
: __clobber_all);
|
||||
}
|
||||
|
||||
SEC("cgroup/skb")
|
||||
__description("invalid access of flow_keys for CGROUP_SKB")
|
||||
__failure __msg("invalid bpf_context access")
|
||||
__failure_unpriv
|
||||
__naked void flow_keys_for_cgroup_skb(void)
|
||||
{
|
||||
asm volatile (" \
|
||||
r0 = *(u32*)(r1 + %[__sk_buff_flow_keys]); \
|
||||
r0 = 0; \
|
||||
exit; \
|
||||
" :
|
||||
: __imm_const(__sk_buff_flow_keys, offsetof(struct __sk_buff, flow_keys))
|
||||
: __clobber_all);
|
||||
}
|
||||
|
||||
SEC("cgroup/skb")
|
||||
__description("invalid write access to napi_id for CGROUP_SKB")
|
||||
__failure __msg("invalid bpf_context access")
|
||||
__failure_unpriv
|
||||
__naked void napi_id_for_cgroup_skb(void)
|
||||
{
|
||||
asm volatile (" \
|
||||
r9 = *(u32*)(r1 + %[__sk_buff_napi_id]); \
|
||||
*(u32*)(r1 + %[__sk_buff_napi_id]) = r9; \
|
||||
r0 = 0; \
|
||||
exit; \
|
||||
" :
|
||||
: __imm_const(__sk_buff_napi_id, offsetof(struct __sk_buff, napi_id))
|
||||
: __clobber_all);
|
||||
}
|
||||
|
||||
SEC("cgroup/skb")
|
||||
__description("write tstamp from CGROUP_SKB")
|
||||
__success __failure_unpriv
|
||||
__msg_unpriv("invalid bpf_context access off=152 size=8")
|
||||
__retval(0)
|
||||
__naked void write_tstamp_from_cgroup_skb(void)
|
||||
{
|
||||
asm volatile (" \
|
||||
r0 = 0; \
|
||||
*(u64*)(r1 + %[__sk_buff_tstamp]) = r0; \
|
||||
r0 = 0; \
|
||||
exit; \
|
||||
" :
|
||||
: __imm_const(__sk_buff_tstamp, offsetof(struct __sk_buff, tstamp))
|
||||
: __clobber_all);
|
||||
}
|
||||
|
||||
SEC("cgroup/skb")
|
||||
__description("read tstamp from CGROUP_SKB")
|
||||
__success __success_unpriv __retval(0)
|
||||
__naked void read_tstamp_from_cgroup_skb(void)
|
||||
{
|
||||
asm volatile (" \
|
||||
r0 = *(u64*)(r1 + %[__sk_buff_tstamp]); \
|
||||
r0 = 0; \
|
||||
exit; \
|
||||
" :
|
||||
: __imm_const(__sk_buff_tstamp, offsetof(struct __sk_buff, tstamp))
|
||||
: __clobber_all);
|
||||
}
|
||||
|
||||
char _license[] SEC("license") = "GPL";
|
@ -1,197 +0,0 @@
|
||||
{
|
||||
"direct packet read test#1 for CGROUP_SKB",
|
||||
.insns = {
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_2, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, data)),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_3, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, data_end)),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_4, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, len)),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_5, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, pkt_type)),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_6, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, mark)),
|
||||
BPF_STX_MEM(BPF_W, BPF_REG_1, BPF_REG_6,
|
||||
offsetof(struct __sk_buff, mark)),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_7, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, queue_mapping)),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_8, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, protocol)),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_9, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, vlan_present)),
|
||||
BPF_MOV64_REG(BPF_REG_0, BPF_REG_2),
|
||||
BPF_ALU64_IMM(BPF_ADD, BPF_REG_0, 8),
|
||||
BPF_JMP_REG(BPF_JGT, BPF_REG_0, BPF_REG_3, 1),
|
||||
BPF_LDX_MEM(BPF_B, BPF_REG_0, BPF_REG_2, 0),
|
||||
BPF_MOV64_IMM(BPF_REG_0, 0),
|
||||
BPF_EXIT_INSN(),
|
||||
},
|
||||
.result = ACCEPT,
|
||||
.result_unpriv = REJECT,
|
||||
.errstr_unpriv = "invalid bpf_context access off=76 size=4",
|
||||
.prog_type = BPF_PROG_TYPE_CGROUP_SKB,
|
||||
},
|
||||
{
|
||||
"direct packet read test#2 for CGROUP_SKB",
|
||||
.insns = {
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_4, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, vlan_tci)),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_5, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, vlan_proto)),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_6, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, priority)),
|
||||
BPF_STX_MEM(BPF_W, BPF_REG_1, BPF_REG_6,
|
||||
offsetof(struct __sk_buff, priority)),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_7, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, ingress_ifindex)),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_8, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, tc_index)),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_9, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, hash)),
|
||||
BPF_MOV64_IMM(BPF_REG_0, 0),
|
||||
BPF_EXIT_INSN(),
|
||||
},
|
||||
.result = ACCEPT,
|
||||
.prog_type = BPF_PROG_TYPE_CGROUP_SKB,
|
||||
},
|
||||
{
|
||||
"direct packet read test#3 for CGROUP_SKB",
|
||||
.insns = {
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_4, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, cb[0])),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_5, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, cb[1])),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_6, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, cb[2])),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_7, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, cb[3])),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_8, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, cb[4])),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_9, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, napi_id)),
|
||||
BPF_STX_MEM(BPF_W, BPF_REG_1, BPF_REG_4,
|
||||
offsetof(struct __sk_buff, cb[0])),
|
||||
BPF_STX_MEM(BPF_W, BPF_REG_1, BPF_REG_5,
|
||||
offsetof(struct __sk_buff, cb[1])),
|
||||
BPF_STX_MEM(BPF_W, BPF_REG_1, BPF_REG_6,
|
||||
offsetof(struct __sk_buff, cb[2])),
|
||||
BPF_STX_MEM(BPF_W, BPF_REG_1, BPF_REG_7,
|
||||
offsetof(struct __sk_buff, cb[3])),
|
||||
BPF_STX_MEM(BPF_W, BPF_REG_1, BPF_REG_8,
|
||||
offsetof(struct __sk_buff, cb[4])),
|
||||
BPF_MOV64_IMM(BPF_REG_0, 0),
|
||||
BPF_EXIT_INSN(),
|
||||
},
|
||||
.result = ACCEPT,
|
||||
.prog_type = BPF_PROG_TYPE_CGROUP_SKB,
|
||||
},
|
||||
{
|
||||
"direct packet read test#4 for CGROUP_SKB",
|
||||
.insns = {
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_2, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, family)),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_3, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, remote_ip4)),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_4, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, local_ip4)),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_5, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, remote_ip6[0])),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_5, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, remote_ip6[1])),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_5, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, remote_ip6[2])),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_5, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, remote_ip6[3])),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_6, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, local_ip6[0])),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_6, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, local_ip6[1])),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_6, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, local_ip6[2])),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_6, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, local_ip6[3])),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_7, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, remote_port)),
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_8, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, local_port)),
|
||||
BPF_MOV64_IMM(BPF_REG_0, 0),
|
||||
BPF_EXIT_INSN(),
|
||||
},
|
||||
.result = ACCEPT,
|
||||
.prog_type = BPF_PROG_TYPE_CGROUP_SKB,
|
||||
},
|
||||
{
|
||||
"invalid access of tc_classid for CGROUP_SKB",
|
||||
.insns = {
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_0, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, tc_classid)),
|
||||
BPF_MOV64_IMM(BPF_REG_0, 0),
|
||||
BPF_EXIT_INSN(),
|
||||
},
|
||||
.result = REJECT,
|
||||
.errstr = "invalid bpf_context access",
|
||||
.prog_type = BPF_PROG_TYPE_CGROUP_SKB,
|
||||
},
|
||||
{
|
||||
"invalid access of data_meta for CGROUP_SKB",
|
||||
.insns = {
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_0, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, data_meta)),
|
||||
BPF_MOV64_IMM(BPF_REG_0, 0),
|
||||
BPF_EXIT_INSN(),
|
||||
},
|
||||
.result = REJECT,
|
||||
.errstr = "invalid bpf_context access",
|
||||
.prog_type = BPF_PROG_TYPE_CGROUP_SKB,
|
||||
},
|
||||
{
|
||||
"invalid access of flow_keys for CGROUP_SKB",
|
||||
.insns = {
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_0, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, flow_keys)),
|
||||
BPF_MOV64_IMM(BPF_REG_0, 0),
|
||||
BPF_EXIT_INSN(),
|
||||
},
|
||||
.result = REJECT,
|
||||
.errstr = "invalid bpf_context access",
|
||||
.prog_type = BPF_PROG_TYPE_CGROUP_SKB,
|
||||
},
|
||||
{
|
||||
"invalid write access to napi_id for CGROUP_SKB",
|
||||
.insns = {
|
||||
BPF_LDX_MEM(BPF_W, BPF_REG_9, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, napi_id)),
|
||||
BPF_STX_MEM(BPF_W, BPF_REG_1, BPF_REG_9,
|
||||
offsetof(struct __sk_buff, napi_id)),
|
||||
BPF_MOV64_IMM(BPF_REG_0, 0),
|
||||
BPF_EXIT_INSN(),
|
||||
},
|
||||
.result = REJECT,
|
||||
.errstr = "invalid bpf_context access",
|
||||
.prog_type = BPF_PROG_TYPE_CGROUP_SKB,
|
||||
},
|
||||
{
|
||||
"write tstamp from CGROUP_SKB",
|
||||
.insns = {
|
||||
BPF_MOV64_IMM(BPF_REG_0, 0),
|
||||
BPF_STX_MEM(BPF_DW, BPF_REG_1, BPF_REG_0,
|
||||
offsetof(struct __sk_buff, tstamp)),
|
||||
BPF_MOV64_IMM(BPF_REG_0, 0),
|
||||
BPF_EXIT_INSN(),
|
||||
},
|
||||
.result = ACCEPT,
|
||||
.result_unpriv = REJECT,
|
||||
.errstr_unpriv = "invalid bpf_context access off=152 size=8",
|
||||
.prog_type = BPF_PROG_TYPE_CGROUP_SKB,
|
||||
},
|
||||
{
|
||||
"read tstamp from CGROUP_SKB",
|
||||
.insns = {
|
||||
BPF_LDX_MEM(BPF_DW, BPF_REG_0, BPF_REG_1,
|
||||
offsetof(struct __sk_buff, tstamp)),
|
||||
BPF_MOV64_IMM(BPF_REG_0, 0),
|
||||
BPF_EXIT_INSN(),
|
||||
},
|
||||
.result = ACCEPT,
|
||||
.prog_type = BPF_PROG_TYPE_CGROUP_SKB,
|
||||
},
|
Loading…
Reference in New Issue
Block a user