ret_t
const struct btf_type *ret_t;
ret_t = btf_type_by_id(ret_btf, ret_btf_id);
if (!ret_t || !__btf_type_is_struct(ret_t)) {
if (ret_t->size > BPF_GLOBAL_PERCPU_MA_MAX_SIZE) {
ret_t->size, BPF_GLOBAL_PERCPU_MA_MAX_SIZE);
err = bpf_mem_alloc_percpu_unit_init(&bpf_global_percpu_ma, ret_t->size);
if (!__btf_type_is_scalar_struct(env, ret_btf, ret_t, 0)) {
insn_aux->obj_new_size = ret_t->size;
ret_t = btf_type_by_id(desc_btf, meta->arg_constant.value);
if (!ret_t) {
} else if (btf_type_is_struct(ret_t)) {
} else if (btf_type_is_void(ret_t)) {
typedef int ret_t;
static const ret_t CONTINUE_PROCESSING = -1;
ret_t __ret = x; \
static INLINING ret_t accept_locally(struct __sk_buff *skb, encap_headers_t *encap)
static INLINING ret_t forward_with_gre(struct __sk_buff *skb, encap_headers_t *encap,
static INLINING ret_t forward_to_next_hop(struct __sk_buff *skb, encap_headers_t *encap,
static INLINING ret_t skip_next_hops(buf_t *pkt, int n)
static INLINING ret_t get_next_hop(buf_t *pkt, encap_headers_t *encap,
static const ret_t CONTINUE_PROCESSING = -1;
ret_t __ret = x; \
static ret_t accept_locally(struct __sk_buff *skb, encap_headers_t *encap)
static ret_t forward_with_gre(struct __sk_buff *skb, struct bpf_dynptr *dynptr,
static ret_t forward_to_next_hop(struct __sk_buff *skb, struct bpf_dynptr *dynptr,
static ret_t skip_next_hops(__u64 *offset, int n)
static ret_t get_next_hop(struct bpf_dynptr *dynptr, __u64 *offset, encap_headers_t *encap,
ret_t ret;