BPF_F_USER_STACK
bool user = flags & BPF_F_USER_STACK;
bool user = flags & BPF_F_USER_STACK;
if (unlikely(flags & ~(BPF_F_SKIP_FIELD_MASK | BPF_F_USER_STACK |
if (unlikely(flags & ~(BPF_F_SKIP_FIELD_MASK | BPF_F_USER_STACK |
user = flags & BPF_F_USER_STACK;
bool user = flags & BPF_F_USER_STACK;
if (unlikely(flags & ~(BPF_F_SKIP_FIELD_MASK | BPF_F_USER_STACK |
if (unlikely(flags & ~(BPF_F_SKIP_FIELD_MASK | BPF_F_USER_STACK |
user = flags & BPF_F_USER_STACK;
#define USER_STACKID_FLAGS (0 | BPF_F_FAST_STACK_CMP | BPF_F_USER_STACK)
BPF_F_FAST_STACK_CMP | BPF_F_USER_STACK);
if (stack_id > 0 && bpf_get_stack(ctx, &pelem->stack, MAX_STACKS * sizeof(u64), BPF_F_USER_STACK)) {
MAX_STACK_TRACE_DEPTH * SIZE_OF_ULONG, BPF_F_USER_STACK);
val = bpf_get_stackid(ctx, &stackmap, BPF_F_USER_STACK);
val = bpf_get_stack(ctx, trace, sizeof(stack_trace_t), BPF_F_USER_STACK);
event->user_stack_id = bpf_get_stackid(ctx, &stackmap, BPF_F_USER_STACK);
sample->user_stack_id = bpf_get_stackid(ctx, &stacks_1, BPF_F_USER_STACK);
sample->user_stack_id = bpf_get_stackid(ctx, &stacks_0, BPF_F_USER_STACK);
BPF_F_USER_STACK | BPF_F_USER_BUILD_ID);
BPF_F_USER_STACK | BPF_F_USER_BUILD_ID);
BPF_F_USER_STACK);
BPF_F_USER_STACK | BPF_F_USER_BUILD_ID);
usize = bpf_get_stack(ctx, raw_data, max_len, BPF_F_USER_STACK);
key = bpf_get_stackid(args, &stackmap, BPF_F_USER_STACK);
BPF_F_USER_STACK | BPF_F_USER_BUILD_ID);
entry1_len = bpf_get_stack(ctx, &entry_stack1, SZ, BPF_F_USER_STACK);
entry1_recur_len = bpf_get_stack(ctx, &entry_stack1_recur, SZ, BPF_F_USER_STACK);
exit1_recur_len = bpf_get_stack(ctx, &exit_stack1_recur, SZ, BPF_F_USER_STACK);
exit1_len = bpf_get_stack(ctx, &exit_stack1, SZ, BPF_F_USER_STACK);
entry2_len = bpf_get_stack(ctx, &entry_stack2, SZ, BPF_F_USER_STACK);
entry3_len = bpf_get_stack(ctx, &entry_stack3, SZ, BPF_F_USER_STACK);
entry4_len = bpf_get_stack(ctx, &entry_stack4, SZ, BPF_F_USER_STACK);
exit4_len = bpf_get_stack(ctx, &exit_stack4, SZ, BPF_F_USER_STACK);
usdt_len = bpf_get_stack(ctx, &usdt_stack, SZ, BPF_F_USER_STACK);