Symbol: stacktrace
arch/alpha/lib/stacktrace.c
94
instr * prologue = (instr *)stacktrace;
arch/arm64/kvm/hyp/nvhe/stacktrace.c
106
stacktrace[*idx] = where;
arch/arm64/kvm/hyp/nvhe/stacktrace.c
107
stacktrace[++*idx] = 0UL;
arch/arm64/kvm/hyp/nvhe/stacktrace.c
96
unsigned long *stacktrace = this_cpu_ptr(pkvm_stacktrace);
arch/arm64/kvm/stacktrace.c
216
unsigned long *stacktrace
arch/arm64/kvm/stacktrace.c
223
i < ARRAY_SIZE(kvm_nvhe_sym(pkvm_stacktrace)) && stacktrace[i];
arch/arm64/kvm/stacktrace.c
225
kvm_nvhe_dump_backtrace_entry((void *)hyp_offset, stacktrace[i]);
kernel/trace/trace_events.c
200
__generic_field(char *, stacktrace, FILTER_STACKTRACE);
tools/mm/page_owner_sort.c
135
return strcmp(l1->stacktrace, l2->stacktrace);
tools/mm/page_owner_sort.c
32
char *stacktrace;
tools/mm/page_owner_sort.c
497
list[list_size].stacktrace = strchr(list[list_size].txt, '\n') ?: "";
tools/mm/page_owner_sort.c
498
if (*list[list_size].stacktrace == '\n')
tools/mm/page_owner_sort.c
499
list[list_size].stacktrace++;
tools/mm/page_owner_sort.c
869
fprintf(fout, ":\n%s", list[i].stacktrace);
tools/perf/util/bpf_skel/lock_contention.bpf.c
469
static inline s32 get_owner_stack_id(u64 *stacktrace)
tools/perf/util/bpf_skel/lock_contention.bpf.c
474
id = bpf_map_lookup_elem(&owner_stacks, stacktrace);
tools/perf/util/bpf_skel/lock_contention.bpf.c
480
bpf_map_update_elem(&owner_stacks, stacktrace, &new_id, BPF_NOEXIST);
tools/perf/util/bpf_skel/lock_contention.bpf.c
482
id = bpf_map_lookup_elem(&owner_stacks, stacktrace);
tools/testing/selftests/bpf/bench.c
355
env.stacktrace = true;
tools/testing/selftests/bpf/bench.h
29
bool stacktrace;
tools/testing/selftests/bpf/benchs/bench_trigger.c
149
ctx.skel->rodata->stacktrace = env.stacktrace;
tools/testing/selftests/bpf/progs/trigger_bench.c
28
volatile const int stacktrace;
tools/testing/selftests/bpf/progs/trigger_bench.c
41
if (!stacktrace)