Symbol: __static_call_return0
arch/arm64/include/asm/static_call.h
26
ARCH_DEFINE_STATIC_CALL_TRAMP(name, __static_call_return0)
arch/arm64/include/asm/static_call.h
29
ARCH_DEFINE_STATIC_CALL_TRAMP(name, __static_call_return0)
arch/arm64/kernel/static_call.c
12
func = __static_call_return0;
arch/powerpc/kernel/static_call.c
10
bool is_ret0 = (func == __static_call_return0);
arch/x86/events/core.c
2202
x86_pmu.guest_get_msrs = (void *)&__static_call_return0;
arch/x86/include/asm/static_call.h
58
ARCH_DEFINE_STATIC_CALL_TRAMP(name, __static_call_return0)
arch/x86/kernel/static_call.c
73
if (func == &__static_call_return0) {
arch/x86/kvm/x86.c
10012
(void *)__static_call_return0);
include/linux/static_call.h
185
extern long __static_call_return0(void);
include/linux/static_call.h
206
.func = __static_call_return0, \
include/linux/static_call.h
251
.func = __static_call_return0, \
include/linux/static_call.h
271
extern long __static_call_return0(void);
include/linux/static_call.h
310
__DEFINE_STATIC_CALL(name, _func, __static_call_return0)
kernel/events/core.c
7701
static_call_update(__perf_guest_state, (void *)&__static_call_return0);
kernel/events/core.c
7702
static_call_update(__perf_guest_get_ip, (void *)&__static_call_return0);
kernel/events/core.c
7703
static_call_update(__perf_guest_handle_intel_pt_intr, (void *)&__static_call_return0);
kernel/events/core.c
7704
static_call_update(__perf_guest_handle_mediated_pmi, (void *)&__static_call_return0);
kernel/sched/core.c
7449
# define cond_resched_dynamic_disabled ((void *)&__static_call_return0)
kernel/sched/core.c
7454
# define might_resched_dynamic_disabled ((void *)&__static_call_return0)
kernel/static_call.c
8
EXPORT_SYMBOL_GPL(__static_call_return0);