return_address
void *return_address(unsigned int);
#define ftrace_return_address(n) return_address(n)
EXPORT_SYMBOL_GPL(return_address);
#define ftrace_return_address(n) return_address(n)
extern void *return_address(unsigned int);
EXPORT_SYMBOL_GPL(return_address);
NOKPROBE_SYMBOL(return_address);
u32 return_address;
return_address = *vcpu_pc(vcpu);
return_address += return_offsets[vect_offset >> 2][is_thumb];
vcpu_gp_regs(vcpu)->compat_lr_abt = return_address;
vcpu_gp_regs(vcpu)->compat_lr_und = return_address;
extern unsigned long return_address(unsigned int);
#define ftrace_return_address(n) return_address(n)
extern void *return_address(unsigned int level);
#define ftrace_return_address(n) return_address(n)
EXPORT_SYMBOL_GPL(return_address);
NOKPROBE_SYMBOL(return_address);
#define ftrace_return_address(n) return_address(n)
unsigned long return_address;
OFFSET(__SFVDSO_RETURN_ADDRESS, stack_frame_vdso_wrapper, return_address);
if (__get_user(ip, &sf_vdso->return_address))
extern void *return_address(unsigned int);
#define ftrace_return_address(n) return_address(n)
EXPORT_SYMBOL_GPL(return_address);
unsigned long return_address;
if (__get_user(frame.return_address, &fp->return_address))
perf_callchain_store(entry, cs_base + frame.return_address);
if (__get_user(frame.return_address, &fp->return_address))
perf_callchain_store(entry, frame.return_address);
unsigned long return_address;
u32 return_address;
unsigned long return_address;
unsigned long return_address;
extern unsigned long return_address(unsigned level);
#define ftrace_return_address(n) return_address(n)
EXPORT_SYMBOL(return_address);
return_address = *(unsigned long *)(unsigned long)ctx->uc_mcontext.gregs[REG_ESP];
ip, return_address);
if (ip == return_address) {
static unsigned long return_address;
if (ip == return_address) {