kaslr_offset
if (kaslr_offset() < MIN_KIMG_ALIGN) {
relocate_kernel(kaslr_offset);
u64 kaslr_offset = pa_base % MIN_KIMG_ALIGN;
kaslr_offset |= kaslr_seed & ~(MIN_KIMG_ALIGN - 1);
va_base = KIMAGE_VADDR + kaslr_offset;
map_kernel(kaslr_offset, va_base - pa_base, root_level);
static void __init map_kernel(u64 kaslr_offset, u64 va_offset, int root_level)
const unsigned long offset = kaslr_offset();
vmcoreinfo_append_str("KERNELOFFSET=%lx\n", kaslr_offset());
(void *)(panic_addr + kaslr_offset()));
u64 hyp_offset = elr_in_kimg - kaslr_offset() - elr_virt;
dump_kernel_instr(panic_addr + kaslr_offset());
kvm_err(" [<%016lx>] %pB\n", where, (void *)(where + kaslr_offset()));
kaslr_offset(), KERNELBASE);
if (IS_ENABLED(CONFIG_RANDOMIZE_BASE) && kaslr_offset() > 0)
vmcoreinfo_append_str("KERNELOFFSET=%lx\n", kaslr_offset());
unsigned long kaslr_offset(void);
unsigned long kaslr_offset;
#define __kaslr_offset vm_layout.kaslr_offset
os_info_entry_add_val(OS_INFO_KASLR_OFFSET, kaslr_offset());
vmcoreinfo_append_str("KERNELOFFSET=%lx\n", kaslr_offset());
kaslr_offset(),
vmcoreinfo_append_str("KERNELOFFSET=%lx\n", kaslr_offset());
#define get_boot_seed() kaslr_offset()
ip -= kaslr_offset();