reloc_offset
static size_t reloc_offset;
reloc_offset, HYP_SECTION_SYMBOL_PREFIX, sh_orig_name,
reloc_offset += 4;
#define RELOCATED(x) ((void *)((long)x + reloc_offset))
*new_addr = (unsigned long)reloc_offset;
static unsigned long reloc_offset;
reloc_offset = (unsigned long)_text - VMLINUX_LOAD_ADDRESS;
reloc_offset += random_offset;
update_reloc_offset(&reloc_offset, random_offset);
if (reloc_offset)
if (reloc_offset > 0) {
pr_cont("Kernel relocated by 0x%lx\n", reloc_offset);
addr = (u64 *)(*relr + reloc_offset);
*addr++ += reloc_offset;
*p += reloc_offset;
extern unsigned long reloc_offset(void);
unsigned long kva, offset = reloc_offset();
unsigned long offset = reloc_offset();
unsigned long offset = reloc_offset();
uintptr_t reloc_offset = kernel_map.virt_addr - KERNEL_LINK_ADDR;
relocated_addr += reloc_offset;
if (r->reloc_offset > bo->obj->base.size - sizeof(*ptr)) {
ptr[off] = bo->mapping->iova + r->reloc_offset;
iova += submit_reloc.reloc_offset;
__u64 reloc_offset; /* in, offset from start of reloc_bo */
__u64 reloc_offset; /* in, offset from start of reloc_bo */
(reloc_offset(reloc) - reloc_offset(table));
if (reloc_offset(reloc) > offset)
scan_for_insn(reloc->sec->base, reloc_offset(reloc),
return addend + insn_off + insn_len - reloc_offset(reloc);
addend += insn->offset + insn->len - reloc_offset(reloc);
if (table_size && reloc_offset(reloc) - reloc_offset(table) >= table_size)
if (prev_offset && reloc_offset(reloc) != prev_offset + arch_reloc_size(reloc))
prev_offset = reloc_offset(reloc);
reloc_offset(reloc) + 1,
(insn->offset + insn->len) - (reloc_offset(reloc) + 1))) {
WARN_FUNC(reloc->sec->base, reloc_offset(reloc),
sec->name, (unsigned long long)reloc_offset(reloc));
idx = (reloc_offset(reloc) - sym->offset) / sizeof(unsigned long);
off = reloc_offset(reloc) + 1;
if (reloc_offset(reloc) >= offset &&
reloc_offset(reloc) < offset + len) {
if (!r || reloc_offset(reloc) < reloc_offset(r))
reloc && reloc_offset(reloc) < sym->offset + sym->len; \
return sec_offset_hash(reloc->sec, reloc_offset(reloc));
ERROR_FUNC(patched_reloc->sec->base, reloc_offset(patched_reloc),
if (reloc_offset(patched_reloc) < start ||
reloc_offset(patched_reloc) >= end)
ERROR_FUNC(patched_rsec->base, reloc_offset(patched_reloc),
offset = out_sym->offset + (reloc_offset(patched_reloc) - patched_sym->offset);