MODULES_VADDR
#if TASK_SIZE > MODULES_VADDR
#define XIP_VIRT_ADDR(physaddr) (MODULES_VADDR + ((physaddr) & 0x000fffff))
{ MODULES_VADDR, "Modules" },
BUILD_BUG_ON(TASK_SIZE > MODULES_VADDR);
BUG_ON(TASK_SIZE > MODULES_VADDR);
.start = MODULES_VADDR,
create_mapping((void *)MODULES_VADDR, (void *)(MODULES_END));
for (addr = KASAN_SHADOW_END; addr < MODULES_VADDR; addr += PMD_SIZE)
for (addr = 0; addr < MODULES_VADDR; addr += PMD_SIZE)
map.virtual = MODULES_VADDR;
if (!range_in_range(start, size, MODULES_VADDR, MODULES_END) &&
new_p4d = p4d_alloc(mm, new_pgd + pgd_index(MODULES_VADDR),
MODULES_VADDR);
new_pud = pud_alloc(mm, new_p4d, MODULES_VADDR);
(pmd_index(MODULES_VADDR) - pmd_index(TASK_SIZE))
#define MODULES_END (MODULES_VADDR + MODULES_VSIZE)
vmcoreinfo_append_str("NUMBER(MODULES_VADDR)=0x%lx\n", MODULES_VADDR);
mod_shadow_start = (u64)kasan_mem_to_shadow((void *)MODULES_VADDR);
{ MODULES_VADDR, "Modules start" },
#define XKVRANGE_VC_START MODULES_VADDR
#define XKVRANGE_VC_SIZE round_up(KFENCE_AREA_END - MODULES_VADDR + 1, PGDIR_SIZE)
#define MODULES_END (MODULES_VADDR + SZ_256M)
#if defined(CONFIG_EXECMEM) && defined(MODULES_VADDR)
.start = MODULES_VADDR,
kasan_map_populate((unsigned long)mem_to_shadow((void *)MODULES_VADDR),
#ifdef MODULES_VADDR
if (unlikely(address >= MODULES_VADDR && address < MODULES_END))
#ifdef MODULES_VADDR
.start = MODULES_VADDR,
.start = MODULES_VADDR,
#define MODULES_BASE (MODULES_VADDR & ~(UL(SZ_4M) - 1))
#define MODULES_BASE (MODULES_VADDR & ~(UL(SZ_256M) - 1))
#ifdef MODULES_VADDR
const unsigned long border = MODULES_VADDR;
if (addr < ALIGN_DOWN(MODULES_VADDR, SZ_256M))
#ifdef MODULES_VADDR
k_start = (unsigned long)kasan_mem_to_shadow((void *)MODULES_VADDR);
for (va = ALIGN_DOWN(MODULES_VADDR, PGDIR_SIZE); va < MODULES_END; va += PGDIR_SIZE)
#ifdef MODULES_VADDR
if (MODULES_VADDR < PAGE_OFFSET && MODULES_END > limit) {
fallback_start = MODULES_VADDR;
start = MODULES_VADDR;
#ifdef MODULES_VADDR
address_markers[i++].start_address = MODULES_VADDR;
#ifdef MODULES_VADDR
vmcoreinfo_append_str("NUMBER(MODULES_VADDR)=0x%lx\n", MODULES_VADDR);
print_ml("modules", (unsigned long)MODULES_VADDR,
preallocate_pgd_pages_range(MODULES_VADDR, MODULES_END, "bpf/modules");
.start = MODULES_VADDR,
(void *)kasan_mem_to_shadow((void *)MODULES_VADDR),
kasan_mem_to_shadow((const void *)MODULES_VADDR + SZ_2G));
address_markers[MODULES_MAPPING_NR].start_address = MODULES_VADDR;
unsigned long __bootdata_preserved(MODULES_VADDR);
MODULES_VADDR = MODULES_END - MODULES_LEN;
VMALLOC_END = MODULES_VADDR;
boot_debug("modules area: 0x%016lx-0x%016lx\n", MODULES_VADDR, MODULES_END);
if (addr < (void *)MODULES_VADDR)
extern unsigned long MODULES_VADDR;
#define MODULES_VADDR MODULES_VADDR
unsigned long __bootdata_preserved(MODULES_VADDR);
rc |= add_marker(MODULES_VADDR, MODULES_END, "Modules Area");
start = MODULES_VADDR + module_load_offset;
.start = MODULES_VADDR,
(tpc >= MODULES_VADDR && tpc < MODULES_END)) {
hv_hypercall_pg = __vmalloc_node_range(PAGE_SIZE, 1, MODULES_VADDR,
#define MODULES_LEN (MODULES_VADDR - MODULES_END)
#define MODULES_LEN (MODULES_END - MODULES_VADDR)
BUILD_BUG_ON(MODULES_VADDR < __START_KERNEL_map);
BUILD_BUG_ON(MODULES_VADDR - __START_KERNEL_map < KERNEL_IMAGE_SIZE);
BUILD_BUG_ON((MODULES_VADDR & ~PMD_MASK) != 0);
BUILD_BUG_ON(!(MODULES_VADDR > __START_KERNEL));
MODULES_VADDR-1);
[MODULES_VADDR_NR] = { MODULES_VADDR, "Modules" },
(address >= MODULES_VADDR && address <= MODULES_END)) {
start = MODULES_VADDR + offset;
#if defined(CONFIG_MODULES) && defined(MODULES_VADDR)
if (MODULES_VADDR != VMALLOC_START && MODULES_END != VMALLOC_END) {
kclist_add(&kcore_modules, (void *)MODULES_VADDR,
MODULES_END - MODULES_VADDR, KCORE_VMALLOC);
#if defined(MODULES_VADDR)
return MODULES_END - MODULES_VADDR;
#if defined(MODULES_VADDR)
u64 addr = MODULES_VADDR;
return ((u64)vaddr >= MODULES_VADDR) && ((u64)vaddr < MODULES_END);
off = addr64 - MODULES_VADDR;
#if defined(CONFIG_EXECMEM) && defined(MODULES_VADDR)
if (addr >= MODULES_VADDR && addr < MODULES_END)