Symbol: MODULES_VADDR
arch/arm/include/asm/memory.h
66
#if TASK_SIZE > MODULES_VADDR
arch/arm/include/asm/memory.h
84
#define XIP_VIRT_ADDR(physaddr) (MODULES_VADDR + ((physaddr) & 0x000fffff))
arch/arm/mm/dump.c
26
{ MODULES_VADDR, "Modules" },
arch/arm/mm/init.c
239
BUILD_BUG_ON(TASK_SIZE > MODULES_VADDR);
arch/arm/mm/init.c
240
BUG_ON(TASK_SIZE > MODULES_VADDR);
arch/arm/mm/init.c
463
.start = MODULES_VADDR,
arch/arm/mm/kasan_init.c
285
create_mapping((void *)MODULES_VADDR, (void *)(MODULES_END));
arch/arm/mm/mmu.c
1301
for (addr = KASAN_SHADOW_END; addr < MODULES_VADDR; addr += PMD_SIZE)
arch/arm/mm/mmu.c
1304
for (addr = 0; addr < MODULES_VADDR; addr += PMD_SIZE)
arch/arm/mm/mmu.c
1599
map.virtual = MODULES_VADDR;
arch/arm/mm/pageattr.c
66
if (!range_in_range(start, size, MODULES_VADDR, MODULES_END) &&
arch/arm/mm/pgd.c
55
new_p4d = p4d_alloc(mm, new_pgd + pgd_index(MODULES_VADDR),
arch/arm/mm/pgd.c
56
MODULES_VADDR);
arch/arm/mm/pgd.c
60
new_pud = pud_alloc(mm, new_p4d, MODULES_VADDR);
arch/arm/mm/pgd.c
77
(pmd_index(MODULES_VADDR) - pmd_index(TASK_SIZE))
arch/arm64/include/asm/memory.h
47
#define MODULES_END (MODULES_VADDR + MODULES_VSIZE)
arch/arm64/kernel/vmcore_info.c
24
vmcoreinfo_append_str("NUMBER(MODULES_VADDR)=0x%lx\n", MODULES_VADDR);
arch/arm64/mm/kasan_init.c
312
mod_shadow_start = (u64)kasan_mem_to_shadow((void *)MODULES_VADDR);
arch/arm64/mm/ptdump.c
388
{ MODULES_VADDR, "Modules start" },
arch/loongarch/include/asm/kasan.h
53
#define XKVRANGE_VC_START MODULES_VADDR
arch/loongarch/include/asm/kasan.h
54
#define XKVRANGE_VC_SIZE round_up(KFENCE_AREA_END - MODULES_VADDR + 1, PGDIR_SIZE)
arch/loongarch/include/asm/pgtable.h
96
#define MODULES_END (MODULES_VADDR + SZ_256M)
arch/loongarch/mm/init.c
223
#if defined(CONFIG_EXECMEM) && defined(MODULES_VADDR)
arch/loongarch/mm/init.c
231
.start = MODULES_VADDR,
arch/loongarch/mm/kasan_init.c
316
kasan_map_populate((unsigned long)mem_to_shadow((void *)MODULES_VADDR),
arch/mips/mm/fault.c
86
#ifdef MODULES_VADDR
arch/mips/mm/fault.c
87
if (unlikely(address >= MODULES_VADDR && address < MODULES_END))
arch/mips/mm/init.c
566
#ifdef MODULES_VADDR
arch/mips/mm/init.c
574
.start = MODULES_VADDR,
arch/nios2/mm/init.c
144
.start = MODULES_VADDR,
arch/powerpc/include/asm/task_size_32.h
15
#define MODULES_BASE (MODULES_VADDR & ~(UL(SZ_4M) - 1))
arch/powerpc/include/asm/task_size_32.h
23
#define MODULES_BASE (MODULES_VADDR & ~(UL(SZ_256M) - 1))
arch/powerpc/include/asm/uaccess.h
507
#ifdef MODULES_VADDR
arch/powerpc/include/asm/uaccess.h
508
const unsigned long border = MODULES_VADDR;
arch/powerpc/mm/book3s32/mmu.c
189
if (addr < ALIGN_DOWN(MODULES_VADDR, SZ_256M))
arch/powerpc/mm/kasan/init_32.c
119
#ifdef MODULES_VADDR
arch/powerpc/mm/kasan/init_32.c
120
k_start = (unsigned long)kasan_mem_to_shadow((void *)MODULES_VADDR);
arch/powerpc/mm/mem.c
390
for (va = ALIGN_DOWN(MODULES_VADDR, PGDIR_SIZE); va < MODULES_END; va += PGDIR_SIZE)
arch/powerpc/mm/mem.c
408
#ifdef MODULES_VADDR
arch/powerpc/mm/mem.c
412
if (MODULES_VADDR < PAGE_OFFSET && MODULES_END > limit) {
arch/powerpc/mm/mem.c
414
fallback_start = MODULES_VADDR;
arch/powerpc/mm/mem.c
417
start = MODULES_VADDR;
arch/powerpc/mm/ptdump/ptdump.c
267
#ifdef MODULES_VADDR
arch/powerpc/mm/ptdump/ptdump.c
268
address_markers[i++].start_address = MODULES_VADDR;
arch/powerpc/mm/ptdump/ptdump.c
77
#ifdef MODULES_VADDR
arch/riscv/kernel/vmcore_info.c
17
vmcoreinfo_append_str("NUMBER(MODULES_VADDR)=0x%lx\n", MODULES_VADDR);
arch/riscv/mm/init.c
151
print_ml("modules", (unsigned long)MODULES_VADDR,
arch/riscv/mm/init.c
1522
preallocate_pgd_pages_range(MODULES_VADDR, MODULES_END, "bpf/modules");
arch/riscv/mm/init.c
1541
.start = MODULES_VADDR,
arch/riscv/mm/kasan_init.c
503
(void *)kasan_mem_to_shadow((void *)MODULES_VADDR),
arch/riscv/mm/kasan_init.c
523
kasan_mem_to_shadow((const void *)MODULES_VADDR + SZ_2G));
arch/riscv/mm/ptdump.c
442
address_markers[MODULES_MAPPING_NR].start_address = MODULES_VADDR;
arch/s390/boot/startup.c
38
unsigned long __bootdata_preserved(MODULES_VADDR);
arch/s390/boot/startup.c
398
MODULES_VADDR = MODULES_END - MODULES_LEN;
arch/s390/boot/startup.c
399
VMALLOC_END = MODULES_VADDR;
arch/s390/boot/startup.c
402
boot_debug("modules area: 0x%016lx-0x%016lx\n", MODULES_VADDR, MODULES_END);
arch/s390/include/asm/pgtable.h
103
if (addr < (void *)MODULES_VADDR)
arch/s390/include/asm/pgtable.h
94
extern unsigned long MODULES_VADDR;
arch/s390/include/asm/pgtable.h
96
#define MODULES_VADDR MODULES_VADDR
arch/s390/kernel/setup.c
175
unsigned long __bootdata_preserved(MODULES_VADDR);
arch/s390/mm/dump_pagetables.c
336
rc |= add_marker(MODULES_VADDR, MODULES_END, "Modules Area");
arch/s390/mm/init.c
302
start = MODULES_VADDR + module_load_offset;
arch/sparc/mm/execmem.c
12
.start = MODULES_VADDR,
arch/sparc/mm/fault_64.c
306
(tpc >= MODULES_VADDR && tpc < MODULES_END)) {
arch/x86/hyperv/hv_init.c
525
hv_hypercall_pg = __vmalloc_node_range(PAGE_SIZE, 1, MODULES_VADDR,
arch/x86/include/asm/pgtable_32_areas.h
49
#define MODULES_LEN (MODULES_VADDR - MODULES_END)
arch/x86/include/asm/pgtable_64_types.h
178
#define MODULES_LEN (MODULES_END - MODULES_VADDR)
arch/x86/kernel/head64.c
228
BUILD_BUG_ON(MODULES_VADDR < __START_KERNEL_map);
arch/x86/kernel/head64.c
229
BUILD_BUG_ON(MODULES_VADDR - __START_KERNEL_map < KERNEL_IMAGE_SIZE);
arch/x86/kernel/head64.c
232
BUILD_BUG_ON((MODULES_VADDR & ~PMD_MASK) != 0);
arch/x86/kernel/head64.c
233
BUILD_BUG_ON(!(MODULES_VADDR > __START_KERNEL));
arch/x86/kernel/setup.c
839
MODULES_VADDR-1);
arch/x86/mm/dump_pagetables.c
105
[MODULES_VADDR_NR] = { MODULES_VADDR, "Modules" },
arch/x86/mm/fault.c
437
(address >= MODULES_VADDR && address <= MODULES_END)) {
arch/x86/mm/init.c
1075
start = MODULES_VADDR + offset;
fs/proc/kcore.c
705
#if defined(CONFIG_MODULES) && defined(MODULES_VADDR)
fs/proc/kcore.c
712
if (MODULES_VADDR != VMALLOC_START && MODULES_END != VMALLOC_END) {
fs/proc/kcore.c
713
kclist_add(&kcore_modules, (void *)MODULES_VADDR,
fs/proc/kcore.c
714
MODULES_END - MODULES_VADDR, KCORE_VMALLOC);
kernel/bpf/core.c
1030
#if defined(MODULES_VADDR)
kernel/bpf/core.c
1031
return MODULES_END - MODULES_VADDR;
kernel/bpf/verifier.c
22858
#if defined(MODULES_VADDR)
kernel/bpf/verifier.c
22859
u64 addr = MODULES_VADDR;
mm/kmsan/kmsan.h
179
return ((u64)vaddr >= MODULES_VADDR) && ((u64)vaddr < MODULES_END);
mm/kmsan/shadow.c
66
off = addr64 - MODULES_VADDR;
mm/vmalloc.c
785
#if defined(CONFIG_EXECMEM) && defined(MODULES_VADDR)
mm/vmalloc.c
787
if (addr >= MODULES_VADDR && addr < MODULES_END)