Symbol: MODULES_END
arch/arm/mm/init.c
464
.end = MODULES_END,
arch/arm/mm/kasan_init.c
285
create_mapping((void *)MODULES_VADDR, (void *)(MODULES_END));
arch/arm/mm/pageattr.c
66
if (!range_in_range(start, size, MODULES_VADDR, MODULES_END) &&
arch/arm64/include/asm/memory.h
46
#define KIMAGE_VADDR (MODULES_END)
arch/arm64/include/asm/pgtable.h
23
#define VMALLOC_START (MODULES_END)
arch/arm64/kernel/vmcore_info.c
25
vmcoreinfo_append_str("NUMBER(MODULES_END)=0x%lx\n", MODULES_END);
arch/arm64/mm/init.c
461
BUILD_BUG_ON(KIMAGE_VADDR != MODULES_END);
arch/arm64/mm/kasan_init.c
348
BUILD_BUG_ON(VMALLOC_START != MODULES_END);
arch/arm64/mm/ptdump.c
389
{ MODULES_END, "Modules end" },
arch/loongarch/include/asm/pgtable.h
104
#define VMALLOC_START MODULES_END
arch/loongarch/mm/init.c
232
.end = MODULES_END,
arch/loongarch/mm/kasan_init.c
317
(unsigned long)mem_to_shadow((void *)MODULES_END), NUMA_NO_NODE);
arch/mips/mm/fault.c
87
if (unlikely(address >= MODULES_VADDR && address < MODULES_END))
arch/mips/mm/init.c
575
.end = MODULES_END,
arch/nios2/mm/init.c
145
.end = MODULES_END,
arch/powerpc/include/asm/task_size_32.h
14
#define MODULES_VADDR (MODULES_END - MODULES_SIZE)
arch/powerpc/include/asm/task_size_32.h
22
#define MODULES_VADDR (MODULES_END - MODULES_SIZE)
arch/powerpc/mm/book3s32/mmu.c
191
if (addr > ALIGN(MODULES_END, SZ_256M) - 1)
arch/powerpc/mm/kasan/init_32.c
121
k_end = (unsigned long)kasan_mem_to_shadow((void *)MODULES_END);
arch/powerpc/mm/mem.c
390
for (va = ALIGN_DOWN(MODULES_VADDR, PGDIR_SIZE); va < MODULES_END; va += PGDIR_SIZE)
arch/powerpc/mm/mem.c
412
if (MODULES_VADDR < PAGE_OFFSET && MODULES_END > limit) {
arch/powerpc/mm/mem.c
415
fallback_end = MODULES_END;
arch/powerpc/mm/mem.c
420
end = MODULES_END;
arch/powerpc/mm/ptdump/ptdump.c
269
address_markers[i++].start_address = MODULES_END;
arch/riscv/include/asm/pgtable.h
50
#define BPF_JIT_REGION_END (MODULES_END)
arch/riscv/kernel/vmcore_info.c
18
vmcoreinfo_append_str("NUMBER(MODULES_END)=0x%lx\n", MODULES_END);
arch/riscv/mm/init.c
152
(unsigned long)MODULES_END);
arch/riscv/mm/init.c
1522
preallocate_pgd_pages_range(MODULES_VADDR, MODULES_END, "bpf/modules");
arch/riscv/mm/init.c
1542
.end = MODULES_END,
arch/riscv/mm/kasan_init.c
504
(void *)kasan_mem_to_shadow((void *)MODULES_END));
arch/riscv/mm/kasan_init.c
522
kasan_populate(kasan_mem_to_shadow((const void *)MODULES_END),
arch/s390/boot/startup.c
39
unsigned long __bootdata_preserved(MODULES_END);
arch/s390/boot/startup.c
397
MODULES_END = round_down(kernel_start, _SEGMENT_SIZE);
arch/s390/boot/startup.c
398
MODULES_VADDR = MODULES_END - MODULES_LEN;
arch/s390/boot/startup.c
402
boot_debug("modules area: 0x%016lx-0x%016lx\n", MODULES_VADDR, MODULES_END);
arch/s390/boot/vmem.c
130
kasan_populate(VMALLOC_START, MODULES_END, POPULATE_KASAN_SHALLOW);
arch/s390/include/asm/pgtable.h
105
if (addr > (void *)MODULES_END)
arch/s390/include/asm/pgtable.h
95
extern unsigned long MODULES_END;
arch/s390/include/asm/pgtable.h
97
#define MODULES_END MODULES_END
arch/s390/kernel/setup.c
176
unsigned long __bootdata_preserved(MODULES_END);
arch/s390/mm/dump_pagetables.c
336
rc |= add_marker(MODULES_VADDR, MODULES_END, "Modules Area");
arch/s390/mm/init.c
309
.end = MODULES_END,
arch/sparc/mm/execmem.c
13
.end = MODULES_END,
arch/sparc/mm/fault_64.c
306
(tpc >= MODULES_VADDR && tpc < MODULES_END)) {
arch/x86/hyperv/hv_init.c
526
MODULES_END, GFP_KERNEL, PAGE_KERNEL_ROX,
arch/x86/include/asm/pgtable_32_areas.h
49
#define MODULES_LEN (MODULES_VADDR - MODULES_END)
arch/x86/include/asm/pgtable_64_types.h
178
#define MODULES_LEN (MODULES_END - MODULES_VADDR)
arch/x86/kernel/head64.c
234
MAYBE_BUILD_BUG_ON(!(((MODULES_END - 1) & PGDIR_MASK) ==
arch/x86/kernel/head64.c
236
BUILD_BUG_ON(__fix_to_virt(__end_of_fixed_addresses) <= MODULES_END);
arch/x86/mm/dump_pagetables.c
106
[MODULES_END_NR] = { MODULES_END, "End Modules" },
arch/x86/mm/fault.c
437
(address >= MODULES_VADDR && address <= MODULES_END)) {
arch/x86/mm/init.c
1091
.end = MODULES_END,
arch/x86/mm/init.c
1098
.end = MODULES_END,
arch/x86/mm/init.c
1105
.end = MODULES_END,
arch/x86/mm/init.c
1112
.end = MODULES_END,
arch/x86/mm/init.c
1119
.end = MODULES_END,
arch/x86/mm/kasan_init_64.c
428
kasan_populate_early_shadow(kasan_mem_to_shadow((void *)MODULES_END),
fs/proc/kcore.c
712
if (MODULES_VADDR != VMALLOC_START && MODULES_END != VMALLOC_END) {
fs/proc/kcore.c
714
MODULES_END - MODULES_VADDR, KCORE_VMALLOC);
kernel/bpf/core.c
1031
return MODULES_END - MODULES_VADDR;
mm/kmsan/kmsan.h
179
return ((u64)vaddr >= MODULES_VADDR) && ((u64)vaddr < MODULES_END);
mm/vmalloc.c
787
if (addr >= MODULES_VADDR && addr < MODULES_END)