Symbol: VMALLOC_END
arch/arc/mm/highmem.c
67
BUILD_BUG_ON(PAGE_OFFSET < (VMALLOC_END + FIXMAP_SIZE + PKMAP_SIZE));
arch/arm/kernel/traps.c
961
if (start < VMALLOC_END && end > VMALLOC_START)
arch/arm/mm/init.c
457
fallback_end = VMALLOC_END;
arch/arm/mm/iomap.c
41
(unsigned long)addr < VMALLOC_END)
arch/arm/mm/ioremap.c
145
memcpy_pgd(mm, VMALLOC_START, VMALLOC_END);
arch/arm/mm/ioremap.c
150
arm_kasan_mem_to_shadow(VMALLOC_END);
arch/arm/mm/kasan_init.c
246
kasan_mem_to_shadow((void *)VMALLOC_END));
arch/arm/mm/kasan_init.c
248
kasan_populate_early_shadow(kasan_mem_to_shadow((void *)VMALLOC_END),
arch/arm/mm/mmu.c
1171
vmalloc_max = VMALLOC_END - (PAGE_OFFSET + SZ_32M + VMALLOC_OFFSET);
arch/arm/mm/mmu.c
1198
vmalloc_limit = (u64)VMALLOC_END - vmalloc_size - VMALLOC_OFFSET -
arch/arm/mm/mmu.c
987
(md->virtual < VMALLOC_START || md->virtual >= VMALLOC_END)) {
arch/arm/mm/pageattr.c
67
!range_in_range(start, size, VMALLOC_START, VMALLOC_END))
arch/arm64/kernel/pi/kaslr_early.c
60
range = (VMALLOC_END - KIMAGE_VADDR) / 2;
arch/arm64/kernel/vmcore_info.c
26
vmcoreinfo_append_str("NUMBER(VMALLOC_END)=0x%lx\n", VMALLOC_END);
arch/arm64/mm/init.c
532
.end = VMALLOC_END,
arch/arm64/mm/init.c
538
.end = VMALLOC_END,
arch/arm64/mm/kasan_init.c
314
vmalloc_shadow_end = (u64)kasan_mem_to_shadow((void *)VMALLOC_END);
arch/arm64/mm/ptdump.c
391
{ VMALLOC_END, "vmalloc() end" },
arch/arm64/net/bpf_jit_comp.c
2282
return VMALLOC_END - VMALLOC_START;
arch/csky/mm/fault.c
214
if (unlikely((addr >= VMALLOC_START) && (addr <= VMALLOC_END))) {
arch/loongarch/include/asm/pgtable.h
116
#define vmemmap ((struct page *)((VMALLOC_END + PMD_SIZE) & PMD_MASK))
arch/mips/mm/fault.c
84
if (unlikely(address >= VMALLOC_START && address <= VMALLOC_END))
arch/nios2/mm/fault.c
67
if (unlikely(address >= VMALLOC_START && address <= VMALLOC_END)) {
arch/parisc/mm/init.c
587
(void*)VMALLOC_START, (void*)VMALLOC_END,
arch/parisc/mm/init.c
588
(VMALLOC_END - VMALLOC_START) >> 20,
arch/parisc/mm/init.c
991
.end = VMALLOC_END,
arch/powerpc/include/asm/book3s/32/pgtable.h
169
#define IOREMAP_END VMALLOC_END
arch/powerpc/include/asm/nohash/32/pgtable.h
84
#define IOREMAP_END VMALLOC_END
arch/powerpc/include/asm/nohash/64/pgtable.h
66
#define VMEMMAP_BASE VMALLOC_END
arch/powerpc/mm/book3s64/hash_utils.c
1883
} else if (ea < VMALLOC_END) {
arch/powerpc/mm/kasan/init_32.c
115
unsigned long k_end = (unsigned long)kasan_mem_to_shadow((void *)VMALLOC_END);
arch/powerpc/mm/mem.c
423
end = VMALLOC_END;
arch/powerpc/mm/mem.c
440
.end = VMALLOC_END,
arch/powerpc/mm/mem.c
446
.end = VMALLOC_END,
arch/powerpc/mm/ptdump/hashpagetable.c
390
if (addr < VMALLOC_END)
arch/powerpc/mm/ptdump/hashpagetable.c
511
address_markers[2].start_address = VMALLOC_END;
arch/powerpc/mm/ptdump/ptdump.c
272
address_markers[i++].start_address = VMALLOC_END;
arch/riscv/include/asm/pgtable.h
53
#define BPF_JIT_REGION_END (VMALLOC_END)
arch/riscv/include/asm/pgtable.h
64
#define MODULES_END VMALLOC_END
arch/riscv/kernel/vmcore_info.c
11
vmcoreinfo_append_str("NUMBER(VMALLOC_END)=0x%lx\n", VMALLOC_END);
arch/riscv/mm/fault.c
312
unlikely(addr >= VMALLOC_START && addr < VMALLOC_END)) {
arch/riscv/mm/init.c
149
(unsigned long)VMALLOC_END);
arch/riscv/mm/init.c
1520
preallocate_pgd_pages_range(VMALLOC_START, VMALLOC_END, "vmalloc");
arch/riscv/mm/init.c
1548
.end = VMALLOC_END,
arch/riscv/mm/kasan_init.c
500
(void *)kasan_mem_to_shadow((void *)VMALLOC_END));
arch/riscv/mm/kasan_init.c
507
(void *)kasan_mem_to_shadow((void *)VMALLOC_END));
arch/riscv/mm/ptdump.c
435
address_markers[VMALLOC_END_NR].start_address = VMALLOC_END;
arch/s390/boot/startup.c
35
unsigned long __bootdata_preserved(VMALLOC_END);
arch/s390/boot/startup.c
399
VMALLOC_END = MODULES_VADDR;
arch/s390/boot/startup.c
401
VMALLOC_END -= MODULES_LEN * 2;
arch/s390/boot/startup.c
405
vsize = (VMALLOC_END - FIXMAP_SIZE) / 2;
arch/s390/boot/startup.c
411
VMALLOC_END -= vmalloc_size * 2;
arch/s390/boot/startup.c
413
VMALLOC_START = VMALLOC_END - vmalloc_size;
arch/s390/boot/startup.c
414
boot_debug("vmalloc area: 0x%016lx-0x%016lx\n", VMALLOC_START, VMALLOC_END);
arch/s390/include/asm/pgtable.h
111
#define KMSAN_VMALLOC_SIZE (VMALLOC_END - VMALLOC_START)
arch/s390/include/asm/pgtable.h
112
#define KMSAN_VMALLOC_SHADOW_START VMALLOC_END
arch/s390/include/asm/pgtable.h
89
extern unsigned long VMALLOC_END;
arch/s390/kernel/setup.c
168
unsigned long __bootdata_preserved(VMALLOC_END);
arch/s390/kernel/setup.c
169
EXPORT_SYMBOL(VMALLOC_END);
arch/s390/mm/dump_pagetables.c
340
rc |= add_marker(VMALLOC_START, VMALLOC_END, "vmalloc Area");
arch/sh/mm/fault.c
170
#define __FAULT_ADDR_LIMIT VMALLOC_END
arch/sh/mm/init.c
353
(unsigned long)VMALLOC_START, VMALLOC_END,
arch/sh/mm/init.c
354
(VMALLOC_END - VMALLOC_START) >> 20,
arch/sparc/include/asm/pgtable_32.h
427
#define MODULES_END VMALLOC_END
arch/sparc/include/asm/pgtable_64.h
46
#define VMEMMAP_BASE VMALLOC_END
arch/sparc/include/asm/pgtable_64.h
84
extern unsigned long VMALLOC_END;
arch/sparc/mm/init_64.c
1940
unsigned long VMALLOC_END = 0x0000010000000000UL;
arch/sparc/mm/init_64.c
1941
EXPORT_SYMBOL(VMALLOC_END);
arch/sparc/mm/init_64.c
2011
VMALLOC_END = ((sparc64_va_hole_bottom >> 1) +
arch/sparc/mm/init_64.c
2017
VMALLOC_START, VMALLOC_END);
arch/um/include/asm/pgtable.h
56
#define MODULES_END VMALLOC_END
arch/x86/include/asm/pgtable_32_areas.h
48
#define MODULES_END VMALLOC_END
arch/x86/include/asm/pgtable_32_areas.h
51
#define MAXMEM (VMALLOC_END - PAGE_OFFSET - __VMALLOC_RESERVE)
arch/x86/include/asm/pgtable_64_types.h
167
#define KMSAN_MODULES_SHADOW_START (VMALLOC_END + KMSAN_VMALLOC_ORIGIN_OFFSET + 1)
arch/x86/mm/dump_pagetables.c
533
address_markers[VMALLOC_END_NR].start_address = VMALLOC_END;
arch/x86/mm/fault.c
239
if (!(address >= VMALLOC_START && address < VMALLOC_END))
arch/x86/mm/fault.c
270
addr >= TASK_SIZE_MAX && addr < VMALLOC_END;
arch/x86/mm/init_32.c
704
BUILD_BUG_ON(VMALLOC_END > PKMAP_BASE);
arch/x86/mm/init_32.c
707
BUILD_BUG_ON(VMALLOC_START >= VMALLOC_END);
arch/x86/mm/init_32.c
713
BUG_ON(VMALLOC_END > PKMAP_BASE);
arch/x86/mm/init_32.c
715
BUG_ON(VMALLOC_START >= VMALLOC_END);
arch/x86/mm/kasan_init_64.c
402
kasan_mem_to_shadow((void *)VMALLOC_END));
arch/x86/mm/kasan_init_64.c
406
kasan_mem_to_shadow((void *)VMALLOC_END));
arch/x86/mm/kasan_init_64.c
409
kasan_mem_to_shadow((void *)VMALLOC_END + 1),
arch/xtensa/mm/init.c
93
VMALLOC_START, VMALLOC_END,
arch/xtensa/mm/init.c
94
(VMALLOC_END - VMALLOC_START) >> 20,
drivers/md/dm-stats.c
94
if (a > (VMALLOC_END - VMALLOC_START) / DM_STATS_VMALLOC_FACTOR)
drivers/pcmcia/electra_cf.c
216
cf->io_base = (unsigned long)cf->io_virt - VMALLOC_END;
fs/proc/kcore.c
712
if (MODULES_VADDR != VMALLOC_START && MODULES_END != VMALLOC_END) {
fs/proc/kcore.c
734
VMALLOC_END - VMALLOC_START, KCORE_VMALLOC);
include/linux/ioremap.h
16
#define IOREMAP_END VMALLOC_END
include/linux/vmalloc.h
287
#define VMALLOC_TOTAL (VMALLOC_END - VMALLOC_START)
kernel/bpf/core.c
1033
return VMALLOC_END - VMALLOC_START;
kernel/bpf/syscall.c
404
return __vmalloc_node_range(size, align, VMALLOC_START, VMALLOC_END,
kernel/events/core.c
8383
!(virt >= VMALLOC_START && virt < VMALLOC_END))
kernel/liveupdate/kexec_handover.c
1131
vm_flags, VMALLOC_START, VMALLOC_END,
kernel/scs.c
46
s = __vmalloc_node_range(SCS_SIZE, 1, VMALLOC_START, VMALLOC_END,
mm/execmem.c
568
info->ranges[EXECMEM_DEFAULT].end = VMALLOC_END;
mm/kmsan/kmsan.h
184
return ((u64)addr >= VMALLOC_START) && ((u64)addr < VMALLOC_END);
mm/slub.c
6780
return __vmalloc_node_range_noprof(size, align, VMALLOC_START, VMALLOC_END,
mm/vmalloc.c
2001
if (vstart != VMALLOC_START || vend != VMALLOC_END ||
mm/vmalloc.c
2720
VMALLOC_START, VMALLOC_END,
mm/vmalloc.c
3034
BUG_ON(addr > VMALLOC_END);
mm/vmalloc.c
3082
VMALLOC_START, VMALLOC_END,
mm/vmalloc.c
3185
BUG_ON(addr > VMALLOC_END - vm->size);
mm/vmalloc.c
3275
VMALLOC_START, VMALLOC_END,
mm/vmalloc.c
3284
VMALLOC_START, VMALLOC_END,
mm/vmalloc.c
4124
return __vmalloc_node_range_noprof(size, align, VMALLOC_START, VMALLOC_END,
mm/vmalloc.c
4181
return __vmalloc_node_range_noprof(size, 1, VMALLOC_START, VMALLOC_END,
mm/vmalloc.c
4218
return __vmalloc_node_range_noprof(size, SHMLBA, VMALLOC_START, VMALLOC_END,
mm/vmalloc.c
4411
return __vmalloc_node_range_noprof(size, SHMLBA, VMALLOC_START, VMALLOC_END,
mm/vmalloc.c
4844
unsigned long vmalloc_end = VMALLOC_END & ~(align - 1);
mm/vmalloc.c
4888
const unsigned long vmalloc_end = VMALLOC_END & ~(align - 1);
mm/vmalloc.c
83
return addr >= VMALLOC_START && addr < VMALLOC_END;