Symbol: VMALLOC_START
arch/alpha/kernel/core_marvel.c
779
if (addr >= VMALLOC_START)
arch/alpha/kernel/core_marvel.c
788
if (addr >= VMALLOC_START)
arch/alpha/kernel/core_titan.c
553
if (addr >= VMALLOC_START)
arch/alpha/kernel/core_titan.c
562
if (addr >= VMALLOC_START)
arch/alpha/mm/init.c
131
if (srm_fixup(VMALLOC_START, (unsigned long)hwrpb))
arch/alpha/mm/init.c
136
(VMALLOC_START + (unsigned long)crb->dispatch_va
arch/alpha/mm/init.c
139
(VMALLOC_START + (unsigned long)crb->fixup_va
arch/alpha/mm/init.c
160
pgd = pgd_offset_k(VMALLOC_START);
arch/alpha/mm/init.c
161
p4d = p4d_offset(pgd, VMALLOC_START);
arch/alpha/mm/init.c
162
pud = pud_offset(p4d, VMALLOC_START);
arch/alpha/mm/init.c
164
pmd = pmd_offset(pud, VMALLOC_START);
arch/arc/include/asm/entry-compact.h
151
brlo sp, VMALLOC_START, 88f
arch/arc/include/asm/processor.h
92
#define VMALLOC_END (VMALLOC_START + VMALLOC_SIZE)
arch/arc/include/asm/processor.h
94
#define USER_KERNEL_GUTTER (VMALLOC_START - TASK_SIZE)
arch/arc/mm/fault.c
89
if (address >= VMALLOC_START && !user_mode(regs)) {
arch/arm/kernel/traps.c
961
if (start < VMALLOC_END && end > VMALLOC_START)
arch/arm/mm/dump.c
441
address_markers[4].start_address = VMALLOC_START;
arch/arm/mm/dump.c
443
address_markers[2].start_address = VMALLOC_START;
arch/arm/mm/init.c
456
fallback_start = VMALLOC_START;
arch/arm/mm/iomap.c
40
if ((unsigned long)addr >= VMALLOC_START &&
arch/arm/mm/ioremap.c
145
memcpy_pgd(mm, VMALLOC_START, VMALLOC_END);
arch/arm/mm/ioremap.c
148
arm_kasan_mem_to_shadow(VMALLOC_START);
arch/arm/mm/kasan_init.c
245
kasan_populate_early_shadow(kasan_mem_to_shadow((void *)VMALLOC_START),
arch/arm/mm/mmu.c
1327
addr < VMALLOC_START; addr += PMD_SIZE)
arch/arm/mm/mmu.c
1382
for (addr = VMALLOC_START; addr < (FIXADDR_TOP & PMD_MASK); addr += PMD_SIZE)
arch/arm/mm/mmu.c
987
(md->virtual < VMALLOC_START || md->virtual >= VMALLOC_END)) {
arch/arm/mm/pageattr.c
67
!range_in_range(start, size, VMALLOC_START, VMALLOC_END))
arch/arm64/mm/init.c
538
.start = VMALLOC_START,
arch/arm64/mm/init.c
544
.start = VMALLOC_START,
arch/arm64/mm/kasan_init.c
348
BUILD_BUG_ON(VMALLOC_START != MODULES_END);
arch/arm64/mm/ptdump.c
390
{ VMALLOC_START, "vmalloc() area" },
arch/arm64/net/bpf_jit_comp.c
2282
return VMALLOC_END - VMALLOC_START;
arch/csky/mm/fault.c
214
if (unlikely((addr >= VMALLOC_START) && (addr <= VMALLOC_END))) {
arch/loongarch/mm/kasan_init.c
300
kasan_populate_early_shadow(mem_to_shadow((void *)VMALLOC_START),
arch/m68k/mm/mcfmmu.c
25
#define KMAPAREA(x) ((x >= VMALLOC_START) && (x < KMAP_END))
arch/mips/include/asm/pgtable-64.h
148
VMALLOC_START != CKSSEG
arch/mips/mm/fault.c
84
if (unlikely(address >= VMALLOC_START && address <= VMALLOC_END))
arch/nios2/mm/fault.c
67
if (unlikely(address >= VMALLOC_START && address <= VMALLOC_END)) {
arch/openrisc/mm/fault.c
77
if (address >= VMALLOC_START &&
arch/parisc/mm/init.c
587
(void*)VMALLOC_START, (void*)VMALLOC_END,
arch/parisc/mm/init.c
588
(VMALLOC_END - VMALLOC_START) >> 20,
arch/parisc/mm/init.c
990
.start = VMALLOC_START,
arch/powerpc/include/asm/book3s/32/pgtable.h
168
#define IOREMAP_START VMALLOC_START
arch/powerpc/include/asm/nohash/32/pgtable.h
83
#define IOREMAP_START VMALLOC_START
arch/powerpc/include/asm/nohash/64/pgtable.h
38
#define VMALLOC_END (VMALLOC_START + VMALLOC_SIZE)
arch/powerpc/mm/kasan/init_32.c
114
unsigned long k_start = (unsigned long)kasan_mem_to_shadow((void *)VMALLOC_START);
arch/powerpc/mm/kasan/init_book3e_64.c
118
kasan_remove_zero_shadow((void *)VMALLOC_START, VMALLOC_SIZE);
arch/powerpc/mm/mem.c
422
start = VMALLOC_START;
arch/powerpc/mm/mem.c
439
.start = VMALLOC_START,
arch/powerpc/mm/mem.c
445
.start = VMALLOC_START,
arch/powerpc/mm/ptdump/hashpagetable.c
510
address_markers[1].start_address = VMALLOC_START;
arch/powerpc/mm/ptdump/ptdump.c
271
address_markers[i++].start_address = VMALLOC_START;
arch/riscv/include/asm/pgtable.h
63
#define MODULES_VADDR VMALLOC_START
arch/riscv/include/asm/pgtable.h
87
#define VMEMMAP_END VMALLOC_START
arch/riscv/include/asm/pgtable.h
88
#define VMEMMAP_START (VMALLOC_START - VMEMMAP_SIZE)
arch/riscv/kernel/traps.c
454
if (pc < VMALLOC_START)
arch/riscv/mm/fault.c
312
unlikely(addr >= VMALLOC_START && addr < VMALLOC_END)) {
arch/riscv/mm/init.c
148
print_ml("vmalloc", (unsigned long)VMALLOC_START,
arch/riscv/mm/init.c
1520
preallocate_pgd_pages_range(VMALLOC_START, VMALLOC_END, "vmalloc");
arch/riscv/mm/init.c
1547
.start = VMALLOC_START,
arch/riscv/mm/kasan_init.c
495
(void *)kasan_mem_to_shadow((void *)VMALLOC_START));
arch/riscv/mm/kasan_init.c
499
(void *)kasan_mem_to_shadow((void *)VMALLOC_START),
arch/riscv/mm/kasan_init.c
506
kasan_populate_early_shadow((void *)kasan_mem_to_shadow((void *)VMALLOC_START),
arch/riscv/mm/ptdump.c
434
address_markers[VMALLOC_START_NR].start_address = VMALLOC_START;
arch/s390/boot/startup.c
34
unsigned long __bootdata_preserved(VMALLOC_START);
arch/s390/boot/startup.c
413
VMALLOC_START = VMALLOC_END - vmalloc_size;
arch/s390/boot/startup.c
414
boot_debug("vmalloc area: 0x%016lx-0x%016lx\n", VMALLOC_START, VMALLOC_END);
arch/s390/boot/startup.c
416
__memcpy_real_area = round_down(VMALLOC_START - MEMCPY_REAL_SIZE, PAGE_SIZE);
arch/s390/boot/vmem.c
130
kasan_populate(VMALLOC_START, MODULES_END, POPULATE_KASAN_SHALLOW);
arch/s390/boot/vmem.c
132
kasan_populate((unsigned long)__identity_va(ident_map_size), VMALLOC_START,
arch/s390/include/asm/pgtable.h
111
#define KMSAN_VMALLOC_SIZE (VMALLOC_END - VMALLOC_START)
arch/s390/include/asm/pgtable.h
88
extern unsigned long VMALLOC_START;
arch/s390/kernel/setup.c
165
unsigned long __bootdata_preserved(VMALLOC_START);
arch/s390/kernel/setup.c
166
EXPORT_SYMBOL(VMALLOC_START);
arch/s390/mm/dump_pagetables.c
340
rc |= add_marker(VMALLOC_START, VMALLOC_END, "vmalloc Area");
arch/sh/mm/fault.c
183
if (!(address >= VMALLOC_START && address < __FAULT_ADDR_LIMIT))
arch/sh/mm/init.c
353
(unsigned long)VMALLOC_START, VMALLOC_END,
arch/sh/mm/init.c
354
(VMALLOC_END - VMALLOC_START) >> 20,
arch/sparc/include/asm/pgtable_32.h
426
#define MODULES_VADDR VMALLOC_START
arch/sparc/mm/init_64.c
2017
VMALLOC_START, VMALLOC_END);
arch/um/include/asm/pgtable.h
55
#define MODULES_VADDR VMALLOC_START
arch/um/kernel/um_arch.c
381
start_vm = VMALLOC_START;
arch/x86/include/asm/pgtable_32_areas.h
47
#define MODULES_VADDR VMALLOC_START
arch/x86/include/asm/pgtable_64_types.h
130
#define VMEMORY_END (VMALLOC_START + (VMALLOC_SIZE_TB << 40) - 1)
arch/x86/include/asm/pgtable_64_types.h
151
#define VMALLOC_END (VMALLOC_START + VMALLOC_QUARTER_SIZE - 1)
arch/x86/include/asm/pgtable_64_types.h
160
#define KMSAN_VMALLOC_SHADOW_START (VMALLOC_START + KMSAN_VMALLOC_SHADOW_OFFSET)
arch/x86/include/asm/pgtable_64_types.h
161
#define KMSAN_VMALLOC_ORIGIN_START (VMALLOC_START + KMSAN_VMALLOC_ORIGIN_OFFSET)
arch/x86/mm/dump_pagetables.c
521
address_markers[VMALLOC_START_NR].start_address = VMALLOC_START;
arch/x86/mm/dump_pagetables.c
532
address_markers[VMALLOC_START_NR].start_address = VMALLOC_START;
arch/x86/mm/fault.c
239
if (!(address >= VMALLOC_START && address < VMALLOC_END))
arch/x86/mm/init_32.c
707
BUILD_BUG_ON(VMALLOC_START >= VMALLOC_END);
arch/x86/mm/init_32.c
715
BUG_ON(VMALLOC_START >= VMALLOC_END);
arch/x86/mm/init_32.c
716
BUG_ON((unsigned long)high_memory > VMALLOC_START);
arch/x86/mm/init_64.c
1323
for (addr = VMALLOC_START; addr <= VMEMORY_END; addr = ALIGN(addr + 1, PGDIR_SIZE)) {
arch/x86/mm/kasan_init_64.c
392
kasan_mem_to_shadow((void *)VMALLOC_START));
arch/x86/mm/kasan_init_64.c
401
kasan_mem_to_shadow((void *)VMALLOC_START),
arch/x86/mm/kasan_init_64.c
405
kasan_mem_to_shadow((void *)VMALLOC_START),
arch/xtensa/include/asm/pgtable.h
68
#define VMALLOC_END (VMALLOC_START + 0x07FEFFFF)
arch/xtensa/include/asm/pgtable.h
69
#define TLBTEMP_BASE_1 (VMALLOC_START + 0x08000000)
arch/xtensa/mm/init.c
93
VMALLOC_START, VMALLOC_END,
arch/xtensa/mm/init.c
94
(VMALLOC_END - VMALLOC_START) >> 20,
arch/xtensa/mm/kasan_init.c
75
BUILD_BUG_ON(VMALLOC_START < KASAN_START_VADDR);
arch/xtensa/mm/kasan_init.c
81
populate(kasan_mem_to_shadow((void *)VMALLOC_START),
drivers/md/dm-stats.c
94
if (a > (VMALLOC_END - VMALLOC_START) / DM_STATS_VMALLOC_FACTOR)
fs/proc/kcore.c
231
if (VMALLOC_START > ent->addr) {
fs/proc/kcore.c
232
if (VMALLOC_START - ent->addr < ent->size)
fs/proc/kcore.c
233
ent->size = VMALLOC_START - ent->addr;
fs/proc/kcore.c
712
if (MODULES_VADDR != VMALLOC_START && MODULES_END != VMALLOC_END) {
fs/proc/kcore.c
733
kclist_add(&kcore_vmalloc, (void *)VMALLOC_START,
fs/proc/kcore.c
734
VMALLOC_END - VMALLOC_START, KCORE_VMALLOC);
include/linux/ioremap.h
15
#define IOREMAP_START VMALLOC_START
include/linux/vmalloc.h
287
#define VMALLOC_TOTAL (VMALLOC_END - VMALLOC_START)
kernel/bpf/core.c
1033
return VMALLOC_END - VMALLOC_START;
kernel/bpf/syscall.c
404
return __vmalloc_node_range(size, align, VMALLOC_START, VMALLOC_END,
kernel/bpf/verifier.c
22861
u64 addr = VMALLOC_START;
kernel/events/core.c
8382
!(virt >= VMALLOC_START && virt < VMALLOC_END))
kernel/liveupdate/kexec_handover.c
1131
vm_flags, VMALLOC_START, VMALLOC_END,
kernel/scs.c
46
s = __vmalloc_node_range(SCS_SIZE, 1, VMALLOC_START, VMALLOC_END,
kernel/vmcore_info.c
172
vmcoreinfo_append_str("NUMBER(VMALLOC_START)=0x%lx\n", (unsigned long) VMALLOC_START);
lib/test_debug_virtual.c
27
va = (void *)VMALLOC_START;
mm/execmem.c
567
info->ranges[EXECMEM_DEFAULT].start = VMALLOC_START;
mm/kmsan/kmsan.h
184
return ((u64)addr >= VMALLOC_START) && ((u64)addr < VMALLOC_END);
mm/kmsan/shadow.c
61
off = addr64 - VMALLOC_START;
mm/slub.c
6780
return __vmalloc_node_range_noprof(size, align, VMALLOC_START, VMALLOC_END,
mm/vmalloc.c
1661
vstart = VMALLOC_START + rnd;
mm/vmalloc.c
2001
if (vstart != VMALLOC_START || vend != VMALLOC_END ||
mm/vmalloc.c
2681
addr -= VMALLOC_START & ~(VMAP_BLOCK_SIZE-1);
mm/vmalloc.c
2720
VMALLOC_START, VMALLOC_END,
mm/vmalloc.c
3033
BUG_ON(addr < VMALLOC_START);
mm/vmalloc.c
3082
VMALLOC_START, VMALLOC_END,
mm/vmalloc.c
3174
unsigned long addr = ALIGN(VMALLOC_START, align);
mm/vmalloc.c
3275
VMALLOC_START, VMALLOC_END,
mm/vmalloc.c
3284
VMALLOC_START, VMALLOC_END,
mm/vmalloc.c
4124
return __vmalloc_node_range_noprof(size, align, VMALLOC_START, VMALLOC_END,
mm/vmalloc.c
4181
return __vmalloc_node_range_noprof(size, 1, VMALLOC_START, VMALLOC_END,
mm/vmalloc.c
4218
return __vmalloc_node_range_noprof(size, SHMLBA, VMALLOC_START, VMALLOC_END,
mm/vmalloc.c
4411
return __vmalloc_node_range_noprof(size, SHMLBA, VMALLOC_START, VMALLOC_END,
mm/vmalloc.c
4887
const unsigned long vmalloc_start = ALIGN(VMALLOC_START, align);
mm/vmalloc.c
83
return addr >= VMALLOC_START && addr < VMALLOC_END;