VMALLOC_END
BUILD_BUG_ON(PAGE_OFFSET < (VMALLOC_END + FIXMAP_SIZE + PKMAP_SIZE));
if (start < VMALLOC_END && end > VMALLOC_START)
fallback_end = VMALLOC_END;
(unsigned long)addr < VMALLOC_END)
memcpy_pgd(mm, VMALLOC_START, VMALLOC_END);
arm_kasan_mem_to_shadow(VMALLOC_END);
kasan_mem_to_shadow((void *)VMALLOC_END));
kasan_populate_early_shadow(kasan_mem_to_shadow((void *)VMALLOC_END),
vmalloc_max = VMALLOC_END - (PAGE_OFFSET + SZ_32M + VMALLOC_OFFSET);
vmalloc_limit = (u64)VMALLOC_END - vmalloc_size - VMALLOC_OFFSET -
(md->virtual < VMALLOC_START || md->virtual >= VMALLOC_END)) {
!range_in_range(start, size, VMALLOC_START, VMALLOC_END))
range = (VMALLOC_END - KIMAGE_VADDR) / 2;
vmcoreinfo_append_str("NUMBER(VMALLOC_END)=0x%lx\n", VMALLOC_END);
.end = VMALLOC_END,
.end = VMALLOC_END,
vmalloc_shadow_end = (u64)kasan_mem_to_shadow((void *)VMALLOC_END);
{ VMALLOC_END, "vmalloc() end" },
return VMALLOC_END - VMALLOC_START;
if (unlikely((addr >= VMALLOC_START) && (addr <= VMALLOC_END))) {
#define vmemmap ((struct page *)((VMALLOC_END + PMD_SIZE) & PMD_MASK))
if (unlikely(address >= VMALLOC_START && address <= VMALLOC_END))
if (unlikely(address >= VMALLOC_START && address <= VMALLOC_END)) {
(void*)VMALLOC_START, (void*)VMALLOC_END,
(VMALLOC_END - VMALLOC_START) >> 20,
.end = VMALLOC_END,
#define IOREMAP_END VMALLOC_END
#define IOREMAP_END VMALLOC_END
#define VMEMMAP_BASE VMALLOC_END
} else if (ea < VMALLOC_END) {
unsigned long k_end = (unsigned long)kasan_mem_to_shadow((void *)VMALLOC_END);
end = VMALLOC_END;
.end = VMALLOC_END,
.end = VMALLOC_END,
if (addr < VMALLOC_END)
address_markers[2].start_address = VMALLOC_END;
address_markers[i++].start_address = VMALLOC_END;
#define BPF_JIT_REGION_END (VMALLOC_END)
#define MODULES_END VMALLOC_END
vmcoreinfo_append_str("NUMBER(VMALLOC_END)=0x%lx\n", VMALLOC_END);
unlikely(addr >= VMALLOC_START && addr < VMALLOC_END)) {
(unsigned long)VMALLOC_END);
preallocate_pgd_pages_range(VMALLOC_START, VMALLOC_END, "vmalloc");
.end = VMALLOC_END,
(void *)kasan_mem_to_shadow((void *)VMALLOC_END));
(void *)kasan_mem_to_shadow((void *)VMALLOC_END));
address_markers[VMALLOC_END_NR].start_address = VMALLOC_END;
unsigned long __bootdata_preserved(VMALLOC_END);
VMALLOC_END = MODULES_VADDR;
VMALLOC_END -= MODULES_LEN * 2;
vsize = (VMALLOC_END - FIXMAP_SIZE) / 2;
VMALLOC_END -= vmalloc_size * 2;
VMALLOC_START = VMALLOC_END - vmalloc_size;
boot_debug("vmalloc area: 0x%016lx-0x%016lx\n", VMALLOC_START, VMALLOC_END);
#define KMSAN_VMALLOC_SIZE (VMALLOC_END - VMALLOC_START)
#define KMSAN_VMALLOC_SHADOW_START VMALLOC_END
extern unsigned long VMALLOC_END;
unsigned long __bootdata_preserved(VMALLOC_END);
EXPORT_SYMBOL(VMALLOC_END);
rc |= add_marker(VMALLOC_START, VMALLOC_END, "vmalloc Area");
#define __FAULT_ADDR_LIMIT VMALLOC_END
(unsigned long)VMALLOC_START, VMALLOC_END,
(VMALLOC_END - VMALLOC_START) >> 20,
#define MODULES_END VMALLOC_END
#define VMEMMAP_BASE VMALLOC_END
extern unsigned long VMALLOC_END;
unsigned long VMALLOC_END = 0x0000010000000000UL;
EXPORT_SYMBOL(VMALLOC_END);
VMALLOC_END = ((sparc64_va_hole_bottom >> 1) +
VMALLOC_START, VMALLOC_END);
#define MODULES_END VMALLOC_END
#define MODULES_END VMALLOC_END
#define MAXMEM (VMALLOC_END - PAGE_OFFSET - __VMALLOC_RESERVE)
#define KMSAN_MODULES_SHADOW_START (VMALLOC_END + KMSAN_VMALLOC_ORIGIN_OFFSET + 1)
address_markers[VMALLOC_END_NR].start_address = VMALLOC_END;
if (!(address >= VMALLOC_START && address < VMALLOC_END))
addr >= TASK_SIZE_MAX && addr < VMALLOC_END;
BUILD_BUG_ON(VMALLOC_END > PKMAP_BASE);
BUILD_BUG_ON(VMALLOC_START >= VMALLOC_END);
BUG_ON(VMALLOC_END > PKMAP_BASE);
BUG_ON(VMALLOC_START >= VMALLOC_END);
kasan_mem_to_shadow((void *)VMALLOC_END));
kasan_mem_to_shadow((void *)VMALLOC_END));
kasan_mem_to_shadow((void *)VMALLOC_END + 1),
VMALLOC_START, VMALLOC_END,
(VMALLOC_END - VMALLOC_START) >> 20,
if (a > (VMALLOC_END - VMALLOC_START) / DM_STATS_VMALLOC_FACTOR)
cf->io_base = (unsigned long)cf->io_virt - VMALLOC_END;
if (MODULES_VADDR != VMALLOC_START && MODULES_END != VMALLOC_END) {
VMALLOC_END - VMALLOC_START, KCORE_VMALLOC);
#define IOREMAP_END VMALLOC_END
#define VMALLOC_TOTAL (VMALLOC_END - VMALLOC_START)
return VMALLOC_END - VMALLOC_START;
return __vmalloc_node_range(size, align, VMALLOC_START, VMALLOC_END,
!(virt >= VMALLOC_START && virt < VMALLOC_END))
vm_flags, VMALLOC_START, VMALLOC_END,
s = __vmalloc_node_range(SCS_SIZE, 1, VMALLOC_START, VMALLOC_END,
info->ranges[EXECMEM_DEFAULT].end = VMALLOC_END;
return ((u64)addr >= VMALLOC_START) && ((u64)addr < VMALLOC_END);
return __vmalloc_node_range_noprof(size, align, VMALLOC_START, VMALLOC_END,
if (vstart != VMALLOC_START || vend != VMALLOC_END ||
VMALLOC_START, VMALLOC_END,
BUG_ON(addr > VMALLOC_END);
VMALLOC_START, VMALLOC_END,
BUG_ON(addr > VMALLOC_END - vm->size);
VMALLOC_START, VMALLOC_END,
VMALLOC_START, VMALLOC_END,
return __vmalloc_node_range_noprof(size, align, VMALLOC_START, VMALLOC_END,
return __vmalloc_node_range_noprof(size, 1, VMALLOC_START, VMALLOC_END,
return __vmalloc_node_range_noprof(size, SHMLBA, VMALLOC_START, VMALLOC_END,
return __vmalloc_node_range_noprof(size, SHMLBA, VMALLOC_START, VMALLOC_END,
unsigned long vmalloc_end = VMALLOC_END & ~(align - 1);
const unsigned long vmalloc_end = VMALLOC_END & ~(align - 1);
return addr >= VMALLOC_START && addr < VMALLOC_END;