VMALLOC_START
if (addr >= VMALLOC_START)
if (addr >= VMALLOC_START)
if (addr >= VMALLOC_START)
if (addr >= VMALLOC_START)
if (srm_fixup(VMALLOC_START, (unsigned long)hwrpb))
(VMALLOC_START + (unsigned long)crb->dispatch_va
(VMALLOC_START + (unsigned long)crb->fixup_va
pgd = pgd_offset_k(VMALLOC_START);
p4d = p4d_offset(pgd, VMALLOC_START);
pud = pud_offset(p4d, VMALLOC_START);
pmd = pmd_offset(pud, VMALLOC_START);
brlo sp, VMALLOC_START, 88f
#define VMALLOC_END (VMALLOC_START + VMALLOC_SIZE)
#define USER_KERNEL_GUTTER (VMALLOC_START - TASK_SIZE)
if (address >= VMALLOC_START && !user_mode(regs)) {
if (start < VMALLOC_END && end > VMALLOC_START)
address_markers[4].start_address = VMALLOC_START;
address_markers[2].start_address = VMALLOC_START;
fallback_start = VMALLOC_START;
if ((unsigned long)addr >= VMALLOC_START &&
memcpy_pgd(mm, VMALLOC_START, VMALLOC_END);
arm_kasan_mem_to_shadow(VMALLOC_START);
kasan_populate_early_shadow(kasan_mem_to_shadow((void *)VMALLOC_START),
addr < VMALLOC_START; addr += PMD_SIZE)
for (addr = VMALLOC_START; addr < (FIXADDR_TOP & PMD_MASK); addr += PMD_SIZE)
(md->virtual < VMALLOC_START || md->virtual >= VMALLOC_END)) {
!range_in_range(start, size, VMALLOC_START, VMALLOC_END))
.start = VMALLOC_START,
.start = VMALLOC_START,
BUILD_BUG_ON(VMALLOC_START != MODULES_END);
{ VMALLOC_START, "vmalloc() area" },
return VMALLOC_END - VMALLOC_START;
if (unlikely((addr >= VMALLOC_START) && (addr <= VMALLOC_END))) {
kasan_populate_early_shadow(mem_to_shadow((void *)VMALLOC_START),
#define KMAPAREA(x) ((x >= VMALLOC_START) && (x < KMAP_END))
VMALLOC_START != CKSSEG
if (unlikely(address >= VMALLOC_START && address <= VMALLOC_END))
if (unlikely(address >= VMALLOC_START && address <= VMALLOC_END)) {
if (address >= VMALLOC_START &&
(void*)VMALLOC_START, (void*)VMALLOC_END,
(VMALLOC_END - VMALLOC_START) >> 20,
.start = VMALLOC_START,
#define IOREMAP_START VMALLOC_START
#define IOREMAP_START VMALLOC_START
#define VMALLOC_END (VMALLOC_START + VMALLOC_SIZE)
unsigned long k_start = (unsigned long)kasan_mem_to_shadow((void *)VMALLOC_START);
kasan_remove_zero_shadow((void *)VMALLOC_START, VMALLOC_SIZE);
start = VMALLOC_START;
.start = VMALLOC_START,
.start = VMALLOC_START,
address_markers[1].start_address = VMALLOC_START;
address_markers[i++].start_address = VMALLOC_START;
#define MODULES_VADDR VMALLOC_START
#define VMEMMAP_END VMALLOC_START
#define VMEMMAP_START (VMALLOC_START - VMEMMAP_SIZE)
if (pc < VMALLOC_START)
unlikely(addr >= VMALLOC_START && addr < VMALLOC_END)) {
print_ml("vmalloc", (unsigned long)VMALLOC_START,
preallocate_pgd_pages_range(VMALLOC_START, VMALLOC_END, "vmalloc");
.start = VMALLOC_START,
(void *)kasan_mem_to_shadow((void *)VMALLOC_START));
(void *)kasan_mem_to_shadow((void *)VMALLOC_START),
kasan_populate_early_shadow((void *)kasan_mem_to_shadow((void *)VMALLOC_START),
address_markers[VMALLOC_START_NR].start_address = VMALLOC_START;
unsigned long __bootdata_preserved(VMALLOC_START);
VMALLOC_START = VMALLOC_END - vmalloc_size;
boot_debug("vmalloc area: 0x%016lx-0x%016lx\n", VMALLOC_START, VMALLOC_END);
__memcpy_real_area = round_down(VMALLOC_START - MEMCPY_REAL_SIZE, PAGE_SIZE);
kasan_populate(VMALLOC_START, MODULES_END, POPULATE_KASAN_SHALLOW);
kasan_populate((unsigned long)__identity_va(ident_map_size), VMALLOC_START,
#define KMSAN_VMALLOC_SIZE (VMALLOC_END - VMALLOC_START)
extern unsigned long VMALLOC_START;
unsigned long __bootdata_preserved(VMALLOC_START);
EXPORT_SYMBOL(VMALLOC_START);
rc |= add_marker(VMALLOC_START, VMALLOC_END, "vmalloc Area");
if (!(address >= VMALLOC_START && address < __FAULT_ADDR_LIMIT))
(unsigned long)VMALLOC_START, VMALLOC_END,
(VMALLOC_END - VMALLOC_START) >> 20,
#define MODULES_VADDR VMALLOC_START
VMALLOC_START, VMALLOC_END);
#define MODULES_VADDR VMALLOC_START
start_vm = VMALLOC_START;
#define MODULES_VADDR VMALLOC_START
#define VMEMORY_END (VMALLOC_START + (VMALLOC_SIZE_TB << 40) - 1)
#define VMALLOC_END (VMALLOC_START + VMALLOC_QUARTER_SIZE - 1)
#define KMSAN_VMALLOC_SHADOW_START (VMALLOC_START + KMSAN_VMALLOC_SHADOW_OFFSET)
#define KMSAN_VMALLOC_ORIGIN_START (VMALLOC_START + KMSAN_VMALLOC_ORIGIN_OFFSET)
address_markers[VMALLOC_START_NR].start_address = VMALLOC_START;
address_markers[VMALLOC_START_NR].start_address = VMALLOC_START;
if (!(address >= VMALLOC_START && address < VMALLOC_END))
BUILD_BUG_ON(VMALLOC_START >= VMALLOC_END);
BUG_ON(VMALLOC_START >= VMALLOC_END);
BUG_ON((unsigned long)high_memory > VMALLOC_START);
for (addr = VMALLOC_START; addr <= VMEMORY_END; addr = ALIGN(addr + 1, PGDIR_SIZE)) {
kasan_mem_to_shadow((void *)VMALLOC_START));
kasan_mem_to_shadow((void *)VMALLOC_START),
kasan_mem_to_shadow((void *)VMALLOC_START),
#define VMALLOC_END (VMALLOC_START + 0x07FEFFFF)
#define TLBTEMP_BASE_1 (VMALLOC_START + 0x08000000)
VMALLOC_START, VMALLOC_END,
(VMALLOC_END - VMALLOC_START) >> 20,
BUILD_BUG_ON(VMALLOC_START < KASAN_START_VADDR);
populate(kasan_mem_to_shadow((void *)VMALLOC_START),
if (a > (VMALLOC_END - VMALLOC_START) / DM_STATS_VMALLOC_FACTOR)
if (VMALLOC_START > ent->addr) {
if (VMALLOC_START - ent->addr < ent->size)
ent->size = VMALLOC_START - ent->addr;
if (MODULES_VADDR != VMALLOC_START && MODULES_END != VMALLOC_END) {
kclist_add(&kcore_vmalloc, (void *)VMALLOC_START,
VMALLOC_END - VMALLOC_START, KCORE_VMALLOC);
#define IOREMAP_START VMALLOC_START
#define VMALLOC_TOTAL (VMALLOC_END - VMALLOC_START)
return VMALLOC_END - VMALLOC_START;
return __vmalloc_node_range(size, align, VMALLOC_START, VMALLOC_END,
u64 addr = VMALLOC_START;
!(virt >= VMALLOC_START && virt < VMALLOC_END))
vm_flags, VMALLOC_START, VMALLOC_END,
s = __vmalloc_node_range(SCS_SIZE, 1, VMALLOC_START, VMALLOC_END,
vmcoreinfo_append_str("NUMBER(VMALLOC_START)=0x%lx\n", (unsigned long) VMALLOC_START);
va = (void *)VMALLOC_START;
info->ranges[EXECMEM_DEFAULT].start = VMALLOC_START;
return ((u64)addr >= VMALLOC_START) && ((u64)addr < VMALLOC_END);
off = addr64 - VMALLOC_START;
return __vmalloc_node_range_noprof(size, align, VMALLOC_START, VMALLOC_END,
vstart = VMALLOC_START + rnd;
if (vstart != VMALLOC_START || vend != VMALLOC_END ||
addr -= VMALLOC_START & ~(VMAP_BLOCK_SIZE-1);
VMALLOC_START, VMALLOC_END,
BUG_ON(addr < VMALLOC_START);
VMALLOC_START, VMALLOC_END,
unsigned long addr = ALIGN(VMALLOC_START, align);
VMALLOC_START, VMALLOC_END,
VMALLOC_START, VMALLOC_END,
return __vmalloc_node_range_noprof(size, align, VMALLOC_START, VMALLOC_END,
return __vmalloc_node_range_noprof(size, 1, VMALLOC_START, VMALLOC_END,
return __vmalloc_node_range_noprof(size, SHMLBA, VMALLOC_START, VMALLOC_END,
return __vmalloc_node_range_noprof(size, SHMLBA, VMALLOC_START, VMALLOC_END,
const unsigned long vmalloc_start = ALIGN(VMALLOC_START, align);
return addr >= VMALLOC_START && addr < VMALLOC_END;