VA_BITS
#if VA_BITS > 48
#define vabits_actual ((u64)VA_BITS)
#define PAGE_OFFSET (_PAGE_OFFSET(VA_BITS))
#if VA_BITS > 48
#define VA_BITS_MIN (VA_BITS)
#define PTRS_PER_PGD (1 << (VA_BITS - PGDIR_SHIFT))
return vabits_actual == VA_BITS;
#if VA_BITS == VA_BITS_MIN
return vabits_actual == VA_BITS;
#define TASK_SIZE_MAX (UL(1) << VA_BITS)
int va_bits = VA_BITS;
VMCOREINFO_NUMBER(VA_BITS);
if ((base ^ io_map_base) & BIT(VA_BITS - 1))
BUILD_BUG_ON(!IS_ALIGNED(_KASAN_SHADOW_START(VA_BITS), SHADOW_ALIGN));
u64 vabits = IS_ENABLED(CONFIG_ARM64_64K_PAGES) ? VA_BITS
#define XRANGE_SHADOW_SHIFT min(cpu_vabits, VA_BITS)
#define TASK_SIZE64 (0x1UL << ((cpu_vabits > VA_BITS) ? VA_BITS : cpu_vabits))
#define ESID_BITS (VA_BITS - (SID_SHIFT + CONTEXT_BITS))
#define ESID_BITS_1T (VA_BITS - (SID_SHIFT_1T + CONTEXT_BITS))
#define VSID_BITS_256M (VA_BITS - SID_SHIFT)
#define VSID_BITS_1T (VA_BITS - SID_SHIFT_1T)
unsigned long va_bits = VA_BITS;
unsigned long va_bits = VA_BITS;
#define KASAN_SHADOW_SIZE (UL(1) << ((VA_BITS - 1) - KASAN_SHADOW_SCALE_SHIFT))
#define KERN_VIRT_START (-(BIT(VA_BITS)) + TASK_SIZE)
#define MMAP_VA_BITS_64 ((VA_BITS >= VA_BITS_SV48) ? VA_BITS_SV48 : VA_BITS)
(VA_BITS - PAGE_SHIFT - 1 + STRUCT_PAGE_MAX_SHIFT)
VMCOREINFO_NUMBER(VA_BITS);
current->comm, PAGE_SIZE / SZ_1K, VA_BITS,
#define VPTE_SIZE (1UL << (VA_BITS - PAGE_SHIFT + 3))
#define VPTE_SIZE (1 << (VA_BITS - PAGE_SHIFT + 3))
pgtbl_cfg.ias = min_t(unsigned long, ias, VA_BITS);