Symbol: vabits_actual
arch/arm64/include/asm/memory.h
105
#define KASAN_SHADOW_START _KASAN_SHADOW_START(vabits_actual)
arch/arm64/include/asm/mmu_context.h
98
__cpu_set_tcr_t0sz(TCR_T0SZ(vabits_actual));
arch/arm64/include/asm/pgtable.h
1076
return vabits_actual == VA_BITS;
arch/arm64/include/asm/pgtable.h
27
#define VMEMMAP_UNUSED_NPAGES ((_PAGE_OFFSET(vabits_actual) - PAGE_OFFSET) >> PAGE_SHIFT)
arch/arm64/include/asm/pgtable.h
948
return vabits_actual == VA_BITS;
arch/arm64/include/asm/pointer_auth.h
17
#define ptrauth_user_pac_mask() GENMASK_ULL(54, vabits_actual)
arch/arm64/include/asm/pointer_auth.h
18
#define ptrauth_kernel_pac_mask() GENMASK_ULL(63, vabits_actual)
arch/arm64/include/asm/processor.h
56
#define TASK_SIZE_64 (UL(1) << vabits_actual)
arch/arm64/kvm/mmu.c
823
.ia_bits = vabits_actual,
arch/arm64/kvm/va_layout.c
68
return max(IDMAP_VA_BITS, vabits_actual);
arch/arm64/mm/fault.c
153
vabits_actual, mm_to_pgd_phys(mm));
arch/arm64/mm/init.c
195
s64 linear_region_size = PAGE_END - _PAGE_OFFSET(vabits_actual);
arch/arm64/mm/init.c
205
if (IS_ENABLED(CONFIG_KVM) && vabits_actual == 52 &&
arch/arm64/mm/init.c
244
if (IS_ENABLED(CONFIG_ARM64_VA_BITS_52) && (vabits_actual != 52))
arch/arm64/mm/init.c
245
memstart_addr -= _PAGE_OFFSET(vabits_actual) - _PAGE_OFFSET(52);
arch/arm64/mm/kasan_init.c
193
int shift = (ARM64_HW_PGTABLE_LEVELS(vabits_actual) - 1) * PTDESC_TABLE_SHIFT;
arch/arm64/mm/kasan_init.c
247
: vabits_actual;
arch/arm64/mm/kasan_init.c
272
int shift = (ARM64_HW_PGTABLE_LEVELS(vabits_actual) - 2) * PTDESC_TABLE_SHIFT;
arch/arm64/mm/mmu.c
1940
phys_addr_t start_linear_pa = __pa(_PAGE_OFFSET(vabits_actual));
arch/arm64/mm/mmu.c
913
unsigned long lstart = _PAGE_OFFSET(vabits_actual);
arch/arm64/mm/ptdump.c
357
{_PAGE_OFFSET(vabits_actual), ~0UL},
arch/arm64/mm/ptdump.c
379
u64 page_offset = _PAGE_OFFSET(vabits_actual);
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-sva.c
216
if (vabits_actual == 52) {
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-sva.c
85
64ULL - vabits_actual) |