Symbol: host_mmu
arch/arm64/kvm/hyp/include/nvhe/mem_protect.h
24
extern struct host_mmu host_mmu;
arch/arm64/kvm/hyp/include/nvhe/mem_protect.h
66
__load_stage2(&host_mmu.arch.mmu, &host_mmu.arch);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
101
kvm_pgtable_stage2_free_unlinked(&host_mmu.mm_ops, addr, level);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
115
host_mmu.mm_ops = (struct kvm_pgtable_mm_ops) {
arch/arm64/kvm/hyp/nvhe/mem_protect.c
1235
selftest_vm.kvm.arch.mmu.vtcr = host_mmu.arch.mmu.vtcr;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
137
host_mmu.arch.mmu.vtcr = kvm_get_vtcr(id_aa64mmfr0_el1_sys_val,
arch/arm64/kvm/hyp/nvhe/mem_protect.c
145
struct kvm_s2_mmu *mmu = &host_mmu.arch.mmu;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
149
hyp_spin_lock_init(&host_mmu.lock);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
150
mmu->arch = &host_mmu.arch;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
156
ret = __kvm_pgtable_stage2_init(&host_mmu.pgt, mmu,
arch/arm64/kvm/hyp/nvhe/mem_protect.c
157
&host_mmu.mm_ops, KVM_HOST_S2_FLAGS,
arch/arm64/kvm/hyp/nvhe/mem_protect.c
162
mmu->pgd_phys = __hyp_pa(host_mmu.pgt.pgd);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
163
mmu->pgt = &host_mmu.pgt;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
24
struct host_mmu host_mmu;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
318
struct kvm_s2_mmu *mmu = &host_mmu.arch.mmu;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
339
__load_stage2(&host_mmu.arch.mmu, &host_mmu.arch);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
357
struct kvm_pgtable *pgt = &host_mmu.pgt;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
380
u64 limit = BIT(kvm_phys_shift(&host_mmu.arch.mmu) - PAGE_SHIFT);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
45
hyp_spin_lock(&host_mmu.lock);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
464
return kvm_pgtable_stage2_map(&host_mmu.pgt, start, end - start, start,
arch/arm64/kvm/hyp/nvhe/mem_protect.c
477
hyp_assert_lock_held(&host_mmu.lock); \
arch/arm64/kvm/hyp/nvhe/mem_protect.c
50
hyp_spin_unlock(&host_mmu.lock);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
501
hyp_assert_lock_held(&host_mmu.lock);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
502
ret = kvm_pgtable_get_leaf(&host_mmu.pgt, addr, &pte, &level);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
551
ret = host_stage2_try(kvm_pgtable_stage2_set_owner, &host_mmu.pgt,
arch/arm64/kvm/hyp/nvhe/mem_protect.c
669
hyp_assert_lock_held(&host_mmu.lock);
arch/arm64/kvm/hyp/nvhe/pkvm.c
434
mmu->vtcr = host_mmu.arch.mmu.vtcr;
arch/arm64/kvm/hyp/nvhe/pkvm.c
763
pgd_size = kvm_pgtable_stage2_pgd_size(host_mmu.arch.mmu.vtcr);
arch/arm64/kvm/hyp/nvhe/tlb.c
23
struct kvm_s2_mmu *host_s2_mmu = &host_mmu.arch.mmu;