Symbol: kvm_s2_mmu
arch/arm64/include/asm/kvm_asm.h
221
struct kvm_s2_mmu;
arch/arm64/include/asm/kvm_asm.h
236
extern void __kvm_flush_cpu_context(struct kvm_s2_mmu *mmu);
arch/arm64/include/asm/kvm_asm.h
237
extern void __kvm_tlb_flush_vmid_ipa(struct kvm_s2_mmu *mmu, phys_addr_t ipa,
arch/arm64/include/asm/kvm_asm.h
239
extern void __kvm_tlb_flush_vmid_ipa_nsh(struct kvm_s2_mmu *mmu,
arch/arm64/include/asm/kvm_asm.h
242
extern void __kvm_tlb_flush_vmid_range(struct kvm_s2_mmu *mmu,
arch/arm64/include/asm/kvm_asm.h
244
extern void __kvm_tlb_flush_vmid(struct kvm_s2_mmu *mmu);
arch/arm64/include/asm/kvm_asm.h
246
extern int __kvm_tlbi_s1e2(struct kvm_s2_mmu *mmu, u64 va, u64 sys_encoding);
arch/arm64/include/asm/kvm_host.h
296
struct kvm_s2_mmu mmu;
arch/arm64/include/asm/kvm_host.h
311
struct kvm_s2_mmu *nested_mmus;
arch/arm64/include/asm/kvm_host.h
842
struct kvm_s2_mmu *hw_mmu;
arch/arm64/include/asm/kvm_mmu.h
172
void kvm_stage2_unmap_range(struct kvm_s2_mmu *mmu, phys_addr_t start,
arch/arm64/include/asm/kvm_mmu.h
174
void kvm_stage2_flush_range(struct kvm_s2_mmu *mmu, phys_addr_t addr, phys_addr_t end);
arch/arm64/include/asm/kvm_mmu.h
175
void kvm_stage2_wp_range(struct kvm_s2_mmu *mmu, phys_addr_t addr, phys_addr_t end);
arch/arm64/include/asm/kvm_mmu.h
178
int kvm_init_stage2_mmu(struct kvm *kvm, struct kvm_s2_mmu *mmu, unsigned long type);
arch/arm64/include/asm/kvm_mmu.h
180
void kvm_free_stage2_pgd(struct kvm_s2_mmu *mmu);
arch/arm64/include/asm/kvm_mmu.h
305
static __always_inline u64 kvm_get_vttbr(struct kvm_s2_mmu *mmu)
arch/arm64/include/asm/kvm_mmu.h
321
static __always_inline void __load_stage2(struct kvm_s2_mmu *mmu,
arch/arm64/include/asm/kvm_mmu.h
335
static inline struct kvm *kvm_s2_mmu_to_kvm(struct kvm_s2_mmu *mmu)
arch/arm64/include/asm/kvm_mmu.h
346
static inline bool kvm_s2_mmu_valid(struct kvm_s2_mmu *mmu)
arch/arm64/include/asm/kvm_mmu.h
351
static inline bool kvm_is_nested_s2_mmu(struct kvm *kvm, struct kvm_s2_mmu *mmu)
arch/arm64/include/asm/kvm_nested.h
170
unsigned long compute_tlb_inval_range(struct kvm_s2_mmu *mmu, u64 val);
arch/arm64/include/asm/kvm_nested.h
70
extern void kvm_init_nested_s2_mmu(struct kvm_s2_mmu *mmu);
arch/arm64/include/asm/kvm_nested.h
71
extern struct kvm_s2_mmu *lookup_s2_mmu(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_nested.h
77
void (*)(struct kvm_s2_mmu *,
arch/arm64/include/asm/kvm_pgtable.h
449
struct kvm_s2_mmu *mmu;
arch/arm64/include/asm/kvm_pgtable.h
550
int __kvm_pgtable_stage2_init(struct kvm_pgtable *pgt, struct kvm_s2_mmu *mmu,
arch/arm64/include/asm/kvm_pgtable.h
555
static inline int kvm_pgtable_stage2_init(struct kvm_pgtable *pgt, struct kvm_s2_mmu *mmu,
arch/arm64/include/asm/kvm_pgtable.h
880
void kvm_tlb_flush_vmid_range(struct kvm_s2_mmu *mmu,
arch/arm64/include/asm/kvm_pkvm.h
205
int pkvm_pgtable_stage2_init(struct kvm_pgtable *pgt, struct kvm_s2_mmu *mmu,
arch/arm64/kvm/arm.c
632
struct kvm_s2_mmu *mmu;
arch/arm64/kvm/at.c
1329
struct kvm_s2_mmu *mmu;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
391
DECLARE_REG(struct kvm_s2_mmu *, mmu, host_ctxt, 1);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
400
DECLARE_REG(struct kvm_s2_mmu *, mmu, host_ctxt, 1);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
410
DECLARE_REG(struct kvm_s2_mmu *, mmu, host_ctxt, 1);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
419
DECLARE_REG(struct kvm_s2_mmu *, mmu, host_ctxt, 1);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
442
DECLARE_REG(struct kvm_s2_mmu *, mmu, host_ctxt, 1);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
145
struct kvm_s2_mmu *mmu = &host_mmu.arch.mmu;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
259
struct kvm_s2_mmu *mmu = &vm->kvm.arch.mmu;
arch/arm64/kvm/hyp/nvhe/mem_protect.c
318
struct kvm_s2_mmu *mmu = &host_mmu.arch.mmu;
arch/arm64/kvm/hyp/nvhe/pkvm.c
419
struct kvm_s2_mmu *mmu = &hyp_vm->kvm.arch.mmu;
arch/arm64/kvm/hyp/nvhe/switch.c
246
struct kvm_s2_mmu *mmu;
arch/arm64/kvm/hyp/nvhe/tlb.c
120
struct kvm_s2_mmu *mmu = cxt->mmu;
arch/arm64/kvm/hyp/nvhe/tlb.c
14
struct kvm_s2_mmu *mmu;
arch/arm64/kvm/hyp/nvhe/tlb.c
148
void __kvm_tlb_flush_vmid_ipa(struct kvm_s2_mmu *mmu,
arch/arm64/kvm/hyp/nvhe/tlb.c
178
void __kvm_tlb_flush_vmid_ipa_nsh(struct kvm_s2_mmu *mmu,
arch/arm64/kvm/hyp/nvhe/tlb.c
19
static void enter_vmid_context(struct kvm_s2_mmu *mmu,
arch/arm64/kvm/hyp/nvhe/tlb.c
208
void __kvm_tlb_flush_vmid_range(struct kvm_s2_mmu *mmu,
arch/arm64/kvm/hyp/nvhe/tlb.c
23
struct kvm_s2_mmu *host_s2_mmu = &host_mmu.arch.mmu;
arch/arm64/kvm/hyp/nvhe/tlb.c
235
void __kvm_tlb_flush_vmid(struct kvm_s2_mmu *mmu)
arch/arm64/kvm/hyp/nvhe/tlb.c
249
void __kvm_flush_cpu_context(struct kvm_s2_mmu *mmu)
arch/arm64/kvm/hyp/pgtable.c
1150
struct kvm_s2_mmu *mmu = pgt->mmu;
arch/arm64/kvm/hyp/pgtable.c
1485
struct kvm_s2_mmu *mmu;
arch/arm64/kvm/hyp/pgtable.c
1524
mmu = container_of(mc, struct kvm_s2_mmu, split_page_cache);
arch/arm64/kvm/hyp/pgtable.c
1563
int __kvm_pgtable_stage2_init(struct kvm_pgtable *pgt, struct kvm_s2_mmu *mmu,
arch/arm64/kvm/hyp/pgtable.c
589
struct kvm_s2_mmu *mmu;
arch/arm64/kvm/hyp/pgtable.c
650
void kvm_tlb_flush_vmid_range(struct kvm_s2_mmu *mmu,
arch/arm64/kvm/hyp/pgtable.c
829
struct kvm_s2_mmu *mmu)
arch/arm64/kvm/hyp/pgtable.c
893
struct kvm_s2_mmu *mmu,
arch/arm64/kvm/hyp/vhe/tlb.c
124
void __kvm_tlb_flush_vmid_ipa_nsh(struct kvm_s2_mmu *mmu,
arch/arm64/kvm/hyp/vhe/tlb.c
14
struct kvm_s2_mmu *mmu;
arch/arm64/kvm/hyp/vhe/tlb.c
156
void __kvm_tlb_flush_vmid_range(struct kvm_s2_mmu *mmu,
arch/arm64/kvm/hyp/vhe/tlb.c
185
void __kvm_tlb_flush_vmid(struct kvm_s2_mmu *mmu)
arch/arm64/kvm/hyp/vhe/tlb.c
20
static void enter_vmid_context(struct kvm_s2_mmu *mmu,
arch/arm64/kvm/hyp/vhe/tlb.c
201
void __kvm_flush_cpu_context(struct kvm_s2_mmu *mmu)
arch/arm64/kvm/hyp/vhe/tlb.c
232
int __kvm_tlbi_s1e2(struct kvm_s2_mmu *mmu, u64 va, u64 sys_encoding)
arch/arm64/kvm/hyp/vhe/tlb.c
92
void __kvm_tlb_flush_vmid_ipa(struct kvm_s2_mmu *mmu,
arch/arm64/kvm/mmu.c
1097
void kvm_free_stage2_pgd(struct kvm_s2_mmu *mmu)
arch/arm64/kvm/mmu.c
1183
struct kvm_s2_mmu *mmu = &kvm->arch.mmu;
arch/arm64/kvm/mmu.c
1221
void kvm_stage2_wp_range(struct kvm_s2_mmu *mmu, phys_addr_t addr, phys_addr_t end)
arch/arm64/kvm/mmu.c
1949
struct kvm_s2_mmu *mmu;
arch/arm64/kvm/mmu.c
328
static void __unmap_stage2_range(struct kvm_s2_mmu *mmu, phys_addr_t start, u64 size,
arch/arm64/kvm/mmu.c
340
void kvm_stage2_unmap_range(struct kvm_s2_mmu *mmu, phys_addr_t start,
arch/arm64/kvm/mmu.c
346
void kvm_stage2_flush_range(struct kvm_s2_mmu *mmu, phys_addr_t addr, phys_addr_t end)
arch/arm64/kvm/mmu.c
62
static int stage2_apply_range(struct kvm_s2_mmu *mmu, phys_addr_t addr,
arch/arm64/kvm/mmu.c
875
static int kvm_init_ipa_range(struct kvm_s2_mmu *mmu, unsigned long type)
arch/arm64/kvm/mmu.c
954
int kvm_init_stage2_mmu(struct kvm *kvm, struct kvm_s2_mmu *mmu, unsigned long type)
arch/arm64/kvm/nested.c
1140
struct kvm_s2_mmu *mmu = &kvm->arch.nested_mmus[i];
arch/arm64/kvm/nested.c
1159
struct kvm_s2_mmu *mmu = &kvm->arch.nested_mmus[i];
arch/arm64/kvm/nested.c
1178
struct kvm_s2_mmu *mmu = &kvm->arch.nested_mmus[i];
arch/arm64/kvm/nested.c
1190
struct kvm_s2_mmu *mmu = &kvm->arch.nested_mmus[i];
arch/arm64/kvm/nested.c
1838
struct kvm_s2_mmu *mmu = vcpu->arch.hw_mmu;
arch/arm64/kvm/nested.c
513
static u8 get_guest_mapping_ttl(struct kvm_s2_mmu *mmu, u64 addr)
arch/arm64/kvm/nested.c
53
static int init_nested_s2_mmu(struct kvm *kvm, struct kvm_s2_mmu *mmu)
arch/arm64/kvm/nested.c
583
unsigned long compute_tlb_inval_range(struct kvm_s2_mmu *mmu, u64 val)
arch/arm64/kvm/nested.c
640
void (*tlbi_callback)(struct kvm_s2_mmu *,
arch/arm64/kvm/nested.c
646
struct kvm_s2_mmu *mmu = &kvm->arch.nested_mmus[i];
arch/arm64/kvm/nested.c
658
struct kvm_s2_mmu *lookup_s2_mmu(struct kvm_vcpu *vcpu)
arch/arm64/kvm/nested.c
688
struct kvm_s2_mmu *mmu = &kvm->arch.nested_mmus[i];
arch/arm64/kvm/nested.c
707
static struct kvm_s2_mmu *get_s2_mmu_nested(struct kvm_vcpu *vcpu)
arch/arm64/kvm/nested.c
71
struct kvm_s2_mmu *tmp;
arch/arm64/kvm/nested.c
710
struct kvm_s2_mmu *s2_mmu;
arch/arm64/kvm/nested.c
768
void kvm_init_nested_s2_mmu(struct kvm_s2_mmu *mmu)
arch/arm64/kvm/pkvm.c
316
int pkvm_pgtable_stage2_init(struct kvm_pgtable *pgt, struct kvm_s2_mmu *mmu,
arch/arm64/kvm/ptdump.c
118
struct kvm_s2_mmu *mmu = &kvm->arch.mmu;
arch/arm64/kvm/ptdump.c
153
struct kvm_s2_mmu *mmu = &kvm->arch.mmu;
arch/arm64/kvm/sys_regs.c
3903
static void s2_mmu_unmap_range(struct kvm_s2_mmu *mmu,
arch/arm64/kvm/sys_regs.c
3985
static void s2_mmu_unmap_ipa(struct kvm_s2_mmu *mmu,
arch/arm64/kvm/sys_regs.c
4031
static void s2_mmu_tlbi_s1e1(struct kvm_s2_mmu *mmu,