Symbol: hpa_t
arch/powerpc/kvm/e500_mmu_host.c
588
hpa_t addr;
arch/s390/kvm/vsie.c
666
static int pin_guest_page(struct kvm *kvm, gpa_t gpa, hpa_t *hpa)
arch/s390/kvm/vsie.c
673
*hpa = (hpa_t)page_to_phys(page) + (gpa & ~PAGE_MASK);
arch/s390/kvm/vsie.c
678
static void unpin_guest_page(struct kvm *kvm, gpa_t gpa, hpa_t hpa)
arch/s390/kvm/vsie.c
689
hpa_t hpa;
arch/s390/kvm/vsie.c
746
hpa_t hpa;
arch/s390/kvm/vsie.c
863
hpa_t hpa = virt_to_phys(vsie_page->scb_o);
arch/s390/kvm/vsie.c
879
hpa_t hpa;
arch/x86/include/asm/kvm_host.h
1102
hpa_t hv_root_tdp;
arch/x86/include/asm/kvm_host.h
151
#define INVALID_PAGE (~(hpa_t)0)
arch/x86/include/asm/kvm_host.h
1595
hpa_t hv_root_tdp;
arch/x86/include/asm/kvm_host.h
1860
void (*load_mmu_pgd)(struct kvm_vcpu *vcpu, hpa_t root_hpa,
arch/x86/include/asm/kvm_host.h
441
hpa_t hpa;
arch/x86/include/asm/kvm_host.h
475
hpa_t mirror_root_hpa;
arch/x86/kvm/kvm_onhyperv.c
112
void hv_track_root_tdp(struct kvm_vcpu *vcpu, hpa_t root_tdp)
arch/x86/kvm/kvm_onhyperv.c
27
static inline int hv_remote_flush_root_tdp(hpa_t root_tdp,
arch/x86/kvm/kvm_onhyperv.c
44
hpa_t root;
arch/x86/kvm/kvm_onhyperv.h
12
void hv_track_root_tdp(struct kvm_vcpu *vcpu, hpa_t root_tdp);
arch/x86/kvm/kvm_onhyperv.h
13
static inline hpa_t hv_get_partition_assist_page(struct kvm_vcpu *vcpu)
arch/x86/kvm/kvm_onhyperv.h
39
static inline void hv_track_root_tdp(struct kvm_vcpu *vcpu, hpa_t root_tdp)
arch/x86/kvm/mmu/mmu.c
158
hpa_t shadow_addr;
arch/x86/kvm/mmu/mmu.c
2477
struct kvm_vcpu *vcpu, hpa_t root,
arch/x86/kvm/mmu/mmu.c
3783
static void mmu_free_root_page(struct kvm *kvm, hpa_t *root_hpa,
arch/x86/kvm/mmu/mmu.c
3877
hpa_t root_hpa;
arch/x86/kvm/mmu/mmu.c
3900
static hpa_t mmu_alloc_root(struct kvm_vcpu *vcpu, gfn_t gfn, int quadrant,
arch/x86/kvm/mmu/mmu.c
3922
hpa_t root;
arch/x86/kvm/mmu/mmu.c
4063
hpa_t root;
arch/x86/kvm/mmu/mmu.c
4247
static bool is_unsync_root(hpa_t root)
arch/x86/kvm/mmu/mmu.c
4296
hpa_t root = vcpu->arch.mmu->root.hpa;
arch/x86/kvm/mmu/mmu.c
4312
hpa_t root = vcpu->arch.mmu->pae_root[i];
arch/x86/kvm/mmu/mmu.c
6094
static bool is_obsolete_root(struct kvm *kvm, hpa_t root_hpa)
arch/x86/kvm/mmu/mmu.c
6488
u64 addr, hpa_t root_hpa)
arch/x86/kvm/mmu/spte.h
249
static inline hpa_t kvm_mmu_get_dummy_root(void)
arch/x86/kvm/mmu/spte.h
254
static inline bool kvm_mmu_is_dummy_root(hpa_t shadow_page)
arch/x86/kvm/mmu/spte.h
259
static inline struct kvm_mmu_page *to_shadow_page(hpa_t shadow_page)
arch/x86/kvm/mmu/spte.h
276
static inline struct kvm_mmu_page *root_to_sp(hpa_t root)
arch/x86/kvm/svm/svm.c
4033
hpa_t root_tdp = vcpu->arch.mmu->root.hpa;
arch/x86/kvm/svm/svm.c
4442
static void svm_load_mmu_pgd(struct kvm_vcpu *vcpu, hpa_t root_hpa,
arch/x86/kvm/svm/svm_onhyperv.c
21
hpa_t partition_assist_page = hv_get_partition_assist_page(vcpu);
arch/x86/kvm/vmx/main.c
600
static void vt_load_mmu_pgd(struct kvm_vcpu *vcpu, hpa_t root_hpa,
arch/x86/kvm/vmx/nested.c
387
static bool nested_ept_root_matches(hpa_t root_hpa, u64 root_eptp, u64 eptp)
arch/x86/kvm/vmx/tdx.c
1613
void tdx_load_mmu_pgd(struct kvm_vcpu *vcpu, hpa_t root_hpa, int pgd_level)
arch/x86/kvm/vmx/vmx.c
3388
static u64 construct_eptp(hpa_t root_hpa)
arch/x86/kvm/vmx/vmx.c
3412
static void vmx_flush_tlb_ept_root(hpa_t root_hpa)
arch/x86/kvm/vmx/vmx.c
3598
void vmx_load_mmu_pgd(struct kvm_vcpu *vcpu, hpa_t root_hpa, int root_level)
arch/x86/kvm/vmx/vmx.c
630
hpa_t partition_assist_page = hv_get_partition_assist_page(vcpu);
arch/x86/kvm/vmx/x86_ops.h
155
void tdx_load_mmu_pgd(struct kvm_vcpu *vcpu, hpa_t root_hpa, int root_level);
arch/x86/kvm/vmx/x86_ops.h
68
void vmx_load_mmu_pgd(struct kvm_vcpu *vcpu, hpa_t root_hpa, int root_level);
include/linux/kvm_host.h
1914
static inline hpa_t pfn_to_hpa(kvm_pfn_t pfn)
include/linux/kvm_host.h
1916
return (hpa_t)pfn << PAGE_SHIFT;