Symbol: kvm_pfn_t
arch/arm64/include/asm/kvm_pgtable.h
151
static inline kvm_pfn_t kvm_pte_to_pfn(kvm_pte_t pte)
arch/arm64/kvm/mmu.c
1403
unsigned long hva, kvm_pfn_t *pfnp,
arch/arm64/kvm/mmu.c
1406
kvm_pfn_t pfn = *pfnp;
arch/arm64/kvm/mmu.c
1470
static void sanitise_mte_tags(struct kvm *kvm, kvm_pfn_t pfn,
arch/arm64/kvm/mmu.c
1579
kvm_pfn_t pfn;
arch/arm64/kvm/mmu.c
1659
kvm_pfn_t pfn;
arch/loongarch/kvm/mmu.c
777
kvm_pfn_t pfn;
arch/mips/kvm/mmu.c
555
kvm_pfn_t pfn;
arch/powerpc/include/asm/kvm_book3s.h
237
extern kvm_pfn_t kvmppc_gpa_to_pfn(struct kvm_vcpu *vcpu, gpa_t gpa,
arch/powerpc/include/asm/kvm_ppc.h
930
static inline void kvmppc_mmu_flush_icache(kvm_pfn_t pfn)
arch/powerpc/kvm/book3s.c
424
kvm_pfn_t kvmppc_gpa_to_pfn(struct kvm_vcpu *vcpu, gpa_t gpa, bool writing,
arch/powerpc/kvm/book3s.c
437
kvm_pfn_t pfn;
arch/powerpc/kvm/book3s.c
439
pfn = (kvm_pfn_t)virt_to_phys((void*)shared_page) >> PAGE_SHIFT;
arch/powerpc/kvm/book3s_32_mmu_host.c
132
kvm_pfn_t hpaddr;
arch/powerpc/kvm/book3s_64_mmu_host.c
75
kvm_pfn_t hpaddr;
arch/powerpc/kvm/book3s_64_mmu_radix.c
836
kvm_pfn_t pfn;
arch/powerpc/kvm/e500.h
43
kvm_pfn_t pfn; /* valid only for TLB0, except briefly */
arch/powerpc/kvm/e500_mmu_host.c
167
kvm_pfn_t pfn;
arch/powerpc/kvm/e500_mmu_host.c
169
pfn = (kvm_pfn_t)virt_to_phys((void *)shared_page) >> PAGE_SHIFT;
arch/powerpc/kvm/e500_mmu_host.c
250
kvm_pfn_t pfn, unsigned int wimg,
arch/powerpc/kvm/e500_mmu_host.c
307
kvm_pfn_t pfn = tlbe->pfn;
arch/powerpc/kvm/trace_pr.h
32
TP_PROTO(int rflags, ulong hpteg, ulong va, kvm_pfn_t hpaddr,
arch/riscv/kvm/mmu.c
422
kvm_pfn_t *hfnp, gpa_t *gpa)
arch/riscv/kvm/mmu.c
424
kvm_pfn_t hfn = *hfnp;
arch/riscv/kvm/mmu.c
453
kvm_pfn_t hfn;
arch/s390/kvm/dat.h
393
static inline union pte _pte(kvm_pfn_t pfn, bool writable, bool dirty, bool special)
arch/s390/kvm/dat.h
407
static inline union crste _crste_fc0(kvm_pfn_t pfn, int tt)
arch/s390/kvm/dat.h
431
static inline union crste _crste_fc1(kvm_pfn_t pfn, int tt, bool writable, bool dirty)
arch/s390/kvm/dat.h
490
kvm_pfn_t pfn; /* Host PFN */
arch/s390/kvm/gmap.c
1006
kvm_pfn_t pfn, int level, bool wr)
arch/s390/kvm/gmap.h
103
kvm_pfn_t pfn, int level, bool wr);
arch/x86/include/asm/kvm_host.h
1962
int (*gmem_prepare)(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order);
arch/x86/include/asm/kvm_host.h
1963
void (*gmem_invalidate)(kvm_pfn_t start, kvm_pfn_t end);
arch/x86/include/asm/kvm_host.h
1964
int (*gmem_max_mapping_level)(struct kvm *kvm, kvm_pfn_t pfn, bool is_private);
arch/x86/kvm/mmu.h
258
int kvm_tdp_mmu_map_private_pfn(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_pfn_t pfn);
arch/x86/kvm/mmu/mmu.c
3032
kvm_pfn_t pfn, struct kvm_page_fault *fault)
arch/x86/kvm/mmu/mmu.c
3317
kvm_pfn_t pfn;
arch/x86/kvm/mmu/mmu.c
3383
kvm_pfn_t mask;
arch/x86/kvm/mmu/mmu.c
5036
int kvm_tdp_mmu_map_private_pfn(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_pfn_t pfn)
arch/x86/kvm/mmu/mmu_internal.h
281
kvm_pfn_t pfn;
arch/x86/kvm/mmu/spte.c
107
static bool __kvm_is_mmio_pfn(kvm_pfn_t pfn)
arch/x86/kvm/mmu/spte.c
128
static bool kvm_is_mmio_pfn(kvm_pfn_t pfn, int *is_host_mmio)
arch/x86/kvm/mmu/spte.c
188
unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn,
arch/x86/kvm/mmu/spte.h
363
static inline kvm_pfn_t spte_to_pfn(u64 pte)
arch/x86/kvm/mmu/spte.h
545
unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn,
arch/x86/kvm/svm/sev.c
2280
static int sev_gmem_post_populate(struct kvm *kvm, gfn_t gfn, kvm_pfn_t pfn,
arch/x86/kvm/svm/sev.c
3702
static int snp_rmptable_psmash(kvm_pfn_t pfn)
arch/x86/kvm/svm/sev.c
3967
kvm_pfn_t pfn;
arch/x86/kvm/svm/sev.c
4850
kvm_pfn_t pfn;
arch/x86/kvm/svm/sev.c
4936
static bool is_pfn_range_shared(kvm_pfn_t start, kvm_pfn_t end)
arch/x86/kvm/svm/sev.c
4938
kvm_pfn_t pfn = start;
arch/x86/kvm/svm/sev.c
4971
static bool is_large_rmp_possible(struct kvm *kvm, kvm_pfn_t pfn, int order)
arch/x86/kvm/svm/sev.c
4973
kvm_pfn_t pfn_aligned = ALIGN_DOWN(pfn, PTRS_PER_PMD);
arch/x86/kvm/svm/sev.c
4987
int sev_gmem_prepare(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order)
arch/x86/kvm/svm/sev.c
4990
kvm_pfn_t pfn_aligned;
arch/x86/kvm/svm/sev.c
5034
void sev_gmem_invalidate(kvm_pfn_t start, kvm_pfn_t end)
arch/x86/kvm/svm/sev.c
5036
kvm_pfn_t pfn;
arch/x86/kvm/svm/sev.c
5097
int sev_gmem_max_mapping_level(struct kvm *kvm, kvm_pfn_t pfn, bool is_private)
arch/x86/kvm/svm/svm.h
897
int sev_gmem_prepare(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order);
arch/x86/kvm/svm/svm.h
898
void sev_gmem_invalidate(kvm_pfn_t start, kvm_pfn_t end);
arch/x86/kvm/svm/svm.h
899
int sev_gmem_max_mapping_level(struct kvm *kvm, kvm_pfn_t pfn, bool is_private);
arch/x86/kvm/svm/svm.h
923
static inline int sev_gmem_prepare(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order)
arch/x86/kvm/svm/svm.h
927
static inline void sev_gmem_invalidate(kvm_pfn_t start, kvm_pfn_t end) {}
arch/x86/kvm/svm/svm.h
928
static inline int sev_gmem_max_mapping_level(struct kvm *kvm, kvm_pfn_t pfn, bool is_private)
arch/x86/kvm/vmx/main.c
842
static int vt_gmem_max_mapping_level(struct kvm *kvm, kvm_pfn_t pfn,
arch/x86/kvm/vmx/tdx.c
1625
kvm_pfn_t pfn)
arch/x86/kvm/vmx/tdx.c
1649
enum pg_level level, kvm_pfn_t pfn)
arch/x86/kvm/vmx/tdx.c
1672
kvm_pfn_t pfn = spte_to_pfn(mirror_spte);
arch/x86/kvm/vmx/tdx.c
3119
static int tdx_gmem_post_populate(struct kvm *kvm, gfn_t gfn, kvm_pfn_t pfn,
arch/x86/kvm/vmx/tdx.c
3285
int tdx_gmem_max_mapping_level(struct kvm *kvm, kvm_pfn_t pfn, bool is_private)
arch/x86/kvm/vmx/vmx.c
7037
kvm_pfn_t pfn;
arch/x86/kvm/vmx/x86_ops.h
156
int tdx_gmem_max_mapping_level(struct kvm *kvm, kvm_pfn_t pfn, bool is_private);
arch/x86/kvm/x86.c
14070
int kvm_arch_gmem_prepare(struct kvm *kvm, gfn_t gfn, kvm_pfn_t pfn, int max_order)
arch/x86/kvm/x86.c
14077
void kvm_arch_gmem_invalidate(kvm_pfn_t start, kvm_pfn_t end)
include/linux/kvm_host.h
107
static inline bool is_error_pfn(kvm_pfn_t pfn)
include/linux/kvm_host.h
116
static inline bool is_sigpending_pfn(kvm_pfn_t pfn)
include/linux/kvm_host.h
126
static inline bool is_error_noslot_pfn(kvm_pfn_t pfn)
include/linux/kvm_host.h
1303
kvm_pfn_t __kvm_faultin_pfn(const struct kvm_memory_slot *slot, gfn_t gfn,
include/linux/kvm_host.h
1307
static inline kvm_pfn_t kvm_faultin_pfn(struct kvm_vcpu *vcpu, gfn_t gfn,
include/linux/kvm_host.h
132
static inline bool is_noslot_pfn(kvm_pfn_t pfn)
include/linux/kvm_host.h
1914
static inline hpa_t pfn_to_hpa(kvm_pfn_t pfn)
include/linux/kvm_host.h
2542
gfn_t gfn, kvm_pfn_t *pfn, struct page **page,
include/linux/kvm_host.h
2547
kvm_pfn_t *pfn, struct page **page,
include/linux/kvm_host.h
2556
int kvm_arch_gmem_prepare(struct kvm *kvm, gfn_t gfn, kvm_pfn_t pfn, int max_order);
include/linux/kvm_host.h
2581
typedef int (*kvm_gmem_populate_cb)(struct kvm *kvm, gfn_t gfn, kvm_pfn_t pfn,
include/linux/kvm_host.h
2589
void kvm_arch_gmem_invalidate(kvm_pfn_t start, kvm_pfn_t end);
include/linux/kvm_host.h
295
kvm_pfn_t pfn;
include/linux/kvm_host.h
296
kvm_pfn_t gfn;
include/linux/kvm_types.h
94
kvm_pfn_t pfn;
virt/kvm/guest_memfd.c
52
static inline kvm_pfn_t folio_file_pfn(struct folio *folio, pgoff_t index)
virt/kvm/guest_memfd.c
529
kvm_pfn_t pfn = page_to_pfn(page);
virt/kvm/guest_memfd.c
66
kvm_pfn_t pfn = folio_file_pfn(folio, index);
virt/kvm/guest_memfd.c
754
pgoff_t index, kvm_pfn_t *pfn,
virt/kvm/guest_memfd.c
789
gfn_t gfn, kvm_pfn_t *pfn, struct page **page,
virt/kvm/guest_memfd.c
830
kvm_pfn_t pfn;
virt/kvm/kvm_main.c
2834
static kvm_pfn_t kvm_resolve_pfn(struct kvm_follow_pfn *kfp, struct page *page,
virt/kvm/kvm_main.c
2837
kvm_pfn_t pfn;
virt/kvm/kvm_main.c
2858
static bool hva_to_pfn_fast(struct kvm_follow_pfn *kfp, kvm_pfn_t *pfn)
virt/kvm/kvm_main.c
2890
static int hva_to_pfn_slow(struct kvm_follow_pfn *kfp, kvm_pfn_t *pfn)
virt/kvm/kvm_main.c
2947
struct kvm_follow_pfn *kfp, kvm_pfn_t *p_pfn)
virt/kvm/kvm_main.c
2992
kvm_pfn_t hva_to_pfn(struct kvm_follow_pfn *kfp)
virt/kvm/kvm_main.c
2995
kvm_pfn_t pfn;
virt/kvm/kvm_main.c
3037
static kvm_pfn_t kvm_follow_pfn(struct kvm_follow_pfn *kfp)
virt/kvm/kvm_main.c
3056
kvm_pfn_t __kvm_faultin_pfn(const struct kvm_memory_slot *slot, gfn_t gfn,
virt/kvm/kvm_mm.h
56
kvm_pfn_t hva_to_pfn(struct kvm_follow_pfn *kfp);
virt/kvm/pfncache.c
111
static void gpc_unmap(kvm_pfn_t pfn, void *khva)
virt/kvm/pfncache.c
155
static kvm_pfn_t hva_to_pfn_retry(struct gfn_to_pfn_cache *gpc)
virt/kvm/pfncache.c
159
kvm_pfn_t new_pfn = KVM_PFN_ERR_FAULT;
virt/kvm/pfncache.c
261
kvm_pfn_t old_pfn;
virt/kvm/pfncache.c
450
kvm_pfn_t old_pfn;
virt/kvm/pfncache.c
99
static void *gpc_map(kvm_pfn_t pfn)