arch/arm64/kvm/at.c
1742
gfn_t gfn;
arch/arm64/kvm/mmu.c
1299
gfn_t gfn_offset, unsigned long mask)
arch/arm64/kvm/mmu.c
1580
gfn_t gfn;
arch/arm64/kvm/mmu.c
1658
gfn_t gfn;
arch/arm64/kvm/mmu.c
185
gfn_t gfn, u64 nr_pages)
arch/arm64/kvm/mmu.c
2058
gfn_t gfn;
arch/arm64/kvm/vgic/vgic-its.c
850
gfn_t gfn = gpa >> PAGE_SHIFT;
arch/loongarch/kvm/mmu.c
292
static void kvm_flush_range(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn, int lock)
arch/loongarch/kvm/mmu.c
341
static int kvm_mkclean_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn)
arch/loongarch/kvm/mmu.c
363
struct kvm_memory_slot *slot, gfn_t gfn_offset, unsigned long mask)
arch/loongarch/kvm/mmu.c
366
gfn_t base_gfn = slot->base_gfn + gfn_offset;
arch/loongarch/kvm/mmu.c
367
gfn_t start = base_gfn + __ffs(mask);
arch/loongarch/kvm/mmu.c
368
gfn_t end = base_gfn + __fls(mask) + 1;
arch/loongarch/kvm/mmu.c
556
gfn_t gfn = gpa >> PAGE_SHIFT;
arch/loongarch/kvm/mmu.c
57
gfn_t offset;
arch/loongarch/kvm/mmu.c
666
static int host_pfn_mapping_level(struct kvm *kvm, gfn_t gfn,
arch/loongarch/kvm/mmu.c
728
static kvm_pte_t *kvm_split_huge(struct kvm_vcpu *vcpu, kvm_pte_t *ptep, gfn_t gfn)
arch/loongarch/kvm/mmu.c
779
gfn_t gfn = gpa >> PAGE_SHIFT;
arch/mips/include/asm/kvm_host.h
808
bool kvm_mips_flush_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn);
arch/mips/include/asm/kvm_host.h
809
int kvm_mips_mkclean_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn);
arch/mips/kvm/mmu.c
270
bool kvm_mips_flush_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn)
arch/mips/kvm/mmu.c
396
int kvm_mips_mkclean_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn)
arch/mips/kvm/mmu.c
416
gfn_t gfn_offset, unsigned long mask)
arch/mips/kvm/mmu.c
418
gfn_t base_gfn = slot->base_gfn + gfn_offset;
arch/mips/kvm/mmu.c
419
gfn_t start = base_gfn + __ffs(mask);
arch/mips/kvm/mmu.c
420
gfn_t end = base_gfn + __fls(mask);
arch/mips/kvm/mmu.c
433
static int kvm_mips_mkold_gpa_pt(struct kvm *kvm, gfn_t start_gfn,
arch/mips/kvm/mmu.c
434
gfn_t end_gfn)
arch/mips/kvm/mmu.c
485
gfn_t gfn = gpa >> PAGE_SHIFT;
arch/mips/kvm/mmu.c
553
gfn_t gfn = gpa >> PAGE_SHIFT;
arch/powerpc/kvm/book3s.c
428
gfn_t gfn = gpa >> PAGE_SHIFT;
arch/powerpc/kvm/book3s_64_mmu_hv.c
842
gfn_t gfn;
arch/powerpc/kvm/book3s_64_mmu_hv.c
939
gfn_t gfn;
arch/powerpc/kvm/book3s_xive_native.c
573
gfn_t gfn;
arch/powerpc/kvm/booke.c
1287
gfn_t gfn;
arch/powerpc/kvm/booke.c
1344
gfn_t gfn;
arch/powerpc/kvm/e500_mmu_host.c
321
u64 gvaddr, gfn_t gfn, struct kvm_book3e_206_tlb_entry *gtlbe,
arch/powerpc/kvm/e500_mmu_host.c
519
u64 gvaddr, gfn_t gfn, struct kvm_book3e_206_tlb_entry *gtlbe,
arch/powerpc/kvm/e500_mmu_host.c
571
gfn_t gfn = gpaddr >> PAGE_SHIFT;
arch/riscv/kvm/mmu.c
104
gfn_t gfn_offset,
arch/riscv/kvm/mmu.c
456
gfn_t gfn = gpa >> PAGE_SHIFT;
arch/riscv/kvm/tlb.c
460
int kvm_arch_flush_remote_tlbs_range(struct kvm *kvm, gfn_t gfn, u64 nr_pages)
arch/riscv/kvm/vcpu_exit.c
22
gfn_t gfn;
arch/riscv/kvm/vcpu_sbi_sta.c
36
gfn_t gfn;
arch/s390/kvm/dat.c
1000
static long dat_set_pn_crste(union crste *crstep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
1018
static long dat_set_pn_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
1032
int dat_set_prefix_notif_bit(union asce asce, gfn_t gfn)
arch/s390/kvm/dat.c
1063
int dat_perform_essa(union asce asce, gfn_t gfn, int orc, union essa_state *state, bool *dirty)
arch/s390/kvm/dat.c
1144
static long dat_reset_cmma_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
1158
long dat_reset_cmma(union asce asce, gfn_t start)
arch/s390/kvm/dat.c
1169
gfn_t start;
arch/s390/kvm/dat.c
1170
gfn_t end;
arch/s390/kvm/dat.c
1176
static long __dat_peek_cmma_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
1189
static long __dat_peek_cmma_crste(union crste *crstep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
1198
int dat_peek_cmma(gfn_t start, union asce asce, unsigned int *count, u8 *values)
arch/s390/kvm/dat.c
1216
static long __dat_get_cmma_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
1244
int dat_get_cmma(union asce asce, gfn_t *start, unsigned int *count, u8 *values, atomic64_t *rem)
arch/s390/kvm/dat.c
1271
static long __dat_set_cmma_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
1304
int dat_set_cmma_bits(struct kvm_s390_mmu_cache *mc, union asce asce, gfn_t gfn,
arch/s390/kvm/dat.c
1311
gfn_t cur;
arch/s390/kvm/dat.c
153
gfn_t gfn, union asce asce)
arch/s390/kvm/dat.c
188
union pgste __dat_ptep_xchg(union pte *ptep, union pgste pgste, union pte new, gfn_t gfn,
arch/s390/kvm/dat.c
233
static int dat_split_ste(struct kvm_s390_mmu_cache *mc, union pmd *pmdp, gfn_t gfn,
arch/s390/kvm/dat.c
302
gfn_t gfn, union asce asce, bool uses_skeys)
arch/s390/kvm/dat.c
378
int dat_entry_walk(struct kvm_s390_mmu_cache *mc, gfn_t gfn, union asce asce, int flags,
arch/s390/kvm/dat.c
513
static long dat_pte_walk_range(gfn_t gfn, gfn_t end, struct page_table *table, struct dat_walk *w)
arch/s390/kvm/dat.c
533
static long dat_crste_walk_range(gfn_t start, gfn_t end, struct crst_table *table,
arch/s390/kvm/dat.c
539
gfn_t cur, next;
arch/s390/kvm/dat.c
592
long _dat_walk_gfn_range(gfn_t start, gfn_t end, union asce asce,
arch/s390/kvm/dat.c
613
int dat_get_storage_key(union asce asce, gfn_t gfn, union skey *skey)
arch/s390/kvm/dat.c
653
int dat_set_storage_key(struct kvm_s390_mmu_cache *mc, union asce asce, gfn_t gfn,
arch/s390/kvm/dat.c
708
int dat_cond_set_storage_key(struct kvm_s390_mmu_cache *mmc, union asce asce, gfn_t gfn,
arch/s390/kvm/dat.c
755
int dat_reset_reference_bit(union asce asce, gfn_t gfn)
arch/s390/kvm/dat.c
788
static long dat_reset_skeys_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
806
static long dat_reset_skeys_crste(union crste *crstep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
825
long dat_reset_skeys(union asce asce, gfn_t start)
arch/s390/kvm/dat.c
841
static long _dat_slot_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
857
static long _dat_slot_crste(union crste *crstep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/dat.c
895
int dat_set_slot(struct kvm_s390_mmu_cache *mc, union asce asce, gfn_t start, gfn_t end,
arch/s390/kvm/dat.c
968
static long _dat_test_young_pte(union pte *ptep, gfn_t start, gfn_t end, struct dat_walk *walk)
arch/s390/kvm/dat.c
973
static long _dat_test_young_crste(union crste *crstep, gfn_t start, gfn_t end,
arch/s390/kvm/dat.c
995
bool dat_test_age_gfn(union asce asce, gfn_t start, gfn_t end)
arch/s390/kvm/dat.h
348
typedef long (*dat_walk_op)(union crste *crste, gfn_t gfn, gfn_t next, struct dat_walk *w);
arch/s390/kvm/dat.h
360
long (*pte_entry)(union pte *pte, gfn_t gfn, gfn_t next, struct dat_walk *w);
arch/s390/kvm/dat.h
368
gfn_t start;
arch/s390/kvm/dat.h
369
gfn_t end;
arch/s390/kvm/dat.h
489
gfn_t gfn; /* Guest frame */
arch/s390/kvm/dat.h
515
gfn_t gfn, union asce asce, bool uses_skeys);
arch/s390/kvm/dat.h
516
bool dat_crstep_xchg_atomic(union crste *crstep, union crste old, union crste new, gfn_t gfn,
arch/s390/kvm/dat.h
518
void dat_crstep_xchg(union crste *crstep, union crste new, gfn_t gfn, union asce asce);
arch/s390/kvm/dat.h
520
long _dat_walk_gfn_range(gfn_t start, gfn_t end, union asce asce,
arch/s390/kvm/dat.h
523
int dat_entry_walk(struct kvm_s390_mmu_cache *mc, gfn_t gfn, union asce asce, int flags,
arch/s390/kvm/dat.h
528
int dat_get_storage_key(union asce asce, gfn_t gfn, union skey *skey);
arch/s390/kvm/dat.h
529
int dat_set_storage_key(struct kvm_s390_mmu_cache *mc, union asce asce, gfn_t gfn,
arch/s390/kvm/dat.h
531
int dat_cond_set_storage_key(struct kvm_s390_mmu_cache *mmc, union asce asce, gfn_t gfn,
arch/s390/kvm/dat.h
533
int dat_reset_reference_bit(union asce asce, gfn_t gfn);
arch/s390/kvm/dat.h
534
long dat_reset_skeys(union asce asce, gfn_t start);
arch/s390/kvm/dat.h
539
int dat_set_slot(struct kvm_s390_mmu_cache *mc, union asce asce, gfn_t start, gfn_t end,
arch/s390/kvm/dat.h
541
int dat_set_prefix_notif_bit(union asce asce, gfn_t gfn);
arch/s390/kvm/dat.h
542
bool dat_test_age_gfn(union asce asce, gfn_t start, gfn_t end);
arch/s390/kvm/dat.h
544
int dat_perform_essa(union asce asce, gfn_t gfn, int orc, union essa_state *state, bool *dirty);
arch/s390/kvm/dat.h
545
long dat_reset_cmma(union asce asce, gfn_t start_gfn);
arch/s390/kvm/dat.h
546
int dat_peek_cmma(gfn_t start, union asce asce, unsigned int *count, u8 *values);
arch/s390/kvm/dat.h
547
int dat_get_cmma(union asce asce, gfn_t *start, unsigned int *count, u8 *values, atomic64_t *rem);
arch/s390/kvm/dat.h
548
int dat_set_cmma_bits(struct kvm_s390_mmu_cache *mc, union asce asce, gfn_t gfn,
arch/s390/kvm/dat.h
586
static inline bool crdte_crste(union crste *crstep, union crste old, union crste new, gfn_t gfn,
arch/s390/kvm/dat.h
603
static __always_inline void idte_crste(union crste *crstep, gfn_t gfn, unsigned long opt,
arch/s390/kvm/dat.h
638
static inline gfn_t asce_end(union asce asce)
arch/s390/kvm/dat.h
664
static inline bool asce_contains_gfn(union asce asce, gfn_t gfn)
arch/s390/kvm/dat.h
777
static inline phys_addr_t large_pud_to_phys(union pud pud, gfn_t gfn)
arch/s390/kvm/dat.h
782
static inline phys_addr_t large_pmd_to_phys(union pmd pmd, gfn_t gfn)
arch/s390/kvm/dat.h
787
static inline phys_addr_t large_crste_to_phys(union crste crste, gfn_t gfn)
arch/s390/kvm/dat.h
875
static inline void dat_ptep_xchg(union pte *ptep, union pte new, gfn_t gfn, union asce asce,
arch/s390/kvm/dat.h
885
static inline void dat_ptep_clear(union pte *ptep, gfn_t gfn, union asce asce, bool has_skeys)
arch/s390/kvm/dat.h
928
gfn_t gfn, union asce asce)
arch/s390/kvm/dat.h
934
gfn_t gfn, union asce asce)
arch/s390/kvm/dat.h
939
static inline union crste dat_crstep_clear_atomic(union crste *crstep, gfn_t gfn, union asce asce)
arch/s390/kvm/dat.h
954
static inline int dat_delete_slot(struct kvm_s390_mmu_cache *mc, union asce asce, gfn_t start,
arch/s390/kvm/dat.h
960
static inline int dat_create_slot(struct kvm_s390_mmu_cache *mc, union asce asce, gfn_t start,
arch/s390/kvm/diag.c
21
static void do_discard_gfn_range(struct kvm_vcpu *vcpu, gfn_t gfn_start, gfn_t gfn_end)
arch/s390/kvm/faultin.c
129
int kvm_s390_get_guest_page(struct kvm *kvm, struct guest_fault *f, gfn_t gfn, bool w)
arch/s390/kvm/faultin.h
17
int kvm_s390_get_guest_page(struct kvm *kvm, struct guest_fault *f, gfn_t gfn, bool w);
arch/s390/kvm/faultin.h
20
gfn_t gfn, bool wr)
arch/s390/kvm/faultin.h
71
gfn_t start, int n_pages, bool write_attempt)
arch/s390/kvm/gaccess.c
1468
gfn_t gfn;
arch/s390/kvm/gaccess.c
822
const gfn_t gfn = gpa_to_gfn(gpa);
arch/s390/kvm/gmap.c
1005
int gmap_protect_rmap(struct kvm_s390_mmu_cache *mc, struct gmap *sg, gfn_t p_gfn, gfn_t r_gfn,
arch/s390/kvm/gmap.c
1040
static long __set_cmma_dirty_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/gmap.c
1051
gfn_t gfn = 0;
arch/s390/kvm/gmap.c
1061
static void gmap_unshadow_level(struct gmap *sg, gfn_t r_gfn, int level)
arch/s390/kvm/gmap.c
1109
void _gmap_handle_vsie_unshadow_event(struct gmap *parent, gfn_t gfn)
arch/s390/kvm/gmap.c
1113
gfn_t start, end;
arch/s390/kvm/gmap.c
113
struct gmap *gmap_new_child(struct gmap *parent, gfn_t limit)
arch/s390/kvm/gmap.c
126
int gmap_set_limit(struct gmap *gmap, gfn_t limit)
arch/s390/kvm/gmap.c
254
bool _gmap_unmap_prefix(struct gmap *gmap, gfn_t gfn, gfn_t end, bool hint)
arch/s390/kvm/gmap.c
258
gfn_t prefix_gfn;
arch/s390/kvm/gmap.c
282
static long gmap_clear_young_pte(union pte *ptep, gfn_t gfn, gfn_t end, struct dat_walk *walk)
arch/s390/kvm/gmap.c
311
static long gmap_clear_young_crste(union crste *crstep, gfn_t gfn, gfn_t end, struct dat_walk *walk)
arch/s390/kvm/gmap.c
32
static int gmap_limit_to_type(gfn_t limit)
arch/s390/kvm/gmap.c
349
bool gmap_age_gfn(struct gmap *gmap, gfn_t start, gfn_t end)
arch/s390/kvm/gmap.c
371
static long _gmap_unmap_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *w)
arch/s390/kvm/gmap.c
393
static long _gmap_unmap_crste(union crste *crstep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/gmap.c
423
bool gmap_unmap_gfn_range(struct gmap *gmap, struct kvm_memory_slot *slot, gfn_t start, gfn_t end)
arch/s390/kvm/gmap.c
441
static union pgste __pte_test_and_clear_softdirty(union pte *ptep, union pgste pgste, gfn_t gfn,
arch/s390/kvm/gmap.c
464
static long _pte_test_and_clear_softdirty(union pte *ptep, gfn_t gfn, gfn_t end,
arch/s390/kvm/gmap.c
476
static long _crste_test_and_clear_softdirty(union crste *table, gfn_t gfn, gfn_t end,
arch/s390/kvm/gmap.c
509
void gmap_sync_dirty_log(struct gmap *gmap, gfn_t start, gfn_t end)
arch/s390/kvm/gmap.c
52
struct gmap *gmap_new(struct kvm *kvm, gfn_t limit)
arch/s390/kvm/gmap.c
620
static inline bool gmap_2g_allowed(struct gmap *gmap, gfn_t gfn)
arch/s390/kvm/gmap.c
625
static inline bool gmap_1m_allowed(struct gmap *gmap, gfn_t gfn)
arch/s390/kvm/gmap.c
700
gfn_t p_gfn, gfn_t c_gfn, bool force_alloc)
arch/s390/kvm/gmap.c
770
gfn_t gfn;
arch/s390/kvm/gmap.c
802
int gmap_ucas_map(struct gmap *gmap, gfn_t p_gfn, gfn_t c_gfn, unsigned long count)
arch/s390/kvm/gmap.c
830
static void gmap_ucas_unmap_one(struct gmap *gmap, gfn_t c_gfn)
arch/s390/kvm/gmap.c
843
void gmap_ucas_unmap(struct gmap *gmap, gfn_t c_gfn, unsigned long count)
arch/s390/kvm/gmap.c
851
static long _gmap_split_crste(union crste *crstep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/gmap.c
881
gfn_t start = 0;
arch/s390/kvm/gmap.c
893
gfn_t start = 0;
arch/s390/kvm/gmap.c
924
static long _destroy_pages_pte(union pte *ptep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/gmap.c
934
static long _destroy_pages_crste(union crste *crstep, gfn_t gfn, gfn_t next, struct dat_walk *walk)
arch/s390/kvm/gmap.c
951
int gmap_pv_destroy_range(struct gmap *gmap, gfn_t start, gfn_t end, bool interruptible)
arch/s390/kvm/gmap.c
970
int gmap_insert_rmap(struct gmap *sg, gfn_t p_gfn, gfn_t r_gfn, int level)
arch/s390/kvm/gmap.h
100
int gmap_pv_destroy_range(struct gmap *gmap, gfn_t start, gfn_t end, bool interruptible);
arch/s390/kvm/gmap.h
101
int gmap_insert_rmap(struct gmap *sg, gfn_t p_gfn, gfn_t r_gfn, int level);
arch/s390/kvm/gmap.h
102
int gmap_protect_rmap(struct kvm_s390_mmu_cache *mc, struct gmap *sg, gfn_t p_gfn, gfn_t r_gfn,
arch/s390/kvm/gmap.h
105
void _gmap_handle_vsie_unshadow_event(struct gmap *parent, gfn_t gfn);
arch/s390/kvm/gmap.h
152
static inline void gmap_handle_vsie_unshadow_event(struct gmap *parent, gfn_t gfn)
arch/s390/kvm/gmap.h
158
static inline bool gmap_mkold_prefix(struct gmap *gmap, gfn_t gfn, gfn_t end)
arch/s390/kvm/gmap.h
163
static inline bool gmap_unmap_prefix(struct gmap *gmap, gfn_t gfn, gfn_t end)
arch/s390/kvm/gmap.h
169
union pgste pgste, gfn_t gfn, bool needs_lock)
arch/s390/kvm/gmap.h
194
union pgste pgste, gfn_t gfn)
arch/s390/kvm/gmap.h
201
gfn_t gfn, bool needs_lock)
arch/s390/kvm/gmap.h
232
gfn_t gfn)
arch/s390/kvm/gmap.h
85
bool _gmap_unmap_prefix(struct gmap *gmap, gfn_t gfn, gfn_t end, bool hint);
arch/s390/kvm/gmap.h
86
bool gmap_age_gfn(struct gmap *gmap, gfn_t start, gfn_t end);
arch/s390/kvm/gmap.h
87
bool gmap_unmap_gfn_range(struct gmap *gmap, struct kvm_memory_slot *slot, gfn_t start, gfn_t end);
arch/s390/kvm/gmap.h
89
struct gmap *gmap_new(struct kvm *kvm, gfn_t limit);
arch/s390/kvm/gmap.h
90
struct gmap *gmap_new_child(struct gmap *parent, gfn_t limit);
arch/s390/kvm/gmap.h
94
void gmap_sync_dirty_log(struct gmap *gmap, gfn_t start, gfn_t end);
arch/s390/kvm/gmap.h
95
int gmap_set_limit(struct gmap *gmap, gfn_t limit);
arch/s390/kvm/gmap.h
97
int gmap_ucas_map(struct gmap *gmap, gfn_t p_gfn, gfn_t c_gfn, unsigned long count);
arch/s390/kvm/gmap.h
98
void gmap_ucas_unmap(struct gmap *gmap, gfn_t c_gfn, unsigned long count);
arch/s390/kvm/kvm-s390.c
4240
gfn_t gfn;
arch/s390/kvm/kvm-s390.c
561
gfn_t gfn = gpa_to_gfn(addr);
arch/s390/kvm/kvm-s390.c
722
gfn_t last_gfn = memslot->base_gfn + memslot->npages;
arch/s390/kvm/kvm-s390.c
978
gfn_t start_gfn = 0;
arch/x86/include/asm/kvm_host.h
1011
gfn_t mmio_gfn;
arch/x86/include/asm/kvm_host.h
1033
gfn_t gfns[ASYNC_PF_PER_VCPU];
arch/x86/include/asm/kvm_host.h
1630
gfn_t gfn_direct_bits;
arch/x86/include/asm/kvm_host.h
1797
int (*flush_remote_tlbs_range)(struct kvm *kvm, gfn_t gfn,
arch/x86/include/asm/kvm_host.h
1798
gfn_t nr_pages);
arch/x86/include/asm/kvm_host.h
1858
u8 (*get_mt_mask)(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio);
arch/x86/include/asm/kvm_host.h
1864
int (*link_external_spt)(struct kvm *kvm, gfn_t gfn, enum pg_level level,
arch/x86/include/asm/kvm_host.h
1867
int (*set_external_spte)(struct kvm *kvm, gfn_t gfn, enum pg_level level,
arch/x86/include/asm/kvm_host.h
1871
int (*free_external_spt)(struct kvm *kvm, gfn_t gfn, enum pg_level level,
arch/x86/include/asm/kvm_host.h
1875
void (*remove_external_spte)(struct kvm *kvm, gfn_t gfn, enum pg_level level,
arch/x86/include/asm/kvm_host.h
1962
int (*gmem_prepare)(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order);
arch/x86/include/asm/kvm_host.h
1999
gfn_t gfn;
arch/x86/include/asm/kvm_host.h
2045
static inline int kvm_arch_flush_remote_tlbs_range(struct kvm *kvm, gfn_t gfn,
arch/x86/include/asm/kvm_host.h
2096
void kvm_zap_gfn_range(struct kvm *kvm, gfn_t gfn_start, gfn_t gfn_end);
arch/x86/include/asm/kvm_host.h
2436
extern bool kvm_find_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn);
arch/x86/include/asm/kvm_page_track.h
43
void (*track_remove_region)(gfn_t gfn, unsigned long nr_pages,
arch/x86/include/asm/kvm_page_track.h
52
int kvm_write_track_add_gfn(struct kvm *kvm, gfn_t gfn);
arch/x86/include/asm/kvm_page_track.h
53
int kvm_write_track_remove_gfn(struct kvm *kvm, gfn_t gfn);
arch/x86/kvm/hyperv.c
764
gfn_t msg_page_gfn;
arch/x86/kvm/kvm_onhyperv.c
95
int hv_flush_remote_tlbs_range(struct kvm *kvm, gfn_t start_gfn, gfn_t nr_pages)
arch/x86/kvm/kvm_onhyperv.h
10
int hv_flush_remote_tlbs_range(struct kvm *kvm, gfn_t gfn, gfn_t nr_pages);
arch/x86/kvm/mmu.h
258
int kvm_tdp_mmu_map_private_pfn(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_pfn_t pfn);
arch/x86/kvm/mmu.h
265
static inline gfn_t gfn_to_index(gfn_t gfn, gfn_t base_gfn, int level)
arch/x86/kvm/mmu.h
309
static inline gfn_t kvm_gfn_direct_bits(const struct kvm *kvm)
arch/x86/kvm/mmu.h
321
static inline bool kvm_is_gfn_alias(struct kvm *kvm, gfn_t gfn)
arch/x86/kvm/mmu.h
61
static inline gfn_t kvm_mmu_max_gfn(void)
arch/x86/kvm/mmu/mmu.c
1175
static struct kvm_rmap_head *gfn_to_rmap(gfn_t gfn, int level,
arch/x86/kvm/mmu/mmu.c
1189
gfn_t gfn;
arch/x86/kvm/mmu/mmu.c
1379
gfn_t gfn_offset, unsigned long mask)
arch/x86/kvm/mmu/mmu.c
1402
gfn_t gfn_offset, unsigned long mask)
arch/x86/kvm/mmu/mmu.c
1425
gfn_t gfn_offset, unsigned long mask)
arch/x86/kvm/mmu/mmu.c
1439
gfn_t start = slot->base_gfn + gfn_offset + __ffs(mask);
arch/x86/kvm/mmu/mmu.c
1440
gfn_t end = slot->base_gfn + gfn_offset + __fls(mask);
arch/x86/kvm/mmu/mmu.c
1514
gfn_t start_gfn;
arch/x86/kvm/mmu/mmu.c
1515
gfn_t end_gfn;
arch/x86/kvm/mmu/mmu.c
1520
gfn_t gfn;
arch/x86/kvm/mmu/mmu.c
1540
gfn_t start_gfn, gfn_t end_gfn)
arch/x86/kvm/mmu/mmu.c
1589
gfn_t start_gfn, gfn_t end_gfn,
arch/x86/kvm/mmu/mmu.c
1639
gfn_t start, gfn_t end, bool can_yield,
arch/x86/kvm/mmu/mmu.c
1682
u64 *spte, gfn_t gfn, unsigned int access)
arch/x86/kvm/mmu/mmu.c
1704
u64 *spte, gfn_t gfn, unsigned int access)
arch/x86/kvm/mmu/mmu.c
1720
gfn_t gfn;
arch/x86/kvm/mmu/mmu.c
1830
static unsigned kvm_page_table_hashfn(gfn_t gfn)
arch/x86/kvm/mmu/mmu.c
1992
static struct hlist_head *kvm_get_mmu_page_hash(struct kvm *kvm, gfn_t gfn)
arch/x86/kvm/mmu/mmu.c
2257
gfn_t gfn,
arch/x86/kvm/mmu/mmu.c
2342
gfn_t gfn,
arch/x86/kvm/mmu/mmu.c
2379
gfn_t gfn,
arch/x86/kvm/mmu/mmu.c
2405
gfn_t gfn,
arch/x86/kvm/mmu/mmu.c
2464
u64 *sptep, gfn_t gfn,
arch/x86/kvm/mmu/mmu.c
281
static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index);
arch/x86/kvm/mmu/mmu.c
287
gfn_t gfn = kvm_mmu_page_get_gfn(sp, spte_index(sptep));
arch/x86/kvm/mmu/mmu.c
2930
gfn_t gfn, bool synchronizing, bool prefetch)
arch/x86/kvm/mmu/mmu.c
301
static gfn_t get_mmio_spte_gfn(u64 spte)
arch/x86/kvm/mmu/mmu.c
3031
u64 *sptep, unsigned int pte_access, gfn_t gfn,
arch/x86/kvm/mmu/mmu.c
3106
static bool kvm_mmu_prefetch_sptes(struct kvm_vcpu *vcpu, gfn_t gfn, u64 *sptep,
arch/x86/kvm/mmu/mmu.c
3147
gfn_t gfn = kvm_mmu_page_get_gfn(sp, spte_index(start));
arch/x86/kvm/mmu/mmu.c
3231
static int host_pfn_mapping_level(struct kvm *kvm, gfn_t gfn,
arch/x86/kvm/mmu/mmu.c
3313
const struct kvm_memory_slot *slot, gfn_t gfn,
arch/x86/kvm/mmu/mmu.c
3346
const struct kvm_memory_slot *slot, gfn_t gfn)
arch/x86/kvm/mmu/mmu.c
3441
gfn_t base_gfn = fault->gfn;
arch/x86/kvm/mmu/mmu.c
3480
static void kvm_send_hwpoison_signal(struct kvm_memory_slot *slot, gfn_t gfn)
arch/x86/kvm/mmu/mmu.c
3900
static hpa_t mmu_alloc_root(struct kvm_vcpu *vcpu, gfn_t gfn, int quadrant,
arch/x86/kvm/mmu/mmu.c
4061
gfn_t root_gfn, root_pgd;
arch/x86/kvm/mmu/mmu.c
4456
gfn_t gfn = get_mmio_spte_gfn(spte);
arch/x86/kvm/mmu/mmu.c
5036
int kvm_tdp_mmu_map_private_pfn(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_pfn_t pfn)
arch/x86/kvm/mmu/mmu.c
5253
static bool sync_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn,
arch/x86/kvm/mmu/mmu.c
6247
gfn_t gfn = gpa >> PAGE_SHIFT;
arch/x86/kvm/mmu/mmu.c
639
static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index)
arch/x86/kvm/mmu/mmu.c
677
gfn_t gfn, unsigned int access)
arch/x86/kvm/mmu/mmu.c
6888
static bool kvm_rmap_zap_gfn_range(struct kvm *kvm, gfn_t gfn_start, gfn_t gfn_end)
arch/x86/kvm/mmu/mmu.c
6894
gfn_t start, end;
arch/x86/kvm/mmu/mmu.c
6922
void kvm_zap_gfn_range(struct kvm *kvm, gfn_t gfn_start, gfn_t gfn_end)
arch/x86/kvm/mmu/mmu.c
698
gfn_t gfn = kvm_mmu_page_get_gfn(sp, index);
arch/x86/kvm/mmu/mmu.c
7033
gfn_t gfn;
arch/x86/kvm/mmu/mmu.c
7064
gfn_t gfn;
arch/x86/kvm/mmu/mmu.c
707
static struct kvm_lpage_info *lpage_info_slot(gfn_t gfn,
arch/x86/kvm/mmu/mmu.c
7106
gfn_t gfn;
arch/x86/kvm/mmu/mmu.c
7187
gfn_t start, gfn_t end,
arch/x86/kvm/mmu/mmu.c
725
gfn_t gfn, int count)
arch/x86/kvm/mmu/mmu.c
739
void kvm_mmu_gfn_disallow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn)
arch/x86/kvm/mmu/mmu.c
7395
gfn_t gfn = slot->base_gfn + i;
arch/x86/kvm/mmu/mmu.c
744
void kvm_mmu_gfn_allow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn)
arch/x86/kvm/mmu/mmu.c
753
gfn_t gfn;
arch/x86/kvm/mmu/mmu.c
7896
static bool hugepage_test_mixed(struct kvm_memory_slot *slot, gfn_t gfn,
arch/x86/kvm/mmu/mmu.c
7902
static void hugepage_clear_mixed(struct kvm_memory_slot *slot, gfn_t gfn,
arch/x86/kvm/mmu/mmu.c
7908
static void hugepage_set_mixed(struct kvm_memory_slot *slot, gfn_t gfn,
arch/x86/kvm/mmu/mmu.c
7947
gfn_t start = gfn_round_for_level(range->start, level);
arch/x86/kvm/mmu/mmu.c
7948
gfn_t end = gfn_round_for_level(range->end - 1, level);
arch/x86/kvm/mmu/mmu.c
7949
gfn_t nr_pages = KVM_PAGES_PER_HPAGE(level);
arch/x86/kvm/mmu/mmu.c
7978
gfn_t gfn, int level, unsigned long attrs)
arch/x86/kvm/mmu/mmu.c
8018
gfn_t nr_pages = KVM_PAGES_PER_HPAGE(level);
arch/x86/kvm/mmu/mmu.c
8019
gfn_t gfn = gfn_round_for_level(range->start, level);
arch/x86/kvm/mmu/mmu.c
8075
gfn_t end = gfn_round_for_level(slot->base_gfn + slot->npages, level);
arch/x86/kvm/mmu/mmu.c
8076
gfn_t start = gfn_round_for_level(slot->base_gfn, level);
arch/x86/kvm/mmu/mmu.c
8077
gfn_t nr_pages = KVM_PAGES_PER_HPAGE(level);
arch/x86/kvm/mmu/mmu.c
8078
gfn_t gfn;
arch/x86/kvm/mmu/mmu.c
812
gfn_t gfn;
arch/x86/kvm/mmu/mmu.c
843
gfn_t gfn,
arch/x86/kvm/mmu/mmu_internal.h
171
static inline gfn_t kvm_gfn_root_bits(const struct kvm *kvm, const struct kvm_mmu_page *root)
arch/x86/kvm/mmu/mmu_internal.h
197
static inline gfn_t gfn_round_for_level(gfn_t gfn, int level)
arch/x86/kvm/mmu/mmu_internal.h
203
gfn_t gfn, bool synchronizing, bool prefetch);
arch/x86/kvm/mmu/mmu_internal.h
205
void kvm_mmu_gfn_disallow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn);
arch/x86/kvm/mmu/mmu_internal.h
206
void kvm_mmu_gfn_allow_lpage(const struct kvm_memory_slot *slot, gfn_t gfn);
arch/x86/kvm/mmu/mmu_internal.h
212
static inline void kvm_flush_remote_tlbs_gfn(struct kvm *kvm, gfn_t gfn, int level)
arch/x86/kvm/mmu/mmu_internal.h
274
gfn_t gfn;
arch/x86/kvm/mmu/mmu_internal.h
405
const struct kvm_memory_slot *slot, gfn_t gfn);
arch/x86/kvm/mmu/mmu_internal.h
73
gfn_t gfn;
arch/x86/kvm/mmu/mmutrace.h
216
TP_PROTO(u64 *sptep, gfn_t gfn, u64 spte),
arch/x86/kvm/mmu/mmutrace.h
221
__field(gfn_t, gfn)
arch/x86/kvm/mmu/mmutrace.h
239
TP_PROTO(u64 addr, gfn_t gfn, unsigned access),
arch/x86/kvm/mmu/mmutrace.h
244
__field(gfn_t, gfn)
arch/x86/kvm/mmu/mmutrace.h
339
TP_PROTO(int level, gfn_t gfn, u64 *sptep),
arch/x86/kvm/mmu/mmutrace.h
397
TP_PROTO(int as_id, gfn_t gfn, int level, u64 old_spte, u64 new_spte),
arch/x86/kvm/mmu/page_track.c
114
struct kvm_memory_slot *slot, gfn_t gfn)
arch/x86/kvm/mmu/page_track.c
137
const struct kvm_memory_slot *slot, gfn_t gfn)
arch/x86/kvm/mmu/page_track.c
322
int kvm_write_track_add_gfn(struct kvm *kvm, gfn_t gfn)
arch/x86/kvm/mmu/page_track.c
352
int kvm_write_track_remove_gfn(struct kvm *kvm, gfn_t gfn)
arch/x86/kvm/mmu/page_track.c
75
static void update_gfn_write_track(struct kvm_memory_slot *slot, gfn_t gfn,
arch/x86/kvm/mmu/page_track.c
91
gfn_t gfn)
arch/x86/kvm/mmu/page_track.h
19
gfn_t gfn);
arch/x86/kvm/mmu/page_track.h
21
struct kvm_memory_slot *slot, gfn_t gfn);
arch/x86/kvm/mmu/page_track.h
24
const struct kvm_memory_slot *slot, gfn_t gfn);
arch/x86/kvm/mmu/paging_tmpl.h
104
static gfn_t gpte_to_gfn_lvl(pt_element_t gpte, int lvl)
arch/x86/kvm/mmu/paging_tmpl.h
205
gfn_t table_gfn;
arch/x86/kvm/mmu/paging_tmpl.h
309
gfn_t table_gfn;
arch/x86/kvm/mmu/paging_tmpl.h
322
gfn_t gfn;
arch/x86/kvm/mmu/paging_tmpl.h
536
gfn_t gfn;
arch/x86/kvm/mmu/paging_tmpl.h
621
gfn_t base_gfn = fault->gfn;
arch/x86/kvm/mmu/paging_tmpl.h
653
gfn_t table_gfn;
arch/x86/kvm/mmu/paging_tmpl.h
83
gfn_t table_gfn[PT_MAX_FULL_LEVELS];
arch/x86/kvm/mmu/paging_tmpl.h
904
gfn_t gfn;
arch/x86/kvm/mmu/paging_tmpl.h
91
gfn_t gfn;
arch/x86/kvm/mmu/paging_tmpl.h
96
static inline gfn_t pse36_gfn_delta(u32 gpte)
arch/x86/kvm/mmu/spte.c
188
unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn,
arch/x86/kvm/mmu/spte.h
545
unsigned int pte_access, gfn_t gfn, kvm_pfn_t pfn,
arch/x86/kvm/mmu/tdp_iter.c
40
int min_level, gfn_t next_last_level_gfn, gfn_t gfn_bits)
arch/x86/kvm/mmu/tdp_iter.h
139
int min_level, gfn_t next_last_level_gfn, gfn_t gfn_bits);
arch/x86/kvm/mmu/tdp_iter.h
81
gfn_t next_last_level_gfn;
arch/x86/kvm/mmu/tdp_iter.h
87
gfn_t yielded_gfn;
arch/x86/kvm/mmu/tdp_iter.h
93
gfn_t gfn;
arch/x86/kvm/mmu/tdp_iter.h
95
gfn_t gfn_bits;
arch/x86/kvm/mmu/tdp_mmu.c
1026
bool kvm_tdp_mmu_zap_leafs(struct kvm *kvm, gfn_t start, gfn_t end, bool flush)
arch/x86/kvm/mmu/tdp_mmu.c
1450
gfn_t start, gfn_t end, int min_level)
arch/x86/kvm/mmu/tdp_mmu.c
1560
gfn_t start, gfn_t end,
arch/x86/kvm/mmu/tdp_mmu.c
1642
gfn_t start, gfn_t end,
arch/x86/kvm/mmu/tdp_mmu.c
1669
gfn_t start, gfn_t end)
arch/x86/kvm/mmu/tdp_mmu.c
1715
gfn_t gfn, unsigned long mask, bool wrprot)
arch/x86/kvm/mmu/tdp_mmu.c
1761
gfn_t gfn, unsigned long mask,
arch/x86/kvm/mmu/tdp_mmu.c
1775
gfn_t start = parent->gfn;
arch/x86/kvm/mmu/tdp_mmu.c
1776
gfn_t end = start + KVM_PAGES_PER_HPAGE(parent->level);
arch/x86/kvm/mmu/tdp_mmu.c
1800
gfn_t start = slot->base_gfn;
arch/x86/kvm/mmu/tdp_mmu.c
1801
gfn_t end = start + slot->npages;
arch/x86/kvm/mmu/tdp_mmu.c
1884
gfn_t gfn, int min_level)
arch/x86/kvm/mmu/tdp_mmu.c
1920
struct kvm_memory_slot *slot, gfn_t gfn,
arch/x86/kvm/mmu/tdp_mmu.c
1944
gfn_t gfn = addr >> PAGE_SHIFT;
arch/x86/kvm/mmu/tdp_mmu.c
1968
u64 *kvm_tdp_mmu_fast_pf_get_last_sptep(struct kvm_vcpu *vcpu, gfn_t gfn,
arch/x86/kvm/mmu/tdp_mmu.c
225
gfn_t gfn, union kvm_mmu_page_role role)
arch/x86/kvm/mmu/tdp_mmu.c
323
static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn,
arch/x86/kvm/mmu/tdp_mmu.c
362
static void remove_external_spte(struct kvm *kvm, gfn_t gfn, u64 old_spte,
arch/x86/kvm/mmu/tdp_mmu.c
402
gfn_t base_gfn = sp->gfn;
arch/x86/kvm/mmu/tdp_mmu.c
411
gfn_t gfn = base_gfn + i * KVM_PAGES_PER_HPAGE(level);
arch/x86/kvm/mmu/tdp_mmu.c
498
static void *get_external_spt(gfn_t gfn, u64 new_spte, int level)
arch/x86/kvm/mmu/tdp_mmu.c
512
gfn_t gfn, u64 old_spte,
arch/x86/kvm/mmu/tdp_mmu.c
566
static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn,
arch/x86/kvm/mmu/tdp_mmu.c
743
u64 old_spte, u64 new_spte, gfn_t gfn, int level)
arch/x86/kvm/mmu/tdp_mmu.c
842
static inline gfn_t tdp_mmu_max_gfn_exclusive(void)
arch/x86/kvm/mmu/tdp_mmu.c
981
gfn_t start, gfn_t end, bool can_yield, bool flush)
arch/x86/kvm/mmu/tdp_mmu.h
113
u64 *kvm_tdp_mmu_fast_pf_get_last_sptep(struct kvm_vcpu *vcpu, gfn_t gfn,
arch/x86/kvm/mmu/tdp_mmu.h
66
bool kvm_tdp_mmu_zap_leafs(struct kvm *kvm, gfn_t start, gfn_t end, bool flush);
arch/x86/kvm/mmu/tdp_mmu.h
87
gfn_t gfn, unsigned long mask,
arch/x86/kvm/mmu/tdp_mmu.h
93
struct kvm_memory_slot *slot, gfn_t gfn,
arch/x86/kvm/mmu/tdp_mmu.h
98
gfn_t start, gfn_t end,
arch/x86/kvm/svm/sev.c
2280
static int sev_gmem_post_populate(struct kvm *kvm, gfn_t gfn, kvm_pfn_t pfn,
arch/x86/kvm/svm/sev.c
3968
gfn_t gfn;
arch/x86/kvm/svm/sev.c
4851
gfn_t gfn;
arch/x86/kvm/svm/sev.c
4987
int sev_gmem_prepare(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order)
arch/x86/kvm/svm/sev.c
4991
gfn_t gfn_aligned;
arch/x86/kvm/svm/svm.h
897
int sev_gmem_prepare(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order);
arch/x86/kvm/svm/svm.h
923
static inline int sev_gmem_prepare(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order)
arch/x86/kvm/vmx/tdx.c
1624
static int tdx_mem_page_add(struct kvm *kvm, gfn_t gfn, enum pg_level level,
arch/x86/kvm/vmx/tdx.c
1648
static int tdx_mem_page_aug(struct kvm *kvm, gfn_t gfn,
arch/x86/kvm/vmx/tdx.c
1668
static int tdx_sept_set_private_spte(struct kvm *kvm, gfn_t gfn,
arch/x86/kvm/vmx/tdx.c
1699
static int tdx_sept_link_private_spt(struct kvm *kvm, gfn_t gfn,
arch/x86/kvm/vmx/tdx.c
1764
static int tdx_sept_free_private_spt(struct kvm *kvm, gfn_t gfn,
arch/x86/kvm/vmx/tdx.c
1787
static void tdx_sept_remove_private_spte(struct kvm *kvm, gfn_t gfn,
arch/x86/kvm/vmx/tdx.c
3119
static int tdx_gmem_post_populate(struct kvm *kvm, gfn_t gfn, kvm_pfn_t pfn,
arch/x86/kvm/vmx/vmx.c
7031
const gfn_t gfn = APIC_DEFAULT_PHYS_BASE >> PAGE_SHIFT;
arch/x86/kvm/vmx/vmx.c
7919
u8 vmx_get_mt_mask(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio)
arch/x86/kvm/vmx/x86_ops.h
105
u8 vmx_get_mt_mask(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio);
arch/x86/kvm/x86.c
1062
gfn_t pdpt_gfn = cr3 >> PAGE_SHIFT;
arch/x86/kvm/x86.c
13806
static inline u32 kvm_async_pf_hash_fn(gfn_t gfn)
arch/x86/kvm/x86.c
13818
static void kvm_add_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn)
arch/x86/kvm/x86.c
13828
static u32 kvm_async_pf_gfn_slot(struct kvm_vcpu *vcpu, gfn_t gfn)
arch/x86/kvm/x86.c
13841
bool kvm_find_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn)
arch/x86/kvm/x86.c
13846
static void kvm_del_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn)
arch/x86/kvm/x86.c
14070
int kvm_arch_gmem_prepare(struct kvm *kvm, gfn_t gfn, kvm_pfn_t pfn, int max_order)
arch/x86/kvm/x86.h
360
gva_t gva, gfn_t gfn, unsigned access)
arch/x86/kvm/xen.c
780
gfn_t gfn = data->u.shared_info.gfn;
drivers/gpu/drm/i915/gvt/kvmgt.c
101
gfn_t gfn;
drivers/gpu/drm/i915/gvt/kvmgt.c
112
static void kvmgt_page_track_remove_region(gfn_t gfn, unsigned long nr_pages,
drivers/gpu/drm/i915/gvt/kvmgt.c
1576
static void kvmgt_page_track_remove_region(gfn_t gfn, unsigned long nr_pages,
drivers/gpu/drm/i915/gvt/kvmgt.c
229
static struct gvt_dma *__gvt_cache_find_gfn(struct intel_vgpu *vgpu, gfn_t gfn)
drivers/gpu/drm/i915/gvt/kvmgt.c
247
static int __gvt_cache_add(struct intel_vgpu *vgpu, gfn_t gfn,
drivers/gpu/drm/i915/gvt/kvmgt.c
350
__kvmgt_protect_table_find(struct intel_vgpu *info, gfn_t gfn)
drivers/gpu/drm/i915/gvt/kvmgt.c
366
static bool kvmgt_gfn_is_write_protected(struct intel_vgpu *info, gfn_t gfn)
drivers/gpu/drm/i915/gvt/kvmgt.c
374
static void kvmgt_protect_table_add(struct intel_vgpu *info, gfn_t gfn)
drivers/gpu/drm/i915/gvt/kvmgt.c
389
static void kvmgt_protect_table_del(struct intel_vgpu *info, gfn_t gfn)
drivers/gpu/drm/i915/gvt/kvmgt.c
93
gfn_t gfn;
include/linux/kvm_host.h
1135
gfn_t start)
include/linux/kvm_host.h
1193
static inline bool kvm_memslot_iter_is_valid(struct kvm_memslot_iter *iter, gfn_t end)
include/linux/kvm_host.h
1211
struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn);
include/linux/kvm_host.h
1213
struct kvm_memory_slot *kvm_vcpu_gfn_to_memslot(struct kvm_vcpu *vcpu, gfn_t gfn);
include/linux/kvm_host.h
1251
int kvm_prefetch_pages(struct kvm_memory_slot *slot, gfn_t gfn,
include/linux/kvm_host.h
1254
struct page *__gfn_to_page(struct kvm *kvm, gfn_t gfn, bool write);
include/linux/kvm_host.h
1255
static inline struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn)
include/linux/kvm_host.h
1260
unsigned long gfn_to_hva(struct kvm *kvm, gfn_t gfn);
include/linux/kvm_host.h
1261
unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable);
include/linux/kvm_host.h
1262
unsigned long gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn);
include/linux/kvm_host.h
1263
unsigned long gfn_to_hva_memslot_prot(struct kvm_memory_slot *slot, gfn_t gfn,
include/linux/kvm_host.h
1303
kvm_pfn_t __kvm_faultin_pfn(const struct kvm_memory_slot *slot, gfn_t gfn,
include/linux/kvm_host.h
1307
static inline kvm_pfn_t kvm_faultin_pfn(struct kvm_vcpu *vcpu, gfn_t gfn,
include/linux/kvm_host.h
1315
int kvm_read_guest_page(struct kvm *kvm, gfn_t gfn, void *data, int offset,
include/linux/kvm_host.h
1323
int kvm_write_guest_page(struct kvm *kvm, gfn_t gfn, const void *data,
include/linux/kvm_host.h
1378
bool kvm_is_visible_gfn(struct kvm *kvm, gfn_t gfn);
include/linux/kvm_host.h
1379
bool kvm_vcpu_is_visible_gfn(struct kvm_vcpu *vcpu, gfn_t gfn);
include/linux/kvm_host.h
1380
unsigned long kvm_host_page_size(struct kvm_vcpu *vcpu, gfn_t gfn);
include/linux/kvm_host.h
1381
void mark_page_dirty_in_slot(struct kvm *kvm, const struct kvm_memory_slot *memslot, gfn_t gfn);
include/linux/kvm_host.h
1382
void mark_page_dirty(struct kvm *kvm, gfn_t gfn);
include/linux/kvm_host.h
1383
void kvm_vcpu_mark_page_dirty(struct kvm_vcpu *vcpu, gfn_t gfn);
include/linux/kvm_host.h
1408
unsigned long kvm_vcpu_gfn_to_hva(struct kvm_vcpu *vcpu, gfn_t gfn);
include/linux/kvm_host.h
1409
unsigned long kvm_vcpu_gfn_to_hva_prot(struct kvm_vcpu *vcpu, gfn_t gfn, bool *writable);
include/linux/kvm_host.h
1410
int kvm_vcpu_read_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, void *data, int offset,
include/linux/kvm_host.h
1416
int kvm_vcpu_write_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, const void *data,
include/linux/kvm_host.h
1545
void kvm_flush_remote_tlbs_range(struct kvm *kvm, gfn_t gfn, u64 nr_pages);
include/linux/kvm_host.h
1558
void kvm_mmu_invalidate_range_add(struct kvm *kvm, gfn_t start, gfn_t end);
include/linux/kvm_host.h
1574
gfn_t gfn_offset,
include/linux/kvm_host.h
1691
gfn_t gfn, u64 nr_pages)
include/linux/kvm_host.h
1696
int kvm_arch_flush_remote_tlbs_range(struct kvm *kvm, gfn_t gfn, u64 nr_pages);
include/linux/kvm_host.h
1808
try_get_memslot(struct kvm_memory_slot *slot, gfn_t gfn)
include/linux/kvm_host.h
1827
search_memslots(struct kvm_memslots *slots, gfn_t gfn, bool approx)
include/linux/kvm_host.h
1848
____gfn_to_memslot(struct kvm_memslots *slots, gfn_t gfn, bool approx)
include/linux/kvm_host.h
1872
__gfn_to_memslot(struct kvm_memslots *slots, gfn_t gfn)
include/linux/kvm_host.h
1878
__gfn_to_hva_memslot(const struct kvm_memory_slot *slot, gfn_t gfn)
include/linux/kvm_host.h
1891
static inline int memslot_id(struct kvm *kvm, gfn_t gfn)
include/linux/kvm_host.h
1896
static inline gfn_t
include/linux/kvm_host.h
1899
gfn_t gfn_offset = (hva - slot->userspace_addr) >> PAGE_SHIFT;
include/linux/kvm_host.h
1904
static inline gpa_t gfn_to_gpa(gfn_t gfn)
include/linux/kvm_host.h
1909
static inline gfn_t gpa_to_gfn(gpa_t gpa)
include/linux/kvm_host.h
1911
return (gfn_t)(gpa >> PAGE_SHIFT);
include/linux/kvm_host.h
2129
gfn_t gfn)
include/linux/kvm_host.h
2165
gfn_t gfn)
include/linux/kvm_host.h
2517
static inline unsigned long kvm_get_memory_attributes(struct kvm *kvm, gfn_t gfn)
include/linux/kvm_host.h
2522
bool kvm_range_has_memory_attributes(struct kvm *kvm, gfn_t start, gfn_t end,
include/linux/kvm_host.h
2529
static inline bool kvm_mem_is_private(struct kvm *kvm, gfn_t gfn)
include/linux/kvm_host.h
2534
static inline bool kvm_mem_is_private(struct kvm *kvm, gfn_t gfn)
include/linux/kvm_host.h
2542
gfn_t gfn, kvm_pfn_t *pfn, struct page **page,
include/linux/kvm_host.h
2546
struct kvm_memory_slot *slot, gfn_t gfn,
include/linux/kvm_host.h
2556
int kvm_arch_gmem_prepare(struct kvm *kvm, gfn_t gfn, kvm_pfn_t pfn, int max_order);
include/linux/kvm_host.h
2581
typedef int (*kvm_gmem_populate_cb)(struct kvm *kvm, gfn_t gfn, kvm_pfn_t pfn,
include/linux/kvm_host.h
2584
long kvm_gmem_populate(struct kvm *kvm, gfn_t gfn, void __user *src, long npages,
include/linux/kvm_host.h
267
gfn_t start;
include/linux/kvm_host.h
268
gfn_t end;
include/linux/kvm_host.h
596
gfn_t base_gfn;
include/linux/kvm_host.h
853
gfn_t mmu_invalidate_range_start;
include/linux/kvm_host.h
854
gfn_t mmu_invalidate_range_end;
include/trace/events/kvm.h
368
TP_PROTO(gfn_t start, gfn_t end, unsigned long attr),
include/trace/events/kvm.h
372
__field(gfn_t, start)
include/trace/events/kvm.h
373
__field(gfn_t, end)
virt/kvm/guest_memfd.c
57
static pgoff_t kvm_gmem_get_index(struct kvm_memory_slot *slot, gfn_t gfn)
virt/kvm/guest_memfd.c
652
BUILD_BUG_ON(sizeof(gfn_t) != sizeof(slot->gmem.pgoff));
virt/kvm/guest_memfd.c
67
gfn_t gfn = slot->base_gfn + index - slot->gmem.pgoff;
virt/kvm/guest_memfd.c
789
gfn_t gfn, kvm_pfn_t *pfn, struct page **page,
virt/kvm/guest_memfd.c
825
struct file *file, gfn_t gfn, struct page *src_page,
virt/kvm/guest_memfd.c
86
gfn_t gfn, struct folio *folio)
virt/kvm/guest_memfd.c
861
long kvm_gmem_populate(struct kvm *kvm, gfn_t start_gfn, void __user *src, long npages,
virt/kvm/kvm_main.c
1989
gfn_t start, gfn_t end)
virt/kvm/kvm_main.c
2008
gfn_t base_gfn;
virt/kvm/kvm_main.c
2275
gfn_t offset;
virt/kvm/kvm_main.c
2344
gfn_t offset;
virt/kvm/kvm_main.c
2441
bool kvm_range_has_memory_attributes(struct kvm *kvm, gfn_t start, gfn_t end,
virt/kvm/kvm_main.c
2544
static int kvm_vm_set_mem_attributes(struct kvm *kvm, gfn_t start, gfn_t end,
virt/kvm/kvm_main.c
2609
gfn_t start, end;
virt/kvm/kvm_main.c
2635
struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn)
virt/kvm/kvm_main.c
2641
struct kvm_memory_slot *kvm_vcpu_gfn_to_memslot(struct kvm_vcpu *vcpu, gfn_t gfn)
virt/kvm/kvm_main.c
2675
bool kvm_is_visible_gfn(struct kvm *kvm, gfn_t gfn)
virt/kvm/kvm_main.c
2683
bool kvm_vcpu_is_visible_gfn(struct kvm_vcpu *vcpu, gfn_t gfn)
virt/kvm/kvm_main.c
2691
unsigned long kvm_host_page_size(struct kvm_vcpu *vcpu, gfn_t gfn)
virt/kvm/kvm_main.c
2720
static unsigned long __gfn_to_hva_many(const struct kvm_memory_slot *slot, gfn_t gfn,
virt/kvm/kvm_main.c
2721
gfn_t *nr_pages, bool write)
virt/kvm/kvm_main.c
2735
static unsigned long gfn_to_hva_many(struct kvm_memory_slot *slot, gfn_t gfn,
virt/kvm/kvm_main.c
2736
gfn_t *nr_pages)
virt/kvm/kvm_main.c
2742
gfn_t gfn)
virt/kvm/kvm_main.c
2748
unsigned long gfn_to_hva(struct kvm *kvm, gfn_t gfn)
virt/kvm/kvm_main.c
2754
unsigned long kvm_vcpu_gfn_to_hva(struct kvm_vcpu *vcpu, gfn_t gfn)
virt/kvm/kvm_main.c
2769
gfn_t gfn, bool *writable)
virt/kvm/kvm_main.c
2779
unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable)
virt/kvm/kvm_main.c
2786
unsigned long kvm_vcpu_gfn_to_hva_prot(struct kvm_vcpu *vcpu, gfn_t gfn, bool *writable)
virt/kvm/kvm_main.c
3056
kvm_pfn_t __kvm_faultin_pfn(const struct kvm_memory_slot *slot, gfn_t gfn,
virt/kvm/kvm_main.c
3078
int kvm_prefetch_pages(struct kvm_memory_slot *slot, gfn_t gfn,
virt/kvm/kvm_main.c
3082
gfn_t entry = 0;
virt/kvm/kvm_main.c
3102
struct page *__gfn_to_page(struct kvm *kvm, gfn_t gfn, bool write)
virt/kvm/kvm_main.c
3117
int __kvm_vcpu_map(struct kvm_vcpu *vcpu, gfn_t gfn, struct kvm_host_map *map,
virt/kvm/kvm_main.c
315
void kvm_flush_remote_tlbs_range(struct kvm *kvm, gfn_t gfn, u64 nr_pages)
virt/kvm/kvm_main.c
3188
static int __kvm_read_guest_page(struct kvm_memory_slot *slot, gfn_t gfn,
virt/kvm/kvm_main.c
3206
int kvm_read_guest_page(struct kvm *kvm, gfn_t gfn, void *data, int offset,
virt/kvm/kvm_main.c
3215
int kvm_vcpu_read_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, void *data,
virt/kvm/kvm_main.c
3226
gfn_t gfn = gpa >> PAGE_SHIFT;
virt/kvm/kvm_main.c
3246
gfn_t gfn = gpa >> PAGE_SHIFT;
virt/kvm/kvm_main.c
3264
static int __kvm_read_guest_atomic(struct kvm_memory_slot *slot, gfn_t gfn,
virt/kvm/kvm_main.c
3287
gfn_t gfn = gpa >> PAGE_SHIFT;
virt/kvm/kvm_main.c
3297
struct kvm_memory_slot *memslot, gfn_t gfn,
virt/kvm/kvm_main.c
3316
int kvm_write_guest_page(struct kvm *kvm, gfn_t gfn,
virt/kvm/kvm_main.c
3325
int kvm_vcpu_write_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn,
virt/kvm/kvm_main.c
3337
gfn_t gfn = gpa >> PAGE_SHIFT;
virt/kvm/kvm_main.c
3358
gfn_t gfn = gpa >> PAGE_SHIFT;
virt/kvm/kvm_main.c
3381
gfn_t start_gfn = gpa >> PAGE_SHIFT;
virt/kvm/kvm_main.c
3382
gfn_t end_gfn = (gpa + len - 1) >> PAGE_SHIFT;
virt/kvm/kvm_main.c
3383
gfn_t nr_pages_needed = end_gfn - start_gfn + 1;
virt/kvm/kvm_main.c
3384
gfn_t nr_pages_avail;
virt/kvm/kvm_main.c
3503
gfn_t gfn = gpa >> PAGE_SHIFT;
virt/kvm/kvm_main.c
3522
gfn_t gfn)
virt/kvm/kvm_main.c
3545
void mark_page_dirty(struct kvm *kvm, gfn_t gfn)
virt/kvm/kvm_main.c
3554
void kvm_vcpu_mark_page_dirty(struct kvm_vcpu *vcpu, gfn_t gfn)
virt/kvm/kvm_main.c
694
void kvm_mmu_invalidate_range_add(struct kvm *kvm, gfn_t start, gfn_t end)
virt/kvm/kvm_mm.h
26
const gfn_t gfn;
virt/kvm/pfncache.c
299
gfn_t gfn = gpa_to_gfn(gpa);