Symbol: gva_t
arch/mips/include/asm/kvm_host.h
738
gpa_t (*gva_to_gpa)(gva_t gva);
arch/mips/kvm/vz.c
197
static gpa_t kvm_vz_gva_to_gpa_cb(gva_t gva)
arch/powerpc/include/asm/kvm_book3s.h
161
extern long kvmppc_hv_find_lock_hpte(struct kvm *kvm, gva_t eaddr,
arch/powerpc/include/asm/kvm_book3s.h
164
unsigned long gpa, gva_t ea, int is_store);
arch/powerpc/include/asm/kvm_book3s.h
180
gva_t eaddr, void *to, void *from,
arch/powerpc/include/asm/kvm_book3s.h
182
extern long kvmhv_copy_from_guest_radix(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/include/asm/kvm_book3s.h
184
extern long kvmhv_copy_to_guest_radix(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/include/asm/kvm_book3s.h
186
extern int kvmppc_mmu_walk_radix_tree(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/include/asm/kvm_book3s.h
189
extern int kvmppc_mmu_radix_translate_table(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/include/asm/kvm_book3s.h
192
extern int kvmppc_mmu_radix_xlate(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/include/asm/kvm_host.h
394
int (*slbfee)(struct kvm_vcpu *vcpu, gva_t eaddr, ulong *ret_slb);
arch/powerpc/include/asm/kvm_host.h
400
int (*xlate)(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/include/asm/kvm_host.h
404
u64 (*ea_to_vp)(struct kvm_vcpu *vcpu, gva_t eaddr, bool data);
arch/powerpc/include/asm/kvm_host.h
737
gva_t vaddr_accessed;
arch/powerpc/include/asm/kvm_ppc.h
110
extern int kvmppc_mmu_dtlb_index(struct kvm_vcpu *vcpu, gva_t eaddr);
arch/powerpc/include/asm/kvm_ppc.h
111
extern int kvmppc_mmu_itlb_index(struct kvm_vcpu *vcpu, gva_t eaddr);
arch/powerpc/include/asm/kvm_ppc.h
113
gva_t eaddr);
arch/powerpc/kvm/book3s_32_mmu.c
121
static u32 kvmppc_mmu_book3s_32_get_ptem(u32 sre, gva_t eaddr, bool primary)
arch/powerpc/kvm/book3s_32_mmu.c
127
static int kvmppc_mmu_book3s_32_xlate_bat(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_32_mmu.c
182
static int kvmppc_mmu_book3s_32_xlate_pte(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_32_mmu.c
290
static int kvmppc_mmu_book3s_32_xlate(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_32_mmu.c
69
static int kvmppc_mmu_book3s_32_xlate_bat(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_32_mmu.c
75
static u32 find_sr(struct kvm_vcpu *vcpu, gva_t eaddr)
arch/powerpc/kvm/book3s_32_mmu.c
80
static u64 kvmppc_mmu_book3s_32_ea_to_vp(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_32_mmu.c
94
u32 sre, gva_t eaddr,
arch/powerpc/kvm/book3s_64_mmu.c
109
static u32 kvmppc_mmu_book3s_64_get_page(struct kvmppc_slb *slbe, gva_t eaddr)
arch/powerpc/kvm/book3s_64_mmu.c
117
struct kvmppc_slb *slbe, gva_t eaddr,
arch/powerpc/kvm/book3s_64_mmu.c
155
static u64 kvmppc_mmu_book3s_64_get_avpn(struct kvmppc_slb *slbe, gva_t eaddr)
arch/powerpc/kvm/book3s_64_mmu.c
191
static int kvmppc_mmu_book3s_64_xlate(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_64_mmu.c
29
gva_t eaddr)
arch/powerpc/kvm/book3s_64_mmu.c
412
static int kvmppc_mmu_book3s_64_slbfee(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_64_mmu.c
73
static u64 kvmppc_slb_calc_vpn(struct kvmppc_slb *slb, gva_t eaddr)
arch/powerpc/kvm/book3s_64_mmu.c
81
static u64 kvmppc_mmu_book3s_64_ea_to_vp(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_64_mmu_hv.c
312
gva_t eaddr)
arch/powerpc/kvm/book3s_64_mmu_hv.c
341
static int kvmppc_mmu_book3s_64_hv_xlate(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_64_mmu_hv.c
434
unsigned long gpa, gva_t ea, int is_store)
arch/powerpc/kvm/book3s_64_mmu_radix.c
124
long kvmhv_copy_from_guest_radix(struct kvm_vcpu *vcpu, gva_t eaddr, void *to,
arch/powerpc/kvm/book3s_64_mmu_radix.c
136
long kvmhv_copy_to_guest_radix(struct kvm_vcpu *vcpu, gva_t eaddr, void *from,
arch/powerpc/kvm/book3s_64_mmu_radix.c
142
int kvmppc_mmu_walk_radix_tree(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_64_mmu_radix.c
238
int kvmppc_mmu_radix_translate_table(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_64_mmu_radix.c
269
int kvmppc_mmu_radix_xlate(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_64_mmu_radix.c
36
gva_t eaddr, void *to, void *from,
arch/powerpc/kvm/book3s_64_mmu_radix.c
99
static long kvmhv_copy_tofrom_guest_radix(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_hv_nested.c
601
gva_t eaddr = kvmppc_get_gpr(vcpu, 6);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1097
long kvmppc_hv_find_lock_hpte(struct kvm *kvm, gva_t eaddr, unsigned long slb_v,
arch/powerpc/kvm/e500.h
127
int kvmppc_e500_emul_tlbivax(struct kvm_vcpu *vcpu, gva_t ea);
arch/powerpc/kvm/e500.h
128
int kvmppc_e500_emul_tlbilx(struct kvm_vcpu *vcpu, int type, gva_t ea);
arch/powerpc/kvm/e500.h
129
int kvmppc_e500_emul_tlbsx(struct kvm_vcpu *vcpu, gva_t ea);
arch/powerpc/kvm/e500.h
154
static inline gva_t get_tlb_eaddr(const struct kvm_book3e_206_tlb_entry *tlbe)
arch/powerpc/kvm/e500.h
165
static inline gva_t get_tlb_end(const struct kvm_book3e_206_tlb_entry *tlbe)
arch/powerpc/kvm/e500_emulate.c
135
gva_t ea;
arch/powerpc/kvm/e500_mmu.c
127
gva_t eaddr, int as)
arch/powerpc/kvm/e500_mmu.c
155
gva_t eaddr;
arch/powerpc/kvm/e500_mmu.c
245
int kvmppc_e500_emul_tlbivax(struct kvm_vcpu *vcpu, gva_t ea)
arch/powerpc/kvm/e500_mmu.c
293
gva_t ea)
arch/powerpc/kvm/e500_mmu.c
307
int kvmppc_e500_emul_tlbilx(struct kvm_vcpu *vcpu, int type, gva_t ea)
arch/powerpc/kvm/e500_mmu.c
341
int kvmppc_e500_emul_tlbsx(struct kvm_vcpu *vcpu, gva_t ea)
arch/powerpc/kvm/e500_mmu.c
454
gva_t eaddr, unsigned int pid, int as)
arch/powerpc/kvm/e500_mmu.c
473
gva_t eaddr;
arch/powerpc/kvm/e500_mmu.c
48
static int tlb0_set_base(gva_t addr, int sets, int ways)
arch/powerpc/kvm/e500_mmu.c
495
int kvmppc_mmu_itlb_index(struct kvm_vcpu *vcpu, gva_t eaddr)
arch/powerpc/kvm/e500_mmu.c
502
int kvmppc_mmu_dtlb_index(struct kvm_vcpu *vcpu, gva_t eaddr)
arch/powerpc/kvm/e500_mmu.c
524
gva_t eaddr)
arch/powerpc/kvm/e500_mmu.c
58
static int gtlb0_set_base(struct kvmppc_vcpu_e500 *vcpu_e500, gva_t addr)
arch/powerpc/kvm/e500_mmu.c
81
gva_t eaddr, int tlbsel, unsigned int pid, int as)
arch/powerpc/kvm/e500_mmu_host.c
587
gva_t geaddr;
arch/powerpc/kvm/e500mc.c
60
gva_t eaddr;
arch/x86/include/asm/kvm_host.h
1807
void (*flush_tlb_gva)(struct kvm_vcpu *vcpu, gva_t addr);
arch/x86/include/asm/kvm_host.h
1960
gva_t (*get_untagged_addr)(struct kvm_vcpu *vcpu, gva_t gva, unsigned int flags);
arch/x86/include/asm/kvm_host.h
2292
gpa_t kvm_mmu_gva_to_gpa_read(struct kvm_vcpu *vcpu, gva_t gva,
arch/x86/include/asm/kvm_host.h
2294
gpa_t kvm_mmu_gva_to_gpa_write(struct kvm_vcpu *vcpu, gva_t gva,
arch/x86/include/asm/kvm_host.h
2296
gpa_t kvm_mmu_gva_to_gpa_system(struct kvm_vcpu *vcpu, gva_t gva,
arch/x86/include/asm/kvm_host.h
2322
void kvm_mmu_invlpg(struct kvm_vcpu *vcpu, gva_t gva);
arch/x86/include/asm/kvm_host.h
2325
void kvm_mmu_invpcid_gva(struct kvm_vcpu *vcpu, gva_t gva, unsigned long pcid);
arch/x86/kvm/emulate.c
2008
gva_t cs_addr;
arch/x86/kvm/emulate.c
2009
gva_t eip_addr;
arch/x86/kvm/hyperv.c
1971
gva_t gva;
arch/x86/kvm/kvm_emulate.h
243
gva_t (*get_untagged_addr)(struct x86_emulate_ctxt *ctxt, gva_t addr,
arch/x86/kvm/kvm_emulate.h
246
bool (*is_canonical_addr)(struct x86_emulate_ctxt *ctxt, gva_t addr,
arch/x86/kvm/mmu/mmu.c
3514
gva_t gva = fault->is_tdp ? 0 : fault->addr;
arch/x86/kvm/mmu/mmu.c
4496
static void shadow_page_table_clear_flood(struct kvm_vcpu *vcpu, gva_t addr)
arch/x86/kvm/mmu/mmu.c
6553
void kvm_mmu_invlpg(struct kvm_vcpu *vcpu, gva_t gva)
arch/x86/kvm/mmu/mmu.c
6571
void kvm_mmu_invpcid_gva(struct kvm_vcpu *vcpu, gva_t gva, unsigned long pcid)
arch/x86/kvm/svm/svm.c
2366
gva_t gva = kvm_rax_read(vcpu);
arch/x86/kvm/svm/svm.c
3172
gva_t gva;
arch/x86/kvm/svm/svm.c
4062
static void svm_flush_tlb_gva(struct kvm_vcpu *vcpu, gva_t gva)
arch/x86/kvm/trace.h
956
TP_PROTO(gva_t gva, gpa_t gpa, bool write, bool gpa_match),
arch/x86/kvm/trace.h
960
__field(gva_t, gva)
arch/x86/kvm/vmx/main.c
533
static void vt_flush_tlb_gva(struct kvm_vcpu *vcpu, gva_t addr)
arch/x86/kvm/vmx/nested.c
5219
u32 vmx_instruction_info, bool wr, int len, gva_t *ret)
arch/x86/kvm/vmx/nested.c
5221
gva_t off;
arch/x86/kvm/vmx/nested.c
5251
off = (gva_t)sign_extend64(off, 31);
arch/x86/kvm/vmx/nested.c
5253
off = (gva_t)sign_extend64(off, 15);
arch/x86/kvm/vmx/nested.c
5346
gva_t gva;
arch/x86/kvm/vmx/nested.c
5641
gva_t gva = 0;
arch/x86/kvm/vmx/nested.c
5747
gva_t gva;
arch/x86/kvm/vmx/nested.c
5934
gva_t gva;
arch/x86/kvm/vmx/nested.c
5962
gva_t gva;
arch/x86/kvm/vmx/nested.c
6041
gva_t gva;
arch/x86/kvm/vmx/nested.h
54
u32 vmx_instruction_info, bool wr, int len, gva_t *ret);
arch/x86/kvm/vmx/sgx.c
105
static int sgx_inject_fault(struct kvm_vcpu *vcpu, gva_t gva, int trapnr)
arch/x86/kvm/vmx/sgx.c
144
gva_t secs_gva)
arch/x86/kvm/vmx/sgx.c
219
gva_t pageinfo_gva, secs_gva;
arch/x86/kvm/vmx/sgx.c
220
gva_t metadata_gva, contents_gva;
arch/x86/kvm/vmx/sgx.c
25
int size, int alignment, gva_t *gva)
arch/x86/kvm/vmx/sgx.c
301
gva_t sig_gva, secs_gva, token_gva;
arch/x86/kvm/vmx/sgx.c
74
static int sgx_gva_to_gpa(struct kvm_vcpu *vcpu, gva_t gva, bool write,
arch/x86/kvm/vmx/vmx.c
3437
void vmx_flush_tlb_gva(struct kvm_vcpu *vcpu, gva_t addr)
arch/x86/kvm/vmx/vmx.c
571
noinline void invvpid_error(unsigned long ext, u16 vpid, gva_t gva)
arch/x86/kvm/vmx/vmx.c
6219
gva_t gva;
arch/x86/kvm/vmx/vmx.c
8590
gva_t vmx_get_untagged_addr(struct kvm_vcpu *vcpu, gva_t gva, unsigned int flags)
arch/x86/kvm/vmx/vmx.h
397
gva_t vmx_get_untagged_addr(struct kvm_vcpu *vcpu, gva_t gva, unsigned int flags);
arch/x86/kvm/vmx/vmx_ops.h
17
void invvpid_error(unsigned long ext, u16 vpid, gva_t gva);
arch/x86/kvm/vmx/vmx_ops.h
306
static inline void __invvpid(unsigned long ext, u16 vpid, gva_t gva)
arch/x86/kvm/vmx/vmx_ops.h
347
static inline void vpid_sync_vcpu_addr(int vpid, gva_t addr)
arch/x86/kvm/vmx/x86_ops.h
84
void vmx_flush_tlb_gva(struct kvm_vcpu *vcpu, gva_t addr);
arch/x86/kvm/x86.c
14110
void kvm_fixup_and_inject_pf_error(struct kvm_vcpu *vcpu, gva_t gva, u16 error_code)
arch/x86/kvm/x86.c
14164
int kvm_handle_invpcid(struct kvm_vcpu *vcpu, unsigned long type, gva_t gva)
arch/x86/kvm/x86.c
7840
gpa_t kvm_mmu_gva_to_gpa_read(struct kvm_vcpu *vcpu, gva_t gva,
arch/x86/kvm/x86.c
7850
gpa_t kvm_mmu_gva_to_gpa_write(struct kvm_vcpu *vcpu, gva_t gva,
arch/x86/kvm/x86.c
7862
gpa_t kvm_mmu_gva_to_gpa_system(struct kvm_vcpu *vcpu, gva_t gva,
arch/x86/kvm/x86.c
7870
static int kvm_read_guest_virt_helper(gva_t addr, void *val, unsigned int bytes,
arch/x86/kvm/x86.c
7903
gva_t addr, void *val, unsigned int bytes,
arch/x86/kvm/x86.c
7930
gva_t addr, void *val, unsigned int bytes,
arch/x86/kvm/x86.c
7948
gva_t addr, void *val, unsigned int bytes,
arch/x86/kvm/x86.c
7962
static int kvm_write_guest_virt_helper(gva_t addr, void *val, unsigned int bytes,
arch/x86/kvm/x86.c
7992
static int emulator_write_std(struct x86_emulate_ctxt *ctxt, gva_t addr, void *val,
arch/x86/kvm/x86.c
8008
int kvm_write_guest_virt_system(struct kvm_vcpu *vcpu, gva_t addr, void *val,
arch/x86/kvm/x86.c
8877
static gva_t emulator_get_untagged_addr(struct x86_emulate_ctxt *ctxt,
arch/x86/kvm/x86.c
8878
gva_t addr, unsigned int flags)
arch/x86/kvm/x86.c
8888
gva_t addr, unsigned int flags)
arch/x86/kvm/x86.h
360
gva_t gva, gfn_t gfn, unsigned access)
arch/x86/kvm/x86.h
386
#define MMIO_GVA_ANY (~(gva_t)0)
arch/x86/kvm/x86.h
388
static inline void vcpu_clear_mmio_info(struct kvm_vcpu *vcpu, gva_t gva)
arch/x86/kvm/x86.h
456
gva_t addr, void *val, unsigned int bytes,
arch/x86/kvm/x86.h
460
gva_t addr, void *val, unsigned int bytes,
arch/x86/kvm/x86.h
470
void kvm_fixup_and_inject_pf_error(struct kvm_vcpu *vcpu, gva_t gva, u16 error_code);
arch/x86/kvm/x86.h
656
int kvm_handle_invpcid(struct kvm_vcpu *vcpu, unsigned long type, gva_t gva);
arch/x86/kvm/xen.c
1525
if (kvm_read_guest_virt(vcpu, (gva_t)sched_poll.ports, ports,