Symbol: gpa_to_gfn
arch/arm64/kvm/guest.c
1001
gfn = gpa_to_gfn(guest_ipa);
arch/loongarch/kvm/vcpu.c
1839
mark_page_dirty_in_slot(vcpu->kvm, ghc->memslot, gpa_to_gfn(ghc->gpa));
arch/loongarch/kvm/vcpu.c
206
mark_page_dirty_in_slot(vcpu->kvm, ghc->memslot, gpa_to_gfn(ghc->gpa));
arch/powerpc/kvm/book3s_xive_native.c
647
gfn = gpa_to_gfn(kvm_eq.qaddr);
arch/powerpc/kvm/book3s_xive_native.c
908
mark_page_dirty(vcpu->kvm, gpa_to_gfn(q->guest_qaddr));
arch/s390/kvm/diag.c
59
do_discard_gfn_range(vcpu, gpa_to_gfn(start), gpa_to_gfn(end));
arch/s390/kvm/diag.c
67
do_discard_gfn_range(vcpu, gpa_to_gfn(start), gpa_to_gfn(prefix));
arch/s390/kvm/diag.c
72
do_discard_gfn_range(vcpu, gpa_to_gfn(prefix) + 2, gpa_to_gfn(end));
arch/s390/kvm/faultin.h
32
rc = kvm_s390_get_guest_page(kvm, f, gpa_to_gfn(gaddr), false);
arch/s390/kvm/gaccess.c
1148
.gfn = gpa_to_gfn(gpa),
arch/s390/kvm/gaccess.c
1300
return kvm_s390_get_guest_page(kvm, entries + LEVEL_MEM, gpa_to_gfn(saddr), false);
arch/s390/kvm/gaccess.c
1305
if (!asce_contains_gfn(asce, gpa_to_gfn(saddr)))
arch/s390/kvm/gaccess.c
1433
rc = gmap_insert_rmap(sg, f->gfn, gpa_to_gfn(raddr), TABLE_TYPE_PAGE_TABLE);
arch/s390/kvm/gaccess.c
1458
pgste = __dat_ptep_xchg(ptep, pgste, newpte, gpa_to_gfn(raddr), sg->asce, uses_skeys(sg));
arch/s390/kvm/gaccess.c
1474
gfn = f->gfn & gpa_to_gfn(is_pmd(*table) ? _SEGMENT_MASK : _REGION3_MASK);
arch/s390/kvm/gaccess.c
1476
rc = gmap_insert_rmap(sg, gfn, gpa_to_gfn(raddr), host->h.tt);
arch/s390/kvm/gaccess.c
1499
gfn = gpa_to_gfn(raddr);
arch/s390/kvm/gaccess.c
1520
rc = dat_entry_walk(NULL, gpa_to_gfn(saddr), sg->asce, DAT_WALK_ANY, TABLE_TYPE_PAGE_TABLE,
arch/s390/kvm/gaccess.c
1544
rc = gmap_protect_rmap(mc, sg, entries[i].gfn, gpa_to_gfn(saddr),
arch/s390/kvm/gaccess.c
1565
rc = dat_entry_walk(mc, gpa_to_gfn(saddr), sg->asce, flags, l, &table, &ptep);
arch/s390/kvm/gaccess.c
660
r = dat_get_storage_key(kvm->arch.gmap->asce, gpa_to_gfn(gpa), &storage_key);
arch/s390/kvm/gaccess.c
719
r = dat_get_storage_key(vcpu->arch.gmap->asce, gpa_to_gfn(gpa), &storage_key);
arch/s390/kvm/gaccess.c
822
const gfn_t gfn = gpa_to_gfn(gpa);
arch/s390/kvm/gaccess.c
893
.gfn = gpa_to_gfn(gpa),
arch/s390/kvm/gmap.c
265
prefix_gfn = gpa_to_gfn(kvm_s390_get_prefix(vcpu));
arch/s390/kvm/gmap.c
742
rc = dat_entry_walk(NULL, gpa_to_gfn(*gaddr), gmap->asce, DAT_WALK_CONTINUE,
arch/s390/kvm/gmap.c
773
gfn = gpa_to_gfn(*gaddr);
arch/s390/kvm/gmap.c
787
rc = gmap_ucas_map_one(mc, gmap, gpa_to_gfn(translated_address), gfn, true);
arch/s390/kvm/intercept.c
373
rc = kvm_s390_faultin_gfn_simple(vcpu, NULL, gpa_to_gfn(srcaddr), false);
arch/s390/kvm/intercept.c
385
rc = kvm_s390_faultin_gfn_simple(vcpu, NULL, gpa_to_gfn(dstaddr), true);
arch/s390/kvm/kvm-s390.c
1013
ret = gmap_set_limit(kvm->arch.gmap, gpa_to_gfn(new_limit));
arch/s390/kvm/kvm-s390.c
3262
kvm->arch.gmap = gmap_new(kvm, gpa_to_gfn(kvm->arch.mem_limit));
arch/s390/kvm/kvm-s390.c
4245
gfn = gpa_to_gfn(gaddr);
arch/s390/kvm/kvm-s390.c
4539
f.gfn = gpa_to_gfn(gaddr);
arch/s390/kvm/kvm-s390.c
5498
r = gmap_ucas_map(vcpu->arch.gmap, gpa_to_gfn(ucas.user_addr),
arch/s390/kvm/kvm-s390.c
5499
gpa_to_gfn(ucas.vcpu_addr),
arch/s390/kvm/kvm-s390.c
5516
gmap_ucas_unmap(vcpu->arch.gmap, gpa_to_gfn(ucas.vcpu_addr),
arch/s390/kvm/kvm-s390.c
5530
r = kvm_s390_faultin_gfn_simple(vcpu, NULL, gpa_to_gfn(gaddr), false);
arch/s390/kvm/kvm-s390.c
561
gfn_t gfn = gpa_to_gfn(addr);
arch/s390/kvm/kvm-s390.h
294
hva_t hva = gfn_to_hva(kvm, gpa_to_gfn(gpa));
arch/s390/kvm/pci.c
250
hva = gfn_to_hva(kvm, gpa_to_gfn((gpa_t)fib->fmt0.aibv));
arch/s390/kvm/pci.c
265
hva = gfn_to_hva(kvm, gpa_to_gfn((gpa_t)fib->fmt0.aisb));
arch/s390/kvm/priv.c
1115
gpa_to_gfn(start), key,
arch/s390/kvm/priv.c
1192
if (dat_entry_walk(NULL, gpa_to_gfn(cbrl[i]), vcpu->arch.gmap->asce,
arch/s390/kvm/priv.c
1474
gfn_to_hva_prot(vcpu->kvm, gpa_to_gfn(gpa), &writable);
arch/s390/kvm/priv.c
279
rc = dat_get_storage_key(vcpu->arch.gmap->asce, gpa_to_gfn(gaddr), &key);
arch/s390/kvm/priv.c
310
rc = dat_reset_reference_bit(vcpu->arch.gmap->asce, gpa_to_gfn(gaddr));
arch/s390/kvm/priv.c
363
gpa_to_gfn(start), key, &oldkey,
arch/s390/kvm/pv.c
133
.gfn = gpa_to_gfn(gaddr),
arch/s390/kvm/pv.c
189
page = gfn_to_page(kvm, gpa_to_gfn(gaddr));
arch/s390/kvm/pv.c
489
gmap_pv_destroy_range(kvm->arch.gmap, 0, gpa_to_gfn(SZ_2G), false);
arch/s390/kvm/pv.c
785
ret = kvm_s390_faultin_gfn_simple(NULL, kvm, gpa_to_gfn(addr), true);
arch/s390/kvm/vsie.c
670
page = gfn_to_page(kvm, gpa_to_gfn(gpa));
arch/s390/kvm/vsie.c
682
mark_page_dirty(kvm, gpa_to_gfn(gpa));
arch/x86/kvm/mmu/mmu.c
2895
for_each_gfn_valid_sp_with_gptes(kvm, sp, gpa_to_gfn(gpa))
arch/x86/kvm/mmu/mmu.c
4989
if (kvm_is_gfn_alias(vcpu->kvm, gpa_to_gfn(range->gpa)))
arch/x86/kvm/mmu/mmu.c
5002
kvm_mem_is_private(vcpu->kvm, gpa_to_gfn(range->gpa)))
arch/x86/kvm/mmu/mmu.c
6418
kvm_mem_is_private(vcpu->kvm, gpa_to_gfn(cr2_or_gpa)))
arch/x86/kvm/mmu/mmu_internal.h
372
fault.gfn = gpa_to_gfn(fault.addr) & ~kvm_gfn_direct_bits(vcpu->kvm);
arch/x86/kvm/mmu/paging_tmpl.h
395
slot = kvm_vcpu_gfn_to_memslot(vcpu, gpa_to_gfn(real_gpa));
arch/x86/kvm/mmu/paging_tmpl.h
399
host_addr = gfn_to_hva_memslot_prot(slot, gpa_to_gfn(real_gpa),
arch/x86/kvm/svm/nested.c
1022
ret = kvm_vcpu_map(vcpu, gpa_to_gfn(vmcb12_gpa), &map);
arch/x86/kvm/svm/nested.c
1137
rc = kvm_vcpu_map(vcpu, gpa_to_gfn(svm->nested.vmcb12_gpa), &map);
arch/x86/kvm/svm/nested.c
69
ret = kvm_vcpu_read_guest_page(vcpu, gpa_to_gfn(cr3), &pdpte,
arch/x86/kvm/svm/sev.c
3993
gfn = gpa_to_gfn(svm->sev_es.snp_vmsa_gpa);
arch/x86/kvm/svm/svm.c
2162
ret = kvm_vcpu_map(vcpu, gpa_to_gfn(svm->vmcb->save.rax), &map);
arch/x86/kvm/svm/svm.c
4823
if (kvm_vcpu_map(vcpu, gpa_to_gfn(svm->nested.hsave_msr), &map_save))
arch/x86/kvm/svm/svm.c
4857
if (kvm_vcpu_map(vcpu, gpa_to_gfn(smram64->svm_guest_vmcb_gpa), &map))
arch/x86/kvm/svm/svm.c
4861
if (kvm_vcpu_map(vcpu, gpa_to_gfn(svm->nested.hsave_msr), &map_save))
arch/x86/kvm/vmx/nested.c
2141
if (kvm_vcpu_map(vcpu, gpa_to_gfn(evmcs_gpa),
arch/x86/kvm/vmx/nested.c
3440
if (!kvm_vcpu_map(vcpu, gpa_to_gfn(vmcs12->apic_access_addr), map)) {
arch/x86/kvm/vmx/nested.c
3456
if (!kvm_vcpu_map(vcpu, gpa_to_gfn(vmcs12->virtual_apic_page_addr), map)) {
arch/x86/kvm/vmx/nested.c
3482
if (!kvm_vcpu_map(vcpu, gpa_to_gfn(vmcs12->posted_intr_desc_addr), map)) {
arch/x86/kvm/vmx/nested.c
3562
if (kvm_write_guest_page(vcpu->kvm, gpa_to_gfn(dst), &gpa,
arch/x86/kvm/vmx/nested.c
699
if (kvm_vcpu_map_readonly(vcpu, gpa_to_gfn(vmcs12->msr_bitmap), &map))
arch/x86/kvm/vmx/tdx.c
1955
slot = kvm_vcpu_gfn_to_memslot(vcpu, gpa_to_gfn(gpa));
arch/x86/kvm/vmx/tdx.c
3198
gmem_ret = kvm_gmem_populate(kvm, gpa_to_gfn(region.gpa),
arch/x86/kvm/vmx/tdx.c
58
#define TDX_SHARED_BIT_PWL_5 gpa_to_gfn(BIT_ULL(51))
arch/x86/kvm/vmx/tdx.c
59
#define TDX_SHARED_BIT_PWL_4 gpa_to_gfn(BIT_ULL(47))
arch/x86/kvm/x86.c
10449
gpa_to_gfn(gpa) + npages <= gpa_to_gfn(gpa)) {
arch/x86/kvm/x86.c
1078
ret = kvm_vcpu_read_guest_page(vcpu, gpa_to_gfn(real_gpa), pdpte,
arch/x86/kvm/x86.c
10994
unsigned long gfn = gpa_to_gfn(APIC_DEFAULT_PHYS_BASE);
arch/x86/kvm/x86.c
14032
kvm_zap_gfn_range(kvm, gpa_to_gfn(0), gpa_to_gfn(~0ULL));
arch/x86/kvm/x86.c
3841
mark_page_dirty_in_slot(vcpu->kvm, ghc->memslot, gpa_to_gfn(ghc->gpa));
arch/x86/kvm/x86.c
5282
mark_page_dirty_in_slot(vcpu->kvm, ghc->memslot, gpa_to_gfn(ghc->gpa));
arch/x86/kvm/x86.c
8345
hva = kvm_vcpu_gfn_to_hva(vcpu, gpa_to_gfn(gpa));
arch/x86/kvm/x86.c
8378
kvm_vcpu_mark_page_dirty(vcpu, gpa_to_gfn(gpa));
arch/x86/kvm/xen.c
865
data->u.shared_info.gfn = gpa_to_gfn(kvm->arch.xen.shinfo_cache.gpa);
include/linux/kvm_host.h
1921
unsigned long hva = gfn_to_hva(kvm, gpa_to_gfn(gpa));
include/linux/kvm_host.h
1933
mark_page_dirty_in_slot(gpc->kvm, gpc->memslot, gpa_to_gfn(gpc->gpa));
virt/kvm/pfncache.c
166
.gfn = gpa_to_gfn(gpc->gpa),
virt/kvm/pfncache.c
299
gfn_t gfn = gpa_to_gfn(gpa);