gpa_to_gfn
gfn = gpa_to_gfn(guest_ipa);
mark_page_dirty_in_slot(vcpu->kvm, ghc->memslot, gpa_to_gfn(ghc->gpa));
mark_page_dirty_in_slot(vcpu->kvm, ghc->memslot, gpa_to_gfn(ghc->gpa));
gfn = gpa_to_gfn(kvm_eq.qaddr);
mark_page_dirty(vcpu->kvm, gpa_to_gfn(q->guest_qaddr));
do_discard_gfn_range(vcpu, gpa_to_gfn(start), gpa_to_gfn(end));
do_discard_gfn_range(vcpu, gpa_to_gfn(start), gpa_to_gfn(prefix));
do_discard_gfn_range(vcpu, gpa_to_gfn(prefix) + 2, gpa_to_gfn(end));
rc = kvm_s390_get_guest_page(kvm, f, gpa_to_gfn(gaddr), false);
.gfn = gpa_to_gfn(gpa),
return kvm_s390_get_guest_page(kvm, entries + LEVEL_MEM, gpa_to_gfn(saddr), false);
if (!asce_contains_gfn(asce, gpa_to_gfn(saddr)))
rc = gmap_insert_rmap(sg, f->gfn, gpa_to_gfn(raddr), TABLE_TYPE_PAGE_TABLE);
pgste = __dat_ptep_xchg(ptep, pgste, newpte, gpa_to_gfn(raddr), sg->asce, uses_skeys(sg));
gfn = f->gfn & gpa_to_gfn(is_pmd(*table) ? _SEGMENT_MASK : _REGION3_MASK);
rc = gmap_insert_rmap(sg, gfn, gpa_to_gfn(raddr), host->h.tt);
gfn = gpa_to_gfn(raddr);
rc = dat_entry_walk(NULL, gpa_to_gfn(saddr), sg->asce, DAT_WALK_ANY, TABLE_TYPE_PAGE_TABLE,
rc = gmap_protect_rmap(mc, sg, entries[i].gfn, gpa_to_gfn(saddr),
rc = dat_entry_walk(mc, gpa_to_gfn(saddr), sg->asce, flags, l, &table, &ptep);
r = dat_get_storage_key(kvm->arch.gmap->asce, gpa_to_gfn(gpa), &storage_key);
r = dat_get_storage_key(vcpu->arch.gmap->asce, gpa_to_gfn(gpa), &storage_key);
const gfn_t gfn = gpa_to_gfn(gpa);
.gfn = gpa_to_gfn(gpa),
prefix_gfn = gpa_to_gfn(kvm_s390_get_prefix(vcpu));
rc = dat_entry_walk(NULL, gpa_to_gfn(*gaddr), gmap->asce, DAT_WALK_CONTINUE,
gfn = gpa_to_gfn(*gaddr);
rc = gmap_ucas_map_one(mc, gmap, gpa_to_gfn(translated_address), gfn, true);
rc = kvm_s390_faultin_gfn_simple(vcpu, NULL, gpa_to_gfn(srcaddr), false);
rc = kvm_s390_faultin_gfn_simple(vcpu, NULL, gpa_to_gfn(dstaddr), true);
ret = gmap_set_limit(kvm->arch.gmap, gpa_to_gfn(new_limit));
kvm->arch.gmap = gmap_new(kvm, gpa_to_gfn(kvm->arch.mem_limit));
gfn = gpa_to_gfn(gaddr);
f.gfn = gpa_to_gfn(gaddr);
r = gmap_ucas_map(vcpu->arch.gmap, gpa_to_gfn(ucas.user_addr),
gpa_to_gfn(ucas.vcpu_addr),
gmap_ucas_unmap(vcpu->arch.gmap, gpa_to_gfn(ucas.vcpu_addr),
r = kvm_s390_faultin_gfn_simple(vcpu, NULL, gpa_to_gfn(gaddr), false);
gfn_t gfn = gpa_to_gfn(addr);
hva_t hva = gfn_to_hva(kvm, gpa_to_gfn(gpa));
hva = gfn_to_hva(kvm, gpa_to_gfn((gpa_t)fib->fmt0.aibv));
hva = gfn_to_hva(kvm, gpa_to_gfn((gpa_t)fib->fmt0.aisb));
gpa_to_gfn(start), key,
if (dat_entry_walk(NULL, gpa_to_gfn(cbrl[i]), vcpu->arch.gmap->asce,
gfn_to_hva_prot(vcpu->kvm, gpa_to_gfn(gpa), &writable);
rc = dat_get_storage_key(vcpu->arch.gmap->asce, gpa_to_gfn(gaddr), &key);
rc = dat_reset_reference_bit(vcpu->arch.gmap->asce, gpa_to_gfn(gaddr));
gpa_to_gfn(start), key, &oldkey,
.gfn = gpa_to_gfn(gaddr),
page = gfn_to_page(kvm, gpa_to_gfn(gaddr));
gmap_pv_destroy_range(kvm->arch.gmap, 0, gpa_to_gfn(SZ_2G), false);
ret = kvm_s390_faultin_gfn_simple(NULL, kvm, gpa_to_gfn(addr), true);
page = gfn_to_page(kvm, gpa_to_gfn(gpa));
mark_page_dirty(kvm, gpa_to_gfn(gpa));
for_each_gfn_valid_sp_with_gptes(kvm, sp, gpa_to_gfn(gpa))
if (kvm_is_gfn_alias(vcpu->kvm, gpa_to_gfn(range->gpa)))
kvm_mem_is_private(vcpu->kvm, gpa_to_gfn(range->gpa)))
kvm_mem_is_private(vcpu->kvm, gpa_to_gfn(cr2_or_gpa)))
fault.gfn = gpa_to_gfn(fault.addr) & ~kvm_gfn_direct_bits(vcpu->kvm);
slot = kvm_vcpu_gfn_to_memslot(vcpu, gpa_to_gfn(real_gpa));
host_addr = gfn_to_hva_memslot_prot(slot, gpa_to_gfn(real_gpa),
ret = kvm_vcpu_map(vcpu, gpa_to_gfn(vmcb12_gpa), &map);
rc = kvm_vcpu_map(vcpu, gpa_to_gfn(svm->nested.vmcb12_gpa), &map);
ret = kvm_vcpu_read_guest_page(vcpu, gpa_to_gfn(cr3), &pdpte,
gfn = gpa_to_gfn(svm->sev_es.snp_vmsa_gpa);
ret = kvm_vcpu_map(vcpu, gpa_to_gfn(svm->vmcb->save.rax), &map);
if (kvm_vcpu_map(vcpu, gpa_to_gfn(svm->nested.hsave_msr), &map_save))
if (kvm_vcpu_map(vcpu, gpa_to_gfn(smram64->svm_guest_vmcb_gpa), &map))
if (kvm_vcpu_map(vcpu, gpa_to_gfn(svm->nested.hsave_msr), &map_save))
if (kvm_vcpu_map(vcpu, gpa_to_gfn(evmcs_gpa),
if (!kvm_vcpu_map(vcpu, gpa_to_gfn(vmcs12->apic_access_addr), map)) {
if (!kvm_vcpu_map(vcpu, gpa_to_gfn(vmcs12->virtual_apic_page_addr), map)) {
if (!kvm_vcpu_map(vcpu, gpa_to_gfn(vmcs12->posted_intr_desc_addr), map)) {
if (kvm_write_guest_page(vcpu->kvm, gpa_to_gfn(dst), &gpa,
if (kvm_vcpu_map_readonly(vcpu, gpa_to_gfn(vmcs12->msr_bitmap), &map))
slot = kvm_vcpu_gfn_to_memslot(vcpu, gpa_to_gfn(gpa));
gmem_ret = kvm_gmem_populate(kvm, gpa_to_gfn(region.gpa),
#define TDX_SHARED_BIT_PWL_5 gpa_to_gfn(BIT_ULL(51))
#define TDX_SHARED_BIT_PWL_4 gpa_to_gfn(BIT_ULL(47))
gpa_to_gfn(gpa) + npages <= gpa_to_gfn(gpa)) {
ret = kvm_vcpu_read_guest_page(vcpu, gpa_to_gfn(real_gpa), pdpte,
unsigned long gfn = gpa_to_gfn(APIC_DEFAULT_PHYS_BASE);
kvm_zap_gfn_range(kvm, gpa_to_gfn(0), gpa_to_gfn(~0ULL));
mark_page_dirty_in_slot(vcpu->kvm, ghc->memslot, gpa_to_gfn(ghc->gpa));
mark_page_dirty_in_slot(vcpu->kvm, ghc->memslot, gpa_to_gfn(ghc->gpa));
hva = kvm_vcpu_gfn_to_hva(vcpu, gpa_to_gfn(gpa));
kvm_vcpu_mark_page_dirty(vcpu, gpa_to_gfn(gpa));
data->u.shared_info.gfn = gpa_to_gfn(kvm->arch.xen.shinfo_cache.gpa);
unsigned long hva = gfn_to_hva(kvm, gpa_to_gfn(gpa));
mark_page_dirty_in_slot(gpc->kvm, gpc->memslot, gpa_to_gfn(gpc->gpa));
.gfn = gpa_to_gfn(gpc->gpa),
gfn_t gfn = gpa_to_gfn(gpa);