INVALID_GPA
return INVALID_GPA;
vcpu_arch->steal.base = INVALID_GPA;
return (vcpu_arch->steal.base != INVALID_GPA);
if (gpa != INVALID_GPA)
if (hpfar == INVALID_GPA)
if (ipa != INVALID_GPA) {
if (KVM_BUG_ON(ipa == INVALID_GPA, vcpu->kvm))
if (base == INVALID_GPA)
if (vcpu->arch.steal.base != INVALID_GPA)
if (base == INVALID_GPA)
if (vcpu->arch.steal.base != INVALID_GPA)
vcpu->arch.flush_gpa = INVALID_GPA;
ghc->gpa = INVALID_GPA;
ghc->gpa = INVALID_GPA;
if (vcpu->arch.flush_gpa != INVALID_GPA) {
vcpu->arch.flush_gpa = INVALID_GPA;
kvpmu->snapshot_addr = INVALID_GPA;
if (kvpmu->snapshot_addr == INVALID_GPA) {
if (snap_flag_set && kvpmu->snapshot_addr == INVALID_GPA) {
kvpmu->snapshot_addr = INVALID_GPA;
vcpu->arch.sta.shmem = INVALID_GPA;
if (shmem == INVALID_GPA)
vcpu->arch.sta.shmem = INVALID_GPA;
vcpu->arch.sta.shmem = INVALID_GPA;
if (unlikely(hc->ingpa == INVALID_GPA))
if (gpa == INVALID_GPA)
arch.cr3 = (unsigned long)INVALID_GPA;
if (unlikely(real_gpa == INVALID_GPA))
if (real_gpa == INVALID_GPA)
gpa_t gpa = INVALID_GPA;
svm->nested.ctl.bus_lock_rip = INVALID_GPA;
svm->nested.last_vmcb12_gpa = INVALID_GPA;
svm->nested.vmcb12_gpa = INVALID_GPA;
svm->nested.vmcb12_gpa = INVALID_GPA;
svm->nested.last_vmcb12_gpa = INVALID_GPA;
if (vmx->nested.current_vmptr == INVALID_GPA &&
vmx->nested.current_vmptr = INVALID_GPA;
vmcs_write64(VMCS_LINK_POINTER, INVALID_GPA);
vmcs_write64(ENCLS_EXITING_BITMAP, INVALID_GPA);
vmcs_write64(VMCS_LINK_POINTER, INVALID_GPA);
hv_vcpu->nested.pa_page_gpa = INVALID_GPA;
if (vmcs12->vmcs_link_pointer == INVALID_GPA)
vmcs_write64(VIRTUAL_APIC_PAGE_ADDR, INVALID_GPA);
vmx->nested.vmxon_ptr = INVALID_GPA;
vmx->nested.current_vmptr = INVALID_GPA;
vmx->nested.current_vmptr == INVALID_GPA))
vmx->nested.shadow_vmcs12_cache.gpa = INVALID_GPA;
if (vmx->nested.current_vmptr == INVALID_GPA)
vmx->nested.current_vmptr = INVALID_GPA;
if (vmx->nested.current_vmptr == INVALID_GPA ||
get_vmcs12(vcpu)->vmcs_link_pointer == INVALID_GPA))
if (vmx->nested.current_vmptr == INVALID_GPA ||
get_vmcs12(vcpu)->vmcs_link_pointer == INVALID_GPA))
last_bitmap = INVALID_GPA;
.hdr.vmx.vmxon_pa = INVALID_GPA,
.hdr.vmx.vmcs12_pa = INVALID_GPA,
vmcs12->vmcs_link_pointer != INVALID_GPA)
vmcs12->vmcs_link_pointer != INVALID_GPA) {
vmcs12->vmcs_link_pointer != INVALID_GPA) {
if (kvm_state->hdr.vmx.vmxon_pa == INVALID_GPA) {
if (kvm_state->hdr.vmx.vmcs12_pa != INVALID_GPA)
if (kvm_state->hdr.vmx.vmxon_pa == INVALID_GPA)
(kvm_state->hdr.vmx.vmcs12_pa != INVALID_GPA))
if (kvm_state->hdr.vmx.vmcs12_pa != INVALID_GPA) {
vmcs12->vmcs_link_pointer != INVALID_GPA) {
vmcs12->vmcs_link_pointer == INVALID_GPA)
vmcs12->vmcs_link_pointer == INVALID_GPA)
if (*gpa == INVALID_GPA) {
vmcs_write64(VMCS_LINK_POINTER, INVALID_GPA); /* 22.3.1.5 */
vmx->nested.vmxon_ptr = INVALID_GPA;
vmx->nested.current_vmptr = INVALID_GPA;
kvm_prepare_event_vectoring_exit(vcpu, INVALID_GPA);
if (real_gpa == INVALID_GPA)
tr->valid = gpa != INVALID_GPA;
mmu->gva_to_gpa(vcpu, mmu, gva, access, &fault) != INVALID_GPA) {
if (gpa == INVALID_GPA)
if (unlikely(gpa == INVALID_GPA))
if (gpa == INVALID_GPA)
if (*gpa == INVALID_GPA)
if (gpa == INVALID_GPA ||
return gpa == INVALID_GPA;
if (WARN_ON_ONCE(kvm->mmu_invalidate_range_start == INVALID_GPA ||
kvm->mmu_invalidate_range_end == INVALID_GPA))
kvm->mmu_invalidate_range_start = INVALID_GPA;
kvm->mmu_invalidate_range_end = INVALID_GPA;
if (likely(kvm->mmu_invalidate_range_start == INVALID_GPA)) {
WARN_ON_ONCE(kvm->mmu_invalidate_range_start == INVALID_GPA);
gpc->gpa = INVALID_GPA;
gpc->gpa = INVALID_GPA;
return __kvm_gpc_activate(gpc, INVALID_GPA, uhva, len);