kvm_vcpu_get_esr
u64 esr = kvm_vcpu_get_esr(vcpu);
u64 esr = kvm_vcpu_get_esr(vcpu);
return kvm_vcpu_get_esr(vcpu) & ESR_ELx_xVC_IMM_MASK;
return !!(kvm_vcpu_get_esr(vcpu) & ESR_ELx_ISV);
return kvm_vcpu_get_esr(vcpu) & (ESR_ELx_CM | ESR_ELx_WNR | ESR_ELx_FSC);
return !!(kvm_vcpu_get_esr(vcpu) & ESR_ELx_SSE);
return !!(kvm_vcpu_get_esr(vcpu) & ESR_ELx_SF);
return (kvm_vcpu_get_esr(vcpu) & ESR_ELx_SRT_MASK) >> ESR_ELx_SRT_SHIFT;
return !!(kvm_vcpu_get_esr(vcpu) & ESR_ELx_S1PTW);
return kvm_vcpu_get_esr(vcpu) & ESR_ELx_WNR;
return !!(kvm_vcpu_get_esr(vcpu) & ESR_ELx_CM);
return 1 << ((kvm_vcpu_get_esr(vcpu) & ESR_ELx_SAS) >> ESR_ELx_SAS_SHIFT);
return !!(kvm_vcpu_get_esr(vcpu) & ESR_ELx_IL);
return ESR_ELx_EC(kvm_vcpu_get_esr(vcpu));
return kvm_vcpu_get_esr(vcpu) & ESR_ELx_FSC;
return esr_fsc_is_permission_fault(kvm_vcpu_get_esr(vcpu));
return esr_fsc_is_translation_fault(kvm_vcpu_get_esr(vcpu));
unsigned long esr = kvm_vcpu_get_esr(vcpu);
u64 esr = kvm_vcpu_get_esr(vcpu);
esr = kvm_vcpu_get_esr(vcpu);
kvm_inject_sync(vcpu, kvm_vcpu_get_esr(vcpu));
kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu));
kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu));
esr = kvm_vcpu_get_esr(vcpu);
return kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu));
u64 esr = kvm_vcpu_get_esr(vcpu);
return kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu));
u64 esr = kvm_vcpu_get_esr(vcpu);
u64 esr = kvm_vcpu_get_esr(vcpu);
return kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu));
kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu));
if (esr_iss_is_eretax(kvm_vcpu_get_esr(vcpu)) &&
kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu));
kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu));
u64 esr = kvm_vcpu_get_esr(vcpu);
u64 esr = kvm_vcpu_get_esr(vcpu);
kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu));
kvm_handle_guest_serror(vcpu, kvm_vcpu_get_esr(vcpu));
u32 sysreg = esr_sys64_to_sysreg(kvm_vcpu_get_esr(vcpu));
sysreg = esr_sys64_to_sysreg(kvm_vcpu_get_esr(vcpu));
u32 sysreg = esr_sys64_to_sysreg(kvm_vcpu_get_esr(vcpu));
inject_sync64(vcpu, kvm_vcpu_get_esr(vcpu));
unsigned long esr = kvm_vcpu_get_esr(vcpu);
esr = kvm_vcpu_get_esr(vcpu);
u64 esr = kvm_vcpu_get_esr(vcpu);
esr = kvm_vcpu_get_esr(vcpu);
u64 esr = kvm_vcpu_get_esr(vcpu);
instr = esr_sys64_to_sysreg(kvm_vcpu_get_esr(vcpu));
u64 esr = kvm_vcpu_get_esr(vcpu);
u32 sysreg = esr_sys64_to_sysreg(kvm_vcpu_get_esr(vcpu));
u8 ec = ESR_ELx_EC(kvm_vcpu_get_esr(vcpu));
esr = kvm_vcpu_get_esr(vcpu);
if (esr_fsc_is_excl_atomic_fault(kvm_vcpu_get_esr(s2fd->vcpu))) {
u64 esr = kvm_vcpu_get_esr(vcpu);
esr = kvm_vcpu_get_esr(vcpu);
trace_kvm_guest_fault(*vcpu_pc(vcpu), kvm_vcpu_get_esr(vcpu),
(unsigned long)kvm_vcpu_get_esr(vcpu));
u64 esr = kvm_vcpu_get_esr(vcpu);
u64 esr = kvm_vcpu_get_esr(vcpu);
esr = kvm_vcpu_get_esr(vcpu) & ~ESR_ELx_FSC;
result->esr |= (kvm_vcpu_get_esr(vcpu) & ~ESR_ELx_FSC);
if (esr_iss_is_eretab(kvm_vcpu_get_esr(vcpu)))
u64 esr = kvm_vcpu_get_esr(vcpu);
u64 esr = kvm_vcpu_get_esr(vcpu);
u64 esr = kvm_vcpu_get_esr(vcpu);
params = esr_cp1x_32_to_params(kvm_vcpu_get_esr(vcpu));
params = esr_cp1x_32_to_params(kvm_vcpu_get_esr(vcpu));
unsigned long esr = kvm_vcpu_get_esr(vcpu);