Symbol: kvm_vcpu_get_esr
arch/arm64/include/asm/kvm_emulate.h
322
u64 esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/include/asm/kvm_emulate.h
335
u64 esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/include/asm/kvm_emulate.h
365
return kvm_vcpu_get_esr(vcpu) & ESR_ELx_xVC_IMM_MASK;
arch/arm64/include/asm/kvm_emulate.h
370
return !!(kvm_vcpu_get_esr(vcpu) & ESR_ELx_ISV);
arch/arm64/include/asm/kvm_emulate.h
375
return kvm_vcpu_get_esr(vcpu) & (ESR_ELx_CM | ESR_ELx_WNR | ESR_ELx_FSC);
arch/arm64/include/asm/kvm_emulate.h
380
return !!(kvm_vcpu_get_esr(vcpu) & ESR_ELx_SSE);
arch/arm64/include/asm/kvm_emulate.h
385
return !!(kvm_vcpu_get_esr(vcpu) & ESR_ELx_SF);
arch/arm64/include/asm/kvm_emulate.h
390
return (kvm_vcpu_get_esr(vcpu) & ESR_ELx_SRT_MASK) >> ESR_ELx_SRT_SHIFT;
arch/arm64/include/asm/kvm_emulate.h
395
return !!(kvm_vcpu_get_esr(vcpu) & ESR_ELx_S1PTW);
arch/arm64/include/asm/kvm_emulate.h
401
return kvm_vcpu_get_esr(vcpu) & ESR_ELx_WNR;
arch/arm64/include/asm/kvm_emulate.h
406
return !!(kvm_vcpu_get_esr(vcpu) & ESR_ELx_CM);
arch/arm64/include/asm/kvm_emulate.h
411
return 1 << ((kvm_vcpu_get_esr(vcpu) & ESR_ELx_SAS) >> ESR_ELx_SAS_SHIFT);
arch/arm64/include/asm/kvm_emulate.h
417
return !!(kvm_vcpu_get_esr(vcpu) & ESR_ELx_IL);
arch/arm64/include/asm/kvm_emulate.h
422
return ESR_ELx_EC(kvm_vcpu_get_esr(vcpu));
arch/arm64/include/asm/kvm_emulate.h
437
return kvm_vcpu_get_esr(vcpu) & ESR_ELx_FSC;
arch/arm64/include/asm/kvm_emulate.h
443
return esr_fsc_is_permission_fault(kvm_vcpu_get_esr(vcpu));
arch/arm64/include/asm/kvm_emulate.h
449
return esr_fsc_is_translation_fault(kvm_vcpu_get_esr(vcpu));
arch/arm64/include/asm/kvm_emulate.h
455
unsigned long esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/include/asm/kvm_emulate.h
476
u64 esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/kvm/emulate-nested.c
2551
esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/kvm/emulate-nested.c
2665
kvm_inject_sync(vcpu, kvm_vcpu_get_esr(vcpu));
arch/arm64/kvm/emulate-nested.c
2691
kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu));
arch/arm64/kvm/emulate-nested.c
2699
kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu));
arch/arm64/kvm/emulate-nested.c
2766
esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/kvm/handle_exit.c
108
return kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu));
arch/arm64/kvm/handle_exit.c
132
u64 esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/kvm/handle_exit.c
136
return kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu));
arch/arm64/kvm/handle_exit.c
194
u64 esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/kvm/handle_exit.c
218
u64 esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/kvm/handle_exit.c
234
return kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu));
arch/arm64/kvm/handle_exit.c
261
kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu));
arch/arm64/kvm/handle_exit.c
273
if (esr_iss_is_eretax(kvm_vcpu_get_esr(vcpu)) &&
arch/arm64/kvm/handle_exit.c
291
kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu));
arch/arm64/kvm/handle_exit.c
303
kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu));
arch/arm64/kvm/handle_exit.c
321
u64 esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/kvm/handle_exit.c
409
u64 esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/kvm/handle_exit.c
49
kvm_inject_nested_sync(vcpu, kvm_vcpu_get_esr(vcpu));
arch/arm64/kvm/handle_exit.c
507
kvm_handle_guest_serror(vcpu, kvm_vcpu_get_esr(vcpu));
arch/arm64/kvm/hyp/include/hyp/switch.h
648
u32 sysreg = esr_sys64_to_sysreg(kvm_vcpu_get_esr(vcpu));
arch/arm64/kvm/hyp/include/hyp/switch.h
736
sysreg = esr_sys64_to_sysreg(kvm_vcpu_get_esr(vcpu));
arch/arm64/kvm/hyp/include/hyp/switch.h
778
u32 sysreg = esr_sys64_to_sysreg(kvm_vcpu_get_esr(vcpu));
arch/arm64/kvm/hyp/nvhe/sys_regs.c
363
inject_sync64(vcpu, kvm_vcpu_get_esr(vcpu));
arch/arm64/kvm/hyp/nvhe/sys_regs.c
577
unsigned long esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/kvm/hyp/vgic-v3-sr.c
1153
esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/kvm/hyp/vgic-v3-sr.c
548
u64 esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/kvm/hyp/vhe/switch.c
274
esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/kvm/hyp/vhe/switch.c
339
u64 esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/kvm/hyp/vhe/switch.c
415
instr = esr_sys64_to_sysreg(kvm_vcpu_get_esr(vcpu));
arch/arm64/kvm/hyp/vhe/switch.c
441
u64 esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/kvm/hyp/vhe/switch.c
463
u32 sysreg = esr_sys64_to_sysreg(kvm_vcpu_get_esr(vcpu));
arch/arm64/kvm/hypercalls.c
248
u8 ec = ESR_ELx_EC(kvm_vcpu_get_esr(vcpu));
arch/arm64/kvm/mmio.c
164
esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/kvm/mmu.c
1935
if (esr_fsc_is_excl_atomic_fault(kvm_vcpu_get_esr(s2fd->vcpu))) {
arch/arm64/kvm/mmu.c
2144
u64 esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/kvm/mmu.c
2217
esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/kvm/mmu.c
2244
trace_kvm_guest_fault(*vcpu_pc(vcpu), kvm_vcpu_get_esr(vcpu),
arch/arm64/kvm/mmu.c
2255
(unsigned long)kvm_vcpu_get_esr(vcpu));
arch/arm64/kvm/nested.c
1350
u64 esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/kvm/nested.c
1381
u64 esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/kvm/nested.c
145
esr = kvm_vcpu_get_esr(vcpu) & ~ESR_ELx_FSC;
arch/arm64/kvm/nested.c
423
result->esr |= (kvm_vcpu_get_esr(vcpu) & ~ESR_ELx_FSC);
arch/arm64/kvm/pauth.c
139
if (esr_iss_is_eretab(kvm_vcpu_get_esr(vcpu)))
arch/arm64/kvm/pauth.c
159
u64 esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/kvm/sys_regs.c
4855
u64 esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/kvm/sys_regs.c
4953
u64 esr = kvm_vcpu_get_esr(vcpu);
arch/arm64/kvm/sys_regs.c
5047
params = esr_cp1x_32_to_params(kvm_vcpu_get_esr(vcpu));
arch/arm64/kvm/sys_regs.c
5074
params = esr_cp1x_32_to_params(kvm_vcpu_get_esr(vcpu));
arch/arm64/kvm/sys_regs.c
5306
unsigned long esr = kvm_vcpu_get_esr(vcpu);