Symbol: to_tdx
arch/x86/kvm/vmx/tdx.c
1001
u64 vp_enter_ret = to_tdx(vcpu)->vp_enter_ret;
arch/x86/kvm/vmx/tdx.c
1067
struct vcpu_tdx *tdx = to_tdx(vcpu);
arch/x86/kvm/vmx/tdx.c
1124
td_management_write8(to_tdx(vcpu), TD_VCPU_PEND_NMI, 1);
arch/x86/kvm/vmx/tdx.c
1170
kvm_rax_write(vcpu, to_tdx(vcpu)->vp_enter_args.r10);
arch/x86/kvm/vmx/tdx.c
1171
kvm_rbx_write(vcpu, to_tdx(vcpu)->vp_enter_args.r11);
arch/x86/kvm/vmx/tdx.c
1172
kvm_rcx_write(vcpu, to_tdx(vcpu)->vp_enter_args.r12);
arch/x86/kvm/vmx/tdx.c
1173
kvm_rdx_write(vcpu, to_tdx(vcpu)->vp_enter_args.r13);
arch/x86/kvm/vmx/tdx.c
1174
kvm_rsi_write(vcpu, to_tdx(vcpu)->vp_enter_args.r14);
arch/x86/kvm/vmx/tdx.c
1189
struct vcpu_tdx *tdx = to_tdx(vcpu);
arch/x86/kvm/vmx/tdx.c
1245
struct vcpu_tdx *tdx = to_tdx(vcpu);
arch/x86/kvm/vmx/tdx.c
1290
struct vcpu_tdx *tdx = to_tdx(vcpu);
arch/x86/kvm/vmx/tdx.c
1317
struct vcpu_tdx *tdx = to_tdx(vcpu);
arch/x86/kvm/vmx/tdx.c
1357
struct vcpu_tdx *tdx = to_tdx(vcpu);
arch/x86/kvm/vmx/tdx.c
1435
struct vcpu_tdx *tdx = to_tdx(vcpu);
arch/x86/kvm/vmx/tdx.c
1494
struct vcpu_tdx *tdx = to_tdx(vcpu);
arch/x86/kvm/vmx/tdx.c
1513
struct vcpu_tdx *tdx = to_tdx(vcpu);
arch/x86/kvm/vmx/tdx.c
1549
struct vcpu_tdx *tdx = to_tdx(vcpu);
arch/x86/kvm/vmx/tdx.c
1573
struct vcpu_tdx *tdx = to_tdx(vcpu);
arch/x86/kvm/vmx/tdx.c
1621
td_vmcs_write64(to_tdx(vcpu), SHARED_EPT_POINTER, root_hpa);
arch/x86/kvm/vmx/tdx.c
1842
struct vcpu_tdx *tdx = to_tdx(vcpu);
arch/x86/kvm/vmx/tdx.c
1852
u64 eeq_type = to_tdx(vcpu)->ext_exit_qualification & TDX_EXT_EXIT_QUAL_TYPE_MASK;
arch/x86/kvm/vmx/tdx.c
1864
gpa_t gpa = to_tdx(vcpu)->exit_gpa;
arch/x86/kvm/vmx/tdx.c
1980
struct vcpu_tdx *tdx = to_tdx(vcpu);
arch/x86/kvm/vmx/tdx.c
2088
struct vcpu_tdx *tdx = to_tdx(vcpu);
arch/x86/kvm/vmx/tdx.c
246
return to_tdx(vcpu)->vp_enter_args.r10;
arch/x86/kvm/vmx/tdx.c
251
return to_tdx(vcpu)->vp_enter_args.r11;
arch/x86/kvm/vmx/tdx.c
257
to_tdx(vcpu)->vp_enter_args.r10 = val;
arch/x86/kvm/vmx/tdx.c
263
to_tdx(vcpu)->vp_enter_args.r11 = val;
arch/x86/kvm/vmx/tdx.c
285
list_del(&to_tdx(vcpu)->cpu_list);
arch/x86/kvm/vmx/tdx.c
2885
struct vcpu_tdx *tdx = to_tdx(vcpu);
arch/x86/kvm/vmx/tdx.c
3070
struct vcpu_tdx *tdx = to_tdx(vcpu);
arch/x86/kvm/vmx/tdx.c
3159
struct vcpu_tdx *tdx = to_tdx(vcpu);
arch/x86/kvm/vmx/tdx.c
395
if (to_tdx(vcpu)->state != VCPU_TD_STATE_UNINITIALIZED) {
arch/x86/kvm/vmx/tdx.c
402
err = tdh_vp_flush(&to_tdx(vcpu)->vp);
arch/x86/kvm/vmx/tdx.c
679
struct vcpu_tdx *tdx = to_tdx(vcpu);
arch/x86/kvm/vmx/tdx.c
725
struct vcpu_tdx *tdx = to_tdx(vcpu);
arch/x86/kvm/vmx/tdx.c
753
!to_tdx(vcpu)->vp_enter_args.r12;
arch/x86/kvm/vmx/tdx.c
771
to_tdx(vcpu)->vp_enter_args.r12)
arch/x86/kvm/vmx/tdx.c
775
td_state_non_arch_read64(to_tdx(vcpu), TD_VCPU_STATE_DETAILS_NON_ARCH);
arch/x86/kvm/vmx/tdx.c
874
struct vcpu_tdx *tdx = to_tdx(vcpu);
arch/x86/kvm/vmx/tdx.c
910
if (unlikely(to_tdx(vcpu)->state != VCPU_TD_STATE_INITIALIZED ||
arch/x86/kvm/vmx/tdx.c
937
struct vcpu_tdx *tdx = to_tdx(vcpu);
arch/x86/kvm/vmx/tdx.c
974
struct vcpu_tdx *tdx = to_tdx(vcpu);