to_kvm_tdx
struct kvm_tdx *kvm_tdx = to_kvm_tdx(vcpu->kvm);
if (unlikely(READ_ONCE(to_kvm_tdx(vcpu->kvm)->wait_for_sept_zap)))
struct kvm_tdx *kvm_tdx = to_kvm_tdx(kvm);
struct kvm_tdx *kvm_tdx = to_kvm_tdx(kvm);
struct kvm_tdx *kvm_tdx = to_kvm_tdx(kvm);
err = tdh_mem_sept_add(&to_kvm_tdx(kvm)->td, gpa, tdx_level, page, &entry,
struct kvm_tdx *kvm_tdx = to_kvm_tdx(kvm);
struct kvm_tdx *kvm_tdx = to_kvm_tdx(kvm);
struct kvm_tdx *kvm_tdx = to_kvm_tdx(kvm);
if (KVM_BUG_ON(!is_hkid_assigned(to_kvm_tdx(kvm)), kvm))
struct kvm_tdx *kvm_tdx = to_kvm_tdx(kvm);
struct kvm_tdx *kvm_tdx = to_kvm_tdx(vcpu->kvm);
struct kvm_tdx *kvm_tdx = to_kvm_tdx(kvm);
struct kvm_tdx *kvm_tdx = to_kvm_tdx(kvm);
struct kvm_tdx *kvm_tdx = to_kvm_tdx(vcpu->kvm);
struct kvm_tdx *__kvm_tdx = to_kvm_tdx(kvm); \
struct kvm_tdx *kvm_tdx = to_kvm_tdx(kvm);
struct kvm_tdx *kvm_tdx = to_kvm_tdx(kvm);
struct kvm_tdx *kvm_tdx = to_kvm_tdx(kvm);
struct kvm_tdx *kvm_tdx = to_kvm_tdx(vcpu->kvm);
struct kvm_tdx *kvm_tdx = to_kvm_tdx(kvm);
struct kvm_tdx *kvm_tdx = to_kvm_tdx(kvm);
struct kvm_tdx *kvm_tdx = to_kvm_tdx(kvm);
struct kvm_tdx *kvm_tdx = to_kvm_tdx(kvm);
struct kvm_tdx *kvm_tdx = to_kvm_tdx(vcpu->kvm);
!(to_kvm_tdx(vcpu->kvm)->attributes & TDX_TD_ATTR_DEBUG);
if (vcpu->cpu == cpu || !is_hkid_assigned(to_kvm_tdx(vcpu->kvm)))
struct kvm_tdx *kvm_tdx = to_kvm_tdx(vcpu->kvm);
to_kvm_tdx(vcpu->kvm)->state != TD_STATE_RUNNABLE))