Symbol: to_hv_vcpu
arch/x86/kvm/hyperv.c
1382
if (unlikely(!host && !hv_check_msr_access(to_hv_vcpu(vcpu), msr)))
arch/x86/kvm/hyperv.c
1521
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.c
1638
if (unlikely(!host && !hv_check_msr_access(to_hv_vcpu(vcpu), msr)))
arch/x86/kvm/hyperv.c
1691
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.c
1937
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.c
1968
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.c
2009
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.c
213
if (!vcpu || !to_hv_vcpu(vcpu))
arch/x86/kvm/hyperv.c
2167
hv_v = to_hv_vcpu(v);
arch/x86/kvm/hyperv.c
2225
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.c
223
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.c
2308
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.c
2356
if (!to_hv_vcpu(vcpu)) {
arch/x86/kvm/hyperv.c
2366
hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.c
2392
kvm_read_guest(vcpu->kvm, to_hv_vcpu(vcpu)->nested.pa_page_gpa,
arch/x86/kvm/hyperv.c
248
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.c
2530
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.c
330
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.c
349
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.c
371
to_hv_vcpu(vcpu)->vp_index, msr, data);
arch/x86/kvm/hyperv.c
597
to_hv_vcpu(vcpu)->stimer_pending_bitmap);
arch/x86/kvm/hyperv.c
612
to_hv_vcpu(vcpu)->stimer_pending_bitmap);
arch/x86/kvm/hyperv.c
694
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.c
869
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.c
902
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.c
917
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.c
930
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.c
965
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.h
113
return to_hv_vcpu(vcpu) && test_bit(vector, to_hv_synic(vcpu)->vec_bitmap);
arch/x86/kvm/hyperv.h
118
return to_hv_vcpu(vcpu) &&
arch/x86/kvm/hyperv.h
130
return &to_hv_vcpu(vcpu)->stimer[timer_index];
arch/x86/kvm/hyperv.h
144
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.h
159
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.h
200
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.h
211
if (!to_hv_vcpu(vcpu) || !kvm_check_request(KVM_REQ_HV_TLB_FLUSH, vcpu))
arch/x86/kvm/hyperv.h
221
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.h
229
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.h
246
if (!to_hv_vcpu(vcpu))
arch/x86/kvm/hyperv.h
264
if (to_hv_vcpu(vcpu) && tdp_enabled)
arch/x86/kvm/hyperv.h
71
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/hyperv.h
90
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/svm/hyperv.h
19
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/svm/hyperv.h
33
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/vmx/hyperv.c
19
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/vmx/hyperv.c
214
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/vmx/hyperv.c
97
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/vmx/nested.c
1694
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(&vmx->vcpu);
arch/x86/kvm/vmx/nested.c
236
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);
arch/x86/kvm/x86.c
11046
if (to_hv_vcpu(vcpu)) {
arch/x86/kvm/x86.c
11214
struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(vcpu);