arch/x86/kvm/vmx/hyperv.c
202
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/hyperv.c
215
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/hyperv.c
42
(!vcpu || to_vmx(vcpu)->nested.enlightened_vmcs_enabled))
arch/x86/kvm/vmx/hyperv.h
52
to_vmx(vcpu)->nested.enlightened_vmcs_enabled;
arch/x86/kvm/vmx/nested.c
1075
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
1221
(nested_cpu_has_vpid(vmcs12) && to_vmx(vcpu)->nested.vpid02);
arch/x86/kvm/vmx/nested.c
1228
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
1480
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
188
if (nested_vmx_is_evmptr12_set(to_vmx(vcpu)))
arch/x86/kvm/vmx/nested.c
189
to_vmx(vcpu)->nested.need_vmcs12_to_shadow_sync = true;
arch/x86/kvm/vmx/nested.c
196
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
2123
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
2211
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
2235
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
2252
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
237
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
254
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
2709
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
2868
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
2917
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
2980
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
299
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
2997
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
3091
void *vapic = to_vmx(vcpu)->nested.virtual_apic_map.hva;
arch/x86/kvm/vmx/nested.c
3109
if ((exec_controls_get(to_vmx(vcpu)) & CPU_BASED_TPR_SHADOW) &&
arch/x86/kvm/vmx/nested.c
324
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
3248
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
3352
if (to_vmx(vcpu)->nested.nested_run_pending &&
arch/x86/kvm/vmx/nested.c
338
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
3392
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
3422
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
3537
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
3580
if (!to_vmx(vcpu)->nested.vmxon) {
arch/x86/kvm/vmx/nested.c
3609
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
3773
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
4008
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
4154
to_vmx(vcpu)->nested.preemption_timer_expired;
arch/x86/kvm/vmx/nested.c
4159
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
417
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
4285
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
4496
hrtimer_get_remaining(&to_vmx(vcpu)->nested.preemption_timer);
arch/x86/kvm/vmx/nested.c
4557
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
4602
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
4630
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
4689
(vm_entry_controls_get(to_vmx(vcpu)) & VM_ENTRY_IA32E_MODE);
arch/x86/kvm/vmx/nested.c
470
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
4938
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
5052
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
5087
hrtimer_cancel(&to_vmx(vcpu)->nested.preemption_timer);
arch/x86/kvm/vmx/nested.c
5373
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
5396
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
5449
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
5537
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
5579
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
5637
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
5743
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
5860
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
5932
gpa_t current_vmptr = to_vmx(vcpu)->nested.current_vmptr;
arch/x86/kvm/vmx/nested.c
5940
if (unlikely(nested_vmx_is_evmptr12_valid(to_vmx(vcpu))))
arch/x86/kvm/vmx/nested.c
5958
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
6038
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
6170
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
642
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
6677
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
672
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
6748
vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
6853
to_vmx(vcpu)->nested.nested_run_pending = 0;
arch/x86/kvm/vmx/nested.c
6885
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
792
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
811
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
894
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.h
125
return vmx_misc_cr3_count(to_vmx(vcpu)->nested.msrs.misc_low);
arch/x86/kvm/vmx/nested.h
135
return to_vmx(vcpu)->nested.msrs.misc_low &
arch/x86/kvm/vmx/nested.h
141
return to_vmx(vcpu)->nested.msrs.misc_low & VMX_MISC_ZERO_LEN_INS;
arch/x86/kvm/vmx/nested.h
146
return to_vmx(vcpu)->nested.msrs.procbased_ctls_high &
arch/x86/kvm/vmx/nested.h
152
return to_vmx(vcpu)->nested.msrs.secondary_ctls_high &
arch/x86/kvm/vmx/nested.h
283
u64 fixed0 = to_vmx(vcpu)->nested.msrs.cr0_fixed0;
arch/x86/kvm/vmx/nested.h
284
u64 fixed1 = to_vmx(vcpu)->nested.msrs.cr0_fixed1;
arch/x86/kvm/vmx/nested.h
287
if (to_vmx(vcpu)->nested.msrs.secondary_ctls_high &
arch/x86/kvm/vmx/nested.h
297
u64 fixed0 = to_vmx(vcpu)->nested.msrs.cr0_fixed0;
arch/x86/kvm/vmx/nested.h
298
u64 fixed1 = to_vmx(vcpu)->nested.msrs.cr0_fixed1;
arch/x86/kvm/vmx/nested.h
305
u64 fixed0 = to_vmx(vcpu)->nested.msrs.cr4_fixed0;
arch/x86/kvm/vmx/nested.h
306
u64 fixed1 = to_vmx(vcpu)->nested.msrs.cr4_fixed1;
arch/x86/kvm/vmx/nested.h
314
return to_vmx(vcpu)->nested.msrs.basic & VMX_BASIC_NO_HW_ERROR_CODE_CC;
arch/x86/kvm/vmx/nested.h
40
if (to_vmx(vcpu)->fail || vm_exit_reason == -1 ||
arch/x86/kvm/vmx/nested.h
63
return to_vmx(vcpu)->nested.cached_vmcs12;
arch/x86/kvm/vmx/nested.h
71
return to_vmx(vcpu)->nested.cached_shadow_vmcs12;
arch/x86/kvm/vmx/nested.h
81
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.h
90
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/pmu_intel.c
44
return &to_vmx(vcpu)->lbr_desc;
arch/x86/kvm/vmx/pmu_intel.c
52
return &to_vmx(vcpu)->lbr_desc.records;
arch/x86/kvm/vmx/sgx.c
300
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/sgx.c
378
return (to_vmx(vcpu)->msr_ia32_feature_control & bits) == bits;
arch/x86/kvm/vmx/sgx.c
430
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
1351
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
1496
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
1512
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
1528
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
1600
vmx_prepare_switch_to_host(to_vmx(vcpu));
arch/x86/kvm/vmx/vmx.c
1606
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
1617
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
1633
vmx_switch_loaded_vmcs(vcpu, &to_vmx(vcpu)->nested.vmcs02);
arch/x86/kvm/vmx/vmx.c
1645
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
1663
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
1726
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
1813
if ((to_vmx(vcpu)->idt_vectoring_info & VECTORING_INFO_VALID_MASK) &&
arch/x86/kvm/vmx/vmx.c
1891
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
1942
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
2126
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
2160
msr_info->data = to_vmx(vcpu)->spec_ctrl;
arch/x86/kvm/vmx/vmx.c
2192
msr_info->data = to_vmx(vcpu)->msr_ia32_sgxlepubkeyhash
arch/x86/kvm/vmx/vmx.c
2332
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
2405
if (intel_pmu_lbr_is_enabled(vcpu) && !to_vmx(vcpu)->lbr_desc.event &&
arch/x86/kvm/vmx/vmx.c
2484
!(to_vmx(vcpu)->msr_ia32_feature_control &
arch/x86/kvm/vmx/vmx.c
2661
if (!(exec_controls_get(to_vmx(vcpu)) & CPU_BASED_CR3_LOAD_EXITING))
arch/x86/kvm/vmx/vmx.c
3194
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
3264
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
3311
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
3338
vmx_segment_cache_clear(to_vmx(vcpu));
arch/x86/kvm/vmx/vmx.c
3360
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
3385
return to_vmx(vcpu)->vpid;
arch/x86/kvm/vmx/vmx.c
3496
if (to_vmx(vcpu)->nested.vmxon)
arch/x86/kvm/vmx/vmx.c
3504
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
3637
if (to_vmx(vcpu)->nested.vmxon && !nested_cr4_valid(vcpu, cr4))
arch/x86/kvm/vmx/vmx.c
3646
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
3709
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
3747
if (to_vmx(vcpu)->rmode.vm86_active) {
arch/x86/kvm/vmx/vmx.c
3751
return vmx_read_guest_seg_base(to_vmx(vcpu), seg);
arch/x86/kvm/vmx/vmx.c
3756
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
3798
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
3837
to_vmx(vcpu)->vt.emulation_required = vmx_emulation_required(vcpu);
arch/x86/kvm/vmx/vmx.c
3842
u32 ar = vmx_read_guest_seg_ar(to_vmx(vcpu), VCPU_SREG_CS);
arch/x86/kvm/vmx/vmx.c
4188
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
4220
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
4273
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
4293
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
4378
!to_vmx(vcpu)->spec_ctrl);
arch/x86/kvm/vmx/vmx.c
4419
exec_controls_changebit(to_vmx(vcpu), CPU_BASED_RDPMC_EXITING,
arch/x86/kvm/vmx/vmx.c
4432
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
4676
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
5061
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
5093
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
5160
exec_controls_setbit(to_vmx(vcpu), CPU_BASED_INTR_WINDOW_EXITING);
arch/x86/kvm/vmx/vmx.c
5171
exec_controls_setbit(to_vmx(vcpu), CPU_BASED_NMI_WINDOW_EXITING);
arch/x86/kvm/vmx/vmx.c
5176
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
5204
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
5235
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
5249
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
5272
if (!enable_vnmi && to_vmx(vcpu)->loaded_vmcs->soft_vnmi_blocked)
arch/x86/kvm/vmx/vmx.c
5282
if (to_vmx(vcpu)->nested.nested_run_pending)
arch/x86/kvm/vmx/vmx.c
5309
if (to_vmx(vcpu)->nested.nested_run_pending)
arch/x86/kvm/vmx/vmx.c
5356
to_vmx(vcpu)->vcpu.arch.event_exit_inst_len =
arch/x86/kvm/vmx/vmx.c
5479
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
5858
exec_controls_clearbit(to_vmx(vcpu), CPU_BASED_MOV_DR_EXITING);
arch/x86/kvm/vmx/vmx.c
5891
exec_controls_setbit(to_vmx(vcpu), CPU_BASED_MOV_DR_EXITING);
arch/x86/kvm/vmx/vmx.c
5913
exec_controls_clearbit(to_vmx(vcpu), CPU_BASED_INTR_WINDOW_EXITING);
arch/x86/kvm/vmx/vmx.c
5980
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
6046
if (!(to_vmx(vcpu)->idt_vectoring_info & VECTORING_INFO_VALID_MASK) &&
arch/x86/kvm/vmx/vmx.c
6094
exec_controls_clearbit(to_vmx(vcpu), CPU_BASED_NMI_WINDOW_EXITING);
arch/x86/kvm/vmx/vmx.c
6108
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
6135
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
6258
if (!(to_vmx(vcpu)->idt_vectoring_info & VECTORING_INFO_VALID_MASK) &&
arch/x86/kvm/vmx/vmx.c
6274
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
635
evmcs = (struct hv_enlightened_vmcs *)to_vmx(vcpu)->loaded_vmcs->vmcs;
arch/x86/kvm/vmx/vmx.c
6467
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
6504
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
6543
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
6578
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
6783
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
6972
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
7043
if (!(secondary_exec_controls_get(to_vmx(vcpu)) &
arch/x86/kvm/vmx/vmx.c
7445
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
7467
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
7569
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
7608
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
7760
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
7777
vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
7960
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
8005
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
8074
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
8421
vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
8456
to_vmx(vcpu)->hv_deadline_tsc = -1;
arch/x86/kvm/vmx/vmx.c
8462
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
8483
to_vmx(vcpu)->msr_ia32_feature_control_valid_bits |=
arch/x86/kvm/vmx/vmx.c
8486
to_vmx(vcpu)->msr_ia32_feature_control_valid_bits &=
arch/x86/kvm/vmx/vmx.c
8494
if (to_vmx(vcpu)->nested.nested_run_pending)
arch/x86/kvm/vmx/vmx.c
8501
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
8522
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
8553
return to_vmx(vcpu)->nested.vmxon && !is_guest_mode(vcpu);
arch/x86/kvm/vmx/vmx.c
8559
struct hrtimer *timer = &to_vmx(vcpu)->nested.preemption_timer;
arch/x86/kvm/vmx/vmx.c
951
if (to_vmx(vcpu)->rmode.vm86_active)
arch/x86/kvm/vmx/vmx.h
698
(secondary_exec_controls_get(to_vmx(vcpu)) &