arch/x86/kvm/svm/avic.c
1006
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/avic.c
1072
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/avic.c
1126
u64 entry = to_svm(vcpu)->avic_physical_id_entry;
arch/x86/kvm/svm/avic.c
1152
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/avic.c
384
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/avic.c
611
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/avic.c
713
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/avic.c
728
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/avic.c
747
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/avic.c
759
u32 offset = to_svm(vcpu)->vmcb->control.exit_info_1 &
arch/x86/kvm/svm/avic.c
811
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/avic.c
869
raw_spin_lock_irqsave(&to_svm(vcpu)->ir_list_lock, flags);
arch/x86/kvm/svm/avic.c
871
raw_spin_unlock_irqrestore(&to_svm(vcpu)->ir_list_lock, flags);
arch/x86/kvm/svm/avic.c
895
.vapic_addr = avic_get_backing_page_address(to_svm(vcpu)),
arch/x86/kvm/svm/avic.c
898
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/avic.c
977
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/hyperv.c
11
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/hyperv.h
17
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/hyperv.h
31
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/nested.c
1345
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/nested.c
1351
nested_svm_simple_vmexit(to_svm(vcpu), SVM_EXIT_SHUTDOWN);
arch/x86/kvm/svm/nested.c
1407
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/nested.c
1530
if (to_svm(vcpu)->vmcb->save.cpl) {
arch/x86/kvm/svm/nested.c
1541
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/nested.c
1549
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/nested.c
1589
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/nested.c
1699
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/nested.c
1761
svm = to_svm(vcpu);
arch/x86/kvm/svm/nested.c
1814
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/nested.c
1948
!nested_npt_enabled(to_svm(vcpu)) && is_pae_paging(vcpu))
arch/x86/kvm/svm/nested.c
281
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/nested.c
39
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/nested.c
407
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/nested.c
415
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/nested.c
60
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/nested.c
78
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/nested.c
85
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/nested.c
946
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/nested.c
997
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/sev.c
2014
dst_svm = to_svm(dst_vcpu);
arch/x86/kvm/svm/sev.c
2026
src_svm = to_svm(src_vcpu);
arch/x86/kvm/svm/sev.c
2454
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/sev.c
3261
svm = to_svm(vcpu);
arch/x86/kvm/svm/sev.c
3721
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/sev.c
3814
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/sev.c
3964
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/sev.c
4052
target_svm = to_svm(target_vcpu);
arch/x86/kvm/svm/sev.c
4163
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/sev.c
4391
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/sev.c
4691
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/sev.c
4776
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/sev.c
5114
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/sev.c
974
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
1007
if (to_svm(vcpu)->guest_state_loaded)
arch/x86/kvm/svm/svm.c
1015
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
1069
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
1223
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
1237
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
1266
svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
1308
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
1380
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
1420
to_svm(vcpu)->guest_state_loaded = false;
arch/x86/kvm/svm/svm.c
1444
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
1459
if (to_svm(vcpu)->nmi_singlestep)
arch/x86/kvm/svm/svm.c
1467
to_svm(vcpu)->vmcb->save.rflags = rflags;
arch/x86/kvm/svm/svm.c
1472
struct vmcb *vmcb = to_svm(vcpu)->vmcb;
arch/x86/kvm/svm/svm.c
1553
struct vmcb_save_area *save = &to_svm(vcpu)->vmcb->save;
arch/x86/kvm/svm/svm.c
1554
struct vmcb_save_area *save01 = &to_svm(vcpu)->vmcb01.ptr->save;
arch/x86/kvm/svm/svm.c
1641
var->dpl = to_svm(vcpu)->vmcb->save.cpl;
arch/x86/kvm/svm/svm.c
1648
struct vmcb_save_area *save = &to_svm(vcpu)->vmcb->save;
arch/x86/kvm/svm/svm.c
1664
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
1672
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
1681
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
1689
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
1698
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
1721
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
1794
to_svm(vcpu)->vmcb->save.cr4 = cr4;
arch/x86/kvm/svm/svm.c
1795
vmcb_mark_dirty(to_svm(vcpu)->vmcb, VMCB_CR);
arch/x86/kvm/svm/svm.c
1804
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
1834
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
1859
struct vmcb *vmcb = to_svm(vcpu)->vmcb;
arch/x86/kvm/svm/svm.c
1872
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
1893
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
1904
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
1920
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
1971
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
2003
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
205
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
2084
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
2113
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
2154
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
2246
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
2269
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
2348
svm_set_gif(to_svm(vcpu), true);
arch/x86/kvm/svm/svm.c
2360
svm_set_gif(to_svm(vcpu), false);
arch/x86/kvm/svm/svm.c
2373
trace_kvm_invlpga(to_svm(vcpu)->vmcb->save.rip, asid, gva);
arch/x86/kvm/svm/svm.c
2383
trace_kvm_skinit(to_svm(vcpu)->vmcb->save.rip, kvm_rax_read(vcpu));
arch/x86/kvm/svm/svm.c
2391
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
2469
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
2488
kvm_mmu_invlpg(vcpu, to_svm(vcpu)->vmcb->control.exit_info_1);
arch/x86/kvm/svm/svm.c
2505
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
2528
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
256
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
2602
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
2637
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
266
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
2704
msr_info.data = to_svm(vcpu)->vmcb->control.exit_info_1 & ~EFER_SVME;
arch/x86/kvm/svm/svm.c
2736
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
279
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
2864
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
2874
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
2899
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
3123
if (to_svm(vcpu)->vmcb->control.exit_info_1)
arch/x86/kvm/svm/svm.c
3132
svm_clear_vintr(to_svm(vcpu));
arch/x86/kvm/svm/svm.c
3170
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
3203
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
3218
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
324
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
3312
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
3571
struct vmcb_control_area *control = &to_svm(vcpu)->vmcb->control;
arch/x86/kvm/svm/svm.c
3587
struct vmcb_control_area *control = &to_svm(vcpu)->vmcb->control;
arch/x86/kvm/svm/svm.c
3601
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
3644
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
3669
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
369
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
3690
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
3700
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
3724
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
3794
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
3817
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
3827
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
3846
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
3863
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
3878
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
3904
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
3924
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
3954
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
4009
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
4064
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
4078
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
4091
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
4107
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
4136
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
4208
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
4227
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
4259
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
4291
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
4445
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
4504
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
4606
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
4737
switch (to_svm(vcpu)->vmcb->control.exit_code) {
arch/x86/kvm/svm/svm.c
4758
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
4769
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
4785
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
4837
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
4903
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
4918
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
5070
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
661
void *msrpm = is_guest_mode(vcpu) ? to_svm(vcpu)->nested.msrpm :
arch/x86/kvm/svm/svm.c
662
to_svm(vcpu)->msrpm;
arch/x86/kvm/svm/svm.c
669
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
712
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
770
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
857
to_svm(vcpu)->vmcb->control.virt_ext |= LBR_CTL_ENABLE_MASK;
arch/x86/kvm/svm/svm.c
869
to_svm(vcpu)->vmcb->control.virt_ext &= ~LBR_CTL_ENABLE_MASK;
arch/x86/kvm/svm/svm.c
874
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
909
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
930
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
983
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
990
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.c
997
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm.h
757
struct vcpu_svm *svm = to_svm(vcpu);
arch/x86/kvm/svm/svm_onhyperv.c
26
hve = &to_svm(vcpu)->vmcb->control.hv_enlightenments;
arch/x86/kvm/svm/svm_onhyperv.c
32
vmcb_mark_dirty(to_svm(vcpu)->vmcb, HV_VMCB_NESTED_ENLIGHTENMENTS);
arch/x86/kvm/svm/svm_onhyperv.h
20
struct hv_vmcb_enlightenments *hve = &to_svm(vcpu)->vmcb->control.hv_enlightenments;
arch/x86/kvm/svm/svm_onhyperv.h
44
struct vmcb *vmcb = to_svm(vcpu)->vmcb;