arch/x86/include/asm/svm.h
566
BUILD_BUG_ON(offsetof(struct vmcb, save) != EXPECTED_VMCB_CONTROL_AREA_SIZE);
arch/x86/kvm/svm/avic.c
1153
struct vmcb *vmcb = svm->vmcb01.ptr;
arch/x86/kvm/svm/avic.c
1171
vmcb_mark_dirty(vmcb, VMCB_AVIC);
arch/x86/kvm/svm/avic.c
188
struct vmcb *vmcb = svm->vmcb01.ptr;
arch/x86/kvm/svm/avic.c
191
vmcb->control.int_ctl &= ~(AVIC_ENABLE_MASK | X2APIC_MODE_MASK);
arch/x86/kvm/svm/avic.c
192
vmcb->control.avic_physical_id &= ~AVIC_PHYSICAL_MAX_INDEX_MASK;
arch/x86/kvm/svm/avic.c
193
vmcb->control.avic_physical_id |= avic_get_max_physical_id(vcpu);
arch/x86/kvm/svm/avic.c
194
vmcb->control.int_ctl |= AVIC_ENABLE_MASK;
arch/x86/kvm/svm/avic.c
206
vmcb->control.int_ctl |= X2APIC_MODE_MASK;
arch/x86/kvm/svm/avic.c
224
struct vmcb *vmcb = svm->vmcb01.ptr;
arch/x86/kvm/svm/avic.c
226
vmcb->control.int_ctl &= ~(AVIC_ENABLE_MASK | X2APIC_MODE_MASK);
arch/x86/kvm/svm/avic.c
227
vmcb->control.avic_physical_id &= ~AVIC_PHYSICAL_MAX_INDEX_MASK;
arch/x86/kvm/svm/avic.c
365
void avic_init_vmcb(struct vcpu_svm *svm, struct vmcb *vmcb)
arch/x86/kvm/svm/avic.c
369
vmcb->control.avic_backing_page = avic_get_backing_page_address(svm);
arch/x86/kvm/svm/avic.c
370
vmcb->control.avic_logical_id = __sme_set(__pa(kvm_svm->avic_logical_id_table));
arch/x86/kvm/svm/avic.c
371
vmcb->control.avic_physical_id = __sme_set(__pa(kvm_svm->avic_physical_id_table));
arch/x86/kvm/svm/avic.c
372
vmcb->control.avic_vapic_bar = APIC_DEFAULT_PHYS_BASE;
arch/x86/kvm/svm/avic.c
612
u32 icrh = svm->vmcb->control.exit_info_1 >> 32;
arch/x86/kvm/svm/avic.c
613
u32 icrl = svm->vmcb->control.exit_info_1;
arch/x86/kvm/svm/avic.c
614
u32 id = svm->vmcb->control.exit_info_2 >> 32;
arch/x86/kvm/svm/avic.c
615
u32 index = svm->vmcb->control.exit_info_2 & AVIC_PHYSICAL_MAX_INDEX_MASK;
arch/x86/kvm/svm/avic.c
759
u32 offset = to_svm(vcpu)->vmcb->control.exit_info_1 &
arch/x86/kvm/svm/avic.c
813
u32 offset = svm->vmcb->control.exit_info_1 &
arch/x86/kvm/svm/avic.c
815
u32 vector = svm->vmcb->control.exit_info_2 &
arch/x86/kvm/svm/avic.c
817
bool write = (svm->vmcb->control.exit_info_1 >> 32) &
arch/x86/kvm/svm/hyperv.c
19
svm->vmcb->control.exit_code = HV_SVM_EXITCODE_ENL;
arch/x86/kvm/svm/hyperv.c
20
svm->vmcb->control.exit_info_1 = HV_SVM_ENL_EXITCODE_TRAP_AFTER_FLUSH;
arch/x86/kvm/svm/hyperv.c
21
svm->vmcb->control.exit_info_2 = 0;
arch/x86/kvm/svm/nested.c
1002
struct vmcb *vmcb01 = svm->vmcb01.ptr;
arch/x86/kvm/svm/nested.c
1021
vmcb12_gpa = svm->vmcb->save.rax;
arch/x86/kvm/svm/nested.c
1073
svm->vmcb->control.exit_code = SVM_EXIT_ERR;
arch/x86/kvm/svm/nested.c
1074
svm->vmcb->control.exit_info_1 = 0;
arch/x86/kvm/svm/nested.c
1075
svm->vmcb->control.exit_info_2 = 0;
arch/x86/kvm/svm/nested.c
1112
void svm_copy_vmloadsave_state(struct vmcb *to_vmcb, struct vmcb *from_vmcb)
arch/x86/kvm/svm/nested.c
1131
struct vmcb *vmcb01 = svm->vmcb01.ptr;
arch/x86/kvm/svm/nested.c
1132
struct vmcb *vmcb02 = svm->nested.vmcb02.ptr;
arch/x86/kvm/svm/nested.c
1133
struct vmcb *vmcb12;
arch/x86/kvm/svm/nested.c
131
vmcb_mark_dirty(svm->vmcb, VMCB_INTERCEPTS);
arch/x86/kvm/svm/nested.c
136
c = &svm->vmcb->control;
arch/x86/kvm/svm/nested.c
1384
if (WARN_ON_ONCE(svm->vmcb != svm->vmcb01.ptr))
arch/x86/kvm/svm/nested.c
1418
vmcb_mark_all_dirty(svm->vmcb);
arch/x86/kvm/svm/nested.c
1441
write = svm->vmcb->control.exit_info_1 & 1;
arch/x86/kvm/svm/nested.c
1464
port = svm->vmcb->control.exit_info_1 >> 16;
arch/x86/kvm/svm/nested.c
1465
size = (svm->vmcb->control.exit_info_1 & SVM_IOIO_SIZE_MASK) >>
arch/x86/kvm/svm/nested.c
1481
u64 exit_code = svm->vmcb->control.exit_code;
arch/x86/kvm/svm/nested.c
1530
if (to_svm(vcpu)->vmcb->save.cpl) {
arch/x86/kvm/svm/nested.c
1550
struct vmcb *vmcb = svm->vmcb;
arch/x86/kvm/svm/nested.c
1552
vmcb->control.exit_code = SVM_EXIT_EXCP_BASE + ex->vector;
arch/x86/kvm/svm/nested.c
1555
vmcb->control.exit_info_1 = ex->error_code;
arch/x86/kvm/svm/nested.c
1563
vmcb->control.exit_info_2 = ex->payload;
arch/x86/kvm/svm/nested.c
1565
vmcb->control.exit_info_2 = vcpu->arch.cr2;
arch/x86/kvm/svm/nested.c
1653
trace_kvm_nested_intr_vmexit(svm->vmcb->save.rip);
arch/x86/kvm/svm/nested.c
1663
u32 exit_code = svm->vmcb->control.exit_code;
arch/x86/kvm/svm/nested.c
1755
struct vmcb __user *user_vmcb = (struct vmcb __user *)
arch/x86/kvm/svm/nested.c
1815
struct vmcb __user *user_vmcb = (struct vmcb __user *)
arch/x86/kvm/svm/nested.c
1918
nested_vmcb02_prepare_control(svm, svm->vmcb->save.rip, svm->vmcb->save.cs.base);
arch/x86/kvm/svm/nested.c
326
svm->vmcb->control.msrpm_base_pa = __sme_set(__pa(svm->nested.msrpm));
arch/x86/kvm/svm/nested.c
40
struct vmcb *vmcb = svm->vmcb;
arch/x86/kvm/svm/nested.c
42
if (vmcb->control.exit_code != SVM_EXIT_NPF) {
arch/x86/kvm/svm/nested.c
47
vmcb->control.exit_code = SVM_EXIT_NPF;
arch/x86/kvm/svm/nested.c
48
vmcb->control.exit_info_1 = (1ULL << 32);
arch/x86/kvm/svm/nested.c
49
vmcb->control.exit_info_2 = fault->address;
arch/x86/kvm/svm/nested.c
52
vmcb->control.exit_info_1 &= ~0xffffffffULL;
arch/x86/kvm/svm/nested.c
53
vmcb->control.exit_info_1 |= fault->error_code;
arch/x86/kvm/svm/nested.c
531
svm->nested.ctl.event_inj = svm->vmcb->control.event_inj;
arch/x86/kvm/svm/nested.c
532
svm->nested.ctl.event_inj_err = svm->vmcb->control.event_inj_err;
arch/x86/kvm/svm/nested.c
555
svm->nested.ctl.int_ctl |= svm->vmcb->control.int_ctl & mask;
arch/x86/kvm/svm/nested.c
563
struct vmcb *vmcb12)
arch/x86/kvm/svm/nested.c
651
static void nested_vmcb02_prepare_save(struct vcpu_svm *svm, struct vmcb *vmcb12)
arch/x86/kvm/svm/nested.c
654
struct vmcb *vmcb01 = svm->vmcb01.ptr;
arch/x86/kvm/svm/nested.c
655
struct vmcb *vmcb02 = svm->nested.vmcb02.ptr;
arch/x86/kvm/svm/nested.c
761
struct vmcb *vmcb01 = svm->vmcb01.ptr;
arch/x86/kvm/svm/nested.c
762
struct vmcb *vmcb02 = svm->nested.vmcb02.ptr;
arch/x86/kvm/svm/nested.c
931
static void nested_svm_copy_common_state(struct vmcb *from_vmcb, struct vmcb *to_vmcb)
arch/x86/kvm/svm/nested.c
944
struct vmcb *vmcb12, bool from_vmrun)
arch/x86/kvm/svm/nested.c
949
trace_kvm_nested_vmenter(svm->vmcb->save.rip,
arch/x86/kvm/svm/nested.c
969
WARN_ON(svm->vmcb == svm->nested.vmcb02.ptr);
arch/x86/kvm/svm/nested.c
999
struct vmcb *vmcb12;
arch/x86/kvm/svm/sev.c
2034
dst_svm->vmcb->control.ghcb_gpa = src_svm->vmcb->control.ghcb_gpa;
arch/x86/kvm/svm/sev.c
2035
dst_svm->vmcb->control.vmsa_pa = src_svm->vmcb->control.vmsa_pa;
arch/x86/kvm/svm/sev.c
2039
src_svm->vmcb->control.ghcb_gpa = INVALID_PAGE;
arch/x86/kvm/svm/sev.c
2040
src_svm->vmcb->control.vmsa_pa = INVALID_PAGE;
arch/x86/kvm/svm/sev.c
3287
struct vmcb_control_area *control = &svm->vmcb->control;
arch/x86/kvm/svm/sev.c
3304
pr_err("GHCB (GPA=%016llx) snapshot:\n", svm->vmcb->control.ghcb_gpa);
arch/x86/kvm/svm/sev.c
3337
struct vmcb_control_area *control = &svm->vmcb->control;
arch/x86/kvm/svm/sev.c
3364
svm->vmcb->save.cpl = kvm_ghcb_get_cpl_if_valid(svm);
arch/x86/kvm/svm/sev.c
3384
struct vmcb_control_area *control = &svm->vmcb->control;
arch/x86/kvm/svm/sev.c
3570
if (sev_es_guest(kvm) && !VALID_PAGE(svm->vmcb->control.vmsa_pa))
arch/x86/kvm/svm/sev.c
3591
if (sd->sev_vmcbs[asid] == svm->vmcb &&
arch/x86/kvm/svm/sev.c
3595
sd->sev_vmcbs[asid] = svm->vmcb;
arch/x86/kvm/svm/sev.c
3596
svm->vmcb->control.tlb_ctl = TLB_CONTROL_FLUSH_ASID;
arch/x86/kvm/svm/sev.c
3597
vmcb_mark_dirty(svm->vmcb, VMCB_ASID);
arch/x86/kvm/svm/sev.c
3604
struct vmcb_control_area *control = &svm->vmcb->control;
arch/x86/kvm/svm/sev.c
3688
svm->vmcb->control.ghcb_gpa &= ~(mask << pos);
arch/x86/kvm/svm/sev.c
3689
svm->vmcb->control.ghcb_gpa |= (value & mask) << pos;
arch/x86/kvm/svm/sev.c
3694
return (svm->vmcb->control.ghcb_gpa >> pos) & mask;
arch/x86/kvm/svm/sev.c
3699
svm->vmcb->control.ghcb_gpa = value;
arch/x86/kvm/svm/sev.c
3982
svm->vmcb->control.vmsa_pa = INVALID_PAGE;
arch/x86/kvm/svm/sev.c
3988
vmcb_mark_all_dirty(svm->vmcb);
arch/x86/kvm/svm/sev.c
4019
svm->vmcb->control.vmsa_pa = pfn_to_hpa(pfn);
arch/x86/kvm/svm/sev.c
4041
request = lower_32_bits(svm->vmcb->control.exit_info_1);
arch/x86/kvm/svm/sev.c
4042
apic_id = upper_32_bits(svm->vmcb->control.exit_info_1);
arch/x86/kvm/svm/sev.c
4065
if (!page_address_valid(vcpu, svm->vmcb->control.exit_info_2)) {
arch/x86/kvm/svm/sev.c
4067
svm->vmcb->control.exit_info_2);
arch/x86/kvm/svm/sev.c
4078
if (IS_ALIGNED(svm->vmcb->control.exit_info_2, PMD_SIZE)) {
arch/x86/kvm/svm/sev.c
4081
svm->vmcb->control.exit_info_2);
arch/x86/kvm/svm/sev.c
4085
target_svm->sev_es.snp_vmsa_gpa = svm->vmcb->control.exit_info_2;
arch/x86/kvm/svm/sev.c
4164
struct vmcb_control_area *control = &svm->vmcb->control;
arch/x86/kvm/svm/sev.c
4247
struct vmcb_control_area *control = &svm->vmcb->control;
arch/x86/kvm/svm/sev.c
4392
struct vmcb_control_area *control = &svm->vmcb->control;
arch/x86/kvm/svm/sev.c
4539
if (svm->vmcb->control.exit_info_2 > INT_MAX)
arch/x86/kvm/svm/sev.c
4542
count = svm->vmcb->control.exit_info_2;
arch/x86/kvm/svm/sev.c
4600
struct vmcb *vmcb = svm->vmcb01.ptr;
arch/x86/kvm/svm/sev.c
4602
svm->vmcb->control.nested_ctl |= SVM_NESTED_CTL_SEV_ES_ENABLE;
arch/x86/kvm/svm/sev.c
4613
svm->vmcb->control.vmsa_pa = __pa(svm->sev_es.vmsa);
arch/x86/kvm/svm/sev.c
4615
svm->vmcb->control.vmsa_pa = INVALID_PAGE;
arch/x86/kvm/svm/sev.c
4619
svm->vmcb->control.allowed_sev_features = sev->vmsa_features |
arch/x86/kvm/svm/sev.c
4638
vmcb->control.intercepts[INTERCEPT_DR] = 0;
arch/x86/kvm/svm/sev.c
4640
vmcb_set_intercept(&vmcb->control, INTERCEPT_DR7_READ);
arch/x86/kvm/svm/sev.c
4641
vmcb_set_intercept(&vmcb->control, INTERCEPT_DR7_WRITE);
arch/x86/kvm/svm/sev.c
4673
svm->vmcb->control.nested_ctl |= SVM_NESTED_CTL_SEV_ENABLE;
arch/x86/kvm/svm/sev.c
5149
dbg.src_addr = svm->vmcb->control.vmsa_pa;
arch/x86/kvm/svm/sev.c
5180
dbg.src_addr = svm->vmcb->control.vmsa_pa;
arch/x86/kvm/svm/sev.c
886
if (svm->vcpu.guest_debug || (svm->vmcb->save.dr7 & ~DR7_FIXED_1))
arch/x86/kvm/svm/sev.c
895
memcpy(save, &svm->vmcb->save, sizeof(svm->vmcb->save));
arch/x86/kvm/svm/svm.c
1000
svm->vmcb->control.tsc_offset = vcpu->arch.tsc_offset;
arch/x86/kvm/svm/svm.c
1001
vmcb_mark_dirty(svm->vmcb, VMCB_INTERCEPTS);
arch/x86/kvm/svm/svm.c
1070
struct vmcb *vmcb = svm->vmcb01.ptr;
arch/x86/kvm/svm/svm.c
1071
struct vmcb_control_area *control = &vmcb->control;
arch/x86/kvm/svm/svm.c
1072
struct vmcb_save_area *save = &vmcb->save;
arch/x86/kvm/svm/svm.c
1189
svm->vmcb->control.erap_ctl |= ERAP_CONTROL_ALLOW_LARGER_RAP;
arch/x86/kvm/svm/svm.c
1192
avic_init_vmcb(svm, vmcb);
arch/x86/kvm/svm/svm.c
1195
svm->vmcb->control.int_ctl |= V_NMI_ENABLE_MASK;
arch/x86/kvm/svm/svm.c
1200
svm->vmcb->control.int_ctl |= V_GIF_ENABLE_MASK;
arch/x86/kvm/svm/svm.c
1204
svm->vmcb->control.virt_ext |= VIRTUAL_VMLOAD_VMSAVE_ENABLE_MASK;
arch/x86/kvm/svm/svm.c
1212
svm_hv_init_vmcb(vmcb);
arch/x86/kvm/svm/svm.c
1216
vmcb_mark_all_dirty(vmcb);
arch/x86/kvm/svm/svm.c
1251
svm->vmcb = target_vmcb->ptr;
arch/x86/kvm/svm/svm.c
1445
unsigned long rflags = svm->vmcb->save.rflags;
arch/x86/kvm/svm/svm.c
1467
to_svm(vcpu)->vmcb->save.rflags = rflags;
arch/x86/kvm/svm/svm.c
1472
struct vmcb *vmcb = to_svm(vcpu)->vmcb;
arch/x86/kvm/svm/svm.c
1475
? vmcb->control.int_state & SVM_GUEST_INTERRUPT_MASK
arch/x86/kvm/svm/svm.c
1522
control = &svm->vmcb->control;
arch/x86/kvm/svm/svm.c
1527
vmcb_mark_dirty(svm->vmcb, VMCB_INTR);
arch/x86/kvm/svm/svm.c
1535
svm->vmcb->control.int_ctl &= ~V_IRQ_INJECTION_BITS_MASK;
arch/x86/kvm/svm/svm.c
1539
WARN_ON((svm->vmcb->control.int_ctl & V_TPR_MASK) !=
arch/x86/kvm/svm/svm.c
1542
svm->vmcb->control.int_ctl |= svm->nested.ctl.int_ctl &
arch/x86/kvm/svm/svm.c
1545
svm->vmcb->control.int_vector = svm->nested.ctl.int_vector;
arch/x86/kvm/svm/svm.c
1548
vmcb_mark_dirty(svm->vmcb, VMCB_INTR);
arch/x86/kvm/svm/svm.c
1553
struct vmcb_save_area *save = &to_svm(vcpu)->vmcb->save;
arch/x86/kvm/svm/svm.c
1641
var->dpl = to_svm(vcpu)->vmcb->save.cpl;
arch/x86/kvm/svm/svm.c
1648
struct vmcb_save_area *save = &to_svm(vcpu)->vmcb->save;
arch/x86/kvm/svm/svm.c
1666
dt->size = svm->vmcb->save.idtr.limit;
arch/x86/kvm/svm/svm.c
1667
dt->address = svm->vmcb->save.idtr.base;
arch/x86/kvm/svm/svm.c
1674
svm->vmcb->save.idtr.limit = dt->size;
arch/x86/kvm/svm/svm.c
1675
svm->vmcb->save.idtr.base = dt->address ;
arch/x86/kvm/svm/svm.c
1676
vmcb_mark_dirty(svm->vmcb, VMCB_DT);
arch/x86/kvm/svm/svm.c
1683
dt->size = svm->vmcb->save.gdtr.limit;
arch/x86/kvm/svm/svm.c
1684
dt->address = svm->vmcb->save.gdtr.base;
arch/x86/kvm/svm/svm.c
1691
svm->vmcb->save.gdtr.limit = dt->size;
arch/x86/kvm/svm/svm.c
1692
svm->vmcb->save.gdtr.base = dt->address ;
arch/x86/kvm/svm/svm.c
1693
vmcb_mark_dirty(svm->vmcb, VMCB_DT);
arch/x86/kvm/svm/svm.c
1709
svm->vmcb->save.cr3 = cr3;
arch/x86/kvm/svm/svm.c
1710
vmcb_mark_dirty(svm->vmcb, VMCB_CR);
arch/x86/kvm/svm/svm.c
1730
svm->vmcb->save.efer |= EFER_LMA | EFER_LME;
arch/x86/kvm/svm/svm.c
1736
svm->vmcb->save.efer &= ~(EFER_LMA | EFER_LME);
arch/x86/kvm/svm/svm.c
1756
svm->vmcb->save.cr0 = hcr0;
arch/x86/kvm/svm/svm.c
1757
vmcb_mark_dirty(svm->vmcb, VMCB_CR);
arch/x86/kvm/svm/svm.c
1794
to_svm(vcpu)->vmcb->save.cr4 = cr4;
arch/x86/kvm/svm/svm.c
1795
vmcb_mark_dirty(to_svm(vcpu)->vmcb, VMCB_CR);
arch/x86/kvm/svm/svm.c
1827
svm->vmcb->save.cpl = (var->dpl & 3);
arch/x86/kvm/svm/svm.c
1829
vmcb_mark_dirty(svm->vmcb, VMCB_SEG);
arch/x86/kvm/svm/svm.c
1849
svm->vmcb->control.tlb_ctl = TLB_CONTROL_FLUSH_ALL_ASID;
arch/x86/kvm/svm/svm.c
1850
vmcb_mark_dirty(svm->vmcb, VMCB_ASID);
arch/x86/kvm/svm/svm.c
1859
struct vmcb *vmcb = to_svm(vcpu)->vmcb;
arch/x86/kvm/svm/svm.c
1864
if (unlikely(value != vmcb->save.dr6)) {
arch/x86/kvm/svm/svm.c
1865
vmcb->save.dr6 = value;
arch/x86/kvm/svm/svm.c
1866
vmcb_mark_dirty(vmcb, VMCB_DR);
arch/x86/kvm/svm/svm.c
1885
vcpu->arch.dr6 = svm->vmcb->save.dr6;
arch/x86/kvm/svm/svm.c
1886
vcpu->arch.dr7 = svm->vmcb->save.dr7;
arch/x86/kvm/svm/svm.c
1898
svm->vmcb->save.dr7 = value;
arch/x86/kvm/svm/svm.c
1899
vmcb_mark_dirty(svm->vmcb, VMCB_DR);
arch/x86/kvm/svm/svm.c
1906
u64 fault_address = svm->vmcb->control.exit_info_2;
arch/x86/kvm/svm/svm.c
1907
u64 error_code = svm->vmcb->control.exit_info_1;
arch/x86/kvm/svm/svm.c
1911
svm->vmcb->control.insn_bytes : NULL,
arch/x86/kvm/svm/svm.c
1912
svm->vmcb->control.insn_len);
arch/x86/kvm/svm/svm.c
1923
u64 error_code = svm->vmcb->control.exit_info_1;
arch/x86/kvm/svm/svm.c
1924
gpa_t gpa = svm->vmcb->control.exit_info_2;
arch/x86/kvm/svm/svm.c
1946
if (nrips && svm->vmcb->control.next_rip &&
arch/x86/kvm/svm/svm.c
1959
svm->vmcb->control.insn_bytes : NULL,
arch/x86/kvm/svm/svm.c
1960
svm->vmcb->control.insn_len);
arch/x86/kvm/svm/svm.c
1976
u32 payload = svm->vmcb->save.dr6 ^ DR6_ACTIVE_LOW;
arch/x86/kvm/svm/svm.c
1990
kvm_run->debug.arch.dr6 = svm->vmcb->save.dr6;
arch/x86/kvm/svm/svm.c
1991
kvm_run->debug.arch.dr7 = svm->vmcb->save.dr7;
arch/x86/kvm/svm/svm.c
1993
svm->vmcb->save.cs.base + svm->vmcb->save.rip;
arch/x86/kvm/svm/svm.c
2007
kvm_run->debug.arch.pc = svm->vmcb->save.cs.base + svm->vmcb->save.rip;
arch/x86/kvm/svm/svm.c
2099
clear_page(svm->vmcb);
arch/x86/kvm/svm/svm.c
2114
u32 io_info = svm->vmcb->control.exit_info_1; /* address size bug? */
arch/x86/kvm/svm/svm.c
2131
svm->next_rip = svm->vmcb->control.exit_info_2;
arch/x86/kvm/svm/svm.c
2155
struct vmcb *vmcb12;
arch/x86/kvm/svm/svm.c
2162
ret = kvm_vcpu_map(vcpu, gpa_to_gfn(svm->vmcb->save.rax), &map);
arch/x86/kvm/svm/svm.c
2270
u32 error_code = svm->vmcb->control.exit_info_1;
arch/x86/kvm/svm/svm.c
2296
if (svm->vmcb->save.rax & ~PAGE_MASK)
arch/x86/kvm/svm/svm.c
2373
trace_kvm_invlpga(to_svm(vcpu)->vmcb->save.rip, asid, gva);
arch/x86/kvm/svm/svm.c
2383
trace_kvm_skinit(to_svm(vcpu)->vmcb->save.rip, kvm_rax_read(vcpu));
arch/x86/kvm/svm/svm.c
2394
int int_type = svm->vmcb->control.exit_int_info &
arch/x86/kvm/svm/svm.c
2396
int int_vec = svm->vmcb->control.exit_int_info & SVM_EVTINJ_VEC_MASK;
arch/x86/kvm/svm/svm.c
2398
svm->vmcb->control.exit_int_info & SVM_EXITINTINFO_TYPE_MASK;
arch/x86/kvm/svm/svm.c
2400
svm->vmcb->control.exit_int_info & SVM_EXITINTINFO_VALID;
arch/x86/kvm/svm/svm.c
2404
tss_selector = (u16)svm->vmcb->control.exit_info_1;
arch/x86/kvm/svm/svm.c
2406
if (svm->vmcb->control.exit_info_2 &
arch/x86/kvm/svm/svm.c
2409
else if (svm->vmcb->control.exit_info_2 &
arch/x86/kvm/svm/svm.c
2423
if (svm->vmcb->control.exit_info_2 &
arch/x86/kvm/svm/svm.c
2427
(u32)svm->vmcb->control.exit_info_2;
arch/x86/kvm/svm/svm.c
2488
kvm_mmu_invlpg(vcpu, to_svm(vcpu)->vmcb->control.exit_info_1);
arch/x86/kvm/svm/svm.c
249
svm->vmcb->save.efer = efer | EFER_SVME;
arch/x86/kvm/svm/svm.c
250
vmcb_mark_dirty(svm->vmcb, VMCB_CR);
arch/x86/kvm/svm/svm.c
2517
svm->vmcb->control.exit_code = SVM_EXIT_CR0_SEL_WRITE;
arch/x86/kvm/svm/svm.c
2536
if (unlikely((svm->vmcb->control.exit_info_1 & CR_VALID) == 0))
arch/x86/kvm/svm/svm.c
2539
reg = svm->vmcb->control.exit_info_1 & SVM_EXITINFO_REG_MASK;
arch/x86/kvm/svm/svm.c
2540
if (svm->vmcb->control.exit_code == SVM_EXIT_CR0_SEL_WRITE)
arch/x86/kvm/svm/svm.c
2543
cr = svm->vmcb->control.exit_code - SVM_EXIT_READ_CR0;
arch/x86/kvm/svm/svm.c
259
if (svm->vmcb->control.int_state & SVM_INTERRUPT_SHADOW_MASK)
arch/x86/kvm/svm/svm.c
2607
new_value = (unsigned long)svm->vmcb->control.exit_info_1;
arch/x86/kvm/svm/svm.c
2609
cr = svm->vmcb->control.exit_code - SVM_EXIT_CR0_WRITE_TRAP;
arch/x86/kvm/svm/svm.c
2662
reg = svm->vmcb->control.exit_info_1 & SVM_EXITINFO_REG_MASK;
arch/x86/kvm/svm/svm.c
2663
dr = svm->vmcb->control.exit_code - SVM_EXIT_READ_DR0;
arch/x86/kvm/svm/svm.c
269
svm->vmcb->control.int_state &= ~SVM_INTERRUPT_SHADOW_MASK;
arch/x86/kvm/svm/svm.c
2704
msr_info.data = to_svm(vcpu)->vmcb->control.exit_info_1 & ~EFER_SVME;
arch/x86/kvm/svm/svm.c
271
svm->vmcb->control.int_state |= SVM_INTERRUPT_SHADOW_MASK;
arch/x86/kvm/svm/svm.c
2787
msr_info->data = svm->vmcb->save.s_cet;
arch/x86/kvm/svm/svm.c
2790
msr_info->data = svm->vmcb->save.isst_addr;
arch/x86/kvm/svm/svm.c
2793
msr_info->data = svm->vmcb->save.ssp;
arch/x86/kvm/svm/svm.c
2799
msr_info->data = svm->vmcb->save.dbgctl;
arch/x86/kvm/svm/svm.c
2802
msr_info->data = svm->vmcb->save.br_from;
arch/x86/kvm/svm/svm.c
2805
msr_info->data = svm->vmcb->save.br_to;
arch/x86/kvm/svm/svm.c
2808
msr_info->data = svm->vmcb->save.last_excp_from;
arch/x86/kvm/svm/svm.c
2811
msr_info->data = svm->vmcb->save.last_excp_to;
arch/x86/kvm/svm/svm.c
2825
msr_info->data = svm->vmcb->save.spec_ctrl;
arch/x86/kvm/svm/svm.c
289
if (nrips && svm->vmcb->control.next_rip != 0) {
arch/x86/kvm/svm/svm.c
291
svm->next_rip = svm->vmcb->control.next_rip;
arch/x86/kvm/svm/svm.c
2946
vmcb_mark_dirty(svm->vmcb, VMCB_NPT);
arch/x86/kvm/svm/svm.c
2957
svm->vmcb->save.spec_ctrl = data;
arch/x86/kvm/svm/svm.c
296
old_rflags = svm->vmcb->save.rflags;
arch/x86/kvm/svm/svm.c
302
svm->vmcb->save.rflags = old_rflags;
arch/x86/kvm/svm/svm.c
3028
svm->vmcb->save.s_cet = data;
arch/x86/kvm/svm/svm.c
3032
svm->vmcb->save.isst_addr = data;
arch/x86/kvm/svm/svm.c
3036
svm->vmcb->save.ssp = data;
arch/x86/kvm/svm/svm.c
3079
if (svm->vmcb->save.dbgctl == data)
arch/x86/kvm/svm/svm.c
3082
svm->vmcb->save.dbgctl = data;
arch/x86/kvm/svm/svm.c
3083
vmcb_mark_dirty(svm->vmcb, VMCB_LBR);
arch/x86/kvm/svm/svm.c
3123
if (to_svm(vcpu)->vmcb->control.exit_info_1)
arch/x86/kvm/svm/svm.c
3184
type = svm->vmcb->control.exit_info_2;
arch/x86/kvm/svm/svm.c
3185
gva = svm->vmcb->control.exit_info_1;
arch/x86/kvm/svm/svm.c
3211
svm->vmcb->control.bus_lock_counter = 1;
arch/x86/kvm/svm/svm.c
3313
struct vmcb_control_area *control = &svm->vmcb->control;
arch/x86/kvm/svm/svm.c
3314
struct vmcb_save_area *save = &svm->vmcb->save;
arch/x86/kvm/svm/svm.c
353
svm->soft_int_csbase = svm->vmcb->save.cs.base;
arch/x86/kvm/svm/svm.c
3571
struct vmcb_control_area *control = &to_svm(vcpu)->vmcb->control;
arch/x86/kvm/svm/svm.c
3587
struct vmcb_control_area *control = &to_svm(vcpu)->vmcb->control;
arch/x86/kvm/svm/svm.c
3607
vcpu->arch.cr0 = svm->vmcb->save.cr0;
arch/x86/kvm/svm/svm.c
3609
vcpu->arch.cr3 = svm->vmcb->save.cr3;
arch/x86/kvm/svm/svm.c
361
svm->vmcb->control.next_rip = rip;
arch/x86/kvm/svm/svm.c
3626
if (svm_is_vmrun_failure(svm->vmcb->control.exit_code)) {
arch/x86/kvm/svm/svm.c
3629
= svm->vmcb->control.exit_code;
arch/x86/kvm/svm/svm.c
3638
return svm_invoke_exit_handler(vcpu, svm->vmcb->control.exit_code);
arch/x86/kvm/svm/svm.c
3653
vmcb_mark_all_dirty(svm->vmcb);
arch/x86/kvm/svm/svm.c
3671
svm->vmcb->control.event_inj = SVM_EVTINJ_VALID | SVM_EVTINJ_TYPE_NMI;
arch/x86/kvm/svm/svm.c
3695
return !!(svm->vmcb->control.int_ctl & V_NMI_PENDING_MASK);
arch/x86/kvm/svm/svm.c
3705
if (svm->vmcb->control.int_ctl & V_NMI_PENDING_MASK)
arch/x86/kvm/svm/svm.c
3708
svm->vmcb->control.int_ctl |= V_NMI_PENDING_MASK;
arch/x86/kvm/svm/svm.c
3709
vmcb_mark_dirty(svm->vmcb, VMCB_INTR);
arch/x86/kvm/svm/svm.c
3739
svm->vmcb->control.event_inj = intr->nr | SVM_EVTINJ_VALID | type;
arch/x86/kvm/svm/svm.c
377
svm->vmcb->control.event_inj = ex->vector
arch/x86/kvm/svm/svm.c
381
svm->vmcb->control.event_inj_err = ex->error_code;
arch/x86/kvm/svm/svm.c
3820
return svm->vmcb->control.int_ctl & V_NMI_BLOCKING_MASK;
arch/x86/kvm/svm/svm.c
3831
svm->vmcb->control.int_ctl |= V_NMI_BLOCKING_MASK;
arch/x86/kvm/svm/svm.c
3833
svm->vmcb->control.int_ctl &= ~V_NMI_BLOCKING_MASK;
arch/x86/kvm/svm/svm.c
3847
struct vmcb *vmcb = svm->vmcb;
arch/x86/kvm/svm/svm.c
3858
return vmcb->control.int_state & SVM_INTERRUPT_SHADOW_MASK;
arch/x86/kvm/svm/svm.c
3879
struct vmcb *vmcb = svm->vmcb;
arch/x86/kvm/svm/svm.c
3899
return (vmcb->control.int_state & SVM_INTERRUPT_SHADOW_MASK);
arch/x86/kvm/svm/svm.c
4004
svm->vmcb->save.rflags |= (X86_EFLAGS_TF | X86_EFLAGS_RF);
arch/x86/kvm/svm/svm.c
4026
svm->vmcb->control.tlb_ctl = TLB_CONTROL_FLUSH_ASID;
arch/x86/kvm/svm/svm.c
4066
invlpga(gva, svm->vmcb->control.asid);
arch/x86/kvm/svm/svm.c
4084
int cr8 = svm->vmcb->control.int_ctl & V_TPR_MASK;
arch/x86/kvm/svm/svm.c
4098
svm->vmcb->control.int_ctl &= ~V_TPR_MASK;
arch/x86/kvm/svm/svm.c
4099
svm->vmcb->control.int_ctl |= cr8 & V_TPR_MASK;
arch/x86/kvm/svm/svm.c
4120
svm->vmcb->control.next_rip = svm->soft_int_next_rip;
arch/x86/kvm/svm/svm.c
4139
u32 exitintinfo = svm->vmcb->control.exit_int_info;
arch/x86/kvm/svm/svm.c
4187
error_code = svm->vmcb->control.exit_int_info_err;
arch/x86/kvm/svm/svm.c
4209
struct vmcb_control_area *control = &svm->vmcb->control;
arch/x86/kvm/svm/svm.c
4228
struct vmcb_control_area *control = &svm->vmcb->control;
arch/x86/kvm/svm/svm.c
4296
svm->vmcb->save.rax = vcpu->arch.regs[VCPU_REGS_RAX];
arch/x86/kvm/svm/svm.c
4297
svm->vmcb->save.rsp = vcpu->arch.regs[VCPU_REGS_RSP];
arch/x86/kvm/svm/svm.c
4298
svm->vmcb->save.rip = vcpu->arch.regs[VCPU_REGS_RIP];
arch/x86/kvm/svm/svm.c
4306
if (svm->nmi_singlestep && svm->vmcb->control.event_inj) {
arch/x86/kvm/svm/svm.c
4328
if (unlikely(svm->asid != svm->vmcb->control.asid)) {
arch/x86/kvm/svm/svm.c
4329
svm->vmcb->control.asid = svm->asid;
arch/x86/kvm/svm/svm.c
4330
vmcb_mark_dirty(svm->vmcb, VMCB_ASID);
arch/x86/kvm/svm/svm.c
4332
svm->vmcb->save.cr2 = vcpu->arch.cr2;
arch/x86/kvm/svm/svm.c
4336
svm->vmcb->control.erap_ctl |= ERAP_CONTROL_CLEAR_RAP;
arch/x86/kvm/svm/svm.c
4338
svm_hv_update_vp_id(svm->vmcb, vcpu);
arch/x86/kvm/svm/svm.c
4358
if (!(svm->vmcb->control.virt_ext & LBR_CTL_ENABLE_MASK) &&
arch/x86/kvm/svm/svm.c
4359
vcpu->arch.host_debugctl != svm->vmcb->save.dbgctl)
arch/x86/kvm/svm/svm.c
4360
update_debugctlmsr(svm->vmcb->save.dbgctl);
arch/x86/kvm/svm/svm.c
4379
vcpu->arch.cr2 = svm->vmcb->save.cr2;
arch/x86/kvm/svm/svm.c
4380
vcpu->arch.regs[VCPU_REGS_RAX] = svm->vmcb->save.rax;
arch/x86/kvm/svm/svm.c
4381
vcpu->arch.regs[VCPU_REGS_RSP] = svm->vmcb->save.rsp;
arch/x86/kvm/svm/svm.c
4382
vcpu->arch.regs[VCPU_REGS_RIP] = svm->vmcb->save.rip;
arch/x86/kvm/svm/svm.c
4386
if (unlikely(svm->vmcb->control.exit_code == SVM_EXIT_NMI))
arch/x86/kvm/svm/svm.c
4389
if (!(svm->vmcb->control.virt_ext & LBR_CTL_ENABLE_MASK) &&
arch/x86/kvm/svm/svm.c
4390
vcpu->arch.host_debugctl != svm->vmcb->save.dbgctl)
arch/x86/kvm/svm/svm.c
4397
if (unlikely(svm->vmcb->control.exit_code == SVM_EXIT_NMI))
arch/x86/kvm/svm/svm.c
4408
!svm_is_vmrun_failure(svm->vmcb->control.exit_code))
arch/x86/kvm/svm/svm.c
4414
svm->vmcb->control.tlb_ctl = TLB_CONTROL_DO_NOTHING;
arch/x86/kvm/svm/svm.c
4421
svm->vmcb->control.erap_ctl &= ~ERAP_CONTROL_CLEAR_RAP;
arch/x86/kvm/svm/svm.c
4423
vmcb_mark_all_clean(svm->vmcb);
arch/x86/kvm/svm/svm.c
4426
if (svm->vmcb->control.exit_code == SVM_EXIT_EXCP_BASE + PF_VECTOR)
arch/x86/kvm/svm/svm.c
4449
svm->vmcb->control.nested_cr3 = __sme_set(root_hpa);
arch/x86/kvm/svm/svm.c
4450
vmcb_mark_dirty(svm->vmcb, VMCB_NPT);
arch/x86/kvm/svm/svm.c
4463
svm->vmcb->save.cr3 = cr3;
arch/x86/kvm/svm/svm.c
4464
vmcb_mark_dirty(svm->vmcb, VMCB_CR);
arch/x86/kvm/svm/svm.c
4609
struct vmcb *vmcb = svm->vmcb;
arch/x86/kvm/svm/svm.c
4674
vmcb->control.exit_info_1 = 1;
arch/x86/kvm/svm/svm.c
4676
vmcb->control.exit_info_1 = 0;
arch/x86/kvm/svm/svm.c
4713
vmcb->control.exit_info_1 = exit_info;
arch/x86/kvm/svm/svm.c
4714
vmcb->control.exit_info_2 = info->next_rip;
arch/x86/kvm/svm/svm.c
4724
vmcb->control.next_rip = info->next_rip;
arch/x86/kvm/svm/svm.c
4725
vmcb->control.exit_code = icpt_info.exit_code;
arch/x86/kvm/svm/svm.c
4737
switch (to_svm(vcpu)->vmcb->control.exit_code) {
arch/x86/kvm/svm/svm.c
4803
svm->vmcb->save.rax = vcpu->arch.regs[VCPU_REGS_RAX];
arch/x86/kvm/svm/svm.c
4804
svm->vmcb->save.rsp = vcpu->arch.regs[VCPU_REGS_RSP];
arch/x86/kvm/svm/svm.c
4805
svm->vmcb->save.rip = vcpu->arch.regs[VCPU_REGS_RIP];
arch/x86/kvm/svm/svm.c
4826
BUILD_BUG_ON(offsetof(struct vmcb, save) != 0x400);
arch/x86/kvm/svm/svm.c
4839
struct vmcb *vmcb12;
arch/x86/kvm/svm/svm.c
4923
if ((svm->vmcb->control.exit_int_info & SVM_EXITINTINFO_TYPE_MASK) &&
arch/x86/kvm/svm/svm.c
5023
error_code = svm->vmcb->control.exit_info_1;
arch/x86/kvm/svm/svm.c
619
struct vmcb *vmcb = svm->vmcb01.ptr;
arch/x86/kvm/svm/svm.c
621
vmcb_set_intercept(&vmcb->control, INTERCEPT_DR0_READ);
arch/x86/kvm/svm/svm.c
622
vmcb_set_intercept(&vmcb->control, INTERCEPT_DR1_READ);
arch/x86/kvm/svm/svm.c
623
vmcb_set_intercept(&vmcb->control, INTERCEPT_DR2_READ);
arch/x86/kvm/svm/svm.c
624
vmcb_set_intercept(&vmcb->control, INTERCEPT_DR3_READ);
arch/x86/kvm/svm/svm.c
625
vmcb_set_intercept(&vmcb->control, INTERCEPT_DR4_READ);
arch/x86/kvm/svm/svm.c
626
vmcb_set_intercept(&vmcb->control, INTERCEPT_DR5_READ);
arch/x86/kvm/svm/svm.c
627
vmcb_set_intercept(&vmcb->control, INTERCEPT_DR6_READ);
arch/x86/kvm/svm/svm.c
628
vmcb_set_intercept(&vmcb->control, INTERCEPT_DR0_WRITE);
arch/x86/kvm/svm/svm.c
629
vmcb_set_intercept(&vmcb->control, INTERCEPT_DR1_WRITE);
arch/x86/kvm/svm/svm.c
630
vmcb_set_intercept(&vmcb->control, INTERCEPT_DR2_WRITE);
arch/x86/kvm/svm/svm.c
631
vmcb_set_intercept(&vmcb->control, INTERCEPT_DR3_WRITE);
arch/x86/kvm/svm/svm.c
632
vmcb_set_intercept(&vmcb->control, INTERCEPT_DR4_WRITE);
arch/x86/kvm/svm/svm.c
633
vmcb_set_intercept(&vmcb->control, INTERCEPT_DR5_WRITE);
arch/x86/kvm/svm/svm.c
634
vmcb_set_intercept(&vmcb->control, INTERCEPT_DR6_WRITE);
arch/x86/kvm/svm/svm.c
635
vmcb_set_intercept(&vmcb->control, INTERCEPT_DR7_READ);
arch/x86/kvm/svm/svm.c
636
vmcb_set_intercept(&vmcb->control, INTERCEPT_DR7_WRITE);
arch/x86/kvm/svm/svm.c
643
struct vmcb *vmcb = svm->vmcb01.ptr;
arch/x86/kvm/svm/svm.c
645
vmcb->control.intercepts[INTERCEPT_DR] = 0;
arch/x86/kvm/svm/svm.c
713
bool intercept = !(svm->vmcb->control.virt_ext & LBR_CTL_ENABLE_MASK);
arch/x86/kvm/svm/svm.c
844
void svm_copy_lbrs(struct vmcb *to_vmcb, struct vmcb *from_vmcb)
arch/x86/kvm/svm/svm.c
857
to_svm(vcpu)->vmcb->control.virt_ext |= LBR_CTL_ENABLE_MASK;
arch/x86/kvm/svm/svm.c
869
to_svm(vcpu)->vmcb->control.virt_ext &= ~LBR_CTL_ENABLE_MASK;
arch/x86/kvm/svm/svm.c
875
bool current_enable_lbrv = svm->vmcb->control.virt_ext & LBR_CTL_ENABLE_MASK;
arch/x86/kvm/svm/svm.c
876
bool enable_lbrv = (svm->vmcb->save.dbgctl & DEBUGCTLMSR_LBR) ||
arch/x86/kvm/svm/svm.c
901
svm->vmcb->save.rflags &= ~X86_EFLAGS_TF;
arch/x86/kvm/svm/svm.c
903
svm->vmcb->save.rflags &= ~X86_EFLAGS_RF;
arch/x86/kvm/svm/svm.c
910
struct vmcb_control_area *control = &svm->vmcb->control;
arch/x86/kvm/svm/svm.c
922
vmcb_mark_dirty(svm->vmcb, VMCB_INTERCEPTS);
arch/x86/kvm/svm/svm.c
931
struct vmcb_control_area *control = &svm->vmcb->control;
arch/x86/kvm/svm/svm.c
943
vmcb_mark_dirty(svm->vmcb, VMCB_INTERCEPTS);
arch/x86/kvm/svm/svm.h
136
struct vmcb *ptr;
arch/x86/kvm/svm/svm.h
255
struct vmcb *vmcb;
arch/x86/kvm/svm/svm.h
351
struct vmcb *save_area;
arch/x86/kvm/svm/svm.h
355
struct vmcb **sev_vmcbs;
arch/x86/kvm/svm/svm.h
402
static inline void vmcb_mark_all_dirty(struct vmcb *vmcb)
arch/x86/kvm/svm/svm.h
404
vmcb->control.clean = 0;
arch/x86/kvm/svm/svm.h
407
static inline void vmcb_mark_all_clean(struct vmcb *vmcb)
arch/x86/kvm/svm/svm.h
409
vmcb->control.clean = VMCB_ALL_CLEAN_MASK
arch/x86/kvm/svm/svm.h
413
static inline void vmcb_mark_dirty(struct vmcb *vmcb, int bit)
arch/x86/kvm/svm/svm.h
415
vmcb->control.clean &= ~(1 << bit);
arch/x86/kvm/svm/svm.h
418
static inline bool vmcb_is_dirty(struct vmcb *vmcb, int bit)
arch/x86/kvm/svm/svm.h
420
return !test_bit(bit, (unsigned long *)&vmcb->control.clean);
arch/x86/kvm/svm/svm.h
491
struct vmcb *vmcb = svm->vmcb01.ptr;
arch/x86/kvm/svm/svm.h
494
vmcb_set_intercept(&vmcb->control, INTERCEPT_EXCEPTION_OFFSET + bit);
arch/x86/kvm/svm/svm.h
501
struct vmcb *vmcb = svm->vmcb01.ptr;
arch/x86/kvm/svm/svm.h
504
vmcb_clr_intercept(&vmcb->control, INTERCEPT_EXCEPTION_OFFSET + bit);
arch/x86/kvm/svm/svm.h
511
struct vmcb *vmcb = svm->vmcb01.ptr;
arch/x86/kvm/svm/svm.h
513
vmcb_set_intercept(&vmcb->control, bit);
arch/x86/kvm/svm/svm.h
520
struct vmcb *vmcb = svm->vmcb01.ptr;
arch/x86/kvm/svm/svm.h
522
vmcb_clr_intercept(&vmcb->control, bit);
arch/x86/kvm/svm/svm.h
529
return vmcb_is_intercept(&svm->vmcb->control, bit);
arch/x86/kvm/svm/svm.h
538
static inline struct vmcb *get_vgif_vmcb(struct vcpu_svm *svm)
arch/x86/kvm/svm/svm.h
551
struct vmcb *vmcb = get_vgif_vmcb(svm);
arch/x86/kvm/svm/svm.h
553
if (vmcb)
arch/x86/kvm/svm/svm.h
554
vmcb->control.int_ctl |= V_GIF_MASK;
arch/x86/kvm/svm/svm.h
561
struct vmcb *vmcb = get_vgif_vmcb(svm);
arch/x86/kvm/svm/svm.h
563
if (vmcb)
arch/x86/kvm/svm/svm.h
564
vmcb->control.int_ctl &= ~V_GIF_MASK;
arch/x86/kvm/svm/svm.h
571
struct vmcb *vmcb = get_vgif_vmcb(svm);
arch/x86/kvm/svm/svm.h
573
if (vmcb)
arch/x86/kvm/svm/svm.h
574
return !!(vmcb->control.int_ctl & V_GIF_MASK);
arch/x86/kvm/svm/svm.h
599
static inline struct vmcb *get_vnmi_vmcb_l1(struct vcpu_svm *svm)
arch/x86/kvm/svm/svm.h
612
struct vmcb *vmcb = get_vnmi_vmcb_l1(svm);
arch/x86/kvm/svm/svm.h
614
if (vmcb)
arch/x86/kvm/svm/svm.h
615
return !!(vmcb->control.int_ctl & V_NMI_ENABLE_MASK);
arch/x86/kvm/svm/svm.h
717
void svm_copy_lbrs(struct vmcb *to_vmcb, struct vmcb *from_vmcb);
arch/x86/kvm/svm/svm.h
780
u64 vmcb_gpa, struct vmcb *vmcb12, bool from_vmrun);
arch/x86/kvm/svm/svm.h
787
void svm_copy_vmloadsave_state(struct vmcb *to_vmcb, struct vmcb *from_vmcb);
arch/x86/kvm/svm/svm.h
792
svm->vmcb->control.exit_code = exit_code;
arch/x86/kvm/svm/svm.h
793
svm->vmcb->control.exit_info_1 = 0;
arch/x86/kvm/svm/svm.h
794
svm->vmcb->control.exit_info_2 = 0;
arch/x86/kvm/svm/svm.h
839
void avic_init_vmcb(struct vcpu_svm *svm, struct vmcb *vmcb);
arch/x86/kvm/svm/svm_onhyperv.c
26
hve = &to_svm(vcpu)->vmcb->control.hv_enlightenments;
arch/x86/kvm/svm/svm_onhyperv.c
32
vmcb_mark_dirty(to_svm(vcpu)->vmcb, HV_VMCB_NESTED_ENLIGHTENMENTS);
arch/x86/kvm/svm/svm_onhyperv.h
20
struct hv_vmcb_enlightenments *hve = &to_svm(vcpu)->vmcb->control.hv_enlightenments;
arch/x86/kvm/svm/svm_onhyperv.h
26
static inline void svm_hv_init_vmcb(struct vmcb *vmcb)
arch/x86/kvm/svm/svm_onhyperv.h
28
struct hv_vmcb_enlightenments *hve = &vmcb->control.hv_enlightenments;
arch/x86/kvm/svm/svm_onhyperv.h
30
BUILD_BUG_ON(sizeof(vmcb->control.hv_enlightenments) !=
arch/x86/kvm/svm/svm_onhyperv.h
31
sizeof(vmcb->control.reserved_sw));
arch/x86/kvm/svm/svm_onhyperv.h
44
struct vmcb *vmcb = to_svm(vcpu)->vmcb;
arch/x86/kvm/svm/svm_onhyperv.h
45
struct hv_vmcb_enlightenments *hve = &vmcb->control.hv_enlightenments;
arch/x86/kvm/svm/svm_onhyperv.h
48
vmcb_mark_dirty(vmcb, HV_VMCB_NESTED_ENLIGHTENMENTS);
arch/x86/kvm/svm/svm_onhyperv.h
51
static inline void svm_hv_update_vp_id(struct vmcb *vmcb, struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm_onhyperv.h
53
struct hv_vmcb_enlightenments *hve = &vmcb->control.hv_enlightenments;
arch/x86/kvm/svm/svm_onhyperv.h
58
vmcb_mark_dirty(vmcb, HV_VMCB_NESTED_ENLIGHTENMENTS);
arch/x86/kvm/svm/svm_onhyperv.h
68
static inline void svm_hv_init_vmcb(struct vmcb *vmcb)
arch/x86/kvm/svm/svm_onhyperv.h
81
static inline void svm_hv_update_vp_id(struct vmcb *vmcb,
arch/x86/kvm/trace.h
703
TP_PROTO(__u64 rip, __u64 vmcb, __u64 nested_rip, __u32 int_ctl,
arch/x86/kvm/trace.h
706
TP_ARGS(rip, vmcb, nested_rip, int_ctl, event_inj, tdp_enabled,
arch/x86/kvm/trace.h
711
__field( __u64, vmcb )
arch/x86/kvm/trace.h
722
__entry->vmcb = vmcb;
arch/x86/kvm/trace.h
735
__entry->vmcb,
tools/testing/selftests/kvm/include/x86/svm_util.h
17
struct vmcb *vmcb; /* gva */
tools/testing/selftests/kvm/include/x86/svm_util.h
61
void run_guest(struct vmcb *vmcb, uint64_t vmcb_gpa);
tools/testing/selftests/kvm/lib/x86/memstress.c
64
run_guest(svm->vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/lib/x86/memstress.c
65
GUEST_ASSERT_EQ(svm->vmcb->control.exit_code, SVM_EXIT_VMMCALL);
tools/testing/selftests/kvm/lib/x86/svm.c
100
memset(vmcb, 0, sizeof(*vmcb));
tools/testing/selftests/kvm/lib/x86/svm.c
124
vmcb->save.rip = (u64)guest_rip;
tools/testing/selftests/kvm/lib/x86/svm.c
125
vmcb->save.rsp = (u64)guest_rsp;
tools/testing/selftests/kvm/lib/x86/svm.c
161
void run_guest(struct vmcb *vmcb, uint64_t vmcb_gpa)
tools/testing/selftests/kvm/lib/x86/svm.c
177
: : [vmcb] "r" (vmcb), [vmcb_gpa] "a" (vmcb_gpa)
tools/testing/selftests/kvm/lib/x86/svm.c
36
svm->vmcb = (void *)vm_vaddr_alloc_page(vm);
tools/testing/selftests/kvm/lib/x86/svm.c
37
svm->vmcb_hva = addr_gva2hva(vm, (uintptr_t)svm->vmcb);
tools/testing/selftests/kvm/lib/x86/svm.c
38
svm->vmcb_gpa = addr_gva2gpa(vm, (uintptr_t)svm->vmcb);
tools/testing/selftests/kvm/lib/x86/svm.c
86
struct vmcb *vmcb = svm->vmcb;
tools/testing/selftests/kvm/lib/x86/svm.c
88
struct vmcb_save_area *save = &vmcb->save;
tools/testing/selftests/kvm/lib/x86/svm.c
89
struct vmcb_control_area *ctrl = &vmcb->control;
tools/testing/selftests/kvm/x86/aperfmperf_test.c
68
struct vmcb *vmcb = svm->vmcb;
tools/testing/selftests/kvm/x86/aperfmperf_test.c
71
run_guest(vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
102
vmcb->control.intercept |= 1ULL << INTERCEPT_MSR_PROT;
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
104
run_guest(vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
105
GUEST_ASSERT(vmcb->control.exit_code == SVM_EXIT_MSR);
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
106
vmcb->save.rip += 2; /* rdmsr */
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
110
run_guest(vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
111
GUEST_ASSERT(vmcb->control.exit_code == SVM_EXIT_MSR);
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
112
vmcb->save.rip += 2; /* rdmsr */
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
117
vmcb->control.clean |= HV_VMCB_NESTED_ENLIGHTENMENTS;
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
118
run_guest(vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
120
GUEST_ASSERT(vmcb->control.exit_code == SVM_EXIT_VMMCALL);
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
121
vmcb->save.rip += 3; /* vmcall */
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
124
vmcb->control.clean &= ~HV_VMCB_NESTED_ENLIGHTENMENTS;
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
125
run_guest(vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
126
GUEST_ASSERT(vmcb->control.exit_code == SVM_EXIT_MSR);
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
127
vmcb->save.rip += 2; /* rdmsr */
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
134
run_guest(vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
135
GUEST_ASSERT(vmcb->control.exit_code == SVM_EXIT_MSR);
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
136
vmcb->save.rip += 2; /* rdmsr */
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
139
run_guest(vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
140
GUEST_ASSERT(vmcb->control.exit_code == HV_SVM_EXITCODE_ENL);
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
141
GUEST_ASSERT(vmcb->control.exit_info_1 == HV_SVM_ENL_EXITCODE_TRAP_AFTER_FLUSH);
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
143
run_guest(vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
144
GUEST_ASSERT(vmcb->control.exit_code == SVM_EXIT_VMMCALL);
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
73
struct vmcb *vmcb = svm->vmcb;
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
74
struct hv_vmcb_enlightenments *hve = &vmcb->control.hv_enlightenments;
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
96
run_guest(vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
97
GUEST_ASSERT(vmcb->control.exit_code == SVM_EXIT_VMMCALL);
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
99
vmcb->save.rip += 3;
tools/testing/selftests/kvm/x86/kvm_buslock_test.c
40
struct vmcb *vmcb = svm->vmcb;
tools/testing/selftests/kvm/x86/kvm_buslock_test.c
43
run_guest(vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/nested_close_kvm_test.c
56
run_guest(svm->vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/nested_dirty_log_test.c
129
run_guest(svm->vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/nested_dirty_log_test.c
131
GUEST_ASSERT_EQ(svm->vmcb->control.exit_code, SVM_EXIT_VMMCALL);
tools/testing/selftests/kvm/x86/nested_emulation_test.c
100
run_guest(vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/nested_emulation_test.c
101
exit_reason = vmcb->control.exit_code;
tools/testing/selftests/kvm/x86/nested_emulation_test.c
102
exit_insn_len = vmcb->control.next_rip - vmcb->save.rip;
tools/testing/selftests/kvm/x86/nested_emulation_test.c
103
GUEST_ASSERT_EQ(vmcb->save.rip, (u64)l2_instruction);
tools/testing/selftests/kvm/x86/nested_emulation_test.c
58
struct vmcb *vmcb = svm->vmcb;
tools/testing/selftests/kvm/x86/nested_emulation_test.c
61
vmcb->save.idtr.limit = 0;
tools/testing/selftests/kvm/x86/nested_emulation_test.c
62
vmcb->save.rip = (u64)l2_guest_code;
tools/testing/selftests/kvm/x86/nested_emulation_test.c
64
vmcb->control.intercept |= BIT_ULL(INTERCEPT_SHUTDOWN) |
tools/testing/selftests/kvm/x86/nested_emulation_test.c
67
vmcb->control.intercept_exceptions = 0;
tools/testing/selftests/kvm/x86/nested_emulation_test.c
98
struct vmcb *vmcb = svm->vmcb;
tools/testing/selftests/kvm/x86/nested_exceptions_test.c
77
struct vmcb *vmcb = svm->vmcb;
tools/testing/selftests/kvm/x86/nested_exceptions_test.c
78
struct vmcb_control_area *ctrl = &vmcb->control;
tools/testing/selftests/kvm/x86/nested_exceptions_test.c
80
vmcb->save.rip = (u64)l2_code;
tools/testing/selftests/kvm/x86/nested_exceptions_test.c
81
run_guest(vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/nested_exceptions_test.c
93
struct vmcb_control_area *ctrl = &svm->vmcb->control;
tools/testing/selftests/kvm/x86/nested_exceptions_test.c
97
svm->vmcb->save.idtr.limit = 0;
tools/testing/selftests/kvm/x86/nested_invalid_cr3_test.c
30
save_cr3 = svm->vmcb->save.cr3;
tools/testing/selftests/kvm/x86/nested_invalid_cr3_test.c
31
svm->vmcb->save.cr3 = -1ull;
tools/testing/selftests/kvm/x86/nested_invalid_cr3_test.c
32
run_guest(svm->vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/nested_invalid_cr3_test.c
33
GUEST_ASSERT(svm->vmcb->control.exit_code == SVM_EXIT_ERR);
tools/testing/selftests/kvm/x86/nested_invalid_cr3_test.c
36
svm->vmcb->save.cr3 = save_cr3;
tools/testing/selftests/kvm/x86/nested_invalid_cr3_test.c
37
run_guest(svm->vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/nested_invalid_cr3_test.c
38
GUEST_ASSERT(svm->vmcb->control.exit_code == SVM_EXIT_VMMCALL);
tools/testing/selftests/kvm/x86/nested_tsc_adjust_test.c
111
svm->vmcb->control.tsc_offset = TSC_OFFSET_VALUE;
tools/testing/selftests/kvm/x86/nested_tsc_adjust_test.c
112
run_guest(svm->vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/nested_tsc_adjust_test.c
113
GUEST_ASSERT(svm->vmcb->control.exit_code == SVM_EXIT_VMMCALL);
tools/testing/selftests/kvm/x86/nested_tsc_scaling_test.c
97
run_guest(svm->vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/nested_tsc_scaling_test.c
98
GUEST_ASSERT(svm->vmcb->control.exit_code == SVM_EXIT_VMMCALL);
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
100
svm->vmcb->save.rip = (u64)l2_guest_code_vmload;
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
101
run_guest(svm->vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
102
GUEST_ASSERT_EQ(svm->vmcb->control.exit_code, SVM_EXIT_VMLOAD);
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
105
svm->vmcb->control.intercept &= ~(BIT_ULL(INTERCEPT_VMSAVE) |
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
112
svm->vmcb->save.rip = (u64)l2_guest_code_vmcb0;
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
113
svm->vmcb->control.virt_ext &= ~VIRTUAL_VMLOAD_VMSAVE_ENABLE_MASK;
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
114
run_guest(svm->vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
115
GUEST_ASSERT_EQ(svm->vmcb->control.exit_code, SVM_EXIT_VMMCALL);
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
121
svm->vmcb->save.rip = (u64)l2_guest_code_vmcb1;
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
122
svm->vmcb->control.virt_ext |= VIRTUAL_VMLOAD_VMSAVE_ENABLE_MASK;
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
123
run_guest(svm->vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
124
GUEST_ASSERT_EQ(svm->vmcb->control.exit_code, SVM_EXIT_VMMCALL);
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
132
struct vmcb *test_vmcb[2];
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
153
test_vmcb[i] = (struct vmcb *)addr_gva2hva(vm, TEST_VMCB_GVA(i));
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
79
svm->vmcb->control.intercept |= (BIT_ULL(INTERCEPT_VMSAVE) |
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
83
svm->vmcb->control.virt_ext &= ~VIRTUAL_VMLOAD_VMSAVE_ENABLE_MASK;
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
85
svm->vmcb->save.rip = (u64)l2_guest_code_vmsave;
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
86
run_guest(svm->vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
87
GUEST_ASSERT_EQ(svm->vmcb->control.exit_code, SVM_EXIT_VMSAVE);
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
89
svm->vmcb->save.rip = (u64)l2_guest_code_vmload;
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
90
run_guest(svm->vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
91
GUEST_ASSERT_EQ(svm->vmcb->control.exit_code, SVM_EXIT_VMLOAD);
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
94
svm->vmcb->control.virt_ext |= VIRTUAL_VMLOAD_VMSAVE_ENABLE_MASK;
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
96
svm->vmcb->save.rip = (u64)l2_guest_code_vmsave;
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
97
run_guest(svm->vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
98
GUEST_ASSERT_EQ(svm->vmcb->control.exit_code, SVM_EXIT_VMSAVE);
tools/testing/selftests/kvm/x86/smm_test.c
100
run_guest(svm->vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/smm_test.c
101
run_guest(svm->vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/state_test.c
37
struct vmcb *vmcb = svm->vmcb;
tools/testing/selftests/kvm/x86/state_test.c
45
run_guest(vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/state_test.c
46
GUEST_ASSERT(vmcb->control.exit_code == SVM_EXIT_VMMCALL);
tools/testing/selftests/kvm/x86/state_test.c
48
vmcb->save.rip += 3;
tools/testing/selftests/kvm/x86/state_test.c
49
run_guest(vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/state_test.c
50
GUEST_ASSERT(vmcb->control.exit_code == SVM_EXIT_VMMCALL);
tools/testing/selftests/kvm/x86/svm_int_ctl_test.c
59
struct vmcb *vmcb = svm->vmcb;
tools/testing/selftests/kvm/x86/svm_int_ctl_test.c
68
vmcb->control.int_ctl &= ~V_INTR_MASKING_MASK;
tools/testing/selftests/kvm/x86/svm_int_ctl_test.c
71
vmcb->control.intercept &= ~(BIT(INTERCEPT_INTR) | BIT(INTERCEPT_VINTR));
tools/testing/selftests/kvm/x86/svm_int_ctl_test.c
74
vmcb->control.int_ctl |= V_IRQ_MASK | (0x1 << V_INTR_PRIO_SHIFT);
tools/testing/selftests/kvm/x86/svm_int_ctl_test.c
75
vmcb->control.int_vector = VINTR_IRQ_NUMBER;
tools/testing/selftests/kvm/x86/svm_int_ctl_test.c
77
run_guest(vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/svm_int_ctl_test.c
78
GUEST_ASSERT(vmcb->control.exit_code == SVM_EXIT_VMMCALL);
tools/testing/selftests/kvm/x86/svm_nested_shutdown_test.c
24
struct vmcb *vmcb = svm->vmcb;
tools/testing/selftests/kvm/x86/svm_nested_shutdown_test.c
29
vmcb->control.intercept &= ~(BIT(INTERCEPT_SHUTDOWN));
tools/testing/selftests/kvm/x86/svm_nested_shutdown_test.c
36
run_guest(vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/svm_nested_soft_inject_test.c
101
vmcb->control.next_rip = vmcb->save.rip;
tools/testing/selftests/kvm/x86/svm_nested_soft_inject_test.c
104
run_guest(vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/svm_nested_soft_inject_test.c
105
__GUEST_ASSERT(vmcb->control.exit_code == SVM_EXIT_VMMCALL,
tools/testing/selftests/kvm/x86/svm_nested_soft_inject_test.c
107
vmcb->control.exit_code,
tools/testing/selftests/kvm/x86/svm_nested_soft_inject_test.c
108
vmcb->control.exit_info_1, vmcb->control.exit_info_2);
tools/testing/selftests/kvm/x86/svm_nested_soft_inject_test.c
124
vmcb->save.rip += 3;
tools/testing/selftests/kvm/x86/svm_nested_soft_inject_test.c
127
vmcb->save.idtr.base = idt_alt;
tools/testing/selftests/kvm/x86/svm_nested_soft_inject_test.c
128
vmcb->control.clean = 0; /* &= ~BIT(VMCB_DT) would be enough */
tools/testing/selftests/kvm/x86/svm_nested_soft_inject_test.c
130
vmcb->control.event_inj = BP_VECTOR | SVM_EVTINJ_VALID | SVM_EVTINJ_TYPE_EXEPT;
tools/testing/selftests/kvm/x86/svm_nested_soft_inject_test.c
132
vmcb->control.next_rip = vmcb->save.rip + 2;
tools/testing/selftests/kvm/x86/svm_nested_soft_inject_test.c
134
run_guest(vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/svm_nested_soft_inject_test.c
135
__GUEST_ASSERT(vmcb->control.exit_code == SVM_EXIT_HLT,
tools/testing/selftests/kvm/x86/svm_nested_soft_inject_test.c
137
vmcb->control.exit_code,
tools/testing/selftests/kvm/x86/svm_nested_soft_inject_test.c
138
vmcb->control.exit_info_1, vmcb->control.exit_info_2);
tools/testing/selftests/kvm/x86/svm_nested_soft_inject_test.c
83
struct vmcb *vmcb = svm->vmcb;
tools/testing/selftests/kvm/x86/svm_nested_soft_inject_test.c
93
vmcb->control.intercept_exceptions |= BIT(PF_VECTOR) | BIT(UD_VECTOR);
tools/testing/selftests/kvm/x86/svm_nested_soft_inject_test.c
94
vmcb->control.intercept |= BIT(INTERCEPT_NMI) | BIT(INTERCEPT_HLT);
tools/testing/selftests/kvm/x86/svm_nested_soft_inject_test.c
97
vmcb->control.event_inj = SVM_EVTINJ_VALID | SVM_EVTINJ_TYPE_NMI;
tools/testing/selftests/kvm/x86/svm_nested_soft_inject_test.c
99
vmcb->control.event_inj = INT_NR | SVM_EVTINJ_VALID | SVM_EVTINJ_TYPE_SOFT;
tools/testing/selftests/kvm/x86/svm_vmcall_test.c
24
struct vmcb *vmcb = svm->vmcb;
tools/testing/selftests/kvm/x86/svm_vmcall_test.c
30
run_guest(vmcb, svm->vmcb_gpa);
tools/testing/selftests/kvm/x86/svm_vmcall_test.c
32
GUEST_ASSERT(vmcb->control.exit_code == SVM_EXIT_VMMCALL);
tools/testing/selftests/kvm/x86/triple_fault_event_test.c
45
struct vmcb *vmcb = svm->vmcb;
tools/testing/selftests/kvm/x86/triple_fault_event_test.c
51
vmcb->control.intercept &= ~(BIT(INTERCEPT_SHUTDOWN));
tools/testing/selftests/kvm/x86/triple_fault_event_test.c
53
run_guest(vmcb, svm->vmcb_gpa);