Symbol: is_guest_mode
arch/x86/include/asm/irq_remapping.h
40
bool is_guest_mode;
arch/x86/kvm/hyperv.c
1976
tlb_flush_fifo = kvm_hv_get_tlb_flush_fifo(vcpu, is_guest_mode(vcpu));
arch/x86/kvm/hyperv.c
2043
if (!hc->fast && is_guest_mode(vcpu)) {
arch/x86/kvm/hyperv.c
2065
is_guest_mode(vcpu));
arch/x86/kvm/hyperv.c
2096
flush_ex.flags, is_guest_mode(vcpu));
arch/x86/kvm/hyperv.c
2140
if (all_cpus && !is_guest_mode(vcpu)) {
arch/x86/kvm/hyperv.c
2148
} else if (!is_guest_mode(vcpu)) {
arch/x86/kvm/hyperv.c
2390
if (hv_result_success(result) && is_guest_mode(vcpu) &&
arch/x86/kvm/hyperv.h
198
bool is_guest_mode)
arch/x86/kvm/hyperv.h
201
int i = is_guest_mode ? HV_L2_TLB_FLUSH_FIFO :
arch/x86/kvm/hyperv.h
214
tlb_flush_fifo = kvm_hv_get_tlb_flush_fifo(vcpu, is_guest_mode(vcpu));
arch/x86/kvm/irq.c
101
if (!is_guest_mode(v) && kvm_vcpu_apicv_active(v))
arch/x86/kvm/lapic.c
3533
if (is_guest_mode(vcpu)) {
arch/x86/kvm/mmu/mmu.c
4699
if (is_guest_mode(vcpu))
arch/x86/kvm/mmu/mmu.c
5705
role.base.guest_mode = is_guest_mode(vcpu);
arch/x86/kvm/mmu/mmu.c
6392
if (!is_guest_mode(vcpu))
arch/x86/kvm/smm.c
660
if (ret != X86EMUL_CONTINUE && is_guest_mode(vcpu))
arch/x86/kvm/svm/avic.c
236
if (is_guest_mode(&svm->vcpu) &&
arch/x86/kvm/svm/avic.c
663
if (is_guest_mode(vcpu))
arch/x86/kvm/svm/avic.c
894
.is_guest_mode = kvm_vcpu_apicv_active(vcpu),
arch/x86/kvm/svm/nested.c
133
if (!is_guest_mode(&svm->vcpu))
arch/x86/kvm/svm/nested.c
1409
if (is_guest_mode(vcpu)) {
arch/x86/kvm/svm/nested.c
1767
if (is_guest_mode(vcpu)) {
arch/x86/kvm/svm/nested.c
1782
if (!is_guest_mode(vcpu))
arch/x86/kvm/svm/nested.c
1902
if (is_guest_mode(vcpu))
arch/x86/kvm/svm/nested.c
1944
if (WARN_ON(!is_guest_mode(vcpu)))
arch/x86/kvm/svm/svm.c
1536
if (is_guest_mode(&svm->vcpu)) {
arch/x86/kvm/svm/svm.c
1940
if ((error_code & PFERR_RSVD_MASK) && !is_guest_mode(vcpu)) {
arch/x86/kvm/svm/svm.c
2249
if (is_guest_mode(vcpu)) {
arch/x86/kvm/svm/svm.c
2291
if (!is_guest_mode(vcpu))
arch/x86/kvm/svm/svm.c
2509
if (!is_guest_mode(vcpu) ||
arch/x86/kvm/svm/svm.c
2934
is_guest_mode(vcpu))
arch/x86/kvm/svm/svm.c
2944
if (is_guest_mode(vcpu))
arch/x86/kvm/svm/svm.c
3226
if (is_guest_mode(vcpu))
arch/x86/kvm/svm/svm.c
3612
if (is_guest_mode(vcpu)) {
arch/x86/kvm/svm/svm.c
3852
if (is_guest_mode(vcpu) && nested_exit_on_nmi(svm))
arch/x86/kvm/svm/svm.c
3871
if (for_injection && is_guest_mode(vcpu) && nested_exit_on_nmi(svm))
arch/x86/kvm/svm/svm.c
3884
if (is_guest_mode(vcpu)) {
arch/x86/kvm/svm/svm.c
3916
if (for_injection && is_guest_mode(vcpu) && nested_exit_on_intr(svm))
arch/x86/kvm/svm/svm.c
3945
if (!is_guest_mode(vcpu))
arch/x86/kvm/svm/svm.c
4237
if (is_guest_mode(vcpu))
arch/x86/kvm/svm/svm.c
4403
if (is_guest_mode(vcpu)) {
arch/x86/kvm/svm/svm.c
4777
if (for_injection && is_guest_mode(vcpu) && nested_exit_on_smi(svm))
arch/x86/kvm/svm/svm.c
4789
if (!is_guest_mode(vcpu))
arch/x86/kvm/svm/svm.c
661
void *msrpm = is_guest_mode(vcpu) ? to_svm(vcpu)->nested.msrpm :
arch/x86/kvm/svm/svm.c
877
(is_guest_mode(vcpu) && guest_cpu_cap_has(vcpu, X86_FEATURE_LBRV) &&
arch/x86/kvm/svm/svm.h
543
if (is_guest_mode(&svm->vcpu) && !nested_vgif_enabled(svm))
arch/x86/kvm/svm/svm.h
604
if (is_guest_mode(&svm->vcpu))
arch/x86/kvm/svm/svm.h
759
return is_guest_mode(vcpu) && (svm->nested.ctl.int_ctl & V_INTR_MASKING_MASK);
arch/x86/kvm/vmx/nested.c
3528
if (is_guest_mode(vcpu) && !nested_get_vmcs12_pages(vcpu))
arch/x86/kvm/vmx/nested.c
3540
if (WARN_ON_ONCE(!is_guest_mode(vcpu)))
arch/x86/kvm/vmx/nested.c
5633
struct vmcs12 *vmcs12 = is_guest_mode(vcpu) ? get_shadow_vmcs12(vcpu)
arch/x86/kvm/vmx/nested.c
5657
(is_guest_mode(vcpu) &&
arch/x86/kvm/vmx/nested.c
5665
if (!is_guest_mode(vcpu) && is_vmcs12_ext_field(field))
arch/x86/kvm/vmx/nested.c
5681
if (WARN_ON_ONCE(is_guest_mode(vcpu)))
arch/x86/kvm/vmx/nested.c
5739
struct vmcs12 *vmcs12 = is_guest_mode(vcpu) ? get_shadow_vmcs12(vcpu)
arch/x86/kvm/vmx/nested.c
5767
(is_guest_mode(vcpu) &&
arch/x86/kvm/vmx/nested.c
5801
if (!is_guest_mode(vcpu) && !is_shadow_field_rw(field))
arch/x86/kvm/vmx/nested.c
5823
if (!is_guest_mode(vcpu) && !is_shadow_field_rw(field)) {
arch/x86/kvm/vmx/nested.c
6178
if (WARN_ON_ONCE(!is_guest_mode(vcpu))) {
arch/x86/kvm/vmx/nested.c
6763
if (is_guest_mode(vcpu) &&
arch/x86/kvm/vmx/nested.c
6775
if (is_guest_mode(vcpu)) {
arch/x86/kvm/vmx/nested.c
6810
if (is_guest_mode(vcpu)) {
arch/x86/kvm/vmx/nested.c
6852
if (is_guest_mode(vcpu)) {
arch/x86/kvm/vmx/sgx.c
504
if (!vmcs12 && is_guest_mode(vcpu))
arch/x86/kvm/vmx/vmx.c
1619
if (!is_guest_mode(vcpu)) {
arch/x86/kvm/vmx/vmx.c
1630
if (!is_guest_mode(vcpu))
arch/x86/kvm/vmx/vmx.c
1893
if (!is_guest_mode(vcpu))
arch/x86/kvm/vmx/vmx.c
2375
if (is_guest_mode(vcpu))
arch/x86/kvm/vmx/vmx.c
2380
if (is_guest_mode(vcpu)) {
arch/x86/kvm/vmx/vmx.c
2387
if (is_guest_mode(vcpu)) {
arch/x86/kvm/vmx/vmx.c
2399
if (is_guest_mode(vcpu) && get_vmcs12(vcpu)->vm_exit_controls &
arch/x86/kvm/vmx/vmx.c
2418
if (is_guest_mode(vcpu) &&
arch/x86/kvm/vmx/vmx.c
2475
if (is_guest_mode(vcpu) &&
arch/x86/kvm/vmx/vmx.c
3274
WARN_ON_ONCE(is_guest_mode(vcpu));
arch/x86/kvm/vmx/vmx.c
3383
if (is_guest_mode(vcpu) && nested_cpu_has_vpid(get_vmcs12(vcpu)))
arch/x86/kvm/vmx/vmx.c
3493
if (is_guest_mode(vcpu))
arch/x86/kvm/vmx/vmx.c
3566
} else if (!is_guest_mode(vcpu)) {
arch/x86/kvm/vmx/vmx.c
3666
} else if (!is_guest_mode(vcpu) ||
arch/x86/kvm/vmx/vmx.c
4441
if (is_guest_mode(vcpu) &&
arch/x86/kvm/vmx/vmx.c
4615
if (is_guest_mode(&vmx->vcpu))
arch/x86/kvm/vmx/vmx.c
5269
if (is_guest_mode(vcpu) && nested_exit_on_nmi(vcpu))
arch/x86/kvm/vmx/vmx.c
5286
if (for_injection && is_guest_mode(vcpu) && nested_exit_on_nmi(vcpu))
arch/x86/kvm/vmx/vmx.c
5301
if (is_guest_mode(vcpu) && nested_exit_on_intr(vcpu))
arch/x86/kvm/vmx/vmx.c
5316
if (for_injection && is_guest_mode(vcpu) && nested_exit_on_intr(vcpu))
arch/x86/kvm/vmx/vmx.c
5688
if (is_guest_mode(vcpu)) {
arch/x86/kvm/vmx/vmx.c
5714
if (is_guest_mode(vcpu)) {
arch/x86/kvm/vmx/vmx.c
6080
if (!is_guest_mode(vcpu) &&
arch/x86/kvm/vmx/vmx.c
6295
if (is_guest_mode(vcpu))
arch/x86/kvm/vmx/vmx.c
6309
WARN_ON_ONCE(!is_guest_mode(vcpu));
arch/x86/kvm/vmx/vmx.c
6796
if (enable_pml && !is_guest_mode(vcpu))
arch/x86/kvm/vmx/vmx.c
6808
if (is_guest_mode(vcpu)) {
arch/x86/kvm/vmx/vmx.c
6960
if (is_guest_mode(vcpu) &&
arch/x86/kvm/vmx/vmx.c
7010
if (!is_guest_mode(vcpu))
arch/x86/kvm/vmx/vmx.c
7176
if (!is_guest_mode(vcpu) && kvm_vcpu_apicv_active(vcpu))
arch/x86/kvm/vmx/vmx.c
7531
if (is_guest_mode(vcpu) &&
arch/x86/kvm/vmx/vmx.c
7728
if (is_guest_mode(vcpu)) {
arch/x86/kvm/vmx/vmx.c
8510
vmx->nested.smm.guest_mode = is_guest_mode(vcpu);
arch/x86/kvm/vmx/vmx.c
8553
return to_vmx(vcpu)->nested.vmxon && !is_guest_mode(vcpu);
arch/x86/kvm/vmx/vmx.c
8558
if (is_guest_mode(vcpu)) {
arch/x86/kvm/vmx/vmx.c
961
if (is_guest_mode(vcpu))
arch/x86/kvm/vmx/vmx.h
697
return enable_unrestricted_guest && (!is_guest_mode(vcpu) ||
arch/x86/kvm/x86.c
10542
if (is_guest_mode(vcpu))
arch/x86/kvm/x86.c
10652
if (is_guest_mode(vcpu))
arch/x86/kvm/x86.c
10820
if (is_guest_mode(vcpu) &&
arch/x86/kvm/x86.c
11034
if (is_guest_mode(vcpu))
arch/x86/kvm/x86.c
11153
if (is_guest_mode(vcpu))
arch/x86/kvm/x86.c
11575
if (is_guest_mode(vcpu) &&
arch/x86/kvm/x86.c
11635
if (is_guest_mode(vcpu)) {
arch/x86/kvm/x86.c
11994
if (vcpu->arch.exception_from_userspace && is_guest_mode(vcpu) &&
arch/x86/kvm/x86.c
12947
if (is_guest_mode(vcpu))
arch/x86/kvm/x86.c
12952
WARN_ON_ONCE(is_guest_mode(vcpu) || is_smm(vcpu));
arch/x86/kvm/x86.c
13911
if (is_guest_mode(vcpu)) {
arch/x86/kvm/x86.c
2724
if (is_guest_mode(vcpu))
arch/x86/kvm/x86.c
2740
if (is_guest_mode(vcpu))
arch/x86/kvm/x86.c
850
if (is_guest_mode(vcpu) &&
arch/x86/kvm/x86.c
867
if (!is_guest_mode(vcpu))
arch/x86/kvm/x86.c
9164
if (!is_guest_mode(vcpu) && kvm_x86_call(get_cpl)(vcpu) == 0) {
arch/x86/kvm/x86.c
9424
(WARN_ON_ONCE(is_guest_mode(vcpu)) ||
arch/x86/kvm/x86.c
9558
r = x86_emulate_insn(ctxt, is_guest_mode(vcpu) &&
arch/x86/kvm/x86.c
979
if (is_guest_mode(vcpu) && fault->async_page_fault)
arch/x86/kvm/x86.h
188
return vcpu->arch.last_vmentry_cpu == -1 && !is_guest_mode(vcpu);
drivers/iommu/amd/iommu.c
4081
if (pi_data->is_guest_mode)