Symbol: loaded_vmcs
arch/x86/kvm/vmx/nested.c
1650
vmcs_load(vmx->loaded_vmcs->vmcs);
arch/x86/kvm/vmx/nested.c
1686
vmcs_load(vmx->loaded_vmcs->vmcs);
arch/x86/kvm/vmx/nested.c
2366
static void prepare_vmcs02_early(struct vcpu_vmx *vmx, struct loaded_vmcs *vmcs01,
arch/x86/kvm/vmx/nested.c
2525
vmx->loaded_vmcs->nmi_known_unmasked =
arch/x86/kvm/vmx/nested.c
279
struct loaded_vmcs *prev)
arch/x86/kvm/vmx/nested.c
287
dest = &vmx->loaded_vmcs->host_state;
arch/x86/kvm/vmx/nested.c
297
static void vmx_switch_vmcs(struct kvm_vcpu *vcpu, struct loaded_vmcs *vmcs)
arch/x86/kvm/vmx/nested.c
300
struct loaded_vmcs *prev;
arch/x86/kvm/vmx/nested.c
303
if (WARN_ON_ONCE(vmx->loaded_vmcs == vmcs))
arch/x86/kvm/vmx/nested.c
307
prev = vmx->loaded_vmcs;
arch/x86/kvm/vmx/nested.c
308
vmx->loaded_vmcs = vmcs;
arch/x86/kvm/vmx/nested.c
340
if (WARN_ON_ONCE(vmx->loaded_vmcs != &vmx->vmcs01))
arch/x86/kvm/vmx/nested.c
4609
WARN_ON_ONCE(vmx->loaded_vmcs != &vmx->vmcs01);
arch/x86/kvm/vmx/nested.c
4612
vmx->loaded_vmcs = &vmx->nested.vmcs02;
arch/x86/kvm/vmx/nested.c
4617
vmx->loaded_vmcs = &vmx->vmcs01;
arch/x86/kvm/vmx/nested.c
5374
struct loaded_vmcs *loaded_vmcs = vmx->loaded_vmcs;
arch/x86/kvm/vmx/nested.c
5384
if (WARN_ON(loaded_vmcs != &vmx->vmcs01 || loaded_vmcs->shadow_vmcs))
arch/x86/kvm/vmx/nested.c
5385
return loaded_vmcs->shadow_vmcs;
arch/x86/kvm/vmx/nested.c
5387
loaded_vmcs->shadow_vmcs = alloc_vmcs(true);
arch/x86/kvm/vmx/nested.c
5388
if (loaded_vmcs->shadow_vmcs)
arch/x86/kvm/vmx/nested.c
5389
vmcs_clear(loaded_vmcs->shadow_vmcs);
arch/x86/kvm/vmx/nested.c
5391
return loaded_vmcs->shadow_vmcs;
arch/x86/kvm/vmx/nested.c
5835
vmcs_load(vmx->loaded_vmcs->vmcs);
arch/x86/kvm/vmx/vmx.c
1000
return vmx_test_msr_bitmap_write(vmx->loaded_vmcs->msr_bitmap, msr);
arch/x86/kvm/vmx/vmx.c
1007
if (vmx->loaded_vmcs->launched)
arch/x86/kvm/vmx/vmx.c
1384
host_state = &vmx->loaded_vmcs->host_state;
arch/x86/kvm/vmx/vmx.c
1429
host_state = &vmx->loaded_vmcs->host_state;
arch/x86/kvm/vmx/vmx.c
1529
bool already_loaded = vmx->loaded_vmcs->cpu == cpu;
arch/x86/kvm/vmx/vmx.c
1533
loaded_vmcs_clear(vmx->loaded_vmcs);
arch/x86/kvm/vmx/vmx.c
1544
list_add(&vmx->loaded_vmcs->loaded_vmcss_on_cpu_link,
arch/x86/kvm/vmx/vmx.c
1550
if (prev != vmx->loaded_vmcs->vmcs) {
arch/x86/kvm/vmx/vmx.c
1551
per_cpu(current_vmcs, cpu) = vmx->loaded_vmcs->vmcs;
arch/x86/kvm/vmx/vmx.c
1552
vmcs_load(vmx->loaded_vmcs->vmcs);
arch/x86/kvm/vmx/vmx.c
1578
vmx->loaded_vmcs->cpu = cpu;
arch/x86/kvm/vmx/vmx.c
1604
struct loaded_vmcs *vmcs)
arch/x86/kvm/vmx/vmx.c
1610
vmx->loaded_vmcs = vmcs;
arch/x86/kvm/vmx/vmx.c
1620
WARN_ON_ONCE(vmx->loaded_vmcs != &vmx->vmcs01);
arch/x86/kvm/vmx/vmx.c
1624
WARN_ON_ONCE(vmx->loaded_vmcs != &vmx->nested.vmcs02);
arch/x86/kvm/vmx/vmx.c
3037
struct loaded_vmcs *v, *n;
arch/x86/kvm/vmx/vmx.c
3087
void free_loaded_vmcs(struct loaded_vmcs *loaded_vmcs)
arch/x86/kvm/vmx/vmx.c
3089
if (!loaded_vmcs->vmcs)
arch/x86/kvm/vmx/vmx.c
3091
loaded_vmcs_clear(loaded_vmcs);
arch/x86/kvm/vmx/vmx.c
3092
free_vmcs(loaded_vmcs->vmcs);
arch/x86/kvm/vmx/vmx.c
3093
loaded_vmcs->vmcs = NULL;
arch/x86/kvm/vmx/vmx.c
3094
if (loaded_vmcs->msr_bitmap)
arch/x86/kvm/vmx/vmx.c
3095
free_page((unsigned long)loaded_vmcs->msr_bitmap);
arch/x86/kvm/vmx/vmx.c
3096
WARN_ON(loaded_vmcs->shadow_vmcs != NULL);
arch/x86/kvm/vmx/vmx.c
3099
int alloc_loaded_vmcs(struct loaded_vmcs *loaded_vmcs)
arch/x86/kvm/vmx/vmx.c
3101
loaded_vmcs->vmcs = alloc_vmcs(false);
arch/x86/kvm/vmx/vmx.c
3102
if (!loaded_vmcs->vmcs)
arch/x86/kvm/vmx/vmx.c
3105
vmcs_clear(loaded_vmcs->vmcs);
arch/x86/kvm/vmx/vmx.c
3107
loaded_vmcs->shadow_vmcs = NULL;
arch/x86/kvm/vmx/vmx.c
3108
loaded_vmcs->hv_timer_soft_disabled = false;
arch/x86/kvm/vmx/vmx.c
3109
loaded_vmcs->cpu = -1;
arch/x86/kvm/vmx/vmx.c
3110
loaded_vmcs->launched = 0;
arch/x86/kvm/vmx/vmx.c
3113
loaded_vmcs->msr_bitmap = (unsigned long *)
arch/x86/kvm/vmx/vmx.c
3115
if (!loaded_vmcs->msr_bitmap)
arch/x86/kvm/vmx/vmx.c
3117
memset(loaded_vmcs->msr_bitmap, 0xff, PAGE_SIZE);
arch/x86/kvm/vmx/vmx.c
3120
memset(&loaded_vmcs->host_state, 0, sizeof(struct vmcs_host_state));
arch/x86/kvm/vmx/vmx.c
3121
memset(&loaded_vmcs->controls_shadow, 0,
arch/x86/kvm/vmx/vmx.c
3127
free_loaded_vmcs(loaded_vmcs);
arch/x86/kvm/vmx/vmx.c
4529
vmx->loaded_vmcs->host_state.cr3 = cr3;
arch/x86/kvm/vmx/vmx.c
4534
vmx->loaded_vmcs->host_state.cr4 = cr4;
arch/x86/kvm/vmx/vmx.c
5215
vmx->loaded_vmcs->soft_vnmi_blocked = 1;
arch/x86/kvm/vmx/vmx.c
5216
vmx->loaded_vmcs->vnmi_blocked_time = 0;
arch/x86/kvm/vmx/vmx.c
5220
vmx->loaded_vmcs->nmi_known_unmasked = false;
arch/x86/kvm/vmx/vmx.c
5239
return vmx->loaded_vmcs->soft_vnmi_blocked;
arch/x86/kvm/vmx/vmx.c
5240
if (vmx->loaded_vmcs->nmi_known_unmasked)
arch/x86/kvm/vmx/vmx.c
5243
vmx->loaded_vmcs->nmi_known_unmasked = !masked;
arch/x86/kvm/vmx/vmx.c
5252
if (vmx->loaded_vmcs->soft_vnmi_blocked != masked) {
arch/x86/kvm/vmx/vmx.c
5253
vmx->loaded_vmcs->soft_vnmi_blocked = masked;
arch/x86/kvm/vmx/vmx.c
5254
vmx->loaded_vmcs->vnmi_blocked_time = 0;
arch/x86/kvm/vmx/vmx.c
5257
vmx->loaded_vmcs->nmi_known_unmasked = !masked;
arch/x86/kvm/vmx/vmx.c
5272
if (!enable_vnmi && to_vmx(vcpu)->loaded_vmcs->soft_vnmi_blocked)
arch/x86/kvm/vmx/vmx.c
6281
if (unlikely(vmx->loaded_vmcs->hv_timer_soft_disabled))
arch/x86/kvm/vmx/vmx.c
635
evmcs = (struct hv_enlightened_vmcs *)to_vmx(vcpu)->loaded_vmcs->vmcs;
arch/x86/kvm/vmx/vmx.c
6607
vmx->loaded_vmcs->vmcs, vcpu->arch.last_vmentry_cpu);
arch/x86/kvm/vmx/vmx.c
6884
vmx->loaded_vmcs->soft_vnmi_blocked)) {
arch/x86/kvm/vmx/vmx.c
6886
vmx->loaded_vmcs->soft_vnmi_blocked = 0;
arch/x86/kvm/vmx/vmx.c
6887
} else if (vmx->loaded_vmcs->vnmi_blocked_time > 1000000000LL &&
arch/x86/kvm/vmx/vmx.c
6898
vmx->loaded_vmcs->soft_vnmi_blocked = 0;
arch/x86/kvm/vmx/vmx.c
7311
if (vmx->loaded_vmcs->nmi_known_unmasked)
arch/x86/kvm/vmx/vmx.c
7332
vmx->loaded_vmcs->nmi_known_unmasked =
arch/x86/kvm/vmx/vmx.c
7335
} else if (unlikely(vmx->loaded_vmcs->soft_vnmi_blocked))
arch/x86/kvm/vmx/vmx.c
7336
vmx->loaded_vmcs->vnmi_blocked_time +=
arch/x86/kvm/vmx/vmx.c
7338
vmx->loaded_vmcs->entry_time));
arch/x86/kvm/vmx/vmx.c
7473
vmx->loaded_vmcs->hv_timer_soft_disabled = false;
arch/x86/kvm/vmx/vmx.c
7484
vmx->loaded_vmcs->hv_timer_soft_disabled = false;
arch/x86/kvm/vmx/vmx.c
7485
} else if (!vmx->loaded_vmcs->hv_timer_soft_disabled) {
arch/x86/kvm/vmx/vmx.c
7487
vmx->loaded_vmcs->hv_timer_soft_disabled = true;
arch/x86/kvm/vmx/vmx.c
7493
if (unlikely(host_rsp != vmx->loaded_vmcs->host_state.rsp)) {
arch/x86/kvm/vmx/vmx.c
7494
vmx->loaded_vmcs->host_state.rsp = host_rsp;
arch/x86/kvm/vmx/vmx.c
7613
vmx->loaded_vmcs->soft_vnmi_blocked))
arch/x86/kvm/vmx/vmx.c
7614
vmx->loaded_vmcs->entry_time = ktime_get();
arch/x86/kvm/vmx/vmx.c
7666
if (unlikely(cr3 != vmx->loaded_vmcs->host_state.cr3)) {
arch/x86/kvm/vmx/vmx.c
7668
vmx->loaded_vmcs->host_state.cr3 = cr3;
arch/x86/kvm/vmx/vmx.c
7672
if (unlikely(cr4 != vmx->loaded_vmcs->host_state.cr4)) {
arch/x86/kvm/vmx/vmx.c
7674
vmx->loaded_vmcs->host_state.cr4 = cr4;
arch/x86/kvm/vmx/vmx.c
7748
vmx->loaded_vmcs->launched = 1;
arch/x86/kvm/vmx/vmx.c
7766
free_loaded_vmcs(vmx->loaded_vmcs);
arch/x86/kvm/vmx/vmx.c
7827
vmx->loaded_vmcs = &vmx->vmcs01;
arch/x86/kvm/vmx/vmx.c
7862
free_loaded_vmcs(vmx->loaded_vmcs);
arch/x86/kvm/vmx/vmx.c
815
struct loaded_vmcs *v;
arch/x86/kvm/vmx/vmx.c
840
struct loaded_vmcs *loaded_vmcs = arg;
arch/x86/kvm/vmx/vmx.c
843
if (loaded_vmcs->cpu != cpu)
arch/x86/kvm/vmx/vmx.c
845
if (per_cpu(current_vmcs, cpu) == loaded_vmcs->vmcs)
arch/x86/kvm/vmx/vmx.c
848
vmcs_clear(loaded_vmcs->vmcs);
arch/x86/kvm/vmx/vmx.c
849
if (loaded_vmcs->shadow_vmcs && loaded_vmcs->launched)
arch/x86/kvm/vmx/vmx.c
850
vmcs_clear(loaded_vmcs->shadow_vmcs);
arch/x86/kvm/vmx/vmx.c
852
list_del(&loaded_vmcs->loaded_vmcss_on_cpu_link);
arch/x86/kvm/vmx/vmx.c
863
loaded_vmcs->cpu = -1;
arch/x86/kvm/vmx/vmx.c
864
loaded_vmcs->launched = 0;
arch/x86/kvm/vmx/vmx.c
867
static void loaded_vmcs_clear(struct loaded_vmcs *loaded_vmcs)
arch/x86/kvm/vmx/vmx.c
869
int cpu = loaded_vmcs->cpu;
arch/x86/kvm/vmx/vmx.c
873
__loaded_vmcs_clear, loaded_vmcs, 1);
arch/x86/kvm/vmx/vmx.h
147
struct loaded_vmcs vmcs02;
arch/x86/kvm/vmx/vmx.h
232
struct loaded_vmcs vmcs01;
arch/x86/kvm/vmx/vmx.h
233
struct loaded_vmcs *loaded_vmcs;
arch/x86/kvm/vmx/vmx.h
583
if (vmx->loaded_vmcs->controls_shadow.lname != val) { \
arch/x86/kvm/vmx/vmx.h
585
vmx->loaded_vmcs->controls_shadow.lname = val; \
arch/x86/kvm/vmx/vmx.h
588
static inline u##bits __##lname##_controls_get(struct loaded_vmcs *vmcs) \
arch/x86/kvm/vmx/vmx.h
594
return __##lname##_controls_get(vmx->loaded_vmcs); \
arch/x86/kvm/vmx/vmx.h
671
int alloc_loaded_vmcs(struct loaded_vmcs *loaded_vmcs);
arch/x86/kvm/vmx/vmx.h
672
void free_loaded_vmcs(struct loaded_vmcs *loaded_vmcs);