Symbol: vmcs
arch/loongarch/include/asm/kvm_host.h
134
struct kvm_context __percpu *vmcs;
arch/loongarch/kvm/main.c
19
static struct kvm_context __percpu *vmcs;
arch/loongarch/kvm/main.c
210
context = per_cpu_ptr(vcpu->kvm->arch.vmcs, cpu);
arch/loongarch/kvm/main.c
239
context = per_cpu_ptr(vcpu->kvm->arch.vmcs, cpu);
arch/loongarch/kvm/main.c
283
kvm->arch.vmcs = vmcs;
arch/loongarch/kvm/main.c
333
this_cpu_ptr(vmcs)->last_vcpu = NULL;
arch/loongarch/kvm/main.c
355
vmcs = alloc_percpu(struct kvm_context);
arch/loongarch/kvm/main.c
356
if (!vmcs) {
arch/loongarch/kvm/main.c
363
free_percpu(vmcs);
arch/loongarch/kvm/main.c
364
vmcs = NULL;
arch/loongarch/kvm/main.c
379
free_percpu(vmcs);
arch/loongarch/kvm/main.c
380
vmcs = NULL;
arch/loongarch/kvm/main.c
399
context = per_cpu_ptr(vmcs, cpu);
arch/loongarch/kvm/main.c
427
if (vmcs)
arch/loongarch/kvm/main.c
428
free_percpu(vmcs);
arch/loongarch/kvm/vcpu.c
1608
context = per_cpu_ptr(vcpu->kvm->arch.vmcs, cpu);
arch/loongarch/kvm/vcpu.c
1630
context = per_cpu_ptr(vcpu->kvm->arch.vmcs, cpu);
arch/loongarch/kvm/vcpu.c
45
context = this_cpu_ptr(vcpu->kvm->arch.vmcs);
arch/loongarch/kvm/vcpu.c
60
context = this_cpu_ptr(vcpu->kvm->arch.vmcs);
arch/x86/kvm/vmx/nested.c
1630
struct vmcs *shadow_vmcs = vmx->vmcs01.shadow_vmcs;
arch/x86/kvm/vmx/nested.c
1650
vmcs_load(vmx->loaded_vmcs->vmcs);
arch/x86/kvm/vmx/nested.c
1665
struct vmcs *shadow_vmcs = vmx->vmcs01.shadow_vmcs;
arch/x86/kvm/vmx/nested.c
1686
vmcs_load(vmx->loaded_vmcs->vmcs);
arch/x86/kvm/vmx/nested.c
297
static void vmx_switch_vmcs(struct kvm_vcpu *vcpu, struct loaded_vmcs *vmcs)
arch/x86/kvm/vmx/nested.c
303
if (WARN_ON_ONCE(vmx->loaded_vmcs == vmcs))
arch/x86/kvm/vmx/nested.c
308
vmx->loaded_vmcs = vmcs;
arch/x86/kvm/vmx/nested.c
5371
static struct vmcs *alloc_shadow_vmcs(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
5835
vmcs_load(vmx->loaded_vmcs->vmcs);
arch/x86/kvm/vmx/tdx.h
182
TDX_BUILD_TDVPS_ACCESSORS(16, VMCS, vmcs);
arch/x86/kvm/vmx/tdx.h
183
TDX_BUILD_TDVPS_ACCESSORS(32, VMCS, vmcs);
arch/x86/kvm/vmx/tdx.h
184
TDX_BUILD_TDVPS_ACCESSORS(64, VMCS, vmcs);
arch/x86/kvm/vmx/vmcs.h
36
DECLARE_PER_CPU(struct vmcs *, current_vmcs);
arch/x86/kvm/vmx/vmcs.h
71
struct vmcs *vmcs;
arch/x86/kvm/vmx/vmcs.h
72
struct vmcs *shadow_vmcs;
arch/x86/kvm/vmx/vmx.c
1530
struct vmcs *prev;
arch/x86/kvm/vmx/vmx.c
1550
if (prev != vmx->loaded_vmcs->vmcs) {
arch/x86/kvm/vmx/vmx.c
1551
per_cpu(current_vmcs, cpu) = vmx->loaded_vmcs->vmcs;
arch/x86/kvm/vmx/vmx.c
1552
vmcs_load(vmx->loaded_vmcs->vmcs);
arch/x86/kvm/vmx/vmx.c
1604
struct loaded_vmcs *vmcs)
arch/x86/kvm/vmx/vmx.c
1610
vmx->loaded_vmcs = vmcs;
arch/x86/kvm/vmx/vmx.c
3056
struct vmcs *alloc_vmcs_cpu(bool shadow, int cpu, gfp_t flags)
arch/x86/kvm/vmx/vmx.c
3060
struct vmcs *vmcs;
arch/x86/kvm/vmx/vmx.c
3065
vmcs = page_address(pages);
arch/x86/kvm/vmx/vmx.c
3066
memset(vmcs, 0, vmx_basic_vmcs_size(vmcs_config.basic));
arch/x86/kvm/vmx/vmx.c
3070
vmcs->hdr.revision_id = KVM_EVMCS_VERSION;
arch/x86/kvm/vmx/vmx.c
3072
vmcs->hdr.revision_id = vmx_basic_vmcs_revision_id(vmcs_config.basic);
arch/x86/kvm/vmx/vmx.c
3075
vmcs->hdr.shadow_vmcs = 1;
arch/x86/kvm/vmx/vmx.c
3076
return vmcs;
arch/x86/kvm/vmx/vmx.c
3079
void free_vmcs(struct vmcs *vmcs)
arch/x86/kvm/vmx/vmx.c
3081
free_page((unsigned long)vmcs);
arch/x86/kvm/vmx/vmx.c
3089
if (!loaded_vmcs->vmcs)
arch/x86/kvm/vmx/vmx.c
3092
free_vmcs(loaded_vmcs->vmcs);
arch/x86/kvm/vmx/vmx.c
3093
loaded_vmcs->vmcs = NULL;
arch/x86/kvm/vmx/vmx.c
3101
loaded_vmcs->vmcs = alloc_vmcs(false);
arch/x86/kvm/vmx/vmx.c
3102
if (!loaded_vmcs->vmcs)
arch/x86/kvm/vmx/vmx.c
3105
vmcs_clear(loaded_vmcs->vmcs);
arch/x86/kvm/vmx/vmx.c
3146
struct vmcs *vmcs;
arch/x86/kvm/vmx/vmx.c
3148
vmcs = alloc_vmcs_cpu(false, cpu, GFP_KERNEL);
arch/x86/kvm/vmx/vmx.c
3149
if (!vmcs) {
arch/x86/kvm/vmx/vmx.c
3165
vmcs->hdr.revision_id = vmx_basic_vmcs_revision_id(vmcs_config.basic);
arch/x86/kvm/vmx/vmx.c
3167
per_cpu(vmxarea, cpu) = vmcs;
arch/x86/kvm/vmx/vmx.c
4176
struct hv_enlightened_vmcs *evmcs = (void *)vmx->vmcs01.vmcs;
arch/x86/kvm/vmx/vmx.c
559
noinline void vmclear_error(struct vmcs *vmcs, u64 phys_addr)
arch/x86/kvm/vmx/vmx.c
562
vmcs, phys_addr, vmcs_read32(VM_INSTRUCTION_ERROR));
arch/x86/kvm/vmx/vmx.c
565
noinline void vmptrld_error(struct vmcs *vmcs, u64 phys_addr)
arch/x86/kvm/vmx/vmx.c
568
vmcs, phys_addr, vmcs_read32(VM_INSTRUCTION_ERROR));
arch/x86/kvm/vmx/vmx.c
582
static DEFINE_PER_CPU(struct vmcs *, vmxarea);
arch/x86/kvm/vmx/vmx.c
583
DEFINE_PER_CPU(struct vmcs *, current_vmcs);
arch/x86/kvm/vmx/vmx.c
635
evmcs = (struct hv_enlightened_vmcs *)to_vmx(vcpu)->loaded_vmcs->vmcs;
arch/x86/kvm/vmx/vmx.c
6607
vmx->loaded_vmcs->vmcs, vcpu->arch.last_vmentry_cpu);
arch/x86/kvm/vmx/vmx.c
7822
struct hv_enlightened_vmcs *evmcs = (void *)vmx->vmcs01.vmcs;
arch/x86/kvm/vmx/vmx.c
830
vmcs_clear(v->vmcs);
arch/x86/kvm/vmx/vmx.c
845
if (per_cpu(current_vmcs, cpu) == loaded_vmcs->vmcs)
arch/x86/kvm/vmx/vmx.c
848
vmcs_clear(loaded_vmcs->vmcs);
arch/x86/kvm/vmx/vmx.h
588
static inline u##bits __##lname##_controls_get(struct loaded_vmcs *vmcs) \
arch/x86/kvm/vmx/vmx.h
590
return vmcs->controls_shadow.lname; \
arch/x86/kvm/vmx/vmx.h
669
struct vmcs *alloc_vmcs_cpu(bool shadow, int cpu, gfp_t flags);
arch/x86/kvm/vmx/vmx.h
670
void free_vmcs(struct vmcs *vmcs);
arch/x86/kvm/vmx/vmx.h
674
static inline struct vmcs *alloc_vmcs(bool shadow)
arch/x86/kvm/vmx/vmx_ops.h
15
void vmclear_error(struct vmcs *vmcs, u64 phys_addr);
arch/x86/kvm/vmx/vmx_ops.h
16
void vmptrld_error(struct vmcs *vmcs, u64 phys_addr);
arch/x86/kvm/vmx/vmx_ops.h
289
static inline void vmcs_clear(struct vmcs *vmcs)
arch/x86/kvm/vmx/vmx_ops.h
291
u64 phys_addr = __pa(vmcs);
arch/x86/kvm/vmx/vmx_ops.h
293
vmx_asm1(vmclear, "m"(phys_addr), vmcs, phys_addr);
arch/x86/kvm/vmx/vmx_ops.h
296
static inline void vmcs_load(struct vmcs *vmcs)
arch/x86/kvm/vmx/vmx_ops.h
298
u64 phys_addr = __pa(vmcs);
arch/x86/kvm/vmx/vmx_ops.h
303
vmx_asm1(vmptrld, "m"(phys_addr), vmcs, phys_addr);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
124
struct intel_pt_vmcs_info *(*findnew_vmcs_info)(void *data, uint64_t vmcs);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
167
uint64_t vmcs;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2320
bool tsc, pip, vmcs, tma, psbend;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2358
data->vmcs = true;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2518
payload, decoder->packet.payload, vmcs_info->vmcs,
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2526
.vmcs = NO_VMCS,
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2546
vmcs_info->vmcs, vmcs_info->tsc_offset);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2601
uint64_t vmcs;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2629
vmcs = data->vmcs ? data->vmcs_packet.payload : decoder->vmcs;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2630
if (vmcs == NO_VMCS)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2631
vmcs = 0;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2633
vmcs_info = decoder->findnew_vmcs_info(decoder->data, vmcs);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2697
if (vmcs_info && vmcs_info->vmcs) {
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2736
vmcs_info->vmcs);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2754
uint64_t vmcs;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2756
vmcs = decoder->vmcs;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2757
if (vmcs == NO_VMCS)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2758
vmcs = 0;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2760
vmcs_info = decoder->findnew_vmcs_info(decoder->data, vmcs);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2784
guest_tsc, host_tsc, vmcs_info->vmcs,
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2888
decoder->vmcs = decoder->packet.payload;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.h
210
uint64_t vmcs;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.h
280
struct intel_pt_vmcs_info *(*findnew_vmcs_info)(void *data, uint64_t vmcs);
tools/perf/util/intel-pt.c
319
u64 vmcs,
tools/perf/util/intel-pt.c
330
if (v->vmcs == vmcs)
tools/perf/util/intel-pt.c
333
if (vmcs < v->vmcs)
tools/perf/util/intel-pt.c
341
v->vmcs = vmcs;
tools/perf/util/intel-pt.c
352
static struct intel_pt_vmcs_info *intel_pt_findnew_vmcs_info(void *data, uint64_t vmcs)
tools/perf/util/intel-pt.c
357
if (!vmcs && !pt->dflt_tsc_offset)
tools/perf/util/intel-pt.c
360
return intel_pt_findnew_vmcs(&pt->vmcs_info, vmcs, pt->dflt_tsc_offset);
tools/perf/util/intel-pt.c
4302
u64 tsc_offset, vmcs;
tools/perf/util/intel-pt.c
4322
vmcs = strtoull(p, &p, 0);
tools/perf/util/intel-pt.c
4325
if (!vmcs)
tools/perf/util/intel-pt.c
4327
vmcs_info = intel_pt_findnew_vmcs(&pt->vmcs_info, vmcs, tsc_offset);
tools/testing/selftests/kvm/include/x86/evmcs.h
248
static inline int evmcs_vmptrld(uint64_t vmcs_pa, void *vmcs)
tools/testing/selftests/kvm/include/x86/evmcs.h
253
current_evmcs = vmcs;
tools/testing/selftests/kvm/include/x86/vmx.h
505
void *vmcs;
tools/testing/selftests/kvm/lib/x86/vmx.c
173
*(uint32_t *)(vmx->vmcs) = vmcs_revision();
tools/testing/selftests/kvm/lib/x86/vmx.c
93
vmx->vmcs = (void *)vm_vaddr_alloc_page(vm);
tools/testing/selftests/kvm/lib/x86/vmx.c
94
vmx->vmcs_hva = addr_gva2hva(vm, (uintptr_t)vmx->vmcs);
tools/testing/selftests/kvm/lib/x86/vmx.c
95
vmx->vmcs_gpa = addr_gva2gpa(vm, (uintptr_t)vmx->vmcs);