arch/x86/include/asm/msr.h
32
struct msr_info info;
arch/x86/kvm/pmu.c
820
int kvm_pmu_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/pmu.c
823
u32 msr = msr_info->index;
arch/x86/kvm/pmu.c
828
msr_info->data = pmu->global_status;
arch/x86/kvm/pmu.c
832
msr_info->data = pmu->global_ctrl;
arch/x86/kvm/pmu.c
837
msr_info->data = 0;
arch/x86/kvm/pmu.c
840
return kvm_pmu_call(get_msr)(vcpu, msr_info);
arch/x86/kvm/pmu.c
846
int kvm_pmu_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/pmu.c
849
u32 msr = msr_info->index;
arch/x86/kvm/pmu.c
850
u64 data = msr_info->data;
arch/x86/kvm/pmu.c
859
if (!msr_info->host_initiated)
arch/x86/kvm/pmu.c
864
if (!msr_info->host_initiated)
arch/x86/kvm/pmu.c
900
if (!msr_info->host_initiated)
arch/x86/kvm/pmu.c
904
if (!msr_info->host_initiated)
arch/x86/kvm/pmu.c
908
kvm_pmu_mark_pmc_in_use(vcpu, msr_info->index);
arch/x86/kvm/pmu.c
909
return kvm_pmu_call(set_msr)(vcpu, msr_info);
arch/x86/kvm/pmu.h
252
int kvm_pmu_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info);
arch/x86/kvm/pmu.h
253
int kvm_pmu_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info);
arch/x86/kvm/pmu.h
32
int (*get_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr_info);
arch/x86/kvm/pmu.h
33
int (*set_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr_info);
arch/x86/kvm/svm/pmu.c
128
static int amd_pmu_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/svm/pmu.c
132
u32 msr = msr_info->index;
arch/x86/kvm/svm/pmu.c
137
msr_info->data = pmc_read_counter(pmc);
arch/x86/kvm/svm/pmu.c
143
msr_info->data = pmc->eventsel;
arch/x86/kvm/svm/pmu.c
150
static int amd_pmu_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/svm/pmu.c
154
u32 msr = msr_info->index;
arch/x86/kvm/svm/pmu.c
155
u64 data = msr_info->data;
arch/x86/kvm/svm/svm.c
2693
struct msr_data msr_info;
arch/x86/kvm/svm/svm.c
2702
msr_info.host_initiated = false;
arch/x86/kvm/svm/svm.c
2703
msr_info.index = MSR_EFER;
arch/x86/kvm/svm/svm.c
2704
msr_info.data = to_svm(vcpu)->vmcb->control.exit_info_1 & ~EFER_SVME;
arch/x86/kvm/svm/svm.c
2705
ret = kvm_set_msr_common(vcpu, &msr_info);
arch/x86/kvm/svm/svm.c
2727
struct msr_data *msr_info)
arch/x86/kvm/svm/svm.c
2730
msr_info->index != MSR_IA32_XSS &&
arch/x86/kvm/svm/svm.c
2731
!msr_write_intercepted(vcpu, msr_info->index);
arch/x86/kvm/svm/svm.c
2734
static int svm_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/svm/svm.c
2738
if (sev_es_prevent_msr_access(vcpu, msr_info)) {
arch/x86/kvm/svm/svm.c
2739
msr_info->data = 0;
arch/x86/kvm/svm/svm.c
2743
switch (msr_info->index) {
arch/x86/kvm/svm/svm.c
2745
if (!msr_info->host_initiated &&
arch/x86/kvm/svm/svm.c
2748
msr_info->data = svm->tsc_ratio_msr;
arch/x86/kvm/svm/svm.c
2751
msr_info->data = svm->vmcb01.ptr->save.star;
arch/x86/kvm/svm/svm.c
2755
msr_info->data = svm->vmcb01.ptr->save.lstar;
arch/x86/kvm/svm/svm.c
2758
msr_info->data = svm->vmcb01.ptr->save.cstar;
arch/x86/kvm/svm/svm.c
2761
msr_info->data = svm->vmcb01.ptr->save.gs.base;
arch/x86/kvm/svm/svm.c
2764
msr_info->data = svm->vmcb01.ptr->save.fs.base;
arch/x86/kvm/svm/svm.c
2767
msr_info->data = svm->vmcb01.ptr->save.kernel_gs_base;
arch/x86/kvm/svm/svm.c
2770
msr_info->data = svm->vmcb01.ptr->save.sfmask;
arch/x86/kvm/svm/svm.c
2774
msr_info->data = svm->vmcb01.ptr->save.sysenter_cs;
arch/x86/kvm/svm/svm.c
2777
msr_info->data = (u32)svm->vmcb01.ptr->save.sysenter_eip;
arch/x86/kvm/svm/svm.c
2779
msr_info->data |= (u64)svm->sysenter_eip_hi << 32;
arch/x86/kvm/svm/svm.c
2782
msr_info->data = svm->vmcb01.ptr->save.sysenter_esp;
arch/x86/kvm/svm/svm.c
2784
msr_info->data |= (u64)svm->sysenter_esp_hi << 32;
arch/x86/kvm/svm/svm.c
2787
msr_info->data = svm->vmcb->save.s_cet;
arch/x86/kvm/svm/svm.c
2790
msr_info->data = svm->vmcb->save.isst_addr;
arch/x86/kvm/svm/svm.c
2793
msr_info->data = svm->vmcb->save.ssp;
arch/x86/kvm/svm/svm.c
2796
msr_info->data = svm->tsc_aux;
arch/x86/kvm/svm/svm.c
2799
msr_info->data = svm->vmcb->save.dbgctl;
arch/x86/kvm/svm/svm.c
2802
msr_info->data = svm->vmcb->save.br_from;
arch/x86/kvm/svm/svm.c
2805
msr_info->data = svm->vmcb->save.br_to;
arch/x86/kvm/svm/svm.c
2808
msr_info->data = svm->vmcb->save.last_excp_from;
arch/x86/kvm/svm/svm.c
2811
msr_info->data = svm->vmcb->save.last_excp_to;
arch/x86/kvm/svm/svm.c
2814
msr_info->data = svm->nested.hsave_msr;
arch/x86/kvm/svm/svm.c
2817
msr_info->data = svm->nested.vm_cr_msr;
arch/x86/kvm/svm/svm.c
2820
if (!msr_info->host_initiated &&
arch/x86/kvm/svm/svm.c
2825
msr_info->data = svm->vmcb->save.spec_ctrl;
arch/x86/kvm/svm/svm.c
2827
msr_info->data = svm->spec_ctrl;
arch/x86/kvm/svm/svm.c
2830
if (!msr_info->host_initiated &&
arch/x86/kvm/svm/svm.c
2834
msr_info->data = svm->virt_spec_ctrl;
arch/x86/kvm/svm/svm.c
2844
return kvm_get_msr_common(vcpu, msr_info);
arch/x86/kvm/svm/svm.c
2846
msr_info->data = 0;
arch/x86/kvm/svm/svm.c
2850
msr_info->data = 0x1E;
arch/x86/kvm/svm/svm.c
2854
msr_info->data = svm->msr_decfg;
arch/x86/kvm/svm/svm.c
2857
return kvm_get_msr_common(vcpu, msr_info);
arch/x86/kvm/vmx/main.c
163
static int vt_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/vmx/main.c
166
return tdx_set_msr(vcpu, msr_info);
arch/x86/kvm/vmx/main.c
168
return vmx_set_msr(vcpu, msr_info);
arch/x86/kvm/vmx/main.c
183
static int vt_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/vmx/main.c
186
return tdx_get_msr(vcpu, msr_info);
arch/x86/kvm/vmx/main.c
188
return vmx_get_msr(vcpu, msr_info);
arch/x86/kvm/vmx/pmu_intel.c
292
struct msr_data *msr_info, bool read)
arch/x86/kvm/vmx/pmu_intel.c
295
u32 index = msr_info->index;
arch/x86/kvm/vmx/pmu_intel.c
312
rdmsrq(index, msr_info->data);
arch/x86/kvm/vmx/pmu_intel.c
314
wrmsrq(index, msr_info->data);
arch/x86/kvm/vmx/pmu_intel.c
324
msr_info->data = 0;
arch/x86/kvm/vmx/pmu_intel.c
328
static int intel_pmu_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/vmx/pmu_intel.c
332
u32 msr = msr_info->index;
arch/x86/kvm/vmx/pmu_intel.c
336
msr_info->data = pmu->fixed_ctr_ctrl;
arch/x86/kvm/vmx/pmu_intel.c
339
msr_info->data = pmu->pebs_enable;
arch/x86/kvm/vmx/pmu_intel.c
342
msr_info->data = pmu->ds_area;
arch/x86/kvm/vmx/pmu_intel.c
345
msr_info->data = pmu->pebs_data_cfg;
arch/x86/kvm/vmx/pmu_intel.c
351
msr_info->data =
arch/x86/kvm/vmx/pmu_intel.c
356
msr_info->data =
arch/x86/kvm/vmx/pmu_intel.c
360
msr_info->data = pmc->eventsel;
arch/x86/kvm/vmx/pmu_intel.c
362
} else if (intel_pmu_handle_lbr_msrs_access(vcpu, msr_info, true)) {
arch/x86/kvm/vmx/pmu_intel.c
371
static int intel_pmu_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/vmx/pmu_intel.c
375
u32 msr = msr_info->index;
arch/x86/kvm/vmx/pmu_intel.c
376
u64 data = msr_info->data;
arch/x86/kvm/vmx/pmu_intel.c
416
if (!msr_info->host_initiated &&
arch/x86/kvm/vmx/pmu_intel.c
438
} else if (intel_pmu_handle_lbr_msrs_access(vcpu, msr_info, false)) {
arch/x86/kvm/vmx/vmx.c
2124
int vmx_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/vmx/vmx.c
2130
switch (msr_info->index) {
arch/x86/kvm/vmx/vmx.c
2133
msr_info->data = vmcs_readl(GUEST_FS_BASE);
arch/x86/kvm/vmx/vmx.c
2136
msr_info->data = vmcs_readl(GUEST_GS_BASE);
arch/x86/kvm/vmx/vmx.c
2139
msr_info->data = vmx_read_guest_kernel_gs_base(vmx);
arch/x86/kvm/vmx/vmx.c
2143
return kvm_get_msr_common(vcpu, msr_info);
arch/x86/kvm/vmx/vmx.c
2145
if (!msr_info->host_initiated &&
arch/x86/kvm/vmx/vmx.c
2150
if (!msr_info->host_initiated && !vmx_has_waitpkg(vmx))
arch/x86/kvm/vmx/vmx.c
2153
msr_info->data = vmx->msr_ia32_umwait_control;
arch/x86/kvm/vmx/vmx.c
2156
if (!msr_info->host_initiated &&
arch/x86/kvm/vmx/vmx.c
2160
msr_info->data = to_vmx(vcpu)->spec_ctrl;
arch/x86/kvm/vmx/vmx.c
2163
msr_info->data = vmcs_read32(GUEST_SYSENTER_CS);
arch/x86/kvm/vmx/vmx.c
2166
msr_info->data = vmcs_readl(GUEST_SYSENTER_EIP);
arch/x86/kvm/vmx/vmx.c
2169
msr_info->data = vmcs_readl(GUEST_SYSENTER_ESP);
arch/x86/kvm/vmx/vmx.c
2173
(!msr_info->host_initiated &&
arch/x86/kvm/vmx/vmx.c
2176
msr_info->data = vmcs_read64(GUEST_BNDCFGS);
arch/x86/kvm/vmx/vmx.c
2179
if (!msr_info->host_initiated &&
arch/x86/kvm/vmx/vmx.c
2183
msr_info->data = vcpu->arch.mcg_ext_ctl;
arch/x86/kvm/vmx/vmx.c
2186
msr_info->data = vmx->msr_ia32_feature_control;
arch/x86/kvm/vmx/vmx.c
2189
if (!msr_info->host_initiated &&
arch/x86/kvm/vmx/vmx.c
2192
msr_info->data = to_vmx(vcpu)->msr_ia32_sgxlepubkeyhash
arch/x86/kvm/vmx/vmx.c
2193
[msr_info->index - MSR_IA32_SGXLEPUBKEYHASH0];
arch/x86/kvm/vmx/vmx.c
2198
if (vmx_get_vmx_msr(&vmx->nested.msrs, msr_info->index,
arch/x86/kvm/vmx/vmx.c
2199
&msr_info->data))
arch/x86/kvm/vmx/vmx.c
2209
if (!msr_info->host_initiated && guest_cpu_cap_has_evmcs(vcpu))
arch/x86/kvm/vmx/vmx.c
2210
nested_evmcs_filter_control_msr(vcpu, msr_info->index,
arch/x86/kvm/vmx/vmx.c
2211
&msr_info->data);
arch/x86/kvm/vmx/vmx.c
2217
msr_info->data = vmx->pt_desc.guest.ctl;
arch/x86/kvm/vmx/vmx.c
2222
msr_info->data = vmx->pt_desc.guest.status;
arch/x86/kvm/vmx/vmx.c
2229
msr_info->data = vmx->pt_desc.guest.cr3_match;
arch/x86/kvm/vmx/vmx.c
2238
msr_info->data = vmx->pt_desc.guest.output_base;
arch/x86/kvm/vmx/vmx.c
2247
msr_info->data = vmx->pt_desc.guest.output_mask;
arch/x86/kvm/vmx/vmx.c
2250
index = msr_info->index - MSR_IA32_RTIT_ADDR0_A;
arch/x86/kvm/vmx/vmx.c
2255
msr_info->data = vmx->pt_desc.guest.addr_b[index / 2];
arch/x86/kvm/vmx/vmx.c
2257
msr_info->data = vmx->pt_desc.guest.addr_a[index / 2];
arch/x86/kvm/vmx/vmx.c
2260
msr_info->data = vmcs_readl(GUEST_S_CET);
arch/x86/kvm/vmx/vmx.c
2263
msr_info->data = vmcs_readl(GUEST_SSP);
arch/x86/kvm/vmx/vmx.c
2266
msr_info->data = vmcs_readl(GUEST_INTR_SSP_TABLE);
arch/x86/kvm/vmx/vmx.c
2269
msr_info->data = vmx_guest_debugctl_read();
arch/x86/kvm/vmx/vmx.c
2273
msr = vmx_find_uret_msr(vmx, msr_info->index);
arch/x86/kvm/vmx/vmx.c
2275
msr_info->data = msr->data;
arch/x86/kvm/vmx/vmx.c
2278
return kvm_get_msr_common(vcpu, msr_info);
arch/x86/kvm/vmx/vmx.c
2330
int vmx_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/vmx/vmx.c
2335
u32 msr_index = msr_info->index;
arch/x86/kvm/vmx/vmx.c
2336
u64 data = msr_info->data;
arch/x86/kvm/vmx/vmx.c
2341
ret = kvm_set_msr_common(vcpu, msr_info);
arch/x86/kvm/vmx/vmx.c
2356
ret = kvm_set_msr_common(vcpu, msr_info);
arch/x86/kvm/vmx/vmx.c
2394
if (!vmx_is_valid_debugctl(vcpu, data, msr_info->host_initiated))
arch/x86/kvm/vmx/vmx.c
2397
data &= vmx_get_supported_debugctl(vcpu, msr_info->host_initiated);
arch/x86/kvm/vmx/vmx.c
2411
(!msr_info->host_initiated &&
arch/x86/kvm/vmx/vmx.c
2426
if (!msr_info->host_initiated && !vmx_has_waitpkg(vmx))
arch/x86/kvm/vmx/vmx.c
2436
if (!msr_info->host_initiated &&
arch/x86/kvm/vmx/vmx.c
2464
if (!msr_info->host_initiated &&
arch/x86/kvm/vmx/vmx.c
2471
ret = kvm_set_msr_common(vcpu, msr_info);
arch/x86/kvm/vmx/vmx.c
2483
if ((!msr_info->host_initiated &&
arch/x86/kvm/vmx/vmx.c
2491
if (!is_vmx_feature_control_msr_valid(vmx, msr_info))
arch/x86/kvm/vmx/vmx.c
2495
if (msr_info->host_initiated && data == 0)
arch/x86/kvm/vmx/vmx.c
2513
if (!msr_info->host_initiated &&
arch/x86/kvm/vmx/vmx.c
2522
if (!msr_info->host_initiated)
arch/x86/kvm/vmx/vmx.c
2576
index = msr_info->index - MSR_IA32_RTIT_ADDR0_A;
arch/x86/kvm/vmx/vmx.c
2614
ret = kvm_set_msr_common(vcpu, msr_info);
arch/x86/kvm/vmx/vmx.c
2623
ret = kvm_set_msr_common(vcpu, msr_info);
arch/x86/kvm/vmx/x86_ops.h
33
int vmx_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info);
arch/x86/kvm/vmx/x86_ops.h
59
int vmx_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info);
arch/x86/kvm/x86.c
1756
static int set_efer(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/x86.c
1759
u64 efer = msr_info->data;
arch/x86/kvm/x86.c
1765
if (!msr_info->host_initiated) {
arch/x86/kvm/x86.c
3557
static int set_msr_mce(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/x86.c
3561
u32 msr = msr_info->index;
arch/x86/kvm/x86.c
3562
u64 data = msr_info->data;
arch/x86/kvm/x86.c
3571
(data || !msr_info->host_initiated))
arch/x86/kvm/x86.c
3582
if (!(mcg_cap & MCG_CMCI_P) && (data || !msr_info->host_initiated))
arch/x86/kvm/x86.c
3615
if (!msr_info->host_initiated && is_mci_status_msr(msr) &&
arch/x86/kvm/x86.c
3878
struct msr_data *msr_info,
arch/x86/kvm/x86.c
3883
KVM_BUG_ON(!is_xstate_managed_msr(vcpu, msr_info->index), vcpu->kvm);
arch/x86/kvm/x86.c
3888
rdmsrq(msr_info->index, msr_info->data);
arch/x86/kvm/x86.c
3890
wrmsrq(msr_info->index, msr_info->data);
arch/x86/kvm/x86.c
3894
static void kvm_set_xstate_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/x86.c
3896
kvm_access_xstate_msr(vcpu, msr_info, MSR_TYPE_W);
arch/x86/kvm/x86.c
3899
static void kvm_get_xstate_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/x86.c
3901
kvm_access_xstate_msr(vcpu, msr_info, MSR_TYPE_R);
arch/x86/kvm/x86.c
3904
int kvm_set_msr_common(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/x86.c
3906
u32 msr = msr_info->index;
arch/x86/kvm/x86.c
3907
u64 data = msr_info->data;
arch/x86/kvm/x86.c
3915
!msr_info->host_initiated)
arch/x86/kvm/x86.c
3930
if (msr_info->host_initiated)
arch/x86/kvm/x86.c
3934
if (!msr_info->host_initiated ||
arch/x86/kvm/x86.c
3940
if (!msr_info->host_initiated ||
arch/x86/kvm/x86.c
3962
if (!msr_info->host_initiated) {
arch/x86/kvm/x86.c
3990
if (!msr_info->host_initiated &&
arch/x86/kvm/x86.c
4002
return set_efer(vcpu, msr_info);
arch/x86/kvm/x86.c
4035
return kvm_apic_set_base(vcpu, data, msr_info->host_initiated);
arch/x86/kvm/x86.c
4043
if (!msr_info->host_initiated) {
arch/x86/kvm/x86.c
4057
if (!msr_info->host_initiated) {
arch/x86/kvm/x86.c
4079
if (!IS_ENABLED(CONFIG_KVM_SMM) || !msr_info->host_initiated)
arch/x86/kvm/x86.c
4087
if (msr_info->host_initiated) {
arch/x86/kvm/x86.c
4107
if (!msr_info->host_initiated)
arch/x86/kvm/x86.c
4129
kvm_write_system_time(vcpu, data, false, msr_info->host_initiated);
arch/x86/kvm/x86.c
4135
kvm_write_system_time(vcpu, data, true, msr_info->host_initiated);
arch/x86/kvm/x86.c
4205
return set_msr_mce(vcpu, msr_info);
arch/x86/kvm/x86.c
4212
return kvm_pmu_set_msr(vcpu, msr_info);
arch/x86/kvm/x86.c
4239
msr_info->host_initiated);
arch/x86/kvm/x86.c
4258
if (!msr_info->host_initiated)
arch/x86/kvm/x86.c
4271
if (!msr_info->host_initiated &&
arch/x86/kvm/x86.c
4281
if (!msr_info->host_initiated &&
arch/x86/kvm/x86.c
4293
kvm_set_xstate_msr(vcpu, msr_info);
arch/x86/kvm/x86.c
4297
return kvm_pmu_set_msr(vcpu, msr_info);
arch/x86/kvm/x86.c
4355
int kvm_get_msr_common(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/x86.c
4357
switch (msr_info->index) {
arch/x86/kvm/x86.c
4387
msr_info->data = 0;
arch/x86/kvm/x86.c
4393
if (kvm_pmu_is_valid_msr(vcpu, msr_info->index))
arch/x86/kvm/x86.c
4394
return kvm_pmu_get_msr(vcpu, msr_info);
arch/x86/kvm/x86.c
4395
msr_info->data = 0;
arch/x86/kvm/x86.c
4398
msr_info->data = vcpu->arch.microcode_version;
arch/x86/kvm/x86.c
4403
msr_info->data = vcpu->arch.arch_capabilities;
arch/x86/kvm/x86.c
4408
msr_info->data = vcpu->arch.perf_capabilities;
arch/x86/kvm/x86.c
4411
msr_info->data = vcpu->arch.msr_ia32_power_ctl;
arch/x86/kvm/x86.c
4425
if (msr_info->host_initiated) {
arch/x86/kvm/x86.c
4433
msr_info->data = kvm_scale_tsc(rdtsc(), ratio) + offset;
arch/x86/kvm/x86.c
4437
msr_info->data = vcpu->arch.pat;
arch/x86/kvm/x86.c
4442
return kvm_mtrr_get_msr(vcpu, msr_info->index, &msr_info->data);
arch/x86/kvm/x86.c
4444
msr_info->data = 3;
arch/x86/kvm/x86.c
4458
msr_info->data = 1 << 24;
arch/x86/kvm/x86.c
4461
msr_info->data = vcpu->arch.apic_base;
arch/x86/kvm/x86.c
4464
return kvm_x2apic_msr_read(vcpu, msr_info->index, &msr_info->data);
arch/x86/kvm/x86.c
4466
msr_info->data = kvm_get_lapic_tscdeadline_msr(vcpu);
arch/x86/kvm/x86.c
4469
msr_info->data = (u64)vcpu->arch.ia32_tsc_adjust_msr;
arch/x86/kvm/x86.c
4472
msr_info->data = vcpu->arch.ia32_misc_enable_msr;
arch/x86/kvm/x86.c
4475
if (!IS_ENABLED(CONFIG_KVM_SMM) || !msr_info->host_initiated)
arch/x86/kvm/x86.c
4477
msr_info->data = vcpu->arch.smbase;
arch/x86/kvm/x86.c
4480
msr_info->data = vcpu->arch.smi_count;
arch/x86/kvm/x86.c
4484
msr_info->data = 1000ULL;
arch/x86/kvm/x86.c
4486
msr_info->data |= (((uint64_t)4ULL) << 40);
arch/x86/kvm/x86.c
4489
msr_info->data = vcpu->arch.efer;
arch/x86/kvm/x86.c
4495
msr_info->data = vcpu->kvm->arch.wall_clock;
arch/x86/kvm/x86.c
4501
msr_info->data = vcpu->kvm->arch.wall_clock;
arch/x86/kvm/x86.c
4507
msr_info->data = vcpu->arch.time;
arch/x86/kvm/x86.c
4513
msr_info->data = vcpu->arch.time;
arch/x86/kvm/x86.c
4519
msr_info->data = vcpu->arch.apf.msr_en_val;
arch/x86/kvm/x86.c
4525
msr_info->data = vcpu->arch.apf.msr_int_val;
arch/x86/kvm/x86.c
4531
msr_info->data = 0;
arch/x86/kvm/x86.c
4537
msr_info->data = vcpu->arch.st.msr_val;
arch/x86/kvm/x86.c
4543
msr_info->data = vcpu->arch.pv_eoi.msr_val;
arch/x86/kvm/x86.c
4549
msr_info->data = vcpu->arch.msr_kvm_poll_control;
arch/x86/kvm/x86.c
4558
return get_msr_mce(vcpu, msr_info->index, &msr_info->data,
arch/x86/kvm/x86.c
4559
msr_info->host_initiated);
arch/x86/kvm/x86.c
4561
if (!msr_info->host_initiated &&
arch/x86/kvm/x86.c
4564
msr_info->data = vcpu->arch.ia32_xss;
arch/x86/kvm/x86.c
4576
msr_info->data = 0x20000000;
arch/x86/kvm/x86.c
4590
msr_info->index, &msr_info->data,
arch/x86/kvm/x86.c
4591
msr_info->host_initiated);
arch/x86/kvm/x86.c
4604
msr_info->data = 0xbe702111;
arch/x86/kvm/x86.c
4609
msr_info->data = vcpu->arch.osvw.length;
arch/x86/kvm/x86.c
4614
msr_info->data = vcpu->arch.osvw.status;
arch/x86/kvm/x86.c
4617
if (!msr_info->host_initiated &&
arch/x86/kvm/x86.c
4620
msr_info->data = vcpu->arch.msr_platform_info;
arch/x86/kvm/x86.c
4623
msr_info->data = vcpu->arch.msr_misc_features_enables;
arch/x86/kvm/x86.c
4626
msr_info->data = vcpu->arch.msr_hwcr;
arch/x86/kvm/x86.c
4630
if (!msr_info->host_initiated &&
arch/x86/kvm/x86.c
4634
msr_info->data = vcpu->arch.guest_fpu.fpstate->xfd;
arch/x86/kvm/x86.c
4637
if (!msr_info->host_initiated &&
arch/x86/kvm/x86.c
4641
msr_info->data = vcpu->arch.guest_fpu.xfd_err;
arch/x86/kvm/x86.c
4646
kvm_get_xstate_msr(vcpu, msr_info);
arch/x86/kvm/x86.c
4649
if (kvm_pmu_is_valid_msr(vcpu, msr_info->index))
arch/x86/kvm/x86.c
4650
return kvm_pmu_get_msr(vcpu, msr_info);
arch/x86/lib/msr-smp.c
10
struct msr_info *rv = info;
arch/x86/lib/msr-smp.c
101
struct msr_info rv;
arch/x86/lib/msr-smp.c
146
struct msr_info msr;
arch/x86/lib/msr-smp.c
162
struct msr_info *rv = info;
arch/x86/lib/msr-smp.c
194
struct msr_info rv;
arch/x86/lib/msr-smp.c
210
struct msr_info rv;
arch/x86/lib/msr-smp.c
23
struct msr_info *rv = info;
arch/x86/lib/msr-smp.c
37
struct msr_info rv;
arch/x86/lib/msr-smp.c
53
struct msr_info rv;
arch/x86/lib/msr-smp.c
68
struct msr_info rv;
arch/x86/lib/msr-smp.c
84
struct msr_info rv;