Symbol: kvm_pmu_ops
arch/x86/include/asm/kvm_host.h
1994
struct kvm_pmu_ops *pmu_ops;
arch/x86/include/asm/kvm_host.h
624
struct kvm_pmu_ops;
arch/x86/kvm/pmu.c
100
WARN_ON(!kvm_pmu_ops.func); __KVM_X86_PMU_OP(func)
arch/x86/kvm/pmu.c
106
void kvm_init_pmu_capability(struct kvm_pmu_ops *pmu_ops)
arch/x86/kvm/pmu.c
1145
u64 mask = kvm_pmu_ops.EVENTSEL_EVENT |
arch/x86/kvm/pmu.c
1170
if (filter->events[i] & ~(kvm_pmu_ops.EVENTSEL_EVENT |
arch/x86/kvm/pmu.c
1288
return kvm_pmu_ops.FIXED_COUNTER_BASE + idx * kvm_pmu_ops.MSR_STRIDE;
arch/x86/kvm/pmu.c
1293
return kvm_pmu_ops.GP_COUNTER_BASE + idx * kvm_pmu_ops.MSR_STRIDE;
arch/x86/kvm/pmu.c
1298
return kvm_pmu_ops.GP_EVENTSEL_BASE + idx * kvm_pmu_ops.MSR_STRIDE;
arch/x86/kvm/pmu.c
1350
wrmsrq(kvm_pmu_ops.PERF_GLOBAL_CTRL, 0);
arch/x86/kvm/pmu.c
452
u64 event_select = eventsel & kvm_pmu_ops.EVENTSEL_EVENT;
arch/x86/kvm/pmu.c
684
if (!kvm_pmu_ops.check_rdpmc_early)
arch/x86/kvm/pmu.c
85
static struct kvm_pmu_ops kvm_pmu_ops __read_mostly;
arch/x86/kvm/pmu.c
89
*(((struct kvm_pmu_ops *)0)->func));
arch/x86/kvm/pmu.c
93
void kvm_pmu_ops_update(const struct kvm_pmu_ops *pmu_ops)
arch/x86/kvm/pmu.c
95
memcpy(&kvm_pmu_ops, pmu_ops, sizeof(kvm_pmu_ops));
arch/x86/kvm/pmu.c
98
static_call_update(kvm_x86_pmu_##func, kvm_pmu_ops.func);
arch/x86/kvm/pmu.h
198
void kvm_init_pmu_capability(struct kvm_pmu_ops *pmu_ops);
arch/x86/kvm/pmu.h
268
extern struct kvm_pmu_ops intel_pmu_ops;
arch/x86/kvm/pmu.h
269
extern struct kvm_pmu_ops amd_pmu_ops;
arch/x86/kvm/pmu.h
56
void kvm_pmu_ops_update(const struct kvm_pmu_ops *pmu_ops);
arch/x86/kvm/svm/pmu.c
263
struct kvm_pmu_ops amd_pmu_ops __initdata = {
arch/x86/kvm/vmx/pmu_intel.c
824
struct kvm_pmu_ops intel_pmu_ops __initdata = {