Symbol: kvm_pmc
arch/arm64/kvm/pmu-emul.c
107
static bool kvm_pmu_counter_can_chain(struct kvm_pmc *pmc)
arch/arm64/kvm/pmu-emul.c
123
static u64 kvm_pmc_read_evtreg(const struct kvm_pmc *pmc)
arch/arm64/kvm/pmu-emul.c
128
static u64 kvm_pmu_get_pmc_value(struct kvm_pmc *pmc)
arch/arm64/kvm/pmu-emul.c
1319
struct kvm_pmc *pmc = kvm_vcpu_idx_to_pmc(vcpu, i);
arch/arm64/kvm/pmu-emul.c
160
static void kvm_pmu_set_pmc_value(struct kvm_pmc *pmc, u64 val, bool force)
arch/arm64/kvm/pmu-emul.c
215
static void kvm_pmu_release_perf_event(struct kvm_pmc *pmc)
arch/arm64/kvm/pmu-emul.c
23
static void kvm_pmu_create_perf_event(struct kvm_pmc *pmc);
arch/arm64/kvm/pmu-emul.c
230
static void kvm_pmu_stop_counter(struct kvm_pmc *pmc)
arch/arm64/kvm/pmu-emul.c
24
static void kvm_pmu_release_perf_event(struct kvm_pmc *pmc);
arch/arm64/kvm/pmu-emul.c
25
static bool kvm_pmu_counter_is_enabled(struct kvm_pmc *pmc);
arch/arm64/kvm/pmu-emul.c
328
static void kvm_pmc_enable_perf_event(struct kvm_pmc *pmc)
arch/arm64/kvm/pmu-emul.c
33
static struct kvm_vcpu *kvm_pmc_to_vcpu(const struct kvm_pmc *pmc)
arch/arm64/kvm/pmu-emul.c
340
static void kvm_pmc_disable_perf_event(struct kvm_pmc *pmc)
arch/arm64/kvm/pmu-emul.c
354
struct kvm_pmc *pmc = kvm_vcpu_idx_to_pmc(vcpu, i);
arch/arm64/kvm/pmu-emul.c
38
static struct kvm_pmc *kvm_vcpu_idx_to_pmc(struct kvm_vcpu *vcpu, int cnt_idx)
arch/arm64/kvm/pmu-emul.c
493
struct kvm_pmc *pmc = kvm_vcpu_idx_to_pmc(vcpu, i);
arch/arm64/kvm/pmu-emul.c
522
static u64 compute_period(struct kvm_pmc *pmc, u64 counter)
arch/arm64/kvm/pmu-emul.c
541
struct kvm_pmc *pmc = perf_event->overflow_handler_context;
arch/arm64/kvm/pmu-emul.c
622
static bool kvm_pmu_counter_is_enabled(struct kvm_pmc *pmc)
arch/arm64/kvm/pmu-emul.c
636
static bool kvm_pmc_counts_at_el0(struct kvm_pmc *pmc)
arch/arm64/kvm/pmu-emul.c
645
static bool kvm_pmc_counts_at_el1(struct kvm_pmc *pmc)
arch/arm64/kvm/pmu-emul.c
654
static bool kvm_pmc_counts_at_el2(struct kvm_pmc *pmc)
arch/arm64/kvm/pmu-emul.c
683
static void kvm_pmu_create_perf_event(struct kvm_pmc *pmc)
arch/arm64/kvm/pmu-emul.c
778
struct kvm_pmc *pmc = kvm_vcpu_idx_to_pmc(vcpu, select_idx);
arch/arm64/kvm/pmu-emul.c
87
static bool kvm_pmc_is_64bit(struct kvm_pmc *pmc)
arch/arm64/kvm/pmu-emul.c
95
static bool kvm_pmc_has_64bit_overflow(struct kvm_pmc *pmc)
arch/riscv/include/asm/kvm_vcpu_pmu.h
45
struct kvm_pmc pmc[RISCV_KVM_MAX_COUNTERS];
arch/riscv/kvm/vcpu_pmu.c
210
struct kvm_pmc *pmc;
arch/riscv/kvm/vcpu_pmu.c
241
struct kvm_pmc *pmc;
arch/riscv/kvm/vcpu_pmu.c
280
struct kvm_pmc *pmc = perf_event->overflow_handler_context;
arch/riscv/kvm/vcpu_pmu.c
317
static long kvm_pmu_create_perf_event(struct kvm_pmc *pmc, struct perf_event_attr *attr,
arch/riscv/kvm/vcpu_pmu.c
38
static u64 kvm_pmu_get_sample_period(struct kvm_pmc *pmc)
arch/riscv/kvm/vcpu_pmu.c
544
struct kvm_pmc *pmc;
arch/riscv/kvm/vcpu_pmu.c
622
struct kvm_pmc *pmc;
arch/riscv/kvm/vcpu_pmu.c
724
struct kvm_pmc *pmc = NULL;
arch/riscv/kvm/vcpu_pmu.c
80
static void kvm_pmu_release_perf_event(struct kvm_pmc *pmc)
arch/riscv/kvm/vcpu_pmu.c
823
struct kvm_pmc *pmc;
arch/riscv/kvm/vcpu_pmu.c
890
struct kvm_pmc *pmc;
arch/x86/include/asm/kvm_host.h
578
struct kvm_pmc gp_counters[KVM_MAX_NR_GP_COUNTERS];
arch/x86/include/asm/kvm_host.h
579
struct kvm_pmc fixed_counters[KVM_MAX_NR_FIXED_COUNTERS];
arch/x86/kvm/pmu.c
1010
struct kvm_pmc *pmc = NULL;
arch/x86/kvm/pmu.c
1034
static bool pmc_is_pmi_enabled(struct kvm_pmc *pmc)
arch/x86/kvm/pmu.c
1046
static void kvm_pmu_incr_counter(struct kvm_pmc *pmc)
arch/x86/kvm/pmu.c
1075
static inline bool cpl_is_matched(struct kvm_pmc *pmc)
arch/x86/kvm/pmu.c
1107
struct kvm_pmc *pmc;
arch/x86/kvm/pmu.c
1304
struct kvm_pmc *pmc;
arch/x86/kvm/pmu.c
1362
struct kvm_pmc *pmc;
arch/x86/kvm/pmu.c
173
static inline void __kvm_perf_overflow(struct kvm_pmc *pmc, bool in_pmi)
arch/x86/kvm/pmu.c
204
struct kvm_pmc *pmc = perf_event->overflow_handler_context;
arch/x86/kvm/pmu.c
219
static u64 pmc_get_pebs_precise_level(struct kvm_pmc *pmc)
arch/x86/kvm/pmu.c
241
static u64 get_sample_period(struct kvm_pmc *pmc, u64 counter_value)
arch/x86/kvm/pmu.c
250
static int pmc_reprogram_counter(struct kvm_pmc *pmc, u32 type, u64 config,
arch/x86/kvm/pmu.c
304
static bool pmc_pause_counter(struct kvm_pmc *pmc)
arch/x86/kvm/pmu.c
330
static bool pmc_resume_counter(struct kvm_pmc *pmc)
arch/x86/kvm/pmu.c
352
static void pmc_release_perf_event(struct kvm_pmc *pmc)
arch/x86/kvm/pmu.c
362
static void pmc_stop_counter(struct kvm_pmc *pmc)
arch/x86/kvm/pmu.c
370
static void pmc_update_sample_period(struct kvm_pmc *pmc)
arch/x86/kvm/pmu.c
380
void pmc_write_counter(struct kvm_pmc *pmc, u64 val)
arch/x86/kvm/pmu.c
508
static bool pmc_is_event_allowed(struct kvm_pmc *pmc)
arch/x86/kvm/pmu.c
523
static void kvm_mediated_pmu_refresh_event_filter(struct kvm_pmc *pmc)
arch/x86/kvm/pmu.c
542
static int reprogram_counter(struct kvm_pmc *pmc)
arch/x86/kvm/pmu.c
593
static bool pmc_is_event_match(struct kvm_pmc *pmc, u64 eventsel)
arch/x86/kvm/pmu.c
612
void kvm_pmu_recalc_pmc_emulation(struct kvm_pmu *pmu, struct kvm_pmc *pmc)
arch/x86/kvm/pmu.c
638
struct kvm_pmc *pmc;
arch/x86/kvm/pmu.c
727
struct kvm_pmc *pmc;
arch/x86/kvm/pmu.c
814
struct kvm_pmc *pmc = kvm_pmu_call(msr_idx_to_pmc)(vcpu, msr);
arch/x86/kvm/pmu.c
918
struct kvm_pmc *pmc;
arch/x86/kvm/pmu.h
111
static inline u64 pmc_bitmask(struct kvm_pmc *pmc)
arch/x86/kvm/pmu.h
118
static inline u64 pmc_read_counter(struct kvm_pmc *pmc)
arch/x86/kvm/pmu.h
134
void pmc_write_counter(struct kvm_pmc *pmc, u64 val);
arch/x86/kvm/pmu.h
136
static inline bool pmc_is_gp(struct kvm_pmc *pmc)
arch/x86/kvm/pmu.h
141
static inline bool pmc_is_fixed(struct kvm_pmc *pmc)
arch/x86/kvm/pmu.h
156
static inline struct kvm_pmc *get_gp_pmc(struct kvm_pmu *pmu, u32 msr,
arch/x86/kvm/pmu.h
170
static inline struct kvm_pmc *get_fixed_pmc(struct kvm_pmu *pmu, u32 msr)
arch/x86/kvm/pmu.h
184
static inline bool pmc_is_locally_enabled(struct kvm_pmc *pmc)
arch/x86/kvm/pmu.h
200
void kvm_pmu_recalc_pmc_emulation(struct kvm_pmu *pmu, struct kvm_pmc *pmc);
arch/x86/kvm/pmu.h
202
static inline void kvm_pmu_request_counter_reprogram(struct kvm_pmc *pmc)
arch/x86/kvm/pmu.h
227
static inline bool pmc_is_globally_enabled(struct kvm_pmc *pmc)
arch/x86/kvm/pmu.h
27
struct kvm_pmc *(*rdpmc_ecx_to_pmc)(struct kvm_vcpu *vcpu,
arch/x86/kvm/pmu.h
29
struct kvm_pmc *(*msr_idx_to_pmc)(struct kvm_vcpu *vcpu, u32 msr);
arch/x86/kvm/pmu.h
93
static inline struct kvm_pmc *kvm_pmc_idx_to_pmc(struct kvm_pmu *pmu, int idx)
arch/x86/kvm/svm/pmu.c
131
struct kvm_pmc *pmc;
arch/x86/kvm/svm/pmu.c
153
struct kvm_pmc *pmc;
arch/x86/kvm/svm/pmu.c
28
static struct kvm_pmc *amd_pmu_get_pmc(struct kvm_pmu *pmu, int pmc_idx)
arch/x86/kvm/svm/pmu.c
38
static inline struct kvm_pmc *get_gp_pmc_amd(struct kvm_pmu *pmu, u32 msr,
arch/x86/kvm/svm/pmu.c
87
static struct kvm_pmc *amd_rdpmc_ecx_to_pmc(struct kvm_vcpu *vcpu,
arch/x86/kvm/svm/pmu.c
93
static struct kvm_pmc *amd_msr_idx_to_pmc(struct kvm_vcpu *vcpu, u32 msr)
arch/x86/kvm/svm/pmu.c
96
struct kvm_pmc *pmc;
arch/x86/kvm/vmx/pmu_intel.c
132
static inline struct kvm_pmc *get_fw_gp_pmc(struct kvm_pmu *pmu, u32 msr)
arch/x86/kvm/vmx/pmu_intel.c
205
static struct kvm_pmc *intel_msr_idx_to_pmc(struct kvm_vcpu *vcpu, u32 msr)
arch/x86/kvm/vmx/pmu_intel.c
208
struct kvm_pmc *pmc;
arch/x86/kvm/vmx/pmu_intel.c
331
struct kvm_pmc *pmc;
arch/x86/kvm/vmx/pmu_intel.c
374
struct kvm_pmc *pmc;
arch/x86/kvm/vmx/pmu_intel.c
59
struct kvm_pmc *pmc;
arch/x86/kvm/vmx/pmu_intel.c
741
struct kvm_pmc *pmc = NULL;
arch/x86/kvm/vmx/pmu_intel.c
79
static struct kvm_pmc *intel_rdpmc_ecx_to_pmc(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/pmu_intel.c
84
struct kvm_pmc *counters;
include/kvm/arm_pmu.h
29
struct kvm_pmc pmc[KVM_ARMV8_PMU_MAX_COUNTERS];