kvm_pmc
static bool kvm_pmu_counter_can_chain(struct kvm_pmc *pmc)
static u64 kvm_pmc_read_evtreg(const struct kvm_pmc *pmc)
static u64 kvm_pmu_get_pmc_value(struct kvm_pmc *pmc)
struct kvm_pmc *pmc = kvm_vcpu_idx_to_pmc(vcpu, i);
static void kvm_pmu_set_pmc_value(struct kvm_pmc *pmc, u64 val, bool force)
static void kvm_pmu_release_perf_event(struct kvm_pmc *pmc)
static void kvm_pmu_create_perf_event(struct kvm_pmc *pmc);
static void kvm_pmu_stop_counter(struct kvm_pmc *pmc)
static void kvm_pmu_release_perf_event(struct kvm_pmc *pmc);
static bool kvm_pmu_counter_is_enabled(struct kvm_pmc *pmc);
static void kvm_pmc_enable_perf_event(struct kvm_pmc *pmc)
static struct kvm_vcpu *kvm_pmc_to_vcpu(const struct kvm_pmc *pmc)
static void kvm_pmc_disable_perf_event(struct kvm_pmc *pmc)
struct kvm_pmc *pmc = kvm_vcpu_idx_to_pmc(vcpu, i);
static struct kvm_pmc *kvm_vcpu_idx_to_pmc(struct kvm_vcpu *vcpu, int cnt_idx)
struct kvm_pmc *pmc = kvm_vcpu_idx_to_pmc(vcpu, i);
static u64 compute_period(struct kvm_pmc *pmc, u64 counter)
struct kvm_pmc *pmc = perf_event->overflow_handler_context;
static bool kvm_pmu_counter_is_enabled(struct kvm_pmc *pmc)
static bool kvm_pmc_counts_at_el0(struct kvm_pmc *pmc)
static bool kvm_pmc_counts_at_el1(struct kvm_pmc *pmc)
static bool kvm_pmc_counts_at_el2(struct kvm_pmc *pmc)
static void kvm_pmu_create_perf_event(struct kvm_pmc *pmc)
struct kvm_pmc *pmc = kvm_vcpu_idx_to_pmc(vcpu, select_idx);
static bool kvm_pmc_is_64bit(struct kvm_pmc *pmc)
static bool kvm_pmc_has_64bit_overflow(struct kvm_pmc *pmc)
struct kvm_pmc pmc[RISCV_KVM_MAX_COUNTERS];
struct kvm_pmc *pmc;
struct kvm_pmc *pmc;
struct kvm_pmc *pmc = perf_event->overflow_handler_context;
static long kvm_pmu_create_perf_event(struct kvm_pmc *pmc, struct perf_event_attr *attr,
static u64 kvm_pmu_get_sample_period(struct kvm_pmc *pmc)
struct kvm_pmc *pmc;
struct kvm_pmc *pmc;
struct kvm_pmc *pmc = NULL;
static void kvm_pmu_release_perf_event(struct kvm_pmc *pmc)
struct kvm_pmc *pmc;
struct kvm_pmc *pmc;
struct kvm_pmc gp_counters[KVM_MAX_NR_GP_COUNTERS];
struct kvm_pmc fixed_counters[KVM_MAX_NR_FIXED_COUNTERS];
struct kvm_pmc *pmc = NULL;
static bool pmc_is_pmi_enabled(struct kvm_pmc *pmc)
static void kvm_pmu_incr_counter(struct kvm_pmc *pmc)
static inline bool cpl_is_matched(struct kvm_pmc *pmc)
struct kvm_pmc *pmc;
struct kvm_pmc *pmc;
struct kvm_pmc *pmc;
static inline void __kvm_perf_overflow(struct kvm_pmc *pmc, bool in_pmi)
struct kvm_pmc *pmc = perf_event->overflow_handler_context;
static u64 pmc_get_pebs_precise_level(struct kvm_pmc *pmc)
static u64 get_sample_period(struct kvm_pmc *pmc, u64 counter_value)
static int pmc_reprogram_counter(struct kvm_pmc *pmc, u32 type, u64 config,
static bool pmc_pause_counter(struct kvm_pmc *pmc)
static bool pmc_resume_counter(struct kvm_pmc *pmc)
static void pmc_release_perf_event(struct kvm_pmc *pmc)
static void pmc_stop_counter(struct kvm_pmc *pmc)
static void pmc_update_sample_period(struct kvm_pmc *pmc)
void pmc_write_counter(struct kvm_pmc *pmc, u64 val)
static bool pmc_is_event_allowed(struct kvm_pmc *pmc)
static void kvm_mediated_pmu_refresh_event_filter(struct kvm_pmc *pmc)
static int reprogram_counter(struct kvm_pmc *pmc)
static bool pmc_is_event_match(struct kvm_pmc *pmc, u64 eventsel)
void kvm_pmu_recalc_pmc_emulation(struct kvm_pmu *pmu, struct kvm_pmc *pmc)
struct kvm_pmc *pmc;
struct kvm_pmc *pmc;
struct kvm_pmc *pmc = kvm_pmu_call(msr_idx_to_pmc)(vcpu, msr);
struct kvm_pmc *pmc;
static inline u64 pmc_bitmask(struct kvm_pmc *pmc)
static inline u64 pmc_read_counter(struct kvm_pmc *pmc)
void pmc_write_counter(struct kvm_pmc *pmc, u64 val);
static inline bool pmc_is_gp(struct kvm_pmc *pmc)
static inline bool pmc_is_fixed(struct kvm_pmc *pmc)
static inline struct kvm_pmc *get_gp_pmc(struct kvm_pmu *pmu, u32 msr,
static inline struct kvm_pmc *get_fixed_pmc(struct kvm_pmu *pmu, u32 msr)
static inline bool pmc_is_locally_enabled(struct kvm_pmc *pmc)
void kvm_pmu_recalc_pmc_emulation(struct kvm_pmu *pmu, struct kvm_pmc *pmc);
static inline void kvm_pmu_request_counter_reprogram(struct kvm_pmc *pmc)
static inline bool pmc_is_globally_enabled(struct kvm_pmc *pmc)
struct kvm_pmc *(*rdpmc_ecx_to_pmc)(struct kvm_vcpu *vcpu,
struct kvm_pmc *(*msr_idx_to_pmc)(struct kvm_vcpu *vcpu, u32 msr);
static inline struct kvm_pmc *kvm_pmc_idx_to_pmc(struct kvm_pmu *pmu, int idx)
struct kvm_pmc *pmc;
struct kvm_pmc *pmc;
static struct kvm_pmc *amd_pmu_get_pmc(struct kvm_pmu *pmu, int pmc_idx)
static inline struct kvm_pmc *get_gp_pmc_amd(struct kvm_pmu *pmu, u32 msr,
static struct kvm_pmc *amd_rdpmc_ecx_to_pmc(struct kvm_vcpu *vcpu,
static struct kvm_pmc *amd_msr_idx_to_pmc(struct kvm_vcpu *vcpu, u32 msr)
struct kvm_pmc *pmc;
static inline struct kvm_pmc *get_fw_gp_pmc(struct kvm_pmu *pmu, u32 msr)
static struct kvm_pmc *intel_msr_idx_to_pmc(struct kvm_vcpu *vcpu, u32 msr)
struct kvm_pmc *pmc;
struct kvm_pmc *pmc;
struct kvm_pmc *pmc;
struct kvm_pmc *pmc;
struct kvm_pmc *pmc = NULL;
static struct kvm_pmc *intel_rdpmc_ecx_to_pmc(struct kvm_vcpu *vcpu,
struct kvm_pmc *counters;
struct kvm_pmc pmc[KVM_ARMV8_PMU_MAX_COUNTERS];