PERF_COUNT_HW_MAX
#define PERF_COUNT_ARC_DCLM (PERF_COUNT_HW_MAX + 0)
#define PERF_COUNT_ARC_DCSM (PERF_COUNT_HW_MAX + 1)
#define PERF_COUNT_ARC_ICM (PERF_COUNT_HW_MAX + 2)
#define PERF_COUNT_ARC_BPOK (PERF_COUNT_HW_MAX + 3)
#define PERF_COUNT_ARC_EDTLB (PERF_COUNT_HW_MAX + 4)
#define PERF_COUNT_ARC_EITLB (PERF_COUNT_HW_MAX + 5)
#define PERF_COUNT_ARC_LDC (PERF_COUNT_HW_MAX + 6)
#define PERF_COUNT_ARC_STC (PERF_COUNT_HW_MAX + 7)
#define PERF_COUNT_ARC_HW_MAX (PERF_COUNT_HW_MAX + 8)
if (event->attr.config >= PERF_COUNT_HW_MAX)
static const int csky_pmu_hw_map[PERF_COUNT_HW_MAX] = {
if (event->attr.config >= PERF_COUNT_HW_MAX)
[0 ... PERF_COUNT_HW_MAX - 1] = {HW_OP_UNSUPPORTED}
const struct loongarch_perf_event (*general_event_map)[PERF_COUNT_HW_MAX];
static const struct loongarch_perf_event loongson_event_map[PERF_COUNT_HW_MAX] = {
if (event->attr.config >= PERF_COUNT_HW_MAX)
if (event->attr.config >= PERF_COUNT_HW_MAX)
const struct mips_perf_event (*general_event_map)[PERF_COUNT_HW_MAX];
[PERF_COUNT_HW_MAX] = {
[PERF_COUNT_HW_MAX] = {
static const struct mips_perf_event i6x00_event_map[PERF_COUNT_HW_MAX] = {
static const struct mips_perf_event loongson3_event_map1[PERF_COUNT_HW_MAX] = {
static const struct mips_perf_event loongson3_event_map2[PERF_COUNT_HW_MAX] = {
static const struct mips_perf_event loongson3_event_map3[PERF_COUNT_HW_MAX] = {
static const struct mips_perf_event octeon_event_map[PERF_COUNT_HW_MAX] = {
[PERF_COUNT_HW_MAX] = {
static const u64 amd_perfmon_event_map[PERF_COUNT_HW_MAX] =
static const u64 amd_zen1_perfmon_event_map[PERF_COUNT_HW_MAX] =
static const u64 amd_zen2_perfmon_event_map[PERF_COUNT_HW_MAX] =
static const u64 amd_zen4_perfmon_event_map[PERF_COUNT_HW_MAX] =
static u64 intel_perfmon_event_map[PERF_COUNT_HW_MAX] __read_mostly =
static u64 p4_general_events[PERF_COUNT_HW_MAX] = {
static u64 zx_pmon_event_map[PERF_COUNT_HW_MAX] __read_mostly = {
static const unsigned m1_pmu_perf_map[PERF_COUNT_HW_MAX] = {
armpmu_map_hw_event(const unsigned (*event_map)[PERF_COUNT_HW_MAX], u64 config)
if (config >= PERF_COUNT_HW_MAX)
const unsigned (*event_map)[PERF_COUNT_HW_MAX],
[PERF_COUNT_HW_MAX],
static const unsigned armv8_pmuv3_perf_map[PERF_COUNT_HW_MAX] = {
static const unsigned armv6_perf_map[PERF_COUNT_HW_MAX] = {
static const unsigned armv7_a8_perf_map[PERF_COUNT_HW_MAX] = {
static const unsigned armv7_a9_perf_map[PERF_COUNT_HW_MAX] = {
static const unsigned armv7_a5_perf_map[PERF_COUNT_HW_MAX] = {
static const unsigned armv7_a15_perf_map[PERF_COUNT_HW_MAX] = {
static const unsigned armv7_a7_perf_map[PERF_COUNT_HW_MAX] = {
static const unsigned armv7_a12_perf_map[PERF_COUNT_HW_MAX] = {
static const unsigned krait_perf_map[PERF_COUNT_HW_MAX] = {
static const unsigned krait_perf_map_no_branch[PERF_COUNT_HW_MAX] = {
static const unsigned scorpion_perf_map[PERF_COUNT_HW_MAX] = {
static const unsigned xscale_perf_map[PERF_COUNT_HW_MAX] = {
if (config >= PERF_COUNT_HW_MAX)
const unsigned (*event_map)[PERF_COUNT_HW_MAX],
[0 ... PERF_COUNT_HW_MAX - 1] = HW_OP_UNSUPPORTED
const char *event_symbols_hw[PERF_COUNT_HW_MAX] = {
if (event < PERF_COUNT_HW_MAX && evsel__hw_names[event])
err = perf_evsel__name_array_test(evsel__hw_names, PERF_COUNT_HW_MAX);
evsel->alternate_hw_config = PERF_COUNT_HW_MAX;
counter->alternate_hw_config != PERF_COUNT_HW_MAX) {
const char *const evsel__hw_names[PERF_COUNT_HW_MAX] = {
if (config < PERF_COUNT_HW_MAX && evsel__hw_names[config])
extern const char *const evsel__hw_names[PERF_COUNT_HW_MAX];
if (term->val.num >= PERF_COUNT_HW_MAX) {
cpus, /*alternate_hw_config=*/PERF_COUNT_HW_MAX) ? 0 : -ENOMEM;
u64 alternate_hw_config = PERF_COUNT_HW_MAX;
/*alternate_hw_config=*/PERF_COUNT_HW_MAX);
&config_terms, /*alternate_hw_config=*/PERF_COUNT_HW_MAX);
assert(term->val.num < PERF_COUNT_HW_MAX);