Symbol: hybrid_pmu
arch/x86/events/core.c
1961
if (!(x86_pmu.hybrid_pmu[i].pmu_type & pmu_attr->pmu_type))
arch/x86/events/core.c
1963
if (x86_pmu.hybrid_pmu[i].pmu_type & pmu->pmu_type) {
arch/x86/events/core.c
214
cntr_mask |= x86_pmu.hybrid_pmu[i].cntr_mask64;
arch/x86/events/core.c
2237
struct x86_hybrid_pmu *hybrid_pmu;
arch/x86/events/core.c
2241
hybrid_pmu = &x86_pmu.hybrid_pmu[i];
arch/x86/events/core.c
2243
hybrid_pmu->pmu = pmu;
arch/x86/events/core.c
2244
hybrid_pmu->pmu.type = -1;
arch/x86/events/core.c
2245
hybrid_pmu->pmu.attr_update = x86_pmu.attr_update;
arch/x86/events/core.c
2246
hybrid_pmu->pmu.capabilities |= PERF_PMU_CAP_EXTENDED_HW_TYPE;
arch/x86/events/core.c
2248
err = perf_pmu_register(&hybrid_pmu->pmu, hybrid_pmu->name,
arch/x86/events/core.c
2249
(hybrid_pmu->pmu_type == hybrid_big) ? PERF_TYPE_RAW : -1);
arch/x86/events/core.c
2256
perf_pmu_unregister(&x86_pmu.hybrid_pmu[j].pmu);
arch/x86/events/core.c
2258
kfree(x86_pmu.hybrid_pmu);
arch/x86/events/core.c
2259
x86_pmu.hybrid_pmu = NULL;
arch/x86/events/core.c
2403
h_pmu = hybrid_pmu(event_pmu);
arch/x86/events/core.c
2520
pmu = hybrid_pmu(event->pmu);
arch/x86/events/intel/core.c
4493
return hybrid_pmu(event->pmu)->pmu_type == hybrid_big;
arch/x86/events/intel/core.c
5306
struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu);
arch/x86/events/intel/core.c
5383
struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu);
arch/x86/events/intel/core.c
5396
struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu);
arch/x86/events/intel/core.c
5422
struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu);
arch/x86/events/intel/core.c
5432
struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu);
arch/x86/events/intel/core.c
6007
enum hybrid_pmu_type pmu_type = x86_pmu.hybrid_pmu[i].pmu_type;
arch/x86/events/intel/core.c
6011
return &x86_pmu.hybrid_pmu[i];
arch/x86/events/intel/core.c
6014
return &x86_pmu.hybrid_pmu[i];
arch/x86/events/intel/core.c
6018
return &x86_pmu.hybrid_pmu[i];
arch/x86/events/intel/core.c
6020
return &x86_pmu.hybrid_pmu[i];
arch/x86/events/intel/core.c
6197
cpumask_clear_cpu(cpu, &hybrid_pmu(cpuc->pmu)->supported_cpus);
arch/x86/events/intel/core.c
6230
struct x86_hybrid_pmu *hpmu = hybrid_pmu(pmu);
arch/x86/events/intel/core.c
7394
x86_pmu.hybrid_pmu = kzalloc_objs(struct x86_hybrid_pmu,
arch/x86/events/intel/core.c
7396
if (!x86_pmu.hybrid_pmu)
arch/x86/events/intel/core.c
7403
pmu = &x86_pmu.hybrid_pmu[idx++];
arch/x86/events/intel/core.c
8227
pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_CORE_IDX];
arch/x86/events/intel/core.c
8259
pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_ATOM_IDX];
arch/x86/events/intel/core.c
8284
pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_CORE_IDX];
arch/x86/events/intel/core.c
8289
pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_ATOM_IDX];
arch/x86/events/intel/core.c
8323
pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_CORE_IDX];
arch/x86/events/intel/core.c
8327
pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_ATOM_IDX];
arch/x86/events/intel/core.c
8347
pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_CORE_IDX];
arch/x86/events/intel/core.c
8351
pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_ATOM_IDX];
arch/x86/events/intel/core.c
8355
pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_TINY_IDX];
arch/x86/events/intel/core.c
8381
pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_CORE_IDX];
arch/x86/events/intel/core.c
8385
pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_ATOM_IDX];
arch/x86/events/intel/ds.c
184
data_source = x86_pmu.hybrid_pmu[X86_HYBRID_PMU_CORE_IDX].pebs_data_source;
arch/x86/events/intel/ds.c
188
data_source = x86_pmu.hybrid_pmu[X86_HYBRID_PMU_ATOM_IDX].pebs_data_source;
arch/x86/events/intel/ds.c
207
data_source = x86_pmu.hybrid_pmu[X86_HYBRID_PMU_CORE_IDX].pebs_data_source;
arch/x86/events/intel/ds.c
211
data_source = x86_pmu.hybrid_pmu[X86_HYBRID_PMU_ATOM_IDX].pebs_data_source;
arch/x86/events/intel/ds.c
222
data_source = x86_pmu.hybrid_pmu[X86_HYBRID_PMU_TINY_IDX].pebs_data_source;
arch/x86/events/intel/ds.c
258
data_source = x86_pmu.hybrid_pmu[X86_HYBRID_PMU_CORE_IDX].pebs_data_source;
arch/x86/events/intel/ds.c
261
data_source = x86_pmu.hybrid_pmu[X86_HYBRID_PMU_ATOM_IDX].pebs_data_source;
arch/x86/events/intel/ds.c
461
hybrid_pmu(event->pmu)->pmu_type == hybrid_big);
arch/x86/events/intel/ds.c
574
struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu);
arch/x86/events/intel/ds.c
584
struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu);
arch/x86/events/intel/ds.c
632
struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu);
arch/x86/events/perf_event.h
1045
struct x86_hybrid_pmu *hybrid_pmu;
arch/x86/events/perf_event.h
793
__Fp = &hybrid_pmu(_pmu)->_field; \
arch/x86/events/perf_event.h
803
__Fp = &hybrid_pmu(_pmu)->_var; \
arch/x86/events/perf_event.h
813
__Fp = hybrid_pmu(_pmu)->_field; \