hybrid_pmu
if (!(x86_pmu.hybrid_pmu[i].pmu_type & pmu_attr->pmu_type))
if (x86_pmu.hybrid_pmu[i].pmu_type & pmu->pmu_type) {
cntr_mask |= x86_pmu.hybrid_pmu[i].cntr_mask64;
struct x86_hybrid_pmu *hybrid_pmu;
hybrid_pmu = &x86_pmu.hybrid_pmu[i];
hybrid_pmu->pmu = pmu;
hybrid_pmu->pmu.type = -1;
hybrid_pmu->pmu.attr_update = x86_pmu.attr_update;
hybrid_pmu->pmu.capabilities |= PERF_PMU_CAP_EXTENDED_HW_TYPE;
err = perf_pmu_register(&hybrid_pmu->pmu, hybrid_pmu->name,
(hybrid_pmu->pmu_type == hybrid_big) ? PERF_TYPE_RAW : -1);
perf_pmu_unregister(&x86_pmu.hybrid_pmu[j].pmu);
kfree(x86_pmu.hybrid_pmu);
x86_pmu.hybrid_pmu = NULL;
h_pmu = hybrid_pmu(event_pmu);
pmu = hybrid_pmu(event->pmu);
return hybrid_pmu(event->pmu)->pmu_type == hybrid_big;
struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu);
struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu);
struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu);
struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu);
struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu);
enum hybrid_pmu_type pmu_type = x86_pmu.hybrid_pmu[i].pmu_type;
return &x86_pmu.hybrid_pmu[i];
return &x86_pmu.hybrid_pmu[i];
return &x86_pmu.hybrid_pmu[i];
return &x86_pmu.hybrid_pmu[i];
cpumask_clear_cpu(cpu, &hybrid_pmu(cpuc->pmu)->supported_cpus);
struct x86_hybrid_pmu *hpmu = hybrid_pmu(pmu);
x86_pmu.hybrid_pmu = kzalloc_objs(struct x86_hybrid_pmu,
if (!x86_pmu.hybrid_pmu)
pmu = &x86_pmu.hybrid_pmu[idx++];
pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_CORE_IDX];
pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_ATOM_IDX];
pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_CORE_IDX];
pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_ATOM_IDX];
pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_CORE_IDX];
pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_ATOM_IDX];
pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_CORE_IDX];
pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_ATOM_IDX];
pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_TINY_IDX];
pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_CORE_IDX];
pmu = &x86_pmu.hybrid_pmu[X86_HYBRID_PMU_ATOM_IDX];
data_source = x86_pmu.hybrid_pmu[X86_HYBRID_PMU_CORE_IDX].pebs_data_source;
data_source = x86_pmu.hybrid_pmu[X86_HYBRID_PMU_ATOM_IDX].pebs_data_source;
data_source = x86_pmu.hybrid_pmu[X86_HYBRID_PMU_CORE_IDX].pebs_data_source;
data_source = x86_pmu.hybrid_pmu[X86_HYBRID_PMU_ATOM_IDX].pebs_data_source;
data_source = x86_pmu.hybrid_pmu[X86_HYBRID_PMU_TINY_IDX].pebs_data_source;
data_source = x86_pmu.hybrid_pmu[X86_HYBRID_PMU_CORE_IDX].pebs_data_source;
data_source = x86_pmu.hybrid_pmu[X86_HYBRID_PMU_ATOM_IDX].pebs_data_source;
hybrid_pmu(event->pmu)->pmu_type == hybrid_big);
struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu);
struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu);
struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu);
struct x86_hybrid_pmu *hybrid_pmu;
__Fp = &hybrid_pmu(_pmu)->_field; \
__Fp = &hybrid_pmu(_pmu)->_var; \
__Fp = hybrid_pmu(_pmu)->_field; \