cache_op
unsigned int cache_type, cache_op, cache_result;
cache_op = (config >> 8) & 0xff;
if (cache_op >= PERF_COUNT_HW_CACHE_OP_MAX)
ret = arc_pmu_cache_map[cache_type][cache_op][cache_result];
cache_type, cache_op, cache_result, ret,
unsigned int cache_type, cache_op, cache_result;
cache_op = (config >> 8) & 0xff;
if (cache_op >= PERF_COUNT_HW_CACHE_OP_MAX)
return csky_pmu_cache_map[cache_type][cache_op][cache_result];
cache_op(page_to_phys(page), size, dma_wbinv_set_zero_range);
cache_op(paddr, size, dma_wb_range);
cache_op(paddr, size, dma_wbinv_range);
cache_op(paddr, size, dma_inv_range);
cache_op(Index_Writeback_Inv_LEAF0, addr);
cache_op(Index_Writeback_Inv_LEAF1, addr);
cache_op(Index_Writeback_Inv_LEAF2, addr);
cache_op(Index_Writeback_Inv_LEAF3, addr);
cache_op(Index_Writeback_Inv_LEAF4, addr);
cache_op(Index_Writeback_Inv_LEAF5, addr);
unsigned int cache_type, cache_op, cache_result;
cache_op = (config >> 8) & 0xff;
if (cache_op >= PERF_COUNT_HW_CACHE_OP_MAX)
[cache_op]
cache_op(Index_Store_Tag_S, ZSCM_REG_BASE + offset);
cache_op(Index_Load_Tag_S, ZSCM_REG_BASE + offset);
cache_op(Hit_Invalidate_SD, addr);
cache_op(Hit_Writeback_Inv_SD, addr);
cache_op(Page_Invalidate_T, addr);
prot##cache_op(hitop, addr); \
cache_op(Index_Invalidate_I, addr);
cache_op(Index_Writeback_Inv_D, addr);
cache_op(Index_Writeback_Inv_SD, addr);
cache_op(Hit_Invalidate_I_Loongson2, addr);
cache_op(Hit_Invalidate_I, addr);
cache_op(Hit_Writeback_Inv_D, addr);
cache_op(Hit_Invalidate_D, addr);
cache_op(Index_Load_Tag_D, INDEX_8);
cache_op(Index_Store_Tag_D, INDEX_8);
cache_op(Index_Store_Tag_D, INDEX_0);
unsigned int cache_type, cache_op, cache_result;
cache_op = (config >> 8) & 0xff;
if (cache_op >= PERF_COUNT_HW_CACHE_OP_MAX)
[cache_op]
cache_op(Index_Load_Tag_D, CKSEG0 | offset);
cache_op(Index_Store_Tag_I, CKSEG0|offset);
cache_op(Index_Load_Tag_I, CKSEG0 | offset);
cache_op(Index_Store_Tag_D, CKSEG0 | offset);
case cache_op:
cache_op(Index_Store_Tag_I, begin);
cache_op(Index_Store_Tag_D, begin);
cache_op(Index_Store_Tag_SD, begin);
cache_op(Index_Load_Tag_SD, addr);
cache_op(Hit_Writeback_Inv_SD, addr & almask);
cache_op(Hit_Writeback_Inv_SD, (addr + size - 1) & almask);
cache_op(R5K_Page_Invalidate_S, start);
cache_op(R5K_Page_Invalidate_S, a);
cache_op(Index_Store_Tag_T, CKSEG0ADDR(i));
cache_op(Index_Store_Tag_SD, CKSEG0ADDR(i));
cache_op(Index_Store_Tag_T, begin);
cache_op(Index_Load_Tag_T, addr);
cache_op(Page_Invalidate_T, start);
[insn_cache] = {M(cache_op, 0, 0, 0, 0, 0), RS | RT | SIMM},
cache_op(Index_Load_Tag_I, addr);
cache_op(Index_Load_Tag_I, addr | 1L);
cache_op(Index_Load_Tag_S, addr);
cache_op(Index_Load_Tag_S, addr | 1L);
cache_op(Index_Load_Tag_D, addr);
cache_op(Index_Load_Tag_D, addr | 1L);
cache_op(Index_Writeback_Inv_D, addr | 0);
cache_op(Index_Writeback_Inv_D, addr | 1);
cache_op(Index_Writeback_Inv_D, addr | 2);
cache_op(Index_Writeback_Inv_D, addr | 3);
cache_op = (sbi_event_code & SBI_PMU_EVENT_CACHE_OP_ID_CODE_MASK) >>
cache_op >= PERF_COUNT_HW_CACHE_OP_MAX ||
config = cache_type | (cache_op << 8) | (cache_result << 16);
unsigned int cache_type, cache_op, cache_result;
unsigned int cache_type, cache_op, cache_result;
cache_op = (config >> 8) & 0xff;
if (cache_op >= PERF_COUNT_HW_CACHE_OP_MAX)
pmap = &((*sparc_pmu->cache_map)[cache_type][cache_op][cache_result]);
unsigned int cache_type, cache_op, cache_result;
cache_op = (config >> 8) & 0xff;
if (cache_op >= PERF_COUNT_HW_CACHE_OP_MAX)
cache_op = array_index_nospec(cache_op, PERF_COUNT_HW_CACHE_OP_MAX);
val = hybrid_var(event->pmu, hw_cache_event_ids)[cache_type][cache_op][cache_result];
attr->config1 = hybrid_var(event->pmu, hw_cache_extra_regs)[cache_type][cache_op][cache_result];
unsigned int cache_type, cache_op, cache_result;
cache_op = (config >> 8) & 0xff;
cache_op >= C(OP_MAX) ||
ret = xtensa_cache_ctl[cache_type][cache_op][cache_result];
unsigned int cache_type, cache_op, cache_result, ret;
cache_op = (config >> 8) & 0xff;
if (cache_op >= PERF_COUNT_HW_CACHE_OP_MAX)
ret = (int)(*cache_map)[cache_type][cache_op][cache_result];
unsigned int cache_type, cache_op, cache_result, ret;
cache_op = (config >> 8) & 0xff;
if (cache_op >= PERF_COUNT_HW_CACHE_OP_MAX)
ret = pmu_cache_event_map[cache_type][cache_op][cache_result].event_idx;
int cache_type, cache_op, cache_result;
cache_op = (term->val.num >> 8) & 0xFF;
cache_op >= PERF_COUNT_HW_CACHE_OP_MAX ||
int len, cache_type = -1, cache_op = -1, cache_result = -1;
cache_op = parse_aliases(str, evsel__hw_cache_op,
if (cache_op >= 0) {
if (!evsel__is_cache_op_valid(cache_type, cache_op))
if (cache_op < 0) {
cache_op = parse_aliases(str, evsel__hw_cache_op,
if (cache_op >= 0) {
if (!evsel__is_cache_op_valid(cache_type, cache_op))
if (cache_op == -1)
cache_op = PERF_COUNT_HW_CACHE_OP_READ;
*config = cache_type | (cache_op << 8) | (cache_result << 16);
int cache_op = (term->val.num >> 8) & 0xFF;
assert(cache_op < PERF_COUNT_HW_CACHE_OP_MAX);