Symbol: cache_map
arch/sparc/kernel/perf_event.c
1201
if (!sparc_pmu->cache_map)
arch/sparc/kernel/perf_event.c
1216
pmap = &((*sparc_pmu->cache_map)[cache_type][cache_op][cache_result]);
arch/sparc/kernel/perf_event.c
159
const cache_map_t *cache_map;
arch/sparc/kernel/perf_event.c
323
.cache_map = &ultra3_cache_map,
arch/sparc/kernel/perf_event.c
461
.cache_map = &niagara1_cache_map,
arch/sparc/kernel/perf_event.c
596
.cache_map = &niagara2_cache_map,
arch/sparc/kernel/perf_event.c
753
.cache_map = &niagara4_cache_map,
arch/sparc/kernel/perf_event.c
783
.cache_map = &niagara4_cache_map,
arch/x86/kernel/cpu/mtrr/generic.c
169
memmove(cache_map + idx, cache_map + idx + 1,
arch/x86/kernel/cpu/mtrr/generic.c
170
sizeof(*cache_map) * (cache_map_n - idx));
arch/x86/kernel/cpu/mtrr/generic.c
192
struct cache_map *prev = cache_map + idx - 1;
arch/x86/kernel/cpu/mtrr/generic.c
199
struct cache_map *next = cache_map + idx;
arch/x86/kernel/cpu/mtrr/generic.c
206
cache_map[idx - 1].end = cache_map[idx].end;
arch/x86/kernel/cpu/mtrr/generic.c
211
cache_map[idx - 1].end = end;
arch/x86/kernel/cpu/mtrr/generic.c
215
cache_map[idx].start = start;
arch/x86/kernel/cpu/mtrr/generic.c
227
memmove(cache_map + idx + 1, cache_map + idx,
arch/x86/kernel/cpu/mtrr/generic.c
228
sizeof(*cache_map) * (cache_map_n - idx));
arch/x86/kernel/cpu/mtrr/generic.c
231
cache_map[idx].start = start;
arch/x86/kernel/cpu/mtrr/generic.c
232
cache_map[idx].end = end;
arch/x86/kernel/cpu/mtrr/generic.c
233
cache_map[idx].type = type;
arch/x86/kernel/cpu/mtrr/generic.c
234
cache_map[idx].fixed = 0;
arch/x86/kernel/cpu/mtrr/generic.c
243
int ret = start != cache_map[idx].start;
arch/x86/kernel/cpu/mtrr/generic.c
246
if (start == cache_map[idx].start && end == cache_map[idx].end) {
arch/x86/kernel/cpu/mtrr/generic.c
248
} else if (start == cache_map[idx].start) {
arch/x86/kernel/cpu/mtrr/generic.c
249
cache_map[idx].start = end;
arch/x86/kernel/cpu/mtrr/generic.c
250
} else if (end == cache_map[idx].end) {
arch/x86/kernel/cpu/mtrr/generic.c
251
cache_map[idx].end = start;
arch/x86/kernel/cpu/mtrr/generic.c
253
tmp = cache_map[idx].end;
arch/x86/kernel/cpu/mtrr/generic.c
254
cache_map[idx].end = start;
arch/x86/kernel/cpu/mtrr/generic.c
255
add_map_entry_at(end, tmp, cache_map[idx].type, idx + 1);
arch/x86/kernel/cpu/mtrr/generic.c
273
if (start >= cache_map[i].end)
arch/x86/kernel/cpu/mtrr/generic.c
276
if (start < cache_map[i].start) {
arch/x86/kernel/cpu/mtrr/generic.c
278
tmp = min(end, cache_map[i].start);
arch/x86/kernel/cpu/mtrr/generic.c
284
new_type = get_effective_type(type, cache_map[i].type);
arch/x86/kernel/cpu/mtrr/generic.c
285
old_type = cache_map[i].type;
arch/x86/kernel/cpu/mtrr/generic.c
287
if (cache_map[i].fixed || new_type == old_type) {
arch/x86/kernel/cpu/mtrr/generic.c
289
start = cache_map[i].end;
arch/x86/kernel/cpu/mtrr/generic.c
294
tmp = min(end, cache_map[i].end);
arch/x86/kernel/cpu/mtrr/generic.c
318
cache_map[cache_map_n - 1].fixed = 1;
arch/x86/kernel/cpu/mtrr/generic.c
383
cache_map[i].fixed = 1;
arch/x86/kernel/cpu/mtrr/generic.c
395
cache_map[i].start, cache_map[i].end - 1,
arch/x86/kernel/cpu/mtrr/generic.c
396
mtrr_attrib_to_str(cache_map[i].type));
arch/x86/kernel/cpu/mtrr/generic.c
407
cache_map = NULL;
arch/x86/kernel/cpu/mtrr/generic.c
413
cache_map = kzalloc_objs(*cache_map, new_size);
arch/x86/kernel/cpu/mtrr/generic.c
414
if (cache_map) {
arch/x86/kernel/cpu/mtrr/generic.c
415
memmove(cache_map, init_cache_map,
arch/x86/kernel/cpu/mtrr/generic.c
416
cache_map_n * sizeof(*cache_map));
arch/x86/kernel/cpu/mtrr/generic.c
531
if (start >= cache_map[i].end)
arch/x86/kernel/cpu/mtrr/generic.c
535
if (start < cache_map[i].start) {
arch/x86/kernel/cpu/mtrr/generic.c
539
if (end <= cache_map[i].start)
arch/x86/kernel/cpu/mtrr/generic.c
544
type = type_merge(type, cache_map[i].type, uniform);
arch/x86/kernel/cpu/mtrr/generic.c
546
start = cache_map[i].end;
arch/x86/kernel/cpu/mtrr/generic.c
81
static struct cache_map init_cache_map[CACHE_MAP_MAX] __initdata;
arch/x86/kernel/cpu/mtrr/generic.c
82
static struct cache_map *cache_map __refdata = init_cache_map;
drivers/md/dm-cache-target.c
3517
.map = cache_map,
drivers/perf/arm_pmu.c
125
armpmu_map_cache_event(const unsigned (*cache_map)
drivers/perf/arm_pmu.c
145
if (!cache_map)
drivers/perf/arm_pmu.c
148
ret = (int)(*cache_map)[cache_type][cache_op][cache_result];
drivers/perf/arm_pmu.c
180
const unsigned (*cache_map)
drivers/perf/arm_pmu.c
196
return armpmu_map_cache_event(cache_map, config);
include/linux/perf/arm_pmu.h
142
const unsigned (*cache_map)[PERF_COUNT_HW_CACHE_MAX]