cpumap
static const BIN_ATTR_RO(cpumap, CPUMAP_FILE_MAX_BYTES);
if (pool->cpumap)
dma_free_coherent(ctlr->dev, pool->mem_size, pool->cpumap,
pool->cpumap = dma_alloc_coherent(ctlr->dev, pool->mem_size,
pool->iomap = (void __iomem __force *)pool->cpumap;
void *cpumap; /* dma_alloc map */
struct cpumap;
struct irq_matrix *matrix, struct cpumap *cmap),
struct irq_matrix *matrix, struct cpumap *cmap),
struct irq_matrix *matrix, struct cpumap *cmap),
struct irq_matrix *matrix, struct cpumap *cmap),
struct irq_matrix *matrix, struct cpumap *cmap),
struct irq_matrix *matrix, struct cpumap *cmap),
struct cpumap *cmap),
struct cpumap *cm = this_cpu_ptr(m->maps);
static unsigned int matrix_alloc_area(struct irq_matrix *m, struct cpumap *cm,
struct cpumap *cm;
struct cpumap *cm;
struct cpumap *cm = this_cpu_ptr(m->maps);
struct cpumap *cm = per_cpu_ptr(m->maps, cpu);
struct cpumap *cm = per_cpu_ptr(m->maps, cpu);
struct cpumap *cm;
struct cpumap __percpu *maps;
struct cpumap *cm = this_cpu_ptr(m->maps);
struct cpumap *cm;
struct cpumap *cm = per_cpu_ptr(m->maps, cpu);
struct cpumap *cm = this_cpu_ptr(m->maps);
struct cpumap *cm = this_cpu_ptr(m->maps);
struct cpumap *cm = per_cpu_ptr(m->maps, cpu);
struct cpumap *cm = per_cpu_ptr(m->maps, cpu);
struct cpumap *cm = this_cpu_ptr(m->maps);
static int set_tracing_cpumask(struct perf_cpu_map *cpumap)
last_cpu = perf_cpu_map__cpu(cpumap, perf_cpu_map__nr(cpumap) - 1).cpu;
cpu_map__snprint_mask(cpumap, cpumask, mask_size);
struct perf_cpu_map *cpumap = ftrace->evlist->core.user_requested_cpus;
return set_tracing_cpumask(cpumap);
struct perf_cpu_map *cpumap = perf_cpu_map__new_online_cpus();
ret = set_tracing_cpumask(cpumap);
perf_cpu_map__put(cpumap);
void cpu_map__set_affinity(const struct perf_cpu_map *cpumap)
perf_cpu_map__for_each_cpu_skip_any(cpu, idx, cpumap)
void cpu_map__set_affinity(const struct perf_cpu_map *cpumap);
struct bpf_cpu_map *cpumap = (struct bpf_cpu_map *)&m_cpumap;
VERIFY(check_default(&cpumap->map, map));