MAX_NR_CPUS
__uint(max_entries, MAX_NR_CPUS);
static int inner_lru_map_fds[MAX_NR_CPUS];
assert(cpu < MAX_NR_CPUS);
WARN_ONCE(end_cpu >= MAX_NR_CPUS, "Perf can support %d CPUs. "
"Consider raising MAX_NR_CPUS\n", MAX_NR_CPUS);
DECLARE_BITMAP(cpu_bitmap, MAX_NR_CPUS);
static DECLARE_BITMAP(cpu_bitmap, MAX_NR_CPUS);
for (i = 0; i < MAX_NR_CPUS; i++) {
u64 idle_time = stat->cpus_runtime[MAX_NR_CPUS].idle;
u64 hardirq_time = stat->cpus_runtime[MAX_NR_CPUS].irq;
u64 softirq_time = stat->cpus_runtime[MAX_NR_CPUS].softirq;
int cpus_nr = bitmap_weight(stat->all_cpus_bitmap, MAX_NR_CPUS);
u64 cpus_total_time = stat->cpus_runtime[MAX_NR_CPUS].total;
BUG_ON(work->cpu >= MAX_NR_CPUS);
stat->cpus_runtime[MAX_NR_CPUS].total += work->total_runtime;
stat->cpus_runtime[MAX_NR_CPUS].idle += work->total_runtime;
stat->cpus_runtime[MAX_NR_CPUS].irq += work->total_runtime;
stat->cpus_runtime[MAX_NR_CPUS].softirq += work->total_runtime;
stat->cpus_runtime[MAX_NR_CPUS].load += work->total_runtime;
cpus_runtime = zalloc(sizeof(struct __top_cpus_runtime) * (MAX_NR_CPUS + 1));
bitmap_zero(kwork->top_stat.all_cpus_bitmap, MAX_NR_CPUS);
DECLARE_BITMAP(cpu_bitmap, MAX_NR_CPUS);
DECLARE_BITMAP(cpu_bitmap, MAX_NR_CPUS);
static DECLARE_BITMAP(cpu_bitmap, MAX_NR_CPUS);
static DECLARE_BITMAP(cpu_bitmap, MAX_NR_CPUS);
__uint(max_entries, MAX_NR_CPUS);
char dname[DNAME_LEN], cpumask[MAX_NR_CPUS];
char cpulist[MAX_NR_CPUS];
if (__bitmap_weight(symbol_conf.parallelism_filter, MAX_NR_CPUS + 1) == 0)
DECLARE_BITMAP(all_cpus_bitmap, MAX_NR_CPUS);
DECLARE_BITMAP(cpu_bitmap, MAX_NR_CPUS);
int nr_cpus = min(perf_session__env(session)->nr_cpus_avail, MAX_NR_CPUS);
typedef struct { DECLARE_BITMAP(bits, MAX_NR_CPUS); } cpumask_t;
nr_cpus = min(env->nr_cpus_online, MAX_NR_CPUS);
bitmap_fill(symbol_conf.parallelism_filter, MAX_NR_CPUS + 1);
if (cpu.cpu <= 0 || cpu.cpu > MAX_NR_CPUS) {
DECLARE_BITMAP(parallelism_filter, MAX_NR_CPUS + 1);
char cpus[MAX_NR_CPUS];