Symbol: NR_CPUS
arch/alpha/include/asm/smp.h
41
extern struct cpuinfo_alpha cpu_data[NR_CPUS];
arch/alpha/kernel/core_t2.c
413
for (i = 0; i < NR_CPUS; i++) {
arch/alpha/kernel/irq.c
61
cpu = (cpu < (NR_CPUS-1) ? cpu + 1 : 0);
arch/alpha/kernel/smp.c
247
for (i = 0; i < NR_CPUS; i++) {
arch/alpha/kernel/smp.c
485
for(cpu = 0; cpu < NR_CPUS; cpu++)
arch/alpha/kernel/smp.c
55
struct cpuinfo_alpha cpu_data[NR_CPUS];
arch/alpha/kernel/smp.c
61
} ipi_data[NR_CPUS] __cacheline_aligned;
arch/alpha/kernel/smp.c
643
for (cpu = 0; cpu < NR_CPUS; cpu++) {
arch/alpha/kernel/smp.c
690
for (cpu = 0; cpu < NR_CPUS; cpu++) {
arch/alpha/kernel/smp.c
744
for (cpu = 0; cpu < NR_CPUS; cpu++) {
arch/alpha/mm/tlbflush.c
102
for (cpu = 0; cpu < NR_CPUS; cpu++) {
arch/arc/include/asm/highmem.h
20
#define FIX_KMAP_SLOTS (KM_MAX_IDX * NR_CPUS)
arch/arc/include/asm/mmu.h
14
unsigned long asid[NR_CPUS]; /* 8 bit MMU PID + Generation cycle */
arch/arc/kernel/setup.c
45
struct task_struct *_current_task[NR_CPUS]; /* For stack switching */
arch/arc/kernel/smp.c
67
NR_CPUS);
arch/arm/common/bL_switcher.c
138
static int bL_switcher_cpu_pairing[NR_CPUS];
arch/arm/common/bL_switcher.c
268
static struct bL_thread bL_threads[NR_CPUS];
arch/arm/common/bL_switcher.c
379
static unsigned int bL_switcher_cpu_original_cluster[NR_CPUS];
arch/arm/include/asm/fixmap.h
17
FIX_KMAP_END = FIX_KMAP_BEGIN + (KM_MAX_IDX * NR_CPUS) - 1,
arch/arm/kernel/cpuidle.c
15
static struct cpuidle_ops cpuidle_ops[NR_CPUS] __ro_after_init;
arch/arm/kernel/devtree.c
78
u32 tmp_map[NR_CPUS] = { [0 ... NR_CPUS-1] = MPIDR_INVALID };
arch/arm/kernel/setup.c
110
struct processor *cpu_vtable[NR_CPUS] = {
arch/arm/kernel/setup.c
144
static struct stack stacks[NR_CPUS];
arch/arm/kernel/setup.c
537
if (cpu >= NR_CPUS) {
arch/arm/kernel/setup.c
594
u32 __cpu_logical_map[NR_CPUS] = { [0 ... NR_CPUS-1] = MPIDR_INVALID };
arch/arm/mach-imx/platsmp.c
108
for (i = ncores; i < NR_CPUS; i++)
arch/arm/mach-imx/platsmp.c
60
for (i = ncores; i < NR_CPUS; i++)
arch/arm/mach-omap2/omap-secure.c
62
static u32 buf[NR_CPUS][5];
arch/arm/mach-shmobile/platsmp-apmu.c
29
} apmu_cpus[NR_CPUS];
arch/arm/vfp/vfpmodule.c
111
thread->vfpstate.hard.cpu = NR_CPUS;
arch/arm/vfp/vfpmodule.c
142
vfp->hard.cpu = NR_CPUS;
arch/arm/vfp/vfpmodule.c
164
thread->vfpstate.hard.cpu = NR_CPUS;
arch/arm/vfp/vfpmodule.c
56
union vfp_state *vfp_current_hw_state[NR_CPUS];
arch/arm64/include/asm/fixmap.h
54
FIX_VNCR = FIX_VNCR_END + NR_CPUS,
arch/arm64/include/asm/kvm_host.h
812
extern u64 kvm_nvhe_sym(hyp_cpu_logical_map)[NR_CPUS];
arch/arm64/include/asm/kvm_nested.h
405
BUG_ON(__c >= NR_CPUS); \
arch/arm64/include/asm/memory.h
439
# define INIT_MEMBLOCK_RESERVED_REGIONS (INIT_MEMBLOCK_REGIONS + NR_CPUS + 1)
arch/arm64/include/asm/processor.h
299
.fpsimd_cpu = NR_CPUS, \
arch/arm64/include/asm/smp.h
37
extern u64 __cpu_logical_map[NR_CPUS];
arch/arm64/kernel/acpi_numa.c
30
static int acpi_early_node_map[NR_CPUS] __initdata = { [0 ... NR_CPUS - 1] = NUMA_NO_NODE };
arch/arm64/kernel/acpi_parking_protocol.c
27
static struct cpu_mailbox_entry cpu_mailbox_entries[NR_CPUS];
arch/arm64/kernel/cpu_ops.c
23
static const struct cpu_operations *cpu_ops[NR_CPUS] __ro_after_init;
arch/arm64/kernel/fpsimd.c
1752
t->thread.fpsimd_cpu = NR_CPUS;
arch/arm64/kernel/setup.c
274
u64 __cpu_logical_map[NR_CPUS] = { [0 ... NR_CPUS-1] = INVALID_HWID };
arch/arm64/kernel/smp.c
478
for (i = 1; (i < cpu) && (i < NR_CPUS); i++)
arch/arm64/kernel/smp.c
561
static struct acpi_madt_generic_interrupt cpu_madt_gicc[NR_CPUS];
arch/arm64/kernel/smp.c
608
if (cpu_count >= NR_CPUS)
arch/arm64/kernel/smp.c
722
if (cpu_count >= NR_CPUS)
arch/arm64/kernel/smp_spin_table.c
25
static phys_addr_t cpu_release_addr[NR_CPUS];
arch/arm64/kvm/hyp/nvhe/hyp-smp.c
17
u64 __ro_after_init hyp_cpu_logical_map[NR_CPUS] = { [0 ... NR_CPUS-1] = INVALID_HWID };
arch/arm64/kvm/hyp/nvhe/hyp-smp.c
26
unsigned long __ro_after_init kvm_arm_hyp_percpu_base[NR_CPUS];
arch/arm64/kvm/hyp/nvhe/psci-relay.c
86
for (i = 0; i < NR_CPUS; i++) {
arch/csky/include/asm/fixmap.h
19
FIX_KMAP_END = FIX_KMAP_BEGIN + (KM_MAX_IDX * NR_CPUS) - 1,
arch/csky/kernel/process.c
17
struct cpuinfo_csky cpu_data[NR_CPUS];
arch/csky/kernel/smp.c
186
if (cpu >= NR_CPUS)
arch/hexagon/kernel/smp.c
243
for (i = 0; i < NR_CPUS; i++)
arch/loongarch/include/asm/bootinfo.h
27
#define NR_WORDS DIV_ROUND_UP(NR_CPUS, BITS_PER_LONG)
arch/loongarch/include/asm/irq.h
69
#define NR_IRQS (64 + NR_VECTORS * (NR_CPUS + MAX_IO_PICS))
arch/loongarch/include/asm/mmu.h
12
u64 asid[NR_CPUS];
arch/loongarch/include/asm/setup.h
18
extern unsigned long pcpu_handlers[NR_CPUS];
arch/loongarch/include/asm/smp.h
61
extern int __cpu_number_map[NR_CPUS];
arch/loongarch/include/asm/smp.h
65
extern int __cpu_logical_map[NR_CPUS];
arch/loongarch/include/asm/sparsemem.h
24
#define INIT_MEMBLOCK_RESERVED_REGIONS (INIT_MEMBLOCK_REGIONS + NR_CPUS)
arch/loongarch/include/asm/vdso/arch_data.h
20
struct vdso_pcpu_data pdata[NR_CPUS];
arch/loongarch/kernel/acpi.c
158
for (i = 0; i < NR_CPUS; i++) {
arch/loongarch/kernel/acpi.c
67
if (num_processors >= NR_CPUS) {
arch/loongarch/kernel/acpi.c
69
" processor 0x%x ignored.\n", NR_CPUS, cpuid);
arch/loongarch/kernel/acpi.c
81
cpu = find_first_zero_bit(cpumask_bits(cpu_present_mask), NR_CPUS);
arch/loongarch/kernel/acpi.c
87
cpu = find_first_zero_bit(cpumask_bits(cpu_possible_mask), NR_CPUS);
arch/loongarch/kernel/cpu-probe.c
351
const char *__cpu_family[NR_CPUS];
arch/loongarch/kernel/cpu-probe.c
352
const char *__cpu_full_name[NR_CPUS];
arch/loongarch/kernel/env.c
56
for (cpu = 0; cpu < NR_CPUS; cpu++)
arch/loongarch/kernel/hw_breakpoint.c
536
for (cpu = 1; cpu < NR_CPUS; cpu++) {
arch/loongarch/kernel/irq.c
92
int nr_io_pics = bitmap_weight(loongson_sysconf.cores_io_master, NR_CPUS);
arch/loongarch/kernel/numa.c
231
for (i = 0; i < NR_CPUS; i++)
arch/loongarch/kernel/numa.c
44
unsigned long __per_cpu_offset[NR_CPUS] __read_mostly;
arch/loongarch/kernel/setup.c
583
for (; i < NR_CPUS; i++) {
arch/loongarch/kernel/setup.c
64
struct cpuinfo_loongarch cpu_data[NR_CPUS] __read_mostly;
arch/loongarch/kernel/smp.c
333
cpu = find_first_zero_bit(cpumask_bits(cpu_present_mask), NR_CPUS);
arch/loongarch/kernel/smp.c
39
int __cpu_number_map[NR_CPUS]; /* Map physical to logical */
arch/loongarch/kernel/smp.c
42
int __cpu_logical_map[NR_CPUS]; /* Map logical to physical */
arch/loongarch/kernel/smp.c
46
cpumask_t cpu_sibling_map[NR_CPUS] __read_mostly;
arch/loongarch/kernel/smp.c
50
cpumask_t cpu_llc_shared_map[NR_CPUS] __read_mostly;
arch/loongarch/kernel/smp.c
54
cpumask_t cpu_core_map[NR_CPUS] __read_mostly;
arch/loongarch/kernel/smp.c
64
cpumask_t cpu_foreign_map[NR_CPUS] __read_mostly;
arch/loongarch/mm/tlb.c
263
unsigned long pcpu_handlers[NR_CPUS];
arch/m68k/kernel/setup_no.c
200
return *pos < NR_CPUS ? ((void *) 0x12345678) : NULL;
arch/microblaze/kernel/cpu/mb.c
140
return i < NR_CPUS ? (void *) (i + 1) : NULL;
arch/mips/cavium-octeon/smp.c
153
for (id = 0; id < NR_CPUS; id++) {
arch/mips/cavium-octeon/smp.c
163
for (id = 0; id < NR_CPUS; id++) {
arch/mips/cavium-octeon/smp.c
180
id < num_cores && id < NR_CPUS; id++) {
arch/mips/include/asm/fixmap.h
55
FIX_KMAP_END = FIX_KMAP_BEGIN + (KM_MAX_IDX * NR_CPUS) - 1,
arch/mips/include/asm/irq.h
22
extern void *irq_stack[NR_CPUS];
arch/mips/include/asm/kvm_host.h
351
u32 vzguestid[NR_CPUS];
arch/mips/include/asm/mach-ip27/topology.h
14
extern struct cpuinfo_ip27 sn_cpu_info[NR_CPUS];
arch/mips/include/asm/mmu.h
11
u64 asid[NR_CPUS];
arch/mips/include/asm/smp.h
44
extern int __cpu_logical_map[NR_CPUS];
arch/mips/kernel/cpu-probe.c
1843
const char *__cpu_name[NR_CPUS];
arch/mips/kernel/cpu-r3k-probe.c
64
const char *__cpu_name[NR_CPUS];
arch/mips/kernel/irq.c
29
void *irq_stack[NR_CPUS];
arch/mips/kernel/setup.c
54
struct cpuinfo_mips cpu_data[NR_CPUS] __read_mostly;
arch/mips/kernel/setup.c
733
for (; i < NR_CPUS; i++)
arch/mips/kernel/setup.c
789
unsigned long kernelsp[NR_CPUS];
arch/mips/kernel/smp-cps.c
241
for (v = 0; v < min_t(int, core_vpes, NR_CPUS - nvpes); v++) {
arch/mips/kernel/smp-cps.c
255
for (v = 0; v < min_t(unsigned, nvpes, NR_CPUS); v++) {
arch/mips/kernel/smp-cps.c
655
for (remote = 0; remote < NR_CPUS; remote++) {
arch/mips/kernel/smp-cps.c
661
if (remote >= NR_CPUS) {
arch/mips/kernel/smp.c
44
int __cpu_logical_map[NR_CPUS]; /* Map logical to physical */
arch/mips/kernel/smp.c
52
cpumask_t cpu_sibling_map[NR_CPUS] __read_mostly;
arch/mips/kernel/smp.c
56
cpumask_t cpu_core_map[NR_CPUS] __read_mostly;
arch/mips/kernel/smp.c
68
cpumask_t cpu_foreign_map[NR_CPUS] __read_mostly;
arch/mips/kvm/vz.c
39
static struct kvm_vcpu *last_vcpu[NR_CPUS];
arch/mips/kvm/vz.c
41
static struct kvm_vcpu *last_exec_vcpu[NR_CPUS];
arch/mips/lantiq/irq.c
65
static void __iomem *ltq_icu_membase[NR_CPUS];
arch/mips/loongson64/env.c
263
if (ecpu->nr_cpus > NR_CPUS || ecpu->nr_cpus == 0)
arch/mips/loongson64/env.c
264
loongson_sysconf.nr_cpus = NR_CPUS;
arch/mips/mm/c-octeon.c
29
unsigned long long cache_err_dcache[NR_CPUS];
arch/mips/mm/init.c
507
unsigned long __per_cpu_offset[NR_CPUS] __read_mostly;
arch/mips/mm/init.c
544
unsigned long pgd_current[NR_CPUS];
arch/mips/mm/tlbex.c
73
static struct tlb_reg_save handler_reg_save[NR_CPUS];
arch/mips/sgi-ip27/ip27-init.c
42
struct cpuinfo_ip27 sn_cpu_info[NR_CPUS];
arch/mips/sgi-ip27/ip27-smp.c
44
(cpus_found != NR_CPUS)) {
arch/mips/sibyte/bcm1480/irq.c
163
for (i=0; i<NR_CPUS; i++) {
arch/mips/sibyte/bcm1480/smp.c
134
for (i = 1, num = 0; i < NR_CPUS; i++) {
arch/mips/sibyte/sb1250/irq.c
144
for (i=0; i<NR_CPUS; i++) {
arch/mips/sibyte/sb1250/smp.c
123
for (i = 1, num = 0; i < NR_CPUS; i++) {
arch/openrisc/include/asm/cpuinfo.h
37
extern struct cpuinfo_or1k cpuinfo_or1k[NR_CPUS];
arch/openrisc/kernel/process.c
52
struct thread_info *current_thread_info_set[NR_CPUS] = { &init_thread_info, };
arch/openrisc/kernel/setup.c
94
struct cpuinfo_or1k cpuinfo_or1k[NR_CPUS];
arch/openrisc/kernel/smp.c
77
if (cpu_id < NR_CPUS)
arch/openrisc/mm/fault.c
32
volatile pgd_t *current_pgd[NR_CPUS];
arch/parisc/kernel/processor.c
146
if (cpu_info.cpu_num >= NR_CPUS) {
arch/parisc/kernel/processor.c
150
&dev->hpa.start, cpu_info.cpu_num, NR_CPUS);
arch/powerpc/include/asm/fixmap.h
52
FIX_KMAP_END = FIX_KMAP_BEGIN + (KM_MAX_IDX * NR_CPUS) - 1,
arch/powerpc/include/asm/irq.h
41
extern void *critirq_ctx[NR_CPUS];
arch/powerpc/include/asm/irq.h
42
extern void *dbgirq_ctx[NR_CPUS];
arch/powerpc/include/asm/irq.h
43
extern void *mcheckirq_ctx[NR_CPUS];
arch/powerpc/include/asm/irq.h
49
extern void *hardirq_ctx[NR_CPUS];
arch/powerpc/include/asm/irq.h
50
extern void *softirq_ctx[NR_CPUS];
arch/powerpc/include/asm/kvm_book3s_64.h
34
short prev_cpu[NR_CPUS];
arch/powerpc/include/asm/kvm_host.h
32
#define KVM_MAX_VCPUS NR_CPUS
arch/powerpc/include/asm/kvm_host.h
33
#define KVM_MAX_VCORES NR_CPUS
arch/powerpc/kernel/head_64.S
408
#if (NR_CPUS == 1) || defined(CONFIG_FORCE_NR_CPUS)
arch/powerpc/kernel/interrupt.c
25
unsigned long global_dbcr0[NR_CPUS];
arch/powerpc/kernel/irq.c
337
void *critirq_ctx[NR_CPUS] __read_mostly;
arch/powerpc/kernel/irq.c
338
void *dbgirq_ctx[NR_CPUS] __read_mostly;
arch/powerpc/kernel/irq.c
339
void *mcheckirq_ctx[NR_CPUS] __read_mostly;
arch/powerpc/kernel/irq.c
342
void *softirq_ctx[NR_CPUS] __read_mostly;
arch/powerpc/kernel/irq.c
343
void *hardirq_ctx[NR_CPUS] __read_mostly;
arch/powerpc/kernel/prom_init.c
1108
.max_cpus = cpu_to_be32(NR_CPUS), /* number of cores supported */
arch/powerpc/kernel/prom_init.c
1419
cores = DIV_ROUND_UP(NR_CPUS, prom_count_smt_threads());
arch/powerpc/kernel/prom_init.c
1421
cores, NR_CPUS);
arch/powerpc/kernel/setup_32.c
162
void *emergency_ctx[NR_CPUS] __ro_after_init = {[0] = &init_stack};
arch/powerpc/kernel/setup_32.c
59
int smp_hw_index[NR_CPUS];
arch/powerpc/kernel/setup_64.c
828
unsigned long __per_cpu_offset[NR_CPUS] __read_mostly;
arch/powerpc/kernel/smp.c
115
static struct thread_groups_list tgl[NR_CPUS] __initdata;
arch/powerpc/kernel/smp.c
139
volatile unsigned int cpu_callin_map[NR_CPUS];
arch/powerpc/kernel/smp.c
701
static struct task_struct *current_set[NR_CPUS];
arch/powerpc/kernel/tau_6xx.c
41
} tau[NR_CPUS];
arch/powerpc/kernel/traps.c
270
IS_ENABLED(CONFIG_SMP) ? (" NR_CPUS=" __stringify(NR_CPUS)) : "",
arch/powerpc/kvm/book3s_hv_nested.c
329
if (l2_hv.vcpu_token >= NR_CPUS)
arch/powerpc/lib/locks.c
29
BUG_ON(holder_cpu >= NR_CPUS);
arch/powerpc/lib/locks.c
55
BUG_ON(holder_cpu >= NR_CPUS);
arch/powerpc/mm/book3s64/hash_utils.c
1339
tmp = memblock_phys_alloc_range(sizeof(struct stress_hpt_struct) * NR_CPUS,
arch/powerpc/mm/book3s64/hash_utils.c
1342
memset((void *)tmp, 0xff, sizeof(struct stress_hpt_struct) * NR_CPUS);
arch/powerpc/mm/nohash/mmu_context.c
375
init_mm.context.active = NR_CPUS;
arch/powerpc/mm/nohash/mmu_context.c
74
static unsigned long *stale_map[NR_CPUS];
arch/powerpc/mm/numa.c
44
int numa_cpu_lookup_table[NR_CPUS];
arch/powerpc/platforms/86xx/mpc86xx_smp.c
40
if (nr < 0 || nr >= NR_CPUS)
arch/powerpc/platforms/86xx/mpc86xx_smp.c
64
if (nr < 0 || nr >= NR_CPUS)
arch/powerpc/platforms/cell/spu_base.c
97
int nr = (NR_CPUS > 1) ? NR_CPUS : NR_CPUS + 1;
arch/powerpc/platforms/powermac/smp.c
1016
for (cpu = 1; cpu < 4 && cpu < NR_CPUS; ++cpu)
arch/powerpc/platforms/powermac/smp.c
323
if (ncpus > NR_CPUS)
arch/powerpc/platforms/powermac/smp.c
324
ncpus = NR_CPUS;
arch/powerpc/platforms/powermac/smp.c
341
extern volatile unsigned int cpu_callin_map[NR_CPUS];
arch/powerpc/platforms/pseries/lpar.c
171
#define NR_CPUS_H NR_CPUS
arch/powerpc/sysdev/mpic.c
632
for (i = 0; i < min(32, NR_CPUS) && cpu_possible(i); ++i, cpumask >>= 1)
arch/powerpc/sysdev/xics/icp-native.c
46
static struct icp_ipl __iomem *icp_native_regs[NR_CPUS];
arch/powerpc/xmon/xmon.c
1339
last_cpu = first_cpu = NR_CPUS;
arch/powerpc/xmon/xmon.c
96
static long *xmon_fault_jmp[NR_CPUS];
arch/riscv/include/asm/cacheflush.h
44
extern u64 new_vmalloc[NR_CPUS / sizeof(u64) + 1];
arch/riscv/include/asm/cpufeature.h
37
extern struct riscv_isainfo hart_isa[NR_CPUS];
arch/riscv/include/asm/smp.h
25
extern unsigned long __cpuid_to_hartid_map[NR_CPUS];
arch/riscv/include/asm/vendor_extensions.h
27
struct riscv_isavendorinfo per_hart_isa_bitmap[NR_CPUS];
arch/riscv/kernel/acpi.c
36
static struct acpi_madt_rintc cpu_madt_rintc[NR_CPUS];
arch/riscv/kernel/acpi_numa.c
31
static int acpi_early_node_map[NR_CPUS] __initdata = { [0 ... NR_CPUS - 1] = NUMA_NO_NODE };
arch/riscv/kernel/cpu_ops_sbi.c
24
static struct sbi_hart_boot_data boot_data[NR_CPUS];
arch/riscv/kernel/cpu_ops_spinwait.c
17
void *__cpu_spinwait_stack_pointer[NR_CPUS] __section(".data");
arch/riscv/kernel/cpu_ops_spinwait.c
18
void *__cpu_spinwait_task_pointer[NR_CPUS] __section(".data");
arch/riscv/kernel/cpu_ops_spinwait.c
32
if (hartid == INVALID_HARTID || hartid >= (unsigned long) NR_CPUS)
arch/riscv/kernel/cpufeature.c
45
struct riscv_isainfo hart_isa[NR_CPUS];
arch/riscv/kernel/smp.c
54
unsigned long __cpuid_to_hartid_map[NR_CPUS] __ro_after_init = {
arch/riscv/kernel/smp.c
55
[0 ... NR_CPUS-1] = INVALID_HARTID
arch/riscv/kernel/smp.c
75
for (i = 0; i < NR_CPUS; i++)
arch/riscv/kernel/smpboot.c
104
if (cpu_count >= NR_CPUS) {
arch/riscv/kernel/smpboot.c
142
if (cpuid >= NR_CPUS) {
arch/riscv/mm/init.c
40
u64 new_vmalloc[NR_CPUS / sizeof(u64) + 1];
arch/s390/include/asm/abs_lowcore.h
8
#define ABS_LOWCORE_MAP_SIZE (NR_CPUS * sizeof(struct lowcore))
arch/s390/include/asm/topology.h
27
extern struct cpu_topology_s390 cpu_topology[NR_CPUS];
arch/s390/kernel/setup.c
179
struct lowcore *lowcore_ptr[NR_CPUS];
arch/s390/kernel/topology.c
64
struct cpu_topology_s390 cpu_topology[NR_CPUS];
arch/s390/kernel/vmcore_info.c
15
VMCOREINFO_LENGTH(lowcore_ptr, NR_CPUS);
arch/s390/mm/init.c
175
unsigned long __per_cpu_offset[NR_CPUS] __read_mostly;
arch/sh/boards/of-generic.c
69
if (id < NR_CPUS) {
arch/sh/include/asm/fixmap.h
51
FIX_CMAP_END = FIX_CMAP_BEGIN + (FIX_N_COLOURS * NR_CPUS) - 1,
arch/sh/include/asm/smp.h
18
extern int __cpu_number_map[NR_CPUS];
arch/sh/include/asm/smp.h
22
extern int __cpu_logical_map[NR_CPUS];
arch/sh/include/asm/topology.h
22
extern cpumask_t cpu_core_map[NR_CPUS];
arch/sh/kernel/cpu/sh2/smp-j2.c
73
for (i=max; i<NR_CPUS; i++) {
arch/sh/kernel/cpu/sh4/sq.c
257
static struct kobject *sq_kobject[NR_CPUS];
arch/sh/kernel/cpu/sh4a/smp-shx3.c
60
for (i = 1, num = 0; i < NR_CPUS; i++) {
arch/sh/kernel/irq.c
66
static union irq_ctx *hardirq_ctx[NR_CPUS] __read_mostly;
arch/sh/kernel/irq.c
67
static union irq_ctx *softirq_ctx[NR_CPUS] __read_mostly;
arch/sh/kernel/irq.c
69
static char softirq_stack[NR_CPUS * THREAD_SIZE] __page_aligned_bss;
arch/sh/kernel/irq.c
70
static char hardirq_stack[NR_CPUS * THREAD_SIZE] __page_aligned_bss;
arch/sh/kernel/setup.c
54
struct sh_cpuinfo cpu_data[NR_CPUS] __read_mostly = {
arch/sh/kernel/smp.c
33
int __cpu_number_map[NR_CPUS]; /* Map physical to logical */
arch/sh/kernel/smp.c
34
int __cpu_logical_map[NR_CPUS]; /* Map logical to physical */
arch/sh/kernel/topology.c
18
cpumask_t cpu_core_map[NR_CPUS];
arch/sparc/include/asm/irq_64.h
93
extern void *hardirq_stack[NR_CPUS];
arch/sparc/include/asm/irq_64.h
94
extern void *softirq_stack[NR_CPUS];
arch/sparc/include/asm/ptrace.h
61
extern union global_cpu_snapshot global_cpu_snapshot[NR_CPUS];
arch/sparc/include/asm/smp_32.h
32
extern volatile unsigned long cpu_callin_map[NR_CPUS];
arch/sparc/include/asm/smp_64.h
35
extern cpumask_t cpu_core_map[NR_CPUS];
arch/sparc/include/asm/switch_to_32.h
7
extern struct thread_info *current_set[NR_CPUS];
arch/sparc/include/asm/topology_64.h
53
extern cpumask_t cpu_core_map[NR_CPUS];
arch/sparc/include/asm/topology_64.h
54
extern cpumask_t cpu_core_sib_map[NR_CPUS];
arch/sparc/include/asm/topology_64.h
55
extern cpumask_t cpu_core_sib_cache_map[NR_CPUS];
arch/sparc/include/asm/trap_block.h
56
extern struct trap_per_cpu trap_block[NR_CPUS];
arch/sparc/include/asm/trap_block.h
60
extern u64 cpu_mondo_counter[NR_CPUS];
arch/sparc/include/asm/vaddrs.h
41
FIX_KMAP_END = (KM_MAX_IDX * NR_CPUS),
arch/sparc/kernel/cpumap.c
59
static u16 cpu_distribution_map[NR_CPUS];
arch/sparc/kernel/irq_64.c
1037
BUILD_BUG_ON((NR_CPUS * sizeof(u16)) > PAGE_SIZE);
arch/sparc/kernel/irq_64.c
807
void *hardirq_stack[NR_CPUS];
arch/sparc/kernel/irq_64.c
808
void *softirq_stack[NR_CPUS];
arch/sparc/kernel/leon_smp.c
167
(unsigned int)nrcpu, (unsigned int)NR_CPUS,
arch/sparc/kernel/leon_smp.c
237
for (i = 0; i < NR_CPUS; i++) {
arch/sparc/kernel/leon_smp.c
368
unsigned long processors_in[NR_CPUS]; /* Set when ipi entered. */
arch/sparc/kernel/leon_smp.c
369
unsigned long processors_out[NR_CPUS]; /* Set when ipi exited. */
arch/sparc/kernel/leon_smp.c
380
register int high = NR_CPUS - 1;
arch/sparc/kernel/leon_smp.c
54
extern volatile unsigned long cpu_callin_map[NR_CPUS];
arch/sparc/kernel/mdesc.c
1027
if (*id < NR_CPUS)
arch/sparc/kernel/mdesc.c
1118
if (cpuid >= NR_CPUS) {
arch/sparc/kernel/mdesc.c
1121
cpuid, NR_CPUS);
arch/sparc/kernel/mdesc.c
805
if (max_cpu > NR_CPUS)
arch/sparc/kernel/mdesc.c
806
max_cpu = NR_CPUS;
arch/sparc/kernel/mdesc.c
808
max_cpu = NR_CPUS;
arch/sparc/kernel/process_32.c
67
struct thread_info *current_set[NR_CPUS];
arch/sparc/kernel/process_64.c
188
union global_cpu_snapshot global_cpu_snapshot[NR_CPUS];
arch/sparc/kernel/prom_64.c
457
if (cpuid >= NR_CPUS) {
arch/sparc/kernel/prom_64.c
460
cpuid, NR_CPUS);
arch/sparc/kernel/setup_64.c
362
if (cpu >= NR_CPUS) {
arch/sparc/kernel/setup_64.c
364
cpu, NR_CPUS);
arch/sparc/kernel/smp_32.c
185
if (cpuid >= NR_CPUS)
arch/sparc/kernel/smp_32.c
229
if (mid < NR_CPUS) {
arch/sparc/kernel/smp_32.c
241
if (cpuid >= NR_CPUS) {
arch/sparc/kernel/smp_32.c
42
volatile unsigned long cpu_callin_map[NR_CPUS] = {0,};
arch/sparc/kernel/smp_64.c
62
cpumask_t cpu_core_map[NR_CPUS] __read_mostly =
arch/sparc/kernel/smp_64.c
63
{ [0 ... NR_CPUS-1] = CPU_MASK_NONE };
arch/sparc/kernel/smp_64.c
65
cpumask_t cpu_core_sib_map[NR_CPUS] __read_mostly = {
arch/sparc/kernel/smp_64.c
66
[0 ... NR_CPUS-1] = CPU_MASK_NONE };
arch/sparc/kernel/smp_64.c
68
cpumask_t cpu_core_sib_cache_map[NR_CPUS] __read_mostly = {
arch/sparc/kernel/smp_64.c
69
[0 ... NR_CPUS - 1] = CPU_MASK_NONE };
arch/sparc/kernel/sun4d_smp.c
277
unsigned char processors_in[NR_CPUS]; /* Set when ipi entered. */
arch/sparc/kernel/sun4d_smp.c
278
unsigned char processors_out[NR_CPUS]; /* Set when ipi exited. */
arch/sparc/kernel/sun4d_smp.c
371
static int cpu_tick[NR_CPUS];
arch/sparc/kernel/sun4d_smp.c
411
for (i = 0; i < NR_CPUS; i++) {
arch/sparc/kernel/traps_64.c
2819
u64 cpu_mondo_counter[NR_CPUS] = {0};
arch/sparc/kernel/traps_64.c
2820
struct trap_per_cpu trap_block[NR_CPUS];
arch/sparc/kernel/traps_64.c
865
for (i = 0; i < NR_CPUS; i++) {
arch/sparc/kernel/traps_64.c
900
sz = NR_CPUS * (2 * sizeof(struct cheetah_err_info));
arch/sparc/kernel/traps_64.c
917
for (i = 0; i < 2 * NR_CPUS; i++)
arch/sparc/mm/init_64.c
1099
for (i = 0; i < NR_CPUS; i++)
arch/sparc/mm/init_64.c
224
((1UL<<ilog2(roundup_pow_of_two(NR_CPUS)))-1UL)
arch/sparc/mm/init_64.c
2307
ilog2(roundup_pow_of_two(NR_CPUS)) > 32);
arch/sparc/mm/init_64.c
2309
BUILD_BUG_ON(NR_CPUS > 4096);
arch/sparc/mm/init_64.c
928
int numa_cpu_lookup_table[NR_CPUS];
arch/um/include/asm/current.h
13
extern struct task_struct *cpu_tasks[NR_CPUS];
arch/um/include/linux/smp-internal.h
15
extern char cpu_irqstacks[NR_CPUS][THREAD_SIZE] __aligned(THREAD_SIZE);
arch/um/kernel/process.c
46
struct task_struct *cpu_tasks[NR_CPUS] = {
arch/um/kernel/process.c
47
[0 ... NR_CPUS - 1] = &init_task,
arch/um/kernel/skas/process.c
30
char cpu_irqstacks[NR_CPUS][THREAD_SIZE] __aligned(THREAD_SIZE);
arch/um/kernel/smp.c
117
static int cpu_states[NR_CPUS];
arch/um/kernel/smp.c
217
for (; cpu < NR_CPUS; cpu++)
arch/um/kernel/smp.c
230
uml_ncpus = clamp(uml_ncpus, 1, NR_CPUS);
arch/um/kernel/time.c
762
static struct clock_event_device timer_clockevent[NR_CPUS];
arch/x86/include/asm/cpumask.h
15
#if NR_CPUS > 1
arch/x86/include/asm/fixmap.h
103
FIX_KMAP_END = FIX_KMAP_BEGIN + (KM_MAX_IDX * NR_CPUS) - 1,
arch/x86/include/asm/irq_vectors.h
132
#define CPU_VECTOR_LIMIT (64 * NR_CPUS)
arch/x86/include/asm/percpu.h
615
__typeof__(_type) _name##_early_map[NR_CPUS] __initdata = \
arch/x86/include/asm/percpu.h
616
{ [0 ... NR_CPUS-1] = _initvalue }; \
arch/x86/include/asm/percpu.h
621
__typeof__(_type) _name##_early_map[NR_CPUS] __initdata = \
arch/x86/include/asm/percpu.h
622
{ [0 ... NR_CPUS-1] = _initvalue }; \
arch/x86/include/asm/pgtable_32_areas.h
27
#define CPU_ENTRY_AREA_PAGES (NR_CPUS * DIV_ROUND_UP(sizeof(struct cpu_entry_area), PAGE_SIZE))
arch/x86/include/asm/pgtable_areas.h
16
(CPU_ENTRY_AREA_SIZE * NR_CPUS) - \
arch/x86/kernel/cpu/microcode/core.c
58
u32 microcode_rev[NR_CPUS] = {};
arch/x86/kernel/cpu/microcode/core.c
73
struct ucode_cpu_info ucode_cpu_info[NR_CPUS];
arch/x86/kernel/cpu/microcode/internal.h
49
extern u32 microcode_rev[NR_CPUS];
arch/x86/kernel/cpu/topology.c
415
static unsigned int max_possible_cpus __initdata = NR_CPUS;
arch/x86/kernel/cpu/topology.c
50
u32 cpuid_to_apicid[] __ro_after_init = { [0 ... NR_CPUS - 1] = BAD_APICID, };
arch/x86/kernel/cpuid.c
155
if (__register_chrdev(CPUID_MAJOR, 0, NR_CPUS,
arch/x86/kernel/cpuid.c
176
__unregister_chrdev(CPUID_MAJOR, 0, NR_CPUS, "cpu/cpuid");
arch/x86/kernel/cpuid.c
185
__unregister_chrdev(CPUID_MAJOR, 0, NR_CPUS, "cpu/cpuid");
arch/x86/kernel/kgdb.c
494
static DECLARE_BITMAP(was_in_debug_nmi, NR_CPUS);
arch/x86/kernel/msr.c
266
if (__register_chrdev(MSR_MAJOR, 0, NR_CPUS, "cpu/msr", &msr_fops)) {
arch/x86/kernel/msr.c
284
__unregister_chrdev(MSR_MAJOR, 0, NR_CPUS, "cpu/msr");
arch/x86/kernel/msr.c
293
__unregister_chrdev(MSR_MAJOR, 0, NR_CPUS, "cpu/msr");
arch/x86/kernel/nmi_selftest.c
27
static DECLARE_BITMAP(nmi_ipi_mask, NR_CPUS) __initdata;
arch/x86/kernel/setup_percpu.c
118
NR_CPUS, nr_cpumask_bits, nr_cpu_ids, nr_node_ids);
arch/x86/kernel/setup_percpu.c
32
unsigned long __per_cpu_offset[NR_CPUS] __ro_after_init;
arch/x86/kernel/vsmp_64.c
106
if (setup_max_cpus != NR_CPUS)
arch/x86/xen/mmu_pv.c
1334
DECLARE_BITMAP(mask, NR_CPUS);
arch/xtensa/include/asm/fixmap.h
25
(KM_MAX_IDX * NR_CPUS * DCACHE_N_COLORS) - 1,
arch/xtensa/include/asm/mmu.h
17
unsigned long asid[NR_CPUS];
arch/xtensa/kernel/smp.c
97
if (ncpus > NR_CPUS) {
arch/xtensa/kernel/smp.c
98
ncpus = NR_CPUS;
drivers/acpi/acpi_pad.c
221
static struct task_struct *ps_tsks[NR_CPUS];
drivers/acpi/acpi_pad.c
88
static unsigned long cpu_weight[NR_CPUS];
drivers/acpi/acpi_pad.c
89
static int tsk_in_cpu[NR_CPUS] = {[0 ... NR_CPUS-1] = -1};
drivers/acpi/acpi_pad.c
90
static DECLARE_BITMAP(pad_busy_cpus_bits, NR_CPUS);
drivers/acpi/acpi_processor.c
680
[0 ... NR_CPUS - 1] = -1,
drivers/acpi/acpi_processor.c
685
[0 ... NR_CPUS - 1] = -1,
drivers/acpi/acpi_processor.c
692
if (nr_unique_ids == NR_CPUS||nr_duplicate_ids == NR_CPUS)
drivers/base/arch_numa.c
144
unsigned long __per_cpu_offset[NR_CPUS] __read_mostly;
drivers/base/arch_numa.c
19
static int cpu_to_node_map[NR_CPUS] = { [0 ... NR_CPUS-1] = NUMA_NO_NODE };
drivers/base/arch_topology.c
702
struct cpu_topology cpu_topology[NR_CPUS];
drivers/base/cpu.c
240
return sysfs_emit(buf, "%d\n", NR_CPUS - 1);
drivers/base/test/test_async_driver_probe.c
291
id = NR_CPUS * 2;
drivers/base/test/test_async_driver_probe.c
76
static struct platform_device *async_dev[NR_CPUS * 2];
drivers/block/mtip32xx/mtip32xx.c
104
static u32 cpu_use[NR_CPUS];
drivers/block/rnbd/rnbd-clt.h
83
DECLARE_BITMAP(cpu_queues_bm, NR_CPUS);
drivers/bus/arm-cci.c
148
static struct cpu_port cpu_port[NR_CPUS];
drivers/clocksource/timer-econet-en751221.c
23
#define ECONET_NUM_BLOCKS DIV_ROUND_UP(NR_CPUS, 2)
drivers/cpufreq/e_powersaver.c
43
static struct eps_cpu_data *eps_cpu[NR_CPUS];
drivers/cpufreq/p4-clockmod.c
47
static int has_N44_O17_errata[NR_CPUS];
drivers/cpufreq/sparc-us2e-cpufreq.c
326
us2e_freq_table = kzalloc_objs(*us2e_freq_table, NR_CPUS);
drivers/cpufreq/sparc-us3-cpufreq.c
174
us3_freq_table = kzalloc_objs(*us3_freq_table, NR_CPUS);
drivers/cpuidle/coupled.c
99
int requested_state[NR_CPUS];
drivers/edac/octeon_edac-pc.c
26
extern unsigned long long cache_err_dcache[NR_CPUS];
drivers/hwtracing/coresight/coresight-cti-core.c
49
static struct cti_drvdata *cti_cpu_drvdata[NR_CPUS];
drivers/hwtracing/coresight/coresight-etm3x-core.c
45
static struct etm_drvdata *etmdrvdata[NR_CPUS];
drivers/hwtracing/coresight/coresight-etm4x-core.c
64
static struct etmv4_drvdata *etmdrvdata[NR_CPUS];
drivers/infiniband/hw/hfi1/sdma.c
812
.max_size = NR_CPUS,
drivers/infiniband/sw/siw/siw_main.c
61
struct task_struct *siw_tx_thread[NR_CPUS];
drivers/irqchip/irq-bcm6345-l1.c
80
struct bcm6345_l1_cpu *cpus[NR_CPUS];
drivers/irqchip/irq-bcm7038-l1.c
41
struct bcm7038_l1_cpu *cpus[NR_CPUS];
drivers/irqchip/irq-loongson-eiointc.c
51
#define MAX_EIO_NODES (NR_CPUS / CORES_PER_EIO_NODE)
drivers/irqchip/irq-mips-gic.c
139
BUG_ON(cpu >= NR_CPUS);
drivers/irqchip/irq-mips-gic.c
385
if (cpu >= NR_CPUS)
drivers/md/dm-stats.c
58
struct dm_stat_percpu *stat_percpu[NR_CPUS];
drivers/net/wireless/ath/ath9k/hw.c
436
if (NR_CPUS > 1 && ah->config.serialize_regmode == SER_REG_MODE_AUTO) {
drivers/net/wireless/ath/ath9k/init.c
178
if (NR_CPUS > 1 && ah->config.serialize_regmode == SER_REG_MODE_ON) {
drivers/net/wireless/ath/ath9k/init.c
194
if (NR_CPUS > 1 && ah->config.serialize_regmode == SER_REG_MODE_ON) {
drivers/net/wireless/ath/ath9k/init.c
235
if (NR_CPUS > 1 && ah->config.serialize_regmode == SER_REG_MODE_ON) {
drivers/nvme/host/tcp.c
57
static atomic_t nvme_tcp_cpu_queues[NR_CPUS];
drivers/scsi/qla2xxx/qla_inline.h
624
ha->qp_cpu_map = kzalloc_objs(struct qla_qpair *, NR_CPUS);
drivers/soc/fsl/dpio/dpio-service.c
53
static struct dpaa2_io *dpio_by_cpu[NR_CPUS];
drivers/soc/fsl/qbman/bman_portal.c
33
static struct bman_portal *affine_bportals[NR_CPUS];
drivers/soc/fsl/qbman/qman.c
1002
static u16 affine_channels[NR_CPUS];
drivers/soc/fsl/qbman/qman.c
1004
struct qman_portal *affine_portals[NR_CPUS];
drivers/soc/fsl/qbman/qman_priv.h
271
extern struct qman_portal *affine_portals[NR_CPUS];
drivers/tty/mips_ejtag_fdc.c
297
void __iomem *regs[NR_CPUS];
drivers/watchdog/octeon-wdt-main.c
84
static unsigned int per_cpu_countdown[NR_CPUS];
drivers/xen/events/events_base.c
171
static atomic_t channels_on_cpu[NR_CPUS];
drivers/xen/xen-acpi-processor.c
300
return NR_CPUS;
fs/gfs2/rgrp.c
1888
if (found > NR_CPUS)
include/asm-generic/percpu.h
34
extern unsigned long __per_cpu_offset[NR_CPUS];
include/linux/arch_topology.h
74
extern struct cpu_topology cpu_topology[NR_CPUS];
include/linux/blockgroup_lock.h
14
#define NR_BG_LOCKS (4 << ilog2(NR_CPUS < 32 ? NR_CPUS : 32))
include/linux/cpumask.h
1122
extern const DECLARE_BITMAP(cpu_all_bits, NR_CPUS);
include/linux/cpumask.h
1128
#if NR_CPUS == 1
include/linux/cpumask.h
1195
cpu_bit_bitmap[BITS_PER_LONG+1][BITS_TO_LONGS(NR_CPUS)];
include/linux/cpumask.h
1204
#if NR_CPUS > 1
include/linux/cpumask.h
1301
#if NR_CPUS <= BITS_PER_LONG
include/linux/cpumask.h
1304
[BITS_TO_LONGS(NR_CPUS)-1] = BITMAP_LAST_WORD_MASK(NR_CPUS) \
include/linux/cpumask.h
1311
[0 ... BITS_TO_LONGS(NR_CPUS)-2] = ~0UL, \
include/linux/cpumask.h
1312
[BITS_TO_LONGS(NR_CPUS)-1] = BITMAP_LAST_WORD_MASK(NR_CPUS) \
include/linux/cpumask.h
1379
#if NR_CPUS <= BITS_PER_LONG
include/linux/cpumask.h
1382
[BITS_TO_LONGS(NR_CPUS)-1] = BITMAP_LAST_WORD_MASK(NR_CPUS) \
include/linux/cpumask.h
1387
[0 ... BITS_TO_LONGS(NR_CPUS)-2] = ~0UL, \
include/linux/cpumask.h
1388
[BITS_TO_LONGS(NR_CPUS)-1] = BITMAP_LAST_WORD_MASK(NR_CPUS) \
include/linux/cpumask.h
1394
[0 ... BITS_TO_LONGS(NR_CPUS)-1] = 0UL \
include/linux/cpumask.h
1417
#define CPUMAP_FILE_MAX_BYTES (((NR_CPUS * 9)/32 > PAGE_SIZE) \
include/linux/cpumask.h
1418
? (NR_CPUS * 9)/32 - 1 : PAGE_SIZE)
include/linux/cpumask.h
1419
#define CPULIST_FILE_MAX_BYTES (((NR_CPUS * 7)/2 > PAGE_SIZE) ? (NR_CPUS * 7)/2 : PAGE_SIZE)
include/linux/cpumask.h
257
#if NR_CPUS == 1
include/linux/cpumask.h
29
#if (NR_CPUS == 1) || defined(CONFIG_FORCE_NR_CPUS)
include/linux/cpumask.h
30
#define nr_cpu_ids ((unsigned int)NR_CPUS)
include/linux/cpumask.h
37
#if (NR_CPUS == 1) || defined(CONFIG_FORCE_NR_CPUS)
include/linux/cpumask.h
586
[0 ... BITS_TO_LONGS(NR_CPUS)-1] = 0UL \
include/linux/cpumask.h
68
#if NR_CPUS <= BITS_PER_LONG
include/linux/cpumask.h
69
#define small_cpumask_bits ((unsigned int)NR_CPUS)
include/linux/cpumask.h
70
#define large_cpumask_bits ((unsigned int)NR_CPUS)
include/linux/cpumask.h
71
#elif NR_CPUS <= 4*BITS_PER_LONG
include/linux/cpumask.h
73
#define large_cpumask_bits ((unsigned int)NR_CPUS)
include/linux/cpumask_types.h
9
typedef struct cpumask { DECLARE_BITMAP(bits, NR_CPUS); } cpumask_t;
include/linux/kernfs.h
63
#define NR_KERNFS_LOCK_BITS (2 * (ilog2(NR_CPUS < 32 ? NR_CPUS : 32)))
include/linux/pstore.h
224
#if NR_CPUS <= 2 && defined(CONFIG_ARM_THUMB)
include/linux/pstore.h
226
#elif NR_CPUS <= 4 && defined(CONFIG_ARM)
include/linux/rcu_node_tree.h
54
#if NR_CPUS <= RCU_FANOUT_1
include/linux/rcu_node_tree.h
61
#elif NR_CPUS <= RCU_FANOUT_2
include/linux/rcu_node_tree.h
64
# define NUM_RCU_LVL_1 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_1)
include/linux/rcu_node_tree.h
69
#elif NR_CPUS <= RCU_FANOUT_3
include/linux/rcu_node_tree.h
72
# define NUM_RCU_LVL_1 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_2)
include/linux/rcu_node_tree.h
73
# define NUM_RCU_LVL_2 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_1)
include/linux/rcu_node_tree.h
78
#elif NR_CPUS <= RCU_FANOUT_4
include/linux/rcu_node_tree.h
81
# define NUM_RCU_LVL_1 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_3)
include/linux/rcu_node_tree.h
82
# define NUM_RCU_LVL_2 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_2)
include/linux/rcu_node_tree.h
83
# define NUM_RCU_LVL_3 DIV_ROUND_UP(NR_CPUS, RCU_FANOUT_1)
include/linux/workqueue.h
92
WORK_CPU_UNBOUND = NR_CPUS,
include/net/busy_poll.h
25
#define MIN_NAPI_ID ((unsigned int)(NR_CPUS + 1))
init/init_task.c
113
.nr_cpus_allowed= NR_CPUS,
kernel/bpf/cpumap.c
98
if (attr->max_entries > NR_CPUS)
kernel/cgroup/cpuset-v1.c
779
.max_write_len = (100U + 6 * NR_CPUS),
kernel/cgroup/cpuset.c
3366
.max_write_len = (100U + 6 * NR_CPUS),
kernel/cgroup/cpuset.c
3405
.max_write_len = (100U + 6 * NR_CPUS),
kernel/cpu.c
3067
const unsigned long cpu_bit_bitmap[BITS_PER_LONG+1][BITS_TO_LONGS(NR_CPUS)] = {
kernel/cpu.c
3078
const DECLARE_BITMAP(cpu_all_bits, NR_CPUS) = CPU_BITS_ALL;
kernel/cpu.c
3084
unsigned int __num_possible_cpus __ro_after_init = NR_CPUS;
kernel/debug/debug_core.c
65
struct debuggerinfo_struct kgdb_info[NR_CPUS];
kernel/debug/gdbstub.c
439
if (tid < -1 && tid > -NR_CPUS - 2) {
kernel/debug/kdb/kdb_main.c
2091
for (start_cpu = -1, i = 0; i < NR_CPUS; i++) {
kernel/rcu/rcutorture.c
166
static struct task_struct *boost_tasks[NR_CPUS];
kernel/rcu/tree.c
4798
nr_cpu_ids == NR_CPUS)
kernel/rcu/tree_plugin.h
67
if (nr_cpu_ids != NR_CPUS)
kernel/rcu/tree_plugin.h
68
pr_info("\tRCU restricting CPUs from NR_CPUS=%d to nr_cpu_ids=%u.\n", NR_CPUS, nr_cpu_ids);
kernel/relay.c
629
if (!chan || cpu >= NR_CPUS)
kernel/sched/fair.c
11033
if (p->nr_cpus_allowed != NR_CPUS) {
kernel/smp.c
928
unsigned int setup_max_cpus = NR_CPUS;
kernel/smp.c
979
#if (NR_CPUS > 1) && !defined(CONFIG_FORCE_NR_CPUS)
kernel/smp.c
981
unsigned int nr_cpu_ids __read_mostly = NR_CPUS;
kernel/smp.c
988
set_nr_cpu_ids(find_last_bit(cpumask_bits(cpu_possible_mask), NR_CPUS) + 1);
kernel/taskstats.c
30
#define TASKSTATS_CPUMASK_MAXLEN (100+6*NR_CPUS)
kernel/torture.c
184
static int torture_online_cpus = NR_CPUS;
kernel/trace/ring_buffer.c
7530
static struct task_struct *rb_threads[NR_CPUS] __initdata;
kernel/trace/ring_buffer.c
7550
static struct rb_test_data rb_data[NR_CPUS] __initdata;
kernel/trace/trace_kdb.c
108
if (err || cpu_file >= NR_CPUS || cpu_file < 0 ||
kernel/watchdog.c
1212
.maxlen = NR_CPUS,
lib/nmi_backtrace.c
25
static DECLARE_BITMAP(backtrace_mask, NR_CPUS) __read_mostly;
mm/percpu.c
2434
ai->groups[0].cpu_map[unit] = NR_CPUS;
mm/percpu.c
2500
if (gi->cpu_map[unit] != NR_CPUS)
mm/percpu.c
2622
pcpu_low_unit_cpu = NR_CPUS;
mm/percpu.c
2623
pcpu_high_unit_cpu = NR_CPUS;
mm/percpu.c
2633
if (cpu == NR_CPUS)
mm/percpu.c
2644
if (pcpu_low_unit_cpu == NR_CPUS ||
mm/percpu.c
2647
if (pcpu_high_unit_cpu == NR_CPUS ||
mm/percpu.c
2810
static int group_map[NR_CPUS] __initdata;
mm/percpu.c
2811
static int group_cnt[NR_CPUS] __initdata;
mm/percpu.c
3046
unsigned int cpu = NR_CPUS;
mm/percpu.c
3049
for (i = 0; i < gi->nr_units && cpu == NR_CPUS; i++)
mm/percpu.c
3051
BUG_ON(cpu == NR_CPUS);
mm/percpu.c
3091
if (gi->cpu_map[i] == NR_CPUS) {
mm/percpu.c
3301
unsigned long __per_cpu_offset[NR_CPUS] __read_mostly;
mm/slub.c
8640
DECLARE_BITMAP(cpus, NR_CPUS);
mm/vmalloc.c
2579
VMALLOC_PAGES / roundup_pow_of_two(NR_CPUS) / 16))
net/bridge/netfilter/ebtables.c
1127
NR_CPUS - SMP_CACHE_BYTES) / sizeof(struct ebt_counter))
net/bridge/netfilter/ebtables.c
2235
NR_CPUS - SMP_CACHE_BYTES) / sizeof(struct ebt_counter))
net/core/dev.c
184
static unsigned int napi_gen_id = NR_CPUS;
net/core/dev.c
4720
if (sender_cpu >= (u32)NR_CPUS)
net/iucv/iucv.c
131
static struct iucv_irq_data *iucv_irq_data[NR_CPUS];
net/iucv/iucv.c
312
static union iucv_param *iucv_param[NR_CPUS];
net/iucv/iucv.c
313
static union iucv_param *iucv_param_irq[NR_CPUS];
net/netfilter/ipvs/ip_vs_ctl.c
2248
.maxlen = NR_CPUS, /* unused */
tools/perf/trace/beauty/arch/x86/include/asm/irq_vectors.h
132
#define CPU_VECTOR_LIMIT (64 * NR_CPUS)
tools/sched_ext/include/scx/common.bpf.h
39
#ifndef NR_CPUS
tools/testing/selftests/bpf/bpf_arena_alloc.h
16
static void __arena * __arena page_frag_cur_page[NR_CPUS];
tools/testing/selftests/bpf/bpf_arena_alloc.h
17
static int __arena page_frag_cur_offset[NR_CPUS];
tools/testing/selftests/bpf/progs/bpf_hashmap_lookup.c
18
#define CPU_MASK (NR_CPUS-1)
tools/testing/selftests/bpf/progs/bpf_hashmap_lookup.c
23
u32 __attribute__((__aligned__(8))) key[NR_CPUS];
tools/testing/selftests/bpf/progs/bpf_hashmap_lookup.c
26
u64 __attribute__((__aligned__(256))) percpu_times_index[NR_CPUS];
tools/testing/selftests/bpf/progs/bpf_hashmap_lookup.c
27
u64 __attribute__((__aligned__(256))) percpu_times[NR_CPUS][NR_SLOTS];
tools/testing/selftests/livepatch/test_modules/test_klp_syscall.c
30
static int klp_pids[NR_CPUS];