arch/arm/kernel/machine_kexec.c
47
if (num_possible_cpus() > 1 && platform_can_secondary_boot() &&
arch/arm/kernel/setup.c
676
if (mpidr_hash_size() > 4 * num_possible_cpus())
arch/arm/kernel/smp.c
505
unsigned int ncores = num_possible_cpus();
arch/arm/mach-exynos/exynos.c
121
for (core_id = 0; core_id < num_possible_cpus(); core_id++) {
arch/arm/mach-mvebu/platsmp.c
93
unsigned int ncores = num_possible_cpus();
arch/arm/mm/cache-b15-rac.c
304
if (WARN(num_possible_cpus() > 4, "RAC only supports 4 CPUs\n"))
arch/arm64/kernel/setup.c
166
if (mpidr_hash_size() > 4 * num_possible_cpus())
arch/arm64/kernel/smp.c
1307
bool smp_spin_tables = (num_possible_cpus() > 1 && !have_cpu_die());
arch/arm64/kvm/arm.c
2468
num_possible_cpus(), kern_hyp_va(per_cpu_base),
arch/arm64/kvm/vmid.c
183
WARN_ON(NUM_USER_VMIDS - 1 <= num_possible_cpus());
arch/arm64/mm/context.c
387
WARN_ON(num_available_asids - 1 <= num_possible_cpus());
arch/arm64/mm/context.c
396
max_pinned_asids = num_available_asids - num_possible_cpus() - 2;
arch/csky/kernel/perf_event.c
1167
irqs = min(pmu_device->num_resources, num_possible_cpus());
arch/csky/mm/asid.c
179
WARN_ON(NUM_CTXT_ASIDS(info) - 1 <= num_possible_cpus());
arch/csky/mm/context.c
31
BUG_ON(((1 << CONFIG_CPU_ASID_BITS) - 1) <= num_possible_cpus());
arch/microblaze/include/asm/fixmap.h
50
FIX_KMAP_END = FIX_KMAP_BEGIN + (KM_MAX_IDX * num_possible_cpus()) - 1,
arch/mips/kernel/csrc-r4k.c
74
if (num_possible_cpus() > 1 &&
arch/mips/kernel/i8253.c
32
if (num_possible_cpus() > 1 || /* PIT does not scale! */
arch/mips/kernel/perf_event_mipsxx.c
138
if (num_possible_cpus() > 1)
arch/mips/kernel/perf_event_mipsxx.c
707
if (num_possible_cpus() > 1)
arch/mips/kernel/setup.c
726
int i, possible = num_possible_cpus();
arch/mips/kernel/smp.c
340
if (num_possible_cpus() == 1)
arch/mips/loongson2ef/common/cs5536/cs5536_mfgpt.c
197
if (num_possible_cpus() > 1) /* MFGPT does not scale! */
arch/mips/mm/context.c
277
WARN_ON(num_mmids <= num_possible_cpus());
arch/mips/mm/tlbex.c
346
if (num_possible_cpus() > 1) {
arch/nios2/kernel/cpuinfo.c
172
return i < num_possible_cpus() ? (void *) (i + 1) : NULL;
arch/powerpc/kernel/smp.c
1167
int idx = DIV_ROUND_UP(num_possible_cpus(), threads_per_core);
arch/powerpc/kexec/crash.c
415
phdr_cnt = num_possible_cpus() + 1;
arch/powerpc/kexec/file_load_64.c
658
possible_cpu_nodes = num_possible_cpus() / threads_per_core;
arch/powerpc/perf/imc-pmu.c
1570
int i, nr_cores = DIV_ROUND_UP(num_possible_cpus(), threads_per_core);
arch/powerpc/perf/imc-pmu.c
1714
nr_cores = DIV_ROUND_UP(num_possible_cpus(), threads_per_core);
arch/powerpc/perf/imc-pmu.c
1752
nr_cores = DIV_ROUND_UP(num_possible_cpus(), threads_per_core);
arch/powerpc/platforms/85xx/smp.c
283
WARN_ON(nr < 0 || nr >= num_possible_cpus());
arch/powerpc/platforms/pseries/hotplug-cpu.c
254
" supports %d logical cpus.\n", np, num_possible_cpus());
arch/powerpc/platforms/pseries/hotplug-cpu.c
372
if (cpu == num_possible_cpus()) {
arch/powerpc/platforms/pseries/hotplug-cpu.c
423
if (cpu == num_possible_cpus())
arch/powerpc/platforms/pseries/lpar.c
210
vcpu_associativity = kcalloc(num_possible_cpus() / threads_per_core,
arch/powerpc/platforms/pseries/lparcfg.c
543
partition_potential_processors = num_possible_cpus();
arch/powerpc/platforms/pseries/setup.c
130
int nr_cpus = num_possible_cpus();
arch/powerpc/sysdev/mpic.c
1443
BUG_ON(num_possible_cpus() > MPIC_MAX_CPUS);
arch/powerpc/sysdev/mpic.c
1515
name, vers, (unsigned long long)mpic->paddr, num_possible_cpus());
arch/powerpc/sysdev/mpic.c
1890
nr_cpus = num_possible_cpus();
arch/powerpc/xmon/xmon.c
1336
if (!scanhex(&cpu) || cpu >= num_possible_cpus()) {
arch/powerpc/xmon/xmon.c
2736
if (num_possible_cpus() == 0) {
arch/powerpc/xmon/xmon.c
2758
if (scanhex(&num) && num < num_possible_cpus())
arch/powerpc/xmon/xmon.c
2831
if (scanhex(&num) && num < num_possible_cpus())
arch/powerpc/xmon/xmon.c
3684
if (scanhex(&cpu) && cpu < num_possible_cpus()) {
arch/riscv/kernel/acpi.c
183
if (cpuid >= 0 && cpuid < num_possible_cpus())
arch/riscv/kernel/unaligned_access_speed.c
141
unsigned int cpu_count = num_possible_cpus();
arch/riscv/kvm/vmid.c
39
if ((1UL << vmid_bits) < num_possible_cpus())
arch/riscv/mm/context.c
255
if (num_asids > (2 * num_possible_cpus())) {
arch/s390/appldata/appldata_os.c
166
max_size = struct_size(appldata_os_data, os_cpu, num_possible_cpus());
arch/s390/hypfs/hypfs_diag0c.c
38
cpu_vec = kmalloc_objs(*cpu_vec, num_possible_cpus());
arch/s390/pci/pci_irq.c
557
zpci_sbv = airq_iv_create(num_possible_cpus(), 0, NULL);
arch/s390/pci/pci_irq.c
562
iib.diib.nr_cpus = num_possible_cpus();
arch/s390/pci/pci_irq.c
566
zpci_ibv = kzalloc_objs(*zpci_ibv, num_possible_cpus());
arch/sparc/kernel/cpumap.c
138
for (i = 0; i < num_possible_cpus(); i++) {
arch/sparc/kernel/cpumap.c
233
for (last_cpu = (num_possible_cpus() - 1); last_cpu >= 0; last_cpu--) {
arch/sparc/kernel/cpumap.c
381
for (i = 0; i < num_possible_cpus(); i++) {
arch/sparc/kernel/mdesc.c
888
if (*id < num_possible_cpus())
arch/sparc/kernel/mdesc.c
897
if (*id < num_possible_cpus()) {
arch/sparc/kernel/mdesc.c
989
if (*id < num_possible_cpus())
arch/sparc/kernel/prom_64.c
489
if (num_possible_cpus() < nr_cpu_ids)
arch/x86/events/amd/core.c
1537
if (num_possible_cpus() == 1) {
arch/x86/kernel/alternative.c
2174
if (num_possible_cpus() == 1)
arch/x86/kernel/alternative.c
2220
BUG_ON(num_possible_cpus() == 1);
arch/x86/kernel/alternative.c
2452
if (!uniproc_patched || num_possible_cpus() == 1) {
arch/x86/kernel/apic/apic.c
982
if (num_possible_cpus() > 1) {
arch/x86/kernel/apic/apic.c
991
if (num_possible_cpus() > 1)
arch/x86/kernel/apic/x2apic_uv_x.c
1687
pr_info("UV: Found %d hubs, %d nodes, %d CPUs\n", uv_num_possible_blades(), num_possible_nodes(), num_possible_cpus());
arch/x86/kernel/cpu/mce/dev-mcelog.c
305
if (m.extcpu >= num_possible_cpus() || !cpu_online(m.extcpu))
arch/x86/kernel/cpu/mce/genpool.c
130
mce_numrecords = max(MCE_MIN_ENTRIES, num_possible_cpus() * MCE_PER_CPU);
arch/x86/kernel/hpet.c
734
if (++hpet_base.nr_clockevents == num_possible_cpus())
arch/x86/kernel/i8253.c
68
if (num_possible_cpus() > 1 || is_hpet_enabled() ||
arch/x86/kernel/kvm.c
1121
if (num_possible_cpus() == 1) {
arch/x86/kernel/kvm.c
498
(num_possible_cpus() != 1));
arch/x86/kernel/kvm.c
504
(num_possible_cpus() != 1));
arch/x86/kernel/kvm.c
513
(num_possible_cpus() != 1));
arch/x86/kernel/kvmclock.c
210
if (HVC_BOOT_ARRAY_SIZE >= num_possible_cpus())
arch/x86/kernel/kvmclock.c
213
ncpus = num_possible_cpus() - HVC_BOOT_ARRAY_SIZE;
arch/x86/kernel/smpboot.c
941
width = num_digits(num_possible_cpus()) + 1; /* + '#' sign */
arch/x86/xen/mmu_pv.c
1338
sizeof(args->mask[0]) * BITS_TO_LONGS(num_possible_cpus());
arch/x86/xen/smp_pv.c
213
while ((num_possible_cpus() > 1) && (num_possible_cpus() > max_cpus)) {
arch/x86/xen/spinlock.c
126
if (num_possible_cpus() == 1)
drivers/base/arch_topology.c
283
raw_capacity = kcalloc(num_possible_cpus(),
drivers/base/arch_topology.c
336
raw_capacity = kcalloc(num_possible_cpus(), sizeof(*raw_capacity),
drivers/block/rnbd/rnbd-clt.c
792
bitmap_zero(sess->cpu_queues_bm, num_possible_cpus());
drivers/clk/mvebu/clk-cpu.c
173
int ncpus = num_possible_cpus();
drivers/clk/samsung/clk-exynos4.c
1186
if (num_possible_cpus() == 4)
drivers/clocksource/ingenic-timer.c
289
tcu = kzalloc_flex(*tcu, timers, num_possible_cpus());
drivers/clocksource/ingenic-timer.c
298
num_possible_cpus() + 1);
drivers/clocksource/ingenic-timer.c
304
soc_info->num_channels - num_possible_cpus() + 1) {
drivers/clocksource/ingenic-timer.c
315
for (cpu = 0; cpu < num_possible_cpus(); cpu++) {
drivers/clocksource/timer-econet-en751221.c
173
int num_blocks = DIV_ROUND_UP(num_possible_cpus(), 2);
drivers/clocksource/timer-nxp-stm.c
422
if (stm_instances >= num_possible_cpus())
drivers/clocksource/timer-nxp-stm.c
469
if (stm_instances == num_possible_cpus()) {
drivers/clocksource/timer-rtl-otto.c
264
unsigned int clkidx = num_possible_cpus();
drivers/cpufreq/armada-8k-cpufreq.c
104
int opps_index, nb_cpus = num_possible_cpus();
drivers/cpufreq/armada-8k-cpufreq.c
145
nb_cpus = num_possible_cpus();
drivers/cpufreq/intel_pstate.c
3849
_all_cpu_data = vzalloc(array_size(sizeof(void *), num_possible_cpus()));
drivers/cpufreq/powernv-cpufreq.c
1051
chip = kcalloc(num_possible_cpus(), sizeof(*chip), GFP_KERNEL);
drivers/cpufreq/qcom-cpufreq-nvmem.c
503
drv = devm_kzalloc(&pdev->dev, struct_size(drv, cpus, num_possible_cpus()),
drivers/cpufreq/sun50i-cpufreq-nvmem.c
247
opp_tokens = kzalloc_objs(*opp_tokens, num_possible_cpus());
drivers/crypto/intel/iaa/iaa_crypto_main.c
1863
nr_cpus = num_possible_cpus();
drivers/edac/a72_edac.c
138
num_possible_cpus(), "L", 2, 1,
drivers/edac/octeon_edac-pc.c
94
p->ed = edac_device_alloc_ctl_info(0, "cpu", num_possible_cpus(),
drivers/hv/mshv_vtl_main.c
468
if (input.cpu >= num_possible_cpus() || !cpu_online(input.cpu))
drivers/hwtracing/stm/ftrace.c
69
stm_ftrace.data.nr_chans = roundup_pow_of_two(num_possible_cpus());
drivers/infiniband/hw/erdma/erdma_main.c
127
int expect_irq_num = min(num_possible_cpus() + 1, ERDMA_NUM_MSIX_VEC);
drivers/infiniband/sw/rxe/rxe_verbs.c
1543
dev->num_comp_vectors = num_possible_cpus();
drivers/infiniband/sw/siw/siw_main.c
314
base_dev->num_comp_vectors = num_possible_cpus();
drivers/infiniband/ulp/ipoib/ipoib_netlink.c
46
return min_t(unsigned int, num_possible_cpus(), 128);
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-sva.c
256
num_possible_cpus() - 2);
drivers/irqchip/irq-gic.c
428
if (!mask && num_possible_cpus() > 1)
drivers/irqchip/irq-ls-scfg-msi.c
370
cpu_num = num_possible_cpus();
drivers/irqchip/irq-mips-gic.c
867
num_ipis = 2 * num_possible_cpus();
drivers/irqchip/irq-ompic.c
167
if (resource_size(&res) < (num_possible_cpus() * OMPIC_CPUBYTES)) {
drivers/irqchip/irq-ompic.c
170
(num_possible_cpus() * OMPIC_CPUBYTES));
drivers/irqchip/irq-riscv-imsic-platform.c
342
imsic->fwnode, num_possible_cpus() * (global->nr_ids - 1));
drivers/md/dm-stats.c
310
num_possible_cpus() * (percpu_alloc_size + histogram_alloc_size)))
drivers/net/caif/caif_virtio.c
386
if (cfv->allocsz <= (num_possible_cpus() + 1) * cfv->ndev->mtu)
drivers/net/caif/caif_virtio.c
390
if (cfv->allocsz <= num_possible_cpus() * cfv->ndev->mtu) {
drivers/net/caif/caif_virtio.c
422
cfv->reserved_size = num_possible_cpus() * cfv->ndev->mtu;
drivers/net/ethernet/amazon/ena/ena_netdev.c
35
#define ENA_MAX_RINGS min_t(unsigned int, ENA_MAX_NUM_IO_QUEUES, num_possible_cpus())
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13075
if (bp->tx_nr_rings_xdp < num_possible_cpus()) {
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
349
unsigned int ppmax = (*total) / num_possible_cpus();
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
961
channels = kcalloc(num_possible_cpus(), sizeof(u16), GFP_KERNEL);
drivers/net/ethernet/freescale/dpaa/dpaa_eth.h
187
return num_possible_cpus();
drivers/net/ethernet/freescale/dpaa/dpaa_ethtool.c
424
needs_revert = kzalloc_objs(bool, num_possible_cpus());
drivers/net/ethernet/freescale/enetc/enetc.c
3158
priv->min_num_stack_tx_queues = num_possible_cpus();
drivers/net/ethernet/freescale/enetc/enetc.c
3270
int num_xdp_tx_queues = prog ? num_possible_cpus() : 0;
drivers/net/ethernet/freescale/enetc/enetc.c
3590
priv->min_num_stack_tx_queues = num_possible_cpus();
drivers/net/ethernet/freescale/enetc/enetc.c
3591
first_xdp_tx_ring = priv->num_tx_rings - num_possible_cpus();
drivers/net/ethernet/freescale/enetc/enetc.c
89
return num_tx_rings - num_possible_cpus();
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1265
if (q_num == num_possible_cpus()) {
drivers/net/ethernet/intel/ice/ice_lib.c
99
vsi->txq_map = devm_kcalloc(dev, (vsi->alloc_txq + num_possible_cpus()),
drivers/net/ethernet/intel/ice/ice_main.c
2940
u16 cpus = num_possible_cpus();
drivers/net/ethernet/marvell/mvpp2/mvpp2_cls.c
1458
int nrxqs, cpu, cpus = num_possible_cpus();
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3340
if (queue < num_possible_cpus())
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
340
nrxqs = (num_possible_cpus() + 3) & ~0x3;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
5031
if (percpu && port->ntxqs >= num_possible_cpus() * 2)
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
5346
if (port->ntxqs < num_possible_cpus() * 2) {
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
7031
else if (port->ntxqs >= num_possible_cpus() * 2)
drivers/net/ethernet/sfc/efx_channels.c
146
n_xdp_tx = num_possible_cpus();
drivers/net/ethernet/sfc/siena/efx_channels.c
147
n_xdp_tx = num_possible_cpus();
drivers/net/veth.c
1968
int queues = num_possible_cpus();
drivers/net/wireless/ath/ath9k/hw.c
433
if (num_possible_cpus() > 1)
drivers/net/wireless/marvell/mwifiex/main.c
1548
if ((num_possible_cpus() > 1) || adapter->iface_type == MWIFIEX_USB)
drivers/net/wireless/marvell/mwifiex/main.c
1714
if ((num_possible_cpus() > 1) || adapter->iface_type == MWIFIEX_USB)
drivers/nvdimm/region.c
25
&& nd_region->num_lanes < num_possible_cpus()
drivers/nvdimm/region.c
29
num_possible_cpus());
drivers/pci/controller/pci-xgene-msi.c
27
#define NR_MSI_VEC (NR_MSI_BITS / num_possible_cpus())
drivers/pci/controller/pci-xgene-msi.c
329
err = irq_set_affinity(irq, cpumask_of(i % num_possible_cpus()));
drivers/pci/controller/pcie-iproc-msi.c
538
msi->nr_cpus = num_possible_cpus();
drivers/perf/arm_pmu.c
647
if (err && num_possible_cpus() > 1) {
drivers/platform/x86/intel/speed_select_if/isst_if_common.c
298
cpu >= nr_cpu_ids || cpu >= num_possible_cpus())
drivers/platform/x86/intel/speed_select_if/isst_if_common.c
372
cpu >= nr_cpu_ids || cpu >= num_possible_cpus())
drivers/platform/x86/intel/speed_select_if/isst_if_common.c
428
isst_cpu_info = kzalloc_objs(*isst_cpu_info, num_possible_cpus());
drivers/platform/x86/intel/speed_select_if/isst_if_common.c
465
cpu_map->logical_cpu >= num_possible_cpus())
drivers/platform/x86/intel/tpmi_power_domains.c
120
if (cpu_no >= num_possible_cpus())
drivers/platform/x86/intel/tpmi_power_domains.c
129
if (cpu_no >= num_possible_cpus())
drivers/platform/x86/intel/tpmi_power_domains.c
142
if (cpu_no >= num_possible_cpus())
drivers/scsi/bnx2fc/bnx2fc_hwi.c
993
unsigned int cpu = wqe % num_possible_cpus();
drivers/scsi/bnx2fc/bnx2fc_io.c
17
#define RESERVE_FREE_LIST_INDEX num_possible_cpus()
drivers/scsi/bnx2fc/bnx2fc_io.c
222
int arr_sz = num_possible_cpus() + 1;
drivers/scsi/bnx2fc/bnx2fc_io.c
287
num_possible_cpus()]);
drivers/scsi/bnx2fc/bnx2fc_io.c
290
&cmgr->free_list[num_possible_cpus()]);
drivers/scsi/bnx2fc/bnx2fc_io.c
377
for (i = 0; i < num_possible_cpus() + 1; i++) {
drivers/scsi/bnx2fc/bnx2fc_io.c
523
index = io_req->xid % num_possible_cpus();
drivers/scsi/elx/efct/efct_hw.c
306
cpus = num_possible_cpus();
drivers/scsi/elx/efct/efct_hw_queues.c
102
hw->wq_cpu_array = kcalloc(num_possible_cpus(), sizeof(void *),
drivers/soc/bcm/brcmstb/biuctrl.c
184
if (WARN(num_possible_cpus() > 4, "RAC only supports 4 CPUs\n"))
drivers/soc/fsl/dpio/dpio-service.c
143
if (desc->cpu != DPAA2_IO_ANY_CPU && desc->cpu >= num_possible_cpus()) {
drivers/soc/fsl/dpio/dpio-service.c
63
if (cpu != DPAA2_IO_ANY_CPU && cpu >= num_possible_cpus())
drivers/soc/fsl/qbman/qman.c
1021
for (i = 0; i < num_possible_cpus(); i++) {
drivers/soc/fsl/qbman/qman.c
1087
for (i = 0; i < num_possible_cpus(); i++) {
drivers/soc/samsung/exynos-pmu.c
454
pmu_context->in_cpuhp = devm_bitmap_zalloc(dev, num_possible_cpus(),
drivers/soc/tegra/flowctrl.c
118
for (i = 0; i < num_possible_cpus(); i++) {
drivers/thermal/thermal_trace.h
122
__bitmask(cpumask, num_possible_cpus())
drivers/thermal/thermal_trace.h
130
num_possible_cpus());
drivers/ufs/core/ufs-mcq.c
199
rw_queues = num_possible_cpus();
drivers/ufs/core/ufs-mcq.c
214
num_possible_cpus());
drivers/ufs/core/ufs-mcq.c
42
num_possible_cpus());
drivers/ufs/core/ufs-mcq.c
58
num_possible_cpus());
drivers/ufs/core/ufs-mcq.c
74
num_possible_cpus());
drivers/virt/acrn/hsm.c
444
if (cpu >= num_possible_cpus() || cpu == 0 || !cpu_is_hotpluggable(cpu))
drivers/xen/time.c
98
num_possible_cpus(), GFP_ATOMIC);
fs/aio.c
753
nr_events = max(nr_events, num_possible_cpus() * 4);
fs/aio.c
796
ctx->req_batch = (ctx->nr_events - 1) / (num_possible_cpus() * 4);
fs/erofs/decompressor_deflate.c
48
z_erofs_deflate_nstrms = num_possible_cpus();
fs/erofs/decompressor_lzma.c
54
z_erofs_lzma_nstrms = num_possible_cpus();
fs/erofs/decompressor_zstd.c
56
z_erofs_zstd_nstrms = num_possible_cpus();
fs/erofs/zdata.c
327
num_possible_cpus(), GFP_ATOMIC);
fs/erofs/zdata.c
461
WQ_UNBOUND | WQ_HIGHPRI, num_possible_cpus());
fs/erofs/zutil.c
125
unsigned int i, total = num_possible_cpus();
fs/ext4/mballoc.c
3820
sbi->s_mb_nr_global_goals = umin(num_possible_cpus(),
fs/fuse/dev_uring.c
230
size_t nr_queues = num_possible_cpus();
fs/squashfs/decompressor_multi_percpu.c
102
return num_possible_cpus();
fs/xfs/xfs_sysfs.c
215
if (val < -1 || val > num_possible_cpus())
include/linux/bpf.h
4005
if ((flags & BPF_F_CPU) && cpu >= num_possible_cpus())
include/linux/kdb.h
189
if (cpu > num_possible_cpus())
include/linux/mm_types.h
1562
return cpumask_size() + bitmap_size(num_possible_cpus());
kernel/bpf/arraymap.c
708
buf_size = array->elem_size * num_possible_cpus();
kernel/bpf/arraymap.c
790
usage += entries * elem_size * num_possible_cpus();
kernel/bpf/bpf_local_storage.c
873
nbuckets = roundup_pow_of_two(num_possible_cpus());
kernel/bpf/bpf_lru_list.c
573
lru->target_free = clamp((nr_elems / num_possible_cpus()) / 2,
kernel/bpf/bpf_lru_list.c
585
pcpu_entries = nr_elems / num_possible_cpus();
kernel/bpf/hashtab.c
1819
value_size = size * num_possible_cpus();
kernel/bpf/hashtab.c
2223
buf_size = round_up(map->value_size, 8) * num_possible_cpus();
kernel/bpf/hashtab.c
226
num_entries += num_possible_cpus();
kernel/bpf/hashtab.c
2327
num_entries += num_possible_cpus();
kernel/bpf/hashtab.c
2332
usage += value_size * num_possible_cpus() * num_entries;
kernel/bpf/hashtab.c
2334
usage += sizeof(struct htab_elem *) * num_possible_cpus();
kernel/bpf/hashtab.c
2344
usage += value_size * num_possible_cpus() * num_entries;
kernel/bpf/hashtab.c
246
num_entries += num_possible_cpus();
kernel/bpf/hashtab.c
320
num_entries += num_possible_cpus();
kernel/bpf/hashtab.c
568
num_possible_cpus());
kernel/bpf/hashtab.c
571
num_possible_cpus());
kernel/bpf/local_storage.c
496
*pages = round_up(round_up(size, 8) * num_possible_cpus(),
kernel/bpf/map_iter.c
131
value_size = round_up(map->value_size, 8) * num_possible_cpus();
kernel/bpf/percpu_freelist.c
84
n = nr_elems / num_possible_cpus();
kernel/bpf/percpu_freelist.c
85
m = nr_elems % num_possible_cpus();
kernel/bpf/syscall.c
145
return round_up(map->value_size, 8) * num_possible_cpus();
kernel/crash_core.c
181
unsigned long nr_cpus = num_possible_cpus(), nr_phdr, elf_sz;
kernel/debug/kdb/kdb_bt.c
108
if (cpu >= num_possible_cpus() || !cpu_online(cpu)) {
kernel/dma/swiotlb.c
380
swiotlb_adjust_nareas(num_possible_cpus());
kernel/dma/swiotlb.c
460
swiotlb_adjust_nareas(num_possible_cpus());
kernel/futex/core.c
1981
hashsize = 256 * num_possible_cpus();
kernel/kprobes.c
2285
rp->maxactive = max_t(unsigned int, 10, 2*num_possible_cpus());
kernel/locking/qspinlock_paravirt.h
186
int pv_hash_size = ALIGN(4 * num_possible_cpus(), PV_HE_PER_LINE);
kernel/padata.c
1108
possible_cpus = num_possible_cpus();
kernel/pid.c
816
PIDS_PER_CPU_DEFAULT * num_possible_cpus()));
kernel/pid.c
848
PIDS_PER_CPU_DEFAULT * num_possible_cpus()));
kernel/pid.c
850
PIDS_PER_CPU_MIN * num_possible_cpus());
kernel/printk/printk.c
1085
if (num_possible_cpus() == 1)
kernel/printk/printk.c
1088
cpu_extra = (num_possible_cpus() - 1) * __LOG_CPU_MAX_BUF_LEN;
kernel/rcu/tasks.h
263
num_possible_cpus());
kernel/rcu/tasks.h
548
if (index < num_possible_cpus()) {
kernel/rcu/tasks.h
554
if (index < num_possible_cpus()) {
kernel/rcu/tree.c
4909
if (num_possible_cpus() <= WAKE_FROM_GP_CPU_THRESHOLD)
kernel/reboot.c
1139
if (cpu >= num_possible_cpus()) {
kernel/reboot.c
1142
cpu, num_possible_cpus());
kernel/reboot.c
1344
if (cpunum >= num_possible_cpus())
kernel/sched/core.c
10470
max_cids = min(opt_cids + (opt_cids / 4), num_possible_cpus());
kernel/sched/core.c
10480
return max(min(opt_cids - opt_cids / 4, num_possible_cpus() / 2), 1);
kernel/sched/core.c
10860
bitmap_zero(mm_cidmask(mm), num_possible_cpus());
kernel/sched/ext_idle.c
413
return p->nr_cpus_allowed >= num_possible_cpus();
kernel/sched/sched.h
3853
cid = __mm_get_cid(mm, num_possible_cpus());
kernel/time/clockevents.c
459
WARN_ON(num_possible_cpus() > 1);
kernel/time/posix-timers.c
1554
size = roundup_pow_of_two(512 * num_possible_cpus());
kernel/time/tick-sched.c
1617
do_div(offset, num_possible_cpus());
kernel/time/timer_migration.c
1957
unsigned int ncpus = num_possible_cpus();
lib/bucket_locks.c
23
unsigned int nr_pcpus = num_possible_cpus();
lib/debugobjects.c
1517
extras = num_possible_cpus() * ODEBUG_BATCH_SIZE;
lib/objpool.c
139
pool->nr_possible_cpus = num_possible_cpus();
lib/test_objpool.c
287
int max = num_possible_cpus() << 3;
lib/test_objpool.c
453
int max = num_possible_cpus() << 3;
mm/hugetlb.c
4188
num_fault_mutexes = roundup_pow_of_two(8 * num_possible_cpus());
mm/percpu-internal.h
162
return size * num_possible_cpus() + extra_size;
mm/percpu.c
2473
v = num_possible_cpus();
mm/percpu.c
2893
if (wasted > num_possible_cpus() / 3)
mm/percpu.c
3213
nr_g0_units = roundup(num_possible_cpus(), upa);
mm/percpu.c
3222
pages_size = PFN_ALIGN(unit_pages * num_possible_cpus() *
mm/percpu.c
3228
for (unit = 0; unit < num_possible_cpus(); unit++) {
mm/percpu.c
3247
vm.size = num_possible_cpus() * ai->unit_size;
mm/percpu.c
3250
for (unit = 0; unit < num_possible_cpus(); unit++) {
mm/vmalloc.c
5366
int n = clamp_t(unsigned int, num_possible_cpus(), 1, 128);
net/core/dev.c
2884
if (num_possible_cpus() > 1)
net/ipv4/inet_hashtables.c
1309
nblocks = max(2U * L1_CACHE_BYTES / locksz, 1U) * num_possible_cpus();
net/ipv4/tcp_sigpool.c
59
size_t stf_sz = struct_size(stf, scratches, num_possible_cpus());
net/xfrm/xfrm_user.c
1860
if (pcpu_num >= num_possible_cpus()) {
net/xfrm/xfrm_user.c
3014
if (x->pcpu_num >= num_possible_cpus()) {
net/xfrm/xfrm_user.c
959
if (x->pcpu_num >= num_possible_cpus()) {
samples/trace_events/trace-events-sample.h
304
__bitmask( cpus, num_possible_cpus() )
samples/trace_events/trace-events-sample.h
318
__assign_bitmask(cpus, cpumask_bits(mask), num_possible_cpus());
tools/sched_ext/scx_userland.bpf.c
125
return p->nr_cpus_allowed < num_possible_cpus;
tools/sched_ext/scx_userland.bpf.c
314
if (num_possible_cpus == 0) {
tools/sched_ext/scx_userland.bpf.c
316
num_possible_cpus);
tools/sched_ext/scx_userland.bpf.c
36
const volatile u32 num_possible_cpus = 64;
tools/sched_ext/scx_userland.c
388
skel->rodata->num_possible_cpus = libbpf_num_possible_cpus();
tools/sched_ext/scx_userland.c
389
assert(skel->rodata->num_possible_cpus > 0);