CPU_COUNT
cpus = CPU_COUNT(&mask);
return (CPU_COUNT(&cpus));
if (CPU_COUNT(mask) < ncpu)
if (CPU_COUNT(mask) < ncpu) {
ncpu = CPU_COUNT(&mask);
ATF_REQUIRE_EQ(1, CPU_COUNT(&mask));
count = CPU_COUNT(mask);
ATF_REQUIRE_EQ(count - 1, CPU_COUNT(mask));
ATF_REQUIRE(sysconf(_SC_NPROCESSORS_ONLN) > CPU_COUNT(&jmask));
ATF_REQUIRE_EQ(CPU_COUNT(&jmask) + 1, ninfo.nproc_init);
ATF_REQUIRE_EQ(CPU_COUNT(&jmask) + 1, ninfo.nproc_global);
ATF_REQUIRE_EQ(CPU_COUNT(&jmask), ninfo.nproc_final);
dom->cg_count = CPU_COUNT(&dom->cg_mask);
if (CPU_COUNT(&all_cpus) == 1)
DBGPR("ncpu %d intrcpu %d\n", mp_ncpus, CPU_COUNT(&lcpus));
pdata->tx_ring_count = min(CPU_COUNT(&lcpus), pdata->hw_feat.tx_ch_cnt);
pdata->rx_ring_count = min(CPU_COUNT(&lcpus), pdata->hw_feat.rx_ch_cnt);
DPRINTF(dev, "We have %d cpu on this dev\n", CPU_COUNT(&sc->cpus));
queues = CPU_COUNT(&cpus);
queues = CPU_COUNT(&cpus);
rdma = min(CPU_COUNT(&cpus), ICE_RDMA_MAX_MSIX);
ncpus = CPU_COUNT(&map);
n = n % CPU_COUNT(&ctx->ifc_cpus);
cores_consumed = CPU_COUNT(&assigned_cpus);
MPASS(CPU_COUNT(&ctx->ifc_cpus) > 0);
queues = imin(CPU_COUNT(&ctx->ifc_cpus), queues);
CPU_COUNT(&ctx->ifc_cpus), queuemsgs, admincnt);
group->cg_count = CPU_COUNT(cpus);
ncores = CPU_COUNT(&domcpus) / smp_threads_per_core;
KASSERT(CPU_COUNT(&domcpus) % smp_threads_per_core == 0,
int i = mftb() % CPU_COUNT(&cpumask);
domain_cpus = CPU_COUNT(&cpuset_domain[vmd->vmd_domain]);
eligible_cpus -= CPU_COUNT(&cpuset_domain[i]);
ncpus = CPU_COUNT(&cpumask);