ncpus
__cpuset_alloc(size_t ncpus)
return (malloc(CPU_ALLOC_SIZE(ncpus)));
#define ncpus JEMALLOC_N(ncpus)
ncpus = aux->a_un.a_val;
ncpus = aux->a_un.a_val;
if (ncpus != 0) {
*(int *)buf = ncpus;
static int pagesize, osreldate, canary_len, ncpus, pagesizes_len, bsdflags;
int fd, ncpus;
if (_elf_aux_info(AT_NCPUS, &ncpus, sizeof(ncpus)) != 0 ||
len = ncpus * sizeof(*pvclock_timeinfos);
int rid, err, ncpus, i;
ncpus = OF_getencprop_alloc_multi(node, "interrupt-affinity",
if (ncpus < 0) {
if (has_affinity && i >= ncpus) {
for (i = 0; i < sc->config->ncpus; i++) {
int ncpus;
.ncpus = 4,
acpi_pxm_init(int ncpus, vm_paddr_t maxphys)
max_cpus = ncpus;
int acpi_pxm_init(int ncpus, vm_paddr_t maxphys);
int error, i, ncpus, nclasses;
ncpus = pmc_cpu_max();
error = pmc_tsc_initialize(pmc_mdep, ncpus);
error = pmc_ibs_initialize(pmc_mdep, ncpus);
pmc_ibs_initialize(struct pmc_mdep *pmc_mdep, int ncpus)
ncpus = pmc_cpu_max();
error = pmc_tsc_initialize(pmc_mdep, ncpus);
error = pmc_core_initialize(pmc_mdep, ncpus, verov);
error = pmc_uncore_initialize(pmc_mdep, ncpus);
int error, family, model, nclasses, ncpus, stepping, verov;
int domain, ncpus, total;
ncpus = pmc_dom_hdrs[domain]->pdbh_ncpus;
total = ncpus * pmc_nlogbuffers_pcpu;
u_int ncpus;
bps->kthreads = nm_os_malloc(sizeof(struct nm_bdg_kthread) * bps->ncpus);
for (i = 0; i < bps->ncpus; i++) {
int all = (bps->ncpus == 1 &&
for (i = 0; i < bps->ncpus; i++) {
for (i = 0; i < bps->ncpus; i++) {
bps->ncpus = req_cpus;
int curcpumap, ncpus = 0;
ncpus = CPU_COUNT(&map);
if (ncpus == 0)
smp_rv_ncpus = ncpus;
while (atomic_load_acq_int(&smp_rv_waiters[3]) < ncpus)
SYSCTL_INT(_net_inet_rss, OID_AUTO, ncpus, CTLFLAG_RD, &rss_ncpus, 0,
uint32_t ncpus = mp_ncpus ? mp_ncpus : MAXCPU;
pace->rp_num_hptss = ncpus;
int ncpus, i, error;
ncpus = 0;
if (CPU_ISSET(cpu, &cpumask)) ncpus++;
i = mftb() % ncpus;
ncpus = 0;
if (ncpus == i)
ncpus++;
int ncpus, i, error = -1;
ncpus = 0;
if (CPU_ISSET(cpu, &cpumask)) ncpus++;
i = mftb() % ncpus;
ncpus = 0;
if (ncpus == i)
ncpus++;
u_int ncpus = mp_ncpus ? mp_ncpus : MAXCPU;
taskqueue_start_threads(&kfrwk.kfrwk_tq, ncpus, PI_NET, "[kt_frwk task]");
kfrwk.kfrwk_waiting = ncpus;
u_int ncpus = mp_ncpus ? mp_ncpus : MAXCPU;
if (ncpus != kfrwk.kfrwk_waiting) {
if (ncpus < 2) {
long ncpus = 0;
ncpus = sysconf(_SC_NPROCESSORS_ONLN);
if (ncpus < 0) {
u_int counter, ncpus;
len = sizeof(ncpus);
if (sysctlbyname("kern.smp.cpus", &ncpus, &len, NULL, 0) < 0)
if (len != sizeof(ncpus))
if (ncpus < 2)
int ncpus;
sysctl(mib, 2, &ncpus, &len, NULL, 0);
ncpus = sysconf(_SC_NPROCESSORS_ONLN);
ncpus = sysinfo.dwNumberOfProcessors;
ncpus = 1;
return (ncpus);
int ncpus;
len = sizeof(ncpus);
if (sysctlbyname(SYSCTLNAME_CPUS, &ncpus, &len, NULL, 0) < 0)
if (len != sizeof(ncpus))
CPU_SET(workernum % ncpus, &mask);
int ncpus;
len = sizeof(ncpus);
if (sysctlbyname(SYSCTLNAME_CPUS, &ncpus, &len, NULL, 0) < 0)
if (len != sizeof(ncpus))
CPU_SET(workernum % ncpus, &mask);
num_cpus = statics->ncpus;
lcpustates = calloc(num_cpustates * sizeof(int), statics->ncpus);
static int ncpus;
if (ncpus == 1)
y_mem += ncpus - 1;
y_arc += ncpus - 1;
y_carc += ncpus - 1;
y_swap += ncpus - 1;
y_idlecursor += ncpus - 1;
y_message += ncpus - 1;
y_header += ncpus - 1;
y_procs += ncpus - 1;
Header_lines += ncpus - 1;
ncpus = CPU_COUNT(&cpumask);
assert(ncpus > 0);
pcpu_cp_old = calloc(ncpus * CPUSTATES, sizeof(long));
pcpu_cp_diff = calloc(ncpus * CPUSTATES, sizeof(long));
pcpu_cpu_states = calloc(ncpus * CPUSTATES, sizeof(int));
statics->ncpus = ncpus;
si->ncpus = ncpus;
si->ncpus = 1;
int ncpus;
int ncpus;
uint64_t ncpus;
ncpus = (uint64_t)cpu_sockets * cpu_cores * cpu_threads;
if (ncpus > UINT16_MAX)
(uintmax_t)ncpus);
if (guest_ncpus != (int)ncpus)
guest_ncpus = ncpus;
int64_t ncpus;
JSON_GET_INT_OR_RETURN(JSON_NCPUS_KEY, obj, &ncpus, 0);
return ((int)ncpus);
static int ncpus = 0;
ncpus = cp_times_len / (sizeof(long) * CPUSTATES);
cp_times_len = sizeof(long) * CPUSTATES * ncpus;
for (cpu = 0; cpu < ncpus; cpu++) {