CPU_SETSIZE
printset((struct bitset *)&mask, CPU_SETSIZE);
for (cpu = 0; cpu < CPU_SETSIZE; cpu++) {
return (parselist(list, (struct bitset *)mask, CPU_SETSIZE));
if (curnum > CPU_SETSIZE)
"Only %d cpus supported", CPU_SETSIZE);
snprintf(testvalue, sizeof(testvalue), "%d", CPU_SETSIZE + 1);
for (int i = 0; i < __bitset_words(CPU_SETSIZE); i++) {
(2 * __bitset_words(CPU_SETSIZE) - 1) * (_BITSET_BITS / 8),
(__bitset_words(CPU_SETSIZE) - 1);
bwp += (2 * __bitset_words(CPU_SETSIZE) - 1);
bwp += (2 * __bitset_words(CPU_SETSIZE) - 1) *
__bitset_words(CPU_SETSIZE) - 1;
#define VM_MAXCPU MIN(0xffff - 1, CPU_SETSIZE)
CPU_SETSIZE);
p = buf[bitset_strscan((struct bitset *)set, CPU_SETSIZE, buf)];
ddb_display_bitset((const struct bitset *)set, CPU_SETSIZE);
for (n = 0; n < CPU_SETSIZE; n++) {
for (i = 0; i < CPU_SETSIZE; i++) {
#ifndef CPU_SETSIZE
__BITSET_DEFINE(_cpuset, CPU_SETSIZE);
#define _NCPUWORDS __bitset_words(CPU_SETSIZE)
#define CPU_CLR(n, p) __BIT_CLR(CPU_SETSIZE, n, p)
#define CPU_COPY(f, t) __BIT_COPY(CPU_SETSIZE, f, t)
#define CPU_ISSET(n, p) __BIT_ISSET(CPU_SETSIZE, n, p)
#define CPU_SET(n, p) __BIT_SET(CPU_SETSIZE, n, p)
#define CPU_ZERO(p) __BIT_ZERO(CPU_SETSIZE, p)
#define CPU_FILL(p) __BIT_FILL(CPU_SETSIZE, p)
#define CPU_SETOF(n, p) __BIT_SETOF(CPU_SETSIZE, n, p)
#define CPU_EQUAL(p, c) (__BIT_CMP(CPU_SETSIZE, p, c) == 0)
#define CPU_EMPTY(p) __BIT_EMPTY(CPU_SETSIZE, p)
#define CPU_ISFULLSET(p) __BIT_ISFULLSET(CPU_SETSIZE, p)
#define CPU_SUBSET(p, c) __BIT_SUBSET(CPU_SETSIZE, p, c)
#define CPU_OVERLAP(p, c) __BIT_OVERLAP(CPU_SETSIZE, p, c)
#define CPU_CMP(p, c) __BIT_CMP(CPU_SETSIZE, p, c)
#define CPU_OR(d, s1, s2) __BIT_OR2(CPU_SETSIZE, d, s1, s2)
#define CPU_ORNOT(d, s1, s2) __BIT_ORNOT2(CPU_SETSIZE, d, s1, s2)
#define CPU_AND(d, s1, s2) __BIT_AND2(CPU_SETSIZE, d, s1, s2)
#define CPU_ANDNOT(d, s1, s2) __BIT_ANDNOT2(CPU_SETSIZE, d, s1, s2)
#define CPU_XOR(d, s1, s2) __BIT_XOR2(CPU_SETSIZE, d, s1, s2)
#define CPU_CLR_ATOMIC(n, p) __BIT_CLR_ATOMIC(CPU_SETSIZE, n, p)
#define CPU_TEST_CLR_ATOMIC(n, p) __BIT_TEST_CLR_ATOMIC(CPU_SETSIZE, n, p)
#define CPU_SET_ATOMIC(n, p) __BIT_SET_ATOMIC(CPU_SETSIZE, n, p)
#define CPU_SET_ATOMIC_ACQ(n, p) __BIT_SET_ATOMIC_ACQ(CPU_SETSIZE, n, p)
#define CPU_TEST_SET_ATOMIC(n, p) __BIT_TEST_SET_ATOMIC(CPU_SETSIZE, n, p)
#define CPU_AND_ATOMIC(n, p) __BIT_AND_ATOMIC(CPU_SETSIZE, n, p)
#define CPU_OR_ATOMIC(d, s) __BIT_OR_ATOMIC(CPU_SETSIZE, d, s)
#define CPU_COPY_STORE_REL(f, t) __BIT_COPY_STORE_REL(CPU_SETSIZE, f, t)
#define CPU_FFS(p) __BIT_FFS(CPU_SETSIZE, p)
#define CPU_FLS(p) __BIT_FLS(CPU_SETSIZE, p)
#define CPU_FOREACH_ISSET(i, p) __BIT_FOREACH_ISSET(CPU_SETSIZE, i, p)
#define CPU_FOREACH_ISCLR(i, p) __BIT_FOREACH_ISCLR(CPU_SETSIZE, i, p)
#define CPU_COUNT(p) ((int)__BIT_COUNT(CPU_SETSIZE, p))
#if defined(CPU_SETSIZE) && 0
#if defined(CPU_SETSIZE) && 0
if (sched_setaffinity(0, CPU_SETSIZE, &mask) < 0)
#if defined(CPU_SETSIZE) && 0
if (sched_setaffinity(0, CPU_SETSIZE, &mask) < 0)
for (ncpu = 0, i = 0; i < CPU_SETSIZE; i++) {
for (i = 0, j = 0; i < CPU_SETSIZE; i++)
for (cpu = 0; cpu < CPU_SETSIZE; cpu++) {
maxid = MIN(size / CPUSTATES / sizeof(long) - 1, CPU_SETSIZE - 1);
if (pcpu < 0 || pcpu >= CPU_SETSIZE) {
"0 to %d\n", pcpu, CPU_SETSIZE - 1);
if (pcpu < 0 || pcpu >= CPU_SETSIZE)
pcpu, CPU_SETSIZE - 1);
for (i = 0; i < CPU_SETSIZE; i++) {