_NPCPV
PV_STAT(counter_u64_add(pv_entry_spare, -_NPCPV));
PV_STAT(counter_u64_add(pv_entry_spare, -_NPCPV));
PV_STAT(counter_u64_add(pv_entry_spare, _NPCPV - 1));
for (reclaimed = false; avail < needed; avail += _NPCPV) {
PV_STAT(counter_u64_add(pv_entry_spare, _NPCPV));
pv_entry_max = roundup(pv_entry_max, _NPCPV);
pv_maxchunks = MAX(pv_entry_max / _NPCPV, maxproc);
PV_STAT(pv_entry_spare -= _NPCPV);
PV_STAT(pv_entry_spare -= _NPCPV);
PV_STAT(pv_entry_spare += _NPCPV - 1);
PV_STAT(atomic_subtract_int(&pv_entry_spare, _NPCPV));
PV_STAT(atomic_subtract_int(&pv_entry_spare, _NPCPV));
PV_STAT(atomic_add_int(&pv_entry_spare, _NPCPV - 1));
for (reclaimed = false; avail < needed; avail += _NPCPV) {
PV_STAT(atomic_add_int(&pv_entry_spare, _NPCPV));
pv_maxchunks = MAX(pv_entry_max / _NPCPV, maxproc);
PV_STAT(pv_entry_spare -= _NPCPV);
PV_STAT(pv_entry_spare -= _NPCPV);
PV_STAT(pv_entry_spare += _NPCPV - 1);
pv_entry_max = roundup(pv_entry_max, _NPCPV);
#define PC_FREE1 ((1ul << (_NPCPV % 64)) - 1)
for (reclaimed = false; avail < needed; avail += _NPCPV) {
PV_STAT(atomic_add_int(&pv_entry_spare, _NPCPV));
PV_STAT(atomic_subtract_int(&pv_entry_spare, _NPCPV));
PV_STAT(atomic_subtract_int(&pv_entry_spare, _NPCPV));
PV_STAT(atomic_add_int(&pv_entry_spare, _NPCPV - 1));
#define _NPCM howmany(_NPCPV, 64)
struct pv_entry pc_pventry[_NPCPV];
PV_STAT(atomic_subtract_int(&pv_entry_spare, _NPCPV));
PV_STAT(atomic_add_int(&pv_entry_spare, _NPCPV - 1));
for (reclaimed = false; avail < needed; avail += _NPCPV) {
PV_STAT(atomic_add_int(&pv_entry_spare, _NPCPV));
struct pv_entry pc_pventry[_NPCPV];
#ifndef _NPCPV
#define _NPCM howmany(_NPCPV, __LONG_WIDTH__)
#define PC_FREEL ((1ul << (_NPCPV % __LONG_WIDTH__)) - 1)