zpcpu_get
*(zpcpu_get(c)) = 0;
*zpcpu_get(base) = __n; \
pcidp = zpcpu_get(pmap->pm_pcidp);
atomic_store_64((uint64_t *)zpcpu_get(arg), 0);
atomic_add_64((uint64_t *)zpcpu_get(c), inc);
atomic_add_64((uint64_t *)zpcpu_get(c), inc);
*(uint64_t *)zpcpu_get(c) += (inc); \
*(uint64_t *)zpcpu_get(c) += inc;
mtsp = zpcpu_get(mtip->mti_stats);
mtsp = zpcpu_get(mtip->mti_stats);
return (zpcpu_get(rms->pcpu));
return (zpcpu_get(epoch->e_pcpu_record));
KASSERT((zpcpu_get(smr)->c_flags & SMR_LAZY) == 0,
self = zpcpu_get(smr);
KASSERT(!wait || (zpcpu_get(smr)->c_flags & SMR_LAZY) == 0,
self = zpcpu_get(smr);
*zpcpu_get(peer->last_active) = time_uptime;
pcpu = zpcpu_get(pfcu64->pfcu64_pcpu);
*zpcpu_get((prule)->timestamp) = time_second; \
ip->ip_id = htons((*(uint64_t *)zpcpu_get(V_ip_id)) & 0xffff);
*(uint64_t *)zpcpu_get(c) += (inc); \
*(zpcpu_get((counter_u64_t *)arg)) = 0;
atomic_add_64((uint64_t *)zpcpu_get(c), inc);
(curthread->td_critnest != 0 && zpcpu_get((smr))->c_seq != SMR_SEQ_INVALID)
#define vfs_mount_pcpu(mp) zpcpu_get(mp->mnt_pcpu)
__typeof(val) *_ptr = zpcpu_get(base); \
__typeof(val) *_ptr = zpcpu_get(base); \
__typeof(val) *_ptr = zpcpu_get(base); \
__typeof(val) *_ptr = zpcpu_get(base); \
smr = zpcpu_get(smr);
smr = zpcpu_get(smr);
smr = zpcpu_get(smr);
smr = zpcpu_get(smr);
return (smr_shared_current(zpcpu_get(smr)->c_shared));
zpcpu_get(smrs_smr)->c_seq, smrs->generation, smrs->count);