atomic_add_64_nv
event->event_id = atomic_add_64_nv(&event_id_counter, 1);
size = atomic_add_64_nv((uint64_t *)&rp->r_size, 0);
if (atomic_add_64_nv((&(vp->v_mmap_write)), 0))
if ((atomic_add_64_nv((&(vp->v_mmap_read)), 0)) &&
(atomic_add_64_nv((&(vp->v_mmap_write)), 0)))
if ((atomic_add_64_nv((&(vp->v_mmap_read)), 0)) ||
(atomic_add_64_nv((&(vp->v_mmap_write)), 0)))
if (atomic_add_64_nv((&(vp->v_mmap_read)), 0))
ASSERT3S(atomic_add_64_nv(&arc_loaned_bytes, 0), >=, 0);
ASSERT3S(atomic_add_64_nv(&arc_loaned_bytes, 0), >=, 0);
object = atomic_add_64_nv(cpuobj, dn_slots) - dn_slots;
if (atomic_add_64_nv(&mc->mc_aliquot, asize) >=
count = atomic_add_64_nv(&(rc)->rc_count, number);
count = atomic_add_64_nv(&(rc)->rc_count, -number);
atomic_add_64_nv(&(rc)->rc_count, number)
atomic_add_64_nv(&(rc)->rc_count, -number)
SE_SEQ(ev) = eid->eid_seq = atomic_add_64_nv(&kernel_event_id,
return (atomic_add_64_nv(&kernel_event_id, (uint64_t)1));
extern uint64_t atomic_add_64_nv(volatile uint64_t *, int64_t);