atomic_fetchadd_int
if (atomic_fetchadd_int(&np->refcount, -1) == 1)
if (atomic_fetchadd_int(&_thread_active_threads, -1) == 1) {
if (atomic_fetchadd_int(&wnested, 1) == 0)
else if (atomic_fetchadd_int(&wnested, -1) == 1)
old = atomic_fetchadd_int(count, -1);
#define atomic_fetchadd_32 atomic_fetchadd_int
nev = atomic_fetchadd_int(&sc->nev, 1);
nev = atomic_fetchadd_int(&sc->nev, -1);
if (atomic_fetchadd_int(&beio->refcnt, diff) + diff == 0)
io->scsiio.tag_num = atomic_fetchadd_int(&cfi->cur_tag_num, 1);
if (atomic_fetchadd_int(&tio->list->tbdio, -1) == 1)
scsi_u64to8b(atomic_fetchadd_int(&id, 1), &token->body[0]);
io->scsiio.tag_num = atomic_fetchadd_int(&tsoftc->cur_tag_num, 1);
cnt = atomic_fetchadd_int(&softc->ref_flags[token], 1);
cnt = atomic_fetchadd_int(&softc->ref_flags[token], -1);
cnt = atomic_fetchadd_int(&softc->ref_flags[token], 1);
cnt = atomic_fetchadd_int(&softc->ref_flags[token], -1);
cnt = atomic_fetchadd_int(&softc->ref_flags[token], -1);
return atomic_fetchadd_int(&v->counter, -1) - 1;
return i + atomic_fetchadd_int(&v->counter, i);
return atomic_fetchadd_int(&v->counter, -i) - i;
return atomic_fetchadd_int(&v->counter, 1) + 1;
atomic_fetchadd_int(&sc->ec_sci_pend, 1) == 0) {
atomic_fetchadd_int(&sc->ec_sci_pend, 1) == 0) {
while (!atomic_fetchadd_int(&sc->packet_reply,0)) {
if (atomic_fetchadd_int(&icp->ref_cnt, -1) != 1)
if (atomic_fetchadd_int(&icp->ref_cnt, -1) == 1)
if (atomic_fetchadd_int(&icp->ref_cnt, -1) == 1)
if (atomic_fetchadd_int(&icp->ref_cnt, -1) == 1)
if (atomic_fetchadd_int(&e->refcnt, 1) == 0) {
if (atomic_fetchadd_int(&e->refcnt, -1) == 1)
if (atomic_fetchadd_int(&clm->refcount, -1) == 1) {
if (atomic_fetchadd_int(&clm->refcount, -1) == 1) {
if (atomic_fetchadd_int(&e->refcnt, -1) == 1) {
if (atomic_fetchadd_int(&drb->refs, -1) == 1)
if (atomic_fetchadd_int(&synqe->ok_to_respond, 1) == 0) {
q_idx = atomic_fetchadd_int(&vi->txq_rr, 1);
q_idx = atomic_fetchadd_int(&vi->rxq_rr, 1);
if (atomic_fetchadd_int(&e->refcnt, 1) == 0) /* 0 -> 1 transition */
#define atomic_add_return(v, p) (atomic_fetchadd_int(p, v) + (v))
#define atomic_sub_return(v, p) (atomic_fetchadd_int(p, -(v)) - (v))
idx = atomic_fetchadd_int(&dp_bus_counter, 1);
old = atomic_fetchadd_int(&bo->list_kref, -count);
unsigned pool_offset = atomic_fetchadd_int(&start_pool, 1);
thread_id = atomic_fetchadd_int(&ctx->thread_counter, 1);
thread_id = atomic_fetchadd_int(&ctx->thread_counter, 1);
rid = atomic_fetchadd_int(&sc->hn_rndis_rid, 1);
sc->hn_cpu = atomic_fetchadd_int(&hn_cpu_index, ring_cnt) % mp_ncpus;
if (atomic_fetchadd_int(&txd->refs, -1) != 1)
gpadl = atomic_fetchadd_int(&sc->vmbus_gpadl, 1);
cpu = atomic_fetchadd_int(&vmbus_chan_nextcpu, 1) % mp_ncpus;
refs = atomic_fetchadd_int(&chan->ch_refs, -1);
if (atomic_fetchadd_int(&isp->ref_cnt, -1) == 1)
if (atomic_fetchadd_int(&isp->ref_cnt, -1) == 1)
#define atomic_add_return(v, p) (atomic_fetchadd_int(p, v) + (v))
#define atomic_sub_return(v, p) (atomic_fetchadd_int(p, -(v)) - (v))
return 1 + atomic_fetchadd_int(&(v)->val, 1);
if (atomic_fetchadd_int(&ndisk->cur_depth, -1) == 1 &&
if (atomic_fetchadd_int(&ndisk->ordered_in_flight, 1) == 0 &&
if (atomic_fetchadd_int(&ndisk->cur_depth, -1) == 1 &&
inbed = atomic_fetchadd_int(&parent->bio_inbed, 1) + 1;
tth->idx = atomic_fetchadd_int(&io_test->td_idx, 1);
idx = atomic_fetchadd_int(&io_test->td_idx, 1);
lev = atomic_fetchadd_int(&sc->grab_level, 1);
head = atomic_fetchadd_int(&sc_cnputc_loghead, 1);
vd->vd_unit = atomic_fetchadd_int(&vt_unit, 1);
if (atomic_fetchadd_int(&reqlist->pendcnt, -1) == 1)
atomic_fetchadd_int(&autofs_softc->sc_last_request_id, 1);
anp->an_fileno = atomic_fetchadd_int(&->am_last_fileno, 1);
nextconn = atomic_fetchadd_int(&nmp->nm_nextaconn, 1);
writerefs = atomic_fetchadd_int(&vp->v_writecount, ap->a_inc);
if (atomic_fetchadd_int(&g_eli_devs, -1) == 1)
if (atomic_fetchadd_int(&g_eli_umaoutstanding, -1) == 1 &&
active = atomic_fetchadd_int(&sc->sc_active, -1) - 1;
active = atomic_fetchadd_int(&sc->sc_active, SC_A_DESTROY);
atomic_fetchadd_int(&NVIDIANodeID, 1));
atomic_fetchadd_int(&NVIDIANodeID, 1));
return (atomic_fetchadd_int((volatile u_int *)p, (u_int)v));
#define atomic_fetchadd_32 atomic_fetchadd_int
*st->oldval = atomic_fetchadd_int((int *)kva, st->oparg);
openfiles_new = atomic_fetchadd_int(&openfiles, 1) + 1;
old = atomic_fetchadd_int(&kq_ncallouts, -1);
if (atomic_fetchadd_int(&kq_ncallouts, 1) + 1 > kq_calloutmax) {
if (atomic_fetchadd_int(&kq_ncallouts, 1) + 1 > kq_calloutmax) {
gen = atomic_fetchadd_int(&exec_args_gen, 1) + 1;
nprocs_new = atomic_fetchadd_int(&nprocs, 1) + 1;
if (*refcnt == 1 || atomic_fetchadd_int(refcnt, -1) == 1) {
if (*refcnt == 1 || atomic_fetchadd_int(refcnt, -1) == 1) {
old = atomic_fetchadd_int(&rms->debug_readers, 1);
old = atomic_fetchadd_int(&rms->debug_readers, -1);
if (atomic_fetchadd_int(&oid->oid_running, -1) == 1 &&
nthreads_new = atomic_fetchadd_int(&nthreads, 1) + 1;
if (atomic_fetchadd_int(&ds->start_count, 1) == ds->end_count) {
if (atomic_fetchadd_int(&epoch->e_drain_count, -1) == 1) {
return (atomic_fetchadd_int(&s->s_wr.seq, SMR_SEQ_INCR) + SMR_SEQ_INCR);
old = atomic_fetchadd_int(&node->mn_refcount, -1);
freebufs = atomic_fetchadd_int(&bd->bd_freebuffers, -1);
atomic_fetchadd_int(&bufobj_cleanq, 1) % buf_domains;
num = atomic_fetchadd_int(&bd->bd_numdirtybuffers, -1);
num = atomic_fetchadd_int(&bd->bd_numdirtybuffers, 1);
if (atomic_fetchadd_int(&bd->bd_running, 1) == 0) {
c = atomic_fetchadd_int(&neg_cycle, 1) + 1;
count = atomic_fetchadd_int(&inotify_watches, 1);
old = atomic_fetchadd_int(&vp->v_usecount, 1);
old = atomic_fetchadd_int(&vp->v_holdcnt, -1);
old = atomic_fetchadd_int(&vp->v_holdcnt, 1);
int old = atomic_fetchadd_int(&vp->v_holdcnt, 1);
if (da == NULL || atomic_fetchadd_int(&da->refcnt, -1) > 1)
#define SCTP_DECREMENT_AND_CHECK_REFCOUNT(addr) (atomic_fetchadd_int(addr, -1) == 1)
oldval = atomic_fetchadd_int(addr, -val); \
oldval = atomic_fetchadd_int(addr, -val); \
chk->rec.data.tsn = atomic_fetchadd_int(&asoc->sending_seq, 1);
chk->rec.data.tsn = atomic_fetchadd_int(&stcb->asoc.sending_seq, 1);
refcnt = atomic_fetchadd_int(&log_entry->tlm_refcnt, 1);
refcnt = atomic_fetchadd_int(&log_entry->tlm_refcnt, -1);
atomic_fetchadd_int(&tlb->tlb_reqcnt, 1);
res = atomic_fetchadd_int(&tcp_trace_point_count, -1);
blk->tfb_id = atomic_fetchadd_int(&next_tcp_stack_id, 1);
atomic_fetchadd_int(&number_of_tcp_connections_pacing, 1);
atomic_fetchadd_int(&number_of_dgp_connections, 1);
ret = atomic_fetchadd_int(&number_of_dgp_connections, -1);
ret = atomic_fetchadd_int(&number_of_tcp_connections_pacing, -1);
if (atomic_fetchadd_int(tp->t_tfo_pending, 1) <=
inp->inp_flowid = atomic_fetchadd_int(&udp_flowid, 1);
xid = atomic_fetchadd_int(&nlm_xid, 1);
xid = atomic_fetchadd_int(&nlm_xid, 1);
xid = atomic_fetchadd_int(&nlm_xid, 1);
xid = atomic_fetchadd_int(&nlm_xid, 1);
atomic_fetchadd_int(&moea64_bpvo_pool_index, 1)];
va = atomic_fetchadd_int(&tlb1_map_base, size);
#define atomic_fetchadd_32 atomic_fetchadd_int
idx = atomic_fetchadd_int(&intrcnt_index, 1);
i->cntindex = atomic_fetchadd_int(&intrcnt_index, 1);
g = atomic_fetchadd_int(&pool->sp_nextgroup, 1) % pool->sp_groupcount;
old = atomic_fetchadd_int(&bc->__count, n);
old = atomic_fetchadd_int(&bc->__count, -n);
old = atomic_fetchadd_int(count, -n);
old = atomic_fetchadd_int(count, 1);
old = atomic_fetchadd_int(count, n);
if (atomic_fetchadd_int(&uma_reclaim_needed, 1) == 0)
old = atomic_fetchadd_int(&m->ref_count, 1);
old = atomic_fetchadd_int(&m->ref_count, -val);
old_vote = atomic_fetchadd_int(&vm_pageout_oom_vote, 1);
if (atomic_fetchadd_int(&vmd->vmd_pageout_wanted, 1) == 0) {
old = atomic_fetchadd_int(&vmd->vmd_free_count, adj);
cnt = atomic_fetchadd_int(&cur->count, -1);