atomic_cmpset_acq_int
if (atomic_cmpset_acq_int(&(*sem)->count, val, val - 1))
if (atomic_cmpset_acq_int(&(*sem)->count, val, val - 1))
if (atomic_cmpset_acq_int(&sem->_kern._count, val, val - 1))
if (atomic_cmpset_acq_int(&sem->_kern._count, val,
if (!atomic_cmpset_acq_int(&p->state, GMON_PROF_ON, GMON_PROF_BUSY))
if (atomic_cmpset_acq_int(&l->lock, 0, WAFLAG))
#define atomic_cmpset_acq_32 atomic_cmpset_acq_int
} while (atomic_cmpset_acq_int(addr, old, new) == 0);
while ((atomic_cmpset_acq_int(&sc->inflight, INFLIGHT_NONE,
if (atomic_cmpset_acq_int(&sc->semaphore, 0, 1) == 0) {
if (atomic_cmpset_acq_int(&sc->recovery_mode, 0, 1)) {
atomic_cmpset_acq_int(&sc->recovery_mode, 1, 0);
#define NM_ATOMIC_TEST_AND_SET(p) (!atomic_cmpset_acq_int((p), 0, 1))
if (atomic_cmpset_acq_int(&NFSD_VNET(nfsrv_dnsnamelen), 0, 0) ==
atomic_cmpset_acq_int(&onethread, 0, 1) != 0) {
if (atomic_cmpset_acq_int(&onethread, 0, 1) == 0)
#define atomic_cmpset_acq_32 atomic_cmpset_acq_int
atomic_cmpset_acq_int((volatile u_int *)(dst), (u_int)(old), \
if (atomic_cmpset_acq_int(&devpoll_run, 0, 1)) {
while (atomic_cmpset_acq_int(&ithd->it_need, 1, 0) != 0) {
if (atomic_cmpset_acq_int(&kexec_loading, false, true) == 0)
atomic_cmpset_acq_int(&cr->cr_lock, 0, 1)) {
if (atomic_cmpset_acq_int(&ngq->q_flags, t,
} else if (atomic_cmpset_acq_int(&ngq->q_flags, OP_PENDING,
if (atomic_cmpset_acq_int(&node->nd_input_queue.q_flags, t,
if (atomic_cmpset_acq_int(&node->nd_input_queue.q_flags, 0,
} while(atomic_cmpset_acq_int(cntr, old, new) == 0);
} while (atomic_cmpset_acq_int(cntr, old, new) == 0);
#define atomic_cmpset_acq_32 atomic_cmpset_acq_int
#define atomic_cmpset_acq_ptr atomic_cmpset_acq_int
if (atomic_cmpset_acq_int(&m->busy_lock, VPB_UNBUSIED,
while (atomic_cmpset_acq_int(&ap_boot_lock, 0, 1) == 0)
if (atomic_cmpset_acq_int(&nmi_kdb_lock, 0, 1)) {
while (!atomic_cmpset_acq_int(&nmi_kdb_lock, 0, 1))