atomic_cas_uint
locked = atomic_cas_uint(&cmtx->lock,
locked = atomic_cas_uint(&cmtx->lock,
locked = atomic_cas_uint(&cmtx->lock,
locked = atomic_cas_uint(&cmtx->lock,
if (atomic_cas_uint(&cmtx->lock,
if (atomic_cas_uint(&mutex->lock, UNLOCKED, LOCKED) == UNLOCKED) {
lock = atomic_cas_uint(&mutex->lock, UNLOCKED, LOCKED);
} while (atomic_cas_uint(&rwlock->value, val, val + 1) != val);
if (atomic_cas_uint(&rwlock->value, val, new) == val) {
if (atomic_cas_uint(&rwlock->value, UNLOCKED, WRITER) != UNLOCKED)
if (atomic_cas_uint(&rwlock->value, val, new) == val) {
} while (atomic_cas_uint(&rwlock->value, val, new) != val);
if (atomic_cas_uint(&sem->value, val, val - 1) == val) {
if (atomic_cas_uint(&sem->value, val, val - 1) == val) {
if ((int)atomic_cas_uint(connectp, -1, sock) != -1) {
if ((int)atomic_cas_uint(acceptp, -1, sock) != -1) {
if ((int)atomic_cas_uint(sockp, sock, -1) != sock) {
if ((int)atomic_cas_uint(sockp, sock, -1) != sock) {
if ((int)atomic_cas_uint(sockp, sock, -1) != sock) {
if ((int)atomic_cas_uint(connectp, -1, sock) != -1) {
if ((int)atomic_cas_uint(acceptp, -1, sock) != -1) {
if ((int)atomic_cas_uint(sockp, sock, -1) != sock) {
if ((int)atomic_cas_uint(sockp, sock, -1) != sock) {
if ((int)atomic_cas_uint(sockp, sock, -1) != sock) {
if (atomic_cas_uint(&pj->pj_state,
} while (atomic_cas_uint(lock, old, new) != old);
} while (atomic_cas_uint(lock, old, new) != old);
while (atomic_cas_uint(&tlb_shoot_lock.lock, 0, 1) != 0) {
while (atomic_cas_uint(&tsc_egress_barrier, 2, 0) != 2)
if (atomic_cas_uint(&vcpu->vc_state, old, next) != old) {
} while (atomic_cas_uint(lock, old, new) != old);
} while (atomic_cas_uint(lock, old, new) != old);
while (atomic_cas_uint(&tlb_shoot_lock.lock, 0, 1) != 0) {
((unsigned long)atomic_cas_uint((unsigned int *)p, (unsigned int)o, \
((void *)atomic_cas_uint((void *)p, (unsigned int)o, (unsigned int)n))
return atomic_cas_uint((unsigned int *)pte, o, n);
if (atomic_cas_uint(&usedsr[tblidx], bits,
if (atomic_cas_uint(&pmap_vsid[vsid / 32], bits,
if (atomic_cas_uint(&pmap_vsid[vsid / 32], bits,
_def_atomic_op_nv(_atomic_add_int_nv, unsigned int, atomic_cas_uint, +)
_def_atomic_op_nv(_atomic_sub_int_nv, unsigned int, atomic_cas_uint, -)
r = atomic_cas_uint(uip, e, e | v);
r = atomic_cas_uint(uip, e, e & ~v);
_def_atomic_swap(_atomic_swap_uint, unsigned int, atomic_cas_uint)
previous = atomic_cas_uint(s, expected, expected | (1UL << bit));
previous = atomic_cas_uint(s, expected, expected & ~(1UL << bit));
while (atomic_cas_uint(ptr, flags, GTF_invalid) != flags) {
if (atomic_cas_uint(&vm->vm_dying, 0, 1) == 1) {
if (atomic_cas_uint(&nfp->f_count, count, count + 1) == count)
} while (atomic_cas_uint(&fp->f_flag, prev, tmp) != prev);
ogen = atomic_cas_uint(&pcl->pcl_gen, gen, ngen);
if (atomic_cas_uint(&pr->ps_siglist, ps_siglist,
} while (atomic_cas_uint(valp, oldval, newval) != oldval);
} while (atomic_cas_uint(&securelevel, oldval, newval) !=
oldval == atomic_cas_uint(&pool_debug, oldval, newval))
} while (atomic_cas_uint(line, prev, prev | LOCATION_REPORTED) != prev);
if (atomic_cas_uint(&kubsan_slot, nslots, 0) != nslots) {
atomic_cas_uint(&rt->rt_mtu, 0, LOMTU);
atomic_cas_uint(&r->rule_flag, rule_flag,
atomic_cas_uint(&rt->rt_mtu, mtu, nhmtu);
atomic_cas_uint(&rt->rt_mtu, rtmtu, mtu);
oldval == atomic_cas_uint(&ip_mtudisc, oldval, newval) &&
atomic_cas_uint(&ro->ro_rt->rt_mtu, rtmtu,
atomic_cas_uint(&rt->rt_mtu, rtmtu, mtu);
atomic_cas_uint(&rt->rt_mtu, rtmtu, mtu);
#ifndef atomic_cas_uint