Symbol: atomic_cas_uint
lib/libc/thread/rthread.c
106
locked = atomic_cas_uint(&cmtx->lock,
lib/libc/thread/rthread.c
118
locked = atomic_cas_uint(&cmtx->lock,
lib/libc/thread/rthread.c
131
locked = atomic_cas_uint(&cmtx->lock,
lib/libc/thread/rthread.c
144
locked = atomic_cas_uint(&cmtx->lock,
lib/libc/thread/rthread.c
92
if (atomic_cas_uint(&cmtx->lock,
lib/libc/thread/rthread_mutex.c
101
if (atomic_cas_uint(&mutex->lock, UNLOCKED, LOCKED) == UNLOCKED) {
lib/libc/thread/rthread_mutex.c
179
lock = atomic_cas_uint(&mutex->lock, UNLOCKED, LOCKED);
lib/librthread/rthread_rwlock.c
108
} while (atomic_cas_uint(&rwlock->value, val, val + 1) != val);
lib/librthread/rthread_rwlock.c
149
if (atomic_cas_uint(&rwlock->value, val, new) == val) {
lib/librthread/rthread_rwlock.c
183
if (atomic_cas_uint(&rwlock->value, UNLOCKED, WRITER) != UNLOCKED)
lib/librthread/rthread_rwlock.c
226
if (atomic_cas_uint(&rwlock->value, val, new) == val) {
lib/librthread/rthread_rwlock.c
273
} while (atomic_cas_uint(&rwlock->value, val, new) != val);
lib/librthread/rthread_sem.c
292
if (atomic_cas_uint(&sem->value, val, val - 1) == val) {
lib/librthread/rthread_sem.c
67
if (atomic_cas_uint(&sem->value, val, val - 1) == val) {
regress/sys/netinet/tcpthread/tcpthread.c
131
if ((int)atomic_cas_uint(connectp, -1, sock) != -1) {
regress/sys/netinet/tcpthread/tcpthread.c
212
if ((int)atomic_cas_uint(acceptp, -1, sock) != -1) {
regress/sys/netinet/tcpthread/tcpthread.c
269
if ((int)atomic_cas_uint(sockp, sock, -1) != sock) {
regress/sys/netinet/tcpthread/tcpthread.c
307
if ((int)atomic_cas_uint(sockp, sock, -1) != sock) {
regress/sys/netinet/tcpthread/tcpthread.c
458
if ((int)atomic_cas_uint(sockp, sock, -1) != sock) {
regress/sys/netinet/udpthread/udpthread.c
148
if ((int)atomic_cas_uint(connectp, -1, sock) != -1) {
regress/sys/netinet/udpthread/udpthread.c
263
if ((int)atomic_cas_uint(acceptp, -1, sock) != -1) {
regress/sys/netinet/udpthread/udpthread.c
320
if ((int)atomic_cas_uint(sockp, sock, -1) != sock) {
regress/sys/netinet/udpthread/udpthread.c
358
if ((int)atomic_cas_uint(sockp, sock, -1) != sock) {
regress/sys/netinet/udpthread/udpthread.c
508
if ((int)atomic_cas_uint(sockp, sock, -1) != sock) {
sys/arch/alpha/alpha/pmap.c
3419
if (atomic_cas_uint(&pj->pj_state,
sys/arch/amd64/amd64/acpi_machdep.c
307
} while (atomic_cas_uint(lock, old, new) != old);
sys/arch/amd64/amd64/acpi_machdep.c
325
} while (atomic_cas_uint(lock, old, new) != old);
sys/arch/amd64/amd64/pmap.c
3216
while (atomic_cas_uint(&tlb_shoot_lock.lock, 0, 1) != 0) {
sys/arch/amd64/amd64/tsc.c
476
while (atomic_cas_uint(&tsc_egress_barrier, 2, 0) != 2)
sys/arch/amd64/amd64/vmm_machdep.c
3370
if (atomic_cas_uint(&vcpu->vc_state, old, next) != old) {
sys/arch/i386/i386/acpi_machdep.c
249
} while (atomic_cas_uint(lock, old, new) != old);
sys/arch/i386/i386/acpi_machdep.c
267
} while (atomic_cas_uint(lock, old, new) != old);
sys/arch/i386/i386/pmap.c
2659
while (atomic_cas_uint(&tlb_shoot_lock.lock, 0, 1) != 0) {
sys/arch/m88k/include/atomic.h
138
((unsigned long)atomic_cas_uint((unsigned int *)p, (unsigned int)o, \
sys/arch/m88k/include/atomic.h
141
((void *)atomic_cas_uint((void *)p, (unsigned int)o, (unsigned int)n))
sys/arch/mips64/mips64/pmap.c
236
return atomic_cas_uint((unsigned int *)pte, o, n);
sys/arch/powerpc/powerpc/pmap.c
1119
if (atomic_cas_uint(&usedsr[tblidx], bits,
sys/arch/powerpc64/powerpc64/pmap.c
397
if (atomic_cas_uint(&pmap_vsid[vsid / 32], bits,
sys/arch/powerpc64/powerpc64/pmap.c
414
if (atomic_cas_uint(&pmap_vsid[vsid / 32], bits,
sys/arch/sparc64/include/atomic.h
108
_def_atomic_op_nv(_atomic_add_int_nv, unsigned int, atomic_cas_uint, +)
sys/arch/sparc64/include/atomic.h
110
_def_atomic_op_nv(_atomic_sub_int_nv, unsigned int, atomic_cas_uint, -)
sys/arch/sparc64/include/atomic.h
137
r = atomic_cas_uint(uip, e, e | v);
sys/arch/sparc64/include/atomic.h
149
r = atomic_cas_uint(uip, e, e & ~v);
sys/arch/sparc64/include/atomic.h
70
_def_atomic_swap(_atomic_swap_uint, unsigned int, atomic_cas_uint)
sys/dev/pci/if_icevar.h
4095
previous = atomic_cas_uint(s, expected, expected | (1UL << bit));
sys/dev/pci/if_icevar.h
4113
previous = atomic_cas_uint(s, expected, expected & ~(1UL << bit));
sys/dev/pv/xen.c
1201
while (atomic_cas_uint(ptr, flags, GTF_invalid) != flags) {
sys/dev/vmm/vmm.c
727
if (atomic_cas_uint(&vm->vm_dying, 0, 1) == 1) {
sys/kern/kern_descrip.c
216
if (atomic_cas_uint(&nfp->f_count, count, count + 1) == count)
sys/kern/kern_descrip.c
494
} while (atomic_cas_uint(&fp->f_flag, prev, tmp) != prev);
sys/kern/kern_lock.c
715
ogen = atomic_cas_uint(&pcl->pcl_gen, gen, ngen);
sys/kern/kern_sig.c
1341
if (atomic_cas_uint(&pr->ps_siglist, ps_siglist,
sys/kern/kern_sysctl.c
1156
} while (atomic_cas_uint(valp, oldval, newval) != oldval);
sys/kern/kern_sysctl.c
1223
} while (atomic_cas_uint(&securelevel, oldval, newval) !=
sys/kern/kern_sysctl.c
688
oldval == atomic_cas_uint(&pool_debug, oldval, newval))
sys/kern/subr_kubsan.c
523
} while (atomic_cas_uint(line, prev, prev | LOCATION_REPORTED) != prev);
sys/kern/subr_kubsan.c
744
if (atomic_cas_uint(&kubsan_slot, nslots, 0) != nslots) {
sys/net/if_loop.c
269
atomic_cas_uint(&rt->rt_mtu, 0, LOMTU);
sys/net/pf.c
4765
atomic_cas_uint(&r->rule_flag, rule_flag,
sys/net/route.c
564
atomic_cas_uint(&rt->rt_mtu, mtu, nhmtu);
sys/netinet/ip_icmp.c
1080
atomic_cas_uint(&rt->rt_mtu, rtmtu, mtu);
sys/netinet/ip_input.c
1751
oldval == atomic_cas_uint(&ip_mtudisc, oldval, newval) &&
sys/netinet/ip_output.c
474
atomic_cas_uint(&ro->ro_rt->rt_mtu, rtmtu,
sys/netinet6/icmp6.c
1006
atomic_cas_uint(&rt->rt_mtu, rtmtu, mtu);
sys/netinet6/ip6_output.c
1050
atomic_cas_uint(&rt->rt_mtu, rtmtu, mtu);
sys/sys/atomic.h
33
#ifndef atomic_cas_uint