atomic_cas_ptr
if (atomic_cas_ptr(evcpp, NULL, evc) != NULL)
char *old = atomic_cas_ptr(&ktest_cached_module_dir, NULL, path);
} while (atomic_cas_ptr(headp, head, sdc) != head);
} while (atomic_cas_ptr(headp, tail, head) != tail);
(void) atomic_cas_ptr(&sysent[sysnum].sy_callc,
(void) atomic_cas_ptr(&sysent32[sysnum].sy_callc,
(void) atomic_cas_ptr(&sysent[sysnum].sy_callc,
(void) atomic_cas_ptr(&sysent32[sysnum].sy_callc,
if (atomic_cas_ptr(hp, NULL, head) != NULL) {
} while (atomic_cas_ptr(&v->v_op, r, newops) != r);
} while (atomic_cas_ptr(&v->vfs_op, r, newops) != r);
if (atomic_cas_ptr(&vp->v_fopdata, NULL, npvp) != NULL) {
if (atomic_cas_ptr(&fop_femop, NULL, femp) != NULL) {
if (atomic_cas_ptr(&fop_fsemop, NULL, fsemp) != NULL) {
return (atomic_cas_ptr(&inst->sofi_cookie, old, new));
atomic_cas_ptr(&vfsp->vfs_op, op, vfsops) == op) {
if (vp->v_femhead != NULL || atomic_cas_ptr(&vp->v_op, op, vnodeops) !=
oldq = atomic_cas_ptr((void **)pfkey_qp, NULL, OTHERQ(q));
(void) atomic_cas_ptr(&connp->conn_sqp, sqp,
(void) atomic_cas_ptr(&oce_dev_list[dev_index], NULL, dev);
if (atomic_cas_ptr(&ksyms_clones[clone].ksyms_base, 0, addr) ==
if (atomic_cas_ptr(&smb_clones[c].c_hdl, NULL, ksmbios) == NULL)
if (atomic_cas_ptr(&hp->tail, bp, nbp) == bp) {
atomic_cas_ptr(&cyc_coverage[ndx].cyv_why, NULL, why) !=
if (atomic_cas_ptr(&eqp->eq_pend, old, eep) == old)
if (atomic_cas_ptr(&eqp->eq_pend, eep, NULL) == eep)
if (atomic_cas_ptr(&eqp->eq_pend, old, eqep) == old)
(void) atomic_cas_ptr((void *)&fm_panicstr, NULL, (void *)format);
if (atomic_cas_ptr(&panic_mutex_addr, NULL, lp) == NULL)
if (atomic_cas_ptr(&panic_rwlock_addr, NULL, lp) == NULL)
extern void *atomic_cas_ptr(volatile void *, void *, void *);
osp = atomic_cas_ptr(&ao_shared[chipid], NULL, sp);
osp = atomic_cas_ptr(&authamd_shared[procnodeid], NULL, sp);
osp = atomic_cas_ptr(&gcpu_shared[chipid], NULL, sp);
cp1 = atomic_cas_ptr(&acpidev_dr_capacity_curr, NULL, cp);
if (atomic_cas_ptr(&cmi_chip_tab[chipid], NULL, pg) != NULL)
} while (atomic_cas_ptr(queue, old_head, msg) != old_head);
} while (atomic_cas_ptr(queue, old_head, old_head->xc_next) !=
(void) atomic_cas_ptr(&cpup->cpu_m.xc_msgbox,
atomic_cas_ptr(volatile void *target, void *cmp, void *new)
tmp = atomic_cas_ptr(&p->p_utraps[idx], ov, nv);
tmp = atomic_cas_ptr(&p->p_utraps, NULL, sv);
tmp = atomic_cas_ptr(&p->p_utraps[idx], ov, nvp);
tmp = atomic_cas_ptr(&p->p_utraps, NULL, sv);
prcp = atomic_cas_ptr((void *)&prom_cpu, NULL, cp);
prcp = atomic_cas_ptr((void *)&prom_cpu, NULL, cp);
if (atomic_cas_ptr(&leaky_bucket_timeout_id,
if (atomic_cas_ptr(&ppmap_vaddrs[index],
if (atomic_cas_ptr(myslot, NULL, va) == NULL)
if (atomic_cas_ptr(&ppmap_vaddrs[nset], va, NULL) ==