Symbol: atomic_cmpxchg
arch/mips/net/bpf_jit_comp32.c
892
emit_mov_i(ctx, MIPS_R_T9, (u32)&atomic_cmpxchg);
arch/sh/kernel/ftrace.c
110
old = atomic_cmpxchg(&nmi_running, old, new);
arch/sh/kernel/ftrace.c
149
if (!atomic_cmpxchg(&nmi_running, 0, MOD_CODE_WRITE_FLAG))
arch/sh/kernel/ftrace.c
154
} while (atomic_cmpxchg(&nmi_running, 0, MOD_CODE_WRITE_FLAG));
drivers/accel/habanalabs/common/command_buffer.c
316
rc = atomic_cmpxchg(&cb->is_handle_destroyed, 0, 1);
drivers/accel/habanalabs/common/command_submission.c
1698
do_ctx_switch = atomic_cmpxchg(&ctx->thread_ctx_switch_token, 1, 0);
drivers/accel/habanalabs/common/command_submission.c
849
rc = atomic_cmpxchg(&hdev->captured_err_info.cs_timeout.write_enable, 1, 0);
drivers/accel/habanalabs/common/device.c
2695
if (atomic_cmpxchg(&hdev->captured_err_info.razwi_info.razwi_detected, 0, 1))
drivers/accel/habanalabs/common/device.c
2784
if (atomic_cmpxchg(&pgf_info->page_fault_detected, 0, 1))
drivers/accel/habanalabs/common/device.c
2809
if (atomic_cmpxchg(&info->event_detected, 0, 1))
drivers/accel/habanalabs/common/device.c
2831
if (atomic_cmpxchg(&info->event_detected, 0, 1))
drivers/accel/habanalabs/common/device.c
2855
if (atomic_cmpxchg(&info->event_detected, 0, 1))
drivers/accel/habanalabs/common/irq.c
278
if (atomic_cmpxchg(&free_node->in_use, 0, 1)) {
drivers/accel/habanalabs/common/memory_mgr.c
271
if (atomic_cmpxchg(&buf->mmap, 0, 1)) {
drivers/accel/ivpu/ivpu_pm.c
189
if (atomic_cmpxchg(&vdev->pm->reset_pending, 0, 1) == 0) {
drivers/block/drbd/drbd_actlog.c
87
(r = atomic_cmpxchg(&device->md_io.in_use, 0, 1)) == 0 ||
drivers/char/hw_random/ba431-rng.c
109
if (atomic_cmpxchg(&ba431->reset_pending, 0, 1))
drivers/char/hw_random/cctrng.c
248
if (atomic_cmpxchg(&drvdata->pending_hw, 0, 1) == 0) {
drivers/char/hw_random/cctrng.c
378
if (atomic_cmpxchg(&drvdata->pending_hw, 0, 1) == 0) {
drivers/char/ipmi/ipmi_watchdog.c
599
} else if (atomic_cmpxchg(&pretimeout_since_last_heartbeat, 1, 0)) {
drivers/cxl/core/region.c
2670
if (atomic_cmpxchg(&cxlrd->region_id, id, rc) != id) {
drivers/firewire/core-device.c
1120
if (atomic_cmpxchg(&reused->state,
drivers/firewire/core-device.c
1195
if (atomic_cmpxchg(&device->state,
drivers/firewire/core-device.c
1269
if (atomic_cmpxchg(&device->state,
drivers/firewire/core-device.c
1295
if (atomic_cmpxchg(&device->state,
drivers/firewire/core-device.c
1392
if (atomic_cmpxchg(&device->state,
drivers/firewire/core-device.c
940
if (atomic_cmpxchg(&device->state,
drivers/firmware/arm_scmi/bus.c
442
atomic_cmpxchg(&scmi_syspower_registered, 0, 1)) {
drivers/firmware/arm_scmi/driver.c
982
ret = atomic_cmpxchg(&xfer->busy, SCMI_XFER_FREE, SCMI_XFER_BUSY);
drivers/firmware/arm_scmi/transports/smc.c
109
ret = atomic_cmpxchg(inflight, INFLIGHT_NONE, xfer->hdr.seq);
drivers/gpio/gpiolib-cdev.c
2334
int abiv = atomic_cmpxchg(&cdata->watch_abi_version, 0, version);
drivers/gpu/drm/amd/amdgpu/amdgpu_aca.c
402
ret = atomic_cmpxchg(&aca->ue_update_flag, 0, 1) == 0;
drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c
230
} while (atomic_cmpxchg(&drv->last_seq, last_seq, seq) != last_seq);
drivers/gpu/drm/amd/amdgpu/amdgpu_mca.c
343
ret = atomic_cmpxchg(&mca->ue_update_flag, 0, 1) == 0;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
3644
if (amdgpu_ras_is_rma(adev) && atomic_cmpxchg(&ras->rma_in_recovery, 0, 1) == 0)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
4859
if (atomic_cmpxchg(&amdgpu_ras_in_intr, 0, 1) == 0) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5088
if (atomic_cmpxchg(&ras->in_recovery, 0, 1) == 0) {
drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c
861
if (!atomic_cmpxchg(&adev->uvd.handles[i], 0, handle)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c
894
atomic_cmpxchg(&adev->uvd.handles[i], handle, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c
760
if (!atomic_cmpxchg(&p->adev->vce.handles[i], 0, handle)) {
drivers/gpu/drm/amd/amdkfd/kfd_int_process_v9.c
159
old_poison = atomic_cmpxchg(&p->poison, 0, 1);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1943
if (atomic_cmpxchg(&prange->invalid, invalid, 0) != invalid)
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1947
if (atomic_cmpxchg(&svms->evicted_ranges, evicted_ranges, 0) !=
drivers/gpu/drm/i915/gt/intel_context.h
197
if (atomic_cmpxchg(&ce->pin_count, 1, 2) == 1) {
drivers/gpu/drm/msm/adreno/a5xx_preempt.c
15
enum preempt_state cur = atomic_cmpxchg(&a5xx_gpu->preempt_state,
drivers/gpu/drm/msm/adreno/a6xx_preempt.c
19
enum a6xx_preempt_state cur = atomic_cmpxchg(&a6xx_gpu->preempt_state,
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.c
147
if (atomic_cmpxchg(&cgrp->rc, NVKM_CGRP_RC_NONE, NVKM_CGRP_RC_PENDING) != NVKM_CGRP_RC_NONE)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.c
76
state = atomic_cmpxchg(&cgrp->rc, NVKM_CGRP_RC_PENDING, NVKM_CGRP_RC_RUNNING);
drivers/gpu/drm/omapdrm/dss/dsi.c
4283
old = atomic_cmpxchg(&dsi->do_ext_te_update, 1, 0);
drivers/gpu/drm/omapdrm/dss/dsi.c
4298
old = atomic_cmpxchg(&dsi->do_ext_te_update, 1, 0);
drivers/gpu/drm/panthor/panthor_device.h
288
if (!atomic_cmpxchg(&ptdev->reset.pending, 0, 1) &&
drivers/gpu/drm/panthor/panthor_device.h
325
if (ret && atomic_cmpxchg(&ptdev->pm.recovery_needed, 1, 0) == 1)
drivers/gpu/drm/radeon/radeon_uvd.c
513
if (!atomic_cmpxchg(&p->rdev->uvd.handles[i], 0, handle)) {
drivers/gpu/drm/radeon/radeon_uvd.c
549
atomic_cmpxchg(&p->rdev->uvd.handles[i], handle, 0);
drivers/gpu/drm/radeon/radeon_vce.c
539
if (!atomic_cmpxchg(&p->rdev->vce.handles[i], 0, handle)) {
drivers/gpu/drm/radeon/radeon_vce.c
680
atomic_cmpxchg(&p->rdev->vce.handles[i], handle, 0);
drivers/gpu/drm/vmwgfx/vmwgfx_mksstat.h
116
const pid_t pid = atomic_cmpxchg(&dev_priv->mksstat_kern_pids[_##kern_cntr.slot], current->pid, MKSSTAT_PID_RESERVED); \
drivers/gpu/drm/vmwgfx/vmwgfx_msg.c
1108
pid = atomic_cmpxchg(&dev_priv->mksstat_user_pids[slot], pgid, MKSSTAT_PID_RESERVED);
drivers/gpu/drm/vmwgfx/vmwgfx_msg.c
771
if (!atomic_cmpxchg(&dev_priv->mksstat_kern_pids[slot], 0, MKSSTAT_PID_RESERVED)) {
drivers/gpu/drm/vmwgfx/vmwgfx_msg.c
846
const pid_t pid1 = atomic_cmpxchg(&dev_priv->mksstat_user_pids[i], pid0, MKSSTAT_PID_RESERVED);
drivers/gpu/drm/vmwgfx/vmwgfx_msg.c
876
const pid_t pid1 = atomic_cmpxchg(&dev_priv->mksstat_kern_pids[i], pid0, MKSSTAT_PID_RESERVED);
drivers/gpu/drm/vmwgfx/vmwgfx_msg.c
975
if (!atomic_cmpxchg(&dev_priv->mksstat_user_pids[slot], 0, MKSSTAT_PID_RESERVED))
drivers/gpu/drm/vmwgfx/vmwgfx_vkms.c
552
ret = atomic_cmpxchg(&du->vkms.atomic_lock,
drivers/gpu/drm/vmwgfx/vmwgfx_vkms.c
586
ret = atomic_cmpxchg(&du->vkms.atomic_lock,
drivers/gpu/drm/vmwgfx/vmwgfx_vkms.c
619
ret = atomic_cmpxchg(&du->vkms.atomic_lock,
drivers/gpu/host1x/hw/syncpt_hw.c
62
} while ((u32)atomic_cmpxchg(&sp->min_val, old, live) != old);
drivers/hwtracing/intel_th/msu.c
1341
if (WARN_ON_ONCE(atomic_cmpxchg(&msc->user_count, -1, 0) != -1))
drivers/hwtracing/intel_th/msu.c
1363
count = atomic_cmpxchg(&msc->user_count, 0, -1);
drivers/infiniband/core/rdma_core.c
81
return atomic_cmpxchg(&uobj->usecnt, 0, -1) == 0 ? 0 : -EBUSY;
drivers/infiniband/hw/ionic/ionic_admin.c
731
old_state = atomic_cmpxchg(&dev->admin_state, IONIC_ADMIN_ACTIVE,
drivers/infiniband/hw/irdma/utils.c
2408
if (atomic_cmpxchg(&cq->armed, 1, 0))
drivers/infiniband/hw/irdma/verbs.c
4516
if (!atomic_cmpxchg(&iwcq->armed, 0, 1) || promo_event) {
drivers/infiniband/hw/ocrdma/ocrdma_ah.c
175
if (atomic_cmpxchg(&dev->update_sl, 1, 0))
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
2502
if (atomic_cmpxchg(&dev->update_sl, 1, 0))
drivers/irqchip/irq-mvebu-icu.c
118
if (atomic_cmpxchg(&msi_data->initialized, false, true))
drivers/mailbox/mtk-gpueb-mailbox.c
100
if (!atomic_cmpxchg(&ch->rx_status, 0, GPUEB_MBOX_FULL | GPUEB_MBOX_BLOCKED))
drivers/mailbox/mtk-gpueb-mailbox.c
112
status = atomic_cmpxchg(&ch->rx_status, GPUEB_MBOX_FULL | GPUEB_MBOX_BLOCKED,
drivers/md/bcache/alloc.c
97
} while (atomic_cmpxchg(&c->rescale, r, r + next) != r);
drivers/md/bcache/io.c
104
count = atomic_cmpxchg(&ca->io_count, old, new);
drivers/md/bcache/io.c
113
errors = atomic_cmpxchg(&ca->io_errors,
drivers/md/dm-vdo/data-vio.c
353
actual = atomic_cmpxchg(&data_vio->compression.status, expected, replacement);
drivers/md/dm-vdo/data-vio.c
657
if (atomic_cmpxchg(&pool->processing, false, true))
drivers/md/dm-vdo/dedupe.c
329
return (atomic_cmpxchg(&context->state, old, new) == old);
drivers/md/dm-vdo/dedupe.c
334
return (atomic_cmpxchg(&zone->timer_state, old, new) == old);
drivers/md/dm-vdo/dm-vdo-target.c
1074
if (atomic_cmpxchg(&vdo->processing_message, 0, 1) != 0)
drivers/md/dm-vdo/dm-vdo-target.c
1216
if (atomic_cmpxchg(&admin->busy, 0, 1) != 0) {
drivers/md/dm-vdo/funnel-workqueue.c
146
if ((atomic_read(&queue->idle) != 1) || (atomic_cmpxchg(&queue->idle, 1, 0) != 1))
drivers/md/dm-vdo/recovery-journal.c
1659
prior_state = atomic_cmpxchg(&counter->state, LOCK_COUNTER_STATE_SUSPENDED,
drivers/md/dm-vdo/recovery-journal.c
177
prior_state = atomic_cmpxchg(&journal->lock_counter.state,
drivers/md/dm-vdo/recovery-journal.c
312
prior_state = atomic_cmpxchg(&counter->state, LOCK_COUNTER_STATE_NOT_NOTIFYING,
drivers/md/dm-vdo/slab-depot.c
2770
atomic_cmpxchg(&allocator->depot->vdo->state, VDO_RECOVERING,
drivers/media/pci/saa7164/saa7164-encoder.c
742
if (atomic_cmpxchg(&fh->v4l_reading, 1, 0) == 1) {
drivers/media/pci/saa7164/saa7164-encoder.c
810
if (atomic_cmpxchg(&fh->v4l_reading, 0, 1) == 0) {
drivers/media/pci/saa7164/saa7164-encoder.c
910
if (atomic_cmpxchg(&fh->v4l_reading, 0, 1) == 0) {
drivers/media/pci/saa7164/saa7164-vbi.c
445
if (atomic_cmpxchg(&fh->v4l_reading, 1, 0) == 1) {
drivers/media/pci/saa7164/saa7164-vbi.c
512
if (atomic_cmpxchg(&fh->v4l_reading, 0, 1) == 0) {
drivers/media/pci/saa7164/saa7164-vbi.c
613
if (atomic_cmpxchg(&fh->v4l_reading, 0, 1) == 0) {
drivers/media/platform/ti/omap3isp/isp.c
1248
if (atomic_cmpxchg(stopping, 1, 0)) {
drivers/media/rc/redrat3.c
856
if (brightness != LED_OFF && atomic_cmpxchg(&rr3->flash, 0, 1) == 0) {
drivers/net/bonding/bond_3ad.c
2498
if (atomic_cmpxchg(&BOND_AD_INFO(bond).agg_select_timer,
drivers/net/ethernet/aquantia/atlantic/aq_utils.h
21
} while (atomic_cmpxchg(flags, flags_old, flags_new) != flags_old);
drivers/net/ethernet/aquantia/atlantic/aq_utils.h
31
} while (atomic_cmpxchg(flags, flags_old, flags_new) != flags_old);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
4173
old = atomic_cmpxchg((v), c, c + a);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
4201
old = atomic_cmpxchg((v), c, c - a);
drivers/net/ethernet/cavium/liquidio/lio_main.c
4081
atomic_cmpxchg(octeon_dev->adapter_fw_state,
drivers/net/ethernet/cavium/liquidio/lio_main.c
4086
fw_state = atomic_cmpxchg(octeon_dev->adapter_fw_state,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
163
if (atomic_cmpxchg(&tx->dqo_compl.free_pending_packets,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
90
if (atomic_cmpxchg(&tx->dqo_compl.free_tx_qpl_buf_head,
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
27
state = atomic_cmpxchg(&pf->ipsec.cpt_state,
drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c
2689
if (atomic_cmpxchg(&rep->rep_data[rep_type].state,
drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c
2699
if (atomic_cmpxchg(&rep->rep_data[rep_type].state,
drivers/net/ethernet/sfc/falcon/farch.c
1099
if (atomic_cmpxchg(&tx_queue->flush_outstanding, 1, 0)) {
drivers/net/ethernet/sfc/falcon/farch.c
633
} else if (atomic_cmpxchg(&tx_queue->flush_outstanding,
drivers/net/ethernet/sfc/siena/farch.c
1093
if (atomic_cmpxchg(&tx_queue->flush_outstanding, 1, 0))
drivers/net/ethernet/sfc/siena/farch.c
621
} else if (atomic_cmpxchg(&tx_queue->flush_outstanding,
drivers/net/ipa/ipa_modem.c
306
state = atomic_cmpxchg(&ipa->modem_state, IPA_MODEM_STATE_STOPPED,
drivers/net/ipa/ipa_modem.c
357
state = atomic_cmpxchg(&ipa->modem_state, IPA_MODEM_STATE_RUNNING,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c
345
if (atomic_cmpxchg(&array[*idx].allocated, 0, 1) == 0)
drivers/net/wireless/mediatek/mt76/tx.c
237
atomic_cmpxchg(&wcid->non_aql_packets, pending, 0);
drivers/net/wireless/realtek/rtw89/chan.c
270
cur = atomic_cmpxchg(&hal->roc_chanctx_idx,
drivers/net/wireless/realtek/rtw89/chan.c
281
cur = atomic_cmpxchg(&hal->roc_chanctx_idx, idx,
drivers/net/wireless/realtek/rtw89/core.c
5656
cur = atomic_cmpxchg(&wait->cond, RTW89_WAIT_COND_IDLE, cond);
drivers/net/wireless/realtek/rtw89/core.c
5733
cur = atomic_cmpxchg(&wait->cond, cond, RTW89_WAIT_COND_IDLE);
drivers/platform/chrome/wilco_ec/event.c
312
if (atomic_cmpxchg(&dev_data->available, 1, 0) == 0)
drivers/platform/chrome/wilco_ec/telemetry.c
246
if (atomic_cmpxchg(&dev_data->available, 1, 0) == 0)
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
678
} while (atomic_cmpxchg(&service->poll_flags, value,
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
684
} while (atomic_cmpxchg(&state->poll_services[index],
drivers/rapidio/rio.c
2125
if (atomic_cmpxchg(&port->state,
drivers/s390/block/dasd.c
2147
if (atomic_cmpxchg (&device->tasklet_scheduled, 0, 1) != 0)
drivers/s390/block/dasd.c
3001
if (atomic_cmpxchg(&block->tasklet_scheduled, 0, 1) != 0)
drivers/s390/cio/ccwgroup.c
177
if (atomic_cmpxchg(&gdev->onoff, 0, 1) != 0)
drivers/s390/cio/ccwgroup.c
63
if (atomic_cmpxchg(&gdev->onoff, 0, 1) != 0)
drivers/s390/cio/ccwgroup.c
93
if (atomic_cmpxchg(&gdev->onoff, 0, 1) != 0)
drivers/s390/cio/device.c
1334
if (atomic_cmpxchg(&cdev->private->onoff, 0, 1) != 0)
drivers/s390/cio/device.c
497
if (atomic_cmpxchg(&cdev->private->onoff, 0, 1) != 0)
drivers/s390/cio/qdio_thinint.c
53
if (!atomic_cmpxchg(&q_indicators[i].count, 0, 1))
drivers/s390/net/qeth_core_main.c
1801
switch (atomic_cmpxchg(&card->qdio.state, QETH_QDIO_ESTABLISHED,
drivers/s390/net/qeth_core_main.c
2638
if (atomic_cmpxchg(&card->qdio.state, QETH_QDIO_UNINITIALIZED,
drivers/s390/net/qeth_core_main.c
5116
if (atomic_cmpxchg(&card->qdio.state, QETH_QDIO_ALLOCATED,
drivers/scsi/lpfc/lpfc_init.c
15742
if (atomic_cmpxchg(&phba->dbg_log_dmping, 0, 1) != 0)
drivers/scsi/qla2xxx/qla_inline.h
122
if (old_val == atomic_cmpxchg(&fcport->shadow_disc_state,
drivers/scsi/qla2xxx/qla_inline.h
475
old = atomic_cmpxchg((v), c, dec);
drivers/scsi/scsi_lib.c
634
if (old && atomic_cmpxchg(&sdev->restarts, old, 0) == old)
drivers/target/target_core_transport.c
3194
if (!atomic_cmpxchg(&cmd_cnt->stopped, 0, 1))
drivers/vhost/vdpa.c
1414
opened = atomic_cmpxchg(&v->opened, 0, 1);
drivers/vhost/vdpa.c
1634
opened = atomic_cmpxchg(&v->opened, 0, 1);
drivers/video/fbdev/omap2/omapfb/displays/panel-dsi-cm.c
824
old = atomic_cmpxchg(&ddata->do_update, 1, 0);
fs/afs/validation.c
359
atomic_cmpxchg(&volume->cb_v_check, cb_v_check, op->cb_v_break);
fs/dlm/midcomms.c
416
} while (atomic_cmpxchg(&node->ulp_delivered, oval, nval) != oval);
fs/dlm/midcomms.c
532
} while (atomic_cmpxchg(&node->seq_next, oval, nval) != oval);
fs/erofs/data.c
274
} while (atomic_cmpxchg((atomic_t *)&folio->private, orig, v) != orig);
fs/hfsplus/xattr.c
219
if (old_state != atomic_cmpxchg(&sbi->attr_tree_state,
fs/mbcache.c
267
if (atomic_cmpxchg(&entry->e_refcnt, 2, 0) != 2)
fs/mbcache.c
314
atomic_cmpxchg(&entry->e_refcnt, 1, 0) != 1) {
fs/ubifs/journal.c
371
if (atomic_cmpxchg(&c->need_wait_space, 0, 1) == 0) {
include/linux/page_ref.h
274
int ret = likely(atomic_cmpxchg(&page->_refcount, count, 0) == count);
include/linux/sbitmap.h
542
atomic_cmpxchg(index, old, new);
io_uring/poll.c
629
if (atomic_cmpxchg(&req->poll_refs, 1, 0) != 1)
kernel/bpf/bpf_local_storage.c
511
atomic_cmpxchg(&selem->state, SELEM_UNLINKED, SELEM_TOFREE) == SELEM_UNLINKED)
kernel/bpf/core.c
2255
BPF_R0 = (u32) atomic_cmpxchg(
kernel/irq_work.c
228
(void)atomic_cmpxchg(&work->node.a_flags, flags, flags & ~IRQ_WORK_BUSY);
kernel/jump_label.c
167
if (!atomic_cmpxchg(&key->enabled, 0, -1)) {
kernel/jump_label.c
239
if (atomic_cmpxchg(&key->enabled, 1, 0) == 1)
kernel/kprobes.c
701
if (atomic_cmpxchg(&optimizer_state,
kernel/liveupdate/luo_core.c
346
if (atomic_cmpxchg(&ldev->in_use, 0, 1))
kernel/sched/psi.c
746
atomic_cmpxchg(&group->rtpoll_wakeup, 1, 0) ||
kernel/trace/trace_recursion_record.c
114
atomic_cmpxchg(&nr_records, i, index + 1);
lib/closure.c
213
} while ((v = atomic_cmpxchg(&cl->remaining, old, new)) != old);
lib/crypto/tests/hash-test-template.h
382
if (atomic_cmpxchg(&ctx->in_use, 0, 1) == 0)
mm/huge_memory.c
289
if (atomic_cmpxchg(&huge_zero_refcount, 1, 0) == 1) {
mm/kasan/kasan_test_c.c
788
KUNIT_EXPECT_KASAN_FAIL(test, atomic_cmpxchg(unsafe, 21, 42));
mm/page_reporting.c
344
state = atomic_cmpxchg(&prdev->state, state, PAGE_REPORTING_IDLE);
mm/rmap.c
822
atomic_cmpxchg(&mm->tlb_flush_batched, batch,
net/mac80211/sta_info.c
2467
atomic_cmpxchg(&sta->airtime[ac].aql_tx_pending,
net/mac80211/sta_info.c
2478
atomic_cmpxchg(&local->aql_ac_pending_airtime[ac],
net/netfilter/nft_numgen.c
31
} while (atomic_cmpxchg(priv->counter, oval, nval) != oval);
net/netfilter/xt_statistic.c
44
} while (atomic_cmpxchg(&info->master->count, oval, nval) != oval);
net/rds/ib_send.c
401
if (atomic_cmpxchg(&ic->i_credits, oldval, newval) != oldval)
net/rds/rds.h
831
return atomic_cmpxchg(&cp->cp_state, old, new) == old;
net/tipc/crypto.c
1416
!atomic_cmpxchg(&rx->key_distr, 0, KEY_DISTR_SCHED)) {
net/tipc/crypto.c
1433
atomic_cmpxchg(&rx->peer_rx_active, cur, new) == cur) {
net/tipc/crypto.c
2366
if (atomic_cmpxchg(&rx->key_distr,
net/tipc/crypto.c
2380
atomic_cmpxchg(&rx->key_distr, KEY_DISTR_COMPL, 0);
net/tipc/crypto.c
479
} while (atomic_cmpxchg(&tmp->users, cur, val) != cur);
rust/helpers/atomic.c
401
return atomic_cmpxchg(v, old, new);
security/integrity/ima/ima_policy.c
846
atomic_cmpxchg(&ima_setxattr_allowed_hash_algorithms,
tools/include/linux/atomic.h
11
#define atomic_cmpxchg_relaxed atomic_cmpxchg
tools/include/linux/atomic.h
12
#define atomic_cmpxchg_release atomic_cmpxchg
tools/include/linux/atomic.h
19
ret = atomic_cmpxchg(ptr, old, new);