Symbol: cmpxchg
arch/arc/kernel/smp.c
267
} while (cmpxchg(ipi_data_ptr, old, new) != old);
arch/arm64/kvm/hyp/pgtable.c
811
return cmpxchg(ctx->ptep, ctx->old, new) == ctx->old;
arch/mips/include/asm/pgtable.h
184
cmpxchg(&buddy->pte, 0, _PAGE_GLOBAL);
arch/mips/kvm/vz.c
2936
if (cmpxchg(&kvm_vz_guest_vtlb_size, 0, guest_mmu_size) &&
arch/powerpc/kernel/fadump.c
702
old_cpu = cmpxchg(&crashing_cpu, -1, this_cpu);
arch/powerpc/kvm/book3s_hv_builtin.c
262
} while (cmpxchg(&vc->entry_exit_map, ee, ee | me) != ee);
arch/powerpc/kvm/book3s_hv_builtin.c
292
} while (cmpxchg(&vc->entry_exit_map, ee,
arch/powerpc/kvm/book3s_hv_rm_xics.c
704
} while (cmpxchg(&state->pq_state, pq_old, pq_new) != pq_old);
arch/powerpc/kvm/book3s_hv_rm_xics.c
874
} while (cmpxchg(&state->pq_state, pq_old, pq_new) != pq_old);
arch/powerpc/kvm/book3s_xics.c
112
} while (cmpxchg(&state->pq_state, pq_old, pq_new) != pq_old);
arch/powerpc/kvm/book3s_xics.c
807
} while (cmpxchg(&state->pq_state, pq_old, pq_new) != pq_old);
arch/powerpc/kvm/booke.c
648
} while (cmpxchg(&vcpu->arch.tsr, tsr, new_tsr) != tsr);
arch/powerpc/platforms/powernv/idle.c
270
tmp = cmpxchg(state, s, new);
arch/powerpc/platforms/powernv/pci-ioda-tce.c
106
oldtce = be64_to_cpu(cmpxchg(&tmp[n], 0,
arch/powerpc/platforms/powernv/rng.c
157
if (cmpxchg(&ppc_md.get_random_seed, pnv_get_random_long_early,
arch/powerpc/xmon/xmon.c
419
last_speaker = cmpxchg(&xmon_speaker, 0, me);
arch/powerpc/xmon/xmon.c
435
prev = cmpxchg(&xmon_speaker, last_speaker, me);
arch/s390/include/asm/percpu.h
133
ret__ = cmpxchg(ptr__, oval, nval); \
arch/s390/include/asm/percpu.h
30
prev__ = cmpxchg(ptr__, old__, new__); \
arch/s390/kvm/interrupt.c
74
rc = cmpxchg(&sigp_ctrl->value, old_val.value, new_val.value);
arch/sh/kernel/cpu/sh2/smp-j2.c
118
while (cmpxchg(pmsg, old, old|(1U<<message)) != old);
arch/sh/kernel/cpu/sh2/smp-j2.c
28
while (cmpxchg(pmsg, messages, 0) != messages);
arch/um/kernel/skas/uaccess.c
355
ret = cmpxchg(uaddr, oldval, newval);
arch/x86/events/amd/core.c
509
old = cmpxchg(nb->owners + idx, NULL, event);
arch/x86/events/amd/core.c
521
cmpxchg(nb->owners + new, event, NULL);
arch/x86/kvm/x86.c
12857
cmpxchg(per_cpu_ptr(&last_vcpu, cpu), vcpu, NULL);
block/blk-crypto-fallback.c
164
cmpxchg(&src_bio->bi_status, 0, enc_bio->bi_status);
block/blk-mq.c
3469
cmpxchg(&drv_tags->rqs[i], rq, NULL);
block/blk-mq.c
3939
cmpxchg(&tags->rqs[i], flush_rq, NULL);
drivers/accessibility/speakup/selection.c
124
if (cmpxchg(&speakup_paste_work.tty, NULL, tty)) {
drivers/accessibility/speakup/selection.c
71
if (cmpxchg(&speakup_sel_work.tty, NULL, tty)) {
drivers/block/drbd/drbd_worker.c
2034
} while (cmpxchg(flags, old, new) != old);
drivers/block/null_blk/main.c
563
cmpxchg(&t_dev->badblocks.shift, -1, 0);
drivers/dax/super.c
103
if (!cmpxchg(&dax_dev->holder_data, NULL, holder))
drivers/dax/super.c
117
cmpxchg(&dax_dev->holder_data, holder, NULL) == holder)
drivers/dma-buf/dma-fence-array.c
36
cmpxchg(&array->base.error, PENDING_ERROR, error);
drivers/dma-buf/dma-fence-array.c
42
cmpxchg(&array->base.error, PENDING_ERROR, 0);
drivers/dma-buf/dma-fence-chain.c
65
tmp = unrcu_pointer(cmpxchg(&chain->prev, RCU_INITIALIZER(prev),
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1442
if (cmpxchg(&vm->process_info, NULL, *process_info) != NULL) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c
257
if (cmpxchg(&ctx->entities[hw_ip][ring], NULL, entity))
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
7475
} while (cmpxchg((struct dma_fence __force **)&adev->gang_submit,
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2131
} while (cmpxchg(&max_svm_range_pages, max_pages, _pages) != max_pages);
drivers/gpu/drm/i915/gem/i915_gem_mman.c
957
cmpxchg(&i915->gem.mmap_singleton, file, NULL);
drivers/gpu/drm/i915/gt/intel_gt_buffer_pool.c
197
if (cmpxchg(&node->age, age, 0) == age) {
drivers/gpu/drm/i915/gt/intel_gt_requests.c
100
if (cmpxchg(&tl->retire, NULL, STUB)) /* already queued */
drivers/gpu/drm/i915/i915_active.c
1062
while (cmpxchg(__active_fence_slot(active), prev, fence) != prev) {
drivers/gpu/drm/i915/i915_gpu_error.c
2246
cmpxchg(&i915->gpu_error.first_error, NULL, error))
drivers/gpu/drm/i915/i915_gpu_error.c
2273
cmpxchg(&gt->i915->gpu_error.first_error, NULL, error);
drivers/gpu/drm/i915/i915_gpu_error.c
998
if (cmpxchg(&error->sgl, NULL, m.sgl))
drivers/gpu/drm/i915/i915_request.c
169
!cmpxchg(&rq->engine->request_pool, NULL, rq))
drivers/gpu/drm/i915/i915_sw_fence.h
120
cmpxchg(&fence->error, 0, error);
drivers/gpu/drm/i915/i915_vma.c
625
if (unlikely(cmpxchg(&vma->iomap, NULL, ptr))) {
drivers/gpu/drm/scheduler/sched_entity.c
309
last_user = cmpxchg(&entity->last_user, current->group_leader, NULL);
drivers/gpu/drm/v3d/v3d_perfmon.c
317
cmpxchg(&v3d->global_perfmon, perfmon, NULL);
drivers/gpu/drm/v3d/v3d_perfmon.c
404
cmpxchg(&v3d->global_perfmon, perfmon, NULL);
drivers/gpu/drm/v3d/v3d_perfmon.c
491
if (cmpxchg(&v3d->global_perfmon, NULL, perfmon))
drivers/gpu/drm/vmwgfx/vmwgfx_cmd.c
162
if (fifo_mem && cmpxchg(fifo_mem + SVGA_FIFO_BUSY, 0, 1) == 0)
drivers/infiniband/hw/hfi1/rc.c
3127
(u64)cmpxchg((u64 *)qp->r_sge.sge.vaddr,
drivers/infiniband/sw/rdmavt/qp.c
3085
(u64)cmpxchg((u64 *)qp->r_sge.sge.vaddr,
drivers/iommu/s390-iommu.c
339
old_rse = cmpxchg(rsep, ZPCI_TABLE_INVALID, rse);
drivers/iommu/s390-iommu.c
369
old_rfe = cmpxchg(rfep, ZPCI_TABLE_INVALID, rfe);
drivers/iommu/s390-iommu.c
400
old_rte = cmpxchg(rtep, ZPCI_TABLE_INVALID, rte);
drivers/iommu/s390-iommu.c
426
old_ste = cmpxchg(step, ZPCI_TABLE_INVALID, ste);
drivers/iommu/sun50i-iommu.c
577
old_dte = cmpxchg(dte_addr, 0, dte);
drivers/md/dm-bufio.c
1146
(void)cmpxchg(&dm_bufio_cache_size, 0,
drivers/md/dm-bufio.c
1432
(void)cmpxchg(&c->async_write_error, 0,
drivers/md/dm-integrity.c
384
if (!cmpxchg(&ic->failed, 0, err))
drivers/md/dm-writecache.c
449
if (!cmpxchg(&(wc)->error, 0, err)) \
drivers/md/dm.c
165
(void)cmpxchg(module_param, param, modified_param);
drivers/md/dm.c
183
(void)cmpxchg(module_param, param, modified_param);
drivers/misc/genwqe/card_ddcb.c
301
icrc_hsi_shi = cmpxchg(&prev_ddcb->icrc_hsi_shi_32, old, new);
drivers/misc/genwqe/card_ddcb.c
653
icrc_hsi_shi = cmpxchg(&pddcb->icrc_hsi_shi_32,
drivers/misc/sgi-xp/xpc_main.c
1081
if (cmpxchg(&xpc_die_disconnecting, 0, 1))
drivers/misc/sgi-xp/xpc_uv.c
1286
if (func != NULL && cmpxchg(&msg_slot->func, func, NULL) == func) {
drivers/misc/sgi-xp/xpc_uv.c
1514
if (cmpxchg(&msg_slot->func, func, NULL) != func) {
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
616
(void)cmpxchg(&t->tid_tab[tid].ctx, ctx, NULL);
drivers/net/ethernet/fungible/funcore/fun_dev.c
420
return cmpxchg(&cmd_ctx->cb_data, cb_data, NULL) == cb_data;
drivers/net/ethernet/huawei/hinic/hinic_hw_dev.c
611
cb_state = cmpxchg(&nic_cb->cb_state,
drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c
253
eqe_state = cmpxchg(&hwe_cb->hwe_state,
drivers/net/ethernet/huawei/hinic/hinic_hw_eqs.c
298
eqe_state = cmpxchg(&ceq_cb->ceqe_state,
drivers/net/ethernet/huawei/hinic/hinic_hw_mgmt.c
410
cb_state = cmpxchg(&mgmt_cb->state,
drivers/net/ethernet/meta/fbnic/fbnic_phylink.c
292
if (cmpxchg(&fbd->pmd_state, FBNIC_PMD_TRAINING,
drivers/net/ethernet/meta/fbnic/fbnic_phylink.c
306
if (cmpxchg(&fbd->pmd_state, FBNIC_PMD_LINK_READY,
drivers/net/ethernet/sfc/mcdi.c
411
return cmpxchg(&mcdi->state,
drivers/net/ethernet/sfc/mcdi.c
422
cmpxchg(&mcdi->state,
drivers/net/ethernet/sfc/mcdi.c
454
if (cmpxchg(&mcdi->state,
drivers/net/ethernet/sfc/mcdi.c
506
if (cmpxchg(&mcdi->state,
drivers/net/ethernet/sfc/ptp.c
2009
(void) cmpxchg(&channel->sync_events_state, SYNC_EVENTS_REQUESTED,
drivers/net/ethernet/sfc/siena/mcdi.c
415
return cmpxchg(&mcdi->state,
drivers/net/ethernet/sfc/siena/mcdi.c
426
cmpxchg(&mcdi->state,
drivers/net/ethernet/sfc/siena/mcdi.c
458
if (cmpxchg(&mcdi->state,
drivers/net/ethernet/sfc/siena/mcdi.c
510
if (cmpxchg(&mcdi->state,
drivers/net/ethernet/sfc/siena/ptp.c
1967
(void) cmpxchg(&channel->sync_events_state, SYNC_EVENTS_REQUESTED,
drivers/nvme/target/fabrics-cmd.c
204
old = cmpxchg(&req->sq->ctrl, NULL, ctrl);
drivers/nvme/target/fabrics-cmd.c
212
old = cmpxchg(&req->cq->ctrl, NULL, ctrl);
drivers/pci/controller/pci-hyperv.c
3546
if (cmpxchg(&hbus->survey_event, NULL, &comp))
drivers/pci/pci.h
752
old = cmpxchg(&dev->error_state, pci_channel_io_normal,
drivers/pci/pci.h
756
old = cmpxchg(&dev->error_state, pci_channel_io_frozen,
drivers/phy/renesas/phy-rcar-gen2.c
90
if (cmpxchg(&channel->selected_phy, -1, phy->number) != -1)
drivers/platform/surface/aggregator/controller.c
61
while (unlikely((ret = cmpxchg(&c->value, old, new)) != old)) {
drivers/platform/surface/aggregator/controller.c
91
while (unlikely((ret = cmpxchg(&c->value, old, new)) != old)) {
drivers/platform/surface/aggregator/ssh_request_layer.c
391
if (cmpxchg(&rqst->packet.ptl, NULL, &rtl->ptl)) {
drivers/platform/surface/aggregator/ssh_request_layer.c
608
flags = cmpxchg(&r->state, fixed, SSH_REQUEST_SF_LOCKED_BIT);
drivers/s390/net/qeth_core.h
877
return cmpxchg(&channel->active_cmd, NULL, cmd) == NULL;
drivers/scsi/elx/efct/efct_hw.c
1970
if (cmpxchg(&io_to_abort->abort_in_progress, false, true)) {
drivers/scsi/libsas/sas_init.c
655
if (cmpxchg(&phy->in_shutdown, 0, 1) == 0) {
drivers/scsi/lpfc/lpfc_sli.c
14927
if (cmpxchg(&cq->queue_claimed, 0, 1) != 0)
drivers/scsi/lpfc/lpfc_sli.c
640
if (cmpxchg(&eq->queue_claimed, 0, 1) != 0)
drivers/scsi/scsi_lib.c
2918
if (cmpxchg(&sdev->queue_stopped, 1, 0))
drivers/scsi/scsi_lib.c
2930
if (!cmpxchg(&sdev->queue_stopped, 0, 1))
drivers/scsi/smartpqi/smartpqi_init.c
6514
if (cmpxchg(&scmd->host_scribble, PQI_NO_COMPLETION, (void *)&wait) == NULL) {
drivers/scsi/smartpqi/smartpqi_init.c
6524
if (cmpxchg(&tmf_work->scmd, NULL, scmd) == NULL) {
drivers/spi/spi-bcm2835.c
652
if (cmpxchg(&bs->rx_dma_active, true, false))
drivers/spi/spi-bcm2835.c
838
cmpxchg(&bs->rx_dma_active, true, false)) {
drivers/target/iscsi/iscsi_target.c
4248
cmpxchg(&conn->tx_thread_active, true, false)) {
drivers/target/iscsi/iscsi_target.c
4254
cmpxchg(&conn->rx_thread_active, true, false)) {
drivers/target/iscsi/iscsi_target.c
4538
sleep = cmpxchg(&conn->tx_thread_active, true, false);
drivers/target/iscsi/iscsi_target.c
4558
sleep = cmpxchg(&conn->tx_thread_active, true, false);
drivers/target/target_core_transport.c
775
if (cmpxchg(&cmd->lun_ref_active, true, false))
drivers/tty/tty_audit.c
194
if (cmpxchg(&current->signal->tty_audit_buf, NULL, buf) != NULL)
drivers/usb/gadget/function/f_fs.c
935
if (cmpxchg(&epfile->read_buffer, NULL, buf))
drivers/usb/gadget/function/f_fs.c
973
if (cmpxchg(&epfile->read_buffer, NULL, buf))
drivers/usb/gadget/function/f_fs.c
99
cmpxchg(&ffs->setup_state, FFS_SETUP_CANCELLED, FFS_NO_SETUP);
drivers/vdpa/vdpa_user/iova_domain.c
130
if (cmpxchg(&head_map->bounce_page, NULL, tmp_page))
drivers/xen/events/events_base.c
242
if (cmpxchg(&evtchn_to_irq[row], NULL, evtchn_row) != NULL)
fs/btrfs/bio.c
136
cmpxchg(&bbio->status, BLK_STS_OK, status);
fs/btrfs/raid56.c
226
x = cmpxchg(&info->stripe_hash_table, NULL, table);
fs/ceph/file.c
1401
cmpxchg(&aio_req->error, 0, rc);
fs/crypto/bio.c
67
cmpxchg(&done->status, 0, bio->bi_status);
fs/debugfs/file.c
120
d_fsd = cmpxchg(&dentry->d_fsdata, NULL, fsd);
fs/erofs/zdata.c
851
if (!cmpxchg(&fe->pcl->next, NULL, fe->head)) {
fs/ext4/orphan.c
80
} while (cmpxchg(&bdata[j], (__le32)0, cpu_to_le32(inode->i_ino)) !=
fs/fcntl.c
110
if (unlikely(cmpxchg(&file->f_owner, NULL, f_owner)))
fs/fs-writeback.c
301
if (unlikely(cmpxchg(&inode->i_wb, NULL, wb)))
fs/gfs2/glock.c
690
if (cmpxchg(&sdp->sd_log_error, 0, ret)) {
fs/gfs2/log.c
115
if (!cmpxchg(&sdp->sd_log_error, 0, -EIO))
fs/gfs2/log.c
325
!cmpxchg(&sdp->sd_log_error, 0, -EIO))
fs/gfs2/lops.c
209
if (!cmpxchg(&sdp->sd_log_error, 0, err))
fs/gfs2/quota.c
1556
if (!cmpxchg(&sdp->sd_log_error, 0, error))
fs/iomap/direct-io.c
176
cmpxchg(&dio->error, 0, ret);
fs/iomap/ioend.c
310
cmpxchg(&ioend->io_error, 0, error);
fs/jbd2/transaction.c
129
oldts = cmpxchg(&transaction->t_max_wait, oldts, newts);
fs/libfs.c
2205
old = cmpxchg(stashed, NULL, dentry);
fs/libfs.c
2292
cmpxchg(stashed, dentry, NULL);
fs/netfs/buffered_read.c
320
cmpxchg(&rreq->error, 0, ret);
fs/nfs/filelayout/filelayout.c
581
if (cmpxchg(&fl->dsaddr, NULL, dsaddr) != NULL)
fs/nfs/flexfilelayout/flexfilelayoutdev.c
339
if (cmpxchg(&mirror->dss[dss_id].mirror_ds, NULL, mirror_ds) &&
fs/nfs/nfs3acl.c
25
cmpxchg(p, ACL_NOT_CACHED, sentinel);
fs/nfs/nfs3acl.c
34
if (cmpxchg(p, sentinel, acl) != sentinel)
fs/nfs/nfs3acl.c
43
cmpxchg(p, sentinel, ACL_NOT_CACHED);
fs/nfs/nfs3proc.c
903
cmpxchg(&server->read_hdrsize, 0, hdr->res.replen);
fs/nfs/nfs4idmap.c
563
if (cmpxchg(&idmap->idmap_upcall_data, data, NULL) == data)
fs/nfsd/localio.c
100
new = unrcu_pointer(cmpxchg(pnf, NULL, RCU_INITIALIZER(localio)));
fs/nfsd/nfs4state.c
4939
cmpxchg(&so->so_replay.rp_locked,
fs/nfsd/state.h
876
cmpxchg(&clp->cl_state, NFSD4_COURTESY, NFSD4_EXPIRABLE);
fs/notify/mark.c
657
if (cmpxchg(&sb->s_fsnotify_info, NULL, sbinfo)) {
fs/notify/mark.c
786
if (cmpxchg(connp, NULL, conn)) {
fs/posix_acl.c
146
cmpxchg(p, ACL_NOT_CACHED, sentinel);
fs/posix_acl.c
169
cmpxchg(p, sentinel, ACL_NOT_CACHED);
fs/userfaultfd.c
2023
if (cmpxchg(&ctx->features, 0, ctx_features) != 0)
include/linux/kvm_host.h
567
return cmpxchg(&vcpu->mode, IN_GUEST_MODE, EXITING_GUEST_MODE);
include/linux/netpoll.h
86
while (cmpxchg(&napi->poll_owner, -1, owner) != -1)
include/linux/skmsg.h
530
if (cmpxchg(pprog, old, prog) != old)
kernel/acct.c
196
cmpxchg(&acct->ns->bacct, pin, NULL);
kernel/bpf/bpf_local_storage.c
607
prev_storage = cmpxchg(owner_storage_ptr, NULL, storage);
kernel/bpf/bpf_struct_ops.c
929
prev_state = cmpxchg(&st_map->kvalue.common.state,
kernel/bpf/devmap.c
1169
odev = unrcu_pointer(cmpxchg(&dtab->netdev_map[i], RCU_INITIALIZER(dev), NULL));
kernel/bpf/helpers.c
1355
old_cb = cmpxchg(&async->cb, NULL, cb);
kernel/bpf/helpers.c
2397
if (cmpxchg(&node->owner, NULL, BPF_PTR_POISON)) {
kernel/bpf/helpers.c
2516
if (cmpxchg(&node->owner, NULL, BPF_PTR_POISON)) {
kernel/bpf/helpers.c
4206
state = cmpxchg(&ctx->state, BPF_TW_SCHEDULING, BPF_TW_RUNNING);
kernel/bpf/helpers.c
4208
state = cmpxchg(&ctx->state, BPF_TW_SCHEDULED, BPF_TW_RUNNING);
kernel/bpf/helpers.c
4221
(void)cmpxchg(&ctx->state, BPF_TW_RUNNING, BPF_TW_STANDBY);
kernel/bpf/helpers.c
4234
if (cmpxchg(&ctx->state, BPF_TW_PENDING, BPF_TW_SCHEDULING) != BPF_TW_PENDING) {
kernel/bpf/helpers.c
4246
(void)cmpxchg(&ctx->state, BPF_TW_SCHEDULING, BPF_TW_STANDBY);
kernel/bpf/helpers.c
4258
state = cmpxchg(&ctx->state, BPF_TW_SCHEDULING, BPF_TW_SCHEDULED);
kernel/bpf/helpers.c
4281
old_ctx = cmpxchg(&twk->ctx, NULL, ctx);
kernel/bpf/helpers.c
4307
if (cmpxchg(&ctx->state, BPF_TW_STANDBY, BPF_TW_PENDING) != BPF_TW_STANDBY) {
kernel/debug/kdb/kdb_io.c
649
old_cpu = cmpxchg(&kdb_printf_cpu, -1, this_cpu);
kernel/exit.c
1199
if (cmpxchg(&p->exit_state, EXIT_ZOMBIE, state) != EXIT_ZOMBIE)
kernel/kcsan/kcsan_test.c
588
KCSAN_EXPECT_READ_BARRIER(cmpxchg(&test_var, 0, 0), true);
kernel/kcsan/kcsan_test.c
633
KCSAN_EXPECT_WRITE_BARRIER(cmpxchg(&test_var, 0, 0), true);
kernel/kcsan/kcsan_test.c
678
KCSAN_EXPECT_RW_BARRIER(cmpxchg(&test_var, 0, 0), true);
kernel/kcsan/selftest.c
153
KCSAN_CHECK_READ_BARRIER(cmpxchg(&test_var, 0, 0));
kernel/kcsan/selftest.c
182
KCSAN_CHECK_WRITE_BARRIER(cmpxchg(&test_var, 0, 0));
kernel/kcsan/selftest.c
214
KCSAN_CHECK_RW_BARRIER(cmpxchg(&test_var, 0, 0));
kernel/locking/osq_lock.c
165
cmpxchg(&prev->next, node, NULL) == node)
kernel/locking/qspinlock_paravirt.h
330
cmpxchg(&pn->state, VCPU_HALTED, VCPU_RUNNING);
kernel/rcu/rcutorture.c
1260
cmpxchg(&last_persist, lp, j) == lp) {
kernel/rcu/srcutree.c
776
old_read_flavor = cmpxchg(&sdp->srcu_reader_flavor, 0, read_flavor);
kernel/rcu/tree_stall.h
835
cmpxchg(&rcu_state.jiffies_stall, js, jn) == js) {
kernel/sched/fair.c
1946
cmpxchg(&pgdat->nbp_rl_start, start, now) == start)
kernel/sched/fair.c
1966
cmpxchg(&pgdat->nbp_th_start, start, now) == start) {
kernel/sched/fair.c
3491
if (cmpxchg(&vma->numab_state, NULL, ptr)) {
kernel/sched/psi.c
1500
if (cmpxchg(&t->event, 1, 0) == 1)
kernel/sched/psi.c
513
if (cmpxchg(&t->event, 0, 1) == 0) {
kernel/trace/ring_buffer.c
1353
ret = cmpxchg((unsigned long *)&list->next,
kernel/trace/ring_buffer.c
3305
(void)cmpxchg(&meta->head_buffer, old_head, new_head);
kernel/trace/trace_functions.c
669
new_count = cmpxchg(count, old_count, new_count);
kernel/trace/trace_recursion_record.c
112
cmpxchg(&recursed_functions[index].ip, ip, 0);
kernel/trace/trace_recursion_record.c
78
old = cmpxchg(&recursed_functions[index].ip, 0, ip);
kernel/trace/tracing_map.c
566
if (!cmpxchg(&entry->key, 0, key_hash)) {
lib/atomic64_test.c
86
FAMILY_TEST(TEST_ARGS, bit, cmpxchg, \
lib/atomic64_test.c
88
FAMILY_TEST(TEST_ARGS, bit, cmpxchg, \
lib/cmpxchg-emu.c
41
ret = data_race(cmpxchg(p32, old32.w, new32.w)); // Overridden above.
lib/errseq.c
202
cmpxchg(eseq, old, new);
lib/errseq.c
96
cur = cmpxchg(eseq, old, new);
lib/lru_cache.c
59
val = cmpxchg(&lc->flags, 0, LC_LOCKED);
lib/rhashtable.c
141
if (cmpxchg((union nested_table **)prev, NULL, ntbl) == NULL)
lib/rhashtable.c
312
if (cmpxchg((struct bucket_table **)&old_tbl->future_tbl, NULL,
mm/huge_memory.c
230
if (cmpxchg(&huge_zero_folio, NULL, zero_folio)) {
mm/kmemleak.c
2248
if (cmpxchg(&kmemleak_error, 0, 1))
mm/memcontrol.c
1080
if (cmpxchg(&iter->position, pos, next) != pos) {
mm/memcontrol.c
1132
cmpxchg(&iter->position, dead_memcg, NULL);
mm/oom_kill.c
778
if (!cmpxchg(&tsk->signal->oom_mm, NULL, mm))
mm/page_io.c
336
if (cmpxchg(&sio_pool, NULL, pool))
mm/slub.c
2072
return cmpxchg(&slab->obj_exts, 0, OBJEXTS_ALLOC_FAIL) == 0;
mm/slub.c
2212
} else if (cmpxchg(&slab->obj_exts, old_exts, new_exts) != old_exts) {
mm/swapfile.c
2764
cmpxchg(&pcp_si[i], si, NULL);
net/can/isotp.c
1205
cmpxchg(&so->tx.state, ISOTP_IDLE, ISOTP_SHUTDOWN) != ISOTP_IDLE)
net/can/isotp.c
966
while (cmpxchg(&so->tx.state, ISOTP_IDLE, ISOTP_SENDING) != ISOTP_IDLE) {
net/core/dev.c
11831
if (p && cmpxchg(&dev->core_stats, NULL, p))
net/core/dev.c
5294
} else if (!cmpxchg(&sd->defer_ipi_scheduled, 0, 1)) {
net/core/dev.c
6949
if (cmpxchg(&napi->state, val,
net/core/dst.c
204
prev = cmpxchg(&dst->_metrics, old, new);
net/core/dst.c
227
prev = cmpxchg(&dst->_metrics, old, new);
net/core/filter.c
4964
if (cmpxchg(&md_dst, NULL, tmp))
net/core/lwtunnel.c
103
ret = (cmpxchg((const struct lwtunnel_encap_ops **)
net/core/lwtunnel.c
88
return !cmpxchg((const struct lwtunnel_encap_ops **)
net/core/netpoll.c
169
if (cmpxchg(&napi->poll_owner, -1, cpu) == -1) {
net/core/sock.c
1263
cmpxchg(&sk->sk_pacing_status,
net/core/sock_diag.c
210
return !cmpxchg((const struct sock_diag_handler **)
net/ipv4/gre_demux.c
36
return (cmpxchg((const struct gre_protocol **)&gre_proto[version], NULL, proto) == NULL) ?
net/ipv4/gre_demux.c
48
ret = (cmpxchg((const struct gre_protocol **)&gre_proto[version], proto, NULL) == proto) ?
net/ipv4/icmp.c
258
if (cmpxchg(&net->ipv4.icmp_global_stamp, oldstamp, now) == oldstamp) {
net/ipv4/igmp.c
821
if (cmpxchg(&in_dev->mr_ifc_count,
net/ipv4/inet_diag.c
1055
return !cmpxchg((const struct inet_diag_handler **)&inet_diag_table[type],
net/ipv4/ip_tunnel.c
465
return !cmpxchg((const struct ip_tunnel_encap_ops **)
net/ipv4/ip_tunnel.c
479
ret = (cmpxchg((const struct ip_tunnel_encap_ops **)
net/ipv4/protocol.c
34
return !cmpxchg((const struct net_protocol **)&inet_protos[protocol],
net/ipv4/protocol.c
41
return !cmpxchg((const struct net_offload **)&inet_offloads[protocol],
net/ipv4/protocol.c
50
ret = (cmpxchg((const struct net_protocol **)&inet_protos[protocol],
net/ipv4/protocol.c
63
ret = (cmpxchg((const struct net_offload **)&inet_offloads[protocol],
net/ipv4/route.c
1506
prev = cmpxchg(p, orig, rt);
net/ipv4/route.c
482
if (old != now && cmpxchg(p_tstamp, old, now) == old)
net/ipv4/tcp.c
341
if (!cmpxchg(&tcp_memory_pressure, 0, val))
net/ipv4/tcp.c
3647
if (cmpxchg(&__tcp_tx_delay_enabled, 0, 1) == 0) {
net/ipv4/tcp_bbr.c
1078
cmpxchg(&sk->sk_pacing_status, SK_PACING_NONE, SK_PACING_NEEDED);
net/ipv6/ip6_fib.c
744
if (cmpxchg(&f6i->fib6_metrics, dflt, p) != dflt)
net/ipv6/ip6_icmp.c
18
return (cmpxchg((ip6_icmp_send_t **)&ip6_icmp_send, NULL, fn) == NULL) ?
net/ipv6/ip6_icmp.c
27
ret = (cmpxchg((ip6_icmp_send_t **)&ip6_icmp_send, fn, NULL) == fn) ?
net/ipv6/ip6_tunnel.c
1789
return !cmpxchg((const struct ip6_tnl_encap_ops **)
net/ipv6/ip6_tunnel.c
1803
ret = (cmpxchg((const struct ip6_tnl_encap_ops **)
net/ipv6/protocol.c
30
return !cmpxchg((const struct inet6_protocol **)&inet6_protos[protocol],
net/ipv6/protocol.c
39
ret = (cmpxchg((const struct inet6_protocol **)&inet6_protos[protocol],
net/ipv6/protocol.c
54
return !cmpxchg((const struct net_offload **)&inet6_offloads[protocol],
net/ipv6/protocol.c
63
ret = (cmpxchg((const struct net_offload **)&inet6_offloads[protocol],
net/ipv6/route.c
1473
prev = cmpxchg(p, NULL, pcpu_rt);
net/ipv6/route.c
697
if (!work || cmpxchg(&fib6_nh->last_probe,
net/mptcp/mib.c
102
if (cmpxchg(&net->mib.mptcp_statistics, NULL, mib))
net/netfilter/ipvs/ip_vs_sync.c
531
n = cmpxchg(&cp->sync_endtime, orig, n);
net/netfilter/nf_bpf_link.c
121
if (!cmpxchg(&nf_link->dead, 0, 1)) {
net/netfilter/nf_conntrack_ecache.c
160
} while (cmpxchg(&e->missed, old, want) != old);
net/netfilter/nf_conntrack_labels.c
23
} while (cmpxchg(address, old, tmp) != old);
net/netfilter/nf_flow_table_core.c
551
cmpxchg(&ct->timeout, expires, new_timeout);
net/netfilter/xt_limit.c
89
} while (cmpxchg(&priv->credit, old_credit, new_credit) != old_credit);
net/rds/ib_frmr.c
141
if (cmpxchg(&frmr->fr_state, FRMR_IS_FREE, FRMR_IS_INUSE) != FRMR_IS_FREE) {
net/rds/ib_frmr.c
40
if (cmpxchg(&ibmr->u.frmr.fr_state,
net/rds/ib_recv.c
509
old = cmpxchg(&cache->xfer, NULL, chpfirst);
net/rxrpc/io_thread.c
366
old_id = cmpxchg(&conn->service_id, conn->orig_service_id,
net/smc/smc_core.c
2121
if (cmpxchg(&buf_slot->used, 0, 1) == 0) {
net/smc/smc_llc.h
50
(cmpxchg(state, SMC_LNK_ACTIVE, SMC_LNK_INACTIVE) == SMC_LNK_ACTIVE)
net/sunrpc/auth.c
113
old = cmpxchg((const struct rpc_authops ** __force)&auth_flavors[flavor], NULL, ops);
net/sunrpc/auth.c
129
old = cmpxchg((const struct rpc_authops ** __force)&auth_flavors[flavor], ops, NULL);
net/sunrpc/auth_gss/auth_gss.c
2133
seq_xmit = cmpxchg(&ctx->gc_seq_xmit, tmp, *req->rq_seqnos);
net/sunrpc/auth_gss/svcauth_gss.c
1359
ret = cmpxchg(&sn->use_gss_proxy, -1, type);
net/sunrpc/sched.c
70
if (cmpxchg(&task->tk_rpc_status, 0, rpc_status) == 0)
net/sunrpc/svcauth.c
147
old = cmpxchg((struct auth_ops ** __force)&authtab[flavor], NULL, aops);
net/tipc/net.c
135
if (cmpxchg(&tn->node_addr, 0, addr))
net/xfrm/xfrm_interface_bpf.c
77
if (cmpxchg(&xfrm_bpf_md_dst, NULL, tmp))
sound/soc/dwc/dwc-pcm.c
114
cmpxchg(&dev->tx_ptr, ptr, new_ptr);
sound/soc/dwc/dwc-pcm.c
119
cmpxchg(&dev->rx_ptr, ptr, new_ptr);
sound/soc/intel/keembay/kmb_platform.c
238
cmpxchg(&kmb_i2s->tx_ptr, ptr, new_ptr);
sound/soc/intel/keembay/kmb_platform.c
243
cmpxchg(&kmb_i2s->rx_ptr, ptr, new_ptr);
sound/soc/sof/ipc3-dtrace.c
256
u32 ret = cmpxchg(&priv->host_offset, host_offset, new_offset);
sound/soc/xtensa/xtfpga-i2s.c
169
cmpxchg(&i2s->tx_ptr, tx_ptr, new_tx_ptr);
tools/arch/x86/include/asm/atomic.h
71
return cmpxchg(&v->counter, old, new);
tools/include/asm-generic/atomic-gcc.h
70
return cmpxchg(&(v)->counter, oldval, newval);
tools/testing/selftests/bpf/bpf_atomic.h
52
__unqual_typeof(*(p)) __r = cmpxchg(p, __o, new); \
tools/testing/selftests/bpf/prog_tests/arena_atomics.c
113
prog_fd = bpf_program__fd(skel->progs.cmpxchg);
tools/testing/selftests/bpf/prog_tests/atomics.c
129
prog_fd = skel->progs.cmpxchg.prog_fd;
tools/testing/selftests/bpf/progs/arena_atomics.c
178
int cmpxchg(const void *ctx)
tools/testing/selftests/bpf/progs/atomics.c
136
int cmpxchg(const void *ctx)
tools/testing/selftests/powerpc/benchmarks/context_switch.c
288
c = cmpxchg(m, 0, 1);