atomic_try_cmpxchg
if (val || !atomic_try_cmpxchg(&lock->val, &val, _Q_LOCKED_VAL)) {
if (val || !atomic_try_cmpxchg(&lock->val, &val, _Q_LOCKED_VAL)) {
} while (!atomic_try_cmpxchg(&eilvt_offsets[offset], &rsvd, new));
if (!atomic_try_cmpxchg(&nmi_cpu, &old_cpu, this_cpu))
if (!atomic_try_cmpxchg(&stopping_cpu, &old_cpu, this_cpu))
while (old && !atomic_try_cmpxchg(&blkg->use_delay, &old, old - 1))
if (!old && atomic_try_cmpxchg(&blkg->use_delay, &old, -1))
if (old && atomic_try_cmpxchg(&blkg->use_delay, &old, 0))
if (!atomic_try_cmpxchg(&iolat->scale_cookie, &our_cookie, cur_cookie)) {
} while (!atomic_try_cmpxchg(v, &cur, cur + 1));
if (atomic_try_cmpxchg(&cxlrd->region_id, &id, cxlr->id)) {
} while (!atomic_try_cmpxchg(&fence->pending, &pending, pending + 1));
} while (!atomic_try_cmpxchg(&vma->flags, &bound, bound + 1));
if (can_wake && atomic_try_cmpxchg(&q->signaled, &signaled, 0)) {
if (atomic_try_cmpxchg(&q->signaled, &signaled, entry->index)) {
if (atomic_try_cmpxchg(&q->signaled, &signaled, entry->index)) {
if (atomic_try_cmpxchg(&q->signaled, &signaled, entry->index)) {
} while (!atomic_try_cmpxchg(&trans_info->tre_avail, &avail, new));
if (atomic_try_cmpxchg(&fst->seq, &old_seq, seq) ||
} while (!atomic_try_cmpxchg(&ctx->reqs_available,
if (atomic_try_cmpxchg(&eb->refs.refs, &refs, refs - 1))
if (atomic_try_cmpxchg(&sdp->sd_log_revokes_available,
if (atomic_try_cmpxchg(&sdp->sd_log_blks_free, &free_blocks,
if (atomic_try_cmpxchg(&sdp->sd_log_blks_free,
return atomic_try_cmpxchg(&lock->val, &old, old + (1<<16)); /* SC, for RCsc */
} while (!likely(atomic_try_cmpxchg(&key->enabled, &v, v + 1)));
} while (!atomic_try_cmpxchg(uref, &old, old - 1));
if (!atomic_try_cmpxchg(&rb->busy, &busy, 1))
if (val || !atomic_try_cmpxchg(&lock->val, &val, 1)) {
if (!atomic_try_cmpxchg(st, &sync, next_state))
} while (!atomic_try_cmpxchg(st, &sync, SYNC_STATE_SHOULD_DIE));
if (!atomic_try_cmpxchg(st, &sync, SYNC_STATE_KICKED))
} while (!atomic_try_cmpxchg(&q->requeue_state, &old, new));
} while (!atomic_try_cmpxchg(&q->requeue_state, &old, new));
} while (!atomic_try_cmpxchg(&q->requeue_state, &old, new));
} while (!likely(atomic_try_cmpxchg(&key->enabled, &v, v + 1)));
} while (!likely(atomic_try_cmpxchg(&key->enabled, &v, v - 1)));
} while (!atomic_try_cmpxchg(&last_used_nid, &old_node, nid));
if (!atomic_try_cmpxchg(&panic_redirect_cpu, &old_cpu, this_cpu))
return atomic_try_cmpxchg(&panic_cpu, &old_cpu, this_cpu);
return atomic_try_cmpxchg(&ACCESS_PRIVATE(con, nbcon_state), &cur->atom, new->atom);
if (atomic_try_cmpxchg(&sch->exit_kind, &kind, SCX_EXIT_DONE))
if (!atomic_try_cmpxchg(&sch->exit_kind, &none, kind))
return atomic_try_cmpxchg(&scx_enable_state_var, &from_v, to);
if (inp || !atomic_try_cmpxchg(&in_progress, &inp, 1))
if (atomic_try_cmpxchg(&group->migr_state, &curstate.state, newstate.state)) {
} while (!atomic_try_cmpxchg(&group->migr_state, &curstate.state, newstate.state));
} while (!atomic_try_cmpxchg(&buffer->record_disabled, &rd, new_rd));
} while (!atomic_try_cmpxchg(&buffer->record_disabled, &rd, new_rd));
} while (!atomic_try_cmpxchg(&sbq->wakeup_cnt,
KUNIT_EXPECT_KASAN_FAIL(test, atomic_try_cmpxchg(unsafe, safe, 42));
KUNIT_EXPECT_KASAN_FAIL(test, atomic_try_cmpxchg(safe, unsafe, 42));
if (!atomic_try_cmpxchg(&mm->tlb_flush_batched, &batch, 1))
} while (!atomic_try_cmpxchg(&sc->ids, &old_ids, new_ids));
if (atomic_try_cmpxchg(&netstamp_wanted, &wanted, wanted + 1))
if (atomic_try_cmpxchg(&netstamp_wanted, &wanted, wanted - 1))
} while (!atomic_try_cmpxchg(&net->ipv4.icmp_global_credit, &old, new));
} while (!atomic_try_cmpxchg(&net->ipv6.fib6_sernum, &old, new));
if (!atomic_try_cmpxchg(&conn->active, &active, -1))
return atomic_try_cmpxchg(v, old, new);
} while (!atomic_try_cmpxchg(&mbochs_avail_mbytes, &avail_mbytes,
} while (!atomic_try_cmpxchg(&mdev_avail_ports,
return atomic_try_cmpxchg(&ep->state, &old, new);
} while (!atomic_try_cmpxchg(v, &c, c + 1));