atomic_fetch_inc
atomic_fetch_inc(&vcpu->kvm->arch.vgic.active_spis) == 0);
atomic_fetch_inc(&dev->open_count);
if (atomic_fetch_inc(&vdev->job_timeout_counter) >= inference_max_retries) {
atomic_fetch_inc(&dev->open_count);
if (!atomic_fetch_inc(&vma->open_count))
if (!atomic_fetch_inc(&obj->mm.shrink_pin) &&
if (atomic_fetch_inc(&pt->used) >> gen8_pd_shift(1) &&
if (!atomic_fetch_inc(&tl->active_count))
if (atomic_fetch_inc(&ring->pin_count))
if (!atomic_fetch_inc(&slpc->num_waiters)) {
if (atomic_fetch_inc(&rps->num_waiters))
if (atomic_fetch_inc(&tl->pin_count)) {
if (!atomic_fetch_inc(&tl->active_count)) {
if (!atomic_fetch_inc(&ref->count))
if (!atomic_fetch_inc(&t->count))
if (!atomic_fetch_inc(&config_desc->active_cnt)) {
if (atomic_fetch_inc(&dd->ipoib_rsm_usr_num) == 0) {
if (atomic_fetch_inc(&rx->netdevs))
if (atomic_fetch_inc(&rx->enabled))
unsigned long packet_count = atomic_fetch_inc(&hist.count);
val = atomic_fetch_inc(count);
alloc = !atomic_fetch_inc(&desc->its_vmapp_cmd.vpe->vmapp_count);
ev.u.frame_sync.frame_sequence = atomic_fetch_inc(&stream->sequence);
if (!atomic_fetch_inc(&nn->ktls_no_space))
if (!atomic_fetch_inc(&ccmni_ctl->napi_usr_refcnt))
idx = (u32)atomic_fetch_inc(&scan_info->next_nsid);
this_id = atomic_fetch_inc(&id);
ni->id = atomic_fetch_inc(&id);
atomic_fetch_inc(&pmu_idx[0]));
atomic_fetch_inc(&pmu_idx[pmu_type]));
atomic_fetch_inc(&pmu_generic_idx));
if (atomic_fetch_inc(&msc->online_refs) == 0)
if (all_devices_probed && !atomic_fetch_inc(&once))
ev.u.frame_sync.frame_sequence = atomic_fetch_inc(&stream->sequence);
buf_id = atomic_fetch_inc(&stream->buf_id);
h->u.h_seq = cpu_to_le32(atomic_fetch_inc(&rd->node->seq_send));
mh->seq = atomic_fetch_inc(&mh->node->seq_send);
return !(atomic_fetch_inc(&req->poll_refs) & IO_POLL_REF_MASK);
return !(atomic_fetch_inc(&req->poll_refs) & IO_POLL_REF_MASK);
if (atomic_fetch_inc(&iw->refs) & IO_WAITID_REF_MASK)
if (atomic_fetch_inc(&iw->refs) & IO_WAITID_REF_MASK)
unsigned int id = atomic_fetch_inc(&sig->next_posix_timer_id) & INT_MAX;
if (!atomic_fetch_inc(&watchdog_cpus))
htonl(atomic_fetch_inc(&tunnel->o_seqno)) : 0);
htonl(atomic_fetch_inc(&tunnel->o_seqno)) : 0);
htonl(atomic_fetch_inc(&tunnel->o_seqno)));
pkt_cnt = atomic_fetch_inc(&ilwt->pkt_cnt);
htonl(atomic_fetch_inc(&t->o_seqno)));
htonl(atomic_fetch_inc(&tunnel->o_seqno)) :
htonl(atomic_fetch_inc(&tunnel->o_seqno)) :
if (!atomic_fetch_inc(&local->phy->hold_txs))
return atomic_fetch_inc(v);