atomic_inc_not_zero
if (!atomic_inc_not_zero(&active_events)) {
if (!atomic_inc_not_zero(&active_events)) {
if (!atomic_inc_not_zero(&num_events)) {
if (!atomic_inc_not_zero(&num_events)) {
if (!atomic_inc_not_zero(&kvm->mm->context.protected_count))
if (!atomic_inc_not_zero(&num_events)) {
if (atomic_inc_not_zero(&active_events))
if (!atomic_inc_not_zero(&pmc_refcount)) {
if (!atomic_inc_not_zero(&x86_pmu.lbr_exclusive[what])) {
if (atomic_inc_not_zero(&srso_nr_vms))
return atomic_inc_not_zero(&req->ref);
retval = atomic_inc_not_zero(&dev->power.usage_count);
if (!atomic_inc_not_zero(&priv->chan[ch].submit_count)) {
if (!atomic_inc_not_zero(&aux_dev->usecount))
if (!atomic_inc_not_zero(&aux_dev->usecount))
if (atomic_inc_not_zero(&obj->mm.pages_pin_count))
if (!atomic_inc_not_zero(&obj->mm.pages_pin_count)) {
return atomic_inc_not_zero(&ce->pin_count);
return atomic_inc_not_zero(&wf->count);
if (unlikely(!atomic_inc_not_zero(&wf->count)))
if (atomic_inc_not_zero(&pfdev->cycle_counter.use_count))
if (atomic_inc_not_zero(&sh->count))
if (!atomic_inc_not_zero(&ch->kthreads_assigned)) {
if (WARN_ON(!atomic_inc_not_zero(&ctx->num_inflight)))
if (!atomic_inc_not_zero(active_events)) {
if (atomic_inc_not_zero(&hvc_needs_init)) {
if (atomic_inc_not_zero(&realm->nref))
if (unlikely(!atomic_inc_not_zero(&p->count)))
} while (unlikely(!atomic_inc_not_zero(&wpa->bucket->count)));
if (unlikely(!__kernfs_active(kn) || !atomic_inc_not_zero(&kn->count)))
atomic_inc_not_zero(&entry->e_refcnt))
atomic_inc_not_zero(&entry->e_refcnt))
if (!atomic_inc_not_zero(&sp->so_count))
if (d != NULL && !atomic_inc_not_zero(&d->ref))
if (!atomic_inc_not_zero(&sb->s_active))
if ((sb != old_sb) && !atomic_inc_not_zero(&sb->s_active))
if (!atomic_inc_not_zero(&share->refcount))
else if (!atomic_inc_not_zero(&tcon->refcount))
if (!atomic_inc_not_zero(&opinfo->refcount))
if (!atomic_inc_not_zero(&opinfo->refcount))
if (!atomic_inc_not_zero(&opinfo->refcount))
if (!atomic_inc_not_zero(&brk_op->refcount))
if (opinfo && !atomic_inc_not_zero(&opinfo->refcount))
!atomic_inc_not_zero(&opinfo->refcount))
if (!atomic_inc_not_zero(&opinfo->refcount))
if (atomic_inc_not_zero(&ci->m_count))
if (!atomic_inc_not_zero(&fp->refcount))
active = atomic_inc_not_zero(&sb->s_active);
active = atomic_inc_not_zero(&sb->s_active);
if (atomic_inc_not_zero(&sb->s_active)) {
if (!atomic_inc_not_zero(&xg->xg_active_ref))
if (!atomic_inc_not_zero(&xg->xg_active_ref))
if (atomic_inc_not_zero(&xg->xg_active_ref))
if (!atomic_inc_not_zero(&oz->oz_ref))
if (!atomic_inc_not_zero(&oz->oz_ref))
return atomic_inc_not_zero(&mm->mm_users);
return atomic_inc_not_zero(&req->refs);
if (atomic_inc_not_zero(&perf_sched_count))
WARN_ON_ONCE(!atomic_inc_not_zero(&epc->refcount));
if (atomic_inc_not_zero(&nr_include_guest_events))
if (atomic_inc_not_zero(&nr_mediated_pmu_vms))
atomic_inc_not_zero(&module->refcnt) != 0))
return atomic_inc_not_zero(&rtsp->rts_refctr);
if (atomic_inc_not_zero(&sched_core_count))
if (likely(atomic_inc_not_zero(&huge_zero_refcount)))
KUNIT_EXPECT_KASAN_FAIL(test, atomic_inc_not_zero(unsafe));
return atomic_inc_not_zero(&object->use_count);
if (!atomic_inc_not_zero(&anon_vma->refcount)) {
if (!atomic_inc_not_zero(&anon_vma->refcount)) {
atomic_inc_not_zero(&pvcc->inflight))
atomic_inc_not_zero(&pvcc->inflight))
if (fl->label == label && atomic_inc_not_zero(&fl->users)) {
if (!atomic_inc_not_zero(&fl1->users))
if (fl && !atomic_inc_not_zero(&fl->users))
return atomic_inc_not_zero(v);
if (!atomic_inc_not_zero(&codec->in_pm))
if (!atomic_inc_not_zero(&codec->in_pm)) {