mutex_trylock
if (!mutex_trylock(&victim->state_mutex)) {
if (mutex_trylock(&hotplug_mutex)) {
if (mutex_trylock(&hotplug_mutex)) {
locked = mutex_trylock(&ctrblk_lock);
locked = mutex_trylock(&ctrblk_lock);
locked = mutex_trylock(&ctrblk_lock);
if (!mutex_trylock(&kvm->lock))
if (!mutex_trylock(&alloc->mutex))
if (mutex_trylock(&device_hotplug_lock))
trylock_failed = !mutex_trylock(&b->bm_change);
if (!mutex_trylock(device->state_mutex)) {
if (!mutex_trylock(&cmd->lock))
if (!mutex_trylock(&sess->lock))
if (mutex_trylock(&mc_bus->scan_mutex)) {
if (!mutex_trylock(&oppanel_mutex)) {
if (!mutex_trylock(&prepare_lock)) {
while (!mutex_trylock(&vf_info->pfvf_mig_lock)) {
if (!mutex_trylock(&rng->lock) && !wait)
if (mutex_trylock(&tegra->devfreq->lock)) {
if (!mutex_trylock(&rwtm->busy))
if (!mutex_trylock(&bridge->mutex))
if (!mutex_trylock(&mgr->ref_mutex)) {
if (!mutex_trylock(®ion->mutex)) {
retval = mutex_trylock(&a_priv->bulk_transfer_lock);
retval = mutex_trylock(&ni_priv->addressed_transfer_lock);
if (!mutex_trylock(&adev->psp.mutex))
if (mutex_trylock(&vm->eviction_lock)) {
return mutex_trylock(&i2c_to_aux(i2c)->hw_mutex);
if (!mutex_trylock(&dev->mode_config.mutex)) {
return mutex_trylock(&display->gmbus.mutex);
if (!mutex_trylock(&ce->timeline->mutex)) {
if (!mutex_trylock(&tl->mutex)) {
if (!mutex_trylock(&tl->mutex)) {
if (mutex_trylock(&tl->mutex)) {
if (!mutex_trylock(&tl->mutex)) {
if (mutex_trylock(&vma->vm->mutex)) {
if (trylock_vm && !mutex_trylock(&vm->mutex)) {
if (flags & INTEL_WAKEREF_PUT_ASYNC || !mutex_trylock(&wf->mutex)) {
if (!mutex_trylock(&disp->mutex)) {
else if (!mutex_trylock(&sn->svmm->mutex))
if (!mutex_trylock(&cli->mutex))
if (!mutex_trylock(&pfdev->shrinker_lock))
if (!mutex_trylock(&bo->mappings.lock))
if (!mutex_trylock(&pfdev->shrinker_lock))
else if (!mutex_trylock(®ion->mreg_mutex))
if (!mutex_trylock(&watchdog_data_mutex))
if (!mutex_trylock(&watchdog_data_mutex))
if (!mutex_trylock(&debug_lock))
if (!mutex_trylock(&hisi_ptt->filter_lock)) {
return mutex_trylock(&atr->lock);
if (mutex_trylock(&data->lock)) {
if (!mutex_trylock(&rdev->pacing.dbq_lock))
if (!mutex_trylock(&rdev->pacing.dbq_lock))
while (!mutex_trylock(&ppd->hls_lock))
while (!mutex_trylock(&dd->pport->hls_lock))
if (!mutex_trylock(&cyapa->state_sync_lock)) {
else if (!mutex_trylock(&c->bucket_lock))
if (mutex_trylock(&b->c->bucket_lock)) {
if (mutex_trylock(&b->c->bucket_lock)) {
if (mutex_trylock(&bch_register_lock)) {
ret = mutex_trylock(&mddev->reconfig_mutex);
mutex_trylock(&conf->cache_size_mutex)) {
if (mutex_trylock(&conf->cache_size_mutex)) {
if (mutex_trylock(&ir->lock)) {
mutex_locked = mutex_trylock(&dev->mfc_mutex);
if (mutex_trylock(&dev->lock)) {
const bool unlock = mutex_trylock(&ictx->lock);
if (out_dev && dev != out_dev && !mutex_trylock(&out_dev->mutex))
if (!mutex_trylock(&dev->mutex)) {
if (!mutex_trylock(&dev->mutex)) {
if (!mutex_trylock(&dev->mutex)) {
if (!mutex_trylock(&dev->mutex)) {
if (!mutex_trylock(&dev->lock))
if (!mutex_trylock(&dev->lock))
if (!mutex_trylock(&dev->lock))
if (!mutex_trylock(&dev->lock))
if (!mutex_trylock(&dev->lock))
if (!mutex_trylock(&ioc->ioctl_cmds.mutex))
mutex_trylock(mutex);
if (!mutex_trylock(&mc13xxx->lock)) {
if (mutex_trylock(&ucr->dev_mutex)) {
if (gts && mutex_trylock(>s->ts_ctxlock)) {
return mutex_trylock(>s->ts_ctxlock);
if (mutex_trylock(®istration->mutex) == 0)
ret = mutex_trylock(&contender->mutex);
ret = mutex_trylock(&contender->mutex);
if (mutex_trylock(&chip->ppu_mutex)) {
if (!mutex_trylock(&uld_mutex)) {
if (!mutex_trylock(&uld_mutex)) {
if (!mutex_trylock(&uld_mutex)) {
if (!mutex_trylock(&uld_mutex)) {
if (!mutex_trylock(&np->adapter->vport_ctrl_lock))
while (!mutex_trylock(&zt->priv->nfp_fl_lock)) {
if (mutex_trylock(&efx->mac_lock)) {
if (mutex_trylock(&efx->mac_lock)) {
if (!mutex_trylock(&efx->rps_mutex))
if (mutex_trylock(&efx->mac_lock)) {
if (!mutex_trylock(&efx->rps_mutex))
if (!mutex_trylock(&wl->assoc_stat_lock)) {
if (mutex_trylock(&phydev->lock)) {
if (!mutex_trylock(&ppp_mutex)) {
if (!mutex_trylock(&tp->control)) {
if (mutex_trylock(&ar->mutex) == 0) {
if (!mutex_trylock(&priv->conf_mutex))
if (!mutex_trylock(&dev->shutdown_lock))
if (!mutex_trylock(&dev->shutdown_lock))
if (mutex_trylock(&queue->send_mutex)) {
empty && mutex_trylock(&queue->send_mutex)) {
if (!mutex_trylock(&zdev->state_lock))
if (mutex_trylock(&dasd_pe_handler_mutex)) {
if (!mutex_trylock(&private->io_mutex))
if (!mutex_trylock(&private->io_mutex))
if (mutex_trylock(&ap_scan_bus_mutex)) {
if (mutex_trylock(&ap_scan_bus_mutex)) {
if (!mutex_trylock(&card->sbp_lock)) {
if (!mutex_trylock(&vhost->passthru_mutex))
if (!mutex_trylock(&mrioc->reset_mutex)) {
if (mutex_trylock(&ioc->hostdiag_unlock_mutex)) {
if (!mutex_trylock(&ioc->ctl_cmds.mutex)) {
mutex_acquired = mutex_trylock(&ctrl_info->scan_mutex);
mutex_acquired = mutex_trylock(&ctrl_info->scan_mutex);
if ((n_tx + n_rx) > SPI_BUFSIZ || !mutex_trylock(&lock)) {
if (!mutex_trylock(&fifo->read_lock))
if (!mutex_trylock(&fifo->write_lock))
if (!mutex_trylock(&list->mtx)) {
if (!mutex_trylock(&list->mtx)) {
if (!mutex_trylock(&ctrlr->mtx)) {
if (!mutex_trylock(&powerclamp_lock))
if (!mutex_trylock(&rt->tb->lock))
if (!mutex_trylock(&rt->tb->lock)) {
if (!mutex_trylock(&rt->tb->lock))
if (!mutex_trylock(&rt->tb->lock)) {
if (!mutex_trylock(&rt->tb->lock))
if (!mutex_trylock(&sw->tb->lock))
if (!mutex_trylock(&sw->tb->lock))
if (!mutex_trylock(&sw->tb->lock))
if (!mutex_trylock(&sw->tb->lock)) {
if (!mutex_trylock(&sw->tb->lock))
if (!mutex_trylock(&sw->tb->lock)) {
if (!mutex_trylock(&sw->tb->lock))
if (!mutex_trylock(&ldata->atomic_read_lock))
if (!mutex_trylock(&tty->atomic_write_lock)) {
? mutex_trylock(mutex) ? 0 : -EAGAIN
if (!mutex_trylock(&epdata->lock))
result = mutex_trylock(&dev->lock);
retval = mutex_trylock(&dev->mutex);
mutex_locked = mutex_trylock(&con->lock) != 0;
} else if (!mutex_trylock(&mres->lock)) {
if (!mutex_trylock(&mvdev->state_mutex)) {
if (!mutex_trylock(&virtvdev->state_mutex)) {
if (!mutex_trylock(&xe_vdev->state_mutex)) {
if (!mutex_trylock(&vq->mutex))
if (!mutex_trylock(&vb->balloon_lock))
if (!mutex_trylock(&w1dev->rx_mutex))
if (!mutex_trylock(&system_transition_mutex))
if (!mutex_trylock(&ctx->ring_lock)) {
if (!mutex_trylock(&head->mutex)) {
if (!mutex_trylock(&fs_info->reclaim_bgs_lock))
if (!mutex_trylock(&fs_info->reclaim_bgs_lock)) {
if (!mutex_trylock(&fs_info->reclaim_bgs_lock))
if (mutex_trylock(&head->mutex))
if (!mutex_trylock(&fs_info->cleaner_mutex))
if (!mutex_trylock(&head->mutex)) {
if (!mutex_trylock(&head->mutex)) {
if (!mutex_trylock(&head->mutex))
if (!mutex_trylock(&head->mutex)) {
if (!mutex_trylock(&dentry->d_sb->s_vfs_rename_mutex))
if (mutex_trylock(&(*msg_ctx)->mux)) {
DBG_BUGON(!mutex_trylock(&pcl->lock));
if (!mutex_trylock(&sbi->umount_mutex)) {
if (mutex_trylock(mutex))
if (!mutex_trylock(&eti->extent_tree_lock))
if (!mutex_trylock(&eti->extent_tree_lock))
if (!mutex_trylock(&nm_i->build_lock))
if (!mutex_trylock(&sbi->umount_mutex)) {
if (!mutex_trylock(&sbi->umount_mutex)) {
if (!mutex_trylock(&sbi->umount_mutex)) {
if (!mutex_trylock(&sbi->umount_mutex)) {
if (!mutex_trylock(&sdp->sd_freeze_mutex))
if (!mutex_trylock(&sdp->sd_freeze_mutex))
} else if (mutex_trylock(&jfs_ip->commit_mutex)) {
if (!mutex_trylock(&ictx->wb_lock)) {
if (!mutex_trylock(&ictx->wb_lock)) {
if (mutex_trylock(&nfsd_mutex)) {
if (mutex_trylock(&nfsd_mutex)) {
return mutex_trylock(&ni->ni_lock);
mlog_bug_on_msg(!mutex_trylock(&oi->ip_io_mutex),
if (!mutex_trylock(&ui->ui_mutex))
if (mutex_trylock(&c->bu_mutex))
if (!mutex_trylock(&c->umount_mutex)) {
if (!mutex_trylock(&c->tnc_mutex)) {
if (!mutex_trylock(&c->umount_mutex)) {
if (!mutex_trylock(&dqp->q_qlock))
if (!mutex_trylock(&mp->m_growlock))
if (!mutex_trylock(&mp->m_growlock))
if (!mutex_trylock(&mp->m_growlock))
return mutex_trylock(&dev->mutex);
extern int mutex_trylock(struct mutex *lock) __cond_acquires(true, lock);
#define mutex_trylock_nest_lock(lock, nest_lock) mutex_trylock(lock)
DEFINE_LOCK_GUARD_1_COND(mutex, _try, mutex_trylock(_T->lock))
return mutex_trylock(&dev->lock);
if (mutex_trylock(&ctx->uring_lock)) {
if (!mutex_trylock(&octx->uring_lock))
if (!mutex_trylock(&tr->mutex)) {
EXPORT_SYMBOL(mutex_trylock);
return mutex_trylock(&ww->base);
EXPORT_SYMBOL(mutex_trylock);
if (!mutex_trylock(&run_lock)) {
if (!mutex_trylock(&system_transition_mutex))
if (!mutex_trylock(&system_transition_mutex))
if (!mutex_trylock(&system_transition_mutex))
if (mutex_trylock(&boost_mutex)) {
mutex_trylock(&rcu_state.exp_mutex))
if (!mutex_trylock(&rcu_state.nocb_mutex))
if (!mutex_trylock(&rcu_state.nocb_mutex)) {
while (!mutex_trylock(&stop_cpus_mutex))
#define ww_mutex_base_trylock(b) mutex_trylock(b)
if (mutex_trylock(&d->mtx)) {
else if (!mutex_trylock(&dmirror->mutex))
else if (!mutex_trylock(&dmirror->mutex))
if (!mutex_trylock(&damon_sysfs_lock))
if (!mutex_trylock(&damon_sysfs_lock))
if (!mutex_trylock(&damon_sysfs_lock))
if (!mutex_trylock(&damon_sysfs_lock))
if (!mutex_trylock(&damon_sysfs_lock))
if (!mutex_trylock(&damon_sysfs_lock))
if (!mutex_trylock(&damon_sysfs_lock))
if (!mutex_trylock(&damon_sysfs_lock))
if (!mutex_trylock(&damon_sysfs_lock))
if (!mutex_trylock(&damon_sysfs_lock))
if (!mutex_trylock(&damon_sysfs_lock))
if (!mutex_trylock(&damon_sysfs_lock))
if (!mutex_trylock(&percpu_charge_mutex))
if (unlikely(!mutex_trylock(&pcpu_drain_mutex))) {
if (!mutex_trylock(&oom_lock)) {
if (!mutex_trylock(&slab_mutex)) {
if (!reclaimable && mutex_trylock(&oom_lock)) {
return mutex_trylock(&rtnl_mutex);
return mutex_trylock(&devlink->lock);
if (mutex_trylock(&ipvs->sync_mutex))
while (!mutex_trylock(&pool->flush_lock)) {
if (!mutex_trylock(&call->user_mutex)) {
if (mutex_trylock(&tls_ctx->tx_lock)) {
ret = mutex_trylock(&vsock->tx_lock);
return mutex_trylock(lock);
if (!mutex_trylock(&mdev_state->state_mutex)) {
if (!mutex_trylock(&tomoyo_gc_mutex))
if (!(mutex_trylock(&runtime->oss.params_lock)))
trylock = nonatomic ? mutex_trylock(&group->mutex) :
if (mutex_trylock(&us122l->mutex)) {
if (mutex_trylock(&ui__lock)) {
return mutex_trylock(mutex);
bool mutex_trylock(struct mutex *mtx) EXCLUSIVE_TRYLOCK_FUNCTION(true, *mtx);