mtx_trylock
int mtx_trylock(mtx_t *__mtx)
#define PMAP_TRYLOCK(pmap) mtx_trylock(&(pmap)->pm_mtx)
#define PMAP_TRYLOCK(pmap) mtx_trylock(&(pmap)->pm_mtx)
#define PMAP_TRYLOCK(pmap) mtx_trylock(&(pmap)->pm_mtx)
__ret = mtx_trylock(_l); \
#define ARCMSR_LOCK_TRY(l) mtx_trylock(l)
mtx_trylock(&(_sc)->sc_tx_mtx))
#define BXE_CORE_TRYLOCK(sc) mtx_trylock(&sc->core_mtx)
#define BXE_FP_TX_TRYLOCK(fp) mtx_trylock(&fp->tx_mtx)
if (mtx_trylock(lock)) {
#define TXQ_TRYLOCK(qs) mtx_trylock(&(qs)->lock)
#define FL_TRYLOCK(fl) mtx_trylock(&(fl)->fl_lock)
#define EQ_TRYLOCK(eq) mtx_trylock(&(eq)->eq_lock)
#define ENA_RING_MTX_TRYLOCK(_ring) mtx_trylock(&(_ring)->ring_mtx)
mtx_trylock(&(_sc)->sc_mtx)
mtx_trylock(&(_sc)->sc_mtx)
mtx_trylock(&(_sc)->sc_mtx)
mtx_trylock(&(_sc)->sc_mtx)
mtx_trylock(&(_sc)->sc_mtx)
mtx_trylock(&(_sc)->sc_mtx)
#define RTL_TRYLOCK(_sc) mtx_trylock(&(_sc)->sc_mtx)
mtx_trylock(&(_sc)->sc_mtx)
#define GVE_RING_TRYLOCK(ring) mtx_trylock(&(ring)->ring_mtx)
if (mtx_trylock(&txr->hn_tx_lock)) {
if (mtx_trylock(&txr->hn_tx_lock)) {
if (mtx_trylock(&txr->hn_tx_lock)) {
if (mtx_trylock(&txr->hn_tx_lock)) {
#define IAVF_VC_TRYLOCK(_sc) mtx_trylock(&(_sc)->vc_mtx)
if (!mtx_trylock(&ioat->cleanup_lock))
if (!mtx_trylock(&iq->iq_flush_running_lock))
if (mtx_trylock(&iq->enq_lock)) {
if (mtx_trylock(&iq->enq_lock)) {
if (mtx_trylock(&iq->enq_lock)) {
if (is_drbr_empty && mtx_trylock(&txq->txq_mtx)) {
if (mtx_trylock(&tx->mtx)) {
return (mtx_trylock(&sc->tx_ring[q].ring_mtx));
#define PTNET_Q_TRYLOCK(_pq) mtx_trylock(&(_pq)->lock)
if (mtx_trylock(&q->tx_lock)) {
#define TRY_LOCK(lock) mtx_trylock(&(lock)->mutex)
int rc = mtx_trylock(&(lock)->lock);
if (mtx_trylock(&fp->tx_mtx)) {
if (mtx_trylock(&fp->tx_mtx)) {
if(mtx_trylock(&fp->tx_mtx)) {
if (!mtx_trylock(&ha->tx_lock)) {
if (!mtx_trylock(&ha->rxj_lock))
if (!mtx_trylock(&ha->rx_lock))
if (!mtx_trylock(&ha->tx_lock)) {
if (!mtx_trylock(&ha->tx_lock)) {
if (mtx_trylock(&random_reseed_mtx) == thrd_busy)
mtx_trylock(&(_txq)->lock)
#define CHN_TRYLOCK(c) mtx_trylock(&(c)->lock)
#define PCM_SG_TRYLOCK() mtx_trylock(&snd_pcm_syncgroups_mtx)
#define PCM_TRYLOCK(d) mtx_trylock(&(d)->lock)
sp->kbd_locked = !kdb_active && mtx_trylock(&Giant);
#define VTNET_TXQ_TRYLOCK(_txq) mtx_trylock(&(_txq)->vtntx_mtx)
#define NICVF_TX_TRYLOCK(sq) mtx_trylock(&(sq)->mtx)
#define XN_TX_TRYLOCK(_q) mtx_trylock(&(_q)->lock)
#define PMAP_TRYLOCK(pmap) mtx_trylock(&(pmap)->pm_mtx)
return (mtx_trylock((struct mtx *)lock));
if (!mtx_trylock(&rm->rm_lock_mtx))
#define UNP_PCB_TRYLOCK(unp) mtx_trylock(&(unp)->unp_mtx)
if (!mtx_trylock(vlp1))
if (!mtx_trylock(vlp2)) {
if (!mtx_trylock(&nl->nl_evict_lock)) {
if (!mtx_trylock(vlp1))
if (mtx_trylock(vlp))
if (!mtx_trylock(&vnode_list_mtx)) {
if (mtx_trylock(&hsh->mtx) == 0)
if (mtx_trylock(&hsh->mtx) == 0)
if (mtx_trylock(&hsh->mtx) == 0) {
#define IPQ_TRYLOCK(i) mtx_trylock(&V_ipq[i].lock)
mtx_trylock(&(_tcb)->tcb_mtx)
#define HPTS_TRYLOCK(hpts) mtx_trylock(&(hpts)->p_mtx)
#define IP6QB_TRYLOCK(_b) mtx_trylock(&V_ip6qb[(_b)].lock)
#define ND6_ONLINK_TRYLOCK() mtx_trylock(&V_nd6_onlink_mtx)
#define PMAP_TRYLOCK(pmap) mtx_trylock(&(pmap)->pm_mtx)
#define PMAP_TRYLOCK(pmap) mtx_trylock(&(pmap)->pm_mtx)
#define MNT_ITRYLOCK(mp) mtx_trylock(&(mp)->mnt_mtx)
#define PROC_TRYLOCK(p) mtx_trylock(&(p)->p_mtx)
#define SIGIO_TRYLOCK() mtx_trylock(&sigio_lock)
#define SOLISTEN_TRYLOCK(sol) mtx_trylock(&(sol)->so_lock)
#define VI_TRYLOCK(vp) mtx_trylock(&(vp)->v_interlock)
#define vm_pagequeue_trylock(pq) mtx_trylock(&(pq)->pq_mutex)
mtx_trylock(vm_domain_free_lockptr((d)))
#define vm_reserv_trylock(rv) mtx_trylock(vm_reserv_lockptr(rv))