sx_xlocked
int sx_xlocked(struct sx *);
#define rw_write_held(lock) sx_xlocked(lock)
#define rw_iswriter(lock) sx_xlocked(lock)
return (sx_xlocked(&m->sx));
sx_xlocked(&(_m)->base.sx)
if (!sx_xlocked(&dev->mode_config.mutex))
#define E6000SW_LOCKED(_sc) sx_xlocked(&(_sc)->sx)
MPASS(in_epoch(global_epoch_preempt) || sx_xlocked(&pmc_sx));
MPASS(in_epoch(global_epoch_preempt) || sx_xlocked(&pmc_sx));
is_locked = sx_xlocked(sc->iflib_ctx_lock);
allocated = sx_xlocked(&sc->call_lock) != 0;
allocated = sx_xlocked(&sc->call_lock) != 0;
#define PRIV_LOCKED(priv) sx_xlocked(&(priv)->state_lock)
#define CONG_LOCKED(dev) sx_xlocked(&(dev)->congestion.lock)
sx_xlocked(&pf->lock), ("%s: Called without pf->lock", __func__));
if (sx_xlocked(&udev->enum_sx))
if (sx_xlocked(&udev->enum_sx))
return (sx_xlocked(&udev->enum_sx));
if (sx_xlocked(&udev->ctrl_sx))
if (sx_xlocked(&topology_lock)) {
#define g_topology_locked() sx_xlocked(&topology_lock)
sx_xlocked(&sc->sc_lock) ? M_WAITOK : M_NOWAIT);
if (!sx_xlocked(FILEDESC_LOCK(fdp))) {
sx_xlocked(&allprison_lock));
locked = sx_xlocked(&kld_sx);
xlocked = sx_xlocked(&proctree_lock);
return (sx_xlocked(&rm->rm_lock_sx));
if (sx_xlocked(sx)) {
int locked = sx_xlocked(&mqfs->mi_lock);
MPASS(in_epoch(net_epoch_preempt) || sx_xlocked(&(_sc)->sc_sx))
sx_xlocked(&carp_sx), ("cif_vrs not locked")); \
if (sx_xlocked(sx)) \
return (sx_xlocked(&map->lock));