rq_lock
spin_lock_init(&qp->rq_lock);
spin_lock_irqsave(&qp->rq_lock, flags);
spin_unlock_irqrestore(&qp->rq_lock, flags);
spinlock_t rq_lock; /* protect rq */
spin_lock_init(&qp->rq_lock);
spin_lock(&qp->rq_lock);
spin_unlock(&qp->rq_lock);
spin_lock(&qp->rq_lock);
spin_unlock(&qp->rq_lock);
spin_lock_irqsave(&qp->rq_lock, irqflags);
spin_unlock_irqrestore(&qp->rq_lock, irqflags);
spin_unlock_irqrestore(&qp->rq_lock, irqflags);
spin_lock(&qp->rq_lock);
spin_unlock(&qp->rq_lock);
spin_lock(&qp->rq_lock);
spin_unlock(&qp->rq_lock);
spin_lock(&qp->rq_lock);
spin_unlock(&qp->rq_lock);
spinlock_t rq_lock; /* for posting and polling */
spinlock_t rq_lock;
spin_lock_irqsave(&qp->rq_lock, flags);
spin_unlock_irqrestore(&qp->rq_lock, flags);
spin_lock_init(&qp->rq_lock);
mutex_lock(&cq->rq_lock);
mutex_unlock(&cq->rq_lock);
mutex_lock(&cq->rq_lock);
mutex_unlock(&cq->rq_lock);
mutex_init(&cq->rq_lock);
mutex_destroy(&cq->rq_lock);
struct mutex rq_lock; /* Receive queue lock */
scoped_guard(spinlock_bh, &zcrx->rq_lock) {
spin_lock_init(&ifq->rq_lock);
guard(spinlock_bh)(&ifq->rq_lock);
spinlock_t rq_lock ____cacheline_aligned_in_smp;
rq_lock(rq, rf);
rq_lock(rq, &rf);
rq_lock(rq, &rf);
rq_lock(rq, &rf);
rq_lock(rq, &rf);
rq_lock(rq, &rf);
rq_lock(rq, &rf);
rq_lock(rq, &rf);
scoped_guard (rq_lock, rq) {
rq_lock(rq, &rf);
rq_lock(rq, &rf);
rq_lock(rq, &rf);
rq_lock(rq, &rf);
rq_lock(env->dst_rq, &rf);
rq_lock(rq, &rf);
DEFINE_LOCK_GUARD_1(rq_lock, struct rq,
rq_lock(_T->lock, &_T->rf),
DECLARE_LOCK_GUARD_1_ATTRS(rq_lock, __acquires(__rq_lockp(_T)), __releases(__rq_lockp(*(struct rq **)_T)));
#define class_rq_lock_constructor(_T) WITH_LOCK_GUARD_1_ATTRS(rq_lock, _T)
rq_lock(rq, rf);