Symbol: rq_lock
drivers/infiniband/hw/bnxt_re/ib_verbs.c
1730
spin_lock_init(&qp->rq_lock);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3048
spin_lock_irqsave(&qp->rq_lock, flags);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3091
spin_unlock_irqrestore(&qp->rq_lock, flags);
drivers/infiniband/hw/bnxt_re/ib_verbs.h
89
spinlock_t rq_lock; /* protect rq */
drivers/infiniband/hw/ionic/ionic_controlpath.c
2175
spin_lock_init(&qp->rq_lock);
drivers/infiniband/hw/ionic/ionic_controlpath.c
2435
spin_lock(&qp->rq_lock);
drivers/infiniband/hw/ionic/ionic_controlpath.c
2441
spin_unlock(&qp->rq_lock);
drivers/infiniband/hw/ionic/ionic_controlpath.c
2507
spin_lock(&qp->rq_lock);
drivers/infiniband/hw/ionic/ionic_controlpath.c
2517
spin_unlock(&qp->rq_lock);
drivers/infiniband/hw/ionic/ionic_datapath.c
1323
spin_lock_irqsave(&qp->rq_lock, irqflags);
drivers/infiniband/hw/ionic/ionic_datapath.c
1341
spin_unlock_irqrestore(&qp->rq_lock, irqflags);
drivers/infiniband/hw/ionic/ionic_datapath.c
1344
spin_unlock_irqrestore(&qp->rq_lock, irqflags);
drivers/infiniband/hw/ionic/ionic_datapath.c
1347
spin_lock(&qp->rq_lock);
drivers/infiniband/hw/ionic/ionic_datapath.c
1366
spin_unlock(&qp->rq_lock);
drivers/infiniband/hw/ionic/ionic_datapath.c
528
spin_lock(&qp->rq_lock);
drivers/infiniband/hw/ionic/ionic_datapath.c
530
spin_unlock(&qp->rq_lock);
drivers/infiniband/hw/ionic/ionic_datapath.c
618
spin_lock(&qp->rq_lock);
drivers/infiniband/hw/ionic/ionic_datapath.c
620
spin_unlock(&qp->rq_lock);
drivers/infiniband/hw/ionic/ionic_ibdev.h
276
spinlock_t rq_lock; /* for posting and polling */
drivers/infiniband/sw/siw/siw.h
449
spinlock_t rq_lock;
drivers/infiniband/sw/siw/siw_verbs.c
1072
spin_lock_irqsave(&qp->rq_lock, flags);
drivers/infiniband/sw/siw/siw_verbs.c
1100
spin_unlock_irqrestore(&qp->rq_lock, flags);
drivers/infiniband/sw/siw/siw_verbs.c
369
spin_lock_init(&qp->rq_lock);
drivers/net/ethernet/intel/ice/ice_controlq.c
1190
mutex_lock(&cq->rq_lock);
drivers/net/ethernet/intel/ice/ice_controlq.c
1258
mutex_unlock(&cq->rq_lock);
drivers/net/ethernet/intel/ice/ice_controlq.c
540
mutex_lock(&cq->rq_lock);
drivers/net/ethernet/intel/ice/ice_controlq.c
562
mutex_unlock(&cq->rq_lock);
drivers/net/ethernet/intel/ice/ice_controlq.c
789
mutex_init(&cq->rq_lock);
drivers/net/ethernet/intel/ice/ice_controlq.c
827
mutex_destroy(&cq->rq_lock);
drivers/net/ethernet/intel/ice/ice_controlq.h
101
struct mutex rq_lock; /* Receive queue lock */
io_uring/zcrx.c
1152
scoped_guard(spinlock_bh, &zcrx->rq_lock) {
io_uring/zcrx.c
522
spin_lock_init(&ifq->rq_lock);
io_uring/zcrx.c
955
guard(spinlock_bh)(&ifq->rq_lock);
io_uring/zcrx.h
47
spinlock_t rq_lock ____cacheline_aligned_in_smp;
kernel/sched/core.c
2460
rq_lock(rq, rf);
kernel/sched/core.c
2541
rq_lock(rq, &rf);
kernel/sched/core.c
3916
rq_lock(rq, &rf);
kernel/sched/core.c
5561
rq_lock(rq, &rf);
kernel/sched/core.c
6814
rq_lock(rq, &rf);
kernel/sched/core.c
8100
rq_lock(rq, &rf);
kernel/sched/core.c
892
rq_lock(rq, &rf);
kernel/sched/core.c
916
rq_lock(rq, &rf);
kernel/sched/deadline.c
1155
scoped_guard (rq_lock, rq) {
kernel/sched/deadline.c
1974
rq_lock(rq, &rf);
kernel/sched/deadline.c
2469
rq_lock(rq, &rf);
kernel/sched/fair.c
6101
rq_lock(rq, &rf);
kernel/sched/fair.c
9803
rq_lock(rq, &rf);
kernel/sched/fair.c
9819
rq_lock(env->dst_rq, &rf);
kernel/sched/rt.c
816
rq_lock(rq, &rf);
kernel/sched/sched.h
1974
DEFINE_LOCK_GUARD_1(rq_lock, struct rq,
kernel/sched/sched.h
1975
rq_lock(_T->lock, &_T->rf),
kernel/sched/sched.h
1979
DECLARE_LOCK_GUARD_1_ATTRS(rq_lock, __acquires(__rq_lockp(_T)), __releases(__rq_lockp(*(struct rq **)_T)));
kernel/sched/sched.h
1980
#define class_rq_lock_constructor(_T) WITH_LOCK_GUARD_1_ATTRS(rq_lock, _T)
kernel/sched/sched.h
2005
rq_lock(rq, rf);