Symbol: rq_depth
block/blk-rq-qos.c
110
bool rq_depth_calc_max_depth(struct rq_depth *rqd)
block/blk-rq-qos.c
158
bool rq_depth_scale_up(struct rq_depth *rqd)
block/blk-rq-qos.c
177
bool rq_depth_scale_down(struct rq_depth *rqd, bool hard_throttle)
block/blk-rq-qos.h
100
bool rq_depth_scale_down(struct rq_depth *rqd, bool hard_throttle);
block/blk-rq-qos.h
101
bool rq_depth_calc_max_depth(struct rq_depth *rqd);
block/blk-rq-qos.h
99
bool rq_depth_scale_up(struct rq_depth *rqd);
block/blk-wbt.c
310
struct rq_depth *rqd = &rwb->rq_depth;
block/blk-wbt.c
363
struct rq_depth *rqd = &rwb->rq_depth;
block/blk-wbt.c
373
} else if (rwb->rq_depth.max_depth <= 2) {
block/blk-wbt.c
374
rwb->wb_normal = rwb->rq_depth.max_depth;
block/blk-wbt.c
377
rwb->wb_normal = (rwb->rq_depth.max_depth + 1) / 2;
block/blk-wbt.c
378
rwb->wb_background = (rwb->rq_depth.max_depth + 3) / 4;
block/blk-wbt.c
384
if (!rq_depth_scale_up(&rwb->rq_depth))
block/blk-wbt.c
394
if (!rq_depth_scale_down(&rwb->rq_depth, hard_throttle))
block/blk-wbt.c
403
struct rq_depth *rqd = &rwb->rq_depth;
block/blk-wbt.c
428
struct rq_depth *rqd = &rwb->rq_depth;
block/blk-wbt.c
485
struct rq_depth *rqd = &rwb->rq_depth;
block/blk-wbt.c
553
limit = rwb->rq_depth.max_depth;
block/blk-wbt.c
811
RQWB(rqos)->rq_depth.queue_depth = blk_queue_depth(rqos->disk->queue);
block/blk-wbt.c
93
struct rq_depth rq_depth;
block/blk-wbt.c
956
rwb->rq_depth.default_depth = RWB_DEF_DEPTH;
block/blk-wbt.c
958
rwb->rq_depth.queue_depth = blk_queue_depth(q);
drivers/infiniband/hw/efa/efa_com_cmd.c
32
params->rq_depth;
drivers/infiniband/hw/efa/efa_com_cmd.h
26
u32 rq_depth;
drivers/infiniband/hw/efa/efa_verbs.c
765
create_qp_params.rq_depth = init_attr->cap.max_recv_wr;
drivers/infiniband/hw/hns/hns_roce_hw_v2.h
1215
__le16 rq_depth;
drivers/infiniband/hw/irdma/uk.c
1677
u32 *rq_depth, u8 *rq_shift)
drivers/infiniband/hw/irdma/uk.c
1690
*rq_shift, rq_depth);
drivers/infiniband/hw/irdma/user.h
448
u32 *rq_depth, u8 *rq_shift);
drivers/infiniband/hw/irdma/user.h
560
u32 rq_depth;
drivers/infiniband/hw/irdma/verbs.c
667
ret = irdma_uk_calc_depth_shift_rq(ukinfo, &ukinfo->rq_depth,
drivers/infiniband/hw/irdma/verbs.c
675
(ukinfo->rq_depth - IRDMA_RQ_RSVD) >> ukinfo->rq_shift;
drivers/infiniband/hw/irdma/verbs.c
677
ukinfo->rq_size = ukinfo->rq_depth >> ukinfo->rq_shift;
drivers/infiniband/hw/irdma/verbs.c
707
status = irdma_uk_calc_depth_shift_rq(ukinfo, &ukinfo->rq_depth,
drivers/infiniband/hw/irdma/verbs.c
718
kzalloc_objs(*iwqp->kqp.rq_wrid_mem, ukinfo->rq_depth);
drivers/infiniband/hw/irdma/verbs.c
729
size = (ukinfo->sq_depth + ukinfo->rq_depth) * IRDMA_QP_WQE_MIN_SIZE;
drivers/infiniband/hw/irdma/verbs.c
747
ukinfo->shadow_area = ukinfo->rq[ukinfo->rq_depth].elem;
drivers/infiniband/hw/irdma/verbs.c
749
info->rq_pa + (ukinfo->rq_depth * IRDMA_QP_WQE_MIN_SIZE);
drivers/infiniband/hw/irdma/verbs.c
751
ukinfo->rq_size = ukinfo->rq_depth >> ukinfo->rq_shift;
drivers/infiniband/hw/irdma/verbs.c
755
iwqp->max_recv_wr = (ukinfo->rq_depth - IRDMA_RQ_RSVD) >> ukinfo->rq_shift;
drivers/net/ethernet/fungible/funcore/fun_dev.c
232
.rq_depth = areq->rq_depth,
drivers/net/ethernet/fungible/funcore/fun_dev.c
280
if (areq->rq_depth) {
drivers/net/ethernet/fungible/funcore/fun_dev.c
579
if (cq_count < 2 || sq_count < 2 + !!fdev->admin_q->rq_depth)
drivers/net/ethernet/fungible/funcore/fun_dev.h
94
u16 rq_depth;
drivers/net/ethernet/fungible/funcore/fun_queue.c
175
for (i = 0; i < funq->rq_depth; i++) {
drivers/net/ethernet/fungible/funcore/fun_queue.c
192
for (i = 0; i < funq->rq_depth; i++) {
drivers/net/ethernet/fungible/funcore/fun_queue.c
209
funq->rq_tail = funq->rq_depth - 1;
drivers/net/ethernet/fungible/funcore/fun_queue.c
222
if (++funq->rq_buf_idx == funq->rq_depth)
drivers/net/ethernet/fungible/funcore/fun_queue.c
344
funq->rq_depth;
drivers/net/ethernet/fungible/funcore/fun_queue.c
375
funq->rqes = fun_alloc_ring_mem(funq->fdev->dev, funq->rq_depth,
drivers/net/ethernet/fungible/funcore/fun_queue.c
396
fun_free_ring_mem(dev, funq->rq_depth, sizeof(*funq->rqes),
drivers/net/ethernet/fungible/funcore/fun_queue.c
419
if (req->rq_depth) {
drivers/net/ethernet/fungible/funcore/fun_queue.c
458
if (req->rq_depth) {
drivers/net/ethernet/fungible/funcore/fun_queue.c
460
funq->rq_depth = req->rq_depth;
drivers/net/ethernet/fungible/funcore/fun_queue.c
492
funq->rq_depth, funq->rq_dma_addr, 0, 0,
drivers/net/ethernet/fungible/funcore/fun_queue.h
124
u32 rq_depth;
drivers/net/ethernet/fungible/funcore/fun_queue.h
47
u32 rq_depth;
drivers/net/ethernet/fungible/funeth/funeth.h
116
unsigned int rq_depth;
drivers/net/ethernet/fungible/funeth/funeth.h
76
unsigned int rq_depth;
drivers/net/ethernet/fungible/funeth/funeth_ethtool.c
566
ring->rx_pending = fp->rq_depth;
drivers/net/ethernet/fungible/funeth/funeth_ethtool.c
594
fp->rq_depth == ring->rx_pending)
drivers/net/ethernet/fungible/funeth/funeth_ethtool.c
600
.rq_depth = ring->rx_pending,
drivers/net/ethernet/fungible/funeth/funeth_ethtool.c
610
fp->rq_depth = ring->rx_pending;
drivers/net/ethernet/fungible/funeth/funeth_ethtool.c
611
fp->cq_depth = 2 * fp->rq_depth;
drivers/net/ethernet/fungible/funeth/funeth_main.c
1631
.rq_depth = fp->rq_depth,
drivers/net/ethernet/fungible/funeth/funeth_main.c
1771
fp->rq_depth = min_t(unsigned int, RQ_DEPTH, fdev->q_depth);
drivers/net/ethernet/fungible/funeth/funeth_main.c
1991
.rq_depth = ADMIN_RQ_DEPTH,
drivers/net/ethernet/fungible/funeth/funeth_main.c
509
qset->rq_depth, qset->rxq_start, qset->state);
drivers/net/ethernet/fungible/funeth/funeth_main.c
842
.rq_depth = fp->rq_depth,
drivers/net/ethernet/huawei/hinic/hinic_dev.h
100
u16 rq_depth;
drivers/net/ethernet/huawei/hinic/hinic_ethtool.c
557
ring->rx_pending = nic_dev->rq_depth;
drivers/net/ethernet/huawei/hinic/hinic_ethtool.c
600
new_rq_depth == nic_dev->rq_depth)
drivers/net/ethernet/huawei/hinic/hinic_ethtool.c
605
nic_dev->sq_depth, nic_dev->rq_depth,
drivers/net/ethernet/huawei/hinic/hinic_ethtool.c
609
nic_dev->rq_depth = new_rq_depth;
drivers/net/ethernet/huawei/hinic/hinic_hw_dev.c
271
unsigned int rq_depth)
drivers/net/ethernet/huawei/hinic/hinic_hw_dev.c
285
hw_ioctxt.rq_depth = ilog2(rq_depth);
drivers/net/ethernet/huawei/hinic/hinic_hw_dev.c
438
int hinic_hwdev_ifup(struct hinic_hwdev *hwdev, u16 sq_depth, u16 rq_depth)
drivers/net/ethernet/huawei/hinic/hinic_hw_dev.c
462
func_to_io->rq_depth = rq_depth;
drivers/net/ethernet/huawei/hinic/hinic_hw_dev.c
489
err = set_hw_ioctxt(hwdev, sq_depth, rq_depth);
drivers/net/ethernet/huawei/hinic/hinic_hw_dev.h
286
u16 rq_depth;
drivers/net/ethernet/huawei/hinic/hinic_hw_dev.h
627
int hinic_hwdev_ifup(struct hinic_hwdev *hwdev, u16 sq_depth, u16 rq_depth);
drivers/net/ethernet/huawei/hinic/hinic_hw_io.c
293
func_to_io->rq_depth, HINIC_RQ_WQE_SIZE);
drivers/net/ethernet/huawei/hinic/hinic_hw_io.h
74
u16 rq_depth;
drivers/net/ethernet/huawei/hinic/hinic_hw_mbox.c
1289
((hw_ctxt)->rq_depth >= HINIC_QUEUE_MIN_DEPTH && \
drivers/net/ethernet/huawei/hinic/hinic_hw_mbox.c
1290
(hw_ctxt)->rq_depth <= HINIC_QUEUE_MAX_DEPTH && \
drivers/net/ethernet/huawei/hinic/hinic_hw_mbox.c
1300
if (!hw_ctxt->rq_depth && !hw_ctxt->sq_depth &&
drivers/net/ethernet/huawei/hinic/hinic_main.c
1215
nic_dev->rq_depth = HINIC_RQ_DEPTH;
drivers/net/ethernet/huawei/hinic/hinic_main.c
425
nic_dev->rq_depth);
drivers/net/ethernet/huawei/hinic/hinic_port.c
491
rq_num.rq_depth = ilog2(nic_dev->rq_depth);
drivers/net/ethernet/huawei/hinic/hinic_port.h
317
u32 rq_depth;
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
482
int hinic3_set_root_ctxt(struct hinic3_hwdev *hwdev, u32 rq_depth, u32 sq_depth,
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.c
501
root_ctxt.rq_depth = ilog2(rq_depth);
drivers/net/ethernet/huawei/hinic3/hinic3_hw_comm.h
49
int hinic3_set_root_ctxt(struct hinic3_hwdev *hwdev, u32 rq_depth, u32 sq_depth,
drivers/net/ethernet/huawei/hinic3/hinic3_hw_intf.h
237
u16 rq_depth;
drivers/net/ethernet/huawei/hinic3/hinic3_main.c
182
nic_dev->q_params.rq_depth = HINIC3_RQ_DEPTH;
drivers/net/ethernet/huawei/hinic3/hinic3_netdev_ops.c
155
q_params->rq_depth, q_params->rxqs_res);
drivers/net/ethernet/huawei/hinic3/hinic3_netdev_ops.c
182
hinic3_free_rxqs_res(netdev, q_params->num_qps, q_params->rq_depth,
drivers/net/ethernet/huawei/hinic3/hinic3_netdev_ops.c
210
q_params->rq_depth, q_params->rxqs_res);
drivers/net/ethernet/huawei/hinic3/hinic3_netdev_ops.c
263
qp_params->rq_depth = trxq_params->rq_depth;
drivers/net/ethernet/huawei/hinic3/hinic3_nic_dev.h
82
u32 rq_depth;
drivers/net/ethernet/huawei/hinic3/hinic3_nic_io.c
341
u16 q_id, u32 rq_depth, u16 rq_msix_idx)
drivers/net/ethernet/huawei/hinic3/hinic3_nic_io.c
348
err = hinic3_wq_create(hwdev, &rq->wq, rq_depth,
drivers/net/ethernet/huawei/hinic3/hinic3_nic_io.c
363
u32 rq_depth, u16 qp_msix_idx)
drivers/net/ethernet/huawei/hinic3/hinic3_nic_io.c
374
err = hinic3_create_rq(hwdev, rq, q_id, rq_depth, qp_msix_idx);
drivers/net/ethernet/huawei/hinic3/hinic3_nic_io.c
425
qp_params->sq_depth, qp_params->rq_depth,
drivers/net/ethernet/huawei/hinic3/hinic3_nic_io.c
836
u32 rq_depth;
drivers/net/ethernet/huawei/hinic3/hinic3_nic_io.c
853
rq_depth = nic_io->rq[0].wq.q_depth << HINIC3_NORMAL_RQ_WQE;
drivers/net/ethernet/huawei/hinic3/hinic3_nic_io.c
855
err = hinic3_set_root_ctxt(hwdev, rq_depth, nic_io->sq[0].wq.q_depth,
drivers/net/ethernet/huawei/hinic3/hinic3_nic_io.h
100
u32 rq_depth;
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
179
u32 rq_depth, u16 buf_len)
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
181
u32 free_wqebbs = rq_depth - 1;
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
411
u32 rq_depth, struct hinic3_dyna_rxq_res *rxqs_res)
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
413
u64 cqe_mem_size = sizeof(struct hinic3_rq_cqe) * rq_depth;
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
422
rqres->rx_info = kzalloc_objs(*rqres->rx_info, rq_depth);
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
436
pp_params.pool_size = rq_depth * nic_dev->rx_buf_len /
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
449
pkt_idx = hinic3_alloc_rx_buffers(rqres, rq_depth,
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
470
hinic3_free_rxqs_res(netdev, idx, rq_depth, rxqs_res);
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
476
u32 rq_depth, struct hinic3_dyna_rxq_res *rxqs_res)
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
478
u64 cqe_mem_size = sizeof(struct hinic3_rq_cqe) * rq_depth;
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
486
hinic3_free_rx_buffers(rqres, rq_depth);
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
496
u32 rq_depth, struct hinic3_dyna_rxq_res *rxqs_res)
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
514
rxq->q_depth = rq_depth;
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
80
rxq->q_depth = nic_dev->q_params.rq_depth;
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
81
rxq->q_mask = nic_dev->q_params.rq_depth - 1;
drivers/net/ethernet/huawei/hinic3/hinic3_rx.h
118
u32 rq_depth, struct hinic3_dyna_rxq_res *rxqs_res);
drivers/net/ethernet/huawei/hinic3/hinic3_rx.h
120
u32 rq_depth, struct hinic3_dyna_rxq_res *rxqs_res);
drivers/net/ethernet/huawei/hinic3/hinic3_rx.h
122
u32 rq_depth, struct hinic3_dyna_rxq_res *rxqs_res);
fs/9p/v9fs.c
380
rdma_opts->rq_depth = result.uint_32;
fs/9p/vfs_super.c
344
ctx->rdma_opts.rq_depth = P9_RDMA_RQ_DEPTH;
include/net/9p/client.h
177
int rq_depth;
net/9p/trans_rdma.c
117
if (rdma->rq_depth != P9_RDMA_RQ_DEPTH)
net/9p/trans_rdma.c
118
seq_printf(m, ",rq=%u", rdma->rq_depth);
net/9p/trans_rdma.c
470
rdma->rq_depth = opts->rq_depth;
net/9p/trans_rdma.c
475
sema_init(&rdma->rq_sem, rdma->rq_depth);
net/9p/trans_rdma.c
585
opts.sq_depth + opts.rq_depth + 1,
net/9p/trans_rdma.c
600
qp_attr.cap.max_recv_wr = opts.rq_depth;
net/9p/trans_rdma.c
81
int rq_depth;
net/sunrpc/xprtrdma/svc_rdma_transport.c
409
unsigned int ctxts, rq_depth, maxpayload;
net/sunrpc/xprtrdma/svc_rdma_transport.c
456
rq_depth = newxprt->sc_max_requests + newxprt->sc_max_bc_requests +
net/sunrpc/xprtrdma/svc_rdma_transport.c
458
if (rq_depth > dev->attrs.max_qp_wr) {
net/sunrpc/xprtrdma/svc_rdma_transport.c
459
rq_depth = dev->attrs.max_qp_wr;
net/sunrpc/xprtrdma/svc_rdma_transport.c
461
newxprt->sc_max_requests = rq_depth - 2;
net/sunrpc/xprtrdma/svc_rdma_transport.c
476
newxprt->sc_sq_depth = rq_depth +
net/sunrpc/xprtrdma/svc_rdma_transport.c
492
ib_alloc_cq_any(dev, newxprt, rq_depth, IB_POLL_WORKQUEUE);
net/sunrpc/xprtrdma/svc_rdma_transport.c
502
qp_attr.cap.max_recv_wr = rq_depth;
net/sunrpc/xprtrdma/svc_rdma_transport.c
514
newxprt->sc_sq_depth, rq_depth);