Symbol: cq_idx
drivers/accel/habanalabs/common/device.c
2246
hdev->completion_queue[i].cq_idx = i;
drivers/accel/habanalabs/common/habanalabs.h
1121
u32 cq_idx;
drivers/accel/habanalabs/common/habanalabs.h
1727
u32 (*get_queue_id_for_cq)(struct hl_device *hdev, u32 cq_idx);
drivers/accel/habanalabs/common/irq.c
86
queue_work(hdev->cq_wq[cq->cq_idx], &job->finish_work);
drivers/accel/habanalabs/gaudi/gaudi.c
8437
static u32 gaudi_get_queue_id_for_cq(struct hl_device *hdev, u32 cq_idx)
drivers/accel/habanalabs/gaudi/gaudi.c
8439
return gaudi_cq_assignment[cq_idx];
drivers/accel/habanalabs/gaudi2/gaudi2.c
11319
static u32 gaudi2_get_queue_id_for_cq(struct hl_device *hdev, u32 cq_idx)
drivers/accel/habanalabs/goya/goya.c
5248
u32 goya_get_queue_id_for_cq(struct hl_device *hdev, u32 cq_idx)
drivers/accel/habanalabs/goya/goya.c
5250
return cq_idx;
drivers/accel/habanalabs/goya/goyaP.h
247
u32 goya_get_queue_id_for_cq(struct hl_device *hdev, u32 cq_idx);
drivers/infiniband/hw/efa/efa.h
107
u16 cq_idx;
drivers/infiniband/hw/efa/efa_admin_cmds_defs.h
560
u16 cq_idx;
drivers/infiniband/hw/efa/efa_admin_cmds_defs.h
578
u16 cq_idx;
drivers/infiniband/hw/efa/efa_com_cmd.c
194
result->cq_idx = cmd_completion.cq_idx;
drivers/infiniband/hw/efa/efa_com_cmd.c
211
destroy_cmd.cq_idx = params->cq_idx;
drivers/infiniband/hw/efa/efa_com_cmd.c
223
params->cq_idx, err);
drivers/infiniband/hw/efa/efa_com_cmd.h
86
u16 cq_idx;
drivers/infiniband/hw/efa/efa_com_cmd.h
94
u16 cq_idx;
drivers/infiniband/hw/efa/efa_verbs.c
1063
static int efa_destroy_cq_idx(struct efa_dev *dev, int cq_idx)
drivers/infiniband/hw/efa/efa_verbs.c
1065
struct efa_com_destroy_cq_params params = { .cq_idx = cq_idx };
drivers/infiniband/hw/efa/efa_verbs.c
1083
cq->cq_idx, cq->cpu_addr, cq->size, &cq->dma_addr);
drivers/infiniband/hw/efa/efa_verbs.c
1085
efa_destroy_cq_idx(dev, cq->cq_idx);
drivers/infiniband/hw/efa/efa_verbs.c
1088
xa_erase(&dev->cqs_xa, cq->cq_idx);
drivers/infiniband/hw/efa/efa_verbs.c
1257
resp.cq_idx = result.cq_idx;
drivers/infiniband/hw/efa/efa_verbs.c
1258
cq->cq_idx = result.cq_idx;
drivers/infiniband/hw/efa/efa_verbs.c
1267
cq->cq_idx);
drivers/infiniband/hw/efa/efa_verbs.c
1272
err = xa_err(xa_store(&dev->cqs_xa, cq->cq_idx, cq, GFP_KERNEL));
drivers/infiniband/hw/efa/efa_verbs.c
1275
cq->cq_idx);
drivers/infiniband/hw/efa/efa_verbs.c
1291
cq->cq_idx, result.actual_depth, &cq->dma_addr, cq->cpu_addr);
drivers/infiniband/hw/efa/efa_verbs.c
1297
xa_erase(&dev->cqs_xa, cq->cq_idx);
drivers/infiniband/hw/efa/efa_verbs.c
1301
efa_destroy_cq_idx(dev, cq->cq_idx);
drivers/infiniband/hw/efa/efa_verbs.c
760
create_qp_params.send_cq_idx = to_ecq(init_attr->send_cq)->cq_idx;
drivers/infiniband/hw/efa/efa_verbs.c
761
create_qp_params.recv_cq_idx = to_ecq(init_attr->recv_cq)->cq_idx;
drivers/infiniband/hw/irdma/ctrl.c
4460
u32 *cq_idx)
drivers/infiniband/hw/irdma/ctrl.c
4474
*cq_idx = temp;
drivers/infiniband/hw/irdma/ctrl.c
4475
if (*cq_idx >= dev->hmc_info->hmc_obj[IRDMA_HMC_IW_CQ].cnt)
drivers/infiniband/hw/irdma/ctrl.c
4476
*cq_idx = IRDMA_INVALID_CQ_IDX;
drivers/infiniband/hw/irdma/ctrl.c
4481
} while (*cq_idx == IRDMA_INVALID_CQ_IDX);
drivers/infiniband/hw/irdma/ctrl.c
4500
u32 cq_idx;
drivers/infiniband/hw/irdma/ctrl.c
4515
cq_idx = temp;
drivers/infiniband/hw/irdma/ctrl.c
4516
if (cq_idx == cq->cq_uk.cq_id)
drivers/infiniband/hw/irdma/hw.c
110
struct irdma_sc_dev *dev, u32 cq_idx)
drivers/infiniband/hw/irdma/hw.c
113
struct irdma_cq *icq = READ_ONCE(rf->cq_table[cq_idx]);
drivers/infiniband/hw/irdma/hw.c
119
cq_idx);
drivers/infiniband/hw/irdma/hw.c
144
struct irdma_sc_dev *dev, u32 cq_idx)
drivers/infiniband/hw/irdma/hw.c
148
if (cq_idx == IRDMA_RSVD_CQ_ID_CQP) {
drivers/infiniband/hw/irdma/hw.c
153
} else if (cq_idx == IRDMA_RSVD_CQ_ID_ILQ ||
drivers/infiniband/hw/irdma/hw.c
154
cq_idx == IRDMA_RSVD_CQ_ID_IEQ) {
drivers/infiniband/hw/irdma/hw.c
156
cq = (cq_idx == IRDMA_RSVD_CQ_ID_ILQ) ?
drivers/infiniband/hw/irdma/hw.c
179
u32 cq_idx;
drivers/infiniband/hw/irdma/hw.c
185
if (!irdma_sc_process_ceq(dev, sc_ceq, &cq_idx)) {
drivers/infiniband/hw/irdma/hw.c
191
if (likely(cq_idx > IRDMA_RSVD_CQ_ID_IEQ)) {
drivers/infiniband/hw/irdma/hw.c
192
irdma_process_normal_ceqe(rf, dev, cq_idx);
drivers/infiniband/hw/irdma/hw.c
199
irdma_process_reserved_ceqe(rf, dev, cq_idx);
drivers/infiniband/hw/irdma/type.h
1351
u32 *cq_idx);
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
965
static void ocrdma_qp_cq_handler(struct ocrdma_dev *dev, u16 cq_idx)
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
970
if (cq_idx >= OCRDMA_MAX_CQ)
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
973
cq = dev->cq_tbl[cq_idx];
drivers/infiniband/hw/usnic/usnic_abi.h
83
u32 cq_idx[USNIC_QP_GRP_MAX_CQS];
drivers/infiniband/hw/usnic/usnic_ib_verbs.c
154
resp.cq_idx[i] = chunk->res[i]->vnic_idx;
drivers/media/platform/mediatek/mdp3/mdp_sm_mt8183.h
114
u32 cq_idx;
drivers/media/platform/mediatek/mdp3/mdp_sm_mt8195.h
245
u32 cq_idx;
drivers/net/ethernet/amazon/ena/ena_admin_defs.h
266
u16 cq_idx;
drivers/net/ethernet/amazon/ena/ena_admin_defs.h
354
u16 cq_idx;
drivers/net/ethernet/amazon/ena/ena_admin_defs.h
369
u16 cq_idx;
drivers/net/ethernet/amazon/ena/ena_admin_defs.h
963
u16 cq_idx;
drivers/net/ethernet/amazon/ena/ena_com.c
1174
struct ena_com_io_sq *io_sq, u16 cq_idx)
drivers/net/ethernet/amazon/ena/ena_com.c
1205
create_cmd.cq_idx = cq_idx;
drivers/net/ethernet/amazon/ena/ena_com.c
1261
rss->rss_ind_tbl[i].cq_idx = io_sq->idx;
drivers/net/ethernet/amazon/ena/ena_com.c
1370
io_cq->idx = cmd_completion.cq_idx;
drivers/net/ethernet/amazon/ena/ena_com.c
1446
destroy_cmd.cq_idx = io_cq->idx;
drivers/net/ethernet/cavium/thunder/nic.h
123
u8 cq_idx; /* Completion queue index */
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1043
int qidx = cq_poll->cq_idx;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1473
cq_poll->cq_idx = qidx;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
845
static int nicvf_cq_intr_handler(struct net_device *netdev, u8 cq_idx,
drivers/net/ethernet/cavium/thunder/nicvf_main.c
853
struct cmp_queue *cq = &qs->cq[cq_idx];
drivers/net/ethernet/cavium/thunder/nicvf_main.c
856
struct snd_queue *sq = &qs->sq[cq_idx];
drivers/net/ethernet/cavium/thunder/nicvf_main.c
857
struct rcv_queue *rq = &qs->rq[cq_idx];
drivers/net/ethernet/cavium/thunder/nicvf_main.c
864
cqe_count = nicvf_queue_reg_read(nic, NIC_QSET_CQ_0_7_STATUS, cq_idx);
drivers/net/ethernet/cavium/thunder/nicvf_main.c
870
cqe_head = nicvf_queue_reg_read(nic, NIC_QSET_CQ_0_7_HEAD, cq_idx) >> 9;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
911
cq_idx, processed_cqe);
drivers/net/ethernet/cavium/thunder/nicvf_main.c
921
txq_idx = nicvf_netdev_qidx(nic, cq_idx);
drivers/net/ethernet/cavium/thunder/nicvf_main.c
925
nicvf_xdp_sq_doorbell(nic, sq, cq_idx);
drivers/net/ethernet/cavium/thunder/nicvf_main.c
965
work_done = nicvf_cq_intr_handler(netdev, cq->cq_idx, napi, budget);
drivers/net/ethernet/cavium/thunder/nicvf_main.c
972
cq->cq_idx);
drivers/net/ethernet/cavium/thunder/nicvf_main.c
973
nicvf_clear_intr(nic, NICVF_INTR_CQ, cq->cq_idx);
drivers/net/ethernet/cavium/thunder/nicvf_main.c
975
cq->cq_idx, cq_head);
drivers/net/ethernet/cavium/thunder/nicvf_main.c
976
nicvf_enable_intr(nic, NICVF_INTR_CQ, cq->cq_idx);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
764
rq->cq_idx = qidx;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
780
(rq->cq_idx << 16) | (rq->cont_rbdr_qs << 9) |
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
877
sq->cq_idx = qidx;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
884
mbx.sq.cfg = (sq->cq_qs << 3) | sq->cq_idx;
drivers/net/ethernet/cavium/thunder/nicvf_queues.h
245
u8 cq_idx; /* CQ index (0 to 7) in the QS */
drivers/net/ethernet/cavium/thunder/nicvf_queues.h
267
u8 cq_idx; /* CQ index (0 to 7) in the above QS */
drivers/net/ethernet/chelsio/cxgb3/t3_cpl.h
234
__be16 cq_idx;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
133
pool = &pfvf->qset.pool[cq->cq_idx];
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
139
__cn10k_aura_freeptr(pfvf, cq->cq_idx, ptrs,
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
149
__cn10k_aura_freeptr(pfvf, cq->cq_idx, ptrs,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1060
cq->cq_idx = qidx;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
589
cq->cq_idx, cq->pool_ptrs - 1)))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1305
otx2_free_rcv_seg(pfvf, cqe, cq->cq_idx);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1315
((u64)cq->cq_idx << 32) | processed_cqe);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1329
qidx = cq->cq_idx - pfvf->hw.rx_queues;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1361
((u64)cq->cq_idx << 32) | processed_cqe);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1498
int qidx = cq->cq_idx;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
346
if (otx2_check_rcv_errors(pfvf, cqe, cq->cq_idx)) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
370
seg_size[seg], parse, cq->cq_idx))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
378
skb_record_rx_queue(skb, cq->cq_idx);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
431
((u64)cq->cq_idx << 32) | processed_cqe);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
445
otx2_aura_freeptr(pfvf, cq->cq_idx, bufptr + OTX2_HEAD_ROOM);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
474
qidx = cq->cq_idx - pfvf->hw.rx_queues;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
513
((u64)cq->cq_idx << 32) | processed_cqe);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
525
qidx = cq->cq_idx - pfvf->hw.rx_queues;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
571
int workdone = 0, cq_idx, i;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
581
cq_idx = cq_poll->cq_ids[i];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
582
if (unlikely(cq_idx == CINT_INVALID_CQ))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
584
cq = &qset->cq[cq_idx];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
60
u64 incr = (u64)(cq->cq_idx) << 32;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
610
pool = &pfvf->qset.pool[cq->cq_idx];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
617
work = &pfvf->refill_wrk[cq->cq_idx];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.h
142
u8 cq_idx;
drivers/net/ethernet/mellanox/mlx4/en_cq.c
126
cq_idx = cq_idx % priv->rx_ring_num;
drivers/net/ethernet/mellanox/mlx4/en_cq.c
127
rx_cq = priv->rx_cq[cq_idx];
drivers/net/ethernet/mellanox/mlx4/en_cq.c
146
cq->cq_idx = cq_idx;
drivers/net/ethernet/mellanox/mlx4/en_cq.c
155
netif_queue_set_napi(cq->dev, cq_idx, NETDEV_QUEUE_TYPE_TX, &cq->napi);
drivers/net/ethernet/mellanox/mlx4/en_cq.c
160
cq_idx);
drivers/net/ethernet/mellanox/mlx4/en_cq.c
163
netif_queue_set_napi(cq->dev, cq_idx, NETDEV_QUEUE_TYPE_RX, &cq->napi);
drivers/net/ethernet/mellanox/mlx4/en_cq.c
206
netif_queue_set_napi(cq->dev, cq->cq_idx, qtype, NULL);
drivers/net/ethernet/mellanox/mlx4/en_cq.c
90
int cq_idx)
drivers/net/ethernet/mellanox/mlx4/mlx4_en.h
374
int cq_idx;
drivers/net/ethernet/mellanox/mlx4/mlx4_en.h
687
int cq_idx);
drivers/nvme/host/pci.c
514
nvmeq->dbbuf_cq_db = &dev->dbbuf_dbs[cq_idx(qid, dev->db_stride)];
drivers/nvme/host/pci.c
516
nvmeq->dbbuf_cq_ei = &dev->dbbuf_eis[cq_idx(qid, dev->db_stride)];
drivers/scsi/lpfc/lpfc_debugfs.h
659
int cq_idx;
drivers/scsi/lpfc/lpfc_debugfs.h
661
for (cq_idx = 0; cq_idx < phba->cfg_hdw_queue; cq_idx++)
drivers/scsi/lpfc/lpfc_debugfs.h
662
if (phba->sli4_hba.hdwq[cq_idx].io_cq->queue_id == qid)
drivers/scsi/lpfc/lpfc_debugfs.h
665
if (cq_idx < phba->cfg_hdw_queue) {
drivers/scsi/lpfc/lpfc_debugfs.h
666
pr_err("IO CQ[Idx:%d|Qid:%d]\n", cq_idx, qid);
drivers/scsi/lpfc/lpfc_debugfs.h
667
lpfc_debug_dump_q(phba->sli4_hba.hdwq[cq_idx].io_cq);
drivers/scsi/qedf/qedf_io.c
597
uint16_t cq_idx = smp_processor_id() % qedf->num_queues;
drivers/scsi/qedf/qedf_io.c
630
io_req->task_params->cq_rss_number = cq_idx;
drivers/scsi/qedi/qedi_fw.c
1993
u16 cq_idx;
drivers/scsi/qedi/qedi_fw.c
2021
cq_idx = smp_processor_id() % qedi->num_queues;
drivers/scsi/qedi/qedi_fw.c
2086
task_params.cq_rss_number = cq_idx;
drivers/scsi/snic/snic_scsi.c
1179
unsigned int cq_idx,
drivers/scsi/snic/snic_scsi.c
1262
unsigned int cq_idx;
drivers/scsi/snic/snic_scsi.c
1266
for (cq_idx = snic->wq_count; cq_idx < snic->cq_count; cq_idx++) {
drivers/scsi/snic/snic_scsi.c
1267
nent_per_cq = vnic_cq_fw_service(&snic->cq[cq_idx],
include/uapi/rdma/efa-abi.h
78
__u16 cq_idx;