Symbol: cqe
drivers/crypto/hisilicon/debugfs.c
360
cqe_curr = qp->cqe + cqe_id;
drivers/crypto/hisilicon/qm.c
1002
while (QM_CQE_PHASE(cqe) == qp->qp_status.cqc_phase) {
drivers/crypto/hisilicon/qm.c
1005
le16_to_cpu(cqe->sq_head));
drivers/crypto/hisilicon/qm.c
1007
cqe = qp->cqe + qp->qp_status.cq_head;
drivers/crypto/hisilicon/qm.c
2108
memset(qp->cqe, 0, sizeof(struct qm_cqe) * qp->cq_depth);
drivers/crypto/hisilicon/qm.c
2640
struct qm_cqe *cqe = qp->cqe + qp->qp_status.cq_head;
drivers/crypto/hisilicon/qm.c
2643
while (QM_CQE_PHASE(cqe) == qp->qp_status.cqc_phase) {
drivers/crypto/hisilicon/qm.c
2647
cqe = qp->cqe + qp->qp_status.cq_head;
drivers/crypto/hisilicon/qm.c
3044
qp->cqe = qp->qdma.va + off;
drivers/crypto/hisilicon/qm.c
62
#define QM_CQE_PHASE(cqe) (le16_to_cpu((cqe)->w7) & 0x1)
drivers/crypto/hisilicon/qm.c
999
struct qm_cqe *cqe = qp->cqe + qp->qp_status.cq_head;
drivers/dma/hisi_dma.c
732
struct hisi_dma_cqe *cqe;
drivers/dma/hisi_dma.c
738
cqe = chan->cq + chan->cq_head;
drivers/dma/hisi_dma.c
744
if (FIELD_GET(STATUS_MASK, cqe->w0) == STATUS_SUCC) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/rpc.c
355
struct r535_gsp_msg *cqe;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/rpc.c
394
cqe = (void *)((u8 *)gsp->shm.cmdq.ptr + 0x1000 + wptr * 0x1000);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/rpc.c
398
memcpy(cqe, (u8 *)msg + off, size);
drivers/infiniband/core/cq.c
217
.cqe = nr_cqe,
drivers/infiniband/core/cq.c
474
if (cq->cqe_used + nr_cqe > cq->cqe)
drivers/infiniband/core/mad.c
1032
mad_send_wr->mad_list.cqe.done = ib_mad_send_done;
drivers/infiniband/core/mad.c
1033
mad_send_wr->send_wr.wr.wr_cqe = &mad_send_wr->mad_list.cqe;
drivers/infiniband/core/mad.c
2279
container_of(wc->wr_cqe, struct ib_mad_list_head, cqe);
drivers/infiniband/core/mad.c
2505
container_of(wc->wr_cqe, struct ib_mad_list_head, cqe);
drivers/infiniband/core/mad.c
2590
container_of(wc->wr_cqe, struct ib_mad_list_head, cqe);
drivers/infiniband/core/mad.c
2967
mad_priv->header.mad_list.cqe.done = ib_mad_recv_done;
drivers/infiniband/core/mad.c
2968
recv_wr.wr_cqe = &mad_priv->header.mad_list.cqe;
drivers/infiniband/core/mad.c
572
static void build_smp_wc(struct ib_qp *qp, struct ib_cqe *cqe, u16 slid,
drivers/infiniband/core/mad.c
576
wc->wr_cqe = cqe;
drivers/infiniband/core/mad.c
918
mad_send_wr->mad_list.cqe.done = ib_mad_send_done;
drivers/infiniband/core/mad.c
920
mad_send_wr->send_wr.wr.wr_cqe = &mad_send_wr->mad_list.cqe;
drivers/infiniband/core/mad_priv.h
67
struct ib_cqe cqe;
drivers/infiniband/core/nldev.c
639
if (nla_put_u32(msg, RDMA_NLDEV_ATTR_RES_CQE, cq->cqe))
drivers/infiniband/core/rw.c
870
u32 port_num, struct ib_cqe *cqe, struct ib_send_wr *chain_wr)
drivers/infiniband/core/rw.c
909
last_wr->wr_cqe = cqe;
drivers/infiniband/core/rw.c
932
struct ib_cqe *cqe, struct ib_send_wr *chain_wr)
drivers/infiniband/core/rw.c
936
first_wr = rdma_rw_ctx_wrs(ctx, qp, port_num, cqe, chain_wr);
drivers/infiniband/core/uverbs_cmd.c
1052
attr.cqe = cmd->cqe;
drivers/infiniband/core/uverbs_cmd.c
1083
resp.base.cqe = cq->cqe;
drivers/infiniband/core/uverbs_cmd.c
1110
cmd_ex.cqe = cmd.cqe;
drivers/infiniband/core/uverbs_cmd.c
1150
ret = cq->device->ops.resize_cq(cq, cmd.cqe, &attrs->driver_udata);
drivers/infiniband/core/uverbs_cmd.c
1154
resp.cqe = cq->cqe;
drivers/infiniband/core/uverbs_std_types_cq.c
204
ret = uverbs_copy_to(attrs, UVERBS_ATTR_CREATE_CQ_RESP_CQE, &cq->cqe,
drivers/infiniband/core/uverbs_std_types_cq.c
205
sizeof(cq->cqe));
drivers/infiniband/core/uverbs_std_types_cq.c
87
ret = uverbs_copy_from(&attr.cqe, attrs,
drivers/infiniband/core/verbs.c
2258
int ib_resize_cq(struct ib_cq *cq, int cqe)
drivers/infiniband/core/verbs.c
2264
cq->device->ops.resize_cq(cq, cqe, NULL) : -EOPNOTSUPP;
drivers/infiniband/core/verbs.c
2875
struct ib_cqe cqe;
drivers/infiniband/core/verbs.c
2881
struct ib_drain_cqe *cqe = container_of(wc->wr_cqe, struct ib_drain_cqe,
drivers/infiniband/core/verbs.c
2882
cqe);
drivers/infiniband/core/verbs.c
2884
complete(&cqe->done);
drivers/infiniband/core/verbs.c
2898
{ .wr_cqe = &sdrain.cqe, },
drivers/infiniband/core/verbs.c
2910
sdrain.cqe.done = ib_drain_qp_done;
drivers/infiniband/core/verbs.c
2943
rwr.wr_cqe = &rdrain.cqe;
drivers/infiniband/core/verbs.c
2944
rdrain.cqe.done = ib_drain_qp_done;
drivers/infiniband/core/verbs.c
3017
while (polled != cq->cqe) {
drivers/infiniband/core/verbs.c
3018
n = ib_process_cq_direct(cq, cq->cqe - polled);
drivers/infiniband/hw/bnxt_re/bnxt_re.h
101
struct bnxt_qplib_cqe cqe;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3155
int cqe = attr->cqe;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3163
if (cqe < 1 || cqe > dev_attr->max_cq_wqes) {
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3172
entries = bnxt_re_init_depth(cqe + 1, uctx);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3215
cq->ib_cq.cqe = entries;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3274
int bnxt_re_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata)
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3300
if (cqe < 1 || cqe > dev_attr->max_cq_wqes) {
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3302
cq->qplib_cq.id, cqe);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3307
entries = bnxt_re_init_depth(cqe + 1, uctx);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3343
cq->ib_cq.cqe = cq->resize_cqe;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3442
static void bnxt_re_process_req_wc(struct ib_wc *wc, struct bnxt_qplib_cqe *cqe)
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3444
switch (cqe->type) {
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3483
wc->status = __req_to_ib_wc_status(cqe->status);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3576
struct bnxt_qplib_cqe *cqe)
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3597
tbl_idx = cqe->wr_id;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3610
memcpy(&sqp_entry->cqe, cqe, sizeof(struct bnxt_qplib_cqe));
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3615
pkt_type = bnxt_re_check_packet_type(cqe->raweth_qp1_flags,
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3616
cqe->raweth_qp1_flags2);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3687
struct bnxt_qplib_cqe *cqe)
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3690
wc->status = __rawqp1_to_ib_wc_status(cqe->status);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3734
struct bnxt_qplib_cqe *cqe)
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3737
wc->status = __rc_to_ib_wc_status(cqe->status);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3739
if (cqe->flags & CQ_RES_RC_FLAGS_IMM)
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3741
if (cqe->flags & CQ_RES_RC_FLAGS_INV)
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3743
if ((cqe->flags & (CQ_RES_RC_FLAGS_RDMA | CQ_RES_RC_FLAGS_IMM)) ==
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3750
struct bnxt_qplib_cqe *cqe)
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3761
tbl_idx = cqe->wr_id;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3765
orig_cqe = &sqp_entry->cqe;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3798
struct bnxt_qplib_cqe *cqe)
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3806
wc->status = __rc_to_ib_wc_status(cqe->status);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3808
if (cqe->flags & CQ_RES_UD_FLAGS_IMM)
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3813
memcpy(wc->smac, cqe->smac, ETH_ALEN);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3815
if (cqe->flags & CQ_RES_UD_FLAGS_META_FORMAT_VLAN) {
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3816
vlan_id = (cqe->cfa_meta & 0xFFF);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3823
nw_type = (cqe->flags & CQ_RES_UD_FLAGS_ROCE_IP_VER_MASK) >>
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3857
struct bnxt_qplib_cqe *cqe;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3881
cqe = &cq->cql[0];
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3884
ncqe = bnxt_qplib_poll_cq(&cq->qplib_cq, cqe, budget, &lib_qp);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3899
cqe + ncqe,
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3905
for (i = 0; i < ncqe; i++, cqe++) {
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3909
wc->wr_id = cqe->wr_id;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3910
wc->byte_len = cqe->length;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3913
(unsigned long)(cqe->qp_handle),
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3916
if (cqe->flags & CQ_RES_RC_FLAGS_IMM)
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3917
wc->ex.imm_data = cpu_to_be32(cqe->immdata);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3919
wc->ex.invalidate_rkey = cqe->invrkey;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3920
wc->src_qp = cqe->src_qp;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3921
memcpy(wc->smac, cqe->smac, ETH_ALEN);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3923
wc->vendor_err = cqe->status;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3925
switch (cqe->opcode) {
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3936
bnxt_re_process_req_wc(wc, cqe);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3939
if (!cqe->status) {
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3943
(qp, cqe);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3948
cqe->status = -1;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3954
tbl_idx = cqe->wr_id;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3957
bnxt_re_process_res_rawqp1_wc(wc, cqe);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3960
bnxt_re_process_res_rc_wc(wc, cqe);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3969
if (cqe->status) {
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3973
(qp, wc, cqe);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3977
bnxt_re_process_res_ud_wc(qp, wc, cqe);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3982
cqe->opcode);
drivers/infiniband/hw/bnxt_re/ib_verbs.h
250
int bnxt_re_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1557
struct cq_req *cqe = (struct cq_req *)hw_cqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1559
if (qp == le64_to_cpu(cqe->qp_handle))
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1560
cqe->qp_handle = 0;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1567
struct cq_res_rc *cqe = (struct cq_res_rc *)hw_cqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1569
if (qp == le64_to_cpu(cqe->qp_handle))
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1570
cqe->qp_handle = 0;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2373
struct bnxt_qplib_cqe *cqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2379
cqe = *pcqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2389
memset(cqe, 0, sizeof(*cqe));
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2390
cqe->status = CQ_REQ_STATUS_WORK_REQUEST_FLUSHED_ERR;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2391
cqe->opcode = CQ_BASE_CQE_TYPE_REQ;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2392
cqe->qp_handle = (u64)(unsigned long)qp;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2393
cqe->wr_id = sq->swq[last].wr_id;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2394
cqe->src_qp = qp->id;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2395
cqe->type = sq->swq[last].type;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2396
cqe++;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2403
*pcqe = cqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2414
struct bnxt_qplib_cqe *cqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2434
cqe = *pcqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2439
memset(cqe, 0, sizeof(*cqe));
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2440
cqe->status =
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2442
cqe->opcode = opcode;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2443
cqe->qp_handle = (unsigned long)qp;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2444
cqe->wr_id = rq->swq[last].wr_id;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2445
cqe++;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2451
*pcqe = cqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2603
struct bnxt_qplib_cqe *cqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2643
cqe = *pcqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2650
memset(cqe, 0, sizeof(*cqe));
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2651
cqe->opcode = CQ_BASE_CQE_TYPE_REQ;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2652
cqe->qp_handle = (u64)(unsigned long)qp;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2653
cqe->src_qp = qp->id;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2654
cqe->wr_id = swq->wr_id;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2655
if (cqe->wr_id == BNXT_QPLIB_FENCE_WRID)
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2657
cqe->type = swq->type;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2665
cqe->status = hwcqe->status;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2668
sq->swq_last, cqe->wr_id, cqe->status);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2669
cqe++;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2684
cqe->status = CQ_REQ_STATUS_OK;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2685
cqe++;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2697
*pcqe = cqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2729
struct bnxt_qplib_cqe *cqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2746
cqe = *pcqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2747
cqe->opcode = hwcqe->cqe_type_toggle & CQ_BASE_CQE_TYPE_MASK;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2748
cqe->length = le32_to_cpu(hwcqe->length);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2749
cqe->invrkey = le32_to_cpu(hwcqe->imm_data_or_inv_r_key);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2750
cqe->mr_handle = le64_to_cpu(hwcqe->mr_handle);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2751
cqe->flags = le16_to_cpu(hwcqe->flags);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2752
cqe->status = hwcqe->status;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2753
cqe->qp_handle = (u64)(unsigned long)qp;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2757
if (cqe->flags & CQ_RES_RC_FLAGS_SRQ_SRQ) {
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2767
cqe->wr_id = srq->swq[wr_id_idx].wr_id;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2769
cqe++;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2771
*pcqe = cqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2785
cqe->wr_id = swq->wr_id;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2786
cqe++;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2791
*pcqe = cqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2809
struct bnxt_qplib_cqe *cqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2825
cqe = *pcqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2826
cqe->opcode = hwcqe->cqe_type_toggle & CQ_BASE_CQE_TYPE_MASK;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2827
cqe->length = le16_to_cpu(hwcqe->length) & CQ_RES_UD_LENGTH_MASK;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2828
cqe->cfa_meta = le16_to_cpu(hwcqe->cfa_metadata);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2829
cqe->invrkey = le32_to_cpu(hwcqe->imm_data);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2830
cqe->flags = le16_to_cpu(hwcqe->flags);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2831
cqe->status = hwcqe->status;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2832
cqe->qp_handle = (u64)(unsigned long)qp;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2834
memcpy(cqe->smac, hwcqe->src_mac, ETH_ALEN);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2837
cqe->src_qp = le16_to_cpu(hwcqe->src_qp_low) |
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2842
if (cqe->flags & CQ_RES_RC_FLAGS_SRQ_SRQ) {
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2853
cqe->wr_id = srq->swq[wr_id_idx].wr_id;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2855
cqe++;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2857
*pcqe = cqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2872
cqe->wr_id = swq->wr_id;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2873
cqe++;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2878
*pcqe = cqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2909
struct bnxt_qplib_cqe *cqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2923
cqe = *pcqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2924
cqe->opcode = hwcqe->cqe_type_toggle & CQ_BASE_CQE_TYPE_MASK;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2925
cqe->flags = le16_to_cpu(hwcqe->flags);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2926
cqe->qp_handle = (u64)(unsigned long)qp;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2931
cqe->src_qp = qp->id;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2932
if (qp->id == 1 && !cqe->length) {
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2934
cqe->length = 296;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2936
cqe->length = le16_to_cpu(hwcqe->length);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2938
cqe->pkey_index = qp->pkey_index;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2939
memcpy(cqe->smac, qp->smac, 6);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2941
cqe->raweth_qp1_flags = le16_to_cpu(hwcqe->raweth_qp1_flags);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2942
cqe->raweth_qp1_flags2 = le32_to_cpu(hwcqe->raweth_qp1_flags2);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2943
cqe->raweth_qp1_metadata = le32_to_cpu(hwcqe->raweth_qp1_metadata);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2945
if (cqe->flags & CQ_RES_RAWETH_QP1_FLAGS_SRQ_SRQ) {
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2958
cqe->wr_id = srq->swq[wr_id_idx].wr_id;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2960
cqe++;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2962
*pcqe = cqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2976
cqe->wr_id = swq->wr_id;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2977
cqe++;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2982
*pcqe = cqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
3001
struct bnxt_qplib_cqe *cqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
3037
cqe = *pcqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
3043
memset(cqe, 0, sizeof(*cqe));
drivers/infiniband/hw/bnxt_re/qplib_fp.c
3044
cqe->status = CQ_REQ_STATUS_OK;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
3045
cqe->opcode = CQ_BASE_CQE_TYPE_REQ;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
3046
cqe->qp_handle = (u64)(unsigned long)qp;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
3047
cqe->src_qp = qp->id;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
3048
cqe->wr_id = sq->swq[swq_last].wr_id;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
3049
cqe->type = sq->swq[swq_last].type;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
3050
cqe++;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
3057
*pcqe = cqe;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
3113
struct bnxt_qplib_cqe *cqe,
drivers/infiniband/hw/bnxt_re/qplib_fp.c
3123
__flush_sq(&qp->sq, qp, &cqe, &budget);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
3128
__flush_rq(&qp->rq, qp, &cqe, &budget);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
3135
int bnxt_qplib_poll_cq(struct bnxt_qplib_cq *cq, struct bnxt_qplib_cqe *cqe,
drivers/infiniband/hw/bnxt_re/qplib_fp.c
3163
&cqe, &budget,
drivers/infiniband/hw/bnxt_re/qplib_fp.c
3169
hw_cqe, &cqe,
drivers/infiniband/hw/bnxt_re/qplib_fp.c
3174
(cq, (struct cq_res_ud *)hw_cqe, &cqe,
drivers/infiniband/hw/bnxt_re/qplib_fp.c
3180
hw_cqe, &cqe, &budget);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
3185
&cqe, &budget);
drivers/infiniband/hw/bnxt_re/qplib_fp.h
588
int bnxt_qplib_poll_cq(struct bnxt_qplib_cq *cq, struct bnxt_qplib_cqe *cqe,
drivers/infiniband/hw/bnxt_re/qplib_fp.h
600
struct bnxt_qplib_cqe *cqe,
drivers/infiniband/hw/cxgb4/cq.c
1002
int entries = attr->cqe;
drivers/infiniband/hw/cxgb4/cq.c
1087
chp->ibcq.cqe = entries - 2;
drivers/infiniband/hw/cxgb4/cq.c
186
struct t4_cqe cqe;
drivers/infiniband/hw/cxgb4/cq.c
190
memset(&cqe, 0, sizeof(cqe));
drivers/infiniband/hw/cxgb4/cq.c
191
cqe.header = cpu_to_be32(CQE_STATUS_V(T4_ERR_SWFLUSH) |
drivers/infiniband/hw/cxgb4/cq.c
196
cqe.bits_type_ts = cpu_to_be64(CQE_GENBIT_V((u64)cq->gen));
drivers/infiniband/hw/cxgb4/cq.c
198
cqe.u.srcqe.abs_rqe_idx = cpu_to_be32(srqidx);
drivers/infiniband/hw/cxgb4/cq.c
199
cq->sw_queue[cq->sw_pidx] = cqe;
drivers/infiniband/hw/cxgb4/cq.c
220
struct t4_cqe cqe;
drivers/infiniband/hw/cxgb4/cq.c
224
memset(&cqe, 0, sizeof(cqe));
drivers/infiniband/hw/cxgb4/cq.c
225
cqe.header = cpu_to_be32(CQE_STATUS_V(T4_ERR_SWFLUSH) |
drivers/infiniband/hw/cxgb4/cq.c
230
CQE_WRID_SQ_IDX(&cqe) = swcqe->idx;
drivers/infiniband/hw/cxgb4/cq.c
231
cqe.bits_type_ts = cpu_to_be64(CQE_GENBIT_V((u64)cq->gen));
drivers/infiniband/hw/cxgb4/cq.c
232
cq->sw_queue[cq->sw_pidx] = cqe;
drivers/infiniband/hw/cxgb4/cq.c
288
swsqe->cqe.header |= htonl(CQE_SWCQE_V(1));
drivers/infiniband/hw/cxgb4/cq.c
289
cq->sw_queue[cq->sw_pidx] = swsqe->cqe;
drivers/infiniband/hw/cxgb4/cq.c
405
swsqe->cqe = *hw_cqe;
drivers/infiniband/hw/cxgb4/cq.c
422
static int cqe_completes_wr(struct t4_cqe *cqe, struct t4_wq *wq)
drivers/infiniband/hw/cxgb4/cq.c
424
if (DRAIN_CQE(cqe)) {
drivers/infiniband/hw/cxgb4/cq.c
429
if (CQE_OPCODE(cqe) == FW_RI_TERMINATE)
drivers/infiniband/hw/cxgb4/cq.c
432
if ((CQE_OPCODE(cqe) == FW_RI_RDMA_WRITE) && RQ_TYPE(cqe))
drivers/infiniband/hw/cxgb4/cq.c
435
if ((CQE_OPCODE(cqe) == FW_RI_READ_RESP) && SQ_TYPE(cqe))
drivers/infiniband/hw/cxgb4/cq.c
438
if (CQE_SEND_OPCODE(cqe) && RQ_TYPE(cqe) && t4_rq_empty(wq))
drivers/infiniband/hw/cxgb4/cq.c
445
struct t4_cqe *cqe;
drivers/infiniband/hw/cxgb4/cq.c
452
cqe = &cq->sw_queue[ptr];
drivers/infiniband/hw/cxgb4/cq.c
453
if (RQ_TYPE(cqe) && (CQE_OPCODE(cqe) != FW_RI_READ_RESP) &&
drivers/infiniband/hw/cxgb4/cq.c
454
(CQE_QPID(cqe) == wq->sq.qid) && cqe_completes_wr(cqe, wq))
drivers/infiniband/hw/cxgb4/cq.c
544
static int poll_cq(struct t4_wq *wq, struct t4_cq *cq, struct t4_cqe *cqe,
drivers/infiniband/hw/cxgb4/cq.c
592
*cqe = *hw_cqe;
drivers/infiniband/hw/cxgb4/cq.c
686
swsqe->cqe = *hw_cqe;
drivers/infiniband/hw/cxgb4/cq.c
693
*cqe = *hw_cqe;
drivers/infiniband/hw/cxgb4/cq.c
757
struct t4_cqe cqe;
drivers/infiniband/hw/cxgb4/cq.c
764
ret = poll_cq(wq, &(chp->cq), &cqe, &cqe_flushed, &cookie, &credit,
drivers/infiniband/hw/cxgb4/cq.c
771
wc->vendor_err = CQE_STATUS(&cqe);
drivers/infiniband/hw/cxgb4/cq.c
782
CQE_QPID(&cqe),
drivers/infiniband/hw/cxgb4/cq.c
783
CQE_TYPE(&cqe), CQE_OPCODE(&cqe),
drivers/infiniband/hw/cxgb4/cq.c
784
CQE_STATUS(&cqe), CQE_LEN(&cqe),
drivers/infiniband/hw/cxgb4/cq.c
785
CQE_WRID_HI(&cqe), CQE_WRID_LOW(&cqe),
drivers/infiniband/hw/cxgb4/cq.c
788
if (CQE_TYPE(&cqe) == 0) {
drivers/infiniband/hw/cxgb4/cq.c
789
if (!CQE_STATUS(&cqe))
drivers/infiniband/hw/cxgb4/cq.c
790
wc->byte_len = CQE_LEN(&cqe);
drivers/infiniband/hw/cxgb4/cq.c
794
switch (CQE_OPCODE(&cqe)) {
drivers/infiniband/hw/cxgb4/cq.c
801
wc->ex.invalidate_rkey = CQE_WRID_STAG(&cqe);
drivers/infiniband/hw/cxgb4/cq.c
807
wc->ex.imm_data = CQE_IMM_DATA(&cqe);
drivers/infiniband/hw/cxgb4/cq.c
812
CQE_OPCODE(&cqe), CQE_QPID(&cqe));
drivers/infiniband/hw/cxgb4/cq.c
817
switch (CQE_OPCODE(&cqe)) {
drivers/infiniband/hw/cxgb4/cq.c
824
wc->byte_len = CQE_LEN(&cqe);
drivers/infiniband/hw/cxgb4/cq.c
843
if (CQE_STATUS(&cqe) != T4_ERR_SUCCESS)
drivers/infiniband/hw/cxgb4/cq.c
845
CQE_WRID_FR_STAG(&cqe));
drivers/infiniband/hw/cxgb4/cq.c
849
CQE_OPCODE(&cqe), CQE_QPID(&cqe));
drivers/infiniband/hw/cxgb4/cq.c
859
switch (CQE_STATUS(&cqe)) {
drivers/infiniband/hw/cxgb4/cq.c
904
CQE_STATUS(&cqe), CQE_QPID(&cqe));
drivers/infiniband/hw/cxgb4/device.c
105
le.cqe_sge_ts = CQE_TS(cqe);
drivers/infiniband/hw/cxgb4/device.c
106
if (SQ_TYPE(cqe)) {
drivers/infiniband/hw/cxgb4/device.c
108
le.opcode = CQE_OPCODE(cqe);
drivers/infiniband/hw/cxgb4/device.c
111
le.wr_id = CQE_WRID_SQ_IDX(cqe);
drivers/infiniband/hw/cxgb4/device.c
117
le.wr_id = CQE_WRID_MSN(cqe);
drivers/infiniband/hw/cxgb4/device.c
92
void c4iw_log_wr_stats(struct t4_wq *wq, struct t4_cqe *cqe)
drivers/infiniband/hw/cxgb4/iw_cxgb4.h
1059
extern void c4iw_log_wr_stats(struct t4_wq *wq, struct t4_cqe *cqe);
drivers/infiniband/hw/cxgb4/qp.c
1001
cqe.bits_type_ts = cpu_to_be64(CQE_GENBIT_V((u64)cq->gen));
drivers/infiniband/hw/cxgb4/qp.c
1002
cq->sw_queue[cq->sw_pidx] = cqe;
drivers/infiniband/hw/cxgb4/qp.c
1035
struct t4_cqe cqe = {};
drivers/infiniband/hw/cxgb4/qp.c
1043
cqe.u.drain_cookie = wr->wr_id;
drivers/infiniband/hw/cxgb4/qp.c
1044
cqe.header = cpu_to_be32(CQE_STATUS_V(T4_ERR_SWFLUSH) |
drivers/infiniband/hw/cxgb4/qp.c
1052
cqe.bits_type_ts = cpu_to_be64(CQE_GENBIT_V((u64)cq->gen));
drivers/infiniband/hw/cxgb4/qp.c
1053
cq->sw_queue[cq->sw_pidx] = cqe;
drivers/infiniband/hw/cxgb4/qp.c
979
struct t4_cqe cqe = {};
drivers/infiniband/hw/cxgb4/qp.c
992
cqe.u.drain_cookie = wr->wr_id;
drivers/infiniband/hw/cxgb4/qp.c
993
cqe.header = cpu_to_be32(CQE_STATUS_V(T4_ERR_SWFLUSH) |
drivers/infiniband/hw/cxgb4/restrack.c
105
rdma_nl_put_driver_u32(msg, "cqe_status", CQE_STATUS(&sqe->cqe)))
drivers/infiniband/hw/cxgb4/restrack.c
307
static int fill_cqe(struct sk_buff *msg, struct t4_cqe *cqe, u16 idx,
drivers/infiniband/hw/cxgb4/restrack.c
313
be32_to_cpu(cqe->header)))
drivers/infiniband/hw/cxgb4/restrack.c
315
if (rdma_nl_put_driver_u32(msg, "len", be32_to_cpu(cqe->len)))
drivers/infiniband/hw/cxgb4/restrack.c
318
be32_to_cpu(cqe->u.gen.wrid_hi)))
drivers/infiniband/hw/cxgb4/restrack.c
321
be32_to_cpu(cqe->u.gen.wrid_low)))
drivers/infiniband/hw/cxgb4/restrack.c
324
be64_to_cpu(cqe->bits_type_ts)))
drivers/infiniband/hw/cxgb4/t4.h
308
struct t4_cqe cqe;
drivers/infiniband/hw/cxgb4/t4.h
770
static inline int t4_valid_cqe(struct t4_cq *cq, struct t4_cqe *cqe)
drivers/infiniband/hw/cxgb4/t4.h
772
return (CQE_GENBIT(cqe) == cq->gen);
drivers/infiniband/hw/cxgb4/t4.h
780
static inline int t4_next_hw_cqe(struct t4_cq *cq, struct t4_cqe **cqe)
drivers/infiniband/hw/cxgb4/t4.h
798
*cqe = &cq->queue[cq->cidx];
drivers/infiniband/hw/cxgb4/t4.h
805
static inline int t4_next_cqe(struct t4_cq *cq, struct t4_cqe **cqe)
drivers/infiniband/hw/cxgb4/t4.h
812
*cqe = &cq->sw_queue[cq->sw_cidx];
drivers/infiniband/hw/cxgb4/t4.h
814
ret = t4_next_hw_cqe(cq, cqe);
drivers/infiniband/hw/efa/efa_com.c
413
struct efa_admin_acq_entry *cqe)
drivers/infiniband/hw/efa/efa_com.c
418
cmd_id = EFA_GET(&cqe->acq_common_descriptor.command,
drivers/infiniband/hw/efa/efa_com.c
431
memcpy(comp_ctx->user_cqe, cqe, comp_ctx->comp_size);
drivers/infiniband/hw/efa/efa_com.c
441
struct efa_admin_acq_entry *cqe;
drivers/infiniband/hw/efa/efa_com.c
453
cqe = &aq->cq.entries[ci];
drivers/infiniband/hw/efa/efa_com.c
456
while ((READ_ONCE(cqe->acq_common_descriptor.flags) &
drivers/infiniband/hw/efa/efa_com.c
463
err = efa_com_handle_single_admin_completion(aq, cqe);
drivers/infiniband/hw/efa/efa_com.c
474
cqe = &aq->cq.entries[ci];
drivers/infiniband/hw/efa/efa_verbs.c
1147
int entries = attr->cqe;
drivers/infiniband/hw/efa/efa_verbs.c
1259
cq->ibcq.cqe = result.actual_depth;
drivers/infiniband/hw/erdma/erdma_cmdq.c
249
__be32 *cqe = get_queue_entry(cmdq->cq.qbuf, cmdq->cq.ci,
drivers/infiniband/hw/erdma/erdma_cmdq.c
252
be32_to_cpu(READ_ONCE(*cqe)));
drivers/infiniband/hw/erdma/erdma_cmdq.c
254
return owner ^ !!(cmdq->cq.ci & cmdq->cq.depth) ? cqe : NULL;
drivers/infiniband/hw/erdma/erdma_cmdq.c
285
__be32 *cqe;
drivers/infiniband/hw/erdma/erdma_cmdq.c
289
cqe = get_next_valid_cmdq_cqe(cmdq);
drivers/infiniband/hw/erdma/erdma_cmdq.c
290
if (!cqe)
drivers/infiniband/hw/erdma/erdma_cmdq.c
296
hdr0 = be32_to_cpu(*cqe);
drivers/infiniband/hw/erdma/erdma_cmdq.c
297
sqe_idx = be32_to_cpu(*(cqe + 1));
drivers/infiniband/hw/erdma/erdma_cmdq.c
310
be32_to_cpu_array(comp_wait->comp_data, cqe + 2, 4);
drivers/infiniband/hw/erdma/erdma_cq.c
108
static void erdma_process_ud_cqe(struct erdma_cqe *cqe, struct ib_wc *wc)
drivers/infiniband/hw/erdma/erdma_cq.c
11
__be32 *cqe = get_queue_entry(cq->kern_cq.qbuf, cq->kern_cq.ci,
drivers/infiniband/hw/erdma/erdma_cq.c
113
ud_info = be32_to_cpu(cqe->ud.info);
drivers/infiniband/hw/erdma/erdma_cq.c
131
struct erdma_cqe *cqe;
drivers/infiniband/hw/erdma/erdma_cq.c
138
cqe = get_next_valid_cqe(cq);
drivers/infiniband/hw/erdma/erdma_cq.c
139
if (!cqe)
drivers/infiniband/hw/erdma/erdma_cq.c
14
be32_to_cpu(READ_ONCE(*cqe)));
drivers/infiniband/hw/erdma/erdma_cq.c
147
qpn = be32_to_cpu(cqe->qpn);
drivers/infiniband/hw/erdma/erdma_cq.c
148
wqe_idx = be32_to_cpu(cqe->qe_idx);
drivers/infiniband/hw/erdma/erdma_cq.c
149
cqe_hdr = be32_to_cpu(cqe->hdr);
drivers/infiniband/hw/erdma/erdma_cq.c
16
return owner ^ !!(cq->kern_cq.ci & cq->depth) ? cqe : NULL;
drivers/infiniband/hw/erdma/erdma_cq.c
174
wc->byte_len = be32_to_cpu(cqe->size);
drivers/infiniband/hw/erdma/erdma_cq.c
180
wc->ex.imm_data = cpu_to_be32(le32_to_cpu(cqe->imm_data));
drivers/infiniband/hw/erdma/erdma_cq.c
183
wc->ex.invalidate_rkey = be32_to_cpu(cqe->inv_rkey);
drivers/infiniband/hw/erdma/erdma_cq.c
189
erdma_process_ud_cqe(cqe, wc);
drivers/infiniband/hw/erdma/erdma_cq.c
228
struct erdma_cqe *cqe, *dst_cqe;
drivers/infiniband/hw/erdma/erdma_cq.c
238
while (ncqe < cq->depth && (cqe = get_next_valid_cqe(cq)) != NULL) {
drivers/infiniband/hw/erdma/erdma_cq.c
245
cqe = get_queue_entry(cq->kern_cq.qbuf, cur_cq_ci, cq->depth,
drivers/infiniband/hw/erdma/erdma_cq.c
248
if (be32_to_cpu(cqe->qpn) == qpn) {
drivers/infiniband/hw/erdma/erdma_cq.c
256
cqe->hdr = cpu_to_be32(
drivers/infiniband/hw/erdma/erdma_cq.c
257
(be32_to_cpu(cqe->hdr) &
drivers/infiniband/hw/erdma/erdma_cq.c
260
memcpy(dst_cqe, cqe, sizeof(*cqe));
drivers/infiniband/hw/erdma/erdma_verbs.c
1960
unsigned int depth = attr->cqe;
drivers/infiniband/hw/erdma/erdma_verbs.c
1969
cq->ibcq.cqe = depth;
drivers/infiniband/hw/hns/hns_roce_cq.c
338
if (!attr->cqe || attr->cqe > hr_dev->caps.max_cqes) {
drivers/infiniband/hw/hns/hns_roce_cq.c
340
attr->cqe, hr_dev->caps.max_cqes);
drivers/infiniband/hw/hns/hns_roce_cq.c
375
hr_cq->ib_cq.cqe = cq_entries - 1; /* used as cqe index */
drivers/infiniband/hw/hns/hns_roce_cq.c
436
set_cq_param(hr_cq, attr->cqe, attr->comp_vector, &ucmd);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
1012
{ .wr_cqe = &sdrain.cqe, },
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
1026
sdrain.cqe.done = hns_roce_drain_qp_done;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
2913
cq_init_attr.cqe = HNS_ROCE_FREE_MR_USED_CQE_NUM;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
3782
struct hns_roce_v2_cqe *cqe = get_cqe_v2(hr_cq, n & hr_cq->ib_cq.cqe);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
3785
return (hr_reg_read(cqe, CQE_OWNER) ^ !!(n & hr_cq->cq_depth)) ? cqe :
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
3810
struct hns_roce_v2_cqe *cqe, *dest;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
3818
if (prod_index > hr_cq->cons_index + hr_cq->ib_cq.cqe)
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
3827
cqe = get_cqe_v2(hr_cq, prod_index & hr_cq->ib_cq.cqe);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
3828
if (hr_reg_read(cqe, CQE_LCL_QPN) == qpn) {
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
3829
if (srq && hr_reg_read(cqe, CQE_S_R)) {
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
3830
wqe_index = hr_reg_read(cqe, CQE_WQE_IDX);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
3836
hr_cq->ib_cq.cqe);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
3838
memcpy(dest, cqe, hr_cq->cqe_size);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4005
struct hns_roce_cq *cq, struct hns_roce_v2_cqe *cqe,
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4030
u32 cqe_status = hr_reg_read(cqe, CQE_STATUS);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4046
print_hex_dump(KERN_DEBUG, "", DUMP_PREFIX_NONE, 16, 4, cqe,
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4048
wc->vendor_err = hr_reg_read(cqe, CQE_SUB_STATUS);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4061
static int get_cur_qp(struct hns_roce_cq *hr_cq, struct hns_roce_v2_cqe *cqe,
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4068
qpn = hr_reg_read(cqe, CQE_LCL_QPN);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4133
static void fill_send_wc(struct ib_wc *wc, struct hns_roce_v2_cqe *cqe)
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4140
hr_opcode = hr_reg_read(cqe, CQE_OPCODE);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4143
wc->byte_len = le32_to_cpu(cqe->byte_cnt);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4166
static int fill_recv_wc(struct ib_wc *wc, struct hns_roce_v2_cqe *cqe)
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4171
wc->byte_len = le32_to_cpu(cqe->byte_cnt);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4173
hr_opcode = hr_reg_read(cqe, CQE_OPCODE);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4178
wc->ex.imm_data = cpu_to_be32(le32_to_cpu(cqe->immtdata));
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4182
wc->ex.invalidate_rkey = le32_to_cpu(cqe->rkey);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4194
wc->sl = hr_reg_read(cqe, CQE_SL);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4195
wc->src_qp = hr_reg_read(cqe, CQE_RMT_QPN);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4197
wc->wc_flags |= hr_reg_read(cqe, CQE_GRH) ? IB_WC_GRH : 0;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4198
wc->port_num = hr_reg_read(cqe, CQE_PORTN);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4201
if (hr_reg_read(cqe, CQE_VID_VLD)) {
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4202
wc->vlan_id = hr_reg_read(cqe, CQE_VID);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4208
wc->network_hdr_type = hr_reg_read(cqe, CQE_PORT_TYPE);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4219
struct hns_roce_v2_cqe *cqe;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4225
cqe = get_sw_cqe_v2(hr_cq, hr_cq->cons_index);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4226
if (!cqe)
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4233
ret = get_cur_qp(hr_cq, cqe, &qp);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4240
wqe_idx = hr_reg_read(cqe, CQE_WQE_IDX);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4242
is_send = !hr_reg_read(cqe, CQE_S_R);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4256
fill_send_wc(wc, cqe);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4268
ret = fill_recv_wc(wc, cqe);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4271
get_cqe_status(hr_dev, qp, hr_cq, cqe, wc);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
900
struct ib_cqe cqe;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
906
struct hns_roce_drain_cqe *cqe = container_of(wc->wr_cqe,
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
908
cqe);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
909
complete(&cqe->done);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
983
rwr.wr_cqe = &rdrain.cqe;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
984
rdrain.cqe.done = hns_roce_drain_qp_done;
drivers/infiniband/hw/ionic/ionic_admin.c
100
struct ionic_v1_cqe *cqe;
drivers/infiniband/hw/ionic/ionic_admin.c
128
while (ionic_admin_next_cqe(dev, cq, &cqe)) {
drivers/infiniband/hw/ionic/ionic_admin.c
129
qtf = ionic_v1_cqe_qtf(cqe);
drivers/infiniband/hw/ionic/ionic_admin.c
145
if (unlikely(be16_to_cpu(cqe->admin.cmd_idx) != aq->q.cons)) {
drivers/infiniband/hw/ionic/ionic_admin.c
148
be16_to_cpu(cqe->admin.cmd_idx),
drivers/infiniband/hw/ionic/ionic_admin.c
164
wr->cqe = *cqe;
drivers/infiniband/hw/ionic/ionic_admin.c
437
} else if (ionic_v1_cqe_error(&wr->cqe)) {
drivers/infiniband/hw/ionic/ionic_admin.c
440
be32_to_cpu(wr->cqe.status_length));
drivers/infiniband/hw/ionic/ionic_admin.c
515
.cqe = IONIC_AQ_DEPTH,
drivers/infiniband/hw/ionic/ionic_admin.c
80
struct ionic_v1_cqe **cqe)
drivers/infiniband/hw/ionic/ionic_admin.c
89
*cqe = qcqe;
drivers/infiniband/hw/ionic/ionic_controlpath.c
1261
vcq->ibcq.cqe = attr->cqe;
drivers/infiniband/hw/ionic/ionic_controlpath.c
129
attr->cqe + IONIC_CQ_GRACE,
drivers/infiniband/hw/ionic/ionic_controlpath.c
92
if (attr->cqe < 1 || attr->cqe + IONIC_CQ_GRACE > 0xffff) {
drivers/infiniband/hw/ionic/ionic_datapath.c
131
struct ionic_qp *cqe_qp, struct ionic_v1_cqe *cqe,
drivers/infiniband/hw/ionic/ionic_datapath.c
145
st_len = be32_to_cpu(cqe->status_length);
drivers/infiniband/hw/ionic/ionic_datapath.c
148
if (ionic_v1_cqe_error(cqe) && st_len == IONIC_STS_WQE_FLUSHED_ERR) {
drivers/infiniband/hw/ionic/ionic_datapath.c
16
struct ionic_v1_cqe **cqe)
drivers/infiniband/hw/ionic/ionic_datapath.c
164
if (unlikely(cqe->recv.wqe_id >> qp->rq.depth_log2)) {
drivers/infiniband/hw/ionic/ionic_datapath.c
167
qp->qpid, (unsigned long long)cqe->recv.wqe_id);
drivers/infiniband/hw/ionic/ionic_datapath.c
172
meta = &qp->rq_meta[cqe->recv.wqe_id];
drivers/infiniband/hw/ionic/ionic_datapath.c
176
qp->qpid, (unsigned long long)cqe->recv.wqe_id);
drivers/infiniband/hw/ionic/ionic_datapath.c
189
if (ionic_v1_cqe_error(cqe)) {
drivers/infiniband/hw/ionic/ionic_datapath.c
200
cqe, BIT(cq->q.stride_log2), true);
drivers/infiniband/hw/ionic/ionic_datapath.c
207
src_qpn = be32_to_cpu(cqe->recv.src_qpn_op);
drivers/infiniband/hw/ionic/ionic_datapath.c
218
wc->ex.imm_data = cqe->recv.imm_data_rkey; /* be32 in wc */
drivers/infiniband/hw/ionic/ionic_datapath.c
222
wc->ex.imm_data = cqe->recv.imm_data_rkey; /* be32 in wc */
drivers/infiniband/hw/ionic/ionic_datapath.c
226
wc->ex.invalidate_rkey = be32_to_cpu(cqe->recv.imm_data_rkey);
drivers/infiniband/hw/ionic/ionic_datapath.c
236
ether_addr_copy(wc->smac, cqe->recv.src_mac);
drivers/infiniband/hw/ionic/ionic_datapath.c
239
if (ionic_v1_cqe_recv_is_ipv4(cqe))
drivers/infiniband/hw/ionic/ionic_datapath.c
244
if (ionic_v1_cqe_recv_is_vlan(cqe))
drivers/infiniband/hw/ionic/ionic_datapath.c
248
vlan_tag = be16_to_cpu(cqe->recv.vlan_tag);
drivers/infiniband/hw/ionic/ionic_datapath.c
26
*cqe = qcqe;
drivers/infiniband/hw/ionic/ionic_datapath.c
367
static int ionic_comp_msn(struct ionic_qp *qp, struct ionic_v1_cqe *cqe)
drivers/infiniband/hw/ionic/ionic_datapath.c
376
cqe_seq = be32_to_cpu(cqe->send.msg_msn) & qp->sq.mask;
drivers/infiniband/hw/ionic/ionic_datapath.c
385
qp->qpid, be32_to_cpu(cqe->send.msg_msn),
drivers/infiniband/hw/ionic/ionic_datapath.c
392
if (ionic_v1_cqe_error(cqe)) {
drivers/infiniband/hw/ionic/ionic_datapath.c
396
meta->len = be32_to_cpu(cqe->status_length);
drivers/infiniband/hw/ionic/ionic_datapath.c
402
cqe, sizeof(*cqe), true);
drivers/infiniband/hw/ionic/ionic_datapath.c
408
static int ionic_comp_npg(struct ionic_qp *qp, struct ionic_v1_cqe *cqe)
drivers/infiniband/hw/ionic/ionic_datapath.c
417
st_len = be32_to_cpu(cqe->status_length);
drivers/infiniband/hw/ionic/ionic_datapath.c
419
if (ionic_v1_cqe_error(cqe) && st_len == IONIC_STS_WQE_FLUSHED_ERR) {
drivers/infiniband/hw/ionic/ionic_datapath.c
433
cqe_idx = cqe->send.npg_wqe_id & qp->sq.mask;
drivers/infiniband/hw/ionic/ionic_datapath.c
437
if (ionic_v1_cqe_error(cqe)) {
drivers/infiniband/hw/ionic/ionic_datapath.c
444
cqe, sizeof(*cqe), true);
drivers/infiniband/hw/ionic/ionic_datapath.c
475
struct ionic_v1_cqe *cqe;
drivers/infiniband/hw/ionic/ionic_datapath.c
505
while (likely(ionic_next_cqe(dev, cq, &cqe))) {
drivers/infiniband/hw/ionic/ionic_datapath.c
509
qtf = ionic_v1_cqe_qtf(cqe);
drivers/infiniband/hw/ionic/ionic_datapath.c
529
rc = ionic_poll_recv(dev, cq, qp, cqe, wc + npolled);
drivers/infiniband/hw/ionic/ionic_datapath.c
541
rc = ionic_comp_msn(qp, cqe);
drivers/infiniband/hw/ionic/ionic_datapath.c
561
rc = ionic_comp_npg(qp, cqe);
drivers/infiniband/hw/ionic/ionic_fw.h
384
static inline bool ionic_v1_cqe_color(struct ionic_v1_cqe *cqe)
drivers/infiniband/hw/ionic/ionic_fw.h
386
return cqe->qid_type_flags & cpu_to_be32(IONIC_V1_CQE_COLOR);
drivers/infiniband/hw/ionic/ionic_fw.h
389
static inline bool ionic_v1_cqe_error(struct ionic_v1_cqe *cqe)
drivers/infiniband/hw/ionic/ionic_fw.h
391
return cqe->qid_type_flags & cpu_to_be32(IONIC_V1_CQE_ERROR);
drivers/infiniband/hw/ionic/ionic_fw.h
394
static inline bool ionic_v1_cqe_recv_is_ipv4(struct ionic_v1_cqe *cqe)
drivers/infiniband/hw/ionic/ionic_fw.h
396
return cqe->recv.src_qpn_op & cpu_to_be32(IONIC_V1_CQE_RECV_IS_IPV4);
drivers/infiniband/hw/ionic/ionic_fw.h
399
static inline bool ionic_v1_cqe_recv_is_vlan(struct ionic_v1_cqe *cqe)
drivers/infiniband/hw/ionic/ionic_fw.h
401
return cqe->recv.src_qpn_op & cpu_to_be32(IONIC_V1_CQE_RECV_IS_VLAN);
drivers/infiniband/hw/ionic/ionic_fw.h
404
static inline void ionic_v1_cqe_clean(struct ionic_v1_cqe *cqe)
drivers/infiniband/hw/ionic/ionic_fw.h
406
cqe->qid_type_flags |= cpu_to_be32(~0u << IONIC_V1_CQE_QID_SHIFT);
drivers/infiniband/hw/ionic/ionic_fw.h
409
static inline u32 ionic_v1_cqe_qtf(struct ionic_v1_cqe *cqe)
drivers/infiniband/hw/ionic/ionic_fw.h
411
return be32_to_cpu(cqe->qid_type_flags);
drivers/infiniband/hw/ionic/ionic_ibdev.h
142
struct ionic_v1_cqe cqe;
drivers/infiniband/hw/irdma/ctrl.c
3978
__le64 *cqe;
drivers/infiniband/hw/irdma/ctrl.c
3987
cqe = IRDMA_GET_CURRENT_EXTENDED_CQ_ELEM(&ccq->cq_uk);
drivers/infiniband/hw/irdma/ctrl.c
3989
cqe = IRDMA_GET_CURRENT_CQ_ELEM(&ccq->cq_uk);
drivers/infiniband/hw/irdma/ctrl.c
3991
get_64bit_val(cqe, 24, &temp);
drivers/infiniband/hw/irdma/ctrl.c
3999
get_64bit_val(cqe, 8, &qp_ctx);
drivers/infiniband/hw/irdma/ctrl.c
4014
get_64bit_val(cqe, 16, &temp1);
drivers/infiniband/hw/irdma/puda.c
216
__le64 *cqe;
drivers/infiniband/hw/irdma/puda.c
227
cqe = IRDMA_GET_CURRENT_CQ_ELEM(&cq->cq_uk);
drivers/infiniband/hw/irdma/puda.c
228
get_64bit_val(cqe, 24, &qword3);
drivers/infiniband/hw/irdma/puda.c
259
print_hex_dump_debug("PUDA: PUDA CQE", DUMP_PREFIX_OFFSET, 16, 8, cqe,
drivers/infiniband/hw/irdma/puda.c
274
get_64bit_val(cqe, 0, &qword0);
drivers/infiniband/hw/irdma/puda.c
275
get_64bit_val(cqe, 16, &qword2);
drivers/infiniband/hw/irdma/puda.c
282
get_64bit_val(cqe, 8, &comp_ctx);
drivers/infiniband/hw/irdma/uk.c
1118
__le64 *cqe;
drivers/infiniband/hw/irdma/uk.c
1123
cqe = IRDMA_GET_CURRENT_EXTENDED_CQ_ELEM(cq);
drivers/infiniband/hw/irdma/uk.c
1125
cqe = IRDMA_GET_CURRENT_CQ_ELEM(cq);
drivers/infiniband/hw/irdma/uk.c
1127
get_64bit_val(cqe, 24, &qword3);
drivers/infiniband/hw/irdma/uk.c
1142
__le64 *cqe;
drivers/infiniband/hw/irdma/uk.c
1156
cqe = IRDMA_GET_CURRENT_EXTENDED_CQ_ELEM(cq);
drivers/infiniband/hw/irdma/uk.c
1158
cqe = IRDMA_GET_CURRENT_CQ_ELEM(cq);
drivers/infiniband/hw/irdma/uk.c
1160
get_64bit_val(cqe, 24, &qword3);
drivers/infiniband/hw/irdma/uk.c
1174
ext_cqe = (__le64 *)((u8 *)cqe + 32);
drivers/infiniband/hw/irdma/uk.c
1223
get_64bit_val(cqe, 8, &comp_ctx);
drivers/infiniband/hw/irdma/uk.c
1225
get_64bit_val(cqe, 40, (u64 *)&qp);
drivers/infiniband/hw/irdma/uk.c
1241
set_64bit_val(cqe, 24, qword3);
drivers/infiniband/hw/irdma/uk.c
1265
get_64bit_val(cqe, 0, &qword0);
drivers/infiniband/hw/irdma/uk.c
1266
get_64bit_val(cqe, 16, &qword2);
drivers/infiniband/hw/irdma/uk.c
1271
get_64bit_val(cqe, 8, &comp_ctx);
drivers/infiniband/hw/irdma/uk.c
1288
get_64bit_val(cqe, 8, &info->wr_id);
drivers/infiniband/hw/irdma/uk.c
1431
set_64bit_val(cqe, 24, qword3);
drivers/infiniband/hw/irdma/uk.c
1778
__le64 *cqe;
drivers/infiniband/hw/irdma/uk.c
1787
cqe = ((struct irdma_extended_cqe *)(cq->cq_base))[cq_head].buf;
drivers/infiniband/hw/irdma/uk.c
1789
cqe = cq->cq_base[cq_head].buf;
drivers/infiniband/hw/irdma/uk.c
1790
get_64bit_val(cqe, 24, &qword3);
drivers/infiniband/hw/irdma/uk.c
1799
get_64bit_val(cqe, 8, &comp_ctx);
drivers/infiniband/hw/irdma/uk.c
1801
set_64bit_val(cqe, 8, 0);
drivers/infiniband/hw/irdma/verbs.c
2153
ibcq->cqe = info.cq_size - 1;
drivers/infiniband/hw/irdma/verbs.c
2491
int entries = attr->cqe;
drivers/infiniband/hw/irdma/verbs.c
2522
iwcq->ibcq.cqe = info.cq_uk_init_info.cq_size;
drivers/infiniband/hw/mana/cq.c
189
static inline void handle_ud_sq_cqe(struct mana_ib_qp *qp, struct gdma_comp *cqe)
drivers/infiniband/hw/mana/cq.c
191
struct mana_rdma_cqe *rdma_cqe = (struct mana_rdma_cqe *)cqe->cqe_data;
drivers/infiniband/hw/mana/cq.c
205
static inline void handle_ud_rq_cqe(struct mana_ib_qp *qp, struct gdma_comp *cqe)
drivers/infiniband/hw/mana/cq.c
207
struct mana_rdma_cqe *rdma_cqe = (struct mana_rdma_cqe *)cqe->cqe_data;
drivers/infiniband/hw/mana/cq.c
223
static void mana_handle_cqe(struct mana_ib_dev *mdev, struct gdma_comp *cqe)
drivers/infiniband/hw/mana/cq.c
225
struct mana_ib_qp *qp = mana_get_qp_ref(mdev, cqe->wq_num, cqe->is_sq);
drivers/infiniband/hw/mana/cq.c
231
if (cqe->is_sq)
drivers/infiniband/hw/mana/cq.c
232
handle_ud_sq_cqe(qp, cqe);
drivers/infiniband/hw/mana/cq.c
234
handle_ud_rq_cqe(qp, cqe);
drivers/infiniband/hw/mana/cq.c
39
if ((!is_rnic_cq && attr->cqe > mdev->adapter_caps.max_qp_wr) ||
drivers/infiniband/hw/mana/cq.c
40
attr->cqe > U32_MAX / COMP_ENTRY_SIZE) {
drivers/infiniband/hw/mana/cq.c
41
ibdev_dbg(ibdev, "CQE %d exceeding limit\n", attr->cqe);
drivers/infiniband/hw/mana/cq.c
45
cq->cqe = attr->cqe;
drivers/infiniband/hw/mana/cq.c
46
err = mana_ib_create_queue(mdev, ucmd.buf_addr, cq->cqe * COMP_ENTRY_SIZE,
drivers/infiniband/hw/mana/cq.c
57
if (attr->cqe > U32_MAX / COMP_ENTRY_SIZE / 2 + 1) {
drivers/infiniband/hw/mana/cq.c
58
ibdev_dbg(ibdev, "CQE %d exceeding limit\n", attr->cqe);
drivers/infiniband/hw/mana/cq.c
61
buf_size = MANA_PAGE_ALIGN(roundup_pow_of_two(attr->cqe * COMP_ENTRY_SIZE));
drivers/infiniband/hw/mana/cq.c
62
cq->cqe = buf_size / COMP_ENTRY_SIZE;
drivers/infiniband/hw/mana/mana_ib.h
146
int cqe;
drivers/infiniband/hw/mana/qp.c
189
cq_spec.queue_size = cq->cqe * COMP_ENTRY_SIZE;
drivers/infiniband/hw/mana/qp.c
340
cq_spec.queue_size = send_cq->cqe * COMP_ENTRY_SIZE;
drivers/infiniband/hw/mlx4/cq.c
133
static void mlx4_ib_free_cq_buf(struct mlx4_ib_dev *dev, struct mlx4_ib_cq_buf *buf, int cqe)
drivers/infiniband/hw/mlx4/cq.c
135
mlx4_buf_free(dev->dev, (cqe + 1) * buf->entry_size, &buf->buf);
drivers/infiniband/hw/mlx4/cq.c
140
struct ib_umem **umem, u64 buf_addr, int cqe)
drivers/infiniband/hw/mlx4/cq.c
147
*umem = ib_umem_get(&dev->ib_dev, buf_addr, cqe * cqe_size,
drivers/infiniband/hw/mlx4/cq.c
183
int entries = attr->cqe;
drivers/infiniband/hw/mlx4/cq.c
200
cq->ibcq.cqe = entries - 1;
drivers/infiniband/hw/mlx4/cq.c
286
mlx4_ib_free_cq_buf(dev, &cq->buf, cq->ibcq.cqe);
drivers/infiniband/hw/mlx4/cq.c
314
cq->resize_buf->cqe = entries - 1;
drivers/infiniband/hw/mlx4/cq.c
343
cq->resize_buf->cqe = entries - 1;
drivers/infiniband/hw/mlx4/cq.c
361
struct mlx4_cqe *cqe, *new_cqe;
drivers/infiniband/hw/mlx4/cq.c
367
cqe = get_cqe(cq, i & cq->ibcq.cqe);
drivers/infiniband/hw/mlx4/cq.c
368
cqe += cqe_inc;
drivers/infiniband/hw/mlx4/cq.c
370
while ((cqe->owner_sr_opcode & MLX4_CQE_OPCODE_MASK) != MLX4_CQE_OPCODE_RESIZE) {
drivers/infiniband/hw/mlx4/cq.c
372
(i + 1) & cq->resize_buf->cqe);
drivers/infiniband/hw/mlx4/cq.c
373
memcpy(new_cqe, get_cqe(cq, i & cq->ibcq.cqe), cqe_size);
drivers/infiniband/hw/mlx4/cq.c
376
new_cqe->owner_sr_opcode = (cqe->owner_sr_opcode & ~MLX4_CQE_OWNER_MASK) |
drivers/infiniband/hw/mlx4/cq.c
377
(((i + 1) & (cq->resize_buf->cqe + 1)) ? MLX4_CQE_OWNER_MASK : 0);
drivers/infiniband/hw/mlx4/cq.c
378
cqe = get_cqe(cq, ++i & cq->ibcq.cqe);
drivers/infiniband/hw/mlx4/cq.c
379
cqe += cqe_inc;
drivers/infiniband/hw/mlx4/cq.c
399
if (entries == ibcq->cqe + 1) {
drivers/infiniband/hw/mlx4/cq.c
435
cq->ibcq.cqe = cq->resize_buf->cqe;
drivers/infiniband/hw/mlx4/cq.c
450
tmp_cqe = cq->ibcq.cqe;
drivers/infiniband/hw/mlx4/cq.c
452
cq->ibcq.cqe = cq->resize_buf->cqe;
drivers/infiniband/hw/mlx4/cq.c
469
cq->resize_buf->cqe);
drivers/infiniband/hw/mlx4/cq.c
498
mlx4_ib_free_cq_buf(dev, &mcq->buf, cq->cqe);
drivers/infiniband/hw/mlx4/cq.c
505
static void dump_cqe(void *cqe)
drivers/infiniband/hw/mlx4/cq.c
507
__be32 *buf = cqe;
drivers/infiniband/hw/mlx4/cq.c
515
static void mlx4_ib_handle_error_cqe(struct mlx4_err_cqe *cqe,
drivers/infiniband/hw/mlx4/cq.c
518
if (cqe->syndrome == MLX4_CQE_SYNDROME_LOCAL_QP_OP_ERR) {
drivers/infiniband/hw/mlx4/cq.c
522
be32_to_cpu(cqe->my_qpn), be16_to_cpu(cqe->wqe_index),
drivers/infiniband/hw/mlx4/cq.c
523
cqe->vendor_err_syndrome,
drivers/infiniband/hw/mlx4/cq.c
524
cqe->owner_sr_opcode & ~MLX4_CQE_OWNER_MASK);
drivers/infiniband/hw/mlx4/cq.c
525
dump_cqe(cqe);
drivers/infiniband/hw/mlx4/cq.c
528
switch (cqe->syndrome) {
drivers/infiniband/hw/mlx4/cq.c
573
wc->vendor_err = cqe->vendor_err_syndrome;
drivers/infiniband/hw/mlx4/cq.c
586
unsigned tail, struct mlx4_cqe *cqe, int is_eth)
drivers/infiniband/hw/mlx4/cq.c
665
struct mlx4_cqe *cqe;
drivers/infiniband/hw/mlx4/cq.c
678
cqe = next_cqe_sw(cq);
drivers/infiniband/hw/mlx4/cq.c
679
if (!cqe)
drivers/infiniband/hw/mlx4/cq.c
683
cqe++;
drivers/infiniband/hw/mlx4/cq.c
693
is_send = cqe->owner_sr_opcode & MLX4_CQE_IS_SEND_MASK;
drivers/infiniband/hw/mlx4/cq.c
694
is_error = (cqe->owner_sr_opcode & MLX4_CQE_OPCODE_MASK) ==
drivers/infiniband/hw/mlx4/cq.c
698
if (unlikely((cqe->owner_sr_opcode & MLX4_CQE_OPCODE_MASK) == MLX4_CQE_OPCODE_RESIZE)) {
drivers/infiniband/hw/mlx4/cq.c
702
mlx4_ib_free_cq_buf(dev, &cq->buf, cq->ibcq.cqe);
drivers/infiniband/hw/mlx4/cq.c
704
cq->ibcq.cqe = cq->resize_buf->cqe;
drivers/infiniband/hw/mlx4/cq.c
714
(be32_to_cpu(cqe->vlan_my_qpn) & MLX4_CQE_QPN_MASK) != (*cur_qp)->mqp.qpn) {
drivers/infiniband/hw/mlx4/cq.c
721
be32_to_cpu(cqe->vlan_my_qpn));
drivers/infiniband/hw/mlx4/cq.c
729
g_mlpath_rqpn = be32_to_cpu(cqe->g_mlpath_rqpn);
drivers/infiniband/hw/mlx4/cq.c
739
wqe_ctr = be16_to_cpu(cqe->wqe_index);
drivers/infiniband/hw/mlx4/cq.c
746
wqe_ctr = be16_to_cpu(cqe->wqe_index);
drivers/infiniband/hw/mlx4/cq.c
751
wqe_ctr = be16_to_cpu(cqe->wqe_index);
drivers/infiniband/hw/mlx4/cq.c
762
mlx4_ib_handle_error_cqe((struct mlx4_err_cqe *) cqe, wc);
drivers/infiniband/hw/mlx4/cq.c
770
switch (cqe->owner_sr_opcode & MLX4_CQE_OPCODE_MASK) {
drivers/infiniband/hw/mlx4/cq.c
786
wc->byte_len = be32_to_cpu(cqe->byte_cnt);
drivers/infiniband/hw/mlx4/cq.c
81
struct mlx4_cqe *cqe = get_cqe(cq, n & cq->ibcq.cqe);
drivers/infiniband/hw/mlx4/cq.c
815
wc->byte_len = be32_to_cpu(cqe->byte_cnt);
drivers/infiniband/hw/mlx4/cq.c
817
switch (cqe->owner_sr_opcode & MLX4_CQE_OPCODE_MASK) {
drivers/infiniband/hw/mlx4/cq.c
82
struct mlx4_cqe *tcqe = ((cq->buf.entry_size == 64) ? (cqe + 1) : cqe);
drivers/infiniband/hw/mlx4/cq.c
821
wc->ex.imm_data = cqe->immed_rss_invalid;
drivers/infiniband/hw/mlx4/cq.c
826
wc->ex.invalidate_rkey = be32_to_cpu(cqe->immed_rss_invalid);
drivers/infiniband/hw/mlx4/cq.c
835
wc->ex.imm_data = cqe->immed_rss_invalid;
drivers/infiniband/hw/mlx4/cq.c
846
use_tunnel_data(*cur_qp, cq, wc, tail, cqe,
drivers/infiniband/hw/mlx4/cq.c
85
!!(n & (cq->ibcq.cqe + 1))) ? NULL : cqe;
drivers/infiniband/hw/mlx4/cq.c
852
g_mlpath_rqpn = be32_to_cpu(cqe->g_mlpath_rqpn);
drivers/infiniband/hw/mlx4/cq.c
856
wc->pkey_index = be32_to_cpu(cqe->immed_rss_invalid) & 0x7f;
drivers/infiniband/hw/mlx4/cq.c
857
wc->wc_flags |= mlx4_ib_ipoib_csum_ok(cqe->status,
drivers/infiniband/hw/mlx4/cq.c
858
cqe->badfcs_enc,
drivers/infiniband/hw/mlx4/cq.c
859
cqe->checksum) ? IB_WC_IP_CSUM_OK : 0;
drivers/infiniband/hw/mlx4/cq.c
862
wc->sl = be16_to_cpu(cqe->sl_vid) >> 13;
drivers/infiniband/hw/mlx4/cq.c
863
if (be32_to_cpu(cqe->vlan_my_qpn) &
drivers/infiniband/hw/mlx4/cq.c
865
wc->vlan_id = be16_to_cpu(cqe->sl_vid) &
drivers/infiniband/hw/mlx4/cq.c
870
memcpy(wc->smac, cqe->smac, ETH_ALEN);
drivers/infiniband/hw/mlx4/cq.c
873
wc->slid = be16_to_cpu(cqe->rlid);
drivers/infiniband/hw/mlx4/cq.c
874
wc->sl = be16_to_cpu(cqe->sl_vid) >> 12;
drivers/infiniband/hw/mlx4/cq.c
924
struct mlx4_cqe *cqe, *dest;
drivers/infiniband/hw/mlx4/cq.c
936
if (prod_index == cq->mcq.cons_index + cq->ibcq.cqe)
drivers/infiniband/hw/mlx4/cq.c
944
cqe = get_cqe(cq, prod_index & cq->ibcq.cqe);
drivers/infiniband/hw/mlx4/cq.c
945
cqe += cqe_inc;
drivers/infiniband/hw/mlx4/cq.c
947
if ((be32_to_cpu(cqe->vlan_my_qpn) & MLX4_CQE_QPN_MASK) == qpn) {
drivers/infiniband/hw/mlx4/cq.c
948
if (srq && !(cqe->owner_sr_opcode & MLX4_CQE_IS_SEND_MASK))
drivers/infiniband/hw/mlx4/cq.c
949
mlx4_ib_free_srq_wqe(srq, be16_to_cpu(cqe->wqe_index));
drivers/infiniband/hw/mlx4/cq.c
952
dest = get_cqe(cq, (prod_index + nfreed) & cq->ibcq.cqe);
drivers/infiniband/hw/mlx4/cq.c
956
memcpy(dest, cqe, sizeof *cqe);
drivers/infiniband/hw/mlx4/mad.c
2010
cq_attr.cqe = cq_size;
drivers/infiniband/hw/mlx4/main.c
1240
cq_attr.cqe = 1;
drivers/infiniband/hw/mlx4/mlx4_ib.h
113
int cqe;
drivers/infiniband/hw/mlx4/qp.c
4390
struct ib_cqe cqe;
drivers/infiniband/hw/mlx4/qp.c
4396
struct mlx4_ib_drain_cqe *cqe = container_of(wc->wr_cqe,
drivers/infiniband/hw/mlx4/qp.c
4398
cqe);
drivers/infiniband/hw/mlx4/qp.c
4400
complete(&cqe->done);
drivers/infiniband/hw/mlx4/qp.c
4462
{ .wr_cqe = &sdrain.cqe, },
drivers/infiniband/hw/mlx4/qp.c
4476
sdrain.cqe.done = mlx4_ib_drain_qp_done;
drivers/infiniband/hw/mlx4/qp.c
4505
rwr.wr_cqe = &rdrain.cqe;
drivers/infiniband/hw/mlx4/qp.c
4506
rdrain.cqe.done = mlx4_ib_drain_qp_done;
drivers/infiniband/hw/mlx5/cq.c
1097
void *cqe, *dest;
drivers/infiniband/hw/mlx5/cq.c
1112
if (prod_index == cq->mcq.cons_index + cq->ibcq.cqe)
drivers/infiniband/hw/mlx5/cq.c
1119
cqe = get_cqe(cq, prod_index & cq->ibcq.cqe);
drivers/infiniband/hw/mlx5/cq.c
1120
cqe64 = (cq->mcq.cqe_sz == 64) ? cqe : cqe + 64;
drivers/infiniband/hw/mlx5/cq.c
1126
dest = get_cqe(cq, (prod_index + nfreed) & cq->ibcq.cqe);
drivers/infiniband/hw/mlx5/cq.c
1129
memcpy(dest, cqe, cq->mcq.cqe_sz);
drivers/infiniband/hw/mlx5/cq.c
120
static void handle_good_req(struct ib_wc *wc, struct mlx5_cqe64 *cqe,
drivers/infiniband/hw/mlx5/cq.c
124
switch (be32_to_cpu(cqe->sop_drop_qpn) >> 24) {
drivers/infiniband/hw/mlx5/cq.c
1317
if (entries == ibcq->cqe + 1)
drivers/infiniband/hw/mlx5/cq.c
1391
cq->ibcq.cqe = entries - 1;
drivers/infiniband/hw/mlx5/cq.c
140
wc->byte_len = be32_to_cpu(cqe->byte_cnt);
drivers/infiniband/hw/mlx5/cq.c
1410
cq->ibcq.cqe = entries - 1;
drivers/infiniband/hw/mlx5/cq.c
169
static void handle_responder(struct ib_wc *wc, struct mlx5_cqe64 *cqe,
drivers/infiniband/hw/mlx5/cq.c
185
msrq = mlx5_cmd_get_srq(dev, be32_to_cpu(cqe->srqn));
drivers/infiniband/hw/mlx5/cq.c
192
wqe_ctr = be16_to_cpu(cqe->wqe_counter);
drivers/infiniband/hw/mlx5/cq.c
203
wc->byte_len = be32_to_cpu(cqe->byte_cnt);
drivers/infiniband/hw/mlx5/cq.c
205
switch (get_cqe_opcode(cqe)) {
drivers/infiniband/hw/mlx5/cq.c
209
wc->ex.imm_data = cqe->immediate;
drivers/infiniband/hw/mlx5/cq.c
214
if (unlikely(!((cqe->hds_ip_ext & CQE_L3_OK) &&
drivers/infiniband/hw/mlx5/cq.c
215
(cqe->hds_ip_ext & CQE_L4_OK))))
drivers/infiniband/hw/mlx5/cq.c
221
wc->ex.imm_data = cqe->immediate;
drivers/infiniband/hw/mlx5/cq.c
226
wc->ex.invalidate_rkey = be32_to_cpu(cqe->inval_rkey);
drivers/infiniband/hw/mlx5/cq.c
229
wc->src_qp = be32_to_cpu(cqe->flags_rqpn) & 0xffffff;
drivers/infiniband/hw/mlx5/cq.c
230
wc->dlid_path_bits = cqe->ml_path;
drivers/infiniband/hw/mlx5/cq.c
231
g = (be32_to_cpu(cqe->flags_rqpn) >> 28) & 3;
drivers/infiniband/hw/mlx5/cq.c
234
u16 pkey = be32_to_cpu(cqe->pkey) & 0xffff;
drivers/infiniband/hw/mlx5/cq.c
243
wc->slid = be16_to_cpu(cqe->slid);
drivers/infiniband/hw/mlx5/cq.c
244
wc->sl = (be32_to_cpu(cqe->flags_rqpn) >> 24) & 0xf;
drivers/infiniband/hw/mlx5/cq.c
249
vlan_present = cqe->l4_l3_hdr_type & 0x1;
drivers/infiniband/hw/mlx5/cq.c
250
roce_packet_type = (be32_to_cpu(cqe->flags_rqpn) >> 24) & 0x3;
drivers/infiniband/hw/mlx5/cq.c
252
wc->vlan_id = (be16_to_cpu(cqe->vlan_info)) & 0xfff;
drivers/infiniband/hw/mlx5/cq.c
253
wc->sl = (be16_to_cpu(cqe->vlan_info) >> 13) & 0x7;
drivers/infiniband/hw/mlx5/cq.c
273
static void dump_cqe(struct mlx5_ib_dev *dev, struct mlx5_err_cqe *cqe,
drivers/infiniband/hw/mlx5/cq.c
279
cqe, sizeof(*cqe), false);
drivers/infiniband/hw/mlx5/cq.c
283
struct mlx5_err_cqe *cqe,
drivers/infiniband/hw/mlx5/cq.c
288
switch (cqe->syndrome) {
drivers/infiniband/hw/mlx5/cq.c
339
wc->vendor_err = cqe->vendor_err_synd;
drivers/infiniband/hw/mlx5/cq.c
341
dump_cqe(dev, cqe, wc, dump);
drivers/infiniband/hw/mlx5/cq.c
365
static void get_sig_err_item(struct mlx5_sig_err_cqe *cqe,
drivers/infiniband/hw/mlx5/cq.c
368
u16 syndrome = be16_to_cpu(cqe->syndrome);
drivers/infiniband/hw/mlx5/cq.c
376
item->expected = be32_to_cpu(cqe->expected_trans_sig) >> 16;
drivers/infiniband/hw/mlx5/cq.c
377
item->actual = be32_to_cpu(cqe->actual_trans_sig) >> 16;
drivers/infiniband/hw/mlx5/cq.c
381
item->expected = be32_to_cpu(cqe->expected_reftag);
drivers/infiniband/hw/mlx5/cq.c
382
item->actual = be32_to_cpu(cqe->actual_reftag);
drivers/infiniband/hw/mlx5/cq.c
386
item->expected = be32_to_cpu(cqe->expected_trans_sig) & 0xffff;
drivers/infiniband/hw/mlx5/cq.c
387
item->actual = be32_to_cpu(cqe->actual_trans_sig) & 0xffff;
drivers/infiniband/hw/mlx5/cq.c
393
item->sig_err_offset = be64_to_cpu(cqe->err_offset);
drivers/infiniband/hw/mlx5/cq.c
394
item->key = be32_to_cpu(cqe->mkey);
drivers/infiniband/hw/mlx5/cq.c
462
void *cqe;
drivers/infiniband/hw/mlx5/cq.c
466
cqe = next_cqe_sw(cq);
drivers/infiniband/hw/mlx5/cq.c
467
if (!cqe)
drivers/infiniband/hw/mlx5/cq.c
470
cqe64 = (cq->mcq.cqe_sz == 64) ? cqe : cqe + 64;
drivers/infiniband/hw/mlx5/cq.c
84
void *cqe = get_cqe(cq, n & cq->ibcq.cqe);
drivers/infiniband/hw/mlx5/cq.c
87
cqe64 = (cq->mcq.cqe_sz == 64) ? cqe : cqe + 64;
drivers/infiniband/hw/mlx5/cq.c
877
void *cqe;
drivers/infiniband/hw/mlx5/cq.c
881
cqe = mlx5_frag_buf_get_wqe(&buf->fbc, i);
drivers/infiniband/hw/mlx5/cq.c
882
cqe64 = buf->cqe_size == 64 ? cqe : cqe + 64;
drivers/infiniband/hw/mlx5/cq.c
90
!((cqe64->op_own & MLX5_CQE_OWNER_MASK) ^ !!(n & (cq->ibcq.cqe + 1)))) {
drivers/infiniband/hw/mlx5/cq.c
91
return cqe;
drivers/infiniband/hw/mlx5/cq.c
957
int entries = attr->cqe;
drivers/infiniband/hw/mlx5/cq.c
981
cq->ibcq.cqe = entries - 1;
drivers/infiniband/hw/mlx5/gsi.c
36
struct ib_cqe cqe;
drivers/infiniband/hw/mlx5/gsi.c
386
gsi_wr->cqe.done = &handle_single_completion;
drivers/infiniband/hw/mlx5/gsi.c
387
wr->wr.wr_cqe = &gsi_wr->cqe;
drivers/infiniband/hw/mlx5/gsi.c
72
container_of(wc->wr_cqe, struct mlx5_ib_gsi_wr, cqe);
drivers/infiniband/hw/mlx5/main.c
3290
struct ib_cq_init_attr cq_attr = {.cqe = 1};
drivers/infiniband/hw/mlx5/mlx5_ib.h
760
struct ib_cqe cqe;
drivers/infiniband/hw/mlx5/qp.c
5714
struct ib_cqe cqe;
drivers/infiniband/hw/mlx5/qp.c
5720
struct mlx5_ib_drain_cqe *cqe = container_of(wc->wr_cqe,
drivers/infiniband/hw/mlx5/qp.c
5722
cqe);
drivers/infiniband/hw/mlx5/qp.c
5724
complete(&cqe->done);
drivers/infiniband/hw/mlx5/qp.c
5786
{ .wr_cqe = &sdrain.cqe, },
drivers/infiniband/hw/mlx5/qp.c
5800
sdrain.cqe.done = mlx5_ib_drain_qp_done;
drivers/infiniband/hw/mlx5/qp.c
5829
rwr.wr_cqe = &rdrain.cqe;
drivers/infiniband/hw/mlx5/qp.c
5830
rdrain.cqe.done = mlx5_ib_drain_qp_done;
drivers/infiniband/hw/mlx5/umr.c
237
static int mlx5r_umr_post_send(struct ib_qp *ibqp, u32 mkey, struct ib_cqe *cqe,
drivers/infiniband/hw/mlx5/umr.c
271
id.ib_cqe = cqe;
drivers/infiniband/hw/mlx5/umr.c
304
err = mlx5r_umr_post_send(umrc->qp, mkey, &umr_context->cqe, wqe,
drivers/infiniband/hw/mlx5/umr.c
341
container_of(wc->wr_cqe, struct mlx5_ib_umr_context, cqe);
drivers/infiniband/hw/mlx5/umr.c
349
context->cqe.done = mlx5r_umr_done;
drivers/infiniband/hw/mlx5/umr.c
381
err = mlx5r_umr_post_send(umrc->qp, mkey, &umr_context.cqe, wqe,
drivers/infiniband/hw/mlx5/umr.h
83
struct ib_cqe cqe;
drivers/infiniband/hw/mthca/mthca_cq.c
174
static inline struct mthca_cqe *cqe_sw(struct mthca_cqe *cqe)
drivers/infiniband/hw/mthca/mthca_cq.c
176
return MTHCA_CQ_ENTRY_OWNER_HW & cqe->owner ? NULL : cqe;
drivers/infiniband/hw/mthca/mthca_cq.c
181
return cqe_sw(get_cqe(cq, cq->cons_index & cq->ibcq.cqe));
drivers/infiniband/hw/mthca/mthca_cq.c
184
static inline void set_cqe_hw(struct mthca_cqe *cqe)
drivers/infiniband/hw/mthca/mthca_cq.c
186
cqe->owner = MTHCA_CQ_ENTRY_OWNER_HW;
drivers/infiniband/hw/mthca/mthca_cq.c
191
__be32 *cqe = cqe_ptr;
drivers/infiniband/hw/mthca/mthca_cq.c
193
(void) cqe; /* avoid warning if mthca_dbg compiled away... */
drivers/infiniband/hw/mthca/mthca_cq.c
195
be32_to_cpu(cqe[0]), be32_to_cpu(cqe[1]), be32_to_cpu(cqe[2]),
drivers/infiniband/hw/mthca/mthca_cq.c
196
be32_to_cpu(cqe[3]), be32_to_cpu(cqe[4]), be32_to_cpu(cqe[5]),
drivers/infiniband/hw/mthca/mthca_cq.c
197
be32_to_cpu(cqe[6]), be32_to_cpu(cqe[7]));
drivers/infiniband/hw/mthca/mthca_cq.c
264
static inline int is_recv_cqe(struct mthca_cqe *cqe)
drivers/infiniband/hw/mthca/mthca_cq.c
266
if ((cqe->opcode & MTHCA_ERROR_CQE_OPCODE_MASK) ==
drivers/infiniband/hw/mthca/mthca_cq.c
268
return !(cqe->opcode & 0x01);
drivers/infiniband/hw/mthca/mthca_cq.c
270
return !(cqe->is_send & 0x80);
drivers/infiniband/hw/mthca/mthca_cq.c
276
struct mthca_cqe *cqe;
drivers/infiniband/hw/mthca/mthca_cq.c
290
cqe_sw(get_cqe(cq, prod_index & cq->ibcq.cqe));
drivers/infiniband/hw/mthca/mthca_cq.c
292
if (prod_index == cq->cons_index + cq->ibcq.cqe)
drivers/infiniband/hw/mthca/mthca_cq.c
304
cqe = get_cqe(cq, prod_index & cq->ibcq.cqe);
drivers/infiniband/hw/mthca/mthca_cq.c
305
if (cqe->my_qpn == cpu_to_be32(qpn)) {
drivers/infiniband/hw/mthca/mthca_cq.c
306
if (srq && is_recv_cqe(cqe))
drivers/infiniband/hw/mthca/mthca_cq.c
307
mthca_free_srq_wqe(srq, be32_to_cpu(cqe->wqe));
drivers/infiniband/hw/mthca/mthca_cq.c
310
memcpy(get_cqe(cq, (prod_index + nfreed) & cq->ibcq.cqe),
drivers/infiniband/hw/mthca/mthca_cq.c
311
cqe, MTHCA_CQ_ENTRY_SIZE);
drivers/infiniband/hw/mthca/mthca_cq.c
316
set_cqe_hw(get_cqe(cq, (cq->cons_index + i) & cq->ibcq.cqe));
drivers/infiniband/hw/mthca/mthca_cq.c
336
cq->ibcq.cqe < cq->resize_buf->cqe) {
drivers/infiniband/hw/mthca/mthca_cq.c
337
cq->cons_index &= cq->ibcq.cqe;
drivers/infiniband/hw/mthca/mthca_cq.c
338
if (cqe_sw(get_cqe(cq, cq->ibcq.cqe)))
drivers/infiniband/hw/mthca/mthca_cq.c
339
cq->cons_index -= cq->ibcq.cqe + 1;
drivers/infiniband/hw/mthca/mthca_cq.c
342
for (i = cq->cons_index; cqe_sw(get_cqe(cq, i & cq->ibcq.cqe)); ++i)
drivers/infiniband/hw/mthca/mthca_cq.c
344
i & cq->resize_buf->cqe),
drivers/infiniband/hw/mthca/mthca_cq.c
345
get_cqe(cq, i & cq->ibcq.cqe), MTHCA_CQ_ENTRY_SIZE);
drivers/infiniband/hw/mthca/mthca_cq.c
366
void mthca_free_cq_buf(struct mthca_dev *dev, struct mthca_cq_buf *buf, int cqe)
drivers/infiniband/hw/mthca/mthca_cq.c
368
mthca_buf_free(dev, (cqe + 1) * MTHCA_CQ_ENTRY_SIZE, &buf->queue,
drivers/infiniband/hw/mthca/mthca_cq.c
374
struct mthca_err_cqe *cqe,
drivers/infiniband/hw/mthca/mthca_cq.c
380
if (cqe->syndrome == SYNDROME_LOCAL_QP_OP_ERR) {
drivers/infiniband/hw/mthca/mthca_cq.c
383
be32_to_cpu(cqe->my_qpn), be32_to_cpu(cqe->wqe),
drivers/infiniband/hw/mthca/mthca_cq.c
385
dump_cqe(dev, cqe);
drivers/infiniband/hw/mthca/mthca_cq.c
392
switch (cqe->syndrome) {
drivers/infiniband/hw/mthca/mthca_cq.c
452
entry->vendor_err = cqe->vendor_err;
drivers/infiniband/hw/mthca/mthca_cq.c
468
if (!(new_wqe & cpu_to_be32(0x3f)) || (!cqe->db_cnt && dbd))
drivers/infiniband/hw/mthca/mthca_cq.c
471
be16_add_cpu(&cqe->db_cnt, -dbd);
drivers/infiniband/hw/mthca/mthca_cq.c
472
cqe->wqe = new_wqe;
drivers/infiniband/hw/mthca/mthca_cq.c
473
cqe->syndrome = SYNDROME_WR_FLUSH_ERR;
drivers/infiniband/hw/mthca/mthca_cq.c
485
struct mthca_cqe *cqe;
drivers/infiniband/hw/mthca/mthca_cq.c
493
cqe = next_cqe_sw(cq);
drivers/infiniband/hw/mthca/mthca_cq.c
494
if (!cqe)
drivers/infiniband/hw/mthca/mthca_cq.c
505
cq->cqn, cq->cons_index, be32_to_cpu(cqe->my_qpn),
drivers/infiniband/hw/mthca/mthca_cq.c
506
be32_to_cpu(cqe->wqe));
drivers/infiniband/hw/mthca/mthca_cq.c
507
dump_cqe(dev, cqe);
drivers/infiniband/hw/mthca/mthca_cq.c
510
is_error = (cqe->opcode & MTHCA_ERROR_CQE_OPCODE_MASK) ==
drivers/infiniband/hw/mthca/mthca_cq.c
512
is_send = is_error ? cqe->opcode & 0x01 : cqe->is_send & 0x80;
drivers/infiniband/hw/mthca/mthca_cq.c
514
if (!*cur_qp || be32_to_cpu(cqe->my_qpn) != (*cur_qp)->qpn) {
drivers/infiniband/hw/mthca/mthca_cq.c
521
be32_to_cpu(cqe->my_qpn) &
drivers/infiniband/hw/mthca/mthca_cq.c
525
be32_to_cpu(cqe->my_qpn) & 0xffffff);
drivers/infiniband/hw/mthca/mthca_cq.c
535
wqe_index = ((be32_to_cpu(cqe->wqe) - (*cur_qp)->send_wqe_offset)
drivers/infiniband/hw/mthca/mthca_cq.c
541
u32 wqe = be32_to_cpu(cqe->wqe);
drivers/infiniband/hw/mthca/mthca_cq.c
549
wqe = be32_to_cpu(cqe->wqe);
drivers/infiniband/hw/mthca/mthca_cq.c
572
(struct mthca_err_cqe *) cqe,
drivers/infiniband/hw/mthca/mthca_cq.c
579
switch (cqe->opcode) {
drivers/infiniband/hw/mthca/mthca_cq.c
596
entry->byte_len = be32_to_cpu(cqe->byte_cnt);
drivers/infiniband/hw/mthca/mthca_cq.c
611
entry->byte_len = be32_to_cpu(cqe->byte_cnt);
drivers/infiniband/hw/mthca/mthca_cq.c
612
switch (cqe->opcode & 0x1f) {
drivers/infiniband/hw/mthca/mthca_cq.c
616
entry->ex.imm_data = cqe->imm_etype_pkey_eec;
drivers/infiniband/hw/mthca/mthca_cq.c
622
entry->ex.imm_data = cqe->imm_etype_pkey_eec;
drivers/infiniband/hw/mthca/mthca_cq.c
630
entry->slid = be16_to_cpu(cqe->rlid);
drivers/infiniband/hw/mthca/mthca_cq.c
631
entry->sl = cqe->sl_ipok >> 4;
drivers/infiniband/hw/mthca/mthca_cq.c
632
entry->src_qp = be32_to_cpu(cqe->rqpn) & 0xffffff;
drivers/infiniband/hw/mthca/mthca_cq.c
633
entry->dlid_path_bits = cqe->g_mlpath & 0x7f;
drivers/infiniband/hw/mthca/mthca_cq.c
634
entry->pkey_index = be32_to_cpu(cqe->imm_etype_pkey_eec) >> 16;
drivers/infiniband/hw/mthca/mthca_cq.c
635
entry->wc_flags |= cqe->g_mlpath & 0x80 ? IB_WC_GRH : 0;
drivers/infiniband/hw/mthca/mthca_cq.c
636
checksum = (be32_to_cpu(cqe->rqpn) >> 24) |
drivers/infiniband/hw/mthca/mthca_cq.c
637
((be32_to_cpu(cqe->my_ee) >> 16) & 0xff00);
drivers/infiniband/hw/mthca/mthca_cq.c
638
entry->wc_flags |= (cqe->sl_ipok & 1 && checksum == 0xffff) ?
drivers/infiniband/hw/mthca/mthca_cq.c
646
set_cqe_hw(cqe);
drivers/infiniband/hw/mthca/mthca_cq.c
698
cq->cons_index &= cq->ibcq.cqe;
drivers/infiniband/hw/mthca/mthca_cq.c
701
cq->cons_index & cq->resize_buf->cqe))) {
drivers/infiniband/hw/mthca/mthca_cq.c
706
tcqe = cq->ibcq.cqe;
drivers/infiniband/hw/mthca/mthca_cq.c
708
cq->ibcq.cqe = cq->resize_buf->cqe;
drivers/infiniband/hw/mthca/mthca_cq.c
711
cq->resize_buf->cqe = tcqe;
drivers/infiniband/hw/mthca/mthca_cq.c
776
cq->ibcq.cqe = nent - 1;
drivers/infiniband/hw/mthca/mthca_cq.c
867
mthca_free_cq_buf(dev, &cq->buf, cq->ibcq.cqe);
drivers/infiniband/hw/mthca/mthca_cq.c
941
mthca_free_cq_buf(dev, &cq->buf, cq->ibcq.cqe);
drivers/infiniband/hw/mthca/mthca_dev.h
499
void mthca_free_cq_buf(struct mthca_dev *dev, struct mthca_cq_buf *buf, int cqe);
drivers/infiniband/hw/mthca/mthca_provider.c
582
int entries = attr->cqe;
drivers/infiniband/hw/mthca/mthca_provider.c
689
cq->resize_buf->cqe = entries - 1;
drivers/infiniband/hw/mthca/mthca_provider.c
712
if (entries == ibcq->cqe + 1) {
drivers/infiniband/hw/mthca/mthca_provider.c
735
cq->resize_buf->cqe);
drivers/infiniband/hw/mthca/mthca_provider.c
752
tcqe = cq->ibcq.cqe;
drivers/infiniband/hw/mthca/mthca_provider.c
754
cq->ibcq.cqe = cq->resize_buf->cqe;
drivers/infiniband/hw/mthca/mthca_provider.c
757
tcqe = cq->resize_buf->cqe;
drivers/infiniband/hw/mthca/mthca_provider.c
766
ibcq->cqe = entries - 1;
drivers/infiniband/hw/mthca/mthca_provider.h
175
int cqe;
drivers/infiniband/hw/ocrdma/ocrdma.h
494
static inline int is_cqe_valid(struct ocrdma_cq *cq, struct ocrdma_cqe *cqe)
drivers/infiniband/hw/ocrdma/ocrdma.h
497
cqe_valid = le32_to_cpu(cqe->flags_status_srcqpn) & OCRDMA_CQE_VALID;
drivers/infiniband/hw/ocrdma/ocrdma.h
501
static inline int is_cqe_for_sq(struct ocrdma_cqe *cqe)
drivers/infiniband/hw/ocrdma/ocrdma.h
503
return (le32_to_cpu(cqe->flags_status_srcqpn) &
drivers/infiniband/hw/ocrdma/ocrdma.h
507
static inline int is_cqe_invalidated(struct ocrdma_cqe *cqe)
drivers/infiniband/hw/ocrdma/ocrdma.h
509
return (le32_to_cpu(cqe->flags_status_srcqpn) &
drivers/infiniband/hw/ocrdma/ocrdma.h
513
static inline int is_cqe_imm(struct ocrdma_cqe *cqe)
drivers/infiniband/hw/ocrdma/ocrdma.h
515
return (le32_to_cpu(cqe->flags_status_srcqpn) &
drivers/infiniband/hw/ocrdma/ocrdma.h
519
static inline int is_cqe_wr_imm(struct ocrdma_cqe *cqe)
drivers/infiniband/hw/ocrdma/ocrdma.h
521
return (le32_to_cpu(cqe->flags_status_srcqpn) &
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
123
struct ocrdma_mcqe *cqe = (struct ocrdma_mcqe *)
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
126
if (!(le32_to_cpu(cqe->valid_ae_cmpl_cons) & OCRDMA_MCQE_VALID_MASK))
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
128
return cqe;
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
676
struct ocrdma_ae_mcqe *cqe)
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
685
int type = (cqe->valid_ae_event & OCRDMA_AE_MCQE_EVENT_TYPE_MASK) >>
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
687
u16 qpid = cqe->qpvalid_qpid & OCRDMA_AE_MCQE_QPID_MASK;
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
688
u16 cqid = cqe->cqvalid_cqid & OCRDMA_AE_MCQE_CQID_MASK;
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
695
if (cqe->qpvalid_qpid & OCRDMA_AE_MCQE_QPVALID) {
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
705
if (cqe->cqvalid_cqid & OCRDMA_AE_MCQE_CQVALID) {
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
802
struct ocrdma_ae_mcqe *cqe)
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
805
int type = (cqe->valid_ae_event & OCRDMA_AE_MCQE_EVENT_TYPE_MASK) >>
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
810
evt = (struct ocrdma_ae_pvid_mcqe *)cqe;
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
828
struct ocrdma_ae_mcqe *cqe)
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
833
evt = (struct ocrdma_ae_lnkst_mcqe *)cqe;
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
846
struct ocrdma_ae_mcqe *cqe = ae_cqe;
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
847
u32 evt_code = (cqe->valid_ae_event & OCRDMA_AE_MCQE_EVENT_CODE_MASK) >>
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
851
ocrdma_process_link_state(dev, cqe);
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
854
ocrdma_dispatch_ibevent(dev, cqe);
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
857
ocrdma_process_grp5_aync(dev, cqe);
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
865
static void ocrdma_process_mcqe(struct ocrdma_dev *dev, struct ocrdma_mcqe *cqe)
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
867
if (dev->mqe_ctx.tag == cqe->tag_lo && dev->mqe_ctx.cmd_done == false) {
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
868
dev->mqe_ctx.cqe_status = (cqe->status &
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
871
(cqe->status & OCRDMA_MCQE_ESTATUS_MASK)
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
877
__func__, cqe->tag_lo, dev->mqe_ctx.tag);
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
883
struct ocrdma_mcqe *cqe;
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
886
cqe = ocrdma_get_mcqe(dev);
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
887
if (cqe == NULL)
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
889
ocrdma_le32_to_cpu(cqe, sizeof(*cqe));
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
891
if (cqe->valid_ae_cmpl_cons & OCRDMA_MCQE_AE_MASK)
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
892
ocrdma_process_acqe(dev, cqe);
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
893
else if (cqe->valid_ae_cmpl_cons & OCRDMA_MCQE_CMPL_MASK)
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
894
ocrdma_process_mcqe(dev, cqe);
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
895
memset(cqe, 0, sizeof(struct ocrdma_mcqe));
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1027
ibcq->cqe = new_cnt;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1038
struct ocrdma_cqe *cqe = NULL;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1040
cqe = cq->va;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1048
if (is_cqe_valid(cq, cqe))
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1050
cqe++;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1596
struct ocrdma_cqe *cqe;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1615
cqe = cq->va + cur_getp;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1620
qpn = cqe->cmn.qpn & OCRDMA_CQE_QPN_MASK;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1626
if (is_cqe_for_sq(cqe)) {
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1630
wqe_idx = (le32_to_cpu(cqe->rq.buftag_qpn) >>
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1646
cqe->cmn.qpn = 0;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2441
struct ocrdma_cqe *cqe)
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2443
if (is_cqe_for_sq(cqe)) {
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2444
cqe->flags_status_srcqpn = cpu_to_le32(le32_to_cpu(
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2445
cqe->flags_status_srcqpn) &
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2447
cqe->flags_status_srcqpn = cpu_to_le32(le32_to_cpu(
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2448
cqe->flags_status_srcqpn) |
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2453
cqe->flags_status_srcqpn = cpu_to_le32(le32_to_cpu(
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2454
cqe->flags_status_srcqpn) &
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2456
cqe->flags_status_srcqpn = cpu_to_le32(le32_to_cpu(
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2457
cqe->flags_status_srcqpn) |
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2461
cqe->flags_status_srcqpn = cpu_to_le32(le32_to_cpu(
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2462
cqe->flags_status_srcqpn) &
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2464
cqe->flags_status_srcqpn = cpu_to_le32(le32_to_cpu(
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2465
cqe->flags_status_srcqpn) |
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2472
static bool ocrdma_update_err_cqe(struct ib_wc *ibwc, struct ocrdma_cqe *cqe,
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2489
ocrdma_set_cqe_status_flushed(qp, cqe);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2494
static int ocrdma_update_err_rcqe(struct ib_wc *ibwc, struct ocrdma_cqe *cqe,
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2501
return ocrdma_update_err_cqe(ibwc, cqe, qp, status);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2504
static int ocrdma_update_err_scqe(struct ib_wc *ibwc, struct ocrdma_cqe *cqe,
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2510
return ocrdma_update_err_cqe(ibwc, cqe, qp, status);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2515
struct ocrdma_cqe *cqe, struct ib_wc *ibwc,
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2520
int status = (le32_to_cpu(cqe->flags_status_srcqpn) &
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2535
expand = ocrdma_update_err_rcqe(ibwc, cqe, qp, status);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2553
expand = ocrdma_update_err_scqe(ibwc, cqe, qp, status);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2559
struct ocrdma_cqe *cqe,
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2575
wqe_idx = (le32_to_cpu(cqe->wq.wqeidx) &
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2584
static bool ocrdma_poll_scqe(struct ocrdma_qp *qp, struct ocrdma_cqe *cqe,
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2590
status = (le32_to_cpu(cqe->flags_status_srcqpn) &
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2594
expand = ocrdma_poll_success_scqe(qp, cqe, ibwc, polled);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2596
expand = ocrdma_poll_err_scqe(qp, cqe, ibwc, polled, stop);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2601
struct ocrdma_cqe *cqe)
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2606
status = (le32_to_cpu(cqe->flags_status_srcqpn) &
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2608
ibwc->src_qp = le32_to_cpu(cqe->flags_status_srcqpn) &
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2612
ibwc->byte_len = (le32_to_cpu(cqe->ud.rxlen_pkey) >>
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2617
hdr_type = (le32_to_cpu(cqe->ud.rxlen_pkey) >>
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2628
struct ocrdma_cqe *cqe,
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2636
wqe_idx = (le32_to_cpu(cqe->rq.buftag_qpn) >>
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2647
static bool ocrdma_poll_err_rcqe(struct ocrdma_qp *qp, struct ocrdma_cqe *cqe,
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2664
expand = ocrdma_update_err_scqe(ibwc, cqe, qp, status);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2677
expand = ocrdma_update_err_rcqe(ibwc, cqe, qp, status);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2683
struct ocrdma_cqe *cqe, struct ib_wc *ibwc)
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2693
ocrdma_update_ud_rcqe(dev, ibwc, cqe);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2695
ibwc->byte_len = le32_to_cpu(cqe->rq.rxlen);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2697
if (is_cqe_imm(cqe)) {
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2698
ibwc->ex.imm_data = htonl(le32_to_cpu(cqe->rq.lkey_immdt));
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2700
} else if (is_cqe_wr_imm(cqe)) {
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2702
ibwc->ex.imm_data = htonl(le32_to_cpu(cqe->rq.lkey_immdt));
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2704
} else if (is_cqe_invalidated(cqe)) {
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2705
ibwc->ex.invalidate_rkey = le32_to_cpu(cqe->rq.lkey_immdt);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2709
ocrdma_update_free_srq_cqe(ibwc, cqe, qp);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2716
static bool ocrdma_poll_rcqe(struct ocrdma_qp *qp, struct ocrdma_cqe *cqe,
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2724
status = (le32_to_cpu(cqe->flags_status_srcqpn) &
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2728
status = (le32_to_cpu(cqe->flags_status_srcqpn) &
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2734
ocrdma_poll_success_rcqe(qp, cqe, ibwc);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2736
expand = ocrdma_poll_err_rcqe(qp, cqe, ibwc, polled, stop,
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2742
static void ocrdma_change_cq_phase(struct ocrdma_cq *cq, struct ocrdma_cqe *cqe,
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2750
cqe->flags_status_srcqpn = 0;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2763
struct ocrdma_cqe *cqe;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2768
cqe = cq->va + cur_getp;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2770
if (!is_cqe_valid(cq, cqe))
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2772
qpn = (le32_to_cpu(cqe->cmn.qpn) & OCRDMA_CQE_QPN_MASK);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2779
if (is_cqe_for_sq(cqe)) {
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2780
expand = ocrdma_poll_scqe(qp, cqe, ibwc, &polled,
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2783
expand = ocrdma_poll_rcqe(qp, cqe, ibwc, &polled,
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2791
cqe->cmn.qpn = 0;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2795
ocrdma_change_cq_phase(cq, cqe, cur_getp);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
973
int entries = attr->cqe;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.h
74
int ocrdma_resize_cq(struct ib_cq *, int cqe, struct ib_udata *);
drivers/infiniband/hw/qedr/verbs.c
4010
static int is_valid_cqe(struct qedr_cq *cq, union rdma_cqe *cqe)
drivers/infiniband/hw/qedr/verbs.c
4012
struct rdma_cqe_requester *resp_cqe = &cqe->req;
drivers/infiniband/hw/qedr/verbs.c
4018
static struct qedr_qp *cqe_get_qp(union rdma_cqe *cqe)
drivers/infiniband/hw/qedr/verbs.c
4020
struct rdma_cqe_requester *resp_cqe = &cqe->req;
drivers/infiniband/hw/qedr/verbs.c
4029
static enum rdma_cqe_type cqe_get_type(union rdma_cqe *cqe)
drivers/infiniband/hw/qedr/verbs.c
4031
struct rdma_cqe_requester *resp_cqe = &cqe->req;
drivers/infiniband/hw/qedr/verbs.c
4416
union rdma_cqe *cqe;
drivers/infiniband/hw/qedr/verbs.c
4433
cqe = cq->latest_cqe;
drivers/infiniband/hw/qedr/verbs.c
4435
while (num_entries && is_valid_cqe(cq, cqe)) {
drivers/infiniband/hw/qedr/verbs.c
4442
qp = cqe_get_qp(cqe);
drivers/infiniband/hw/qedr/verbs.c
4444
WARN(1, "Error: CQE QP pointer is NULL. CQE=%p\n", cqe);
drivers/infiniband/hw/qedr/verbs.c
4450
switch (cqe_get_type(cqe)) {
drivers/infiniband/hw/qedr/verbs.c
4453
&cqe->req);
drivers/infiniband/hw/qedr/verbs.c
4454
try_consume_req_cqe(cq, qp, &cqe->req, &update);
drivers/infiniband/hw/qedr/verbs.c
4458
&cqe->resp, &update);
drivers/infiniband/hw/qedr/verbs.c
4462
wc, &cqe->resp);
drivers/infiniband/hw/qedr/verbs.c
4468
cqe_get_type(cqe));
drivers/infiniband/hw/qedr/verbs.c
4474
cqe = get_cqe(cq);
drivers/infiniband/hw/qedr/verbs.c
921
int entries = attr->cqe;
drivers/infiniband/hw/qedr/verbs.c
976
cq->ibcq.cqe = chain_entries;
drivers/infiniband/hw/qedr/verbs.c
988
cq->ibcq.cqe = cq->pbl.capacity;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
106
int entries = attr->cqe;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
133
cq->ibcq.cqe = entries;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
187
cmd->cqe = entries;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
196
cq->ibcq.cqe = resp->cqe;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
289
cq->ibcq.cqe, &head);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
294
cq->ibcq.cqe);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
295
struct pvrdma_cqe *cqe;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
299
(cq->ibcq.cqe - head + tail);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
303
curr = cq->ibcq.cqe - 1;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
305
tail = cq->ibcq.cqe - 1;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
309
cqe = get_cqe(cq, tail);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
310
*cqe = *curr_cqe;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
316
cq->ibcq.cqe);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
330
struct pvrdma_cqe *cqe;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
334
cq->ibcq.cqe, &head);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
348
cqe = get_cqe(cq, head);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
352
if (dev->qp_tbl[cqe->qp & 0xffff])
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
353
*cur_qp = (struct pvrdma_qp *)dev->qp_tbl[cqe->qp & 0xffff];
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
357
wc->opcode = pvrdma_wc_opcode_to_ib(cqe->opcode);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
358
wc->status = pvrdma_wc_status_to_ib(cqe->status);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
359
wc->wr_id = cqe->wr_id;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
361
wc->byte_len = cqe->byte_len;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
362
wc->ex.imm_data = cqe->imm_data;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
363
wc->src_qp = cqe->src_qp;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
364
wc->wc_flags = pvrdma_wc_flags_to_ib(cqe->wc_flags);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
365
wc->pkey_index = cqe->pkey_index;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
366
wc->slid = cqe->slid;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
367
wc->sl = cqe->sl;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
368
wc->dlid_path_bits = cqe->dlid_path_bits;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
369
wc->port_num = cqe->port_num;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
370
wc->vendor_err = cqe->vendor_err;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
371
wc->network_hdr_type = pvrdma_network_type_to_ib(cqe->network_hdr_type);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
374
pvrdma_idx_ring_inc(&cq->ring_state->rx.cons_head, cq->ibcq.cqe);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
83
cq->ibcq.cqe, &head);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_dev_api.h
483
u32 cqe;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_dev_api.h
491
u32 cqe;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_dev_api.h
497
u32 cqe;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_dev_api.h
502
u32 cqe;
drivers/infiniband/sw/rdmavt/cq.c
169
unsigned int entries = attr->cqe;
drivers/infiniband/sw/rdmavt/cq.c
251
cq->ibcq.cqe = entries;
drivers/infiniband/sw/rdmavt/cq.c
340
int rvt_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata)
drivers/infiniband/sw/rdmavt/cq.c
352
if (cqe < 1 || cqe > rdi->dparms.props.max_cqe)
drivers/infiniband/sw/rdmavt/cq.c
359
sz = sizeof(struct ib_uverbs_wc) * (cqe + 1);
drivers/infiniband/sw/rdmavt/cq.c
365
sz = sizeof(struct ib_wc) * (cqe + 1);
drivers/infiniband/sw/rdmavt/cq.c
395
if (head > (u32)cq->ibcq.cqe)
drivers/infiniband/sw/rdmavt/cq.c
396
head = (u32)cq->ibcq.cqe;
drivers/infiniband/sw/rdmavt/cq.c
397
if (tail > (u32)cq->ibcq.cqe)
drivers/infiniband/sw/rdmavt/cq.c
398
tail = (u32)cq->ibcq.cqe;
drivers/infiniband/sw/rdmavt/cq.c
400
n = cq->ibcq.cqe + 1 + head - tail;
drivers/infiniband/sw/rdmavt/cq.c
403
if (unlikely((u32)cqe < n)) {
drivers/infiniband/sw/rdmavt/cq.c
412
if (tail == (u32)cq->ibcq.cqe)
drivers/infiniband/sw/rdmavt/cq.c
417
cq->ibcq.cqe = cqe;
drivers/infiniband/sw/rdmavt/cq.c
494
if (tail > (u32)cq->ibcq.cqe)
drivers/infiniband/sw/rdmavt/cq.c
495
tail = (u32)cq->ibcq.cqe;
drivers/infiniband/sw/rdmavt/cq.c
502
if (tail >= cq->ibcq.cqe)
drivers/infiniband/sw/rdmavt/cq.c
55
if (head >= (unsigned)cq->ibcq.cqe) {
drivers/infiniband/sw/rdmavt/cq.c
56
head = cq->ibcq.cqe;
drivers/infiniband/sw/rdmavt/cq.h
16
int rvt_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata);
drivers/infiniband/sw/rdmavt/trace_cq.h
41
__field(unsigned int, cqe)
drivers/infiniband/sw/rdmavt/trace_cq.h
48
__entry->cqe = attr->cqe;
drivers/infiniband/sw/rdmavt/trace_cq.h
55
__entry->ip ? "true" : "false", __entry->cqe,
drivers/infiniband/sw/rxe/rxe_comp.c
397
struct rxe_cqe *cqe)
drivers/infiniband/sw/rxe/rxe_comp.c
399
struct ib_wc *wc = &cqe->ibwc;
drivers/infiniband/sw/rxe/rxe_comp.c
400
struct ib_uverbs_wc *uwc = &cqe->uibwc;
drivers/infiniband/sw/rxe/rxe_comp.c
402
memset(cqe, 0, sizeof(*cqe));
drivers/infiniband/sw/rxe/rxe_comp.c
446
struct rxe_cqe cqe;
drivers/infiniband/sw/rxe/rxe_comp.c
455
make_send_cqe(qp, wqe, &cqe);
drivers/infiniband/sw/rxe/rxe_comp.c
460
rxe_cq_post(qp->scq, &cqe, 0);
drivers/infiniband/sw/rxe/rxe_comp.c
562
struct rxe_cqe cqe = {};
drivers/infiniband/sw/rxe/rxe_comp.c
563
struct ib_wc *wc = &cqe.ibwc;
drivers/infiniband/sw/rxe/rxe_comp.c
564
struct ib_uverbs_wc *uwc = &cqe.uibwc;
drivers/infiniband/sw/rxe/rxe_comp.c
577
err = rxe_cq_post(qp->scq, &cqe, 0);
drivers/infiniband/sw/rxe/rxe_cq.c
109
memcpy(addr, cqe, sizeof(*cqe));
drivers/infiniband/sw/rxe/rxe_cq.c
12
int cqe, int comp_vector)
drivers/infiniband/sw/rxe/rxe_cq.c
16
if (cqe <= 0) {
drivers/infiniband/sw/rxe/rxe_cq.c
17
rxe_dbg_dev(rxe, "cqe(%d) <= 0\n", cqe);
drivers/infiniband/sw/rxe/rxe_cq.c
21
if (cqe > rxe->attr.max_cqe) {
drivers/infiniband/sw/rxe/rxe_cq.c
23
cqe, rxe->attr.max_cqe);
drivers/infiniband/sw/rxe/rxe_cq.c
29
if (cqe < count) {
drivers/infiniband/sw/rxe/rxe_cq.c
31
cqe, count);
drivers/infiniband/sw/rxe/rxe_cq.c
42
int rxe_cq_from_init(struct rxe_dev *rxe, struct rxe_cq *cq, int cqe,
drivers/infiniband/sw/rxe/rxe_cq.c
50
cq->queue = rxe_queue_init(rxe, &cqe,
drivers/infiniband/sw/rxe/rxe_cq.c
65
cq->ibcq.cqe = cqe;
drivers/infiniband/sw/rxe/rxe_cq.c
69
int rxe_cq_resize_queue(struct rxe_cq *cq, int cqe,
drivers/infiniband/sw/rxe/rxe_cq.c
75
err = rxe_queue_resize(cq->queue, (unsigned int *)&cqe,
drivers/infiniband/sw/rxe/rxe_cq.c
79
cq->ibcq.cqe = cqe;
drivers/infiniband/sw/rxe/rxe_cq.c
85
int rxe_cq_post(struct rxe_cq *cq, struct rxe_cqe *cqe, int solicited)
drivers/infiniband/sw/rxe/rxe_loc.h
22
int cqe, int comp_vector);
drivers/infiniband/sw/rxe/rxe_loc.h
24
int rxe_cq_from_init(struct rxe_dev *rxe, struct rxe_cq *cq, int cqe,
drivers/infiniband/sw/rxe/rxe_loc.h
32
int rxe_cq_post(struct rxe_cq *cq, struct rxe_cqe *cqe, int solicited);
drivers/infiniband/sw/rxe/rxe_resp.c
1072
struct rxe_cqe cqe;
drivers/infiniband/sw/rxe/rxe_resp.c
1073
struct ib_wc *wc = &cqe.ibwc;
drivers/infiniband/sw/rxe/rxe_resp.c
1074
struct ib_uverbs_wc *uwc = &cqe.uibwc;
drivers/infiniband/sw/rxe/rxe_resp.c
1082
memset(&cqe, 0, sizeof(cqe));
drivers/infiniband/sw/rxe/rxe_resp.c
1164
if (rxe_cq_post(qp->rcq, &cqe, pkt ? bth_se(pkt) : 1))
drivers/infiniband/sw/rxe/rxe_resp.c
1446
struct rxe_cqe cqe = {};
drivers/infiniband/sw/rxe/rxe_resp.c
1447
struct ib_wc *wc = &cqe.ibwc;
drivers/infiniband/sw/rxe/rxe_resp.c
1448
struct ib_uverbs_wc *uwc = &cqe.uibwc;
drivers/infiniband/sw/rxe/rxe_resp.c
1461
err = rxe_cq_post(qp->rcq, &cqe, 0);
drivers/infiniband/sw/rxe/rxe_verbs.c
1100
err = rxe_cq_chk_attr(rxe, NULL, attr->cqe, attr->comp_vector);
drivers/infiniband/sw/rxe/rxe_verbs.c
1112
err = rxe_cq_from_init(rxe, cq, attr->cqe, attr->comp_vector, udata,
drivers/infiniband/sw/rxe/rxe_verbs.c
1130
static int rxe_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata)
drivers/infiniband/sw/rxe/rxe_verbs.c
1146
err = rxe_cq_chk_attr(rxe, cq, cqe, 0);
drivers/infiniband/sw/rxe/rxe_verbs.c
1152
err = rxe_cq_resize_queue(cq, cqe, uresp, udata);
drivers/infiniband/sw/rxe/rxe_verbs.c
1169
struct rxe_cqe *cqe;
drivers/infiniband/sw/rxe/rxe_verbs.c
1174
cqe = queue_head(cq->queue, QUEUE_TYPE_TO_ULP);
drivers/infiniband/sw/rxe/rxe_verbs.c
1175
if (!cqe)
drivers/infiniband/sw/rxe/rxe_verbs.c
1178
memcpy(wc++, &cqe->ibwc, sizeof(*wc));
drivers/infiniband/sw/siw/siw_cq.c
50
struct siw_cqe *cqe;
drivers/infiniband/sw/siw/siw_cq.c
55
cqe = &cq->queue[cq->cq_get % cq->num_cqe];
drivers/infiniband/sw/siw/siw_cq.c
56
if (READ_ONCE(cqe->flags) & SIW_WQE_VALID) {
drivers/infiniband/sw/siw/siw_cq.c
58
wc->wr_id = cqe->id;
drivers/infiniband/sw/siw/siw_cq.c
59
wc->byte_len = cqe->bytes;
drivers/infiniband/sw/siw/siw_cq.c
67
if (cqe->flags & SIW_WQE_REM_INVAL) {
drivers/infiniband/sw/siw/siw_cq.c
68
wc->ex.invalidate_rkey = cqe->inval_stag;
drivers/infiniband/sw/siw/siw_cq.c
71
wc->qp = cqe->base_qp;
drivers/infiniband/sw/siw/siw_cq.c
72
wc->opcode = map_wc_opcode[cqe->opcode];
drivers/infiniband/sw/siw/siw_cq.c
73
wc->status = map_cqe_status[cqe->status].ib;
drivers/infiniband/sw/siw/siw_cq.c
76
cq->cq_get % cq->num_cqe, cqe->opcode,
drivers/infiniband/sw/siw/siw_cq.c
77
cqe->flags, (void *)(uintptr_t)cqe->id);
drivers/infiniband/sw/siw/siw_cq.c
86
u8 opcode = cqe->opcode;
drivers/infiniband/sw/siw/siw_cq.c
87
u16 status = cqe->status;
drivers/infiniband/sw/siw/siw_cq.c
99
WRITE_ONCE(cqe->flags, 0);
drivers/infiniband/sw/siw/siw_qp.c
1039
struct siw_cqe *cqe;
drivers/infiniband/sw/siw/siw_qp.c
1046
cqe = &cq->queue[idx];
drivers/infiniband/sw/siw/siw_qp.c
1048
if (!READ_ONCE(cqe->flags)) {
drivers/infiniband/sw/siw/siw_qp.c
1051
cqe->id = sqe->id;
drivers/infiniband/sw/siw/siw_qp.c
1052
cqe->opcode = sqe->opcode;
drivers/infiniband/sw/siw/siw_qp.c
1053
cqe->status = status;
drivers/infiniband/sw/siw/siw_qp.c
1054
cqe->imm_data = 0;
drivers/infiniband/sw/siw/siw_qp.c
1055
cqe->bytes = bytes;
drivers/infiniband/sw/siw/siw_qp.c
1058
cqe->base_qp = &qp->base_qp;
drivers/infiniband/sw/siw/siw_qp.c
1060
cqe->qp_id = qp_id(qp);
drivers/infiniband/sw/siw/siw_qp.c
1063
WRITE_ONCE(cqe->flags, SIW_WQE_VALID);
drivers/infiniband/sw/siw/siw_qp.c
1096
struct siw_cqe *cqe;
drivers/infiniband/sw/siw/siw_qp.c
1103
cqe = &cq->queue[idx];
drivers/infiniband/sw/siw/siw_qp.c
1105
if (!READ_ONCE(cqe->flags)) {
drivers/infiniband/sw/siw/siw_qp.c
1109
cqe->id = rqe->id;
drivers/infiniband/sw/siw/siw_qp.c
1110
cqe->opcode = SIW_OP_RECEIVE;
drivers/infiniband/sw/siw/siw_qp.c
1111
cqe->status = status;
drivers/infiniband/sw/siw/siw_qp.c
1112
cqe->imm_data = 0;
drivers/infiniband/sw/siw/siw_qp.c
1113
cqe->bytes = bytes;
drivers/infiniband/sw/siw/siw_qp.c
1116
cqe->base_qp = &qp->base_qp;
drivers/infiniband/sw/siw/siw_qp.c
1119
cqe->inval_stag = inval_stag;
drivers/infiniband/sw/siw/siw_qp.c
1122
cqe->qp_id = qp_id(qp);
drivers/infiniband/sw/siw/siw_qp.c
1125
WRITE_ONCE(cqe->flags, cqe_flags);
drivers/infiniband/sw/siw/siw_verbs.c
1148
int rv, size = attr->cqe;
drivers/infiniband/sw/siw/siw_verbs.c
1164
cq->base_cq.cqe = size;
drivers/infiniband/ulp/ipoib/ipoib_verbs.c
176
cq_attr.cqe = size;
drivers/infiniband/ulp/ipoib/ipoib_verbs.c
185
cq_attr.cqe = ipoib_sendq_size;
drivers/infiniband/ulp/iser/iscsi_iser.h
244
struct ib_cqe cqe;
drivers/infiniband/ulp/iser/iscsi_iser.h
271
struct ib_cqe cqe;
drivers/infiniband/ulp/iser/iscsi_iser.h
291
struct ib_cqe cqe;
drivers/infiniband/ulp/iser/iscsi_iser.h
552
iser_rx(struct ib_cqe *cqe)
drivers/infiniband/ulp/iser/iscsi_iser.h
554
return container_of(cqe, struct iser_rx_desc, cqe);
drivers/infiniband/ulp/iser/iscsi_iser.h
558
iser_tx(struct ib_cqe *cqe)
drivers/infiniband/ulp/iser/iscsi_iser.h
560
return container_of(cqe, struct iser_tx_desc, cqe);
drivers/infiniband/ulp/iser/iscsi_iser.h
564
iser_login(struct ib_cqe *cqe)
drivers/infiniband/ulp/iser/iscsi_iser.h
566
return container_of(cqe, struct iser_login_desc, cqe);
drivers/infiniband/ulp/iser/iser_initiator.c
150
tx_desc->cqe.done = done;
drivers/infiniband/ulp/iser/iser_initiator.c
257
rx_desc->cqe.done = iser_task_rsp;
drivers/infiniband/ulp/iser/iser_initiator.c
436
tx_desc->cqe.done = iser_dataout_comp;
drivers/infiniband/ulp/iser/iser_memory.c
237
struct ib_cqe *cqe, struct ib_send_wr *next_wr)
drivers/infiniband/ulp/iser/iser_memory.c
240
inv_wr->wr_cqe = cqe;
drivers/infiniband/ulp/iser/iser_memory.c
254
struct ib_cqe *cqe = &iser_task->iser_conn->ib_conn.reg_cqe;
drivers/infiniband/ulp/iser/iser_memory.c
268
iser_inv_rkey(&tx_desc->inv_wr, mr, cqe, &wr->wr);
drivers/infiniband/ulp/iser/iser_memory.c
283
wr->wr.wr_cqe = cqe;
drivers/infiniband/ulp/iser/iser_memory.c
311
struct ib_cqe *cqe = &iser_task->iser_conn->ib_conn.reg_cqe;
drivers/infiniband/ulp/iser/iser_memory.c
317
iser_inv_rkey(&tx_desc->inv_wr, mr, cqe, &wr->wr);
drivers/infiniband/ulp/iser/iser_memory.c
330
wr->wr.wr_cqe = cqe;
drivers/infiniband/ulp/iser/iser_verbs.c
817
desc->cqe.done = iser_login_rsp;
drivers/infiniband/ulp/iser/iser_verbs.c
818
wr.wr_cqe = &desc->cqe;
drivers/infiniband/ulp/iser/iser_verbs.c
836
rx_desc->cqe.done = iser_task_rsp;
drivers/infiniband/ulp/iser/iser_verbs.c
837
wr.wr_cqe = &rx_desc->cqe;
drivers/infiniband/ulp/iser/iser_verbs.c
868
wr->wr_cqe = &tx_desc->cqe;
drivers/infiniband/ulp/isert/ib_isert.c
2035
struct ib_cqe *cqe, struct ib_send_wr *chain_wr)
drivers/infiniband/ulp/isert/ib_isert.c
2083
ret = rdma_rw_ctx_post(&cmd->rw, conn->qp, port_num, cqe, chain_wr);
drivers/infiniband/ulp/isert/ib_isert.c
2095
struct ib_cqe *cqe = NULL;
drivers/infiniband/ulp/isert/ib_isert.c
2104
cqe = &isert_cmd->tx_desc.tx_cqe;
drivers/infiniband/ulp/isert/ib_isert.c
2124
rc = isert_rdma_rw_ctx_post(isert_cmd, isert_conn, cqe, chain_wr);
drivers/infiniband/ulp/isert/ib_isert.h
135
static inline struct iser_tx_desc *cqe_to_tx_desc(struct ib_cqe *cqe)
drivers/infiniband/ulp/isert/ib_isert.h
137
return container_of(cqe, struct iser_tx_desc, tx_cqe);
drivers/infiniband/ulp/isert/ib_isert.h
96
static inline struct iser_rx_desc *cqe_to_rx_desc(struct ib_cqe *cqe)
drivers/infiniband/ulp/isert/ib_isert.h
98
return container_of(cqe, struct iser_rx_desc, rx_cqe);
drivers/infiniband/ulp/rtrs/rtrs-clt.c
2421
iu = container_of(wc->wr_cqe, struct rtrs_iu, cqe);
drivers/infiniband/ulp/rtrs/rtrs-clt.c
2512
iu = container_of(wc->wr_cqe, struct rtrs_iu, cqe);
drivers/infiniband/ulp/rtrs/rtrs-clt.c
514
cqe);
drivers/infiniband/ulp/rtrs/rtrs-clt.c
534
iu = container_of(wc->wr_cqe, struct rtrs_iu, cqe);
drivers/infiniband/ulp/rtrs/rtrs-clt.c
584
static int rtrs_post_recv_empty_x2(struct rtrs_con *con, struct ib_cqe *cqe)
drivers/infiniband/ulp/rtrs/rtrs-clt.c
592
wr->wr_cqe = cqe;
drivers/infiniband/ulp/rtrs/rtrs-pri.h
128
struct ib_cqe cqe;
drivers/infiniband/ulp/rtrs/rtrs-pri.h
323
int rtrs_post_recv_empty(struct rtrs_con *con, struct ib_cqe *cqe);
drivers/infiniband/ulp/rtrs/rtrs-pri.h
331
void rtrs_init_hb(struct rtrs_path *path, struct ib_cqe *cqe,
drivers/infiniband/ulp/rtrs/rtrs-srv.c
725
iu = container_of(wc->wr_cqe, struct rtrs_iu, cqe);
drivers/infiniband/ulp/rtrs/rtrs-srv.c
936
iu = container_of(wc->wr_cqe, struct rtrs_iu, cqe);
drivers/infiniband/ulp/rtrs/rtrs.c
102
.wr_cqe = cqe,
drivers/infiniband/ulp/rtrs/rtrs.c
143
.wr_cqe = &iu->cqe,
drivers/infiniband/ulp/rtrs/rtrs.c
165
.wr.wr_cqe = &iu->cqe,
drivers/infiniband/ulp/rtrs/rtrs.c
188
struct ib_cqe *cqe,
drivers/infiniband/ulp/rtrs/rtrs.c
201
.wr.wr_cqe = cqe,
drivers/infiniband/ulp/rtrs/rtrs.c
389
void rtrs_init_hb(struct rtrs_path *path, struct ib_cqe *cqe,
drivers/infiniband/ulp/rtrs/rtrs.c
394
path->hb_cqe = cqe;
drivers/infiniband/ulp/rtrs/rtrs.c
45
iu->cqe.done = done;
drivers/infiniband/ulp/rtrs/rtrs.c
88
.wr_cqe = &iu->cqe,
drivers/infiniband/ulp/rtrs/rtrs.c
97
int rtrs_post_recv_empty(struct rtrs_con *con, struct ib_cqe *cqe)
drivers/infiniband/ulp/srp/ib_srp.c
1871
struct srp_iu *iu = container_of(wc->wr_cqe, struct srp_iu, cqe);
drivers/infiniband/ulp/srp/ib_srp.c
1902
iu->cqe.done = srp_send_done;
drivers/infiniband/ulp/srp/ib_srp.c
1905
wr.wr_cqe = &iu->cqe;
drivers/infiniband/ulp/srp/ib_srp.c
1924
iu->cqe.done = srp_recv_done;
drivers/infiniband/ulp/srp/ib_srp.c
1927
wr.wr_cqe = &iu->cqe;
drivers/infiniband/ulp/srp/ib_srp.c
2062
struct srp_iu *iu = container_of(wc->wr_cqe, struct srp_iu, cqe);
drivers/infiniband/ulp/srp/ib_srp.h
279
struct ib_cqe cqe;
drivers/infiniband/ulp/srpt/ib_srpt.c
1758
container_of(wc->wr_cqe, struct srpt_recv_ioctx, ioctx.cqe);
drivers/infiniband/ulp/srpt/ib_srpt.c
1819
container_of(wc->wr_cqe, struct srpt_send_ioctx, ioctx.cqe);
drivers/infiniband/ulp/srpt/ib_srpt.c
1929
__func__, ch->cq->cqe, qp_init->cap.max_send_sge,
drivers/infiniband/ulp/srpt/ib_srpt.c
2808
struct ib_cqe *cqe = &ioctx->rdma_cqe;
drivers/infiniband/ulp/srpt/ib_srpt.c
2828
cqe->done = srpt_rdma_read_done;
drivers/infiniband/ulp/srpt/ib_srpt.c
2833
cqe, first_wr);
drivers/infiniband/ulp/srpt/ib_srpt.c
2834
cqe = NULL;
drivers/infiniband/ulp/srpt/ib_srpt.c
2937
ioctx->ioctx.cqe.done = srpt_send_done;
drivers/infiniband/ulp/srpt/ib_srpt.c
2939
send_wr.wr_cqe = &ioctx->ioctx.cqe;
drivers/infiniband/ulp/srpt/ib_srpt.c
899
ioctx->ioctx.cqe.done = srpt_recv_done;
drivers/infiniband/ulp/srpt/ib_srpt.c
900
wr.wr_cqe = &ioctx->ioctx.cqe;
drivers/infiniband/ulp/srpt/ib_srpt.h
170
struct ib_cqe cqe;
drivers/mmc/core/core.c
264
bool cqe)
drivers/mmc/core/core.c
274
mmc_hostname(host), cqe ? "CQE direct " : "",
drivers/mmc/core/core.c
276
} else if (cqe) {
drivers/net/ethernet/amazon/ena/ena_com.c
430
struct ena_admin_acq_entry *cqe)
drivers/net/ethernet/amazon/ena/ena_com.c
435
cmd_id = cqe->acq_common_descriptor.command &
drivers/net/ethernet/amazon/ena/ena_com.c
447
comp_ctx->comp_status = cqe->acq_common_descriptor.status;
drivers/net/ethernet/amazon/ena/ena_com.c
450
memcpy(comp_ctx->user_cqe, (void *)cqe, comp_ctx->comp_size);
drivers/net/ethernet/amazon/ena/ena_com.c
458
struct ena_admin_acq_entry *cqe = NULL;
drivers/net/ethernet/amazon/ena/ena_com.c
466
cqe = &admin_queue->cq.entries[head_masked];
drivers/net/ethernet/amazon/ena/ena_com.c
469
while ((READ_ONCE(cqe->acq_common_descriptor.flags) &
drivers/net/ethernet/amazon/ena/ena_com.c
475
ena_com_handle_single_admin_completion(admin_queue, cqe);
drivers/net/ethernet/amazon/ena/ena_com.c
484
cqe = &admin_queue->cq.entries[head_masked];
drivers/net/ethernet/broadcom/bnx2x/bnx2x.h
815
#define BNX2X_RX_SUM_FIX(cqe) \
drivers/net/ethernet/broadcom/bnx2x/bnx2x.h
816
BNX2X_PRS_FLAG_OVERETH_IPV4(cqe->fast_path_cqe.pars_flags.flags)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
1003
frag_size = le16_to_cpu(cqe->end_agg_cqe.pkt_len) -
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
1014
&cqe->end_agg_cqe, comp_ring_cons);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
1020
bnx2x_update_sge_prod(fp, pages, &cqe->end_agg_cqe);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
1090
bnx2x_csum_validate(skb, cqe, fp,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
1096
if (unlikely(cqe->fast_path_cqe.type_error_flags &
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
1124
cqe = &fp->rx_comp_ring[comp_ring_cons];
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
1125
cqe_fp = &cqe->fast_path_cqe;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
356
struct eth_end_agg_rx_cqe *cqe)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
369
RX_SGE(le16_to_cpu(cqe->sgl_or_raw_data.sgl[i])));
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
372
sge_len - 1, le16_to_cpu(cqe->sgl_or_raw_data.sgl[sge_len - 1]));
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
377
le16_to_cpu(cqe->sgl_or_raw_data.sgl[sge_len - 1]));
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
411
const struct eth_fast_path_rx_cqe *cqe,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
416
(cqe->status_flags & ETH_FAST_PATH_RX_CQE_RSS_HASH_FLG)) {
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
419
htype = cqe->status_flags & ETH_FAST_PATH_RX_CQE_RSS_HASH_TYPE;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
424
return le32_to_cpu(cqe->rss_hash_result);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
432
struct eth_fast_path_rx_cqe *cqe)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
475
le16_to_cpu(cqe->pars_flags.flags);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
476
tpa_info->vlan_tag = le16_to_cpu(cqe->vlan_tag);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
478
tpa_info->len_on_bd = le16_to_cpu(cqe->len_on_bd);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
479
tpa_info->placement_offset = cqe->placement_offset;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
480
tpa_info->rxhash = bnx2x_get_rxhash(bp, cqe, &tpa_info->rxhash_type);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
482
u16 gro_size = le16_to_cpu(cqe->pkt_len_or_gro_seg_len);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
590
struct eth_end_agg_rx_cqe *cqe,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
599
frag_size = le16_to_cpu(cqe->pkt_len) - len_on_bd;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
609
le16_to_cpu(cqe->pkt_len),
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
610
le16_to_cpu(cqe->num_of_coalesced_segs));
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
616
BNX2X_ERR("cqe->pkt_len = %d\n", cqe->pkt_len);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
624
u16 sge_idx = RX_SGE(le16_to_cpu(cqe->sgl_or_raw_data.sgl[j]));
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
769
struct eth_end_agg_rx_cqe *cqe,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
816
skb, cqe, cqe_idx)) {
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
871
void bnx2x_csum_validate(struct sk_buff *skb, union eth_rx_cqe *cqe,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
880
if (cqe->fast_path_cqe.status_flags &
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
886
if (cqe->fast_path_cqe.type_error_flags &
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
900
union eth_rx_cqe *cqe;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
917
cqe = &fp->rx_comp_ring[comp_ring_cons];
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
918
cqe_fp = &cqe->fast_path_cqe;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
966
bnx2x_sp_event(fp, cqe);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
997
queue = cqe->end_agg_cqe.queue_index;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h
779
union eth_rx_cqe *cqe;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h
783
cqe = &fp->rx_comp_ring[cons];
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h
784
cqe_fp = &cqe->fast_path_cqe;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c
2506
union eth_rx_cqe *cqe;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c
2658
cqe = &fp_rx->rx_comp_ring[RCQ_BD(fp_rx->rx_comp_cons)];
drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c
2659
cqe_fp_flags = cqe->fast_path_cqe.type_error_flags;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c
2664
len = le16_to_cpu(cqe->fast_path_cqe.pkt_len_or_gro_seg_len);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c
2672
data = rx_buf->data + NET_SKB_PAD + cqe->fast_path_cqe.placement_offset;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
1145
u32 *cqe = (u32 *)&fp->rx_comp_ring[j];
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
1148
i, j, cqe[0], cqe[1], cqe[2], cqe[3]);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
1619
union event_ring_elem *cqe,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
1639
if (cqe->message.error)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.h
438
union event_ring_elem *cqe,
drivers/net/ethernet/broadcom/cnic.c
2921
union eth_rx_cqe *cqe, *cqe_ring = (union eth_rx_cqe *)
drivers/net/ethernet/broadcom/cnic.c
2937
cqe = &cqe_ring[sw_cons & BNX2X_MAX_RCQ_DESC_CNT];
drivers/net/ethernet/broadcom/cnic.c
2938
cqe_fp_flags = cqe->fast_path_cqe.type_error_flags;
drivers/net/ethernet/broadcom/cnic.c
2940
cmd = le32_to_cpu(cqe->ramrod_cqe.conn_and_cmd_data);
drivers/net/ethernet/fungible/funcore/fun_dev.h
37
typedef void (*fun_admin_event_cb)(struct fun_dev *fdev, void *cqe);
drivers/net/ethernet/fungible/funcore/fun_queue.c
289
void *cqe;
drivers/net/ethernet/fungible/funcore/fun_queue.c
295
cqe = funq->cqes + (funq->cq_head << funq->cqe_size_log2);
drivers/net/ethernet/fungible/funcore/fun_queue.c
296
info = funq_cqe_info(funq, cqe);
drivers/net/ethernet/fungible/funcore/fun_queue.c
310
rsp = cqe;
drivers/net/ethernet/fungible/funcore/fun_queue.c
317
rsp = cqe;
drivers/net/ethernet/fungible/funcore/fun_queue.h
104
void *cqe)
drivers/net/ethernet/fungible/funcore/fun_queue.h
106
return cqe + funq->cqe_info_offset;
drivers/net/ethernet/fungible/funeth/funeth_rx.c
206
static const void *cqe_to_info(const void *cqe)
drivers/net/ethernet/fungible/funeth/funeth_rx.c
208
return cqe + FUNETH_CQE_INFO_OFFSET;
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
324
cqe_size = wq->q_depth * sizeof(*rq->cqe);
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
325
rq->cqe = vzalloc(cqe_size);
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
326
if (!rq->cqe)
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
335
rq->cqe[i] = dma_alloc_coherent(&pdev->dev,
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
336
sizeof(*rq->cqe[i]),
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
338
if (!rq->cqe[i])
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
346
dma_free_coherent(&pdev->dev, sizeof(*rq->cqe[j]), rq->cqe[j],
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
352
vfree(rq->cqe);
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
368
dma_free_coherent(&pdev->dev, sizeof(*rq->cqe[i]), rq->cqe[i],
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
372
vfree(rq->cqe);
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
842
struct hinic_rq_cqe *cqe;
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
850
cqe = rq->cqe[*cons_idx];
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
852
status = be32_to_cpu(cqe->status);
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
902
struct hinic_rq_cqe *cqe = rq->cqe[cons_idx];
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
903
u32 status = be32_to_cpu(cqe->status);
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
908
cqe->status = cpu_to_be32(status);
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
925
struct hinic_rq_cqe *cqe = rq->cqe[cons_idx];
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
926
u32 len = be32_to_cpu(cqe->len);
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
945
struct hinic_rq_cqe *cqe = rq->cqe[prod_idx];
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
957
hinic_set_sge(&cqe_sect->sge, cqe_dma, sizeof(*cqe));
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.h
113
struct hinic_rq_cqe **cqe;
drivers/net/ethernet/huawei/hinic/hinic_rx.c
363
struct hinic_rq_cqe *cqe;
drivers/net/ethernet/huawei/hinic/hinic_rx.c
387
cqe = rq->cqe[ci];
drivers/net/ethernet/huawei/hinic/hinic_rx.c
388
status = be32_to_cpu(cqe->status);
drivers/net/ethernet/huawei/hinic/hinic_rx.c
410
offload_type = be32_to_cpu(cqe->offload_type);
drivers/net/ethernet/huawei/hinic/hinic_rx.c
411
vlan_len = be32_to_cpu(cqe->len);
drivers/net/ethernet/huawei/hinic/hinic_rx.c
432
LRO_PKT_HDR_LEN(cqe));
drivers/net/ethernet/huawei/hinic/hinic_rx.c
439
cqe->status = 0;
drivers/net/ethernet/huawei/hinic/hinic_rx.c
45
#define LRO_PKT_HDR_LEN(cqe) \
drivers/net/ethernet/huawei/hinic/hinic_rx.c
46
(HINIC_GET_RX_PKT_TYPE(be32_to_cpu((cqe)->offload_type)) == \
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
31
#define HINIC3_LRO_PKT_HDR_LEN(cqe) \
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
32
(RQ_CQE_OFFOLAD_TYPE_GET((cqe)->offload_type, IP_TYPE) == \
drivers/net/ethernet/ibm/ehea/ehea_main.c
532
static inline int ehea_check_cqe(struct ehea_cqe *cqe, int *rq_num)
drivers/net/ethernet/ibm/ehea/ehea_main.c
534
*rq_num = (cqe->type & EHEA_CQE_TYPE_RQ) >> 5;
drivers/net/ethernet/ibm/ehea/ehea_main.c
535
if ((cqe->status & EHEA_CQE_STAT_ERR_MASK) == 0)
drivers/net/ethernet/ibm/ehea/ehea_main.c
537
if (((cqe->status & EHEA_CQE_STAT_ERR_TCP) != 0) &&
drivers/net/ethernet/ibm/ehea/ehea_main.c
538
(cqe->header_length == 0))
drivers/net/ethernet/ibm/ehea/ehea_main.c
544
struct sk_buff *skb, struct ehea_cqe *cqe,
drivers/net/ethernet/ibm/ehea/ehea_main.c
547
int length = cqe->num_bytes_transfered - 4; /*remove CRC */
drivers/net/ethernet/ibm/ehea/ehea_main.c
554
if (cqe->status & EHEA_CQE_BLIND_CKSUM) {
drivers/net/ethernet/ibm/ehea/ehea_main.c
556
skb->csum = csum_unfold(~cqe->inet_checksum_value);
drivers/net/ethernet/ibm/ehea/ehea_main.c
565
struct ehea_cqe *cqe)
drivers/net/ethernet/ibm/ehea/ehea_main.c
567
int skb_index = EHEA_BMASK_GET(EHEA_WR_ID_INDEX, cqe->wr_id);
drivers/net/ethernet/ibm/ehea/ehea_main.c
618
struct ehea_cqe *cqe, int *processed_rq2,
drivers/net/ethernet/ibm/ehea/ehea_main.c
623
if (cqe->status & EHEA_CQE_STAT_ERR_TCP)
drivers/net/ethernet/ibm/ehea/ehea_main.c
625
if (cqe->status & EHEA_CQE_STAT_ERR_IP)
drivers/net/ethernet/ibm/ehea/ehea_main.c
627
if (cqe->status & EHEA_CQE_STAT_ERR_CRC)
drivers/net/ethernet/ibm/ehea/ehea_main.c
632
skb = get_skb_by_index(pr->rq2_skba.arr, pr->rq2_skba.len, cqe);
drivers/net/ethernet/ibm/ehea/ehea_main.c
636
skb = get_skb_by_index(pr->rq3_skba.arr, pr->rq3_skba.len, cqe);
drivers/net/ethernet/ibm/ehea/ehea_main.c
640
if (cqe->status & EHEA_CQE_STAT_FAT_ERR_MASK) {
drivers/net/ethernet/ibm/ehea/ehea_main.c
644
ehea_dump(cqe, sizeof(*cqe), "CQE");
drivers/net/ethernet/ibm/ehea/ehea_main.c
659
struct ehea_cqe *cqe;
drivers/net/ethernet/ibm/ehea/ehea_main.c
674
cqe = ehea_poll_rq1(qp, &wqe_index);
drivers/net/ethernet/ibm/ehea/ehea_main.c
675
while ((processed < budget) && cqe) {
drivers/net/ethernet/ibm/ehea/ehea_main.c
680
ehea_dump(cqe, sizeof(*cqe), "CQE");
drivers/net/ethernet/ibm/ehea/ehea_main.c
684
if (!ehea_check_cqe(cqe, &rq)) {
drivers/net/ethernet/ibm/ehea/ehea_main.c
699
skb_copy_to_linear_data(skb, ((char *)cqe) + 64,
drivers/net/ethernet/ibm/ehea/ehea_main.c
700
cqe->num_bytes_transfered - 4);
drivers/net/ethernet/ibm/ehea/ehea_main.c
701
ehea_fill_skb(dev, skb, cqe, pr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
705
skb_arr_rq2_len, cqe);
drivers/net/ethernet/ibm/ehea/ehea_main.c
711
ehea_fill_skb(dev, skb, cqe, pr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
716
skb_arr_rq3_len, cqe);
drivers/net/ethernet/ibm/ehea/ehea_main.c
722
ehea_fill_skb(dev, skb, cqe, pr);
drivers/net/ethernet/ibm/ehea/ehea_main.c
728
if (cqe->status & EHEA_CQE_VLAN_TAG_XTRACT)
drivers/net/ethernet/ibm/ehea/ehea_main.c
730
cqe->vlan_tag);
drivers/net/ethernet/ibm/ehea/ehea_main.c
735
port_reset = ehea_treat_poll_error(pr, rq, cqe,
drivers/net/ethernet/ibm/ehea/ehea_main.c
741
cqe = ehea_poll_rq1(qp, &wqe_index);
drivers/net/ethernet/ibm/ehea/ehea_main.c
805
struct ehea_cqe *cqe;
drivers/net/ethernet/ibm/ehea/ehea_main.c
813
cqe = ehea_poll_cq(send_cq);
drivers/net/ethernet/ibm/ehea/ehea_main.c
814
while (cqe && (quota > 0)) {
drivers/net/ethernet/ibm/ehea/ehea_main.c
820
if (cqe->wr_id == SWQE_RESTART_CHECK) {
drivers/net/ethernet/ibm/ehea/ehea_main.c
826
if (cqe->status & EHEA_CQE_STAT_ERR_MASK) {
drivers/net/ethernet/ibm/ehea/ehea_main.c
828
cqe->status);
drivers/net/ethernet/ibm/ehea/ehea_main.c
831
ehea_dump(cqe, sizeof(*cqe), "Send CQE");
drivers/net/ethernet/ibm/ehea/ehea_main.c
833
if (cqe->status & EHEA_CQE_STAT_RESET_MASK) {
drivers/net/ethernet/ibm/ehea/ehea_main.c
841
ehea_dump(cqe, sizeof(*cqe), "CQE");
drivers/net/ethernet/ibm/ehea/ehea_main.c
843
if (likely(EHEA_BMASK_GET(EHEA_WR_ID_TYPE, cqe->wr_id)
drivers/net/ethernet/ibm/ehea/ehea_main.c
846
index = EHEA_BMASK_GET(EHEA_WR_ID_INDEX, cqe->wr_id);
drivers/net/ethernet/ibm/ehea/ehea_main.c
852
swqe_av += EHEA_BMASK_GET(EHEA_WR_ID_REFILL, cqe->wr_id);
drivers/net/ethernet/ibm/ehea/ehea_main.c
855
cqe = ehea_poll_cq(send_cq);
drivers/net/ethernet/ibm/ehea/ehea_main.c
872
return cqe;
drivers/net/ethernet/ibm/ehea/ehea_main.c
882
struct ehea_cqe *cqe;
drivers/net/ethernet/ibm/ehea/ehea_main.c
897
cqe = ehea_poll_rq1(pr->qp, &wqe_index);
drivers/net/ethernet/ibm/ehea/ehea_main.c
900
if (!cqe && !cqe_skb)
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
364
struct ehea_cq *ehea_create_cq(struct ehea_adapter *adapter, int cqe,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1100
err = qmem_alloc(pfvf->dev, &cq->cqe, cq->cqe_cnt, cq->cqe_size);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1105
cq->cqe_base = cq->cqe->base;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1122
aq->cq.base = cq->cqe->iova;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
1558
qmem_free(pf->dev, cq->cqe);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
107
struct nix_cqe_tx_s *cqe,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
110
struct nix_send_comp_s *snd_comp = &cqe->comp;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1278
struct nix_cqe_rx_s *cqe;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
128
struct nix_cqe_tx_s *cqe,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1298
cqe = (struct nix_cqe_rx_s *)otx2_get_next_cqe(cq);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1302
if (!cqe)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1304
if (cqe->sg.segs > 1) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1305
otx2_free_rcv_seg(pfvf, cqe, cq->cq_idx);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1308
iova = cqe->sg.seg_addr - OTX2_HEAD_ROOM;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
131
struct nix_send_comp_s *snd_comp = &cqe->comp;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1323
struct nix_cqe_tx_s *cqe;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1336
cqe = (struct nix_cqe_tx_s *)otx2_get_next_cqe(cq);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1340
if (!cqe)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1342
sg = &sq->sg[cqe->comp.sqe_id];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1490
struct nix_cqe_rx_s *cqe,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1511
xsk_buff->data_end = xsk_buff->data + cqe->sg.seg_size;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1516
iova = cqe->sg.seg_addr - OTX2_HEAD_ROOM;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1524
cqe->sg.seg_size, true);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1538
cqe->sg.seg_addr,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
1539
cqe->sg.seg_size,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
225
struct nix_cqe_rx_s *cqe, struct sk_buff *skb)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
241
hash = cqe->hdr.flow_tag;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
246
static void otx2_free_rcv_seg(struct otx2_nic *pfvf, struct nix_cqe_rx_s *cqe,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
249
struct nix_rx_sg_s *sg = &cqe->sg;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
255
end = start + ((cqe->parse.desc_sizem1 + 1) * 16);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
267
struct nix_cqe_rx_s *cqe, int qidx)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
270
struct nix_rx_parse_s *parse = &cqe->parse;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
325
if (cqe->sg.segs)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
326
otx2_free_rcv_seg(pfvf, cqe, qidx);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
333
struct nix_cqe_rx_s *cqe, bool *need_xdp_flush)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
335
struct nix_rx_parse_s *parse = &cqe->parse;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
336
struct nix_rx_sg_s *sg = &cqe->sg;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
346
if (otx2_check_rcv_errors(pfvf, cqe, cq->cq_idx)) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
354
if (otx2_xdp_rcv_pkt_handler(pfvf, pfvf->xdp_prog, cqe, cq,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
363
end = start + ((cqe->parse.desc_sizem1 + 1) * 16);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
375
otx2_set_rxhash(pfvf, cqe, skb);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
398
struct nix_cqe_rx_s *cqe;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
409
cqe = (struct nix_cqe_rx_s *)CQE_ADDR(cq, cq->cq_head);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
410
if (cqe->hdr.cqe_type == NIX_XQE_TYPE_INVALID ||
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
411
!cqe->sg.seg_addr) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
419
otx2_rcv_pkt_handler(pfvf, napi, cq, cqe, &need_xdp_flush);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
42
struct nix_cqe_rx_s *cqe,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
421
cqe->hdr.cqe_type = NIX_XQE_TYPE_INVALID;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
422
cqe->sg.seg_addr = 0x00;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
469
struct nix_cqe_tx_s *cqe;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
490
cqe = (struct nix_cqe_tx_s *)otx2_get_next_cqe(cq);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
491
if (unlikely(!cqe)) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
498
otx2_xdp_snd_pkt_handler(pfvf, sq, cqe, &xsk_frames);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
501
cqe, budget, &tx_pkts, &tx_bytes);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
503
cqe->hdr.cqe_type = NIX_XQE_TYPE_INVALID;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.h
153
struct qmem *cqe;
drivers/net/ethernet/mellanox/mlx4/en_clock.c
50
u64 mlx4_en_get_cqe_ts(struct mlx4_cqe *cqe)
drivers/net/ethernet/mellanox/mlx4/en_clock.c
53
struct mlx4_ts_cqe *ts_cqe = (struct mlx4_ts_cqe *)cqe;
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1680
struct mlx4_cqe *cqe = NULL;
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1682
cqe = mlx4_en_get_cqe(cq->buf, j, priv->cqe_size) +
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
1684
cqe->owner_sr_opcode = MLX4_CQE_OWNER_MASK;
drivers/net/ethernet/mellanox/mlx4/en_rx.c
596
static int check_csum(struct mlx4_cqe *cqe, struct sk_buff *skb, void *va,
drivers/net/ethernet/mellanox/mlx4/en_rx.c
614
hw_checksum = csum_unfold((__force __sum16)cqe->checksum);
drivers/net/ethernet/mellanox/mlx4/en_rx.c
616
if (cqe->vlan_my_qpn & cpu_to_be32(MLX4_CQE_CVLAN_PRESENT_MASK) &&
drivers/net/ethernet/mellanox/mlx4/en_rx.c
623
if (cqe->status & cpu_to_be16(MLX4_CQE_STATUS_IPV6))
drivers/net/ethernet/mellanox/mlx4/en_rx.c
637
struct mlx4_cqe *cqe;
drivers/net/ethernet/mellanox/mlx4/en_rx.c
651
mlx4_en_get_cqe_ts(_ctx->cqe));
drivers/net/ethernet/mellanox/mlx4/en_rx.c
659
struct mlx4_cqe *cqe = _ctx->cqe;
drivers/net/ethernet/mellanox/mlx4/en_rx.c
666
*hash = be32_to_cpu(cqe->immed_rss_invalid);
drivers/net/ethernet/mellanox/mlx4/en_rx.c
667
status = cqe->status;
drivers/net/ethernet/mellanox/mlx4/en_rx.c
676
if (cqe->ipv6_ext_mask)
drivers/net/ethernet/mellanox/mlx4/en_rx.c
694
struct mlx4_cqe *cqe;
drivers/net/ethernet/mellanox/mlx4/en_rx.c
712
cqe = mlx4_en_get_cqe(cq->buf, index, priv->cqe_size) + factor;
drivers/net/ethernet/mellanox/mlx4/en_rx.c
715
while (XNOR(cqe->owner_sr_opcode & MLX4_CQE_OWNER_MASK,
drivers/net/ethernet/mellanox/mlx4/en_rx.c
734
if (unlikely((cqe->owner_sr_opcode & MLX4_CQE_OPCODE_MASK) ==
drivers/net/ethernet/mellanox/mlx4/en_rx.c
737
((struct mlx4_err_cqe *)cqe)->vendor_err_syndrome,
drivers/net/ethernet/mellanox/mlx4/en_rx.c
738
((struct mlx4_err_cqe *)cqe)->syndrome);
drivers/net/ethernet/mellanox/mlx4/en_rx.c
741
if (unlikely(cqe->badfcs_enc & MLX4_CQE_BAD_FCS)) {
drivers/net/ethernet/mellanox/mlx4/en_rx.c
784
length = be32_to_cpu(cqe->byte_cnt);
drivers/net/ethernet/mellanox/mlx4/en_rx.c
804
mxbuf.cqe = cqe;
drivers/net/ethernet/mellanox/mlx4/en_rx.c
862
u64 timestamp = mlx4_en_get_cqe_ts(cqe);
drivers/net/ethernet/mellanox/mlx4/en_rx.c
875
if ((cqe->status & cpu_to_be16(MLX4_CQE_STATUS_TCP |
drivers/net/ethernet/mellanox/mlx4/en_rx.c
877
(cqe->status & cpu_to_be16(MLX4_CQE_STATUS_IPOK)) &&
drivers/net/ethernet/mellanox/mlx4/en_rx.c
878
cqe->checksum == cpu_to_be16(0xffff)) {
drivers/net/ethernet/mellanox/mlx4/en_rx.c
882
(cqe->vlan_my_qpn & cpu_to_be32(MLX4_CQE_L2_TUNNEL));
drivers/net/ethernet/mellanox/mlx4/en_rx.c
890
(cqe->status & cpu_to_be16(MLX4_CQE_STATUS_IP_ANY))))
drivers/net/ethernet/mellanox/mlx4/en_rx.c
892
if (check_csum(cqe, skb, va, dev->features))
drivers/net/ethernet/mellanox/mlx4/en_rx.c
907
be32_to_cpu(cqe->immed_rss_invalid),
drivers/net/ethernet/mellanox/mlx4/en_rx.c
910
if ((cqe->vlan_my_qpn &
drivers/net/ethernet/mellanox/mlx4/en_rx.c
914
be16_to_cpu(cqe->sl_vid));
drivers/net/ethernet/mellanox/mlx4/en_rx.c
915
else if ((cqe->vlan_my_qpn &
drivers/net/ethernet/mellanox/mlx4/en_rx.c
919
be16_to_cpu(cqe->sl_vid));
drivers/net/ethernet/mellanox/mlx4/en_rx.c
934
cqe = mlx4_en_get_cqe(cq->buf, index, priv->cqe_size) + factor;
drivers/net/ethernet/mellanox/mlx4/en_tx.c
430
struct mlx4_cqe *cqe;
drivers/net/ethernet/mellanox/mlx4/en_tx.c
454
cqe = mlx4_en_get_cqe(buf, index, priv->cqe_size) + factor;
drivers/net/ethernet/mellanox/mlx4/en_tx.c
461
while (XNOR(cqe->owner_sr_opcode & MLX4_CQE_OWNER_MASK,
drivers/net/ethernet/mellanox/mlx4/en_tx.c
471
if (unlikely((cqe->owner_sr_opcode & MLX4_CQE_OPCODE_MASK) ==
drivers/net/ethernet/mellanox/mlx4/en_tx.c
474
mlx4_en_handle_err_cqe(priv, (struct mlx4_err_cqe *)cqe, index,
drivers/net/ethernet/mellanox/mlx4/en_tx.c
478
new_index = be16_to_cpu(cqe->wqe_index) & size_mask;
drivers/net/ethernet/mellanox/mlx4/en_tx.c
487
timestamp = mlx4_en_get_cqe_ts(cqe);
drivers/net/ethernet/mellanox/mlx4/en_tx.c
507
cqe = mlx4_en_get_cqe(buf, index, priv->cqe_size) + factor;
drivers/net/ethernet/mellanox/mlx4/mlx4_en.h
800
u64 mlx4_en_get_cqe_ts(struct mlx4_cqe *cqe);
drivers/net/ethernet/mellanox/mlx5/core/en.h
526
struct mlx5_cqe64 *cqe;
drivers/net/ethernet/mellanox/mlx5/core/en/ptp.c
190
struct mlx5_cqe64 *cqe,
drivers/net/ethernet/mellanox/mlx5/core/en/ptp.c
196
u8 metadata_id = PTP_WQE_CTR2IDX(be16_to_cpu(cqe->wqe_counter));
drivers/net/ethernet/mellanox/mlx5/core/en/ptp.c
197
bool is_err_cqe = !!MLX5E_RX_ERR_CQE(cqe);
drivers/net/ethernet/mellanox/mlx5/core/en/ptp.c
219
hwtstamp = mlx5e_cqe_ts_to_ns(sq->ptp_cyc2time, sq->clock, get_cqe_ts(cqe));
drivers/net/ethernet/mellanox/mlx5/core/en/ptp.c
222
ptpsq->cq_stats->cqe++;
drivers/net/ethernet/mellanox/mlx5/core/en/ptp.c
240
struct mlx5_cqe64 *cqe;
drivers/net/ethernet/mellanox/mlx5/core/en/ptp.c
248
cqe = mlx5_cqwq_get_cqe(cqwq);
drivers/net/ethernet/mellanox/mlx5/core/en/ptp.c
249
if (!cqe)
drivers/net/ethernet/mellanox/mlx5/core/en/ptp.c
255
mlx5e_ptp_handle_ts_cqe(ptpsq, cqe,
drivers/net/ethernet/mellanox/mlx5/core/en/ptp.c
257
} while ((++work_done < budget) && (cqe = mlx5_cqwq_get_cqe(cqwq)));
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
700
void mlx5e_rep_tc_receive(struct mlx5_cqe64 *cqe, struct mlx5e_rq *rq,
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
712
reg_c0 = (be32_to_cpu(cqe->sop_drop_qpn) & MLX5E_TC_FLOW_ID_MASK);
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
724
reg_c1 = be32_to_cpu(cqe->ft_metadata);
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
744
if (!mlx5e_tc_update_skb(cqe, skb, mapping_ctx, reg_c0, ct_priv,
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.h
39
void mlx5e_rep_tc_receive(struct mlx5_cqe64 *cqe, struct mlx5e_rq *rq,
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.h
68
mlx5e_rep_tc_receive(struct mlx5_cqe64 *cqe, struct mlx5e_rq *rq,
drivers/net/ethernet/mellanox/mlx5/core/en/txrx.h
231
static inline u16 mlx5e_shampo_get_cqe_header_index(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/en/txrx.h
233
return be16_to_cpu(cqe->shampo.header_entry_index) & (rq->mpwqe.shampo->hd_per_wq - 1);
drivers/net/ethernet/mellanox/mlx5/core/en/txrx.h
37
#define MLX5E_RX_ERR_CQE(cqe) (get_cqe_opcode(cqe) != MLX5_CQE_RESP_SEND)
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
186
_ctx->rq->clock, get_cqe_ts(_ctx->cqe));
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
241
const struct mlx5_cqe64 *cqe = _ctx->cqe;
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
247
*hash = be32_to_cpu(cqe->rss_hash_result);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
249
hash_type = cqe->rss_hash_type;
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
263
const struct mlx5_cqe64 *cqe = _ctx->cqe;
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
265
if (!cqe_has_vlan(cqe))
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
269
*vlan_tci = be16_to_cpu(cqe->vlan_info);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
280
struct mlx5_cqe64 *cqe;
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
289
ts = get_cqe_ts(priv->cqe);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
658
struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
723
.cqe = cqe,
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
747
struct mlx5_cqe64 *cqe;
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
759
cqe = mlx5_cqwq_get_cqe(&cq->wq);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
760
if (!cqe)
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
776
wqe_counter = be16_to_cpu(cqe->wqe_counter);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
785
mlx5e_free_xdpsq_desc(sq, wi, &xsk_frames, &bq, cq, cqe);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
788
if (unlikely(get_cqe_opcode(cqe) != MLX5_CQE_REQ)) {
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
791
get_cqe_opcode(cqe));
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
793
(struct mlx5_err_cqe *)cqe);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
796
} while ((++i < MLX5E_TX_CQ_POLL_BUDGET) && (cqe = mlx5_cqwq_get_cqe(&cq->wq)));
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.c
252
struct mlx5_cqe64 *cqe,
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.c
274
mxbuf->cqe = cqe;
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.c
309
struct mlx5_cqe64 *cqe,
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.c
323
mxbuf->cqe = cqe;
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.h
16
struct mlx5_cqe64 *cqe,
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.h
22
struct mlx5_cqe64 *cqe,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.h
153
static inline bool mlx5_ipsec_is_rx_flow(struct mlx5_cqe64 *cqe) { return false; }
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.h
75
static inline bool mlx5_ipsec_is_rx_flow(struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.h
77
return MLX5_IPSEC_METADATA_MARKER(be32_to_cpu(cqe->ft_metadata));
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_rx.c
591
struct mlx5_cqe64 *cqe, u32 *cqe_bcnt)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_rx.c
595
switch (get_cqe_tls_offload(cqe)) {
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_txrx.h
23
struct mlx5_cqe64 *cqe, u32 *cqe_bcnt);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_txrx.h
92
struct mlx5_cqe64 *cqe,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/macsec.c
1681
struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/macsec.c
1684
u32 macsec_meta_data = be32_to_cpu(cqe->ft_metadata);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/macsec.h
32
static inline bool mlx5e_macsec_is_rx_flow(struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/macsec.h
34
return MLX5_MACSEC_METADATA_MARKER(be32_to_cpu(cqe->ft_metadata));
drivers/net/ethernet/mellanox/mlx5/core/en_accel/macsec.h
38
struct mlx5_cqe64 *cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/macsec.h
46
static inline bool mlx5e_macsec_is_rx_flow(struct mlx5_cqe64 *cqe) { return false; }
drivers/net/ethernet/mellanox/mlx5/core/en_accel/macsec.h
49
struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/psp_rxtx.c
123
struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/psp_rxtx.c
125
u32 psp_meta_data = be32_to_cpu(cqe->ft_metadata);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/psp_rxtx.h
109
static inline bool mlx5e_psp_is_rx_flow(struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/psp_rxtx.h
116
struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/psp_rxtx.h
84
static inline bool mlx5e_psp_is_rx_flow(struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/psp_rxtx.h
86
return MLX5_PSP_METADATA_MARKER(be32_to_cpu(cqe->ft_metadata));
drivers/net/ethernet/mellanox/mlx5/core/en_accel/psp_rxtx.h
90
struct mlx5_cqe64 *cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1168
struct mlx5_cqe64 *cqe;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1171
while ((cqe = mlx5_cqwq_get_cqe_enhanced_comp(cqwq)))
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1174
while ((cqe = mlx5_cqwq_get_cqe(cqwq)))
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
2159
struct mlx5_cqe64 *cqe = mlx5_cqwq_get_wqe(&cq->wq, i);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
2161
cqe->op_own = 0xf1;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
2162
cqe->validity_iteration_count = 0xff;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1001
ipv4->ttl = cqe->lro.min_ttl;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1007
mlx5e_lro_update_tcp_hdr(cqe, tcp);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1009
csum_unfold((__force __sum16)cqe->check_sum));
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1020
ipv6->hop_limit = cqe->lro.min_ttl;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1023
mlx5e_lro_update_tcp_hdr(cqe, tcp);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1025
csum_unfold((__force __sum16)cqe->check_sum));
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1073
struct mlx5_cqe64 *cqe,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1077
u32 header_index = mlx5e_shampo_get_cqe_header_index(rq, cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1087
static void *mlx5e_shampo_get_hdr(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1093
mlx5e_shampo_get_hd_buf_info(rq, cqe, &di, &head_offset);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1101
static void mlx5e_shampo_update_fin_psh_flags(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1109
last_hd_addr = mlx5e_shampo_get_hdr(rq, cqe, len);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1116
struct mlx5_cqe64 *cqe, bool match)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1124
mlx5e_shampo_update_fin_psh_flags(rq, cqe, tcp);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1144
struct mlx5_cqe64 *cqe, bool match)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1152
mlx5e_shampo_update_fin_psh_flags(rq, cqe, tcp);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1164
static void mlx5e_shampo_update_hdr(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe, bool match)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1181
mlx5e_shampo_update_ipv4_tcp_hdr(rq, ipv4, cqe, match);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1191
mlx5e_shampo_update_ipv6_tcp_hdr(rq, ipv6, cqe, match);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1197
static inline void mlx5e_skb_set_hash(struct mlx5_cqe64 *cqe,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1200
u8 cht = cqe->rss_hash_type;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1204
skb_set_hash(skb, be32_to_cpu(cqe->rss_hash_result), ht);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1323
struct mlx5_cqe64 *cqe,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1343
get_cqe_tls_offload(cqe))
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1363
skb->csum = csum_unfold((__force __sum16)cqe->check_sum);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1365
if (unlikely(mlx5e_psp_is_rx_flow(cqe))) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1379
if (likely((cqe->hds_ip_ext & CQE_L3_OK) &&
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1380
(cqe->hds_ip_ext & CQE_L4_OK))) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1382
if (cqe_is_tunneled(cqe)) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
139
struct mlx5_cqe64 *cqe = mlx5_cqwq_get_wqe(wq, ci);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1398
static inline bool mlx5e_build_rx_skb(struct mlx5_cqe64 *cqe,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1403
u8 lro_num_seg = get_cqe_lro_num_seg(cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1409
if (unlikely(get_cqe_tls_offload(cqe)))
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
141
cqe->op_own = op_own;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1410
mlx5e_ktls_handle_rx_skb(rq, skb, cqe, &cqe_bcnt);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1412
if (unlikely(mlx5e_psp_is_rx_flow(cqe))) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1413
if (mlx5e_psp_offload_handle_rx_skb(netdev, skb, cqe))
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1417
if (unlikely(mlx5_ipsec_is_rx_flow(cqe)))
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1419
be32_to_cpu(cqe->ft_metadata));
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1421
if (unlikely(mlx5e_macsec_is_rx_flow(cqe)))
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1422
mlx5e_macsec_offload_handle_rx_skb(netdev, skb, cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1425
unsigned int hdrlen = mlx5e_lro_update_hdr(skb, cqe, cqe_bcnt);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1439
rq->clock, get_cqe_ts(cqe));
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1443
mlx5e_skb_set_hash(cqe, skb);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1445
if (cqe_has_vlan(cqe)) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1447
be16_to_cpu(cqe->vlan_info));
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1451
skb->mark = be32_to_cpu(cqe->sop_drop_qpn) & MLX5E_TC_FLOW_ID_MASK;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1453
mlx5e_handle_csum(netdev, cqe, rq, skb, !!lro_num_seg);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1455
if (unlikely(cqe->ml_path & MLX5E_CE_BIT_MASK))
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1467
struct mlx5_cqe64 *cqe,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
147
struct mlx5_cqe64 *cqe = mlx5_cqwq_get_wqe(wq, ci);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1478
if (mlx5e_build_rx_skb(cqe, cqe_bcnt, rq, skb))
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
149
cqe->op_own = op_own;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1490
struct mlx5_cqe64 *cqe,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1498
return mlx5e_build_rx_skb(cqe, cqe_bcnt, rq, skb);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1522
static void mlx5e_fill_mxbuf(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1528
mxbuf->cqe = cqe;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1534
struct mlx5_cqe64 *cqe, u32 cqe_bcnt)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1559
mlx5e_fill_mxbuf(rq, cqe, va, rx_headroom, rq->buff.frame0_sz,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1582
struct mlx5_cqe64 *cqe, u32 cqe_bcnt)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1609
mlx5e_fill_mxbuf(rq, cqe, va, rx_headroom, rq->buff.frame0_sz,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1677
static void trigger_report(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1679
struct mlx5_err_cqe *err_cqe = (struct mlx5_err_cqe *)cqe;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1689
static void mlx5e_handle_rx_err_cqe(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1691
trigger_report(rq, cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1695
static void mlx5e_handle_rx_cqe(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1703
ci = mlx5_wq_cyc_ctr2ix(wq, be16_to_cpu(cqe->wqe_counter));
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1705
cqe_bcnt = be32_to_cpu(cqe->byte_cnt);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1707
if (unlikely(MLX5E_RX_ERR_CQE(cqe))) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1708
mlx5e_handle_rx_err_cqe(rq, cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1716
rq, wi, cqe, cqe_bcnt);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1724
if (mlx5e_complete_rx_cqe(rq, cqe, cqe_bcnt, skb))
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1727
if (mlx5e_cqe_regb_chain(cqe))
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1728
if (!mlx5e_tc_update_skb_nic(cqe, skb)) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1740
static void mlx5e_handle_rx_cqe_rep(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1752
ci = mlx5_wq_cyc_ctr2ix(wq, be16_to_cpu(cqe->wqe_counter));
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1754
cqe_bcnt = be32_to_cpu(cqe->byte_cnt);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1756
if (unlikely(MLX5E_RX_ERR_CQE(cqe))) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1757
mlx5e_handle_rx_err_cqe(rq, cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1764
rq, wi, cqe, cqe_bcnt);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1772
if (mlx5e_complete_rx_cqe(rq, cqe, cqe_bcnt, skb))
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1778
mlx5e_rep_tc_receive(cqe, rq, skb);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1784
static void mlx5e_handle_rx_cqe_mpwrq_rep(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1786
u16 cstrides = mpwrq_get_cqe_consumed_strides(cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1787
u16 wqe_id = be16_to_cpu(cqe->wqe_id);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1789
u16 stride_ix = mpwrq_get_cqe_stride_index(cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1800
if (unlikely(MLX5E_RX_ERR_CQE(cqe))) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1801
mlx5e_handle_rx_err_cqe(rq, cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1805
if (unlikely(mpwrq_is_filler_cqe(cqe))) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1813
cqe_bcnt = mpwrq_get_cqe_byte_cnt(cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1818
rq, wi, cqe, cqe_bcnt, head_offset, page_idx);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1822
if (mlx5e_complete_rx_cqe(rq, cqe, cqe_bcnt, skb))
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1825
mlx5e_rep_tc_receive(cqe, rq, skb);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1833
mlx5_wq_ll_pop(wq, cqe->wqe_id, &wqe->next.next_wqe_index);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1865
struct mlx5_cqe64 *cqe, u16 cqe_bcnt, u32 head_offset,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1885
u8 lro_num_seg = get_cqe_lro_num_seg(cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1932
mlx5e_fill_mxbuf(rq, cqe, va, linear_hr, linear_frame_sz,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
197
struct mlx5_cqe64 *cqe,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
204
left = get_cqe_enhanced_num_mini_cqes(cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2045
struct mlx5_cqe64 *cqe, u16 cqe_bcnt, u32 head_offset,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2077
mlx5e_fill_mxbuf(rq, cqe, va, rx_headroom, rq->buff.frame0_sz,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2103
struct mlx5_cqe64 *cqe, u16 header_index)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2105
u16 head_size = cqe->shampo.header_size;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2120
mlx5e_shampo_get_hd_buf_info(rq, cqe, &di, &head_offset);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
214
memcpy(cqd->mini_arr, cqe, sizeof(struct mlx5_cqe64));
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2143
mlx5e_shampo_flush_skb(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe, bool match)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2156
mlx5e_shampo_update_hdr(rq, cqe, match);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2175
static void mlx5e_handle_rx_cqe_mpwrq_shampo(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2177
u16 data_bcnt = mpwrq_get_cqe_byte_cnt(cqe) - cqe->shampo.header_size;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2178
u16 header_index = mlx5e_shampo_get_cqe_header_index(rq, cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2179
u32 wqe_offset = be32_to_cpu(cqe->shampo.data_offset);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2180
u16 cstrides = mpwrq_get_cqe_consumed_strides(cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2182
u32 cqe_bcnt = mpwrq_get_cqe_byte_cnt(cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2183
u16 wqe_id = be16_to_cpu(cqe->wqe_id);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2185
u16 head_size = cqe->shampo.header_size;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2187
bool flush = cqe->shampo.flush;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2188
bool match = cqe->shampo.match;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2197
if (unlikely(MLX5E_RX_ERR_CQE(cqe))) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2198
mlx5e_handle_rx_err_cqe(rq, cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2202
if (unlikely(mpwrq_is_filler_cqe(cqe))) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2210
mlx5e_shampo_flush_skb(rq, cqe, match);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2215
*skb = mlx5e_skb_from_cqe_shampo(rq, wi, cqe, header_index);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2225
*skb = mlx5e_skb_from_cqe_mpwrq_nonlinear(rq, wi, cqe,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2246
last_hd_addr = mlx5e_shampo_get_hdr(rq, cqe, len);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2267
if (mlx5e_shampo_complete_rx_cqe(rq, cqe, cqe_bcnt, *skb)) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2272
mlx5e_shampo_flush_skb(rq, cqe, match);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2282
mlx5_wq_ll_pop(wq, cqe->wqe_id, &wqe->next.next_wqe_index);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2285
static void mlx5e_handle_rx_cqe_mpwrq(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2287
u16 cstrides = mpwrq_get_cqe_consumed_strides(cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2288
u16 wqe_id = be16_to_cpu(cqe->wqe_id);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2290
u16 stride_ix = mpwrq_get_cqe_stride_index(cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2301
if (unlikely(MLX5E_RX_ERR_CQE(cqe))) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2302
mlx5e_handle_rx_err_cqe(rq, cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2306
if (unlikely(mpwrq_is_filler_cqe(cqe))) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2314
cqe_bcnt = mpwrq_get_cqe_byte_cnt(cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2320
rq, wi, cqe, cqe_bcnt, head_offset,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2325
if (mlx5e_complete_rx_cqe(rq, cqe, cqe_bcnt, skb))
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2328
if (mlx5e_cqe_regb_chain(cqe))
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2329
if (!mlx5e_tc_update_skb_nic(cqe, skb)) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2342
mlx5_wq_ll_pop(wq, cqe->wqe_id, &wqe->next.next_wqe_index);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2349
struct mlx5_cqe64 *cqe, *title_cqe = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2353
cqe = mlx5_cqwq_get_cqe_enhanced_comp(cqwq);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2354
if (!cqe)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2358
(mlx5_get_cqe_format(cqe) == MLX5_COMPRESSED)) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2364
if (mlx5_get_cqe_format(cqe) == MLX5_COMPRESSED) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2371
mlx5e_decompress_enhanced_cqe(rq, cqwq, cqe,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2375
title_cqe = cqe;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2380
rq, cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2383
(cqe = mlx5_cqwq_get_cqe_enhanced_comp(cqwq)));
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2398
struct mlx5_cqe64 *cqe;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2404
while (work_done < budget_rem && (cqe = mlx5_cqwq_get_cqe(cqwq))) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2405
if (mlx5_get_cqe_format(cqe) == MLX5_COMPRESSED) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2415
rq, cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2462
struct mlx5_cqe64 *cqe,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2475
qpn = be32_to_cpu(cqe->sop_drop_qpn) & 0xffffff;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2491
flags_rqpn = be32_to_cpu(cqe->flags_rqpn);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2517
skb->csum = csum_unfold((__force __sum16)cqe->check_sum);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2526
rq->clock, get_cqe_ts(cqe));
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2530
mlx5e_skb_set_hash(cqe, skb);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2544
static void mlx5i_handle_rx_cqe(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2552
ci = mlx5_wq_cyc_ctr2ix(wq, be16_to_cpu(cqe->wqe_counter));
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2554
cqe_bcnt = be32_to_cpu(cqe->byte_cnt);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2556
if (unlikely(MLX5E_RX_ERR_CQE(cqe))) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2564
rq, wi, cqe, cqe_bcnt);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2568
mlx5i_complete_rx_cqe(rq, cqe, cqe_bcnt, skb);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2634
static void mlx5e_trap_handle_rx_cqe(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2643
trap_id = get_cqe_flow_tag(cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2644
ci = mlx5_wq_cyc_ctr2ix(wq, be16_to_cpu(cqe->wqe_counter));
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2646
cqe_bcnt = be32_to_cpu(cqe->byte_cnt);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2648
if (unlikely(MLX5E_RX_ERR_CQE(cqe))) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2653
skb = mlx5e_skb_from_cqe_nonlinear(rq, wi, cqe, cqe_bcnt);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2657
if (mlx5e_complete_rx_cqe(rq, cqe, cqe_bcnt, skb))
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
66
struct mlx5_cqe64 *cqe, u16 cqe_bcnt, u32 head_offset,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
70
struct mlx5_cqe64 *cqe, u16 cqe_bcnt, u32 head_offset,
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
72
static void mlx5e_handle_rx_cqe(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
73
static void mlx5e_handle_rx_cqe_mpwrq(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
74
static void mlx5e_handle_rx_cqe_mpwrq_shampo(struct mlx5e_rq *rq, struct mlx5_cqe64 *cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
790
struct mlx5_cqe64 *cqe;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
797
cqe = mlx5_cqwq_get_cqe(&cq->wq);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
798
if (likely(!cqe))
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
813
wqe_counter = be16_to_cpu(cqe->wqe_counter);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
825
if (last_wqe && unlikely(get_cqe_opcode(cqe) != MLX5_CQE_REQ)) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
828
get_cqe_opcode(cqe));
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
834
(struct mlx5_err_cqe *)cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
863
} while ((++i < MLX5E_TX_CQ_POLL_BUDGET) && (cqe = mlx5_cqwq_get_cqe(&cq->wq)));
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
91
struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
96
memcpy(title, cqe, sizeof(struct mlx5_cqe64));
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
962
static void mlx5e_lro_update_tcp_hdr(struct mlx5_cqe64 *cqe, struct tcphdr *tcp)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
964
u8 l4_hdr_type = get_cqe_l4_hdr_type(cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
969
tcp->psh = get_cqe_lro_tcppsh(cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
973
tcp->ack_seq = cqe->lro.ack_seq_num;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
974
tcp->window = cqe->lro.tcp_win;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
979
struct mlx5_cqe64 *cqe,
drivers/net/ethernet/mellanox/mlx5/core/en_stats.c
1191
ts_stats->pkts += stats->cqe;
drivers/net/ethernet/mellanox/mlx5/core/en_stats.c
2383
{ MLX5E_DECLARE_PTP_CQ_STAT(struct mlx5e_ptp_cq_stats, cqe) },
drivers/net/ethernet/mellanox/mlx5/core/en_stats.h
471
u64 cqe;
drivers/net/ethernet/mellanox/mlx5/core/en_tc.c
5769
bool mlx5e_tc_update_skb(struct mlx5_cqe64 *cqe, struct sk_buff *skb,
drivers/net/ethernet/mellanox/mlx5/core/en_tc.c
5806
bool mlx5e_tc_update_skb_nic(struct mlx5_cqe64 *cqe, struct sk_buff *skb)
drivers/net/ethernet/mellanox/mlx5/core/en_tc.c
5814
reg_b = be32_to_cpu(cqe->ft_metadata);
drivers/net/ethernet/mellanox/mlx5/core/en_tc.c
5822
return mlx5e_tc_update_skb(cqe, skb, mapping_ctx, mapped_obj_id, ct_priv, zone_restore_id,
drivers/net/ethernet/mellanox/mlx5/core/en_tc.h
371
static inline bool mlx5e_cqe_regb_chain(struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/en_tc.h
375
reg_b = be32_to_cpu(cqe->ft_metadata);
drivers/net/ethernet/mellanox/mlx5/core/en_tc.h
387
bool mlx5e_tc_update_skb_nic(struct mlx5_cqe64 *cqe, struct sk_buff *skb);
drivers/net/ethernet/mellanox/mlx5/core/en_tc.h
388
bool mlx5e_tc_update_skb(struct mlx5_cqe64 *cqe, struct sk_buff *skb,
drivers/net/ethernet/mellanox/mlx5/core/en_tc.h
396
static inline bool mlx5e_cqe_regb_chain(struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/en_tc.h
399
mlx5e_tc_update_skb_nic(struct mlx5_cqe64 *cqe, struct sk_buff *skb)
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
718
struct mlx5_cqe64 *cqe, int napi_budget)
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
722
u64 ts = get_cqe_ts(cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
738
struct mlx5_cqe64 *cqe, int napi_budget)
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
745
mlx5e_consume_skb(sq, skb, cqe, napi_budget);
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
764
struct mlx5_cqe64 *cqe;
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
776
cqe = mlx5_cqwq_get_cqe(&cq->wq);
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
777
if (!cqe)
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
802
wqe_counter = be16_to_cpu(cqe->wqe_counter);
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
814
mlx5e_consume_skb(sq, wi->skb, cqe, napi_budget);
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
827
mlx5e_tx_wi_consume_fifo_skbs(sq, wi, cqe, napi_budget);
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
834
if (unlikely(get_cqe_opcode(cqe) == MLX5_CQE_REQ_ERR)) {
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
838
(struct mlx5_err_cqe *)cqe);
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
845
} while ((++i < MLX5E_TX_CQ_POLL_BUDGET) && (cqe = mlx5_cqwq_get_cqe(&cq->wq)));
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
251
struct mlx5_cqe64 *cqe, u8 status)
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
256
ix = be16_to_cpu(cqe->wqe_counter) & (conn->qp.rq.size - 1);
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
276
buf->sg[0].size = be32_to_cpu(cqe->byte_cnt);
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
291
struct mlx5_cqe64 *cqe, u8 status)
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
299
ix = be16_to_cpu(cqe->wqe_counter) & (conn->qp.sq.size - 1);
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
333
struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
337
opcode = get_cqe_opcode(cqe);
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
341
status = ((struct mlx5_err_cqe *)cqe)->syndrome;
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
344
mlx5_fpga_conn_sq_cqe(conn, cqe, status);
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
348
status = ((struct mlx5_err_cqe *)cqe)->syndrome;
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
351
mlx5_fpga_conn_rq_cqe(conn, cqe, status);
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
368
struct mlx5_cqe64 *cqe;
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
371
cqe = mlx5_cqwq_get_cqe(&conn->cq.wq);
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
372
if (!cqe)
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
377
mlx5_fpga_conn_handle_cqe(conn, cqe);
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
418
struct mlx5_cqe64 *cqe;
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
443
cqe = mlx5_cqwq_get_wqe(&conn->cq.wq, i);
drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c
444
cqe->op_own = MLX5_CQE_INVALID << 4 | MLX5_CQE_OWNER_MASK;
drivers/net/ethernet/mellanox/mlx5/core/lib/aso.c
394
struct mlx5_cqe64 *cqe;
drivers/net/ethernet/mellanox/mlx5/core/lib/aso.c
396
cqe = mlx5_cqwq_get_cqe(&cq->wq);
drivers/net/ethernet/mellanox/mlx5/core/lib/aso.c
397
if (!cqe)
drivers/net/ethernet/mellanox/mlx5/core/lib/aso.c
405
if (unlikely(get_cqe_opcode(cqe) != MLX5_CQE_REQ)) {
drivers/net/ethernet/mellanox/mlx5/core/lib/aso.c
409
get_cqe_opcode(cqe));
drivers/net/ethernet/mellanox/mlx5/core/lib/aso.c
411
err_cqe = (struct mlx5_err_cqe *)cqe;
drivers/net/ethernet/mellanox/mlx5/core/lib/aso.c
65
struct mlx5_cqe64 *cqe = mlx5_cqwq_get_wqe(&cq->wq, i);
drivers/net/ethernet/mellanox/mlx5/core/lib/aso.c
67
cqe->op_own = 0xf1;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1221
struct mlx5_cqe64 cqe;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1263
ret = mlx5hws_cmd_generate_wqe(mdev, &attr, &cqe);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1269
if ((get_cqe_opcode(&cqe) == MLX5_CQE_REQ) &&
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1270
(be32_to_cpu(cqe.byte_cnt) >> 31 == 0)) {
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
349
struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
351
u8 wqe_opcode = cqe ? be32_to_cpu(cqe->sop_drop_qpn) >> 24 : 0;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
353
u32 opcode = cqe ? get_cqe_opcode(cqe) : 0;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
396
if (!cqe) {
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
408
struct mlx5_err_cqe *err_cqe = (struct mlx5_err_cqe *)cqe;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
430
HWS_PTR_TO_ID(rule), be32_to_cpu(cqe->byte_cnt));
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
434
(be32_to_cpu(cqe->byte_cnt) & 0x80000000) ?
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
439
((be32_to_cpu(cqe->byte_cnt) & 0x00000003) == 1) ?
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
441
((be32_to_cpu(cqe->byte_cnt) & 0x00000003) == 2) ?
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
445
HWS_PTR_TO_ID(rule), be32_to_cpu(cqe->sop_drop_qpn));
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
454
be32_to_cpu(cqe->sop_drop_qpn) & 0xffffff);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
461
struct mlx5_cqe64 *cqe)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
473
hws_send_engine_dump_error_cqe(queue, priv, cqe);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
520
struct mlx5_cqe64 *cqe,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
529
if (!cqe || (likely(be32_to_cpu(cqe->byte_cnt) >> 31 == 0) &&
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
530
likely(get_cqe_opcode(cqe) == MLX5_CQE_REQ))) {
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
539
&status, cqe);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
606
struct mlx5_cqe64 *cqe;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
610
cqe = mlx5_cqwq_get_cqe(&cq->wq);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
611
if (!cqe)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
614
cqe_opcode = get_cqe_opcode(cqe);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
621
wqe_cnt = be16_to_cpu(cqe->wqe_counter) & sq->buf_mask;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
631
hws_send_engine_update(queue, cqe, priv, res, polled, res_nb, wqe_cnt);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
884
struct mlx5_cqe64 *cqe;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
900
cqe = mlx5_cqwq_get_wqe(&cq->wq, i);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
901
cqe->op_own = 0xf1;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_send.c
1059
struct mlx5_cqe64 *cqe;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_send.c
1082
cqe = mlx5_cqwq_get_wqe(&cq->wq, i);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_send.c
1083
cqe->op_own = MLX5_CQE_INVALID << 4 | MLX5_CQE_OWNER_MASK;
drivers/net/ethernet/mellanox/mlx5/core/wc.c
327
struct mlx5_cqe64 *cqe;
drivers/net/ethernet/mellanox/mlx5/core/wc.c
329
cqe = mlx5_cqwq_get_cqe(&cq->wq);
drivers/net/ethernet/mellanox/mlx5/core/wc.c
330
if (!cqe)
drivers/net/ethernet/mellanox/mlx5/core/wc.c
338
if (get_cqe_opcode(cqe) == MLX5_CQE_REQ) {
drivers/net/ethernet/mellanox/mlx5/core/wc.c
339
int wqe_counter = be16_to_cpu(cqe->wqe_counter);
drivers/net/ethernet/mellanox/mlx5/core/wc.c
66
struct mlx5_cqe64 *cqe = mlx5_cqwq_get_wqe(&cq->wq, i);
drivers/net/ethernet/mellanox/mlx5/core/wc.c
68
cqe->op_own = 0xf1;
drivers/net/ethernet/mellanox/mlx5/core/wq.h
202
struct mlx5_cqe64 *cqe = mlx5_frag_buf_get_wqe(&wq->fbc, ix);
drivers/net/ethernet/mellanox/mlx5/core/wq.h
205
cqe += wq->fbc.log_stride == 7;
drivers/net/ethernet/mellanox/mlx5/core/wq.h
207
return cqe;
drivers/net/ethernet/mellanox/mlx5/core/wq.h
233
struct mlx5_cqe64 *cqe = mlx5_cqwq_get_wqe(wq, ci);
drivers/net/ethernet/mellanox/mlx5/core/wq.h
234
u8 cqe_ownership_bit = cqe->op_own & MLX5_CQE_OWNER_MASK;
drivers/net/ethernet/mellanox/mlx5/core/wq.h
243
return cqe;
drivers/net/ethernet/mellanox/mlx5/core/wq.h
251
struct mlx5_cqe64 *cqe;
drivers/net/ethernet/mellanox/mlx5/core/wq.h
253
cqe = mlx5_cqwq_get_wqe(wq, ci);
drivers/net/ethernet/mellanox/mlx5/core/wq.h
254
if (cqe->validity_iteration_count != sw_validity_iteration_count)
drivers/net/ethernet/mellanox/mlx5/core/wq.h
260
return cqe;
drivers/net/ethernet/mellanox/mlxsw/pci.c
635
enum mlxsw_pci_cqe_v cqe_v, char *cqe)
drivers/net/ethernet/mellanox/mlxsw/pci.c
642
ts_type = mlxsw_pci_cqe2_time_stamp_type_get(cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
648
mlxsw_skb_cb(skb)->cqe_ts.sec = mlxsw_pci_cqe2_time_stamp_sec_get(cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
650
mlxsw_pci_cqe2_time_stamp_nsec_get(cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
657
char *cqe, int budget)
drivers/net/ethernet/mellanox/mlxsw/pci.c
676
mlxsw_pci_skb_cb_ts_set(mlxsw_pci, skb, cqe_v, cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
692
const char *cqe)
drivers/net/ethernet/mellanox/mlxsw/pci.c
696
if (mlxsw_pci_cqe2_tx_lag_get(cqe)) {
drivers/net/ethernet/mellanox/mlxsw/pci.c
698
cb->rx_md_info.tx_lag_id = mlxsw_pci_cqe2_tx_lag_id_get(cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
700
mlxsw_pci_cqe2_tx_lag_subport_get(cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
704
mlxsw_pci_cqe2_tx_system_port_get(cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
714
static void mlxsw_pci_cqe_rdq_md_init(struct sk_buff *skb, const char *cqe)
drivers/net/ethernet/mellanox/mlxsw/pci.c
718
cb->rx_md_info.tx_congestion = mlxsw_pci_cqe2_mirror_cong_get(cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
725
cb->rx_md_info.latency = mlxsw_pci_cqe2_mirror_latency_get(cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
731
cb->rx_md_info.tx_tc = mlxsw_pci_cqe2_mirror_tclass_get(cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
737
mlxsw_pci_cqe_rdq_md_tx_port_init(skb, cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
744
enum mlxsw_pci_cqe_v cqe_v, char *cqe)
drivers/net/ethernet/mellanox/mlxsw/pci.c
760
byte_count = mlxsw_pci_cqe_byte_count_get(cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
761
if (mlxsw_pci_cqe_crc_get(cqe_v, cqe))
drivers/net/ethernet/mellanox/mlxsw/pci.c
782
if (mlxsw_pci_cqe_lag_get(cqe_v, cqe)) {
drivers/net/ethernet/mellanox/mlxsw/pci.c
784
rx_info.u.lag_id = mlxsw_pci_cqe_lag_id_get(cqe_v, cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
786
mlxsw_pci_cqe_lag_subport_get(cqe_v, cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
789
rx_info.u.sys_port = mlxsw_pci_cqe_system_port_get(cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
792
rx_info.trap_id = mlxsw_pci_cqe_trap_id_get(cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
799
cookie_index = mlxsw_pci_cqe2_user_def_val_orig_pkt_len_get(cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
804
rx_info.mirror_reason = mlxsw_pci_cqe2_mirror_reason_get(cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
805
mlxsw_pci_cqe_rdq_md_init(skb, cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
808
mlxsw_pci_cqe_rdq_md_tx_port_init(skb, cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
811
mlxsw_pci_skb_cb_ts_set(mlxsw_pci, skb, cqe_v, cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
854
char *cqe;
drivers/net/ethernet/mellanox/mlxsw/pci.c
860
while ((cqe = mlxsw_pci_cq_sw_cqe_get(q))) {
drivers/net/ethernet/mellanox/mlxsw/pci.c
861
u16 wqe_counter = mlxsw_pci_cqe_wqe_counter_get(cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
862
u8 sendq = mlxsw_pci_cqe_sr_get(q->u.cq.v, cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
863
u8 dqn = mlxsw_pci_cqe_dqn_get(q->u.cq.v, cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
876
wqe_counter, q->u.cq.v, cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
916
char *cqe;
drivers/net/ethernet/mellanox/mlxsw/pci.c
918
while ((cqe = mlxsw_pci_cq_sw_cqe_get(q))) {
drivers/net/ethernet/mellanox/mlxsw/pci.c
919
u16 wqe_counter = mlxsw_pci_cqe_wqe_counter_get(cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
920
u8 sendq = mlxsw_pci_cqe_sr_get(q->u.cq.v, cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
921
u8 dqn = mlxsw_pci_cqe_dqn_get(q->u.cq.v, cqe);
drivers/net/ethernet/mellanox/mlxsw/pci.c
934
memcpy(ncqe, cqe, q->elem_size);
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
116
static inline u32 mlxsw_pci_cqe_##name##_get(enum mlxsw_pci_cqe_v v, char *cqe) \
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
121
return mlxsw_pci_cqe##v0##_##name##_get(cqe); \
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
123
return mlxsw_pci_cqe##v1##_##name##_get(cqe); \
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
125
return mlxsw_pci_cqe##v2##_##name##_get(cqe); \
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
129
char *cqe, u32 val) \
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
134
mlxsw_pci_cqe##v0##_##name##_set(cqe, val); \
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
137
mlxsw_pci_cqe##v1##_##name##_set(cqe, val); \
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
140
mlxsw_pci_cqe##v2##_##name##_set(cqe, val); \
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
158
MLXSW_ITEM32(pci, cqe, system_port, 0x00, 0, 16);
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
169
MLXSW_ITEM32(pci, cqe, wqe_counter, 0x04, 16, 16);
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
176
MLXSW_ITEM32(pci, cqe, byte_count, 0x04, 0, 14);
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
190
MLXSW_ITEM32(pci, cqe, trap_id, 0x08, 0, 10);
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
286
static inline u16 mlxsw_pci_cqe2_mirror_cong_get(const char *cqe)
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
288
u16 cong_high = mlxsw_pci_cqe2_mirror_cong_high_get(cqe);
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
289
u16 cong_low = mlxsw_pci_cqe2_mirror_cong_low_get(cqe);
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
343
static inline u64 mlxsw_pci_cqe2_time_stamp_get(const char *cqe)
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
345
u64 ts_high = mlxsw_pci_cqe2_time_stamp_high_get(cqe);
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
346
u64 ts_low = mlxsw_pci_cqe2_time_stamp_low_get(cqe);
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
351
static inline u8 mlxsw_pci_cqe2_time_stamp_sec_get(const char *cqe)
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
353
u64 full_ts = mlxsw_pci_cqe2_time_stamp_get(cqe);
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
358
static inline u32 mlxsw_pci_cqe2_time_stamp_nsec_get(const char *cqe)
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
360
u64 full_ts = mlxsw_pci_cqe2_time_stamp_get(cqe);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1459
struct gdma_cqe *cqe;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1461
cqe = &cq_cqe[cq->head % num_cqe];
drivers/net/ethernet/microsoft/mana/gdma_main.c
1462
owner_bits = cqe->cqe_info.owner_bits;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1479
comp->wq_num = cqe->cqe_info.wq_num;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1480
comp->is_sq = cqe->cqe_info.is_sq;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1481
memcpy(comp->cqe_data, cqe->cqe_data, GDMA_COMP_DATA_SIZE);
drivers/net/ethernet/microsoft/mana/mana_en.c
1925
struct mana_rxcomp_oob *cqe, struct mana_rxq *rxq)
drivers/net/ethernet/microsoft/mana/mana_en.c
1929
uint pkt_len = cqe->ppi[0].pkt_len;
drivers/net/ethernet/microsoft/mana/mana_en.c
1967
if ((ndev->features & NETIF_F_RXCSUM) && cqe->rx_iphdr_csum_succeed) {
drivers/net/ethernet/microsoft/mana/mana_en.c
1968
if (cqe->rx_tcp_csum_succeed || cqe->rx_udp_csum_succeed)
drivers/net/ethernet/microsoft/mana/mana_en.c
1972
if (cqe->rx_hashtype != 0 && (ndev->features & NETIF_F_RXHASH)) {
drivers/net/ethernet/microsoft/mana/mana_en.c
1973
hash_value = cqe->ppi[0].pkt_hash;
drivers/net/ethernet/microsoft/mana/mana_en.c
1975
if (cqe->rx_hashtype & MANA_HASH_L4)
drivers/net/ethernet/microsoft/mana/mana_en.c
1981
if (cqe->rx_vlantag_present) {
drivers/net/ethernet/microsoft/mana/mana_en.c
1982
u16 vlan_tci = cqe->rx_vlan_id;
drivers/net/ethernet/microsoft/mana/mana_en.c
2100
struct gdma_comp *cqe)
drivers/net/ethernet/microsoft/mana/mana_en.c
2102
struct mana_rxcomp_oob *oob = (struct mana_rxcomp_oob *)cqe->cqe_data;
drivers/net/ethernet/qlogic/qed/qed_l2.c
2854
u8 rss_id, struct eth_slow_path_rx_cqe *cqe)
drivers/net/ethernet/qlogic/qed/qed_l2.c
2857
cqe);
drivers/net/ethernet/qlogic/qed/qed_ll2.c
530
union core_rx_cqe_union *cqe = NULL;
drivers/net/ethernet/qlogic/qed/qed_ll2.c
551
cqe =
drivers/net/ethernet/qlogic/qed/qed_ll2.c
559
cq_old_idx, cq_new_idx, cqe->rx_cqe_sp.type);
drivers/net/ethernet/qlogic/qed/qed_ll2.c
561
switch (cqe->rx_cqe_sp.type) {
drivers/net/ethernet/qlogic/qed/qed_ll2.c
564
cqe, &flags);
drivers/net/ethernet/qlogic/qed/qed_ll2.c
569
cqe, &flags,
drivers/net/ethernet/qlogic/qed/qed_ll2.c
652
union core_rx_cqe_union *cqe = NULL;
drivers/net/ethernet/qlogic/qed/qed_ll2.c
668
cqe = qed_chain_consume(&p_rx->rcq_chain);
drivers/net/ethernet/qlogic/qed/qed_ll2.c
670
cqe_type = cqe->rx_cqe_sp.type;
drivers/net/ethernet/qlogic/qed/qed_ll2.c
674
&cqe->rx_cqe_sp))
drivers/net/ethernet/qlogic/qed/qed_ll2.c
683
p_cqe_fp = &cqe->rx_cqe_fp;
drivers/net/ethernet/qlogic/qed/qed_sp.h
43
struct eth_slow_path_rx_cqe *cqe);
drivers/net/ethernet/qlogic/qed/qed_spq.c
453
struct eth_slow_path_rx_cqe *cqe,
drivers/net/ethernet/qlogic/qed/qed_spq.c
463
return qed_spq_completion(p_hwfn, cqe->echo, 0, NULL);
drivers/net/ethernet/qlogic/qed/qed_spq.c
467
struct eth_slow_path_rx_cqe *cqe)
drivers/net/ethernet/qlogic/qed/qed_spq.c
471
rc = qed_cqe_completion(p_hwfn, cqe, PROTOCOLID_ETH);
drivers/net/ethernet/qlogic/qed/qed_spq.c
475
cqe->ramrod_cmd_id);
drivers/net/ethernet/qlogic/qede/qede_ethtool.c
1552
union eth_rx_cqe *cqe;
drivers/net/ethernet/qlogic/qede/qede_ethtool.c
1579
cqe = (union eth_rx_cqe *)qed_chain_consume(&rxq->rx_comp_ring);
drivers/net/ethernet/qlogic/qede/qede_ethtool.c
1584
fp_cqe = &cqe->fast_path_regular;
drivers/net/ethernet/qlogic/qede/qede_fp.c
1000
if (unlikely(cqe->num_of_bds != tpa_info->frag_id + 1))
drivers/net/ethernet/qlogic/qede/qede_fp.c
1003
cqe->num_of_bds, tpa_info->frag_id);
drivers/net/ethernet/qlogic/qede/qede_fp.c
1004
if (unlikely(skb->len != le16_to_cpu(cqe->total_packet_len)))
drivers/net/ethernet/qlogic/qede/qede_fp.c
1007
le16_to_cpu(cqe->total_packet_len), skb->len);
drivers/net/ethernet/qlogic/qede/qede_fp.c
1016
NAPI_GRO_CB(skb)->count = le16_to_cpu(cqe->num_of_coalesced_segs);
drivers/net/ethernet/qlogic/qede/qede_fp.c
1065
static bool qede_pkt_is_ip_fragmented(struct eth_fast_path_rx_reg_cqe *cqe,
drivers/net/ethernet/qlogic/qede/qede_fp.c
1068
u8 tun_pars_flg = cqe->tunnel_pars_flags.flags;
drivers/net/ethernet/qlogic/qede/qede_fp.c
1085
struct eth_fast_path_rx_reg_cqe *cqe,
drivers/net/ethernet/qlogic/qede/qede_fp.c
1164
qede_recycle_rx_bd_ring(rxq, cqe->bd_num);
drivers/net/ethernet/qlogic/qede/qede_fp.c
1173
struct eth_fast_path_rx_reg_cqe *cqe,
drivers/net/ethernet/qlogic/qede/qede_fp.c
1176
u16 pkt_len = le16_to_cpu(cqe->pkt_len);
drivers/net/ethernet/qlogic/qede/qede_fp.c
1184
for (num_frags = cqe->bd_num - 1; num_frags > 0; num_frags--) {
drivers/net/ethernet/qlogic/qede/qede_fp.c
1227
union eth_rx_cqe *cqe,
drivers/net/ethernet/qlogic/qede/qede_fp.c
1232
qede_tpa_start(edev, rxq, &cqe->fast_path_tpa_start);
drivers/net/ethernet/qlogic/qede/qede_fp.c
1235
qede_tpa_cont(edev, rxq, &cqe->fast_path_tpa_cont);
drivers/net/ethernet/qlogic/qede/qede_fp.c
1238
return qede_tpa_end(edev, fp, &cqe->fast_path_tpa_end);
drivers/net/ethernet/qlogic/qede/qede_fp.c
1252
union eth_rx_cqe *cqe;
drivers/net/ethernet/qlogic/qede/qede_fp.c
1259
cqe = (union eth_rx_cqe *)qed_chain_consume(&rxq->rx_comp_ring);
drivers/net/ethernet/qlogic/qede/qede_fp.c
1260
cqe_type = cqe->fast_path_regular.type;
drivers/net/ethernet/qlogic/qede/qede_fp.c
1266
sp_cqe = (struct eth_slow_path_rx_cqe *)cqe;
drivers/net/ethernet/qlogic/qede/qede_fp.c
1273
return qede_rx_process_tpa_cqe(edev, fp, rxq, cqe, cqe_type);
drivers/net/ethernet/qlogic/qede/qede_fp.c
1281
fp_cqe = &cqe->fast_path_regular;
drivers/net/ethernet/qlogic/qede/qede_fp.c
1292
flags = cqe->fast_path_regular.pars_flags.flags;
drivers/net/ethernet/qlogic/qede/qede_fp.c
1332
qede_ptp_record_rx_ts(edev, cqe, skb);
drivers/net/ethernet/qlogic/qede/qede_fp.c
651
struct eth_fast_path_rx_tpa_start_cqe *cqe)
drivers/net/ethernet/qlogic/qede/qede_fp.c
653
u16 parsing_flags = le16_to_cpu(cqe->pars_flags.flags);
drivers/net/ethernet/qlogic/qede/qede_fp.c
661
skb_shinfo(skb)->gso_size = __le16_to_cpu(cqe->len_on_first_bd) -
drivers/net/ethernet/qlogic/qede/qede_fp.c
662
cqe->header_len;
drivers/net/ethernet/qlogic/qede/qede_fp.c
834
struct eth_fast_path_rx_tpa_start_cqe *cqe)
drivers/net/ethernet/qlogic/qede/qede_fp.c
836
struct qede_agg_info *tpa_info = &rxq->tpa_info[cqe->tpa_agg_index];
drivers/net/ethernet/qlogic/qede/qede_fp.c
841
pad = cqe->placement_offset + rxq->rx_headroom;
drivers/net/ethernet/qlogic/qede/qede_fp.c
844
le16_to_cpu(cqe->len_on_first_bd),
drivers/net/ethernet/qlogic/qede/qede_fp.c
865
if ((le16_to_cpu(cqe->pars_flags.flags) >>
drivers/net/ethernet/qlogic/qede/qede_fp.c
868
tpa_info->vlan_tag = le16_to_cpu(cqe->vlan_tag);
drivers/net/ethernet/qlogic/qede/qede_fp.c
872
qede_get_rxhash(tpa_info->skb, cqe->bitfields, cqe->rss_hash);
drivers/net/ethernet/qlogic/qede/qede_fp.c
875
qede_set_gro_params(edev, tpa_info->skb, cqe);
drivers/net/ethernet/qlogic/qede/qede_fp.c
878
if (likely(cqe->bw_ext_bd_len_list[0]))
drivers/net/ethernet/qlogic/qede/qede_fp.c
879
qede_fill_frag_skb(edev, rxq, cqe->tpa_agg_index,
drivers/net/ethernet/qlogic/qede/qede_fp.c
880
le16_to_cpu(cqe->bw_ext_bd_len_list[0]));
drivers/net/ethernet/qlogic/qede/qede_fp.c
882
if (unlikely(cqe->bw_ext_bd_len_list[1])) {
drivers/net/ethernet/qlogic/qede/qede_fp.c
960
struct eth_fast_path_rx_tpa_cont_cqe *cqe)
drivers/net/ethernet/qlogic/qede/qede_fp.c
964
for (i = 0; cqe->len_list[i] && i < ARRAY_SIZE(cqe->len_list); i++)
drivers/net/ethernet/qlogic/qede/qede_fp.c
965
qede_fill_frag_skb(edev, rxq, cqe->tpa_agg_index,
drivers/net/ethernet/qlogic/qede/qede_fp.c
966
le16_to_cpu(cqe->len_list[i]));
drivers/net/ethernet/qlogic/qede/qede_fp.c
975
struct eth_fast_path_rx_tpa_end_cqe *cqe)
drivers/net/ethernet/qlogic/qede/qede_fp.c
982
tpa_info = &rxq->tpa_info[cqe->tpa_agg_index];
drivers/net/ethernet/qlogic/qede/qede_fp.c
989
for (i = 0; cqe->len_list[i] && i < ARRAY_SIZE(cqe->len_list); i++)
drivers/net/ethernet/qlogic/qede/qede_fp.c
990
qede_fill_frag_skb(edev, rxq, cqe->tpa_agg_index,
drivers/net/ethernet/qlogic/qede/qede_fp.c
991
le16_to_cpu(cqe->len_list[i]));
drivers/net/ethernet/qlogic/qede/qede_ptp.h
27
union eth_rx_cqe *cqe,
drivers/net/ethernet/qlogic/qede/qede_ptp.h
31
if (unlikely(le16_to_cpu(cqe->fast_path_regular.pars_flags.flags) &
drivers/net/ethernet/qlogic/qede/qede_ptp.h
33
if (likely(le16_to_cpu(cqe->fast_path_regular.pars_flags.flags)
drivers/nvme/host/apple.c
620
struct nvme_completion *cqe = &q->cqes[idx];
drivers/nvme/host/apple.c
621
__u16 command_id = READ_ONCE(cqe->command_id);
drivers/nvme/host/apple.c
633
if (!nvme_try_complete_req(req, cqe->status, cqe->result) &&
drivers/nvme/host/fc.c
1903
struct nvme_completion *cqe = &op->rsp_iu.cqe;
drivers/nvme/host/fc.c
2008
sqe->common.command_id != cqe->command_id)) {
drivers/nvme/host/fc.c
2019
cqe->command_id);
drivers/nvme/host/fc.c
2022
result = cqe->result;
drivers/nvme/host/fc.c
2023
status = cqe->status;
drivers/nvme/host/pci.c
1534
struct nvme_completion *cqe = &nvmeq->cqes[idx];
drivers/nvme/host/pci.c
1535
__u16 command_id = READ_ONCE(cqe->command_id);
drivers/nvme/host/pci.c
1546
cqe->status, &cqe->result);
drivers/nvme/host/pci.c
1554
command_id, le16_to_cpu(cqe->sq_id));
drivers/nvme/host/pci.c
1558
trace_nvme_sq(req, cqe->sq_head, nvmeq->sq_tail);
drivers/nvme/host/pci.c
1559
if (!nvme_try_complete_req(req, cqe->status, cqe->result) &&
drivers/nvme/host/rdma.c
1581
container_of(wc->wr_cqe, struct nvme_rdma_qe, cqe);
drivers/nvme/host/rdma.c
1603
wr.wr_cqe = &qe->cqe;
drivers/nvme/host/rdma.c
1633
qe->cqe.done = nvme_rdma_recv_done;
drivers/nvme/host/rdma.c
1636
wr.wr_cqe = &qe->cqe;
drivers/nvme/host/rdma.c
1681
sqe->cqe.done = nvme_rdma_async_done;
drivers/nvme/host/rdma.c
1691
struct nvme_completion *cqe, struct ib_wc *wc)
drivers/nvme/host/rdma.c
1696
rq = nvme_find_rq(nvme_rdma_tagset(queue), cqe->command_id);
drivers/nvme/host/rdma.c
1700
cqe->command_id, queue->qp->qp_num);
drivers/nvme/host/rdma.c
1706
req->status = cqe->status;
drivers/nvme/host/rdma.c
1707
req->result = cqe->result;
drivers/nvme/host/rdma.c
1737
container_of(wc->wr_cqe, struct nvme_rdma_qe, cqe);
drivers/nvme/host/rdma.c
1740
struct nvme_completion *cqe = qe->data;
drivers/nvme/host/rdma.c
1764
cqe->command_id)))
drivers/nvme/host/rdma.c
1765
nvme_complete_async_event(&queue->ctrl->ctrl, cqe->status,
drivers/nvme/host/rdma.c
1766
&cqe->result);
drivers/nvme/host/rdma.c
1768
nvme_rdma_process_nvme_rsp(queue, cqe, wc);
drivers/nvme/host/rdma.c
2045
sqe->cqe.done = nvme_rdma_send_done;
drivers/nvme/host/rdma.c
51
struct ib_cqe cqe;
drivers/nvme/host/tcp.c
623
struct nvme_completion *cqe)
drivers/nvme/host/tcp.c
628
rq = nvme_find_rq(nvme_tcp_tagset(queue), cqe->command_id);
drivers/nvme/host/tcp.c
632
cqe->command_id, nvme_tcp_queue_id(queue));
drivers/nvme/host/tcp.c
639
req->status = cqe->status;
drivers/nvme/host/tcp.c
641
if (!nvme_try_complete_req(rq, req->status, cqe->result))
drivers/nvme/host/tcp.c
685
struct nvme_completion *cqe = &pdu->cqe;
drivers/nvme/host/tcp.c
695
cqe->command_id)))
drivers/nvme/host/tcp.c
696
nvme_complete_async_event(&queue->ctrl->ctrl, cqe->status,
drivers/nvme/host/tcp.c
697
&cqe->result);
drivers/nvme/host/tcp.c
699
ret = nvme_tcp_process_nvme_cqe(queue, cqe);
drivers/nvme/target/core.c
1167
req->cqe->result.u64 = 0;
drivers/nvme/target/core.c
1168
req->cqe->status = 0;
drivers/nvme/target/core.c
1169
req->cqe->sq_head = 0;
drivers/nvme/target/core.c
1472
req->cqe->result.u32 = IPO_IATTR_CONNECT_DATA(subsysnqn);
drivers/nvme/target/core.c
1494
req->cqe->result.u32 = IPO_IATTR_CONNECT_DATA(cntlid);
drivers/nvme/target/core.c
167
trace_nvmet_async_event(ctrl, req->cqe->result.u32);
drivers/nvme/target/core.c
745
req->cqe->sq_head = cpu_to_le16(req->sq->sqhd & 0x0000FFFF);
drivers/nvme/target/core.c
754
req->cqe->status = cpu_to_le16(status << 1);
drivers/nvme/target/core.c
774
req->cqe->status |= cpu_to_le16(1 << 14);
drivers/nvme/target/core.c
784
req->cqe->sq_id = cpu_to_le16(req->sq->qid);
drivers/nvme/target/core.c
785
req->cqe->command_id = req->cmd->common.command_id;
drivers/nvme/target/fabrics-cmd.c
184
req->cqe->result.u32 = IPO_IATTR_CONNECT_SQE(sqsize);
drivers/nvme/target/fabrics-cmd.c
200
req->cqe->result.u32 = IPO_IATTR_CONNECT_SQE(sqsize);
drivers/nvme/target/fabrics-cmd.c
225
req->cqe->sq_head = cpu_to_le16(0xffff);
drivers/nvme/target/fabrics-cmd.c
332
req->cqe->result.u32 = args.result;
drivers/nvme/target/fabrics-cmd.c
376
req->cqe->result.u32 = IPO_IATTR_CONNECT_SQE(qid);
drivers/nvme/target/fabrics-cmd.c
385
req->cqe->result.u32 = cpu_to_le32(nvmet_connect_result(ctrl, req->sq));
drivers/nvme/target/fabrics-cmd.c
84
req->cqe->result.u64 = cpu_to_le64(val);
drivers/nvme/target/fc.c
2153
struct nvme_completion *cqe = &ersp->cqe;
drivers/nvme/target/fc.c
2154
u32 *cqewd = (u32 *)cqe;
drivers/nvme/target/fc.c
2186
(le16_to_cpu(cqe->status) & 0xFFFE) || cqewd[0] || cqewd[1] ||
drivers/nvme/target/fc.c
2188
queue_90percent_full(fod->queue, le16_to_cpu(cqe->sq_head)))
drivers/nvme/target/fc.c
2459
struct nvme_completion *cqe = &fod->rspiubuf.cqe;
drivers/nvme/target/fc.c
2469
fod->queue->sqhd = cqe->sq_head;
drivers/nvme/target/fc.c
2479
memset(cqe, 0, sizeof(*cqe));
drivers/nvme/target/fc.c
2480
cqe->sq_head = fod->queue->sqhd; /* echo last cqe sqhd */
drivers/nvme/target/fc.c
2481
cqe->sq_id = cpu_to_le16(fod->queue->qid);
drivers/nvme/target/fc.c
2482
cqe->command_id = sqe->command_id;
drivers/nvme/target/fc.c
2483
cqe->status = cpu_to_le16(status);
drivers/nvme/target/fc.c
2555
fod->req.cqe = &fod->rspiubuf.cqe;
drivers/nvme/target/loop.c
106
cqe->command_id))) {
drivers/nvme/target/loop.c
107
nvme_complete_async_event(&queue->ctrl->ctrl, cqe->status,
drivers/nvme/target/loop.c
108
&cqe->result);
drivers/nvme/target/loop.c
112
rq = nvme_find_rq(nvme_loop_tagset(queue), cqe->command_id);
drivers/nvme/target/loop.c
116
cqe->command_id, nvme_loop_queue_idx(queue));
drivers/nvme/target/loop.c
120
if (!nvme_try_complete_req(rq, cqe->status, cqe->result))
drivers/nvme/target/loop.c
197
iod->req.cqe = &iod->cqe;
drivers/nvme/target/loop.c
21
struct nvme_completion cqe;
drivers/nvme/target/loop.c
97
struct nvme_completion *cqe = req->cqe;
drivers/nvme/target/nvmet.h
448
struct nvme_completion *cqe;
drivers/nvme/target/nvmet.h
508
req->cqe->result.u32 = cpu_to_le32(result);
drivers/nvme/target/passthru.c
241
req->cqe->result = nvme_req(rq)->result;
drivers/nvme/target/passthru.c
255
req->cqe->result = nvme_req(rq)->result;
drivers/nvme/target/pci-epf.c
1175
iod->status = le16_to_cpu(iod->cqe.status) >> 1;
drivers/nvme/target/pci-epf.c
1244
iod->status = le16_to_cpu(req->cqe->status) >> 1;
drivers/nvme/target/pci-epf.c
126
struct nvme_completion cqe;
drivers/nvme/target/pci-epf.c
1759
struct nvme_completion *cqe;
drivers/nvme/target/pci-epf.c
1790
cqe = &iod->cqe;
drivers/nvme/target/pci-epf.c
1791
cqe->sq_head = cpu_to_le16(READ_ONCE(iod->sq->head));
drivers/nvme/target/pci-epf.c
1792
cqe->sq_id = cpu_to_le16(iod->sq->qid);
drivers/nvme/target/pci-epf.c
1793
cqe->command_id = iod->cmd.common.command_id;
drivers/nvme/target/pci-epf.c
1794
cqe->status = cpu_to_le16((iod->status << 1) | cq->phase);
drivers/nvme/target/pci-epf.c
1799
le64_to_cpu(cqe->result.u64), cq->head, cq->tail,
drivers/nvme/target/pci-epf.c
1803
cqe, cq->qes);
drivers/nvme/target/pci-epf.c
686
iod->req.cqe = &iod->cqe;
drivers/nvme/target/rdma.c
1021
container_of(wc->wr_cqe, struct nvmet_rdma_cmd, cqe);
drivers/nvme/target/rdma.c
1317
__func__, queue->cq->cqe, qp_attr.cap.max_send_sge,
drivers/nvme/target/rdma.c
208
!rsp->req.cqe->status &&
drivers/nvme/target/rdma.c
335
c->cqe.done = nvmet_rdma_recv_done;
drivers/nvme/target/rdma.c
337
c->wr.wr_cqe = &c->cqe;
drivers/nvme/target/rdma.c
404
r->req.cqe = kmalloc_obj(*r->req.cqe);
drivers/nvme/target/rdma.c
405
if (!r->req.cqe)
drivers/nvme/target/rdma.c
408
r->send_sge.addr = ib_dma_map_single(ndev->device, r->req.cqe,
drivers/nvme/target/rdma.c
409
sizeof(*r->req.cqe), DMA_TO_DEVICE);
drivers/nvme/target/rdma.c
415
r->send_sge.length = sizeof(*r->req.cqe);
drivers/nvme/target/rdma.c
434
kfree(r->req.cqe);
drivers/nvme/target/rdma.c
443
sizeof(*r->req.cqe), DMA_TO_DEVICE);
drivers/nvme/target/rdma.c
444
kfree(r->req.cqe);
drivers/nvme/target/rdma.c
48
struct ib_cqe cqe;
drivers/nvme/target/rdma.c
812
rsp->req.cqe->status = cpu_to_le16(status << 1);
drivers/nvme/target/tcp.c
1075
le16_to_cpu(req->cqe->status));
drivers/nvme/target/tcp.c
1472
c->req.cqe = &c->rsp_pdu->cqe;
drivers/nvme/target/tcp.c
237
return nvmet_tcp_has_data_in(cmd) && !cmd->req.cqe->status;
drivers/nvme/target/tcp.c
244
!cmd->req.cqe->status;
drivers/nvme/target/tcp.c
491
pdu->command_id = cmd->req.cqe->command_id;
drivers/nvme/target/trace.h
122
__entry->cid = req->cqe->command_id;
drivers/nvme/target/trace.h
123
__entry->result = le64_to_cpu(req->cqe->result.u64);
drivers/nvme/target/trace.h
124
__entry->status = le16_to_cpu(req->cqe->status) >> 1;
drivers/nvme/target/zns.c
525
req->cqe->result.u64 =
drivers/scsi/bnx2fc/bnx2fc_hwi.c
1068
struct fcoe_cqe *cqe;
drivers/scsi/bnx2fc/bnx2fc_hwi.c
1087
cqe = &cq[cq_cons];
drivers/scsi/bnx2fc/bnx2fc_hwi.c
1089
while (((wqe = cqe->wqe) & FCOE_CQE_TOGGLE_BIT) ==
drivers/scsi/bnx2fc/bnx2fc_hwi.c
1101
cqe++;
drivers/scsi/bnx2fc/bnx2fc_hwi.c
1107
cqe = cq;
drivers/scsi/bnx2i/bnx2i.h
650
struct cqe *cq_virt;
drivers/scsi/bnx2i/bnx2i.h
654
struct cqe *cq_prod_qe;
drivers/scsi/bnx2i/bnx2i.h
655
struct cqe *cq_cons_qe;
drivers/scsi/bnx2i/bnx2i.h
656
struct cqe *cq_first_qe;
drivers/scsi/bnx2i/bnx2i.h
657
struct cqe *cq_last_qe;
drivers/scsi/bnx2i/bnx2i.h
774
struct cqe cqe;
drivers/scsi/bnx2i/bnx2i.h
870
struct cqe *cqe);
drivers/scsi/bnx2i/bnx2i_hwi.c
1338
struct cqe *cqe)
drivers/scsi/bnx2i/bnx2i_hwi.c
1348
resp_cqe = (struct bnx2i_cmd_response *)cqe;
drivers/scsi/bnx2i/bnx2i_hwi.c
1383
resp_cqe = (struct bnx2i_cmd_response *)cqe;
drivers/scsi/bnx2i/bnx2i_hwi.c
1435
struct cqe *cqe)
drivers/scsi/bnx2i/bnx2i_hwi.c
1444
login = (struct bnx2i_login_response *) cqe;
drivers/scsi/bnx2i/bnx2i_hwi.c
1503
struct cqe *cqe)
drivers/scsi/bnx2i/bnx2i_hwi.c
1512
text = (struct bnx2i_text_response *) cqe;
drivers/scsi/bnx2i/bnx2i_hwi.c
1564
struct cqe *cqe)
drivers/scsi/bnx2i/bnx2i_hwi.c
1571
tmf_cqe = (struct bnx2i_tmf_response *)cqe;
drivers/scsi/bnx2i/bnx2i_hwi.c
1603
struct cqe *cqe)
drivers/scsi/bnx2i/bnx2i_hwi.c
1610
logout = (struct bnx2i_logout_response *) cqe;
drivers/scsi/bnx2i/bnx2i_hwi.c
1649
struct cqe *cqe)
drivers/scsi/bnx2i/bnx2i_hwi.c
1655
nop_in = (struct bnx2i_nop_in_msg *)cqe;
drivers/scsi/bnx2i/bnx2i_hwi.c
1690
struct cqe *cqe)
drivers/scsi/bnx2i/bnx2i_hwi.c
1698
nop_in = (struct bnx2i_nop_in_msg *)cqe;
drivers/scsi/bnx2i/bnx2i_hwi.c
1742
struct cqe *cqe)
drivers/scsi/bnx2i/bnx2i_hwi.c
1750
async_cqe = (struct bnx2i_async_msg *)cqe;
drivers/scsi/bnx2i/bnx2i_hwi.c
1792
struct cqe *cqe)
drivers/scsi/bnx2i/bnx2i_hwi.c
1798
reject = (struct bnx2i_reject_msg *) cqe;
drivers/scsi/bnx2i/bnx2i_hwi.c
1829
struct cqe *cqe)
drivers/scsi/bnx2i/bnx2i_hwi.c
1835
cmd_clean_rsp = (struct bnx2i_cleanup_response *)cqe;
drivers/scsi/bnx2i/bnx2i_hwi.c
1871
&work->cqe);
drivers/scsi/bnx2i/bnx2i_hwi.c
1902
struct bnx2i_nop_in_msg *cqe)
drivers/scsi/bnx2i/bnx2i_hwi.c
1912
cqe->itt & ISCSI_CMD_RESPONSE_INDEX);
drivers/scsi/bnx2i/bnx2i_hwi.c
1933
memcpy(&bnx2i_work->cqe, cqe, sizeof(struct cqe));
drivers/scsi/bnx2i/bnx2i_hwi.c
1943
bnx2i_process_scsi_cmd_resp(session, bnx2i_conn, (struct cqe *)cqe);
drivers/scsi/bnx2i/bnx2i_init.c
444
work->bnx2i_conn, &work->cqe);
drivers/scsi/elx/efct/efct_hw.c
1847
efct_hw_wq_process_abort(void *arg, u8 *cqe, int status)
drivers/scsi/elx/efct/efct_hw.c
1861
ext = sli_fc_ext_status(&hw->sli, cqe);
drivers/scsi/elx/efct/efct_hw.c
2091
void (*callback)(void *arg, u8 *cqe, int status),
drivers/scsi/elx/efct/efct_hw.c
2315
u8 cqe[sizeof(struct sli4_mcqe)];
drivers/scsi/elx/efct/efct_hw.c
2324
while (!sli_cq_read(&hw->sli, cq->queue, cqe)) {
drivers/scsi/elx/efct/efct_hw.c
2327
status = sli_cq_parse(&hw->sli, cq->queue, cqe, &ctype, &rid);
drivers/scsi/elx/efct/efct_hw.c
2350
sli_cqe_async(&hw->sli, cqe);
drivers/scsi/elx/efct/efct_hw.c
2360
efct_hw_wq_process(hw, cq, cqe, status, rid);
drivers/scsi/elx/efct/efct_hw.c
2382
efct_hw_rqpair_process_rq(hw, cq, cqe);
drivers/scsi/elx/efct/efct_hw.c
2385
efct_hw_xabt_process(hw, cq, cqe, rid);
drivers/scsi/elx/efct/efct_hw.c
2413
u8 *cqe, int status, u16 rid)
drivers/scsi/elx/efct/efct_hw.c
2435
(*wqcb->callback)(wqcb->arg, cqe, status);
drivers/scsi/elx/efct/efct_hw.c
2440
u8 *cqe, u16 rid)
drivers/scsi/elx/efct/efct_hw.c
2972
void (*callback)(void *arg, u8 *cqe, int status),
drivers/scsi/elx/efct/efct_hw.c
361
efct_hw_wq_process_io(void *arg, u8 *cqe, int status)
drivers/scsi/elx/efct/efct_hw.c
365
struct sli4_fc_wcqe *wcqe = (void *)cqe;
drivers/scsi/elx/efct/efct_hw.c
379
sli_fc_els_did(&hw->sli, cqe, &ext);
drivers/scsi/elx/efct/efct_hw.c
380
len = sli_fc_response_length(&hw->sli, cqe);
drivers/scsi/elx/efct/efct_hw.c
386
len = sli_fc_response_length(&hw->sli, cqe);
drivers/scsi/elx/efct/efct_hw.c
389
len = sli_fc_io_length(&hw->sli, cqe);
drivers/scsi/elx/efct/efct_hw.c
392
len = sli_fc_io_length(&hw->sli, cqe);
drivers/scsi/elx/efct/efct_hw.c
406
ext = sli_fc_ext_status(&hw->sli, cqe);
drivers/scsi/elx/efct/efct_hw.h
335
void (*callback)(void *arg, u8 *cqe, int status);
drivers/scsi/elx/efct/efct_hw.h
648
void (*callback)(void *arg, u8 *cqe,
drivers/scsi/elx/efct/efct_hw.h
658
struct hw_cq *cq, u8 *cqe);
drivers/scsi/elx/efct/efct_hw.h
685
u8 *cqe, int status, u16 rid);
drivers/scsi/elx/efct/efct_hw.h
688
u8 *cqe, u16 rid);
drivers/scsi/elx/efct/efct_hw.h
698
void (*callback)(void *arg, u8 *cqe, int status),
drivers/scsi/elx/efct/efct_hw_queues.c
522
u8 *cqe)
drivers/scsi/elx/efct/efct_hw_queues.c
533
rq_status = sli_fc_rqe_rqid_and_index(&hw->sli, cqe,
drivers/scsi/elx/efct/efct_hw_queues.c
590
sli_fc_rqe_length(&hw->sli, cqe, &h_len, &p_len);
drivers/scsi/elx/efct/efct_hw_queues.c
593
seq->fcfi = sli_fc_rqe_fcfi(&hw->sli, cqe);
drivers/scsi/elx/efct/efct_unsol.c
208
efct_sframe_common_send_cb(void *arg, u8 *cqe, int status)
drivers/scsi/elx/libefc_sli/sli4.c
1342
sli_cq_parse(struct sli4 *sli4, struct sli4_queue *cq, u8 *cqe,
drivers/scsi/elx/libefc_sli/sli4.c
1347
if (!cq || !cqe || !etype) {
drivers/scsi/elx/libefc_sli/sli4.c
1349
sli4, cq, cqe, etype, q_id);
drivers/scsi/elx/libefc_sli/sli4.c
1355
struct sli4_mcqe *mcqe = (void *)cqe;
drivers/scsi/elx/libefc_sli/sli4.c
1365
rc = sli_fc_cqe_parse(sli4, cq, cqe, etype, q_id);
drivers/scsi/elx/libefc_sli/sli4.c
2570
u8 *cqe, enum sli4_qentry *etype, u16 *r_id)
drivers/scsi/elx/libefc_sli/sli4.c
2572
u8 code = cqe[SLI4_CQE_CODE_OFFSET];
drivers/scsi/elx/libefc_sli/sli4.c
2578
struct sli4_fc_wcqe *wcqe = (void *)cqe;
drivers/scsi/elx/libefc_sli/sli4.c
2594
((u32 *)cqe)[0], ((u32 *)cqe)[1],
drivers/scsi/elx/libefc_sli/sli4.c
2595
((u32 *)cqe)[2], ((u32 *)cqe)[3]);
drivers/scsi/elx/libefc_sli/sli4.c
2602
struct sli4_fc_async_rcqe *rcqe = (void *)cqe;
drivers/scsi/elx/libefc_sli/sli4.c
2611
struct sli4_fc_async_rcqe_v1 *rcqe = (void *)cqe;
drivers/scsi/elx/libefc_sli/sli4.c
2620
struct sli4_fc_optimized_write_cmd_cqe *optcqe = (void *)cqe;
drivers/scsi/elx/libefc_sli/sli4.c
2629
struct sli4_fc_optimized_write_data_cqe *dcqe = (void *)cqe;
drivers/scsi/elx/libefc_sli/sli4.c
2642
((u32 *)cqe)[3],
drivers/scsi/elx/libefc_sli/sli4.c
2649
struct sli4_fc_coalescing_rcqe *rcqe = (void *)cqe;
drivers/scsi/elx/libefc_sli/sli4.c
2658
struct sli4_fc_xri_aborted_cqe *xa = (void *)cqe;
drivers/scsi/elx/libefc_sli/sli4.c
2667
struct sli4_fc_wqec *wqec = (void *)cqe;
drivers/scsi/elx/libefc_sli/sli4.c
2686
sli_fc_response_length(struct sli4 *sli4, u8 *cqe)
drivers/scsi/elx/libefc_sli/sli4.c
2688
struct sli4_fc_wcqe *wcqe = (void *)cqe;
drivers/scsi/elx/libefc_sli/sli4.c
2694
sli_fc_io_length(struct sli4 *sli4, u8 *cqe)
drivers/scsi/elx/libefc_sli/sli4.c
2696
struct sli4_fc_wcqe *wcqe = (void *)cqe;
drivers/scsi/elx/libefc_sli/sli4.c
2702
sli_fc_els_did(struct sli4 *sli4, u8 *cqe, u32 *d_id)
drivers/scsi/elx/libefc_sli/sli4.c
2704
struct sli4_fc_wcqe *wcqe = (void *)cqe;
drivers/scsi/elx/libefc_sli/sli4.c
2715
sli_fc_ext_status(struct sli4 *sli4, u8 *cqe)
drivers/scsi/elx/libefc_sli/sli4.c
2717
struct sli4_fc_wcqe *wcqe = (void *)cqe;
drivers/scsi/elx/libefc_sli/sli4.c
2746
sli_fc_rqe_rqid_and_index(struct sli4 *sli4, u8 *cqe, u16 *rq_id, u32 *index)
drivers/scsi/elx/libefc_sli/sli4.c
2755
code = cqe[SLI4_CQE_CODE_OFFSET];
drivers/scsi/elx/libefc_sli/sli4.c
2759
struct sli4_fc_async_rcqe *rcqe = (void *)cqe;
drivers/scsi/elx/libefc_sli/sli4.c
2781
struct sli4_fc_async_rcqe_v1 *rcqe_v1 = (void *)cqe;
drivers/scsi/elx/libefc_sli/sli4.c
2803
struct sli4_fc_optimized_write_cmd_cqe *optcqe = (void *)cqe;
drivers/scsi/elx/libefc_sli/sli4.c
2824
struct sli4_fc_coalescing_rcqe *rcqe = (void *)cqe;
drivers/scsi/elx/libefc_sli/sli4.c
2846
struct sli4_fc_async_rcqe *rcqe = (void *)cqe;
drivers/scsi/elx/libefc_sli/sli4.c
2906
void *cqe = (u8 *)sli4->bmbx.virt + SLI4_BMBX_SIZE;
drivers/scsi/elx/libefc_sli/sli4.c
2925
if (le32_to_cpu(((struct sli4_mcqe *)cqe)->dw3_flags) &
drivers/scsi/elx/libefc_sli/sli4.c
2927
return sli_cqe_mq(sli4, cqe);
drivers/scsi/elx/libefc_sli/sli4.h
3830
sli_fc_rqe_length(struct sli4 *sli4, void *cqe, u32 *len_hdr,
drivers/scsi/elx/libefc_sli/sli4.h
3833
struct sli4_fc_async_rcqe *rcqe = cqe;
drivers/scsi/elx/libefc_sli/sli4.h
3847
sli_fc_rqe_fcfi(struct sli4 *sli4, void *cqe)
drivers/scsi/elx/libefc_sli/sli4.h
3849
u8 code = ((u8 *)cqe)[SLI4_CQE_CODE_OFFSET];
drivers/scsi/elx/libefc_sli/sli4.h
3854
struct sli4_fc_async_rcqe *rcqe = cqe;
drivers/scsi/elx/libefc_sli/sli4.h
3860
struct sli4_fc_async_rcqe_v1 *rcqev1 = cqe;
drivers/scsi/elx/libefc_sli/sli4.h
3866
struct sli4_fc_optimized_write_cmd_cqe *opt_wr = cqe;
drivers/scsi/elx/libefc_sli/sli4.h
4026
sli_cq_parse(struct sli4 *sli4, struct sli4_queue *cq, u8 *cqe,
drivers/scsi/elx/libefc_sli/sli4.h
4038
u8 *cqe, enum sli4_qentry *etype,
drivers/scsi/elx/libefc_sli/sli4.h
4040
u32 sli_fc_response_length(struct sli4 *sli4, u8 *cqe);
drivers/scsi/elx/libefc_sli/sli4.h
4041
u32 sli_fc_io_length(struct sli4 *sli4, u8 *cqe);
drivers/scsi/elx/libefc_sli/sli4.h
4042
int sli_fc_els_did(struct sli4 *sli4, u8 *cqe, u32 *d_id);
drivers/scsi/elx/libefc_sli/sli4.h
4043
u32 sli_fc_ext_status(struct sli4 *sli4, u8 *cqe);
drivers/scsi/elx/libefc_sli/sli4.h
4045
sli_fc_rqe_rqid_and_index(struct sli4 *sli4, u8 *cqe, u16 *rq_id, u32 *index);
drivers/scsi/lpfc/lpfc_init.c
7327
switch (bf_get(lpfc_trailer_code, &cq_event->cqe.mcqe_cmpl)) {
drivers/scsi/lpfc/lpfc_init.c
7330
&cq_event->cqe.acqe_link);
drivers/scsi/lpfc/lpfc_init.c
7333
lpfc_sli4_async_fip_evt(phba, &cq_event->cqe.acqe_fip);
drivers/scsi/lpfc/lpfc_init.c
7337
&cq_event->cqe.acqe_dcbx);
drivers/scsi/lpfc/lpfc_init.c
7341
&cq_event->cqe.acqe_grp5);
drivers/scsi/lpfc/lpfc_init.c
7344
lpfc_sli4_async_fc_evt(phba, &cq_event->cqe.acqe_fc);
drivers/scsi/lpfc/lpfc_init.c
7347
lpfc_sli4_async_sli_evt(phba, &cq_event->cqe.acqe_sli);
drivers/scsi/lpfc/lpfc_init.c
7354
&cq_event->cqe.mcqe_cmpl));
drivers/scsi/lpfc/lpfc_init.c
938
switch (bf_get(lpfc_wcqe_c_code, &cq_event->cqe.wcqe_cmpl)) {
drivers/scsi/lpfc/lpfc_nvme.c
1031
ptr = (uint32_t *)&ep->cqe.result.u64;
drivers/scsi/lpfc/lpfc_nvme.c
1035
ep->cqe.sq_head = sqhd;
drivers/scsi/lpfc/lpfc_nvme.c
1036
ep->cqe.sq_id = nCmd->sqid;
drivers/scsi/lpfc/lpfc_nvme.c
1037
ep->cqe.command_id = cid;
drivers/scsi/lpfc/lpfc_nvme.c
1038
ep->cqe.status = 0;
drivers/scsi/lpfc/lpfc_nvme.c
1087
ep->cqe.status);
drivers/scsi/lpfc/lpfc_nvme.c
1100
ep->cqe.status);
drivers/scsi/lpfc/lpfc_sli.c
14207
lpfc_sli4_els_xri_aborted(phba, &cq_event->cqe.wcqe_axri);
drivers/scsi/lpfc/lpfc_sli.c
14241
wcqe = &irspiocbq->cq_event.cqe.wcqe_cmpl;
drivers/scsi/lpfc/lpfc_sli.c
14288
memcpy(&cq_event->cqe, entry, size);
drivers/scsi/lpfc/lpfc_sli.c
14491
struct lpfc_cqe *cqe)
drivers/scsi/lpfc/lpfc_sli.c
14499
lpfc_sli4_pcimem_bcopy(cqe, &mcqe, sizeof(struct lpfc_mcqe));
drivers/scsi/lpfc/lpfc_sli.c
14556
memcpy(&irspiocbq->cq_event.cqe.wcqe_cmpl, wcqe, sizeof(*wcqe));
drivers/scsi/lpfc/lpfc_sli.c
14703
memcpy(&dma_buf->cq_event.cqe.rcqe_cmpl, rcqe, sizeof(*rcqe));
drivers/scsi/lpfc/lpfc_sli.c
14797
struct lpfc_cqe *cqe)
drivers/scsi/lpfc/lpfc_sli.c
14803
lpfc_sli4_pcimem_bcopy(cqe, &cqevt, sizeof(struct lpfc_cqe));
drivers/scsi/lpfc/lpfc_sli.c
14919
struct lpfc_cqe *cqe;
drivers/scsi/lpfc/lpfc_sli.c
14932
cqe = lpfc_sli4_cq_get(cq);
drivers/scsi/lpfc/lpfc_sli.c
14933
while (cqe) {
drivers/scsi/lpfc/lpfc_sli.c
14934
workposted |= handler(phba, cq, cqe);
drivers/scsi/lpfc/lpfc_sli.c
14935
__lpfc_sli4_consume_cqe(phba, cq, cqe);
drivers/scsi/lpfc/lpfc_sli.c
14951
cqe = lpfc_sli4_cq_get(cq);
drivers/scsi/lpfc/lpfc_sli.c
15344
struct lpfc_cqe *cqe)
drivers/scsi/lpfc/lpfc_sli.c
15350
lpfc_sli4_pcimem_bcopy(cqe, &wcqe, sizeof(struct lpfc_cqe));
drivers/scsi/lpfc/lpfc_sli.c
19328
&seq_dmabuf->cq_event.cqe.rcqe_cmpl);
drivers/scsi/lpfc/lpfc_sli.c
19361
&hbq_buf->cq_event.cqe.rcqe_cmpl);
drivers/scsi/lpfc/lpfc_sli.c
19381
&hbq_buf->cq_event.cqe.rcqe_cmpl);
drivers/scsi/lpfc/lpfc_sli.c
19473
frame_len = bf_get(lpfc_rcqe_length, &dmabuf->cq_event.cqe.rcqe_cmpl);
drivers/scsi/lpfc/lpfc_sli.c
19602
&dmabuf->cq_event.cqe.rcqe_cmpl) == CQE_CODE_RECEIVE_V1))
drivers/scsi/lpfc/lpfc_sli.c
19604
&dmabuf->cq_event.cqe.rcqe_cmpl);
drivers/scsi/lpfc/lpfc_sli.c
19607
&dmabuf->cq_event.cqe.rcqe_cmpl);
drivers/scsi/lpfc/lpfc_sli.c
19614
&dmabuf->cq_event.cqe.rcqe_cmpl));
drivers/scsi/lpfc/lpfc_sli.c
3223
size = bf_get(lpfc_rcqe_length, &nvmebuf->cq_event.cqe.rcqe_cmpl);
drivers/scsi/lpfc/lpfc_sli.c
4532
switch (bf_get(lpfc_wcqe_c_code, &cq_event->cqe.wcqe_cmpl)) {
drivers/scsi/lpfc/lpfc_sli.c
594
struct lpfc_cqe *cqe = NULL;
drivers/scsi/lpfc/lpfc_sli.c
613
cqe = lpfc_sli4_cq_get(cq);
drivers/scsi/lpfc/lpfc_sli.c
614
while (cqe) {
drivers/scsi/lpfc/lpfc_sli.c
615
__lpfc_sli4_consume_cqe(phba, cq, cqe);
drivers/scsi/lpfc/lpfc_sli.c
617
cqe = lpfc_sli4_cq_get(cq);
drivers/scsi/lpfc/lpfc_sli.c
687
struct lpfc_cqe *cqe;
drivers/scsi/lpfc/lpfc_sli.c
692
cqe = lpfc_sli4_qe(q, q->host_index);
drivers/scsi/lpfc/lpfc_sli.c
695
if (bf_get_le32(lpfc_cqe_valid, cqe) != q->qe_valid)
drivers/scsi/lpfc/lpfc_sli.c
707
return cqe;
drivers/scsi/lpfc/lpfc_sli.c
712
struct lpfc_cqe *cqe)
drivers/scsi/lpfc/lpfc_sli.c
715
bf_set_le32(lpfc_cqe_valid, cqe, 0);
drivers/scsi/lpfc/lpfc_sli.c
82
struct lpfc_queue *cq, struct lpfc_cqe *cqe);
drivers/scsi/lpfc/lpfc_sli.c
94
struct lpfc_cqe *cqe);
drivers/scsi/lpfc/lpfc_sli.h
64
} cqe;
drivers/scsi/qedf/qedf.h
259
struct fcoe_cqe cqe;
drivers/scsi/qedf/qedf.h
496
extern void qedf_scsi_completion(struct qedf_ctx *qedf, struct fcoe_cqe *cqe,
drivers/scsi/qedf/qedf.h
499
struct fcoe_cqe *cqe, struct qedf_ioreq *io_req);
drivers/scsi/qedf/qedf.h
501
struct fcoe_cqe *cqe, struct qedf_ioreq *io_req);
drivers/scsi/qedf/qedf.h
506
extern void qedf_process_abts_compl(struct qedf_ctx *qedf, struct fcoe_cqe *cqe,
drivers/scsi/qedf/qedf.h
519
extern void qedf_process_els_compl(struct qedf_ctx *qedf, struct fcoe_cqe *cqe,
drivers/scsi/qedf/qedf.h
526
struct fcoe_cqe *cqe, struct qedf_ioreq *io_req);
drivers/scsi/qedf/qedf.h
528
extern void qedf_process_tmf_compl(struct qedf_ctx *qedf, struct fcoe_cqe *cqe,
drivers/scsi/qedf/qedf.h
530
extern void qedf_process_cqe(struct qedf_ctx *qedf, struct fcoe_cqe *cqe);
drivers/scsi/qedf/qedf.h
539
struct fcoe_cqe *cqe);
drivers/scsi/qedf/qedf.h
545
struct fcoe_cqe *cqe, struct qedf_ioreq *io_req);
drivers/scsi/qedf/qedf_els.c
142
void qedf_process_els_compl(struct qedf_ctx *qedf, struct fcoe_cqe *cqe,
drivers/scsi/qedf/qedf_els.c
179
mp_info = &cqe->cqe_info.midpath_info;
drivers/scsi/qedf/qedf_els.c
744
struct fcoe_cqe *cqe, struct qedf_ioreq *io_req)
drivers/scsi/qedf/qedf_els.c
752
if (io_req->event == QEDF_IOREQ_EV_ELS_TMO || !cqe) {
drivers/scsi/qedf/qedf_io.c
1117
void qedf_scsi_completion(struct qedf_ctx *qedf, struct fcoe_cqe *cqe,
drivers/scsi/qedf/qedf_io.c
1131
if (!cqe)
drivers/scsi/qedf/qedf_io.c
1144
fcp_rsp = &cqe->cqe_info.rsp_info;
drivers/scsi/qedf/qedf_io.c
1198
fw_residual_flag = GET_FIELD(cqe->cqe_info.rsp_info.fw_error_flags,
drivers/scsi/qedf/qedf_io.c
1205
cqe->cqe_info.rsp_info.fw_residual, sc_cmd->cmnd[2],
drivers/scsi/qedf/qedf_io.c
1428
void qedf_process_warning_compl(struct qedf_ctx *qedf, struct fcoe_cqe *cqe,
drivers/scsi/qedf/qedf_io.c
1436
if (!cqe) {
drivers/scsi/qedf/qedf_io.c
1447
le32_to_cpu(cqe->cqe_info.err_info.err_warn_bitmap_hi),
drivers/scsi/qedf/qedf_io.c
1448
le32_to_cpu(cqe->cqe_info.err_info.err_warn_bitmap_lo));
drivers/scsi/qedf/qedf_io.c
1451
le32_to_cpu(cqe->cqe_info.err_info.tx_buf_off),
drivers/scsi/qedf/qedf_io.c
1452
le32_to_cpu(cqe->cqe_info.err_info.rx_buf_off),
drivers/scsi/qedf/qedf_io.c
1453
le32_to_cpu(cqe->cqe_info.err_info.rx_id));
drivers/scsi/qedf/qedf_io.c
1457
((u64)cqe->cqe_info.err_info.err_warn_bitmap_hi << 32) |
drivers/scsi/qedf/qedf_io.c
1458
(u64)cqe->cqe_info.err_info.err_warn_bitmap_lo;
drivers/scsi/qedf/qedf_io.c
1473
cqe->cqe_info.err_info.rx_buf_off;
drivers/scsi/qedf/qedf_io.c
1475
cqe->cqe_info.err_info.tx_buf_off;
drivers/scsi/qedf/qedf_io.c
1476
io_req->rx_id = cqe->cqe_info.err_info.rx_id;
drivers/scsi/qedf/qedf_io.c
1498
void qedf_process_error_detect(struct qedf_ctx *qedf, struct fcoe_cqe *cqe,
drivers/scsi/qedf/qedf_io.c
1513
if (!cqe) {
drivers/scsi/qedf/qedf_io.c
1523
le32_to_cpu(cqe->cqe_info.err_info.err_warn_bitmap_hi),
drivers/scsi/qedf/qedf_io.c
1524
le32_to_cpu(cqe->cqe_info.err_info.err_warn_bitmap_lo));
drivers/scsi/qedf/qedf_io.c
1527
le32_to_cpu(cqe->cqe_info.err_info.tx_buf_off),
drivers/scsi/qedf/qedf_io.c
1528
le32_to_cpu(cqe->cqe_info.err_info.rx_buf_off),
drivers/scsi/qedf/qedf_io.c
1529
le32_to_cpu(cqe->cqe_info.err_info.rx_id));
drivers/scsi/qedf/qedf_io.c
1953
void qedf_process_abts_compl(struct qedf_ctx *qedf, struct fcoe_cqe *cqe,
drivers/scsi/qedf/qedf_io.c
1963
r_ctl = cqe->cqe_info.abts_info.r_ctl;
drivers/scsi/qedf/qedf_io.c
2274
void qedf_process_cleanup_compl(struct qedf_ctx *qedf, struct fcoe_cqe *cqe,
drivers/scsi/qedf/qedf_io.c
2472
void qedf_process_tmf_compl(struct qedf_ctx *qedf, struct fcoe_cqe *cqe,
drivers/scsi/qedf/qedf_io.c
2479
fcp_rsp = &cqe->cqe_info.rsp_info;
drivers/scsi/qedf/qedf_io.c
2486
struct fcoe_cqe *cqe)
drivers/scsi/qedf/qedf_io.c
2489
uint16_t pktlen = cqe->cqe_info.unsolic_info.pkt_len;
drivers/scsi/qedf/qedf_io.c
2498
p_bd_info = &cqe->cqe_info.unsolic_info.bd_info;
drivers/scsi/qedf/qedf_io.c
2569
memcpy(&io_work->cqe, cqe, sizeof(struct fcoe_cqe));
drivers/scsi/qedf/qedf_main.c
2238
struct fcoe_cqe *cqe;
drivers/scsi/qedf/qedf_main.c
2262
cqe = &que->cq[que->cq_cons_idx];
drivers/scsi/qedf/qedf_main.c
2264
comp_type = (cqe->cqe_data >> FCOE_CQE_CQE_TYPE_SHIFT) &
drivers/scsi/qedf/qedf_main.c
2274
qedf_process_unsol_compl(qedf, fp->sb_id, cqe);
drivers/scsi/qedf/qedf_main.c
2282
xid = cqe->cqe_data & FCOE_CQE_TASK_ID_MASK;
drivers/scsi/qedf/qedf_main.c
2310
memcpy(&io_work->cqe, cqe, sizeof(struct fcoe_cqe));
drivers/scsi/qedf/qedf_main.c
2709
comp_type = (io_work->cqe.cqe_data >>
drivers/scsi/qedf/qedf_main.c
2716
qedf_process_cqe(io_work->qedf, &io_work->cqe);
drivers/scsi/qedf/qedf_main.c
2814
void qedf_process_cqe(struct qedf_ctx *qedf, struct fcoe_cqe *cqe)
drivers/scsi/qedf/qedf_main.c
2823
comp_type = (cqe->cqe_data >> FCOE_CQE_CQE_TYPE_SHIFT) &
drivers/scsi/qedf/qedf_main.c
2826
xid = cqe->cqe_data & FCOE_CQE_TASK_ID_MASK;
drivers/scsi/qedf/qedf_main.c
2864
qedf_scsi_completion(qedf, cqe, io_req);
drivers/scsi/qedf/qedf_main.c
2867
qedf_process_els_compl(qedf, cqe, io_req);
drivers/scsi/qedf/qedf_main.c
2870
qedf_process_tmf_compl(qedf, cqe, io_req);
drivers/scsi/qedf/qedf_main.c
2873
qedf_process_seq_cleanup_compl(qedf, cqe, io_req);
drivers/scsi/qedf/qedf_main.c
2881
qedf_process_error_detect(qedf, cqe, io_req);
drivers/scsi/qedf/qedf_main.c
2887
qedf_process_cleanup_compl(qedf, cqe, io_req);
drivers/scsi/qedf/qedf_main.c
2893
qedf_process_abts_compl(qedf, cqe, io_req);
drivers/scsi/qedf/qedf_main.c
2909
qedf_process_warning_compl(qedf, cqe, io_req);
drivers/scsi/qedi/qedi.h
178
struct iscsi_cqe_solicited cqe;
drivers/scsi/qedi/qedi.h
373
union iscsi_cqe cqe;
drivers/scsi/qedi/qedi_fw.c
109
resp_hdr_ptr->itt = build_itt(cqe->cqe_solicited.itid,
drivers/scsi/qedi/qedi_fw.c
178
union iscsi_cqe *cqe,
drivers/scsi/qedi/qedi_fw.c
190
cqe_tmp_response = &cqe->cqe_common.iscsi_hdr.tmf_response;
drivers/scsi/qedi/qedi_fw.c
214
resp_hdr_ptr->itt = build_itt(cqe->cqe_solicited.itid,
drivers/scsi/qedi/qedi_fw.c
258
union iscsi_cqe *cqe,
drivers/scsi/qedi/qedi_fw.c
272
cqe_login_response = &cqe->cqe_common.iscsi_hdr.login_response;
drivers/scsi/qedi/qedi_fw.c
285
resp_hdr_ptr->itt = build_itt(cqe->cqe_solicited.itid,
drivers/scsi/qedi/qedi_fw.c
31
union iscsi_cqe *cqe,
drivers/scsi/qedi/qedi_fw.c
320
struct iscsi_cqe_unsolicited *cqe,
drivers/scsi/qedi/qedi_fw.c
331
idx = cqe->rqe_opaque;
drivers/scsi/qedi/qedi_fw.c
340
"rqe_opaque [0x%p], idx [%d]\n", cqe->rqe_opaque, idx);
drivers/scsi/qedi/qedi_fw.c
343
"unsol_cqe_type = %d\n", cqe->unsol_cqe_type);
drivers/scsi/qedi/qedi_fw.c
344
switch (cqe->unsol_cqe_type) {
drivers/scsi/qedi/qedi_fw.c
359
struct iscsi_cqe_unsolicited *cqe,
drivers/scsi/qedi/qedi_fw.c
366
idx = cqe->rqe_opaque;
drivers/scsi/qedi/qedi_fw.c
397
struct iscsi_cqe_unsolicited *cqe,
drivers/scsi/qedi/qedi_fw.c
404
qedi_get_rq_bdq_buf(qedi, cqe, bdq_data, pdu_len);
drivers/scsi/qedi/qedi_fw.c
405
qedi_put_rq_bdq_buf(qedi, cqe, (num_bdqs + 1));
drivers/scsi/qedi/qedi_fw.c
409
union iscsi_cqe *cqe,
drivers/scsi/qedi/qedi_fw.c
42
cqe_logout_response = &cqe->cqe_common.iscsi_hdr.logout_response;
drivers/scsi/qedi/qedi_fw.c
425
cqe_nop_in = &cqe->cqe_common.iscsi_hdr.nop_in;
drivers/scsi/qedi/qedi_fw.c
439
if (cqe->cqe_common.cqe_type == ISCSI_CQE_TYPE_UNSOLICITED) {
drivers/scsi/qedi/qedi_fw.c
441
qedi_unsol_pdu_adjust_bdq(qedi, &cqe->cqe_unsolicited,
drivers/scsi/qedi/qedi_fw.c
453
hdr->itt = build_itt(cqe->cqe_solicited.itid,
drivers/scsi/qedi/qedi_fw.c
480
union iscsi_cqe *cqe,
drivers/scsi/qedi/qedi_fw.c
496
cqe_async_msg = &cqe->cqe_common.iscsi_hdr.async_msg;
drivers/scsi/qedi/qedi_fw.c
50
resp_hdr->itt = build_itt(cqe->cqe_solicited.itid, conn->session->age);
drivers/scsi/qedi/qedi_fw.c
501
if (cqe->cqe_common.cqe_type == ISCSI_CQE_TYPE_UNSOLICITED) {
drivers/scsi/qedi/qedi_fw.c
503
qedi_unsol_pdu_adjust_bdq(qedi, &cqe->cqe_unsolicited,
drivers/scsi/qedi/qedi_fw.c
534
union iscsi_cqe *cqe,
drivers/scsi/qedi/qedi_fw.c
547
cqe_reject = &cqe->cqe_common.iscsi_hdr.reject;
drivers/scsi/qedi/qedi_fw.c
552
if (cqe->cqe_common.cqe_type == ISCSI_CQE_TYPE_UNSOLICITED) {
drivers/scsi/qedi/qedi_fw.c
554
qedi_unsol_pdu_adjust_bdq(qedi, &cqe->cqe_unsolicited,
drivers/scsi/qedi/qedi_fw.c
576
union iscsi_cqe *cqe,
drivers/scsi/qedi/qedi_fw.c
590
iscsi_cid = cqe->cqe_common.conn_id;
drivers/scsi/qedi/qedi_fw.c
593
cqe_data_in = &cqe->cqe_common.iscsi_hdr.data_in;
drivers/scsi/qedi/qedi_fw.c
595
cqe->cqe_common.error_bitmap.error_bits.cqe_error_status_bits;
drivers/scsi/qedi/qedi_fw.c
625
hdr->itt = build_itt(cqe->cqe_solicited.itid, conn->session->age);
drivers/scsi/qedi/qedi_fw.c
674
union iscsi_cqe *cqe,
drivers/scsi/qedi/qedi_fw.c
681
hdr_opcode = cqe->cqe_common.iscsi_hdr.common.hdr_first_byte;
drivers/scsi/qedi/qedi_fw.c
687
qedi_scsi_completion(qedi, cqe, task, iscsi_conn);
drivers/scsi/qedi/qedi_fw.c
690
qedi_process_login_resp(qedi, cqe, task, conn);
drivers/scsi/qedi/qedi_fw.c
693
qedi_process_tmf_resp(qedi, cqe, task, conn);
drivers/scsi/qedi/qedi_fw.c
696
qedi_process_text_resp(qedi, cqe, task, conn);
drivers/scsi/qedi/qedi_fw.c
699
qedi_process_logout_resp(qedi, cqe, task, conn);
drivers/scsi/qedi/qedi_fw.c
702
qedi_process_nopin_mesg(qedi, cqe, task, conn, que_idx);
drivers/scsi/qedi/qedi_fw.c
710
struct iscsi_cqe_solicited *cqe,
drivers/scsi/qedi/qedi_fw.c
720
cqe->itid, cmd->task_id);
drivers/scsi/qedi/qedi_fw.c
730
struct iscsi_cqe_solicited *cqe,
drivers/scsi/qedi/qedi_fw.c
734
u32 proto_itt = cqe->itid;
drivers/scsi/qedi/qedi_fw.c
743
iscsi_cid = cqe->conn_id;
drivers/scsi/qedi/qedi_fw.c
747
"icid not found 0x%x\n", cqe->conn_id);
drivers/scsi/qedi/qedi_fw.c
816
cqe->itid, qedi_conn->iscsi_conn_id);
drivers/scsi/qedi/qedi_fw.c
82
union iscsi_cqe *cqe,
drivers/scsi/qedi/qedi_fw.c
824
union iscsi_cqe *cqe = &work->cqe;
drivers/scsi/qedi/qedi_fw.c
836
comp_type = cqe->cqe_common.cqe_type;
drivers/scsi/qedi/qedi_fw.c
837
hdr_opcode = cqe->cqe_common.iscsi_hdr.common.hdr_first_byte;
drivers/scsi/qedi/qedi_fw.c
839
cqe->cqe_common.error_bitmap.error_bits.cqe_error_status_bits;
drivers/scsi/qedi/qedi_fw.c
843
cqe->cqe_common.conn_id, comp_type, hdr_opcode);
drivers/scsi/qedi/qedi_fw.c
850
iscsi_cid = cqe->cqe_common.conn_id;
drivers/scsi/qedi/qedi_fw.c
881
(cqe->cqe_solicited.itid != (u16)RESERVED_ITT)) {
drivers/scsi/qedi/qedi_fw.c
882
qedi_process_nopin_local_cmpl(qedi, &cqe->cqe_solicited,
drivers/scsi/qedi/qedi_fw.c
885
cqe->cqe_solicited.itid =
drivers/scsi/qedi/qedi_fw.c
886
qedi_get_itt(cqe->cqe_solicited);
drivers/scsi/qedi/qedi_fw.c
888
qedi_mtask_completion(qedi, cqe, task, q_conn, que_idx);
drivers/scsi/qedi/qedi_fw.c
894
qedi_process_nopin_mesg(qedi, cqe, task, q_conn,
drivers/scsi/qedi/qedi_fw.c
898
qedi_process_async_mesg(qedi, cqe, task, q_conn,
drivers/scsi/qedi/qedi_fw.c
902
qedi_process_reject_mesg(qedi, cqe, task, q_conn,
drivers/scsi/qedi/qedi_fw.c
912
qedi_process_cmd_cleanup_resp(qedi, &cqe->cqe_solicited, conn);
drivers/scsi/qedi/qedi_fw.c
97
cqe_text_response = &cqe->cqe_common.iscsi_hdr.text_response;
drivers/scsi/qedi/qedi_iscsi.h
234
#define qedi_get_itt(cqe) (cqe.iscsi_hdr.cmd.itt >> 16)
drivers/scsi/qedi/qedi_main.c
1205
static int qedi_queue_cqe(struct qedi_ctx *qedi, union iscsi_cqe *cqe,
drivers/scsi/qedi/qedi_main.c
1214
iscsi_cid = cqe->cqe_common.conn_id;
drivers/scsi/qedi/qedi_main.c
1223
switch (cqe->cqe_common.cqe_type) {
drivers/scsi/qedi/qedi_main.c
1226
qedi_cmd = qedi_get_cmd_from_tid(qedi, cqe->cqe_solicited.itid);
drivers/scsi/qedi/qedi_main.c
1233
memcpy(&qedi_cmd->cqe_work.cqe, cqe, sizeof(union iscsi_cqe));
drivers/scsi/qedi/qedi_main.c
1248
memcpy(&qedi_work->cqe, cqe, sizeof(union iscsi_cqe));
drivers/scsi/qedi/qedi_main.c
1269
union iscsi_cqe *cqe;
drivers/scsi/qedi/qedi_main.c
1293
cqe = &que->cq[que->cq_cons_idx];
drivers/scsi/qedi/qedi_main.c
1297
cqe, prod_idx, que->cq_cons_idx);
drivers/scsi/qedi/qedi_main.c
1299
ret = qedi_queue_cqe(qedi, cqe, fp->sb_id, p);
drivers/scsi/qedi/qedi_main.c
1303
que->cq_cons_idx, cqe->cqe_common.conn_id);
drivers/ufs/core/ufs-mcq.c
279
static int ufshcd_mcq_get_tag(struct ufs_hba *hba, struct cq_entry *cqe)
drivers/ufs/core/ufs-mcq.c
287
addr = (le64_to_cpu(cqe->command_desc_base_addr) & CQE_UCD_BA) -
drivers/ufs/core/ufs-mcq.c
296
struct cq_entry *cqe = ufshcd_mcq_cur_cqe(hwq);
drivers/ufs/core/ufs-mcq.c
298
if (cqe->command_desc_base_addr) {
drivers/ufs/core/ufs-mcq.c
299
int tag = ufshcd_mcq_get_tag(hba, cqe);
drivers/ufs/core/ufs-mcq.c
301
ufshcd_compl_one_cqe(hba, tag, cqe);
drivers/ufs/core/ufs-mcq.c
303
cqe->command_desc_base_addr = 0;
drivers/ufs/core/ufshcd-priv.h
424
struct cq_entry *cqe = q->cqe_base_addr;
drivers/ufs/core/ufshcd-priv.h
426
return cqe + q->cq_head_slot;
drivers/ufs/core/ufshcd-priv.h
67
struct cq_entry *cqe);
drivers/ufs/core/ufshcd.c
5444
struct cq_entry *cqe)
drivers/ufs/core/ufshcd.c
5464
ocs = ufshcd_get_tr_ocs(lrbp, cqe);
drivers/ufs/core/ufshcd.c
5542
if (cqe)
drivers/ufs/core/ufshcd.c
5543
ufshcd_hex_dump("UPIU CQE: ", cqe, sizeof(struct cq_entry));
drivers/ufs/core/ufshcd.c
5626
struct cq_entry *cqe)
drivers/ufs/core/ufshcd.c
5633
le64_to_cpu(cqe->command_desc_base_addr)))
drivers/ufs/core/ufshcd.c
5644
cmd->result = ufshcd_transfer_rsp_status(hba, cmd, cqe);
drivers/ufs/core/ufshcd.c
5647
if (cqe) {
drivers/ufs/core/ufshcd.c
5648
ocs = cqe->overall_status & MASK_OCS;
drivers/ufs/core/ufshcd.c
859
struct cq_entry *cqe)
drivers/ufs/core/ufshcd.c
861
if (cqe)
drivers/ufs/core/ufshcd.c
862
return cqe->overall_status & MASK_OCS;
drivers/vdpa/mlx5/net/mlx5_vnet.c
346
void *cqe;
drivers/vdpa/mlx5/net/mlx5_vnet.c
350
cqe = get_cqe(vcq, i);
drivers/vdpa/mlx5/net/mlx5_vnet.c
351
cqe64 = cqe;
drivers/vdpa/mlx5/net/mlx5_vnet.c
358
struct mlx5_cqe64 *cqe64 = get_cqe(cq, n & (cq->cqe - 1));
drivers/vdpa/mlx5/net/mlx5_vnet.c
361
!((cqe64->op_own & MLX5_CQE_OWNER_MASK) ^ !!(n & cq->cqe)))
drivers/vdpa/mlx5/net/mlx5_vnet.c
577
vcq->cqe = num_ent;
drivers/vdpa/mlx5/net/mlx5_vnet.c
74
int cqe;
drivers/vfio/pci/mlx5/cmd.c
1076
void *cqe;
drivers/vfio/pci/mlx5/cmd.c
1080
cqe = mlx5_frag_buf_get_wqe(&buf->fbc, i);
drivers/vfio/pci/mlx5/cmd.c
1081
cqe64 = buf->cqe_size == 64 ? cqe : cqe + 64;
drivers/vfio/pci/mlx5/cmd.c
1656
mlx5vf_rq_cqe(struct mlx5_vhca_qp *qp, struct mlx5_cqe64 *cqe,
drivers/vfio/pci/mlx5/cmd.c
1663
*tracker_status = be32_to_cpu(cqe->immediate) >> 28;
drivers/vfio/pci/mlx5/cmd.c
1664
size = be32_to_cpu(cqe->byte_cnt);
drivers/vfio/pci/mlx5/cmd.c
1665
ix = be16_to_cpu(cqe->wqe_counter) & (qp->rq.wqe_cnt - 1);
drivers/vfio/pci/mlx5/cmd.c
1683
void *cqe = get_cqe(cq, n & (cq->ncqe - 1));
drivers/vfio/pci/mlx5/cmd.c
1686
cqe64 = (cq->mcq.cqe_sz == 64) ? cqe : cqe + 64;
drivers/vfio/pci/mlx5/cmd.c
1700
struct mlx5_cqe64 *cqe;
drivers/vfio/pci/mlx5/cmd.c
1703
cqe = get_sw_cqe(cq, cq->mcq.cons_index);
drivers/vfio/pci/mlx5/cmd.c
1704
if (!cqe)
drivers/vfio/pci/mlx5/cmd.c
1713
opcode = get_cqe_opcode(cqe);
drivers/vfio/pci/mlx5/cmd.c
1716
mlx5vf_rq_cqe(qp, cqe, dirty, tracker_status);
fs/smb/client/smbdirect.c
1242
request->cqe.done = send_done;
fs/smb/client/smbdirect.c
1264
request->wr.wr_cqe = &request->cqe;
fs/smb/client/smbdirect.c
1305
last->wr.wr_cqe = &last->cqe;
fs/smb/client/smbdirect.c
1638
response->cqe.done = recv_done;
fs/smb/client/smbdirect.c
1640
recv_wr.wr_cqe = &response->cqe;
fs/smb/client/smbdirect.c
2620
container_of(wc->wr_cqe, struct smbdirect_mr_io, cqe);
fs/smb/client/smbdirect.c
2961
mr->cqe.done = register_mr_done;
fs/smb/client/smbdirect.c
2962
reg_wr->wr.wr_cqe = &mr->cqe;
fs/smb/client/smbdirect.c
3028
struct ib_cqe *cqe;
fs/smb/client/smbdirect.c
3030
cqe = wc->wr_cqe;
fs/smb/client/smbdirect.c
3031
smbdirect_mr = container_of(cqe, struct smbdirect_mr_io, cqe);
fs/smb/client/smbdirect.c
3065
mr->cqe.done = local_inv_done;
fs/smb/client/smbdirect.c
3066
wr->wr_cqe = &mr->cqe;
fs/smb/client/smbdirect.c
564
container_of(wc->wr_cqe, struct smbdirect_send_io, cqe);
fs/smb/client/smbdirect.c
768
container_of(wc->wr_cqe, struct smbdirect_recv_io, cqe);
fs/smb/common/smbdirect/smbdirect_socket.h
465
struct ib_cqe cqe;
fs/smb/common/smbdirect/smbdirect_socket.h
508
struct ib_cqe cqe;
fs/smb/common/smbdirect/smbdirect_socket.h
538
struct ib_cqe cqe;
fs/smb/common/smbdirect/smbdirect_socket.h
565
struct ib_cqe cqe;
fs/smb/server/transport_rdma.c
1068
sendmsg = container_of(wc->wr_cqe, struct smbdirect_send_io, cqe);
fs/smb/server/transport_rdma.c
1232
last->wr.wr_cqe = &last->cqe;
fs/smb/server/transport_rdma.c
1474
msg->cqe.done = send_done;
fs/smb/server/transport_rdma.c
1496
msg->wr.wr_cqe = &msg->cqe;
fs/smb/server/transport_rdma.c
1775
container_of(wc->wr_cqe, struct smbdirect_rw_io, cqe);
fs/smb/server/transport_rdma.c
1865
msg->cqe.done = is_read ? read_done : write_done;
fs/smb/server/transport_rdma.c
1902
&msg->cqe, first_wr);
fs/smb/server/transport_rdma.c
2173
recvmsg->cqe.done = smb_direct_negotiate_recv_done;
fs/smb/server/transport_rdma.c
2631
recv_io->cqe.done = recv_done;
fs/smb/server/transport_rdma.c
623
recvmsg = container_of(wc->wr_cqe, struct smbdirect_recv_io, cqe);
fs/smb/server/transport_rdma.c
749
container_of(wc->wr_cqe, struct smbdirect_recv_io, cqe);
fs/smb/server/transport_rdma.c
756
recv_io->cqe.done = recv_done;
fs/smb/server/transport_rdma.c
881
wr.wr_cqe = &recvmsg->cqe;
include/linux/hisi_acc_qm.h
463
struct qm_cqe *cqe;
include/linux/io_uring_types.h
714
struct io_cqe cqe;
include/linux/io_uring_types.h
774
struct io_uring_cqe cqe;
include/linux/mlx5/device.h
1000
static inline bool mpwrq_is_filler_cqe(struct mlx5_cqe64 *cqe)
include/linux/mlx5/device.h
1002
struct mpwrq_cqe_bc *bc = (struct mpwrq_cqe_bc *)&cqe->byte_cnt;
include/linux/mlx5/device.h
1007
static inline u16 mpwrq_get_cqe_stride_index(struct mlx5_cqe64 *cqe)
include/linux/mlx5/device.h
1009
return be16_to_cpu(cqe->wqe_counter);
include/linux/mlx5/device.h
909
static inline u8 mlx5_get_cqe_format(struct mlx5_cqe64 *cqe)
include/linux/mlx5/device.h
911
return (cqe->op_own >> 2) & 0x3;
include/linux/mlx5/device.h
914
static inline u8 get_cqe_opcode(struct mlx5_cqe64 *cqe)
include/linux/mlx5/device.h
916
return cqe->op_own >> 4;
include/linux/mlx5/device.h
919
static inline u8 get_cqe_enhanced_num_mini_cqes(struct mlx5_cqe64 *cqe)
include/linux/mlx5/device.h
922
return get_cqe_opcode(cqe) + 1;
include/linux/mlx5/device.h
925
static inline u8 get_cqe_lro_tcppsh(struct mlx5_cqe64 *cqe)
include/linux/mlx5/device.h
927
return (cqe->lro.tcppsh_abort_dupack >> 6) & 1;
include/linux/mlx5/device.h
930
static inline u8 get_cqe_l4_hdr_type(struct mlx5_cqe64 *cqe)
include/linux/mlx5/device.h
932
return (cqe->l4_l3_hdr_type >> 4) & 0x7;
include/linux/mlx5/device.h
935
static inline bool cqe_is_tunneled(struct mlx5_cqe64 *cqe)
include/linux/mlx5/device.h
937
return cqe->tls_outer_l3_tunneled & 0x1;
include/linux/mlx5/device.h
940
static inline u8 get_cqe_tls_offload(struct mlx5_cqe64 *cqe)
include/linux/mlx5/device.h
942
return (cqe->tls_outer_l3_tunneled >> 3) & 0x3;
include/linux/mlx5/device.h
945
static inline bool cqe_has_vlan(const struct mlx5_cqe64 *cqe)
include/linux/mlx5/device.h
947
return cqe->l4_l3_hdr_type & 0x1;
include/linux/mlx5/device.h
950
static inline u64 get_cqe_ts(struct mlx5_cqe64 *cqe)
include/linux/mlx5/device.h
954
hi = be32_to_cpu(cqe->timestamp_h);
include/linux/mlx5/device.h
955
lo = be32_to_cpu(cqe->timestamp_l);
include/linux/mlx5/device.h
960
static inline u16 get_cqe_flow_tag(struct mlx5_cqe64 *cqe)
include/linux/mlx5/device.h
962
return be32_to_cpu(cqe->sop_drop_qpn) & 0xFFF;
include/linux/mlx5/device.h
965
static inline u8 get_cqe_lro_num_seg(struct mlx5_cqe64 *cqe)
include/linux/mlx5/device.h
967
return be32_to_cpu(cqe->srqn) >> 24;
include/linux/mlx5/device.h
981
static inline u16 mpwrq_get_cqe_byte_cnt(struct mlx5_cqe64 *cqe)
include/linux/mlx5/device.h
983
struct mpwrq_cqe_bc *bc = (struct mpwrq_cqe_bc *)&cqe->byte_cnt;
include/linux/mlx5/device.h
993
static inline u16 mpwrq_get_cqe_consumed_strides(struct mlx5_cqe64 *cqe)
include/linux/mlx5/device.h
995
struct mpwrq_cqe_bc *bc = (struct mpwrq_cqe_bc *)&cqe->byte_cnt;
include/linux/nvme-fc.h
79
struct nvme_completion cqe;
include/linux/nvme-tcp.h
151
struct nvme_completion cqe;
include/linux/qed/eth_common.h
327
union eth_rx_cqe cqe;
include/linux/qed/qed_eth_if.h
319
struct eth_slow_path_rx_cqe *cqe);
include/rdma/ib_verbs.h
1635
int cqe;
include/rdma/ib_verbs.h
2546
int (*resize_cq)(struct ib_cq *cq, int cqe, struct ib_udata *udata);
include/rdma/ib_verbs.h
374
unsigned int cqe;
include/rdma/ib_verbs.h
4068
int ib_resize_cq(struct ib_cq *cq, int cqe);
include/rdma/rw.h
83
u32 port_num, struct ib_cqe *cqe, struct ib_send_wr *chain_wr);
include/rdma/rw.h
85
struct ib_cqe *cqe, struct ib_send_wr *chain_wr);
include/trace/events/io_uring.h
124
__entry->user_data = req->cqe.user_data;
include/trace/events/io_uring.h
161
__entry->user_data = req->cqe.user_data;
include/trace/events/io_uring.h
202
__entry->data = req->cqe.user_data;
include/trace/events/io_uring.h
302
__entry->user_data = req->cqe.user_data;
include/trace/events/io_uring.h
323
TP_PROTO(struct io_ring_ctx *ctx, void *req, struct io_uring_cqe *cqe),
include/trace/events/io_uring.h
325
TP_ARGS(ctx, req, cqe),
include/trace/events/io_uring.h
340
__entry->user_data = cqe->user_data;
include/trace/events/io_uring.h
341
__entry->res = cqe->res;
include/trace/events/io_uring.h
342
__entry->cflags = cqe->flags;
include/trace/events/io_uring.h
343
__entry->extra1 = ctx->flags & IORING_SETUP_CQE32 || cqe->flags & IORING_CQE_F_32 ? cqe->big_cqe[0] : 0;
include/trace/events/io_uring.h
344
__entry->extra2 = ctx->flags & IORING_SETUP_CQE32 || cqe->flags & IORING_CQE_F_32 ? cqe->big_cqe[1] : 0;
include/trace/events/io_uring.h
384
__entry->user_data = req->cqe.user_data;
include/trace/events/io_uring.h
428
__entry->user_data = req->cqe.user_data;
include/trace/events/io_uring.h
468
__entry->user_data = req->cqe.user_data;
include/uapi/rdma/ib_user_verbs.h
435
__u32 cqe;
include/uapi/rdma/ib_user_verbs.h
449
__u32 cqe;
include/uapi/rdma/ib_user_verbs.h
459
__u32 cqe;
include/uapi/rdma/ib_user_verbs.h
472
__u32 cqe;
include/uapi/rdma/ib_user_verbs.h
477
__u32 cqe;
io_uring/bpf_filter.c
33
bctx->user_data = req->cqe.user_data;
io_uring/cancel.c
59
if (match_user_data && req->cqe.user_data != cd->data)
io_uring/cmd_net.c
61
struct io_uring_cqe cqe[2];
io_uring/cmd_net.c
76
cqe->user_data = 0;
io_uring/cmd_net.c
77
cqe->res = tskey;
io_uring/cmd_net.c
78
cqe->flags = IORING_CQE_F_MORE | ctx_cqe32_flags(cmd_to_io_kiocb(cmd)->ctx);
io_uring/cmd_net.c
79
cqe->flags |= tstype << IORING_TIMESTAMP_TYPE_SHIFT;
io_uring/cmd_net.c
81
cqe->flags |= IORING_CQE_F_TSTAMP_HW;
io_uring/cmd_net.c
83
iots = (struct io_timespec *)&cqe[1];
io_uring/cmd_net.c
86
return io_uring_cmd_post_mshot_cqe32(cmd, issue_flags, cqe);
io_uring/fdinfo.c
156
struct io_uring_cqe *cqe;
io_uring/fdinfo.c
159
cqe = &r->cqes[(cq_head & cq_mask)];
io_uring/fdinfo.c
160
if (cqe->flags & IORING_CQE_F_32 || ctx->flags & IORING_SETUP_CQE32)
io_uring/fdinfo.c
163
cq_head & cq_mask, cqe->user_data, cqe->res,
io_uring/fdinfo.c
164
cqe->flags);
io_uring/fdinfo.c
167
cqe->big_cqe[0], cqe->big_cqe[1]);
io_uring/fdinfo.c
244
struct io_uring_cqe *cqe = &ocqe->cqe;
io_uring/fdinfo.c
247
cqe->user_data, cqe->res, cqe->flags);
io_uring/io_uring.c
1030
io_req_defer_failed(req, req->cqe.res);
io_uring/io_uring.c
1139
io_cqe_overflow(ctx, &req->cqe, &req->big_cqe);
io_uring/io_uring.c
1141
io_cqe_overflow_locked(ctx, &req->cqe, &req->big_cqe);
io_uring/io_uring.c
1357
req->file = io_file_get_fixed(req, req->cqe.fd, issue_flags);
io_uring/io_uring.c
1359
req->file = io_file_get_normal(req, req->cqe.fd);
io_uring/io_uring.c
1654
io_req_defer_failed(req, req->cqe.res);
io_uring/io_uring.c
1729
req->cqe.user_data = READ_ONCE(sqe->user_data);
io_uring/io_uring.c
1799
req->cqe.fd = READ_ONCE(sqe->fd);
io_uring/io_uring.c
533
struct io_uring_cqe *cqe;
io_uring/io_uring.c
539
if (ocqe->cqe.flags & IORING_CQE_F_32 ||
io_uring/io_uring.c
548
if (!io_get_cqe_overflow(ctx, &cqe, true, is_cqe32))
io_uring/io_uring.c
550
memcpy(cqe, &ocqe->cqe, cqe_size);
io_uring/io_uring.c
655
struct io_cqe *cqe,
io_uring/io_uring.c
662
if (cqe->flags & IORING_CQE_F_32 || ctx->flags & IORING_SETUP_CQE32) {
io_uring/io_uring.c
668
trace_io_uring_cqe_overflow(ctx, cqe->user_data, cqe->res, cqe->flags, ocqe);
io_uring/io_uring.c
670
ocqe->cqe.user_data = cqe->user_data;
io_uring/io_uring.c
671
ocqe->cqe.res = cqe->res;
io_uring/io_uring.c
672
ocqe->cqe.flags = cqe->flags;
io_uring/io_uring.c
674
ocqe->cqe.big_cqe[0] = big_cqe->extra1;
io_uring/io_uring.c
675
ocqe->cqe.big_cqe[1] = big_cqe->extra2;
io_uring/io_uring.c
690
struct io_uring_cqe *cqe = &ctx->rings->cqes[off];
io_uring/io_uring.c
692
cqe->user_data = 0;
io_uring/io_uring.c
693
cqe->res = 0;
io_uring/io_uring.c
694
cqe->flags = IORING_CQE_F_SKIP;
io_uring/io_uring.c
751
struct io_uring_cqe *cqe;
io_uring/io_uring.c
755
if (unlikely(!io_get_cqe(ctx, &cqe, true)))
io_uring/io_uring.c
758
memcpy(cqe, src_cqe, 2 * sizeof(*cqe));
io_uring/io_uring.c
759
trace_io_uring_complete(ctx, NULL, cqe);
io_uring/io_uring.c
767
struct io_uring_cqe *cqe;
io_uring/io_uring.c
769
if (likely(io_get_cqe(ctx, &cqe, cqe32))) {
io_uring/io_uring.c
770
WRITE_ONCE(cqe->user_data, user_data);
io_uring/io_uring.c
771
WRITE_ONCE(cqe->res, res);
io_uring/io_uring.c
772
WRITE_ONCE(cqe->flags, cflags);
io_uring/io_uring.c
775
WRITE_ONCE(cqe->big_cqe[0], 0);
io_uring/io_uring.c
776
WRITE_ONCE(cqe->big_cqe[1], 0);
io_uring/io_uring.c
779
trace_io_uring_complete(ctx, NULL, cqe);
io_uring/io_uring.c
790
static __cold void io_cqe_overflow(struct io_ring_ctx *ctx, struct io_cqe *cqe,
io_uring/io_uring.c
795
ocqe = io_alloc_ocqe(ctx, cqe, big_cqe, GFP_KERNEL);
io_uring/io_uring.c
802
struct io_cqe *cqe,
io_uring/io_uring.c
807
ocqe = io_alloc_ocqe(ctx, cqe, big_cqe, GFP_NOWAIT);
io_uring/io_uring.c
818
struct io_cqe cqe = io_init_cqe(user_data, res, cflags);
io_uring/io_uring.c
820
filled = io_cqe_overflow_locked(ctx, &cqe, NULL);
io_uring/io_uring.c
836
struct io_cqe cqe = io_init_cqe(user_data, res, cflags);
io_uring/io_uring.c
838
io_cqe_overflow(ctx, &cqe, NULL);
io_uring/io_uring.c
865
posted = io_fill_cqe_aux(ctx, req->cqe.user_data, res, cflags);
io_uring/io_uring.c
868
posted = io_fill_cqe_aux(ctx, req->cqe.user_data, res, cflags);
io_uring/io_uring.c
879
bool io_req_post_cqe32(struct io_kiocb *req, struct io_uring_cqe cqe[2])
io_uring/io_uring.c
887
cqe[0].user_data = req->cqe.user_data;
io_uring/io_uring.c
890
posted = io_fill_cqe_aux32(ctx, cqe);
io_uring/io_uring.c
893
posted = io_fill_cqe_aux32(ctx, cqe);
io_uring/io_uring.h
297
bool is_cqe32 = req->cqe.flags & IORING_CQE_F_32;
io_uring/io_uring.h
298
struct io_uring_cqe *cqe;
io_uring/io_uring.h
304
if (unlikely(!io_get_cqe(ctx, &cqe, is_cqe32)))
io_uring/io_uring.h
307
memcpy(cqe, &req->cqe, sizeof(*cqe));
io_uring/io_uring.h
309
memcpy(cqe->big_cqe, &req->big_cqe, sizeof(*cqe));
io_uring/io_uring.h
314
trace_io_uring_complete(req->ctx, req, cqe);
io_uring/io_uring.h
329
req->cqe.res = res;
io_uring/io_uring.h
330
req->cqe.flags = cflags;
io_uring/io_uring.h
343
req->cqe.res = res;
io_uring/io_uring.h
344
req->cqe.flags = cflags | ctx_cqe32_flags(req->ctx);
io_uring/msg_ring.c
78
io_add_aux_cqe(ctx, req->cqe.user_data, req->cqe.res, req->cqe.flags);
io_uring/msg_ring.c
87
req->cqe.user_data = user_data;
io_uring/net.c
1349
notif->cqe.user_data = req->cqe.user_data;
io_uring/net.c
1350
notif->cqe.res = 0;
io_uring/net.c
1351
notif->cqe.flags = IORING_CQE_F_NOTIF;
io_uring/net.c
1605
req->cqe.res = sr->done_io;
io_uring/net.c
1609
req->cqe.flags |= IORING_CQE_F_MORE;
io_uring/net.c
877
cflags = req->cqe.flags | (cflags & CQE_F_MASK);
io_uring/net.c
890
req->cqe.flags = cflags & ~CQE_F_MASK;
io_uring/notif.c
31
notif->cqe.res |= IORING_NOTIF_USAGE_ZC_COPIED;
io_uring/poll.c
124
u32 index = hash_long(req->cqe.user_data, table->hash_bits);
io_uring/poll.c
263
req->cqe.res = 0;
io_uring/poll.c
266
req->cqe.res = 0;
io_uring/poll.c
279
if (!req->cqe.res) {
io_uring/poll.c
281
req->cqe.res = vfs_poll(req->file, &pt) & req->apoll_events;
io_uring/poll.c
288
if (unlikely(!req->cqe.res)) {
io_uring/poll.c
300
__poll_t mask = mangle_poll(req->cqe.res &
io_uring/poll.c
311
if ((req->cqe.res & (POLLHUP | POLLRDHUP)) && v != 1)
io_uring/poll.c
324
req->cqe.res = 0;
io_uring/poll.c
357
req->cqe.res = mangle_poll(req->cqe.res & poll->events);
io_uring/poll.c
362
req->cqe.res = ret;
io_uring/poll.c
366
io_req_set_res(req, req->cqe.res, 0);
io_uring/poll.c
762
if (cd->data != req->cqe.user_data)
io_uring/poll.c
942
preq->cqe.user_data = poll_update->new_user_data;
io_uring/poll.c
955
if (preq->cqe.res < 0)
io_uring/rw.c
1007
req->cqe.res = iov_iter_count(&io->iter);
io_uring/rw.c
1144
req->cqe.res = iov_iter_count(&io->iter);
io_uring/rw.c
1165
ret = rw_verify_area(WRITE, req->file, ppos, req->cqe.res);
io_uring/rw.c
1194
if (ret2 != req->cqe.res && ret2 >= 0 && need_complete_io(req)) {
io_uring/rw.c
1196
req->cqe.res, ret2);
io_uring/rw.c
1248
res = io_fixup_rw_res(req, req->cqe.res);
io_uring/rw.c
1249
io_req_set_res(req, res, req->cqe.flags);
io_uring/rw.c
1382
req->cqe.flags = io_put_kbuf(req, req->cqe.res, NULL);
io_uring/rw.c
552
if (res == req->cqe.res)
io_uring/rw.c
558
req->cqe.res = res;
io_uring/rw.c
583
req->cqe.flags |= io_put_kbuf(req, req->cqe.res, NULL);
io_uring/rw.c
607
if (unlikely(res != req->cqe.res)) {
io_uring/rw.c
611
req->cqe.res = res;
io_uring/rw.c
933
req->cqe.res = iov_iter_count(&io->iter);
io_uring/rw.c
947
ret = rw_verify_area(READ, req->file, ppos, req->cqe.res);
io_uring/rw.c
974
} else if (ret == req->cqe.res || ret <= 0 || !force_nonblock ||
io_uring/timeout.c
171
res = link->cqe.res;
io_uring/timeout.c
334
.data = prev->cqe.user_data,
io_uring/timeout.c
408
if (user_data == tmp->cqe.user_data) {
io_uring/uring_cmd.c
166
req->cqe.flags |= IORING_CQE_F_32;
io_uring/uring_cmd.c
351
struct io_uring_cqe cqe[2])
io_uring/uring_cmd.c
357
return io_req_post_cqe32(req, cqe);
io_uring/uring_cmd.h
21
struct io_uring_cqe cqe[2]);
io_uring/zcrx.c
1199
struct io_uring_cqe *cqe;
io_uring/zcrx.c
1202
if (!io_defer_get_uncommited_cqe(ctx, &cqe))
io_uring/zcrx.c
1205
cqe->user_data = req->cqe.user_data;
io_uring/zcrx.c
1206
cqe->res = len;
io_uring/zcrx.c
1207
cqe->flags = IORING_CQE_F_MORE;
io_uring/zcrx.c
1209
cqe->flags |= IORING_CQE_F_32;
io_uring/zcrx.c
1213
rcqe = (struct io_uring_zcrx_cqe *)(cqe + 1);
net/9p/trans_rdma.c
101
struct ib_cqe cqe;
net/9p/trans_rdma.c
183
container_of(wc->wr_cqe, struct p9_rdma_context, cqe);
net/9p/trans_rdma.c
234
container_of(wc->wr_cqe, struct p9_rdma_context, cqe);
net/9p/trans_rdma.c
284
c->cqe.done = recv_done;
net/9p/trans_rdma.c
291
wr.wr_cqe = &c->cqe;
net/9p/trans_rdma.c
381
c->cqe.done = send_done;
net/9p/trans_rdma.c
388
wr.wr_cqe = &c->cqe;
net/rds/ib_cm.c
543
cq_attr.cqe = ic->i_send_ring.w_nr + fr_queue_space + 1;
net/rds/ib_cm.c
557
cq_attr.cqe = ic->i_recv_ring.w_nr;
net/smc/smc_ib.c
844
.cqe = SMC_MAX_CQE, .comp_vector = 0 };
net/smc/smc_ib.c
856
cqattr.cqe = (0x00000001 << smc_order) * PAGE_SIZE - 2;
net/sunrpc/xprtrdma/frwr_ops.c
361
struct ib_cqe *cqe = wc->wr_cqe;
net/sunrpc/xprtrdma/frwr_ops.c
362
struct rpcrdma_mr *mr = container_of(cqe, struct rpcrdma_mr, mr_cqe);
net/sunrpc/xprtrdma/frwr_ops.c
456
struct ib_cqe *cqe = wc->wr_cqe;
net/sunrpc/xprtrdma/frwr_ops.c
457
struct rpcrdma_mr *mr = container_of(cqe, struct rpcrdma_mr, mr_cqe);
net/sunrpc/xprtrdma/frwr_ops.c
475
struct ib_cqe *cqe = wc->wr_cqe;
net/sunrpc/xprtrdma/frwr_ops.c
476
struct rpcrdma_mr *mr = container_of(cqe, struct rpcrdma_mr, mr_cqe);
net/sunrpc/xprtrdma/frwr_ops.c
572
struct ib_cqe *cqe = wc->wr_cqe;
net/sunrpc/xprtrdma/frwr_ops.c
573
struct rpcrdma_mr *mr = container_of(cqe, struct rpcrdma_mr, mr_cqe);
net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
335
struct ib_cqe *cqe = wc->wr_cqe;
net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
341
ctxt = container_of(cqe, struct svc_rdma_recv_ctxt, rc_cqe);
net/sunrpc/xprtrdma/svc_rdma_rw.c
278
struct ib_cqe *cqe = wc->wr_cqe;
net/sunrpc/xprtrdma/svc_rdma_rw.c
280
container_of(cqe, struct svc_rdma_chunk_ctxt, cc_cqe);
net/sunrpc/xprtrdma/svc_rdma_rw.c
307
struct ib_cqe *cqe = wc->wr_cqe;
net/sunrpc/xprtrdma/svc_rdma_rw.c
309
container_of(cqe, struct svc_rdma_chunk_ctxt, cc_cqe);
net/sunrpc/xprtrdma/svc_rdma_rw.c
341
struct ib_cqe *cqe = wc->wr_cqe;
net/sunrpc/xprtrdma/svc_rdma_rw.c
343
container_of(cqe, struct svc_rdma_chunk_ctxt, cc_cqe);
net/sunrpc/xprtrdma/svc_rdma_rw.c
389
struct ib_cqe *cqe;
net/sunrpc/xprtrdma/svc_rdma_rw.c
398
cqe = &cc->cc_cqe;
net/sunrpc/xprtrdma/svc_rdma_rw.c
404
rdma->sc_port_num, cqe, first_wr);
net/sunrpc/xprtrdma/svc_rdma_rw.c
405
cqe = NULL;
net/sunrpc/xprtrdma/svc_rdma_rw.c
704
struct ib_cqe *cqe;
net/sunrpc/xprtrdma/svc_rdma_rw.c
719
cqe = &cc->cc_cqe;
net/sunrpc/xprtrdma/svc_rdma_rw.c
725
rdma->sc_port_num, cqe, first_wr);
net/sunrpc/xprtrdma/svc_rdma_rw.c
726
cqe = NULL;
net/sunrpc/xprtrdma/svc_rdma_sendto.c
308
struct ib_cqe *cqe = wc->wr_cqe;
net/sunrpc/xprtrdma/svc_rdma_sendto.c
310
container_of(cqe, struct svc_rdma_send_ctxt, sc_cqe);
net/sunrpc/xprtrdma/verbs.c
144
struct ib_cqe *cqe = wc->wr_cqe;
net/sunrpc/xprtrdma/verbs.c
146
container_of(cqe, struct rpcrdma_sendctx, sc_cqe);
net/sunrpc/xprtrdma/verbs.c
163
struct ib_cqe *cqe = wc->wr_cqe;
net/sunrpc/xprtrdma/verbs.c
164
struct rpcrdma_rep *rep = container_of(cqe, struct rpcrdma_rep,
tools/testing/selftests/bpf/prog_tests/sockopt.c
1016
err = io_uring_wait_cqe(&ring, &cqe);
tools/testing/selftests/bpf/prog_tests/sockopt.c
1020
err = cqe->res;
tools/testing/selftests/bpf/prog_tests/sockopt.c
995
struct io_uring_cqe *cqe;
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
253
static void process_accept(struct io_uring *ring, struct io_uring_cqe *cqe)
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
255
if (cqe->res < 0)
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
260
connfd = cqe->res;
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
267
static void process_recvzc(struct io_uring *ring, struct io_uring_cqe *cqe)
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
278
if (cqe->res == 0 && cqe->flags == 0 && cfg_oneshot_recvs == 0) {
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
283
if (cqe->res < 0)
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
284
error(1, 0, "recvzc(): %d", cqe->res);
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
287
if (cqe->res == 0 && cqe->flags == 0 && cfg_oneshot_recvs) {
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
291
} else if (!(cqe->flags & IORING_CQE_F_MORE)) {
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
295
rcqe = (struct io_uring_zcrx_cqe *)(cqe + 1);
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
297
n = cqe->res;
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
309
rqe->len = cqe->res;
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
315
struct io_uring_cqe *cqe;
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
322
io_uring_for_each_cqe(ring, head, cqe) {
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
323
if (cqe->user_data == 1)
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
324
process_accept(ring, cqe);
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
325
else if (cqe->user_data == 2)
tools/testing/selftests/drivers/net/hw/iou-zcrx.c
326
process_recvzc(ring, cqe);
tools/testing/selftests/mm/cow.c
414
struct io_uring_cqe *cqe;
tools/testing/selftests/mm/cow.c
532
ret = io_uring_wait_cqe(&ring, &cqe);
tools/testing/selftests/mm/cow.c
539
if (cqe->res != size) {
tools/testing/selftests/mm/cow.c
544
io_uring_cqe_seen(&ring, cqe);
tools/testing/selftests/net/io_uring_zerocopy_tx.c
157
ret = io_uring_wait_cqe(&ring, &cqe);
tools/testing/selftests/net/io_uring_zerocopy_tx.c
161
if (cqe->user_data != NONZC_TAG &&
tools/testing/selftests/net/io_uring_zerocopy_tx.c
162
cqe->user_data != ZC_TAG)
tools/testing/selftests/net/io_uring_zerocopy_tx.c
165
if (cqe->flags & IORING_CQE_F_NOTIF) {
tools/testing/selftests/net/io_uring_zerocopy_tx.c
166
if (cqe->flags & IORING_CQE_F_MORE)
tools/testing/selftests/net/io_uring_zerocopy_tx.c
175
if (cqe->flags & IORING_CQE_F_MORE) {
tools/testing/selftests/net/io_uring_zerocopy_tx.c
176
if (cqe->user_data != ZC_TAG)
tools/testing/selftests/net/io_uring_zerocopy_tx.c
177
error(1, -cqe->res, "unexpected F_MORE");
tools/testing/selftests/net/io_uring_zerocopy_tx.c
180
if (cqe->res >= 0) {
tools/testing/selftests/net/io_uring_zerocopy_tx.c
182
bytes += cqe->res;
tools/testing/selftests/net/io_uring_zerocopy_tx.c
183
} else if (cqe->res != -EAGAIN) {
tools/testing/selftests/net/io_uring_zerocopy_tx.c
184
error(1, -cqe->res, "send failed");
tools/testing/selftests/net/io_uring_zerocopy_tx.c
191
ret = io_uring_wait_cqe(&ring, &cqe);
tools/testing/selftests/net/io_uring_zerocopy_tx.c
194
if (cqe->flags & IORING_CQE_F_MORE)
tools/testing/selftests/net/io_uring_zerocopy_tx.c
196
if (!(cqe->flags & IORING_CQE_F_NOTIF))
tools/testing/selftests/net/io_uring_zerocopy_tx.c
97
struct io_uring_cqe *cqe;
tools/testing/selftests/ublk/batch.c
319
const struct io_uring_cqe *cqe)
tools/testing/selftests/ublk/batch.c
321
unsigned short buf_idx = user_data_to_tag(cqe->user_data);
tools/testing/selftests/ublk/batch.c
323
unsigned end = start + cqe->res;
tools/testing/selftests/ublk/batch.c
327
if (cqe->res < 0)
tools/testing/selftests/ublk/batch.c
331
ublk_err("%s: fetch duplicated ios offset %u count %u\n", __func__, start, cqe->res);
tools/testing/selftests/ublk/batch.c
408
const struct io_uring_cqe *cqe,
tools/testing/selftests/ublk/batch.c
411
unsigned short buf_idx = user_data_to_tag(cqe->user_data);
tools/testing/selftests/ublk/batch.c
414
ublk_assert(cqe->res == 0);
tools/testing/selftests/ublk/batch.c
416
int nr_elem = user_data_to_tgt_data(cqe->user_data);
tools/testing/selftests/ublk/batch.c
418
ublk_assert(cqe->res == t->commit_buf_elem_size * nr_elem);
tools/testing/selftests/ublk/batch.c
426
const struct io_uring_cqe *cqe)
tools/testing/selftests/ublk/batch.c
428
unsigned op = user_data_to_op(cqe->user_data);
tools/testing/selftests/ublk/batch.c
436
ublk_batch_compl_commit_cmd(t, cqe, op);
tools/testing/selftests/ublk/batch.c
441
q_id = user_data_to_q_id(cqe->user_data);
tools/testing/selftests/ublk/batch.c
443
buf_idx = ublk_compl_batch_fetch(t, q, cqe);
tools/testing/selftests/ublk/batch.c
445
if (cqe->res < 0 && cqe->res != -ENOBUFS) {
tools/testing/selftests/ublk/batch.c
448
} else if (!(cqe->flags & IORING_CQE_F_MORE) || cqe->res == -ENOBUFS) {
tools/testing/selftests/ublk/fault_inject.c
62
const struct io_uring_cqe *cqe)
tools/testing/selftests/ublk/fault_inject.c
64
unsigned tag = user_data_to_tag(cqe->user_data);
tools/testing/selftests/ublk/fault_inject.c
67
if (cqe->res != -ETIME)
tools/testing/selftests/ublk/fault_inject.c
68
ublk_err("%s: unexpected cqe res %d\n", __func__, cqe->res);
tools/testing/selftests/ublk/file_backed.c
131
const struct io_uring_cqe *cqe)
tools/testing/selftests/ublk/file_backed.c
133
unsigned tag = user_data_to_tag(cqe->user_data);
tools/testing/selftests/ublk/file_backed.c
134
unsigned op = user_data_to_op(cqe->user_data);
tools/testing/selftests/ublk/file_backed.c
137
if (cqe->res < 0) {
tools/testing/selftests/ublk/file_backed.c
138
io->result = cqe->res;
tools/testing/selftests/ublk/file_backed.c
140
__func__, op, cqe->user_data);
tools/testing/selftests/ublk/file_backed.c
142
__s32 data_len = user_data_to_tgt_data(cqe->user_data)
tools/testing/selftests/ublk/file_backed.c
143
? ublk_integrity_data_len(q, cqe->res)
tools/testing/selftests/ublk/file_backed.c
144
: cqe->res;
tools/testing/selftests/ublk/kublk.c
75
struct io_uring_cqe *cqe;
tools/testing/selftests/ublk/kublk.c
827
struct io_uring_cqe *cqe)
tools/testing/selftests/ublk/kublk.c
829
if (cqe->res < 0 && cqe->res != -EAGAIN)
tools/testing/selftests/ublk/kublk.c
831
__func__, cqe->res, q->q_id,
tools/testing/selftests/ublk/kublk.c
832
user_data_to_tag(cqe->user_data),
tools/testing/selftests/ublk/kublk.c
833
user_data_to_op(cqe->user_data));
tools/testing/selftests/ublk/kublk.c
836
q->tgt_ops->tgt_io_done(t, q, cqe);
tools/testing/selftests/ublk/kublk.c
841
const struct io_uring_cqe *cqe)
tools/testing/selftests/ublk/kublk.c
843
int fetch = (cqe->res != UBLK_IO_RES_ABORT) &&
tools/testing/selftests/ublk/kublk.c
845
unsigned tag = user_data_to_tag(cqe->user_data);
tools/testing/selftests/ublk/kublk.c
855
if (cqe->res == UBLK_IO_RES_OK) {
tools/testing/selftests/ublk/kublk.c
863
} else if (cqe->res == UBLK_IO_RES_NEED_GET_DATA) {
tools/testing/selftests/ublk/kublk.c
880
struct io_uring_cqe *cqe, void *data)
tools/testing/selftests/ublk/kublk.c
883
unsigned q_id = user_data_to_q_id(cqe->user_data);
tools/testing/selftests/ublk/kublk.c
884
unsigned cmd_op = user_data_to_op(cqe->user_data);
tools/testing/selftests/ublk/kublk.c
886
if (cqe->res < 0 && cqe->res != -ENODEV && cqe->res != -ENOBUFS)
tools/testing/selftests/ublk/kublk.c
888
cqe->res, cqe->user_data, t->state);
tools/testing/selftests/ublk/kublk.c
892
__func__, cqe->res, t->idx, q_id,
tools/testing/selftests/ublk/kublk.c
893
user_data_to_tag(cqe->user_data),
tools/testing/selftests/ublk/kublk.c
894
cmd_op, cqe->user_data, is_target_io(cqe->user_data),
tools/testing/selftests/ublk/kublk.c
895
user_data_to_tgt_data(cqe->user_data),
tools/testing/selftests/ublk/kublk.c
899
if (is_target_io(cqe->user_data)) {
tools/testing/selftests/ublk/kublk.c
900
ublksrv_handle_tgt_cqe(t, &dev->q[q_id], cqe);
tools/testing/selftests/ublk/kublk.c
905
ublk_batch_compl_cmd(t, cqe);
tools/testing/selftests/ublk/kublk.c
907
ublk_handle_uring_cmd(t, &dev->q[q_id], cqe);
tools/testing/selftests/ublk/kublk.c
912
struct io_uring_cqe *cqe;
tools/testing/selftests/ublk/kublk.c
916
io_uring_for_each_cqe(&t->ring, head, cqe) {
tools/testing/selftests/ublk/kublk.c
917
ublk_handle_cqe(t, cqe, NULL);
tools/testing/selftests/ublk/kublk.c
92
ret = io_uring_wait_cqe(&dev->ring, &cqe);
tools/testing/selftests/ublk/kublk.c
97
io_uring_cqe_seen(&dev->ring, cqe);
tools/testing/selftests/ublk/kublk.c
99
return cqe->res;
tools/testing/selftests/ublk/kublk.h
556
const struct io_uring_cqe *cqe);
tools/testing/selftests/ublk/null.c
101
if (cqe->res < 0 || op != ublk_cmd_op_nr(UBLK_U_IO_UNREGISTER_IO_BUF)) {
tools/testing/selftests/ublk/null.c
103
io->result = cqe->res;
tools/testing/selftests/ublk/null.c
104
if (cqe->res < 0)
tools/testing/selftests/ublk/null.c
106
__func__, op, cqe->user_data);
tools/testing/selftests/ublk/null.c
95
const struct io_uring_cqe *cqe)
tools/testing/selftests/ublk/null.c
97
unsigned tag = user_data_to_tag(cqe->user_data);
tools/testing/selftests/ublk/null.c
98
unsigned op = user_data_to_op(cqe->user_data);
tools/testing/selftests/ublk/stripe.c
235
const struct io_uring_cqe *cqe)
tools/testing/selftests/ublk/stripe.c
237
unsigned tag = user_data_to_tag(cqe->user_data);
tools/testing/selftests/ublk/stripe.c
239
unsigned op = user_data_to_op(cqe->user_data);
tools/testing/selftests/ublk/stripe.c
241
int res = cqe->res;
tools/testing/selftests/ublk/stripe.c
256
unsigned seq = user_data_to_tgt_data(cqe->user_data);
tools/testing/selftests/x86/lam.c
571
struct io_uring_cqe *cqe;
tools/testing/selftests/x86/lam.c
582
cqe = &cring->queue.cqes[head & *s->cq_ring.ring_mask];
tools/testing/selftests/x86/lam.c
583
fi = (struct file_io *)cqe->user_data;
tools/testing/selftests/x86/lam.c
584
if (cqe->res < 0)
tools/testing/vsock/vsock_uring_test.c
102
io_uring_cqe_seen(&ring, cqe);
tools/testing/vsock/vsock_uring_test.c
144
struct io_uring_cqe *cqe;
tools/testing/vsock/vsock_uring_test.c
156
if (io_uring_wait_cqe(&ring, &cqe))
tools/testing/vsock/vsock_uring_test.c
159
recv_len += cqe->res;
tools/testing/vsock/vsock_uring_test.c
160
io_uring_cqe_seen(&ring, cqe);
tools/testing/vsock/vsock_uring_test.c
63
struct io_uring_cqe *cqe;
tools/testing/vsock/vsock_uring_test.c
99
if (io_uring_wait_cqe(&ring, &cqe))