rx_qp
gsi->rx_qp = ib_create_qp(pd, &hw_init_attr);
if (IS_ERR(gsi->rx_qp)) {
gsi->rx_qp);
ret = PTR_ERR(gsi->rx_qp);
ret = ib_destroy_qp(gsi->rx_qp);
gsi->rx_qp = NULL;
struct ib_pd *pd = gsi->rx_qp->pd;
.event_handler = gsi->rx_qp->event_handler,
.qp_context = gsi->rx_qp->qp_context,
.recv_cq = gsi->rx_qp->recv_cq,
struct ib_device *device = gsi->rx_qp->device;
ret = ib_modify_qp(gsi->rx_qp, attr, attr_mask);
if (to_mqp(gsi->rx_qp)->state != IB_QPS_RTS)
ret = ib_query_qp(gsi->rx_qp, qp_attr, qp_attr_mask, qp_init_attr);
struct mlx5_ib_dev *dev = to_mdev(gsi->rx_qp->device);
struct mlx5_ib_dev *dev = to_mdev(gsi->rx_qp->device);
return gsi->rx_qp;
return ib_post_recv(gsi->rx_qp, wr, bad_wr);
struct ib_qp *rx_qp;
siw_init_terminate(rx_qp(srx), TERM_ERROR_LAYER_DDP,
siw_init_terminate(rx_qp(srx), TERM_ERROR_LAYER_RDMAP,
siw_init_terminate(rx_qp(srx), TERM_ERROR_LAYER_RDMAP,
siw_dbg_qp(rx_qp(srx), "new header, opcode %u\n", opcode);
siw_dbg_qp(rx_qp(srx), "kva: 0x%p, len: %u\n", kva, len);
qp_id(rx_qp(srx)), __func__, len, kva, rv);
qp_id(rx_qp(srx)), sink_stag, srx->ddp_stag);
qp_id(rx_qp(srx)), (unsigned long long)sink_to,
qp_id(rx_qp(srx)),
siw_init_terminate(rx_qp(srx), TERM_ERROR_LAYER_DDP,
qp_id(rx_qp(srx)), sink_stag,
qp_id(rx_qp(srx)),
siw_init_terminate(rx_qp(srx), TERM_ERROR_LAYER_DDP,
qp_id(rx_qp(srx)), ddp_qn);
qp_id(rx_qp(srx)), ddp_msn,
qp_id(rx_qp(srx)), ddp_mo, wqe->processed);
siw_dbg_qp(rx_qp(srx), "receive space short: %d - %d < %d\n",
siw_init_terminate(rx_qp(srx), TERM_ERROR_LAYER_DDP,
__func__, qp_id(rx_qp(srx)),
siw_dbg_qp(rx_qp(srx), "page %p, bytes=%u\n", p, bytes);
qp_id(rx_qp(srx)), __func__, len, p, rv);
if (rdma_is_kernel_res(&rx_qp(srx)->base_qp.res)) {
struct siw_qp *qp = rx_qp(srx);