root/drivers/infiniband/hw/irdma/virtchnl.c
// SPDX-License-Identifier: GPL-2.0 or Linux-OpenIB
/* Copyright (c) 2015 - 2024 Intel Corporation */

#include "osdep.h"
#include "hmc.h"
#include "defs.h"
#include "type.h"
#include "protos.h"
#include "virtchnl.h"
#include "ws.h"
#include "i40iw_hw.h"
#include "ig3rdma_hw.h"

struct vchnl_reg_map_elem {
        u16 reg_id;
        u16 reg_idx;
        bool pg_rel;
};

struct vchnl_regfld_map_elem {
        u16 regfld_id;
        u16 regfld_idx;
};

static struct vchnl_reg_map_elem vchnl_reg_map[] = {
        {IRDMA_VCHNL_REG_ID_CQPTAIL, IRDMA_CQPTAIL, false},
        {IRDMA_VCHNL_REG_ID_CQPDB, IRDMA_CQPDB, false},
        {IRDMA_VCHNL_REG_ID_CCQPSTATUS, IRDMA_CCQPSTATUS, false},
        {IRDMA_VCHNL_REG_ID_CCQPHIGH, IRDMA_CCQPHIGH, false},
        {IRDMA_VCHNL_REG_ID_CCQPLOW, IRDMA_CCQPLOW, false},
        {IRDMA_VCHNL_REG_ID_CQARM, IRDMA_CQARM, false},
        {IRDMA_VCHNL_REG_ID_CQACK, IRDMA_CQACK, false},
        {IRDMA_VCHNL_REG_ID_AEQALLOC, IRDMA_AEQALLOC, false},
        {IRDMA_VCHNL_REG_ID_CQPERRCODES, IRDMA_CQPERRCODES, false},
        {IRDMA_VCHNL_REG_ID_WQEALLOC, IRDMA_WQEALLOC, false},
        {IRDMA_VCHNL_REG_ID_DB_ADDR_OFFSET, IRDMA_DB_ADDR_OFFSET, false },
        {IRDMA_VCHNL_REG_ID_DYN_CTL, IRDMA_GLINT_DYN_CTL, false },
        {IRDMA_VCHNL_REG_INV_ID, IRDMA_VCHNL_REG_INV_ID, false }
};

static struct vchnl_regfld_map_elem vchnl_regfld_map[] = {
        {IRDMA_VCHNL_REGFLD_ID_CCQPSTATUS_CQP_OP_ERR, IRDMA_CCQPSTATUS_CCQP_ERR_M},
        {IRDMA_VCHNL_REGFLD_ID_CCQPSTATUS_CCQP_DONE, IRDMA_CCQPSTATUS_CCQP_DONE_M},
        {IRDMA_VCHNL_REGFLD_ID_CQPSQ_STAG_PDID, IRDMA_CQPSQ_STAG_PDID_M},
        {IRDMA_VCHNL_REGFLD_ID_CQPSQ_CQ_CEQID, IRDMA_CQPSQ_CQ_CEQID_M},
        {IRDMA_VCHNL_REGFLD_ID_CQPSQ_CQ_CQID, IRDMA_CQPSQ_CQ_CQID_M},
        {IRDMA_VCHNL_REGFLD_ID_COMMIT_FPM_CQCNT, IRDMA_COMMIT_FPM_CQCNT_M},
        {IRDMA_VCHNL_REGFLD_ID_UPESD_HMCN_ID, IRDMA_CQPSQ_UPESD_HMCFNID_M},
        {IRDMA_VCHNL_REGFLD_INV_ID, IRDMA_VCHNL_REGFLD_INV_ID}
};

#define IRDMA_VCHNL_REG_COUNT ARRAY_SIZE(vchnl_reg_map)
#define IRDMA_VCHNL_REGFLD_COUNT ARRAY_SIZE(vchnl_regfld_map)
#define IRDMA_VCHNL_REGFLD_BUF_SIZE \
        (IRDMA_VCHNL_REG_COUNT * sizeof(struct irdma_vchnl_reg_info) + \
         IRDMA_VCHNL_REGFLD_COUNT * sizeof(struct irdma_vchnl_reg_field_info))
#define IRDMA_REGMAP_RESP_BUF_SIZE (IRDMA_VCHNL_RESP_MIN_SIZE + IRDMA_VCHNL_REGFLD_BUF_SIZE)

/**
 * irdma_sc_vchnl_init - Initialize dev virtchannel and get hw_rev
 * @dev: dev structure to update
 * @info: virtchannel info parameters to fill into the dev structure
 */
int irdma_sc_vchnl_init(struct irdma_sc_dev *dev,
                        struct irdma_vchnl_init_info *info)
{
        dev->vchnl_up = true;
        dev->privileged = info->privileged;
        dev->is_pf = info->is_pf;
        dev->hw_attrs.uk_attrs.hw_rev = info->hw_rev;

        if (!dev->privileged) {
                int ret = irdma_vchnl_req_get_ver(dev, IRDMA_VCHNL_CHNL_VER_MAX,
                                                  &dev->vchnl_ver);

                ibdev_dbg(to_ibdev(dev),
                          "DEV: Get Channel version ret = %d, version is %u\n",
                          ret, dev->vchnl_ver);

                if (ret)
                        return ret;

                ret = irdma_vchnl_req_get_caps(dev);
                if (ret)
                        return ret;

                dev->hw_attrs.uk_attrs.hw_rev = dev->vc_caps.hw_rev;
        }

        return 0;
}

/**
 * irdma_vchnl_req_verify_resp - Verify requested response size
 * @vchnl_req: vchnl message requested
 * @resp_len: response length sent from vchnl peer
 */
static int irdma_vchnl_req_verify_resp(struct irdma_vchnl_req *vchnl_req,
                                       u16 resp_len)
{
        switch (vchnl_req->vchnl_msg->op_code) {
        case IRDMA_VCHNL_OP_GET_VER:
        case IRDMA_VCHNL_OP_GET_HMC_FCN:
        case IRDMA_VCHNL_OP_PUT_HMC_FCN:
                if (resp_len != vchnl_req->parm_len)
                        return -EBADMSG;
                break;
        case IRDMA_VCHNL_OP_GET_RDMA_CAPS:
                if (resp_len < IRDMA_VCHNL_OP_GET_RDMA_CAPS_MIN_SIZE)
                        return -EBADMSG;
                break;
        case IRDMA_VCHNL_OP_GET_REG_LAYOUT:
        case IRDMA_VCHNL_OP_QUEUE_VECTOR_MAP:
        case IRDMA_VCHNL_OP_QUEUE_VECTOR_UNMAP:
        case IRDMA_VCHNL_OP_ADD_VPORT:
        case IRDMA_VCHNL_OP_DEL_VPORT:
                break;
        default:
                return -EOPNOTSUPP;
        }

        return 0;
}

static void irdma_free_vchnl_req_msg(struct irdma_vchnl_req *vchnl_req)
{
        kfree(vchnl_req->vchnl_msg);
}

static int irdma_alloc_vchnl_req_msg(struct irdma_vchnl_req *vchnl_req,
                                     struct irdma_vchnl_req_init_info *info)
{
        struct irdma_vchnl_op_buf *vchnl_msg;

        vchnl_msg = kzalloc(IRDMA_VCHNL_MAX_MSG_SIZE, GFP_KERNEL);

        if (!vchnl_msg)
                return -ENOMEM;

        vchnl_msg->op_ctx = (uintptr_t)vchnl_req;
        vchnl_msg->buf_len = sizeof(*vchnl_msg) + info->req_parm_len;
        if (info->req_parm_len)
                memcpy(vchnl_msg->buf, info->req_parm, info->req_parm_len);
        vchnl_msg->op_code = info->op_code;
        vchnl_msg->op_ver = info->op_ver;

        vchnl_req->vchnl_msg = vchnl_msg;
        vchnl_req->parm = info->resp_parm;
        vchnl_req->parm_len = info->resp_parm_len;

        return 0;
}

static int irdma_vchnl_req_send_sync(struct irdma_sc_dev *dev,
                                     struct irdma_vchnl_req_init_info *info)
{
        u16 resp_len = sizeof(dev->vc_recv_buf);
        struct irdma_vchnl_req vchnl_req = {};
        u16 msg_len;
        u8 *msg;
        int ret;

        ret = irdma_alloc_vchnl_req_msg(&vchnl_req, info);
        if (ret)
                return ret;

        msg_len = vchnl_req.vchnl_msg->buf_len;
        msg = (u8 *)vchnl_req.vchnl_msg;

        mutex_lock(&dev->vchnl_mutex);
        ret = ig3rdma_vchnl_send_sync(dev, msg, msg_len, dev->vc_recv_buf,
                                      &resp_len);
        dev->vc_recv_len = resp_len;
        if (ret)
                goto exit;

        ret = irdma_vchnl_req_get_resp(dev, &vchnl_req);
exit:
        mutex_unlock(&dev->vchnl_mutex);
        ibdev_dbg(to_ibdev(dev),
                  "VIRT: virtual channel send %s caller: %pS ret=%d op=%u op_ver=%u req_len=%u parm_len=%u resp_len=%u\n",
                  !ret ? "SUCCEEDS" : "FAILS", __builtin_return_address(0),
                  ret, vchnl_req.vchnl_msg->op_code,
                  vchnl_req.vchnl_msg->op_ver, vchnl_req.vchnl_msg->buf_len,
                  vchnl_req.parm_len, vchnl_req.resp_len);
        irdma_free_vchnl_req_msg(&vchnl_req);

        return ret;
}

/**
 * irdma_vchnl_req_get_reg_layout - Get Register Layout
 * @dev: RDMA device pointer
 */
int irdma_vchnl_req_get_reg_layout(struct irdma_sc_dev *dev)
{
        u16 reg_idx, reg_id, tmp_reg_id, regfld_idx, regfld_id, tmp_regfld_id;
        struct irdma_vchnl_reg_field_info *regfld_array = NULL;
        u8 resp_buffer[IRDMA_REGMAP_RESP_BUF_SIZE] = {};
        struct vchnl_regfld_map_elem *regfld_map_array;
        struct irdma_vchnl_req_init_info info = {};
        struct vchnl_reg_map_elem *reg_map_array;
        struct irdma_vchnl_reg_info *reg_array;
        u8 num_bits, shift_cnt;
        u16 buf_len = 0;
        u64 bitmask;
        u32 rindex;
        int ret;

        if (!dev->vchnl_up)
                return -EBUSY;

        info.op_code = IRDMA_VCHNL_OP_GET_REG_LAYOUT;
        info.op_ver = IRDMA_VCHNL_OP_GET_REG_LAYOUT_V0;
        info.resp_parm = resp_buffer;
        info.resp_parm_len = sizeof(resp_buffer);

        ret = irdma_vchnl_req_send_sync(dev, &info);

        if (ret)
                return ret;

        /* parse the response buffer and update reg info*/
        /* Parse registers till invalid */
        /* Parse register fields till invalid */
        reg_array = (struct irdma_vchnl_reg_info *)resp_buffer;
        for (rindex = 0; rindex < IRDMA_VCHNL_REG_COUNT; rindex++) {
                buf_len += sizeof(struct irdma_vchnl_reg_info);
                if (buf_len >= sizeof(resp_buffer))
                        return -ENOMEM;

                regfld_array =
                        (struct irdma_vchnl_reg_field_info *)&reg_array[rindex + 1];
                reg_id = reg_array[rindex].reg_id;
                if (reg_id == IRDMA_VCHNL_REG_INV_ID)
                        break;

                reg_id &= ~IRDMA_VCHNL_REG_PAGE_REL;
                if (reg_id >= IRDMA_VCHNL_REG_COUNT)
                        return -EINVAL;

                /* search regmap for register index in hw_regs.*/
                reg_map_array = vchnl_reg_map;
                do {
                        tmp_reg_id = reg_map_array->reg_id;
                        if (tmp_reg_id == reg_id)
                                break;

                        reg_map_array++;
                } while (tmp_reg_id != IRDMA_VCHNL_REG_INV_ID);
                if (tmp_reg_id != reg_id)
                        continue;

                reg_idx = reg_map_array->reg_idx;

                /* Page relative, DB Offset do not need bar offset */
                if (reg_idx == IRDMA_DB_ADDR_OFFSET ||
                    (reg_array[rindex].reg_id & IRDMA_VCHNL_REG_PAGE_REL)) {
                        dev->hw_regs[reg_idx] =
                                (u32 __iomem *)(uintptr_t)reg_array[rindex].reg_offset;
                        continue;
                }

                /* Update the local HW struct */
                dev->hw_regs[reg_idx] = ig3rdma_get_reg_addr(dev->hw,
                                                reg_array[rindex].reg_offset);
                if (!dev->hw_regs[reg_idx])
                        return -EINVAL;
        }

        if (!regfld_array)
                return -ENOMEM;

        /* set up doorbell variables using mapped DB page */
        dev->wqe_alloc_db = dev->hw_regs[IRDMA_WQEALLOC];
        dev->cq_arm_db = dev->hw_regs[IRDMA_CQARM];
        dev->aeq_alloc_db = dev->hw_regs[IRDMA_AEQALLOC];
        dev->cqp_db = dev->hw_regs[IRDMA_CQPDB];
        dev->cq_ack_db = dev->hw_regs[IRDMA_CQACK];

        for (rindex = 0; rindex < IRDMA_VCHNL_REGFLD_COUNT; rindex++) {
                buf_len += sizeof(struct irdma_vchnl_reg_field_info);
                if ((buf_len - 1) > sizeof(resp_buffer))
                        break;

                if (regfld_array[rindex].fld_id == IRDMA_VCHNL_REGFLD_INV_ID)
                        break;

                regfld_id = regfld_array[rindex].fld_id;
                regfld_map_array = vchnl_regfld_map;
                do {
                        tmp_regfld_id = regfld_map_array->regfld_id;
                        if (tmp_regfld_id == regfld_id)
                                break;

                        regfld_map_array++;
                } while (tmp_regfld_id != IRDMA_VCHNL_REGFLD_INV_ID);

                if (tmp_regfld_id != regfld_id)
                        continue;

                regfld_idx = regfld_map_array->regfld_idx;

                num_bits = regfld_array[rindex].fld_bits;
                shift_cnt = regfld_array[rindex].fld_shift;
                if ((num_bits + shift_cnt > 64) || !num_bits) {
                        ibdev_dbg(to_ibdev(dev),
                                  "ERR: Invalid field mask id %d bits %d shift %d",
                                  regfld_id, num_bits, shift_cnt);

                        continue;
                }

                bitmask = (1ULL << num_bits) - 1;
                dev->hw_masks[regfld_idx] = bitmask << shift_cnt;
                dev->hw_shifts[regfld_idx] = shift_cnt;
        }

        return 0;
}

int irdma_vchnl_req_add_vport(struct irdma_sc_dev *dev, u16 vport_id,
                              u32 qp1_id, struct irdma_qos *qos)
{
        struct irdma_vchnl_resp_vport_info resp_vport = { 0 };
        struct irdma_vchnl_req_vport_info req_vport = { 0 };
        struct irdma_vchnl_req_init_info info = { 0 };
        int ret, i;

        if (!dev->vchnl_up)
                return -EBUSY;

        info.op_code = IRDMA_VCHNL_OP_ADD_VPORT;
        info.op_ver = IRDMA_VCHNL_OP_ADD_VPORT_V0;
        req_vport.vport_id = vport_id;
        req_vport.qp1_id = qp1_id;
        info.req_parm_len = sizeof(req_vport);
        info.req_parm = &req_vport;
        info.resp_parm = &resp_vport;
        info.resp_parm_len = sizeof(resp_vport);

        ret = irdma_vchnl_req_send_sync(dev, &info);
        if (ret)
                return ret;

        for (i = 0;  i < IRDMA_MAX_USER_PRIORITY; i++) {
                qos[i].qs_handle = resp_vport.qs_handle[i];
                qos[i].valid = true;
        }

        return 0;
}

int irdma_vchnl_req_del_vport(struct irdma_sc_dev *dev, u16 vport_id, u32 qp1_id)
{
        struct irdma_vchnl_req_init_info info = { 0 };
        struct irdma_vchnl_req_vport_info req_vport = { 0 };

        if (!dev->vchnl_up)
                return -EBUSY;

        info.op_code = IRDMA_VCHNL_OP_DEL_VPORT;
        info.op_ver = IRDMA_VCHNL_OP_DEL_VPORT_V0;
        req_vport.vport_id = vport_id;
        req_vport.qp1_id = qp1_id;
        info.req_parm_len = sizeof(req_vport);
        info.req_parm = &req_vport;

        return irdma_vchnl_req_send_sync(dev, &info);
}

/**
 * irdma_vchnl_req_aeq_vec_map - Map AEQ to vector on this function
 * @dev: RDMA device pointer
 * @v_idx: vector index
 */
int irdma_vchnl_req_aeq_vec_map(struct irdma_sc_dev *dev, u32 v_idx)
{
        struct irdma_vchnl_req_init_info info = {};
        struct irdma_vchnl_qvlist_info *qvl;
        struct irdma_vchnl_qv_info *qv;
        u16 qvl_size, num_vectors = 1;
        int ret;

        if (!dev->vchnl_up)
                return -EBUSY;

        qvl_size = struct_size(qvl, qv_info, num_vectors);

        qvl = kzalloc(qvl_size, GFP_KERNEL);
        if (!qvl)
                return -ENOMEM;

        qvl->num_vectors = 1;
        qv = qvl->qv_info;

        qv->ceq_idx = IRDMA_Q_INVALID_IDX;
        qv->v_idx = v_idx;
        qv->itr_idx = IRDMA_IDX_ITR0;

        info.op_code = IRDMA_VCHNL_OP_QUEUE_VECTOR_MAP;
        info.op_ver = IRDMA_VCHNL_OP_QUEUE_VECTOR_MAP_V0;
        info.req_parm = qvl;
        info.req_parm_len = qvl_size;

        ret = irdma_vchnl_req_send_sync(dev, &info);
        kfree(qvl);

        return ret;
}

/**
 * irdma_vchnl_req_ceq_vec_map - Map CEQ to vector on this function
 * @dev: RDMA device pointer
 * @ceq_id: CEQ index
 * @v_idx: vector index
 */
int irdma_vchnl_req_ceq_vec_map(struct irdma_sc_dev *dev, u16 ceq_id, u32 v_idx)
{
        struct irdma_vchnl_req_init_info info = {};
        struct irdma_vchnl_qvlist_info *qvl;
        struct irdma_vchnl_qv_info *qv;
        u16 qvl_size, num_vectors = 1;
        int ret;

        if (!dev->vchnl_up)
                return -EBUSY;

        qvl_size = struct_size(qvl, qv_info, num_vectors);

        qvl = kzalloc(qvl_size, GFP_KERNEL);
        if (!qvl)
                return -ENOMEM;

        qvl->num_vectors = num_vectors;
        qv = qvl->qv_info;

        qv->aeq_idx = IRDMA_Q_INVALID_IDX;
        qv->ceq_idx = ceq_id;
        qv->v_idx = v_idx;
        qv->itr_idx = IRDMA_IDX_ITR0;

        info.op_code = IRDMA_VCHNL_OP_QUEUE_VECTOR_MAP;
        info.op_ver = IRDMA_VCHNL_OP_QUEUE_VECTOR_MAP_V0;
        info.req_parm = qvl;
        info.req_parm_len = qvl_size;

        ret = irdma_vchnl_req_send_sync(dev, &info);
        kfree(qvl);

        return ret;
}

/**
 * irdma_vchnl_req_get_ver - Request Channel version
 * @dev: RDMA device pointer
 * @ver_req: Virtual channel version requested
 * @ver_res: Virtual channel version response
 */
int irdma_vchnl_req_get_ver(struct irdma_sc_dev *dev, u16 ver_req, u32 *ver_res)
{
        struct irdma_vchnl_req_init_info info = {};
        int ret;

        if (!dev->vchnl_up)
                return -EBUSY;

        info.op_code = IRDMA_VCHNL_OP_GET_VER;
        info.op_ver = ver_req;
        info.resp_parm = ver_res;
        info.resp_parm_len = sizeof(*ver_res);

        ret = irdma_vchnl_req_send_sync(dev, &info);
        if (ret)
                return ret;

        if (*ver_res < IRDMA_VCHNL_CHNL_VER_MIN) {
                ibdev_dbg(to_ibdev(dev),
                          "VIRT: %s unsupported vchnl version 0x%0x\n",
                          __func__, *ver_res);
                return -EOPNOTSUPP;
        }

        return 0;
}

/**
 * irdma_vchnl_req_get_hmc_fcn - Request VF HMC Function
 * @dev: RDMA device pointer
 */
int irdma_vchnl_req_get_hmc_fcn(struct irdma_sc_dev *dev)
{
        struct irdma_vchnl_req_hmc_info req_hmc = {};
        struct irdma_vchnl_resp_hmc_info resp_hmc = {};
        struct irdma_vchnl_req_init_info info = {};
        int ret;

        if (!dev->vchnl_up)
                return -EBUSY;

        info.op_code = IRDMA_VCHNL_OP_GET_HMC_FCN;
        if (dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_3) {
                info.op_ver = IRDMA_VCHNL_OP_GET_HMC_FCN_V2;
                req_hmc.protocol_used = dev->protocol_used;
                info.req_parm_len = sizeof(req_hmc);
                info.req_parm = &req_hmc;
                info.resp_parm = &resp_hmc;
                info.resp_parm_len = sizeof(resp_hmc);
        }

        ret = irdma_vchnl_req_send_sync(dev, &info);

        if (ret)
                return ret;

        if (dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_3) {
                int i;

                dev->hmc_fn_id = resp_hmc.hmc_func;

                for (i = 0;  i < IRDMA_MAX_USER_PRIORITY; i++) {
                        dev->qos[i].qs_handle = resp_hmc.qs_handle[i];
                        dev->qos[i].valid = true;
                }
        }
        return 0;
}

/**
 * irdma_vchnl_req_put_hmc_fcn - Free VF HMC Function
 * @dev: RDMA device pointer
 */
int irdma_vchnl_req_put_hmc_fcn(struct irdma_sc_dev *dev)
{
        struct irdma_vchnl_req_init_info info = {};

        if (!dev->vchnl_up)
                return -EBUSY;

        info.op_code = IRDMA_VCHNL_OP_PUT_HMC_FCN;
        info.op_ver = IRDMA_VCHNL_OP_PUT_HMC_FCN_V0;

        return irdma_vchnl_req_send_sync(dev, &info);
}

/**
 * irdma_vchnl_req_get_caps - Request RDMA capabilities
 * @dev: RDMA device pointer
 */
int irdma_vchnl_req_get_caps(struct irdma_sc_dev *dev)
{
        struct irdma_vchnl_req_init_info info = {};
        int ret;

        if (!dev->vchnl_up)
                return -EBUSY;

        info.op_code = IRDMA_VCHNL_OP_GET_RDMA_CAPS;
        info.op_ver = IRDMA_VCHNL_OP_GET_RDMA_CAPS_V0;
        info.resp_parm = &dev->vc_caps;
        info.resp_parm_len = sizeof(dev->vc_caps);

        ret = irdma_vchnl_req_send_sync(dev, &info);

        if (ret)
                return ret;

        if (dev->vc_caps.hw_rev > IRDMA_GEN_MAX ||
            dev->vc_caps.hw_rev < IRDMA_GEN_2) {
                ibdev_dbg(to_ibdev(dev),
                          "ERR: %s unsupported hw_rev version 0x%0x\n",
                          __func__, dev->vc_caps.hw_rev);
                return -EOPNOTSUPP;
        }

        return 0;
}

/**
 * irdma_vchnl_req_get_resp - Receive the inbound vchnl response.
 * @dev: Dev pointer
 * @vchnl_req: Vchannel request
 */
int irdma_vchnl_req_get_resp(struct irdma_sc_dev *dev,
                             struct irdma_vchnl_req *vchnl_req)
{
        struct irdma_vchnl_resp_buf *vchnl_msg_resp =
                (struct irdma_vchnl_resp_buf *)dev->vc_recv_buf;
        u16 resp_len;
        int ret;

        if ((uintptr_t)vchnl_req != (uintptr_t)vchnl_msg_resp->op_ctx) {
                ibdev_dbg(to_ibdev(dev),
                          "VIRT: error vchnl context value does not match\n");
                return -EBADMSG;
        }

        resp_len = dev->vc_recv_len - sizeof(*vchnl_msg_resp);
        resp_len = min(resp_len, vchnl_req->parm_len);

        ret = irdma_vchnl_req_verify_resp(vchnl_req, resp_len);
        if (ret)
                return ret;

        ret = (int)vchnl_msg_resp->op_ret;
        if (ret)
                return ret;

        vchnl_req->resp_len = 0;
        if (vchnl_req->parm_len && vchnl_req->parm && resp_len) {
                memcpy(vchnl_req->parm, vchnl_msg_resp->buf, resp_len);
                vchnl_req->resp_len = resp_len;
                ibdev_dbg(to_ibdev(dev), "VIRT: Got response, data size %u\n",
                          resp_len);
        }

        return 0;
}