nvme_completion
memset(q->cqes, 0, depth * sizeof(struct nvme_completion));
struct nvme_completion *cqes;
depth * sizeof(struct nvme_completion),
struct nvme_completion *hcqe = &q->cqes[q->cq_head];
struct nvme_completion *cqe = &q->cqes[idx];
struct nvme_completion *cqe = &op->rsp_iu.cqe;
struct nvme_completion *hcqe = &nvmeq->cqes[nvmeq->cq_head];
struct nvme_completion *cqe = &nvmeq->cqes[idx];
#define CQ_SIZE(q) ((q)->q_depth * sizeof(struct nvme_completion))
struct nvme_completion *cqes;
list.length = sizeof(struct nvme_completion);
struct nvme_completion *cqe, struct ib_wc *wc)
struct nvme_completion *cqe = qe->data;
const size_t len = sizeof(struct nvme_completion);
sizeof(struct nvme_completion), DMA_FROM_DEVICE);
sizeof(struct nvme_completion), DMA_FROM_DEVICE);
sizeof(struct nvme_completion), DMA_FROM_DEVICE);
struct nvme_completion *cqe)
struct nvme_completion *cqe = &pdu->cqe;
id->iorcsz = cpu_to_le32(sizeof(struct nvme_completion) / 16);
struct nvme_completion *cqe = &ersp->cqe;
struct nvme_completion *cqe = &fod->rspiubuf.cqe;
struct nvme_completion cqe;
struct nvme_completion *cqe = req->cqe;
struct nvme_completion *cqe;
id->iorcsz = cpu_to_le32(sizeof(struct nvme_completion) / 16);
struct nvme_completion cqe;
cq->qes = sizeof(struct nvme_completion);
struct nvme_completion *cqe;
if (ctrl->io_cqes < sizeof(struct nvme_completion)) {
ctrl->io_cqes, sizeof(struct nvme_completion));
struct nvme_completion cqe;
struct nvme_completion cqe;