nvmet_fc_fcp_iod
struct nvmet_fc_fcp_iod fod[] /* __counted_by(sqsize) */;
nvmet_fc_alloc_tgt_pgs(struct nvmet_fc_fcp_iod *fod)
nvmet_fc_free_tgt_pgs(struct nvmet_fc_fcp_iod *fod)
static inline struct nvmet_fc_fcp_iod *
struct nvmet_fc_fcp_iod *fod)
return container_of(nvme_req, struct nvmet_fc_fcp_iod, req);
struct nvmet_fc_fcp_iod *fod)
struct nvmet_fc_fcp_iod *fod)
struct nvmet_fc_fcp_iod *fod, u8 op)
__nvmet_fc_fod_op_abort(struct nvmet_fc_fcp_iod *fod, bool abort)
nvmet_fc_fod_op_done(struct nvmet_fc_fcp_iod *fod)
struct nvmet_fc_fcp_iod *fod = fcpreq->nvmet_fc_private;
struct nvmet_fc_fcp_iod *fod, int status)
struct nvmet_fc_fcp_iod *fod);
struct nvmet_fc_fcp_iod *fod = nvmet_req_to_fod(nvme_req);
struct nvmet_fc_fcp_iod *fod)
struct nvmet_fc_fcp_iod *fod;
struct nvmet_fc_fcp_iod *fod = fcpreq->nvmet_fc_private;
struct nvmet_fc_fcp_iod *fod = queue->fod;
struct nvmet_fc_fcp_iod *fod = queue->fod;
static struct nvmet_fc_fcp_iod *
struct nvmet_fc_fcp_iod *fod;
struct nvmet_fc_fcp_iod, fcp_list);
struct nvmet_fc_fcp_iod *fod = fcpreq->nvmet_fc_private;
struct nvmet_fc_fcp_iod *fod =
container_of(work, struct nvmet_fc_fcp_iod, defer_work);
struct nvmet_fc_fcp_iod *fod)
struct nvmet_fc_fcp_iod *fod = queue->fod;