root/drivers/scsi/bnx2fc/bnx2fc_els.c
/*
 * bnx2fc_els.c: QLogic Linux FCoE offload driver.
 * This file contains helper routines that handle ELS requests
 * and responses.
 *
 * Copyright (c) 2008-2013 Broadcom Corporation
 * Copyright (c) 2014-2016 QLogic Corporation
 * Copyright (c) 2016-2017 Cavium Inc.
 *
 * This program is free software; you can redistribute it and/or modify
 * it under the terms of the GNU General Public License as published by
 * the Free Software Foundation.
 *
 * Written by: Bhanu Prakash Gollapudi (bprakash@broadcom.com)
 */

#include "bnx2fc.h"

static void bnx2fc_logo_resp(struct fc_seq *seq, struct fc_frame *fp,
                             void *arg);
static void bnx2fc_flogi_resp(struct fc_seq *seq, struct fc_frame *fp,
                              void *arg);
static int bnx2fc_initiate_els(struct bnx2fc_rport *tgt, unsigned int op,
                        void *data, u32 data_len,
                        void (*cb_func)(struct bnx2fc_els_cb_arg *cb_arg),
                        struct bnx2fc_els_cb_arg *cb_arg, u32 timer_msec);

static void bnx2fc_rrq_compl(struct bnx2fc_els_cb_arg *cb_arg)
{
        struct bnx2fc_cmd *orig_io_req;
        struct bnx2fc_cmd *rrq_req;
        int rc = 0;

        BUG_ON(!cb_arg);
        rrq_req = cb_arg->io_req;
        orig_io_req = cb_arg->aborted_io_req;
        BUG_ON(!orig_io_req);
        BNX2FC_ELS_DBG("rrq_compl: orig xid = 0x%x, rrq_xid = 0x%x\n",
                   orig_io_req->xid, rrq_req->xid);

        kref_put(&orig_io_req->refcount, bnx2fc_cmd_release);

        if (test_and_clear_bit(BNX2FC_FLAG_ELS_TIMEOUT, &rrq_req->req_flags)) {
                /*
                 * els req is timed out. cleanup the IO with FW and
                 * drop the completion. Remove from active_cmd_queue.
                 */
                BNX2FC_ELS_DBG("rrq xid - 0x%x timed out, clean it up\n",
                           rrq_req->xid);

                if (rrq_req->on_active_queue) {
                        list_del_init(&rrq_req->link);
                        rrq_req->on_active_queue = 0;
                        rc = bnx2fc_initiate_cleanup(rrq_req);
                        BUG_ON(rc);
                }
        }
        kfree(cb_arg);
}
int bnx2fc_send_rrq(struct bnx2fc_cmd *aborted_io_req)
{

        struct fc_els_rrq rrq;
        struct bnx2fc_rport *tgt = aborted_io_req->tgt;
        struct fc_lport *lport = NULL;
        struct bnx2fc_els_cb_arg *cb_arg = NULL;
        u32 sid = 0;
        u32 r_a_tov = 0;
        unsigned long start = jiffies;
        int rc;

        if (!test_bit(BNX2FC_FLAG_SESSION_READY, &tgt->flags))
                return -EINVAL;

        lport = tgt->rdata->local_port;
        sid = tgt->sid;
        r_a_tov = lport->r_a_tov;

        BNX2FC_ELS_DBG("Sending RRQ orig_xid = 0x%x\n",
                   aborted_io_req->xid);
        memset(&rrq, 0, sizeof(rrq));

        cb_arg = kzalloc_obj(struct bnx2fc_els_cb_arg, GFP_NOIO);
        if (!cb_arg) {
                printk(KERN_ERR PFX "Unable to allocate cb_arg for RRQ\n");
                rc = -ENOMEM;
                goto rrq_err;
        }

        cb_arg->aborted_io_req = aborted_io_req;

        rrq.rrq_cmd = ELS_RRQ;
        hton24(rrq.rrq_s_id, sid);
        rrq.rrq_ox_id = htons(aborted_io_req->xid);
        rrq.rrq_rx_id = htons(aborted_io_req->task->rxwr_txrd.var_ctx.rx_id);

retry_rrq:
        rc = bnx2fc_initiate_els(tgt, ELS_RRQ, &rrq, sizeof(rrq),
                                 bnx2fc_rrq_compl, cb_arg,
                                 r_a_tov);
        if (rc == -ENOMEM) {
                if (time_after(jiffies, start + (10 * HZ))) {
                        BNX2FC_ELS_DBG("rrq Failed\n");
                        rc = FAILED;
                        goto rrq_err;
                }
                msleep(20);
                goto retry_rrq;
        }
rrq_err:
        if (rc) {
                BNX2FC_ELS_DBG("RRQ failed - release orig io req 0x%x\n",
                        aborted_io_req->xid);
                kfree(cb_arg);
                spin_lock_bh(&tgt->tgt_lock);
                kref_put(&aborted_io_req->refcount, bnx2fc_cmd_release);
                spin_unlock_bh(&tgt->tgt_lock);
        }
        return rc;
}

static void bnx2fc_l2_els_compl(struct bnx2fc_els_cb_arg *cb_arg)
{
        struct bnx2fc_cmd *els_req;
        struct bnx2fc_rport *tgt;
        struct bnx2fc_mp_req *mp_req;
        struct fc_frame_header *fc_hdr;
        unsigned char *buf;
        void *resp_buf;
        u32 resp_len, hdr_len;
        u16 l2_oxid;
        int frame_len;
        int rc = 0;

        l2_oxid = cb_arg->l2_oxid;
        BNX2FC_ELS_DBG("ELS COMPL - l2_oxid = 0x%x\n", l2_oxid);

        els_req = cb_arg->io_req;
        if (test_and_clear_bit(BNX2FC_FLAG_ELS_TIMEOUT, &els_req->req_flags)) {
                /*
                 * els req is timed out. cleanup the IO with FW and
                 * drop the completion. libfc will handle the els timeout
                 */
                if (els_req->on_active_queue) {
                        list_del_init(&els_req->link);
                        els_req->on_active_queue = 0;
                        rc = bnx2fc_initiate_cleanup(els_req);
                        BUG_ON(rc);
                }
                goto free_arg;
        }

        tgt = els_req->tgt;
        mp_req = &(els_req->mp_req);
        fc_hdr = &(mp_req->resp_fc_hdr);
        resp_len = mp_req->resp_len;
        resp_buf = mp_req->resp_buf;

        buf = kzalloc(PAGE_SIZE, GFP_ATOMIC);
        if (!buf) {
                printk(KERN_ERR PFX "Unable to alloc mp buf\n");
                goto free_arg;
        }
        hdr_len = sizeof(*fc_hdr);
        if (hdr_len + resp_len > PAGE_SIZE) {
                printk(KERN_ERR PFX "l2_els_compl: resp len is "
                                    "beyond page size\n");
                goto free_buf;
        }
        memcpy(buf, fc_hdr, hdr_len);
        memcpy(buf + hdr_len, resp_buf, resp_len);
        frame_len = hdr_len + resp_len;

        bnx2fc_process_l2_frame_compl(tgt, buf, frame_len, l2_oxid);

free_buf:
        kfree(buf);
free_arg:
        kfree(cb_arg);
}

int bnx2fc_send_adisc(struct bnx2fc_rport *tgt, struct fc_frame *fp)
{
        struct fc_els_adisc *adisc;
        struct fc_frame_header *fh;
        struct bnx2fc_els_cb_arg *cb_arg;
        struct fc_lport *lport = tgt->rdata->local_port;
        u32 r_a_tov = lport->r_a_tov;
        int rc;

        fh = fc_frame_header_get(fp);
        cb_arg = kzalloc_obj(struct bnx2fc_els_cb_arg, GFP_ATOMIC);
        if (!cb_arg) {
                printk(KERN_ERR PFX "Unable to allocate cb_arg for ADISC\n");
                return -ENOMEM;
        }

        cb_arg->l2_oxid = ntohs(fh->fh_ox_id);

        BNX2FC_ELS_DBG("send ADISC: l2_oxid = 0x%x\n", cb_arg->l2_oxid);
        adisc = fc_frame_payload_get(fp, sizeof(*adisc));
        /* adisc is initialized by libfc */
        rc = bnx2fc_initiate_els(tgt, ELS_ADISC, adisc, sizeof(*adisc),
                                 bnx2fc_l2_els_compl, cb_arg, 2 * r_a_tov);
        if (rc)
                kfree(cb_arg);
        return rc;
}

int bnx2fc_send_logo(struct bnx2fc_rport *tgt, struct fc_frame *fp)
{
        struct fc_els_logo *logo;
        struct fc_frame_header *fh;
        struct bnx2fc_els_cb_arg *cb_arg;
        struct fc_lport *lport = tgt->rdata->local_port;
        u32 r_a_tov = lport->r_a_tov;
        int rc;

        fh = fc_frame_header_get(fp);
        cb_arg = kzalloc_obj(struct bnx2fc_els_cb_arg, GFP_ATOMIC);
        if (!cb_arg) {
                printk(KERN_ERR PFX "Unable to allocate cb_arg for LOGO\n");
                return -ENOMEM;
        }

        cb_arg->l2_oxid = ntohs(fh->fh_ox_id);

        BNX2FC_ELS_DBG("Send LOGO: l2_oxid = 0x%x\n", cb_arg->l2_oxid);
        logo = fc_frame_payload_get(fp, sizeof(*logo));
        /* logo is initialized by libfc */
        rc = bnx2fc_initiate_els(tgt, ELS_LOGO, logo, sizeof(*logo),
                                 bnx2fc_l2_els_compl, cb_arg, 2 * r_a_tov);
        if (rc)
                kfree(cb_arg);
        return rc;
}

int bnx2fc_send_rls(struct bnx2fc_rport *tgt, struct fc_frame *fp)
{
        struct fc_els_rls *rls;
        struct fc_frame_header *fh;
        struct bnx2fc_els_cb_arg *cb_arg;
        struct fc_lport *lport = tgt->rdata->local_port;
        u32 r_a_tov = lport->r_a_tov;
        int rc;

        fh = fc_frame_header_get(fp);
        cb_arg = kzalloc_obj(struct bnx2fc_els_cb_arg, GFP_ATOMIC);
        if (!cb_arg) {
                printk(KERN_ERR PFX "Unable to allocate cb_arg for LOGO\n");
                return -ENOMEM;
        }

        cb_arg->l2_oxid = ntohs(fh->fh_ox_id);

        rls = fc_frame_payload_get(fp, sizeof(*rls));
        /* rls is initialized by libfc */
        rc = bnx2fc_initiate_els(tgt, ELS_RLS, rls, sizeof(*rls),
                                  bnx2fc_l2_els_compl, cb_arg, 2 * r_a_tov);
        if (rc)
                kfree(cb_arg);
        return rc;
}

static void bnx2fc_srr_compl(struct bnx2fc_els_cb_arg *cb_arg)
{
        struct bnx2fc_mp_req *mp_req;
        struct fc_frame_header *fc_hdr, *fh;
        struct bnx2fc_cmd *srr_req;
        struct bnx2fc_cmd *orig_io_req;
        struct fc_frame *fp;
        unsigned char *buf;
        void *resp_buf;
        u32 resp_len, hdr_len;
        u8 opcode;
        int rc = 0;

        orig_io_req = cb_arg->aborted_io_req;
        srr_req = cb_arg->io_req;
        if (test_and_clear_bit(BNX2FC_FLAG_ELS_TIMEOUT, &srr_req->req_flags)) {
                /* SRR timedout */
                BNX2FC_IO_DBG(srr_req, "srr timed out, abort "
                       "orig_io - 0x%x\n",
                        orig_io_req->xid);
                rc = bnx2fc_initiate_abts(srr_req);
                if (rc != SUCCESS) {
                        BNX2FC_IO_DBG(srr_req, "srr_compl: initiate_abts "
                                "failed. issue cleanup\n");
                        bnx2fc_initiate_cleanup(srr_req);
                }
                if (test_bit(BNX2FC_FLAG_IO_COMPL, &orig_io_req->req_flags) ||
                    test_bit(BNX2FC_FLAG_ISSUE_ABTS, &orig_io_req->req_flags)) {
                        BNX2FC_IO_DBG(srr_req, "srr_compl:xid 0x%x flags = %lx",
                                      orig_io_req->xid, orig_io_req->req_flags);
                        goto srr_compl_done;
                }
                orig_io_req->srr_retry++;
                if (orig_io_req->srr_retry <= SRR_RETRY_COUNT) {
                        struct bnx2fc_rport *tgt = orig_io_req->tgt;
                        spin_unlock_bh(&tgt->tgt_lock);
                        rc = bnx2fc_send_srr(orig_io_req,
                                             orig_io_req->srr_offset,
                                             orig_io_req->srr_rctl);
                        spin_lock_bh(&tgt->tgt_lock);
                        if (!rc)
                                goto srr_compl_done;
                }

                rc = bnx2fc_initiate_abts(orig_io_req);
                if (rc != SUCCESS) {
                        BNX2FC_IO_DBG(srr_req, "srr_compl: initiate_abts "
                                "failed xid = 0x%x. issue cleanup\n",
                                orig_io_req->xid);
                        bnx2fc_initiate_cleanup(orig_io_req);
                }
                goto srr_compl_done;
        }
        if (test_bit(BNX2FC_FLAG_IO_COMPL, &orig_io_req->req_flags) ||
            test_bit(BNX2FC_FLAG_ISSUE_ABTS, &orig_io_req->req_flags)) {
                BNX2FC_IO_DBG(srr_req, "srr_compl:xid - 0x%x flags = %lx",
                              orig_io_req->xid, orig_io_req->req_flags);
                goto srr_compl_done;
        }
        mp_req = &(srr_req->mp_req);
        fc_hdr = &(mp_req->resp_fc_hdr);
        resp_len = mp_req->resp_len;
        resp_buf = mp_req->resp_buf;

        hdr_len = sizeof(*fc_hdr);
        buf = kzalloc(PAGE_SIZE, GFP_ATOMIC);
        if (!buf) {
                printk(KERN_ERR PFX "srr buf: mem alloc failure\n");
                goto srr_compl_done;
        }
        memcpy(buf, fc_hdr, hdr_len);
        memcpy(buf + hdr_len, resp_buf, resp_len);

        fp = fc_frame_alloc(NULL, resp_len);
        if (!fp) {
                printk(KERN_ERR PFX "fc_frame_alloc failure\n");
                goto free_buf;
        }

        fh = (struct fc_frame_header *) fc_frame_header_get(fp);
        /* Copy FC Frame header and payload into the frame */
        memcpy(fh, buf, hdr_len + resp_len);

        opcode = fc_frame_payload_op(fp);
        switch (opcode) {
        case ELS_LS_ACC:
                BNX2FC_IO_DBG(srr_req, "SRR success\n");
                break;
        case ELS_LS_RJT:
                BNX2FC_IO_DBG(srr_req, "SRR rejected\n");
                rc = bnx2fc_initiate_abts(orig_io_req);
                if (rc != SUCCESS) {
                        BNX2FC_IO_DBG(srr_req, "srr_compl: initiate_abts "
                                "failed xid = 0x%x. issue cleanup\n",
                                orig_io_req->xid);
                        bnx2fc_initiate_cleanup(orig_io_req);
                }
                break;
        default:
                BNX2FC_IO_DBG(srr_req, "srr compl - invalid opcode = %d\n",
                        opcode);
                break;
        }
        fc_frame_free(fp);
free_buf:
        kfree(buf);
srr_compl_done:
        kref_put(&orig_io_req->refcount, bnx2fc_cmd_release);
}

static void bnx2fc_rec_compl(struct bnx2fc_els_cb_arg *cb_arg)
{
        struct bnx2fc_cmd *orig_io_req, *new_io_req;
        struct bnx2fc_cmd *rec_req;
        struct bnx2fc_mp_req *mp_req;
        struct fc_frame_header *fc_hdr, *fh;
        struct fc_els_ls_rjt *rjt;
        struct fc_els_rec_acc *acc;
        struct bnx2fc_rport *tgt;
        struct fcoe_err_report_entry *err_entry;
        struct scsi_cmnd *sc_cmd;
        enum fc_rctl r_ctl;
        unsigned char *buf;
        void *resp_buf;
        struct fc_frame *fp;
        u8 opcode;
        u32 offset;
        u32 e_stat;
        u32 resp_len, hdr_len;
        int rc = 0;
        bool send_seq_clnp = false;
        bool abort_io = false;

        BNX2FC_MISC_DBG("Entered rec_compl callback\n");
        rec_req = cb_arg->io_req;
        orig_io_req = cb_arg->aborted_io_req;
        BNX2FC_IO_DBG(rec_req, "rec_compl: orig xid = 0x%x", orig_io_req->xid);
        tgt = orig_io_req->tgt;

        /* Handle REC timeout case */
        if (test_and_clear_bit(BNX2FC_FLAG_ELS_TIMEOUT, &rec_req->req_flags)) {
                BNX2FC_IO_DBG(rec_req, "timed out, abort "
                       "orig_io - 0x%x\n",
                        orig_io_req->xid);
                /* els req is timed out. send abts for els */
                rc = bnx2fc_initiate_abts(rec_req);
                if (rc != SUCCESS) {
                        BNX2FC_IO_DBG(rec_req, "rec_compl: initiate_abts "
                                "failed. issue cleanup\n");
                        bnx2fc_initiate_cleanup(rec_req);
                }
                orig_io_req->rec_retry++;
                /* REC timedout. send ABTS to the orig IO req */
                if (orig_io_req->rec_retry <= REC_RETRY_COUNT) {
                        spin_unlock_bh(&tgt->tgt_lock);
                        rc = bnx2fc_send_rec(orig_io_req);
                        spin_lock_bh(&tgt->tgt_lock);
                        if (!rc)
                                goto rec_compl_done;
                }
                rc = bnx2fc_initiate_abts(orig_io_req);
                if (rc != SUCCESS) {
                        BNX2FC_IO_DBG(rec_req, "rec_compl: initiate_abts "
                                "failed xid = 0x%x. issue cleanup\n",
                                orig_io_req->xid);
                        bnx2fc_initiate_cleanup(orig_io_req);
                }
                goto rec_compl_done;
        }

        if (test_bit(BNX2FC_FLAG_IO_COMPL, &orig_io_req->req_flags)) {
                BNX2FC_IO_DBG(rec_req, "completed"
                       "orig_io - 0x%x\n",
                        orig_io_req->xid);
                goto rec_compl_done;
        }
        if (test_bit(BNX2FC_FLAG_ISSUE_ABTS, &orig_io_req->req_flags)) {
                BNX2FC_IO_DBG(rec_req, "abts in prog "
                       "orig_io - 0x%x\n",
                        orig_io_req->xid);
                goto rec_compl_done;
        }

        mp_req = &(rec_req->mp_req);
        fc_hdr = &(mp_req->resp_fc_hdr);
        resp_len = mp_req->resp_len;
        acc = resp_buf = mp_req->resp_buf;

        hdr_len = sizeof(*fc_hdr);

        buf = kzalloc(PAGE_SIZE, GFP_ATOMIC);
        if (!buf) {
                printk(KERN_ERR PFX "rec buf: mem alloc failure\n");
                goto rec_compl_done;
        }
        memcpy(buf, fc_hdr, hdr_len);
        memcpy(buf + hdr_len, resp_buf, resp_len);

        fp = fc_frame_alloc(NULL, resp_len);
        if (!fp) {
                printk(KERN_ERR PFX "fc_frame_alloc failure\n");
                goto free_buf;
        }

        fh = (struct fc_frame_header *) fc_frame_header_get(fp);
        /* Copy FC Frame header and payload into the frame */
        memcpy(fh, buf, hdr_len + resp_len);

        opcode = fc_frame_payload_op(fp);
        if (opcode == ELS_LS_RJT) {
                BNX2FC_IO_DBG(rec_req, "opcode is RJT\n");
                rjt = fc_frame_payload_get(fp, sizeof(*rjt));
                if ((rjt->er_reason == ELS_RJT_LOGIC ||
                    rjt->er_reason == ELS_RJT_UNAB) &&
                    rjt->er_explan == ELS_EXPL_OXID_RXID) {
                        BNX2FC_IO_DBG(rec_req, "handle CMD LOST case\n");
                        new_io_req = bnx2fc_cmd_alloc(tgt);
                        if (!new_io_req)
                                goto abort_io;
                        new_io_req->sc_cmd = orig_io_req->sc_cmd;
                        /* cleanup orig_io_req that is with the FW */
                        set_bit(BNX2FC_FLAG_CMD_LOST,
                                &orig_io_req->req_flags);
                        bnx2fc_initiate_cleanup(orig_io_req);
                        /* Post a new IO req with the same sc_cmd */
                        BNX2FC_IO_DBG(rec_req, "Post IO request again\n");
                        rc = bnx2fc_post_io_req(tgt, new_io_req);
                        if (!rc)
                                goto free_frame;
                        BNX2FC_IO_DBG(rec_req, "REC: io post err\n");
                }
abort_io:
                rc = bnx2fc_initiate_abts(orig_io_req);
                if (rc != SUCCESS) {
                        BNX2FC_IO_DBG(rec_req, "rec_compl: initiate_abts "
                                "failed. issue cleanup\n");
                        bnx2fc_initiate_cleanup(orig_io_req);
                }
        } else if (opcode == ELS_LS_ACC) {
                /* REVISIT: Check if the exchange is already aborted */
                offset = ntohl(acc->reca_fc4value);
                e_stat = ntohl(acc->reca_e_stat);
                if (e_stat & ESB_ST_SEQ_INIT)  {
                        BNX2FC_IO_DBG(rec_req, "target has the seq init\n");
                        goto free_frame;
                }
                BNX2FC_IO_DBG(rec_req, "e_stat = 0x%x, offset = 0x%x\n",
                        e_stat, offset);
                /* Seq initiative is with us */
                err_entry = (struct fcoe_err_report_entry *)
                             &orig_io_req->err_entry;
                sc_cmd = orig_io_req->sc_cmd;
                if (sc_cmd->sc_data_direction == DMA_TO_DEVICE) {
                        /* SCSI WRITE command */
                        if (offset == orig_io_req->data_xfer_len) {
                                BNX2FC_IO_DBG(rec_req, "WRITE - resp lost\n");
                                /* FCP_RSP lost */
                                r_ctl = FC_RCTL_DD_CMD_STATUS;
                                offset = 0;
                        } else  {
                                /* start transmitting from offset */
                                BNX2FC_IO_DBG(rec_req, "XFER_RDY/DATA lost\n");
                                send_seq_clnp = true;
                                r_ctl = FC_RCTL_DD_DATA_DESC;
                                if (bnx2fc_initiate_seq_cleanup(orig_io_req,
                                                                offset, r_ctl))
                                        abort_io = true;
                                /* XFER_RDY */
                        }
                } else {
                        /* SCSI READ command */
                        if (err_entry->data.rx_buf_off ==
                                        orig_io_req->data_xfer_len) {
                                /* FCP_RSP lost */
                                BNX2FC_IO_DBG(rec_req, "READ - resp lost\n");
                                r_ctl = FC_RCTL_DD_CMD_STATUS;
                                offset = 0;
                        } else  {
                                /* request retransmission from this offset */
                                send_seq_clnp = true;
                                offset = err_entry->data.rx_buf_off;
                                BNX2FC_IO_DBG(rec_req, "RD DATA lost\n");
                                /* FCP_DATA lost */
                                r_ctl = FC_RCTL_DD_SOL_DATA;
                                if (bnx2fc_initiate_seq_cleanup(orig_io_req,
                                                                offset, r_ctl))
                                        abort_io = true;
                        }
                }
                if (abort_io) {
                        rc = bnx2fc_initiate_abts(orig_io_req);
                        if (rc != SUCCESS) {
                                BNX2FC_IO_DBG(rec_req, "rec_compl:initiate_abts"
                                              " failed. issue cleanup\n");
                                bnx2fc_initiate_cleanup(orig_io_req);
                        }
                } else if (!send_seq_clnp) {
                        BNX2FC_IO_DBG(rec_req, "Send SRR - FCP_RSP\n");
                        spin_unlock_bh(&tgt->tgt_lock);
                        rc = bnx2fc_send_srr(orig_io_req, offset, r_ctl);
                        spin_lock_bh(&tgt->tgt_lock);

                        if (rc) {
                                BNX2FC_IO_DBG(rec_req, "Unable to send SRR"
                                        " IO will abort\n");
                        }
                }
        }
free_frame:
        fc_frame_free(fp);
free_buf:
        kfree(buf);
rec_compl_done:
        kref_put(&orig_io_req->refcount, bnx2fc_cmd_release);
        kfree(cb_arg);
}

int bnx2fc_send_rec(struct bnx2fc_cmd *orig_io_req)
{
        struct fc_els_rec rec;
        struct bnx2fc_rport *tgt = orig_io_req->tgt;
        struct fc_lport *lport = tgt->rdata->local_port;
        struct bnx2fc_els_cb_arg *cb_arg = NULL;
        u32 sid = tgt->sid;
        u32 r_a_tov = lport->r_a_tov;
        int rc;

        BNX2FC_IO_DBG(orig_io_req, "Sending REC\n");
        memset(&rec, 0, sizeof(rec));

        cb_arg = kzalloc_obj(struct bnx2fc_els_cb_arg, GFP_ATOMIC);
        if (!cb_arg) {
                printk(KERN_ERR PFX "Unable to allocate cb_arg for REC\n");
                rc = -ENOMEM;
                goto rec_err;
        }
        kref_get(&orig_io_req->refcount);

        cb_arg->aborted_io_req = orig_io_req;

        rec.rec_cmd = ELS_REC;
        hton24(rec.rec_s_id, sid);
        rec.rec_ox_id = htons(orig_io_req->xid);
        rec.rec_rx_id = htons(orig_io_req->task->rxwr_txrd.var_ctx.rx_id);

        rc = bnx2fc_initiate_els(tgt, ELS_REC, &rec, sizeof(rec),
                                 bnx2fc_rec_compl, cb_arg,
                                 r_a_tov);
        if (rc) {
                BNX2FC_IO_DBG(orig_io_req, "REC failed - release\n");
                spin_lock_bh(&tgt->tgt_lock);
                kref_put(&orig_io_req->refcount, bnx2fc_cmd_release);
                spin_unlock_bh(&tgt->tgt_lock);
                kfree(cb_arg);
        }
rec_err:
        return rc;
}

int bnx2fc_send_srr(struct bnx2fc_cmd *orig_io_req, u32 offset, u8 r_ctl)
{
        struct fcp_srr srr;
        struct bnx2fc_rport *tgt = orig_io_req->tgt;
        struct fc_lport *lport = tgt->rdata->local_port;
        struct bnx2fc_els_cb_arg *cb_arg = NULL;
        u32 r_a_tov = lport->r_a_tov;
        int rc;

        BNX2FC_IO_DBG(orig_io_req, "Sending SRR\n");
        memset(&srr, 0, sizeof(srr));

        cb_arg = kzalloc_obj(struct bnx2fc_els_cb_arg, GFP_ATOMIC);
        if (!cb_arg) {
                printk(KERN_ERR PFX "Unable to allocate cb_arg for SRR\n");
                rc = -ENOMEM;
                goto srr_err;
        }
        kref_get(&orig_io_req->refcount);

        cb_arg->aborted_io_req = orig_io_req;

        srr.srr_op = ELS_SRR;
        srr.srr_ox_id = htons(orig_io_req->xid);
        srr.srr_rx_id = htons(orig_io_req->task->rxwr_txrd.var_ctx.rx_id);
        srr.srr_rel_off = htonl(offset);
        srr.srr_r_ctl = r_ctl;
        orig_io_req->srr_offset = offset;
        orig_io_req->srr_rctl = r_ctl;

        rc = bnx2fc_initiate_els(tgt, ELS_SRR, &srr, sizeof(srr),
                                 bnx2fc_srr_compl, cb_arg,
                                 r_a_tov);
        if (rc) {
                BNX2FC_IO_DBG(orig_io_req, "SRR failed - release\n");
                spin_lock_bh(&tgt->tgt_lock);
                kref_put(&orig_io_req->refcount, bnx2fc_cmd_release);
                spin_unlock_bh(&tgt->tgt_lock);
                kfree(cb_arg);
        } else
                set_bit(BNX2FC_FLAG_SRR_SENT, &orig_io_req->req_flags);

srr_err:
        return rc;
}

static int bnx2fc_initiate_els(struct bnx2fc_rport *tgt, unsigned int op,
                        void *data, u32 data_len,
                        void (*cb_func)(struct bnx2fc_els_cb_arg *cb_arg),
                        struct bnx2fc_els_cb_arg *cb_arg, u32 timer_msec)
{
        struct fcoe_port *port = tgt->port;
        struct bnx2fc_interface *interface = port->priv;
        struct fc_rport *rport = tgt->rport;
        struct fc_lport *lport = port->lport;
        struct bnx2fc_cmd *els_req;
        struct bnx2fc_mp_req *mp_req;
        struct fc_frame_header *fc_hdr;
        struct fcoe_task_ctx_entry *task;
        struct fcoe_task_ctx_entry *task_page;
        int rc = 0;
        int task_idx, index;
        u32 did, sid;
        u16 xid;

        rc = fc_remote_port_chkready(rport);
        if (rc) {
                printk(KERN_ERR PFX "els 0x%x: rport not ready\n", op);
                rc = -EINVAL;
                goto els_err;
        }
        if (lport->state != LPORT_ST_READY || !(lport->link_up)) {
                printk(KERN_ERR PFX "els 0x%x: link is not ready\n", op);
                rc = -EINVAL;
                goto els_err;
        }
        if (!(test_bit(BNX2FC_FLAG_SESSION_READY, &tgt->flags))) {
                printk(KERN_ERR PFX "els 0x%x: tgt not ready\n", op);
                rc = -EINVAL;
                goto els_err;
        }
        els_req = bnx2fc_elstm_alloc(tgt, BNX2FC_ELS);
        if (!els_req) {
                rc = -ENOMEM;
                goto els_err;
        }

        els_req->sc_cmd = NULL;
        els_req->port = port;
        els_req->tgt = tgt;
        els_req->cb_func = cb_func;
        cb_arg->io_req = els_req;
        els_req->cb_arg = cb_arg;
        els_req->data_xfer_len = data_len;

        mp_req = (struct bnx2fc_mp_req *)&(els_req->mp_req);
        rc = bnx2fc_init_mp_req(els_req);
        if (rc == FAILED) {
                printk(KERN_ERR PFX "ELS MP request init failed\n");
                spin_lock_bh(&tgt->tgt_lock);
                kref_put(&els_req->refcount, bnx2fc_cmd_release);
                spin_unlock_bh(&tgt->tgt_lock);
                rc = -ENOMEM;
                goto els_err;
        } else {
                /* rc SUCCESS */
                rc = 0;
        }

        /* Set the data_xfer_len to the size of ELS payload */
        mp_req->req_len = data_len;
        els_req->data_xfer_len = mp_req->req_len;

        /* Fill ELS Payload */
        if ((op >= ELS_LS_RJT) && (op <= ELS_AUTH_ELS)) {
                memcpy(mp_req->req_buf, data, data_len);
        } else {
                printk(KERN_ERR PFX "Invalid ELS op 0x%x\n", op);
                els_req->cb_func = NULL;
                els_req->cb_arg = NULL;
                spin_lock_bh(&tgt->tgt_lock);
                kref_put(&els_req->refcount, bnx2fc_cmd_release);
                spin_unlock_bh(&tgt->tgt_lock);
                rc = -EINVAL;
        }

        if (rc)
                goto els_err;

        /* Fill FC header */
        fc_hdr = &(mp_req->req_fc_hdr);

        did = tgt->rport->port_id;
        sid = tgt->sid;

        if (op == ELS_SRR)
                __fc_fill_fc_hdr(fc_hdr, FC_RCTL_ELS4_REQ, did, sid,
                                   FC_TYPE_FCP, FC_FC_FIRST_SEQ |
                                   FC_FC_END_SEQ | FC_FC_SEQ_INIT, 0);
        else
                __fc_fill_fc_hdr(fc_hdr, FC_RCTL_ELS_REQ, did, sid,
                                   FC_TYPE_ELS, FC_FC_FIRST_SEQ |
                                   FC_FC_END_SEQ | FC_FC_SEQ_INIT, 0);

        /* Obtain exchange id */
        xid = els_req->xid;
        task_idx = xid/BNX2FC_TASKS_PER_PAGE;
        index = xid % BNX2FC_TASKS_PER_PAGE;

        /* Initialize task context for this IO request */
        task_page = (struct fcoe_task_ctx_entry *)
                        interface->hba->task_ctx[task_idx];
        task = &(task_page[index]);
        bnx2fc_init_mp_task(els_req, task);

        spin_lock_bh(&tgt->tgt_lock);

        if (!test_bit(BNX2FC_FLAG_SESSION_READY, &tgt->flags)) {
                printk(KERN_ERR PFX "initiate_els.. session not ready\n");
                els_req->cb_func = NULL;
                els_req->cb_arg = NULL;
                kref_put(&els_req->refcount, bnx2fc_cmd_release);
                spin_unlock_bh(&tgt->tgt_lock);
                return -EINVAL;
        }

        if (timer_msec)
                bnx2fc_cmd_timer_set(els_req, timer_msec);
        bnx2fc_add_2_sq(tgt, xid);

        els_req->on_active_queue = 1;
        list_add_tail(&els_req->link, &tgt->els_queue);

        /* Ring doorbell */
        bnx2fc_ring_doorbell(tgt);
        spin_unlock_bh(&tgt->tgt_lock);

els_err:
        return rc;
}

void bnx2fc_process_els_compl(struct bnx2fc_cmd *els_req,
                              struct fcoe_task_ctx_entry *task, u8 num_rq)
{
        struct bnx2fc_mp_req *mp_req;
        struct fc_frame_header *fc_hdr;
        u64 *hdr;
        u64 *temp_hdr;

        BNX2FC_ELS_DBG("Entered process_els_compl xid = 0x%x"
                        "cmd_type = %d\n", els_req->xid, els_req->cmd_type);

        if (test_and_set_bit(BNX2FC_FLAG_ELS_DONE,
                             &els_req->req_flags)) {
                BNX2FC_ELS_DBG("Timer context finished processing this "
                           "els - 0x%x\n", els_req->xid);
                /* This IO doesn't receive cleanup completion */
                kref_put(&els_req->refcount, bnx2fc_cmd_release);
                return;
        }

        /* Cancel the timeout_work, as we received the response */
        if (cancel_delayed_work(&els_req->timeout_work))
                kref_put(&els_req->refcount,
                         bnx2fc_cmd_release); /* drop timer hold */

        if (els_req->on_active_queue) {
                list_del_init(&els_req->link);
                els_req->on_active_queue = 0;
        }

        mp_req = &(els_req->mp_req);
        fc_hdr = &(mp_req->resp_fc_hdr);

        hdr = (u64 *)fc_hdr;
        temp_hdr = (u64 *)
                &task->rxwr_only.union_ctx.comp_info.mp_rsp.fc_hdr;
        hdr[0] = cpu_to_be64(temp_hdr[0]);
        hdr[1] = cpu_to_be64(temp_hdr[1]);
        hdr[2] = cpu_to_be64(temp_hdr[2]);

        mp_req->resp_len =
                task->rxwr_only.union_ctx.comp_info.mp_rsp.mp_payload_len;

        /* Parse ELS response */
        if ((els_req->cb_func) && (els_req->cb_arg)) {
                els_req->cb_func(els_req->cb_arg);
                els_req->cb_arg = NULL;
        }

        kref_put(&els_req->refcount, bnx2fc_cmd_release);
}

#define         BNX2FC_FCOE_MAC_METHOD_GRANGED_MAC      1
#define         BNX2FC_FCOE_MAC_METHOD_FCF_MAP          2
#define         BNX2FC_FCOE_MAC_METHOD_FCOE_SET_MAC     3
static void bnx2fc_flogi_resp(struct fc_seq *seq, struct fc_frame *fp,
                              void *arg)
{
        struct fcoe_ctlr *fip = arg;
        struct fc_exch *exch = fc_seq_exch(seq);
        struct fc_lport *lport = exch->lp;

        struct fc_frame_header *fh;
        u8 *granted_mac;
        u8 fcoe_mac[6];
        u8 fc_map[3];
        int method;

        if (IS_ERR(fp))
                goto done;

        fh = fc_frame_header_get(fp);
        granted_mac = fr_cb(fp)->granted_mac;

        /*
         * We set the source MAC for FCoE traffic based on the Granted MAC
         * address from the switch.
         *
         * If granted_mac is non-zero, we use that.
         * If the granted_mac is zeroed out, create the FCoE MAC based on
         * the sel_fcf->fc_map and the d_id fo the FLOGI frame.
         * If sel_fcf->fc_map is 0, then we use the default FCF-MAC plus the
         * d_id of the FLOGI frame.
         */
        if (!is_zero_ether_addr(granted_mac)) {
                ether_addr_copy(fcoe_mac, granted_mac);
                method = BNX2FC_FCOE_MAC_METHOD_GRANGED_MAC;
        } else if (fip->sel_fcf && fip->sel_fcf->fc_map != 0) {
                hton24(fc_map, fip->sel_fcf->fc_map);
                fcoe_mac[0] = fc_map[0];
                fcoe_mac[1] = fc_map[1];
                fcoe_mac[2] = fc_map[2];
                fcoe_mac[3] = fh->fh_d_id[0];
                fcoe_mac[4] = fh->fh_d_id[1];
                fcoe_mac[5] = fh->fh_d_id[2];
                method = BNX2FC_FCOE_MAC_METHOD_FCF_MAP;
        } else {
                fc_fcoe_set_mac(fcoe_mac, fh->fh_d_id);
                method = BNX2FC_FCOE_MAC_METHOD_FCOE_SET_MAC;
        }

        BNX2FC_HBA_DBG(lport, "fcoe_mac=%pM method=%d\n", fcoe_mac, method);
        fip->update_mac(lport, fcoe_mac);
done:
        fc_lport_flogi_resp(seq, fp, lport);
}

static void bnx2fc_logo_resp(struct fc_seq *seq, struct fc_frame *fp,
                             void *arg)
{
        struct fcoe_ctlr *fip = arg;
        struct fc_exch *exch = fc_seq_exch(seq);
        struct fc_lport *lport = exch->lp;
        static u8 zero_mac[ETH_ALEN] = { 0 };

        if (!IS_ERR(fp))
                fip->update_mac(lport, zero_mac);
        fc_lport_logo_resp(seq, fp, lport);
}

struct fc_seq *bnx2fc_elsct_send(struct fc_lport *lport, u32 did,
                                      struct fc_frame *fp, unsigned int op,
                                      void (*resp)(struct fc_seq *,
                                                   struct fc_frame *,
                                                   void *),
                                      void *arg, u32 timeout)
{
        struct fcoe_port *port = lport_priv(lport);
        struct bnx2fc_interface *interface = port->priv;
        struct fcoe_ctlr *fip = bnx2fc_to_ctlr(interface);
        struct fc_frame_header *fh = fc_frame_header_get(fp);

        switch (op) {
        case ELS_FLOGI:
        case ELS_FDISC:
                return fc_elsct_send(lport, did, fp, op, bnx2fc_flogi_resp,
                                     fip, timeout);
        case ELS_LOGO:
                /* only hook onto fabric logouts, not port logouts */
                if (ntoh24(fh->fh_d_id) != FC_FID_FLOGI)
                        break;
                return fc_elsct_send(lport, did, fp, op, bnx2fc_logo_resp,
                                     fip, timeout);
        }
        return fc_elsct_send(lport, did, fp, op, resp, arg, timeout);
}