root/sys/dev/cxgbe/crypto/t6_kern_tls.c
/*-
 * SPDX-License-Identifier: BSD-2-Clause
 *
 * Copyright (c) 2018-2019 Chelsio Communications, Inc.
 * All rights reserved.
 * Written by: John Baldwin <jhb@FreeBSD.org>
 *
 * Redistribution and use in source and binary forms, with or without
 * modification, are permitted provided that the following conditions
 * are met:
 * 1. Redistributions of source code must retain the above copyright
 *    notice, this list of conditions and the following disclaimer.
 * 2. Redistributions in binary form must reproduce the above copyright
 *    notice, this list of conditions and the following disclaimer in the
 *    documentation and/or other materials provided with the distribution.
 *
 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
 * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
 * SUCH DAMAGE.
 */

#include "opt_inet.h"
#include "opt_inet6.h"
#include "opt_kern_tls.h"

#include <sys/param.h>
#include <sys/ktr.h>
#include <sys/ktls.h>
#include <sys/sglist.h>
#include <sys/socket.h>
#include <sys/socketvar.h>
#include <sys/sockbuf.h>
#include <netinet/in.h>
#include <netinet/in_pcb.h>
#include <netinet/ip.h>
#include <netinet/ip6.h>
#include <netinet/tcp_var.h>
#include <opencrypto/cryptodev.h>
#include <opencrypto/xform.h>

#include "common/common.h"
#include "common/t4_regs.h"
#include "common/t4_regs_values.h"
#include "common/t4_tcb.h"
#include "t4_l2t.h"
#include "t4_clip.h"
#include "t4_mp_ring.h"
#include "crypto/t4_crypto.h"

#if defined(INET) || defined(INET6)

#define TLS_HEADER_LENGTH               5

struct tls_scmd {
        __be32 seqno_numivs;
        __be32 ivgen_hdrlen;
};

struct tlspcb {
        struct m_snd_tag com;
        struct vi_info *vi;     /* virtual interface */
        struct adapter *sc;
        struct l2t_entry *l2te; /* L2 table entry used by this connection */
        struct sge_txq *txq;
        int tid;                /* Connection identifier */

        int tx_key_addr;
        bool inline_key;
        bool using_timestamps;
        unsigned char enc_mode;

        struct tls_scmd scmd0;
        struct tls_scmd scmd0_short;

        unsigned int tx_key_info_size;

        uint32_t prev_seq;
        uint32_t prev_ack;
        uint32_t prev_tsecr;
        uint16_t prev_win;
        uint16_t prev_mss;

        /* Only used outside of setup and teardown when using inline keys. */
        struct tls_keyctx keyctx;

        /* Fields only used during setup and teardown. */
        struct inpcb *inp;      /* backpointer to host stack's PCB */
        struct sge_wrq *ctrlq;
        struct clip_entry *ce;  /* CLIP table entry used by this tid */

        bool open_pending;
};

static void t6_tls_tag_free(struct m_snd_tag *mst);
static int ktls_setup_keys(struct tlspcb *tlsp,
    const struct ktls_session *tls, struct sge_txq *txq);

static const struct if_snd_tag_sw t6_tls_tag_sw = {
        .snd_tag_free = t6_tls_tag_free,
        .type = IF_SND_TAG_TYPE_TLS
};

static inline struct tlspcb *
mst_to_tls(struct m_snd_tag *t)
{
        return (__containerof(t, struct tlspcb, com));
}

static struct tlspcb *
alloc_tlspcb(if_t ifp, struct vi_info *vi, int flags)
{
        struct port_info *pi = vi->pi;
        struct adapter *sc = pi->adapter;
        struct tlspcb *tlsp;

        tlsp = malloc(sizeof(*tlsp), M_CXGBE, M_ZERO | flags);
        if (tlsp == NULL)
                return (NULL);

        m_snd_tag_init(&tlsp->com, ifp, &t6_tls_tag_sw);
        tlsp->vi = vi;
        tlsp->sc = sc;
        tlsp->ctrlq = &sc->sge.ctrlq[pi->port_id];
        tlsp->tid = -1;
        tlsp->tx_key_addr = -1;

        return (tlsp);
}

static int
ktls_act_open_cpl_size(bool isipv6)
{

        if (isipv6)
                return (sizeof(struct cpl_t6_act_open_req6));
        else
                return (sizeof(struct cpl_t6_act_open_req));
}

static void
mk_ktls_act_open_req(struct adapter *sc, struct vi_info *vi, struct inpcb *inp,
    struct tlspcb *tlsp, int atid, void *dst)
{
        struct tcpcb *tp = intotcpcb(inp);
        struct cpl_t6_act_open_req *cpl6;
        struct cpl_act_open_req *cpl;
        uint64_t options;
        int qid_atid;

        cpl6 = dst;
        cpl = (struct cpl_act_open_req *)cpl6;
        INIT_TP_WR(cpl6, 0);
        qid_atid = V_TID_QID(sc->sge.fwq.abs_id) | V_TID_TID(atid) |
            V_TID_COOKIE(CPL_COOKIE_KERN_TLS);
        OPCODE_TID(cpl) = htobe32(MK_OPCODE_TID(CPL_ACT_OPEN_REQ,
                qid_atid));
        inp_4tuple_get(inp, &cpl->local_ip, &cpl->local_port,
            &cpl->peer_ip, &cpl->peer_port);

        options = F_TCAM_BYPASS | V_ULP_MODE(ULP_MODE_NONE);
        options |= V_SMAC_SEL(vi->smt_idx) | V_TX_CHAN(vi->pi->tx_chan);
        options |= F_NON_OFFLOAD;
        cpl->opt0 = htobe64(options);

        options = V_TX_QUEUE(TX_MODQ(vi->pi->tx_chan));
        if (tp->t_flags & TF_REQ_TSTMP)
                options |= F_TSTAMPS_EN;
        cpl->opt2 = htobe32(options);
}

static void
mk_ktls_act_open_req6(struct adapter *sc, struct vi_info *vi,
    struct inpcb *inp, struct tlspcb *tlsp, int atid, void *dst)
{
        struct tcpcb *tp = intotcpcb(inp);
        struct cpl_t6_act_open_req6 *cpl6;
        struct cpl_act_open_req6 *cpl;
        uint64_t options;
        int qid_atid;

        cpl6 = dst;
        cpl = (struct cpl_act_open_req6 *)cpl6;
        INIT_TP_WR(cpl6, 0);
        qid_atid = V_TID_QID(sc->sge.fwq.abs_id) | V_TID_TID(atid) |
            V_TID_COOKIE(CPL_COOKIE_KERN_TLS);
        OPCODE_TID(cpl) = htobe32(MK_OPCODE_TID(CPL_ACT_OPEN_REQ6,
                qid_atid));
        cpl->local_port = inp->inp_lport;
        cpl->local_ip_hi = *(uint64_t *)&inp->in6p_laddr.s6_addr[0];
        cpl->local_ip_lo = *(uint64_t *)&inp->in6p_laddr.s6_addr[8];
        cpl->peer_port = inp->inp_fport;
        cpl->peer_ip_hi = *(uint64_t *)&inp->in6p_faddr.s6_addr[0];
        cpl->peer_ip_lo = *(uint64_t *)&inp->in6p_faddr.s6_addr[8];

        options = F_TCAM_BYPASS | V_ULP_MODE(ULP_MODE_NONE);
        options |= V_SMAC_SEL(vi->smt_idx) | V_TX_CHAN(vi->pi->tx_chan);
        options |= F_NON_OFFLOAD;
        cpl->opt0 = htobe64(options);

        options = V_TX_QUEUE(TX_MODQ(vi->pi->tx_chan));
        if (tp->t_flags & TF_REQ_TSTMP)
                options |= F_TSTAMPS_EN;
        cpl->opt2 = htobe32(options);
}

static int
send_ktls_act_open_req(struct adapter *sc, struct vi_info *vi,
    struct inpcb *inp, struct tlspcb *tlsp, int atid)
{
        struct wrqe *wr;
        bool isipv6;

        isipv6 = (inp->inp_vflag & INP_IPV6) != 0;
        if (isipv6) {
                tlsp->ce = t4_get_clip_entry(sc, &inp->in6p_laddr, true);
                if (tlsp->ce == NULL)
                        return (ENOENT);
        }

        wr = alloc_wrqe(ktls_act_open_cpl_size(isipv6), tlsp->ctrlq);
        if (wr == NULL) {
                CTR2(KTR_CXGBE, "%s: atid %d failed to alloc WR", __func__,
                    atid);
                return (ENOMEM);
        }

        if (isipv6)
                mk_ktls_act_open_req6(sc, vi, inp, tlsp, atid, wrtod(wr));
        else
                mk_ktls_act_open_req(sc, vi, inp, tlsp, atid, wrtod(wr));

        tlsp->open_pending = true;
        t4_wrq_tx(sc, wr);
        return (0);
}

static int
ktls_act_open_rpl(struct sge_iq *iq, const struct rss_header *rss,
    struct mbuf *m)
{
        struct adapter *sc = iq->adapter;
        const struct cpl_act_open_rpl *cpl = (const void *)(rss + 1);
        u_int atid = G_TID_TID(G_AOPEN_ATID(be32toh(cpl->atid_status)));
        u_int status = G_AOPEN_STATUS(be32toh(cpl->atid_status));
        struct tlspcb *tlsp = lookup_atid(sc, atid);
        struct inpcb *inp = tlsp->inp;

        CTR3(KTR_CXGBE, "%s: atid %d status %d", __func__, atid, status);
        free_atid(sc, atid);
        if (status == 0)
                tlsp->tid = GET_TID(cpl);

        INP_WLOCK(inp);
        tlsp->open_pending = false;
        wakeup(tlsp);
        INP_WUNLOCK(inp);
        return (0);
}

/* SET_TCB_FIELD sent as a ULP command looks like this */
#define LEN__SET_TCB_FIELD_ULP (sizeof(struct ulp_txpkt) + \
    sizeof(struct ulptx_idata) + sizeof(struct cpl_set_tcb_field_core))

_Static_assert((LEN__SET_TCB_FIELD_ULP + sizeof(struct ulptx_idata)) % 16 == 0,
    "CPL_SET_TCB_FIELD ULP command not 16-byte aligned");

static void
write_set_tcb_field_ulp(struct tlspcb *tlsp, void *dst, struct sge_txq *txq,
    uint16_t word, uint64_t mask, uint64_t val)
{
        struct ulp_txpkt *txpkt;
        struct ulptx_idata *idata;
        struct cpl_set_tcb_field_core *cpl;

        /* ULP_TXPKT */
        txpkt = dst;
        txpkt->cmd_dest = htobe32(V_ULPTX_CMD(ULP_TX_PKT) |
            V_ULP_TXPKT_DATAMODIFY(0) |
            V_ULP_TXPKT_CHANNELID(tlsp->vi->pi->port_id) | V_ULP_TXPKT_DEST(0) |
            V_ULP_TXPKT_FID(txq->eq.cntxt_id) | V_ULP_TXPKT_RO(1));
        txpkt->len = htobe32(howmany(LEN__SET_TCB_FIELD_ULP, 16));

        /* ULPTX_IDATA sub-command */
        idata = (struct ulptx_idata *)(txpkt + 1);
        idata->cmd_more = htobe32(V_ULPTX_CMD(ULP_TX_SC_IMM));
        idata->len = htobe32(sizeof(*cpl));

        /* CPL_SET_TCB_FIELD */
        cpl = (struct cpl_set_tcb_field_core *)(idata + 1);
        OPCODE_TID(cpl) = htobe32(MK_OPCODE_TID(CPL_SET_TCB_FIELD, tlsp->tid));
        cpl->reply_ctrl = htobe16(F_NO_REPLY);
        cpl->word_cookie = htobe16(V_WORD(word));
        cpl->mask = htobe64(mask);
        cpl->val = htobe64(val);

        /* ULPTX_NOOP */
        idata = (struct ulptx_idata *)(cpl + 1);
        idata->cmd_more = htobe32(V_ULPTX_CMD(ULP_TX_SC_NOOP));
        idata->len = htobe32(0);
}

static int
ktls_set_tcb_fields(struct tlspcb *tlsp, struct tcpcb *tp, struct sge_txq *txq)
{
        struct fw_ulptx_wr *wr;
        struct mbuf *m;
        char *dst;
        void *items[1];
        int error, len;

        len = sizeof(*wr) + 3 * roundup2(LEN__SET_TCB_FIELD_ULP, 16);
        if (tp->t_flags & TF_REQ_TSTMP)
                len += roundup2(LEN__SET_TCB_FIELD_ULP, 16);
        m = alloc_wr_mbuf(len, M_NOWAIT);
        if (m == NULL) {
                CTR2(KTR_CXGBE, "%s: tid %d failed to alloc WR mbuf", __func__,
                    tlsp->tid);
                return (ENOMEM);
        }
        m->m_pkthdr.snd_tag = m_snd_tag_ref(&tlsp->com);
        m->m_pkthdr.csum_flags |= CSUM_SND_TAG;

        /* FW_ULPTX_WR */
        wr = mtod(m, void *);
        wr->op_to_compl = htobe32(V_FW_WR_OP(FW_ULPTX_WR));
        wr->flowid_len16 = htobe32(F_FW_ULPTX_WR_DATA |
            V_FW_WR_LEN16(len / 16));
        wr->cookie = 0;
        dst = (char *)(wr + 1);

        /* Clear TF_NON_OFFLOAD and set TF_CORE_BYPASS */
        write_set_tcb_field_ulp(tlsp, dst, txq, W_TCB_T_FLAGS,
            V_TCB_T_FLAGS(V_TF_CORE_BYPASS(1) | V_TF_NON_OFFLOAD(1)),
            V_TCB_T_FLAGS(V_TF_CORE_BYPASS(1)));
        dst += roundup2(LEN__SET_TCB_FIELD_ULP, 16);

        /* Clear the SND_UNA_RAW, SND_NXT_RAW, and SND_MAX_RAW offsets. */
        write_set_tcb_field_ulp(tlsp, dst, txq, W_TCB_SND_UNA_RAW,
            V_TCB_SND_NXT_RAW(M_TCB_SND_NXT_RAW) |
            V_TCB_SND_UNA_RAW(M_TCB_SND_UNA_RAW),
            V_TCB_SND_NXT_RAW(0) | V_TCB_SND_UNA_RAW(0));
        dst += roundup2(LEN__SET_TCB_FIELD_ULP, 16);

        write_set_tcb_field_ulp(tlsp, dst, txq, W_TCB_SND_MAX_RAW,
            V_TCB_SND_MAX_RAW(M_TCB_SND_MAX_RAW), V_TCB_SND_MAX_RAW(0));
        dst += roundup2(LEN__SET_TCB_FIELD_ULP, 16);

        if (tp->t_flags & TF_REQ_TSTMP) {
                write_set_tcb_field_ulp(tlsp, dst, txq, W_TCB_TIMESTAMP_OFFSET,
                    V_TCB_TIMESTAMP_OFFSET(M_TCB_TIMESTAMP_OFFSET),
                    V_TCB_TIMESTAMP_OFFSET(tp->ts_offset >> 28));
                dst += roundup2(LEN__SET_TCB_FIELD_ULP, 16);
        }

        KASSERT(dst - (char *)wr == len, ("%s: length mismatch", __func__));

        items[0] = m;
        error = mp_ring_enqueue(txq->r, items, 1, 1);
        if (error)
                m_free(m);
        return (error);
}

int
t6_tls_tag_alloc(if_t ifp, union if_snd_tag_alloc_params *params,
    struct m_snd_tag **pt)
{
        const struct ktls_session *tls;
        struct tlspcb *tlsp;
        struct adapter *sc;
        struct vi_info *vi;
        struct inpcb *inp;
        struct tcpcb *tp;
        struct sge_txq *txq;
        int atid, error, explicit_iv_size, keyid, mac_first;

        tls = params->tls.tls;

        /* Only TLS 1.1 and TLS 1.2 are currently supported. */
        if (tls->params.tls_vmajor != TLS_MAJOR_VER_ONE ||
            tls->params.tls_vminor < TLS_MINOR_VER_ONE ||
            tls->params.tls_vminor > TLS_MINOR_VER_TWO)
                return (EPROTONOSUPPORT);

        /* Sanity check values in *tls. */
        switch (tls->params.cipher_algorithm) {
        case CRYPTO_AES_CBC:
                /* XXX: Explicitly ignore any provided IV. */
                switch (tls->params.cipher_key_len) {
                case 128 / 8:
                case 192 / 8:
                case 256 / 8:
                        break;
                default:
                        return (EINVAL);
                }
                switch (tls->params.auth_algorithm) {
                case CRYPTO_SHA1_HMAC:
                case CRYPTO_SHA2_256_HMAC:
                case CRYPTO_SHA2_384_HMAC:
                        break;
                default:
                        return (EPROTONOSUPPORT);
                }
                explicit_iv_size = AES_BLOCK_LEN;
                mac_first = 1;
                break;
        case CRYPTO_AES_NIST_GCM_16:
                if (tls->params.iv_len != SALT_SIZE)
                        return (EINVAL);
                switch (tls->params.cipher_key_len) {
                case 128 / 8:
                case 192 / 8:
                case 256 / 8:
                        break;
                default:
                        return (EINVAL);
                }
                explicit_iv_size = 8;
                mac_first = 0;
                break;
        default:
                return (EPROTONOSUPPORT);
        }

        vi = if_getsoftc(ifp);
        sc = vi->adapter;

        tlsp = alloc_tlspcb(ifp, vi, M_WAITOK);

        atid = alloc_atid(sc, tlsp);
        if (atid < 0) {
                error = ENOMEM;
                goto failed;
        }

        if (sc->tlst.inline_keys)
                keyid = -1;
        else
                keyid = t4_alloc_tls_keyid(sc);
        if (keyid < 0) {
                CTR2(KTR_CXGBE, "%s: atid %d using immediate key ctx", __func__,
                    atid);
                tlsp->inline_key = true;
        } else {
                tlsp->tx_key_addr = keyid;
                CTR3(KTR_CXGBE, "%s: atid %d allocated TX key addr %#x",
                    __func__,
                    atid, tlsp->tx_key_addr);
        }

        inp = params->tls.inp;
        INP_RLOCK(inp);
        if (inp->inp_flags & INP_DROPPED) {
                INP_RUNLOCK(inp);
                error = ECONNRESET;
                goto failed;
        }
        tlsp->inp = inp;

        tp = intotcpcb(inp);
        if (tp->t_flags & TF_REQ_TSTMP) {
                tlsp->using_timestamps = true;
                if ((tp->ts_offset & 0xfffffff) != 0) {
                        INP_RUNLOCK(inp);
                        error = EINVAL;
                        goto failed;
                }
        } else
                tlsp->using_timestamps = false;

        error = send_ktls_act_open_req(sc, vi, inp, tlsp, atid);
        if (error) {
                INP_RUNLOCK(inp);
                goto failed;
        }

        /* Wait for reply to active open. */
        CTR2(KTR_CXGBE, "%s: atid %d sent CPL_ACT_OPEN_REQ", __func__,
            atid);
        while (tlsp->open_pending) {
                /*
                 * XXX: PCATCH?  We would then have to discard the PCB
                 * when the completion CPL arrived.
                 */
                error = rw_sleep(tlsp, &inp->inp_lock, 0, "t6tlsop", 0);
        }

        atid = -1;
        if (tlsp->tid < 0) {
                INP_RUNLOCK(inp);
                error = ENOMEM;
                goto failed;
        }

        if (inp->inp_flags & INP_DROPPED) {
                INP_RUNLOCK(inp);
                error = ECONNRESET;
                goto failed;
        }

        txq = &sc->sge.txq[vi->first_txq];
        if (inp->inp_flowtype != M_HASHTYPE_NONE)
                txq += ((inp->inp_flowid % (vi->ntxq - vi->rsrv_noflowq)) +
                    vi->rsrv_noflowq);
        tlsp->txq = txq;

        error = ktls_set_tcb_fields(tlsp, tp, txq);
        INP_RUNLOCK(inp);
        if (error)
                goto failed;

        error = ktls_setup_keys(tlsp, tls, txq);
        if (error)
                goto failed;

        tlsp->enc_mode = t4_tls_cipher_mode(tls);
        tlsp->tx_key_info_size = t4_tls_key_info_size(tls);

        /* The SCMD fields used when encrypting a full TLS record. */
        tlsp->scmd0.seqno_numivs = htobe32(V_SCMD_SEQ_NO_CTRL(3) |
            V_SCMD_PROTO_VERSION(t4_tls_proto_ver(tls)) |
            V_SCMD_ENC_DEC_CTRL(SCMD_ENCDECCTRL_ENCRYPT) |
            V_SCMD_CIPH_AUTH_SEQ_CTRL((mac_first == 0)) |
            V_SCMD_CIPH_MODE(tlsp->enc_mode) |
            V_SCMD_AUTH_MODE(t4_tls_auth_mode(tls)) |
            V_SCMD_HMAC_CTRL(t4_tls_hmac_ctrl(tls)) |
            V_SCMD_IV_SIZE(explicit_iv_size / 2) | V_SCMD_NUM_IVS(1));

        tlsp->scmd0.ivgen_hdrlen = V_SCMD_IV_GEN_CTRL(0) |
            V_SCMD_TLS_FRAG_ENABLE(0);
        if (tlsp->inline_key)
                tlsp->scmd0.ivgen_hdrlen |= V_SCMD_KEY_CTX_INLINE(1);
        tlsp->scmd0.ivgen_hdrlen = htobe32(tlsp->scmd0.ivgen_hdrlen);

        /*
         * The SCMD fields used when encrypting a partial TLS record
         * (no trailer and possibly a truncated payload).
         */
        tlsp->scmd0_short.seqno_numivs = V_SCMD_SEQ_NO_CTRL(0) |
            V_SCMD_PROTO_VERSION(SCMD_PROTO_VERSION_GENERIC) |
            V_SCMD_ENC_DEC_CTRL(SCMD_ENCDECCTRL_ENCRYPT) |
            V_SCMD_CIPH_AUTH_SEQ_CTRL((mac_first == 0)) |
            V_SCMD_AUTH_MODE(SCMD_AUTH_MODE_NOP) |
            V_SCMD_HMAC_CTRL(SCMD_HMAC_CTRL_NOP) |
            V_SCMD_IV_SIZE(AES_BLOCK_LEN / 2) | V_SCMD_NUM_IVS(0);
        if (tlsp->enc_mode == SCMD_CIPH_MODE_AES_GCM)
                tlsp->scmd0_short.seqno_numivs |=
                    V_SCMD_CIPH_MODE(SCMD_CIPH_MODE_AES_CTR);
        else
                tlsp->scmd0_short.seqno_numivs |=
                    V_SCMD_CIPH_MODE(tlsp->enc_mode);
        tlsp->scmd0_short.seqno_numivs =
            htobe32(tlsp->scmd0_short.seqno_numivs);

        tlsp->scmd0_short.ivgen_hdrlen = V_SCMD_IV_GEN_CTRL(0) |
            V_SCMD_TLS_FRAG_ENABLE(0) |
            V_SCMD_AADIVDROP(1);
        if (tlsp->inline_key)
                tlsp->scmd0_short.ivgen_hdrlen |= V_SCMD_KEY_CTX_INLINE(1);

        TXQ_LOCK(txq);
        if (tlsp->enc_mode == SCMD_CIPH_MODE_AES_GCM)
                txq->kern_tls_gcm++;
        else
                txq->kern_tls_cbc++;
        TXQ_UNLOCK(txq);
        *pt = &tlsp->com;
        return (0);

failed:
        if (atid >= 0)
                free_atid(sc, atid);
        m_snd_tag_rele(&tlsp->com);
        return (error);
}

static int
ktls_setup_keys(struct tlspcb *tlsp, const struct ktls_session *tls,
    struct sge_txq *txq)
{
        struct tls_key_req *kwr;
        struct tls_keyctx *kctx;
        void *items[1];
        struct mbuf *m;
        int error;

        /*
         * Store the salt and keys in the key context.  For
         * connections with an inline key, this key context is passed
         * as immediate data in each work request.  For connections
         * storing the key in DDR, a work request is used to store a
         * copy of the key context in DDR.
         */
        t4_tls_key_ctx(tls, KTLS_TX, &tlsp->keyctx);
        if (tlsp->inline_key)
                return (0);

        /* Populate key work request. */
        m = alloc_wr_mbuf(TLS_KEY_WR_SZ, M_NOWAIT);
        if (m == NULL) {
                CTR2(KTR_CXGBE, "%s: tid %d failed to alloc WR mbuf", __func__,
                    tlsp->tid);
                return (ENOMEM);
        }
        m->m_pkthdr.snd_tag = m_snd_tag_ref(&tlsp->com);
        m->m_pkthdr.csum_flags |= CSUM_SND_TAG;
        kwr = mtod(m, void *);
        memset(kwr, 0, TLS_KEY_WR_SZ);

        t4_write_tlskey_wr(tls, KTLS_TX, tlsp->tid, 0, tlsp->tx_key_addr, kwr);
        kctx = (struct tls_keyctx *)(kwr + 1);
        memcpy(kctx, &tlsp->keyctx, sizeof(*kctx));

        /*
         * Place the key work request in the transmit queue.  It
         * should be sent to the NIC before any TLS packets using this
         * session.
         */
        items[0] = m;
        error = mp_ring_enqueue(txq->r, items, 1, 1);
        if (error)
                m_free(m);
        else
                CTR2(KTR_CXGBE, "%s: tid %d sent key WR", __func__, tlsp->tid);
        return (error);
}

static u_int
ktls_base_wr_size(struct tlspcb *tlsp)
{
        u_int wr_len;

        wr_len = sizeof(struct fw_ulptx_wr);    // 16
        wr_len += sizeof(struct ulp_txpkt);     // 8
        wr_len += sizeof(struct ulptx_idata);   // 8
        wr_len += sizeof(struct cpl_tx_sec_pdu);// 32
        if (tlsp->inline_key)
                wr_len += tlsp->tx_key_info_size;
        else {
                wr_len += sizeof(struct ulptx_sc_memrd);// 8
                wr_len += sizeof(struct ulptx_idata);   // 8
        }
        wr_len += sizeof(struct cpl_tx_data);   // 16
        return (wr_len);
}

/* How many bytes of TCP payload to send for a given TLS record. */
static u_int
ktls_tcp_payload_length(struct tlspcb *tlsp, struct mbuf *m_tls)
{
        struct tls_record_layer *hdr;
        u_int plen, mlen;

        M_ASSERTEXTPG(m_tls);
        hdr = (void *)m_tls->m_epg_hdr;
        plen = ntohs(hdr->tls_length);

        /*
         * What range of the TLS record is the mbuf requesting to be
         * sent.
         */
        mlen = mtod(m_tls, vm_offset_t) + m_tls->m_len;

        /* Always send complete records. */
        if (mlen == TLS_HEADER_LENGTH + plen)
                return (mlen);

        /*
         * If the host stack has asked to send part of the trailer,
         * trim the length to avoid sending any of the trailer.  There
         * is no way to send a partial trailer currently.
         */
        if (mlen > TLS_HEADER_LENGTH + plen - m_tls->m_epg_trllen)
                mlen = TLS_HEADER_LENGTH + plen - m_tls->m_epg_trllen;


        /*
         * For AES-CBC adjust the ciphertext length for the block
         * size.
         */
        if (tlsp->enc_mode == SCMD_CIPH_MODE_AES_CBC &&
            mlen > TLS_HEADER_LENGTH) {
                mlen = TLS_HEADER_LENGTH + rounddown(mlen - TLS_HEADER_LENGTH,
                    AES_BLOCK_LEN);
        }

#ifdef VERBOSE_TRACES
        CTR4(KTR_CXGBE, "%s: tid %d short TLS record (%u vs %u)",
            __func__, tlsp->tid, mlen, TLS_HEADER_LENGTH + plen);
#endif
        return (mlen);
}

/*
 * For a "short" TLS record, determine the offset into the TLS record
 * payload to send.  This offset does not include the TLS header, but
 * a non-zero offset implies that a header will not be sent.
 */
static u_int
ktls_payload_offset(struct tlspcb *tlsp, struct mbuf *m_tls)
{
        struct tls_record_layer *hdr;
        u_int offset, plen;
#ifdef INVARIANTS
        u_int mlen;
#endif

        M_ASSERTEXTPG(m_tls);
        hdr = (void *)m_tls->m_epg_hdr;
        plen = ntohs(hdr->tls_length);
#ifdef INVARIANTS
        mlen = mtod(m_tls, vm_offset_t) + m_tls->m_len;
        MPASS(mlen < TLS_HEADER_LENGTH + plen);
#endif
        if (mtod(m_tls, vm_offset_t) <= m_tls->m_epg_hdrlen)
                return (0);
        if (tlsp->enc_mode == SCMD_CIPH_MODE_AES_GCM) {
                /*
                 * Always send something.  This function is only called
                 * if we aren't sending the tag at all, but if the
                 * request starts in the tag then we are in an odd
                 * state where would effectively send nothing.  Cap
                 * the offset at the last byte of the record payload
                 * to send the last cipher block.
                 */
                offset = min(mtod(m_tls, vm_offset_t) - m_tls->m_epg_hdrlen,
                    (plen - TLS_HEADER_LENGTH - m_tls->m_epg_trllen) - 1);
                return (rounddown(offset, AES_BLOCK_LEN));
        }
        return (0);
}

static u_int
ktls_sgl_size(u_int nsegs)
{
        u_int wr_len;

        /* First segment is part of ulptx_sgl. */
        nsegs--;

        wr_len = sizeof(struct ulptx_sgl);
        wr_len += 8 * ((3 * nsegs) / 2 + (nsegs & 1));
        return (wr_len);
}

static int
ktls_wr_len(struct tlspcb *tlsp, struct mbuf *m, struct mbuf *m_tls,
    int *nsegsp)
{
        struct tls_record_layer *hdr;
        u_int imm_len, offset, plen, wr_len, tlen;

        M_ASSERTEXTPG(m_tls);

        /*
         * Determine the size of the TLS record payload to send
         * excluding header and trailer.
         */
        tlen = ktls_tcp_payload_length(tlsp, m_tls);
        if (tlen <= m_tls->m_epg_hdrlen) {
                /*
                 * For requests that only want to send the TLS header,
                 * send a tunnelled packet as immediate data.
                 */
                wr_len = sizeof(struct fw_eth_tx_pkt_wr) +
                    sizeof(struct cpl_tx_pkt_core) +
                    roundup2(m->m_len + m_tls->m_len, 16);
                if (wr_len > SGE_MAX_WR_LEN) {
                        CTR3(KTR_CXGBE,
                    "%s: tid %d TLS header-only packet too long (len %d)",
                            __func__, tlsp->tid, m->m_len + m_tls->m_len);
                }

                /* This should always be the last TLS record in a chain. */
                MPASS(m_tls->m_next == NULL);
                *nsegsp = 0;
                return (wr_len);
        }

        hdr = (void *)m_tls->m_epg_hdr;
        plen = TLS_HEADER_LENGTH + ntohs(hdr->tls_length) - m_tls->m_epg_trllen;
        if (tlen < plen) {
                plen = tlen;
                offset = ktls_payload_offset(tlsp, m_tls);
        } else
                offset = 0;

        /* Calculate the size of the work request. */
        wr_len = ktls_base_wr_size(tlsp);

        /*
         * Full records and short records with an offset of 0 include
         * the TLS header as immediate data.  Short records include a
         * raw AES IV as immediate data.
         */
        imm_len = 0;
        if (offset == 0)
                imm_len += m_tls->m_epg_hdrlen;
        if (plen == tlen)
                imm_len += AES_BLOCK_LEN;
        wr_len += roundup2(imm_len, 16);

        /* TLS record payload via DSGL. */
        *nsegsp = sglist_count_mbuf_epg(m_tls, m_tls->m_epg_hdrlen + offset,
            plen - (m_tls->m_epg_hdrlen + offset));
        wr_len += ktls_sgl_size(*nsegsp);

        wr_len = roundup2(wr_len, 16);
        return (wr_len);
}

/*
 * See if we have any TCP options requiring a dedicated options-only
 * packet.
 */
static int
ktls_has_tcp_options(struct tcphdr *tcp)
{
        u_char *cp;
        int cnt, opt, optlen;

        cp = (u_char *)(tcp + 1);
        cnt = tcp->th_off * 4 - sizeof(struct tcphdr);
        for (; cnt > 0; cnt -= optlen, cp += optlen) {
                opt = cp[0];
                if (opt == TCPOPT_EOL)
                        break;
                if (opt == TCPOPT_NOP)
                        optlen = 1;
                else {
                        if (cnt < 2)
                                break;
                        optlen = cp[1];
                        if (optlen < 2 || optlen > cnt)
                                break;
                }
                switch (opt) {
                case TCPOPT_NOP:
                case TCPOPT_TIMESTAMP:
                        break;
                default:
                        return (1);
                }
        }
        return (0);
}

/*
 * Find the TCP timestamp option.
 */
static void *
ktls_find_tcp_timestamps(struct tcphdr *tcp)
{
        u_char *cp;
        int cnt, opt, optlen;

        cp = (u_char *)(tcp + 1);
        cnt = tcp->th_off * 4 - sizeof(struct tcphdr);
        for (; cnt > 0; cnt -= optlen, cp += optlen) {
                opt = cp[0];
                if (opt == TCPOPT_EOL)
                        break;
                if (opt == TCPOPT_NOP)
                        optlen = 1;
                else {
                        if (cnt < 2)
                                break;
                        optlen = cp[1];
                        if (optlen < 2 || optlen > cnt)
                                break;
                }
                if (opt == TCPOPT_TIMESTAMP && optlen == TCPOLEN_TIMESTAMP)
                        return (cp + 2);
        }
        return (NULL);
}

int
t6_ktls_parse_pkt(struct mbuf *m)
{
        struct tlspcb *tlsp;
        struct ether_header *eh;
        struct ip *ip;
        struct ip6_hdr *ip6;
        struct tcphdr *tcp;
        struct mbuf *m_tls;
        void *items[1];
        int nsegs;
        u_int wr_len, tot_len;

        /*
         * Locate headers in initial mbuf.
         *
         * XXX: This assumes all of the headers are in the initial mbuf.
         * Could perhaps use m_advance() like parse_pkt() if that turns
         * out to not be true.
         */
        M_ASSERTPKTHDR(m);
        MPASS(m->m_pkthdr.snd_tag != NULL);
        tlsp = mst_to_tls(m->m_pkthdr.snd_tag);

        if (m->m_len <= sizeof(*eh) + sizeof(*ip)) {
                CTR2(KTR_CXGBE, "%s: tid %d header mbuf too short", __func__,
                    tlsp->tid);
                return (EINVAL);
        }
        eh = mtod(m, struct ether_header *);
        if (ntohs(eh->ether_type) != ETHERTYPE_IP &&
            ntohs(eh->ether_type) != ETHERTYPE_IPV6) {
                CTR2(KTR_CXGBE, "%s: tid %d mbuf not ETHERTYPE_IP{,V6}",
                    __func__, tlsp->tid);
                return (EINVAL);
        }
        m->m_pkthdr.l2hlen = sizeof(*eh);

        /* XXX: Reject unsupported IP options? */
        if (ntohs(eh->ether_type) == ETHERTYPE_IP) {
                ip = (struct ip *)(eh + 1);
                if (ip->ip_p != IPPROTO_TCP) {
                        CTR2(KTR_CXGBE, "%s: tid %d mbuf not IPPROTO_TCP",
                            __func__, tlsp->tid);
                        return (EINVAL);
                }
                m->m_pkthdr.l3hlen = ip->ip_hl * 4;
        } else {
                ip6 = (struct ip6_hdr *)(eh + 1);
                if (ip6->ip6_nxt != IPPROTO_TCP) {
                        CTR3(KTR_CXGBE, "%s: tid %d mbuf not IPPROTO_TCP (%u)",
                            __func__, tlsp->tid, ip6->ip6_nxt);
                        return (EINVAL);
                }
                m->m_pkthdr.l3hlen = sizeof(struct ip6_hdr);
        }
        if (m->m_len < m->m_pkthdr.l2hlen + m->m_pkthdr.l3hlen +
            sizeof(*tcp)) {
                CTR2(KTR_CXGBE, "%s: tid %d header mbuf too short (2)",
                    __func__, tlsp->tid);
                return (EINVAL);
        }
        tcp = (struct tcphdr *)((char *)(eh + 1) + m->m_pkthdr.l3hlen);
        m->m_pkthdr.l4hlen = tcp->th_off * 4;

        /* Bail if there is TCP payload before the TLS record. */
        if (m->m_len != m->m_pkthdr.l2hlen + m->m_pkthdr.l3hlen +
            m->m_pkthdr.l4hlen) {
                CTR6(KTR_CXGBE,
                    "%s: tid %d header mbuf bad length (%d + %d + %d != %d)",
                    __func__, tlsp->tid, m->m_pkthdr.l2hlen,
                    m->m_pkthdr.l3hlen, m->m_pkthdr.l4hlen, m->m_len);
                return (EINVAL);
        }

        /* Assume all headers are in 'm' for now. */
        MPASS(m->m_next != NULL);
        MPASS(m->m_next->m_flags & M_EXTPG);

        tot_len = 0;

        /*
         * Each of the remaining mbufs in the chain should reference a
         * TLS record.
         */
        for (m_tls = m->m_next; m_tls != NULL; m_tls = m_tls->m_next) {
                MPASS(m_tls->m_flags & M_EXTPG);

                wr_len = ktls_wr_len(tlsp, m, m_tls, &nsegs);
#ifdef VERBOSE_TRACES
                CTR4(KTR_CXGBE, "%s: tid %d wr_len %d nsegs %d", __func__,
                    tlsp->tid, wr_len, nsegs);
#endif
                if (wr_len > SGE_MAX_WR_LEN || nsegs > TX_SGL_SEGS)
                        return (EFBIG);
                tot_len += roundup2(wr_len, EQ_ESIZE);

                /*
                 * Store 'nsegs' for the first TLS record in the
                 * header mbuf's metadata.
                 */
                if (m_tls == m->m_next)
                        set_mbuf_nsegs(m, nsegs);
        }

        MPASS(tot_len != 0);

        /*
         * See if we have any TCP options or a FIN requiring a
         * dedicated packet.
         */
        if ((tcp_get_flags(tcp) & TH_FIN) != 0 || ktls_has_tcp_options(tcp)) {
                wr_len = sizeof(struct fw_eth_tx_pkt_wr) +
                    sizeof(struct cpl_tx_pkt_core) + roundup2(m->m_len, 16);
                if (wr_len > SGE_MAX_WR_LEN) {
                        CTR3(KTR_CXGBE,
                            "%s: tid %d options-only packet too long (len %d)",
                            __func__, tlsp->tid, m->m_len);
                        return (EINVAL);
                }
                tot_len += roundup2(wr_len, EQ_ESIZE);
        }

        /* Include room for a TP work request to program an L2T entry. */
        tot_len += EQ_ESIZE;

        /*
         * Include room for a ULPTX work request including up to 5
         * CPL_SET_TCB_FIELD commands before the first TLS work
         * request.
         */
        wr_len = sizeof(struct fw_ulptx_wr) +
            5 * roundup2(LEN__SET_TCB_FIELD_ULP, 16);

        /*
         * If timestamps are present, reserve 1 more command for
         * setting the echoed timestamp.
         */
        if (tlsp->using_timestamps)
                wr_len += roundup2(LEN__SET_TCB_FIELD_ULP, 16);

        tot_len += roundup2(wr_len, EQ_ESIZE);

        set_mbuf_len16(m, tot_len / 16);
#ifdef VERBOSE_TRACES
        CTR4(KTR_CXGBE, "%s: tid %d len16 %d nsegs %d", __func__,
            tlsp->tid, mbuf_len16(m), mbuf_nsegs(m));
#endif
        items[0] = m;
        return (mp_ring_enqueue(tlsp->txq->r, items, 1, 256));
}

/*
 * If the SGL ends on an address that is not 16 byte aligned, this function will
 * add a 0 filled flit at the end.
 */
static void
write_gl_to_buf(struct sglist *gl, caddr_t to)
{
        struct sglist_seg *seg;
        __be64 *flitp;
        struct ulptx_sgl *usgl;
        int i, nflits, nsegs;

        KASSERT(((uintptr_t)to & 0xf) == 0,
            ("%s: SGL must start at a 16 byte boundary: %p", __func__, to));

        nsegs = gl->sg_nseg;
        MPASS(nsegs > 0);

        nflits = (3 * (nsegs - 1)) / 2 + ((nsegs - 1) & 1) + 2;
        flitp = (__be64 *)to;
        seg = &gl->sg_segs[0];
        usgl = (void *)flitp;

        usgl->cmd_nsge = htobe32(V_ULPTX_CMD(ULP_TX_SC_DSGL) |
            V_ULPTX_NSGE(nsegs));
        usgl->len0 = htobe32(seg->ss_len);
        usgl->addr0 = htobe64(seg->ss_paddr);
        seg++;

        for (i = 0; i < nsegs - 1; i++, seg++) {
                usgl->sge[i / 2].len[i & 1] = htobe32(seg->ss_len);
                usgl->sge[i / 2].addr[i & 1] = htobe64(seg->ss_paddr);
        }
        if (i & 1)
                usgl->sge[i / 2].len[1] = htobe32(0);
        flitp += nflits;

        if (nflits & 1) {
                MPASS(((uintptr_t)flitp) & 0xf);
                *flitp++ = 0;
        }

        MPASS((((uintptr_t)flitp) & 0xf) == 0);
}

static inline void
copy_to_txd(struct sge_eq *eq, caddr_t from, caddr_t *to, int len)
{

        MPASS((uintptr_t)(*to) >= (uintptr_t)&eq->desc[0]);
        MPASS((uintptr_t)(*to) < (uintptr_t)&eq->desc[eq->sidx]);

        if (__predict_true((uintptr_t)(*to) + len <=
            (uintptr_t)&eq->desc[eq->sidx])) {
                bcopy(from, *to, len);
                (*to) += len;
                if ((uintptr_t)(*to) == (uintptr_t)&eq->desc[eq->sidx])
                        (*to) = (caddr_t)eq->desc;
        } else {
                int portion = (uintptr_t)&eq->desc[eq->sidx] - (uintptr_t)(*to);

                bcopy(from, *to, portion);
                from += portion;
                portion = len - portion;        /* remaining */
                bcopy(from, (void *)eq->desc, portion);
                (*to) = (caddr_t)eq->desc + portion;
        }
}

static int
ktls_write_tcp_options(struct sge_txq *txq, void *dst, struct mbuf *m,
    u_int available, u_int pidx)
{
        struct tx_sdesc *txsd;
        struct fw_eth_tx_pkt_wr *wr;
        struct cpl_tx_pkt_core *cpl;
        uint32_t ctrl;
        uint64_t ctrl1;
        int len16, ndesc, pktlen;
        struct ether_header *eh;
        struct ip *ip, newip;
        struct ip6_hdr *ip6, newip6;
        struct tcphdr *tcp, newtcp;
        caddr_t out;

        TXQ_LOCK_ASSERT_OWNED(txq);
        M_ASSERTPKTHDR(m);

        wr = dst;
        pktlen = m->m_len;
        ctrl = sizeof(struct cpl_tx_pkt_core) + pktlen;
        len16 = howmany(sizeof(struct fw_eth_tx_pkt_wr) + ctrl, 16);
        ndesc = tx_len16_to_desc(len16);
        MPASS(ndesc <= available);

        /* Firmware work request header */
        wr->op_immdlen = htobe32(V_FW_WR_OP(FW_ETH_TX_PKT_WR) |
            V_FW_ETH_TX_PKT_WR_IMMDLEN(ctrl));

        ctrl = V_FW_WR_LEN16(len16);
        wr->equiq_to_len16 = htobe32(ctrl);
        wr->r3 = 0;

        cpl = (void *)(wr + 1);

        /* CPL header */
        cpl->ctrl0 = txq->cpl_ctrl0;
        cpl->pack = 0;
        cpl->len = htobe16(pktlen);

        out = (void *)(cpl + 1);

        /* Copy over Ethernet header. */
        eh = mtod(m, struct ether_header *);
        copy_to_txd(&txq->eq, (caddr_t)eh, &out, m->m_pkthdr.l2hlen);

        /* Fixup length in IP header and copy out. */
        if (ntohs(eh->ether_type) == ETHERTYPE_IP) {
                ip = (void *)((char *)eh + m->m_pkthdr.l2hlen);
                newip = *ip;
                newip.ip_len = htons(pktlen - m->m_pkthdr.l2hlen);
                copy_to_txd(&txq->eq, (caddr_t)&newip, &out, sizeof(newip));
                if (m->m_pkthdr.l3hlen > sizeof(*ip))
                        copy_to_txd(&txq->eq, (caddr_t)(ip + 1), &out,
                            m->m_pkthdr.l3hlen - sizeof(*ip));
                ctrl1 = V_TXPKT_CSUM_TYPE(TX_CSUM_TCPIP) |
                    V_T6_TXPKT_ETHHDR_LEN(m->m_pkthdr.l2hlen - ETHER_HDR_LEN) |
                    V_TXPKT_IPHDR_LEN(m->m_pkthdr.l3hlen);
        } else {
                ip6 = (void *)((char *)eh + m->m_pkthdr.l2hlen);
                newip6 = *ip6;
                newip6.ip6_plen = htons(pktlen - m->m_pkthdr.l2hlen -
                    sizeof(*ip6));
                copy_to_txd(&txq->eq, (caddr_t)&newip6, &out, sizeof(newip6));
                MPASS(m->m_pkthdr.l3hlen == sizeof(*ip6));
                ctrl1 = V_TXPKT_CSUM_TYPE(TX_CSUM_TCPIP6) |
                    V_T6_TXPKT_ETHHDR_LEN(m->m_pkthdr.l2hlen - ETHER_HDR_LEN) |
                    V_TXPKT_IPHDR_LEN(m->m_pkthdr.l3hlen);
        }
        cpl->ctrl1 = htobe64(ctrl1);
        txq->txcsum++;

        /* Clear PUSH and FIN in the TCP header if present. */
        tcp = (void *)((char *)eh + m->m_pkthdr.l2hlen + m->m_pkthdr.l3hlen);
        newtcp = *tcp;
        tcp_set_flags(&newtcp, tcp_get_flags(&newtcp) & ~(TH_PUSH | TH_FIN));
        copy_to_txd(&txq->eq, (caddr_t)&newtcp, &out, sizeof(newtcp));

        /* Copy rest of packet. */
        copy_to_txd(&txq->eq, (caddr_t)(tcp + 1), &out, pktlen -
            (m->m_pkthdr.l2hlen + m->m_pkthdr.l3hlen + sizeof(*tcp)));
        txq->imm_wrs++;

        txq->txpkt_wrs++;

        txq->kern_tls_options++;

        txsd = &txq->sdesc[pidx];
        txsd->m = NULL;
        txsd->desc_used = ndesc;

        return (ndesc);
}

static int
ktls_write_tunnel_packet(struct sge_txq *txq, void *dst, struct mbuf *m,
    struct mbuf *m_tls, u_int available, tcp_seq tcp_seqno, u_int pidx)
{
        struct tx_sdesc *txsd;
        struct fw_eth_tx_pkt_wr *wr;
        struct cpl_tx_pkt_core *cpl;
        uint32_t ctrl;
        uint64_t ctrl1;
        int len16, ndesc, pktlen;
        struct ether_header *eh;
        struct ip *ip, newip;
        struct ip6_hdr *ip6, newip6;
        struct tcphdr *tcp, newtcp;
        caddr_t out;

        TXQ_LOCK_ASSERT_OWNED(txq);
        M_ASSERTPKTHDR(m);

        /* Locate the template TLS header. */
        M_ASSERTEXTPG(m_tls);

        /* This should always be the last TLS record in a chain. */
        MPASS(m_tls->m_next == NULL);

        wr = dst;
        pktlen = m->m_len + m_tls->m_len;
        ctrl = sizeof(struct cpl_tx_pkt_core) + pktlen;
        len16 = howmany(sizeof(struct fw_eth_tx_pkt_wr) + ctrl, 16);
        ndesc = tx_len16_to_desc(len16);
        MPASS(ndesc <= available);

        /* Firmware work request header */
        wr->op_immdlen = htobe32(V_FW_WR_OP(FW_ETH_TX_PKT_WR) |
            V_FW_ETH_TX_PKT_WR_IMMDLEN(ctrl));

        ctrl = V_FW_WR_LEN16(len16);
        wr->equiq_to_len16 = htobe32(ctrl);
        wr->r3 = 0;

        cpl = (void *)(wr + 1);

        /* CPL header */
        cpl->ctrl0 = txq->cpl_ctrl0;
        cpl->pack = 0;
        cpl->len = htobe16(pktlen);

        out = (void *)(cpl + 1);

        /* Copy over Ethernet header. */
        eh = mtod(m, struct ether_header *);
        copy_to_txd(&txq->eq, (caddr_t)eh, &out, m->m_pkthdr.l2hlen);

        /* Fixup length in IP header and copy out. */
        if (ntohs(eh->ether_type) == ETHERTYPE_IP) {
                ip = (void *)((char *)eh + m->m_pkthdr.l2hlen);
                newip = *ip;
                newip.ip_len = htons(pktlen - m->m_pkthdr.l2hlen);
                copy_to_txd(&txq->eq, (caddr_t)&newip, &out, sizeof(newip));
                if (m->m_pkthdr.l3hlen > sizeof(*ip))
                        copy_to_txd(&txq->eq, (caddr_t)(ip + 1), &out,
                            m->m_pkthdr.l3hlen - sizeof(*ip));
                ctrl1 = V_TXPKT_CSUM_TYPE(TX_CSUM_TCPIP) |
                    V_T6_TXPKT_ETHHDR_LEN(m->m_pkthdr.l2hlen - ETHER_HDR_LEN) |
                    V_TXPKT_IPHDR_LEN(m->m_pkthdr.l3hlen);
        } else {
                ip6 = (void *)((char *)eh + m->m_pkthdr.l2hlen);
                newip6 = *ip6;
                newip6.ip6_plen = htons(pktlen - m->m_pkthdr.l2hlen -
                    sizeof(*ip6));
                copy_to_txd(&txq->eq, (caddr_t)&newip6, &out, sizeof(newip6));
                MPASS(m->m_pkthdr.l3hlen == sizeof(*ip6));
                ctrl1 = V_TXPKT_CSUM_TYPE(TX_CSUM_TCPIP6) |
                    V_T6_TXPKT_ETHHDR_LEN(m->m_pkthdr.l2hlen - ETHER_HDR_LEN) |
                    V_TXPKT_IPHDR_LEN(m->m_pkthdr.l3hlen);
        }
        cpl->ctrl1 = htobe64(ctrl1);
        txq->txcsum++;

        /* Set sequence number in TCP header. */
        tcp = (void *)((char *)eh + m->m_pkthdr.l2hlen + m->m_pkthdr.l3hlen);
        newtcp = *tcp;
        newtcp.th_seq = htonl(tcp_seqno + mtod(m_tls, vm_offset_t));
        copy_to_txd(&txq->eq, (caddr_t)&newtcp, &out, sizeof(newtcp));

        /* Copy rest of TCP header. */
        copy_to_txd(&txq->eq, (caddr_t)(tcp + 1), &out, m->m_len -
            (m->m_pkthdr.l2hlen + m->m_pkthdr.l3hlen + sizeof(*tcp)));

        /* Copy the subset of the TLS header requested. */
        copy_to_txd(&txq->eq, (char *)m_tls->m_epg_hdr +
            mtod(m_tls, vm_offset_t), &out, m_tls->m_len);
        txq->imm_wrs++;

        txq->txpkt_wrs++;

        txq->kern_tls_header++;

        txsd = &txq->sdesc[pidx];
        txsd->m = m;
        txsd->desc_used = ndesc;

        return (ndesc);
}

_Static_assert(sizeof(struct cpl_set_tcb_field) <= EQ_ESIZE,
    "CPL_SET_TCB_FIELD must be smaller than a single TX descriptor");
_Static_assert(W_TCB_SND_UNA_RAW == W_TCB_SND_NXT_RAW,
    "SND_NXT_RAW and SND_UNA_RAW are in different words");

static int
ktls_write_tls_wr(struct tlspcb *tlsp, struct sge_txq *txq, void *dst,
    struct mbuf *m, struct tcphdr *tcp, struct mbuf *m_tls, u_int available,
    tcp_seq tcp_seqno, uint32_t *tsopt, u_int pidx, bool set_l2t_idx)
{
        struct sge_eq *eq = &txq->eq;
        struct tx_sdesc *txsd;
        struct fw_ulptx_wr *wr;
        struct ulp_txpkt *txpkt;
        struct ulptx_sc_memrd *memrd;
        struct ulptx_idata *idata;
        struct cpl_tx_sec_pdu *sec_pdu;
        struct cpl_tx_data *tx_data;
        struct tls_record_layer *hdr;
        char *iv, *out;
        u_int aad_start, aad_stop;
        u_int auth_start, auth_stop, auth_insert;
        u_int cipher_start, cipher_stop, iv_offset;
        u_int imm_len, mss, ndesc, offset, plen, tlen, twr_len, wr_len;
        u_int fields, nsegs, tx_max_offset, tx_max;
        bool first_wr, last_wr, using_scratch;

        ndesc = 0;
        MPASS(tlsp->txq == txq);

        first_wr = (tlsp->prev_seq == 0 && tlsp->prev_ack == 0 &&
            tlsp->prev_win == 0);

        /*
         * Use the per-txq scratch pad if near the end of the ring to
         * simplify handling of wrap-around.  This uses a simple but
         * not quite perfect test of using the scratch buffer if we
         * can't fit a maximal work request in without wrapping.
         */
        using_scratch = (eq->sidx - pidx < SGE_MAX_WR_LEN / EQ_ESIZE);

        /* Locate the TLS header. */
        M_ASSERTEXTPG(m_tls);
        hdr = (void *)m_tls->m_epg_hdr;
        plen = TLS_HEADER_LENGTH + ntohs(hdr->tls_length) - m_tls->m_epg_trllen;

        /* Determine how much of the TLS record to send. */
        tlen = ktls_tcp_payload_length(tlsp, m_tls);
        if (tlen <= m_tls->m_epg_hdrlen) {
                /*
                 * For requests that only want to send the TLS header,
                 * send a tunnelled packet as immediate data.
                 */
#ifdef VERBOSE_TRACES
                CTR3(KTR_CXGBE, "%s: tid %d header-only TLS record %u",
                    __func__, tlsp->tid, (u_int)m_tls->m_epg_seqno);
#endif
                return (ktls_write_tunnel_packet(txq, dst, m, m_tls, available,
                    tcp_seqno, pidx));
        }
        if (tlen < plen) {
                plen = tlen;
                offset = ktls_payload_offset(tlsp, m_tls);
#ifdef VERBOSE_TRACES
                CTR4(KTR_CXGBE, "%s: tid %d short TLS record %u with offset %u",
                    __func__, tlsp->tid, (u_int)m_tls->m_epg_seqno, offset);
#endif
                if (m_tls->m_next == NULL && (tcp_get_flags(tcp) & TH_FIN) != 0) {
                        txq->kern_tls_fin_short++;
#ifdef INVARIANTS
                        panic("%s: FIN on short TLS record", __func__);
#endif
                }
        } else
                offset = 0;

        /*
         * This is the last work request for a given TLS mbuf chain if
         * it is the last mbuf in the chain and FIN is not set.  If
         * FIN is set, then ktls_write_tcp_fin() will write out the
         * last work request.
         */
        last_wr = m_tls->m_next == NULL && (tcp_get_flags(tcp) & TH_FIN) == 0;

        /*
         * The host stack may ask us to not send part of the start of
         * a TLS record.  (For example, the stack might have
         * previously sent a "short" TLS record and might later send
         * down an mbuf that requests to send the remainder of the TLS
         * record.)  The crypto engine must process a TLS record from
         * the beginning if computing a GCM tag or HMAC, so we always
         * send the TLS record from the beginning as input to the
         * crypto engine and via CPL_TX_DATA to TP.  However, TP will
         * drop individual packets after they have been chopped up
         * into MSS-sized chunks if the entire sequence range of those
         * packets is less than SND_UNA.  SND_UNA is computed as
         * TX_MAX - SND_UNA_RAW.  Thus, use the offset stored in
         * m_data to set TX_MAX to the first byte in the TCP sequence
         * space the host actually wants us to send and set
         * SND_UNA_RAW to 0.
         *
         * If the host sends us back to back requests that span the
         * trailer of a single TLS record (first request ends "in" the
         * trailer and second request starts at the next byte but
         * still "in" the trailer), the initial bytes of the trailer
         * that the first request drops will not be retransmitted.  If
         * the host uses the same requests when retransmitting the
         * connection will hang.  To handle this, always transmit the
         * full trailer for a request that begins "in" the trailer
         * (the second request in the example above).  This should
         * also help to avoid retransmits for the common case.
         *
         * A similar condition exists when using CBC for back to back
         * requests that span a single AES block.  The first request
         * will be truncated to end at the end of the previous AES
         * block.  To handle this, always begin transmission at the
         * start of the current AES block.
         */
        tx_max_offset = mtod(m_tls, vm_offset_t);
        if (tx_max_offset > TLS_HEADER_LENGTH + ntohs(hdr->tls_length) -
            m_tls->m_epg_trllen) {
                /* Always send the full trailer. */
                tx_max_offset = TLS_HEADER_LENGTH + ntohs(hdr->tls_length) -
                    m_tls->m_epg_trllen;
        }
        if (tlsp->enc_mode == SCMD_CIPH_MODE_AES_CBC &&
            tx_max_offset > TLS_HEADER_LENGTH) {
                /* Always send all of the first AES block. */
                tx_max_offset = TLS_HEADER_LENGTH +
                    rounddown(tx_max_offset - TLS_HEADER_LENGTH,
                    AES_BLOCK_LEN);
        }
        tx_max = tcp_seqno + tx_max_offset;

        /*
         * Update TCB fields.  Reserve space for the FW_ULPTX_WR header
         * but don't populate it until we know how many field updates
         * are required.
         */
        if (using_scratch)
                wr = (void *)txq->ss;
        else
                wr = dst;
        out = (void *)(wr + 1);
        fields = 0;
        if (set_l2t_idx) {
                KASSERT(m->m_next == m_tls,
                    ("trying to set L2T_IX for subsequent TLS WR"));
#ifdef VERBOSE_TRACES
                CTR3(KTR_CXGBE, "%s: tid %d set L2T_IX to %d", __func__,
                    tlsp->tid, tlsp->l2te->idx);
#endif
                write_set_tcb_field_ulp(tlsp, out, txq, W_TCB_L2T_IX,
                    V_TCB_L2T_IX(M_TCB_L2T_IX), V_TCB_L2T_IX(tlsp->l2te->idx));
                out += roundup2(LEN__SET_TCB_FIELD_ULP, 16);
                fields++;
        }
        if (tsopt != NULL && tlsp->prev_tsecr != ntohl(tsopt[1])) {
                KASSERT(m->m_next == m_tls,
                    ("trying to set T_RTSEQ_RECENT for subsequent TLS WR"));
#ifdef VERBOSE_TRACES
                CTR2(KTR_CXGBE, "%s: tid %d wrote updated T_RTSEQ_RECENT",
                    __func__, tlsp->tid);
#endif
                write_set_tcb_field_ulp(tlsp, out, txq, W_TCB_T_RTSEQ_RECENT,
                    V_TCB_T_RTSEQ_RECENT(M_TCB_T_RTSEQ_RECENT),
                    V_TCB_T_RTSEQ_RECENT(ntohl(tsopt[1])));
                out += roundup2(LEN__SET_TCB_FIELD_ULP, 16);
                fields++;

                tlsp->prev_tsecr = ntohl(tsopt[1]);
        }

        if (first_wr || tlsp->prev_seq != tx_max) {
                KASSERT(m->m_next == m_tls,
                    ("trying to set TX_MAX for subsequent TLS WR"));
#ifdef VERBOSE_TRACES
                CTR4(KTR_CXGBE,
                    "%s: tid %d setting TX_MAX to %u (tcp_seqno %u)",
                    __func__, tlsp->tid, tx_max, tcp_seqno);
#endif
                write_set_tcb_field_ulp(tlsp, out, txq, W_TCB_TX_MAX,
                    V_TCB_TX_MAX(M_TCB_TX_MAX), V_TCB_TX_MAX(tx_max));
                out += roundup2(LEN__SET_TCB_FIELD_ULP, 16);
                fields++;
        }

        /*
         * If there is data to drop at the beginning of this TLS
         * record or if this is a retransmit,
         * reset SND_UNA_RAW to 0 so that SND_UNA == TX_MAX.
         */
        if (tlsp->prev_seq != tx_max || mtod(m_tls, vm_offset_t) != 0) {
                KASSERT(m->m_next == m_tls,
                    ("trying to clear SND_UNA_RAW for subsequent TLS WR"));
#ifdef VERBOSE_TRACES
                CTR2(KTR_CXGBE, "%s: tid %d clearing SND_UNA_RAW", __func__,
                    tlsp->tid);
#endif
                write_set_tcb_field_ulp(tlsp, out, txq, W_TCB_SND_UNA_RAW,
                    V_TCB_SND_UNA_RAW(M_TCB_SND_UNA_RAW),
                    V_TCB_SND_UNA_RAW(0));
                out += roundup2(LEN__SET_TCB_FIELD_ULP, 16);
                fields++;
        }

        /*
         * Store the expected sequence number of the next byte after
         * this record.
         */
        tlsp->prev_seq = tcp_seqno + tlen;

        if (first_wr || tlsp->prev_ack != ntohl(tcp->th_ack)) {
                KASSERT(m->m_next == m_tls,
                    ("trying to set RCV_NXT for subsequent TLS WR"));
                write_set_tcb_field_ulp(tlsp, out, txq, W_TCB_RCV_NXT,
                    V_TCB_RCV_NXT(M_TCB_RCV_NXT),
                    V_TCB_RCV_NXT(ntohl(tcp->th_ack)));
                out += roundup2(LEN__SET_TCB_FIELD_ULP, 16);
                fields++;

                tlsp->prev_ack = ntohl(tcp->th_ack);
        }

        if (first_wr || tlsp->prev_win != ntohs(tcp->th_win)) {
                KASSERT(m->m_next == m_tls,
                    ("trying to set RCV_WND for subsequent TLS WR"));
                write_set_tcb_field_ulp(tlsp, out, txq, W_TCB_RCV_WND,
                    V_TCB_RCV_WND(M_TCB_RCV_WND),
                    V_TCB_RCV_WND(ntohs(tcp->th_win)));
                out += roundup2(LEN__SET_TCB_FIELD_ULP, 16);
                fields++;

                tlsp->prev_win = ntohs(tcp->th_win);
        }

        /* Use cached value for first record in chain. */
        if (m->m_next == m_tls)
                nsegs = mbuf_nsegs(m);
        else
                nsegs = sglist_count_mbuf_epg(m_tls, m_tls->m_epg_hdrlen +
                    offset, plen - (m_tls->m_epg_hdrlen + offset));

        /* Calculate the size of the TLS work request. */
        twr_len = ktls_base_wr_size(tlsp);

        imm_len = 0;
        if (offset == 0)
                imm_len += m_tls->m_epg_hdrlen;
        if (plen == tlen)
                imm_len += AES_BLOCK_LEN;
        twr_len += roundup2(imm_len, 16);
        twr_len += ktls_sgl_size(nsegs);

        /*
         * If any field updates were required, determine if they can
         * be included in the TLS work request.  If not, use the
         * FW_ULPTX_WR work request header at 'wr' as a dedicated work
         * request for the field updates and start a new work request
         * for the TLS work request afterward.
         */
        if (fields != 0) {
                wr_len = fields * roundup2(LEN__SET_TCB_FIELD_ULP, 16);
                if (twr_len + wr_len <= SGE_MAX_WR_LEN &&
                    tlsp->sc->tlst.combo_wrs) {
                        wr_len += twr_len;
                        txpkt = (void *)out;
                } else {
                        wr_len += sizeof(*wr);
                        wr->op_to_compl = htobe32(V_FW_WR_OP(FW_ULPTX_WR));
                        wr->flowid_len16 = htobe32(F_FW_ULPTX_WR_DATA |
                            V_FW_WR_LEN16(wr_len / 16));
                        wr->cookie = 0;

                        /*
                         * If we were using scratch space, copy the
                         * field updates work request to the ring.
                         */
                        if (using_scratch) {
                                out = dst;
                                copy_to_txd(eq, txq->ss, &out, wr_len);
                        }

                        ndesc = howmany(wr_len, EQ_ESIZE);
                        MPASS(ndesc <= available);

                        txq->raw_wrs++;
                        txsd = &txq->sdesc[pidx];
                        txsd->m = NULL;
                        txsd->desc_used = ndesc;
                        IDXINCR(pidx, ndesc, eq->sidx);
                        dst = &eq->desc[pidx];

                        /*
                         * Determine if we should use scratch space
                         * for the TLS work request based on the
                         * available space after advancing pidx for
                         * the field updates work request.
                         */
                        wr_len = twr_len;
                        using_scratch = (eq->sidx - pidx <
                            howmany(wr_len, EQ_ESIZE));
                        if (using_scratch)
                                wr = (void *)txq->ss;
                        else
                                wr = dst;
                        txpkt = (void *)(wr + 1);
                }
        } else {
                wr_len = twr_len;
                txpkt = (void *)out;
        }

        wr_len = roundup2(wr_len, 16);
        MPASS(ndesc + howmany(wr_len, EQ_ESIZE) <= available);

        /* FW_ULPTX_WR */
        wr->op_to_compl = htobe32(V_FW_WR_OP(FW_ULPTX_WR));
        wr->flowid_len16 = htobe32(F_FW_ULPTX_WR_DATA |
            V_FW_WR_LEN16(wr_len / 16));
        wr->cookie = 0;

        /* ULP_TXPKT */
        txpkt->cmd_dest = htobe32(V_ULPTX_CMD(ULP_TX_PKT) |
            V_ULP_TXPKT_DATAMODIFY(0) |
            V_ULP_TXPKT_CHANNELID(tlsp->vi->pi->port_id) | V_ULP_TXPKT_DEST(0) |
            V_ULP_TXPKT_FID(txq->eq.cntxt_id) | V_ULP_TXPKT_RO(1));
        txpkt->len = htobe32(howmany(twr_len - sizeof(*wr), 16));

        /* ULPTX_IDATA sub-command */
        idata = (void *)(txpkt + 1);
        idata->cmd_more = htobe32(V_ULPTX_CMD(ULP_TX_SC_IMM) |
            V_ULP_TX_SC_MORE(1));
        idata->len = sizeof(struct cpl_tx_sec_pdu);

        /*
         * The key context, CPL_TX_DATA, and immediate data are part
         * of this ULPTX_IDATA when using an inline key.  When reading
         * the key from memory, the CPL_TX_DATA and immediate data are
         * part of a separate ULPTX_IDATA.
         */
        if (tlsp->inline_key)
                idata->len += tlsp->tx_key_info_size +
                    sizeof(struct cpl_tx_data) + imm_len;
        idata->len = htobe32(idata->len);

        /* CPL_TX_SEC_PDU */
        sec_pdu = (void *)(idata + 1);

        /*
         * For short records, AAD is counted as header data in SCMD0,
         * the IV is next followed by a cipher region for the payload.
         */
        if (plen == tlen) {
                aad_start = 0;
                aad_stop = 0;
                iv_offset = 1;
                auth_start = 0;
                auth_stop = 0;
                auth_insert = 0;
                cipher_start = AES_BLOCK_LEN + 1;
                cipher_stop = 0;

                sec_pdu->pldlen = htobe32(16 + plen -
                    (m_tls->m_epg_hdrlen + offset));

                /* These two flits are actually a CPL_TLS_TX_SCMD_FMT. */
                sec_pdu->seqno_numivs = tlsp->scmd0_short.seqno_numivs;
                sec_pdu->ivgen_hdrlen = htobe32(
                    tlsp->scmd0_short.ivgen_hdrlen |
                    V_SCMD_HDR_LEN(offset == 0 ? m_tls->m_epg_hdrlen : 0));

                txq->kern_tls_short++;
        } else {
                /*
                 * AAD is TLS header.  IV is after AAD.  The cipher region
                 * starts after the IV.  See comments in ccr_authenc() and
                 * ccr_gmac() in t4_crypto.c regarding cipher and auth
                 * start/stop values.
                 */
                aad_start = 1;
                aad_stop = TLS_HEADER_LENGTH;
                iv_offset = TLS_HEADER_LENGTH + 1;
                cipher_start = m_tls->m_epg_hdrlen + 1;
                if (tlsp->enc_mode == SCMD_CIPH_MODE_AES_GCM) {
                        cipher_stop = 0;
                        auth_start = cipher_start;
                        auth_stop = 0;
                        auth_insert = 0;
                } else {
                        cipher_stop = 0;
                        auth_start = cipher_start;
                        auth_stop = 0;
                        auth_insert = 0;
                }

                sec_pdu->pldlen = htobe32(plen);

                /* These two flits are actually a CPL_TLS_TX_SCMD_FMT. */
                sec_pdu->seqno_numivs = tlsp->scmd0.seqno_numivs;
                sec_pdu->ivgen_hdrlen = tlsp->scmd0.ivgen_hdrlen;

                if (mtod(m_tls, vm_offset_t) == 0)
                        txq->kern_tls_full++;
                else
                        txq->kern_tls_partial++;
        }
        sec_pdu->op_ivinsrtofst = htobe32(
            V_CPL_TX_SEC_PDU_OPCODE(CPL_TX_SEC_PDU) |
            V_CPL_TX_SEC_PDU_CPLLEN(2) | V_CPL_TX_SEC_PDU_PLACEHOLDER(0) |
            V_CPL_TX_SEC_PDU_IVINSRTOFST(iv_offset));
        sec_pdu->aadstart_cipherstop_hi = htobe32(
            V_CPL_TX_SEC_PDU_AADSTART(aad_start) |
            V_CPL_TX_SEC_PDU_AADSTOP(aad_stop) |
            V_CPL_TX_SEC_PDU_CIPHERSTART(cipher_start) |
            V_CPL_TX_SEC_PDU_CIPHERSTOP_HI(cipher_stop >> 4));
        sec_pdu->cipherstop_lo_authinsert = htobe32(
            V_CPL_TX_SEC_PDU_CIPHERSTOP_LO(cipher_stop & 0xf) |
            V_CPL_TX_SEC_PDU_AUTHSTART(auth_start) |
            V_CPL_TX_SEC_PDU_AUTHSTOP(auth_stop) |
            V_CPL_TX_SEC_PDU_AUTHINSERT(auth_insert));

        sec_pdu->scmd1 = htobe64(m_tls->m_epg_seqno);

        /* Key context */
        out = (void *)(sec_pdu + 1);
        if (tlsp->inline_key) {
                memcpy(out, &tlsp->keyctx, tlsp->tx_key_info_size);
                out += tlsp->tx_key_info_size;
        } else {
                /* ULPTX_SC_MEMRD to read key context. */
                memrd = (void *)out;
                memrd->cmd_to_len = htobe32(V_ULPTX_CMD(ULP_TX_SC_MEMRD) |
                    V_ULP_TX_SC_MORE(1) |
                    V_ULPTX_LEN16(tlsp->tx_key_info_size >> 4));
                memrd->addr = htobe32(tlsp->tx_key_addr >> 5);

                /* ULPTX_IDATA for CPL_TX_DATA and TLS header. */
                idata = (void *)(memrd + 1);
                idata->cmd_more = htobe32(V_ULPTX_CMD(ULP_TX_SC_IMM) |
                    V_ULP_TX_SC_MORE(1));
                idata->len = htobe32(sizeof(struct cpl_tx_data) + imm_len);

                out = (void *)(idata + 1);
        }

        /* CPL_TX_DATA */
        tx_data = (void *)out;
        OPCODE_TID(tx_data) = htonl(MK_OPCODE_TID(CPL_TX_DATA, tlsp->tid));
        if (m->m_pkthdr.csum_flags & CSUM_TSO) {
                mss = m->m_pkthdr.tso_segsz;
                tlsp->prev_mss = mss;
        } else if (tlsp->prev_mss != 0)
                mss = tlsp->prev_mss;
        else
                mss = if_getmtu(tlsp->vi->ifp) -
                    (m->m_pkthdr.l3hlen + m->m_pkthdr.l4hlen);
        if (offset == 0) {
                tx_data->len = htobe32(V_TX_DATA_MSS(mss) | V_TX_LENGTH(tlen));
                tx_data->rsvd = htobe32(tcp_seqno);
        } else {
                tx_data->len = htobe32(V_TX_DATA_MSS(mss) |
                    V_TX_LENGTH(tlen - (m_tls->m_epg_hdrlen + offset)));
                tx_data->rsvd = htobe32(tcp_seqno + m_tls->m_epg_hdrlen + offset);
        }
        tx_data->flags = htobe32(F_TX_BYPASS);
        if (last_wr && tcp_get_flags(tcp) & TH_PUSH)
                tx_data->flags |= htobe32(F_TX_PUSH | F_TX_SHOVE);

        /* Populate the TLS header */
        out = (void *)(tx_data + 1);
        if (offset == 0) {
                memcpy(out, m_tls->m_epg_hdr, m_tls->m_epg_hdrlen);
                out += m_tls->m_epg_hdrlen;
        }

        /* AES IV for a short record. */
        if (plen == tlen) {
                iv = out;
                if (tlsp->enc_mode == SCMD_CIPH_MODE_AES_GCM) {
                        memcpy(iv, tlsp->keyctx.u.txhdr.txsalt, SALT_SIZE);
                        memcpy(iv + 4, hdr + 1, 8);
                        *(uint32_t *)(iv + 12) = htobe32(2 +
                            offset / AES_BLOCK_LEN);
                } else
                        memcpy(iv, hdr + 1, AES_BLOCK_LEN);
                out += AES_BLOCK_LEN;
        }

        if (imm_len % 16 != 0) {
                if (imm_len % 8 != 0) {
                        /* Zero pad to an 8-byte boundary. */
                        memset(out, 0, 8 - (imm_len % 8));
                        out += 8 - (imm_len % 8);
                }

                /*
                 * Insert a ULP_TX_SC_NOOP if needed so the SGL is
                 * 16-byte aligned.
                 */
                if (imm_len % 16 <= 8) {
                        idata = (void *)out;
                        idata->cmd_more = htobe32(V_ULPTX_CMD(ULP_TX_SC_NOOP));
                        idata->len = htobe32(0);
                        out = (void *)(idata + 1);
                }
        }

        /* SGL for record payload */
        sglist_reset(txq->gl);
        if (sglist_append_mbuf_epg(txq->gl, m_tls, m_tls->m_epg_hdrlen + offset,
            plen - (m_tls->m_epg_hdrlen + offset)) != 0) {
#ifdef INVARIANTS
                panic("%s: failed to append sglist", __func__);
#endif
        }
        write_gl_to_buf(txq->gl, out);

        if (using_scratch) {
                out = dst;
                copy_to_txd(eq, txq->ss, &out, wr_len);
        }

        ndesc += howmany(wr_len, EQ_ESIZE);
        MPASS(ndesc <= available);

        txq->kern_tls_records++;
        txq->kern_tls_octets += tlen - mtod(m_tls, vm_offset_t);
        if (mtod(m_tls, vm_offset_t) != 0) {
                if (offset == 0)
                        txq->kern_tls_waste += mtod(m_tls, vm_offset_t);
                else
                        txq->kern_tls_waste += mtod(m_tls, vm_offset_t) -
                            (m_tls->m_epg_hdrlen + offset);
        }

        txsd = &txq->sdesc[pidx];
        if (last_wr)
                txsd->m = m;
        else
                txsd->m = NULL;
        txsd->desc_used = howmany(wr_len, EQ_ESIZE);

        return (ndesc);
}

static int
ktls_write_tcp_fin(struct sge_txq *txq, void *dst, struct mbuf *m,
    u_int available, tcp_seq tcp_seqno, u_int pidx)
{
        struct tx_sdesc *txsd;
        struct fw_eth_tx_pkt_wr *wr;
        struct cpl_tx_pkt_core *cpl;
        uint32_t ctrl;
        uint64_t ctrl1;
        int len16, ndesc, pktlen;
        struct ether_header *eh;
        struct ip *ip, newip;
        struct ip6_hdr *ip6, newip6;
        struct tcphdr *tcp, newtcp;
        caddr_t out;

        TXQ_LOCK_ASSERT_OWNED(txq);
        M_ASSERTPKTHDR(m);

        wr = dst;
        pktlen = m->m_len;
        ctrl = sizeof(struct cpl_tx_pkt_core) + pktlen;
        len16 = howmany(sizeof(struct fw_eth_tx_pkt_wr) + ctrl, 16);
        ndesc = tx_len16_to_desc(len16);
        MPASS(ndesc <= available);

        /* Firmware work request header */
        wr->op_immdlen = htobe32(V_FW_WR_OP(FW_ETH_TX_PKT_WR) |
            V_FW_ETH_TX_PKT_WR_IMMDLEN(ctrl));

        ctrl = V_FW_WR_LEN16(len16);
        wr->equiq_to_len16 = htobe32(ctrl);
        wr->r3 = 0;

        cpl = (void *)(wr + 1);

        /* CPL header */
        cpl->ctrl0 = txq->cpl_ctrl0;
        cpl->pack = 0;
        cpl->len = htobe16(pktlen);

        out = (void *)(cpl + 1);

        /* Copy over Ethernet header. */
        eh = mtod(m, struct ether_header *);
        copy_to_txd(&txq->eq, (caddr_t)eh, &out, m->m_pkthdr.l2hlen);

        /* Fixup length in IP header and copy out. */
        if (ntohs(eh->ether_type) == ETHERTYPE_IP) {
                ip = (void *)((char *)eh + m->m_pkthdr.l2hlen);
                newip = *ip;
                newip.ip_len = htons(pktlen - m->m_pkthdr.l2hlen);
                copy_to_txd(&txq->eq, (caddr_t)&newip, &out, sizeof(newip));
                if (m->m_pkthdr.l3hlen > sizeof(*ip))
                        copy_to_txd(&txq->eq, (caddr_t)(ip + 1), &out,
                            m->m_pkthdr.l3hlen - sizeof(*ip));
                ctrl1 = V_TXPKT_CSUM_TYPE(TX_CSUM_TCPIP) |
                    V_T6_TXPKT_ETHHDR_LEN(m->m_pkthdr.l2hlen - ETHER_HDR_LEN) |
                    V_TXPKT_IPHDR_LEN(m->m_pkthdr.l3hlen);
        } else {
                ip6 = (void *)((char *)eh + m->m_pkthdr.l2hlen);
                newip6 = *ip6;
                newip6.ip6_plen = htons(pktlen - m->m_pkthdr.l2hlen -
                    sizeof(*ip6));
                copy_to_txd(&txq->eq, (caddr_t)&newip6, &out, sizeof(newip6));
                MPASS(m->m_pkthdr.l3hlen == sizeof(*ip6));
                ctrl1 = V_TXPKT_CSUM_TYPE(TX_CSUM_TCPIP6) |
                    V_T6_TXPKT_ETHHDR_LEN(m->m_pkthdr.l2hlen - ETHER_HDR_LEN) |
                    V_TXPKT_IPHDR_LEN(m->m_pkthdr.l3hlen);
        }
        cpl->ctrl1 = htobe64(ctrl1);
        txq->txcsum++;

        /* Set sequence number in TCP header. */
        tcp = (void *)((char *)eh + m->m_pkthdr.l2hlen + m->m_pkthdr.l3hlen);
        newtcp = *tcp;
        newtcp.th_seq = htonl(tcp_seqno);
        copy_to_txd(&txq->eq, (caddr_t)&newtcp, &out, sizeof(newtcp));

        /* Copy rest of packet. */
        copy_to_txd(&txq->eq, (caddr_t)(tcp + 1), &out, m->m_len -
            (m->m_pkthdr.l2hlen + m->m_pkthdr.l3hlen + sizeof(*tcp)));
        txq->imm_wrs++;

        txq->txpkt_wrs++;

        txq->kern_tls_fin++;

        txsd = &txq->sdesc[pidx];
        txsd->m = m;
        txsd->desc_used = ndesc;

        return (ndesc);
}

int
t6_ktls_write_wr(struct sge_txq *txq, void *dst, struct mbuf *m,
    u_int available)
{
        struct sge_eq *eq = &txq->eq;
        struct tx_sdesc *txsd;
        struct tlspcb *tlsp;
        struct tcphdr *tcp;
        struct mbuf *m_tls;
        struct ether_header *eh;
        tcp_seq tcp_seqno;
        u_int ndesc, pidx, totdesc;
        uint16_t vlan_tag;
        bool has_fin, set_l2t_idx;
        void *tsopt;

        M_ASSERTPKTHDR(m);
        MPASS(m->m_pkthdr.snd_tag != NULL);
        tlsp = mst_to_tls(m->m_pkthdr.snd_tag);

        totdesc = 0;
        eh = mtod(m, struct ether_header *);
        tcp = (struct tcphdr *)((char *)eh + m->m_pkthdr.l2hlen +
            m->m_pkthdr.l3hlen);
        pidx = eq->pidx;
        has_fin = (tcp_get_flags(tcp) & TH_FIN) != 0;

        /*
         * If this TLS record has a FIN, then we will send any
         * requested options as part of the FIN packet.
         */
        if (!has_fin && ktls_has_tcp_options(tcp)) {
                ndesc = ktls_write_tcp_options(txq, dst, m, available, pidx);
                totdesc += ndesc;
                IDXINCR(pidx, ndesc, eq->sidx);
                dst = &eq->desc[pidx];
#ifdef VERBOSE_TRACES
                CTR2(KTR_CXGBE, "%s: tid %d wrote TCP options packet", __func__,
                    tlsp->tid);
#endif
        }

        /*
         * Allocate a new L2T entry if necessary.  This may write out
         * a work request to the txq.
         */
        if (m->m_flags & M_VLANTAG)
                vlan_tag = m->m_pkthdr.ether_vtag;
        else
                vlan_tag = 0xfff;
        set_l2t_idx = false;
        if (tlsp->l2te == NULL || tlsp->l2te->vlan != vlan_tag ||
            memcmp(tlsp->l2te->dmac, eh->ether_dhost, ETHER_ADDR_LEN) != 0) {
                set_l2t_idx = true;
                if (tlsp->l2te)
                        t4_l2t_release(tlsp->l2te);
                tlsp->l2te = t4_l2t_alloc_tls(tlsp->sc, txq, dst, &ndesc,
                    vlan_tag, tlsp->vi->pi->hw_port, eh->ether_dhost);
                if (tlsp->l2te == NULL)
                        CXGBE_UNIMPLEMENTED("failed to allocate TLS L2TE");
                if (ndesc != 0) {
                        MPASS(ndesc <= available - totdesc);

                        txq->raw_wrs++;
                        txsd = &txq->sdesc[pidx];
                        txsd->m = NULL;
                        txsd->desc_used = ndesc;
                        totdesc += ndesc;
                        IDXINCR(pidx, ndesc, eq->sidx);
                        dst = &eq->desc[pidx];
                }
        }

        /*
         * Iterate over each TLS record constructing a work request
         * for that record.
         */
        for (m_tls = m->m_next; m_tls != NULL; m_tls = m_tls->m_next) {
                MPASS(m_tls->m_flags & M_EXTPG);

                /*
                 * Determine the initial TCP sequence number for this
                 * record.
                 */
                tsopt = NULL;
                if (m_tls == m->m_next) {
                        tcp_seqno = ntohl(tcp->th_seq) -
                            mtod(m_tls, vm_offset_t);
                        if (tlsp->using_timestamps)
                                tsopt = ktls_find_tcp_timestamps(tcp);
                } else {
                        MPASS(mtod(m_tls, vm_offset_t) == 0);
                        tcp_seqno = tlsp->prev_seq;
                }

                ndesc = ktls_write_tls_wr(tlsp, txq, dst, m, tcp, m_tls,
                    available - totdesc, tcp_seqno, tsopt, pidx, set_l2t_idx);
                totdesc += ndesc;
                IDXINCR(pidx, ndesc, eq->sidx);
                dst = &eq->desc[pidx];

                /* Only need to set the L2T index once. */
                set_l2t_idx = false;
        }

        if (has_fin) {
                /*
                 * If the TCP header for this chain has FIN sent, then
                 * explicitly send a packet that has FIN set.  This
                 * will also have PUSH set if requested.  This assumes
                 * we sent at least one TLS record work request and
                 * uses the TCP sequence number after that reqeust as
                 * the sequence number for the FIN packet.
                 */
                ndesc = ktls_write_tcp_fin(txq, dst, m, available,
                    tlsp->prev_seq, pidx);
                totdesc += ndesc;
        }

        MPASS(totdesc <= available);
        return (totdesc);
}

static void
t6_tls_tag_free(struct m_snd_tag *mst)
{
        struct adapter *sc;
        struct tlspcb *tlsp;

        tlsp = mst_to_tls(mst);
        sc = tlsp->sc;

        CTR2(KTR_CXGBE, "%s: tid %d", __func__, tlsp->tid);

        if (tlsp->l2te)
                t4_l2t_release(tlsp->l2te);
        if (tlsp->tid >= 0)
                release_tid(sc, tlsp->tid, tlsp->ctrlq);
        if (tlsp->ce)
                t4_release_clip_entry(sc, tlsp->ce);
        if (tlsp->tx_key_addr >= 0)
                t4_free_tls_keyid(sc, tlsp->tx_key_addr);

        zfree(tlsp, M_CXGBE);
}

void
t6_ktls_modload(void)
{

        t4_register_shared_cpl_handler(CPL_ACT_OPEN_RPL, ktls_act_open_rpl,
            CPL_COOKIE_KERN_TLS);
}

void
t6_ktls_modunload(void)
{

        t4_register_shared_cpl_handler(CPL_ACT_OPEN_RPL, NULL,
            CPL_COOKIE_KERN_TLS);
}

#else

int
t6_tls_tag_alloc(if_t ifp, union if_snd_tag_alloc_params *params,
    struct m_snd_tag **pt)
{
        return (ENXIO);
}

int
t6_ktls_parse_pkt(struct mbuf *m)
{
        return (EINVAL);
}

int
t6_ktls_write_wr(struct sge_txq *txq, void *dst, struct mbuf *m,
    u_int available)
{
        panic("can't happen");
}

void
t6_ktls_modload(void)
{
}

void
t6_ktls_modunload(void)
{
}

#endif