root/usr/src/uts/common/sys/nxge/nxge_txdma.h
/*
 * CDDL HEADER START
 *
 * The contents of this file are subject to the terms of the
 * Common Development and Distribution License (the "License").
 * You may not use this file except in compliance with the License.
 *
 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
 * or http://www.opensolaris.org/os/licensing.
 * See the License for the specific language governing permissions
 * and limitations under the License.
 *
 * When distributing Covered Code, include this CDDL HEADER in each
 * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
 * If applicable, add the following below this CDDL HEADER, with the
 * fields enclosed by brackets "[]" replaced with your own identifying
 * information: Portions Copyright [yyyy] [name of copyright owner]
 *
 * CDDL HEADER END
 */
/*
 * Copyright 2008 Sun Microsystems, Inc.  All rights reserved.
 * Use is subject to license terms.
 */

#ifndef _SYS_NXGE_NXGE_TXDMA_H
#define _SYS_NXGE_NXGE_TXDMA_H

#ifdef  __cplusplus
extern "C" {
#endif

#include <sys/taskq.h>
#include <sys/nxge/nxge_txdma_hw.h>
#include <npi_txdma.h>

#define TXDMA_PORT_BITMAP(nxgep)                (nxgep->pt_config.tx_dma_map)

#define TXDMA_RECLAIM_PENDING_DEFAULT           64
#define TX_FULL_MARK                            3

/*
 * Transmit load balancing definitions.
 */
#define NXGE_TX_LB_TCPUDP                       0       /* default policy */
#define NXGE_TX_LB_HASH                         1       /* from the hint data */
#define NXGE_TX_LB_DEST_MAC                     2       /* Dest. MAC */

/*
 * Descriptor ring empty:
 *              (1) head index is equal to tail index.
 *              (2) wrapped around bits are the same.
 * Descriptor ring full:
 *              (1) head index is equal to tail index.
 *              (2) wrapped around bits are different.
 *
 */
#define TXDMA_RING_EMPTY(head, head_wrap, tail, tail_wrap)      \
        ((head == tail && head_wrap == tail_wrap) ? B_TRUE : B_FALSE)

#define TXDMA_RING_FULL(head, head_wrap, tail, tail_wrap)       \
        ((head == tail && head_wrap != tail_wrap) ? B_TRUE : B_FALSE)

#define TXDMA_DESC_NEXT_INDEX(index, entries, wrap_mask) \
                        ((index + entries) & wrap_mask)

#define TXDMA_DRR_WEIGHT_DEFAULT        0x001f

typedef enum {
        NXGE_USE_SERIAL = 0,
        NXGE_USE_START,
} nxge_tx_mode_t;

typedef struct _tx_msg_t {
        nxge_os_block_mv_t      flags;          /* DMA, BCOPY, DVMA (?) */
        nxge_os_dma_common_t    buf_dma;        /* premapped buffer blocks */
        nxge_os_dma_handle_t    buf_dma_handle; /* premapped buffer handle */
        nxge_os_dma_handle_t    dma_handle;     /* DMA handle for normal send */
        nxge_os_dma_handle_t    dvma_handle;    /* Fast DVMA  handle */
        struct _tx_msg_t        *nextp;

        p_mblk_t                tx_message;
        uint32_t                tx_msg_size;
        size_t                  bytes_used;
        int                     head;
        int                     tail;
} tx_msg_t, *p_tx_msg_t;

/*
 * TX  Statistics.
 */
typedef struct _nxge_tx_ring_stats_t {
        uint64_t        opackets;
        uint64_t        obytes;
        uint64_t        oerrors;

        uint32_t        tx_inits;
        uint32_t        tx_no_buf;

        uint32_t                mbox_err;
        uint32_t                pkt_size_err;
        uint32_t                tx_ring_oflow;
        uint32_t                pre_buf_par_err;
        uint32_t                nack_pref;
        uint32_t                nack_pkt_rd;
        uint32_t                conf_part_err;
        uint32_t                pkt_part_err;
        uint32_t                tx_starts;
        uint32_t                tx_nocanput;
        uint32_t                tx_msgdup_fail;
        uint32_t                tx_allocb_fail;
        uint32_t                tx_no_desc;
        uint32_t                tx_dma_bind_fail;
        uint32_t                tx_uflo;

        uint32_t                tx_hdr_pkts;
        uint32_t                tx_ddi_pkts;
        uint32_t                tx_dvma_pkts;

        uint32_t                tx_max_pend;
        uint32_t                tx_jumbo_pkts;

        txdma_ring_errlog_t     errlog;
} nxge_tx_ring_stats_t, *p_nxge_tx_ring_stats_t;

typedef struct _tx_ring_t {
        nxge_os_dma_common_t    tdc_desc;
        struct _nxge_t          *nxgep;
        p_tx_msg_t              tx_msg_ring;
        uint32_t                tnblocks;
        tx_rng_cfig_t           tx_ring_cfig;
        tx_ring_hdl_t           tx_ring_hdl;
        tx_ring_kick_t          tx_ring_kick;
        tx_cs_t                 tx_cs;
        tx_dma_ent_msk_t        tx_evmask;
        txdma_mbh_t             tx_mbox_mbh;
        txdma_mbl_t             tx_mbox_mbl;
        log_page_vld_t          page_valid;
        log_page_mask_t         page_mask_1;
        log_page_mask_t         page_mask_2;
        log_page_value_t        page_value_1;
        log_page_value_t        page_value_2;
        log_page_relo_t         page_reloc_1;
        log_page_relo_t         page_reloc_2;
        log_page_hdl_t          page_hdl;
        txc_dma_max_burst_t     max_burst;
        boolean_t               cfg_set;
#define NXGE_TX_RING_ONLINE     0x00
#define NXGE_TX_RING_OFFLINING  0x01
#define NXGE_TX_RING_OFFLINED   0x02
        uint32_t                tx_ring_offline;
        boolean_t               tx_ring_busy;

        nxge_os_mutex_t         lock;
        mac_ring_handle_t       tx_ring_handle;
        ddi_taskq_t             *taskq;
        uint16_t                index;
        uint16_t                tdc;
        struct nxge_tdc_cfg     *tdc_p;
        int                     tx_ring_size;
        uint32_t                num_chunks;

        uint_t                  tx_wrap_mask;
        uint_t                  rd_index;
        uint_t                  wr_index;
        boolean_t               wr_index_wrap;
        tx_ring_hdl_t           ring_head;
        tx_ring_kick_t          ring_kick_tail;
        txdma_mailbox_t         tx_mbox;

        int                     descs_pending;
        boolean_t               queueing;

        nxge_os_mutex_t         sq_lock;
        p_mblk_t                head;
        p_mblk_t                tail;

        uint16_t                ldg_group_id;
        p_nxge_tx_ring_stats_t tdc_stats;

        nxge_os_mutex_t         dvma_lock;
        uint_t                  dvma_wr_index;
        uint_t                  dvma_rd_index;
        uint_t                  dvma_pending;
        uint_t                  dvma_available;
        uint_t                  dvma_wrap_mask;

        nxge_os_dma_handle_t    *dvma_ring;

#if     defined(sun4v) && defined(NIU_LP_WORKAROUND)
        uint64_t                hv_tx_buf_base_ioaddr_pp;
        uint64_t                hv_tx_buf_ioaddr_size;
        uint64_t                hv_tx_cntl_base_ioaddr_pp;
        uint64_t                hv_tx_cntl_ioaddr_size;
        boolean_t               hv_set;
#endif
} tx_ring_t, *p_tx_ring_t;


/* Transmit Mailbox */
typedef struct _tx_mbox_t {
        nxge_os_mutex_t         lock;
        uint16_t                index;
        struct _nxge_t          *nxgep;
        uint16_t                tdc;
        nxge_os_dma_common_t    tx_mbox;
        txdma_mbl_t             tx_mbox_l;
        txdma_mbh_t             tx_mbox_h;
} tx_mbox_t, *p_tx_mbox_t;

typedef struct _tx_rings_t {
        p_tx_ring_t             *rings;
        boolean_t               txdesc_allocated;
        uint32_t                ndmas;
        nxge_os_dma_common_t    tdc_dma;
        nxge_os_dma_common_t    tdc_mbox;
} tx_rings_t, *p_tx_rings_t;


typedef struct _tx_mbox_areas_t {
        p_tx_mbox_t             *txmbox_areas_p;
        boolean_t               txmbox_allocated;
} tx_mbox_areas_t, *p_tx_mbox_areas_t;

/*
 * Transmit prototypes.
 */
nxge_status_t nxge_init_txdma_channels(p_nxge_t);
void nxge_uninit_txdma_channels(p_nxge_t);

nxge_status_t nxge_init_txdma_channel(p_nxge_t, int);
void nxge_uninit_txdma_channel(p_nxge_t, int);

void nxge_setup_dma_common(p_nxge_dma_common_t, p_nxge_dma_common_t,
                uint32_t, uint32_t);
nxge_status_t nxge_reset_txdma_channel(p_nxge_t, uint16_t,
        uint64_t);
nxge_status_t nxge_init_txdma_channel_event_mask(p_nxge_t,
        uint16_t, p_tx_dma_ent_msk_t);
nxge_status_t nxge_init_txdma_channel_cntl_stat(p_nxge_t,
        uint16_t, uint64_t);
nxge_status_t nxge_enable_txdma_channel(p_nxge_t, uint16_t,
        p_tx_ring_t, p_tx_mbox_t);

p_mblk_t nxge_tx_pkt_header_reserve(p_mblk_t, uint8_t *);
int nxge_tx_pkt_nmblocks(p_mblk_t, int *);
boolean_t nxge_txdma_reclaim(p_nxge_t, p_tx_ring_t, int);

void nxge_fill_tx_hdr(p_mblk_t, boolean_t, boolean_t,
        int, uint8_t, p_tx_pkt_hdr_all_t, t_uscalar_t, t_uscalar_t);

nxge_status_t nxge_txdma_hw_mode(p_nxge_t, boolean_t);
void nxge_hw_start_tx(p_nxge_t);
void nxge_txdma_stop(p_nxge_t);
void nxge_txdma_stop_start(p_nxge_t);
void nxge_fixup_txdma_rings(p_nxge_t);
void nxge_txdma_hw_kick(p_nxge_t);
void nxge_txdma_fix_channel(p_nxge_t, uint16_t);
void nxge_txdma_fixup_channel(p_nxge_t, p_tx_ring_t,
        uint16_t);
void nxge_txdma_hw_kick_channel(p_nxge_t, p_tx_ring_t,
        uint16_t);

void nxge_txdma_regs_dump(p_nxge_t, int);
void nxge_txdma_regs_dump_channels(p_nxge_t);

void nxge_check_tx_hang(p_nxge_t);
void nxge_fixup_hung_txdma_rings(p_nxge_t);

void nxge_reclaim_rings(p_nxge_t);
int nxge_txdma_channel_hung(p_nxge_t,
        p_tx_ring_t tx_ring_p, uint16_t);
int nxge_txdma_hung(p_nxge_t);
int nxge_txdma_stop_inj_err(p_nxge_t, int);
void nxge_txdma_inject_err(p_nxge_t, uint32_t, uint8_t);

extern nxge_status_t nxge_alloc_tx_mem_pool(p_nxge_t);
extern nxge_status_t nxge_alloc_txb(p_nxge_t nxgep, int channel);
extern void nxge_free_txb(p_nxge_t nxgep, int channel);

#ifdef  __cplusplus
}
#endif

#endif  /* _SYS_NXGE_NXGE_TXDMA_H */