Symbol: tx_ring
sys/dev/ic/acx.c
1952
ACX_TX_RING_SIZE, (caddr_t *)&rd->tx_ring, BUS_DMA_NOWAIT);
sys/dev/ic/acx.c
1961
rd->tx_ring, ACX_TX_RING_SIZE, NULL, BUS_DMA_WAITOK);
sys/dev/ic/acx.c
2003
bd->tx_buf[i].tb_desc1 = &rd->tx_ring[i * 2];
sys/dev/ic/acx.c
2004
bd->tx_buf[i].tb_desc2 = &rd->tx_ring[(i * 2) + 1];
sys/dev/ic/acx.c
2022
if (rd->tx_ring != NULL) {
sys/dev/ic/acx.c
2068
bzero(&rd->tx_ring[i], sizeof(struct acx_host_desc));
sys/dev/ic/acx.c
2069
rd->tx_ring[i].h_ctrl = htole16(DESC_CTRL_HOSTOWN);
sys/dev/ic/acx.c
2072
rd->tx_ring[i].h_next_desc = htole32(rd->tx_ring_paddr);
sys/dev/ic/acx.c
2074
rd->tx_ring[i].h_next_desc = htole32(paddr);
sys/dev/ic/acx.c
595
bzero(rd->tx_ring, ACX_TX_RING_SIZE);
sys/dev/ic/acxvar.h
241
struct acx_host_desc *tx_ring;
sys/dev/ic/qwx.c
10319
qwx_dp_srng_cleanup(sc, &dp->tx_ring[i].tcl_data_ring);
sys/dev/ic/qwx.c
10320
qwx_dp_srng_cleanup(sc, &dp->tx_ring[i].tcl_comp_ring);
sys/dev/ic/qwx.c
10774
ret = qwx_dp_srng_setup(sc, &dp->tx_ring[i].tcl_data_ring,
sys/dev/ic/qwx.c
10782
ret = qwx_dp_srng_setup(sc, &dp->tx_ring[i].tcl_comp_ring,
sys/dev/ic/qwx.c
10790
srng = &sc->hal.srng_list[dp->tx_ring[i].tcl_data_ring.ring_id];
sys/dev/ic/qwx.c
10795
dp->tx_ring[i].tcl_data_ring.ring_id);
sys/dev/ic/qwx.c
10870
qwx_dp_tx_ring_free_tx_data(struct qwx_softc *sc, struct dp_tx_ring *tx_ring)
sys/dev/ic/qwx.c
10874
if (tx_ring->data == NULL)
sys/dev/ic/qwx.c
10878
struct qwx_tx_data *tx_data = &tx_ring->data[i];
sys/dev/ic/qwx.c
10888
free(tx_ring->data, M_DEVBUF,
sys/dev/ic/qwx.c
10890
tx_ring->data = NULL;
sys/dev/ic/qwx.c
10894
qwx_dp_tx_ring_alloc_tx_data(struct qwx_softc *sc, struct dp_tx_ring *tx_ring)
sys/dev/ic/qwx.c
10898
tx_ring->data = mallocarray(sc->hw_params.tx_ring_size,
sys/dev/ic/qwx.c
10900
if (tx_ring->data == NULL)
sys/dev/ic/qwx.c
10904
struct qwx_tx_data *tx_data = &tx_ring->data[i];
sys/dev/ic/qwx.c
10961
idr_init(&dp->tx_ring[i].txbuf_idr);
sys/dev/ic/qwx.c
10962
spin_lock_init(&dp->tx_ring[i].tx_idr_lock);
sys/dev/ic/qwx.c
10964
ret = qwx_dp_tx_ring_alloc_tx_data(sc, &dp->tx_ring[i]);
sys/dev/ic/qwx.c
10968
dp->tx_ring[i].cur = 0;
sys/dev/ic/qwx.c
10969
dp->tx_ring[i].queued = 0;
sys/dev/ic/qwx.c
10970
dp->tx_ring[i].tcl_data_ring_id = i;
sys/dev/ic/qwx.c
10971
dp->tx_ring[i].tx_status_head = 0;
sys/dev/ic/qwx.c
10972
dp->tx_ring[i].tx_status_tail = DP_TX_COMP_RING_SIZE - 1;
sys/dev/ic/qwx.c
10973
dp->tx_ring[i].tx_status = malloc(size, M_DEVBUF,
sys/dev/ic/qwx.c
10975
if (!dp->tx_ring[i].tx_status) {
sys/dev/ic/qwx.c
11051
spin_lock_bh(&dp->tx_ring[i].tx_idr_lock);
sys/dev/ic/qwx.c
11052
idr_for_each(&dp->tx_ring[i].txbuf_idr,
sys/dev/ic/qwx.c
11054
idr_destroy(&dp->tx_ring[i].txbuf_idr);
sys/dev/ic/qwx.c
11055
spin_unlock_bh(&dp->tx_ring[i].tx_idr_lock);
sys/dev/ic/qwx.c
11057
qwx_dp_tx_ring_free_tx_data(sc, &dp->tx_ring[i]);
sys/dev/ic/qwx.c
11058
free(dp->tx_ring[i].tx_status, M_DEVBUF,
sys/dev/ic/qwx.c
11060
dp->tx_ring[i].tx_status = NULL;
sys/dev/ic/qwx.c
15902
struct dp_tx_ring *tx_ring)
sys/dev/ic/qwx.c
15910
tx_data = &tx_ring->data[msdu_id];
sys/dev/ic/qwx.c
15917
if (tx_ring->queued > 0)
sys/dev/ic/qwx.c
15918
tx_ring->queued--;
sys/dev/ic/qwx.c
15928
qwx_dp_tx_htt_tx_complete_buf(struct qwx_softc *sc, struct dp_tx_ring *tx_ring,
sys/dev/ic/qwx.c
15932
qwx_dp_tx_free_txbuf(sc, ts->msdu_id, tx_ring);
sys/dev/ic/qwx.c
15937
uint8_t mac_id, uint32_t msdu_id, struct dp_tx_ring *tx_ring)
sys/dev/ic/qwx.c
15963
qwx_dp_tx_htt_tx_complete_buf(sc, tx_ring, &ts);
sys/dev/ic/qwx.c
15967
qwx_dp_tx_free_txbuf(sc, msdu_id, tx_ring);
sys/dev/ic/qwx.c
16052
qwx_dp_tx_complete_msdu(struct qwx_softc *sc, struct dp_tx_ring *tx_ring,
sys/dev/ic/qwx.c
16056
struct qwx_tx_data *tx_data = &tx_ring->data[msdu_id];
sys/dev/ic/qwx.c
16070
if (tx_ring->queued > 0)
sys/dev/ic/qwx.c
16071
tx_ring->queued--;
sys/dev/ic/qwx.c
16114
int hal_ring_id = dp->tx_ring[ring_id].tcl_comp_ring.ring_id;
sys/dev/ic/qwx.c
16117
struct dp_tx_ring *tx_ring = &dp->tx_ring[ring_id];
sys/dev/ic/qwx.c
16126
while ((QWX_TX_COMPL_NEXT(tx_ring->tx_status_head) !=
sys/dev/ic/qwx.c
16127
tx_ring->tx_status_tail) &&
sys/dev/ic/qwx.c
16129
memcpy(&tx_ring->tx_status[tx_ring->tx_status_head], desc,
sys/dev/ic/qwx.c
16131
tx_ring->tx_status_head =
sys/dev/ic/qwx.c
16132
QWX_TX_COMPL_NEXT(tx_ring->tx_status_head);
sys/dev/ic/qwx.c
16136
(ATH11K_TX_COMPL_NEXT(tx_ring->tx_status_head) ==
sys/dev/ic/qwx.c
16137
tx_ring->tx_status_tail))) {
sys/dev/ic/qwx.c
16146
while (QWX_TX_COMPL_NEXT(tx_ring->tx_status_tail) !=
sys/dev/ic/qwx.c
16147
tx_ring->tx_status_head) {
sys/dev/ic/qwx.c
16151
tx_ring->tx_status_tail =
sys/dev/ic/qwx.c
16152
QWX_TX_COMPL_NEXT(tx_ring->tx_status_tail);
sys/dev/ic/qwx.c
16153
tx_status = &tx_ring->tx_status[tx_ring->tx_status_tail];
sys/dev/ic/qwx.c
16167
(void *)tx_status, mac_id, msdu_id, tx_ring);
sys/dev/ic/qwx.c
16171
spin_lock(&tx_ring->tx_idr_lock);
sys/dev/ic/qwx.c
16172
msdu = idr_remove(&tx_ring->txbuf_idr, msdu_id);
sys/dev/ic/qwx.c
16176
spin_unlock(&tx_ring->tx_idr_lock);
sys/dev/ic/qwx.c
16180
spin_unlock(&tx_ring->tx_idr_lock);
sys/dev/ic/qwx.c
16186
qwx_dp_tx_complete_msdu(sc, tx_ring, msdu_id, &ts);
sys/dev/ic/qwx.c
16189
if (tx_ring->queued < sc->hw_params.tx_ring_size - 1) {
sys/dev/ic/qwx.c
25154
struct dp_tx_ring *tx_ring;
sys/dev/ic/qwx.c
25179
tx_ring = &dp->tx_ring[ti.ring_id];
sys/dev/ic/qwx.c
25181
if (tx_ring->queued >= sc->hw_params.tx_ring_size) {
sys/dev/ic/qwx.c
25186
msdu_id = tx_ring->cur;
sys/dev/ic/qwx.c
25187
tx_data = &tx_ring->data[msdu_id];
sys/dev/ic/qwx.c
25309
hal_ring_id = tx_ring->tcl_data_ring.ring_id;
sys/dev/ic/qwx.c
25345
tx_ring->queued++;
sys/dev/ic/qwx.c
25346
tx_ring->cur = (tx_ring->cur + 1) % sc->hw_params.tx_ring_size;
sys/dev/ic/qwx.c
25348
if (tx_ring->queued >= sc->hw_params.tx_ring_size - 1)
sys/dev/ic/qwxvar.h
1115
struct dp_tx_ring tx_ring[DP_TCL_NUM_RING_MAX];
sys/dev/ic/qwz.c
13524
struct dp_tx_ring *tx_ring)
sys/dev/ic/qwz.c
13531
tx_data = &tx_ring->data[msdu_id];
sys/dev/ic/qwz.c
13537
if (tx_ring->queued > 0)
sys/dev/ic/qwz.c
13538
tx_ring->queued--;
sys/dev/ic/qwz.c
13542
qwz_dp_tx_htt_tx_complete_buf(struct qwz_softc *sc, struct dp_tx_ring *tx_ring,
sys/dev/ic/qwz.c
13546
qwz_dp_tx_free_txbuf(sc, ts->msdu_id, tx_ring);
sys/dev/ic/qwz.c
13551
uint8_t mac_id, uint32_t msdu_id, struct dp_tx_ring *tx_ring)
sys/dev/ic/qwz.c
13577
qwz_dp_tx_htt_tx_complete_buf(sc, tx_ring, &ts);
sys/dev/ic/qwz.c
13581
qwz_dp_tx_free_txbuf(sc, msdu_id, tx_ring);
sys/dev/ic/qwz.c
13666
qwz_dp_tx_complete_msdu(struct qwz_softc *sc, struct dp_tx_ring *tx_ring,
sys/dev/ic/qwz.c
13670
struct qwz_tx_data *tx_data = &tx_ring->data[msdu_id];
sys/dev/ic/qwz.c
13692
if (tx_ring->queued > 0)
sys/dev/ic/qwz.c
13693
tx_ring->queued--;
sys/dev/ic/qwz.c
13704
int hal_ring_id = dp->tx_ring[ring_id].tcl_comp_ring.ring_id;
sys/dev/ic/qwz.c
13707
struct dp_tx_ring *tx_ring = &dp->tx_ring[ring_id];
sys/dev/ic/qwz.c
13716
while ((QWZ_TX_COMPL_NEXT(tx_ring->tx_status_head) !=
sys/dev/ic/qwz.c
13717
tx_ring->tx_status_tail) &&
sys/dev/ic/qwz.c
13719
memcpy(&tx_ring->tx_status[tx_ring->tx_status_head], desc,
sys/dev/ic/qwz.c
13721
tx_ring->tx_status_head =
sys/dev/ic/qwz.c
13722
QWZ_TX_COMPL_NEXT(tx_ring->tx_status_head);
sys/dev/ic/qwz.c
13726
(QWZ_TX_COMPL_NEXT(tx_ring->tx_status_head) ==
sys/dev/ic/qwz.c
13727
tx_ring->tx_status_tail))) {
sys/dev/ic/qwz.c
13736
while (QWZ_TX_COMPL_NEXT(tx_ring->tx_status_tail) !=
sys/dev/ic/qwz.c
13737
tx_ring->tx_status_head) {
sys/dev/ic/qwz.c
13741
tx_ring->tx_status_tail =
sys/dev/ic/qwz.c
13742
QWZ_TX_COMPL_NEXT(tx_ring->tx_status_tail);
sys/dev/ic/qwz.c
13743
tx_status = &tx_ring->tx_status[tx_ring->tx_status_tail];
sys/dev/ic/qwz.c
13757
(void *)tx_status, mac_id, msdu_id, tx_ring);
sys/dev/ic/qwz.c
13761
spin_lock(&tx_ring->tx_idr_lock);
sys/dev/ic/qwz.c
13762
msdu = idr_remove(&tx_ring->txbuf_idr, msdu_id);
sys/dev/ic/qwz.c
13766
spin_unlock(&tx_ring->tx_idr_lock);
sys/dev/ic/qwz.c
13770
spin_unlock(&tx_ring->tx_idr_lock);
sys/dev/ic/qwz.c
13776
qwz_dp_tx_complete_msdu(sc, tx_ring, msdu_id, &ts);
sys/dev/ic/qwz.c
13779
if (tx_ring->queued < sc->hw_params.tx_ring_size - 1) {
sys/dev/ic/qwz.c
22273
struct dp_tx_ring *tx_ring;
sys/dev/ic/qwz.c
22298
tx_ring = &dp->tx_ring[ti.ring_id];
sys/dev/ic/qwz.c
22300
if (tx_ring->queued >= sc->hw_params.tx_ring_size) {
sys/dev/ic/qwz.c
22305
msdu_id = tx_ring->cur;
sys/dev/ic/qwz.c
22306
tx_data = &tx_ring->data[msdu_id];
sys/dev/ic/qwz.c
22430
hal_ring_id = tx_ring->tcl_data_ring.ring_id;
sys/dev/ic/qwz.c
22465
tx_ring->queued++;
sys/dev/ic/qwz.c
22466
tx_ring->cur = (tx_ring->cur + 1) % sc->hw_params.tx_ring_size;
sys/dev/ic/qwz.c
22468
if (tx_ring->queued >= sc->hw_params.tx_ring_size - 1)
sys/dev/ic/qwz.c
7584
qwz_dp_srng_cleanup(sc, &dp->tx_ring[i].tcl_data_ring);
sys/dev/ic/qwz.c
7585
qwz_dp_srng_cleanup(sc, &dp->tx_ring[i].tcl_comp_ring);
sys/dev/ic/qwz.c
8003
ret = qwz_dp_srng_setup(sc, &dp->tx_ring[i].tcl_data_ring,
sys/dev/ic/qwz.c
8011
ret = qwz_dp_srng_setup(sc, &dp->tx_ring[i].tcl_comp_ring,
sys/dev/ic/qwz.c
8019
srng = &sc->hal.srng_list[dp->tx_ring[i].tcl_data_ring.ring_id];
sys/dev/ic/qwz.c
8101
qwz_dp_tx_ring_free_tx_data(struct qwz_softc *sc, struct dp_tx_ring *tx_ring)
sys/dev/ic/qwz.c
8105
if (tx_ring->data == NULL)
sys/dev/ic/qwz.c
8109
struct qwz_tx_data *tx_data = &tx_ring->data[i];
sys/dev/ic/qwz.c
8119
free(tx_ring->data, M_DEVBUF,
sys/dev/ic/qwz.c
8121
tx_ring->data = NULL;
sys/dev/ic/qwz.c
8125
qwz_dp_tx_ring_alloc_tx_data(struct qwz_softc *sc, struct dp_tx_ring *tx_ring)
sys/dev/ic/qwz.c
8129
tx_ring->data = mallocarray(sc->hw_params.tx_ring_size,
sys/dev/ic/qwz.c
8131
if (tx_ring->data == NULL)
sys/dev/ic/qwz.c
8135
struct qwz_tx_data *tx_data = &tx_ring->data[i];
sys/dev/ic/qwz.c
8677
idr_init(&dp->tx_ring[i].txbuf_idr);
sys/dev/ic/qwz.c
8678
spin_lock_init(&dp->tx_ring[i].tx_idr_lock);
sys/dev/ic/qwz.c
8680
ret = qwz_dp_tx_ring_alloc_tx_data(sc, &dp->tx_ring[i]);
sys/dev/ic/qwz.c
8684
dp->tx_ring[i].cur = 0;
sys/dev/ic/qwz.c
8685
dp->tx_ring[i].queued = 0;
sys/dev/ic/qwz.c
8686
dp->tx_ring[i].tcl_data_ring_id = i;
sys/dev/ic/qwz.c
8687
dp->tx_ring[i].tx_status_head = 0;
sys/dev/ic/qwz.c
8688
dp->tx_ring[i].tx_status_tail = DP_TX_COMP_RING_SIZE - 1;
sys/dev/ic/qwz.c
8689
dp->tx_ring[i].tx_status = malloc(size, M_DEVBUF,
sys/dev/ic/qwz.c
8691
if (!dp->tx_ring[i].tx_status) {
sys/dev/ic/qwz.c
8782
spin_lock_bh(&dp->tx_ring[i].tx_idr_lock);
sys/dev/ic/qwz.c
8783
idr_for_each(&dp->tx_ring[i].txbuf_idr,
sys/dev/ic/qwz.c
8785
idr_destroy(&dp->tx_ring[i].txbuf_idr);
sys/dev/ic/qwz.c
8786
spin_unlock_bh(&dp->tx_ring[i].tx_idr_lock);
sys/dev/ic/qwz.c
8788
qwz_dp_tx_ring_free_tx_data(sc, &dp->tx_ring[i]);
sys/dev/ic/qwz.c
8789
free(dp->tx_ring[i].tx_status, M_DEVBUF,
sys/dev/ic/qwz.c
8791
dp->tx_ring[i].tx_status = NULL;
sys/dev/ic/qwzvar.h
1253
struct dp_tx_ring tx_ring[DP_TCL_NUM_RING_MAX];
sys/dev/pci/if_bnxt.c
1000
tx->tx_ring.vaddr = BNXT_DMA_KVA(tx->tx_ring_mem);
sys/dev/pci/if_bnxt.c
1001
tx->tx_ring.paddr = BNXT_DMA_DVA(tx->tx_ring_mem);
sys/dev/pci/if_bnxt.c
1003
&tx->tx_ring, cp->ring.phys_id,
sys/dev/pci/if_bnxt.c
1009
bnxt_write_tx_doorbell(sc, &tx->tx_ring, 0);
sys/dev/pci/if_bnxt.c
1089
tx->tx_slots = mallocarray(sizeof(*bs), tx->tx_ring.ring_size,
sys/dev/pci/if_bnxt.c
1096
for (i = 0; i < tx->tx_ring.ring_size; i++) {
sys/dev/pci/if_bnxt.c
1133
bnxt_free_slots(sc, tx->tx_slots, i, tx->tx_ring.ring_size);
sys/dev/pci/if_bnxt.c
1152
&tx->tx_ring);
sys/dev/pci/if_bnxt.c
1185
bnxt_free_slots(sc, tx->tx_slots, tx->tx_ring.ring_size,
sys/dev/pci/if_bnxt.c
1186
tx->tx_ring.ring_size);
sys/dev/pci/if_bnxt.c
1203
&tx->tx_ring);
sys/dev/pci/if_bnxt.c
1553
free += tx->tx_ring.ring_size;
sys/dev/pci/if_bnxt.c
1612
if (idx == tx->tx_ring.ring_size)
sys/dev/pci/if_bnxt.c
1665
if (idx == tx->tx_ring.ring_size)
sys/dev/pci/if_bnxt.c
1682
if (idx == tx->tx_ring.ring_size)
sys/dev/pci/if_bnxt.c
1686
if (++tx->tx_prod >= tx->tx_ring.ring_size)
sys/dev/pci/if_bnxt.c
1696
bnxt_write_tx_doorbell(sc, &tx->tx_ring, idx);
sys/dev/pci/if_bnxt.c
232
struct bnxt_ring tx_ring;
sys/dev/pci/if_bnxt.c
2762
if (idx >= tx->tx_ring.ring_size)
sys/dev/pci/if_bnxt.c
2763
idx -= tx->tx_ring.ring_size;
sys/dev/pci/if_bnxt.c
2766
if (++tx->tx_cons >= tx->tx_ring.ring_size)
sys/dev/pci/if_bnxt.c
996
tx->tx_ring.phys_id = (uint16_t)HWRM_NA_SIGNATURE;
sys/dev/pci/if_bnxt.c
997
tx->tx_ring.id = BNXT_TX_RING_ID_BASE + bq->q_index;
sys/dev/pci/if_bnxt.c
998
tx->tx_ring.doorbell = tx->tx_ring.id * 0x80;
sys/dev/pci/if_bnxt.c
999
tx->tx_ring.ring_size = PAGE_SIZE / sizeof(struct tx_bd_short);
sys/dev/pci/if_et.c
1134
struct et_txdesc_ring *tx_ring = &sc->sc_tx_ring;
sys/dev/pci/if_et.c
1147
bzero(tx_ring->tr_desc, ET_TX_RING_SIZE);
sys/dev/pci/if_et.c
1148
bus_dmamap_sync(sc->sc_dmat, tx_ring->tr_dmap, 0,
sys/dev/pci/if_et.c
1149
tx_ring->tr_dmap->dm_mapsize, BUS_DMASYNC_PREWRITE);
sys/dev/pci/if_et.c
1292
struct et_txdesc_ring *tx_ring = &sc->sc_tx_ring;
sys/dev/pci/if_et.c
1296
bzero(tx_ring->tr_desc, ET_TX_RING_SIZE);
sys/dev/pci/if_et.c
1297
bus_dmamap_sync(sc->sc_dmat, tx_ring->tr_dmap, 0,
sys/dev/pci/if_et.c
1298
tx_ring->tr_dmap->dm_mapsize, BUS_DMASYNC_PREWRITE);
sys/dev/pci/if_et.c
1415
struct et_txdesc_ring *tx_ring = &sc->sc_tx_ring;
sys/dev/pci/if_et.c
1428
CSR_WRITE_4(sc, ET_TX_RING_HI, ET_ADDR_HI(tx_ring->tr_paddr));
sys/dev/pci/if_et.c
1429
CSR_WRITE_4(sc, ET_TX_RING_LO, ET_ADDR_LO(tx_ring->tr_paddr));
sys/dev/pci/if_et.c
1441
tx_ring->tr_ready_index = 0;
sys/dev/pci/if_et.c
1442
tx_ring->tr_ready_wrap = 0;
sys/dev/pci/if_et.c
1762
struct et_txdesc_ring *tx_ring = &sc->sc_tx_ring;
sys/dev/pci/if_et.c
1775
KKASSERT(tx_ring->tr_ready_index < ET_TX_NDESC);
sys/dev/pci/if_et.c
1776
first_idx = tx_ring->tr_ready_index;
sys/dev/pci/if_et.c
1824
td = &tx_ring->tr_desc[idx];
sys/dev/pci/if_et.c
1835
KKASSERT(tx_ring->tr_ready_index < ET_TX_NDESC);
sys/dev/pci/if_et.c
1836
if (++tx_ring->tr_ready_index == ET_TX_NDESC) {
sys/dev/pci/if_et.c
1837
tx_ring->tr_ready_index = 0;
sys/dev/pci/if_et.c
1838
tx_ring->tr_ready_wrap ^= 1;
sys/dev/pci/if_et.c
1841
td = &tx_ring->tr_desc[first_idx];
sys/dev/pci/if_et.c
1852
bus_dmamap_sync(sc->sc_dmat, tx_ring->tr_dmap, 0,
sys/dev/pci/if_et.c
1853
tx_ring->tr_dmap->dm_mapsize, BUS_DMASYNC_PREWRITE);
sys/dev/pci/if_et.c
1856
tx_ready_pos = __SHIFTIN(tx_ring->tr_ready_index,
sys/dev/pci/if_et.c
1858
if (tx_ring->tr_ready_wrap)
sys/dev/pci/if_et.c
1875
struct et_txdesc_ring *tx_ring = &sc->sc_tx_ring;
sys/dev/pci/if_et.c
1893
bzero(&tx_ring->tr_desc[tbd->tbd_start_index],
sys/dev/pci/if_et.c
1895
bus_dmamap_sync(sc->sc_dmat, tx_ring->tr_dmap, 0,
sys/dev/pci/if_et.c
1896
tx_ring->tr_dmap->dm_mapsize, BUS_DMASYNC_PREWRITE);
sys/dev/pci/if_et.c
596
struct et_txdesc_ring *tx_ring = &sc->sc_tx_ring;
sys/dev/pci/if_et.c
606
(void **)&tx_ring->tr_desc, &tx_ring->tr_paddr, &tx_ring->tr_dmap,
sys/dev/pci/if_et.c
607
&tx_ring->tr_seg);
sys/dev/pci/if_et.c
683
struct et_txdesc_ring *tx_ring = &sc->sc_tx_ring;
sys/dev/pci/if_et.c
692
et_dma_mem_destroy(sc, tx_ring->tr_desc, tx_ring->tr_dmap);
sys/dev/pci/if_ngbe.c
1050
struct tx_ring *txr = &sc->tx_rings[i];
sys/dev/pci/if_ngbe.c
1164
struct tx_ring *txr;
sys/dev/pci/if_ngbe.c
1177
sc->tx_rings = mallocarray(sc->sc_nqueues, sizeof(struct tx_ring),
sys/dev/pci/if_ngbe.c
1249
free(sc->tx_rings, M_DEVBUF, sc->sc_nqueues * sizeof(struct tx_ring));
sys/dev/pci/if_ngbe.c
1300
struct tx_ring *txr = sc->tx_rings;
sys/dev/pci/if_ngbe.c
1308
ngbe_free_transmit_buffers(struct tx_ring *txr)
sys/dev/pci/if_ngbe.c
137
int ngbe_encap(struct tx_ring *, struct mbuf *);
sys/dev/pci/if_ngbe.c
1379
ngbe_allocate_transmit_buffers(struct tx_ring *txr)
sys/dev/pci/if_ngbe.c
1446
ngbe_setup_transmit_ring(struct tx_ring *txr)
sys/dev/pci/if_ngbe.c
1489
struct tx_ring *txr = sc->tx_rings;
sys/dev/pci/if_ngbe.c
1707
struct tx_ring *txr;
sys/dev/pci/if_ngbe.c
1793
struct tx_ring *txr = nq->txr;
sys/dev/pci/if_ngbe.c
210
int ngbe_tx_ctx_setup(struct tx_ring *, struct mbuf *,
sys/dev/pci/if_ngbe.c
212
void ngbe_txeof(struct tx_ring *);
sys/dev/pci/if_ngbe.c
2629
ngbe_encap(struct tx_ring *txr, struct mbuf *m)
sys/dev/pci/if_ngbe.c
4414
ngbe_tx_ctx_setup(struct tx_ring *txr, struct mbuf *m, uint32_t *cmd_type_len,
sys/dev/pci/if_ngbe.c
4456
ngbe_txeof(struct tx_ring *txr)
sys/dev/pci/if_ngbe.c
695
struct tx_ring *txr = ifq->ifq_softc;
sys/dev/pci/if_ngbe.c
87
void ngbe_free_transmit_buffers(struct tx_ring *);
sys/dev/pci/if_ngbe.c
889
struct tx_ring *txr = sc->tx_rings;
sys/dev/pci/if_ngbe.c
89
int ngbe_allocate_transmit_buffers(struct tx_ring *);
sys/dev/pci/if_ngbe.c
91
int ngbe_setup_transmit_ring(struct tx_ring *);
sys/dev/pci/if_ngbereg.h
1032
struct tx_ring *txr;
sys/dev/pci/if_ngbereg.h
1062
struct tx_ring *tx_rings;
sys/dev/pci/if_rtwn.c
1011
tx_ring = &sc->tx_ring[qid];
sys/dev/pci/if_rtwn.c
1012
data = &tx_ring->tx_data[tx_ring->cur];
sys/dev/pci/if_rtwn.c
1019
txd = &tx_ring->desc[tx_ring->cur];
sys/dev/pci/if_rtwn.c
1158
bus_dmamap_sync(sc->sc_dmat, tx_ring->map, 0, MCLBYTES,
sys/dev/pci/if_rtwn.c
1185
tx_ring->cur = (tx_ring->cur + 1) % RTWN_TX_LIST_COUNT;
sys/dev/pci/if_rtwn.c
1186
tx_ring->queued++;
sys/dev/pci/if_rtwn.c
1188
if (tx_ring->queued >= (RTWN_TX_LIST_COUNT - 1))
sys/dev/pci/if_rtwn.c
1202
struct rtwn_tx_ring *tx_ring = &sc->tx_ring[qid];
sys/dev/pci/if_rtwn.c
1207
bus_dmamap_sync(sc->sc_dmat, tx_ring->map, 0, MCLBYTES,
sys/dev/pci/if_rtwn.c
1211
tx_data = &tx_ring->tx_data[i];
sys/dev/pci/if_rtwn.c
1215
tx_desc = &tx_ring->desc[i];
sys/dev/pci/if_rtwn.c
1226
tx_ring->queued--;
sys/dev/pci/if_rtwn.c
1232
if (tx_ring->queued < (RTWN_TX_LIST_COUNT - 1))
sys/dev/pci/if_rtwn.c
167
struct rtwn_tx_ring tx_ring[RTWN_NTXQUEUES];
sys/dev/pci/if_rtwn.c
2061
sc->tx_ring[RTWN_BK_QUEUE].map->dm_segs[0].ds_addr);
sys/dev/pci/if_rtwn.c
2063
sc->tx_ring[RTWN_BE_QUEUE].map->dm_segs[0].ds_addr);
sys/dev/pci/if_rtwn.c
2065
sc->tx_ring[RTWN_VI_QUEUE].map->dm_segs[0].ds_addr);
sys/dev/pci/if_rtwn.c
2067
sc->tx_ring[RTWN_VO_QUEUE].map->dm_segs[0].ds_addr);
sys/dev/pci/if_rtwn.c
2069
sc->tx_ring[RTWN_BEACON_QUEUE].map->dm_segs[0].ds_addr);
sys/dev/pci/if_rtwn.c
2071
sc->tx_ring[RTWN_MGNT_QUEUE].map->dm_segs[0].ds_addr);
sys/dev/pci/if_rtwn.c
2073
sc->tx_ring[RTWN_HIGH_QUEUE].map->dm_segs[0].ds_addr);
sys/dev/pci/if_rtwn.c
623
struct rtwn_tx_ring *tx_ring = &sc->tx_ring[qid];
sys/dev/pci/if_rtwn.c
630
BUS_DMA_NOWAIT, &tx_ring->map);
sys/dev/pci/if_rtwn.c
639
&tx_ring->seg, 1, &tx_ring->nsegs, BUS_DMA_NOWAIT | BUS_DMA_ZERO);
sys/dev/pci/if_rtwn.c
646
error = bus_dmamem_map(sc->sc_dmat, &tx_ring->seg, tx_ring->nsegs,
sys/dev/pci/if_rtwn.c
648
(caddr_t *)&tx_ring->desc, BUS_DMA_NOWAIT);
sys/dev/pci/if_rtwn.c
650
bus_dmamem_free(sc->sc_dmat, &tx_ring->seg, tx_ring->nsegs);
sys/dev/pci/if_rtwn.c
656
error = bus_dmamap_load(sc->sc_dmat, tx_ring->map, tx_ring->desc,
sys/dev/pci/if_rtwn.c
666
struct r92c_tx_desc_pci *desc = &tx_ring->desc[i];
sys/dev/pci/if_rtwn.c
669
desc->nextdescaddr = htole32(tx_ring->map->dm_segs[0].ds_addr
sys/dev/pci/if_rtwn.c
673
tx_data = &tx_ring->tx_data[i];
sys/dev/pci/if_rtwn.c
694
struct rtwn_tx_ring *tx_ring = &sc->tx_ring[qid];
sys/dev/pci/if_rtwn.c
698
struct r92c_tx_desc_pci *desc = &tx_ring->desc[i];
sys/dev/pci/if_rtwn.c
699
struct rtwn_tx_data *tx_data = &tx_ring->tx_data[i];
sys/dev/pci/if_rtwn.c
714
bus_dmamap_sync(sc->sc_dmat, tx_ring->map, 0, MCLBYTES,
sys/dev/pci/if_rtwn.c
718
tx_ring->queued = 0;
sys/dev/pci/if_rtwn.c
719
tx_ring->cur = 0;
sys/dev/pci/if_rtwn.c
725
struct rtwn_tx_ring *tx_ring = &sc->tx_ring[qid];
sys/dev/pci/if_rtwn.c
729
if (tx_ring->map != NULL) {
sys/dev/pci/if_rtwn.c
730
if (tx_ring->desc != NULL) {
sys/dev/pci/if_rtwn.c
731
bus_dmamap_unload(sc->sc_dmat, tx_ring->map);
sys/dev/pci/if_rtwn.c
732
bus_dmamem_unmap(sc->sc_dmat, (caddr_t)tx_ring->desc,
sys/dev/pci/if_rtwn.c
735
bus_dmamem_free(sc->sc_dmat, &tx_ring->seg, tx_ring->nsegs);
sys/dev/pci/if_rtwn.c
737
bus_dmamap_destroy(sc->sc_dmat, tx_ring->map);
sys/dev/pci/if_rtwn.c
741
tx_data = &tx_ring->tx_data[i];
sys/dev/pci/if_rtwn.c
752
tx_ring->queued = 0;
sys/dev/pci/if_rtwn.c
753
tx_ring->cur = 0;
sys/dev/pci/if_rtwn.c
984
struct rtwn_tx_ring *tx_ring;