tx_stats
struct iflib_dma_info tx_stats[BNXT_MAX_NUM_QUEUES];
struct ctx_hw_stats *tx_stats = (void *)softc->tx_stats[txr].idi_vaddr;
"ucast_pkts", CTLFLAG_RD, &tx_stats->tx_ucast_pkts,
"mcast_pkts", CTLFLAG_RD, &tx_stats->tx_mcast_pkts,
"bcast_pkts", CTLFLAG_RD, &tx_stats->tx_bcast_pkts,
&tx_stats->tx_discard_pkts, "discarded transmit packets");
"error_pkts", CTLFLAG_RD, &tx_stats->tx_error_pkts,
"ucast_bytes", CTLFLAG_RD, &tx_stats->tx_ucast_bytes,
"mcast_bytes", CTLFLAG_RD, &tx_stats->tx_mcast_bytes,
"bcast_bytes", CTLFLAG_RD, &tx_stats->tx_bcast_bytes,
softc->tx_stats[i].idi_paddr);
&softc->tx_stats[i], 0);
bus_dmamap_sync(softc->tx_stats[i].idi_tag, softc->tx_stats[i].idi_map,
iflib_dma_free(&softc->tx_stats[i]);
iflib_dma_free(&softc->tx_stats[i]);
static const char *tx_stats[MAX_PM_NSTATS] = {
sbuf_printf(sb, "\n%-13s %10u %20ju", tx_stats[i], tx_cnt[i],
sbuf_printf(sb, "\n%-13s %10u %20ju", tx_stats[i], tx_cnt[i],
MPASS(i < nitems(tx_stats));
sbuf_printf(sb, "\n%-13s %10u %20ju", tx_stats[i], tx_cnt[i],
counter_u64_add(tx_ring->tx_stats.unmask_interrupt_num, 1);
counter_u64_add(tx_ring->tx_stats.missing_tx_comp, new_missed_tx);
ena_alloc_counters((counter_u64_t *)&txr->tx_stats,
sizeof(txr->tx_stats));
ena_free_counters((counter_u64_t *)&txr->tx_stats,
sizeof(txr->tx_stats));
counter_u64_add(tx_ring->tx_stats.bad_req_id, 1);
ena_reset_counters((counter_u64_t *)&tx_ring->tx_stats,
sizeof(tx_ring->tx_stats));
struct ena_stats_tx tx_stats;
counter_u64_add(tx_ring->tx_stats.doorbells, 1);
counter_u64_add(tx_ring->tx_stats.prepare_ctx_err, 1);
counter_u64_add_protected(tx_ring->tx_stats.cnt, 1);
counter_u64_add_protected(tx_ring->tx_stats.bytes,
counter_u64_add(tx_ring->tx_stats.queue_stop, 1);
counter_u64_add(tx_ring->tx_stats.queue_wakeup, 1);
counter_u64_add(tx_ring->tx_stats.unmask_interrupt_num, 1);
counter_u64_add(tx_ring->tx_stats.queue_wakeup, 1);
counter_u64_add(tx_ring->tx_stats.collapse, 1);
counter_u64_add(tx_ring->tx_stats.collapse_err, 1);
counter_u64_add(tx_ring->tx_stats.llq_buffer_copy, 1);
counter_u64_add(tx_ring->tx_stats.dma_mapping_err, 1);
counter_u64_add(tx_ring->tx_stats.prepare_ctx_err, 1);
counter_u64_add_protected(tx_ring->tx_stats.cnt, 1);
counter_u64_add_protected(tx_ring->tx_stats.bytes, packet_len);
counter_u64_add(tx_ring->tx_stats.llq_buffer_copy, 1);
struct ena_stats_tx *tx_stats;
tx_stats = &tx_ring->tx_stats;
CTLFLAG_RD, &tx_stats->cnt, "Packets sent");
CTLFLAG_RD, &tx_stats->bytes, "Bytes sent");
"prepare_ctx_err", CTLFLAG_RD, &tx_stats->prepare_ctx_err,
"dma_mapping_err", CTLFLAG_RD, &tx_stats->dma_mapping_err,
CTLFLAG_RD, &tx_stats->doorbells, "Queue doorbells");
"missing_tx_comp", CTLFLAG_RD, &tx_stats->missing_tx_comp,
CTLFLAG_RD, &tx_stats->bad_req_id, "Bad request id count");
CTLFLAG_RD, &tx_stats->collapse, "Mbuf collapse count");
"mbuf_collapse_err", CTLFLAG_RD, &tx_stats->collapse_err,
CTLFLAG_RD, &tx_stats->queue_wakeup, "Queue wakeups");
CTLFLAG_RD, &tx_stats->queue_stop, "Queue stops");
"llq_buffer_copy", CTLFLAG_RD, &tx_stats->llq_buffer_copy,
&tx_stats->unmask_interrupt_num,
struct tx_stats stats;
struct mana_stats *tx_stats)
counter_u64_add(tx_stats->collapse, 1);
counter_u64_add(tx_stats->collapse_err, 1);
struct mana_stats *tx_stats;
tx_stats = &txq->stats;
counter_u64_add(tx_stats->stop, 1);
uint64_t stops = counter_u64_fetch(tx_stats->stop);
uint64_t wakeups = counter_u64_fetch(tx_stats->wakeup);
counter_u64_add(tx_stats->alt_chg, 1);
err = mana_tx_map_mbuf(apc, tx_info, &mbuf, &pkg, tx_stats);
counter_u64_add(tx_stats->dma_mapping_err, 1);
counter_u64_add_protected(tx_stats->packets, packets);
counter_u64_add_protected(tx_stats->bytes, bytes);
struct mana_stats *tx_stats, *rx_stats;
tx_stats = &txq->stats;
CTLFLAG_RD, &tx_stats->packets, "Packets sent");
CTLFLAG_RD, &tx_stats->bytes, "Bytes sent");
CTLFLAG_RD, &tx_stats->wakeup, "Queue wakeups");
CTLFLAG_RD, &tx_stats->stop, "Queue stops");
CTLFLAG_RD, &tx_stats->collapse, "Mbuf collapse count");
&tx_stats->collapse_err, "Mbuf collapse failures");
&tx_stats->dma_mapping_err, "DMA mapping failures");
&tx_stats->alt_chg, "Switch to alternative txq");
&tx_stats->alt_reset, "Reset to self txq");
&tx_stats->cqe_err, "Error CQE count");
&tx_stats->cqe_unknown_type, "Unknown CQE count");
wq->tx_stats.tx_reqs++;
wq->tx_stats.tx_wrbs += num_wqes;
wq->tx_stats.tx_bytes += m->m_pkthdr.len;
wq->tx_stats.tx_pkts++;
sc->wq[def_q]->tx_stats.tx_stops ++;
wq->tx_stats.tx_compl++;
wq->tx_stats.tx_stops ++;
tx_reqs = wq->tx_stats.tx_reqs;
struct oce_tx_queue_stats tx_stats;
CTLFLAG_RD, &sc->wq[i]->tx_stats.tx_pkts,
CTLFLAG_RD, &sc->wq[i]->tx_stats.tx_bytes,
CTLFLAG_RD, &sc->wq[i]->tx_stats.tx_reqs, 0,
CTLFLAG_RD, &sc->wq[i]->tx_stats.tx_stops, 0,
CTLFLAG_RD, &sc->wq[i]->tx_stats.tx_wrbs, 0,
CTLFLAG_RD, &sc->wq[i]->tx_stats.tx_compl, 0,
&sc->wq[i]->tx_stats.ipv6_ext_hdr_tx_drop, 0,
adapter_stats->tx.t_tx_reqs += sc->wq[i]->tx_stats.tx_reqs;
adapter_stats->tx.t_tx_stops += sc->wq[i]->tx_stats.tx_stops;
adapter_stats->tx.t_tx_wrbs += sc->wq[i]->tx_stats.tx_wrbs;
adapter_stats->tx.t_tx_compl += sc->wq[i]->tx_stats.tx_compl;
adapter_stats->tx.t_tx_bytes += sc->wq[i]->tx_stats.tx_bytes;
adapter_stats->tx.t_tx_pkts += sc->wq[i]->tx_stats.tx_pkts;
sc->wq[i]->tx_stats.ipv6_ext_hdr_tx_drop;
CTLFLAG_RD, &sc->wq[i]->tx_stats.tx_pkts,
CTLFLAG_RD, &sc->wq[i]->tx_stats.tx_bytes,
CTLFLAG_RD, &sc->wq[i]->tx_stats.tx_reqs, 0,
CTLFLAG_RD, &sc->wq[i]->tx_stats.tx_stops, 0,
CTLFLAG_RD, &sc->wq[i]->tx_stats.tx_wrbs, 0,
CTLFLAG_RD, &sc->wq[i]->tx_stats.tx_compl, 0,
&sc->wq[i]->tx_stats.ipv6_ext_hdr_tx_drop, 0,
nic->bgx_stats.tx_stats[bgx->idx] = bgx->stats;
uint64_t tx_stats[BGX_TX_STATS_COUNT];