sys/dev/al_eth/al_eth.c
1000
struct al_eth_ring *ring = &adapter->rx_ring[i];
sys/dev/al_eth/al_eth.c
1041
struct al_eth_ring *rx_ring,
sys/dev/al_eth/al_eth.c
1067
error = bus_dmamap_load_mbuf_sg(rx_ring->dma_buf_tag, rx_info->dma_map,
sys/dev/al_eth/al_eth.c
1070
device_printf(rx_ring->dev, "failed to map mbuf, error = %d\n",
sys/dev/al_eth/al_eth.c
1088
struct al_eth_ring *rx_ring = &adapter->rx_ring[qid];
sys/dev/al_eth/al_eth.c
1092
next_to_use = rx_ring->next_to_use;
sys/dev/al_eth/al_eth.c
1097
&rx_ring->rx_buffer_info[next_to_use];
sys/dev/al_eth/al_eth.c
1100
rx_ring, rx_info) < 0)) {
sys/dev/al_eth/al_eth.c
1106
rc = al_eth_rx_buffer_add(rx_ring->dma_q,
sys/dev/al_eth/al_eth.c
1114
next_to_use = AL_ETH_RX_RING_IDX_NEXT(rx_ring, next_to_use);
sys/dev/al_eth/al_eth.c
1120
qid, i, al_udma_available_get(rx_ring->dma_q));
sys/dev/al_eth/al_eth.c
1123
al_eth_rx_buffer_action(rx_ring->dma_q, i);
sys/dev/al_eth/al_eth.c
1125
rx_ring->next_to_use = next_to_use;
sys/dev/al_eth/al_eth.c
1451
struct al_eth_ring *rx_ring = arg;
sys/dev/al_eth/al_eth.c
1455
device_printf_dbg(rx_ring->dev, "%s for ring ID = %d\n", __func__,
sys/dev/al_eth/al_eth.c
1456
rx_ring->ring_id);
sys/dev/al_eth/al_eth.c
1462
if ((napi == 0) || (napi && rx_ring->enqueue_is_running == 0))
sys/dev/al_eth/al_eth.c
1463
taskqueue_enqueue(rx_ring->enqueue_tq, &rx_ring->enqueue_task);
sys/dev/al_eth/al_eth.c
1516
struct al_eth_ring *rx_ring, struct al_eth_pkt *hal_pkt,
sys/dev/al_eth/al_eth.c
1521
&rx_ring->rx_buffer_info[*next_to_clean];
sys/dev/al_eth/al_eth.c
1529
*next_to_clean = AL_ETH_RX_RING_IDX_NEXT(rx_ring,
sys/dev/al_eth/al_eth.c
1537
mbuf->m_pkthdr.rcvif = rx_ring->netdev;
sys/dev/al_eth/al_eth.c
1556
smbuf->m_pkthdr.rcvif = rx_ring->netdev;
sys/dev/al_eth/al_eth.c
1562
*next_to_clean = AL_ETH_RX_RING_IDX_NEXT(rx_ring,
sys/dev/al_eth/al_eth.c
1570
bus_dmamap_unload(rx_ring->dma_buf_tag, rx_info->dma_map);
sys/dev/al_eth/al_eth.c
1573
*next_to_clean = AL_ETH_RX_RING_IDX_NEXT(rx_ring, *next_to_clean);
sys/dev/al_eth/al_eth.c
1581
struct al_eth_ring *rx_ring = arg;
sys/dev/al_eth/al_eth.c
1583
unsigned int qid = rx_ring->ring_id;
sys/dev/al_eth/al_eth.c
1584
struct al_eth_pkt *hal_pkt = &rx_ring->hal_pkt;
sys/dev/al_eth/al_eth.c
1585
uint16_t next_to_clean = rx_ring->next_to_clean;
sys/dev/al_eth/al_eth.c
1591
rx_ring->enqueue_is_running = 1;
sys/dev/al_eth/al_eth.c
1598
descs = al_eth_pkt_rx(rx_ring->dma_q, hal_pkt);
sys/dev/al_eth/al_eth.c
1602
device_printf_dbg(rx_ring->dev, "rx_poll: q %d got packet "
sys/dev/al_eth/al_eth.c
1604
device_printf_dbg(rx_ring->dev, "rx_poll: q %d flags %x. "
sys/dev/al_eth/al_eth.c
1611
device_printf(rx_ring->dev, "receive packet with error. "
sys/dev/al_eth/al_eth.c
1613
next_to_clean = AL_ETH_RX_RING_IDX_ADD(rx_ring,
sys/dev/al_eth/al_eth.c
1619
mbuf = al_eth_rx_mbuf(rx_ring->adapter, rx_ring, hal_pkt, descs,
sys/dev/al_eth/al_eth.c
1624
next_to_clean = AL_ETH_RX_RING_IDX_ADD(rx_ring,
sys/dev/al_eth/al_eth.c
1629
if (__predict_true(if_getcapenable(rx_ring->netdev) & IFCAP_RXCSUM ||
sys/dev/al_eth/al_eth.c
1630
if_getcapenable(rx_ring->netdev) & IFCAP_RXCSUM_IPV6)) {
sys/dev/al_eth/al_eth.c
1631
al_eth_rx_checksum(rx_ring->adapter, hal_pkt, mbuf);
sys/dev/al_eth/al_eth.c
1642
if ((rx_ring->lro_enabled != 0) &&
sys/dev/al_eth/al_eth.c
1651
if (rx_ring->lro.lro_cnt != 0) {
sys/dev/al_eth/al_eth.c
1652
if (tcp_lro_rx(&rx_ring->lro, mbuf, 0) == 0)
sys/dev/al_eth/al_eth.c
1658
if_input(rx_ring->netdev, mbuf);
sys/dev/al_eth/al_eth.c
1662
rx_ring->next_to_clean = next_to_clean;
sys/dev/al_eth/al_eth.c
1664
refill_required = al_udma_available_get(rx_ring->dma_q);
sys/dev/al_eth/al_eth.c
1665
refill_actual = al_eth_refill_rx_bufs(rx_ring->adapter, qid,
sys/dev/al_eth/al_eth.c
1669
device_printf_dbg(rx_ring->dev,
sys/dev/al_eth/al_eth.c
1673
tcp_lro_flush_all(&rx_ring->lro);
sys/dev/al_eth/al_eth.c
1676
rx_ring->enqueue_is_running = 0;
sys/dev/al_eth/al_eth.c
1680
al_eth_irq_config(rx_ring->unmask_reg_offset, rx_ring->unmask_val);
sys/dev/al_eth/al_eth.c
2116
adapter->irq_tbl[irq_idx].data = &adapter->rx_ring[i];
sys/dev/al_eth/al_eth.c
2448
struct al_eth_ring *rx_ring = &adapter->rx_ring[qid];
sys/dev/al_eth/al_eth.c
2449
device_t dev = rx_ring->dev;
sys/dev/al_eth/al_eth.c
2450
struct al_udma_q_params *q_params = &rx_ring->q_params;
sys/dev/al_eth/al_eth.c
2454
size = sizeof(struct al_eth_rx_buffer) * rx_ring->sw_count;
sys/dev/al_eth/al_eth.c
2459
rx_ring->rx_buffer_info = malloc(size, M_IFAL, M_ZERO | M_WAITOK);
sys/dev/al_eth/al_eth.c
2460
rx_ring->descs_size = rx_ring->hw_count * sizeof(union al_udma_desc);
sys/dev/al_eth/al_eth.c
2461
q_params->size = rx_ring->hw_count;
sys/dev/al_eth/al_eth.c
2466
(void**)&q_params->desc_base, rx_ring->descs_size);
sys/dev/al_eth/al_eth.c
2473
rx_ring->cdescs_size = rx_ring->hw_count * q_params->cdesc_size;
sys/dev/al_eth/al_eth.c
2477
(void**)&q_params->cdesc_base, rx_ring->cdescs_size);
sys/dev/al_eth/al_eth.c
2483
NET_TASK_INIT(&rx_ring->enqueue_task, 0, al_eth_rx_recv_work, rx_ring);
sys/dev/al_eth/al_eth.c
2484
rx_ring->enqueue_tq = taskqueue_create_fast("al_rx_enque", M_NOWAIT,
sys/dev/al_eth/al_eth.c
2485
taskqueue_thread_enqueue, &rx_ring->enqueue_tq);
sys/dev/al_eth/al_eth.c
2486
taskqueue_start_threads(&rx_ring->enqueue_tq, 1, PI_NET, "%s rxeq",
sys/dev/al_eth/al_eth.c
2501
&rx_ring->dma_buf_tag);
sys/dev/al_eth/al_eth.c
2508
for (size = 0; size < rx_ring->sw_count; size++) {
sys/dev/al_eth/al_eth.c
2509
ret = bus_dmamap_create(rx_ring->dma_buf_tag, 0,
sys/dev/al_eth/al_eth.c
2510
&rx_ring->rx_buffer_info[size].dma_map);
sys/dev/al_eth/al_eth.c
2518
memset(q_params->cdesc_base, 0, rx_ring->cdescs_size);
sys/dev/al_eth/al_eth.c
2522
int err = tcp_lro_init(&rx_ring->lro);
sys/dev/al_eth/al_eth.c
2529
rx_ring->lro_enabled = true;
sys/dev/al_eth/al_eth.c
2530
rx_ring->lro.ifp = adapter->netdev;
sys/dev/al_eth/al_eth.c
2534
rx_ring->next_to_clean = 0;
sys/dev/al_eth/al_eth.c
2535
rx_ring->next_to_use = 0;
sys/dev/al_eth/al_eth.c
2550
struct al_eth_ring *rx_ring = &adapter->rx_ring[qid];
sys/dev/al_eth/al_eth.c
2551
struct al_udma_q_params *q_params = &rx_ring->q_params;
sys/dev/al_eth/al_eth.c
2555
while (taskqueue_cancel(rx_ring->enqueue_tq,
sys/dev/al_eth/al_eth.c
2556
&rx_ring->enqueue_task, NULL)) {
sys/dev/al_eth/al_eth.c
2557
taskqueue_drain(rx_ring->enqueue_tq, &rx_ring->enqueue_task);
sys/dev/al_eth/al_eth.c
2560
taskqueue_free(rx_ring->enqueue_tq);
sys/dev/al_eth/al_eth.c
2562
for (size = 0; size < rx_ring->sw_count; size++) {
sys/dev/al_eth/al_eth.c
2563
m_freem(rx_ring->rx_buffer_info[size].m);
sys/dev/al_eth/al_eth.c
2564
rx_ring->rx_buffer_info[size].m = NULL;
sys/dev/al_eth/al_eth.c
2565
bus_dmamap_unload(rx_ring->dma_buf_tag,
sys/dev/al_eth/al_eth.c
2566
rx_ring->rx_buffer_info[size].dma_map);
sys/dev/al_eth/al_eth.c
2567
bus_dmamap_destroy(rx_ring->dma_buf_tag,
sys/dev/al_eth/al_eth.c
2568
rx_ring->rx_buffer_info[size].dma_map);
sys/dev/al_eth/al_eth.c
2570
bus_dma_tag_destroy(rx_ring->dma_buf_tag);
sys/dev/al_eth/al_eth.c
2572
free(rx_ring->rx_buffer_info, M_IFAL);
sys/dev/al_eth/al_eth.c
2573
rx_ring->rx_buffer_info = NULL;
sys/dev/al_eth/al_eth.c
2594
tcp_lro_free(&rx_ring->lro);
sys/dev/al_eth/al_eth.c
2609
if (adapter->rx_ring[i].q_params.desc_base != 0)
sys/dev/al_eth/al_eth.c
3051
q_params = &adapter->rx_ring[qid].q_params;
sys/dev/al_eth/al_eth.h
152
#define AL_ETH_RX_RING_IDX_NEXT(rx_ring, idx) (((idx) + 1) & (AL_ETH_DEFAULT_RX_DESCS - 1))
sys/dev/al_eth/al_eth.h
153
#define AL_ETH_RX_RING_IDX_ADD(rx_ring, idx, n) (((idx) + (n)) & (AL_ETH_DEFAULT_RX_DESCS - 1))
sys/dev/al_eth/al_eth.h
277
struct al_eth_ring rx_ring[AL_ETH_NUM_QUEUES];
sys/dev/axgbe/if_axgbe_pci.c
1756
struct xgbe_ring *rx_ring;
sys/dev/axgbe/if_axgbe_pci.c
1774
rx_ring = (struct xgbe_ring*)malloc(nrxqs *
sys/dev/axgbe/if_axgbe_pci.c
1777
if (rx_ring == NULL) {
sys/dev/axgbe/if_axgbe_pci.c
1782
channel->rx_ring = rx_ring;
sys/dev/axgbe/if_axgbe_pci.c
1784
for (j = 0; j < nrxqs; j++, rx_ring++) {
sys/dev/axgbe/if_axgbe_pci.c
1785
rx_ring->rdata =
sys/dev/axgbe/if_axgbe_pci.c
1790
rx_ring->rdesc = (struct xgbe_ring_desc *)va[i*nrxqs + j];
sys/dev/axgbe/if_axgbe_pci.c
1791
rx_ring->rdesc_paddr = pa[i*nrxqs + j];
sys/dev/axgbe/if_axgbe_pci.c
1792
rx_ring->rdesc_count = scctx->isc_nrxd[j];
sys/dev/axgbe/if_axgbe_pci.c
1793
spin_lock_init(&rx_ring->lock);
sys/dev/axgbe/if_axgbe_pci.c
1807
rx_ring = channel->rx_ring;
sys/dev/axgbe/if_axgbe_pci.c
1808
for (k = 0; k < nrxqs ; k++, rx_ring++) {
sys/dev/axgbe/if_axgbe_pci.c
1809
if (rx_ring && rx_ring->rdata)
sys/dev/axgbe/if_axgbe_pci.c
1810
free(rx_ring->rdata, M_AXGBE);
sys/dev/axgbe/if_axgbe_pci.c
1812
free(channel->rx_ring, M_AXGBE);
sys/dev/axgbe/if_axgbe_pci.c
1814
channel->rx_ring = NULL;
sys/dev/axgbe/if_axgbe_pci.c
1830
struct xgbe_ring *rx_ring;
sys/dev/axgbe/if_axgbe_pci.c
1850
rx_ring = channel->rx_ring;
sys/dev/axgbe/if_axgbe_pci.c
1851
for (j = 0; j < sctx->isc_nrxqs ; j++, rx_ring++) {
sys/dev/axgbe/if_axgbe_pci.c
1852
if (rx_ring && rx_ring->rdata)
sys/dev/axgbe/if_axgbe_pci.c
1853
free(rx_ring->rdata, M_AXGBE);
sys/dev/axgbe/if_axgbe_pci.c
1855
free(channel->rx_ring, M_AXGBE);
sys/dev/axgbe/if_axgbe_pci.c
1856
channel->rx_ring = NULL;
sys/dev/axgbe/if_axgbe_pci.c
2061
if (channel->tx_ring && channel->rx_ring)
sys/dev/axgbe/if_axgbe_pci.c
2065
else if (channel->rx_ring)
sys/dev/axgbe/if_axgbe_pci.c
2081
if (channel->tx_ring && channel->rx_ring)
sys/dev/axgbe/if_axgbe_pci.c
2085
else if (channel->rx_ring)
sys/dev/axgbe/if_axgbe_pci.c
809
channel->rx_ring = NULL;
sys/dev/axgbe/xgbe-desc.c
179
ring = channel->rx_ring;
sys/dev/axgbe/xgbe-dev.c
1329
struct xgbe_ring *ring = channel->rx_ring;
sys/dev/axgbe/xgbe-dev.c
1353
struct xgbe_ring *ring = channel->rx_ring;
sys/dev/axgbe/xgbe-dev.c
185
if (pdata->channel[i]->rx_ring)
sys/dev/axgbe/xgbe-dev.c
259
if (!pdata->channel[i]->rx_ring)
sys/dev/axgbe/xgbe-dev.c
2610
if (!pdata->channel[i]->rx_ring)
sys/dev/axgbe/xgbe-dev.c
2649
if (!pdata->channel[i]->rx_ring)
sys/dev/axgbe/xgbe-dev.c
2701
if (!pdata->channel[i]->rx_ring)
sys/dev/axgbe/xgbe-dev.c
2715
if (!pdata->channel[i]->rx_ring)
sys/dev/axgbe/xgbe-dev.c
281
if (!pdata->channel[i]->rx_ring)
sys/dev/axgbe/xgbe-dev.c
318
if (!pdata->channel[i]->rx_ring)
sys/dev/axgbe/xgbe-dev.c
651
if (channel->rx_ring) {
sys/dev/axgbe/xgbe-txrx.c
448
struct xgbe_ring *ring = channel->rx_ring;
sys/dev/axgbe/xgbe-txrx.c
539
struct xgbe_ring *ring = channel->rx_ring;
sys/dev/axgbe/xgbe-txrx.c
563
struct xgbe_ring *ring = channel->rx_ring;
sys/dev/axgbe/xgbe-txrx.c
701
struct xgbe_ring *ring = channel->rx_ring;
sys/dev/axgbe/xgbe.h
541
struct xgbe_ring *rx_ring;
sys/dev/bnxt/bnxt_en/bnxt_txrx.c
296
struct bnxt_ring *rx_ring;
sys/dev/bnxt/bnxt_en/bnxt_txrx.c
315
rx_ring = &softc->rx_rings[rxqid];
sys/dev/bnxt/bnxt_en/bnxt_txrx.c
319
rx_ring = &softc->ag_rings[rxqid];
sys/dev/bnxt/bnxt_en/bnxt_txrx.c
322
rxbd = (void *)rx_ring->vaddr;
sys/dev/bnxt/bnxt_en/bnxt_txrx.c
333
if (++pidx == rx_ring->ring_size) {
sys/dev/bnxt/bnxt_en/bnxt_txrx.c
335
rx_ring->epoch_bit = !rx_ring->epoch_bit;
sys/dev/bnxt/bnxt_en/bnxt_txrx.c
338
rx_ring->epoch_arr[pidx] = rx_ring->epoch_bit;
sys/dev/bnxt/bnxt_en/bnxt_txrx.c
349
struct bnxt_ring *rx_ring;
sys/dev/bnxt/bnxt_en/bnxt_txrx.c
352
rx_ring = &softc->rx_rings[rxqid];
sys/dev/bnxt/bnxt_en/bnxt_txrx.c
354
rx_ring = &softc->ag_rings[rxqid];
sys/dev/bnxt/bnxt_en/bnxt_txrx.c
362
softc->db_ops.bnxt_db_rx(rx_ring, pidx);
sys/dev/e1000/em_txrx.c
544
struct rx_ring *rxr = &que->rxr;
sys/dev/e1000/em_txrx.c
573
struct rx_ring *rxr = &que->rxr;
sys/dev/e1000/em_txrx.c
601
struct rx_ring *rxr = &que->rxr;
sys/dev/e1000/em_txrx.c
612
struct rx_ring *rxr = &que->rxr;
sys/dev/e1000/em_txrx.c
637
struct rx_ring *rxr = &que->rxr;
sys/dev/e1000/em_txrx.c
662
struct rx_ring *rxr = &que->rxr;
sys/dev/e1000/em_txrx.c
726
struct rx_ring *rxr = &que->rxr;
sys/dev/e1000/if_em.c
1653
struct tx_ring *txr, struct rx_ring *rxr)
sys/dev/e1000/if_em.c
1815
struct rx_ring *rxr = &que->rxr;
sys/dev/e1000/if_em.c
1915
struct rx_ring *rxr = &que->rxr;
sys/dev/e1000/if_em.c
3568
struct rx_ring *rxr = &que->rxr;
sys/dev/e1000/if_em.c
3877
struct rx_ring *rxr = &que->rxr;
sys/dev/e1000/if_em.c
3948
struct rx_ring *rxr = &que->rxr;
sys/dev/e1000/if_em.c
465
struct tx_ring *, struct rx_ring *);
sys/dev/e1000/if_em.c
5027
struct rx_ring *rxr = &rx_que->rxr;
sys/dev/e1000/if_em.c
5720
struct rx_ring *rxr = &sc->rx_queues->rxr;
sys/dev/e1000/if_em.c
854
struct rx_ring *rxr = &rx_que->rxr;
sys/dev/e1000/if_em.h
467
struct rx_ring rxr;
sys/dev/e1000/igb_txrx.c
370
struct rx_ring *rxr = &que->rxr;
sys/dev/e1000/igb_txrx.c
395
struct rx_ring *rxr = &que->rxr;
sys/dev/e1000/igb_txrx.c
406
struct rx_ring *rxr = &que->rxr;
sys/dev/e1000/igb_txrx.c
438
struct rx_ring *rxr = &que->rxr;
sys/dev/ena/ena.c
1022
ena_alloc_rx_mbuf(struct ena_adapter *adapter, struct ena_ring *rx_ring,
sys/dev/ena/ena.c
1037
rx_ring->rx_mbuf_sz);
sys/dev/ena/ena.c
1040
counter_u64_add(rx_ring->rx_stats.mjum_alloc_fail, 1);
sys/dev/ena/ena.c
1043
counter_u64_add(rx_ring->rx_stats.mbuf_alloc_fail, 1);
sys/dev/ena/ena.c
1048
mlen = rx_ring->rx_mbuf_sz;
sys/dev/ena/ena.c
1062
counter_u64_add(rx_ring->rx_stats.dma_mapping_err, 1);
sys/dev/ena/ena.c
1085
ena_free_rx_mbuf(struct ena_adapter *adapter, struct ena_ring *rx_ring,
sys/dev/ena/ena.c
1108
ena_refill_rx_bufs(struct ena_ring *rx_ring, uint32_t num)
sys/dev/ena/ena.c
1110
struct ena_adapter *adapter = rx_ring->adapter;
sys/dev/ena/ena.c
1116
ena_log_io(adapter->pdev, DBG, "refill qid: %d\n", rx_ring->qid);
sys/dev/ena/ena.c
1118
next_to_use = rx_ring->next_to_use;
sys/dev/ena/ena.c
1126
req_id = rx_ring->free_rx_ids[next_to_use];
sys/dev/ena/ena.c
1127
rx_info = &rx_ring->rx_buffer_info[req_id];
sys/dev/ena/ena.c
1129
if (ena_rx_ring_in_netmap(adapter, rx_ring->qid))
sys/dev/ena/ena.c
1130
rc = ena_netmap_alloc_rx_slot(adapter, rx_ring,
sys/dev/ena/ena.c
1134
rc = ena_alloc_rx_mbuf(adapter, rx_ring, rx_info);
sys/dev/ena/ena.c
1138
rx_ring->qid);
sys/dev/ena/ena.c
1141
rc = ena_com_add_single_rx_desc(rx_ring->ena_com_io_sq,
sys/dev/ena/ena.c
1146
rx_ring->qid);
sys/dev/ena/ena.c
1150
rx_ring->ring_size);
sys/dev/ena/ena.c
1154
counter_u64_add(rx_ring->rx_stats.refil_partial, 1);
sys/dev/ena/ena.c
1157
rx_ring->qid, i, num);
sys/dev/ena/ena.c
1161
ena_com_write_sq_doorbell(rx_ring->ena_com_io_sq);
sys/dev/ena/ena.c
1163
rx_ring->next_to_use = next_to_use;
sys/dev/ena/ena.c
1441
struct ena_ring *rx_ring = &adapter->rx_ring[qid];
sys/dev/ena/ena.c
1444
for (i = 0; i < rx_ring->ring_size; i++) {
sys/dev/ena/ena.c
1445
struct ena_rx_buffer *rx_info = &rx_ring->rx_buffer_info[i];
sys/dev/ena/ena.c
1448
ena_free_rx_mbuf(adapter, rx_ring, rx_info);
sys/dev/ena/ena.c
1453
ena_netmap_free_rx_slot(adapter, rx_ring,
sys/dev/ena/ena.c
1468
struct ena_ring *rx_ring;
sys/dev/ena/ena.c
1472
rx_ring = &adapter->rx_ring[i];
sys/dev/ena/ena.c
1473
bufs_num = rx_ring->ring_size - 1;
sys/dev/ena/ena.c
1474
rc = ena_refill_rx_bufs(rx_ring, bufs_num);
sys/dev/ena/ena.c
1481
rx_ring->initialized = true;
sys/dev/ena/ena.c
1647
ring = &adapter->rx_ring[i];
sys/dev/ena/ena.c
2148
adapter->rx_ring[i].ring_size = new_rx_size;
sys/dev/ena/ena.c
2207
cur_rx_ring_size = adapter->rx_ring[0].ring_size;
sys/dev/ena/ena.c
3149
struct ena_ring *rx_ring)
sys/dev/ena/ena.c
3151
if (likely(atomic_load_8(&rx_ring->first_interrupt)))
sys/dev/ena/ena.c
3154
if (ena_com_cq_empty(rx_ring->ena_com_io_cq))
sys/dev/ena/ena.c
3157
rx_ring->no_interrupt_event_cnt++;
sys/dev/ena/ena.c
3159
if (rx_ring->no_interrupt_event_cnt ==
sys/dev/ena/ena.c
3163
rx_ring->qid);
sys/dev/ena/ena.c
3296
struct ena_ring *rx_ring;
sys/dev/ena/ena.c
3315
rx_ring = &adapter->rx_ring[i];
sys/dev/ena/ena.c
3321
rc = check_for_rx_interrupt_queue(adapter, rx_ring);
sys/dev/ena/ena.c
3350
struct ena_ring *rx_ring;
sys/dev/ena/ena.c
3360
rx_ring = &adapter->rx_ring[i];
sys/dev/ena/ena.c
3363
rx_ring->ena_com_io_sq);
sys/dev/ena/ena.c
3364
if (unlikely(refill_required == (rx_ring->ring_size - 1))) {
sys/dev/ena/ena.c
3365
rx_ring->empty_rx_queue++;
sys/dev/ena/ena.c
3367
if (rx_ring->empty_rx_queue >= EMPTY_RX_REFILL) {
sys/dev/ena/ena.c
3368
counter_u64_add(rx_ring->rx_stats.empty_rx_ring,
sys/dev/ena/ena.c
3375
taskqueue_enqueue(rx_ring->que->cleanup_tq,
sys/dev/ena/ena.c
3376
&rx_ring->que->cleanup_task);
sys/dev/ena/ena.c
3377
rx_ring->empty_rx_queue = 0;
sys/dev/ena/ena.c
3380
rx_ring->empty_rx_queue = 0;
sys/dev/ena/ena.c
400
rxr = &adapter->rx_ring[i];
sys/dev/ena/ena.c
414
que->rx_ring = rxr;
sys/dev/ena/ena.c
432
rxr = &adapter->rx_ring[i];
sys/dev/ena/ena.c
475
struct ena_ring *rxr = &adapter->rx_ring[qid];
sys/dev/ena/ena.c
873
struct ena_ring *rx_ring = que->rx_ring;
sys/dev/ena/ena.c
876
size = sizeof(struct ena_rx_buffer) * rx_ring->ring_size;
sys/dev/ena/ena.c
880
rx_ring->initialized = false;
sys/dev/ena/ena.c
889
rx_ring->rx_buffer_info = malloc(size, M_DEVBUF, M_WAITOK | M_ZERO);
sys/dev/ena/ena.c
891
size = sizeof(uint16_t) * rx_ring->ring_size;
sys/dev/ena/ena.c
892
rx_ring->free_rx_ids = malloc(size, M_DEVBUF, M_WAITOK);
sys/dev/ena/ena.c
894
for (i = 0; i < rx_ring->ring_size; i++)
sys/dev/ena/ena.c
895
rx_ring->free_rx_ids[i] = i;
sys/dev/ena/ena.c
898
ena_reset_counters((counter_u64_t *)&rx_ring->rx_stats,
sys/dev/ena/ena.c
899
sizeof(rx_ring->rx_stats));
sys/dev/ena/ena.c
901
rx_ring->next_to_clean = 0;
sys/dev/ena/ena.c
902
rx_ring->next_to_use = 0;
sys/dev/ena/ena.c
905
for (i = 0; i < rx_ring->ring_size; i++) {
sys/dev/ena/ena.c
907
&(rx_ring->rx_buffer_info[i].map));
sys/dev/ena/ena.c
917
int err = tcp_lro_init(&rx_ring->lro);
sys/dev/ena/ena.c
924
rx_ring->lro.ifp = adapter->ifp;
sys/dev/ena/ena.c
933
rx_ring->rx_buffer_info[i].map);
sys/dev/ena/ena.c
936
free(rx_ring->free_rx_ids, M_DEVBUF);
sys/dev/ena/ena.c
937
rx_ring->free_rx_ids = NULL;
sys/dev/ena/ena.c
938
free(rx_ring->rx_buffer_info, M_DEVBUF);
sys/dev/ena/ena.c
939
rx_ring->rx_buffer_info = NULL;
sys/dev/ena/ena.c
953
struct ena_ring *rx_ring = &adapter->rx_ring[qid];
sys/dev/ena/ena.c
956
for (int i = 0; i < rx_ring->ring_size; i++) {
sys/dev/ena/ena.c
958
rx_ring->rx_buffer_info[i].map, BUS_DMASYNC_POSTREAD);
sys/dev/ena/ena.c
959
m_freem(rx_ring->rx_buffer_info[i].mbuf);
sys/dev/ena/ena.c
960
rx_ring->rx_buffer_info[i].mbuf = NULL;
sys/dev/ena/ena.c
962
rx_ring->rx_buffer_info[i].map);
sys/dev/ena/ena.c
964
rx_ring->rx_buffer_info[i].map);
sys/dev/ena/ena.c
968
tcp_lro_free(&rx_ring->lro);
sys/dev/ena/ena.c
971
free(rx_ring->rx_buffer_info, M_DEVBUF);
sys/dev/ena/ena.c
972
rx_ring->rx_buffer_info = NULL;
sys/dev/ena/ena.c
974
free(rx_ring->free_rx_ids, M_DEVBUF);
sys/dev/ena/ena.c
975
rx_ring->free_rx_ids = NULL;
sys/dev/ena/ena.h
223
struct ena_ring *rx_ring;
sys/dev/ena/ena.h
490
struct ena_ring rx_ring[ENA_MAX_NUM_IO_QUEUES]
sys/dev/ena/ena.h
580
int ena_refill_rx_bufs(struct ena_ring *rx_ring, uint32_t num);
sys/dev/ena/ena_datapath.c
100
rxc = ena_rx_cleanup(rx_ring);
sys/dev/ena/ena_datapath.c
344
ena_rx_hash_mbuf(struct ena_ring *rx_ring, struct ena_com_rx_ctx *ena_rx_ctx,
sys/dev/ena/ena_datapath.c
347
struct ena_adapter *adapter = rx_ring->adapter;
sys/dev/ena/ena_datapath.c
401
mbuf->m_pkthdr.flowid = rx_ring->qid;
sys/dev/ena/ena_datapath.c
415
ena_rx_mbuf(struct ena_ring *rx_ring, struct ena_com_rx_buf_info *ena_bufs,
sys/dev/ena/ena_datapath.c
426
adapter = rx_ring->adapter;
sys/dev/ena/ena_datapath.c
431
rx_info = &rx_ring->rx_buffer_info[req_id];
sys/dev/ena/ena_datapath.c
434
rx_ring->qid, req_id);
sys/dev/ena/ena_datapath.c
451
mbuf->m_pkthdr.rcvif = rx_ring->que->adapter->ifp;
sys/dev/ena/ena_datapath.c
454
ena_rx_hash_mbuf(rx_ring, ena_rx_ctx, mbuf);
sys/dev/ena/ena_datapath.c
460
bus_dmamap_unload(rx_ring->adapter->rx_buf_tag, rx_info->map);
sys/dev/ena/ena_datapath.c
463
rx_ring->free_rx_ids[ntc] = req_id;
sys/dev/ena/ena_datapath.c
464
ntc = ENA_RX_RING_IDX_NEXT(ntc, rx_ring->ring_size);
sys/dev/ena/ena_datapath.c
474
rx_info = &rx_ring->rx_buffer_info[req_id];
sys/dev/ena/ena_datapath.c
478
rx_ring->qid, req_id);
sys/dev/ena/ena_datapath.c
497
counter_u64_add(rx_ring->rx_stats.mbuf_alloc_fail, 1);
sys/dev/ena/ena_datapath.c
506
bus_dmamap_unload(rx_ring->adapter->rx_buf_tag, rx_info->map);
sys/dev/ena/ena_datapath.c
510
rx_ring->free_rx_ids[ntc] = req_id;
sys/dev/ena/ena_datapath.c
511
ntc = ENA_RX_RING_IDX_NEXT(ntc, rx_ring->ring_size);
sys/dev/ena/ena_datapath.c
523
ena_rx_checksum(struct ena_ring *rx_ring, struct ena_com_rx_ctx *ena_rx_ctx,
sys/dev/ena/ena_datapath.c
526
device_t pdev = rx_ring->adapter->pdev;
sys/dev/ena/ena_datapath.c
533
counter_u64_add(rx_ring->rx_stats.csum_bad, 1);
sys/dev/ena/ena_datapath.c
544
counter_u64_add(rx_ring->rx_stats.csum_bad, 1);
sys/dev/ena/ena_datapath.c
549
counter_u64_add(rx_ring->rx_stats.csum_good, 1);
sys/dev/ena/ena_datapath.c
559
ena_rx_cleanup(struct ena_ring *rx_ring)
sys/dev/ena/ena_datapath.c
581
adapter = rx_ring->que->adapter;
sys/dev/ena/ena_datapath.c
584
qid = rx_ring->que->id;
sys/dev/ena/ena_datapath.c
588
next_to_clean = rx_ring->next_to_clean;
sys/dev/ena/ena_datapath.c
591
if (netmap_rx_irq(adapter->ifp, rx_ring->qid, &done) != NM_IRQ_PASS)
sys/dev/ena/ena_datapath.c
598
ena_rx_ctx.ena_bufs = rx_ring->ena_bufs;
sys/dev/ena/ena_datapath.c
608
counter_u64_add(rx_ring->rx_stats.bad_desc_num,
sys/dev/ena/ena_datapath.c
614
counter_u64_add(rx_ring->rx_stats.bad_req_id,
sys/dev/ena/ena_datapath.c
627
rx_ring->qid, ena_rx_ctx.descs, ena_rx_ctx.l3_proto,
sys/dev/ena/ena_datapath.c
631
mbuf = ena_rx_mbuf(rx_ring, rx_ring->ena_bufs, &ena_rx_ctx,
sys/dev/ena/ena_datapath.c
638
rx_ring->free_rx_ids[next_to_clean] =
sys/dev/ena/ena_datapath.c
639
rx_ring->ena_bufs[i].req_id;
sys/dev/ena/ena_datapath.c
641
next_to_clean, rx_ring->ring_size);
sys/dev/ena/ena_datapath.c
648
ena_rx_checksum(rx_ring, &ena_rx_ctx, mbuf);
sys/dev/ena/ena_datapath.c
652
counter_u64_add_protected(rx_ring->rx_stats.bytes,
sys/dev/ena/ena_datapath.c
671
if ((rx_ring->lro.lro_cnt != 0) &&
sys/dev/ena/ena_datapath.c
672
(tcp_lro_rx(&rx_ring->lro, mbuf, 0) == 0))
sys/dev/ena/ena_datapath.c
682
counter_u64_add_protected(rx_ring->rx_stats.cnt, 1);
sys/dev/ena/ena_datapath.c
687
rx_ring->next_to_clean = next_to_clean;
sys/dev/ena/ena_datapath.c
691
rx_ring->ring_size / ENA_RX_REFILL_THRESH_DIVIDER,
sys/dev/ena/ena_datapath.c
695
ena_refill_rx_bufs(rx_ring, refill_required);
sys/dev/ena/ena_datapath.c
698
tcp_lro_flush_all(&rx_ring->lro);
sys/dev/ena/ena_datapath.c
72
struct ena_ring *rx_ring;
sys/dev/ena/ena_datapath.c
79
rx_ring = que->rx_ring;
sys/dev/ena/ena_datapath.c
97
atomic_store_8(&rx_ring->first_interrupt, 1);
sys/dev/ena/ena_netmap.c
104
ena_netmap_alloc_rx_slot(struct ena_adapter *adapter, struct ena_ring *rx_ring,
sys/dev/ena/ena_netmap.c
119
qid = rx_ring->qid;
sys/dev/ena/ena_netmap.c
129
if ((nm_i == head) && rx_ring->initialized) {
sys/dev/ena/ena_netmap.c
169
ena_netmap_free_rx_slot(struct ena_adapter *adapter, struct ena_ring *rx_ring,
sys/dev/ena/ena_netmap.c
188
qid = rx_ring->qid;
sys/dev/ena/ena_netmap.c
875
ctx.ring = &ctx.adapter->rx_ring[kring->ring_id];
sys/dev/ena/ena_netmap.h
51
struct ena_ring *rx_ring, struct ena_rx_buffer *rx_info);
sys/dev/ena/ena_netmap.h
53
struct ena_ring *rx_ring, struct ena_rx_buffer *rx_info);
sys/dev/ena/ena_sysctl.c
240
struct ena_ring *rx_ring;
sys/dev/ena/ena_sysctl.c
268
rx_ring = adapter->rx_ring;
sys/dev/ena/ena_sysctl.c
309
for (i = 0; i < adapter->num_io_queues; ++i, ++tx_ring, ++rx_ring) {
sys/dev/ena/ena_sysctl.c
372
rx_stats = &rx_ring->rx_stats;
sys/dev/et/if_et.c
1037
struct et_rxdesc_ring *rx_ring;
sys/dev/et/if_et.c
1095
rx_ring = &sc->sc_rx_ring[0];
sys/dev/et/if_et.c
1096
et_dma_ring_free(sc, &rx_ring->rr_dtag, (void *)&rx_ring->rr_desc,
sys/dev/et/if_et.c
1097
rx_ring->rr_dmap, &rx_ring->rr_paddr);
sys/dev/et/if_et.c
1099
rx_ring = &sc->sc_rx_ring[1];
sys/dev/et/if_et.c
1100
et_dma_ring_free(sc, &rx_ring->rr_dtag, (void *)&rx_ring->rr_desc,
sys/dev/et/if_et.c
1101
rx_ring->rr_dmap, &rx_ring->rr_paddr);
sys/dev/et/if_et.c
1513
struct et_rxdesc_ring *rx_ring;
sys/dev/et/if_et.c
1518
rx_ring = &sc->sc_rx_ring[0];
sys/dev/et/if_et.c
1523
bus_dmamap_sync(sc->sc_rx_mini_tag, rx_ring->rr_dmap,
sys/dev/et/if_et.c
1532
rx_ring = &sc->sc_rx_ring[1];
sys/dev/et/if_et.c
1537
bus_dmamap_sync(sc->sc_rx_tag, rx_ring->rr_dmap,
sys/dev/et/if_et.c
1727
struct et_rxdesc_ring *rx_ring;
sys/dev/et/if_et.c
1760
rx_ring = &sc->sc_rx_ring[1];
sys/dev/et/if_et.c
1761
CSR_WRITE_4(sc, ET_RX_RING1_HI, ET_ADDR_HI(rx_ring->rr_paddr));
sys/dev/et/if_et.c
1762
CSR_WRITE_4(sc, ET_RX_RING1_LO, ET_ADDR_LO(rx_ring->rr_paddr));
sys/dev/et/if_et.c
1768
rx_ring->rr_index = 0;
sys/dev/et/if_et.c
1769
rx_ring->rr_wrap = 1;
sys/dev/et/if_et.c
1774
rx_ring = &sc->sc_rx_ring[0];
sys/dev/et/if_et.c
1775
CSR_WRITE_4(sc, ET_RX_RING0_HI, ET_ADDR_HI(rx_ring->rr_paddr));
sys/dev/et/if_et.c
1776
CSR_WRITE_4(sc, ET_RX_RING0_LO, ET_ADDR_LO(rx_ring->rr_paddr));
sys/dev/et/if_et.c
1782
rx_ring->rr_index = 0;
sys/dev/et/if_et.c
1783
rx_ring->rr_wrap = 1;
sys/dev/et/if_et.c
2026
struct et_rxdesc_ring *rx_ring;
sys/dev/et/if_et.c
2114
rx_ring = &sc->sc_rx_ring[ring_idx];
sys/dev/et/if_et.c
2115
if (buf_idx != rx_ring->rr_index) {
sys/dev/et/if_et.c
2118
ring_idx, buf_idx, rx_ring->rr_index);
sys/dev/et/if_et.c
2121
MPASS(rx_ring->rr_index < ET_RX_NDESC);
sys/dev/et/if_et.c
2122
if (++rx_ring->rr_index == ET_RX_NDESC) {
sys/dev/et/if_et.c
2123
rx_ring->rr_index = 0;
sys/dev/et/if_et.c
2124
rx_ring->rr_wrap ^= 1;
sys/dev/et/if_et.c
2126
rxring_pos = rx_ring->rr_index & ET_RX_RING_POS_INDEX_MASK;
sys/dev/et/if_et.c
2127
if (rx_ring->rr_wrap)
sys/dev/et/if_et.c
2129
CSR_WRITE_4(sc, rx_ring->rr_posreg, rxring_pos);
sys/dev/et/if_et.c
858
struct et_rxdesc_ring *rx_ring;
sys/dev/et/if_et.c
892
rx_ring = &sc->sc_rx_ring[0];
sys/dev/et/if_et.c
894
&rx_ring->rr_dtag, (uint8_t **)&rx_ring->rr_desc, &rx_ring->rr_dmap,
sys/dev/et/if_et.c
895
&rx_ring->rr_paddr, "RX ring 0");
sys/dev/et/if_et.c
896
rx_ring->rr_posreg = ET_RX_RING0_POS;
sys/dev/et/if_et.c
901
rx_ring = &sc->sc_rx_ring[1];
sys/dev/et/if_et.c
903
&rx_ring->rr_dtag, (uint8_t **)&rx_ring->rr_desc, &rx_ring->rr_dmap,
sys/dev/et/if_et.c
904
&rx_ring->rr_paddr, "RX ring 1");
sys/dev/et/if_et.c
905
rx_ring->rr_posreg = ET_RX_RING1_POS;
sys/dev/iavf/iavf_iflib.h
220
struct rx_ring rxr;
sys/dev/iavf/iavf_lib.c
789
struct rx_ring *rxr = &que->rxr;
sys/dev/iavf/iavf_txrx_iflib.c
574
struct rx_ring *rxr = &((vsi->rx_queues[iru->iru_qsidx]).rxr);
sys/dev/iavf/iavf_txrx_iflib.c
605
struct rx_ring *rxr = &vsi->rx_queues[rxqid].rxr;
sys/dev/iavf/iavf_txrx_iflib.c
626
struct rx_ring *rxr = &vsi->rx_queues[rxqid].rxr;
sys/dev/iavf/iavf_txrx_iflib.c
669
struct rx_ring *rxr = &que->rxr;
sys/dev/iavf/iavf_vc_iflib.c
61
struct rx_ring *rxr;
sys/dev/iavf/if_iavf_iflib.c
1053
struct rx_ring *rxr = &que->rxr;
sys/dev/iavf/if_iavf_iflib.c
1709
struct rx_ring *rxr = &que->rxr;
sys/dev/iavf/if_iavf_iflib.c
1940
struct rx_ring *rxr;
sys/dev/iavf/if_iavf_iflib.c
689
struct rx_ring *rxr;
sys/dev/igc/if_igc.c
1051
struct rx_ring *rxr = &que->rxr;
sys/dev/igc/if_igc.c
1125
struct rx_ring *rxr = &que->rxr;
sys/dev/igc/if_igc.c
146
struct tx_ring *, struct rx_ring *);
sys/dev/igc/if_igc.c
2076
struct rx_ring *rxr = &que->rxr;
sys/dev/igc/if_igc.c
2284
struct rx_ring *rxr = &que->rxr;
sys/dev/igc/if_igc.c
2742
struct rx_ring *rxr = &rx_que->rxr;
sys/dev/igc/if_igc.c
3280
struct rx_ring *rxr = &sc->rx_queues->rxr;
sys/dev/igc/if_igc.c
396
struct rx_ring *rxr = &rx_que->rxr;
sys/dev/igc/if_igc.c
908
struct tx_ring *txr, struct rx_ring *rxr)
sys/dev/igc/if_igc.h
291
struct rx_ring rxr;
sys/dev/igc/igc_txrx.c
398
struct rx_ring *rxr = &que->rxr;
sys/dev/igc/igc_txrx.c
423
struct rx_ring *rxr = &que->rxr;
sys/dev/igc/igc_txrx.c
434
struct rx_ring *rxr = &que->rxr;
sys/dev/igc/igc_txrx.c
466
struct rx_ring *rxr = &que->rxr;
sys/dev/ixgbe/if_ix.c
2061
struct rx_ring *rxr = &rx_que->rxr;
sys/dev/ixgbe/if_ix.c
2255
struct rx_ring *rxr = ((struct rx_ring *)oidp->oid_arg1);
sys/dev/ixgbe/if_ix.c
2281
struct rx_ring *rxr = ((struct rx_ring *)oidp->oid_arg1);
sys/dev/ixgbe/if_ix.c
2348
struct rx_ring *rxr;
sys/dev/ixgbe/if_ix.c
2598
struct rx_ring *rxr = &que->rxr;
sys/dev/ixgbe/if_ix.c
3882
struct rx_ring *rxr = &rx_que->rxr;
sys/dev/ixgbe/if_ix.c
4052
struct rx_ring *rxr = &rx_que->rxr;
sys/dev/ixgbe/if_ix.c
5031
struct rx_ring *rxr;
sys/dev/ixgbe/if_ix.c
5057
struct rx_ring *rxr;
sys/dev/ixgbe/if_ix.c
571
struct rx_ring *rxr = &que->rxr;
sys/dev/ixgbe/if_ix.c
780
struct rx_ring *rxr = &que->rxr;
sys/dev/ixgbe/if_ixv.c
1431
struct rx_ring *rxr = &que->rxr;
sys/dev/ixgbe/if_ixv.c
1869
struct rx_ring *rxr = &rx_que->rxr;
sys/dev/ixgbe/if_ixv.c
329
struct rx_ring *rxr = &que->rxr;
sys/dev/ixgbe/ix_txrx.c
328
struct rx_ring *rxr = &que->rxr;
sys/dev/ixgbe/ix_txrx.c
354
struct rx_ring *rxr = &que->rxr;
sys/dev/ixgbe/ix_txrx.c
367
struct rx_ring *rxr = &que->rxr;
sys/dev/ixgbe/ix_txrx.c
402
struct rx_ring *rxr = &que->rxr;
sys/dev/ixgbe/ixgbe.h
335
struct rx_ring rxr;
sys/dev/ixl/if_ixl.c
1294
struct rx_ring *rxr = &que->rxr;
sys/dev/ixl/ixl.h
393
struct rx_ring rxr;
sys/dev/ixl/ixl_pf_iflib.c
567
struct rx_ring *rxr = &rx_que->rxr;
sys/dev/ixl/ixl_pf_iflib.c
625
struct rx_ring *rxr = &que->rxr;
sys/dev/ixl/ixl_pf_iflib.c
65
struct rx_ring *rxr = &que->rxr;
sys/dev/ixl/ixl_txrx.c
551
struct rx_ring *rxr = &((vsi->rx_queues[iru->iru_qsidx]).rxr);
sys/dev/ixl/ixl_txrx.c
572
struct rx_ring *rxr = &vsi->rx_queues[rxqid].rxr;
sys/dev/ixl/ixl_txrx.c
581
struct rx_ring *rxr = &vsi->rx_queues[rxqid].rxr;
sys/dev/ixl/ixl_txrx.c
662
struct rx_ring *rxr = &que->rxr;
sys/dev/ixl/ixl_txrx.c
907
struct rx_ring *rxr;
sys/dev/mlx4/mlx4_en/en.h
595
struct mlx4_en_rx_ring *rx_ring[MAX_RX_RINGS];
sys/dev/mlx4/mlx4_en/en.h
721
struct mlx4_en_rx_ring *rx_ring = priv->rx_ring[cq->ring];
sys/dev/mlx4/mlx4_en/en.h
726
rx_ring->yields++;
sys/dev/mlx4/mlx4_en/en.h
876
struct mlx4_en_rx_ring *rx_ring);
sys/dev/mlx4/mlx4_en/mlx4_en_cq.c
155
cq->size = priv->rx_ring[cq->ring]->actual_size;
sys/dev/mlx4/mlx4_en/mlx4_en_netdev.c
1135
rx_packets = priv->rx_ring[ring]->packets;
sys/dev/mlx4/mlx4_en/mlx4_en_netdev.c
1136
rx_bytes = priv->rx_ring[ring]->bytes;
sys/dev/mlx4/mlx4_en/mlx4_en_netdev.c
1315
priv->rx_ring[i]->cqn = cq->mcq.cqn;
sys/dev/mlx4/mlx4_en/mlx4_en_netdev.c
1447
mlx4_en_deactivate_rx_ring(priv, priv->rx_ring[i]);
sys/dev/mlx4/mlx4_en/mlx4_en_netdev.c
1556
mlx4_en_deactivate_rx_ring(priv, priv->rx_ring[i]);
sys/dev/mlx4/mlx4_en/mlx4_en_netdev.c
1625
priv->rx_ring[i]->bytes = 0;
sys/dev/mlx4/mlx4_en/mlx4_en_netdev.c
1626
priv->rx_ring[i]->packets = 0;
sys/dev/mlx4/mlx4_en/mlx4_en_netdev.c
1627
priv->rx_ring[i]->csum_ok = 0;
sys/dev/mlx4/mlx4_en/mlx4_en_netdev.c
1628
priv->rx_ring[i]->csum_none = 0;
sys/dev/mlx4/mlx4_en/mlx4_en_netdev.c
1683
if (priv->rx_ring[i])
sys/dev/mlx4/mlx4_en/mlx4_en_netdev.c
1684
mlx4_en_destroy_rx_ring(priv, &priv->rx_ring[i],
sys/dev/mlx4/mlx4_en/mlx4_en_netdev.c
1706
if (mlx4_en_create_rx_ring(priv, &priv->rx_ring[i],
sys/dev/mlx4/mlx4_en/mlx4_en_netdev.c
1734
if (priv->rx_ring[i])
sys/dev/mlx4/mlx4_en/mlx4_en_netdev.c
1735
mlx4_en_destroy_rx_ring(priv, &priv->rx_ring[i],
sys/dev/mlx4/mlx4_en/mlx4_en_netdev.c
2361
priv->rx_ring[0]->actual_size : priv->rx_ring[0]->size) &&
sys/dev/mlx4/mlx4_en/mlx4_en_netdev.c
2732
struct mlx4_en_rx_ring *rx_ring;
sys/dev/mlx4/mlx4_en/mlx4_en_netdev.c
2875
rx_ring = priv->rx_ring[i];
sys/dev/mlx4/mlx4_en/mlx4_en_netdev.c
2881
CTLFLAG_RD, &rx_ring->packets, 0, "RX packets");
sys/dev/mlx4/mlx4_en/mlx4_en_netdev.c
2883
CTLFLAG_RD, &rx_ring->bytes, 0, "RX bytes");
sys/dev/mlx4/mlx4_en/mlx4_en_netdev.c
2885
CTLFLAG_RD, &rx_ring->errors, 0, "RX soft errors");
sys/dev/mlx4/mlx4_en/mlx4_en_netdev.c
69
struct mlx4_en_rx_ring *rx_ring = priv->rx_ring[cq->ring];
sys/dev/mlx4/mlx4_en/mlx4_en_netdev.c
81
rx_ring->cleaned += done;
sys/dev/mlx4/mlx4_en/mlx4_en_netdev.c
83
rx_ring->misses++;
sys/dev/mlx4/mlx4_en/mlx4_en_port.c
165
const struct mlx4_en_rx_ring *ring = priv->rx_ring[i];
sys/dev/mlx4/mlx4_en/mlx4_en_port.c
214
priv->port_stats.rx_chksum_good += priv->rx_ring[i]->csum_ok;
sys/dev/mlx4/mlx4_en/mlx4_en_port.c
215
priv->port_stats.rx_chksum_none += priv->rx_ring[i]->csum_none;
sys/dev/mlx4/mlx4_en/mlx4_en_port.c
439
priv->pkstats.rx_packets += priv->rx_ring[i]->packets;
sys/dev/mlx4/mlx4_en/mlx4_en_port.c
440
priv->pkstats.rx_bytes += priv->rx_ring[i]->bytes;
sys/dev/mlx4/mlx4_en/mlx4_en_port.c
441
priv->port_stats.rx_chksum_good += priv->rx_ring[i]->csum_ok;
sys/dev/mlx4/mlx4_en/mlx4_en_port.c
442
priv->port_stats.rx_chksum_none += priv->rx_ring[i]->csum_none;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
1052
priv->rx_ring[i]->qpn = rss_map->base_qpn + i;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
1053
err = mlx4_en_config_rss_qp(priv, priv->rx_ring[i]->qpn,
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
1054
priv->rx_ring[i],
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
1071
priv->rx_ring[0]->cqn, -1, &context);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
265
ring = priv->rx_ring[ring_ind];
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
291
ring = priv->rx_ring[ring_ind];
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
473
ring = priv->rx_ring[ring_ind];
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
520
ring = priv->rx_ring[ring_ind];
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
530
mlx4_en_free_rx_buf(priv, priv->rx_ring[ring_ind]);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
535
ring = priv->rx_ring[ring_ind];
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
742
struct mlx4_en_rx_ring *ring = priv->rx_ring[cq->ring];
sys/dev/neta/if_mvneta.c
1604
mtx_lock(&sc->rx_ring[q].ring_mtx);
sys/dev/neta/if_mvneta.c
1613
mtx_unlock(&sc->rx_ring[q].ring_mtx);
sys/dev/neta/if_mvnetavar.h
213
KASSERT(mtx_owned(&(sc)->rx_ring[(q)].ring_mtx),\
sys/dev/neta/if_mvnetavar.h
289
struct mvneta_rx_ring rx_ring[MVNETA_RX_QNUM_MAX];
sys/dev/neta/if_mvnetavar.h
317
(&(sc)->rx_ring[(q)])
sys/dev/nfe/if_nfe.c
2834
struct nfe_rx_ring *rx_ring;
sys/dev/nfe/if_nfe.c
2860
rx_ring = &sc->rxq;
sys/dev/nfe/if_nfe.c
2862
rdata = &rx_ring->data[i];
sys/dev/nfe/if_nfe.c
2864
bus_dmamap_sync(rx_ring->rx_data_tag,
sys/dev/nfe/if_nfe.c
2866
bus_dmamap_unload(rx_ring->rx_data_tag,
sys/dev/qat/qat_common/adf_freebsd_uio_cleanup.c
67
int rx_ring = i + tx_rx_gap;
sys/dev/qat/qat_common/adf_freebsd_uio_cleanup.c
69
if (!test_bit(rx_ring, &orphan->rx_mask)) {
sys/dev/qlxgbe/ql_def.h
199
qla_rx_ring_t rx_ring[MAX_RDS_RINGS];
sys/dev/qlxgbe/ql_isr.c
202
qla_rx_ring_t *rx_ring;
sys/dev/qlxgbe/ql_isr.c
209
rx_ring = &ha->rx_ring[r_idx];
sys/dev/qlxgbe/ql_isr.c
225
rxb = &rx_ring->rx_buf[sgc->handle[i] & 0x7FFF];
sys/dev/qlxgbe/ql_isr.c
68
qla_rx_ring_t *rx_ring;
sys/dev/qlxgbe/ql_isr.c
81
rx_ring = &ha->rx_ring[r_idx];
sys/dev/qlxgbe/ql_isr.c
84
rxb = &rx_ring->rx_buf[sgc->handle[i] & 0x7FFF];
sys/dev/qlxgbe/ql_os.c
1786
qla_rx_ring_t *rx_ring;
sys/dev/qlxgbe/ql_os.c
1789
rx_ring = &ha->rx_ring[r];
sys/dev/qlxgbe/ql_os.c
1792
rxb = &rx_ring->rx_buf[i];
sys/dev/qlxgbe/ql_os.c
1805
rxb = &ha->rx_ring[k].rx_buf[j];
sys/dev/qlxgbe/ql_os.c
1813
rx_ring->rx_buf[j].map);
sys/dev/qlxgbe/ql_os.c
1823
rx_ring = &ha->rx_ring[r];
sys/dev/qlxgbe/ql_os.c
1826
rxb = &rx_ring->rx_buf[i];
sys/dev/qlxgbe/ql_os.c
1860
rxb = &ha->rx_ring[r].rx_buf[i];
sys/dev/qlxgbe/ql_os.c
1895
bzero((void *)ha->rx_ring, (sizeof(qla_rx_ring_t) * MAX_RDS_RINGS));
sys/dev/qlxgbe/ql_os.c
1920
bzero((void *)ha->rx_ring, (sizeof(qla_rx_ring_t) * MAX_RDS_RINGS));
sys/dev/qlxge/qls_dbg.c
274
qls_dump_buf32(ha, "cq_icb", ha->rx_ring[0].cq_icb_vaddr,
sys/dev/qlxge/qls_dbg.c
278
(void *)ha->rx_ring[0].lbq_addr_tbl_paddr);
sys/dev/qlxge/qls_dbg.c
280
qls_dump_buf32(ha, "lbq_addr_tbl", ha->rx_ring[0].lbq_addr_tbl_vaddr,
sys/dev/qlxge/qls_dbg.c
284
(void *)ha->rx_ring[0].lbq_paddr);
sys/dev/qlxge/qls_dbg.c
286
qls_dump_buf32(ha, "lbq", ha->rx_ring[0].lbq_vaddr,
sys/dev/qlxge/qls_dbg.c
290
(void *)ha->rx_ring[0].sbq_addr_tbl_paddr);
sys/dev/qlxge/qls_dbg.c
292
qls_dump_buf32(ha, "sbq_addr_tbl", ha->rx_ring[0].sbq_addr_tbl_vaddr,
sys/dev/qlxge/qls_dbg.c
296
(void *)ha->rx_ring[0].sbq_paddr);
sys/dev/qlxge/qls_dbg.c
298
qls_dump_buf32(ha, "sbq", ha->rx_ring[0].sbq_vaddr,
sys/dev/qlxge/qls_dbg.c
302
(void *)ha->rx_ring[0].lb_paddr);
sys/dev/qlxge/qls_def.h
309
qla_rx_ring_t rx_ring[MAX_RX_RINGS];
sys/dev/qlxge/qls_hw.c
1049
rxr = &ha->rx_ring[cid];
sys/dev/qlxge/qls_hw.c
1538
if (ha->rx_ring[ridx].flags.cq_dma) {
sys/dev/qlxge/qls_hw.c
1539
qls_free_dmabuf(ha, &ha->rx_ring[ridx].cq_dma);
sys/dev/qlxge/qls_hw.c
1540
ha->rx_ring[ridx].flags.cq_dma = 0;
sys/dev/qlxge/qls_hw.c
1543
if (ha->rx_ring[ridx].flags.lbq_dma) {
sys/dev/qlxge/qls_hw.c
1544
qls_free_dmabuf(ha, &ha->rx_ring[ridx].lbq_dma);
sys/dev/qlxge/qls_hw.c
1545
ha->rx_ring[ridx].flags.lbq_dma = 0;
sys/dev/qlxge/qls_hw.c
1548
if (ha->rx_ring[ridx].flags.sbq_dma) {
sys/dev/qlxge/qls_hw.c
1549
qls_free_dmabuf(ha, &ha->rx_ring[ridx].sbq_dma);
sys/dev/qlxge/qls_hw.c
1550
ha->rx_ring[ridx].flags.sbq_dma = 0;
sys/dev/qlxge/qls_hw.c
1553
if (ha->rx_ring[ridx].flags.lb_dma) {
sys/dev/qlxge/qls_hw.c
1554
qls_free_dmabuf(ha, &ha->rx_ring[ridx].lb_dma);
sys/dev/qlxge/qls_hw.c
1555
ha->rx_ring[ridx].flags.lb_dma = 0;
sys/dev/qlxge/qls_hw.c
1586
ha->rx_ring[ridx].cq_dma.alignment = 128;
sys/dev/qlxge/qls_hw.c
1587
ha->rx_ring[ridx].cq_dma.size =
sys/dev/qlxge/qls_hw.c
1590
ret = qls_alloc_dmabuf(ha, &ha->rx_ring[ridx].cq_dma);
sys/dev/qlxge/qls_hw.c
1596
ha->rx_ring[ridx].flags.cq_dma = 1;
sys/dev/qlxge/qls_hw.c
1598
ha->rx_ring[ridx].lbq_dma.alignment = 8;
sys/dev/qlxge/qls_hw.c
1599
ha->rx_ring[ridx].lbq_dma.size = QLA_LGBQ_AND_TABLE_SIZE;
sys/dev/qlxge/qls_hw.c
1601
ret = qls_alloc_dmabuf(ha, &ha->rx_ring[ridx].lbq_dma);
sys/dev/qlxge/qls_hw.c
1607
ha->rx_ring[ridx].flags.lbq_dma = 1;
sys/dev/qlxge/qls_hw.c
1609
ha->rx_ring[ridx].sbq_dma.alignment = 8;
sys/dev/qlxge/qls_hw.c
1610
ha->rx_ring[ridx].sbq_dma.size = QLA_SMBQ_AND_TABLE_SIZE;
sys/dev/qlxge/qls_hw.c
1612
ret = qls_alloc_dmabuf(ha, &ha->rx_ring[ridx].sbq_dma);
sys/dev/qlxge/qls_hw.c
1618
ha->rx_ring[ridx].flags.sbq_dma = 1;
sys/dev/qlxge/qls_hw.c
1620
ha->rx_ring[ridx].lb_dma.alignment = 8;
sys/dev/qlxge/qls_hw.c
1621
ha->rx_ring[ridx].lb_dma.size = (QLA_LGB_SIZE * QLA_NUM_LGB_ENTRIES);
sys/dev/qlxge/qls_hw.c
1623
ret = qls_alloc_dmabuf(ha, &ha->rx_ring[ridx].lb_dma);
sys/dev/qlxge/qls_hw.c
1628
ha->rx_ring[ridx].flags.lb_dma = 1;
sys/dev/qlxge/qls_hw.c
1630
bzero(ha->rx_ring[ridx].cq_dma.dma_b, ha->rx_ring[ridx].cq_dma.size);
sys/dev/qlxge/qls_hw.c
1631
bzero(ha->rx_ring[ridx].lbq_dma.dma_b, ha->rx_ring[ridx].lbq_dma.size);
sys/dev/qlxge/qls_hw.c
1632
bzero(ha->rx_ring[ridx].sbq_dma.dma_b, ha->rx_ring[ridx].sbq_dma.size);
sys/dev/qlxge/qls_hw.c
1633
bzero(ha->rx_ring[ridx].lb_dma.dma_b, ha->rx_ring[ridx].lb_dma.size);
sys/dev/qlxge/qls_hw.c
1636
ha->rx_ring[ridx].cq_base_vaddr = ha->rx_ring[ridx].cq_dma.dma_b;
sys/dev/qlxge/qls_hw.c
1637
ha->rx_ring[ridx].cq_base_paddr = ha->rx_ring[ridx].cq_dma.dma_addr;
sys/dev/qlxge/qls_hw.c
1639
v_addr = ha->rx_ring[ridx].cq_dma.dma_b;
sys/dev/qlxge/qls_hw.c
1640
p_addr = ha->rx_ring[ridx].cq_dma.dma_addr;
sys/dev/qlxge/qls_hw.c
1646
ha->rx_ring[ridx].cq_icb_vaddr = v_addr;
sys/dev/qlxge/qls_hw.c
1647
ha->rx_ring[ridx].cq_icb_paddr = p_addr;
sys/dev/qlxge/qls_hw.c
1653
ha->rx_ring[ridx].cqi_vaddr = (uint32_t *)v_addr;
sys/dev/qlxge/qls_hw.c
1654
ha->rx_ring[ridx].cqi_paddr = p_addr;
sys/dev/qlxge/qls_hw.c
1656
v_addr = ha->rx_ring[ridx].lbq_dma.dma_b;
sys/dev/qlxge/qls_hw.c
1657
p_addr = ha->rx_ring[ridx].lbq_dma.dma_addr;
sys/dev/qlxge/qls_hw.c
1660
ha->rx_ring[ridx].lbq_addr_tbl_vaddr = v_addr;
sys/dev/qlxge/qls_hw.c
1661
ha->rx_ring[ridx].lbq_addr_tbl_paddr = p_addr;
sys/dev/qlxge/qls_hw.c
1664
ha->rx_ring[ridx].lbq_vaddr = v_addr + PAGE_SIZE;
sys/dev/qlxge/qls_hw.c
1665
ha->rx_ring[ridx].lbq_paddr = p_addr + PAGE_SIZE;
sys/dev/qlxge/qls_hw.c
1667
v_addr = ha->rx_ring[ridx].sbq_dma.dma_b;
sys/dev/qlxge/qls_hw.c
1668
p_addr = ha->rx_ring[ridx].sbq_dma.dma_addr;
sys/dev/qlxge/qls_hw.c
1671
ha->rx_ring[ridx].sbq_addr_tbl_vaddr = v_addr;
sys/dev/qlxge/qls_hw.c
1672
ha->rx_ring[ridx].sbq_addr_tbl_paddr = p_addr;
sys/dev/qlxge/qls_hw.c
1675
ha->rx_ring[ridx].sbq_vaddr = v_addr + PAGE_SIZE;
sys/dev/qlxge/qls_hw.c
1676
ha->rx_ring[ridx].sbq_paddr = p_addr + PAGE_SIZE;
sys/dev/qlxge/qls_hw.c
1678
ha->rx_ring[ridx].lb_vaddr = ha->rx_ring[ridx].lb_dma.dma_b;
sys/dev/qlxge/qls_hw.c
1679
ha->rx_ring[ridx].lb_paddr = ha->rx_ring[ridx].lb_dma.dma_addr;
sys/dev/qlxge/qls_hw.c
1683
p_addr = ha->rx_ring[ridx].lbq_paddr;
sys/dev/qlxge/qls_hw.c
1684
bq_e = ha->rx_ring[ridx].lbq_addr_tbl_vaddr;
sys/dev/qlxge/qls_hw.c
1689
p_addr = ha->rx_ring[ridx].lb_paddr;
sys/dev/qlxge/qls_hw.c
1690
bq_e = ha->rx_ring[ridx].lbq_vaddr;
sys/dev/qlxge/qls_hw.c
1702
p_addr = ha->rx_ring[ridx].sbq_paddr;
sys/dev/qlxge/qls_hw.c
1703
bq_e = ha->rx_ring[ridx].sbq_addr_tbl_vaddr;
sys/dev/qlxge/qls_hw.c
928
ha->rx_ring[0].cq_db_offset));
sys/dev/qlxge/qls_hw.c
934
Q81_WR_LBQ_PROD_IDX(i, ha->rx_ring[i].lbq_in);
sys/dev/qlxge/qls_hw.c
935
Q81_WR_SBQ_PROD_IDX(i, ha->rx_ring[i].sbq_in);
sys/dev/qlxge/qls_hw.h
934
(ha->rx_ring[cq_idx].cq_db_offset + Q81_COMPQ_INDEX_REG), idx)
sys/dev/qlxge/qls_hw.h
937
(ha->rx_ring[cq_idx].cq_db_offset + Q81_COMPQ_INDEX_REG))
sys/dev/qlxge/qls_hw.h
940
(ha->rx_ring[cq_idx].cq_db_offset + Q81_COMPQ_VALID_REG),\
sys/dev/qlxge/qls_hw.h
944
(ha->rx_ring[cq_idx].cq_db_offset + Q81_COMPQ_VALID_REG),\
sys/dev/qlxge/qls_hw.h
948
(ha->rx_ring[cq_idx].cq_db_offset + Q81_LRGBQ_INDEX_REG), idx)
sys/dev/qlxge/qls_hw.h
951
(ha->rx_ring[cq_idx].cq_db_offset + Q81_LRGBQ_INDEX_REG))
sys/dev/qlxge/qls_hw.h
954
(ha->rx_ring[cq_idx].cq_db_offset + Q81_SMBQ_INDEX_REG), idx)
sys/dev/qlxge/qls_hw.h
957
(ha->rx_ring[cq_idx].cq_db_offset + Q81_SMBQ_INDEX_REG))
sys/dev/qlxge/qls_isr.c
134
rxr = &ha->rx_ring[rxr_idx];
sys/dev/qlxge/qls_isr.c
235
struct lro_ctrl *lro = &ha->rx_ring[cq_idx].lro;
sys/dev/qlxge/qls_isr.c
238
cq_b = ha->rx_ring[cq_idx].cq_base_vaddr;
sys/dev/qlxge/qls_isr.c
240
cq_comp_idx = *(ha->rx_ring[cq_idx].cqi_vaddr);
sys/dev/qlxge/qls_isr.c
242
i = ha->rx_ring[cq_idx].cq_next;
sys/dev/qlxge/qls_isr.c
263
__func__, i, (*(ha->rx_ring[cq_idx].cqi_vaddr)),
sys/dev/qlxge/qls_isr.c
279
cq_comp_idx = *(ha->rx_ring[cq_idx].cqi_vaddr);
sys/dev/qlxge/qls_isr.c
292
ha->rx_ring[cq_idx].cq_next = cq_comp_idx;
sys/dev/qlxge/qls_isr.c
295
Q81_WR_CQ_CONS_IDX(cq_idx, (ha->rx_ring[cq_idx].cq_next));
sys/dev/qlxge/qls_isr.c
80
rxr = &ha->rx_ring[r_idx];
sys/dev/qlxge/qls_os.c
1259
qla_rx_ring_t *rx_ring;
sys/dev/qlxge/qls_os.c
1262
rx_ring = &ha->rx_ring[r];
sys/dev/qlxge/qls_os.c
1265
rxb = &rx_ring->rx_buf[i];
sys/dev/qlxge/qls_os.c
1274
rxb = &rx_ring->rx_buf[j];
sys/dev/qlxge/qls_os.c
1281
rx_ring = &ha->rx_ring[r];
sys/dev/qlxge/qls_os.c
1283
sbq_e = rx_ring->sbq_vaddr;
sys/dev/qlxge/qls_os.c
1285
rxb = &rx_ring->rx_buf[0];
sys/dev/qlxge/qls_os.c
1323
rxr = &ha->rx_ring[r];
sys/dev/qlxge/qls_os.c
1346
rxr = &ha->rx_ring[r];
sys/dev/qlxge/qls_os.c
1453
lro = &ha->rx_ring[i].lro;
sys/dev/qlxge/qls_os.c
1479
lro = &ha->rx_ring[i].lro;
sys/dev/qlxge/qls_os.c
203
(void *)ha->rx_ring[i].rx_int));
sys/dev/qlxge/qls_os.c
208
(void *)ha->rx_ring[i].rss_int));
sys/dev/qlxge/qls_os.c
213
ha->rx_ring[i].lbq_next);
sys/dev/qlxge/qls_os.c
218
ha->rx_ring[i].lbq_free);
sys/dev/qlxge/qls_os.c
223
ha->rx_ring[i].lbq_in);
sys/dev/qlxge/qls_os.c
228
ha->rx_ring[i].sbq_next);
sys/dev/qlxge/qls_os.c
233
ha->rx_ring[i].sbq_free);
sys/dev/qlxge/qls_os.c
238
ha->rx_ring[i].sbq_in);
sys/dev/rtwn/pci/rtwn_pci_attach.c
125
struct rtwn_rx_ring *rx_ring = &pc->rx_ring;
sys/dev/rtwn/pci/rtwn_pci_attach.c
134
size, 1, size, 0, NULL, NULL, &rx_ring->desc_dmat);
sys/dev/rtwn/pci/rtwn_pci_attach.c
140
error = bus_dmamem_alloc(rx_ring->desc_dmat, (void **)&rx_ring->desc,
sys/dev/rtwn/pci/rtwn_pci_attach.c
142
&rx_ring->desc_map);
sys/dev/rtwn/pci/rtwn_pci_attach.c
147
error = bus_dmamap_load(rx_ring->desc_dmat, rx_ring->desc_map,
sys/dev/rtwn/pci/rtwn_pci_attach.c
148
rx_ring->desc, size, rtwn_pci_dma_map_addr, &rx_ring->paddr, 0);
sys/dev/rtwn/pci/rtwn_pci_attach.c
153
bus_dmamap_sync(rx_ring->desc_dmat, rx_ring->desc_map,
sys/dev/rtwn/pci/rtwn_pci_attach.c
159
MJUMPAGESIZE, 1, MJUMPAGESIZE, 0, NULL, NULL, &rx_ring->data_dmat);
sys/dev/rtwn/pci/rtwn_pci_attach.c
167
rx_data = &rx_ring->rx_data[i];
sys/dev/rtwn/pci/rtwn_pci_attach.c
168
error = bus_dmamap_create(rx_ring->data_dmat, 0, &rx_data->map);
sys/dev/rtwn/pci/rtwn_pci_attach.c
184
error = bus_dmamap_load(rx_ring->data_dmat, rx_data->map,
sys/dev/rtwn/pci/rtwn_pci_attach.c
193
rtwn_pci_setup_rx_desc(pc, &rx_ring->desc[i], rx_data->paddr,
sys/dev/rtwn/pci/rtwn_pci_attach.c
196
rx_ring->cur = 0;
sys/dev/rtwn/pci/rtwn_pci_attach.c
209
struct rtwn_rx_ring *rx_ring = &pc->rx_ring;
sys/dev/rtwn/pci/rtwn_pci_attach.c
214
rx_data = &rx_ring->rx_data[i];
sys/dev/rtwn/pci/rtwn_pci_attach.c
215
rtwn_pci_setup_rx_desc(pc, &rx_ring->desc[i],
sys/dev/rtwn/pci/rtwn_pci_attach.c
218
rx_ring->cur = 0;
sys/dev/rtwn/pci/rtwn_pci_attach.c
225
struct rtwn_rx_ring *rx_ring = &pc->rx_ring;
sys/dev/rtwn/pci/rtwn_pci_attach.c
229
if (rx_ring->desc_dmat != NULL) {
sys/dev/rtwn/pci/rtwn_pci_attach.c
230
if (rx_ring->desc != NULL) {
sys/dev/rtwn/pci/rtwn_pci_attach.c
231
bus_dmamap_sync(rx_ring->desc_dmat,
sys/dev/rtwn/pci/rtwn_pci_attach.c
232
rx_ring->desc_map,
sys/dev/rtwn/pci/rtwn_pci_attach.c
234
bus_dmamap_unload(rx_ring->desc_dmat,
sys/dev/rtwn/pci/rtwn_pci_attach.c
235
rx_ring->desc_map);
sys/dev/rtwn/pci/rtwn_pci_attach.c
236
bus_dmamem_free(rx_ring->desc_dmat, rx_ring->desc,
sys/dev/rtwn/pci/rtwn_pci_attach.c
237
rx_ring->desc_map);
sys/dev/rtwn/pci/rtwn_pci_attach.c
238
rx_ring->desc = NULL;
sys/dev/rtwn/pci/rtwn_pci_attach.c
240
bus_dma_tag_destroy(rx_ring->desc_dmat);
sys/dev/rtwn/pci/rtwn_pci_attach.c
241
rx_ring->desc_dmat = NULL;
sys/dev/rtwn/pci/rtwn_pci_attach.c
245
rx_data = &rx_ring->rx_data[i];
sys/dev/rtwn/pci/rtwn_pci_attach.c
248
bus_dmamap_sync(rx_ring->data_dmat,
sys/dev/rtwn/pci/rtwn_pci_attach.c
250
bus_dmamap_unload(rx_ring->data_dmat, rx_data->map);
sys/dev/rtwn/pci/rtwn_pci_attach.c
254
bus_dmamap_destroy(rx_ring->data_dmat, rx_data->map);
sys/dev/rtwn/pci/rtwn_pci_attach.c
257
if (rx_ring->data_dmat != NULL) {
sys/dev/rtwn/pci/rtwn_pci_attach.c
258
bus_dma_tag_destroy(rx_ring->data_dmat);
sys/dev/rtwn/pci/rtwn_pci_attach.c
259
rx_ring->data_dmat = NULL;
sys/dev/rtwn/pci/rtwn_pci_attach.c
527
(uintmax_t)pc->rx_ring.paddr);
sys/dev/rtwn/pci/rtwn_pci_attach.c
549
rtwn_pci_write_4(sc, R92C_RX_DESA, pc->rx_ring.paddr);
sys/dev/rtwn/pci/rtwn_pci_rx.c
182
struct rtwn_rx_ring *ring = &pc->rx_ring;
sys/dev/rtwn/pci/rtwn_pci_rx.c
348
struct rtwn_rx_ring *ring = &pc->rx_ring;
sys/dev/rtwn/pci/rtwn_pci_rx.c
86
struct rtwn_rx_ring *ring = &pc->rx_ring;
sys/dev/rtwn/pci/rtwn_pci_var.h
114
struct rtwn_rx_ring rx_ring;
sys/dev/sfxge/sfxge.c
87
#define SFXGE_PARAM_RX_RING SFXGE_PARAM(rx_ring)
sys/dev/sfxge/sfxge.c
90
SYSCTL_INT(_hw_sfxge, OID_AUTO, rx_ring, CTLFLAG_RDTUN,
sys/dev/thunderbolt/nhi.c
1091
rxd = &r->rx_ring[r->rx_pi];
sys/dev/thunderbolt/nhi.c
426
r->rx_ring = (union nhi_ring_desc *)(ring);
sys/dev/thunderbolt/nhi.c
431
r->rx_ring, r->rx_ring_busaddr);
sys/dev/thunderbolt/nhi.c
669
desc = &rp->rx_ring[ci].rx;
sys/dev/thunderbolt/nhi_var.h
82
union nhi_ring_desc *rx_ring;
sys/dev/xen/netback/netback.c
2318
rxb = &xnb->ring_configs[XNB_RING_TYPE_RX].back_ring.rx_ring;
sys/dev/xen/netback/netback.c
318
netif_rx_back_ring_t rx_ring;
sys/dev/xen/netback/netback.c
543
&xnb->ring_configs[XNB_RING_TYPE_RX].back_ring.rx_ring;
sys/dev/xen/netback/netback.c
733
BACK_RING_INIT(&ring->back_ring.rx_ring,
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib.h
263
struct ipoib_cm_rx_buf *rx_ring;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib.h
366
struct ipoib_rx_buf *rx_ring;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
130
rx_req = (struct ipoib_rx_buf *)&rx->rx_ring[id];
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
143
m_freem(rx->rx_ring[id].mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
144
rx->rx_ring[id].mb = NULL;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
158
struct ipoib_cm_rx_buf *rx_ring)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
163
if (rx_ring[i].mb) {
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
164
ipoib_cm_dma_unmap_rx(priv, &rx_ring[i]);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
165
m_freem(rx_ring[i].mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
168
kfree(rx_ring);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
310
rx->rx_ring = kzalloc(ipoib_recvq_size * sizeof *rx->rx_ring, GFP_KERNEL);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
311
if (!rx->rx_ring) {
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
317
memset(rx->rx_ring, 0, ipoib_recvq_size * sizeof *rx->rx_ring);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
340
if (!ipoib_cm_alloc_rx_mb(priv, &rx->rx_ring[i])) {
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
367
ipoib_cm_free_rx_ring(priv, rx->rx_ring);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
479
struct ipoib_cm_rx_buf *rx_ring;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
507
rx_ring = has_srq ? priv->cm.srq_ring : p->rx_ring;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
509
mb = rx_ring[wr_id].mb;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
539
memcpy(&saverx, &rx_ring[wr_id], sizeof(saverx));
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
540
newmb = ipoib_cm_alloc_rx_mb(priv, &rx_ring[wr_id]);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
548
memcpy(&rx_ring[wr_id], &saverx, sizeof(saverx));
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
777
ipoib_cm_free_rx_ring(priv, rx->rx_ring);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
155
rx_req = &priv->rx_ring[id];
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
166
ipoib_dma_unmap_rx(priv, &priv->rx_ring[id]);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
167
m_freem(priv->rx_ring[id].mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
168
priv->rx_ring[id].mb = NULL;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
177
return ipoib_alloc_map_mb(priv, &priv->rx_ring[id],
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
217
mb = priv->rx_ring[wr_id].mb;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
227
ipoib_dma_unmap_rx(priv, &priv->rx_ring[wr_id]);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
229
priv->rx_ring[wr_id].mb = NULL;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
241
memcpy(&saverx, &priv->rx_ring[wr_id], sizeof(saverx));
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
247
memcpy(&priv->rx_ring[wr_id], &saverx, sizeof(saverx));
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
711
if (priv->rx_ring[i].mb)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
818
rx_req = &priv->rx_ring[i];
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
821
ipoib_dma_unmap_rx(priv, &priv->rx_ring[i]);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
794
priv->rx_ring = kzalloc(ipoib_recvq_size * sizeof *priv->rx_ring,
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
796
if (!priv->rx_ring) {
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
821
kfree(priv->rx_ring);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
868
kfree(priv->rx_ring);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
871
priv->rx_ring = NULL;
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
148
#define rx_ring_posted(ssk) ring_posted(ssk->rx_ring)
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
359
struct sdp_rx_ring rx_ring;
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
423
static inline void rx_ring_unlock(struct sdp_rx_ring *rx_ring)
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
425
rw_runlock(&rx_ring->destroyed_lock);
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
428
static inline int rx_ring_trylock(struct sdp_rx_ring *rx_ring)
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
430
rw_rlock(&rx_ring->destroyed_lock);
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
431
if (rx_ring->destroyed) {
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
432
rx_ring_unlock(rx_ring);
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
438
static inline void rx_ring_destroy_lock(struct sdp_rx_ring *rx_ring)
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
440
rw_wlock(&rx_ring->destroyed_lock);
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
441
rx_ring->destroyed = 1;
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
442
rw_wunlock(&rx_ring->destroyed_lock);
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
450
ib_req_notify_cq(ssk->rx_ring.cq, IB_CQ_NEXT_COMP);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
187
ring_tail(ssk->rx_ring) >= ssk->recv_request_head &&
sys/ofed/drivers/infiniband/ulp/sdp/sdp_cma.c
117
qp_init_attr.recv_cq = ssk->rx_ring.cq;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
1471
if (!rx_ring_trylock(&ssk->rx_ring)) {
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
1495
rx_ring_unlock(&ssk->rx_ring);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
1685
ib_modify_cq(ssk->rx_ring.cq, sdp_mod_count, sdp_mod_usec);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
185
rw_destroy(&ssk->rx_ring.destroyed_lock);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
109
int id = ring_head(ssk->rx_ring);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
125
rx_req = ssk->rx_ring.buffer + (id & (SDP_RX_SIZE - 1));
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
156
atomic_inc(&ssk->rx_ring.head);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
286
ssk->recv_request_head = ring_head(ssk->rx_ring) + 1;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
288
ssk->recv_request_head = ring_tail(ssk->rx_ring);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
308
if (unlikely(id != ring_tail(ssk->rx_ring))) {
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
310
id, ring_tail(ssk->rx_ring));
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
315
rx_req = &ssk->rx_ring.buffer[id & (SDP_RX_SIZE - 1)];
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
319
atomic_inc(&ssk->rx_ring.tail);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
534
struct ib_cq *cq = ssk->rx_ring.cq;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
574
if (unlikely(!ssk->rx_ring.cq)) {
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
610
if (!ssk->rx_ring.cq || !ssk->tx_ring.cq)
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
638
if (!rx_ring_trylock(&ssk->rx_ring)) {
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
655
rx_ring_unlock(&ssk->rx_ring);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
666
KASSERT(cq == ssk->rx_ring.cq,
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
681
mb = sdp_recv_completion(ssk, ring_tail(ssk->rx_ring));
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
691
ssk->rx_ring.buffer = NULL;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
692
ssk->rx_ring.destroyed = 0;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
693
rw_init(&ssk->rx_ring.destroyed_lock, "sdp rx lock");
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
714
atomic_set(&ssk->rx_ring.head, 1);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
715
atomic_set(&ssk->rx_ring.tail, 1);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
717
ssk->rx_ring.buffer = malloc(sizeof(*ssk->rx_ring.buffer) * SDP_RX_SIZE,
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
728
sdp_sk(ssk->socket)->rx_ring.cq = rx_cq;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
734
free(ssk->rx_ring.buffer, M_SDP);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
735
ssk->rx_ring.buffer = NULL;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
744
rx_ring_destroy_lock(&ssk->rx_ring);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
746
if (ssk->rx_ring.buffer) {
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
748
free(ssk->rx_ring.buffer, M_SDP);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
749
ssk->rx_ring.buffer = NULL;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
752
if (ssk->rx_ring.cq) {
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
753
ib_destroy_cq(ssk->rx_ring.cq);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
754
ssk->rx_ring.cq = NULL;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
757
WARN_ON(ring_head(ssk->rx_ring) != ring_tail(ssk->rx_ring));
usr.sbin/virtual_oss/virtual_oss/httpd.c
726
vring_get_read(&pvc->rx_ring[1], &ptr, &len);
usr.sbin/virtual_oss/virtual_oss/httpd.c
729
vring_reset(&pvc->rx_ring[1]);
usr.sbin/virtual_oss/virtual_oss/httpd.c
774
vring_inc_read(&pvc->rx_ring[1], len);
usr.sbin/virtual_oss/virtual_oss/int.h
182
struct virtual_ring rx_ring[2];
usr.sbin/virtual_oss/virtual_oss/main.c
117
return (vring_total_read_len(&pvc->rx_ring[1]));
usr.sbin/virtual_oss/virtual_oss/main.c
1371
vring_reset(&pvc->rx_ring[1]);
usr.sbin/virtual_oss/virtual_oss/main.c
272
vring_free(&pvc->rx_ring[0]);
usr.sbin/virtual_oss/virtual_oss/main.c
273
vring_free(&pvc->rx_ring[1]);
usr.sbin/virtual_oss/virtual_oss/main.c
357
vring_free(&pvc->rx_ring[0]);
usr.sbin/virtual_oss/virtual_oss/main.c
358
vring_free(&pvc->rx_ring[1]);
usr.sbin/virtual_oss/virtual_oss/main.c
426
if (vring_alloc(&pvc->rx_ring[0], bufsize_internal))
usr.sbin/virtual_oss/virtual_oss/main.c
428
if (vring_alloc(&pvc->rx_ring[1], bufsize))
usr.sbin/virtual_oss/virtual_oss/main.c
451
vring_free(&pvc->rx_ring[1]);
usr.sbin/virtual_oss/virtual_oss/main.c
453
vring_free(&pvc->rx_ring[0]);
usr.sbin/virtual_oss/virtual_oss/main.c
551
size = vring_write_zero(&pvc->rx_ring[0], size);
usr.sbin/virtual_oss/virtual_oss/main.c
568
vring_get_write(&pvc->rx_ring[1], &ptr, &len);
usr.sbin/virtual_oss/virtual_oss/main.c
580
vring_inc_write(&pvc->rx_ring[1], len);
usr.sbin/virtual_oss/virtual_oss/main.c
703
vring_get_read(&pvc->rx_ring[0], &src_ptr, &src_len);
usr.sbin/virtual_oss/virtual_oss/main.c
704
vring_get_write(&pvc->rx_ring[1], &dst_ptr, &dst_len);
usr.sbin/virtual_oss/virtual_oss/main.c
725
vring_inc_read(&pvc->rx_ring[0], src_len);
usr.sbin/virtual_oss/virtual_oss/main.c
726
vring_inc_write(&pvc->rx_ring[1], dst_len);
usr.sbin/virtual_oss/virtual_oss/main.c
743
vring_get_read(&pvc->rx_ring[0], &src_ptr, &src_len);
usr.sbin/virtual_oss/virtual_oss/main.c
744
vring_get_write(&pvc->rx_ring[1], &dst_ptr, &dst_len);
usr.sbin/virtual_oss/virtual_oss/main.c
793
vring_inc_read(&pvc->rx_ring[0], src_len);
usr.sbin/virtual_oss/virtual_oss/main.c
794
vring_inc_write(&pvc->rx_ring[1], dst_len);
usr.sbin/virtual_oss/virtual_oss/main.c
839
vring_get_read(&pvc->rx_ring[1], &buf_ptr, &buf_len);
usr.sbin/virtual_oss/virtual_oss/main.c
875
vring_inc_read(&pvc->rx_ring[1], buf_len);
usr.sbin/virtual_oss/virtual_oss/virtual_oss.c
499
vclient_write_linear(pvc, &pvc->rx_ring[0],
usr.sbin/virtual_oss/virtual_oss/virtual_oss.c
827
vclient_write_linear(pvc, &pvc->rx_ring[0],