sys/dev/age/if_age.c
2997
rxd->rx_desc = &rd->age_rx_ring[i];
sys/dev/age/if_age.c
3056
struct rx_desc *desc;
sys/dev/age/if_age.c
3091
desc = rxd->rx_desc;
sys/dev/age/if_agevar.h
117
struct rx_desc *age_rx_ring;
sys/dev/age/if_agevar.h
130
(sizeof(struct rx_desc) * AGE_RX_RING_CNT)
sys/dev/age/if_agevar.h
80
struct rx_desc *rx_desc;
sys/dev/alc/if_alc.c
3510
rxd->rx_desc->addr = htole64(segs[0].ds_addr);
sys/dev/alc/if_alc.c
4490
rxd->rx_desc = &rd->alc_rx_ring[i];
sys/dev/alc/if_alcvar.h
36
#define ALC_RX_RING_ALIGN sizeof(struct rx_desc)
sys/dev/alc/if_alcvar.h
65
(sizeof(struct rx_desc) * ALC_RX_RING_CNT)
sys/dev/alc/if_alcvar.h
87
struct rx_desc *rx_desc;
sys/dev/alc/if_alcvar.h
98
struct rx_desc *alc_rx_ring;
sys/dev/aq/aq_ring.c
240
aq_rx_desc_t *rx_desc;
sys/dev/aq/aq_ring.c
251
rx_desc = (aq_rx_desc_t *) &ring->rx_descs[pidx];
sys/dev/aq/aq_ring.c
252
rx_desc->read.buf_addr = htole64(iru->iru_paddrs[i]);
sys/dev/aq/aq_ring.c
253
rx_desc->read.hdr_addr = 0;
sys/dev/aq/aq_ring.c
277
aq_rx_desc_t *rx_desc = (aq_rx_desc_t *) ring->rx_descs;
sys/dev/aq/aq_ring.c
285
(volatile uint64_t*)&rx_desc[i]);
sys/dev/aq/aq_ring.c
286
if (!rx_desc[i].wb.dd)
sys/dev/aq/aq_ring.c
289
if (rx_desc[i].wb.eop) {
sys/dev/aq/aq_ring.c
296
if (rx_desc[i].wb.rsc_cnt) {
sys/dev/aq/aq_ring.c
297
i = rx_desc[i].wb.next_desp;
sys/dev/aq/aq_ring.c
313
aq_rx_set_cso_flags(aq_rx_desc_t *rx_desc, if_rxd_info_t ri)
sys/dev/aq/aq_ring.c
315
if ((rx_desc->wb.pkt_type & 0x3) == 0) { // IPv4
sys/dev/aq/aq_ring.c
316
if (rx_desc->wb.rx_cntl & BIT(0)) { // IPv4 csum checked
sys/dev/aq/aq_ring.c
318
if (!(rx_desc->wb.rx_stat & BIT(1)))
sys/dev/aq/aq_ring.c
322
if (rx_desc->wb.rx_cntl & BIT(1)) { // TCP/UDP csum checked
sys/dev/aq/aq_ring.c
324
if (!(rx_desc->wb.rx_stat & BIT(2)) && // L4 csum error
sys/dev/aq/aq_ring.c
325
(rx_desc->wb.rx_stat & BIT(3))) { // L4 csum valid
sys/dev/aq/aq_ring.c
348
aq_rx_desc_t *rx_desc;
sys/dev/aq/aq_ring.c
359
rx_desc = (aq_rx_desc_t *) &ring->rx_descs[cidx];
sys/dev/aq/aq_ring.c
362
(volatile uint64_t *)rx_desc);
sys/dev/aq/aq_ring.c
364
if ((rx_desc->wb.rx_stat & BIT(0)) != 0) {
sys/dev/aq/aq_ring.c
370
if (!rx_desc->wb.eop) {
sys/dev/aq/aq_ring.c
373
total_len = le32toh(rx_desc->wb.pkt_len);
sys/dev/aq/aq_ring.c
380
if ((rx_desc->wb.pkt_type & 0x60) != 0) {
sys/dev/aq/aq_ring.c
382
ri->iri_vtag = le32toh(rx_desc->wb.vlan);
sys/dev/aq/aq_ring.c
387
} while (!rx_desc->wb.eop);
sys/dev/aq/aq_ring.c
390
aq_rx_set_cso_flags(rx_desc, ri);
sys/dev/aq/aq_ring.c
392
ri->iri_rsstype = bsd_rss_type[rx_desc->wb.rss_type & 0xF];
sys/dev/aq/aq_ring.c
394
ri->iri_flowid = le32toh(rx_desc->wb.rss_hash);
sys/dev/cadence/if_cgem.c
378
struct cgem_rx_desc *rx_desc;
sys/dev/cadence/if_cgem.c
392
rx_desc = sc->null_qs;
sys/dev/cadence/if_cgem.c
393
rx_desc->addr = CGEM_RXDESC_OWN | CGEM_RXDESC_WRAP;
sys/dev/cadence/if_cgem.c
394
tx_desc = (struct cgem_tx_desc *)(rx_desc + 1);
sys/dev/cxgb/cxgb_adapter.h
199
struct rx_desc;
sys/dev/cxgb/cxgb_adapter.h
217
struct rx_desc *desc;
sys/dev/cxgb/cxgb_sge.c
2437
if ((ret = alloc_ring(sc, p->fl_size, sizeof(struct rx_desc),
sys/dev/cxgb/cxgb_sge.c
2446
if ((ret = alloc_ring(sc, p->jumbo_size, sizeof(struct rx_desc),
sys/dev/cxgb/cxgb_sge.c
3715
memcpy(data, &qs->fl[qnum].desc[idx], sizeof(struct rx_desc));
sys/dev/cxgb/cxgb_sge.c
3716
return sizeof(struct rx_desc);
sys/dev/cxgb/cxgb_sge.c
706
struct rx_desc *d = &q->desc[q->pidx];
sys/dev/cxgb/cxgb_sge.c
843
struct rx_desc *from = &q->desc[idx];
sys/dev/cxgb/cxgb_sge.c
844
struct rx_desc *to = &q->desc[q->pidx];
sys/dev/iser/iser_initiator.c
202
struct iser_rx_desc *rx_desc;
sys/dev/iser/iser_initiator.c
221
rx_desc = iser_conn->rx_descs;
sys/dev/iser/iser_initiator.c
223
for (i = 0; i < iser_conn->qp_max_recv_dtos; i++, rx_desc++) {
sys/dev/iser/iser_initiator.c
224
dma_addr = ib_dma_map_single(device->ib_device, (void *)rx_desc,
sys/dev/iser/iser_initiator.c
229
rx_desc->dma_addr = dma_addr;
sys/dev/iser/iser_initiator.c
231
rx_sg = &rx_desc->rx_sg;
sys/dev/iser/iser_initiator.c
232
rx_sg->addr = rx_desc->dma_addr;
sys/dev/iser/iser_initiator.c
242
rx_desc = iser_conn->rx_descs;
sys/dev/iser/iser_initiator.c
243
for (j = 0; j < i; j++, rx_desc++)
sys/dev/iser/iser_initiator.c
244
ib_dma_unmap_single(device->ib_device, rx_desc->dma_addr,
sys/dev/iser/iser_initiator.c
260
struct iser_rx_desc *rx_desc;
sys/dev/iser/iser_initiator.c
266
rx_desc = iser_conn->rx_descs;
sys/dev/iser/iser_initiator.c
267
for (i = 0; i < iser_conn->qp_max_recv_dtos; i++, rx_desc++)
sys/dev/iser/iser_initiator.c
268
ib_dma_unmap_single(device->ib_device, rx_desc->dma_addr,
sys/dev/iser/iser_initiator.c
465
iser_rcv_completion(struct iser_rx_desc *rx_desc,
sys/dev/iser/iser_initiator.c
479
if ((char *)rx_desc == iser_conn->login_resp_buf) {
sys/dev/iser/iser_initiator.c
483
rx_dma = rx_desc->dma_addr;
sys/dev/iser/iser_initiator.c
490
hdr = &rx_desc->iscsi_header;
sys/dev/iser/iser_initiator.c
501
response->ip_data_mbuf = (struct mbuf *)(rx_desc->data);
sys/dev/iser/iser_verbs.c
119
struct iser_rx_desc *rx_desc;
sys/dev/iser/iser_verbs.c
124
rx_desc = (struct iser_rx_desc *)(uintptr_t)wc->wr_id;
sys/dev/iser/iser_verbs.c
125
iser_rcv_completion(rx_desc, wc->byte_len,
sys/dev/iser/iser_verbs.c
884
struct iser_rx_desc *rx_desc;
sys/dev/iser/iser_verbs.c
887
rx_desc = &iser_conn->rx_descs[my_rx_head];
sys/dev/iser/iser_verbs.c
888
rx_wr->wr_id = (uintptr_t)rx_desc;
sys/dev/iser/iser_verbs.c
889
rx_wr->sg_list = &rx_desc->rx_sg;
sys/dev/jme/if_jme.c
3125
rxd->rx_desc = &rd->jme_rx_ring[i];
sys/dev/jme/if_jme.c
3176
desc = rxd->rx_desc;
sys/dev/jme/if_jmevar.h
109
struct jme_desc *rx_desc;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
110
mlx4_en_alloc_buf(struct mlx4_en_rx_ring *ring, struct mlx4_en_rx_desc *rx_desc,
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
175
rx_desc->data[0].addr = cpu_to_be64(segs[0].ds_addr);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
178
rx_desc->data[i].byte_count = cpu_to_be32(segs[i].ds_len);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
179
rx_desc->data[i].lkey = ring->rx_mr_key_be;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
180
rx_desc->data[i].addr = cpu_to_be64(segs[i].ds_addr);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
183
rx_desc->data[i].byte_count = 0;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
184
rx_desc->data[i].lkey = cpu_to_be32(MLX4_EN_MEMTYPE_PAD);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
185
rx_desc->data[i].addr = 0;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
205
rx_desc->data[0].addr = cpu_to_be64(ring->spare.segs[0].ds_addr);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
209
rx_desc->data[i].byte_count = cpu_to_be32(ring->spare.segs[i].ds_len);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
210
rx_desc->data[i].lkey = ring->rx_mr_key_be;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
211
rx_desc->data[i].addr = cpu_to_be64(ring->spare.segs[i].ds_addr);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
213
rx_desc->data[i].byte_count = 0;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
214
rx_desc->data[i].lkey = cpu_to_be32(MLX4_EN_MEMTYPE_PAD);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
215
rx_desc->data[i].addr = 0;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
236
struct mlx4_en_rx_desc *rx_desc =
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
242
if (mlx4_en_alloc_buf(ring, rx_desc, mb_list)) {
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
52
struct mlx4_en_rx_desc *rx_desc =
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
57
rx_desc->data[0].byte_count = cpu_to_be32(priv->rx_mb_size - MLX4_NET_IP_ALIGN);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
58
rx_desc->data[0].lkey = cpu_to_be32(priv->mdev->mr.key);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
624
struct mlx4_en_rx_desc *rx_desc, struct mlx4_en_rx_mbuf *mb_list,
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
654
if (mlx4_en_alloc_buf(ring, rx_desc, mb_list))
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
66
rx_desc->data[i].byte_count = 0;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
67
rx_desc->data[i].lkey = cpu_to_be32(MLX4_EN_MEMTYPE_PAD);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
68
rx_desc->data[i].addr = 0;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
744
struct mlx4_en_rx_desc *rx_desc;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
770
rx_desc = ((struct mlx4_en_rx_desc *)ring->buf) + index;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
786
mb = mlx4_en_rx_mb(priv, ring, rx_desc, mb_list, length);
sys/dev/rtwn/pci/rtwn_pci_rx.c
100
le32toh(rx_desc->rxdw2), le32toh(rx_desc->rxdw3),
sys/dev/rtwn/pci/rtwn_pci_rx.c
101
le32toh(rx_desc->rxdw4), le32toh(rx_desc->tsf_low),
sys/dev/rtwn/pci/rtwn_pci_rx.c
102
le32toh(rx_desc->rxbufaddr), le32toh(rx_desc->rxbufaddr64));
sys/dev/rtwn/pci/rtwn_pci_rx.c
104
rxdw0 = le32toh(rx_desc->rxdw0);
sys/dev/rtwn/pci/rtwn_pci_rx.c
156
ni = rtwn_rx_common(sc, m, rx_desc);
sys/dev/rtwn/pci/rtwn_pci_rx.c
183
struct rtwn_rx_stat_pci *rx_desc = &ring->desc[ring->cur];
sys/dev/rtwn/pci/rtwn_pci_rx.c
196
memcpy(pc->pc_rx_buf, rx_desc, desc_size);
sys/dev/rtwn/pci/rtwn_pci_rx.c
198
rxdw0 = le32toh(rx_desc->rxdw0);
sys/dev/rtwn/pci/rtwn_pci_rx.c
349
struct rtwn_rx_stat_pci *rx_desc;
sys/dev/rtwn/pci/rtwn_pci_rx.c
356
rx_desc = &ring->desc[ring->cur];
sys/dev/rtwn/pci/rtwn_pci_rx.c
359
if (le32toh(rx_desc->rxdw0) & RTWN_RXDW0_OWN)
sys/dev/rtwn/pci/rtwn_pci_rx.c
384
rtwn_pci_setup_rx_desc(pc, rx_desc, rx_data->paddr,
sys/dev/rtwn/pci/rtwn_pci_rx.c
394
if (le32toh(rx_desc->rxdw0) & RTWN_RXDW0_OWN)
sys/dev/rtwn/pci/rtwn_pci_rx.c
87
struct rtwn_rx_stat_pci *rx_desc = &ring->desc[ring->cur];
sys/dev/rtwn/pci/rtwn_pci_rx.c
99
__func__, le32toh(rx_desc->rxdw0), le32toh(rx_desc->rxdw1),
sys/dev/sfxge/sfxge_ev.c
110
struct sfxge_rx_sw_desc *rx_desc;
sys/dev/sfxge/sfxge_ev.c
143
rx_desc = &rxq->queue[id];
sys/dev/sfxge/sfxge_ev.c
145
prefetch_read_many(rx_desc->mbuf);
sys/dev/sfxge/sfxge_ev.c
148
rx_desc = &rxq->queue[id];
sys/dev/sfxge/sfxge_ev.c
149
KASSERT(rx_desc->flags == EFX_DISCARD,
sys/dev/sfxge/sfxge_ev.c
151
rx_desc->flags = flags;
sys/dev/sfxge/sfxge_ev.c
154
rx_desc->size = (uint16_t)size;
sys/dev/sfxge/sfxge_rx.c
245
struct sfxge_rx_sw_desc *rx_desc;
sys/dev/sfxge/sfxge_rx.c
250
rx_desc = &rxq->queue[id];
sys/dev/sfxge/sfxge_rx.c
251
KASSERT(rx_desc->mbuf == NULL, ("rx_desc->mbuf != NULL"));
sys/dev/sfxge/sfxge_rx.c
253
rx_desc->flags = EFX_DISCARD;
sys/dev/sfxge/sfxge_rx.c
254
m = rx_desc->mbuf = m_getjcl(M_NOWAIT, MT_DATA, M_PKTHDR,
sys/dev/sfxge/sfxge_rx.c
321
sfxge_rx_deliver(struct sfxge_rxq *rxq, struct sfxge_rx_sw_desc *rx_desc)
sys/dev/sfxge/sfxge_rx.c
324
struct mbuf *m = rx_desc->mbuf;
sys/dev/sfxge/sfxge_rx.c
325
int flags = rx_desc->flags;
sys/dev/sfxge/sfxge_rx.c
348
m->m_len = rx_desc->size - sc->rx_prefix_size;
sys/dev/sfxge/sfxge_rx.c
351
__sfxge_rx_deliver(sc, rx_desc->mbuf);
sys/dev/sfxge/sfxge_rx.c
353
rx_desc->flags = EFX_DISCARD;
sys/dev/sfxge/sfxge_rx.c
354
rx_desc->mbuf = NULL;
sys/dev/sfxge/sfxge_rx.c
818
struct sfxge_rx_sw_desc *rx_desc;
sys/dev/sfxge/sfxge_rx.c
821
rx_desc = &rxq->queue[id];
sys/dev/sfxge/sfxge_rx.c
822
m = rx_desc->mbuf;
sys/dev/sfxge/sfxge_rx.c
827
if (rx_desc->flags & (EFX_ADDR_MISMATCH | EFX_DISCARD))
sys/dev/sfxge/sfxge_rx.c
831
if (rx_desc->flags & EFX_PKT_PREFIX_LEN) {
sys/dev/sfxge/sfxge_rx.c
839
rx_desc->size = (int)tmp_size + sc->rx_prefix_size;
sys/dev/sfxge/sfxge_rx.c
844
switch (rx_desc->flags & (EFX_PKT_IPV4 | EFX_PKT_IPV6)) {
sys/dev/sfxge/sfxge_rx.c
847
rx_desc->flags &=
sys/dev/sfxge/sfxge_rx.c
852
rx_desc->flags &= ~EFX_CKSUM_TCPUDP;
sys/dev/sfxge/sfxge_rx.c
886
prev = rx_desc;
sys/dev/sfxge/sfxge_rx.c
892
rx_desc->mbuf = NULL;
sys/dev/sis/if_sis.c
1389
rxd->rx_desc = &sc->sis_rx_list[i];
sys/dev/sis/if_sis.c
1394
rxd->rx_desc->sis_next = htole32(SIS_ADDR_LO(next));
sys/dev/sis/if_sis.c
1441
rxd->rx_desc->sis_ptr = htole32(SIS_ADDR_LO(segs[0].ds_addr));
sys/dev/sis/if_sis.c
1442
rxd->rx_desc->sis_cmdsts = htole32(SIS_RXLEN);
sys/dev/sis/if_sis.c
1450
rxd->rx_desc->sis_cmdsts = htole32(SIS_RXLEN);
sys/dev/sis/if_sisreg.h
451
struct sis_desc *rx_desc;
sys/dev/tsec/if_tsec.c
1277
struct tsec_desc *rx_desc;
sys/dev/tsec/if_tsec.c
1298
rx_desc = TSEC_GET_CUR_RX_DESC(sc);
sys/dev/tsec/if_tsec.c
1299
flags = rx_desc->flags;
sys/dev/tsec/if_tsec.c
1319
rx_desc->length = 0;
sys/dev/tsec/if_tsec.c
1320
rx_desc->flags = (rx_desc->flags &
sys/dev/tsec/if_tsec.c
1334
m->m_len = rx_desc->length;
sys/dev/tsec/if_tsec.c
1365
rx_desc->bufptr = rx_data[i].paddr;
sys/dev/tsec/if_tsec.c
1366
rx_desc->length = 0;
sys/dev/tsec/if_tsec.c
1367
rx_desc->flags = (rx_desc->flags & ~TSEC_RXBD_ZEROONINIT) |
sys/dev/tsec/if_tsec.c
370
struct tsec_desc *rx_desc = sc->tsec_rx_vaddr;
sys/dev/tsec/if_tsec.c
504
rx_desc[i].bufptr = sc->rx_data[i].paddr;
sys/dev/tsec/if_tsec.c
505
rx_desc[i].length = 0;
sys/dev/tsec/if_tsec.c
506
rx_desc[i].flags = TSEC_RXBD_E | TSEC_RXBD_I |
sys/dev/vge/if_vge.c
1206
rxd->rx_desc->vge_sts = 0;
sys/dev/vge/if_vge.c
1207
rxd->rx_desc->vge_ctl = 0;
sys/dev/vge/if_vge.c
1219
rxd->rx_desc->vge_sts = htole32(VGE_RDSTS_OWN);
sys/dev/vge/if_vge.c
1271
rxd->rx_desc->vge_sts = 0;
sys/dev/vge/if_vge.c
1272
rxd->rx_desc->vge_ctl = 0;
sys/dev/vge/if_vge.c
1273
rxd->rx_desc->vge_addrlo = htole32(VGE_ADDR_LO(segs[0].ds_addr));
sys/dev/vge/if_vge.c
1274
rxd->rx_desc->vge_addrhi = htole32(VGE_ADDR_HI(segs[0].ds_addr) |
sys/dev/vge/if_vge.c
1287
rxd->rx_desc->vge_sts = htole32(VGE_RDSTS_OWN);
sys/dev/vge/if_vge.c
1343
rxd->rx_desc = &rd->vge_rx_ring[i];
sys/dev/vge/if_vgevar.h
98
struct vge_rx_desc *rx_desc;
sys/dev/vte/if_vte.c
1454
rxd->rx_desc->drbp = htole32(segs[0].ds_addr);
sys/dev/vte/if_vte.c
1455
rxd->rx_desc->drlen = htole16(VTE_RX_LEN(segs[0].ds_len));
sys/dev/vte/if_vte.c
1456
rxd->rx_desc->drst = htole16(VTE_DRST_RX_OWN);
sys/dev/vte/if_vte.c
1499
status = le16toh(rxd->rx_desc->drst);
sys/dev/vte/if_vte.c
1502
total_len = VTE_RX_LEN(le16toh(rxd->rx_desc->drlen));
sys/dev/vte/if_vte.c
1506
rxd->rx_desc->drlen =
sys/dev/vte/if_vte.c
1508
rxd->rx_desc->drst = htole16(VTE_DRST_RX_OWN);
sys/dev/vte/if_vte.c
1513
rxd->rx_desc->drlen =
sys/dev/vte/if_vte.c
1515
rxd->rx_desc->drst = htole16(VTE_DRST_RX_OWN);
sys/dev/vte/if_vte.c
1941
rxd->rx_desc = desc;
sys/dev/vte/if_vtevar.h
60
struct vte_rx_desc *rx_desc;