Symbol: rx_ring
usr/src/grub/grub-0.97/netboot/epic100.c
250
outl(virt_to_le32desc(&rx_ring), prcdar);
usr/src/grub/grub-0.97/netboot/epic100.c
269
rx_ring[i].status = cpu_to_le32(RRING_OWN); /* Owned by Epic chip */
usr/src/grub/grub-0.97/netboot/epic100.c
270
rx_ring[i].buflength = cpu_to_le32(PKT_BUF_SZ);
usr/src/grub/grub-0.97/netboot/epic100.c
271
rx_ring[i].bufaddr = virt_to_bus(&rx_packet[i * PKT_BUF_SZ]);
usr/src/grub/grub-0.97/netboot/epic100.c
272
rx_ring[i].next = virt_to_le32desc(&rx_ring[i + 1]) ;
usr/src/grub/grub-0.97/netboot/epic100.c
275
rx_ring[i-1].next = virt_to_le32desc(&rx_ring[0]);
usr/src/grub/grub-0.97/netboot/epic100.c
375
if ((rx_ring[entry].status & cpu_to_le32(RRING_OWN)) == RRING_OWN)
usr/src/grub/grub-0.97/netboot/epic100.c
380
status = le32_to_cpu(rx_ring[entry].status);
usr/src/grub/grub-0.97/netboot/epic100.c
397
nic->packetlen = le32_to_cpu((rx_ring[entry].buflength))- 4;
usr/src/grub/grub-0.97/netboot/epic100.c
406
rx_ring[entry].status = RRING_OWN;
usr/src/grub/grub-0.97/netboot/epic100.c
85
static struct epic_rx_desc rx_ring[RX_RING_SIZE]
usr/src/grub/grub-0.97/netboot/forcedeth.c
302
static struct ring_desc rx_ring[RX_RING];
usr/src/grub/grub-0.97/netboot/forcedeth.c
325
struct ring_desc *rx_ring;
usr/src/grub/grub-0.97/netboot/forcedeth.c
530
rx_ring[i].PacketBuffer =
usr/src/grub/grub-0.97/netboot/forcedeth.c
532
rx_ring[i].Length = cpu_to_le16(RX_NIC_BUFSIZE);
usr/src/grub/grub-0.97/netboot/forcedeth.c
534
rx_ring[i].Flags = cpu_to_le16(NV_RX_AVAIL);
usr/src/grub/grub-0.97/netboot/forcedeth.c
595
rx_ring[i].Flags = 0;
usr/src/grub/grub-0.97/netboot/forcedeth.c
742
writel((u32) virt_to_le32desc(&rx_ring[0]),
usr/src/grub/grub-0.97/netboot/forcedeth.c
811
prd = &rx_ring[i];
usr/src/grub/grub-0.97/netboot/ns83820.c
331
writel(virt_to_le32desc(&rx_ring[ns->cur_rx]), ns->base + RXDP); \
usr/src/grub/grub-0.97/netboot/ns83820.c
383
static struct ring_desc rx_ring[NR_RX_DESC]
usr/src/grub/grub-0.97/netboot/ns83820.c
514
rx_ring[i].link = virt_to_le32desc(&rx_ring[i + 1]);
usr/src/grub/grub-0.97/netboot/ns83820.c
515
rx_ring[i].bufptr =
usr/src/grub/grub-0.97/netboot/ns83820.c
517
rx_ring[i].cmdsts = cpu_to_le32(REAL_RX_BUF_SIZE);
usr/src/grub/grub-0.97/netboot/ns83820.c
518
rx_ring[i].extsts = cpu_to_le32(0);
usr/src/grub/grub-0.97/netboot/ns83820.c
523
writel(virt_to_le32desc(&rx_ring[0]), ns->base + RXDP);
usr/src/grub/grub-0.97/netboot/ns83820.c
664
if (rx_ring[i].cmdsts == CMDSTS_OWN) {
usr/src/grub/grub-0.97/netboot/ns83820.c
666
rx_ring[i].cmdsts = cpu_to_le32(REAL_RX_BUF_SIZE);
usr/src/grub/grub-0.97/netboot/ns83820.c
683
cmdsts = le32_to_cpu(rx_ring[entry].cmdsts);
usr/src/grub/grub-0.97/netboot/ns83820.c
698
rx_ring[entry].cmdsts = cpu_to_le32(CMDSTS_OWN);
usr/src/grub/grub-0.97/netboot/pcnet32.c
222
u32 rx_ring;
usr/src/grub/grub-0.97/netboot/pcnet32.c
242
static struct pcnet32_rx_head rx_ring[RX_RING_SIZE]
usr/src/grub/grub-0.97/netboot/pcnet32.c
412
rx_ring[i].base = (u32) virt_to_le32desc(&rxb[i]);
usr/src/grub/grub-0.97/netboot/pcnet32.c
413
rx_ring[i].buf_length = le16_to_cpu(-PKT_BUF_SZ);
usr/src/grub/grub-0.97/netboot/pcnet32.c
414
rx_ring[i].status = le16_to_cpu(0x8000);
usr/src/grub/grub-0.97/netboot/pcnet32.c
429
lp->init_block.rx_ring = (u32) virt_to_le32desc(&rx_ring[0]);
usr/src/grub/grub-0.97/netboot/pcnet32.c
552
status = ((short) le16_to_cpu(rx_ring[entry].status) >> 8);
usr/src/grub/grub-0.97/netboot/pcnet32.c
561
(le32_to_cpu(rx_ring[entry].msg_length) & 0xfff) - 4;
usr/src/grub/grub-0.97/netboot/pcnet32.c
566
rx_ring[entry].buf_length = le16_to_cpu(-PKT_BUF_SZ);
usr/src/grub/grub-0.97/netboot/pcnet32.c
567
rx_ring[entry].status |= le16_to_cpu(0x8000); /* prime for next receive */
usr/src/grub/grub-0.97/netboot/pcnet32.c
884
lp->init_block.rx_ring = virt_to_bus(&rx_ring);
usr/src/grub/grub-0.97/netboot/r8169.c
299
static u8 rx_ring[NUM_RX_DESC * sizeof(struct TxDesc) + 256];
usr/src/grub/grub-0.97/netboot/r8169.c
656
tpc->RxDescArrays = rx_ring;
usr/src/grub/grub-0.97/netboot/rtl8139.c
173
static unsigned char rx_ring[RX_BUF_LEN+16] __attribute__((aligned(4)));
usr/src/grub/grub-0.97/netboot/rtl8139.c
343
printf("rx ring address is %X\n",(unsigned long)rx_ring);
usr/src/grub/grub-0.97/netboot/rtl8139.c
345
outl((unsigned long)virt_to_bus(rx_ring), nic->ioaddr + RxBuf);
usr/src/grub/grub-0.97/netboot/rtl8139.c
445
rx_status = *(unsigned int*)(rx_ring + ring_offs);
usr/src/grub/grub-0.97/netboot/rtl8139.c
461
memcpy(nic->packet, rx_ring + ring_offs + 4, semi_count);
usr/src/grub/grub-0.97/netboot/rtl8139.c
462
memcpy(nic->packet+semi_count, rx_ring, rx_size-4-semi_count);
usr/src/grub/grub-0.97/netboot/rtl8139.c
467
memcpy(nic->packet, rx_ring + ring_offs + 4, nic->packetlen);
usr/src/grub/grub-0.97/netboot/rtl8139.c
474
(unsigned long)(rx_ring+ring_offs+4),
usr/src/grub/grub-0.97/netboot/sundance.c
261
static struct netdev_desc rx_ring[RX_RING_SIZE];
usr/src/grub/grub-0.97/netboot/sundance.c
354
rx_ring[i].next_desc = virt_to_le32desc(&rx_ring[i + 1]);
usr/src/grub/grub-0.97/netboot/sundance.c
355
rx_ring[i].status = 0;
usr/src/grub/grub-0.97/netboot/sundance.c
356
rx_ring[i].length = 0;
usr/src/grub/grub-0.97/netboot/sundance.c
357
rx_ring[i].addr = 0;
usr/src/grub/grub-0.97/netboot/sundance.c
361
rx_ring[i - 1].next_desc = virt_to_le32desc(&rx_ring[0]);
usr/src/grub/grub-0.97/netboot/sundance.c
364
rx_ring[i].addr = virt_to_le32desc(&rxb[i * PKT_BUF_SZ]);
usr/src/grub/grub-0.97/netboot/sundance.c
365
rx_ring[i].length = cpu_to_le32(PKT_BUF_SZ | LastFrag);
usr/src/grub/grub-0.97/netboot/sundance.c
394
outl(virt_to_le32desc(&rx_ring[0]), BASE + RxListPtr);
usr/src/grub/grub-0.97/netboot/sundance.c
474
u32 frame_status = le32_to_cpu(rx_ring[entry].status);
usr/src/grub/grub-0.97/netboot/sundance.c
503
rx_ring[entry].length = cpu_to_le32(PKT_BUF_SZ | LastFrag);
usr/src/grub/grub-0.97/netboot/sundance.c
504
rx_ring[entry].status = 0;
usr/src/grub/grub-0.97/netboot/tlan.c
201
struct TLanList rx_ring[TLAN_NUM_RX_LISTS];
usr/src/grub/grub-0.97/netboot/tlan.c
295
priv->rx_head_desc = &rx_ring[0];
usr/src/grub/grub-0.97/netboot/tlan.c
299
rx_ring[i].forward = virt_to_le32desc(&rx_ring[i + 1]);
usr/src/grub/grub-0.97/netboot/tlan.c
300
rx_ring[i].cStat = TLAN_CSTAT_READY;
usr/src/grub/grub-0.97/netboot/tlan.c
301
rx_ring[i].frameSize = TLAN_MAX_FRAME_SIZE;
usr/src/grub/grub-0.97/netboot/tlan.c
302
rx_ring[i].buffer[0].count =
usr/src/grub/grub-0.97/netboot/tlan.c
304
rx_ring[i].buffer[0].address =
usr/src/grub/grub-0.97/netboot/tlan.c
306
rx_ring[i].buffer[1].count = 0;
usr/src/grub/grub-0.97/netboot/tlan.c
307
rx_ring[i].buffer[1].address = 0;
usr/src/grub/grub-0.97/netboot/tlan.c
311
rx_ring[i - 1].forward = virt_to_le32desc(&rx_ring[0]);
usr/src/grub/grub-0.97/netboot/tlan.c
511
outl(virt_to_bus(&rx_ring), BASE + TLAN_CH_PARM);
usr/src/grub/grub-0.97/netboot/tlan.c
539
u16 tmpCStat = le32_to_cpu(rx_ring[entry].cStat);
usr/src/grub/grub-0.97/netboot/tlan.c
554
framesize = rx_ring[entry].frameSize;
usr/src/grub/grub-0.97/netboot/tlan.c
565
rx_ring[entry].cStat = 0;
usr/src/grub/grub-0.97/netboot/tlan.c
573
if ((rx_ring[entry].cStat & TLAN_CSTAT_READY) ==
usr/src/grub/grub-0.97/netboot/tlan.c
601
rx_ring[entry].frameSize = TLAN_MAX_FRAME_SIZE;
usr/src/grub/grub-0.97/netboot/tlan.c
602
rx_ring[entry].cStat = TLAN_CSTAT_READY;
usr/src/grub/grub-0.97/netboot/tlan.c
911
priv->rxList = rx_ring;
usr/src/grub/grub-0.97/netboot/tulip.c
1034
outl(virt_to_le32desc(&rx_ring[0]), ioaddr + CSR3);
usr/src/grub/grub-0.97/netboot/tulip.c
1138
if (rx_ring[tp->cur_rx].status & 0x80000000)
usr/src/grub/grub-0.97/netboot/tulip.c
1147
nic->packetlen = (rx_ring[tp->cur_rx].status & 0x3FFF0000) >> 16;
usr/src/grub/grub-0.97/netboot/tulip.c
1150
if (rx_ring[tp->cur_rx].status & 0x00008000) {
usr/src/grub/grub-0.97/netboot/tulip.c
1152
rx_ring[tp->cur_rx].status = 0x80000000;
usr/src/grub/grub-0.97/netboot/tulip.c
1161
rx_ring[tp->cur_rx].status = 0x80000000;
usr/src/grub/grub-0.97/netboot/tulip.c
402
static struct tulip_rx_desc rx_ring[RX_RING_SIZE] __attribute__ ((aligned(4)));
usr/src/grub/grub-0.97/netboot/tulip.c
908
rx_ring[i].status = cpu_to_le32(0x80000000);
usr/src/grub/grub-0.97/netboot/tulip.c
909
rx_ring[i].length = cpu_to_le32(BUFLEN);
usr/src/grub/grub-0.97/netboot/tulip.c
910
rx_ring[i].buffer1 = virt_to_le32desc(&rxb[i * BUFLEN]);
usr/src/grub/grub-0.97/netboot/tulip.c
911
rx_ring[i].buffer2 = virt_to_le32desc(&rx_ring[i+1]);
usr/src/grub/grub-0.97/netboot/tulip.c
914
rx_ring[i-1].length = cpu_to_le32(DESC_RING_WRAP | BUFLEN);
usr/src/grub/grub-0.97/netboot/tulip.c
915
rx_ring[i-1].buffer2 = virt_to_le32desc(&rx_ring[0]);
usr/src/grub/grub-0.97/netboot/via-rhine.c
1121
tp->rx_ring = (struct rhine_rx_desc *) bus_to_virt (j);
usr/src/grub/grub-0.97/netboot/via-rhine.c
1160
outl (virt_to_bus (tp->rx_ring), dwCurrentRxDescAddr);
usr/src/grub/grub-0.97/netboot/via-rhine.c
1199
if (tp->rx_ring[tp->cur_rx].rx_status.bits.own_bit == 0)
usr/src/grub/grub-0.97/netboot/via-rhine.c
1218
rxstatus = tp->rx_ring[tp->cur_rx].rx_status.lw;
usr/src/grub/grub-0.97/netboot/via-rhine.c
1232
nic->packetlen = tp->rx_ring[tp->cur_rx].rx_status.bits.frame_length;
usr/src/grub/grub-0.97/netboot/via-rhine.c
1236
tp->rx_ring[tp->cur_rx].rx_status.bits.own_bit = 1;
usr/src/grub/grub-0.97/netboot/via-rhine.c
631
struct rhine_rx_desc *rx_ring;
usr/src/grub/grub-0.97/netboot/via-rhine.c
683
tp->rx_ring[i].rx_status.bits.own_bit = 1;
usr/src/grub/grub-0.97/netboot/via-rhine.c
684
tp->rx_ring[i].rx_ctrl.bits.rx_buf_size = 1536;
usr/src/grub/grub-0.97/netboot/via-rhine.c
686
tp->rx_ring[i].buf_addr_1 = virt_to_bus (tp->rx_buffs[i]);
usr/src/grub/grub-0.97/netboot/via-rhine.c
687
tp->rx_ring[i].buf_addr_2 = virt_to_bus (&tp->rx_ring[i + 1]);
usr/src/grub/grub-0.97/netboot/via-rhine.c
692
tp->rx_ring[i - 1].buf_addr_2 = virt_to_bus (&tp->rx_ring[0]);
usr/src/grub/grub-0.97/netboot/w89c840.c
231
struct w840_rx_desc rx_ring[RX_RING_SIZE];
usr/src/grub/grub-0.97/netboot/w89c840.c
313
writel(virt_to_bus(w840private.rx_ring), ioaddr + RxRingPtr);
usr/src/grub/grub-0.97/netboot/w89c840.c
452
memcpy(nic->packet, le32desc_to_virt(w840private.rx_ring[entry].buffer1), pkt_len);
usr/src/grub/grub-0.97/netboot/w89c840.c
456
w840private.rx_ring[entry].status = DescOwn;
usr/src/grub/grub-0.97/netboot/w89c840.c
474
w840private.rx_head_desc = &w840private.rx_ring[entry];
usr/src/grub/grub-0.97/netboot/w89c840.c
915
w840private.rx_head_desc = &w840private.rx_ring[0];
usr/src/grub/grub-0.97/netboot/w89c840.c
922
w840private.rx_ring[i].length = w840private.rx_buf_sz;
usr/src/grub/grub-0.97/netboot/w89c840.c
923
w840private.rx_ring[i].status = 0;
usr/src/grub/grub-0.97/netboot/w89c840.c
924
w840private.rx_ring[i].next_desc = virt_to_le32desc(&w840private.rx_ring[i+1]);
usr/src/grub/grub-0.97/netboot/w89c840.c
926
w840private.rx_ring[i].buffer1 = virt_to_le32desc(p + (PKT_BUF_SZ * i));
usr/src/grub/grub-0.97/netboot/w89c840.c
927
w840private.rx_ring[i].status = DescOwn | DescIntr;
usr/src/grub/grub-0.97/netboot/w89c840.c
931
w840private.rx_ring[i-1].length |= DescEndRing;
usr/src/grub/grub-0.97/netboot/w89c840.c
932
w840private.rx_ring[i-1].next_desc = virt_to_le32desc(&w840private.rx_ring[0]);
usr/src/uts/common/inet/ip/ip_squeue.c
447
ip_squeue_bind_ring(ill_t *ill, ill_rx_ring_t *rx_ring, processorid_t cpuid)
usr/src/uts/common/inet/ip/ip_squeue.c
450
ASSERT(rx_ring->rr_ill == ill);
usr/src/uts/common/inet/ip/ip_squeue.c
453
if (rx_ring->rr_ring_state == RR_FREE ||
usr/src/uts/common/inet/ip/ip_squeue.c
454
rx_ring->rr_ring_state == RR_FREE_INPROG) {
usr/src/uts/common/inet/ip/ip_squeue.c
459
if (ip_squeue_cpu_move(rx_ring->rr_sqp, cpuid) != -1)
usr/src/uts/common/inet/ip/ip_squeue.c
460
rx_ring->rr_ring_state = RR_SQUEUE_BOUND;
usr/src/uts/common/inet/ip/ip_squeue.c
469
ill_rx_ring_t *rx_ring, *ring_tbl;
usr/src/uts/common/inet/ip/ip_squeue.c
482
rx_ring = &ring_tbl[ip_rx_index];
usr/src/uts/common/inet/ip/ip_squeue.c
483
if (rx_ring->rr_ring_state == RR_FREE)
usr/src/uts/common/inet/ip/ip_squeue.c
503
bzero(rx_ring, sizeof (ill_rx_ring_t));
usr/src/uts/common/inet/ip/ip_squeue.c
504
rx_ring->rr_rx = mrfp->mrf_receive;
usr/src/uts/common/inet/ip/ip_squeue.c
505
rx_ring->rr_ip_accept = (ill->ill_isv6 != 0) ?
usr/src/uts/common/inet/ip/ip_squeue.c
509
rx_ring->rr_intr_handle = mrfp->mrf_intr_handle;
usr/src/uts/common/inet/ip/ip_squeue.c
510
rx_ring->rr_intr_enable = (ip_mac_intr_enable_t)mrfp->mrf_intr_enable;
usr/src/uts/common/inet/ip/ip_squeue.c
511
rx_ring->rr_intr_disable =
usr/src/uts/common/inet/ip/ip_squeue.c
513
rx_ring->rr_rx_handle = mrfp->mrf_rx_arg;
usr/src/uts/common/inet/ip/ip_squeue.c
514
rx_ring->rr_ill = ill;
usr/src/uts/common/inet/ip/ip_squeue.c
521
sq->sq_rx_ring = rx_ring;
usr/src/uts/common/inet/ip/ip_squeue.c
522
rx_ring->rr_sqp = sq;
usr/src/uts/common/inet/ip/ip_squeue.c
526
rx_ring->rr_ring_state = RR_SQUEUE_UNBOUND;
usr/src/uts/common/inet/ip/ip_squeue.c
536
(void) ip_squeue_bind_ring(ill, rx_ring, mrfp->mrf_cpu_id);
usr/src/uts/common/inet/ip/ip_squeue.c
539
return (rx_ring);
usr/src/uts/common/inet/ip/ip_squeue.c
547
ip_squeue_clean_ring(ill_t *ill, ill_rx_ring_t *rx_ring)
usr/src/uts/common/inet/ip/ip_squeue.c
552
ASSERT(rx_ring != NULL);
usr/src/uts/common/inet/ip/ip_squeue.c
556
if (rx_ring->rr_ring_state == RR_FREE) {
usr/src/uts/common/inet/ip/ip_squeue.c
560
rx_ring->rr_ring_state = RR_FREE_INPROG;
usr/src/uts/common/inet/ip/ip_squeue.c
561
sqp = rx_ring->rr_sqp;
usr/src/uts/common/inet/ip/ip_squeue.c
599
rx_ring->rr_ring_state = RR_FREE;
usr/src/uts/common/inet/ip/ip_squeue.c
608
ip_squeue_quiesce_ring(ill_t *ill, ill_rx_ring_t *rx_ring)
usr/src/uts/common/inet/ip/ip_squeue.c
613
ASSERT(rx_ring != NULL);
usr/src/uts/common/inet/ip/ip_squeue.c
615
sqp = rx_ring->rr_sqp;
usr/src/uts/common/inet/ip/ip_squeue.c
630
ip_squeue_restart_ring(ill_t *ill, ill_rx_ring_t *rx_ring)
usr/src/uts/common/inet/ip/ip_squeue.c
635
ASSERT(rx_ring != NULL);
usr/src/uts/common/inet/ip/ip_squeue.c
637
sqp = rx_ring->rr_sqp;
usr/src/uts/common/inet/ip/ip_squeue.c
663
ill_rx_ring_t *rx_ring;
usr/src/uts/common/inet/ip/ip_squeue.c
666
rx_ring = &ill->ill_dld_capab->idc_poll.idp_ring_tbl[idx];
usr/src/uts/common/inet/ip/ip_squeue.c
667
ip_squeue_clean_ring(ill, rx_ring);
usr/src/uts/common/inet/squeue.c
1082
ill_rx_ring_t *rx_ring;
usr/src/uts/common/inet/squeue.c
1131
rx_ring = sqp->sq_rx_ring;
usr/src/uts/common/inet/squeue.c
1132
ill = rx_ring->rr_ill;
usr/src/uts/common/inet/squeue.c
1155
rx_ring->rr_sqp = NULL;
usr/src/uts/common/inet/squeue.c
1161
rx_ring->rr_rx_handle = NULL;
usr/src/uts/common/inet/squeue.c
1162
rx_ring->rr_intr_handle = NULL;
usr/src/uts/common/inet/squeue.c
1163
rx_ring->rr_intr_enable = NULL;
usr/src/uts/common/inet/squeue.c
1164
rx_ring->rr_intr_disable = NULL;
usr/src/uts/common/inet/squeue.c
183
#define SQS_POLLING_ON(sqp, sq_poll_capable, rx_ring) { \
usr/src/uts/common/inet/squeue.c
186
ASSERT(rx_ring != NULL); \
usr/src/uts/common/inet/squeue.c
189
if (rx_ring->rr_intr_disable(rx_ring->rr_intr_handle)) \
usr/src/uts/common/inet/squeue.c
195
#define SQS_POLLING_OFF(sqp, sq_poll_capable, rx_ring) { \
usr/src/uts/common/inet/squeue.c
198
ASSERT(rx_ring != NULL); \
usr/src/uts/common/inet/squeue.c
202
rx_ring->rr_intr_enable(rx_ring->rr_intr_handle); \
usr/src/uts/common/io/aggr/aggr_grp.c
2249
aggr_pseudo_rx_ring_t *rx_ring = (aggr_pseudo_rx_ring_t *)rdriver;
usr/src/uts/common/io/aggr/aggr_grp.c
2251
if (rx_ring->arr_hw_rh != NULL) {
usr/src/uts/common/io/aggr/aggr_grp.c
2252
*val = mac_pseudo_rx_ring_stat_get(rx_ring->arr_hw_rh, stat);
usr/src/uts/common/io/aggr/aggr_grp.c
2254
aggr_port_t *port = rx_ring->arr_port;
usr/src/uts/common/io/aggr/aggr_grp.c
2581
aggr_pseudo_rx_ring_t *rx_ring;
usr/src/uts/common/io/aggr/aggr_grp.c
2587
rx_ring = rx_group->arg_rings + index;
usr/src/uts/common/io/aggr/aggr_grp.c
2588
rx_ring->arr_rh = rh;
usr/src/uts/common/io/aggr/aggr_grp.c
2594
aggr_mac_intr.mi_handle = (mac_intr_handle_t)rx_ring;
usr/src/uts/common/io/aggr/aggr_grp.c
2599
infop->mri_driver = (mac_ring_driver_t)rx_ring;
usr/src/uts/common/io/bge/bge_kstats.c
1170
recv_ring_t *rx_ring = (recv_ring_t *)rh;
usr/src/uts/common/io/bge/bge_kstats.c
1174
*val = rx_ring->rx_bytes;
usr/src/uts/common/io/bge/bge_kstats.c
1178
*val = rx_ring->rx_pkts;
usr/src/uts/common/io/bge/bge_main2.c
1687
recv_ring_t *rx_ring;
usr/src/uts/common/io/bge/bge_main2.c
1689
rx_ring = (recv_ring_t *)rh;
usr/src/uts/common/io/bge/bge_main2.c
1690
mutex_enter(rx_ring->rx_lock);
usr/src/uts/common/io/bge/bge_main2.c
1691
rx_ring->ring_gen_num = mr_gen_num;
usr/src/uts/common/io/bge/bge_main2.c
1692
mutex_exit(rx_ring->rx_lock);
usr/src/uts/common/io/bge/bge_main2.c
1710
recv_ring_t *rx_ring;
usr/src/uts/common/io/bge/bge_main2.c
1714
rx_ring = &bgep->recv[rg_index];
usr/src/uts/common/io/bge/bge_main2.c
1715
rx_ring->ring_handle = rh;
usr/src/uts/common/io/bge/bge_main2.c
1717
infop->mri_driver = (mac_ring_driver_t)rx_ring;
usr/src/uts/common/io/bge/bge_main2.c
1724
mintr->mi_handle = (mac_intr_handle_t)rx_ring;
usr/src/uts/common/io/bge/bge_main2.c
1751
recv_ring_t *rx_ring;
usr/src/uts/common/io/bge/bge_main2.c
1755
rx_ring = &bgep->recv[rg_index];
usr/src/uts/common/io/bge/bge_main2.c
1756
rx_ring->ring_group_handle = gh;
usr/src/uts/common/io/bge/bge_main2.c
1758
infop->mgi_driver = (mac_group_driver_t)rx_ring;
usr/src/uts/common/io/e1000g/e1000g_alloc.c
1253
Adapter = rx_data->rx_ring->adapter;
usr/src/uts/common/io/e1000g/e1000g_alloc.c
1290
Adapter = rx_data->rx_ring->adapter;
usr/src/uts/common/io/e1000g/e1000g_alloc.c
1331
Adapter = rx_data->rx_ring->adapter;
usr/src/uts/common/io/e1000g/e1000g_alloc.c
241
rx_data = Adapter->rx_ring->rx_data;
usr/src/uts/common/io/e1000g/e1000g_alloc.c
259
rx_data = Adapter->rx_ring->rx_data;
usr/src/uts/common/io/e1000g/e1000g_alloc.c
484
Adapter = rx_data->rx_ring->adapter;
usr/src/uts/common/io/e1000g/e1000g_alloc.c
710
rx_data = Adapter->rx_ring->rx_data;
usr/src/uts/common/io/e1000g/e1000g_alloc.c
770
rx_data = Adapter->rx_ring->rx_data;
usr/src/uts/common/io/e1000g/e1000g_main.c
1269
e1000g_rx_ring_t *rx_ring;
usr/src/uts/common/io/e1000g/e1000g_main.c
1287
rx_ring = Adapter->rx_ring;
usr/src/uts/common/io/e1000g/e1000g_main.c
1289
mutex_init(&rx_ring->rx_lock, NULL,
usr/src/uts/common/io/e1000g/e1000g_main.c
1300
e1000g_rx_ring_t *rx_ring;
usr/src/uts/common/io/e1000g/e1000g_main.c
1309
rx_ring = Adapter->rx_ring;
usr/src/uts/common/io/e1000g/e1000g_main.c
1310
mutex_destroy(&rx_ring->rx_lock);
usr/src/uts/common/io/e1000g/e1000g_main.c
1680
e1000g_rx_ring_t *rx_ring;
usr/src/uts/common/io/e1000g/e1000g_main.c
1683
rx_ring = Adapter->rx_ring;
usr/src/uts/common/io/e1000g/e1000g_main.c
1691
rx_data->rx_ring = rx_ring;
usr/src/uts/common/io/e1000g/e1000g_main.c
1698
rx_ring->rx_data = rx_data;
usr/src/uts/common/io/e1000g/e1000g_main.c
1898
e1000g_rx_ring_t *rx_ring = (e1000g_rx_ring_t *)arg;
usr/src/uts/common/io/e1000g/e1000g_main.c
1903
adapter = rx_ring->adapter;
usr/src/uts/common/io/e1000g/e1000g_main.c
1912
mutex_enter(&rx_ring->rx_lock);
usr/src/uts/common/io/e1000g/e1000g_main.c
1913
mp = e1000g_receive(rx_ring, &tail, bytes_to_pickup);
usr/src/uts/common/io/e1000g/e1000g_main.c
1914
mutex_exit(&rx_ring->rx_lock);
usr/src/uts/common/io/e1000g/e1000g_main.c
1997
rx_data = Adapter->rx_ring->rx_data;
usr/src/uts/common/io/e1000g/e1000g_main.c
2122
rx_data = Adapter->rx_ring->rx_data;
usr/src/uts/common/io/e1000g/e1000g_main.c
2148
e1000g_rx_data_t *rx_data = Adapter->rx_ring->rx_data;
usr/src/uts/common/io/e1000g/e1000g_main.c
2439
e1000g_rx_ring_t *rx_ring;
usr/src/uts/common/io/e1000g/e1000g_main.c
2441
rx_ring = Adapter->rx_ring;
usr/src/uts/common/io/e1000g/e1000g_main.c
2442
mutex_enter(&rx_ring->rx_lock);
usr/src/uts/common/io/e1000g/e1000g_main.c
2449
if (!rx_ring->poll_flag)
usr/src/uts/common/io/e1000g/e1000g_main.c
2450
mp = e1000g_receive(rx_ring, &tail,
usr/src/uts/common/io/e1000g/e1000g_main.c
2452
mutex_exit(&rx_ring->rx_lock);
usr/src/uts/common/io/e1000g/e1000g_main.c
2455
mac_rx_ring(Adapter->mh, rx_ring->mrh,
usr/src/uts/common/io/e1000g/e1000g_main.c
2456
mp, rx_ring->ring_gen_num);
usr/src/uts/common/io/e1000g/e1000g_main.c
2906
e1000g_rx_ring_t *rx_ring = adapter->rx_ring;
usr/src/uts/common/io/e1000g/e1000g_main.c
2916
mutex_enter(&rx_ring->rx_lock);
usr/src/uts/common/io/e1000g/e1000g_main.c
2917
rx_ring->mrh = NULL;
usr/src/uts/common/io/e1000g/e1000g_main.c
2919
mutex_exit(&rx_ring->rx_lock);
usr/src/uts/common/io/e1000g/e1000g_main.c
2927
e1000g_rx_ring_t *rx_ring = adapter->rx_ring;
usr/src/uts/common/io/e1000g/e1000g_main.c
2929
mutex_enter(&rx_ring->rx_lock);
usr/src/uts/common/io/e1000g/e1000g_main.c
2936
rx_ring->mrh = rx_ring->mrh_init;
usr/src/uts/common/io/e1000g/e1000g_main.c
2937
mutex_exit(&rx_ring->rx_lock);
usr/src/uts/common/io/e1000g/e1000g_main.c
2950
e1000g_rx_ring_t *rx_ring = (e1000g_rx_ring_t *)intrh;
usr/src/uts/common/io/e1000g/e1000g_main.c
2951
struct e1000g *adapter = rx_ring->adapter;
usr/src/uts/common/io/e1000g/e1000g_main.c
2962
mutex_enter(&rx_ring->rx_lock);
usr/src/uts/common/io/e1000g/e1000g_main.c
2963
rx_ring->poll_flag = 0;
usr/src/uts/common/io/e1000g/e1000g_main.c
2964
mutex_exit(&rx_ring->rx_lock);
usr/src/uts/common/io/e1000g/e1000g_main.c
2983
e1000g_rx_ring_t *rx_ring = (e1000g_rx_ring_t *)intrh;
usr/src/uts/common/io/e1000g/e1000g_main.c
2984
struct e1000g *adapter = rx_ring->adapter;
usr/src/uts/common/io/e1000g/e1000g_main.c
2993
mutex_enter(&rx_ring->rx_lock);
usr/src/uts/common/io/e1000g/e1000g_main.c
2994
rx_ring->poll_flag = 1;
usr/src/uts/common/io/e1000g/e1000g_main.c
2995
mutex_exit(&rx_ring->rx_lock);
usr/src/uts/common/io/e1000g/e1000g_main.c
3104
e1000g_rx_ring_t *rx_ring = (e1000g_rx_ring_t *)rh;
usr/src/uts/common/io/e1000g/e1000g_main.c
3106
mutex_enter(&rx_ring->rx_lock);
usr/src/uts/common/io/e1000g/e1000g_main.c
3107
rx_ring->ring_gen_num = mr_gen_num;
usr/src/uts/common/io/e1000g/e1000g_main.c
3108
mutex_exit(&rx_ring->rx_lock);
usr/src/uts/common/io/e1000g/e1000g_main.c
3126
e1000g_rx_ring_t *rx_ring = Adapter->rx_ring;
usr/src/uts/common/io/e1000g/e1000g_main.c
3135
rx_ring->mrh = rx_ring->mrh_init = rh;
usr/src/uts/common/io/e1000g/e1000g_main.c
3136
infop->mri_driver = (mac_ring_driver_t)rx_ring;
usr/src/uts/common/io/e1000g/e1000g_main.c
3144
mintr->mi_handle = (mac_intr_handle_t)rx_ring;
usr/src/uts/common/io/e1000g/e1000g_main.c
437
Adapter->rx_ring->adapter = Adapter;
usr/src/uts/common/io/e1000g/e1000g_rx.c
123
Adapter = rx_data->rx_ring->adapter;
usr/src/uts/common/io/e1000g/e1000g_rx.c
161
rx_data = Adapter->rx_ring->rx_data;
usr/src/uts/common/io/e1000g/e1000g_rx.c
388
Adapter = rx_data->rx_ring->adapter;
usr/src/uts/common/io/e1000g/e1000g_rx.c
433
e1000g_receive(e1000g_rx_ring_t *rx_ring, mblk_t **tail, uint_t sz)
usr/src/uts/common/io/e1000g/e1000g_rx.c
463
Adapter = rx_ring->adapter;
usr/src/uts/common/io/e1000g/e1000g_rx.c
464
rx_data = rx_ring->rx_data;
usr/src/uts/common/io/e1000g/e1000g_rx.c
482
E1000G_DEBUG_STAT(rx_ring->stat_none);
usr/src/uts/common/io/e1000g/e1000g_rx.c
590
E1000G_STAT(rx_ring->stat_error);
usr/src/uts/common/io/e1000g/e1000g_rx.c
614
E1000G_STAT(rx_ring->stat_crc_only_pkt);
usr/src/uts/common/io/e1000g/e1000g_rx.c
682
E1000G_DEBUG_STAT(rx_ring->stat_no_freepkt);
usr/src/uts/common/io/e1000g/e1000g_rx.c
702
E1000G_STAT(rx_ring->stat_allocb_fail);
usr/src/uts/common/io/e1000g/e1000g_rx.c
785
E1000G_STAT(rx_ring->stat_size_error);
usr/src/uts/common/io/e1000g/e1000g_rx.c
802
E1000G_DEBUG_STAT_COND(rx_ring->stat_multi_desc,
usr/src/uts/common/io/e1000g/e1000g_stat.c
155
e1000g_rx_ring_t *rx_ring;
usr/src/uts/common/io/e1000g/e1000g_stat.c
172
rx_ring = Adapter->rx_ring;
usr/src/uts/common/io/e1000g/e1000g_stat.c
174
rx_data = rx_ring->rx_data;
usr/src/uts/common/io/e1000g/e1000g_stat.c
181
e1000g_ksp->rx_error.value.ul = rx_ring->stat_error;
usr/src/uts/common/io/e1000g/e1000g_stat.c
182
e1000g_ksp->rx_allocb_fail.value.ul = rx_ring->stat_allocb_fail;
usr/src/uts/common/io/e1000g/e1000g_stat.c
183
e1000g_ksp->rx_size_error.value.ul = rx_ring->stat_size_error;
usr/src/uts/common/io/e1000g/e1000g_stat.c
192
e1000g_ksp->rx_none.value.ul = rx_ring->stat_none;
usr/src/uts/common/io/e1000g/e1000g_stat.c
193
e1000g_ksp->rx_multi_desc.value.ul = rx_ring->stat_multi_desc;
usr/src/uts/common/io/e1000g/e1000g_stat.c
194
e1000g_ksp->rx_no_freepkt.value.ul = rx_ring->stat_no_freepkt;
usr/src/uts/common/io/e1000g/e1000g_stat.c
938
e1000g_rx_ring_t *rx_ring = (e1000g_rx_ring_t *)rh;
usr/src/uts/common/io/e1000g/e1000g_stat.c
939
struct e1000g *Adapter = rx_ring->adapter;
usr/src/uts/common/io/e1000g/e1000g_sw.h
1062
mblk_t *e1000g_receive(e1000g_rx_ring_t *rx_ring, mblk_t **tail, uint_t sz);
usr/src/uts/common/io/e1000g/e1000g_sw.h
789
struct _e1000g_rx_ring *rx_ring;
usr/src/uts/common/io/e1000g/e1000g_sw.h
916
e1000g_rx_ring_t rx_ring[1];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1007
ql_add_sbuf_to_free_list(struct rx_ring *rx_ring,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1013
free_idx = rx_ring->sbq_free_tail;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1014
rx_ring->sbuf_free[free_idx] = sbq_desc;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1015
ASSERT(rx_ring->sbuf_free_count <= rx_ring->sbq_len);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1017
if (free_idx >= rx_ring->sbq_len)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1019
rx_ring->sbq_free_tail = free_idx;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1020
atomic_inc_32(&rx_ring->sbuf_free_count);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1027
ql_get_sbuf_from_free_list(struct rx_ring *rx_ring)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1032
free_idx = rx_ring->sbq_free_head;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1034
sbq_desc = rx_ring->sbuf_free[free_idx];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1035
rx_ring->sbuf_free[free_idx] = NULL;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1038
if (free_idx >= rx_ring->sbq_len)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1040
rx_ring->sbq_free_head = free_idx;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1041
atomic_dec_32(&rx_ring->sbuf_free_count);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1050
ql_add_lbuf_to_in_use_list(struct rx_ring *rx_ring,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1055
inuse_idx = rx_ring->lbq_use_tail;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1057
rx_ring->lbuf_in_use[inuse_idx] = lbq_desc;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1059
if (inuse_idx >= rx_ring->lbq_len)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1061
rx_ring->lbq_use_tail = inuse_idx;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1062
atomic_inc_32(&rx_ring->lbuf_in_use_count);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1069
ql_get_lbuf_from_in_use_list(struct rx_ring *rx_ring)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1075
inuse_idx = rx_ring->lbq_use_head;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1076
lbq_desc = rx_ring->lbuf_in_use[inuse_idx];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1077
rx_ring->lbuf_in_use[inuse_idx] = NULL;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1081
if (inuse_idx >= rx_ring->lbq_len)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1083
rx_ring->lbq_use_head = inuse_idx;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1084
atomic_dec_32(&rx_ring->lbuf_in_use_count);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1085
atomic_inc_32(&rx_ring->rx_indicate);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1093
rx_ring->lbq_buf_size, 0, &lbq_desc->rx_recycle);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1103
ql_add_lbuf_to_free_list(struct rx_ring *rx_ring,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1109
free_idx = rx_ring->lbq_free_tail;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1110
rx_ring->lbuf_free[free_idx] = lbq_desc;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1112
if (free_idx >= rx_ring->lbq_len)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1114
rx_ring->lbq_free_tail = free_idx;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1115
atomic_inc_32(&rx_ring->lbuf_free_count);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1116
ASSERT(rx_ring->lbuf_free_count <= rx_ring->lbq_len);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1123
ql_get_lbuf_from_free_list(struct rx_ring *rx_ring)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1128
free_idx = rx_ring->lbq_free_head;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1130
lbq_desc = rx_ring->lbuf_free[free_idx];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1131
rx_ring->lbuf_free[free_idx] = NULL;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1135
if (free_idx >= rx_ring->lbq_len)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1137
rx_ring->lbq_free_head = free_idx;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1138
atomic_dec_32(&rx_ring->lbuf_free_count);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1149
struct rx_ring *rx_ring = sbq_desc->rx_ring;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1151
qlge_t *qlge = (qlge_t *)rx_ring->qlge;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1155
mutex_enter(&rx_ring->sbq_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1165
atomic_dec_32(&rx_ring->rx_indicate);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1166
mutex_exit(&rx_ring->sbq_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1170
if (rx_ring->rx_indicate == 0)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1179
temp = ql_read_doorbell_reg(qlge, rx_ring->sbq_prod_idx_db_reg);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1189
if (num_free_buffers < qlge->rx_sb_low_count[rx_ring->cq_id])
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1190
qlge->rx_sb_low_count[rx_ring->cq_id] = num_free_buffers;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1195
if (rx_ring->rx_indicate > 0xFF000000)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1198
rx_ring->cq_id, rx_ring->rx_indicate, rx_ring->mac_flags,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1204
rx_ring->sbq_buf_size, 0, &sbq_desc->rx_recycle);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1206
rx_ring->rx_failed_sbq_allocs++;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1211
atomic_dec_32(&rx_ring->rx_indicate);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1213
ql_add_sbuf_to_free_list(rx_ring, sbq_desc);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1216
if ((rx_ring->sbuf_free_count >= MIN_BUFFERS_FREE_COUNT) &&
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1218
sbq_entry = rx_ring->sbq_dma.vaddr;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1219
sbq_entry += rx_ring->sbq_prod_idx;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1221
while (rx_ring->sbuf_free_count > MIN_BUFFERS_ARM_COUNT) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1223
sbq_desc = ql_get_sbuf_from_free_list(rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1227
rx_ring->sbq_prod_idx++;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1228
if (rx_ring->sbq_prod_idx >= rx_ring->sbq_len) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1229
rx_ring->sbq_prod_idx = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1230
sbq_entry = rx_ring->sbq_dma.vaddr;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1233
ql_add_sbuf_to_in_use_list(rx_ring, sbq_desc);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1237
ql_update_sbq_prod_idx(qlge, rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1240
mutex_exit(&rx_ring->sbq_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1242
__func__, qlge->instance, rx_ring->sbuf_free_count));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1264
struct rx_ring *rx_ring = lbq_desc->rx_ring;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1266
qlge_t *qlge = rx_ring->qlge;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1269
mutex_enter(&rx_ring->lbq_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1278
atomic_dec_32(&rx_ring->rx_indicate);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1279
mutex_exit(&rx_ring->lbq_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1283
if (rx_ring->rx_indicate == 0)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1292
temp = ql_read_doorbell_reg(qlge, rx_ring->lbq_prod_idx_db_reg);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1303
if (num_free_buffers < qlge->rx_lb_low_count[rx_ring->cq_id]) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1304
qlge->rx_lb_low_count[rx_ring->cq_id] = num_free_buffers;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1309
if (rx_ring->rx_indicate > 0xFF000000)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1312
rx_ring->cq_id, rx_ring->rx_indicate, rx_ring->mac_flags,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1318
rx_ring->lbq_buf_size, 0, &lbq_desc->rx_recycle);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1320
rx_ring->rx_failed_lbq_allocs++;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1325
atomic_dec_32(&rx_ring->rx_indicate);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1327
ql_add_lbuf_to_free_list(rx_ring, lbq_desc);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1330
if ((rx_ring->lbuf_free_count >= MIN_BUFFERS_FREE_COUNT) &&
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1332
lbq_entry = rx_ring->lbq_dma.vaddr;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1333
lbq_entry += rx_ring->lbq_prod_idx;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1334
while (rx_ring->lbuf_free_count > MIN_BUFFERS_ARM_COUNT) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1336
lbq_desc = ql_get_lbuf_from_free_list(rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1340
rx_ring->lbq_prod_idx++;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1341
if (rx_ring->lbq_prod_idx >= rx_ring->lbq_len) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1342
rx_ring->lbq_prod_idx = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1343
lbq_entry = rx_ring->lbq_dma.vaddr;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1347
ql_add_lbuf_to_in_use_list(rx_ring, lbq_desc);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1351
ql_update_lbq_prod_idx(rx_ring->qlge, rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1354
mutex_exit(&rx_ring->lbq_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1356
__func__, rx_ring->lbuf_free_count));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1375
ql_free_sbq_buffers(struct rx_ring *rx_ring)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1379
uint32_t j = rx_ring->sbq_free_head;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1382
for (i = 0; i < rx_ring->sbuf_free_count; i++) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1383
sbq_desc = rx_ring->sbuf_free[j];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1386
if (j >= rx_ring->sbq_len) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1394
rx_ring->sbuf_free_count = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1396
j = rx_ring->sbq_use_head;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1397
for (i = 0; i < rx_ring->sbuf_in_use_count; i++) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1398
sbq_desc = rx_ring->sbuf_in_use[j];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1401
if (j >= rx_ring->sbq_len) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1409
rx_ring->sbuf_in_use_count = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1411
sbq_desc = &rx_ring->sbq_desc[0];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1412
for (i = 0; i < rx_ring->sbq_len; i++, sbq_desc++) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1429
rx_ring->sbuf_free_count, rx_ring->sbuf_in_use_count, force_cnt);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1431
if (rx_ring->sbuf_in_use != NULL) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1432
kmem_free(rx_ring->sbuf_in_use, (rx_ring->sbq_len *
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1434
rx_ring->sbuf_in_use = NULL;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1437
if (rx_ring->sbuf_free != NULL) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1438
kmem_free(rx_ring->sbuf_free, (rx_ring->sbq_len *
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1440
rx_ring->sbuf_free = NULL;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1446
ql_alloc_sbufs(qlge_t *qlge, struct rx_ring *rx_ring)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1452
rx_ring->sbq_use_head = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1453
rx_ring->sbq_use_tail = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1454
rx_ring->sbuf_in_use_count = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1455
rx_ring->sbq_free_head = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1456
rx_ring->sbq_free_tail = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1457
rx_ring->sbuf_free_count = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1458
rx_ring->sbuf_free = kmem_zalloc(rx_ring->sbq_len *
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1460
if (rx_ring->sbuf_free == NULL) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1467
rx_ring->sbuf_in_use = kmem_zalloc(rx_ring->sbq_len *
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1469
if (rx_ring->sbuf_in_use == NULL) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1476
sbq_desc = &rx_ring->sbq_desc[0];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1478
for (i = 0; i < rx_ring->sbq_len; i++, sbq_desc++) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1484
(size_t)rx_ring->sbq_buf_size, /* mem size */
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1498
sbq_desc->rx_ring = rx_ring;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1504
rx_ring->sbq_buf_size, 0, &sbq_desc->rx_recycle);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1509
ql_add_sbuf_to_free_list(rx_ring, sbq_desc);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1515
ql_free_sbq_buffers(rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1520
ql_free_lbq_buffers(struct rx_ring *rx_ring)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1526
j = rx_ring->lbq_free_head;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1527
for (i = 0; i < rx_ring->lbuf_free_count; i++) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1528
lbq_desc = rx_ring->lbuf_free[j];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1531
if (j >= rx_ring->lbq_len)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1538
rx_ring->lbuf_free_count = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1540
j = rx_ring->lbq_use_head;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1541
for (i = 0; i < rx_ring->lbuf_in_use_count; i++) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1542
lbq_desc = rx_ring->lbuf_in_use[j];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1545
if (j >= rx_ring->lbq_len) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1553
rx_ring->lbuf_in_use_count = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1555
lbq_desc = &rx_ring->lbq_desc[0];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1556
for (i = 0; i < rx_ring->lbq_len; i++, lbq_desc++) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1572
rx_ring->lbuf_free_count, rx_ring->lbuf_in_use_count,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1576
if (rx_ring->lbuf_in_use != NULL) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1577
kmem_free(rx_ring->lbuf_in_use, (rx_ring->lbq_len *
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1579
rx_ring->lbuf_in_use = NULL;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1582
if (rx_ring->lbuf_free != NULL) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1583
kmem_free(rx_ring->lbuf_free, (rx_ring->lbq_len *
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1585
rx_ring->lbuf_free = NULL;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1591
ql_alloc_lbufs(qlge_t *qlge, struct rx_ring *rx_ring)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1598
rx_ring->lbq_use_head = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1599
rx_ring->lbq_use_tail = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1600
rx_ring->lbuf_in_use_count = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1601
rx_ring->lbq_free_head = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1602
rx_ring->lbq_free_tail = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1603
rx_ring->lbuf_free_count = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1604
rx_ring->lbuf_free = kmem_zalloc(rx_ring->lbq_len *
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1606
if (rx_ring->lbuf_free == NULL) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1613
rx_ring->lbuf_in_use = kmem_zalloc(rx_ring->lbq_len *
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1616
if (rx_ring->lbuf_in_use == NULL) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1626
lbq_desc = &rx_ring->lbq_desc[0];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1627
for (i = 0; i < rx_ring->lbq_len; i++, lbq_desc++) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1628
rx_ring->lbq_buf_size = lbq_buf_size;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1634
(size_t)rx_ring->lbq_buf_size, /* mem size */
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1648
lbq_desc->rx_ring = rx_ring;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1654
rx_ring->lbq_buf_size, 0, &lbq_desc->rx_recycle);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1659
ql_add_lbuf_to_free_list(rx_ring, lbq_desc);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1665
ql_free_lbq_buffers(rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1676
struct rx_ring *rx_ring;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1679
rx_ring = &qlge->rx_ring[i];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1680
if (rx_ring->type != TX_Q) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1681
ql_free_lbq_buffers(rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1682
ql_free_sbq_buffers(rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1693
struct rx_ring *rx_ring;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1697
rx_ring = &qlge->rx_ring[i];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1698
if (rx_ring->type != TX_Q) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1699
if (ql_alloc_sbufs(qlge, rx_ring) != DDI_SUCCESS)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1701
if (ql_alloc_lbufs(qlge, rx_ring) != DDI_SUCCESS)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1707
if (qlge->rx_ring[i].type == RX_Q) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1725
ql_init_lbq_ring(struct rx_ring *rx_ring)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
173
ql_read_sh_reg(qlge_t *qlge, struct rx_ring *rx_ring)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1730
bzero(rx_ring->lbq_desc, rx_ring->lbq_len * sizeof (struct bq_desc));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1731
for (i = 0; i < rx_ring->lbq_len; i++) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1732
lbq_desc = &rx_ring->lbq_desc[i];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1741
ql_init_sbq_ring(struct rx_ring *rx_ring)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1746
bzero(rx_ring->sbq_desc, rx_ring->sbq_len * sizeof (struct bq_desc));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1747
for (i = 0; i < rx_ring->sbq_len; i++) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1748
sbq_desc = &rx_ring->sbq_desc[i];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
179
rx_ring->prod_idx_sh_reg_offset,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
183
(uint32_t *)rx_ring->prod_idx_sh_reg);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1874
ql_build_rx_mp(qlge_t *qlge, struct rx_ring *rx_ring,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1901
rx_ring->frame_too_long++;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1903
rx_ring->frame_too_short++;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1905
rx_ring->fcs_err++;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1925
(rx_ring->sbuf_in_use_count <= qlge->rx_copy_threshold) ||
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1926
(rx_ring->lbuf_in_use_count <= qlge->rx_copy_threshold)) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1951
sbq_desc = ql_get_sbuf_from_in_use_list(rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
1964
__func__, qlge->instance, rx_ring->cq_id);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2022
sbq_desc = ql_get_sbuf_from_in_use_list(rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2042
__func__, qlge->instance, rx_ring->cq_id);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2090
lbq_desc = ql_get_lbuf_from_in_use_list(rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2110
__func__, qlge->instance, rx_ring->cq_id);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2165
sbq_desc = ql_get_sbuf_from_in_use_list(rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2180
lbq_desc = ql_get_lbuf_from_in_use_list(rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2193
size = (payload_len < rx_ring->lbq_buf_size)?
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2194
payload_len : rx_ring->lbq_buf_size;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2304
ql_update_cq(struct rx_ring *rx_ring)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2306
rx_ring->cnsmr_idx++;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2307
rx_ring->curr_entry++;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2308
if (rx_ring->cnsmr_idx >= rx_ring->cq_len) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2309
rx_ring->cnsmr_idx = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2310
rx_ring->curr_entry = rx_ring->cq_dma.vaddr;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2318
ql_write_cq_idx(struct rx_ring *rx_ring)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2320
qlge_t *qlge = rx_ring->qlge;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2322
ql_write_doorbell_reg(qlge, rx_ring->cnsmr_idx_db_reg,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2323
rx_ring->cnsmr_idx);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2441
ql_ring_rx(struct rx_ring *rx_ring, int poll_bytes)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2443
qlge_t *qlge = rx_ring->qlge;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2444
uint32_t prod = ql_read_sh_reg(qlge, rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2461
temp = ql_read_doorbell_reg(qlge, rx_ring->cnsmr_idx_db_reg);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2471
if (num_free_entries < qlge->cq_low_count[rx_ring->cq_id])
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2472
qlge->cq_low_count[rx_ring->cq_id] = num_free_entries;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2478
while ((prod != rx_ring->cnsmr_idx)) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2481
__func__, rx_ring->cq_id, prod, rx_ring->cnsmr_idx));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2483
net_rsp = (struct ib_mac_iocb_rsp *)rx_ring->curr_entry;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2484
(void) ddi_dma_sync(rx_ring->cq_dma.dma_handle,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2486
(uintptr_t)rx_ring->cq_dma.vaddr),
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2489
rx_ring->curr_entry, 8, (size_t)sizeof (*net_rsp));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2506
mp = ql_build_rx_mp(qlge, rx_ring, net_rsp);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2508
if (rx_ring->mac_flags != QL_MAC_STARTED) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2515
ASSERT(rx_ring->rx_indicate
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2516
<= rx_ring->cq_len);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2520
__func__, rx_ring->mac_flags,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2521
rx_ring->rx_indicate);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2527
rx_ring->cq_id));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2528
rx_ring->rx_pkt_dropped_mac_unenabled++;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2531
mutex_exit(&rx_ring->rx_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2533
mutex_enter(&rx_ring->rx_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2545
rx_ring->rx_packets++;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2546
rx_ring->rx_bytes += length;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2553
rx_ring->cq_id));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2554
rx_ring->rx_packets_dropped_no_buffer++;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2568
"opcode = %x.", __func__, rx_ring->cq_id,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2573
ql_update_cq(rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2574
prod = ql_read_sh_reg(qlge, rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2580
rx_ring->hist[7]++;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2582
rx_ring->hist[6]++;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2584
rx_ring->hist[5]++;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2586
rx_ring->hist[4]++;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2588
rx_ring->hist[3]++;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2590
rx_ring->hist[2]++;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2592
rx_ring->hist[1]++;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2594
rx_ring->hist[0]++;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2598
ql_write_cq_idx(rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2601
ql_enable_completion_interrupt(rx_ring->qlge, rx_ring->irq);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2675
ql_clean_outbound_rx_ring(struct rx_ring *rx_ring)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2677
qlge_t *qlge = rx_ring->qlge;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2678
uint32_t prod = ql_read_sh_reg(qlge, rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2684
mutex_enter(&rx_ring->rx_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2692
temp = ql_read_doorbell_reg(qlge, rx_ring->cnsmr_idx_db_reg);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2702
if (num_free_entries < qlge->cq_low_count[rx_ring->cq_id])
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2703
qlge->cq_low_count[rx_ring->cq_id] = num_free_entries;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2708
while (prod != rx_ring->cnsmr_idx) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2712
rx_ring->cq_id, prod, rx_ring->cnsmr_idx));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2714
net_rsp = (struct ob_mac_iocb_rsp *)rx_ring->curr_entry;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2715
(void) ddi_dma_sync(rx_ring->cq_dma.dma_handle,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2717
(uintptr_t)rx_ring->cq_dma.vaddr),
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2722
rx_ring->curr_entry, 8,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2741
ql_update_cq(rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2742
prod = ql_read_sh_reg(qlge, rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2744
ql_write_cq_idx(rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2746
mutex_exit(&rx_ring->rx_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2748
net_rsp = (struct ob_mac_iocb_rsp *)rx_ring->curr_entry;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2867
struct rx_ring *rx_ring = (struct rx_ring *)((void *)arg1);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2868
struct rx_ring *ob_ring;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2869
qlge_t *qlge = rx_ring->qlge;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2879
++qlge->rx_interrupts[rx_ring->cq_id];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2895
ob_ring = &qlge->rx_ring[qlge->isr_stride];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2904
rx_ring = &qlge->rx_ring[0];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2905
prod = ql_read_sh_reg(qlge, rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2907
prod, rx_ring->cnsmr_idx));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2909
if (prod != rx_ring->cnsmr_idx) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2912
mutex_enter(&rx_ring->rx_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2913
mp = ql_ring_rx(rx_ring, QLGE_POLL_ALL);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2914
mutex_exit(&rx_ring->rx_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2917
RX_UPSTREAM(rx_ring, mp);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2965
rx_ring = &qlge->rx_ring[i];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2967
if (ql_read_sh_reg(qlge, rx_ring) !=
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2968
rx_ring->cnsmr_idx) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2973
rx_ring->irq);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2974
if (rx_ring->type == TX_Q) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2976
rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2978
rx_ring->qlge, rx_ring->irq);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2980
mutex_enter(&rx_ring->rx_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2981
mp = ql_ring_rx(rx_ring, QLGE_POLL_ALL);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2982
mutex_exit(&rx_ring->rx_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2984
RX_UPSTREAM(rx_ring, mp);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2986
if (rx_ring->mac_flags ==
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
2991
rx_ring->rx_indicate);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3011
struct rx_ring *rx_ring = (struct rx_ring *)((void *)arg1);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3012
qlge_t *qlge = rx_ring->qlge;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3015
++qlge->rx_interrupts[rx_ring->cq_id];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3016
(void) ql_clean_outbound_rx_ring(rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3017
ql_enable_completion_interrupt(rx_ring->qlge, rx_ring->irq);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3029
struct rx_ring *rx_ring = (struct rx_ring *)((void *)arg1);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3030
struct rx_ring *ob_ring;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3031
qlge_t *qlge = rx_ring->qlge;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3035
QL_PRINT(DBG_INTR, ("%s for ring %d\n", __func__, rx_ring->cq_id));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3037
ql_disable_completion_interrupt(qlge, rx_ring->irq);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3043
ob_ring = rx_ring + qlge->isr_stride;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3051
++qlge->rx_interrupts[rx_ring->cq_id];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3053
mutex_enter(&rx_ring->rx_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3054
mp = ql_ring_rx(rx_ring, QLGE_POLL_ALL);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3055
mutex_exit(&rx_ring->rx_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3058
RX_UPSTREAM(rx_ring, mp);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3069
struct rx_ring *rx_ring = (struct rx_ring *)arg;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3070
qlge_t *qlge = rx_ring->qlge;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3076
__func__, rx_ring->cq_id, n_bytes));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3078
++qlge->rx_polls[rx_ring->cq_id];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3082
mutex_enter(&rx_ring->rx_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3083
mp = ql_ring_rx(rx_ring, n_bytes);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3084
mutex_exit(&rx_ring->rx_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3086
if ((rx_ring->cq_id == 0) && (mp == NULL)) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3126
struct rx_ring *rx_ring = (struct rx_ring *)((void *)arg1);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3127
qlge_t *qlge = rx_ring->qlge;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3131
QL_PRINT(DBG_INTR, ("%s for ring %d\n", __func__, rx_ring->cq_id));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3133
++qlge->rx_interrupts[rx_ring->cq_id];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3135
mutex_enter(&rx_ring->rx_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3136
mp = ql_ring_rx(rx_ring, QLGE_POLL_ALL);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3137
mutex_exit(&rx_ring->rx_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3140
RX_UPSTREAM(rx_ring, mp);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3487
ql_free_rx_resources(struct rx_ring *rx_ring)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3490
if (rx_ring->sbq_dma.dma_handle) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3491
ql_free_phys(&rx_ring->sbq_dma.dma_handle,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3492
&rx_ring->sbq_dma.acc_handle);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3493
bzero(&rx_ring->sbq_dma, sizeof (rx_ring->sbq_dma));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3497
if (rx_ring->sbq_desc != NULL) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3498
kmem_free(rx_ring->sbq_desc, rx_ring->sbq_len *
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3500
rx_ring->sbq_desc = NULL;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3504
if (rx_ring->lbq_dma.dma_handle) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3505
ql_free_phys(&rx_ring->lbq_dma.dma_handle,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3506
&rx_ring->lbq_dma.acc_handle);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3507
bzero(&rx_ring->lbq_dma, sizeof (rx_ring->lbq_dma));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3511
if (rx_ring->lbq_desc != NULL) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3512
kmem_free(rx_ring->lbq_desc, rx_ring->lbq_len *
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3514
rx_ring->lbq_desc = NULL;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3518
if (rx_ring->cqicb_dma.dma_handle) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3519
ql_free_phys(&rx_ring->cqicb_dma.dma_handle,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3520
&rx_ring->cqicb_dma.acc_handle);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3521
bzero(&rx_ring->cqicb_dma, sizeof (rx_ring->cqicb_dma));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3524
if (rx_ring->cq_dma.dma_handle) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3525
ql_free_phys(&rx_ring->cq_dma.dma_handle,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3526
&rx_ring->cq_dma.acc_handle);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3527
bzero(&rx_ring->cq_dma, sizeof (rx_ring->cq_dma));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3536
ql_alloc_rx_resources(qlge_t *qlge, struct rx_ring *rx_ring)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3540
if (ql_alloc_phys(qlge->dip, &rx_ring->cq_dma.dma_handle,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3543
&rx_ring->cq_dma.acc_handle,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3544
(size_t)rx_ring->cq_size, /* mem size */
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3546
(caddr_t *)&rx_ring->cq_dma.vaddr,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3548
bzero(&rx_ring->cq_dma, sizeof (rx_ring->cq_dma));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3553
rx_ring->cq_dma.dma_addr = dma_cookie.dmac_laddress;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3555
if (rx_ring->sbq_len != 0) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3559
if (ql_alloc_phys(qlge->dip, &rx_ring->sbq_dma.dma_handle,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3562
&rx_ring->sbq_dma.acc_handle,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3563
(size_t)rx_ring->sbq_size, /* mem size */
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3565
(caddr_t *)&rx_ring->sbq_dma.vaddr,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3567
bzero(&rx_ring->sbq_dma, sizeof (rx_ring->sbq_dma));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3573
rx_ring->sbq_dma.dma_addr = dma_cookie.dmac_laddress;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3578
rx_ring->sbq_desc =
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3579
kmem_zalloc(rx_ring->sbq_len * sizeof (struct bq_desc),
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3581
if (rx_ring->sbq_desc == NULL) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3587
ql_init_sbq_ring(rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3590
if (rx_ring->lbq_len != 0) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3594
if (ql_alloc_phys(qlge->dip, &rx_ring->lbq_dma.dma_handle,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3597
&rx_ring->lbq_dma.acc_handle,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3598
(size_t)rx_ring->lbq_size, /* mem size */
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3600
(caddr_t *)&rx_ring->lbq_dma.vaddr,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3602
bzero(&rx_ring->lbq_dma, sizeof (rx_ring->lbq_dma));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3607
rx_ring->lbq_dma.dma_addr = dma_cookie.dmac_laddress;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3612
rx_ring->lbq_desc =
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3613
kmem_zalloc(rx_ring->lbq_len * sizeof (struct bq_desc),
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3615
if (rx_ring->lbq_desc == NULL) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3621
ql_init_lbq_ring(rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3624
if (ql_alloc_phys(qlge->dip, &rx_ring->cqicb_dma.dma_handle,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3627
&rx_ring->cqicb_dma.acc_handle,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3630
(caddr_t *)&rx_ring->cqicb_dma.vaddr,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3632
bzero(&rx_ring->cqicb_dma, sizeof (rx_ring->cqicb_dma));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3637
rx_ring->cqicb_dma.dma_addr = dma_cookie.dmac_laddress;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3642
ql_free_rx_resources(rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3669
ql_free_rx_resources(&qlge->rx_ring[i]);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3692
if (ql_alloc_rx_resources(qlge, &qlge->rx_ring[i]) != 0) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3930
(void *)&qlge->rx_ring[vector], NULL);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3933
vector, &qlge->rx_ring[vector]));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3956
(caddr_t)&qlge->rx_ring[0], NULL);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
3973
(caddr_t)&qlge->rx_ring[0], NULL);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
4023
qlge->rx_ring[i].irq = i;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
4050
} else if (qlge->rx_ring[i].type == TX_Q) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
4072
qlge->rx_ring[i].irq = iv;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
4093
if (qlge->rx_ring[i].type == TX_Q) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
4131
qlge->rx_ring[i].irq = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
4443
struct rx_ring *rx_ring;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
4452
rx_ring = &qlge->rx_ring[i];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
4453
mutex_destroy(&rx_ring->rx_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
4454
mutex_destroy(&rx_ring->sbq_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
4455
mutex_destroy(&rx_ring->lbq_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5674
struct rx_ring *rx_ring;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5687
if (qlge->rx_ring[0].lbq_buf_size != lbq_buf_size) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5690
qlge->rx_ring[0].lbq_buf_size, lbq_buf_size);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5703
rx_ring = &qlge->rx_ring[i];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5704
if (rx_ring->rx_indicate == 0)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5716
rx_ring = &qlge->rx_ring[i];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5717
if (rx_ring->type != TX_Q) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5718
ql_free_sbq_buffers(rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5719
ql_free_lbq_buffers(rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5724
rx_ring = &qlge->rx_ring[i];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5725
if (rx_ring->type != TX_Q) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5726
(void) ql_alloc_sbufs(qlge, rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5727
(void) ql_alloc_lbufs(qlge, rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5781
struct rx_ring *rx_ring;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5794
rx_ring = &qlge->rx_ring[k];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5795
if (rx_ring->type != TX_Q) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5796
j = rx_ring->lbq_use_head;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5800
k, rx_ring->lbuf_in_use_count,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5801
rx_ring->lbuf_free_count,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5802
rx_ring->lbuf_in_use_count +
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5803
rx_ring->lbuf_free_count);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5805
for (i = 0; i < rx_ring->lbuf_in_use_count; i++) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5806
lbq_desc = rx_ring->lbuf_in_use[j];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5808
if (j >= rx_ring->lbq_len) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5812
atomic_inc_32(&rx_ring->rx_indicate);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5816
rx_ring->lbq_use_head = j;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5817
rx_ring->lbq_use_tail = j;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5818
rx_ring->lbuf_in_use_count = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5819
j = rx_ring->sbq_use_head;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5823
k, rx_ring->sbuf_in_use_count,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5824
rx_ring->sbuf_free_count,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5825
rx_ring->sbuf_in_use_count +
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5826
rx_ring->sbuf_free_count);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5828
for (i = 0; i < rx_ring->sbuf_in_use_count; i++) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5829
sbq_desc = rx_ring->sbuf_in_use[j];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5831
if (j >= rx_ring->sbq_len) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5835
atomic_inc_32(&rx_ring->rx_indicate);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5839
rx_ring->sbq_use_head = j;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5840
rx_ring->sbq_use_tail = j;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
5841
rx_ring->sbuf_in_use_count = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6130
struct rx_ring *rx_ring;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6184
rx_ring = &qlge->rx_ring[i];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6185
val32 += rx_ring->rx_packets_dropped_no_buffer;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6287
struct rx_ring *rx_ring;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6327
rx_ring = &qlge->rx_ring[i];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6328
bzero((void *)rx_ring, sizeof (*rx_ring));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6329
rx_ring->qlge = qlge;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6330
rx_ring->cq_id = i;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6332
rx_ring->cpu = (i) % qlge->rx_ring_count;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6334
rx_ring->cpu = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6343
rx_ring->cq_len = qlge->rx_ring_size;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6344
rx_ring->cq_size = (uint32_t)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6345
(rx_ring->cq_len * sizeof (struct net_rsp_iocb));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6346
rx_ring->lbq_len = NUM_LARGE_BUFFERS;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6347
rx_ring->lbq_size = (uint32_t)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6348
(rx_ring->lbq_len * sizeof (uint64_t));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6349
rx_ring->lbq_buf_size = lbq_buf_size;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6350
rx_ring->sbq_len = NUM_SMALL_BUFFERS;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6351
rx_ring->sbq_size = (uint32_t)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6352
(rx_ring->sbq_len * sizeof (uint64_t));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6353
rx_ring->sbq_buf_size = SMALL_BUFFER_SIZE * 2;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6354
rx_ring->type = RX_Q;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6359
rx_ring->cq_id, rx_ring->cpu));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6365
QL_PRINT(DBG_INIT, ("rx_ring 0x%p i %d\n", rx_ring, i));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6366
rx_ring->cq_len = qlge->tx_ring_size;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6367
rx_ring->cq_size = (uint32_t)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6368
(rx_ring->cq_len * sizeof (struct net_rsp_iocb));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6369
rx_ring->lbq_len = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6370
rx_ring->lbq_size = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6371
rx_ring->lbq_buf_size = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6372
rx_ring->sbq_len = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6373
rx_ring->sbq_size = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6374
rx_ring->sbq_buf_size = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6375
rx_ring->type = TX_Q;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6380
rx_ring->cq_id, rx_ring->cpu));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6388
ql_start_rx_ring(qlge_t *qlge, struct rx_ring *rx_ring)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6390
struct cqicb_t *cqicb = (struct cqicb_t *)rx_ring->cqicb_dma.vaddr;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6392
(rx_ring->cq_id * sizeof (uint64_t) * RX_TX_RING_SHADOW_SPACE)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6396
(rx_ring->cq_id * sizeof (uint64_t) * RX_TX_RING_SHADOW_SPACE)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6401
(rx_ring->cq_id * sizeof (uint64_t) * BUF_Q_PTR_SPACE);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6404
(rx_ring->cq_id * sizeof (uint64_t) * BUF_Q_PTR_SPACE);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6406
qlge->doorbell_reg_iobase + (VM_PAGE_SIZE * (128 + rx_ring->cq_id));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6414
rx_ring->prod_idx_sh_reg = shadow_reg;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6415
rx_ring->prod_idx_sh_reg_dma = shadow_reg_dma;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6416
rx_ring->prod_idx_sh_reg_offset = (off_t)(((rx_ring->cq_id *
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6419
rx_ring->lbq_base_indirect = (uint64_t *)(void *)buf_q_base_reg;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6420
rx_ring->lbq_base_indirect_dma = buf_q_base_reg_dma;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6423
" phys_addr 0x%lx\n", __func__, rx_ring->cq_id,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6424
rx_ring->prod_idx_sh_reg, rx_ring->prod_idx_sh_reg_dma));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6428
rx_ring->sbq_base_indirect = (uint64_t *)(void *)buf_q_base_reg;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6429
rx_ring->sbq_base_indirect_dma = buf_q_base_reg_dma;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6432
rx_ring->cnsmr_idx_db_reg = (uint32_t *)(void *)doorbell_area;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6433
rx_ring->cnsmr_idx = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6434
*rx_ring->prod_idx_sh_reg = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6435
rx_ring->curr_entry = rx_ring->cq_dma.vaddr;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6438
rx_ring->valid_db_reg = (uint32_t *)(void *)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6442
rx_ring->lbq_prod_idx_db_reg = (uint32_t *)(void *)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6446
rx_ring->sbq_prod_idx_db_reg = (uint32_t *)(void *)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6451
cqicb->msix_vect = (uint8_t)rx_ring->irq;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6453
bq_len = (uint16_t)((rx_ring->cq_len == 65536) ?
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6454
(uint16_t)0 : (uint16_t)rx_ring->cq_len);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6458
cpu_to_le32(LS_64BITS(rx_ring->cq_dma.dma_addr));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6460
cpu_to_le32(MS_64BITS(rx_ring->cq_dma.dma_addr));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6463
cpu_to_le32(LS_64BITS(rx_ring->prod_idx_sh_reg_dma));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6465
cpu_to_le32(MS_64BITS(rx_ring->prod_idx_sh_reg_dma));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6473
if (rx_ring->lbq_len) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6476
tmp = (uint64_t)rx_ring->lbq_dma.dma_addr;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6477
base_indirect_ptr = (uint64_t *)rx_ring->lbq_base_indirect;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6485
((rx_ring->lbq_len * sizeof (uint64_t)) / VM_PAGE_SIZE)));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6488
cpu_to_le32(LS_64BITS(rx_ring->lbq_base_indirect_dma));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6490
cpu_to_le32(MS_64BITS(rx_ring->lbq_base_indirect_dma));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6491
bq_len = (uint16_t)((rx_ring->lbq_buf_size == 65536) ?
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6492
(uint16_t)0 : (uint16_t)rx_ring->lbq_buf_size);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6494
bq_len = (uint16_t)((rx_ring->lbq_len == 65536) ? (uint16_t)0 :
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6495
(uint16_t)rx_ring->lbq_len);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6497
rx_ring->lbq_prod_idx = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6498
rx_ring->lbq_curr_idx = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6500
if (rx_ring->sbq_len) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6503
tmp = (uint64_t)rx_ring->sbq_dma.dma_addr;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6504
base_indirect_ptr = (uint64_t *)rx_ring->sbq_base_indirect;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6513
(((rx_ring->sbq_len * sizeof (uint64_t)) / VM_PAGE_SIZE)));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6516
cpu_to_le32(LS_64BITS(rx_ring->sbq_base_indirect_dma));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6518
cpu_to_le32(MS_64BITS(rx_ring->sbq_base_indirect_dma));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6520
cpu_to_le16((uint16_t)(rx_ring->sbq_buf_size/2));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6521
bq_len = (uint16_t)((rx_ring->sbq_len == 65536) ?
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6522
(uint16_t)0 : (uint16_t)rx_ring->sbq_len);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6524
rx_ring->sbq_prod_idx = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6525
rx_ring->sbq_curr_idx = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6527
switch (rx_ring->type) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6554
rx_ring->type);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6557
rx_ring->cq_id));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6559
err = ql_write_cfg(qlge, CFG_LCQ, rx_ring->cqicb_dma.dma_addr,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6560
rx_ring->cq_id);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6566
rx_ring->rx_packets_dropped_no_buffer = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6567
rx_ring->rx_pkt_dropped_mac_unenabled = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6568
rx_ring->rx_failed_sbq_allocs = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6569
rx_ring->rx_failed_lbq_allocs = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6570
rx_ring->rx_packets = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6571
rx_ring->rx_bytes = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6572
rx_ring->frame_too_long = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6573
rx_ring->frame_too_short = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
6574
rx_ring->fcs_err = 0;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7048
status = ql_start_rx_ring(qlge, &qlge->rx_ring[i]);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7115
ql_arm_sbuf(qlge_t *qlge, struct rx_ring *rx_ring)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7119
uint64_t *sbq_entry = rx_ring->sbq_dma.vaddr;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7122
if (rx_ring->sbuf_free_count > rx_ring->sbq_len-MIN_BUFFERS_ARM_COUNT)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7123
arm_count = (rx_ring->sbq_len-MIN_BUFFERS_ARM_COUNT);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7126
arm_count = (rx_ring->sbuf_free_count / 16) * 16;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7132
sbq_desc = ql_get_sbuf_from_free_list(rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7140
ql_add_sbuf_to_in_use_list(rx_ring, sbq_desc);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7141
rx_ring->sbq_prod_idx++;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7143
ql_update_sbq_prod_idx(qlge, rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7152
ql_arm_lbuf(qlge_t *qlge, struct rx_ring *rx_ring)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7156
uint64_t *lbq_entry = rx_ring->lbq_dma.vaddr;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7159
if (rx_ring->lbuf_free_count > rx_ring->lbq_len-MIN_BUFFERS_ARM_COUNT)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7160
arm_count = (rx_ring->lbq_len-MIN_BUFFERS_ARM_COUNT);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7163
arm_count = (rx_ring->lbuf_free_count / 16) * 16;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7169
lbq_desc = ql_get_lbuf_from_free_list(rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7177
ql_add_lbuf_to_in_use_list(rx_ring, lbq_desc);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7178
rx_ring->lbq_prod_idx++;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7180
ql_update_lbq_prod_idx(qlge, rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7203
if (qlge->rx_ring[i].type != TX_Q) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7212
if (qlge->rx_ring[i].type != TX_Q) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7213
ql_arm_sbuf(qlge, &qlge->rx_ring[i]);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7214
ql_arm_lbuf(qlge, &qlge->rx_ring[i]);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7228
if (qlge->rx_ring[i].valid_db_reg)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7230
qlge->rx_ring[i].valid_db_reg,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7241
mutex_enter(&qlge->rx_ring[i].rx_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7242
qlge->rx_ring[i].mac_flags = QL_MAC_STARTED;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7243
mutex_exit(&qlge->rx_ring[i].rx_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7269
struct rx_ring *rx_ring;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7279
rx_ring = &qlge->rx_ring[i];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7280
mutex_init(&rx_ring->rx_lock, NULL, MUTEX_DRIVER,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7282
mutex_init(&rx_ring->sbq_lock, NULL, MUTEX_DRIVER,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7284
mutex_init(&rx_ring->lbq_lock, NULL, MUTEX_DRIVER,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
73
int ql_clean_outbound_rx_ring(struct rx_ring *rx_ring);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7734
struct rx_ring *ob_ring;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7735
ob_ring = &qlge->rx_ring[tx_ring->cq_id];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7777
volatile struct rx_ring *rx_ring;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7785
rx_ring = &qlge->rx_ring[i+1];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7787
rx_ring->cnsmr_idx_db_reg);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7832
if (qlge->rx_ring[i].valid_db_reg) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7834
qlge->rx_ring[i].valid_db_reg, 0);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7872
mutex_enter(&qlge->rx_ring[i].rx_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7873
qlge->rx_ring[i].mac_flags = QL_MAC_STOPPED;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
7874
mutex_exit(&qlge->rx_ring[i].rx_lock);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
8006
if (qlge->rx_ring[i].valid_db_reg) {
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
8008
qlge->rx_ring[i].valid_db_reg, 0);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
928
ql_update_sbq_prod_idx(qlge_t *qlge, struct rx_ring *rx_ring)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
932
rx_ring->sbq_prod_idx));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
933
ql_write_doorbell_reg(qlge, rx_ring->sbq_prod_idx_db_reg,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
934
rx_ring->sbq_prod_idx);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
941
ql_update_lbq_prod_idx(qlge_t *qlge, struct rx_ring *rx_ring)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
945
rx_ring->lbq_prod_idx));
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
946
ql_write_doorbell_reg(qlge, rx_ring->lbq_prod_idx_db_reg,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
947
rx_ring->lbq_prod_idx);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
955
ql_add_sbuf_to_in_use_list(struct rx_ring *rx_ring,
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
958
uint32_t inuse_idx = rx_ring->sbq_use_tail;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
960
rx_ring->sbuf_in_use[inuse_idx] = sbq_desc;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
962
if (inuse_idx >= rx_ring->sbq_len)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
964
rx_ring->sbq_use_tail = inuse_idx;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
965
atomic_inc_32(&rx_ring->sbuf_in_use_count);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
966
ASSERT(rx_ring->sbuf_in_use_count <= rx_ring->sbq_len);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
973
ql_get_sbuf_from_in_use_list(struct rx_ring *rx_ring)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
979
inuse_idx = rx_ring->sbq_use_head;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
980
sbq_desc = rx_ring->sbuf_in_use[inuse_idx];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
981
rx_ring->sbuf_in_use[inuse_idx] = NULL;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
985
if (inuse_idx >= rx_ring->sbq_len)
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
987
rx_ring->sbq_use_head = inuse_idx;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
988
atomic_dec_32(&rx_ring->sbuf_in_use_count);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
989
atomic_inc_32(&rx_ring->rx_indicate);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge.c
996
rx_ring->sbq_buf_size, 0, &sbq_desc->rx_recycle);
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge_gld.c
222
struct rx_ring *rx_ring;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge_gld.c
316
rx_ring = &qlge->rx_ring[i];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge_gld.c
317
cur_stats->rbytes += rx_ring->rx_bytes;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge_gld.c
325
rx_ring = &qlge->rx_ring[i];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge_gld.c
326
cur_stats->rpackets += rx_ring->rx_packets;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge_gld.c
334
rx_ring = &qlge->rx_ring[i];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge_gld.c
335
cur_stats->crc += rx_ring->fcs_err;
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge_gld.c
343
rx_ring = &qlge->rx_ring[i];
usr/src/uts/common/io/fibre-channel/fca/qlge/qlge_gld.c
345
rx_ring->frame_too_long;
usr/src/uts/common/io/igb/igb_buf.c
127
igb_rx_ring_t *rx_ring;
usr/src/uts/common/io/igb/igb_buf.c
136
rx_ring = &igb->rx_rings[i];
usr/src/uts/common/io/igb/igb_buf.c
137
rx_data = rx_ring->rx_data;
usr/src/uts/common/io/igb/igb_buf.c
174
igb_rx_ring_t *rx_ring;
usr/src/uts/common/io/igb/igb_buf.c
183
rx_ring = &igb->rx_rings[i];
usr/src/uts/common/io/igb/igb_buf.c
184
rx_data = rx_ring->rx_data;
usr/src/uts/common/io/igb/igb_buf.c
330
igb_alloc_rx_ring_data(igb_rx_ring_t *rx_ring)
usr/src/uts/common/io/igb/igb_buf.c
333
igb_t *igb = rx_ring->igb;
usr/src/uts/common/io/igb/igb_buf.c
347
rx_data->rx_ring = rx_ring;
usr/src/uts/common/io/igb/igb_buf.c
397
rx_ring->rx_data = rx_data;
usr/src/uts/common/io/igb/igb_buf.c
450
igb_t *igb = rx_data->rx_ring->igb;
usr/src/uts/common/io/igb/igb_buf.c
834
igb_t *igb = rx_data->rx_ring->igb;
usr/src/uts/common/io/igb/igb_buf.c
901
igb = rx_data->rx_ring->igb;
usr/src/uts/common/io/igb/igb_gld.c
696
igb_rx_ring_t *rx_ring = (igb_rx_ring_t *)intrh;
usr/src/uts/common/io/igb/igb_gld.c
697
igb_t *igb = rx_ring->igb;
usr/src/uts/common/io/igb/igb_gld.c
699
uint32_t index = rx_ring->index;
usr/src/uts/common/io/igb/igb_gld.c
724
igb_rx_ring_t *rx_ring = (igb_rx_ring_t *)intrh;
usr/src/uts/common/io/igb/igb_gld.c
725
igb_t *igb = rx_ring->igb;
usr/src/uts/common/io/igb/igb_gld.c
727
uint32_t index = rx_ring->index;
usr/src/uts/common/io/igb/igb_gld.c
753
igb_rx_ring_t *rx_ring;
usr/src/uts/common/io/igb/igb_gld.c
757
rx_ring = &igb->rx_rings[i];
usr/src/uts/common/io/igb/igb_gld.c
758
if (rx_ring->group_index == gindex)
usr/src/uts/common/io/igb/igb_gld.c
770
igb_rx_ring_t *rx_ring = (igb_rx_ring_t *)rh;
usr/src/uts/common/io/igb/igb_gld.c
772
mutex_enter(&rx_ring->rx_lock);
usr/src/uts/common/io/igb/igb_gld.c
773
rx_ring->ring_gen_num = mr_gen_num;
usr/src/uts/common/io/igb/igb_gld.c
774
mutex_exit(&rx_ring->rx_lock);
usr/src/uts/common/io/igb/igb_gld.c
791
igb_rx_ring_t *rx_ring;
usr/src/uts/common/io/igb/igb_gld.c
802
rx_ring = &igb->rx_rings[global_index];
usr/src/uts/common/io/igb/igb_gld.c
803
rx_ring->ring_handle = rh;
usr/src/uts/common/io/igb/igb_gld.c
805
infop->mri_driver = (mac_ring_driver_t)rx_ring;
usr/src/uts/common/io/igb/igb_gld.c
811
mintr->mi_handle = (mac_intr_handle_t)rx_ring;
usr/src/uts/common/io/igb/igb_gld.c
816
igb->htable[rx_ring->intr_vector];
usr/src/uts/common/io/igb/igb_main.c
1043
igb_rx_ring_t *rx_ring;
usr/src/uts/common/io/igb/igb_main.c
1089
rx_ring = &igb->rx_rings[i];
usr/src/uts/common/io/igb/igb_main.c
1090
rx_ring->index = i;
usr/src/uts/common/io/igb/igb_main.c
1091
rx_ring->igb = igb;
usr/src/uts/common/io/igb/igb_main.c
1128
igb_rx_ring_t *rx_ring;
usr/src/uts/common/io/igb/igb_main.c
1133
rx_ring = &igb->rx_rings[i];
usr/src/uts/common/io/igb/igb_main.c
1134
mutex_init(&rx_ring->rx_lock, NULL,
usr/src/uts/common/io/igb/igb_main.c
1166
igb_rx_ring_t *rx_ring;
usr/src/uts/common/io/igb/igb_main.c
1171
rx_ring = &igb->rx_rings[i];
usr/src/uts/common/io/igb/igb_main.c
1172
mutex_destroy(&rx_ring->rx_lock);
usr/src/uts/common/io/igb/igb_main.c
2077
igb_rx_ring_t *rx_ring;
usr/src/uts/common/io/igb/igb_main.c
2081
rx_ring = &igb->rx_rings[i];
usr/src/uts/common/io/igb/igb_main.c
2082
if (igb_alloc_rx_ring_data(rx_ring) != IGB_SUCCESS)
usr/src/uts/common/io/igb/igb_main.c
2095
igb_rx_ring_t *rx_ring;
usr/src/uts/common/io/igb/igb_main.c
2100
rx_ring = &igb->rx_rings[i];
usr/src/uts/common/io/igb/igb_main.c
2103
rx_data = rx_ring->rx_data;
usr/src/uts/common/io/igb/igb_main.c
2110
rx_ring->rx_data = NULL;
usr/src/uts/common/io/igb/igb_main.c
2137
igb_setup_rx_ring(igb_rx_ring_t *rx_ring)
usr/src/uts/common/io/igb/igb_main.c
2139
igb_t *igb = rx_ring->igb;
usr/src/uts/common/io/igb/igb_main.c
2140
igb_rx_data_t *rx_data = rx_ring->rx_data;
usr/src/uts/common/io/igb/igb_main.c
2150
ASSERT(mutex_owned(&rx_ring->rx_lock));
usr/src/uts/common/io/igb/igb_main.c
2169
E1000_WRITE_REG(hw, E1000_RDBAH(rx_ring->index), buf_high);
usr/src/uts/common/io/igb/igb_main.c
2170
E1000_WRITE_REG(hw, E1000_RDBAL(rx_ring->index), buf_low);
usr/src/uts/common/io/igb/igb_main.c
2176
E1000_WRITE_REG(hw, E1000_RDLEN(rx_ring->index), size);
usr/src/uts/common/io/igb/igb_main.c
2181
E1000_WRITE_REG(hw, E1000_SRRCTL(rx_ring->index),
usr/src/uts/common/io/igb/igb_main.c
2188
rxdctl = E1000_READ_REG(hw, E1000_RXDCTL(rx_ring->index));
usr/src/uts/common/io/igb/igb_main.c
2194
E1000_WRITE_REG(hw, E1000_RXDCTL(rx_ring->index), rxdctl);
usr/src/uts/common/io/igb/igb_main.c
2202
igb_rx_ring_t *rx_ring;
usr/src/uts/common/io/igb/igb_main.c
2247
rx_ring = &igb->rx_rings[i];
usr/src/uts/common/io/igb/igb_main.c
2248
igb_setup_rx_ring(rx_ring);
usr/src/uts/common/io/igb/igb_main.c
2253
rx_ring->group_index = i / ring_per_group;
usr/src/uts/common/io/igb/igb_main.c
2311
rx_ring = &igb->rx_rings[i];
usr/src/uts/common/io/igb/igb_main.c
2312
rx_data = rx_ring->rx_data;
usr/src/uts/common/io/igb/igb_main.c
4055
igb_intr_rx_work(igb_rx_ring_t *rx_ring)
usr/src/uts/common/io/igb/igb_main.c
4059
mutex_enter(&rx_ring->rx_lock);
usr/src/uts/common/io/igb/igb_main.c
4060
mp = igb_rx(rx_ring, IGB_NO_POLL);
usr/src/uts/common/io/igb/igb_main.c
4061
mutex_exit(&rx_ring->rx_lock);
usr/src/uts/common/io/igb/igb_main.c
4064
mac_rx_ring(rx_ring->igb->mac_hdl, rx_ring->ring_handle, mp,
usr/src/uts/common/io/igb/igb_main.c
4065
rx_ring->ring_gen_num);
usr/src/uts/common/io/igb/igb_main.c
4285
igb_rx_ring_t *rx_ring = (igb_rx_ring_t *)arg1;
usr/src/uts/common/io/igb/igb_main.c
4293
igb_intr_rx_work(rx_ring);
usr/src/uts/common/io/igb/igb_main.c
4617
igb_rx_ring_t *rx_ring;
usr/src/uts/common/io/igb/igb_main.c
4643
rx_ring = &igb->rx_rings[i];
usr/src/uts/common/io/igb/igb_main.c
4647
(void *)rx_ring, NULL);
usr/src/uts/common/io/igb/igb_main.c
4660
rx_ring->intr_vector = vector;
usr/src/uts/common/io/igb/igb_main.c
4703
rx_ring = &igb->rx_rings[0];
usr/src/uts/common/io/igb/igb_main.c
4704
rx_ring->intr_vector = vector;
usr/src/uts/common/io/igb/igb_main.c
4721
rx_ring = &igb->rx_rings[0];
usr/src/uts/common/io/igb/igb_main.c
4722
rx_ring->intr_vector = vector;
usr/src/uts/common/io/igb/igb_rx.c
137
igb_t *igb = rx_data->rx_ring->igb;
usr/src/uts/common/io/igb/igb_rx.c
183
igb_t *igb = rx_data->rx_ring->igb;
usr/src/uts/common/io/igb/igb_rx.c
287
igb_rx_ring_t *rx_ring = (igb_rx_ring_t *)arg;
usr/src/uts/common/io/igb/igb_rx.c
292
if ((bytes == 0) || (rx_ring->igb->igb_state & IGB_SUSPENDED) ||
usr/src/uts/common/io/igb/igb_rx.c
293
!(rx_ring->igb->igb_state & IGB_STARTED))
usr/src/uts/common/io/igb/igb_rx.c
296
mutex_enter(&rx_ring->rx_lock);
usr/src/uts/common/io/igb/igb_rx.c
297
mp = igb_rx(rx_ring, bytes);
usr/src/uts/common/io/igb/igb_rx.c
298
mutex_exit(&rx_ring->rx_lock);
usr/src/uts/common/io/igb/igb_rx.c
312
igb_rx(igb_rx_ring_t *rx_ring, int poll_bytes)
usr/src/uts/common/io/igb/igb_rx.c
325
igb_t *igb = rx_ring->igb;
usr/src/uts/common/io/igb/igb_rx.c
326
igb_rx_data_t *rx_data = rx_ring->rx_data;
usr/src/uts/common/io/igb/igb_rx.c
366
IGB_DEBUG_STAT(rx_ring->stat_frame_error);
usr/src/uts/common/io/igb/igb_rx.c
370
IGB_DEBUG_STAT_COND(rx_ring->stat_cksum_error,
usr/src/uts/common/io/igb/igb_rx.c
380
IGB_DEBUG_STAT(rx_ring->stat_pkt_cnt);
usr/src/uts/common/io/igb/igb_rx.c
411
rx_ring->rx_pkts++;
usr/src/uts/common/io/igb/igb_rx.c
412
rx_ring->rx_bytes += pkt_len;
usr/src/uts/common/io/igb/igb_rx.c
430
IGB_DEBUG_STAT(rx_ring->stat_exceed_pkt);
usr/src/uts/common/io/igb/igb_rx.c
447
E1000_WRITE_REG(&igb->hw, E1000_RDT(rx_ring->index), rx_tail);
usr/src/uts/common/io/igb/igb_rx.c
46
igb_rx_ring_t *rx_ring;
usr/src/uts/common/io/igb/igb_rx.c
54
rx_ring = rx_data->rx_ring;
usr/src/uts/common/io/igb/igb_rx.c
55
igb = rx_ring->igb;
usr/src/uts/common/io/igb/igb_stat.c
279
igb_rx_ring_t *rx_ring = (igb_rx_ring_t *)rh;
usr/src/uts/common/io/igb/igb_stat.c
283
*val = rx_ring->rx_bytes;
usr/src/uts/common/io/igb/igb_stat.c
287
*val = rx_ring->rx_pkts;
usr/src/uts/common/io/igb/igb_sw.h
520
struct igb_rx_ring *rx_ring; /* Pointer to rx ring */
usr/src/uts/common/io/igb/igb_sw.h
845
int igb_alloc_rx_ring_data(igb_rx_ring_t *rx_ring);
usr/src/uts/common/io/igc/igc_gld.c
607
igc_rx_ring_t *rx_ring = (igc_rx_ring_t *)rh;
usr/src/uts/common/io/igc/igc_gld.c
609
mutex_enter(&rx_ring->irr_lock);
usr/src/uts/common/io/igc/igc_gld.c
610
rx_ring->irr_gen = gen;
usr/src/uts/common/io/igc/igc_gld.c
611
mutex_exit(&rx_ring->irr_lock);
usr/src/uts/common/io/igc/igc_gld.c
637
igc_rx_ring_t *rx_ring = (igc_rx_ring_t *)rh;
usr/src/uts/common/io/igc/igc_gld.c
641
*val = rx_ring->irr_stat.irs_rbytes.value.ui64;
usr/src/uts/common/io/igc/igc_gld.c
644
*val = rx_ring->irr_stat.irs_ipackets.value.ui64;
usr/src/uts/common/io/ixgbe/ixgbe_buf.c
125
ixgbe_rx_ring_t *rx_ring;
usr/src/uts/common/io/ixgbe/ixgbe_buf.c
134
rx_ring = &ixgbe->rx_rings[i];
usr/src/uts/common/io/ixgbe/ixgbe_buf.c
135
rx_data = rx_ring->rx_data;
usr/src/uts/common/io/ixgbe/ixgbe_buf.c
171
ixgbe_rx_ring_t *rx_ring;
usr/src/uts/common/io/ixgbe/ixgbe_buf.c
180
rx_ring = &ixgbe->rx_rings[i];
usr/src/uts/common/io/ixgbe/ixgbe_buf.c
181
rx_data = rx_ring->rx_data;
usr/src/uts/common/io/ixgbe/ixgbe_buf.c
198
ixgbe_alloc_rx_ring_data(ixgbe_rx_ring_t *rx_ring)
usr/src/uts/common/io/ixgbe/ixgbe_buf.c
201
ixgbe_t *ixgbe = rx_ring->ixgbe;
usr/src/uts/common/io/ixgbe/ixgbe_buf.c
214
rx_data->rx_ring = rx_ring;
usr/src/uts/common/io/ixgbe/ixgbe_buf.c
264
rx_ring->rx_data = rx_data;
usr/src/uts/common/io/ixgbe/ixgbe_buf.c
446
ixgbe_t *ixgbe = rx_data->rx_ring->ixgbe;
usr/src/uts/common/io/ixgbe/ixgbe_buf.c
828
ixgbe_t *ixgbe = rx_data->rx_ring->ixgbe;
usr/src/uts/common/io/ixgbe/ixgbe_buf.c
896
ixgbe = rx_data->rx_ring->ixgbe;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
1189
ixgbe_rx_ring_t *rx_ring;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
1240
rx_ring = &ixgbe->rx_rings[i];
usr/src/uts/common/io/ixgbe/ixgbe_main.c
1241
rx_ring->index = i;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
1242
rx_ring->ixgbe = ixgbe;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
1243
rx_ring->group_index = i / ring_per_group;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
1244
rx_ring->hw_index = ixgbe_get_hw_rx_index(ixgbe, i);
usr/src/uts/common/io/ixgbe/ixgbe_main.c
1289
ixgbe_rx_ring_t *rx_ring;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
1294
rx_ring = &ixgbe->rx_rings[i];
usr/src/uts/common/io/ixgbe/ixgbe_main.c
1295
mutex_init(&rx_ring->rx_lock, NULL,
usr/src/uts/common/io/ixgbe/ixgbe_main.c
1324
ixgbe_rx_ring_t *rx_ring;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
1329
rx_ring = &ixgbe->rx_rings[i];
usr/src/uts/common/io/ixgbe/ixgbe_main.c
1330
mutex_destroy(&rx_ring->rx_lock);
usr/src/uts/common/io/ixgbe/ixgbe_main.c
2562
ixgbe_rx_ring_t *rx_ring;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
2566
rx_ring = &ixgbe->rx_rings[i];
usr/src/uts/common/io/ixgbe/ixgbe_main.c
2567
if (ixgbe_alloc_rx_ring_data(rx_ring) != IXGBE_SUCCESS)
usr/src/uts/common/io/ixgbe/ixgbe_main.c
2580
ixgbe_rx_ring_t *rx_ring;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
2585
rx_ring = &ixgbe->rx_rings[i];
usr/src/uts/common/io/ixgbe/ixgbe_main.c
2588
rx_data = rx_ring->rx_data;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
2595
rx_ring->rx_data = NULL;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
2625
ixgbe_setup_rx_ring(ixgbe_rx_ring_t *rx_ring)
usr/src/uts/common/io/ixgbe/ixgbe_main.c
2627
ixgbe_t *ixgbe = rx_ring->ixgbe;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
2628
ixgbe_rx_data_t *rx_data = rx_ring->rx_data;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
2638
ASSERT(mutex_owned(&rx_ring->rx_lock));
usr/src/uts/common/io/ixgbe/ixgbe_main.c
2653
IXGBE_WRITE_REG(hw, IXGBE_RDLEN(rx_ring->hw_index), size);
usr/src/uts/common/io/ixgbe/ixgbe_main.c
2660
IXGBE_WRITE_REG(hw, IXGBE_RDBAH(rx_ring->hw_index), buf_high);
usr/src/uts/common/io/ixgbe/ixgbe_main.c
2661
IXGBE_WRITE_REG(hw, IXGBE_RDBAL(rx_ring->hw_index), buf_low);
usr/src/uts/common/io/ixgbe/ixgbe_main.c
2666
IXGBE_WRITE_REG(hw, IXGBE_RDT(rx_ring->hw_index),
usr/src/uts/common/io/ixgbe/ixgbe_main.c
2668
IXGBE_WRITE_REG(hw, IXGBE_RDH(rx_ring->hw_index), 0);
usr/src/uts/common/io/ixgbe/ixgbe_main.c
2679
reg_val = IXGBE_READ_REG(hw, IXGBE_RXDCTL(rx_ring->hw_index));
usr/src/uts/common/io/ixgbe/ixgbe_main.c
2686
IXGBE_WRITE_REG(hw, IXGBE_RXDCTL(rx_ring->hw_index), reg_val);
usr/src/uts/common/io/ixgbe/ixgbe_main.c
2701
IXGBE_WRITE_REG(hw, IXGBE_SRRCTL(rx_ring->hw_index), reg_val);
usr/src/uts/common/io/ixgbe/ixgbe_main.c
2707
ixgbe_rx_ring_t *rx_ring;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
2820
rx_ring = &ixgbe->rx_rings[i];
usr/src/uts/common/io/ixgbe/ixgbe_main.c
2821
ixgbe_setup_rx_ring(rx_ring);
usr/src/uts/common/io/ixgbe/ixgbe_main.c
4848
ixgbe_intr_rx_work(ixgbe_rx_ring_t *rx_ring)
usr/src/uts/common/io/ixgbe/ixgbe_main.c
4852
mutex_enter(&rx_ring->rx_lock);
usr/src/uts/common/io/ixgbe/ixgbe_main.c
4854
mp = ixgbe_ring_rx(rx_ring, IXGBE_POLL_NULL);
usr/src/uts/common/io/ixgbe/ixgbe_main.c
4855
mutex_exit(&rx_ring->rx_lock);
usr/src/uts/common/io/ixgbe/ixgbe_main.c
4858
mac_rx_ring(rx_ring->ixgbe->mac_hdl, rx_ring->ring_handle, mp,
usr/src/uts/common/io/ixgbe/ixgbe_main.c
4859
rx_ring->ring_gen_num);
usr/src/uts/common/io/ixgbe/ixgbe_main.c
5146
ixgbe_rx_ring_t *rx_ring;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
5193
rx_ring = &ixgbe->rx_rings[0];
usr/src/uts/common/io/ixgbe/ixgbe_main.c
5194
mp = ixgbe_ring_rx(rx_ring, IXGBE_POLL_NULL);
usr/src/uts/common/io/ixgbe/ixgbe_main.c
5257
mac_rx_ring(rx_ring->ixgbe->mac_hdl, rx_ring->ring_handle, mp,
usr/src/uts/common/io/ixgbe/ixgbe_main.c
5258
rx_ring->ring_gen_num);
usr/src/uts/common/io/ixgbe/ixgbe_main.c
6552
ixgbe_rx_ring_t *rx_ring = (ixgbe_rx_ring_t *)rh;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
6554
mutex_enter(&rx_ring->rx_lock);
usr/src/uts/common/io/ixgbe/ixgbe_main.c
6555
rx_ring->ring_gen_num = mr_gen_num;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
6556
mutex_exit(&rx_ring->rx_lock);
usr/src/uts/common/io/ixgbe/ixgbe_main.c
6566
ixgbe_rx_ring_t *rx_ring;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
6570
rx_ring = &ixgbe->rx_rings[i];
usr/src/uts/common/io/ixgbe/ixgbe_main.c
6571
if (rx_ring->group_index == gindex)
usr/src/uts/common/io/ixgbe/ixgbe_main.c
6602
ixgbe_rx_ring_t *rx_ring = &ixgbe->rx_rings[global_ring_index];
usr/src/uts/common/io/ixgbe/ixgbe_main.c
6603
rx_ring->ring_handle = rh;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
6605
infop->mri_driver = (mac_ring_driver_t)rx_ring;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
6611
mintr->mi_handle = (mac_intr_handle_t)rx_ring;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
6617
ixgbe->htable[rx_ring->intr_vector];
usr/src/uts/common/io/ixgbe/ixgbe_main.c
6696
ixgbe_rx_ring_t *rx_ring = (ixgbe_rx_ring_t *)intrh;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
6697
ixgbe_t *ixgbe = rx_ring->ixgbe;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
6698
int r_idx = rx_ring->index;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
6699
int hw_r_idx = rx_ring->hw_index;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
6700
int v_idx = rx_ring->intr_vector;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
6739
ixgbe_rx_ring_t *rx_ring = (ixgbe_rx_ring_t *)intrh;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
6740
ixgbe_t *ixgbe = rx_ring->ixgbe;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
6741
int r_idx = rx_ring->index;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
6742
int hw_r_idx = rx_ring->hw_index;
usr/src/uts/common/io/ixgbe/ixgbe_main.c
6743
int v_idx = rx_ring->intr_vector;
usr/src/uts/common/io/ixgbe/ixgbe_rx.c
146
ixgbe = rx_data->rx_ring->ixgbe;
usr/src/uts/common/io/ixgbe/ixgbe_rx.c
190
ixgbe_t *ixgbe = rx_data->rx_ring->ixgbe;
usr/src/uts/common/io/ixgbe/ixgbe_rx.c
277
ixgbe_t *ixgbe = rx_data->rx_ring->ixgbe;
usr/src/uts/common/io/ixgbe/ixgbe_rx.c
395
ixgbe = rx_data->rx_ring->ixgbe;
usr/src/uts/common/io/ixgbe/ixgbe_rx.c
53
ixgbe_rx_ring_t *rx_ring;
usr/src/uts/common/io/ixgbe/ixgbe_rx.c
537
ixgbe_ring_rx(ixgbe_rx_ring_t *rx_ring, int poll_bytes)
usr/src/uts/common/io/ixgbe/ixgbe_rx.c
555
ixgbe_t *ixgbe = rx_ring->ixgbe;
usr/src/uts/common/io/ixgbe/ixgbe_rx.c
564
rx_data = rx_ring->rx_data;
usr/src/uts/common/io/ixgbe/ixgbe_rx.c
599
rx_ring->stat_frame_error++;
usr/src/uts/common/io/ixgbe/ixgbe_rx.c
605
rx_ring->stat_cksum_error++;
usr/src/uts/common/io/ixgbe/ixgbe_rx.c
61
rx_ring = rx_data->rx_ring;
usr/src/uts/common/io/ixgbe/ixgbe_rx.c
62
ixgbe = rx_ring->ixgbe;
usr/src/uts/common/io/ixgbe/ixgbe_rx.c
714
rx_ring->stat_exceed_pkt++;
usr/src/uts/common/io/ixgbe/ixgbe_rx.c
722
rx_ring->stat_rbytes += received_bytes;
usr/src/uts/common/io/ixgbe/ixgbe_rx.c
723
rx_ring->stat_ipackets += pkt_num;
usr/src/uts/common/io/ixgbe/ixgbe_rx.c
738
IXGBE_WRITE_REG(&ixgbe->hw, IXGBE_RDT(rx_ring->hw_index), rx_tail);
usr/src/uts/common/io/ixgbe/ixgbe_rx.c
751
ixgbe_rx_ring_t *rx_ring = (ixgbe_rx_ring_t *)arg;
usr/src/uts/common/io/ixgbe/ixgbe_rx.c
759
mutex_enter(&rx_ring->rx_lock);
usr/src/uts/common/io/ixgbe/ixgbe_rx.c
760
mp = ixgbe_ring_rx(rx_ring, n_bytes);
usr/src/uts/common/io/ixgbe/ixgbe_rx.c
761
mutex_exit(&rx_ring->rx_lock);
usr/src/uts/common/io/ixgbe/ixgbe_stat.c
779
ixgbe_rx_ring_t *rx_ring = (ixgbe_rx_ring_t *)rh;
usr/src/uts/common/io/ixgbe/ixgbe_stat.c
780
ixgbe_t *ixgbe = rx_ring->ixgbe;
usr/src/uts/common/io/ixgbe/ixgbe_stat.c
788
*val = rx_ring->stat_rbytes;
usr/src/uts/common/io/ixgbe/ixgbe_stat.c
792
*val = rx_ring->stat_ipackets;
usr/src/uts/common/io/ixgbe/ixgbe_sw.h
581
struct ixgbe_rx_ring *rx_ring; /* Pointer to rx ring */
usr/src/uts/common/io/ixgbe/ixgbe_sw.h
887
int ixgbe_alloc_rx_ring_data(ixgbe_rx_ring_t *rx_ring);
usr/src/uts/common/io/mac/mac.c
2121
mac_rx_ring_quiesce(mac_ring_t *rx_ring, uint_t ring_flag)
usr/src/uts/common/io/mac/mac.c
2123
ASSERT(rx_ring->mr_classify_type == MAC_HW_CLASSIFIER);
usr/src/uts/common/io/mac/mac.c
2126
mutex_enter(&rx_ring->mr_lock);
usr/src/uts/common/io/mac/mac.c
2127
rx_ring->mr_flag |= ring_flag;
usr/src/uts/common/io/mac/mac.c
2128
while (rx_ring->mr_refcnt != 0)
usr/src/uts/common/io/mac/mac.c
2129
cv_wait(&rx_ring->mr_cv, &rx_ring->mr_lock);
usr/src/uts/common/io/mac/mac.c
2130
mutex_exit(&rx_ring->mr_lock);
usr/src/uts/common/io/qede/qede.h
1083
qede_update_rx_q_producer(qede_rx_ring_t *rx_ring);
usr/src/uts/common/io/qede/qede.h
1086
qede_put_to_passive_list(qede_rx_ring_t *rx_ring, qede_rx_buffer_t *rx_buffer);
usr/src/uts/common/io/qede/qede.h
201
#define RX_RING_MASK (rx_ring->rx_buf_count - 1)
usr/src/uts/common/io/qede/qede.h
378
struct qede_rx_ring *rx_ring;
usr/src/uts/common/io/qede/qede.h
435
struct qede_rx_ring *rx_ring;
usr/src/uts/common/io/qede/qede.h
459
struct qede_rx_ring *rx_ring;
usr/src/uts/common/io/qede/qede_dbg.c
263
qede_print_rx_ring(qede_rx_ring_t *rx_ring)
usr/src/uts/common/io/qede/qede_fp.c
105
qede_print_rx_indexes(qede_rx_ring_t *rx_ring)
usr/src/uts/common/io/qede/qede_fp.c
107
u16 hw_bd_cons = HOST_TO_LE_16(*rx_ring->hw_cons_ptr);
usr/src/uts/common/io/qede/qede_fp.c
108
u16 sw_bd_cons = ecore_chain_get_cons_idx(&rx_ring->rx_cqe_ring);
usr/src/uts/common/io/qede/qede_fp.c
110
hw_bd_cons &= (rx_ring->qede->rx_ring_size - 1);
usr/src/uts/common/io/qede/qede_fp.c
111
sw_bd_cons &= (rx_ring->qede->rx_ring_size - 1);
usr/src/uts/common/io/qede/qede_fp.c
218
qede_has_rx_work(qede_rx_ring_t *rx_ring)
usr/src/uts/common/io/qede/qede_fp.c
220
u16 hw_bd_cons = HOST_TO_LE_16(*rx_ring->hw_cons_ptr);
usr/src/uts/common/io/qede/qede_fp.c
221
u16 sw_bd_cons = ecore_chain_get_cons_idx(&rx_ring->rx_cqe_ring);
usr/src/uts/common/io/qede/qede_fp.c
261
qede_get_next_rx_buffer(qede_rx_ring_t *rx_ring,
usr/src/uts/common/io/qede/qede_fp.c
267
rx_buffer = qede_get_from_active_list(rx_ring, &num_entries);
usr/src/uts/common/io/qede/qede_fp.c
269
ecore_chain_consume(&rx_ring->rx_bd_ring);
usr/src/uts/common/io/qede/qede_fp.c
276
qede_get_next_lro_buffer(qede_rx_ring_t *rx_ring,
usr/src/uts/common/io/qede/qede_fp.c
280
qede_get_next_rx_buffer(rx_ring,
usr/src/uts/common/io/qede/qede_fp.c
290
qede_lro_start(qede_rx_ring_t *rx_ring,
usr/src/uts/common/io/qede/qede_fp.c
296
lro_info = &rx_ring->lro_info[cqe->tpa_agg_index];
usr/src/uts/common/io/qede/qede_fp.c
310
rx_ring->lro_active_count++;
usr/src/uts/common/io/qede/qede_fp.c
325
qede_get_next_lro_buffer(rx_ring, lro_info);
usr/src/uts/common/io/qede/qede_fp.c
337
qede_get_next_lro_buffer(rx_ring, lro_info);
usr/src/uts/common/io/qede/qede_fp.c
343
qede_lro_cont(qede_rx_ring_t *rx_ring,
usr/src/uts/common/io/qede/qede_fp.c
349
lro_info = &rx_ring->lro_info[cqe->tpa_agg_index];
usr/src/uts/common/io/qede/qede_fp.c
362
qede_get_next_lro_buffer(rx_ring, lro_info);
usr/src/uts/common/io/qede/qede_fp.c
367
qede_lro_end(qede_rx_ring_t *rx_ring,
usr/src/uts/common/io/qede/qede_fp.c
376
uint32_t rx_buf_size = rx_ring->rx_buf_size;
usr/src/uts/common/io/qede/qede_fp.c
379
lro_info = &rx_ring->lro_info[cqe->tpa_agg_index];
usr/src/uts/common/io/qede/qede_fp.c
398
qede_get_next_lro_buffer(rx_ring, lro_info);
usr/src/uts/common/io/qede/qede_fp.c
404
rx_ring->rx_low_buffer_threshold) {
usr/src/uts/common/io/qede/qede_fp.c
410
rx_ring->rx_low_water_cnt++;
usr/src/uts/common/io/qede/qede_fp.c
461
rx_ring->rx_lro_pkt_cnt++;
usr/src/uts/common/io/qede/qede_fp.c
462
rx_ring->lro_active_count--;
usr/src/uts/common/io/qede/qede_fp.c
467
qede_dump_mblk_chain_bcont_ptr(rx_ring->qede, head);
usr/src/uts/common/io/qede/qede_fp.c
481
qede_reg_jumbo_cqe(qede_rx_ring_t *rx_ring,
usr/src/uts/common/io/qede/qede_fp.c
489
uint32_t rx_buf_size = rx_ring->rx_buf_size, bd_len;
usr/src/uts/common/io/qede/qede_fp.c
508
qede_get_next_rx_buffer(rx_ring, &free_buffer_count);
usr/src/uts/common/io/qede/qede_fp.c
516
rx_ring->rx_low_buffer_threshold) {
usr/src/uts/common/io/qede/qede_fp.c
520
rx_ring->rx_low_water_cnt++;
usr/src/uts/common/io/qede/qede_fp.c
570
qede_dump_mblk_chain_bcont_ptr(rx_ring->qede, head);
usr/src/uts/common/io/qede/qede_fp.c
573
rx_ring->rx_jumbo_pkt_cnt++;
usr/src/uts/common/io/qede/qede_fp.c
578
qede_reg_cqe(qede_rx_ring_t *rx_ring,
usr/src/uts/common/io/qede/qede_fp.c
582
qede_t *qede = rx_ring->qede;
usr/src/uts/common/io/qede/qede_fp.c
606
return (qede_reg_jumbo_cqe(rx_ring, cqe));
usr/src/uts/common/io/qede/qede_fp.c
610
rx_buffer = qede_get_next_rx_buffer(rx_ring,
usr/src/uts/common/io/qede/qede_fp.c
614
rx_ring->rx_low_buffer_threshold) {
usr/src/uts/common/io/qede/qede_fp.c
616
rx_ring->rx_low_water_cnt++;
usr/src/uts/common/io/qede/qede_fp.c
627
if (pkt_len <= rx_ring->rx_copy_threshold) {
usr/src/uts/common/io/qede/qede_fp.c
639
rx_ring->qede->instance);
usr/src/uts/common/io/qede/qede_fp.c
650
rx_ring->rx_copy_cnt++;
usr/src/uts/common/io/qede/qede_fp.c
671
qede_dump_mblk_chain_bnext_ptr(rx_ring->qede, mp);
usr/src/uts/common/io/qede/qede_fp.c
675
rx_ring->rx_reg_pkt_cnt++;
usr/src/uts/common/io/qede/qede_fp.c
689
qede_process_rx_ring(qede_rx_ring_t *rx_ring, int nbytes, int npkts)
usr/src/uts/common/io/qede/qede_fp.c
692
u16 last_cqe_consumer = rx_ring->last_cqe_consumer;
usr/src/uts/common/io/qede/qede_fp.c
698
hw_comp_cons = HOST_TO_LE_16(*rx_ring->hw_cons_ptr);
usr/src/uts/common/io/qede/qede_fp.c
701
sw_comp_cons = ecore_chain_get_cons_idx(&rx_ring->rx_cqe_ring);
usr/src/uts/common/io/qede/qede_fp.c
710
ecore_chain_consume(&rx_ring->rx_cqe_ring);
usr/src/uts/common/io/qede/qede_fp.c
713
(void) ddi_dma_sync(rx_ring->rx_cqe_dmah,
usr/src/uts/common/io/qede/qede_fp.c
721
ecore_eth_cqe_completion(&rx_ring->qede->edev.hwfns[0],
usr/src/uts/common/io/qede/qede_fp.c
725
mp = qede_reg_cqe(rx_ring,
usr/src/uts/common/io/qede/qede_fp.c
730
qede_lro_start(rx_ring,
usr/src/uts/common/io/qede/qede_fp.c
734
qede_lro_cont(rx_ring,
usr/src/uts/common/io/qede/qede_fp.c
738
mp = qede_lro_end(rx_ring,
usr/src/uts/common/io/qede/qede_fp.c
746
rx_ring->qede->instance,
usr/src/uts/common/io/qede/qede_fp.c
759
rx_ring->rx_drop_cnt++;
usr/src/uts/common/io/qede/qede_fp.c
772
ecore_chain_recycle_consumed(&rx_ring->rx_cqe_ring);
usr/src/uts/common/io/qede/qede_fp.c
774
sw_comp_cons = ecore_chain_get_cons_idx(&rx_ring->rx_cqe_ring);
usr/src/uts/common/io/qede/qede_fp.c
775
if (!(qede_has_rx_work(rx_ring))) {
usr/src/uts/common/io/qede/qede_fp.c
776
ecore_sb_update_sb_idx(rx_ring->fp->sb_info);
usr/src/uts/common/io/qede/qede_fp.c
778
hw_comp_cons = HOST_TO_LE_16(*rx_ring->hw_cons_ptr);
usr/src/uts/common/io/qede/qede_fp.c
780
rx_ring->rx_pkt_cnt += pkt_cnt;
usr/src/uts/common/io/qede/qede_fp.c
781
rx_ring->rx_byte_cnt += byte_cnt;
usr/src/uts/common/io/qede/qede_fp.c
792
qede_replenish_rx_buffers(rx_ring);
usr/src/uts/common/io/qede/qede_fp.c
793
qede_update_rx_q_producer(rx_ring);
usr/src/uts/common/io/qede/qede_fp.c
794
rx_ring->last_cqe_consumer = last_cqe_consumer;
usr/src/uts/common/io/qede/qede_fp.c
805
qede_rx_ring_t *rx_ring;
usr/src/uts/common/io/qede/qede_fp.c
808
rx_ring = fp->rx_ring;
usr/src/uts/common/io/qede/qede_fp.c
822
if (!(qede_has_rx_work(rx_ring))) {
usr/src/uts/common/io/qede/qede_fp.c
826
rx_ring = fp->rx_ring;
usr/src/uts/common/io/qede/qede_fp.c
827
if (qede_has_rx_work(rx_ring)) {
usr/src/uts/common/io/qede/qede_fp.c
828
mutex_enter(&rx_ring->rx_lock);
usr/src/uts/common/io/qede/qede_fp.c
829
mp = qede_process_rx_ring(rx_ring,
usr/src/uts/common/io/qede/qede_fp.c
834
mutex_exit(&rx_ring->rx_lock);
usr/src/uts/common/io/qede/qede_fp.h
39
#define RX_INDICATE_UPSTREAM(rx_ring, mp) \
usr/src/uts/common/io/qede/qede_fp.h
40
mac_rx_ring(rx_ring->qede->mac_handle, \
usr/src/uts/common/io/qede/qede_fp.h
41
rx_ring->mac_ring_handle, mp, \
usr/src/uts/common/io/qede/qede_fp.h
42
rx_ring->mr_gen_num)
usr/src/uts/common/io/qede/qede_gld.c
282
fp->rx_ring->rx_poll_cnt++;
usr/src/uts/common/io/qede/qede_gld.c
307
fp->rx_ring->intrEnableCnt++;
usr/src/uts/common/io/qede/qede_gld.c
330
fp->rx_ring->intrDisableCnt++;
usr/src/uts/common/io/qede/qede_gld.c
345
qede_rx_ring_t *rx_ring = fp->rx_ring;
usr/src/uts/common/io/qede/qede_gld.c
353
*val = rx_ring->rx_byte_cnt;
usr/src/uts/common/io/qede/qede_gld.c
356
*val = rx_ring->rx_pkt_cnt;
usr/src/uts/common/io/qede/qede_gld.c
371
qede_rx_ring_t *rx_ring;
usr/src/uts/common/io/qede/qede_gld.c
376
rx_ring = fp->rx_ring;
usr/src/uts/common/io/qede/qede_gld.c
378
if (rx_ring->group_index == gindex) {
usr/src/uts/common/io/qede/qede_gld.c
393
qede_rx_ring_t *rx_ring = fp->rx_ring;
usr/src/uts/common/io/qede/qede_gld.c
397
rx_ring->mac_ring_started = B_FALSE;
usr/src/uts/common/io/qede/qede_gld.c
405
qede_rx_ring_t *rx_ring = fp->rx_ring;
usr/src/uts/common/io/qede/qede_gld.c
409
rx_ring->mr_gen_num = mr_gen_num;
usr/src/uts/common/io/qede/qede_gld.c
410
rx_ring->mac_ring_started = B_TRUE;
usr/src/uts/common/io/qede/qede_gld.c
411
rx_ring->intrDisableCnt = 0;
usr/src/uts/common/io/qede/qede_gld.c
412
rx_ring->intrEnableCnt = 0;
usr/src/uts/common/io/qede/qede_gld.c
440
qede_rx_ring_t *rx_ring;
usr/src/uts/common/io/qede/qede_gld.c
455
rx_ring = fp->rx_ring;
usr/src/uts/common/io/qede/qede_gld.c
458
rx_ring->mac_ring_handle = rh;
usr/src/uts/common/io/qede/qede_gld.c
461
rx_ring->rss_id, rh);
usr/src/uts/common/io/qede/qede_gld.c
690
qede_rx_ring_t *rx_ring;
usr/src/uts/common/io/qede/qede_kstat.c
260
qede_rx_ring_t *rx_ring;
usr/src/uts/common/io/qede/qede_kstat.c
332
rx_ring = fp->rx_ring;
usr/src/uts/common/io/qede/qede_kstat.c
333
qede->rxTotalPkts += rx_ring->rx_pkt_cnt;
usr/src/uts/common/io/qede/qede_kstat.c
334
qede->rxTotalBytes += rx_ring->rx_byte_cnt;
usr/src/uts/common/io/qede/qede_kstat.c
335
qede->rxTotalDiscards += rx_ring->rx_drop_cnt;
usr/src/uts/common/io/qede/qede_kstat.c
578
qede_rx_ring_t *rx_ring = &qede->rx_array[idx];
usr/src/uts/common/io/qede/qede_kstat.c
590
pStats->rxLowWater.value.ui64 = rx_ring->rx_low_buffer_threshold;
usr/src/uts/common/io/qede/qede_kstat.c
595
ecore_chain_get_elem_left(&rx_ring->rx_bd_ring);
usr/src/uts/common/io/qede/qede_kstat.c
603
ecore_chain_get_prod_idx(&rx_ring->rx_bd_ring) &
usr/src/uts/common/io/qede/qede_kstat.c
604
(rx_ring->qede->rx_ring_size - 1);
usr/src/uts/common/io/qede/qede_kstat.c
606
ecore_chain_get_cons_idx(&rx_ring->rx_bd_ring) &
usr/src/uts/common/io/qede/qede_kstat.c
607
(rx_ring->qede->rx_ring_size - 1);
usr/src/uts/common/io/qede/qede_kstat.c
609
ecore_chain_get_elem_left(&rx_ring->rx_cqe_ring);
usr/src/uts/common/io/qede/qede_kstat.c
617
ecore_chain_get_prod_idx(&rx_ring->rx_cqe_ring) &
usr/src/uts/common/io/qede/qede_kstat.c
618
(rx_ring->qede->rx_ring_size - 1);
usr/src/uts/common/io/qede/qede_kstat.c
620
ecore_chain_get_cons_idx(&rx_ring->rx_cqe_ring) &
usr/src/uts/common/io/qede/qede_kstat.c
621
(rx_ring->qede->rx_ring_size - 1);
usr/src/uts/common/io/qede/qede_kstat.c
623
(rx_ring->hw_cons_ptr != NULL) ?
usr/src/uts/common/io/qede/qede_kstat.c
624
HOST_TO_LE_16(*rx_ring->hw_cons_ptr) &
usr/src/uts/common/io/qede/qede_kstat.c
625
(rx_ring->qede->rx_ring_size - 1): 0;
usr/src/uts/common/io/qede/qede_kstat.c
627
rx_ring->rx_buf_area->passive_buf_list.num_entries;
usr/src/uts/common/io/qede/qede_kstat.c
629
rx_ring->rx_buf_area->active_buf_list.num_entries;
usr/src/uts/common/io/qede/qede_kstat.c
631
rx_ring->rx_buf_area->buf_upstream;
usr/src/uts/common/io/qede/qede_kstat.c
633
rx_ring->rx_copy_cnt;
usr/src/uts/common/io/qede/qede_kstat.c
635
rx_ring->rx_drop_cnt;
usr/src/uts/common/io/qede/qede_kstat.c
637
rx_ring->rx_low_water_cnt;
usr/src/uts/common/io/qede/qede_kstat.c
639
rx_ring->fp->disabled_by_poll;
usr/src/uts/common/io/qede/qede_kstat.c
641
rx_ring->rx_poll_cnt;;
usr/src/uts/common/io/qede/qede_kstat.c
643
rx_ring->intrDisableCnt;
usr/src/uts/common/io/qede/qede_kstat.c
645
rx_ring->intrEnableCnt;
usr/src/uts/common/io/qede/qede_kstat.c
647
rx_ring->mr_gen_num;
usr/src/uts/common/io/qede/qede_kstat.c
649
rx_ring->rx_lro_pkt_cnt;
usr/src/uts/common/io/qede/qede_kstat.c
651
rx_ring->rx_pkt_cnt;
usr/src/uts/common/io/qede/qede_kstat.c
653
rx_ring->rx_byte_cnt;
usr/src/uts/common/io/qede/qede_kstat.c
655
rx_ring->rx_reg_pkt_cnt;
usr/src/uts/common/io/qede/qede_kstat.c
657
rx_ring->rx_jumbo_pkt_cnt;
usr/src/uts/common/io/qede/qede_main.c
1389
rss_params.rss_ind_table[i] = (void *)(fp->rx_ring->p_cid);
usr/src/uts/common/io/qede/qede_main.c
1589
qede_update_rx_q_producer(qede_rx_ring_t *rx_ring)
usr/src/uts/common/io/qede/qede_main.c
1591
u16 bd_prod = ecore_chain_get_prod_idx(&rx_ring->rx_bd_ring);
usr/src/uts/common/io/qede/qede_main.c
1592
u16 cqe_prod = ecore_chain_get_prod_idx(&rx_ring->rx_cqe_ring);
usr/src/uts/common/io/qede/qede_main.c
1599
UPDATE_RX_PROD(rx_ring, rx_prod_cmd);
usr/src/uts/common/io/qede/qede_main.c
1612
qede_rx_ring_t *rx_ring;
usr/src/uts/common/io/qede/qede_main.c
1629
rx_ring = fp->rx_ring;
usr/src/uts/common/io/qede/qede_main.c
1650
if (rx_ring->queue_started == B_TRUE) {
usr/src/uts/common/io/qede/qede_main.c
1653
p_rx_cid = rx_ring->p_cid;
usr/src/uts/common/io/qede/qede_main.c
1663
rx_ring->queue_started = B_FALSE;
usr/src/uts/common/io/qede/qede_main.c
1691
qede_rx_ring_t *rx_ring;
usr/src/uts/common/io/qede/qede_main.c
1709
rx_ring = fp->rx_ring;
usr/src/uts/common/io/qede/qede_main.c
1717
p_phys_table = ecore_chain_get_pbl_phys(&rx_ring->rx_cqe_ring);
usr/src/uts/common/io/qede/qede_main.c
1718
page_cnt = ecore_chain_get_page_cnt(&rx_ring->rx_cqe_ring);
usr/src/uts/common/io/qede/qede_main.c
1724
rx_ring->rx_bd_ring.p_phys_addr,
usr/src/uts/common/io/qede/qede_main.c
1729
rx_ring->hw_rxq_prod_addr = rx_ret_params.p_prod;
usr/src/uts/common/io/qede/qede_main.c
1730
rx_ring->p_cid = rx_ret_params.p_handle;
usr/src/uts/common/io/qede/qede_main.c
1736
rx_ring->hw_cons_ptr = &fp->sb_info->sb_virt->pi_array[RX_PI];
usr/src/uts/common/io/qede/qede_main.c
1739
*rx_ring->hw_cons_ptr = 0;
usr/src/uts/common/io/qede/qede_main.c
1741
qede_update_rx_q_producer(rx_ring);
usr/src/uts/common/io/qede/qede_main.c
1742
rx_ring->queue_started = B_TRUE;
usr/src/uts/common/io/qede/qede_main.c
1799
qede_free_mag_elem(qede_rx_ring_t *rx_ring, qede_rx_buffer_t *rx_buffer,
usr/src/uts/common/io/qede/qede_main.c
1815
qede_free_lro_rx_buffers(qede_rx_ring_t *rx_ring)
usr/src/uts/common/io/qede/qede_main.c
1821
lro_info = &rx_ring->lro_info[i];
usr/src/uts/common/io/qede/qede_main.c
1843
qede_rx_ring_t *rx_ring = rx_buf_area->rx_ring;
usr/src/uts/common/io/qede/qede_main.c
1851
for (i = 0; i < rx_ring->rx_buf_count; i += bufs_per_page) {
usr/src/uts/common/io/qede/qede_main.c
1907
qede_free_rx_buffers(qede_t *qede, qede_rx_ring_t *rx_ring)
usr/src/uts/common/io/qede/qede_main.c
1909
qede_free_lro_rx_buffers(rx_ring);
usr/src/uts/common/io/qede/qede_main.c
1910
qede_rx_buf_area_t *rx_buf_area = rx_ring->rx_buf_area;
usr/src/uts/common/io/qede/qede_main.c
1917
qede_rx_ring_t *rx_ring;
usr/src/uts/common/io/qede/qede_main.c
1923
rx_ring = fp->rx_ring;
usr/src/uts/common/io/qede/qede_main.c
1924
rx_ring->rx_buf_area->inactive = 1;
usr/src/uts/common/io/qede/qede_main.c
1926
qede_free_rx_buffers(qede, rx_ring);
usr/src/uts/common/io/qede/qede_main.c
1929
if (rx_ring->rx_bd_ring.p_virt_addr) {
usr/src/uts/common/io/qede/qede_main.c
1930
ecore_chain_free(&qede->edev, &rx_ring->rx_bd_ring);
usr/src/uts/common/io/qede/qede_main.c
1931
rx_ring->rx_bd_ring.p_virt_addr = NULL;
usr/src/uts/common/io/qede/qede_main.c
1934
if (rx_ring->rx_cqe_ring.p_virt_addr) {
usr/src/uts/common/io/qede/qede_main.c
1935
ecore_chain_free(&qede->edev, &rx_ring->rx_cqe_ring);
usr/src/uts/common/io/qede/qede_main.c
1936
rx_ring->rx_cqe_ring.p_virt_addr = NULL;
usr/src/uts/common/io/qede/qede_main.c
1937
if (rx_ring->rx_cqe_ring.pbl_sp.p_virt_table) {
usr/src/uts/common/io/qede/qede_main.c
1938
rx_ring->rx_cqe_ring.pbl_sp.p_virt_table = NULL;
usr/src/uts/common/io/qede/qede_main.c
1941
rx_ring->hw_cons_ptr = NULL;
usr/src/uts/common/io/qede/qede_main.c
1942
rx_ring->hw_rxq_prod_addr = NULL;
usr/src/uts/common/io/qede/qede_main.c
1943
rx_ring->sw_rx_cons = 0;
usr/src/uts/common/io/qede/qede_main.c
1944
rx_ring->sw_rx_prod = 0;
usr/src/uts/common/io/qede/qede_main.c
1950
qede_init_bd(qede_t *qede, qede_rx_ring_t *rx_ring)
usr/src/uts/common/io/qede/qede_main.c
1956
qede_rx_buf_area_t *rx_buf_area = rx_ring->rx_buf_area;
usr/src/uts/common/io/qede/qede_main.c
1959
for (i = 0; i < rx_ring->rx_buf_count; i++) {
usr/src/uts/common/io/qede/qede_main.c
1963
bd = ecore_chain_produce(&rx_ring->rx_bd_ring);
usr/src/uts/common/io/qede/qede_main.c
1984
qede_get_from_active_list(qede_rx_ring_t *rx_ring,
usr/src/uts/common/io/qede/qede_main.c
1989
&rx_ring->rx_buf_area->active_buf_list;
usr/src/uts/common/io/qede/qede_main.c
1998
atomic_inc_32(&rx_ring->rx_buf_area->buf_upstream);
usr/src/uts/common/io/qede/qede_main.c
2005
rx_ring->rx_buf_size, 0, &rx_buffer->recycle);
usr/src/uts/common/io/qede/qede_main.c
2016
qede_get_from_passive_list(qede_rx_ring_t *rx_ring)
usr/src/uts/common/io/qede/qede_main.c
2019
&rx_ring->rx_buf_area->passive_buf_list;
usr/src/uts/common/io/qede/qede_main.c
2042
qede_put_to_active_list(qede_rx_ring_t *rx_ring, qede_rx_buffer_t *rx_buffer)
usr/src/uts/common/io/qede/qede_main.c
2045
&rx_ring->rx_buf_area->active_buf_list;
usr/src/uts/common/io/qede/qede_main.c
2056
qede_replenish_rx_buffers(qede_rx_ring_t *rx_ring)
usr/src/uts/common/io/qede/qede_main.c
2080
if (mutex_tryenter(&rx_ring->rx_replen_lock) == 0) {
usr/src/uts/common/io/qede/qede_main.c
2081
qede_info(rx_ring->qede, "!%s(%d): Failed to take"
usr/src/uts/common/io/qede/qede_main.c
2083
__func__, rx_ring->qede->instance);
usr/src/uts/common/io/qede/qede_main.c
2087
rx_buffer = qede_get_from_passive_list(rx_ring);
usr/src/uts/common/io/qede/qede_main.c
2090
bd = ecore_chain_produce(&rx_ring->rx_bd_ring);
usr/src/uts/common/io/qede/qede_main.c
2092
qede_info(rx_ring->qede, "!%s(%d): bd = null",
usr/src/uts/common/io/qede/qede_main.c
2093
__func__, rx_ring->qede->instance);
usr/src/uts/common/io/qede/qede_main.c
2094
qede_put_to_passive_list(rx_ring, rx_buffer);
usr/src/uts/common/io/qede/qede_main.c
2107
qede_put_to_active_list(rx_ring, rx_buffer);
usr/src/uts/common/io/qede/qede_main.c
2110
rx_buffer = qede_get_from_passive_list(rx_ring);
usr/src/uts/common/io/qede/qede_main.c
2112
mutex_exit(&rx_ring->rx_replen_lock);
usr/src/uts/common/io/qede/qede_main.c
2119
qede_put_to_passive_list(qede_rx_ring_t *rx_ring, qede_rx_buffer_t *rx_buffer)
usr/src/uts/common/io/qede/qede_main.c
2122
&rx_ring->rx_buf_area->passive_buf_list;
usr/src/uts/common/io/qede/qede_main.c
2123
qede_rx_buf_area_t *rx_buf_area = rx_ring->rx_buf_area;
usr/src/uts/common/io/qede/qede_main.c
2134
if (passive_buf_list->num_entries > rx_ring->rx_buf_count) {
usr/src/uts/common/io/qede/qede_main.c
2136
qede_info(rx_ring->qede, "ERROR: num_entries (%d)"
usr/src/uts/common/io/qede/qede_main.c
2139
rx_ring->rx_buf_count);
usr/src/uts/common/io/qede/qede_main.c
2150
qede_rx_ring_t *rx_ring = rx_buffer->rx_ring;
usr/src/uts/common/io/qede/qede_main.c
2152
qede_t *qede = rx_ring->qede;
usr/src/uts/common/io/qede/qede_main.c
2166
rx_ring->rx_buf_size, 0, &rx_buffer->recycle);
usr/src/uts/common/io/qede/qede_main.c
2171
num_entries = qede_put_to_passive_list(rx_ring, rx_buffer);
usr/src/uts/common/io/qede/qede_main.c
2173
if(mutex_tryenter(&rx_ring->rx_lock) != 0) {
usr/src/uts/common/io/qede/qede_main.c
2174
qede_replenish_rx_buffers(rx_ring);
usr/src/uts/common/io/qede/qede_main.c
2175
qede_update_rx_q_producer(rx_ring);
usr/src/uts/common/io/qede/qede_main.c
2176
mutex_exit(&rx_ring->rx_lock);
usr/src/uts/common/io/qede/qede_main.c
2203
qede_info(rx_ring->qede, "rx_buffer %p"
usr/src/uts/common/io/qede/qede_main.c
2212
qede_rx_ring_t *rx_ring = rx_buffer->rx_ring;
usr/src/uts/common/io/qede/qede_main.c
2214
qede_t *qede = rx_ring->qede;
usr/src/uts/common/io/qede/qede_main.c
2242
qede_put_to_passive_list(rx_ring, rx_buffer);
usr/src/uts/common/io/qede/qede_main.c
2251
qede_info(rx_ring->qede, "Free up rx_buffer %p, index %d"
usr/src/uts/common/io/qede/qede_main.c
2266
qede_info(rx_ring->qede, "rx_buffer %p"
usr/src/uts/common/io/qede/qede_main.c
2274
qede_alloc_rx_buffers(qede_t *qede, qede_rx_ring_t *rx_ring)
usr/src/uts/common/io/qede/qede_main.c
2278
qede_rx_buf_area_t *rx_buf_area = rx_ring->rx_buf_area;
usr/src/uts/common/io/qede/qede_main.c
2289
if (rx_ring->rx_buf_size > page_size) {
usr/src/uts/common/io/qede/qede_main.c
2291
buf_size = rx_ring->rx_buf_size;
usr/src/uts/common/io/qede/qede_main.c
2306
for (i = 0; i < rx_ring->rx_buf_count; i += bufs_per_page) {
usr/src/uts/common/io/qede/qede_main.c
2334
dma_info->offset = j * rx_ring->rx_buf_size;
usr/src/uts/common/io/qede/qede_main.c
2340
rx_ring->rx_buf_size, 0,
usr/src/uts/common/io/qede/qede_main.c
2346
rx_buffer->rx_ring = rx_ring;
usr/src/uts/common/io/qede/qede_main.c
2352
vaddr += rx_ring->rx_buf_size;
usr/src/uts/common/io/qede/qede_main.c
2353
dma_addr += rx_ring->rx_buf_size;
usr/src/uts/common/io/qede/qede_main.c
2355
rx_ring->sw_rx_prod++;
usr/src/uts/common/io/qede/qede_main.c
2362
ret = qede_init_bd(qede, rx_ring);
usr/src/uts/common/io/qede/qede_main.c
2371
qede_free_rx_buffers(qede, rx_ring);
usr/src/uts/common/io/qede/qede_main.c
2378
qede_rx_ring_t *rx_ring;
usr/src/uts/common/io/qede/qede_main.c
2385
rx_ring = fp->rx_ring;
usr/src/uts/common/io/qede/qede_main.c
2399
rx_buf_area->rx_ring = rx_ring;
usr/src/uts/common/io/qede/qede_main.c
2400
rx_ring->rx_buf_area = rx_buf_area;
usr/src/uts/common/io/qede/qede_main.c
2408
&rx_ring->rx_bd_ring,
usr/src/uts/common/io/qede/qede_main.c
2422
&rx_ring->rx_cqe_ring,
usr/src/uts/common/io/qede/qede_main.c
2430
if (qede_alloc_rx_buffers(qede, rx_ring) != DDI_SUCCESS) {
usr/src/uts/common/io/qede/qede_main.c
2844
qede_rx_ring_t *rx_ring;
usr/src/uts/common/io/qede/qede_main.c
2847
rx_ring = fp->rx_ring;
usr/src/uts/common/io/qede/qede_main.c
2851
(void *)rx_ring->rx_bd_ring.p_phys_addr,
usr/src/uts/common/io/qede/qede_main.c
2852
&rx_ring->rx_bd_dmah);
usr/src/uts/common/io/qede/qede_main.c
2856
rx_ring->rx_bd_ring.p_phys_addr);
usr/src/uts/common/io/qede/qede_main.c
2862
(void *)rx_ring->rx_cqe_ring.p_phys_addr,
usr/src/uts/common/io/qede/qede_main.c
2863
&rx_ring->rx_cqe_dmah);
usr/src/uts/common/io/qede/qede_main.c
2867
rx_ring->rx_cqe_ring.p_phys_addr);
usr/src/uts/common/io/qede/qede_main.c
2872
(void *)rx_ring->rx_cqe_ring.pbl_sp.p_phys_table,
usr/src/uts/common/io/qede/qede_main.c
2873
&rx_ring->rx_cqe_pbl_dmah);
usr/src/uts/common/io/qede/qede_main.c
2877
rx_ring->rx_cqe_ring.pbl_sp.p_phys_table);
usr/src/uts/common/io/qede/qede_main.c
2950
qede_rx_ring_t *rx_ring;
usr/src/uts/common/io/qede/qede_main.c
2987
rx_ring = &qede->rx_array[i];
usr/src/uts/common/io/qede/qede_main.c
2988
fp->rx_ring = rx_ring;
usr/src/uts/common/io/qede/qede_main.c
2989
rx_ring->fp = fp;
usr/src/uts/common/io/qede/qede_main.c
2990
rx_ring->rx_buf_count = qede->rx_buf_count;
usr/src/uts/common/io/qede/qede_main.c
2991
rx_ring->rx_buf_size = qede->rx_buf_size;
usr/src/uts/common/io/qede/qede_main.c
2992
rx_ring->qede = qede;
usr/src/uts/common/io/qede/qede_main.c
2993
rx_ring->sw_rx_cons = 0;
usr/src/uts/common/io/qede/qede_main.c
2994
rx_ring->rx_copy_threshold = qede->rx_copy_threshold;
usr/src/uts/common/io/qede/qede_main.c
2995
rx_ring->rx_low_buffer_threshold =
usr/src/uts/common/io/qede/qede_main.c
2997
rx_ring->queue_started = B_FALSE;
usr/src/uts/common/io/qede/qede_main.c
317
qede_rx_ring_t *rx_ring;
usr/src/uts/common/io/qede/qede_main.c
327
rx_ring = fp->rx_ring;
usr/src/uts/common/io/qede/qede_main.c
328
mutex_destroy(&rx_ring->rx_lock);
usr/src/uts/common/io/qede/qede_main.c
329
mutex_destroy(&rx_ring->rx_replen_lock);
usr/src/uts/common/io/qede/qede_main.c
345
qede_rx_ring_t *rx_ring;
usr/src/uts/common/io/qede/qede_main.c
358
rx_ring = fp->rx_ring;
usr/src/uts/common/io/qede/qede_main.c
359
mutex_init(&rx_ring->rx_lock, NULL,
usr/src/uts/common/io/qede/qede_main.c
361
mutex_init(&rx_ring->rx_replen_lock, NULL,
usr/src/uts/common/io/qede/qede_main.c
386
qede_rx_ring_t *rx_ring;
usr/src/uts/common/io/qede/qede_main.c
396
rx_ring = &qede->rx_array[i];
usr/src/uts/common/io/qede/qede_main.c
404
fp->rx_ring = rx_ring;
usr/src/uts/common/io/qede/qede_main.c
405
rx_ring->group_index = 0;
usr/src/uts/common/io/qede/qede_main.c
690
qede_rx_ring_t *rx_ring;
usr/src/uts/common/io/qede/qede_main.c
700
rx_ring = fp->rx_ring;
usr/src/uts/common/io/qede/qede_main.c
717
mac_rx_ring(rx_ring->qede->mac_handle,
usr/src/uts/common/io/qede/qede_main.c
718
rx_ring->mac_ring_handle,
usr/src/uts/common/io/qede/qede_main.c
720
rx_ring->mr_gen_num);
usr/src/uts/common/io/rge/rge.h
428
rge_bd_t *rx_ring;
usr/src/uts/common/io/rge/rge_main.c
350
bdp = rgep->rx_ring;
usr/src/uts/common/io/rge/rge_main.c
536
rgep->rx_ring = rgep->rx_desc.mem_va;
usr/src/uts/common/io/rge/rge_rxtx.c
129
hw_rbd_p = &rgep->rx_ring[slot];
usr/src/uts/common/io/rge/rge_rxtx.c
169
hw_rbd_p = &rgep->rx_ring[slot];
usr/src/uts/common/io/rge/rge_rxtx.c
315
hw_rbd_p = &rgep->rx_ring[slot];
usr/src/uts/common/io/rge/rge_rxtx.c
332
hw_rbd_p = &rgep->rx_ring[slot];
usr/src/uts/common/io/rtls/rtls.c
1224
rx_ptr = rtlsp->rx_ring + cur_rx;
usr/src/uts/common/io/rtls/rtls.c
1315
bcopy(rtlsp->rx_ring, mp->b_wptr, wrap_size);
usr/src/uts/common/io/rtls/rtls.c
1596
rtlsp->rx_ring = (uint8_t *)rtlsp->dma_area_rx.mem_va;
usr/src/uts/common/io/rtls/rtls.c
1616
rtlsp->rx_ring = NULL;
usr/src/uts/common/io/rtls/rtls.h
187
uint8_t *rx_ring;
usr/src/uts/common/io/sfe/sfe.c
1195
rdp = (void *)&dp->rx_ring[SFE_DESC_SIZE * slot];
usr/src/uts/common/io/sfe/sfe.c
1326
rdp = (void *)&dp->rx_ring[SFE_DESC_SIZE * slot];
usr/src/uts/common/io/sfe/sfe.c
1432
rdp = (void *)&dp->rx_ring[SFE_DESC_SIZE * slot];
usr/src/uts/common/io/sfe/sfe.c
1441
&dp->rx_ring[SFE_DESC_SIZE * SLOT(slot - 1, rx_ring_size)];
usr/src/uts/common/io/sfe/sfe.c
1459
rdp = (void *)&dp->rx_ring[SFE_DESC_SIZE * slot];
usr/src/uts/common/io/sfe/sfe_util.c
670
dp->rx_ring = ring;
usr/src/uts/common/io/sfe/sfe_util.c
674
dp->tx_ring = dp->rx_ring + dp->rx_desc_size;
usr/src/uts/common/io/sfe/sfe_util.h
203
caddr_t rx_ring;
usr/src/uts/common/io/xge/drv/xgell.c
1400
xgell_rx_ring_t *rx_ring = (xgell_rx_ring_t *)rh;
usr/src/uts/common/io/xge/drv/xgell.c
1402
rx_ring->ring_gen_num = mr_gen_num;
usr/src/uts/common/io/xge/drv/xgell.c
1446
xgell_rx_ring_t *rx_ring;
usr/src/uts/common/io/xge/drv/xgell.c
1455
rx_ring = lldev->rx_ring + rg_index;
usr/src/uts/common/io/xge/drv/xgell.c
1457
rx_ring = lldev->rx_ring + index;
usr/src/uts/common/io/xge/drv/xgell.c
1459
rx_ring->ring_handle = rh;
usr/src/uts/common/io/xge/drv/xgell.c
1461
infop->mri_driver = (mac_ring_driver_t)rx_ring;
usr/src/uts/common/io/xge/drv/xgell.c
1468
mintr->mi_handle = (mac_intr_handle_t)rx_ring;
usr/src/uts/common/io/xge/drv/xgell.c
1505
xgell_rx_ring_t *rx_ring;
usr/src/uts/common/io/xge/drv/xgell.c
1509
rx_ring = lldev->rx_ring + index;
usr/src/uts/common/io/xge/drv/xgell.c
1511
rx_ring->group_handle = gh;
usr/src/uts/common/io/xge/drv/xgell.c
1513
infop->mgi_driver = (mac_group_driver_t)rx_ring;
usr/src/uts/common/io/xge/drv/xgell.c
1585
xgell_rx_ring_open(xgell_rx_ring_t *rx_ring)
usr/src/uts/common/io/xge/drv/xgell.c
1589
xgelldev_t *lldev = rx_ring->lldev;
usr/src/uts/common/io/xge/drv/xgell.c
1592
if (rx_ring->live)
usr/src/uts/common/io/xge/drv/xgell.c
1596
if (!xgell_rx_create_buffer_pool(rx_ring)) {
usr/src/uts/common/io/xge/drv/xgell.c
1598
rx_ring->index);
usr/src/uts/common/io/xge/drv/xgell.c
1603
attr.post_qid = rx_ring->index;
usr/src/uts/common/io/xge/drv/xgell.c
1611
attr.userdata = rx_ring;
usr/src/uts/common/io/xge/drv/xgell.c
1613
status = xge_hal_channel_open(lldev->devh, &attr, &rx_ring->channelh,
usr/src/uts/common/io/xge/drv/xgell.c
1618
(void) xgell_rx_destroy_buffer_pool(rx_ring);
usr/src/uts/common/io/xge/drv/xgell.c
1622
xgell_rx_ring_maddr_init(rx_ring);
usr/src/uts/common/io/xge/drv/xgell.c
1624
mutex_init(&rx_ring->ring_lock, NULL, MUTEX_DRIVER,
usr/src/uts/common/io/xge/drv/xgell.c
1627
rx_ring->poll_bytes = -1;
usr/src/uts/common/io/xge/drv/xgell.c
1628
rx_ring->polled_bytes = 0;
usr/src/uts/common/io/xge/drv/xgell.c
1629
rx_ring->poll_mp = NULL;
usr/src/uts/common/io/xge/drv/xgell.c
1630
rx_ring->live = B_TRUE;
usr/src/uts/common/io/xge/drv/xgell.c
1633
rx_ring->index);
usr/src/uts/common/io/xge/drv/xgell.c
1639
xgell_rx_ring_close(xgell_rx_ring_t *rx_ring)
usr/src/uts/common/io/xge/drv/xgell.c
1641
if (!rx_ring->live)
usr/src/uts/common/io/xge/drv/xgell.c
1643
xge_hal_channel_close(rx_ring->channelh, XGE_HAL_CHANNEL_OC_NORMAL);
usr/src/uts/common/io/xge/drv/xgell.c
1644
rx_ring->channelh = NULL;
usr/src/uts/common/io/xge/drv/xgell.c
1646
if (xgell_rx_destroy_buffer_pool(rx_ring))
usr/src/uts/common/io/xge/drv/xgell.c
1647
rx_ring->live = B_FALSE;
usr/src/uts/common/io/xge/drv/xgell.c
1649
mutex_destroy(&rx_ring->ring_lock);
usr/src/uts/common/io/xge/drv/xgell.c
1661
xgell_rx_ring_t *rx_ring;
usr/src/uts/common/io/xge/drv/xgell.c
1673
rx_ring = &lldev->rx_ring[i];
usr/src/uts/common/io/xge/drv/xgell.c
1674
rx_ring->index = i;
usr/src/uts/common/io/xge/drv/xgell.c
1675
rx_ring->lldev = lldev;
usr/src/uts/common/io/xge/drv/xgell.c
1676
rx_ring->live = B_FALSE;
usr/src/uts/common/io/xge/drv/xgell.c
1678
if (!xgell_rx_ring_open(rx_ring))
usr/src/uts/common/io/xge/drv/xgell.c
1690
xgell_rx_ring_t *rx_ring;
usr/src/uts/common/io/xge/drv/xgell.c
1700
rx_ring = &lldev->rx_ring[i];
usr/src/uts/common/io/xge/drv/xgell.c
1702
if (rx_ring->live) {
usr/src/uts/common/io/xge/drv/xgell.c
1703
xgell_rx_ring_close(rx_ring);
usr/src/uts/common/io/xge/drv/xgell.c
2274
xgell_rx_ring_t *rx_ring = (xgell_rx_ring_t *)rh;
usr/src/uts/common/io/xge/drv/xgell.c
2278
*val = rx_ring->rx_bytes;
usr/src/uts/common/io/xge/drv/xgell.c
2282
*val = rx_ring->rx_pkts;
usr/src/uts/common/io/xge/drv/xgell.h
370
xgell_rx_ring_t rx_ring[XGELL_RX_RING_NUM_DEFAULT];
usr/src/uts/common/sys/fibre-channel/fca/qlge/qlge.h
191
#define RX_UPSTREAM(rx_ring, mp) mac_rx(rx_ring->qlge->mh, \
usr/src/uts/common/sys/fibre-channel/fca/qlge/qlge.h
192
rx_ring->qlge->handle, mp);
usr/src/uts/common/sys/fibre-channel/fca/qlge/qlge.h
397
struct rx_ring *rx_ring;
usr/src/uts/common/sys/fibre-channel/fca/qlge/qlge.h
799
struct rx_ring rx_ring[MAX_RX_RINGS];
usr/src/uts/sun/io/eri/eri.c
1929
uint64_t rx_ring;
usr/src/uts/sun/io/eri/eri.c
1938
rx_ring = ERI_IOPBIOADDR(erip, erip->rmdp);
usr/src/uts/sun/io/eri/eri.c
1939
PUT_ERXREG(rxring_lo, (uint32_t)(rx_ring));
usr/src/uts/sun/io/eri/eri.c
1940
PUT_ERXREG(rxring_hi, (uint32_t)(rx_ring >> 32));