drivers/net/ethernet/3com/3c515.c
1320
int entry = vp->cur_rx % RX_RING_SIZE;
drivers/net/ethernet/3com/3c515.c
1382
entry = (++vp->cur_rx) % RX_RING_SIZE;
drivers/net/ethernet/3com/3c515.c
1387
entry = vp->dirty_rx % RX_RING_SIZE;
drivers/net/ethernet/3com/3c515.c
1437
for (i = 0; i < RX_RING_SIZE; i++)
drivers/net/ethernet/3com/3c515.c
298
struct boom_rx_desc rx_ring[RX_RING_SIZE];
drivers/net/ethernet/3com/3c515.c
301
struct sk_buff *rx_skbuff[RX_RING_SIZE];
drivers/net/ethernet/3com/3c515.c
807
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/3com/3c515.c
809
if (i < (RX_RING_SIZE - 1))
drivers/net/ethernet/3com/3c59x.c
1211
vp->rx_ring = dma_alloc_coherent(gendev, sizeof(struct boom_rx_desc) * RX_RING_SIZE
drivers/net/ethernet/3com/3c59x.c
1218
vp->tx_ring = (struct boom_tx_desc *)(vp->rx_ring + RX_RING_SIZE);
drivers/net/ethernet/3com/3c59x.c
1219
vp->tx_ring_dma = vp->rx_ring_dma + sizeof(struct boom_rx_desc) * RX_RING_SIZE;
drivers/net/ethernet/3com/3c59x.c
1477
sizeof(struct boom_rx_desc) * RX_RING_SIZE +
drivers/net/ethernet/3com/3c59x.c
1683
for (i = 0; i < RX_RING_SIZE; i++) /* AKPM: this is done in vortex_open, too */
drivers/net/ethernet/3com/3c59x.c
1736
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/3com/3c59x.c
1755
if (i != RX_RING_SIZE) {
drivers/net/ethernet/3com/3c59x.c
1769
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/3com/3c59x.c
2594
int entry = vp->cur_rx % RX_RING_SIZE;
drivers/net/ethernet/3com/3c59x.c
2597
int rx_work_limit = RX_RING_SIZE;
drivers/net/ethernet/3com/3c59x.c
2680
entry = (++vp->cur_rx) % RX_RING_SIZE;
drivers/net/ethernet/3com/3c59x.c
2755
for (i = 0; i < RX_RING_SIZE; i++)
drivers/net/ethernet/3com/3c59x.c
3285
sizeof(struct boom_rx_desc) * RX_RING_SIZE +
drivers/net/ethernet/3com/3c59x.c
602
struct sk_buff* rx_skbuff[RX_RING_SIZE];
drivers/net/ethernet/adaptec/starfire.c
1130
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/adaptec/starfire.c
1148
np->dirty_rx = (unsigned int)(i - RX_RING_SIZE);
drivers/net/ethernet/adaptec/starfire.c
1151
for ( ; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/adaptec/starfire.c
1157
np->rx_ring[RX_RING_SIZE - 1].rxaddr |= cpu_to_dma(RxDescEndRing);
drivers/net/ethernet/adaptec/starfire.c
127
#if RX_RING_SIZE > 256
drivers/net/ethernet/adaptec/starfire.c
1583
entry = np->dirty_rx % RX_RING_SIZE;
drivers/net/ethernet/adaptec/starfire.c
1600
if (entry == RX_RING_SIZE - 1)
drivers/net/ethernet/adaptec/starfire.c
1962
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/adaptec/starfire.c
522
struct rx_ring_info rx_info[RX_RING_SIZE];
drivers/net/ethernet/adaptec/starfire.c
881
rx_ring_size = sizeof(struct starfire_rx_desc) * RX_RING_SIZE;
drivers/net/ethernet/alteon/acenic.c
1481
ace_load_std_rx_ring(dev, RX_RING_SIZE);
drivers/net/ethernet/alteon/acenic.c
1575
ace_load_std_rx_ring(dev, RX_RING_SIZE - cur_size);
drivers/net/ethernet/alteon/acenic.c
2173
RX_RING_SIZE - cur_size);
drivers/net/ethernet/alteon/acenic.c
332
#define RX_LOW_STD_THRES (3*RX_RING_SIZE)/4
drivers/net/ethernet/amd/7990.c
284
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/amd/7990.c
98
for (t = 0; t < RX_RING_SIZE; t++) { \
drivers/net/ethernet/amd/7990.h
42
#define RX_RING_MOD_MASK (RX_RING_SIZE - 1)
drivers/net/ethernet/amd/7990.h
88
volatile struct lance_rx_desc brx_ring[RX_RING_SIZE];
drivers/net/ethernet/amd/7990.h
91
volatile char rx_buf[RX_RING_SIZE][RX_BUFF_SIZE];
drivers/net/ethernet/amd/a2065.c
251
char buf[RX_RING_SIZE + 1];
drivers/net/ethernet/amd/a2065.c
253
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/amd/a2065.c
260
buf[RX_RING_SIZE] = 0;
drivers/net/ethernet/amd/a2065.c
76
#define RX_RING_MOD_MASK (RX_RING_SIZE - 1)
drivers/net/ethernet/amd/a2065.c
96
struct lance_rx_desc brx_ring[RX_RING_SIZE];
drivers/net/ethernet/amd/a2065.c
99
char rx_buf[RX_RING_SIZE][RX_BUFF_SIZE];
drivers/net/ethernet/amd/ariadne.c
101
struct RDRE rx_ring[RX_RING_SIZE];
drivers/net/ethernet/amd/ariadne.c
103
u_short rx_buff[RX_RING_SIZE][PKT_BUF_SIZE / sizeof(u_short)];
drivers/net/ethernet/amd/ariadne.c
145
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/amd/ariadne.c
164
int entry = priv->cur_rx % RX_RING_SIZE;
drivers/net/ethernet/amd/ariadne.c
197
for (i = 0; i < RX_RING_SIZE; i++)
drivers/net/ethernet/amd/ariadne.c
198
if (lowb(priv->rx_ring[(entry + i) % RX_RING_SIZE]->RMD1) & RF_OWN)
drivers/net/ethernet/amd/ariadne.c
201
if (i > RX_RING_SIZE - 2) {
drivers/net/ethernet/amd/ariadne.c
227
entry = (++priv->cur_rx) % RX_RING_SIZE;
drivers/net/ethernet/amd/ariadne.c
230
priv->cur_rx = priv->cur_rx % RX_RING_SIZE;
drivers/net/ethernet/amd/ariadne.c
468
lance->RDP = swapw(((u_short)-RX_RING_SIZE));
drivers/net/ethernet/amd/ariadne.c
89
volatile struct RDRE *rx_ring[RX_RING_SIZE];
drivers/net/ethernet/amd/ariadne.c
91
volatile u_short *rx_buff[RX_RING_SIZE];
drivers/net/ethernet/amd/atarilance.c
1000
for( i = 0; i < RX_RING_SIZE; i++ )
drivers/net/ethernet/amd/atarilance.c
1005
if (i > RX_RING_SIZE - 2) {
drivers/net/ethernet/amd/atarilance.c
118
#define RX_RING_MOD_MASK (RX_RING_SIZE - 1)
drivers/net/ethernet/amd/atarilance.c
159
struct lance_rx_head rx_head[RX_RING_SIZE];
drivers/net/ethernet/amd/atarilance.c
715
for( i = 0; i < RX_RING_SIZE; i++ ) {
drivers/net/ethernet/amd/atarilance.c
751
for( i = 0 ; i < RX_RING_SIZE; i++ )
drivers/net/ethernet/amd/declance.c
1088
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/amd/declance.c
1098
2 * RX_RING_SIZE * RX_BUFF_SIZE +
drivers/net/ethernet/amd/declance.c
1102
RX_RING_SIZE * RX_BUFF_SIZE +
drivers/net/ethernet/amd/declance.c
1133
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/amd/declance.c
1143
RX_RING_SIZE * RX_BUFF_SIZE +
drivers/net/ethernet/amd/declance.c
1147
RX_RING_SIZE * RX_BUFF_SIZE +
drivers/net/ethernet/amd/declance.c
1164
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/amd/declance.c
1174
2 * RX_RING_SIZE * RX_BUFF_SIZE +
drivers/net/ethernet/amd/declance.c
1178
RX_RING_SIZE * RX_BUFF_SIZE +
drivers/net/ethernet/amd/declance.c
160
#define RX_RING_MOD_MASK (RX_RING_SIZE - 1)
drivers/net/ethernet/amd/declance.c
222
struct lance_rx_desc brx_ring[RX_RING_SIZE];
drivers/net/ethernet/amd/declance.c
267
char *rx_buf_ptr_cpu[RX_RING_SIZE];
drivers/net/ethernet/amd/declance.c
271
uint rx_buf_ptr_lnc[RX_RING_SIZE];
drivers/net/ethernet/amd/declance.c
511
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/amd/declance.c
570
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/amd/lance.c
1194
for (i=0; i < RX_RING_SIZE; i++)
drivers/net/ethernet/amd/lance.c
1198
if (i > RX_RING_SIZE -2)
drivers/net/ethernet/amd/lance.c
199
#define RX_RING_MOD_MASK (RX_RING_SIZE - 1)
drivers/net/ethernet/amd/lance.c
238
struct lance_rx_head rx_ring[RX_RING_SIZE];
drivers/net/ethernet/amd/lance.c
245
struct sk_buff* rx_skbuff[RX_RING_SIZE];
drivers/net/ethernet/amd/lance.c
560
lp->rx_buffs = (unsigned long)kmalloc_array(RX_RING_SIZE, PKT_BUF_SZ,
drivers/net/ethernet/amd/lance.c
849
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/amd/lance.c
875
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/amd/lance.c
938
for (i = 0; i < RX_RING_SIZE; i++)
drivers/net/ethernet/amd/pcnet32.c
1861
lp->rx_ring_size = RX_RING_SIZE; /* default rx ring size */
drivers/net/ethernet/amd/sun3lance.c
101
#define RX_RING_MOD_MASK (RX_RING_SIZE - 1)
drivers/net/ethernet/amd/sun3lance.c
144
struct lance_rx_head rx_head[RX_RING_SIZE];
drivers/net/ethernet/amd/sun3lance.c
145
char rx_data[RX_RING_SIZE][PKT_BUF_SZ];
drivers/net/ethernet/amd/sun3lance.c
469
for( i = 0; i < RX_RING_SIZE; i++ ) {
drivers/net/ethernet/amd/sun3lance.c
543
for( i = 0 ; i < RX_RING_SIZE; i++ )
drivers/net/ethernet/amd/sunlance.c
182
#define RX_RING_MOD_MASK (RX_RING_SIZE - 1)
drivers/net/ethernet/amd/sunlance.c
222
struct lance_rx_desc brx_ring[RX_RING_SIZE];
drivers/net/ethernet/amd/sunlance.c
227
u8 rx_buf [RX_RING_SIZE][RX_BUFF_SIZE];
drivers/net/ethernet/amd/sunlance.c
350
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/amd/sunlance.c
408
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/dec/tulip/interrupt.c
116
int entry = tp->cur_rx % RX_RING_SIZE;
drivers/net/ethernet/dec/tulip/interrupt.c
127
if (budget >=RX_RING_SIZE) budget--;
drivers/net/ethernet/dec/tulip/interrupt.c
148
if (tp->dirty_rx + RX_RING_SIZE == tp->cur_rx)
drivers/net/ethernet/dec/tulip/interrupt.c
264
entry = (++tp->cur_rx) % RX_RING_SIZE;
drivers/net/ethernet/dec/tulip/interrupt.c
265
if (tp->cur_rx - tp->dirty_rx > RX_RING_SIZE/4)
drivers/net/ethernet/dec/tulip/interrupt.c
321
if (tp->rx_buffers[tp->dirty_rx % RX_RING_SIZE].skb == NULL)
drivers/net/ethernet/dec/tulip/interrupt.c
343
if (tp->cur_rx - tp->dirty_rx > RX_RING_SIZE/2 ||
drivers/net/ethernet/dec/tulip/interrupt.c
344
tp->rx_buffers[tp->dirty_rx % RX_RING_SIZE].skb == NULL)
drivers/net/ethernet/dec/tulip/interrupt.c
347
if (tp->rx_buffers[tp->dirty_rx % RX_RING_SIZE].skb == NULL)
drivers/net/ethernet/dec/tulip/interrupt.c
372
int entry = tp->cur_rx % RX_RING_SIZE;
drivers/net/ethernet/dec/tulip/interrupt.c
373
int rx_work_limit = tp->dirty_rx + RX_RING_SIZE - tp->cur_rx;
drivers/net/ethernet/dec/tulip/interrupt.c
490
entry = (++tp->cur_rx) % RX_RING_SIZE;
drivers/net/ethernet/dec/tulip/interrupt.c
532
int maxrx = RX_RING_SIZE;
drivers/net/ethernet/dec/tulip/interrupt.c
67
entry = tp->dirty_rx % RX_RING_SIZE;
drivers/net/ethernet/dec/tulip/interrupt.c
788
entry = tp->dirty_rx % RX_RING_SIZE;
drivers/net/ethernet/dec/tulip/tulip.h
411
struct ring_info rx_buffers[RX_RING_SIZE];
drivers/net/ethernet/dec/tulip/tulip_core.c
1430
sizeof(struct tulip_rx_desc) * RX_RING_SIZE +
drivers/net/ethernet/dec/tulip/tulip_core.c
1435
tp->tx_ring = (struct tulip_tx_desc *)(tp->rx_ring + RX_RING_SIZE);
drivers/net/ethernet/dec/tulip/tulip_core.c
1436
tp->tx_ring_dma = tp->rx_ring_dma + sizeof(struct tulip_rx_desc) * RX_RING_SIZE;
drivers/net/ethernet/dec/tulip/tulip_core.c
567
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/dec/tulip/tulip_core.c
584
for (i = 0; i < RX_RING_SIZE; i++)
drivers/net/ethernet/dec/tulip/tulip_core.c
612
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/dec/tulip/tulip_core.c
623
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/dec/tulip/tulip_core.c
639
tp->dirty_rx = (unsigned int)(i - RX_RING_SIZE);
drivers/net/ethernet/dec/tulip/tulip_core.c
787
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/dec/tulip/winbond-840.c
1164
int entry = np->cur_rx % RX_RING_SIZE;
drivers/net/ethernet/dec/tulip/winbond-840.c
1165
int work_limit = np->dirty_rx + RX_RING_SIZE - np->cur_rx;
drivers/net/ethernet/dec/tulip/winbond-840.c
1247
entry = (++np->cur_rx) % RX_RING_SIZE;
drivers/net/ethernet/dec/tulip/winbond-840.c
1254
entry = np->dirty_rx % RX_RING_SIZE;
drivers/net/ethernet/dec/tulip/winbond-840.c
1504
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/dec/tulip/winbond-840.c
286
dma_addr_t rx_addr[RX_RING_SIZE];
drivers/net/ethernet/dec/tulip/winbond-840.c
291
struct sk_buff* rx_skbuff[RX_RING_SIZE];
drivers/net/ethernet/dec/tulip/winbond-840.c
787
np->tx_ring = (struct w840_tx_desc*)&np->rx_ring[RX_RING_SIZE];
drivers/net/ethernet/dec/tulip/winbond-840.c
790
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/dec/tulip/winbond-840.c
799
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/dec/tulip/winbond-840.c
813
np->dirty_rx = (unsigned int)(i - RX_RING_SIZE);
drivers/net/ethernet/dec/tulip/winbond-840.c
824
iowrite32(np->ring_dma_addr+sizeof(struct w840_rx_desc)*RX_RING_SIZE,
drivers/net/ethernet/dec/tulip/winbond-840.c
833
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/dec/tulip/winbond-840.c
923
for (i = 0; i < RX_RING_SIZE; i++)
drivers/net/ethernet/dec/tulip/winbond-840.c
965
sizeof(struct w840_rx_desc) * RX_RING_SIZE +
drivers/net/ethernet/dec/tulip/winbond-840.c
977
sizeof(struct w840_rx_desc) * RX_RING_SIZE +
drivers/net/ethernet/dlink/dl2k.c
1008
entry = (entry + 1) % RX_RING_SIZE;
drivers/net/ethernet/dlink/dl2k.c
1035
entry = (entry + 1) % RX_RING_SIZE;
drivers/net/ethernet/dlink/dl2k.c
447
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/dlink/dl2k.c
483
for (i = 0; i < RX_RING_SIZE; i++)
drivers/net/ethernet/dlink/dl2k.c
505
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/dlink/dl2k.c
521
((i + 1) % RX_RING_SIZE) *
drivers/net/ethernet/dlink/dl2k.c
678
if (np->cur_rx - np->old_rx >= RX_RING_SIZE) {
drivers/net/ethernet/dlink/dl2k.c
683
entry = np->old_rx % RX_RING_SIZE;
drivers/net/ethernet/dlink/dl2k.c
940
int entry = np->cur_rx % RX_RING_SIZE;
drivers/net/ethernet/dlink/dl2k.h
368
struct sk_buff *rx_skbuff[RX_RING_SIZE];
drivers/net/ethernet/dlink/dl2k.h
39
#define RX_TOTAL_SIZE RX_RING_SIZE*sizeof(struct netdev_desc)
drivers/net/ethernet/dlink/sundance.c
1026
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/dlink/sundance.c
1028
((i+1)%RX_RING_SIZE)*sizeof(*np->rx_ring));
drivers/net/ethernet/dlink/sundance.c
1035
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/dlink/sundance.c
1054
np->dirty_rx = (unsigned int)(i - RX_RING_SIZE);
drivers/net/ethernet/dlink/sundance.c
1325
int entry = np->cur_rx % RX_RING_SIZE;
drivers/net/ethernet/dlink/sundance.c
1396
entry = (entry + 1) % RX_RING_SIZE;
drivers/net/ethernet/dlink/sundance.c
1422
for (;(np->cur_rx - np->dirty_rx + RX_RING_SIZE) % RX_RING_SIZE > 0;
drivers/net/ethernet/dlink/sundance.c
1423
np->dirty_rx = (np->dirty_rx + 1) % RX_RING_SIZE) {
drivers/net/ethernet/dlink/sundance.c
1427
entry = np->dirty_rx % RX_RING_SIZE;
drivers/net/ethernet/dlink/sundance.c
1893
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/dlink/sundance.c
365
struct sk_buff* rx_skbuff[RX_RING_SIZE];
drivers/net/ethernet/dlink/sundance.c
70
#define RX_TOTAL_SIZE RX_RING_SIZE*sizeof(struct netdev_desc)
drivers/net/ethernet/fealnx.c
1054
while (np->really_rx_count != RX_RING_SIZE) {
drivers/net/ethernet/fealnx.c
1201
for (i = 0; i < RX_RING_SIZE; i++)
drivers/net/ethernet/fealnx.c
1238
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/fealnx.c
1252
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/fealnx.c
1414
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/fealnx.c
1909
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/fealnx.c
59
#define RX_TOTAL_SIZE RX_RING_SIZE*sizeof(struct fealnx_desc)
drivers/net/ethernet/freescale/fec.h
563
union fec_rx_buffer rx_buf[RX_RING_SIZE];
drivers/net/ethernet/freescale/fec_main.c
3984
fep->rx_queue[i]->bd.ring_size = RX_RING_SIZE;
drivers/net/ethernet/freescale/fs_enet/fs_enet-main.c
889
fpi->rx_ring = RX_RING_SIZE;
drivers/net/ethernet/i825xx/82596.c
331
struct i596_rfd rfds[RX_RING_SIZE];
drivers/net/ethernet/i825xx/82596.c
332
struct i596_rbd rbds[RX_RING_SIZE];
drivers/net/ethernet/i825xx/82596.c
369
static int rx_ring_size = RX_RING_SIZE;
drivers/net/ethernet/i825xx/lib82596.c
307
struct i596_rfd rfds[RX_RING_SIZE] __attribute__((aligned(32)));
drivers/net/ethernet/i825xx/lib82596.c
308
struct i596_rbd rbds[RX_RING_SIZE] __attribute__((aligned(32)));
drivers/net/ethernet/i825xx/lib82596.c
359
static int rx_ring_size = RX_RING_SIZE;
drivers/net/ethernet/natsemi/natsemi.c
1719
writel(np->ring_dma + RX_RING_SIZE * sizeof(struct netdev_desc),
drivers/net/ethernet/natsemi/natsemi.c
1872
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/natsemi/natsemi.c
1917
sizeof(struct netdev_desc) * (RX_RING_SIZE + TX_RING_SIZE),
drivers/net/ethernet/natsemi/natsemi.c
1921
np->tx_ring = &np->rx_ring[RX_RING_SIZE];
drivers/net/ethernet/natsemi/natsemi.c
1932
int entry = np->dirty_rx % RX_RING_SIZE;
drivers/net/ethernet/natsemi/natsemi.c
1951
if (np->cur_rx - np->dirty_rx == RX_RING_SIZE) {
drivers/net/ethernet/natsemi/natsemi.c
1979
*((i+1)%TX_RING_SIZE+RX_RING_SIZE));
drivers/net/ethernet/natsemi/natsemi.c
1985
np->cur_rx = RX_RING_SIZE;
drivers/net/ethernet/natsemi/natsemi.c
1995
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/natsemi/natsemi.c
1998
*((i+1)%RX_RING_SIZE));
drivers/net/ethernet/natsemi/natsemi.c
2029
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/natsemi/natsemi.c
2052
sizeof(struct netdev_desc) * (RX_RING_SIZE + TX_RING_SIZE),
drivers/net/ethernet/natsemi/natsemi.c
2063
np->cur_rx = RX_RING_SIZE;
drivers/net/ethernet/natsemi/natsemi.c
2066
for (i = 0; i < RX_RING_SIZE; i++)
drivers/net/ethernet/natsemi/natsemi.c
2206
prefetch(&np->rx_skbuff[np->cur_rx % RX_RING_SIZE]);
drivers/net/ethernet/natsemi/natsemi.c
2280
int entry = np->cur_rx % RX_RING_SIZE;
drivers/net/ethernet/natsemi/natsemi.c
2281
int boguscnt = np->dirty_rx + RX_RING_SIZE - np->cur_rx;
drivers/net/ethernet/natsemi/natsemi.c
2381
entry = (++np->cur_rx) % RX_RING_SIZE;
drivers/net/ethernet/natsemi/natsemi.c
544
struct sk_buff *rx_skbuff[RX_RING_SIZE];
drivers/net/ethernet/natsemi/natsemi.c
545
dma_addr_t rx_dma[RX_RING_SIZE];
drivers/net/ethernet/packetengines/hamachi.c
1062
for (i = 0; i < RX_RING_SIZE; i++)
drivers/net/ethernet/packetengines/hamachi.c
1084
for (i = 0; i < RX_RING_SIZE; i++)
drivers/net/ethernet/packetengines/hamachi.c
1121
for (i = 0; i < RX_RING_SIZE; i++){
drivers/net/ethernet/packetengines/hamachi.c
1133
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/packetengines/hamachi.c
1148
hmp->dirty_rx = (unsigned int)(i - RX_RING_SIZE);
drivers/net/ethernet/packetengines/hamachi.c
1150
hmp->rx_ring[RX_RING_SIZE-1].status_n_length |= cpu_to_le32(DescEndRing);
drivers/net/ethernet/packetengines/hamachi.c
1184
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/packetengines/hamachi.c
1189
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/packetengines/hamachi.c
1203
hmp->dirty_rx = (unsigned int)(i - RX_RING_SIZE);
drivers/net/ethernet/packetengines/hamachi.c
1204
hmp->rx_ring[RX_RING_SIZE-1].status_n_length |= cpu_to_le32(DescEndRing);
drivers/net/ethernet/packetengines/hamachi.c
122
#define RX_TOTAL_SIZE RX_RING_SIZE*sizeof(struct hamachi_desc)
drivers/net/ethernet/packetengines/hamachi.c
1411
int entry = hmp->cur_rx % RX_RING_SIZE;
drivers/net/ethernet/packetengines/hamachi.c
1412
int boguscnt = (hmp->dirty_rx + RX_RING_SIZE) - hmp->cur_rx;
drivers/net/ethernet/packetengines/hamachi.c
1444
dev->name, desc, &hmp->rx_ring[hmp->cur_rx % RX_RING_SIZE]);
drivers/net/ethernet/packetengines/hamachi.c
1447
le32_to_cpu(hmp->rx_ring[(hmp->cur_rx+1) % RX_RING_SIZE].status_n_length) & 0xffff0000,
drivers/net/ethernet/packetengines/hamachi.c
1448
le32_to_cpu(hmp->rx_ring[(hmp->cur_rx+1) % RX_RING_SIZE].status_n_length) & 0x0000ffff,
drivers/net/ethernet/packetengines/hamachi.c
1449
le32_to_cpu(hmp->rx_ring[(hmp->cur_rx-1) % RX_RING_SIZE].status_n_length));
drivers/net/ethernet/packetengines/hamachi.c
1586
entry = (++hmp->cur_rx) % RX_RING_SIZE;
drivers/net/ethernet/packetengines/hamachi.c
1593
entry = hmp->dirty_rx % RX_RING_SIZE;
drivers/net/ethernet/packetengines/hamachi.c
1608
if (entry >= RX_RING_SIZE-1)
drivers/net/ethernet/packetengines/hamachi.c
1694
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/packetengines/hamachi.c
1718
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/packetengines/hamachi.c
486
struct sk_buff* rx_skbuff[RX_RING_SIZE];
drivers/net/ethernet/packetengines/yellowfin.c
1048
int entry = yp->cur_rx % RX_RING_SIZE;
drivers/net/ethernet/packetengines/yellowfin.c
1049
int boguscnt = yp->dirty_rx + RX_RING_SIZE - yp->cur_rx;
drivers/net/ethernet/packetengines/yellowfin.c
1159
entry = (++yp->cur_rx) % RX_RING_SIZE;
drivers/net/ethernet/packetengines/yellowfin.c
1164
entry = yp->dirty_rx % RX_RING_SIZE;
drivers/net/ethernet/packetengines/yellowfin.c
1182
yp->rx_ring[RX_RING_SIZE - 1].dbdma_cmd =
drivers/net/ethernet/packetengines/yellowfin.c
1244
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/packetengines/yellowfin.c
1267
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/packetengines/yellowfin.c
311
struct sk_buff* rx_skbuff[RX_RING_SIZE];
drivers/net/ethernet/packetengines/yellowfin.c
701
for (i = 0; i < RX_RING_SIZE; i++)
drivers/net/ethernet/packetengines/yellowfin.c
737
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/packetengines/yellowfin.c
741
((i+1)%RX_RING_SIZE)*sizeof(struct yellowfin_desc));
drivers/net/ethernet/packetengines/yellowfin.c
744
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/packetengines/yellowfin.c
755
if (i != RX_RING_SIZE) {
drivers/net/ethernet/packetengines/yellowfin.c
761
yp->dirty_rx = (unsigned int)(i - RX_RING_SIZE);
drivers/net/ethernet/packetengines/yellowfin.c
78
#define RX_TOTAL_SIZE RX_RING_SIZE*sizeof(struct yellowfin_desc)
drivers/net/ethernet/pasemi/pasemi_mac.c
1133
pasemi_mac_replenish_rx_ring(dev, RX_RING_SIZE);
drivers/net/ethernet/pasemi/pasemi_mac.c
1136
RX_RING_SIZE>>1);
drivers/net/ethernet/pasemi/pasemi_mac.c
1314
pasemi_mac_clean_rx(rx_ring(mac), RX_RING_SIZE);
drivers/net/ethernet/pasemi/pasemi_mac.c
1618
pasemi_mac_clean_rx(rx_ring(mac), RX_RING_SIZE);
drivers/net/ethernet/pasemi/pasemi_mac.c
1650
pasemi_mac_replenish_rx_ring(dev, RX_RING_SIZE-1);
drivers/net/ethernet/pasemi/pasemi_mac.c
383
ring->size = RX_RING_SIZE;
drivers/net/ethernet/pasemi/pasemi_mac.c
384
ring->ring_info = kzalloc_objs(struct pasemi_mac_buffer, RX_RING_SIZE);
drivers/net/ethernet/pasemi/pasemi_mac.c
390
if (pasemi_dma_alloc_ring(&ring->chan, RX_RING_SIZE))
drivers/net/ethernet/pasemi/pasemi_mac.c
394
RX_RING_SIZE * sizeof(u64),
drivers/net/ethernet/pasemi/pasemi_mac.c
404
PAS_DMA_RXCHAN_BASEU_SIZ(RX_RING_SIZE >> 3));
drivers/net/ethernet/pasemi/pasemi_mac.c
418
PAS_DMA_RXINT_BASEU_SIZ(RX_RING_SIZE >> 3));
drivers/net/ethernet/pasemi/pasemi_mac.c
546
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/pasemi/pasemi_mac.c
557
for (i = 0; i < RX_RING_SIZE; i++)
drivers/net/ethernet/pasemi/pasemi_mac.c
565
dma_free_coherent(&mac->dma_pdev->dev, RX_RING_SIZE * sizeof(u64),
drivers/net/ethernet/pasemi/pasemi_mac.c
619
(RX_RING_SIZE - 1);
drivers/net/ethernet/pasemi/pasemi_mac.c
784
if (n > RX_RING_SIZE) {
drivers/net/ethernet/pasemi/pasemi_mac.c
787
n &= (RX_RING_SIZE-1);
drivers/net/ethernet/pasemi/pasemi_mac.h
96
#define RX_DESC(rx, num) ((rx)->chan.ring_virt[(num) & (RX_RING_SIZE-1)])
drivers/net/ethernet/pasemi/pasemi_mac.h
97
#define RX_DESC_INFO(rx, num) ((rx)->ring_info[(num) & (RX_RING_SIZE-1)])
drivers/net/ethernet/pasemi/pasemi_mac.h
98
#define RX_BUFF(rx, num) ((rx)->buffers[(num) & (RX_RING_SIZE-1)])
drivers/net/ethernet/pasemi/pasemi_mac_ethtool.c
80
ering->rx_max_pending = RX_RING_SIZE/4;
drivers/net/ethernet/qlogic/qede/qede.h
590
#define NUM_RX_BDS_MAX (RX_RING_SIZE - 1)
drivers/net/ethernet/qlogic/qede/qede_main.c
1554
.num_elems = RX_RING_SIZE,
drivers/net/ethernet/qlogic/qede/qede_main.c
1583
size = sizeof(*rxq->sw_rx_ring) * RX_RING_SIZE;
drivers/net/ethernet/qlogic/qede/qede_main.c
2871
RX_RING_SIZE - 100)
drivers/net/ethernet/renesas/rswitch_main.c
655
err = rswitch_gwca_queue_alloc(ndev, priv, rdev->rx_queue, false, RX_RING_SIZE);
drivers/net/ethernet/renesas/sh_eth.c
3280
mdp->num_rx_ring = RX_RING_SIZE;
drivers/net/ethernet/sgi/ioc3-eth.c
885
ip->rxr = dma_alloc_coherent(ip->dma_dev, RX_RING_SIZE, &ip->rxr_dma,
drivers/net/ethernet/sgi/ioc3-eth.c
955
dma_free_coherent(ip->dma_dev, RX_RING_SIZE, ip->rxr,
drivers/net/ethernet/sgi/ioc3-eth.c
970
dma_free_coherent(ip->dma_dev, RX_RING_SIZE, ip->rxr, ip->rxr_dma);
drivers/net/ethernet/smsc/epic100.c
1131
int entry = ep->cur_rx % RX_RING_SIZE;
drivers/net/ethernet/smsc/epic100.c
1132
int rx_work_limit = ep->dirty_rx + RX_RING_SIZE - ep->cur_rx;
drivers/net/ethernet/smsc/epic100.c
1202
entry = (++ep->cur_rx) % RX_RING_SIZE;
drivers/net/ethernet/smsc/epic100.c
1207
entry = ep->dirty_rx % RX_RING_SIZE;
drivers/net/ethernet/smsc/epic100.c
1291
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/smsc/epic100.c
250
struct sk_buff* rx_skbuff[RX_RING_SIZE];
drivers/net/ethernet/smsc/epic100.c
55
#define RX_TOTAL_SIZE RX_RING_SIZE*sizeof(struct epic_rx_desc)
drivers/net/ethernet/smsc/epic100.c
764
epic_rx(dev, RX_RING_SIZE);
drivers/net/ethernet/smsc/epic100.c
800
(ep->cur_rx % RX_RING_SIZE) * sizeof(struct epic_rx_desc));
drivers/net/ethernet/smsc/epic100.c
897
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/smsc/epic100.c
908
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/smsc/epic100.c
920
ep->dirty_rx = (unsigned int)(i - RX_RING_SIZE);
drivers/net/ethernet/smsc/smsc9420.c
1213
pd->rx_buffers = kmalloc_objs(struct smsc9420_ring_info, RX_RING_SIZE);
drivers/net/ethernet/smsc/smsc9420.c
1218
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/smsc/smsc9420.c
1225
pd->rx_ring[RX_RING_SIZE - 1].length = (PKT_BUF_SZ | RDES1_RER_);
drivers/net/ethernet/smsc/smsc9420.c
1228
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/smsc/smsc9420.c
1546
sizeof(struct smsc9420_dma_desc) * (RX_RING_SIZE + TX_RING_SIZE),
drivers/net/ethernet/smsc/smsc9420.c
1553
pd->tx_ring = (pd->rx_ring + RX_RING_SIZE);
drivers/net/ethernet/smsc/smsc9420.c
1555
sizeof(struct smsc9420_dma_desc) * RX_RING_SIZE;
drivers/net/ethernet/smsc/smsc9420.c
1604
sizeof(struct smsc9420_dma_desc) * (RX_RING_SIZE + TX_RING_SIZE),
drivers/net/ethernet/smsc/smsc9420.c
1638
sizeof(struct smsc9420_dma_desc) * (RX_RING_SIZE + TX_RING_SIZE),
drivers/net/ethernet/smsc/smsc9420.c
532
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/smsc/smsc9420.c
813
pd->rx_ring_tail = (pd->rx_ring_tail + 1) % RX_RING_SIZE;
drivers/net/ethernet/smsc/smsc9420.c
835
pd->rx_ring_head = (pd->rx_ring_head + 1) % RX_RING_SIZE;
drivers/net/ethernet/sun/sunbmac.c
197
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/sun/sunbmac.c
227
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/sun/sunbmac.h
254
#define NEXT_RX(num) (((num) + 1) & (RX_RING_SIZE - 1))
drivers/net/ethernet/sun/sunbmac.h
256
#define PREV_RX(num) (((num) - 1) & (RX_RING_SIZE - 1))
drivers/net/ethernet/sun/sunbmac.h
299
struct sk_buff *rx_skbs[RX_RING_SIZE];
drivers/net/ethernet/sun/sungem.c
1210
writel(RX_RING_SIZE - 4, gp->regs + RXDMA_KICK);
drivers/net/ethernet/sun/sungem.c
1562
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/sun/sungem.c
1619
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/sun/sungem.c
1752
writel(RX_RING_SIZE - 4, gp->regs + RXDMA_KICK);
drivers/net/ethernet/sun/sungem.c
418
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/sun/sungem.c
435
writel(RX_RING_SIZE - 4, gp->regs + RXDMA_KICK);
drivers/net/ethernet/sun/sungem.c
769
if (work_done >= RX_RING_SIZE || work_done >= work_to_do)
drivers/net/ethernet/sun/sungem.h
1014
struct sk_buff *rx_skbs[RX_RING_SIZE];
drivers/net/ethernet/sun/sungem.h
908
#if RX_RING_SIZE == 32
drivers/net/ethernet/sun/sungem.h
910
#elif RX_RING_SIZE == 64
drivers/net/ethernet/sun/sungem.h
912
#elif RX_RING_SIZE == 128
drivers/net/ethernet/sun/sungem.h
914
#elif RX_RING_SIZE == 256
drivers/net/ethernet/sun/sungem.h
916
#elif RX_RING_SIZE == 512
drivers/net/ethernet/sun/sungem.h
918
#elif RX_RING_SIZE == 1024
drivers/net/ethernet/sun/sungem.h
920
#elif RX_RING_SIZE == 2048
drivers/net/ethernet/sun/sungem.h
922
#elif RX_RING_SIZE == 4096
drivers/net/ethernet/sun/sungem.h
924
#elif RX_RING_SIZE == 8192
drivers/net/ethernet/sun/sungem.h
931
#define NEXT_RX(N) (((N) + 1) & (RX_RING_SIZE - 1))
drivers/net/ethernet/sun/sungem.h
949
#if RX_RING_SIZE < 128
drivers/net/ethernet/sun/sungem.h
952
#define INIT_BLOCK_RX_RING_SIZE RX_RING_SIZE
drivers/net/ethernet/sun/sunhme.c
1161
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/sun/sunhme.c
1222
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/sun/sunhme.h
342
#if (RX_RING_SIZE == 32)
drivers/net/ethernet/sun/sunhme.h
345
#if (RX_RING_SIZE == 64)
drivers/net/ethernet/sun/sunhme.h
348
#if (RX_RING_SIZE == 128)
drivers/net/ethernet/sun/sunhme.h
351
#if (RX_RING_SIZE == 256)
drivers/net/ethernet/sun/sunhme.h
360
#define NEXT_RX(num) (((num) + 1) & (RX_RING_SIZE - 1))
drivers/net/ethernet/sun/sunhme.h
362
#define PREV_RX(num) (((num) - 1) & (RX_RING_SIZE - 1))
drivers/net/ethernet/sun/sunhme.h
417
struct sk_buff *rx_skbs[RX_RING_SIZE];
drivers/net/ethernet/sun/sunqe.c
134
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/sun/sunqe.c
426
&qbufs->rx_buf[elem & (RX_RING_SIZE - 1)][0];
drivers/net/ethernet/sun/sunqe.c
428
qebuf_offset(rx_buf, (elem & (RX_RING_SIZE - 1)));
drivers/net/ethernet/sun/sunqe.c
430
&rxbase[(elem+RX_RING_SIZE)&(RX_RING_MAXSIZE-1)];
drivers/net/ethernet/sun/sunqe.h
328
u8 rx_buf[RX_RING_SIZE][PKT_BUF_SZ];
drivers/net/ethernet/via/via-rhine.c
1146
RX_RING_SIZE * sizeof(struct rx_desc) +
drivers/net/ethernet/via/via-rhine.c
1161
RX_RING_SIZE * sizeof(struct rx_desc) +
drivers/net/ethernet/via/via-rhine.c
1169
rp->tx_ring = ring + RX_RING_SIZE * sizeof(struct rx_desc);
drivers/net/ethernet/via/via-rhine.c
1171
rp->tx_ring_dma = ring_dma + RX_RING_SIZE * sizeof(struct rx_desc);
drivers/net/ethernet/via/via-rhine.c
1182
RX_RING_SIZE * sizeof(struct rx_desc) +
drivers/net/ethernet/via/via-rhine.c
1227
for (i = 0; i < RX_RING_SIZE; i++)
drivers/net/ethernet/via/via-rhine.c
1253
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/via/via-rhine.c
1264
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/via/via-rhine.c
1288
for (i = 0; i < RX_RING_SIZE; i++) {
drivers/net/ethernet/via/via-rhine.c
2039
int entry = rp->cur_rx % RX_RING_SIZE;
drivers/net/ethernet/via/via-rhine.c
2139
entry = (++rp->cur_rx) % RX_RING_SIZE;
drivers/net/ethernet/via/via-rhine.c
442
struct sk_buff *rx_skbuff[RX_RING_SIZE];
drivers/net/ethernet/via/via-rhine.c
443
dma_addr_t rx_skbuff_dma[RX_RING_SIZE];