drivers/net/ethernet/3com/3c515.c
1000
int entry = vp->cur_tx % TX_RING_SIZE;
drivers/net/ethernet/3com/3c515.c
1008
prev_entry = &vp->tx_ring[(vp->cur_tx - 1) % TX_RING_SIZE];
drivers/net/ethernet/3com/3c515.c
1038
if (vp->cur_tx - vp->dirty_tx > TX_RING_SIZE - 1)
drivers/net/ethernet/3com/3c515.c
1163
int entry = dirty_tx % TX_RING_SIZE;
drivers/net/ethernet/3com/3c515.c
1174
if (lp->tx_full && (lp->cur_tx - dirty_tx <= TX_RING_SIZE - 1)) {
drivers/net/ethernet/3com/3c515.c
1445
for (i = 0; i < TX_RING_SIZE; i++)
drivers/net/ethernet/3com/3c515.c
299
struct boom_tx_desc tx_ring[TX_RING_SIZE];
drivers/net/ethernet/3com/3c515.c
302
struct sk_buff *tx_skbuff[TX_RING_SIZE];
drivers/net/ethernet/3com/3c515.c
832
for (i = 0; i < TX_RING_SIZE; i++)
drivers/net/ethernet/3com/3c515.c
970
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/3com/3c59x.c
1212
+ sizeof(struct boom_tx_desc) * TX_RING_SIZE,
drivers/net/ethernet/3com/3c59x.c
1478
sizeof(struct boom_tx_desc) * TX_RING_SIZE,
drivers/net/ethernet/3com/3c59x.c
1685
for (i = 0; i < TX_RING_SIZE; i++)
drivers/net/ethernet/3com/3c59x.c
1915
iowrite32(vp->tx_ring_dma + (vp->dirty_tx % TX_RING_SIZE) * sizeof(struct boom_tx_desc),
drivers/net/ethernet/3com/3c59x.c
1917
if (vp->cur_tx - vp->dirty_tx < TX_RING_SIZE) {
drivers/net/ethernet/3com/3c59x.c
2116
int entry = vp->cur_tx % TX_RING_SIZE;
drivers/net/ethernet/3com/3c59x.c
2118
struct boom_tx_desc *prev_entry = &vp->tx_ring[(vp->cur_tx-1) % TX_RING_SIZE];
drivers/net/ethernet/3com/3c59x.c
2137
if (vp->cur_tx - vp->dirty_tx >= TX_RING_SIZE) {
drivers/net/ethernet/3com/3c59x.c
2225
if (vp->cur_tx - vp->dirty_tx > TX_RING_SIZE - 1) {
drivers/net/ethernet/3com/3c59x.c
2422
int entry = dirty_tx % TX_RING_SIZE;
drivers/net/ethernet/3com/3c59x.c
2461
if (vp->cur_tx - dirty_tx <= TX_RING_SIZE - 1) {
drivers/net/ethernet/3com/3c59x.c
2764
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/3com/3c59x.c
2800
vp->dirty_tx, vp->dirty_tx % TX_RING_SIZE,
drivers/net/ethernet/3com/3c59x.c
2801
vp->cur_tx, vp->cur_tx % TX_RING_SIZE);
drivers/net/ethernet/3com/3c59x.c
2804
&vp->tx_ring[vp->dirty_tx % TX_RING_SIZE]);
drivers/net/ethernet/3com/3c59x.c
2806
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/3com/3c59x.c
3286
sizeof(struct boom_tx_desc) * TX_RING_SIZE,
drivers/net/ethernet/3com/3c59x.c
603
struct sk_buff* tx_skbuff[TX_RING_SIZE];
drivers/net/ethernet/adaptec/starfire.c
1165
for (i = 0; i < TX_RING_SIZE; i++)
drivers/net/ethernet/adaptec/starfire.c
1182
if ((np->cur_tx - np->dirty_tx) + skb_num_frags(skb) * 2 > TX_RING_SIZE) {
drivers/net/ethernet/adaptec/starfire.c
1195
entry = np->cur_tx % TX_RING_SIZE;
drivers/net/ethernet/adaptec/starfire.c
1203
if (entry >= TX_RING_SIZE - skb_num_frags(skb)) {
drivers/net/ethernet/adaptec/starfire.c
1242
np->tx_info[entry].used_slots = TX_RING_SIZE - entry;
drivers/net/ethernet/adaptec/starfire.c
1251
if (np->cur_tx % (TX_RING_SIZE / 2) == 0)
drivers/net/ethernet/adaptec/starfire.c
1264
if ((np->cur_tx - np->dirty_tx) + 4 > TX_RING_SIZE)
drivers/net/ethernet/adaptec/starfire.c
1270
entry = prev_tx % TX_RING_SIZE;
drivers/net/ethernet/adaptec/starfire.c
1277
entry = (entry + np->tx_info[entry].used_slots) % TX_RING_SIZE;
drivers/net/ethernet/adaptec/starfire.c
1362
entry = (entry + np->tx_info[entry].used_slots) % TX_RING_SIZE;
drivers/net/ethernet/adaptec/starfire.c
1383
(np->cur_tx - np->dirty_tx + 4 < TX_RING_SIZE)) {
drivers/net/ethernet/adaptec/starfire.c
1973
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/adaptec/starfire.c
523
struct tx_ring_info tx_info[TX_RING_SIZE];
drivers/net/ethernet/adaptec/starfire.c
880
tx_ring_size = ((sizeof(starfire_tx_desc) * TX_RING_SIZE + QUEUE_ALIGN - 1) / QUEUE_ALIGN) * QUEUE_ALIGN;
drivers/net/ethernet/amd/7990.c
104
for (t = 0; t < TX_RING_SIZE; t++) { \
drivers/net/ethernet/amd/7990.h
41
#define TX_RING_MOD_MASK (TX_RING_SIZE - 1)
drivers/net/ethernet/amd/7990.h
87
volatile struct lance_tx_desc btx_ring[TX_RING_SIZE];
drivers/net/ethernet/amd/7990.h
90
volatile char tx_buf[TX_RING_SIZE][TX_BUFF_SIZE];
drivers/net/ethernet/amd/a2065.c
100
char tx_buf[TX_RING_SIZE][TX_BUFF_SIZE];
drivers/net/ethernet/amd/a2065.c
75
#define TX_RING_MOD_MASK (TX_RING_SIZE - 1)
drivers/net/ethernet/amd/a2065.c
97
struct lance_tx_desc btx_ring[TX_RING_SIZE];
drivers/net/ethernet/amd/ariadne.c
100
struct TDRE tx_ring[TX_RING_SIZE];
drivers/net/ethernet/amd/ariadne.c
102
u_short tx_buff[TX_RING_SIZE][PKT_BUF_SIZE / sizeof(u_short)];
drivers/net/ethernet/amd/ariadne.c
129
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/amd/ariadne.c
309
int entry = dirty_tx % TX_RING_SIZE;
drivers/net/ethernet/amd/ariadne.c
345
if (priv->cur_tx - dirty_tx >= TX_RING_SIZE) {
drivers/net/ethernet/amd/ariadne.c
349
dirty_tx += TX_RING_SIZE;
drivers/net/ethernet/amd/ariadne.c
354
dirty_tx > priv->cur_tx - TX_RING_SIZE + 2) {
drivers/net/ethernet/amd/ariadne.c
470
lance->RDP = swapw(((u_short)-TX_RING_SIZE));
drivers/net/ethernet/amd/ariadne.c
576
entry = priv->cur_tx % TX_RING_SIZE;
drivers/net/ethernet/amd/ariadne.c
597
if ((priv->cur_tx >= TX_RING_SIZE) &&
drivers/net/ethernet/amd/ariadne.c
598
(priv->dirty_tx >= TX_RING_SIZE)) {
drivers/net/ethernet/amd/ariadne.c
603
priv->cur_tx -= TX_RING_SIZE;
drivers/net/ethernet/amd/ariadne.c
604
priv->dirty_tx -= TX_RING_SIZE;
drivers/net/ethernet/amd/ariadne.c
612
if (lowb(priv->tx_ring[(entry + 1) % TX_RING_SIZE]->TMD1) != 0) {
drivers/net/ethernet/amd/ariadne.c
88
volatile struct TDRE *tx_ring[TX_RING_SIZE];
drivers/net/ethernet/amd/ariadne.c
90
volatile u_short *tx_buff[TX_RING_SIZE];
drivers/net/ethernet/amd/atarilance.c
114
#define TX_RING_MOD_MASK (TX_RING_SIZE - 1)
drivers/net/ethernet/amd/atarilance.c
158
struct lance_tx_head tx_head[TX_RING_SIZE];
drivers/net/ethernet/amd/atarilance.c
705
for( i = 0; i < TX_RING_SIZE; i++ ) {
drivers/net/ethernet/amd/atarilance.c
756
for( i = 0 ; i < TX_RING_SIZE; i++ )
drivers/net/ethernet/amd/atarilance.c
830
while( lp->cur_tx >= TX_RING_SIZE && lp->dirty_tx >= TX_RING_SIZE ) {
drivers/net/ethernet/amd/atarilance.c
831
lp->cur_tx -= TX_RING_SIZE;
drivers/net/ethernet/amd/atarilance.c
832
lp->dirty_tx -= TX_RING_SIZE;
drivers/net/ethernet/amd/atarilance.c
921
if (lp->cur_tx - dirty_tx >= TX_RING_SIZE) {
drivers/net/ethernet/amd/atarilance.c
925
dirty_tx += TX_RING_SIZE;
drivers/net/ethernet/amd/atarilance.c
930
dirty_tx > lp->cur_tx - TX_RING_SIZE + 2) {
drivers/net/ethernet/amd/declance.c
1095
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/amd/declance.c
1140
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/amd/declance.c
1171
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/amd/declance.c
157
#define TX_RING_MOD_MASK (TX_RING_SIZE - 1)
drivers/net/ethernet/amd/declance.c
223
struct lance_tx_desc btx_ring[TX_RING_SIZE];
drivers/net/ethernet/amd/declance.c
268
char *tx_buf_ptr_cpu[TX_RING_SIZE];
drivers/net/ethernet/amd/declance.c
272
uint tx_buf_ptr_lnc[TX_RING_SIZE];
drivers/net/ethernet/amd/declance.c
495
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/amd/lance.c
1017
if ((lp->cur_tx - lp->dirty_tx) >= TX_RING_SIZE)
drivers/net/ethernet/amd/lance.c
1100
if (lp->cur_tx - dirty_tx >= TX_RING_SIZE) {
drivers/net/ethernet/amd/lance.c
1104
dirty_tx += TX_RING_SIZE;
drivers/net/ethernet/amd/lance.c
1110
dirty_tx > lp->cur_tx - TX_RING_SIZE + 2)
drivers/net/ethernet/amd/lance.c
195
#define TX_RING_MOD_MASK (TX_RING_SIZE - 1)
drivers/net/ethernet/amd/lance.c
239
struct lance_tx_head tx_ring[TX_RING_SIZE];
drivers/net/ethernet/amd/lance.c
243
struct sk_buff* tx_skbuff[TX_RING_SIZE];
drivers/net/ethernet/amd/lance.c
565
lp->tx_bounce_buffs = kmalloc_array(TX_RING_SIZE, PKT_BUF_SZ,
drivers/net/ethernet/amd/lance.c
856
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/amd/lance.c
893
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/amd/lance.c
942
for (i = 0; i < TX_RING_SIZE; i++)
drivers/net/ethernet/amd/pcnet32.c
1860
lp->tx_ring_size = TX_RING_SIZE; /* default tx ring size */
drivers/net/ethernet/amd/sun3lance.c
143
struct lance_tx_head tx_head[TX_RING_SIZE];
drivers/net/ethernet/amd/sun3lance.c
146
char tx_data[TX_RING_SIZE][PKT_BUF_SZ];
drivers/net/ethernet/amd/sun3lance.c
460
for( i = 0; i < TX_RING_SIZE; i++ ) {
drivers/net/ethernet/amd/sun3lance.c
548
for( i = 0 ; i < TX_RING_SIZE; i++ )
drivers/net/ethernet/amd/sun3lance.c
97
#define TX_RING_MOD_MASK (TX_RING_SIZE - 1)
drivers/net/ethernet/amd/sunlance.c
177
#define TX_RING_MOD_MASK (TX_RING_SIZE - 1)
drivers/net/ethernet/amd/sunlance.c
223
struct lance_tx_desc btx_ring[TX_RING_SIZE];
drivers/net/ethernet/amd/sunlance.c
225
u8 tx_buf [TX_RING_SIZE][TX_BUFF_SIZE];
drivers/net/ethernet/amd/sunlance.c
340
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/amd/sunlance.c
396
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/dec/tulip/interrupt.c
533
int maxtx = TX_RING_SIZE;
drivers/net/ethernet/dec/tulip/interrupt.c
534
int maxoi = TX_RING_SIZE;
drivers/net/ethernet/dec/tulip/interrupt.c
597
int entry = dirty_tx % TX_RING_SIZE;
drivers/net/ethernet/dec/tulip/interrupt.c
652
if (tp->cur_tx - dirty_tx > TX_RING_SIZE) {
drivers/net/ethernet/dec/tulip/interrupt.c
656
dirty_tx += TX_RING_SIZE;
drivers/net/ethernet/dec/tulip/interrupt.c
660
if (tp->cur_tx - dirty_tx < TX_RING_SIZE - 2)
drivers/net/ethernet/dec/tulip/tulip.h
409
struct ring_info tx_buffers[TX_RING_SIZE];
drivers/net/ethernet/dec/tulip/tulip_core.c
1129
if (tp->cur_tx - tp->dirty_tx > TX_RING_SIZE - 2) {
drivers/net/ethernet/dec/tulip/tulip_core.c
1137
entry = tp->cur_tx++ % TX_RING_SIZE;
drivers/net/ethernet/dec/tulip/tulip_core.c
1144
(entry == TX_RING_SIZE-1) ? cpu_to_le32(DESC_RING_WRAP) : 0;
drivers/net/ethernet/dec/tulip/tulip_core.c
1148
entry = tp->cur_tx++ % TX_RING_SIZE;
drivers/net/ethernet/dec/tulip/tulip_core.c
1159
if (entry == TX_RING_SIZE-1)
drivers/net/ethernet/dec/tulip/tulip_core.c
1167
if (tp->cur_tx - tp->dirty_tx >= TX_RING_SIZE - 2)
drivers/net/ethernet/dec/tulip/tulip_core.c
1431
sizeof(struct tulip_tx_desc) * TX_RING_SIZE,
drivers/net/ethernet/dec/tulip/tulip_core.c
587
for (i = 0; i < TX_RING_SIZE; i++)
drivers/net/ethernet/dec/tulip/tulip_core.c
643
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/dec/tulip/tulip_core.c
664
entry = tp->cur_tx % TX_RING_SIZE;
drivers/net/ethernet/dec/tulip/tulip_core.c
672
if (tp->cur_tx - tp->dirty_tx < TX_RING_SIZE/2) {/* Typical path */
drivers/net/ethernet/dec/tulip/tulip_core.c
674
} else if (tp->cur_tx - tp->dirty_tx == TX_RING_SIZE/2) {
drivers/net/ethernet/dec/tulip/tulip_core.c
676
} else if (tp->cur_tx - tp->dirty_tx < TX_RING_SIZE - 2) {
drivers/net/ethernet/dec/tulip/tulip_core.c
682
if (entry == TX_RING_SIZE-1)
drivers/net/ethernet/dec/tulip/tulip_core.c
707
int entry = dirty_tx % TX_RING_SIZE;
drivers/net/ethernet/dec/tulip/tulip_core.c
805
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/dec/tulip/winbond-840.c
1007
if(entry == TX_RING_SIZE-1)
drivers/net/ethernet/dec/tulip/winbond-840.c
1049
int entry = np->dirty_tx % TX_RING_SIZE;
drivers/net/ethernet/dec/tulip/winbond-840.c
1499
for (i = 0; i < TX_RING_SIZE; i++)
drivers/net/ethernet/dec/tulip/winbond-840.c
288
dma_addr_t tx_addr[TX_RING_SIZE];
drivers/net/ethernet/dec/tulip/winbond-840.c
293
struct sk_buff* tx_skbuff[TX_RING_SIZE];
drivers/net/ethernet/dec/tulip/winbond-840.c
816
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/dec/tulip/winbond-840.c
843
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/dec/tulip/winbond-840.c
927
for (i = 0; i < TX_RING_SIZE; i++)
drivers/net/ethernet/dec/tulip/winbond-840.c
966
sizeof(struct w840_tx_desc) * TX_RING_SIZE,
drivers/net/ethernet/dec/tulip/winbond-840.c
978
sizeof(struct w840_tx_desc) * TX_RING_SIZE,
drivers/net/ethernet/dec/tulip/winbond-840.c
992
entry = np->cur_tx % TX_RING_SIZE;
drivers/net/ethernet/dlink/dl2k.c
223
else if (tx_coalesce > TX_RING_SIZE-1)
drivers/net/ethernet/dlink/dl2k.c
224
tx_coalesce = TX_RING_SIZE - 1;
drivers/net/ethernet/dlink/dl2k.c
459
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/dlink/dl2k.c
480
for (i = 0; i < TX_RING_SIZE; i++)
drivers/net/ethernet/dlink/dl2k.c
497
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/dlink/dl2k.c
500
((i + 1) % TX_RING_SIZE) *
drivers/net/ethernet/dlink/dl2k.c
736
entry = np->cur_tx % TX_RING_SIZE;
drivers/net/ethernet/dlink/dl2k.c
772
np->cur_tx = (np->cur_tx + 1) % TX_RING_SIZE;
drivers/net/ethernet/dlink/dl2k.c
773
if ((np->cur_tx - np->old_tx + TX_RING_SIZE) % TX_RING_SIZE
drivers/net/ethernet/dlink/dl2k.c
834
int entry = np->old_tx % TX_RING_SIZE;
drivers/net/ethernet/dlink/dl2k.c
858
entry = (entry + 1) % TX_RING_SIZE;
drivers/net/ethernet/dlink/dl2k.c
870
((np->cur_tx - np->old_tx + TX_RING_SIZE) % TX_RING_SIZE
drivers/net/ethernet/dlink/dl2k.h
36
#define TX_QUEUE_LEN (TX_RING_SIZE - 1) /* Limit ring entries actually used.*/
drivers/net/ethernet/dlink/dl2k.h
369
struct sk_buff *tx_skbuff[TX_RING_SIZE];
drivers/net/ethernet/dlink/dl2k.h
38
#define TX_TOTAL_SIZE TX_RING_SIZE*sizeof(struct netdev_desc)
drivers/net/ethernet/dlink/sundance.c
1056
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/dlink/sundance.c
1065
unsigned head = np->cur_task % TX_RING_SIZE;
drivers/net/ethernet/dlink/sundance.c
1067
&np->tx_ring[(np->cur_tx - 1) % TX_RING_SIZE];
drivers/net/ethernet/dlink/sundance.c
1071
int entry = np->cur_task % TX_RING_SIZE;
drivers/net/ethernet/dlink/sundance.c
1096
entry = np->cur_tx % TX_RING_SIZE;
drivers/net/ethernet/dlink/sundance.c
1151
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/dlink/sundance.c
1262
int entry = np->dirty_tx % TX_RING_SIZE;
drivers/net/ethernet/dlink/sundance.c
1272
TX_RING_SIZE)
drivers/net/ethernet/dlink/sundance.c
1288
int entry = np->dirty_tx % TX_RING_SIZE;
drivers/net/ethernet/dlink/sundance.c
1874
for (i = 0; i < TX_RING_SIZE; i++)
drivers/net/ethernet/dlink/sundance.c
1905
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/dlink/sundance.c
366
struct sk_buff* tx_skbuff[TX_RING_SIZE];
drivers/net/ethernet/dlink/sundance.c
66
#define TX_QUEUE_LEN (TX_RING_SIZE - 1) /* Limit ring entries actually used. */
drivers/net/ethernet/dlink/sundance.c
69
#define TX_TOTAL_SIZE TX_RING_SIZE*sizeof(struct netdev_desc)
drivers/net/ethernet/dlink/sundance.c
977
for (i=0; i<TX_RING_SIZE; i++) {
drivers/net/ethernet/dlink/sundance.c
990
np->cur_tx, np->cur_tx % TX_RING_SIZE,
drivers/net/ethernet/dlink/sundance.c
991
np->dirty_tx, np->dirty_tx % TX_RING_SIZE);
drivers/net/ethernet/fealnx.c
1206
for (i = 0; i < TX_RING_SIZE; i++)
drivers/net/ethernet/fealnx.c
1274
np->free_tx_count = TX_RING_SIZE;
drivers/net/ethernet/fealnx.c
1276
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/fealnx.c
1382
np->free_tx_count = TX_RING_SIZE;
drivers/net/ethernet/fealnx.c
1384
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/fealnx.c
1400
np->tx_ring[TX_RING_SIZE - 1].next_desc = np->tx_ring_dma;
drivers/net/ethernet/fealnx.c
1401
np->tx_ring[TX_RING_SIZE - 1].next_desc_logical = &np->tx_ring[0];
drivers/net/ethernet/fealnx.c
1922
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/fealnx.c
58
#define TX_TOTAL_SIZE TX_RING_SIZE*sizeof(struct fealnx_desc)
drivers/net/ethernet/freescale/fec.h
543
unsigned char *tx_bounce[TX_RING_SIZE];
drivers/net/ethernet/freescale/fec.h
544
struct fec_tx_buffer tx_buf[TX_RING_SIZE];
drivers/net/ethernet/freescale/fec_main.c
3962
txq->bd.ring_size = TX_RING_SIZE;
drivers/net/ethernet/freescale/fs_enet/fs_enet-main.c
890
fpi->tx_ring = TX_RING_SIZE;
drivers/net/ethernet/freescale/fs_enet/fs_enet-main.c
96
int tx_left = TX_RING_SIZE;
drivers/net/ethernet/i825xx/82596.c
1076
if (++lp->next_tx_cmd == TX_RING_SIZE)
drivers/net/ethernet/i825xx/82596.c
333
struct tx_cmd tx_cmds[TX_RING_SIZE];
drivers/net/ethernet/i825xx/82596.c
334
struct i596_tbd tbds[TX_RING_SIZE];
drivers/net/ethernet/i825xx/82596.c
371
static int max_cmd_backlog = TX_RING_SIZE-1;
drivers/net/ethernet/i825xx/lib82596.c
1026
if (++lp->next_tx_cmd == TX_RING_SIZE)
drivers/net/ethernet/i825xx/lib82596.c
309
struct tx_cmd tx_cmds[TX_RING_SIZE] __attribute__((aligned(32)));
drivers/net/ethernet/i825xx/lib82596.c
310
struct i596_tbd tbds[TX_RING_SIZE] __attribute__((aligned(32)));
drivers/net/ethernet/i825xx/lib82596.c
361
static int max_cmd_backlog = TX_RING_SIZE-1;
drivers/net/ethernet/natsemi/natsemi.c
1865
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/natsemi/natsemi.c
1917
sizeof(struct netdev_desc) * (RX_RING_SIZE + TX_RING_SIZE),
drivers/net/ethernet/natsemi/natsemi.c
1975
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/natsemi/natsemi.c
1979
*((i+1)%TX_RING_SIZE+RX_RING_SIZE));
drivers/net/ethernet/natsemi/natsemi.c
2011
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/natsemi/natsemi.c
2052
sizeof(struct netdev_desc) * (RX_RING_SIZE + TX_RING_SIZE),
drivers/net/ethernet/natsemi/natsemi.c
2080
for (i=0;i<TX_RING_SIZE;i++)
drivers/net/ethernet/natsemi/natsemi.c
2097
entry = np->cur_tx % TX_RING_SIZE;
drivers/net/ethernet/natsemi/natsemi.c
2144
int entry = np->dirty_tx % TX_RING_SIZE;
drivers/net/ethernet/natsemi/natsemi.c
547
struct sk_buff *tx_skbuff[TX_RING_SIZE];
drivers/net/ethernet/natsemi/natsemi.c
548
dma_addr_t tx_dma[TX_RING_SIZE];
drivers/net/ethernet/packetengines/hamachi.c
1002
int entry = hmp->dirty_tx % TX_RING_SIZE;
drivers/net/ethernet/packetengines/hamachi.c
1017
if (entry >= TX_RING_SIZE-1)
drivers/net/ethernet/packetengines/hamachi.c
1018
hmp->tx_ring[TX_RING_SIZE-1].status_n_length |=
drivers/net/ethernet/packetengines/hamachi.c
1067
for (i = 0; i < TX_RING_SIZE; i++)
drivers/net/ethernet/packetengines/hamachi.c
1090
for (i = 0; i < TX_RING_SIZE; i++){
drivers/net/ethernet/packetengines/hamachi.c
1093
if (i >= TX_RING_SIZE - 1)
drivers/net/ethernet/packetengines/hamachi.c
1206
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/packetengines/hamachi.c
121
#define TX_TOTAL_SIZE TX_RING_SIZE*sizeof(struct hamachi_desc)
drivers/net/ethernet/packetengines/hamachi.c
1211
hmp->tx_ring[TX_RING_SIZE-1].status_n_length |= cpu_to_le32(DescEndRing);
drivers/net/ethernet/packetengines/hamachi.c
1243
entry = hmp->cur_tx % TX_RING_SIZE;
drivers/net/ethernet/packetengines/hamachi.c
1261
if (entry >= TX_RING_SIZE-1) /* Wrap ring */
drivers/net/ethernet/packetengines/hamachi.c
1286
if ((hmp->cur_tx - hmp->dirty_tx) < (TX_RING_SIZE - 4))
drivers/net/ethernet/packetengines/hamachi.c
1341
int entry = hmp->dirty_tx % TX_RING_SIZE;
drivers/net/ethernet/packetengines/hamachi.c
1357
if (entry >= TX_RING_SIZE-1)
drivers/net/ethernet/packetengines/hamachi.c
1358
hmp->tx_ring[TX_RING_SIZE-1].status_n_length |=
drivers/net/ethernet/packetengines/hamachi.c
1362
if (hmp->cur_tx - hmp->dirty_tx < TX_RING_SIZE - 4){
drivers/net/ethernet/packetengines/hamachi.c
1688
for (i = 0; i < TX_RING_SIZE; i++)
drivers/net/ethernet/packetengines/hamachi.c
1730
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/packetengines/hamachi.c
487
struct sk_buff* tx_skbuff[TX_RING_SIZE];
drivers/net/ethernet/packetengines/yellowfin.c
1005
if (yp->cur_tx - dirty_tx > TX_RING_SIZE) {
drivers/net/ethernet/packetengines/yellowfin.c
1008
dirty_tx += TX_RING_SIZE;
drivers/net/ethernet/packetengines/yellowfin.c
1020
yp->tx_tail_desc = &yp->tx_status[dirty_tx % TX_RING_SIZE];
drivers/net/ethernet/packetengines/yellowfin.c
1231
for (i = 0; i < TX_RING_SIZE*2; i++)
drivers/net/ethernet/packetengines/yellowfin.c
1237
for (i = 0; i < TX_RING_SIZE; i++)
drivers/net/ethernet/packetengines/yellowfin.c
1275
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/packetengines/yellowfin.c
312
struct sk_buff* tx_skbuff[TX_RING_SIZE];
drivers/net/ethernet/packetengines/yellowfin.c
705
for (i = 0; i < TX_RING_SIZE; i++)
drivers/net/ethernet/packetengines/yellowfin.c
76
#define STATUS_TOTAL_SIZE TX_RING_SIZE*sizeof(struct tx_status_words)
drivers/net/ethernet/packetengines/yellowfin.c
766
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/packetengines/yellowfin.c
77
#define TX_TOTAL_SIZE 2*TX_RING_SIZE*sizeof(struct yellowfin_desc)
drivers/net/ethernet/packetengines/yellowfin.c
770
((i+1)%TX_RING_SIZE)*sizeof(struct yellowfin_desc));
drivers/net/ethernet/packetengines/yellowfin.c
777
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/packetengines/yellowfin.c
803
((j+1)%(2*TX_RING_SIZE))*sizeof(struct yellowfin_desc));
drivers/net/ethernet/packetengines/yellowfin.c
826
entry = yp->cur_tx % TX_RING_SIZE;
drivers/net/ethernet/packetengines/yellowfin.c
847
if (entry >= TX_RING_SIZE-1) {
drivers/net/ethernet/packetengines/yellowfin.c
850
yp->tx_ring[TX_RING_SIZE-1].dbdma_cmd =
drivers/net/ethernet/packetengines/yellowfin.c
868
unsigned next_entry = yp->cur_tx % TX_RING_SIZE;
drivers/net/ethernet/packetengines/yellowfin.c
928
int entry = yp->dirty_tx % TX_RING_SIZE;
drivers/net/ethernet/packetengines/yellowfin.c
956
int entry = dirty_tx % TX_RING_SIZE;
drivers/net/ethernet/pasemi/pasemi_mac.c
1520
txring->next_to_fill = (fill + nfrags + 1) & (TX_RING_SIZE-1);
drivers/net/ethernet/pasemi/pasemi_mac.c
465
ring->size = TX_RING_SIZE;
drivers/net/ethernet/pasemi/pasemi_mac.c
466
ring->ring_info = kzalloc_objs(struct pasemi_mac_buffer, TX_RING_SIZE);
drivers/net/ethernet/pasemi/pasemi_mac.c
471
if (pasemi_dma_alloc_ring(&ring->chan, TX_RING_SIZE))
drivers/net/ethernet/pasemi/pasemi_mac.c
477
val |= PAS_DMA_TXCHAN_BASEU_SIZ(TX_RING_SIZE >> 3);
drivers/net/ethernet/pasemi/pasemi_mac.c
519
limit += TX_RING_SIZE;
drivers/net/ethernet/pasemi/pasemi_mac.c
522
info = &txring->ring_info[(i+1) & (TX_RING_SIZE-1)];
drivers/net/ethernet/pasemi/pasemi_mac.c
527
(TX_RING_SIZE-1)].dma;
drivers/net/ethernet/pasemi/pasemi_mac.c
837
ring_limit += TX_RING_SIZE;
drivers/net/ethernet/pasemi/pasemi_mac.c
885
txring->next_to_clean = i & (TX_RING_SIZE-1);
drivers/net/ethernet/pasemi/pasemi_mac.h
20
#define CS_RING_SIZE (TX_RING_SIZE*2)
drivers/net/ethernet/pasemi/pasemi_mac.h
94
#define TX_DESC(tx, num) ((tx)->chan.ring_virt[(num) & (TX_RING_SIZE-1)])
drivers/net/ethernet/pasemi/pasemi_mac.h
95
#define TX_DESC_INFO(tx, num) ((tx)->ring_info[(num) & (TX_RING_SIZE-1)])
drivers/net/ethernet/pasemi/pasemi_mac_ethtool.c
78
ering->tx_max_pending = TX_RING_SIZE/2;
drivers/net/ethernet/qlogic/qede/qede.h
597
#define NUM_TX_BDS_MAX (TX_RING_SIZE - 1)
drivers/net/ethernet/renesas/rswitch.h
37
#define TS_RING_SIZE (TX_RING_SIZE * RSWITCH_NUM_PORTS)
drivers/net/ethernet/renesas/rswitch_main.c
621
err = rswitch_gwca_queue_alloc(ndev, priv, rdev->tx_queue, true, TX_RING_SIZE);
drivers/net/ethernet/renesas/sh_eth.c
3279
mdp->num_tx_ring = TX_RING_SIZE;
drivers/net/ethernet/sgi/ioc3-eth.c
894
ip->tx_ring = dma_alloc_coherent(ip->dma_dev, TX_RING_SIZE + SZ_16K - 1,
drivers/net/ethernet/sgi/ioc3-eth.c
958
dma_free_coherent(ip->dma_dev, TX_RING_SIZE + SZ_16K - 1, ip->tx_ring,
drivers/net/ethernet/sgi/ioc3-eth.c
971
dma_free_coherent(ip->dma_dev, TX_RING_SIZE + SZ_16K - 1, ip->tx_ring, ip->txr_dma);
drivers/net/ethernet/smsc/epic100.c
1020
int entry = dirty_tx % TX_RING_SIZE;
drivers/net/ethernet/smsc/epic100.c
1043
if (cur_tx - dirty_tx > TX_RING_SIZE) {
drivers/net/ethernet/smsc/epic100.c
1046
dirty_tx += TX_RING_SIZE;
drivers/net/ethernet/smsc/epic100.c
1303
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/smsc/epic100.c
248
struct sk_buff* tx_skbuff[TX_RING_SIZE];
drivers/net/ethernet/smsc/epic100.c
54
#define TX_TOTAL_SIZE TX_RING_SIZE*sizeof(struct epic_tx_desc)
drivers/net/ethernet/smsc/epic100.c
802
(ep->dirty_tx % TX_RING_SIZE) * sizeof(struct epic_tx_desc));
drivers/net/ethernet/smsc/epic100.c
924
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/smsc/epic100.c
950
entry = ep->cur_tx % TX_RING_SIZE;
drivers/net/ethernet/smsc/smsc9420.c
1182
pd->tx_buffers = kmalloc_objs(struct smsc9420_ring_info, TX_RING_SIZE);
drivers/net/ethernet/smsc/smsc9420.c
1187
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/smsc/smsc9420.c
1195
pd->tx_ring[TX_RING_SIZE - 1].length = TDES1_TER_;
drivers/net/ethernet/smsc/smsc9420.c
1546
sizeof(struct smsc9420_dma_desc) * (RX_RING_SIZE + TX_RING_SIZE),
drivers/net/ethernet/smsc/smsc9420.c
1604
sizeof(struct smsc9420_dma_desc) * (RX_RING_SIZE + TX_RING_SIZE),
drivers/net/ethernet/smsc/smsc9420.c
1638
sizeof(struct smsc9420_dma_desc) * (RX_RING_SIZE + TX_RING_SIZE),
drivers/net/ethernet/smsc/smsc9420.c
498
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/smsc/smsc9420.c
921
pd->tx_ring_tail = (pd->tx_ring_tail + 1) % TX_RING_SIZE;
drivers/net/ethernet/smsc/smsc9420.c
933
(((pd->tx_ring_head + 2) % TX_RING_SIZE) == pd->tx_ring_tail);
drivers/net/ethernet/smsc/smsc9420.c
960
if (unlikely(index == (TX_RING_SIZE - 1)))
drivers/net/ethernet/smsc/smsc9420.c
968
pd->tx_ring_head = (pd->tx_ring_head + 1) % TX_RING_SIZE;
drivers/net/ethernet/sun/sunbmac.c
204
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/sun/sunbmac.c
249
for (i = 0; i < TX_RING_SIZE; i++)
drivers/net/ethernet/sun/sunbmac.h
255
#define NEXT_TX(num) (((num) + 1) & (TX_RING_SIZE - 1))
drivers/net/ethernet/sun/sunbmac.h
257
#define PREV_TX(num) (((num) - 1) & (TX_RING_SIZE - 1))
drivers/net/ethernet/sun/sunbmac.h
261
(bp)->tx_old + (TX_RING_SIZE - 1) - (bp)->tx_new : \
drivers/net/ethernet/sun/sunbmac.h
300
struct sk_buff *tx_skbs[TX_RING_SIZE];
drivers/net/ethernet/sun/sungem.c
1580
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/sun/sungem.c
1589
int ent = i & (TX_RING_SIZE - 1);
drivers/net/ethernet/sun/sungem.c
1643
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/sun/sungem.c
653
last &= (TX_RING_SIZE - 1);
drivers/net/ethernet/sun/sungem.c
973
if (!(entry & ((TX_RING_SIZE>>1)-1)))
drivers/net/ethernet/sun/sungem.h
1015
struct sk_buff *tx_skbs[TX_RING_SIZE];
drivers/net/ethernet/sun/sungem.h
886
#if TX_RING_SIZE == 32
drivers/net/ethernet/sun/sungem.h
888
#elif TX_RING_SIZE == 64
drivers/net/ethernet/sun/sungem.h
890
#elif TX_RING_SIZE == 128
drivers/net/ethernet/sun/sungem.h
892
#elif TX_RING_SIZE == 256
drivers/net/ethernet/sun/sungem.h
894
#elif TX_RING_SIZE == 512
drivers/net/ethernet/sun/sungem.h
896
#elif TX_RING_SIZE == 1024
drivers/net/ethernet/sun/sungem.h
898
#elif TX_RING_SIZE == 2048
drivers/net/ethernet/sun/sungem.h
900
#elif TX_RING_SIZE == 4096
drivers/net/ethernet/sun/sungem.h
902
#elif TX_RING_SIZE == 8192
drivers/net/ethernet/sun/sungem.h
930
#define NEXT_TX(N) (((N) + 1) & (TX_RING_SIZE - 1))
drivers/net/ethernet/sun/sungem.h
935
(GP)->tx_old + (TX_RING_SIZE - 1) - (GP)->tx_new : \
drivers/net/ethernet/sun/sungem.h
943
#if TX_RING_SIZE < 128
drivers/net/ethernet/sun/sungem.h
946
#define INIT_BLOCK_TX_RING_SIZE TX_RING_SIZE
drivers/net/ethernet/sun/sunhme.c
1176
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/sun/sunhme.c
1249
for (i = 0; i < TX_RING_SIZE; i++)
drivers/net/ethernet/sun/sunhme.c
1443
HMD("tx rsize=%d oreg[%08x]\n", (int)TX_RING_SIZE,
drivers/net/ethernet/sun/sunhme.c
1445
hme_write32(hp, etxregs + ETX_RSIZE, (TX_RING_SIZE >> ETX_RSIZE_SHIFT) - 1);
drivers/net/ethernet/sun/sunhme.c
1719
last &= (TX_RING_SIZE - 1);
drivers/net/ethernet/sun/sunhme.h
334
#if (TX_RING_SIZE < 16 || TX_RING_SIZE > 256 || (TX_RING_SIZE % 16) != 0)
drivers/net/ethernet/sun/sunhme.h
361
#define NEXT_TX(num) (((num) + 1) & (TX_RING_SIZE - 1))
drivers/net/ethernet/sun/sunhme.h
363
#define PREV_TX(num) (((num) - 1) & (TX_RING_SIZE - 1))
drivers/net/ethernet/sun/sunhme.h
367
(hp)->tx_old + (TX_RING_SIZE - 1) - (hp)->tx_new : \
drivers/net/ethernet/sun/sunhme.h
418
struct sk_buff *tx_skbs[TX_RING_SIZE];
drivers/net/ethernet/sun/sunqe.c
588
txbuf = &qbufs->tx_buf[entry & (TX_RING_SIZE - 1)][0];
drivers/net/ethernet/sun/sunqe.c
590
qebuf_offset(tx_buf, (entry & (TX_RING_SIZE - 1)));
drivers/net/ethernet/sun/sunqe.h
301
(qp)->tx_old + (TX_RING_SIZE - 1) - (qp)->tx_new : \
drivers/net/ethernet/sun/sunqe.h
326
u8 tx_buf[TX_RING_SIZE][PKT_BUF_SZ];
drivers/net/ethernet/via/via-rhine.c
1147
TX_RING_SIZE * sizeof(struct tx_desc),
drivers/net/ethernet/via/via-rhine.c
1156
PKT_BUF_SZ * TX_RING_SIZE,
drivers/net/ethernet/via/via-rhine.c
1162
TX_RING_SIZE * sizeof(struct tx_desc),
drivers/net/ethernet/via/via-rhine.c
1183
TX_RING_SIZE * sizeof(struct tx_desc),
drivers/net/ethernet/via/via-rhine.c
1188
dma_free_coherent(hwdev, PKT_BUF_SZ * TX_RING_SIZE,
drivers/net/ethernet/via/via-rhine.c
1309
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/via/via-rhine.c
1329
for (i = 0; i < TX_RING_SIZE; i++) {
drivers/net/ethernet/via/via-rhine.c
1787
entry = rp->cur_tx % TX_RING_SIZE;
drivers/net/ethernet/via/via-rhine.c
1937
unsigned int entry = dirty_tx % TX_RING_SIZE;
drivers/net/ethernet/via/via-rhine.c
2152
int entry = rp->dirty_tx % TX_RING_SIZE;
drivers/net/ethernet/via/via-rhine.c
446
struct sk_buff *tx_skbuff[TX_RING_SIZE];
drivers/net/ethernet/via/via-rhine.c
447
dma_addr_t tx_skbuff_dma[TX_RING_SIZE];
drivers/net/ethernet/via/via-rhine.c
450
unsigned char *tx_buf[TX_RING_SIZE];
drivers/net/ethernet/via/via-rhine.c
78
#define TX_QUEUE_LEN (TX_RING_SIZE - 6) /* Limit ring entries actually used. */