TX_RING_SIZE
for (i = 0; i < TX_RING_SIZE; i++) {
entry = cur_tx % TX_RING_SIZE;
static struct epic_tx_desc tx_ring[TX_RING_SIZE]
static unsigned char tx_packet[PKT_BUF_SZ * TX_RING_SIZE];
#define TX_RING_MOD_MASK (TX_RING_SIZE - 1)
static unsigned char txb[PKT_BUF_SZ * TX_RING_SIZE];
static struct pcnet32_tx_head tx_ring[TX_RING_SIZE]
struct sk_buff *tx_skbuff[TX_RING_SIZE];
for (i = 0; i < TX_RING_SIZE; i++) {
static struct netdev_desc tx_ring[TX_RING_SIZE];
static unsigned char txb[PKT_BUF_SZ * TX_RING_SIZE];
static struct tulip_tx_desc tx_ring[TX_RING_SIZE] __attribute__ ((aligned(4)));
static char desc1[TX_RING_SIZE * sizeof (struct rhine_tx_desc) + 32];
static char desc2[TX_RING_SIZE * sizeof (struct rhine_tx_desc) + 32];
for (i = 0; i < TX_RING_SIZE; i++)
entry = tp->cur_tx % TX_RING_SIZE;
char *tx_buffs[TX_RING_SIZE];
struct sk_buff *tx_skbuff[TX_RING_SIZE];
for (i = 0; i < TX_RING_SIZE; i++)
struct w840_tx_desc tx_ring[TX_RING_SIZE];
static char tx_packet[PKT_BUF_SZ * TX_RING_SIZE];
entry = w840private.cur_tx % TX_RING_SIZE;
if (entry >= TX_RING_SIZE-1) /* Wrap ring */
for (i = 0; i < TX_RING_SIZE; i++) {
gcp->gc_tx_ring_size = TX_RING_SIZE;
#ifndef TX_RING_SIZE
for (i = 0; i < TX_RING_SIZE; i++) {
TxRingSize = TX_RING_SIZE;
pMdl->TxRingSize = TX_RING_SIZE;
#define TX_RING_MOD_MASK (2 * TX_RING_SIZE - 1)
#define TX_RESCHEDULE_THRESHOLD (TX_RING_SIZE >> 1)
while ((pTx_desc->Tx_OWN == 0) && (desc_count < TX_RING_SIZE)) {
if (desc_count == TX_RING_SIZE) {
length = sizeof (struct tx_desc) * TX_RING_SIZE + ALIGNMENT;
pMil->pNonphysical->TxDescQEnd = &(pMil->Tx_desc[TX_RING_SIZE -1]);
for (i = 0; i < TX_RING_SIZE; i++) {