drivers/cdrom/cdrom.c
684
struct rwrt_feature_desc *rfd)
drivers/cdrom/cdrom.c
700
memcpy(rfd, &buffer[sizeof(struct feature_header)], sizeof (*rfd));
drivers/cdrom/cdrom.c
731
struct rwrt_feature_desc rfd;
drivers/cdrom/cdrom.c
736
if ((ret = cdrom_get_random_writable(cdi, &rfd)))
drivers/cdrom/cdrom.c
739
if (CDF_RWRT == be16_to_cpu(rfd.feature_code))
drivers/cdrom/cdrom.c
843
struct rwrt_feature_desc rfd;
drivers/cdrom/cdrom.c
849
if ((ret = cdrom_get_random_writable(cdi, &rfd)))
drivers/cdrom/cdrom.c
851
else if (CDF_RWRT == be16_to_cpu(rfd.feature_code))
drivers/cdrom/cdrom.c
852
ret = !rfd.curr;
drivers/net/ethernet/agere/et131x.c
2008
struct rfd *rfd;
drivers/net/ethernet/agere/et131x.c
2016
rfd = list_entry(rx_ring->recv_list.next,
drivers/net/ethernet/agere/et131x.c
2017
struct rfd, list_node);
drivers/net/ethernet/agere/et131x.c
2019
list_del(&rfd->list_node);
drivers/net/ethernet/agere/et131x.c
2020
rfd->skb = NULL;
drivers/net/ethernet/agere/et131x.c
2021
kfree(rfd);
drivers/net/ethernet/agere/et131x.c
2086
struct rfd *rfd;
drivers/net/ethernet/agere/et131x.c
2092
rfd = kzalloc_obj(*rfd, GFP_ATOMIC | GFP_DMA);
drivers/net/ethernet/agere/et131x.c
2093
if (!rfd)
drivers/net/ethernet/agere/et131x.c
2096
rfd->skb = NULL;
drivers/net/ethernet/agere/et131x.c
2099
list_add_tail(&rfd->list_node, &rx_ring->recv_list);
drivers/net/ethernet/agere/et131x.c
2123
static void nic_return_rfd(struct et131x_adapter *adapter, struct rfd *rfd)
drivers/net/ethernet/agere/et131x.c
2127
u16 buff_index = rfd->bufferindex;
drivers/net/ethernet/agere/et131x.c
2128
u8 ring_index = rfd->ringindex;
drivers/net/ethernet/agere/et131x.c
2168
list_add_tail(&rfd->list_node, &rx_local->recv_list);
drivers/net/ethernet/agere/et131x.c
2182
static struct rfd *nic_rx_pkts(struct et131x_adapter *adapter)
drivers/net/ethernet/agere/et131x.c
2187
struct rfd *rfd;
drivers/net/ethernet/agere/et131x.c
2245
rfd = list_entry(element, struct rfd, list_node);
drivers/net/ethernet/agere/et131x.c
2247
if (!rfd) {
drivers/net/ethernet/agere/et131x.c
2252
list_del(&rfd->list_node);
drivers/net/ethernet/agere/et131x.c
2257
rfd->bufferindex = buff_index;
drivers/net/ethernet/agere/et131x.c
2258
rfd->ringindex = ring_index;
drivers/net/ethernet/agere/et131x.c
2266
rfd->len = 0;
drivers/net/ethernet/agere/et131x.c
2273
rfd->len = len;
drivers/net/ethernet/agere/et131x.c
2275
skb = dev_alloc_skb(rfd->len + 2);
drivers/net/ethernet/agere/et131x.c
2279
adapter->netdev->stats.rx_bytes += rfd->len;
drivers/net/ethernet/agere/et131x.c
2281
skb_put_data(skb, fbr->virt[buff_index], rfd->len);
drivers/net/ethernet/agere/et131x.c
2288
nic_return_rfd(adapter, rfd);
drivers/net/ethernet/agere/et131x.c
2289
return rfd;
drivers/net/ethernet/agere/et131x.c
2294
struct rfd *rfd = NULL;
drivers/net/ethernet/agere/et131x.c
2311
rfd = nic_rx_pkts(adapter);
drivers/net/ethernet/agere/et131x.c
2313
if (rfd == NULL)
drivers/net/ethernet/agere/et131x.c
2323
rfd->len == 0)
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2200
unsigned int rfa, rfd;
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2207
rfd = rfa + frame_fifo_size;
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2208
if (rfd > XGMAC_FLOW_CONTROL_MAX)
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2209
rfd = XGMAC_FLOW_CONTROL_MAX;
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2245
rfd = rfa + (frame_fifo_size / 2);
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2252
rfd = rfa + frame_fifo_size;
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2257
pdata->rx_rfd[queue] = XGMAC_FLOW_CONTROL_VALUE(rfd);
drivers/net/ethernet/atheros/alx/alx.h
60
struct alx_rfd *rfd;
drivers/net/ethernet/atheros/alx/main.c
122
rfd->addr = cpu_to_le64(dma);
drivers/net/ethernet/atheros/alx/main.c
641
rxq->rfd = alx->descmem.virt + offset;
drivers/net/ethernet/atheros/alx/main.c
85
struct alx_rfd *rfd = &rxq->rfd[cur];
drivers/net/ethernet/i825xx/82596.c
1207
lp->scb.rfd = I596_NULL;
drivers/net/ethernet/i825xx/82596.c
293
struct i596_rfd *rfd;
drivers/net/ethernet/i825xx/82596.c
459
struct i596_rfd *rfd;
drivers/net/ethernet/i825xx/82596.c
469
lp->scb.cmd, lp->scb.rfd);
drivers/net/ethernet/i825xx/82596.c
480
rfd = lp->rfd_head;
drivers/net/ethernet/i825xx/82596.c
481
printk(KERN_ERR "rfd_head = %p\n", rfd);
drivers/net/ethernet/i825xx/82596.c
485
rfd, rfd->stat, rfd->cmd, rfd->b_next, rfd->rbd,
drivers/net/ethernet/i825xx/82596.c
486
rfd->count);
drivers/net/ethernet/i825xx/82596.c
487
rfd = rfd->v_next;
drivers/net/ethernet/i825xx/82596.c
488
} while (rfd != lp->rfd_head);
drivers/net/ethernet/i825xx/82596.c
543
struct i596_rfd *rfd;
drivers/net/ethernet/i825xx/82596.c
574
for (i = 0, rfd = lp->rfds; i < rx_ring_size; i++, rfd++) {
drivers/net/ethernet/i825xx/82596.c
575
rfd->rbd = I596_NULL;
drivers/net/ethernet/i825xx/82596.c
576
rfd->v_next = rfd+1;
drivers/net/ethernet/i825xx/82596.c
577
rfd->v_prev = rfd-1;
drivers/net/ethernet/i825xx/82596.c
578
rfd->b_next = WSWAPrfd(virt_to_bus(rfd+1));
drivers/net/ethernet/i825xx/82596.c
579
rfd->cmd = CMD_FLEX;
drivers/net/ethernet/i825xx/82596.c
582
lp->scb.rfd = WSWAPrfd(virt_to_bus(lp->rfds));
drivers/net/ethernet/i825xx/82596.c
583
rfd = lp->rfds;
drivers/net/ethernet/i825xx/82596.c
584
rfd->rbd = lp->rbd_head;
drivers/net/ethernet/i825xx/82596.c
585
rfd->v_prev = lp->rfds + rx_ring_size - 1;
drivers/net/ethernet/i825xx/82596.c
586
rfd = lp->rfds + rx_ring_size - 1;
drivers/net/ethernet/i825xx/82596.c
587
rfd->v_next = lp->rfds;
drivers/net/ethernet/i825xx/82596.c
588
rfd->b_next = WSWAPrfd(virt_to_bus(lp->rfds));
drivers/net/ethernet/i825xx/82596.c
589
rfd->cmd = CMD_EOL|CMD_FLEX;
drivers/net/ethernet/i825xx/82596.c
608
lp->scb.rfd = WSWAPrfd(virt_to_bus(lp->rfds));
drivers/net/ethernet/i825xx/82596.c
747
struct i596_rfd *rfd;
drivers/net/ethernet/i825xx/82596.c
754
rfd = lp->rfd_head; /* Ref next frame to check */
drivers/net/ethernet/i825xx/82596.c
756
while ((rfd->stat) & STAT_C) { /* Loop while complete frames */
drivers/net/ethernet/i825xx/82596.c
757
if (rfd->rbd == I596_NULL)
drivers/net/ethernet/i825xx/82596.c
759
else if (rfd->rbd == lp->rbd_head->b_addr)
drivers/net/ethernet/i825xx/82596.c
767
rfd, rfd->rbd, rfd->stat));
drivers/net/ethernet/i825xx/82596.c
769
if (rbd != I596_NULL && ((rfd->stat) & STAT_OK)) {
drivers/net/ethernet/i825xx/82596.c
827
dev->name, rfd->stat));
drivers/net/ethernet/i825xx/82596.c
829
if ((rfd->stat) & 0x0001)
drivers/net/ethernet/i825xx/82596.c
831
if ((rfd->stat) & 0x0080)
drivers/net/ethernet/i825xx/82596.c
833
if ((rfd->stat) & 0x0100)
drivers/net/ethernet/i825xx/82596.c
835
if ((rfd->stat) & 0x0200)
drivers/net/ethernet/i825xx/82596.c
837
if ((rfd->stat) & 0x0400)
drivers/net/ethernet/i825xx/82596.c
839
if ((rfd->stat) & 0x0800)
drivers/net/ethernet/i825xx/82596.c
841
if ((rfd->stat) & 0x1000)
drivers/net/ethernet/i825xx/82596.c
854
rfd->rbd = I596_NULL;
drivers/net/ethernet/i825xx/82596.c
855
rfd->stat = 0;
drivers/net/ethernet/i825xx/82596.c
856
rfd->cmd = CMD_EOL|CMD_FLEX;
drivers/net/ethernet/i825xx/82596.c
857
rfd->count = 0;
drivers/net/ethernet/i825xx/82596.c
861
rfd->v_prev->cmd = CMD_FLEX;
drivers/net/ethernet/i825xx/82596.c
865
lp->scb.rfd = rfd->b_next;
drivers/net/ethernet/i825xx/82596.c
866
lp->rfd_head = rfd->v_next;
drivers/net/ethernet/i825xx/82596.c
867
rfd = lp->rfd_head;
drivers/net/ethernet/i825xx/lib82596.c
1098
lp->dma->scb.rfd = I596_NULL;
drivers/net/ethernet/i825xx/lib82596.c
277
u32 rfd;
drivers/net/ethernet/i825xx/lib82596.c
438
struct i596_rfd *rfd;
drivers/net/ethernet/i825xx/lib82596.c
448
SWAP16(dma->scb.cmd), SWAP32(dma->scb.rfd));
drivers/net/ethernet/i825xx/lib82596.c
463
rfd = lp->rfd_head;
drivers/net/ethernet/i825xx/lib82596.c
464
printk(KERN_DEBUG "rfd_head = %p\n", rfd);
drivers/net/ethernet/i825xx/lib82596.c
469
rfd, SWAP16(rfd->stat), SWAP16(rfd->cmd),
drivers/net/ethernet/i825xx/lib82596.c
470
SWAP32(rfd->b_next), SWAP32(rfd->rbd),
drivers/net/ethernet/i825xx/lib82596.c
471
SWAP16(rfd->count));
drivers/net/ethernet/i825xx/lib82596.c
472
rfd = rfd->v_next;
drivers/net/ethernet/i825xx/lib82596.c
473
} while (rfd != lp->rfd_head);
drivers/net/ethernet/i825xx/lib82596.c
492
struct i596_rfd *rfd;
drivers/net/ethernet/i825xx/lib82596.c
521
for (i = 0, rfd = dma->rfds; i < rx_ring_size; i++, rfd++) {
drivers/net/ethernet/i825xx/lib82596.c
522
rfd->rbd = I596_NULL;
drivers/net/ethernet/i825xx/lib82596.c
523
rfd->v_next = rfd+1;
drivers/net/ethernet/i825xx/lib82596.c
524
rfd->v_prev = rfd-1;
drivers/net/ethernet/i825xx/lib82596.c
525
rfd->b_next = SWAP32(virt_to_dma(lp, rfd+1));
drivers/net/ethernet/i825xx/lib82596.c
526
rfd->cmd = SWAP16(CMD_FLEX);
drivers/net/ethernet/i825xx/lib82596.c
529
dma->scb.rfd = SWAP32(virt_to_dma(lp, dma->rfds));
drivers/net/ethernet/i825xx/lib82596.c
530
rfd = dma->rfds;
drivers/net/ethernet/i825xx/lib82596.c
531
rfd->rbd = SWAP32(virt_to_dma(lp, lp->rbd_head));
drivers/net/ethernet/i825xx/lib82596.c
532
rfd->v_prev = dma->rfds + rx_ring_size - 1;
drivers/net/ethernet/i825xx/lib82596.c
533
rfd = dma->rfds + rx_ring_size - 1;
drivers/net/ethernet/i825xx/lib82596.c
534
rfd->v_next = dma->rfds;
drivers/net/ethernet/i825xx/lib82596.c
535
rfd->b_next = SWAP32(virt_to_dma(lp, dma->rfds));
drivers/net/ethernet/i825xx/lib82596.c
536
rfd->cmd = SWAP16(CMD_EOL|CMD_FLEX);
drivers/net/ethernet/i825xx/lib82596.c
573
dma->scb.rfd = SWAP32(virt_to_dma(lp, dma->rfds));
drivers/net/ethernet/i825xx/lib82596.c
654
dma->scb.rfd = SWAP32(virt_to_dma(lp, dma->rfds));
drivers/net/ethernet/i825xx/lib82596.c
678
struct i596_rfd *rfd;
drivers/net/ethernet/i825xx/lib82596.c
687
rfd = lp->rfd_head; /* Ref next frame to check */
drivers/net/ethernet/i825xx/lib82596.c
689
dma_sync_cpu(dev, rfd, sizeof(struct i596_rfd));
drivers/net/ethernet/i825xx/lib82596.c
690
while (rfd->stat & SWAP16(STAT_C)) { /* Loop while complete frames */
drivers/net/ethernet/i825xx/lib82596.c
691
if (rfd->rbd == I596_NULL)
drivers/net/ethernet/i825xx/lib82596.c
693
else if (rfd->rbd == lp->rbd_head->b_addr) {
drivers/net/ethernet/i825xx/lib82596.c
703
rfd, rfd->rbd, rfd->stat));
drivers/net/ethernet/i825xx/lib82596.c
705
if (rbd != NULL && (rfd->stat & SWAP16(STAT_OK))) {
drivers/net/ethernet/i825xx/lib82596.c
772
dev->name, rfd->stat));
drivers/net/ethernet/i825xx/lib82596.c
774
if (rfd->stat & SWAP16(0x0100))
drivers/net/ethernet/i825xx/lib82596.c
776
if (rfd->stat & SWAP16(0x8000))
drivers/net/ethernet/i825xx/lib82596.c
778
if (rfd->stat & SWAP16(0x0001))
drivers/net/ethernet/i825xx/lib82596.c
780
if (rfd->stat & SWAP16(0x0002))
drivers/net/ethernet/i825xx/lib82596.c
782
if (rfd->stat & SWAP16(0x0004))
drivers/net/ethernet/i825xx/lib82596.c
784
if (rfd->stat & SWAP16(0x0008))
drivers/net/ethernet/i825xx/lib82596.c
786
if (rfd->stat & SWAP16(0x0010))
drivers/net/ethernet/i825xx/lib82596.c
800
rfd->rbd = I596_NULL;
drivers/net/ethernet/i825xx/lib82596.c
801
rfd->stat = 0;
drivers/net/ethernet/i825xx/lib82596.c
802
rfd->cmd = SWAP16(CMD_EOL|CMD_FLEX);
drivers/net/ethernet/i825xx/lib82596.c
803
rfd->count = 0;
drivers/net/ethernet/i825xx/lib82596.c
807
lp->dma->scb.rfd = rfd->b_next;
drivers/net/ethernet/i825xx/lib82596.c
808
lp->rfd_head = rfd->v_next;
drivers/net/ethernet/i825xx/lib82596.c
809
dma_sync_dev(dev, rfd, sizeof(struct i596_rfd));
drivers/net/ethernet/i825xx/lib82596.c
813
rfd->v_prev->cmd = SWAP16(CMD_FLEX);
drivers/net/ethernet/i825xx/lib82596.c
814
dma_sync_dev(dev, rfd->v_prev, sizeof(struct i596_rfd));
drivers/net/ethernet/i825xx/lib82596.c
815
rfd = lp->rfd_head;
drivers/net/ethernet/i825xx/lib82596.c
816
dma_sync_cpu(dev, rfd, sizeof(struct i596_rfd));
drivers/net/ethernet/i825xx/sun3_82586.c
638
volatile struct rfd_struct *rfd = (struct rfd_struct *)ptr;
drivers/net/ethernet/i825xx/sun3_82586.c
643
memset((char *) rfd,0,sizeof(struct rfd_struct)*(p->num_recv_buffs+rfdadd));
drivers/net/ethernet/i825xx/sun3_82586.c
644
p->rfd_first = rfd;
drivers/net/ethernet/i825xx/sun3_82586.c
647
rfd[i].next = make16(rfd + (i+1) % (p->num_recv_buffs+rfdadd) );
drivers/net/ethernet/i825xx/sun3_82586.c
648
rfd[i].rbd_offset = 0xffff;
drivers/net/ethernet/i825xx/sun3_82586.c
650
rfd[p->num_recv_buffs-1+rfdadd].last = RFD_SUSP; /* RU suspend */
drivers/net/ethernet/i825xx/sun3_82586.c
652
ptr = (void *) (rfd + (p->num_recv_buffs + rfdadd) );
drivers/net/ethernet/intel/e100.c
1915
#define RFD_BUF_LEN (sizeof(struct rfd) + VLAN_ETH_FRAME_LEN + ETH_FCS_LEN)
drivers/net/ethernet/intel/e100.c
1922
skb_copy_to_linear_data(rx->skb, &nic->blank_rfd, sizeof(struct rfd));
drivers/net/ethernet/intel/e100.c
1937
struct rfd *prev_rfd = (struct rfd *)rx->prev->skb->data;
drivers/net/ethernet/intel/e100.c
1941
sizeof(struct rfd),
drivers/net/ethernet/intel/e100.c
1953
struct rfd *rfd = (struct rfd *)skb->data;
drivers/net/ethernet/intel/e100.c
1962
sizeof(struct rfd), DMA_BIDIRECTIONAL);
drivers/net/ethernet/intel/e100.c
1963
rfd_status = le16_to_cpu(rfd->status);
drivers/net/ethernet/intel/e100.c
1976
if ((le16_to_cpu(rfd->command) & cb_el) &&
drivers/net/ethernet/intel/e100.c
1982
sizeof(struct rfd),
drivers/net/ethernet/intel/e100.c
1990
actual_size = le16_to_cpu(rfd->actual_size) & 0x3FFF;
drivers/net/ethernet/intel/e100.c
1991
if (unlikely(actual_size > RFD_BUF_LEN - sizeof(struct rfd)))
drivers/net/ethernet/intel/e100.c
1992
actual_size = RFD_BUF_LEN - sizeof(struct rfd);
drivers/net/ethernet/intel/e100.c
2004
if ((le16_to_cpu(rfd->command) & cb_el) &&
drivers/net/ethernet/intel/e100.c
2012
skb_reserve(skb, sizeof(struct rfd));
drivers/net/ethernet/intel/e100.c
2053
struct rfd *old_before_last_rfd, *new_before_last_rfd;
drivers/net/ethernet/intel/e100.c
2074
old_before_last_rfd = (struct rfd *)old_before_last_rx->skb->data;
drivers/net/ethernet/intel/e100.c
2094
(struct rfd *)new_before_last_rx->skb->data;
drivers/net/ethernet/intel/e100.c
2099
sizeof(struct rfd),
drivers/net/ethernet/intel/e100.c
2108
sizeof(struct rfd),
drivers/net/ethernet/intel/e100.c
2114
sizeof(struct rfd),
drivers/net/ethernet/intel/e100.c
2154
struct rfd *before_last;
drivers/net/ethernet/intel/e100.c
2178
before_last = (struct rfd *)rx->skb->data;
drivers/net/ethernet/intel/e100.c
2182
sizeof(struct rfd), DMA_BIDIRECTIONAL);
drivers/net/ethernet/intel/e100.c
2384
if (memcmp(nic->rx_to_clean->skb->data + sizeof(struct rfd),
drivers/net/ethernet/intel/e100.c
546
struct rfd blank_rfd;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1057
struct emac_buffer *rfbuf = rx_q->rfd.rfbuff;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1063
if (++consume_idx == rx_q->rfd.count)
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1067
rx_q->rfd.consume_idx = consume_idx;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1068
rx_q->rfd.process_idx = consume_idx;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1159
proc_idx = (rx_q->rfd.process_idx << rx_q->process_shft) &
drivers/net/ethernet/qualcomm/emac/emac-mac.c
235
#define EMAC_RFD(RXQ, SIZE, IDX) ((RXQ)->rfd.v_addr + (SIZE * (IDX)))
drivers/net/ethernet/qualcomm/emac/emac-mac.c
238
#define GET_RFD_BUFFER(RXQ, IDX) (&((RXQ)->rfd.rfbuff[(IDX)]))
drivers/net/ethernet/qualcomm/emac/emac-mac.c
315
writel(upper_32_bits(adpt->rx_q.rfd.dma_addr),
drivers/net/ethernet/qualcomm/emac/emac-mac.c
318
writel(lower_32_bits(adpt->rx_q.rfd.dma_addr),
drivers/net/ethernet/qualcomm/emac/emac-mac.c
323
writel(adpt->rx_q.rfd.count & RFD_RING_SIZE_BMSK,
drivers/net/ethernet/qualcomm/emac/emac-mac.c
629
if (!rx_q->rfd.rfbuff)
drivers/net/ethernet/qualcomm/emac/emac-mac.c
632
for (i = 0; i < rx_q->rfd.count; i++) {
drivers/net/ethernet/qualcomm/emac/emac-mac.c
646
size = sizeof(struct emac_buffer) * rx_q->rfd.count;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
647
memset(rx_q->rfd.rfbuff, 0, size);
drivers/net/ethernet/qualcomm/emac/emac-mac.c
654
memset(rx_q->rfd.v_addr, 0, rx_q->rfd.size);
drivers/net/ethernet/qualcomm/emac/emac-mac.c
655
rx_q->rfd.produce_idx = 0;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
656
rx_q->rfd.consume_idx = 0;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
703
kfree(rx_q->rfd.rfbuff);
drivers/net/ethernet/qualcomm/emac/emac-mac.c
704
rx_q->rfd.rfbuff = NULL;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
706
rx_q->rfd.v_addr = NULL;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
707
rx_q->rfd.dma_addr = 0;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
708
rx_q->rfd.size = 0;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
723
size = sizeof(struct emac_buffer) * rx_q->rfd.count;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
724
rx_q->rfd.rfbuff = kzalloc_node(size, GFP_KERNEL, node);
drivers/net/ethernet/qualcomm/emac/emac-mac.c
725
if (!rx_q->rfd.rfbuff)
drivers/net/ethernet/qualcomm/emac/emac-mac.c
729
rx_q->rfd.size = rx_q->rfd.count * (adpt->rfd_size * 4);
drivers/net/ethernet/qualcomm/emac/emac-mac.c
735
rx_q->rfd.dma_addr = ring_header->dma_addr + ring_header->used;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
736
rx_q->rfd.v_addr = ring_header->v_addr + ring_header->used;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
737
ring_header->used += ALIGN(rx_q->rfd.size, 8);
drivers/net/ethernet/qualcomm/emac/emac-mac.c
742
rx_q->rfd.produce_idx = 0;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
743
rx_q->rfd.consume_idx = 0;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
760
adpt->rx_q.rfd.count = adpt->rx_desc_cnt;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
838
adpt->rx_q.rfd.produce_idx = 0;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
839
adpt->rx_q.rfd.consume_idx = 0;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
840
for (i = 0; i < adpt->rx_q.rfd.count; i++)
drivers/net/ethernet/qualcomm/emac/emac-mac.c
841
adpt->rx_q.rfd.rfbuff[i].dma_addr = 0;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
849
u32 *hw_rfd = EMAC_RFD(rx_q, adpt->rfd_size, rx_q->rfd.produce_idx);
drivers/net/ethernet/qualcomm/emac/emac-mac.c
854
if (++rx_q->rfd.produce_idx == rx_q->rfd.count)
drivers/net/ethernet/qualcomm/emac/emac-mac.c
855
rx_q->rfd.produce_idx = 0;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
867
next_produce_idx = rx_q->rfd.produce_idx + 1;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
868
if (next_produce_idx == rx_q->rfd.count)
drivers/net/ethernet/qualcomm/emac/emac-mac.c
871
curr_rxbuf = GET_RFD_BUFFER(rx_q, rx_q->rfd.produce_idx);
drivers/net/ethernet/qualcomm/emac/emac-mac.c
897
next_produce_idx = rx_q->rfd.produce_idx + 1;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
898
if (next_produce_idx == rx_q->rfd.count)
drivers/net/ethernet/qualcomm/emac/emac-mac.c
901
curr_rxbuf = GET_RFD_BUFFER(rx_q, rx_q->rfd.produce_idx);
drivers/net/ethernet/qualcomm/emac/emac-mac.c
907
u32 prod_idx = (rx_q->rfd.produce_idx << rx_q->produce_shift) &
drivers/net/ethernet/qualcomm/emac/emac-mac.h
175
struct emac_rfd_ring rfd;
drivers/net/ethernet/stmicro/stmmac/dwmac4_dma.c
269
unsigned int rfd, rfa;
drivers/net/ethernet/stmicro/stmmac/dwmac4_dma.c
284
rfd = 0x03; /* Full-2.5K */
drivers/net/ethernet/stmicro/stmmac/dwmac4_dma.c
289
rfd = 0x07; /* Full-4.5K */
drivers/net/ethernet/stmicro/stmmac/dwmac4_dma.c
294
mtl_rx_op = u32_replace_bits(mtl_rx_op, rfd,
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_dma.c
143
unsigned int rfd, rfa;
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_dma.c
158
rfd = 0x03; /* Full-2.5K */
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_dma.c
163
rfd = 0x07; /* Full-4.5K */
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_dma.c
168
flow = u32_replace_bits(flow, rfd, XGMAC_RFD);
fs/9p/v9fs.c
370
fd_opts->rfd = result.uint_32;
fs/9p/vfs_super.c
337
ctx->fd_opts.rfd = ~0;
include/net/9p/client.h
119
int rfd;
include/net/9p/client.h
158
int rfd;
net/9p/trans_fd.c
712
if (clnt->trans_opts.fd.rfd != ~0)
net/9p/trans_fd.c
713
seq_printf(m, ",rfd=%u", clnt->trans_opts.fd.rfd);
net/9p/trans_fd.c
720
static int p9_fd_open(struct p9_client *client, int rfd, int wfd)
net/9p/trans_fd.c
726
ts->rd = fget(rfd);
net/9p/trans_fd.c
978
client->trans_opts.fd.rfd = opts.rfd;
net/9p/trans_fd.c
981
if (opts.rfd == ~0 || opts.wfd == ~0) {
net/9p/trans_fd.c
986
err = p9_fd_open(client, opts.rfd, opts.wfd);