RX_BUF_SIZE
unsigned int rx_buf_size_idx = d->dma_addr & RX_BUF_SIZE;
iowrite32(FTGMAC100_RBSR_SIZE(RX_BUF_SIZE),
skb = netdev_alloc_skb_ip_align(netdev, RX_BUF_SIZE);
map = dma_map_single(priv->dev, skb->data, RX_BUF_SIZE,
dma_unmap_single(priv->dev, map, RX_BUF_SIZE, DMA_FROM_DEVICE);
dma_unmap_single(priv->dev, map, RX_BUF_SIZE, DMA_FROM_DEVICE);
dma_free_coherent(priv->dev, RX_BUF_SIZE,
RX_BUF_SIZE,
#define MAX_PKT_SIZE RX_BUF_SIZE /* multi-segment not supported */
#if RX_BUF_SIZE > 0x7ff || RX_BUF_SIZE > PAGE_SIZE
dma_unmap_page(priv->dev, map, RX_BUF_SIZE, DMA_FROM_DEVICE);
map = dma_map_page(priv->dev, page, 0, RX_BUF_SIZE, DMA_FROM_DEVICE);
ftmac100_rxdes_set_buffer_size(rxdes, RX_BUF_SIZE);
dma_unmap_page(priv->dev, map, RX_BUF_SIZE, DMA_FROM_DEVICE);
val = RX_BUF_SIZE >> PPE_BUF_SIZE_SHIFT;
RX_BUF_SIZE, DMA_FROM_DEVICE);
RX_BUF_SIZE, DMA_FROM_DEVICE);
RX_BUF_SIZE, DMA_FROM_DEVICE);
RX_BUF_SIZE, DMA_FROM_DEVICE);
priv->rx_buf_size = RX_BUF_SIZE +
if (dev->mtu > RX_BUF_SIZE)
skge->rx_buf_size = RX_BUF_SIZE;
moxart_desc_write(RX_BUF_SIZE & RX_DESC1_BUF_SIZE_MASK,
if (len > RX_BUF_SIZE)
len = RX_BUF_SIZE;
priv->rx_buf_size = RX_BUF_SIZE;
#if RX_BUF_SIZE >= RX_BUF_SIZE_MAX
#define REAL_RX_BUF_SIZE (RX_BUF_SIZE + 14) /* rx/tx mac addr + type */
dma_unmap_single(&dev->pci_dev->dev, bufptr, RX_BUF_SIZE,
new_skb = alloc_skb(RX_BUF_SIZE + IOC3_DMA_XFER_LEN - 1, GFP_ATOMIC);
RX_BUF_SIZE, DMA_FROM_DEVICE);
RX_BUF_SIZE, DMA_FROM_DEVICE);
RX_BUF_SIZE, DMA_FROM_DEVICE);
tp->rx_buf_sz = (mtu > RX_BUF_SIZE) ? mtu + ETH_HLEN + 8 : RX_BUF_SIZE;
if ((skb = netdev_alloc_skb(net_dev, RX_BUF_SIZE)) == NULL) {
sis_priv->rx_ring[i].cmdsts = RX_BUF_SIZE;
RX_BUF_SIZE,
sis_priv->rx_ring[entry].cmdsts = RX_BUF_SIZE;
RX_BUF_SIZE, DMA_FROM_DEVICE);
if ((skb = netdev_alloc_skb(net_dev, RX_BUF_SIZE)) == NULL) {
sis_priv->rx_ring[entry].cmdsts = RX_BUF_SIZE;
skb->data, RX_BUF_SIZE,
skb = netdev_alloc_skb(net_dev, RX_BUF_SIZE);
sis_priv->rx_ring[entry].cmdsts = RX_BUF_SIZE;
skb->data, RX_BUF_SIZE,
RX_BUF_SIZE, DMA_FROM_DEVICE);
RX_BUF_SIZE, DMA_FROM_DEVICE);
RX_BUF_SIZE);
skb = netdev_alloc_skb(dev, RX_BUF_SIZE);
*dma_handle = dma_map_single(&hwdev->dev, skb->data, RX_BUF_SIZE,
dma_unmap_single(&hwdev->dev, dma_handle, RX_BUF_SIZE,
RX_BUF_SIZE);
rx_buf = kzalloc(RX_BUF_SIZE, GFP_KERNEL);
priv->rx_size = RX_BUF_SIZE * RX_BUFS_COUNT;
udma_writel(priv, REGS_DMA_RX, UDMA_RX_BUFFER_SIZE, RX_BUF_SIZE);
daddr += RX_BUF_SIZE;
priv->rx_bufs + (index * RX_BUF_SIZE),
else if (length != RX_BUF_SIZE)
if (*buff_index > (RX_BUF_SIZE - UART_BURST_SIZE))
if (*buff_index >= RX_BUF_SIZE)
unsigned char rx_buff[RX_BUF_SIZE];
valid_byte_count < RX_BUF_SIZE) {
pinfo->rx_fifosize = RX_BUF_SIZE;
pinfo->rx_fifosize = RX_BUF_SIZE;
qe_port->rx_fifosize = RX_BUF_SIZE;
u8 buf[RX_BUF_SIZE];
u8 rx_data[RX_BUF_SIZE];