MAX_FRAME_SIZE
MAX_FRAME_SIZE+NET_IP_ALIGN,
MAX_FRAME_SIZE,
MAX_FRAME_SIZE,
skb = netdev_alloc_skb(greth->netdev, MAX_FRAME_SIZE+NET_IP_ALIGN);
MAX_FRAME_SIZE+NET_IP_ALIGN,
greth->rx_bufs[i] = kmalloc(MAX_FRAME_SIZE, GFP_KERNEL);
MAX_FRAME_SIZE,
greth->tx_bufs[i] = kmalloc(MAX_FRAME_SIZE, GFP_KERNEL);
MAX_FRAME_SIZE,
if (unlikely(skb->len > MAX_FRAME_SIZE)) {
if (unlikely(skb->len > MAX_FRAME_SIZE)) {
dma_sync_single_for_device(greth->dev, dma_addr, MAX_FRAME_SIZE, DMA_FROM_DEVICE);
if (!bad && (newskb=netdev_alloc_skb(dev, MAX_FRAME_SIZE + NET_IP_ALIGN))) {
MAX_FRAME_SIZE + NET_IP_ALIGN,
MAX_FRAME_SIZE + NET_IP_ALIGN,
bgx_reg_modify(bgx, lmacid, BGX_GMP_GMI_RXX_JABBER, MAX_FRAME_SIZE);
bgx_reg_modify(bgx, lmacid, BGX_SMUX_RX_JABBER, MAX_FRAME_SIZE);
MAX_FRAME_SIZE + 2 + sizeof(struct cpl_rx_pkt);
V_RXMAXFRAMERSIZE(MAX_FRAME_SIZE) | F_RXENFRAMER);
if (mtu > MAX_FRAME_SIZE - 4)
u32 len = MAX_FRAME_SIZE;
dma_unmap_single(priv->dev, addr, MAX_FRAME_SIZE,
if (unlikely(skb->len > MAX_FRAME_SIZE)) {
dma_unmap_single(priv->dev, dma_addr, MAX_FRAME_SIZE,
val |= MAX_FRAME_SIZE;
writew(MAX_FRAME_SIZE, ioaddr + MaxFrameSize); /* dev->mtu+14 ??? */
if ((rx_status & TOOLONG) && data_size <= MAX_FRAME_SIZE)
#define TX_BUF_SIZE (MAX_FRAME_SIZE+18)
#define RX_BUF_SIZE (MAX_FRAME_SIZE+18)
smc->os.MaxFrameSize = MAX_FRAME_SIZE;
dma_free_coherent(&pdev->dev, MAX_FRAME_SIZE,
MAX_FRAME_SIZE,
bp->LocalRxBuffer = dma_alloc_coherent(&bp->pdev.dev, MAX_FRAME_SIZE,
printk("LocalRxBuffer: %d byte\n", MAX_FRAME_SIZE);
dma_free_coherent(&bp->pdev.dev, MAX_FRAME_SIZE,