maplen
u32 maplen;
maplen = skb_headlen(skb);
paddr = dma_map_single(&sdev->pdev->dev, skb->data, maplen,
dma_unmap_len_set(buff, map_len, maplen);
desc->totlen = cpu_to_le32(maplen);
desc->len = cpu_to_le32(maplen);
unsigned int maplen = SLIC_RX_BUFF_SIZE;
skb = alloc_skb(maplen + ALIGN_MASK, gfp);
paddr = dma_map_single(&sdev->pdev->dev, skb->data, maplen,
dma_unmap_len_set(buff, map_len, maplen);
if (dma_unmap_len(info, maplen)) {
dma_unmap_len(info, maplen),
dma_unmap_len_set(info, maplen, 0);
if (dma_unmap_len(info, maplen)) {
dma_unmap_len(info, maplen),
dma_unmap_len_set(info, maplen, 0);
dma_unmap_len_set(info, maplen, skb->len);
dma_unmap_len_set(info, maplen, skb_frag_size(frag));
DEFINE_DMA_UNMAP_LEN(maplen);
int maplen, f, first_idx = txq->write_idx;
maplen = skb_headlen(skb);
dma = dma_map_single(txq->dev, skb->data, maplen,
dma_unmap_len_set(&txq->bufs[txq->write_idx], size, maplen);
tpd->len = cpu_to_le16(maplen);
maplen = skb_frag_size(frag);
maplen, DMA_TO_DEVICE);
dma_unmap_len_set(&txq->bufs[txq->write_idx], size, maplen);
tpd->len = cpu_to_le16(maplen);
dma_unmap_len(e, maplen),
dma_unmap_len_set(e, maplen, len);
dma_unmap_len_set(e, maplen, skb_frag_size(frag));
dma_unmap_len(e, maplen), DMA_TO_DEVICE);
dma_unmap_len(e, maplen), DMA_TO_DEVICE);
dma_unmap_len(e, maplen), DMA_TO_DEVICE);
dma_unmap_len(e, maplen), DMA_TO_DEVICE);
dma_unmap_len(e, maplen),
dma_unmap_len(e, maplen),
dma_unmap_len(&ee, maplen), DMA_FROM_DEVICE);
dma_unmap_len_set(e, maplen, bufsize);
DEFINE_DMA_UNMAP_LEN(maplen);
dma_unmap_len(re, maplen), DMA_TO_DEVICE);
dma_unmap_len(re, maplen), DMA_TO_DEVICE);
dma_unmap_len_set(re, maplen, len);
dma_unmap_len_set(re, maplen, skb_frag_size(frag));
DEFINE_DMA_UNMAP_LEN(maplen);
dma_unmap_len_set(lrg_buf_cb, maplen,
dma_unmap_len(&tx_cb->map[0], maplen), DMA_TO_DEVICE);
dma_unmap_len(&tx_cb->map[i], maplen),
dma_unmap_len(lrg_buf_cb2, maplen), DMA_FROM_DEVICE);
dma_unmap_len(lrg_buf_cb2, maplen), DMA_FROM_DEVICE);
dma_unmap_len_set(&tx_cb->map[seg], maplen, len);
dma_unmap_len_set(&tx_cb->map[seg], maplen,
dma_unmap_len_set(&tx_cb->map[seg], maplen, skb_frag_size(frag));
dma_unmap_len(&tx_cb->map[seg], maplen),
dma_unmap_len(&tx_cb->map[seg], maplen),
dma_unmap_addr(&tx_cb->map[0], maplen),
dma_unmap_len(lrg_buf_cb, maplen),
dma_unmap_len_set(lrg_buf_cb, maplen,
dma_unmap_len_set(lrg_buf_cb, maplen,
dma_unmap_len(&tx_cb->map[0], maplen),
dma_unmap_len(&tx_cb->map[j], maplen),
DEFINE_DMA_UNMAP_LEN(maplen);
DEFINE_DMA_UNMAP_LEN(maplen);
u32 hlen, maplen;
maplen = skb_headlen(skb);
if (unlikely(sizeof(struct iphdr) + sizeof(struct tcphdr) > maplen))
if (hlen + sizeof(struct tcphdr) > maplen)
size = maplen(map, mr);
paend = map->addr + offset + maplen(map, mr);
u32 maplen;
maplen = ceph_decode_32(&p);
doutc(cl, "epoch %u len %d\n", epoch, (int)maplen);
u32 nr_maps, maplen;
maplen = ceph_decode_32(&p);
ceph_decode_need(&p, end, maplen, bad);
epoch, maplen);
err = handle_one_map(osdc, p, p + maplen, true,
epoch, maplen);
p += maplen;
maplen = ceph_decode_32(&p);
ceph_decode_need(&p, end, maplen, bad);
epoch, maplen);
"older than our %u\n", epoch, maplen,
dout("taking full map %u len %d\n", epoch, maplen);
err = handle_one_map(osdc, p, p + maplen, false,
p += maplen;
h->maplen = st.st_size;
sz = (h->maplen + page_size - 1) & ~(page_size - 1);
h->numlines = countlines(h->map, h->maplen);
fill_lines(h->lines, h->numlines, h->map, h->maplen);
map_total_sz += h->maplen;
p = memchr(l, '\n', sf->map + sf->maplen - l);
size_t maplen;
static int countlines(char *map, int maplen)
char *end = map + maplen;
if (maplen == 0)
static void fill_lines(char **lines, int maxline, char *map, int maplen)
char *end = map + maplen;
if (maplen == 0 || maxline == 0)
map_total_sz -= sf->maplen;
munmap(sf->map, sf->maplen);