sg_dma_len
tmp_sg[sg_indx].size = sg_dma_len(sg);
len = sg_dma_len(sg) >> shift;
chunk_pages = DIV_ROUND_UP(sg_dma_len(sg), iwmr->page_size);
return sg_dma_len(&iter->chunk->mem[iter->page_idx]);
sg_dma_len(mem) = PAGE_SIZE << order;
if (sg_dma_len(&chunk->mem[i]) > dma_offset)
dma_offset -= sg_dma_len(&chunk->mem[i]);
len = sg_dma_len(sg) >> mtt->page_shift;
len = sg_dma_len(sg) >> umem_page_shift;
len = sg_dma_len(sg) >> PAGE_SHIFT;
klms[i].bcount = cpu_to_be32(sg_dma_len(sg) - sg_offset);
mr->ibmr.length += sg_dma_len(sg);
sg_dma_len(mem) = PAGE_SIZE << order;
if (sg_dma_len(&chunk->mem[i]) > dma_offset)
dma_offset -= sg_dma_len(&chunk->mem[i]);
return sg_dma_len(&iter->chunk->mem[iter->page_idx]);
len = sg_dma_len(sg) >> shift;
pages = sg_dma_len(sg) >> shift;
n += sg_dma_len(sg) >> shift;
if (unlikely(sg_nents <= 0 || sg_offset > sg_dma_len(&sgl[0])))
unsigned int dma_len = sg_dma_len(sg) - sg_offset;
return sg_dma_len(sg);