dma_get_mask
dma_get_mask(dev), direction, attrs);
dma_get_mask(dev), direction, attrs);
ret = iommu_map_phys(dev, tbl, phys, size, dma_get_mask(dev),
ret = ppc_iommu_map_sg(dev, tbl, sglist, nelems, dma_get_mask(dev),
dma_dev->src_addr_widths = CCP_DMA_WIDTH(dma_get_mask(ccp->dev));
dma_dev->dst_addr_widths = CCP_DMA_WIDTH(dma_get_mask(ccp->dev));
u64 dma_mask = dma_get_mask(pdev->dev.parent);
u64 dma_mask = dma_get_mask(&dev->dev);
u64 dma_mask = dma_get_mask(pdev->dev.parent);
dma_addr_t iova, dma_mask = dma_get_mask(dev);
iova = iommu_dma_alloc_iova(domain, iova_len, dma_get_mask(dev), dev);
dma_get_mask(dev), dev);
iova = iommu_dma_alloc_iova(domain, size, dma_get_mask(dev), dev);
hw->mac.ops.set_dma_mask(hw, dma_get_mask(&pdev->dev));
if (dma_get_mask(&pdev->dev) > DMA_BIT_MASK(32)) {
sdma->dma_mask = dma_get_mask(dev);
u64 dma_mask = dma_get_mask(&pdev->dev);
if (dma_get_mask(&pdev->dev) == DMA_BIT_MASK(64)) {
ret = dma_coerce_mask_and_coherent(dev, dma_get_mask(rproc->dev.parent));
dma_get_mask(rproc->dev.parent), ERR_PTR(ret));
old_dma_mask = dma_get_mask(map->dev);
if (min_not_zero(dma_get_mask(dev), dev->bus_dma_limit) <