dma_to_phys
struct page *page = phys_to_page(dma_to_phys(dev, handle));
cflush.a.dev_bus_addr = dma_to_phys(dev, handle);
unsigned int bfn = XEN_PFN_DOWN(dma_to_phys(dev, dev_addr));
.dma_to_phys = octeon_big_dma_to_phys,
.dma_to_phys = octeon_small_dma_to_phys,
return octeon_pci_dma_ops->dma_to_phys(dev, daddr);
phys_addr_t (*dma_to_phys)(struct device *dev, dma_addr_t daddr);
.dma_to_phys = octeon_gen1_dma_to_phys,
.dma_to_phys = octeon_gen2_dma_to_phys,
phys_addr_t dma_to_phys(struct device *dev, dma_addr_t daddr);
paddr = dma_to_phys(dev_bogus, expect_dma_addr);
gbe_mem_phys = dma_to_phys(&p_dev->dev, gbe_dma_addr);
unsigned long pfn = bfn_to_pfn(XEN_PFN_DOWN(dma_to_phys(dev, addr)));
unsigned long bfn = XEN_PFN_DOWN(dma_to_phys(dev, dma_addr));
if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dev_addr))))
if (pfn_valid(PFN_DOWN(dma_to_phys(hwdev, dev_addr))))
if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dma_addr))))
if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dma_addr))))
return xen_bus_to_phys(dev, dma_to_phys(dev, dma_addr));
return pfn_to_page(PHYS_PFN(dma_to_phys(dev, dma_addr)));
phys_addr_t paddr = dma_to_phys(dev, sg_dma_address(sg));
phys_addr_t paddr = dma_to_phys(dev, sg_dma_address(sg));
unsigned long pfn = PHYS_PFN(dma_to_phys(dev, dma_addr));
*phys_limit = dma_to_phys(dev, dma_limit);
swiotlb_find_pool(dev, dma_to_phys(dev, dma_addr));
phys = dma_to_phys(dev, addr);
phys_addr_t paddr = dma_to_phys(dev, addr);
phys_addr_t paddr = dma_to_phys(dev, addr);