map_pages
unsigned long map_pages;
map_pages(start, __pa(start), end-start,
map_pages(init_end, __pa(init_end), kernel_end - init_end,
map_pages(init_begin, __pa(init_begin), init_end - init_begin,
map_pages(init_begin, __pa(init_begin), init_end - init_begin,
map_pages(start, __pa(start), end - start, PAGE_KERNEL, 0);
map_pages((unsigned long)__va(start_paddr), start_paddr,
map_pages(initrd_start, __pa(initrd_start),
map_pages(linux_gateway_page_addr, __pa(&linux_gateway_page),
map_pages:
goto map_pages;
ret = ops->map_pages(ops, addr, phys, size, 1, prot, GFP_KERNEL, &mapped);
ret = ops->map_pages(ops, addr, phys, pgsize, count,
ret = ops->map_pages(ops, iova, paddr, pgsize, pgcount, prot,
ret = ops->map_pages(ops, iova, paddr, pgsize, pgcount, prot,
.map_pages = apple_dart_map_pages,
return ops->map_pages(ops, iova, paddr, pgsize, pgcount, prot, gfp,
return ops->map_pages(ops, iova, paddr, pgsize, pgcount, prot, gfp, mapped);
.map_pages = arm_smmu_map_pages,
ret = ops->map_pages(ops, iova, paddr, pgsize, pgcount, prot, gfp, mapped);
.map_pages = arm_smmu_map_pages,
ret = ops->map_pages(ops, iova, paddr, pgsize, pgcount, prot, GFP_ATOMIC, mapped);
.map_pages = qcom_iommu_map,
.map_pages = exynos_iommu_map,
int DOMAIN_NS(map_pages)(struct iommu_domain *domain, unsigned long iova,
EXPORT_SYMBOL_NS_GPL(DOMAIN_NS(map_pages), "GENERIC_PT_IOMMU");
if (ops->map_pages(ops, iova, iova, size, 1,
if (ops->map_pages(ops, iova, iova, size, 1,
if (ops->map_pages(ops, iova, iova, size, 1,
if (!ops->map_pages(ops, iova, iova + size, size, 1,
.map_pages = arm_v7s_map_pages,
if (ops->map_pages(ops, iova, iova, size, 1,
if (!ops->map_pages(ops, iova, iova + size, size, 1,
if (ops->map_pages(ops, iova, iova, size, 1, IOMMU_WRITE,
.map_pages = arm_lpae_map_pages,
.map_pages = dart_map_pages,
if (WARN_ON(!ops->map_pages || domain->pgsize_bitmap == 0UL))
ret = ops->map_pages(domain, iova, paddr, pgsize, count, prot,
return domain->iop->map_pages(domain->iop, iova, paddr, pgsize, pgcount,
.map_pages = ipmmu_map,
ret = priv->iop->map_pages(priv->iop, iova, pa, pgsize, pgcount, prot,
.map_pages = msm_iommu_map,
.map_pages = mtk_iommu_map,
return dom->iop->map_pages(dom->iop, iova, paddr, pgsize, pgcount, prot, gfp, mapped);
.map_pages = mtk_iommu_v1_map,
.map_pages = omap_iommu_map,
.map_pages = riscv_iommu_map_pages,
.map_pages = rk_iommu_map,
.map_pages = s390_iommu_map_pages, \
.map_pages = sprd_iommu_map,
.map_pages = sun50i_iommu_map,
.map_pages = tegra_smmu_map,
.map_pages = viommu_map_pages,
.map_pages = filemap_map_pages,
.map_pages = afs_vm_map_pages,
.map_pages = filemap_map_pages,
.map_pages = filemap_map_pages,
.map_pages = filemap_map_pages,
.map_pages = filemap_map_pages,
.map_pages = filemap_map_pages,
.map_pages = filemap_map_pages,
.map_pages = filemap_map_pages,
.map_pages = filemap_map_pages,
.map_pages = filemap_map_pages,
.map_pages = filemap_map_pages,
.map_pages = filemap_map_pages,
.map_pages = filemap_map_pages,
.map_pages = filemap_map_pages,
.map_pages = filemap_map_pages,
.map_pages = filemap_map_pages,
.map_pages = filemap_map_pages,
.map_pages = filemap_map_pages,
.map_pages = &pt_iommu_##fmt##_map_pages, \
int (*map_pages)(struct io_pgtable_ops *ops, unsigned long iova,
int (*map_pages)(struct iommu_domain *domain, unsigned long iova,
vm_fault_t (*map_pages)(struct vm_fault *vmf,
addr = map_pages(start, size);
.map_pages = filemap_map_pages,
if (vma->vm_ops->map_pages || !(vmf->flags & FAULT_FLAG_VMA_LOCK))
ret = vmf->vma->vm_ops->map_pages(vmf,
if (!vmf->vma->vm_ops->map_pages)
.map_pages = filemap_map_pages,
.map_pages = filemap_map_pages,
vm_fault_t (*map_pages)(struct vm_fault *vmf,