IOMMU_WRITE
return prot | IOMMU_READ | IOMMU_WRITE;
return prot | IOMMU_WRITE;
IOMMU_READ | IOMMU_WRITE);
int prot = IOMMU_READ | IOMMU_WRITE;
int prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO;
IOMMU_READ | IOMMU_WRITE, &cur_page);
if (prot & IOMMU_WRITE)
prot |= IOMMU_WRITE;
prot = IOMMU_READ | IOMMU_WRITE;
PAGE_SIZE, IOMMU_READ | IOMMU_WRITE,
int prot = IOMMU_READ | IOMMU_WRITE | IOMMU_CACHE;
IOMMU_WRITE | IOMMU_READ | IOMMU_CACHE | IOMMU_NOEXEC, sgt);
prot |= IOMMU_READ | IOMMU_WRITE;
int prot = IOMMU_READ | IOMMU_WRITE;
size, IOMMU_READ | IOMMU_WRITE, GFP_KERNEL);
int prot = IOMMU_READ | IOMMU_WRITE;
if (interval->flags & IOMMU_WRITE)
flags |= (uiomr->writable) ? IOMMU_WRITE : 0;
(writable) ? IOMMU_WRITE : 0,
IOMMU_WRITE,
(writable) ? IOMMU_WRITE : 0);
prot |= IOMMU_WRITE;
int prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO;
int prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO;
int prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO;
int prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO;
return prot | IOMMU_READ | IOMMU_WRITE;
return prot | IOMMU_WRITE;
#define SYSMMU_SUPPORTED_PROT_BITS (IOMMU_READ | IOMMU_WRITE)
if (iommu_prot & IOMMU_WRITE)
if (!(iommu_prot & (IOMMU_READ | IOMMU_WRITE)))
if (iommu_prot & IOMMU_WRITE)
!(iommu_prot & IOMMU_WRITE)) {
if (iommu_prot & IOMMU_WRITE)
if (WARN_ON(!(prot & (IOMMU_READ | IOMMU_WRITE))))
ret = iommu_map(&priv->domain, va, pa, len, IOMMU_READ | IOMMU_WRITE,
for (prot = 0; prot <= (IOMMU_READ | IOMMU_WRITE | IOMMU_CACHE |
IOMMU_READ | IOMMU_WRITE);
IOMMU_READ | IOMMU_WRITE));
IOMMU_READ | IOMMU_WRITE, GFP_KERNEL);
IOMMU_WRITE, GFP_KERNEL, &mapped))
IOMMU_READ | IOMMU_WRITE |
IOMMU_READ | IOMMU_WRITE |
if (!(prot & IOMMU_WRITE))
if (!(prot & (IOMMU_READ | IOMMU_WRITE)))
IOMMU_READ | IOMMU_WRITE |
if (ops->map_pages(ops, iova, iova, size, 1, IOMMU_WRITE,
if (!(prot & IOMMU_WRITE) && (prot & IOMMU_READ))
if (prot & IOMMU_WRITE)
if (!(iommu_prot & (IOMMU_READ | IOMMU_WRITE)))
if (!(prot & IOMMU_WRITE))
if (!(prot & IOMMU_WRITE))
if (!(iommu_prot & (IOMMU_READ | IOMMU_WRITE)))
return area->iommu_prot & IOMMU_WRITE;
IOMMU_WRITE | IOMMU_READ | IOMMU_MMIO,
if ((iommu_prot & IOMMU_WRITE) && !pages->writable)
pages = iopt_alloc_user_pages(uptr, length, iommu_prot & IOMMU_WRITE);
iommu_prot & IOMMU_WRITE);
iommu_prot & IOMMU_WRITE);
iommu_prot |= IOMMU_WRITE;
iommu_prot |= IOMMU_WRITE;
int prot = IOMMU_WRITE | IOMMU_READ;
int prot = IOMMU_READ | IOMMU_WRITE;
if (!(prot & IOMMU_WRITE))
flags |= (prot & IOMMU_WRITE) ? RK_PTE_PAGE_WRITABLE : 0;
flags |= (prot & IOMMU_WRITE) ? RK_PTE_PAGE_WRITABLE : 0;
if (!(prot & IOMMU_WRITE))
if ((prot & (IOMMU_READ | IOMMU_WRITE)) == (IOMMU_READ | IOMMU_WRITE))
else if (prot & IOMMU_WRITE)
if (prot & IOMMU_WRITE)
unsigned long prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO;
(prot & IOMMU_WRITE ? VIRTIO_IOMMU_MAP_F_WRITE : 0) |
int prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO;
IOMMU_READ | IOMMU_WRITE);
IOMMU_READ | IOMMU_WRITE | IOMMU_PRIV, GFP_KERNEL);
rcb_size, IOMMU_READ | IOMMU_WRITE, 0);
ret = iommu_map(domain, iova, phys, size, IOMMU_READ | IOMMU_WRITE,
ret = iommu_map(domain, iova, phys, size, IOMMU_READ | IOMMU_WRITE,
IOMMU_READ | IOMMU_WRITE, GFP_KERNEL);
IOMMU_READ | IOMMU_WRITE, GFP_KERNEL);
IOMMU_READ | IOMMU_WRITE, GFP_KERNEL);
adsp->mem_size, IOMMU_READ | IOMMU_WRITE,
IOMMU_READ | IOMMU_WRITE, GFP_KERNEL);
IOMMU_READ | IOMMU_WRITE,
IOMMU_READ | IOMMU_WRITE, &h_page);
prot |= IOMMU_WRITE;
if ((write && !(dma->prot & IOMMU_WRITE)) ||
unpin_user_pages_dirty_lock(&page, 1, prot & IOMMU_WRITE);
if (prot & IOMMU_WRITE)
prot & IOMMU_WRITE);
prot & IOMMU_WRITE);
(prot & IOMMU_WRITE) ? IOMMUFD_ACCESS_RW_WRITE : 0);
flags |= IOMMU_WRITE;
flags |= (IOMMU_WRITE | IOMMU_READ);
int prot = IOMMU_READ | IOMMU_WRITE;
prot = IOMMU_READ | IOMMU_WRITE;