#include <linux/dma-buf.h>
#include <linux/dma-resv.h>
#include <linux/file.h>
#include <linux/highmem.h>
#include <linux/iommu.h>
#include <linux/iommufd.h>
#include <linux/kthread.h>
#include <linux/overflow.h>
#include <linux/slab.h>
#include <linux/sched/mm.h>
#include <linux/vfio_pci_core.h>
#include "double_span.h"
#include "io_pagetable.h"
#ifndef CONFIG_IOMMUFD_TEST
#define TEMP_MEMORY_LIMIT 65536
#else
#define TEMP_MEMORY_LIMIT iommufd_test_memory_limit
#endif
#define BATCH_BACKUP_SIZE 32
static void *temp_kmalloc(size_t *size, void *backup, size_t backup_len)
{
void *res;
if (WARN_ON(*size == 0))
return NULL;
if (*size < backup_len)
return backup;
if (!backup && iommufd_should_fail())
return NULL;
*size = min_t(size_t, *size, TEMP_MEMORY_LIMIT);
res = kmalloc(*size, GFP_KERNEL | __GFP_NOWARN | __GFP_NORETRY);
if (res)
return res;
*size = PAGE_SIZE;
if (backup_len) {
res = kmalloc(*size, GFP_KERNEL | __GFP_NOWARN | __GFP_NORETRY);
if (res)
return res;
*size = backup_len;
return backup;
}
return kmalloc(*size, GFP_KERNEL);
}
void interval_tree_double_span_iter_update(
struct interval_tree_double_span_iter *iter)
{
unsigned long last_hole = ULONG_MAX;
unsigned int i;
for (i = 0; i != ARRAY_SIZE(iter->spans); i++) {
if (interval_tree_span_iter_done(&iter->spans[i])) {
iter->is_used = -1;
return;
}
if (iter->spans[i].is_hole) {
last_hole = min(last_hole, iter->spans[i].last_hole);
continue;
}
iter->is_used = i + 1;
iter->start_used = iter->spans[i].start_used;
iter->last_used = min(iter->spans[i].last_used, last_hole);
return;
}
iter->is_used = 0;
iter->start_hole = iter->spans[0].start_hole;
iter->last_hole =
min(iter->spans[0].last_hole, iter->spans[1].last_hole);
}
void interval_tree_double_span_iter_first(
struct interval_tree_double_span_iter *iter,
struct rb_root_cached *itree1, struct rb_root_cached *itree2,
unsigned long first_index, unsigned long last_index)
{
unsigned int i;
iter->itrees[0] = itree1;
iter->itrees[1] = itree2;
for (i = 0; i != ARRAY_SIZE(iter->spans); i++)
interval_tree_span_iter_first(&iter->spans[i], iter->itrees[i],
first_index, last_index);
interval_tree_double_span_iter_update(iter);
}
void interval_tree_double_span_iter_next(
struct interval_tree_double_span_iter *iter)
{
unsigned int i;
if (iter->is_used == -1 ||
iter->last_hole == iter->spans[0].last_index) {
iter->is_used = -1;
return;
}
for (i = 0; i != ARRAY_SIZE(iter->spans); i++)
interval_tree_span_iter_advance(
&iter->spans[i], iter->itrees[i], iter->last_hole + 1);
interval_tree_double_span_iter_update(iter);
}
static void iopt_pages_add_npinned(struct iopt_pages *pages, size_t npages)
{
int rc;
rc = check_add_overflow(pages->npinned, npages, &pages->npinned);
if (IS_ENABLED(CONFIG_IOMMUFD_TEST))
WARN_ON(rc || pages->npinned > pages->npages);
}
static void iopt_pages_sub_npinned(struct iopt_pages *pages, size_t npages)
{
int rc;
rc = check_sub_overflow(pages->npinned, npages, &pages->npinned);
if (IS_ENABLED(CONFIG_IOMMUFD_TEST))
WARN_ON(rc || pages->npinned > pages->npages);
}
static void iopt_pages_err_unpin(struct iopt_pages *pages,
unsigned long start_index,
unsigned long last_index,
struct page **page_list)
{
unsigned long npages = last_index - start_index + 1;
unpin_user_pages(page_list, npages);
iopt_pages_sub_npinned(pages, npages);
}
static unsigned long iopt_area_index_to_iova(struct iopt_area *area,
unsigned long index)
{
if (IS_ENABLED(CONFIG_IOMMUFD_TEST))
WARN_ON(index < iopt_area_index(area) ||
index > iopt_area_last_index(area));
index -= iopt_area_index(area);
if (index == 0)
return iopt_area_iova(area);
return iopt_area_iova(area) - area->page_offset + index * PAGE_SIZE;
}
static unsigned long iopt_area_index_to_iova_last(struct iopt_area *area,
unsigned long index)
{
if (IS_ENABLED(CONFIG_IOMMUFD_TEST))
WARN_ON(index < iopt_area_index(area) ||
index > iopt_area_last_index(area));
if (index == iopt_area_last_index(area))
return iopt_area_last_iova(area);
return iopt_area_iova(area) - area->page_offset +
(index - iopt_area_index(area) + 1) * PAGE_SIZE - 1;
}
static void iommu_unmap_nofail(struct iommu_domain *domain, unsigned long iova,
size_t size)
{
size_t ret;
ret = iommu_unmap(domain, iova, size);
WARN_ON(ret != size);
}
static void iopt_area_unmap_domain_range(struct iopt_area *area,
struct iommu_domain *domain,
unsigned long start_index,
unsigned long last_index)
{
unsigned long start_iova = iopt_area_index_to_iova(area, start_index);
iommu_unmap_nofail(domain, start_iova,
iopt_area_index_to_iova_last(area, last_index) -
start_iova + 1);
}
static struct iopt_area *iopt_pages_find_domain_area(struct iopt_pages *pages,
unsigned long index)
{
struct interval_tree_node *node;
node = interval_tree_iter_first(&pages->domains_itree, index, index);
if (!node)
return NULL;
return container_of(node, struct iopt_area, pages_node);
}
enum batch_kind {
BATCH_CPU_MEMORY = 0,
BATCH_MMIO,
};
struct pfn_batch {
unsigned long *pfns;
u32 *npfns;
unsigned int array_size;
unsigned int end;
unsigned int total_pfns;
enum batch_kind kind;
};
enum { MAX_NPFNS = type_max(typeof(((struct pfn_batch *)0)->npfns[0])) };
static void batch_clear(struct pfn_batch *batch)
{
batch->total_pfns = 0;
batch->end = 0;
batch->pfns[0] = 0;
batch->npfns[0] = 0;
batch->kind = 0;
}
static void batch_clear_carry(struct pfn_batch *batch, unsigned int keep_pfns)
{
if (!keep_pfns)
return batch_clear(batch);
if (IS_ENABLED(CONFIG_IOMMUFD_TEST))
WARN_ON(!batch->end ||
batch->npfns[batch->end - 1] < keep_pfns);
batch->total_pfns = keep_pfns;
batch->pfns[0] = batch->pfns[batch->end - 1] +
(batch->npfns[batch->end - 1] - keep_pfns);
batch->npfns[0] = keep_pfns;
batch->end = 1;
}
static void batch_skip_carry(struct pfn_batch *batch, unsigned int skip_pfns)
{
if (!batch->total_pfns)
return;
if (IS_ENABLED(CONFIG_IOMMUFD_TEST))
WARN_ON(batch->total_pfns != batch->npfns[0]);
skip_pfns = min(batch->total_pfns, skip_pfns);
batch->pfns[0] += skip_pfns;
batch->npfns[0] -= skip_pfns;
batch->total_pfns -= skip_pfns;
}
static int __batch_init(struct pfn_batch *batch, size_t max_pages, void *backup,
size_t backup_len)
{
const size_t elmsz = sizeof(*batch->pfns) + sizeof(*batch->npfns);
size_t size = max_pages * elmsz;
batch->pfns = temp_kmalloc(&size, backup, backup_len);
if (!batch->pfns)
return -ENOMEM;
if (IS_ENABLED(CONFIG_IOMMUFD_TEST) && WARN_ON(size < elmsz))
return -EINVAL;
batch->array_size = size / elmsz;
batch->npfns = (u32 *)(batch->pfns + batch->array_size);
batch_clear(batch);
return 0;
}
static int batch_init(struct pfn_batch *batch, size_t max_pages)
{
return __batch_init(batch, max_pages, NULL, 0);
}
static void batch_init_backup(struct pfn_batch *batch, size_t max_pages,
void *backup, size_t backup_len)
{
__batch_init(batch, max_pages, backup, backup_len);
}
static void batch_destroy(struct pfn_batch *batch, void *backup)
{
if (batch->pfns != backup)
kfree(batch->pfns);
}
static bool batch_add_pfn_num(struct pfn_batch *batch, unsigned long pfn,
u32 nr, enum batch_kind kind)
{
unsigned int end = batch->end;
if (batch->kind != kind) {
if (batch->end != 0)
return false;
batch->kind = kind;
}
if (end && pfn == batch->pfns[end - 1] + batch->npfns[end - 1] &&
nr <= MAX_NPFNS - batch->npfns[end - 1]) {
batch->npfns[end - 1] += nr;
} else if (end < batch->array_size) {
batch->pfns[end] = pfn;
batch->npfns[end] = nr;
batch->end++;
} else {
return false;
}
batch->total_pfns += nr;
return true;
}
static void batch_remove_pfn_num(struct pfn_batch *batch, unsigned long nr)
{
batch->npfns[batch->end - 1] -= nr;
if (batch->npfns[batch->end - 1] == 0)
batch->end--;
batch->total_pfns -= nr;
}
static bool batch_add_pfn(struct pfn_batch *batch, unsigned long pfn)
{
return batch_add_pfn_num(batch, pfn, 1, BATCH_CPU_MEMORY);
}
static void batch_from_domain(struct pfn_batch *batch,
struct iommu_domain *domain,
struct iopt_area *area, unsigned long start_index,
unsigned long last_index)
{
unsigned int page_offset = 0;
unsigned long iova;
phys_addr_t phys;
iova = iopt_area_index_to_iova(area, start_index);
if (start_index == iopt_area_index(area))
page_offset = area->page_offset;
while (start_index <= last_index) {
phys = iommu_iova_to_phys(domain, iova) - page_offset;
if (!batch_add_pfn(batch, PHYS_PFN(phys)))
return;
iova += PAGE_SIZE - page_offset;
page_offset = 0;
start_index++;
}
}
static struct page **raw_pages_from_domain(struct iommu_domain *domain,
struct iopt_area *area,
unsigned long start_index,
unsigned long last_index,
struct page **out_pages)
{
unsigned int page_offset = 0;
unsigned long iova;
phys_addr_t phys;
iova = iopt_area_index_to_iova(area, start_index);
if (start_index == iopt_area_index(area))
page_offset = area->page_offset;
while (start_index <= last_index) {
phys = iommu_iova_to_phys(domain, iova) - page_offset;
*(out_pages++) = pfn_to_page(PHYS_PFN(phys));
iova += PAGE_SIZE - page_offset;
page_offset = 0;
start_index++;
}
return out_pages;
}
static void batch_from_domain_continue(struct pfn_batch *batch,
struct iommu_domain *domain,
struct iopt_area *area,
unsigned long start_index,
unsigned long last_index)
{
unsigned int array_size = batch->array_size;
batch->array_size = batch->end;
batch_from_domain(batch, domain, area, start_index, last_index);
batch->array_size = array_size;
}
static int batch_iommu_map_small(struct iommu_domain *domain,
unsigned long iova, phys_addr_t paddr,
size_t size, int prot)
{
unsigned long start_iova = iova;
int rc;
if (IS_ENABLED(CONFIG_IOMMUFD_TEST))
WARN_ON(paddr % PAGE_SIZE || iova % PAGE_SIZE ||
size % PAGE_SIZE);
while (size) {
rc = iommu_map(domain, iova, paddr, PAGE_SIZE, prot,
GFP_KERNEL_ACCOUNT);
if (rc)
goto err_unmap;
iova += PAGE_SIZE;
paddr += PAGE_SIZE;
size -= PAGE_SIZE;
}
return 0;
err_unmap:
if (start_iova != iova)
iommu_unmap_nofail(domain, start_iova, iova - start_iova);
return rc;
}
static int batch_to_domain(struct pfn_batch *batch, struct iommu_domain *domain,
struct iopt_area *area, unsigned long start_index)
{
bool disable_large_pages = area->iopt->disable_large_pages;
unsigned long last_iova = iopt_area_last_iova(area);
int iommu_prot = area->iommu_prot;
unsigned int page_offset = 0;
unsigned long start_iova;
unsigned long next_iova;
unsigned int cur = 0;
unsigned long iova;
int rc;
if (batch->kind == BATCH_MMIO) {
iommu_prot &= ~IOMMU_CACHE;
iommu_prot |= IOMMU_MMIO;
}
if (start_index == iopt_area_index(area))
page_offset = area->page_offset;
next_iova = iova = start_iova =
iopt_area_index_to_iova(area, start_index);
while (cur < batch->end) {
next_iova = min(last_iova + 1,
next_iova + batch->npfns[cur] * PAGE_SIZE -
page_offset);
if (disable_large_pages)
rc = batch_iommu_map_small(
domain, iova,
PFN_PHYS(batch->pfns[cur]) + page_offset,
next_iova - iova, iommu_prot);
else
rc = iommu_map(domain, iova,
PFN_PHYS(batch->pfns[cur]) + page_offset,
next_iova - iova, iommu_prot,
GFP_KERNEL_ACCOUNT);
if (rc)
goto err_unmap;
iova = next_iova;
page_offset = 0;
cur++;
}
return 0;
err_unmap:
if (start_iova != iova)
iommu_unmap_nofail(domain, start_iova, iova - start_iova);
return rc;
}
static void batch_from_xarray(struct pfn_batch *batch, struct xarray *xa,
unsigned long start_index,
unsigned long last_index)
{
XA_STATE(xas, xa, start_index);
void *entry;
rcu_read_lock();
while (true) {
entry = xas_next(&xas);
if (xas_retry(&xas, entry))
continue;
WARN_ON(!xa_is_value(entry));
if (!batch_add_pfn(batch, xa_to_value(entry)) ||
start_index == last_index)
break;
start_index++;
}
rcu_read_unlock();
}
static void batch_from_xarray_clear(struct pfn_batch *batch, struct xarray *xa,
unsigned long start_index,
unsigned long last_index)
{
XA_STATE(xas, xa, start_index);
void *entry;
xas_lock(&xas);
while (true) {
entry = xas_next(&xas);
if (xas_retry(&xas, entry))
continue;
WARN_ON(!xa_is_value(entry));
if (!batch_add_pfn(batch, xa_to_value(entry)))
break;
xas_store(&xas, NULL);
if (start_index == last_index)
break;
start_index++;
}
xas_unlock(&xas);
}
static void clear_xarray(struct xarray *xa, unsigned long start_index,
unsigned long last_index)
{
XA_STATE(xas, xa, start_index);
void *entry;
xas_lock(&xas);
xas_for_each(&xas, entry, last_index)
xas_store(&xas, NULL);
xas_unlock(&xas);
}
static int pages_to_xarray(struct xarray *xa, unsigned long start_index,
unsigned long last_index, struct page **pages)
{
struct page **end_pages = pages + (last_index - start_index) + 1;
struct page **half_pages = pages + (end_pages - pages) / 2;
XA_STATE(xas, xa, start_index);
do {
void *old;
xas_lock(&xas);
while (pages != end_pages) {
if (pages == half_pages && iommufd_should_fail()) {
xas_set_err(&xas, -EINVAL);
xas_unlock(&xas);
xas_nomem(&xas, GFP_KERNEL);
goto err_clear;
}
old = xas_store(&xas, xa_mk_value(page_to_pfn(*pages)));
if (xas_error(&xas))
break;
WARN_ON(old);
pages++;
xas_next(&xas);
}
xas_unlock(&xas);
} while (xas_nomem(&xas, GFP_KERNEL));
err_clear:
if (xas_error(&xas)) {
if (xas.xa_index != start_index)
clear_xarray(xa, start_index, xas.xa_index - 1);
return xas_error(&xas);
}
return 0;
}
static void batch_from_pages(struct pfn_batch *batch, struct page **pages,
size_t npages)
{
struct page **end = pages + npages;
for (; pages != end; pages++)
if (!batch_add_pfn(batch, page_to_pfn(*pages)))
break;
}
static int batch_from_folios(struct pfn_batch *batch, struct folio ***folios_p,
unsigned long *offset_p, unsigned long npages)
{
int rc = 0;
struct folio **folios = *folios_p;
unsigned long offset = *offset_p;
while (npages) {
struct folio *folio = *folios;
unsigned long nr = folio_nr_pages(folio) - offset;
unsigned long pfn = page_to_pfn(folio_page(folio, offset));
nr = min(nr, npages);
npages -= nr;
if (!batch_add_pfn_num(batch, pfn, nr, BATCH_CPU_MEMORY))
break;
if (nr > 1) {
rc = folio_add_pins(folio, nr - 1);
if (rc) {
batch_remove_pfn_num(batch, nr);
goto out;
}
}
folios++;
offset = 0;
}
out:
*folios_p = folios;
*offset_p = offset;
return rc;
}
static void batch_unpin(struct pfn_batch *batch, struct iopt_pages *pages,
unsigned int first_page_off, size_t npages)
{
unsigned int cur = 0;
while (first_page_off) {
if (batch->npfns[cur] > first_page_off)
break;
first_page_off -= batch->npfns[cur];
cur++;
}
while (npages) {
size_t to_unpin = min_t(size_t, npages,
batch->npfns[cur] - first_page_off);
unpin_user_page_range_dirty_lock(
pfn_to_page(batch->pfns[cur] + first_page_off),
to_unpin, pages->writable);
iopt_pages_sub_npinned(pages, to_unpin);
cur++;
first_page_off = 0;
npages -= to_unpin;
}
}
static void copy_data_page(struct page *page, void *data, unsigned long offset,
size_t length, unsigned int flags)
{
void *mem;
mem = kmap_local_page(page);
if (flags & IOMMUFD_ACCESS_RW_WRITE) {
memcpy(mem + offset, data, length);
set_page_dirty_lock(page);
} else {
memcpy(data, mem + offset, length);
}
kunmap_local(mem);
}
static unsigned long batch_rw(struct pfn_batch *batch, void *data,
unsigned long offset, unsigned long length,
unsigned int flags)
{
unsigned long copied = 0;
unsigned int npage = 0;
unsigned int cur = 0;
while (cur < batch->end) {
unsigned long bytes = min(length, PAGE_SIZE - offset);
copy_data_page(pfn_to_page(batch->pfns[cur] + npage), data,
offset, bytes, flags);
offset = 0;
length -= bytes;
data += bytes;
copied += bytes;
npage++;
if (npage == batch->npfns[cur]) {
npage = 0;
cur++;
}
if (!length)
break;
}
return copied;
}
struct pfn_reader_user {
struct page **upages;
size_t upages_len;
unsigned long upages_start;
unsigned long upages_end;
unsigned int gup_flags;
int locked;
struct file *file;
struct folio **ufolios;
size_t ufolios_len;
unsigned long ufolios_offset;
struct folio **ufolios_next;
};
static void pfn_reader_user_init(struct pfn_reader_user *user,
struct iopt_pages *pages)
{
user->upages = NULL;
user->upages_len = 0;
user->upages_start = 0;
user->upages_end = 0;
user->locked = -1;
user->gup_flags = FOLL_LONGTERM;
if (pages->writable)
user->gup_flags |= FOLL_WRITE;
user->file = (pages->type == IOPT_ADDRESS_FILE) ? pages->file : NULL;
user->ufolios = NULL;
user->ufolios_len = 0;
user->ufolios_next = NULL;
user->ufolios_offset = 0;
}
static void pfn_reader_user_destroy(struct pfn_reader_user *user,
struct iopt_pages *pages)
{
if (user->locked != -1) {
if (user->locked)
mmap_read_unlock(pages->source_mm);
if (!user->file && pages->source_mm != current->mm)
mmput(pages->source_mm);
user->locked = -1;
}
kfree(user->upages);
user->upages = NULL;
kfree(user->ufolios);
user->ufolios = NULL;
}
static long pin_memfd_pages(struct pfn_reader_user *user, unsigned long start,
unsigned long npages)
{
unsigned long i;
unsigned long offset;
unsigned long npages_out = 0;
struct page **upages = user->upages;
unsigned long end = start + (npages << PAGE_SHIFT) - 1;
long nfolios = user->ufolios_len / sizeof(*user->ufolios);
nfolios = memfd_pin_folios(user->file, start, end, user->ufolios,
nfolios, &offset);
if (nfolios <= 0)
return nfolios;
offset >>= PAGE_SHIFT;
user->ufolios_next = user->ufolios;
user->ufolios_offset = offset;
for (i = 0; i < nfolios; i++) {
struct folio *folio = user->ufolios[i];
unsigned long nr = folio_nr_pages(folio);
unsigned long npin = min(nr - offset, npages);
npages -= npin;
npages_out += npin;
if (upages) {
if (npin == 1) {
*upages++ = folio_page(folio, offset);
} else {
int rc = folio_add_pins(folio, npin - 1);
if (rc)
return rc;
while (npin--)
*upages++ = folio_page(folio, offset++);
}
}
offset = 0;
}
return npages_out;
}
static int pfn_reader_user_pin(struct pfn_reader_user *user,
struct iopt_pages *pages,
unsigned long start_index,
unsigned long last_index)
{
bool remote_mm = pages->source_mm != current->mm;
unsigned long npages = last_index - start_index + 1;
unsigned long start;
unsigned long unum;
uintptr_t uptr;
long rc;
if (IS_ENABLED(CONFIG_IOMMUFD_TEST) &&
WARN_ON(last_index < start_index))
return -EINVAL;
if (!user->file && !user->upages) {
user->upages_len = npages * sizeof(*user->upages);
user->upages = temp_kmalloc(&user->upages_len, NULL, 0);
if (!user->upages)
return -ENOMEM;
}
if (user->file && !user->ufolios) {
user->ufolios_len = npages * sizeof(*user->ufolios);
user->ufolios = temp_kmalloc(&user->ufolios_len, NULL, 0);
if (!user->ufolios)
return -ENOMEM;
}
if (user->locked == -1) {
if (!user->file && remote_mm) {
if (!mmget_not_zero(pages->source_mm))
return -EFAULT;
}
user->locked = 0;
}
unum = user->file ? user->ufolios_len / sizeof(*user->ufolios) :
user->upages_len / sizeof(*user->upages);
npages = min_t(unsigned long, npages, unum);
if (iommufd_should_fail())
return -EFAULT;
if (user->file) {
start = pages->start + (start_index * PAGE_SIZE);
rc = pin_memfd_pages(user, start, npages);
} else if (!remote_mm) {
uptr = (uintptr_t)(pages->uptr + start_index * PAGE_SIZE);
rc = pin_user_pages_fast(uptr, npages, user->gup_flags,
user->upages);
} else {
uptr = (uintptr_t)(pages->uptr + start_index * PAGE_SIZE);
if (!user->locked) {
mmap_read_lock(pages->source_mm);
user->locked = 1;
}
rc = pin_user_pages_remote(pages->source_mm, uptr, npages,
user->gup_flags, user->upages,
&user->locked);
}
if (rc <= 0) {
if (WARN_ON(!rc))
return -EFAULT;
return rc;
}
iopt_pages_add_npinned(pages, rc);
user->upages_start = start_index;
user->upages_end = start_index + rc;
return 0;
}
static int incr_user_locked_vm(struct iopt_pages *pages, unsigned long npages)
{
unsigned long lock_limit;
unsigned long cur_pages;
unsigned long new_pages;
lock_limit = task_rlimit(pages->source_task, RLIMIT_MEMLOCK) >>
PAGE_SHIFT;
cur_pages = atomic_long_read(&pages->source_user->locked_vm);
do {
new_pages = cur_pages + npages;
if (new_pages > lock_limit)
return -ENOMEM;
} while (!atomic_long_try_cmpxchg(&pages->source_user->locked_vm,
&cur_pages, new_pages));
return 0;
}
static void decr_user_locked_vm(struct iopt_pages *pages, unsigned long npages)
{
if (WARN_ON(atomic_long_read(&pages->source_user->locked_vm) < npages))
return;
atomic_long_sub(npages, &pages->source_user->locked_vm);
}
static int update_mm_locked_vm(struct iopt_pages *pages, unsigned long npages,
bool inc, struct pfn_reader_user *user)
{
bool do_put = false;
int rc;
if (user && user->locked) {
mmap_read_unlock(pages->source_mm);
user->locked = 0;
} else if ((!user || (!user->upages && !user->ufolios)) &&
pages->source_mm != current->mm) {
if (!mmget_not_zero(pages->source_mm))
return -EINVAL;
do_put = true;
}
mmap_write_lock(pages->source_mm);
rc = __account_locked_vm(pages->source_mm, npages, inc,
pages->source_task, false);
mmap_write_unlock(pages->source_mm);
if (do_put)
mmput(pages->source_mm);
return rc;
}
int iopt_pages_update_pinned(struct iopt_pages *pages, unsigned long npages,
bool inc, struct pfn_reader_user *user)
{
int rc = 0;
switch (pages->account_mode) {
case IOPT_PAGES_ACCOUNT_NONE:
break;
case IOPT_PAGES_ACCOUNT_USER:
if (inc)
rc = incr_user_locked_vm(pages, npages);
else
decr_user_locked_vm(pages, npages);
break;
case IOPT_PAGES_ACCOUNT_MM:
rc = update_mm_locked_vm(pages, npages, inc, user);
break;
}
if (rc)
return rc;
pages->last_npinned = pages->npinned;
if (inc)
atomic64_add(npages, &pages->source_mm->pinned_vm);
else
atomic64_sub(npages, &pages->source_mm->pinned_vm);
return 0;
}
static void update_unpinned(struct iopt_pages *pages)
{
if (WARN_ON(pages->npinned > pages->last_npinned))
return;
if (pages->npinned == pages->last_npinned)
return;
iopt_pages_update_pinned(pages, pages->last_npinned - pages->npinned,
false, NULL);
}
static int pfn_reader_user_update_pinned(struct pfn_reader_user *user,
struct iopt_pages *pages)
{
unsigned long npages;
bool inc;
lockdep_assert_held(&pages->mutex);
if (pages->npinned == pages->last_npinned)
return 0;
if (pages->npinned < pages->last_npinned) {
npages = pages->last_npinned - pages->npinned;
inc = false;
} else {
if (iommufd_should_fail())
return -ENOMEM;
npages = pages->npinned - pages->last_npinned;
inc = true;
}
return iopt_pages_update_pinned(pages, npages, inc, user);
}
struct pfn_reader_dmabuf {
struct phys_vec phys;
unsigned long start_offset;
};
static int pfn_reader_dmabuf_init(struct pfn_reader_dmabuf *dmabuf,
struct iopt_pages *pages)
{
if (WARN_ON(iopt_dmabuf_revoked(pages)))
return -EINVAL;
dmabuf->phys = pages->dmabuf.phys;
dmabuf->start_offset = pages->dmabuf.start;
return 0;
}
static int pfn_reader_fill_dmabuf(struct pfn_reader_dmabuf *dmabuf,
struct pfn_batch *batch,
unsigned long start_index,
unsigned long last_index)
{
unsigned long start = dmabuf->start_offset + start_index * PAGE_SIZE;
batch_add_pfn_num(batch, PHYS_PFN(dmabuf->phys.paddr + start),
last_index - start_index + 1, BATCH_MMIO);
return 0;
}
struct pfn_reader {
struct iopt_pages *pages;
struct interval_tree_double_span_iter span;
struct pfn_batch batch;
unsigned long batch_start_index;
unsigned long batch_end_index;
unsigned long last_index;
union {
struct pfn_reader_user user;
struct pfn_reader_dmabuf dmabuf;
};
};
static int pfn_reader_update_pinned(struct pfn_reader *pfns)
{
return pfn_reader_user_update_pinned(&pfns->user, pfns->pages);
}
static void pfn_reader_unpin(struct pfn_reader *pfns)
{
unsigned long last = pfns->batch_end_index - 1;
unsigned long start = pfns->batch_start_index;
struct interval_tree_double_span_iter span;
struct iopt_pages *pages = pfns->pages;
lockdep_assert_held(&pages->mutex);
interval_tree_for_each_double_span(&span, &pages->access_itree,
&pages->domains_itree, start, last) {
if (span.is_used)
continue;
batch_unpin(&pfns->batch, pages, span.start_hole - start,
span.last_hole - span.start_hole + 1);
}
}
static int pfn_reader_fill_span(struct pfn_reader *pfns)
{
struct interval_tree_double_span_iter *span = &pfns->span;
unsigned long start_index = pfns->batch_end_index;
struct pfn_reader_user *user;
unsigned long npages;
struct iopt_area *area;
int rc;
if (IS_ENABLED(CONFIG_IOMMUFD_TEST) &&
WARN_ON(span->last_used < start_index))
return -EINVAL;
if (span->is_used == 1) {
batch_from_xarray(&pfns->batch, &pfns->pages->pinned_pfns,
start_index, span->last_used);
return 0;
}
if (span->is_used == 2) {
area = iopt_pages_find_domain_area(pfns->pages, start_index);
if (WARN_ON(!area))
return -EINVAL;
batch_from_domain(
&pfns->batch, area->storage_domain, area, start_index,
min(iopt_area_last_index(area), span->last_used));
return 0;
}
if (iopt_is_dmabuf(pfns->pages))
return pfn_reader_fill_dmabuf(&pfns->dmabuf, &pfns->batch,
start_index, span->last_hole);
user = &pfns->user;
if (start_index >= user->upages_end) {
rc = pfn_reader_user_pin(user, pfns->pages, start_index,
span->last_hole);
if (rc)
return rc;
}
npages = user->upages_end - start_index;
start_index -= user->upages_start;
rc = 0;
if (!user->file)
batch_from_pages(&pfns->batch, user->upages + start_index,
npages);
else
rc = batch_from_folios(&pfns->batch, &user->ufolios_next,
&user->ufolios_offset, npages);
return rc;
}
static bool pfn_reader_done(struct pfn_reader *pfns)
{
return pfns->batch_start_index == pfns->last_index + 1;
}
static int pfn_reader_next(struct pfn_reader *pfns)
{
int rc;
batch_clear(&pfns->batch);
pfns->batch_start_index = pfns->batch_end_index;
while (pfns->batch_end_index != pfns->last_index + 1) {
unsigned int npfns = pfns->batch.total_pfns;
if (IS_ENABLED(CONFIG_IOMMUFD_TEST) &&
WARN_ON(interval_tree_double_span_iter_done(&pfns->span)))
return -EINVAL;
rc = pfn_reader_fill_span(pfns);
if (rc)
return rc;
if (WARN_ON(!pfns->batch.total_pfns))
return -EINVAL;
pfns->batch_end_index =
pfns->batch_start_index + pfns->batch.total_pfns;
if (pfns->batch_end_index == pfns->span.last_used + 1)
interval_tree_double_span_iter_next(&pfns->span);
if (npfns == pfns->batch.total_pfns)
return 0;
}
return 0;
}
static int pfn_reader_init(struct pfn_reader *pfns, struct iopt_pages *pages,
unsigned long start_index, unsigned long last_index)
{
int rc;
lockdep_assert_held(&pages->mutex);
pfns->pages = pages;
pfns->batch_start_index = start_index;
pfns->batch_end_index = start_index;
pfns->last_index = last_index;
if (iopt_is_dmabuf(pages))
pfn_reader_dmabuf_init(&pfns->dmabuf, pages);
else
pfn_reader_user_init(&pfns->user, pages);
rc = batch_init(&pfns->batch, last_index - start_index + 1);
if (rc)
return rc;
interval_tree_double_span_iter_first(&pfns->span, &pages->access_itree,
&pages->domains_itree, start_index,
last_index);
return 0;
}
static void pfn_reader_release_pins(struct pfn_reader *pfns)
{
struct iopt_pages *pages = pfns->pages;
struct pfn_reader_user *user;
if (iopt_is_dmabuf(pages))
return;
user = &pfns->user;
if (user->upages_end > pfns->batch_end_index) {
unsigned long npages = user->upages_end - pfns->batch_end_index;
unsigned long start_index = pfns->batch_end_index -
user->upages_start;
if (!user->file) {
unpin_user_pages(user->upages + start_index, npages);
} else {
long n = user->ufolios_len / sizeof(*user->ufolios);
unpin_folios(user->ufolios_next,
user->ufolios + n - user->ufolios_next);
}
iopt_pages_sub_npinned(pages, npages);
user->upages_end = pfns->batch_end_index;
}
if (pfns->batch_start_index != pfns->batch_end_index) {
pfn_reader_unpin(pfns);
pfns->batch_start_index = pfns->batch_end_index;
}
}
static void pfn_reader_destroy(struct pfn_reader *pfns)
{
struct iopt_pages *pages = pfns->pages;
pfn_reader_release_pins(pfns);
if (!iopt_is_dmabuf(pfns->pages))
pfn_reader_user_destroy(&pfns->user, pfns->pages);
batch_destroy(&pfns->batch, NULL);
WARN_ON(pages->last_npinned != pages->npinned);
}
static int pfn_reader_first(struct pfn_reader *pfns, struct iopt_pages *pages,
unsigned long start_index, unsigned long last_index)
{
int rc;
if (IS_ENABLED(CONFIG_IOMMUFD_TEST) &&
WARN_ON(last_index < start_index))
return -EINVAL;
rc = pfn_reader_init(pfns, pages, start_index, last_index);
if (rc)
return rc;
rc = pfn_reader_next(pfns);
if (rc) {
pfn_reader_destroy(pfns);
return rc;
}
return 0;
}
static struct iopt_pages *iopt_alloc_pages(unsigned long start_byte,
unsigned long length, bool writable)
{
struct iopt_pages *pages;
if (length > SIZE_MAX - PAGE_SIZE || length == 0)
return ERR_PTR(-EINVAL);
pages = kzalloc_obj(*pages, GFP_KERNEL_ACCOUNT);
if (!pages)
return ERR_PTR(-ENOMEM);
kref_init(&pages->kref);
xa_init_flags(&pages->pinned_pfns, XA_FLAGS_ACCOUNT);
mutex_init(&pages->mutex);
pages->source_mm = current->mm;
mmgrab(pages->source_mm);
pages->npages = DIV_ROUND_UP(length + start_byte, PAGE_SIZE);
pages->access_itree = RB_ROOT_CACHED;
pages->domains_itree = RB_ROOT_CACHED;
pages->writable = writable;
if (capable(CAP_IPC_LOCK))
pages->account_mode = IOPT_PAGES_ACCOUNT_NONE;
else
pages->account_mode = IOPT_PAGES_ACCOUNT_USER;
pages->source_task = current->group_leader;
get_task_struct(current->group_leader);
pages->source_user = get_uid(current_user());
return pages;
}
struct iopt_pages *iopt_alloc_user_pages(void __user *uptr,
unsigned long length, bool writable)
{
struct iopt_pages *pages;
unsigned long end;
void __user *uptr_down =
(void __user *)ALIGN_DOWN((uintptr_t)uptr, PAGE_SIZE);
if (check_add_overflow((unsigned long)uptr, length, &end))
return ERR_PTR(-EOVERFLOW);
pages = iopt_alloc_pages(uptr - uptr_down, length, writable);
if (IS_ERR(pages))
return pages;
pages->uptr = uptr_down;
pages->type = IOPT_ADDRESS_USER;
return pages;
}
struct iopt_pages *iopt_alloc_file_pages(struct file *file,
unsigned long start_byte,
unsigned long start,
unsigned long length, bool writable)
{
struct iopt_pages *pages;
pages = iopt_alloc_pages(start_byte, length, writable);
if (IS_ERR(pages))
return pages;
pages->file = get_file(file);
pages->start = start - start_byte;
pages->type = IOPT_ADDRESS_FILE;
return pages;
}
static void iopt_revoke_notify(struct dma_buf_attachment *attach)
{
struct iopt_pages *pages = attach->importer_priv;
struct iopt_pages_dmabuf_track *track;
guard(mutex)(&pages->mutex);
if (iopt_dmabuf_revoked(pages))
return;
list_for_each_entry(track, &pages->dmabuf.tracker, elm) {
struct iopt_area *area = track->area;
iopt_area_unmap_domain_range(area, track->domain,
iopt_area_index(area),
iopt_area_last_index(area));
}
pages->dmabuf.phys.len = 0;
}
static struct dma_buf_attach_ops iopt_dmabuf_attach_revoke_ops = {
.allow_peer2peer = true,
.move_notify = iopt_revoke_notify,
};
static int
sym_vfio_pci_dma_buf_iommufd_map(struct dma_buf_attachment *attachment,
struct phys_vec *phys)
{
typeof(&vfio_pci_dma_buf_iommufd_map) fn;
int rc;
rc = iommufd_test_dma_buf_iommufd_map(attachment, phys);
if (rc != -EOPNOTSUPP)
return rc;
if (!IS_ENABLED(CONFIG_VFIO_PCI_DMABUF))
return -EOPNOTSUPP;
fn = symbol_get(vfio_pci_dma_buf_iommufd_map);
if (!fn)
return -EOPNOTSUPP;
rc = fn(attachment, phys);
symbol_put(vfio_pci_dma_buf_iommufd_map);
return rc;
}
static int iopt_map_dmabuf(struct iommufd_ctx *ictx, struct iopt_pages *pages,
struct dma_buf *dmabuf)
{
struct dma_buf_attachment *attach;
int rc;
attach = dma_buf_dynamic_attach(dmabuf, iommufd_global_device(),
&iopt_dmabuf_attach_revoke_ops, pages);
if (IS_ERR(attach))
return PTR_ERR(attach);
dma_resv_lock(dmabuf->resv, NULL);
if (IS_ENABLED(CONFIG_LOCKDEP)) {
mutex_lock(&pages->mutex);
mutex_unlock(&pages->mutex);
}
rc = sym_vfio_pci_dma_buf_iommufd_map(attach, &pages->dmabuf.phys);
if (rc)
goto err_detach;
dma_resv_unlock(dmabuf->resv);
pages->dmabuf.attach = attach;
return 0;
err_detach:
dma_resv_unlock(dmabuf->resv);
dma_buf_detach(dmabuf, attach);
return rc;
}
struct iopt_pages *iopt_alloc_dmabuf_pages(struct iommufd_ctx *ictx,
struct dma_buf *dmabuf,
unsigned long start_byte,
unsigned long start,
unsigned long length, bool writable)
{
static struct lock_class_key pages_dmabuf_mutex_key;
struct iopt_pages *pages;
int rc;
if (!IS_ENABLED(CONFIG_DMA_SHARED_BUFFER))
return ERR_PTR(-EOPNOTSUPP);
if (dmabuf->size <= (start + length - 1) ||
length / PAGE_SIZE >= MAX_NPFNS)
return ERR_PTR(-EINVAL);
pages = iopt_alloc_pages(start_byte, length, writable);
if (IS_ERR(pages))
return pages;
lockdep_set_class(&pages->mutex, &pages_dmabuf_mutex_key);
pages->account_mode = IOPT_PAGES_ACCOUNT_NONE;
pages->type = IOPT_ADDRESS_DMABUF;
pages->dmabuf.start = start - start_byte;
INIT_LIST_HEAD(&pages->dmabuf.tracker);
rc = iopt_map_dmabuf(ictx, pages, dmabuf);
if (rc) {
iopt_put_pages(pages);
return ERR_PTR(rc);
}
return pages;
}
int iopt_dmabuf_track_domain(struct iopt_pages *pages, struct iopt_area *area,
struct iommu_domain *domain)
{
struct iopt_pages_dmabuf_track *track;
lockdep_assert_held(&pages->mutex);
if (WARN_ON(!iopt_is_dmabuf(pages)))
return -EINVAL;
list_for_each_entry(track, &pages->dmabuf.tracker, elm)
if (WARN_ON(track->domain == domain && track->area == area))
return -EINVAL;
track = kzalloc_obj(*track);
if (!track)
return -ENOMEM;
track->domain = domain;
track->area = area;
list_add_tail(&track->elm, &pages->dmabuf.tracker);
return 0;
}
void iopt_dmabuf_untrack_domain(struct iopt_pages *pages,
struct iopt_area *area,
struct iommu_domain *domain)
{
struct iopt_pages_dmabuf_track *track;
lockdep_assert_held(&pages->mutex);
WARN_ON(!iopt_is_dmabuf(pages));
list_for_each_entry(track, &pages->dmabuf.tracker, elm) {
if (track->domain == domain && track->area == area) {
list_del(&track->elm);
kfree(track);
return;
}
}
WARN_ON(true);
}
int iopt_dmabuf_track_all_domains(struct iopt_area *area,
struct iopt_pages *pages)
{
struct iopt_pages_dmabuf_track *track;
struct iommu_domain *domain;
unsigned long index;
int rc;
list_for_each_entry(track, &pages->dmabuf.tracker, elm)
if (WARN_ON(track->area == area))
return -EINVAL;
xa_for_each(&area->iopt->domains, index, domain) {
rc = iopt_dmabuf_track_domain(pages, area, domain);
if (rc)
goto err_untrack;
}
return 0;
err_untrack:
iopt_dmabuf_untrack_all_domains(area, pages);
return rc;
}
void iopt_dmabuf_untrack_all_domains(struct iopt_area *area,
struct iopt_pages *pages)
{
struct iopt_pages_dmabuf_track *track;
struct iopt_pages_dmabuf_track *tmp;
list_for_each_entry_safe(track, tmp, &pages->dmabuf.tracker,
elm) {
if (track->area == area) {
list_del(&track->elm);
kfree(track);
}
}
}
void iopt_release_pages(struct kref *kref)
{
struct iopt_pages *pages = container_of(kref, struct iopt_pages, kref);
WARN_ON(!RB_EMPTY_ROOT(&pages->access_itree.rb_root));
WARN_ON(!RB_EMPTY_ROOT(&pages->domains_itree.rb_root));
WARN_ON(pages->npinned);
WARN_ON(!xa_empty(&pages->pinned_pfns));
mmdrop(pages->source_mm);
mutex_destroy(&pages->mutex);
put_task_struct(pages->source_task);
free_uid(pages->source_user);
if (iopt_is_dmabuf(pages) && pages->dmabuf.attach) {
struct dma_buf *dmabuf = pages->dmabuf.attach->dmabuf;
dma_buf_detach(dmabuf, pages->dmabuf.attach);
dma_buf_put(dmabuf);
WARN_ON(!list_empty(&pages->dmabuf.tracker));
} else if (pages->type == IOPT_ADDRESS_FILE) {
fput(pages->file);
}
kfree(pages);
}
static void
iopt_area_unpin_domain(struct pfn_batch *batch, struct iopt_area *area,
struct iopt_pages *pages, struct iommu_domain *domain,
unsigned long start_index, unsigned long last_index,
unsigned long *unmapped_end_index,
unsigned long real_last_index)
{
while (start_index <= last_index) {
unsigned long batch_last_index;
if (*unmapped_end_index <= last_index) {
unsigned long start =
max(start_index, *unmapped_end_index);
if (IS_ENABLED(CONFIG_IOMMUFD_TEST) &&
batch->total_pfns)
WARN_ON(*unmapped_end_index -
batch->total_pfns !=
start_index);
batch_from_domain(batch, domain, area, start,
last_index);
batch_last_index = start_index + batch->total_pfns - 1;
} else {
batch_last_index = last_index;
}
if (IS_ENABLED(CONFIG_IOMMUFD_TEST))
WARN_ON(batch_last_index > real_last_index);
if (batch_last_index == last_index &&
last_index != real_last_index)
batch_from_domain_continue(batch, domain, area,
last_index + 1,
real_last_index);
if (*unmapped_end_index <= batch_last_index) {
iopt_area_unmap_domain_range(
area, domain, *unmapped_end_index,
start_index + batch->total_pfns - 1);
*unmapped_end_index = start_index + batch->total_pfns;
}
batch_unpin(batch, pages, 0,
batch_last_index - start_index + 1);
start_index = batch_last_index + 1;
batch_clear_carry(batch,
*unmapped_end_index - batch_last_index - 1);
}
}
static void __iopt_area_unfill_domain(struct iopt_area *area,
struct iopt_pages *pages,
struct iommu_domain *domain,
unsigned long last_index)
{
struct interval_tree_double_span_iter span;
unsigned long start_index = iopt_area_index(area);
unsigned long unmapped_end_index = start_index;
u64 backup[BATCH_BACKUP_SIZE];
struct pfn_batch batch;
lockdep_assert_held(&pages->mutex);
if (iopt_is_dmabuf(pages)) {
if (WARN_ON(iopt_dmabuf_revoked(pages)))
return;
iopt_area_unmap_domain_range(area, domain, start_index,
last_index);
return;
}
batch_init_backup(&batch, last_index + 1, backup, sizeof(backup));
interval_tree_for_each_double_span(&span, &pages->domains_itree,
&pages->access_itree, start_index,
last_index) {
if (span.is_used) {
batch_skip_carry(&batch,
span.last_used - span.start_used + 1);
continue;
}
iopt_area_unpin_domain(&batch, area, pages, domain,
span.start_hole, span.last_hole,
&unmapped_end_index, last_index);
}
if (unmapped_end_index != last_index + 1)
iopt_area_unmap_domain_range(area, domain, unmapped_end_index,
last_index);
WARN_ON(batch.total_pfns);
batch_destroy(&batch, backup);
update_unpinned(pages);
}
static void iopt_area_unfill_partial_domain(struct iopt_area *area,
struct iopt_pages *pages,
struct iommu_domain *domain,
unsigned long end_index)
{
if (end_index != iopt_area_index(area))
__iopt_area_unfill_domain(area, pages, domain, end_index - 1);
}
void iopt_area_unmap_domain(struct iopt_area *area, struct iommu_domain *domain)
{
iommu_unmap_nofail(domain, iopt_area_iova(area),
iopt_area_length(area));
}
void iopt_area_unfill_domain(struct iopt_area *area, struct iopt_pages *pages,
struct iommu_domain *domain)
{
if (iopt_dmabuf_revoked(pages))
return;
__iopt_area_unfill_domain(area, pages, domain,
iopt_area_last_index(area));
}
int iopt_area_fill_domain(struct iopt_area *area, struct iommu_domain *domain)
{
unsigned long done_end_index;
struct pfn_reader pfns;
int rc;
lockdep_assert_held(&area->pages->mutex);
if (iopt_dmabuf_revoked(area->pages))
return 0;
rc = pfn_reader_first(&pfns, area->pages, iopt_area_index(area),
iopt_area_last_index(area));
if (rc)
return rc;
while (!pfn_reader_done(&pfns)) {
done_end_index = pfns.batch_start_index;
rc = batch_to_domain(&pfns.batch, domain, area,
pfns.batch_start_index);
if (rc)
goto out_unmap;
done_end_index = pfns.batch_end_index;
rc = pfn_reader_next(&pfns);
if (rc)
goto out_unmap;
}
rc = pfn_reader_update_pinned(&pfns);
if (rc)
goto out_unmap;
goto out_destroy;
out_unmap:
pfn_reader_release_pins(&pfns);
iopt_area_unfill_partial_domain(area, area->pages, domain,
done_end_index);
out_destroy:
pfn_reader_destroy(&pfns);
return rc;
}
int iopt_area_fill_domains(struct iopt_area *area, struct iopt_pages *pages)
{
unsigned long done_first_end_index;
unsigned long done_all_end_index;
struct iommu_domain *domain;
unsigned long unmap_index;
struct pfn_reader pfns;
unsigned long index;
int rc;
lockdep_assert_held(&area->iopt->domains_rwsem);
if (xa_empty(&area->iopt->domains))
return 0;
mutex_lock(&pages->mutex);
if (iopt_is_dmabuf(pages)) {
rc = iopt_dmabuf_track_all_domains(area, pages);
if (rc)
goto out_unlock;
}
if (!iopt_dmabuf_revoked(pages)) {
rc = pfn_reader_first(&pfns, pages, iopt_area_index(area),
iopt_area_last_index(area));
if (rc)
goto out_untrack;
while (!pfn_reader_done(&pfns)) {
done_first_end_index = pfns.batch_end_index;
done_all_end_index = pfns.batch_start_index;
xa_for_each(&area->iopt->domains, index, domain) {
rc = batch_to_domain(&pfns.batch, domain, area,
pfns.batch_start_index);
if (rc)
goto out_unmap;
}
done_all_end_index = done_first_end_index;
rc = pfn_reader_next(&pfns);
if (rc)
goto out_unmap;
}
rc = pfn_reader_update_pinned(&pfns);
if (rc)
goto out_unmap;
pfn_reader_destroy(&pfns);
}
area->storage_domain = xa_load(&area->iopt->domains, 0);
interval_tree_insert(&area->pages_node, &pages->domains_itree);
mutex_unlock(&pages->mutex);
return 0;
out_unmap:
pfn_reader_release_pins(&pfns);
xa_for_each(&area->iopt->domains, unmap_index, domain) {
unsigned long end_index;
if (unmap_index < index)
end_index = done_first_end_index;
else
end_index = done_all_end_index;
if (unmap_index != area->iopt->next_domain_id - 1) {
if (end_index != iopt_area_index(area))
iopt_area_unmap_domain_range(
area, domain, iopt_area_index(area),
end_index - 1);
} else {
iopt_area_unfill_partial_domain(area, pages, domain,
end_index);
}
}
pfn_reader_destroy(&pfns);
out_untrack:
if (iopt_is_dmabuf(pages))
iopt_dmabuf_untrack_all_domains(area, pages);
out_unlock:
mutex_unlock(&pages->mutex);
return rc;
}
void iopt_area_unfill_domains(struct iopt_area *area, struct iopt_pages *pages)
{
struct io_pagetable *iopt = area->iopt;
struct iommu_domain *domain;
unsigned long index;
lockdep_assert_held(&iopt->domains_rwsem);
mutex_lock(&pages->mutex);
if (!area->storage_domain)
goto out_unlock;
xa_for_each(&iopt->domains, index, domain) {
if (domain == area->storage_domain)
continue;
if (!iopt_dmabuf_revoked(pages))
iopt_area_unmap_domain_range(
area, domain, iopt_area_index(area),
iopt_area_last_index(area));
}
if (IS_ENABLED(CONFIG_IOMMUFD_TEST))
WARN_ON(RB_EMPTY_NODE(&area->pages_node.rb));
interval_tree_remove(&area->pages_node, &pages->domains_itree);
iopt_area_unfill_domain(area, pages, area->storage_domain);
if (iopt_is_dmabuf(pages))
iopt_dmabuf_untrack_all_domains(area, pages);
area->storage_domain = NULL;
out_unlock:
mutex_unlock(&pages->mutex);
}
static void iopt_pages_unpin_xarray(struct pfn_batch *batch,
struct iopt_pages *pages,
unsigned long start_index,
unsigned long end_index)
{
while (start_index <= end_index) {
batch_from_xarray_clear(batch, &pages->pinned_pfns, start_index,
end_index);
batch_unpin(batch, pages, 0, batch->total_pfns);
start_index += batch->total_pfns;
batch_clear(batch);
}
}
void iopt_pages_unfill_xarray(struct iopt_pages *pages,
unsigned long start_index,
unsigned long last_index)
{
struct interval_tree_double_span_iter span;
u64 backup[BATCH_BACKUP_SIZE];
struct pfn_batch batch;
bool batch_inited = false;
lockdep_assert_held(&pages->mutex);
interval_tree_for_each_double_span(&span, &pages->access_itree,
&pages->domains_itree, start_index,
last_index) {
if (!span.is_used) {
if (!batch_inited) {
batch_init_backup(&batch,
last_index - start_index + 1,
backup, sizeof(backup));
batch_inited = true;
}
iopt_pages_unpin_xarray(&batch, pages, span.start_hole,
span.last_hole);
} else if (span.is_used == 2) {
clear_xarray(&pages->pinned_pfns, span.start_used,
span.last_used);
}
}
if (batch_inited)
batch_destroy(&batch, backup);
update_unpinned(pages);
}
void iopt_pages_fill_from_xarray(struct iopt_pages *pages,
unsigned long start_index,
unsigned long last_index,
struct page **out_pages)
{
XA_STATE(xas, &pages->pinned_pfns, start_index);
void *entry;
rcu_read_lock();
while (start_index <= last_index) {
entry = xas_next(&xas);
if (xas_retry(&xas, entry))
continue;
WARN_ON(!xa_is_value(entry));
*(out_pages++) = pfn_to_page(xa_to_value(entry));
start_index++;
}
rcu_read_unlock();
}
static int iopt_pages_fill_from_domain(struct iopt_pages *pages,
unsigned long start_index,
unsigned long last_index,
struct page **out_pages)
{
while (start_index != last_index + 1) {
unsigned long domain_last;
struct iopt_area *area;
area = iopt_pages_find_domain_area(pages, start_index);
if (WARN_ON(!area))
return -EINVAL;
domain_last = min(iopt_area_last_index(area), last_index);
out_pages = raw_pages_from_domain(area->storage_domain, area,
start_index, domain_last,
out_pages);
start_index = domain_last + 1;
}
return 0;
}
static int iopt_pages_fill(struct iopt_pages *pages,
struct pfn_reader_user *user,
unsigned long start_index,
unsigned long last_index,
struct page **out_pages)
{
unsigned long cur_index = start_index;
int rc;
while (cur_index != last_index + 1) {
user->upages = out_pages + (cur_index - start_index);
rc = pfn_reader_user_pin(user, pages, cur_index, last_index);
if (rc)
goto out_unpin;
cur_index = user->upages_end;
}
return 0;
out_unpin:
if (start_index != cur_index)
iopt_pages_err_unpin(pages, start_index, cur_index - 1,
out_pages);
return rc;
}
int iopt_pages_fill_xarray(struct iopt_pages *pages, unsigned long start_index,
unsigned long last_index, struct page **out_pages)
{
struct interval_tree_double_span_iter span;
unsigned long xa_end = start_index;
struct pfn_reader_user user;
int rc;
lockdep_assert_held(&pages->mutex);
pfn_reader_user_init(&user, pages);
user.upages_len = (last_index - start_index + 1) * sizeof(*out_pages);
interval_tree_for_each_double_span(&span, &pages->access_itree,
&pages->domains_itree, start_index,
last_index) {
struct page **cur_pages;
if (span.is_used == 1) {
cur_pages = out_pages + (span.start_used - start_index);
iopt_pages_fill_from_xarray(pages, span.start_used,
span.last_used, cur_pages);
continue;
}
if (span.is_used == 2) {
cur_pages = out_pages + (span.start_used - start_index);
iopt_pages_fill_from_domain(pages, span.start_used,
span.last_used, cur_pages);
rc = pages_to_xarray(&pages->pinned_pfns,
span.start_used, span.last_used,
cur_pages);
if (rc)
goto out_clean_xa;
xa_end = span.last_used + 1;
continue;
}
cur_pages = out_pages + (span.start_hole - start_index);
rc = iopt_pages_fill(pages, &user, span.start_hole,
span.last_hole, cur_pages);
if (rc)
goto out_clean_xa;
rc = pages_to_xarray(&pages->pinned_pfns, span.start_hole,
span.last_hole, cur_pages);
if (rc) {
iopt_pages_err_unpin(pages, span.start_hole,
span.last_hole, cur_pages);
goto out_clean_xa;
}
xa_end = span.last_hole + 1;
}
rc = pfn_reader_user_update_pinned(&user, pages);
if (rc)
goto out_clean_xa;
user.upages = NULL;
pfn_reader_user_destroy(&user, pages);
return 0;
out_clean_xa:
if (start_index != xa_end)
iopt_pages_unfill_xarray(pages, start_index, xa_end - 1);
user.upages = NULL;
pfn_reader_user_destroy(&user, pages);
return rc;
}
static int iopt_pages_rw_slow(struct iopt_pages *pages,
unsigned long start_index,
unsigned long last_index, unsigned long offset,
void *data, unsigned long length,
unsigned int flags)
{
struct pfn_reader pfns;
int rc;
mutex_lock(&pages->mutex);
rc = pfn_reader_first(&pfns, pages, start_index, last_index);
if (rc)
goto out_unlock;
while (!pfn_reader_done(&pfns)) {
unsigned long done;
done = batch_rw(&pfns.batch, data, offset, length, flags);
data += done;
length -= done;
offset = 0;
pfn_reader_unpin(&pfns);
rc = pfn_reader_next(&pfns);
if (rc)
goto out_destroy;
}
if (WARN_ON(length != 0))
rc = -EINVAL;
out_destroy:
pfn_reader_destroy(&pfns);
out_unlock:
mutex_unlock(&pages->mutex);
return rc;
}
static int iopt_pages_rw_page(struct iopt_pages *pages, unsigned long index,
unsigned long offset, void *data,
unsigned long length, unsigned int flags)
{
struct page *page = NULL;
int rc;
if (IS_ENABLED(CONFIG_IOMMUFD_TEST) &&
WARN_ON(pages->type != IOPT_ADDRESS_USER))
return -EINVAL;
if (!mmget_not_zero(pages->source_mm))
return iopt_pages_rw_slow(pages, index, index, offset, data,
length, flags);
if (iommufd_should_fail()) {
rc = -EINVAL;
goto out_mmput;
}
mmap_read_lock(pages->source_mm);
rc = pin_user_pages_remote(
pages->source_mm, (uintptr_t)(pages->uptr + index * PAGE_SIZE),
1, (flags & IOMMUFD_ACCESS_RW_WRITE) ? FOLL_WRITE : 0, &page,
NULL);
mmap_read_unlock(pages->source_mm);
if (rc != 1) {
if (WARN_ON(rc >= 0))
rc = -EINVAL;
goto out_mmput;
}
copy_data_page(page, data, offset, length, flags);
unpin_user_page(page);
rc = 0;
out_mmput:
mmput(pages->source_mm);
return rc;
}
int iopt_pages_rw_access(struct iopt_pages *pages, unsigned long start_byte,
void *data, unsigned long length, unsigned int flags)
{
unsigned long start_index = start_byte / PAGE_SIZE;
unsigned long last_index = (start_byte + length - 1) / PAGE_SIZE;
bool change_mm = current->mm != pages->source_mm;
int rc = 0;
if (IS_ENABLED(CONFIG_IOMMUFD_TEST) &&
(flags & __IOMMUFD_ACCESS_RW_SLOW_PATH))
change_mm = true;
if ((flags & IOMMUFD_ACCESS_RW_WRITE) && !pages->writable)
return -EPERM;
if (iopt_is_dmabuf(pages))
return -EINVAL;
if (pages->type != IOPT_ADDRESS_USER)
return iopt_pages_rw_slow(pages, start_index, last_index,
start_byte % PAGE_SIZE, data, length,
flags);
if (!(flags & IOMMUFD_ACCESS_RW_KTHREAD) && change_mm) {
if (start_index == last_index)
return iopt_pages_rw_page(pages, start_index,
start_byte % PAGE_SIZE, data,
length, flags);
return iopt_pages_rw_slow(pages, start_index, last_index,
start_byte % PAGE_SIZE, data, length,
flags);
}
if (change_mm) {
if (!mmget_not_zero(pages->source_mm))
return iopt_pages_rw_slow(pages, start_index,
last_index,
start_byte % PAGE_SIZE, data,
length, flags);
kthread_use_mm(pages->source_mm);
}
if (flags & IOMMUFD_ACCESS_RW_WRITE) {
if (copy_to_user(pages->uptr + start_byte, data, length))
rc = -EFAULT;
} else {
if (copy_from_user(data, pages->uptr + start_byte, length))
rc = -EFAULT;
}
if (change_mm) {
kthread_unuse_mm(pages->source_mm);
mmput(pages->source_mm);
}
return rc;
}
static struct iopt_pages_access *
iopt_pages_get_exact_access(struct iopt_pages *pages, unsigned long index,
unsigned long last)
{
struct interval_tree_node *node;
lockdep_assert_held(&pages->mutex);
for (node = interval_tree_iter_first(&pages->access_itree, index, last);
node; node = interval_tree_iter_next(node, index, last))
if (node->start == index && node->last == last)
return container_of(node, struct iopt_pages_access,
node);
return NULL;
}
int iopt_area_add_access(struct iopt_area *area, unsigned long start_index,
unsigned long last_index, struct page **out_pages,
unsigned int flags, bool lock_area)
{
struct iopt_pages *pages = area->pages;
struct iopt_pages_access *access;
int rc;
if ((flags & IOMMUFD_ACCESS_RW_WRITE) && !pages->writable)
return -EPERM;
mutex_lock(&pages->mutex);
access = iopt_pages_get_exact_access(pages, start_index, last_index);
if (access) {
area->num_accesses++;
if (lock_area)
area->num_locks++;
access->users++;
iopt_pages_fill_from_xarray(pages, start_index, last_index,
out_pages);
mutex_unlock(&pages->mutex);
return 0;
}
access = kzalloc_obj(*access, GFP_KERNEL_ACCOUNT);
if (!access) {
rc = -ENOMEM;
goto err_unlock;
}
rc = iopt_pages_fill_xarray(pages, start_index, last_index, out_pages);
if (rc)
goto err_free;
access->node.start = start_index;
access->node.last = last_index;
access->users = 1;
area->num_accesses++;
if (lock_area)
area->num_locks++;
interval_tree_insert(&access->node, &pages->access_itree);
mutex_unlock(&pages->mutex);
return 0;
err_free:
kfree(access);
err_unlock:
mutex_unlock(&pages->mutex);
return rc;
}
void iopt_area_remove_access(struct iopt_area *area, unsigned long start_index,
unsigned long last_index, bool unlock_area)
{
struct iopt_pages *pages = area->pages;
struct iopt_pages_access *access;
mutex_lock(&pages->mutex);
access = iopt_pages_get_exact_access(pages, start_index, last_index);
if (WARN_ON(!access))
goto out_unlock;
WARN_ON(area->num_accesses == 0 || access->users == 0);
if (unlock_area) {
WARN_ON(area->num_locks == 0);
area->num_locks--;
}
area->num_accesses--;
access->users--;
if (access->users)
goto out_unlock;
interval_tree_remove(&access->node, &pages->access_itree);
iopt_pages_unfill_xarray(pages, start_index, last_index);
kfree(access);
out_unlock:
mutex_unlock(&pages->mutex);
}