root/drivers/iommu/iommufd/iommufd_private.h
/* SPDX-License-Identifier: GPL-2.0-only */
/* Copyright (c) 2021-2022, NVIDIA CORPORATION & AFFILIATES
 */
#ifndef __IOMMUFD_PRIVATE_H
#define __IOMMUFD_PRIVATE_H

#include <linux/iommu.h>
#include <linux/iommufd.h>
#include <linux/iova_bitmap.h>
#include <linux/maple_tree.h>
#include <linux/rwsem.h>
#include <linux/uaccess.h>
#include <linux/xarray.h>
#include <uapi/linux/iommufd.h>

#include "../iommu-priv.h"

struct iommu_domain;
struct iommu_group;
struct iommu_option;
struct iommufd_device;
struct dma_buf_attachment;

struct iommufd_sw_msi_map {
        struct list_head sw_msi_item;
        phys_addr_t sw_msi_start;
        phys_addr_t msi_addr;
        unsigned int pgoff;
        unsigned int id;
};

/* Bitmap of struct iommufd_sw_msi_map::id */
struct iommufd_sw_msi_maps {
        DECLARE_BITMAP(bitmap, 64);
};

#ifdef CONFIG_IRQ_MSI_IOMMU
int iommufd_sw_msi_install(struct iommufd_ctx *ictx,
                           struct iommufd_hwpt_paging *hwpt_paging,
                           struct iommufd_sw_msi_map *msi_map);
#endif

struct iommufd_ctx {
        struct file *file;
        struct xarray objects;
        struct xarray groups;
        wait_queue_head_t destroy_wait;
        struct rw_semaphore ioas_creation_lock;
        struct maple_tree mt_mmap;

        struct mutex sw_msi_lock;
        struct list_head sw_msi_list;
        unsigned int sw_msi_id;

        u8 account_mode;
        /* Compatibility with VFIO no iommu */
        u8 no_iommu_mode;
        struct iommufd_ioas *vfio_ioas;
};

/* Entry for iommufd_ctx::mt_mmap */
struct iommufd_mmap {
        struct iommufd_object *owner;

        /* Page-shifted start position in mt_mmap to validate vma->vm_pgoff */
        unsigned long vm_pgoff;

        /* Physical range for io_remap_pfn_range() */
        phys_addr_t mmio_addr;
        size_t length;
};

/*
 * The IOVA to PFN map. The map automatically copies the PFNs into multiple
 * domains and permits sharing of PFNs between io_pagetable instances. This
 * supports both a design where IOAS's are 1:1 with a domain (eg because the
 * domain is HW customized), or where the IOAS is 1:N with multiple generic
 * domains.  The io_pagetable holds an interval tree of iopt_areas which point
 * to shared iopt_pages which hold the pfns mapped to the page table.
 *
 * The locking order is domains_rwsem -> iova_rwsem -> pages::mutex
 */
struct io_pagetable {
        struct rw_semaphore domains_rwsem;
        struct xarray domains;
        struct xarray access_list;
        unsigned int next_domain_id;

        struct rw_semaphore iova_rwsem;
        struct rb_root_cached area_itree;
        /* IOVA that cannot become reserved, struct iopt_allowed */
        struct rb_root_cached allowed_itree;
        /* IOVA that cannot be allocated, struct iopt_reserved */
        struct rb_root_cached reserved_itree;
        u8 disable_large_pages;
        unsigned long iova_alignment;
};

void iopt_init_table(struct io_pagetable *iopt);
void iopt_destroy_table(struct io_pagetable *iopt);
int iopt_get_pages(struct io_pagetable *iopt, unsigned long iova,
                   unsigned long length, struct list_head *pages_list);
void iopt_free_pages_list(struct list_head *pages_list);
enum {
        IOPT_ALLOC_IOVA = 1 << 0,
};
int iopt_map_user_pages(struct iommufd_ctx *ictx, struct io_pagetable *iopt,
                        unsigned long *iova, void __user *uptr,
                        unsigned long length, int iommu_prot,
                        unsigned int flags);
int iopt_map_file_pages(struct iommufd_ctx *ictx, struct io_pagetable *iopt,
                        unsigned long *iova, int fd,
                        unsigned long start, unsigned long length,
                        int iommu_prot, unsigned int flags);
int iopt_map_pages(struct io_pagetable *iopt, struct list_head *pages_list,
                   unsigned long length, unsigned long *dst_iova,
                   int iommu_prot, unsigned int flags);
int iopt_unmap_iova(struct io_pagetable *iopt, unsigned long iova,
                    unsigned long length, unsigned long *unmapped);
int iopt_unmap_all(struct io_pagetable *iopt, unsigned long *unmapped);

int iopt_read_and_clear_dirty_data(struct io_pagetable *iopt,
                                   struct iommu_domain *domain,
                                   unsigned long flags,
                                   struct iommu_hwpt_get_dirty_bitmap *bitmap);
int iopt_set_dirty_tracking(struct io_pagetable *iopt,
                            struct iommu_domain *domain, bool enable);

void iommufd_access_notify_unmap(struct io_pagetable *iopt, unsigned long iova,
                                 unsigned long length);
int iopt_table_add_domain(struct io_pagetable *iopt,
                          struct iommu_domain *domain);
void iopt_table_remove_domain(struct io_pagetable *iopt,
                              struct iommu_domain *domain);
int iopt_table_enforce_dev_resv_regions(struct io_pagetable *iopt,
                                        struct device *dev,
                                        phys_addr_t *sw_msi_start);
int iopt_set_allow_iova(struct io_pagetable *iopt,
                        struct rb_root_cached *allowed_iova);
int iopt_reserve_iova(struct io_pagetable *iopt, unsigned long start,
                      unsigned long last, void *owner);
void iopt_remove_reserved_iova(struct io_pagetable *iopt, void *owner);
int iopt_cut_iova(struct io_pagetable *iopt, unsigned long *iovas,
                  size_t num_iovas);
void iopt_enable_large_pages(struct io_pagetable *iopt);
int iopt_disable_large_pages(struct io_pagetable *iopt);

struct iommufd_ucmd {
        struct iommufd_ctx *ictx;
        void __user *ubuffer;
        u32 user_size;
        void *cmd;
        struct iommufd_object *new_obj;
};

int iommufd_vfio_ioctl(struct iommufd_ctx *ictx, unsigned int cmd,
                       unsigned long arg);

/* Copy the response in ucmd->cmd back to userspace. */
static inline int iommufd_ucmd_respond(struct iommufd_ucmd *ucmd,
                                       size_t cmd_len)
{
        if (copy_to_user(ucmd->ubuffer, ucmd->cmd,
                         min_t(size_t, ucmd->user_size, cmd_len)))
                return -EFAULT;
        return 0;
}

static inline bool iommufd_lock_obj(struct iommufd_object *obj)
{
        if (!refcount_inc_not_zero(&obj->users))
                return false;
        if (!refcount_inc_not_zero(&obj->wait_cnt)) {
                /*
                 * If the caller doesn't already have a ref on obj this must be
                 * called under the xa_lock. Otherwise the caller is holding a
                 * ref on users. Thus it cannot be one before this decrement.
                 */
                refcount_dec(&obj->users);
                return false;
        }
        return true;
}

struct iommufd_object *iommufd_get_object(struct iommufd_ctx *ictx, u32 id,
                                          enum iommufd_object_type type);
static inline void iommufd_put_object(struct iommufd_ctx *ictx,
                                      struct iommufd_object *obj)
{
        /*
         * Users first, then wait_cnt so that REMOVE_WAIT never sees a spurious
         * !0 users with a 0 wait_cnt.
         */
        refcount_dec(&obj->users);
        if (refcount_dec_and_test(&obj->wait_cnt))
                wake_up_interruptible_all(&ictx->destroy_wait);
}

void iommufd_object_abort(struct iommufd_ctx *ictx, struct iommufd_object *obj);
void iommufd_object_abort_and_destroy(struct iommufd_ctx *ictx,
                                      struct iommufd_object *obj);
void iommufd_object_finalize(struct iommufd_ctx *ictx,
                             struct iommufd_object *obj);

enum {
        REMOVE_WAIT             = BIT(0),
        REMOVE_OBJ_TOMBSTONE    = BIT(1),
};
int iommufd_object_remove(struct iommufd_ctx *ictx,
                          struct iommufd_object *to_destroy, u32 id,
                          unsigned int flags);

/*
 * The caller holds a users refcount and wants to destroy the object. At this
 * point the caller has no wait_cnt reference and at least the xarray will be
 * holding one.
 */
static inline void iommufd_object_destroy_user(struct iommufd_ctx *ictx,
                                               struct iommufd_object *obj)
{
        int ret;

        ret = iommufd_object_remove(ictx, obj, obj->id, REMOVE_WAIT);

        /*
         * If there is a bug and we couldn't destroy the object then we did put
         * back the caller's users refcount and will eventually try to free it
         * again during close.
         */
        WARN_ON(ret);
}

/*
 * Similar to iommufd_object_destroy_user(), except that the object ID is left
 * reserved/tombstoned.
 */
static inline void iommufd_object_tombstone_user(struct iommufd_ctx *ictx,
                                                 struct iommufd_object *obj)
{
        int ret;

        ret = iommufd_object_remove(ictx, obj, obj->id,
                                    REMOVE_WAIT | REMOVE_OBJ_TOMBSTONE);

        /*
         * If there is a bug and we couldn't destroy the object then we did put
         * back the caller's users refcount and will eventually try to free it
         * again during close.
         */
        WARN_ON(ret);
}

/*
 * The HWPT allocated by autodomains is used in possibly many devices and
 * is automatically destroyed when its refcount reaches zero.
 *
 * If userspace uses the HWPT manually, even for a short term, then it will
 * disrupt this refcounting and the auto-free in the kernel will not work.
 * Userspace that tries to use the automatically allocated HWPT must be careful
 * to ensure that it is consistently destroyed, eg by not racing accesses
 * and by not attaching an automatic HWPT to a device manually.
 */
static inline void
iommufd_object_put_and_try_destroy(struct iommufd_ctx *ictx,
                                   struct iommufd_object *obj)
{
        iommufd_object_remove(ictx, obj, obj->id, 0);
}

/*
 * Callers of these normal object allocators must call iommufd_object_finalize()
 * to finalize the object, or call iommufd_object_abort_and_destroy() to revert
 * the allocation.
 */
struct iommufd_object *_iommufd_object_alloc(struct iommufd_ctx *ictx,
                                             size_t size,
                                             enum iommufd_object_type type);

#define __iommufd_object_alloc(ictx, ptr, type, obj)                           \
        container_of(_iommufd_object_alloc(                                    \
                             ictx,                                             \
                             sizeof(*(ptr)) + BUILD_BUG_ON_ZERO(               \
                                                      offsetof(typeof(*(ptr)), \
                                                               obj) != 0),     \
                             type),                                            \
                     typeof(*(ptr)), obj)

#define iommufd_object_alloc(ictx, ptr, type) \
        __iommufd_object_alloc(ictx, ptr, type, obj)

/*
 * Callers of these _ucmd allocators should not call iommufd_object_finalize()
 * or iommufd_object_abort_and_destroy(), as the core automatically does that.
 */
struct iommufd_object *
_iommufd_object_alloc_ucmd(struct iommufd_ucmd *ucmd, size_t size,
                           enum iommufd_object_type type);

#define __iommufd_object_alloc_ucmd(ucmd, ptr, type, obj)                      \
        container_of(_iommufd_object_alloc_ucmd(                               \
                             ucmd,                                             \
                             sizeof(*(ptr)) + BUILD_BUG_ON_ZERO(               \
                                                      offsetof(typeof(*(ptr)), \
                                                               obj) != 0),     \
                             type),                                            \
                     typeof(*(ptr)), obj)

#define iommufd_object_alloc_ucmd(ucmd, ptr, type) \
        __iommufd_object_alloc_ucmd(ucmd, ptr, type, obj)

/*
 * The IO Address Space (IOAS) pagetable is a virtual page table backed by the
 * io_pagetable object. It is a user controlled mapping of IOVA -> PFNs. The
 * mapping is copied into all of the associated domains and made available to
 * in-kernel users.
 *
 * Every iommu_domain that is created is wrapped in a iommufd_hw_pagetable
 * object. When we go to attach a device to an IOAS we need to get an
 * iommu_domain and wrapping iommufd_hw_pagetable for it.
 *
 * An iommu_domain & iommfd_hw_pagetable will be automatically selected
 * for a device based on the hwpt_list. If no suitable iommu_domain
 * is found a new iommu_domain will be created.
 */
struct iommufd_ioas {
        struct iommufd_object obj;
        struct io_pagetable iopt;
        struct mutex mutex;
        struct list_head hwpt_list;
};

static inline struct iommufd_ioas *iommufd_get_ioas(struct iommufd_ctx *ictx,
                                                    u32 id)
{
        return container_of(iommufd_get_object(ictx, id, IOMMUFD_OBJ_IOAS),
                            struct iommufd_ioas, obj);
}

struct iommufd_ioas *iommufd_ioas_alloc(struct iommufd_ctx *ictx);
int iommufd_ioas_alloc_ioctl(struct iommufd_ucmd *ucmd);
void iommufd_ioas_destroy(struct iommufd_object *obj);
int iommufd_ioas_iova_ranges(struct iommufd_ucmd *ucmd);
int iommufd_ioas_allow_iovas(struct iommufd_ucmd *ucmd);
int iommufd_ioas_map(struct iommufd_ucmd *ucmd);
int iommufd_ioas_map_file(struct iommufd_ucmd *ucmd);
int iommufd_ioas_change_process(struct iommufd_ucmd *ucmd);
int iommufd_ioas_copy(struct iommufd_ucmd *ucmd);
int iommufd_ioas_unmap(struct iommufd_ucmd *ucmd);
int iommufd_ioas_option(struct iommufd_ucmd *ucmd);
int iommufd_option_rlimit_mode(struct iommu_option *cmd,
                               struct iommufd_ctx *ictx);

int iommufd_vfio_ioas(struct iommufd_ucmd *ucmd);
int iommufd_check_iova_range(struct io_pagetable *iopt,
                             struct iommu_hwpt_get_dirty_bitmap *bitmap);

/*
 * A HW pagetable is called an iommu_domain inside the kernel. This user object
 * allows directly creating and inspecting the domains. Domains that have kernel
 * owned page tables will be associated with an iommufd_ioas that provides the
 * IOVA to PFN map.
 */
struct iommufd_hw_pagetable {
        struct iommufd_object obj;
        struct iommu_domain *domain;
        struct iommufd_fault *fault;
        bool pasid_compat : 1;
};

struct iommufd_hwpt_paging {
        struct iommufd_hw_pagetable common;
        struct iommufd_ioas *ioas;
        bool auto_domain : 1;
        bool enforce_cache_coherency : 1;
        bool nest_parent : 1;
        /* Head at iommufd_ioas::hwpt_list */
        struct list_head hwpt_item;
        struct iommufd_sw_msi_maps present_sw_msi;
};

struct iommufd_hwpt_nested {
        struct iommufd_hw_pagetable common;
        struct iommufd_hwpt_paging *parent;
        struct iommufd_viommu *viommu;
};

static inline bool hwpt_is_paging(struct iommufd_hw_pagetable *hwpt)
{
        return hwpt->obj.type == IOMMUFD_OBJ_HWPT_PAGING;
}

static inline struct iommufd_hwpt_paging *
to_hwpt_paging(struct iommufd_hw_pagetable *hwpt)
{
        return container_of(hwpt, struct iommufd_hwpt_paging, common);
}

static inline struct iommufd_hwpt_nested *
to_hwpt_nested(struct iommufd_hw_pagetable *hwpt)
{
        return container_of(hwpt, struct iommufd_hwpt_nested, common);
}

static inline struct iommufd_hwpt_paging *
find_hwpt_paging(struct iommufd_hw_pagetable *hwpt)
{
        switch (hwpt->obj.type) {
        case IOMMUFD_OBJ_HWPT_PAGING:
                return to_hwpt_paging(hwpt);
        case IOMMUFD_OBJ_HWPT_NESTED:
                return to_hwpt_nested(hwpt)->parent;
        default:
                return NULL;
        }
}

static inline struct iommufd_hwpt_paging *
iommufd_get_hwpt_paging(struct iommufd_ucmd *ucmd, u32 id)
{
        return container_of(iommufd_get_object(ucmd->ictx, id,
                                               IOMMUFD_OBJ_HWPT_PAGING),
                            struct iommufd_hwpt_paging, common.obj);
}

static inline struct iommufd_hw_pagetable *
iommufd_get_hwpt_nested(struct iommufd_ucmd *ucmd, u32 id)
{
        return container_of(iommufd_get_object(ucmd->ictx, id,
                                               IOMMUFD_OBJ_HWPT_NESTED),
                            struct iommufd_hw_pagetable, obj);
}

int iommufd_hwpt_set_dirty_tracking(struct iommufd_ucmd *ucmd);
int iommufd_hwpt_get_dirty_bitmap(struct iommufd_ucmd *ucmd);

struct iommufd_hwpt_paging *
iommufd_hwpt_paging_alloc(struct iommufd_ctx *ictx, struct iommufd_ioas *ioas,
                          struct iommufd_device *idev, ioasid_t pasid,
                          u32 flags, bool immediate_attach,
                          const struct iommu_user_data *user_data);
int iommufd_hw_pagetable_attach(struct iommufd_hw_pagetable *hwpt,
                                struct iommufd_device *idev, ioasid_t pasid);
struct iommufd_hw_pagetable *
iommufd_hw_pagetable_detach(struct iommufd_device *idev, ioasid_t pasid);
void iommufd_hwpt_paging_destroy(struct iommufd_object *obj);
void iommufd_hwpt_paging_abort(struct iommufd_object *obj);
void iommufd_hwpt_nested_destroy(struct iommufd_object *obj);
void iommufd_hwpt_nested_abort(struct iommufd_object *obj);
int iommufd_hwpt_alloc(struct iommufd_ucmd *ucmd);
int iommufd_hwpt_invalidate(struct iommufd_ucmd *ucmd);

static inline void iommufd_hw_pagetable_put(struct iommufd_ctx *ictx,
                                            struct iommufd_hw_pagetable *hwpt)
{
        if (hwpt->obj.type == IOMMUFD_OBJ_HWPT_PAGING) {
                struct iommufd_hwpt_paging *hwpt_paging = to_hwpt_paging(hwpt);

                if (hwpt_paging->auto_domain) {
                        lockdep_assert_not_held(&hwpt_paging->ioas->mutex);
                        iommufd_object_put_and_try_destroy(ictx, &hwpt->obj);
                        return;
                }
        }
        refcount_dec(&hwpt->obj.users);
}

struct iommufd_attach;

struct iommufd_group {
        struct kref ref;
        struct mutex lock;
        struct iommufd_ctx *ictx;
        struct iommu_group *group;
        struct xarray pasid_attach;
        struct iommufd_sw_msi_maps required_sw_msi;
        phys_addr_t sw_msi_start;
};

/*
 * A iommufd_device object represents the binding relationship between a
 * consuming driver and the iommufd. These objects are created/destroyed by
 * external drivers, not by userspace.
 */
struct iommufd_device {
        struct iommufd_object obj;
        struct iommufd_ctx *ictx;
        struct iommufd_group *igroup;
        struct list_head group_item;
        /* always the physical device */
        struct device *dev;
        bool enforce_cache_coherency;
        struct iommufd_vdevice *vdev;
        bool destroying;
};

static inline struct iommufd_device *
iommufd_get_device(struct iommufd_ucmd *ucmd, u32 id)
{
        return container_of(iommufd_get_object(ucmd->ictx, id,
                                               IOMMUFD_OBJ_DEVICE),
                            struct iommufd_device, obj);
}

void iommufd_device_pre_destroy(struct iommufd_object *obj);
void iommufd_device_destroy(struct iommufd_object *obj);
int iommufd_get_hw_info(struct iommufd_ucmd *ucmd);

struct device *iommufd_global_device(void);

struct iommufd_access {
        struct iommufd_object obj;
        struct iommufd_ctx *ictx;
        struct iommufd_ioas *ioas;
        struct iommufd_ioas *ioas_unpin;
        struct mutex ioas_lock;
        const struct iommufd_access_ops *ops;
        void *data;
        unsigned long iova_alignment;
        u32 iopt_access_list_id;
};

int iopt_add_access(struct io_pagetable *iopt, struct iommufd_access *access);
void iopt_remove_access(struct io_pagetable *iopt,
                        struct iommufd_access *access, u32 iopt_access_list_id);
void iommufd_access_destroy_object(struct iommufd_object *obj);

/* iommufd_access for internal use */
static inline bool iommufd_access_is_internal(struct iommufd_access *access)
{
        return !access->ictx;
}

struct iommufd_access *iommufd_access_create_internal(struct iommufd_ctx *ictx);

static inline void
iommufd_access_destroy_internal(struct iommufd_ctx *ictx,
                                struct iommufd_access *access)
{
        iommufd_object_destroy_user(ictx, &access->obj);
}

int iommufd_access_attach_internal(struct iommufd_access *access,
                                   struct iommufd_ioas *ioas);

static inline void iommufd_access_detach_internal(struct iommufd_access *access)
{
        iommufd_access_detach(access);
}

struct iommufd_eventq {
        struct iommufd_object obj;
        struct iommufd_ctx *ictx;
        struct file *filep;

        spinlock_t lock; /* protects the deliver list */
        struct list_head deliver;

        struct wait_queue_head wait_queue;
};

struct iommufd_attach_handle {
        struct iommu_attach_handle handle;
        struct iommufd_device *idev;
};

/* Convert an iommu attach handle to iommufd handle. */
#define to_iommufd_handle(hdl)  container_of(hdl, struct iommufd_attach_handle, handle)

/*
 * An iommufd_fault object represents an interface to deliver I/O page faults
 * to the user space. These objects are created/destroyed by the user space and
 * associated with hardware page table objects during page-table allocation.
 */
struct iommufd_fault {
        struct iommufd_eventq common;
        struct mutex mutex; /* serializes response flows */
        struct xarray response;
};

static inline struct iommufd_fault *
eventq_to_fault(struct iommufd_eventq *eventq)
{
        return container_of(eventq, struct iommufd_fault, common);
}

static inline struct iommufd_fault *
iommufd_get_fault(struct iommufd_ucmd *ucmd, u32 id)
{
        return container_of(iommufd_get_object(ucmd->ictx, id,
                                               IOMMUFD_OBJ_FAULT),
                            struct iommufd_fault, common.obj);
}

int iommufd_fault_alloc(struct iommufd_ucmd *ucmd);
void iommufd_fault_destroy(struct iommufd_object *obj);
int iommufd_fault_iopf_handler(struct iopf_group *group);
void iommufd_auto_response_faults(struct iommufd_hw_pagetable *hwpt,
                                  struct iommufd_attach_handle *handle);

/* An iommufd_vevent represents a vIOMMU event in an iommufd_veventq */
struct iommufd_vevent {
        struct iommufd_vevent_header header;
        struct list_head node; /* for iommufd_eventq::deliver */
        ssize_t data_len;
        u64 event_data[] __counted_by(data_len);
};

#define vevent_for_lost_events_header(vevent) \
        (vevent->header.flags & IOMMU_VEVENTQ_FLAG_LOST_EVENTS)

/*
 * An iommufd_veventq object represents an interface to deliver vIOMMU events to
 * the user space. It is created/destroyed by the user space and associated with
 * a vIOMMU object during the allocations.
 */
struct iommufd_veventq {
        struct iommufd_eventq common;
        struct iommufd_viommu *viommu;
        struct list_head node; /* for iommufd_viommu::veventqs */

        enum iommu_veventq_type type;
        unsigned int depth;

        /* Use common.lock for protection */
        u32 num_events;
        u32 sequence;

        /* Must be last as it ends in a flexible-array member. */
        struct iommufd_vevent lost_events_header;
};

static inline struct iommufd_veventq *
eventq_to_veventq(struct iommufd_eventq *eventq)
{
        return container_of(eventq, struct iommufd_veventq, common);
}

static inline struct iommufd_veventq *
iommufd_get_veventq(struct iommufd_ucmd *ucmd, u32 id)
{
        return container_of(iommufd_get_object(ucmd->ictx, id,
                                               IOMMUFD_OBJ_VEVENTQ),
                            struct iommufd_veventq, common.obj);
}

int iommufd_veventq_alloc(struct iommufd_ucmd *ucmd);
void iommufd_veventq_destroy(struct iommufd_object *obj);
void iommufd_veventq_abort(struct iommufd_object *obj);

static inline void iommufd_vevent_handler(struct iommufd_veventq *veventq,
                                          struct iommufd_vevent *vevent)
{
        struct iommufd_eventq *eventq = &veventq->common;

        lockdep_assert_held(&eventq->lock);

        /*
         * Remove the lost_events_header and add the new node at the same time.
         * Note the new node can be lost_events_header, for a sequence update.
         */
        if (list_is_last(&veventq->lost_events_header.node, &eventq->deliver))
                list_del(&veventq->lost_events_header.node);
        list_add_tail(&vevent->node, &eventq->deliver);
        vevent->header.sequence = veventq->sequence;
        veventq->sequence = (veventq->sequence + 1) & INT_MAX;

        wake_up_interruptible(&eventq->wait_queue);
}

static inline struct iommufd_viommu *
iommufd_get_viommu(struct iommufd_ucmd *ucmd, u32 id)
{
        return container_of(iommufd_get_object(ucmd->ictx, id,
                                               IOMMUFD_OBJ_VIOMMU),
                            struct iommufd_viommu, obj);
}

static inline struct iommufd_veventq *
iommufd_viommu_find_veventq(struct iommufd_viommu *viommu,
                            enum iommu_veventq_type type)
{
        struct iommufd_veventq *veventq, *next;

        lockdep_assert_held(&viommu->veventqs_rwsem);

        list_for_each_entry_safe(veventq, next, &viommu->veventqs, node) {
                if (veventq->type == type)
                        return veventq;
        }
        return NULL;
}

int iommufd_viommu_alloc_ioctl(struct iommufd_ucmd *ucmd);
void iommufd_viommu_destroy(struct iommufd_object *obj);
int iommufd_vdevice_alloc_ioctl(struct iommufd_ucmd *ucmd);
void iommufd_vdevice_destroy(struct iommufd_object *obj);
void iommufd_vdevice_abort(struct iommufd_object *obj);
int iommufd_hw_queue_alloc_ioctl(struct iommufd_ucmd *ucmd);
void iommufd_hw_queue_destroy(struct iommufd_object *obj);

static inline struct iommufd_vdevice *
iommufd_get_vdevice(struct iommufd_ctx *ictx, u32 id)
{
        return container_of(iommufd_get_object(ictx, id,
                                               IOMMUFD_OBJ_VDEVICE),
                            struct iommufd_vdevice, obj);
}

#ifdef CONFIG_IOMMUFD_TEST
int iommufd_test(struct iommufd_ucmd *ucmd);
void iommufd_selftest_destroy(struct iommufd_object *obj);
extern size_t iommufd_test_memory_limit;
void iommufd_test_syz_conv_iova_id(struct iommufd_ucmd *ucmd,
                                   unsigned int ioas_id, u64 *iova, u32 *flags);
bool iommufd_should_fail(void);
int __init iommufd_test_init(void);
void iommufd_test_exit(void);
bool iommufd_selftest_is_mock_dev(struct device *dev);
int iommufd_test_dma_buf_iommufd_map(struct dma_buf_attachment *attachment,
                                     struct phys_vec *phys);
#else
static inline void iommufd_test_syz_conv_iova_id(struct iommufd_ucmd *ucmd,
                                                 unsigned int ioas_id,
                                                 u64 *iova, u32 *flags)
{
}
static inline bool iommufd_should_fail(void)
{
        return false;
}
static inline int __init iommufd_test_init(void)
{
        return 0;
}
static inline void iommufd_test_exit(void)
{
}
static inline bool iommufd_selftest_is_mock_dev(struct device *dev)
{
        return false;
}
static inline int
iommufd_test_dma_buf_iommufd_map(struct dma_buf_attachment *attachment,
                                 struct phys_vec *phys)
{
        return -EOPNOTSUPP;
}
#endif
#endif