root/drivers/gpu/drm/imagination/pvr_drv.c
// SPDX-License-Identifier: GPL-2.0-only OR MIT
/* Copyright (c) 2023 Imagination Technologies Ltd. */

#include "pvr_context.h"
#include "pvr_debugfs.h"
#include "pvr_device.h"
#include "pvr_drv.h"
#include "pvr_free_list.h"
#include "pvr_gem.h"
#include "pvr_hwrt.h"
#include "pvr_job.h"
#include "pvr_mmu.h"
#include "pvr_power.h"
#include "pvr_rogue_defs.h"
#include "pvr_rogue_fwif_client.h"
#include "pvr_rogue_fwif_shared.h"
#include "pvr_vm.h"

#include <uapi/drm/pvr_drm.h>

#include <drm/drm_device.h>
#include <drm/drm_drv.h>
#include <drm/drm_file.h>
#include <drm/drm_gem.h>
#include <drm/drm_ioctl.h>

#include <linux/err.h>
#include <linux/export.h>
#include <linux/fs.h>
#include <linux/kernel.h>
#include <linux/list.h>
#include <linux/mod_devicetable.h>
#include <linux/module.h>
#include <linux/moduleparam.h>
#include <linux/of_device.h>
#include <linux/of_platform.h>
#include <linux/platform_device.h>
#include <linux/pm_runtime.h>
#include <linux/xarray.h>

/**
 * DOC: PowerVR (Series 6 and later) and IMG Graphics Driver
 *
 * This driver supports the following PowerVR/IMG graphics cores from Imagination Technologies:
 *
 * * AXE-1-16M (found in Texas Instruments AM62)
 * * BXS-4-64 MC1 (found in Texas Instruments J721S2/AM68)
 */

/**
 * pvr_ioctl_create_bo() - IOCTL to create a GEM buffer object.
 * @drm_dev: [IN] Target DRM device.
 * @raw_args: [IN/OUT] Arguments passed to this IOCTL. This must be of type
 * &struct drm_pvr_ioctl_create_bo_args.
 * @file: [IN] DRM file-private data.
 *
 * Called from userspace with %DRM_IOCTL_PVR_CREATE_BO.
 *
 * Return:
 *  * 0 on success,
 *  * -%EINVAL if the value of &drm_pvr_ioctl_create_bo_args.size is zero
 *    or wider than &typedef size_t,
 *  * -%EINVAL if any bits in &drm_pvr_ioctl_create_bo_args.flags that are
 *    reserved or undefined are set,
 *  * -%EINVAL if any padding fields in &drm_pvr_ioctl_create_bo_args are not
 *    zero,
 *  * Any error encountered while creating the object (see
 *    pvr_gem_object_create()), or
 *  * Any error encountered while transferring ownership of the object into a
 *    userspace-accessible handle (see pvr_gem_object_into_handle()).
 */
static int
pvr_ioctl_create_bo(struct drm_device *drm_dev, void *raw_args,
                    struct drm_file *file)
{
        struct drm_pvr_ioctl_create_bo_args *args = raw_args;
        struct pvr_device *pvr_dev = to_pvr_device(drm_dev);
        struct pvr_file *pvr_file = to_pvr_file(file);

        struct pvr_gem_object *pvr_obj;
        size_t sanitized_size;

        int idx;
        int err;

        if (!drm_dev_enter(drm_dev, &idx))
                return -EIO;

        /* All padding fields must be zeroed. */
        if (args->_padding_c != 0) {
                err = -EINVAL;
                goto err_drm_dev_exit;
        }

        /*
         * On 64-bit platforms (our primary target), size_t is a u64. However,
         * on other architectures we have to check for overflow when casting
         * down to size_t from u64.
         *
         * We also disallow zero-sized allocations, and reserved (kernel-only)
         * flags.
         */
        if (args->size > SIZE_MAX || args->size == 0 || args->flags &
            ~DRM_PVR_BO_FLAGS_MASK || args->size & (PVR_DEVICE_PAGE_SIZE - 1)) {
                err = -EINVAL;
                goto err_drm_dev_exit;
        }

        sanitized_size = (size_t)args->size;

        /*
         * Create a buffer object and transfer ownership to a userspace-
         * accessible handle.
         */
        pvr_obj = pvr_gem_object_create(pvr_dev, sanitized_size, args->flags);
        if (IS_ERR(pvr_obj)) {
                err = PTR_ERR(pvr_obj);
                goto err_drm_dev_exit;
        }

        /* This function will not modify &args->handle unless it succeeds. */
        err = pvr_gem_object_into_handle(pvr_obj, pvr_file, &args->handle);
        if (err)
                goto err_destroy_obj;

        drm_dev_exit(idx);

        return 0;

err_destroy_obj:
        /*
         * GEM objects are refcounted, so there is no explicit destructor
         * function. Instead, we release the singular reference we currently
         * hold on the object and let GEM take care of the rest.
         */
        pvr_gem_object_put(pvr_obj);

err_drm_dev_exit:
        drm_dev_exit(idx);

        return err;
}

/**
 * pvr_ioctl_get_bo_mmap_offset() - IOCTL to generate a "fake" offset to be
 * used when calling mmap() from userspace to map the given GEM buffer object
 * @drm_dev: [IN] DRM device (unused).
 * @raw_args: [IN/OUT] Arguments passed to this IOCTL. This must be of type
 *                     &struct drm_pvr_ioctl_get_bo_mmap_offset_args.
 * @file: [IN] DRM file private data.
 *
 * Called from userspace with %DRM_IOCTL_PVR_GET_BO_MMAP_OFFSET.
 *
 * This IOCTL does *not* perform an mmap. See the docs on
 * &struct drm_pvr_ioctl_get_bo_mmap_offset_args for details.
 *
 * Return:
 *  * 0 on success,
 *  * -%ENOENT if the handle does not reference a valid GEM buffer object,
 *  * -%EINVAL if any padding fields in &struct
 *    drm_pvr_ioctl_get_bo_mmap_offset_args are not zero, or
 *  * Any error returned by drm_gem_create_mmap_offset().
 */
static int
pvr_ioctl_get_bo_mmap_offset(struct drm_device *drm_dev, void *raw_args,
                             struct drm_file *file)
{
        struct drm_pvr_ioctl_get_bo_mmap_offset_args *args = raw_args;
        struct pvr_file *pvr_file = to_pvr_file(file);
        struct pvr_gem_object *pvr_obj;
        struct drm_gem_object *gem_obj;
        int idx;
        int ret;

        if (!drm_dev_enter(drm_dev, &idx))
                return -EIO;

        /* All padding fields must be zeroed. */
        if (args->_padding_4 != 0) {
                ret = -EINVAL;
                goto err_drm_dev_exit;
        }

        /*
         * Obtain a kernel reference to the buffer object. This reference is
         * counted and must be manually dropped before returning. If a buffer
         * object cannot be found for the specified handle, return -%ENOENT (No
         * such file or directory).
         */
        pvr_obj = pvr_gem_object_from_handle(pvr_file, args->handle);
        if (!pvr_obj) {
                ret = -ENOENT;
                goto err_drm_dev_exit;
        }

        gem_obj = gem_from_pvr_gem(pvr_obj);

        /*
         * Allocate a fake offset which can be used in userspace calls to mmap
         * on the DRM device file. If this fails, return the error code. This
         * operation is idempotent.
         */
        ret = drm_gem_create_mmap_offset(gem_obj);
        if (ret != 0) {
                /* Drop our reference to the buffer object. */
                drm_gem_object_put(gem_obj);
                goto err_drm_dev_exit;
        }

        /*
         * Read out the fake offset allocated by the earlier call to
         * drm_gem_create_mmap_offset.
         */
        args->offset = drm_vma_node_offset_addr(&gem_obj->vma_node);

        /* Drop our reference to the buffer object. */
        pvr_gem_object_put(pvr_obj);

err_drm_dev_exit:
        drm_dev_exit(idx);

        return ret;
}

static __always_inline __maybe_unused u64
pvr_fw_version_packed(u32 major, u32 minor)
{
        return ((u64)major << 32) | minor;
}

static u32
rogue_get_common_store_partition_space_size(struct pvr_device *pvr_dev)
{
        u32 max_partitions = 0;
        u32 tile_size_x = 0;
        u32 tile_size_y = 0;

        PVR_FEATURE_VALUE(pvr_dev, tile_size_x, &tile_size_x);
        PVR_FEATURE_VALUE(pvr_dev, tile_size_y, &tile_size_y);
        PVR_FEATURE_VALUE(pvr_dev, max_partitions, &max_partitions);

        if (tile_size_x == 16 && tile_size_y == 16) {
                u32 usc_min_output_registers_per_pix = 0;

                PVR_FEATURE_VALUE(pvr_dev, usc_min_output_registers_per_pix,
                                  &usc_min_output_registers_per_pix);

                return tile_size_x * tile_size_y * max_partitions *
                       usc_min_output_registers_per_pix;
        }

        return max_partitions * 1024;
}

static u32
rogue_get_common_store_alloc_region_size(struct pvr_device *pvr_dev)
{
        u32 common_store_size_in_dwords = 512 * 4 * 4;
        u32 alloc_region_size;

        PVR_FEATURE_VALUE(pvr_dev, common_store_size_in_dwords, &common_store_size_in_dwords);

        alloc_region_size = common_store_size_in_dwords - (256U * 4U) -
                            rogue_get_common_store_partition_space_size(pvr_dev);

        if (PVR_HAS_QUIRK(pvr_dev, 44079)) {
                u32 common_store_split_point = (768U * 4U * 4U);

                return min(common_store_split_point - (256U * 4U), alloc_region_size);
        }

        return alloc_region_size;
}

static inline u32
rogue_get_num_phantoms(struct pvr_device *pvr_dev)
{
        u32 num_clusters = 1;

        PVR_FEATURE_VALUE(pvr_dev, num_clusters, &num_clusters);

        return ROGUE_REQ_NUM_PHANTOMS(num_clusters);
}

static inline u32
rogue_get_max_coeffs(struct pvr_device *pvr_dev)
{
        u32 max_coeff_additional_portion = ROGUE_MAX_VERTEX_SHARED_REGISTERS;
        u32 pending_allocation_shared_regs = 2U * 1024U;
        u32 pending_allocation_coeff_regs = 0U;
        u32 num_phantoms = rogue_get_num_phantoms(pvr_dev);
        u32 tiles_in_flight = 0;
        u32 max_coeff_pixel_portion;

        PVR_FEATURE_VALUE(pvr_dev, isp_max_tiles_in_flight, &tiles_in_flight);
        max_coeff_pixel_portion = DIV_ROUND_UP(tiles_in_flight, num_phantoms);
        max_coeff_pixel_portion *= ROGUE_MAX_PIXEL_SHARED_REGISTERS;

        /*
         * Compute tasks on cores with BRN48492 and without compute overlap may lock
         * up without two additional lines of coeffs.
         */
        if (PVR_HAS_QUIRK(pvr_dev, 48492) && !PVR_HAS_FEATURE(pvr_dev, compute_overlap))
                pending_allocation_coeff_regs = 2U * 1024U;

        if (PVR_HAS_ENHANCEMENT(pvr_dev, 38748))
                pending_allocation_shared_regs = 0;

        if (PVR_HAS_ENHANCEMENT(pvr_dev, 38020))
                max_coeff_additional_portion += ROGUE_MAX_COMPUTE_SHARED_REGISTERS;

        return rogue_get_common_store_alloc_region_size(pvr_dev) + pending_allocation_coeff_regs -
                (max_coeff_pixel_portion + max_coeff_additional_portion +
                 pending_allocation_shared_regs);
}

static inline u32
rogue_get_cdm_max_local_mem_size_regs(struct pvr_device *pvr_dev)
{
        u32 available_coeffs_in_dwords = rogue_get_max_coeffs(pvr_dev);

        if (PVR_HAS_QUIRK(pvr_dev, 48492) && PVR_HAS_FEATURE(pvr_dev, roguexe) &&
            !PVR_HAS_FEATURE(pvr_dev, compute_overlap)) {
                /* Driver must not use the 2 reserved lines. */
                available_coeffs_in_dwords -= ROGUE_CSRM_LINE_SIZE_IN_DWORDS * 2;
        }

        /*
         * The maximum amount of local memory available to a kernel is the minimum
         * of the total number of coefficient registers available and the max common
         * store allocation size which can be made by the CDM.
         *
         * If any coeff lines are reserved for tessellation or pixel then we need to
         * subtract those too.
         */
        return min(available_coeffs_in_dwords, (u32)ROGUE_MAX_PER_KERNEL_LOCAL_MEM_SIZE_REGS);
}

/**
 * pvr_dev_query_gpu_info_get()
 * @pvr_dev: Device pointer.
 * @args: [IN] Device query arguments containing a pointer to a userspace
 *        struct drm_pvr_dev_query_gpu_info.
 *
 * If the query object pointer is NULL, the size field is updated with the
 * expected size of the query object.
 *
 * Returns:
 *  * 0 on success, or if size is requested using a NULL pointer, or
 *  * -%E2BIG if the indicated length of the allocation is less than is
 *    required to contain the copied data, or
 *  * -%EFAULT if local memory could not be copied to userspace.
 */
static int
pvr_dev_query_gpu_info_get(struct pvr_device *pvr_dev,
                           struct drm_pvr_ioctl_dev_query_args *args)
{
        struct drm_pvr_dev_query_gpu_info gpu_info = {0};
        int err;

        if (!args->pointer) {
                args->size = sizeof(struct drm_pvr_dev_query_gpu_info);
                return 0;
        }

        gpu_info.gpu_id =
                pvr_gpu_id_to_packed_bvnc(&pvr_dev->gpu_id);
        gpu_info.num_phantoms = rogue_get_num_phantoms(pvr_dev);

        err = PVR_UOBJ_SET(args->pointer, args->size, gpu_info);
        if (err < 0)
                return err;

        if (args->size > sizeof(gpu_info))
                args->size = sizeof(gpu_info);
        return 0;
}

/**
 * pvr_dev_query_runtime_info_get()
 * @pvr_dev: Device pointer.
 * @args: [IN] Device query arguments containing a pointer to a userspace
 *        struct drm_pvr_dev_query_runtime_info.
 *
 * If the query object pointer is NULL, the size field is updated with the
 * expected size of the query object.
 *
 * Returns:
 *  * 0 on success, or if size is requested using a NULL pointer, or
 *  * -%E2BIG if the indicated length of the allocation is less than is
 *    required to contain the copied data, or
 *  * -%EFAULT if local memory could not be copied to userspace.
 */
static int
pvr_dev_query_runtime_info_get(struct pvr_device *pvr_dev,
                               struct drm_pvr_ioctl_dev_query_args *args)
{
        struct drm_pvr_dev_query_runtime_info runtime_info = {0};
        int err;

        if (!args->pointer) {
                args->size = sizeof(struct drm_pvr_dev_query_runtime_info);
                return 0;
        }

        runtime_info.free_list_min_pages =
                pvr_get_free_list_min_pages(pvr_dev);
        runtime_info.free_list_max_pages =
                ROGUE_PM_MAX_FREELIST_SIZE / ROGUE_PM_PAGE_SIZE;
        runtime_info.common_store_alloc_region_size =
                rogue_get_common_store_alloc_region_size(pvr_dev);
        runtime_info.common_store_partition_space_size =
                rogue_get_common_store_partition_space_size(pvr_dev);
        runtime_info.max_coeffs = rogue_get_max_coeffs(pvr_dev);
        runtime_info.cdm_max_local_mem_size_regs =
                rogue_get_cdm_max_local_mem_size_regs(pvr_dev);

        err = PVR_UOBJ_SET(args->pointer, args->size, runtime_info);
        if (err < 0)
                return err;

        if (args->size > sizeof(runtime_info))
                args->size = sizeof(runtime_info);
        return 0;
}

/**
 * pvr_dev_query_quirks_get() - Unpack array of quirks at the address given
 * in a struct drm_pvr_dev_query_quirks, or gets the amount of space required
 * for it.
 * @pvr_dev: Device pointer.
 * @args: [IN] Device query arguments containing a pointer to a userspace
 *        struct drm_pvr_dev_query_query_quirks.
 *
 * If the query object pointer is NULL, the size field is updated with the
 * expected size of the query object.
 * If the userspace pointer in the query object is NULL, or the count is
 * short, no data is copied.
 * The count field will be updated to that copied, or if either pointer is
 * NULL, that which would have been copied.
 * The size field in the query object will be updated to the size copied.
 *
 * Returns:
 *  * 0 on success, or if size/count is requested using a NULL pointer, or
 *  * -%EINVAL if args contained non-zero reserved fields, or
 *  * -%E2BIG if the indicated length of the allocation is less than is
 *    required to contain the copied data, or
 *  * -%EFAULT if local memory could not be copied to userspace.
 */
static int
pvr_dev_query_quirks_get(struct pvr_device *pvr_dev,
                         struct drm_pvr_ioctl_dev_query_args *args)
{
        /*
         * @FIXME - hardcoding of numbers here is intended as an
         * intermediate step so the UAPI can be fixed, but requires a
         * a refactor in the future to store them in a more appropriate
         * location
         */
        static const u32 umd_quirks_musthave[] = {
                47217,
                49927,
                62269,
        };
        static const u32 umd_quirks[] = {
                48545,
                51764,
        };
        struct drm_pvr_dev_query_quirks query;
        u32 out[ARRAY_SIZE(umd_quirks_musthave) + ARRAY_SIZE(umd_quirks)];
        size_t out_musthave_count = 0;
        size_t out_count = 0;
        int err;

        if (!args->pointer) {
                args->size = sizeof(struct drm_pvr_dev_query_quirks);
                return 0;
        }

        err = PVR_UOBJ_GET(query, args->size, args->pointer);

        if (err < 0)
                return err;
        if (query._padding_c)
                return -EINVAL;

        for (int i = 0; i < ARRAY_SIZE(umd_quirks_musthave); i++) {
                if (pvr_device_has_uapi_quirk(pvr_dev, umd_quirks_musthave[i])) {
                        out[out_count++] = umd_quirks_musthave[i];
                        out_musthave_count++;
                }
        }

        for (int i = 0; i < ARRAY_SIZE(umd_quirks); i++) {
                if (pvr_device_has_uapi_quirk(pvr_dev, umd_quirks[i]))
                        out[out_count++] = umd_quirks[i];
        }

        if (!query.quirks)
                goto copy_out;
        if (query.count < out_count)
                return -E2BIG;

        if (copy_to_user(u64_to_user_ptr(query.quirks), out,
                         out_count * sizeof(u32))) {
                return -EFAULT;
        }

        query.musthave_count = out_musthave_count;

copy_out:
        query.count = out_count;
        err = PVR_UOBJ_SET(args->pointer, args->size, query);
        if (err < 0)
                return err;

        args->size = sizeof(query);
        return 0;
}

/**
 * pvr_dev_query_enhancements_get() - Unpack array of enhancements at the
 * address given in a struct drm_pvr_dev_query_enhancements, or gets the amount
 * of space required for it.
 * @pvr_dev: Device pointer.
 * @args: [IN] Device query arguments containing a pointer to a userspace
 *        struct drm_pvr_dev_query_enhancements.
 *
 * If the query object pointer is NULL, the size field is updated with the
 * expected size of the query object.
 * If the userspace pointer in the query object is NULL, or the count is
 * short, no data is copied.
 * The count field will be updated to that copied, or if either pointer is
 * NULL, that which would have been copied.
 * The size field in the query object will be updated to the size copied.
 *
 * Returns:
 *  * 0 on success, or if size/count is requested using a NULL pointer, or
 *  * -%EINVAL if args contained non-zero reserved fields, or
 *  * -%E2BIG if the indicated length of the allocation is less than is
 *    required to contain the copied data, or
 *  * -%EFAULT if local memory could not be copied to userspace.
 */
static int
pvr_dev_query_enhancements_get(struct pvr_device *pvr_dev,
                               struct drm_pvr_ioctl_dev_query_args *args)
{
        /*
         * @FIXME - hardcoding of numbers here is intended as an
         * intermediate step so the UAPI can be fixed, but requires a
         * a refactor in the future to store them in a more appropriate
         * location
         */
        const u32 umd_enhancements[] = {
                35421,
                42064,
        };
        struct drm_pvr_dev_query_enhancements query;
        u32 out[ARRAY_SIZE(umd_enhancements)];
        size_t out_idx = 0;
        int err;

        if (!args->pointer) {
                args->size = sizeof(struct drm_pvr_dev_query_enhancements);
                return 0;
        }

        err = PVR_UOBJ_GET(query, args->size, args->pointer);

        if (err < 0)
                return err;
        if (query._padding_a)
                return -EINVAL;
        if (query._padding_c)
                return -EINVAL;

        for (int i = 0; i < ARRAY_SIZE(umd_enhancements); i++) {
                if (pvr_device_has_uapi_enhancement(pvr_dev, umd_enhancements[i]))
                        out[out_idx++] = umd_enhancements[i];
        }

        if (!query.enhancements)
                goto copy_out;
        if (query.count < out_idx)
                return -E2BIG;

        if (copy_to_user(u64_to_user_ptr(query.enhancements), out,
                         out_idx * sizeof(u32))) {
                return -EFAULT;
        }

copy_out:
        query.count = out_idx;
        err = PVR_UOBJ_SET(args->pointer, args->size, query);
        if (err < 0)
                return err;

        args->size = sizeof(query);
        return 0;
}

/**
 * pvr_ioctl_dev_query() - IOCTL to copy information about a device
 * @drm_dev: [IN] DRM device.
 * @raw_args: [IN/OUT] Arguments passed to this IOCTL. This must be of type
 *                     &struct drm_pvr_ioctl_dev_query_args.
 * @file: [IN] DRM file private data.
 *
 * Called from userspace with %DRM_IOCTL_PVR_DEV_QUERY.
 * If the given receiving struct pointer is NULL, or the indicated size is too
 * small, the expected size of the struct type will be returned in the size
 * argument field.
 *
 * Return:
 *  * 0 on success or when fetching the size with args->pointer == NULL, or
 *  * -%E2BIG if the indicated size of the receiving struct is less than is
 *    required to contain the copied data, or
 *  * -%EINVAL if the indicated struct type is unknown, or
 *  * -%ENOMEM if local memory could not be allocated, or
 *  * -%EFAULT if local memory could not be copied to userspace.
 */
static int
pvr_ioctl_dev_query(struct drm_device *drm_dev, void *raw_args,
                    struct drm_file *file)
{
        struct pvr_device *pvr_dev = to_pvr_device(drm_dev);
        struct drm_pvr_ioctl_dev_query_args *args = raw_args;
        int idx;
        int ret = -EINVAL;

        if (!drm_dev_enter(drm_dev, &idx))
                return -EIO;

        switch ((enum drm_pvr_dev_query)args->type) {
        case DRM_PVR_DEV_QUERY_GPU_INFO_GET:
                ret = pvr_dev_query_gpu_info_get(pvr_dev, args);
                break;

        case DRM_PVR_DEV_QUERY_RUNTIME_INFO_GET:
                ret = pvr_dev_query_runtime_info_get(pvr_dev, args);
                break;

        case DRM_PVR_DEV_QUERY_QUIRKS_GET:
                ret = pvr_dev_query_quirks_get(pvr_dev, args);
                break;

        case DRM_PVR_DEV_QUERY_ENHANCEMENTS_GET:
                ret = pvr_dev_query_enhancements_get(pvr_dev, args);
                break;

        case DRM_PVR_DEV_QUERY_HEAP_INFO_GET:
                ret = pvr_heap_info_get(pvr_dev, args);
                break;

        case DRM_PVR_DEV_QUERY_STATIC_DATA_AREAS_GET:
                ret = pvr_static_data_areas_get(pvr_dev, args);
                break;
        }

        drm_dev_exit(idx);

        return ret;
}

/**
 * pvr_ioctl_create_context() - IOCTL to create a context
 * @drm_dev: [IN] DRM device.
 * @raw_args: [IN/OUT] Arguments passed to this IOCTL. This must be of type
 *                     &struct drm_pvr_ioctl_create_context_args.
 * @file: [IN] DRM file private data.
 *
 * Called from userspace with %DRM_IOCTL_PVR_CREATE_CONTEXT.
 *
 * Return:
 *  * 0 on success, or
 *  * -%EINVAL if provided arguments are invalid, or
 *  * -%EFAULT if arguments can't be copied from userspace, or
 *  * Any error returned by pvr_create_render_context().
 */
static int
pvr_ioctl_create_context(struct drm_device *drm_dev, void *raw_args,
                         struct drm_file *file)
{
        struct drm_pvr_ioctl_create_context_args *args = raw_args;
        struct pvr_file *pvr_file = file->driver_priv;
        int idx;
        int ret;

        if (!drm_dev_enter(drm_dev, &idx))
                return -EIO;

        ret = pvr_context_create(pvr_file, args);

        drm_dev_exit(idx);

        return ret;
}

/**
 * pvr_ioctl_destroy_context() - IOCTL to destroy a context
 * @drm_dev: [IN] DRM device.
 * @raw_args: [IN/OUT] Arguments passed to this IOCTL. This must be of type
 *                     &struct drm_pvr_ioctl_destroy_context_args.
 * @file: [IN] DRM file private data.
 *
 * Called from userspace with %DRM_IOCTL_PVR_DESTROY_CONTEXT.
 *
 * Return:
 *  * 0 on success, or
 *  * -%EINVAL if context not in context list.
 */
static int
pvr_ioctl_destroy_context(struct drm_device *drm_dev, void *raw_args,
                          struct drm_file *file)
{
        struct drm_pvr_ioctl_destroy_context_args *args = raw_args;
        struct pvr_file *pvr_file = file->driver_priv;

        if (args->_padding_4)
                return -EINVAL;

        return pvr_context_destroy(pvr_file, args->handle);
}

/**
 * pvr_ioctl_create_free_list() - IOCTL to create a free list
 * @drm_dev: [IN] DRM device.
 * @raw_args: [IN/OUT] Arguments passed to this IOCTL. This must be of type
 *                     &struct drm_pvr_ioctl_create_free_list_args.
 * @file: [IN] DRM file private data.
 *
 * Called from userspace with %DRM_IOCTL_PVR_CREATE_FREE_LIST.
 *
 * Return:
 *  * 0 on success, or
 *  * Any error returned by pvr_free_list_create().
 */
static int
pvr_ioctl_create_free_list(struct drm_device *drm_dev, void *raw_args,
                           struct drm_file *file)
{
        struct drm_pvr_ioctl_create_free_list_args *args = raw_args;
        struct pvr_file *pvr_file = to_pvr_file(file);
        struct pvr_free_list *free_list;
        int idx;
        int err;

        if (!drm_dev_enter(drm_dev, &idx))
                return -EIO;

        free_list = pvr_free_list_create(pvr_file, args);
        if (IS_ERR(free_list)) {
                err = PTR_ERR(free_list);
                goto err_drm_dev_exit;
        }

        /* Allocate object handle for userspace. */
        err = xa_alloc(&pvr_file->free_list_handles,
                       &args->handle,
                       free_list,
                       xa_limit_32b,
                       GFP_KERNEL);
        if (err < 0)
                goto err_cleanup;

        drm_dev_exit(idx);

        return 0;

err_cleanup:
        pvr_free_list_put(free_list);

err_drm_dev_exit:
        drm_dev_exit(idx);

        return err;
}

/**
 * pvr_ioctl_destroy_free_list() - IOCTL to destroy a free list
 * @drm_dev: [IN] DRM device.
 * @raw_args: [IN] Arguments passed to this IOCTL. This must be of type
 *                 &struct drm_pvr_ioctl_destroy_free_list_args.
 * @file: [IN] DRM file private data.
 *
 * Called from userspace with %DRM_IOCTL_PVR_DESTROY_FREE_LIST.
 *
 * Return:
 *  * 0 on success, or
 *  * -%EINVAL if free list not in object list.
 */
static int
pvr_ioctl_destroy_free_list(struct drm_device *drm_dev, void *raw_args,
                            struct drm_file *file)
{
        struct drm_pvr_ioctl_destroy_free_list_args *args = raw_args;
        struct pvr_file *pvr_file = to_pvr_file(file);
        struct pvr_free_list *free_list;

        if (args->_padding_4)
                return -EINVAL;

        free_list = xa_erase(&pvr_file->free_list_handles, args->handle);
        if (!free_list)
                return -EINVAL;

        pvr_free_list_put(free_list);
        return 0;
}

/**
 * pvr_ioctl_create_hwrt_dataset() - IOCTL to create a HWRT dataset
 * @drm_dev: [IN] DRM device.
 * @raw_args: [IN/OUT] Arguments passed to this IOCTL. This must be of type
 *                     &struct drm_pvr_ioctl_create_hwrt_dataset_args.
 * @file: [IN] DRM file private data.
 *
 * Called from userspace with %DRM_IOCTL_PVR_CREATE_HWRT_DATASET.
 *
 * Return:
 *  * 0 on success, or
 *  * Any error returned by pvr_hwrt_dataset_create().
 */
static int
pvr_ioctl_create_hwrt_dataset(struct drm_device *drm_dev, void *raw_args,
                              struct drm_file *file)
{
        struct drm_pvr_ioctl_create_hwrt_dataset_args *args = raw_args;
        struct pvr_file *pvr_file = to_pvr_file(file);
        struct pvr_hwrt_dataset *hwrt;
        int idx;
        int err;

        if (!drm_dev_enter(drm_dev, &idx))
                return -EIO;

        hwrt = pvr_hwrt_dataset_create(pvr_file, args);
        if (IS_ERR(hwrt)) {
                err = PTR_ERR(hwrt);
                goto err_drm_dev_exit;
        }

        /* Allocate object handle for userspace. */
        err = xa_alloc(&pvr_file->hwrt_handles,
                       &args->handle,
                       hwrt,
                       xa_limit_32b,
                       GFP_KERNEL);
        if (err < 0)
                goto err_cleanup;

        drm_dev_exit(idx);

        return 0;

err_cleanup:
        pvr_hwrt_dataset_put(hwrt);

err_drm_dev_exit:
        drm_dev_exit(idx);

        return err;
}

/**
 * pvr_ioctl_destroy_hwrt_dataset() - IOCTL to destroy a HWRT dataset
 * @drm_dev: [IN] DRM device.
 * @raw_args: [IN] Arguments passed to this IOCTL. This must be of type
 *                 &struct drm_pvr_ioctl_destroy_hwrt_dataset_args.
 * @file: [IN] DRM file private data.
 *
 * Called from userspace with %DRM_IOCTL_PVR_DESTROY_HWRT_DATASET.
 *
 * Return:
 *  * 0 on success, or
 *  * -%EINVAL if HWRT dataset not in object list.
 */
static int
pvr_ioctl_destroy_hwrt_dataset(struct drm_device *drm_dev, void *raw_args,
                               struct drm_file *file)
{
        struct drm_pvr_ioctl_destroy_hwrt_dataset_args *args = raw_args;
        struct pvr_file *pvr_file = to_pvr_file(file);
        struct pvr_hwrt_dataset *hwrt;

        if (args->_padding_4)
                return -EINVAL;

        hwrt = xa_erase(&pvr_file->hwrt_handles, args->handle);
        if (!hwrt)
                return -EINVAL;

        pvr_hwrt_dataset_put(hwrt);
        return 0;
}

/**
 * pvr_ioctl_create_vm_context() - IOCTL to create a VM context
 * @drm_dev: [IN] DRM device.
 * @raw_args: [IN/OUT] Arguments passed to this IOCTL. This must be of type
 *                     &struct drm_pvr_ioctl_create_vm_context_args.
 * @file: [IN] DRM file private data.
 *
 * Called from userspace with %DRM_IOCTL_PVR_CREATE_VM_CONTEXT.
 *
 * Return:
 *  * 0 on success, or
 *  * Any error returned by pvr_vm_create_context().
 */
static int
pvr_ioctl_create_vm_context(struct drm_device *drm_dev, void *raw_args,
                            struct drm_file *file)
{
        struct drm_pvr_ioctl_create_vm_context_args *args = raw_args;
        struct pvr_file *pvr_file = to_pvr_file(file);
        struct pvr_vm_context *vm_ctx;
        int idx;
        int err;

        if (!drm_dev_enter(drm_dev, &idx))
                return -EIO;

        if (args->_padding_4) {
                err = -EINVAL;
                goto err_drm_dev_exit;
        }

        vm_ctx = pvr_vm_create_context(pvr_file->pvr_dev, true);
        if (IS_ERR(vm_ctx)) {
                err = PTR_ERR(vm_ctx);
                goto err_drm_dev_exit;
        }

        /* Allocate object handle for userspace. */
        err = xa_alloc(&pvr_file->vm_ctx_handles,
                       &args->handle,
                       vm_ctx,
                       xa_limit_32b,
                       GFP_KERNEL);
        if (err < 0)
                goto err_cleanup;

        drm_dev_exit(idx);

        return 0;

err_cleanup:
        pvr_vm_context_put(vm_ctx);

err_drm_dev_exit:
        drm_dev_exit(idx);

        return err;
}

/**
 * pvr_ioctl_destroy_vm_context() - IOCTL to destroy a VM context
* @drm_dev: [IN] DRM device.
* @raw_args: [IN] Arguments passed to this IOCTL. This must be of type
*                 &struct drm_pvr_ioctl_destroy_vm_context_args.
* @file: [IN] DRM file private data.
*
* Called from userspace with %DRM_IOCTL_PVR_DESTROY_VM_CONTEXT.
*
* Return:
*  * 0 on success, or
*  * -%EINVAL if object not in object list.
 */
static int
pvr_ioctl_destroy_vm_context(struct drm_device *drm_dev, void *raw_args,
                             struct drm_file *file)
{
        struct drm_pvr_ioctl_destroy_vm_context_args *args = raw_args;
        struct pvr_file *pvr_file = to_pvr_file(file);
        struct pvr_vm_context *vm_ctx;

        if (args->_padding_4)
                return -EINVAL;

        vm_ctx = xa_erase(&pvr_file->vm_ctx_handles, args->handle);
        if (!vm_ctx)
                return -EINVAL;

        pvr_vm_context_put(vm_ctx);
        return 0;
}

/**
 * pvr_ioctl_vm_map() - IOCTL to map buffer to GPU address space.
 * @drm_dev: [IN] DRM device.
 * @raw_args: [IN] Arguments passed to this IOCTL. This must be of type
 *                 &struct drm_pvr_ioctl_vm_map_args.
 * @file: [IN] DRM file private data.
 *
 * Called from userspace with %DRM_IOCTL_PVR_VM_MAP.
 *
 * Return:
 *  * 0 on success,
 *  * -%EINVAL if &drm_pvr_ioctl_vm_op_map_args.flags is not zero,
 *  * -%EINVAL if the bounds specified by &drm_pvr_ioctl_vm_op_map_args.offset
 *    and &drm_pvr_ioctl_vm_op_map_args.size are not valid or do not fall
 *    within the buffer object specified by
 *    &drm_pvr_ioctl_vm_op_map_args.handle,
 *  * -%EINVAL if the bounds specified by
 *    &drm_pvr_ioctl_vm_op_map_args.device_addr and
 *    &drm_pvr_ioctl_vm_op_map_args.size do not form a valid device-virtual
 *    address range which falls entirely within a single heap, or
 *  * -%ENOENT if &drm_pvr_ioctl_vm_op_map_args.handle does not refer to a
 *    valid PowerVR buffer object.
 */
static int
pvr_ioctl_vm_map(struct drm_device *drm_dev, void *raw_args,
                 struct drm_file *file)
{
        struct pvr_device *pvr_dev = to_pvr_device(drm_dev);
        struct drm_pvr_ioctl_vm_map_args *args = raw_args;
        struct pvr_file *pvr_file = to_pvr_file(file);
        struct pvr_vm_context *vm_ctx;

        struct pvr_gem_object *pvr_obj;
        size_t pvr_obj_size;

        u64 offset_plus_size;
        int idx;
        int err;

        if (!drm_dev_enter(drm_dev, &idx))
                return -EIO;

        /* Initial validation of args. */
        if (args->_padding_14) {
                err = -EINVAL;
                goto err_drm_dev_exit;
        }

        if (args->flags != 0 ||
            check_add_overflow(args->offset, args->size, &offset_plus_size) ||
            !pvr_find_heap_containing(pvr_dev, args->device_addr, args->size)) {
                err = -EINVAL;
                goto err_drm_dev_exit;
        }

        vm_ctx = pvr_vm_context_lookup(pvr_file, args->vm_context_handle);
        if (!vm_ctx) {
                err = -EINVAL;
                goto err_drm_dev_exit;
        }

        pvr_obj = pvr_gem_object_from_handle(pvr_file, args->handle);
        if (!pvr_obj) {
                err = -ENOENT;
                goto err_put_vm_context;
        }

        pvr_obj_size = pvr_gem_object_size(pvr_obj);

        /*
         * Validate offset and size args. The alignment of these will be
         * checked when mapping; for now just check that they're within valid
         * bounds
         */
        if (args->offset >= pvr_obj_size || offset_plus_size > pvr_obj_size) {
                err = -EINVAL;
                goto err_put_pvr_object;
        }

        err = pvr_vm_map(vm_ctx, pvr_obj, args->offset,
                         args->device_addr, args->size);
        if (err)
                goto err_put_pvr_object;

        /*
         * In order to set up the mapping, we needed a reference to &pvr_obj.
         * However, pvr_vm_map() obtains and stores its own reference, so we
         * must release ours before returning.
         */

err_put_pvr_object:
        pvr_gem_object_put(pvr_obj);

err_put_vm_context:
        pvr_vm_context_put(vm_ctx);

err_drm_dev_exit:
        drm_dev_exit(idx);

        return err;
}

/**
 * pvr_ioctl_vm_unmap() - IOCTL to unmap buffer from GPU address space.
 * @drm_dev: [IN] DRM device.
 * @raw_args: [IN] Arguments passed to this IOCTL. This must be of type
 *                 &struct drm_pvr_ioctl_vm_unmap_args.
 * @file: [IN] DRM file private data.
 *
 * Called from userspace with %DRM_IOCTL_PVR_VM_UNMAP.
 *
 * Return:
 *  * 0 on success,
 *  * -%EINVAL if &drm_pvr_ioctl_vm_op_unmap_args.device_addr is not a valid
 *    device page-aligned device-virtual address, or
 *  * -%ENOENT if there is currently no PowerVR buffer object mapped at
 *    &drm_pvr_ioctl_vm_op_unmap_args.device_addr.
 */
static int
pvr_ioctl_vm_unmap(struct drm_device *drm_dev, void *raw_args,
                   struct drm_file *file)
{
        struct drm_pvr_ioctl_vm_unmap_args *args = raw_args;
        struct pvr_file *pvr_file = to_pvr_file(file);
        struct pvr_vm_context *vm_ctx;
        int err;

        /* Initial validation of args. */
        if (args->_padding_4)
                return -EINVAL;

        vm_ctx = pvr_vm_context_lookup(pvr_file, args->vm_context_handle);
        if (!vm_ctx)
                return -EINVAL;

        err = pvr_vm_unmap(vm_ctx, args->device_addr, args->size);

        pvr_vm_context_put(vm_ctx);

        return err;
}

/*
 * pvr_ioctl_submit_job() - IOCTL to submit a job to the GPU
 * @drm_dev: [IN] DRM device.
 * @raw_args: [IN] Arguments passed to this IOCTL. This must be of type
 *                 &struct drm_pvr_ioctl_submit_job_args.
 * @file: [IN] DRM file private data.
 *
 * Called from userspace with %DRM_IOCTL_PVR_SUBMIT_JOB.
 *
 * Return:
 *  * 0 on success, or
 *  * -%EINVAL if arguments are invalid.
 */
static int
pvr_ioctl_submit_jobs(struct drm_device *drm_dev, void *raw_args,
                      struct drm_file *file)
{
        struct drm_pvr_ioctl_submit_jobs_args *args = raw_args;
        struct pvr_device *pvr_dev = to_pvr_device(drm_dev);
        struct pvr_file *pvr_file = to_pvr_file(file);
        int idx;
        int err;

        if (!drm_dev_enter(drm_dev, &idx))
                return -EIO;

        err = pvr_submit_jobs(pvr_dev, pvr_file, args);

        drm_dev_exit(idx);

        return err;
}

int
pvr_get_uobj(u64 usr_ptr, u32 usr_stride, u32 min_stride, u32 obj_size, void *out)
{
        if (usr_stride < min_stride)
                return -EINVAL;

        return copy_struct_from_user(out, obj_size, u64_to_user_ptr(usr_ptr), usr_stride);
}

int
pvr_set_uobj(u64 usr_ptr, u32 usr_stride, u32 min_stride, u32 obj_size, const void *in)
{
        if (usr_stride < min_stride)
                return -EINVAL;

        if (copy_to_user(u64_to_user_ptr(usr_ptr), in, min_t(u32, usr_stride, obj_size)))
                return -EFAULT;

        if (usr_stride > obj_size &&
            clear_user(u64_to_user_ptr(usr_ptr + obj_size), usr_stride - obj_size)) {
                return -EFAULT;
        }

        return 0;
}

int
pvr_get_uobj_array(const struct drm_pvr_obj_array *in, u32 min_stride, u32 obj_size, void **out)
{
        int ret = 0;
        void *out_alloc;

        if (in->stride < min_stride)
                return -EINVAL;

        if (!in->count)
                return 0;

        out_alloc = kvmalloc_array(in->count, obj_size, GFP_KERNEL);
        if (!out_alloc)
                return -ENOMEM;

        if (obj_size == in->stride) {
                if (copy_from_user(out_alloc, u64_to_user_ptr(in->array),
                                   (unsigned long)obj_size * in->count))
                        ret = -EFAULT;
        } else {
                void __user *in_ptr = u64_to_user_ptr(in->array);
                void *out_ptr = out_alloc;

                for (u32 i = 0; i < in->count; i++) {
                        ret = copy_struct_from_user(out_ptr, obj_size, in_ptr, in->stride);
                        if (ret)
                                break;

                        out_ptr += obj_size;
                        in_ptr += in->stride;
                }
        }

        if (ret) {
                kvfree(out_alloc);
                return ret;
        }

        *out = out_alloc;
        return 0;
}

int
pvr_set_uobj_array(const struct drm_pvr_obj_array *out, u32 min_stride, u32 obj_size,
                   const void *in)
{
        if (out->stride < min_stride)
                return -EINVAL;

        if (!out->count)
                return 0;

        if (obj_size == out->stride) {
                if (copy_to_user(u64_to_user_ptr(out->array), in,
                                 (unsigned long)obj_size * out->count))
                        return -EFAULT;
        } else {
                u32 cpy_elem_size = min_t(u32, out->stride, obj_size);
                void __user *out_ptr = u64_to_user_ptr(out->array);
                const void *in_ptr = in;

                for (u32 i = 0; i < out->count; i++) {
                        if (copy_to_user(out_ptr, in_ptr, cpy_elem_size))
                                return -EFAULT;

                        out_ptr += obj_size;
                        in_ptr += out->stride;
                }

                if (out->stride > obj_size &&
                    clear_user(u64_to_user_ptr(out->array + obj_size),
                               out->stride - obj_size)) {
                        return -EFAULT;
                }
        }

        return 0;
}

#define DRM_PVR_IOCTL(_name, _func, _flags) \
        DRM_IOCTL_DEF_DRV(PVR_##_name, pvr_ioctl_##_func, _flags)

/* clang-format off */

static const struct drm_ioctl_desc pvr_drm_driver_ioctls[] = {
        DRM_PVR_IOCTL(DEV_QUERY, dev_query, DRM_RENDER_ALLOW),
        DRM_PVR_IOCTL(CREATE_BO, create_bo, DRM_RENDER_ALLOW),
        DRM_PVR_IOCTL(GET_BO_MMAP_OFFSET, get_bo_mmap_offset, DRM_RENDER_ALLOW),
        DRM_PVR_IOCTL(CREATE_VM_CONTEXT, create_vm_context, DRM_RENDER_ALLOW),
        DRM_PVR_IOCTL(DESTROY_VM_CONTEXT, destroy_vm_context, DRM_RENDER_ALLOW),
        DRM_PVR_IOCTL(VM_MAP, vm_map, DRM_RENDER_ALLOW),
        DRM_PVR_IOCTL(VM_UNMAP, vm_unmap, DRM_RENDER_ALLOW),
        DRM_PVR_IOCTL(CREATE_CONTEXT, create_context, DRM_RENDER_ALLOW),
        DRM_PVR_IOCTL(DESTROY_CONTEXT, destroy_context, DRM_RENDER_ALLOW),
        DRM_PVR_IOCTL(CREATE_FREE_LIST, create_free_list, DRM_RENDER_ALLOW),
        DRM_PVR_IOCTL(DESTROY_FREE_LIST, destroy_free_list, DRM_RENDER_ALLOW),
        DRM_PVR_IOCTL(CREATE_HWRT_DATASET, create_hwrt_dataset, DRM_RENDER_ALLOW),
        DRM_PVR_IOCTL(DESTROY_HWRT_DATASET, destroy_hwrt_dataset, DRM_RENDER_ALLOW),
        DRM_PVR_IOCTL(SUBMIT_JOBS, submit_jobs, DRM_RENDER_ALLOW),
};

/* clang-format on */

#undef DRM_PVR_IOCTL

/**
 * pvr_drm_driver_open() - Driver callback when a new &struct drm_file is opened
 * @drm_dev: [IN] DRM device.
 * @file: [IN] DRM file private data.
 *
 * Allocates powervr-specific file private data (&struct pvr_file).
 *
 * Registered in &pvr_drm_driver.
 *
 * Return:
 *  * 0 on success,
 *  * -%ENOMEM if the allocation of a &struct ipvr_file fails, or
 *  * Any error returned by pvr_memory_context_init().
 */
static int
pvr_drm_driver_open(struct drm_device *drm_dev, struct drm_file *file)
{
        struct pvr_device *pvr_dev = to_pvr_device(drm_dev);
        struct pvr_file *pvr_file;

        pvr_file = kzalloc_obj(*pvr_file);
        if (!pvr_file)
                return -ENOMEM;

        /*
         * Store reference to base DRM file private data for use by
         * from_pvr_file.
         */
        pvr_file->file = file;

        /*
         * Store reference to powervr-specific outer device struct in file
         * private data for convenient access.
         */
        pvr_file->pvr_dev = pvr_dev;

        INIT_LIST_HEAD(&pvr_file->contexts);

        xa_init_flags(&pvr_file->ctx_handles, XA_FLAGS_ALLOC1);
        xa_init_flags(&pvr_file->free_list_handles, XA_FLAGS_ALLOC1);
        xa_init_flags(&pvr_file->hwrt_handles, XA_FLAGS_ALLOC1);
        xa_init_flags(&pvr_file->vm_ctx_handles, XA_FLAGS_ALLOC1);

        /*
         * Store reference to powervr-specific file private data in DRM file
         * private data.
         */
        file->driver_priv = pvr_file;

        return 0;
}

/**
 * pvr_drm_driver_postclose() - One of the driver callbacks when a &struct
 * drm_file is closed.
 * @drm_dev: [IN] DRM device (unused).
 * @file: [IN] DRM file private data.
 *
 * Frees powervr-specific file private data (&struct pvr_file).
 *
 * Registered in &pvr_drm_driver.
 */
static void
pvr_drm_driver_postclose(__always_unused struct drm_device *drm_dev,
                         struct drm_file *file)
{
        struct pvr_file *pvr_file = to_pvr_file(file);

        /* Kill remaining contexts. */
        pvr_destroy_contexts_for_file(pvr_file);

        /* Drop references on any remaining objects. */
        pvr_destroy_free_lists_for_file(pvr_file);
        pvr_destroy_hwrt_datasets_for_file(pvr_file);
        pvr_destroy_vm_contexts_for_file(pvr_file);

        kfree(pvr_file);
        file->driver_priv = NULL;
}

DEFINE_DRM_GEM_FOPS(pvr_drm_driver_fops);

static struct drm_driver pvr_drm_driver = {
        .driver_features = DRIVER_GEM | DRIVER_GEM_GPUVA | DRIVER_RENDER |
                           DRIVER_SYNCOBJ | DRIVER_SYNCOBJ_TIMELINE,
        .open = pvr_drm_driver_open,
        .postclose = pvr_drm_driver_postclose,
        .ioctls = pvr_drm_driver_ioctls,
        .num_ioctls = ARRAY_SIZE(pvr_drm_driver_ioctls),
        .fops = &pvr_drm_driver_fops,
#if defined(CONFIG_DEBUG_FS)
        .debugfs_init = pvr_debugfs_init,
#endif

        .name = PVR_DRIVER_NAME,
        .desc = PVR_DRIVER_DESC,
        .major = PVR_DRIVER_MAJOR,
        .minor = PVR_DRIVER_MINOR,
        .patchlevel = PVR_DRIVER_PATCHLEVEL,

        .gem_prime_import_sg_table = drm_gem_shmem_prime_import_sg_table,
        .gem_create_object = pvr_gem_create_object,
};

static int
pvr_probe(struct platform_device *plat_dev)
{
        struct pvr_device *pvr_dev;
        struct drm_device *drm_dev;
        int err;

        pvr_dev = devm_drm_dev_alloc(&plat_dev->dev, &pvr_drm_driver,
                                     struct pvr_device, base);
        if (IS_ERR(pvr_dev))
                return PTR_ERR(pvr_dev);

        drm_dev = &pvr_dev->base;

        platform_set_drvdata(plat_dev, drm_dev);

        err = pvr_power_domains_init(pvr_dev);
        if (err)
                return err;

        init_rwsem(&pvr_dev->reset_sem);

        pvr_context_device_init(pvr_dev);

        err = pvr_queue_device_init(pvr_dev);
        if (err)
                goto err_context_fini;

        devm_pm_runtime_enable(&plat_dev->dev);
        pm_runtime_mark_last_busy(&plat_dev->dev);

        pm_runtime_set_autosuspend_delay(&plat_dev->dev, 50);
        pm_runtime_use_autosuspend(&plat_dev->dev);
        pvr_watchdog_init(pvr_dev);

        err = pvr_device_init(pvr_dev);
        if (err)
                goto err_watchdog_fini;

        err = drm_dev_register(drm_dev, 0);
        if (err)
                goto err_device_fini;

        xa_init_flags(&pvr_dev->free_list_ids, XA_FLAGS_ALLOC1);
        xa_init_flags(&pvr_dev->job_ids, XA_FLAGS_ALLOC1);

        return 0;

err_device_fini:
        pvr_device_fini(pvr_dev);

err_watchdog_fini:
        pvr_watchdog_fini(pvr_dev);

        pvr_queue_device_fini(pvr_dev);

err_context_fini:
        pvr_context_device_fini(pvr_dev);

        pvr_power_domains_fini(pvr_dev);

        return err;
}

static void pvr_remove(struct platform_device *plat_dev)
{
        struct drm_device *drm_dev = platform_get_drvdata(plat_dev);
        struct pvr_device *pvr_dev = to_pvr_device(drm_dev);

        WARN_ON(!xa_empty(&pvr_dev->job_ids));
        WARN_ON(!xa_empty(&pvr_dev->free_list_ids));

        xa_destroy(&pvr_dev->job_ids);
        xa_destroy(&pvr_dev->free_list_ids);

        pm_runtime_suspend(drm_dev->dev);
        pvr_device_fini(pvr_dev);
        drm_dev_unplug(drm_dev);
        pvr_watchdog_fini(pvr_dev);
        pvr_queue_device_fini(pvr_dev);
        pvr_context_device_fini(pvr_dev);
        pvr_power_domains_fini(pvr_dev);
}

static const struct pvr_device_data pvr_device_data_manual = {
        .pwr_ops = &pvr_power_sequence_ops_manual,
};

static const struct pvr_device_data pvr_device_data_pwrseq = {
        .pwr_ops = &pvr_power_sequence_ops_pwrseq,
};

static const struct of_device_id dt_match[] = {
        {
                .compatible = "thead,th1520-gpu",
                .data = &pvr_device_data_pwrseq,
        },
        {
                .compatible = "img,img-rogue",
                .data = &pvr_device_data_manual,
        },

        /*
         * This legacy compatible string was introduced early on before the more generic
         * "img,img-rogue" was added. Keep it around here for compatibility, but never use
         * "img,img-axe" in new devicetrees.
         */
        {
                .compatible = "img,img-axe",
                .data = &pvr_device_data_manual,
        },
        {}
};
MODULE_DEVICE_TABLE(of, dt_match);

static const struct dev_pm_ops pvr_pm_ops = {
        RUNTIME_PM_OPS(pvr_power_device_suspend, pvr_power_device_resume, pvr_power_device_idle)
};

static struct platform_driver pvr_driver = {
        .probe = pvr_probe,
        .remove = pvr_remove,
        .driver = {
                .name = PVR_DRIVER_NAME,
                .pm = &pvr_pm_ops,
                .of_match_table = dt_match,
        },
};
module_platform_driver(pvr_driver);

MODULE_AUTHOR("Imagination Technologies Ltd.");
MODULE_DESCRIPTION(PVR_DRIVER_DESC);
MODULE_LICENSE("Dual MIT/GPL");
MODULE_IMPORT_NS("DMA_BUF");
MODULE_FIRMWARE("powervr/rogue_33.15.11.3_v1.fw");
MODULE_FIRMWARE("powervr/rogue_36.52.104.182_v1.fw");
MODULE_FIRMWARE("powervr/rogue_36.53.104.796_v1.fw");