root/drivers/gpu/drm/v3d/v3d_bo.c
// SPDX-License-Identifier: GPL-2.0+
/* Copyright (C) 2015-2018 Broadcom */

/**
 * DOC: V3D GEM BO management support
 *
 * Compared to VC4 (V3D 2.x), V3D 3.3 introduces an MMU between the
 * GPU and the bus, allowing us to use shmem objects for our storage
 * instead of CMA.
 *
 * Physically contiguous objects may still be imported to V3D, but the
 * driver doesn't allocate physically contiguous objects on its own.
 * Display engines requiring physically contiguous allocations should
 * look into Mesa's "renderonly" support (as used by the Mesa pl111
 * driver) for an example of how to integrate with V3D.
 */

#include <linux/dma-buf.h>
#include <linux/vmalloc.h>

#include <drm/drm_print.h>

#include "v3d_drv.h"
#include "uapi/drm/v3d_drm.h"

static enum drm_gem_object_status v3d_gem_status(struct drm_gem_object *obj)
{
        struct v3d_bo *bo = to_v3d_bo(obj);
        enum drm_gem_object_status res = 0;

        if (bo->base.pages)
                res |= DRM_GEM_OBJECT_RESIDENT;

        return res;
}

/* Called DRM core on the last userspace/kernel unreference of the
 * BO.
 */
void v3d_free_object(struct drm_gem_object *obj)
{
        struct v3d_dev *v3d = to_v3d_dev(obj->dev);
        struct v3d_bo *bo = to_v3d_bo(obj);

        if (bo->vaddr)
                v3d_put_bo_vaddr(bo);

        v3d_mmu_remove_ptes(bo);

        mutex_lock(&v3d->bo_lock);
        v3d->bo_stats.num_allocated--;
        v3d->bo_stats.pages_allocated -= obj->size >> V3D_MMU_PAGE_SHIFT;
        mutex_unlock(&v3d->bo_lock);

        spin_lock(&v3d->mm_lock);
        drm_mm_remove_node(&bo->node);
        spin_unlock(&v3d->mm_lock);

        /* GPU execution may have dirtied any pages in the BO. */
        bo->base.pages_mark_dirty_on_put = true;

        drm_gem_shmem_free(&bo->base);
}

static const struct drm_gem_object_funcs v3d_gem_funcs = {
        .free = v3d_free_object,
        .print_info = drm_gem_shmem_object_print_info,
        .pin = drm_gem_shmem_object_pin,
        .unpin = drm_gem_shmem_object_unpin,
        .get_sg_table = drm_gem_shmem_object_get_sg_table,
        .vmap = drm_gem_shmem_object_vmap,
        .vunmap = drm_gem_shmem_object_vunmap,
        .mmap = drm_gem_shmem_object_mmap,
        .status = v3d_gem_status,
        .vm_ops = &drm_gem_shmem_vm_ops,
};

/* gem_create_object function for allocating a BO struct and doing
 * early setup.
 */
struct drm_gem_object *v3d_create_object(struct drm_device *dev, size_t size)
{
        struct v3d_bo *bo;
        struct drm_gem_object *obj;

        if (size == 0)
                return ERR_PTR(-EINVAL);

        bo = kzalloc_obj(*bo);
        if (!bo)
                return ERR_PTR(-ENOMEM);
        obj = &bo->base.base;

        obj->funcs = &v3d_gem_funcs;
        bo->base.map_wc = true;
        INIT_LIST_HEAD(&bo->unref_head);

        return &bo->base.base;
}

static int
v3d_bo_create_finish(struct drm_gem_object *obj)
{
        struct v3d_dev *v3d = to_v3d_dev(obj->dev);
        struct v3d_bo *bo = to_v3d_bo(obj);
        struct sg_table *sgt;
        u64 align;
        int ret;

        /* So far we pin the BO in the MMU for its lifetime, so use
         * shmem's helper for getting a lifetime sgt.
         */
        sgt = drm_gem_shmem_get_pages_sgt(&bo->base);
        if (IS_ERR(sgt))
                return PTR_ERR(sgt);

        if (!drm_gem_get_huge_mnt(obj->dev))
                align = SZ_4K;
        else if (obj->size >= SZ_1M)
                align = SZ_1M;
        else if (obj->size >= SZ_64K)
                align = SZ_64K;
        else
                align = SZ_4K;

        spin_lock(&v3d->mm_lock);
        /* Allocate the object's space in the GPU's page tables.
         * Inserting PTEs will happen later, but the offset is for the
         * lifetime of the BO.
         */
        ret = drm_mm_insert_node_generic(&v3d->mm, &bo->node,
                                         obj->size >> V3D_MMU_PAGE_SHIFT,
                                         align >> V3D_MMU_PAGE_SHIFT, 0, 0);
        spin_unlock(&v3d->mm_lock);
        if (ret)
                return ret;

        /* Track stats for /debug/dri/n/bo_stats. */
        mutex_lock(&v3d->bo_lock);
        v3d->bo_stats.num_allocated++;
        v3d->bo_stats.pages_allocated += obj->size >> V3D_MMU_PAGE_SHIFT;
        mutex_unlock(&v3d->bo_lock);

        v3d_mmu_insert_ptes(bo);

        return 0;
}

struct v3d_bo *v3d_bo_create(struct drm_device *dev, struct drm_file *file_priv,
                             size_t unaligned_size)
{
        struct drm_gem_shmem_object *shmem_obj;
        struct v3d_bo *bo;
        int ret;

        shmem_obj = drm_gem_shmem_create(dev, unaligned_size);
        if (IS_ERR(shmem_obj))
                return ERR_CAST(shmem_obj);
        bo = to_v3d_bo(&shmem_obj->base);
        bo->vaddr = NULL;

        ret = v3d_bo_create_finish(&shmem_obj->base);
        if (ret)
                goto free_obj;

        return bo;

free_obj:
        drm_gem_shmem_free(shmem_obj);
        return ERR_PTR(ret);
}

struct drm_gem_object *
v3d_prime_import_sg_table(struct drm_device *dev,
                          struct dma_buf_attachment *attach,
                          struct sg_table *sgt)
{
        struct drm_gem_object *obj;
        int ret;

        obj = drm_gem_shmem_prime_import_sg_table(dev, attach, sgt);
        if (IS_ERR(obj))
                return obj;

        ret = v3d_bo_create_finish(obj);
        if (ret) {
                drm_gem_shmem_free(&to_v3d_bo(obj)->base);
                return ERR_PTR(ret);
        }

        return obj;
}

void v3d_get_bo_vaddr(struct v3d_bo *bo)
{
        struct drm_gem_shmem_object *obj = &bo->base;

        bo->vaddr = vmap(obj->pages, obj->base.size >> PAGE_SHIFT, VM_MAP,
                         pgprot_writecombine(PAGE_KERNEL));
}

void v3d_put_bo_vaddr(struct v3d_bo *bo)
{
        vunmap(bo->vaddr);
        bo->vaddr = NULL;
}

int v3d_create_bo_ioctl(struct drm_device *dev, void *data,
                        struct drm_file *file_priv)
{
        struct drm_v3d_create_bo *args = data;
        struct v3d_bo *bo = NULL;
        int ret;

        if (args->flags != 0) {
                drm_dbg(dev, "unknown create_bo flags: %d\n", args->flags);
                return -EINVAL;
        }

        bo = v3d_bo_create(dev, file_priv, PAGE_ALIGN(args->size));
        if (IS_ERR(bo))
                return PTR_ERR(bo);

        args->offset = bo->node.start << V3D_MMU_PAGE_SHIFT;

        ret = drm_gem_handle_create(file_priv, &bo->base.base, &args->handle);
        drm_gem_object_put(&bo->base.base);

        return ret;
}

int v3d_mmap_bo_ioctl(struct drm_device *dev, void *data,
                      struct drm_file *file_priv)
{
        struct drm_v3d_mmap_bo *args = data;
        struct drm_gem_object *gem_obj;

        if (args->flags != 0) {
                drm_dbg(dev, "unknown mmap_bo flags: %d\n", args->flags);
                return -EINVAL;
        }

        gem_obj = drm_gem_object_lookup(file_priv, args->handle);
        if (!gem_obj) {
                drm_dbg(dev, "Failed to look up GEM BO %d\n", args->handle);
                return -ENOENT;
        }

        args->offset = drm_vma_node_offset_addr(&gem_obj->vma_node);
        drm_gem_object_put(gem_obj);

        return 0;
}

int v3d_get_bo_offset_ioctl(struct drm_device *dev, void *data,
                            struct drm_file *file_priv)
{
        struct drm_v3d_get_bo_offset *args = data;
        struct drm_gem_object *gem_obj;
        struct v3d_bo *bo;

        gem_obj = drm_gem_object_lookup(file_priv, args->handle);
        if (!gem_obj) {
                drm_dbg(dev, "Failed to look up GEM BO %d\n", args->handle);
                return -ENOENT;
        }
        bo = to_v3d_bo(gem_obj);

        args->offset = bo->node.start << V3D_MMU_PAGE_SHIFT;

        drm_gem_object_put(gem_obj);
        return 0;
}

int
v3d_wait_bo_ioctl(struct drm_device *dev, void *data,
                  struct drm_file *file_priv)
{
        int ret;
        struct drm_v3d_wait_bo *args = data;
        ktime_t start = ktime_get();
        u64 delta_ns;
        unsigned long timeout_jiffies =
                nsecs_to_jiffies_timeout(args->timeout_ns);

        if (args->pad != 0)
                return -EINVAL;

        ret = drm_gem_dma_resv_wait(file_priv, args->handle,
                                    true, timeout_jiffies);

        /* Decrement the user's timeout, in case we got interrupted
         * such that the ioctl will be restarted.
         */
        delta_ns = ktime_to_ns(ktime_sub(ktime_get(), start));
        if (delta_ns < args->timeout_ns)
                args->timeout_ns -= delta_ns;
        else
                args->timeout_ns = 0;

        /* Asked to wait beyond the jiffy/scheduler precision? */
        if (ret == -ETIME && args->timeout_ns)
                ret = -EAGAIN;

        return ret;
}