root/drivers/remoteproc/remoteproc_virtio.c
// SPDX-License-Identifier: GPL-2.0-only
/*
 * Remote processor messaging transport (OMAP platform-specific bits)
 *
 * Copyright (C) 2011 Texas Instruments, Inc.
 * Copyright (C) 2011 Google, Inc.
 *
 * Ohad Ben-Cohen <ohad@wizery.com>
 * Brian Swetland <swetland@google.com>
 */

#include <linux/dma-direct.h>
#include <linux/dma-map-ops.h>
#include <linux/dma-mapping.h>
#include <linux/export.h>
#include <linux/of_reserved_mem.h>
#include <linux/platform_device.h>
#include <linux/remoteproc.h>
#include <linux/virtio.h>
#include <linux/virtio_config.h>
#include <linux/virtio_ids.h>
#include <linux/virtio_ring.h>
#include <linux/err.h>
#include <linux/kref.h>
#include <linux/slab.h>

#include "remoteproc_internal.h"

static int copy_dma_range_map(struct device *to, struct device *from)
{
        const struct bus_dma_region *map = from->dma_range_map, *new_map, *r;
        int num_ranges = 0;

        if (!map)
                return 0;

        for (r = map; r->size; r++)
                num_ranges++;

        new_map = kmemdup(map, array_size(num_ranges + 1, sizeof(*map)),
                          GFP_KERNEL);
        if (!new_map)
                return -ENOMEM;
        to->dma_range_map = new_map;
        return 0;
}

static struct rproc_vdev *vdev_to_rvdev(struct virtio_device *vdev)
{
        struct platform_device *pdev;

        pdev = container_of(vdev->dev.parent, struct platform_device, dev);

        return platform_get_drvdata(pdev);
}

static  struct rproc *vdev_to_rproc(struct virtio_device *vdev)
{
        struct rproc_vdev *rvdev = vdev_to_rvdev(vdev);

        return rvdev->rproc;
}

/* kick the remote processor, and let it know which virtqueue to poke at */
static bool rproc_virtio_notify(struct virtqueue *vq)
{
        struct rproc_vring *rvring = vq->priv;
        struct rproc *rproc = rvring->rvdev->rproc;
        int notifyid = rvring->notifyid;

        dev_dbg(&rproc->dev, "kicking vq index: %d\n", notifyid);

        rproc->ops->kick(rproc, notifyid);
        return true;
}

/**
 * rproc_vq_interrupt() - tell remoteproc that a virtqueue is interrupted
 * @rproc: handle to the remote processor
 * @notifyid: index of the signalled virtqueue (unique per this @rproc)
 *
 * This function should be called by the platform-specific rproc driver,
 * when the remote processor signals that a specific virtqueue has pending
 * messages available.
 *
 * Return: IRQ_NONE if no message was found in the @notifyid virtqueue,
 * and otherwise returns IRQ_HANDLED.
 */
irqreturn_t rproc_vq_interrupt(struct rproc *rproc, int notifyid)
{
        struct rproc_vring *rvring;

        dev_dbg(&rproc->dev, "vq index %d is interrupted\n", notifyid);

        rvring = idr_find(&rproc->notifyids, notifyid);
        if (!rvring || !rvring->vq)
                return IRQ_NONE;

        return vring_interrupt(0, rvring->vq);
}
EXPORT_SYMBOL(rproc_vq_interrupt);

static struct virtqueue *rp_find_vq(struct virtio_device *vdev,
                                    unsigned int id,
                                    void (*callback)(struct virtqueue *vq),
                                    const char *name, bool ctx)
{
        struct rproc_vdev *rvdev = vdev_to_rvdev(vdev);
        struct rproc *rproc = vdev_to_rproc(vdev);
        struct device *dev = &rproc->dev;
        struct rproc_mem_entry *mem;
        struct rproc_vring *rvring;
        struct fw_rsc_vdev *rsc;
        struct virtqueue *vq;
        void *addr;
        int num, size;

        /* we're temporarily limited to two virtqueues per rvdev */
        if (id >= ARRAY_SIZE(rvdev->vring))
                return ERR_PTR(-EINVAL);

        if (!name)
                return NULL;

        /* Search allocated memory region by name */
        mem = rproc_find_carveout_by_name(rproc, "vdev%dvring%d", rvdev->index,
                                          id);
        if (!mem || !mem->va)
                return ERR_PTR(-ENOMEM);

        rvring = &rvdev->vring[id];
        addr = mem->va;
        num = rvring->num;

        /* zero vring */
        size = vring_size(num, rvring->align);
        memset(addr, 0, size);

        dev_dbg(dev, "vring%d: va %p qsz %d notifyid %d\n",
                id, addr, num, rvring->notifyid);

        /*
         * Create the new vq, and tell virtio we're not interested in
         * the 'weak' smp barriers, since we're talking with a real device.
         */
        vq = vring_new_virtqueue(id, num, rvring->align, vdev, false, ctx,
                                 addr, rproc_virtio_notify, callback, name);
        if (!vq) {
                dev_err(dev, "vring_new_virtqueue %s failed\n", name);
                rproc_free_vring(rvring);
                return ERR_PTR(-ENOMEM);
        }

        vq->num_max = num;

        rvring->vq = vq;
        vq->priv = rvring;

        /* Update vring in resource table */
        rsc = (void *)rproc->table_ptr + rvdev->rsc_offset;
        rsc->vring[id].da = mem->da;

        return vq;
}

static void __rproc_virtio_del_vqs(struct virtio_device *vdev)
{
        struct virtqueue *vq, *n;
        struct rproc_vring *rvring;

        list_for_each_entry_safe(vq, n, &vdev->vqs, list) {
                rvring = vq->priv;
                rvring->vq = NULL;
                vring_del_virtqueue(vq);
        }
}

static void rproc_virtio_del_vqs(struct virtio_device *vdev)
{
        __rproc_virtio_del_vqs(vdev);
}

static int rproc_virtio_find_vqs(struct virtio_device *vdev, unsigned int nvqs,
                                 struct virtqueue *vqs[],
                                 struct virtqueue_info vqs_info[],
                                 struct irq_affinity *desc)
{
        int i, ret, queue_idx = 0;

        for (i = 0; i < nvqs; ++i) {
                struct virtqueue_info *vqi = &vqs_info[i];

                if (!vqi->name) {
                        vqs[i] = NULL;
                        continue;
                }

                vqs[i] = rp_find_vq(vdev, queue_idx++, vqi->callback,
                                    vqi->name, vqi->ctx);
                if (IS_ERR(vqs[i])) {
                        ret = PTR_ERR(vqs[i]);
                        goto error;
                }
        }

        return 0;

error:
        __rproc_virtio_del_vqs(vdev);
        return ret;
}

static u8 rproc_virtio_get_status(struct virtio_device *vdev)
{
        struct rproc_vdev *rvdev = vdev_to_rvdev(vdev);
        struct fw_rsc_vdev *rsc;

        rsc = (void *)rvdev->rproc->table_ptr + rvdev->rsc_offset;

        return rsc->status;
}

static void rproc_virtio_set_status(struct virtio_device *vdev, u8 status)
{
        struct rproc_vdev *rvdev = vdev_to_rvdev(vdev);
        struct fw_rsc_vdev *rsc;

        rsc = (void *)rvdev->rproc->table_ptr + rvdev->rsc_offset;

        rsc->status = status;
        dev_dbg(&vdev->dev, "status: %d\n", status);
}

static void rproc_virtio_reset(struct virtio_device *vdev)
{
        struct rproc_vdev *rvdev = vdev_to_rvdev(vdev);
        struct fw_rsc_vdev *rsc;

        rsc = (void *)rvdev->rproc->table_ptr + rvdev->rsc_offset;

        rsc->status = 0;
        dev_dbg(&vdev->dev, "reset !\n");
}

/* provide the vdev features as retrieved from the firmware */
static u64 rproc_virtio_get_features(struct virtio_device *vdev)
{
        struct rproc_vdev *rvdev = vdev_to_rvdev(vdev);
        struct fw_rsc_vdev *rsc;

        rsc = (void *)rvdev->rproc->table_ptr + rvdev->rsc_offset;

        return rsc->dfeatures;
}

static void rproc_transport_features(struct virtio_device *vdev)
{
        /*
         * Packed ring isn't enabled on remoteproc for now,
         * because remoteproc uses vring_new_virtqueue() which
         * creates virtio rings on preallocated memory.
         */
        __virtio_clear_bit(vdev, VIRTIO_F_RING_PACKED);
}

static int rproc_virtio_finalize_features(struct virtio_device *vdev)
{
        struct rproc_vdev *rvdev = vdev_to_rvdev(vdev);
        struct fw_rsc_vdev *rsc;

        rsc = (void *)rvdev->rproc->table_ptr + rvdev->rsc_offset;

        /* Give virtio_ring a chance to accept features */
        vring_transport_features(vdev);

        /* Give virtio_rproc a chance to accept features. */
        rproc_transport_features(vdev);

        /* Make sure we don't have any features > 32 bits! */
        BUG_ON((u32)vdev->features != vdev->features);

        /*
         * Remember the finalized features of our vdev, and provide it
         * to the remote processor once it is powered on.
         */
        rsc->gfeatures = vdev->features;

        return 0;
}

static void rproc_virtio_get(struct virtio_device *vdev, unsigned int offset,
                             void *buf, unsigned int len)
{
        struct rproc_vdev *rvdev = vdev_to_rvdev(vdev);
        struct fw_rsc_vdev *rsc;
        void *cfg;

        rsc = (void *)rvdev->rproc->table_ptr + rvdev->rsc_offset;
        cfg = &rsc->vring[rsc->num_of_vrings];

        if (offset + len > rsc->config_len || offset + len < len) {
                dev_err(&vdev->dev, "rproc_virtio_get: access out of bounds\n");
                return;
        }

        memcpy(buf, cfg + offset, len);
}

static void rproc_virtio_set(struct virtio_device *vdev, unsigned int offset,
                             const void *buf, unsigned int len)
{
        struct rproc_vdev *rvdev = vdev_to_rvdev(vdev);
        struct fw_rsc_vdev *rsc;
        void *cfg;

        rsc = (void *)rvdev->rproc->table_ptr + rvdev->rsc_offset;
        cfg = &rsc->vring[rsc->num_of_vrings];

        if (offset + len > rsc->config_len || offset + len < len) {
                dev_err(&vdev->dev, "rproc_virtio_set: access out of bounds\n");
                return;
        }

        memcpy(cfg + offset, buf, len);
}

static const struct virtio_config_ops rproc_virtio_config_ops = {
        .get_features   = rproc_virtio_get_features,
        .finalize_features = rproc_virtio_finalize_features,
        .find_vqs       = rproc_virtio_find_vqs,
        .del_vqs        = rproc_virtio_del_vqs,
        .reset          = rproc_virtio_reset,
        .set_status     = rproc_virtio_set_status,
        .get_status     = rproc_virtio_get_status,
        .get            = rproc_virtio_get,
        .set            = rproc_virtio_set,
};

/*
 * This function is called whenever vdev is released, and is responsible
 * to decrement the remote processor's refcount which was taken when vdev was
 * added.
 *
 * Never call this function directly; it will be called by the driver
 * core when needed.
 */
static void rproc_virtio_dev_release(struct device *dev)
{
        struct virtio_device *vdev = dev_to_virtio(dev);
        struct rproc_vdev *rvdev = vdev_to_rvdev(vdev);

        kfree(vdev);

        of_reserved_mem_device_release(&rvdev->pdev->dev);
        dma_release_coherent_memory(&rvdev->pdev->dev);

        put_device(&rvdev->pdev->dev);
}

/**
 * rproc_add_virtio_dev() - register an rproc-induced virtio device
 * @rvdev: the remote vdev
 * @id: the device type identification (used to match it with a driver).
 *
 * This function registers a virtio device. This vdev's partent is
 * the rproc device.
 *
 * Return: 0 on success or an appropriate error value otherwise
 */
static int rproc_add_virtio_dev(struct rproc_vdev *rvdev, int id)
{
        struct rproc *rproc = rvdev->rproc;
        struct device *dev = &rvdev->pdev->dev;
        struct virtio_device *vdev;
        struct rproc_mem_entry *mem;
        int ret;

        if (rproc->ops->kick == NULL) {
                ret = -EINVAL;
                dev_err(dev, ".kick method not defined for %s\n", rproc->name);
                goto out;
        }

        /* Try to find dedicated vdev buffer carveout */
        mem = rproc_find_carveout_by_name(rproc, "vdev%dbuffer", rvdev->index);
        if (mem) {
                phys_addr_t pa;

                if (mem->of_resm_idx != -1) {
                        struct device_node *np = rproc->dev.parent->of_node;

                        /* Associate reserved memory to vdev device */
                        ret = of_reserved_mem_device_init_by_idx(dev, np,
                                                                 mem->of_resm_idx);
                        if (ret) {
                                dev_err(dev, "Can't associate reserved memory\n");
                                goto out;
                        }
                } else {
                        if (mem->va) {
                                dev_warn(dev, "vdev %d buffer already mapped\n",
                                         rvdev->index);
                                pa = rproc_va_to_pa(mem->va);
                        } else {
                                /* Use dma address as carveout no memmapped yet */
                                pa = (phys_addr_t)mem->dma;
                        }

                        /* Associate vdev buffer memory pool to vdev subdev */
                        ret = dma_declare_coherent_memory(dev, pa,
                                                           mem->da,
                                                           mem->len);
                        if (ret < 0) {
                                dev_err(dev, "Failed to associate buffer\n");
                                goto out;
                        }
                }
        } else {
                struct device_node *np = rproc->dev.parent->of_node;

                /*
                 * If we don't have dedicated buffer, just attempt to re-assign
                 * the reserved memory from our parent. A default memory-region
                 * at index 0 from the parent's memory-regions is assigned for
                 * the rvdev dev to allocate from. Failure is non-critical and
                 * the allocations will fall back to global pools, so don't
                 * check return value either.
                 */
                of_reserved_mem_device_init_by_idx(dev, np, 0);
        }

        /* Allocate virtio device */
        vdev = kzalloc_obj(*vdev);
        if (!vdev) {
                ret = -ENOMEM;
                goto out;
        }
        vdev->id.device = id,
        vdev->config = &rproc_virtio_config_ops,
        vdev->dev.parent = dev;
        vdev->dev.release = rproc_virtio_dev_release;

        /* Reference the vdev and vring allocations */
        get_device(dev);

        ret = register_virtio_device(vdev);
        if (ret) {
                put_device(&vdev->dev);
                dev_err(dev, "failed to register vdev: %d\n", ret);
                goto out;
        }

        dev_info(dev, "registered %s (type %d)\n", dev_name(&vdev->dev), id);

out:
        return ret;
}

/**
 * rproc_remove_virtio_dev() - remove an rproc-induced virtio device
 * @dev: the virtio device
 * @data: must be null
 *
 * This function unregisters an existing virtio device.
 *
 * Return: 0
 */
static int rproc_remove_virtio_dev(struct device *dev, void *data)
{
        struct virtio_device *vdev = dev_to_virtio(dev);

        unregister_virtio_device(vdev);
        return 0;
}

static int rproc_vdev_do_start(struct rproc_subdev *subdev)
{
        struct rproc_vdev *rvdev = container_of(subdev, struct rproc_vdev, subdev);

        return rproc_add_virtio_dev(rvdev, rvdev->id);
}

static void rproc_vdev_do_stop(struct rproc_subdev *subdev, bool crashed)
{
        struct rproc_vdev *rvdev = container_of(subdev, struct rproc_vdev, subdev);
        struct device *dev = &rvdev->pdev->dev;
        int ret;

        ret = device_for_each_child(dev, NULL, rproc_remove_virtio_dev);
        if (ret)
                dev_warn(dev, "can't remove vdev child device: %d\n", ret);
}

static int rproc_virtio_probe(struct platform_device *pdev)
{
        struct device *dev = &pdev->dev;
        struct rproc_vdev_data *rvdev_data = dev->platform_data;
        struct rproc_vdev *rvdev;
        struct rproc *rproc = container_of(dev->parent, struct rproc, dev);
        struct fw_rsc_vdev *rsc;
        int i, ret;

        if (!rvdev_data)
                return -EINVAL;

        rvdev = devm_kzalloc(dev, sizeof(*rvdev), GFP_KERNEL);
        if (!rvdev)
                return -ENOMEM;

        rvdev->id = rvdev_data->id;
        rvdev->rproc = rproc;
        rvdev->index = rvdev_data->index;

        ret = copy_dma_range_map(dev, rproc->dev.parent);
        if (ret)
                return ret;

        /* Make device dma capable by inheriting from parent's capabilities */
        set_dma_ops(dev, get_dma_ops(rproc->dev.parent));

        ret = dma_coerce_mask_and_coherent(dev, dma_get_mask(rproc->dev.parent));
        if (ret) {
                dev_warn(dev, "Failed to set DMA mask %llx. Trying to continue... (%pe)\n",
                         dma_get_mask(rproc->dev.parent), ERR_PTR(ret));
        }

        platform_set_drvdata(pdev, rvdev);
        rvdev->pdev = pdev;

        rsc = rvdev_data->rsc;

        /* parse the vrings */
        for (i = 0; i < rsc->num_of_vrings; i++) {
                ret = rproc_parse_vring(rvdev, rsc, i);
                if (ret)
                        return ret;
        }

        /* remember the resource offset*/
        rvdev->rsc_offset = rvdev_data->rsc_offset;

        /* allocate the vring resources */
        for (i = 0; i < rsc->num_of_vrings; i++) {
                ret = rproc_alloc_vring(rvdev, i);
                if (ret)
                        goto unwind_vring_allocations;
        }

        rproc_add_rvdev(rproc, rvdev);

        rvdev->subdev.start = rproc_vdev_do_start;
        rvdev->subdev.stop = rproc_vdev_do_stop;

        rproc_add_subdev(rproc, &rvdev->subdev);

        /*
         * We're indirectly making a non-temporary copy of the rproc pointer
         * here, because the platform device or the vdev device will indirectly
         * access the wrapping rproc.
         *
         * Therefore we must increment the rproc refcount here, and decrement
         * it _only_ on platform remove.
         */
        get_device(&rproc->dev);

        return 0;

unwind_vring_allocations:
        for (i--; i >= 0; i--)
                rproc_free_vring(&rvdev->vring[i]);

        return ret;
}

static void rproc_virtio_remove(struct platform_device *pdev)
{
        struct rproc_vdev *rvdev = dev_get_drvdata(&pdev->dev);
        struct rproc *rproc = rvdev->rproc;
        struct rproc_vring *rvring;
        int id;

        for (id = 0; id < ARRAY_SIZE(rvdev->vring); id++) {
                rvring = &rvdev->vring[id];
                rproc_free_vring(rvring);
        }

        rproc_remove_subdev(rproc, &rvdev->subdev);
        rproc_remove_rvdev(rvdev);

        put_device(&rproc->dev);
}

/* Platform driver */
static struct platform_driver rproc_virtio_driver = {
        .probe          = rproc_virtio_probe,
        .remove         = rproc_virtio_remove,
        .driver         = {
                .name   = "rproc-virtio",
        },
};
builtin_platform_driver(rproc_virtio_driver);