root/drivers/dma/st_fdma.c
// SPDX-License-Identifier: GPL-2.0-or-later
/*
 * DMA driver for STMicroelectronics STi FDMA controller
 *
 * Copyright (C) 2014 STMicroelectronics
 *
 * Author: Ludovic Barre <Ludovic.barre@st.com>
 *         Peter Griffin <peter.griffin@linaro.org>
 */

#include <linux/init.h>
#include <linux/module.h>
#include <linux/of.h>
#include <linux/of_dma.h>
#include <linux/platform_device.h>
#include <linux/property.h>
#include <linux/interrupt.h>
#include <linux/remoteproc.h>
#include <linux/slab.h>

#include "st_fdma.h"

static inline struct st_fdma_chan *to_st_fdma_chan(struct dma_chan *c)
{
        return container_of(c, struct st_fdma_chan, vchan.chan);
}

static struct st_fdma_desc *to_st_fdma_desc(struct virt_dma_desc *vd)
{
        return container_of(vd, struct st_fdma_desc, vdesc);
}

static int st_fdma_dreq_get(struct st_fdma_chan *fchan)
{
        struct st_fdma_dev *fdev = fchan->fdev;
        u32 req_line_cfg = fchan->cfg.req_line;
        u32 dreq_line;
        int try = 0;

        /*
         * dreq_mask is shared for n channels of fdma, so all accesses must be
         * atomic. if the dreq_mask is changed between ffz and set_bit,
         * we retry
         */
        do {
                if (fdev->dreq_mask == ~0L) {
                        dev_err(fdev->dev, "No req lines available\n");
                        return -EINVAL;
                }

                if (try || req_line_cfg >= ST_FDMA_NR_DREQS) {
                        dev_err(fdev->dev, "Invalid or used req line\n");
                        return -EINVAL;
                } else {
                        dreq_line = req_line_cfg;
                }

                try++;
        } while (test_and_set_bit(dreq_line, &fdev->dreq_mask));

        dev_dbg(fdev->dev, "get dreq_line:%d mask:%#lx\n",
                dreq_line, fdev->dreq_mask);

        return dreq_line;
}

static void st_fdma_dreq_put(struct st_fdma_chan *fchan)
{
        struct st_fdma_dev *fdev = fchan->fdev;

        dev_dbg(fdev->dev, "put dreq_line:%#lx\n", fchan->dreq_line);
        clear_bit(fchan->dreq_line, &fdev->dreq_mask);
}

static void st_fdma_xfer_desc(struct st_fdma_chan *fchan)
{
        struct virt_dma_desc *vdesc;
        unsigned long nbytes, ch_cmd, cmd;

        vdesc = vchan_next_desc(&fchan->vchan);
        if (!vdesc)
                return;

        fchan->fdesc = to_st_fdma_desc(vdesc);
        nbytes = fchan->fdesc->node[0].desc->nbytes;
        cmd = FDMA_CMD_START(fchan->vchan.chan.chan_id);
        ch_cmd = fchan->fdesc->node[0].pdesc | FDMA_CH_CMD_STA_START;

        /* start the channel for the descriptor */
        fnode_write(fchan, nbytes, FDMA_CNTN_OFST);
        fchan_write(fchan, ch_cmd, FDMA_CH_CMD_OFST);
        writel(cmd,
                fchan->fdev->slim_rproc->peri + FDMA_CMD_SET_OFST);

        dev_dbg(fchan->fdev->dev, "start chan:%d\n", fchan->vchan.chan.chan_id);
}

static void st_fdma_ch_sta_update(struct st_fdma_chan *fchan,
                                  unsigned long int_sta)
{
        unsigned long ch_sta, ch_err;
        int ch_id = fchan->vchan.chan.chan_id;
        struct st_fdma_dev *fdev = fchan->fdev;

        ch_sta = fchan_read(fchan, FDMA_CH_CMD_OFST);
        ch_err = ch_sta & FDMA_CH_CMD_ERR_MASK;
        ch_sta &= FDMA_CH_CMD_STA_MASK;

        if (int_sta & FDMA_INT_STA_ERR) {
                dev_warn(fdev->dev, "chan:%d, error:%ld\n", ch_id, ch_err);
                fchan->status = DMA_ERROR;
                return;
        }

        switch (ch_sta) {
        case FDMA_CH_CMD_STA_PAUSED:
                fchan->status = DMA_PAUSED;
                break;

        case FDMA_CH_CMD_STA_RUNNING:
                fchan->status = DMA_IN_PROGRESS;
                break;
        }
}

static irqreturn_t st_fdma_irq_handler(int irq, void *dev_id)
{
        struct st_fdma_dev *fdev = dev_id;
        irqreturn_t ret = IRQ_NONE;
        struct st_fdma_chan *fchan = &fdev->chans[0];
        unsigned long int_sta, clr;

        int_sta = fdma_read(fdev, FDMA_INT_STA_OFST);
        clr = int_sta;

        for (; int_sta != 0 ; int_sta >>= 2, fchan++) {
                if (!(int_sta & (FDMA_INT_STA_CH | FDMA_INT_STA_ERR)))
                        continue;

                spin_lock(&fchan->vchan.lock);
                st_fdma_ch_sta_update(fchan, int_sta);

                if (fchan->fdesc) {
                        if (!fchan->fdesc->iscyclic) {
                                list_del(&fchan->fdesc->vdesc.node);
                                vchan_cookie_complete(&fchan->fdesc->vdesc);
                                fchan->fdesc = NULL;
                                fchan->status = DMA_COMPLETE;
                        } else {
                                vchan_cyclic_callback(&fchan->fdesc->vdesc);
                        }

                        /* Start the next descriptor (if available) */
                        if (!fchan->fdesc)
                                st_fdma_xfer_desc(fchan);
                }

                spin_unlock(&fchan->vchan.lock);
                ret = IRQ_HANDLED;
        }

        fdma_write(fdev, clr, FDMA_INT_CLR_OFST);

        return ret;
}

static struct dma_chan *st_fdma_of_xlate(struct of_phandle_args *dma_spec,
                                         struct of_dma *ofdma)
{
        struct st_fdma_dev *fdev = ofdma->of_dma_data;
        struct dma_chan *chan;
        struct st_fdma_chan *fchan;
        int ret;

        if (dma_spec->args_count < 1)
                return ERR_PTR(-EINVAL);

        if (fdev->dma_device.dev->of_node != dma_spec->np)
                return ERR_PTR(-EINVAL);

        ret = rproc_boot(fdev->slim_rproc->rproc);
        if (ret == -ENOENT)
                return ERR_PTR(-EPROBE_DEFER);
        else if (ret)
                return ERR_PTR(ret);

        chan = dma_get_any_slave_channel(&fdev->dma_device);
        if (!chan)
                goto err_chan;

        fchan = to_st_fdma_chan(chan);

        fchan->cfg.of_node = dma_spec->np;
        fchan->cfg.req_line = dma_spec->args[0];
        fchan->cfg.req_ctrl = 0;
        fchan->cfg.type = ST_FDMA_TYPE_FREE_RUN;

        if (dma_spec->args_count > 1)
                fchan->cfg.req_ctrl = dma_spec->args[1]
                        & FDMA_REQ_CTRL_CFG_MASK;

        if (dma_spec->args_count > 2)
                fchan->cfg.type = dma_spec->args[2];

        if (fchan->cfg.type == ST_FDMA_TYPE_FREE_RUN) {
                fchan->dreq_line = 0;
        } else {
                fchan->dreq_line = st_fdma_dreq_get(fchan);
                if (IS_ERR_VALUE(fchan->dreq_line)) {
                        chan = ERR_PTR(fchan->dreq_line);
                        goto err_chan;
                }
        }

        dev_dbg(fdev->dev, "xlate req_line:%d type:%d req_ctrl:%#lx\n",
                fchan->cfg.req_line, fchan->cfg.type, fchan->cfg.req_ctrl);

        return chan;

err_chan:
        rproc_shutdown(fdev->slim_rproc->rproc);
        return chan;

}

static void st_fdma_free_desc(struct virt_dma_desc *vdesc)
{
        struct st_fdma_desc *fdesc;
        int i;

        fdesc = to_st_fdma_desc(vdesc);
        for (i = 0; i < fdesc->n_nodes; i++)
                dma_pool_free(fdesc->fchan->node_pool, fdesc->node[i].desc,
                              fdesc->node[i].pdesc);
        kfree(fdesc);
}

static struct st_fdma_desc *st_fdma_alloc_desc(struct st_fdma_chan *fchan,
                                               int sg_len)
{
        struct st_fdma_desc *fdesc;
        int i;

        fdesc = kzalloc_flex(*fdesc, node, sg_len, GFP_NOWAIT);
        if (!fdesc)
                return NULL;

        fdesc->fchan = fchan;
        fdesc->n_nodes = sg_len;
        for (i = 0; i < sg_len; i++) {
                fdesc->node[i].desc = dma_pool_alloc(fchan->node_pool,
                                GFP_NOWAIT, &fdesc->node[i].pdesc);
                if (!fdesc->node[i].desc)
                        goto err;
        }
        return fdesc;

err:
        while (--i >= 0)
                dma_pool_free(fchan->node_pool, fdesc->node[i].desc,
                              fdesc->node[i].pdesc);
        kfree(fdesc);
        return NULL;
}

static int st_fdma_alloc_chan_res(struct dma_chan *chan)
{
        struct st_fdma_chan *fchan = to_st_fdma_chan(chan);

        /* Create the dma pool for descriptor allocation */
        fchan->node_pool = dma_pool_create(dev_name(&chan->dev->device),
                                            fchan->fdev->dev,
                                            sizeof(struct st_fdma_hw_node),
                                            __alignof__(struct st_fdma_hw_node),
                                            0);

        if (!fchan->node_pool) {
                dev_err(fchan->fdev->dev, "unable to allocate desc pool\n");
                return -ENOMEM;
        }

        dev_dbg(fchan->fdev->dev, "alloc ch_id:%d type:%d\n",
                fchan->vchan.chan.chan_id, fchan->cfg.type);

        return 0;
}

static void st_fdma_free_chan_res(struct dma_chan *chan)
{
        struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
        struct rproc *rproc = fchan->fdev->slim_rproc->rproc;
        unsigned long flags;

        dev_dbg(fchan->fdev->dev, "%s: freeing chan:%d\n",
                __func__, fchan->vchan.chan.chan_id);

        if (fchan->cfg.type != ST_FDMA_TYPE_FREE_RUN)
                st_fdma_dreq_put(fchan);

        spin_lock_irqsave(&fchan->vchan.lock, flags);
        fchan->fdesc = NULL;
        spin_unlock_irqrestore(&fchan->vchan.lock, flags);

        dma_pool_destroy(fchan->node_pool);
        fchan->node_pool = NULL;
        memset(&fchan->cfg, 0, sizeof(struct st_fdma_cfg));

        rproc_shutdown(rproc);
}

static struct dma_async_tx_descriptor *st_fdma_prep_dma_memcpy(
        struct dma_chan *chan,  dma_addr_t dst, dma_addr_t src,
        size_t len, unsigned long flags)
{
        struct st_fdma_chan *fchan;
        struct st_fdma_desc *fdesc;
        struct st_fdma_hw_node *hw_node;

        if (!len)
                return NULL;

        fchan = to_st_fdma_chan(chan);

        /* We only require a single descriptor */
        fdesc = st_fdma_alloc_desc(fchan, 1);
        if (!fdesc) {
                dev_err(fchan->fdev->dev, "no memory for desc\n");
                return NULL;
        }

        hw_node = fdesc->node[0].desc;
        hw_node->next = 0;
        hw_node->control = FDMA_NODE_CTRL_REQ_MAP_FREE_RUN;
        hw_node->control |= FDMA_NODE_CTRL_SRC_INCR;
        hw_node->control |= FDMA_NODE_CTRL_DST_INCR;
        hw_node->control |= FDMA_NODE_CTRL_INT_EON;
        hw_node->nbytes = len;
        hw_node->saddr = src;
        hw_node->daddr = dst;
        hw_node->generic.length = len;
        hw_node->generic.sstride = 0;
        hw_node->generic.dstride = 0;

        return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags);
}

static int config_reqctrl(struct st_fdma_chan *fchan,
                          enum dma_transfer_direction direction)
{
        u32 maxburst = 0, addr = 0;
        enum dma_slave_buswidth width;
        int ch_id = fchan->vchan.chan.chan_id;
        struct st_fdma_dev *fdev = fchan->fdev;

        switch (direction) {

        case DMA_DEV_TO_MEM:
                fchan->cfg.req_ctrl &= ~FDMA_REQ_CTRL_WNR;
                maxburst = fchan->scfg.src_maxburst;
                width = fchan->scfg.src_addr_width;
                addr = fchan->scfg.src_addr;
                break;

        case DMA_MEM_TO_DEV:
                fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_WNR;
                maxburst = fchan->scfg.dst_maxburst;
                width = fchan->scfg.dst_addr_width;
                addr = fchan->scfg.dst_addr;
                break;

        default:
                return -EINVAL;
        }

        fchan->cfg.req_ctrl &= ~FDMA_REQ_CTRL_OPCODE_MASK;

        switch (width) {

        case DMA_SLAVE_BUSWIDTH_1_BYTE:
                fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_OPCODE_LD_ST1;
                break;

        case DMA_SLAVE_BUSWIDTH_2_BYTES:
                fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_OPCODE_LD_ST2;
                break;

        case DMA_SLAVE_BUSWIDTH_4_BYTES:
                fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_OPCODE_LD_ST4;
                break;

        case DMA_SLAVE_BUSWIDTH_8_BYTES:
                fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_OPCODE_LD_ST8;
                break;

        default:
                return -EINVAL;
        }

        fchan->cfg.req_ctrl &= ~FDMA_REQ_CTRL_NUM_OPS_MASK;
        fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_NUM_OPS(maxburst-1);
        dreq_write(fchan, fchan->cfg.req_ctrl, FDMA_REQ_CTRL_OFST);

        fchan->cfg.dev_addr = addr;
        fchan->cfg.dir = direction;

        dev_dbg(fdev->dev, "chan:%d config_reqctrl:%#x req_ctrl:%#lx\n",
                ch_id, addr, fchan->cfg.req_ctrl);

        return 0;
}

static void fill_hw_node(struct st_fdma_hw_node *hw_node,
                        struct st_fdma_chan *fchan,
                        enum dma_transfer_direction direction)
{
        if (direction == DMA_MEM_TO_DEV) {
                hw_node->control |= FDMA_NODE_CTRL_SRC_INCR;
                hw_node->control |= FDMA_NODE_CTRL_DST_STATIC;
                hw_node->daddr = fchan->cfg.dev_addr;
        } else {
                hw_node->control |= FDMA_NODE_CTRL_SRC_STATIC;
                hw_node->control |= FDMA_NODE_CTRL_DST_INCR;
                hw_node->saddr = fchan->cfg.dev_addr;
        }

        hw_node->generic.sstride = 0;
        hw_node->generic.dstride = 0;
}

static inline struct st_fdma_chan *st_fdma_prep_common(struct dma_chan *chan,
                size_t len, enum dma_transfer_direction direction)
{
        struct st_fdma_chan *fchan;

        if (!chan || !len)
                return NULL;

        fchan = to_st_fdma_chan(chan);

        if (!is_slave_direction(direction)) {
                dev_err(fchan->fdev->dev, "bad direction?\n");
                return NULL;
        }

        return fchan;
}

static struct dma_async_tx_descriptor *st_fdma_prep_dma_cyclic(
                struct dma_chan *chan, dma_addr_t buf_addr, size_t len,
                size_t period_len, enum dma_transfer_direction direction,
                unsigned long flags)
{
        struct st_fdma_chan *fchan;
        struct st_fdma_desc *fdesc;
        int sg_len, i;

        fchan = st_fdma_prep_common(chan, len, direction);
        if (!fchan)
                return NULL;

        if (!period_len)
                return NULL;

        if (config_reqctrl(fchan, direction)) {
                dev_err(fchan->fdev->dev, "bad width or direction\n");
                return NULL;
        }

        /* the buffer length must be a multiple of period_len */
        if (len % period_len != 0) {
                dev_err(fchan->fdev->dev, "len is not multiple of period\n");
                return NULL;
        }

        sg_len = len / period_len;
        fdesc = st_fdma_alloc_desc(fchan, sg_len);
        if (!fdesc) {
                dev_err(fchan->fdev->dev, "no memory for desc\n");
                return NULL;
        }

        fdesc->iscyclic = true;

        for (i = 0; i < sg_len; i++) {
                struct st_fdma_hw_node *hw_node = fdesc->node[i].desc;

                hw_node->next = fdesc->node[(i + 1) % sg_len].pdesc;

                hw_node->control =
                        FDMA_NODE_CTRL_REQ_MAP_DREQ(fchan->dreq_line);
                hw_node->control |= FDMA_NODE_CTRL_INT_EON;

                fill_hw_node(hw_node, fchan, direction);

                if (direction == DMA_MEM_TO_DEV)
                        hw_node->saddr = buf_addr + (i * period_len);
                else
                        hw_node->daddr = buf_addr + (i * period_len);

                hw_node->nbytes = period_len;
                hw_node->generic.length = period_len;
        }

        return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags);
}

static struct dma_async_tx_descriptor *st_fdma_prep_slave_sg(
                struct dma_chan *chan, struct scatterlist *sgl,
                unsigned int sg_len, enum dma_transfer_direction direction,
                unsigned long flags, void *context)
{
        struct st_fdma_chan *fchan;
        struct st_fdma_desc *fdesc;
        struct st_fdma_hw_node *hw_node;
        struct scatterlist *sg;
        int i;

        fchan = st_fdma_prep_common(chan, sg_len, direction);
        if (!fchan)
                return NULL;

        if (!sgl)
                return NULL;

        fdesc = st_fdma_alloc_desc(fchan, sg_len);
        if (!fdesc) {
                dev_err(fchan->fdev->dev, "no memory for desc\n");
                return NULL;
        }

        fdesc->iscyclic = false;

        for_each_sg(sgl, sg, sg_len, i) {
                hw_node = fdesc->node[i].desc;

                hw_node->next = fdesc->node[(i + 1) % sg_len].pdesc;
                hw_node->control = FDMA_NODE_CTRL_REQ_MAP_DREQ(fchan->dreq_line);

                fill_hw_node(hw_node, fchan, direction);

                if (direction == DMA_MEM_TO_DEV)
                        hw_node->saddr = sg_dma_address(sg);
                else
                        hw_node->daddr = sg_dma_address(sg);

                hw_node->nbytes = sg_dma_len(sg);
                hw_node->generic.length = sg_dma_len(sg);
        }

        /* interrupt at end of last node */
        hw_node->control |= FDMA_NODE_CTRL_INT_EON;

        return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags);
}

static size_t st_fdma_desc_residue(struct st_fdma_chan *fchan,
                                   struct virt_dma_desc *vdesc,
                                   bool in_progress)
{
        struct st_fdma_desc *fdesc = fchan->fdesc;
        size_t residue = 0;
        dma_addr_t cur_addr = 0;
        int i;

        if (in_progress) {
                cur_addr = fchan_read(fchan, FDMA_CH_CMD_OFST);
                cur_addr &= FDMA_CH_CMD_DATA_MASK;
        }

        for (i = fchan->fdesc->n_nodes - 1 ; i >= 0; i--) {
                if (cur_addr == fdesc->node[i].pdesc) {
                        residue += fnode_read(fchan, FDMA_CNTN_OFST);
                        break;
                }
                residue += fdesc->node[i].desc->nbytes;
        }

        return residue;
}

static enum dma_status st_fdma_tx_status(struct dma_chan *chan,
                                         dma_cookie_t cookie,
                                         struct dma_tx_state *txstate)
{
        struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
        struct virt_dma_desc *vd;
        enum dma_status ret;
        unsigned long flags;

        ret = dma_cookie_status(chan, cookie, txstate);
        if (ret == DMA_COMPLETE || !txstate)
                return ret;

        spin_lock_irqsave(&fchan->vchan.lock, flags);
        vd = vchan_find_desc(&fchan->vchan, cookie);
        if (fchan->fdesc && cookie == fchan->fdesc->vdesc.tx.cookie)
                txstate->residue = st_fdma_desc_residue(fchan, vd, true);
        else if (vd)
                txstate->residue = st_fdma_desc_residue(fchan, vd, false);
        else
                txstate->residue = 0;

        spin_unlock_irqrestore(&fchan->vchan.lock, flags);

        return ret;
}

static void st_fdma_issue_pending(struct dma_chan *chan)
{
        struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
        unsigned long flags;

        spin_lock_irqsave(&fchan->vchan.lock, flags);

        if (vchan_issue_pending(&fchan->vchan) && !fchan->fdesc)
                st_fdma_xfer_desc(fchan);

        spin_unlock_irqrestore(&fchan->vchan.lock, flags);
}

static int st_fdma_pause(struct dma_chan *chan)
{
        unsigned long flags;
        struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
        int ch_id = fchan->vchan.chan.chan_id;
        unsigned long cmd = FDMA_CMD_PAUSE(ch_id);

        dev_dbg(fchan->fdev->dev, "pause chan:%d\n", ch_id);

        spin_lock_irqsave(&fchan->vchan.lock, flags);
        if (fchan->fdesc)
                fdma_write(fchan->fdev, cmd, FDMA_CMD_SET_OFST);
        spin_unlock_irqrestore(&fchan->vchan.lock, flags);

        return 0;
}

static int st_fdma_resume(struct dma_chan *chan)
{
        unsigned long flags;
        unsigned long val;
        struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
        int ch_id = fchan->vchan.chan.chan_id;

        dev_dbg(fchan->fdev->dev, "resume chan:%d\n", ch_id);

        spin_lock_irqsave(&fchan->vchan.lock, flags);
        if (fchan->fdesc) {
                val = fchan_read(fchan, FDMA_CH_CMD_OFST);
                val &= FDMA_CH_CMD_DATA_MASK;
                fchan_write(fchan, val, FDMA_CH_CMD_OFST);
        }
        spin_unlock_irqrestore(&fchan->vchan.lock, flags);

        return 0;
}

static int st_fdma_terminate_all(struct dma_chan *chan)
{
        unsigned long flags;
        LIST_HEAD(head);
        struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
        int ch_id = fchan->vchan.chan.chan_id;
        unsigned long cmd = FDMA_CMD_PAUSE(ch_id);

        dev_dbg(fchan->fdev->dev, "terminate chan:%d\n", ch_id);

        spin_lock_irqsave(&fchan->vchan.lock, flags);
        fdma_write(fchan->fdev, cmd, FDMA_CMD_SET_OFST);
        fchan->fdesc = NULL;
        vchan_get_all_descriptors(&fchan->vchan, &head);
        spin_unlock_irqrestore(&fchan->vchan.lock, flags);
        vchan_dma_desc_free_list(&fchan->vchan, &head);

        return 0;
}

static int st_fdma_slave_config(struct dma_chan *chan,
                                struct dma_slave_config *slave_cfg)
{
        struct st_fdma_chan *fchan = to_st_fdma_chan(chan);

        memcpy(&fchan->scfg, slave_cfg, sizeof(fchan->scfg));
        return 0;
}

static const struct st_fdma_driverdata fdma_mpe31_stih407_11 = {
        .name = "STiH407",
        .id = 0,
};

static const struct st_fdma_driverdata fdma_mpe31_stih407_12 = {
        .name = "STiH407",
        .id = 1,
};

static const struct st_fdma_driverdata fdma_mpe31_stih407_13 = {
        .name = "STiH407",
        .id = 2,
};

static const struct of_device_id st_fdma_match[] = {
        { .compatible = "st,stih407-fdma-mpe31-11"
          , .data = &fdma_mpe31_stih407_11 },
        { .compatible = "st,stih407-fdma-mpe31-12"
          , .data = &fdma_mpe31_stih407_12 },
        { .compatible = "st,stih407-fdma-mpe31-13"
          , .data = &fdma_mpe31_stih407_13 },
        {},
};
MODULE_DEVICE_TABLE(of, st_fdma_match);

static int st_fdma_parse_dt(struct platform_device *pdev,
                        const struct st_fdma_driverdata *drvdata,
                        struct st_fdma_dev *fdev)
{
        snprintf(fdev->fw_name, FW_NAME_SIZE, "fdma_%s_%d.elf",
                drvdata->name, drvdata->id);

        return of_property_read_u32(pdev->dev.of_node, "dma-channels",
                                    &fdev->nr_channels);
}
#define FDMA_DMA_BUSWIDTHS      (BIT(DMA_SLAVE_BUSWIDTH_1_BYTE) | \
                                 BIT(DMA_SLAVE_BUSWIDTH_2_BYTES) | \
                                 BIT(DMA_SLAVE_BUSWIDTH_3_BYTES) | \
                                 BIT(DMA_SLAVE_BUSWIDTH_4_BYTES))

static void st_fdma_free(struct st_fdma_dev *fdev)
{
        struct st_fdma_chan *fchan;
        int i;

        for (i = 0; i < fdev->nr_channels; i++) {
                fchan = &fdev->chans[i];
                list_del(&fchan->vchan.chan.device_node);
                tasklet_kill(&fchan->vchan.task);
        }
}

static int st_fdma_probe(struct platform_device *pdev)
{
        struct st_fdma_dev *fdev;
        struct device_node *np = pdev->dev.of_node;
        const struct st_fdma_driverdata *drvdata;
        int ret, i;

        drvdata = device_get_match_data(&pdev->dev);

        fdev = devm_kzalloc(&pdev->dev, sizeof(*fdev), GFP_KERNEL);
        if (!fdev)
                return -ENOMEM;

        ret = st_fdma_parse_dt(pdev, drvdata, fdev);
        if (ret) {
                dev_err(&pdev->dev, "unable to find platform data\n");
                goto err;
        }

        fdev->chans = devm_kcalloc(&pdev->dev, fdev->nr_channels,
                                   sizeof(struct st_fdma_chan), GFP_KERNEL);
        if (!fdev->chans)
                return -ENOMEM;

        fdev->dev = &pdev->dev;
        fdev->drvdata = drvdata;
        platform_set_drvdata(pdev, fdev);

        fdev->irq = platform_get_irq(pdev, 0);
        if (fdev->irq < 0)
                return -EINVAL;

        ret = devm_request_irq(&pdev->dev, fdev->irq, st_fdma_irq_handler, 0,
                               dev_name(&pdev->dev), fdev);
        if (ret) {
                dev_err(&pdev->dev, "Failed to request irq (%d)\n", ret);
                goto err;
        }

        fdev->slim_rproc = st_slim_rproc_alloc(pdev, fdev->fw_name);
        if (IS_ERR(fdev->slim_rproc)) {
                ret = PTR_ERR(fdev->slim_rproc);
                dev_err(&pdev->dev, "slim_rproc_alloc failed (%d)\n", ret);
                goto err;
        }

        /* Initialise list of FDMA channels */
        INIT_LIST_HEAD(&fdev->dma_device.channels);
        for (i = 0; i < fdev->nr_channels; i++) {
                struct st_fdma_chan *fchan = &fdev->chans[i];

                fchan->fdev = fdev;
                fchan->vchan.desc_free = st_fdma_free_desc;
                vchan_init(&fchan->vchan, &fdev->dma_device);
        }

        /* Initialise the FDMA dreq (reserve 0 & 31 for FDMA use) */
        fdev->dreq_mask = BIT(0) | BIT(31);

        dma_cap_set(DMA_SLAVE, fdev->dma_device.cap_mask);
        dma_cap_set(DMA_CYCLIC, fdev->dma_device.cap_mask);
        dma_cap_set(DMA_MEMCPY, fdev->dma_device.cap_mask);

        fdev->dma_device.dev = &pdev->dev;
        fdev->dma_device.device_alloc_chan_resources = st_fdma_alloc_chan_res;
        fdev->dma_device.device_free_chan_resources = st_fdma_free_chan_res;
        fdev->dma_device.device_prep_dma_cyclic = st_fdma_prep_dma_cyclic;
        fdev->dma_device.device_prep_slave_sg = st_fdma_prep_slave_sg;
        fdev->dma_device.device_prep_dma_memcpy = st_fdma_prep_dma_memcpy;
        fdev->dma_device.device_tx_status = st_fdma_tx_status;
        fdev->dma_device.device_issue_pending = st_fdma_issue_pending;
        fdev->dma_device.device_terminate_all = st_fdma_terminate_all;
        fdev->dma_device.device_config = st_fdma_slave_config;
        fdev->dma_device.device_pause = st_fdma_pause;
        fdev->dma_device.device_resume = st_fdma_resume;

        fdev->dma_device.src_addr_widths = FDMA_DMA_BUSWIDTHS;
        fdev->dma_device.dst_addr_widths = FDMA_DMA_BUSWIDTHS;
        fdev->dma_device.directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV);
        fdev->dma_device.residue_granularity = DMA_RESIDUE_GRANULARITY_BURST;

        ret = dmaenginem_async_device_register(&fdev->dma_device);
        if (ret) {
                dev_err(&pdev->dev,
                        "Failed to register DMA device (%d)\n", ret);
                goto err_rproc;
        }

        ret = of_dma_controller_register(np, st_fdma_of_xlate, fdev);
        if (ret) {
                dev_err(&pdev->dev,
                        "Failed to register controller (%d)\n", ret);
                goto err_rproc;
        }

        dev_info(&pdev->dev, "ST FDMA engine driver, irq:%d\n", fdev->irq);

        return 0;

err_rproc:
        st_fdma_free(fdev);
        st_slim_rproc_put(fdev->slim_rproc);
err:
        return ret;
}

static void st_fdma_remove(struct platform_device *pdev)
{
        struct st_fdma_dev *fdev = platform_get_drvdata(pdev);

        devm_free_irq(&pdev->dev, fdev->irq, fdev);
        st_slim_rproc_put(fdev->slim_rproc);
        of_dma_controller_free(pdev->dev.of_node);
}

static struct platform_driver st_fdma_platform_driver = {
        .driver = {
                .name = DRIVER_NAME,
                .of_match_table = st_fdma_match,
        },
        .probe = st_fdma_probe,
        .remove = st_fdma_remove,
};
module_platform_driver(st_fdma_platform_driver);

MODULE_LICENSE("GPL v2");
MODULE_DESCRIPTION("STMicroelectronics FDMA engine driver");
MODULE_AUTHOR("Ludovic.barre <Ludovic.barre@st.com>");
MODULE_AUTHOR("Peter Griffin <peter.griffin@linaro.org>");