root/drivers/spi/spi-uniphier.c
// SPDX-License-Identifier: GPL-2.0
// spi-uniphier.c - Socionext UniPhier SPI controller driver
// Copyright 2012      Panasonic Corporation
// Copyright 2016-2018 Socionext Inc.

#include <linux/kernel.h>
#include <linux/bitfield.h>
#include <linux/bitops.h>
#include <linux/clk.h>
#include <linux/delay.h>
#include <linux/dmaengine.h>
#include <linux/interrupt.h>
#include <linux/io.h>
#include <linux/module.h>
#include <linux/platform_device.h>
#include <linux/spi/spi.h>

#include <linux/unaligned.h>

#define SSI_TIMEOUT_MS          2000
#define SSI_POLL_TIMEOUT_US     200
#define SSI_MAX_CLK_DIVIDER     254
#define SSI_MIN_CLK_DIVIDER     4

struct uniphier_spi_priv {
        void __iomem *base;
        dma_addr_t base_dma_addr;
        struct clk *clk;
        struct spi_controller *host;
        struct completion xfer_done;

        int error;
        unsigned int tx_bytes;
        unsigned int rx_bytes;
        const u8 *tx_buf;
        u8 *rx_buf;
        atomic_t dma_busy;

        bool is_save_param;
        u8 bits_per_word;
        u16 mode;
        u32 speed_hz;
};

#define SSI_CTL                 0x00
#define   SSI_CTL_EN            BIT(0)

#define SSI_CKS                 0x04
#define   SSI_CKS_CKRAT_MASK    GENMASK(7, 0)
#define   SSI_CKS_CKPHS         BIT(14)
#define   SSI_CKS_CKINIT        BIT(13)
#define   SSI_CKS_CKDLY         BIT(12)

#define SSI_TXWDS               0x08
#define   SSI_TXWDS_WDLEN_MASK  GENMASK(13, 8)
#define   SSI_TXWDS_TDTF_MASK   GENMASK(7, 6)
#define   SSI_TXWDS_DTLEN_MASK  GENMASK(5, 0)

#define SSI_RXWDS               0x0c
#define   SSI_RXWDS_DTLEN_MASK  GENMASK(5, 0)

#define SSI_FPS                 0x10
#define   SSI_FPS_FSPOL         BIT(15)
#define   SSI_FPS_FSTRT         BIT(14)

#define SSI_SR                  0x14
#define   SSI_SR_BUSY           BIT(7)
#define   SSI_SR_RNE            BIT(0)

#define SSI_IE                  0x18
#define   SSI_IE_TCIE           BIT(4)
#define   SSI_IE_RCIE           BIT(3)
#define   SSI_IE_TXRE           BIT(2)
#define   SSI_IE_RXRE           BIT(1)
#define   SSI_IE_RORIE          BIT(0)
#define   SSI_IE_ALL_MASK       GENMASK(4, 0)

#define SSI_IS                  0x1c
#define   SSI_IS_RXRS           BIT(9)
#define   SSI_IS_RCID           BIT(3)
#define   SSI_IS_RORID          BIT(0)

#define SSI_IC                  0x1c
#define   SSI_IC_TCIC           BIT(4)
#define   SSI_IC_RCIC           BIT(3)
#define   SSI_IC_RORIC          BIT(0)

#define SSI_FC                  0x20
#define   SSI_FC_TXFFL          BIT(12)
#define   SSI_FC_TXFTH_MASK     GENMASK(11, 8)
#define   SSI_FC_RXFFL          BIT(4)
#define   SSI_FC_RXFTH_MASK     GENMASK(3, 0)

#define SSI_TXDR                0x24
#define SSI_RXDR                0x24

#define SSI_FIFO_DEPTH          8U
#define SSI_FIFO_BURST_NUM      1

#define SSI_DMA_RX_BUSY         BIT(1)
#define SSI_DMA_TX_BUSY         BIT(0)

static inline unsigned int bytes_per_word(unsigned int bits)
{
        return bits <= 8 ? 1 : (bits <= 16 ? 2 : 4);
}

static inline void uniphier_spi_irq_enable(struct uniphier_spi_priv *priv,
                                           u32 mask)
{
        u32 val;

        val = readl(priv->base + SSI_IE);
        val |= mask;
        writel(val, priv->base + SSI_IE);
}

static inline void uniphier_spi_irq_disable(struct uniphier_spi_priv *priv,
                                            u32 mask)
{
        u32 val;

        val = readl(priv->base + SSI_IE);
        val &= ~mask;
        writel(val, priv->base + SSI_IE);
}

static void uniphier_spi_set_mode(struct spi_device *spi)
{
        struct uniphier_spi_priv *priv = spi_controller_get_devdata(spi->controller);
        u32 val1, val2;

        /*
         * clock setting
         * CKPHS    capture timing. 0:rising edge, 1:falling edge
         * CKINIT   clock initial level. 0:low, 1:high
         * CKDLY    clock delay. 0:no delay, 1:delay depending on FSTRT
         *          (FSTRT=0: 1 clock, FSTRT=1: 0.5 clock)
         *
         * frame setting
         * FSPOL    frame signal porarity. 0: low, 1: high
         * FSTRT    start frame timing
         *          0: rising edge of clock, 1: falling edge of clock
         */
        switch (spi->mode & SPI_MODE_X_MASK) {
        case SPI_MODE_0:
                /* CKPHS=1, CKINIT=0, CKDLY=1, FSTRT=0 */
                val1 = SSI_CKS_CKPHS | SSI_CKS_CKDLY;
                val2 = 0;
                break;
        case SPI_MODE_1:
                /* CKPHS=0, CKINIT=0, CKDLY=0, FSTRT=1 */
                val1 = 0;
                val2 = SSI_FPS_FSTRT;
                break;
        case SPI_MODE_2:
                /* CKPHS=0, CKINIT=1, CKDLY=1, FSTRT=1 */
                val1 = SSI_CKS_CKINIT | SSI_CKS_CKDLY;
                val2 = SSI_FPS_FSTRT;
                break;
        case SPI_MODE_3:
                /* CKPHS=1, CKINIT=1, CKDLY=0, FSTRT=0 */
                val1 = SSI_CKS_CKPHS | SSI_CKS_CKINIT;
                val2 = 0;
                break;
        }

        if (!(spi->mode & SPI_CS_HIGH))
                val2 |= SSI_FPS_FSPOL;

        writel(val1, priv->base + SSI_CKS);
        writel(val2, priv->base + SSI_FPS);

        val1 = 0;
        if (spi->mode & SPI_LSB_FIRST)
                val1 |= FIELD_PREP(SSI_TXWDS_TDTF_MASK, 1);
        writel(val1, priv->base + SSI_TXWDS);
        writel(val1, priv->base + SSI_RXWDS);
}

static void uniphier_spi_set_transfer_size(struct spi_device *spi, int size)
{
        struct uniphier_spi_priv *priv = spi_controller_get_devdata(spi->controller);
        u32 val;

        val = readl(priv->base + SSI_TXWDS);
        val &= ~(SSI_TXWDS_WDLEN_MASK | SSI_TXWDS_DTLEN_MASK);
        val |= FIELD_PREP(SSI_TXWDS_WDLEN_MASK, size);
        val |= FIELD_PREP(SSI_TXWDS_DTLEN_MASK, size);
        writel(val, priv->base + SSI_TXWDS);

        val = readl(priv->base + SSI_RXWDS);
        val &= ~SSI_RXWDS_DTLEN_MASK;
        val |= FIELD_PREP(SSI_RXWDS_DTLEN_MASK, size);
        writel(val, priv->base + SSI_RXWDS);
}

static void uniphier_spi_set_baudrate(struct spi_device *spi,
                                      unsigned int speed)
{
        struct uniphier_spi_priv *priv = spi_controller_get_devdata(spi->controller);
        u32 val, ckdiv;

        /*
         * the supported rates are even numbers from 4 to 254. (4,6,8...254)
         * round up as we look for equal or less speed
         */
        ckdiv = DIV_ROUND_UP(clk_get_rate(priv->clk), speed);
        ckdiv = round_up(ckdiv, 2);

        val = readl(priv->base + SSI_CKS);
        val &= ~SSI_CKS_CKRAT_MASK;
        val |= ckdiv & SSI_CKS_CKRAT_MASK;
        writel(val, priv->base + SSI_CKS);
}

static void uniphier_spi_setup_transfer(struct spi_device *spi,
                                       struct spi_transfer *t)
{
        struct uniphier_spi_priv *priv = spi_controller_get_devdata(spi->controller);
        u32 val;

        priv->error = 0;
        priv->tx_buf = t->tx_buf;
        priv->rx_buf = t->rx_buf;
        priv->tx_bytes = priv->rx_bytes = t->len;

        if (!priv->is_save_param || priv->mode != spi->mode) {
                uniphier_spi_set_mode(spi);
                priv->mode = spi->mode;
                priv->is_save_param = false;
        }

        if (!priv->is_save_param || priv->bits_per_word != t->bits_per_word) {
                uniphier_spi_set_transfer_size(spi, t->bits_per_word);
                priv->bits_per_word = t->bits_per_word;
        }

        if (!priv->is_save_param || priv->speed_hz != t->speed_hz) {
                uniphier_spi_set_baudrate(spi, t->speed_hz);
                priv->speed_hz = t->speed_hz;
        }

        priv->is_save_param = true;

        /* reset FIFOs */
        val = SSI_FC_TXFFL | SSI_FC_RXFFL;
        writel(val, priv->base + SSI_FC);
}

static void uniphier_spi_send(struct uniphier_spi_priv *priv)
{
        int wsize;
        u32 val = 0;

        wsize = min(bytes_per_word(priv->bits_per_word), priv->tx_bytes);
        priv->tx_bytes -= wsize;

        if (priv->tx_buf) {
                switch (wsize) {
                case 1:
                        val = *priv->tx_buf;
                        break;
                case 2:
                        val = get_unaligned_le16(priv->tx_buf);
                        break;
                case 4:
                        val = get_unaligned_le32(priv->tx_buf);
                        break;
                }

                priv->tx_buf += wsize;
        }

        writel(val, priv->base + SSI_TXDR);
}

static void uniphier_spi_recv(struct uniphier_spi_priv *priv)
{
        int rsize;
        u32 val;

        rsize = min(bytes_per_word(priv->bits_per_word), priv->rx_bytes);
        priv->rx_bytes -= rsize;

        val = readl(priv->base + SSI_RXDR);

        if (priv->rx_buf) {
                switch (rsize) {
                case 1:
                        *priv->rx_buf = val;
                        break;
                case 2:
                        put_unaligned_le16(val, priv->rx_buf);
                        break;
                case 4:
                        put_unaligned_le32(val, priv->rx_buf);
                        break;
                }

                priv->rx_buf += rsize;
        }
}

static void uniphier_spi_set_fifo_threshold(struct uniphier_spi_priv *priv,
                                            unsigned int threshold)
{
        u32 val;

        val = readl(priv->base + SSI_FC);
        val &= ~(SSI_FC_TXFTH_MASK | SSI_FC_RXFTH_MASK);
        val |= FIELD_PREP(SSI_FC_TXFTH_MASK, SSI_FIFO_DEPTH - threshold);
        val |= FIELD_PREP(SSI_FC_RXFTH_MASK, threshold);
        writel(val, priv->base + SSI_FC);
}

static void uniphier_spi_fill_tx_fifo(struct uniphier_spi_priv *priv)
{
        unsigned int fifo_threshold, fill_words;
        unsigned int bpw = bytes_per_word(priv->bits_per_word);

        fifo_threshold = DIV_ROUND_UP(priv->rx_bytes, bpw);
        fifo_threshold = min(fifo_threshold, SSI_FIFO_DEPTH);

        uniphier_spi_set_fifo_threshold(priv, fifo_threshold);

        fill_words = fifo_threshold -
                DIV_ROUND_UP(priv->rx_bytes - priv->tx_bytes, bpw);

        while (fill_words--)
                uniphier_spi_send(priv);
}

static void uniphier_spi_set_cs(struct spi_device *spi, bool enable)
{
        struct uniphier_spi_priv *priv = spi_controller_get_devdata(spi->controller);
        u32 val;

        val = readl(priv->base + SSI_FPS);

        if (enable)
                val |= SSI_FPS_FSPOL;
        else
                val &= ~SSI_FPS_FSPOL;

        writel(val, priv->base + SSI_FPS);
}

static bool uniphier_spi_can_dma(struct spi_controller *host,
                                 struct spi_device *spi,
                                 struct spi_transfer *t)
{
        struct uniphier_spi_priv *priv = spi_controller_get_devdata(host);
        unsigned int bpw = bytes_per_word(priv->bits_per_word);

        if ((!host->dma_tx && !host->dma_rx)
            || (!host->dma_tx && t->tx_buf)
            || (!host->dma_rx && t->rx_buf))
                return false;

        return DIV_ROUND_UP(t->len, bpw) > SSI_FIFO_DEPTH;
}

static void uniphier_spi_dma_rxcb(void *data)
{
        struct spi_controller *host = data;
        struct uniphier_spi_priv *priv = spi_controller_get_devdata(host);
        int state = atomic_fetch_andnot(SSI_DMA_RX_BUSY, &priv->dma_busy);

        uniphier_spi_irq_disable(priv, SSI_IE_RXRE);

        if (!(state & SSI_DMA_TX_BUSY))
                spi_finalize_current_transfer(host);
}

static void uniphier_spi_dma_txcb(void *data)
{
        struct spi_controller *host = data;
        struct uniphier_spi_priv *priv = spi_controller_get_devdata(host);
        int state = atomic_fetch_andnot(SSI_DMA_TX_BUSY, &priv->dma_busy);

        uniphier_spi_irq_disable(priv, SSI_IE_TXRE);

        if (!(state & SSI_DMA_RX_BUSY))
                spi_finalize_current_transfer(host);
}

static int uniphier_spi_transfer_one_dma(struct spi_controller *host,
                                         struct spi_device *spi,
                                         struct spi_transfer *t)
{
        struct uniphier_spi_priv *priv = spi_controller_get_devdata(host);
        struct dma_async_tx_descriptor *rxdesc = NULL, *txdesc = NULL;
        int buswidth;

        atomic_set(&priv->dma_busy, 0);

        uniphier_spi_set_fifo_threshold(priv, SSI_FIFO_BURST_NUM);

        if (priv->bits_per_word <= 8)
                buswidth = DMA_SLAVE_BUSWIDTH_1_BYTE;
        else if (priv->bits_per_word <= 16)
                buswidth = DMA_SLAVE_BUSWIDTH_2_BYTES;
        else
                buswidth = DMA_SLAVE_BUSWIDTH_4_BYTES;

        if (priv->rx_buf) {
                struct dma_slave_config rxconf = {
                        .direction = DMA_DEV_TO_MEM,
                        .src_addr = priv->base_dma_addr + SSI_RXDR,
                        .src_addr_width = buswidth,
                        .src_maxburst = SSI_FIFO_BURST_NUM,
                };

                dmaengine_slave_config(host->dma_rx, &rxconf);

                rxdesc = dmaengine_prep_slave_sg(
                        host->dma_rx,
                        t->rx_sg.sgl, t->rx_sg.nents,
                        DMA_DEV_TO_MEM, DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
                if (!rxdesc)
                        goto out_err_prep;

                rxdesc->callback = uniphier_spi_dma_rxcb;
                rxdesc->callback_param = host;

                uniphier_spi_irq_enable(priv, SSI_IE_RXRE);
                atomic_or(SSI_DMA_RX_BUSY, &priv->dma_busy);

                dmaengine_submit(rxdesc);
                dma_async_issue_pending(host->dma_rx);
        }

        if (priv->tx_buf) {
                struct dma_slave_config txconf = {
                        .direction = DMA_MEM_TO_DEV,
                        .dst_addr = priv->base_dma_addr + SSI_TXDR,
                        .dst_addr_width = buswidth,
                        .dst_maxburst = SSI_FIFO_BURST_NUM,
                };

                dmaengine_slave_config(host->dma_tx, &txconf);

                txdesc = dmaengine_prep_slave_sg(
                        host->dma_tx,
                        t->tx_sg.sgl, t->tx_sg.nents,
                        DMA_MEM_TO_DEV, DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
                if (!txdesc)
                        goto out_err_prep;

                txdesc->callback = uniphier_spi_dma_txcb;
                txdesc->callback_param = host;

                uniphier_spi_irq_enable(priv, SSI_IE_TXRE);
                atomic_or(SSI_DMA_TX_BUSY, &priv->dma_busy);

                dmaengine_submit(txdesc);
                dma_async_issue_pending(host->dma_tx);
        }

        /* signal that we need to wait for completion */
        return (priv->tx_buf || priv->rx_buf);

out_err_prep:
        if (rxdesc)
                dmaengine_terminate_sync(host->dma_rx);

        return -EINVAL;
}

static int uniphier_spi_transfer_one_irq(struct spi_controller *host,
                                         struct spi_device *spi,
                                         struct spi_transfer *t)
{
        struct uniphier_spi_priv *priv = spi_controller_get_devdata(host);
        struct device *dev = host->dev.parent;
        unsigned long time_left;

        reinit_completion(&priv->xfer_done);

        uniphier_spi_fill_tx_fifo(priv);

        uniphier_spi_irq_enable(priv, SSI_IE_RCIE | SSI_IE_RORIE);

        time_left = wait_for_completion_timeout(&priv->xfer_done,
                                        msecs_to_jiffies(SSI_TIMEOUT_MS));

        uniphier_spi_irq_disable(priv, SSI_IE_RCIE | SSI_IE_RORIE);

        if (!time_left) {
                dev_err(dev, "transfer timeout.\n");
                return -ETIMEDOUT;
        }

        return priv->error;
}

static int uniphier_spi_transfer_one_poll(struct spi_controller *host,
                                          struct spi_device *spi,
                                          struct spi_transfer *t)
{
        struct uniphier_spi_priv *priv = spi_controller_get_devdata(host);
        int loop = SSI_POLL_TIMEOUT_US * 10;

        while (priv->tx_bytes) {
                uniphier_spi_fill_tx_fifo(priv);

                while ((priv->rx_bytes - priv->tx_bytes) > 0) {
                        while (!(readl(priv->base + SSI_SR) & SSI_SR_RNE)
                                                                && loop--)
                                ndelay(100);

                        if (loop == -1)
                                goto irq_transfer;

                        uniphier_spi_recv(priv);
                }
        }

        return 0;

irq_transfer:
        return uniphier_spi_transfer_one_irq(host, spi, t);
}

static int uniphier_spi_transfer_one(struct spi_controller *host,
                                     struct spi_device *spi,
                                     struct spi_transfer *t)
{
        struct uniphier_spi_priv *priv = spi_controller_get_devdata(host);
        unsigned long threshold;
        bool use_dma;

        /* Terminate and return success for 0 byte length transfer */
        if (!t->len)
                return 0;

        uniphier_spi_setup_transfer(spi, t);

        use_dma = host->can_dma ? host->can_dma(host, spi, t) : false;
        if (use_dma)
                return uniphier_spi_transfer_one_dma(host, spi, t);

        /*
         * If the transfer operation will take longer than
         * SSI_POLL_TIMEOUT_US, it should use irq.
         */
        threshold = DIV_ROUND_UP(SSI_POLL_TIMEOUT_US * priv->speed_hz,
                                        USEC_PER_SEC * BITS_PER_BYTE);
        if (t->len > threshold)
                return uniphier_spi_transfer_one_irq(host, spi, t);
        else
                return uniphier_spi_transfer_one_poll(host, spi, t);
}

static int uniphier_spi_prepare_transfer_hardware(struct spi_controller *host)
{
        struct uniphier_spi_priv *priv = spi_controller_get_devdata(host);

        writel(SSI_CTL_EN, priv->base + SSI_CTL);

        return 0;
}

static int uniphier_spi_unprepare_transfer_hardware(struct spi_controller *host)
{
        struct uniphier_spi_priv *priv = spi_controller_get_devdata(host);

        writel(0, priv->base + SSI_CTL);

        return 0;
}

static void uniphier_spi_handle_err(struct spi_controller *host,
                                    struct spi_message *msg)
{
        struct uniphier_spi_priv *priv = spi_controller_get_devdata(host);
        u32 val;

        /* stop running spi transfer */
        writel(0, priv->base + SSI_CTL);

        /* reset FIFOs */
        val = SSI_FC_TXFFL | SSI_FC_RXFFL;
        writel(val, priv->base + SSI_FC);

        uniphier_spi_irq_disable(priv, SSI_IE_ALL_MASK);

        if (atomic_read(&priv->dma_busy) & SSI_DMA_TX_BUSY) {
                dmaengine_terminate_async(host->dma_tx);
                atomic_andnot(SSI_DMA_TX_BUSY, &priv->dma_busy);
        }

        if (atomic_read(&priv->dma_busy) & SSI_DMA_RX_BUSY) {
                dmaengine_terminate_async(host->dma_rx);
                atomic_andnot(SSI_DMA_RX_BUSY, &priv->dma_busy);
        }
}

static irqreturn_t uniphier_spi_handler(int irq, void *dev_id)
{
        struct uniphier_spi_priv *priv = dev_id;
        u32 val, stat;

        stat = readl(priv->base + SSI_IS);
        val = SSI_IC_TCIC | SSI_IC_RCIC | SSI_IC_RORIC;
        writel(val, priv->base + SSI_IC);

        /* rx fifo overrun */
        if (stat & SSI_IS_RORID) {
                priv->error = -EIO;
                goto done;
        }

        /* rx complete */
        if ((stat & SSI_IS_RCID) && (stat & SSI_IS_RXRS)) {
                while ((readl(priv->base + SSI_SR) & SSI_SR_RNE) &&
                                (priv->rx_bytes - priv->tx_bytes) > 0)
                        uniphier_spi_recv(priv);

                if ((readl(priv->base + SSI_SR) & SSI_SR_RNE) ||
                                (priv->rx_bytes != priv->tx_bytes)) {
                        priv->error = -EIO;
                        goto done;
                } else if (priv->rx_bytes == 0)
                        goto done;

                /* next tx transfer */
                uniphier_spi_fill_tx_fifo(priv);

                return IRQ_HANDLED;
        }

        return IRQ_NONE;

done:
        complete(&priv->xfer_done);
        return IRQ_HANDLED;
}

static int uniphier_spi_probe(struct platform_device *pdev)
{
        struct uniphier_spi_priv *priv;
        struct spi_controller *host;
        struct resource *res;
        struct dma_slave_caps caps;
        u32 dma_tx_burst = 0, dma_rx_burst = 0;
        unsigned long clk_rate;
        int irq;
        int ret;

        host = spi_alloc_host(&pdev->dev, sizeof(*priv));
        if (!host)
                return -ENOMEM;

        platform_set_drvdata(pdev, host);

        priv = spi_controller_get_devdata(host);
        priv->host = host;
        priv->is_save_param = false;

        priv->base = devm_platform_get_and_ioremap_resource(pdev, 0, &res);
        if (IS_ERR(priv->base)) {
                ret = PTR_ERR(priv->base);
                goto out_host_put;
        }
        priv->base_dma_addr = res->start;

        priv->clk = devm_clk_get(&pdev->dev, NULL);
        if (IS_ERR(priv->clk)) {
                dev_err(&pdev->dev, "failed to get clock\n");
                ret = PTR_ERR(priv->clk);
                goto out_host_put;
        }

        ret = clk_prepare_enable(priv->clk);
        if (ret)
                goto out_host_put;

        irq = platform_get_irq(pdev, 0);
        if (irq < 0) {
                ret = irq;
                goto out_disable_clk;
        }

        ret = devm_request_irq(&pdev->dev, irq, uniphier_spi_handler,
                               0, "uniphier-spi", priv);
        if (ret) {
                dev_err(&pdev->dev, "failed to request IRQ\n");
                goto out_disable_clk;
        }

        init_completion(&priv->xfer_done);

        clk_rate = clk_get_rate(priv->clk);

        host->max_speed_hz = DIV_ROUND_UP(clk_rate, SSI_MIN_CLK_DIVIDER);
        host->min_speed_hz = DIV_ROUND_UP(clk_rate, SSI_MAX_CLK_DIVIDER);
        host->mode_bits = SPI_CPOL | SPI_CPHA | SPI_CS_HIGH | SPI_LSB_FIRST;
        host->bus_num = pdev->id;
        host->bits_per_word_mask = SPI_BPW_RANGE_MASK(1, 32);

        host->set_cs = uniphier_spi_set_cs;
        host->transfer_one = uniphier_spi_transfer_one;
        host->prepare_transfer_hardware
                                = uniphier_spi_prepare_transfer_hardware;
        host->unprepare_transfer_hardware
                                = uniphier_spi_unprepare_transfer_hardware;
        host->handle_err = uniphier_spi_handle_err;
        host->can_dma = uniphier_spi_can_dma;

        host->num_chipselect = 1;
        host->flags = SPI_CONTROLLER_MUST_RX | SPI_CONTROLLER_MUST_TX;

        host->dma_tx = dma_request_chan(&pdev->dev, "tx");
        if (IS_ERR_OR_NULL(host->dma_tx)) {
                if (PTR_ERR(host->dma_tx) == -EPROBE_DEFER) {
                        ret = -EPROBE_DEFER;
                        goto out_disable_clk;
                }
                host->dma_tx = NULL;
                dma_tx_burst = INT_MAX;
        } else {
                ret = dma_get_slave_caps(host->dma_tx, &caps);
                if (ret) {
                        dev_err(&pdev->dev, "failed to get TX DMA capacities: %d\n",
                                ret);
                        goto out_release_dma;
                }
                dma_tx_burst = caps.max_burst;
        }

        host->dma_rx = dma_request_chan(&pdev->dev, "rx");
        if (IS_ERR_OR_NULL(host->dma_rx)) {
                if (PTR_ERR(host->dma_rx) == -EPROBE_DEFER) {
                        ret = -EPROBE_DEFER;
                        goto out_release_dma;
                }
                host->dma_rx = NULL;
                dma_rx_burst = INT_MAX;
        } else {
                ret = dma_get_slave_caps(host->dma_rx, &caps);
                if (ret) {
                        dev_err(&pdev->dev, "failed to get RX DMA capacities: %d\n",
                                ret);
                        goto out_release_dma;
                }
                dma_rx_burst = caps.max_burst;
        }

        host->max_dma_len = min(dma_tx_burst, dma_rx_burst);

        ret = devm_spi_register_controller(&pdev->dev, host);
        if (ret)
                goto out_release_dma;

        return 0;

out_release_dma:
        if (!IS_ERR_OR_NULL(host->dma_rx)) {
                dma_release_channel(host->dma_rx);
                host->dma_rx = NULL;
        }
        if (!IS_ERR_OR_NULL(host->dma_tx)) {
                dma_release_channel(host->dma_tx);
                host->dma_tx = NULL;
        }

out_disable_clk:
        clk_disable_unprepare(priv->clk);

out_host_put:
        spi_controller_put(host);
        return ret;
}

static void uniphier_spi_remove(struct platform_device *pdev)
{
        struct spi_controller *host = platform_get_drvdata(pdev);
        struct uniphier_spi_priv *priv = spi_controller_get_devdata(host);

        if (host->dma_tx)
                dma_release_channel(host->dma_tx);
        if (host->dma_rx)
                dma_release_channel(host->dma_rx);

        clk_disable_unprepare(priv->clk);
}

static const struct of_device_id uniphier_spi_match[] = {
        { .compatible = "socionext,uniphier-scssi" },
        { /* sentinel */ }
};
MODULE_DEVICE_TABLE(of, uniphier_spi_match);

static struct platform_driver uniphier_spi_driver = {
        .probe = uniphier_spi_probe,
        .remove = uniphier_spi_remove,
        .driver = {
                .name = "uniphier-spi",
                .of_match_table = uniphier_spi_match,
        },
};
module_platform_driver(uniphier_spi_driver);

MODULE_AUTHOR("Kunihiko Hayashi <hayashi.kunihiko@socionext.com>");
MODULE_AUTHOR("Keiji Hayashibara <hayashibara.keiji@socionext.com>");
MODULE_DESCRIPTION("Socionext UniPhier SPI controller driver");
MODULE_LICENSE("GPL v2");