root/drivers/dma/fsldma.h
/* SPDX-License-Identifier: GPL-2.0-or-later */
/*
 * Copyright (C) 2007-2010 Freescale Semiconductor, Inc. All rights reserved.
 *
 * Author:
 *   Zhang Wei <wei.zhang@freescale.com>, Jul 2007
 *   Ebony Zhu <ebony.zhu@freescale.com>, May 2007
 */
#ifndef __DMA_FSLDMA_H
#define __DMA_FSLDMA_H

#include <linux/device.h>
#include <linux/dmapool.h>
#include <linux/dmaengine.h>

/* Define data structures needed by Freescale
 * MPC8540 and MPC8349 DMA controller.
 */
#define FSL_DMA_MR_CS           0x00000001
#define FSL_DMA_MR_CC           0x00000002
#define FSL_DMA_MR_CA           0x00000008
#define FSL_DMA_MR_EIE          0x00000040
#define FSL_DMA_MR_XFE          0x00000020
#define FSL_DMA_MR_EOLNIE       0x00000100
#define FSL_DMA_MR_EOLSIE       0x00000080
#define FSL_DMA_MR_EOSIE        0x00000200
#define FSL_DMA_MR_CDSM         0x00000010
#define FSL_DMA_MR_CTM          0x00000004
#define FSL_DMA_MR_EMP_EN       0x00200000
#define FSL_DMA_MR_EMS_EN       0x00040000
#define FSL_DMA_MR_DAHE         0x00002000
#define FSL_DMA_MR_SAHE         0x00001000

#define FSL_DMA_MR_SAHTS_MASK   0x0000C000
#define FSL_DMA_MR_DAHTS_MASK   0x00030000
#define FSL_DMA_MR_BWC_MASK     0x0f000000

/*
 * Bandwidth/pause control determines how many bytes a given
 * channel is allowed to transfer before the DMA engine pauses
 * the current channel and switches to the next channel
 */
#define FSL_DMA_MR_BWC         0x0A000000

/* Special MR definition for MPC8349 */
#define FSL_DMA_MR_EOTIE        0x00000080
#define FSL_DMA_MR_PRC_RM       0x00000800

#define FSL_DMA_SR_CH           0x00000020
#define FSL_DMA_SR_PE           0x00000010
#define FSL_DMA_SR_CB           0x00000004
#define FSL_DMA_SR_TE           0x00000080
#define FSL_DMA_SR_EOSI         0x00000002
#define FSL_DMA_SR_EOLSI        0x00000001
#define FSL_DMA_SR_EOCDI        0x00000001
#define FSL_DMA_SR_EOLNI        0x00000008

#define FSL_DMA_SATR_SBPATMU                    0x20000000
#define FSL_DMA_SATR_STRANSINT_RIO              0x00c00000
#define FSL_DMA_SATR_SREADTYPE_SNOOP_READ       0x00050000
#define FSL_DMA_SATR_SREADTYPE_BP_IORH          0x00020000
#define FSL_DMA_SATR_SREADTYPE_BP_NREAD         0x00040000
#define FSL_DMA_SATR_SREADTYPE_BP_MREAD         0x00070000

#define FSL_DMA_DATR_DBPATMU                    0x20000000
#define FSL_DMA_DATR_DTRANSINT_RIO              0x00c00000
#define FSL_DMA_DATR_DWRITETYPE_SNOOP_WRITE     0x00050000
#define FSL_DMA_DATR_DWRITETYPE_BP_FLUSH        0x00010000

#define FSL_DMA_EOL             ((u64)0x1)
#define FSL_DMA_SNEN            ((u64)0x10)
#define FSL_DMA_EOSIE           0x8
#define FSL_DMA_NLDA_MASK       (~(u64)0x1f)

#define FSL_DMA_BCR_MAX_CNT     0x03ffffffu

#define FSL_DMA_DGSR_TE         0x80
#define FSL_DMA_DGSR_CH         0x20
#define FSL_DMA_DGSR_PE         0x10
#define FSL_DMA_DGSR_EOLNI      0x08
#define FSL_DMA_DGSR_CB         0x04
#define FSL_DMA_DGSR_EOSI       0x02
#define FSL_DMA_DGSR_EOLSI      0x01

#define FSL_DMA_BUSWIDTHS       (BIT(DMA_SLAVE_BUSWIDTH_1_BYTE) | \
                                BIT(DMA_SLAVE_BUSWIDTH_2_BYTES) | \
                                BIT(DMA_SLAVE_BUSWIDTH_4_BYTES) | \
                                BIT(DMA_SLAVE_BUSWIDTH_8_BYTES))
typedef u64 __bitwise v64;
typedef u32 __bitwise v32;

struct fsl_dma_ld_hw {
        v64 src_addr;
        v64 dst_addr;
        v64 next_ln_addr;
        v32 count;
        v32 reserve;
} __attribute__((aligned(32)));

struct fsl_desc_sw {
        struct fsl_dma_ld_hw hw;
        struct list_head node;
        struct list_head tx_list;
        struct dma_async_tx_descriptor async_tx;
} __attribute__((aligned(32)));

struct fsldma_chan_regs {
        u32 mr;         /* 0x00 - Mode Register */
        u32 sr;         /* 0x04 - Status Register */
        u64 cdar;       /* 0x08 - Current descriptor address register */
        u64 sar;        /* 0x10 - Source Address Register */
        u64 dar;        /* 0x18 - Destination Address Register */
        u32 bcr;        /* 0x20 - Byte Count Register */
        u64 ndar;       /* 0x24 - Next Descriptor Address Register */
};

struct fsldma_chan;
#define FSL_DMA_MAX_CHANS_PER_DEVICE 8

struct fsldma_device {
        void __iomem *regs;     /* DGSR register base */
        struct device *dev;
        struct dma_device common;
        struct fsldma_chan *chan[FSL_DMA_MAX_CHANS_PER_DEVICE];
        u32 feature;            /* The same as DMA channels */
        int irq;                /* Channel IRQ */
        int addr_bits;          /* DMA addressing bits supported */
};

/* Define macros for fsldma_chan->feature property */
#define FSL_DMA_LITTLE_ENDIAN   0x00000000
#define FSL_DMA_BIG_ENDIAN      0x00000001

#define FSL_DMA_IP_MASK         0x00000ff0
#define FSL_DMA_IP_85XX         0x00000010
#define FSL_DMA_IP_83XX         0x00000020

#define FSL_DMA_CHAN_PAUSE_EXT  0x00001000
#define FSL_DMA_CHAN_START_EXT  0x00002000

#ifdef CONFIG_PM
struct fsldma_chan_regs_save {
        u32 mr;
};

enum fsldma_pm_state {
        RUNNING = 0,
        SUSPENDED,
};
#endif

struct fsldma_chan {
        char name[8];                   /* Channel name */
        struct fsldma_chan_regs __iomem *regs;
        spinlock_t desc_lock;           /* Descriptor operation lock */
        /*
         * Descriptors which are queued to run, but have not yet been
         * submitted to the hardware for execution
         */
        struct list_head ld_pending;
        /*
         * Descriptors which are currently being executed by the hardware
         */
        struct list_head ld_running;
        /*
         * Descriptors which have finished execution by the hardware. These
         * descriptors have already had their cleanup actions run. They are
         * waiting for the ACK bit to be set by the async_tx API.
         */
        struct list_head ld_completed;  /* Link descriptors queue */
        struct dma_chan common;         /* DMA common channel */
        struct dma_pool *desc_pool;     /* Descriptors pool */
        struct device *dev;             /* Channel device */
        int irq;                        /* Channel IRQ */
        int id;                         /* Raw id of this channel */
        struct tasklet_struct tasklet;
        u32 feature;
        bool idle;                      /* DMA controller is idle */
#ifdef CONFIG_PM
        struct fsldma_chan_regs_save regs_save;
        enum fsldma_pm_state pm_state;
#endif

        void (*toggle_ext_pause)(struct fsldma_chan *fsl_chan, int enable);
        void (*toggle_ext_start)(struct fsldma_chan *fsl_chan, int enable);
        void (*set_src_loop_size)(struct fsldma_chan *fsl_chan, int size);
        void (*set_dst_loop_size)(struct fsldma_chan *fsl_chan, int size);
        void (*set_request_count)(struct fsldma_chan *fsl_chan, int size);
};

#define to_fsl_chan(chan) container_of(chan, struct fsldma_chan, common)
#define to_fsl_desc(lh) container_of(lh, struct fsl_desc_sw, node)
#define tx_to_fsl_desc(tx) container_of(tx, struct fsl_desc_sw, async_tx)

#ifdef  CONFIG_PPC
#define fsl_ioread32(p)         in_le32(p)
#define fsl_ioread32be(p)       in_be32(p)
#define fsl_iowrite32(v, p)     out_le32(p, v)
#define fsl_iowrite32be(v, p)   out_be32(p, v)

#ifdef __powerpc64__
#define fsl_ioread64(p)         in_le64(p)
#define fsl_ioread64be(p)       in_be64(p)
#define fsl_iowrite64(v, p)     out_le64(p, v)
#define fsl_iowrite64be(v, p)   out_be64(p, v)
#else
static u64 fsl_ioread64(const u64 __iomem *addr)
{
        u32 val_lo = in_le32((u32 __iomem *)addr);
        u32 val_hi = in_le32((u32 __iomem *)addr + 1);

        return ((u64)val_hi << 32) + val_lo;
}

static void fsl_iowrite64(u64 val, u64 __iomem *addr)
{
        out_le32((u32 __iomem *)addr + 1, val >> 32);
        out_le32((u32 __iomem *)addr, (u32)val);
}

static u64 fsl_ioread64be(const u64 __iomem *addr)
{
        u32 val_hi = in_be32((u32 __iomem *)addr);
        u32 val_lo = in_be32((u32 __iomem *)addr + 1);

        return ((u64)val_hi << 32) + val_lo;
}

static void fsl_iowrite64be(u64 val, u64 __iomem *addr)
{
        out_be32((u32 __iomem *)addr, val >> 32);
        out_be32((u32 __iomem *)addr + 1, (u32)val);
}
#endif
#endif

#if defined(CONFIG_ARM64) || defined(CONFIG_ARM)
#define fsl_ioread32(p)         ioread32(p)
#define fsl_ioread32be(p)       ioread32be(p)
#define fsl_iowrite32(v, p)     iowrite32(v, p)
#define fsl_iowrite32be(v, p)   iowrite32be(v, p)
#define fsl_ioread64(p)         ioread64(p)
#define fsl_ioread64be(p)       ioread64be(p)
#define fsl_iowrite64(v, p)     iowrite64(v, p)
#define fsl_iowrite64be(v, p)   iowrite64be(v, p)
#endif

#define FSL_DMA_IN(fsl_dma, addr, width)                        \
                (((fsl_dma)->feature & FSL_DMA_BIG_ENDIAN) ?    \
                        fsl_ioread##width##be(addr) : fsl_ioread##width(addr))

#define FSL_DMA_OUT(fsl_dma, addr, val, width)                  \
                (((fsl_dma)->feature & FSL_DMA_BIG_ENDIAN) ?    \
                        fsl_iowrite##width##be(val, addr) : fsl_iowrite \
                ##width(val, addr))

#define DMA_TO_CPU(fsl_chan, d, width)                                  \
                (((fsl_chan)->feature & FSL_DMA_BIG_ENDIAN) ?           \
                        be##width##_to_cpu((__force __be##width)(v##width)d) : \
                        le##width##_to_cpu((__force __le##width)(v##width)d))
#define CPU_TO_DMA(fsl_chan, c, width)                                  \
                (((fsl_chan)->feature & FSL_DMA_BIG_ENDIAN) ?           \
                        (__force v##width)cpu_to_be##width(c) :         \
                        (__force v##width)cpu_to_le##width(c))

#endif  /* __DMA_FSLDMA_H */