root/drivers/media/platform/cadence/cdns-csi2rx.c
// SPDX-License-Identifier: GPL-2.0+
/*
 * Driver for Cadence MIPI-CSI2 RX Controller v1.3
 *
 * Copyright (C) 2017 Cadence Design Systems Inc.
 */

#include <linux/bitfield.h>
#include <linux/clk.h>
#include <linux/delay.h>
#include <linux/export.h>
#include <linux/io.h>
#include <linux/iopoll.h>
#include <linux/module.h>
#include <linux/of.h>
#include <linux/of_graph.h>
#include <linux/phy/phy.h>
#include <linux/platform_device.h>
#include <linux/reset.h>
#include <linux/slab.h>

#include <media/cadence/cdns-csi2rx.h>
#include <media/v4l2-ctrls.h>
#include <media/v4l2-device.h>
#include <media/v4l2-fwnode.h>
#include <media/v4l2-subdev.h>

#define CSI2RX_DEVICE_CFG_REG                   0x000

#define CSI2RX_SOFT_RESET_REG                   0x004
#define CSI2RX_SOFT_RESET_PROTOCOL                      BIT(1)
#define CSI2RX_SOFT_RESET_FRONT                         BIT(0)

#define CSI2RX_STATIC_CFG_REG                   0x008
#define CSI2RX_STATIC_CFG_DLANE_MAP(llane, plane)       ((plane) << (16 + (llane) * 4))
#define CSI2RX_STATIC_CFG_LANES_MASK                    GENMASK(11, 8)

#define CSI2RX_DPHY_LANE_CTRL_REG               0x40
#define CSI2RX_DPHY_CL_RST                      BIT(16)
#define CSI2RX_DPHY_DL_RST(i)                   BIT((i) + 12)
#define CSI2RX_DPHY_CL_EN                       BIT(4)
#define CSI2RX_DPHY_DL_EN(i)                    BIT(i)

#define CSI2RX_STREAM_BASE(n)           (((n) + 1) * 0x100)

#define CSI2RX_STREAM_CTRL_REG(n)               (CSI2RX_STREAM_BASE(n) + 0x000)
#define CSI2RX_STREAM_CTRL_SOFT_RST                     BIT(4)
#define CSI2RX_STREAM_CTRL_STOP                         BIT(1)
#define CSI2RX_STREAM_CTRL_START                        BIT(0)

#define CSI2RX_STREAM_STATUS_REG(n)             (CSI2RX_STREAM_BASE(n) + 0x004)
#define CSI2RX_STREAM_STATUS_RDY                        BIT(31)

#define CSI2RX_STREAM_DATA_CFG_REG(n)           (CSI2RX_STREAM_BASE(n) + 0x008)
#define CSI2RX_STREAM_DATA_CFG_VC_SELECT(n)             BIT((n) + 16)

#define CSI2RX_STREAM_CFG_REG(n)                (CSI2RX_STREAM_BASE(n) + 0x00c)
#define CSI2RX_STREAM_CFG_FIFO_MODE_LARGE_BUF           BIT(8)
#define CSI2RX_STREAM_CFG_NUM_PIXELS_MASK               GENMASK(5, 4)
#define CSI2RX_STREAM_CFG_NUM_PIXELS(n)                 ((n) >> 1U)

#define CSI2RX_LANES_MAX        4
#define CSI2RX_STREAMS_MAX      4

#define CSI2RX_ERROR_IRQS_REG                   0x28
#define CSI2RX_ERROR_IRQS_MASK_REG              0x2C

#define CSI2RX_STREAM3_FIFO_OVERFLOW_IRQ        BIT(19)
#define CSI2RX_STREAM2_FIFO_OVERFLOW_IRQ        BIT(18)
#define CSI2RX_STREAM1_FIFO_OVERFLOW_IRQ        BIT(17)
#define CSI2RX_STREAM0_FIFO_OVERFLOW_IRQ        BIT(16)
#define CSI2RX_FRONT_TRUNC_HDR_IRQ              BIT(12)
#define CSI2RX_PROT_TRUNCATED_PACKET_IRQ        BIT(11)
#define CSI2RX_FRONT_LP_NO_PAYLOAD_IRQ          BIT(10)
#define CSI2RX_SP_INVALID_RCVD_IRQ              BIT(9)
#define CSI2RX_DATA_ID_IRQ                      BIT(7)
#define CSI2RX_HEADER_CORRECTED_ECC_IRQ BIT(6)
#define CSI2RX_HEADER_ECC_IRQ                   BIT(5)
#define CSI2RX_PAYLOAD_CRC_IRQ                  BIT(4)

#define CSI2RX_ECC_ERRORS               GENMASK(7, 4)
#define CSI2RX_PACKET_ERRORS            GENMASK(12, 9)

enum csi2rx_pads {
        CSI2RX_PAD_SINK,
        CSI2RX_PAD_SOURCE_STREAM0,
        CSI2RX_PAD_SOURCE_STREAM1,
        CSI2RX_PAD_SOURCE_STREAM2,
        CSI2RX_PAD_SOURCE_STREAM3,
        CSI2RX_PAD_MAX,
};

struct csi2rx_fmt {
        u32                             code;
        /* width of a single pixel on CSI-2 bus */
        u8                              bpp;
        /* max pixels per clock supported on output bus */
        u8                              max_pixels;
};

struct csi2rx_event {
        u32 mask;
        const char *name;
};

static const struct csi2rx_event csi2rx_events[] = {
        { CSI2RX_STREAM3_FIFO_OVERFLOW_IRQ, "Overflow of the Stream 3 FIFO detected" },
        { CSI2RX_STREAM2_FIFO_OVERFLOW_IRQ, "Overflow of the Stream 2 FIFO detected" },
        { CSI2RX_STREAM1_FIFO_OVERFLOW_IRQ, "Overflow of the Stream 1 FIFO detected" },
        { CSI2RX_STREAM0_FIFO_OVERFLOW_IRQ, "Overflow of the Stream 0 FIFO detected" },
        { CSI2RX_FRONT_TRUNC_HDR_IRQ, "A truncated header [short or long] has been received" },
        { CSI2RX_PROT_TRUNCATED_PACKET_IRQ, "A truncated long packet has been received" },
        { CSI2RX_FRONT_LP_NO_PAYLOAD_IRQ, "A truncated long packet has been received. No payload" },
        { CSI2RX_SP_INVALID_RCVD_IRQ, "A reserved or invalid short packet has been received" },
        { CSI2RX_DATA_ID_IRQ, "Data ID error in the header packet" },
        { CSI2RX_HEADER_CORRECTED_ECC_IRQ, "ECC error detected and corrected" },
        { CSI2RX_HEADER_ECC_IRQ, "Unrecoverable ECC error" },
        { CSI2RX_PAYLOAD_CRC_IRQ, "CRC error" },
};

#define CSI2RX_NUM_EVENTS               ARRAY_SIZE(csi2rx_events)

struct csi2rx_priv {
        struct device                   *dev;
        unsigned int                    count;
        int                             error_irq;

        /*
         * Used to prevent race conditions between multiple,
         * concurrent calls to start and stop.
         */
        struct mutex                    lock;

        void __iomem                    *base;
        struct clk                      *sys_clk;
        struct clk                      *p_clk;
        struct clk                      *pixel_clk[CSI2RX_STREAMS_MAX];
        struct reset_control            *sys_rst;
        struct reset_control            *p_rst;
        struct reset_control            *pixel_rst[CSI2RX_STREAMS_MAX];
        struct phy                      *dphy;

        u8                              num_pixels[CSI2RX_STREAMS_MAX];
        u8                              lanes[CSI2RX_LANES_MAX];
        u8                              num_lanes;
        u8                              max_lanes;
        u8                              max_streams;
        bool                            has_internal_dphy;
        u32                             events[CSI2RX_NUM_EVENTS];

        struct v4l2_subdev              subdev;
        struct v4l2_async_notifier      notifier;
        struct media_pad                pads[CSI2RX_PAD_MAX];

        /* Remote source */
        struct v4l2_subdev              *source_subdev;
        int                             source_pad;
};

static const struct csi2rx_fmt formats[] = {
        { .code = MEDIA_BUS_FMT_YUYV8_1X16, .bpp = 16, .max_pixels = 2, },
        { .code = MEDIA_BUS_FMT_UYVY8_1X16, .bpp = 16, .max_pixels = 2, },
        { .code = MEDIA_BUS_FMT_YVYU8_1X16, .bpp = 16, .max_pixels = 2, },
        { .code = MEDIA_BUS_FMT_VYUY8_1X16, .bpp = 16, .max_pixels = 2, },
        { .code = MEDIA_BUS_FMT_SBGGR8_1X8, .bpp = 8, .max_pixels = 4, },
        { .code = MEDIA_BUS_FMT_SGBRG8_1X8, .bpp = 8, .max_pixels = 4, },
        { .code = MEDIA_BUS_FMT_SGRBG8_1X8, .bpp = 8, .max_pixels = 4, },
        { .code = MEDIA_BUS_FMT_SRGGB8_1X8, .bpp = 8, .max_pixels = 4, },
        { .code = MEDIA_BUS_FMT_Y8_1X8,     .bpp = 8, .max_pixels = 4, },
        { .code = MEDIA_BUS_FMT_SBGGR10_1X10, .bpp = 10, .max_pixels = 2, },
        { .code = MEDIA_BUS_FMT_SGBRG10_1X10, .bpp = 10, .max_pixels = 2, },
        { .code = MEDIA_BUS_FMT_SGRBG10_1X10, .bpp = 10, .max_pixels = 2, },
        { .code = MEDIA_BUS_FMT_SRGGB10_1X10, .bpp = 10, .max_pixels = 2, },
        { .code = MEDIA_BUS_FMT_RGB565_1X16,  .bpp = 16, .max_pixels = 1, },
        { .code = MEDIA_BUS_FMT_RGB888_1X24,  .bpp = 24, .max_pixels = 1, },
        { .code = MEDIA_BUS_FMT_BGR888_1X24,  .bpp = 24, .max_pixels = 1, },
};

static void csi2rx_configure_error_irq_mask(void __iomem *base,
                                            struct csi2rx_priv *csi2rx)
{
        u32 error_irq_mask = 0;

        error_irq_mask |= CSI2RX_ECC_ERRORS;
        error_irq_mask |= CSI2RX_PACKET_ERRORS;

        /*
         * Iterate through all source pads and check if they are linked
         * to an active remote pad. If an active remote pad is found,
         * calculate the corresponding bit position and set it in
         * mask, enabling the stream overflow error in the mask.
         */
        for (int i = CSI2RX_PAD_SOURCE_STREAM0; i < CSI2RX_PAD_MAX; i++) {
                struct media_pad *remote_pad;

                remote_pad = media_pad_remote_pad_first(&csi2rx->pads[i]);
                if (remote_pad) {
                        int pad = i - CSI2RX_PAD_SOURCE_STREAM0;
                        u32 bit_mask = CSI2RX_STREAM0_FIFO_OVERFLOW_IRQ << pad;

                        error_irq_mask |= bit_mask;
                }
        }

        writel(error_irq_mask, base + CSI2RX_ERROR_IRQS_MASK_REG);
}

static irqreturn_t csi2rx_irq_handler(int irq, void *dev_id)
{
        struct csi2rx_priv *csi2rx = dev_id;
        int i;
        u32 error_status, error_mask;

        error_status = readl(csi2rx->base + CSI2RX_ERROR_IRQS_REG);
        error_mask = readl(csi2rx->base + CSI2RX_ERROR_IRQS_MASK_REG);

        for (i = 0; i < CSI2RX_NUM_EVENTS; i++)
                if ((error_status & csi2rx_events[i].mask) &&
                    (error_mask & csi2rx_events[i].mask))
                        csi2rx->events[i]++;

        writel(error_status, csi2rx->base + CSI2RX_ERROR_IRQS_REG);

        return IRQ_HANDLED;
}

static const struct csi2rx_fmt *csi2rx_get_fmt_by_code(u32 code)
{
        unsigned int i;

        for (i = 0; i < ARRAY_SIZE(formats); i++)
                if (formats[i].code == code)
                        return &formats[i];

        return NULL;
}

static inline
struct csi2rx_priv *v4l2_subdev_to_csi2rx(struct v4l2_subdev *subdev)
{
        return container_of(subdev, struct csi2rx_priv, subdev);
}

static void csi2rx_reset(struct csi2rx_priv *csi2rx)
{
        unsigned int i;

        /* Reset module */
        writel(CSI2RX_SOFT_RESET_PROTOCOL | CSI2RX_SOFT_RESET_FRONT,
               csi2rx->base + CSI2RX_SOFT_RESET_REG);
        /* Reset individual streams. */
        for (i = 0; i < csi2rx->max_streams; i++) {
                writel(CSI2RX_STREAM_CTRL_SOFT_RST,
                       csi2rx->base + CSI2RX_STREAM_CTRL_REG(i));
        }

        usleep_range(10, 20);

        /* Clear resets */
        writel(0, csi2rx->base + CSI2RX_SOFT_RESET_REG);
        for (i = 0; i < csi2rx->max_streams; i++)
                writel(0, csi2rx->base + CSI2RX_STREAM_CTRL_REG(i));
}

static int csi2rx_configure_ext_dphy(struct csi2rx_priv *csi2rx)
{
        struct media_pad *src_pad =
                &csi2rx->source_subdev->entity.pads[csi2rx->source_pad];
        union phy_configure_opts opts = { };
        struct phy_configure_opts_mipi_dphy *cfg = &opts.mipi_dphy;
        struct v4l2_subdev_format sd_fmt = {
                .which  = V4L2_SUBDEV_FORMAT_ACTIVE,
                .pad    = CSI2RX_PAD_SINK,
        };
        const struct csi2rx_fmt *fmt;
        s64 link_freq;
        int ret;

        ret = v4l2_subdev_call_state_active(&csi2rx->subdev, pad, get_fmt,
                                            &sd_fmt);
        if (ret < 0)
                return ret;

        fmt = csi2rx_get_fmt_by_code(sd_fmt.format.code);

        link_freq = v4l2_get_link_freq(src_pad,
                                       fmt->bpp, 2 * csi2rx->num_lanes);
        if (link_freq < 0)
                return link_freq;

        ret = phy_mipi_dphy_get_default_config_for_hsclk(link_freq,
                                                         csi2rx->num_lanes, cfg);
        if (ret)
                return ret;

        ret = phy_power_on(csi2rx->dphy);
        if (ret)
                return ret;

        ret = phy_configure(csi2rx->dphy, &opts);
        if (ret) {
                phy_power_off(csi2rx->dphy);
                return ret;
        }

        return 0;
}

static int csi2rx_start(struct csi2rx_priv *csi2rx)
{
        unsigned int i;
        unsigned long lanes_used = 0;
        u32 reg;
        int ret;

        ret = clk_prepare_enable(csi2rx->p_clk);
        if (ret)
                return ret;

        reset_control_deassert(csi2rx->p_rst);
        csi2rx_reset(csi2rx);

        if (csi2rx->error_irq >= 0)
                csi2rx_configure_error_irq_mask(csi2rx->base, csi2rx);

        reg = csi2rx->num_lanes << 8;
        for (i = 0; i < csi2rx->num_lanes; i++) {
                reg |= CSI2RX_STATIC_CFG_DLANE_MAP(i, csi2rx->lanes[i]);
                set_bit(csi2rx->lanes[i], &lanes_used);
        }

        /*
         * Even the unused lanes need to be mapped. In order to avoid
         * to map twice to the same physical lane, keep the lanes used
         * in the previous loop, and only map unused physical lanes to
         * the rest of our logical lanes.
         */
        for (i = csi2rx->num_lanes; i < csi2rx->max_lanes; i++) {
                unsigned int idx = find_first_zero_bit(&lanes_used,
                                                       csi2rx->max_lanes);
                set_bit(idx, &lanes_used);
                reg |= CSI2RX_STATIC_CFG_DLANE_MAP(i, i + 1);
        }

        writel(reg, csi2rx->base + CSI2RX_STATIC_CFG_REG);

        /* Enable DPHY clk and data lanes. */
        if (csi2rx->dphy) {
                reg = CSI2RX_DPHY_CL_EN | CSI2RX_DPHY_CL_RST;
                for (i = 0; i < csi2rx->num_lanes; i++) {
                        reg |= CSI2RX_DPHY_DL_EN(csi2rx->lanes[i] - 1);
                        reg |= CSI2RX_DPHY_DL_RST(csi2rx->lanes[i] - 1);
                }

                writel(reg, csi2rx->base + CSI2RX_DPHY_LANE_CTRL_REG);

                ret = csi2rx_configure_ext_dphy(csi2rx);
                if (ret) {
                        dev_err(csi2rx->dev,
                                "Failed to configure external DPHY: %d\n", ret);
                        goto err_disable_pclk;
                }
        }

        /*
         * Create a static mapping between the CSI virtual channels
         * and the output stream.
         *
         * This should be enhanced, but v4l2 lacks the support for
         * changing that mapping dynamically.
         *
         * We also cannot enable and disable independent streams here,
         * hence the reference counting.
         */
        for (i = 0; i < csi2rx->max_streams; i++) {
                ret = clk_prepare_enable(csi2rx->pixel_clk[i]);
                if (ret)
                        goto err_disable_pixclk;

                reset_control_deassert(csi2rx->pixel_rst[i]);

                writel(CSI2RX_STREAM_CFG_FIFO_MODE_LARGE_BUF |
                               FIELD_PREP(CSI2RX_STREAM_CFG_NUM_PIXELS_MASK,
                                          csi2rx->num_pixels[i]),
                       csi2rx->base + CSI2RX_STREAM_CFG_REG(i));

                /*
                 * Enable one virtual channel. When multiple virtual channels
                 * are supported this will have to be changed.
                 */
                writel(CSI2RX_STREAM_DATA_CFG_VC_SELECT(0),
                       csi2rx->base + CSI2RX_STREAM_DATA_CFG_REG(i));

                writel(CSI2RX_STREAM_CTRL_START,
                       csi2rx->base + CSI2RX_STREAM_CTRL_REG(i));
        }

        ret = clk_prepare_enable(csi2rx->sys_clk);
        if (ret)
                goto err_disable_pixclk;

        reset_control_deassert(csi2rx->sys_rst);

        ret = v4l2_subdev_call(csi2rx->source_subdev, video, s_stream, true);
        if (ret)
                goto err_disable_sysclk;

        clk_disable_unprepare(csi2rx->p_clk);

        return 0;

err_disable_sysclk:
        clk_disable_unprepare(csi2rx->sys_clk);
err_disable_pixclk:
        for (; i > 0; i--) {
                reset_control_assert(csi2rx->pixel_rst[i - 1]);
                clk_disable_unprepare(csi2rx->pixel_clk[i - 1]);
        }

        if (csi2rx->dphy) {
                writel(0, csi2rx->base + CSI2RX_DPHY_LANE_CTRL_REG);
                phy_power_off(csi2rx->dphy);
        }
err_disable_pclk:
        clk_disable_unprepare(csi2rx->p_clk);

        return ret;
}

static void csi2rx_stop(struct csi2rx_priv *csi2rx)
{
        unsigned int i;
        u32 val;
        int ret;

        clk_prepare_enable(csi2rx->p_clk);
        reset_control_assert(csi2rx->sys_rst);
        clk_disable_unprepare(csi2rx->sys_clk);

        writel(0, csi2rx->base + CSI2RX_ERROR_IRQS_MASK_REG);

        for (i = 0; i < csi2rx->max_streams; i++) {
                writel(CSI2RX_STREAM_CTRL_STOP,
                       csi2rx->base + CSI2RX_STREAM_CTRL_REG(i));

                ret = readl_relaxed_poll_timeout(csi2rx->base +
                                                 CSI2RX_STREAM_STATUS_REG(i),
                                                 val,
                                                 !(val & CSI2RX_STREAM_STATUS_RDY),
                                                 10, 10000);
                if (ret)
                        dev_warn(csi2rx->dev,
                                 "Failed to stop streaming on pad%u\n", i);

                reset_control_assert(csi2rx->pixel_rst[i]);
                clk_disable_unprepare(csi2rx->pixel_clk[i]);
        }

        reset_control_assert(csi2rx->p_rst);
        clk_disable_unprepare(csi2rx->p_clk);

        if (v4l2_subdev_call(csi2rx->source_subdev, video, s_stream, false))
                dev_warn(csi2rx->dev, "Couldn't disable our subdev\n");

        if (csi2rx->dphy) {
                writel(0, csi2rx->base + CSI2RX_DPHY_LANE_CTRL_REG);

                if (phy_power_off(csi2rx->dphy))
                        dev_warn(csi2rx->dev, "Couldn't power off DPHY\n");
        }
}

static int csi2rx_log_status(struct v4l2_subdev *sd)
{
        struct csi2rx_priv *csi2rx = v4l2_subdev_to_csi2rx(sd);
        unsigned int i;

        for (i = 0; i < CSI2RX_NUM_EVENTS; i++) {
                if (csi2rx->events[i])
                        dev_info(csi2rx->dev, "%s events: %d\n",
                                 csi2rx_events[i].name,
                                 csi2rx->events[i]);
        }

        return 0;
}

static int csi2rx_s_stream(struct v4l2_subdev *subdev, int enable)
{
        struct csi2rx_priv *csi2rx = v4l2_subdev_to_csi2rx(subdev);
        int ret = 0;

        mutex_lock(&csi2rx->lock);

        if (enable) {
                /*
                 * If we're not the first users, there's no need to
                 * enable the whole controller.
                 */
                if (!csi2rx->count) {
                        ret = csi2rx_start(csi2rx);
                        if (ret)
                                goto out;
                }

                csi2rx->count++;
        } else {
                csi2rx->count--;

                /*
                 * Let the last user turn off the lights.
                 */
                if (!csi2rx->count)
                        csi2rx_stop(csi2rx);
        }

out:
        mutex_unlock(&csi2rx->lock);
        return ret;
}

static int csi2rx_enum_mbus_code(struct v4l2_subdev *subdev,
                                 struct v4l2_subdev_state *state,
                                 struct v4l2_subdev_mbus_code_enum *code_enum)
{
        if (code_enum->index >= ARRAY_SIZE(formats))
                return -EINVAL;

        code_enum->code = formats[code_enum->index].code;

        return 0;
}

static int csi2rx_set_fmt(struct v4l2_subdev *subdev,
                          struct v4l2_subdev_state *state,
                          struct v4l2_subdev_format *format)
{
        struct v4l2_mbus_framefmt *fmt;
        unsigned int i;

        /* No transcoding, source and sink formats must match. */
        if (format->pad != CSI2RX_PAD_SINK)
                return v4l2_subdev_get_fmt(subdev, state, format);

        if (!csi2rx_get_fmt_by_code(format->format.code))
                format->format.code = formats[0].code;

        format->format.field = V4L2_FIELD_NONE;

        /* Set sink format */
        fmt = v4l2_subdev_state_get_format(state, format->pad);
        *fmt = format->format;

        /* Propagate to source formats */
        for (i = CSI2RX_PAD_SOURCE_STREAM0; i < CSI2RX_PAD_MAX; i++) {
                fmt = v4l2_subdev_state_get_format(state, i);
                *fmt = format->format;
        }

        return 0;
}

static int csi2rx_init_state(struct v4l2_subdev *subdev,
                             struct v4l2_subdev_state *state)
{
        struct v4l2_subdev_format format = {
                .pad = CSI2RX_PAD_SINK,
                .format = {
                        .width = 640,
                        .height = 480,
                        .code = MEDIA_BUS_FMT_UYVY8_1X16,
                        .field = V4L2_FIELD_NONE,
                        .colorspace = V4L2_COLORSPACE_SRGB,
                        .ycbcr_enc = V4L2_YCBCR_ENC_601,
                        .quantization = V4L2_QUANTIZATION_LIM_RANGE,
                        .xfer_func = V4L2_XFER_FUNC_SRGB,
                },
        };

        return csi2rx_set_fmt(subdev, state, &format);
}

int cdns_csi2rx_negotiate_ppc(struct v4l2_subdev *subdev, unsigned int pad,
                              u8 *ppc)
{
        struct csi2rx_priv *csi2rx = v4l2_subdev_to_csi2rx(subdev);
        const struct csi2rx_fmt *csi_fmt;
        struct v4l2_subdev_state *state;
        struct v4l2_mbus_framefmt *fmt;

        if (!ppc || pad < CSI2RX_PAD_SOURCE_STREAM0 || pad >= CSI2RX_PAD_MAX)
                return -EINVAL;

        state = v4l2_subdev_lock_and_get_active_state(subdev);
        fmt = v4l2_subdev_state_get_format(state, pad);
        csi_fmt = csi2rx_get_fmt_by_code(fmt->code);

        /* Reduce requested PPC if it is too high */
        *ppc = min(*ppc, csi_fmt->max_pixels);

        v4l2_subdev_unlock_state(state);

        csi2rx->num_pixels[pad - CSI2RX_PAD_SOURCE_STREAM0] =
                CSI2RX_STREAM_CFG_NUM_PIXELS(*ppc);

        return 0;
}
EXPORT_SYMBOL_FOR_MODULES(cdns_csi2rx_negotiate_ppc, "j721e-csi2rx");

static const struct v4l2_subdev_pad_ops csi2rx_pad_ops = {
        .enum_mbus_code = csi2rx_enum_mbus_code,
        .get_fmt        = v4l2_subdev_get_fmt,
        .set_fmt        = csi2rx_set_fmt,
};

static const struct v4l2_subdev_video_ops csi2rx_video_ops = {
        .s_stream       = csi2rx_s_stream,
};

static const struct v4l2_subdev_core_ops csi2rx_core_ops = {
        .log_status     = csi2rx_log_status,
};

static const struct v4l2_subdev_ops csi2rx_subdev_ops = {
        .core           = &csi2rx_core_ops,
        .video          = &csi2rx_video_ops,
        .pad            = &csi2rx_pad_ops,
};

static const struct v4l2_subdev_internal_ops csi2rx_internal_ops = {
        .init_state     = csi2rx_init_state,
};

static const struct media_entity_operations csi2rx_media_ops = {
        .link_validate = v4l2_subdev_link_validate,
        .get_fwnode_pad = v4l2_subdev_get_fwnode_pad_1_to_1,
};

static int csi2rx_async_bound(struct v4l2_async_notifier *notifier,
                              struct v4l2_subdev *s_subdev,
                              struct v4l2_async_connection *asd)
{
        struct v4l2_subdev *subdev = notifier->sd;
        struct csi2rx_priv *csi2rx = v4l2_subdev_to_csi2rx(subdev);

        csi2rx->source_pad = media_entity_get_fwnode_pad(&s_subdev->entity,
                                                         asd->match.fwnode,
                                                         MEDIA_PAD_FL_SOURCE);
        if (csi2rx->source_pad < 0) {
                dev_err(csi2rx->dev, "Couldn't find output pad for subdev %s\n",
                        s_subdev->name);
                return csi2rx->source_pad;
        }

        csi2rx->source_subdev = s_subdev;

        dev_dbg(csi2rx->dev, "Bound %s pad: %d\n", s_subdev->name,
                csi2rx->source_pad);

        return media_create_pad_link(&csi2rx->source_subdev->entity,
                                     csi2rx->source_pad,
                                     &csi2rx->subdev.entity, 0,
                                     MEDIA_LNK_FL_ENABLED |
                                     MEDIA_LNK_FL_IMMUTABLE);
}

static const struct v4l2_async_notifier_operations csi2rx_notifier_ops = {
        .bound          = csi2rx_async_bound,
};

static int csi2rx_get_resources(struct csi2rx_priv *csi2rx,
                                struct platform_device *pdev)
{
        unsigned char i;
        u32 dev_cfg;
        int ret;

        csi2rx->base = devm_platform_ioremap_resource(pdev, 0);
        if (IS_ERR(csi2rx->base))
                return PTR_ERR(csi2rx->base);

        csi2rx->sys_clk = devm_clk_get(&pdev->dev, "sys_clk");
        if (IS_ERR(csi2rx->sys_clk)) {
                dev_err(&pdev->dev, "Couldn't get sys clock\n");
                return PTR_ERR(csi2rx->sys_clk);
        }

        csi2rx->p_clk = devm_clk_get(&pdev->dev, "p_clk");
        if (IS_ERR(csi2rx->p_clk)) {
                dev_err(&pdev->dev, "Couldn't get P clock\n");
                return PTR_ERR(csi2rx->p_clk);
        }

        csi2rx->sys_rst = devm_reset_control_get_optional_exclusive(&pdev->dev,
                                                                    "sys");
        if (IS_ERR(csi2rx->sys_rst))
                return PTR_ERR(csi2rx->sys_rst);

        csi2rx->p_rst = devm_reset_control_get_optional_exclusive(&pdev->dev,
                                                                  "reg_bank");
        if (IS_ERR(csi2rx->p_rst))
                return PTR_ERR(csi2rx->p_rst);

        csi2rx->dphy = devm_phy_optional_get(&pdev->dev, "dphy");
        if (IS_ERR(csi2rx->dphy)) {
                dev_err(&pdev->dev, "Couldn't get external D-PHY\n");
                return PTR_ERR(csi2rx->dphy);
        }

        ret = clk_prepare_enable(csi2rx->p_clk);
        if (ret) {
                dev_err(&pdev->dev, "Couldn't prepare and enable P clock\n");
                return ret;
        }

        dev_cfg = readl(csi2rx->base + CSI2RX_DEVICE_CFG_REG);
        clk_disable_unprepare(csi2rx->p_clk);

        csi2rx->max_lanes = dev_cfg & 7;
        if (csi2rx->max_lanes > CSI2RX_LANES_MAX) {
                dev_err(&pdev->dev, "Invalid number of lanes: %u\n",
                        csi2rx->max_lanes);
                return -EINVAL;
        }

        csi2rx->max_streams = (dev_cfg >> 4) & 7;
        if (csi2rx->max_streams > CSI2RX_STREAMS_MAX) {
                dev_err(&pdev->dev, "Invalid number of streams: %u\n",
                        csi2rx->max_streams);
                return -EINVAL;
        }

        csi2rx->has_internal_dphy = dev_cfg & BIT(3) ? true : false;

        /*
         * FIXME: Once we'll have internal D-PHY support, the check
         * will need to be removed.
         */
        if (!csi2rx->dphy && csi2rx->has_internal_dphy) {
                dev_err(&pdev->dev, "Internal D-PHY not supported yet\n");
                return -EINVAL;
        }

        for (i = 0; i < csi2rx->max_streams; i++) {
                char name[16];

                snprintf(name, sizeof(name), "pixel_if%u_clk", i);
                csi2rx->pixel_clk[i] = devm_clk_get(&pdev->dev, name);
                if (IS_ERR(csi2rx->pixel_clk[i])) {
                        dev_err(&pdev->dev, "Couldn't get clock %s\n", name);
                        return PTR_ERR(csi2rx->pixel_clk[i]);
                }

                snprintf(name, sizeof(name), "pixel_if%u", i);
                csi2rx->pixel_rst[i] =
                        devm_reset_control_get_optional_exclusive(&pdev->dev,
                                                                  name);
                if (IS_ERR(csi2rx->pixel_rst[i]))
                        return PTR_ERR(csi2rx->pixel_rst[i]);
        }

        return 0;
}

static int csi2rx_parse_dt(struct csi2rx_priv *csi2rx)
{
        struct v4l2_fwnode_endpoint v4l2_ep = { .bus_type = 0 };
        struct v4l2_async_connection *asd;
        struct fwnode_handle *fwh;
        struct device_node *ep;
        int ret;

        ep = of_graph_get_endpoint_by_regs(csi2rx->dev->of_node, 0, 0);
        if (!ep)
                return -EINVAL;

        fwh = of_fwnode_handle(ep);
        ret = v4l2_fwnode_endpoint_parse(fwh, &v4l2_ep);
        if (ret) {
                dev_err(csi2rx->dev, "Could not parse v4l2 endpoint\n");
                of_node_put(ep);
                return ret;
        }

        if (v4l2_ep.bus_type != V4L2_MBUS_CSI2_DPHY) {
                dev_err(csi2rx->dev, "Unsupported media bus type: 0x%x\n",
                        v4l2_ep.bus_type);
                of_node_put(ep);
                return -EINVAL;
        }

        memcpy(csi2rx->lanes, v4l2_ep.bus.mipi_csi2.data_lanes,
               sizeof(csi2rx->lanes));
        csi2rx->num_lanes = v4l2_ep.bus.mipi_csi2.num_data_lanes;
        if (csi2rx->num_lanes > csi2rx->max_lanes) {
                dev_err(csi2rx->dev, "Unsupported number of data-lanes: %d\n",
                        csi2rx->num_lanes);
                of_node_put(ep);
                return -EINVAL;
        }

        v4l2_async_subdev_nf_init(&csi2rx->notifier, &csi2rx->subdev);

        asd = v4l2_async_nf_add_fwnode_remote(&csi2rx->notifier, fwh,
                                              struct v4l2_async_connection);
        of_node_put(ep);
        if (IS_ERR(asd)) {
                v4l2_async_nf_cleanup(&csi2rx->notifier);
                return PTR_ERR(asd);
        }

        csi2rx->notifier.ops = &csi2rx_notifier_ops;

        ret = v4l2_async_nf_register(&csi2rx->notifier);
        if (ret)
                v4l2_async_nf_cleanup(&csi2rx->notifier);

        return ret;
}

static int csi2rx_probe(struct platform_device *pdev)
{
        struct csi2rx_priv *csi2rx;
        unsigned int i;
        int ret;

        csi2rx = kzalloc_obj(*csi2rx);
        if (!csi2rx)
                return -ENOMEM;
        platform_set_drvdata(pdev, csi2rx);
        csi2rx->dev = &pdev->dev;
        mutex_init(&csi2rx->lock);

        ret = csi2rx_get_resources(csi2rx, pdev);
        if (ret)
                goto err_free_priv;

        ret = csi2rx_parse_dt(csi2rx);
        if (ret)
                goto err_free_priv;

        csi2rx->subdev.owner = THIS_MODULE;
        csi2rx->subdev.dev = &pdev->dev;
        v4l2_subdev_init(&csi2rx->subdev, &csi2rx_subdev_ops);
        csi2rx->subdev.internal_ops = &csi2rx_internal_ops;
        v4l2_set_subdevdata(&csi2rx->subdev, &pdev->dev);
        snprintf(csi2rx->subdev.name, sizeof(csi2rx->subdev.name),
                 "%s.%s", KBUILD_MODNAME, dev_name(&pdev->dev));

        /* Create our media pads */
        csi2rx->subdev.entity.function = MEDIA_ENT_F_VID_IF_BRIDGE;
        csi2rx->pads[CSI2RX_PAD_SINK].flags = MEDIA_PAD_FL_SINK;
        for (i = CSI2RX_PAD_SOURCE_STREAM0; i < CSI2RX_PAD_MAX; i++)
                csi2rx->pads[i].flags = MEDIA_PAD_FL_SOURCE;
        csi2rx->subdev.flags |= V4L2_SUBDEV_FL_HAS_DEVNODE;
        csi2rx->subdev.entity.ops = &csi2rx_media_ops;

        ret = media_entity_pads_init(&csi2rx->subdev.entity, CSI2RX_PAD_MAX,
                                     csi2rx->pads);
        if (ret)
                goto err_cleanup;

        csi2rx->error_irq = platform_get_irq_byname_optional(pdev, "error_irq");

        if (csi2rx->error_irq < 0) {
                dev_dbg(csi2rx->dev, "Optional interrupt not defined, proceeding without it\n");
        } else {
                ret = devm_request_irq(csi2rx->dev, csi2rx->error_irq,
                                       csi2rx_irq_handler, 0,
                                       dev_name(&pdev->dev), csi2rx);
                if (ret) {
                        dev_err(csi2rx->dev,
                                "Unable to request interrupt: %d\n", ret);
                        goto err_cleanup;
                }
        }

        ret = v4l2_subdev_init_finalize(&csi2rx->subdev);
        if (ret)
                goto err_cleanup;

        ret = v4l2_async_register_subdev(&csi2rx->subdev);
        if (ret < 0)
                goto err_free_state;

        dev_info(&pdev->dev,
                 "Probed CSI2RX with %u/%u lanes, %u streams, %s D-PHY\n",
                 csi2rx->num_lanes, csi2rx->max_lanes, csi2rx->max_streams,
                 csi2rx->dphy ? "external" :
                 csi2rx->has_internal_dphy ? "internal" : "no");

        return 0;

err_free_state:
        v4l2_subdev_cleanup(&csi2rx->subdev);
err_cleanup:
        v4l2_async_nf_unregister(&csi2rx->notifier);
        v4l2_async_nf_cleanup(&csi2rx->notifier);
        media_entity_cleanup(&csi2rx->subdev.entity);
err_free_priv:
        kfree(csi2rx);
        return ret;
}

static void csi2rx_remove(struct platform_device *pdev)
{
        struct csi2rx_priv *csi2rx = platform_get_drvdata(pdev);

        v4l2_async_nf_unregister(&csi2rx->notifier);
        v4l2_async_nf_cleanup(&csi2rx->notifier);
        v4l2_async_unregister_subdev(&csi2rx->subdev);
        v4l2_subdev_cleanup(&csi2rx->subdev);
        media_entity_cleanup(&csi2rx->subdev.entity);
        kfree(csi2rx);
}

static const struct of_device_id csi2rx_of_table[] = {
        { .compatible = "starfive,jh7110-csi2rx" },
        { .compatible = "cdns,csi2rx" },
        { },
};
MODULE_DEVICE_TABLE(of, csi2rx_of_table);

static struct platform_driver csi2rx_driver = {
        .probe  = csi2rx_probe,
        .remove = csi2rx_remove,

        .driver = {
                .name           = "cdns-csi2rx",
                .of_match_table = csi2rx_of_table,
        },
};
module_platform_driver(csi2rx_driver);
MODULE_AUTHOR("Maxime Ripard <maxime.ripard@bootlin.com>");
MODULE_DESCRIPTION("Cadence CSI2-RX controller");
MODULE_LICENSE("GPL");