root/drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c
// SPDX-License-Identifier: GPL-2.0-only
/*
 * Copyright (c) 2015-2018, 2020-2021 The Linux Foundation. All rights reserved.
 */

#define pr_fmt(fmt)     "[drm:%s:%d] " fmt, __func__, __LINE__
#include <linux/delay.h>
#include <linux/string_choices.h>
#include "dpu_encoder_phys.h"
#include "dpu_hw_interrupts.h"
#include "dpu_hw_pingpong.h"
#include "dpu_core_irq.h"
#include "dpu_formats.h"
#include "dpu_trace.h"
#include "disp/msm_disp_snapshot.h"

#include <drm/drm_managed.h>

#define DPU_DEBUG_CMDENC(e, fmt, ...) DPU_DEBUG("enc%d intf%d " fmt, \
                (e) && (e)->base.parent ? \
                (e)->base.parent->base.id : -1, \
                (e) ? (e)->base.hw_intf->idx - INTF_0 : -1, ##__VA_ARGS__)

#define DPU_ERROR_CMDENC(e, fmt, ...) DPU_ERROR("enc%d intf%d " fmt, \
                (e) && (e)->base.parent ? \
                (e)->base.parent->base.id : -1, \
                (e) ? (e)->base.hw_intf->idx - INTF_0 : -1, ##__VA_ARGS__)

#define to_dpu_encoder_phys_cmd(x) \
        container_of(x, struct dpu_encoder_phys_cmd, base)

#define PP_TIMEOUT_MAX_TRIALS   10

/*
 * Tearcheck sync start and continue thresholds are empirically found
 * based on common panels In the future, may want to allow panels to override
 * these default values
 */
#define DEFAULT_TEARCHECK_SYNC_THRESH_START     4
#define DEFAULT_TEARCHECK_SYNC_THRESH_CONTINUE  4

static void dpu_encoder_phys_cmd_enable_te(struct dpu_encoder_phys *phys_enc);

static bool dpu_encoder_phys_cmd_is_master(struct dpu_encoder_phys *phys_enc)
{
        return (phys_enc->split_role != ENC_ROLE_SLAVE);
}

static void _dpu_encoder_phys_cmd_update_intf_cfg(
                struct dpu_encoder_phys *phys_enc)
{
        struct dpu_encoder_phys_cmd *cmd_enc =
                        to_dpu_encoder_phys_cmd(phys_enc);
        struct dpu_hw_ctl *ctl;
        struct dpu_hw_intf_cfg intf_cfg = { 0 };
        struct dpu_hw_intf_cmd_mode_cfg cmd_mode_cfg = {};

        ctl = phys_enc->hw_ctl;
        if (!ctl->ops.setup_intf_cfg)
                return;

        intf_cfg.intf = phys_enc->hw_intf->idx;
        if (phys_enc->split_role == ENC_ROLE_MASTER)
                intf_cfg.intf_master = phys_enc->hw_intf->idx;
        intf_cfg.intf_mode_sel = DPU_CTL_MODE_SEL_CMD;
        intf_cfg.stream_sel = cmd_enc->stream_sel;
        intf_cfg.mode_3d = dpu_encoder_helper_get_3d_blend_mode(phys_enc);
        intf_cfg.dsc = dpu_encoder_helper_get_dsc(phys_enc);
        ctl->ops.setup_intf_cfg(ctl, &intf_cfg);

        /* setup which pp blk will connect to this intf */
        if (phys_enc->dpu_kms->catalog->mdss_ver->core_major_ver >= 5 &&
            phys_enc->hw_intf->ops.bind_pingpong_blk)
                phys_enc->hw_intf->ops.bind_pingpong_blk(
                                phys_enc->hw_intf,
                                phys_enc->hw_pp->idx);

        if (intf_cfg.dsc != 0)
                cmd_mode_cfg.data_compress = true;

        cmd_mode_cfg.wide_bus_en = dpu_encoder_is_widebus_enabled(phys_enc->parent);

        if (phys_enc->hw_intf->ops.program_intf_cmd_cfg)
                phys_enc->hw_intf->ops.program_intf_cmd_cfg(phys_enc->hw_intf, &cmd_mode_cfg);
}

static void dpu_encoder_phys_cmd_pp_tx_done_irq(void *arg)
{
        struct dpu_encoder_phys *phys_enc = arg;
        unsigned long lock_flags;
        int new_cnt;
        u32 event = DPU_ENCODER_FRAME_EVENT_DONE;

        if (!phys_enc->hw_pp)
                return;

        DPU_ATRACE_BEGIN("pp_done_irq");
        /* notify all synchronous clients first, then asynchronous clients */
        dpu_encoder_frame_done_callback(phys_enc->parent, phys_enc, event);

        spin_lock_irqsave(phys_enc->enc_spinlock, lock_flags);
        new_cnt = atomic_add_unless(&phys_enc->pending_kickoff_cnt, -1, 0);
        spin_unlock_irqrestore(phys_enc->enc_spinlock, lock_flags);

        trace_dpu_enc_phys_cmd_pp_tx_done(DRMID(phys_enc->parent),
                                          phys_enc->hw_pp->idx - PINGPONG_0,
                                          new_cnt, event);

        /* Signal any waiting atomic commit thread */
        wake_up_all(&phys_enc->pending_kickoff_wq);
        DPU_ATRACE_END("pp_done_irq");
}

static void dpu_encoder_phys_cmd_te_rd_ptr_irq(void *arg)
{
        struct dpu_encoder_phys *phys_enc = arg;
        struct dpu_encoder_phys_cmd *cmd_enc;

        DPU_ATRACE_BEGIN("rd_ptr_irq");
        cmd_enc = to_dpu_encoder_phys_cmd(phys_enc);

        dpu_encoder_vblank_callback(phys_enc->parent, phys_enc);

        atomic_add_unless(&cmd_enc->pending_vblank_cnt, -1, 0);
        wake_up_all(&cmd_enc->pending_vblank_wq);
        DPU_ATRACE_END("rd_ptr_irq");
}

static void dpu_encoder_phys_cmd_ctl_start_irq(void *arg)
{
        struct dpu_encoder_phys *phys_enc = arg;

        DPU_ATRACE_BEGIN("ctl_start_irq");

        atomic_add_unless(&phys_enc->pending_ctlstart_cnt, -1, 0);

        /* Signal any waiting ctl start interrupt */
        wake_up_all(&phys_enc->pending_kickoff_wq);
        DPU_ATRACE_END("ctl_start_irq");
}

static void dpu_encoder_phys_cmd_underrun_irq(void *arg)
{
        struct dpu_encoder_phys *phys_enc = arg;

        dpu_encoder_underrun_callback(phys_enc->parent, phys_enc);
}

static void dpu_encoder_phys_cmd_atomic_mode_set(
                struct dpu_encoder_phys *phys_enc,
                struct drm_crtc_state *crtc_state,
                struct drm_connector_state *conn_state)
{
        phys_enc->irq[INTR_IDX_CTL_START] = phys_enc->hw_ctl->caps->intr_start;

        phys_enc->irq[INTR_IDX_PINGPONG] = phys_enc->hw_pp->caps->intr_done;

        if (phys_enc->has_intf_te)
                phys_enc->irq[INTR_IDX_RDPTR] = phys_enc->hw_intf->cap->intr_tear_rd_ptr;
        else
                phys_enc->irq[INTR_IDX_RDPTR] = phys_enc->hw_pp->caps->intr_rdptr;

        phys_enc->irq[INTR_IDX_UNDERRUN] = phys_enc->hw_intf->cap->intr_underrun;
}

static int _dpu_encoder_phys_cmd_handle_ppdone_timeout(
                struct dpu_encoder_phys *phys_enc)
{
        struct dpu_encoder_phys_cmd *cmd_enc =
                        to_dpu_encoder_phys_cmd(phys_enc);
        u32 frame_event = DPU_ENCODER_FRAME_EVENT_ERROR;
        bool do_log = false;
        struct drm_encoder *drm_enc;

        if (!phys_enc->hw_pp)
                return -EINVAL;

        drm_enc = phys_enc->parent;

        cmd_enc->pp_timeout_report_cnt++;
        if (cmd_enc->pp_timeout_report_cnt == PP_TIMEOUT_MAX_TRIALS) {
                frame_event |= DPU_ENCODER_FRAME_EVENT_PANEL_DEAD;
                do_log = true;
        } else if (cmd_enc->pp_timeout_report_cnt == 1) {
                do_log = true;
        }

        trace_dpu_enc_phys_cmd_pdone_timeout(DRMID(drm_enc),
                     phys_enc->hw_pp->idx - PINGPONG_0,
                     cmd_enc->pp_timeout_report_cnt,
                     atomic_read(&phys_enc->pending_kickoff_cnt),
                     frame_event);

        /* to avoid flooding, only log first time, and "dead" time */
        if (do_log) {
                DRM_ERROR("id:%d pp:%d kickoff timeout %d cnt %d koff_cnt %d\n",
                          DRMID(drm_enc),
                          phys_enc->hw_pp->idx - PINGPONG_0,
                          phys_enc->hw_ctl->idx - CTL_0,
                          cmd_enc->pp_timeout_report_cnt,
                          atomic_read(&phys_enc->pending_kickoff_cnt));
                msm_disp_snapshot_state(drm_enc->dev);
                dpu_core_irq_unregister_callback(phys_enc->dpu_kms,
                                phys_enc->irq[INTR_IDX_RDPTR]);
        }

        atomic_add_unless(&phys_enc->pending_kickoff_cnt, -1, 0);

        /* request a ctl reset before the next kickoff */
        phys_enc->enable_state = DPU_ENC_ERR_NEEDS_HW_RESET;

        dpu_encoder_frame_done_callback(phys_enc->parent, phys_enc, frame_event);

        return -ETIMEDOUT;
}

static int _dpu_encoder_phys_cmd_wait_for_idle(
                struct dpu_encoder_phys *phys_enc)
{
        struct dpu_encoder_phys_cmd *cmd_enc =
                        to_dpu_encoder_phys_cmd(phys_enc);
        struct dpu_encoder_wait_info wait_info;
        int ret;

        wait_info.wq = &phys_enc->pending_kickoff_wq;
        wait_info.atomic_cnt = &phys_enc->pending_kickoff_cnt;
        wait_info.timeout_ms = KICKOFF_TIMEOUT_MS;

        ret = dpu_encoder_helper_wait_for_irq(phys_enc,
                        phys_enc->irq[INTR_IDX_PINGPONG],
                        dpu_encoder_phys_cmd_pp_tx_done_irq,
                        &wait_info);
        if (ret == -ETIMEDOUT)
                _dpu_encoder_phys_cmd_handle_ppdone_timeout(phys_enc);
        else if (!ret)
                cmd_enc->pp_timeout_report_cnt = 0;

        return ret;
}

static int dpu_encoder_phys_cmd_control_vblank_irq(
                struct dpu_encoder_phys *phys_enc,
                bool enable)
{
        int ret = 0;
        int refcount;

        if (!phys_enc->hw_pp) {
                DPU_ERROR("invalid encoder\n");
                return -EINVAL;
        }

        mutex_lock(&phys_enc->vblank_ctl_lock);
        refcount = phys_enc->vblank_refcount;

        /* Slave encoders don't report vblank */
        if (!dpu_encoder_phys_cmd_is_master(phys_enc))
                goto end;

        /* protect against negative */
        if (!enable && refcount == 0) {
                ret = -EINVAL;
                goto end;
        }

        DRM_DEBUG_KMS("id:%u pp:%d enable=%s/%d\n", DRMID(phys_enc->parent),
                      phys_enc->hw_pp->idx - PINGPONG_0,
                      str_true_false(enable), refcount);

        if (enable) {
                if (phys_enc->vblank_refcount == 0)
                        ret = dpu_core_irq_register_callback(phys_enc->dpu_kms,
                                        phys_enc->irq[INTR_IDX_RDPTR],
                                        dpu_encoder_phys_cmd_te_rd_ptr_irq,
                                        phys_enc);
                if (!ret)
                        phys_enc->vblank_refcount++;
        } else if (!enable) {
                if (phys_enc->vblank_refcount == 1)
                        ret = dpu_core_irq_unregister_callback(phys_enc->dpu_kms,
                                        phys_enc->irq[INTR_IDX_RDPTR]);
                if (!ret)
                        phys_enc->vblank_refcount--;
        }

end:
        mutex_unlock(&phys_enc->vblank_ctl_lock);
        if (ret) {
                DRM_ERROR("vblank irq err id:%u pp:%d ret:%d, enable %s/%d\n",
                          DRMID(phys_enc->parent),
                          phys_enc->hw_pp->idx - PINGPONG_0, ret,
                          str_true_false(enable), refcount);
        }

        return ret;
}

static void dpu_encoder_phys_cmd_irq_enable(struct dpu_encoder_phys *phys_enc)
{
        trace_dpu_enc_phys_cmd_irq_enable(DRMID(phys_enc->parent),
                                          phys_enc->hw_pp->idx - PINGPONG_0,
                                          phys_enc->vblank_refcount);

        dpu_core_irq_register_callback(phys_enc->dpu_kms,
                                       phys_enc->irq[INTR_IDX_PINGPONG],
                                       dpu_encoder_phys_cmd_pp_tx_done_irq,
                                       phys_enc);
        dpu_core_irq_register_callback(phys_enc->dpu_kms,
                                       phys_enc->irq[INTR_IDX_UNDERRUN],
                                       dpu_encoder_phys_cmd_underrun_irq,
                                       phys_enc);
        dpu_encoder_phys_cmd_control_vblank_irq(phys_enc, true);

        if (dpu_encoder_phys_cmd_is_master(phys_enc) && phys_enc->irq[INTR_IDX_CTL_START])
                dpu_core_irq_register_callback(phys_enc->dpu_kms,
                                               phys_enc->irq[INTR_IDX_CTL_START],
                                               dpu_encoder_phys_cmd_ctl_start_irq,
                                               phys_enc);
}

static void dpu_encoder_phys_cmd_irq_disable(struct dpu_encoder_phys *phys_enc)
{
        trace_dpu_enc_phys_cmd_irq_disable(DRMID(phys_enc->parent),
                                           phys_enc->hw_pp->idx - PINGPONG_0,
                                           phys_enc->vblank_refcount);

        if (dpu_encoder_phys_cmd_is_master(phys_enc) && phys_enc->irq[INTR_IDX_CTL_START])
                dpu_core_irq_unregister_callback(phys_enc->dpu_kms,
                                                 phys_enc->irq[INTR_IDX_CTL_START]);

        dpu_core_irq_unregister_callback(phys_enc->dpu_kms, phys_enc->irq[INTR_IDX_UNDERRUN]);
        dpu_encoder_phys_cmd_control_vblank_irq(phys_enc, false);
        dpu_core_irq_unregister_callback(phys_enc->dpu_kms, phys_enc->irq[INTR_IDX_PINGPONG]);
}

static void dpu_encoder_phys_cmd_tearcheck_config(
                struct dpu_encoder_phys *phys_enc)
{
        struct dpu_encoder_phys_cmd *cmd_enc =
                to_dpu_encoder_phys_cmd(phys_enc);
        struct dpu_hw_tear_check tc_cfg = { 0 };
        struct drm_display_mode *mode;
        bool tc_enable = true;
        unsigned long vsync_hz;
        struct dpu_kms *dpu_kms;

        /*
         * TODO: if/when resource allocation is refactored, move this to a
         * place where the driver can actually return an error.
         */
        if (!phys_enc->has_intf_te &&
            (!phys_enc->hw_pp ||
             !phys_enc->hw_pp->ops.enable_tearcheck)) {
                DPU_DEBUG_CMDENC(cmd_enc, "tearcheck not supported\n");
                return;
        }

        DPU_DEBUG_CMDENC(cmd_enc, "intf %d pp %d\n",
                         phys_enc->hw_intf ? phys_enc->hw_intf->idx - INTF_0 : -1,
                         phys_enc->hw_pp ? phys_enc->hw_pp->idx - PINGPONG_0 : -1);

        mode = &phys_enc->cached_mode;

        dpu_kms = phys_enc->dpu_kms;

        /*
         * TE default: dsi byte clock calculated base on 70 fps;
         * around 14 ms to complete a kickoff cycle if te disabled;
         * vclk_line base on 60 fps; write is faster than read;
         * init == start == rdptr;
         *
         * vsync_count is ratio of MDP VSYNC clock frequency to LCD panel
         * frequency divided by the no. of rows (lines) in the LCDpanel.
         */
        vsync_hz = dpu_kms_get_clk_rate(dpu_kms, "vsync");
        if (!vsync_hz) {
                DPU_DEBUG_CMDENC(cmd_enc, "invalid - no vsync clock\n");
                return;
        }

        tc_cfg.vsync_count = vsync_hz /
                                (mode->vtotal * drm_mode_vrefresh(mode));

        /*
         * Set the sync_cfg_height to twice vtotal so that if we lose a
         * TE event coming from the display TE pin we won't stall immediately
         */
        tc_cfg.hw_vsync_mode = 1;
        tc_cfg.sync_cfg_height = mode->vtotal * 2;
        tc_cfg.vsync_init_val = mode->vdisplay;
        tc_cfg.sync_threshold_start = DEFAULT_TEARCHECK_SYNC_THRESH_START;
        tc_cfg.sync_threshold_continue = DEFAULT_TEARCHECK_SYNC_THRESH_CONTINUE;
        tc_cfg.start_pos = mode->vdisplay;
        tc_cfg.rd_ptr_irq = mode->vdisplay + 1;

        DPU_DEBUG_CMDENC(cmd_enc,
                "tc vsync_clk_speed_hz %lu vtotal %u vrefresh %u\n",
                vsync_hz, mode->vtotal, drm_mode_vrefresh(mode));
        DPU_DEBUG_CMDENC(cmd_enc,
                "tc enable %u start_pos %u rd_ptr_irq %u\n",
                tc_enable, tc_cfg.start_pos, tc_cfg.rd_ptr_irq);
        DPU_DEBUG_CMDENC(cmd_enc,
                "tc hw_vsync_mode %u vsync_count %u vsync_init_val %u\n",
                tc_cfg.hw_vsync_mode, tc_cfg.vsync_count,
                tc_cfg.vsync_init_val);
        DPU_DEBUG_CMDENC(cmd_enc,
                "tc cfgheight %u thresh_start %u thresh_cont %u\n",
                tc_cfg.sync_cfg_height, tc_cfg.sync_threshold_start,
                tc_cfg.sync_threshold_continue);

        if (phys_enc->has_intf_te)
                phys_enc->hw_intf->ops.enable_tearcheck(phys_enc->hw_intf, &tc_cfg);
        else
                phys_enc->hw_pp->ops.enable_tearcheck(phys_enc->hw_pp, &tc_cfg);
}

static void _dpu_encoder_phys_cmd_pingpong_config(
                struct dpu_encoder_phys *phys_enc)
{
        struct dpu_encoder_phys_cmd *cmd_enc =
                to_dpu_encoder_phys_cmd(phys_enc);

        if (!phys_enc->hw_pp || !phys_enc->hw_ctl->ops.setup_intf_cfg) {
                DPU_ERROR("invalid arg(s), enc %d\n", phys_enc != NULL);
                return;
        }

        DPU_DEBUG_CMDENC(cmd_enc, "pp %d, enabling mode:\n",
                        phys_enc->hw_pp->idx - PINGPONG_0);
        drm_mode_debug_printmodeline(&phys_enc->cached_mode);

        _dpu_encoder_phys_cmd_update_intf_cfg(phys_enc);
        dpu_encoder_phys_cmd_tearcheck_config(phys_enc);
}

static bool dpu_encoder_phys_cmd_needs_single_flush(
                struct dpu_encoder_phys *phys_enc)
{
        /**
         * we do separate flush for each CTL and let
         * CTL_START synchronize them
         */
        return false;
}

static void dpu_encoder_phys_cmd_enable_helper(
                struct dpu_encoder_phys *phys_enc)
{
        struct dpu_hw_ctl *ctl;

        if (!phys_enc->hw_pp) {
                DPU_ERROR("invalid arg(s), encoder %d\n", phys_enc != NULL);
                return;
        }

        dpu_encoder_helper_split_config(phys_enc, phys_enc->hw_intf->idx);

        _dpu_encoder_phys_cmd_pingpong_config(phys_enc);

        ctl = phys_enc->hw_ctl;
        ctl->ops.update_pending_flush_intf(ctl, phys_enc->hw_intf->idx);
}

static void dpu_encoder_phys_cmd_enable(struct dpu_encoder_phys *phys_enc)
{
        struct dpu_encoder_phys_cmd *cmd_enc =
                to_dpu_encoder_phys_cmd(phys_enc);

        if (!phys_enc->hw_pp) {
                DPU_ERROR("invalid phys encoder\n");
                return;
        }

        DPU_DEBUG_CMDENC(cmd_enc, "pp %d\n", phys_enc->hw_pp->idx - PINGPONG_0);

        if (phys_enc->enable_state == DPU_ENC_ENABLED) {
                DPU_ERROR("already enabled\n");
                return;
        }

        dpu_encoder_phys_cmd_enable_helper(phys_enc);
        phys_enc->enable_state = DPU_ENC_ENABLED;
}

static void _dpu_encoder_phys_cmd_connect_te(
                struct dpu_encoder_phys *phys_enc, bool enable)
{
        if (phys_enc->has_intf_te) {
                if (!phys_enc->hw_intf || !phys_enc->hw_intf->ops.connect_external_te)
                        return;

                trace_dpu_enc_phys_cmd_connect_te(DRMID(phys_enc->parent), enable);
                phys_enc->hw_intf->ops.connect_external_te(phys_enc->hw_intf, enable);
        } else {
                if (!phys_enc->hw_pp || !phys_enc->hw_pp->ops.connect_external_te)
                        return;

                trace_dpu_enc_phys_cmd_connect_te(DRMID(phys_enc->parent), enable);
                phys_enc->hw_pp->ops.connect_external_te(phys_enc->hw_pp, enable);
        }
}

static void dpu_encoder_phys_cmd_prepare_idle_pc(
                struct dpu_encoder_phys *phys_enc)
{
        _dpu_encoder_phys_cmd_connect_te(phys_enc, false);
}

static int dpu_encoder_phys_cmd_get_line_count(
                struct dpu_encoder_phys *phys_enc)
{
        struct dpu_hw_pingpong *hw_pp;
        struct dpu_hw_intf *hw_intf;

        if (!dpu_encoder_phys_cmd_is_master(phys_enc))
                return -EINVAL;

        if (phys_enc->has_intf_te) {
                hw_intf = phys_enc->hw_intf;
                if (!hw_intf || !hw_intf->ops.get_line_count)
                        return -EINVAL;
                return hw_intf->ops.get_line_count(hw_intf);
        }

        hw_pp = phys_enc->hw_pp;
        if (!hw_pp || !hw_pp->ops.get_line_count)
                return -EINVAL;
        return hw_pp->ops.get_line_count(hw_pp);
}

static void dpu_encoder_phys_cmd_disable(struct dpu_encoder_phys *phys_enc)
{
        struct dpu_encoder_phys_cmd *cmd_enc =
                to_dpu_encoder_phys_cmd(phys_enc);
        struct dpu_hw_ctl *ctl;

        if (phys_enc->enable_state == DPU_ENC_DISABLED) {
                DPU_ERROR_CMDENC(cmd_enc, "already disabled\n");
                return;
        }

        if (phys_enc->has_intf_te) {
                DRM_DEBUG_KMS("id:%u intf:%d state:%d\n", DRMID(phys_enc->parent),
                              phys_enc->hw_intf->idx - INTF_0,
                              phys_enc->enable_state);

                if (phys_enc->hw_intf->ops.disable_tearcheck)
                        phys_enc->hw_intf->ops.disable_tearcheck(phys_enc->hw_intf);
        } else {
                if (!phys_enc->hw_pp) {
                        DPU_ERROR("invalid encoder\n");
                        return;
                }

                DRM_DEBUG_KMS("id:%u pp:%d state:%d\n", DRMID(phys_enc->parent),
                              phys_enc->hw_pp->idx - PINGPONG_0,
                              phys_enc->enable_state);

                if (phys_enc->hw_pp->ops.disable_tearcheck)
                        phys_enc->hw_pp->ops.disable_tearcheck(phys_enc->hw_pp);
        }

        if (phys_enc->hw_intf->ops.bind_pingpong_blk) {
                phys_enc->hw_intf->ops.bind_pingpong_blk(
                                phys_enc->hw_intf,
                                PINGPONG_NONE);

                ctl = phys_enc->hw_ctl;
                ctl->ops.update_pending_flush_intf(ctl, phys_enc->hw_intf->idx);
        }

        phys_enc->enable_state = DPU_ENC_DISABLED;
}

static void dpu_encoder_phys_cmd_prepare_for_kickoff(
                struct dpu_encoder_phys *phys_enc)
{
        struct dpu_encoder_phys_cmd *cmd_enc =
                        to_dpu_encoder_phys_cmd(phys_enc);
        int ret;

        if (!phys_enc->hw_pp) {
                DPU_ERROR("invalid encoder\n");
                return;
        }
        DRM_DEBUG_KMS("id:%u pp:%d pending_cnt:%d\n", DRMID(phys_enc->parent),
                      phys_enc->hw_pp->idx - PINGPONG_0,
                      atomic_read(&phys_enc->pending_kickoff_cnt));

        /*
         * Mark kickoff request as outstanding. If there are more than one,
         * outstanding, then we have to wait for the previous one to complete
         */
        ret = _dpu_encoder_phys_cmd_wait_for_idle(phys_enc);
        if (ret) {
                /* force pending_kickoff_cnt 0 to discard failed kickoff */
                atomic_set(&phys_enc->pending_kickoff_cnt, 0);
                DRM_ERROR("failed wait_for_idle: id:%u ret:%d pp:%d\n",
                          DRMID(phys_enc->parent), ret,
                          phys_enc->hw_pp->idx - PINGPONG_0);
        }

        dpu_encoder_phys_cmd_enable_te(phys_enc);

        DPU_DEBUG_CMDENC(cmd_enc, "pp:%d pending_cnt %d\n",
                        phys_enc->hw_pp->idx - PINGPONG_0,
                        atomic_read(&phys_enc->pending_kickoff_cnt));
}

static void dpu_encoder_phys_cmd_enable_te(struct dpu_encoder_phys *phys_enc)
{
        if (!phys_enc)
                return;
        if (!dpu_encoder_phys_cmd_is_master(phys_enc))
                return;

        if (phys_enc->has_intf_te) {
                if (!phys_enc->hw_intf->ops.disable_autorefresh)
                        return;

                phys_enc->hw_intf->ops.disable_autorefresh(
                                phys_enc->hw_intf,
                                DRMID(phys_enc->parent),
                                phys_enc->cached_mode.vdisplay);
        } else {
                if (!phys_enc->hw_pp ||
                    !phys_enc->hw_pp->ops.disable_autorefresh)
                        return;

                phys_enc->hw_pp->ops.disable_autorefresh(
                                phys_enc->hw_pp,
                                DRMID(phys_enc->parent),
                                phys_enc->cached_mode.vdisplay);
        }
}

static int _dpu_encoder_phys_cmd_wait_for_ctl_start(
                struct dpu_encoder_phys *phys_enc)
{
        struct dpu_encoder_phys_cmd *cmd_enc =
                        to_dpu_encoder_phys_cmd(phys_enc);
        struct dpu_encoder_wait_info wait_info;
        int ret;

        wait_info.wq = &phys_enc->pending_kickoff_wq;
        wait_info.atomic_cnt = &phys_enc->pending_ctlstart_cnt;
        wait_info.timeout_ms = KICKOFF_TIMEOUT_MS;

        ret = dpu_encoder_helper_wait_for_irq(phys_enc,
                        phys_enc->irq[INTR_IDX_CTL_START],
                        dpu_encoder_phys_cmd_ctl_start_irq,
                        &wait_info);
        if (ret == -ETIMEDOUT) {
                DPU_ERROR_CMDENC(cmd_enc, "ctl start interrupt wait failed\n");
                ret = -EINVAL;
        } else if (!ret)
                ret = 0;

        return ret;
}

static int dpu_encoder_phys_cmd_wait_for_tx_complete(
                struct dpu_encoder_phys *phys_enc)
{
        int rc;

        rc = _dpu_encoder_phys_cmd_wait_for_idle(phys_enc);
        if (rc) {
                DRM_ERROR("failed wait_for_idle: id:%u ret:%d intf:%d\n",
                          DRMID(phys_enc->parent), rc,
                          phys_enc->hw_intf->idx - INTF_0);
        }

        return rc;
}

static int dpu_encoder_phys_cmd_wait_for_commit_done(
                struct dpu_encoder_phys *phys_enc)
{
        /* only required for master controller */
        if (!dpu_encoder_phys_cmd_is_master(phys_enc))
                return 0;

        if (phys_enc->irq[INTR_IDX_CTL_START] &&
            !phys_enc->hw_ctl->ops.is_started(phys_enc->hw_ctl))
                return _dpu_encoder_phys_cmd_wait_for_ctl_start(phys_enc);

        return dpu_encoder_phys_cmd_wait_for_tx_complete(phys_enc);
}

static void dpu_encoder_phys_cmd_handle_post_kickoff(
                struct dpu_encoder_phys *phys_enc)
{
        /**
         * re-enable external TE, either for the first time after enabling
         * or if disabled for Autorefresh
         */
        _dpu_encoder_phys_cmd_connect_te(phys_enc, true);
}

static void dpu_encoder_phys_cmd_trigger_start(
                struct dpu_encoder_phys *phys_enc)
{
        dpu_encoder_helper_trigger_start(phys_enc);
}

static void dpu_encoder_phys_cmd_init_ops(
                struct dpu_encoder_phys_ops *ops)
{
        ops->is_master = dpu_encoder_phys_cmd_is_master;
        ops->atomic_mode_set = dpu_encoder_phys_cmd_atomic_mode_set;
        ops->enable = dpu_encoder_phys_cmd_enable;
        ops->disable = dpu_encoder_phys_cmd_disable;
        ops->control_vblank_irq = dpu_encoder_phys_cmd_control_vblank_irq;
        ops->wait_for_commit_done = dpu_encoder_phys_cmd_wait_for_commit_done;
        ops->prepare_for_kickoff = dpu_encoder_phys_cmd_prepare_for_kickoff;
        ops->wait_for_tx_complete = dpu_encoder_phys_cmd_wait_for_tx_complete;
        ops->trigger_start = dpu_encoder_phys_cmd_trigger_start;
        ops->needs_single_flush = dpu_encoder_phys_cmd_needs_single_flush;
        ops->irq_enable = dpu_encoder_phys_cmd_irq_enable;
        ops->irq_disable = dpu_encoder_phys_cmd_irq_disable;
        ops->restore = dpu_encoder_phys_cmd_enable_helper;
        ops->prepare_idle_pc = dpu_encoder_phys_cmd_prepare_idle_pc;
        ops->handle_post_kickoff = dpu_encoder_phys_cmd_handle_post_kickoff;
        ops->get_line_count = dpu_encoder_phys_cmd_get_line_count;
}

/**
 * dpu_encoder_phys_cmd_init - Construct a new command mode physical encoder
 * @dev:  Corresponding device for devres management
 * @p:  Pointer to init params structure
 * Return: Error code or newly allocated encoder
 */
struct dpu_encoder_phys *dpu_encoder_phys_cmd_init(struct drm_device *dev,
                struct dpu_enc_phys_init_params *p)
{
        struct dpu_encoder_phys *phys_enc = NULL;
        struct dpu_encoder_phys_cmd *cmd_enc = NULL;

        DPU_DEBUG("intf\n");

        cmd_enc = drmm_kzalloc(dev, sizeof(*cmd_enc), GFP_KERNEL);
        if (!cmd_enc) {
                DPU_ERROR("failed to allocate\n");
                return ERR_PTR(-ENOMEM);
        }
        phys_enc = &cmd_enc->base;

        dpu_encoder_phys_init(phys_enc, p);

        mutex_init(&phys_enc->vblank_ctl_lock);
        phys_enc->vblank_refcount = 0;

        dpu_encoder_phys_cmd_init_ops(&phys_enc->ops);
        phys_enc->intf_mode = INTF_MODE_CMD;
        cmd_enc->stream_sel = 0;

        if (!phys_enc->hw_intf) {
                DPU_ERROR_CMDENC(cmd_enc, "no INTF provided\n");
                return ERR_PTR(-EINVAL);
        }

        /* DPU before 5.0 use PINGPONG for TE handling */
        if (phys_enc->dpu_kms->catalog->mdss_ver->core_major_ver >= 5)
                phys_enc->has_intf_te = true;

        if (phys_enc->has_intf_te && !phys_enc->hw_intf->ops.enable_tearcheck) {
                DPU_ERROR_CMDENC(cmd_enc, "tearcheck not supported\n");
                return ERR_PTR(-EINVAL);
        }

        atomic_set(&cmd_enc->pending_vblank_cnt, 0);
        init_waitqueue_head(&cmd_enc->pending_vblank_wq);

        DPU_DEBUG_CMDENC(cmd_enc, "created\n");

        return phys_enc;
}