root/drivers/media/platform/verisilicon/hantro_g2_vp9_dec.c
// SPDX-License-Identifier: GPL-2.0
/*
 * Hantro VP9 codec driver
 *
 * Copyright (C) 2021 Collabora Ltd.
 */
#include "media/videobuf2-core.h"
#include "media/videobuf2-dma-contig.h"
#include "media/videobuf2-v4l2.h"
#include <linux/kernel.h>
#include <linux/vmalloc.h>
#include <media/v4l2-mem2mem.h>
#include <media/v4l2-vp9.h>

#include "hantro.h"
#include "hantro_vp9.h"
#include "hantro_g2_regs.h"

enum hantro_ref_frames {
        INTRA_FRAME = 0,
        LAST_FRAME = 1,
        GOLDEN_FRAME = 2,
        ALTREF_FRAME = 3,
        MAX_REF_FRAMES = 4
};

static int start_prepare_run(struct hantro_ctx *ctx, const struct v4l2_ctrl_vp9_frame **dec_params)
{
        const struct v4l2_ctrl_vp9_compressed_hdr *prob_updates;
        struct hantro_vp9_dec_hw_ctx *vp9_ctx = &ctx->vp9_dec;
        struct v4l2_ctrl *ctrl;
        unsigned int fctx_idx;

        /* v4l2-specific stuff */
        hantro_start_prepare_run(ctx);

        ctrl = v4l2_ctrl_find(&ctx->ctrl_handler, V4L2_CID_STATELESS_VP9_FRAME);
        if (WARN_ON(!ctrl))
                return -EINVAL;
        *dec_params = ctrl->p_cur.p;

        ctrl = v4l2_ctrl_find(&ctx->ctrl_handler, V4L2_CID_STATELESS_VP9_COMPRESSED_HDR);
        if (WARN_ON(!ctrl))
                return -EINVAL;
        prob_updates = ctrl->p_cur.p;
        vp9_ctx->cur.tx_mode = prob_updates->tx_mode;

        /*
         * vp9 stuff
         *
         * by this point the userspace has done all parts of 6.2 uncompressed_header()
         * except this fragment:
         * if ( FrameIsIntra || error_resilient_mode ) {
         *      setup_past_independence ( )
         *      if ( frame_type == KEY_FRAME || error_resilient_mode == 1 ||
         *           reset_frame_context == 3 ) {
         *              for ( i = 0; i < 4; i ++ ) {
         *                      save_probs( i )
         *              }
         *      } else if ( reset_frame_context == 2 ) {
         *              save_probs( frame_context_idx )
         *      }
         *      frame_context_idx = 0
         * }
         */
        fctx_idx = v4l2_vp9_reset_frame_ctx(*dec_params, vp9_ctx->frame_context);
        vp9_ctx->cur.frame_context_idx = fctx_idx;

        /* 6.1 frame(sz): load_probs() and load_probs2() */
        vp9_ctx->probability_tables = vp9_ctx->frame_context[fctx_idx];

        /*
         * The userspace has also performed 6.3 compressed_header(), but handling the
         * probs in a special way. All probs which need updating, except MV-related,
         * have been read from the bitstream and translated through inv_map_table[],
         * but no 6.3.6 inv_recenter_nonneg(v, m) has been performed. The values passed
         * by userspace are either translated values (there are no 0 values in
         * inv_map_table[]), or zero to indicate no update. All MV-related probs which need
         * updating have been read from the bitstream and (mv_prob << 1) | 1 has been
         * performed. The values passed by userspace are either new values
         * to replace old ones (the above mentioned shift and bitwise or never result in
         * a zero) or zero to indicate no update.
         * fw_update_probs() performs actual probs updates or leaves probs as-is
         * for values for which a zero was passed from userspace.
         */
        v4l2_vp9_fw_update_probs(&vp9_ctx->probability_tables, prob_updates, *dec_params);

        return 0;
}

static struct hantro_decoded_buffer *
get_ref_buf(struct hantro_ctx *ctx, struct vb2_v4l2_buffer *dst, u64 timestamp)
{
        struct v4l2_m2m_ctx *m2m_ctx = ctx->fh.m2m_ctx;
        struct vb2_queue *cap_q = &m2m_ctx->cap_q_ctx.q;
        struct vb2_buffer *buf;

        /*
         * If a ref is unused or invalid, address of current destination
         * buffer is returned.
         */
        buf = vb2_find_buffer(cap_q, timestamp);
        if (!buf)
                buf = &dst->vb2_buf;

        return vb2_to_hantro_decoded_buf(buf);
}

static void update_dec_buf_info(struct hantro_decoded_buffer *buf,
                                const struct v4l2_ctrl_vp9_frame *dec_params)
{
        buf->vp9.width = dec_params->frame_width_minus_1 + 1;
        buf->vp9.height = dec_params->frame_height_minus_1 + 1;
        buf->vp9.bit_depth = dec_params->bit_depth;
}

static void update_ctx_cur_info(struct hantro_vp9_dec_hw_ctx *vp9_ctx,
                                struct hantro_decoded_buffer *buf,
                                const struct v4l2_ctrl_vp9_frame *dec_params)
{
        vp9_ctx->cur.valid = true;
        vp9_ctx->cur.reference_mode = dec_params->reference_mode;
        vp9_ctx->cur.interpolation_filter = dec_params->interpolation_filter;
        vp9_ctx->cur.flags = dec_params->flags;
        vp9_ctx->cur.timestamp = buf->base.vb.vb2_buf.timestamp;
}

static void config_output(struct hantro_ctx *ctx,
                          struct hantro_decoded_buffer *dst,
                          const struct v4l2_ctrl_vp9_frame *dec_params)
{
        dma_addr_t luma_addr, chroma_addr, mv_addr;

        hantro_reg_write(ctx->dev, &g2_out_dis, 0);
        if (!ctx->dev->variant->legacy_regs)
                hantro_reg_write(ctx->dev, &g2_output_format, 0);

        luma_addr = hantro_get_dec_buf_addr(ctx, &dst->base.vb.vb2_buf);
        hantro_write_addr(ctx->dev, G2_OUT_LUMA_ADDR, luma_addr);

        chroma_addr = luma_addr + hantro_g2_chroma_offset(ctx);
        hantro_write_addr(ctx->dev, G2_OUT_CHROMA_ADDR, chroma_addr);
        dst->vp9.chroma_offset = hantro_g2_chroma_offset(ctx);

        mv_addr = luma_addr + hantro_g2_motion_vectors_offset(ctx);
        hantro_write_addr(ctx->dev, G2_OUT_MV_ADDR, mv_addr);
        dst->vp9.mv_offset = hantro_g2_motion_vectors_offset(ctx);
}

struct hantro_vp9_ref_reg {
        const struct hantro_reg width;
        const struct hantro_reg height;
        const struct hantro_reg hor_scale;
        const struct hantro_reg ver_scale;
        u32 y_base;
        u32 c_base;
};

static void config_ref(struct hantro_ctx *ctx,
                       struct hantro_decoded_buffer *dst,
                       const struct hantro_vp9_ref_reg *ref_reg,
                       const struct v4l2_ctrl_vp9_frame *dec_params,
                       u64 ref_ts)
{
        struct hantro_decoded_buffer *buf;
        dma_addr_t luma_addr, chroma_addr;
        u32 refw, refh;

        buf = get_ref_buf(ctx, &dst->base.vb, ref_ts);
        refw = buf->vp9.width;
        refh = buf->vp9.height;

        hantro_reg_write(ctx->dev, &ref_reg->width, refw);
        hantro_reg_write(ctx->dev, &ref_reg->height, refh);

        hantro_reg_write(ctx->dev, &ref_reg->hor_scale, (refw << 14) / dst->vp9.width);
        hantro_reg_write(ctx->dev, &ref_reg->ver_scale, (refh << 14) / dst->vp9.height);

        luma_addr = hantro_get_dec_buf_addr(ctx, &buf->base.vb.vb2_buf);
        hantro_write_addr(ctx->dev, ref_reg->y_base, luma_addr);

        chroma_addr = luma_addr + buf->vp9.chroma_offset;
        hantro_write_addr(ctx->dev, ref_reg->c_base, chroma_addr);
}

static void config_ref_registers(struct hantro_ctx *ctx,
                                 const struct v4l2_ctrl_vp9_frame *dec_params,
                                 struct hantro_decoded_buffer *dst,
                                 struct hantro_decoded_buffer *mv_ref)
{
        static const struct hantro_vp9_ref_reg ref_regs[] = {
                {
                        /* Last */
                        .width = vp9_lref_width,
                        .height = vp9_lref_height,
                        .hor_scale = vp9_lref_hor_scale,
                        .ver_scale = vp9_lref_ver_scale,
                        .y_base = G2_REF_LUMA_ADDR(0),
                        .c_base = G2_REF_CHROMA_ADDR(0),
                }, {
                        /* Golden */
                        .width = vp9_gref_width,
                        .height = vp9_gref_height,
                        .hor_scale = vp9_gref_hor_scale,
                        .ver_scale = vp9_gref_ver_scale,
                        .y_base = G2_REF_LUMA_ADDR(4),
                        .c_base = G2_REF_CHROMA_ADDR(4),
                }, {
                        /* Altref */
                        .width = vp9_aref_width,
                        .height = vp9_aref_height,
                        .hor_scale = vp9_aref_hor_scale,
                        .ver_scale = vp9_aref_ver_scale,
                        .y_base = G2_REF_LUMA_ADDR(5),
                        .c_base = G2_REF_CHROMA_ADDR(5),
                },
        };
        dma_addr_t mv_addr;

        config_ref(ctx, dst, &ref_regs[0], dec_params, dec_params->last_frame_ts);
        config_ref(ctx, dst, &ref_regs[1], dec_params, dec_params->golden_frame_ts);
        config_ref(ctx, dst, &ref_regs[2], dec_params, dec_params->alt_frame_ts);

        mv_addr = hantro_get_dec_buf_addr(ctx, &mv_ref->base.vb.vb2_buf) +
                  mv_ref->vp9.mv_offset;
        hantro_write_addr(ctx->dev, G2_REF_MV_ADDR(0), mv_addr);

        hantro_reg_write(ctx->dev, &vp9_last_sign_bias,
                         dec_params->ref_frame_sign_bias & V4L2_VP9_SIGN_BIAS_LAST ? 1 : 0);

        hantro_reg_write(ctx->dev, &vp9_gref_sign_bias,
                         dec_params->ref_frame_sign_bias & V4L2_VP9_SIGN_BIAS_GOLDEN ? 1 : 0);

        hantro_reg_write(ctx->dev, &vp9_aref_sign_bias,
                         dec_params->ref_frame_sign_bias & V4L2_VP9_SIGN_BIAS_ALT ? 1 : 0);
}

static void recompute_tile_info(unsigned short *tile_info, unsigned int tiles, unsigned int sbs)
{
        int i;
        unsigned int accumulated = 0;
        unsigned int next_accumulated;

        for (i = 1; i <= tiles; ++i) {
                next_accumulated = i * sbs / tiles;
                *tile_info++ = next_accumulated - accumulated;
                accumulated = next_accumulated;
        }
}

static void
recompute_tile_rc_info(struct hantro_ctx *ctx,
                       unsigned int tile_r, unsigned int tile_c,
                       unsigned int sbs_r, unsigned int sbs_c)
{
        struct hantro_vp9_dec_hw_ctx *vp9_ctx = &ctx->vp9_dec;

        recompute_tile_info(vp9_ctx->tile_r_info, tile_r, sbs_r);
        recompute_tile_info(vp9_ctx->tile_c_info, tile_c, sbs_c);

        vp9_ctx->last_tile_r = tile_r;
        vp9_ctx->last_tile_c = tile_c;
        vp9_ctx->last_sbs_r = sbs_r;
        vp9_ctx->last_sbs_c = sbs_c;
}

static inline unsigned int first_tile_row(unsigned int tile_r, unsigned int sbs_r)
{
        if (tile_r == sbs_r + 1)
                return 1;

        if (tile_r == sbs_r + 2)
                return 2;

        return 0;
}

static void
fill_tile_info(struct hantro_ctx *ctx,
               unsigned int tile_r, unsigned int tile_c,
               unsigned int sbs_r, unsigned int sbs_c,
               unsigned short *tile_mem)
{
        struct hantro_vp9_dec_hw_ctx *vp9_ctx = &ctx->vp9_dec;
        unsigned int i, j;
        bool first = true;

        for (i = first_tile_row(tile_r, sbs_r); i < tile_r; ++i) {
                unsigned short r_info = vp9_ctx->tile_r_info[i];

                if (first) {
                        if (i > 0)
                                r_info += vp9_ctx->tile_r_info[0];
                        if (i == 2)
                                r_info += vp9_ctx->tile_r_info[1];
                        first = false;
                }
                for (j = 0; j < tile_c; ++j) {
                        *tile_mem++ = vp9_ctx->tile_c_info[j];
                        *tile_mem++ = r_info;
                }
        }
}

static void
config_tiles(struct hantro_ctx *ctx,
             const struct v4l2_ctrl_vp9_frame *dec_params,
             struct hantro_decoded_buffer *dst)
{
        struct hantro_vp9_dec_hw_ctx *vp9_ctx = &ctx->vp9_dec;
        struct hantro_aux_buf *misc = &vp9_ctx->misc;
        struct hantro_aux_buf *tile_edge = &vp9_ctx->tile_edge;
        dma_addr_t addr;
        unsigned short *tile_mem;
        unsigned int rows, cols;

        addr = misc->dma + vp9_ctx->tile_info_offset;
        hantro_write_addr(ctx->dev, G2_TILE_SIZES_ADDR, addr);

        tile_mem = misc->cpu + vp9_ctx->tile_info_offset;
        if (dec_params->tile_cols_log2 || dec_params->tile_rows_log2) {
                unsigned int tile_r = (1 << dec_params->tile_rows_log2);
                unsigned int tile_c = (1 << dec_params->tile_cols_log2);
                unsigned int sbs_r = hantro_vp9_num_sbs(dst->vp9.height);
                unsigned int sbs_c = hantro_vp9_num_sbs(dst->vp9.width);

                if (tile_r != vp9_ctx->last_tile_r || tile_c != vp9_ctx->last_tile_c ||
                    sbs_r != vp9_ctx->last_sbs_r || sbs_c != vp9_ctx->last_sbs_c)
                        recompute_tile_rc_info(ctx, tile_r, tile_c, sbs_r, sbs_c);

                fill_tile_info(ctx, tile_r, tile_c, sbs_r, sbs_c, tile_mem);

                cols = tile_c;
                rows = tile_r;
                hantro_reg_write(ctx->dev, &g2_tile_e, 1);
        } else {
                tile_mem[0] = hantro_vp9_num_sbs(dst->vp9.width);
                tile_mem[1] = hantro_vp9_num_sbs(dst->vp9.height);

                cols = 1;
                rows = 1;
                hantro_reg_write(ctx->dev, &g2_tile_e, 0);
        }

        if (ctx->dev->variant->legacy_regs) {
                hantro_reg_write(ctx->dev, &g2_num_tile_cols_old, cols);
                hantro_reg_write(ctx->dev, &g2_num_tile_rows_old, rows);
        } else {
                hantro_reg_write(ctx->dev, &g2_num_tile_cols, cols);
                hantro_reg_write(ctx->dev, &g2_num_tile_rows, rows);
        }

        /* provide aux buffers even if no tiles are used */
        addr = tile_edge->dma;
        hantro_write_addr(ctx->dev, G2_TILE_FILTER_ADDR, addr);

        addr = tile_edge->dma + vp9_ctx->bsd_ctrl_offset;
        hantro_write_addr(ctx->dev, G2_TILE_BSD_ADDR, addr);
}

static void
update_feat_and_flag(struct hantro_vp9_dec_hw_ctx *vp9_ctx,
                     const struct v4l2_vp9_segmentation *seg,
                     unsigned int feature,
                     unsigned int segid)
{
        u8 mask = V4L2_VP9_SEGMENT_FEATURE_ENABLED(feature);

        vp9_ctx->feature_data[segid][feature] = seg->feature_data[segid][feature];
        vp9_ctx->feature_enabled[segid] &= ~mask;
        vp9_ctx->feature_enabled[segid] |= (seg->feature_enabled[segid] & mask);
}

static inline s16 clip3(s16 x, s16 y, s16 z)
{
        return (z < x) ? x : (z > y) ? y : z;
}

static s16 feat_val_clip3(s16 feat_val, s16 feature_data, bool absolute, u8 clip)
{
        if (absolute)
                return feature_data;

        return clip3(0, 255, feat_val + feature_data);
}

static void config_segment(struct hantro_ctx *ctx, const struct v4l2_ctrl_vp9_frame *dec_params)
{
        struct hantro_vp9_dec_hw_ctx *vp9_ctx = &ctx->vp9_dec;
        const struct v4l2_vp9_segmentation *seg;
        s16 feat_val;
        unsigned char feat_id;
        unsigned int segid;
        bool segment_enabled, absolute, update_data;

        static const struct hantro_reg seg_regs[8][V4L2_VP9_SEG_LVL_MAX] = {
                { vp9_quant_seg0, vp9_filt_level_seg0, vp9_refpic_seg0, vp9_skip_seg0 },
                { vp9_quant_seg1, vp9_filt_level_seg1, vp9_refpic_seg1, vp9_skip_seg1 },
                { vp9_quant_seg2, vp9_filt_level_seg2, vp9_refpic_seg2, vp9_skip_seg2 },
                { vp9_quant_seg3, vp9_filt_level_seg3, vp9_refpic_seg3, vp9_skip_seg3 },
                { vp9_quant_seg4, vp9_filt_level_seg4, vp9_refpic_seg4, vp9_skip_seg4 },
                { vp9_quant_seg5, vp9_filt_level_seg5, vp9_refpic_seg5, vp9_skip_seg5 },
                { vp9_quant_seg6, vp9_filt_level_seg6, vp9_refpic_seg6, vp9_skip_seg6 },
                { vp9_quant_seg7, vp9_filt_level_seg7, vp9_refpic_seg7, vp9_skip_seg7 },
        };

        segment_enabled = !!(dec_params->seg.flags & V4L2_VP9_SEGMENTATION_FLAG_ENABLED);
        hantro_reg_write(ctx->dev, &vp9_segment_e, segment_enabled);
        hantro_reg_write(ctx->dev, &vp9_segment_upd_e,
                         !!(dec_params->seg.flags & V4L2_VP9_SEGMENTATION_FLAG_UPDATE_MAP));
        hantro_reg_write(ctx->dev, &vp9_segment_temp_upd_e,
                         !!(dec_params->seg.flags & V4L2_VP9_SEGMENTATION_FLAG_TEMPORAL_UPDATE));

        seg = &dec_params->seg;
        absolute = !!(seg->flags & V4L2_VP9_SEGMENTATION_FLAG_ABS_OR_DELTA_UPDATE);
        update_data = !!(seg->flags & V4L2_VP9_SEGMENTATION_FLAG_UPDATE_DATA);

        for (segid = 0; segid < 8; ++segid) {
                /* Quantizer segment feature */
                feat_id = V4L2_VP9_SEG_LVL_ALT_Q;
                feat_val = dec_params->quant.base_q_idx;
                if (segment_enabled) {
                        if (update_data)
                                update_feat_and_flag(vp9_ctx, seg, feat_id, segid);
                        if (v4l2_vp9_seg_feat_enabled(vp9_ctx->feature_enabled, feat_id, segid))
                                feat_val = feat_val_clip3(feat_val,
                                                          vp9_ctx->feature_data[segid][feat_id],
                                                          absolute, 255);
                }
                hantro_reg_write(ctx->dev, &seg_regs[segid][feat_id], feat_val);

                /* Loop filter segment feature */
                feat_id = V4L2_VP9_SEG_LVL_ALT_L;
                feat_val = dec_params->lf.level;
                if (segment_enabled) {
                        if (update_data)
                                update_feat_and_flag(vp9_ctx, seg, feat_id, segid);
                        if (v4l2_vp9_seg_feat_enabled(vp9_ctx->feature_enabled, feat_id, segid))
                                feat_val = feat_val_clip3(feat_val,
                                                          vp9_ctx->feature_data[segid][feat_id],
                                                          absolute, 63);
                }
                hantro_reg_write(ctx->dev, &seg_regs[segid][feat_id], feat_val);

                /* Reference frame segment feature */
                feat_id = V4L2_VP9_SEG_LVL_REF_FRAME;
                feat_val = 0;
                if (segment_enabled) {
                        if (update_data)
                                update_feat_and_flag(vp9_ctx, seg, feat_id, segid);
                        if (!(dec_params->flags & V4L2_VP9_FRAME_FLAG_KEY_FRAME) &&
                            v4l2_vp9_seg_feat_enabled(vp9_ctx->feature_enabled, feat_id, segid))
                                feat_val = vp9_ctx->feature_data[segid][feat_id] + 1;
                }
                hantro_reg_write(ctx->dev, &seg_regs[segid][feat_id], feat_val);

                /* Skip segment feature */
                feat_id = V4L2_VP9_SEG_LVL_SKIP;
                feat_val = 0;
                if (segment_enabled) {
                        if (update_data)
                                update_feat_and_flag(vp9_ctx, seg, feat_id, segid);
                        feat_val = v4l2_vp9_seg_feat_enabled(vp9_ctx->feature_enabled,
                                                             feat_id, segid) ? 1 : 0;
                }
                hantro_reg_write(ctx->dev, &seg_regs[segid][feat_id], feat_val);
        }
}

static void config_loop_filter(struct hantro_ctx *ctx, const struct v4l2_ctrl_vp9_frame *dec_params)
{
        bool d = dec_params->lf.flags & V4L2_VP9_LOOP_FILTER_FLAG_DELTA_ENABLED;

        hantro_reg_write(ctx->dev, &vp9_filt_level, dec_params->lf.level);
        hantro_reg_write(ctx->dev, &g2_out_filtering_dis, dec_params->lf.level == 0);
        hantro_reg_write(ctx->dev, &vp9_filt_sharpness, dec_params->lf.sharpness);

        hantro_reg_write(ctx->dev, &vp9_filt_ref_adj_0, d ? dec_params->lf.ref_deltas[0] : 0);
        hantro_reg_write(ctx->dev, &vp9_filt_ref_adj_1, d ? dec_params->lf.ref_deltas[1] : 0);
        hantro_reg_write(ctx->dev, &vp9_filt_ref_adj_2, d ? dec_params->lf.ref_deltas[2] : 0);
        hantro_reg_write(ctx->dev, &vp9_filt_ref_adj_3, d ? dec_params->lf.ref_deltas[3] : 0);
        hantro_reg_write(ctx->dev, &vp9_filt_mb_adj_0, d ? dec_params->lf.mode_deltas[0] : 0);
        hantro_reg_write(ctx->dev, &vp9_filt_mb_adj_1, d ? dec_params->lf.mode_deltas[1] : 0);
}

static void config_picture_dimensions(struct hantro_ctx *ctx, struct hantro_decoded_buffer *dst)
{
        u32 pic_w_4x4, pic_h_4x4;

        hantro_reg_write(ctx->dev, &g2_pic_width_in_cbs, (dst->vp9.width + 7) / 8);
        hantro_reg_write(ctx->dev, &g2_pic_height_in_cbs, (dst->vp9.height + 7) / 8);
        pic_w_4x4 = roundup(dst->vp9.width, 8) >> 2;
        pic_h_4x4 = roundup(dst->vp9.height, 8) >> 2;
        hantro_reg_write(ctx->dev, &g2_pic_width_4x4, pic_w_4x4);
        hantro_reg_write(ctx->dev, &g2_pic_height_4x4, pic_h_4x4);
}

static void
config_bit_depth(struct hantro_ctx *ctx, const struct v4l2_ctrl_vp9_frame *dec_params)
{
        if (ctx->dev->variant->legacy_regs) {
                hantro_reg_write(ctx->dev, &g2_bit_depth_y, dec_params->bit_depth);
                hantro_reg_write(ctx->dev, &g2_bit_depth_c, dec_params->bit_depth);
                hantro_reg_write(ctx->dev, &g2_pix_shift, 0);
        } else {
                hantro_reg_write(ctx->dev, &g2_bit_depth_y_minus8, dec_params->bit_depth - 8);
                hantro_reg_write(ctx->dev, &g2_bit_depth_c_minus8, dec_params->bit_depth - 8);
        }
}

static inline bool is_lossless(const struct v4l2_vp9_quantization *quant)
{
        return quant->base_q_idx == 0 && quant->delta_q_uv_ac == 0 &&
               quant->delta_q_uv_dc == 0 && quant->delta_q_y_dc == 0;
}

static void
config_quant(struct hantro_ctx *ctx, const struct v4l2_ctrl_vp9_frame *dec_params)
{
        hantro_reg_write(ctx->dev, &vp9_qp_delta_y_dc, dec_params->quant.delta_q_y_dc);
        hantro_reg_write(ctx->dev, &vp9_qp_delta_ch_dc, dec_params->quant.delta_q_uv_dc);
        hantro_reg_write(ctx->dev, &vp9_qp_delta_ch_ac, dec_params->quant.delta_q_uv_ac);
        hantro_reg_write(ctx->dev, &vp9_lossless_e, is_lossless(&dec_params->quant));
}

static u32
hantro_interp_filter_from_v4l2(unsigned int interpolation_filter)
{
        switch (interpolation_filter) {
        case V4L2_VP9_INTERP_FILTER_EIGHTTAP:
                return 0x1;
        case V4L2_VP9_INTERP_FILTER_EIGHTTAP_SMOOTH:
                return 0;
        case V4L2_VP9_INTERP_FILTER_EIGHTTAP_SHARP:
                return 0x2;
        case V4L2_VP9_INTERP_FILTER_BILINEAR:
                return 0x3;
        case V4L2_VP9_INTERP_FILTER_SWITCHABLE:
                return 0x4;
        }

        return 0;
}

static void
config_others(struct hantro_ctx *ctx, const struct v4l2_ctrl_vp9_frame *dec_params,
              bool intra_only, bool resolution_change)
{
        struct hantro_vp9_dec_hw_ctx *vp9_ctx = &ctx->vp9_dec;

        hantro_reg_write(ctx->dev, &g2_idr_pic_e, intra_only);

        hantro_reg_write(ctx->dev, &vp9_transform_mode, vp9_ctx->cur.tx_mode);

        hantro_reg_write(ctx->dev, &vp9_mcomp_filt_type, intra_only ?
                0 : hantro_interp_filter_from_v4l2(dec_params->interpolation_filter));

        hantro_reg_write(ctx->dev, &vp9_high_prec_mv_e,
                         !!(dec_params->flags & V4L2_VP9_FRAME_FLAG_ALLOW_HIGH_PREC_MV));

        hantro_reg_write(ctx->dev, &vp9_comp_pred_mode, dec_params->reference_mode);

        hantro_reg_write(ctx->dev, &g2_tempor_mvp_e,
                         !(dec_params->flags & V4L2_VP9_FRAME_FLAG_ERROR_RESILIENT) &&
                         !(dec_params->flags & V4L2_VP9_FRAME_FLAG_KEY_FRAME) &&
                         !(vp9_ctx->last.flags & V4L2_VP9_FRAME_FLAG_KEY_FRAME) &&
                         !(dec_params->flags & V4L2_VP9_FRAME_FLAG_INTRA_ONLY) &&
                         !resolution_change &&
                         vp9_ctx->last.flags & V4L2_VP9_FRAME_FLAG_SHOW_FRAME
        );

        hantro_reg_write(ctx->dev, &g2_write_mvs_e,
                         !(dec_params->flags & V4L2_VP9_FRAME_FLAG_KEY_FRAME));
}

static void
config_compound_reference(struct hantro_ctx *ctx,
                          const struct v4l2_ctrl_vp9_frame *dec_params)
{
        u32 comp_fixed_ref, comp_var_ref[2];
        bool last_ref_frame_sign_bias;
        bool golden_ref_frame_sign_bias;
        bool alt_ref_frame_sign_bias;
        bool comp_ref_allowed = 0;

        comp_fixed_ref = 0;
        comp_var_ref[0] = 0;
        comp_var_ref[1] = 0;

        last_ref_frame_sign_bias = dec_params->ref_frame_sign_bias & V4L2_VP9_SIGN_BIAS_LAST;
        golden_ref_frame_sign_bias = dec_params->ref_frame_sign_bias & V4L2_VP9_SIGN_BIAS_GOLDEN;
        alt_ref_frame_sign_bias = dec_params->ref_frame_sign_bias & V4L2_VP9_SIGN_BIAS_ALT;

        /* 6.3.12 Frame reference mode syntax */
        comp_ref_allowed |= golden_ref_frame_sign_bias != last_ref_frame_sign_bias;
        comp_ref_allowed |= alt_ref_frame_sign_bias != last_ref_frame_sign_bias;

        if (comp_ref_allowed) {
                if (last_ref_frame_sign_bias ==
                    golden_ref_frame_sign_bias) {
                        comp_fixed_ref = ALTREF_FRAME;
                        comp_var_ref[0] = LAST_FRAME;
                        comp_var_ref[1] = GOLDEN_FRAME;
                } else if (last_ref_frame_sign_bias ==
                           alt_ref_frame_sign_bias) {
                        comp_fixed_ref = GOLDEN_FRAME;
                        comp_var_ref[0] = LAST_FRAME;
                        comp_var_ref[1] = ALTREF_FRAME;
                } else {
                        comp_fixed_ref = LAST_FRAME;
                        comp_var_ref[0] = GOLDEN_FRAME;
                        comp_var_ref[1] = ALTREF_FRAME;
                }
        }

        hantro_reg_write(ctx->dev, &vp9_comp_pred_fixed_ref, comp_fixed_ref);
        hantro_reg_write(ctx->dev, &vp9_comp_pred_var_ref0, comp_var_ref[0]);
        hantro_reg_write(ctx->dev, &vp9_comp_pred_var_ref1, comp_var_ref[1]);
}

#define INNER_LOOP \
do {                                                                    \
        for (m = 0; m < ARRAY_SIZE(adaptive->coef[0][0][0][0]); ++m) {  \
                memcpy(adaptive->coef[i][j][k][l][m],                   \
                       probs->coef[i][j][k][l][m],                      \
                       sizeof(probs->coef[i][j][k][l][m]));             \
                                                                        \
                adaptive->coef[i][j][k][l][m][3] = 0;                   \
        }                                                               \
} while (0)

static void config_probs(struct hantro_ctx *ctx, const struct v4l2_ctrl_vp9_frame *dec_params)
{
        struct hantro_vp9_dec_hw_ctx *vp9_ctx = &ctx->vp9_dec;
        struct hantro_aux_buf *misc = &vp9_ctx->misc;
        struct hantro_g2_all_probs *all_probs = misc->cpu;
        struct hantro_g2_probs *adaptive;
        struct hantro_g2_mv_probs *mv;
        const struct v4l2_vp9_segmentation *seg = &dec_params->seg;
        const struct v4l2_vp9_frame_context *probs = &vp9_ctx->probability_tables;
        int i, j, k, l, m;

        for (i = 0; i < ARRAY_SIZE(all_probs->kf_y_mode_prob); ++i)
                for (j = 0; j < ARRAY_SIZE(all_probs->kf_y_mode_prob[0]); ++j) {
                        memcpy(all_probs->kf_y_mode_prob[i][j],
                               v4l2_vp9_kf_y_mode_prob[i][j],
                               ARRAY_SIZE(all_probs->kf_y_mode_prob[i][j]));

                        all_probs->kf_y_mode_prob_tail[i][j][0] =
                                v4l2_vp9_kf_y_mode_prob[i][j][8];
                }

        memcpy(all_probs->mb_segment_tree_probs, seg->tree_probs,
               sizeof(all_probs->mb_segment_tree_probs));

        memcpy(all_probs->segment_pred_probs, seg->pred_probs,
               sizeof(all_probs->segment_pred_probs));

        for (i = 0; i < ARRAY_SIZE(all_probs->kf_uv_mode_prob); ++i) {
                memcpy(all_probs->kf_uv_mode_prob[i], v4l2_vp9_kf_uv_mode_prob[i],
                       ARRAY_SIZE(all_probs->kf_uv_mode_prob[i]));

                all_probs->kf_uv_mode_prob_tail[i][0] = v4l2_vp9_kf_uv_mode_prob[i][8];
        }

        adaptive = &all_probs->probs;

        for (i = 0; i < ARRAY_SIZE(adaptive->inter_mode); ++i) {
                memcpy(adaptive->inter_mode[i], probs->inter_mode[i],
                       ARRAY_SIZE(probs->inter_mode[i]));

                adaptive->inter_mode[i][3] = 0;
        }

        memcpy(adaptive->is_inter, probs->is_inter, sizeof(adaptive->is_inter));

        for (i = 0; i < ARRAY_SIZE(adaptive->uv_mode); ++i) {
                memcpy(adaptive->uv_mode[i], probs->uv_mode[i],
                       sizeof(adaptive->uv_mode[i]));
                adaptive->uv_mode_tail[i][0] = probs->uv_mode[i][8];
        }

        memcpy(adaptive->tx8, probs->tx8, sizeof(adaptive->tx8));
        memcpy(adaptive->tx16, probs->tx16, sizeof(adaptive->tx16));
        memcpy(adaptive->tx32, probs->tx32, sizeof(adaptive->tx32));

        for (i = 0; i < ARRAY_SIZE(adaptive->y_mode); ++i) {
                memcpy(adaptive->y_mode[i], probs->y_mode[i],
                       ARRAY_SIZE(adaptive->y_mode[i]));

                adaptive->y_mode_tail[i][0] = probs->y_mode[i][8];
        }

        for (i = 0; i < ARRAY_SIZE(adaptive->partition[0]); ++i) {
                memcpy(adaptive->partition[0][i], v4l2_vp9_kf_partition_probs[i],
                       sizeof(v4l2_vp9_kf_partition_probs[i]));

                adaptive->partition[0][i][3] = 0;
        }

        for (i = 0; i < ARRAY_SIZE(adaptive->partition[1]); ++i) {
                memcpy(adaptive->partition[1][i], probs->partition[i],
                       sizeof(probs->partition[i]));

                adaptive->partition[1][i][3] = 0;
        }

        memcpy(adaptive->interp_filter, probs->interp_filter,
               sizeof(adaptive->interp_filter));

        memcpy(adaptive->comp_mode, probs->comp_mode, sizeof(adaptive->comp_mode));

        memcpy(adaptive->skip, probs->skip, sizeof(adaptive->skip));

        mv = &adaptive->mv;

        memcpy(mv->joint, probs->mv.joint, sizeof(mv->joint));
        memcpy(mv->sign, probs->mv.sign, sizeof(mv->sign));
        memcpy(mv->class0_bit, probs->mv.class0_bit, sizeof(mv->class0_bit));
        memcpy(mv->fr, probs->mv.fr, sizeof(mv->fr));
        memcpy(mv->class0_hp, probs->mv.class0_hp, sizeof(mv->class0_hp));
        memcpy(mv->hp, probs->mv.hp, sizeof(mv->hp));
        memcpy(mv->classes, probs->mv.classes, sizeof(mv->classes));
        memcpy(mv->class0_fr, probs->mv.class0_fr, sizeof(mv->class0_fr));
        memcpy(mv->bits, probs->mv.bits, sizeof(mv->bits));

        memcpy(adaptive->single_ref, probs->single_ref, sizeof(adaptive->single_ref));

        memcpy(adaptive->comp_ref, probs->comp_ref, sizeof(adaptive->comp_ref));

        for (i = 0; i < ARRAY_SIZE(adaptive->coef); ++i)
                for (j = 0; j < ARRAY_SIZE(adaptive->coef[0]); ++j)
                        for (k = 0; k < ARRAY_SIZE(adaptive->coef[0][0]); ++k)
                                for (l = 0; l < ARRAY_SIZE(adaptive->coef[0][0][0]); ++l)
                                        INNER_LOOP;

        hantro_write_addr(ctx->dev, G2_VP9_PROBS_ADDR, misc->dma);
}

static void config_counts(struct hantro_ctx *ctx)
{
        struct hantro_vp9_dec_hw_ctx *vp9_dec = &ctx->vp9_dec;
        struct hantro_aux_buf *misc = &vp9_dec->misc;
        dma_addr_t addr = misc->dma + vp9_dec->ctx_counters_offset;

        hantro_write_addr(ctx->dev, G2_VP9_CTX_COUNT_ADDR, addr);
}

static void config_seg_map(struct hantro_ctx *ctx,
                           const struct v4l2_ctrl_vp9_frame *dec_params,
                           bool intra_only, bool update_map)
{
        struct hantro_vp9_dec_hw_ctx *vp9_ctx = &ctx->vp9_dec;
        struct hantro_aux_buf *segment_map = &vp9_ctx->segment_map;
        dma_addr_t addr;

        if (intra_only ||
            (dec_params->flags & V4L2_VP9_FRAME_FLAG_ERROR_RESILIENT)) {
                memset(segment_map->cpu, 0, segment_map->size);
                memset(vp9_ctx->feature_data, 0, sizeof(vp9_ctx->feature_data));
                memset(vp9_ctx->feature_enabled, 0, sizeof(vp9_ctx->feature_enabled));
        }

        addr = segment_map->dma + vp9_ctx->active_segment * vp9_ctx->segment_map_size;
        hantro_write_addr(ctx->dev, G2_VP9_SEGMENT_READ_ADDR, addr);

        addr = segment_map->dma + (1 - vp9_ctx->active_segment) * vp9_ctx->segment_map_size;
        hantro_write_addr(ctx->dev, G2_VP9_SEGMENT_WRITE_ADDR, addr);

        if (update_map)
                vp9_ctx->active_segment = 1 - vp9_ctx->active_segment;
}

static void
config_source(struct hantro_ctx *ctx, const struct v4l2_ctrl_vp9_frame *dec_params,
              struct vb2_v4l2_buffer *vb2_src)
{
        dma_addr_t stream_base, tmp_addr;
        unsigned int headers_size;
        u32 src_len, start_bit, src_buf_len;

        headers_size = dec_params->uncompressed_header_size
                     + dec_params->compressed_header_size;

        stream_base = vb2_dma_contig_plane_dma_addr(&vb2_src->vb2_buf, 0);

        tmp_addr = stream_base + headers_size;
        if (ctx->dev->variant->legacy_regs)
                hantro_write_addr(ctx->dev, G2_STREAM_ADDR, (tmp_addr & ~0xf));
        else
                hantro_write_addr(ctx->dev, G2_STREAM_ADDR, stream_base);

        start_bit = (tmp_addr & 0xf) * 8;
        hantro_reg_write(ctx->dev, &g2_start_bit, start_bit);

        src_len = vb2_get_plane_payload(&vb2_src->vb2_buf, 0);
        src_len += start_bit / 8 - headers_size;
        hantro_reg_write(ctx->dev, &g2_stream_len, src_len);

        if (!ctx->dev->variant->legacy_regs) {
                tmp_addr &= ~0xf;
                hantro_reg_write(ctx->dev, &g2_strm_start_offset, tmp_addr - stream_base);
                src_buf_len = vb2_plane_size(&vb2_src->vb2_buf, 0);
                hantro_reg_write(ctx->dev, &g2_strm_buffer_len, src_buf_len);
        }
}

static void
config_registers(struct hantro_ctx *ctx, const struct v4l2_ctrl_vp9_frame *dec_params,
                 struct vb2_v4l2_buffer *vb2_src, struct vb2_v4l2_buffer *vb2_dst)
{
        struct hantro_decoded_buffer *dst, *last, *mv_ref;
        struct hantro_vp9_dec_hw_ctx *vp9_ctx = &ctx->vp9_dec;
        const struct v4l2_vp9_segmentation *seg;
        bool intra_only, resolution_change;

        /* vp9 stuff */
        dst = vb2_to_hantro_decoded_buf(&vb2_dst->vb2_buf);

        if (vp9_ctx->last.valid)
                last = get_ref_buf(ctx, &dst->base.vb, vp9_ctx->last.timestamp);
        else
                last = dst;

        update_dec_buf_info(dst, dec_params);
        update_ctx_cur_info(vp9_ctx, dst, dec_params);
        seg = &dec_params->seg;

        intra_only = !!(dec_params->flags &
                        (V4L2_VP9_FRAME_FLAG_KEY_FRAME |
                        V4L2_VP9_FRAME_FLAG_INTRA_ONLY));

        if (!intra_only &&
            !(dec_params->flags & V4L2_VP9_FRAME_FLAG_ERROR_RESILIENT) &&
            vp9_ctx->last.valid)
                mv_ref = last;
        else
                mv_ref = dst;

        resolution_change = dst->vp9.width != last->vp9.width ||
                            dst->vp9.height != last->vp9.height;

        /* configure basic registers */
        hantro_reg_write(ctx->dev, &g2_mode, VP9_DEC_MODE);
        if (!ctx->dev->variant->legacy_regs) {
                hantro_reg_write(ctx->dev, &g2_strm_swap, 0xf);
                hantro_reg_write(ctx->dev, &g2_dirmv_swap, 0xf);
                hantro_reg_write(ctx->dev, &g2_compress_swap, 0xf);
                hantro_reg_write(ctx->dev, &g2_ref_compress_bypass, 1);
        } else {
                hantro_reg_write(ctx->dev, &g2_strm_swap_old, 0x1f);
                hantro_reg_write(ctx->dev, &g2_pic_swap, 0x10);
                hantro_reg_write(ctx->dev, &g2_dirmv_swap_old, 0x10);
                hantro_reg_write(ctx->dev, &g2_tab0_swap_old, 0x10);
                hantro_reg_write(ctx->dev, &g2_tab1_swap_old, 0x10);
                hantro_reg_write(ctx->dev, &g2_tab2_swap_old, 0x10);
                hantro_reg_write(ctx->dev, &g2_tab3_swap_old, 0x10);
                hantro_reg_write(ctx->dev, &g2_rscan_swap, 0x10);
        }
        hantro_reg_write(ctx->dev, &g2_buswidth, BUS_WIDTH_128);
        hantro_reg_write(ctx->dev, &g2_max_burst, 16);
        hantro_reg_write(ctx->dev, &g2_apf_threshold, 8);
        hantro_reg_write(ctx->dev, &g2_clk_gate_e, 1);
        hantro_reg_write(ctx->dev, &g2_max_cb_size, 6);
        hantro_reg_write(ctx->dev, &g2_min_cb_size, 3);
        if (ctx->dev->variant->double_buffer)
                hantro_reg_write(ctx->dev, &g2_double_buffer_e, 1);

        config_output(ctx, dst, dec_params);

        if (!intra_only)
                config_ref_registers(ctx, dec_params, dst, mv_ref);

        config_tiles(ctx, dec_params, dst);
        config_segment(ctx, dec_params);
        config_loop_filter(ctx, dec_params);
        config_picture_dimensions(ctx, dst);
        config_bit_depth(ctx, dec_params);
        config_quant(ctx, dec_params);
        config_others(ctx, dec_params, intra_only, resolution_change);
        config_compound_reference(ctx, dec_params);
        config_probs(ctx, dec_params);
        config_counts(ctx);
        config_seg_map(ctx, dec_params, intra_only,
                       seg->flags & V4L2_VP9_SEGMENTATION_FLAG_UPDATE_MAP);
        config_source(ctx, dec_params, vb2_src);
}

int hantro_g2_vp9_dec_run(struct hantro_ctx *ctx)
{
        const struct v4l2_ctrl_vp9_frame *decode_params;
        struct vb2_v4l2_buffer *src;
        struct vb2_v4l2_buffer *dst;
        int ret;

        ret = start_prepare_run(ctx, &decode_params);
        if (ret) {
                hantro_end_prepare_run(ctx);
                return ret;
        }

        src = hantro_get_src_buf(ctx);
        dst = hantro_get_dst_buf(ctx);

        config_registers(ctx, decode_params, src, dst);

        hantro_end_prepare_run(ctx);

        vdpu_write(ctx->dev, G2_REG_INTERRUPT_DEC_E, G2_REG_INTERRUPT);

        return 0;
}

#define copy_tx_and_skip(p1, p2)                                \
do {                                                            \
        memcpy((p1)->tx8, (p2)->tx8, sizeof((p1)->tx8));        \
        memcpy((p1)->tx16, (p2)->tx16, sizeof((p1)->tx16));     \
        memcpy((p1)->tx32, (p2)->tx32, sizeof((p1)->tx32));     \
        memcpy((p1)->skip, (p2)->skip, sizeof((p1)->skip));     \
} while (0)

void hantro_g2_vp9_dec_done(struct hantro_ctx *ctx)
{
        struct hantro_vp9_dec_hw_ctx *vp9_ctx = &ctx->vp9_dec;
        unsigned int fctx_idx;

        if (!(vp9_ctx->cur.flags & V4L2_VP9_FRAME_FLAG_REFRESH_FRAME_CTX))
                goto out_update_last;

        fctx_idx = vp9_ctx->cur.frame_context_idx;

        if (!(vp9_ctx->cur.flags & V4L2_VP9_FRAME_FLAG_PARALLEL_DEC_MODE)) {
                /* error_resilient_mode == 0 && frame_parallel_decoding_mode == 0 */
                struct v4l2_vp9_frame_context *probs = &vp9_ctx->probability_tables;
                bool frame_is_intra = vp9_ctx->cur.flags &
                    (V4L2_VP9_FRAME_FLAG_KEY_FRAME | V4L2_VP9_FRAME_FLAG_INTRA_ONLY);
                struct tx_and_skip {
                        u8 tx8[2][1];
                        u8 tx16[2][2];
                        u8 tx32[2][3];
                        u8 skip[3];
                } _tx_skip, *tx_skip = &_tx_skip;
                struct v4l2_vp9_frame_symbol_counts *counts;
                struct symbol_counts *hantro_cnts;
                u32 tx16p[2][4];
                int i;

                /* buffer the forward-updated TX and skip probs */
                if (frame_is_intra)
                        copy_tx_and_skip(tx_skip, probs);

                /* 6.1.2 refresh_probs(): load_probs() and load_probs2() */
                *probs = vp9_ctx->frame_context[fctx_idx];

                /* if FrameIsIntra then undo the effect of load_probs2() */
                if (frame_is_intra)
                        copy_tx_and_skip(probs, tx_skip);

                counts = &vp9_ctx->cnts;
                hantro_cnts = vp9_ctx->misc.cpu + vp9_ctx->ctx_counters_offset;
                for (i = 0; i < ARRAY_SIZE(tx16p); ++i) {
                        memcpy(tx16p[i],
                               hantro_cnts->tx16x16_count[i],
                               sizeof(hantro_cnts->tx16x16_count[0]));
                        tx16p[i][3] = 0;
                }
                counts->tx16p = &tx16p;

                v4l2_vp9_adapt_coef_probs(probs, counts,
                                          !vp9_ctx->last.valid ||
                                          vp9_ctx->last.flags & V4L2_VP9_FRAME_FLAG_KEY_FRAME,
                                          frame_is_intra);

                if (!frame_is_intra) {
                        /* load_probs2() already done */
                        u32 mv_mode[7][4];

                        for (i = 0; i < ARRAY_SIZE(mv_mode); ++i) {
                                mv_mode[i][0] = hantro_cnts->inter_mode_counts[i][1][0];
                                mv_mode[i][1] = hantro_cnts->inter_mode_counts[i][2][0];
                                mv_mode[i][2] = hantro_cnts->inter_mode_counts[i][0][0];
                                mv_mode[i][3] = hantro_cnts->inter_mode_counts[i][2][1];
                        }
                        counts->mv_mode = &mv_mode;
                        v4l2_vp9_adapt_noncoef_probs(&vp9_ctx->probability_tables, counts,
                                                     vp9_ctx->cur.reference_mode,
                                                     vp9_ctx->cur.interpolation_filter,
                                                     vp9_ctx->cur.tx_mode, vp9_ctx->cur.flags);
                }
        }

        vp9_ctx->frame_context[fctx_idx] = vp9_ctx->probability_tables;

out_update_last:
        vp9_ctx->last = vp9_ctx->cur;
}