media: hantro: merge Rockchip platform drivers
authorAlex Bee <knaerzche@gmail.com>
Mon, 14 Jun 2021 21:32:13 +0000 (23:32 +0200)
committerMauro Carvalho Chehab <mchehab+huawei@kernel.org>
Thu, 17 Jun 2021 08:58:57 +0000 (10:58 +0200)
Merge the two Rockchip platform drivers into one as it was suggested at
[1] and [2].
This will hopefully make it easier to add new variants (which are surely
to come for Rockchip).
Also rename from "rk3288" to "v(d/e)pu1" and "rk3399" to "v(d/e)pu2"
where applicable, as this is the dicition the vendor uses and will
also refelect the variants that get added later in this series. Rename
from "rk3288" to "rockchip" if applicable to both hardware versions.

[1] https://patchwork.kernel.org/project/linux-rockchip/patch/20210107134101.195426-6-paul.kocialkowski@bootlin.com/
[2] https://patchwork.kernel.org/project/linux-rockchip/patch/20210525152225.154302-5-knaerzche@gmail.com/

Signed-off-by: Alex Bee <knaerzche@gmail.com>
Reviewed-by: Ezequiel Garcia <ezequiel@collabora.com>
Signed-off-by: Hans Verkuil <hverkuil-cisco@xs4all.nl>
Signed-off-by: Mauro Carvalho Chehab <mchehab+huawei@kernel.org>
13 files changed:
drivers/staging/media/hantro/Makefile
drivers/staging/media/hantro/hantro_hw.h
drivers/staging/media/hantro/rk3288_vpu_hw.c [deleted file]
drivers/staging/media/hantro/rk3399_vpu_hw.c [deleted file]
drivers/staging/media/hantro/rk3399_vpu_hw_jpeg_enc.c [deleted file]
drivers/staging/media/hantro/rk3399_vpu_hw_mpeg2_dec.c [deleted file]
drivers/staging/media/hantro/rk3399_vpu_hw_vp8_dec.c [deleted file]
drivers/staging/media/hantro/rk3399_vpu_regs.h [deleted file]
drivers/staging/media/hantro/rockchip_vpu2_hw_jpeg_enc.c [new file with mode: 0644]
drivers/staging/media/hantro/rockchip_vpu2_hw_mpeg2_dec.c [new file with mode: 0644]
drivers/staging/media/hantro/rockchip_vpu2_hw_vp8_dec.c [new file with mode: 0644]
drivers/staging/media/hantro/rockchip_vpu2_regs.h [new file with mode: 0644]
drivers/staging/media/hantro/rockchip_vpu_hw.c [new file with mode: 0644]

index 23bfc423b23c33af320eb7691f2235030f305778..287370188d2aed0541cde1b300666388b115a9aa 100644 (file)
@@ -12,9 +12,9 @@ hantro-vpu-y += \
                hantro_g1_mpeg2_dec.o \
                hantro_g2_hevc_dec.o \
                hantro_g1_vp8_dec.o \
-               rk3399_vpu_hw_jpeg_enc.o \
-               rk3399_vpu_hw_mpeg2_dec.o \
-               rk3399_vpu_hw_vp8_dec.o \
+               rockchip_vpu2_hw_jpeg_enc.o \
+               rockchip_vpu2_hw_mpeg2_dec.o \
+               rockchip_vpu2_hw_vp8_dec.o \
                hantro_jpeg.o \
                hantro_h264.o \
                hantro_hevc.o \
@@ -28,5 +28,4 @@ hantro-vpu-$(CONFIG_VIDEO_HANTRO_SAMA5D4) += \
                sama5d4_vdec_hw.o
 
 hantro-vpu-$(CONFIG_VIDEO_HANTRO_ROCKCHIP) += \
-               rk3288_vpu_hw.o \
-               rk3399_vpu_hw.o
+               rockchip_vpu_hw.o
index 7fa67d0c7e0fbf1f49aaf57b38096202285e6df2..a7b75b05e849343172e8bb0861887df3e05b326f 100644 (file)
@@ -191,16 +191,16 @@ struct hantro_codec_ops {
 /**
  * enum hantro_enc_fmt - source format ID for hardware registers.
  *
- * @RK3288_VPU_ENC_FMT_YUV420P: Y/CbCr 4:2:0 planar format
- * @RK3288_VPU_ENC_FMT_YUV420SP: Y/CbCr 4:2:0 semi-planar format
- * @RK3288_VPU_ENC_FMT_YUYV422: YUV 4:2:2 packed format (YUYV)
- * @RK3288_VPU_ENC_FMT_UYVY422: YUV 4:2:2 packed format (UYVY)
+ * @ROCKCHIP_VPU_ENC_FMT_YUV420P: Y/CbCr 4:2:0 planar format
+ * @ROCKCHIP_VPU_ENC_FMT_YUV420SP: Y/CbCr 4:2:0 semi-planar format
+ * @ROCKCHIP_VPU_ENC_FMT_YUYV422: YUV 4:2:2 packed format (YUYV)
+ * @ROCKCHIP_VPU_ENC_FMT_UYVY422: YUV 4:2:2 packed format (UYVY)
  */
 enum hantro_enc_fmt {
-       RK3288_VPU_ENC_FMT_YUV420P = 0,
-       RK3288_VPU_ENC_FMT_YUV420SP = 1,
-       RK3288_VPU_ENC_FMT_YUYV422 = 2,
-       RK3288_VPU_ENC_FMT_UYVY422 = 3,
+       ROCKCHIP_VPU_ENC_FMT_YUV420P = 0,
+       ROCKCHIP_VPU_ENC_FMT_YUV420SP = 1,
+       ROCKCHIP_VPU_ENC_FMT_YUYV422 = 2,
+       ROCKCHIP_VPU_ENC_FMT_UYVY422 = 3,
 };
 
 extern const struct hantro_variant imx8mq_vpu_g2_variant;
@@ -225,7 +225,7 @@ irqreturn_t hantro_g1_irq(int irq, void *dev_id);
 void hantro_g1_reset(struct hantro_ctx *ctx);
 
 int hantro_h1_jpeg_enc_run(struct hantro_ctx *ctx);
-int rk3399_vpu_jpeg_enc_run(struct hantro_ctx *ctx);
+int rockchip_vpu2_jpeg_enc_run(struct hantro_ctx *ctx);
 int hantro_jpeg_enc_init(struct hantro_ctx *ctx);
 void hantro_jpeg_enc_exit(struct hantro_ctx *ctx);
 void hantro_jpeg_enc_done(struct hantro_ctx *ctx);
@@ -274,14 +274,14 @@ hantro_h264_mv_size(unsigned int width, unsigned int height)
 }
 
 int hantro_g1_mpeg2_dec_run(struct hantro_ctx *ctx);
-int rk3399_vpu_mpeg2_dec_run(struct hantro_ctx *ctx);
+int rockchip_vpu2_mpeg2_dec_run(struct hantro_ctx *ctx);
 void hantro_mpeg2_dec_copy_qtable(u8 *qtable,
                                  const struct v4l2_ctrl_mpeg2_quantisation *ctrl);
 int hantro_mpeg2_dec_init(struct hantro_ctx *ctx);
 void hantro_mpeg2_dec_exit(struct hantro_ctx *ctx);
 
 int hantro_g1_vp8_dec_run(struct hantro_ctx *ctx);
-int rk3399_vpu_vp8_dec_run(struct hantro_ctx *ctx);
+int rockchip_vpu2_vp8_dec_run(struct hantro_ctx *ctx);
 int hantro_vp8_dec_init(struct hantro_ctx *ctx);
 void hantro_vp8_dec_exit(struct hantro_ctx *ctx);
 void hantro_vp8_prob_update(struct hantro_ctx *ctx,
diff --git a/drivers/staging/media/hantro/rk3288_vpu_hw.c b/drivers/staging/media/hantro/rk3288_vpu_hw.c
deleted file mode 100644 (file)
index fefd452..0000000
+++ /dev/null
@@ -1,208 +0,0 @@
-// SPDX-License-Identifier: GPL-2.0
-/*
- * Hantro VPU codec driver
- *
- * Copyright (C) 2018 Rockchip Electronics Co., Ltd.
- *     Jeffy Chen <jeffy.chen@rock-chips.com>
- */
-
-#include <linux/clk.h>
-
-#include "hantro.h"
-#include "hantro_jpeg.h"
-#include "hantro_h1_regs.h"
-
-#define RK3288_ACLK_MAX_FREQ (400 * 1000 * 1000)
-
-/*
- * Supported formats.
- */
-
-static const struct hantro_fmt rk3288_vpu_enc_fmts[] = {
-       {
-               .fourcc = V4L2_PIX_FMT_YUV420M,
-               .codec_mode = HANTRO_MODE_NONE,
-               .enc_fmt = RK3288_VPU_ENC_FMT_YUV420P,
-       },
-       {
-               .fourcc = V4L2_PIX_FMT_NV12M,
-               .codec_mode = HANTRO_MODE_NONE,
-               .enc_fmt = RK3288_VPU_ENC_FMT_YUV420SP,
-       },
-       {
-               .fourcc = V4L2_PIX_FMT_YUYV,
-               .codec_mode = HANTRO_MODE_NONE,
-               .enc_fmt = RK3288_VPU_ENC_FMT_YUYV422,
-       },
-       {
-               .fourcc = V4L2_PIX_FMT_UYVY,
-               .codec_mode = HANTRO_MODE_NONE,
-               .enc_fmt = RK3288_VPU_ENC_FMT_UYVY422,
-       },
-       {
-               .fourcc = V4L2_PIX_FMT_JPEG,
-               .codec_mode = HANTRO_MODE_JPEG_ENC,
-               .max_depth = 2,
-               .header_size = JPEG_HEADER_SIZE,
-               .frmsize = {
-                       .min_width = 96,
-                       .max_width = 8192,
-                       .step_width = MB_DIM,
-                       .min_height = 32,
-                       .max_height = 8192,
-                       .step_height = MB_DIM,
-               },
-       },
-};
-
-static const struct hantro_fmt rk3288_vpu_postproc_fmts[] = {
-       {
-               .fourcc = V4L2_PIX_FMT_YUYV,
-               .codec_mode = HANTRO_MODE_NONE,
-       },
-};
-
-static const struct hantro_fmt rk3288_vpu_dec_fmts[] = {
-       {
-               .fourcc = V4L2_PIX_FMT_NV12,
-               .codec_mode = HANTRO_MODE_NONE,
-       },
-       {
-               .fourcc = V4L2_PIX_FMT_H264_SLICE,
-               .codec_mode = HANTRO_MODE_H264_DEC,
-               .max_depth = 2,
-               .frmsize = {
-                       .min_width = 48,
-                       .max_width = 4096,
-                       .step_width = MB_DIM,
-                       .min_height = 48,
-                       .max_height = 2304,
-                       .step_height = MB_DIM,
-               },
-       },
-       {
-               .fourcc = V4L2_PIX_FMT_MPEG2_SLICE,
-               .codec_mode = HANTRO_MODE_MPEG2_DEC,
-               .max_depth = 2,
-               .frmsize = {
-                       .min_width = 48,
-                       .max_width = 1920,
-                       .step_width = MB_DIM,
-                       .min_height = 48,
-                       .max_height = 1088,
-                       .step_height = MB_DIM,
-               },
-       },
-       {
-               .fourcc = V4L2_PIX_FMT_VP8_FRAME,
-               .codec_mode = HANTRO_MODE_VP8_DEC,
-               .max_depth = 2,
-               .frmsize = {
-                       .min_width = 48,
-                       .max_width = 3840,
-                       .step_width = MB_DIM,
-                       .min_height = 48,
-                       .max_height = 2160,
-                       .step_height = MB_DIM,
-               },
-       },
-};
-
-static irqreturn_t rk3288_vepu_irq(int irq, void *dev_id)
-{
-       struct hantro_dev *vpu = dev_id;
-       enum vb2_buffer_state state;
-       u32 status;
-
-       status = vepu_read(vpu, H1_REG_INTERRUPT);
-       state = (status & H1_REG_INTERRUPT_FRAME_RDY) ?
-               VB2_BUF_STATE_DONE : VB2_BUF_STATE_ERROR;
-
-       vepu_write(vpu, 0, H1_REG_INTERRUPT);
-       vepu_write(vpu, 0, H1_REG_AXI_CTRL);
-
-       hantro_irq_done(vpu, state);
-
-       return IRQ_HANDLED;
-}
-
-static int rk3288_vpu_hw_init(struct hantro_dev *vpu)
-{
-       /* Bump ACLK to max. possible freq. to improve performance. */
-       clk_set_rate(vpu->clocks[0].clk, RK3288_ACLK_MAX_FREQ);
-       return 0;
-}
-
-static void rk3288_vpu_enc_reset(struct hantro_ctx *ctx)
-{
-       struct hantro_dev *vpu = ctx->dev;
-
-       vepu_write(vpu, H1_REG_INTERRUPT_DIS_BIT, H1_REG_INTERRUPT);
-       vepu_write(vpu, 0, H1_REG_ENC_CTRL);
-       vepu_write(vpu, 0, H1_REG_AXI_CTRL);
-}
-
-/*
- * Supported codec ops.
- */
-
-static const struct hantro_codec_ops rk3288_vpu_codec_ops[] = {
-       [HANTRO_MODE_JPEG_ENC] = {
-               .run = hantro_h1_jpeg_enc_run,
-               .reset = rk3288_vpu_enc_reset,
-               .init = hantro_jpeg_enc_init,
-               .done = hantro_jpeg_enc_done,
-               .exit = hantro_jpeg_enc_exit,
-       },
-       [HANTRO_MODE_H264_DEC] = {
-               .run = hantro_g1_h264_dec_run,
-               .reset = hantro_g1_reset,
-               .init = hantro_h264_dec_init,
-               .exit = hantro_h264_dec_exit,
-       },
-       [HANTRO_MODE_MPEG2_DEC] = {
-               .run = hantro_g1_mpeg2_dec_run,
-               .reset = hantro_g1_reset,
-               .init = hantro_mpeg2_dec_init,
-               .exit = hantro_mpeg2_dec_exit,
-       },
-       [HANTRO_MODE_VP8_DEC] = {
-               .run = hantro_g1_vp8_dec_run,
-               .reset = hantro_g1_reset,
-               .init = hantro_vp8_dec_init,
-               .exit = hantro_vp8_dec_exit,
-       },
-};
-
-/*
- * VPU variant.
- */
-
-static const struct hantro_irq rk3288_irqs[] = {
-       { "vepu", rk3288_vepu_irq },
-       { "vdpu", hantro_g1_irq },
-};
-
-static const char * const rk3288_clk_names[] = {
-       "aclk", "hclk"
-};
-
-const struct hantro_variant rk3288_vpu_variant = {
-       .enc_offset = 0x0,
-       .enc_fmts = rk3288_vpu_enc_fmts,
-       .num_enc_fmts = ARRAY_SIZE(rk3288_vpu_enc_fmts),
-       .dec_offset = 0x400,
-       .dec_fmts = rk3288_vpu_dec_fmts,
-       .num_dec_fmts = ARRAY_SIZE(rk3288_vpu_dec_fmts),
-       .postproc_fmts = rk3288_vpu_postproc_fmts,
-       .num_postproc_fmts = ARRAY_SIZE(rk3288_vpu_postproc_fmts),
-       .postproc_regs = &hantro_g1_postproc_regs,
-       .codec = HANTRO_JPEG_ENCODER | HANTRO_MPEG2_DECODER |
-                HANTRO_VP8_DECODER | HANTRO_H264_DECODER,
-       .codec_ops = rk3288_vpu_codec_ops,
-       .irqs = rk3288_irqs,
-       .num_irqs = ARRAY_SIZE(rk3288_irqs),
-       .init = rk3288_vpu_hw_init,
-       .clk_names = rk3288_clk_names,
-       .num_clocks = ARRAY_SIZE(rk3288_clk_names)
-};
diff --git a/drivers/staging/media/hantro/rk3399_vpu_hw.c b/drivers/staging/media/hantro/rk3399_vpu_hw.c
deleted file mode 100644 (file)
index 7a7962c..0000000
+++ /dev/null
@@ -1,222 +0,0 @@
-// SPDX-License-Identifier: GPL-2.0
-/*
- * Hantro VPU codec driver
- *
- * Copyright (C) 2018 Rockchip Electronics Co., Ltd.
- *     Jeffy Chen <jeffy.chen@rock-chips.com>
- */
-
-#include <linux/clk.h>
-
-#include "hantro.h"
-#include "hantro_jpeg.h"
-#include "rk3399_vpu_regs.h"
-
-#define RK3399_ACLK_MAX_FREQ (400 * 1000 * 1000)
-
-/*
- * Supported formats.
- */
-
-static const struct hantro_fmt rk3399_vpu_enc_fmts[] = {
-       {
-               .fourcc = V4L2_PIX_FMT_YUV420M,
-               .codec_mode = HANTRO_MODE_NONE,
-               .enc_fmt = RK3288_VPU_ENC_FMT_YUV420P,
-       },
-       {
-               .fourcc = V4L2_PIX_FMT_NV12M,
-               .codec_mode = HANTRO_MODE_NONE,
-               .enc_fmt = RK3288_VPU_ENC_FMT_YUV420SP,
-       },
-       {
-               .fourcc = V4L2_PIX_FMT_YUYV,
-               .codec_mode = HANTRO_MODE_NONE,
-               .enc_fmt = RK3288_VPU_ENC_FMT_YUYV422,
-       },
-       {
-               .fourcc = V4L2_PIX_FMT_UYVY,
-               .codec_mode = HANTRO_MODE_NONE,
-               .enc_fmt = RK3288_VPU_ENC_FMT_UYVY422,
-       },
-       {
-               .fourcc = V4L2_PIX_FMT_JPEG,
-               .codec_mode = HANTRO_MODE_JPEG_ENC,
-               .max_depth = 2,
-               .header_size = JPEG_HEADER_SIZE,
-               .frmsize = {
-                       .min_width = 96,
-                       .max_width = 8192,
-                       .step_width = MB_DIM,
-                       .min_height = 32,
-                       .max_height = 8192,
-                       .step_height = MB_DIM,
-               },
-       },
-};
-
-static const struct hantro_fmt rk3399_vpu_dec_fmts[] = {
-       {
-               .fourcc = V4L2_PIX_FMT_NV12,
-               .codec_mode = HANTRO_MODE_NONE,
-       },
-       {
-               .fourcc = V4L2_PIX_FMT_MPEG2_SLICE,
-               .codec_mode = HANTRO_MODE_MPEG2_DEC,
-               .max_depth = 2,
-               .frmsize = {
-                       .min_width = 48,
-                       .max_width = 1920,
-                       .step_width = MB_DIM,
-                       .min_height = 48,
-                       .max_height = 1088,
-                       .step_height = MB_DIM,
-               },
-       },
-       {
-               .fourcc = V4L2_PIX_FMT_VP8_FRAME,
-               .codec_mode = HANTRO_MODE_VP8_DEC,
-               .max_depth = 2,
-               .frmsize = {
-                       .min_width = 48,
-                       .max_width = 3840,
-                       .step_width = MB_DIM,
-                       .min_height = 48,
-                       .max_height = 2160,
-                       .step_height = MB_DIM,
-               },
-       },
-};
-
-static irqreturn_t rk3399_vepu_irq(int irq, void *dev_id)
-{
-       struct hantro_dev *vpu = dev_id;
-       enum vb2_buffer_state state;
-       u32 status;
-
-       status = vepu_read(vpu, VEPU_REG_INTERRUPT);
-       state = (status & VEPU_REG_INTERRUPT_FRAME_READY) ?
-               VB2_BUF_STATE_DONE : VB2_BUF_STATE_ERROR;
-
-       vepu_write(vpu, 0, VEPU_REG_INTERRUPT);
-       vepu_write(vpu, 0, VEPU_REG_AXI_CTRL);
-
-       hantro_irq_done(vpu, state);
-
-       return IRQ_HANDLED;
-}
-
-static irqreturn_t rk3399_vdpu_irq(int irq, void *dev_id)
-{
-       struct hantro_dev *vpu = dev_id;
-       enum vb2_buffer_state state;
-       u32 status;
-
-       status = vdpu_read(vpu, VDPU_REG_INTERRUPT);
-       state = (status & VDPU_REG_INTERRUPT_DEC_IRQ) ?
-               VB2_BUF_STATE_DONE : VB2_BUF_STATE_ERROR;
-
-       vdpu_write(vpu, 0, VDPU_REG_INTERRUPT);
-       vdpu_write(vpu, 0, VDPU_REG_AXI_CTRL);
-
-       hantro_irq_done(vpu, state);
-
-       return IRQ_HANDLED;
-}
-
-static int rk3399_vpu_hw_init(struct hantro_dev *vpu)
-{
-       /* Bump ACLK to max. possible freq. to improve performance. */
-       clk_set_rate(vpu->clocks[0].clk, RK3399_ACLK_MAX_FREQ);
-       return 0;
-}
-
-static void rk3399_vpu_enc_reset(struct hantro_ctx *ctx)
-{
-       struct hantro_dev *vpu = ctx->dev;
-
-       vepu_write(vpu, VEPU_REG_INTERRUPT_DIS_BIT, VEPU_REG_INTERRUPT);
-       vepu_write(vpu, 0, VEPU_REG_ENCODE_START);
-       vepu_write(vpu, 0, VEPU_REG_AXI_CTRL);
-}
-
-static void rk3399_vpu_dec_reset(struct hantro_ctx *ctx)
-{
-       struct hantro_dev *vpu = ctx->dev;
-
-       vdpu_write(vpu, VDPU_REG_INTERRUPT_DEC_IRQ_DIS, VDPU_REG_INTERRUPT);
-       vdpu_write(vpu, 0, VDPU_REG_EN_FLAGS);
-       vdpu_write(vpu, 1, VDPU_REG_SOFT_RESET);
-}
-
-/*
- * Supported codec ops.
- */
-
-static const struct hantro_codec_ops rk3399_vpu_codec_ops[] = {
-       [HANTRO_MODE_JPEG_ENC] = {
-               .run = rk3399_vpu_jpeg_enc_run,
-               .reset = rk3399_vpu_enc_reset,
-               .init = hantro_jpeg_enc_init,
-               .exit = hantro_jpeg_enc_exit,
-       },
-       [HANTRO_MODE_MPEG2_DEC] = {
-               .run = rk3399_vpu_mpeg2_dec_run,
-               .reset = rk3399_vpu_dec_reset,
-               .init = hantro_mpeg2_dec_init,
-               .exit = hantro_mpeg2_dec_exit,
-       },
-       [HANTRO_MODE_VP8_DEC] = {
-               .run = rk3399_vpu_vp8_dec_run,
-               .reset = rk3399_vpu_dec_reset,
-               .init = hantro_vp8_dec_init,
-               .exit = hantro_vp8_dec_exit,
-       },
-};
-
-/*
- * VPU variant.
- */
-
-static const struct hantro_irq rk3399_irqs[] = {
-       { "vepu", rk3399_vepu_irq },
-       { "vdpu", rk3399_vdpu_irq },
-};
-
-static const char * const rk3399_clk_names[] = {
-       "aclk", "hclk"
-};
-
-const struct hantro_variant rk3399_vpu_variant = {
-       .enc_offset = 0x0,
-       .enc_fmts = rk3399_vpu_enc_fmts,
-       .num_enc_fmts = ARRAY_SIZE(rk3399_vpu_enc_fmts),
-       .dec_offset = 0x400,
-       .dec_fmts = rk3399_vpu_dec_fmts,
-       .num_dec_fmts = ARRAY_SIZE(rk3399_vpu_dec_fmts),
-       .codec = HANTRO_JPEG_ENCODER | HANTRO_MPEG2_DECODER |
-                HANTRO_VP8_DECODER,
-       .codec_ops = rk3399_vpu_codec_ops,
-       .irqs = rk3399_irqs,
-       .num_irqs = ARRAY_SIZE(rk3399_irqs),
-       .init = rk3399_vpu_hw_init,
-       .clk_names = rk3399_clk_names,
-       .num_clocks = ARRAY_SIZE(rk3399_clk_names)
-};
-
-static const struct hantro_irq rk3328_irqs[] = {
-       { "vdpu", rk3399_vdpu_irq },
-};
-
-const struct hantro_variant rk3328_vpu_variant = {
-       .dec_offset = 0x400,
-       .dec_fmts = rk3399_vpu_dec_fmts,
-       .num_dec_fmts = ARRAY_SIZE(rk3399_vpu_dec_fmts),
-       .codec = HANTRO_MPEG2_DECODER | HANTRO_VP8_DECODER,
-       .codec_ops = rk3399_vpu_codec_ops,
-       .irqs = rk3328_irqs,
-       .num_irqs = ARRAY_SIZE(rk3328_irqs),
-       .init = rk3399_vpu_hw_init,
-       .clk_names = rk3399_clk_names,
-       .num_clocks = ARRAY_SIZE(rk3399_clk_names),
-};
diff --git a/drivers/staging/media/hantro/rk3399_vpu_hw_jpeg_enc.c b/drivers/staging/media/hantro/rk3399_vpu_hw_jpeg_enc.c
deleted file mode 100644 (file)
index 3a27ebe..0000000
+++ /dev/null
@@ -1,173 +0,0 @@
-// SPDX-License-Identifier: GPL-2.0
-/*
- * Hantro VPU codec driver
- *
- * Copyright (C) 2018 Rockchip Electronics Co., Ltd.
- *
- * JPEG encoder
- * ------------
- * The VPU JPEG encoder produces JPEG baseline sequential format.
- * The quantization coefficients are 8-bit values, complying with
- * the baseline specification. Therefore, it requires
- * luma and chroma quantization tables. The hardware does entropy
- * encoding using internal Huffman tables, as specified in the JPEG
- * specification.
- *
- * In other words, only the luma and chroma quantization tables are
- * required for the encoding operation.
- *
- * Quantization luma table values are written to registers
- * VEPU_swreg_0-VEPU_swreg_15, and chroma table values to
- * VEPU_swreg_16-VEPU_swreg_31. A special order is needed, neither
- * zigzag, nor linear.
- */
-
-#include <asm/unaligned.h>
-#include <media/v4l2-mem2mem.h>
-#include "hantro_jpeg.h"
-#include "hantro.h"
-#include "hantro_v4l2.h"
-#include "hantro_hw.h"
-#include "rk3399_vpu_regs.h"
-
-#define VEPU_JPEG_QUANT_TABLE_COUNT 16
-
-static void rk3399_vpu_set_src_img_ctrl(struct hantro_dev *vpu,
-                                       struct hantro_ctx *ctx)
-{
-       struct v4l2_pix_format_mplane *pix_fmt = &ctx->src_fmt;
-       u32 reg;
-
-       /*
-        * The pix fmt width/height are already macroblock aligned
-        * by .vidioc_s_fmt_vid_cap_mplane() callback
-        */
-       reg = VEPU_REG_IN_IMG_CTRL_ROW_LEN(pix_fmt->width);
-       vepu_write_relaxed(vpu, reg, VEPU_REG_INPUT_LUMA_INFO);
-
-       reg = VEPU_REG_IN_IMG_CTRL_OVRFLR_D4(0) |
-             VEPU_REG_IN_IMG_CTRL_OVRFLB(0);
-       /*
-        * This register controls the input crop, as the offset
-        * from the right/bottom within the last macroblock. The offset from the
-        * right must be divided by 4 and so the crop must be aligned to 4 pixels
-        * horizontally.
-        */
-       vepu_write_relaxed(vpu, reg, VEPU_REG_ENC_OVER_FILL_STRM_OFFSET);
-
-       reg = VEPU_REG_IN_IMG_CTRL_FMT(ctx->vpu_src_fmt->enc_fmt);
-       vepu_write_relaxed(vpu, reg, VEPU_REG_ENC_CTRL1);
-}
-
-static void rk3399_vpu_jpeg_enc_set_buffers(struct hantro_dev *vpu,
-                                           struct hantro_ctx *ctx,
-                                           struct vb2_buffer *src_buf)
-{
-       struct v4l2_pix_format_mplane *pix_fmt = &ctx->src_fmt;
-       dma_addr_t src[3];
-
-       WARN_ON(pix_fmt->num_planes > 3);
-
-       vepu_write_relaxed(vpu, ctx->jpeg_enc.bounce_buffer.dma,
-                          VEPU_REG_ADDR_OUTPUT_STREAM);
-       vepu_write_relaxed(vpu, ctx->jpeg_enc.bounce_buffer.size,
-                          VEPU_REG_STR_BUF_LIMIT);
-
-       if (pix_fmt->num_planes == 1) {
-               src[0] = vb2_dma_contig_plane_dma_addr(src_buf, 0);
-               vepu_write_relaxed(vpu, src[0], VEPU_REG_ADDR_IN_PLANE_0);
-       } else if (pix_fmt->num_planes == 2) {
-               src[0] = vb2_dma_contig_plane_dma_addr(src_buf, 0);
-               src[1] = vb2_dma_contig_plane_dma_addr(src_buf, 1);
-               vepu_write_relaxed(vpu, src[0], VEPU_REG_ADDR_IN_PLANE_0);
-               vepu_write_relaxed(vpu, src[1], VEPU_REG_ADDR_IN_PLANE_1);
-       } else {
-               src[0] = vb2_dma_contig_plane_dma_addr(src_buf, 0);
-               src[1] = vb2_dma_contig_plane_dma_addr(src_buf, 1);
-               src[2] = vb2_dma_contig_plane_dma_addr(src_buf, 2);
-               vepu_write_relaxed(vpu, src[0], VEPU_REG_ADDR_IN_PLANE_0);
-               vepu_write_relaxed(vpu, src[1], VEPU_REG_ADDR_IN_PLANE_1);
-               vepu_write_relaxed(vpu, src[2], VEPU_REG_ADDR_IN_PLANE_2);
-       }
-}
-
-static void
-rk3399_vpu_jpeg_enc_set_qtable(struct hantro_dev *vpu,
-                              unsigned char *luma_qtable,
-                              unsigned char *chroma_qtable)
-{
-       u32 reg, i;
-       __be32 *luma_qtable_p;
-       __be32 *chroma_qtable_p;
-
-       luma_qtable_p = (__be32 *)luma_qtable;
-       chroma_qtable_p = (__be32 *)chroma_qtable;
-
-       /*
-        * Quantization table registers must be written in contiguous blocks.
-        * DO NOT collapse the below two "for" loops into one.
-        */
-       for (i = 0; i < VEPU_JPEG_QUANT_TABLE_COUNT; i++) {
-               reg = get_unaligned_be32(&luma_qtable_p[i]);
-               vepu_write_relaxed(vpu, reg, VEPU_REG_JPEG_LUMA_QUAT(i));
-       }
-
-       for (i = 0; i < VEPU_JPEG_QUANT_TABLE_COUNT; i++) {
-               reg = get_unaligned_be32(&chroma_qtable_p[i]);
-               vepu_write_relaxed(vpu, reg, VEPU_REG_JPEG_CHROMA_QUAT(i));
-       }
-}
-
-int rk3399_vpu_jpeg_enc_run(struct hantro_ctx *ctx)
-{
-       struct hantro_dev *vpu = ctx->dev;
-       struct vb2_v4l2_buffer *src_buf, *dst_buf;
-       struct hantro_jpeg_ctx jpeg_ctx;
-       u32 reg;
-
-       src_buf = hantro_get_src_buf(ctx);
-       dst_buf = hantro_get_dst_buf(ctx);
-
-       hantro_start_prepare_run(ctx);
-
-       memset(&jpeg_ctx, 0, sizeof(jpeg_ctx));
-       jpeg_ctx.buffer = vb2_plane_vaddr(&dst_buf->vb2_buf, 0);
-       jpeg_ctx.width = ctx->dst_fmt.width;
-       jpeg_ctx.height = ctx->dst_fmt.height;
-       jpeg_ctx.quality = ctx->jpeg_quality;
-       hantro_jpeg_header_assemble(&jpeg_ctx);
-
-       /* Switch to JPEG encoder mode before writing registers */
-       vepu_write_relaxed(vpu, VEPU_REG_ENCODE_FORMAT_JPEG,
-                          VEPU_REG_ENCODE_START);
-
-       rk3399_vpu_set_src_img_ctrl(vpu, ctx);
-       rk3399_vpu_jpeg_enc_set_buffers(vpu, ctx, &src_buf->vb2_buf);
-       rk3399_vpu_jpeg_enc_set_qtable(vpu,
-                                      hantro_jpeg_get_qtable(0),
-                                      hantro_jpeg_get_qtable(1));
-
-       reg = VEPU_REG_OUTPUT_SWAP32
-               | VEPU_REG_OUTPUT_SWAP16
-               | VEPU_REG_OUTPUT_SWAP8
-               | VEPU_REG_INPUT_SWAP8
-               | VEPU_REG_INPUT_SWAP16
-               | VEPU_REG_INPUT_SWAP32;
-       /* Make sure that all registers are written at this point. */
-       vepu_write(vpu, reg, VEPU_REG_DATA_ENDIAN);
-
-       reg = VEPU_REG_AXI_CTRL_BURST_LEN(16);
-       vepu_write_relaxed(vpu, reg, VEPU_REG_AXI_CTRL);
-
-       reg = VEPU_REG_MB_WIDTH(MB_WIDTH(ctx->src_fmt.width))
-               | VEPU_REG_MB_HEIGHT(MB_HEIGHT(ctx->src_fmt.height))
-               | VEPU_REG_FRAME_TYPE_INTRA
-               | VEPU_REG_ENCODE_FORMAT_JPEG
-               | VEPU_REG_ENCODE_ENABLE;
-
-       /* Kick the watchdog and start encoding */
-       hantro_end_prepare_run(ctx);
-       vepu_write(vpu, reg, VEPU_REG_ENCODE_START);
-
-       return 0;
-}
diff --git a/drivers/staging/media/hantro/rk3399_vpu_hw_mpeg2_dec.c b/drivers/staging/media/hantro/rk3399_vpu_hw_mpeg2_dec.c
deleted file mode 100644 (file)
index 683982c..0000000
+++ /dev/null
@@ -1,249 +0,0 @@
-// SPDX-License-Identifier: GPL-2.0
-/*
- * Hantro VPU codec driver
- *
- * Copyright (C) 2018 Rockchip Electronics Co., Ltd.
- */
-
-#include <asm/unaligned.h>
-#include <linux/bitfield.h>
-#include <media/v4l2-mem2mem.h>
-#include "hantro.h"
-#include "hantro_hw.h"
-
-#define VDPU_SWREG(nr)                 ((nr) * 4)
-
-#define VDPU_REG_DEC_OUT_BASE          VDPU_SWREG(63)
-#define VDPU_REG_RLC_VLC_BASE          VDPU_SWREG(64)
-#define VDPU_REG_QTABLE_BASE           VDPU_SWREG(61)
-#define VDPU_REG_REFER0_BASE           VDPU_SWREG(131)
-#define VDPU_REG_REFER2_BASE           VDPU_SWREG(134)
-#define VDPU_REG_REFER3_BASE           VDPU_SWREG(135)
-#define VDPU_REG_REFER1_BASE           VDPU_SWREG(148)
-#define VDPU_REG_DEC_E(v)              ((v) ? BIT(0) : 0)
-
-#define VDPU_REG_DEC_ADV_PRE_DIS(v)    ((v) ? BIT(11) : 0)
-#define VDPU_REG_DEC_SCMD_DIS(v)       ((v) ? BIT(10) : 0)
-#define VDPU_REG_FILTERING_DIS(v)      ((v) ? BIT(8) : 0)
-#define VDPU_REG_DEC_LATENCY(v)                (((v) << 1) & GENMASK(6, 1))
-
-#define VDPU_REG_INIT_QP(v)            (((v) << 25) & GENMASK(30, 25))
-#define VDPU_REG_STREAM_LEN(v)         (((v) << 0) & GENMASK(23, 0))
-
-#define VDPU_REG_APF_THRESHOLD(v)      (((v) << 17) & GENMASK(30, 17))
-#define VDPU_REG_STARTMB_X(v)          (((v) << 8) & GENMASK(16, 8))
-#define VDPU_REG_STARTMB_Y(v)          (((v) << 0) & GENMASK(7, 0))
-
-#define VDPU_REG_DEC_MODE(v)           (((v) << 0) & GENMASK(3, 0))
-
-#define VDPU_REG_DEC_STRENDIAN_E(v)    ((v) ? BIT(5) : 0)
-#define VDPU_REG_DEC_STRSWAP32_E(v)    ((v) ? BIT(4) : 0)
-#define VDPU_REG_DEC_OUTSWAP32_E(v)    ((v) ? BIT(3) : 0)
-#define VDPU_REG_DEC_INSWAP32_E(v)     ((v) ? BIT(2) : 0)
-#define VDPU_REG_DEC_OUT_ENDIAN(v)     ((v) ? BIT(1) : 0)
-#define VDPU_REG_DEC_IN_ENDIAN(v)      ((v) ? BIT(0) : 0)
-
-#define VDPU_REG_DEC_DATA_DISC_E(v)    ((v) ? BIT(22) : 0)
-#define VDPU_REG_DEC_MAX_BURST(v)      (((v) << 16) & GENMASK(20, 16))
-#define VDPU_REG_DEC_AXI_WR_ID(v)      (((v) << 8) & GENMASK(15, 8))
-#define VDPU_REG_DEC_AXI_RD_ID(v)      (((v) << 0) & GENMASK(7, 0))
-
-#define VDPU_REG_RLC_MODE_E(v)         ((v) ? BIT(20) : 0)
-#define VDPU_REG_PIC_INTERLACE_E(v)    ((v) ? BIT(17) : 0)
-#define VDPU_REG_PIC_FIELDMODE_E(v)    ((v) ? BIT(16) : 0)
-#define VDPU_REG_PIC_B_E(v)            ((v) ? BIT(15) : 0)
-#define VDPU_REG_PIC_INTER_E(v)                ((v) ? BIT(14) : 0)
-#define VDPU_REG_PIC_TOPFIELD_E(v)     ((v) ? BIT(13) : 0)
-#define VDPU_REG_FWD_INTERLACE_E(v)    ((v) ? BIT(12) : 0)
-#define VDPU_REG_WRITE_MVS_E(v)                ((v) ? BIT(10) : 0)
-#define VDPU_REG_DEC_TIMEOUT_E(v)      ((v) ? BIT(5) : 0)
-#define VDPU_REG_DEC_CLK_GATE_E(v)     ((v) ? BIT(4) : 0)
-
-#define VDPU_REG_PIC_MB_WIDTH(v)       (((v) << 23) & GENMASK(31, 23))
-#define VDPU_REG_PIC_MB_HEIGHT_P(v)    (((v) << 11) & GENMASK(18, 11))
-#define VDPU_REG_ALT_SCAN_E(v)         ((v) ? BIT(6) : 0)
-#define VDPU_REG_TOPFIELDFIRST_E(v)    ((v) ? BIT(5) : 0)
-
-#define VDPU_REG_STRM_START_BIT(v)     (((v) << 26) & GENMASK(31, 26))
-#define VDPU_REG_QSCALE_TYPE(v)                ((v) ? BIT(24) : 0)
-#define VDPU_REG_CON_MV_E(v)           ((v) ? BIT(4) : 0)
-#define VDPU_REG_INTRA_DC_PREC(v)      (((v) << 2) & GENMASK(3, 2))
-#define VDPU_REG_INTRA_VLC_TAB(v)      ((v) ? BIT(1) : 0)
-#define VDPU_REG_FRAME_PRED_DCT(v)     ((v) ? BIT(0) : 0)
-
-#define VDPU_REG_ALT_SCAN_FLAG_E(v)    ((v) ? BIT(19) : 0)
-#define VDPU_REG_FCODE_FWD_HOR(v)      (((v) << 15) & GENMASK(18, 15))
-#define VDPU_REG_FCODE_FWD_VER(v)      (((v) << 11) & GENMASK(14, 11))
-#define VDPU_REG_FCODE_BWD_HOR(v)      (((v) << 7) & GENMASK(10, 7))
-#define VDPU_REG_FCODE_BWD_VER(v)      (((v) << 3) & GENMASK(6, 3))
-#define VDPU_REG_MV_ACCURACY_FWD(v)    ((v) ? BIT(2) : 0)
-#define VDPU_REG_MV_ACCURACY_BWD(v)    ((v) ? BIT(1) : 0)
-
-static void
-rk3399_vpu_mpeg2_dec_set_quantisation(struct hantro_dev *vpu,
-                                     struct hantro_ctx *ctx)
-{
-       struct v4l2_ctrl_mpeg2_quantisation *q;
-
-       q = hantro_get_ctrl(ctx, V4L2_CID_STATELESS_MPEG2_QUANTISATION);
-       hantro_mpeg2_dec_copy_qtable(ctx->mpeg2_dec.qtable.cpu, q);
-       vdpu_write_relaxed(vpu, ctx->mpeg2_dec.qtable.dma, VDPU_REG_QTABLE_BASE);
-}
-
-static void
-rk3399_vpu_mpeg2_dec_set_buffers(struct hantro_dev *vpu,
-                                struct hantro_ctx *ctx,
-                                struct vb2_buffer *src_buf,
-                                struct vb2_buffer *dst_buf,
-                                const struct v4l2_ctrl_mpeg2_sequence *seq,
-                                const struct v4l2_ctrl_mpeg2_picture *pic)
-{
-       dma_addr_t forward_addr = 0, backward_addr = 0;
-       dma_addr_t current_addr, addr;
-
-       switch (pic->picture_coding_type) {
-       case V4L2_MPEG2_PIC_CODING_TYPE_B:
-               backward_addr = hantro_get_ref(ctx, pic->backward_ref_ts);
-               fallthrough;
-       case V4L2_MPEG2_PIC_CODING_TYPE_P:
-               forward_addr = hantro_get_ref(ctx, pic->forward_ref_ts);
-       }
-
-       /* Source bitstream buffer */
-       addr = vb2_dma_contig_plane_dma_addr(src_buf, 0);
-       vdpu_write_relaxed(vpu, addr, VDPU_REG_RLC_VLC_BASE);
-
-       /* Destination frame buffer */
-       addr = vb2_dma_contig_plane_dma_addr(dst_buf, 0);
-       current_addr = addr;
-
-       if (pic->picture_structure == V4L2_MPEG2_PIC_BOTTOM_FIELD)
-               addr += ALIGN(ctx->dst_fmt.width, 16);
-       vdpu_write_relaxed(vpu, addr, VDPU_REG_DEC_OUT_BASE);
-
-       if (!forward_addr)
-               forward_addr = current_addr;
-       if (!backward_addr)
-               backward_addr = current_addr;
-
-       /* Set forward ref frame (top/bottom field) */
-       if (pic->picture_structure == V4L2_MPEG2_PIC_FRAME ||
-           pic->picture_coding_type == V4L2_MPEG2_PIC_CODING_TYPE_B ||
-           (pic->picture_structure == V4L2_MPEG2_PIC_TOP_FIELD &&
-            pic->flags & V4L2_MPEG2_PIC_TOP_FIELD) ||
-           (pic->picture_structure == V4L2_MPEG2_PIC_BOTTOM_FIELD &&
-            !(pic->flags & V4L2_MPEG2_PIC_TOP_FIELD))) {
-               vdpu_write_relaxed(vpu, forward_addr, VDPU_REG_REFER0_BASE);
-               vdpu_write_relaxed(vpu, forward_addr, VDPU_REG_REFER1_BASE);
-       } else if (pic->picture_structure == V4L2_MPEG2_PIC_TOP_FIELD) {
-               vdpu_write_relaxed(vpu, forward_addr, VDPU_REG_REFER0_BASE);
-               vdpu_write_relaxed(vpu, current_addr, VDPU_REG_REFER1_BASE);
-       } else if (pic->picture_structure == V4L2_MPEG2_PIC_BOTTOM_FIELD) {
-               vdpu_write_relaxed(vpu, current_addr, VDPU_REG_REFER0_BASE);
-               vdpu_write_relaxed(vpu, forward_addr, VDPU_REG_REFER1_BASE);
-       }
-
-       /* Set backward ref frame (top/bottom field) */
-       vdpu_write_relaxed(vpu, backward_addr, VDPU_REG_REFER2_BASE);
-       vdpu_write_relaxed(vpu, backward_addr, VDPU_REG_REFER3_BASE);
-}
-
-int rk3399_vpu_mpeg2_dec_run(struct hantro_ctx *ctx)
-{
-       struct hantro_dev *vpu = ctx->dev;
-       struct vb2_v4l2_buffer *src_buf, *dst_buf;
-       const struct v4l2_ctrl_mpeg2_sequence *seq;
-       const struct v4l2_ctrl_mpeg2_picture *pic;
-       u32 reg;
-
-       src_buf = hantro_get_src_buf(ctx);
-       dst_buf = hantro_get_dst_buf(ctx);
-
-       hantro_start_prepare_run(ctx);
-
-       seq = hantro_get_ctrl(ctx,
-                             V4L2_CID_STATELESS_MPEG2_SEQUENCE);
-       pic = hantro_get_ctrl(ctx,
-                             V4L2_CID_STATELESS_MPEG2_PICTURE);
-
-       reg = VDPU_REG_DEC_ADV_PRE_DIS(0) |
-             VDPU_REG_DEC_SCMD_DIS(0) |
-             VDPU_REG_FILTERING_DIS(1) |
-             VDPU_REG_DEC_LATENCY(0);
-       vdpu_write_relaxed(vpu, reg, VDPU_SWREG(50));
-
-       reg = VDPU_REG_INIT_QP(1) |
-             VDPU_REG_STREAM_LEN(vb2_get_plane_payload(&src_buf->vb2_buf, 0));
-       vdpu_write_relaxed(vpu, reg, VDPU_SWREG(51));
-
-       reg = VDPU_REG_APF_THRESHOLD(8) |
-             VDPU_REG_STARTMB_X(0) |
-             VDPU_REG_STARTMB_Y(0);
-       vdpu_write_relaxed(vpu, reg, VDPU_SWREG(52));
-
-       reg = VDPU_REG_DEC_MODE(5);
-       vdpu_write_relaxed(vpu, reg, VDPU_SWREG(53));
-
-       reg = VDPU_REG_DEC_STRENDIAN_E(1) |
-             VDPU_REG_DEC_STRSWAP32_E(1) |
-             VDPU_REG_DEC_OUTSWAP32_E(1) |
-             VDPU_REG_DEC_INSWAP32_E(1) |
-             VDPU_REG_DEC_OUT_ENDIAN(1) |
-             VDPU_REG_DEC_IN_ENDIAN(1);
-       vdpu_write_relaxed(vpu, reg, VDPU_SWREG(54));
-
-       reg = VDPU_REG_DEC_DATA_DISC_E(0) |
-             VDPU_REG_DEC_MAX_BURST(16) |
-             VDPU_REG_DEC_AXI_WR_ID(0) |
-             VDPU_REG_DEC_AXI_RD_ID(0);
-       vdpu_write_relaxed(vpu, reg, VDPU_SWREG(56));
-
-       reg = VDPU_REG_RLC_MODE_E(0) |
-             VDPU_REG_PIC_INTERLACE_E(!(seq->flags & V4L2_MPEG2_SEQ_FLAG_PROGRESSIVE)) |
-             VDPU_REG_PIC_FIELDMODE_E(pic->picture_structure != V4L2_MPEG2_PIC_FRAME) |
-             VDPU_REG_PIC_B_E(pic->picture_coding_type == V4L2_MPEG2_PIC_CODING_TYPE_B) |
-             VDPU_REG_PIC_INTER_E(pic->picture_coding_type != V4L2_MPEG2_PIC_CODING_TYPE_I) |
-             VDPU_REG_PIC_TOPFIELD_E(pic->picture_structure == V4L2_MPEG2_PIC_TOP_FIELD) |
-             VDPU_REG_FWD_INTERLACE_E(0) |
-             VDPU_REG_WRITE_MVS_E(0) |
-             VDPU_REG_DEC_TIMEOUT_E(1) |
-             VDPU_REG_DEC_CLK_GATE_E(1);
-       vdpu_write_relaxed(vpu, reg, VDPU_SWREG(57));
-
-       reg = VDPU_REG_PIC_MB_WIDTH(MB_WIDTH(ctx->dst_fmt.width)) |
-             VDPU_REG_PIC_MB_HEIGHT_P(MB_HEIGHT(ctx->dst_fmt.height)) |
-             VDPU_REG_ALT_SCAN_E(pic->flags & V4L2_MPEG2_PIC_FLAG_ALT_SCAN) |
-             VDPU_REG_TOPFIELDFIRST_E(pic->flags & V4L2_MPEG2_PIC_FLAG_TOP_FIELD_FIRST);
-       vdpu_write_relaxed(vpu, reg, VDPU_SWREG(120));
-
-       reg = VDPU_REG_STRM_START_BIT(0) |
-             VDPU_REG_QSCALE_TYPE(pic->flags & V4L2_MPEG2_PIC_FLAG_Q_SCALE_TYPE) |
-             VDPU_REG_CON_MV_E(pic->flags & V4L2_MPEG2_PIC_FLAG_CONCEALMENT_MV) |
-             VDPU_REG_INTRA_DC_PREC(pic->intra_dc_precision) |
-             VDPU_REG_INTRA_VLC_TAB(pic->flags & V4L2_MPEG2_PIC_FLAG_INTRA_VLC) |
-             VDPU_REG_FRAME_PRED_DCT(pic->flags & V4L2_MPEG2_PIC_FLAG_FRAME_PRED_DCT);
-       vdpu_write_relaxed(vpu, reg, VDPU_SWREG(122));
-
-       reg = VDPU_REG_ALT_SCAN_FLAG_E(pic->flags & V4L2_MPEG2_PIC_FLAG_ALT_SCAN) |
-             VDPU_REG_FCODE_FWD_HOR(pic->f_code[0][0]) |
-             VDPU_REG_FCODE_FWD_VER(pic->f_code[0][1]) |
-             VDPU_REG_FCODE_BWD_HOR(pic->f_code[1][0]) |
-             VDPU_REG_FCODE_BWD_VER(pic->f_code[1][1]) |
-             VDPU_REG_MV_ACCURACY_FWD(1) |
-             VDPU_REG_MV_ACCURACY_BWD(1);
-       vdpu_write_relaxed(vpu, reg, VDPU_SWREG(136));
-
-       rk3399_vpu_mpeg2_dec_set_quantisation(vpu, ctx);
-
-       rk3399_vpu_mpeg2_dec_set_buffers(vpu, ctx, &src_buf->vb2_buf,
-                                        &dst_buf->vb2_buf,
-                                        seq, pic);
-
-       /* Kick the watchdog and start decoding */
-       hantro_end_prepare_run(ctx);
-
-       reg = vdpu_read(vpu, VDPU_SWREG(57)) | VDPU_REG_DEC_E(1);
-       vdpu_write(vpu, reg, VDPU_SWREG(57));
-
-       return 0;
-}
diff --git a/drivers/staging/media/hantro/rk3399_vpu_hw_vp8_dec.c b/drivers/staging/media/hantro/rk3399_vpu_hw_vp8_dec.c
deleted file mode 100644 (file)
index e5d20fe..0000000
+++ /dev/null
@@ -1,594 +0,0 @@
-// SPDX-License-Identifier: GPL-2.0
-/*
- * Rockchip VPU codec vp8 decode driver
- *
- * Copyright (C) 2014 Rockchip Electronics Co., Ltd.
- *     ZhiChao Yu <zhichao.yu@rock-chips.com>
- *
- * Copyright (C) 2014 Google LLC.
- *      Tomasz Figa <tfiga@chromium.org>
- *
- * Copyright (C) 2015 Rockchip Electronics Co., Ltd.
- *      Alpha Lin <alpha.lin@rock-chips.com>
- */
-
-#include <media/v4l2-mem2mem.h>
-
-#include "hantro_hw.h"
-#include "hantro.h"
-#include "hantro_g1_regs.h"
-
-#define VDPU_REG_DEC_CTRL0                     0x0c8
-#define VDPU_REG_STREAM_LEN                    0x0cc
-#define VDPU_REG_DEC_FORMAT                    0x0d4
-#define     VDPU_REG_DEC_CTRL0_DEC_MODE(x)             (((x) & 0xf) << 0)
-#define VDPU_REG_DATA_ENDIAN                   0x0d8
-#define     VDPU_REG_CONFIG_DEC_STRENDIAN_E            BIT(5)
-#define     VDPU_REG_CONFIG_DEC_STRSWAP32_E            BIT(4)
-#define     VDPU_REG_CONFIG_DEC_OUTSWAP32_E            BIT(3)
-#define     VDPU_REG_CONFIG_DEC_INSWAP32_E             BIT(2)
-#define     VDPU_REG_CONFIG_DEC_OUT_ENDIAN             BIT(1)
-#define     VDPU_REG_CONFIG_DEC_IN_ENDIAN              BIT(0)
-#define VDPU_REG_AXI_CTRL                      0x0e0
-#define     VDPU_REG_CONFIG_DEC_MAX_BURST(x)           (((x) & 0x1f) << 16)
-#define VDPU_REG_EN_FLAGS                      0x0e4
-#define     VDPU_REG_DEC_CTRL0_PIC_INTER_E             BIT(14)
-#define     VDPU_REG_CONFIG_DEC_TIMEOUT_E              BIT(5)
-#define     VDPU_REG_CONFIG_DEC_CLK_GATE_E             BIT(4)
-#define VDPU_REG_PRED_FLT                      0x0ec
-#define VDPU_REG_ADDR_QTABLE                   0x0f4
-#define VDPU_REG_ADDR_DST                      0x0fc
-#define VDPU_REG_ADDR_STR                      0x100
-#define VDPU_REG_VP8_PIC_MB_SIZE               0x1e0
-#define VDPU_REG_VP8_DCT_START_BIT             0x1e4
-#define     VDPU_REG_DEC_CTRL4_VC1_HEIGHT_EXT          BIT(13)
-#define     VDPU_REG_DEC_CTRL4_BILIN_MC_E              BIT(12)
-#define VDPU_REG_VP8_CTRL0                     0x1e8
-#define VDPU_REG_VP8_DATA_VAL                  0x1f0
-#define VDPU_REG_PRED_FLT7                     0x1f4
-#define VDPU_REG_PRED_FLT8                     0x1f8
-#define VDPU_REG_PRED_FLT9                     0x1fc
-#define VDPU_REG_PRED_FLT10                    0x200
-#define VDPU_REG_FILTER_LEVEL                  0x204
-#define VDPU_REG_VP8_QUANTER0                  0x208
-#define VDPU_REG_VP8_ADDR_REF0                 0x20c
-#define VDPU_REG_FILTER_MB_ADJ                 0x210
-#define     VDPU_REG_REF_PIC_FILT_TYPE_E               BIT(31)
-#define     VDPU_REG_REF_PIC_FILT_SHARPNESS(x)         (((x) & 0x7) << 28)
-#define VDPU_REG_FILTER_REF_ADJ                        0x214
-#define VDPU_REG_VP8_ADDR_REF2_5(i)            (0x218 + ((i) * 0x4))
-#define     VDPU_REG_VP8_GREF_SIGN_BIAS                        BIT(0)
-#define     VDPU_REG_VP8_AREF_SIGN_BIAS                        BIT(0)
-#define VDPU_REG_VP8_DCT_BASE(i)               \
-               (0x230 + ((((i) < 5) ? (i) : ((i) + 1)) * 0x4))
-#define VDPU_REG_VP8_ADDR_CTRL_PART            0x244
-#define VDPU_REG_VP8_SEGMENT_VAL               0x254
-#define     VDPU_REG_FWD_PIC1_SEGMENT_BASE(x)          ((x) << 0)
-#define     VDPU_REG_FWD_PIC1_SEGMENT_UPD_E            BIT(1)
-#define     VDPU_REG_FWD_PIC1_SEGMENT_E                        BIT(0)
-#define VDPU_REG_VP8_DCT_START_BIT2            0x258
-#define VDPU_REG_VP8_QUANTER1                  0x25c
-#define VDPU_REG_VP8_QUANTER2                  0x260
-#define VDPU_REG_PRED_FLT1                     0x264
-#define VDPU_REG_PRED_FLT2                     0x268
-#define VDPU_REG_PRED_FLT3                     0x26c
-#define VDPU_REG_PRED_FLT4                     0x270
-#define VDPU_REG_PRED_FLT5                     0x274
-#define VDPU_REG_PRED_FLT6                     0x278
-
-static const struct hantro_reg vp8_dec_dct_base[8] = {
-       { VDPU_REG_ADDR_STR, 0, 0xffffffff },
-       { VDPU_REG_VP8_DCT_BASE(0), 0, 0xffffffff },
-       { VDPU_REG_VP8_DCT_BASE(1), 0, 0xffffffff },
-       { VDPU_REG_VP8_DCT_BASE(2), 0, 0xffffffff },
-       { VDPU_REG_VP8_DCT_BASE(3), 0, 0xffffffff },
-       { VDPU_REG_VP8_DCT_BASE(4), 0, 0xffffffff },
-       { VDPU_REG_VP8_DCT_BASE(5), 0, 0xffffffff },
-       { VDPU_REG_VP8_DCT_BASE(6), 0, 0xffffffff },
-};
-
-static const struct hantro_reg vp8_dec_lf_level[4] = {
-       { VDPU_REG_FILTER_LEVEL, 18, 0x3f },
-       { VDPU_REG_FILTER_LEVEL, 12, 0x3f },
-       { VDPU_REG_FILTER_LEVEL, 6, 0x3f },
-       { VDPU_REG_FILTER_LEVEL, 0, 0x3f },
-};
-
-static const struct hantro_reg vp8_dec_mb_adj[4] = {
-       { VDPU_REG_FILTER_MB_ADJ, 21, 0x7f },
-       { VDPU_REG_FILTER_MB_ADJ, 14, 0x7f },
-       { VDPU_REG_FILTER_MB_ADJ, 7, 0x7f },
-       { VDPU_REG_FILTER_MB_ADJ, 0, 0x7f },
-};
-
-static const struct hantro_reg vp8_dec_ref_adj[4] = {
-       { VDPU_REG_FILTER_REF_ADJ, 21, 0x7f },
-       { VDPU_REG_FILTER_REF_ADJ, 14, 0x7f },
-       { VDPU_REG_FILTER_REF_ADJ, 7, 0x7f },
-       { VDPU_REG_FILTER_REF_ADJ, 0, 0x7f },
-};
-
-static const struct hantro_reg vp8_dec_quant[4] = {
-       { VDPU_REG_VP8_QUANTER0, 11, 0x7ff },
-       { VDPU_REG_VP8_QUANTER0, 0, 0x7ff },
-       { VDPU_REG_VP8_QUANTER1, 11, 0x7ff },
-       { VDPU_REG_VP8_QUANTER1, 0, 0x7ff },
-};
-
-static const struct hantro_reg vp8_dec_quant_delta[5] = {
-       { VDPU_REG_VP8_QUANTER0, 27, 0x1f },
-       { VDPU_REG_VP8_QUANTER0, 22, 0x1f },
-       { VDPU_REG_VP8_QUANTER1, 27, 0x1f },
-       { VDPU_REG_VP8_QUANTER1, 22, 0x1f },
-       { VDPU_REG_VP8_QUANTER2, 27, 0x1f },
-};
-
-static const struct hantro_reg vp8_dec_dct_start_bits[8] = {
-       { VDPU_REG_VP8_CTRL0, 26, 0x3f },
-       { VDPU_REG_VP8_DCT_START_BIT, 26, 0x3f },
-       { VDPU_REG_VP8_DCT_START_BIT, 20, 0x3f },
-       { VDPU_REG_VP8_DCT_START_BIT2, 24, 0x3f },
-       { VDPU_REG_VP8_DCT_START_BIT2, 18, 0x3f },
-       { VDPU_REG_VP8_DCT_START_BIT2, 12, 0x3f },
-       { VDPU_REG_VP8_DCT_START_BIT2, 6, 0x3f },
-       { VDPU_REG_VP8_DCT_START_BIT2, 0, 0x3f },
-};
-
-static const struct hantro_reg vp8_dec_pred_bc_tap[8][6] = {
-       {
-               { 0, 0, 0},
-               { VDPU_REG_PRED_FLT, 22, 0x3ff },
-               { VDPU_REG_PRED_FLT, 12, 0x3ff },
-               { VDPU_REG_PRED_FLT, 2, 0x3ff },
-               { VDPU_REG_PRED_FLT1, 22, 0x3ff },
-               { 0, 0, 0},
-       }, {
-               { 0, 0, 0},
-               { VDPU_REG_PRED_FLT1, 12, 0x3ff },
-               { VDPU_REG_PRED_FLT1, 2, 0x3ff },
-               { VDPU_REG_PRED_FLT2, 22, 0x3ff },
-               { VDPU_REG_PRED_FLT2, 12, 0x3ff },
-               { 0, 0, 0},
-       }, {
-               { VDPU_REG_PRED_FLT10, 10, 0x3 },
-               { VDPU_REG_PRED_FLT2, 2, 0x3ff },
-               { VDPU_REG_PRED_FLT3, 22, 0x3ff },
-               { VDPU_REG_PRED_FLT3, 12, 0x3ff },
-               { VDPU_REG_PRED_FLT3, 2, 0x3ff },
-               { VDPU_REG_PRED_FLT10, 8, 0x3},
-       }, {
-               { 0, 0, 0},
-               { VDPU_REG_PRED_FLT4, 22, 0x3ff },
-               { VDPU_REG_PRED_FLT4, 12, 0x3ff },
-               { VDPU_REG_PRED_FLT4, 2, 0x3ff },
-               { VDPU_REG_PRED_FLT5, 22, 0x3ff },
-               { 0, 0, 0},
-       }, {
-               { VDPU_REG_PRED_FLT10, 6, 0x3 },
-               { VDPU_REG_PRED_FLT5, 12, 0x3ff },
-               { VDPU_REG_PRED_FLT5, 2, 0x3ff },
-               { VDPU_REG_PRED_FLT6, 22, 0x3ff },
-               { VDPU_REG_PRED_FLT6, 12, 0x3ff },
-               { VDPU_REG_PRED_FLT10, 4, 0x3 },
-       }, {
-               { 0, 0, 0},
-               { VDPU_REG_PRED_FLT6, 2, 0x3ff },
-               { VDPU_REG_PRED_FLT7, 22, 0x3ff },
-               { VDPU_REG_PRED_FLT7, 12, 0x3ff },
-               { VDPU_REG_PRED_FLT7, 2, 0x3ff },
-               { 0, 0, 0},
-       }, {
-               { VDPU_REG_PRED_FLT10, 2, 0x3 },
-               { VDPU_REG_PRED_FLT8, 22, 0x3ff },
-               { VDPU_REG_PRED_FLT8, 12, 0x3ff },
-               { VDPU_REG_PRED_FLT8, 2, 0x3ff },
-               { VDPU_REG_PRED_FLT9, 22, 0x3ff },
-               { VDPU_REG_PRED_FLT10, 0, 0x3 },
-       }, {
-               { 0, 0, 0},
-               { VDPU_REG_PRED_FLT9, 12, 0x3ff },
-               { VDPU_REG_PRED_FLT9, 2, 0x3ff },
-               { VDPU_REG_PRED_FLT10, 22, 0x3ff },
-               { VDPU_REG_PRED_FLT10, 12, 0x3ff },
-               { 0, 0, 0},
-       },
-};
-
-static const struct hantro_reg vp8_dec_mb_start_bit = {
-       .base = VDPU_REG_VP8_CTRL0,
-       .shift = 18,
-       .mask = 0x3f
-};
-
-static const struct hantro_reg vp8_dec_mb_aligned_data_len = {
-       .base = VDPU_REG_VP8_DATA_VAL,
-       .shift = 0,
-       .mask = 0x3fffff
-};
-
-static const struct hantro_reg vp8_dec_num_dct_partitions = {
-       .base = VDPU_REG_VP8_DATA_VAL,
-       .shift = 24,
-       .mask = 0xf
-};
-
-static const struct hantro_reg vp8_dec_stream_len = {
-       .base = VDPU_REG_STREAM_LEN,
-       .shift = 0,
-       .mask = 0xffffff
-};
-
-static const struct hantro_reg vp8_dec_mb_width = {
-       .base = VDPU_REG_VP8_PIC_MB_SIZE,
-       .shift = 23,
-       .mask = 0x1ff
-};
-
-static const struct hantro_reg vp8_dec_mb_height = {
-       .base = VDPU_REG_VP8_PIC_MB_SIZE,
-       .shift = 11,
-       .mask = 0xff
-};
-
-static const struct hantro_reg vp8_dec_mb_width_ext = {
-       .base = VDPU_REG_VP8_PIC_MB_SIZE,
-       .shift = 3,
-       .mask = 0x7
-};
-
-static const struct hantro_reg vp8_dec_mb_height_ext = {
-       .base = VDPU_REG_VP8_PIC_MB_SIZE,
-       .shift = 0,
-       .mask = 0x7
-};
-
-static const struct hantro_reg vp8_dec_bool_range = {
-       .base = VDPU_REG_VP8_CTRL0,
-       .shift = 0,
-       .mask = 0xff
-};
-
-static const struct hantro_reg vp8_dec_bool_value = {
-       .base = VDPU_REG_VP8_CTRL0,
-       .shift = 8,
-       .mask = 0xff
-};
-
-static const struct hantro_reg vp8_dec_filter_disable = {
-       .base = VDPU_REG_DEC_CTRL0,
-       .shift = 8,
-       .mask = 1
-};
-
-static const struct hantro_reg vp8_dec_skip_mode = {
-       .base = VDPU_REG_DEC_CTRL0,
-       .shift = 9,
-       .mask = 1
-};
-
-static const struct hantro_reg vp8_dec_start_dec = {
-       .base = VDPU_REG_EN_FLAGS,
-       .shift = 0,
-       .mask = 1
-};
-
-static void cfg_lf(struct hantro_ctx *ctx,
-                  const struct v4l2_ctrl_vp8_frame *hdr)
-{
-       const struct v4l2_vp8_segment *seg = &hdr->segment;
-       const struct v4l2_vp8_loop_filter *lf = &hdr->lf;
-       struct hantro_dev *vpu = ctx->dev;
-       unsigned int i;
-       u32 reg;
-
-       if (!(seg->flags & V4L2_VP8_SEGMENT_FLAG_ENABLED)) {
-               hantro_reg_write(vpu, &vp8_dec_lf_level[0], lf->level);
-       } else if (seg->flags & V4L2_VP8_SEGMENT_FLAG_DELTA_VALUE_MODE) {
-               for (i = 0; i < 4; i++) {
-                       u32 lf_level = clamp(lf->level + seg->lf_update[i],
-                                            0, 63);
-
-                       hantro_reg_write(vpu, &vp8_dec_lf_level[i], lf_level);
-               }
-       } else {
-               for (i = 0; i < 4; i++)
-                       hantro_reg_write(vpu, &vp8_dec_lf_level[i],
-                                        seg->lf_update[i]);
-       }
-
-       reg = VDPU_REG_REF_PIC_FILT_SHARPNESS(lf->sharpness_level);
-       if (lf->flags & V4L2_VP8_LF_FILTER_TYPE_SIMPLE)
-               reg |= VDPU_REG_REF_PIC_FILT_TYPE_E;
-       vdpu_write_relaxed(vpu, reg, VDPU_REG_FILTER_MB_ADJ);
-
-       if (lf->flags & V4L2_VP8_LF_ADJ_ENABLE) {
-               for (i = 0; i < 4; i++) {
-                       hantro_reg_write(vpu, &vp8_dec_mb_adj[i],
-                                        lf->mb_mode_delta[i]);
-                       hantro_reg_write(vpu, &vp8_dec_ref_adj[i],
-                                        lf->ref_frm_delta[i]);
-               }
-       }
-}
-
-static void cfg_qp(struct hantro_ctx *ctx,
-                  const struct v4l2_ctrl_vp8_frame *hdr)
-{
-       const struct v4l2_vp8_quantization *q = &hdr->quant;
-       const struct v4l2_vp8_segment *seg = &hdr->segment;
-       struct hantro_dev *vpu = ctx->dev;
-       unsigned int i;
-
-       if (!(seg->flags & V4L2_VP8_SEGMENT_FLAG_ENABLED)) {
-               hantro_reg_write(vpu, &vp8_dec_quant[0], q->y_ac_qi);
-       } else if (seg->flags & V4L2_VP8_SEGMENT_FLAG_DELTA_VALUE_MODE) {
-               for (i = 0; i < 4; i++) {
-                       u32 quant = clamp(q->y_ac_qi + seg->quant_update[i],
-                                         0, 127);
-
-                       hantro_reg_write(vpu, &vp8_dec_quant[i], quant);
-               }
-       } else {
-               for (i = 0; i < 4; i++)
-                       hantro_reg_write(vpu, &vp8_dec_quant[i],
-                                        seg->quant_update[i]);
-       }
-
-       hantro_reg_write(vpu, &vp8_dec_quant_delta[0], q->y_dc_delta);
-       hantro_reg_write(vpu, &vp8_dec_quant_delta[1], q->y2_dc_delta);
-       hantro_reg_write(vpu, &vp8_dec_quant_delta[2], q->y2_ac_delta);
-       hantro_reg_write(vpu, &vp8_dec_quant_delta[3], q->uv_dc_delta);
-       hantro_reg_write(vpu, &vp8_dec_quant_delta[4], q->uv_ac_delta);
-}
-
-static void cfg_parts(struct hantro_ctx *ctx,
-                     const struct v4l2_ctrl_vp8_frame *hdr)
-{
-       struct hantro_dev *vpu = ctx->dev;
-       struct vb2_v4l2_buffer *vb2_src;
-       u32 first_part_offset = V4L2_VP8_FRAME_IS_KEY_FRAME(hdr) ? 10 : 3;
-       u32 mb_size, mb_offset_bytes, mb_offset_bits, mb_start_bits;
-       u32 dct_size_part_size, dct_part_offset;
-       dma_addr_t src_dma;
-       u32 dct_part_total_len = 0;
-       u32 count = 0;
-       unsigned int i;
-
-       vb2_src = hantro_get_src_buf(ctx);
-       src_dma = vb2_dma_contig_plane_dma_addr(&vb2_src->vb2_buf, 0);
-
-       /*
-        * Calculate control partition mb data info
-        * @first_part_header_bits:     bits offset of mb data from first
-        *                              part start pos
-        * @mb_offset_bits:             bits offset of mb data from src_dma
-        *                              base addr
-        * @mb_offset_byte:             bytes offset of mb data from src_dma
-        *                              base addr
-        * @mb_start_bits:              bits offset of mb data from mb data
-        *                              64bits alignment addr
-        */
-       mb_offset_bits = first_part_offset * 8 +
-                        hdr->first_part_header_bits + 8;
-       mb_offset_bytes = mb_offset_bits / 8;
-       mb_start_bits = mb_offset_bits -
-                       (mb_offset_bytes & (~DEC_8190_ALIGN_MASK)) * 8;
-       mb_size = hdr->first_part_size -
-                 (mb_offset_bytes - first_part_offset) +
-                 (mb_offset_bytes & DEC_8190_ALIGN_MASK);
-
-       /* Macroblock data aligned base addr */
-       vdpu_write_relaxed(vpu, (mb_offset_bytes & (~DEC_8190_ALIGN_MASK)) +
-                          src_dma, VDPU_REG_VP8_ADDR_CTRL_PART);
-       hantro_reg_write(vpu, &vp8_dec_mb_start_bit, mb_start_bits);
-       hantro_reg_write(vpu, &vp8_dec_mb_aligned_data_len, mb_size);
-
-       /*
-        * Calculate DCT partition info
-        * @dct_size_part_size: Containing sizes of DCT part, every DCT part
-        *                      has 3 bytes to store its size, except the last
-        *                      DCT part
-        * @dct_part_offset:    bytes offset of DCT parts from src_dma base addr
-        * @dct_part_total_len: total size of all DCT parts
-        */
-       dct_size_part_size = (hdr->num_dct_parts - 1) * 3;
-       dct_part_offset = first_part_offset + hdr->first_part_size;
-       for (i = 0; i < hdr->num_dct_parts; i++)
-               dct_part_total_len += hdr->dct_part_sizes[i];
-       dct_part_total_len += dct_size_part_size;
-       dct_part_total_len += (dct_part_offset & DEC_8190_ALIGN_MASK);
-
-       /* Number of DCT partitions */
-       hantro_reg_write(vpu, &vp8_dec_num_dct_partitions,
-                        hdr->num_dct_parts - 1);
-
-       /* DCT partition length */
-       hantro_reg_write(vpu, &vp8_dec_stream_len, dct_part_total_len);
-
-       /* DCT partitions base address */
-       for (i = 0; i < hdr->num_dct_parts; i++) {
-               u32 byte_offset = dct_part_offset + dct_size_part_size + count;
-               u32 base_addr = byte_offset + src_dma;
-
-               hantro_reg_write(vpu, &vp8_dec_dct_base[i],
-                                base_addr & (~DEC_8190_ALIGN_MASK));
-
-               hantro_reg_write(vpu, &vp8_dec_dct_start_bits[i],
-                                (byte_offset & DEC_8190_ALIGN_MASK) * 8);
-
-               count += hdr->dct_part_sizes[i];
-       }
-}
-
-/*
- * prediction filter taps
- * normal 6-tap filters
- */
-static void cfg_tap(struct hantro_ctx *ctx,
-                   const struct v4l2_ctrl_vp8_frame *hdr)
-{
-       struct hantro_dev *vpu = ctx->dev;
-       int i, j;
-
-       if ((hdr->version & 0x03) != 0)
-               return; /* Tap filter not used. */
-
-       for (i = 0; i < 8; i++) {
-               for (j = 0; j < 6; j++) {
-                       if (vp8_dec_pred_bc_tap[i][j].base != 0)
-                               hantro_reg_write(vpu,
-                                                &vp8_dec_pred_bc_tap[i][j],
-                                                hantro_vp8_dec_mc_filter[i][j]);
-               }
-       }
-}
-
-static void cfg_ref(struct hantro_ctx *ctx,
-                   const struct v4l2_ctrl_vp8_frame *hdr)
-{
-       struct hantro_dev *vpu = ctx->dev;
-       struct vb2_v4l2_buffer *vb2_dst;
-       dma_addr_t ref;
-
-       vb2_dst = hantro_get_dst_buf(ctx);
-
-       ref = hantro_get_ref(ctx, hdr->last_frame_ts);
-       if (!ref)
-               ref = vb2_dma_contig_plane_dma_addr(&vb2_dst->vb2_buf, 0);
-       vdpu_write_relaxed(vpu, ref, VDPU_REG_VP8_ADDR_REF0);
-
-       ref = hantro_get_ref(ctx, hdr->golden_frame_ts);
-       WARN_ON(!ref && hdr->golden_frame_ts);
-       if (!ref)
-               ref = vb2_dma_contig_plane_dma_addr(&vb2_dst->vb2_buf, 0);
-       if (hdr->flags & V4L2_VP8_FRAME_FLAG_SIGN_BIAS_GOLDEN)
-               ref |= VDPU_REG_VP8_GREF_SIGN_BIAS;
-       vdpu_write_relaxed(vpu, ref, VDPU_REG_VP8_ADDR_REF2_5(2));
-
-       ref = hantro_get_ref(ctx, hdr->alt_frame_ts);
-       WARN_ON(!ref && hdr->alt_frame_ts);
-       if (!ref)
-               ref = vb2_dma_contig_plane_dma_addr(&vb2_dst->vb2_buf, 0);
-       if (hdr->flags & V4L2_VP8_FRAME_FLAG_SIGN_BIAS_ALT)
-               ref |= VDPU_REG_VP8_AREF_SIGN_BIAS;
-       vdpu_write_relaxed(vpu, ref, VDPU_REG_VP8_ADDR_REF2_5(3));
-}
-
-static void cfg_buffers(struct hantro_ctx *ctx,
-                       const struct v4l2_ctrl_vp8_frame *hdr)
-{
-       const struct v4l2_vp8_segment *seg = &hdr->segment;
-       struct hantro_dev *vpu = ctx->dev;
-       struct vb2_v4l2_buffer *vb2_dst;
-       dma_addr_t dst_dma;
-       u32 reg;
-
-       vb2_dst = hantro_get_dst_buf(ctx);
-
-       /* Set probability table buffer address */
-       vdpu_write_relaxed(vpu, ctx->vp8_dec.prob_tbl.dma,
-                          VDPU_REG_ADDR_QTABLE);
-
-       /* Set segment map address */
-       reg = VDPU_REG_FWD_PIC1_SEGMENT_BASE(ctx->vp8_dec.segment_map.dma);
-       if (seg->flags & V4L2_VP8_SEGMENT_FLAG_ENABLED) {
-               reg |= VDPU_REG_FWD_PIC1_SEGMENT_E;
-               if (seg->flags & V4L2_VP8_SEGMENT_FLAG_UPDATE_MAP)
-                       reg |= VDPU_REG_FWD_PIC1_SEGMENT_UPD_E;
-       }
-       vdpu_write_relaxed(vpu, reg, VDPU_REG_VP8_SEGMENT_VAL);
-
-       /* set output frame buffer address */
-       dst_dma = vb2_dma_contig_plane_dma_addr(&vb2_dst->vb2_buf, 0);
-       vdpu_write_relaxed(vpu, dst_dma, VDPU_REG_ADDR_DST);
-}
-
-int rk3399_vpu_vp8_dec_run(struct hantro_ctx *ctx)
-{
-       const struct v4l2_ctrl_vp8_frame *hdr;
-       struct hantro_dev *vpu = ctx->dev;
-       size_t height = ctx->dst_fmt.height;
-       size_t width = ctx->dst_fmt.width;
-       u32 mb_width, mb_height;
-       u32 reg;
-
-       hantro_start_prepare_run(ctx);
-
-       hdr = hantro_get_ctrl(ctx, V4L2_CID_STATELESS_VP8_FRAME);
-       if (WARN_ON(!hdr))
-               return -EINVAL;
-
-       /* Reset segment_map buffer in keyframe */
-       if (V4L2_VP8_FRAME_IS_KEY_FRAME(hdr) && ctx->vp8_dec.segment_map.cpu)
-               memset(ctx->vp8_dec.segment_map.cpu, 0,
-                      ctx->vp8_dec.segment_map.size);
-
-       hantro_vp8_prob_update(ctx, hdr);
-
-       /*
-        * Extensive testing shows that the hardware does not properly
-        * clear the internal state from previous a decoding run. This
-        * causes corruption in decoded frames for multi-instance use cases.
-        * A soft reset before programming the registers has been found
-        * to resolve those problems.
-        */
-       ctx->codec_ops->reset(ctx);
-
-       reg = VDPU_REG_CONFIG_DEC_TIMEOUT_E
-               | VDPU_REG_CONFIG_DEC_CLK_GATE_E;
-       if (!V4L2_VP8_FRAME_IS_KEY_FRAME(hdr))
-               reg |= VDPU_REG_DEC_CTRL0_PIC_INTER_E;
-       vdpu_write_relaxed(vpu, reg, VDPU_REG_EN_FLAGS);
-
-       reg = VDPU_REG_CONFIG_DEC_STRENDIAN_E
-               | VDPU_REG_CONFIG_DEC_INSWAP32_E
-               | VDPU_REG_CONFIG_DEC_STRSWAP32_E
-               | VDPU_REG_CONFIG_DEC_OUTSWAP32_E
-               | VDPU_REG_CONFIG_DEC_IN_ENDIAN
-               | VDPU_REG_CONFIG_DEC_OUT_ENDIAN;
-       vdpu_write_relaxed(vpu, reg, VDPU_REG_DATA_ENDIAN);
-
-       reg = VDPU_REG_CONFIG_DEC_MAX_BURST(16);
-       vdpu_write_relaxed(vpu, reg, VDPU_REG_AXI_CTRL);
-
-       reg = VDPU_REG_DEC_CTRL0_DEC_MODE(10);
-       vdpu_write_relaxed(vpu, reg, VDPU_REG_DEC_FORMAT);
-
-       if (!(hdr->flags & V4L2_VP8_FRAME_FLAG_MB_NO_SKIP_COEFF))
-               hantro_reg_write(vpu, &vp8_dec_skip_mode, 1);
-       if (hdr->lf.level == 0)
-               hantro_reg_write(vpu, &vp8_dec_filter_disable, 1);
-
-       /* Frame dimensions */
-       mb_width = MB_WIDTH(width);
-       mb_height = MB_HEIGHT(height);
-
-       hantro_reg_write(vpu, &vp8_dec_mb_width, mb_width);
-       hantro_reg_write(vpu, &vp8_dec_mb_height, mb_height);
-       hantro_reg_write(vpu, &vp8_dec_mb_width_ext, mb_width >> 9);
-       hantro_reg_write(vpu, &vp8_dec_mb_height_ext, mb_height >> 8);
-
-       /* Boolean decoder */
-       hantro_reg_write(vpu, &vp8_dec_bool_range, hdr->coder_state.range);
-       hantro_reg_write(vpu, &vp8_dec_bool_value, hdr->coder_state.value);
-
-       reg = vdpu_read(vpu, VDPU_REG_VP8_DCT_START_BIT);
-       if (hdr->version != 3)
-               reg |= VDPU_REG_DEC_CTRL4_VC1_HEIGHT_EXT;
-       if (hdr->version & 0x3)
-               reg |= VDPU_REG_DEC_CTRL4_BILIN_MC_E;
-       vdpu_write_relaxed(vpu, reg, VDPU_REG_VP8_DCT_START_BIT);
-
-       cfg_lf(ctx, hdr);
-       cfg_qp(ctx, hdr);
-       cfg_parts(ctx, hdr);
-       cfg_tap(ctx, hdr);
-       cfg_ref(ctx, hdr);
-       cfg_buffers(ctx, hdr);
-
-       hantro_end_prepare_run(ctx);
-
-       hantro_reg_write(vpu, &vp8_dec_start_dec, 1);
-
-       return 0;
-}
diff --git a/drivers/staging/media/hantro/rk3399_vpu_regs.h b/drivers/staging/media/hantro/rk3399_vpu_regs.h
deleted file mode 100644 (file)
index 88d0969..0000000
+++ /dev/null
@@ -1,600 +0,0 @@
-/* SPDX-License-Identifier: GPL-2.0 */
-/*
- * Hantro VPU codec driver
- *
- * Copyright (C) 2018 Rockchip Electronics Co., Ltd.
- *     Alpha Lin <alpha.lin@rock-chips.com>
- */
-
-#ifndef RK3399_VPU_REGS_H_
-#define RK3399_VPU_REGS_H_
-
-/* Encoder registers. */
-#define VEPU_REG_VP8_QUT_1ST(i)                        (0x000 + ((i) * 0x24))
-#define     VEPU_REG_VP8_QUT_DC_Y2(x)                  (((x) & 0x3fff) << 16)
-#define     VEPU_REG_VP8_QUT_DC_Y1(x)                  (((x) & 0x3fff) << 0)
-#define VEPU_REG_VP8_QUT_2ND(i)                        (0x004 + ((i) * 0x24))
-#define     VEPU_REG_VP8_QUT_AC_Y1(x)                  (((x) & 0x3fff) << 16)
-#define     VEPU_REG_VP8_QUT_DC_CHR(x)                 (((x) & 0x3fff) << 0)
-#define VEPU_REG_VP8_QUT_3RD(i)                        (0x008 + ((i) * 0x24))
-#define     VEPU_REG_VP8_QUT_AC_CHR(x)                 (((x) & 0x3fff) << 16)
-#define     VEPU_REG_VP8_QUT_AC_Y2(x)                  (((x) & 0x3fff) << 0)
-#define VEPU_REG_VP8_QUT_4TH(i)                        (0x00c + ((i) * 0x24))
-#define     VEPU_REG_VP8_QUT_ZB_DC_CHR(x)              (((x) & 0x1ff) << 18)
-#define     VEPU_REG_VP8_QUT_ZB_DC_Y2(x)               (((x) & 0x1ff) << 9)
-#define     VEPU_REG_VP8_QUT_ZB_DC_Y1(x)               (((x) & 0x1ff) << 0)
-#define VEPU_REG_VP8_QUT_5TH(i)                        (0x010 + ((i) * 0x24))
-#define     VEPU_REG_VP8_QUT_ZB_AC_CHR(x)              (((x) & 0x1ff) << 18)
-#define     VEPU_REG_VP8_QUT_ZB_AC_Y2(x)               (((x) & 0x1ff) << 9)
-#define     VEPU_REG_VP8_QUT_ZB_AC_Y1(x)               (((x) & 0x1ff) << 0)
-#define VEPU_REG_VP8_QUT_6TH(i)                        (0x014 + ((i) * 0x24))
-#define     VEPU_REG_VP8_QUT_RND_DC_CHR(x)             (((x) & 0xff) << 16)
-#define     VEPU_REG_VP8_QUT_RND_DC_Y2(x)              (((x) & 0xff) << 8)
-#define     VEPU_REG_VP8_QUT_RND_DC_Y1(x)              (((x) & 0xff) << 0)
-#define VEPU_REG_VP8_QUT_7TH(i)                        (0x018 + ((i) * 0x24))
-#define     VEPU_REG_VP8_QUT_RND_AC_CHR(x)             (((x) & 0xff) << 16)
-#define     VEPU_REG_VP8_QUT_RND_AC_Y2(x)              (((x) & 0xff) << 8)
-#define     VEPU_REG_VP8_QUT_RND_AC_Y1(x)              (((x) & 0xff) << 0)
-#define VEPU_REG_VP8_QUT_8TH(i)                        (0x01c + ((i) * 0x24))
-#define     VEPU_REG_VP8_SEG_FILTER_LEVEL(x)           (((x) & 0x3f) << 25)
-#define     VEPU_REG_VP8_DEQUT_DC_CHR(x)               (((x) & 0xff) << 17)
-#define     VEPU_REG_VP8_DEQUT_DC_Y2(x)                        (((x) & 0x1ff) << 8)
-#define     VEPU_REG_VP8_DEQUT_DC_Y1(x)                        (((x) & 0xff) << 0)
-#define VEPU_REG_VP8_QUT_9TH(i)                        (0x020 + ((i) * 0x24))
-#define     VEPU_REG_VP8_DEQUT_AC_CHR(x)               (((x) & 0x1ff) << 18)
-#define     VEPU_REG_VP8_DEQUT_AC_Y2(x)                        (((x) & 0x1ff) << 9)
-#define     VEPU_REG_VP8_DEQUT_AC_Y1(x)                        (((x) & 0x1ff) << 0)
-#define VEPU_REG_ADDR_VP8_SEG_MAP              0x06c
-#define VEPU_REG_VP8_INTRA_4X4_PENALTY(i)      (0x070 + ((i) * 0x4))
-#define     VEPU_REG_VP8_INTRA_4X4_PENALTY_0(x)                (((x) & 0xfff) << 0)
-#define     VEPU_REG_VP8_INTRA_4x4_PENALTY_1(x)                (((x) & 0xfff) << 16)
-#define VEPU_REG_VP8_INTRA_16X16_PENALTY(i)    (0x084 + ((i) * 0x4))
-#define     VEPU_REG_VP8_INTRA_16X16_PENALTY_0(x)      (((x) & 0xfff) << 0)
-#define     VEPU_REG_VP8_INTRA_16X16_PENALTY_1(x)      (((x) & 0xfff) << 16)
-#define VEPU_REG_VP8_CONTROL                   0x0a0
-#define     VEPU_REG_VP8_LF_MODE_DELTA_BPRED(x)                (((x) & 0x1f) << 24)
-#define     VEPU_REG_VP8_LF_REF_DELTA_INTRA_MB(x)      (((x) & 0x7f) << 16)
-#define     VEPU_REG_VP8_INTER_TYPE_BIT_COST(x)                (((x) & 0xfff) << 0)
-#define VEPU_REG_VP8_REF_FRAME_VAL             0x0a4
-#define     VEPU_REG_VP8_COEF_DMV_PENALTY(x)           (((x) & 0xfff) << 16)
-#define     VEPU_REG_VP8_REF_FRAME(x)                  (((x) & 0xfff) << 0)
-#define VEPU_REG_VP8_LOOP_FILTER_REF_DELTA     0x0a8
-#define     VEPU_REG_VP8_LF_REF_DELTA_ALT_REF(x)       (((x) & 0x7f) << 16)
-#define     VEPU_REG_VP8_LF_REF_DELTA_LAST_REF(x)      (((x) & 0x7f) << 8)
-#define     VEPU_REG_VP8_LF_REF_DELTA_GOLDEN(x)                (((x) & 0x7f) << 0)
-#define VEPU_REG_VP8_LOOP_FILTER_MODE_DELTA    0x0ac
-#define     VEPU_REG_VP8_LF_MODE_DELTA_SPLITMV(x)      (((x) & 0x7f) << 16)
-#define     VEPU_REG_VP8_LF_MODE_DELTA_ZEROMV(x)       (((x) & 0x7f) << 8)
-#define     VEPU_REG_VP8_LF_MODE_DELTA_NEWMV(x)                (((x) & 0x7f) << 0)
-#define        VEPU_REG_JPEG_LUMA_QUAT(i)              (0x000 + ((i) * 0x4))
-#define        VEPU_REG_JPEG_CHROMA_QUAT(i)            (0x040 + ((i) * 0x4))
-#define VEPU_REG_INTRA_SLICE_BITMAP(i)         (0x0b0 + ((i) * 0x4))
-#define VEPU_REG_ADDR_VP8_DCT_PART(i)          (0x0b0 + ((i) * 0x4))
-#define VEPU_REG_INTRA_AREA_CTRL               0x0b8
-#define     VEPU_REG_INTRA_AREA_TOP(x)                 (((x) & 0xff) << 24)
-#define     VEPU_REG_INTRA_AREA_BOTTOM(x)              (((x) & 0xff) << 16)
-#define     VEPU_REG_INTRA_AREA_LEFT(x)                        (((x) & 0xff) << 8)
-#define     VEPU_REG_INTRA_AREA_RIGHT(x)               (((x) & 0xff) << 0)
-#define VEPU_REG_CIR_INTRA_CTRL                        0x0bc
-#define     VEPU_REG_CIR_INTRA_FIRST_MB(x)             (((x) & 0xffff) << 16)
-#define     VEPU_REG_CIR_INTRA_INTERVAL(x)             (((x) & 0xffff) << 0)
-#define VEPU_REG_ADDR_IN_PLANE_0               0x0c0
-#define VEPU_REG_ADDR_IN_PLANE_1               0x0c4
-#define VEPU_REG_ADDR_IN_PLANE_2               0x0c8
-#define VEPU_REG_STR_HDR_REM_MSB               0x0cc
-#define VEPU_REG_STR_HDR_REM_LSB               0x0d0
-#define VEPU_REG_STR_BUF_LIMIT                 0x0d4
-#define VEPU_REG_AXI_CTRL                      0x0d8
-#define     VEPU_REG_AXI_CTRL_READ_ID(x)               (((x) & 0xff) << 24)
-#define     VEPU_REG_AXI_CTRL_WRITE_ID(x)              (((x) & 0xff) << 16)
-#define     VEPU_REG_AXI_CTRL_BURST_LEN(x)             (((x) & 0x3f) << 8)
-#define     VEPU_REG_AXI_CTRL_INCREMENT_MODE(x)                (((x) & 0x01) << 2)
-#define     VEPU_REG_AXI_CTRL_BIRST_DISCARD(x)         (((x) & 0x01) << 1)
-#define     VEPU_REG_AXI_CTRL_BIRST_DISABLE            BIT(0)
-#define VEPU_QP_ADJUST_MAD_DELTA_ROI           0x0dc
-#define     VEPU_REG_ROI_QP_DELTA_1                    (((x) & 0xf) << 12)
-#define     VEPU_REG_ROI_QP_DELTA_2                    (((x) & 0xf) << 8)
-#define     VEPU_REG_MAD_QP_ADJUSTMENT                 (((x) & 0xf) << 0)
-#define VEPU_REG_ADDR_REF_LUMA                 0x0e0
-#define VEPU_REG_ADDR_REF_CHROMA               0x0e4
-#define VEPU_REG_QP_SUM_DIV2                   0x0e8
-#define     VEPU_REG_QP_SUM(x)                         (((x) & 0x001fffff) * 2)
-#define VEPU_REG_ENC_CTRL0                     0x0ec
-#define     VEPU_REG_DISABLE_QUARTER_PIXEL_MV          BIT(28)
-#define     VEPU_REG_DEBLOCKING_FILTER_MODE(x)         (((x) & 0x3) << 24)
-#define     VEPU_REG_CABAC_INIT_IDC(x)                 (((x) & 0x3) << 21)
-#define     VEPU_REG_ENTROPY_CODING_MODE               BIT(20)
-#define     VEPU_REG_H264_TRANS8X8_MODE                        BIT(17)
-#define     VEPU_REG_H264_INTER4X4_MODE                        BIT(16)
-#define     VEPU_REG_H264_STREAM_MODE                  BIT(15)
-#define     VEPU_REG_H264_SLICE_SIZE(x)                        (((x) & 0x7f) << 8)
-#define VEPU_REG_ENC_OVER_FILL_STRM_OFFSET     0x0f0
-#define     VEPU_REG_STREAM_START_OFFSET(x)            (((x) & 0x3f) << 16)
-#define     VEPU_REG_SKIP_MACROBLOCK_PENALTY(x)                (((x) & 0xff) << 8)
-#define     VEPU_REG_IN_IMG_CTRL_OVRFLR_D4(x)          (((x) & 0x3) << 4)
-#define     VEPU_REG_IN_IMG_CTRL_OVRFLB(x)             (((x) & 0xf) << 0)
-#define VEPU_REG_INPUT_LUMA_INFO               0x0f4
-#define     VEPU_REG_IN_IMG_CHROMA_OFFSET(x)           (((x) & 0x7) << 20)
-#define     VEPU_REG_IN_IMG_LUMA_OFFSET(x)             (((x) & 0x7) << 16)
-#define     VEPU_REG_IN_IMG_CTRL_ROW_LEN(x)            (((x) & 0x3fff) << 0)
-#define VEPU_REG_RLC_SUM                       0x0f8
-#define     VEPU_REG_RLC_SUM_OUT(x)                    (((x) & 0x007fffff) * 4)
-#define VEPU_REG_SPLIT_PENALTY_4X4             0x0f8
-#define            VEPU_REG_VP8_SPLIT_PENALTY_4X4              (((x) & 0x1ff) << 19)
-#define VEPU_REG_ADDR_REC_LUMA                 0x0fc
-#define VEPU_REG_ADDR_REC_CHROMA               0x100
-#define VEPU_REG_CHECKPOINT(i)                 (0x104 + ((i) * 0x4))
-#define     VEPU_REG_CHECKPOINT_CHECK0(x)              (((x) & 0xffff))
-#define     VEPU_REG_CHECKPOINT_CHECK1(x)              (((x) & 0xffff) << 16)
-#define     VEPU_REG_CHECKPOINT_RESULT(x) \
-               ((((x) >> (16 - 16 * ((i) & 1))) & 0xffff) * 32)
-#define VEPU_REG_VP8_SEG0_QUANT_AC_Y1          0x104
-#define     VEPU_REG_VP8_SEG0_RND_AC_Y1(x)             (((x) & 0xff) << 23)
-#define     VEPU_REG_VP8_SEG0_ZBIN_AC_Y1(x)            (((x) & 0x1ff) << 14)
-#define     VEPU_REG_VP8_SEG0_QUT_AC_Y1(x)             (((x) & 0x3fff) << 0)
-#define VEPU_REG_VP8_SEG0_QUANT_DC_Y2          0x108
-#define     VEPU_REG_VP8_SEG0_RND_DC_Y2(x)             (((x) & 0xff) << 23)
-#define     VEPU_REG_VP8_SEG0_ZBIN_DC_Y2(x)            (((x) & 0x1ff) << 14)
-#define     VEPU_REG_VP8_SEG0_QUT_DC_Y2(x)             (((x) & 0x3fff) << 0)
-#define VEPU_REG_VP8_SEG0_QUANT_AC_Y2          0x10c
-#define     VEPU_REG_VP8_SEG0_RND_AC_Y2(x)             (((x) & 0xff) << 23)
-#define     VEPU_REG_VP8_SEG0_ZBIN_AC_Y2(x)            (((x) & 0x1ff) << 14)
-#define     VEPU_REG_VP8_SEG0_QUT_AC_Y2(x)             (((x) & 0x3fff) << 0)
-#define VEPU_REG_VP8_SEG0_QUANT_DC_CHR         0x110
-#define     VEPU_REG_VP8_SEG0_RND_DC_CHR(x)            (((x) & 0xff) << 23)
-#define     VEPU_REG_VP8_SEG0_ZBIN_DC_CHR(x)           (((x) & 0x1ff) << 14)
-#define     VEPU_REG_VP8_SEG0_QUT_DC_CHR(x)            (((x) & 0x3fff) << 0)
-#define VEPU_REG_VP8_SEG0_QUANT_AC_CHR         0x114
-#define     VEPU_REG_VP8_SEG0_RND_AC_CHR(x)            (((x) & 0xff) << 23)
-#define     VEPU_REG_VP8_SEG0_ZBIN_AC_CHR(x)           (((x) & 0x1ff) << 14)
-#define     VEPU_REG_VP8_SEG0_QUT_AC_CHR(x)            (((x) & 0x3fff) << 0)
-#define VEPU_REG_VP8_SEG0_QUANT_DQUT           0x118
-#define     VEPU_REG_VP8_MV_REF_IDX1(x)                        (((x) & 0x03) << 26)
-#define     VEPU_REG_VP8_SEG0_DQUT_DC_Y2(x)            (((x) & 0x1ff) << 17)
-#define     VEPU_REG_VP8_SEG0_DQUT_AC_Y1(x)            (((x) & 0x1ff) << 8)
-#define     VEPU_REG_VP8_SEG0_DQUT_DC_Y1(x)            (((x) & 0xff) << 0)
-#define VEPU_REG_CHKPT_WORD_ERR(i)             (0x118 + ((i) * 0x4))
-#define     VEPU_REG_CHKPT_WORD_ERR_CHK0(x)            (((x) & 0xffff))
-#define     VEPU_REG_CHKPT_WORD_ERR_CHK1(x)            (((x) & 0xffff) << 16)
-#define VEPU_REG_VP8_SEG0_QUANT_DQUT_1         0x11c
-#define     VEPU_REG_VP8_SEGMENT_MAP_UPDATE            BIT(30)
-#define     VEPU_REG_VP8_SEGMENT_EN                    BIT(29)
-#define     VEPU_REG_VP8_MV_REF_IDX2_EN                        BIT(28)
-#define     VEPU_REG_VP8_MV_REF_IDX2(x)                        (((x) & 0x03) << 26)
-#define     VEPU_REG_VP8_SEG0_DQUT_AC_CHR(x)           (((x) & 0x1ff) << 17)
-#define     VEPU_REG_VP8_SEG0_DQUT_DC_CHR(x)           (((x) & 0xff) << 9)
-#define     VEPU_REG_VP8_SEG0_DQUT_AC_Y2(x)            (((x) & 0x1ff) << 0)
-#define VEPU_REG_VP8_BOOL_ENC_VALUE            0x120
-#define VEPU_REG_CHKPT_DELTA_QP                        0x124
-#define     VEPU_REG_CHKPT_DELTA_QP_CHK0(x)            (((x) & 0x0f) << 0)
-#define     VEPU_REG_CHKPT_DELTA_QP_CHK1(x)            (((x) & 0x0f) << 4)
-#define     VEPU_REG_CHKPT_DELTA_QP_CHK2(x)            (((x) & 0x0f) << 8)
-#define     VEPU_REG_CHKPT_DELTA_QP_CHK3(x)            (((x) & 0x0f) << 12)
-#define     VEPU_REG_CHKPT_DELTA_QP_CHK4(x)            (((x) & 0x0f) << 16)
-#define     VEPU_REG_CHKPT_DELTA_QP_CHK5(x)            (((x) & 0x0f) << 20)
-#define     VEPU_REG_CHKPT_DELTA_QP_CHK6(x)            (((x) & 0x0f) << 24)
-#define VEPU_REG_VP8_ENC_CTRL2                 0x124
-#define     VEPU_REG_VP8_ZERO_MV_PENALTY_FOR_REF2(x)   (((x) & 0xff) << 24)
-#define     VEPU_REG_VP8_FILTER_SHARPNESS(x)           (((x) & 0x07) << 21)
-#define     VEPU_REG_VP8_FILTER_LEVEL(x)               (((x) & 0x3f) << 15)
-#define     VEPU_REG_VP8_DCT_PARTITION_CNT(x)          (((x) & 0x03) << 13)
-#define     VEPU_REG_VP8_BOOL_ENC_VALUE_BITS(x)                (((x) & 0x1f) << 8)
-#define     VEPU_REG_VP8_BOOL_ENC_RANGE(x)             (((x) & 0xff) << 0)
-#define VEPU_REG_ENC_CTRL1                     0x128
-#define     VEPU_REG_MAD_THRESHOLD(x)                  (((x) & 0x3f) << 24)
-#define     VEPU_REG_COMPLETED_SLICES(x)               (((x) & 0xff) << 16)
-#define     VEPU_REG_IN_IMG_CTRL_FMT(x)                        (((x) & 0xf) << 4)
-#define     VEPU_REG_IN_IMG_ROTATE_MODE(x)             (((x) & 0x3) << 2)
-#define     VEPU_REG_SIZE_TABLE_PRESENT                        BIT(0)
-#define VEPU_REG_INTRA_INTER_MODE              0x12c
-#define     VEPU_REG_INTRA16X16_MODE(x)                        (((x) & 0xffff) << 16)
-#define     VEPU_REG_INTER_MODE(x)                     (((x) & 0xffff) << 0)
-#define VEPU_REG_ENC_CTRL2                     0x130
-#define     VEPU_REG_PPS_INIT_QP(x)                    (((x) & 0x3f) << 26)
-#define     VEPU_REG_SLICE_FILTER_ALPHA(x)             (((x) & 0xf) << 22)
-#define     VEPU_REG_SLICE_FILTER_BETA(x)              (((x) & 0xf) << 18)
-#define     VEPU_REG_CHROMA_QP_OFFSET(x)               (((x) & 0x1f) << 13)
-#define     VEPU_REG_FILTER_DISABLE                    BIT(5)
-#define     VEPU_REG_IDR_PIC_ID(x)                     (((x) & 0xf) << 1)
-#define     VEPU_REG_CONSTRAINED_INTRA_PREDICTION      BIT(0)
-#define VEPU_REG_ADDR_OUTPUT_STREAM            0x134
-#define VEPU_REG_ADDR_OUTPUT_CTRL              0x138
-#define VEPU_REG_ADDR_NEXT_PIC                 0x13c
-#define VEPU_REG_ADDR_MV_OUT                   0x140
-#define VEPU_REG_ADDR_CABAC_TBL                        0x144
-#define VEPU_REG_ROI1                          0x148
-#define     VEPU_REG_ROI1_TOP_MB(x)                    (((x) & 0xff) << 24)
-#define     VEPU_REG_ROI1_BOTTOM_MB(x)                 (((x) & 0xff) << 16)
-#define     VEPU_REG_ROI1_LEFT_MB(x)                   (((x) & 0xff) << 8)
-#define     VEPU_REG_ROI1_RIGHT_MB(x)                  (((x) & 0xff) << 0)
-#define VEPU_REG_ROI2                          0x14c
-#define     VEPU_REG_ROI2_TOP_MB(x)                    (((x) & 0xff) << 24)
-#define     VEPU_REG_ROI2_BOTTOM_MB(x)                 (((x) & 0xff) << 16)
-#define     VEPU_REG_ROI2_LEFT_MB(x)                   (((x) & 0xff) << 8)
-#define     VEPU_REG_ROI2_RIGHT_MB(x)                  (((x) & 0xff) << 0)
-#define VEPU_REG_STABLE_MATRIX(i)              (0x150 + ((i) * 0x4))
-#define VEPU_REG_STABLE_MOTION_SUM             0x174
-#define VEPU_REG_STABILIZATION_OUTPUT          0x178
-#define     VEPU_REG_STABLE_MIN_VALUE(x)               (((x) & 0xffffff) << 8)
-#define     VEPU_REG_STABLE_MODE_SEL(x)                        (((x) & 0x3) << 6)
-#define     VEPU_REG_STABLE_HOR_GMV(x)                 (((x) & 0x3f) << 0)
-#define VEPU_REG_RGB2YUV_CONVERSION_COEF1      0x17c
-#define     VEPU_REG_RGB2YUV_CONVERSION_COEFB(x)       (((x) & 0xffff) << 16)
-#define     VEPU_REG_RGB2YUV_CONVERSION_COEFA(x)       (((x) & 0xffff) << 0)
-#define VEPU_REG_RGB2YUV_CONVERSION_COEF2      0x180
-#define     VEPU_REG_RGB2YUV_CONVERSION_COEFE(x)       (((x) & 0xffff) << 16)
-#define     VEPU_REG_RGB2YUV_CONVERSION_COEFC(x)       (((x) & 0xffff) << 0)
-#define VEPU_REG_RGB2YUV_CONVERSION_COEF3      0x184
-#define     VEPU_REG_RGB2YUV_CONVERSION_COEFF(x)       (((x) & 0xffff) << 0)
-#define VEPU_REG_RGB_MASK_MSB                  0x188
-#define     VEPU_REG_RGB_MASK_B_MSB(x)                 (((x) & 0x1f) << 16)
-#define     VEPU_REG_RGB_MASK_G_MSB(x)                 (((x) & 0x1f) << 8)
-#define     VEPU_REG_RGB_MASK_R_MSB(x)                 (((x) & 0x1f) << 0)
-#define VEPU_REG_MV_PENALTY                    0x18c
-#define     VEPU_REG_1MV_PENALTY(x)                    (((x) & 0x3ff) << 21)
-#define     VEPU_REG_QMV_PENALTY(x)                    (((x) & 0x3ff) << 11)
-#define     VEPU_REG_4MV_PENALTY(x)                    (((x) & 0x3ff) << 1)
-#define     VEPU_REG_SPLIT_MV_MODE_EN                  BIT(0)
-#define VEPU_REG_QP_VAL                                0x190
-#define     VEPU_REG_H264_LUMA_INIT_QP(x)              (((x) & 0x3f) << 26)
-#define     VEPU_REG_H264_QP_MAX(x)                    (((x) & 0x3f) << 20)
-#define     VEPU_REG_H264_QP_MIN(x)                    (((x) & 0x3f) << 14)
-#define     VEPU_REG_H264_CHKPT_DISTANCE(x)            (((x) & 0xfff) << 0)
-#define VEPU_REG_VP8_SEG0_QUANT_DC_Y1          0x190
-#define     VEPU_REG_VP8_SEG0_RND_DC_Y1(x)             (((x) & 0xff) << 23)
-#define     VEPU_REG_VP8_SEG0_ZBIN_DC_Y1(x)            (((x) & 0x1ff) << 14)
-#define     VEPU_REG_VP8_SEG0_QUT_DC_Y1(x)             (((x) & 0x3fff) << 0)
-#define VEPU_REG_MVC_RELATE                    0x198
-#define     VEPU_REG_ZERO_MV_FAVOR_D2(x)               (((x) & 0xf) << 20)
-#define     VEPU_REG_PENALTY_4X4MV(x)                  (((x) & 0x1ff) << 11)
-#define     VEPU_REG_MVC_VIEW_ID(x)                    (((x) & 0x7) << 8)
-#define     VEPU_REG_MVC_ANCHOR_PIC_FLAG               BIT(7)
-#define     VEPU_REG_MVC_PRIORITY_ID(x)                        (((x) & 0x7) << 4)
-#define     VEPU_REG_MVC_TEMPORAL_ID(x)                        (((x) & 0x7) << 1)
-#define     VEPU_REG_MVC_INTER_VIEW_FLAG               BIT(0)
-#define VEPU_REG_ENCODE_START                  0x19c
-#define     VEPU_REG_MB_HEIGHT(x)                      (((x) & 0x1ff) << 20)
-#define     VEPU_REG_MB_WIDTH(x)                       (((x) & 0x1ff) << 8)
-#define     VEPU_REG_FRAME_TYPE_INTER                  (0x0 << 6)
-#define     VEPU_REG_FRAME_TYPE_INTRA                  (0x1 << 6)
-#define     VEPU_REG_FRAME_TYPE_MVCINTER               (0x2 << 6)
-#define     VEPU_REG_ENCODE_FORMAT_JPEG                        (0x2 << 4)
-#define     VEPU_REG_ENCODE_FORMAT_H264                        (0x3 << 4)
-#define     VEPU_REG_ENCODE_ENABLE                     BIT(0)
-#define VEPU_REG_MB_CTRL                       0x1a0
-#define     VEPU_REG_MB_CNT_OUT(x)                     (((x) & 0xffff) << 16)
-#define     VEPU_REG_MB_CNT_SET(x)                     (((x) & 0xffff) << 0)
-#define VEPU_REG_DATA_ENDIAN                   0x1a4
-#define     VEPU_REG_INPUT_SWAP8                       BIT(31)
-#define     VEPU_REG_INPUT_SWAP16                      BIT(30)
-#define     VEPU_REG_INPUT_SWAP32                      BIT(29)
-#define     VEPU_REG_OUTPUT_SWAP8                      BIT(28)
-#define     VEPU_REG_OUTPUT_SWAP16                     BIT(27)
-#define     VEPU_REG_OUTPUT_SWAP32                     BIT(26)
-#define     VEPU_REG_TEST_IRQ                          BIT(24)
-#define     VEPU_REG_TEST_COUNTER(x)                   (((x) & 0xf) << 20)
-#define     VEPU_REG_TEST_REG                          BIT(19)
-#define     VEPU_REG_TEST_MEMORY                       BIT(18)
-#define     VEPU_REG_TEST_LEN(x)                       (((x) & 0x3ffff) << 0)
-#define VEPU_REG_ENC_CTRL3                     0x1a8
-#define     VEPU_REG_PPS_ID(x)                         (((x) & 0xff) << 24)
-#define     VEPU_REG_INTRA_PRED_MODE(x)                        (((x) & 0xff) << 16)
-#define     VEPU_REG_FRAME_NUM(x)                      (((x) & 0xffff) << 0)
-#define VEPU_REG_ENC_CTRL4                     0x1ac
-#define     VEPU_REG_MV_PENALTY_16X8_8X16(x)           (((x) & 0x3ff) << 20)
-#define     VEPU_REG_MV_PENALTY_8X8(x)                 (((x) & 0x3ff) << 10)
-#define     VEPU_REG_MV_PENALTY_8X4_4X8(x)             (((x) & 0x3ff) << 0)
-#define VEPU_REG_ADDR_VP8_PROB_CNT             0x1b0
-#define VEPU_REG_INTERRUPT                     0x1b4
-#define     VEPU_REG_INTERRUPT_NON                     BIT(28)
-#define     VEPU_REG_MV_WRITE_EN                       BIT(24)
-#define     VEPU_REG_RECON_WRITE_DIS                   BIT(20)
-#define     VEPU_REG_INTERRUPT_SLICE_READY_EN          BIT(16)
-#define     VEPU_REG_CLK_GATING_EN                     BIT(12)
-#define     VEPU_REG_INTERRUPT_TIMEOUT_EN              BIT(10)
-#define     VEPU_REG_INTERRUPT_RESET                   BIT(9)
-#define     VEPU_REG_INTERRUPT_DIS_BIT                 BIT(8)
-#define     VEPU_REG_INTERRUPT_TIMEOUT                 BIT(6)
-#define     VEPU_REG_INTERRUPT_BUFFER_FULL             BIT(5)
-#define     VEPU_REG_INTERRUPT_BUS_ERROR               BIT(4)
-#define     VEPU_REG_INTERRUPT_FUSE                    BIT(3)
-#define     VEPU_REG_INTERRUPT_SLICE_READY             BIT(2)
-#define     VEPU_REG_INTERRUPT_FRAME_READY             BIT(1)
-#define     VEPU_REG_INTERRUPT_BIT                     BIT(0)
-#define VEPU_REG_DMV_PENALTY_TBL(i)            (0x1E0 + ((i) * 0x4))
-#define     VEPU_REG_DMV_PENALTY_TABLE_BIT(x, i)        ((x) << (i) * 8)
-#define VEPU_REG_DMV_Q_PIXEL_PENALTY_TBL(i)    (0x260 + ((i) * 0x4))
-#define     VEPU_REG_DMV_Q_PIXEL_PENALTY_TABLE_BIT(x, i)       ((x) << (i) * 8)
-
-/* vpu decoder register */
-#define VDPU_REG_DEC_CTRL0                     0x0c8 // 50
-#define     VDPU_REG_REF_BUF_CTRL2_REFBU2_PICID(x)     (((x) & 0x1f) << 25)
-#define     VDPU_REG_REF_BUF_CTRL2_REFBU2_THR(x)       (((x) & 0xfff) << 13)
-#define     VDPU_REG_CONFIG_TILED_MODE_LSB             BIT(12)
-#define     VDPU_REG_CONFIG_DEC_ADV_PRE_DIS            BIT(11)
-#define     VDPU_REG_CONFIG_DEC_SCMD_DIS               BIT(10)
-#define     VDPU_REG_DEC_CTRL0_SKIP_MODE               BIT(9)
-#define     VDPU_REG_DEC_CTRL0_FILTERING_DIS           BIT(8)
-#define     VDPU_REG_DEC_CTRL0_PIC_FIXED_QUANT         BIT(7)
-#define     VDPU_REG_CONFIG_DEC_LATENCY(x)             (((x) & 0x3f) << 1)
-#define     VDPU_REG_CONFIG_TILED_MODE_MSB(x)          BIT(0)
-#define     VDPU_REG_CONFIG_DEC_OUT_TILED_E            BIT(0)
-#define VDPU_REG_STREAM_LEN                    0x0cc
-#define     VDPU_REG_DEC_CTRL3_INIT_QP(x)              (((x) & 0x3f) << 25)
-#define     VDPU_REG_DEC_STREAM_LEN_HI                 BIT(24)
-#define     VDPU_REG_DEC_CTRL3_STREAM_LEN(x)           (((x) & 0xffffff) << 0)
-#define VDPU_REG_ERROR_CONCEALMENT             0x0d0
-#define     VDPU_REG_REF_BUF_CTRL2_APF_THRESHOLD(x)    (((x) & 0x3fff) << 17)
-#define     VDPU_REG_ERR_CONC_STARTMB_X(x)             (((x) & 0x1ff) << 8)
-#define     VDPU_REG_ERR_CONC_STARTMB_Y(x)             (((x) & 0xff) << 0)
-#define VDPU_REG_DEC_FORMAT                    0x0d4
-#define     VDPU_REG_DEC_CTRL0_DEC_MODE(x)             (((x) & 0xf) << 0)
-#define VDPU_REG_DATA_ENDIAN                   0x0d8
-#define     VDPU_REG_CONFIG_DEC_STRENDIAN_E            BIT(5)
-#define     VDPU_REG_CONFIG_DEC_STRSWAP32_E            BIT(4)
-#define     VDPU_REG_CONFIG_DEC_OUTSWAP32_E            BIT(3)
-#define     VDPU_REG_CONFIG_DEC_INSWAP32_E             BIT(2)
-#define     VDPU_REG_CONFIG_DEC_OUT_ENDIAN             BIT(1)
-#define     VDPU_REG_CONFIG_DEC_IN_ENDIAN              BIT(0)
-#define VDPU_REG_INTERRUPT                     0x0dc
-#define     VDPU_REG_INTERRUPT_DEC_TIMEOUT             BIT(13)
-#define     VDPU_REG_INTERRUPT_DEC_ERROR_INT           BIT(12)
-#define     VDPU_REG_INTERRUPT_DEC_PIC_INF             BIT(10)
-#define     VDPU_REG_INTERRUPT_DEC_SLICE_INT           BIT(9)
-#define     VDPU_REG_INTERRUPT_DEC_ASO_INT             BIT(8)
-#define     VDPU_REG_INTERRUPT_DEC_BUFFER_INT          BIT(6)
-#define     VDPU_REG_INTERRUPT_DEC_BUS_INT             BIT(5)
-#define     VDPU_REG_INTERRUPT_DEC_RDY_INT             BIT(4)
-#define     VDPU_REG_INTERRUPT_DEC_IRQ_DIS             BIT(1)
-#define     VDPU_REG_INTERRUPT_DEC_IRQ                 BIT(0)
-#define VDPU_REG_AXI_CTRL                      0x0e0
-#define     VDPU_REG_AXI_DEC_SEL                       BIT(23)
-#define     VDPU_REG_CONFIG_DEC_DATA_DISC_E            BIT(22)
-#define     VDPU_REG_PARAL_BUS_E(x)                    BIT(21)
-#define     VDPU_REG_CONFIG_DEC_MAX_BURST(x)           (((x) & 0x1f) << 16)
-#define     VDPU_REG_DEC_CTRL0_DEC_AXI_WR_ID(x)                (((x) & 0xff) << 8)
-#define     VDPU_REG_CONFIG_DEC_AXI_RD_ID(x)           (((x) & 0xff) << 0)
-#define VDPU_REG_EN_FLAGS                      0x0e4
-#define     VDPU_REG_AHB_HLOCK_E                       BIT(31)
-#define     VDPU_REG_CACHE_E                           BIT(29)
-#define     VDPU_REG_PREFETCH_SINGLE_CHANNEL_E         BIT(28)
-#define     VDPU_REG_INTRA_3_CYCLE_ENHANCE             BIT(27)
-#define     VDPU_REG_INTRA_DOUBLE_SPEED                        BIT(26)
-#define     VDPU_REG_INTER_DOUBLE_SPEED                        BIT(25)
-#define     VDPU_REG_DEC_CTRL3_START_CODE_E            BIT(22)
-#define     VDPU_REG_DEC_CTRL3_CH_8PIX_ILEAV_E         BIT(21)
-#define     VDPU_REG_DEC_CTRL0_RLC_MODE_E              BIT(20)
-#define     VDPU_REG_DEC_CTRL0_DIVX3_E                 BIT(19)
-#define     VDPU_REG_DEC_CTRL0_PJPEG_E                 BIT(18)
-#define     VDPU_REG_DEC_CTRL0_PIC_INTERLACE_E         BIT(17)
-#define     VDPU_REG_DEC_CTRL0_PIC_FIELDMODE_E         BIT(16)
-#define     VDPU_REG_DEC_CTRL0_PIC_B_E                 BIT(15)
-#define     VDPU_REG_DEC_CTRL0_PIC_INTER_E             BIT(14)
-#define     VDPU_REG_DEC_CTRL0_PIC_TOPFIELD_E          BIT(13)
-#define     VDPU_REG_DEC_CTRL0_FWD_INTERLACE_E         BIT(12)
-#define     VDPU_REG_DEC_CTRL0_SORENSON_E              BIT(11)
-#define     VDPU_REG_DEC_CTRL0_WRITE_MVS_E             BIT(10)
-#define     VDPU_REG_DEC_CTRL0_REF_TOPFIELD_E          BIT(9)
-#define     VDPU_REG_DEC_CTRL0_REFTOPFIRST_E           BIT(8)
-#define     VDPU_REG_DEC_CTRL0_SEQ_MBAFF_E             BIT(7)
-#define     VDPU_REG_DEC_CTRL0_PICORD_COUNT_E          BIT(6)
-#define     VDPU_REG_CONFIG_DEC_TIMEOUT_E              BIT(5)
-#define     VDPU_REG_CONFIG_DEC_CLK_GATE_E             BIT(4)
-#define     VDPU_REG_DEC_CTRL0_DEC_OUT_DIS             BIT(2)
-#define     VDPU_REG_REF_BUF_CTRL2_REFBU2_BUF_E                BIT(1)
-#define     VDPU_REG_INTERRUPT_DEC_E                   BIT(0)
-#define VDPU_REG_SOFT_RESET                    0x0e8
-#define VDPU_REG_PRED_FLT                      0x0ec
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_0_0(x)       (((x) & 0x3ff) << 22)
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_0_1(x)       (((x) & 0x3ff) << 12)
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_0_2(x)       (((x) & 0x3ff) << 2)
-#define VDPU_REG_ADDITIONAL_CHROMA_ADDRESS     0x0f0
-#define VDPU_REG_ADDR_QTABLE                   0x0f4
-#define VDPU_REG_DIRECT_MV_ADDR                        0x0f8
-#define VDPU_REG_ADDR_DST                      0x0fc
-#define VDPU_REG_ADDR_STR                      0x100
-#define VDPU_REG_REFBUF_RELATED                        0x104
-#define VDPU_REG_FWD_PIC(i)                    (0x128 + ((i) * 0x4))
-#define     VDPU_REG_FWD_PIC_PINIT_RLIST_F5(x)         (((x) & 0x1f) << 25)
-#define     VDPU_REG_FWD_PIC_PINIT_RLIST_F4(x)         (((x) & 0x1f) << 20)
-#define     VDPU_REG_FWD_PIC_PINIT_RLIST_F3(x)         (((x) & 0x1f) << 15)
-#define     VDPU_REG_FWD_PIC_PINIT_RLIST_F2(x)         (((x) & 0x1f) << 10)
-#define     VDPU_REG_FWD_PIC_PINIT_RLIST_F1(x)         (((x) & 0x1f) << 5)
-#define     VDPU_REG_FWD_PIC_PINIT_RLIST_F0(x)         (((x) & 0x1f) << 0)
-#define VDPU_REG_REF_PIC(i)                    (0x130 + ((i) * 0x4))
-#define     VDPU_REG_REF_PIC_REFER1_NBR(x)             (((x) & 0xffff) << 16)
-#define     VDPU_REG_REF_PIC_REFER0_NBR(x)             (((x) & 0xffff) << 0)
-#define VDPU_REG_H264_ADDR_REF(i)                      (0x150 + ((i) * 0x4))
-#define     VDPU_REG_ADDR_REF_FIELD_E                  BIT(1)
-#define     VDPU_REG_ADDR_REF_TOPC_E                   BIT(0)
-#define VDPU_REG_INITIAL_REF_PIC_LIST0         0x190
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F5(x)      (((x) & 0x1f) << 25)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F4(x)      (((x) & 0x1f) << 20)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F3(x)      (((x) & 0x1f) << 15)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F2(x)      (((x) & 0x1f) << 10)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F1(x)      (((x) & 0x1f) << 5)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F0(x)      (((x) & 0x1f) << 0)
-#define VDPU_REG_INITIAL_REF_PIC_LIST1         0x194
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F11(x)     (((x) & 0x1f) << 25)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F10(x)     (((x) & 0x1f) << 20)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F9(x)      (((x) & 0x1f) << 15)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F8(x)      (((x) & 0x1f) << 10)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F7(x)      (((x) & 0x1f) << 5)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F6(x)      (((x) & 0x1f) << 0)
-#define VDPU_REG_INITIAL_REF_PIC_LIST2         0x198
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F15(x)     (((x) & 0x1f) << 15)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F14(x)     (((x) & 0x1f) << 10)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F13(x)     (((x) & 0x1f) << 5)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F12(x)     (((x) & 0x1f) << 0)
-#define VDPU_REG_INITIAL_REF_PIC_LIST3         0x19c
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B5(x)      (((x) & 0x1f) << 25)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B4(x)      (((x) & 0x1f) << 20)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B3(x)      (((x) & 0x1f) << 15)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B2(x)      (((x) & 0x1f) << 10)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B1(x)      (((x) & 0x1f) << 5)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B0(x)      (((x) & 0x1f) << 0)
-#define VDPU_REG_INITIAL_REF_PIC_LIST4         0x1a0
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B11(x)     (((x) & 0x1f) << 25)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B10(x)     (((x) & 0x1f) << 20)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B9(x)      (((x) & 0x1f) << 15)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B8(x)      (((x) & 0x1f) << 10)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B7(x)      (((x) & 0x1f) << 5)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B6(x)      (((x) & 0x1f) << 0)
-#define VDPU_REG_INITIAL_REF_PIC_LIST5         0x1a4
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B15(x)     (((x) & 0x1f) << 15)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B14(x)     (((x) & 0x1f) << 10)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B13(x)     (((x) & 0x1f) << 5)
-#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B12(x)     (((x) & 0x1f) << 0)
-#define VDPU_REG_INITIAL_REF_PIC_LIST6         0x1a8
-#define     VDPU_REG_BD_P_REF_PIC_PINIT_RLIST_F3(x)    (((x) & 0x1f) << 15)
-#define     VDPU_REG_BD_P_REF_PIC_PINIT_RLIST_F2(x)    (((x) & 0x1f) << 10)
-#define     VDPU_REG_BD_P_REF_PIC_PINIT_RLIST_F1(x)    (((x) & 0x1f) << 5)
-#define     VDPU_REG_BD_P_REF_PIC_PINIT_RLIST_F0(x)    (((x) & 0x1f) << 0)
-#define VDPU_REG_LT_REF                                0x1ac
-#define VDPU_REG_VALID_REF                     0x1b0
-#define VDPU_REG_H264_PIC_MB_SIZE              0x1b8
-#define     VDPU_REG_DEC_CTRL2_CH_QP_OFFSET2(x)                (((x) & 0x1f) << 22)
-#define     VDPU_REG_DEC_CTRL2_CH_QP_OFFSET(x)         (((x) & 0x1f) << 17)
-#define     VDPU_REG_DEC_CTRL1_PIC_MB_HEIGHT_P(x)      (((x) & 0xff) << 9)
-#define     VDPU_REG_DEC_CTRL1_PIC_MB_WIDTH(x)         (((x) & 0x1ff) << 0)
-#define VDPU_REG_H264_CTRL                     0x1bc
-#define     VDPU_REG_DEC_CTRL4_WEIGHT_BIPR_IDC(x)      (((x) & 0x3) << 16)
-#define     VDPU_REG_DEC_CTRL1_REF_FRAMES(x)           (((x) & 0x1f) << 0)
-#define VDPU_REG_CURRENT_FRAME                 0x1c0
-#define     VDPU_REG_DEC_CTRL5_FILT_CTRL_PRES          BIT(31)
-#define     VDPU_REG_DEC_CTRL5_RDPIC_CNT_PRES          BIT(30)
-#define     VDPU_REG_DEC_CTRL4_FRAMENUM_LEN(x)         (((x) & 0x1f) << 16)
-#define     VDPU_REG_DEC_CTRL4_FRAMENUM(x)             (((x) & 0xffff) << 0)
-#define VDPU_REG_REF_FRAME                     0x1c4
-#define     VDPU_REG_DEC_CTRL5_REFPIC_MK_LEN(x)                (((x) & 0x7ff) << 16)
-#define     VDPU_REG_DEC_CTRL5_IDR_PIC_ID(x)           (((x) & 0xffff) << 0)
-#define VDPU_REG_DEC_CTRL6                     0x1c8
-#define     VDPU_REG_DEC_CTRL6_PPS_ID(x)               (((x) & 0xff) << 24)
-#define     VDPU_REG_DEC_CTRL6_REFIDX1_ACTIVE(x)       (((x) & 0x1f) << 19)
-#define     VDPU_REG_DEC_CTRL6_REFIDX0_ACTIVE(x)       (((x) & 0x1f) << 14)
-#define     VDPU_REG_DEC_CTRL6_POC_LENGTH(x)           (((x) & 0xff) << 0)
-#define VDPU_REG_ENABLE_FLAG                   0x1cc
-#define     VDPU_REG_DEC_CTRL5_IDR_PIC_E               BIT(8)
-#define     VDPU_REG_DEC_CTRL4_DIR_8X8_INFER_E         BIT(7)
-#define     VDPU_REG_DEC_CTRL4_BLACKWHITE_E            BIT(6)
-#define     VDPU_REG_DEC_CTRL4_CABAC_E                 BIT(5)
-#define     VDPU_REG_DEC_CTRL4_WEIGHT_PRED_E           BIT(4)
-#define     VDPU_REG_DEC_CTRL5_CONST_INTRA_E           BIT(3)
-#define     VDPU_REG_DEC_CTRL5_8X8TRANS_FLAG_E         BIT(2)
-#define     VDPU_REG_DEC_CTRL2_TYPE1_QUANT_E           BIT(1)
-#define     VDPU_REG_DEC_CTRL2_FIELDPIC_FLAG_E         BIT(0)
-#define VDPU_REG_VP8_PIC_MB_SIZE               0x1e0
-#define     VDPU_REG_DEC_PIC_MB_WIDTH(x)               (((x) & 0x1ff) << 23)
-#define            VDPU_REG_DEC_MB_WIDTH_OFF(x)                (((x) & 0xf) << 19)
-#define            VDPU_REG_DEC_PIC_MB_HEIGHT_P(x)             (((x) & 0xff) << 11)
-#define     VDPU_REG_DEC_MB_HEIGHT_OFF(x)              (((x) & 0xf) << 7)
-#define     VDPU_REG_DEC_CTRL1_PIC_MB_W_EXT(x)         (((x) & 0x7) << 3)
-#define     VDPU_REG_DEC_CTRL1_PIC_MB_H_EXT(x)         (((x) & 0x7) << 0)
-#define VDPU_REG_VP8_DCT_START_BIT             0x1e4
-#define     VDPU_REG_DEC_CTRL4_DCT1_START_BIT(x)       (((x) & 0x3f) << 26)
-#define     VDPU_REG_DEC_CTRL4_DCT2_START_BIT(x)       (((x) & 0x3f) << 20)
-#define     VDPU_REG_DEC_CTRL4_VC1_HEIGHT_EXT          BIT(13)
-#define     VDPU_REG_DEC_CTRL4_BILIN_MC_E              BIT(12)
-#define VDPU_REG_VP8_CTRL0                     0x1e8
-#define     VDPU_REG_DEC_CTRL2_STRM_START_BIT(x)       (((x) & 0x3f) << 26)
-#define     VDPU_REG_DEC_CTRL2_STRM1_START_BIT(x)      (((x) & 0x3f) << 18)
-#define     VDPU_REG_DEC_CTRL2_BOOLEAN_VALUE(x)                (((x) & 0xff) << 8)
-#define     VDPU_REG_DEC_CTRL2_BOOLEAN_RANGE(x)                (((x) & 0xff) << 0)
-#define VDPU_REG_VP8_DATA_VAL                  0x1f0
-#define     VDPU_REG_DEC_CTRL6_COEFFS_PART_AM(x)       (((x) & 0xf) << 24)
-#define     VDPU_REG_DEC_CTRL6_STREAM1_LEN(x)          (((x) & 0xffffff) << 0)
-#define VDPU_REG_PRED_FLT7                     0x1f4
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_5_1(x)       (((x) & 0x3ff) << 22)
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_5_2(x)       (((x) & 0x3ff) << 12)
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_5_3(x)       (((x) & 0x3ff) << 2)
-#define VDPU_REG_PRED_FLT8                     0x1f8
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_6_0(x)       (((x) & 0x3ff) << 22)
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_6_1(x)       (((x) & 0x3ff) << 12)
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_6_2(x)       (((x) & 0x3ff) << 2)
-#define VDPU_REG_PRED_FLT9                     0x1fc
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_6_3(x)       (((x) & 0x3ff) << 22)
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_7_0(x)       (((x) & 0x3ff) << 12)
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_7_1(x)       (((x) & 0x3ff) << 2)
-#define VDPU_REG_PRED_FLT10                    0x200
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_7_2(x)       (((x) & 0x3ff) << 22)
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_7_3(x)       (((x) & 0x3ff) << 12)
-#define     VDPU_REG_BD_REF_PIC_PRED_TAP_2_M1(x)       (((x) & 0x3) << 10)
-#define     VDPU_REG_BD_REF_PIC_PRED_TAP_2_4(x)                (((x) & 0x3) << 8)
-#define     VDPU_REG_BD_REF_PIC_PRED_TAP_4_M1(x)       (((x) & 0x3) << 6)
-#define     VDPU_REG_BD_REF_PIC_PRED_TAP_4_4(x)                (((x) & 0x3) << 4)
-#define     VDPU_REG_BD_REF_PIC_PRED_TAP_6_M1(x)       (((x) & 0x3) << 2)
-#define     VDPU_REG_BD_REF_PIC_PRED_TAP_6_4(x)                (((x) & 0x3) << 0)
-#define VDPU_REG_FILTER_LEVEL                  0x204
-#define     VDPU_REG_REF_PIC_LF_LEVEL_0(x)             (((x) & 0x3f) << 18)
-#define     VDPU_REG_REF_PIC_LF_LEVEL_1(x)             (((x) & 0x3f) << 12)
-#define     VDPU_REG_REF_PIC_LF_LEVEL_2(x)             (((x) & 0x3f) << 6)
-#define     VDPU_REG_REF_PIC_LF_LEVEL_3(x)             (((x) & 0x3f) << 0)
-#define VDPU_REG_VP8_QUANTER0                  0x208
-#define     VDPU_REG_REF_PIC_QUANT_DELTA_0(x)          (((x) & 0x1f) << 27)
-#define     VDPU_REG_REF_PIC_QUANT_DELTA_1(x)          (((x) & 0x1f) << 22)
-#define     VDPU_REG_REF_PIC_QUANT_0(x)                        (((x) & 0x7ff) << 11)
-#define     VDPU_REG_REF_PIC_QUANT_1(x)                        (((x) & 0x7ff) << 0)
-#define VDPU_REG_VP8_ADDR_REF0                 0x20c
-#define VDPU_REG_FILTER_MB_ADJ                 0x210
-#define     VDPU_REG_REF_PIC_FILT_TYPE_E               BIT(31)
-#define     VDPU_REG_REF_PIC_FILT_SHARPNESS(x)         (((x) & 0x7) << 28)
-#define     VDPU_REG_FILT_MB_ADJ_0(x)                  (((x) & 0x7f) << 21)
-#define     VDPU_REG_FILT_MB_ADJ_1(x)                  (((x) & 0x7f) << 14)
-#define     VDPU_REG_FILT_MB_ADJ_2(x)                  (((x) & 0x7f) << 7)
-#define     VDPU_REG_FILT_MB_ADJ_3(x)                  (((x) & 0x7f) << 0)
-#define VDPU_REG_FILTER_REF_ADJ                        0x214
-#define     VDPU_REG_REF_PIC_ADJ_0(x)                  (((x) & 0x7f) << 21)
-#define     VDPU_REG_REF_PIC_ADJ_1(x)                  (((x) & 0x7f) << 14)
-#define     VDPU_REG_REF_PIC_ADJ_2(x)                  (((x) & 0x7f) << 7)
-#define     VDPU_REG_REF_PIC_ADJ_3(x)                  (((x) & 0x7f) << 0)
-#define VDPU_REG_VP8_ADDR_REF2_5(i)            (0x218 + ((i) * 0x4))
-#define     VDPU_REG_VP8_GREF_SIGN_BIAS                        BIT(0)
-#define     VDPU_REG_VP8_AREF_SIGN_BIAS                        BIT(0)
-#define VDPU_REG_VP8_DCT_BASE(i)               (0x230 + ((i) * 0x4))
-#define VDPU_REG_VP8_ADDR_CTRL_PART            0x244
-#define VDPU_REG_VP8_ADDR_REF1                 0x250
-#define VDPU_REG_VP8_SEGMENT_VAL               0x254
-#define     VDPU_REG_FWD_PIC1_SEGMENT_BASE(x)          ((x) << 0)
-#define     VDPU_REG_FWD_PIC1_SEGMENT_UPD_E            BIT(1)
-#define     VDPU_REG_FWD_PIC1_SEGMENT_E                        BIT(0)
-#define VDPU_REG_VP8_DCT_START_BIT2            0x258
-#define     VDPU_REG_DEC_CTRL7_DCT3_START_BIT(x)       (((x) & 0x3f) << 24)
-#define     VDPU_REG_DEC_CTRL7_DCT4_START_BIT(x)       (((x) & 0x3f) << 18)
-#define     VDPU_REG_DEC_CTRL7_DCT5_START_BIT(x)       (((x) & 0x3f) << 12)
-#define     VDPU_REG_DEC_CTRL7_DCT6_START_BIT(x)       (((x) & 0x3f) << 6)
-#define     VDPU_REG_DEC_CTRL7_DCT7_START_BIT(x)       (((x) & 0x3f) << 0)
-#define VDPU_REG_VP8_QUANTER1                  0x25c
-#define     VDPU_REG_REF_PIC_QUANT_DELTA_2(x)          (((x) & 0x1f) << 27)
-#define     VDPU_REG_REF_PIC_QUANT_DELTA_3(x)          (((x) & 0x1f) << 22)
-#define     VDPU_REG_REF_PIC_QUANT_2(x)                        (((x) & 0x7ff) << 11)
-#define     VDPU_REG_REF_PIC_QUANT_3(x)                        (((x) & 0x7ff) << 0)
-#define VDPU_REG_VP8_QUANTER2                  0x260
-#define     VDPU_REG_REF_PIC_QUANT_DELTA_4(x)          (((x) & 0x1f) << 27)
-#define     VDPU_REG_REF_PIC_QUANT_4(x)                        (((x) & 0x7ff) << 11)
-#define     VDPU_REG_REF_PIC_QUANT_5(x)                        (((x) & 0x7ff) << 0)
-#define VDPU_REG_PRED_FLT1                     0x264
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_0_3(x)       (((x) & 0x3ff) << 22)
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_1_0(x)       (((x) & 0x3ff) << 12)
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_1_1(x)       (((x) & 0x3ff) << 2)
-#define VDPU_REG_PRED_FLT2                     0x268
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_1_2(x)       (((x) & 0x3ff) << 22)
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_1_3(x)       (((x) & 0x3ff) << 12)
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_2_0(x)       (((x) & 0x3ff) << 2)
-#define VDPU_REG_PRED_FLT3                     0x26c
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_2_1(x)       (((x) & 0x3ff) << 22)
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_2_2(x)       (((x) & 0x3ff) << 12)
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_2_3(x)       (((x) & 0x3ff) << 2)
-#define VDPU_REG_PRED_FLT4                     0x270
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_3_0(x)       (((x) & 0x3ff) << 22)
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_3_1(x)       (((x) & 0x3ff) << 12)
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_3_2(x)       (((x) & 0x3ff) << 2)
-#define VDPU_REG_PRED_FLT5                     0x274
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_3_3(x)       (((x) & 0x3ff) << 22)
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_4_0(x)       (((x) & 0x3ff) << 12)
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_4_1(x)       (((x) & 0x3ff) << 2)
-#define VDPU_REG_PRED_FLT6                     0x278
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_4_2(x)       (((x) & 0x3ff) << 22)
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_4_3(x)       (((x) & 0x3ff) << 12)
-#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_5_0(x)       (((x) & 0x3ff) << 2)
-
-#endif /* RK3399_VPU_REGS_H_ */
diff --git a/drivers/staging/media/hantro/rockchip_vpu2_hw_jpeg_enc.c b/drivers/staging/media/hantro/rockchip_vpu2_hw_jpeg_enc.c
new file mode 100644 (file)
index 0000000..991213c
--- /dev/null
@@ -0,0 +1,173 @@
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * Hantro VPU codec driver
+ *
+ * Copyright (C) 2018 Rockchip Electronics Co., Ltd.
+ *
+ * JPEG encoder
+ * ------------
+ * The VPU JPEG encoder produces JPEG baseline sequential format.
+ * The quantization coefficients are 8-bit values, complying with
+ * the baseline specification. Therefore, it requires
+ * luma and chroma quantization tables. The hardware does entropy
+ * encoding using internal Huffman tables, as specified in the JPEG
+ * specification.
+ *
+ * In other words, only the luma and chroma quantization tables are
+ * required for the encoding operation.
+ *
+ * Quantization luma table values are written to registers
+ * VEPU_swreg_0-VEPU_swreg_15, and chroma table values to
+ * VEPU_swreg_16-VEPU_swreg_31. A special order is needed, neither
+ * zigzag, nor linear.
+ */
+
+#include <asm/unaligned.h>
+#include <media/v4l2-mem2mem.h>
+#include "hantro_jpeg.h"
+#include "hantro.h"
+#include "hantro_v4l2.h"
+#include "hantro_hw.h"
+#include "rockchip_vpu2_regs.h"
+
+#define VEPU_JPEG_QUANT_TABLE_COUNT 16
+
+static void rockchip_vpu2_set_src_img_ctrl(struct hantro_dev *vpu,
+                                          struct hantro_ctx *ctx)
+{
+       struct v4l2_pix_format_mplane *pix_fmt = &ctx->src_fmt;
+       u32 reg;
+
+       /*
+        * The pix fmt width/height are already macroblock aligned
+        * by .vidioc_s_fmt_vid_cap_mplane() callback
+        */
+       reg = VEPU_REG_IN_IMG_CTRL_ROW_LEN(pix_fmt->width);
+       vepu_write_relaxed(vpu, reg, VEPU_REG_INPUT_LUMA_INFO);
+
+       reg = VEPU_REG_IN_IMG_CTRL_OVRFLR_D4(0) |
+             VEPU_REG_IN_IMG_CTRL_OVRFLB(0);
+       /*
+        * This register controls the input crop, as the offset
+        * from the right/bottom within the last macroblock. The offset from the
+        * right must be divided by 4 and so the crop must be aligned to 4 pixels
+        * horizontally.
+        */
+       vepu_write_relaxed(vpu, reg, VEPU_REG_ENC_OVER_FILL_STRM_OFFSET);
+
+       reg = VEPU_REG_IN_IMG_CTRL_FMT(ctx->vpu_src_fmt->enc_fmt);
+       vepu_write_relaxed(vpu, reg, VEPU_REG_ENC_CTRL1);
+}
+
+static void rockchip_vpu2_jpeg_enc_set_buffers(struct hantro_dev *vpu,
+                                              struct hantro_ctx *ctx,
+                                              struct vb2_buffer *src_buf)
+{
+       struct v4l2_pix_format_mplane *pix_fmt = &ctx->src_fmt;
+       dma_addr_t src[3];
+
+       WARN_ON(pix_fmt->num_planes > 3);
+
+       vepu_write_relaxed(vpu, ctx->jpeg_enc.bounce_buffer.dma,
+                          VEPU_REG_ADDR_OUTPUT_STREAM);
+       vepu_write_relaxed(vpu, ctx->jpeg_enc.bounce_buffer.size,
+                          VEPU_REG_STR_BUF_LIMIT);
+
+       if (pix_fmt->num_planes == 1) {
+               src[0] = vb2_dma_contig_plane_dma_addr(src_buf, 0);
+               vepu_write_relaxed(vpu, src[0], VEPU_REG_ADDR_IN_PLANE_0);
+       } else if (pix_fmt->num_planes == 2) {
+               src[0] = vb2_dma_contig_plane_dma_addr(src_buf, 0);
+               src[1] = vb2_dma_contig_plane_dma_addr(src_buf, 1);
+               vepu_write_relaxed(vpu, src[0], VEPU_REG_ADDR_IN_PLANE_0);
+               vepu_write_relaxed(vpu, src[1], VEPU_REG_ADDR_IN_PLANE_1);
+       } else {
+               src[0] = vb2_dma_contig_plane_dma_addr(src_buf, 0);
+               src[1] = vb2_dma_contig_plane_dma_addr(src_buf, 1);
+               src[2] = vb2_dma_contig_plane_dma_addr(src_buf, 2);
+               vepu_write_relaxed(vpu, src[0], VEPU_REG_ADDR_IN_PLANE_0);
+               vepu_write_relaxed(vpu, src[1], VEPU_REG_ADDR_IN_PLANE_1);
+               vepu_write_relaxed(vpu, src[2], VEPU_REG_ADDR_IN_PLANE_2);
+       }
+}
+
+static void
+rockchip_vpu2_jpeg_enc_set_qtable(struct hantro_dev *vpu,
+                                 unsigned char *luma_qtable,
+                                 unsigned char *chroma_qtable)
+{
+       u32 reg, i;
+       __be32 *luma_qtable_p;
+       __be32 *chroma_qtable_p;
+
+       luma_qtable_p = (__be32 *)luma_qtable;
+       chroma_qtable_p = (__be32 *)chroma_qtable;
+
+       /*
+        * Quantization table registers must be written in contiguous blocks.
+        * DO NOT collapse the below two "for" loops into one.
+        */
+       for (i = 0; i < VEPU_JPEG_QUANT_TABLE_COUNT; i++) {
+               reg = get_unaligned_be32(&luma_qtable_p[i]);
+               vepu_write_relaxed(vpu, reg, VEPU_REG_JPEG_LUMA_QUAT(i));
+       }
+
+       for (i = 0; i < VEPU_JPEG_QUANT_TABLE_COUNT; i++) {
+               reg = get_unaligned_be32(&chroma_qtable_p[i]);
+               vepu_write_relaxed(vpu, reg, VEPU_REG_JPEG_CHROMA_QUAT(i));
+       }
+}
+
+int rockchip_vpu2_jpeg_enc_run(struct hantro_ctx *ctx)
+{
+       struct hantro_dev *vpu = ctx->dev;
+       struct vb2_v4l2_buffer *src_buf, *dst_buf;
+       struct hantro_jpeg_ctx jpeg_ctx;
+       u32 reg;
+
+       src_buf = hantro_get_src_buf(ctx);
+       dst_buf = hantro_get_dst_buf(ctx);
+
+       hantro_start_prepare_run(ctx);
+
+       memset(&jpeg_ctx, 0, sizeof(jpeg_ctx));
+       jpeg_ctx.buffer = vb2_plane_vaddr(&dst_buf->vb2_buf, 0);
+       jpeg_ctx.width = ctx->dst_fmt.width;
+       jpeg_ctx.height = ctx->dst_fmt.height;
+       jpeg_ctx.quality = ctx->jpeg_quality;
+       hantro_jpeg_header_assemble(&jpeg_ctx);
+
+       /* Switch to JPEG encoder mode before writing registers */
+       vepu_write_relaxed(vpu, VEPU_REG_ENCODE_FORMAT_JPEG,
+                          VEPU_REG_ENCODE_START);
+
+       rockchip_vpu2_set_src_img_ctrl(vpu, ctx);
+       rockchip_vpu2_jpeg_enc_set_buffers(vpu, ctx, &src_buf->vb2_buf);
+       rockchip_vpu2_jpeg_enc_set_qtable(vpu,
+                                         hantro_jpeg_get_qtable(0),
+                                         hantro_jpeg_get_qtable(1));
+
+       reg = VEPU_REG_OUTPUT_SWAP32
+               | VEPU_REG_OUTPUT_SWAP16
+               | VEPU_REG_OUTPUT_SWAP8
+               | VEPU_REG_INPUT_SWAP8
+               | VEPU_REG_INPUT_SWAP16
+               | VEPU_REG_INPUT_SWAP32;
+       /* Make sure that all registers are written at this point. */
+       vepu_write(vpu, reg, VEPU_REG_DATA_ENDIAN);
+
+       reg = VEPU_REG_AXI_CTRL_BURST_LEN(16);
+       vepu_write_relaxed(vpu, reg, VEPU_REG_AXI_CTRL);
+
+       reg = VEPU_REG_MB_WIDTH(MB_WIDTH(ctx->src_fmt.width))
+               | VEPU_REG_MB_HEIGHT(MB_HEIGHT(ctx->src_fmt.height))
+               | VEPU_REG_FRAME_TYPE_INTRA
+               | VEPU_REG_ENCODE_FORMAT_JPEG
+               | VEPU_REG_ENCODE_ENABLE;
+
+       /* Kick the watchdog and start encoding */
+       hantro_end_prepare_run(ctx);
+       vepu_write(vpu, reg, VEPU_REG_ENCODE_START);
+
+       return 0;
+}
diff --git a/drivers/staging/media/hantro/rockchip_vpu2_hw_mpeg2_dec.c b/drivers/staging/media/hantro/rockchip_vpu2_hw_mpeg2_dec.c
new file mode 100644 (file)
index 0000000..b66737f
--- /dev/null
@@ -0,0 +1,248 @@
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * Hantro VPU codec driver
+ *
+ * Copyright (C) 2018 Rockchip Electronics Co., Ltd.
+ */
+
+#include <asm/unaligned.h>
+#include <linux/bitfield.h>
+#include <media/v4l2-mem2mem.h>
+#include "hantro.h"
+#include "hantro_hw.h"
+
+#define VDPU_SWREG(nr)                 ((nr) * 4)
+
+#define VDPU_REG_DEC_OUT_BASE          VDPU_SWREG(63)
+#define VDPU_REG_RLC_VLC_BASE          VDPU_SWREG(64)
+#define VDPU_REG_QTABLE_BASE           VDPU_SWREG(61)
+#define VDPU_REG_REFER0_BASE           VDPU_SWREG(131)
+#define VDPU_REG_REFER2_BASE           VDPU_SWREG(134)
+#define VDPU_REG_REFER3_BASE           VDPU_SWREG(135)
+#define VDPU_REG_REFER1_BASE           VDPU_SWREG(148)
+#define VDPU_REG_DEC_E(v)              ((v) ? BIT(0) : 0)
+
+#define VDPU_REG_DEC_ADV_PRE_DIS(v)    ((v) ? BIT(11) : 0)
+#define VDPU_REG_DEC_SCMD_DIS(v)       ((v) ? BIT(10) : 0)
+#define VDPU_REG_FILTERING_DIS(v)      ((v) ? BIT(8) : 0)
+#define VDPU_REG_DEC_LATENCY(v)                (((v) << 1) & GENMASK(6, 1))
+
+#define VDPU_REG_INIT_QP(v)            (((v) << 25) & GENMASK(30, 25))
+#define VDPU_REG_STREAM_LEN(v)         (((v) << 0) & GENMASK(23, 0))
+
+#define VDPU_REG_APF_THRESHOLD(v)      (((v) << 17) & GENMASK(30, 17))
+#define VDPU_REG_STARTMB_X(v)          (((v) << 8) & GENMASK(16, 8))
+#define VDPU_REG_STARTMB_Y(v)          (((v) << 0) & GENMASK(7, 0))
+
+#define VDPU_REG_DEC_MODE(v)           (((v) << 0) & GENMASK(3, 0))
+
+#define VDPU_REG_DEC_STRENDIAN_E(v)    ((v) ? BIT(5) : 0)
+#define VDPU_REG_DEC_STRSWAP32_E(v)    ((v) ? BIT(4) : 0)
+#define VDPU_REG_DEC_OUTSWAP32_E(v)    ((v) ? BIT(3) : 0)
+#define VDPU_REG_DEC_INSWAP32_E(v)     ((v) ? BIT(2) : 0)
+#define VDPU_REG_DEC_OUT_ENDIAN(v)     ((v) ? BIT(1) : 0)
+#define VDPU_REG_DEC_IN_ENDIAN(v)      ((v) ? BIT(0) : 0)
+
+#define VDPU_REG_DEC_DATA_DISC_E(v)    ((v) ? BIT(22) : 0)
+#define VDPU_REG_DEC_MAX_BURST(v)      (((v) << 16) & GENMASK(20, 16))
+#define VDPU_REG_DEC_AXI_WR_ID(v)      (((v) << 8) & GENMASK(15, 8))
+#define VDPU_REG_DEC_AXI_RD_ID(v)      (((v) << 0) & GENMASK(7, 0))
+
+#define VDPU_REG_RLC_MODE_E(v)         ((v) ? BIT(20) : 0)
+#define VDPU_REG_PIC_INTERLACE_E(v)    ((v) ? BIT(17) : 0)
+#define VDPU_REG_PIC_FIELDMODE_E(v)    ((v) ? BIT(16) : 0)
+#define VDPU_REG_PIC_B_E(v)            ((v) ? BIT(15) : 0)
+#define VDPU_REG_PIC_INTER_E(v)                ((v) ? BIT(14) : 0)
+#define VDPU_REG_PIC_TOPFIELD_E(v)     ((v) ? BIT(13) : 0)
+#define VDPU_REG_FWD_INTERLACE_E(v)    ((v) ? BIT(12) : 0)
+#define VDPU_REG_WRITE_MVS_E(v)                ((v) ? BIT(10) : 0)
+#define VDPU_REG_DEC_TIMEOUT_E(v)      ((v) ? BIT(5) : 0)
+#define VDPU_REG_DEC_CLK_GATE_E(v)     ((v) ? BIT(4) : 0)
+
+#define VDPU_REG_PIC_MB_WIDTH(v)       (((v) << 23) & GENMASK(31, 23))
+#define VDPU_REG_PIC_MB_HEIGHT_P(v)    (((v) << 11) & GENMASK(18, 11))
+#define VDPU_REG_ALT_SCAN_E(v)         ((v) ? BIT(6) : 0)
+#define VDPU_REG_TOPFIELDFIRST_E(v)    ((v) ? BIT(5) : 0)
+
+#define VDPU_REG_STRM_START_BIT(v)     (((v) << 26) & GENMASK(31, 26))
+#define VDPU_REG_QSCALE_TYPE(v)                ((v) ? BIT(24) : 0)
+#define VDPU_REG_CON_MV_E(v)           ((v) ? BIT(4) : 0)
+#define VDPU_REG_INTRA_DC_PREC(v)      (((v) << 2) & GENMASK(3, 2))
+#define VDPU_REG_INTRA_VLC_TAB(v)      ((v) ? BIT(1) : 0)
+#define VDPU_REG_FRAME_PRED_DCT(v)     ((v) ? BIT(0) : 0)
+
+#define VDPU_REG_ALT_SCAN_FLAG_E(v)    ((v) ? BIT(19) : 0)
+#define VDPU_REG_FCODE_FWD_HOR(v)      (((v) << 15) & GENMASK(18, 15))
+#define VDPU_REG_FCODE_FWD_VER(v)      (((v) << 11) & GENMASK(14, 11))
+#define VDPU_REG_FCODE_BWD_HOR(v)      (((v) << 7) & GENMASK(10, 7))
+#define VDPU_REG_FCODE_BWD_VER(v)      (((v) << 3) & GENMASK(6, 3))
+#define VDPU_REG_MV_ACCURACY_FWD(v)    ((v) ? BIT(2) : 0)
+#define VDPU_REG_MV_ACCURACY_BWD(v)    ((v) ? BIT(1) : 0)
+
+static void
+rockchip_vpu2_mpeg2_dec_set_quantisation(struct hantro_dev *vpu,
+                                        struct hantro_ctx *ctx)
+{
+       struct v4l2_ctrl_mpeg2_quantisation *q;
+
+       q = hantro_get_ctrl(ctx, V4L2_CID_STATELESS_MPEG2_QUANTISATION);
+       hantro_mpeg2_dec_copy_qtable(ctx->mpeg2_dec.qtable.cpu, q);
+       vdpu_write_relaxed(vpu, ctx->mpeg2_dec.qtable.dma, VDPU_REG_QTABLE_BASE);
+}
+
+static void
+rockchip_vpu2_mpeg2_dec_set_buffers(struct hantro_dev *vpu,
+                                   struct hantro_ctx *ctx,
+                                   struct vb2_buffer *src_buf,
+                                   struct vb2_buffer *dst_buf,
+                                   const struct v4l2_ctrl_mpeg2_sequence *seq,
+                                   const struct v4l2_ctrl_mpeg2_picture *pic)
+{
+       dma_addr_t forward_addr = 0, backward_addr = 0;
+       dma_addr_t current_addr, addr;
+
+       switch (pic->picture_coding_type) {
+       case V4L2_MPEG2_PIC_CODING_TYPE_B:
+               backward_addr = hantro_get_ref(ctx, pic->backward_ref_ts);
+               fallthrough;
+       case V4L2_MPEG2_PIC_CODING_TYPE_P:
+               forward_addr = hantro_get_ref(ctx, pic->forward_ref_ts);
+       }
+
+       /* Source bitstream buffer */
+       addr = vb2_dma_contig_plane_dma_addr(src_buf, 0);
+       vdpu_write_relaxed(vpu, addr, VDPU_REG_RLC_VLC_BASE);
+
+       /* Destination frame buffer */
+       addr = vb2_dma_contig_plane_dma_addr(dst_buf, 0);
+       current_addr = addr;
+
+       if (pic->picture_structure == V4L2_MPEG2_PIC_BOTTOM_FIELD)
+               addr += ALIGN(ctx->dst_fmt.width, 16);
+       vdpu_write_relaxed(vpu, addr, VDPU_REG_DEC_OUT_BASE);
+
+       if (!forward_addr)
+               forward_addr = current_addr;
+       if (!backward_addr)
+               backward_addr = current_addr;
+
+       /* Set forward ref frame (top/bottom field) */
+       if (pic->picture_structure == V4L2_MPEG2_PIC_FRAME ||
+           pic->picture_coding_type == V4L2_MPEG2_PIC_CODING_TYPE_B ||
+           (pic->picture_structure == V4L2_MPEG2_PIC_TOP_FIELD &&
+            pic->flags & V4L2_MPEG2_PIC_TOP_FIELD) ||
+           (pic->picture_structure == V4L2_MPEG2_PIC_BOTTOM_FIELD &&
+            !(pic->flags & V4L2_MPEG2_PIC_TOP_FIELD))) {
+               vdpu_write_relaxed(vpu, forward_addr, VDPU_REG_REFER0_BASE);
+               vdpu_write_relaxed(vpu, forward_addr, VDPU_REG_REFER1_BASE);
+       } else if (pic->picture_structure == V4L2_MPEG2_PIC_TOP_FIELD) {
+               vdpu_write_relaxed(vpu, forward_addr, VDPU_REG_REFER0_BASE);
+               vdpu_write_relaxed(vpu, current_addr, VDPU_REG_REFER1_BASE);
+       } else if (pic->picture_structure == V4L2_MPEG2_PIC_BOTTOM_FIELD) {
+               vdpu_write_relaxed(vpu, current_addr, VDPU_REG_REFER0_BASE);
+               vdpu_write_relaxed(vpu, forward_addr, VDPU_REG_REFER1_BASE);
+       }
+
+       /* Set backward ref frame (top/bottom field) */
+       vdpu_write_relaxed(vpu, backward_addr, VDPU_REG_REFER2_BASE);
+       vdpu_write_relaxed(vpu, backward_addr, VDPU_REG_REFER3_BASE);
+}
+
+int rockchip_vpu2_mpeg2_dec_run(struct hantro_ctx *ctx)
+{
+       struct hantro_dev *vpu = ctx->dev;
+       struct vb2_v4l2_buffer *src_buf, *dst_buf;
+       const struct v4l2_ctrl_mpeg2_sequence *seq;
+       const struct v4l2_ctrl_mpeg2_picture *pic;
+       u32 reg;
+
+       src_buf = hantro_get_src_buf(ctx);
+       dst_buf = hantro_get_dst_buf(ctx);
+
+       hantro_start_prepare_run(ctx);
+
+       seq = hantro_get_ctrl(ctx,
+                             V4L2_CID_STATELESS_MPEG2_SEQUENCE);
+       pic = hantro_get_ctrl(ctx,
+                             V4L2_CID_STATELESS_MPEG2_PICTURE);
+
+       reg = VDPU_REG_DEC_ADV_PRE_DIS(0) |
+             VDPU_REG_DEC_SCMD_DIS(0) |
+             VDPU_REG_FILTERING_DIS(1) |
+             VDPU_REG_DEC_LATENCY(0);
+       vdpu_write_relaxed(vpu, reg, VDPU_SWREG(50));
+
+       reg = VDPU_REG_INIT_QP(1) |
+             VDPU_REG_STREAM_LEN(vb2_get_plane_payload(&src_buf->vb2_buf, 0));
+       vdpu_write_relaxed(vpu, reg, VDPU_SWREG(51));
+
+       reg = VDPU_REG_APF_THRESHOLD(8) |
+             VDPU_REG_STARTMB_X(0) |
+             VDPU_REG_STARTMB_Y(0);
+       vdpu_write_relaxed(vpu, reg, VDPU_SWREG(52));
+
+       reg = VDPU_REG_DEC_MODE(5);
+       vdpu_write_relaxed(vpu, reg, VDPU_SWREG(53));
+
+       reg = VDPU_REG_DEC_STRENDIAN_E(1) |
+             VDPU_REG_DEC_STRSWAP32_E(1) |
+             VDPU_REG_DEC_OUTSWAP32_E(1) |
+             VDPU_REG_DEC_INSWAP32_E(1) |
+             VDPU_REG_DEC_OUT_ENDIAN(1) |
+             VDPU_REG_DEC_IN_ENDIAN(1);
+       vdpu_write_relaxed(vpu, reg, VDPU_SWREG(54));
+
+       reg = VDPU_REG_DEC_DATA_DISC_E(0) |
+             VDPU_REG_DEC_MAX_BURST(16) |
+             VDPU_REG_DEC_AXI_WR_ID(0) |
+             VDPU_REG_DEC_AXI_RD_ID(0);
+       vdpu_write_relaxed(vpu, reg, VDPU_SWREG(56));
+
+       reg = VDPU_REG_RLC_MODE_E(0) |
+             VDPU_REG_PIC_INTERLACE_E(!(seq->flags & V4L2_MPEG2_SEQ_FLAG_PROGRESSIVE)) |
+             VDPU_REG_PIC_FIELDMODE_E(pic->picture_structure != V4L2_MPEG2_PIC_FRAME) |
+             VDPU_REG_PIC_B_E(pic->picture_coding_type == V4L2_MPEG2_PIC_CODING_TYPE_B) |
+             VDPU_REG_PIC_INTER_E(pic->picture_coding_type != V4L2_MPEG2_PIC_CODING_TYPE_I) |
+             VDPU_REG_PIC_TOPFIELD_E(pic->picture_structure == V4L2_MPEG2_PIC_TOP_FIELD) |
+             VDPU_REG_FWD_INTERLACE_E(0) |
+             VDPU_REG_WRITE_MVS_E(0) |
+             VDPU_REG_DEC_TIMEOUT_E(1) |
+             VDPU_REG_DEC_CLK_GATE_E(1);
+       vdpu_write_relaxed(vpu, reg, VDPU_SWREG(57));
+
+       reg = VDPU_REG_PIC_MB_WIDTH(MB_WIDTH(ctx->dst_fmt.width)) |
+             VDPU_REG_PIC_MB_HEIGHT_P(MB_HEIGHT(ctx->dst_fmt.height)) |
+             VDPU_REG_ALT_SCAN_E(pic->flags & V4L2_MPEG2_PIC_FLAG_ALT_SCAN) |
+             VDPU_REG_TOPFIELDFIRST_E(pic->flags & V4L2_MPEG2_PIC_FLAG_TOP_FIELD_FIRST);
+       vdpu_write_relaxed(vpu, reg, VDPU_SWREG(120));
+
+       reg = VDPU_REG_STRM_START_BIT(0) |
+             VDPU_REG_QSCALE_TYPE(pic->flags & V4L2_MPEG2_PIC_FLAG_Q_SCALE_TYPE) |
+             VDPU_REG_CON_MV_E(pic->flags & V4L2_MPEG2_PIC_FLAG_CONCEALMENT_MV) |
+             VDPU_REG_INTRA_DC_PREC(pic->intra_dc_precision) |
+             VDPU_REG_INTRA_VLC_TAB(pic->flags & V4L2_MPEG2_PIC_FLAG_INTRA_VLC) |
+             VDPU_REG_FRAME_PRED_DCT(pic->flags & V4L2_MPEG2_PIC_FLAG_FRAME_PRED_DCT);
+       vdpu_write_relaxed(vpu, reg, VDPU_SWREG(122));
+
+       reg = VDPU_REG_ALT_SCAN_FLAG_E(pic->flags & V4L2_MPEG2_PIC_FLAG_ALT_SCAN) |
+             VDPU_REG_FCODE_FWD_HOR(pic->f_code[0][0]) |
+             VDPU_REG_FCODE_FWD_VER(pic->f_code[0][1]) |
+             VDPU_REG_FCODE_BWD_HOR(pic->f_code[1][0]) |
+             VDPU_REG_FCODE_BWD_VER(pic->f_code[1][1]) |
+             VDPU_REG_MV_ACCURACY_FWD(1) |
+             VDPU_REG_MV_ACCURACY_BWD(1);
+       vdpu_write_relaxed(vpu, reg, VDPU_SWREG(136));
+
+       rockchip_vpu2_mpeg2_dec_set_quantisation(vpu, ctx);
+
+       rockchip_vpu2_mpeg2_dec_set_buffers(vpu, ctx, &src_buf->vb2_buf,
+                                           &dst_buf->vb2_buf, seq, pic);
+
+       /* Kick the watchdog and start decoding */
+       hantro_end_prepare_run(ctx);
+
+       reg = vdpu_read(vpu, VDPU_SWREG(57)) | VDPU_REG_DEC_E(1);
+       vdpu_write(vpu, reg, VDPU_SWREG(57));
+
+       return 0;
+}
diff --git a/drivers/staging/media/hantro/rockchip_vpu2_hw_vp8_dec.c b/drivers/staging/media/hantro/rockchip_vpu2_hw_vp8_dec.c
new file mode 100644 (file)
index 0000000..951b55f
--- /dev/null
@@ -0,0 +1,594 @@
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * Rockchip VPU codec vp8 decode driver
+ *
+ * Copyright (C) 2014 Rockchip Electronics Co., Ltd.
+ *     ZhiChao Yu <zhichao.yu@rock-chips.com>
+ *
+ * Copyright (C) 2014 Google LLC.
+ *      Tomasz Figa <tfiga@chromium.org>
+ *
+ * Copyright (C) 2015 Rockchip Electronics Co., Ltd.
+ *      Alpha Lin <alpha.lin@rock-chips.com>
+ */
+
+#include <media/v4l2-mem2mem.h>
+
+#include "hantro_hw.h"
+#include "hantro.h"
+#include "hantro_g1_regs.h"
+
+#define VDPU_REG_DEC_CTRL0                     0x0c8
+#define VDPU_REG_STREAM_LEN                    0x0cc
+#define VDPU_REG_DEC_FORMAT                    0x0d4
+#define     VDPU_REG_DEC_CTRL0_DEC_MODE(x)             (((x) & 0xf) << 0)
+#define VDPU_REG_DATA_ENDIAN                   0x0d8
+#define     VDPU_REG_CONFIG_DEC_STRENDIAN_E            BIT(5)
+#define     VDPU_REG_CONFIG_DEC_STRSWAP32_E            BIT(4)
+#define     VDPU_REG_CONFIG_DEC_OUTSWAP32_E            BIT(3)
+#define     VDPU_REG_CONFIG_DEC_INSWAP32_E             BIT(2)
+#define     VDPU_REG_CONFIG_DEC_OUT_ENDIAN             BIT(1)
+#define     VDPU_REG_CONFIG_DEC_IN_ENDIAN              BIT(0)
+#define VDPU_REG_AXI_CTRL                      0x0e0
+#define     VDPU_REG_CONFIG_DEC_MAX_BURST(x)           (((x) & 0x1f) << 16)
+#define VDPU_REG_EN_FLAGS                      0x0e4
+#define     VDPU_REG_DEC_CTRL0_PIC_INTER_E             BIT(14)
+#define     VDPU_REG_CONFIG_DEC_TIMEOUT_E              BIT(5)
+#define     VDPU_REG_CONFIG_DEC_CLK_GATE_E             BIT(4)
+#define VDPU_REG_PRED_FLT                      0x0ec
+#define VDPU_REG_ADDR_QTABLE                   0x0f4
+#define VDPU_REG_ADDR_DST                      0x0fc
+#define VDPU_REG_ADDR_STR                      0x100
+#define VDPU_REG_VP8_PIC_MB_SIZE               0x1e0
+#define VDPU_REG_VP8_DCT_START_BIT             0x1e4
+#define     VDPU_REG_DEC_CTRL4_VC1_HEIGHT_EXT          BIT(13)
+#define     VDPU_REG_DEC_CTRL4_BILIN_MC_E              BIT(12)
+#define VDPU_REG_VP8_CTRL0                     0x1e8
+#define VDPU_REG_VP8_DATA_VAL                  0x1f0
+#define VDPU_REG_PRED_FLT7                     0x1f4
+#define VDPU_REG_PRED_FLT8                     0x1f8
+#define VDPU_REG_PRED_FLT9                     0x1fc
+#define VDPU_REG_PRED_FLT10                    0x200
+#define VDPU_REG_FILTER_LEVEL                  0x204
+#define VDPU_REG_VP8_QUANTER0                  0x208
+#define VDPU_REG_VP8_ADDR_REF0                 0x20c
+#define VDPU_REG_FILTER_MB_ADJ                 0x210
+#define     VDPU_REG_REF_PIC_FILT_TYPE_E               BIT(31)
+#define     VDPU_REG_REF_PIC_FILT_SHARPNESS(x)         (((x) & 0x7) << 28)
+#define VDPU_REG_FILTER_REF_ADJ                        0x214
+#define VDPU_REG_VP8_ADDR_REF2_5(i)            (0x218 + ((i) * 0x4))
+#define     VDPU_REG_VP8_GREF_SIGN_BIAS                        BIT(0)
+#define     VDPU_REG_VP8_AREF_SIGN_BIAS                        BIT(0)
+#define VDPU_REG_VP8_DCT_BASE(i)               \
+               (0x230 + ((((i) < 5) ? (i) : ((i) + 1)) * 0x4))
+#define VDPU_REG_VP8_ADDR_CTRL_PART            0x244
+#define VDPU_REG_VP8_SEGMENT_VAL               0x254
+#define     VDPU_REG_FWD_PIC1_SEGMENT_BASE(x)          ((x) << 0)
+#define     VDPU_REG_FWD_PIC1_SEGMENT_UPD_E            BIT(1)
+#define     VDPU_REG_FWD_PIC1_SEGMENT_E                        BIT(0)
+#define VDPU_REG_VP8_DCT_START_BIT2            0x258
+#define VDPU_REG_VP8_QUANTER1                  0x25c
+#define VDPU_REG_VP8_QUANTER2                  0x260
+#define VDPU_REG_PRED_FLT1                     0x264
+#define VDPU_REG_PRED_FLT2                     0x268
+#define VDPU_REG_PRED_FLT3                     0x26c
+#define VDPU_REG_PRED_FLT4                     0x270
+#define VDPU_REG_PRED_FLT5                     0x274
+#define VDPU_REG_PRED_FLT6                     0x278
+
+static const struct hantro_reg vp8_dec_dct_base[8] = {
+       { VDPU_REG_ADDR_STR, 0, 0xffffffff },
+       { VDPU_REG_VP8_DCT_BASE(0), 0, 0xffffffff },
+       { VDPU_REG_VP8_DCT_BASE(1), 0, 0xffffffff },
+       { VDPU_REG_VP8_DCT_BASE(2), 0, 0xffffffff },
+       { VDPU_REG_VP8_DCT_BASE(3), 0, 0xffffffff },
+       { VDPU_REG_VP8_DCT_BASE(4), 0, 0xffffffff },
+       { VDPU_REG_VP8_DCT_BASE(5), 0, 0xffffffff },
+       { VDPU_REG_VP8_DCT_BASE(6), 0, 0xffffffff },
+};
+
+static const struct hantro_reg vp8_dec_lf_level[4] = {
+       { VDPU_REG_FILTER_LEVEL, 18, 0x3f },
+       { VDPU_REG_FILTER_LEVEL, 12, 0x3f },
+       { VDPU_REG_FILTER_LEVEL, 6, 0x3f },
+       { VDPU_REG_FILTER_LEVEL, 0, 0x3f },
+};
+
+static const struct hantro_reg vp8_dec_mb_adj[4] = {
+       { VDPU_REG_FILTER_MB_ADJ, 21, 0x7f },
+       { VDPU_REG_FILTER_MB_ADJ, 14, 0x7f },
+       { VDPU_REG_FILTER_MB_ADJ, 7, 0x7f },
+       { VDPU_REG_FILTER_MB_ADJ, 0, 0x7f },
+};
+
+static const struct hantro_reg vp8_dec_ref_adj[4] = {
+       { VDPU_REG_FILTER_REF_ADJ, 21, 0x7f },
+       { VDPU_REG_FILTER_REF_ADJ, 14, 0x7f },
+       { VDPU_REG_FILTER_REF_ADJ, 7, 0x7f },
+       { VDPU_REG_FILTER_REF_ADJ, 0, 0x7f },
+};
+
+static const struct hantro_reg vp8_dec_quant[4] = {
+       { VDPU_REG_VP8_QUANTER0, 11, 0x7ff },
+       { VDPU_REG_VP8_QUANTER0, 0, 0x7ff },
+       { VDPU_REG_VP8_QUANTER1, 11, 0x7ff },
+       { VDPU_REG_VP8_QUANTER1, 0, 0x7ff },
+};
+
+static const struct hantro_reg vp8_dec_quant_delta[5] = {
+       { VDPU_REG_VP8_QUANTER0, 27, 0x1f },
+       { VDPU_REG_VP8_QUANTER0, 22, 0x1f },
+       { VDPU_REG_VP8_QUANTER1, 27, 0x1f },
+       { VDPU_REG_VP8_QUANTER1, 22, 0x1f },
+       { VDPU_REG_VP8_QUANTER2, 27, 0x1f },
+};
+
+static const struct hantro_reg vp8_dec_dct_start_bits[8] = {
+       { VDPU_REG_VP8_CTRL0, 26, 0x3f },
+       { VDPU_REG_VP8_DCT_START_BIT, 26, 0x3f },
+       { VDPU_REG_VP8_DCT_START_BIT, 20, 0x3f },
+       { VDPU_REG_VP8_DCT_START_BIT2, 24, 0x3f },
+       { VDPU_REG_VP8_DCT_START_BIT2, 18, 0x3f },
+       { VDPU_REG_VP8_DCT_START_BIT2, 12, 0x3f },
+       { VDPU_REG_VP8_DCT_START_BIT2, 6, 0x3f },
+       { VDPU_REG_VP8_DCT_START_BIT2, 0, 0x3f },
+};
+
+static const struct hantro_reg vp8_dec_pred_bc_tap[8][6] = {
+       {
+               { 0, 0, 0},
+               { VDPU_REG_PRED_FLT, 22, 0x3ff },
+               { VDPU_REG_PRED_FLT, 12, 0x3ff },
+               { VDPU_REG_PRED_FLT, 2, 0x3ff },
+               { VDPU_REG_PRED_FLT1, 22, 0x3ff },
+               { 0, 0, 0},
+       }, {
+               { 0, 0, 0},
+               { VDPU_REG_PRED_FLT1, 12, 0x3ff },
+               { VDPU_REG_PRED_FLT1, 2, 0x3ff },
+               { VDPU_REG_PRED_FLT2, 22, 0x3ff },
+               { VDPU_REG_PRED_FLT2, 12, 0x3ff },
+               { 0, 0, 0},
+       }, {
+               { VDPU_REG_PRED_FLT10, 10, 0x3 },
+               { VDPU_REG_PRED_FLT2, 2, 0x3ff },
+               { VDPU_REG_PRED_FLT3, 22, 0x3ff },
+               { VDPU_REG_PRED_FLT3, 12, 0x3ff },
+               { VDPU_REG_PRED_FLT3, 2, 0x3ff },
+               { VDPU_REG_PRED_FLT10, 8, 0x3},
+       }, {
+               { 0, 0, 0},
+               { VDPU_REG_PRED_FLT4, 22, 0x3ff },
+               { VDPU_REG_PRED_FLT4, 12, 0x3ff },
+               { VDPU_REG_PRED_FLT4, 2, 0x3ff },
+               { VDPU_REG_PRED_FLT5, 22, 0x3ff },
+               { 0, 0, 0},
+       }, {
+               { VDPU_REG_PRED_FLT10, 6, 0x3 },
+               { VDPU_REG_PRED_FLT5, 12, 0x3ff },
+               { VDPU_REG_PRED_FLT5, 2, 0x3ff },
+               { VDPU_REG_PRED_FLT6, 22, 0x3ff },
+               { VDPU_REG_PRED_FLT6, 12, 0x3ff },
+               { VDPU_REG_PRED_FLT10, 4, 0x3 },
+       }, {
+               { 0, 0, 0},
+               { VDPU_REG_PRED_FLT6, 2, 0x3ff },
+               { VDPU_REG_PRED_FLT7, 22, 0x3ff },
+               { VDPU_REG_PRED_FLT7, 12, 0x3ff },
+               { VDPU_REG_PRED_FLT7, 2, 0x3ff },
+               { 0, 0, 0},
+       }, {
+               { VDPU_REG_PRED_FLT10, 2, 0x3 },
+               { VDPU_REG_PRED_FLT8, 22, 0x3ff },
+               { VDPU_REG_PRED_FLT8, 12, 0x3ff },
+               { VDPU_REG_PRED_FLT8, 2, 0x3ff },
+               { VDPU_REG_PRED_FLT9, 22, 0x3ff },
+               { VDPU_REG_PRED_FLT10, 0, 0x3 },
+       }, {
+               { 0, 0, 0},
+               { VDPU_REG_PRED_FLT9, 12, 0x3ff },
+               { VDPU_REG_PRED_FLT9, 2, 0x3ff },
+               { VDPU_REG_PRED_FLT10, 22, 0x3ff },
+               { VDPU_REG_PRED_FLT10, 12, 0x3ff },
+               { 0, 0, 0},
+       },
+};
+
+static const struct hantro_reg vp8_dec_mb_start_bit = {
+       .base = VDPU_REG_VP8_CTRL0,
+       .shift = 18,
+       .mask = 0x3f
+};
+
+static const struct hantro_reg vp8_dec_mb_aligned_data_len = {
+       .base = VDPU_REG_VP8_DATA_VAL,
+       .shift = 0,
+       .mask = 0x3fffff
+};
+
+static const struct hantro_reg vp8_dec_num_dct_partitions = {
+       .base = VDPU_REG_VP8_DATA_VAL,
+       .shift = 24,
+       .mask = 0xf
+};
+
+static const struct hantro_reg vp8_dec_stream_len = {
+       .base = VDPU_REG_STREAM_LEN,
+       .shift = 0,
+       .mask = 0xffffff
+};
+
+static const struct hantro_reg vp8_dec_mb_width = {
+       .base = VDPU_REG_VP8_PIC_MB_SIZE,
+       .shift = 23,
+       .mask = 0x1ff
+};
+
+static const struct hantro_reg vp8_dec_mb_height = {
+       .base = VDPU_REG_VP8_PIC_MB_SIZE,
+       .shift = 11,
+       .mask = 0xff
+};
+
+static const struct hantro_reg vp8_dec_mb_width_ext = {
+       .base = VDPU_REG_VP8_PIC_MB_SIZE,
+       .shift = 3,
+       .mask = 0x7
+};
+
+static const struct hantro_reg vp8_dec_mb_height_ext = {
+       .base = VDPU_REG_VP8_PIC_MB_SIZE,
+       .shift = 0,
+       .mask = 0x7
+};
+
+static const struct hantro_reg vp8_dec_bool_range = {
+       .base = VDPU_REG_VP8_CTRL0,
+       .shift = 0,
+       .mask = 0xff
+};
+
+static const struct hantro_reg vp8_dec_bool_value = {
+       .base = VDPU_REG_VP8_CTRL0,
+       .shift = 8,
+       .mask = 0xff
+};
+
+static const struct hantro_reg vp8_dec_filter_disable = {
+       .base = VDPU_REG_DEC_CTRL0,
+       .shift = 8,
+       .mask = 1
+};
+
+static const struct hantro_reg vp8_dec_skip_mode = {
+       .base = VDPU_REG_DEC_CTRL0,
+       .shift = 9,
+       .mask = 1
+};
+
+static const struct hantro_reg vp8_dec_start_dec = {
+       .base = VDPU_REG_EN_FLAGS,
+       .shift = 0,
+       .mask = 1
+};
+
+static void cfg_lf(struct hantro_ctx *ctx,
+                  const struct v4l2_ctrl_vp8_frame *hdr)
+{
+       const struct v4l2_vp8_segment *seg = &hdr->segment;
+       const struct v4l2_vp8_loop_filter *lf = &hdr->lf;
+       struct hantro_dev *vpu = ctx->dev;
+       unsigned int i;
+       u32 reg;
+
+       if (!(seg->flags & V4L2_VP8_SEGMENT_FLAG_ENABLED)) {
+               hantro_reg_write(vpu, &vp8_dec_lf_level[0], lf->level);
+       } else if (seg->flags & V4L2_VP8_SEGMENT_FLAG_DELTA_VALUE_MODE) {
+               for (i = 0; i < 4; i++) {
+                       u32 lf_level = clamp(lf->level + seg->lf_update[i],
+                                            0, 63);
+
+                       hantro_reg_write(vpu, &vp8_dec_lf_level[i], lf_level);
+               }
+       } else {
+               for (i = 0; i < 4; i++)
+                       hantro_reg_write(vpu, &vp8_dec_lf_level[i],
+                                        seg->lf_update[i]);
+       }
+
+       reg = VDPU_REG_REF_PIC_FILT_SHARPNESS(lf->sharpness_level);
+       if (lf->flags & V4L2_VP8_LF_FILTER_TYPE_SIMPLE)
+               reg |= VDPU_REG_REF_PIC_FILT_TYPE_E;
+       vdpu_write_relaxed(vpu, reg, VDPU_REG_FILTER_MB_ADJ);
+
+       if (lf->flags & V4L2_VP8_LF_ADJ_ENABLE) {
+               for (i = 0; i < 4; i++) {
+                       hantro_reg_write(vpu, &vp8_dec_mb_adj[i],
+                                        lf->mb_mode_delta[i]);
+                       hantro_reg_write(vpu, &vp8_dec_ref_adj[i],
+                                        lf->ref_frm_delta[i]);
+               }
+       }
+}
+
+static void cfg_qp(struct hantro_ctx *ctx,
+                  const struct v4l2_ctrl_vp8_frame *hdr)
+{
+       const struct v4l2_vp8_quantization *q = &hdr->quant;
+       const struct v4l2_vp8_segment *seg = &hdr->segment;
+       struct hantro_dev *vpu = ctx->dev;
+       unsigned int i;
+
+       if (!(seg->flags & V4L2_VP8_SEGMENT_FLAG_ENABLED)) {
+               hantro_reg_write(vpu, &vp8_dec_quant[0], q->y_ac_qi);
+       } else if (seg->flags & V4L2_VP8_SEGMENT_FLAG_DELTA_VALUE_MODE) {
+               for (i = 0; i < 4; i++) {
+                       u32 quant = clamp(q->y_ac_qi + seg->quant_update[i],
+                                         0, 127);
+
+                       hantro_reg_write(vpu, &vp8_dec_quant[i], quant);
+               }
+       } else {
+               for (i = 0; i < 4; i++)
+                       hantro_reg_write(vpu, &vp8_dec_quant[i],
+                                        seg->quant_update[i]);
+       }
+
+       hantro_reg_write(vpu, &vp8_dec_quant_delta[0], q->y_dc_delta);
+       hantro_reg_write(vpu, &vp8_dec_quant_delta[1], q->y2_dc_delta);
+       hantro_reg_write(vpu, &vp8_dec_quant_delta[2], q->y2_ac_delta);
+       hantro_reg_write(vpu, &vp8_dec_quant_delta[3], q->uv_dc_delta);
+       hantro_reg_write(vpu, &vp8_dec_quant_delta[4], q->uv_ac_delta);
+}
+
+static void cfg_parts(struct hantro_ctx *ctx,
+                     const struct v4l2_ctrl_vp8_frame *hdr)
+{
+       struct hantro_dev *vpu = ctx->dev;
+       struct vb2_v4l2_buffer *vb2_src;
+       u32 first_part_offset = V4L2_VP8_FRAME_IS_KEY_FRAME(hdr) ? 10 : 3;
+       u32 mb_size, mb_offset_bytes, mb_offset_bits, mb_start_bits;
+       u32 dct_size_part_size, dct_part_offset;
+       dma_addr_t src_dma;
+       u32 dct_part_total_len = 0;
+       u32 count = 0;
+       unsigned int i;
+
+       vb2_src = hantro_get_src_buf(ctx);
+       src_dma = vb2_dma_contig_plane_dma_addr(&vb2_src->vb2_buf, 0);
+
+       /*
+        * Calculate control partition mb data info
+        * @first_part_header_bits:     bits offset of mb data from first
+        *                              part start pos
+        * @mb_offset_bits:             bits offset of mb data from src_dma
+        *                              base addr
+        * @mb_offset_byte:             bytes offset of mb data from src_dma
+        *                              base addr
+        * @mb_start_bits:              bits offset of mb data from mb data
+        *                              64bits alignment addr
+        */
+       mb_offset_bits = first_part_offset * 8 +
+                        hdr->first_part_header_bits + 8;
+       mb_offset_bytes = mb_offset_bits / 8;
+       mb_start_bits = mb_offset_bits -
+                       (mb_offset_bytes & (~DEC_8190_ALIGN_MASK)) * 8;
+       mb_size = hdr->first_part_size -
+                 (mb_offset_bytes - first_part_offset) +
+                 (mb_offset_bytes & DEC_8190_ALIGN_MASK);
+
+       /* Macroblock data aligned base addr */
+       vdpu_write_relaxed(vpu, (mb_offset_bytes & (~DEC_8190_ALIGN_MASK)) +
+                          src_dma, VDPU_REG_VP8_ADDR_CTRL_PART);
+       hantro_reg_write(vpu, &vp8_dec_mb_start_bit, mb_start_bits);
+       hantro_reg_write(vpu, &vp8_dec_mb_aligned_data_len, mb_size);
+
+       /*
+        * Calculate DCT partition info
+        * @dct_size_part_size: Containing sizes of DCT part, every DCT part
+        *                      has 3 bytes to store its size, except the last
+        *                      DCT part
+        * @dct_part_offset:    bytes offset of DCT parts from src_dma base addr
+        * @dct_part_total_len: total size of all DCT parts
+        */
+       dct_size_part_size = (hdr->num_dct_parts - 1) * 3;
+       dct_part_offset = first_part_offset + hdr->first_part_size;
+       for (i = 0; i < hdr->num_dct_parts; i++)
+               dct_part_total_len += hdr->dct_part_sizes[i];
+       dct_part_total_len += dct_size_part_size;
+       dct_part_total_len += (dct_part_offset & DEC_8190_ALIGN_MASK);
+
+       /* Number of DCT partitions */
+       hantro_reg_write(vpu, &vp8_dec_num_dct_partitions,
+                        hdr->num_dct_parts - 1);
+
+       /* DCT partition length */
+       hantro_reg_write(vpu, &vp8_dec_stream_len, dct_part_total_len);
+
+       /* DCT partitions base address */
+       for (i = 0; i < hdr->num_dct_parts; i++) {
+               u32 byte_offset = dct_part_offset + dct_size_part_size + count;
+               u32 base_addr = byte_offset + src_dma;
+
+               hantro_reg_write(vpu, &vp8_dec_dct_base[i],
+                                base_addr & (~DEC_8190_ALIGN_MASK));
+
+               hantro_reg_write(vpu, &vp8_dec_dct_start_bits[i],
+                                (byte_offset & DEC_8190_ALIGN_MASK) * 8);
+
+               count += hdr->dct_part_sizes[i];
+       }
+}
+
+/*
+ * prediction filter taps
+ * normal 6-tap filters
+ */
+static void cfg_tap(struct hantro_ctx *ctx,
+                   const struct v4l2_ctrl_vp8_frame *hdr)
+{
+       struct hantro_dev *vpu = ctx->dev;
+       int i, j;
+
+       if ((hdr->version & 0x03) != 0)
+               return; /* Tap filter not used. */
+
+       for (i = 0; i < 8; i++) {
+               for (j = 0; j < 6; j++) {
+                       if (vp8_dec_pred_bc_tap[i][j].base != 0)
+                               hantro_reg_write(vpu,
+                                                &vp8_dec_pred_bc_tap[i][j],
+                                                hantro_vp8_dec_mc_filter[i][j]);
+               }
+       }
+}
+
+static void cfg_ref(struct hantro_ctx *ctx,
+                   const struct v4l2_ctrl_vp8_frame *hdr)
+{
+       struct hantro_dev *vpu = ctx->dev;
+       struct vb2_v4l2_buffer *vb2_dst;
+       dma_addr_t ref;
+
+       vb2_dst = hantro_get_dst_buf(ctx);
+
+       ref = hantro_get_ref(ctx, hdr->last_frame_ts);
+       if (!ref)
+               ref = vb2_dma_contig_plane_dma_addr(&vb2_dst->vb2_buf, 0);
+       vdpu_write_relaxed(vpu, ref, VDPU_REG_VP8_ADDR_REF0);
+
+       ref = hantro_get_ref(ctx, hdr->golden_frame_ts);
+       WARN_ON(!ref && hdr->golden_frame_ts);
+       if (!ref)
+               ref = vb2_dma_contig_plane_dma_addr(&vb2_dst->vb2_buf, 0);
+       if (hdr->flags & V4L2_VP8_FRAME_FLAG_SIGN_BIAS_GOLDEN)
+               ref |= VDPU_REG_VP8_GREF_SIGN_BIAS;
+       vdpu_write_relaxed(vpu, ref, VDPU_REG_VP8_ADDR_REF2_5(2));
+
+       ref = hantro_get_ref(ctx, hdr->alt_frame_ts);
+       WARN_ON(!ref && hdr->alt_frame_ts);
+       if (!ref)
+               ref = vb2_dma_contig_plane_dma_addr(&vb2_dst->vb2_buf, 0);
+       if (hdr->flags & V4L2_VP8_FRAME_FLAG_SIGN_BIAS_ALT)
+               ref |= VDPU_REG_VP8_AREF_SIGN_BIAS;
+       vdpu_write_relaxed(vpu, ref, VDPU_REG_VP8_ADDR_REF2_5(3));
+}
+
+static void cfg_buffers(struct hantro_ctx *ctx,
+                       const struct v4l2_ctrl_vp8_frame *hdr)
+{
+       const struct v4l2_vp8_segment *seg = &hdr->segment;
+       struct hantro_dev *vpu = ctx->dev;
+       struct vb2_v4l2_buffer *vb2_dst;
+       dma_addr_t dst_dma;
+       u32 reg;
+
+       vb2_dst = hantro_get_dst_buf(ctx);
+
+       /* Set probability table buffer address */
+       vdpu_write_relaxed(vpu, ctx->vp8_dec.prob_tbl.dma,
+                          VDPU_REG_ADDR_QTABLE);
+
+       /* Set segment map address */
+       reg = VDPU_REG_FWD_PIC1_SEGMENT_BASE(ctx->vp8_dec.segment_map.dma);
+       if (seg->flags & V4L2_VP8_SEGMENT_FLAG_ENABLED) {
+               reg |= VDPU_REG_FWD_PIC1_SEGMENT_E;
+               if (seg->flags & V4L2_VP8_SEGMENT_FLAG_UPDATE_MAP)
+                       reg |= VDPU_REG_FWD_PIC1_SEGMENT_UPD_E;
+       }
+       vdpu_write_relaxed(vpu, reg, VDPU_REG_VP8_SEGMENT_VAL);
+
+       /* set output frame buffer address */
+       dst_dma = vb2_dma_contig_plane_dma_addr(&vb2_dst->vb2_buf, 0);
+       vdpu_write_relaxed(vpu, dst_dma, VDPU_REG_ADDR_DST);
+}
+
+int rockchip_vpu2_vp8_dec_run(struct hantro_ctx *ctx)
+{
+       const struct v4l2_ctrl_vp8_frame *hdr;
+       struct hantro_dev *vpu = ctx->dev;
+       size_t height = ctx->dst_fmt.height;
+       size_t width = ctx->dst_fmt.width;
+       u32 mb_width, mb_height;
+       u32 reg;
+
+       hantro_start_prepare_run(ctx);
+
+       hdr = hantro_get_ctrl(ctx, V4L2_CID_STATELESS_VP8_FRAME);
+       if (WARN_ON(!hdr))
+               return -EINVAL;
+
+       /* Reset segment_map buffer in keyframe */
+       if (V4L2_VP8_FRAME_IS_KEY_FRAME(hdr) && ctx->vp8_dec.segment_map.cpu)
+               memset(ctx->vp8_dec.segment_map.cpu, 0,
+                      ctx->vp8_dec.segment_map.size);
+
+       hantro_vp8_prob_update(ctx, hdr);
+
+       /*
+        * Extensive testing shows that the hardware does not properly
+        * clear the internal state from previous a decoding run. This
+        * causes corruption in decoded frames for multi-instance use cases.
+        * A soft reset before programming the registers has been found
+        * to resolve those problems.
+        */
+       ctx->codec_ops->reset(ctx);
+
+       reg = VDPU_REG_CONFIG_DEC_TIMEOUT_E
+               | VDPU_REG_CONFIG_DEC_CLK_GATE_E;
+       if (!V4L2_VP8_FRAME_IS_KEY_FRAME(hdr))
+               reg |= VDPU_REG_DEC_CTRL0_PIC_INTER_E;
+       vdpu_write_relaxed(vpu, reg, VDPU_REG_EN_FLAGS);
+
+       reg = VDPU_REG_CONFIG_DEC_STRENDIAN_E
+               | VDPU_REG_CONFIG_DEC_INSWAP32_E
+               | VDPU_REG_CONFIG_DEC_STRSWAP32_E
+               | VDPU_REG_CONFIG_DEC_OUTSWAP32_E
+               | VDPU_REG_CONFIG_DEC_IN_ENDIAN
+               | VDPU_REG_CONFIG_DEC_OUT_ENDIAN;
+       vdpu_write_relaxed(vpu, reg, VDPU_REG_DATA_ENDIAN);
+
+       reg = VDPU_REG_CONFIG_DEC_MAX_BURST(16);
+       vdpu_write_relaxed(vpu, reg, VDPU_REG_AXI_CTRL);
+
+       reg = VDPU_REG_DEC_CTRL0_DEC_MODE(10);
+       vdpu_write_relaxed(vpu, reg, VDPU_REG_DEC_FORMAT);
+
+       if (!(hdr->flags & V4L2_VP8_FRAME_FLAG_MB_NO_SKIP_COEFF))
+               hantro_reg_write(vpu, &vp8_dec_skip_mode, 1);
+       if (hdr->lf.level == 0)
+               hantro_reg_write(vpu, &vp8_dec_filter_disable, 1);
+
+       /* Frame dimensions */
+       mb_width = MB_WIDTH(width);
+       mb_height = MB_HEIGHT(height);
+
+       hantro_reg_write(vpu, &vp8_dec_mb_width, mb_width);
+       hantro_reg_write(vpu, &vp8_dec_mb_height, mb_height);
+       hantro_reg_write(vpu, &vp8_dec_mb_width_ext, mb_width >> 9);
+       hantro_reg_write(vpu, &vp8_dec_mb_height_ext, mb_height >> 8);
+
+       /* Boolean decoder */
+       hantro_reg_write(vpu, &vp8_dec_bool_range, hdr->coder_state.range);
+       hantro_reg_write(vpu, &vp8_dec_bool_value, hdr->coder_state.value);
+
+       reg = vdpu_read(vpu, VDPU_REG_VP8_DCT_START_BIT);
+       if (hdr->version != 3)
+               reg |= VDPU_REG_DEC_CTRL4_VC1_HEIGHT_EXT;
+       if (hdr->version & 0x3)
+               reg |= VDPU_REG_DEC_CTRL4_BILIN_MC_E;
+       vdpu_write_relaxed(vpu, reg, VDPU_REG_VP8_DCT_START_BIT);
+
+       cfg_lf(ctx, hdr);
+       cfg_qp(ctx, hdr);
+       cfg_parts(ctx, hdr);
+       cfg_tap(ctx, hdr);
+       cfg_ref(ctx, hdr);
+       cfg_buffers(ctx, hdr);
+
+       hantro_end_prepare_run(ctx);
+
+       hantro_reg_write(vpu, &vp8_dec_start_dec, 1);
+
+       return 0;
+}
diff --git a/drivers/staging/media/hantro/rockchip_vpu2_regs.h b/drivers/staging/media/hantro/rockchip_vpu2_regs.h
new file mode 100644 (file)
index 0000000..49e4088
--- /dev/null
@@ -0,0 +1,600 @@
+/* SPDX-License-Identifier: GPL-2.0 */
+/*
+ * Hantro VPU codec driver
+ *
+ * Copyright (C) 2018 Rockchip Electronics Co., Ltd.
+ *     Alpha Lin <alpha.lin@rock-chips.com>
+ */
+
+#ifndef ROCKCHIP_VPU2_REGS_H_
+#define ROCKCHIP_VPU2_REGS_H_
+
+/* Encoder registers. */
+#define VEPU_REG_VP8_QUT_1ST(i)                        (0x000 + ((i) * 0x24))
+#define     VEPU_REG_VP8_QUT_DC_Y2(x)                  (((x) & 0x3fff) << 16)
+#define     VEPU_REG_VP8_QUT_DC_Y1(x)                  (((x) & 0x3fff) << 0)
+#define VEPU_REG_VP8_QUT_2ND(i)                        (0x004 + ((i) * 0x24))
+#define     VEPU_REG_VP8_QUT_AC_Y1(x)                  (((x) & 0x3fff) << 16)
+#define     VEPU_REG_VP8_QUT_DC_CHR(x)                 (((x) & 0x3fff) << 0)
+#define VEPU_REG_VP8_QUT_3RD(i)                        (0x008 + ((i) * 0x24))
+#define     VEPU_REG_VP8_QUT_AC_CHR(x)                 (((x) & 0x3fff) << 16)
+#define     VEPU_REG_VP8_QUT_AC_Y2(x)                  (((x) & 0x3fff) << 0)
+#define VEPU_REG_VP8_QUT_4TH(i)                        (0x00c + ((i) * 0x24))
+#define     VEPU_REG_VP8_QUT_ZB_DC_CHR(x)              (((x) & 0x1ff) << 18)
+#define     VEPU_REG_VP8_QUT_ZB_DC_Y2(x)               (((x) & 0x1ff) << 9)
+#define     VEPU_REG_VP8_QUT_ZB_DC_Y1(x)               (((x) & 0x1ff) << 0)
+#define VEPU_REG_VP8_QUT_5TH(i)                        (0x010 + ((i) * 0x24))
+#define     VEPU_REG_VP8_QUT_ZB_AC_CHR(x)              (((x) & 0x1ff) << 18)
+#define     VEPU_REG_VP8_QUT_ZB_AC_Y2(x)               (((x) & 0x1ff) << 9)
+#define     VEPU_REG_VP8_QUT_ZB_AC_Y1(x)               (((x) & 0x1ff) << 0)
+#define VEPU_REG_VP8_QUT_6TH(i)                        (0x014 + ((i) * 0x24))
+#define     VEPU_REG_VP8_QUT_RND_DC_CHR(x)             (((x) & 0xff) << 16)
+#define     VEPU_REG_VP8_QUT_RND_DC_Y2(x)              (((x) & 0xff) << 8)
+#define     VEPU_REG_VP8_QUT_RND_DC_Y1(x)              (((x) & 0xff) << 0)
+#define VEPU_REG_VP8_QUT_7TH(i)                        (0x018 + ((i) * 0x24))
+#define     VEPU_REG_VP8_QUT_RND_AC_CHR(x)             (((x) & 0xff) << 16)
+#define     VEPU_REG_VP8_QUT_RND_AC_Y2(x)              (((x) & 0xff) << 8)
+#define     VEPU_REG_VP8_QUT_RND_AC_Y1(x)              (((x) & 0xff) << 0)
+#define VEPU_REG_VP8_QUT_8TH(i)                        (0x01c + ((i) * 0x24))
+#define     VEPU_REG_VP8_SEG_FILTER_LEVEL(x)           (((x) & 0x3f) << 25)
+#define     VEPU_REG_VP8_DEQUT_DC_CHR(x)               (((x) & 0xff) << 17)
+#define     VEPU_REG_VP8_DEQUT_DC_Y2(x)                        (((x) & 0x1ff) << 8)
+#define     VEPU_REG_VP8_DEQUT_DC_Y1(x)                        (((x) & 0xff) << 0)
+#define VEPU_REG_VP8_QUT_9TH(i)                        (0x020 + ((i) * 0x24))
+#define     VEPU_REG_VP8_DEQUT_AC_CHR(x)               (((x) & 0x1ff) << 18)
+#define     VEPU_REG_VP8_DEQUT_AC_Y2(x)                        (((x) & 0x1ff) << 9)
+#define     VEPU_REG_VP8_DEQUT_AC_Y1(x)                        (((x) & 0x1ff) << 0)
+#define VEPU_REG_ADDR_VP8_SEG_MAP              0x06c
+#define VEPU_REG_VP8_INTRA_4X4_PENALTY(i)      (0x070 + ((i) * 0x4))
+#define     VEPU_REG_VP8_INTRA_4X4_PENALTY_0(x)                (((x) & 0xfff) << 0)
+#define     VEPU_REG_VP8_INTRA_4x4_PENALTY_1(x)                (((x) & 0xfff) << 16)
+#define VEPU_REG_VP8_INTRA_16X16_PENALTY(i)    (0x084 + ((i) * 0x4))
+#define     VEPU_REG_VP8_INTRA_16X16_PENALTY_0(x)      (((x) & 0xfff) << 0)
+#define     VEPU_REG_VP8_INTRA_16X16_PENALTY_1(x)      (((x) & 0xfff) << 16)
+#define VEPU_REG_VP8_CONTROL                   0x0a0
+#define     VEPU_REG_VP8_LF_MODE_DELTA_BPRED(x)                (((x) & 0x1f) << 24)
+#define     VEPU_REG_VP8_LF_REF_DELTA_INTRA_MB(x)      (((x) & 0x7f) << 16)
+#define     VEPU_REG_VP8_INTER_TYPE_BIT_COST(x)                (((x) & 0xfff) << 0)
+#define VEPU_REG_VP8_REF_FRAME_VAL             0x0a4
+#define     VEPU_REG_VP8_COEF_DMV_PENALTY(x)           (((x) & 0xfff) << 16)
+#define     VEPU_REG_VP8_REF_FRAME(x)                  (((x) & 0xfff) << 0)
+#define VEPU_REG_VP8_LOOP_FILTER_REF_DELTA     0x0a8
+#define     VEPU_REG_VP8_LF_REF_DELTA_ALT_REF(x)       (((x) & 0x7f) << 16)
+#define     VEPU_REG_VP8_LF_REF_DELTA_LAST_REF(x)      (((x) & 0x7f) << 8)
+#define     VEPU_REG_VP8_LF_REF_DELTA_GOLDEN(x)                (((x) & 0x7f) << 0)
+#define VEPU_REG_VP8_LOOP_FILTER_MODE_DELTA    0x0ac
+#define     VEPU_REG_VP8_LF_MODE_DELTA_SPLITMV(x)      (((x) & 0x7f) << 16)
+#define     VEPU_REG_VP8_LF_MODE_DELTA_ZEROMV(x)       (((x) & 0x7f) << 8)
+#define     VEPU_REG_VP8_LF_MODE_DELTA_NEWMV(x)                (((x) & 0x7f) << 0)
+#define        VEPU_REG_JPEG_LUMA_QUAT(i)              (0x000 + ((i) * 0x4))
+#define        VEPU_REG_JPEG_CHROMA_QUAT(i)            (0x040 + ((i) * 0x4))
+#define VEPU_REG_INTRA_SLICE_BITMAP(i)         (0x0b0 + ((i) * 0x4))
+#define VEPU_REG_ADDR_VP8_DCT_PART(i)          (0x0b0 + ((i) * 0x4))
+#define VEPU_REG_INTRA_AREA_CTRL               0x0b8
+#define     VEPU_REG_INTRA_AREA_TOP(x)                 (((x) & 0xff) << 24)
+#define     VEPU_REG_INTRA_AREA_BOTTOM(x)              (((x) & 0xff) << 16)
+#define     VEPU_REG_INTRA_AREA_LEFT(x)                        (((x) & 0xff) << 8)
+#define     VEPU_REG_INTRA_AREA_RIGHT(x)               (((x) & 0xff) << 0)
+#define VEPU_REG_CIR_INTRA_CTRL                        0x0bc
+#define     VEPU_REG_CIR_INTRA_FIRST_MB(x)             (((x) & 0xffff) << 16)
+#define     VEPU_REG_CIR_INTRA_INTERVAL(x)             (((x) & 0xffff) << 0)
+#define VEPU_REG_ADDR_IN_PLANE_0               0x0c0
+#define VEPU_REG_ADDR_IN_PLANE_1               0x0c4
+#define VEPU_REG_ADDR_IN_PLANE_2               0x0c8
+#define VEPU_REG_STR_HDR_REM_MSB               0x0cc
+#define VEPU_REG_STR_HDR_REM_LSB               0x0d0
+#define VEPU_REG_STR_BUF_LIMIT                 0x0d4
+#define VEPU_REG_AXI_CTRL                      0x0d8
+#define     VEPU_REG_AXI_CTRL_READ_ID(x)               (((x) & 0xff) << 24)
+#define     VEPU_REG_AXI_CTRL_WRITE_ID(x)              (((x) & 0xff) << 16)
+#define     VEPU_REG_AXI_CTRL_BURST_LEN(x)             (((x) & 0x3f) << 8)
+#define     VEPU_REG_AXI_CTRL_INCREMENT_MODE(x)                (((x) & 0x01) << 2)
+#define     VEPU_REG_AXI_CTRL_BIRST_DISCARD(x)         (((x) & 0x01) << 1)
+#define     VEPU_REG_AXI_CTRL_BIRST_DISABLE            BIT(0)
+#define VEPU_QP_ADJUST_MAD_DELTA_ROI           0x0dc
+#define     VEPU_REG_ROI_QP_DELTA_1                    (((x) & 0xf) << 12)
+#define     VEPU_REG_ROI_QP_DELTA_2                    (((x) & 0xf) << 8)
+#define     VEPU_REG_MAD_QP_ADJUSTMENT                 (((x) & 0xf) << 0)
+#define VEPU_REG_ADDR_REF_LUMA                 0x0e0
+#define VEPU_REG_ADDR_REF_CHROMA               0x0e4
+#define VEPU_REG_QP_SUM_DIV2                   0x0e8
+#define     VEPU_REG_QP_SUM(x)                         (((x) & 0x001fffff) * 2)
+#define VEPU_REG_ENC_CTRL0                     0x0ec
+#define     VEPU_REG_DISABLE_QUARTER_PIXEL_MV          BIT(28)
+#define     VEPU_REG_DEBLOCKING_FILTER_MODE(x)         (((x) & 0x3) << 24)
+#define     VEPU_REG_CABAC_INIT_IDC(x)                 (((x) & 0x3) << 21)
+#define     VEPU_REG_ENTROPY_CODING_MODE               BIT(20)
+#define     VEPU_REG_H264_TRANS8X8_MODE                        BIT(17)
+#define     VEPU_REG_H264_INTER4X4_MODE                        BIT(16)
+#define     VEPU_REG_H264_STREAM_MODE                  BIT(15)
+#define     VEPU_REG_H264_SLICE_SIZE(x)                        (((x) & 0x7f) << 8)
+#define VEPU_REG_ENC_OVER_FILL_STRM_OFFSET     0x0f0
+#define     VEPU_REG_STREAM_START_OFFSET(x)            (((x) & 0x3f) << 16)
+#define     VEPU_REG_SKIP_MACROBLOCK_PENALTY(x)                (((x) & 0xff) << 8)
+#define     VEPU_REG_IN_IMG_CTRL_OVRFLR_D4(x)          (((x) & 0x3) << 4)
+#define     VEPU_REG_IN_IMG_CTRL_OVRFLB(x)             (((x) & 0xf) << 0)
+#define VEPU_REG_INPUT_LUMA_INFO               0x0f4
+#define     VEPU_REG_IN_IMG_CHROMA_OFFSET(x)           (((x) & 0x7) << 20)
+#define     VEPU_REG_IN_IMG_LUMA_OFFSET(x)             (((x) & 0x7) << 16)
+#define     VEPU_REG_IN_IMG_CTRL_ROW_LEN(x)            (((x) & 0x3fff) << 0)
+#define VEPU_REG_RLC_SUM                       0x0f8
+#define     VEPU_REG_RLC_SUM_OUT(x)                    (((x) & 0x007fffff) * 4)
+#define VEPU_REG_SPLIT_PENALTY_4X4             0x0f8
+#define            VEPU_REG_VP8_SPLIT_PENALTY_4X4              (((x) & 0x1ff) << 19)
+#define VEPU_REG_ADDR_REC_LUMA                 0x0fc
+#define VEPU_REG_ADDR_REC_CHROMA               0x100
+#define VEPU_REG_CHECKPOINT(i)                 (0x104 + ((i) * 0x4))
+#define     VEPU_REG_CHECKPOINT_CHECK0(x)              (((x) & 0xffff))
+#define     VEPU_REG_CHECKPOINT_CHECK1(x)              (((x) & 0xffff) << 16)
+#define     VEPU_REG_CHECKPOINT_RESULT(x) \
+               ((((x) >> (16 - 16 * ((i) & 1))) & 0xffff) * 32)
+#define VEPU_REG_VP8_SEG0_QUANT_AC_Y1          0x104
+#define     VEPU_REG_VP8_SEG0_RND_AC_Y1(x)             (((x) & 0xff) << 23)
+#define     VEPU_REG_VP8_SEG0_ZBIN_AC_Y1(x)            (((x) & 0x1ff) << 14)
+#define     VEPU_REG_VP8_SEG0_QUT_AC_Y1(x)             (((x) & 0x3fff) << 0)
+#define VEPU_REG_VP8_SEG0_QUANT_DC_Y2          0x108
+#define     VEPU_REG_VP8_SEG0_RND_DC_Y2(x)             (((x) & 0xff) << 23)
+#define     VEPU_REG_VP8_SEG0_ZBIN_DC_Y2(x)            (((x) & 0x1ff) << 14)
+#define     VEPU_REG_VP8_SEG0_QUT_DC_Y2(x)             (((x) & 0x3fff) << 0)
+#define VEPU_REG_VP8_SEG0_QUANT_AC_Y2          0x10c
+#define     VEPU_REG_VP8_SEG0_RND_AC_Y2(x)             (((x) & 0xff) << 23)
+#define     VEPU_REG_VP8_SEG0_ZBIN_AC_Y2(x)            (((x) & 0x1ff) << 14)
+#define     VEPU_REG_VP8_SEG0_QUT_AC_Y2(x)             (((x) & 0x3fff) << 0)
+#define VEPU_REG_VP8_SEG0_QUANT_DC_CHR         0x110
+#define     VEPU_REG_VP8_SEG0_RND_DC_CHR(x)            (((x) & 0xff) << 23)
+#define     VEPU_REG_VP8_SEG0_ZBIN_DC_CHR(x)           (((x) & 0x1ff) << 14)
+#define     VEPU_REG_VP8_SEG0_QUT_DC_CHR(x)            (((x) & 0x3fff) << 0)
+#define VEPU_REG_VP8_SEG0_QUANT_AC_CHR         0x114
+#define     VEPU_REG_VP8_SEG0_RND_AC_CHR(x)            (((x) & 0xff) << 23)
+#define     VEPU_REG_VP8_SEG0_ZBIN_AC_CHR(x)           (((x) & 0x1ff) << 14)
+#define     VEPU_REG_VP8_SEG0_QUT_AC_CHR(x)            (((x) & 0x3fff) << 0)
+#define VEPU_REG_VP8_SEG0_QUANT_DQUT           0x118
+#define     VEPU_REG_VP8_MV_REF_IDX1(x)                        (((x) & 0x03) << 26)
+#define     VEPU_REG_VP8_SEG0_DQUT_DC_Y2(x)            (((x) & 0x1ff) << 17)
+#define     VEPU_REG_VP8_SEG0_DQUT_AC_Y1(x)            (((x) & 0x1ff) << 8)
+#define     VEPU_REG_VP8_SEG0_DQUT_DC_Y1(x)            (((x) & 0xff) << 0)
+#define VEPU_REG_CHKPT_WORD_ERR(i)             (0x118 + ((i) * 0x4))
+#define     VEPU_REG_CHKPT_WORD_ERR_CHK0(x)            (((x) & 0xffff))
+#define     VEPU_REG_CHKPT_WORD_ERR_CHK1(x)            (((x) & 0xffff) << 16)
+#define VEPU_REG_VP8_SEG0_QUANT_DQUT_1         0x11c
+#define     VEPU_REG_VP8_SEGMENT_MAP_UPDATE            BIT(30)
+#define     VEPU_REG_VP8_SEGMENT_EN                    BIT(29)
+#define     VEPU_REG_VP8_MV_REF_IDX2_EN                        BIT(28)
+#define     VEPU_REG_VP8_MV_REF_IDX2(x)                        (((x) & 0x03) << 26)
+#define     VEPU_REG_VP8_SEG0_DQUT_AC_CHR(x)           (((x) & 0x1ff) << 17)
+#define     VEPU_REG_VP8_SEG0_DQUT_DC_CHR(x)           (((x) & 0xff) << 9)
+#define     VEPU_REG_VP8_SEG0_DQUT_AC_Y2(x)            (((x) & 0x1ff) << 0)
+#define VEPU_REG_VP8_BOOL_ENC_VALUE            0x120
+#define VEPU_REG_CHKPT_DELTA_QP                        0x124
+#define     VEPU_REG_CHKPT_DELTA_QP_CHK0(x)            (((x) & 0x0f) << 0)
+#define     VEPU_REG_CHKPT_DELTA_QP_CHK1(x)            (((x) & 0x0f) << 4)
+#define     VEPU_REG_CHKPT_DELTA_QP_CHK2(x)            (((x) & 0x0f) << 8)
+#define     VEPU_REG_CHKPT_DELTA_QP_CHK3(x)            (((x) & 0x0f) << 12)
+#define     VEPU_REG_CHKPT_DELTA_QP_CHK4(x)            (((x) & 0x0f) << 16)
+#define     VEPU_REG_CHKPT_DELTA_QP_CHK5(x)            (((x) & 0x0f) << 20)
+#define     VEPU_REG_CHKPT_DELTA_QP_CHK6(x)            (((x) & 0x0f) << 24)
+#define VEPU_REG_VP8_ENC_CTRL2                 0x124
+#define     VEPU_REG_VP8_ZERO_MV_PENALTY_FOR_REF2(x)   (((x) & 0xff) << 24)
+#define     VEPU_REG_VP8_FILTER_SHARPNESS(x)           (((x) & 0x07) << 21)
+#define     VEPU_REG_VP8_FILTER_LEVEL(x)               (((x) & 0x3f) << 15)
+#define     VEPU_REG_VP8_DCT_PARTITION_CNT(x)          (((x) & 0x03) << 13)
+#define     VEPU_REG_VP8_BOOL_ENC_VALUE_BITS(x)                (((x) & 0x1f) << 8)
+#define     VEPU_REG_VP8_BOOL_ENC_RANGE(x)             (((x) & 0xff) << 0)
+#define VEPU_REG_ENC_CTRL1                     0x128
+#define     VEPU_REG_MAD_THRESHOLD(x)                  (((x) & 0x3f) << 24)
+#define     VEPU_REG_COMPLETED_SLICES(x)               (((x) & 0xff) << 16)
+#define     VEPU_REG_IN_IMG_CTRL_FMT(x)                        (((x) & 0xf) << 4)
+#define     VEPU_REG_IN_IMG_ROTATE_MODE(x)             (((x) & 0x3) << 2)
+#define     VEPU_REG_SIZE_TABLE_PRESENT                        BIT(0)
+#define VEPU_REG_INTRA_INTER_MODE              0x12c
+#define     VEPU_REG_INTRA16X16_MODE(x)                        (((x) & 0xffff) << 16)
+#define     VEPU_REG_INTER_MODE(x)                     (((x) & 0xffff) << 0)
+#define VEPU_REG_ENC_CTRL2                     0x130
+#define     VEPU_REG_PPS_INIT_QP(x)                    (((x) & 0x3f) << 26)
+#define     VEPU_REG_SLICE_FILTER_ALPHA(x)             (((x) & 0xf) << 22)
+#define     VEPU_REG_SLICE_FILTER_BETA(x)              (((x) & 0xf) << 18)
+#define     VEPU_REG_CHROMA_QP_OFFSET(x)               (((x) & 0x1f) << 13)
+#define     VEPU_REG_FILTER_DISABLE                    BIT(5)
+#define     VEPU_REG_IDR_PIC_ID(x)                     (((x) & 0xf) << 1)
+#define     VEPU_REG_CONSTRAINED_INTRA_PREDICTION      BIT(0)
+#define VEPU_REG_ADDR_OUTPUT_STREAM            0x134
+#define VEPU_REG_ADDR_OUTPUT_CTRL              0x138
+#define VEPU_REG_ADDR_NEXT_PIC                 0x13c
+#define VEPU_REG_ADDR_MV_OUT                   0x140
+#define VEPU_REG_ADDR_CABAC_TBL                        0x144
+#define VEPU_REG_ROI1                          0x148
+#define     VEPU_REG_ROI1_TOP_MB(x)                    (((x) & 0xff) << 24)
+#define     VEPU_REG_ROI1_BOTTOM_MB(x)                 (((x) & 0xff) << 16)
+#define     VEPU_REG_ROI1_LEFT_MB(x)                   (((x) & 0xff) << 8)
+#define     VEPU_REG_ROI1_RIGHT_MB(x)                  (((x) & 0xff) << 0)
+#define VEPU_REG_ROI2                          0x14c
+#define     VEPU_REG_ROI2_TOP_MB(x)                    (((x) & 0xff) << 24)
+#define     VEPU_REG_ROI2_BOTTOM_MB(x)                 (((x) & 0xff) << 16)
+#define     VEPU_REG_ROI2_LEFT_MB(x)                   (((x) & 0xff) << 8)
+#define     VEPU_REG_ROI2_RIGHT_MB(x)                  (((x) & 0xff) << 0)
+#define VEPU_REG_STABLE_MATRIX(i)              (0x150 + ((i) * 0x4))
+#define VEPU_REG_STABLE_MOTION_SUM             0x174
+#define VEPU_REG_STABILIZATION_OUTPUT          0x178
+#define     VEPU_REG_STABLE_MIN_VALUE(x)               (((x) & 0xffffff) << 8)
+#define     VEPU_REG_STABLE_MODE_SEL(x)                        (((x) & 0x3) << 6)
+#define     VEPU_REG_STABLE_HOR_GMV(x)                 (((x) & 0x3f) << 0)
+#define VEPU_REG_RGB2YUV_CONVERSION_COEF1      0x17c
+#define     VEPU_REG_RGB2YUV_CONVERSION_COEFB(x)       (((x) & 0xffff) << 16)
+#define     VEPU_REG_RGB2YUV_CONVERSION_COEFA(x)       (((x) & 0xffff) << 0)
+#define VEPU_REG_RGB2YUV_CONVERSION_COEF2      0x180
+#define     VEPU_REG_RGB2YUV_CONVERSION_COEFE(x)       (((x) & 0xffff) << 16)
+#define     VEPU_REG_RGB2YUV_CONVERSION_COEFC(x)       (((x) & 0xffff) << 0)
+#define VEPU_REG_RGB2YUV_CONVERSION_COEF3      0x184
+#define     VEPU_REG_RGB2YUV_CONVERSION_COEFF(x)       (((x) & 0xffff) << 0)
+#define VEPU_REG_RGB_MASK_MSB                  0x188
+#define     VEPU_REG_RGB_MASK_B_MSB(x)                 (((x) & 0x1f) << 16)
+#define     VEPU_REG_RGB_MASK_G_MSB(x)                 (((x) & 0x1f) << 8)
+#define     VEPU_REG_RGB_MASK_R_MSB(x)                 (((x) & 0x1f) << 0)
+#define VEPU_REG_MV_PENALTY                    0x18c
+#define     VEPU_REG_1MV_PENALTY(x)                    (((x) & 0x3ff) << 21)
+#define     VEPU_REG_QMV_PENALTY(x)                    (((x) & 0x3ff) << 11)
+#define     VEPU_REG_4MV_PENALTY(x)                    (((x) & 0x3ff) << 1)
+#define     VEPU_REG_SPLIT_MV_MODE_EN                  BIT(0)
+#define VEPU_REG_QP_VAL                                0x190
+#define     VEPU_REG_H264_LUMA_INIT_QP(x)              (((x) & 0x3f) << 26)
+#define     VEPU_REG_H264_QP_MAX(x)                    (((x) & 0x3f) << 20)
+#define     VEPU_REG_H264_QP_MIN(x)                    (((x) & 0x3f) << 14)
+#define     VEPU_REG_H264_CHKPT_DISTANCE(x)            (((x) & 0xfff) << 0)
+#define VEPU_REG_VP8_SEG0_QUANT_DC_Y1          0x190
+#define     VEPU_REG_VP8_SEG0_RND_DC_Y1(x)             (((x) & 0xff) << 23)
+#define     VEPU_REG_VP8_SEG0_ZBIN_DC_Y1(x)            (((x) & 0x1ff) << 14)
+#define     VEPU_REG_VP8_SEG0_QUT_DC_Y1(x)             (((x) & 0x3fff) << 0)
+#define VEPU_REG_MVC_RELATE                    0x198
+#define     VEPU_REG_ZERO_MV_FAVOR_D2(x)               (((x) & 0xf) << 20)
+#define     VEPU_REG_PENALTY_4X4MV(x)                  (((x) & 0x1ff) << 11)
+#define     VEPU_REG_MVC_VIEW_ID(x)                    (((x) & 0x7) << 8)
+#define     VEPU_REG_MVC_ANCHOR_PIC_FLAG               BIT(7)
+#define     VEPU_REG_MVC_PRIORITY_ID(x)                        (((x) & 0x7) << 4)
+#define     VEPU_REG_MVC_TEMPORAL_ID(x)                        (((x) & 0x7) << 1)
+#define     VEPU_REG_MVC_INTER_VIEW_FLAG               BIT(0)
+#define VEPU_REG_ENCODE_START                  0x19c
+#define     VEPU_REG_MB_HEIGHT(x)                      (((x) & 0x1ff) << 20)
+#define     VEPU_REG_MB_WIDTH(x)                       (((x) & 0x1ff) << 8)
+#define     VEPU_REG_FRAME_TYPE_INTER                  (0x0 << 6)
+#define     VEPU_REG_FRAME_TYPE_INTRA                  (0x1 << 6)
+#define     VEPU_REG_FRAME_TYPE_MVCINTER               (0x2 << 6)
+#define     VEPU_REG_ENCODE_FORMAT_JPEG                        (0x2 << 4)
+#define     VEPU_REG_ENCODE_FORMAT_H264                        (0x3 << 4)
+#define     VEPU_REG_ENCODE_ENABLE                     BIT(0)
+#define VEPU_REG_MB_CTRL                       0x1a0
+#define     VEPU_REG_MB_CNT_OUT(x)                     (((x) & 0xffff) << 16)
+#define     VEPU_REG_MB_CNT_SET(x)                     (((x) & 0xffff) << 0)
+#define VEPU_REG_DATA_ENDIAN                   0x1a4
+#define     VEPU_REG_INPUT_SWAP8                       BIT(31)
+#define     VEPU_REG_INPUT_SWAP16                      BIT(30)
+#define     VEPU_REG_INPUT_SWAP32                      BIT(29)
+#define     VEPU_REG_OUTPUT_SWAP8                      BIT(28)
+#define     VEPU_REG_OUTPUT_SWAP16                     BIT(27)
+#define     VEPU_REG_OUTPUT_SWAP32                     BIT(26)
+#define     VEPU_REG_TEST_IRQ                          BIT(24)
+#define     VEPU_REG_TEST_COUNTER(x)                   (((x) & 0xf) << 20)
+#define     VEPU_REG_TEST_REG                          BIT(19)
+#define     VEPU_REG_TEST_MEMORY                       BIT(18)
+#define     VEPU_REG_TEST_LEN(x)                       (((x) & 0x3ffff) << 0)
+#define VEPU_REG_ENC_CTRL3                     0x1a8
+#define     VEPU_REG_PPS_ID(x)                         (((x) & 0xff) << 24)
+#define     VEPU_REG_INTRA_PRED_MODE(x)                        (((x) & 0xff) << 16)
+#define     VEPU_REG_FRAME_NUM(x)                      (((x) & 0xffff) << 0)
+#define VEPU_REG_ENC_CTRL4                     0x1ac
+#define     VEPU_REG_MV_PENALTY_16X8_8X16(x)           (((x) & 0x3ff) << 20)
+#define     VEPU_REG_MV_PENALTY_8X8(x)                 (((x) & 0x3ff) << 10)
+#define     VEPU_REG_MV_PENALTY_8X4_4X8(x)             (((x) & 0x3ff) << 0)
+#define VEPU_REG_ADDR_VP8_PROB_CNT             0x1b0
+#define VEPU_REG_INTERRUPT                     0x1b4
+#define     VEPU_REG_INTERRUPT_NON                     BIT(28)
+#define     VEPU_REG_MV_WRITE_EN                       BIT(24)
+#define     VEPU_REG_RECON_WRITE_DIS                   BIT(20)
+#define     VEPU_REG_INTERRUPT_SLICE_READY_EN          BIT(16)
+#define     VEPU_REG_CLK_GATING_EN                     BIT(12)
+#define     VEPU_REG_INTERRUPT_TIMEOUT_EN              BIT(10)
+#define     VEPU_REG_INTERRUPT_RESET                   BIT(9)
+#define     VEPU_REG_INTERRUPT_DIS_BIT                 BIT(8)
+#define     VEPU_REG_INTERRUPT_TIMEOUT                 BIT(6)
+#define     VEPU_REG_INTERRUPT_BUFFER_FULL             BIT(5)
+#define     VEPU_REG_INTERRUPT_BUS_ERROR               BIT(4)
+#define     VEPU_REG_INTERRUPT_FUSE                    BIT(3)
+#define     VEPU_REG_INTERRUPT_SLICE_READY             BIT(2)
+#define     VEPU_REG_INTERRUPT_FRAME_READY             BIT(1)
+#define     VEPU_REG_INTERRUPT_BIT                     BIT(0)
+#define VEPU_REG_DMV_PENALTY_TBL(i)            (0x1E0 + ((i) * 0x4))
+#define     VEPU_REG_DMV_PENALTY_TABLE_BIT(x, i)        ((x) << (i) * 8)
+#define VEPU_REG_DMV_Q_PIXEL_PENALTY_TBL(i)    (0x260 + ((i) * 0x4))
+#define     VEPU_REG_DMV_Q_PIXEL_PENALTY_TABLE_BIT(x, i)       ((x) << (i) * 8)
+
+/* vpu decoder register */
+#define VDPU_REG_DEC_CTRL0                     0x0c8 // 50
+#define     VDPU_REG_REF_BUF_CTRL2_REFBU2_PICID(x)     (((x) & 0x1f) << 25)
+#define     VDPU_REG_REF_BUF_CTRL2_REFBU2_THR(x)       (((x) & 0xfff) << 13)
+#define     VDPU_REG_CONFIG_TILED_MODE_LSB             BIT(12)
+#define     VDPU_REG_CONFIG_DEC_ADV_PRE_DIS            BIT(11)
+#define     VDPU_REG_CONFIG_DEC_SCMD_DIS               BIT(10)
+#define     VDPU_REG_DEC_CTRL0_SKIP_MODE               BIT(9)
+#define     VDPU_REG_DEC_CTRL0_FILTERING_DIS           BIT(8)
+#define     VDPU_REG_DEC_CTRL0_PIC_FIXED_QUANT         BIT(7)
+#define     VDPU_REG_CONFIG_DEC_LATENCY(x)             (((x) & 0x3f) << 1)
+#define     VDPU_REG_CONFIG_TILED_MODE_MSB(x)          BIT(0)
+#define     VDPU_REG_CONFIG_DEC_OUT_TILED_E            BIT(0)
+#define VDPU_REG_STREAM_LEN                    0x0cc
+#define     VDPU_REG_DEC_CTRL3_INIT_QP(x)              (((x) & 0x3f) << 25)
+#define     VDPU_REG_DEC_STREAM_LEN_HI                 BIT(24)
+#define     VDPU_REG_DEC_CTRL3_STREAM_LEN(x)           (((x) & 0xffffff) << 0)
+#define VDPU_REG_ERROR_CONCEALMENT             0x0d0
+#define     VDPU_REG_REF_BUF_CTRL2_APF_THRESHOLD(x)    (((x) & 0x3fff) << 17)
+#define     VDPU_REG_ERR_CONC_STARTMB_X(x)             (((x) & 0x1ff) << 8)
+#define     VDPU_REG_ERR_CONC_STARTMB_Y(x)             (((x) & 0xff) << 0)
+#define VDPU_REG_DEC_FORMAT                    0x0d4
+#define     VDPU_REG_DEC_CTRL0_DEC_MODE(x)             (((x) & 0xf) << 0)
+#define VDPU_REG_DATA_ENDIAN                   0x0d8
+#define     VDPU_REG_CONFIG_DEC_STRENDIAN_E            BIT(5)
+#define     VDPU_REG_CONFIG_DEC_STRSWAP32_E            BIT(4)
+#define     VDPU_REG_CONFIG_DEC_OUTSWAP32_E            BIT(3)
+#define     VDPU_REG_CONFIG_DEC_INSWAP32_E             BIT(2)
+#define     VDPU_REG_CONFIG_DEC_OUT_ENDIAN             BIT(1)
+#define     VDPU_REG_CONFIG_DEC_IN_ENDIAN              BIT(0)
+#define VDPU_REG_INTERRUPT                     0x0dc
+#define     VDPU_REG_INTERRUPT_DEC_TIMEOUT             BIT(13)
+#define     VDPU_REG_INTERRUPT_DEC_ERROR_INT           BIT(12)
+#define     VDPU_REG_INTERRUPT_DEC_PIC_INF             BIT(10)
+#define     VDPU_REG_INTERRUPT_DEC_SLICE_INT           BIT(9)
+#define     VDPU_REG_INTERRUPT_DEC_ASO_INT             BIT(8)
+#define     VDPU_REG_INTERRUPT_DEC_BUFFER_INT          BIT(6)
+#define     VDPU_REG_INTERRUPT_DEC_BUS_INT             BIT(5)
+#define     VDPU_REG_INTERRUPT_DEC_RDY_INT             BIT(4)
+#define     VDPU_REG_INTERRUPT_DEC_IRQ_DIS             BIT(1)
+#define     VDPU_REG_INTERRUPT_DEC_IRQ                 BIT(0)
+#define VDPU_REG_AXI_CTRL                      0x0e0
+#define     VDPU_REG_AXI_DEC_SEL                       BIT(23)
+#define     VDPU_REG_CONFIG_DEC_DATA_DISC_E            BIT(22)
+#define     VDPU_REG_PARAL_BUS_E(x)                    BIT(21)
+#define     VDPU_REG_CONFIG_DEC_MAX_BURST(x)           (((x) & 0x1f) << 16)
+#define     VDPU_REG_DEC_CTRL0_DEC_AXI_WR_ID(x)                (((x) & 0xff) << 8)
+#define     VDPU_REG_CONFIG_DEC_AXI_RD_ID(x)           (((x) & 0xff) << 0)
+#define VDPU_REG_EN_FLAGS                      0x0e4
+#define     VDPU_REG_AHB_HLOCK_E                       BIT(31)
+#define     VDPU_REG_CACHE_E                           BIT(29)
+#define     VDPU_REG_PREFETCH_SINGLE_CHANNEL_E         BIT(28)
+#define     VDPU_REG_INTRA_3_CYCLE_ENHANCE             BIT(27)
+#define     VDPU_REG_INTRA_DOUBLE_SPEED                        BIT(26)
+#define     VDPU_REG_INTER_DOUBLE_SPEED                        BIT(25)
+#define     VDPU_REG_DEC_CTRL3_START_CODE_E            BIT(22)
+#define     VDPU_REG_DEC_CTRL3_CH_8PIX_ILEAV_E         BIT(21)
+#define     VDPU_REG_DEC_CTRL0_RLC_MODE_E              BIT(20)
+#define     VDPU_REG_DEC_CTRL0_DIVX3_E                 BIT(19)
+#define     VDPU_REG_DEC_CTRL0_PJPEG_E                 BIT(18)
+#define     VDPU_REG_DEC_CTRL0_PIC_INTERLACE_E         BIT(17)
+#define     VDPU_REG_DEC_CTRL0_PIC_FIELDMODE_E         BIT(16)
+#define     VDPU_REG_DEC_CTRL0_PIC_B_E                 BIT(15)
+#define     VDPU_REG_DEC_CTRL0_PIC_INTER_E             BIT(14)
+#define     VDPU_REG_DEC_CTRL0_PIC_TOPFIELD_E          BIT(13)
+#define     VDPU_REG_DEC_CTRL0_FWD_INTERLACE_E         BIT(12)
+#define     VDPU_REG_DEC_CTRL0_SORENSON_E              BIT(11)
+#define     VDPU_REG_DEC_CTRL0_WRITE_MVS_E             BIT(10)
+#define     VDPU_REG_DEC_CTRL0_REF_TOPFIELD_E          BIT(9)
+#define     VDPU_REG_DEC_CTRL0_REFTOPFIRST_E           BIT(8)
+#define     VDPU_REG_DEC_CTRL0_SEQ_MBAFF_E             BIT(7)
+#define     VDPU_REG_DEC_CTRL0_PICORD_COUNT_E          BIT(6)
+#define     VDPU_REG_CONFIG_DEC_TIMEOUT_E              BIT(5)
+#define     VDPU_REG_CONFIG_DEC_CLK_GATE_E             BIT(4)
+#define     VDPU_REG_DEC_CTRL0_DEC_OUT_DIS             BIT(2)
+#define     VDPU_REG_REF_BUF_CTRL2_REFBU2_BUF_E                BIT(1)
+#define     VDPU_REG_INTERRUPT_DEC_E                   BIT(0)
+#define VDPU_REG_SOFT_RESET                    0x0e8
+#define VDPU_REG_PRED_FLT                      0x0ec
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_0_0(x)       (((x) & 0x3ff) << 22)
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_0_1(x)       (((x) & 0x3ff) << 12)
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_0_2(x)       (((x) & 0x3ff) << 2)
+#define VDPU_REG_ADDITIONAL_CHROMA_ADDRESS     0x0f0
+#define VDPU_REG_ADDR_QTABLE                   0x0f4
+#define VDPU_REG_DIRECT_MV_ADDR                        0x0f8
+#define VDPU_REG_ADDR_DST                      0x0fc
+#define VDPU_REG_ADDR_STR                      0x100
+#define VDPU_REG_REFBUF_RELATED                        0x104
+#define VDPU_REG_FWD_PIC(i)                    (0x128 + ((i) * 0x4))
+#define     VDPU_REG_FWD_PIC_PINIT_RLIST_F5(x)         (((x) & 0x1f) << 25)
+#define     VDPU_REG_FWD_PIC_PINIT_RLIST_F4(x)         (((x) & 0x1f) << 20)
+#define     VDPU_REG_FWD_PIC_PINIT_RLIST_F3(x)         (((x) & 0x1f) << 15)
+#define     VDPU_REG_FWD_PIC_PINIT_RLIST_F2(x)         (((x) & 0x1f) << 10)
+#define     VDPU_REG_FWD_PIC_PINIT_RLIST_F1(x)         (((x) & 0x1f) << 5)
+#define     VDPU_REG_FWD_PIC_PINIT_RLIST_F0(x)         (((x) & 0x1f) << 0)
+#define VDPU_REG_REF_PIC(i)                    (0x130 + ((i) * 0x4))
+#define     VDPU_REG_REF_PIC_REFER1_NBR(x)             (((x) & 0xffff) << 16)
+#define     VDPU_REG_REF_PIC_REFER0_NBR(x)             (((x) & 0xffff) << 0)
+#define VDPU_REG_H264_ADDR_REF(i)                      (0x150 + ((i) * 0x4))
+#define     VDPU_REG_ADDR_REF_FIELD_E                  BIT(1)
+#define     VDPU_REG_ADDR_REF_TOPC_E                   BIT(0)
+#define VDPU_REG_INITIAL_REF_PIC_LIST0         0x190
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F5(x)      (((x) & 0x1f) << 25)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F4(x)      (((x) & 0x1f) << 20)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F3(x)      (((x) & 0x1f) << 15)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F2(x)      (((x) & 0x1f) << 10)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F1(x)      (((x) & 0x1f) << 5)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F0(x)      (((x) & 0x1f) << 0)
+#define VDPU_REG_INITIAL_REF_PIC_LIST1         0x194
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F11(x)     (((x) & 0x1f) << 25)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F10(x)     (((x) & 0x1f) << 20)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F9(x)      (((x) & 0x1f) << 15)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F8(x)      (((x) & 0x1f) << 10)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F7(x)      (((x) & 0x1f) << 5)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F6(x)      (((x) & 0x1f) << 0)
+#define VDPU_REG_INITIAL_REF_PIC_LIST2         0x198
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F15(x)     (((x) & 0x1f) << 15)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F14(x)     (((x) & 0x1f) << 10)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F13(x)     (((x) & 0x1f) << 5)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_F12(x)     (((x) & 0x1f) << 0)
+#define VDPU_REG_INITIAL_REF_PIC_LIST3         0x19c
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B5(x)      (((x) & 0x1f) << 25)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B4(x)      (((x) & 0x1f) << 20)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B3(x)      (((x) & 0x1f) << 15)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B2(x)      (((x) & 0x1f) << 10)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B1(x)      (((x) & 0x1f) << 5)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B0(x)      (((x) & 0x1f) << 0)
+#define VDPU_REG_INITIAL_REF_PIC_LIST4         0x1a0
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B11(x)     (((x) & 0x1f) << 25)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B10(x)     (((x) & 0x1f) << 20)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B9(x)      (((x) & 0x1f) << 15)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B8(x)      (((x) & 0x1f) << 10)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B7(x)      (((x) & 0x1f) << 5)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B6(x)      (((x) & 0x1f) << 0)
+#define VDPU_REG_INITIAL_REF_PIC_LIST5         0x1a4
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B15(x)     (((x) & 0x1f) << 15)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B14(x)     (((x) & 0x1f) << 10)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B13(x)     (((x) & 0x1f) << 5)
+#define     VDPU_REG_BD_REF_PIC_BINIT_RLIST_B12(x)     (((x) & 0x1f) << 0)
+#define VDPU_REG_INITIAL_REF_PIC_LIST6         0x1a8
+#define     VDPU_REG_BD_P_REF_PIC_PINIT_RLIST_F3(x)    (((x) & 0x1f) << 15)
+#define     VDPU_REG_BD_P_REF_PIC_PINIT_RLIST_F2(x)    (((x) & 0x1f) << 10)
+#define     VDPU_REG_BD_P_REF_PIC_PINIT_RLIST_F1(x)    (((x) & 0x1f) << 5)
+#define     VDPU_REG_BD_P_REF_PIC_PINIT_RLIST_F0(x)    (((x) & 0x1f) << 0)
+#define VDPU_REG_LT_REF                                0x1ac
+#define VDPU_REG_VALID_REF                     0x1b0
+#define VDPU_REG_H264_PIC_MB_SIZE              0x1b8
+#define     VDPU_REG_DEC_CTRL2_CH_QP_OFFSET2(x)                (((x) & 0x1f) << 22)
+#define     VDPU_REG_DEC_CTRL2_CH_QP_OFFSET(x)         (((x) & 0x1f) << 17)
+#define     VDPU_REG_DEC_CTRL1_PIC_MB_HEIGHT_P(x)      (((x) & 0xff) << 9)
+#define     VDPU_REG_DEC_CTRL1_PIC_MB_WIDTH(x)         (((x) & 0x1ff) << 0)
+#define VDPU_REG_H264_CTRL                     0x1bc
+#define     VDPU_REG_DEC_CTRL4_WEIGHT_BIPR_IDC(x)      (((x) & 0x3) << 16)
+#define     VDPU_REG_DEC_CTRL1_REF_FRAMES(x)           (((x) & 0x1f) << 0)
+#define VDPU_REG_CURRENT_FRAME                 0x1c0
+#define     VDPU_REG_DEC_CTRL5_FILT_CTRL_PRES          BIT(31)
+#define     VDPU_REG_DEC_CTRL5_RDPIC_CNT_PRES          BIT(30)
+#define     VDPU_REG_DEC_CTRL4_FRAMENUM_LEN(x)         (((x) & 0x1f) << 16)
+#define     VDPU_REG_DEC_CTRL4_FRAMENUM(x)             (((x) & 0xffff) << 0)
+#define VDPU_REG_REF_FRAME                     0x1c4
+#define     VDPU_REG_DEC_CTRL5_REFPIC_MK_LEN(x)                (((x) & 0x7ff) << 16)
+#define     VDPU_REG_DEC_CTRL5_IDR_PIC_ID(x)           (((x) & 0xffff) << 0)
+#define VDPU_REG_DEC_CTRL6                     0x1c8
+#define     VDPU_REG_DEC_CTRL6_PPS_ID(x)               (((x) & 0xff) << 24)
+#define     VDPU_REG_DEC_CTRL6_REFIDX1_ACTIVE(x)       (((x) & 0x1f) << 19)
+#define     VDPU_REG_DEC_CTRL6_REFIDX0_ACTIVE(x)       (((x) & 0x1f) << 14)
+#define     VDPU_REG_DEC_CTRL6_POC_LENGTH(x)           (((x) & 0xff) << 0)
+#define VDPU_REG_ENABLE_FLAG                   0x1cc
+#define     VDPU_REG_DEC_CTRL5_IDR_PIC_E               BIT(8)
+#define     VDPU_REG_DEC_CTRL4_DIR_8X8_INFER_E         BIT(7)
+#define     VDPU_REG_DEC_CTRL4_BLACKWHITE_E            BIT(6)
+#define     VDPU_REG_DEC_CTRL4_CABAC_E                 BIT(5)
+#define     VDPU_REG_DEC_CTRL4_WEIGHT_PRED_E           BIT(4)
+#define     VDPU_REG_DEC_CTRL5_CONST_INTRA_E           BIT(3)
+#define     VDPU_REG_DEC_CTRL5_8X8TRANS_FLAG_E         BIT(2)
+#define     VDPU_REG_DEC_CTRL2_TYPE1_QUANT_E           BIT(1)
+#define     VDPU_REG_DEC_CTRL2_FIELDPIC_FLAG_E         BIT(0)
+#define VDPU_REG_VP8_PIC_MB_SIZE               0x1e0
+#define     VDPU_REG_DEC_PIC_MB_WIDTH(x)               (((x) & 0x1ff) << 23)
+#define            VDPU_REG_DEC_MB_WIDTH_OFF(x)                (((x) & 0xf) << 19)
+#define            VDPU_REG_DEC_PIC_MB_HEIGHT_P(x)             (((x) & 0xff) << 11)
+#define     VDPU_REG_DEC_MB_HEIGHT_OFF(x)              (((x) & 0xf) << 7)
+#define     VDPU_REG_DEC_CTRL1_PIC_MB_W_EXT(x)         (((x) & 0x7) << 3)
+#define     VDPU_REG_DEC_CTRL1_PIC_MB_H_EXT(x)         (((x) & 0x7) << 0)
+#define VDPU_REG_VP8_DCT_START_BIT             0x1e4
+#define     VDPU_REG_DEC_CTRL4_DCT1_START_BIT(x)       (((x) & 0x3f) << 26)
+#define     VDPU_REG_DEC_CTRL4_DCT2_START_BIT(x)       (((x) & 0x3f) << 20)
+#define     VDPU_REG_DEC_CTRL4_VC1_HEIGHT_EXT          BIT(13)
+#define     VDPU_REG_DEC_CTRL4_BILIN_MC_E              BIT(12)
+#define VDPU_REG_VP8_CTRL0                     0x1e8
+#define     VDPU_REG_DEC_CTRL2_STRM_START_BIT(x)       (((x) & 0x3f) << 26)
+#define     VDPU_REG_DEC_CTRL2_STRM1_START_BIT(x)      (((x) & 0x3f) << 18)
+#define     VDPU_REG_DEC_CTRL2_BOOLEAN_VALUE(x)                (((x) & 0xff) << 8)
+#define     VDPU_REG_DEC_CTRL2_BOOLEAN_RANGE(x)                (((x) & 0xff) << 0)
+#define VDPU_REG_VP8_DATA_VAL                  0x1f0
+#define     VDPU_REG_DEC_CTRL6_COEFFS_PART_AM(x)       (((x) & 0xf) << 24)
+#define     VDPU_REG_DEC_CTRL6_STREAM1_LEN(x)          (((x) & 0xffffff) << 0)
+#define VDPU_REG_PRED_FLT7                     0x1f4
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_5_1(x)       (((x) & 0x3ff) << 22)
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_5_2(x)       (((x) & 0x3ff) << 12)
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_5_3(x)       (((x) & 0x3ff) << 2)
+#define VDPU_REG_PRED_FLT8                     0x1f8
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_6_0(x)       (((x) & 0x3ff) << 22)
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_6_1(x)       (((x) & 0x3ff) << 12)
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_6_2(x)       (((x) & 0x3ff) << 2)
+#define VDPU_REG_PRED_FLT9                     0x1fc
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_6_3(x)       (((x) & 0x3ff) << 22)
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_7_0(x)       (((x) & 0x3ff) << 12)
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_7_1(x)       (((x) & 0x3ff) << 2)
+#define VDPU_REG_PRED_FLT10                    0x200
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_7_2(x)       (((x) & 0x3ff) << 22)
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_7_3(x)       (((x) & 0x3ff) << 12)
+#define     VDPU_REG_BD_REF_PIC_PRED_TAP_2_M1(x)       (((x) & 0x3) << 10)
+#define     VDPU_REG_BD_REF_PIC_PRED_TAP_2_4(x)                (((x) & 0x3) << 8)
+#define     VDPU_REG_BD_REF_PIC_PRED_TAP_4_M1(x)       (((x) & 0x3) << 6)
+#define     VDPU_REG_BD_REF_PIC_PRED_TAP_4_4(x)                (((x) & 0x3) << 4)
+#define     VDPU_REG_BD_REF_PIC_PRED_TAP_6_M1(x)       (((x) & 0x3) << 2)
+#define     VDPU_REG_BD_REF_PIC_PRED_TAP_6_4(x)                (((x) & 0x3) << 0)
+#define VDPU_REG_FILTER_LEVEL                  0x204
+#define     VDPU_REG_REF_PIC_LF_LEVEL_0(x)             (((x) & 0x3f) << 18)
+#define     VDPU_REG_REF_PIC_LF_LEVEL_1(x)             (((x) & 0x3f) << 12)
+#define     VDPU_REG_REF_PIC_LF_LEVEL_2(x)             (((x) & 0x3f) << 6)
+#define     VDPU_REG_REF_PIC_LF_LEVEL_3(x)             (((x) & 0x3f) << 0)
+#define VDPU_REG_VP8_QUANTER0                  0x208
+#define     VDPU_REG_REF_PIC_QUANT_DELTA_0(x)          (((x) & 0x1f) << 27)
+#define     VDPU_REG_REF_PIC_QUANT_DELTA_1(x)          (((x) & 0x1f) << 22)
+#define     VDPU_REG_REF_PIC_QUANT_0(x)                        (((x) & 0x7ff) << 11)
+#define     VDPU_REG_REF_PIC_QUANT_1(x)                        (((x) & 0x7ff) << 0)
+#define VDPU_REG_VP8_ADDR_REF0                 0x20c
+#define VDPU_REG_FILTER_MB_ADJ                 0x210
+#define     VDPU_REG_REF_PIC_FILT_TYPE_E               BIT(31)
+#define     VDPU_REG_REF_PIC_FILT_SHARPNESS(x)         (((x) & 0x7) << 28)
+#define     VDPU_REG_FILT_MB_ADJ_0(x)                  (((x) & 0x7f) << 21)
+#define     VDPU_REG_FILT_MB_ADJ_1(x)                  (((x) & 0x7f) << 14)
+#define     VDPU_REG_FILT_MB_ADJ_2(x)                  (((x) & 0x7f) << 7)
+#define     VDPU_REG_FILT_MB_ADJ_3(x)                  (((x) & 0x7f) << 0)
+#define VDPU_REG_FILTER_REF_ADJ                        0x214
+#define     VDPU_REG_REF_PIC_ADJ_0(x)                  (((x) & 0x7f) << 21)
+#define     VDPU_REG_REF_PIC_ADJ_1(x)                  (((x) & 0x7f) << 14)
+#define     VDPU_REG_REF_PIC_ADJ_2(x)                  (((x) & 0x7f) << 7)
+#define     VDPU_REG_REF_PIC_ADJ_3(x)                  (((x) & 0x7f) << 0)
+#define VDPU_REG_VP8_ADDR_REF2_5(i)            (0x218 + ((i) * 0x4))
+#define     VDPU_REG_VP8_GREF_SIGN_BIAS                        BIT(0)
+#define     VDPU_REG_VP8_AREF_SIGN_BIAS                        BIT(0)
+#define VDPU_REG_VP8_DCT_BASE(i)               (0x230 + ((i) * 0x4))
+#define VDPU_REG_VP8_ADDR_CTRL_PART            0x244
+#define VDPU_REG_VP8_ADDR_REF1                 0x250
+#define VDPU_REG_VP8_SEGMENT_VAL               0x254
+#define     VDPU_REG_FWD_PIC1_SEGMENT_BASE(x)          ((x) << 0)
+#define     VDPU_REG_FWD_PIC1_SEGMENT_UPD_E            BIT(1)
+#define     VDPU_REG_FWD_PIC1_SEGMENT_E                        BIT(0)
+#define VDPU_REG_VP8_DCT_START_BIT2            0x258
+#define     VDPU_REG_DEC_CTRL7_DCT3_START_BIT(x)       (((x) & 0x3f) << 24)
+#define     VDPU_REG_DEC_CTRL7_DCT4_START_BIT(x)       (((x) & 0x3f) << 18)
+#define     VDPU_REG_DEC_CTRL7_DCT5_START_BIT(x)       (((x) & 0x3f) << 12)
+#define     VDPU_REG_DEC_CTRL7_DCT6_START_BIT(x)       (((x) & 0x3f) << 6)
+#define     VDPU_REG_DEC_CTRL7_DCT7_START_BIT(x)       (((x) & 0x3f) << 0)
+#define VDPU_REG_VP8_QUANTER1                  0x25c
+#define     VDPU_REG_REF_PIC_QUANT_DELTA_2(x)          (((x) & 0x1f) << 27)
+#define     VDPU_REG_REF_PIC_QUANT_DELTA_3(x)          (((x) & 0x1f) << 22)
+#define     VDPU_REG_REF_PIC_QUANT_2(x)                        (((x) & 0x7ff) << 11)
+#define     VDPU_REG_REF_PIC_QUANT_3(x)                        (((x) & 0x7ff) << 0)
+#define VDPU_REG_VP8_QUANTER2                  0x260
+#define     VDPU_REG_REF_PIC_QUANT_DELTA_4(x)          (((x) & 0x1f) << 27)
+#define     VDPU_REG_REF_PIC_QUANT_4(x)                        (((x) & 0x7ff) << 11)
+#define     VDPU_REG_REF_PIC_QUANT_5(x)                        (((x) & 0x7ff) << 0)
+#define VDPU_REG_PRED_FLT1                     0x264
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_0_3(x)       (((x) & 0x3ff) << 22)
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_1_0(x)       (((x) & 0x3ff) << 12)
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_1_1(x)       (((x) & 0x3ff) << 2)
+#define VDPU_REG_PRED_FLT2                     0x268
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_1_2(x)       (((x) & 0x3ff) << 22)
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_1_3(x)       (((x) & 0x3ff) << 12)
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_2_0(x)       (((x) & 0x3ff) << 2)
+#define VDPU_REG_PRED_FLT3                     0x26c
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_2_1(x)       (((x) & 0x3ff) << 22)
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_2_2(x)       (((x) & 0x3ff) << 12)
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_2_3(x)       (((x) & 0x3ff) << 2)
+#define VDPU_REG_PRED_FLT4                     0x270
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_3_0(x)       (((x) & 0x3ff) << 22)
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_3_1(x)       (((x) & 0x3ff) << 12)
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_3_2(x)       (((x) & 0x3ff) << 2)
+#define VDPU_REG_PRED_FLT5                     0x274
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_3_3(x)       (((x) & 0x3ff) << 22)
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_4_0(x)       (((x) & 0x3ff) << 12)
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_4_1(x)       (((x) & 0x3ff) << 2)
+#define VDPU_REG_PRED_FLT6                     0x278
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_4_2(x)       (((x) & 0x3ff) << 22)
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_4_3(x)       (((x) & 0x3ff) << 12)
+#define     VDPU_REG_PRED_FLT_PRED_BC_TAP_5_0(x)       (((x) & 0x3ff) << 2)
+
+#endif /* ROCKCHIP_VPU2_REGS_H_ */
diff --git a/drivers/staging/media/hantro/rockchip_vpu_hw.c b/drivers/staging/media/hantro/rockchip_vpu_hw.c
new file mode 100644 (file)
index 0000000..bf760e8
--- /dev/null
@@ -0,0 +1,356 @@
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * Hantro VPU codec driver
+ *
+ * Copyright (C) 2018 Rockchip Electronics Co., Ltd.
+ *     Jeffy Chen <jeffy.chen@rock-chips.com>
+ */
+
+#include <linux/clk.h>
+
+#include "hantro.h"
+#include "hantro_jpeg.h"
+#include "hantro_h1_regs.h"
+#include "rockchip_vpu2_regs.h"
+
+#define RK3288_ACLK_MAX_FREQ (400 * 1000 * 1000)
+
+/*
+ * Supported formats.
+ */
+
+static const struct hantro_fmt rockchip_vpu_enc_fmts[] = {
+       {
+               .fourcc = V4L2_PIX_FMT_YUV420M,
+               .codec_mode = HANTRO_MODE_NONE,
+               .enc_fmt = ROCKCHIP_VPU_ENC_FMT_YUV420P,
+       },
+       {
+               .fourcc = V4L2_PIX_FMT_NV12M,
+               .codec_mode = HANTRO_MODE_NONE,
+               .enc_fmt = ROCKCHIP_VPU_ENC_FMT_YUV420SP,
+       },
+       {
+               .fourcc = V4L2_PIX_FMT_YUYV,
+               .codec_mode = HANTRO_MODE_NONE,
+               .enc_fmt = ROCKCHIP_VPU_ENC_FMT_YUYV422,
+       },
+       {
+               .fourcc = V4L2_PIX_FMT_UYVY,
+               .codec_mode = HANTRO_MODE_NONE,
+               .enc_fmt = ROCKCHIP_VPU_ENC_FMT_UYVY422,
+       },
+       {
+               .fourcc = V4L2_PIX_FMT_JPEG,
+               .codec_mode = HANTRO_MODE_JPEG_ENC,
+               .max_depth = 2,
+               .header_size = JPEG_HEADER_SIZE,
+               .frmsize = {
+                       .min_width = 96,
+                       .max_width = 8192,
+                       .step_width = MB_DIM,
+                       .min_height = 32,
+                       .max_height = 8192,
+                       .step_height = MB_DIM,
+               },
+       },
+};
+
+static const struct hantro_fmt rockchip_vpu1_postproc_fmts[] = {
+       {
+               .fourcc = V4L2_PIX_FMT_YUYV,
+               .codec_mode = HANTRO_MODE_NONE,
+       },
+};
+
+static const struct hantro_fmt rk3288_vpu_dec_fmts[] = {
+       {
+               .fourcc = V4L2_PIX_FMT_NV12,
+               .codec_mode = HANTRO_MODE_NONE,
+       },
+       {
+               .fourcc = V4L2_PIX_FMT_H264_SLICE,
+               .codec_mode = HANTRO_MODE_H264_DEC,
+               .max_depth = 2,
+               .frmsize = {
+                       .min_width = 48,
+                       .max_width = 4096,
+                       .step_width = MB_DIM,
+                       .min_height = 48,
+                       .max_height = 2304,
+                       .step_height = MB_DIM,
+               },
+       },
+       {
+               .fourcc = V4L2_PIX_FMT_MPEG2_SLICE,
+               .codec_mode = HANTRO_MODE_MPEG2_DEC,
+               .max_depth = 2,
+               .frmsize = {
+                       .min_width = 48,
+                       .max_width = 1920,
+                       .step_width = MB_DIM,
+                       .min_height = 48,
+                       .max_height = 1088,
+                       .step_height = MB_DIM,
+               },
+       },
+       {
+               .fourcc = V4L2_PIX_FMT_VP8_FRAME,
+               .codec_mode = HANTRO_MODE_VP8_DEC,
+               .max_depth = 2,
+               .frmsize = {
+                       .min_width = 48,
+                       .max_width = 3840,
+                       .step_width = MB_DIM,
+                       .min_height = 48,
+                       .max_height = 2160,
+                       .step_height = MB_DIM,
+               },
+       },
+};
+
+static const struct hantro_fmt rk3399_vpu_dec_fmts[] = {
+       {
+               .fourcc = V4L2_PIX_FMT_NV12,
+               .codec_mode = HANTRO_MODE_NONE,
+       },
+       {
+               .fourcc = V4L2_PIX_FMT_MPEG2_SLICE,
+               .codec_mode = HANTRO_MODE_MPEG2_DEC,
+               .max_depth = 2,
+               .frmsize = {
+                       .min_width = 48,
+                       .max_width = 1920,
+                       .step_width = MB_DIM,
+                       .min_height = 48,
+                       .max_height = 1088,
+                       .step_height = MB_DIM,
+               },
+       },
+       {
+               .fourcc = V4L2_PIX_FMT_VP8_FRAME,
+               .codec_mode = HANTRO_MODE_VP8_DEC,
+               .max_depth = 2,
+               .frmsize = {
+                       .min_width = 48,
+                       .max_width = 3840,
+                       .step_width = MB_DIM,
+                       .min_height = 48,
+                       .max_height = 2160,
+                       .step_height = MB_DIM,
+               },
+       },
+};
+
+static irqreturn_t rockchip_vpu1_vepu_irq(int irq, void *dev_id)
+{
+       struct hantro_dev *vpu = dev_id;
+       enum vb2_buffer_state state;
+       u32 status;
+
+       status = vepu_read(vpu, H1_REG_INTERRUPT);
+       state = (status & H1_REG_INTERRUPT_FRAME_RDY) ?
+               VB2_BUF_STATE_DONE : VB2_BUF_STATE_ERROR;
+
+       vepu_write(vpu, 0, H1_REG_INTERRUPT);
+       vepu_write(vpu, 0, H1_REG_AXI_CTRL);
+
+       hantro_irq_done(vpu, state);
+
+       return IRQ_HANDLED;
+}
+
+static irqreturn_t rockchip_vpu2_vdpu_irq(int irq, void *dev_id)
+{
+       struct hantro_dev *vpu = dev_id;
+       enum vb2_buffer_state state;
+       u32 status;
+
+       status = vdpu_read(vpu, VDPU_REG_INTERRUPT);
+       state = (status & VDPU_REG_INTERRUPT_DEC_IRQ) ?
+               VB2_BUF_STATE_DONE : VB2_BUF_STATE_ERROR;
+
+       vdpu_write(vpu, 0, VDPU_REG_INTERRUPT);
+       vdpu_write(vpu, 0, VDPU_REG_AXI_CTRL);
+
+       hantro_irq_done(vpu, state);
+
+       return IRQ_HANDLED;
+}
+
+static irqreturn_t rockchip_vpu2_vepu_irq(int irq, void *dev_id)
+{
+       struct hantro_dev *vpu = dev_id;
+       enum vb2_buffer_state state;
+       u32 status;
+
+       status = vepu_read(vpu, VEPU_REG_INTERRUPT);
+       state = (status & VEPU_REG_INTERRUPT_FRAME_READY) ?
+               VB2_BUF_STATE_DONE : VB2_BUF_STATE_ERROR;
+
+       vepu_write(vpu, 0, VEPU_REG_INTERRUPT);
+       vepu_write(vpu, 0, VEPU_REG_AXI_CTRL);
+
+       hantro_irq_done(vpu, state);
+
+       return IRQ_HANDLED;
+}
+
+static int rockchip_vpu_hw_init(struct hantro_dev *vpu)
+{
+       /* Bump ACLK to max. possible freq. to improve performance. */
+       clk_set_rate(vpu->clocks[0].clk, RK3288_ACLK_MAX_FREQ);
+       return 0;
+}
+
+static void rockchip_vpu1_enc_reset(struct hantro_ctx *ctx)
+{
+       struct hantro_dev *vpu = ctx->dev;
+
+       vepu_write(vpu, H1_REG_INTERRUPT_DIS_BIT, H1_REG_INTERRUPT);
+       vepu_write(vpu, 0, H1_REG_ENC_CTRL);
+       vepu_write(vpu, 0, H1_REG_AXI_CTRL);
+}
+
+static void rockchip_vpu2_dec_reset(struct hantro_ctx *ctx)
+{
+       struct hantro_dev *vpu = ctx->dev;
+
+       vdpu_write(vpu, VDPU_REG_INTERRUPT_DEC_IRQ_DIS, VDPU_REG_INTERRUPT);
+       vdpu_write(vpu, 0, VDPU_REG_EN_FLAGS);
+       vdpu_write(vpu, 1, VDPU_REG_SOFT_RESET);
+}
+
+static void rockchip_vpu2_enc_reset(struct hantro_ctx *ctx)
+{
+       struct hantro_dev *vpu = ctx->dev;
+
+       vepu_write(vpu, VEPU_REG_INTERRUPT_DIS_BIT, VEPU_REG_INTERRUPT);
+       vepu_write(vpu, 0, VEPU_REG_ENCODE_START);
+       vepu_write(vpu, 0, VEPU_REG_AXI_CTRL);
+}
+
+/*
+ * Supported codec ops.
+ */
+
+static const struct hantro_codec_ops rk3288_vpu_codec_ops[] = {
+       [HANTRO_MODE_JPEG_ENC] = {
+               .run = hantro_h1_jpeg_enc_run,
+               .reset = rockchip_vpu1_enc_reset,
+               .init = hantro_jpeg_enc_init,
+               .done = hantro_jpeg_enc_done,
+               .exit = hantro_jpeg_enc_exit,
+       },
+       [HANTRO_MODE_H264_DEC] = {
+               .run = hantro_g1_h264_dec_run,
+               .reset = hantro_g1_reset,
+               .init = hantro_h264_dec_init,
+               .exit = hantro_h264_dec_exit,
+       },
+       [HANTRO_MODE_MPEG2_DEC] = {
+               .run = hantro_g1_mpeg2_dec_run,
+               .reset = hantro_g1_reset,
+               .init = hantro_mpeg2_dec_init,
+               .exit = hantro_mpeg2_dec_exit,
+       },
+       [HANTRO_MODE_VP8_DEC] = {
+               .run = hantro_g1_vp8_dec_run,
+               .reset = hantro_g1_reset,
+               .init = hantro_vp8_dec_init,
+               .exit = hantro_vp8_dec_exit,
+       },
+};
+
+static const struct hantro_codec_ops rk3399_vpu_codec_ops[] = {
+       [HANTRO_MODE_JPEG_ENC] = {
+               .run = rockchip_vpu2_jpeg_enc_run,
+               .reset = rockchip_vpu2_enc_reset,
+               .init = hantro_jpeg_enc_init,
+               .exit = hantro_jpeg_enc_exit,
+       },
+       [HANTRO_MODE_MPEG2_DEC] = {
+               .run = rockchip_vpu2_mpeg2_dec_run,
+               .reset = rockchip_vpu2_dec_reset,
+               .init = hantro_mpeg2_dec_init,
+               .exit = hantro_mpeg2_dec_exit,
+       },
+       [HANTRO_MODE_VP8_DEC] = {
+               .run = rockchip_vpu2_vp8_dec_run,
+               .reset = rockchip_vpu2_dec_reset,
+               .init = hantro_vp8_dec_init,
+               .exit = hantro_vp8_dec_exit,
+       },
+};
+
+/*
+ * VPU variant.
+ */
+
+static const struct hantro_irq rockchip_vpu1_irqs[] = {
+       { "vepu", rockchip_vpu1_vepu_irq },
+       { "vdpu", hantro_g1_irq },
+};
+
+static const struct hantro_irq rockchip_vdpu2_irqs[] = {
+       { "vdpu", rockchip_vpu2_vdpu_irq },
+};
+
+static const struct hantro_irq rockchip_vpu2_irqs[] = {
+       { "vepu", rockchip_vpu2_vepu_irq },
+       { "vdpu", rockchip_vpu2_vdpu_irq },
+};
+
+static const char * const rockchip_vpu_clk_names[] = {
+       "aclk", "hclk"
+};
+
+const struct hantro_variant rk3288_vpu_variant = {
+       .enc_offset = 0x0,
+       .enc_fmts = rockchip_vpu_enc_fmts,
+       .num_enc_fmts = ARRAY_SIZE(rockchip_vpu_enc_fmts),
+       .dec_offset = 0x400,
+       .dec_fmts = rk3288_vpu_dec_fmts,
+       .num_dec_fmts = ARRAY_SIZE(rk3288_vpu_dec_fmts),
+       .postproc_fmts = rockchip_vpu1_postproc_fmts,
+       .num_postproc_fmts = ARRAY_SIZE(rockchip_vpu1_postproc_fmts),
+       .postproc_regs = &hantro_g1_postproc_regs,
+       .codec = HANTRO_JPEG_ENCODER | HANTRO_MPEG2_DECODER |
+                HANTRO_VP8_DECODER | HANTRO_H264_DECODER,
+       .codec_ops = rk3288_vpu_codec_ops,
+       .irqs = rockchip_vpu1_irqs,
+       .num_irqs = ARRAY_SIZE(rockchip_vpu1_irqs),
+       .init = rockchip_vpu_hw_init,
+       .clk_names = rockchip_vpu_clk_names,
+       .num_clocks = ARRAY_SIZE(rockchip_vpu_clk_names)
+};
+
+const struct hantro_variant rk3328_vpu_variant = {
+       .dec_offset = 0x400,
+       .dec_fmts = rk3399_vpu_dec_fmts,
+       .num_dec_fmts = ARRAY_SIZE(rk3399_vpu_dec_fmts),
+       .codec = HANTRO_MPEG2_DECODER | HANTRO_VP8_DECODER,
+       .codec_ops = rk3399_vpu_codec_ops,
+       .irqs = rockchip_vdpu2_irqs,
+       .num_irqs = ARRAY_SIZE(rockchip_vdpu2_irqs),
+       .init = rockchip_vpu_hw_init,
+       .clk_names = rockchip_vpu_clk_names,
+       .num_clocks = ARRAY_SIZE(rockchip_vpu_clk_names),
+};
+
+const struct hantro_variant rk3399_vpu_variant = {
+       .enc_offset = 0x0,
+       .enc_fmts = rockchip_vpu_enc_fmts,
+       .num_enc_fmts = ARRAY_SIZE(rockchip_vpu_enc_fmts),
+       .dec_offset = 0x400,
+       .dec_fmts = rk3399_vpu_dec_fmts,
+       .num_dec_fmts = ARRAY_SIZE(rk3399_vpu_dec_fmts),
+       .codec = HANTRO_JPEG_ENCODER | HANTRO_MPEG2_DECODER |
+                HANTRO_VP8_DECODER,
+       .codec_ops = rk3399_vpu_codec_ops,
+       .irqs = rockchip_vpu2_irqs,
+       .num_irqs = ARRAY_SIZE(rockchip_vpu2_irqs),
+       .init = rockchip_vpu_hw_init,
+       .clk_names = rockchip_vpu_clk_names,
+       .num_clocks = ARRAY_SIZE(rockchip_vpu_clk_names)
+};