cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

rkvdec-vp9.c (33033B)


      1// SPDX-License-Identifier: GPL-2.0
      2/*
      3 * Rockchip Video Decoder VP9 backend
      4 *
      5 * Copyright (C) 2019 Collabora, Ltd.
      6 *	Boris Brezillon <boris.brezillon@collabora.com>
      7 * Copyright (C) 2021 Collabora, Ltd.
      8 *	Andrzej Pietrasiewicz <andrzej.p@collabora.com>
      9 *
     10 * Copyright (C) 2016 Rockchip Electronics Co., Ltd.
     11 *	Alpha Lin <Alpha.Lin@rock-chips.com>
     12 */
     13
     14/*
     15 * For following the vp9 spec please start reading this driver
     16 * code from rkvdec_vp9_run() followed by rkvdec_vp9_done().
     17 */
     18
     19#include <linux/kernel.h>
     20#include <linux/vmalloc.h>
     21#include <media/v4l2-mem2mem.h>
     22#include <media/v4l2-vp9.h>
     23
     24#include "rkvdec.h"
     25#include "rkvdec-regs.h"
     26
     27#define RKVDEC_VP9_PROBE_SIZE		4864
     28#define RKVDEC_VP9_COUNT_SIZE		13232
     29#define RKVDEC_VP9_MAX_SEGMAP_SIZE	73728
     30
     31struct rkvdec_vp9_intra_mode_probs {
     32	u8 y_mode[105];
     33	u8 uv_mode[23];
     34};
     35
     36struct rkvdec_vp9_intra_only_frame_probs {
     37	u8 coef_intra[4][2][128];
     38	struct rkvdec_vp9_intra_mode_probs intra_mode[10];
     39};
     40
     41struct rkvdec_vp9_inter_frame_probs {
     42	u8 y_mode[4][9];
     43	u8 comp_mode[5];
     44	u8 comp_ref[5];
     45	u8 single_ref[5][2];
     46	u8 inter_mode[7][3];
     47	u8 interp_filter[4][2];
     48	u8 padding0[11];
     49	u8 coef[2][4][2][128];
     50	u8 uv_mode_0_2[3][9];
     51	u8 padding1[5];
     52	u8 uv_mode_3_5[3][9];
     53	u8 padding2[5];
     54	u8 uv_mode_6_8[3][9];
     55	u8 padding3[5];
     56	u8 uv_mode_9[9];
     57	u8 padding4[7];
     58	u8 padding5[16];
     59	struct {
     60		u8 joint[3];
     61		u8 sign[2];
     62		u8 classes[2][10];
     63		u8 class0_bit[2];
     64		u8 bits[2][10];
     65		u8 class0_fr[2][2][3];
     66		u8 fr[2][3];
     67		u8 class0_hp[2];
     68		u8 hp[2];
     69	} mv;
     70};
     71
     72struct rkvdec_vp9_probs {
     73	u8 partition[16][3];
     74	u8 pred[3];
     75	u8 tree[7];
     76	u8 skip[3];
     77	u8 tx32[2][3];
     78	u8 tx16[2][2];
     79	u8 tx8[2][1];
     80	u8 is_inter[4];
     81	/* 128 bit alignment */
     82	u8 padding0[3];
     83	union {
     84		struct rkvdec_vp9_inter_frame_probs inter;
     85		struct rkvdec_vp9_intra_only_frame_probs intra_only;
     86	};
     87};
     88
     89/* Data structure describing auxiliary buffer format. */
     90struct rkvdec_vp9_priv_tbl {
     91	struct rkvdec_vp9_probs probs;
     92	u8 segmap[2][RKVDEC_VP9_MAX_SEGMAP_SIZE];
     93};
     94
     95struct rkvdec_vp9_refs_counts {
     96	u32 eob[2];
     97	u32 coeff[3];
     98};
     99
    100struct rkvdec_vp9_inter_frame_symbol_counts {
    101	u32 partition[16][4];
    102	u32 skip[3][2];
    103	u32 inter[4][2];
    104	u32 tx32p[2][4];
    105	u32 tx16p[2][4];
    106	u32 tx8p[2][2];
    107	u32 y_mode[4][10];
    108	u32 uv_mode[10][10];
    109	u32 comp[5][2];
    110	u32 comp_ref[5][2];
    111	u32 single_ref[5][2][2];
    112	u32 mv_mode[7][4];
    113	u32 filter[4][3];
    114	u32 mv_joint[4];
    115	u32 sign[2][2];
    116	/* add 1 element for align */
    117	u32 classes[2][11 + 1];
    118	u32 class0[2][2];
    119	u32 bits[2][10][2];
    120	u32 class0_fp[2][2][4];
    121	u32 fp[2][4];
    122	u32 class0_hp[2][2];
    123	u32 hp[2][2];
    124	struct rkvdec_vp9_refs_counts ref_cnt[2][4][2][6][6];
    125};
    126
    127struct rkvdec_vp9_intra_frame_symbol_counts {
    128	u32 partition[4][4][4];
    129	u32 skip[3][2];
    130	u32 intra[4][2];
    131	u32 tx32p[2][4];
    132	u32 tx16p[2][4];
    133	u32 tx8p[2][2];
    134	struct rkvdec_vp9_refs_counts ref_cnt[2][4][2][6][6];
    135};
    136
    137struct rkvdec_vp9_run {
    138	struct rkvdec_run base;
    139	const struct v4l2_ctrl_vp9_frame *decode_params;
    140};
    141
    142struct rkvdec_vp9_frame_info {
    143	u32 valid : 1;
    144	u32 segmapid : 1;
    145	u32 frame_context_idx : 2;
    146	u32 reference_mode : 2;
    147	u32 tx_mode : 3;
    148	u32 interpolation_filter : 3;
    149	u32 flags;
    150	u64 timestamp;
    151	struct v4l2_vp9_segmentation seg;
    152	struct v4l2_vp9_loop_filter lf;
    153};
    154
    155struct rkvdec_vp9_ctx {
    156	struct rkvdec_aux_buf priv_tbl;
    157	struct rkvdec_aux_buf count_tbl;
    158	struct v4l2_vp9_frame_symbol_counts inter_cnts;
    159	struct v4l2_vp9_frame_symbol_counts intra_cnts;
    160	struct v4l2_vp9_frame_context probability_tables;
    161	struct v4l2_vp9_frame_context frame_context[4];
    162	struct rkvdec_vp9_frame_info cur;
    163	struct rkvdec_vp9_frame_info last;
    164};
    165
    166static void write_coeff_plane(const u8 coef[6][6][3], u8 *coeff_plane)
    167{
    168	unsigned int idx = 0, byte_count = 0;
    169	int k, m, n;
    170	u8 p;
    171
    172	for (k = 0; k < 6; k++) {
    173		for (m = 0; m < 6; m++) {
    174			for (n = 0; n < 3; n++) {
    175				p = coef[k][m][n];
    176				coeff_plane[idx++] = p;
    177				byte_count++;
    178				if (byte_count == 27) {
    179					idx += 5;
    180					byte_count = 0;
    181				}
    182			}
    183		}
    184	}
    185}
    186
    187static void init_intra_only_probs(struct rkvdec_ctx *ctx,
    188				  const struct rkvdec_vp9_run *run)
    189{
    190	struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv;
    191	struct rkvdec_vp9_priv_tbl *tbl = vp9_ctx->priv_tbl.cpu;
    192	struct rkvdec_vp9_intra_only_frame_probs *rkprobs;
    193	const struct v4l2_vp9_frame_context *probs;
    194	unsigned int i, j, k;
    195
    196	rkprobs = &tbl->probs.intra_only;
    197	probs = &vp9_ctx->probability_tables;
    198
    199	/*
    200	 * intra only 149 x 128 bits ,aligned to 152 x 128 bits coeff related
    201	 * prob 64 x 128 bits
    202	 */
    203	for (i = 0; i < ARRAY_SIZE(probs->coef); i++) {
    204		for (j = 0; j < ARRAY_SIZE(probs->coef[0]); j++)
    205			write_coeff_plane(probs->coef[i][j][0],
    206					  rkprobs->coef_intra[i][j]);
    207	}
    208
    209	/* intra mode prob  80 x 128 bits */
    210	for (i = 0; i < ARRAY_SIZE(v4l2_vp9_kf_y_mode_prob); i++) {
    211		unsigned int byte_count = 0;
    212		int idx = 0;
    213
    214		/* vp9_kf_y_mode_prob */
    215		for (j = 0; j < ARRAY_SIZE(v4l2_vp9_kf_y_mode_prob[0]); j++) {
    216			for (k = 0; k < ARRAY_SIZE(v4l2_vp9_kf_y_mode_prob[0][0]);
    217			     k++) {
    218				u8 val = v4l2_vp9_kf_y_mode_prob[i][j][k];
    219
    220				rkprobs->intra_mode[i].y_mode[idx++] = val;
    221				byte_count++;
    222				if (byte_count == 27) {
    223					byte_count = 0;
    224					idx += 5;
    225				}
    226			}
    227		}
    228
    229	}
    230
    231	for (i = 0; i < sizeof(v4l2_vp9_kf_uv_mode_prob); ++i) {
    232		const u8 *ptr = (const u8 *)v4l2_vp9_kf_uv_mode_prob;
    233
    234		rkprobs->intra_mode[i / 23].uv_mode[i % 23] = ptr[i];
    235	}
    236}
    237
    238static void init_inter_probs(struct rkvdec_ctx *ctx,
    239			     const struct rkvdec_vp9_run *run)
    240{
    241	struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv;
    242	struct rkvdec_vp9_priv_tbl *tbl = vp9_ctx->priv_tbl.cpu;
    243	struct rkvdec_vp9_inter_frame_probs *rkprobs;
    244	const struct v4l2_vp9_frame_context *probs;
    245	unsigned int i, j, k;
    246
    247	rkprobs = &tbl->probs.inter;
    248	probs = &vp9_ctx->probability_tables;
    249
    250	/*
    251	 * inter probs
    252	 * 151 x 128 bits, aligned to 152 x 128 bits
    253	 * inter only
    254	 * intra_y_mode & inter_block info 6 x 128 bits
    255	 */
    256
    257	memcpy(rkprobs->y_mode, probs->y_mode, sizeof(rkprobs->y_mode));
    258	memcpy(rkprobs->comp_mode, probs->comp_mode,
    259	       sizeof(rkprobs->comp_mode));
    260	memcpy(rkprobs->comp_ref, probs->comp_ref,
    261	       sizeof(rkprobs->comp_ref));
    262	memcpy(rkprobs->single_ref, probs->single_ref,
    263	       sizeof(rkprobs->single_ref));
    264	memcpy(rkprobs->inter_mode, probs->inter_mode,
    265	       sizeof(rkprobs->inter_mode));
    266	memcpy(rkprobs->interp_filter, probs->interp_filter,
    267	       sizeof(rkprobs->interp_filter));
    268
    269	/* 128 x 128 bits coeff related */
    270	for (i = 0; i < ARRAY_SIZE(probs->coef); i++) {
    271		for (j = 0; j < ARRAY_SIZE(probs->coef[0]); j++) {
    272			for (k = 0; k < ARRAY_SIZE(probs->coef[0][0]); k++)
    273				write_coeff_plane(probs->coef[i][j][k],
    274						  rkprobs->coef[k][i][j]);
    275		}
    276	}
    277
    278	/* intra uv mode 6 x 128 */
    279	memcpy(rkprobs->uv_mode_0_2, &probs->uv_mode[0],
    280	       sizeof(rkprobs->uv_mode_0_2));
    281	memcpy(rkprobs->uv_mode_3_5, &probs->uv_mode[3],
    282	       sizeof(rkprobs->uv_mode_3_5));
    283	memcpy(rkprobs->uv_mode_6_8, &probs->uv_mode[6],
    284	       sizeof(rkprobs->uv_mode_6_8));
    285	memcpy(rkprobs->uv_mode_9, &probs->uv_mode[9],
    286	       sizeof(rkprobs->uv_mode_9));
    287
    288	/* mv related 6 x 128 */
    289	memcpy(rkprobs->mv.joint, probs->mv.joint,
    290	       sizeof(rkprobs->mv.joint));
    291	memcpy(rkprobs->mv.sign, probs->mv.sign,
    292	       sizeof(rkprobs->mv.sign));
    293	memcpy(rkprobs->mv.classes, probs->mv.classes,
    294	       sizeof(rkprobs->mv.classes));
    295	memcpy(rkprobs->mv.class0_bit, probs->mv.class0_bit,
    296	       sizeof(rkprobs->mv.class0_bit));
    297	memcpy(rkprobs->mv.bits, probs->mv.bits,
    298	       sizeof(rkprobs->mv.bits));
    299	memcpy(rkprobs->mv.class0_fr, probs->mv.class0_fr,
    300	       sizeof(rkprobs->mv.class0_fr));
    301	memcpy(rkprobs->mv.fr, probs->mv.fr,
    302	       sizeof(rkprobs->mv.fr));
    303	memcpy(rkprobs->mv.class0_hp, probs->mv.class0_hp,
    304	       sizeof(rkprobs->mv.class0_hp));
    305	memcpy(rkprobs->mv.hp, probs->mv.hp,
    306	       sizeof(rkprobs->mv.hp));
    307}
    308
    309static void init_probs(struct rkvdec_ctx *ctx,
    310		       const struct rkvdec_vp9_run *run)
    311{
    312	const struct v4l2_ctrl_vp9_frame *dec_params;
    313	struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv;
    314	struct rkvdec_vp9_priv_tbl *tbl = vp9_ctx->priv_tbl.cpu;
    315	struct rkvdec_vp9_probs *rkprobs = &tbl->probs;
    316	const struct v4l2_vp9_segmentation *seg;
    317	const struct v4l2_vp9_frame_context *probs;
    318	bool intra_only;
    319
    320	dec_params = run->decode_params;
    321	probs = &vp9_ctx->probability_tables;
    322	seg = &dec_params->seg;
    323
    324	memset(rkprobs, 0, sizeof(*rkprobs));
    325
    326	intra_only = !!(dec_params->flags &
    327			(V4L2_VP9_FRAME_FLAG_KEY_FRAME |
    328			 V4L2_VP9_FRAME_FLAG_INTRA_ONLY));
    329
    330	/* sb info  5 x 128 bit */
    331	memcpy(rkprobs->partition,
    332	       intra_only ? v4l2_vp9_kf_partition_probs : probs->partition,
    333	       sizeof(rkprobs->partition));
    334
    335	memcpy(rkprobs->pred, seg->pred_probs, sizeof(rkprobs->pred));
    336	memcpy(rkprobs->tree, seg->tree_probs, sizeof(rkprobs->tree));
    337	memcpy(rkprobs->skip, probs->skip, sizeof(rkprobs->skip));
    338	memcpy(rkprobs->tx32, probs->tx32, sizeof(rkprobs->tx32));
    339	memcpy(rkprobs->tx16, probs->tx16, sizeof(rkprobs->tx16));
    340	memcpy(rkprobs->tx8, probs->tx8, sizeof(rkprobs->tx8));
    341	memcpy(rkprobs->is_inter, probs->is_inter, sizeof(rkprobs->is_inter));
    342
    343	if (intra_only)
    344		init_intra_only_probs(ctx, run);
    345	else
    346		init_inter_probs(ctx, run);
    347}
    348
    349struct rkvdec_vp9_ref_reg {
    350	u32 reg_frm_size;
    351	u32 reg_hor_stride;
    352	u32 reg_y_stride;
    353	u32 reg_yuv_stride;
    354	u32 reg_ref_base;
    355};
    356
    357static struct rkvdec_vp9_ref_reg ref_regs[] = {
    358	{
    359		.reg_frm_size = RKVDEC_REG_VP9_FRAME_SIZE(0),
    360		.reg_hor_stride = RKVDEC_VP9_HOR_VIRSTRIDE(0),
    361		.reg_y_stride = RKVDEC_VP9_LAST_FRAME_YSTRIDE,
    362		.reg_yuv_stride = RKVDEC_VP9_LAST_FRAME_YUVSTRIDE,
    363		.reg_ref_base = RKVDEC_REG_VP9_LAST_FRAME_BASE,
    364	},
    365	{
    366		.reg_frm_size = RKVDEC_REG_VP9_FRAME_SIZE(1),
    367		.reg_hor_stride = RKVDEC_VP9_HOR_VIRSTRIDE(1),
    368		.reg_y_stride = RKVDEC_VP9_GOLDEN_FRAME_YSTRIDE,
    369		.reg_yuv_stride = 0,
    370		.reg_ref_base = RKVDEC_REG_VP9_GOLDEN_FRAME_BASE,
    371	},
    372	{
    373		.reg_frm_size = RKVDEC_REG_VP9_FRAME_SIZE(2),
    374		.reg_hor_stride = RKVDEC_VP9_HOR_VIRSTRIDE(2),
    375		.reg_y_stride = RKVDEC_VP9_ALTREF_FRAME_YSTRIDE,
    376		.reg_yuv_stride = 0,
    377		.reg_ref_base = RKVDEC_REG_VP9_ALTREF_FRAME_BASE,
    378	}
    379};
    380
    381static struct rkvdec_decoded_buffer *
    382get_ref_buf(struct rkvdec_ctx *ctx, struct vb2_v4l2_buffer *dst, u64 timestamp)
    383{
    384	struct v4l2_m2m_ctx *m2m_ctx = ctx->fh.m2m_ctx;
    385	struct vb2_queue *cap_q = &m2m_ctx->cap_q_ctx.q;
    386	int buf_idx;
    387
    388	/*
    389	 * If a ref is unused or invalid, address of current destination
    390	 * buffer is returned.
    391	 */
    392	buf_idx = vb2_find_timestamp(cap_q, timestamp, 0);
    393	if (buf_idx < 0)
    394		return vb2_to_rkvdec_decoded_buf(&dst->vb2_buf);
    395
    396	return vb2_to_rkvdec_decoded_buf(vb2_get_buffer(cap_q, buf_idx));
    397}
    398
    399static dma_addr_t get_mv_base_addr(struct rkvdec_decoded_buffer *buf)
    400{
    401	unsigned int aligned_pitch, aligned_height, yuv_len;
    402
    403	aligned_height = round_up(buf->vp9.height, 64);
    404	aligned_pitch = round_up(buf->vp9.width * buf->vp9.bit_depth, 512) / 8;
    405	yuv_len = (aligned_height * aligned_pitch * 3) / 2;
    406
    407	return vb2_dma_contig_plane_dma_addr(&buf->base.vb.vb2_buf, 0) +
    408	       yuv_len;
    409}
    410
    411static void config_ref_registers(struct rkvdec_ctx *ctx,
    412				 const struct rkvdec_vp9_run *run,
    413				 struct rkvdec_decoded_buffer *ref_buf,
    414				 struct rkvdec_vp9_ref_reg *ref_reg)
    415{
    416	unsigned int aligned_pitch, aligned_height, y_len, yuv_len;
    417	struct rkvdec_dev *rkvdec = ctx->dev;
    418
    419	aligned_height = round_up(ref_buf->vp9.height, 64);
    420	writel_relaxed(RKVDEC_VP9_FRAMEWIDTH(ref_buf->vp9.width) |
    421		       RKVDEC_VP9_FRAMEHEIGHT(ref_buf->vp9.height),
    422		       rkvdec->regs + ref_reg->reg_frm_size);
    423
    424	writel_relaxed(vb2_dma_contig_plane_dma_addr(&ref_buf->base.vb.vb2_buf, 0),
    425		       rkvdec->regs + ref_reg->reg_ref_base);
    426
    427	if (&ref_buf->base.vb == run->base.bufs.dst)
    428		return;
    429
    430	aligned_pitch = round_up(ref_buf->vp9.width * ref_buf->vp9.bit_depth, 512) / 8;
    431	y_len = aligned_height * aligned_pitch;
    432	yuv_len = (y_len * 3) / 2;
    433
    434	writel_relaxed(RKVDEC_HOR_Y_VIRSTRIDE(aligned_pitch / 16) |
    435		       RKVDEC_HOR_UV_VIRSTRIDE(aligned_pitch / 16),
    436		       rkvdec->regs + ref_reg->reg_hor_stride);
    437	writel_relaxed(RKVDEC_VP9_REF_YSTRIDE(y_len / 16),
    438		       rkvdec->regs + ref_reg->reg_y_stride);
    439
    440	if (!ref_reg->reg_yuv_stride)
    441		return;
    442
    443	writel_relaxed(RKVDEC_VP9_REF_YUVSTRIDE(yuv_len / 16),
    444		       rkvdec->regs + ref_reg->reg_yuv_stride);
    445}
    446
    447static void config_seg_registers(struct rkvdec_ctx *ctx, unsigned int segid)
    448{
    449	struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv;
    450	const struct v4l2_vp9_segmentation *seg;
    451	struct rkvdec_dev *rkvdec = ctx->dev;
    452	s16 feature_val;
    453	int feature_id;
    454	u32 val = 0;
    455
    456	seg = vp9_ctx->last.valid ? &vp9_ctx->last.seg : &vp9_ctx->cur.seg;
    457	feature_id = V4L2_VP9_SEG_LVL_ALT_Q;
    458	if (v4l2_vp9_seg_feat_enabled(seg->feature_enabled, feature_id, segid)) {
    459		feature_val = seg->feature_data[segid][feature_id];
    460		val |= RKVDEC_SEGID_FRAME_QP_DELTA_EN(1) |
    461		       RKVDEC_SEGID_FRAME_QP_DELTA(feature_val);
    462	}
    463
    464	feature_id = V4L2_VP9_SEG_LVL_ALT_L;
    465	if (v4l2_vp9_seg_feat_enabled(seg->feature_enabled, feature_id, segid)) {
    466		feature_val = seg->feature_data[segid][feature_id];
    467		val |= RKVDEC_SEGID_FRAME_LOOPFILTER_VALUE_EN(1) |
    468		       RKVDEC_SEGID_FRAME_LOOPFILTER_VALUE(feature_val);
    469	}
    470
    471	feature_id = V4L2_VP9_SEG_LVL_REF_FRAME;
    472	if (v4l2_vp9_seg_feat_enabled(seg->feature_enabled, feature_id, segid)) {
    473		feature_val = seg->feature_data[segid][feature_id];
    474		val |= RKVDEC_SEGID_REFERINFO_EN(1) |
    475		       RKVDEC_SEGID_REFERINFO(feature_val);
    476	}
    477
    478	feature_id = V4L2_VP9_SEG_LVL_SKIP;
    479	if (v4l2_vp9_seg_feat_enabled(seg->feature_enabled, feature_id, segid))
    480		val |= RKVDEC_SEGID_FRAME_SKIP_EN(1);
    481
    482	if (!segid &&
    483	    (seg->flags & V4L2_VP9_SEGMENTATION_FLAG_ABS_OR_DELTA_UPDATE))
    484		val |= RKVDEC_SEGID_ABS_DELTA(1);
    485
    486	writel_relaxed(val, rkvdec->regs + RKVDEC_VP9_SEGID_GRP(segid));
    487}
    488
    489static void update_dec_buf_info(struct rkvdec_decoded_buffer *buf,
    490				const struct v4l2_ctrl_vp9_frame *dec_params)
    491{
    492	buf->vp9.width = dec_params->frame_width_minus_1 + 1;
    493	buf->vp9.height = dec_params->frame_height_minus_1 + 1;
    494	buf->vp9.bit_depth = dec_params->bit_depth;
    495}
    496
    497static void update_ctx_cur_info(struct rkvdec_vp9_ctx *vp9_ctx,
    498				struct rkvdec_decoded_buffer *buf,
    499				const struct v4l2_ctrl_vp9_frame *dec_params)
    500{
    501	vp9_ctx->cur.valid = true;
    502	vp9_ctx->cur.reference_mode = dec_params->reference_mode;
    503	vp9_ctx->cur.interpolation_filter = dec_params->interpolation_filter;
    504	vp9_ctx->cur.flags = dec_params->flags;
    505	vp9_ctx->cur.timestamp = buf->base.vb.vb2_buf.timestamp;
    506	vp9_ctx->cur.seg = dec_params->seg;
    507	vp9_ctx->cur.lf = dec_params->lf;
    508}
    509
    510static void update_ctx_last_info(struct rkvdec_vp9_ctx *vp9_ctx)
    511{
    512	vp9_ctx->last = vp9_ctx->cur;
    513}
    514
    515static void config_registers(struct rkvdec_ctx *ctx,
    516			     const struct rkvdec_vp9_run *run)
    517{
    518	unsigned int y_len, uv_len, yuv_len, bit_depth, aligned_height, aligned_pitch, stream_len;
    519	const struct v4l2_ctrl_vp9_frame *dec_params;
    520	struct rkvdec_decoded_buffer *ref_bufs[3];
    521	struct rkvdec_decoded_buffer *dst, *last, *mv_ref;
    522	struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv;
    523	u32 val, last_frame_info = 0;
    524	const struct v4l2_vp9_segmentation *seg;
    525	struct rkvdec_dev *rkvdec = ctx->dev;
    526	dma_addr_t addr;
    527	bool intra_only;
    528	unsigned int i;
    529
    530	dec_params = run->decode_params;
    531	dst = vb2_to_rkvdec_decoded_buf(&run->base.bufs.dst->vb2_buf);
    532	ref_bufs[0] = get_ref_buf(ctx, &dst->base.vb, dec_params->last_frame_ts);
    533	ref_bufs[1] = get_ref_buf(ctx, &dst->base.vb, dec_params->golden_frame_ts);
    534	ref_bufs[2] = get_ref_buf(ctx, &dst->base.vb, dec_params->alt_frame_ts);
    535
    536	if (vp9_ctx->last.valid)
    537		last = get_ref_buf(ctx, &dst->base.vb, vp9_ctx->last.timestamp);
    538	else
    539		last = dst;
    540
    541	update_dec_buf_info(dst, dec_params);
    542	update_ctx_cur_info(vp9_ctx, dst, dec_params);
    543	seg = &dec_params->seg;
    544
    545	intra_only = !!(dec_params->flags &
    546			(V4L2_VP9_FRAME_FLAG_KEY_FRAME |
    547			 V4L2_VP9_FRAME_FLAG_INTRA_ONLY));
    548
    549	writel_relaxed(RKVDEC_MODE(RKVDEC_MODE_VP9),
    550		       rkvdec->regs + RKVDEC_REG_SYSCTRL);
    551
    552	bit_depth = dec_params->bit_depth;
    553	aligned_height = round_up(ctx->decoded_fmt.fmt.pix_mp.height, 64);
    554
    555	aligned_pitch = round_up(ctx->decoded_fmt.fmt.pix_mp.width *
    556				 bit_depth,
    557				 512) / 8;
    558	y_len = aligned_height * aligned_pitch;
    559	uv_len = y_len / 2;
    560	yuv_len = y_len + uv_len;
    561
    562	writel_relaxed(RKVDEC_Y_HOR_VIRSTRIDE(aligned_pitch / 16) |
    563		       RKVDEC_UV_HOR_VIRSTRIDE(aligned_pitch / 16),
    564		       rkvdec->regs + RKVDEC_REG_PICPAR);
    565	writel_relaxed(RKVDEC_Y_VIRSTRIDE(y_len / 16),
    566		       rkvdec->regs + RKVDEC_REG_Y_VIRSTRIDE);
    567	writel_relaxed(RKVDEC_YUV_VIRSTRIDE(yuv_len / 16),
    568		       rkvdec->regs + RKVDEC_REG_YUV_VIRSTRIDE);
    569
    570	stream_len = vb2_get_plane_payload(&run->base.bufs.src->vb2_buf, 0);
    571	writel_relaxed(RKVDEC_STRM_LEN(stream_len),
    572		       rkvdec->regs + RKVDEC_REG_STRM_LEN);
    573
    574	/*
    575	 * Reset count buffer, because decoder only output intra related syntax
    576	 * counts when decoding intra frame, but update entropy need to update
    577	 * all the probabilities.
    578	 */
    579	if (intra_only)
    580		memset(vp9_ctx->count_tbl.cpu, 0, vp9_ctx->count_tbl.size);
    581
    582	vp9_ctx->cur.segmapid = vp9_ctx->last.segmapid;
    583	if (!intra_only &&
    584	    !(dec_params->flags & V4L2_VP9_FRAME_FLAG_ERROR_RESILIENT) &&
    585	    (!(seg->flags & V4L2_VP9_SEGMENTATION_FLAG_ENABLED) ||
    586	     (seg->flags & V4L2_VP9_SEGMENTATION_FLAG_UPDATE_MAP)))
    587		vp9_ctx->cur.segmapid++;
    588
    589	for (i = 0; i < ARRAY_SIZE(ref_bufs); i++)
    590		config_ref_registers(ctx, run, ref_bufs[i], &ref_regs[i]);
    591
    592	for (i = 0; i < 8; i++)
    593		config_seg_registers(ctx, i);
    594
    595	writel_relaxed(RKVDEC_VP9_TX_MODE(vp9_ctx->cur.tx_mode) |
    596		       RKVDEC_VP9_FRAME_REF_MODE(dec_params->reference_mode),
    597		       rkvdec->regs + RKVDEC_VP9_CPRHEADER_CONFIG);
    598
    599	if (!intra_only) {
    600		const struct v4l2_vp9_loop_filter *lf;
    601		s8 delta;
    602
    603		if (vp9_ctx->last.valid)
    604			lf = &vp9_ctx->last.lf;
    605		else
    606			lf = &vp9_ctx->cur.lf;
    607
    608		val = 0;
    609		for (i = 0; i < ARRAY_SIZE(lf->ref_deltas); i++) {
    610			delta = lf->ref_deltas[i];
    611			val |= RKVDEC_REF_DELTAS_LASTFRAME(i, delta);
    612		}
    613
    614		writel_relaxed(val,
    615			       rkvdec->regs + RKVDEC_VP9_REF_DELTAS_LASTFRAME);
    616
    617		for (i = 0; i < ARRAY_SIZE(lf->mode_deltas); i++) {
    618			delta = lf->mode_deltas[i];
    619			last_frame_info |= RKVDEC_MODE_DELTAS_LASTFRAME(i,
    620									delta);
    621		}
    622	}
    623
    624	if (vp9_ctx->last.valid && !intra_only &&
    625	    vp9_ctx->last.seg.flags & V4L2_VP9_SEGMENTATION_FLAG_ENABLED)
    626		last_frame_info |= RKVDEC_SEG_EN_LASTFRAME;
    627
    628	if (vp9_ctx->last.valid &&
    629	    vp9_ctx->last.flags & V4L2_VP9_FRAME_FLAG_SHOW_FRAME)
    630		last_frame_info |= RKVDEC_LAST_SHOW_FRAME;
    631
    632	if (vp9_ctx->last.valid &&
    633	    vp9_ctx->last.flags &
    634	    (V4L2_VP9_FRAME_FLAG_KEY_FRAME | V4L2_VP9_FRAME_FLAG_INTRA_ONLY))
    635		last_frame_info |= RKVDEC_LAST_INTRA_ONLY;
    636
    637	if (vp9_ctx->last.valid &&
    638	    last->vp9.width == dst->vp9.width &&
    639	    last->vp9.height == dst->vp9.height)
    640		last_frame_info |= RKVDEC_LAST_WIDHHEIGHT_EQCUR;
    641
    642	writel_relaxed(last_frame_info,
    643		       rkvdec->regs + RKVDEC_VP9_INFO_LASTFRAME);
    644
    645	writel_relaxed(stream_len - dec_params->compressed_header_size -
    646		       dec_params->uncompressed_header_size,
    647		       rkvdec->regs + RKVDEC_VP9_LASTTILE_SIZE);
    648
    649	for (i = 0; !intra_only && i < ARRAY_SIZE(ref_bufs); i++) {
    650		unsigned int refw = ref_bufs[i]->vp9.width;
    651		unsigned int refh = ref_bufs[i]->vp9.height;
    652		u32 hscale, vscale;
    653
    654		hscale = (refw << 14) /	dst->vp9.width;
    655		vscale = (refh << 14) / dst->vp9.height;
    656		writel_relaxed(RKVDEC_VP9_REF_HOR_SCALE(hscale) |
    657			       RKVDEC_VP9_REF_VER_SCALE(vscale),
    658			       rkvdec->regs + RKVDEC_VP9_REF_SCALE(i));
    659	}
    660
    661	addr = vb2_dma_contig_plane_dma_addr(&dst->base.vb.vb2_buf, 0);
    662	writel_relaxed(addr, rkvdec->regs + RKVDEC_REG_DECOUT_BASE);
    663	addr = vb2_dma_contig_plane_dma_addr(&run->base.bufs.src->vb2_buf, 0);
    664	writel_relaxed(addr, rkvdec->regs + RKVDEC_REG_STRM_RLC_BASE);
    665	writel_relaxed(vp9_ctx->priv_tbl.dma +
    666		       offsetof(struct rkvdec_vp9_priv_tbl, probs),
    667		       rkvdec->regs + RKVDEC_REG_CABACTBL_PROB_BASE);
    668	writel_relaxed(vp9_ctx->count_tbl.dma,
    669		       rkvdec->regs + RKVDEC_REG_VP9COUNT_BASE);
    670
    671	writel_relaxed(vp9_ctx->priv_tbl.dma +
    672		       offsetof(struct rkvdec_vp9_priv_tbl, segmap) +
    673		       (RKVDEC_VP9_MAX_SEGMAP_SIZE * vp9_ctx->cur.segmapid),
    674		       rkvdec->regs + RKVDEC_REG_VP9_SEGIDCUR_BASE);
    675	writel_relaxed(vp9_ctx->priv_tbl.dma +
    676		       offsetof(struct rkvdec_vp9_priv_tbl, segmap) +
    677		       (RKVDEC_VP9_MAX_SEGMAP_SIZE * (!vp9_ctx->cur.segmapid)),
    678		       rkvdec->regs + RKVDEC_REG_VP9_SEGIDLAST_BASE);
    679
    680	if (!intra_only &&
    681	    !(dec_params->flags & V4L2_VP9_FRAME_FLAG_ERROR_RESILIENT) &&
    682	    vp9_ctx->last.valid)
    683		mv_ref = last;
    684	else
    685		mv_ref = dst;
    686
    687	writel_relaxed(get_mv_base_addr(mv_ref),
    688		       rkvdec->regs + RKVDEC_VP9_REF_COLMV_BASE);
    689
    690	writel_relaxed(ctx->decoded_fmt.fmt.pix_mp.width |
    691		       (ctx->decoded_fmt.fmt.pix_mp.height << 16),
    692		       rkvdec->regs + RKVDEC_REG_PERFORMANCE_CYCLE);
    693}
    694
    695static int validate_dec_params(struct rkvdec_ctx *ctx,
    696			       const struct v4l2_ctrl_vp9_frame *dec_params)
    697{
    698	unsigned int aligned_width, aligned_height;
    699
    700	/* We only support profile 0. */
    701	if (dec_params->profile != 0) {
    702		dev_err(ctx->dev->dev, "unsupported profile %d\n",
    703			dec_params->profile);
    704		return -EINVAL;
    705	}
    706
    707	aligned_width = round_up(dec_params->frame_width_minus_1 + 1, 64);
    708	aligned_height = round_up(dec_params->frame_height_minus_1 + 1, 64);
    709
    710	/*
    711	 * Userspace should update the capture/decoded format when the
    712	 * resolution changes.
    713	 */
    714	if (aligned_width != ctx->decoded_fmt.fmt.pix_mp.width ||
    715	    aligned_height != ctx->decoded_fmt.fmt.pix_mp.height) {
    716		dev_err(ctx->dev->dev,
    717			"unexpected bitstream resolution %dx%d\n",
    718			dec_params->frame_width_minus_1 + 1,
    719			dec_params->frame_height_minus_1 + 1);
    720		return -EINVAL;
    721	}
    722
    723	return 0;
    724}
    725
    726static int rkvdec_vp9_run_preamble(struct rkvdec_ctx *ctx,
    727				   struct rkvdec_vp9_run *run)
    728{
    729	const struct v4l2_ctrl_vp9_frame *dec_params;
    730	const struct v4l2_ctrl_vp9_compressed_hdr *prob_updates;
    731	struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv;
    732	struct v4l2_ctrl *ctrl;
    733	unsigned int fctx_idx;
    734	int ret;
    735
    736	/* v4l2-specific stuff */
    737	rkvdec_run_preamble(ctx, &run->base);
    738
    739	ctrl = v4l2_ctrl_find(&ctx->ctrl_hdl,
    740			      V4L2_CID_STATELESS_VP9_FRAME);
    741	if (WARN_ON(!ctrl))
    742		return -EINVAL;
    743	dec_params = ctrl->p_cur.p;
    744
    745	ret = validate_dec_params(ctx, dec_params);
    746	if (ret)
    747		return ret;
    748
    749	run->decode_params = dec_params;
    750
    751	ctrl = v4l2_ctrl_find(&ctx->ctrl_hdl, V4L2_CID_STATELESS_VP9_COMPRESSED_HDR);
    752	if (WARN_ON(!ctrl))
    753		return -EINVAL;
    754	prob_updates = ctrl->p_cur.p;
    755	vp9_ctx->cur.tx_mode = prob_updates->tx_mode;
    756
    757	/*
    758	 * vp9 stuff
    759	 *
    760	 * by this point the userspace has done all parts of 6.2 uncompressed_header()
    761	 * except this fragment:
    762	 * if ( FrameIsIntra || error_resilient_mode ) {
    763	 *	setup_past_independence ( )
    764	 *	if ( frame_type == KEY_FRAME || error_resilient_mode == 1 ||
    765	 *	     reset_frame_context == 3 ) {
    766	 *		for ( i = 0; i < 4; i ++ ) {
    767	 *			save_probs( i )
    768	 *		}
    769	 *	} else if ( reset_frame_context == 2 ) {
    770	 *		save_probs( frame_context_idx )
    771	 *	}
    772	 *	frame_context_idx = 0
    773	 * }
    774	 */
    775	fctx_idx = v4l2_vp9_reset_frame_ctx(dec_params, vp9_ctx->frame_context);
    776	vp9_ctx->cur.frame_context_idx = fctx_idx;
    777
    778	/* 6.1 frame(sz): load_probs() and load_probs2() */
    779	vp9_ctx->probability_tables = vp9_ctx->frame_context[fctx_idx];
    780
    781	/*
    782	 * The userspace has also performed 6.3 compressed_header(), but handling the
    783	 * probs in a special way. All probs which need updating, except MV-related,
    784	 * have been read from the bitstream and translated through inv_map_table[],
    785	 * but no 6.3.6 inv_recenter_nonneg(v, m) has been performed. The values passed
    786	 * by userspace are either translated values (there are no 0 values in
    787	 * inv_map_table[]), or zero to indicate no update. All MV-related probs which need
    788	 * updating have been read from the bitstream and (mv_prob << 1) | 1 has been
    789	 * performed. The values passed by userspace are either new values
    790	 * to replace old ones (the above mentioned shift and bitwise or never result in
    791	 * a zero) or zero to indicate no update.
    792	 * fw_update_probs() performs actual probs updates or leaves probs as-is
    793	 * for values for which a zero was passed from userspace.
    794	 */
    795	v4l2_vp9_fw_update_probs(&vp9_ctx->probability_tables, prob_updates, dec_params);
    796
    797	return 0;
    798}
    799
    800static int rkvdec_vp9_run(struct rkvdec_ctx *ctx)
    801{
    802	struct rkvdec_dev *rkvdec = ctx->dev;
    803	struct rkvdec_vp9_run run = { };
    804	int ret;
    805
    806	ret = rkvdec_vp9_run_preamble(ctx, &run);
    807	if (ret) {
    808		rkvdec_run_postamble(ctx, &run.base);
    809		return ret;
    810	}
    811
    812	/* Prepare probs. */
    813	init_probs(ctx, &run);
    814
    815	/* Configure hardware registers. */
    816	config_registers(ctx, &run);
    817
    818	rkvdec_run_postamble(ctx, &run.base);
    819
    820	schedule_delayed_work(&rkvdec->watchdog_work, msecs_to_jiffies(2000));
    821
    822	writel(1, rkvdec->regs + RKVDEC_REG_PREF_LUMA_CACHE_COMMAND);
    823	writel(1, rkvdec->regs + RKVDEC_REG_PREF_CHR_CACHE_COMMAND);
    824
    825	writel(0xe, rkvdec->regs + RKVDEC_REG_STRMD_ERR_EN);
    826	/* Start decoding! */
    827	writel(RKVDEC_INTERRUPT_DEC_E | RKVDEC_CONFIG_DEC_CLK_GATE_E |
    828	       RKVDEC_TIMEOUT_E | RKVDEC_BUF_EMPTY_E,
    829	       rkvdec->regs + RKVDEC_REG_INTERRUPT);
    830
    831	return 0;
    832}
    833
    834#define copy_tx_and_skip(p1, p2)				\
    835do {								\
    836	memcpy((p1)->tx8, (p2)->tx8, sizeof((p1)->tx8));	\
    837	memcpy((p1)->tx16, (p2)->tx16, sizeof((p1)->tx16));	\
    838	memcpy((p1)->tx32, (p2)->tx32, sizeof((p1)->tx32));	\
    839	memcpy((p1)->skip, (p2)->skip, sizeof((p1)->skip));	\
    840} while (0)
    841
    842static void rkvdec_vp9_done(struct rkvdec_ctx *ctx,
    843			    struct vb2_v4l2_buffer *src_buf,
    844			    struct vb2_v4l2_buffer *dst_buf,
    845			    enum vb2_buffer_state result)
    846{
    847	struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv;
    848	unsigned int fctx_idx;
    849
    850	/* v4l2-specific stuff */
    851	if (result == VB2_BUF_STATE_ERROR)
    852		goto out_update_last;
    853
    854	/*
    855	 * vp9 stuff
    856	 *
    857	 * 6.1.2 refresh_probs()
    858	 *
    859	 * In the spec a complementary condition goes last in 6.1.2 refresh_probs(),
    860	 * but it makes no sense to perform all the activities from the first "if"
    861	 * there if we actually are not refreshing the frame context. On top of that,
    862	 * because of 6.2 uncompressed_header() whenever error_resilient_mode == 1,
    863	 * refresh_frame_context == 0. Consequently, if we don't jump to out_update_last
    864	 * it means error_resilient_mode must be 0.
    865	 */
    866	if (!(vp9_ctx->cur.flags & V4L2_VP9_FRAME_FLAG_REFRESH_FRAME_CTX))
    867		goto out_update_last;
    868
    869	fctx_idx = vp9_ctx->cur.frame_context_idx;
    870
    871	if (!(vp9_ctx->cur.flags & V4L2_VP9_FRAME_FLAG_PARALLEL_DEC_MODE)) {
    872		/* error_resilient_mode == 0 && frame_parallel_decoding_mode == 0 */
    873		struct v4l2_vp9_frame_context *probs = &vp9_ctx->probability_tables;
    874		bool frame_is_intra = vp9_ctx->cur.flags &
    875		    (V4L2_VP9_FRAME_FLAG_KEY_FRAME | V4L2_VP9_FRAME_FLAG_INTRA_ONLY);
    876		struct tx_and_skip {
    877			u8 tx8[2][1];
    878			u8 tx16[2][2];
    879			u8 tx32[2][3];
    880			u8 skip[3];
    881		} _tx_skip, *tx_skip = &_tx_skip;
    882		struct v4l2_vp9_frame_symbol_counts *counts;
    883
    884		/* buffer the forward-updated TX and skip probs */
    885		if (frame_is_intra)
    886			copy_tx_and_skip(tx_skip, probs);
    887
    888		/* 6.1.2 refresh_probs(): load_probs() and load_probs2() */
    889		*probs = vp9_ctx->frame_context[fctx_idx];
    890
    891		/* if FrameIsIntra then undo the effect of load_probs2() */
    892		if (frame_is_intra)
    893			copy_tx_and_skip(probs, tx_skip);
    894
    895		counts = frame_is_intra ? &vp9_ctx->intra_cnts : &vp9_ctx->inter_cnts;
    896		v4l2_vp9_adapt_coef_probs(probs, counts,
    897					  !vp9_ctx->last.valid ||
    898					  vp9_ctx->last.flags & V4L2_VP9_FRAME_FLAG_KEY_FRAME,
    899					  frame_is_intra);
    900		if (!frame_is_intra) {
    901			const struct rkvdec_vp9_inter_frame_symbol_counts *inter_cnts;
    902			u32 classes[2][11];
    903			int i;
    904
    905			inter_cnts = vp9_ctx->count_tbl.cpu;
    906			for (i = 0; i < ARRAY_SIZE(classes); ++i)
    907				memcpy(classes[i], inter_cnts->classes[i], sizeof(classes[0]));
    908			counts->classes = &classes;
    909
    910			/* load_probs2() already done */
    911			v4l2_vp9_adapt_noncoef_probs(&vp9_ctx->probability_tables, counts,
    912						     vp9_ctx->cur.reference_mode,
    913						     vp9_ctx->cur.interpolation_filter,
    914						     vp9_ctx->cur.tx_mode, vp9_ctx->cur.flags);
    915		}
    916	}
    917
    918	/* 6.1.2 refresh_probs(): save_probs(fctx_idx) */
    919	vp9_ctx->frame_context[fctx_idx] = vp9_ctx->probability_tables;
    920
    921out_update_last:
    922	update_ctx_last_info(vp9_ctx);
    923}
    924
    925static void rkvdec_init_v4l2_vp9_count_tbl(struct rkvdec_ctx *ctx)
    926{
    927	struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv;
    928	struct rkvdec_vp9_intra_frame_symbol_counts *intra_cnts = vp9_ctx->count_tbl.cpu;
    929	struct rkvdec_vp9_inter_frame_symbol_counts *inter_cnts = vp9_ctx->count_tbl.cpu;
    930	int i, j, k, l, m;
    931
    932	vp9_ctx->inter_cnts.partition = &inter_cnts->partition;
    933	vp9_ctx->inter_cnts.skip = &inter_cnts->skip;
    934	vp9_ctx->inter_cnts.intra_inter = &inter_cnts->inter;
    935	vp9_ctx->inter_cnts.tx32p = &inter_cnts->tx32p;
    936	vp9_ctx->inter_cnts.tx16p = &inter_cnts->tx16p;
    937	vp9_ctx->inter_cnts.tx8p = &inter_cnts->tx8p;
    938
    939	vp9_ctx->intra_cnts.partition = (u32 (*)[16][4])(&intra_cnts->partition);
    940	vp9_ctx->intra_cnts.skip = &intra_cnts->skip;
    941	vp9_ctx->intra_cnts.intra_inter = &intra_cnts->intra;
    942	vp9_ctx->intra_cnts.tx32p = &intra_cnts->tx32p;
    943	vp9_ctx->intra_cnts.tx16p = &intra_cnts->tx16p;
    944	vp9_ctx->intra_cnts.tx8p = &intra_cnts->tx8p;
    945
    946	vp9_ctx->inter_cnts.y_mode = &inter_cnts->y_mode;
    947	vp9_ctx->inter_cnts.uv_mode = &inter_cnts->uv_mode;
    948	vp9_ctx->inter_cnts.comp = &inter_cnts->comp;
    949	vp9_ctx->inter_cnts.comp_ref = &inter_cnts->comp_ref;
    950	vp9_ctx->inter_cnts.single_ref = &inter_cnts->single_ref;
    951	vp9_ctx->inter_cnts.mv_mode = &inter_cnts->mv_mode;
    952	vp9_ctx->inter_cnts.filter = &inter_cnts->filter;
    953	vp9_ctx->inter_cnts.mv_joint = &inter_cnts->mv_joint;
    954	vp9_ctx->inter_cnts.sign = &inter_cnts->sign;
    955	/*
    956	 * rk hardware actually uses "u32 classes[2][11 + 1];"
    957	 * instead of "u32 classes[2][11];", so this must be explicitly
    958	 * copied into vp9_ctx->classes when passing the data to the
    959	 * vp9 library function
    960	 */
    961	vp9_ctx->inter_cnts.class0 = &inter_cnts->class0;
    962	vp9_ctx->inter_cnts.bits = &inter_cnts->bits;
    963	vp9_ctx->inter_cnts.class0_fp = &inter_cnts->class0_fp;
    964	vp9_ctx->inter_cnts.fp = &inter_cnts->fp;
    965	vp9_ctx->inter_cnts.class0_hp = &inter_cnts->class0_hp;
    966	vp9_ctx->inter_cnts.hp = &inter_cnts->hp;
    967
    968#define INNERMOST_LOOP \
    969	do {										\
    970		for (m = 0; m < ARRAY_SIZE(vp9_ctx->inter_cnts.coeff[0][0][0][0]); ++m) {\
    971			vp9_ctx->inter_cnts.coeff[i][j][k][l][m] =			\
    972				&inter_cnts->ref_cnt[k][i][j][l][m].coeff;		\
    973			vp9_ctx->inter_cnts.eob[i][j][k][l][m][0] =			\
    974				&inter_cnts->ref_cnt[k][i][j][l][m].eob[0];		\
    975			vp9_ctx->inter_cnts.eob[i][j][k][l][m][1] =			\
    976				&inter_cnts->ref_cnt[k][i][j][l][m].eob[1];		\
    977											\
    978			vp9_ctx->intra_cnts.coeff[i][j][k][l][m] =			\
    979				&intra_cnts->ref_cnt[k][i][j][l][m].coeff;		\
    980			vp9_ctx->intra_cnts.eob[i][j][k][l][m][0] =			\
    981				&intra_cnts->ref_cnt[k][i][j][l][m].eob[0];		\
    982			vp9_ctx->intra_cnts.eob[i][j][k][l][m][1] =			\
    983				&intra_cnts->ref_cnt[k][i][j][l][m].eob[1];		\
    984		}									\
    985	} while (0)
    986
    987	for (i = 0; i < ARRAY_SIZE(vp9_ctx->inter_cnts.coeff); ++i)
    988		for (j = 0; j < ARRAY_SIZE(vp9_ctx->inter_cnts.coeff[0]); ++j)
    989			for (k = 0; k < ARRAY_SIZE(vp9_ctx->inter_cnts.coeff[0][0]); ++k)
    990				for (l = 0; l < ARRAY_SIZE(vp9_ctx->inter_cnts.coeff[0][0][0]); ++l)
    991					INNERMOST_LOOP;
    992#undef INNERMOST_LOOP
    993}
    994
    995static int rkvdec_vp9_start(struct rkvdec_ctx *ctx)
    996{
    997	struct rkvdec_dev *rkvdec = ctx->dev;
    998	struct rkvdec_vp9_priv_tbl *priv_tbl;
    999	struct rkvdec_vp9_ctx *vp9_ctx;
   1000	unsigned char *count_tbl;
   1001	int ret;
   1002
   1003	vp9_ctx = kzalloc(sizeof(*vp9_ctx), GFP_KERNEL);
   1004	if (!vp9_ctx)
   1005		return -ENOMEM;
   1006
   1007	ctx->priv = vp9_ctx;
   1008
   1009	priv_tbl = dma_alloc_coherent(rkvdec->dev, sizeof(*priv_tbl),
   1010				      &vp9_ctx->priv_tbl.dma, GFP_KERNEL);
   1011	if (!priv_tbl) {
   1012		ret = -ENOMEM;
   1013		goto err_free_ctx;
   1014	}
   1015
   1016	vp9_ctx->priv_tbl.size = sizeof(*priv_tbl);
   1017	vp9_ctx->priv_tbl.cpu = priv_tbl;
   1018	memset(priv_tbl, 0, sizeof(*priv_tbl));
   1019
   1020	count_tbl = dma_alloc_coherent(rkvdec->dev, RKVDEC_VP9_COUNT_SIZE,
   1021				       &vp9_ctx->count_tbl.dma, GFP_KERNEL);
   1022	if (!count_tbl) {
   1023		ret = -ENOMEM;
   1024		goto err_free_priv_tbl;
   1025	}
   1026
   1027	vp9_ctx->count_tbl.size = RKVDEC_VP9_COUNT_SIZE;
   1028	vp9_ctx->count_tbl.cpu = count_tbl;
   1029	memset(count_tbl, 0, sizeof(*count_tbl));
   1030	rkvdec_init_v4l2_vp9_count_tbl(ctx);
   1031
   1032	return 0;
   1033
   1034err_free_priv_tbl:
   1035	dma_free_coherent(rkvdec->dev, vp9_ctx->priv_tbl.size,
   1036			  vp9_ctx->priv_tbl.cpu, vp9_ctx->priv_tbl.dma);
   1037
   1038err_free_ctx:
   1039	kfree(vp9_ctx);
   1040	return ret;
   1041}
   1042
   1043static void rkvdec_vp9_stop(struct rkvdec_ctx *ctx)
   1044{
   1045	struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv;
   1046	struct rkvdec_dev *rkvdec = ctx->dev;
   1047
   1048	dma_free_coherent(rkvdec->dev, vp9_ctx->count_tbl.size,
   1049			  vp9_ctx->count_tbl.cpu, vp9_ctx->count_tbl.dma);
   1050	dma_free_coherent(rkvdec->dev, vp9_ctx->priv_tbl.size,
   1051			  vp9_ctx->priv_tbl.cpu, vp9_ctx->priv_tbl.dma);
   1052	kfree(vp9_ctx);
   1053}
   1054
   1055static int rkvdec_vp9_adjust_fmt(struct rkvdec_ctx *ctx,
   1056				 struct v4l2_format *f)
   1057{
   1058	struct v4l2_pix_format_mplane *fmt = &f->fmt.pix_mp;
   1059
   1060	fmt->num_planes = 1;
   1061	if (!fmt->plane_fmt[0].sizeimage)
   1062		fmt->plane_fmt[0].sizeimage = fmt->width * fmt->height * 2;
   1063	return 0;
   1064}
   1065
   1066const struct rkvdec_coded_fmt_ops rkvdec_vp9_fmt_ops = {
   1067	.adjust_fmt = rkvdec_vp9_adjust_fmt,
   1068	.start = rkvdec_vp9_start,
   1069	.stop = rkvdec_vp9_stop,
   1070	.run = rkvdec_vp9_run,
   1071	.done = rkvdec_vp9_done,
   1072};