// SPDX-License-Identifier: GPL-2.0 /* * Rockchip Video Decoder VP9 backend * * Copyright (C) 2019 Collabora, Ltd. * Boris Brezillon * Copyright (C) 2021 Collabora, Ltd. * Andrzej Pietrasiewicz * * Copyright (C) 2016 Rockchip Electronics Co., Ltd. * Alpha Lin */ /* * For following the vp9 spec please start reading this driver * code from rkvdec_vp9_run() followed by rkvdec_vp9_done(). */ #include #include #include #include #include "rkvdec.h" #include "rkvdec-regs.h" #define RKVDEC_VP9_PROBE_SIZE 4864 #define RKVDEC_VP9_COUNT_SIZE 13232 #define RKVDEC_VP9_MAX_SEGMAP_SIZE 73728 struct rkvdec_vp9_intra_mode_probs { u8 y_mode[105]; u8 uv_mode[23]; }; struct rkvdec_vp9_intra_only_frame_probs { u8 coef_intra[4][2][128]; struct rkvdec_vp9_intra_mode_probs intra_mode[10]; }; struct rkvdec_vp9_inter_frame_probs { u8 y_mode[4][9]; u8 comp_mode[5]; u8 comp_ref[5]; u8 single_ref[5][2]; u8 inter_mode[7][3]; u8 interp_filter[4][2]; u8 padding0[11]; u8 coef[2][4][2][128]; u8 uv_mode_0_2[3][9]; u8 padding1[5]; u8 uv_mode_3_5[3][9]; u8 padding2[5]; u8 uv_mode_6_8[3][9]; u8 padding3[5]; u8 uv_mode_9[9]; u8 padding4[7]; u8 padding5[16]; struct { u8 joint[3]; u8 sign[2]; u8 classes[2][10]; u8 class0_bit[2]; u8 bits[2][10]; u8 class0_fr[2][2][3]; u8 fr[2][3]; u8 class0_hp[2]; u8 hp[2]; } mv; }; struct rkvdec_vp9_probs { u8 partition[16][3]; u8 pred[3]; u8 tree[7]; u8 skip[3]; u8 tx32[2][3]; u8 tx16[2][2]; u8 tx8[2][1]; u8 is_inter[4]; /* 128 bit alignment */ u8 padding0[3]; union { struct rkvdec_vp9_inter_frame_probs inter; struct rkvdec_vp9_intra_only_frame_probs intra_only; }; }; /* Data structure describing auxiliary buffer format. */ struct rkvdec_vp9_priv_tbl { struct rkvdec_vp9_probs probs; u8 segmap[2][RKVDEC_VP9_MAX_SEGMAP_SIZE]; }; struct rkvdec_vp9_refs_counts { u32 eob[2]; u32 coeff[3]; }; struct rkvdec_vp9_inter_frame_symbol_counts { u32 partition[16][4]; u32 skip[3][2]; u32 inter[4][2]; u32 tx32p[2][4]; u32 tx16p[2][4]; u32 tx8p[2][2]; u32 y_mode[4][10]; u32 uv_mode[10][10]; u32 comp[5][2]; u32 comp_ref[5][2]; u32 single_ref[5][2][2]; u32 mv_mode[7][4]; u32 filter[4][3]; u32 mv_joint[4]; u32 sign[2][2]; /* add 1 element for align */ u32 classes[2][11 + 1]; u32 class0[2][2]; u32 bits[2][10][2]; u32 class0_fp[2][2][4]; u32 fp[2][4]; u32 class0_hp[2][2]; u32 hp[2][2]; struct rkvdec_vp9_refs_counts ref_cnt[2][4][2][6][6]; }; struct rkvdec_vp9_intra_frame_symbol_counts { u32 partition[4][4][4]; u32 skip[3][2]; u32 intra[4][2]; u32 tx32p[2][4]; u32 tx16p[2][4]; u32 tx8p[2][2]; struct rkvdec_vp9_refs_counts ref_cnt[2][4][2][6][6]; }; struct rkvdec_vp9_run { struct rkvdec_run base; const struct v4l2_ctrl_vp9_frame *decode_params; }; struct rkvdec_vp9_frame_info { u32 valid : 1; u32 segmapid : 1; u32 frame_context_idx : 2; u32 reference_mode : 2; u32 tx_mode : 3; u32 interpolation_filter : 3; u32 flags; u64 timestamp; struct v4l2_vp9_segmentation seg; struct v4l2_vp9_loop_filter lf; }; struct rkvdec_vp9_ctx { struct rkvdec_aux_buf priv_tbl; struct rkvdec_aux_buf count_tbl; struct v4l2_vp9_frame_symbol_counts inter_cnts; struct v4l2_vp9_frame_symbol_counts intra_cnts; struct v4l2_vp9_frame_context probability_tables; struct v4l2_vp9_frame_context frame_context[4]; struct rkvdec_vp9_frame_info cur; struct rkvdec_vp9_frame_info last; }; static void write_coeff_plane(const u8 coef[6][6][3], u8 *coeff_plane) { unsigned int idx = 0, byte_count = 0; int k, m, n; u8 p; for (k = 0; k < 6; k++) { for (m = 0; m < 6; m++) { for (n = 0; n < 3; n++) { p = coef[k][m][n]; coeff_plane[idx++] = p; byte_count++; if (byte_count == 27) { idx += 5; byte_count = 0; } } } } } static void init_intra_only_probs(struct rkvdec_ctx *ctx, const struct rkvdec_vp9_run *run) { struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv; struct rkvdec_vp9_priv_tbl *tbl = vp9_ctx->priv_tbl.cpu; struct rkvdec_vp9_intra_only_frame_probs *rkprobs; const struct v4l2_vp9_frame_context *probs; unsigned int i, j, k; rkprobs = &tbl->probs.intra_only; probs = &vp9_ctx->probability_tables; /* * intra only 149 x 128 bits ,aligned to 152 x 128 bits coeff related * prob 64 x 128 bits */ for (i = 0; i < ARRAY_SIZE(probs->coef); i++) { for (j = 0; j < ARRAY_SIZE(probs->coef[0]); j++) write_coeff_plane(probs->coef[i][j][0], rkprobs->coef_intra[i][j]); } /* intra mode prob 80 x 128 bits */ for (i = 0; i < ARRAY_SIZE(v4l2_vp9_kf_y_mode_prob); i++) { unsigned int byte_count = 0; int idx = 0; /* vp9_kf_y_mode_prob */ for (j = 0; j < ARRAY_SIZE(v4l2_vp9_kf_y_mode_prob[0]); j++) { for (k = 0; k < ARRAY_SIZE(v4l2_vp9_kf_y_mode_prob[0][0]); k++) { u8 val = v4l2_vp9_kf_y_mode_prob[i][j][k]; rkprobs->intra_mode[i].y_mode[idx++] = val; byte_count++; if (byte_count == 27) { byte_count = 0; idx += 5; } } } } for (i = 0; i < sizeof(v4l2_vp9_kf_uv_mode_prob); ++i) { const u8 *ptr = (const u8 *)v4l2_vp9_kf_uv_mode_prob; rkprobs->intra_mode[i / 23].uv_mode[i % 23] = ptr[i]; } } static void init_inter_probs(struct rkvdec_ctx *ctx, const struct rkvdec_vp9_run *run) { struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv; struct rkvdec_vp9_priv_tbl *tbl = vp9_ctx->priv_tbl.cpu; struct rkvdec_vp9_inter_frame_probs *rkprobs; const struct v4l2_vp9_frame_context *probs; unsigned int i, j, k; rkprobs = &tbl->probs.inter; probs = &vp9_ctx->probability_tables; /* * inter probs * 151 x 128 bits, aligned to 152 x 128 bits * inter only * intra_y_mode & inter_block info 6 x 128 bits */ memcpy(rkprobs->y_mode, probs->y_mode, sizeof(rkprobs->y_mode)); memcpy(rkprobs->comp_mode, probs->comp_mode, sizeof(rkprobs->comp_mode)); memcpy(rkprobs->comp_ref, probs->comp_ref, sizeof(rkprobs->comp_ref)); memcpy(rkprobs->single_ref, probs->single_ref, sizeof(rkprobs->single_ref)); memcpy(rkprobs->inter_mode, probs->inter_mode, sizeof(rkprobs->inter_mode)); memcpy(rkprobs->interp_filter, probs->interp_filter, sizeof(rkprobs->interp_filter)); /* 128 x 128 bits coeff related */ for (i = 0; i < ARRAY_SIZE(probs->coef); i++) { for (j = 0; j < ARRAY_SIZE(probs->coef[0]); j++) { for (k = 0; k < ARRAY_SIZE(probs->coef[0][0]); k++) write_coeff_plane(probs->coef[i][j][k], rkprobs->coef[k][i][j]); } } /* intra uv mode 6 x 128 */ memcpy(rkprobs->uv_mode_0_2, &probs->uv_mode[0], sizeof(rkprobs->uv_mode_0_2)); memcpy(rkprobs->uv_mode_3_5, &probs->uv_mode[3], sizeof(rkprobs->uv_mode_3_5)); memcpy(rkprobs->uv_mode_6_8, &probs->uv_mode[6], sizeof(rkprobs->uv_mode_6_8)); memcpy(rkprobs->uv_mode_9, &probs->uv_mode[9], sizeof(rkprobs->uv_mode_9)); /* mv related 6 x 128 */ memcpy(rkprobs->mv.joint, probs->mv.joint, sizeof(rkprobs->mv.joint)); memcpy(rkprobs->mv.sign, probs->mv.sign, sizeof(rkprobs->mv.sign)); memcpy(rkprobs->mv.classes, probs->mv.classes, sizeof(rkprobs->mv.classes)); memcpy(rkprobs->mv.class0_bit, probs->mv.class0_bit, sizeof(rkprobs->mv.class0_bit)); memcpy(rkprobs->mv.bits, probs->mv.bits, sizeof(rkprobs->mv.bits)); memcpy(rkprobs->mv.class0_fr, probs->mv.class0_fr, sizeof(rkprobs->mv.class0_fr)); memcpy(rkprobs->mv.fr, probs->mv.fr, sizeof(rkprobs->mv.fr)); memcpy(rkprobs->mv.class0_hp, probs->mv.class0_hp, sizeof(rkprobs->mv.class0_hp)); memcpy(rkprobs->mv.hp, probs->mv.hp, sizeof(rkprobs->mv.hp)); } static void init_probs(struct rkvdec_ctx *ctx, const struct rkvdec_vp9_run *run) { const struct v4l2_ctrl_vp9_frame *dec_params; struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv; struct rkvdec_vp9_priv_tbl *tbl = vp9_ctx->priv_tbl.cpu; struct rkvdec_vp9_probs *rkprobs = &tbl->probs; const struct v4l2_vp9_segmentation *seg; const struct v4l2_vp9_frame_context *probs; bool intra_only; dec_params = run->decode_params; probs = &vp9_ctx->probability_tables; seg = &dec_params->seg; memset(rkprobs, 0, sizeof(*rkprobs)); intra_only = !!(dec_params->flags & (V4L2_VP9_FRAME_FLAG_KEY_FRAME | V4L2_VP9_FRAME_FLAG_INTRA_ONLY)); /* sb info 5 x 128 bit */ memcpy(rkprobs->partition, intra_only ? v4l2_vp9_kf_partition_probs : probs->partition, sizeof(rkprobs->partition)); memcpy(rkprobs->pred, seg->pred_probs, sizeof(rkprobs->pred)); memcpy(rkprobs->tree, seg->tree_probs, sizeof(rkprobs->tree)); memcpy(rkprobs->skip, probs->skip, sizeof(rkprobs->skip)); memcpy(rkprobs->tx32, probs->tx32, sizeof(rkprobs->tx32)); memcpy(rkprobs->tx16, probs->tx16, sizeof(rkprobs->tx16)); memcpy(rkprobs->tx8, probs->tx8, sizeof(rkprobs->tx8)); memcpy(rkprobs->is_inter, probs->is_inter, sizeof(rkprobs->is_inter)); if (intra_only) init_intra_only_probs(ctx, run); else init_inter_probs(ctx, run); } struct rkvdec_vp9_ref_reg { u32 reg_frm_size; u32 reg_hor_stride; u32 reg_y_stride; u32 reg_yuv_stride; u32 reg_ref_base; }; static struct rkvdec_vp9_ref_reg ref_regs[] = { { .reg_frm_size = RKVDEC_REG_VP9_FRAME_SIZE(0), .reg_hor_stride = RKVDEC_VP9_HOR_VIRSTRIDE(0), .reg_y_stride = RKVDEC_VP9_LAST_FRAME_YSTRIDE, .reg_yuv_stride = RKVDEC_VP9_LAST_FRAME_YUVSTRIDE, .reg_ref_base = RKVDEC_REG_VP9_LAST_FRAME_BASE, }, { .reg_frm_size = RKVDEC_REG_VP9_FRAME_SIZE(1), .reg_hor_stride = RKVDEC_VP9_HOR_VIRSTRIDE(1), .reg_y_stride = RKVDEC_VP9_GOLDEN_FRAME_YSTRIDE, .reg_yuv_stride = 0, .reg_ref_base = RKVDEC_REG_VP9_GOLDEN_FRAME_BASE, }, { .reg_frm_size = RKVDEC_REG_VP9_FRAME_SIZE(2), .reg_hor_stride = RKVDEC_VP9_HOR_VIRSTRIDE(2), .reg_y_stride = RKVDEC_VP9_ALTREF_FRAME_YSTRIDE, .reg_yuv_stride = 0, .reg_ref_base = RKVDEC_REG_VP9_ALTREF_FRAME_BASE, } }; static struct rkvdec_decoded_buffer * get_ref_buf(struct rkvdec_ctx *ctx, struct vb2_v4l2_buffer *dst, u64 timestamp) { struct v4l2_m2m_ctx *m2m_ctx = ctx->fh.m2m_ctx; struct vb2_queue *cap_q = &m2m_ctx->cap_q_ctx.q; int buf_idx; /* * If a ref is unused or invalid, address of current destination * buffer is returned. */ buf_idx = vb2_find_timestamp(cap_q, timestamp, 0); if (buf_idx < 0) return vb2_to_rkvdec_decoded_buf(&dst->vb2_buf); return vb2_to_rkvdec_decoded_buf(vb2_get_buffer(cap_q, buf_idx)); } static dma_addr_t get_mv_base_addr(struct rkvdec_decoded_buffer *buf) { unsigned int aligned_pitch, aligned_height, yuv_len; aligned_height = round_up(buf->vp9.height, 64); aligned_pitch = round_up(buf->vp9.width * buf->vp9.bit_depth, 512) / 8; yuv_len = (aligned_height * aligned_pitch * 3) / 2; return vb2_dma_contig_plane_dma_addr(&buf->base.vb.vb2_buf, 0) + yuv_len; } static void config_ref_registers(struct rkvdec_ctx *ctx, const struct rkvdec_vp9_run *run, struct rkvdec_decoded_buffer *ref_buf, struct rkvdec_vp9_ref_reg *ref_reg) { unsigned int aligned_pitch, aligned_height, y_len, yuv_len; struct rkvdec_dev *rkvdec = ctx->dev; aligned_height = round_up(ref_buf->vp9.height, 64); writel_relaxed(RKVDEC_VP9_FRAMEWIDTH(ref_buf->vp9.width) | RKVDEC_VP9_FRAMEHEIGHT(ref_buf->vp9.height), rkvdec->regs + ref_reg->reg_frm_size); writel_relaxed(vb2_dma_contig_plane_dma_addr(&ref_buf->base.vb.vb2_buf, 0), rkvdec->regs + ref_reg->reg_ref_base); if (&ref_buf->base.vb == run->base.bufs.dst) return; aligned_pitch = round_up(ref_buf->vp9.width * ref_buf->vp9.bit_depth, 512) / 8; y_len = aligned_height * aligned_pitch; yuv_len = (y_len * 3) / 2; writel_relaxed(RKVDEC_HOR_Y_VIRSTRIDE(aligned_pitch / 16) | RKVDEC_HOR_UV_VIRSTRIDE(aligned_pitch / 16), rkvdec->regs + ref_reg->reg_hor_stride); writel_relaxed(RKVDEC_VP9_REF_YSTRIDE(y_len / 16), rkvdec->regs + ref_reg->reg_y_stride); if (!ref_reg->reg_yuv_stride) return; writel_relaxed(RKVDEC_VP9_REF_YUVSTRIDE(yuv_len / 16), rkvdec->regs + ref_reg->reg_yuv_stride); } static void config_seg_registers(struct rkvdec_ctx *ctx, unsigned int segid) { struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv; const struct v4l2_vp9_segmentation *seg; struct rkvdec_dev *rkvdec = ctx->dev; s16 feature_val; int feature_id; u32 val = 0; seg = vp9_ctx->last.valid ? &vp9_ctx->last.seg : &vp9_ctx->cur.seg; feature_id = V4L2_VP9_SEG_LVL_ALT_Q; if (v4l2_vp9_seg_feat_enabled(seg->feature_enabled, feature_id, segid)) { feature_val = seg->feature_data[segid][feature_id]; val |= RKVDEC_SEGID_FRAME_QP_DELTA_EN(1) | RKVDEC_SEGID_FRAME_QP_DELTA(feature_val); } feature_id = V4L2_VP9_SEG_LVL_ALT_L; if (v4l2_vp9_seg_feat_enabled(seg->feature_enabled, feature_id, segid)) { feature_val = seg->feature_data[segid][feature_id]; val |= RKVDEC_SEGID_FRAME_LOOPFILTER_VALUE_EN(1) | RKVDEC_SEGID_FRAME_LOOPFILTER_VALUE(feature_val); } feature_id = V4L2_VP9_SEG_LVL_REF_FRAME; if (v4l2_vp9_seg_feat_enabled(seg->feature_enabled, feature_id, segid)) { feature_val = seg->feature_data[segid][feature_id]; val |= RKVDEC_SEGID_REFERINFO_EN(1) | RKVDEC_SEGID_REFERINFO(feature_val); } feature_id = V4L2_VP9_SEG_LVL_SKIP; if (v4l2_vp9_seg_feat_enabled(seg->feature_enabled, feature_id, segid)) val |= RKVDEC_SEGID_FRAME_SKIP_EN(1); if (!segid && (seg->flags & V4L2_VP9_SEGMENTATION_FLAG_ABS_OR_DELTA_UPDATE)) val |= RKVDEC_SEGID_ABS_DELTA(1); writel_relaxed(val, rkvdec->regs + RKVDEC_VP9_SEGID_GRP(segid)); } static void update_dec_buf_info(struct rkvdec_decoded_buffer *buf, const struct v4l2_ctrl_vp9_frame *dec_params) { buf->vp9.width = dec_params->frame_width_minus_1 + 1; buf->vp9.height = dec_params->frame_height_minus_1 + 1; buf->vp9.bit_depth = dec_params->bit_depth; } static void update_ctx_cur_info(struct rkvdec_vp9_ctx *vp9_ctx, struct rkvdec_decoded_buffer *buf, const struct v4l2_ctrl_vp9_frame *dec_params) { vp9_ctx->cur.valid = true; vp9_ctx->cur.reference_mode = dec_params->reference_mode; vp9_ctx->cur.interpolation_filter = dec_params->interpolation_filter; vp9_ctx->cur.flags = dec_params->flags; vp9_ctx->cur.timestamp = buf->base.vb.vb2_buf.timestamp; vp9_ctx->cur.seg = dec_params->seg; vp9_ctx->cur.lf = dec_params->lf; } static void update_ctx_last_info(struct rkvdec_vp9_ctx *vp9_ctx) { vp9_ctx->last = vp9_ctx->cur; } static void config_registers(struct rkvdec_ctx *ctx, const struct rkvdec_vp9_run *run) { unsigned int y_len, uv_len, yuv_len, bit_depth, aligned_height, aligned_pitch, stream_len; const struct v4l2_ctrl_vp9_frame *dec_params; struct rkvdec_decoded_buffer *ref_bufs[3]; struct rkvdec_decoded_buffer *dst, *last, *mv_ref; struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv; u32 val, last_frame_info = 0; const struct v4l2_vp9_segmentation *seg; struct rkvdec_dev *rkvdec = ctx->dev; dma_addr_t addr; bool intra_only; unsigned int i; dec_params = run->decode_params; dst = vb2_to_rkvdec_decoded_buf(&run->base.bufs.dst->vb2_buf); ref_bufs[0] = get_ref_buf(ctx, &dst->base.vb, dec_params->last_frame_ts); ref_bufs[1] = get_ref_buf(ctx, &dst->base.vb, dec_params->golden_frame_ts); ref_bufs[2] = get_ref_buf(ctx, &dst->base.vb, dec_params->alt_frame_ts); if (vp9_ctx->last.valid) last = get_ref_buf(ctx, &dst->base.vb, vp9_ctx->last.timestamp); else last = dst; update_dec_buf_info(dst, dec_params); update_ctx_cur_info(vp9_ctx, dst, dec_params); seg = &dec_params->seg; intra_only = !!(dec_params->flags & (V4L2_VP9_FRAME_FLAG_KEY_FRAME | V4L2_VP9_FRAME_FLAG_INTRA_ONLY)); writel_relaxed(RKVDEC_MODE(RKVDEC_MODE_VP9), rkvdec->regs + RKVDEC_REG_SYSCTRL); bit_depth = dec_params->bit_depth; aligned_height = round_up(ctx->decoded_fmt.fmt.pix_mp.height, 64); aligned_pitch = round_up(ctx->decoded_fmt.fmt.pix_mp.width * bit_depth, 512) / 8; y_len = aligned_height * aligned_pitch; uv_len = y_len / 2; yuv_len = y_len + uv_len; writel_relaxed(RKVDEC_Y_HOR_VIRSTRIDE(aligned_pitch / 16) | RKVDEC_UV_HOR_VIRSTRIDE(aligned_pitch / 16), rkvdec->regs + RKVDEC_REG_PICPAR); writel_relaxed(RKVDEC_Y_VIRSTRIDE(y_len / 16), rkvdec->regs + RKVDEC_REG_Y_VIRSTRIDE); writel_relaxed(RKVDEC_YUV_VIRSTRIDE(yuv_len / 16), rkvdec->regs + RKVDEC_REG_YUV_VIRSTRIDE); stream_len = vb2_get_plane_payload(&run->base.bufs.src->vb2_buf, 0); writel_relaxed(RKVDEC_STRM_LEN(stream_len), rkvdec->regs + RKVDEC_REG_STRM_LEN); /* * Reset count buffer, because decoder only output intra related syntax * counts when decoding intra frame, but update entropy need to update * all the probabilities. */ if (intra_only) memset(vp9_ctx->count_tbl.cpu, 0, vp9_ctx->count_tbl.size); vp9_ctx->cur.segmapid = vp9_ctx->last.segmapid; if (!intra_only && !(dec_params->flags & V4L2_VP9_FRAME_FLAG_ERROR_RESILIENT) && (!(seg->flags & V4L2_VP9_SEGMENTATION_FLAG_ENABLED) || (seg->flags & V4L2_VP9_SEGMENTATION_FLAG_UPDATE_MAP))) vp9_ctx->cur.segmapid++; for (i = 0; i < ARRAY_SIZE(ref_bufs); i++) config_ref_registers(ctx, run, ref_bufs[i], &ref_regs[i]); for (i = 0; i < 8; i++) config_seg_registers(ctx, i); writel_relaxed(RKVDEC_VP9_TX_MODE(vp9_ctx->cur.tx_mode) | RKVDEC_VP9_FRAME_REF_MODE(dec_params->reference_mode), rkvdec->regs + RKVDEC_VP9_CPRHEADER_CONFIG); if (!intra_only) { const struct v4l2_vp9_loop_filter *lf; s8 delta; if (vp9_ctx->last.valid) lf = &vp9_ctx->last.lf; else lf = &vp9_ctx->cur.lf; val = 0; for (i = 0; i < ARRAY_SIZE(lf->ref_deltas); i++) { delta = lf->ref_deltas[i]; val |= RKVDEC_REF_DELTAS_LASTFRAME(i, delta); } writel_relaxed(val, rkvdec->regs + RKVDEC_VP9_REF_DELTAS_LASTFRAME); for (i = 0; i < ARRAY_SIZE(lf->mode_deltas); i++) { delta = lf->mode_deltas[i]; last_frame_info |= RKVDEC_MODE_DELTAS_LASTFRAME(i, delta); } } if (vp9_ctx->last.valid && !intra_only && vp9_ctx->last.seg.flags & V4L2_VP9_SEGMENTATION_FLAG_ENABLED) last_frame_info |= RKVDEC_SEG_EN_LASTFRAME; if (vp9_ctx->last.valid && vp9_ctx->last.flags & V4L2_VP9_FRAME_FLAG_SHOW_FRAME) last_frame_info |= RKVDEC_LAST_SHOW_FRAME; if (vp9_ctx->last.valid && vp9_ctx->last.flags & (V4L2_VP9_FRAME_FLAG_KEY_FRAME | V4L2_VP9_FRAME_FLAG_INTRA_ONLY)) last_frame_info |= RKVDEC_LAST_INTRA_ONLY; if (vp9_ctx->last.valid && last->vp9.width == dst->vp9.width && last->vp9.height == dst->vp9.height) last_frame_info |= RKVDEC_LAST_WIDHHEIGHT_EQCUR; writel_relaxed(last_frame_info, rkvdec->regs + RKVDEC_VP9_INFO_LASTFRAME); writel_relaxed(stream_len - dec_params->compressed_header_size - dec_params->uncompressed_header_size, rkvdec->regs + RKVDEC_VP9_LASTTILE_SIZE); for (i = 0; !intra_only && i < ARRAY_SIZE(ref_bufs); i++) { unsigned int refw = ref_bufs[i]->vp9.width; unsigned int refh = ref_bufs[i]->vp9.height; u32 hscale, vscale; hscale = (refw << 14) / dst->vp9.width; vscale = (refh << 14) / dst->vp9.height; writel_relaxed(RKVDEC_VP9_REF_HOR_SCALE(hscale) | RKVDEC_VP9_REF_VER_SCALE(vscale), rkvdec->regs + RKVDEC_VP9_REF_SCALE(i)); } addr = vb2_dma_contig_plane_dma_addr(&dst->base.vb.vb2_buf, 0); writel_relaxed(addr, rkvdec->regs + RKVDEC_REG_DECOUT_BASE); addr = vb2_dma_contig_plane_dma_addr(&run->base.bufs.src->vb2_buf, 0); writel_relaxed(addr, rkvdec->regs + RKVDEC_REG_STRM_RLC_BASE); writel_relaxed(vp9_ctx->priv_tbl.dma + offsetof(struct rkvdec_vp9_priv_tbl, probs), rkvdec->regs + RKVDEC_REG_CABACTBL_PROB_BASE); writel_relaxed(vp9_ctx->count_tbl.dma, rkvdec->regs + RKVDEC_REG_VP9COUNT_BASE); writel_relaxed(vp9_ctx->priv_tbl.dma + offsetof(struct rkvdec_vp9_priv_tbl, segmap) + (RKVDEC_VP9_MAX_SEGMAP_SIZE * vp9_ctx->cur.segmapid), rkvdec->regs + RKVDEC_REG_VP9_SEGIDCUR_BASE); writel_relaxed(vp9_ctx->priv_tbl.dma + offsetof(struct rkvdec_vp9_priv_tbl, segmap) + (RKVDEC_VP9_MAX_SEGMAP_SIZE * (!vp9_ctx->cur.segmapid)), rkvdec->regs + RKVDEC_REG_VP9_SEGIDLAST_BASE); if (!intra_only && !(dec_params->flags & V4L2_VP9_FRAME_FLAG_ERROR_RESILIENT) && vp9_ctx->last.valid) mv_ref = last; else mv_ref = dst; writel_relaxed(get_mv_base_addr(mv_ref), rkvdec->regs + RKVDEC_VP9_REF_COLMV_BASE); writel_relaxed(ctx->decoded_fmt.fmt.pix_mp.width | (ctx->decoded_fmt.fmt.pix_mp.height << 16), rkvdec->regs + RKVDEC_REG_PERFORMANCE_CYCLE); } static int validate_dec_params(struct rkvdec_ctx *ctx, const struct v4l2_ctrl_vp9_frame *dec_params) { unsigned int aligned_width, aligned_height; /* We only support profile 0. */ if (dec_params->profile != 0) { dev_err(ctx->dev->dev, "unsupported profile %d\n", dec_params->profile); return -EINVAL; } aligned_width = round_up(dec_params->frame_width_minus_1 + 1, 64); aligned_height = round_up(dec_params->frame_height_minus_1 + 1, 64); /* * Userspace should update the capture/decoded format when the * resolution changes. */ if (aligned_width != ctx->decoded_fmt.fmt.pix_mp.width || aligned_height != ctx->decoded_fmt.fmt.pix_mp.height) { dev_err(ctx->dev->dev, "unexpected bitstream resolution %dx%d\n", dec_params->frame_width_minus_1 + 1, dec_params->frame_height_minus_1 + 1); return -EINVAL; } return 0; } static int rkvdec_vp9_run_preamble(struct rkvdec_ctx *ctx, struct rkvdec_vp9_run *run) { const struct v4l2_ctrl_vp9_frame *dec_params; const struct v4l2_ctrl_vp9_compressed_hdr *prob_updates; struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv; struct v4l2_ctrl *ctrl; unsigned int fctx_idx; int ret; /* v4l2-specific stuff */ rkvdec_run_preamble(ctx, &run->base); ctrl = v4l2_ctrl_find(&ctx->ctrl_hdl, V4L2_CID_STATELESS_VP9_FRAME); if (WARN_ON(!ctrl)) return -EINVAL; dec_params = ctrl->p_cur.p; ret = validate_dec_params(ctx, dec_params); if (ret) return ret; run->decode_params = dec_params; ctrl = v4l2_ctrl_find(&ctx->ctrl_hdl, V4L2_CID_STATELESS_VP9_COMPRESSED_HDR); if (WARN_ON(!ctrl)) return -EINVAL; prob_updates = ctrl->p_cur.p; vp9_ctx->cur.tx_mode = prob_updates->tx_mode; /* * vp9 stuff * * by this point the userspace has done all parts of 6.2 uncompressed_header() * except this fragment: * if ( FrameIsIntra || error_resilient_mode ) { * setup_past_independence ( ) * if ( frame_type == KEY_FRAME || error_resilient_mode == 1 || * reset_frame_context == 3 ) { * for ( i = 0; i < 4; i ++ ) { * save_probs( i ) * } * } else if ( reset_frame_context == 2 ) { * save_probs( frame_context_idx ) * } * frame_context_idx = 0 * } */ fctx_idx = v4l2_vp9_reset_frame_ctx(dec_params, vp9_ctx->frame_context); vp9_ctx->cur.frame_context_idx = fctx_idx; /* 6.1 frame(sz): load_probs() and load_probs2() */ vp9_ctx->probability_tables = vp9_ctx->frame_context[fctx_idx]; /* * The userspace has also performed 6.3 compressed_header(), but handling the * probs in a special way. All probs which need updating, except MV-related, * have been read from the bitstream and translated through inv_map_table[], * but no 6.3.6 inv_recenter_nonneg(v, m) has been performed. The values passed * by userspace are either translated values (there are no 0 values in * inv_map_table[]), or zero to indicate no update. All MV-related probs which need * updating have been read from the bitstream and (mv_prob << 1) | 1 has been * performed. The values passed by userspace are either new values * to replace old ones (the above mentioned shift and bitwise or never result in * a zero) or zero to indicate no update. * fw_update_probs() performs actual probs updates or leaves probs as-is * for values for which a zero was passed from userspace. */ v4l2_vp9_fw_update_probs(&vp9_ctx->probability_tables, prob_updates, dec_params); return 0; } static int rkvdec_vp9_run(struct rkvdec_ctx *ctx) { struct rkvdec_dev *rkvdec = ctx->dev; struct rkvdec_vp9_run run = { }; int ret; ret = rkvdec_vp9_run_preamble(ctx, &run); if (ret) { rkvdec_run_postamble(ctx, &run.base); return ret; } /* Prepare probs. */ init_probs(ctx, &run); /* Configure hardware registers. */ config_registers(ctx, &run); rkvdec_run_postamble(ctx, &run.base); schedule_delayed_work(&rkvdec->watchdog_work, msecs_to_jiffies(2000)); writel(1, rkvdec->regs + RKVDEC_REG_PREF_LUMA_CACHE_COMMAND); writel(1, rkvdec->regs + RKVDEC_REG_PREF_CHR_CACHE_COMMAND); writel(0xe, rkvdec->regs + RKVDEC_REG_STRMD_ERR_EN); /* Start decoding! */ writel(RKVDEC_INTERRUPT_DEC_E | RKVDEC_CONFIG_DEC_CLK_GATE_E | RKVDEC_TIMEOUT_E | RKVDEC_BUF_EMPTY_E, rkvdec->regs + RKVDEC_REG_INTERRUPT); return 0; } #define copy_tx_and_skip(p1, p2) \ do { \ memcpy((p1)->tx8, (p2)->tx8, sizeof((p1)->tx8)); \ memcpy((p1)->tx16, (p2)->tx16, sizeof((p1)->tx16)); \ memcpy((p1)->tx32, (p2)->tx32, sizeof((p1)->tx32)); \ memcpy((p1)->skip, (p2)->skip, sizeof((p1)->skip)); \ } while (0) static void rkvdec_vp9_done(struct rkvdec_ctx *ctx, struct vb2_v4l2_buffer *src_buf, struct vb2_v4l2_buffer *dst_buf, enum vb2_buffer_state result) { struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv; unsigned int fctx_idx; /* v4l2-specific stuff */ if (result == VB2_BUF_STATE_ERROR) goto out_update_last; /* * vp9 stuff * * 6.1.2 refresh_probs() * * In the spec a complementary condition goes last in 6.1.2 refresh_probs(), * but it makes no sense to perform all the activities from the first "if" * there if we actually are not refreshing the frame context. On top of that, * because of 6.2 uncompressed_header() whenever error_resilient_mode == 1, * refresh_frame_context == 0. Consequently, if we don't jump to out_update_last * it means error_resilient_mode must be 0. */ if (!(vp9_ctx->cur.flags & V4L2_VP9_FRAME_FLAG_REFRESH_FRAME_CTX)) goto out_update_last; fctx_idx = vp9_ctx->cur.frame_context_idx; if (!(vp9_ctx->cur.flags & V4L2_VP9_FRAME_FLAG_PARALLEL_DEC_MODE)) { /* error_resilient_mode == 0 && frame_parallel_decoding_mode == 0 */ struct v4l2_vp9_frame_context *probs = &vp9_ctx->probability_tables; bool frame_is_intra = vp9_ctx->cur.flags & (V4L2_VP9_FRAME_FLAG_KEY_FRAME | V4L2_VP9_FRAME_FLAG_INTRA_ONLY); struct tx_and_skip { u8 tx8[2][1]; u8 tx16[2][2]; u8 tx32[2][3]; u8 skip[3]; } _tx_skip, *tx_skip = &_tx_skip; struct v4l2_vp9_frame_symbol_counts *counts; /* buffer the forward-updated TX and skip probs */ if (frame_is_intra) copy_tx_and_skip(tx_skip, probs); /* 6.1.2 refresh_probs(): load_probs() and load_probs2() */ *probs = vp9_ctx->frame_context[fctx_idx]; /* if FrameIsIntra then undo the effect of load_probs2() */ if (frame_is_intra) copy_tx_and_skip(probs, tx_skip); counts = frame_is_intra ? &vp9_ctx->intra_cnts : &vp9_ctx->inter_cnts; v4l2_vp9_adapt_coef_probs(probs, counts, !vp9_ctx->last.valid || vp9_ctx->last.flags & V4L2_VP9_FRAME_FLAG_KEY_FRAME, frame_is_intra); if (!frame_is_intra) { const struct rkvdec_vp9_inter_frame_symbol_counts *inter_cnts; u32 classes[2][11]; int i; inter_cnts = vp9_ctx->count_tbl.cpu; for (i = 0; i < ARRAY_SIZE(classes); ++i) memcpy(classes[i], inter_cnts->classes[i], sizeof(classes[0])); counts->classes = &classes; /* load_probs2() already done */ v4l2_vp9_adapt_noncoef_probs(&vp9_ctx->probability_tables, counts, vp9_ctx->cur.reference_mode, vp9_ctx->cur.interpolation_filter, vp9_ctx->cur.tx_mode, vp9_ctx->cur.flags); } } /* 6.1.2 refresh_probs(): save_probs(fctx_idx) */ vp9_ctx->frame_context[fctx_idx] = vp9_ctx->probability_tables; out_update_last: update_ctx_last_info(vp9_ctx); } static void rkvdec_init_v4l2_vp9_count_tbl(struct rkvdec_ctx *ctx) { struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv; struct rkvdec_vp9_intra_frame_symbol_counts *intra_cnts = vp9_ctx->count_tbl.cpu; struct rkvdec_vp9_inter_frame_symbol_counts *inter_cnts = vp9_ctx->count_tbl.cpu; int i, j, k, l, m; vp9_ctx->inter_cnts.partition = &inter_cnts->partition; vp9_ctx->inter_cnts.skip = &inter_cnts->skip; vp9_ctx->inter_cnts.intra_inter = &inter_cnts->inter; vp9_ctx->inter_cnts.tx32p = &inter_cnts->tx32p; vp9_ctx->inter_cnts.tx16p = &inter_cnts->tx16p; vp9_ctx->inter_cnts.tx8p = &inter_cnts->tx8p; vp9_ctx->intra_cnts.partition = (u32 (*)[16][4])(&intra_cnts->partition); vp9_ctx->intra_cnts.skip = &intra_cnts->skip; vp9_ctx->intra_cnts.intra_inter = &intra_cnts->intra; vp9_ctx->intra_cnts.tx32p = &intra_cnts->tx32p; vp9_ctx->intra_cnts.tx16p = &intra_cnts->tx16p; vp9_ctx->intra_cnts.tx8p = &intra_cnts->tx8p; vp9_ctx->inter_cnts.y_mode = &inter_cnts->y_mode; vp9_ctx->inter_cnts.uv_mode = &inter_cnts->uv_mode; vp9_ctx->inter_cnts.comp = &inter_cnts->comp; vp9_ctx->inter_cnts.comp_ref = &inter_cnts->comp_ref; vp9_ctx->inter_cnts.single_ref = &inter_cnts->single_ref; vp9_ctx->inter_cnts.mv_mode = &inter_cnts->mv_mode; vp9_ctx->inter_cnts.filter = &inter_cnts->filter; vp9_ctx->inter_cnts.mv_joint = &inter_cnts->mv_joint; vp9_ctx->inter_cnts.sign = &inter_cnts->sign; /* * rk hardware actually uses "u32 classes[2][11 + 1];" * instead of "u32 classes[2][11];", so this must be explicitly * copied into vp9_ctx->classes when passing the data to the * vp9 library function */ vp9_ctx->inter_cnts.class0 = &inter_cnts->class0; vp9_ctx->inter_cnts.bits = &inter_cnts->bits; vp9_ctx->inter_cnts.class0_fp = &inter_cnts->class0_fp; vp9_ctx->inter_cnts.fp = &inter_cnts->fp; vp9_ctx->inter_cnts.class0_hp = &inter_cnts->class0_hp; vp9_ctx->inter_cnts.hp = &inter_cnts->hp; #define INNERMOST_LOOP \ do { \ for (m = 0; m < ARRAY_SIZE(vp9_ctx->inter_cnts.coeff[0][0][0][0]); ++m) {\ vp9_ctx->inter_cnts.coeff[i][j][k][l][m] = \ &inter_cnts->ref_cnt[k][i][j][l][m].coeff; \ vp9_ctx->inter_cnts.eob[i][j][k][l][m][0] = \ &inter_cnts->ref_cnt[k][i][j][l][m].eob[0]; \ vp9_ctx->inter_cnts.eob[i][j][k][l][m][1] = \ &inter_cnts->ref_cnt[k][i][j][l][m].eob[1]; \ \ vp9_ctx->intra_cnts.coeff[i][j][k][l][m] = \ &intra_cnts->ref_cnt[k][i][j][l][m].coeff; \ vp9_ctx->intra_cnts.eob[i][j][k][l][m][0] = \ &intra_cnts->ref_cnt[k][i][j][l][m].eob[0]; \ vp9_ctx->intra_cnts.eob[i][j][k][l][m][1] = \ &intra_cnts->ref_cnt[k][i][j][l][m].eob[1]; \ } \ } while (0) for (i = 0; i < ARRAY_SIZE(vp9_ctx->inter_cnts.coeff); ++i) for (j = 0; j < ARRAY_SIZE(vp9_ctx->inter_cnts.coeff[0]); ++j) for (k = 0; k < ARRAY_SIZE(vp9_ctx->inter_cnts.coeff[0][0]); ++k) for (l = 0; l < ARRAY_SIZE(vp9_ctx->inter_cnts.coeff[0][0][0]); ++l) INNERMOST_LOOP; #undef INNERMOST_LOOP } static int rkvdec_vp9_start(struct rkvdec_ctx *ctx) { struct rkvdec_dev *rkvdec = ctx->dev; struct rkvdec_vp9_priv_tbl *priv_tbl; struct rkvdec_vp9_ctx *vp9_ctx; unsigned char *count_tbl; int ret; vp9_ctx = kzalloc(sizeof(*vp9_ctx), GFP_KERNEL); if (!vp9_ctx) return -ENOMEM; ctx->priv = vp9_ctx; priv_tbl = dma_alloc_coherent(rkvdec->dev, sizeof(*priv_tbl), &vp9_ctx->priv_tbl.dma, GFP_KERNEL); if (!priv_tbl) { ret = -ENOMEM; goto err_free_ctx; } vp9_ctx->priv_tbl.size = sizeof(*priv_tbl); vp9_ctx->priv_tbl.cpu = priv_tbl; memset(priv_tbl, 0, sizeof(*priv_tbl)); count_tbl = dma_alloc_coherent(rkvdec->dev, RKVDEC_VP9_COUNT_SIZE, &vp9_ctx->count_tbl.dma, GFP_KERNEL); if (!count_tbl) { ret = -ENOMEM; goto err_free_priv_tbl; } vp9_ctx->count_tbl.size = RKVDEC_VP9_COUNT_SIZE; vp9_ctx->count_tbl.cpu = count_tbl; memset(count_tbl, 0, sizeof(*count_tbl)); rkvdec_init_v4l2_vp9_count_tbl(ctx); return 0; err_free_priv_tbl: dma_free_coherent(rkvdec->dev, vp9_ctx->priv_tbl.size, vp9_ctx->priv_tbl.cpu, vp9_ctx->priv_tbl.dma); err_free_ctx: kfree(vp9_ctx); return ret; } static void rkvdec_vp9_stop(struct rkvdec_ctx *ctx) { struct rkvdec_vp9_ctx *vp9_ctx = ctx->priv; struct rkvdec_dev *rkvdec = ctx->dev; dma_free_coherent(rkvdec->dev, vp9_ctx->count_tbl.size, vp9_ctx->count_tbl.cpu, vp9_ctx->count_tbl.dma); dma_free_coherent(rkvdec->dev, vp9_ctx->priv_tbl.size, vp9_ctx->priv_tbl.cpu, vp9_ctx->priv_tbl.dma); kfree(vp9_ctx); } static int rkvdec_vp9_adjust_fmt(struct rkvdec_ctx *ctx, struct v4l2_format *f) { struct v4l2_pix_format_mplane *fmt = &f->fmt.pix_mp; fmt->num_planes = 1; if (!fmt->plane_fmt[0].sizeimage) fmt->plane_fmt[0].sizeimage = fmt->width * fmt->height * 2; return 0; } const struct rkvdec_coded_fmt_ops rkvdec_vp9_fmt_ops = { .adjust_fmt = rkvdec_vp9_adjust_fmt, .start = rkvdec_vp9_start, .stop = rkvdec_vp9_stop, .run = rkvdec_vp9_run, .done = rkvdec_vp9_done, };