| Index: source/libvpx/vp9/encoder/vp9_encodemb.c
|
| ===================================================================
|
| --- source/libvpx/vp9/encoder/vp9_encodemb.c (revision 254565)
|
| +++ source/libvpx/vp9/encoder/vp9_encodemb.c (working copy)
|
| @@ -36,22 +36,22 @@
|
| };
|
|
|
| void vp9_subtract_block_c(int rows, int cols,
|
| - int16_t *diff_ptr, ptrdiff_t diff_stride,
|
| - const uint8_t *src_ptr, ptrdiff_t src_stride,
|
| - const uint8_t *pred_ptr, ptrdiff_t pred_stride) {
|
| + int16_t *diff, ptrdiff_t diff_stride,
|
| + const uint8_t *src, ptrdiff_t src_stride,
|
| + const uint8_t *pred, ptrdiff_t pred_stride) {
|
| int r, c;
|
|
|
| for (r = 0; r < rows; r++) {
|
| for (c = 0; c < cols; c++)
|
| - diff_ptr[c] = src_ptr[c] - pred_ptr[c];
|
| + diff[c] = src[c] - pred[c];
|
|
|
| - diff_ptr += diff_stride;
|
| - pred_ptr += pred_stride;
|
| - src_ptr += src_stride;
|
| + diff += diff_stride;
|
| + pred += pred_stride;
|
| + src += src_stride;
|
| }
|
| }
|
|
|
| -static void subtract_plane(MACROBLOCK *x, BLOCK_SIZE bsize, int plane) {
|
| +void vp9_subtract_plane(MACROBLOCK *x, BLOCK_SIZE bsize, int plane) {
|
| struct macroblock_plane *const p = &x->plane[plane];
|
| const struct macroblockd_plane *const pd = &x->e_mbd.plane[plane];
|
| const BLOCK_SIZE plane_bsize = get_plane_block_size(bsize, pd);
|
| @@ -62,22 +62,6 @@
|
| pd->dst.buf, pd->dst.stride);
|
| }
|
|
|
| -void vp9_subtract_sby(MACROBLOCK *x, BLOCK_SIZE bsize) {
|
| - subtract_plane(x, bsize, 0);
|
| -}
|
| -
|
| -void vp9_subtract_sbuv(MACROBLOCK *x, BLOCK_SIZE bsize) {
|
| - int i;
|
| -
|
| - for (i = 1; i < MAX_MB_PLANE; i++)
|
| - subtract_plane(x, bsize, i);
|
| -}
|
| -
|
| -void vp9_subtract_sb(MACROBLOCK *x, BLOCK_SIZE bsize) {
|
| - vp9_subtract_sby(x, bsize);
|
| - vp9_subtract_sbuv(x, bsize);
|
| -}
|
| -
|
| #define RDTRUNC(RM, DM, R, D) ((128 + (R) * (RM)) & 0xFF)
|
| typedef struct vp9_token_state vp9_token_state;
|
|
|
| @@ -121,19 +105,18 @@
|
| return pt;
|
| }
|
|
|
| -static void optimize_b(MACROBLOCK *mb,
|
| - int plane, int block, BLOCK_SIZE plane_bsize,
|
| - ENTROPY_CONTEXT *a, ENTROPY_CONTEXT *l,
|
| - TX_SIZE tx_size) {
|
| +static void optimize_b(int plane, int block, BLOCK_SIZE plane_bsize,
|
| + TX_SIZE tx_size, MACROBLOCK *mb,
|
| + struct optimize_ctx *ctx) {
|
| MACROBLOCKD *const xd = &mb->e_mbd;
|
| struct macroblock_plane *p = &mb->plane[plane];
|
| struct macroblockd_plane *pd = &xd->plane[plane];
|
| const int ref = is_inter_block(&xd->mi_8x8[0]->mbmi);
|
| vp9_token_state tokens[1025][2];
|
| unsigned best_index[1025][2];
|
| - const int16_t *coeff_ptr = BLOCK_OFFSET(mb->plane[plane].coeff, block);
|
| - int16_t *qcoeff_ptr;
|
| - int16_t *dqcoeff_ptr;
|
| + const int16_t *coeff = BLOCK_OFFSET(mb->plane[plane].coeff, block);
|
| + int16_t *qcoeff = BLOCK_OFFSET(p->qcoeff, block);
|
| + int16_t *dqcoeff = BLOCK_OFFSET(pd->dqcoeff, block);
|
| int eob = p->eobs[block], final_eob, sz = 0;
|
| const int i0 = 0;
|
| int rc, x, next, i;
|
| @@ -143,7 +126,6 @@
|
| PLANE_TYPE type = pd->plane_type;
|
| int err_mult = plane_rd_mult[type];
|
| const int default_eob = 16 << (tx_size << 1);
|
| -
|
| const int mul = 1 + (tx_size == TX_32X32);
|
| uint8_t token_cache[1024];
|
| const int16_t *dequant_ptr = pd->dequant;
|
| @@ -151,10 +133,13 @@
|
| const scan_order *so = get_scan(xd, tx_size, type, block);
|
| const int16_t *scan = so->scan;
|
| const int16_t *nb = so->neighbors;
|
| + ENTROPY_CONTEXT *a, *l;
|
| + int tx_x, tx_y;
|
| + txfrm_block_to_raster_xy(plane_bsize, tx_size, block, &tx_x, &tx_y);
|
| + a = &ctx->ta[plane][tx_x];
|
| + l = &ctx->tl[plane][tx_y];
|
|
|
| assert((!type && !plane) || (type && plane));
|
| - dqcoeff_ptr = BLOCK_OFFSET(pd->dqcoeff, block);
|
| - qcoeff_ptr = BLOCK_OFFSET(p->qcoeff, block);
|
| assert(eob <= default_eob);
|
|
|
| /* Now set up a Viterbi trellis to evaluate alternative roundings. */
|
| @@ -172,13 +157,13 @@
|
| next = eob;
|
| for (i = 0; i < eob; i++)
|
| token_cache[scan[i]] = vp9_pt_energy_class[vp9_dct_value_tokens_ptr[
|
| - qcoeff_ptr[scan[i]]].token];
|
| + qcoeff[scan[i]]].token];
|
|
|
| for (i = eob; i-- > i0;) {
|
| int base_bits, d2, dx;
|
|
|
| rc = scan[i];
|
| - x = qcoeff_ptr[rc];
|
| + x = qcoeff[rc];
|
| /* Only add a trellis state for non-zero coefficients. */
|
| if (x) {
|
| int shortcut = 0;
|
| @@ -203,7 +188,7 @@
|
| /* And pick the best. */
|
| best = rd_cost1 < rd_cost0;
|
| base_bits = *(vp9_dct_value_cost_ptr + x);
|
| - dx = mul * (dqcoeff_ptr[rc] - coeff_ptr[rc]);
|
| + dx = mul * (dqcoeff[rc] - coeff[rc]);
|
| d2 = dx * dx;
|
| tokens[i][0].rate = base_bits + (best ? rate1 : rate0);
|
| tokens[i][0].error = d2 + (best ? error1 : error0);
|
| @@ -216,8 +201,8 @@
|
| rate0 = tokens[next][0].rate;
|
| rate1 = tokens[next][1].rate;
|
|
|
| - if ((abs(x)*dequant_ptr[rc != 0] > abs(coeff_ptr[rc]) * mul) &&
|
| - (abs(x)*dequant_ptr[rc != 0] < abs(coeff_ptr[rc]) * mul +
|
| + if ((abs(x)*dequant_ptr[rc != 0] > abs(coeff[rc]) * mul) &&
|
| + (abs(x)*dequant_ptr[rc != 0] < abs(coeff[rc]) * mul +
|
| dequant_ptr[rc != 0]))
|
| shortcut = 1;
|
| else
|
| @@ -306,16 +291,16 @@
|
| UPDATE_RD_COST();
|
| best = rd_cost1 < rd_cost0;
|
| final_eob = i0 - 1;
|
| - vpx_memset(qcoeff_ptr, 0, sizeof(*qcoeff_ptr) * (16 << (tx_size * 2)));
|
| - vpx_memset(dqcoeff_ptr, 0, sizeof(*dqcoeff_ptr) * (16 << (tx_size * 2)));
|
| + vpx_memset(qcoeff, 0, sizeof(*qcoeff) * (16 << (tx_size * 2)));
|
| + vpx_memset(dqcoeff, 0, sizeof(*dqcoeff) * (16 << (tx_size * 2)));
|
| for (i = next; i < eob; i = next) {
|
| x = tokens[i][best].qc;
|
| if (x) {
|
| final_eob = i;
|
| }
|
| rc = scan[i];
|
| - qcoeff_ptr[rc] = x;
|
| - dqcoeff_ptr[rc] = (x * dequant_ptr[rc != 0]) / mul;
|
| + qcoeff[rc] = x;
|
| + dqcoeff[rc] = (x * dequant_ptr[rc != 0]) / mul;
|
|
|
| next = tokens[i][best].next;
|
| best = best_index[i][best];
|
| @@ -326,58 +311,39 @@
|
| *a = *l = (final_eob > 0);
|
| }
|
|
|
| -void vp9_optimize_b(int plane, int block, BLOCK_SIZE plane_bsize,
|
| - TX_SIZE tx_size, MACROBLOCK *mb, struct optimize_ctx *ctx) {
|
| - int x, y;
|
| - txfrm_block_to_raster_xy(plane_bsize, tx_size, block, &x, &y);
|
| - optimize_b(mb, plane, block, plane_bsize,
|
| - &ctx->ta[plane][x], &ctx->tl[plane][y], tx_size);
|
| +static INLINE void fdct32x32(int rd_transform,
|
| + const int16_t *src, int16_t *dst, int src_stride) {
|
| + if (rd_transform)
|
| + vp9_fdct32x32_rd(src, dst, src_stride);
|
| + else
|
| + vp9_fdct32x32(src, dst, src_stride);
|
| }
|
|
|
| -static void optimize_init_b(int plane, BLOCK_SIZE bsize,
|
| - struct encode_b_args *args) {
|
| - const MACROBLOCKD *xd = &args->x->e_mbd;
|
| - const struct macroblockd_plane* const pd = &xd->plane[plane];
|
| - const BLOCK_SIZE plane_bsize = get_plane_block_size(bsize, pd);
|
| - const int num_4x4_w = num_4x4_blocks_wide_lookup[plane_bsize];
|
| - const int num_4x4_h = num_4x4_blocks_high_lookup[plane_bsize];
|
| - const MB_MODE_INFO *mbmi = &xd->mi_8x8[0]->mbmi;
|
| - const TX_SIZE tx_size = plane ? get_uv_tx_size(mbmi) : mbmi->tx_size;
|
| -
|
| - vp9_get_entropy_contexts(tx_size, args->ctx->ta[plane], args->ctx->tl[plane],
|
| - pd->above_context, pd->left_context,
|
| - num_4x4_w, num_4x4_h);
|
| -}
|
| void vp9_xform_quant(MACROBLOCK *x, int plane, int block,
|
| BLOCK_SIZE plane_bsize, TX_SIZE tx_size) {
|
| MACROBLOCKD *const xd = &x->e_mbd;
|
| - struct macroblock_plane *const p = &x->plane[plane];
|
| - struct macroblockd_plane *const pd = &xd->plane[plane];
|
| - int16_t *coeff = BLOCK_OFFSET(p->coeff, block);
|
| - int16_t *qcoeff = BLOCK_OFFSET(p->qcoeff, block);
|
| - int16_t *dqcoeff = BLOCK_OFFSET(pd->dqcoeff, block);
|
| - const scan_order *scan_order;
|
| - uint16_t *eob = &p->eobs[block];
|
| + const struct macroblock_plane *const p = &x->plane[plane];
|
| + const struct macroblockd_plane *const pd = &xd->plane[plane];
|
| + const scan_order *const scan_order = &vp9_default_scan_orders[tx_size];
|
| + int16_t *const coeff = BLOCK_OFFSET(p->coeff, block);
|
| + int16_t *const qcoeff = BLOCK_OFFSET(p->qcoeff, block);
|
| + int16_t *const dqcoeff = BLOCK_OFFSET(pd->dqcoeff, block);
|
| + uint16_t *const eob = &p->eobs[block];
|
| const int diff_stride = 4 * num_4x4_blocks_wide_lookup[plane_bsize];
|
| int i, j;
|
| - int16_t *src_diff;
|
| + const int16_t *src_diff;
|
| txfrm_block_to_raster_xy(plane_bsize, tx_size, block, &i, &j);
|
| src_diff = &p->src_diff[4 * (j * diff_stride + i)];
|
|
|
| switch (tx_size) {
|
| case TX_32X32:
|
| - scan_order = &vp9_default_scan_orders[TX_32X32];
|
| - if (x->use_lp32x32fdct)
|
| - vp9_fdct32x32_rd(src_diff, coeff, diff_stride);
|
| - else
|
| - vp9_fdct32x32(src_diff, coeff, diff_stride);
|
| + fdct32x32(x->use_lp32x32fdct, src_diff, coeff, diff_stride);
|
| vp9_quantize_b_32x32(coeff, 1024, x->skip_block, p->zbin, p->round,
|
| p->quant, p->quant_shift, qcoeff, dqcoeff,
|
| pd->dequant, p->zbin_extra, eob, scan_order->scan,
|
| scan_order->iscan);
|
| break;
|
| case TX_16X16:
|
| - scan_order = &vp9_default_scan_orders[TX_16X16];
|
| vp9_fdct16x16(src_diff, coeff, diff_stride);
|
| vp9_quantize_b(coeff, 256, x->skip_block, p->zbin, p->round,
|
| p->quant, p->quant_shift, qcoeff, dqcoeff,
|
| @@ -385,7 +351,6 @@
|
| scan_order->scan, scan_order->iscan);
|
| break;
|
| case TX_8X8:
|
| - scan_order = &vp9_default_scan_orders[TX_8X8];
|
| vp9_fdct8x8(src_diff, coeff, diff_stride);
|
| vp9_quantize_b(coeff, 64, x->skip_block, p->zbin, p->round,
|
| p->quant, p->quant_shift, qcoeff, dqcoeff,
|
| @@ -393,7 +358,6 @@
|
| scan_order->scan, scan_order->iscan);
|
| break;
|
| case TX_4X4:
|
| - scan_order = &vp9_default_scan_orders[TX_4X4];
|
| x->fwd_txm4x4(src_diff, coeff, diff_stride);
|
| vp9_quantize_b(coeff, 16, x->skip_block, p->zbin, p->round,
|
| p->quant, p->quant_shift, qcoeff, dqcoeff,
|
| @@ -432,7 +396,7 @@
|
| vp9_xform_quant(x, plane, block, plane_bsize, tx_size);
|
|
|
| if (x->optimize && (!x->skip_recode || !x->skip_optimize)) {
|
| - vp9_optimize_b(plane, block, plane_bsize, tx_size, x, ctx);
|
| + optimize_b(plane, block, plane_bsize, tx_size, x, ctx);
|
| } else {
|
| ctx->ta[plane][i] = p->eobs[block] > 0;
|
| ctx->tl[plane][j] = p->eobs[block] > 0;
|
| @@ -466,8 +430,7 @@
|
| }
|
| static void encode_block_pass1(int plane, int block, BLOCK_SIZE plane_bsize,
|
| TX_SIZE tx_size, void *arg) {
|
| - struct encode_b_args *const args = arg;
|
| - MACROBLOCK *const x = args->x;
|
| + MACROBLOCK *const x = (MACROBLOCK *)arg;
|
| MACROBLOCKD *const xd = &x->e_mbd;
|
| struct macroblock_plane *const p = &x->plane[plane];
|
| struct macroblockd_plane *const pd = &xd->plane[plane];
|
| @@ -479,24 +442,14 @@
|
|
|
| vp9_xform_quant(x, plane, block, plane_bsize, tx_size);
|
|
|
| - if (p->eobs[block] == 0)
|
| - return;
|
| -
|
| - xd->itxm_add(dqcoeff, dst, pd->dst.stride, p->eobs[block]);
|
| + if (p->eobs[block] > 0)
|
| + xd->itxm_add(dqcoeff, dst, pd->dst.stride, p->eobs[block]);
|
| }
|
|
|
| -void vp9_encode_sby(MACROBLOCK *x, BLOCK_SIZE bsize) {
|
| - MACROBLOCKD *const xd = &x->e_mbd;
|
| - struct optimize_ctx ctx;
|
| - MB_MODE_INFO *mbmi = &xd->mi_8x8[0]->mbmi;
|
| - struct encode_b_args arg = {x, &ctx, &mbmi->skip};
|
| -
|
| - vp9_subtract_sby(x, bsize);
|
| - if (x->optimize)
|
| - optimize_init_b(0, bsize, &arg);
|
| -
|
| - vp9_foreach_transformed_block_in_plane(xd, bsize, 0, encode_block_pass1,
|
| - &arg);
|
| +void vp9_encode_sby_pass1(MACROBLOCK *x, BLOCK_SIZE bsize) {
|
| + vp9_subtract_plane(x, bsize, 0);
|
| + vp9_foreach_transformed_block_in_plane(&x->e_mbd, bsize, 0,
|
| + encode_block_pass1, x);
|
| }
|
|
|
| void vp9_encode_sb(MACROBLOCK *x, BLOCK_SIZE bsize) {
|
| @@ -504,17 +457,22 @@
|
| struct optimize_ctx ctx;
|
| MB_MODE_INFO *mbmi = &xd->mi_8x8[0]->mbmi;
|
| struct encode_b_args arg = {x, &ctx, &mbmi->skip};
|
| + int plane;
|
|
|
| - if (!x->skip_recode)
|
| - vp9_subtract_sb(x, bsize);
|
| + for (plane = 0; plane < MAX_MB_PLANE; ++plane) {
|
| + if (!x->skip_recode)
|
| + vp9_subtract_plane(x, bsize, plane);
|
|
|
| - if (x->optimize && (!x->skip_recode || !x->skip_optimize)) {
|
| - int i;
|
| - for (i = 0; i < MAX_MB_PLANE; ++i)
|
| - optimize_init_b(i, bsize, &arg);
|
| - }
|
| + if (x->optimize && (!x->skip_recode || !x->skip_optimize)) {
|
| + const struct macroblockd_plane* const pd = &xd->plane[plane];
|
| + const TX_SIZE tx_size = plane ? get_uv_tx_size(mbmi) : mbmi->tx_size;
|
| + vp9_get_entropy_contexts(bsize, tx_size, pd,
|
| + ctx.ta[plane], ctx.tl[plane]);
|
| + }
|
|
|
| - vp9_foreach_transformed_block(xd, bsize, encode_block, &arg);
|
| + vp9_foreach_transformed_block_in_plane(xd, bsize, plane, encode_block,
|
| + &arg);
|
| + }
|
| }
|
|
|
| static void encode_block_intra(int plane, int block, BLOCK_SIZE plane_bsize,
|
| @@ -536,14 +494,16 @@
|
| uint8_t *src, *dst;
|
| int16_t *src_diff;
|
| uint16_t *eob = &p->eobs[block];
|
| + const int src_stride = p->src.stride;
|
| + const int dst_stride = pd->dst.stride;
|
| int i, j;
|
| txfrm_block_to_raster_xy(plane_bsize, tx_size, block, &i, &j);
|
| - dst = &pd->dst.buf[4 * (j * pd->dst.stride + i)];
|
| - src = &p->src.buf[4 * (j * p->src.stride + i)];
|
| + dst = &pd->dst.buf[4 * (j * dst_stride + i)];
|
| + src = &p->src.buf[4 * (j * src_stride + i)];
|
| src_diff = &p->src_diff[4 * (j * diff_stride + i)];
|
|
|
| // if (x->optimize)
|
| - // vp9_optimize_b(plane, block, plane_bsize, tx_size, x, args->ctx);
|
| + // optimize_b(plane, block, plane_bsize, tx_size, x, args->ctx);
|
|
|
| switch (tx_size) {
|
| case TX_32X32:
|
| @@ -551,22 +511,19 @@
|
| mode = plane == 0 ? mbmi->mode : mbmi->uv_mode;
|
| vp9_predict_intra_block(xd, block >> 6, bwl, TX_32X32, mode,
|
| x->skip_encode ? src : dst,
|
| - x->skip_encode ? p->src.stride : pd->dst.stride,
|
| - dst, pd->dst.stride, i, j, plane);
|
| + x->skip_encode ? src_stride : dst_stride,
|
| + dst, dst_stride, i, j, plane);
|
| if (!x->skip_recode) {
|
| vp9_subtract_block(32, 32, src_diff, diff_stride,
|
| - src, p->src.stride, dst, pd->dst.stride);
|
| - if (x->use_lp32x32fdct)
|
| - vp9_fdct32x32_rd(src_diff, coeff, diff_stride);
|
| - else
|
| - vp9_fdct32x32(src_diff, coeff, diff_stride);
|
| + src, src_stride, dst, dst_stride);
|
| + fdct32x32(x->use_lp32x32fdct, src_diff, coeff, diff_stride);
|
| vp9_quantize_b_32x32(coeff, 1024, x->skip_block, p->zbin, p->round,
|
| p->quant, p->quant_shift, qcoeff, dqcoeff,
|
| pd->dequant, p->zbin_extra, eob, scan_order->scan,
|
| scan_order->iscan);
|
| }
|
| if (!x->skip_encode && *eob)
|
| - vp9_idct32x32_add(dqcoeff, dst, pd->dst.stride, *eob);
|
| + vp9_idct32x32_add(dqcoeff, dst, dst_stride, *eob);
|
| break;
|
| case TX_16X16:
|
| tx_type = get_tx_type_16x16(pd->plane_type, xd);
|
| @@ -574,11 +531,11 @@
|
| mode = plane == 0 ? mbmi->mode : mbmi->uv_mode;
|
| vp9_predict_intra_block(xd, block >> 4, bwl, TX_16X16, mode,
|
| x->skip_encode ? src : dst,
|
| - x->skip_encode ? p->src.stride : pd->dst.stride,
|
| - dst, pd->dst.stride, i, j, plane);
|
| + x->skip_encode ? src_stride : dst_stride,
|
| + dst, dst_stride, i, j, plane);
|
| if (!x->skip_recode) {
|
| vp9_subtract_block(16, 16, src_diff, diff_stride,
|
| - src, p->src.stride, dst, pd->dst.stride);
|
| + src, src_stride, dst, dst_stride);
|
| vp9_fht16x16(src_diff, coeff, diff_stride, tx_type);
|
| vp9_quantize_b(coeff, 256, x->skip_block, p->zbin, p->round,
|
| p->quant, p->quant_shift, qcoeff, dqcoeff,
|
| @@ -586,7 +543,7 @@
|
| scan_order->iscan);
|
| }
|
| if (!x->skip_encode && *eob)
|
| - vp9_iht16x16_add(tx_type, dqcoeff, dst, pd->dst.stride, *eob);
|
| + vp9_iht16x16_add(tx_type, dqcoeff, dst, dst_stride, *eob);
|
| break;
|
| case TX_8X8:
|
| tx_type = get_tx_type_8x8(pd->plane_type, xd);
|
| @@ -594,11 +551,11 @@
|
| mode = plane == 0 ? mbmi->mode : mbmi->uv_mode;
|
| vp9_predict_intra_block(xd, block >> 2, bwl, TX_8X8, mode,
|
| x->skip_encode ? src : dst,
|
| - x->skip_encode ? p->src.stride : pd->dst.stride,
|
| - dst, pd->dst.stride, i, j, plane);
|
| + x->skip_encode ? src_stride : dst_stride,
|
| + dst, dst_stride, i, j, plane);
|
| if (!x->skip_recode) {
|
| vp9_subtract_block(8, 8, src_diff, diff_stride,
|
| - src, p->src.stride, dst, pd->dst.stride);
|
| + src, src_stride, dst, dst_stride);
|
| vp9_fht8x8(src_diff, coeff, diff_stride, tx_type);
|
| vp9_quantize_b(coeff, 64, x->skip_block, p->zbin, p->round, p->quant,
|
| p->quant_shift, qcoeff, dqcoeff,
|
| @@ -606,24 +563,20 @@
|
| scan_order->iscan);
|
| }
|
| if (!x->skip_encode && *eob)
|
| - vp9_iht8x8_add(tx_type, dqcoeff, dst, pd->dst.stride, *eob);
|
| + vp9_iht8x8_add(tx_type, dqcoeff, dst, dst_stride, *eob);
|
| break;
|
| case TX_4X4:
|
| tx_type = get_tx_type_4x4(pd->plane_type, xd, block);
|
| scan_order = &vp9_scan_orders[TX_4X4][tx_type];
|
| - if (mbmi->sb_type < BLOCK_8X8 && plane == 0)
|
| - mode = xd->mi_8x8[0]->bmi[block].as_mode;
|
| - else
|
| - mode = plane == 0 ? mbmi->mode : mbmi->uv_mode;
|
| -
|
| + mode = plane == 0 ? get_y_mode(xd->mi_8x8[0], block) : mbmi->uv_mode;
|
| vp9_predict_intra_block(xd, block, bwl, TX_4X4, mode,
|
| x->skip_encode ? src : dst,
|
| - x->skip_encode ? p->src.stride : pd->dst.stride,
|
| - dst, pd->dst.stride, i, j, plane);
|
| + x->skip_encode ? src_stride : dst_stride,
|
| + dst, dst_stride, i, j, plane);
|
|
|
| if (!x->skip_recode) {
|
| vp9_subtract_block(4, 4, src_diff, diff_stride,
|
| - src, p->src.stride, dst, pd->dst.stride);
|
| + src, src_stride, dst, dst_stride);
|
| if (tx_type != DCT_DCT)
|
| vp9_fht4x4(src_diff, coeff, diff_stride, tx_type);
|
| else
|
| @@ -639,9 +592,9 @@
|
| // this is like vp9_short_idct4x4 but has a special case around eob<=1
|
| // which is significant (not just an optimization) for the lossless
|
| // case.
|
| - xd->itxm_add(dqcoeff, dst, pd->dst.stride, *eob);
|
| + xd->itxm_add(dqcoeff, dst, dst_stride, *eob);
|
| else
|
| - vp9_iht4x4_16_add(dqcoeff, dst, pd->dst.stride, tx_type);
|
| + vp9_iht4x4_16_add(dqcoeff, dst, dst_stride, tx_type);
|
| }
|
| break;
|
| default:
|
|
|