| Index: source/libvpx/vp9/encoder/vp9_temporal_filter.c
|
| ===================================================================
|
| --- source/libvpx/vp9/encoder/vp9_temporal_filter.c (revision 292608)
|
| +++ source/libvpx/vp9/encoder/vp9_temporal_filter.c (working copy)
|
| @@ -56,6 +56,34 @@
|
| mv_precision_uv = MV_PRECISION_Q3;
|
| }
|
|
|
| +#if CONFIG_VP9_HIGHBITDEPTH
|
| + if (xd->cur_buf->flags & YV12_FLAG_HIGHBITDEPTH) {
|
| + vp9_highbd_build_inter_predictor(y_mb_ptr, stride,
|
| + &pred[0], 16,
|
| + &mv,
|
| + scale,
|
| + 16, 16,
|
| + which_mv,
|
| + kernel, MV_PRECISION_Q3, x, y, xd->bd);
|
| +
|
| + vp9_highbd_build_inter_predictor(u_mb_ptr, uv_stride,
|
| + &pred[256], uv_block_width,
|
| + &mv,
|
| + scale,
|
| + uv_block_width, uv_block_height,
|
| + which_mv,
|
| + kernel, mv_precision_uv, x, y, xd->bd);
|
| +
|
| + vp9_highbd_build_inter_predictor(v_mb_ptr, uv_stride,
|
| + &pred[512], uv_block_width,
|
| + &mv,
|
| + scale,
|
| + uv_block_width, uv_block_height,
|
| + which_mv,
|
| + kernel, mv_precision_uv, x, y, xd->bd);
|
| + return;
|
| + }
|
| +#endif // CONFIG_VP9_HIGHBITDEPTH
|
| vp9_build_inter_predictor(y_mb_ptr, stride,
|
| &pred[0], 16,
|
| &mv,
|
| @@ -133,6 +161,54 @@
|
| }
|
| }
|
|
|
| +#if CONFIG_VP9_HIGHBITDEPTH
|
| +void vp9_highbd_temporal_filter_apply_c(uint8_t *frame1_8,
|
| + unsigned int stride,
|
| + uint8_t *frame2_8,
|
| + unsigned int block_width,
|
| + unsigned int block_height,
|
| + int strength,
|
| + int filter_weight,
|
| + unsigned int *accumulator,
|
| + uint16_t *count) {
|
| + uint16_t *frame1 = CONVERT_TO_SHORTPTR(frame1_8);
|
| + uint16_t *frame2 = CONVERT_TO_SHORTPTR(frame2_8);
|
| + unsigned int i, j, k;
|
| + int modifier;
|
| + int byte = 0;
|
| + const int rounding = strength > 0 ? 1 << (strength - 1) : 0;
|
| +
|
| + for (i = 0, k = 0; i < block_height; i++) {
|
| + for (j = 0; j < block_width; j++, k++) {
|
| + int src_byte = frame1[byte];
|
| + int pixel_value = *frame2++;
|
| +
|
| + modifier = src_byte - pixel_value;
|
| + // This is an integer approximation of:
|
| + // float coeff = (3.0 * modifer * modifier) / pow(2, strength);
|
| + // modifier = (int)roundf(coeff > 16 ? 0 : 16-coeff);
|
| + modifier *= modifier;
|
| + modifier *= 3;
|
| + modifier += rounding;
|
| + modifier >>= strength;
|
| +
|
| + if (modifier > 16)
|
| + modifier = 16;
|
| +
|
| + modifier = 16 - modifier;
|
| + modifier *= filter_weight;
|
| +
|
| + count[k] += modifier;
|
| + accumulator[k] += modifier * pixel_value;
|
| +
|
| + byte++;
|
| + }
|
| +
|
| + byte += stride - block_width;
|
| + }
|
| +}
|
| +#endif // CONFIG_VP9_HIGHBITDEPTH
|
| +
|
| static int temporal_filter_find_matching_mb_c(VP9_COMP *cpi,
|
| uint8_t *arf_frame_buf,
|
| uint8_t *frame_ptr_buf,
|
| @@ -145,7 +221,7 @@
|
| int bestsme = INT_MAX;
|
| int distortion;
|
| unsigned int sse;
|
| - int sad_list[5];
|
| + int cost_list[5];
|
|
|
| MV best_ref_mv1 = {0, 0};
|
| MV best_ref_mv1_full; /* full-pixel value of best_ref_mv1 */
|
| @@ -169,7 +245,7 @@
|
|
|
| // Ignore mv costing by sending NULL pointer instead of cost arrays
|
| vp9_hex_search(x, &best_ref_mv1_full, step_param, sadpb, 1,
|
| - cond_sad_list(cpi, sad_list),
|
| + cond_cost_list(cpi, cost_list),
|
| &cpi->fn_ptr[BLOCK_16X16], 0, &best_ref_mv1, ref_mv);
|
|
|
| // Ignore mv costing by sending NULL pointer instead of cost array
|
| @@ -179,7 +255,7 @@
|
| x->errorperbit,
|
| &cpi->fn_ptr[BLOCK_16X16],
|
| 0, mv_sf->subpel_iters_per_step,
|
| - cond_sad_list(cpi, sad_list),
|
| + cond_cost_list(cpi, cost_list),
|
| NULL, NULL,
|
| &distortion, &sse, NULL, 0, 0);
|
|
|
| @@ -209,7 +285,13 @@
|
| MACROBLOCKD *mbd = &cpi->mb.e_mbd;
|
| YV12_BUFFER_CONFIG *f = frames[alt_ref_index];
|
| uint8_t *dst1, *dst2;
|
| +#if CONFIG_VP9_HIGHBITDEPTH
|
| + DECLARE_ALIGNED_ARRAY(16, uint16_t, predictor16, 16 * 16 * 3);
|
| + DECLARE_ALIGNED_ARRAY(16, uint8_t, predictor8, 16 * 16 * 3);
|
| + uint8_t *predictor;
|
| +#else
|
| DECLARE_ALIGNED_ARRAY(16, uint8_t, predictor, 16 * 16 * 3);
|
| +#endif
|
| const int mb_uv_height = 16 >> mbd->plane[1].subsampling_y;
|
| const int mb_uv_width = 16 >> mbd->plane[1].subsampling_x;
|
|
|
| @@ -216,6 +298,13 @@
|
| // Save input state
|
| uint8_t* input_buffer[MAX_MB_PLANE];
|
| int i;
|
| +#if CONFIG_VP9_HIGHBITDEPTH
|
| + if (mbd->cur_buf->flags & YV12_FLAG_HIGHBITDEPTH) {
|
| + predictor = CONVERT_TO_BYTEPTR(predictor16);
|
| + } else {
|
| + predictor = predictor8;
|
| + }
|
| +#endif
|
|
|
| for (i = 0; i < MAX_MB_PLANE; i++)
|
| input_buffer[i] = mbd->plane[i].pre[0].buf;
|
| @@ -286,6 +375,44 @@
|
| predictor, scale,
|
| mb_col * 16, mb_row * 16);
|
|
|
| +#if CONFIG_VP9_HIGHBITDEPTH
|
| + if (mbd->cur_buf->flags & YV12_FLAG_HIGHBITDEPTH) {
|
| + int adj_strength = strength + 2 * (mbd->bd - 8);
|
| + // Apply the filter (YUV)
|
| + vp9_highbd_temporal_filter_apply(f->y_buffer + mb_y_offset,
|
| + f->y_stride,
|
| + predictor, 16, 16, adj_strength,
|
| + filter_weight,
|
| + accumulator, count);
|
| + vp9_highbd_temporal_filter_apply(f->u_buffer + mb_uv_offset,
|
| + f->uv_stride, predictor + 256,
|
| + mb_uv_width, mb_uv_height,
|
| + adj_strength,
|
| + filter_weight, accumulator + 256,
|
| + count + 256);
|
| + vp9_highbd_temporal_filter_apply(f->v_buffer + mb_uv_offset,
|
| + f->uv_stride, predictor + 512,
|
| + mb_uv_width, mb_uv_height,
|
| + adj_strength, filter_weight,
|
| + accumulator + 512, count + 512);
|
| + } else {
|
| + // Apply the filter (YUV)
|
| + vp9_temporal_filter_apply(f->y_buffer + mb_y_offset, f->y_stride,
|
| + predictor, 16, 16,
|
| + strength, filter_weight,
|
| + accumulator, count);
|
| + vp9_temporal_filter_apply(f->u_buffer + mb_uv_offset, f->uv_stride,
|
| + predictor + 256,
|
| + mb_uv_width, mb_uv_height, strength,
|
| + filter_weight, accumulator + 256,
|
| + count + 256);
|
| + vp9_temporal_filter_apply(f->v_buffer + mb_uv_offset, f->uv_stride,
|
| + predictor + 512,
|
| + mb_uv_width, mb_uv_height, strength,
|
| + filter_weight, accumulator + 512,
|
| + count + 512);
|
| + }
|
| +#else
|
| // Apply the filter (YUV)
|
| vp9_temporal_filter_apply(f->y_buffer + mb_y_offset, f->y_stride,
|
| predictor, 16, 16,
|
| @@ -301,9 +428,108 @@
|
| mb_uv_width, mb_uv_height, strength,
|
| filter_weight, accumulator + 512,
|
| count + 512);
|
| +#endif // CONFIG_VP9_HIGHBITDEPTH
|
| }
|
| }
|
|
|
| +#if CONFIG_VP9_HIGHBITDEPTH
|
| + if (mbd->cur_buf->flags & YV12_FLAG_HIGHBITDEPTH) {
|
| + uint16_t *dst1_16;
|
| + uint16_t *dst2_16;
|
| + // Normalize filter output to produce AltRef frame
|
| + dst1 = cpi->alt_ref_buffer.y_buffer;
|
| + dst1_16 = CONVERT_TO_SHORTPTR(dst1);
|
| + stride = cpi->alt_ref_buffer.y_stride;
|
| + byte = mb_y_offset;
|
| + for (i = 0, k = 0; i < 16; i++) {
|
| + for (j = 0; j < 16; j++, k++) {
|
| + unsigned int pval = accumulator[k] + (count[k] >> 1);
|
| + pval *= fixed_divide[count[k]];
|
| + pval >>= 19;
|
| +
|
| + dst1_16[byte] = (uint16_t)pval;
|
| +
|
| + // move to next pixel
|
| + byte++;
|
| + }
|
| +
|
| + byte += stride - 16;
|
| + }
|
| +
|
| + dst1 = cpi->alt_ref_buffer.u_buffer;
|
| + dst2 = cpi->alt_ref_buffer.v_buffer;
|
| + dst1_16 = CONVERT_TO_SHORTPTR(dst1);
|
| + dst2_16 = CONVERT_TO_SHORTPTR(dst2);
|
| + stride = cpi->alt_ref_buffer.uv_stride;
|
| + byte = mb_uv_offset;
|
| + for (i = 0, k = 256; i < mb_uv_height; i++) {
|
| + for (j = 0; j < mb_uv_width; j++, k++) {
|
| + int m = k + 256;
|
| +
|
| + // U
|
| + unsigned int pval = accumulator[k] + (count[k] >> 1);
|
| + pval *= fixed_divide[count[k]];
|
| + pval >>= 19;
|
| + dst1_16[byte] = (uint16_t)pval;
|
| +
|
| + // V
|
| + pval = accumulator[m] + (count[m] >> 1);
|
| + pval *= fixed_divide[count[m]];
|
| + pval >>= 19;
|
| + dst2_16[byte] = (uint16_t)pval;
|
| +
|
| + // move to next pixel
|
| + byte++;
|
| + }
|
| +
|
| + byte += stride - mb_uv_width;
|
| + }
|
| + } else {
|
| + // Normalize filter output to produce AltRef frame
|
| + dst1 = cpi->alt_ref_buffer.y_buffer;
|
| + stride = cpi->alt_ref_buffer.y_stride;
|
| + byte = mb_y_offset;
|
| + for (i = 0, k = 0; i < 16; i++) {
|
| + for (j = 0; j < 16; j++, k++) {
|
| + unsigned int pval = accumulator[k] + (count[k] >> 1);
|
| + pval *= fixed_divide[count[k]];
|
| + pval >>= 19;
|
| +
|
| + dst1[byte] = (uint8_t)pval;
|
| +
|
| + // move to next pixel
|
| + byte++;
|
| + }
|
| + byte += stride - 16;
|
| + }
|
| +
|
| + dst1 = cpi->alt_ref_buffer.u_buffer;
|
| + dst2 = cpi->alt_ref_buffer.v_buffer;
|
| + stride = cpi->alt_ref_buffer.uv_stride;
|
| + byte = mb_uv_offset;
|
| + for (i = 0, k = 256; i < mb_uv_height; i++) {
|
| + for (j = 0; j < mb_uv_width; j++, k++) {
|
| + int m = k + 256;
|
| +
|
| + // U
|
| + unsigned int pval = accumulator[k] + (count[k] >> 1);
|
| + pval *= fixed_divide[count[k]];
|
| + pval >>= 19;
|
| + dst1[byte] = (uint8_t)pval;
|
| +
|
| + // V
|
| + pval = accumulator[m] + (count[m] >> 1);
|
| + pval *= fixed_divide[count[m]];
|
| + pval >>= 19;
|
| + dst2[byte] = (uint8_t)pval;
|
| +
|
| + // move to next pixel
|
| + byte++;
|
| + }
|
| + byte += stride - mb_uv_width;
|
| + }
|
| + }
|
| +#else
|
| // Normalize filter output to produce AltRef frame
|
| dst1 = cpi->alt_ref_buffer.y_buffer;
|
| stride = cpi->alt_ref_buffer.y_stride;
|
| @@ -347,6 +573,7 @@
|
| }
|
| byte += stride - mb_uv_width;
|
| }
|
| +#endif // CONFIG_VP9_HIGHBITDEPTH
|
| mb_y_offset += 16;
|
| mb_uv_offset += mb_uv_width;
|
| }
|
| @@ -449,60 +676,69 @@
|
| frames[frames_to_blur - 1 - frame] = &buf->img;
|
| }
|
|
|
| - // Setup scaling factors. Scaling on each of the arnr frames is not supported
|
| - if (is_two_pass_svc(cpi)) {
|
| - // In spatial svc the scaling factors might be less then 1/2. So we will use
|
| - // non-normative scaling.
|
| - int frame_used = 0;
|
| + if (frames_to_blur > 0) {
|
| + // Setup scaling factors. Scaling on each of the arnr frames is not
|
| + // supported.
|
| + if (is_two_pass_svc(cpi)) {
|
| + // In spatial svc the scaling factors might be less then 1/2.
|
| + // So we will use non-normative scaling.
|
| + int frame_used = 0;
|
| #if CONFIG_VP9_HIGHBITDEPTH
|
| - vp9_setup_scale_factors_for_frame(&sf,
|
| - get_frame_new_buffer(cm)->y_crop_width,
|
| - get_frame_new_buffer(cm)->y_crop_height,
|
| - get_frame_new_buffer(cm)->y_crop_width,
|
| - get_frame_new_buffer(cm)->y_crop_height,
|
| - cm->use_highbitdepth);
|
| + vp9_setup_scale_factors_for_frame(
|
| + &sf,
|
| + get_frame_new_buffer(cm)->y_crop_width,
|
| + get_frame_new_buffer(cm)->y_crop_height,
|
| + get_frame_new_buffer(cm)->y_crop_width,
|
| + get_frame_new_buffer(cm)->y_crop_height,
|
| + cm->use_highbitdepth);
|
| #else
|
| - vp9_setup_scale_factors_for_frame(&sf,
|
| - get_frame_new_buffer(cm)->y_crop_width,
|
| - get_frame_new_buffer(cm)->y_crop_height,
|
| - get_frame_new_buffer(cm)->y_crop_width,
|
| - get_frame_new_buffer(cm)->y_crop_height);
|
| -#endif
|
| - for (frame = 0; frame < frames_to_blur; ++frame) {
|
| - if (cm->mi_cols * MI_SIZE != frames[frame]->y_width ||
|
| - cm->mi_rows * MI_SIZE != frames[frame]->y_height) {
|
| - if (vp9_realloc_frame_buffer(&cpi->svc.scaled_frames[frame_used],
|
| - cm->width, cm->height,
|
| - cm->subsampling_x, cm->subsampling_y,
|
| + vp9_setup_scale_factors_for_frame(
|
| + &sf,
|
| + get_frame_new_buffer(cm)->y_crop_width,
|
| + get_frame_new_buffer(cm)->y_crop_height,
|
| + get_frame_new_buffer(cm)->y_crop_width,
|
| + get_frame_new_buffer(cm)->y_crop_height);
|
| +#endif // CONFIG_VP9_HIGHBITDEPTH
|
| +
|
| + for (frame = 0; frame < frames_to_blur; ++frame) {
|
| + if (cm->mi_cols * MI_SIZE != frames[frame]->y_width ||
|
| + cm->mi_rows * MI_SIZE != frames[frame]->y_height) {
|
| + if (vp9_realloc_frame_buffer(&cpi->svc.scaled_frames[frame_used],
|
| + cm->width, cm->height,
|
| + cm->subsampling_x, cm->subsampling_y,
|
| #if CONFIG_VP9_HIGHBITDEPTH
|
| - cm->use_highbitdepth,
|
| + cm->use_highbitdepth,
|
| #endif
|
| - VP9_ENC_BORDER_IN_PIXELS, NULL, NULL,
|
| - NULL))
|
| - vpx_internal_error(&cm->error, VPX_CODEC_MEM_ERROR,
|
| - "Failed to reallocate alt_ref_buffer");
|
| -
|
| - frames[frame] = vp9_scale_if_required(cm, frames[frame],
|
| - &cpi->svc.scaled_frames[frame_used]);
|
| - ++frame_used;
|
| + VP9_ENC_BORDER_IN_PIXELS, NULL, NULL,
|
| + NULL)) {
|
| + vpx_internal_error(&cm->error, VPX_CODEC_MEM_ERROR,
|
| + "Failed to reallocate alt_ref_buffer");
|
| + }
|
| + frames[frame] = vp9_scale_if_required(
|
| + cm, frames[frame], &cpi->svc.scaled_frames[frame_used]);
|
| + ++frame_used;
|
| + }
|
| }
|
| - }
|
| - } else {
|
| - // ARF is produced at the native frame size and resized when coded.
|
| + cm->mi = cm->mip + cm->mi_stride + 1;
|
| + cpi->mb.e_mbd.mi = cm->mi;
|
| + cpi->mb.e_mbd.mi[0].src_mi = &cpi->mb.e_mbd.mi[0];
|
| + } else {
|
| + // ARF is produced at the native frame size and resized when coded.
|
| #if CONFIG_VP9_HIGHBITDEPTH
|
| - vp9_setup_scale_factors_for_frame(&sf,
|
| - frames[0]->y_crop_width,
|
| - frames[0]->y_crop_height,
|
| - frames[0]->y_crop_width,
|
| - frames[0]->y_crop_height,
|
| - cm->use_highbitdepth);
|
| + vp9_setup_scale_factors_for_frame(&sf,
|
| + frames[0]->y_crop_width,
|
| + frames[0]->y_crop_height,
|
| + frames[0]->y_crop_width,
|
| + frames[0]->y_crop_height,
|
| + cm->use_highbitdepth);
|
| #else
|
| - vp9_setup_scale_factors_for_frame(&sf,
|
| - frames[0]->y_crop_width,
|
| - frames[0]->y_crop_height,
|
| - frames[0]->y_crop_width,
|
| - frames[0]->y_crop_height);
|
| -#endif
|
| + vp9_setup_scale_factors_for_frame(&sf,
|
| + frames[0]->y_crop_width,
|
| + frames[0]->y_crop_height,
|
| + frames[0]->y_crop_width,
|
| + frames[0]->y_crop_height);
|
| +#endif // CONFIG_VP9_HIGHBITDEPTH
|
| + }
|
| }
|
|
|
| temporal_filter_iterate_c(cpi, frames, frames_to_blur,
|
|
|