Index: source/libvpx/vp9/common/vp9_reconintra.c |
=================================================================== |
--- source/libvpx/vp9/common/vp9_reconintra.c (revision 292072) |
+++ source/libvpx/vp9/common/vp9_reconintra.c (working copy) |
@@ -40,12 +40,290 @@ |
type##_predictor(dst, stride, size, above, left); \ |
} |
+#if CONFIG_VP9_HIGHBITDEPTH |
+#define intra_pred_high_sized(type, size) \ |
+ void vp9_high_##type##_predictor_##size##x##size##_c( \ |
+ uint16_t *dst, ptrdiff_t stride, const uint16_t *above, \ |
+ const uint16_t *left, int bd) { \ |
+ high_##type##_predictor(dst, stride, size, above, left, bd); \ |
+ } |
+ |
#define intra_pred_allsizes(type) \ |
intra_pred_sized(type, 4) \ |
intra_pred_sized(type, 8) \ |
intra_pred_sized(type, 16) \ |
+ intra_pred_sized(type, 32) \ |
+ intra_pred_high_sized(type, 4) \ |
+ intra_pred_high_sized(type, 8) \ |
+ intra_pred_high_sized(type, 16) \ |
+ intra_pred_high_sized(type, 32) |
+ |
+#else |
+ |
+#define intra_pred_allsizes(type) \ |
+ intra_pred_sized(type, 4) \ |
+ intra_pred_sized(type, 8) \ |
+ intra_pred_sized(type, 16) \ |
intra_pred_sized(type, 32) |
+#endif // CONFIG_VP9_HIGHBITDEPTH |
+#if CONFIG_VP9_HIGHBITDEPTH |
+static INLINE void high_d207_predictor(uint16_t *dst, ptrdiff_t stride, int bs, |
+ const uint16_t *above, |
+ const uint16_t *left, int bd) { |
+ int r, c; |
+ (void) above; |
+ (void) bd; |
+ |
+ // First column. |
+ for (r = 0; r < bs - 1; ++r) { |
+ dst[r * stride] = ROUND_POWER_OF_TWO(left[r] + left[r + 1], 1); |
+ } |
+ dst[(bs - 1) * stride] = left[bs - 1]; |
+ dst++; |
+ |
+ // Second column. |
+ for (r = 0; r < bs - 2; ++r) { |
+ dst[r * stride] = ROUND_POWER_OF_TWO(left[r] + left[r + 1] * 2 + |
+ left[r + 2], 2); |
+ } |
+ dst[(bs - 2) * stride] = ROUND_POWER_OF_TWO(left[bs - 2] + |
+ left[bs - 1] * 3, 2); |
+ dst[(bs - 1) * stride] = left[bs - 1]; |
+ dst++; |
+ |
+ // Rest of last row. |
+ for (c = 0; c < bs - 2; ++c) |
+ dst[(bs - 1) * stride + c] = left[bs - 1]; |
+ |
+ for (r = bs - 2; r >= 0; --r) { |
+ for (c = 0; c < bs - 2; ++c) |
+ dst[r * stride + c] = dst[(r + 1) * stride + c - 2]; |
+ } |
+} |
+ |
+static INLINE void high_d63_predictor(uint16_t *dst, ptrdiff_t stride, int bs, |
+ const uint16_t *above, |
+ const uint16_t *left, int bd) { |
+ int r, c; |
+ (void) left; |
+ (void) bd; |
+ for (r = 0; r < bs; ++r) { |
+ for (c = 0; c < bs; ++c) { |
+ dst[c] = r & 1 ? ROUND_POWER_OF_TWO(above[r/2 + c] + |
+ above[r/2 + c + 1] * 2 + |
+ above[r/2 + c + 2], 2) |
+ : ROUND_POWER_OF_TWO(above[r/2 + c] + |
+ above[r/2 + c + 1], 1); |
+ } |
+ dst += stride; |
+ } |
+} |
+ |
+static INLINE void high_d45_predictor(uint16_t *dst, ptrdiff_t stride, int bs, |
+ const uint16_t *above, |
+ const uint16_t *left, int bd) { |
+ int r, c; |
+ (void) left; |
+ (void) bd; |
+ for (r = 0; r < bs; ++r) { |
+ for (c = 0; c < bs; ++c) { |
+ dst[c] = r + c + 2 < bs * 2 ? ROUND_POWER_OF_TWO(above[r + c] + |
+ above[r + c + 1] * 2 + |
+ above[r + c + 2], 2) |
+ : above[bs * 2 - 1]; |
+ } |
+ dst += stride; |
+ } |
+} |
+ |
+static INLINE void high_d117_predictor(uint16_t *dst, ptrdiff_t stride, |
+ int bs, const uint16_t *above, |
+ const uint16_t *left, int bd) { |
+ int r, c; |
+ (void) bd; |
+ |
+ // first row |
+ for (c = 0; c < bs; c++) |
+ dst[c] = ROUND_POWER_OF_TWO(above[c - 1] + above[c], 1); |
+ dst += stride; |
+ |
+ // second row |
+ dst[0] = ROUND_POWER_OF_TWO(left[0] + above[-1] * 2 + above[0], 2); |
+ for (c = 1; c < bs; c++) |
+ dst[c] = ROUND_POWER_OF_TWO(above[c - 2] + above[c - 1] * 2 + above[c], 2); |
+ dst += stride; |
+ |
+ // the rest of first col |
+ dst[0] = ROUND_POWER_OF_TWO(above[-1] + left[0] * 2 + left[1], 2); |
+ for (r = 3; r < bs; ++r) |
+ dst[(r - 2) * stride] = ROUND_POWER_OF_TWO(left[r - 3] + left[r - 2] * 2 + |
+ left[r - 1], 2); |
+ |
+ // the rest of the block |
+ for (r = 2; r < bs; ++r) { |
+ for (c = 1; c < bs; c++) |
+ dst[c] = dst[-2 * stride + c - 1]; |
+ dst += stride; |
+ } |
+} |
+ |
+static INLINE void high_d135_predictor(uint16_t *dst, ptrdiff_t stride, int bs, |
+ const uint16_t *above, |
+ const uint16_t *left, int bd) { |
+ int r, c; |
+ (void) bd; |
+ dst[0] = ROUND_POWER_OF_TWO(left[0] + above[-1] * 2 + above[0], 2); |
+ for (c = 1; c < bs; c++) |
+ dst[c] = ROUND_POWER_OF_TWO(above[c - 2] + above[c - 1] * 2 + above[c], 2); |
+ |
+ dst[stride] = ROUND_POWER_OF_TWO(above[-1] + left[0] * 2 + left[1], 2); |
+ for (r = 2; r < bs; ++r) |
+ dst[r * stride] = ROUND_POWER_OF_TWO(left[r - 2] + left[r - 1] * 2 + |
+ left[r], 2); |
+ |
+ dst += stride; |
+ for (r = 1; r < bs; ++r) { |
+ for (c = 1; c < bs; c++) |
+ dst[c] = dst[-stride + c - 1]; |
+ dst += stride; |
+ } |
+} |
+ |
+static INLINE void high_d153_predictor(uint16_t *dst, ptrdiff_t stride, int bs, |
+ const uint16_t *above, |
+ const uint16_t *left, int bd) { |
+ int r, c; |
+ (void) bd; |
+ dst[0] = ROUND_POWER_OF_TWO(above[-1] + left[0], 1); |
+ for (r = 1; r < bs; r++) |
+ dst[r * stride] = ROUND_POWER_OF_TWO(left[r - 1] + left[r], 1); |
+ dst++; |
+ |
+ dst[0] = ROUND_POWER_OF_TWO(left[0] + above[-1] * 2 + above[0], 2); |
+ dst[stride] = ROUND_POWER_OF_TWO(above[-1] + left[0] * 2 + left[1], 2); |
+ for (r = 2; r < bs; r++) |
+ dst[r * stride] = ROUND_POWER_OF_TWO(left[r - 2] + left[r - 1] * 2 + |
+ left[r], 2); |
+ dst++; |
+ |
+ for (c = 0; c < bs - 2; c++) |
+ dst[c] = ROUND_POWER_OF_TWO(above[c - 1] + above[c] * 2 + above[c + 1], 2); |
+ dst += stride; |
+ |
+ for (r = 1; r < bs; ++r) { |
+ for (c = 0; c < bs - 2; c++) |
+ dst[c] = dst[-stride + c - 2]; |
+ dst += stride; |
+ } |
+} |
+ |
+static INLINE void high_v_predictor(uint16_t *dst, ptrdiff_t stride, int bs, |
+ const uint16_t *above, |
+ const uint16_t *left, int bd) { |
+ int r; |
+ (void) left; |
+ (void) bd; |
+ for (r = 0; r < bs; r++) { |
+ vpx_memcpy(dst, above, bs * sizeof(uint16_t)); |
+ dst += stride; |
+ } |
+} |
+ |
+static INLINE void high_h_predictor(uint16_t *dst, ptrdiff_t stride, int bs, |
+ const uint16_t *above, const uint16_t *left, |
+ int bd) { |
+ int r; |
+ (void) above; |
+ (void) bd; |
+ for (r = 0; r < bs; r++) { |
+ vpx_memset16(dst, left[r], bs); |
+ dst += stride; |
+ } |
+} |
+ |
+static INLINE void high_tm_predictor(uint16_t *dst, ptrdiff_t stride, int bs, |
+ const uint16_t *above, |
+ const uint16_t *left, int bd) { |
+ int r, c; |
+ int ytop_left = above[-1]; |
+ (void) bd; |
+ |
+ for (r = 0; r < bs; r++) { |
+ for (c = 0; c < bs; c++) |
+ dst[c] = clip_pixel_high(left[r] + above[c] - ytop_left, bd); |
+ dst += stride; |
+ } |
+} |
+ |
+static INLINE void high_dc_128_predictor(uint16_t *dst, ptrdiff_t stride, |
+ int bs, const uint16_t *above, |
+ const uint16_t *left, int bd) { |
+ int r; |
+ (void) above; |
+ (void) left; |
+ |
+ for (r = 0; r < bs; r++) { |
+ vpx_memset16(dst, 128 << (bd - 8), bs); |
+ dst += stride; |
+ } |
+} |
+ |
+static INLINE void high_dc_left_predictor(uint16_t *dst, ptrdiff_t stride, |
+ int bs, const uint16_t *above, |
+ const uint16_t *left, int bd) { |
+ int i, r, expected_dc, sum = 0; |
+ (void) above; |
+ (void) bd; |
+ |
+ for (i = 0; i < bs; i++) |
+ sum += left[i]; |
+ expected_dc = (sum + (bs >> 1)) / bs; |
+ |
+ for (r = 0; r < bs; r++) { |
+ vpx_memset16(dst, expected_dc, bs); |
+ dst += stride; |
+ } |
+} |
+ |
+static INLINE void high_dc_top_predictor(uint16_t *dst, ptrdiff_t stride, |
+ int bs, const uint16_t *above, |
+ const uint16_t *left, int bd) { |
+ int i, r, expected_dc, sum = 0; |
+ (void) left; |
+ (void) bd; |
+ |
+ for (i = 0; i < bs; i++) |
+ sum += above[i]; |
+ expected_dc = (sum + (bs >> 1)) / bs; |
+ |
+ for (r = 0; r < bs; r++) { |
+ vpx_memset16(dst, expected_dc, bs); |
+ dst += stride; |
+ } |
+} |
+ |
+static INLINE void high_dc_predictor(uint16_t *dst, ptrdiff_t stride, |
+ int bs, const uint16_t *above, |
+ const uint16_t *left, int bd) { |
+ int i, r, expected_dc, sum = 0; |
+ const int count = 2 * bs; |
+ (void) bd; |
+ |
+ for (i = 0; i < bs; i++) { |
+ sum += above[i]; |
+ sum += left[i]; |
+ } |
+ |
+ expected_dc = (sum + (count >> 1)) / count; |
+ |
+ for (r = 0; r < bs; r++) { |
+ vpx_memset16(dst, expected_dc, bs); |
+ dst += stride; |
+ } |
+} |
+#endif // CONFIG_VP9_HIGHBITDEPTH |
+ |
static INLINE void d207_predictor(uint8_t *dst, ptrdiff_t stride, int bs, |
const uint8_t *above, const uint8_t *left) { |
int r, c; |
@@ -293,6 +571,14 @@ |
static intra_pred_fn pred[INTRA_MODES][TX_SIZES]; |
static intra_pred_fn dc_pred[2][2][TX_SIZES]; |
+#if CONFIG_VP9_HIGHBITDEPTH |
+typedef void (*intra_high_pred_fn)(uint16_t *dst, ptrdiff_t stride, |
+ const uint16_t *above, const uint16_t *left, |
+ int bd); |
+static intra_high_pred_fn pred_high[INTRA_MODES][4]; |
+static intra_high_pred_fn dc_pred_high[2][2][4]; |
+#endif // CONFIG_VP9_HIGHBITDEPTH |
+ |
void vp9_init_intra_predictors() { |
#define INIT_ALL_SIZES(p, type) \ |
p[TX_4X4] = vp9_##type##_predictor_4x4; \ |
@@ -315,9 +601,164 @@ |
INIT_ALL_SIZES(dc_pred[1][0], dc_left); |
INIT_ALL_SIZES(dc_pred[1][1], dc); |
-#undef INIT_ALL_SIZES |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ INIT_ALL_SIZES(pred_high[V_PRED], high_v); |
+ INIT_ALL_SIZES(pred_high[H_PRED], high_h); |
+ INIT_ALL_SIZES(pred_high[D207_PRED], high_d207); |
+ INIT_ALL_SIZES(pred_high[D45_PRED], high_d45); |
+ INIT_ALL_SIZES(pred_high[D63_PRED], high_d63); |
+ INIT_ALL_SIZES(pred_high[D117_PRED], high_d117); |
+ INIT_ALL_SIZES(pred_high[D135_PRED], high_d135); |
+ INIT_ALL_SIZES(pred_high[D153_PRED], high_d153); |
+ INIT_ALL_SIZES(pred_high[TM_PRED], high_tm); |
+ |
+ INIT_ALL_SIZES(dc_pred_high[0][0], high_dc_128); |
+ INIT_ALL_SIZES(dc_pred_high[0][1], high_dc_top); |
+ INIT_ALL_SIZES(dc_pred_high[1][0], high_dc_left); |
+ INIT_ALL_SIZES(dc_pred_high[1][1], high_dc); |
+#endif // CONFIG_VP9_HIGHBITDEPTH |
+ |
+#undef intra_pred_allsizes |
} |
+#if CONFIG_VP9_HIGHBITDEPTH |
+static void build_intra_predictors_high(const MACROBLOCKD *xd, |
+ const uint8_t *ref8, |
+ int ref_stride, |
+ uint8_t *dst8, |
+ int dst_stride, |
+ PREDICTION_MODE mode, |
+ TX_SIZE tx_size, |
+ int up_available, |
+ int left_available, |
+ int right_available, |
+ int x, int y, |
+ int plane, int bd) { |
+ int i; |
+ uint16_t *dst = CONVERT_TO_SHORTPTR(dst8); |
+ uint16_t *ref = CONVERT_TO_SHORTPTR(ref8); |
+ DECLARE_ALIGNED_ARRAY(16, uint16_t, left_col, 64); |
+ DECLARE_ALIGNED_ARRAY(16, uint16_t, above_data, 128 + 16); |
+ uint16_t *above_row = above_data + 16; |
+ const uint16_t *const_above_row = above_row; |
+ const int bs = 4 << tx_size; |
+ int frame_width, frame_height; |
+ int x0, y0; |
+ const struct macroblockd_plane *const pd = &xd->plane[plane]; |
+ // int base=128; |
+ int base = 128 << (bd - 8); |
+ // 127 127 127 .. 127 127 127 127 127 127 |
+ // 129 A B .. Y Z |
+ // 129 C D .. W X |
+ // 129 E F .. U V |
+ // 129 G H .. S T T T T T |
+ |
+ // Get current frame pointer, width and height. |
+ if (plane == 0) { |
+ frame_width = xd->cur_buf->y_width; |
+ frame_height = xd->cur_buf->y_height; |
+ } else { |
+ frame_width = xd->cur_buf->uv_width; |
+ frame_height = xd->cur_buf->uv_height; |
+ } |
+ |
+ // Get block position in current frame. |
+ x0 = (-xd->mb_to_left_edge >> (3 + pd->subsampling_x)) + x; |
+ y0 = (-xd->mb_to_top_edge >> (3 + pd->subsampling_y)) + y; |
+ |
+ // left |
+ if (left_available) { |
+ if (xd->mb_to_bottom_edge < 0) { |
+ /* slower path if the block needs border extension */ |
+ if (y0 + bs <= frame_height) { |
+ for (i = 0; i < bs; ++i) |
+ left_col[i] = ref[i * ref_stride - 1]; |
+ } else { |
+ const int extend_bottom = frame_height - y0; |
+ for (i = 0; i < extend_bottom; ++i) |
+ left_col[i] = ref[i * ref_stride - 1]; |
+ for (; i < bs; ++i) |
+ left_col[i] = ref[(extend_bottom - 1) * ref_stride - 1]; |
+ } |
+ } else { |
+ /* faster path if the block does not need extension */ |
+ for (i = 0; i < bs; ++i) |
+ left_col[i] = ref[i * ref_stride - 1]; |
+ } |
+ } else { |
+ // TODO(Peter): this value should probably change for high bitdepth |
+ vpx_memset16(left_col, base + 1, bs); |
+ } |
+ |
+ // TODO(hkuang) do not extend 2*bs pixels for all modes. |
+ // above |
+ if (up_available) { |
+ const uint16_t *above_ref = ref - ref_stride; |
+ if (xd->mb_to_right_edge < 0) { |
+ /* slower path if the block needs border extension */ |
+ if (x0 + 2 * bs <= frame_width) { |
+ if (right_available && bs == 4) { |
+ vpx_memcpy(above_row, above_ref, 2 * bs * sizeof(uint16_t)); |
+ } else { |
+ vpx_memcpy(above_row, above_ref, bs * sizeof(uint16_t)); |
+ vpx_memset16(above_row + bs, above_row[bs - 1], bs); |
+ } |
+ } else if (x0 + bs <= frame_width) { |
+ const int r = frame_width - x0; |
+ if (right_available && bs == 4) { |
+ vpx_memcpy(above_row, above_ref, r * sizeof(uint16_t)); |
+ vpx_memset16(above_row + r, above_row[r - 1], |
+ x0 + 2 * bs - frame_width); |
+ } else { |
+ vpx_memcpy(above_row, above_ref, bs * sizeof(uint16_t)); |
+ vpx_memset16(above_row + bs, above_row[bs - 1], bs); |
+ } |
+ } else if (x0 <= frame_width) { |
+ const int r = frame_width - x0; |
+ if (right_available && bs == 4) { |
+ vpx_memcpy(above_row, above_ref, r * sizeof(uint16_t)); |
+ vpx_memset16(above_row + r, above_row[r - 1], |
+ x0 + 2 * bs - frame_width); |
+ } else { |
+ vpx_memcpy(above_row, above_ref, r * sizeof(uint16_t)); |
+ vpx_memset16(above_row + r, above_row[r - 1], |
+ x0 + 2 * bs - frame_width); |
+ } |
+ } |
+ // TODO(Peter) this value should probably change for high bitdepth |
+ above_row[-1] = left_available ? above_ref[-1] : (base+1); |
+ } else { |
+ /* faster path if the block does not need extension */ |
+ if (bs == 4 && right_available && left_available) { |
+ const_above_row = above_ref; |
+ } else { |
+ vpx_memcpy(above_row, above_ref, bs * sizeof(uint16_t)); |
+ if (bs == 4 && right_available) |
+ vpx_memcpy(above_row + bs, above_ref + bs, bs * sizeof(uint16_t)); |
+ else |
+ vpx_memset16(above_row + bs, above_row[bs - 1], bs); |
+ // TODO(Peter): this value should probably change for high bitdepth |
+ above_row[-1] = left_available ? above_ref[-1] : (base+1); |
+ } |
+ } |
+ } else { |
+ vpx_memset16(above_row, base - 1, bs * 2); |
+ // TODO(Peter): this value should probably change for high bitdepth |
+ above_row[-1] = base - 1; |
+ } |
+ |
+ // predict |
+ if (mode == DC_PRED) { |
+ dc_pred_high[left_available][up_available][tx_size](dst, dst_stride, |
+ const_above_row, |
+ left_col, xd->bd); |
+ } else { |
+ pred_high[mode][tx_size](dst, dst_stride, const_above_row, left_col, |
+ xd->bd); |
+ } |
+} |
+#endif // CONFIG_VP9_HIGHBITDEPTH |
+ |
static void build_intra_predictors(const MACROBLOCKD *xd, const uint8_t *ref, |
int ref_stride, uint8_t *dst, int dst_stride, |
PREDICTION_MODE mode, TX_SIZE tx_size, |
@@ -454,6 +895,14 @@ |
const int y = loff * 4; |
assert(bwl >= 0); |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ if (xd->cur_buf->flags & YV12_FLAG_HIGHBITDEPTH) { |
+ build_intra_predictors_high(xd, ref, ref_stride, dst, dst_stride, mode, |
+ tx_size, have_top, have_left, have_right, |
+ x, y, plane, xd->bd); |
+ return; |
+ } |
+#endif |
build_intra_predictors(xd, ref, ref_stride, dst, dst_stride, mode, tx_size, |
have_top, have_left, have_right, x, y, plane); |
} |