| Index: source/libvpx/vp9/common/vp9_convolve.c
|
| ===================================================================
|
| --- source/libvpx/vp9/common/vp9_convolve.c (revision 219822)
|
| +++ source/libvpx/vp9/common/vp9_convolve.c (working copy)
|
| @@ -14,66 +14,45 @@
|
| #include "./vpx_config.h"
|
| #include "./vp9_rtcd.h"
|
| #include "vp9/common/vp9_common.h"
|
| +#include "vp9/common/vp9_filter.h"
|
| #include "vpx/vpx_integer.h"
|
| #include "vpx_ports/mem.h"
|
|
|
| -#define VP9_FILTER_WEIGHT 128
|
| -#define VP9_FILTER_SHIFT 7
|
| -
|
| -/* Assume a bank of 16 filters to choose from. There are two implementations
|
| - * for filter wrapping behavior, since we want to be able to pick which filter
|
| - * to start with. We could either:
|
| - *
|
| - * 1) make filter_ a pointer to the base of the filter array, and then add an
|
| - * additional offset parameter, to choose the starting filter.
|
| - * 2) use a pointer to 2 periods worth of filters, so that even if the original
|
| - * phase offset is at 15/16, we'll have valid data to read. The filter
|
| - * tables become [32][8], and the second half is duplicated.
|
| - * 3) fix the alignment of the filter tables, so that we know the 0/16 is
|
| - * always 256 byte aligned.
|
| - *
|
| - * Implementations 2 and 3 are likely preferable, as they avoid an extra 2
|
| - * parameters, and switching between them is trivial, with the
|
| - * ALIGN_FILTERS_256 macro, below.
|
| - */
|
| - #define ALIGN_FILTERS_256 1
|
| -
|
| static void convolve_horiz_c(const uint8_t *src, ptrdiff_t src_stride,
|
| uint8_t *dst, ptrdiff_t dst_stride,
|
| const int16_t *filter_x0, int x_step_q4,
|
| const int16_t *filter_y, int y_step_q4,
|
| int w, int h, int taps) {
|
| - int x, y, k, sum;
|
| - const int16_t *filter_x_base = filter_x0;
|
| + int x, y, k;
|
|
|
| -#if ALIGN_FILTERS_256
|
| - filter_x_base = (const int16_t *)(((intptr_t)filter_x0) & ~(intptr_t)0xff);
|
| -#endif
|
| + /* NOTE: This assumes that the filter table is 256-byte aligned. */
|
| + /* TODO(agrange) Modify to make independent of table alignment. */
|
| + const int16_t *const filter_x_base =
|
| + (const int16_t *)(((intptr_t)filter_x0) & ~(intptr_t)0xff);
|
|
|
| /* Adjust base pointer address for this source line */
|
| src -= taps / 2 - 1;
|
|
|
| for (y = 0; y < h; ++y) {
|
| - /* Pointer to filter to use */
|
| - const int16_t *filter_x = filter_x0;
|
| -
|
| /* Initial phase offset */
|
| - int x0_q4 = (filter_x - filter_x_base) / taps;
|
| - int x_q4 = x0_q4;
|
| + int x_q4 = (filter_x0 - filter_x_base) / taps;
|
|
|
| for (x = 0; x < w; ++x) {
|
| /* Per-pixel src offset */
|
| - int src_x = (x_q4 - x0_q4) >> 4;
|
| + const int src_x = x_q4 >> SUBPEL_BITS;
|
| + int sum = 0;
|
|
|
| - for (sum = 0, k = 0; k < taps; ++k) {
|
| + /* Pointer to filter to use */
|
| + const int16_t *const filter_x = filter_x_base +
|
| + (x_q4 & SUBPEL_MASK) * taps;
|
| +
|
| + for (k = 0; k < taps; ++k)
|
| sum += src[src_x + k] * filter_x[k];
|
| - }
|
| - sum += (VP9_FILTER_WEIGHT >> 1);
|
| - dst[x] = clip_pixel(sum >> VP9_FILTER_SHIFT);
|
|
|
| - /* Adjust source and filter to use for the next pixel */
|
| + dst[x] = clip_pixel(ROUND_POWER_OF_TWO(sum, FILTER_BITS));
|
| +
|
| + /* Move to the next source pixel */
|
| x_q4 += x_step_q4;
|
| - filter_x = filter_x_base + (x_q4 & 0xf) * taps;
|
| }
|
| src += src_stride;
|
| dst += dst_stride;
|
| @@ -85,37 +64,37 @@
|
| const int16_t *filter_x0, int x_step_q4,
|
| const int16_t *filter_y, int y_step_q4,
|
| int w, int h, int taps) {
|
| - int x, y, k, sum;
|
| - const int16_t *filter_x_base = filter_x0;
|
| + int x, y, k;
|
|
|
| -#if ALIGN_FILTERS_256
|
| - filter_x_base = (const int16_t *)(((intptr_t)filter_x0) & ~(intptr_t)0xff);
|
| -#endif
|
| + /* NOTE: This assumes that the filter table is 256-byte aligned. */
|
| + /* TODO(agrange) Modify to make independent of table alignment. */
|
| + const int16_t *const filter_x_base =
|
| + (const int16_t *)(((intptr_t)filter_x0) & ~(intptr_t)0xff);
|
|
|
| /* Adjust base pointer address for this source line */
|
| src -= taps / 2 - 1;
|
|
|
| for (y = 0; y < h; ++y) {
|
| - /* Pointer to filter to use */
|
| - const int16_t *filter_x = filter_x0;
|
| -
|
| /* Initial phase offset */
|
| - int x0_q4 = (filter_x - filter_x_base) / taps;
|
| - int x_q4 = x0_q4;
|
| + int x_q4 = (filter_x0 - filter_x_base) / taps;
|
|
|
| for (x = 0; x < w; ++x) {
|
| /* Per-pixel src offset */
|
| - int src_x = (x_q4 - x0_q4) >> 4;
|
| + const int src_x = x_q4 >> SUBPEL_BITS;
|
| + int sum = 0;
|
|
|
| - for (sum = 0, k = 0; k < taps; ++k) {
|
| + /* Pointer to filter to use */
|
| + const int16_t *const filter_x = filter_x_base +
|
| + (x_q4 & SUBPEL_MASK) * taps;
|
| +
|
| + for (k = 0; k < taps; ++k)
|
| sum += src[src_x + k] * filter_x[k];
|
| - }
|
| - sum += (VP9_FILTER_WEIGHT >> 1);
|
| - dst[x] = (dst[x] + clip_pixel(sum >> VP9_FILTER_SHIFT) + 1) >> 1;
|
|
|
| - /* Adjust source and filter to use for the next pixel */
|
| + dst[x] = ROUND_POWER_OF_TWO(dst[x] +
|
| + clip_pixel(ROUND_POWER_OF_TWO(sum, FILTER_BITS)), 1);
|
| +
|
| + /* Move to the next source pixel */
|
| x_q4 += x_step_q4;
|
| - filter_x = filter_x_base + (x_q4 & 0xf) * taps;
|
| }
|
| src += src_stride;
|
| dst += dst_stride;
|
| @@ -127,37 +106,37 @@
|
| const int16_t *filter_x, int x_step_q4,
|
| const int16_t *filter_y0, int y_step_q4,
|
| int w, int h, int taps) {
|
| - int x, y, k, sum;
|
| + int x, y, k;
|
|
|
| - const int16_t *filter_y_base = filter_y0;
|
| + /* NOTE: This assumes that the filter table is 256-byte aligned. */
|
| + /* TODO(agrange) Modify to make independent of table alignment. */
|
| + const int16_t *const filter_y_base =
|
| + (const int16_t *)(((intptr_t)filter_y0) & ~(intptr_t)0xff);
|
|
|
| -#if ALIGN_FILTERS_256
|
| - filter_y_base = (const int16_t *)(((intptr_t)filter_y0) & ~(intptr_t)0xff);
|
| -#endif
|
| -
|
| /* Adjust base pointer address for this source column */
|
| src -= src_stride * (taps / 2 - 1);
|
| - for (x = 0; x < w; ++x) {
|
| - /* Pointer to filter to use */
|
| - const int16_t *filter_y = filter_y0;
|
|
|
| + for (x = 0; x < w; ++x) {
|
| /* Initial phase offset */
|
| - int y0_q4 = (filter_y - filter_y_base) / taps;
|
| - int y_q4 = y0_q4;
|
| + int y_q4 = (filter_y0 - filter_y_base) / taps;
|
|
|
| for (y = 0; y < h; ++y) {
|
| /* Per-pixel src offset */
|
| - int src_y = (y_q4 - y0_q4) >> 4;
|
| + const int src_y = y_q4 >> SUBPEL_BITS;
|
| + int sum = 0;
|
|
|
| - for (sum = 0, k = 0; k < taps; ++k) {
|
| + /* Pointer to filter to use */
|
| + const int16_t *const filter_y = filter_y_base +
|
| + (y_q4 & SUBPEL_MASK) * taps;
|
| +
|
| + for (k = 0; k < taps; ++k)
|
| sum += src[(src_y + k) * src_stride] * filter_y[k];
|
| - }
|
| - sum += (VP9_FILTER_WEIGHT >> 1);
|
| - dst[y * dst_stride] = clip_pixel(sum >> VP9_FILTER_SHIFT);
|
|
|
| - /* Adjust source and filter to use for the next pixel */
|
| + dst[y * dst_stride] =
|
| + clip_pixel(ROUND_POWER_OF_TWO(sum, FILTER_BITS));
|
| +
|
| + /* Move to the next source pixel */
|
| y_q4 += y_step_q4;
|
| - filter_y = filter_y_base + (y_q4 & 0xf) * taps;
|
| }
|
| ++src;
|
| ++dst;
|
| @@ -169,38 +148,37 @@
|
| const int16_t *filter_x, int x_step_q4,
|
| const int16_t *filter_y0, int y_step_q4,
|
| int w, int h, int taps) {
|
| - int x, y, k, sum;
|
| + int x, y, k;
|
|
|
| - const int16_t *filter_y_base = filter_y0;
|
| + /* NOTE: This assumes that the filter table is 256-byte aligned. */
|
| + /* TODO(agrange) Modify to make independent of table alignment. */
|
| + const int16_t *const filter_y_base =
|
| + (const int16_t *)(((intptr_t)filter_y0) & ~(intptr_t)0xff);
|
|
|
| -#if ALIGN_FILTERS_256
|
| - filter_y_base = (const int16_t *)(((intptr_t)filter_y0) & ~(intptr_t)0xff);
|
| -#endif
|
| -
|
| /* Adjust base pointer address for this source column */
|
| src -= src_stride * (taps / 2 - 1);
|
| - for (x = 0; x < w; ++x) {
|
| - /* Pointer to filter to use */
|
| - const int16_t *filter_y = filter_y0;
|
|
|
| + for (x = 0; x < w; ++x) {
|
| /* Initial phase offset */
|
| - int y0_q4 = (filter_y - filter_y_base) / taps;
|
| - int y_q4 = y0_q4;
|
| + int y_q4 = (filter_y0 - filter_y_base) / taps;
|
|
|
| for (y = 0; y < h; ++y) {
|
| /* Per-pixel src offset */
|
| - int src_y = (y_q4 - y0_q4) >> 4;
|
| + const int src_y = y_q4 >> SUBPEL_BITS;
|
| + int sum = 0;
|
|
|
| - for (sum = 0, k = 0; k < taps; ++k) {
|
| + /* Pointer to filter to use */
|
| + const int16_t *const filter_y = filter_y_base +
|
| + (y_q4 & SUBPEL_MASK) * taps;
|
| +
|
| + for (k = 0; k < taps; ++k)
|
| sum += src[(src_y + k) * src_stride] * filter_y[k];
|
| - }
|
| - sum += (VP9_FILTER_WEIGHT >> 1);
|
| - dst[y * dst_stride] =
|
| - (dst[y * dst_stride] + clip_pixel(sum >> VP9_FILTER_SHIFT) + 1) >> 1;
|
|
|
| - /* Adjust source and filter to use for the next pixel */
|
| + dst[y * dst_stride] = ROUND_POWER_OF_TWO(dst[y * dst_stride] +
|
| + clip_pixel(ROUND_POWER_OF_TWO(sum, FILTER_BITS)), 1);
|
| +
|
| + /* Move to the next source pixel */
|
| y_q4 += y_step_q4;
|
| - filter_y = filter_y_base + (y_q4 & 0xf) * taps;
|
| }
|
| ++src;
|
| ++dst;
|
| @@ -228,53 +206,20 @@
|
| if (intermediate_height < h)
|
| intermediate_height = h;
|
|
|
| - convolve_horiz_c(src - src_stride * (taps / 2 - 1), src_stride,
|
| - temp, 64,
|
| - filter_x, x_step_q4, filter_y, y_step_q4,
|
| - w, intermediate_height, taps);
|
| - convolve_vert_c(temp + 64 * (taps / 2 - 1), 64, dst, dst_stride,
|
| - filter_x, x_step_q4, filter_y, y_step_q4,
|
| - w, h, taps);
|
| + convolve_horiz_c(src - src_stride * (taps / 2 - 1), src_stride, temp, 64,
|
| + filter_x, x_step_q4, filter_y, y_step_q4, w,
|
| + intermediate_height, taps);
|
| + convolve_vert_c(temp + 64 * (taps / 2 - 1), 64, dst, dst_stride, filter_x,
|
| + x_step_q4, filter_y, y_step_q4, w, h, taps);
|
| }
|
|
|
| -static void convolve_avg_c(const uint8_t *src, ptrdiff_t src_stride,
|
| - uint8_t *dst, ptrdiff_t dst_stride,
|
| - const int16_t *filter_x, int x_step_q4,
|
| - const int16_t *filter_y, int y_step_q4,
|
| - int w, int h, int taps) {
|
| - /* Fixed size intermediate buffer places limits on parameters.
|
| - * Maximum intermediate_height is 135, for y_step_q4 == 32,
|
| - * h == 64, taps == 8.
|
| - */
|
| - uint8_t temp[64 * 135];
|
| - int intermediate_height = MAX(((h * y_step_q4) >> 4), 1) + taps - 1;
|
| -
|
| - assert(w <= 64);
|
| - assert(h <= 64);
|
| - assert(taps <= 8);
|
| - assert(y_step_q4 <= 32);
|
| - assert(x_step_q4 <= 32);
|
| -
|
| - if (intermediate_height < h)
|
| - intermediate_height = h;
|
| -
|
| - convolve_horiz_c(src - src_stride * (taps / 2 - 1), src_stride,
|
| - temp, 64,
|
| - filter_x, x_step_q4, filter_y, y_step_q4,
|
| - w, intermediate_height, taps);
|
| - convolve_avg_vert_c(temp + 64 * (taps / 2 - 1), 64, dst, dst_stride,
|
| - filter_x, x_step_q4, filter_y, y_step_q4,
|
| - w, h, taps);
|
| -}
|
| -
|
| void vp9_convolve8_horiz_c(const uint8_t *src, ptrdiff_t src_stride,
|
| uint8_t *dst, ptrdiff_t dst_stride,
|
| const int16_t *filter_x, int x_step_q4,
|
| const int16_t *filter_y, int y_step_q4,
|
| int w, int h) {
|
| convolve_horiz_c(src, src_stride, dst, dst_stride,
|
| - filter_x, x_step_q4, filter_y, y_step_q4,
|
| - w, h, 8);
|
| + filter_x, x_step_q4, filter_y, y_step_q4, w, h, 8);
|
| }
|
|
|
| void vp9_convolve8_avg_horiz_c(const uint8_t *src, ptrdiff_t src_stride,
|
| @@ -283,8 +228,7 @@
|
| const int16_t *filter_y, int y_step_q4,
|
| int w, int h) {
|
| convolve_avg_horiz_c(src, src_stride, dst, dst_stride,
|
| - filter_x, x_step_q4, filter_y, y_step_q4,
|
| - w, h, 8);
|
| + filter_x, x_step_q4, filter_y, y_step_q4, w, h, 8);
|
| }
|
|
|
| void vp9_convolve8_vert_c(const uint8_t *src, ptrdiff_t src_stride,
|
| @@ -293,8 +237,7 @@
|
| const int16_t *filter_y, int y_step_q4,
|
| int w, int h) {
|
| convolve_vert_c(src, src_stride, dst, dst_stride,
|
| - filter_x, x_step_q4, filter_y, y_step_q4,
|
| - w, h, 8);
|
| + filter_x, x_step_q4, filter_y, y_step_q4, w, h, 8);
|
| }
|
|
|
| void vp9_convolve8_avg_vert_c(const uint8_t *src, ptrdiff_t src_stride,
|
| @@ -303,8 +246,7 @@
|
| const int16_t *filter_y, int y_step_q4,
|
| int w, int h) {
|
| convolve_avg_vert_c(src, src_stride, dst, dst_stride,
|
| - filter_x, x_step_q4, filter_y, y_step_q4,
|
| - w, h, 8);
|
| + filter_x, x_step_q4, filter_y, y_step_q4, w, h, 8);
|
| }
|
|
|
| void vp9_convolve8_c(const uint8_t *src, ptrdiff_t src_stride,
|
| @@ -313,8 +255,7 @@
|
| const int16_t *filter_y, int y_step_q4,
|
| int w, int h) {
|
| convolve_c(src, src_stride, dst, dst_stride,
|
| - filter_x, x_step_q4, filter_y, y_step_q4,
|
| - w, h, 8);
|
| + filter_x, x_step_q4, filter_y, y_step_q4, w, h, 8);
|
| }
|
|
|
| void vp9_convolve8_avg_c(const uint8_t *src, ptrdiff_t src_stride,
|
| @@ -327,16 +268,9 @@
|
| assert(w <= 64);
|
| assert(h <= 64);
|
|
|
| - vp9_convolve8(src, src_stride,
|
| - temp, 64,
|
| - filter_x, x_step_q4,
|
| - filter_y, y_step_q4,
|
| - w, h);
|
| - vp9_convolve_avg(temp, 64,
|
| - dst, dst_stride,
|
| - NULL, 0, /* These unused parameter should be removed! */
|
| - NULL, 0, /* These unused parameter should be removed! */
|
| - w, h);
|
| + vp9_convolve8(src, src_stride, temp, 64,
|
| + filter_x, x_step_q4, filter_y, y_step_q4, w, h);
|
| + vp9_convolve_avg(temp, 64, dst, dst_stride, NULL, 0, NULL, 0, w, h);
|
| }
|
|
|
| void vp9_convolve_copy_c(const uint8_t *src, ptrdiff_t src_stride,
|
| @@ -361,9 +295,9 @@
|
| int x, y;
|
|
|
| for (y = 0; y < h; ++y) {
|
| - for (x = 0; x < w; ++x) {
|
| - dst[x] = (dst[x] + src[x] + 1) >> 1;
|
| - }
|
| + for (x = 0; x < w; ++x)
|
| + dst[x] = ROUND_POWER_OF_TWO(dst[x] + src[x], 1);
|
| +
|
| src += src_stride;
|
| dst += dst_stride;
|
| }
|
|
|