Index: source/libvpx/vp9/common/vp9_convolve.c |
=================================================================== |
--- source/libvpx/vp9/common/vp9_convolve.c (revision 219822) |
+++ source/libvpx/vp9/common/vp9_convolve.c (working copy) |
@@ -14,66 +14,45 @@ |
#include "./vpx_config.h" |
#include "./vp9_rtcd.h" |
#include "vp9/common/vp9_common.h" |
+#include "vp9/common/vp9_filter.h" |
#include "vpx/vpx_integer.h" |
#include "vpx_ports/mem.h" |
-#define VP9_FILTER_WEIGHT 128 |
-#define VP9_FILTER_SHIFT 7 |
- |
-/* Assume a bank of 16 filters to choose from. There are two implementations |
- * for filter wrapping behavior, since we want to be able to pick which filter |
- * to start with. We could either: |
- * |
- * 1) make filter_ a pointer to the base of the filter array, and then add an |
- * additional offset parameter, to choose the starting filter. |
- * 2) use a pointer to 2 periods worth of filters, so that even if the original |
- * phase offset is at 15/16, we'll have valid data to read. The filter |
- * tables become [32][8], and the second half is duplicated. |
- * 3) fix the alignment of the filter tables, so that we know the 0/16 is |
- * always 256 byte aligned. |
- * |
- * Implementations 2 and 3 are likely preferable, as they avoid an extra 2 |
- * parameters, and switching between them is trivial, with the |
- * ALIGN_FILTERS_256 macro, below. |
- */ |
- #define ALIGN_FILTERS_256 1 |
- |
static void convolve_horiz_c(const uint8_t *src, ptrdiff_t src_stride, |
uint8_t *dst, ptrdiff_t dst_stride, |
const int16_t *filter_x0, int x_step_q4, |
const int16_t *filter_y, int y_step_q4, |
int w, int h, int taps) { |
- int x, y, k, sum; |
- const int16_t *filter_x_base = filter_x0; |
+ int x, y, k; |
-#if ALIGN_FILTERS_256 |
- filter_x_base = (const int16_t *)(((intptr_t)filter_x0) & ~(intptr_t)0xff); |
-#endif |
+ /* NOTE: This assumes that the filter table is 256-byte aligned. */ |
+ /* TODO(agrange) Modify to make independent of table alignment. */ |
+ const int16_t *const filter_x_base = |
+ (const int16_t *)(((intptr_t)filter_x0) & ~(intptr_t)0xff); |
/* Adjust base pointer address for this source line */ |
src -= taps / 2 - 1; |
for (y = 0; y < h; ++y) { |
- /* Pointer to filter to use */ |
- const int16_t *filter_x = filter_x0; |
- |
/* Initial phase offset */ |
- int x0_q4 = (filter_x - filter_x_base) / taps; |
- int x_q4 = x0_q4; |
+ int x_q4 = (filter_x0 - filter_x_base) / taps; |
for (x = 0; x < w; ++x) { |
/* Per-pixel src offset */ |
- int src_x = (x_q4 - x0_q4) >> 4; |
+ const int src_x = x_q4 >> SUBPEL_BITS; |
+ int sum = 0; |
- for (sum = 0, k = 0; k < taps; ++k) { |
+ /* Pointer to filter to use */ |
+ const int16_t *const filter_x = filter_x_base + |
+ (x_q4 & SUBPEL_MASK) * taps; |
+ |
+ for (k = 0; k < taps; ++k) |
sum += src[src_x + k] * filter_x[k]; |
- } |
- sum += (VP9_FILTER_WEIGHT >> 1); |
- dst[x] = clip_pixel(sum >> VP9_FILTER_SHIFT); |
- /* Adjust source and filter to use for the next pixel */ |
+ dst[x] = clip_pixel(ROUND_POWER_OF_TWO(sum, FILTER_BITS)); |
+ |
+ /* Move to the next source pixel */ |
x_q4 += x_step_q4; |
- filter_x = filter_x_base + (x_q4 & 0xf) * taps; |
} |
src += src_stride; |
dst += dst_stride; |
@@ -85,37 +64,37 @@ |
const int16_t *filter_x0, int x_step_q4, |
const int16_t *filter_y, int y_step_q4, |
int w, int h, int taps) { |
- int x, y, k, sum; |
- const int16_t *filter_x_base = filter_x0; |
+ int x, y, k; |
-#if ALIGN_FILTERS_256 |
- filter_x_base = (const int16_t *)(((intptr_t)filter_x0) & ~(intptr_t)0xff); |
-#endif |
+ /* NOTE: This assumes that the filter table is 256-byte aligned. */ |
+ /* TODO(agrange) Modify to make independent of table alignment. */ |
+ const int16_t *const filter_x_base = |
+ (const int16_t *)(((intptr_t)filter_x0) & ~(intptr_t)0xff); |
/* Adjust base pointer address for this source line */ |
src -= taps / 2 - 1; |
for (y = 0; y < h; ++y) { |
- /* Pointer to filter to use */ |
- const int16_t *filter_x = filter_x0; |
- |
/* Initial phase offset */ |
- int x0_q4 = (filter_x - filter_x_base) / taps; |
- int x_q4 = x0_q4; |
+ int x_q4 = (filter_x0 - filter_x_base) / taps; |
for (x = 0; x < w; ++x) { |
/* Per-pixel src offset */ |
- int src_x = (x_q4 - x0_q4) >> 4; |
+ const int src_x = x_q4 >> SUBPEL_BITS; |
+ int sum = 0; |
- for (sum = 0, k = 0; k < taps; ++k) { |
+ /* Pointer to filter to use */ |
+ const int16_t *const filter_x = filter_x_base + |
+ (x_q4 & SUBPEL_MASK) * taps; |
+ |
+ for (k = 0; k < taps; ++k) |
sum += src[src_x + k] * filter_x[k]; |
- } |
- sum += (VP9_FILTER_WEIGHT >> 1); |
- dst[x] = (dst[x] + clip_pixel(sum >> VP9_FILTER_SHIFT) + 1) >> 1; |
- /* Adjust source and filter to use for the next pixel */ |
+ dst[x] = ROUND_POWER_OF_TWO(dst[x] + |
+ clip_pixel(ROUND_POWER_OF_TWO(sum, FILTER_BITS)), 1); |
+ |
+ /* Move to the next source pixel */ |
x_q4 += x_step_q4; |
- filter_x = filter_x_base + (x_q4 & 0xf) * taps; |
} |
src += src_stride; |
dst += dst_stride; |
@@ -127,37 +106,37 @@ |
const int16_t *filter_x, int x_step_q4, |
const int16_t *filter_y0, int y_step_q4, |
int w, int h, int taps) { |
- int x, y, k, sum; |
+ int x, y, k; |
- const int16_t *filter_y_base = filter_y0; |
+ /* NOTE: This assumes that the filter table is 256-byte aligned. */ |
+ /* TODO(agrange) Modify to make independent of table alignment. */ |
+ const int16_t *const filter_y_base = |
+ (const int16_t *)(((intptr_t)filter_y0) & ~(intptr_t)0xff); |
-#if ALIGN_FILTERS_256 |
- filter_y_base = (const int16_t *)(((intptr_t)filter_y0) & ~(intptr_t)0xff); |
-#endif |
- |
/* Adjust base pointer address for this source column */ |
src -= src_stride * (taps / 2 - 1); |
- for (x = 0; x < w; ++x) { |
- /* Pointer to filter to use */ |
- const int16_t *filter_y = filter_y0; |
+ for (x = 0; x < w; ++x) { |
/* Initial phase offset */ |
- int y0_q4 = (filter_y - filter_y_base) / taps; |
- int y_q4 = y0_q4; |
+ int y_q4 = (filter_y0 - filter_y_base) / taps; |
for (y = 0; y < h; ++y) { |
/* Per-pixel src offset */ |
- int src_y = (y_q4 - y0_q4) >> 4; |
+ const int src_y = y_q4 >> SUBPEL_BITS; |
+ int sum = 0; |
- for (sum = 0, k = 0; k < taps; ++k) { |
+ /* Pointer to filter to use */ |
+ const int16_t *const filter_y = filter_y_base + |
+ (y_q4 & SUBPEL_MASK) * taps; |
+ |
+ for (k = 0; k < taps; ++k) |
sum += src[(src_y + k) * src_stride] * filter_y[k]; |
- } |
- sum += (VP9_FILTER_WEIGHT >> 1); |
- dst[y * dst_stride] = clip_pixel(sum >> VP9_FILTER_SHIFT); |
- /* Adjust source and filter to use for the next pixel */ |
+ dst[y * dst_stride] = |
+ clip_pixel(ROUND_POWER_OF_TWO(sum, FILTER_BITS)); |
+ |
+ /* Move to the next source pixel */ |
y_q4 += y_step_q4; |
- filter_y = filter_y_base + (y_q4 & 0xf) * taps; |
} |
++src; |
++dst; |
@@ -169,38 +148,37 @@ |
const int16_t *filter_x, int x_step_q4, |
const int16_t *filter_y0, int y_step_q4, |
int w, int h, int taps) { |
- int x, y, k, sum; |
+ int x, y, k; |
- const int16_t *filter_y_base = filter_y0; |
+ /* NOTE: This assumes that the filter table is 256-byte aligned. */ |
+ /* TODO(agrange) Modify to make independent of table alignment. */ |
+ const int16_t *const filter_y_base = |
+ (const int16_t *)(((intptr_t)filter_y0) & ~(intptr_t)0xff); |
-#if ALIGN_FILTERS_256 |
- filter_y_base = (const int16_t *)(((intptr_t)filter_y0) & ~(intptr_t)0xff); |
-#endif |
- |
/* Adjust base pointer address for this source column */ |
src -= src_stride * (taps / 2 - 1); |
- for (x = 0; x < w; ++x) { |
- /* Pointer to filter to use */ |
- const int16_t *filter_y = filter_y0; |
+ for (x = 0; x < w; ++x) { |
/* Initial phase offset */ |
- int y0_q4 = (filter_y - filter_y_base) / taps; |
- int y_q4 = y0_q4; |
+ int y_q4 = (filter_y0 - filter_y_base) / taps; |
for (y = 0; y < h; ++y) { |
/* Per-pixel src offset */ |
- int src_y = (y_q4 - y0_q4) >> 4; |
+ const int src_y = y_q4 >> SUBPEL_BITS; |
+ int sum = 0; |
- for (sum = 0, k = 0; k < taps; ++k) { |
+ /* Pointer to filter to use */ |
+ const int16_t *const filter_y = filter_y_base + |
+ (y_q4 & SUBPEL_MASK) * taps; |
+ |
+ for (k = 0; k < taps; ++k) |
sum += src[(src_y + k) * src_stride] * filter_y[k]; |
- } |
- sum += (VP9_FILTER_WEIGHT >> 1); |
- dst[y * dst_stride] = |
- (dst[y * dst_stride] + clip_pixel(sum >> VP9_FILTER_SHIFT) + 1) >> 1; |
- /* Adjust source and filter to use for the next pixel */ |
+ dst[y * dst_stride] = ROUND_POWER_OF_TWO(dst[y * dst_stride] + |
+ clip_pixel(ROUND_POWER_OF_TWO(sum, FILTER_BITS)), 1); |
+ |
+ /* Move to the next source pixel */ |
y_q4 += y_step_q4; |
- filter_y = filter_y_base + (y_q4 & 0xf) * taps; |
} |
++src; |
++dst; |
@@ -228,53 +206,20 @@ |
if (intermediate_height < h) |
intermediate_height = h; |
- convolve_horiz_c(src - src_stride * (taps / 2 - 1), src_stride, |
- temp, 64, |
- filter_x, x_step_q4, filter_y, y_step_q4, |
- w, intermediate_height, taps); |
- convolve_vert_c(temp + 64 * (taps / 2 - 1), 64, dst, dst_stride, |
- filter_x, x_step_q4, filter_y, y_step_q4, |
- w, h, taps); |
+ convolve_horiz_c(src - src_stride * (taps / 2 - 1), src_stride, temp, 64, |
+ filter_x, x_step_q4, filter_y, y_step_q4, w, |
+ intermediate_height, taps); |
+ convolve_vert_c(temp + 64 * (taps / 2 - 1), 64, dst, dst_stride, filter_x, |
+ x_step_q4, filter_y, y_step_q4, w, h, taps); |
} |
-static void convolve_avg_c(const uint8_t *src, ptrdiff_t src_stride, |
- uint8_t *dst, ptrdiff_t dst_stride, |
- const int16_t *filter_x, int x_step_q4, |
- const int16_t *filter_y, int y_step_q4, |
- int w, int h, int taps) { |
- /* Fixed size intermediate buffer places limits on parameters. |
- * Maximum intermediate_height is 135, for y_step_q4 == 32, |
- * h == 64, taps == 8. |
- */ |
- uint8_t temp[64 * 135]; |
- int intermediate_height = MAX(((h * y_step_q4) >> 4), 1) + taps - 1; |
- |
- assert(w <= 64); |
- assert(h <= 64); |
- assert(taps <= 8); |
- assert(y_step_q4 <= 32); |
- assert(x_step_q4 <= 32); |
- |
- if (intermediate_height < h) |
- intermediate_height = h; |
- |
- convolve_horiz_c(src - src_stride * (taps / 2 - 1), src_stride, |
- temp, 64, |
- filter_x, x_step_q4, filter_y, y_step_q4, |
- w, intermediate_height, taps); |
- convolve_avg_vert_c(temp + 64 * (taps / 2 - 1), 64, dst, dst_stride, |
- filter_x, x_step_q4, filter_y, y_step_q4, |
- w, h, taps); |
-} |
- |
void vp9_convolve8_horiz_c(const uint8_t *src, ptrdiff_t src_stride, |
uint8_t *dst, ptrdiff_t dst_stride, |
const int16_t *filter_x, int x_step_q4, |
const int16_t *filter_y, int y_step_q4, |
int w, int h) { |
convolve_horiz_c(src, src_stride, dst, dst_stride, |
- filter_x, x_step_q4, filter_y, y_step_q4, |
- w, h, 8); |
+ filter_x, x_step_q4, filter_y, y_step_q4, w, h, 8); |
} |
void vp9_convolve8_avg_horiz_c(const uint8_t *src, ptrdiff_t src_stride, |
@@ -283,8 +228,7 @@ |
const int16_t *filter_y, int y_step_q4, |
int w, int h) { |
convolve_avg_horiz_c(src, src_stride, dst, dst_stride, |
- filter_x, x_step_q4, filter_y, y_step_q4, |
- w, h, 8); |
+ filter_x, x_step_q4, filter_y, y_step_q4, w, h, 8); |
} |
void vp9_convolve8_vert_c(const uint8_t *src, ptrdiff_t src_stride, |
@@ -293,8 +237,7 @@ |
const int16_t *filter_y, int y_step_q4, |
int w, int h) { |
convolve_vert_c(src, src_stride, dst, dst_stride, |
- filter_x, x_step_q4, filter_y, y_step_q4, |
- w, h, 8); |
+ filter_x, x_step_q4, filter_y, y_step_q4, w, h, 8); |
} |
void vp9_convolve8_avg_vert_c(const uint8_t *src, ptrdiff_t src_stride, |
@@ -303,8 +246,7 @@ |
const int16_t *filter_y, int y_step_q4, |
int w, int h) { |
convolve_avg_vert_c(src, src_stride, dst, dst_stride, |
- filter_x, x_step_q4, filter_y, y_step_q4, |
- w, h, 8); |
+ filter_x, x_step_q4, filter_y, y_step_q4, w, h, 8); |
} |
void vp9_convolve8_c(const uint8_t *src, ptrdiff_t src_stride, |
@@ -313,8 +255,7 @@ |
const int16_t *filter_y, int y_step_q4, |
int w, int h) { |
convolve_c(src, src_stride, dst, dst_stride, |
- filter_x, x_step_q4, filter_y, y_step_q4, |
- w, h, 8); |
+ filter_x, x_step_q4, filter_y, y_step_q4, w, h, 8); |
} |
void vp9_convolve8_avg_c(const uint8_t *src, ptrdiff_t src_stride, |
@@ -327,16 +268,9 @@ |
assert(w <= 64); |
assert(h <= 64); |
- vp9_convolve8(src, src_stride, |
- temp, 64, |
- filter_x, x_step_q4, |
- filter_y, y_step_q4, |
- w, h); |
- vp9_convolve_avg(temp, 64, |
- dst, dst_stride, |
- NULL, 0, /* These unused parameter should be removed! */ |
- NULL, 0, /* These unused parameter should be removed! */ |
- w, h); |
+ vp9_convolve8(src, src_stride, temp, 64, |
+ filter_x, x_step_q4, filter_y, y_step_q4, w, h); |
+ vp9_convolve_avg(temp, 64, dst, dst_stride, NULL, 0, NULL, 0, w, h); |
} |
void vp9_convolve_copy_c(const uint8_t *src, ptrdiff_t src_stride, |
@@ -361,9 +295,9 @@ |
int x, y; |
for (y = 0; y < h; ++y) { |
- for (x = 0; x < w; ++x) { |
- dst[x] = (dst[x] + src[x] + 1) >> 1; |
- } |
+ for (x = 0; x < w; ++x) |
+ dst[x] = ROUND_POWER_OF_TWO(dst[x] + src[x], 1); |
+ |
src += src_stride; |
dst += dst_stride; |
} |