Index: source/libvpx/vp9/common/vp9_convolve.c |
=================================================================== |
--- source/libvpx/vp9/common/vp9_convolve.c (revision 240950) |
+++ source/libvpx/vp9/common/vp9_convolve.c (working copy) |
@@ -18,40 +18,21 @@ |
#include "vpx/vpx_integer.h" |
#include "vpx_ports/mem.h" |
-static void convolve_horiz_c(const uint8_t *src, ptrdiff_t src_stride, |
- uint8_t *dst, ptrdiff_t dst_stride, |
- const int16_t *filter_x0, int x_step_q4, |
- const int16_t *filter_y, int y_step_q4, |
- int w, int h, int taps) { |
- int x, y, k; |
- |
- /* NOTE: This assumes that the filter table is 256-byte aligned. */ |
- /* TODO(agrange) Modify to make independent of table alignment. */ |
- const int16_t *const filter_x_base = |
- (const int16_t *)(((intptr_t)filter_x0) & ~(intptr_t)0xff); |
- |
- /* Adjust base pointer address for this source line */ |
- src -= taps / 2 - 1; |
- |
+static void convolve_horiz(const uint8_t *src, ptrdiff_t src_stride, |
+ uint8_t *dst, ptrdiff_t dst_stride, |
+ const subpel_kernel *x_filters, |
+ int x0_q4, int x_step_q4, int w, int h) { |
+ int x, y; |
+ src -= SUBPEL_TAPS / 2 - 1; |
for (y = 0; y < h; ++y) { |
- /* Initial phase offset */ |
- int x_q4 = (int)(filter_x0 - filter_x_base) / taps; |
- |
+ int x_q4 = x0_q4; |
for (x = 0; x < w; ++x) { |
- /* Per-pixel src offset */ |
- const int src_x = x_q4 >> SUBPEL_BITS; |
- int sum = 0; |
- |
- /* Pointer to filter to use */ |
- const int16_t *const filter_x = filter_x_base + |
- (x_q4 & SUBPEL_MASK) * taps; |
- |
- for (k = 0; k < taps; ++k) |
- sum += src[src_x + k] * filter_x[k]; |
- |
+ const uint8_t *const src_x = &src[x_q4 >> SUBPEL_BITS]; |
+ const int16_t *const x_filter = x_filters[x_q4 & SUBPEL_MASK]; |
+ int k, sum = 0; |
+ for (k = 0; k < SUBPEL_TAPS; ++k) |
+ sum += src_x[k] * x_filter[k]; |
dst[x] = clip_pixel(ROUND_POWER_OF_TWO(sum, FILTER_BITS)); |
- |
- /* Move to the next source pixel */ |
x_q4 += x_step_q4; |
} |
src += src_stride; |
@@ -59,41 +40,22 @@ |
} |
} |
-static void convolve_avg_horiz_c(const uint8_t *src, ptrdiff_t src_stride, |
- uint8_t *dst, ptrdiff_t dst_stride, |
- const int16_t *filter_x0, int x_step_q4, |
- const int16_t *filter_y, int y_step_q4, |
- int w, int h, int taps) { |
- int x, y, k; |
- |
- /* NOTE: This assumes that the filter table is 256-byte aligned. */ |
- /* TODO(agrange) Modify to make independent of table alignment. */ |
- const int16_t *const filter_x_base = |
- (const int16_t *)(((intptr_t)filter_x0) & ~(intptr_t)0xff); |
- |
- /* Adjust base pointer address for this source line */ |
- src -= taps / 2 - 1; |
- |
+static void convolve_avg_horiz(const uint8_t *src, ptrdiff_t src_stride, |
+ uint8_t *dst, ptrdiff_t dst_stride, |
+ const subpel_kernel *x_filters, |
+ int x0_q4, int x_step_q4, int w, int h) { |
+ int x, y; |
+ src -= SUBPEL_TAPS / 2 - 1; |
for (y = 0; y < h; ++y) { |
- /* Initial phase offset */ |
- int x_q4 = (int)(filter_x0 - filter_x_base) / taps; |
- |
+ int x_q4 = x0_q4; |
for (x = 0; x < w; ++x) { |
- /* Per-pixel src offset */ |
- const int src_x = x_q4 >> SUBPEL_BITS; |
- int sum = 0; |
- |
- /* Pointer to filter to use */ |
- const int16_t *const filter_x = filter_x_base + |
- (x_q4 & SUBPEL_MASK) * taps; |
- |
- for (k = 0; k < taps; ++k) |
- sum += src[src_x + k] * filter_x[k]; |
- |
+ const uint8_t *const src_x = &src[x_q4 >> SUBPEL_BITS]; |
+ const int16_t *const x_filter = x_filters[x_q4 & SUBPEL_MASK]; |
+ int k, sum = 0; |
+ for (k = 0; k < SUBPEL_TAPS; ++k) |
+ sum += src_x[k] * x_filter[k]; |
dst[x] = ROUND_POWER_OF_TWO(dst[x] + |
- clip_pixel(ROUND_POWER_OF_TWO(sum, FILTER_BITS)), 1); |
- |
- /* Move to the next source pixel */ |
+ clip_pixel(ROUND_POWER_OF_TWO(sum, FILTER_BITS)), 1); |
x_q4 += x_step_q4; |
} |
src += src_stride; |
@@ -101,41 +63,22 @@ |
} |
} |
-static void convolve_vert_c(const uint8_t *src, ptrdiff_t src_stride, |
- uint8_t *dst, ptrdiff_t dst_stride, |
- const int16_t *filter_x, int x_step_q4, |
- const int16_t *filter_y0, int y_step_q4, |
- int w, int h, int taps) { |
- int x, y, k; |
+static void convolve_vert(const uint8_t *src, ptrdiff_t src_stride, |
+ uint8_t *dst, ptrdiff_t dst_stride, |
+ const subpel_kernel *y_filters, |
+ int y0_q4, int y_step_q4, int w, int h) { |
+ int x, y; |
+ src -= src_stride * (SUBPEL_TAPS / 2 - 1); |
- /* NOTE: This assumes that the filter table is 256-byte aligned. */ |
- /* TODO(agrange) Modify to make independent of table alignment. */ |
- const int16_t *const filter_y_base = |
- (const int16_t *)(((intptr_t)filter_y0) & ~(intptr_t)0xff); |
- |
- /* Adjust base pointer address for this source column */ |
- src -= src_stride * (taps / 2 - 1); |
- |
for (x = 0; x < w; ++x) { |
- /* Initial phase offset */ |
- int y_q4 = (int)(filter_y0 - filter_y_base) / taps; |
- |
+ int y_q4 = y0_q4; |
for (y = 0; y < h; ++y) { |
- /* Per-pixel src offset */ |
- const int src_y = y_q4 >> SUBPEL_BITS; |
- int sum = 0; |
- |
- /* Pointer to filter to use */ |
- const int16_t *const filter_y = filter_y_base + |
- (y_q4 & SUBPEL_MASK) * taps; |
- |
- for (k = 0; k < taps; ++k) |
- sum += src[(src_y + k) * src_stride] * filter_y[k]; |
- |
- dst[y * dst_stride] = |
- clip_pixel(ROUND_POWER_OF_TWO(sum, FILTER_BITS)); |
- |
- /* Move to the next source pixel */ |
+ const unsigned char *src_y = &src[(y_q4 >> SUBPEL_BITS) * src_stride]; |
+ const int16_t *const y_filter = y_filters[y_q4 & SUBPEL_MASK]; |
+ int k, sum = 0; |
+ for (k = 0; k < SUBPEL_TAPS; ++k) |
+ sum += src_y[k * src_stride] * y_filter[k]; |
+ dst[y * dst_stride] = clip_pixel(ROUND_POWER_OF_TWO(sum, FILTER_BITS)); |
y_q4 += y_step_q4; |
} |
++src; |
@@ -143,41 +86,23 @@ |
} |
} |
-static void convolve_avg_vert_c(const uint8_t *src, ptrdiff_t src_stride, |
- uint8_t *dst, ptrdiff_t dst_stride, |
- const int16_t *filter_x, int x_step_q4, |
- const int16_t *filter_y0, int y_step_q4, |
- int w, int h, int taps) { |
- int x, y, k; |
+static void convolve_avg_vert(const uint8_t *src, ptrdiff_t src_stride, |
+ uint8_t *dst, ptrdiff_t dst_stride, |
+ const subpel_kernel *y_filters, |
+ int y0_q4, int y_step_q4, int w, int h) { |
+ int x, y; |
+ src -= src_stride * (SUBPEL_TAPS / 2 - 1); |
- /* NOTE: This assumes that the filter table is 256-byte aligned. */ |
- /* TODO(agrange) Modify to make independent of table alignment. */ |
- const int16_t *const filter_y_base = |
- (const int16_t *)(((intptr_t)filter_y0) & ~(intptr_t)0xff); |
- |
- /* Adjust base pointer address for this source column */ |
- src -= src_stride * (taps / 2 - 1); |
- |
for (x = 0; x < w; ++x) { |
- /* Initial phase offset */ |
- int y_q4 = (int)(filter_y0 - filter_y_base) / taps; |
- |
+ int y_q4 = y0_q4; |
for (y = 0; y < h; ++y) { |
- /* Per-pixel src offset */ |
- const int src_y = y_q4 >> SUBPEL_BITS; |
- int sum = 0; |
- |
- /* Pointer to filter to use */ |
- const int16_t *const filter_y = filter_y_base + |
- (y_q4 & SUBPEL_MASK) * taps; |
- |
- for (k = 0; k < taps; ++k) |
- sum += src[(src_y + k) * src_stride] * filter_y[k]; |
- |
+ const unsigned char *src_y = &src[(y_q4 >> SUBPEL_BITS) * src_stride]; |
+ const int16_t *const y_filter = y_filters[y_q4 & SUBPEL_MASK]; |
+ int k, sum = 0; |
+ for (k = 0; k < SUBPEL_TAPS; ++k) |
+ sum += src_y[k * src_stride] * y_filter[k]; |
dst[y * dst_stride] = ROUND_POWER_OF_TWO(dst[y * dst_stride] + |
- clip_pixel(ROUND_POWER_OF_TWO(sum, FILTER_BITS)), 1); |
- |
- /* Move to the next source pixel */ |
+ clip_pixel(ROUND_POWER_OF_TWO(sum, FILTER_BITS)), 1); |
y_q4 += y_step_q4; |
} |
++src; |
@@ -185,42 +110,54 @@ |
} |
} |
-static void convolve_c(const uint8_t *src, ptrdiff_t src_stride, |
- uint8_t *dst, ptrdiff_t dst_stride, |
- const int16_t *filter_x, int x_step_q4, |
- const int16_t *filter_y, int y_step_q4, |
- int w, int h, int taps) { |
- /* Fixed size intermediate buffer places limits on parameters. |
- * Maximum intermediate_height is 324, for y_step_q4 == 80, |
- * h == 64, taps == 8. |
- * y_step_q4 of 80 allows for 1/10 scale for 5 layer svc |
- */ |
+static void convolve(const uint8_t *src, ptrdiff_t src_stride, |
+ uint8_t *dst, ptrdiff_t dst_stride, |
+ const subpel_kernel *const x_filters, |
+ int x0_q4, int x_step_q4, |
+ const subpel_kernel *const y_filters, |
+ int y0_q4, int y_step_q4, |
+ int w, int h) { |
+ // Fixed size intermediate buffer places limits on parameters. |
+ // Maximum intermediate_height is 324, for y_step_q4 == 80, |
+ // h == 64, taps == 8. |
+ // y_step_q4 of 80 allows for 1/10 scale for 5 layer svc |
uint8_t temp[64 * 324]; |
- int intermediate_height = (((h - 1) * y_step_q4 + 15) >> 4) + taps; |
+ int intermediate_height = (((h - 1) * y_step_q4 + 15) >> 4) + SUBPEL_TAPS; |
assert(w <= 64); |
assert(h <= 64); |
- assert(taps <= 8); |
assert(y_step_q4 <= 80); |
assert(x_step_q4 <= 80); |
if (intermediate_height < h) |
intermediate_height = h; |
- convolve_horiz_c(src - src_stride * (taps / 2 - 1), src_stride, temp, 64, |
- filter_x, x_step_q4, filter_y, y_step_q4, w, |
- intermediate_height, taps); |
- convolve_vert_c(temp + 64 * (taps / 2 - 1), 64, dst, dst_stride, filter_x, |
- x_step_q4, filter_y, y_step_q4, w, h, taps); |
+ convolve_horiz(src - src_stride * (SUBPEL_TAPS / 2 - 1), src_stride, temp, 64, |
+ x_filters, x0_q4, x_step_q4, w, intermediate_height); |
+ convolve_vert(temp + 64 * (SUBPEL_TAPS / 2 - 1), 64, dst, dst_stride, |
+ y_filters, y0_q4, y_step_q4, w, h); |
} |
+static const subpel_kernel *get_filter_base(const int16_t *filter) { |
+ // NOTE: This assumes that the filter table is 256-byte aligned. |
+ // TODO(agrange) Modify to make independent of table alignment. |
+ return (const subpel_kernel *)(((intptr_t)filter) & ~((intptr_t)0xFF)); |
+} |
+ |
+static int get_filter_offset(const int16_t *f, const subpel_kernel *base) { |
+ return (const subpel_kernel *)(intptr_t)f - base; |
+} |
+ |
void vp9_convolve8_horiz_c(const uint8_t *src, ptrdiff_t src_stride, |
uint8_t *dst, ptrdiff_t dst_stride, |
const int16_t *filter_x, int x_step_q4, |
const int16_t *filter_y, int y_step_q4, |
int w, int h) { |
- convolve_horiz_c(src, src_stride, dst, dst_stride, |
- filter_x, x_step_q4, filter_y, y_step_q4, w, h, 8); |
+ const subpel_kernel *const filters_x = get_filter_base(filter_x); |
+ const int x0_q4 = get_filter_offset(filter_x, filters_x); |
+ |
+ convolve_horiz(src, src_stride, dst, dst_stride, filters_x, |
+ x0_q4, x_step_q4, w, h); |
} |
void vp9_convolve8_avg_horiz_c(const uint8_t *src, ptrdiff_t src_stride, |
@@ -228,8 +165,11 @@ |
const int16_t *filter_x, int x_step_q4, |
const int16_t *filter_y, int y_step_q4, |
int w, int h) { |
- convolve_avg_horiz_c(src, src_stride, dst, dst_stride, |
- filter_x, x_step_q4, filter_y, y_step_q4, w, h, 8); |
+ const subpel_kernel *const filters_x = get_filter_base(filter_x); |
+ const int x0_q4 = get_filter_offset(filter_x, filters_x); |
+ |
+ convolve_avg_horiz(src, src_stride, dst, dst_stride, filters_x, |
+ x0_q4, x_step_q4, w, h); |
} |
void vp9_convolve8_vert_c(const uint8_t *src, ptrdiff_t src_stride, |
@@ -237,8 +177,10 @@ |
const int16_t *filter_x, int x_step_q4, |
const int16_t *filter_y, int y_step_q4, |
int w, int h) { |
- convolve_vert_c(src, src_stride, dst, dst_stride, |
- filter_x, x_step_q4, filter_y, y_step_q4, w, h, 8); |
+ const subpel_kernel *const filters_y = get_filter_base(filter_y); |
+ const int y0_q4 = get_filter_offset(filter_y, filters_y); |
+ convolve_vert(src, src_stride, dst, dst_stride, filters_y, |
+ y0_q4, y_step_q4, w, h); |
} |
void vp9_convolve8_avg_vert_c(const uint8_t *src, ptrdiff_t src_stride, |
@@ -246,8 +188,10 @@ |
const int16_t *filter_x, int x_step_q4, |
const int16_t *filter_y, int y_step_q4, |
int w, int h) { |
- convolve_avg_vert_c(src, src_stride, dst, dst_stride, |
- filter_x, x_step_q4, filter_y, y_step_q4, w, h, 8); |
+ const subpel_kernel *const filters_y = get_filter_base(filter_y); |
+ const int y0_q4 = get_filter_offset(filter_y, filters_y); |
+ convolve_avg_vert(src, src_stride, dst, dst_stride, filters_y, |
+ y0_q4, y_step_q4, w, h); |
} |
void vp9_convolve8_c(const uint8_t *src, ptrdiff_t src_stride, |
@@ -255,8 +199,15 @@ |
const int16_t *filter_x, int x_step_q4, |
const int16_t *filter_y, int y_step_q4, |
int w, int h) { |
- convolve_c(src, src_stride, dst, dst_stride, |
- filter_x, x_step_q4, filter_y, y_step_q4, w, h, 8); |
+ const subpel_kernel *const filters_x = get_filter_base(filter_x); |
+ const int x0_q4 = get_filter_offset(filter_x, filters_x); |
+ |
+ const subpel_kernel *const filters_y = get_filter_base(filter_y); |
+ const int y0_q4 = get_filter_offset(filter_y, filters_y); |
+ |
+ convolve(src, src_stride, dst, dst_stride, |
+ filters_x, x0_q4, x_step_q4, |
+ filters_y, y0_q4, y_step_q4, w, h); |
} |
void vp9_convolve8_avg_c(const uint8_t *src, ptrdiff_t src_stride, |
@@ -269,9 +220,9 @@ |
assert(w <= 64); |
assert(h <= 64); |
- vp9_convolve8(src, src_stride, temp, 64, |
- filter_x, x_step_q4, filter_y, y_step_q4, w, h); |
- vp9_convolve_avg(temp, 64, dst, dst_stride, NULL, 0, NULL, 0, w, h); |
+ vp9_convolve8_c(src, src_stride, temp, 64, |
+ filter_x, x_step_q4, filter_y, y_step_q4, w, h); |
+ vp9_convolve_avg_c(temp, 64, dst, dst_stride, NULL, 0, NULL, 0, w, h); |
} |
void vp9_convolve_copy_c(const uint8_t *src, ptrdiff_t src_stride, |