Index: source/libvpx/vpx_dsp/inv_txfm.c |
diff --git a/source/libvpx/vpx_dsp/inv_txfm.c b/source/libvpx/vpx_dsp/inv_txfm.c |
index 6babf54d0ae030a011bc8d486073d9750f3501a1..3afa8cdc40981a8c1acabfd88d9a780da5fcc7b9 100644 |
--- a/source/libvpx/vpx_dsp/inv_txfm.c |
+++ b/source/libvpx/vpx_dsp/inv_txfm.c |
@@ -13,7 +13,7 @@ |
#include "vpx_dsp/inv_txfm.h" |
-void vp9_iwht4x4_16_add_c(const tran_low_t *input, uint8_t *dest, int stride) { |
+void vpx_iwht4x4_16_add_c(const tran_low_t *input, uint8_t *dest, int stride) { |
/* 4-point reversible, orthonormal inverse Walsh-Hadamard in 3.5 adds, |
0.5 shifts per pixel. */ |
int i; |
@@ -65,7 +65,7 @@ void vp9_iwht4x4_16_add_c(const tran_low_t *input, uint8_t *dest, int stride) { |
} |
} |
-void vp9_iwht4x4_1_add_c(const tran_low_t *in, uint8_t *dest, int dest_stride) { |
+void vpx_iwht4x4_1_add_c(const tran_low_t *in, uint8_t *dest, int dest_stride) { |
int i; |
tran_high_t a1, e1; |
tran_low_t tmp[4]; |
@@ -111,7 +111,7 @@ void idct4_c(const tran_low_t *input, tran_low_t *output) { |
output[3] = WRAPLOW(step[0] - step[3], 8); |
} |
-void vp9_idct4x4_16_add_c(const tran_low_t *input, uint8_t *dest, int stride) { |
+void vpx_idct4x4_16_add_c(const tran_low_t *input, uint8_t *dest, int stride) { |
tran_low_t out[4 * 4]; |
tran_low_t *outptr = out; |
int i, j; |
@@ -136,7 +136,7 @@ void vp9_idct4x4_16_add_c(const tran_low_t *input, uint8_t *dest, int stride) { |
} |
} |
-void vp9_idct4x4_1_add_c(const tran_low_t *input, uint8_t *dest, |
+void vpx_idct4x4_1_add_c(const tran_low_t *input, uint8_t *dest, |
int dest_stride) { |
int i; |
tran_high_t a1; |
@@ -198,7 +198,7 @@ void idct8_c(const tran_low_t *input, tran_low_t *output) { |
output[7] = WRAPLOW(step1[0] - step1[7], 8); |
} |
-void vp9_idct8x8_64_add_c(const tran_low_t *input, uint8_t *dest, int stride) { |
+void vpx_idct8x8_64_add_c(const tran_low_t *input, uint8_t *dest, int stride) { |
tran_low_t out[8 * 8]; |
tran_low_t *outptr = out; |
int i, j; |
@@ -223,7 +223,7 @@ void vp9_idct8x8_64_add_c(const tran_low_t *input, uint8_t *dest, int stride) { |
} |
} |
-void vp9_idct8x8_1_add_c(const tran_low_t *input, uint8_t *dest, int stride) { |
+void vpx_idct8x8_1_add_c(const tran_low_t *input, uint8_t *dest, int stride) { |
int i, j; |
tran_high_t a1; |
tran_low_t out = WRAPLOW(dct_const_round_shift(input[0] * cospi_16_64), 8); |
@@ -350,7 +350,7 @@ void iadst8_c(const tran_low_t *input, tran_low_t *output) { |
output[7] = WRAPLOW(-x1, 8); |
} |
-void vp9_idct8x8_12_add_c(const tran_low_t *input, uint8_t *dest, int stride) { |
+void vpx_idct8x8_12_add_c(const tran_low_t *input, uint8_t *dest, int stride) { |
tran_low_t out[8 * 8] = { 0 }; |
tran_low_t *outptr = out; |
int i, j; |
@@ -541,7 +541,7 @@ void idct16_c(const tran_low_t *input, tran_low_t *output) { |
output[15] = WRAPLOW(step2[0] - step2[15], 8); |
} |
-void vp9_idct16x16_256_add_c(const tran_low_t *input, uint8_t *dest, |
+void vpx_idct16x16_256_add_c(const tran_low_t *input, uint8_t *dest, |
int stride) { |
tran_low_t out[16 * 16]; |
tran_low_t *outptr = out; |
@@ -739,7 +739,7 @@ void iadst16_c(const tran_low_t *input, tran_low_t *output) { |
output[15] = WRAPLOW(-x1, 8); |
} |
-void vp9_idct16x16_10_add_c(const tran_low_t *input, uint8_t *dest, |
+void vpx_idct16x16_10_add_c(const tran_low_t *input, uint8_t *dest, |
int stride) { |
tran_low_t out[16 * 16] = { 0 }; |
tran_low_t *outptr = out; |
@@ -766,7 +766,7 @@ void vp9_idct16x16_10_add_c(const tran_low_t *input, uint8_t *dest, |
} |
} |
-void vp9_idct16x16_1_add_c(const tran_low_t *input, uint8_t *dest, int stride) { |
+void vpx_idct16x16_1_add_c(const tran_low_t *input, uint8_t *dest, int stride) { |
int i, j; |
tran_high_t a1; |
tran_low_t out = WRAPLOW(dct_const_round_shift(input[0] * cospi_16_64), 8); |
@@ -1146,7 +1146,7 @@ void idct32_c(const tran_low_t *input, tran_low_t *output) { |
output[31] = WRAPLOW(step1[0] - step1[31], 8); |
} |
-void vp9_idct32x32_1024_add_c(const tran_low_t *input, uint8_t *dest, |
+void vpx_idct32x32_1024_add_c(const tran_low_t *input, uint8_t *dest, |
int stride) { |
tran_low_t out[32 * 32]; |
tran_low_t *outptr = out; |
@@ -1185,7 +1185,7 @@ void vp9_idct32x32_1024_add_c(const tran_low_t *input, uint8_t *dest, |
} |
} |
-void vp9_idct32x32_34_add_c(const tran_low_t *input, uint8_t *dest, |
+void vpx_idct32x32_34_add_c(const tran_low_t *input, uint8_t *dest, |
int stride) { |
tran_low_t out[32 * 32] = {0}; |
tran_low_t *outptr = out; |
@@ -1212,7 +1212,7 @@ void vp9_idct32x32_34_add_c(const tran_low_t *input, uint8_t *dest, |
} |
} |
-void vp9_idct32x32_1_add_c(const tran_low_t *input, uint8_t *dest, int stride) { |
+void vpx_idct32x32_1_add_c(const tran_low_t *input, uint8_t *dest, int stride) { |
int i, j; |
tran_high_t a1; |
@@ -1228,7 +1228,7 @@ void vp9_idct32x32_1_add_c(const tran_low_t *input, uint8_t *dest, int stride) { |
} |
#if CONFIG_VP9_HIGHBITDEPTH |
-void vp9_highbd_iwht4x4_16_add_c(const tran_low_t *input, uint8_t *dest8, |
+void vpx_highbd_iwht4x4_16_add_c(const tran_low_t *input, uint8_t *dest8, |
int stride, int bd) { |
/* 4-point reversible, orthonormal inverse Walsh-Hadamard in 3.5 adds, |
0.5 shifts per pixel. */ |
@@ -1282,7 +1282,7 @@ void vp9_highbd_iwht4x4_16_add_c(const tran_low_t *input, uint8_t *dest8, |
} |
} |
-void vp9_highbd_iwht4x4_1_add_c(const tran_low_t *in, uint8_t *dest8, |
+void vpx_highbd_iwht4x4_1_add_c(const tran_low_t *in, uint8_t *dest8, |
int dest_stride, int bd) { |
int i; |
tran_high_t a1, e1; |
@@ -1315,7 +1315,7 @@ void vp9_highbd_iwht4x4_1_add_c(const tran_low_t *in, uint8_t *dest8, |
} |
} |
-void vp9_highbd_idct4_c(const tran_low_t *input, tran_low_t *output, int bd) { |
+void vpx_highbd_idct4_c(const tran_low_t *input, tran_low_t *output, int bd) { |
tran_low_t step[4]; |
tran_high_t temp1, temp2; |
(void) bd; |
@@ -1336,7 +1336,7 @@ void vp9_highbd_idct4_c(const tran_low_t *input, tran_low_t *output, int bd) { |
output[3] = WRAPLOW(step[0] - step[3], bd); |
} |
-void vp9_highbd_idct4x4_16_add_c(const tran_low_t *input, uint8_t *dest8, |
+void vpx_highbd_idct4x4_16_add_c(const tran_low_t *input, uint8_t *dest8, |
int stride, int bd) { |
tran_low_t out[4 * 4]; |
tran_low_t *outptr = out; |
@@ -1346,7 +1346,7 @@ void vp9_highbd_idct4x4_16_add_c(const tran_low_t *input, uint8_t *dest8, |
// Rows |
for (i = 0; i < 4; ++i) { |
- vp9_highbd_idct4_c(input, outptr, bd); |
+ vpx_highbd_idct4_c(input, outptr, bd); |
input += 4; |
outptr += 4; |
} |
@@ -1355,7 +1355,7 @@ void vp9_highbd_idct4x4_16_add_c(const tran_low_t *input, uint8_t *dest8, |
for (i = 0; i < 4; ++i) { |
for (j = 0; j < 4; ++j) |
temp_in[j] = out[j * 4 + i]; |
- vp9_highbd_idct4_c(temp_in, temp_out, bd); |
+ vpx_highbd_idct4_c(temp_in, temp_out, bd); |
for (j = 0; j < 4; ++j) { |
dest[j * stride + i] = highbd_clip_pixel_add( |
dest[j * stride + i], ROUND_POWER_OF_TWO(temp_out[j], 4), bd); |
@@ -1363,7 +1363,7 @@ void vp9_highbd_idct4x4_16_add_c(const tran_low_t *input, uint8_t *dest8, |
} |
} |
-void vp9_highbd_idct4x4_1_add_c(const tran_low_t *input, uint8_t *dest8, |
+void vpx_highbd_idct4x4_1_add_c(const tran_low_t *input, uint8_t *dest8, |
int dest_stride, int bd) { |
int i; |
tran_high_t a1; |
@@ -1383,7 +1383,7 @@ void vp9_highbd_idct4x4_1_add_c(const tran_low_t *input, uint8_t *dest8, |
} |
} |
-void vp9_highbd_idct8_c(const tran_low_t *input, tran_low_t *output, int bd) { |
+void vpx_highbd_idct8_c(const tran_low_t *input, tran_low_t *output, int bd) { |
tran_low_t step1[8], step2[8]; |
tran_high_t temp1, temp2; |
// stage 1 |
@@ -1401,7 +1401,7 @@ void vp9_highbd_idct8_c(const tran_low_t *input, tran_low_t *output, int bd) { |
step1[6] = WRAPLOW(highbd_dct_const_round_shift(temp2, bd), bd); |
// stage 2 & stage 3 - even half |
- vp9_highbd_idct4_c(step1, step1, bd); |
+ vpx_highbd_idct4_c(step1, step1, bd); |
// stage 2 - odd half |
step2[4] = WRAPLOW(step1[4] + step1[5], bd); |
@@ -1428,7 +1428,7 @@ void vp9_highbd_idct8_c(const tran_low_t *input, tran_low_t *output, int bd) { |
output[7] = WRAPLOW(step1[0] - step1[7], bd); |
} |
-void vp9_highbd_idct8x8_64_add_c(const tran_low_t *input, uint8_t *dest8, |
+void vpx_highbd_idct8x8_64_add_c(const tran_low_t *input, uint8_t *dest8, |
int stride, int bd) { |
tran_low_t out[8 * 8]; |
tran_low_t *outptr = out; |
@@ -1438,7 +1438,7 @@ void vp9_highbd_idct8x8_64_add_c(const tran_low_t *input, uint8_t *dest8, |
// First transform rows. |
for (i = 0; i < 8; ++i) { |
- vp9_highbd_idct8_c(input, outptr, bd); |
+ vpx_highbd_idct8_c(input, outptr, bd); |
input += 8; |
outptr += 8; |
} |
@@ -1447,7 +1447,7 @@ void vp9_highbd_idct8x8_64_add_c(const tran_low_t *input, uint8_t *dest8, |
for (i = 0; i < 8; ++i) { |
for (j = 0; j < 8; ++j) |
temp_in[j] = out[j * 8 + i]; |
- vp9_highbd_idct8_c(temp_in, temp_out, bd); |
+ vpx_highbd_idct8_c(temp_in, temp_out, bd); |
for (j = 0; j < 8; ++j) { |
dest[j * stride + i] = highbd_clip_pixel_add( |
dest[j * stride + i], ROUND_POWER_OF_TWO(temp_out[j], 5), bd); |
@@ -1455,7 +1455,7 @@ void vp9_highbd_idct8x8_64_add_c(const tran_low_t *input, uint8_t *dest8, |
} |
} |
-void vp9_highbd_idct8x8_1_add_c(const tran_low_t *input, uint8_t *dest8, |
+void vpx_highbd_idct8x8_1_add_c(const tran_low_t *input, uint8_t *dest8, |
int stride, int bd) { |
int i, j; |
tran_high_t a1; |
@@ -1471,7 +1471,7 @@ void vp9_highbd_idct8x8_1_add_c(const tran_low_t *input, uint8_t *dest8, |
} |
} |
-void highbd_iadst4_c(const tran_low_t *input, tran_low_t *output, int bd) { |
+void vpx_highbd_iadst4_c(const tran_low_t *input, tran_low_t *output, int bd) { |
tran_high_t s0, s1, s2, s3, s4, s5, s6, s7; |
tran_low_t x0 = input[0]; |
@@ -1509,7 +1509,7 @@ void highbd_iadst4_c(const tran_low_t *input, tran_low_t *output, int bd) { |
output[3] = WRAPLOW(highbd_dct_const_round_shift(s0 + s1 - s3, bd), bd); |
} |
-void highbd_iadst8_c(const tran_low_t *input, tran_low_t *output, int bd) { |
+void vpx_highbd_iadst8_c(const tran_low_t *input, tran_low_t *output, int bd) { |
tran_high_t s0, s1, s2, s3, s4, s5, s6, s7; |
tran_low_t x0 = input[7]; |
@@ -1586,7 +1586,7 @@ void highbd_iadst8_c(const tran_low_t *input, tran_low_t *output, int bd) { |
output[7] = WRAPLOW(-x1, bd); |
} |
-void vp9_highbd_idct8x8_10_add_c(const tran_low_t *input, uint8_t *dest8, |
+void vpx_highbd_idct8x8_10_add_c(const tran_low_t *input, uint8_t *dest8, |
int stride, int bd) { |
tran_low_t out[8 * 8] = { 0 }; |
tran_low_t *outptr = out; |
@@ -1597,7 +1597,7 @@ void vp9_highbd_idct8x8_10_add_c(const tran_low_t *input, uint8_t *dest8, |
// First transform rows. |
// Only first 4 row has non-zero coefs. |
for (i = 0; i < 4; ++i) { |
- vp9_highbd_idct8_c(input, outptr, bd); |
+ vpx_highbd_idct8_c(input, outptr, bd); |
input += 8; |
outptr += 8; |
} |
@@ -1605,7 +1605,7 @@ void vp9_highbd_idct8x8_10_add_c(const tran_low_t *input, uint8_t *dest8, |
for (i = 0; i < 8; ++i) { |
for (j = 0; j < 8; ++j) |
temp_in[j] = out[j * 8 + i]; |
- vp9_highbd_idct8_c(temp_in, temp_out, bd); |
+ vpx_highbd_idct8_c(temp_in, temp_out, bd); |
for (j = 0; j < 8; ++j) { |
dest[j * stride + i] = highbd_clip_pixel_add( |
dest[j * stride + i], ROUND_POWER_OF_TWO(temp_out[j], 5), bd); |
@@ -1613,7 +1613,7 @@ void vp9_highbd_idct8x8_10_add_c(const tran_low_t *input, uint8_t *dest8, |
} |
} |
-void vp9_highbd_idct16_c(const tran_low_t *input, tran_low_t *output, int bd) { |
+void vpx_highbd_idct16_c(const tran_low_t *input, tran_low_t *output, int bd) { |
tran_low_t step1[16], step2[16]; |
tran_high_t temp1, temp2; |
(void) bd; |
@@ -1779,7 +1779,7 @@ void vp9_highbd_idct16_c(const tran_low_t *input, tran_low_t *output, int bd) { |
output[15] = WRAPLOW(step2[0] - step2[15], bd); |
} |
-void vp9_highbd_idct16x16_256_add_c(const tran_low_t *input, uint8_t *dest8, |
+void vpx_highbd_idct16x16_256_add_c(const tran_low_t *input, uint8_t *dest8, |
int stride, int bd) { |
tran_low_t out[16 * 16]; |
tran_low_t *outptr = out; |
@@ -1789,7 +1789,7 @@ void vp9_highbd_idct16x16_256_add_c(const tran_low_t *input, uint8_t *dest8, |
// First transform rows. |
for (i = 0; i < 16; ++i) { |
- vp9_highbd_idct16_c(input, outptr, bd); |
+ vpx_highbd_idct16_c(input, outptr, bd); |
input += 16; |
outptr += 16; |
} |
@@ -1798,7 +1798,7 @@ void vp9_highbd_idct16x16_256_add_c(const tran_low_t *input, uint8_t *dest8, |
for (i = 0; i < 16; ++i) { |
for (j = 0; j < 16; ++j) |
temp_in[j] = out[j * 16 + i]; |
- vp9_highbd_idct16_c(temp_in, temp_out, bd); |
+ vpx_highbd_idct16_c(temp_in, temp_out, bd); |
for (j = 0; j < 16; ++j) { |
dest[j * stride + i] = highbd_clip_pixel_add( |
dest[j * stride + i], ROUND_POWER_OF_TWO(temp_out[j], 6), bd); |
@@ -1806,7 +1806,7 @@ void vp9_highbd_idct16x16_256_add_c(const tran_low_t *input, uint8_t *dest8, |
} |
} |
-void highbd_iadst16_c(const tran_low_t *input, tran_low_t *output, int bd) { |
+void vpx_highbd_iadst16_c(const tran_low_t *input, tran_low_t *output, int bd) { |
tran_high_t s0, s1, s2, s3, s4, s5, s6, s7, s8; |
tran_high_t s9, s10, s11, s12, s13, s14, s15; |
@@ -1976,7 +1976,7 @@ void highbd_iadst16_c(const tran_low_t *input, tran_low_t *output, int bd) { |
output[15] = WRAPLOW(-x1, bd); |
} |
-void vp9_highbd_idct16x16_10_add_c(const tran_low_t *input, uint8_t *dest8, |
+void vpx_highbd_idct16x16_10_add_c(const tran_low_t *input, uint8_t *dest8, |
int stride, int bd) { |
tran_low_t out[16 * 16] = { 0 }; |
tran_low_t *outptr = out; |
@@ -1987,7 +1987,7 @@ void vp9_highbd_idct16x16_10_add_c(const tran_low_t *input, uint8_t *dest8, |
// First transform rows. Since all non-zero dct coefficients are in |
// upper-left 4x4 area, we only need to calculate first 4 rows here. |
for (i = 0; i < 4; ++i) { |
- vp9_highbd_idct16_c(input, outptr, bd); |
+ vpx_highbd_idct16_c(input, outptr, bd); |
input += 16; |
outptr += 16; |
} |
@@ -1996,7 +1996,7 @@ void vp9_highbd_idct16x16_10_add_c(const tran_low_t *input, uint8_t *dest8, |
for (i = 0; i < 16; ++i) { |
for (j = 0; j < 16; ++j) |
temp_in[j] = out[j*16 + i]; |
- vp9_highbd_idct16_c(temp_in, temp_out, bd); |
+ vpx_highbd_idct16_c(temp_in, temp_out, bd); |
for (j = 0; j < 16; ++j) { |
dest[j * stride + i] = highbd_clip_pixel_add( |
dest[j * stride + i], ROUND_POWER_OF_TWO(temp_out[j], 6), bd); |
@@ -2004,7 +2004,7 @@ void vp9_highbd_idct16x16_10_add_c(const tran_low_t *input, uint8_t *dest8, |
} |
} |
-void vp9_highbd_idct16x16_1_add_c(const tran_low_t *input, uint8_t *dest8, |
+void vpx_highbd_idct16x16_1_add_c(const tran_low_t *input, uint8_t *dest8, |
int stride, int bd) { |
int i, j; |
tran_high_t a1; |
@@ -2021,7 +2021,8 @@ void vp9_highbd_idct16x16_1_add_c(const tran_low_t *input, uint8_t *dest8, |
} |
} |
-void highbd_idct32_c(const tran_low_t *input, tran_low_t *output, int bd) { |
+static void highbd_idct32_c(const tran_low_t *input, |
+ tran_low_t *output, int bd) { |
tran_low_t step1[32], step2[32]; |
tran_high_t temp1, temp2; |
(void) bd; |
@@ -2389,7 +2390,7 @@ void highbd_idct32_c(const tran_low_t *input, tran_low_t *output, int bd) { |
output[31] = WRAPLOW(step1[0] - step1[31], bd); |
} |
-void vp9_highbd_idct32x32_1024_add_c(const tran_low_t *input, uint8_t *dest8, |
+void vpx_highbd_idct32x32_1024_add_c(const tran_low_t *input, uint8_t *dest8, |
int stride, int bd) { |
tran_low_t out[32 * 32]; |
tran_low_t *outptr = out; |
@@ -2429,7 +2430,7 @@ void vp9_highbd_idct32x32_1024_add_c(const tran_low_t *input, uint8_t *dest8, |
} |
} |
-void vp9_highbd_idct32x32_34_add_c(const tran_low_t *input, uint8_t *dest8, |
+void vpx_highbd_idct32x32_34_add_c(const tran_low_t *input, uint8_t *dest8, |
int stride, int bd) { |
tran_low_t out[32 * 32] = {0}; |
tran_low_t *outptr = out; |
@@ -2456,7 +2457,7 @@ void vp9_highbd_idct32x32_34_add_c(const tran_low_t *input, uint8_t *dest8, |
} |
} |
-void vp9_highbd_idct32x32_1_add_c(const tran_low_t *input, uint8_t *dest8, |
+void vpx_highbd_idct32x32_1_add_c(const tran_low_t *input, uint8_t *dest8, |
int stride, int bd) { |
int i, j; |
int a1; |