Index: source/libvpx/test/dct16x16_test.cc |
=================================================================== |
--- source/libvpx/test/dct16x16_test.cc (revision 292072) |
+++ source/libvpx/test/dct16x16_test.cc (working copy) |
@@ -20,12 +20,9 @@ |
#include "./vp9_rtcd.h" |
#include "vp9/common/vp9_entropy.h" |
+#include "vpx/vpx_codec.h" |
#include "vpx/vpx_integer.h" |
-extern "C" { |
-void vp9_idct16x16_256_add_c(const int16_t *input, uint8_t *output, int pitch); |
-} |
- |
using libvpx_test::ACMRandom; |
namespace { |
@@ -258,42 +255,72 @@ |
} |
} |
-typedef void (*FdctFunc)(const int16_t *in, int16_t *out, int stride); |
-typedef void (*IdctFunc)(const int16_t *in, uint8_t *out, int stride); |
-typedef void (*FhtFunc)(const int16_t *in, int16_t *out, int stride, |
+typedef void (*FdctFunc)(const int16_t *in, tran_low_t *out, int stride); |
+typedef void (*IdctFunc)(const tran_low_t *in, uint8_t *out, int stride); |
+typedef void (*FhtFunc)(const int16_t *in, tran_low_t *out, int stride, |
int tx_type); |
-typedef void (*IhtFunc)(const int16_t *in, uint8_t *out, int stride, |
+typedef void (*IhtFunc)(const tran_low_t *in, uint8_t *out, int stride, |
int tx_type); |
-typedef std::tr1::tuple<FdctFunc, IdctFunc, int> Dct16x16Param; |
-typedef std::tr1::tuple<FhtFunc, IhtFunc, int> Ht16x16Param; |
+typedef std::tr1::tuple<FdctFunc, IdctFunc, int, vpx_bit_depth_t> Dct16x16Param; |
+typedef std::tr1::tuple<FhtFunc, IhtFunc, int, vpx_bit_depth_t> Ht16x16Param; |
-void fdct16x16_ref(const int16_t *in, int16_t *out, int stride, |
+void fdct16x16_ref(const int16_t *in, tran_low_t *out, int stride, |
int /*tx_type*/) { |
vp9_fdct16x16_c(in, out, stride); |
} |
-void idct16x16_ref(const int16_t *in, uint8_t *dest, int stride, |
+void idct16x16_ref(const tran_low_t *in, uint8_t *dest, int stride, |
int /*tx_type*/) { |
vp9_idct16x16_256_add_c(in, dest, stride); |
} |
-void fht16x16_ref(const int16_t *in, int16_t *out, int stride, int tx_type) { |
+void fht16x16_ref(const int16_t *in, tran_low_t *out, int stride, |
+ int tx_type) { |
vp9_fht16x16_c(in, out, stride, tx_type); |
} |
-void iht16x16_ref(const int16_t *in, uint8_t *dest, int stride, int tx_type) { |
+void iht16x16_ref(const tran_low_t *in, uint8_t *dest, int stride, |
+ int tx_type) { |
vp9_iht16x16_256_add_c(in, dest, stride, tx_type); |
} |
+#if CONFIG_VP9_HIGHBITDEPTH |
+void idct16x16_10(const tran_low_t *in, uint8_t *out, int stride) { |
+ vp9_high_idct16x16_256_add_c(in, out, stride, 10); |
+} |
+ |
+void idct16x16_12(const tran_low_t *in, uint8_t *out, int stride) { |
+ vp9_high_idct16x16_256_add_c(in, out, stride, 12); |
+} |
+ |
+void idct16x16_10_ref(const tran_low_t *in, uint8_t *out, int stride, |
+ int tx_type) { |
+ idct16x16_10(in, out, stride); |
+} |
+ |
+void idct16x16_12_ref(const tran_low_t *in, uint8_t *out, int stride, |
+ int tx_type) { |
+ idct16x16_12(in, out, stride); |
+} |
+ |
+void iht16x16_10(const tran_low_t *in, uint8_t *out, int stride, int tx_type) { |
+ vp9_high_iht16x16_256_add_c(in, out, stride, tx_type, 10); |
+} |
+ |
+void iht16x16_12(const tran_low_t *in, uint8_t *out, int stride, int tx_type) { |
+ vp9_high_iht16x16_256_add_c(in, out, stride, tx_type, 12); |
+} |
+#endif |
+ |
class Trans16x16TestBase { |
public: |
virtual ~Trans16x16TestBase() {} |
protected: |
- virtual void RunFwdTxfm(int16_t *in, int16_t *out, int stride) = 0; |
+ virtual void RunFwdTxfm(int16_t *in, tran_low_t *out, int stride) = 0; |
- virtual void RunInvTxfm(int16_t *out, uint8_t *dst, int stride) = 0; |
+ virtual void RunInvTxfm(tran_low_t *out, uint8_t *dst, int stride) = 0; |
void RunAccuracyCheck() { |
ACMRandom rnd(ACMRandom::DeterministicSeed()); |
@@ -302,23 +329,48 @@ |
const int count_test_block = 10000; |
for (int i = 0; i < count_test_block; ++i) { |
DECLARE_ALIGNED_ARRAY(16, int16_t, test_input_block, kNumCoeffs); |
- DECLARE_ALIGNED_ARRAY(16, int16_t, test_temp_block, kNumCoeffs); |
+ DECLARE_ALIGNED_ARRAY(16, tran_low_t, test_temp_block, kNumCoeffs); |
DECLARE_ALIGNED_ARRAY(16, uint8_t, dst, kNumCoeffs); |
DECLARE_ALIGNED_ARRAY(16, uint8_t, src, kNumCoeffs); |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ DECLARE_ALIGNED_ARRAY(16, uint16_t, dst16, kNumCoeffs); |
+ DECLARE_ALIGNED_ARRAY(16, uint16_t, src16, kNumCoeffs); |
+#endif |
- // Initialize a test block with input range [-255, 255]. |
+ // Initialize a test block with input range [-mask_, mask_]. |
for (int j = 0; j < kNumCoeffs; ++j) { |
- src[j] = rnd.Rand8(); |
- dst[j] = rnd.Rand8(); |
- test_input_block[j] = src[j] - dst[j]; |
+ if (bit_depth_ == VPX_BITS_8) { |
+ src[j] = rnd.Rand8(); |
+ dst[j] = rnd.Rand8(); |
+ test_input_block[j] = src[j] - dst[j]; |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ } else { |
+ src16[j] = rnd.Rand16() & mask_; |
+ dst16[j] = rnd.Rand16() & mask_; |
+ test_input_block[j] = src16[j] - dst16[j]; |
+#endif |
+ } |
} |
ASM_REGISTER_STATE_CHECK(RunFwdTxfm(test_input_block, |
test_temp_block, pitch_)); |
- ASM_REGISTER_STATE_CHECK(RunInvTxfm(test_temp_block, dst, pitch_)); |
+ if (bit_depth_ == VPX_BITS_8) { |
+ ASM_REGISTER_STATE_CHECK( |
+ RunInvTxfm(test_temp_block, dst, pitch_)); |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ } else { |
+ ASM_REGISTER_STATE_CHECK( |
+ RunInvTxfm(test_temp_block, CONVERT_TO_BYTEPTR(dst16), pitch_)); |
+#endif |
+ } |
for (int j = 0; j < kNumCoeffs; ++j) { |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ const uint32_t diff = |
+ bit_depth_ == VPX_BITS_8 ? dst[j] - src[j] : dst16[j] - src16[j]; |
+#else |
const uint32_t diff = dst[j] - src[j]; |
+#endif |
const uint32_t error = diff * diff; |
if (max_error < error) |
max_error = error; |
@@ -326,10 +378,10 @@ |
} |
} |
- EXPECT_GE(1u, max_error) |
+ EXPECT_GE(1u << 2 * (bit_depth_ - 8), max_error) |
<< "Error: 16x16 FHT/IHT has an individual round trip error > 1"; |
- EXPECT_GE(count_test_block , total_error) |
+ EXPECT_GE(count_test_block << 2 * (bit_depth_ - 8), total_error) |
<< "Error: 16x16 FHT/IHT has average round trip error > 1 per block"; |
} |
@@ -337,13 +389,13 @@ |
ACMRandom rnd(ACMRandom::DeterministicSeed()); |
const int count_test_block = 1000; |
DECLARE_ALIGNED_ARRAY(16, int16_t, input_block, kNumCoeffs); |
- DECLARE_ALIGNED_ARRAY(16, int16_t, output_ref_block, kNumCoeffs); |
- DECLARE_ALIGNED_ARRAY(16, int16_t, output_block, kNumCoeffs); |
+ DECLARE_ALIGNED_ARRAY(16, tran_low_t, output_ref_block, kNumCoeffs); |
+ DECLARE_ALIGNED_ARRAY(16, tran_low_t, output_block, kNumCoeffs); |
for (int i = 0; i < count_test_block; ++i) { |
- // Initialize a test block with input range [-255, 255]. |
+ // Initialize a test block with input range [-mask_, mask_]. |
for (int j = 0; j < kNumCoeffs; ++j) |
- input_block[j] = rnd.Rand8() - rnd.Rand8(); |
+ input_block[j] = (rnd.Rand16() & mask_) - (rnd.Rand16() & mask_); |
fwd_txfm_ref(input_block, output_ref_block, pitch_, tx_type_); |
ASM_REGISTER_STATE_CHECK(RunFwdTxfm(input_block, output_block, pitch_)); |
@@ -359,21 +411,21 @@ |
const int count_test_block = 1000; |
DECLARE_ALIGNED_ARRAY(16, int16_t, input_block, kNumCoeffs); |
DECLARE_ALIGNED_ARRAY(16, int16_t, input_extreme_block, kNumCoeffs); |
- DECLARE_ALIGNED_ARRAY(16, int16_t, output_ref_block, kNumCoeffs); |
- DECLARE_ALIGNED_ARRAY(16, int16_t, output_block, kNumCoeffs); |
+ DECLARE_ALIGNED_ARRAY(16, tran_low_t, output_ref_block, kNumCoeffs); |
+ DECLARE_ALIGNED_ARRAY(16, tran_low_t, output_block, kNumCoeffs); |
for (int i = 0; i < count_test_block; ++i) { |
- // Initialize a test block with input range [-255, 255]. |
+ // Initialize a test block with input range [-mask_, mask_]. |
for (int j = 0; j < kNumCoeffs; ++j) { |
- input_block[j] = rnd.Rand8() - rnd.Rand8(); |
- input_extreme_block[j] = rnd.Rand8() % 2 ? 255 : -255; |
+ input_block[j] = (rnd.Rand16() & mask_) - (rnd.Rand16() & mask_); |
+ input_extreme_block[j] = rnd.Rand8() % 2 ? mask_ : -mask_; |
} |
if (i == 0) { |
for (int j = 0; j < kNumCoeffs; ++j) |
- input_extreme_block[j] = 255; |
+ input_extreme_block[j] = mask_; |
} else if (i == 1) { |
for (int j = 0; j < kNumCoeffs; ++j) |
- input_extreme_block[j] = -255; |
+ input_extreme_block[j] = -mask_; |
} |
fwd_txfm_ref(input_extreme_block, output_ref_block, pitch_, tx_type_); |
@@ -383,7 +435,7 @@ |
// The minimum quant value is 4. |
for (int j = 0; j < kNumCoeffs; ++j) { |
EXPECT_EQ(output_block[j], output_ref_block[j]); |
- EXPECT_GE(4 * DCT_MAX_VALUE, abs(output_block[j])) |
+ EXPECT_GE(4 * DCT_MAX_VALUE << (bit_depth_ - 8), abs(output_block[j])) |
<< "Error: 16x16 FDCT has coefficient larger than 4*DCT_MAX_VALUE"; |
} |
} |
@@ -394,39 +446,65 @@ |
const int count_test_block = 1000; |
DECLARE_ALIGNED_ARRAY(16, int16_t, input_block, kNumCoeffs); |
DECLARE_ALIGNED_ARRAY(16, int16_t, input_extreme_block, kNumCoeffs); |
- DECLARE_ALIGNED_ARRAY(16, int16_t, output_ref_block, kNumCoeffs); |
+ DECLARE_ALIGNED_ARRAY(16, tran_low_t, output_ref_block, kNumCoeffs); |
DECLARE_ALIGNED_ARRAY(16, uint8_t, dst, kNumCoeffs); |
DECLARE_ALIGNED_ARRAY(16, uint8_t, ref, kNumCoeffs); |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ DECLARE_ALIGNED_ARRAY(16, uint16_t, dst16, kNumCoeffs); |
+ DECLARE_ALIGNED_ARRAY(16, uint16_t, ref16, kNumCoeffs); |
+#endif |
for (int i = 0; i < count_test_block; ++i) { |
- // Initialize a test block with input range [-255, 255]. |
+ // Initialize a test block with input range [-mask_, mask_]. |
for (int j = 0; j < kNumCoeffs; ++j) { |
- input_block[j] = rnd.Rand8() - rnd.Rand8(); |
- input_extreme_block[j] = rnd.Rand8() % 2 ? 255 : -255; |
+ if (bit_depth_ == VPX_BITS_8) |
+ input_block[j] = rnd.Rand8() - rnd.Rand8(); |
+ else |
+ input_block[j] = (rnd.Rand16() & mask_) - (rnd.Rand16() & mask_); |
+ input_extreme_block[j] = rnd.Rand8() % 2 ? mask_ : -mask_; |
} |
if (i == 0) |
for (int j = 0; j < kNumCoeffs; ++j) |
- input_extreme_block[j] = 255; |
+ input_extreme_block[j] = mask_; |
if (i == 1) |
for (int j = 0; j < kNumCoeffs; ++j) |
- input_extreme_block[j] = -255; |
+ input_extreme_block[j] = -mask_; |
fwd_txfm_ref(input_extreme_block, output_ref_block, pitch_, tx_type_); |
// clear reconstructed pixel buffers |
vpx_memset(dst, 0, kNumCoeffs * sizeof(uint8_t)); |
vpx_memset(ref, 0, kNumCoeffs * sizeof(uint8_t)); |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ vpx_memset(dst16, 0, kNumCoeffs * sizeof(uint16_t)); |
+ vpx_memset(ref16, 0, kNumCoeffs * sizeof(uint16_t)); |
+#endif |
// quantization with maximum allowed step sizes |
output_ref_block[0] = (output_ref_block[0] / dc_thred) * dc_thred; |
for (int j = 1; j < kNumCoeffs; ++j) |
output_ref_block[j] = (output_ref_block[j] / ac_thred) * ac_thred; |
- inv_txfm_ref(output_ref_block, ref, pitch_, tx_type_); |
- ASM_REGISTER_STATE_CHECK(RunInvTxfm(output_ref_block, dst, pitch_)); |
- |
- for (int j = 0; j < kNumCoeffs; ++j) |
- EXPECT_EQ(ref[j], dst[j]); |
+ if (bit_depth_ == VPX_BITS_8) { |
+ inv_txfm_ref(output_ref_block, ref, pitch_, tx_type_); |
+ ASM_REGISTER_STATE_CHECK(RunInvTxfm(output_ref_block, dst, pitch_)); |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ } else { |
+ inv_txfm_ref(output_ref_block, CONVERT_TO_BYTEPTR(ref16), pitch_, |
+ tx_type_); |
+ ASM_REGISTER_STATE_CHECK(RunInvTxfm(output_ref_block, |
+ CONVERT_TO_BYTEPTR(dst16), pitch_)); |
+#endif |
+ } |
+ if (bit_depth_ == VPX_BITS_8) { |
+ for (int j = 0; j < kNumCoeffs; ++j) |
+ EXPECT_EQ(ref[j], dst[j]); |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ } else { |
+ for (int j = 0; j < kNumCoeffs; ++j) |
+ EXPECT_EQ(ref16[j], dst16[j]); |
+#endif |
+ } |
} |
} |
@@ -434,28 +512,52 @@ |
ACMRandom rnd(ACMRandom::DeterministicSeed()); |
const int count_test_block = 1000; |
DECLARE_ALIGNED_ARRAY(16, int16_t, in, kNumCoeffs); |
- DECLARE_ALIGNED_ARRAY(16, int16_t, coeff, kNumCoeffs); |
+ DECLARE_ALIGNED_ARRAY(16, tran_low_t, coeff, kNumCoeffs); |
DECLARE_ALIGNED_ARRAY(16, uint8_t, dst, kNumCoeffs); |
DECLARE_ALIGNED_ARRAY(16, uint8_t, src, kNumCoeffs); |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ DECLARE_ALIGNED_ARRAY(16, uint16_t, dst16, kNumCoeffs); |
+ DECLARE_ALIGNED_ARRAY(16, uint16_t, src16, kNumCoeffs); |
+#endif |
for (int i = 0; i < count_test_block; ++i) { |
double out_r[kNumCoeffs]; |
// Initialize a test block with input range [-255, 255]. |
for (int j = 0; j < kNumCoeffs; ++j) { |
- src[j] = rnd.Rand8(); |
- dst[j] = rnd.Rand8(); |
- in[j] = src[j] - dst[j]; |
+ if (bit_depth_ == VPX_BITS_8) { |
+ src[j] = rnd.Rand8(); |
+ dst[j] = rnd.Rand8(); |
+ in[j] = src[j] - dst[j]; |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ } else { |
+ src16[j] = rnd.Rand16() & mask_; |
+ dst16[j] = rnd.Rand16() & mask_; |
+ in[j] = src16[j] - dst16[j]; |
+#endif |
+ } |
} |
reference_16x16_dct_2d(in, out_r); |
for (int j = 0; j < kNumCoeffs; ++j) |
coeff[j] = round(out_r[j]); |
- ASM_REGISTER_STATE_CHECK(RunInvTxfm(coeff, dst, 16)); |
+ if (bit_depth_ == VPX_BITS_8) { |
+ ASM_REGISTER_STATE_CHECK(RunInvTxfm(coeff, dst, 16)); |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ } else { |
+ ASM_REGISTER_STATE_CHECK(RunInvTxfm(coeff, CONVERT_TO_BYTEPTR(dst16), |
+ 16)); |
+#endif |
+ } |
for (int j = 0; j < kNumCoeffs; ++j) { |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ const uint32_t diff = |
+ bit_depth_ == VPX_BITS_8 ? dst[j] - src[j] : dst16[j] - src16[j]; |
+#else |
const uint32_t diff = dst[j] - src[j]; |
+#endif |
const uint32_t error = diff * diff; |
EXPECT_GE(1u, error) |
<< "Error: 16x16 IDCT has error " << error |
@@ -465,6 +567,8 @@ |
} |
int pitch_; |
int tx_type_; |
+ vpx_bit_depth_t bit_depth_; |
+ int mask_; |
FhtFunc fwd_txfm_ref; |
IhtFunc inv_txfm_ref; |
}; |
@@ -479,17 +583,34 @@ |
fwd_txfm_ = GET_PARAM(0); |
inv_txfm_ = GET_PARAM(1); |
tx_type_ = GET_PARAM(2); |
+ bit_depth_ = GET_PARAM(3); |
pitch_ = 16; |
fwd_txfm_ref = fdct16x16_ref; |
inv_txfm_ref = idct16x16_ref; |
+ mask_ = (1 << bit_depth_) - 1; |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ switch (bit_depth_) { |
+ case 10: |
+ inv_txfm_ref = idct16x16_10_ref; |
+ break; |
+ case 12: |
+ inv_txfm_ref = idct16x16_12_ref; |
+ break; |
+ default: |
+ inv_txfm_ref = idct16x16_ref; |
+ break; |
+ } |
+#else |
+ inv_txfm_ref = idct16x16_ref; |
+#endif |
} |
virtual void TearDown() { libvpx_test::ClearSystemState(); } |
protected: |
- void RunFwdTxfm(int16_t *in, int16_t *out, int stride) { |
+ void RunFwdTxfm(int16_t *in, tran_low_t *out, int stride) { |
fwd_txfm_(in, out, stride); |
} |
- void RunInvTxfm(int16_t *out, uint8_t *dst, int stride) { |
+ void RunInvTxfm(tran_low_t *out, uint8_t *dst, int stride) { |
inv_txfm_(out, dst, stride); |
} |
@@ -529,17 +650,34 @@ |
fwd_txfm_ = GET_PARAM(0); |
inv_txfm_ = GET_PARAM(1); |
tx_type_ = GET_PARAM(2); |
+ bit_depth_ = GET_PARAM(3); |
pitch_ = 16; |
fwd_txfm_ref = fht16x16_ref; |
inv_txfm_ref = iht16x16_ref; |
+ mask_ = (1 << bit_depth_) - 1; |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ switch (bit_depth_) { |
+ case VPX_BITS_10: |
+ inv_txfm_ref = iht16x16_10; |
+ break; |
+ case VPX_BITS_12: |
+ inv_txfm_ref = iht16x16_12; |
+ break; |
+ default: |
+ inv_txfm_ref = iht16x16_ref; |
+ break; |
+ } |
+#else |
+ inv_txfm_ref = iht16x16_ref; |
+#endif |
} |
virtual void TearDown() { libvpx_test::ClearSystemState(); } |
protected: |
- void RunFwdTxfm(int16_t *in, int16_t *out, int stride) { |
+ void RunFwdTxfm(int16_t *in, tran_low_t *out, int stride) { |
fwd_txfm_(in, out, stride, tx_type_); |
} |
- void RunInvTxfm(int16_t *out, uint8_t *dst, int stride) { |
+ void RunInvTxfm(tran_low_t *out, uint8_t *dst, int stride) { |
inv_txfm_(out, dst, stride, tx_type_); |
} |
@@ -567,45 +705,78 @@ |
using std::tr1::make_tuple; |
+#if CONFIG_VP9_HIGHBITDEPTH |
INSTANTIATE_TEST_CASE_P( |
C, Trans16x16DCT, |
::testing::Values( |
- make_tuple(&vp9_fdct16x16_c, &vp9_idct16x16_256_add_c, 0))); |
+ make_tuple(&vp9_high_fdct16x16_c, &idct16x16_10, 0, VPX_BITS_10), |
+ make_tuple(&vp9_high_fdct16x16_c, &idct16x16_12, 0, VPX_BITS_12), |
+ make_tuple(&vp9_fdct16x16_c, &vp9_idct16x16_256_add_c, 0, VPX_BITS_8))); |
+#else |
INSTANTIATE_TEST_CASE_P( |
+ C, Trans16x16DCT, |
+ ::testing::Values( |
+ make_tuple(&vp9_fdct16x16_c, &vp9_idct16x16_256_add_c, 0, VPX_BITS_8))); |
+#endif |
+ |
+#if CONFIG_VP9_HIGHBITDEPTH |
+INSTANTIATE_TEST_CASE_P( |
C, Trans16x16HT, |
::testing::Values( |
- make_tuple(&vp9_fht16x16_c, &vp9_iht16x16_256_add_c, 0), |
- make_tuple(&vp9_fht16x16_c, &vp9_iht16x16_256_add_c, 1), |
- make_tuple(&vp9_fht16x16_c, &vp9_iht16x16_256_add_c, 2), |
- make_tuple(&vp9_fht16x16_c, &vp9_iht16x16_256_add_c, 3))); |
+ make_tuple(&vp9_high_fht16x16_c, &iht16x16_10, 0, VPX_BITS_10), |
+ make_tuple(&vp9_high_fht16x16_c, &iht16x16_10, 1, VPX_BITS_10), |
+ make_tuple(&vp9_high_fht16x16_c, &iht16x16_10, 2, VPX_BITS_10), |
+ make_tuple(&vp9_high_fht16x16_c, &iht16x16_10, 3, VPX_BITS_10), |
+ make_tuple(&vp9_high_fht16x16_c, &iht16x16_12, 0, VPX_BITS_12), |
+ make_tuple(&vp9_high_fht16x16_c, &iht16x16_12, 1, VPX_BITS_12), |
+ make_tuple(&vp9_high_fht16x16_c, &iht16x16_12, 2, VPX_BITS_12), |
+ make_tuple(&vp9_high_fht16x16_c, &iht16x16_12, 3, VPX_BITS_12), |
+ make_tuple(&vp9_fht16x16_c, &vp9_iht16x16_256_add_c, 0, VPX_BITS_8), |
+ make_tuple(&vp9_fht16x16_c, &vp9_iht16x16_256_add_c, 1, VPX_BITS_8), |
+ make_tuple(&vp9_fht16x16_c, &vp9_iht16x16_256_add_c, 2, VPX_BITS_8), |
+ make_tuple(&vp9_fht16x16_c, &vp9_iht16x16_256_add_c, 3, VPX_BITS_8))); |
+#else |
+INSTANTIATE_TEST_CASE_P( |
+ C, Trans16x16HT, |
+ ::testing::Values( |
+ make_tuple(&vp9_fht16x16_c, &vp9_iht16x16_256_add_c, 0, VPX_BITS_8), |
+ make_tuple(&vp9_fht16x16_c, &vp9_iht16x16_256_add_c, 1, VPX_BITS_8), |
+ make_tuple(&vp9_fht16x16_c, &vp9_iht16x16_256_add_c, 2, VPX_BITS_8), |
+ make_tuple(&vp9_fht16x16_c, &vp9_iht16x16_256_add_c, 3, VPX_BITS_8))); |
+#endif |
-#if HAVE_NEON_ASM |
+#if HAVE_NEON_ASM && !CONFIG_VP9_HIGHBITDEPTH |
INSTANTIATE_TEST_CASE_P( |
NEON, Trans16x16DCT, |
::testing::Values( |
make_tuple(&vp9_fdct16x16_c, |
- &vp9_idct16x16_256_add_neon, 0))); |
+ &vp9_idct16x16_256_add_neon, 0, VPX_BITS_8))); |
#endif |
-#if HAVE_SSE2 |
+#if HAVE_SSE2 && !CONFIG_VP9_HIGHBITDEPTH |
INSTANTIATE_TEST_CASE_P( |
SSE2, Trans16x16DCT, |
::testing::Values( |
make_tuple(&vp9_fdct16x16_sse2, |
- &vp9_idct16x16_256_add_sse2, 0))); |
+ &vp9_idct16x16_256_add_sse2, 0, VPX_BITS_8))); |
INSTANTIATE_TEST_CASE_P( |
SSE2, Trans16x16HT, |
::testing::Values( |
- make_tuple(&vp9_fht16x16_sse2, &vp9_iht16x16_256_add_sse2, 0), |
- make_tuple(&vp9_fht16x16_sse2, &vp9_iht16x16_256_add_sse2, 1), |
- make_tuple(&vp9_fht16x16_sse2, &vp9_iht16x16_256_add_sse2, 2), |
- make_tuple(&vp9_fht16x16_sse2, &vp9_iht16x16_256_add_sse2, 3))); |
+ make_tuple(&vp9_fht16x16_sse2, &vp9_iht16x16_256_add_sse2, 0, |
+ VPX_BITS_8), |
+ make_tuple(&vp9_fht16x16_sse2, &vp9_iht16x16_256_add_sse2, 1, |
+ VPX_BITS_8), |
+ make_tuple(&vp9_fht16x16_sse2, &vp9_iht16x16_256_add_sse2, 2, |
+ VPX_BITS_8), |
+ make_tuple(&vp9_fht16x16_sse2, &vp9_iht16x16_256_add_sse2, 3, |
+ VPX_BITS_8))); |
#endif |
-#if HAVE_SSSE3 |
+#if HAVE_SSSE3 && !CONFIG_VP9_HIGHBITDEPTH |
INSTANTIATE_TEST_CASE_P( |
SSSE3, Trans16x16DCT, |
::testing::Values( |
- make_tuple(&vp9_fdct16x16_c, &vp9_idct16x16_256_add_ssse3, 0))); |
+ make_tuple(&vp9_fdct16x16_c, &vp9_idct16x16_256_add_ssse3, 0, |
+ VPX_BITS_8))); |
#endif |
} // namespace |