Index: source/libvpx/test/variance_test.cc |
=================================================================== |
--- source/libvpx/test/variance_test.cc (revision 293081) |
+++ source/libvpx/test/variance_test.cc (working copy) |
@@ -7,16 +7,18 @@ |
* in the file PATENTS. All contributing project authors may |
* be found in the AUTHORS file in the root of the source tree. |
*/ |
-#include <stdlib.h> |
+ |
+#include <cstdlib> |
#include <new> |
-#include "third_party/googletest/src/include/gtest/gtest.h" |
- |
+#include "test/acm_random.h" |
#include "test/clear_system_state.h" |
#include "test/register_state_check.h" |
+#include "third_party/googletest/src/include/gtest/gtest.h" |
+#include "./vpx_config.h" |
+#include "vpx/vpx_codec.h" |
#include "vpx/vpx_integer.h" |
-#include "./vpx_config.h" |
#include "vpx_mem/vpx_mem.h" |
#if CONFIG_VP8_ENCODER |
# include "./vp8_rtcd.h" |
@@ -26,7 +28,6 @@ |
# include "./vp9_rtcd.h" |
# include "vp9/encoder/vp9_variance.h" |
#endif |
-#include "test/acm_random.h" |
namespace { |
@@ -43,18 +44,50 @@ |
return res; |
} |
-static unsigned int variance_ref(const uint8_t *ref, const uint8_t *src, |
- int l2w, int l2h, unsigned int *sse_ptr) { |
+static unsigned int variance_ref(const uint8_t *src, const uint8_t *ref, |
+ int l2w, int l2h, int src_stride_coeff, |
+ int ref_stride_coeff, uint32_t *sse_ptr, |
+ bool use_high_bit_depth_, |
+ vpx_bit_depth_t bit_depth) { |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ int64_t se = 0; |
+ uint64_t sse = 0; |
+ const int w = 1 << l2w; |
+ const int h = 1 << l2h; |
+ for (int y = 0; y < h; y++) { |
+ for (int x = 0; x < w; x++) { |
+ int diff; |
+ if (!use_high_bit_depth_) { |
+ diff = ref[w * y * ref_stride_coeff + x] - |
+ src[w * y * src_stride_coeff + x]; |
+ se += diff; |
+ sse += diff * diff; |
+ } else { |
+ diff = CONVERT_TO_SHORTPTR(ref)[w * y * ref_stride_coeff + x] - |
+ CONVERT_TO_SHORTPTR(src)[w * y * src_stride_coeff + x]; |
+ se += diff; |
+ sse += diff * diff; |
+ } |
+ } |
+ } |
+ if (bit_depth > VPX_BITS_8) { |
+ sse = ROUND_POWER_OF_TWO(sse, 2 * (bit_depth - 8)); |
+ se = ROUND_POWER_OF_TWO(se, bit_depth - 8); |
+ } |
+#else |
int se = 0; |
unsigned int sse = 0; |
- const int w = 1 << l2w, h = 1 << l2h; |
+ const int w = 1 << l2w; |
+ const int h = 1 << l2h; |
for (int y = 0; y < h; y++) { |
for (int x = 0; x < w; x++) { |
- int diff = ref[w * y + x] - src[w * y + x]; |
+ int diff = ref[w * y * ref_stride_coeff + x] - |
+ src[w * y * src_stride_coeff + x]; |
se += diff; |
sse += diff * diff; |
} |
} |
+#endif // CONFIG_VP9_HIGHBITDEPTH |
*sse_ptr = sse; |
return sse - (((int64_t) se * se) >> (l2w + l2h)); |
} |
@@ -61,13 +94,56 @@ |
static unsigned int subpel_variance_ref(const uint8_t *ref, const uint8_t *src, |
int l2w, int l2h, int xoff, int yoff, |
- unsigned int *sse_ptr) { |
+ unsigned int *sse_ptr, |
+ bool use_high_bit_depth_, |
+ vpx_bit_depth_t bit_depth) { |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ int64_t se = 0; |
+ uint64_t sse = 0; |
+ const int w = 1 << l2w; |
+ const int h = 1 << l2h; |
+ for (int y = 0; y < h; y++) { |
+ for (int x = 0; x < w; x++) { |
+ // Bilinear interpolation at a 16th pel step. |
+ if (!use_high_bit_depth_) { |
+ const int a1 = ref[(w + 1) * (y + 0) + x + 0]; |
+ const int a2 = ref[(w + 1) * (y + 0) + x + 1]; |
+ const int b1 = ref[(w + 1) * (y + 1) + x + 0]; |
+ const int b2 = ref[(w + 1) * (y + 1) + x + 1]; |
+ const int a = a1 + (((a2 - a1) * xoff + 8) >> 4); |
+ const int b = b1 + (((b2 - b1) * xoff + 8) >> 4); |
+ const int r = a + (((b - a) * yoff + 8) >> 4); |
+ const int diff = r - src[w * y + x]; |
+ se += diff; |
+ sse += diff * diff; |
+ } else { |
+ uint16_t *ref16 = CONVERT_TO_SHORTPTR(ref); |
+ uint16_t *src16 = CONVERT_TO_SHORTPTR(src); |
+ const int a1 = ref16[(w + 1) * (y + 0) + x + 0]; |
+ const int a2 = ref16[(w + 1) * (y + 0) + x + 1]; |
+ const int b1 = ref16[(w + 1) * (y + 1) + x + 0]; |
+ const int b2 = ref16[(w + 1) * (y + 1) + x + 1]; |
+ const int a = a1 + (((a2 - a1) * xoff + 8) >> 4); |
+ const int b = b1 + (((b2 - b1) * xoff + 8) >> 4); |
+ const int r = a + (((b - a) * yoff + 8) >> 4); |
+ const int diff = r - src16[w * y + x]; |
+ se += diff; |
+ sse += diff * diff; |
+ } |
+ } |
+ } |
+ if (bit_depth > VPX_BITS_8) { |
+ sse = ROUND_POWER_OF_TWO(sse, 2 * (bit_depth - 8)); |
+ se = ROUND_POWER_OF_TWO(se, bit_depth - 8); |
+ } |
+#else |
int se = 0; |
unsigned int sse = 0; |
- const int w = 1 << l2w, h = 1 << l2h; |
+ const int w = 1 << l2w; |
+ const int h = 1 << l2h; |
for (int y = 0; y < h; y++) { |
for (int x = 0; x < w; x++) { |
- // bilinear interpolation at a 16th pel step |
+ // Bilinear interpolation at a 16th pel step. |
const int a1 = ref[(w + 1) * (y + 0) + x + 0]; |
const int a2 = ref[(w + 1) * (y + 0) + x + 1]; |
const int b1 = ref[(w + 1) * (y + 1) + x + 0]; |
@@ -75,11 +151,12 @@ |
const int a = a1 + (((a2 - a1) * xoff + 8) >> 4); |
const int b = b1 + (((b2 - b1) * xoff + 8) >> 4); |
const int r = a + (((b - a) * yoff + 8) >> 4); |
- int diff = r - src[w * y + x]; |
+ const int diff = r - src[w * y + x]; |
se += diff; |
sse += diff * diff; |
} |
} |
+#endif // CONFIG_VP9_HIGHBITDEPTH |
*sse_ptr = sse; |
return sse - (((int64_t) se * se) >> (l2w + l2h)); |
} |
@@ -130,27 +207,57 @@ |
template<typename VarianceFunctionType> |
class VarianceTest |
- : public ::testing::TestWithParam<tuple<int, int, VarianceFunctionType> > { |
+ : public ::testing::TestWithParam<tuple<int, int, |
+ VarianceFunctionType, int> > { |
public: |
virtual void SetUp() { |
- const tuple<int, int, VarianceFunctionType>& params = this->GetParam(); |
+ const tuple<int, int, VarianceFunctionType, int>& params = this->GetParam(); |
log2width_ = get<0>(params); |
width_ = 1 << log2width_; |
log2height_ = get<1>(params); |
height_ = 1 << log2height_; |
variance_ = get<2>(params); |
+ if (get<3>(params)) { |
+ bit_depth_ = static_cast<vpx_bit_depth_t>(get<3>(params)); |
+ use_high_bit_depth_ = true; |
+ } else { |
+ bit_depth_ = VPX_BITS_8; |
+ use_high_bit_depth_ = false; |
+ } |
+ mask_ = (1 << bit_depth_) - 1; |
rnd_.Reset(ACMRandom::DeterministicSeed()); |
block_size_ = width_ * height_; |
- src_ = reinterpret_cast<uint8_t *>(vpx_memalign(16, block_size_)); |
- ref_ = new uint8_t[block_size_]; |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ if (!use_high_bit_depth_) { |
+ src_ = reinterpret_cast<uint8_t *>(vpx_memalign(16, block_size_ * 2)); |
+ ref_ = new uint8_t[block_size_ * 2]; |
+ } else { |
+ src_ = CONVERT_TO_BYTEPTR(reinterpret_cast<uint16_t *>( |
+ vpx_memalign(16, block_size_ * 2 * sizeof(uint16_t)))); |
+ ref_ = CONVERT_TO_BYTEPTR(new uint16_t[block_size_ * 2]); |
+ } |
+#else |
+ src_ = reinterpret_cast<uint8_t *>(vpx_memalign(16, block_size_ * 2)); |
+ ref_ = new uint8_t[block_size_ * 2]; |
+#endif |
ASSERT_TRUE(src_ != NULL); |
ASSERT_TRUE(ref_ != NULL); |
} |
virtual void TearDown() { |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ if (!use_high_bit_depth_) { |
+ vpx_free(src_); |
+ delete[] ref_; |
+ } else { |
+ vpx_free(CONVERT_TO_SHORTPTR(src_)); |
+ delete[] CONVERT_TO_SHORTPTR(ref_); |
+ } |
+#else |
vpx_free(src_); |
delete[] ref_; |
+#endif |
libvpx_test::ClearSystemState(); |
} |
@@ -157,13 +264,17 @@ |
protected: |
void ZeroTest(); |
void RefTest(); |
+ void RefStrideTest(); |
void OneQuarterTest(); |
ACMRandom rnd_; |
- uint8_t* src_; |
- uint8_t* ref_; |
+ uint8_t *src_; |
+ uint8_t *ref_; |
int width_, log2width_; |
int height_, log2height_; |
+ vpx_bit_depth_t bit_depth_; |
+ int mask_; |
+ bool use_high_bit_depth_; |
int block_size_; |
VarianceFunctionType variance_; |
}; |
@@ -171,14 +282,32 @@ |
template<typename VarianceFunctionType> |
void VarianceTest<VarianceFunctionType>::ZeroTest() { |
for (int i = 0; i <= 255; ++i) { |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ if (!use_high_bit_depth_) { |
+ memset(src_, i, block_size_); |
+ } else { |
+ vpx_memset16(CONVERT_TO_SHORTPTR(src_), i << (bit_depth_ - 8), |
+ block_size_); |
+ } |
+#else |
memset(src_, i, block_size_); |
+#endif |
for (int j = 0; j <= 255; ++j) { |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ if (!use_high_bit_depth_) { |
+ memset(ref_, j, block_size_); |
+ } else { |
+ vpx_memset16(CONVERT_TO_SHORTPTR(ref_), j << (bit_depth_ - 8), |
+ block_size_); |
+ } |
+#else |
memset(ref_, j, block_size_); |
+#endif |
unsigned int sse; |
unsigned int var; |
ASM_REGISTER_STATE_CHECK( |
var = variance_(src_, width_, ref_, width_, &sse)); |
- EXPECT_EQ(0u, var) << "src values: " << i << "ref values: " << j; |
+ EXPECT_EQ(0u, var) << "src values: " << i << " ref values: " << j; |
} |
} |
} |
@@ -187,15 +316,28 @@ |
void VarianceTest<VarianceFunctionType>::RefTest() { |
for (int i = 0; i < 10; ++i) { |
for (int j = 0; j < block_size_; j++) { |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ if (!use_high_bit_depth_) { |
src_[j] = rnd_.Rand8(); |
ref_[j] = rnd_.Rand8(); |
+ } else { |
+ CONVERT_TO_SHORTPTR(src_)[j] = rnd_.Rand16() && mask_; |
+ CONVERT_TO_SHORTPTR(ref_)[j] = rnd_.Rand16() && mask_; |
} |
+#else |
+ src_[j] = rnd_.Rand8(); |
+ ref_[j] = rnd_.Rand8(); |
+#endif |
+ } |
unsigned int sse1, sse2; |
unsigned int var1; |
+ const int stride_coeff = 1; |
ASM_REGISTER_STATE_CHECK( |
var1 = variance_(src_, width_, ref_, width_, &sse1)); |
const unsigned int var2 = variance_ref(src_, ref_, log2width_, |
- log2height_, &sse2); |
+ log2height_, stride_coeff, |
+ stride_coeff, &sse2, |
+ use_high_bit_depth_, bit_depth_); |
EXPECT_EQ(sse1, sse2); |
EXPECT_EQ(var1, var2); |
} |
@@ -202,11 +344,60 @@ |
} |
template<typename VarianceFunctionType> |
+void VarianceTest<VarianceFunctionType>::RefStrideTest() { |
+ for (int i = 0; i < 10; ++i) { |
+ int ref_stride_coeff = i % 2; |
+ int src_stride_coeff = (i >> 1) % 2; |
+ for (int j = 0; j < block_size_; j++) { |
+ int ref_ind = (j / width_) * ref_stride_coeff * width_ + j % width_; |
+ int src_ind = (j / width_) * src_stride_coeff * width_ + j % width_; |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ if (!use_high_bit_depth_) { |
+ src_[src_ind] = rnd_.Rand8(); |
+ ref_[ref_ind] = rnd_.Rand8(); |
+ } else { |
+ CONVERT_TO_SHORTPTR(src_)[src_ind] = rnd_.Rand16() && mask_; |
+ CONVERT_TO_SHORTPTR(ref_)[ref_ind] = rnd_.Rand16() && mask_; |
+ } |
+#else |
+ src_[src_ind] = rnd_.Rand8(); |
+ ref_[ref_ind] = rnd_.Rand8(); |
+#endif |
+ } |
+ unsigned int sse1, sse2; |
+ unsigned int var1; |
+ |
+ ASM_REGISTER_STATE_CHECK( |
+ var1 = variance_(src_, width_ * src_stride_coeff, |
+ ref_, width_ * ref_stride_coeff, &sse1)); |
+ const unsigned int var2 = variance_ref(src_, ref_, log2width_, |
+ log2height_, src_stride_coeff, |
+ ref_stride_coeff, &sse2, |
+ use_high_bit_depth_, bit_depth_); |
+ EXPECT_EQ(sse1, sse2); |
+ EXPECT_EQ(var1, var2); |
+ } |
+} |
+ |
+template<typename VarianceFunctionType> |
void VarianceTest<VarianceFunctionType>::OneQuarterTest() { |
+ const int half = block_size_ / 2; |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ if (!use_high_bit_depth_) { |
+ memset(src_, 255, block_size_); |
+ memset(ref_, 255, half); |
+ memset(ref_ + half, 0, half); |
+ } else { |
+ vpx_memset16(CONVERT_TO_SHORTPTR(src_), 255 << (bit_depth_ - 8), |
+ block_size_); |
+ vpx_memset16(CONVERT_TO_SHORTPTR(ref_), 255 << (bit_depth_ - 8), half); |
+ vpx_memset16(CONVERT_TO_SHORTPTR(ref_) + half, 0, half); |
+ } |
+#else |
memset(src_, 255, block_size_); |
- const int half = block_size_ / 2; |
memset(ref_, 255, half); |
memset(ref_ + half, 0, half); |
+#endif |
unsigned int sse; |
unsigned int var; |
ASM_REGISTER_STATE_CHECK(var = variance_(src_, width_, ref_, width_, &sse)); |
@@ -264,8 +455,10 @@ |
ref_[j] = rnd.Rand8(); |
} |
unsigned int sse1, sse2; |
+ const int stride_coeff = 1; |
ASM_REGISTER_STATE_CHECK(mse_(src_, width_, ref_, width_, &sse1)); |
- variance_ref(src_, ref_, log2width_, log2height_, &sse2); |
+ variance_ref(src_, ref_, log2width_, log2height_, stride_coeff, |
+ stride_coeff, &sse2, false, VPX_BITS_8); |
EXPECT_EQ(sse1, sse2); |
} |
} |
@@ -279,9 +472,10 @@ |
} |
unsigned int sse2; |
unsigned int var1; |
- ASM_REGISTER_STATE_CHECK( |
- var1 = mse_(src_, width_, ref_, width_)); |
- variance_ref(src_, ref_, log2width_, log2height_, &sse2); |
+ const int stride_coeff = 1; |
+ ASM_REGISTER_STATE_CHECK(var1 = mse_(src_, width_, ref_, width_)); |
+ variance_ref(src_, ref_, log2width_, log2height_, stride_coeff, |
+ stride_coeff, &sse2, false, VPX_BITS_8); |
EXPECT_EQ(var1, sse2); |
} |
} |
@@ -308,16 +502,59 @@ |
#endif |
#if CONFIG_VP9_ENCODER |
- |
unsigned int subpel_avg_variance_ref(const uint8_t *ref, |
const uint8_t *src, |
const uint8_t *second_pred, |
int l2w, int l2h, |
int xoff, int yoff, |
- unsigned int *sse_ptr) { |
+ unsigned int *sse_ptr, |
+ bool use_high_bit_depth, |
+ vpx_bit_depth_t bit_depth) { |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ int64_t se = 0; |
+ uint64_t sse = 0; |
+ const int w = 1 << l2w; |
+ const int h = 1 << l2h; |
+ for (int y = 0; y < h; y++) { |
+ for (int x = 0; x < w; x++) { |
+ // bilinear interpolation at a 16th pel step |
+ if (!use_high_bit_depth) { |
+ const int a1 = ref[(w + 1) * (y + 0) + x + 0]; |
+ const int a2 = ref[(w + 1) * (y + 0) + x + 1]; |
+ const int b1 = ref[(w + 1) * (y + 1) + x + 0]; |
+ const int b2 = ref[(w + 1) * (y + 1) + x + 1]; |
+ const int a = a1 + (((a2 - a1) * xoff + 8) >> 4); |
+ const int b = b1 + (((b2 - b1) * xoff + 8) >> 4); |
+ const int r = a + (((b - a) * yoff + 8) >> 4); |
+ const int diff = ((r + second_pred[w * y + x] + 1) >> 1) - src[w * y + x]; |
+ se += diff; |
+ sse += diff * diff; |
+ } else { |
+ uint16_t *ref16 = CONVERT_TO_SHORTPTR(ref); |
+ uint16_t *src16 = CONVERT_TO_SHORTPTR(src); |
+ uint16_t *sec16 = CONVERT_TO_SHORTPTR(second_pred); |
+ const int a1 = ref16[(w + 1) * (y + 0) + x + 0]; |
+ const int a2 = ref16[(w + 1) * (y + 0) + x + 1]; |
+ const int b1 = ref16[(w + 1) * (y + 1) + x + 0]; |
+ const int b2 = ref16[(w + 1) * (y + 1) + x + 1]; |
+ const int a = a1 + (((a2 - a1) * xoff + 8) >> 4); |
+ const int b = b1 + (((b2 - b1) * xoff + 8) >> 4); |
+ const int r = a + (((b - a) * yoff + 8) >> 4); |
+ const int diff = ((r + sec16[w * y + x] + 1) >> 1) - src16[w * y + x]; |
+ se += diff; |
+ sse += diff * diff; |
+ } |
+ } |
+ } |
+ if (bit_depth > 8) { |
+ sse = ROUND_POWER_OF_TWO(sse, 2*(bit_depth-8)); |
+ se = ROUND_POWER_OF_TWO(se, bit_depth-8); |
+ } |
+#else |
int se = 0; |
unsigned int sse = 0; |
- const int w = 1 << l2w, h = 1 << l2h; |
+ const int w = 1 << l2w; |
+ const int h = 1 << l2h; |
for (int y = 0; y < h; y++) { |
for (int x = 0; x < w; x++) { |
// bilinear interpolation at a 16th pel step |
@@ -328,11 +565,12 @@ |
const int a = a1 + (((a2 - a1) * xoff + 8) >> 4); |
const int b = b1 + (((b2 - b1) * xoff + 8) >> 4); |
const int r = a + (((b - a) * yoff + 8) >> 4); |
- int diff = ((r + second_pred[w * y + x] + 1) >> 1) - src[w * y + x]; |
+ const int diff = ((r + second_pred[w * y + x] + 1) >> 1) - src[w * y + x]; |
se += diff; |
sse += diff * diff; |
} |
} |
+#endif // CONFIG_VP9_HIGHBITDEPTH |
*sse_ptr = sse; |
return sse - (((int64_t) se * se) >> (l2w + l2h)); |
} |
@@ -340,10 +578,10 @@ |
template<typename SubpelVarianceFunctionType> |
class SubpelVarianceTest |
: public ::testing::TestWithParam<tuple<int, int, |
- SubpelVarianceFunctionType> > { |
+ SubpelVarianceFunctionType, int> > { |
public: |
virtual void SetUp() { |
- const tuple<int, int, SubpelVarianceFunctionType>& params = |
+ const tuple<int, int, SubpelVarianceFunctionType, int>& params = |
this->GetParam(); |
log2width_ = get<0>(params); |
width_ = 1 << log2width_; |
@@ -350,12 +588,37 @@ |
log2height_ = get<1>(params); |
height_ = 1 << log2height_; |
subpel_variance_ = get<2>(params); |
+ if (get<3>(params)) { |
+ bit_depth_ = (vpx_bit_depth_t) get<3>(params); |
+ use_high_bit_depth_ = true; |
+ } else { |
+ bit_depth_ = VPX_BITS_8; |
+ use_high_bit_depth_ = false; |
+ } |
+ mask_ = (1 << bit_depth_)-1; |
rnd_.Reset(ACMRandom::DeterministicSeed()); |
block_size_ = width_ * height_; |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ if (!use_high_bit_depth_) { |
+ src_ = reinterpret_cast<uint8_t *>(vpx_memalign(16, block_size_)); |
+ sec_ = reinterpret_cast<uint8_t *>(vpx_memalign(16, block_size_)); |
+ ref_ = new uint8_t[block_size_ + width_ + height_ + 1]; |
+ } else { |
+ src_ = CONVERT_TO_BYTEPTR( |
+ reinterpret_cast<uint16_t *>( |
+ vpx_memalign(16, block_size_*sizeof(uint16_t)))); |
+ sec_ = CONVERT_TO_BYTEPTR( |
+ reinterpret_cast<uint16_t *>( |
+ vpx_memalign(16, block_size_*sizeof(uint16_t)))); |
+ ref_ = CONVERT_TO_BYTEPTR( |
+ new uint16_t[block_size_ + width_ + height_ + 1]); |
+ } |
+#else |
src_ = reinterpret_cast<uint8_t *>(vpx_memalign(16, block_size_)); |
sec_ = reinterpret_cast<uint8_t *>(vpx_memalign(16, block_size_)); |
ref_ = new uint8_t[block_size_ + width_ + height_ + 1]; |
+#endif // CONFIG_VP9_HIGHBITDEPTH |
ASSERT_TRUE(src_ != NULL); |
ASSERT_TRUE(sec_ != NULL); |
ASSERT_TRUE(ref_ != NULL); |
@@ -362,22 +625,37 @@ |
} |
virtual void TearDown() { |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ if (!use_high_bit_depth_) { |
+ vpx_free(src_); |
+ delete[] ref_; |
+ vpx_free(sec_); |
+ } else { |
+ vpx_free(CONVERT_TO_SHORTPTR(src_)); |
+ delete[] CONVERT_TO_SHORTPTR(ref_); |
+ vpx_free(CONVERT_TO_SHORTPTR(sec_)); |
+ } |
+#else |
vpx_free(src_); |
delete[] ref_; |
vpx_free(sec_); |
+#endif |
libvpx_test::ClearSystemState(); |
} |
protected: |
void RefTest(); |
+ void ExtremeRefTest(); |
ACMRandom rnd_; |
uint8_t *src_; |
uint8_t *ref_; |
uint8_t *sec_; |
+ bool use_high_bit_depth_; |
+ vpx_bit_depth_t bit_depth_; |
int width_, log2width_; |
int height_, log2height_; |
- int block_size_; |
+ int block_size_, mask_; |
SubpelVarianceFunctionType subpel_variance_; |
}; |
@@ -385,6 +663,23 @@ |
void SubpelVarianceTest<SubpelVarianceFunctionType>::RefTest() { |
for (int x = 0; x < 16; ++x) { |
for (int y = 0; y < 16; ++y) { |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ if (!use_high_bit_depth_) { |
+ for (int j = 0; j < block_size_; j++) { |
+ src_[j] = rnd_.Rand8(); |
+ } |
+ for (int j = 0; j < block_size_ + width_ + height_ + 1; j++) { |
+ ref_[j] = rnd_.Rand8(); |
+ } |
+ } else { |
+ for (int j = 0; j < block_size_; j++) { |
+ CONVERT_TO_SHORTPTR(src_)[j] = rnd_.Rand16() & mask_; |
+ } |
+ for (int j = 0; j < block_size_ + width_ + height_ + 1; j++) { |
+ CONVERT_TO_SHORTPTR(ref_)[j] = rnd_.Rand16() & mask_; |
+ } |
+ } |
+#else |
for (int j = 0; j < block_size_; j++) { |
src_[j] = rnd_.Rand8(); |
} |
@@ -391,12 +686,15 @@ |
for (int j = 0; j < block_size_ + width_ + height_ + 1; j++) { |
ref_[j] = rnd_.Rand8(); |
} |
+#endif // CONFIG_VP9_HIGHBITDEPTH |
unsigned int sse1, sse2; |
unsigned int var1; |
ASM_REGISTER_STATE_CHECK(var1 = subpel_variance_(ref_, width_ + 1, x, y, |
src_, width_, &sse1)); |
const unsigned int var2 = subpel_variance_ref(ref_, src_, log2width_, |
- log2height_, x, y, &sse2); |
+ log2height_, x, y, &sse2, |
+ use_high_bit_depth_, |
+ bit_depth_); |
EXPECT_EQ(sse1, sse2) << "at position " << x << ", " << y; |
EXPECT_EQ(var1, var2) << "at position " << x << ", " << y; |
} |
@@ -403,10 +701,69 @@ |
} |
} |
+template<typename SubpelVarianceFunctionType> |
+void SubpelVarianceTest<SubpelVarianceFunctionType>::ExtremeRefTest() { |
+ // Compare against reference. |
+ // Src: Set the first half of values to 0, the second half to the maximum. |
+ // Ref: Set the first half of values to the maximum, the second half to 0. |
+ for (int x = 0; x < 16; ++x) { |
+ for (int y = 0; y < 16; ++y) { |
+ const int half = block_size_ / 2; |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ if (!use_high_bit_depth_) { |
+ memset(src_, 0, half); |
+ memset(src_ + half, 255, half); |
+ memset(ref_, 255, half); |
+ memset(ref_ + half, 0, half + width_ + height_ + 1); |
+ } else { |
+ vpx_memset16(CONVERT_TO_SHORTPTR(src_), mask_, half); |
+ vpx_memset16(CONVERT_TO_SHORTPTR(src_) + half, 0, half); |
+ vpx_memset16(CONVERT_TO_SHORTPTR(ref_), 0, half); |
+ vpx_memset16(CONVERT_TO_SHORTPTR(ref_) + half, mask_, |
+ half + width_ + height_ + 1); |
+ } |
+#else |
+ memset(src_, 0, half); |
+ memset(src_ + half, 255, half); |
+ memset(ref_, 255, half); |
+ memset(ref_ + half, 0, half + width_ + height_ + 1); |
+#endif // CONFIG_VP9_HIGHBITDEPTH |
+ unsigned int sse1, sse2; |
+ unsigned int var1; |
+ ASM_REGISTER_STATE_CHECK( |
+ var1 = subpel_variance_(ref_, width_ + 1, x, y, src_, width_, &sse1)); |
+ const unsigned int var2 = |
+ subpel_variance_ref(ref_, src_, log2width_, log2height_, x, y, &sse2, |
+ use_high_bit_depth_, bit_depth_); |
+ EXPECT_EQ(sse1, sse2) << "at position " << x << ", " << y; |
+ EXPECT_EQ(var1, var2) << "at position " << x << ", " << y; |
+ } |
+ } |
+} |
+ |
template<> |
void SubpelVarianceTest<vp9_subp_avg_variance_fn_t>::RefTest() { |
for (int x = 0; x < 16; ++x) { |
for (int y = 0; y < 16; ++y) { |
+#if CONFIG_VP9_HIGHBITDEPTH |
+ if (!use_high_bit_depth_) { |
+ for (int j = 0; j < block_size_; j++) { |
+ src_[j] = rnd_.Rand8(); |
+ sec_[j] = rnd_.Rand8(); |
+ } |
+ for (int j = 0; j < block_size_ + width_ + height_ + 1; j++) { |
+ ref_[j] = rnd_.Rand8(); |
+ } |
+ } else { |
+ for (int j = 0; j < block_size_; j++) { |
+ CONVERT_TO_SHORTPTR(src_)[j] = rnd_.Rand16() & mask_; |
+ CONVERT_TO_SHORTPTR(sec_)[j] = rnd_.Rand16() & mask_; |
+ } |
+ for (int j = 0; j < block_size_ + width_ + height_ + 1; j++) { |
+ CONVERT_TO_SHORTPTR(ref_)[j] = rnd_.Rand16() & mask_; |
+ } |
+ } |
+#else |
for (int j = 0; j < block_size_; j++) { |
src_[j] = rnd_.Rand8(); |
sec_[j] = rnd_.Rand8(); |
@@ -414,6 +771,7 @@ |
for (int j = 0; j < block_size_ + width_ + height_ + 1; j++) { |
ref_[j] = rnd_.Rand8(); |
} |
+#endif |
unsigned int sse1, sse2; |
unsigned int var1; |
ASM_REGISTER_STATE_CHECK( |
@@ -421,7 +779,9 @@ |
src_, width_, &sse1, sec_)); |
const unsigned int var2 = subpel_avg_variance_ref(ref_, src_, sec_, |
log2width_, log2height_, |
- x, y, &sse2); |
+ x, y, &sse2, |
+ use_high_bit_depth_, |
+ bit_depth_); |
EXPECT_EQ(sse1, sse2) << "at position " << x << ", " << y; |
EXPECT_EQ(var1, var2) << "at position " << x << ", " << y; |
} |
@@ -468,11 +828,11 @@ |
const vp8_variance_fn_t variance16x16_c = vp8_variance16x16_c; |
INSTANTIATE_TEST_CASE_P( |
C, VP8VarianceTest, |
- ::testing::Values(make_tuple(2, 2, variance4x4_c), |
- make_tuple(3, 3, variance8x8_c), |
- make_tuple(3, 4, variance8x16_c), |
- make_tuple(4, 3, variance16x8_c), |
- make_tuple(4, 4, variance16x16_c))); |
+ ::testing::Values(make_tuple(2, 2, variance4x4_c, 0), |
+ make_tuple(3, 3, variance8x8_c, 0), |
+ make_tuple(3, 4, variance8x16_c, 0), |
+ make_tuple(4, 3, variance16x8_c, 0), |
+ make_tuple(4, 4, variance16x16_c, 0))); |
#if HAVE_NEON |
const vp8_sse_fn_t get4x4sse_cs_neon = vp8_get4x4sse_cs_neon; |
@@ -491,13 +851,12 @@ |
const vp8_variance_fn_t variance16x16_neon = vp8_variance16x16_neon; |
INSTANTIATE_TEST_CASE_P( |
NEON, VP8VarianceTest, |
- ::testing::Values(make_tuple(3, 3, variance8x8_neon), |
- make_tuple(3, 4, variance8x16_neon), |
- make_tuple(4, 3, variance16x8_neon), |
- make_tuple(4, 4, variance16x16_neon))); |
+ ::testing::Values(make_tuple(3, 3, variance8x8_neon, 0), |
+ make_tuple(3, 4, variance8x16_neon, 0), |
+ make_tuple(4, 3, variance16x8_neon, 0), |
+ make_tuple(4, 4, variance16x16_neon, 0))); |
#endif |
- |
#if HAVE_MMX |
const vp8_variance_fn_t variance4x4_mmx = vp8_variance4x4_mmx; |
const vp8_variance_fn_t variance8x8_mmx = vp8_variance8x8_mmx; |
@@ -506,11 +865,11 @@ |
const vp8_variance_fn_t variance16x16_mmx = vp8_variance16x16_mmx; |
INSTANTIATE_TEST_CASE_P( |
MMX, VP8VarianceTest, |
- ::testing::Values(make_tuple(2, 2, variance4x4_mmx), |
- make_tuple(3, 3, variance8x8_mmx), |
- make_tuple(3, 4, variance8x16_mmx), |
- make_tuple(4, 3, variance16x8_mmx), |
- make_tuple(4, 4, variance16x16_mmx))); |
+ ::testing::Values(make_tuple(2, 2, variance4x4_mmx, 0), |
+ make_tuple(3, 3, variance8x8_mmx, 0), |
+ make_tuple(3, 4, variance8x16_mmx, 0), |
+ make_tuple(4, 3, variance16x8_mmx, 0), |
+ make_tuple(4, 4, variance16x16_mmx, 0))); |
#endif |
#if HAVE_SSE2 |
@@ -521,11 +880,11 @@ |
const vp8_variance_fn_t variance16x16_wmt = vp8_variance16x16_wmt; |
INSTANTIATE_TEST_CASE_P( |
SSE2, VP8VarianceTest, |
- ::testing::Values(make_tuple(2, 2, variance4x4_wmt), |
- make_tuple(3, 3, variance8x8_wmt), |
- make_tuple(3, 4, variance8x16_wmt), |
- make_tuple(4, 3, variance16x8_wmt), |
- make_tuple(4, 4, variance16x16_wmt))); |
+ ::testing::Values(make_tuple(2, 2, variance4x4_wmt, 0), |
+ make_tuple(3, 3, variance8x8_wmt, 0), |
+ make_tuple(3, 4, variance8x16_wmt, 0), |
+ make_tuple(4, 3, variance16x8_wmt, 0), |
+ make_tuple(4, 4, variance16x16_wmt, 0))); |
#endif |
#endif // CONFIG_VP8_ENCODER |
@@ -537,7 +896,6 @@ |
namespace vp9 { |
#if CONFIG_VP9_ENCODER |
- |
TEST_P(SumOfSquaresTest, Const) { ConstTest(); } |
TEST_P(SumOfSquaresTest, Ref) { RefTest(); } |
@@ -550,10 +908,27 @@ |
TEST_P(VP9VarianceTest, Zero) { ZeroTest(); } |
TEST_P(VP9VarianceTest, Ref) { RefTest(); } |
+TEST_P(VP9VarianceTest, RefStride) { RefStrideTest(); } |
TEST_P(VP9SubpelVarianceTest, Ref) { RefTest(); } |
+TEST_P(VP9SubpelVarianceTest, ExtremeRef) { ExtremeRefTest(); } |
TEST_P(VP9SubpelAvgVarianceTest, Ref) { RefTest(); } |
TEST_P(VP9VarianceTest, OneQuarter) { OneQuarterTest(); } |
+#if CONFIG_VP9_HIGHBITDEPTH |
+typedef VarianceTest<vp9_variance_fn_t> VP9VarianceHighTest; |
+typedef SubpelVarianceTest<vp9_subpixvariance_fn_t> VP9SubpelVarianceHighTest; |
+typedef SubpelVarianceTest<vp9_subp_avg_variance_fn_t> |
+ VP9SubpelAvgVarianceHighTest; |
+ |
+TEST_P(VP9VarianceHighTest, Zero) { ZeroTest(); } |
+TEST_P(VP9VarianceHighTest, Ref) { RefTest(); } |
+TEST_P(VP9VarianceHighTest, RefStride) { RefStrideTest(); } |
+TEST_P(VP9SubpelVarianceHighTest, Ref) { RefTest(); } |
+TEST_P(VP9SubpelVarianceHighTest, ExtremeRef) { ExtremeRefTest(); } |
+TEST_P(VP9SubpelAvgVarianceHighTest, Ref) { RefTest(); } |
+TEST_P(VP9VarianceHighTest, OneQuarter) { OneQuarterTest(); } |
+#endif // CONFIG_VP9_HIGHBITDEPTH |
+ |
const vp9_variance_fn_t variance4x4_c = vp9_variance4x4_c; |
const vp9_variance_fn_t variance4x8_c = vp9_variance4x8_c; |
const vp9_variance_fn_t variance8x4_c = vp9_variance8x4_c; |
@@ -569,20 +944,115 @@ |
const vp9_variance_fn_t variance64x64_c = vp9_variance64x64_c; |
INSTANTIATE_TEST_CASE_P( |
C, VP9VarianceTest, |
- ::testing::Values(make_tuple(2, 2, variance4x4_c), |
- make_tuple(2, 3, variance4x8_c), |
- make_tuple(3, 2, variance8x4_c), |
- make_tuple(3, 3, variance8x8_c), |
- make_tuple(3, 4, variance8x16_c), |
- make_tuple(4, 3, variance16x8_c), |
- make_tuple(4, 4, variance16x16_c), |
- make_tuple(4, 5, variance16x32_c), |
- make_tuple(5, 4, variance32x16_c), |
- make_tuple(5, 5, variance32x32_c), |
- make_tuple(5, 6, variance32x64_c), |
- make_tuple(6, 5, variance64x32_c), |
- make_tuple(6, 6, variance64x64_c))); |
- |
+ ::testing::Values(make_tuple(2, 2, variance4x4_c, 0), |
+ make_tuple(2, 3, variance4x8_c, 0), |
+ make_tuple(3, 2, variance8x4_c, 0), |
+ make_tuple(3, 3, variance8x8_c, 0), |
+ make_tuple(3, 4, variance8x16_c, 0), |
+ make_tuple(4, 3, variance16x8_c, 0), |
+ make_tuple(4, 4, variance16x16_c, 0), |
+ make_tuple(4, 5, variance16x32_c, 0), |
+ make_tuple(5, 4, variance32x16_c, 0), |
+ make_tuple(5, 5, variance32x32_c, 0), |
+ make_tuple(5, 6, variance32x64_c, 0), |
+ make_tuple(6, 5, variance64x32_c, 0), |
+ make_tuple(6, 6, variance64x64_c, 0))); |
+#if CONFIG_VP9_HIGHBITDEPTH |
+const vp9_variance_fn_t highbd_10_variance4x4_c = vp9_highbd_10_variance4x4_c; |
+const vp9_variance_fn_t highbd_10_variance4x8_c = vp9_highbd_10_variance4x8_c; |
+const vp9_variance_fn_t highbd_10_variance8x4_c = vp9_highbd_10_variance8x4_c; |
+const vp9_variance_fn_t highbd_10_variance8x8_c = vp9_highbd_10_variance8x8_c; |
+const vp9_variance_fn_t highbd_10_variance8x16_c = vp9_highbd_10_variance8x16_c; |
+const vp9_variance_fn_t highbd_10_variance16x8_c = vp9_highbd_10_variance16x8_c; |
+const vp9_variance_fn_t highbd_10_variance16x16_c = |
+ vp9_highbd_10_variance16x16_c; |
+const vp9_variance_fn_t highbd_10_variance16x32_c = |
+ vp9_highbd_10_variance16x32_c; |
+const vp9_variance_fn_t highbd_10_variance32x16_c = |
+ vp9_highbd_10_variance32x16_c; |
+const vp9_variance_fn_t highbd_10_variance32x32_c = |
+ vp9_highbd_10_variance32x32_c; |
+const vp9_variance_fn_t highbd_10_variance32x64_c = |
+ vp9_highbd_10_variance32x64_c; |
+const vp9_variance_fn_t highbd_10_variance64x32_c = |
+ vp9_highbd_10_variance64x32_c; |
+const vp9_variance_fn_t highbd_10_variance64x64_c = |
+ vp9_highbd_10_variance64x64_c; |
+const vp9_variance_fn_t highbd_12_variance4x4_c = vp9_highbd_12_variance4x4_c; |
+const vp9_variance_fn_t highbd_12_variance4x8_c = vp9_highbd_12_variance4x8_c; |
+const vp9_variance_fn_t highbd_12_variance8x4_c = vp9_highbd_12_variance8x4_c; |
+const vp9_variance_fn_t highbd_12_variance8x8_c = vp9_highbd_12_variance8x8_c; |
+const vp9_variance_fn_t highbd_12_variance8x16_c = vp9_highbd_12_variance8x16_c; |
+const vp9_variance_fn_t highbd_12_variance16x8_c = vp9_highbd_12_variance16x8_c; |
+const vp9_variance_fn_t highbd_12_variance16x16_c = |
+ vp9_highbd_12_variance16x16_c; |
+const vp9_variance_fn_t highbd_12_variance16x32_c = |
+ vp9_highbd_12_variance16x32_c; |
+const vp9_variance_fn_t highbd_12_variance32x16_c = |
+ vp9_highbd_12_variance32x16_c; |
+const vp9_variance_fn_t highbd_12_variance32x32_c = |
+ vp9_highbd_12_variance32x32_c; |
+const vp9_variance_fn_t highbd_12_variance32x64_c = |
+ vp9_highbd_12_variance32x64_c; |
+const vp9_variance_fn_t highbd_12_variance64x32_c = |
+ vp9_highbd_12_variance64x32_c; |
+const vp9_variance_fn_t highbd_12_variance64x64_c = |
+ vp9_highbd_12_variance64x64_c; |
+const vp9_variance_fn_t highbd_variance4x4_c = vp9_highbd_variance4x4_c; |
+const vp9_variance_fn_t highbd_variance4x8_c = vp9_highbd_variance4x8_c; |
+const vp9_variance_fn_t highbd_variance8x4_c = vp9_highbd_variance8x4_c; |
+const vp9_variance_fn_t highbd_variance8x8_c = vp9_highbd_variance8x8_c; |
+const vp9_variance_fn_t highbd_variance8x16_c = vp9_highbd_variance8x16_c; |
+const vp9_variance_fn_t highbd_variance16x8_c = vp9_highbd_variance16x8_c; |
+const vp9_variance_fn_t highbd_variance16x16_c = vp9_highbd_variance16x16_c; |
+const vp9_variance_fn_t highbd_variance16x32_c = vp9_highbd_variance16x32_c; |
+const vp9_variance_fn_t highbd_variance32x16_c = vp9_highbd_variance32x16_c; |
+const vp9_variance_fn_t highbd_variance32x32_c = vp9_highbd_variance32x32_c; |
+const vp9_variance_fn_t highbd_variance32x64_c = vp9_highbd_variance32x64_c; |
+const vp9_variance_fn_t highbd_variance64x32_c = vp9_highbd_variance64x32_c; |
+const vp9_variance_fn_t highbd_variance64x64_c = vp9_highbd_variance64x64_c; |
+INSTANTIATE_TEST_CASE_P( |
+ C, VP9VarianceHighTest, |
+ ::testing::Values(make_tuple(2, 2, highbd_10_variance4x4_c, 10), |
+ make_tuple(2, 3, highbd_10_variance4x8_c, 10), |
+ make_tuple(3, 2, highbd_10_variance8x4_c, 10), |
+ make_tuple(3, 3, highbd_10_variance8x8_c, 10), |
+ make_tuple(3, 4, highbd_10_variance8x16_c, 10), |
+ make_tuple(4, 3, highbd_10_variance16x8_c, 10), |
+ make_tuple(4, 4, highbd_10_variance16x16_c, 10), |
+ make_tuple(4, 5, highbd_10_variance16x32_c, 10), |
+ make_tuple(5, 4, highbd_10_variance32x16_c, 10), |
+ make_tuple(5, 5, highbd_10_variance32x32_c, 10), |
+ make_tuple(5, 6, highbd_10_variance32x64_c, 10), |
+ make_tuple(6, 5, highbd_10_variance64x32_c, 10), |
+ make_tuple(6, 6, highbd_10_variance64x64_c, 10), |
+ make_tuple(2, 2, highbd_12_variance4x4_c, 12), |
+ make_tuple(2, 3, highbd_12_variance4x8_c, 12), |
+ make_tuple(3, 2, highbd_12_variance8x4_c, 12), |
+ make_tuple(3, 3, highbd_12_variance8x8_c, 12), |
+ make_tuple(3, 4, highbd_12_variance8x16_c, 12), |
+ make_tuple(4, 3, highbd_12_variance16x8_c, 12), |
+ make_tuple(4, 4, highbd_12_variance16x16_c, 12), |
+ make_tuple(4, 5, highbd_12_variance16x32_c, 12), |
+ make_tuple(5, 4, highbd_12_variance32x16_c, 12), |
+ make_tuple(5, 5, highbd_12_variance32x32_c, 12), |
+ make_tuple(5, 6, highbd_12_variance32x64_c, 12), |
+ make_tuple(6, 5, highbd_12_variance64x32_c, 12), |
+ make_tuple(6, 6, highbd_12_variance64x64_c, 12), |
+ make_tuple(2, 2, highbd_variance4x4_c, 8), |
+ make_tuple(2, 3, highbd_variance4x8_c, 8), |
+ make_tuple(3, 2, highbd_variance8x4_c, 8), |
+ make_tuple(3, 3, highbd_variance8x8_c, 8), |
+ make_tuple(3, 4, highbd_variance8x16_c, 8), |
+ make_tuple(4, 3, highbd_variance16x8_c, 8), |
+ make_tuple(4, 4, highbd_variance16x16_c, 8), |
+ make_tuple(4, 5, highbd_variance16x32_c, 8), |
+ make_tuple(5, 4, highbd_variance32x16_c, 8), |
+ make_tuple(5, 5, highbd_variance32x32_c, 8), |
+ make_tuple(5, 6, highbd_variance32x64_c, 8), |
+ make_tuple(6, 5, highbd_variance64x32_c, 8), |
+ make_tuple(6, 6, highbd_variance64x64_c, 8))); |
+#endif // CONFIG_VP9_HIGHBITDEPTH |
const vp9_subpixvariance_fn_t subpel_variance4x4_c = |
vp9_sub_pixel_variance4x4_c; |
const vp9_subpixvariance_fn_t subpel_variance4x8_c = |
@@ -611,20 +1081,19 @@ |
vp9_sub_pixel_variance64x64_c; |
INSTANTIATE_TEST_CASE_P( |
C, VP9SubpelVarianceTest, |
- ::testing::Values(make_tuple(2, 2, subpel_variance4x4_c), |
- make_tuple(2, 3, subpel_variance4x8_c), |
- make_tuple(3, 2, subpel_variance8x4_c), |
- make_tuple(3, 3, subpel_variance8x8_c), |
- make_tuple(3, 4, subpel_variance8x16_c), |
- make_tuple(4, 3, subpel_variance16x8_c), |
- make_tuple(4, 4, subpel_variance16x16_c), |
- make_tuple(4, 5, subpel_variance16x32_c), |
- make_tuple(5, 4, subpel_variance32x16_c), |
- make_tuple(5, 5, subpel_variance32x32_c), |
- make_tuple(5, 6, subpel_variance32x64_c), |
- make_tuple(6, 5, subpel_variance64x32_c), |
- make_tuple(6, 6, subpel_variance64x64_c))); |
- |
+ ::testing::Values(make_tuple(2, 2, subpel_variance4x4_c, 0), |
+ make_tuple(2, 3, subpel_variance4x8_c, 0), |
+ make_tuple(3, 2, subpel_variance8x4_c, 0), |
+ make_tuple(3, 3, subpel_variance8x8_c, 0), |
+ make_tuple(3, 4, subpel_variance8x16_c, 0), |
+ make_tuple(4, 3, subpel_variance16x8_c, 0), |
+ make_tuple(4, 4, subpel_variance16x16_c, 0), |
+ make_tuple(4, 5, subpel_variance16x32_c, 0), |
+ make_tuple(5, 4, subpel_variance32x16_c, 0), |
+ make_tuple(5, 5, subpel_variance32x32_c, 0), |
+ make_tuple(5, 6, subpel_variance32x64_c, 0), |
+ make_tuple(6, 5, subpel_variance64x32_c, 0), |
+ make_tuple(6, 6, subpel_variance64x64_c, 0))); |
const vp9_subp_avg_variance_fn_t subpel_avg_variance4x4_c = |
vp9_sub_pixel_avg_variance4x4_c; |
const vp9_subp_avg_variance_fn_t subpel_avg_variance4x8_c = |
@@ -653,23 +1122,263 @@ |
vp9_sub_pixel_avg_variance64x64_c; |
INSTANTIATE_TEST_CASE_P( |
C, VP9SubpelAvgVarianceTest, |
- ::testing::Values(make_tuple(2, 2, subpel_avg_variance4x4_c), |
- make_tuple(2, 3, subpel_avg_variance4x8_c), |
- make_tuple(3, 2, subpel_avg_variance8x4_c), |
- make_tuple(3, 3, subpel_avg_variance8x8_c), |
- make_tuple(3, 4, subpel_avg_variance8x16_c), |
- make_tuple(4, 3, subpel_avg_variance16x8_c), |
- make_tuple(4, 4, subpel_avg_variance16x16_c), |
- make_tuple(4, 5, subpel_avg_variance16x32_c), |
- make_tuple(5, 4, subpel_avg_variance32x16_c), |
- make_tuple(5, 5, subpel_avg_variance32x32_c), |
- make_tuple(5, 6, subpel_avg_variance32x64_c), |
- make_tuple(6, 5, subpel_avg_variance64x32_c), |
- make_tuple(6, 6, subpel_avg_variance64x64_c))); |
+ ::testing::Values(make_tuple(2, 2, subpel_avg_variance4x4_c, 0), |
+ make_tuple(2, 3, subpel_avg_variance4x8_c, 0), |
+ make_tuple(3, 2, subpel_avg_variance8x4_c, 0), |
+ make_tuple(3, 3, subpel_avg_variance8x8_c, 0), |
+ make_tuple(3, 4, subpel_avg_variance8x16_c, 0), |
+ make_tuple(4, 3, subpel_avg_variance16x8_c, 0), |
+ make_tuple(4, 4, subpel_avg_variance16x16_c, 0), |
+ make_tuple(4, 5, subpel_avg_variance16x32_c, 0), |
+ make_tuple(5, 4, subpel_avg_variance32x16_c, 0), |
+ make_tuple(5, 5, subpel_avg_variance32x32_c, 0), |
+ make_tuple(5, 6, subpel_avg_variance32x64_c, 0), |
+ make_tuple(6, 5, subpel_avg_variance64x32_c, 0), |
+ make_tuple(6, 6, subpel_avg_variance64x64_c, 0))); |
+#if CONFIG_VP9_HIGHBITDEPTH |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance4x4_c = |
+ vp9_highbd_10_sub_pixel_variance4x4_c; |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance4x8_c = |
+ vp9_highbd_10_sub_pixel_variance4x8_c; |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance8x4_c = |
+ vp9_highbd_10_sub_pixel_variance8x4_c; |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance8x8_c = |
+ vp9_highbd_10_sub_pixel_variance8x8_c; |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance8x16_c = |
+ vp9_highbd_10_sub_pixel_variance8x16_c; |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance16x8_c = |
+ vp9_highbd_10_sub_pixel_variance16x8_c; |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance16x16_c = |
+ vp9_highbd_10_sub_pixel_variance16x16_c; |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance16x32_c = |
+ vp9_highbd_10_sub_pixel_variance16x32_c; |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance32x16_c = |
+ vp9_highbd_10_sub_pixel_variance32x16_c; |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance32x32_c = |
+ vp9_highbd_10_sub_pixel_variance32x32_c; |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance32x64_c = |
+ vp9_highbd_10_sub_pixel_variance32x64_c; |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance64x32_c = |
+ vp9_highbd_10_sub_pixel_variance64x32_c; |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance64x64_c = |
+ vp9_highbd_10_sub_pixel_variance64x64_c; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance4x4_c = |
+ vp9_highbd_12_sub_pixel_variance4x4_c; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance4x8_c = |
+ vp9_highbd_12_sub_pixel_variance4x8_c; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance8x4_c = |
+ vp9_highbd_12_sub_pixel_variance8x4_c; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance8x8_c = |
+ vp9_highbd_12_sub_pixel_variance8x8_c; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance8x16_c = |
+ vp9_highbd_12_sub_pixel_variance8x16_c; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance16x8_c = |
+ vp9_highbd_12_sub_pixel_variance16x8_c; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance16x16_c = |
+ vp9_highbd_12_sub_pixel_variance16x16_c; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance16x32_c = |
+ vp9_highbd_12_sub_pixel_variance16x32_c; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance32x16_c = |
+ vp9_highbd_12_sub_pixel_variance32x16_c; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance32x32_c = |
+ vp9_highbd_12_sub_pixel_variance32x32_c; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance32x64_c = |
+ vp9_highbd_12_sub_pixel_variance32x64_c; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance64x32_c = |
+ vp9_highbd_12_sub_pixel_variance64x32_c; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance64x64_c = |
+ vp9_highbd_12_sub_pixel_variance64x64_c; |
+const vp9_subpixvariance_fn_t highbd_subpel_variance4x4_c = |
+ vp9_highbd_sub_pixel_variance4x4_c; |
+const vp9_subpixvariance_fn_t highbd_subpel_variance4x8_c = |
+ vp9_highbd_sub_pixel_variance4x8_c; |
+const vp9_subpixvariance_fn_t highbd_subpel_variance8x4_c = |
+ vp9_highbd_sub_pixel_variance8x4_c; |
+const vp9_subpixvariance_fn_t highbd_subpel_variance8x8_c = |
+ vp9_highbd_sub_pixel_variance8x8_c; |
+const vp9_subpixvariance_fn_t highbd_subpel_variance8x16_c = |
+ vp9_highbd_sub_pixel_variance8x16_c; |
+const vp9_subpixvariance_fn_t highbd_subpel_variance16x8_c = |
+ vp9_highbd_sub_pixel_variance16x8_c; |
+const vp9_subpixvariance_fn_t highbd_subpel_variance16x16_c = |
+ vp9_highbd_sub_pixel_variance16x16_c; |
+const vp9_subpixvariance_fn_t highbd_subpel_variance16x32_c = |
+ vp9_highbd_sub_pixel_variance16x32_c; |
+const vp9_subpixvariance_fn_t highbd_subpel_variance32x16_c = |
+ vp9_highbd_sub_pixel_variance32x16_c; |
+const vp9_subpixvariance_fn_t highbd_subpel_variance32x32_c = |
+ vp9_highbd_sub_pixel_variance32x32_c; |
+const vp9_subpixvariance_fn_t highbd_subpel_variance32x64_c = |
+ vp9_highbd_sub_pixel_variance32x64_c; |
+const vp9_subpixvariance_fn_t highbd_subpel_variance64x32_c = |
+ vp9_highbd_sub_pixel_variance64x32_c; |
+const vp9_subpixvariance_fn_t highbd_subpel_variance64x64_c = |
+ vp9_highbd_sub_pixel_variance64x64_c; |
+INSTANTIATE_TEST_CASE_P( |
+ C, VP9SubpelVarianceHighTest, |
+ ::testing::Values(make_tuple(2, 2, highbd_10_subpel_variance4x4_c, 10), |
+ make_tuple(2, 3, highbd_10_subpel_variance4x8_c, 10), |
+ make_tuple(3, 2, highbd_10_subpel_variance8x4_c, 10), |
+ make_tuple(3, 3, highbd_10_subpel_variance8x8_c, 10), |
+ make_tuple(3, 4, highbd_10_subpel_variance8x16_c, 10), |
+ make_tuple(4, 3, highbd_10_subpel_variance16x8_c, 10), |
+ make_tuple(4, 4, highbd_10_subpel_variance16x16_c, 10), |
+ make_tuple(4, 5, highbd_10_subpel_variance16x32_c, 10), |
+ make_tuple(5, 4, highbd_10_subpel_variance32x16_c, 10), |
+ make_tuple(5, 5, highbd_10_subpel_variance32x32_c, 10), |
+ make_tuple(5, 6, highbd_10_subpel_variance32x64_c, 10), |
+ make_tuple(6, 5, highbd_10_subpel_variance64x32_c, 10), |
+ make_tuple(6, 6, highbd_10_subpel_variance64x64_c, 10), |
+ make_tuple(2, 2, highbd_12_subpel_variance4x4_c, 12), |
+ make_tuple(2, 3, highbd_12_subpel_variance4x8_c, 12), |
+ make_tuple(3, 2, highbd_12_subpel_variance8x4_c, 12), |
+ make_tuple(3, 3, highbd_12_subpel_variance8x8_c, 12), |
+ make_tuple(3, 4, highbd_12_subpel_variance8x16_c, 12), |
+ make_tuple(4, 3, highbd_12_subpel_variance16x8_c, 12), |
+ make_tuple(4, 4, highbd_12_subpel_variance16x16_c, 12), |
+ make_tuple(4, 5, highbd_12_subpel_variance16x32_c, 12), |
+ make_tuple(5, 4, highbd_12_subpel_variance32x16_c, 12), |
+ make_tuple(5, 5, highbd_12_subpel_variance32x32_c, 12), |
+ make_tuple(5, 6, highbd_12_subpel_variance32x64_c, 12), |
+ make_tuple(6, 5, highbd_12_subpel_variance64x32_c, 12), |
+ make_tuple(6, 6, highbd_12_subpel_variance64x64_c, 12), |
+ make_tuple(2, 2, highbd_subpel_variance4x4_c, 8), |
+ make_tuple(2, 3, highbd_subpel_variance4x8_c, 8), |
+ make_tuple(3, 2, highbd_subpel_variance8x4_c, 8), |
+ make_tuple(3, 3, highbd_subpel_variance8x8_c, 8), |
+ make_tuple(3, 4, highbd_subpel_variance8x16_c, 8), |
+ make_tuple(4, 3, highbd_subpel_variance16x8_c, 8), |
+ make_tuple(4, 4, highbd_subpel_variance16x16_c, 8), |
+ make_tuple(4, 5, highbd_subpel_variance16x32_c, 8), |
+ make_tuple(5, 4, highbd_subpel_variance32x16_c, 8), |
+ make_tuple(5, 5, highbd_subpel_variance32x32_c, 8), |
+ make_tuple(5, 6, highbd_subpel_variance32x64_c, 8), |
+ make_tuple(6, 5, highbd_subpel_variance64x32_c, 8), |
+ make_tuple(6, 6, highbd_subpel_variance64x64_c, 8))); |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance4x4_c = |
+ vp9_highbd_10_sub_pixel_avg_variance4x4_c; |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance4x8_c = |
+ vp9_highbd_10_sub_pixel_avg_variance4x8_c; |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance8x4_c = |
+ vp9_highbd_10_sub_pixel_avg_variance8x4_c; |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance8x8_c = |
+ vp9_highbd_10_sub_pixel_avg_variance8x8_c; |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance8x16_c = |
+ vp9_highbd_10_sub_pixel_avg_variance8x16_c; |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance16x8_c = |
+ vp9_highbd_10_sub_pixel_avg_variance16x8_c; |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance16x16_c = |
+ vp9_highbd_10_sub_pixel_avg_variance16x16_c; |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance16x32_c = |
+ vp9_highbd_10_sub_pixel_avg_variance16x32_c; |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance32x16_c = |
+ vp9_highbd_10_sub_pixel_avg_variance32x16_c; |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance32x32_c = |
+ vp9_highbd_10_sub_pixel_avg_variance32x32_c; |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance32x64_c = |
+ vp9_highbd_10_sub_pixel_avg_variance32x64_c; |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance64x32_c = |
+ vp9_highbd_10_sub_pixel_avg_variance64x32_c; |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance64x64_c = |
+ vp9_highbd_10_sub_pixel_avg_variance64x64_c; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance4x4_c = |
+ vp9_highbd_12_sub_pixel_avg_variance4x4_c; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance4x8_c = |
+ vp9_highbd_12_sub_pixel_avg_variance4x8_c; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance8x4_c = |
+ vp9_highbd_12_sub_pixel_avg_variance8x4_c; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance8x8_c = |
+ vp9_highbd_12_sub_pixel_avg_variance8x8_c; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance8x16_c = |
+ vp9_highbd_12_sub_pixel_avg_variance8x16_c; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance16x8_c = |
+ vp9_highbd_12_sub_pixel_avg_variance16x8_c; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance16x16_c = |
+ vp9_highbd_12_sub_pixel_avg_variance16x16_c; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance16x32_c = |
+ vp9_highbd_12_sub_pixel_avg_variance16x32_c; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance32x16_c = |
+ vp9_highbd_12_sub_pixel_avg_variance32x16_c; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance32x32_c = |
+ vp9_highbd_12_sub_pixel_avg_variance32x32_c; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance32x64_c = |
+ vp9_highbd_12_sub_pixel_avg_variance32x64_c; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance64x32_c = |
+ vp9_highbd_12_sub_pixel_avg_variance64x32_c; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance64x64_c = |
+ vp9_highbd_12_sub_pixel_avg_variance64x64_c; |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance4x4_c = |
+ vp9_highbd_sub_pixel_avg_variance4x4_c; |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance4x8_c = |
+ vp9_highbd_sub_pixel_avg_variance4x8_c; |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance8x4_c = |
+ vp9_highbd_sub_pixel_avg_variance8x4_c; |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance8x8_c = |
+ vp9_highbd_sub_pixel_avg_variance8x8_c; |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance8x16_c = |
+ vp9_highbd_sub_pixel_avg_variance8x16_c; |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance16x8_c = |
+ vp9_highbd_sub_pixel_avg_variance16x8_c; |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance16x16_c = |
+ vp9_highbd_sub_pixel_avg_variance16x16_c; |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance16x32_c = |
+ vp9_highbd_sub_pixel_avg_variance16x32_c; |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance32x16_c = |
+ vp9_highbd_sub_pixel_avg_variance32x16_c; |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance32x32_c = |
+ vp9_highbd_sub_pixel_avg_variance32x32_c; |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance32x64_c = |
+ vp9_highbd_sub_pixel_avg_variance32x64_c; |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance64x32_c = |
+ vp9_highbd_sub_pixel_avg_variance64x32_c; |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance64x64_c = |
+ vp9_highbd_sub_pixel_avg_variance64x64_c; |
+INSTANTIATE_TEST_CASE_P( |
+ C, VP9SubpelAvgVarianceHighTest, |
+ ::testing::Values( |
+ make_tuple(2, 2, highbd_10_subpel_avg_variance4x4_c, 10), |
+ make_tuple(2, 3, highbd_10_subpel_avg_variance4x8_c, 10), |
+ make_tuple(3, 2, highbd_10_subpel_avg_variance8x4_c, 10), |
+ make_tuple(3, 3, highbd_10_subpel_avg_variance8x8_c, 10), |
+ make_tuple(3, 4, highbd_10_subpel_avg_variance8x16_c, 10), |
+ make_tuple(4, 3, highbd_10_subpel_avg_variance16x8_c, 10), |
+ make_tuple(4, 4, highbd_10_subpel_avg_variance16x16_c, 10), |
+ make_tuple(4, 5, highbd_10_subpel_avg_variance16x32_c, 10), |
+ make_tuple(5, 4, highbd_10_subpel_avg_variance32x16_c, 10), |
+ make_tuple(5, 5, highbd_10_subpel_avg_variance32x32_c, 10), |
+ make_tuple(5, 6, highbd_10_subpel_avg_variance32x64_c, 10), |
+ make_tuple(6, 5, highbd_10_subpel_avg_variance64x32_c, 10), |
+ make_tuple(6, 6, highbd_10_subpel_avg_variance64x64_c, 10), |
+ make_tuple(2, 2, highbd_12_subpel_avg_variance4x4_c, 12), |
+ make_tuple(2, 3, highbd_12_subpel_avg_variance4x8_c, 12), |
+ make_tuple(3, 2, highbd_12_subpel_avg_variance8x4_c, 12), |
+ make_tuple(3, 3, highbd_12_subpel_avg_variance8x8_c, 12), |
+ make_tuple(3, 4, highbd_12_subpel_avg_variance8x16_c, 12), |
+ make_tuple(4, 3, highbd_12_subpel_avg_variance16x8_c, 12), |
+ make_tuple(4, 4, highbd_12_subpel_avg_variance16x16_c, 12), |
+ make_tuple(4, 5, highbd_12_subpel_avg_variance16x32_c, 12), |
+ make_tuple(5, 4, highbd_12_subpel_avg_variance32x16_c, 12), |
+ make_tuple(5, 5, highbd_12_subpel_avg_variance32x32_c, 12), |
+ make_tuple(5, 6, highbd_12_subpel_avg_variance32x64_c, 12), |
+ make_tuple(6, 5, highbd_12_subpel_avg_variance64x32_c, 12), |
+ make_tuple(6, 6, highbd_12_subpel_avg_variance64x64_c, 12), |
+ make_tuple(2, 2, highbd_subpel_avg_variance4x4_c, 8), |
+ make_tuple(2, 3, highbd_subpel_avg_variance4x8_c, 8), |
+ make_tuple(3, 2, highbd_subpel_avg_variance8x4_c, 8), |
+ make_tuple(3, 3, highbd_subpel_avg_variance8x8_c, 8), |
+ make_tuple(3, 4, highbd_subpel_avg_variance8x16_c, 8), |
+ make_tuple(4, 3, highbd_subpel_avg_variance16x8_c, 8), |
+ make_tuple(4, 4, highbd_subpel_avg_variance16x16_c, 8), |
+ make_tuple(4, 5, highbd_subpel_avg_variance16x32_c, 8), |
+ make_tuple(5, 4, highbd_subpel_avg_variance32x16_c, 8), |
+ make_tuple(5, 5, highbd_subpel_avg_variance32x32_c, 8), |
+ make_tuple(5, 6, highbd_subpel_avg_variance32x64_c, 8), |
+ make_tuple(6, 5, highbd_subpel_avg_variance64x32_c, 8), |
+ make_tuple(6, 6, highbd_subpel_avg_variance64x64_c, 8))); |
+#endif // CONFIG_VP9_HIGHBITDEPTH |
#if HAVE_SSE2 |
#if CONFIG_USE_X86INC |
- |
INSTANTIATE_TEST_CASE_P(SSE2, SumOfSquaresTest, |
::testing::Values(vp9_get_mb_ss_sse2)); |
@@ -688,20 +1397,19 @@ |
const vp9_variance_fn_t variance64x64_sse2 = vp9_variance64x64_sse2; |
INSTANTIATE_TEST_CASE_P( |
SSE2, VP9VarianceTest, |
- ::testing::Values(make_tuple(2, 2, variance4x4_sse2), |
- make_tuple(2, 3, variance4x8_sse2), |
- make_tuple(3, 2, variance8x4_sse2), |
- make_tuple(3, 3, variance8x8_sse2), |
- make_tuple(3, 4, variance8x16_sse2), |
- make_tuple(4, 3, variance16x8_sse2), |
- make_tuple(4, 4, variance16x16_sse2), |
- make_tuple(4, 5, variance16x32_sse2), |
- make_tuple(5, 4, variance32x16_sse2), |
- make_tuple(5, 5, variance32x32_sse2), |
- make_tuple(5, 6, variance32x64_sse2), |
- make_tuple(6, 5, variance64x32_sse2), |
- make_tuple(6, 6, variance64x64_sse2))); |
- |
+ ::testing::Values(make_tuple(2, 2, variance4x4_sse2, 0), |
+ make_tuple(2, 3, variance4x8_sse2, 0), |
+ make_tuple(3, 2, variance8x4_sse2, 0), |
+ make_tuple(3, 3, variance8x8_sse2, 0), |
+ make_tuple(3, 4, variance8x16_sse2, 0), |
+ make_tuple(4, 3, variance16x8_sse2, 0), |
+ make_tuple(4, 4, variance16x16_sse2, 0), |
+ make_tuple(4, 5, variance16x32_sse2, 0), |
+ make_tuple(5, 4, variance32x16_sse2, 0), |
+ make_tuple(5, 5, variance32x32_sse2, 0), |
+ make_tuple(5, 6, variance32x64_sse2, 0), |
+ make_tuple(6, 5, variance64x32_sse2, 0), |
+ make_tuple(6, 6, variance64x64_sse2, 0))); |
const vp9_subpixvariance_fn_t subpel_variance4x4_sse = |
vp9_sub_pixel_variance4x4_sse; |
const vp9_subpixvariance_fn_t subpel_variance4x8_sse = |
@@ -730,20 +1438,19 @@ |
vp9_sub_pixel_variance64x64_sse2; |
INSTANTIATE_TEST_CASE_P( |
SSE2, VP9SubpelVarianceTest, |
- ::testing::Values(make_tuple(2, 2, subpel_variance4x4_sse), |
- make_tuple(2, 3, subpel_variance4x8_sse), |
- make_tuple(3, 2, subpel_variance8x4_sse2), |
- make_tuple(3, 3, subpel_variance8x8_sse2), |
- make_tuple(3, 4, subpel_variance8x16_sse2), |
- make_tuple(4, 3, subpel_variance16x8_sse2), |
- make_tuple(4, 4, subpel_variance16x16_sse2), |
- make_tuple(4, 5, subpel_variance16x32_sse2), |
- make_tuple(5, 4, subpel_variance32x16_sse2), |
- make_tuple(5, 5, subpel_variance32x32_sse2), |
- make_tuple(5, 6, subpel_variance32x64_sse2), |
- make_tuple(6, 5, subpel_variance64x32_sse2), |
- make_tuple(6, 6, subpel_variance64x64_sse2))); |
- |
+ ::testing::Values(make_tuple(2, 2, subpel_variance4x4_sse, 0), |
+ make_tuple(2, 3, subpel_variance4x8_sse, 0), |
+ make_tuple(3, 2, subpel_variance8x4_sse2, 0), |
+ make_tuple(3, 3, subpel_variance8x8_sse2, 0), |
+ make_tuple(3, 4, subpel_variance8x16_sse2, 0), |
+ make_tuple(4, 3, subpel_variance16x8_sse2, 0), |
+ make_tuple(4, 4, subpel_variance16x16_sse2, 0), |
+ make_tuple(4, 5, subpel_variance16x32_sse2, 0), |
+ make_tuple(5, 4, subpel_variance32x16_sse2, 0), |
+ make_tuple(5, 5, subpel_variance32x32_sse2, 0), |
+ make_tuple(5, 6, subpel_variance32x64_sse2, 0), |
+ make_tuple(6, 5, subpel_variance64x32_sse2, 0), |
+ make_tuple(6, 6, subpel_variance64x64_sse2, 0))); |
const vp9_subp_avg_variance_fn_t subpel_avg_variance4x4_sse = |
vp9_sub_pixel_avg_variance4x4_sse; |
const vp9_subp_avg_variance_fn_t subpel_avg_variance4x8_sse = |
@@ -772,22 +1479,316 @@ |
vp9_sub_pixel_avg_variance64x64_sse2; |
INSTANTIATE_TEST_CASE_P( |
SSE2, VP9SubpelAvgVarianceTest, |
- ::testing::Values(make_tuple(2, 2, subpel_avg_variance4x4_sse), |
- make_tuple(2, 3, subpel_avg_variance4x8_sse), |
- make_tuple(3, 2, subpel_avg_variance8x4_sse2), |
- make_tuple(3, 3, subpel_avg_variance8x8_sse2), |
- make_tuple(3, 4, subpel_avg_variance8x16_sse2), |
- make_tuple(4, 3, subpel_avg_variance16x8_sse2), |
- make_tuple(4, 4, subpel_avg_variance16x16_sse2), |
- make_tuple(4, 5, subpel_avg_variance16x32_sse2), |
- make_tuple(5, 4, subpel_avg_variance32x16_sse2), |
- make_tuple(5, 5, subpel_avg_variance32x32_sse2), |
- make_tuple(5, 6, subpel_avg_variance32x64_sse2), |
- make_tuple(6, 5, subpel_avg_variance64x32_sse2), |
- make_tuple(6, 6, subpel_avg_variance64x64_sse2))); |
-#endif |
-#endif |
- |
+ ::testing::Values(make_tuple(2, 2, subpel_avg_variance4x4_sse, 0), |
+ make_tuple(2, 3, subpel_avg_variance4x8_sse, 0), |
+ make_tuple(3, 2, subpel_avg_variance8x4_sse2, 0), |
+ make_tuple(3, 3, subpel_avg_variance8x8_sse2, 0), |
+ make_tuple(3, 4, subpel_avg_variance8x16_sse2, 0), |
+ make_tuple(4, 3, subpel_avg_variance16x8_sse2, 0), |
+ make_tuple(4, 4, subpel_avg_variance16x16_sse2, 0), |
+ make_tuple(4, 5, subpel_avg_variance16x32_sse2, 0), |
+ make_tuple(5, 4, subpel_avg_variance32x16_sse2, 0), |
+ make_tuple(5, 5, subpel_avg_variance32x32_sse2, 0), |
+ make_tuple(5, 6, subpel_avg_variance32x64_sse2, 0), |
+ make_tuple(6, 5, subpel_avg_variance64x32_sse2, 0), |
+ make_tuple(6, 6, subpel_avg_variance64x64_sse2, 0))); |
+#if CONFIG_VP9_HIGHBITDEPTH |
+const vp9_variance_fn_t highbd_variance8x8_sse2 = vp9_highbd_variance8x8_sse2; |
+const vp9_variance_fn_t highbd_10_variance8x8_sse2 = |
+ vp9_highbd_10_variance8x8_sse2; |
+const vp9_variance_fn_t highbd_12_variance8x8_sse2 = |
+ vp9_highbd_12_variance8x8_sse2; |
+const vp9_variance_fn_t highbd_variance8x16_sse2 = vp9_highbd_variance8x16_sse2; |
+const vp9_variance_fn_t highbd_10_variance8x16_sse2 = |
+ vp9_highbd_10_variance8x16_sse2; |
+const vp9_variance_fn_t highbd_12_variance8x16_sse2 = |
+ vp9_highbd_12_variance8x16_sse2; |
+const vp9_variance_fn_t highbd_variance16x8_sse2 = |
+ vp9_highbd_variance16x8_sse2; |
+const vp9_variance_fn_t highbd_10_variance16x8_sse2 = |
+ vp9_highbd_10_variance16x8_sse2; |
+const vp9_variance_fn_t highbd_12_variance16x8_sse2 = |
+ vp9_highbd_12_variance16x8_sse2; |
+const vp9_variance_fn_t highbd_variance16x16_sse2 = |
+ vp9_highbd_variance16x16_sse2; |
+const vp9_variance_fn_t highbd_10_variance16x16_sse2 = |
+ vp9_highbd_10_variance16x16_sse2; |
+const vp9_variance_fn_t highbd_12_variance16x16_sse2 = |
+ vp9_highbd_12_variance16x16_sse2; |
+const vp9_variance_fn_t highbd_variance16x32_sse2 = |
+ vp9_highbd_variance16x32_sse2; |
+const vp9_variance_fn_t highbd_10_variance16x32_sse2 = |
+ vp9_highbd_10_variance16x32_sse2; |
+const vp9_variance_fn_t highbd_12_variance16x32_sse2 = |
+ vp9_highbd_12_variance16x32_sse2; |
+const vp9_variance_fn_t highbd_variance32x16_sse2 = |
+ vp9_highbd_variance32x16_sse2; |
+const vp9_variance_fn_t highbd_10_variance32x16_sse2 = |
+ vp9_highbd_10_variance32x16_sse2; |
+const vp9_variance_fn_t highbd_12_variance32x16_sse2 = |
+ vp9_highbd_12_variance32x16_sse2; |
+const vp9_variance_fn_t highbd_variance32x32_sse2 = |
+ vp9_highbd_variance32x32_sse2; |
+const vp9_variance_fn_t highbd_10_variance32x32_sse2 = |
+ vp9_highbd_10_variance32x32_sse2; |
+const vp9_variance_fn_t highbd_12_variance32x32_sse2 = |
+ vp9_highbd_12_variance32x32_sse2; |
+const vp9_variance_fn_t highbd_variance32x64_sse2 = |
+ vp9_highbd_variance32x64_sse2; |
+const vp9_variance_fn_t highbd_10_variance32x64_sse2 = |
+ vp9_highbd_10_variance32x64_sse2; |
+const vp9_variance_fn_t highbd_12_variance32x64_sse2 = |
+ vp9_highbd_12_variance32x64_sse2; |
+const vp9_variance_fn_t highbd_variance64x32_sse2 = |
+ vp9_highbd_variance64x32_sse2; |
+const vp9_variance_fn_t highbd_10_variance64x32_sse2 = |
+ vp9_highbd_10_variance64x32_sse2; |
+const vp9_variance_fn_t highbd_12_variance64x32_sse2 = |
+ vp9_highbd_12_variance64x32_sse2; |
+const vp9_variance_fn_t highbd_variance64x64_sse2 = |
+ vp9_highbd_variance64x64_sse2; |
+const vp9_variance_fn_t highbd_10_variance64x64_sse2 = |
+ vp9_highbd_10_variance64x64_sse2; |
+const vp9_variance_fn_t highbd_12_variance64x64_sse2 = |
+ vp9_highbd_12_variance64x64_sse2; |
+INSTANTIATE_TEST_CASE_P( |
+ SSE2, VP9VarianceHighTest, |
+ ::testing::Values(make_tuple(3, 3, highbd_10_variance8x8_sse2, 10), |
+ make_tuple(3, 4, highbd_10_variance8x16_sse2, 10), |
+ make_tuple(4, 3, highbd_10_variance16x8_sse2, 10), |
+ make_tuple(4, 4, highbd_10_variance16x16_sse2, 10), |
+ make_tuple(4, 5, highbd_10_variance16x32_sse2, 10), |
+ make_tuple(5, 4, highbd_10_variance32x16_sse2, 10), |
+ make_tuple(5, 5, highbd_10_variance32x32_sse2, 10), |
+ make_tuple(5, 6, highbd_10_variance32x64_sse2, 10), |
+ make_tuple(6, 5, highbd_10_variance64x32_sse2, 10), |
+ make_tuple(6, 6, highbd_10_variance64x64_sse2, 10), |
+ make_tuple(3, 3, highbd_12_variance8x8_sse2, 12), |
+ make_tuple(3, 4, highbd_12_variance8x16_sse2, 12), |
+ make_tuple(4, 3, highbd_12_variance16x8_sse2, 12), |
+ make_tuple(4, 4, highbd_12_variance16x16_sse2, 12), |
+ make_tuple(4, 5, highbd_12_variance16x32_sse2, 12), |
+ make_tuple(5, 4, highbd_12_variance32x16_sse2, 12), |
+ make_tuple(5, 5, highbd_12_variance32x32_sse2, 12), |
+ make_tuple(5, 6, highbd_12_variance32x64_sse2, 12), |
+ make_tuple(6, 5, highbd_12_variance64x32_sse2, 12), |
+ make_tuple(6, 6, highbd_12_variance64x64_sse2, 12), |
+ make_tuple(3, 3, highbd_variance8x8_sse2, 8), |
+ make_tuple(3, 4, highbd_variance8x16_sse2, 8), |
+ make_tuple(4, 3, highbd_variance16x8_sse2, 8), |
+ make_tuple(4, 4, highbd_variance16x16_sse2, 8), |
+ make_tuple(4, 5, highbd_variance16x32_sse2, 8), |
+ make_tuple(5, 4, highbd_variance32x16_sse2, 8), |
+ make_tuple(5, 5, highbd_variance32x32_sse2, 8), |
+ make_tuple(5, 6, highbd_variance32x64_sse2, 8), |
+ make_tuple(6, 5, highbd_variance64x32_sse2, 8), |
+ make_tuple(6, 6, highbd_variance64x64_sse2, 8))); |
+const vp9_subpixvariance_fn_t highbd_subpel_variance8x4_sse2 = |
+ vp9_highbd_sub_pixel_variance8x4_sse2; |
+const vp9_subpixvariance_fn_t highbd_subpel_variance8x8_sse2 = |
+ vp9_highbd_sub_pixel_variance8x8_sse2; |
+const vp9_subpixvariance_fn_t highbd_subpel_variance8x16_sse2 = |
+ vp9_highbd_sub_pixel_variance8x16_sse2; |
+const vp9_subpixvariance_fn_t highbd_subpel_variance16x8_sse2 = |
+ vp9_highbd_sub_pixel_variance16x8_sse2; |
+const vp9_subpixvariance_fn_t highbd_subpel_variance16x16_sse2 = |
+ vp9_highbd_sub_pixel_variance16x16_sse2; |
+const vp9_subpixvariance_fn_t highbd_subpel_variance16x32_sse2 = |
+ vp9_highbd_sub_pixel_variance16x32_sse2; |
+const vp9_subpixvariance_fn_t highbd_subpel_variance32x16_sse2 = |
+ vp9_highbd_sub_pixel_variance32x16_sse2; |
+const vp9_subpixvariance_fn_t highbd_subpel_variance32x32_sse2 = |
+ vp9_highbd_sub_pixel_variance32x32_sse2; |
+const vp9_subpixvariance_fn_t highbd_subpel_variance32x64_sse2 = |
+ vp9_highbd_sub_pixel_variance32x64_sse2; |
+const vp9_subpixvariance_fn_t highbd_subpel_variance64x32_sse2 = |
+ vp9_highbd_sub_pixel_variance64x32_sse2; |
+const vp9_subpixvariance_fn_t highbd_subpel_variance64x64_sse2 = |
+ vp9_highbd_sub_pixel_variance64x64_sse2; |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance8x4_sse2 = |
+ vp9_highbd_10_sub_pixel_variance8x4_sse2; |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance8x8_sse2 = |
+ vp9_highbd_10_sub_pixel_variance8x8_sse2; |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance8x16_sse2 = |
+ vp9_highbd_10_sub_pixel_variance8x16_sse2; |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance16x8_sse2 = |
+ vp9_highbd_10_sub_pixel_variance16x8_sse2; |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance16x16_sse2 = |
+ vp9_highbd_10_sub_pixel_variance16x16_sse2; |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance16x32_sse2 = |
+ vp9_highbd_10_sub_pixel_variance16x32_sse2; |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance32x16_sse2 = |
+ vp9_highbd_10_sub_pixel_variance32x16_sse2; |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance32x32_sse2 = |
+ vp9_highbd_10_sub_pixel_variance32x32_sse2; |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance32x64_sse2 = |
+ vp9_highbd_10_sub_pixel_variance32x64_sse2; |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance64x32_sse2 = |
+ vp9_highbd_10_sub_pixel_variance64x32_sse2; |
+const vp9_subpixvariance_fn_t highbd_10_subpel_variance64x64_sse2 = |
+ vp9_highbd_10_sub_pixel_variance64x64_sse2; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance8x4_sse2 = |
+ vp9_highbd_12_sub_pixel_variance8x4_sse2; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance8x8_sse2 = |
+ vp9_highbd_12_sub_pixel_variance8x8_sse2; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance8x16_sse2 = |
+ vp9_highbd_12_sub_pixel_variance8x16_sse2; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance16x8_sse2 = |
+ vp9_highbd_12_sub_pixel_variance16x8_sse2; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance16x16_sse2 = |
+ vp9_highbd_12_sub_pixel_variance16x16_sse2; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance16x32_sse2 = |
+ vp9_highbd_12_sub_pixel_variance16x32_sse2; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance32x16_sse2 = |
+ vp9_highbd_12_sub_pixel_variance32x16_sse2; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance32x32_sse2 = |
+ vp9_highbd_12_sub_pixel_variance32x32_sse2; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance32x64_sse2 = |
+ vp9_highbd_12_sub_pixel_variance32x64_sse2; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance64x32_sse2 = |
+ vp9_highbd_12_sub_pixel_variance64x32_sse2; |
+const vp9_subpixvariance_fn_t highbd_12_subpel_variance64x64_sse2 = |
+ vp9_highbd_12_sub_pixel_variance64x64_sse2; |
+INSTANTIATE_TEST_CASE_P( |
+ SSE2, VP9SubpelVarianceHighTest, |
+ ::testing::Values(make_tuple(3, 2, highbd_10_subpel_variance8x4_sse2, 10), |
+ make_tuple(3, 3, highbd_10_subpel_variance8x8_sse2, 10), |
+ make_tuple(3, 4, highbd_10_subpel_variance8x16_sse2, 10), |
+ make_tuple(4, 3, highbd_10_subpel_variance16x8_sse2, 10), |
+ make_tuple(4, 4, highbd_10_subpel_variance16x16_sse2, 10), |
+ make_tuple(4, 5, highbd_10_subpel_variance16x32_sse2, 10), |
+ make_tuple(5, 4, highbd_10_subpel_variance32x16_sse2, 10), |
+ make_tuple(5, 5, highbd_10_subpel_variance32x32_sse2, 10), |
+ make_tuple(5, 6, highbd_10_subpel_variance32x64_sse2, 10), |
+ make_tuple(6, 5, highbd_10_subpel_variance64x32_sse2, 10), |
+ make_tuple(6, 6, highbd_10_subpel_variance64x64_sse2, 10), |
+ make_tuple(3, 2, highbd_12_subpel_variance8x4_sse2, 12), |
+ make_tuple(3, 3, highbd_12_subpel_variance8x8_sse2, 12), |
+ make_tuple(3, 4, highbd_12_subpel_variance8x16_sse2, 12), |
+ make_tuple(4, 3, highbd_12_subpel_variance16x8_sse2, 12), |
+ make_tuple(4, 4, highbd_12_subpel_variance16x16_sse2, 12), |
+ make_tuple(4, 5, highbd_12_subpel_variance16x32_sse2, 12), |
+ make_tuple(5, 4, highbd_12_subpel_variance32x16_sse2, 12), |
+ make_tuple(5, 5, highbd_12_subpel_variance32x32_sse2, 12), |
+ make_tuple(5, 6, highbd_12_subpel_variance32x64_sse2, 12), |
+ make_tuple(6, 5, highbd_12_subpel_variance64x32_sse2, 12), |
+ make_tuple(6, 6, highbd_12_subpel_variance64x64_sse2, 12), |
+ make_tuple(3, 2, highbd_subpel_variance8x4_sse2, 8), |
+ make_tuple(3, 3, highbd_subpel_variance8x8_sse2, 8), |
+ make_tuple(3, 4, highbd_subpel_variance8x16_sse2, 8), |
+ make_tuple(4, 3, highbd_subpel_variance16x8_sse2, 8), |
+ make_tuple(4, 4, highbd_subpel_variance16x16_sse2, 8), |
+ make_tuple(4, 5, highbd_subpel_variance16x32_sse2, 8), |
+ make_tuple(5, 4, highbd_subpel_variance32x16_sse2, 8), |
+ make_tuple(5, 5, highbd_subpel_variance32x32_sse2, 8), |
+ make_tuple(5, 6, highbd_subpel_variance32x64_sse2, 8), |
+ make_tuple(6, 5, highbd_subpel_variance64x32_sse2, 8), |
+ make_tuple(6, 6, highbd_subpel_variance64x64_sse2, 8))); |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance8x4_sse2 = |
+ vp9_highbd_sub_pixel_avg_variance8x4_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance8x8_sse2 = |
+ vp9_highbd_sub_pixel_avg_variance8x8_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance8x16_sse2 = |
+ vp9_highbd_sub_pixel_avg_variance8x16_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance16x8_sse2 = |
+ vp9_highbd_sub_pixel_avg_variance16x8_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance16x16_sse2 = |
+ vp9_highbd_sub_pixel_avg_variance16x16_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance16x32_sse2 = |
+ vp9_highbd_sub_pixel_avg_variance16x32_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance32x16_sse2 = |
+ vp9_highbd_sub_pixel_avg_variance32x16_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance32x32_sse2 = |
+ vp9_highbd_sub_pixel_avg_variance32x32_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance32x64_sse2 = |
+ vp9_highbd_sub_pixel_avg_variance32x64_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance64x32_sse2 = |
+ vp9_highbd_sub_pixel_avg_variance64x32_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_subpel_avg_variance64x64_sse2 = |
+ vp9_highbd_sub_pixel_avg_variance64x64_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance8x4_sse2 = |
+ vp9_highbd_10_sub_pixel_avg_variance8x4_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance8x8_sse2 = |
+ vp9_highbd_10_sub_pixel_avg_variance8x8_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance8x16_sse2 = |
+ vp9_highbd_10_sub_pixel_avg_variance8x16_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance16x8_sse2 = |
+ vp9_highbd_10_sub_pixel_avg_variance16x8_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance16x16_sse2 = |
+ vp9_highbd_10_sub_pixel_avg_variance16x16_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance16x32_sse2 = |
+ vp9_highbd_10_sub_pixel_avg_variance16x32_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance32x16_sse2 = |
+ vp9_highbd_10_sub_pixel_avg_variance32x16_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance32x32_sse2 = |
+ vp9_highbd_10_sub_pixel_avg_variance32x32_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance32x64_sse2 = |
+ vp9_highbd_10_sub_pixel_avg_variance32x64_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance64x32_sse2 = |
+ vp9_highbd_10_sub_pixel_avg_variance64x32_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_10_subpel_avg_variance64x64_sse2 = |
+ vp9_highbd_10_sub_pixel_avg_variance64x64_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance8x4_sse2 = |
+ vp9_highbd_12_sub_pixel_avg_variance8x4_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance8x8_sse2 = |
+ vp9_highbd_12_sub_pixel_avg_variance8x8_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance8x16_sse2 = |
+ vp9_highbd_12_sub_pixel_avg_variance8x16_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance16x8_sse2 = |
+ vp9_highbd_12_sub_pixel_avg_variance16x8_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance16x16_sse2 = |
+ vp9_highbd_12_sub_pixel_avg_variance16x16_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance16x32_sse2 = |
+ vp9_highbd_12_sub_pixel_avg_variance16x32_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance32x16_sse2 = |
+ vp9_highbd_12_sub_pixel_avg_variance32x16_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance32x32_sse2 = |
+ vp9_highbd_12_sub_pixel_avg_variance32x32_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance32x64_sse2 = |
+ vp9_highbd_12_sub_pixel_avg_variance32x64_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance64x32_sse2 = |
+ vp9_highbd_12_sub_pixel_avg_variance64x32_sse2; |
+const vp9_subp_avg_variance_fn_t highbd_12_subpel_avg_variance64x64_sse2 = |
+ vp9_highbd_12_sub_pixel_avg_variance64x64_sse2; |
+INSTANTIATE_TEST_CASE_P( |
+ SSE2, VP9SubpelAvgVarianceHighTest, |
+ ::testing::Values( |
+ make_tuple(3, 2, highbd_10_subpel_avg_variance8x4_sse2, 10), |
+ make_tuple(3, 3, highbd_10_subpel_avg_variance8x8_sse2, 10), |
+ make_tuple(3, 4, highbd_10_subpel_avg_variance8x16_sse2, 10), |
+ make_tuple(4, 3, highbd_10_subpel_avg_variance16x8_sse2, 10), |
+ make_tuple(4, 4, highbd_10_subpel_avg_variance16x16_sse2, 10), |
+ make_tuple(4, 5, highbd_10_subpel_avg_variance16x32_sse2, 10), |
+ make_tuple(5, 4, highbd_10_subpel_avg_variance32x16_sse2, 10), |
+ make_tuple(5, 5, highbd_10_subpel_avg_variance32x32_sse2, 10), |
+ make_tuple(5, 6, highbd_10_subpel_avg_variance32x64_sse2, 10), |
+ make_tuple(6, 5, highbd_10_subpel_avg_variance64x32_sse2, 10), |
+ make_tuple(6, 6, highbd_10_subpel_avg_variance64x64_sse2, 10), |
+ make_tuple(3, 2, highbd_12_subpel_avg_variance8x4_sse2, 12), |
+ make_tuple(3, 3, highbd_12_subpel_avg_variance8x8_sse2, 12), |
+ make_tuple(3, 4, highbd_12_subpel_avg_variance8x16_sse2, 12), |
+ make_tuple(4, 3, highbd_12_subpel_avg_variance16x8_sse2, 12), |
+ make_tuple(4, 4, highbd_12_subpel_avg_variance16x16_sse2, 12), |
+ make_tuple(4, 5, highbd_12_subpel_avg_variance16x32_sse2, 12), |
+ make_tuple(5, 4, highbd_12_subpel_avg_variance32x16_sse2, 12), |
+ make_tuple(5, 5, highbd_12_subpel_avg_variance32x32_sse2, 12), |
+ make_tuple(5, 6, highbd_12_subpel_avg_variance32x64_sse2, 12), |
+ make_tuple(6, 5, highbd_12_subpel_avg_variance64x32_sse2, 12), |
+ make_tuple(6, 6, highbd_12_subpel_avg_variance64x64_sse2, 12), |
+ make_tuple(3, 2, highbd_subpel_avg_variance8x4_sse2, 8), |
+ make_tuple(3, 3, highbd_subpel_avg_variance8x8_sse2, 8), |
+ make_tuple(3, 4, highbd_subpel_avg_variance8x16_sse2, 8), |
+ make_tuple(4, 3, highbd_subpel_avg_variance16x8_sse2, 8), |
+ make_tuple(4, 4, highbd_subpel_avg_variance16x16_sse2, 8), |
+ make_tuple(4, 5, highbd_subpel_avg_variance16x32_sse2, 8), |
+ make_tuple(5, 4, highbd_subpel_avg_variance32x16_sse2, 8), |
+ make_tuple(5, 5, highbd_subpel_avg_variance32x32_sse2, 8), |
+ make_tuple(5, 6, highbd_subpel_avg_variance32x64_sse2, 8), |
+ make_tuple(6, 5, highbd_subpel_avg_variance64x32_sse2, 8), |
+ make_tuple(6, 6, highbd_subpel_avg_variance64x64_sse2, 8))); |
+#endif // CONFIG_VP9_HIGHBITDEPTH |
+#endif // CONFIG_USE_X86INC |
+#endif // HAVE_SSE2 |
#if HAVE_SSSE3 |
#if CONFIG_USE_X86INC |
@@ -819,20 +1820,19 @@ |
vp9_sub_pixel_variance64x64_ssse3; |
INSTANTIATE_TEST_CASE_P( |
SSSE3, VP9SubpelVarianceTest, |
- ::testing::Values(make_tuple(2, 2, subpel_variance4x4_ssse3), |
- make_tuple(2, 3, subpel_variance4x8_ssse3), |
- make_tuple(3, 2, subpel_variance8x4_ssse3), |
- make_tuple(3, 3, subpel_variance8x8_ssse3), |
- make_tuple(3, 4, subpel_variance8x16_ssse3), |
- make_tuple(4, 3, subpel_variance16x8_ssse3), |
- make_tuple(4, 4, subpel_variance16x16_ssse3), |
- make_tuple(4, 5, subpel_variance16x32_ssse3), |
- make_tuple(5, 4, subpel_variance32x16_ssse3), |
- make_tuple(5, 5, subpel_variance32x32_ssse3), |
- make_tuple(5, 6, subpel_variance32x64_ssse3), |
- make_tuple(6, 5, subpel_variance64x32_ssse3), |
- make_tuple(6, 6, subpel_variance64x64_ssse3))); |
- |
+ ::testing::Values(make_tuple(2, 2, subpel_variance4x4_ssse3, 0), |
+ make_tuple(2, 3, subpel_variance4x8_ssse3, 0), |
+ make_tuple(3, 2, subpel_variance8x4_ssse3, 0), |
+ make_tuple(3, 3, subpel_variance8x8_ssse3, 0), |
+ make_tuple(3, 4, subpel_variance8x16_ssse3, 0), |
+ make_tuple(4, 3, subpel_variance16x8_ssse3, 0), |
+ make_tuple(4, 4, subpel_variance16x16_ssse3, 0), |
+ make_tuple(4, 5, subpel_variance16x32_ssse3, 0), |
+ make_tuple(5, 4, subpel_variance32x16_ssse3, 0), |
+ make_tuple(5, 5, subpel_variance32x32_ssse3, 0), |
+ make_tuple(5, 6, subpel_variance32x64_ssse3, 0), |
+ make_tuple(6, 5, subpel_variance64x32_ssse3, 0), |
+ make_tuple(6, 6, subpel_variance64x64_ssse3, 0))); |
const vp9_subp_avg_variance_fn_t subpel_avg_variance4x4_ssse3 = |
vp9_sub_pixel_avg_variance4x4_ssse3; |
const vp9_subp_avg_variance_fn_t subpel_avg_variance4x8_ssse3 = |
@@ -861,21 +1861,21 @@ |
vp9_sub_pixel_avg_variance64x64_ssse3; |
INSTANTIATE_TEST_CASE_P( |
SSSE3, VP9SubpelAvgVarianceTest, |
- ::testing::Values(make_tuple(2, 2, subpel_avg_variance4x4_ssse3), |
- make_tuple(2, 3, subpel_avg_variance4x8_ssse3), |
- make_tuple(3, 2, subpel_avg_variance8x4_ssse3), |
- make_tuple(3, 3, subpel_avg_variance8x8_ssse3), |
- make_tuple(3, 4, subpel_avg_variance8x16_ssse3), |
- make_tuple(4, 3, subpel_avg_variance16x8_ssse3), |
- make_tuple(4, 4, subpel_avg_variance16x16_ssse3), |
- make_tuple(4, 5, subpel_avg_variance16x32_ssse3), |
- make_tuple(5, 4, subpel_avg_variance32x16_ssse3), |
- make_tuple(5, 5, subpel_avg_variance32x32_ssse3), |
- make_tuple(5, 6, subpel_avg_variance32x64_ssse3), |
- make_tuple(6, 5, subpel_avg_variance64x32_ssse3), |
- make_tuple(6, 6, subpel_avg_variance64x64_ssse3))); |
-#endif |
-#endif |
+ ::testing::Values(make_tuple(2, 2, subpel_avg_variance4x4_ssse3, 0), |
+ make_tuple(2, 3, subpel_avg_variance4x8_ssse3, 0), |
+ make_tuple(3, 2, subpel_avg_variance8x4_ssse3, 0), |
+ make_tuple(3, 3, subpel_avg_variance8x8_ssse3, 0), |
+ make_tuple(3, 4, subpel_avg_variance8x16_ssse3, 0), |
+ make_tuple(4, 3, subpel_avg_variance16x8_ssse3, 0), |
+ make_tuple(4, 4, subpel_avg_variance16x16_ssse3, 0), |
+ make_tuple(4, 5, subpel_avg_variance16x32_ssse3, 0), |
+ make_tuple(5, 4, subpel_avg_variance32x16_ssse3, 0), |
+ make_tuple(5, 5, subpel_avg_variance32x32_ssse3, 0), |
+ make_tuple(5, 6, subpel_avg_variance32x64_ssse3, 0), |
+ make_tuple(6, 5, subpel_avg_variance64x32_ssse3, 0), |
+ make_tuple(6, 6, subpel_avg_variance64x64_ssse3, 0))); |
+#endif // CONFIG_USE_X86INC |
+#endif // HAVE_SSSE3 |
#if HAVE_AVX2 |
@@ -886,11 +1886,11 @@ |
const vp9_variance_fn_t variance64x64_avx2 = vp9_variance64x64_avx2; |
INSTANTIATE_TEST_CASE_P( |
AVX2, VP9VarianceTest, |
- ::testing::Values(make_tuple(4, 4, variance16x16_avx2), |
- make_tuple(5, 4, variance32x16_avx2), |
- make_tuple(5, 5, variance32x32_avx2), |
- make_tuple(6, 5, variance64x32_avx2), |
- make_tuple(6, 6, variance64x64_avx2))); |
+ ::testing::Values(make_tuple(4, 4, variance16x16_avx2, 0), |
+ make_tuple(5, 4, variance32x16_avx2, 0), |
+ make_tuple(5, 5, variance32x32_avx2, 0), |
+ make_tuple(6, 5, variance64x32_avx2, 0), |
+ make_tuple(6, 6, variance64x64_avx2, 0))); |
const vp9_subpixvariance_fn_t subpel_variance32x32_avx2 = |
vp9_sub_pixel_variance32x32_avx2; |
@@ -898,8 +1898,8 @@ |
vp9_sub_pixel_variance64x64_avx2; |
INSTANTIATE_TEST_CASE_P( |
AVX2, VP9SubpelVarianceTest, |
- ::testing::Values(make_tuple(5, 5, subpel_variance32x32_avx2), |
- make_tuple(6, 6, subpel_variance64x64_avx2))); |
+ ::testing::Values(make_tuple(5, 5, subpel_variance32x32_avx2, 0), |
+ make_tuple(6, 6, subpel_variance64x64_avx2, 0))); |
const vp9_subp_avg_variance_fn_t subpel_avg_variance32x32_avx2 = |
vp9_sub_pixel_avg_variance32x32_avx2; |
@@ -907,8 +1907,8 @@ |
vp9_sub_pixel_avg_variance64x64_avx2; |
INSTANTIATE_TEST_CASE_P( |
AVX2, VP9SubpelAvgVarianceTest, |
- ::testing::Values(make_tuple(5, 5, subpel_avg_variance32x32_avx2), |
- make_tuple(6, 6, subpel_avg_variance64x64_avx2))); |
+ ::testing::Values(make_tuple(5, 5, subpel_avg_variance32x32_avx2, 0), |
+ make_tuple(6, 6, subpel_avg_variance64x64_avx2, 0))); |
#endif // HAVE_AVX2 |
#if HAVE_NEON |
const vp9_variance_fn_t variance8x8_neon = vp9_variance8x8_neon; |
@@ -916,9 +1916,9 @@ |
const vp9_variance_fn_t variance32x32_neon = vp9_variance32x32_neon; |
INSTANTIATE_TEST_CASE_P( |
NEON, VP9VarianceTest, |
- ::testing::Values(make_tuple(3, 3, variance8x8_neon), |
- make_tuple(4, 4, variance16x16_neon), |
- make_tuple(5, 5, variance32x32_neon))); |
+ ::testing::Values(make_tuple(3, 3, variance8x8_neon, 0), |
+ make_tuple(4, 4, variance16x16_neon, 0), |
+ make_tuple(5, 5, variance32x32_neon, 0))); |
const vp9_subpixvariance_fn_t subpel_variance8x8_neon = |
vp9_sub_pixel_variance8x8_neon; |
@@ -928,12 +1928,11 @@ |
vp9_sub_pixel_variance32x32_neon; |
INSTANTIATE_TEST_CASE_P( |
NEON, VP9SubpelVarianceTest, |
- ::testing::Values(make_tuple(3, 3, subpel_variance8x8_neon), |
- make_tuple(4, 4, subpel_variance16x16_neon), |
- make_tuple(5, 5, subpel_variance32x32_neon))); |
+ ::testing::Values(make_tuple(3, 3, subpel_variance8x8_neon, 0), |
+ make_tuple(4, 4, subpel_variance16x16_neon, 0), |
+ make_tuple(5, 5, subpel_variance32x32_neon, 0))); |
#endif // HAVE_NEON |
#endif // CONFIG_VP9_ENCODER |
} // namespace vp9 |
- |
} // namespace |