| Index: source/libvpx/vp9/encoder/x86/vp9_sad_intrin_avx2.c
|
| diff --git a/source/libvpx/vp9/encoder/x86/vp9_sad_intrin_avx2.c b/source/libvpx/vp9/encoder/x86/vp9_sad_intrin_avx2.c
|
| deleted file mode 100644
|
| index 113193070e143a3d4c06d31f0b20cb41629f795c..0000000000000000000000000000000000000000
|
| --- a/source/libvpx/vp9/encoder/x86/vp9_sad_intrin_avx2.c
|
| +++ /dev/null
|
| @@ -1,180 +0,0 @@
|
| -/*
|
| - * Copyright (c) 2012 The WebM project authors. All Rights Reserved.
|
| - *
|
| - * Use of this source code is governed by a BSD-style license
|
| - * that can be found in the LICENSE file in the root of the source
|
| - * tree. An additional intellectual property rights grant can be found
|
| - * in the file PATENTS. All contributing project authors may
|
| - * be found in the AUTHORS file in the root of the source tree.
|
| - */
|
| -#include <immintrin.h>
|
| -#include "vpx_ports/mem.h"
|
| -
|
| -#define FSAD64_H(h) \
|
| -unsigned int vp9_sad64x##h##_avx2(const uint8_t *src_ptr, \
|
| - int src_stride, \
|
| - const uint8_t *ref_ptr, \
|
| - int ref_stride) { \
|
| - int i, res; \
|
| - __m256i sad1_reg, sad2_reg, ref1_reg, ref2_reg; \
|
| - __m256i sum_sad = _mm256_setzero_si256(); \
|
| - __m256i sum_sad_h; \
|
| - __m128i sum_sad128; \
|
| - for (i = 0 ; i < h ; i++) { \
|
| - ref1_reg = _mm256_loadu_si256((__m256i const *)ref_ptr); \
|
| - ref2_reg = _mm256_loadu_si256((__m256i const *)(ref_ptr + 32)); \
|
| - sad1_reg = _mm256_sad_epu8(ref1_reg, \
|
| - _mm256_loadu_si256((__m256i const *)src_ptr)); \
|
| - sad2_reg = _mm256_sad_epu8(ref2_reg, \
|
| - _mm256_loadu_si256((__m256i const *)(src_ptr + 32))); \
|
| - sum_sad = _mm256_add_epi32(sum_sad, _mm256_add_epi32(sad1_reg, sad2_reg)); \
|
| - ref_ptr+= ref_stride; \
|
| - src_ptr+= src_stride; \
|
| - } \
|
| - sum_sad_h = _mm256_srli_si256(sum_sad, 8); \
|
| - sum_sad = _mm256_add_epi32(sum_sad, sum_sad_h); \
|
| - sum_sad128 = _mm256_extracti128_si256(sum_sad, 1); \
|
| - sum_sad128 = _mm_add_epi32(_mm256_castsi256_si128(sum_sad), sum_sad128); \
|
| - res = _mm_cvtsi128_si32(sum_sad128); \
|
| - return res; \
|
| -}
|
| -
|
| -#define FSAD32_H(h) \
|
| -unsigned int vp9_sad32x##h##_avx2(const uint8_t *src_ptr, \
|
| - int src_stride, \
|
| - const uint8_t *ref_ptr, \
|
| - int ref_stride) { \
|
| - int i, res; \
|
| - __m256i sad1_reg, sad2_reg, ref1_reg, ref2_reg; \
|
| - __m256i sum_sad = _mm256_setzero_si256(); \
|
| - __m256i sum_sad_h; \
|
| - __m128i sum_sad128; \
|
| - int ref2_stride = ref_stride << 1; \
|
| - int src2_stride = src_stride << 1; \
|
| - int max = h >> 1; \
|
| - for (i = 0 ; i < max ; i++) { \
|
| - ref1_reg = _mm256_loadu_si256((__m256i const *)ref_ptr); \
|
| - ref2_reg = _mm256_loadu_si256((__m256i const *)(ref_ptr + ref_stride)); \
|
| - sad1_reg = _mm256_sad_epu8(ref1_reg, \
|
| - _mm256_loadu_si256((__m256i const *)src_ptr)); \
|
| - sad2_reg = _mm256_sad_epu8(ref2_reg, \
|
| - _mm256_loadu_si256((__m256i const *)(src_ptr + src_stride))); \
|
| - sum_sad = _mm256_add_epi32(sum_sad, _mm256_add_epi32(sad1_reg, sad2_reg)); \
|
| - ref_ptr+= ref2_stride; \
|
| - src_ptr+= src2_stride; \
|
| - } \
|
| - sum_sad_h = _mm256_srli_si256(sum_sad, 8); \
|
| - sum_sad = _mm256_add_epi32(sum_sad, sum_sad_h); \
|
| - sum_sad128 = _mm256_extracti128_si256(sum_sad, 1); \
|
| - sum_sad128 = _mm_add_epi32(_mm256_castsi256_si128(sum_sad), sum_sad128); \
|
| - res = _mm_cvtsi128_si32(sum_sad128); \
|
| - return res; \
|
| -}
|
| -
|
| -#define FSAD64 \
|
| -FSAD64_H(64); \
|
| -FSAD64_H(32);
|
| -
|
| -#define FSAD32 \
|
| -FSAD32_H(64); \
|
| -FSAD32_H(32); \
|
| -FSAD32_H(16);
|
| -
|
| -FSAD64;
|
| -FSAD32;
|
| -
|
| -#undef FSAD64
|
| -#undef FSAD32
|
| -#undef FSAD64_H
|
| -#undef FSAD32_H
|
| -
|
| -#define FSADAVG64_H(h) \
|
| -unsigned int vp9_sad64x##h##_avg_avx2(const uint8_t *src_ptr, \
|
| - int src_stride, \
|
| - const uint8_t *ref_ptr, \
|
| - int ref_stride, \
|
| - const uint8_t *second_pred) { \
|
| - int i, res; \
|
| - __m256i sad1_reg, sad2_reg, ref1_reg, ref2_reg; \
|
| - __m256i sum_sad = _mm256_setzero_si256(); \
|
| - __m256i sum_sad_h; \
|
| - __m128i sum_sad128; \
|
| - for (i = 0 ; i < h ; i++) { \
|
| - ref1_reg = _mm256_loadu_si256((__m256i const *)ref_ptr); \
|
| - ref2_reg = _mm256_loadu_si256((__m256i const *)(ref_ptr + 32)); \
|
| - ref1_reg = _mm256_avg_epu8(ref1_reg, \
|
| - _mm256_loadu_si256((__m256i const *)second_pred)); \
|
| - ref2_reg = _mm256_avg_epu8(ref2_reg, \
|
| - _mm256_loadu_si256((__m256i const *)(second_pred +32))); \
|
| - sad1_reg = _mm256_sad_epu8(ref1_reg, \
|
| - _mm256_loadu_si256((__m256i const *)src_ptr)); \
|
| - sad2_reg = _mm256_sad_epu8(ref2_reg, \
|
| - _mm256_loadu_si256((__m256i const *)(src_ptr + 32))); \
|
| - sum_sad = _mm256_add_epi32(sum_sad, _mm256_add_epi32(sad1_reg, sad2_reg)); \
|
| - ref_ptr+= ref_stride; \
|
| - src_ptr+= src_stride; \
|
| - second_pred+= 64; \
|
| - } \
|
| - sum_sad_h = _mm256_srli_si256(sum_sad, 8); \
|
| - sum_sad = _mm256_add_epi32(sum_sad, sum_sad_h); \
|
| - sum_sad128 = _mm256_extracti128_si256(sum_sad, 1); \
|
| - sum_sad128 = _mm_add_epi32(_mm256_castsi256_si128(sum_sad), sum_sad128); \
|
| - res = _mm_cvtsi128_si32(sum_sad128); \
|
| - return res; \
|
| -}
|
| -
|
| -#define FSADAVG32_H(h) \
|
| -unsigned int vp9_sad32x##h##_avg_avx2(const uint8_t *src_ptr, \
|
| - int src_stride, \
|
| - const uint8_t *ref_ptr, \
|
| - int ref_stride, \
|
| - const uint8_t *second_pred) { \
|
| - int i, res; \
|
| - __m256i sad1_reg, sad2_reg, ref1_reg, ref2_reg; \
|
| - __m256i sum_sad = _mm256_setzero_si256(); \
|
| - __m256i sum_sad_h; \
|
| - __m128i sum_sad128; \
|
| - int ref2_stride = ref_stride << 1; \
|
| - int src2_stride = src_stride << 1; \
|
| - int max = h >> 1; \
|
| - for (i = 0 ; i < max ; i++) { \
|
| - ref1_reg = _mm256_loadu_si256((__m256i const *)ref_ptr); \
|
| - ref2_reg = _mm256_loadu_si256((__m256i const *)(ref_ptr + ref_stride)); \
|
| - ref1_reg = _mm256_avg_epu8(ref1_reg, \
|
| - _mm256_loadu_si256((__m256i const *)second_pred)); \
|
| - ref2_reg = _mm256_avg_epu8(ref2_reg, \
|
| - _mm256_loadu_si256((__m256i const *)(second_pred +32))); \
|
| - sad1_reg = _mm256_sad_epu8(ref1_reg, \
|
| - _mm256_loadu_si256((__m256i const *)src_ptr)); \
|
| - sad2_reg = _mm256_sad_epu8(ref2_reg, \
|
| - _mm256_loadu_si256((__m256i const *)(src_ptr + src_stride))); \
|
| - sum_sad = _mm256_add_epi32(sum_sad, \
|
| - _mm256_add_epi32(sad1_reg, sad2_reg)); \
|
| - ref_ptr+= ref2_stride; \
|
| - src_ptr+= src2_stride; \
|
| - second_pred+= 64; \
|
| - } \
|
| - sum_sad_h = _mm256_srli_si256(sum_sad, 8); \
|
| - sum_sad = _mm256_add_epi32(sum_sad, sum_sad_h); \
|
| - sum_sad128 = _mm256_extracti128_si256(sum_sad, 1); \
|
| - sum_sad128 = _mm_add_epi32(_mm256_castsi256_si128(sum_sad), sum_sad128); \
|
| - res = _mm_cvtsi128_si32(sum_sad128); \
|
| - return res; \
|
| -}
|
| -
|
| -#define FSADAVG64 \
|
| -FSADAVG64_H(64); \
|
| -FSADAVG64_H(32);
|
| -
|
| -#define FSADAVG32 \
|
| -FSADAVG32_H(64); \
|
| -FSADAVG32_H(32); \
|
| -FSADAVG32_H(16);
|
| -
|
| -FSADAVG64;
|
| -FSADAVG32;
|
| -
|
| -#undef FSADAVG64
|
| -#undef FSADAVG32
|
| -#undef FSADAVG64_H
|
| -#undef FSADAVG32_H
|
|
|