Index: source/libvpx/third_party/libyuv/source/row_any.cc |
=================================================================== |
--- source/libvpx/third_party/libyuv/source/row_any.cc (revision 0) |
+++ source/libvpx/third_party/libyuv/source/row_any.cc (revision 0) |
@@ -0,0 +1,542 @@ |
+/* |
+ * Copyright 2012 The LibYuv Project Authors. All rights reserved. |
+ * |
+ * Use of this source code is governed by a BSD-style license |
+ * that can be found in the LICENSE file in the root of the source |
+ * tree. An additional intellectual property rights grant can be found |
+ * in the file PATENTS. All contributing project authors may |
+ * be found in the AUTHORS file in the root of the source tree. |
+ */ |
+ |
+#include "third_party/libyuv/include/libyuv/row.h" |
+ |
+#include "third_party/libyuv/include/libyuv/basic_types.h" |
+ |
+#ifdef __cplusplus |
+namespace libyuv { |
+extern "C" { |
+#endif |
+ |
+// TODO(fbarchard): Consider 'any' functions handling any quantity of pixels. |
+// TODO(fbarchard): Consider 'any' functions handling odd alignment. |
+// YUV to RGB does multiple of 8 with SIMD and remainder with C. |
+#define YANY(NAMEANY, I420TORGB_SIMD, I420TORGB_C, UV_SHIFT, BPP, MASK) \ |
+ void NAMEANY(const uint8* y_buf, \ |
+ const uint8* u_buf, \ |
+ const uint8* v_buf, \ |
+ uint8* rgb_buf, \ |
+ int width) { \ |
+ int n = width & ~MASK; \ |
+ I420TORGB_SIMD(y_buf, u_buf, v_buf, rgb_buf, n); \ |
+ I420TORGB_C(y_buf + n, \ |
+ u_buf + (n >> UV_SHIFT), \ |
+ v_buf + (n >> UV_SHIFT), \ |
+ rgb_buf + n * BPP, width & MASK); \ |
+ } |
+ |
+#ifdef HAS_I422TOARGBROW_SSSE3 |
+YANY(I444ToARGBRow_Any_SSSE3, I444ToARGBRow_Unaligned_SSSE3, I444ToARGBRow_C, |
+ 0, 4, 7) |
+YANY(I422ToARGBRow_Any_SSSE3, I422ToARGBRow_Unaligned_SSSE3, I422ToARGBRow_C, |
+ 1, 4, 7) |
+YANY(I411ToARGBRow_Any_SSSE3, I411ToARGBRow_Unaligned_SSSE3, I411ToARGBRow_C, |
+ 2, 4, 7) |
+YANY(I422ToBGRARow_Any_SSSE3, I422ToBGRARow_Unaligned_SSSE3, I422ToBGRARow_C, |
+ 1, 4, 7) |
+YANY(I422ToABGRRow_Any_SSSE3, I422ToABGRRow_Unaligned_SSSE3, I422ToABGRRow_C, |
+ 1, 4, 7) |
+YANY(I422ToRGBARow_Any_SSSE3, I422ToRGBARow_Unaligned_SSSE3, I422ToRGBARow_C, |
+ 1, 4, 7) |
+// I422ToRGB565Row_SSSE3 is unaligned. |
+YANY(I422ToARGB4444Row_Any_SSSE3, I422ToARGB4444Row_SSSE3, I422ToARGB4444Row_C, |
+ 1, 2, 7) |
+YANY(I422ToARGB1555Row_Any_SSSE3, I422ToARGB1555Row_SSSE3, I422ToARGB1555Row_C, |
+ 1, 2, 7) |
+YANY(I422ToRGB565Row_Any_SSSE3, I422ToRGB565Row_SSSE3, I422ToRGB565Row_C, |
+ 1, 2, 7) |
+// I422ToRGB24Row_SSSE3 is unaligned. |
+YANY(I422ToRGB24Row_Any_SSSE3, I422ToRGB24Row_SSSE3, I422ToRGB24Row_C, 1, 3, 7) |
+YANY(I422ToRAWRow_Any_SSSE3, I422ToRAWRow_SSSE3, I422ToRAWRow_C, 1, 3, 7) |
+YANY(I422ToYUY2Row_Any_SSE2, I422ToYUY2Row_SSE2, I422ToYUY2Row_C, 1, 2, 15) |
+YANY(I422ToUYVYRow_Any_SSE2, I422ToUYVYRow_SSE2, I422ToUYVYRow_C, 1, 2, 15) |
+#endif // HAS_I422TOARGBROW_SSSE3 |
+#ifdef HAS_I422TOARGBROW_AVX2 |
+YANY(I422ToARGBRow_Any_AVX2, I422ToARGBRow_AVX2, I422ToARGBRow_C, 1, 4, 15) |
+#endif // HAS_I422TOARGBROW_AVX2 |
+#ifdef HAS_I422TOARGBROW_NEON |
+YANY(I444ToARGBRow_Any_NEON, I444ToARGBRow_NEON, I444ToARGBRow_C, 0, 4, 7) |
+YANY(I422ToARGBRow_Any_NEON, I422ToARGBRow_NEON, I422ToARGBRow_C, 1, 4, 7) |
+YANY(I411ToARGBRow_Any_NEON, I411ToARGBRow_NEON, I411ToARGBRow_C, 2, 4, 7) |
+YANY(I422ToBGRARow_Any_NEON, I422ToBGRARow_NEON, I422ToBGRARow_C, 1, 4, 7) |
+YANY(I422ToABGRRow_Any_NEON, I422ToABGRRow_NEON, I422ToABGRRow_C, 1, 4, 7) |
+YANY(I422ToRGBARow_Any_NEON, I422ToRGBARow_NEON, I422ToRGBARow_C, 1, 4, 7) |
+YANY(I422ToRGB24Row_Any_NEON, I422ToRGB24Row_NEON, I422ToRGB24Row_C, 1, 3, 7) |
+YANY(I422ToRAWRow_Any_NEON, I422ToRAWRow_NEON, I422ToRAWRow_C, 1, 3, 7) |
+YANY(I422ToARGB4444Row_Any_NEON, I422ToARGB4444Row_NEON, I422ToARGB4444Row_C, |
+ 1, 2, 7) |
+YANY(I422ToARGB1555Row_Any_NEON, I422ToARGB1555Row_NEON, I422ToARGB1555Row_C, |
+ 1, 2, 7) |
+YANY(I422ToRGB565Row_Any_NEON, I422ToRGB565Row_NEON, I422ToRGB565Row_C, 1, 2, 7) |
+YANY(I422ToYUY2Row_Any_NEON, I422ToYUY2Row_NEON, I422ToYUY2Row_C, 1, 2, 15) |
+YANY(I422ToUYVYRow_Any_NEON, I422ToUYVYRow_NEON, I422ToUYVYRow_C, 1, 2, 15) |
+#endif // HAS_I422TOARGBROW_NEON |
+#undef YANY |
+ |
+// Wrappers to handle odd width |
+#define NV2NY(NAMEANY, NV12TORGB_SIMD, NV12TORGB_C, UV_SHIFT, BPP) \ |
+ void NAMEANY(const uint8* y_buf, \ |
+ const uint8* uv_buf, \ |
+ uint8* rgb_buf, \ |
+ int width) { \ |
+ int n = width & ~7; \ |
+ NV12TORGB_SIMD(y_buf, uv_buf, rgb_buf, n); \ |
+ NV12TORGB_C(y_buf + n, \ |
+ uv_buf + (n >> UV_SHIFT), \ |
+ rgb_buf + n * BPP, width & 7); \ |
+ } |
+ |
+#ifdef HAS_NV12TOARGBROW_SSSE3 |
+NV2NY(NV12ToARGBRow_Any_SSSE3, NV12ToARGBRow_Unaligned_SSSE3, NV12ToARGBRow_C, |
+ 0, 4) |
+NV2NY(NV21ToARGBRow_Any_SSSE3, NV21ToARGBRow_Unaligned_SSSE3, NV21ToARGBRow_C, |
+ 0, 4) |
+#endif // HAS_NV12TOARGBROW_SSSE3 |
+#ifdef HAS_NV12TOARGBROW_NEON |
+NV2NY(NV12ToARGBRow_Any_NEON, NV12ToARGBRow_NEON, NV12ToARGBRow_C, 0, 4) |
+NV2NY(NV21ToARGBRow_Any_NEON, NV21ToARGBRow_NEON, NV21ToARGBRow_C, 0, 4) |
+#endif // HAS_NV12TOARGBROW_NEON |
+#ifdef HAS_NV12TORGB565ROW_SSSE3 |
+NV2NY(NV12ToRGB565Row_Any_SSSE3, NV12ToRGB565Row_SSSE3, NV12ToRGB565Row_C, |
+ 0, 2) |
+NV2NY(NV21ToRGB565Row_Any_SSSE3, NV21ToRGB565Row_SSSE3, NV21ToRGB565Row_C, |
+ 0, 2) |
+#endif // HAS_NV12TORGB565ROW_SSSE3 |
+#ifdef HAS_NV12TORGB565ROW_NEON |
+NV2NY(NV12ToRGB565Row_Any_NEON, NV12ToRGB565Row_NEON, NV12ToRGB565Row_C, 0, 2) |
+NV2NY(NV21ToRGB565Row_Any_NEON, NV21ToRGB565Row_NEON, NV21ToRGB565Row_C, 0, 2) |
+#endif // HAS_NV12TORGB565ROW_NEON |
+#undef NVANY |
+ |
+#define RGBANY(NAMEANY, ARGBTORGB_SIMD, ARGBTORGB_C, MASK, SBPP, BPP) \ |
+ void NAMEANY(const uint8* src, \ |
+ uint8* dst, \ |
+ int width) { \ |
+ int n = width & ~MASK; \ |
+ ARGBTORGB_SIMD(src, dst, n); \ |
+ ARGBTORGB_C(src + n * SBPP, dst + n * BPP, width & MASK); \ |
+ } |
+ |
+#if defined(HAS_ARGBTORGB24ROW_SSSE3) |
+RGBANY(ARGBToRGB24Row_Any_SSSE3, ARGBToRGB24Row_SSSE3, ARGBToRGB24Row_C, |
+ 15, 4, 3) |
+RGBANY(ARGBToRAWRow_Any_SSSE3, ARGBToRAWRow_SSSE3, ARGBToRAWRow_C, |
+ 15, 4, 3) |
+RGBANY(ARGBToRGB565Row_Any_SSE2, ARGBToRGB565Row_SSE2, ARGBToRGB565Row_C, |
+ 3, 4, 2) |
+RGBANY(ARGBToARGB1555Row_Any_SSE2, ARGBToARGB1555Row_SSE2, ARGBToARGB1555Row_C, |
+ 3, 4, 2) |
+RGBANY(ARGBToARGB4444Row_Any_SSE2, ARGBToARGB4444Row_SSE2, ARGBToARGB4444Row_C, |
+ 3, 4, 2) |
+#endif |
+#if defined(HAS_I400TOARGBROW_SSE2) |
+RGBANY(I400ToARGBRow_Any_SSE2, I400ToARGBRow_Unaligned_SSE2, I400ToARGBRow_C, |
+ 7, 1, 4) |
+#endif |
+#if defined(HAS_YTOARGBROW_SSE2) |
+RGBANY(YToARGBRow_Any_SSE2, YToARGBRow_SSE2, YToARGBRow_C, |
+ 7, 1, 4) |
+RGBANY(YUY2ToARGBRow_Any_SSSE3, YUY2ToARGBRow_Unaligned_SSSE3, YUY2ToARGBRow_C, |
+ 15, 2, 4) |
+RGBANY(UYVYToARGBRow_Any_SSSE3, UYVYToARGBRow_Unaligned_SSSE3, UYVYToARGBRow_C, |
+ 15, 2, 4) |
+// These require alignment on ARGB, so C is used for remainder. |
+RGBANY(RGB24ToARGBRow_Any_SSSE3, RGB24ToARGBRow_SSSE3, RGB24ToARGBRow_C, |
+ 15, 3, 4) |
+RGBANY(RAWToARGBRow_Any_SSSE3, RAWToARGBRow_SSSE3, RAWToARGBRow_C, |
+ 15, 3, 4) |
+RGBANY(RGB565ToARGBRow_Any_SSE2, RGB565ToARGBRow_SSE2, RGB565ToARGBRow_C, |
+ 7, 2, 4) |
+RGBANY(ARGB1555ToARGBRow_Any_SSE2, ARGB1555ToARGBRow_SSE2, ARGB1555ToARGBRow_C, |
+ 7, 2, 4) |
+RGBANY(ARGB4444ToARGBRow_Any_SSE2, ARGB4444ToARGBRow_SSE2, ARGB4444ToARGBRow_C, |
+ 7, 2, 4) |
+#endif |
+#if defined(HAS_ARGBTORGB24ROW_NEON) |
+RGBANY(ARGBToRGB24Row_Any_NEON, ARGBToRGB24Row_NEON, ARGBToRGB24Row_C, 7, 4, 3) |
+RGBANY(ARGBToRAWRow_Any_NEON, ARGBToRAWRow_NEON, ARGBToRAWRow_C, 7, 4, 3) |
+RGBANY(ARGBToRGB565Row_Any_NEON, ARGBToRGB565Row_NEON, ARGBToRGB565Row_C, |
+ 7, 4, 2) |
+RGBANY(ARGBToARGB1555Row_Any_NEON, ARGBToARGB1555Row_NEON, ARGBToARGB1555Row_C, |
+ 7, 4, 2) |
+RGBANY(ARGBToARGB4444Row_Any_NEON, ARGBToARGB4444Row_NEON, ARGBToARGB4444Row_C, |
+ 7, 4, 2) |
+RGBANY(I400ToARGBRow_Any_NEON, I400ToARGBRow_NEON, I400ToARGBRow_C, |
+ 7, 1, 4) |
+RGBANY(YToARGBRow_Any_NEON, YToARGBRow_NEON, YToARGBRow_C, |
+ 7, 1, 4) |
+RGBANY(YUY2ToARGBRow_Any_NEON, YUY2ToARGBRow_NEON, YUY2ToARGBRow_C, |
+ 7, 2, 4) |
+RGBANY(UYVYToARGBRow_Any_NEON, UYVYToARGBRow_NEON, UYVYToARGBRow_C, |
+ 7, 2, 4) |
+#endif |
+#undef RGBANY |
+ |
+// ARGB to Bayer does multiple of 4 pixels, SSSE3 aligned src, unaligned dst. |
+#define BAYERANY(NAMEANY, ARGBTORGB_SIMD, ARGBTORGB_C, MASK, SBPP, BPP) \ |
+ void NAMEANY(const uint8* src, \ |
+ uint8* dst, uint32 selector, \ |
+ int width) { \ |
+ int n = width & ~MASK; \ |
+ ARGBTORGB_SIMD(src, dst, selector, n); \ |
+ ARGBTORGB_C(src + n * SBPP, dst + n * BPP, selector, width & MASK); \ |
+ } |
+ |
+#if defined(HAS_ARGBTOBAYERROW_SSSE3) |
+BAYERANY(ARGBToBayerRow_Any_SSSE3, ARGBToBayerRow_SSSE3, ARGBToBayerRow_C, |
+ 7, 4, 1) |
+#endif |
+#if defined(HAS_ARGBTOBAYERROW_NEON) |
+BAYERANY(ARGBToBayerRow_Any_NEON, ARGBToBayerRow_NEON, ARGBToBayerRow_C, |
+ 7, 4, 1) |
+#endif |
+#if defined(HAS_ARGBTOBAYERGGROW_SSE2) |
+BAYERANY(ARGBToBayerGGRow_Any_SSE2, ARGBToBayerGGRow_SSE2, ARGBToBayerGGRow_C, |
+ 7, 4, 1) |
+#endif |
+#if defined(HAS_ARGBTOBAYERGGROW_NEON) |
+BAYERANY(ARGBToBayerGGRow_Any_NEON, ARGBToBayerGGRow_NEON, ARGBToBayerGGRow_C, |
+ 7, 4, 1) |
+#endif |
+ |
+#undef BAYERANY |
+ |
+// RGB/YUV to Y does multiple of 16 with SIMD and last 16 with SIMD. |
+#define YANY(NAMEANY, ARGBTOY_SIMD, SBPP, BPP, NUM) \ |
+ void NAMEANY(const uint8* src_argb, uint8* dst_y, int width) { \ |
+ ARGBTOY_SIMD(src_argb, dst_y, width - NUM); \ |
+ ARGBTOY_SIMD(src_argb + (width - NUM) * SBPP, \ |
+ dst_y + (width - NUM) * BPP, NUM); \ |
+ } |
+ |
+#ifdef HAS_ARGBTOYROW_AVX2 |
+YANY(ARGBToYRow_Any_AVX2, ARGBToYRow_AVX2, 4, 1, 32) |
+YANY(ARGBToYJRow_Any_AVX2, ARGBToYJRow_AVX2, 4, 1, 32) |
+YANY(YUY2ToYRow_Any_AVX2, YUY2ToYRow_AVX2, 2, 1, 32) |
+YANY(UYVYToYRow_Any_AVX2, UYVYToYRow_AVX2, 2, 1, 32) |
+#endif |
+#ifdef HAS_ARGBTOYROW_SSSE3 |
+YANY(ARGBToYRow_Any_SSSE3, ARGBToYRow_Unaligned_SSSE3, 4, 1, 16) |
+#endif |
+#ifdef HAS_BGRATOYROW_SSSE3 |
+YANY(BGRAToYRow_Any_SSSE3, BGRAToYRow_Unaligned_SSSE3, 4, 1, 16) |
+YANY(ABGRToYRow_Any_SSSE3, ABGRToYRow_Unaligned_SSSE3, 4, 1, 16) |
+YANY(RGBAToYRow_Any_SSSE3, RGBAToYRow_Unaligned_SSSE3, 4, 1, 16) |
+YANY(YUY2ToYRow_Any_SSE2, YUY2ToYRow_Unaligned_SSE2, 2, 1, 16) |
+YANY(UYVYToYRow_Any_SSE2, UYVYToYRow_Unaligned_SSE2, 2, 1, 16) |
+#endif |
+#ifdef HAS_ARGBTOYJROW_SSSE3 |
+YANY(ARGBToYJRow_Any_SSSE3, ARGBToYJRow_Unaligned_SSSE3, 4, 1, 16) |
+#endif |
+#ifdef HAS_ARGBTOYROW_NEON |
+YANY(ARGBToYRow_Any_NEON, ARGBToYRow_NEON, 4, 1, 8) |
+YANY(ARGBToYJRow_Any_NEON, ARGBToYJRow_NEON, 4, 1, 8) |
+YANY(BGRAToYRow_Any_NEON, BGRAToYRow_NEON, 4, 1, 8) |
+YANY(ABGRToYRow_Any_NEON, ABGRToYRow_NEON, 4, 1, 8) |
+YANY(RGBAToYRow_Any_NEON, RGBAToYRow_NEON, 4, 1, 8) |
+YANY(RGB24ToYRow_Any_NEON, RGB24ToYRow_NEON, 3, 1, 8) |
+YANY(RAWToYRow_Any_NEON, RAWToYRow_NEON, 3, 1, 8) |
+YANY(RGB565ToYRow_Any_NEON, RGB565ToYRow_NEON, 2, 1, 8) |
+YANY(ARGB1555ToYRow_Any_NEON, ARGB1555ToYRow_NEON, 2, 1, 8) |
+YANY(ARGB4444ToYRow_Any_NEON, ARGB4444ToYRow_NEON, 2, 1, 8) |
+YANY(YUY2ToYRow_Any_NEON, YUY2ToYRow_NEON, 2, 1, 16) |
+YANY(UYVYToYRow_Any_NEON, UYVYToYRow_NEON, 2, 1, 16) |
+YANY(RGB24ToARGBRow_Any_NEON, RGB24ToARGBRow_NEON, 3, 4, 8) |
+YANY(RAWToARGBRow_Any_NEON, RAWToARGBRow_NEON, 3, 4, 8) |
+YANY(RGB565ToARGBRow_Any_NEON, RGB565ToARGBRow_NEON, 2, 4, 8) |
+YANY(ARGB1555ToARGBRow_Any_NEON, ARGB1555ToARGBRow_NEON, 2, 4, 8) |
+YANY(ARGB4444ToARGBRow_Any_NEON, ARGB4444ToARGBRow_NEON, 2, 4, 8) |
+#endif |
+#undef YANY |
+ |
+#define YANY(NAMEANY, ARGBTOY_SIMD, ARGBTOY_C, SBPP, BPP, MASK) \ |
+ void NAMEANY(const uint8* src_argb, uint8* dst_y, int width) { \ |
+ int n = width & ~MASK; \ |
+ ARGBTOY_SIMD(src_argb, dst_y, n); \ |
+ ARGBTOY_C(src_argb + n * SBPP, \ |
+ dst_y + n * BPP, width & MASK); \ |
+ } |
+ |
+// Attenuate is destructive so last16 method can not be used due to overlap. |
+#ifdef HAS_ARGBATTENUATEROW_SSSE3 |
+YANY(ARGBAttenuateRow_Any_SSSE3, ARGBAttenuateRow_SSSE3, ARGBAttenuateRow_C, |
+ 4, 4, 3) |
+#endif |
+#ifdef HAS_ARGBATTENUATEROW_SSE2 |
+YANY(ARGBAttenuateRow_Any_SSE2, ARGBAttenuateRow_SSE2, ARGBAttenuateRow_C, |
+ 4, 4, 3) |
+#endif |
+#ifdef HAS_ARGBUNATTENUATEROW_SSE2 |
+YANY(ARGBUnattenuateRow_Any_SSE2, ARGBUnattenuateRow_SSE2, ARGBUnattenuateRow_C, |
+ 4, 4, 3) |
+#endif |
+#ifdef HAS_ARGBATTENUATEROW_AVX2 |
+YANY(ARGBAttenuateRow_Any_AVX2, ARGBAttenuateRow_AVX2, ARGBAttenuateRow_C, |
+ 4, 4, 7) |
+#endif |
+#ifdef HAS_ARGBUNATTENUATEROW_AVX2 |
+YANY(ARGBUnattenuateRow_Any_AVX2, ARGBUnattenuateRow_AVX2, ARGBUnattenuateRow_C, |
+ 4, 4, 7) |
+#endif |
+#ifdef HAS_ARGBATTENUATEROW_NEON |
+YANY(ARGBAttenuateRow_Any_NEON, ARGBAttenuateRow_NEON, ARGBAttenuateRow_C, |
+ 4, 4, 7) |
+#endif |
+#undef YANY |
+ |
+// RGB/YUV to UV does multiple of 16 with SIMD and remainder with C. |
+#define UVANY(NAMEANY, ANYTOUV_SIMD, ANYTOUV_C, BPP, MASK) \ |
+ void NAMEANY(const uint8* src_argb, int src_stride_argb, \ |
+ uint8* dst_u, uint8* dst_v, int width) { \ |
+ int n = width & ~MASK; \ |
+ ANYTOUV_SIMD(src_argb, src_stride_argb, dst_u, dst_v, n); \ |
+ ANYTOUV_C(src_argb + n * BPP, src_stride_argb, \ |
+ dst_u + (n >> 1), \ |
+ dst_v + (n >> 1), \ |
+ width & MASK); \ |
+ } |
+ |
+#ifdef HAS_ARGBTOUVROW_AVX2 |
+UVANY(ARGBToUVRow_Any_AVX2, ARGBToUVRow_AVX2, ARGBToUVRow_C, 4, 31) |
+UVANY(YUY2ToUVRow_Any_AVX2, YUY2ToUVRow_AVX2, YUY2ToUVRow_C, 2, 31) |
+UVANY(UYVYToUVRow_Any_AVX2, UYVYToUVRow_AVX2, UYVYToUVRow_C, 2, 31) |
+#endif |
+#ifdef HAS_ARGBTOUVROW_SSSE3 |
+UVANY(ARGBToUVRow_Any_SSSE3, ARGBToUVRow_Unaligned_SSSE3, ARGBToUVRow_C, 4, 15) |
+UVANY(ARGBToUVJRow_Any_SSSE3, ARGBToUVJRow_Unaligned_SSSE3, ARGBToUVJRow_C, |
+ 4, 15) |
+UVANY(BGRAToUVRow_Any_SSSE3, BGRAToUVRow_Unaligned_SSSE3, BGRAToUVRow_C, 4, 15) |
+UVANY(ABGRToUVRow_Any_SSSE3, ABGRToUVRow_Unaligned_SSSE3, ABGRToUVRow_C, 4, 15) |
+UVANY(RGBAToUVRow_Any_SSSE3, RGBAToUVRow_Unaligned_SSSE3, RGBAToUVRow_C, 4, 15) |
+UVANY(YUY2ToUVRow_Any_SSE2, YUY2ToUVRow_Unaligned_SSE2, YUY2ToUVRow_C, 2, 15) |
+UVANY(UYVYToUVRow_Any_SSE2, UYVYToUVRow_Unaligned_SSE2, UYVYToUVRow_C, 2, 15) |
+#endif |
+#ifdef HAS_ARGBTOUVROW_NEON |
+UVANY(ARGBToUVRow_Any_NEON, ARGBToUVRow_NEON, ARGBToUVRow_C, 4, 15) |
+UVANY(ARGBToUVJRow_Any_NEON, ARGBToUVJRow_NEON, ARGBToUVJRow_C, 4, 15) |
+UVANY(BGRAToUVRow_Any_NEON, BGRAToUVRow_NEON, BGRAToUVRow_C, 4, 15) |
+UVANY(ABGRToUVRow_Any_NEON, ABGRToUVRow_NEON, ABGRToUVRow_C, 4, 15) |
+UVANY(RGBAToUVRow_Any_NEON, RGBAToUVRow_NEON, RGBAToUVRow_C, 4, 15) |
+UVANY(RGB24ToUVRow_Any_NEON, RGB24ToUVRow_NEON, RGB24ToUVRow_C, 3, 15) |
+UVANY(RAWToUVRow_Any_NEON, RAWToUVRow_NEON, RAWToUVRow_C, 3, 15) |
+UVANY(RGB565ToUVRow_Any_NEON, RGB565ToUVRow_NEON, RGB565ToUVRow_C, 2, 15) |
+UVANY(ARGB1555ToUVRow_Any_NEON, ARGB1555ToUVRow_NEON, ARGB1555ToUVRow_C, 2, 15) |
+UVANY(ARGB4444ToUVRow_Any_NEON, ARGB4444ToUVRow_NEON, ARGB4444ToUVRow_C, 2, 15) |
+UVANY(YUY2ToUVRow_Any_NEON, YUY2ToUVRow_NEON, YUY2ToUVRow_C, 2, 15) |
+UVANY(UYVYToUVRow_Any_NEON, UYVYToUVRow_NEON, UYVYToUVRow_C, 2, 15) |
+#endif |
+#undef UVANY |
+ |
+#define UV422ANY(NAMEANY, ANYTOUV_SIMD, ANYTOUV_C, BPP, MASK, SHIFT) \ |
+ void NAMEANY(const uint8* src_uv, \ |
+ uint8* dst_u, uint8* dst_v, int width) { \ |
+ int n = width & ~MASK; \ |
+ ANYTOUV_SIMD(src_uv, dst_u, dst_v, n); \ |
+ ANYTOUV_C(src_uv + n * BPP, \ |
+ dst_u + (n >> SHIFT), \ |
+ dst_v + (n >> SHIFT), \ |
+ width & MASK); \ |
+ } |
+ |
+#ifdef HAS_ARGBTOUV444ROW_SSSE3 |
+UV422ANY(ARGBToUV444Row_Any_SSSE3, ARGBToUV444Row_Unaligned_SSSE3, |
+ ARGBToUV444Row_C, 4, 15, 0) |
+#endif |
+#ifdef HAS_YUY2TOUV422ROW_AVX2 |
+UV422ANY(YUY2ToUV422Row_Any_AVX2, YUY2ToUV422Row_AVX2, |
+ YUY2ToUV422Row_C, 2, 31, 1) |
+UV422ANY(UYVYToUV422Row_Any_AVX2, UYVYToUV422Row_AVX2, |
+ UYVYToUV422Row_C, 2, 31, 1) |
+#endif |
+#ifdef HAS_ARGBTOUVROW_SSSE3 |
+UV422ANY(ARGBToUV422Row_Any_SSSE3, ARGBToUV422Row_Unaligned_SSSE3, |
+ ARGBToUV422Row_C, 4, 15, 1) |
+UV422ANY(YUY2ToUV422Row_Any_SSE2, YUY2ToUV422Row_Unaligned_SSE2, |
+ YUY2ToUV422Row_C, 2, 15, 1) |
+UV422ANY(UYVYToUV422Row_Any_SSE2, UYVYToUV422Row_Unaligned_SSE2, |
+ UYVYToUV422Row_C, 2, 15, 1) |
+#endif |
+#ifdef HAS_YUY2TOUV422ROW_NEON |
+UV422ANY(ARGBToUV444Row_Any_NEON, ARGBToUV444Row_NEON, |
+ ARGBToUV444Row_C, 4, 7, 0) |
+UV422ANY(ARGBToUV422Row_Any_NEON, ARGBToUV422Row_NEON, |
+ ARGBToUV422Row_C, 4, 15, 1) |
+UV422ANY(ARGBToUV411Row_Any_NEON, ARGBToUV411Row_NEON, |
+ ARGBToUV411Row_C, 4, 31, 2) |
+UV422ANY(YUY2ToUV422Row_Any_NEON, YUY2ToUV422Row_NEON, |
+ YUY2ToUV422Row_C, 2, 15, 1) |
+UV422ANY(UYVYToUV422Row_Any_NEON, UYVYToUV422Row_NEON, |
+ UYVYToUV422Row_C, 2, 15, 1) |
+#endif |
+#undef UV422ANY |
+ |
+#define SPLITUVROWANY(NAMEANY, ANYTOUV_SIMD, ANYTOUV_C, MASK) \ |
+ void NAMEANY(const uint8* src_uv, \ |
+ uint8* dst_u, uint8* dst_v, int width) { \ |
+ int n = width & ~MASK; \ |
+ ANYTOUV_SIMD(src_uv, dst_u, dst_v, n); \ |
+ ANYTOUV_C(src_uv + n * 2, \ |
+ dst_u + n, \ |
+ dst_v + n, \ |
+ width & MASK); \ |
+ } |
+ |
+#ifdef HAS_SPLITUVROW_SSE2 |
+SPLITUVROWANY(SplitUVRow_Any_SSE2, SplitUVRow_Unaligned_SSE2, SplitUVRow_C, 15) |
+#endif |
+#ifdef HAS_SPLITUVROW_AVX2 |
+SPLITUVROWANY(SplitUVRow_Any_AVX2, SplitUVRow_AVX2, SplitUVRow_C, 31) |
+#endif |
+#ifdef HAS_SPLITUVROW_NEON |
+SPLITUVROWANY(SplitUVRow_Any_NEON, SplitUVRow_NEON, SplitUVRow_C, 15) |
+#endif |
+#ifdef HAS_SPLITUVROW_MIPS_DSPR2 |
+SPLITUVROWANY(SplitUVRow_Any_MIPS_DSPR2, SplitUVRow_Unaligned_MIPS_DSPR2, |
+ SplitUVRow_C, 15) |
+#endif |
+#undef SPLITUVROWANY |
+ |
+#define MERGEUVROW_ANY(NAMEANY, ANYTOUV_SIMD, ANYTOUV_C, MASK) \ |
+ void NAMEANY(const uint8* src_u, const uint8* src_v, \ |
+ uint8* dst_uv, int width) { \ |
+ int n = width & ~MASK; \ |
+ ANYTOUV_SIMD(src_u, src_v, dst_uv, n); \ |
+ ANYTOUV_C(src_u + n, \ |
+ src_v + n, \ |
+ dst_uv + n * 2, \ |
+ width & MASK); \ |
+ } |
+ |
+#ifdef HAS_MERGEUVROW_SSE2 |
+MERGEUVROW_ANY(MergeUVRow_Any_SSE2, MergeUVRow_Unaligned_SSE2, MergeUVRow_C, 15) |
+#endif |
+#ifdef HAS_MERGEUVROW_AVX2 |
+MERGEUVROW_ANY(MergeUVRow_Any_AVX2, MergeUVRow_AVX2, MergeUVRow_C, 31) |
+#endif |
+#ifdef HAS_MERGEUVROW_NEON |
+MERGEUVROW_ANY(MergeUVRow_Any_NEON, MergeUVRow_NEON, MergeUVRow_C, 15) |
+#endif |
+#undef MERGEUVROW_ANY |
+ |
+#define MATHROW_ANY(NAMEANY, ARGBMATH_SIMD, ARGBMATH_C, MASK) \ |
+ void NAMEANY(const uint8* src_argb0, const uint8* src_argb1, \ |
+ uint8* dst_argb, int width) { \ |
+ int n = width & ~MASK; \ |
+ ARGBMATH_SIMD(src_argb0, src_argb1, dst_argb, n); \ |
+ ARGBMATH_C(src_argb0 + n * 4, \ |
+ src_argb1 + n * 4, \ |
+ dst_argb + n * 4, \ |
+ width & MASK); \ |
+ } |
+ |
+#ifdef HAS_ARGBMULTIPLYROW_SSE2 |
+MATHROW_ANY(ARGBMultiplyRow_Any_SSE2, ARGBMultiplyRow_SSE2, ARGBMultiplyRow_C, |
+ 3) |
+#endif |
+#ifdef HAS_ARGBADDROW_SSE2 |
+MATHROW_ANY(ARGBAddRow_Any_SSE2, ARGBAddRow_SSE2, ARGBAddRow_C, 3) |
+#endif |
+#ifdef HAS_ARGBSUBTRACTROW_SSE2 |
+MATHROW_ANY(ARGBSubtractRow_Any_SSE2, ARGBSubtractRow_SSE2, ARGBSubtractRow_C, |
+ 3) |
+#endif |
+#ifdef HAS_ARGBMULTIPLYROW_AVX2 |
+MATHROW_ANY(ARGBMultiplyRow_Any_AVX2, ARGBMultiplyRow_AVX2, ARGBMultiplyRow_C, |
+ 7) |
+#endif |
+#ifdef HAS_ARGBADDROW_AVX2 |
+MATHROW_ANY(ARGBAddRow_Any_AVX2, ARGBAddRow_AVX2, ARGBAddRow_C, 7) |
+#endif |
+#ifdef HAS_ARGBSUBTRACTROW_AVX2 |
+MATHROW_ANY(ARGBSubtractRow_Any_AVX2, ARGBSubtractRow_AVX2, ARGBSubtractRow_C, |
+ 7) |
+#endif |
+#ifdef HAS_ARGBMULTIPLYROW_NEON |
+MATHROW_ANY(ARGBMultiplyRow_Any_NEON, ARGBMultiplyRow_NEON, ARGBMultiplyRow_C, |
+ 7) |
+#endif |
+#ifdef HAS_ARGBADDROW_NEON |
+MATHROW_ANY(ARGBAddRow_Any_NEON, ARGBAddRow_NEON, ARGBAddRow_C, 7) |
+#endif |
+#ifdef HAS_ARGBSUBTRACTROW_NEON |
+MATHROW_ANY(ARGBSubtractRow_Any_NEON, ARGBSubtractRow_NEON, ARGBSubtractRow_C, |
+ 7) |
+#endif |
+#undef MATHROW_ANY |
+ |
+// Shuffle may want to work in place, so last16 method can not be used. |
+#define YANY(NAMEANY, ARGBTOY_SIMD, ARGBTOY_C, SBPP, BPP, MASK) \ |
+ void NAMEANY(const uint8* src_argb, uint8* dst_argb, \ |
+ const uint8* shuffler, int width) { \ |
+ int n = width & ~MASK; \ |
+ ARGBTOY_SIMD(src_argb, dst_argb, shuffler, n); \ |
+ ARGBTOY_C(src_argb + n * SBPP, \ |
+ dst_argb + n * BPP, shuffler, width & MASK); \ |
+ } |
+ |
+#ifdef HAS_ARGBSHUFFLEROW_SSE2 |
+YANY(ARGBShuffleRow_Any_SSE2, ARGBShuffleRow_SSE2, |
+ ARGBShuffleRow_C, 4, 4, 3) |
+#endif |
+#ifdef HAS_ARGBSHUFFLEROW_SSSE3 |
+YANY(ARGBShuffleRow_Any_SSSE3, ARGBShuffleRow_Unaligned_SSSE3, |
+ ARGBShuffleRow_C, 4, 4, 7) |
+#endif |
+#ifdef HAS_ARGBSHUFFLEROW_AVX2 |
+YANY(ARGBShuffleRow_Any_AVX2, ARGBShuffleRow_AVX2, |
+ ARGBShuffleRow_C, 4, 4, 15) |
+#endif |
+#ifdef HAS_ARGBSHUFFLEROW_NEON |
+YANY(ARGBShuffleRow_Any_NEON, ARGBShuffleRow_NEON, |
+ ARGBShuffleRow_C, 4, 4, 3) |
+#endif |
+#undef YANY |
+ |
+// Interpolate may want to work in place, so last16 method can not be used. |
+#define NANY(NAMEANY, TERP_SIMD, TERP_C, SBPP, BPP, MASK) \ |
+ void NAMEANY(uint8* dst_ptr, const uint8* src_ptr, \ |
+ ptrdiff_t src_stride_ptr, int width, \ |
+ int source_y_fraction) { \ |
+ int n = width & ~MASK; \ |
+ TERP_SIMD(dst_ptr, src_ptr, src_stride_ptr, \ |
+ n, source_y_fraction); \ |
+ TERP_C(dst_ptr + n * BPP, \ |
+ src_ptr + n * SBPP, src_stride_ptr, \ |
+ width & MASK, source_y_fraction); \ |
+ } |
+ |
+#ifdef HAS_INTERPOLATEROW_AVX2 |
+NANY(InterpolateRow_Any_AVX2, InterpolateRow_AVX2, |
+ InterpolateRow_C, 1, 1, 32) |
+#endif |
+#ifdef HAS_INTERPOLATEROW_SSSE3 |
+NANY(InterpolateRow_Any_SSSE3, InterpolateRow_Unaligned_SSSE3, |
+ InterpolateRow_C, 1, 1, 15) |
+#endif |
+#ifdef HAS_INTERPOLATEROW_SSE2 |
+NANY(InterpolateRow_Any_SSE2, InterpolateRow_Unaligned_SSE2, |
+ InterpolateRow_C, 1, 1, 15) |
+#endif |
+#ifdef HAS_INTERPOLATEROW_NEON |
+NANY(InterpolateRow_Any_NEON, InterpolateRow_NEON, |
+ InterpolateRow_C, 1, 1, 15) |
+#endif |
+#ifdef HAS_INTERPOLATEROW_MIPS_DSPR2 |
+NANY(InterpolateRow_Any_MIPS_DSPR2, InterpolateRow_MIPS_DSPR2, |
+ InterpolateRow_C, 1, 1, 3) |
+#endif |
+#undef NANY |
+ |
+#ifdef __cplusplus |
+} // extern "C" |
+} // namespace libyuv |
+#endif |