| Index: source/libvpx/third_party/libyuv/source/scale_argb.cc
|
| ===================================================================
|
| --- source/libvpx/third_party/libyuv/source/scale_argb.cc (revision 0)
|
| +++ source/libvpx/third_party/libyuv/source/scale_argb.cc (revision 0)
|
| @@ -0,0 +1,809 @@
|
| +/*
|
| + * Copyright 2011 The LibYuv Project Authors. All rights reserved.
|
| + *
|
| + * Use of this source code is governed by a BSD-style license
|
| + * that can be found in the LICENSE file in the root of the source
|
| + * tree. An additional intellectual property rights grant can be found
|
| + * in the file PATENTS. All contributing project authors may
|
| + * be found in the AUTHORS file in the root of the source tree.
|
| + */
|
| +
|
| +#include "libyuv/scale.h"
|
| +
|
| +#include <assert.h>
|
| +#include <string.h>
|
| +
|
| +#include "libyuv/cpu_id.h"
|
| +#include "libyuv/planar_functions.h" // For CopyARGB
|
| +#include "libyuv/row.h"
|
| +#include "libyuv/scale_row.h"
|
| +
|
| +#ifdef __cplusplus
|
| +namespace libyuv {
|
| +extern "C" {
|
| +#endif
|
| +
|
| +static __inline int Abs(int v) {
|
| + return v >= 0 ? v : -v;
|
| +}
|
| +
|
| +// ScaleARGB ARGB, 1/2
|
| +// This is an optimized version for scaling down a ARGB to 1/2 of
|
| +// its original size.
|
| +static void ScaleARGBDown2(int src_width, int src_height,
|
| + int dst_width, int dst_height,
|
| + int src_stride, int dst_stride,
|
| + const uint8* src_argb, uint8* dst_argb,
|
| + int x, int dx, int y, int dy,
|
| + enum FilterMode filtering) {
|
| + int j;
|
| + int row_stride = src_stride * (dy >> 16);
|
| + void (*ScaleARGBRowDown2)(const uint8* src_argb, ptrdiff_t src_stride,
|
| + uint8* dst_argb, int dst_width) =
|
| + filtering == kFilterNone ? ScaleARGBRowDown2_C :
|
| + (filtering == kFilterLinear ? ScaleARGBRowDown2Linear_C :
|
| + ScaleARGBRowDown2Box_C);
|
| + assert(dx == 65536 * 2); // Test scale factor of 2.
|
| + assert((dy & 0x1ffff) == 0); // Test vertical scale is multiple of 2.
|
| + // Advance to odd row, even column.
|
| + if (filtering == kFilterBilinear) {
|
| + src_argb += (y >> 16) * src_stride + (x >> 16) * 4;
|
| + } else {
|
| + src_argb += (y >> 16) * src_stride + ((x >> 16) - 1) * 4;
|
| + }
|
| +
|
| +#if defined(HAS_SCALEARGBROWDOWN2_SSE2)
|
| + if (TestCpuFlag(kCpuHasSSE2) && IS_ALIGNED(dst_width, 4) &&
|
| + IS_ALIGNED(src_argb, 16) && IS_ALIGNED(row_stride, 16) &&
|
| + IS_ALIGNED(dst_argb, 16) && IS_ALIGNED(dst_stride, 16)) {
|
| + ScaleARGBRowDown2 = filtering == kFilterNone ? ScaleARGBRowDown2_SSE2 :
|
| + (filtering == kFilterLinear ? ScaleARGBRowDown2Linear_SSE2 :
|
| + ScaleARGBRowDown2Box_SSE2);
|
| + }
|
| +#elif defined(HAS_SCALEARGBROWDOWN2_NEON)
|
| + if (TestCpuFlag(kCpuHasNEON) && IS_ALIGNED(dst_width, 8) &&
|
| + IS_ALIGNED(src_argb, 4) && IS_ALIGNED(row_stride, 4)) {
|
| + ScaleARGBRowDown2 = filtering ? ScaleARGBRowDown2Box_NEON :
|
| + ScaleARGBRowDown2_NEON;
|
| + }
|
| +#endif
|
| +
|
| + if (filtering == kFilterLinear) {
|
| + src_stride = 0;
|
| + }
|
| + for (j = 0; j < dst_height; ++j) {
|
| + ScaleARGBRowDown2(src_argb, src_stride, dst_argb, dst_width);
|
| + src_argb += row_stride;
|
| + dst_argb += dst_stride;
|
| + }
|
| +}
|
| +
|
| +// ScaleARGB ARGB, 1/4
|
| +// This is an optimized version for scaling down a ARGB to 1/4 of
|
| +// its original size.
|
| +static void ScaleARGBDown4Box(int src_width, int src_height,
|
| + int dst_width, int dst_height,
|
| + int src_stride, int dst_stride,
|
| + const uint8* src_argb, uint8* dst_argb,
|
| + int x, int dx, int y, int dy) {
|
| + int j;
|
| + // Allocate 2 rows of ARGB.
|
| + const int kRowSize = (dst_width * 2 * 4 + 15) & ~15;
|
| + align_buffer_64(row, kRowSize * 2);
|
| + int row_stride = src_stride * (dy >> 16);
|
| + void (*ScaleARGBRowDown2)(const uint8* src_argb, ptrdiff_t src_stride,
|
| + uint8* dst_argb, int dst_width) = ScaleARGBRowDown2Box_C;
|
| + // Advance to odd row, even column.
|
| + src_argb += (y >> 16) * src_stride + (x >> 16) * 4;
|
| + assert(dx == 65536 * 4); // Test scale factor of 4.
|
| + assert((dy & 0x3ffff) == 0); // Test vertical scale is multiple of 4.
|
| +#if defined(HAS_SCALEARGBROWDOWN2_SSE2)
|
| + if (TestCpuFlag(kCpuHasSSE2) && IS_ALIGNED(dst_width, 4) &&
|
| + IS_ALIGNED(src_argb, 16) && IS_ALIGNED(row_stride, 16) &&
|
| + IS_ALIGNED(dst_argb, 16) && IS_ALIGNED(dst_stride, 16)) {
|
| + ScaleARGBRowDown2 = ScaleARGBRowDown2Box_SSE2;
|
| + }
|
| +#elif defined(HAS_SCALEARGBROWDOWN2_NEON)
|
| + if (TestCpuFlag(kCpuHasNEON) && IS_ALIGNED(dst_width, 8) &&
|
| + IS_ALIGNED(src_argb, 4) && IS_ALIGNED(row_stride, 4)) {
|
| + ScaleARGBRowDown2 = ScaleARGBRowDown2Box_NEON;
|
| + }
|
| +#endif
|
| + for (j = 0; j < dst_height; ++j) {
|
| + ScaleARGBRowDown2(src_argb, src_stride, row, dst_width * 2);
|
| + ScaleARGBRowDown2(src_argb + src_stride * 2, src_stride,
|
| + row + kRowSize, dst_width * 2);
|
| + ScaleARGBRowDown2(row, kRowSize, dst_argb, dst_width);
|
| + src_argb += row_stride;
|
| + dst_argb += dst_stride;
|
| + }
|
| + free_aligned_buffer_64(row);
|
| +}
|
| +
|
| +// ScaleARGB ARGB Even
|
| +// This is an optimized version for scaling down a ARGB to even
|
| +// multiple of its original size.
|
| +static void ScaleARGBDownEven(int src_width, int src_height,
|
| + int dst_width, int dst_height,
|
| + int src_stride, int dst_stride,
|
| + const uint8* src_argb, uint8* dst_argb,
|
| + int x, int dx, int y, int dy,
|
| + enum FilterMode filtering) {
|
| + int j;
|
| + int col_step = dx >> 16;
|
| + int row_stride = (dy >> 16) * src_stride;
|
| + void (*ScaleARGBRowDownEven)(const uint8* src_argb, ptrdiff_t src_stride,
|
| + int src_step, uint8* dst_argb, int dst_width) =
|
| + filtering ? ScaleARGBRowDownEvenBox_C : ScaleARGBRowDownEven_C;
|
| + assert(IS_ALIGNED(src_width, 2));
|
| + assert(IS_ALIGNED(src_height, 2));
|
| + src_argb += (y >> 16) * src_stride + (x >> 16) * 4;
|
| +#if defined(HAS_SCALEARGBROWDOWNEVEN_SSE2)
|
| + if (TestCpuFlag(kCpuHasSSE2) && IS_ALIGNED(dst_width, 4) &&
|
| + IS_ALIGNED(dst_argb, 16) && IS_ALIGNED(dst_stride, 16)) {
|
| + ScaleARGBRowDownEven = filtering ? ScaleARGBRowDownEvenBox_SSE2 :
|
| + ScaleARGBRowDownEven_SSE2;
|
| + }
|
| +#elif defined(HAS_SCALEARGBROWDOWNEVEN_NEON)
|
| + if (TestCpuFlag(kCpuHasNEON) && IS_ALIGNED(dst_width, 4) &&
|
| + IS_ALIGNED(src_argb, 4)) {
|
| + ScaleARGBRowDownEven = filtering ? ScaleARGBRowDownEvenBox_NEON :
|
| + ScaleARGBRowDownEven_NEON;
|
| + }
|
| +#endif
|
| +
|
| + if (filtering == kFilterLinear) {
|
| + src_stride = 0;
|
| + }
|
| + for (j = 0; j < dst_height; ++j) {
|
| + ScaleARGBRowDownEven(src_argb, src_stride, col_step, dst_argb, dst_width);
|
| + src_argb += row_stride;
|
| + dst_argb += dst_stride;
|
| + }
|
| +}
|
| +
|
| +// Scale ARGB down with bilinear interpolation.
|
| +static void ScaleARGBBilinearDown(int src_width, int src_height,
|
| + int dst_width, int dst_height,
|
| + int src_stride, int dst_stride,
|
| + const uint8* src_argb, uint8* dst_argb,
|
| + int x, int dx, int y, int dy,
|
| + enum FilterMode filtering) {
|
| + int j;
|
| + void (*InterpolateRow)(uint8* dst_argb, const uint8* src_argb,
|
| + ptrdiff_t src_stride, int dst_width, int source_y_fraction) =
|
| + InterpolateRow_C;
|
| + void (*ScaleARGBFilterCols)(uint8* dst_argb, const uint8* src_argb,
|
| + int dst_width, int x, int dx) =
|
| + (src_width >= 32768) ? ScaleARGBFilterCols64_C : ScaleARGBFilterCols_C;
|
| + int64 xlast = x + (int64)(dst_width - 1) * dx;
|
| + int64 xl = (dx >= 0) ? x : xlast;
|
| + int64 xr = (dx >= 0) ? xlast : x;
|
| + int clip_src_width;
|
| + xl = (xl >> 16) & ~3; // Left edge aligned.
|
| + xr = (xr >> 16) + 1; // Right most pixel used. Bilinear uses 2 pixels.
|
| + xr = (xr + 1 + 3) & ~3; // 1 beyond 4 pixel aligned right most pixel.
|
| + if (xr > src_width) {
|
| + xr = src_width;
|
| + }
|
| + clip_src_width = (int)(xr - xl) * 4; // Width aligned to 4.
|
| + src_argb += xl * 4;
|
| + x -= (int)(xl << 16);
|
| +#if defined(HAS_INTERPOLATEROW_SSE2)
|
| + if (TestCpuFlag(kCpuHasSSE2) && clip_src_width >= 16) {
|
| + InterpolateRow = InterpolateRow_Any_SSE2;
|
| + if (IS_ALIGNED(clip_src_width, 16)) {
|
| + InterpolateRow = InterpolateRow_Unaligned_SSE2;
|
| + if (IS_ALIGNED(src_argb, 16) && IS_ALIGNED(src_stride, 16)) {
|
| + InterpolateRow = InterpolateRow_SSE2;
|
| + }
|
| + }
|
| + }
|
| +#endif
|
| +#if defined(HAS_INTERPOLATEROW_SSSE3)
|
| + if (TestCpuFlag(kCpuHasSSSE3) && clip_src_width >= 16) {
|
| + InterpolateRow = InterpolateRow_Any_SSSE3;
|
| + if (IS_ALIGNED(clip_src_width, 16)) {
|
| + InterpolateRow = InterpolateRow_Unaligned_SSSE3;
|
| + if (IS_ALIGNED(src_argb, 16) && IS_ALIGNED(src_stride, 16)) {
|
| + InterpolateRow = InterpolateRow_SSSE3;
|
| + }
|
| + }
|
| + }
|
| +#endif
|
| +#if defined(HAS_INTERPOLATEROW_AVX2)
|
| + if (TestCpuFlag(kCpuHasAVX2) && clip_src_width >= 32) {
|
| + InterpolateRow = InterpolateRow_Any_AVX2;
|
| + if (IS_ALIGNED(clip_src_width, 32)) {
|
| + InterpolateRow = InterpolateRow_AVX2;
|
| + }
|
| + }
|
| +#endif
|
| +#if defined(HAS_INTERPOLATEROW_NEON)
|
| + if (TestCpuFlag(kCpuHasNEON) && clip_src_width >= 16) {
|
| + InterpolateRow = InterpolateRow_Any_NEON;
|
| + if (IS_ALIGNED(clip_src_width, 16)) {
|
| + InterpolateRow = InterpolateRow_NEON;
|
| + }
|
| + }
|
| +#endif
|
| +#if defined(HAS_INTERPOLATEROWS_MIPS_DSPR2)
|
| + if (TestCpuFlag(kCpuHasMIPS_DSPR2) && clip_src_width >= 4 &&
|
| + IS_ALIGNED(src_argb, 4) && IS_ALIGNED(src_stride, 4)) {
|
| + InterpolateRow = InterpolateRow_Any_MIPS_DSPR2;
|
| + if (IS_ALIGNED(clip_src_width, 4)) {
|
| + InterpolateRow = InterpolateRow_MIPS_DSPR2;
|
| + }
|
| + }
|
| +#endif
|
| +#if defined(HAS_SCALEARGBFILTERCOLS_SSSE3)
|
| + if (TestCpuFlag(kCpuHasSSSE3) && src_width < 32768) {
|
| + ScaleARGBFilterCols = ScaleARGBFilterCols_SSSE3;
|
| + }
|
| +#endif
|
| + // TODO(fbarchard): Consider not allocating row buffer for kFilterLinear.
|
| + // Allocate a row of ARGB.
|
| + {
|
| + align_buffer_64(row, clip_src_width * 4);
|
| +
|
| + const int max_y = (src_height - 1) << 16;
|
| + if (y > max_y) {
|
| + y = max_y;
|
| + }
|
| + for (j = 0; j < dst_height; ++j) {
|
| + int yi = y >> 16;
|
| + const uint8* src = src_argb + yi * src_stride;
|
| + if (filtering == kFilterLinear) {
|
| + ScaleARGBFilterCols(dst_argb, src, dst_width, x, dx);
|
| + } else {
|
| + int yf = (y >> 8) & 255;
|
| + InterpolateRow(row, src, src_stride, clip_src_width, yf);
|
| + ScaleARGBFilterCols(dst_argb, row, dst_width, x, dx);
|
| + }
|
| + dst_argb += dst_stride;
|
| + y += dy;
|
| + if (y > max_y) {
|
| + y = max_y;
|
| + }
|
| + }
|
| + free_aligned_buffer_64(row);
|
| + }
|
| +}
|
| +
|
| +// Scale ARGB up with bilinear interpolation.
|
| +static void ScaleARGBBilinearUp(int src_width, int src_height,
|
| + int dst_width, int dst_height,
|
| + int src_stride, int dst_stride,
|
| + const uint8* src_argb, uint8* dst_argb,
|
| + int x, int dx, int y, int dy,
|
| + enum FilterMode filtering) {
|
| + int j;
|
| + void (*InterpolateRow)(uint8* dst_argb, const uint8* src_argb,
|
| + ptrdiff_t src_stride, int dst_width, int source_y_fraction) =
|
| + InterpolateRow_C;
|
| + void (*ScaleARGBFilterCols)(uint8* dst_argb, const uint8* src_argb,
|
| + int dst_width, int x, int dx) =
|
| + filtering ? ScaleARGBFilterCols_C : ScaleARGBCols_C;
|
| + const int max_y = (src_height - 1) << 16;
|
| +#if defined(HAS_INTERPOLATEROW_SSE2)
|
| + if (TestCpuFlag(kCpuHasSSE2) && dst_width >= 4) {
|
| + InterpolateRow = InterpolateRow_Any_SSE2;
|
| + if (IS_ALIGNED(dst_width, 4)) {
|
| + InterpolateRow = InterpolateRow_Unaligned_SSE2;
|
| + if (IS_ALIGNED(dst_argb, 16) && IS_ALIGNED(dst_stride, 16)) {
|
| + InterpolateRow = InterpolateRow_SSE2;
|
| + }
|
| + }
|
| + }
|
| +#endif
|
| +#if defined(HAS_INTERPOLATEROW_SSSE3)
|
| + if (TestCpuFlag(kCpuHasSSSE3) && dst_width >= 4) {
|
| + InterpolateRow = InterpolateRow_Any_SSSE3;
|
| + if (IS_ALIGNED(dst_width, 4)) {
|
| + InterpolateRow = InterpolateRow_Unaligned_SSSE3;
|
| + if (IS_ALIGNED(dst_argb, 16) && IS_ALIGNED(dst_stride, 16)) {
|
| + InterpolateRow = InterpolateRow_SSSE3;
|
| + }
|
| + }
|
| + }
|
| +#endif
|
| +#if defined(HAS_INTERPOLATEROW_AVX2)
|
| + if (TestCpuFlag(kCpuHasAVX2) && dst_width >= 8) {
|
| + InterpolateRow = InterpolateRow_Any_AVX2;
|
| + if (IS_ALIGNED(dst_width, 8)) {
|
| + InterpolateRow = InterpolateRow_AVX2;
|
| + }
|
| + }
|
| +#endif
|
| +#if defined(HAS_INTERPOLATEROW_NEON)
|
| + if (TestCpuFlag(kCpuHasNEON) && dst_width >= 4) {
|
| + InterpolateRow = InterpolateRow_Any_NEON;
|
| + if (IS_ALIGNED(dst_width, 4)) {
|
| + InterpolateRow = InterpolateRow_NEON;
|
| + }
|
| + }
|
| +#endif
|
| +#if defined(HAS_INTERPOLATEROWS_MIPS_DSPR2)
|
| + if (TestCpuFlag(kCpuHasMIPS_DSPR2) && dst_width >= 1 &&
|
| + IS_ALIGNED(dst_argb, 4) && IS_ALIGNED(dst_stride, 4)) {
|
| + InterpolateRow = InterpolateRow_MIPS_DSPR2;
|
| + }
|
| +#endif
|
| + if (src_width >= 32768) {
|
| + ScaleARGBFilterCols = filtering ?
|
| + ScaleARGBFilterCols64_C : ScaleARGBCols64_C;
|
| + }
|
| +#if defined(HAS_SCALEARGBFILTERCOLS_SSSE3)
|
| + if (filtering && TestCpuFlag(kCpuHasSSSE3) && src_width < 32768) {
|
| + ScaleARGBFilterCols = ScaleARGBFilterCols_SSSE3;
|
| + }
|
| +#endif
|
| +#if defined(HAS_SCALEARGBCOLS_SSE2)
|
| + if (!filtering && TestCpuFlag(kCpuHasSSE2) && src_width < 32768) {
|
| + ScaleARGBFilterCols = ScaleARGBCols_SSE2;
|
| + }
|
| +#endif
|
| + if (!filtering && src_width * 2 == dst_width && x < 0x8000) {
|
| + ScaleARGBFilterCols = ScaleARGBColsUp2_C;
|
| +#if defined(HAS_SCALEARGBCOLSUP2_SSE2)
|
| + if (TestCpuFlag(kCpuHasSSE2) && IS_ALIGNED(dst_width, 8) &&
|
| + IS_ALIGNED(src_argb, 16) && IS_ALIGNED(src_stride, 16) &&
|
| + IS_ALIGNED(dst_argb, 16) && IS_ALIGNED(dst_stride, 16)) {
|
| + ScaleARGBFilterCols = ScaleARGBColsUp2_SSE2;
|
| + }
|
| +#endif
|
| + }
|
| +
|
| + if (y > max_y) {
|
| + y = max_y;
|
| + }
|
| +
|
| + {
|
| + int yi = y >> 16;
|
| + const uint8* src = src_argb + yi * src_stride;
|
| +
|
| + // Allocate 2 rows of ARGB.
|
| + const int kRowSize = (dst_width * 4 + 15) & ~15;
|
| + align_buffer_64(row, kRowSize * 2);
|
| +
|
| + uint8* rowptr = row;
|
| + int rowstride = kRowSize;
|
| + int lasty = yi;
|
| +
|
| + ScaleARGBFilterCols(rowptr, src, dst_width, x, dx);
|
| + if (src_height > 1) {
|
| + src += src_stride;
|
| + }
|
| + ScaleARGBFilterCols(rowptr + rowstride, src, dst_width, x, dx);
|
| + src += src_stride;
|
| +
|
| + for (j = 0; j < dst_height; ++j) {
|
| + yi = y >> 16;
|
| + if (yi != lasty) {
|
| + if (y > max_y) {
|
| + y = max_y;
|
| + yi = y >> 16;
|
| + src = src_argb + yi * src_stride;
|
| + }
|
| + if (yi != lasty) {
|
| + ScaleARGBFilterCols(rowptr, src, dst_width, x, dx);
|
| + rowptr += rowstride;
|
| + rowstride = -rowstride;
|
| + lasty = yi;
|
| + src += src_stride;
|
| + }
|
| + }
|
| + if (filtering == kFilterLinear) {
|
| + InterpolateRow(dst_argb, rowptr, 0, dst_width * 4, 0);
|
| + } else {
|
| + int yf = (y >> 8) & 255;
|
| + InterpolateRow(dst_argb, rowptr, rowstride, dst_width * 4, yf);
|
| + }
|
| + dst_argb += dst_stride;
|
| + y += dy;
|
| + }
|
| + free_aligned_buffer_64(row);
|
| + }
|
| +}
|
| +
|
| +#ifdef YUVSCALEUP
|
| +// Scale YUV to ARGB up with bilinear interpolation.
|
| +static void ScaleYUVToARGBBilinearUp(int src_width, int src_height,
|
| + int dst_width, int dst_height,
|
| + int src_stride_y,
|
| + int src_stride_u,
|
| + int src_stride_v,
|
| + int dst_stride_argb,
|
| + const uint8* src_y,
|
| + const uint8* src_u,
|
| + const uint8* src_v,
|
| + uint8* dst_argb,
|
| + int x, int dx, int y, int dy,
|
| + enum FilterMode filtering) {
|
| + int j;
|
| + void (*I422ToARGBRow)(const uint8* y_buf,
|
| + const uint8* u_buf,
|
| + const uint8* v_buf,
|
| + uint8* rgb_buf,
|
| + int width) = I422ToARGBRow_C;
|
| +#if defined(HAS_I422TOARGBROW_SSSE3)
|
| + if (TestCpuFlag(kCpuHasSSSE3) && src_width >= 8) {
|
| + I422ToARGBRow = I422ToARGBRow_Any_SSSE3;
|
| + if (IS_ALIGNED(src_width, 8)) {
|
| + I422ToARGBRow = I422ToARGBRow_Unaligned_SSSE3;
|
| + if (IS_ALIGNED(dst_argb, 16) && IS_ALIGNED(dst_stride_argb, 16)) {
|
| + I422ToARGBRow = I422ToARGBRow_SSSE3;
|
| + }
|
| + }
|
| + }
|
| +#endif
|
| +#if defined(HAS_I422TOARGBROW_AVX2)
|
| + if (TestCpuFlag(kCpuHasAVX2) && src_width >= 16) {
|
| + I422ToARGBRow = I422ToARGBRow_Any_AVX2;
|
| + if (IS_ALIGNED(src_width, 16)) {
|
| + I422ToARGBRow = I422ToARGBRow_AVX2;
|
| + }
|
| + }
|
| +#endif
|
| +#if defined(HAS_I422TOARGBROW_NEON)
|
| + if (TestCpuFlag(kCpuHasNEON) && src_width >= 8) {
|
| + I422ToARGBRow = I422ToARGBRow_Any_NEON;
|
| + if (IS_ALIGNED(src_width, 8)) {
|
| + I422ToARGBRow = I422ToARGBRow_NEON;
|
| + }
|
| + }
|
| +#endif
|
| +#if defined(HAS_I422TOARGBROW_MIPS_DSPR2)
|
| + if (TestCpuFlag(kCpuHasMIPS_DSPR2) && IS_ALIGNED(src_width, 4) &&
|
| + IS_ALIGNED(src_y, 4) && IS_ALIGNED(src_stride_y, 4) &&
|
| + IS_ALIGNED(src_u, 2) && IS_ALIGNED(src_stride_u, 2) &&
|
| + IS_ALIGNED(src_v, 2) && IS_ALIGNED(src_stride_v, 2) &&
|
| + IS_ALIGNED(dst_argb, 4) && IS_ALIGNED(dst_stride_argb, 4)) {
|
| + I422ToARGBRow = I422ToARGBRow_MIPS_DSPR2;
|
| + }
|
| +#endif
|
| +
|
| + void (*InterpolateRow)(uint8* dst_argb, const uint8* src_argb,
|
| + ptrdiff_t src_stride, int dst_width, int source_y_fraction) =
|
| + InterpolateRow_C;
|
| +#if defined(HAS_INTERPOLATEROW_SSE2)
|
| + if (TestCpuFlag(kCpuHasSSE2) && dst_width >= 4) {
|
| + InterpolateRow = InterpolateRow_Any_SSE2;
|
| + if (IS_ALIGNED(dst_width, 4)) {
|
| + InterpolateRow = InterpolateRow_Unaligned_SSE2;
|
| + if (IS_ALIGNED(dst_argb, 16) && IS_ALIGNED(dst_stride_argb, 16)) {
|
| + InterpolateRow = InterpolateRow_SSE2;
|
| + }
|
| + }
|
| + }
|
| +#endif
|
| +#if defined(HAS_INTERPOLATEROW_SSSE3)
|
| + if (TestCpuFlag(kCpuHasSSSE3) && dst_width >= 4) {
|
| + InterpolateRow = InterpolateRow_Any_SSSE3;
|
| + if (IS_ALIGNED(dst_width, 4)) {
|
| + InterpolateRow = InterpolateRow_Unaligned_SSSE3;
|
| + if (IS_ALIGNED(dst_argb, 16) && IS_ALIGNED(dst_stride_argb, 16)) {
|
| + InterpolateRow = InterpolateRow_SSSE3;
|
| + }
|
| + }
|
| + }
|
| +#endif
|
| +#if defined(HAS_INTERPOLATEROW_AVX2)
|
| + if (TestCpuFlag(kCpuHasAVX2) && dst_width >= 8) {
|
| + InterpolateRow = InterpolateRow_Any_AVX2;
|
| + if (IS_ALIGNED(dst_width, 8)) {
|
| + InterpolateRow = InterpolateRow_AVX2;
|
| + }
|
| + }
|
| +#endif
|
| +#if defined(HAS_INTERPOLATEROW_NEON)
|
| + if (TestCpuFlag(kCpuHasNEON) && dst_width >= 4) {
|
| + InterpolateRow = InterpolateRow_Any_NEON;
|
| + if (IS_ALIGNED(dst_width, 4)) {
|
| + InterpolateRow = InterpolateRow_NEON;
|
| + }
|
| + }
|
| +#endif
|
| +#if defined(HAS_INTERPOLATEROWS_MIPS_DSPR2)
|
| + if (TestCpuFlag(kCpuHasMIPS_DSPR2) && dst_width >= 1 &&
|
| + IS_ALIGNED(dst_argb, 4) && IS_ALIGNED(dst_stride_argb, 4)) {
|
| + InterpolateRow = InterpolateRow_MIPS_DSPR2;
|
| + }
|
| +#endif
|
| +
|
| + void (*ScaleARGBFilterCols)(uint8* dst_argb, const uint8* src_argb,
|
| + int dst_width, int x, int dx) =
|
| + filtering ? ScaleARGBFilterCols_C : ScaleARGBCols_C;
|
| + if (src_width >= 32768) {
|
| + ScaleARGBFilterCols = filtering ?
|
| + ScaleARGBFilterCols64_C : ScaleARGBCols64_C;
|
| + }
|
| +#if defined(HAS_SCALEARGBFILTERCOLS_SSSE3)
|
| + if (filtering && TestCpuFlag(kCpuHasSSSE3) && src_width < 32768) {
|
| + ScaleARGBFilterCols = ScaleARGBFilterCols_SSSE3;
|
| + }
|
| +#endif
|
| +#if defined(HAS_SCALEARGBCOLS_SSE2)
|
| + if (!filtering && TestCpuFlag(kCpuHasSSE2) && src_width < 32768) {
|
| + ScaleARGBFilterCols = ScaleARGBCols_SSE2;
|
| + }
|
| +#endif
|
| + if (!filtering && src_width * 2 == dst_width && x < 0x8000) {
|
| + ScaleARGBFilterCols = ScaleARGBColsUp2_C;
|
| +#if defined(HAS_SCALEARGBCOLSUP2_SSE2)
|
| + if (TestCpuFlag(kCpuHasSSE2) && IS_ALIGNED(dst_width, 8) &&
|
| + IS_ALIGNED(src_argb, 16) && IS_ALIGNED(src_stride, 16) &&
|
| + IS_ALIGNED(dst_argb, 16) && IS_ALIGNED(dst_stride, 16)) {
|
| + ScaleARGBFilterCols = ScaleARGBColsUp2_SSE2;
|
| + }
|
| +#endif
|
| + }
|
| +
|
| + const int max_y = (src_height - 1) << 16;
|
| + if (y > max_y) {
|
| + y = max_y;
|
| + }
|
| + const int kYShift = 1; // Shift Y by 1 to convert Y plane to UV coordinate.
|
| + int yi = y >> 16;
|
| + int uv_yi = yi >> kYShift;
|
| + const uint8* src_row_y = src_y + yi * src_stride_y;
|
| + const uint8* src_row_u = src_u + uv_yi * src_stride_u;
|
| + const uint8* src_row_v = src_v + uv_yi * src_stride_v;
|
| +
|
| + // Allocate 2 rows of ARGB.
|
| + const int kRowSize = (dst_width * 4 + 15) & ~15;
|
| + align_buffer_64(row, kRowSize * 2);
|
| +
|
| + // Allocate 1 row of ARGB for source conversion.
|
| + align_buffer_64(argb_row, src_width * 4);
|
| +
|
| + uint8* rowptr = row;
|
| + int rowstride = kRowSize;
|
| + int lasty = yi;
|
| +
|
| + // TODO(fbarchard): Convert first 2 rows of YUV to ARGB.
|
| + ScaleARGBFilterCols(rowptr, src_row_y, dst_width, x, dx);
|
| + if (src_height > 1) {
|
| + src_row_y += src_stride_y;
|
| + if (yi & 1) {
|
| + src_row_u += src_stride_u;
|
| + src_row_v += src_stride_v;
|
| + }
|
| + }
|
| + ScaleARGBFilterCols(rowptr + rowstride, src_row_y, dst_width, x, dx);
|
| + if (src_height > 2) {
|
| + src_row_y += src_stride_y;
|
| + if (!(yi & 1)) {
|
| + src_row_u += src_stride_u;
|
| + src_row_v += src_stride_v;
|
| + }
|
| + }
|
| +
|
| + for (j = 0; j < dst_height; ++j) {
|
| + yi = y >> 16;
|
| + if (yi != lasty) {
|
| + if (y > max_y) {
|
| + y = max_y;
|
| + yi = y >> 16;
|
| + uv_yi = yi >> kYShift;
|
| + src_row_y = src_y + yi * src_stride_y;
|
| + src_row_u = src_u + uv_yi * src_stride_u;
|
| + src_row_v = src_v + uv_yi * src_stride_v;
|
| + }
|
| + if (yi != lasty) {
|
| + // TODO(fbarchard): Convert the clipped region of row.
|
| + I422ToARGBRow(src_row_y, src_row_u, src_row_v, argb_row, src_width);
|
| + ScaleARGBFilterCols(rowptr, argb_row, dst_width, x, dx);
|
| + rowptr += rowstride;
|
| + rowstride = -rowstride;
|
| + lasty = yi;
|
| + src_row_y += src_stride_y;
|
| + if (yi & 1) {
|
| + src_row_u += src_stride_u;
|
| + src_row_v += src_stride_v;
|
| + }
|
| + }
|
| + }
|
| + if (filtering == kFilterLinear) {
|
| + InterpolateRow(dst_argb, rowptr, 0, dst_width * 4, 0);
|
| + } else {
|
| + int yf = (y >> 8) & 255;
|
| + InterpolateRow(dst_argb, rowptr, rowstride, dst_width * 4, yf);
|
| + }
|
| + dst_argb += dst_stride_argb;
|
| + y += dy;
|
| + }
|
| + free_aligned_buffer_64(row);
|
| + free_aligned_buffer_64(row_argb);
|
| +}
|
| +#endif
|
| +
|
| +// Scale ARGB to/from any dimensions, without interpolation.
|
| +// Fixed point math is used for performance: The upper 16 bits
|
| +// of x and dx is the integer part of the source position and
|
| +// the lower 16 bits are the fixed decimal part.
|
| +
|
| +static void ScaleARGBSimple(int src_width, int src_height,
|
| + int dst_width, int dst_height,
|
| + int src_stride, int dst_stride,
|
| + const uint8* src_argb, uint8* dst_argb,
|
| + int x, int dx, int y, int dy) {
|
| + int j;
|
| + void (*ScaleARGBCols)(uint8* dst_argb, const uint8* src_argb,
|
| + int dst_width, int x, int dx) =
|
| + (src_width >= 32768) ? ScaleARGBCols64_C : ScaleARGBCols_C;
|
| +#if defined(HAS_SCALEARGBCOLS_SSE2)
|
| + if (TestCpuFlag(kCpuHasSSE2) && src_width < 32768) {
|
| + ScaleARGBCols = ScaleARGBCols_SSE2;
|
| + }
|
| +#endif
|
| + if (src_width * 2 == dst_width && x < 0x8000) {
|
| + ScaleARGBCols = ScaleARGBColsUp2_C;
|
| +#if defined(HAS_SCALEARGBCOLSUP2_SSE2)
|
| + if (TestCpuFlag(kCpuHasSSE2) && IS_ALIGNED(dst_width, 8) &&
|
| + IS_ALIGNED(src_argb, 16) && IS_ALIGNED(src_stride, 16) &&
|
| + IS_ALIGNED(dst_argb, 16) && IS_ALIGNED(dst_stride, 16)) {
|
| + ScaleARGBCols = ScaleARGBColsUp2_SSE2;
|
| + }
|
| +#endif
|
| + }
|
| +
|
| + for (j = 0; j < dst_height; ++j) {
|
| + ScaleARGBCols(dst_argb, src_argb + (y >> 16) * src_stride,
|
| + dst_width, x, dx);
|
| + dst_argb += dst_stride;
|
| + y += dy;
|
| + }
|
| +}
|
| +
|
| +// ScaleARGB a ARGB.
|
| +// This function in turn calls a scaling function
|
| +// suitable for handling the desired resolutions.
|
| +static void ScaleARGB(const uint8* src, int src_stride,
|
| + int src_width, int src_height,
|
| + uint8* dst, int dst_stride,
|
| + int dst_width, int dst_height,
|
| + int clip_x, int clip_y, int clip_width, int clip_height,
|
| + enum FilterMode filtering) {
|
| + // Initial source x/y coordinate and step values as 16.16 fixed point.
|
| + int x = 0;
|
| + int y = 0;
|
| + int dx = 0;
|
| + int dy = 0;
|
| + // ARGB does not support box filter yet, but allow the user to pass it.
|
| + // Simplify filtering when possible.
|
| + filtering = ScaleFilterReduce(src_width, src_height,
|
| + dst_width, dst_height,
|
| + filtering);
|
| +
|
| + // Negative src_height means invert the image.
|
| + if (src_height < 0) {
|
| + src_height = -src_height;
|
| + src = src + (src_height - 1) * src_stride;
|
| + src_stride = -src_stride;
|
| + }
|
| + ScaleSlope(src_width, src_height, dst_width, dst_height, filtering,
|
| + &x, &y, &dx, &dy);
|
| + src_width = Abs(src_width);
|
| + if (clip_x) {
|
| + int64 clipf = (int64)(clip_x) * dx;
|
| + x += (clipf & 0xffff);
|
| + src += (clipf >> 16) * 4;
|
| + dst += clip_x * 4;
|
| + }
|
| + if (clip_y) {
|
| + int64 clipf = (int64)(clip_y) * dy;
|
| + y += (clipf & 0xffff);
|
| + src += (clipf >> 16) * src_stride;
|
| + dst += clip_y * dst_stride;
|
| + }
|
| +
|
| + // Special case for integer step values.
|
| + if (((dx | dy) & 0xffff) == 0) {
|
| + if (!dx || !dy) { // 1 pixel wide and/or tall.
|
| + filtering = kFilterNone;
|
| + } else {
|
| + // Optimized even scale down. ie 2, 4, 6, 8, 10x.
|
| + if (!(dx & 0x10000) && !(dy & 0x10000)) {
|
| + if (dx == 0x20000) {
|
| + // Optimized 1/2 downsample.
|
| + ScaleARGBDown2(src_width, src_height,
|
| + clip_width, clip_height,
|
| + src_stride, dst_stride, src, dst,
|
| + x, dx, y, dy, filtering);
|
| + return;
|
| + }
|
| + if (dx == 0x40000 && filtering == kFilterBox) {
|
| + // Optimized 1/4 box downsample.
|
| + ScaleARGBDown4Box(src_width, src_height,
|
| + clip_width, clip_height,
|
| + src_stride, dst_stride, src, dst,
|
| + x, dx, y, dy);
|
| + return;
|
| + }
|
| + ScaleARGBDownEven(src_width, src_height,
|
| + clip_width, clip_height,
|
| + src_stride, dst_stride, src, dst,
|
| + x, dx, y, dy, filtering);
|
| + return;
|
| + }
|
| + // Optimized odd scale down. ie 3, 5, 7, 9x.
|
| + if ((dx & 0x10000) && (dy & 0x10000)) {
|
| + filtering = kFilterNone;
|
| + if (dx == 0x10000 && dy == 0x10000) {
|
| + // Straight copy.
|
| + ARGBCopy(src + (y >> 16) * src_stride + (x >> 16) * 4, src_stride,
|
| + dst, dst_stride, clip_width, clip_height);
|
| + return;
|
| + }
|
| + }
|
| + }
|
| + }
|
| + if (dx == 0x10000 && (x & 0xffff) == 0) {
|
| + // Arbitrary scale vertically, but unscaled vertically.
|
| + ScalePlaneVertical(src_height,
|
| + clip_width, clip_height,
|
| + src_stride, dst_stride, src, dst,
|
| + x, y, dy, 4, filtering);
|
| + return;
|
| + }
|
| + if (filtering && dy < 65536) {
|
| + ScaleARGBBilinearUp(src_width, src_height,
|
| + clip_width, clip_height,
|
| + src_stride, dst_stride, src, dst,
|
| + x, dx, y, dy, filtering);
|
| + return;
|
| + }
|
| + if (filtering) {
|
| + ScaleARGBBilinearDown(src_width, src_height,
|
| + clip_width, clip_height,
|
| + src_stride, dst_stride, src, dst,
|
| + x, dx, y, dy, filtering);
|
| + return;
|
| + }
|
| + ScaleARGBSimple(src_width, src_height, clip_width, clip_height,
|
| + src_stride, dst_stride, src, dst,
|
| + x, dx, y, dy);
|
| +}
|
| +
|
| +LIBYUV_API
|
| +int ARGBScaleClip(const uint8* src_argb, int src_stride_argb,
|
| + int src_width, int src_height,
|
| + uint8* dst_argb, int dst_stride_argb,
|
| + int dst_width, int dst_height,
|
| + int clip_x, int clip_y, int clip_width, int clip_height,
|
| + enum FilterMode filtering) {
|
| + if (!src_argb || src_width == 0 || src_height == 0 ||
|
| + !dst_argb || dst_width <= 0 || dst_height <= 0 ||
|
| + clip_x < 0 || clip_y < 0 ||
|
| + (clip_x + clip_width) > dst_width ||
|
| + (clip_y + clip_height) > dst_height) {
|
| + return -1;
|
| + }
|
| + ScaleARGB(src_argb, src_stride_argb, src_width, src_height,
|
| + dst_argb, dst_stride_argb, dst_width, dst_height,
|
| + clip_x, clip_y, clip_width, clip_height, filtering);
|
| + return 0;
|
| +}
|
| +
|
| +// Scale an ARGB image.
|
| +LIBYUV_API
|
| +int ARGBScale(const uint8* src_argb, int src_stride_argb,
|
| + int src_width, int src_height,
|
| + uint8* dst_argb, int dst_stride_argb,
|
| + int dst_width, int dst_height,
|
| + enum FilterMode filtering) {
|
| + if (!src_argb || src_width == 0 || src_height == 0 ||
|
| + !dst_argb || dst_width <= 0 || dst_height <= 0) {
|
| + return -1;
|
| + }
|
| + ScaleARGB(src_argb, src_stride_argb, src_width, src_height,
|
| + dst_argb, dst_stride_argb, dst_width, dst_height,
|
| + 0, 0, dst_width, dst_height, filtering);
|
| + return 0;
|
| +}
|
| +
|
| +#ifdef __cplusplus
|
| +} // extern "C"
|
| +} // namespace libyuv
|
| +#endif
|
|
|