| Index: source/libvpx/third_party/libyuv/source/compare.cc
|
| ===================================================================
|
| --- source/libvpx/third_party/libyuv/source/compare.cc (revision 0)
|
| +++ source/libvpx/third_party/libyuv/source/compare.cc (revision 0)
|
| @@ -0,0 +1,325 @@
|
| +/*
|
| + * Copyright 2011 The LibYuv Project Authors. All rights reserved.
|
| + *
|
| + * Use of this source code is governed by a BSD-style license
|
| + * that can be found in the LICENSE file in the root of the source
|
| + * tree. An additional intellectual property rights grant can be found
|
| + * in the file PATENTS. All contributing project authors may
|
| + * be found in the AUTHORS file in the root of the source tree.
|
| + */
|
| +
|
| +#include "libyuv/compare.h"
|
| +
|
| +#include <float.h>
|
| +#include <math.h>
|
| +#ifdef _OPENMP
|
| +#include <omp.h>
|
| +#endif
|
| +
|
| +#include "libyuv/basic_types.h"
|
| +#include "libyuv/cpu_id.h"
|
| +#include "libyuv/row.h"
|
| +
|
| +#ifdef __cplusplus
|
| +namespace libyuv {
|
| +extern "C" {
|
| +#endif
|
| +
|
| +// hash seed of 5381 recommended.
|
| +// Internal C version of HashDjb2 with int sized count for efficiency.
|
| +uint32 HashDjb2_C(const uint8* src, int count, uint32 seed);
|
| +
|
| +// This module is for Visual C x86
|
| +#if !defined(LIBYUV_DISABLE_X86) && \
|
| + (defined(_M_IX86) || \
|
| + (defined(__x86_64__) || (defined(__i386__) && !defined(__pic__))))
|
| +#define HAS_HASHDJB2_SSE41
|
| +uint32 HashDjb2_SSE41(const uint8* src, int count, uint32 seed);
|
| +
|
| +#if _MSC_VER >= 1700
|
| +#define HAS_HASHDJB2_AVX2
|
| +uint32 HashDjb2_AVX2(const uint8* src, int count, uint32 seed);
|
| +#endif
|
| +
|
| +#endif // HAS_HASHDJB2_SSE41
|
| +
|
| +// hash seed of 5381 recommended.
|
| +LIBYUV_API
|
| +uint32 HashDjb2(const uint8* src, uint64 count, uint32 seed) {
|
| + const int kBlockSize = 1 << 15; // 32768;
|
| + int remainder;
|
| + uint32 (*HashDjb2_SSE)(const uint8* src, int count, uint32 seed) = HashDjb2_C;
|
| +#if defined(HAS_HASHDJB2_SSE41)
|
| + if (TestCpuFlag(kCpuHasSSE41)) {
|
| + HashDjb2_SSE = HashDjb2_SSE41;
|
| + }
|
| +#endif
|
| +#if defined(HAS_HASHDJB2_AVX2)
|
| + if (TestCpuFlag(kCpuHasAVX2)) {
|
| + HashDjb2_SSE = HashDjb2_AVX2;
|
| + }
|
| +#endif
|
| +
|
| + while (count >= (uint64)(kBlockSize)) {
|
| + seed = HashDjb2_SSE(src, kBlockSize, seed);
|
| + src += kBlockSize;
|
| + count -= kBlockSize;
|
| + }
|
| + remainder = (int)(count) & ~15;
|
| + if (remainder) {
|
| + seed = HashDjb2_SSE(src, remainder, seed);
|
| + src += remainder;
|
| + count -= remainder;
|
| + }
|
| + remainder = (int)(count) & 15;
|
| + if (remainder) {
|
| + seed = HashDjb2_C(src, remainder, seed);
|
| + }
|
| + return seed;
|
| +}
|
| +
|
| +uint32 SumSquareError_C(const uint8* src_a, const uint8* src_b, int count);
|
| +#if !defined(LIBYUV_DISABLE_NEON) && \
|
| + (defined(__ARM_NEON__) || defined(LIBYUV_NEON))
|
| +#define HAS_SUMSQUAREERROR_NEON
|
| +uint32 SumSquareError_NEON(const uint8* src_a, const uint8* src_b, int count);
|
| +#endif
|
| +#if !defined(LIBYUV_DISABLE_X86) && \
|
| + (defined(_M_IX86) || defined(__x86_64__) || defined(__i386__))
|
| +#define HAS_SUMSQUAREERROR_SSE2
|
| +uint32 SumSquareError_SSE2(const uint8* src_a, const uint8* src_b, int count);
|
| +#endif
|
| +// Visual C 2012 required for AVX2.
|
| +#if !defined(LIBYUV_DISABLE_X86) && defined(_M_IX86) && _MSC_VER >= 1700
|
| +#define HAS_SUMSQUAREERROR_AVX2
|
| +uint32 SumSquareError_AVX2(const uint8* src_a, const uint8* src_b, int count);
|
| +#endif
|
| +
|
| +// TODO(fbarchard): Refactor into row function.
|
| +LIBYUV_API
|
| +uint64 ComputeSumSquareError(const uint8* src_a, const uint8* src_b,
|
| + int count) {
|
| + // SumSquareError returns values 0 to 65535 for each squared difference.
|
| + // Up to 65536 of those can be summed and remain within a uint32.
|
| + // After each block of 65536 pixels, accumulate into a uint64.
|
| + const int kBlockSize = 65536;
|
| + int remainder = count & (kBlockSize - 1) & ~31;
|
| + uint64 sse = 0;
|
| + int i;
|
| + uint32 (*SumSquareError)(const uint8* src_a, const uint8* src_b, int count) =
|
| + SumSquareError_C;
|
| +#if defined(HAS_SUMSQUAREERROR_NEON)
|
| + if (TestCpuFlag(kCpuHasNEON)) {
|
| + SumSquareError = SumSquareError_NEON;
|
| + }
|
| +#endif
|
| +#if defined(HAS_SUMSQUAREERROR_SSE2)
|
| + if (TestCpuFlag(kCpuHasSSE2) &&
|
| + IS_ALIGNED(src_a, 16) && IS_ALIGNED(src_b, 16)) {
|
| + // Note only used for multiples of 16 so count is not checked.
|
| + SumSquareError = SumSquareError_SSE2;
|
| + }
|
| +#endif
|
| +#if defined(HAS_SUMSQUAREERROR_AVX2)
|
| + if (TestCpuFlag(kCpuHasAVX2)) {
|
| + // Note only used for multiples of 32 so count is not checked.
|
| + SumSquareError = SumSquareError_AVX2;
|
| + }
|
| +#endif
|
| +#ifdef _OPENMP
|
| +#pragma omp parallel for reduction(+: sse)
|
| +#endif
|
| + for (i = 0; i < (count - (kBlockSize - 1)); i += kBlockSize) {
|
| + sse += SumSquareError(src_a + i, src_b + i, kBlockSize);
|
| + }
|
| + src_a += count & ~(kBlockSize - 1);
|
| + src_b += count & ~(kBlockSize - 1);
|
| + if (remainder) {
|
| + sse += SumSquareError(src_a, src_b, remainder);
|
| + src_a += remainder;
|
| + src_b += remainder;
|
| + }
|
| + remainder = count & 31;
|
| + if (remainder) {
|
| + sse += SumSquareError_C(src_a, src_b, remainder);
|
| + }
|
| + return sse;
|
| +}
|
| +
|
| +LIBYUV_API
|
| +uint64 ComputeSumSquareErrorPlane(const uint8* src_a, int stride_a,
|
| + const uint8* src_b, int stride_b,
|
| + int width, int height) {
|
| + uint64 sse = 0;
|
| + int h;
|
| + // Coalesce rows.
|
| + if (stride_a == width &&
|
| + stride_b == width) {
|
| + width *= height;
|
| + height = 1;
|
| + stride_a = stride_b = 0;
|
| + }
|
| + for (h = 0; h < height; ++h) {
|
| + sse += ComputeSumSquareError(src_a, src_b, width);
|
| + src_a += stride_a;
|
| + src_b += stride_b;
|
| + }
|
| + return sse;
|
| +}
|
| +
|
| +LIBYUV_API
|
| +double SumSquareErrorToPsnr(uint64 sse, uint64 count) {
|
| + double psnr;
|
| + if (sse > 0) {
|
| + double mse = (double)(count) / (double)(sse);
|
| + psnr = 10.0 * log10(255.0 * 255.0 * mse);
|
| + } else {
|
| + psnr = kMaxPsnr; // Limit to prevent divide by 0
|
| + }
|
| +
|
| + if (psnr > kMaxPsnr)
|
| + psnr = kMaxPsnr;
|
| +
|
| + return psnr;
|
| +}
|
| +
|
| +LIBYUV_API
|
| +double CalcFramePsnr(const uint8* src_a, int stride_a,
|
| + const uint8* src_b, int stride_b,
|
| + int width, int height) {
|
| + const uint64 samples = width * height;
|
| + const uint64 sse = ComputeSumSquareErrorPlane(src_a, stride_a,
|
| + src_b, stride_b,
|
| + width, height);
|
| + return SumSquareErrorToPsnr(sse, samples);
|
| +}
|
| +
|
| +LIBYUV_API
|
| +double I420Psnr(const uint8* src_y_a, int stride_y_a,
|
| + const uint8* src_u_a, int stride_u_a,
|
| + const uint8* src_v_a, int stride_v_a,
|
| + const uint8* src_y_b, int stride_y_b,
|
| + const uint8* src_u_b, int stride_u_b,
|
| + const uint8* src_v_b, int stride_v_b,
|
| + int width, int height) {
|
| + const uint64 sse_y = ComputeSumSquareErrorPlane(src_y_a, stride_y_a,
|
| + src_y_b, stride_y_b,
|
| + width, height);
|
| + const int width_uv = (width + 1) >> 1;
|
| + const int height_uv = (height + 1) >> 1;
|
| + const uint64 sse_u = ComputeSumSquareErrorPlane(src_u_a, stride_u_a,
|
| + src_u_b, stride_u_b,
|
| + width_uv, height_uv);
|
| + const uint64 sse_v = ComputeSumSquareErrorPlane(src_v_a, stride_v_a,
|
| + src_v_b, stride_v_b,
|
| + width_uv, height_uv);
|
| + const uint64 samples = width * height + 2 * (width_uv * height_uv);
|
| + const uint64 sse = sse_y + sse_u + sse_v;
|
| + return SumSquareErrorToPsnr(sse, samples);
|
| +}
|
| +
|
| +static const int64 cc1 = 26634; // (64^2*(.01*255)^2
|
| +static const int64 cc2 = 239708; // (64^2*(.03*255)^2
|
| +
|
| +static double Ssim8x8_C(const uint8* src_a, int stride_a,
|
| + const uint8* src_b, int stride_b) {
|
| + int64 sum_a = 0;
|
| + int64 sum_b = 0;
|
| + int64 sum_sq_a = 0;
|
| + int64 sum_sq_b = 0;
|
| + int64 sum_axb = 0;
|
| +
|
| + int i;
|
| + for (i = 0; i < 8; ++i) {
|
| + int j;
|
| + for (j = 0; j < 8; ++j) {
|
| + sum_a += src_a[j];
|
| + sum_b += src_b[j];
|
| + sum_sq_a += src_a[j] * src_a[j];
|
| + sum_sq_b += src_b[j] * src_b[j];
|
| + sum_axb += src_a[j] * src_b[j];
|
| + }
|
| +
|
| + src_a += stride_a;
|
| + src_b += stride_b;
|
| + }
|
| +
|
| + {
|
| + const int64 count = 64;
|
| + // scale the constants by number of pixels
|
| + const int64 c1 = (cc1 * count * count) >> 12;
|
| + const int64 c2 = (cc2 * count * count) >> 12;
|
| +
|
| + const int64 sum_a_x_sum_b = sum_a * sum_b;
|
| +
|
| + const int64 ssim_n = (2 * sum_a_x_sum_b + c1) *
|
| + (2 * count * sum_axb - 2 * sum_a_x_sum_b + c2);
|
| +
|
| + const int64 sum_a_sq = sum_a*sum_a;
|
| + const int64 sum_b_sq = sum_b*sum_b;
|
| +
|
| + const int64 ssim_d = (sum_a_sq + sum_b_sq + c1) *
|
| + (count * sum_sq_a - sum_a_sq +
|
| + count * sum_sq_b - sum_b_sq + c2);
|
| +
|
| + if (ssim_d == 0.0) {
|
| + return DBL_MAX;
|
| + }
|
| + return ssim_n * 1.0 / ssim_d;
|
| + }
|
| +}
|
| +
|
| +// We are using a 8x8 moving window with starting location of each 8x8 window
|
| +// on the 4x4 pixel grid. Such arrangement allows the windows to overlap
|
| +// block boundaries to penalize blocking artifacts.
|
| +LIBYUV_API
|
| +double CalcFrameSsim(const uint8* src_a, int stride_a,
|
| + const uint8* src_b, int stride_b,
|
| + int width, int height) {
|
| + int samples = 0;
|
| + double ssim_total = 0;
|
| + double (*Ssim8x8)(const uint8* src_a, int stride_a,
|
| + const uint8* src_b, int stride_b) = Ssim8x8_C;
|
| +
|
| + // sample point start with each 4x4 location
|
| + int i;
|
| + for (i = 0; i < height - 8; i += 4) {
|
| + int j;
|
| + for (j = 0; j < width - 8; j += 4) {
|
| + ssim_total += Ssim8x8(src_a + j, stride_a, src_b + j, stride_b);
|
| + samples++;
|
| + }
|
| +
|
| + src_a += stride_a * 4;
|
| + src_b += stride_b * 4;
|
| + }
|
| +
|
| + ssim_total /= samples;
|
| + return ssim_total;
|
| +}
|
| +
|
| +LIBYUV_API
|
| +double I420Ssim(const uint8* src_y_a, int stride_y_a,
|
| + const uint8* src_u_a, int stride_u_a,
|
| + const uint8* src_v_a, int stride_v_a,
|
| + const uint8* src_y_b, int stride_y_b,
|
| + const uint8* src_u_b, int stride_u_b,
|
| + const uint8* src_v_b, int stride_v_b,
|
| + int width, int height) {
|
| + const double ssim_y = CalcFrameSsim(src_y_a, stride_y_a,
|
| + src_y_b, stride_y_b, width, height);
|
| + const int width_uv = (width + 1) >> 1;
|
| + const int height_uv = (height + 1) >> 1;
|
| + const double ssim_u = CalcFrameSsim(src_u_a, stride_u_a,
|
| + src_u_b, stride_u_b,
|
| + width_uv, height_uv);
|
| + const double ssim_v = CalcFrameSsim(src_v_a, stride_v_a,
|
| + src_v_b, stride_v_b,
|
| + width_uv, height_uv);
|
| + return ssim_y * 0.8 + 0.1 * (ssim_u + ssim_v);
|
| +}
|
| +
|
| +#ifdef __cplusplus
|
| +} // extern "C"
|
| +} // namespace libyuv
|
| +#endif
|
|
|