mirror of
https://github.com/mozilla/gecko-dev.git
synced 2024-11-06 17:16:12 +00:00
248 lines
7.9 KiB
C++
248 lines
7.9 KiB
C++
|
/*
|
||
|
* Copyright 2013 The LibYuv Project Authors. All rights reserved.
|
||
|
*
|
||
|
* Use of this source code is governed by a BSD-style license
|
||
|
* that can be found in the LICENSE file in the root of the source
|
||
|
* tree. An additional intellectual property rights grant can be found
|
||
|
* in the file PATENTS. All contributing project authors may
|
||
|
* be found in the AUTHORS file in the root of the source tree.
|
||
|
*/
|
||
|
|
||
|
#include "./psnr.h" // NOLINT
|
||
|
|
||
|
#include <math.h>
|
||
|
|
||
|
#ifdef _OPENMP
|
||
|
#include <omp.h>
|
||
|
#endif
|
||
|
#ifdef _MSC_VER
|
||
|
#include <intrin.h> // For __cpuid()
|
||
|
#endif
|
||
|
|
||
|
#ifdef __cplusplus
|
||
|
extern "C" {
|
||
|
#endif
|
||
|
|
||
|
typedef unsigned int uint32; // NOLINT
|
||
|
#ifdef _MSC_VER
|
||
|
typedef unsigned __int64 uint64;
|
||
|
#else // COMPILER_MSVC
|
||
|
#if defined(__LP64__) && !defined(__OpenBSD__) && !defined(__APPLE__)
|
||
|
typedef unsigned long uint64; // NOLINT
|
||
|
#else // defined(__LP64__) && !defined(__OpenBSD__) && !defined(__APPLE__)
|
||
|
typedef unsigned long long uint64; // NOLINT
|
||
|
#endif // __LP64__
|
||
|
#endif // _MSC_VER
|
||
|
|
||
|
// PSNR formula: psnr = 10 * log10 (Peak Signal^2 * size / sse)
|
||
|
double ComputePSNR(double sse, double size) {
|
||
|
const double kMINSSE = 255.0 * 255.0 * size / pow(10., kMaxPSNR / 10.);
|
||
|
if (sse <= kMINSSE)
|
||
|
sse = kMINSSE; // Produces max PSNR of 128
|
||
|
return 10.0 * log10(65025.0 * size / sse);
|
||
|
}
|
||
|
|
||
|
#if !defined(LIBYUV_DISABLE_NEON) && defined(__ARM_NEON__)
|
||
|
#define HAS_SUMSQUAREERROR_NEON
|
||
|
static uint32 SumSquareError_NEON(const uint8* src_a,
|
||
|
const uint8* src_b, int count) {
|
||
|
volatile uint32 sse;
|
||
|
asm volatile ( // NOLINT
|
||
|
"vmov.u8 q7, #0 \n"
|
||
|
"vmov.u8 q9, #0 \n"
|
||
|
"vmov.u8 q8, #0 \n"
|
||
|
"vmov.u8 q10, #0 \n"
|
||
|
|
||
|
"1: \n"
|
||
|
"vld1.u8 {q0}, [%0]! \n"
|
||
|
"vld1.u8 {q1}, [%1]! \n"
|
||
|
"vsubl.u8 q2, d0, d2 \n"
|
||
|
"vsubl.u8 q3, d1, d3 \n"
|
||
|
"vmlal.s16 q7, d4, d4 \n"
|
||
|
"vmlal.s16 q8, d6, d6 \n"
|
||
|
"vmlal.s16 q8, d5, d5 \n"
|
||
|
"vmlal.s16 q10, d7, d7 \n"
|
||
|
"subs %2, %2, #16 \n"
|
||
|
"bhi 1b \n"
|
||
|
|
||
|
"vadd.u32 q7, q7, q8 \n"
|
||
|
"vadd.u32 q9, q9, q10 \n"
|
||
|
"vadd.u32 q10, q7, q9 \n"
|
||
|
"vpaddl.u32 q1, q10 \n"
|
||
|
"vadd.u64 d0, d2, d3 \n"
|
||
|
"vmov.32 %3, d0[0] \n"
|
||
|
: "+r"(src_a),
|
||
|
"+r"(src_b),
|
||
|
"+r"(count),
|
||
|
"=r"(sse)
|
||
|
:
|
||
|
: "memory", "cc", "q0", "q1", "q2", "q3", "q7", "q8", "q9", "q10");
|
||
|
return sse;
|
||
|
}
|
||
|
#elif !defined(LIBYUV_DISABLE_X86) && defined(_M_IX86) && defined(_MSC_VER)
|
||
|
#define HAS_SUMSQUAREERROR_SSE2
|
||
|
__declspec(naked)
|
||
|
static uint32 SumSquareError_SSE2(const uint8* /*src_a*/,
|
||
|
const uint8* /*src_b*/, int /*count*/) {
|
||
|
__asm {
|
||
|
mov eax, [esp + 4] // src_a
|
||
|
mov edx, [esp + 8] // src_b
|
||
|
mov ecx, [esp + 12] // count
|
||
|
pxor xmm0, xmm0
|
||
|
pxor xmm5, xmm5
|
||
|
sub edx, eax
|
||
|
|
||
|
wloop:
|
||
|
movdqu xmm1, [eax]
|
||
|
movdqu xmm2, [eax + edx]
|
||
|
lea eax, [eax + 16]
|
||
|
movdqu xmm3, xmm1
|
||
|
psubusb xmm1, xmm2
|
||
|
psubusb xmm2, xmm3
|
||
|
por xmm1, xmm2
|
||
|
movdqu xmm2, xmm1
|
||
|
punpcklbw xmm1, xmm5
|
||
|
punpckhbw xmm2, xmm5
|
||
|
pmaddwd xmm1, xmm1
|
||
|
pmaddwd xmm2, xmm2
|
||
|
paddd xmm0, xmm1
|
||
|
paddd xmm0, xmm2
|
||
|
sub ecx, 16
|
||
|
ja wloop
|
||
|
|
||
|
pshufd xmm1, xmm0, 0EEh
|
||
|
paddd xmm0, xmm1
|
||
|
pshufd xmm1, xmm0, 01h
|
||
|
paddd xmm0, xmm1
|
||
|
movd eax, xmm0
|
||
|
ret
|
||
|
}
|
||
|
}
|
||
|
#elif !defined(LIBYUV_DISABLE_X86) && (defined(__x86_64__) || defined(__i386__))
|
||
|
#define HAS_SUMSQUAREERROR_SSE2
|
||
|
static uint32 SumSquareError_SSE2(const uint8* src_a,
|
||
|
const uint8* src_b, int count) {
|
||
|
uint32 sse;
|
||
|
asm volatile ( // NOLINT
|
||
|
"pxor %%xmm0,%%xmm0 \n"
|
||
|
"pxor %%xmm5,%%xmm5 \n"
|
||
|
"sub %0,%1 \n"
|
||
|
|
||
|
"1: \n"
|
||
|
"movdqu (%0),%%xmm1 \n"
|
||
|
"movdqu (%0,%1,1),%%xmm2 \n"
|
||
|
"lea 0x10(%0),%0 \n"
|
||
|
"movdqu %%xmm1,%%xmm3 \n"
|
||
|
"psubusb %%xmm2,%%xmm1 \n"
|
||
|
"psubusb %%xmm3,%%xmm2 \n"
|
||
|
"por %%xmm2,%%xmm1 \n"
|
||
|
"movdqu %%xmm1,%%xmm2 \n"
|
||
|
"punpcklbw %%xmm5,%%xmm1 \n"
|
||
|
"punpckhbw %%xmm5,%%xmm2 \n"
|
||
|
"pmaddwd %%xmm1,%%xmm1 \n"
|
||
|
"pmaddwd %%xmm2,%%xmm2 \n"
|
||
|
"paddd %%xmm1,%%xmm0 \n"
|
||
|
"paddd %%xmm2,%%xmm0 \n"
|
||
|
"sub $0x10,%2 \n"
|
||
|
"ja 1b \n"
|
||
|
|
||
|
"pshufd $0xee,%%xmm0,%%xmm1 \n"
|
||
|
"paddd %%xmm1,%%xmm0 \n"
|
||
|
"pshufd $0x1,%%xmm0,%%xmm1 \n"
|
||
|
"paddd %%xmm1,%%xmm0 \n"
|
||
|
"movd %%xmm0,%3 \n"
|
||
|
|
||
|
: "+r"(src_a), // %0
|
||
|
"+r"(src_b), // %1
|
||
|
"+r"(count), // %2
|
||
|
"=g"(sse) // %3
|
||
|
:
|
||
|
: "memory", "cc"
|
||
|
#if defined(__SSE2__)
|
||
|
, "xmm0", "xmm1", "xmm2", "xmm3", "xmm5"
|
||
|
#endif
|
||
|
); // NOLINT
|
||
|
return sse;
|
||
|
}
|
||
|
#endif // LIBYUV_DISABLE_X86 etc
|
||
|
|
||
|
#if defined(HAS_SUMSQUAREERROR_SSE2)
|
||
|
#if (defined(__pic__) || defined(__APPLE__)) && defined(__i386__)
|
||
|
static __inline void __cpuid(int cpu_info[4], int info_type) {
|
||
|
asm volatile ( // NOLINT
|
||
|
"mov %%ebx, %%edi \n"
|
||
|
"cpuid \n"
|
||
|
"xchg %%edi, %%ebx \n"
|
||
|
: "=a"(cpu_info[0]), "=D"(cpu_info[1]), "=c"(cpu_info[2]), "=d"(cpu_info[3])
|
||
|
: "a"(info_type));
|
||
|
}
|
||
|
#elif defined(__i386__) || defined(__x86_64__)
|
||
|
static __inline void __cpuid(int cpu_info[4], int info_type) {
|
||
|
asm volatile ( // NOLINT
|
||
|
"cpuid \n"
|
||
|
: "=a"(cpu_info[0]), "=b"(cpu_info[1]), "=c"(cpu_info[2]), "=d"(cpu_info[3])
|
||
|
: "a"(info_type));
|
||
|
}
|
||
|
#endif
|
||
|
|
||
|
static int CpuHasSSE2() {
|
||
|
#if defined(__i386__) || defined(__x86_64__) || defined(_M_IX86)
|
||
|
int cpu_info[4];
|
||
|
__cpuid(cpu_info, 1);
|
||
|
if (cpu_info[3] & 0x04000000) {
|
||
|
return 1;
|
||
|
}
|
||
|
#endif
|
||
|
return 0;
|
||
|
}
|
||
|
#endif // HAS_SUMSQUAREERROR_SSE2
|
||
|
|
||
|
static uint32 SumSquareError_C(const uint8* src_a,
|
||
|
const uint8* src_b, int count) {
|
||
|
uint32 sse = 0u;
|
||
|
for (int x = 0; x < count; ++x) {
|
||
|
int diff = src_a[x] - src_b[x];
|
||
|
sse += static_cast<uint32>(diff * diff);
|
||
|
}
|
||
|
return sse;
|
||
|
}
|
||
|
|
||
|
double ComputeSumSquareError(const uint8* src_a,
|
||
|
const uint8* src_b, int count) {
|
||
|
uint32 (*SumSquareError)(const uint8* src_a,
|
||
|
const uint8* src_b, int count) = SumSquareError_C;
|
||
|
#if defined(HAS_SUMSQUAREERROR_NEON)
|
||
|
SumSquareError = SumSquareError_NEON;
|
||
|
#endif
|
||
|
#if defined(HAS_SUMSQUAREERROR_SSE2)
|
||
|
if (CpuHasSSE2()) {
|
||
|
SumSquareError = SumSquareError_SSE2;
|
||
|
}
|
||
|
#endif
|
||
|
const int kBlockSize = 1 << 15;
|
||
|
uint64 sse = 0;
|
||
|
#ifdef _OPENMP
|
||
|
#pragma omp parallel for reduction(+: sse)
|
||
|
#endif
|
||
|
for (int i = 0; i < (count - (kBlockSize - 1)); i += kBlockSize) {
|
||
|
sse += SumSquareError(src_a + i, src_b + i, kBlockSize);
|
||
|
}
|
||
|
src_a += count & ~(kBlockSize - 1);
|
||
|
src_b += count & ~(kBlockSize - 1);
|
||
|
int remainder = count & (kBlockSize - 1) & ~15;
|
||
|
if (remainder) {
|
||
|
sse += SumSquareError(src_a, src_b, remainder);
|
||
|
src_a += remainder;
|
||
|
src_b += remainder;
|
||
|
}
|
||
|
remainder = count & 15;
|
||
|
if (remainder) {
|
||
|
sse += SumSquareError_C(src_a, src_b, remainder);
|
||
|
}
|
||
|
return static_cast<double>(sse);
|
||
|
}
|
||
|
|
||
|
#ifdef __cplusplus
|
||
|
} // extern "C"
|
||
|
#endif
|