mirror of
https://github.com/hrydgard/ppsspp.git
synced 2024-11-23 21:39:52 +00:00
7426e2a8f8
Now that the speed is closer, disable on non-ARMV7 so we can avoid wrong compilations more often.
296 lines
7.8 KiB
C++
296 lines
7.8 KiB
C++
// Copyright (c) 2012- PPSSPP Project.
|
|
|
|
// This program is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU General Public License as published by
|
|
// the Free Software Foundation, version 2.0 or later versions.
|
|
|
|
// This program is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU General Public License 2.0 for more details.
|
|
|
|
// A copy of the GPL 2.0 should have been included with the program.
|
|
// If not, see http://www.gnu.org/licenses/
|
|
|
|
// Official git repository and contact information can be found at
|
|
// https://github.com/hrydgard/ppsspp and http://www.ppsspp.org/.
|
|
|
|
#include "ppsspp_config.h"
|
|
|
|
#if PPSSPP_ARCH(ARM_NEON)
|
|
|
|
#include "ext/xxhash.h"
|
|
|
|
#if defined(_MSC_VER) && PPSSPP_ARCH(ARM64)
|
|
#include <arm64_neon.h>
|
|
#else
|
|
#include <arm_neon.h>
|
|
#endif
|
|
|
|
#include "GPU/GPUState.h"
|
|
#include "GPU/Common/TextureDecoder.h"
|
|
|
|
alignas(16) static const u16 QuickTexHashInitial[8] = {0xc00bU, 0x9bd9U, 0x4b73U, 0xb651U, 0x4d9bU, 0x4309U, 0x0083U, 0x0001U};
|
|
|
|
#ifdef _MSC_VER
|
|
#define __builtin_prefetch(a,b,c)
|
|
#endif
|
|
|
|
u32 QuickTexHashNEON(const void *checkp, u32 size) {
|
|
u32 check = 0;
|
|
__builtin_prefetch(checkp, 0, 0);
|
|
|
|
if (((intptr_t)checkp & 0xf) == 0 && (size & 0x3f) == 0) {
|
|
#if defined(IOS) || PPSSPP_ARCH(ARM64) || defined(_MSC_VER) || !PPSSPP_ARCH(ARMV7)
|
|
uint32x4_t cursor = vdupq_n_u32(0);
|
|
uint16x8_t cursor2 = vld1q_u16(QuickTexHashInitial);
|
|
uint16x8_t update = vdupq_n_u16(0x2455U);
|
|
|
|
const u32 *p = (const u32 *)checkp;
|
|
const u32 *pend = p + size / 4;
|
|
while (p < pend) {
|
|
cursor = vreinterpretq_u32_u16(vmlaq_u16(vreinterpretq_u16_u32(cursor), vreinterpretq_u16_u32(vld1q_u32(&p[4 * 0])), cursor2));
|
|
cursor = veorq_u32(cursor, vld1q_u32(&p[4 * 1]));
|
|
cursor = vaddq_u32(cursor, vld1q_u32(&p[4 * 2]));
|
|
cursor = veorq_u32(cursor, vreinterpretq_u32_u16(vmulq_u16(vreinterpretq_u16_u32(vld1q_u32(&p[4 * 3])), cursor2)));
|
|
cursor2 = vaddq_u16(cursor2, update);
|
|
|
|
p += 4 * 4;
|
|
}
|
|
|
|
cursor = vaddq_u32(cursor, vreinterpretq_u32_u16(cursor2));
|
|
uint32x2_t mixed = vadd_u32(vget_high_u32(cursor), vget_low_u32(cursor));
|
|
check = vget_lane_u32(mixed, 0) + vget_lane_u32(mixed, 1);
|
|
#else
|
|
// TODO: Why does this crash on iOS, but only certain devices?
|
|
// It's faster than the above, but I guess it sucks to be using an iPhone.
|
|
// As of 2020 clang, it's still faster by ~1.4%.
|
|
|
|
// d0/d1 (q0) - cursor
|
|
// d2/d3 (q1) - cursor2
|
|
// d4/d5 (q2) - update
|
|
// d16-d23 (q8-q11) - memory transfer
|
|
asm volatile (
|
|
// Initialize cursor.
|
|
"vmov.i32 q0, #0\n"
|
|
|
|
// Initialize cursor2.
|
|
"movw r0, 0xc00b\n"
|
|
"movt r0, 0x9bd9\n"
|
|
"movw r1, 0x4b73\n"
|
|
"movt r1, 0xb651\n"
|
|
"vmov d2, r0, r1\n"
|
|
"movw r0, 0x4d9b\n"
|
|
"movt r0, 0x4309\n"
|
|
"movw r1, 0x0083\n"
|
|
"movt r1, 0x0001\n"
|
|
"vmov d3, r0, r1\n"
|
|
|
|
// Initialize update.
|
|
"movw r0, 0x2455\n"
|
|
"vdup.i16 q2, r0\n"
|
|
|
|
// This is where we end.
|
|
"add r0, %1, %2\n"
|
|
|
|
// Okay, do the memory hashing.
|
|
"QuickTexHashNEON_next:\n"
|
|
"pld [%2, #0xc0]\n"
|
|
"vldmia %2!, {d16-d23}\n"
|
|
"vmla.i16 q0, q1, q8\n"
|
|
"vmul.i16 q11, q11, q1\n"
|
|
"veor.i32 q0, q0, q9\n"
|
|
"cmp %2, r0\n"
|
|
"vadd.i32 q0, q0, q10\n"
|
|
"vadd.i16 q1, q1, q2\n"
|
|
"veor.i32 q0, q0, q11\n"
|
|
"blo QuickTexHashNEON_next\n"
|
|
|
|
// Now let's get the result.
|
|
"vadd.i32 q0, q0, q1\n"
|
|
"vadd.i32 d0, d0, d1\n"
|
|
"vmov r0, r1, d0\n"
|
|
"add %0, r0, r1\n"
|
|
|
|
: "=r"(check)
|
|
: "r"(size), "r"(checkp)
|
|
: "r0", "r1", "d0", "d1", "d2", "d3", "d4", "d5", "d16", "d17", "d18", "d19", "d20", "d21", "d22", "d23", "cc"
|
|
);
|
|
#endif
|
|
} else {
|
|
const u32 size_u32 = size / 4;
|
|
const u32 *p = (const u32 *)checkp;
|
|
for (u32 i = 0; i < size_u32; i += 4) {
|
|
check += p[i + 0];
|
|
check ^= p[i + 1];
|
|
check += p[i + 2];
|
|
check ^= p[i + 3];
|
|
}
|
|
}
|
|
|
|
return check;
|
|
}
|
|
|
|
void DoUnswizzleTex16NEON(const u8 *texptr, u32 *ydestp, int bxc, int byc, u32 pitch) {
|
|
// ydestp is in 32-bits, so this is convenient.
|
|
const u32 pitchBy32 = pitch >> 2;
|
|
|
|
__builtin_prefetch(texptr, 0, 0);
|
|
__builtin_prefetch(ydestp, 1, 1);
|
|
|
|
const u32 *src = (const u32 *)texptr;
|
|
for (int by = 0; by < byc; by++) {
|
|
u32 *xdest = ydestp;
|
|
for (int bx = 0; bx < bxc; bx++) {
|
|
u32 *dest = xdest;
|
|
for (int n = 0; n < 2; n++) {
|
|
// Textures are always 16-byte aligned so this is fine.
|
|
uint32x4_t temp1 = vld1q_u32(src);
|
|
uint32x4_t temp2 = vld1q_u32(src + 4);
|
|
uint32x4_t temp3 = vld1q_u32(src + 8);
|
|
uint32x4_t temp4 = vld1q_u32(src + 12);
|
|
vst1q_u32(dest, temp1);
|
|
dest += pitchBy32;
|
|
vst1q_u32(dest, temp2);
|
|
dest += pitchBy32;
|
|
vst1q_u32(dest, temp3);
|
|
dest += pitchBy32;
|
|
vst1q_u32(dest, temp4);
|
|
dest += pitchBy32;
|
|
src += 16;
|
|
}
|
|
xdest += 4;
|
|
}
|
|
ydestp += pitchBy32 * 8;
|
|
}
|
|
}
|
|
|
|
static inline bool VectorIsNonZeroNEON(const uint32x4_t &v) {
|
|
u64 low = vgetq_lane_u64(vreinterpretq_u64_u32(v), 0);
|
|
u64 high = vgetq_lane_u64(vreinterpretq_u64_u32(v), 1);
|
|
|
|
return (low | high) != 0;
|
|
}
|
|
|
|
#ifndef _MSC_VER
|
|
// MSVC consider this function the same as the one above! uint16x8_t is typedef'd to the same type as uint32x4_t.
|
|
static inline bool VectorIsNonZeroNEON(const uint16x8_t &v) {
|
|
u64 low = vgetq_lane_u64(vreinterpretq_u64_u16(v), 0);
|
|
u64 high = vgetq_lane_u64(vreinterpretq_u64_u16(v), 1);
|
|
|
|
return (low | high) != 0;
|
|
}
|
|
#endif
|
|
|
|
CheckAlphaResult CheckAlphaRGBA8888NEON(const u32 *pixelData, int stride, int w, int h) {
|
|
const u32 *p = (const u32 *)pixelData;
|
|
|
|
const uint32x4_t mask = vdupq_n_u32(0xFF000000);
|
|
uint32x4_t bits = mask;
|
|
for (int y = 0; y < h; ++y) {
|
|
for (int i = 0; i < w; i += 4) {
|
|
const uint32x4_t a = vld1q_u32(&p[i]);
|
|
bits = vandq_u32(bits, a);
|
|
}
|
|
|
|
uint32x4_t result = veorq_u32(bits, mask);
|
|
if (VectorIsNonZeroNEON(result)) {
|
|
return CHECKALPHA_ANY;
|
|
}
|
|
|
|
p += stride;
|
|
}
|
|
|
|
return CHECKALPHA_FULL;
|
|
}
|
|
|
|
CheckAlphaResult CheckAlphaABGR4444NEON(const u32 *pixelData, int stride, int w, int h) {
|
|
const u16 *p = (const u16 *)pixelData;
|
|
|
|
const uint16x8_t mask = vdupq_n_u16((u16)0x000F);
|
|
uint16x8_t bits = mask;
|
|
for (int y = 0; y < h; ++y) {
|
|
for (int i = 0; i < w; i += 8) {
|
|
const uint16x8_t a = vld1q_u16(&p[i]);
|
|
bits = vandq_u16(bits, a);
|
|
}
|
|
|
|
uint16x8_t result = veorq_u16(bits, mask);
|
|
if (VectorIsNonZeroNEON(result)) {
|
|
return CHECKALPHA_ANY;
|
|
}
|
|
|
|
p += stride;
|
|
}
|
|
|
|
return CHECKALPHA_FULL;
|
|
}
|
|
|
|
CheckAlphaResult CheckAlphaABGR1555NEON(const u32 *pixelData, int stride, int w, int h) {
|
|
const u16 *p = (const u16 *)pixelData;
|
|
|
|
const uint16x8_t mask = vdupq_n_u16((u16)0x0001);
|
|
uint16x8_t bits = mask;
|
|
for (int y = 0; y < h; ++y) {
|
|
for (int i = 0; i < w; i += 8) {
|
|
const uint16x8_t a = vld1q_u16(&p[i]);
|
|
bits = vandq_u16(bits, a);
|
|
}
|
|
|
|
uint16x8_t result = veorq_u16(bits, mask);
|
|
if (VectorIsNonZeroNEON(result)) {
|
|
return CHECKALPHA_ANY;
|
|
}
|
|
|
|
p += stride;
|
|
}
|
|
|
|
return CHECKALPHA_FULL;
|
|
}
|
|
|
|
CheckAlphaResult CheckAlphaRGBA4444NEON(const u32 *pixelData, int stride, int w, int h) {
|
|
const u16 *p = (const u16 *)pixelData;
|
|
|
|
const uint16x8_t mask = vdupq_n_u16((u16)0xF000);
|
|
uint16x8_t bits = mask;
|
|
for (int y = 0; y < h; ++y) {
|
|
for (int i = 0; i < w; i += 8) {
|
|
const uint16x8_t a = vld1q_u16(&p[i]);
|
|
bits = vandq_u16(bits, a);
|
|
}
|
|
|
|
uint16x8_t result = veorq_u16(bits, mask);
|
|
if (VectorIsNonZeroNEON(result)) {
|
|
return CHECKALPHA_ANY;
|
|
}
|
|
|
|
p += stride;
|
|
}
|
|
|
|
return CHECKALPHA_FULL;
|
|
}
|
|
|
|
CheckAlphaResult CheckAlphaRGBA5551NEON(const u32 *pixelData, int stride, int w, int h) {
|
|
const u16 *p = (const u16 *)pixelData;
|
|
|
|
const uint16x8_t mask = vdupq_n_u16((u16)0x8000);
|
|
uint16x8_t bits = mask;
|
|
for (int y = 0; y < h; ++y) {
|
|
for (int i = 0; i < w; i += 8) {
|
|
const uint16x8_t a = vld1q_u16(&p[i]);
|
|
bits = vandq_u16(bits, a);
|
|
}
|
|
|
|
uint16x8_t result = veorq_u16(bits, mask);
|
|
if (VectorIsNonZeroNEON(result)) {
|
|
return CHECKALPHA_ANY;
|
|
}
|
|
|
|
p += stride;
|
|
}
|
|
|
|
return CHECKALPHA_FULL;
|
|
}
|
|
|
|
#endif
|