// Copyright (c) 2013- PPSSPP Project. // This program is free software: you can redistribute it and/or modify // it under the terms of the GNU General Public License as published by // the Free Software Foundation, version 2.0 or later versions. // This program is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU General Public License 2.0 for more details. // A copy of the GPL 2.0 should have been included with the program. // If not, see http://www.gnu.org/licenses/ // Official git repository and contact information can be found at // https://github.com/hrydgard/ppsspp and http://www.ppsspp.org/. #include "ppsspp_config.h" #if PPSSPP_ARCH(X86) || PPSSPP_ARCH(AMD64) #include #include "Common/CPUDetect.h" #include "Core/Config.h" #include "Core/Reporting.h" #include "GPU/GPUState.h" #include "GPU/Common/VertexDecoderCommon.h" // We start out by converting the active matrices into 4x4 which are easier to multiply with // using SSE / NEON and store them here. alignas(16) static float bones[16 * 8]; using namespace Gen; alignas(16) static const float by128[4] = { 1.0f / 128.0f, 1.0f / 128.0f, 1.0f / 128.0f, 1.0f / 128.0f }; alignas(16) static const float by32768[4] = { 1.0f / 32768.0f, 1.0f / 32768.0f, 1.0f / 32768.0f, 1.0f / 32768.0f, }; alignas(16) static const float by128_11[4] = { 1.0f / 128.0f, 1.0f / 128.0f, 1.0f, 1.0f, }; alignas(16) static const float by32768_11[4] = { 1.0f / 32768.0f, 1.0f / 32768.0f, 1.0f, 1.0f, }; alignas(16) static const u32 threeMasks[4] = { 0xFFFFFFFF, 0xFFFFFFFF, 0xFFFFFFFF, 0 }; alignas(16) static const u32 aOne[4] = {0, 0, 0, 0x3F800000}; alignas(16) static const float by16384[4] = { 1.0f / 16384.0f, 1.0f / 16384.0f, 1.0f / 16384.0f, 1.0f / 16384.0f, }; #if PPSSPP_ARCH(AMD64) #ifdef _WIN32 static const X64Reg tempReg1 = RAX; static const X64Reg tempReg2 = R9; static const X64Reg tempReg3 = R10; static const X64Reg srcReg = RCX; static const X64Reg dstReg = RDX; static const X64Reg counterReg = R8; #else static const X64Reg tempReg1 = RAX; static const X64Reg tempReg2 = R9; static const X64Reg tempReg3 = R10; static const X64Reg srcReg = RDI; static const X64Reg dstReg = RSI; static const X64Reg counterReg = RDX; #endif #else static const X64Reg tempReg1 = EAX; static const X64Reg tempReg2 = EBX; static const X64Reg tempReg3 = EDX; static const X64Reg srcReg = ESI; static const X64Reg dstReg = EDI; static const X64Reg counterReg = ECX; #endif // XMM0-XMM5 are volatile on Windows X64 // XMM0-XMM7 are arguments (and thus volatile) on System V ABI (other x64 platforms) static const X64Reg fpScaleOffsetReg = XMM0; static const X64Reg fpScratchReg = XMM1; static const X64Reg fpScratchReg2 = XMM2; static const X64Reg fpScratchReg3 = XMM3; static const X64Reg fpScratchReg4 = XMM4; // We're gonna keep the current skinning matrix in 4 XMM regs. Fortunately we easily // have space for that now. // To debug, just comment them out one at a time until it works. We fall back // on the interpreter if the compiler fails. static const JitLookup jitLookup[] = { {&VertexDecoder::Step_WeightsU8, &VertexDecoderJitCache::Jit_WeightsU8}, {&VertexDecoder::Step_WeightsU16, &VertexDecoderJitCache::Jit_WeightsU16}, {&VertexDecoder::Step_WeightsFloat, &VertexDecoderJitCache::Jit_WeightsFloat}, {&VertexDecoder::Step_WeightsU8Skin, &VertexDecoderJitCache::Jit_WeightsU8Skin}, {&VertexDecoder::Step_WeightsU16Skin, &VertexDecoderJitCache::Jit_WeightsU16Skin}, {&VertexDecoder::Step_WeightsFloatSkin, &VertexDecoderJitCache::Jit_WeightsFloatSkin}, {&VertexDecoder::Step_WeightsU8ToFloat, &VertexDecoderJitCache::Jit_WeightsU8ToFloat}, {&VertexDecoder::Step_WeightsU16ToFloat, &VertexDecoderJitCache::Jit_WeightsU16ToFloat}, {&VertexDecoder::Step_TcFloat, &VertexDecoderJitCache::Jit_TcFloat}, {&VertexDecoder::Step_TcU8ToFloat, &VertexDecoderJitCache::Jit_TcU8ToFloat}, {&VertexDecoder::Step_TcU16ToFloat, &VertexDecoderJitCache::Jit_TcU16ToFloat}, {&VertexDecoder::Step_TcU8Prescale, &VertexDecoderJitCache::Jit_TcU8Prescale}, {&VertexDecoder::Step_TcU16Prescale, &VertexDecoderJitCache::Jit_TcU16Prescale}, {&VertexDecoder::Step_TcFloatPrescale, &VertexDecoderJitCache::Jit_TcFloatPrescale}, {&VertexDecoder::Step_TcU16ThroughToFloat, &VertexDecoderJitCache::Jit_TcU16ThroughToFloat}, {&VertexDecoder::Step_TcFloatThrough, &VertexDecoderJitCache::Jit_TcFloatThrough}, {&VertexDecoder::Step_TcU8MorphToFloat, &VertexDecoderJitCache::Jit_TcU8MorphToFloat}, {&VertexDecoder::Step_TcU16MorphToFloat, &VertexDecoderJitCache::Jit_TcU16MorphToFloat}, {&VertexDecoder::Step_TcFloatMorph, &VertexDecoderJitCache::Jit_TcFloatMorph}, {&VertexDecoder::Step_TcU8PrescaleMorph, &VertexDecoderJitCache::Jit_TcU8PrescaleMorph}, {&VertexDecoder::Step_TcU16PrescaleMorph, &VertexDecoderJitCache::Jit_TcU16PrescaleMorph}, {&VertexDecoder::Step_TcFloatPrescaleMorph, &VertexDecoderJitCache::Jit_TcFloatPrescaleMorph}, {&VertexDecoder::Step_NormalS8, &VertexDecoderJitCache::Jit_NormalS8}, {&VertexDecoder::Step_NormalS8ToFloat, &VertexDecoderJitCache::Jit_NormalS8ToFloat}, {&VertexDecoder::Step_NormalS16, &VertexDecoderJitCache::Jit_NormalS16}, {&VertexDecoder::Step_NormalFloat, &VertexDecoderJitCache::Jit_NormalFloat}, {&VertexDecoder::Step_NormalS8Skin, &VertexDecoderJitCache::Jit_NormalS8Skin}, {&VertexDecoder::Step_NormalS16Skin, &VertexDecoderJitCache::Jit_NormalS16Skin}, {&VertexDecoder::Step_NormalFloatSkin, &VertexDecoderJitCache::Jit_NormalFloatSkin}, {&VertexDecoder::Step_Color8888, &VertexDecoderJitCache::Jit_Color8888}, {&VertexDecoder::Step_Color4444, &VertexDecoderJitCache::Jit_Color4444}, {&VertexDecoder::Step_Color565, &VertexDecoderJitCache::Jit_Color565}, {&VertexDecoder::Step_Color5551, &VertexDecoderJitCache::Jit_Color5551}, {&VertexDecoder::Step_PosS8Through, &VertexDecoderJitCache::Jit_PosS8Through}, {&VertexDecoder::Step_PosS16Through, &VertexDecoderJitCache::Jit_PosS16Through}, {&VertexDecoder::Step_PosFloatThrough, &VertexDecoderJitCache::Jit_PosFloat}, {&VertexDecoder::Step_PosS8, &VertexDecoderJitCache::Jit_PosS8}, {&VertexDecoder::Step_PosS16, &VertexDecoderJitCache::Jit_PosS16}, {&VertexDecoder::Step_PosFloat, &VertexDecoderJitCache::Jit_PosFloat}, {&VertexDecoder::Step_PosS8Skin, &VertexDecoderJitCache::Jit_PosS8Skin}, {&VertexDecoder::Step_PosS16Skin, &VertexDecoderJitCache::Jit_PosS16Skin}, {&VertexDecoder::Step_PosFloatSkin, &VertexDecoderJitCache::Jit_PosFloatSkin}, {&VertexDecoder::Step_NormalS8Morph, &VertexDecoderJitCache::Jit_NormalS8Morph}, {&VertexDecoder::Step_NormalS16Morph, &VertexDecoderJitCache::Jit_NormalS16Morph}, {&VertexDecoder::Step_NormalFloatMorph, &VertexDecoderJitCache::Jit_NormalFloatMorph}, {&VertexDecoder::Step_PosS8Morph, &VertexDecoderJitCache::Jit_PosS8Morph}, {&VertexDecoder::Step_PosS16Morph, &VertexDecoderJitCache::Jit_PosS16Morph}, {&VertexDecoder::Step_PosFloatMorph, &VertexDecoderJitCache::Jit_PosFloatMorph}, {&VertexDecoder::Step_Color8888Morph, &VertexDecoderJitCache::Jit_Color8888Morph}, {&VertexDecoder::Step_Color4444Morph, &VertexDecoderJitCache::Jit_Color4444Morph}, {&VertexDecoder::Step_Color565Morph, &VertexDecoderJitCache::Jit_Color565Morph}, {&VertexDecoder::Step_Color5551Morph, &VertexDecoderJitCache::Jit_Color5551Morph}, }; JittedVertexDecoder VertexDecoderJitCache::Compile(const VertexDecoder &dec, int32_t *jittedSize) { dec_ = &dec; BeginWrite(); const u8 *start = this->AlignCode16(); #if PPSSPP_ARCH(X86) // Store register values PUSH(ESI); PUSH(EDI); PUSH(EBX); PUSH(EBP); // Read parameters int offset = 4; MOV(32, R(srcReg), MDisp(ESP, 16 + offset + 0)); MOV(32, R(dstReg), MDisp(ESP, 16 + offset + 4)); MOV(32, R(counterReg), MDisp(ESP, 16 + offset + 8)); const uint8_t STACK_FIXED_ALLOC = 64; #else // Parameters automatically fall into place. // This will align the stack properly to 16 bytes (the call of this function pushed RIP, which is 8 bytes). const uint8_t STACK_FIXED_ALLOC = 96 + 8; #endif // Allocate temporary storage on the stack. SUB(PTRBITS, R(ESP), Imm8(STACK_FIXED_ALLOC)); // Save XMM4/XMM5 which apparently can be problematic? // Actually, if they are, it must be a compiler bug because they SHOULD be ok. // So I won't bother. MOVUPS(MDisp(ESP, 0), XMM4); MOVUPS(MDisp(ESP, 16), XMM5); MOVUPS(MDisp(ESP, 32), XMM6); MOVUPS(MDisp(ESP, 48), XMM7); #if PPSSPP_ARCH(AMD64) MOVUPS(MDisp(ESP, 64), XMM8); MOVUPS(MDisp(ESP, 80), XMM9); #endif bool prescaleStep = false; // Look for prescaled texcoord steps for (int i = 0; i < dec.numSteps_; i++) { if (dec.steps_[i] == &VertexDecoder::Step_TcU8Prescale || dec.steps_[i] == &VertexDecoder::Step_TcU16Prescale || dec.steps_[i] == &VertexDecoder::Step_TcFloatPrescale) { prescaleStep = true; } if (dec.steps_[i] == &VertexDecoder::Step_TcU8PrescaleMorph || dec.steps_[i] == &VertexDecoder::Step_TcU16PrescaleMorph || dec.steps_[i] == &VertexDecoder::Step_TcFloatPrescaleMorph) { prescaleStep = true; } } // Add code to convert matrices to 4x4. // Later we might want to do this when the matrices are loaded instead. if (dec.weighttype && g_Config.bSoftwareSkinning) { MOV(PTRBITS, R(tempReg1), ImmPtr(&threeMasks)); MOVAPS(XMM4, MatR(tempReg1)); MOV(PTRBITS, R(tempReg1), ImmPtr(&aOne)); MOVUPS(XMM5, MatR(tempReg1)); MOV(PTRBITS, R(tempReg1), ImmPtr(gstate.boneMatrix)); MOV(PTRBITS, R(tempReg2), ImmPtr(bones)); for (int i = 0; i < dec.nweights; i++) { MOVUPS(XMM0, MDisp(tempReg1, (12 * i) * 4)); MOVUPS(XMM1, MDisp(tempReg1, (12 * i + 3) * 4)); MOVUPS(XMM2, MDisp(tempReg1, (12 * i + 3 * 2) * 4)); MOVUPS(XMM3, MDisp(tempReg1, (12 * i + 3 * 3) * 4)); ANDPS(XMM0, R(XMM4)); ANDPS(XMM1, R(XMM4)); ANDPS(XMM2, R(XMM4)); ANDPS(XMM3, R(XMM4)); ORPS(XMM3, R(XMM5)); MOVAPS(MDisp(tempReg2, (16 * i) * 4), XMM0); MOVAPS(MDisp(tempReg2, (16 * i + 4) * 4), XMM1); MOVAPS(MDisp(tempReg2, (16 * i + 8) * 4), XMM2); MOVAPS(MDisp(tempReg2, (16 * i + 12) * 4), XMM3); } } // Keep the scale/offset in a few fp registers if we need it. if (prescaleStep) { MOV(PTRBITS, R(tempReg1), ImmPtr(&gstate_c.uv)); MOVUPS(fpScaleOffsetReg, MatR(tempReg1)); if ((dec.VertexType() & GE_VTYPE_TC_MASK) == GE_VTYPE_TC_8BIT) { MOV(PTRBITS, R(tempReg2), ImmPtr(&by128_11)); MULPS(fpScaleOffsetReg, MatR(tempReg2)); } else if ((dec.VertexType() & GE_VTYPE_TC_MASK) == GE_VTYPE_TC_16BIT) { MOV(PTRBITS, R(tempReg2), ImmPtr(&by32768_11)); MULPS(fpScaleOffsetReg, MatR(tempReg2)); } } // Let's not bother with a proper stack frame. We just grab the arguments and go. JumpTarget loopStart = GetCodePtr(); for (int i = 0; i < dec.numSteps_; i++) { if (!CompileStep(dec, i)) { EndWrite(); // Reset the code ptr and return zero to indicate that we failed. ResetCodePtr(GetOffset(start)); return 0; } } ADD(PTRBITS, R(srcReg), Imm32(dec.VertexSize())); ADD(PTRBITS, R(dstReg), Imm32(dec.decFmt.stride)); SUB(32, R(counterReg), Imm8(1)); J_CC(CC_NZ, loopStart, true); MOVUPS(XMM4, MDisp(ESP, 0)); MOVUPS(XMM5, MDisp(ESP, 16)); MOVUPS(XMM6, MDisp(ESP, 32)); MOVUPS(XMM7, MDisp(ESP, 48)); #if PPSSPP_ARCH(AMD64) MOVUPS(XMM8, MDisp(ESP, 64)); MOVUPS(XMM9, MDisp(ESP, 80)); #endif ADD(PTRBITS, R(ESP), Imm8(STACK_FIXED_ALLOC)); #if PPSSPP_ARCH(X86) // Restore register values POP(EBP); POP(EBX); POP(EDI); POP(ESI); #endif RET(); *jittedSize = GetCodePtr() - start; EndWrite(); return (JittedVertexDecoder)start; } void VertexDecoderJitCache::Jit_WeightsU8() { switch (dec_->nweights) { case 1: MOVZX(32, 8, tempReg1, MDisp(srcReg, dec_->weightoff)); break; case 2: MOVZX(32, 16, tempReg1, MDisp(srcReg, dec_->weightoff)); break; case 3: MOV(32, R(tempReg1), MDisp(srcReg, dec_->weightoff)); AND(32, R(tempReg1), Imm32(0x00FFFFFF)); break; case 4: MOV(32, R(tempReg1), MDisp(srcReg, dec_->weightoff)); break; case 5: MOV(32, R(tempReg1), MDisp(srcReg, dec_->weightoff)); MOVZX(32, 8, tempReg2, MDisp(srcReg, dec_->weightoff + 4)); break; case 6: MOV(32, R(tempReg1), MDisp(srcReg, dec_->weightoff)); MOVZX(32, 16, tempReg2, MDisp(srcReg, dec_->weightoff + 4)); break; case 7: MOV(32, R(tempReg1), MDisp(srcReg, dec_->weightoff)); MOV(32, R(tempReg2), MDisp(srcReg, dec_->weightoff + 4)); AND(32, R(tempReg2), Imm32(0x00FFFFFF)); break; case 8: MOV(32, R(tempReg1), MDisp(srcReg, dec_->weightoff)); MOV(32, R(tempReg2), MDisp(srcReg, dec_->weightoff + 4)); break; } if (dec_->nweights <= 4) { MOV(32, MDisp(dstReg, dec_->decFmt.w0off), R(tempReg1)); } else { MOV(32, MDisp(dstReg, dec_->decFmt.w0off), R(tempReg1)); MOV(32, MDisp(dstReg, dec_->decFmt.w1off), R(tempReg2)); } } void VertexDecoderJitCache::Jit_WeightsU16() { switch (dec_->nweights) { case 1: MOVZX(32, 16, tempReg1, MDisp(srcReg, dec_->weightoff)); MOV(32, MDisp(dstReg, dec_->decFmt.w0off), R(tempReg1)); MOV(32, MDisp(dstReg, dec_->decFmt.w0off + 4), Imm32(0)); return; case 2: MOV(32, R(tempReg1), MDisp(srcReg, dec_->weightoff)); MOV(32, MDisp(dstReg, dec_->decFmt.w0off), R(tempReg1)); MOV(32, MDisp(dstReg, dec_->decFmt.w0off + 4), Imm32(0)); return; case 3: MOV(32, R(tempReg1), MDisp(srcReg, dec_->weightoff)); MOVZX(32, 16, tempReg2, MDisp(srcReg, dec_->weightoff + 4)); MOV(32, MDisp(dstReg, dec_->decFmt.w0off), R(tempReg1)); MOV(32, MDisp(dstReg, dec_->decFmt.w0off + 4), R(tempReg2)); return; case 4: // Anything above 4 will do 4 here, and then the rest after. case 5: case 6: case 7: case 8: MOV(32, R(tempReg1), MDisp(srcReg, dec_->weightoff)); MOV(32, R(tempReg2), MDisp(srcReg, dec_->weightoff + 4)); MOV(32, MDisp(dstReg, dec_->decFmt.w0off), R(tempReg1)); MOV(32, MDisp(dstReg, dec_->decFmt.w0off + 4), R(tempReg2)); break; } // Basic implementation - a short at a time. TODO: Optimize int j; for (j = 4; j < dec_->nweights; j++) { MOV(16, R(tempReg1), MDisp(srcReg, dec_->weightoff + j * 2)); MOV(16, MDisp(dstReg, dec_->decFmt.w0off + j * 2), R(tempReg1)); } while (j & 3) { MOV(16, MDisp(dstReg, dec_->decFmt.w0off + j * 2), Imm16(0)); j++; } } void VertexDecoderJitCache::Jit_WeightsU8ToFloat() { if (dec_->nweights >= 4) { Jit_AnyU8ToFloat(dec_->weightoff, 32); MOVUPS(MDisp(dstReg, dec_->decFmt.w0off), XMM3); if (dec_->nweights > 4) { Jit_AnyU8ToFloat(dec_->weightoff + 4, (dec_->nweights - 4) * 8); MOVUPS(MDisp(dstReg, dec_->decFmt.w1off), XMM3); } } else { Jit_AnyU8ToFloat(dec_->weightoff, dec_->nweights * 8); MOVUPS(MDisp(dstReg, dec_->decFmt.w0off), XMM3); } } void VertexDecoderJitCache::Jit_WeightsU16ToFloat() { if (dec_->nweights >= 4) { Jit_AnyU16ToFloat(dec_->weightoff, 64); MOVUPS(MDisp(dstReg, dec_->decFmt.w0off), XMM3); if (dec_->nweights > 4) { Jit_AnyU16ToFloat(dec_->weightoff + 4 * 2, (dec_->nweights - 4) * 16); MOVUPS(MDisp(dstReg, dec_->decFmt.w1off), XMM3); } } else { Jit_AnyU16ToFloat(dec_->weightoff, dec_->nweights * 16); MOVUPS(MDisp(dstReg, dec_->decFmt.w0off), XMM3); } } void VertexDecoderJitCache::Jit_WeightsFloat() { int j; switch (dec_->nweights) { case 1: // MOVSS: When the source operand is a memory location and destination operand is an XMM register, the three high-order doublewords of the destination operand are cleared to all 0s. MOVSS(XMM3, MDisp(srcReg, dec_->weightoff)); MOVUPS(MDisp(dstReg, dec_->decFmt.w0off), XMM3); break; case 2: MOVQ_xmm(XMM3, MDisp(srcReg, dec_->weightoff)); MOVUPS(MDisp(dstReg, dec_->decFmt.w0off), XMM3); break; case 4: MOVUPS(XMM3, MDisp(srcReg, dec_->weightoff)); MOVUPS(MDisp(dstReg, dec_->decFmt.w0off), XMM3); break; case 5: MOVUPS(XMM3, MDisp(srcReg, dec_->weightoff)); MOVSS(XMM4, MDisp(srcReg, dec_->weightoff + 16)); MOVUPS(MDisp(dstReg, dec_->decFmt.w0off), XMM3); MOVUPS(MDisp(dstReg, dec_->decFmt.w0off + 16), XMM4); break; case 6: MOVUPS(XMM3, MDisp(srcReg, dec_->weightoff)); MOVQ_xmm(XMM4, MDisp(srcReg, dec_->weightoff + 16)); MOVUPS(MDisp(dstReg, dec_->decFmt.w0off), XMM3); MOVUPS(MDisp(dstReg, dec_->decFmt.w0off + 16), XMM4); break; case 8: MOVUPS(XMM3, MDisp(srcReg, dec_->weightoff)); MOVUPS(XMM4, MDisp(srcReg, dec_->weightoff + 16)); MOVUPS(MDisp(dstReg, dec_->decFmt.w0off), XMM3); MOVUPS(MDisp(dstReg, dec_->decFmt.w0off + 16), XMM4); break; default: for (j = 0; j < dec_->nweights; j++) { MOV(32, R(tempReg1), MDisp(srcReg, dec_->weightoff + j * 4)); MOV(32, MDisp(dstReg, dec_->decFmt.w0off + j * 4), R(tempReg1)); } while (j & 3) { // Zero additional weights rounding up to 4. MOV(32, MDisp(dstReg, dec_->decFmt.w0off + j * 4), Imm32(0)); j++; } break; } } void VertexDecoderJitCache::Jit_WeightsU8Skin() { MOV(PTRBITS, R(tempReg2), ImmPtr(&bones)); #if PPSSPP_ARCH(AMD64) if (dec_->nweights > 4) { // This reads 8 bytes, we split the top 4 so we can expand each set of 4. MOVQ_xmm(XMM8, MDisp(srcReg, dec_->weightoff)); PSHUFD(XMM9, R(XMM8), _MM_SHUFFLE(1, 1, 1, 1)); } else { MOVD_xmm(XMM8, MDisp(srcReg, dec_->weightoff)); } if (cpu_info.bSSE4_1) { PMOVZXBD(XMM8, R(XMM8)); } else { PXOR(fpScratchReg, R(fpScratchReg)); PUNPCKLBW(XMM8, R(fpScratchReg)); PUNPCKLWD(XMM8, R(fpScratchReg)); } if (dec_->nweights > 4) { if (cpu_info.bSSE4_1) { PMOVZXBD(XMM9, R(XMM9)); } else { PUNPCKLBW(XMM9, R(fpScratchReg)); PUNPCKLWD(XMM9, R(fpScratchReg)); } } CVTDQ2PS(XMM8, R(XMM8)); if (dec_->nweights > 4) CVTDQ2PS(XMM9, R(XMM9)); if (RipAccessible(&by128)) { MULPS(XMM8, M(&by128)); // rip accessible if (dec_->nweights > 4) MULPS(XMM9, M(&by128)); // rip accessible } else { MOV(PTRBITS, R(tempReg1), ImmPtr(&by128)); MULPS(XMM8, MatR(tempReg1)); if (dec_->nweights > 4) MULPS(XMM9, MatR(tempReg1)); } auto weightToAllLanes = [this](X64Reg dst, int lane) { X64Reg src = lane < 4 ? XMM8 : XMM9; if (dst != INVALID_REG && dst != src) { MOVAPS(dst, R(src)); } else { // INVALID_REG means ruin the existing src (it's not needed any more.) dst = src; } SHUFPS(dst, R(dst), _MM_SHUFFLE(lane % 4, lane % 4, lane % 4, lane % 4)); }; #endif for (int j = 0; j < dec_->nweights; j++) { X64Reg weight = XMM1; #if PPSSPP_ARCH(AMD64) X64Reg weightSrc = j < 4 ? XMM8 : XMM9; if (j == 3 || j == dec_->nweights - 1) { // In the previous iteration, we already spread this value to all lanes. weight = weightSrc; if (j == 0) { // If there's only the one weight, no one shuffled it for us yet. weightToAllLanes(weight, j); } // If we're on #3, prepare #4 if it's the last (and only for that reg, in fact.) if (j == dec_->nweights - 2) { weightToAllLanes(INVALID_REG, j + 1); } } else { weightToAllLanes(weight, j); // To improve latency, we shuffle in the last weight of the reg. // If we're on slot #2, slot #3 will be the last. Otherwise, nweights - 1 is last. if ((j == 2 && dec_->nweights > 3) || (j == dec_->nweights - 2)) { // Prepare the last one now for better latency. weightToAllLanes(INVALID_REG, j + 1); } } #else MOVZX(32, 8, tempReg1, MDisp(srcReg, dec_->weightoff + j)); CVTSI2SS(weight, R(tempReg1)); MULSS(weight, M(&by128)); // rip accessible (x86) SHUFPS(weight, R(weight), _MM_SHUFFLE(0, 0, 0, 0)); #endif if (j == 0) { MOVAPS(XMM4, MDisp(tempReg2, 0)); MOVAPS(XMM5, MDisp(tempReg2, 16)); MOVAPS(XMM6, MDisp(tempReg2, 32)); MOVAPS(XMM7, MDisp(tempReg2, 48)); MULPS(XMM4, R(weight)); MULPS(XMM5, R(weight)); MULPS(XMM6, R(weight)); MULPS(XMM7, R(weight)); } else { MOVAPS(XMM2, MDisp(tempReg2, 0)); MOVAPS(XMM3, MDisp(tempReg2, 16)); MULPS(XMM2, R(weight)); MULPS(XMM3, R(weight)); ADDPS(XMM4, R(XMM2)); ADDPS(XMM5, R(XMM3)); MOVAPS(XMM2, MDisp(tempReg2, 32)); MOVAPS(XMM3, MDisp(tempReg2, 48)); MULPS(XMM2, R(weight)); MULPS(XMM3, R(weight)); ADDPS(XMM6, R(XMM2)); ADDPS(XMM7, R(XMM3)); } ADD(PTRBITS, R(tempReg2), Imm8(4 * 16)); } } void VertexDecoderJitCache::Jit_WeightsU16Skin() { MOV(PTRBITS, R(tempReg2), ImmPtr(&bones)); #if PPSSPP_ARCH(AMD64) if (dec_->nweights > 6) { // Since this is probably not aligned, two MOVQs are better than one MOVDQU. MOVQ_xmm(XMM8, MDisp(srcReg, dec_->weightoff)); MOVQ_xmm(XMM9, MDisp(srcReg, dec_->weightoff + 8)); } else if (dec_->nweights > 4) { // Since this is probably not aligned, two MOVQs are better than one MOVDQU. MOVQ_xmm(XMM8, MDisp(srcReg, dec_->weightoff)); MOVD_xmm(XMM9, MDisp(srcReg, dec_->weightoff + 8)); } else if (dec_->nweights > 2) { MOVQ_xmm(XMM8, MDisp(srcReg, dec_->weightoff)); } else { MOVD_xmm(XMM8, MDisp(srcReg, dec_->weightoff)); } if (cpu_info.bSSE4_1) { PMOVZXWD(XMM8, R(XMM8)); } else { PXOR(fpScratchReg, R(fpScratchReg)); PUNPCKLWD(XMM8, R(fpScratchReg)); } if (dec_->nweights > 4) { if (cpu_info.bSSE4_1) { PMOVZXWD(XMM9, R(XMM9)); } else { PUNPCKLWD(XMM9, R(fpScratchReg)); } } CVTDQ2PS(XMM8, R(XMM8)); if (dec_->nweights > 4) CVTDQ2PS(XMM9, R(XMM9)); if (RipAccessible(&by32768)) { MULPS(XMM8, M(&by32768)); // rip accessible if (dec_->nweights > 4) MULPS(XMM9, M(&by32768)); // rip accessible } else { MOV(PTRBITS, R(tempReg1), ImmPtr(&by32768)); MULPS(XMM8, MatR(tempReg1)); if (dec_->nweights > 4) MULPS(XMM9, MatR(tempReg1)); } auto weightToAllLanes = [this](X64Reg dst, int lane) { X64Reg src = lane < 4 ? XMM8 : XMM9; if (dst != INVALID_REG && dst != src) { MOVAPS(dst, R(src)); } else { // INVALID_REG means ruin the existing src (it's not needed any more.) dst = src; } SHUFPS(dst, R(dst), _MM_SHUFFLE(lane % 4, lane % 4, lane % 4, lane % 4)); }; #endif for (int j = 0; j < dec_->nweights; j++) { X64Reg weight = XMM1; #if PPSSPP_ARCH(AMD64) X64Reg weightSrc = j < 4 ? XMM8 : XMM9; if (j == 3 || j == dec_->nweights - 1) { // In the previous iteration, we already spread this value to all lanes. weight = weightSrc; if (j == 0) { // If there's only the one weight, no one shuffled it for us yet. weightToAllLanes(weight, j); } // If we're on #3, prepare #4 if it's the last (and only for that reg, in fact.) if (j == dec_->nweights - 2) { weightToAllLanes(INVALID_REG, j + 1); } } else { weightToAllLanes(weight, j); // To improve latency, we shuffle in the last weight of the reg. // If we're on slot #2, slot #3 will be the last. Otherwise, nweights - 1 is last. if ((j == 2 && dec_->nweights > 3) || (j == dec_->nweights - 2)) { // Prepare the last one now for better latency. weightToAllLanes(INVALID_REG, j + 1); } } #else MOVZX(32, 16, tempReg1, MDisp(srcReg, dec_->weightoff + j * 2)); CVTSI2SS(weight, R(tempReg1)); MULSS(weight, M(&by32768)); // rip accessible (x86) SHUFPS(weight, R(weight), _MM_SHUFFLE(0, 0, 0, 0)); #endif if (j == 0) { MOVAPS(XMM4, MDisp(tempReg2, 0)); MOVAPS(XMM5, MDisp(tempReg2, 16)); MOVAPS(XMM6, MDisp(tempReg2, 32)); MOVAPS(XMM7, MDisp(tempReg2, 48)); MULPS(XMM4, R(weight)); MULPS(XMM5, R(weight)); MULPS(XMM6, R(weight)); MULPS(XMM7, R(weight)); } else { MOVAPS(XMM2, MDisp(tempReg2, 0)); MOVAPS(XMM3, MDisp(tempReg2, 16)); MULPS(XMM2, R(weight)); MULPS(XMM3, R(weight)); ADDPS(XMM4, R(XMM2)); ADDPS(XMM5, R(XMM3)); MOVAPS(XMM2, MDisp(tempReg2, 32)); MOVAPS(XMM3, MDisp(tempReg2, 48)); MULPS(XMM2, R(weight)); MULPS(XMM3, R(weight)); ADDPS(XMM6, R(XMM2)); ADDPS(XMM7, R(XMM3)); } ADD(PTRBITS, R(tempReg2), Imm8(4 * 16)); } } void VertexDecoderJitCache::Jit_WeightsFloatSkin() { MOV(PTRBITS, R(tempReg2), ImmPtr(&bones)); for (int j = 0; j < dec_->nweights; j++) { MOVSS(XMM1, MDisp(srcReg, dec_->weightoff + j * 4)); SHUFPS(XMM1, R(XMM1), _MM_SHUFFLE(0, 0, 0, 0)); if (j == 0) { MOVAPS(XMM4, MDisp(tempReg2, 0)); MOVAPS(XMM5, MDisp(tempReg2, 16)); MOVAPS(XMM6, MDisp(tempReg2, 32)); MOVAPS(XMM7, MDisp(tempReg2, 48)); MULPS(XMM4, R(XMM1)); MULPS(XMM5, R(XMM1)); MULPS(XMM6, R(XMM1)); MULPS(XMM7, R(XMM1)); } else { MOVAPS(XMM2, MDisp(tempReg2, 0)); MOVAPS(XMM3, MDisp(tempReg2, 16)); MULPS(XMM2, R(XMM1)); MULPS(XMM3, R(XMM1)); ADDPS(XMM4, R(XMM2)); ADDPS(XMM5, R(XMM3)); MOVAPS(XMM2, MDisp(tempReg2, 32)); MOVAPS(XMM3, MDisp(tempReg2, 48)); MULPS(XMM2, R(XMM1)); MULPS(XMM3, R(XMM1)); ADDPS(XMM6, R(XMM2)); ADDPS(XMM7, R(XMM3)); } ADD(PTRBITS, R(tempReg2), Imm8(4 * 16)); } } void VertexDecoderJitCache::Jit_TcU8ToFloat() { Jit_AnyU8ToFloat(dec_->tcoff, 16); MOVQ_xmm(MDisp(dstReg, dec_->decFmt.uvoff), XMM3); } void VertexDecoderJitCache::Jit_TcU16ToFloat() { Jit_AnyU16ToFloat(dec_->tcoff, 32); MOVQ_xmm(MDisp(dstReg, dec_->decFmt.uvoff), XMM3); } void VertexDecoderJitCache::Jit_TcFloat() { #if PPSSPP_ARCH(AMD64) MOV(64, R(tempReg1), MDisp(srcReg, dec_->tcoff)); MOV(64, MDisp(dstReg, dec_->decFmt.uvoff), R(tempReg1)); #else MOV(32, R(tempReg1), MDisp(srcReg, dec_->tcoff)); MOV(32, R(tempReg2), MDisp(srcReg, dec_->tcoff + 4)); MOV(32, MDisp(dstReg, dec_->decFmt.uvoff), R(tempReg1)); MOV(32, MDisp(dstReg, dec_->decFmt.uvoff + 4), R(tempReg2)); #endif } void VertexDecoderJitCache::Jit_TcU8Prescale() { // TODO: The first five instructions could be done in 1 or 2 in SSE4 MOVZX(32, 8, tempReg1, MDisp(srcReg, dec_->tcoff)); MOVZX(32, 8, tempReg2, MDisp(srcReg, dec_->tcoff + 1)); CVTSI2SS(fpScratchReg, R(tempReg1)); CVTSI2SS(fpScratchReg2, R(tempReg2)); UNPCKLPS(fpScratchReg, R(fpScratchReg2)); MULPS(fpScratchReg, R(fpScaleOffsetReg)); SHUFPS(fpScaleOffsetReg, R(fpScaleOffsetReg), _MM_SHUFFLE(1, 0, 3, 2)); ADDPS(fpScratchReg, R(fpScaleOffsetReg)); SHUFPS(fpScaleOffsetReg, R(fpScaleOffsetReg), _MM_SHUFFLE(1, 0, 3, 2)); MOVQ_xmm(MDisp(dstReg, dec_->decFmt.uvoff), fpScratchReg); } void VertexDecoderJitCache::Jit_TcU16Prescale() { PXOR(fpScratchReg2, R(fpScratchReg2)); MOVD_xmm(fpScratchReg, MDisp(srcReg, dec_->tcoff)); PUNPCKLWD(fpScratchReg, R(fpScratchReg2)); CVTDQ2PS(fpScratchReg, R(fpScratchReg)); MULPS(fpScratchReg, R(fpScaleOffsetReg)); SHUFPS(fpScaleOffsetReg, R(fpScaleOffsetReg), _MM_SHUFFLE(1, 0, 3, 2)); ADDPS(fpScratchReg, R(fpScaleOffsetReg)); SHUFPS(fpScaleOffsetReg, R(fpScaleOffsetReg), _MM_SHUFFLE(1, 0, 3, 2)); MOVQ_xmm(MDisp(dstReg, dec_->decFmt.uvoff), fpScratchReg); } void VertexDecoderJitCache::Jit_TcFloatPrescale() { MOVQ_xmm(fpScratchReg, MDisp(srcReg, dec_->tcoff)); MULPS(fpScratchReg, R(fpScaleOffsetReg)); SHUFPS(fpScaleOffsetReg, R(fpScaleOffsetReg), _MM_SHUFFLE(1, 0, 3, 2)); ADDPS(fpScratchReg, R(fpScaleOffsetReg)); SHUFPS(fpScaleOffsetReg, R(fpScaleOffsetReg), _MM_SHUFFLE(1, 0, 3, 2)); MOVQ_xmm(MDisp(dstReg, dec_->decFmt.uvoff), fpScratchReg); } void VertexDecoderJitCache::Jit_TcAnyMorph(int bits) { MOV(PTRBITS, R(tempReg1), ImmPtr(&gstate_c.morphWeights[0])); if (!cpu_info.bSSE4_1) { PXOR(fpScratchReg4, R(fpScratchReg4)); } bool first = true; for (int n = 0; n < dec_->morphcount; ++n) { const X64Reg reg = first ? fpScratchReg : fpScratchReg2; const OpArg src = MDisp(srcReg, dec_->onesize_ * n + dec_->tcoff); // Load the actual values and convert to float. if (bits == 32) { // Two floats: just load as a MOVQ. MOVQ_xmm(reg, src); } else { if (bits == 8) { MOVZX(32, 16, tempReg2, src); MOVD_xmm(reg, R(tempReg2)); } else { MOVD_xmm(reg, src); } if (cpu_info.bSSE4_1) { if (bits == 8) { PMOVZXBD(reg, R(reg)); } else { PMOVZXWD(reg, R(reg)); } } else { if (bits == 8) { PUNPCKLBW(reg, R(fpScratchReg4)); } PUNPCKLWD(reg, R(fpScratchReg4)); } CVTDQ2PS(reg, R(reg)); } // And now scale by the weight. MOVSS(fpScratchReg3, MDisp(tempReg1, n * sizeof(float))); SHUFPS(fpScratchReg3, R(fpScratchReg3), _MM_SHUFFLE(0, 0, 0, 0)); MULPS(reg, R(fpScratchReg3)); if (!first) { ADDPS(fpScratchReg, R(fpScratchReg2)); } else { first = false; } } } void VertexDecoderJitCache::Jit_TcU8MorphToFloat() { Jit_TcAnyMorph(8); // They were all added (weighted) pre-normalize, we normalize once here. if (RipAccessible(&by128)) { MULPS(fpScratchReg, M(&by128)); // rip accessible } else { MOV(PTRBITS, R(tempReg1), ImmPtr(&by128)); MULPS(fpScratchReg, MatR(tempReg1)); } MOVQ_xmm(MDisp(dstReg, dec_->decFmt.uvoff), fpScratchReg); } void VertexDecoderJitCache::Jit_TcU16MorphToFloat() { Jit_TcAnyMorph(16); // They were all added (weighted) pre-normalize, we normalize once here. if (RipAccessible(&by32768)) { MULPS(fpScratchReg, M(&by32768)); // rip accessible } else { MOV(PTRBITS, R(tempReg1), ImmPtr(&by32768)); MULPS(fpScratchReg, MatR(tempReg1)); } MOVQ_xmm(MDisp(dstReg, dec_->decFmt.uvoff), fpScratchReg); } void VertexDecoderJitCache::Jit_TcFloatMorph() { Jit_TcAnyMorph(32); MOVQ_xmm(MDisp(dstReg, dec_->decFmt.uvoff), fpScratchReg); } void VertexDecoderJitCache::Jit_TcU8PrescaleMorph() { Jit_TcAnyMorph(8); // The scale takes into account the u8 normalization. MULPS(fpScratchReg, R(fpScaleOffsetReg)); SHUFPS(fpScaleOffsetReg, R(fpScaleOffsetReg), _MM_SHUFFLE(1, 0, 3, 2)); ADDPS(fpScratchReg, R(fpScaleOffsetReg)); SHUFPS(fpScaleOffsetReg, R(fpScaleOffsetReg), _MM_SHUFFLE(1, 0, 3, 2)); MOVQ_xmm(MDisp(dstReg, dec_->decFmt.uvoff), fpScratchReg); } void VertexDecoderJitCache::Jit_TcU16PrescaleMorph() { Jit_TcAnyMorph(16); // The scale takes into account the u16 normalization. MULPS(fpScratchReg, R(fpScaleOffsetReg)); SHUFPS(fpScaleOffsetReg, R(fpScaleOffsetReg), _MM_SHUFFLE(1, 0, 3, 2)); ADDPS(fpScratchReg, R(fpScaleOffsetReg)); SHUFPS(fpScaleOffsetReg, R(fpScaleOffsetReg), _MM_SHUFFLE(1, 0, 3, 2)); MOVQ_xmm(MDisp(dstReg, dec_->decFmt.uvoff), fpScratchReg); } void VertexDecoderJitCache::Jit_TcFloatPrescaleMorph() { Jit_TcAnyMorph(32); MULPS(fpScratchReg, R(fpScaleOffsetReg)); SHUFPS(fpScaleOffsetReg, R(fpScaleOffsetReg), _MM_SHUFFLE(1, 0, 3, 2)); ADDPS(fpScratchReg, R(fpScaleOffsetReg)); SHUFPS(fpScaleOffsetReg, R(fpScaleOffsetReg), _MM_SHUFFLE(1, 0, 3, 2)); MOVQ_xmm(MDisp(dstReg, dec_->decFmt.uvoff), fpScratchReg); } void VertexDecoderJitCache::Jit_TcU16ThroughToFloat() { PXOR(fpScratchReg2, R(fpScratchReg2)); MOV(32, R(tempReg1), MDisp(srcReg, dec_->tcoff)); MOVD_xmm(fpScratchReg, R(tempReg1)); PUNPCKLWD(fpScratchReg, R(fpScratchReg2)); CVTDQ2PS(fpScratchReg, R(fpScratchReg)); MOVQ_xmm(MDisp(dstReg, dec_->decFmt.uvoff), fpScratchReg); MOV(32, R(tempReg2), R(tempReg1)); SHR(32, R(tempReg2), Imm8(16)); MOV(PTRBITS, R(tempReg3), ImmPtr(&gstate_c.vertBounds)); auto updateSide = [&](X64Reg r, CCFlags skipCC, int offset) { CMP(16, R(r), MDisp(tempReg3, offset)); FixupBranch skip = J_CC(skipCC); MOV(16, MDisp(tempReg3, offset), R(r)); SetJumpTarget(skip); }; // TODO: Can this actually be fast? Hmm, floats aren't better. updateSide(tempReg1, CC_GE, offsetof(KnownVertexBounds, minU)); updateSide(tempReg1, CC_LE, offsetof(KnownVertexBounds, maxU)); updateSide(tempReg2, CC_GE, offsetof(KnownVertexBounds, minV)); updateSide(tempReg2, CC_LE, offsetof(KnownVertexBounds, maxV)); } void VertexDecoderJitCache::Jit_TcFloatThrough() { #if PPSSPP_ARCH(AMD64) MOV(64, R(tempReg1), MDisp(srcReg, dec_->tcoff)); MOV(64, MDisp(dstReg, dec_->decFmt.uvoff), R(tempReg1)); #else MOV(32, R(tempReg1), MDisp(srcReg, dec_->tcoff)); MOV(32, R(tempReg2), MDisp(srcReg, dec_->tcoff + 4)); MOV(32, MDisp(dstReg, dec_->decFmt.uvoff), R(tempReg1)); MOV(32, MDisp(dstReg, dec_->decFmt.uvoff + 4), R(tempReg2)); #endif } void VertexDecoderJitCache::Jit_Color8888() { MOV(32, R(tempReg1), MDisp(srcReg, dec_->coloff)); MOV(32, MDisp(dstReg, dec_->decFmt.c0off), R(tempReg1)); CMP(32, R(tempReg1), Imm32(0xFF000000)); FixupBranch skip = J_CC(CC_AE, false); if (RipAccessible(&gstate_c.vertexFullAlpha)) { MOV(8, M(&gstate_c.vertexFullAlpha), Imm8(0)); // rip accessible } else { MOV(PTRBITS, R(tempReg1), ImmPtr(&gstate_c.vertexFullAlpha)); MOV(8, MatR(tempReg1), Imm8(0)); } SetJumpTarget(skip); } alignas(16) static const u32 color4444mask[4] = { 0xf00ff00f, 0xf00ff00f, 0xf00ff00f, 0xf00ff00f, }; void VertexDecoderJitCache::Jit_Color4444() { // This over-reads slightly, but we assume pos or another component follows anyway. MOVD_xmm(fpScratchReg, MDisp(srcReg, dec_->coloff)); // Spread to RGBA -> R00GB00A. PUNPCKLBW(fpScratchReg, R(fpScratchReg)); if (RipAccessible(&color4444mask[0])) { PAND(fpScratchReg, M(&color4444mask[0])); // rip accessible } else { MOV(PTRBITS, R(tempReg1), ImmPtr(&color4444mask)); PAND(fpScratchReg, MatR(tempReg1)); } MOVSS(fpScratchReg2, R(fpScratchReg)); MOVSS(fpScratchReg3, R(fpScratchReg)); // Create 0R000B00 and 00G000A0. PSRLW(fpScratchReg2, 4); PSLLW(fpScratchReg3, 4); // Combine for the complete set: RRGGBBAA. POR(fpScratchReg, R(fpScratchReg2)); POR(fpScratchReg, R(fpScratchReg3)); MOVD_xmm(R(tempReg1), fpScratchReg); MOV(32, MDisp(dstReg, dec_->decFmt.c0off), R(tempReg1)); CMP(32, R(tempReg1), Imm32(0xFF000000)); FixupBranch skip = J_CC(CC_AE, false); if (RipAccessible(&gstate_c.vertexFullAlpha)) { MOV(8, M(&gstate_c.vertexFullAlpha), Imm8(0)); // rip accessible } else { MOV(PTRBITS, R(tempReg1), ImmPtr(&gstate_c.vertexFullAlpha)); MOV(8, MatR(tempReg1), Imm8(0)); } SetJumpTarget(skip); } void VertexDecoderJitCache::Jit_Color565() { MOVZX(32, 16, tempReg1, MDisp(srcReg, dec_->coloff)); MOV(32, R(tempReg2), R(tempReg1)); AND(32, R(tempReg2), Imm32(0x0000001F)); // B (we do R and B at the same time, they're both 5.) MOV(32, R(tempReg3), R(tempReg1)); AND(32, R(tempReg3), Imm32(0x0000F800)); SHL(32, R(tempReg3), Imm8(5)); OR(32, R(tempReg2), R(tempReg3)); // Expand 5 -> 8. At this point we have 00BB00RR. MOV(32, R(tempReg3), R(tempReg2)); SHL(32, R(tempReg2), Imm8(3)); SHR(32, R(tempReg3), Imm8(2)); OR(32, R(tempReg2), R(tempReg3)); AND(32, R(tempReg2), Imm32(0x00FF00FF)); // Now's as good a time to put in A as any. OR(32, R(tempReg2), Imm32(0xFF000000)); // Last, we need to align, extract, and expand G. // 3 to align to G, and then 2 to expand to 8. SHL(32, R(tempReg1), Imm8(3 + 2)); AND(32, R(tempReg1), Imm32(0x0000FC00)); MOV(32, R(tempReg3), R(tempReg1)); // 2 to account for tempReg1 being preshifted, 4 for expansion. SHR(32, R(tempReg3), Imm8(2 + 4)); OR(32, R(tempReg1), R(tempReg3)); AND(32, R(tempReg1), Imm32(0x0000FF00)); OR(32, R(tempReg2), R(tempReg1)); MOV(32, MDisp(dstReg, dec_->decFmt.c0off), R(tempReg2)); // Never has alpha, no need to update fullAlphaArg. } void VertexDecoderJitCache::Jit_Color5551() { MOVZX(32, 16, tempReg1, MDisp(srcReg, dec_->coloff)); MOV(32, R(tempReg2), R(tempReg1)); MOV(32, R(tempReg3), R(tempReg1)); AND(32, R(tempReg2), Imm32(0x0000001F)); AND(32, R(tempReg3), Imm32(0x000003E0)); SHL(32, R(tempReg3), Imm8(3)); OR(32, R(tempReg2), R(tempReg3)); // This mask intentionally keeps alpha, but it's all the same bit so we can keep it. AND(32, R(tempReg1), Imm32(0x0000FC00)); SHL(32, R(tempReg1), Imm8(6)); // At this point, 0x003F1F1F are potentially set in tempReg2. OR(32, R(tempReg2), R(tempReg1)); // Expand 5 -> 8. After this is just expanding A. MOV(32, R(tempReg3), R(tempReg2)); SHL(32, R(tempReg2), Imm8(3)); SHR(32, R(tempReg3), Imm8(2)); // Chop off the bits that were shifted out (and also alpha.) AND(32, R(tempReg3), Imm32(0x00070707)); OR(32, R(tempReg2), R(tempReg3)); // For A, use sign extend to repeat the bit. B is still here, but it's the same bits. SHL(32, R(tempReg1), Imm8(10)); SAR(32, R(tempReg1), Imm8(7)); OR(32, R(tempReg2), R(tempReg1)); MOV(32, MDisp(dstReg, dec_->decFmt.c0off), R(tempReg2)); // Let's AND to avoid a branch, tempReg1 has alpha only in the top 8 bits. SHR(32, R(tempReg1), Imm8(24)); if (RipAccessible(&gstate_c.vertexFullAlpha)) { AND(8, M(&gstate_c.vertexFullAlpha), R(tempReg1)); // rip accessible } else { MOV(PTRBITS, R(tempReg3), ImmPtr(&gstate_c.vertexFullAlpha)); AND(8, MatR(tempReg3), R(tempReg1)); } } void VertexDecoderJitCache::Jit_Color8888Morph() { MOV(PTRBITS, R(tempReg1), ImmPtr(&gstate_c.morphWeights[0])); if (!cpu_info.bSSE4_1) { PXOR(fpScratchReg4, R(fpScratchReg4)); } bool first = true; for (int n = 0; n < dec_->morphcount; ++n) { const X64Reg reg = first ? fpScratchReg : fpScratchReg2; MOVD_xmm(reg, MDisp(srcReg, dec_->onesize_ * n + dec_->coloff)); if (cpu_info.bSSE4_1) { PMOVZXBD(reg, R(reg)); } else { PUNPCKLBW(reg, R(fpScratchReg4)); PUNPCKLWD(reg, R(fpScratchReg4)); } CVTDQ2PS(reg, R(reg)); // And now the weight. MOVSS(fpScratchReg3, MDisp(tempReg1, n * sizeof(float))); SHUFPS(fpScratchReg3, R(fpScratchReg3), _MM_SHUFFLE(0, 0, 0, 0)); MULPS(reg, R(fpScratchReg3)); if (!first) { ADDPS(fpScratchReg, R(fpScratchReg2)); } else { first = false; } } Jit_WriteMorphColor(dec_->decFmt.c0off); } alignas(16) static const float byColor4444[4] = { 255.0f / 15.0f, 255.0f / 15.0f, 255.0f / 15.0f, 255.0f / 15.0f, }; void VertexDecoderJitCache::Jit_Color4444Morph() { MOV(PTRBITS, R(tempReg1), ImmPtr(&gstate_c.morphWeights[0])); if (!cpu_info.bSSE4_1) { PXOR(fpScratchReg4, R(fpScratchReg4)); } MOV(PTRBITS, R(tempReg2), ImmPtr(color4444mask)); MOVDQA(XMM5, MatR(tempReg2)); MOV(PTRBITS, R(tempReg2), ImmPtr(byColor4444)); MOVAPS(XMM6, MatR(tempReg2)); bool first = true; for (int n = 0; n < dec_->morphcount; ++n) { const X64Reg reg = first ? fpScratchReg : fpScratchReg2; MOVD_xmm(reg, MDisp(srcReg, dec_->onesize_ * n + dec_->coloff)); PUNPCKLBW(reg, R(reg)); PAND(reg, R(XMM5)); MOVSS(fpScratchReg3, R(reg)); PSLLW(fpScratchReg3, 4); POR(reg, R(fpScratchReg3)); PSRLW(reg, 4); if (cpu_info.bSSE4_1) { PMOVZXBD(reg, R(reg)); } else { PUNPCKLBW(reg, R(fpScratchReg4)); PUNPCKLWD(reg, R(fpScratchReg4)); } CVTDQ2PS(reg, R(reg)); MULPS(reg, R(XMM6)); // And now the weight. MOVSS(fpScratchReg3, MDisp(tempReg1, n * sizeof(float))); SHUFPS(fpScratchReg3, R(fpScratchReg3), _MM_SHUFFLE(0, 0, 0, 0)); MULPS(reg, R(fpScratchReg3)); if (!first) { ADDPS(fpScratchReg, R(fpScratchReg2)); } else { first = false; } } Jit_WriteMorphColor(dec_->decFmt.c0off); } // The mask is intentionally in reverse order (but skips A.) alignas(16) static const u32 color565Mask[4] = { 0x0000f800, 0x000007e0, 0x0000001f, 0x00000000, }; alignas(16) static const float byColor565[4] = { 255.0f / 31.0f, 255.0f / 63.0f, 255.0f / 31.0f, 255.0f / 1.0f, }; void VertexDecoderJitCache::Jit_Color565Morph() { MOV(PTRBITS, R(tempReg1), ImmPtr(&gstate_c.morphWeights[0])); MOV(PTRBITS, R(tempReg2), ImmPtr(color565Mask)); MOVDQA(XMM5, MatR(tempReg2)); MOV(PTRBITS, R(tempReg2), ImmPtr(byColor565)); MOVAPS(XMM6, MatR(tempReg2)); bool first = true; for (int n = 0; n < dec_->morphcount; ++n) { const X64Reg reg = first ? fpScratchReg : fpScratchReg3; MOVD_xmm(fpScratchReg2, MDisp(srcReg, dec_->onesize_ * n + dec_->coloff)); // Spread it out into each lane. We end up with it reversed (R high, A low.) // Below, we shift out each lane from low to high and reverse them. PSHUFD(fpScratchReg2, R(fpScratchReg2), _MM_SHUFFLE(0, 0, 0, 0)); PAND(fpScratchReg2, R(XMM5)); // Alpha handled in Jit_WriteMorphColor. // Blue first. MOVSS(reg, R(fpScratchReg2)); PSRLD(reg, 6); PSHUFD(reg, R(reg), _MM_SHUFFLE(3, 0, 0, 0)); // Green, let's shift it into the right lane first. PSRLDQ(fpScratchReg2, 4); MOVSS(reg, R(fpScratchReg2)); PSRLD(reg, 5); PSHUFD(reg, R(reg), _MM_SHUFFLE(3, 2, 0, 0)); // Last one, red. PSRLDQ(fpScratchReg2, 4); MOVSS(reg, R(fpScratchReg2)); CVTDQ2PS(reg, R(reg)); MULPS(reg, R(XMM6)); // And now the weight. MOVSS(fpScratchReg2, MDisp(tempReg1, n * sizeof(float))); SHUFPS(fpScratchReg2, R(fpScratchReg2), _MM_SHUFFLE(0, 0, 0, 0)); MULPS(reg, R(fpScratchReg2)); if (!first) { ADDPS(fpScratchReg, R(fpScratchReg3)); } else { first = false; } } Jit_WriteMorphColor(dec_->decFmt.c0off, false); } // The mask is intentionally in reverse order. alignas(16) static const u32 color5551Mask[4] = { 0x00008000, 0x00007c00, 0x000003e0, 0x0000001f, }; alignas(16) static const float byColor5551[4] = { 255.0f / 31.0f, 255.0f / 31.0f, 255.0f / 31.0f, 255.0f / 1.0f, }; void VertexDecoderJitCache::Jit_Color5551Morph() { MOV(PTRBITS, R(tempReg1), ImmPtr(&gstate_c.morphWeights[0])); MOV(PTRBITS, R(tempReg2), ImmPtr(color5551Mask)); MOVDQA(XMM5, MatR(tempReg2)); MOV(PTRBITS, R(tempReg2), ImmPtr(byColor5551)); MOVAPS(XMM6, MatR(tempReg2)); bool first = true; for (int n = 0; n < dec_->morphcount; ++n) { const X64Reg reg = first ? fpScratchReg : fpScratchReg3; MOVD_xmm(fpScratchReg2, MDisp(srcReg, dec_->onesize_ * n + dec_->coloff)); // Spread it out into each lane. PSHUFD(fpScratchReg2, R(fpScratchReg2), _MM_SHUFFLE(0, 0, 0, 0)); PAND(fpScratchReg2, R(XMM5)); // Alpha first. MOVSS(reg, R(fpScratchReg2)); PSRLD(reg, 5); PSHUFD(reg, R(reg), _MM_SHUFFLE(0, 0, 0, 0)); // Blue, let's shift it into the right lane first. PSRLDQ(fpScratchReg2, 4); MOVSS(reg, R(fpScratchReg2)); PSRLD(reg, 5); PSHUFD(reg, R(reg), _MM_SHUFFLE(3, 0, 0, 0)); // Green. PSRLDQ(fpScratchReg2, 4); MOVSS(reg, R(fpScratchReg2)); PSRLD(reg, 5); PSHUFD(reg, R(reg), _MM_SHUFFLE(3, 2, 0, 0)); // Last one, red. PSRLDQ(fpScratchReg2, 4); MOVSS(reg, R(fpScratchReg2)); CVTDQ2PS(reg, R(reg)); MULPS(reg, R(XMM6)); // And now the weight. MOVSS(fpScratchReg2, MDisp(tempReg1, n * sizeof(float))); SHUFPS(fpScratchReg2, R(fpScratchReg2), _MM_SHUFFLE(0, 0, 0, 0)); MULPS(reg, R(fpScratchReg2)); if (!first) { ADDPS(fpScratchReg, R(fpScratchReg3)); } else { first = false; } } Jit_WriteMorphColor(dec_->decFmt.c0off); } void VertexDecoderJitCache::Jit_WriteMorphColor(int outOff, bool checkAlpha) { // Pack back into a u32, with saturation. CVTPS2DQ(fpScratchReg, R(fpScratchReg)); PACKSSDW(fpScratchReg, R(fpScratchReg)); PACKUSWB(fpScratchReg, R(fpScratchReg)); MOVD_xmm(R(tempReg1), fpScratchReg); // TODO: May be a faster way to do this without the MOVD. if (checkAlpha) { CMP(32, R(tempReg1), Imm32(0xFF000000)); FixupBranch skip = J_CC(CC_AE, false); if (RipAccessible(&gstate_c.vertexFullAlpha)) { MOV(8, M(&gstate_c.vertexFullAlpha), Imm8(0)); // rip accessible } else { MOV(PTRBITS, R(tempReg2), ImmPtr(&gstate_c.vertexFullAlpha)); MOV(8, MatR(tempReg2), Imm8(0)); } SetJumpTarget(skip); } else { // Force alpha to full if we're not checking it. OR(32, R(tempReg1), Imm32(0xFF000000)); } MOV(32, MDisp(dstReg, outOff), R(tempReg1)); } // Copy 3 bytes and then a zero. Might as well copy four. void VertexDecoderJitCache::Jit_NormalS8() { MOV(32, R(tempReg1), MDisp(srcReg, dec_->nrmoff)); AND(32, R(tempReg1), Imm32(0x00FFFFFF)); MOV(32, MDisp(dstReg, dec_->decFmt.nrmoff), R(tempReg1)); } void VertexDecoderJitCache::Jit_NormalS8ToFloat() { Jit_AnyS8ToFloat(dec_->nrmoff); MOVUPS(MDisp(dstReg, dec_->decFmt.nrmoff), XMM3); } // Copy 6 bytes and then 2 zeroes. void VertexDecoderJitCache::Jit_NormalS16() { MOV(32, R(tempReg1), MDisp(srcReg, dec_->nrmoff)); MOVZX(32, 16, tempReg2, MDisp(srcReg, dec_->nrmoff + 4)); MOV(32, MDisp(dstReg, dec_->decFmt.nrmoff), R(tempReg1)); MOV(32, MDisp(dstReg, dec_->decFmt.nrmoff + 4), R(tempReg2)); } void VertexDecoderJitCache::Jit_NormalFloat() { if (cpu_info.Mode64bit) { MOV(64, R(tempReg1), MDisp(srcReg, dec_->nrmoff)); MOV(32, R(tempReg3), MDisp(srcReg, dec_->nrmoff + 8)); MOV(64, MDisp(dstReg, dec_->decFmt.nrmoff), R(tempReg1)); MOV(32, MDisp(dstReg, dec_->decFmt.nrmoff + 8), R(tempReg3)); } else { MOV(32, R(tempReg1), MDisp(srcReg, dec_->nrmoff)); MOV(32, R(tempReg2), MDisp(srcReg, dec_->nrmoff + 4)); MOV(32, R(tempReg3), MDisp(srcReg, dec_->nrmoff + 8)); MOV(32, MDisp(dstReg, dec_->decFmt.nrmoff), R(tempReg1)); MOV(32, MDisp(dstReg, dec_->decFmt.nrmoff + 4), R(tempReg2)); MOV(32, MDisp(dstReg, dec_->decFmt.nrmoff + 8), R(tempReg3)); } } // This could be a bit shorter with AVX 3-operand instructions and FMA. void VertexDecoderJitCache::Jit_WriteMatrixMul(int outOff, bool pos) { MOVAPS(XMM1, R(XMM3)); MOVAPS(XMM2, R(XMM3)); SHUFPS(XMM1, R(XMM1), _MM_SHUFFLE(0, 0, 0, 0)); SHUFPS(XMM2, R(XMM2), _MM_SHUFFLE(1, 1, 1, 1)); SHUFPS(XMM3, R(XMM3), _MM_SHUFFLE(2, 2, 2, 2)); MULPS(XMM1, R(XMM4)); MULPS(XMM2, R(XMM5)); MULPS(XMM3, R(XMM6)); ADDPS(XMM1, R(XMM2)); ADDPS(XMM1, R(XMM3)); if (pos) { ADDPS(XMM1, R(XMM7)); } MOVUPS(MDisp(dstReg, outOff), XMM1); } void VertexDecoderJitCache::Jit_NormalS8Skin() { Jit_AnyS8ToFloat(dec_->nrmoff); Jit_WriteMatrixMul(dec_->decFmt.nrmoff, false); } void VertexDecoderJitCache::Jit_NormalS16Skin() { Jit_AnyS16ToFloat(dec_->nrmoff); Jit_WriteMatrixMul(dec_->decFmt.nrmoff, false); } void VertexDecoderJitCache::Jit_NormalFloatSkin() { MOVUPS(XMM3, MDisp(srcReg, dec_->nrmoff)); Jit_WriteMatrixMul(dec_->decFmt.nrmoff, false); } // Through expands into floats, always. Might want to look at changing this. void VertexDecoderJitCache::Jit_PosS8Through() { DEBUG_LOG_REPORT_ONCE(vertexS8Through, G3D, "Using S8 positions in throughmode"); // SIMD doesn't really matter since this isn't useful on hardware. for (int i = 0; i < 3; i++) { MOVSX(32, 8, tempReg1, MDisp(srcReg, dec_->posoff + i)); CVTSI2SS(fpScratchReg, R(tempReg1)); MOVSS(MDisp(dstReg, dec_->decFmt.posoff + i * 4), fpScratchReg); } } // Through expands into floats, always. Might want to look at changing this. void VertexDecoderJitCache::Jit_PosS16Through() { if (cpu_info.bSSE4_1) { MOVD_xmm(fpScratchReg, MDisp(srcReg, dec_->posoff)); MOVZX(32, 16, tempReg3, MDisp(srcReg, dec_->posoff + 4)); MOVD_xmm(fpScratchReg2, R(tempReg3)); PMOVSXWD(fpScratchReg, R(fpScratchReg)); PUNPCKLQDQ(fpScratchReg, R(fpScratchReg2)); CVTDQ2PS(fpScratchReg, R(fpScratchReg)); MOVUPS(MDisp(dstReg, dec_->decFmt.posoff), fpScratchReg); } else { MOVSX(32, 16, tempReg1, MDisp(srcReg, dec_->posoff)); MOVSX(32, 16, tempReg2, MDisp(srcReg, dec_->posoff + 2)); MOVZX(32, 16, tempReg3, MDisp(srcReg, dec_->posoff + 4)); // NOTE: MOVZX CVTSI2SS(fpScratchReg, R(tempReg1)); MOVSS(MDisp(dstReg, dec_->decFmt.posoff), fpScratchReg); CVTSI2SS(fpScratchReg, R(tempReg2)); MOVSS(MDisp(dstReg, dec_->decFmt.posoff + 4), fpScratchReg); CVTSI2SS(fpScratchReg, R(tempReg3)); MOVSS(MDisp(dstReg, dec_->decFmt.posoff + 8), fpScratchReg); } } void VertexDecoderJitCache::Jit_PosS8() { Jit_AnyS8ToFloat(dec_->posoff); MOVUPS(MDisp(dstReg, dec_->decFmt.posoff), XMM3); } void VertexDecoderJitCache::Jit_PosS16() { Jit_AnyS16ToFloat(dec_->posoff); MOVUPS(MDisp(dstReg, dec_->decFmt.posoff), XMM3); } // Just copy 12 bytes. void VertexDecoderJitCache::Jit_PosFloat() { if (cpu_info.Mode64bit) { MOV(64, R(tempReg1), MDisp(srcReg, dec_->posoff)); MOV(32, R(tempReg3), MDisp(srcReg, dec_->posoff + 8)); MOV(64, MDisp(dstReg, dec_->decFmt.posoff), R(tempReg1)); MOV(32, MDisp(dstReg, dec_->decFmt.posoff + 8), R(tempReg3)); } else { MOV(32, R(tempReg1), MDisp(srcReg, dec_->posoff)); MOV(32, R(tempReg2), MDisp(srcReg, dec_->posoff + 4)); MOV(32, R(tempReg3), MDisp(srcReg, dec_->posoff + 8)); MOV(32, MDisp(dstReg, dec_->decFmt.posoff), R(tempReg1)); MOV(32, MDisp(dstReg, dec_->decFmt.posoff + 4), R(tempReg2)); MOV(32, MDisp(dstReg, dec_->decFmt.posoff + 8), R(tempReg3)); } } void VertexDecoderJitCache::Jit_PosS8Skin() { Jit_AnyS8ToFloat(dec_->posoff); Jit_WriteMatrixMul(dec_->decFmt.posoff, true); } void VertexDecoderJitCache::Jit_PosS16Skin() { Jit_AnyS16ToFloat(dec_->posoff); Jit_WriteMatrixMul(dec_->decFmt.posoff, true); } void VertexDecoderJitCache::Jit_PosFloatSkin() { MOVUPS(XMM3, MDisp(srcReg, dec_->posoff)); Jit_WriteMatrixMul(dec_->decFmt.posoff, true); } void VertexDecoderJitCache::Jit_AnyS8ToFloat(int srcoff) { if (!cpu_info.bSSE4_1) { PXOR(XMM3, R(XMM3)); } MOVD_xmm(XMM1, MDisp(srcReg, srcoff)); if (cpu_info.bSSE4_1) { PMOVSXBD(XMM1, R(XMM1)); } else { PUNPCKLBW(XMM1, R(XMM3)); PUNPCKLWD(XMM1, R(XMM3)); PSLLD(XMM1, 24); PSRAD(XMM1, 24); } CVTDQ2PS(XMM3, R(XMM1)); if (RipAccessible(&by128)) { MULPS(XMM3, M(&by128)); // rip accessible } else { MOV(PTRBITS, R(tempReg1), ImmPtr(&by128)); MULPS(XMM3, MatR(tempReg1)); } } void VertexDecoderJitCache::Jit_AnyS16ToFloat(int srcoff) { if (!cpu_info.bSSE4_1) { PXOR(XMM3, R(XMM3)); } MOVQ_xmm(XMM1, MDisp(srcReg, srcoff)); if (cpu_info.bSSE4_1) { PMOVSXWD(XMM1, R(XMM1)); } else { PUNPCKLWD(XMM1, R(XMM3)); PSLLD(XMM1, 16); PSRAD(XMM1, 16); } CVTDQ2PS(XMM3, R(XMM1)); if (RipAccessible(&by32768)) { MULPS(XMM3, M(&by32768)); // rip accessible } else { MOV(PTRBITS, R(tempReg1), ImmPtr(&by32768)); MULPS(XMM3, MatR(tempReg1)); } } void VertexDecoderJitCache::Jit_AnyU8ToFloat(int srcoff, u32 bits) { _dbg_assert_msg_((bits & ~(32 | 16 | 8)) == 0, "Bits must be a multiple of 8."); _dbg_assert_msg_(bits >= 8 && bits <= 32, "Bits must be a between 8 and 32."); if (!cpu_info.bSSE4_1) { PXOR(XMM3, R(XMM3)); } if (bits == 32) { MOVD_xmm(XMM1, MDisp(srcReg, srcoff)); } else if (bits == 24) { MOV(32, R(tempReg1), MDisp(srcReg, srcoff)); AND(32, R(tempReg1), Imm32(0x00FFFFFF)); MOVD_xmm(XMM1, R(tempReg1)); } else { MOVZX(32, bits, tempReg1, MDisp(srcReg, srcoff)); MOVD_xmm(XMM1, R(tempReg1)); } if (cpu_info.bSSE4_1) { PMOVZXBD(XMM1, R(XMM1)); } else { PUNPCKLBW(XMM1, R(XMM3)); PUNPCKLWD(XMM1, R(XMM3)); } CVTDQ2PS(XMM3, R(XMM1)); if (RipAccessible(&by128)) { MULPS(XMM3, M(&by128)); // rip accessible } else { MOV(PTRBITS, R(tempReg1), ImmPtr(&by128)); MULPS(XMM3, MatR(tempReg1)); } } void VertexDecoderJitCache::Jit_AnyU16ToFloat(int srcoff, u32 bits) { _dbg_assert_msg_((bits & ~(64 | 32 | 16)) == 0, "Bits must be a multiple of 16."); _dbg_assert_msg_(bits >= 16 && bits <= 64, "Bits must be a between 16 and 64."); if (!cpu_info.bSSE4_1) { PXOR(XMM3, R(XMM3)); } if (bits == 64) { MOVQ_xmm(XMM1, MDisp(srcReg, srcoff)); } else if (bits == 48) { MOVD_xmm(XMM1, MDisp(srcReg, srcoff)); PINSRW(XMM1, MDisp(srcReg, srcoff + 4), 2); } else if (bits == 32) { MOVD_xmm(XMM1, MDisp(srcReg, srcoff)); } else if (bits == 16) { MOVZX(32, bits, tempReg1, MDisp(srcReg, srcoff)); MOVD_xmm(XMM1, R(tempReg1)); } if (cpu_info.bSSE4_1) { PMOVZXWD(XMM1, R(XMM1)); } else { PUNPCKLWD(XMM1, R(XMM3)); } CVTDQ2PS(XMM3, R(XMM1)); if (RipAccessible(&by32768)) { MULPS(XMM3, M(&by32768)); // rip accessible } else { MOV(PTRBITS, R(tempReg1), ImmPtr(&by32768)); MULPS(XMM3, MatR(tempReg1)); } } void VertexDecoderJitCache::Jit_AnyS8Morph(int srcoff, int dstoff) { MOV(PTRBITS, R(tempReg1), ImmPtr(&gstate_c.morphWeights[0])); if (!cpu_info.bSSE4_1) { PXOR(fpScratchReg4, R(fpScratchReg4)); } if (RipAccessible(&by128)) { MOVAPS(XMM5, M(&by128)); // rip accessible } else { MOV(PTRBITS, R(tempReg1), ImmPtr(&by128)); MOVAPS(XMM5, MatR(tempReg1)); } // Sum into fpScratchReg. bool first = true; for (int n = 0; n < dec_->morphcount; ++n) { const X64Reg reg = first ? fpScratchReg : fpScratchReg2; // Okay, first convert to floats. MOVD_xmm(reg, MDisp(srcReg, dec_->onesize_ * n + srcoff)); if (cpu_info.bSSE4_1) { PMOVSXBD(reg, R(reg)); } else { PUNPCKLBW(reg, R(fpScratchReg4)); PUNPCKLWD(reg, R(fpScratchReg4)); PSLLD(reg, 24); PSRAD(reg, 24); } CVTDQ2PS(reg, R(reg)); // Now, It's time to multiply by the weight and 1.0f/128.0f. MOVSS(fpScratchReg3, MDisp(tempReg1, sizeof(float) * n)); MULSS(fpScratchReg3, R(XMM5)); SHUFPS(fpScratchReg3, R(fpScratchReg3), _MM_SHUFFLE(0, 0, 0, 0)); MULPS(reg, R(fpScratchReg3)); if (!first) { ADDPS(fpScratchReg, R(fpScratchReg2)); } else { first = false; } } MOVUPS(MDisp(dstReg, dstoff), fpScratchReg); } void VertexDecoderJitCache::Jit_AnyS16Morph(int srcoff, int dstoff) { MOV(PTRBITS, R(tempReg1), ImmPtr(&gstate_c.morphWeights[0])); if (!cpu_info.bSSE4_1) { PXOR(fpScratchReg4, R(fpScratchReg4)); } if (RipAccessible(&by32768)) { MOVAPS(XMM5, M(&by32768)); // rip accessible } else { MOV(PTRBITS, R(tempReg1), ImmPtr(&by32768)); MOVAPS(XMM5, MatR(tempReg1)); } // Sum into fpScratchReg. bool first = true; for (int n = 0; n < dec_->morphcount; ++n) { const X64Reg reg = first ? fpScratchReg : fpScratchReg2; // Okay, first convert to floats. MOVQ_xmm(reg, MDisp(srcReg, dec_->onesize_ * n + srcoff)); if (cpu_info.bSSE4_1) { PMOVSXWD(reg, R(reg)); } else { PUNPCKLWD(reg, R(fpScratchReg4)); PSLLD(reg, 16); PSRAD(reg, 16); } CVTDQ2PS(reg, R(reg)); // Now, It's time to multiply by the weight and 1.0f/32768.0f. MOVSS(fpScratchReg3, MDisp(tempReg1, sizeof(float) * n)); MULSS(fpScratchReg3, R(XMM5)); SHUFPS(fpScratchReg3, R(fpScratchReg3), _MM_SHUFFLE(0, 0, 0, 0)); MULPS(reg, R(fpScratchReg3)); if (!first) { ADDPS(fpScratchReg, R(fpScratchReg2)); } else { first = false; } } MOVUPS(MDisp(dstReg, dstoff), fpScratchReg); } void VertexDecoderJitCache::Jit_AnyFloatMorph(int srcoff, int dstoff) { MOV(PTRBITS, R(tempReg1), ImmPtr(&gstate_c.morphWeights[0])); // Sum into fpScratchReg. bool first = true; for (int n = 0; n < dec_->morphcount; ++n) { const X64Reg reg = first ? fpScratchReg : fpScratchReg2; MOVUPS(reg, MDisp(srcReg, dec_->onesize_ * n + srcoff)); MOVSS(fpScratchReg3, MDisp(tempReg1, sizeof(float) * n)); SHUFPS(fpScratchReg3, R(fpScratchReg3), _MM_SHUFFLE(0, 0, 0, 0)); MULPS(reg, R(fpScratchReg3)); if (!first) { ADDPS(fpScratchReg, R(fpScratchReg2)); } else { first = false; } } MOVUPS(MDisp(dstReg, dstoff), fpScratchReg); } void VertexDecoderJitCache::Jit_PosS8Morph() { Jit_AnyS8Morph(dec_->posoff, dec_->decFmt.posoff); } void VertexDecoderJitCache::Jit_PosS16Morph() { Jit_AnyS16Morph(dec_->posoff, dec_->decFmt.posoff); } void VertexDecoderJitCache::Jit_PosFloatMorph() { Jit_AnyFloatMorph(dec_->posoff, dec_->decFmt.posoff); } void VertexDecoderJitCache::Jit_NormalS8Morph() { Jit_AnyS8Morph(dec_->nrmoff, dec_->decFmt.nrmoff); } void VertexDecoderJitCache::Jit_NormalS16Morph() { Jit_AnyS16Morph(dec_->nrmoff, dec_->decFmt.nrmoff); } void VertexDecoderJitCache::Jit_NormalFloatMorph() { Jit_AnyFloatMorph(dec_->nrmoff, dec_->decFmt.nrmoff); } bool VertexDecoderJitCache::CompileStep(const VertexDecoder &dec, int step) { // See if we find a matching JIT function for (size_t i = 0; i < ARRAY_SIZE(jitLookup); i++) { if (dec.steps_[step] == jitLookup[i].func) { ((*this).*jitLookup[i].jitFunc)(); return true; } } return false; } #endif // PPSSPP_ARCH(X86) || PPSSPP_ARCH(AMD64)