darling-JavaScriptCore/jit/JITArithmetic.cpp

1254 lines
47 KiB
C++
Raw Normal View History

2017-08-12 16:48:01 +00:00
/*
2020-08-29 13:27:11 +00:00
* Copyright (C) 2008-2019 Apple Inc. All rights reserved.
2017-08-12 16:48:01 +00:00
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY APPLE INC. ``AS IS'' AND ANY
* EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR
* CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
* EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
* PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
* PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY
* OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#include "config.h"
#if ENABLE(JIT)
#include "JIT.h"
#include "ArithProfile.h"
#include "BytecodeGenerator.h"
2017-08-12 16:48:01 +00:00
#include "CodeBlock.h"
#include "JITBitAndGenerator.h"
#include "JITBitOrGenerator.h"
#include "JITBitXorGenerator.h"
#include "JITDivGenerator.h"
#include "JITInlines.h"
#include "JITLeftShiftGenerator.h"
#include "JITMathIC.h"
#include "JITOperations.h"
#include "ResultType.h"
#include "SlowPathCall.h"
namespace JSC {
2020-08-29 13:27:11 +00:00
void JIT::emit_op_jless(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
emit_compareAndJump<OpJless>(currentInstruction, LessThan);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emit_op_jlesseq(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
emit_compareAndJump<OpJlesseq>(currentInstruction, LessThanOrEqual);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emit_op_jgreater(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
emit_compareAndJump<OpJgreater>(currentInstruction, GreaterThan);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emit_op_jgreatereq(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
emit_compareAndJump<OpJgreatereq>(currentInstruction, GreaterThanOrEqual);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emit_op_jnless(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
emit_compareAndJump<OpJnless>(currentInstruction, GreaterThanOrEqual);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emit_op_jnlesseq(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
emit_compareAndJump<OpJnlesseq>(currentInstruction, GreaterThan);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emit_op_jngreater(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
emit_compareAndJump<OpJngreater>(currentInstruction, LessThanOrEqual);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emit_op_jngreatereq(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
emit_compareAndJump<OpJngreatereq>(currentInstruction, LessThan);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emitSlow_op_jless(const Instruction* currentInstruction, Vector<SlowCaseEntry>::iterator& iter)
2017-08-12 16:48:01 +00:00
{
emit_compareAndJumpSlow<OpJless>(currentInstruction, DoubleLessThanAndOrdered, operationCompareLess, false, iter);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emitSlow_op_jlesseq(const Instruction* currentInstruction, Vector<SlowCaseEntry>::iterator& iter)
2017-08-12 16:48:01 +00:00
{
emit_compareAndJumpSlow<OpJlesseq>(currentInstruction, DoubleLessThanOrEqualAndOrdered, operationCompareLessEq, false, iter);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emitSlow_op_jgreater(const Instruction* currentInstruction, Vector<SlowCaseEntry>::iterator& iter)
2017-08-12 16:48:01 +00:00
{
emit_compareAndJumpSlow<OpJgreater>(currentInstruction, DoubleGreaterThanAndOrdered, operationCompareGreater, false, iter);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emitSlow_op_jgreatereq(const Instruction* currentInstruction, Vector<SlowCaseEntry>::iterator& iter)
2017-08-12 16:48:01 +00:00
{
emit_compareAndJumpSlow<OpJgreatereq>(currentInstruction, DoubleGreaterThanOrEqualAndOrdered, operationCompareGreaterEq, false, iter);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emitSlow_op_jnless(const Instruction* currentInstruction, Vector<SlowCaseEntry>::iterator& iter)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
emit_compareAndJumpSlow<OpJnless>(currentInstruction, DoubleGreaterThanOrEqualOrUnordered, operationCompareLess, true, iter);
}
2017-08-12 16:48:01 +00:00
2020-08-29 13:27:11 +00:00
void JIT::emitSlow_op_jnlesseq(const Instruction* currentInstruction, Vector<SlowCaseEntry>::iterator& iter)
{
emit_compareAndJumpSlow<OpJnlesseq>(currentInstruction, DoubleGreaterThanOrUnordered, operationCompareLessEq, true, iter);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emitSlow_op_jngreater(const Instruction* currentInstruction, Vector<SlowCaseEntry>::iterator& iter)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
emit_compareAndJumpSlow<OpJngreater>(currentInstruction, DoubleLessThanOrEqualOrUnordered, operationCompareGreater, true, iter);
}
2017-08-12 16:48:01 +00:00
2020-08-29 13:27:11 +00:00
void JIT::emitSlow_op_jngreatereq(const Instruction* currentInstruction, Vector<SlowCaseEntry>::iterator& iter)
{
emit_compareAndJumpSlow<OpJngreatereq>(currentInstruction, DoubleLessThanOrUnordered, operationCompareGreaterEq, true, iter);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emit_op_below(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
emit_compareUnsigned<OpBelow>(currentInstruction, Below);
}
2017-08-12 16:48:01 +00:00
2020-08-29 13:27:11 +00:00
void JIT::emit_op_beloweq(const Instruction* currentInstruction)
{
emit_compareUnsigned<OpBeloweq>(currentInstruction, BelowOrEqual);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emit_op_jbelow(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
emit_compareUnsignedAndJump<OpJbelow>(currentInstruction, Below);
}
2017-08-12 16:48:01 +00:00
2020-08-29 13:27:11 +00:00
void JIT::emit_op_jbeloweq(const Instruction* currentInstruction)
{
emit_compareUnsignedAndJump<OpJbeloweq>(currentInstruction, BelowOrEqual);
2017-08-12 16:48:01 +00:00
}
#if USE(JSVALUE64)
2020-08-29 13:27:11 +00:00
void JIT::emit_op_unsigned(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
auto bytecode = currentInstruction->as<OpUnsigned>();
VirtualRegister result = bytecode.m_dst;
VirtualRegister op1 = bytecode.m_operand;
2017-08-12 16:48:01 +00:00
emitGetVirtualRegister(op1, regT0);
emitJumpSlowCaseIfNotInt(regT0);
addSlowCase(branch32(LessThan, regT0, TrustedImm32(0)));
2020-08-29 13:27:11 +00:00
boxInt32(regT0, JSValueRegs { regT0 });
2017-08-12 16:48:01 +00:00
emitPutVirtualRegister(result, regT0);
}
2020-08-29 13:27:11 +00:00
template<typename Op>
void JIT::emit_compareAndJump(const Instruction* instruction, RelationalCondition condition)
{
auto bytecode = instruction->as<Op>();
VirtualRegister op1 = bytecode.m_lhs;
VirtualRegister op2 = bytecode.m_rhs;
unsigned target = jumpTarget(instruction, bytecode.m_targetLabel);
emit_compareAndJumpImpl(op1, op2, target, condition);
}
void JIT::emit_compareAndJumpImpl(VirtualRegister op1, VirtualRegister op2, unsigned target, RelationalCondition condition)
2017-08-12 16:48:01 +00:00
{
// We generate inline code for the following cases in the fast path:
// - int immediate to constant int immediate
// - constant int immediate to int immediate
// - int immediate to int immediate
2020-08-29 13:27:11 +00:00
bool disallowAllocation = false;
2017-08-12 16:48:01 +00:00
if (isOperandConstantChar(op1)) {
emitGetVirtualRegister(op2, regT0);
2020-08-29 13:27:11 +00:00
addSlowCase(branchIfNotCell(regT0));
2017-08-12 16:48:01 +00:00
JumpList failures;
emitLoadCharacterString(regT0, regT0, failures);
addSlowCase(failures);
2020-08-29 13:27:11 +00:00
addJump(branch32(commute(condition), regT0, Imm32(asString(getConstantOperand(op1))->tryGetValue(disallowAllocation)[0])), target);
2017-08-12 16:48:01 +00:00
return;
}
if (isOperandConstantChar(op2)) {
emitGetVirtualRegister(op1, regT0);
2020-08-29 13:27:11 +00:00
addSlowCase(branchIfNotCell(regT0));
2017-08-12 16:48:01 +00:00
JumpList failures;
emitLoadCharacterString(regT0, regT0, failures);
addSlowCase(failures);
2020-08-29 13:27:11 +00:00
addJump(branch32(condition, regT0, Imm32(asString(getConstantOperand(op2))->tryGetValue(disallowAllocation)[0])), target);
2017-08-12 16:48:01 +00:00
return;
}
if (isOperandConstantInt(op2)) {
emitGetVirtualRegister(op1, regT0);
emitJumpSlowCaseIfNotInt(regT0);
int32_t op2imm = getOperandConstantInt(op2);
addJump(branch32(condition, regT0, Imm32(op2imm)), target);
2020-08-29 13:27:11 +00:00
return;
}
if (isOperandConstantInt(op1)) {
2017-08-12 16:48:01 +00:00
emitGetVirtualRegister(op2, regT1);
emitJumpSlowCaseIfNotInt(regT1);
int32_t op1imm = getOperandConstantInt(op1);
addJump(branch32(commute(condition), regT1, Imm32(op1imm)), target);
2020-08-29 13:27:11 +00:00
return;
}
emitGetVirtualRegisters(op1, regT0, op2, regT1);
emitJumpSlowCaseIfNotInt(regT0);
emitJumpSlowCaseIfNotInt(regT1);
addJump(branch32(condition, regT0, regT1), target);
}
template<typename Op>
void JIT::emit_compareUnsignedAndJump(const Instruction* instruction, RelationalCondition condition)
{
auto bytecode = instruction->as<Op>();
VirtualRegister op1 = bytecode.m_lhs;
VirtualRegister op2 = bytecode.m_rhs;
2020-08-29 13:27:11 +00:00
unsigned target = jumpTarget(instruction, bytecode.m_targetLabel);
emit_compareUnsignedAndJumpImpl(op1, op2, target, condition);
}
void JIT::emit_compareUnsignedAndJumpImpl(VirtualRegister op1, VirtualRegister op2, unsigned target, RelationalCondition condition)
{
2020-08-29 13:27:11 +00:00
if (isOperandConstantInt(op2)) {
emitGetVirtualRegister(op1, regT0);
int32_t op2imm = getOperandConstantInt(op2);
addJump(branch32(condition, regT0, Imm32(op2imm)), target);
} else if (isOperandConstantInt(op1)) {
emitGetVirtualRegister(op2, regT1);
int32_t op1imm = getOperandConstantInt(op1);
addJump(branch32(commute(condition), regT1, Imm32(op1imm)), target);
2017-08-12 16:48:01 +00:00
} else {
emitGetVirtualRegisters(op1, regT0, op2, regT1);
addJump(branch32(condition, regT0, regT1), target);
}
}
2020-08-29 13:27:11 +00:00
template<typename Op>
void JIT::emit_compareUnsigned(const Instruction* instruction, RelationalCondition condition)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
auto bytecode = instruction->as<Op>();
VirtualRegister dst = bytecode.m_dst;
VirtualRegister op1 = bytecode.m_lhs;
VirtualRegister op2 = bytecode.m_rhs;
emit_compareUnsignedImpl(dst, op1, op2, condition);
}
void JIT::emit_compareUnsignedImpl(VirtualRegister dst, VirtualRegister op1, VirtualRegister op2, RelationalCondition condition)
{
2020-08-29 13:27:11 +00:00
if (isOperandConstantInt(op2)) {
emitGetVirtualRegister(op1, regT0);
int32_t op2imm = getOperandConstantInt(op2);
compare32(condition, regT0, Imm32(op2imm), regT0);
} else if (isOperandConstantInt(op1)) {
emitGetVirtualRegister(op2, regT0);
int32_t op1imm = getOperandConstantInt(op1);
compare32(commute(condition), regT0, Imm32(op1imm), regT0);
} else {
emitGetVirtualRegisters(op1, regT0, op2, regT1);
compare32(condition, regT0, regT1, regT0);
}
boxBoolean(regT0, JSValueRegs { regT0 });
emitPutVirtualRegister(dst);
}
template<typename Op, typename SlowOperation>
void JIT::emit_compareAndJumpSlow(const Instruction* instruction, DoubleCondition condition, SlowOperation operation, bool invert, Vector<SlowCaseEntry>::iterator& iter)
2020-08-29 13:27:11 +00:00
{
auto bytecode = instruction->as<Op>();
VirtualRegister op1 = bytecode.m_lhs;
VirtualRegister op2 = bytecode.m_rhs;
2020-08-29 13:27:11 +00:00
unsigned target = jumpTarget(instruction, bytecode.m_targetLabel);
emit_compareAndJumpSlowImpl(op1, op2, target, instruction->size(), condition, operation, invert, iter);
}
template<typename SlowOperation>
void JIT::emit_compareAndJumpSlowImpl(VirtualRegister op1, VirtualRegister op2, unsigned target, size_t instructionSize, DoubleCondition condition, SlowOperation operation, bool invert, Vector<SlowCaseEntry>::iterator& iter)
{
2020-08-29 13:27:11 +00:00
2017-08-12 16:48:01 +00:00
// We generate inline code for the following cases in the slow path:
// - floating-point number to constant int immediate
// - constant int immediate to floating-point number
// - floating-point number to floating-point number.
if (isOperandConstantChar(op1) || isOperandConstantChar(op2)) {
2020-08-29 13:27:11 +00:00
linkAllSlowCases(iter);
2017-08-12 16:48:01 +00:00
emitGetVirtualRegister(op1, argumentGPR0);
emitGetVirtualRegister(op2, argumentGPR1);
callOperation(operation, TrustedImmPtr(m_codeBlock->globalObject()), argumentGPR0, argumentGPR1);
2017-08-12 16:48:01 +00:00
emitJumpSlowToHot(branchTest32(invert ? Zero : NonZero, returnValueGPR), target);
return;
}
if (isOperandConstantInt(op2)) {
2020-08-29 13:27:11 +00:00
linkAllSlowCases(iter);
2017-08-12 16:48:01 +00:00
if (supportsFloatingPoint()) {
2020-08-29 13:27:11 +00:00
Jump fail1 = branchIfNotNumber(regT0);
add64(numberTagRegister, regT0);
2017-08-12 16:48:01 +00:00
move64ToDouble(regT0, fpRegT0);
int32_t op2imm = getConstantOperand(op2).asInt32();
move(Imm32(op2imm), regT1);
convertInt32ToDouble(regT1, fpRegT1);
emitJumpSlowToHot(branchDouble(condition, fpRegT0, fpRegT1), target);
emitJumpSlowToHot(jump(), instructionSize);
2017-08-12 16:48:01 +00:00
fail1.link(this);
}
emitGetVirtualRegister(op2, regT1);
callOperation(operation, TrustedImmPtr(m_codeBlock->globalObject()), regT0, regT1);
2017-08-12 16:48:01 +00:00
emitJumpSlowToHot(branchTest32(invert ? Zero : NonZero, returnValueGPR), target);
2020-08-29 13:27:11 +00:00
return;
}
if (isOperandConstantInt(op1)) {
linkAllSlowCases(iter);
2017-08-12 16:48:01 +00:00
if (supportsFloatingPoint()) {
2020-08-29 13:27:11 +00:00
Jump fail1 = branchIfNotNumber(regT1);
add64(numberTagRegister, regT1);
2017-08-12 16:48:01 +00:00
move64ToDouble(regT1, fpRegT1);
int32_t op1imm = getConstantOperand(op1).asInt32();
move(Imm32(op1imm), regT0);
convertInt32ToDouble(regT0, fpRegT0);
emitJumpSlowToHot(branchDouble(condition, fpRegT0, fpRegT1), target);
emitJumpSlowToHot(jump(), instructionSize);
2017-08-12 16:48:01 +00:00
fail1.link(this);
}
emitGetVirtualRegister(op1, regT2);
callOperation(operation, TrustedImmPtr(m_codeBlock->globalObject()), regT2, regT1);
2017-08-12 16:48:01 +00:00
emitJumpSlowToHot(branchTest32(invert ? Zero : NonZero, returnValueGPR), target);
2020-08-29 13:27:11 +00:00
return;
}
2017-08-12 16:48:01 +00:00
2020-08-29 13:27:11 +00:00
linkSlowCase(iter); // LHS is not Int.
2017-08-12 16:48:01 +00:00
2020-08-29 13:27:11 +00:00
if (supportsFloatingPoint()) {
Jump fail1 = branchIfNotNumber(regT0);
Jump fail2 = branchIfNotNumber(regT1);
Jump fail3 = branchIfInt32(regT1);
add64(numberTagRegister, regT0);
add64(numberTagRegister, regT1);
2020-08-29 13:27:11 +00:00
move64ToDouble(regT0, fpRegT0);
move64ToDouble(regT1, fpRegT1);
2017-08-12 16:48:01 +00:00
2020-08-29 13:27:11 +00:00
emitJumpSlowToHot(branchDouble(condition, fpRegT0, fpRegT1), target);
2017-08-12 16:48:01 +00:00
emitJumpSlowToHot(jump(), instructionSize);
2017-08-12 16:48:01 +00:00
2020-08-29 13:27:11 +00:00
fail1.link(this);
fail2.link(this);
fail3.link(this);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
linkSlowCase(iter); // RHS is not Int.
callOperation(operation, TrustedImmPtr(m_codeBlock->globalObject()), regT0, regT1);
2020-08-29 13:27:11 +00:00
emitJumpSlowToHot(branchTest32(invert ? Zero : NonZero, returnValueGPR), target);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emit_op_inc(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
auto bytecode = currentInstruction->as<OpInc>();
VirtualRegister srcDst = bytecode.m_srcDst;
2017-08-12 16:48:01 +00:00
emitGetVirtualRegister(srcDst, regT0);
emitJumpSlowCaseIfNotInt(regT0);
addSlowCase(branchAdd32(Overflow, TrustedImm32(1), regT0));
2020-08-29 13:27:11 +00:00
boxInt32(regT0, JSValueRegs { regT0 });
2017-08-12 16:48:01 +00:00
emitPutVirtualRegister(srcDst);
}
2020-08-29 13:27:11 +00:00
void JIT::emit_op_dec(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
auto bytecode = currentInstruction->as<OpDec>();
VirtualRegister srcDst = bytecode.m_srcDst;
2017-08-12 16:48:01 +00:00
emitGetVirtualRegister(srcDst, regT0);
emitJumpSlowCaseIfNotInt(regT0);
addSlowCase(branchSub32(Overflow, TrustedImm32(1), regT0));
2020-08-29 13:27:11 +00:00
boxInt32(regT0, JSValueRegs { regT0 });
2017-08-12 16:48:01 +00:00
emitPutVirtualRegister(srcDst);
}
/* ------------------------------ BEGIN: OP_MOD ------------------------------ */
2020-08-29 13:27:11 +00:00
#if CPU(X86_64)
2017-08-12 16:48:01 +00:00
2020-08-29 13:27:11 +00:00
void JIT::emit_op_mod(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
auto bytecode = currentInstruction->as<OpMod>();
VirtualRegister result = bytecode.m_dst;
VirtualRegister op1 = bytecode.m_lhs;
VirtualRegister op2 = bytecode.m_rhs;
2017-08-12 16:48:01 +00:00
// Make sure registers are correct for x86 IDIV instructions.
ASSERT(regT0 == X86Registers::eax);
auto edx = X86Registers::edx;
auto ecx = X86Registers::ecx;
ASSERT(regT4 != edx);
ASSERT(regT4 != ecx);
emitGetVirtualRegisters(op1, regT4, op2, ecx);
emitJumpSlowCaseIfNotInt(regT4);
emitJumpSlowCaseIfNotInt(ecx);
move(regT4, regT0);
addSlowCase(branchTest32(Zero, ecx));
Jump denominatorNotNeg1 = branch32(NotEqual, ecx, TrustedImm32(-1));
addSlowCase(branch32(Equal, regT0, TrustedImm32(-2147483647-1)));
denominatorNotNeg1.link(this);
x86ConvertToDoubleWord32();
x86Div32(ecx);
Jump numeratorPositive = branch32(GreaterThanOrEqual, regT4, TrustedImm32(0));
addSlowCase(branchTest32(Zero, edx));
numeratorPositive.link(this);
2020-08-29 13:27:11 +00:00
boxInt32(edx, JSValueRegs { regT0 });
2017-08-12 16:48:01 +00:00
emitPutVirtualRegister(result);
}
2020-08-29 13:27:11 +00:00
void JIT::emitSlow_op_mod(const Instruction* currentInstruction, Vector<SlowCaseEntry>::iterator& iter)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
linkAllSlowCases(iter);
2017-08-12 16:48:01 +00:00
JITSlowPathCall slowPathCall(this, currentInstruction, slow_path_mod);
slowPathCall.call();
}
2020-08-29 13:27:11 +00:00
#else // CPU(X86_64)
2017-08-12 16:48:01 +00:00
2020-08-29 13:27:11 +00:00
void JIT::emit_op_mod(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
JITSlowPathCall slowPathCall(this, currentInstruction, slow_path_mod);
slowPathCall.call();
}
2020-08-29 13:27:11 +00:00
void JIT::emitSlow_op_mod(const Instruction*, Vector<SlowCaseEntry>::iterator&)
2017-08-12 16:48:01 +00:00
{
UNREACHABLE_FOR_PLATFORM();
}
2020-08-29 13:27:11 +00:00
#endif // CPU(X86_64)
2017-08-12 16:48:01 +00:00
/* ------------------------------ END: OP_MOD ------------------------------ */
#else // USE(JSVALUE64)
template <typename Op>
void JIT::emit_compareAndJump(const Instruction* instruction, RelationalCondition condition)
{
JumpList notInt32Op1;
JumpList notInt32Op2;
auto bytecode = instruction->as<Op>();
VirtualRegister op1 = bytecode.m_lhs;
VirtualRegister op2 = bytecode.m_rhs;
unsigned target = jumpTarget(instruction, bytecode.m_targetLabel);
// Character less.
if (isOperandConstantChar(op1)) {
emitLoad(op2, regT1, regT0);
addSlowCase(branchIfNotCell(regT1));
JumpList failures;
emitLoadCharacterString(regT0, regT0, failures);
addSlowCase(failures);
addJump(branch32(commute(condition), regT0, Imm32(asString(getConstantOperand(op1))->tryGetValue()[0])), target);
return;
}
if (isOperandConstantChar(op2)) {
emitLoad(op1, regT1, regT0);
addSlowCase(branchIfNotCell(regT1));
JumpList failures;
emitLoadCharacterString(regT0, regT0, failures);
addSlowCase(failures);
addJump(branch32(condition, regT0, Imm32(asString(getConstantOperand(op2))->tryGetValue()[0])), target);
return;
}
if (isOperandConstantInt(op1)) {
emitLoad(op2, regT3, regT2);
notInt32Op2.append(branchIfNotInt32(regT3));
addJump(branch32(commute(condition), regT2, Imm32(getConstantOperand(op1).asInt32())), target);
} else if (isOperandConstantInt(op2)) {
emitLoad(op1, regT1, regT0);
notInt32Op1.append(branchIfNotInt32(regT1));
addJump(branch32(condition, regT0, Imm32(getConstantOperand(op2).asInt32())), target);
} else {
emitLoad2(op1, regT1, regT0, op2, regT3, regT2);
notInt32Op1.append(branchIfNotInt32(regT1));
notInt32Op2.append(branchIfNotInt32(regT3));
addJump(branch32(condition, regT0, regT2), target);
}
if (!supportsFloatingPoint()) {
addSlowCase(notInt32Op1);
addSlowCase(notInt32Op2);
return;
}
Jump end = jump();
// Double less.
emitBinaryDoubleOp<Op>(instruction, OperandTypes(), notInt32Op1, notInt32Op2, !isOperandConstantInt(op1), isOperandConstantInt(op1) || !isOperandConstantInt(op2));
end.link(this);
}
template <typename Op>
void JIT::emit_compareUnsignedAndJump(const Instruction* instruction, RelationalCondition condition)
{
auto bytecode = instruction->as<Op>();
VirtualRegister op1 = bytecode.m_lhs;
VirtualRegister op2 = bytecode.m_rhs;
unsigned target = jumpTarget(instruction, bytecode.m_targetLabel);
if (isOperandConstantInt(op1)) {
emitLoad(op2, regT3, regT2);
addJump(branch32(commute(condition), regT2, Imm32(getConstantOperand(op1).asInt32())), target);
} else if (isOperandConstantInt(op2)) {
emitLoad(op1, regT1, regT0);
addJump(branch32(condition, regT0, Imm32(getConstantOperand(op2).asInt32())), target);
} else {
emitLoad2(op1, regT1, regT0, op2, regT3, regT2);
addJump(branch32(condition, regT0, regT2), target);
}
}
template <typename Op>
void JIT::emit_compareUnsigned(const Instruction* instruction, RelationalCondition condition)
{
auto bytecode = instruction->as<Op>();
VirtualRegister dst = bytecode.m_dst;
VirtualRegister op1 = bytecode.m_lhs;
VirtualRegister op2 = bytecode.m_rhs;
if (isOperandConstantInt(op1)) {
emitLoad(op2, regT3, regT2);
compare32(commute(condition), regT2, Imm32(getConstantOperand(op1).asInt32()), regT0);
} else if (isOperandConstantInt(op2)) {
emitLoad(op1, regT1, regT0);
compare32(condition, regT0, Imm32(getConstantOperand(op2).asInt32()), regT0);
} else {
emitLoad2(op1, regT1, regT0, op2, regT3, regT2);
compare32(condition, regT0, regT2, regT0);
}
emitStoreBool(dst, regT0);
}
template <typename Op, typename SlowOperation>
void JIT::emit_compareAndJumpSlow(const Instruction *instruction, DoubleCondition, SlowOperation operation, bool invert, Vector<SlowCaseEntry>::iterator& iter)
{
auto bytecode = instruction->as<Op>();
VirtualRegister op1 = bytecode.m_lhs;
VirtualRegister op2 = bytecode.m_rhs;
unsigned target = jumpTarget(instruction, bytecode.m_targetLabel);
linkAllSlowCases(iter);
emitLoad(op1, regT1, regT0);
emitLoad(op2, regT3, regT2);
callOperation(operation, m_codeBlock->globalObject(), JSValueRegs(regT1, regT0), JSValueRegs(regT3, regT2));
emitJumpSlowToHot(branchTest32(invert ? Zero : NonZero, returnValueGPR), target);
}
template <typename Op>
void JIT::emitBinaryDoubleOp(const Instruction *instruction, OperandTypes types, JumpList& notInt32Op1, JumpList& notInt32Op2, bool op1IsInRegisters, bool op2IsInRegisters)
{
JumpList end;
auto bytecode = instruction->as<Op>();
int opcodeID = Op::opcodeID;
int target = jumpTarget(instruction, bytecode.m_targetLabel);
VirtualRegister op1 = bytecode.m_lhs;
VirtualRegister op2 = bytecode.m_rhs;
if (!notInt32Op1.empty()) {
// Double case 1: Op1 is not int32; Op2 is unknown.
notInt32Op1.link(this);
ASSERT(op1IsInRegisters);
// Verify Op1 is double.
if (!types.first().definitelyIsNumber())
addSlowCase(branch32(Above, regT1, TrustedImm32(JSValue::LowestTag)));
if (!op2IsInRegisters)
emitLoad(op2, regT3, regT2);
Jump doubleOp2 = branch32(Below, regT3, TrustedImm32(JSValue::LowestTag));
if (!types.second().definitelyIsNumber())
addSlowCase(branchIfNotInt32(regT3));
convertInt32ToDouble(regT2, fpRegT0);
Jump doTheMath = jump();
// Load Op2 as double into double register.
doubleOp2.link(this);
emitLoadDouble(op2, fpRegT0);
// Do the math.
doTheMath.link(this);
switch (opcodeID) {
case op_jless:
emitLoadDouble(op1, fpRegT2);
addJump(branchDouble(DoubleLessThanAndOrdered, fpRegT2, fpRegT0), target);
break;
case op_jlesseq:
emitLoadDouble(op1, fpRegT2);
addJump(branchDouble(DoubleLessThanOrEqualAndOrdered, fpRegT2, fpRegT0), target);
break;
case op_jgreater:
emitLoadDouble(op1, fpRegT2);
addJump(branchDouble(DoubleGreaterThanAndOrdered, fpRegT2, fpRegT0), target);
break;
case op_jgreatereq:
emitLoadDouble(op1, fpRegT2);
addJump(branchDouble(DoubleGreaterThanOrEqualAndOrdered, fpRegT2, fpRegT0), target);
break;
case op_jnless:
emitLoadDouble(op1, fpRegT2);
addJump(branchDouble(DoubleLessThanOrEqualOrUnordered, fpRegT0, fpRegT2), target);
break;
case op_jnlesseq:
emitLoadDouble(op1, fpRegT2);
addJump(branchDouble(DoubleLessThanOrUnordered, fpRegT0, fpRegT2), target);
break;
case op_jngreater:
emitLoadDouble(op1, fpRegT2);
addJump(branchDouble(DoubleGreaterThanOrEqualOrUnordered, fpRegT0, fpRegT2), target);
break;
case op_jngreatereq:
emitLoadDouble(op1, fpRegT2);
addJump(branchDouble(DoubleGreaterThanOrUnordered, fpRegT0, fpRegT2), target);
break;
default:
RELEASE_ASSERT_NOT_REACHED();
}
if (!notInt32Op2.empty())
end.append(jump());
}
if (!notInt32Op2.empty()) {
// Double case 2: Op1 is int32; Op2 is not int32.
notInt32Op2.link(this);
ASSERT(op2IsInRegisters);
if (!op1IsInRegisters)
emitLoadPayload(op1, regT0);
convertInt32ToDouble(regT0, fpRegT0);
// Verify op2 is double.
if (!types.second().definitelyIsNumber())
addSlowCase(branch32(Above, regT3, TrustedImm32(JSValue::LowestTag)));
// Do the math.
switch (opcodeID) {
case op_jless:
emitLoadDouble(op2, fpRegT1);
addJump(branchDouble(DoubleLessThanAndOrdered, fpRegT0, fpRegT1), target);
break;
case op_jlesseq:
emitLoadDouble(op2, fpRegT1);
addJump(branchDouble(DoubleLessThanOrEqualAndOrdered, fpRegT0, fpRegT1), target);
break;
case op_jgreater:
emitLoadDouble(op2, fpRegT1);
addJump(branchDouble(DoubleGreaterThanAndOrdered, fpRegT0, fpRegT1), target);
break;
case op_jgreatereq:
emitLoadDouble(op2, fpRegT1);
addJump(branchDouble(DoubleGreaterThanOrEqualAndOrdered, fpRegT0, fpRegT1), target);
break;
case op_jnless:
emitLoadDouble(op2, fpRegT1);
addJump(branchDouble(DoubleLessThanOrEqualOrUnordered, fpRegT1, fpRegT0), target);
break;
case op_jnlesseq:
emitLoadDouble(op2, fpRegT1);
addJump(branchDouble(DoubleLessThanOrUnordered, fpRegT1, fpRegT0), target);
break;
case op_jngreater:
emitLoadDouble(op2, fpRegT1);
addJump(branchDouble(DoubleGreaterThanOrEqualOrUnordered, fpRegT1, fpRegT0), target);
break;
case op_jngreatereq:
emitLoadDouble(op2, fpRegT1);
addJump(branchDouble(DoubleGreaterThanOrUnordered, fpRegT1, fpRegT0), target);
break;
default:
RELEASE_ASSERT_NOT_REACHED();
}
}
end.link(this);
}
2017-08-12 16:48:01 +00:00
#endif // USE(JSVALUE64)
2020-08-29 13:27:11 +00:00
void JIT::emit_op_negate(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
UnaryArithProfile* arithProfile = &currentInstruction->as<OpNegate>().metadata(m_codeBlock).m_arithProfile;
2017-08-12 16:48:01 +00:00
JITNegIC* negateIC = m_codeBlock->addJITNegIC(arithProfile);
m_instructionToMathIC.add(currentInstruction, negateIC);
// FIXME: it would be better to call those operationValueNegate, since the operand can be a BigInt
2020-08-29 13:27:11 +00:00
emitMathICFast<OpNegate>(negateIC, currentInstruction, operationArithNegateProfiled, operationArithNegate);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emitSlow_op_negate(const Instruction* currentInstruction, Vector<SlowCaseEntry>::iterator& iter)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
linkAllSlowCases(iter);
2017-08-12 16:48:01 +00:00
JITNegIC* negIC = bitwise_cast<JITNegIC*>(m_instructionToMathIC.get(currentInstruction));
// FIXME: it would be better to call those operationValueNegate, since the operand can be a BigInt
2020-08-29 13:27:11 +00:00
emitMathICSlow<OpNegate>(negIC, currentInstruction, operationArithNegateProfiledOptimize, operationArithNegateProfiled, operationArithNegateOptimize);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
template<typename Op, typename SnippetGenerator>
void JIT::emitBitBinaryOpFastPath(const Instruction* currentInstruction, ProfilingPolicy profilingPolicy)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
auto bytecode = currentInstruction->as<Op>();
VirtualRegister result = bytecode.m_dst;
VirtualRegister op1 = bytecode.m_lhs;
VirtualRegister op2 = bytecode.m_rhs;
2017-08-12 16:48:01 +00:00
#if USE(JSVALUE64)
JSValueRegs leftRegs = JSValueRegs(regT0);
JSValueRegs rightRegs = JSValueRegs(regT1);
JSValueRegs resultRegs = leftRegs;
GPRReg scratchGPR = regT2;
#else
JSValueRegs leftRegs = JSValueRegs(regT1, regT0);
JSValueRegs rightRegs = JSValueRegs(regT3, regT2);
JSValueRegs resultRegs = leftRegs;
GPRReg scratchGPR = regT4;
#endif
SnippetOperand leftOperand;
SnippetOperand rightOperand;
if (isOperandConstantInt(op1))
leftOperand.setConstInt32(getOperandConstantInt(op1));
else if (isOperandConstantInt(op2))
rightOperand.setConstInt32(getOperandConstantInt(op2));
RELEASE_ASSERT(!leftOperand.isConst() || !rightOperand.isConst());
if (!leftOperand.isConst())
emitGetVirtualRegister(op1, leftRegs);
if (!rightOperand.isConst())
emitGetVirtualRegister(op2, rightRegs);
SnippetGenerator gen(leftOperand, rightOperand, resultRegs, leftRegs, rightRegs, scratchGPR);
gen.generateFastPath(*this);
ASSERT(gen.didEmitFastPath());
gen.endJumpList().link(this);
2020-08-29 13:27:11 +00:00
if (profilingPolicy == ProfilingPolicy::ShouldEmitProfiling)
emitValueProfilingSiteIfProfiledOpcode(bytecode);
2017-08-12 16:48:01 +00:00
emitPutVirtualRegister(result, resultRegs);
addSlowCase(gen.slowPathJumpList());
}
2020-08-29 13:27:11 +00:00
void JIT::emit_op_bitnot(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
auto bytecode = currentInstruction->as<OpBitnot>();
VirtualRegister result = bytecode.m_dst;
VirtualRegister op1 = bytecode.m_operand;
2017-08-12 16:48:01 +00:00
2020-08-29 13:27:11 +00:00
#if USE(JSVALUE64)
JSValueRegs leftRegs = JSValueRegs(regT0);
#else
JSValueRegs leftRegs = JSValueRegs(regT1, regT0);
#endif
2017-08-12 16:48:01 +00:00
2020-08-29 13:27:11 +00:00
emitGetVirtualRegister(op1, leftRegs);
2017-08-12 16:48:01 +00:00
2020-08-29 13:27:11 +00:00
addSlowCase(branchIfNotInt32(leftRegs));
not32(leftRegs.payloadGPR());
#if USE(JSVALUE64)
boxInt32(leftRegs.payloadGPR(), leftRegs);
#endif
2017-08-12 16:48:01 +00:00
2020-08-29 13:27:11 +00:00
emitValueProfilingSiteIfProfiledOpcode(bytecode);
2017-08-12 16:48:01 +00:00
2020-08-29 13:27:11 +00:00
emitPutVirtualRegister(result, leftRegs);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emit_op_bitand(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
emitBitBinaryOpFastPath<OpBitand, JITBitAndGenerator>(currentInstruction, ProfilingPolicy::ShouldEmitProfiling);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emit_op_bitor(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
emitBitBinaryOpFastPath<OpBitor, JITBitOrGenerator>(currentInstruction, ProfilingPolicy::ShouldEmitProfiling);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emit_op_bitxor(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
emitBitBinaryOpFastPath<OpBitxor, JITBitXorGenerator>(currentInstruction, ProfilingPolicy::ShouldEmitProfiling);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emit_op_lshift(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
emitBitBinaryOpFastPath<OpLshift, JITLeftShiftGenerator>(currentInstruction);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emitRightShiftFastPath(const Instruction* currentInstruction, OpcodeID opcodeID)
2017-08-12 16:48:01 +00:00
{
ASSERT(opcodeID == op_rshift || opcodeID == op_urshift);
2020-08-29 13:27:11 +00:00
switch (opcodeID) {
case op_rshift:
emitRightShiftFastPath<OpRshift>(currentInstruction, JITRightShiftGenerator::SignedShift);
break;
case op_urshift:
emitRightShiftFastPath<OpUrshift>(currentInstruction, JITRightShiftGenerator::UnsignedShift);
break;
default:
ASSERT_NOT_REACHED();
}
}
2017-08-12 16:48:01 +00:00
2020-08-29 13:27:11 +00:00
template<typename Op>
void JIT::emitRightShiftFastPath(const Instruction* currentInstruction, JITRightShiftGenerator::ShiftType snippetShiftType)
{
auto bytecode = currentInstruction->as<Op>();
VirtualRegister result = bytecode.m_dst;
VirtualRegister op1 = bytecode.m_lhs;
VirtualRegister op2 = bytecode.m_rhs;
2017-08-12 16:48:01 +00:00
#if USE(JSVALUE64)
JSValueRegs leftRegs = JSValueRegs(regT0);
JSValueRegs rightRegs = JSValueRegs(regT1);
JSValueRegs resultRegs = leftRegs;
GPRReg scratchGPR = regT2;
FPRReg scratchFPR = InvalidFPRReg;
#else
JSValueRegs leftRegs = JSValueRegs(regT1, regT0);
JSValueRegs rightRegs = JSValueRegs(regT3, regT2);
JSValueRegs resultRegs = leftRegs;
GPRReg scratchGPR = regT4;
FPRReg scratchFPR = fpRegT2;
#endif
SnippetOperand leftOperand;
SnippetOperand rightOperand;
if (isOperandConstantInt(op1))
leftOperand.setConstInt32(getOperandConstantInt(op1));
else if (isOperandConstantInt(op2))
rightOperand.setConstInt32(getOperandConstantInt(op2));
RELEASE_ASSERT(!leftOperand.isConst() || !rightOperand.isConst());
if (!leftOperand.isConst())
emitGetVirtualRegister(op1, leftRegs);
if (!rightOperand.isConst())
emitGetVirtualRegister(op2, rightRegs);
JITRightShiftGenerator gen(leftOperand, rightOperand, resultRegs, leftRegs, rightRegs,
fpRegT0, scratchGPR, scratchFPR, snippetShiftType);
gen.generateFastPath(*this);
ASSERT(gen.didEmitFastPath());
gen.endJumpList().link(this);
emitPutVirtualRegister(result, resultRegs);
addSlowCase(gen.slowPathJumpList());
}
2020-08-29 13:27:11 +00:00
void JIT::emit_op_rshift(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
emitRightShiftFastPath(currentInstruction, op_rshift);
}
2020-08-29 13:27:11 +00:00
void JIT::emit_op_urshift(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
emitRightShiftFastPath(currentInstruction, op_urshift);
}
2020-08-29 13:27:11 +00:00
void JIT::emit_op_add(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
BinaryArithProfile* arithProfile = &currentInstruction->as<OpAdd>().metadata(m_codeBlock).m_arithProfile;
2017-08-12 16:48:01 +00:00
JITAddIC* addIC = m_codeBlock->addJITAddIC(arithProfile);
m_instructionToMathIC.add(currentInstruction, addIC);
2020-08-29 13:27:11 +00:00
emitMathICFast<OpAdd>(addIC, currentInstruction, operationValueAddProfiled, operationValueAdd);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emitSlow_op_add(const Instruction* currentInstruction, Vector<SlowCaseEntry>::iterator& iter)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
linkAllSlowCases(iter);
2017-08-12 16:48:01 +00:00
JITAddIC* addIC = bitwise_cast<JITAddIC*>(m_instructionToMathIC.get(currentInstruction));
2020-08-29 13:27:11 +00:00
emitMathICSlow<OpAdd>(addIC, currentInstruction, operationValueAddProfiledOptimize, operationValueAddProfiled, operationValueAddOptimize);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
template <typename Op, typename Generator, typename ProfiledFunction, typename NonProfiledFunction>
void JIT::emitMathICFast(JITUnaryMathIC<Generator>* mathIC, const Instruction* currentInstruction, ProfiledFunction profiledFunction, NonProfiledFunction nonProfiledFunction)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
auto bytecode = currentInstruction->as<Op>();
VirtualRegister result = bytecode.m_dst;
VirtualRegister operand = bytecode.m_operand;
2017-08-12 16:48:01 +00:00
#if USE(JSVALUE64)
// ArithNegate benefits from using the same register as src and dst.
// Since regT1==argumentGPR1, using regT1 avoid shuffling register to call the slow path.
JSValueRegs srcRegs = JSValueRegs(regT1);
JSValueRegs resultRegs = JSValueRegs(regT1);
GPRReg scratchGPR = regT2;
#else
JSValueRegs srcRegs = JSValueRegs(regT1, regT0);
JSValueRegs resultRegs = JSValueRegs(regT3, regT2);
GPRReg scratchGPR = regT4;
#endif
#if ENABLE(MATH_IC_STATS)
auto inlineStart = label();
#endif
mathIC->m_generator = Generator(resultRegs, srcRegs, scratchGPR);
emitGetVirtualRegister(operand, srcRegs);
MathICGenerationState& mathICGenerationState = m_instructionToMathICGenerationState.add(currentInstruction, MathICGenerationState()).iterator->value;
bool generatedInlineCode = mathIC->generateInline(*this, mathICGenerationState);
if (!generatedInlineCode) {
UnaryArithProfile* arithProfile = mathIC->arithProfile();
2017-08-12 16:48:01 +00:00
if (arithProfile && shouldEmitProfiling())
callOperationWithResult(profiledFunction, resultRegs, TrustedImmPtr(m_codeBlock->globalObject()), srcRegs, arithProfile);
2017-08-12 16:48:01 +00:00
else
callOperationWithResult(nonProfiledFunction, resultRegs, TrustedImmPtr(m_codeBlock->globalObject()), srcRegs);
2017-08-12 16:48:01 +00:00
} else
addSlowCase(mathICGenerationState.slowPathJumps);
#if ENABLE(MATH_IC_STATS)
auto inlineEnd = label();
addLinkTask([=] (LinkBuffer& linkBuffer) {
2020-08-29 13:27:11 +00:00
size_t size = linkBuffer.locationOf(inlineEnd).executableAddress<char*>() - linkBuffer.locationOf(inlineStart).executableAddress<char*>();
2017-08-12 16:48:01 +00:00
mathIC->m_generatedCodeSize += size;
});
#endif
emitPutVirtualRegister(result, resultRegs);
}
2020-08-29 13:27:11 +00:00
template <typename Op, typename Generator, typename ProfiledFunction, typename NonProfiledFunction>
void JIT::emitMathICFast(JITBinaryMathIC<Generator>* mathIC, const Instruction* currentInstruction, ProfiledFunction profiledFunction, NonProfiledFunction nonProfiledFunction)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
auto bytecode = currentInstruction->as<Op>();
VirtualRegister result = bytecode.m_dst;
VirtualRegister op1 = bytecode.m_lhs;
VirtualRegister op2 = bytecode.m_rhs;
2017-08-12 16:48:01 +00:00
#if USE(JSVALUE64)
JSValueRegs leftRegs = JSValueRegs(regT1);
JSValueRegs rightRegs = JSValueRegs(regT2);
JSValueRegs resultRegs = JSValueRegs(regT0);
GPRReg scratchGPR = regT3;
FPRReg scratchFPR = fpRegT2;
#else
JSValueRegs leftRegs = JSValueRegs(regT1, regT0);
JSValueRegs rightRegs = JSValueRegs(regT3, regT2);
JSValueRegs resultRegs = leftRegs;
GPRReg scratchGPR = regT4;
FPRReg scratchFPR = fpRegT2;
#endif
SnippetOperand leftOperand(bytecode.m_operandTypes.first());
SnippetOperand rightOperand(bytecode.m_operandTypes.second());
2017-08-12 16:48:01 +00:00
if (isOperandConstantInt(op1))
leftOperand.setConstInt32(getOperandConstantInt(op1));
else if (isOperandConstantInt(op2))
rightOperand.setConstInt32(getOperandConstantInt(op2));
RELEASE_ASSERT(!leftOperand.isConst() || !rightOperand.isConst());
mathIC->m_generator = Generator(leftOperand, rightOperand, resultRegs, leftRegs, rightRegs, fpRegT0, fpRegT1, scratchGPR, scratchFPR);
ASSERT(!(Generator::isLeftOperandValidConstant(leftOperand) && Generator::isRightOperandValidConstant(rightOperand)));
if (!Generator::isLeftOperandValidConstant(leftOperand))
emitGetVirtualRegister(op1, leftRegs);
if (!Generator::isRightOperandValidConstant(rightOperand))
emitGetVirtualRegister(op2, rightRegs);
#if ENABLE(MATH_IC_STATS)
auto inlineStart = label();
#endif
MathICGenerationState& mathICGenerationState = m_instructionToMathICGenerationState.add(currentInstruction, MathICGenerationState()).iterator->value;
bool generatedInlineCode = mathIC->generateInline(*this, mathICGenerationState);
if (!generatedInlineCode) {
if (leftOperand.isConst())
emitGetVirtualRegister(op1, leftRegs);
else if (rightOperand.isConst())
emitGetVirtualRegister(op2, rightRegs);
BinaryArithProfile* arithProfile = mathIC->arithProfile();
2017-08-12 16:48:01 +00:00
if (arithProfile && shouldEmitProfiling())
callOperationWithResult(profiledFunction, resultRegs, TrustedImmPtr(m_codeBlock->globalObject()), leftRegs, rightRegs, arithProfile);
2017-08-12 16:48:01 +00:00
else
callOperationWithResult(nonProfiledFunction, resultRegs, TrustedImmPtr(m_codeBlock->globalObject()), leftRegs, rightRegs);
2017-08-12 16:48:01 +00:00
} else
addSlowCase(mathICGenerationState.slowPathJumps);
#if ENABLE(MATH_IC_STATS)
auto inlineEnd = label();
addLinkTask([=] (LinkBuffer& linkBuffer) {
2020-08-29 13:27:11 +00:00
size_t size = linkBuffer.locationOf(inlineEnd).executableAddress<char*>() - linkBuffer.locationOf(inlineStart).executableAddress<char*>();
2017-08-12 16:48:01 +00:00
mathIC->m_generatedCodeSize += size;
});
#endif
emitPutVirtualRegister(result, resultRegs);
}
2020-08-29 13:27:11 +00:00
template <typename Op, typename Generator, typename ProfiledRepatchFunction, typename ProfiledFunction, typename RepatchFunction>
void JIT::emitMathICSlow(JITUnaryMathIC<Generator>* mathIC, const Instruction* currentInstruction, ProfiledRepatchFunction profiledRepatchFunction, ProfiledFunction profiledFunction, RepatchFunction repatchFunction)
2017-08-12 16:48:01 +00:00
{
MathICGenerationState& mathICGenerationState = m_instructionToMathICGenerationState.find(currentInstruction)->value;
mathICGenerationState.slowPathStart = label();
2020-08-29 13:27:11 +00:00
auto bytecode = currentInstruction->as<Op>();
VirtualRegister result = bytecode.m_dst;
2017-08-12 16:48:01 +00:00
#if USE(JSVALUE64)
JSValueRegs srcRegs = JSValueRegs(regT1);
JSValueRegs resultRegs = JSValueRegs(regT0);
#else
JSValueRegs srcRegs = JSValueRegs(regT1, regT0);
JSValueRegs resultRegs = JSValueRegs(regT3, regT2);
#endif
#if ENABLE(MATH_IC_STATS)
auto slowPathStart = label();
#endif
UnaryArithProfile* arithProfile = mathIC->arithProfile();
2017-08-12 16:48:01 +00:00
if (arithProfile && shouldEmitProfiling()) {
if (mathICGenerationState.shouldSlowPathRepatch)
mathICGenerationState.slowPathCall = callOperationWithResult(reinterpret_cast<J_JITOperation_GJMic>(profiledRepatchFunction), resultRegs, TrustedImmPtr(m_codeBlock->globalObject()), srcRegs, TrustedImmPtr(mathIC));
2017-08-12 16:48:01 +00:00
else
mathICGenerationState.slowPathCall = callOperationWithResult(profiledFunction, resultRegs, TrustedImmPtr(m_codeBlock->globalObject()), srcRegs, arithProfile);
2017-08-12 16:48:01 +00:00
} else
mathICGenerationState.slowPathCall = callOperationWithResult(reinterpret_cast<J_JITOperation_GJMic>(repatchFunction), resultRegs, TrustedImmPtr(m_codeBlock->globalObject()), srcRegs, TrustedImmPtr(mathIC));
2017-08-12 16:48:01 +00:00
#if ENABLE(MATH_IC_STATS)
auto slowPathEnd = label();
addLinkTask([=] (LinkBuffer& linkBuffer) {
2020-08-29 13:27:11 +00:00
size_t size = linkBuffer.locationOf(slowPathEnd).executableAddress<char*>() - linkBuffer.locationOf(slowPathStart).executableAddress<char*>();
2017-08-12 16:48:01 +00:00
mathIC->m_generatedCodeSize += size;
});
#endif
emitPutVirtualRegister(result, resultRegs);
addLinkTask([=] (LinkBuffer& linkBuffer) {
MathICGenerationState& mathICGenerationState = m_instructionToMathICGenerationState.find(currentInstruction)->value;
mathIC->finalizeInlineCode(mathICGenerationState, linkBuffer);
});
}
2020-08-29 13:27:11 +00:00
template <typename Op, typename Generator, typename ProfiledRepatchFunction, typename ProfiledFunction, typename RepatchFunction>
void JIT::emitMathICSlow(JITBinaryMathIC<Generator>* mathIC, const Instruction* currentInstruction, ProfiledRepatchFunction profiledRepatchFunction, ProfiledFunction profiledFunction, RepatchFunction repatchFunction)
2017-08-12 16:48:01 +00:00
{
MathICGenerationState& mathICGenerationState = m_instructionToMathICGenerationState.find(currentInstruction)->value;
mathICGenerationState.slowPathStart = label();
2020-08-29 13:27:11 +00:00
auto bytecode = currentInstruction->as<Op>();
VirtualRegister result = bytecode.m_dst;
VirtualRegister op1 = bytecode.m_lhs;
VirtualRegister op2 = bytecode.m_rhs;
2017-08-12 16:48:01 +00:00
#if USE(JSVALUE64)
JSValueRegs leftRegs = JSValueRegs(regT1);
JSValueRegs rightRegs = JSValueRegs(regT2);
JSValueRegs resultRegs = JSValueRegs(regT0);
#else
JSValueRegs leftRegs = JSValueRegs(regT1, regT0);
JSValueRegs rightRegs = JSValueRegs(regT3, regT2);
JSValueRegs resultRegs = leftRegs;
#endif
SnippetOperand leftOperand(bytecode.m_operandTypes.first());
SnippetOperand rightOperand(bytecode.m_operandTypes.second());
2017-08-12 16:48:01 +00:00
if (isOperandConstantInt(op1))
leftOperand.setConstInt32(getOperandConstantInt(op1));
else if (isOperandConstantInt(op2))
rightOperand.setConstInt32(getOperandConstantInt(op2));
ASSERT(!(Generator::isLeftOperandValidConstant(leftOperand) && Generator::isRightOperandValidConstant(rightOperand)));
if (Generator::isLeftOperandValidConstant(leftOperand))
emitGetVirtualRegister(op1, leftRegs);
else if (Generator::isRightOperandValidConstant(rightOperand))
emitGetVirtualRegister(op2, rightRegs);
#if ENABLE(MATH_IC_STATS)
auto slowPathStart = label();
#endif
BinaryArithProfile* arithProfile = mathIC->arithProfile();
2017-08-12 16:48:01 +00:00
if (arithProfile && shouldEmitProfiling()) {
if (mathICGenerationState.shouldSlowPathRepatch)
mathICGenerationState.slowPathCall = callOperationWithResult(bitwise_cast<J_JITOperation_GJJMic>(profiledRepatchFunction), resultRegs, TrustedImmPtr(m_codeBlock->globalObject()), leftRegs, rightRegs, TrustedImmPtr(mathIC));
2017-08-12 16:48:01 +00:00
else
mathICGenerationState.slowPathCall = callOperationWithResult(profiledFunction, resultRegs, TrustedImmPtr(m_codeBlock->globalObject()), leftRegs, rightRegs, arithProfile);
2017-08-12 16:48:01 +00:00
} else
mathICGenerationState.slowPathCall = callOperationWithResult(bitwise_cast<J_JITOperation_GJJMic>(repatchFunction), resultRegs, TrustedImmPtr(m_codeBlock->globalObject()), leftRegs, rightRegs, TrustedImmPtr(mathIC));
2017-08-12 16:48:01 +00:00
#if ENABLE(MATH_IC_STATS)
auto slowPathEnd = label();
addLinkTask([=] (LinkBuffer& linkBuffer) {
2020-08-29 13:27:11 +00:00
size_t size = linkBuffer.locationOf(slowPathEnd).executableAddress<char*>() - linkBuffer.locationOf(slowPathStart).executableAddress<char*>();
2017-08-12 16:48:01 +00:00
mathIC->m_generatedCodeSize += size;
});
#endif
emitPutVirtualRegister(result, resultRegs);
addLinkTask([=] (LinkBuffer& linkBuffer) {
MathICGenerationState& mathICGenerationState = m_instructionToMathICGenerationState.find(currentInstruction)->value;
mathIC->finalizeInlineCode(mathICGenerationState, linkBuffer);
});
}
2020-08-29 13:27:11 +00:00
void JIT::emit_op_div(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
auto bytecode = currentInstruction->as<OpDiv>();
VirtualRegister result = bytecode.m_dst;
VirtualRegister op1 = bytecode.m_lhs;
VirtualRegister op2 = bytecode.m_rhs;
2017-08-12 16:48:01 +00:00
#if USE(JSVALUE64)
JSValueRegs leftRegs = JSValueRegs(regT0);
JSValueRegs rightRegs = JSValueRegs(regT1);
JSValueRegs resultRegs = leftRegs;
GPRReg scratchGPR = regT2;
#else
JSValueRegs leftRegs = JSValueRegs(regT1, regT0);
JSValueRegs rightRegs = JSValueRegs(regT3, regT2);
JSValueRegs resultRegs = leftRegs;
GPRReg scratchGPR = regT4;
#endif
FPRReg scratchFPR = fpRegT2;
BinaryArithProfile* arithProfile = nullptr;
2017-08-12 16:48:01 +00:00
if (shouldEmitProfiling())
2020-08-29 13:27:11 +00:00
arithProfile = &currentInstruction->as<OpDiv>().metadata(m_codeBlock).m_arithProfile;
2017-08-12 16:48:01 +00:00
SnippetOperand leftOperand(bytecode.m_operandTypes.first());
SnippetOperand rightOperand(bytecode.m_operandTypes.second());
2017-08-12 16:48:01 +00:00
if (isOperandConstantInt(op1))
leftOperand.setConstInt32(getOperandConstantInt(op1));
#if USE(JSVALUE64)
else if (isOperandConstantDouble(op1))
leftOperand.setConstDouble(getOperandConstantDouble(op1));
#endif
else if (isOperandConstantInt(op2))
rightOperand.setConstInt32(getOperandConstantInt(op2));
#if USE(JSVALUE64)
else if (isOperandConstantDouble(op2))
rightOperand.setConstDouble(getOperandConstantDouble(op2));
#endif
RELEASE_ASSERT(!leftOperand.isConst() || !rightOperand.isConst());
if (!leftOperand.isConst())
emitGetVirtualRegister(op1, leftRegs);
if (!rightOperand.isConst())
emitGetVirtualRegister(op2, rightRegs);
JITDivGenerator gen(leftOperand, rightOperand, resultRegs, leftRegs, rightRegs,
fpRegT0, fpRegT1, scratchGPR, scratchFPR, arithProfile);
gen.generateFastPath(*this);
if (gen.didEmitFastPath()) {
gen.endJumpList().link(this);
emitPutVirtualRegister(result, resultRegs);
addSlowCase(gen.slowPathJumpList());
} else {
ASSERT(gen.endJumpList().empty());
ASSERT(gen.slowPathJumpList().empty());
JITSlowPathCall slowPathCall(this, currentInstruction, slow_path_div);
slowPathCall.call();
}
}
2020-08-29 13:27:11 +00:00
void JIT::emit_op_mul(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
BinaryArithProfile* arithProfile = &currentInstruction->as<OpMul>().metadata(m_codeBlock).m_arithProfile;
2017-08-12 16:48:01 +00:00
JITMulIC* mulIC = m_codeBlock->addJITMulIC(arithProfile);
m_instructionToMathIC.add(currentInstruction, mulIC);
2020-08-29 13:27:11 +00:00
emitMathICFast<OpMul>(mulIC, currentInstruction, operationValueMulProfiled, operationValueMul);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emitSlow_op_mul(const Instruction* currentInstruction, Vector<SlowCaseEntry>::iterator& iter)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
linkAllSlowCases(iter);
2017-08-12 16:48:01 +00:00
JITMulIC* mulIC = bitwise_cast<JITMulIC*>(m_instructionToMathIC.get(currentInstruction));
2020-08-29 13:27:11 +00:00
emitMathICSlow<OpMul>(mulIC, currentInstruction, operationValueMulProfiledOptimize, operationValueMulProfiled, operationValueMulOptimize);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emit_op_sub(const Instruction* currentInstruction)
2017-08-12 16:48:01 +00:00
{
BinaryArithProfile* arithProfile = &currentInstruction->as<OpSub>().metadata(m_codeBlock).m_arithProfile;
2017-08-12 16:48:01 +00:00
JITSubIC* subIC = m_codeBlock->addJITSubIC(arithProfile);
m_instructionToMathIC.add(currentInstruction, subIC);
2020-08-29 13:27:11 +00:00
emitMathICFast<OpSub>(subIC, currentInstruction, operationValueSubProfiled, operationValueSub);
2017-08-12 16:48:01 +00:00
}
2020-08-29 13:27:11 +00:00
void JIT::emitSlow_op_sub(const Instruction* currentInstruction, Vector<SlowCaseEntry>::iterator& iter)
2017-08-12 16:48:01 +00:00
{
2020-08-29 13:27:11 +00:00
linkAllSlowCases(iter);
2017-08-12 16:48:01 +00:00
2020-08-29 13:27:11 +00:00
JITSubIC* subIC = bitwise_cast<JITSubIC*>(m_instructionToMathIC.get(currentInstruction));
emitMathICSlow<OpSub>(subIC, currentInstruction, operationValueSubProfiledOptimize, operationValueSubProfiled, operationValueSubOptimize);
2017-08-12 16:48:01 +00:00
}
/* ------------------------------ END: OP_ADD, OP_SUB, OP_MUL, OP_POW ------------------------------ */
} // namespace JSC
#endif // ENABLE(JIT)