mirror of
https://github.com/darlinghq/darling-JavaScriptCore.git
synced 2025-02-17 02:18:14 +00:00
19828 lines
839 KiB
C++
19828 lines
839 KiB
C++
/*
|
|
* Copyright (C) 2013-2020 Apple Inc. All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY APPLE INC. ``AS IS'' AND ANY
|
|
* EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
|
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR
|
|
* CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
|
|
* EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
|
|
* PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
|
|
* PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY
|
|
* OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#include "config.h"
|
|
#include "FTLLowerDFGToB3.h"
|
|
|
|
#if ENABLE(FTL_JIT)
|
|
|
|
#include "AirCode.h"
|
|
#include "AllowMacroScratchRegisterUsage.h"
|
|
#include "AllowMacroScratchRegisterUsageIf.h"
|
|
#include "AtomicsObject.h"
|
|
#include "B3CheckValue.h"
|
|
#include "B3PatchpointValue.h"
|
|
#include "B3SlotBaseValue.h"
|
|
#include "B3StackmapGenerationParams.h"
|
|
#include "B3ValueInlines.h"
|
|
#include "ButterflyInlines.h"
|
|
#include "CallFrameShuffler.h"
|
|
#include "DFGAbstractInterpreterInlines.h"
|
|
#include "DFGCapabilities.h"
|
|
#include "DFGClobberize.h"
|
|
#include "DFGDoesGC.h"
|
|
#include "DFGDominators.h"
|
|
#include "DFGInPlaceAbstractState.h"
|
|
#include "DFGLivenessAnalysisPhase.h"
|
|
#include "DFGMayExit.h"
|
|
#include "DFGOSRAvailabilityAnalysisPhase.h"
|
|
#include "DFGOSRExitFuzz.h"
|
|
#include "DirectArguments.h"
|
|
#include "FTLAbstractHeapRepository.h"
|
|
#include "FTLExceptionTarget.h"
|
|
#include "FTLForOSREntryJITCode.h"
|
|
#include "FTLFormattedValue.h"
|
|
#include "FTLLazySlowPathCall.h"
|
|
#include "FTLLoweredNodeValue.h"
|
|
#include "FTLOperations.h"
|
|
#include "FTLOutput.h"
|
|
#include "FTLPatchpointExceptionHandle.h"
|
|
#include "FTLSnippetParams.h"
|
|
#include "FTLThunks.h"
|
|
#include "FTLWeightedTarget.h"
|
|
#include "JITAddGenerator.h"
|
|
#include "JITBitAndGenerator.h"
|
|
#include "JITBitOrGenerator.h"
|
|
#include "JITBitXorGenerator.h"
|
|
#include "JITDivGenerator.h"
|
|
#include "JITInlineCacheGenerator.h"
|
|
#include "JITLeftShiftGenerator.h"
|
|
#include "JITMathIC.h"
|
|
#include "JITMulGenerator.h"
|
|
#include "JITRightShiftGenerator.h"
|
|
#include "JITSubGenerator.h"
|
|
#include "JSArrayIterator.h"
|
|
#include "JSAsyncFunction.h"
|
|
#include "JSAsyncGenerator.h"
|
|
#include "JSAsyncGeneratorFunction.h"
|
|
#include "JSCInlines.h"
|
|
#include "JSGenerator.h"
|
|
#include "JSGeneratorFunction.h"
|
|
#include "JSImmutableButterfly.h"
|
|
#include "JSInternalPromise.h"
|
|
#include "JSLexicalEnvironment.h"
|
|
#include "JSMapIterator.h"
|
|
#include "JSSetIterator.h"
|
|
#include "LLIntThunks.h"
|
|
#include "OperandsInlines.h"
|
|
#include "ProbeContext.h"
|
|
#include "RegExpObject.h"
|
|
#include "ScratchRegisterAllocator.h"
|
|
#include "SetupVarargsFrame.h"
|
|
#include "ShadowChicken.h"
|
|
#include "StructureStubInfo.h"
|
|
#include "SuperSampler.h"
|
|
#include "ThunkGenerators.h"
|
|
#include "VirtualRegister.h"
|
|
#include <atomic>
|
|
#include <wtf/Box.h>
|
|
#include <wtf/RecursableLambda.h>
|
|
#include <wtf/StdUnorderedSet.h>
|
|
|
|
#undef RELEASE_ASSERT
|
|
#define RELEASE_ASSERT(assertion) do { \
|
|
if (!(assertion)) { \
|
|
WTFReportAssertionFailure(__FILE__, __LINE__, WTF_PRETTY_FUNCTION, #assertion); \
|
|
CRASH(); \
|
|
} \
|
|
} while (0)
|
|
|
|
namespace JSC { namespace FTL {
|
|
|
|
using namespace B3;
|
|
using namespace DFG;
|
|
|
|
namespace {
|
|
|
|
std::atomic<int> compileCounter;
|
|
|
|
#if ASSERT_ENABLED
|
|
NO_RETURN_DUE_TO_CRASH static void ftlUnreachable(
|
|
CodeBlock* codeBlock, BlockIndex blockIndex, unsigned nodeIndex)
|
|
{
|
|
dataLog("Crashing in thought-to-be-unreachable FTL-generated code for ", pointerDump(codeBlock), " at basic block #", blockIndex);
|
|
if (nodeIndex != UINT_MAX)
|
|
dataLog(", node @", nodeIndex);
|
|
dataLog(".\n");
|
|
CRASH();
|
|
}
|
|
#endif // ASSERT_ENABLED
|
|
|
|
// Using this instead of typeCheck() helps to reduce the load on B3, by creating
|
|
// significantly less dead code.
|
|
#define FTL_TYPE_CHECK_WITH_EXIT_KIND(exitKind, lowValue, highValue, typesPassedThrough, failCondition) do { \
|
|
FormattedValue _ftc_lowValue = (lowValue); \
|
|
Edge _ftc_highValue = (highValue); \
|
|
SpeculatedType _ftc_typesPassedThrough = (typesPassedThrough); \
|
|
if (!m_interpreter.needsTypeCheck(_ftc_highValue, _ftc_typesPassedThrough)) \
|
|
break; \
|
|
typeCheck(_ftc_lowValue, _ftc_highValue, _ftc_typesPassedThrough, (failCondition), exitKind); \
|
|
} while (false)
|
|
|
|
#define FTL_TYPE_CHECK(lowValue, highValue, typesPassedThrough, failCondition) \
|
|
FTL_TYPE_CHECK_WITH_EXIT_KIND(BadType, lowValue, highValue, typesPassedThrough, failCondition)
|
|
|
|
class LowerDFGToB3 {
|
|
WTF_MAKE_NONCOPYABLE(LowerDFGToB3);
|
|
public:
|
|
LowerDFGToB3(State& state)
|
|
: m_graph(state.graph)
|
|
, m_ftlState(state)
|
|
, m_out(state)
|
|
, m_proc(*state.proc)
|
|
, m_availabilityCalculator(m_graph)
|
|
, m_state(state.graph)
|
|
, m_interpreter(state.graph, m_state)
|
|
{
|
|
if (Options::validateAbstractInterpreterState()) {
|
|
performLivenessAnalysis(m_graph);
|
|
|
|
// We only use node liveness here, not combined liveness, as we only track
|
|
// AI state for live nodes.
|
|
for (DFG::BasicBlock* block : m_graph.blocksInNaturalOrder()) {
|
|
NodeSet live;
|
|
|
|
for (NodeFlowProjection node : block->ssa->liveAtTail) {
|
|
if (node.kind() == NodeFlowProjection::Primary)
|
|
live.addVoid(node.node());
|
|
}
|
|
|
|
for (unsigned i = block->size(); i--; ) {
|
|
Node* node = block->at(i);
|
|
live.remove(node);
|
|
m_graph.doToChildren(node, [&] (Edge child) {
|
|
live.addVoid(child.node());
|
|
});
|
|
m_liveInToNode.add(node, live);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
void lower()
|
|
{
|
|
State* state = &m_ftlState;
|
|
|
|
CString name;
|
|
if (verboseCompilationEnabled()) {
|
|
name = toCString(
|
|
"jsBody_", ++compileCounter, "_", codeBlock()->inferredName(),
|
|
"_", codeBlock()->hash());
|
|
} else
|
|
name = "jsBody";
|
|
|
|
{
|
|
m_proc.setNumEntrypoints(m_graph.m_numberOfEntrypoints);
|
|
CodeBlock* codeBlock = m_graph.m_codeBlock;
|
|
|
|
Ref<B3::Air::PrologueGenerator> catchPrologueGenerator = createSharedTask<B3::Air::PrologueGeneratorFunction>(
|
|
[codeBlock] (CCallHelpers& jit, B3::Air::Code& code) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
jit.addPtr(CCallHelpers::TrustedImm32(-code.frameSize()), GPRInfo::callFrameRegister, CCallHelpers::stackPointerRegister);
|
|
|
|
jit.emitSave(code.calleeSaveRegisterAtOffsetList());
|
|
jit.emitPutToCallFrameHeader(codeBlock, VirtualRegister(CallFrameSlot::codeBlock));
|
|
});
|
|
|
|
for (unsigned catchEntrypointIndex : m_graph.m_entrypointIndexToCatchBytecodeIndex.keys()) {
|
|
RELEASE_ASSERT(catchEntrypointIndex != 0);
|
|
m_proc.code().setPrologueForEntrypoint(catchEntrypointIndex, catchPrologueGenerator.copyRef());
|
|
}
|
|
|
|
if (m_graph.m_maxLocalsForCatchOSREntry) {
|
|
uint32_t numberOfLiveLocals = std::max(*m_graph.m_maxLocalsForCatchOSREntry, 1u); // Make sure we always allocate a non-null catchOSREntryBuffer.
|
|
m_ftlState.jitCode->common.catchOSREntryBuffer = m_graph.m_vm.scratchBufferForSize(sizeof(JSValue) * numberOfLiveLocals);
|
|
}
|
|
}
|
|
|
|
m_graph.ensureSSADominators();
|
|
|
|
if (verboseCompilationEnabled())
|
|
dataLog("Function ready, beginning lowering.\n");
|
|
|
|
m_out.initialize(m_heaps);
|
|
|
|
// We use prologue frequency for all of the initialization code.
|
|
m_out.setFrequency(1);
|
|
|
|
bool hasMultipleEntrypoints = m_graph.m_numberOfEntrypoints > 1;
|
|
|
|
LBasicBlock prologue = m_out.newBlock();
|
|
LBasicBlock callEntrypointArgumentSpeculations = hasMultipleEntrypoints ? m_out.newBlock() : nullptr;
|
|
m_handleExceptions = m_out.newBlock();
|
|
|
|
for (BlockIndex blockIndex = 0; blockIndex < m_graph.numBlocks(); ++blockIndex) {
|
|
m_highBlock = m_graph.block(blockIndex);
|
|
if (!m_highBlock)
|
|
continue;
|
|
m_out.setFrequency(m_highBlock->executionCount);
|
|
m_blocks.add(m_highBlock, m_out.newBlock());
|
|
}
|
|
|
|
// Back to prologue frequency for any bocks that get sneakily created in the initialization code.
|
|
m_out.setFrequency(1);
|
|
|
|
m_out.appendTo(prologue, hasMultipleEntrypoints ? callEntrypointArgumentSpeculations : m_handleExceptions);
|
|
m_out.initializeConstants(m_proc, prologue);
|
|
createPhiVariables();
|
|
|
|
size_t sizeOfCaptured = sizeof(JSValue) * m_graph.m_nextMachineLocal;
|
|
B3::SlotBaseValue* capturedBase = m_out.lockedStackSlot(sizeOfCaptured);
|
|
m_captured = m_out.add(capturedBase, m_out.constIntPtr(sizeOfCaptured));
|
|
state->capturedValue = capturedBase->slot();
|
|
|
|
auto preOrder = m_graph.blocksInPreOrder();
|
|
|
|
VM* vm = &this->vm();
|
|
|
|
m_callFrame = m_out.framePointer();
|
|
m_vmValue = m_out.constIntPtr(vm);
|
|
m_numberTag = m_out.constInt64(JSValue::NumberTag);
|
|
m_notCellMask = m_out.constInt64(JSValue::NotCellMask);
|
|
if (Options::validateDFGClobberize())
|
|
m_out.store32As8(m_out.int32Zero, m_out.absolute(reinterpret_cast<char*>(vm) + OBJECT_OFFSETOF(VM, didEnterVM)));
|
|
|
|
// Make sure that B3 knows that we really care about the mask registers. This forces the
|
|
// constants to be materialized in registers.
|
|
m_proc.addFastConstant(m_numberTag->key());
|
|
m_proc.addFastConstant(m_notCellMask->key());
|
|
|
|
// We don't want the CodeBlock to have a weak pointer to itself because
|
|
// that would cause it to always get collected.
|
|
m_out.storePtr(m_out.constIntPtr(bitwise_cast<intptr_t>(codeBlock())), addressFor(VirtualRegister(CallFrameSlot::codeBlock)));
|
|
|
|
// Stack Overflow Check.
|
|
unsigned exitFrameSize = m_graph.requiredRegisterCountForExit() * sizeof(Register);
|
|
MacroAssembler::AbsoluteAddress addressOfStackLimit(vm->addressOfSoftStackLimit());
|
|
PatchpointValue* stackOverflowHandler = m_out.patchpoint(Void);
|
|
CallSiteIndex callSiteIndex = callSiteIndexForCodeOrigin(m_ftlState, CodeOrigin(BytecodeIndex(0)));
|
|
stackOverflowHandler->appendSomeRegister(m_callFrame);
|
|
stackOverflowHandler->clobber(RegisterSet::macroScratchRegisters());
|
|
stackOverflowHandler->numGPScratchRegisters = 1;
|
|
stackOverflowHandler->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
GPRReg fp = params[0].gpr();
|
|
GPRReg scratch = params.gpScratch(0);
|
|
|
|
unsigned ftlFrameSize = params.proc().frameSize();
|
|
unsigned maxFrameSize = std::max(exitFrameSize, ftlFrameSize);
|
|
|
|
jit.addPtr(MacroAssembler::TrustedImm32(-maxFrameSize), fp, scratch);
|
|
MacroAssembler::JumpList stackOverflow;
|
|
if (UNLIKELY(maxFrameSize > Options::reservedZoneSize()))
|
|
stackOverflow.append(jit.branchPtr(MacroAssembler::Above, scratch, fp));
|
|
stackOverflow.append(jit.branchPtr(MacroAssembler::Above, addressOfStackLimit, scratch));
|
|
|
|
params.addLatePath([=] (CCallHelpers& jit) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
stackOverflow.link(&jit);
|
|
|
|
// FIXME: We would not have to do this if the stack check was part of the Air
|
|
// prologue. Then, we would know that there is no way for the callee-saves to
|
|
// get clobbered.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=172456
|
|
jit.emitRestore(params.proc().calleeSaveRegisterAtOffsetList());
|
|
|
|
jit.store32(
|
|
MacroAssembler::TrustedImm32(callSiteIndex.bits()),
|
|
CCallHelpers::tagFor(VirtualRegister(CallFrameSlot::argumentCountIncludingThis)));
|
|
jit.copyCalleeSavesToEntryFrameCalleeSavesBuffer(vm->topEntryFrame);
|
|
|
|
jit.move(CCallHelpers::TrustedImmPtr(jit.codeBlock()), GPRInfo::argumentGPR0);
|
|
jit.prepareCallOperation(*vm);
|
|
CCallHelpers::Call throwCall = jit.call(OperationPtrTag);
|
|
|
|
jit.move(CCallHelpers::TrustedImmPtr(vm), GPRInfo::argumentGPR0);
|
|
jit.prepareCallOperation(*vm);
|
|
CCallHelpers::Call lookupExceptionHandlerCall = jit.call(OperationPtrTag);
|
|
jit.jumpToExceptionHandler(*vm);
|
|
|
|
jit.addLinkTask(
|
|
[=] (LinkBuffer& linkBuffer) {
|
|
linkBuffer.link(throwCall, FunctionPtr<OperationPtrTag>(operationThrowStackOverflowError));
|
|
linkBuffer.link(lookupExceptionHandlerCall, FunctionPtr<OperationPtrTag>(operationLookupExceptionHandlerFromCallerFrame));
|
|
});
|
|
});
|
|
});
|
|
|
|
LBasicBlock firstDFGBasicBlock = lowBlock(m_graph.block(0));
|
|
|
|
{
|
|
if (hasMultipleEntrypoints) {
|
|
Vector<LBasicBlock> successors(m_graph.m_numberOfEntrypoints);
|
|
successors[0] = callEntrypointArgumentSpeculations;
|
|
for (unsigned i = 1; i < m_graph.m_numberOfEntrypoints; ++i) {
|
|
// Currently, the only other entrypoint is an op_catch entrypoint.
|
|
// We do OSR entry at op_catch, and we prove argument formats before
|
|
// jumping to FTL code, so we don't need to check argument types here
|
|
// for these entrypoints.
|
|
successors[i] = firstDFGBasicBlock;
|
|
}
|
|
|
|
m_out.entrySwitch(successors);
|
|
m_out.appendTo(callEntrypointArgumentSpeculations, m_handleExceptions);
|
|
}
|
|
|
|
m_node = nullptr;
|
|
m_nodeIndexInGraph = 0;
|
|
m_origin = NodeOrigin(CodeOrigin(BytecodeIndex(0)), CodeOrigin(BytecodeIndex(0)), true);
|
|
|
|
// Check Arguments.
|
|
availabilityMap().clear();
|
|
availabilityMap().m_locals = Operands<Availability>(codeBlock()->numParameters(), 0, 0);
|
|
for (unsigned i = codeBlock()->numParameters(); i--;) {
|
|
availabilityMap().m_locals.argument(i) =
|
|
Availability(FlushedAt(FlushedJSValue, virtualRegisterForArgumentIncludingThis(i)));
|
|
}
|
|
|
|
if (m_graph.m_plan.mode() == FTLForOSREntryMode) {
|
|
auto* jitCode = m_ftlState.jitCode->ftlForOSREntry();
|
|
jitCode->argumentFlushFormats().reserveInitialCapacity(codeBlock()->numParameters());
|
|
for (unsigned i = 0; i < codeBlock()->numParameters(); ++i)
|
|
jitCode->argumentFlushFormats().uncheckedAppend(m_graph.m_argumentFormats[0][i]);
|
|
} else {
|
|
for (unsigned i = codeBlock()->numParameters(); i--;) {
|
|
MethodOfGettingAValueProfile profile(&m_graph.m_profiledBlock->valueProfileForArgument(i));
|
|
VirtualRegister operand = virtualRegisterForArgumentIncludingThis(i);
|
|
LValue jsValue = m_out.load64(addressFor(operand));
|
|
|
|
switch (m_graph.m_argumentFormats[0][i]) {
|
|
case FlushedInt32:
|
|
speculate(BadType, jsValueValue(jsValue), profile, isNotInt32(jsValue));
|
|
break;
|
|
case FlushedBoolean:
|
|
speculate(BadType, jsValueValue(jsValue), profile, isNotBoolean(jsValue));
|
|
break;
|
|
case FlushedCell:
|
|
speculate(BadType, jsValueValue(jsValue), profile, isNotCell(jsValue));
|
|
break;
|
|
case FlushedJSValue:
|
|
break;
|
|
default:
|
|
DFG_CRASH(m_graph, nullptr, "Bad flush format for argument");
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
m_out.jump(firstDFGBasicBlock);
|
|
}
|
|
|
|
|
|
m_out.appendTo(m_handleExceptions, firstDFGBasicBlock);
|
|
Box<CCallHelpers::Label> exceptionHandler = state->exceptionHandler;
|
|
m_out.patchpoint(Void)->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams&) {
|
|
CCallHelpers::Jump jump = jit.jump();
|
|
jit.addLinkTask(
|
|
[=] (LinkBuffer& linkBuffer) {
|
|
linkBuffer.link(jump, linkBuffer.locationOf<ExceptionHandlerPtrTag>(*exceptionHandler));
|
|
});
|
|
});
|
|
m_out.unreachable();
|
|
|
|
for (DFG::BasicBlock* block : preOrder)
|
|
compileBlock(block);
|
|
|
|
// Make sure everything is decorated. This does a bunch of deferred decorating. This has
|
|
// to happen last because our abstract heaps are generated lazily. They have to be
|
|
// generated lazily because we have an infinite number of numbered, indexed, and
|
|
// absolute heaps. We only become aware of the ones we actually mention while lowering.
|
|
m_heaps.computeRangesAndDecorateInstructions();
|
|
|
|
// We create all Phi's up front, but we may then decide not to compile the basic block
|
|
// that would have contained one of them. So this creates orphans, which triggers B3
|
|
// validation failures. Calling this fixes the issue.
|
|
//
|
|
// Note that you should avoid the temptation to make this call conditional upon
|
|
// validation being enabled. B3 makes no guarantees of any kind of correctness when
|
|
// dealing with IR that would have failed validation. For example, it would be valid to
|
|
// write a B3 phase that so aggressively assumes the lack of orphans that it would crash
|
|
// if any orphans were around. We might even have such phases already.
|
|
m_proc.deleteOrphans();
|
|
|
|
// We put the blocks into the B3 procedure in a super weird order. Now we reorder them.
|
|
m_out.applyBlockOrder();
|
|
}
|
|
|
|
private:
|
|
|
|
void createPhiVariables()
|
|
{
|
|
for (BlockIndex blockIndex = m_graph.numBlocks(); blockIndex--;) {
|
|
DFG::BasicBlock* block = m_graph.block(blockIndex);
|
|
if (!block)
|
|
continue;
|
|
for (unsigned nodeIndex = block->size(); nodeIndex--;) {
|
|
Node* node = block->at(nodeIndex);
|
|
if (node->op() != DFG::Phi)
|
|
continue;
|
|
LType type;
|
|
switch (node->flags() & NodeResultMask) {
|
|
case NodeResultDouble:
|
|
type = Double;
|
|
break;
|
|
case NodeResultInt32:
|
|
type = Int32;
|
|
break;
|
|
case NodeResultInt52:
|
|
type = Int64;
|
|
break;
|
|
case NodeResultBoolean:
|
|
type = Int32;
|
|
break;
|
|
case NodeResultJS:
|
|
type = Int64;
|
|
break;
|
|
default:
|
|
DFG_CRASH(m_graph, node, "Bad Phi node result type");
|
|
break;
|
|
}
|
|
m_phis.add(node, m_proc.add<Value>(B3::Phi, type, Origin(node)));
|
|
}
|
|
}
|
|
}
|
|
|
|
void compileBlock(DFG::BasicBlock* block)
|
|
{
|
|
if (!block)
|
|
return;
|
|
|
|
if (verboseCompilationEnabled())
|
|
dataLog("Compiling block ", *block, "\n");
|
|
|
|
m_highBlock = block;
|
|
|
|
// Make sure that any blocks created while lowering code in the high block have the frequency of
|
|
// the high block. This is appropriate because B3 doesn't need precise frequencies. It just needs
|
|
// something roughly approximate for things like register allocation.
|
|
m_out.setFrequency(m_highBlock->executionCount);
|
|
|
|
LBasicBlock lowBlock = m_blocks.get(m_highBlock);
|
|
|
|
m_nextHighBlock = nullptr;
|
|
for (BlockIndex nextBlockIndex = m_highBlock->index + 1; nextBlockIndex < m_graph.numBlocks(); ++nextBlockIndex) {
|
|
m_nextHighBlock = m_graph.block(nextBlockIndex);
|
|
if (m_nextHighBlock)
|
|
break;
|
|
}
|
|
m_nextLowBlock = m_nextHighBlock ? m_blocks.get(m_nextHighBlock) : nullptr;
|
|
|
|
// All of this effort to find the next block gives us the ability to keep the
|
|
// generated IR in roughly program order. This ought not affect the performance
|
|
// of the generated code (since we expect B3 to reorder things) but it will
|
|
// make IR dumps easier to read.
|
|
m_out.appendTo(lowBlock, m_nextLowBlock);
|
|
|
|
if (Options::ftlCrashes())
|
|
m_out.trap();
|
|
|
|
if (!m_highBlock->cfaHasVisited) {
|
|
if (verboseCompilationEnabled())
|
|
dataLog("Bailing because CFA didn't reach.\n");
|
|
crash(m_highBlock, nullptr);
|
|
return;
|
|
}
|
|
|
|
m_aiCheckedNodes.clear();
|
|
|
|
m_availabilityCalculator.beginBlock(m_highBlock);
|
|
|
|
m_state.reset();
|
|
m_state.beginBasicBlock(m_highBlock);
|
|
|
|
if (Options::validateDFGClobberize()) {
|
|
bool clobberedWorld = m_highBlock->predecessors.isEmpty() || m_highBlock->isOSRTarget || m_highBlock->isCatchEntrypoint;
|
|
auto validateClobberize = [&] () {
|
|
clobberedWorld = true;
|
|
};
|
|
|
|
for (auto* predecessor : m_highBlock->predecessors)
|
|
clobberize(m_graph, predecessor->terminal(), [] (auto...) { }, [] (auto...) { }, [] (auto...) { }, validateClobberize);
|
|
|
|
if (!clobberedWorld) {
|
|
LValue didNotEnterVM = m_out.notZero32(m_out.load8ZeroExt32(m_out.absolute(&vm().didEnterVM)));
|
|
auto* check = m_out.speculate(didNotEnterVM);
|
|
check->setGenerator([=] (CCallHelpers& jit, const StackmapGenerationParams&) {
|
|
jit.breakpoint();
|
|
});
|
|
} else
|
|
m_out.store(m_out.int32Zero, m_out.absolute(&vm().didEnterVM));
|
|
}
|
|
|
|
for (unsigned nodeIndex = 0; nodeIndex < m_highBlock->size(); ++nodeIndex) {
|
|
if (!compileNode(nodeIndex))
|
|
break;
|
|
}
|
|
}
|
|
|
|
void safelyInvalidateAfterTermination()
|
|
{
|
|
if (verboseCompilationEnabled())
|
|
dataLog("Bailing.\n");
|
|
crash();
|
|
|
|
// Invalidate dominated blocks. Under normal circumstances we would expect
|
|
// them to be invalidated already. But you can have the CFA become more
|
|
// precise over time because the structures of objects change on the main
|
|
// thread. Failing to do this would result in weird crashes due to a value
|
|
// being used but not defined. Race conditions FTW!
|
|
for (BlockIndex blockIndex = m_graph.numBlocks(); blockIndex--;) {
|
|
DFG::BasicBlock* target = m_graph.block(blockIndex);
|
|
if (!target)
|
|
continue;
|
|
if (m_graph.m_ssaDominators->dominates(m_highBlock, target)) {
|
|
if (verboseCompilationEnabled())
|
|
dataLog("Block ", *target, " will bail also.\n");
|
|
target->cfaHasVisited = false;
|
|
}
|
|
}
|
|
}
|
|
|
|
void validateAIState(Node* node)
|
|
{
|
|
if (!m_graphDump) {
|
|
StringPrintStream out;
|
|
m_graph.dump(out);
|
|
m_graphDump = out.tryToString().value_or("<out of memory while dumping graph>"_s);
|
|
}
|
|
|
|
switch (node->op()) {
|
|
case MovHint:
|
|
case JSConstant:
|
|
case LazyJSConstant:
|
|
case DoubleConstant:
|
|
case Int52Constant:
|
|
case GetStack:
|
|
case PutStack:
|
|
case KillStack:
|
|
case ExitOK:
|
|
return;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
// Before we execute node.
|
|
NodeSet& live = m_liveInToNode.find(node)->value;
|
|
unsigned highParentIndex = node->index();
|
|
{
|
|
uint64_t hash = WTF::intHash(highParentIndex);
|
|
if (hash >= static_cast<uint64_t>((static_cast<double>(std::numeric_limits<unsigned>::max()) + 1) * Options::validateAbstractInterpreterStateProbability()))
|
|
return;
|
|
}
|
|
|
|
for (Node* node : live) {
|
|
if (node->isPhantomAllocation())
|
|
continue;
|
|
|
|
if (node->op() == AssertInBounds)
|
|
continue;
|
|
if (node->op() == CheckInBounds)
|
|
continue;
|
|
|
|
AbstractValue value = m_interpreter.forNode(node);
|
|
{
|
|
auto iter = m_aiCheckedNodes.find(node);
|
|
if (iter != m_aiCheckedNodes.end()) {
|
|
AbstractValue checkedValue = iter->value;
|
|
if (checkedValue == value) {
|
|
if (!(value.m_type & SpecCell))
|
|
continue;
|
|
}
|
|
}
|
|
m_aiCheckedNodes.set(node, value);
|
|
}
|
|
|
|
FlushFormat flushFormat;
|
|
LValue input;
|
|
if (node->hasJSResult()) {
|
|
input = lowJSValue(Edge(node, UntypedUse));
|
|
flushFormat = FlushedJSValue;
|
|
} else if (node->hasDoubleResult()) {
|
|
input = lowDouble(Edge(node, DoubleRepUse));
|
|
flushFormat = FlushedDouble;
|
|
} else if (node->hasInt52Result()) {
|
|
input = strictInt52ToJSValue(lowStrictInt52(Edge(node, Int52RepUse)));
|
|
flushFormat = FlushedInt52;
|
|
} else
|
|
continue;
|
|
|
|
unsigned highChildIndex = node->index();
|
|
|
|
String graphDump = m_graphDump;
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Void);
|
|
patchpoint->effects = Effects::none();
|
|
patchpoint->effects.writesLocalState = true;
|
|
patchpoint->appendSomeRegister(input);
|
|
patchpoint->setGenerator([=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
GPRReg reg = InvalidGPRReg;
|
|
FPRReg fpReg = InvalidFPRReg;
|
|
if (flushFormat == FlushedDouble)
|
|
fpReg = params[0].fpr();
|
|
else
|
|
reg = params[0].gpr();
|
|
jit.probe([=] (Probe::Context& context) {
|
|
JSValue input;
|
|
double doubleInput;
|
|
|
|
auto dumpAndCrash = [&] {
|
|
dataLogLn("Validation failed at node: @", highParentIndex);
|
|
dataLogLn("Failed validating live value: @", highChildIndex);
|
|
dataLogLn();
|
|
dataLogLn("Expected AI value = ", value);
|
|
if (flushFormat != FlushedDouble)
|
|
dataLogLn("Unexpected value = ", input);
|
|
else
|
|
dataLogLn("Unexpected double value = ", doubleInput);
|
|
dataLogLn();
|
|
dataLogLn(graphDump);
|
|
CRASH();
|
|
};
|
|
|
|
if (flushFormat == FlushedDouble) {
|
|
doubleInput = context.fpr(fpReg);
|
|
SpeculatedType type;
|
|
if (!std::isnan(doubleInput))
|
|
type = speculationFromValue(jsDoubleNumber(doubleInput));
|
|
else if (isImpureNaN(doubleInput))
|
|
type = SpecDoubleImpureNaN;
|
|
else
|
|
type = SpecDoublePureNaN;
|
|
|
|
if (!value.couldBeType(type))
|
|
dumpAndCrash();
|
|
} else {
|
|
input = JSValue::decode(context.gpr(reg));
|
|
if (flushFormat == FlushedInt52) {
|
|
RELEASE_ASSERT(input.isAnyInt());
|
|
input = jsDoubleNumber(input.asAnyInt());
|
|
}
|
|
if (!value.validateOSREntryValue(input, flushFormat))
|
|
dumpAndCrash();
|
|
}
|
|
|
|
});
|
|
});
|
|
}
|
|
}
|
|
|
|
bool compileNode(unsigned nodeIndex)
|
|
{
|
|
if (!m_state.isValid()) {
|
|
safelyInvalidateAfterTermination();
|
|
return false;
|
|
}
|
|
|
|
m_node = m_highBlock->at(nodeIndex);
|
|
m_nodeIndexInGraph = m_node->index();
|
|
m_origin = m_node->origin;
|
|
m_out.setOrigin(m_node);
|
|
|
|
if (verboseCompilationEnabled())
|
|
dataLog("Lowering ", m_node, "\n");
|
|
|
|
m_interpreter.startExecuting();
|
|
m_interpreter.executeKnownEdgeTypes(m_node);
|
|
|
|
if (Options::validateAbstractInterpreterState())
|
|
validateAIState(m_node);
|
|
|
|
if constexpr (validateDFGDoesGC) {
|
|
if (Options::validateDoesGC()) {
|
|
bool expectDoesGC = doesGC(m_graph, m_node);
|
|
m_out.store(m_out.constInt32(DoesGCCheck::encode(expectDoesGC, m_node->index(), m_node->op())), m_out.absolute(vm().heap.addressOfDoesGC()));
|
|
}
|
|
}
|
|
|
|
switch (m_node->op()) {
|
|
case DFG::Upsilon:
|
|
compileUpsilon();
|
|
break;
|
|
case DFG::Phi:
|
|
compilePhi();
|
|
break;
|
|
case JSConstant:
|
|
break;
|
|
case DoubleConstant:
|
|
compileDoubleConstant();
|
|
break;
|
|
case Int52Constant:
|
|
compileInt52Constant();
|
|
break;
|
|
case LazyJSConstant:
|
|
compileLazyJSConstant();
|
|
break;
|
|
case DoubleRep:
|
|
compileDoubleRep();
|
|
break;
|
|
case DoubleAsInt32:
|
|
compileDoubleAsInt32();
|
|
break;
|
|
case DFG::ValueRep:
|
|
compileValueRep();
|
|
break;
|
|
case Int52Rep:
|
|
compileInt52Rep();
|
|
break;
|
|
case ValueToInt32:
|
|
compileValueToInt32();
|
|
break;
|
|
case BooleanToNumber:
|
|
compileBooleanToNumber();
|
|
break;
|
|
case ExtractOSREntryLocal:
|
|
compileExtractOSREntryLocal();
|
|
break;
|
|
case ExtractCatchLocal:
|
|
compileExtractCatchLocal();
|
|
break;
|
|
case ClearCatchLocals:
|
|
compileClearCatchLocals();
|
|
break;
|
|
case GetStack:
|
|
compileGetStack();
|
|
break;
|
|
case PutStack:
|
|
compilePutStack();
|
|
break;
|
|
case DFG::Check:
|
|
case CheckVarargs:
|
|
compileNoOp();
|
|
break;
|
|
case ToObject:
|
|
case CallObjectConstructor:
|
|
compileToObjectOrCallObjectConstructor();
|
|
break;
|
|
case ToThis:
|
|
compileToThis();
|
|
break;
|
|
case Inc:
|
|
case Dec:
|
|
compileIncOrDec();
|
|
break;
|
|
case ValueNegate:
|
|
compileValueNegate();
|
|
break;
|
|
case ValueAdd:
|
|
compileValueAdd();
|
|
break;
|
|
case ValueSub:
|
|
compileValueSub();
|
|
break;
|
|
case ValueMul:
|
|
compileValueMul();
|
|
break;
|
|
case StrCat:
|
|
compileStrCat();
|
|
break;
|
|
case ArithAdd:
|
|
case ArithSub:
|
|
compileArithAddOrSub();
|
|
break;
|
|
case ArithClz32:
|
|
compileArithClz32();
|
|
break;
|
|
case ArithMul:
|
|
compileArithMul();
|
|
break;
|
|
case ValueDiv:
|
|
compileValueDiv();
|
|
break;
|
|
case ArithDiv:
|
|
compileArithDiv();
|
|
break;
|
|
case ValueMod:
|
|
compileValueMod();
|
|
break;
|
|
case ArithMod:
|
|
compileArithMod();
|
|
break;
|
|
case ArithMin:
|
|
case ArithMax:
|
|
compileArithMinOrMax();
|
|
break;
|
|
case ArithAbs:
|
|
compileArithAbs();
|
|
break;
|
|
case ValuePow:
|
|
compileValuePow();
|
|
break;
|
|
case ArithPow:
|
|
compileArithPow();
|
|
break;
|
|
case ArithRandom:
|
|
compileArithRandom();
|
|
break;
|
|
case ArithRound:
|
|
compileArithRound();
|
|
break;
|
|
case ArithFloor:
|
|
compileArithFloor();
|
|
break;
|
|
case ArithCeil:
|
|
compileArithCeil();
|
|
break;
|
|
case ArithTrunc:
|
|
compileArithTrunc();
|
|
break;
|
|
case ArithSqrt:
|
|
compileArithSqrt();
|
|
break;
|
|
case ArithFRound:
|
|
compileArithFRound();
|
|
break;
|
|
case ArithNegate:
|
|
compileArithNegate();
|
|
break;
|
|
case ArithUnary:
|
|
compileArithUnary();
|
|
break;
|
|
case ValueBitNot:
|
|
compileValueBitNot();
|
|
break;
|
|
case ArithBitNot:
|
|
compileArithBitNot();
|
|
break;
|
|
case ValueBitAnd:
|
|
compileValueBitAnd();
|
|
break;
|
|
case ArithBitAnd:
|
|
compileArithBitAnd();
|
|
break;
|
|
case ValueBitOr:
|
|
compileValueBitOr();
|
|
break;
|
|
case ArithBitOr:
|
|
compileArithBitOr();
|
|
break;
|
|
case ArithBitXor:
|
|
compileArithBitXor();
|
|
break;
|
|
case ValueBitXor:
|
|
compileValueBitXor();
|
|
break;
|
|
case ValueBitRShift:
|
|
compileValueBitRShift();
|
|
break;
|
|
case ArithBitRShift:
|
|
compileArithBitRShift();
|
|
break;
|
|
case ArithBitLShift:
|
|
compileArithBitLShift();
|
|
break;
|
|
case ValueBitLShift:
|
|
compileValueBitLShift();
|
|
break;
|
|
case BitURShift:
|
|
compileBitURShift();
|
|
break;
|
|
case UInt32ToNumber:
|
|
compileUInt32ToNumber();
|
|
break;
|
|
case CheckStructure:
|
|
compileCheckStructure();
|
|
break;
|
|
case CheckStructureOrEmpty:
|
|
compileCheckStructureOrEmpty();
|
|
break;
|
|
case CheckIsConstant:
|
|
compileCheckIsConstant();
|
|
break;
|
|
case CheckNotEmpty:
|
|
compileCheckNotEmpty();
|
|
break;
|
|
case AssertNotEmpty:
|
|
compileAssertNotEmpty();
|
|
break;
|
|
case CheckBadValue:
|
|
compileCheckBadValue();
|
|
break;
|
|
case CheckIdent:
|
|
compileCheckIdent();
|
|
break;
|
|
case GetExecutable:
|
|
compileGetExecutable();
|
|
break;
|
|
case Arrayify:
|
|
case ArrayifyToStructure:
|
|
compileArrayify();
|
|
break;
|
|
case PutStructure:
|
|
compilePutStructure();
|
|
break;
|
|
case TryGetById:
|
|
compileGetById(AccessType::TryGetById);
|
|
break;
|
|
case GetById:
|
|
case GetByIdFlush:
|
|
compileGetById(AccessType::GetById);
|
|
break;
|
|
case GetByIdWithThis:
|
|
compileGetByIdWithThis();
|
|
break;
|
|
case GetByIdDirect:
|
|
case GetByIdDirectFlush:
|
|
compileGetById(AccessType::GetByIdDirect);
|
|
break;
|
|
case GetPrivateName:
|
|
compileGetPrivateName();
|
|
break;
|
|
case GetPrivateNameById:
|
|
compileGetPrivateNameById();
|
|
break;
|
|
case InById:
|
|
compileInById();
|
|
break;
|
|
case InByVal:
|
|
compileInByVal();
|
|
break;
|
|
case HasOwnProperty:
|
|
compileHasOwnProperty();
|
|
break;
|
|
case PutById:
|
|
case PutByIdDirect:
|
|
case PutByIdFlush:
|
|
compilePutById();
|
|
break;
|
|
case PutByIdWithThis:
|
|
compilePutByIdWithThis();
|
|
break;
|
|
case PutGetterById:
|
|
case PutSetterById:
|
|
compilePutAccessorById();
|
|
break;
|
|
case PutGetterSetterById:
|
|
compilePutGetterSetterById();
|
|
break;
|
|
case PutGetterByVal:
|
|
case PutSetterByVal:
|
|
compilePutAccessorByVal();
|
|
break;
|
|
case DeleteById:
|
|
compileDeleteById();
|
|
break;
|
|
case DeleteByVal:
|
|
compileDeleteByVal();
|
|
break;
|
|
case GetButterfly:
|
|
compileGetButterfly();
|
|
break;
|
|
case ConstantStoragePointer:
|
|
compileConstantStoragePointer();
|
|
break;
|
|
case GetIndexedPropertyStorage:
|
|
compileGetIndexedPropertyStorage();
|
|
break;
|
|
case CheckArray:
|
|
compileCheckArray();
|
|
break;
|
|
case CheckArrayOrEmpty:
|
|
compileCheckArrayOrEmpty();
|
|
break;
|
|
case CheckDetached:
|
|
compileCheckDetached();
|
|
break;
|
|
case GetArrayLength:
|
|
compileGetArrayLength();
|
|
break;
|
|
case GetVectorLength:
|
|
compileGetVectorLength();
|
|
break;
|
|
case AssertInBounds:
|
|
compileAssertInBounds();
|
|
break;
|
|
case CheckInBounds:
|
|
compileCheckInBounds();
|
|
break;
|
|
case GetByVal:
|
|
compileGetByVal();
|
|
break;
|
|
case GetMyArgumentByVal:
|
|
case GetMyArgumentByValOutOfBounds:
|
|
compileGetMyArgumentByVal();
|
|
break;
|
|
case GetByValWithThis:
|
|
compileGetByValWithThis();
|
|
break;
|
|
case PutByVal:
|
|
case PutByValAlias:
|
|
case PutByValDirect:
|
|
compilePutByVal();
|
|
break;
|
|
case PutByValWithThis:
|
|
compilePutByValWithThis();
|
|
break;
|
|
case PutPrivateName:
|
|
compilePutPrivateName();
|
|
break;
|
|
case PutPrivateNameById:
|
|
compilePutPrivateNameById();
|
|
break;
|
|
case AtomicsAdd:
|
|
case AtomicsAnd:
|
|
case AtomicsCompareExchange:
|
|
case AtomicsExchange:
|
|
case AtomicsLoad:
|
|
case AtomicsOr:
|
|
case AtomicsStore:
|
|
case AtomicsSub:
|
|
case AtomicsXor:
|
|
compileAtomicsReadModifyWrite();
|
|
break;
|
|
case AtomicsIsLockFree:
|
|
compileAtomicsIsLockFree();
|
|
break;
|
|
case DefineDataProperty:
|
|
compileDefineDataProperty();
|
|
break;
|
|
case DefineAccessorProperty:
|
|
compileDefineAccessorProperty();
|
|
break;
|
|
case ArrayPush:
|
|
compileArrayPush();
|
|
break;
|
|
case ArrayPop:
|
|
compileArrayPop();
|
|
break;
|
|
case ArraySlice:
|
|
compileArraySlice();
|
|
break;
|
|
case ArrayIndexOf:
|
|
compileArrayIndexOf();
|
|
break;
|
|
case CreateActivation:
|
|
compileCreateActivation();
|
|
break;
|
|
case PushWithScope:
|
|
compilePushWithScope();
|
|
break;
|
|
case NewFunction:
|
|
case NewGeneratorFunction:
|
|
case NewAsyncGeneratorFunction:
|
|
case NewAsyncFunction:
|
|
compileNewFunction();
|
|
break;
|
|
case CreateDirectArguments:
|
|
compileCreateDirectArguments();
|
|
break;
|
|
case CreateScopedArguments:
|
|
compileCreateScopedArguments();
|
|
break;
|
|
case CreateClonedArguments:
|
|
compileCreateClonedArguments();
|
|
break;
|
|
case CreateArgumentsButterfly:
|
|
compileCreateArgumentsButterfly();
|
|
break;
|
|
case ObjectCreate:
|
|
compileObjectCreate();
|
|
break;
|
|
case ObjectKeys:
|
|
case ObjectGetOwnPropertyNames:
|
|
compileObjectKeysOrObjectGetOwnPropertyNames();
|
|
break;
|
|
case NewObject:
|
|
compileNewObject();
|
|
break;
|
|
case NewGenerator:
|
|
compileNewGenerator();
|
|
break;
|
|
case NewAsyncGenerator:
|
|
compileNewAsyncGenerator();
|
|
break;
|
|
case NewInternalFieldObject:
|
|
compileNewInternalFieldObject();
|
|
break;
|
|
case NewStringObject:
|
|
compileNewStringObject();
|
|
break;
|
|
case NewSymbol:
|
|
compileNewSymbol();
|
|
break;
|
|
case NewArray:
|
|
compileNewArray();
|
|
break;
|
|
case NewArrayWithSpread:
|
|
compileNewArrayWithSpread();
|
|
break;
|
|
case CreateThis:
|
|
compileCreateThis();
|
|
break;
|
|
case CreatePromise:
|
|
compileCreatePromise();
|
|
break;
|
|
case CreateGenerator:
|
|
compileCreateGenerator();
|
|
break;
|
|
case CreateAsyncGenerator:
|
|
compileCreateAsyncGenerator();
|
|
break;
|
|
case Spread:
|
|
compileSpread();
|
|
break;
|
|
case NewArrayBuffer:
|
|
compileNewArrayBuffer();
|
|
break;
|
|
case NewArrayWithSize:
|
|
compileNewArrayWithSize();
|
|
break;
|
|
case NewTypedArray:
|
|
compileNewTypedArray();
|
|
break;
|
|
case GetTypedArrayByteOffset:
|
|
compileGetTypedArrayByteOffset();
|
|
break;
|
|
case GetPrototypeOf:
|
|
compileGetPrototypeOf();
|
|
break;
|
|
case AllocatePropertyStorage:
|
|
compileAllocatePropertyStorage();
|
|
break;
|
|
case ReallocatePropertyStorage:
|
|
compileReallocatePropertyStorage();
|
|
break;
|
|
case NukeStructureAndSetButterfly:
|
|
compileNukeStructureAndSetButterfly();
|
|
break;
|
|
case ToNumber:
|
|
compileToNumber();
|
|
break;
|
|
case ToNumeric:
|
|
compileToNumeric();
|
|
break;
|
|
case CallNumberConstructor:
|
|
compileCallNumberConstructor();
|
|
break;
|
|
case ToString:
|
|
case CallStringConstructor:
|
|
case StringValueOf:
|
|
compileToStringOrCallStringConstructorOrStringValueOf();
|
|
break;
|
|
case ToPrimitive:
|
|
compileToPrimitive();
|
|
break;
|
|
case ToPropertyKey:
|
|
compileToPropertyKey();
|
|
break;
|
|
case MakeRope:
|
|
compileMakeRope();
|
|
break;
|
|
case StringCharAt:
|
|
compileStringCharAt();
|
|
break;
|
|
case StringCharCodeAt:
|
|
compileStringCharCodeAt();
|
|
break;
|
|
case StringCodePointAt:
|
|
compileStringCodePointAt();
|
|
break;
|
|
case StringFromCharCode:
|
|
compileStringFromCharCode();
|
|
break;
|
|
case GetByOffset:
|
|
case GetGetterSetterByOffset:
|
|
compileGetByOffset();
|
|
break;
|
|
case GetGetter:
|
|
compileGetGetter();
|
|
break;
|
|
case GetSetter:
|
|
compileGetSetter();
|
|
break;
|
|
case MultiGetByOffset:
|
|
compileMultiGetByOffset();
|
|
break;
|
|
case PutByOffset:
|
|
compilePutByOffset();
|
|
break;
|
|
case MultiPutByOffset:
|
|
compileMultiPutByOffset();
|
|
break;
|
|
case MultiDeleteByOffset:
|
|
compileMultiDeleteByOffset();
|
|
break;
|
|
case MatchStructure:
|
|
compileMatchStructure();
|
|
break;
|
|
case GetGlobalVar:
|
|
case GetGlobalLexicalVariable:
|
|
compileGetGlobalVariable();
|
|
break;
|
|
case PutGlobalVariable:
|
|
compilePutGlobalVariable();
|
|
break;
|
|
case NotifyWrite:
|
|
compileNotifyWrite();
|
|
break;
|
|
case GetCallee:
|
|
compileGetCallee();
|
|
break;
|
|
case SetCallee:
|
|
compileSetCallee();
|
|
break;
|
|
case GetArgumentCountIncludingThis:
|
|
compileGetArgumentCountIncludingThis();
|
|
break;
|
|
case SetArgumentCountIncludingThis:
|
|
compileSetArgumentCountIncludingThis();
|
|
break;
|
|
case GetScope:
|
|
compileGetScope();
|
|
break;
|
|
case SkipScope:
|
|
compileSkipScope();
|
|
break;
|
|
case GetGlobalObject:
|
|
compileGetGlobalObject();
|
|
break;
|
|
case GetGlobalThis:
|
|
compileGetGlobalThis();
|
|
break;
|
|
case GetClosureVar:
|
|
compileGetClosureVar();
|
|
break;
|
|
case PutClosureVar:
|
|
compilePutClosureVar();
|
|
break;
|
|
case GetInternalField:
|
|
compileGetInternalField();
|
|
break;
|
|
case PutInternalField:
|
|
compilePutInternalField();
|
|
break;
|
|
case GetFromArguments:
|
|
compileGetFromArguments();
|
|
break;
|
|
case PutToArguments:
|
|
compilePutToArguments();
|
|
break;
|
|
case GetArgument:
|
|
compileGetArgument();
|
|
break;
|
|
case CompareEq:
|
|
compileCompareEq();
|
|
break;
|
|
case CompareStrictEq:
|
|
compileCompareStrictEq();
|
|
break;
|
|
case CompareLess:
|
|
compileCompareLess();
|
|
break;
|
|
case CompareLessEq:
|
|
compileCompareLessEq();
|
|
break;
|
|
case CompareGreater:
|
|
compileCompareGreater();
|
|
break;
|
|
case CompareGreaterEq:
|
|
compileCompareGreaterEq();
|
|
break;
|
|
case CompareBelow:
|
|
compileCompareBelow();
|
|
break;
|
|
case CompareBelowEq:
|
|
compileCompareBelowEq();
|
|
break;
|
|
case CompareEqPtr:
|
|
compileCompareEqPtr();
|
|
break;
|
|
case SameValue:
|
|
compileSameValue();
|
|
break;
|
|
case LogicalNot:
|
|
compileLogicalNot();
|
|
break;
|
|
case Call:
|
|
case TailCallInlinedCaller:
|
|
case Construct:
|
|
compileCallOrConstruct();
|
|
break;
|
|
case DirectCall:
|
|
case DirectTailCallInlinedCaller:
|
|
case DirectConstruct:
|
|
case DirectTailCall:
|
|
compileDirectCallOrConstruct();
|
|
break;
|
|
case TailCall:
|
|
compileTailCall();
|
|
break;
|
|
case CallVarargs:
|
|
case CallForwardVarargs:
|
|
case TailCallVarargs:
|
|
case TailCallVarargsInlinedCaller:
|
|
case TailCallForwardVarargs:
|
|
case TailCallForwardVarargsInlinedCaller:
|
|
case ConstructVarargs:
|
|
case ConstructForwardVarargs:
|
|
compileCallOrConstructVarargs();
|
|
break;
|
|
case CallEval:
|
|
compileCallEval();
|
|
break;
|
|
case VarargsLength:
|
|
compileVarargsLength();
|
|
break;
|
|
case LoadVarargs:
|
|
compileLoadVarargs();
|
|
break;
|
|
case ForwardVarargs:
|
|
compileForwardVarargs();
|
|
break;
|
|
case DFG::Jump:
|
|
compileJump();
|
|
break;
|
|
case DFG::Branch:
|
|
compileBranch();
|
|
break;
|
|
case DFG::Switch:
|
|
compileSwitch();
|
|
break;
|
|
case DFG::EntrySwitch:
|
|
compileEntrySwitch();
|
|
break;
|
|
case DFG::Return:
|
|
compileReturn();
|
|
break;
|
|
case ForceOSRExit:
|
|
compileForceOSRExit();
|
|
break;
|
|
case CPUIntrinsic:
|
|
#if CPU(X86_64)
|
|
compileCPUIntrinsic();
|
|
#else
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
#endif
|
|
break;
|
|
case Throw:
|
|
compileThrow();
|
|
break;
|
|
case ThrowStaticError:
|
|
compileThrowStaticError();
|
|
break;
|
|
case InvalidationPoint:
|
|
compileInvalidationPoint();
|
|
break;
|
|
case IsEmpty:
|
|
compileIsEmpty();
|
|
break;
|
|
case TypeOfIsUndefined:
|
|
compileTypeOfIsUndefined();
|
|
break;
|
|
case TypeOfIsObject:
|
|
compileTypeOfIsObject();
|
|
break;
|
|
case TypeOfIsFunction:
|
|
compileIsCallable(operationTypeOfIsFunction);
|
|
break;
|
|
case IsUndefinedOrNull:
|
|
compileIsUndefinedOrNull();
|
|
break;
|
|
case IsBoolean:
|
|
compileIsBoolean();
|
|
break;
|
|
case IsNumber:
|
|
compileIsNumber();
|
|
break;
|
|
case IsBigInt:
|
|
compileIsBigInt();
|
|
break;
|
|
case NumberIsInteger:
|
|
compileNumberIsInteger();
|
|
break;
|
|
case IsCellWithType:
|
|
compileIsCellWithType();
|
|
break;
|
|
case MapHash:
|
|
compileMapHash();
|
|
break;
|
|
case NormalizeMapKey:
|
|
compileNormalizeMapKey();
|
|
break;
|
|
case GetMapBucket:
|
|
compileGetMapBucket();
|
|
break;
|
|
case GetMapBucketHead:
|
|
compileGetMapBucketHead();
|
|
break;
|
|
case GetMapBucketNext:
|
|
compileGetMapBucketNext();
|
|
break;
|
|
case LoadKeyFromMapBucket:
|
|
compileLoadKeyFromMapBucket();
|
|
break;
|
|
case LoadValueFromMapBucket:
|
|
compileLoadValueFromMapBucket();
|
|
break;
|
|
case ExtractValueFromWeakMapGet:
|
|
compileExtractValueFromWeakMapGet();
|
|
break;
|
|
case SetAdd:
|
|
compileSetAdd();
|
|
break;
|
|
case MapSet:
|
|
compileMapSet();
|
|
break;
|
|
case WeakMapGet:
|
|
compileWeakMapGet();
|
|
break;
|
|
case WeakSetAdd:
|
|
compileWeakSetAdd();
|
|
break;
|
|
case WeakMapSet:
|
|
compileWeakMapSet();
|
|
break;
|
|
case IsObject:
|
|
compileIsObject();
|
|
break;
|
|
case IsCallable:
|
|
compileIsCallable(operationObjectIsCallable);
|
|
break;
|
|
case IsConstructor:
|
|
compileIsConstructor();
|
|
break;
|
|
case IsTypedArrayView:
|
|
compileIsTypedArrayView();
|
|
break;
|
|
case ParseInt:
|
|
compileParseInt();
|
|
break;
|
|
case TypeOf:
|
|
compileTypeOf();
|
|
break;
|
|
case CheckTypeInfoFlags:
|
|
compileCheckTypeInfoFlags();
|
|
break;
|
|
case OverridesHasInstance:
|
|
compileOverridesHasInstance();
|
|
break;
|
|
case InstanceOf:
|
|
compileInstanceOf();
|
|
break;
|
|
case InstanceOfCustom:
|
|
compileInstanceOfCustom();
|
|
break;
|
|
case CountExecution:
|
|
compileCountExecution();
|
|
break;
|
|
case SuperSamplerBegin:
|
|
compileSuperSamplerBegin();
|
|
break;
|
|
case SuperSamplerEnd:
|
|
compileSuperSamplerEnd();
|
|
break;
|
|
case StoreBarrier:
|
|
case FencedStoreBarrier:
|
|
compileStoreBarrier();
|
|
break;
|
|
case HasIndexedProperty:
|
|
compileHasIndexedProperty(operationHasIndexedProperty);
|
|
break;
|
|
case HasEnumerableIndexedProperty:
|
|
compileHasIndexedProperty(operationHasEnumerableIndexedProperty);
|
|
break;
|
|
case HasEnumerableStructureProperty:
|
|
compileHasEnumerableStructureProperty();
|
|
break;
|
|
case HasEnumerableProperty:
|
|
compileHasEnumerableProperty();
|
|
break;
|
|
case HasOwnStructureProperty:
|
|
compileHasOwnStructureProperty();
|
|
break;
|
|
case InStructureProperty:
|
|
compileInStructureProperty();
|
|
break;
|
|
case GetDirectPname:
|
|
compileGetDirectPname();
|
|
break;
|
|
case GetEnumerableLength:
|
|
compileGetEnumerableLength();
|
|
break;
|
|
case GetPropertyEnumerator:
|
|
compileGetPropertyEnumerator();
|
|
break;
|
|
case GetEnumeratorStructurePname:
|
|
compileGetEnumeratorStructurePname();
|
|
break;
|
|
case GetEnumeratorGenericPname:
|
|
compileGetEnumeratorGenericPname();
|
|
break;
|
|
case ToIndexString:
|
|
compileToIndexString();
|
|
break;
|
|
case CheckStructureImmediate:
|
|
compileCheckStructureImmediate();
|
|
break;
|
|
case MaterializeNewObject:
|
|
compileMaterializeNewObject();
|
|
break;
|
|
case MaterializeCreateActivation:
|
|
compileMaterializeCreateActivation();
|
|
break;
|
|
case MaterializeNewInternalFieldObject:
|
|
compileMaterializeNewInternalFieldObject();
|
|
break;
|
|
case CheckTraps:
|
|
compileCheckTraps();
|
|
break;
|
|
case CreateRest:
|
|
compileCreateRest();
|
|
break;
|
|
case GetRestLength:
|
|
compileGetRestLength();
|
|
break;
|
|
case RegExpExec:
|
|
compileRegExpExec();
|
|
break;
|
|
case RegExpExecNonGlobalOrSticky:
|
|
compileRegExpExecNonGlobalOrSticky();
|
|
break;
|
|
case RegExpTest:
|
|
compileRegExpTest();
|
|
break;
|
|
case RegExpMatchFast:
|
|
compileRegExpMatchFast();
|
|
break;
|
|
case RegExpMatchFastGlobal:
|
|
compileRegExpMatchFastGlobal();
|
|
break;
|
|
case NewRegexp:
|
|
compileNewRegexp();
|
|
break;
|
|
case SetFunctionName:
|
|
compileSetFunctionName();
|
|
break;
|
|
case StringReplace:
|
|
case StringReplaceRegExp:
|
|
compileStringReplace();
|
|
break;
|
|
case GetRegExpObjectLastIndex:
|
|
compileGetRegExpObjectLastIndex();
|
|
break;
|
|
case SetRegExpObjectLastIndex:
|
|
compileSetRegExpObjectLastIndex();
|
|
break;
|
|
case LogShadowChickenPrologue:
|
|
compileLogShadowChickenPrologue();
|
|
break;
|
|
case LogShadowChickenTail:
|
|
compileLogShadowChickenTail();
|
|
break;
|
|
case RecordRegExpCachedResult:
|
|
compileRecordRegExpCachedResult();
|
|
break;
|
|
case ResolveScopeForHoistingFuncDeclInEval:
|
|
compileResolveScopeForHoistingFuncDeclInEval();
|
|
break;
|
|
case ResolveScope:
|
|
compileResolveScope();
|
|
break;
|
|
case GetDynamicVar:
|
|
compileGetDynamicVar();
|
|
break;
|
|
case PutDynamicVar:
|
|
compilePutDynamicVar();
|
|
break;
|
|
case Unreachable:
|
|
compileUnreachable();
|
|
break;
|
|
case StringSlice:
|
|
compileStringSlice();
|
|
break;
|
|
case ToLowerCase:
|
|
compileToLowerCase();
|
|
break;
|
|
case NumberToStringWithRadix:
|
|
compileNumberToStringWithRadix();
|
|
break;
|
|
case NumberToStringWithValidRadixConstant:
|
|
compileNumberToStringWithValidRadixConstant();
|
|
break;
|
|
case CheckJSCast:
|
|
case CheckNotJSCast:
|
|
compileCheckJSCast();
|
|
break;
|
|
case CallDOM:
|
|
compileCallDOM();
|
|
break;
|
|
case CallDOMGetter:
|
|
compileCallDOMGetter();
|
|
break;
|
|
case FilterCallLinkStatus:
|
|
case FilterGetByStatus:
|
|
case FilterPutByIdStatus:
|
|
case FilterInByIdStatus:
|
|
case FilterDeleteByStatus:
|
|
compileFilterICStatus();
|
|
break;
|
|
case DateGetInt32OrNaN:
|
|
case DateGetTime:
|
|
compileDateGet();
|
|
break;
|
|
case DataViewGetInt:
|
|
case DataViewGetFloat:
|
|
compileDataViewGet();
|
|
break;
|
|
case DataViewSet:
|
|
compileDataViewSet();
|
|
break;
|
|
|
|
case LoopHint: {
|
|
compileLoopHint();
|
|
break;
|
|
}
|
|
|
|
case PhantomLocal:
|
|
case MovHint:
|
|
case ExitOK:
|
|
case PhantomNewObject:
|
|
case PhantomNewFunction:
|
|
case PhantomNewGeneratorFunction:
|
|
case PhantomNewAsyncGeneratorFunction:
|
|
case PhantomNewAsyncFunction:
|
|
case PhantomNewInternalFieldObject:
|
|
case PhantomCreateActivation:
|
|
case PhantomDirectArguments:
|
|
case PhantomCreateRest:
|
|
case PhantomSpread:
|
|
case PhantomNewArrayWithSpread:
|
|
case PhantomNewArrayBuffer:
|
|
case PhantomClonedArguments:
|
|
case PhantomNewRegexp:
|
|
case PutHint:
|
|
case BottomValue:
|
|
case KillStack:
|
|
case InitializeEntrypointArguments:
|
|
break;
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Unrecognized node in FTL backend");
|
|
break;
|
|
}
|
|
|
|
if (Options::validateDFGClobberize() && !m_node->isTerminal()) {
|
|
bool clobberedWorld = false;
|
|
auto validateClobberize = [&] () {
|
|
clobberedWorld = true;
|
|
};
|
|
|
|
clobberize(m_graph, m_node, [] (auto...) { }, [] (auto...) { }, [] (auto...) { }, validateClobberize);
|
|
if (!clobberedWorld) {
|
|
LValue didNotEnterVM = m_out.notZero32(m_out.load8ZeroExt32(m_out.absolute(&vm().didEnterVM)));
|
|
auto* check = m_out.speculate(didNotEnterVM);
|
|
check->setGenerator([=] (CCallHelpers& jit, const StackmapGenerationParams&) {
|
|
jit.breakpoint();
|
|
});
|
|
} else
|
|
m_out.store(m_out.int32Zero, m_out.absolute(&vm().didEnterVM));
|
|
}
|
|
|
|
if (m_node->isTerminal())
|
|
return false;
|
|
|
|
if (!m_state.isValid()) {
|
|
safelyInvalidateAfterTermination();
|
|
return false;
|
|
}
|
|
|
|
m_availabilityCalculator.executeNode(m_node);
|
|
m_interpreter.executeEffects(nodeIndex);
|
|
|
|
return true;
|
|
}
|
|
|
|
void compileUpsilon()
|
|
{
|
|
LValue upsilonValue = nullptr;
|
|
switch (m_node->child1().useKind()) {
|
|
case DoubleRepUse:
|
|
upsilonValue = lowDouble(m_node->child1());
|
|
break;
|
|
case Int32Use:
|
|
case KnownInt32Use:
|
|
upsilonValue = lowInt32(m_node->child1());
|
|
break;
|
|
case Int52RepUse:
|
|
upsilonValue = lowInt52(m_node->child1());
|
|
break;
|
|
case BooleanUse:
|
|
case KnownBooleanUse:
|
|
upsilonValue = lowBoolean(m_node->child1());
|
|
break;
|
|
case CellUse:
|
|
case KnownCellUse:
|
|
upsilonValue = lowCell(m_node->child1());
|
|
break;
|
|
case UntypedUse:
|
|
upsilonValue = lowJSValue(m_node->child1());
|
|
break;
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
break;
|
|
}
|
|
ValueFromBlock upsilon = m_out.anchor(upsilonValue);
|
|
LValue phiNode = m_phis.get(m_node->phi());
|
|
m_out.addIncomingToPhi(phiNode, upsilon);
|
|
}
|
|
|
|
void compilePhi()
|
|
{
|
|
LValue phi = m_phis.get(m_node);
|
|
m_out.m_block->append(phi);
|
|
|
|
switch (m_node->flags() & NodeResultMask) {
|
|
case NodeResultDouble:
|
|
setDouble(phi);
|
|
break;
|
|
case NodeResultInt32:
|
|
setInt32(phi);
|
|
break;
|
|
case NodeResultInt52:
|
|
setInt52(phi);
|
|
break;
|
|
case NodeResultBoolean:
|
|
setBoolean(phi);
|
|
break;
|
|
case NodeResultJS:
|
|
setJSValue(phi);
|
|
break;
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad result type");
|
|
break;
|
|
}
|
|
}
|
|
|
|
void compileDoubleConstant()
|
|
{
|
|
setDouble(m_out.constDouble(m_node->asNumber()));
|
|
}
|
|
|
|
void compileInt52Constant()
|
|
{
|
|
int64_t value = m_node->asAnyInt();
|
|
|
|
setInt52(m_out.constInt64(value << JSValue::int52ShiftAmount));
|
|
setStrictInt52(m_out.constInt64(value));
|
|
}
|
|
|
|
void compileLazyJSConstant()
|
|
{
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int64);
|
|
LazyJSValue value = m_node->lazyJSValue();
|
|
patchpoint->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
value.emit(jit, JSValueRegs(params[0].gpr()));
|
|
});
|
|
patchpoint->effects = Effects::none();
|
|
setJSValue(patchpoint);
|
|
}
|
|
|
|
void compileDoubleRep()
|
|
{
|
|
switch (m_node->child1().useKind()) {
|
|
case RealNumberUse: {
|
|
LValue value = lowJSValue(m_node->child1(), ManualOperandSpeculation);
|
|
|
|
LValue doubleValue = unboxDouble(value);
|
|
|
|
LBasicBlock intCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
ValueFromBlock fastResult = m_out.anchor(doubleValue);
|
|
m_out.branch(
|
|
m_out.doubleEqual(doubleValue, doubleValue),
|
|
usually(continuation), rarely(intCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(intCase, continuation);
|
|
|
|
FTL_TYPE_CHECK(
|
|
jsValueValue(value), m_node->child1(), SpecBytecodeRealNumber,
|
|
isNotInt32(value, provenType(m_node->child1()) & ~SpecDoubleReal));
|
|
ValueFromBlock slowResult = m_out.anchor(m_out.intToDouble(unboxInt32(value)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
|
|
setDouble(m_out.phi(Double, fastResult, slowResult));
|
|
return;
|
|
}
|
|
|
|
case NotCellNorBigIntUse:
|
|
case NumberUse: {
|
|
bool shouldConvertNonNumber = m_node->child1().useKind() == NotCellNorBigIntUse;
|
|
|
|
LValue value = lowJSValue(m_node->child1(), ManualOperandSpeculation);
|
|
|
|
LBasicBlock intCase = m_out.newBlock();
|
|
LBasicBlock doubleTesting = m_out.newBlock();
|
|
LBasicBlock doubleCase = m_out.newBlock();
|
|
LBasicBlock nonDoubleCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
isNotInt32(value, provenType(m_node->child1())),
|
|
unsure(doubleTesting), unsure(intCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(intCase, doubleTesting);
|
|
|
|
ValueFromBlock intToDouble = m_out.anchor(
|
|
m_out.intToDouble(unboxInt32(value)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(doubleTesting, doubleCase);
|
|
LValue valueIsNumber = isNumber(value, provenType(m_node->child1()));
|
|
m_out.branch(valueIsNumber, usually(doubleCase), rarely(nonDoubleCase));
|
|
|
|
m_out.appendTo(doubleCase, nonDoubleCase);
|
|
ValueFromBlock unboxedDouble = m_out.anchor(unboxDouble(value));
|
|
m_out.jump(continuation);
|
|
|
|
if (shouldConvertNonNumber) {
|
|
LBasicBlock undefinedCase = m_out.newBlock();
|
|
LBasicBlock testNullCase = m_out.newBlock();
|
|
LBasicBlock nullCase = m_out.newBlock();
|
|
LBasicBlock testBooleanTrueCase = m_out.newBlock();
|
|
LBasicBlock convertBooleanTrueCase = m_out.newBlock();
|
|
LBasicBlock convertBooleanFalseCase = m_out.newBlock();
|
|
|
|
m_out.appendTo(nonDoubleCase, undefinedCase);
|
|
LValue valueIsUndefined = m_out.equal(value, m_out.constInt64(JSValue::ValueUndefined));
|
|
m_out.branch(valueIsUndefined, unsure(undefinedCase), unsure(testNullCase));
|
|
|
|
m_out.appendTo(undefinedCase, testNullCase);
|
|
ValueFromBlock convertedUndefined = m_out.anchor(m_out.constDouble(PNaN));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(testNullCase, nullCase);
|
|
LValue valueIsNull = m_out.equal(value, m_out.constInt64(JSValue::ValueNull));
|
|
m_out.branch(valueIsNull, unsure(nullCase), unsure(testBooleanTrueCase));
|
|
|
|
m_out.appendTo(nullCase, testBooleanTrueCase);
|
|
ValueFromBlock convertedNull = m_out.anchor(m_out.constDouble(0));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(testBooleanTrueCase, convertBooleanTrueCase);
|
|
LValue valueIsBooleanTrue = m_out.equal(value, m_out.constInt64(JSValue::ValueTrue));
|
|
m_out.branch(valueIsBooleanTrue, unsure(convertBooleanTrueCase), unsure(convertBooleanFalseCase));
|
|
|
|
m_out.appendTo(convertBooleanTrueCase, convertBooleanFalseCase);
|
|
ValueFromBlock convertedTrue = m_out.anchor(m_out.constDouble(1));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(convertBooleanFalseCase, continuation);
|
|
|
|
LValue valueIsNotBooleanFalse = m_out.notEqual(value, m_out.constInt64(JSValue::ValueFalse));
|
|
FTL_TYPE_CHECK(jsValueValue(value), m_node->child1(), ~SpecCellCheck & ~SpecBigInt, valueIsNotBooleanFalse);
|
|
ValueFromBlock convertedFalse = m_out.anchor(m_out.constDouble(0));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setDouble(m_out.phi(Double, intToDouble, unboxedDouble, convertedUndefined, convertedNull, convertedTrue, convertedFalse));
|
|
return;
|
|
}
|
|
m_out.appendTo(nonDoubleCase, continuation);
|
|
FTL_TYPE_CHECK(jsValueValue(value), m_node->child1(), SpecBytecodeNumber, m_out.booleanTrue);
|
|
m_out.unreachable();
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
|
|
setDouble(m_out.phi(Double, intToDouble, unboxedDouble));
|
|
return;
|
|
}
|
|
|
|
case Int52RepUse: {
|
|
setDouble(strictInt52ToDouble(lowStrictInt52(m_node->child1())));
|
|
return;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
}
|
|
}
|
|
|
|
void compileDoubleAsInt32()
|
|
{
|
|
LValue integerValue = convertDoubleToInt32(lowDouble(m_node->child1()), shouldCheckNegativeZero(m_node->arithMode()));
|
|
setInt32(integerValue);
|
|
}
|
|
|
|
void compileValueRep()
|
|
{
|
|
switch (m_node->child1().useKind()) {
|
|
case DoubleRepUse: {
|
|
LValue value = lowDouble(m_node->child1());
|
|
|
|
if (m_interpreter.needsTypeCheck(m_node->child1(), ~SpecDoubleImpureNaN)) {
|
|
value = m_out.select(
|
|
m_out.doubleEqual(value, value), value, m_out.constDouble(PNaN));
|
|
}
|
|
|
|
setJSValue(boxDouble(value));
|
|
return;
|
|
}
|
|
|
|
case Int52RepUse: {
|
|
setJSValue(strictInt52ToJSValue(lowStrictInt52(m_node->child1())));
|
|
return;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
}
|
|
}
|
|
|
|
void compileInt52Rep()
|
|
{
|
|
switch (m_node->child1().useKind()) {
|
|
case Int32Use:
|
|
setStrictInt52(m_out.signExt32To64(lowInt32(m_node->child1())));
|
|
return;
|
|
|
|
case AnyIntUse:
|
|
setStrictInt52(
|
|
jsValueToStrictInt52(
|
|
m_node->child1(), lowJSValue(m_node->child1(), ManualOperandSpeculation)));
|
|
return;
|
|
|
|
case DoubleRepAnyIntUse:
|
|
setStrictInt52(
|
|
doubleToStrictInt52(
|
|
m_node->child1(), lowDouble(m_node->child1())));
|
|
return;
|
|
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
}
|
|
}
|
|
|
|
void compileValueToInt32()
|
|
{
|
|
switch (m_node->child1().useKind()) {
|
|
case Int52RepUse:
|
|
setInt32(m_out.castToInt32(lowStrictInt52(m_node->child1())));
|
|
break;
|
|
|
|
case DoubleRepUse:
|
|
setInt32(doubleToInt32(lowDouble(m_node->child1())));
|
|
break;
|
|
|
|
case NumberUse:
|
|
case NotCellNorBigIntUse: {
|
|
LoweredNodeValue value = m_int32Values.get(m_node->child1().node());
|
|
if (isValid(value)) {
|
|
setInt32(value.value());
|
|
break;
|
|
}
|
|
|
|
value = m_jsValueValues.get(m_node->child1().node());
|
|
if (isValid(value)) {
|
|
setInt32(numberOrNotCellNorBigIntToInt32(m_node->child1(), value.value()));
|
|
break;
|
|
}
|
|
|
|
// We'll basically just get here for constants. But it's good to have this
|
|
// catch-all since we often add new representations into the mix.
|
|
setInt32(
|
|
numberOrNotCellNorBigIntToInt32(
|
|
m_node->child1(),
|
|
lowJSValue(m_node->child1(), ManualOperandSpeculation)));
|
|
break;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
break;
|
|
}
|
|
}
|
|
|
|
void compileBooleanToNumber()
|
|
{
|
|
switch (m_node->child1().useKind()) {
|
|
case BooleanUse: {
|
|
setInt32(m_out.zeroExt(lowBoolean(m_node->child1()), Int32));
|
|
return;
|
|
}
|
|
|
|
case UntypedUse: {
|
|
LValue value = lowJSValue(m_node->child1());
|
|
|
|
if (!m_interpreter.needsTypeCheck(m_node->child1(), SpecBoolInt32 | SpecBoolean)) {
|
|
setInt32(m_out.bitAnd(m_out.castToInt32(value), m_out.int32One));
|
|
return;
|
|
}
|
|
|
|
LBasicBlock booleanCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
ValueFromBlock notBooleanResult = m_out.anchor(value);
|
|
m_out.branch(
|
|
isBoolean(value, provenType(m_node->child1())),
|
|
unsure(booleanCase), unsure(continuation));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(booleanCase, continuation);
|
|
ValueFromBlock booleanResult = m_out.anchor(m_out.bitOr(
|
|
m_out.zeroExt(unboxBoolean(value), Int64), m_numberTag));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, booleanResult, notBooleanResult));
|
|
return;
|
|
}
|
|
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
return;
|
|
}
|
|
}
|
|
|
|
void compileExtractOSREntryLocal()
|
|
{
|
|
EncodedJSValue* buffer = static_cast<EncodedJSValue*>(
|
|
m_ftlState.jitCode->ftlForOSREntry()->entryBuffer()->dataBuffer());
|
|
setJSValue(m_out.load64(m_out.absolute(buffer + m_node->unlinkedOperand().virtualRegister().toLocal())));
|
|
}
|
|
|
|
void compileExtractCatchLocal()
|
|
{
|
|
EncodedJSValue* buffer = static_cast<EncodedJSValue*>(m_ftlState.jitCode->common.catchOSREntryBuffer->dataBuffer());
|
|
setJSValue(m_out.load64(m_out.absolute(buffer + m_node->catchOSREntryIndex())));
|
|
}
|
|
|
|
void compileClearCatchLocals()
|
|
{
|
|
ScratchBuffer* scratchBuffer = m_ftlState.jitCode->common.catchOSREntryBuffer;
|
|
ASSERT(scratchBuffer);
|
|
m_out.storePtr(m_out.constIntPtr(0), m_out.absolute(scratchBuffer->addressOfActiveLength()));
|
|
}
|
|
|
|
void compileGetStack()
|
|
{
|
|
StackAccessData* data = m_node->stackAccessData();
|
|
AbstractValue& value = m_state.operand(data->operand);
|
|
|
|
DFG_ASSERT(m_graph, m_node, isConcrete(data->format), data->format);
|
|
|
|
switch (data->format) {
|
|
case FlushedDouble:
|
|
setDouble(m_out.loadDouble(addressFor(data->machineLocal)));
|
|
break;
|
|
case FlushedInt52:
|
|
setInt52(m_out.load64(addressFor(data->machineLocal)));
|
|
break;
|
|
default:
|
|
if (isInt32Speculation(value.m_type))
|
|
setInt32(m_out.load32(payloadFor(data->machineLocal)));
|
|
else
|
|
setJSValue(m_out.load64(addressFor(data->machineLocal)));
|
|
break;
|
|
}
|
|
}
|
|
|
|
void compilePutStack()
|
|
{
|
|
StackAccessData* data = m_node->stackAccessData();
|
|
switch (data->format) {
|
|
case FlushedJSValue: {
|
|
LValue value = lowJSValue(m_node->child1());
|
|
m_out.store64(value, addressFor(data->machineLocal));
|
|
break;
|
|
}
|
|
|
|
case FlushedDouble: {
|
|
LValue value = lowDouble(m_node->child1());
|
|
m_out.storeDouble(value, addressFor(data->machineLocal));
|
|
break;
|
|
}
|
|
|
|
case FlushedInt32: {
|
|
LValue value = lowInt32(m_node->child1());
|
|
m_out.store32(value, payloadFor(data->machineLocal));
|
|
break;
|
|
}
|
|
|
|
case FlushedInt52: {
|
|
LValue value = lowInt52(m_node->child1());
|
|
m_out.store64(value, addressFor(data->machineLocal));
|
|
break;
|
|
}
|
|
|
|
case FlushedCell: {
|
|
LValue value = lowCell(m_node->child1());
|
|
m_out.store64(value, addressFor(data->machineLocal));
|
|
break;
|
|
}
|
|
|
|
case FlushedBoolean: {
|
|
speculateBoolean(m_node->child1());
|
|
m_out.store64(
|
|
lowJSValue(m_node->child1(), ManualOperandSpeculation),
|
|
addressFor(data->machineLocal));
|
|
break;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad flush format");
|
|
break;
|
|
}
|
|
}
|
|
|
|
void compileNoOp()
|
|
{
|
|
DFG_NODE_DO_TO_CHILDREN(m_graph, m_node, speculate);
|
|
}
|
|
|
|
void compileToObjectOrCallObjectConstructor()
|
|
{
|
|
LValue value = lowJSValue(m_node->child1());
|
|
|
|
LBasicBlock isCellCase = m_out.newBlock();
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(isCell(value, provenType(m_node->child1())), usually(isCellCase), rarely(slowCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isCellCase, slowCase);
|
|
ValueFromBlock fastResult = m_out.anchor(value);
|
|
m_out.branch(isObject(value), usually(continuation), rarely(slowCase));
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
|
|
ValueFromBlock slowResult;
|
|
if (m_node->op() == ToObject) {
|
|
UniquedStringImpl* errorMessage = nullptr;
|
|
if (m_node->identifierNumber() != UINT32_MAX)
|
|
errorMessage = m_graph.identifiers()[m_node->identifierNumber()];
|
|
auto* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
slowResult = m_out.anchor(vmCall(Int64, operationToObject, weakPointer(globalObject), value, m_out.constIntPtr(errorMessage)));
|
|
} else
|
|
slowResult = m_out.anchor(vmCall(Int64, operationCallObjectConstructor, frozenPointer(m_node->cellOperand()), value));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, fastResult, slowResult));
|
|
}
|
|
|
|
void compileToThis()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
|
|
LValue value = lowJSValue(m_node->child1());
|
|
|
|
LBasicBlock isCellCase = m_out.newBlock();
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
isCell(value, provenType(m_node->child1())), usually(isCellCase), rarely(slowCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isCellCase, slowCase);
|
|
ValueFromBlock fastResult = m_out.anchor(value);
|
|
m_out.branch(
|
|
m_out.testIsZero32(
|
|
m_out.load8ZeroExt32(value, m_heaps.JSCell_typeInfoFlags),
|
|
m_out.constInt32(OverridesToThis)),
|
|
usually(continuation), rarely(slowCase));
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
J_JITOperation_GJ function;
|
|
if (m_node->ecmaMode().isStrict())
|
|
function = operationToThisStrict;
|
|
else
|
|
function = operationToThis;
|
|
ValueFromBlock slowResult = m_out.anchor(vmCall(Int64, function, weakPointer(globalObject), value));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, fastResult, slowResult));
|
|
}
|
|
|
|
void compileValueAdd()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
|
|
#if USE(BIGINT32)
|
|
// FIXME: Introduce another BigInt32 code generation: binary use kinds are BigIntUse32, but result is SpecAnyInt and accepting overflow.
|
|
// Let's distinguish these modes based on result type information by introducing NodeResultBigInt32.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=210957
|
|
// https://bugs.webkit.org/show_bug.cgi?id=211040
|
|
if (m_node->isBinaryUseKind(BigInt32Use)) {
|
|
LValue left = lowBigInt32(m_node->child1());
|
|
LValue right = lowBigInt32(m_node->child2());
|
|
|
|
LValue unboxedLeft = unboxBigInt32(left);
|
|
LValue unboxedRight = unboxBigInt32(right);
|
|
|
|
CheckValue* result = m_out.speculateAdd(unboxedLeft, unboxedRight);
|
|
blessSpeculation(result, BigInt32Overflow, noValue(), nullptr, m_origin);
|
|
|
|
LValue boxedResult = boxBigInt32(result);
|
|
setJSValue(boxedResult);
|
|
return;
|
|
}
|
|
#endif
|
|
|
|
if (m_node->isBinaryUseKind(HeapBigIntUse)) {
|
|
LValue left = lowHeapBigInt(m_node->child1());
|
|
LValue right = lowHeapBigInt(m_node->child2());
|
|
|
|
LValue result = vmCall(pointerType(), operationAddHeapBigInt, weakPointer(globalObject), left, right);
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
|
|
CodeBlock* baselineCodeBlock = m_ftlState.graph.baselineCodeBlockFor(m_origin.semantic);
|
|
BytecodeIndex bytecodeIndex = m_origin.semantic.bytecodeIndex();
|
|
BinaryArithProfile* arithProfile = baselineCodeBlock->binaryArithProfileForBytecodeIndex(bytecodeIndex);
|
|
auto repatchingFunction = operationValueAddOptimize;
|
|
auto nonRepatchingFunction = operationValueAdd;
|
|
compileBinaryMathIC<JITAddGenerator>(arithProfile, repatchingFunction, nonRepatchingFunction);
|
|
}
|
|
|
|
void compileValueSub()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
|
|
#if USE(BIGINT32)
|
|
// FIXME: Introduce another BigInt32 code generation: binary use kinds are BigIntUse32, but result is SpecAnyInt and accepting overflow.
|
|
// Let's distinguish these modes based on result type information by introducing NodeResultBigInt32.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=210957
|
|
// https://bugs.webkit.org/show_bug.cgi?id=211040
|
|
if (m_node->isBinaryUseKind(BigInt32Use)) {
|
|
LValue left = lowBigInt32(m_node->child1());
|
|
LValue right = lowBigInt32(m_node->child2());
|
|
|
|
LValue unboxedLeft = unboxBigInt32(left);
|
|
LValue unboxedRight = unboxBigInt32(right);
|
|
|
|
CheckValue* result = m_out.speculateSub(unboxedLeft, unboxedRight);
|
|
blessSpeculation(result, BigInt32Overflow, noValue(), nullptr, m_origin);
|
|
|
|
LValue boxedResult = boxBigInt32(result);
|
|
setJSValue(boxedResult);
|
|
return;
|
|
}
|
|
#endif
|
|
|
|
if (m_node->isBinaryUseKind(HeapBigIntUse)) {
|
|
LValue left = lowHeapBigInt(m_node->child1());
|
|
LValue right = lowHeapBigInt(m_node->child2());
|
|
|
|
LValue result = vmCall(pointerType(), operationSubHeapBigInt, weakPointer(globalObject), left, right);
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
|
|
CodeBlock* baselineCodeBlock = m_ftlState.graph.baselineCodeBlockFor(m_origin.semantic);
|
|
BytecodeIndex bytecodeIndex = m_origin.semantic.bytecodeIndex();
|
|
BinaryArithProfile* arithProfile = baselineCodeBlock->binaryArithProfileForBytecodeIndex(bytecodeIndex);
|
|
auto repatchingFunction = operationValueSubOptimize;
|
|
auto nonRepatchingFunction = operationValueSub;
|
|
compileBinaryMathIC<JITSubGenerator>(arithProfile, repatchingFunction, nonRepatchingFunction);
|
|
}
|
|
|
|
void compileValueMul()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
|
|
#if USE(BIGINT32)
|
|
// FIXME: Introduce another BigInt32 code generation: binary use kinds are BigIntUse32, but result is SpecAnyInt and accepting overflow.
|
|
// Let's distinguish these modes based on result type information by introducing NodeResultBigInt32.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=210957
|
|
// https://bugs.webkit.org/show_bug.cgi?id=211040
|
|
if (m_node->isBinaryUseKind(BigInt32Use)) {
|
|
LValue left = lowBigInt32(m_node->child1());
|
|
LValue right = lowBigInt32(m_node->child2());
|
|
|
|
LValue unboxedLeft = unboxBigInt32(left);
|
|
LValue unboxedRight = unboxBigInt32(right);
|
|
|
|
CheckValue* result = m_out.speculateMul(unboxedLeft, unboxedRight);
|
|
blessSpeculation(result, BigInt32Overflow, noValue(), nullptr, m_origin);
|
|
|
|
LValue boxedResult = boxBigInt32(result);
|
|
setJSValue(boxedResult);
|
|
return;
|
|
}
|
|
#endif
|
|
|
|
if (m_node->isBinaryUseKind(HeapBigIntUse)) {
|
|
LValue left = lowHeapBigInt(m_node->child1());
|
|
LValue right = lowHeapBigInt(m_node->child2());
|
|
|
|
LValue result = vmCall(Int64, operationMulHeapBigInt, weakPointer(globalObject), left, right);
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
|
|
CodeBlock* baselineCodeBlock = m_ftlState.graph.baselineCodeBlockFor(m_origin.semantic);
|
|
BytecodeIndex bytecodeIndex = m_origin.semantic.bytecodeIndex();
|
|
BinaryArithProfile* arithProfile = baselineCodeBlock->binaryArithProfileForBytecodeIndex(bytecodeIndex);
|
|
auto repatchingFunction = operationValueMulOptimize;
|
|
auto nonRepatchingFunction = operationValueMul;
|
|
compileBinaryMathIC<JITMulGenerator>(arithProfile, repatchingFunction, nonRepatchingFunction);
|
|
}
|
|
|
|
template <typename Generator, typename Func1, typename Func2,
|
|
typename = std::enable_if_t<std::is_function<typename std::remove_pointer<Func1>::type>::value && std::is_function<typename std::remove_pointer<Func2>::type>::value>>
|
|
void compileUnaryMathIC(UnaryArithProfile* arithProfile, Func1 repatchingFunction, Func2 nonRepatchingFunction)
|
|
{
|
|
Node* node = m_node;
|
|
|
|
LValue operand = lowJSValue(node->child1());
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int64);
|
|
patchpoint->appendSomeRegister(operand);
|
|
patchpoint->append(m_notCellMask, ValueRep::lateReg(GPRInfo::notCellMaskRegister));
|
|
patchpoint->append(m_numberTag, ValueRep::lateReg(GPRInfo::numberTagRegister));
|
|
RefPtr<PatchpointExceptionHandle> exceptionHandle = preparePatchpointForExceptions(patchpoint);
|
|
patchpoint->numGPScratchRegisters = 1;
|
|
patchpoint->clobber(RegisterSet::macroScratchRegisters());
|
|
State* state = &m_ftlState;
|
|
patchpoint->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
Box<CCallHelpers::JumpList> exceptions =
|
|
exceptionHandle->scheduleExitCreation(params)->jumps(jit);
|
|
|
|
#if ENABLE(MATH_IC_STATS)
|
|
auto inlineStart = jit.label();
|
|
#endif
|
|
|
|
Box<MathICGenerationState> mathICGenerationState = Box<MathICGenerationState>::create();
|
|
JITUnaryMathIC<Generator>* mathIC = jit.codeBlock()->addMathIC<Generator>(arithProfile);
|
|
mathIC->m_generator = Generator(JSValueRegs(params[0].gpr()), JSValueRegs(params[1].gpr()), params.gpScratch(0));
|
|
|
|
bool shouldEmitProfiling = false;
|
|
bool generatedInline = mathIC->generateInline(jit, *mathICGenerationState, shouldEmitProfiling);
|
|
|
|
if (generatedInline) {
|
|
ASSERT(!mathICGenerationState->slowPathJumps.empty());
|
|
auto done = jit.label();
|
|
params.addLatePath([=] (CCallHelpers& jit) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
mathICGenerationState->slowPathJumps.link(&jit);
|
|
mathICGenerationState->slowPathStart = jit.label();
|
|
#if ENABLE(MATH_IC_STATS)
|
|
auto slowPathStart = jit.label();
|
|
#endif
|
|
|
|
if (mathICGenerationState->shouldSlowPathRepatch) {
|
|
SlowPathCall call = callOperation(*state, params.unavailableRegisters(), jit, node->origin.semantic, exceptions.get(),
|
|
repatchingFunction, params[0].gpr(), jit.codeBlock()->globalObjectFor(node->origin.semantic), params[1].gpr(), CCallHelpers::TrustedImmPtr(mathIC));
|
|
mathICGenerationState->slowPathCall = call.call();
|
|
} else {
|
|
SlowPathCall call = callOperation(*state, params.unavailableRegisters(), jit, node->origin.semantic,
|
|
exceptions.get(), nonRepatchingFunction, params[0].gpr(), jit.codeBlock()->globalObjectFor(node->origin.semantic), params[1].gpr());
|
|
mathICGenerationState->slowPathCall = call.call();
|
|
}
|
|
jit.jump().linkTo(done, &jit);
|
|
|
|
jit.addLinkTask([=] (LinkBuffer& linkBuffer) {
|
|
mathIC->finalizeInlineCode(*mathICGenerationState, linkBuffer);
|
|
});
|
|
|
|
#if ENABLE(MATH_IC_STATS)
|
|
auto slowPathEnd = jit.label();
|
|
jit.addLinkTask([=] (LinkBuffer& linkBuffer) {
|
|
size_t size = linkBuffer.locationOf(slowPathEnd).executableAddress<char*>() - linkBuffer.locationOf(slowPathStart).executableAddress<char*>();
|
|
mathIC->m_generatedCodeSize += size;
|
|
});
|
|
#endif
|
|
});
|
|
} else {
|
|
callOperation(
|
|
*state, params.unavailableRegisters(), jit, node->origin.semantic, exceptions.get(),
|
|
nonRepatchingFunction, params[0].gpr(), jit.codeBlock()->globalObjectFor(node->origin.semantic), params[1].gpr());
|
|
}
|
|
|
|
#if ENABLE(MATH_IC_STATS)
|
|
auto inlineEnd = jit.label();
|
|
jit.addLinkTask([=] (LinkBuffer& linkBuffer) {
|
|
size_t size = linkBuffer.locationOf(inlineEnd).executableAddress<char*>() - linkBuffer.locationOf(inlineStart).executableAddress<char*>();
|
|
mathIC->m_generatedCodeSize += size;
|
|
});
|
|
#endif
|
|
});
|
|
|
|
setJSValue(patchpoint);
|
|
}
|
|
|
|
template <typename Generator, typename Func1, typename Func2,
|
|
typename = std::enable_if_t<std::is_function<typename std::remove_pointer<Func1>::type>::value && std::is_function<typename std::remove_pointer<Func2>::type>::value>>
|
|
void compileBinaryMathIC(BinaryArithProfile* arithProfile, Func1 repatchingFunction, Func2 nonRepatchingFunction)
|
|
{
|
|
Node* node = m_node;
|
|
|
|
#if USE(BIGINT32)
|
|
if (node->isBinaryUseKind(AnyBigIntUse)) {
|
|
// FIXME: This is not supported by the IC yet.
|
|
LValue left = lowJSValue(node->child1(), ManualOperandSpeculation);
|
|
LValue right = lowJSValue(node->child2(), ManualOperandSpeculation);
|
|
speculate(node, node->child1());
|
|
speculate(node, node->child2());
|
|
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
setJSValue(vmCall(pointerType(), nonRepatchingFunction, weakPointer(globalObject), left, right));
|
|
return;
|
|
}
|
|
#endif
|
|
|
|
LValue left = lowJSValue(node->child1());
|
|
LValue right = lowJSValue(node->child2());
|
|
|
|
SnippetOperand leftOperand(m_state.forNode(node->child1()).resultType());
|
|
SnippetOperand rightOperand(m_state.forNode(node->child2()).resultType());
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int64);
|
|
patchpoint->appendSomeRegister(left);
|
|
patchpoint->appendSomeRegister(right);
|
|
patchpoint->append(m_notCellMask, ValueRep::lateReg(GPRInfo::notCellMaskRegister));
|
|
patchpoint->append(m_numberTag, ValueRep::lateReg(GPRInfo::numberTagRegister));
|
|
RefPtr<PatchpointExceptionHandle> exceptionHandle =
|
|
preparePatchpointForExceptions(patchpoint);
|
|
patchpoint->numGPScratchRegisters = 1;
|
|
patchpoint->numFPScratchRegisters = 2;
|
|
patchpoint->clobber(RegisterSet::macroScratchRegisters());
|
|
State* state = &m_ftlState;
|
|
patchpoint->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
|
|
Box<CCallHelpers::JumpList> exceptions =
|
|
exceptionHandle->scheduleExitCreation(params)->jumps(jit);
|
|
|
|
#if ENABLE(MATH_IC_STATS)
|
|
auto inlineStart = jit.label();
|
|
#endif
|
|
|
|
Box<MathICGenerationState> mathICGenerationState = Box<MathICGenerationState>::create();
|
|
JITBinaryMathIC<Generator>* mathIC = jit.codeBlock()->addMathIC<Generator>(arithProfile);
|
|
mathIC->m_generator = Generator(leftOperand, rightOperand, JSValueRegs(params[0].gpr()),
|
|
JSValueRegs(params[1].gpr()), JSValueRegs(params[2].gpr()), params.fpScratch(0),
|
|
params.fpScratch(1), params.gpScratch(0), InvalidFPRReg);
|
|
|
|
bool shouldEmitProfiling = false;
|
|
bool generatedInline = mathIC->generateInline(jit, *mathICGenerationState, shouldEmitProfiling);
|
|
|
|
if (generatedInline) {
|
|
ASSERT(!mathICGenerationState->slowPathJumps.empty());
|
|
auto done = jit.label();
|
|
params.addLatePath([=] (CCallHelpers& jit) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
mathICGenerationState->slowPathJumps.link(&jit);
|
|
mathICGenerationState->slowPathStart = jit.label();
|
|
#if ENABLE(MATH_IC_STATS)
|
|
auto slowPathStart = jit.label();
|
|
#endif
|
|
|
|
if (mathICGenerationState->shouldSlowPathRepatch) {
|
|
SlowPathCall call = callOperation(*state, params.unavailableRegisters(), jit, node->origin.semantic, exceptions.get(),
|
|
repatchingFunction, params[0].gpr(), jit.codeBlock()->globalObjectFor(node->origin.semantic), params[1].gpr(), params[2].gpr(), CCallHelpers::TrustedImmPtr(mathIC));
|
|
mathICGenerationState->slowPathCall = call.call();
|
|
} else {
|
|
SlowPathCall call = callOperation(*state, params.unavailableRegisters(), jit, node->origin.semantic,
|
|
exceptions.get(), nonRepatchingFunction, params[0].gpr(), jit.codeBlock()->globalObjectFor(node->origin.semantic), params[1].gpr(), params[2].gpr());
|
|
mathICGenerationState->slowPathCall = call.call();
|
|
}
|
|
jit.jump().linkTo(done, &jit);
|
|
|
|
jit.addLinkTask([=] (LinkBuffer& linkBuffer) {
|
|
mathIC->finalizeInlineCode(*mathICGenerationState, linkBuffer);
|
|
});
|
|
|
|
#if ENABLE(MATH_IC_STATS)
|
|
auto slowPathEnd = jit.label();
|
|
jit.addLinkTask([=] (LinkBuffer& linkBuffer) {
|
|
size_t size = linkBuffer.locationOf(slowPathEnd).executableAddress<char*>() - linkBuffer.locationOf(slowPathStart).executableAddress<char*>();
|
|
mathIC->m_generatedCodeSize += size;
|
|
});
|
|
#endif
|
|
});
|
|
} else {
|
|
callOperation(
|
|
*state, params.unavailableRegisters(), jit, node->origin.semantic, exceptions.get(),
|
|
nonRepatchingFunction, params[0].gpr(), jit.codeBlock()->globalObjectFor(node->origin.semantic), params[1].gpr(), params[2].gpr());
|
|
}
|
|
|
|
#if ENABLE(MATH_IC_STATS)
|
|
auto inlineEnd = jit.label();
|
|
jit.addLinkTask([=] (LinkBuffer& linkBuffer) {
|
|
size_t size = linkBuffer.locationOf(inlineEnd).executableAddress<char*>() - linkBuffer.locationOf(inlineStart).executableAddress<char*>();
|
|
mathIC->m_generatedCodeSize += size;
|
|
});
|
|
#endif
|
|
});
|
|
|
|
setJSValue(patchpoint);
|
|
}
|
|
|
|
void compileStrCat()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
|
|
LValue result;
|
|
if (m_node->child3()) {
|
|
result = vmCall(
|
|
Int64, operationStrCat3, weakPointer(globalObject),
|
|
lowJSValue(m_node->child1(), ManualOperandSpeculation),
|
|
lowJSValue(m_node->child2(), ManualOperandSpeculation),
|
|
lowJSValue(m_node->child3(), ManualOperandSpeculation));
|
|
} else {
|
|
result = vmCall(
|
|
Int64, operationStrCat2, weakPointer(globalObject),
|
|
lowJSValue(m_node->child1(), ManualOperandSpeculation),
|
|
lowJSValue(m_node->child2(), ManualOperandSpeculation));
|
|
}
|
|
setJSValue(result);
|
|
}
|
|
|
|
void compileArithAddOrSub()
|
|
{
|
|
bool isSub = m_node->op() == ArithSub;
|
|
switch (m_node->binaryUseKind()) {
|
|
case Int32Use: {
|
|
LValue left = lowInt32(m_node->child1());
|
|
LValue right = lowInt32(m_node->child2());
|
|
|
|
if (!shouldCheckOverflow(m_node->arithMode())) {
|
|
setInt32(isSub ? m_out.sub(left, right) : m_out.add(left, right));
|
|
break;
|
|
}
|
|
|
|
CheckValue* result =
|
|
isSub ? m_out.speculateSub(left, right) : m_out.speculateAdd(left, right);
|
|
blessSpeculation(result, Overflow, noValue(), nullptr, m_origin);
|
|
setInt32(result);
|
|
break;
|
|
}
|
|
|
|
case Int52RepUse: {
|
|
if (!abstractValue(m_node->child1()).couldBeType(SpecNonInt32AsInt52)
|
|
&& !abstractValue(m_node->child2()).couldBeType(SpecNonInt32AsInt52)) {
|
|
Int52Kind kind;
|
|
LValue left = lowWhicheverInt52(m_node->child1(), kind);
|
|
LValue right = lowInt52(m_node->child2(), kind);
|
|
setInt52(isSub ? m_out.sub(left, right) : m_out.add(left, right), kind);
|
|
break;
|
|
}
|
|
|
|
LValue left = lowInt52(m_node->child1());
|
|
LValue right = lowInt52(m_node->child2());
|
|
CheckValue* result =
|
|
isSub ? m_out.speculateSub(left, right) : m_out.speculateAdd(left, right);
|
|
blessSpeculation(result, Overflow, noValue(), nullptr, m_origin);
|
|
setInt52(result);
|
|
break;
|
|
}
|
|
|
|
case DoubleRepUse: {
|
|
LValue C1 = lowDouble(m_node->child1());
|
|
LValue C2 = lowDouble(m_node->child2());
|
|
|
|
setDouble(isSub ? m_out.doubleSub(C1, C2) : m_out.doubleAdd(C1, C2));
|
|
break;
|
|
}
|
|
|
|
case UntypedUse: {
|
|
if (!isSub) {
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
break;
|
|
}
|
|
|
|
CodeBlock* baselineCodeBlock = m_ftlState.graph.baselineCodeBlockFor(m_origin.semantic);
|
|
BytecodeIndex bytecodeIndex = m_origin.semantic.bytecodeIndex();
|
|
BinaryArithProfile* arithProfile = baselineCodeBlock->binaryArithProfileForBytecodeIndex(bytecodeIndex);
|
|
auto repatchingFunction = operationValueSubOptimize;
|
|
auto nonRepatchingFunction = operationValueSub;
|
|
compileBinaryMathIC<JITSubGenerator>(arithProfile, repatchingFunction, nonRepatchingFunction);
|
|
break;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
break;
|
|
}
|
|
}
|
|
|
|
void compileArithClz32()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
if (m_node->child1().useKind() == Int32Use || m_node->child1().useKind() == KnownInt32Use) {
|
|
LValue operand = lowInt32(m_node->child1());
|
|
setInt32(m_out.ctlz32(operand));
|
|
return;
|
|
}
|
|
DFG_ASSERT(m_graph, m_node, m_node->child1().useKind() == UntypedUse, m_node->child1().useKind());
|
|
LValue argument = lowJSValue(m_node->child1());
|
|
LValue result = m_out.castToInt32(vmCall(Int64, operationArithClz32, weakPointer(globalObject), argument));
|
|
setInt32(result);
|
|
}
|
|
|
|
void compileArithMul()
|
|
{
|
|
switch (m_node->binaryUseKind()) {
|
|
case Int32Use: {
|
|
LValue left = lowInt32(m_node->child1());
|
|
LValue right = lowInt32(m_node->child2());
|
|
|
|
LValue result;
|
|
|
|
if (!shouldCheckOverflow(m_node->arithMode()))
|
|
result = m_out.mul(left, right);
|
|
else {
|
|
CheckValue* speculation = m_out.speculateMul(left, right);
|
|
blessSpeculation(speculation, Overflow, noValue(), nullptr, m_origin);
|
|
result = speculation;
|
|
}
|
|
|
|
if (shouldCheckNegativeZero(m_node->arithMode())) {
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
m_out.notZero32(result), usually(continuation), rarely(slowCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(slowCase, continuation);
|
|
speculate(NegativeZero, noValue(), nullptr, m_out.lessThan(left, m_out.int32Zero));
|
|
speculate(NegativeZero, noValue(), nullptr, m_out.lessThan(right, m_out.int32Zero));
|
|
m_out.jump(continuation);
|
|
m_out.appendTo(continuation, lastNext);
|
|
}
|
|
|
|
setInt32(result);
|
|
break;
|
|
}
|
|
|
|
case Int52RepUse: {
|
|
Int52Kind kind;
|
|
LValue left = lowWhicheverInt52(m_node->child1(), kind);
|
|
LValue right = lowInt52(m_node->child2(), opposite(kind));
|
|
|
|
CheckValue* result = m_out.speculateMul(left, right);
|
|
blessSpeculation(result, Overflow, noValue(), nullptr, m_origin);
|
|
|
|
if (shouldCheckNegativeZero(m_node->arithMode())) {
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
m_out.notZero64(result), usually(continuation), rarely(slowCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(slowCase, continuation);
|
|
speculate(NegativeZero, noValue(), nullptr, m_out.lessThan(left, m_out.int64Zero));
|
|
speculate(NegativeZero, noValue(), nullptr, m_out.lessThan(right, m_out.int64Zero));
|
|
m_out.jump(continuation);
|
|
m_out.appendTo(continuation, lastNext);
|
|
}
|
|
|
|
setInt52(result);
|
|
break;
|
|
}
|
|
|
|
case DoubleRepUse: {
|
|
setDouble(
|
|
m_out.doubleMul(lowDouble(m_node->child1()), lowDouble(m_node->child2())));
|
|
break;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
break;
|
|
}
|
|
}
|
|
|
|
void compileValueDiv()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
// FIXME: add a fast path for BigInt32 here
|
|
// https://bugs.webkit.org/show_bug.cgi?id=211041
|
|
if (m_node->isBinaryUseKind(HeapBigIntUse)) {
|
|
LValue left = lowHeapBigInt(m_node->child1());
|
|
LValue right = lowHeapBigInt(m_node->child2());
|
|
|
|
LValue result = vmCall(pointerType(), operationDivHeapBigInt, weakPointer(globalObject), left, right);
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
|
|
emitBinarySnippet<JITDivGenerator, NeedScratchFPR>(operationValueDiv);
|
|
}
|
|
|
|
void compileArithDiv()
|
|
{
|
|
switch (m_node->binaryUseKind()) {
|
|
case Int32Use: {
|
|
LValue numerator = lowInt32(m_node->child1());
|
|
LValue denominator = lowInt32(m_node->child2());
|
|
|
|
if (shouldCheckNegativeZero(m_node->arithMode())) {
|
|
LBasicBlock zeroNumerator = m_out.newBlock();
|
|
LBasicBlock numeratorContinuation = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
m_out.isZero32(numerator),
|
|
rarely(zeroNumerator), usually(numeratorContinuation));
|
|
|
|
LBasicBlock innerLastNext = m_out.appendTo(zeroNumerator, numeratorContinuation);
|
|
|
|
speculate(
|
|
NegativeZero, noValue(), nullptr, m_out.lessThan(denominator, m_out.int32Zero));
|
|
|
|
m_out.jump(numeratorContinuation);
|
|
|
|
m_out.appendTo(numeratorContinuation, innerLastNext);
|
|
}
|
|
|
|
if (shouldCheckOverflow(m_node->arithMode())) {
|
|
LBasicBlock unsafeDenominator = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue adjustedDenominator = m_out.add(denominator, m_out.int32One);
|
|
m_out.branch(
|
|
m_out.above(adjustedDenominator, m_out.int32One),
|
|
usually(continuation), rarely(unsafeDenominator));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(unsafeDenominator, continuation);
|
|
LValue neg2ToThe31 = m_out.constInt32(-2147483647-1);
|
|
speculate(Overflow, noValue(), nullptr, m_out.isZero32(denominator));
|
|
speculate(Overflow, noValue(), nullptr, m_out.equal(numerator, neg2ToThe31));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
LValue result = m_out.div(numerator, denominator);
|
|
speculate(
|
|
Overflow, noValue(), nullptr,
|
|
m_out.notEqual(m_out.mul(result, denominator), numerator));
|
|
setInt32(result);
|
|
} else
|
|
setInt32(m_out.chillDiv(numerator, denominator));
|
|
|
|
break;
|
|
}
|
|
|
|
case DoubleRepUse: {
|
|
setDouble(m_out.doubleDiv(
|
|
lowDouble(m_node->child1()), lowDouble(m_node->child2())));
|
|
break;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
break;
|
|
}
|
|
}
|
|
|
|
void compileValueMod()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
// FIXME: add a BigInt32 fast path here
|
|
if (m_node->binaryUseKind() == HeapBigIntUse) {
|
|
LValue left = lowHeapBigInt(m_node->child1());
|
|
LValue right = lowHeapBigInt(m_node->child2());
|
|
|
|
LValue result = vmCall(pointerType(), operationModHeapBigInt, weakPointer(globalObject), left, right);
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
|
|
DFG_ASSERT(m_graph, m_node, m_node->binaryUseKind() == UntypedUse || m_node->binaryUseKind() == AnyBigIntUse, m_node->binaryUseKind());
|
|
LValue left = lowJSValue(m_node->child1(), ManualOperandSpeculation);
|
|
LValue right = lowJSValue(m_node->child2(), ManualOperandSpeculation);
|
|
speculate(m_node, m_node->child1());
|
|
speculate(m_node, m_node->child2());
|
|
LValue result = vmCall(Int64, operationValueMod, weakPointer(globalObject), left, right);
|
|
setJSValue(result);
|
|
}
|
|
|
|
void compileArithMod()
|
|
{
|
|
switch (m_node->binaryUseKind()) {
|
|
case Int32Use: {
|
|
LValue numerator = lowInt32(m_node->child1());
|
|
LValue denominator = lowInt32(m_node->child2());
|
|
|
|
LValue remainder;
|
|
if (shouldCheckOverflow(m_node->arithMode())) {
|
|
LBasicBlock unsafeDenominator = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue adjustedDenominator = m_out.add(denominator, m_out.int32One);
|
|
m_out.branch(
|
|
m_out.above(adjustedDenominator, m_out.int32One),
|
|
usually(continuation), rarely(unsafeDenominator));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(unsafeDenominator, continuation);
|
|
LValue neg2ToThe31 = m_out.constInt32(-2147483647-1);
|
|
speculate(Overflow, noValue(), nullptr, m_out.isZero32(denominator));
|
|
speculate(Overflow, noValue(), nullptr, m_out.equal(numerator, neg2ToThe31));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
LValue result = m_out.mod(numerator, denominator);
|
|
remainder = result;
|
|
} else
|
|
remainder = m_out.chillMod(numerator, denominator);
|
|
|
|
if (shouldCheckNegativeZero(m_node->arithMode())) {
|
|
LBasicBlock negativeNumerator = m_out.newBlock();
|
|
LBasicBlock numeratorContinuation = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
m_out.lessThan(numerator, m_out.int32Zero),
|
|
unsure(negativeNumerator), unsure(numeratorContinuation));
|
|
|
|
LBasicBlock innerLastNext = m_out.appendTo(negativeNumerator, numeratorContinuation);
|
|
|
|
speculate(NegativeZero, noValue(), nullptr, m_out.isZero32(remainder));
|
|
|
|
m_out.jump(numeratorContinuation);
|
|
|
|
m_out.appendTo(numeratorContinuation, innerLastNext);
|
|
}
|
|
|
|
setInt32(remainder);
|
|
break;
|
|
}
|
|
|
|
case DoubleRepUse: {
|
|
setDouble(
|
|
m_out.doubleMod(lowDouble(m_node->child1()), lowDouble(m_node->child2())));
|
|
break;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
break;
|
|
}
|
|
}
|
|
|
|
void compileArithMinOrMax()
|
|
{
|
|
switch (m_node->binaryUseKind()) {
|
|
case Int32Use: {
|
|
LValue left = lowInt32(m_node->child1());
|
|
LValue right = lowInt32(m_node->child2());
|
|
|
|
setInt32(
|
|
m_out.select(
|
|
m_node->op() == ArithMin
|
|
? m_out.lessThan(left, right)
|
|
: m_out.lessThan(right, left),
|
|
left, right));
|
|
break;
|
|
}
|
|
|
|
case DoubleRepUse: {
|
|
LValue left = lowDouble(m_node->child1());
|
|
LValue right = lowDouble(m_node->child2());
|
|
|
|
LBasicBlock notLessThan = m_out.newBlock();
|
|
LBasicBlock isEqual = m_out.newBlock();
|
|
LBasicBlock notEqual = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
Vector<ValueFromBlock, 2> results;
|
|
|
|
results.append(m_out.anchor(left));
|
|
m_out.branch(
|
|
m_node->op() == ArithMin
|
|
? m_out.doubleLessThan(left, right)
|
|
: m_out.doubleGreaterThan(left, right),
|
|
unsure(continuation), unsure(notLessThan));
|
|
|
|
// The spec for Math.min and Math.max states that +0 is considered to be larger than -0.
|
|
LBasicBlock lastNext = m_out.appendTo(notLessThan, isEqual);
|
|
m_out.branch(
|
|
m_out.doubleEqual(left, right),
|
|
rarely(isEqual), usually(notEqual));
|
|
|
|
lastNext = m_out.appendTo(isEqual, notEqual);
|
|
results.append(m_out.anchor(
|
|
m_node->op() == ArithMin
|
|
? m_out.bitOr(left, right)
|
|
: m_out.bitAnd(left, right)));
|
|
m_out.jump(continuation);
|
|
|
|
lastNext = m_out.appendTo(notEqual, continuation);
|
|
results.append(
|
|
m_out.anchor(
|
|
m_out.select(
|
|
m_node->op() == ArithMin
|
|
? m_out.doubleGreaterThan(left, right)
|
|
: m_out.doubleLessThan(left, right),
|
|
right, m_out.constDouble(PNaN))));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setDouble(m_out.phi(Double, results));
|
|
break;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
break;
|
|
}
|
|
}
|
|
|
|
void compileArithAbs()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
switch (m_node->child1().useKind()) {
|
|
case Int32Use: {
|
|
LValue value = lowInt32(m_node->child1());
|
|
|
|
LValue mask = m_out.aShr(value, m_out.constInt32(31));
|
|
LValue result = m_out.bitXor(mask, m_out.add(mask, value));
|
|
|
|
if (shouldCheckOverflow(m_node->arithMode()))
|
|
speculate(Overflow, noValue(), nullptr, m_out.lessThan(result, m_out.int32Zero));
|
|
|
|
setInt32(result);
|
|
break;
|
|
}
|
|
|
|
case DoubleRepUse: {
|
|
setDouble(m_out.doubleAbs(lowDouble(m_node->child1())));
|
|
break;
|
|
}
|
|
|
|
default: {
|
|
DFG_ASSERT(m_graph, m_node, m_node->child1().useKind() == UntypedUse, m_node->child1().useKind());
|
|
LValue argument = lowJSValue(m_node->child1());
|
|
LValue result = vmCall(Double, operationArithAbs, weakPointer(globalObject), argument);
|
|
setDouble(result);
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
void compileArithUnary()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
if (m_node->child1().useKind() == DoubleRepUse) {
|
|
setDouble(m_out.doubleUnary(m_node->arithUnaryType(), lowDouble(m_node->child1())));
|
|
return;
|
|
}
|
|
LValue argument = lowJSValue(m_node->child1());
|
|
LValue result = vmCall(Double, DFG::arithUnaryOperation(m_node->arithUnaryType()), weakPointer(globalObject), argument);
|
|
setDouble(result);
|
|
}
|
|
|
|
void compileValuePow()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
// FIXME: maybe add a fast path for BigInt32 here
|
|
if (m_node->isBinaryUseKind(HeapBigIntUse)) {
|
|
LValue base = lowHeapBigInt(m_node->child1());
|
|
LValue exponent = lowHeapBigInt(m_node->child2());
|
|
|
|
LValue result = vmCall(pointerType(), operationPowHeapBigInt, weakPointer(globalObject), base, exponent);
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
|
|
ASSERT(m_node->isBinaryUseKind(UntypedUse) || m_node->isBinaryUseKind(AnyBigIntUse));
|
|
LValue base = lowJSValue(m_node->child1(), ManualOperandSpeculation);
|
|
LValue exponent = lowJSValue(m_node->child2(), ManualOperandSpeculation);
|
|
speculate(m_node, m_node->child1());
|
|
speculate(m_node, m_node->child2());
|
|
LValue result = vmCall(Int64, operationValuePow, weakPointer(globalObject), base, exponent);
|
|
setJSValue(result);
|
|
}
|
|
|
|
void compileArithPow()
|
|
{
|
|
if (m_node->child2().useKind() == Int32Use)
|
|
setDouble(m_out.doublePowi(lowDouble(m_node->child1()), lowInt32(m_node->child2())));
|
|
else {
|
|
LValue base = lowDouble(m_node->child1());
|
|
LValue exponent = lowDouble(m_node->child2());
|
|
|
|
LBasicBlock integerExponentIsSmallBlock = m_out.newBlock();
|
|
LBasicBlock integerExponentPowBlock = m_out.newBlock();
|
|
LBasicBlock doubleExponentPowBlockEntry = m_out.newBlock();
|
|
LBasicBlock nanExceptionBaseIsOne = m_out.newBlock();
|
|
LBasicBlock nanExceptionExponentIsInfinity = m_out.newBlock();
|
|
LBasicBlock testExponentIsOneHalf = m_out.newBlock();
|
|
LBasicBlock handleBaseZeroExponentIsOneHalf = m_out.newBlock();
|
|
LBasicBlock handleInfinityForExponentIsOneHalf = m_out.newBlock();
|
|
LBasicBlock exponentIsOneHalfNormal = m_out.newBlock();
|
|
LBasicBlock exponentIsOneHalfInfinity = m_out.newBlock();
|
|
LBasicBlock testExponentIsNegativeOneHalf = m_out.newBlock();
|
|
LBasicBlock testBaseZeroExponentIsNegativeOneHalf = m_out.newBlock();
|
|
LBasicBlock handleBaseZeroExponentIsNegativeOneHalf = m_out.newBlock();
|
|
LBasicBlock handleInfinityForExponentIsNegativeOneHalf = m_out.newBlock();
|
|
LBasicBlock exponentIsNegativeOneHalfNormal = m_out.newBlock();
|
|
LBasicBlock exponentIsNegativeOneHalfInfinity = m_out.newBlock();
|
|
LBasicBlock powBlock = m_out.newBlock();
|
|
LBasicBlock nanExceptionResultIsNaN = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue integerExponent = m_out.doubleToInt(exponent);
|
|
LValue integerExponentConvertedToDouble = m_out.intToDouble(integerExponent);
|
|
LValue exponentIsInteger = m_out.doubleEqual(exponent, integerExponentConvertedToDouble);
|
|
m_out.branch(exponentIsInteger, unsure(integerExponentIsSmallBlock), unsure(doubleExponentPowBlockEntry));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(integerExponentIsSmallBlock, integerExponentPowBlock);
|
|
LValue integerExponentBelowMax = m_out.belowOrEqual(integerExponent, m_out.constInt32(maxExponentForIntegerMathPow));
|
|
m_out.branch(integerExponentBelowMax, usually(integerExponentPowBlock), rarely(doubleExponentPowBlockEntry));
|
|
|
|
m_out.appendTo(integerExponentPowBlock, doubleExponentPowBlockEntry);
|
|
ValueFromBlock powDoubleIntResult = m_out.anchor(m_out.doublePowi(base, integerExponent));
|
|
m_out.jump(continuation);
|
|
|
|
// If y is NaN, the result is NaN.
|
|
m_out.appendTo(doubleExponentPowBlockEntry, nanExceptionBaseIsOne);
|
|
LValue exponentIsNaN;
|
|
if (provenType(m_node->child2()) & SpecDoubleNaN)
|
|
exponentIsNaN = m_out.doubleNotEqualOrUnordered(exponent, exponent);
|
|
else
|
|
exponentIsNaN = m_out.booleanFalse;
|
|
m_out.branch(exponentIsNaN, rarely(nanExceptionResultIsNaN), usually(nanExceptionBaseIsOne));
|
|
|
|
// If abs(x) is 1 and y is +infinity, the result is NaN.
|
|
// If abs(x) is 1 and y is -infinity, the result is NaN.
|
|
|
|
// Test if base == 1.
|
|
m_out.appendTo(nanExceptionBaseIsOne, nanExceptionExponentIsInfinity);
|
|
LValue absoluteBase = m_out.doubleAbs(base);
|
|
LValue absoluteBaseIsOne = m_out.doubleEqual(absoluteBase, m_out.constDouble(1));
|
|
m_out.branch(absoluteBaseIsOne, rarely(nanExceptionExponentIsInfinity), usually(testExponentIsOneHalf));
|
|
|
|
// Test if abs(y) == Infinity.
|
|
m_out.appendTo(nanExceptionExponentIsInfinity, testExponentIsOneHalf);
|
|
LValue absoluteExponent = m_out.doubleAbs(exponent);
|
|
LValue absoluteExponentIsInfinity = m_out.doubleEqual(absoluteExponent, m_out.constDouble(std::numeric_limits<double>::infinity()));
|
|
m_out.branch(absoluteExponentIsInfinity, rarely(nanExceptionResultIsNaN), usually(testExponentIsOneHalf));
|
|
|
|
// If y == 0.5 or y == -0.5, handle it through SQRT.
|
|
// We have be carefuly with -0 and -Infinity.
|
|
|
|
// Test if y == 0.5
|
|
m_out.appendTo(testExponentIsOneHalf, handleBaseZeroExponentIsOneHalf);
|
|
LValue exponentIsOneHalf = m_out.doubleEqual(exponent, m_out.constDouble(0.5));
|
|
m_out.branch(exponentIsOneHalf, rarely(handleBaseZeroExponentIsOneHalf), usually(testExponentIsNegativeOneHalf));
|
|
|
|
// Handle x == -0.
|
|
m_out.appendTo(handleBaseZeroExponentIsOneHalf, handleInfinityForExponentIsOneHalf);
|
|
LValue baseIsZeroExponentIsOneHalf = m_out.doubleEqual(base, m_out.doubleZero);
|
|
ValueFromBlock zeroResultExponentIsOneHalf = m_out.anchor(m_out.doubleZero);
|
|
m_out.branch(baseIsZeroExponentIsOneHalf, rarely(continuation), usually(handleInfinityForExponentIsOneHalf));
|
|
|
|
// Test if abs(x) == Infinity.
|
|
m_out.appendTo(handleInfinityForExponentIsOneHalf, exponentIsOneHalfNormal);
|
|
LValue absoluteBaseIsInfinityOneHalf = m_out.doubleEqual(absoluteBase, m_out.constDouble(std::numeric_limits<double>::infinity()));
|
|
m_out.branch(absoluteBaseIsInfinityOneHalf, rarely(exponentIsOneHalfInfinity), usually(exponentIsOneHalfNormal));
|
|
|
|
// The exponent is 0.5, the base is finite or NaN, we can use SQRT.
|
|
m_out.appendTo(exponentIsOneHalfNormal, exponentIsOneHalfInfinity);
|
|
ValueFromBlock sqrtResult = m_out.anchor(m_out.doubleSqrt(base));
|
|
m_out.jump(continuation);
|
|
|
|
// The exponent is 0.5, the base is infinite, the result is always infinite.
|
|
m_out.appendTo(exponentIsOneHalfInfinity, testExponentIsNegativeOneHalf);
|
|
ValueFromBlock sqrtInfinityResult = m_out.anchor(m_out.constDouble(std::numeric_limits<double>::infinity()));
|
|
m_out.jump(continuation);
|
|
|
|
// Test if y == -0.5
|
|
m_out.appendTo(testExponentIsNegativeOneHalf, testBaseZeroExponentIsNegativeOneHalf);
|
|
LValue exponentIsNegativeOneHalf = m_out.doubleEqual(exponent, m_out.constDouble(-0.5));
|
|
m_out.branch(exponentIsNegativeOneHalf, rarely(testBaseZeroExponentIsNegativeOneHalf), usually(powBlock));
|
|
|
|
// Handle x == -0.
|
|
m_out.appendTo(testBaseZeroExponentIsNegativeOneHalf, handleBaseZeroExponentIsNegativeOneHalf);
|
|
LValue baseIsZeroExponentIsNegativeOneHalf = m_out.doubleEqual(base, m_out.doubleZero);
|
|
m_out.branch(baseIsZeroExponentIsNegativeOneHalf, rarely(handleBaseZeroExponentIsNegativeOneHalf), usually(handleInfinityForExponentIsNegativeOneHalf));
|
|
|
|
m_out.appendTo(handleBaseZeroExponentIsNegativeOneHalf, handleInfinityForExponentIsNegativeOneHalf);
|
|
ValueFromBlock oneOverSqrtZeroResult = m_out.anchor(m_out.constDouble(std::numeric_limits<double>::infinity()));
|
|
m_out.jump(continuation);
|
|
|
|
// Test if abs(x) == Infinity.
|
|
m_out.appendTo(handleInfinityForExponentIsNegativeOneHalf, exponentIsNegativeOneHalfNormal);
|
|
LValue absoluteBaseIsInfinityNegativeOneHalf = m_out.doubleEqual(absoluteBase, m_out.constDouble(std::numeric_limits<double>::infinity()));
|
|
m_out.branch(absoluteBaseIsInfinityNegativeOneHalf, rarely(exponentIsNegativeOneHalfInfinity), usually(exponentIsNegativeOneHalfNormal));
|
|
|
|
// The exponent is -0.5, the base is finite or NaN, we can use 1/SQRT.
|
|
m_out.appendTo(exponentIsNegativeOneHalfNormal, exponentIsNegativeOneHalfInfinity);
|
|
LValue sqrtBase = m_out.doubleSqrt(base);
|
|
ValueFromBlock oneOverSqrtResult = m_out.anchor(m_out.div(m_out.constDouble(1.), sqrtBase));
|
|
m_out.jump(continuation);
|
|
|
|
// The exponent is -0.5, the base is infinite, the result is always zero.
|
|
m_out.appendTo(exponentIsNegativeOneHalfInfinity, powBlock);
|
|
ValueFromBlock oneOverSqrtInfinityResult = m_out.anchor(m_out.doubleZero);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(powBlock, nanExceptionResultIsNaN);
|
|
ValueFromBlock powResult = m_out.anchor(m_out.doubleStdPow(base, exponent));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(nanExceptionResultIsNaN, continuation);
|
|
ValueFromBlock pureNan = m_out.anchor(m_out.constDouble(PNaN));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setDouble(m_out.phi(Double, powDoubleIntResult, zeroResultExponentIsOneHalf, sqrtResult, sqrtInfinityResult, oneOverSqrtZeroResult, oneOverSqrtResult, oneOverSqrtInfinityResult, powResult, pureNan));
|
|
}
|
|
}
|
|
|
|
void compileArithRandom()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
|
|
// Inlined WeakRandom::advance().
|
|
// uint64_t x = m_low;
|
|
void* lowAddress = reinterpret_cast<uint8_t*>(globalObject) + JSGlobalObject::weakRandomOffset() + WeakRandom::lowOffset();
|
|
LValue low = m_out.load64(m_out.absolute(lowAddress));
|
|
// uint64_t y = m_high;
|
|
void* highAddress = reinterpret_cast<uint8_t*>(globalObject) + JSGlobalObject::weakRandomOffset() + WeakRandom::highOffset();
|
|
LValue high = m_out.load64(m_out.absolute(highAddress));
|
|
// m_low = y;
|
|
m_out.store64(high, m_out.absolute(lowAddress));
|
|
|
|
// x ^= x << 23;
|
|
LValue phase1 = m_out.bitXor(m_out.shl(low, m_out.constInt64(23)), low);
|
|
|
|
// x ^= x >> 17;
|
|
LValue phase2 = m_out.bitXor(m_out.lShr(phase1, m_out.constInt64(17)), phase1);
|
|
|
|
// x ^= y ^ (y >> 26);
|
|
LValue phase3 = m_out.bitXor(m_out.bitXor(high, m_out.lShr(high, m_out.constInt64(26))), phase2);
|
|
|
|
// m_high = x;
|
|
m_out.store64(phase3, m_out.absolute(highAddress));
|
|
|
|
// return x + y;
|
|
LValue random64 = m_out.add(phase3, high);
|
|
|
|
// Extract random 53bit. [0, 53] bit is safe integer number ranges in double representation.
|
|
LValue random53 = m_out.bitAnd(random64, m_out.constInt64((1ULL << 53) - 1));
|
|
|
|
LValue double53Integer = m_out.intToDouble(random53);
|
|
|
|
// Convert `(53bit double integer value) / (1 << 53)` to `(53bit double integer value) * (1.0 / (1 << 53))`.
|
|
// In latter case, `1.0 / (1 << 53)` will become a double value represented as (mantissa = 0 & exp = 970, it means 1e-(2**54)).
|
|
static constexpr double scale = 1.0 / (1ULL << 53);
|
|
|
|
// Multiplying 1e-(2**54) with the double integer does not change anything of the mantissa part of the double integer.
|
|
// It just reduces the exp part of the given 53bit double integer.
|
|
// (Except for 0.0. This is specially handled and in this case, exp just becomes 0.)
|
|
// Now we get 53bit precision random double value in [0, 1).
|
|
LValue result = m_out.doubleMul(double53Integer, m_out.constDouble(scale));
|
|
|
|
setDouble(result);
|
|
}
|
|
|
|
void compileArithRound()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
if (m_node->child1().useKind() == DoubleRepUse) {
|
|
LValue result = nullptr;
|
|
if (producesInteger(m_node->arithRoundingMode()) && !shouldCheckNegativeZero(m_node->arithRoundingMode())) {
|
|
LValue value = lowDouble(m_node->child1());
|
|
result = m_out.doubleFloor(m_out.doubleAdd(value, m_out.constDouble(0.5)));
|
|
} else {
|
|
LBasicBlock shouldRoundDown = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue value = lowDouble(m_node->child1());
|
|
LValue integerValue = m_out.doubleCeil(value);
|
|
ValueFromBlock integerValueResult = m_out.anchor(integerValue);
|
|
|
|
LValue ceilMinusHalf = m_out.doubleSub(integerValue, m_out.constDouble(0.5));
|
|
m_out.branch(m_out.doubleGreaterThanOrUnordered(ceilMinusHalf, value), unsure(shouldRoundDown), unsure(continuation));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(shouldRoundDown, continuation);
|
|
LValue integerValueRoundedDown = m_out.doubleSub(integerValue, m_out.constDouble(1));
|
|
ValueFromBlock integerValueRoundedDownResult = m_out.anchor(integerValueRoundedDown);
|
|
m_out.jump(continuation);
|
|
m_out.appendTo(continuation, lastNext);
|
|
|
|
result = m_out.phi(Double, integerValueResult, integerValueRoundedDownResult);
|
|
}
|
|
|
|
if (producesInteger(m_node->arithRoundingMode())) {
|
|
LValue integerValue = convertDoubleToInt32(result, shouldCheckNegativeZero(m_node->arithRoundingMode()));
|
|
setInt32(integerValue);
|
|
} else
|
|
setDouble(result);
|
|
return;
|
|
}
|
|
|
|
DFG_ASSERT(m_graph, m_node, m_node->child1().useKind() == UntypedUse, m_node->child1().useKind());
|
|
LValue argument = lowJSValue(m_node->child1());
|
|
setJSValue(vmCall(Int64, operationArithRound, weakPointer(globalObject), argument));
|
|
}
|
|
|
|
void compileArithFloor()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
if (m_node->child1().useKind() == DoubleRepUse) {
|
|
LValue value = lowDouble(m_node->child1());
|
|
LValue integerValue = m_out.doubleFloor(value);
|
|
if (producesInteger(m_node->arithRoundingMode()))
|
|
setInt32(convertDoubleToInt32(integerValue, shouldCheckNegativeZero(m_node->arithRoundingMode())));
|
|
else
|
|
setDouble(integerValue);
|
|
return;
|
|
}
|
|
DFG_ASSERT(m_graph, m_node, m_node->child1().useKind() == UntypedUse, m_node->child1().useKind());
|
|
LValue argument = lowJSValue(m_node->child1());
|
|
setJSValue(vmCall(Int64, operationArithFloor, weakPointer(globalObject), argument));
|
|
}
|
|
|
|
void compileArithCeil()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
if (m_node->child1().useKind() == DoubleRepUse) {
|
|
LValue value = lowDouble(m_node->child1());
|
|
LValue integerValue = m_out.doubleCeil(value);
|
|
if (producesInteger(m_node->arithRoundingMode()))
|
|
setInt32(convertDoubleToInt32(integerValue, shouldCheckNegativeZero(m_node->arithRoundingMode())));
|
|
else
|
|
setDouble(integerValue);
|
|
return;
|
|
}
|
|
DFG_ASSERT(m_graph, m_node, m_node->child1().useKind() == UntypedUse, m_node->child1().useKind());
|
|
LValue argument = lowJSValue(m_node->child1());
|
|
setJSValue(vmCall(Int64, operationArithCeil, weakPointer(globalObject), argument));
|
|
}
|
|
|
|
void compileArithTrunc()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
if (m_node->child1().useKind() == DoubleRepUse) {
|
|
LValue value = lowDouble(m_node->child1());
|
|
LValue result = m_out.doubleTrunc(value);
|
|
if (producesInteger(m_node->arithRoundingMode()))
|
|
setInt32(convertDoubleToInt32(result, shouldCheckNegativeZero(m_node->arithRoundingMode())));
|
|
else
|
|
setDouble(result);
|
|
return;
|
|
}
|
|
DFG_ASSERT(m_graph, m_node, m_node->child1().useKind() == UntypedUse, m_node->child1().useKind());
|
|
LValue argument = lowJSValue(m_node->child1());
|
|
setJSValue(vmCall(Int64, operationArithTrunc, weakPointer(globalObject), argument));
|
|
}
|
|
|
|
void compileArithSqrt()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
if (m_node->child1().useKind() == DoubleRepUse) {
|
|
setDouble(m_out.doubleSqrt(lowDouble(m_node->child1())));
|
|
return;
|
|
}
|
|
LValue argument = lowJSValue(m_node->child1());
|
|
LValue result = vmCall(Double, operationArithSqrt, weakPointer(globalObject), argument);
|
|
setDouble(result);
|
|
}
|
|
|
|
void compileArithFRound()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
if (m_node->child1().useKind() == DoubleRepUse) {
|
|
setDouble(m_out.fround(lowDouble(m_node->child1())));
|
|
return;
|
|
}
|
|
LValue argument = lowJSValue(m_node->child1());
|
|
LValue result = vmCall(Double, operationArithFRound, weakPointer(globalObject), argument);
|
|
setDouble(result);
|
|
}
|
|
|
|
void compileIncOrDec()
|
|
{
|
|
DFG_ASSERT(m_graph, m_node, m_node->child1().useKind() == UntypedUse);
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue operand = lowJSValue(m_node->child1());
|
|
LValue result = vmCall(Int64, m_node->op() == Inc ? operationInc : operationDec, weakPointer(globalObject), operand);
|
|
setJSValue(result);
|
|
}
|
|
|
|
void compileValueNegate()
|
|
{
|
|
DFG_ASSERT(m_graph, m_node, m_node->child1().useKind() == UntypedUse);
|
|
CodeBlock* baselineCodeBlock = m_ftlState.graph.baselineCodeBlockFor(m_origin.semantic);
|
|
BytecodeIndex bytecodeIndex = m_origin.semantic.bytecodeIndex();
|
|
UnaryArithProfile* arithProfile = baselineCodeBlock->unaryArithProfileForBytecodeIndex(bytecodeIndex);
|
|
auto repatchingFunction = operationArithNegateOptimize;
|
|
auto nonRepatchingFunction = operationArithNegate;
|
|
compileUnaryMathIC<JITNegGenerator>(arithProfile, repatchingFunction, nonRepatchingFunction);
|
|
}
|
|
|
|
void compileArithNegate()
|
|
{
|
|
switch (m_node->child1().useKind()) {
|
|
case Int32Use: {
|
|
LValue value = lowInt32(m_node->child1());
|
|
|
|
LValue result;
|
|
if (!shouldCheckOverflow(m_node->arithMode()))
|
|
result = m_out.neg(value);
|
|
else if (!shouldCheckNegativeZero(m_node->arithMode())) {
|
|
CheckValue* check = m_out.speculateSub(m_out.int32Zero, value);
|
|
blessSpeculation(check, Overflow, noValue(), nullptr, m_origin);
|
|
result = check;
|
|
} else {
|
|
speculate(Overflow, noValue(), nullptr, m_out.testIsZero32(value, m_out.constInt32(0x7fffffff)));
|
|
result = m_out.neg(value);
|
|
}
|
|
|
|
setInt32(result);
|
|
break;
|
|
}
|
|
|
|
case Int52RepUse: {
|
|
if (!abstractValue(m_node->child1()).couldBeType(SpecNonInt32AsInt52)) {
|
|
Int52Kind kind;
|
|
LValue value = lowWhicheverInt52(m_node->child1(), kind);
|
|
LValue result = m_out.neg(value);
|
|
if (shouldCheckNegativeZero(m_node->arithMode()))
|
|
speculate(NegativeZero, noValue(), nullptr, m_out.isZero64(result));
|
|
setInt52(result, kind);
|
|
break;
|
|
}
|
|
|
|
LValue value = lowInt52(m_node->child1());
|
|
CheckValue* result = m_out.speculateSub(m_out.int64Zero, value);
|
|
blessSpeculation(result, Int52Overflow, noValue(), nullptr, m_origin);
|
|
if (shouldCheckNegativeZero(m_node->arithMode()))
|
|
speculate(NegativeZero, noValue(), nullptr, m_out.isZero64(result));
|
|
setInt52(result);
|
|
break;
|
|
}
|
|
|
|
case DoubleRepUse: {
|
|
setDouble(m_out.doubleNeg(lowDouble(m_node->child1())));
|
|
break;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
break;
|
|
}
|
|
}
|
|
|
|
void compileValueBitNot()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
|
|
#if USE(BIGINT32)
|
|
if (m_node->child1().useKind() == BigInt32Use) {
|
|
LValue operand = lowBigInt32(m_node->child1());
|
|
// The following trick relies on details of the representation of BigInt32, and will have to be updated if we move bits around.
|
|
static_assert(JSValue::BigInt32Tag == 0x12);
|
|
static_assert(JSValue::BigInt32Mask == static_cast<int64_t>(0xfffe000000000012));
|
|
uint64_t maskForBigInt32Bits = 0x0000ffffffff0000;
|
|
LValue result = m_out.bitXor(operand, m_out.constInt64(maskForBigInt32Bits));
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
#endif
|
|
|
|
if (m_node->child1().useKind() == HeapBigIntUse) {
|
|
LValue operand = lowHeapBigInt(m_node->child1());
|
|
LValue result = vmCall(pointerType(), operationBitNotHeapBigInt, weakPointer(globalObject), operand);
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
|
|
DFG_ASSERT(m_graph, m_node, m_node->child1().useKind() == UntypedUse || m_node->child1().useKind() == AnyBigIntUse);
|
|
LValue operand = lowJSValue(m_node->child1(), ManualOperandSpeculation);
|
|
speculate(m_node, m_node->child1());
|
|
LValue result = vmCall(Int64, operationValueBitNot, weakPointer(globalObject), operand);
|
|
setJSValue(result);
|
|
}
|
|
|
|
void compileArithBitNot()
|
|
{
|
|
setInt32(m_out.bitNot(lowInt32(m_node->child1())));
|
|
}
|
|
|
|
void compileValueBitAnd()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
|
|
#if USE(BIGINT32)
|
|
if (m_node->isBinaryUseKind(BigInt32Use)) {
|
|
LValue left = lowBigInt32(m_node->child1());
|
|
LValue right = lowBigInt32(m_node->child2());
|
|
// No need to unbox, since the tagging is not affected by bitAnd
|
|
LValue result = m_out.bitAnd(left, right);
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
#endif
|
|
|
|
if (m_node->isBinaryUseKind(HeapBigIntUse)) {
|
|
LValue left = lowHeapBigInt(m_node->child1());
|
|
LValue right = lowHeapBigInt(m_node->child2());
|
|
|
|
LValue result = vmCall(pointerType(), operationBitAndHeapBigInt, weakPointer(globalObject), left, right);
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
|
|
emitBinaryBitOpSnippet<JITBitAndGenerator>(operationValueBitAnd);
|
|
}
|
|
|
|
void compileArithBitAnd()
|
|
{
|
|
setInt32(m_out.bitAnd(lowInt32(m_node->child1()), lowInt32(m_node->child2())));
|
|
}
|
|
|
|
void compileValueBitOr()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
|
|
#if USE(BIGINT32)
|
|
if (m_node->isBinaryUseKind(BigInt32Use)) {
|
|
LValue left = lowBigInt32(m_node->child1());
|
|
LValue right = lowBigInt32(m_node->child2());
|
|
// No need to unbox, since the tagging is not affected by bitAnd
|
|
LValue result = m_out.bitOr(left, right);
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
#endif
|
|
|
|
if (m_node->isBinaryUseKind(HeapBigIntUse)) {
|
|
LValue left = lowHeapBigInt(m_node->child1());
|
|
LValue right = lowHeapBigInt(m_node->child2());
|
|
|
|
LValue result = vmCall(pointerType(), operationBitOrHeapBigInt, weakPointer(globalObject), left, right);
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
|
|
emitBinaryBitOpSnippet<JITBitOrGenerator>(operationValueBitOr);
|
|
}
|
|
|
|
void compileArithBitOr()
|
|
{
|
|
setInt32(m_out.bitOr(lowInt32(m_node->child1()), lowInt32(m_node->child2())));
|
|
}
|
|
|
|
void compileValueBitXor()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
|
|
#if USE(BIGINT32)
|
|
if (m_node->isBinaryUseKind(BigInt32Use)) {
|
|
LValue left = lowBigInt32(m_node->child1());
|
|
LValue right = lowBigInt32(m_node->child2());
|
|
LValue resultMissingTag = m_out.bitXor(left, right);
|
|
LValue result = m_out.bitOr(resultMissingTag, m_out.constInt64(JSValue::BigInt32Tag));
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
#endif
|
|
|
|
if (m_node->isBinaryUseKind(HeapBigIntUse)) {
|
|
LValue left = lowHeapBigInt(m_node->child1());
|
|
LValue right = lowHeapBigInt(m_node->child2());
|
|
|
|
LValue result = vmCall(pointerType(), operationBitXorHeapBigInt, weakPointer(globalObject), left, right);
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
|
|
emitBinaryBitOpSnippet<JITBitXorGenerator>(operationValueBitXor);
|
|
}
|
|
|
|
void compileArithBitXor()
|
|
{
|
|
setInt32(m_out.bitXor(lowInt32(m_node->child1()), lowInt32(m_node->child2())));
|
|
}
|
|
|
|
void compileValueBitRShift()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
|
|
#if USE(BIGINT32)
|
|
if (m_node->isBinaryUseKind(AnyBigIntUse) || m_node->isBinaryUseKind(BigInt32Use)) {
|
|
// FIXME: do something smarter here
|
|
// Things are a bit tricky because a right-shift by a negative number is a left-shift for BigInts.
|
|
// So even a right shift can overflow.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=210847
|
|
|
|
LValue left = lowJSValue(m_node->child1(), ManualOperandSpeculation);
|
|
LValue right = lowJSValue(m_node->child2(), ManualOperandSpeculation);
|
|
speculate(m_node, m_node->child1());
|
|
speculate(m_node, m_node->child2());
|
|
|
|
LValue result = vmCall(pointerType(), operationValueBitRShift, weakPointer(globalObject), left, right);
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
#endif // USE(BIGINT32)
|
|
|
|
if (m_node->isBinaryUseKind(HeapBigIntUse)) {
|
|
LValue left = lowHeapBigInt(m_node->child1());
|
|
LValue right = lowHeapBigInt(m_node->child2());
|
|
|
|
LValue result = vmCall(pointerType(), operationBitRShiftHeapBigInt, weakPointer(globalObject), left, right);
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
|
|
emitRightShiftSnippet(JITRightShiftGenerator::SignedShift);
|
|
}
|
|
|
|
void compileArithBitRShift()
|
|
{
|
|
setInt32(m_out.aShr(
|
|
lowInt32(m_node->child1()),
|
|
m_out.bitAnd(lowInt32(m_node->child2()), m_out.constInt32(31)))); // FIXME: I don't think that the BitAnd is useful, it is included in the semantics of shift in B3
|
|
}
|
|
|
|
void compileArithBitLShift()
|
|
{
|
|
setInt32(m_out.shl(
|
|
lowInt32(m_node->child1()),
|
|
m_out.bitAnd(lowInt32(m_node->child2()), m_out.constInt32(31)))); // FIXME: I don't think that the BitAnd is useful, it is included in the semantics of shift in B3
|
|
}
|
|
|
|
void compileValueBitLShift()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
// FIXME: consider adding a fast path for BigInt32 here.
|
|
if (m_node->isBinaryUseKind(HeapBigIntUse)) {
|
|
LValue left = lowHeapBigInt(m_node->child1());
|
|
LValue right = lowHeapBigInt(m_node->child2());
|
|
|
|
LValue result = vmCall(pointerType(), operationBitLShiftHeapBigInt, weakPointer(globalObject), left, right);
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
|
|
DFG_ASSERT(m_graph, m_node, m_node->isBinaryUseKind(UntypedUse) || m_node->isBinaryUseKind(AnyBigIntUse) || m_node->isBinaryUseKind(BigInt32Use));
|
|
emitBinaryBitOpSnippet<JITLeftShiftGenerator>(operationValueBitLShift);
|
|
}
|
|
|
|
void compileBitURShift()
|
|
{
|
|
if (m_node->isBinaryUseKind(UntypedUse)) {
|
|
emitRightShiftSnippet(JITRightShiftGenerator::UnsignedShift);
|
|
return;
|
|
}
|
|
setInt32(m_out.lShr(
|
|
lowInt32(m_node->child1()),
|
|
m_out.bitAnd(lowInt32(m_node->child2()), m_out.constInt32(31)))); // FIXME: I don't think that the BitAnd is useful, it is included in the semantics of shift in B3
|
|
}
|
|
|
|
void compileUInt32ToNumber()
|
|
{
|
|
LValue value = lowInt32(m_node->child1());
|
|
|
|
if (doesOverflow(m_node->arithMode())) {
|
|
setStrictInt52(m_out.zeroExtPtr(value));
|
|
return;
|
|
}
|
|
|
|
speculate(Overflow, noValue(), nullptr, m_out.lessThan(value, m_out.int32Zero));
|
|
setInt32(value);
|
|
}
|
|
|
|
void compileCheckStructure()
|
|
{
|
|
ExitKind exitKind;
|
|
if (m_node->child1()->hasConstant())
|
|
exitKind = BadConstantCache;
|
|
else
|
|
exitKind = BadCache;
|
|
|
|
switch (m_node->child1().useKind()) {
|
|
case CellUse:
|
|
case KnownCellUse: {
|
|
LValue cell = lowCell(m_node->child1());
|
|
|
|
checkStructure(
|
|
m_out.load32(cell, m_heaps.JSCell_structureID), jsValueValue(cell),
|
|
exitKind, m_node->structureSet(),
|
|
[&] (RegisteredStructure structure) {
|
|
return weakStructureID(structure);
|
|
});
|
|
return;
|
|
}
|
|
|
|
case CellOrOtherUse: {
|
|
LValue value = lowJSValue(m_node->child1(), ManualOperandSpeculation);
|
|
|
|
LBasicBlock cellCase = m_out.newBlock();
|
|
LBasicBlock notCellCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
isCell(value, provenType(m_node->child1())), unsure(cellCase), unsure(notCellCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(cellCase, notCellCase);
|
|
checkStructure(
|
|
m_out.load32(value, m_heaps.JSCell_structureID), jsValueValue(value),
|
|
exitKind, m_node->structureSet(),
|
|
[&] (RegisteredStructure structure) {
|
|
return weakStructureID(structure);
|
|
});
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(notCellCase, continuation);
|
|
FTL_TYPE_CHECK(jsValueValue(value), m_node->child1(), SpecCell | SpecOther, isNotOther(value));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
return;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
return;
|
|
}
|
|
}
|
|
|
|
void compileCheckStructureOrEmpty()
|
|
{
|
|
ExitKind exitKind;
|
|
if (m_node->child1()->hasConstant())
|
|
exitKind = BadConstantCache;
|
|
else
|
|
exitKind = BadCache;
|
|
|
|
LValue cell = lowCell(m_node->child1());
|
|
bool maySeeEmptyValue = m_interpreter.forNode(m_node->child1()).m_type & SpecEmpty;
|
|
LBasicBlock continuation = nullptr;
|
|
LBasicBlock lastNext = nullptr;
|
|
if (maySeeEmptyValue) {
|
|
LBasicBlock notEmpty = m_out.newBlock();
|
|
continuation = m_out.newBlock();
|
|
m_out.branch(m_out.isZero64(cell), unsure(continuation), unsure(notEmpty));
|
|
lastNext = m_out.appendTo(notEmpty, continuation);
|
|
}
|
|
|
|
checkStructure(
|
|
m_out.load32(cell, m_heaps.JSCell_structureID), jsValueValue(cell),
|
|
exitKind, m_node->structureSet(),
|
|
[&] (RegisteredStructure structure) {
|
|
return weakStructureID(structure);
|
|
});
|
|
|
|
if (maySeeEmptyValue) {
|
|
m_out.jump(continuation);
|
|
m_out.appendTo(continuation, lastNext);
|
|
}
|
|
}
|
|
|
|
void compileCheckIsConstant()
|
|
{
|
|
if (m_node->child1().useKind() == CellUse) {
|
|
LValue cell = lowCell(m_node->child1());
|
|
|
|
speculate(
|
|
BadConstantValue, jsValueValue(cell), m_node->child1().node(),
|
|
m_out.notEqual(cell, weakPointer(m_node->cellOperand()->cell())));
|
|
} else {
|
|
LValue value = lowJSValue(m_node->child1());
|
|
|
|
ASSERT(!m_node->constant()->value().isCell() || !m_node->constant()->value());
|
|
speculate(
|
|
BadConstantValue, jsValueValue(value), m_node->child1().node(),
|
|
m_out.notEqual(value, m_out.constInt64(JSValue::encode(m_node->constant()->value()))));
|
|
}
|
|
}
|
|
|
|
void compileCheckBadValue()
|
|
{
|
|
terminate(BadConstantValue);
|
|
}
|
|
|
|
void compileCheckNotEmpty()
|
|
{
|
|
speculate(TDZFailure, noValue(), nullptr, m_out.isZero64(lowJSValue(m_node->child1())));
|
|
}
|
|
|
|
void compileAssertNotEmpty()
|
|
{
|
|
if (!validationEnabled())
|
|
return;
|
|
|
|
LValue val = lowJSValue(m_node->child1());
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Void);
|
|
patchpoint->appendSomeRegister(val);
|
|
patchpoint->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
GPRReg input = params[0].gpr();
|
|
CCallHelpers::Jump done = jit.branchIfNotEmpty(input);
|
|
jit.breakpoint();
|
|
done.link(&jit);
|
|
});
|
|
}
|
|
|
|
void compileCheckIdent()
|
|
{
|
|
UniquedStringImpl* uid = m_node->uidOperand();
|
|
LValue stringImpl;
|
|
if (m_node->child1().useKind() == StringIdentUse)
|
|
stringImpl = lowStringIdent(m_node->child1());
|
|
else {
|
|
ASSERT(m_node->child1().useKind() == SymbolUse);
|
|
stringImpl = m_out.loadPtr(lowSymbol(m_node->child1()), m_heaps.Symbol_symbolImpl);
|
|
}
|
|
speculate(BadIdent, noValue(), nullptr, m_out.notEqual(stringImpl, m_out.constIntPtr(uid)));
|
|
}
|
|
|
|
void compileGetExecutable()
|
|
{
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
LBasicBlock hasRareData = m_out.newBlock();
|
|
LValue cell = lowCell(m_node->child1());
|
|
speculateFunction(m_node->child1(), cell);
|
|
|
|
LValue rareDataTags = m_out.loadPtr(cell, m_heaps.JSFunction_executableOrRareData);
|
|
ValueFromBlock fastExecutable = m_out.anchor(rareDataTags);
|
|
m_out.branch(m_out.testIsZeroPtr(rareDataTags, m_out.constIntPtr(JSFunction::rareDataTag)), unsure(continuation), unsure(hasRareData));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(hasRareData, continuation);
|
|
LValue rareData = m_out.sub(rareDataTags, m_out.constIntPtr(JSFunction::rareDataTag));
|
|
ValueFromBlock slowExecutable = m_out.anchor(m_out.loadPtr(rareData, m_heaps.FunctionRareData_executable));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(pointerType(), fastExecutable, slowExecutable));
|
|
}
|
|
|
|
void compileArrayify()
|
|
{
|
|
LValue cell = lowCell(m_node->child1());
|
|
LValue property = !!m_node->child2() ? lowInt32(m_node->child2()) : nullptr;
|
|
|
|
LBasicBlock unexpectedStructure = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
auto isUnexpectedArray = [&] (LValue cell) {
|
|
if (m_node->op() == Arrayify)
|
|
return m_out.logicalNot(isArrayTypeForArrayify(cell, m_node->arrayMode()));
|
|
|
|
ASSERT(m_node->op() == ArrayifyToStructure);
|
|
return m_out.notEqual(m_out.load32(cell, m_heaps.JSCell_structureID), weakStructureID(m_node->structure()));
|
|
};
|
|
|
|
m_out.branch(isUnexpectedArray(cell), rarely(unexpectedStructure), usually(continuation));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(unexpectedStructure, continuation);
|
|
|
|
if (property) {
|
|
switch (m_node->arrayMode().type()) {
|
|
case Array::Int32:
|
|
case Array::Double:
|
|
case Array::Contiguous:
|
|
speculate(
|
|
Uncountable, noValue(), nullptr,
|
|
m_out.aboveOrEqual(property, m_out.constInt32(MIN_SPARSE_ARRAY_INDEX)));
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
|
|
switch (m_node->arrayMode().type()) {
|
|
case Array::Int32:
|
|
vmCall(Void, operationEnsureInt32, m_vmValue, cell);
|
|
break;
|
|
case Array::Double:
|
|
vmCall(Void, operationEnsureDouble, m_vmValue, cell);
|
|
break;
|
|
case Array::Contiguous:
|
|
vmCall(Void, operationEnsureContiguous, m_vmValue, cell);
|
|
break;
|
|
case Array::ArrayStorage:
|
|
case Array::SlowPutArrayStorage:
|
|
vmCall(Void, operationEnsureArrayStorage, m_vmValue, cell);
|
|
break;
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad array type");
|
|
break;
|
|
}
|
|
|
|
speculate(BadIndexingType, jsValueValue(cell), nullptr, isUnexpectedArray(cell));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
}
|
|
|
|
void compilePutStructure()
|
|
{
|
|
m_ftlState.jitCode->common.notifyCompilingStructureTransition(m_graph.m_plan, codeBlock(), m_node);
|
|
|
|
RegisteredStructure oldStructure = m_node->transition()->previous;
|
|
RegisteredStructure newStructure = m_node->transition()->next;
|
|
ASSERT_UNUSED(oldStructure, oldStructure->indexingMode() == newStructure->indexingMode());
|
|
ASSERT(oldStructure->typeInfo().inlineTypeFlags() == newStructure->typeInfo().inlineTypeFlags());
|
|
ASSERT(oldStructure->typeInfo().type() == newStructure->typeInfo().type());
|
|
|
|
LValue cell = lowCell(m_node->child1());
|
|
|
|
auto& heap = m_node->transition()->next->transitionKind() == TransitionKind::PropertyDeletion ? m_heaps.JSCellHeaderAndNamedProperties : m_heaps.JSCell_structureID;
|
|
TypedPointer pointer { heap, m_out.addPtr(cell, m_heaps.JSCell_structureID.offset()) };
|
|
|
|
m_out.store32(
|
|
weakStructureID(newStructure), pointer);
|
|
}
|
|
|
|
void compileGetById(AccessType type)
|
|
{
|
|
ASSERT(type == AccessType::GetById || type == AccessType::TryGetById || type == AccessType::GetByIdDirect);
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
switch (m_node->child1().useKind()) {
|
|
case CellUse: {
|
|
setJSValue(getById(lowCell(m_node->child1()), type));
|
|
return;
|
|
}
|
|
|
|
case UntypedUse: {
|
|
// This is pretty weird, since we duplicate the slow path both here and in the
|
|
// code generated by the IC. We should investigate making this less bad.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=127830
|
|
LValue value = lowJSValue(m_node->child1());
|
|
|
|
LBasicBlock cellCase = m_out.newBlock();
|
|
LBasicBlock notCellCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
isCell(value, provenType(m_node->child1())), unsure(cellCase), unsure(notCellCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(cellCase, notCellCase);
|
|
ValueFromBlock cellResult = m_out.anchor(getById(value, type));
|
|
m_out.jump(continuation);
|
|
|
|
auto getByIdFunction = appropriateGenericGetByIdFunction(type);
|
|
|
|
m_out.appendTo(notCellCase, continuation);
|
|
ValueFromBlock notCellResult = m_out.anchor(vmCall(
|
|
Int64, getByIdFunction,
|
|
weakPointer(globalObject), value,
|
|
m_out.constIntPtr(m_node->cacheableIdentifier().rawBits())));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, cellResult, notCellResult));
|
|
return;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
return;
|
|
}
|
|
}
|
|
|
|
void compileGetByIdWithThis()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
if (m_node->child1().useKind() == CellUse && m_node->child2().useKind() == CellUse)
|
|
setJSValue(getByIdWithThis(lowCell(m_node->child1()), lowCell(m_node->child2())));
|
|
else {
|
|
LValue base = lowJSValue(m_node->child1());
|
|
LValue thisValue = lowJSValue(m_node->child2());
|
|
|
|
LBasicBlock baseCellCase = m_out.newBlock();
|
|
LBasicBlock notCellCase = m_out.newBlock();
|
|
LBasicBlock thisValueCellCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
isCell(base, provenType(m_node->child1())), unsure(baseCellCase), unsure(notCellCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(baseCellCase, thisValueCellCase);
|
|
|
|
m_out.branch(
|
|
isCell(thisValue, provenType(m_node->child2())), unsure(thisValueCellCase), unsure(notCellCase));
|
|
|
|
m_out.appendTo(thisValueCellCase, notCellCase);
|
|
ValueFromBlock cellResult = m_out.anchor(getByIdWithThis(base, thisValue));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(notCellCase, continuation);
|
|
ValueFromBlock notCellResult = m_out.anchor(vmCall(
|
|
Int64, operationGetByIdWithThisGeneric,
|
|
weakPointer(globalObject), base, thisValue,
|
|
m_out.constIntPtr(m_node->cacheableIdentifier().rawBits())));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, cellResult, notCellResult));
|
|
}
|
|
|
|
}
|
|
|
|
void compileGetByValWithThis()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue base = lowJSValue(m_node->child1());
|
|
LValue thisValue = lowJSValue(m_node->child2());
|
|
LValue subscript = lowJSValue(m_node->child3());
|
|
|
|
LValue result = vmCall(Int64, operationGetByValWithThis, weakPointer(globalObject), base, thisValue, subscript);
|
|
setJSValue(result);
|
|
}
|
|
|
|
LValue getPrivateName(LValue base, LValue property)
|
|
{
|
|
Node* node = m_node;
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int64);
|
|
patchpoint->appendSomeRegister(base);
|
|
patchpoint->appendSomeRegister(property);
|
|
patchpoint->append(m_notCellMask, ValueRep::lateReg(GPRInfo::notCellMaskRegister));
|
|
patchpoint->append(m_numberTag, ValueRep::lateReg(GPRInfo::numberTagRegister));
|
|
patchpoint->clobber(RegisterSet::macroScratchRegisters());
|
|
|
|
RefPtr<PatchpointExceptionHandle> exceptionHandle = preparePatchpointForExceptions(patchpoint);
|
|
|
|
State* state = &m_ftlState;
|
|
bool baseIsCell = abstractValue(node->child1()).isType(SpecCell);
|
|
patchpoint->setGenerator([=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
CallSiteIndex callSiteIndex = state->jitCode->common.codeOrigins->addUniqueCallSiteIndex(node->origin.semantic);
|
|
|
|
// This is the direct exit target for operation calls.
|
|
Box<CCallHelpers::JumpList> exceptions = exceptionHandle->scheduleExitCreation(params)->jumps(jit);
|
|
|
|
// This is the exit for call IC's created by the IC for getters. We don't have
|
|
// to do anything weird other than call this, since it will associate the exit with
|
|
// the callsite index.
|
|
exceptionHandle->scheduleExitCreationForUnwind(params, callSiteIndex);
|
|
|
|
GPRReg resultGPR = params[0].gpr();
|
|
GPRReg baseGPR = params[1].gpr();
|
|
GPRReg propertyGPR = params[2].gpr();
|
|
|
|
auto generator = Box<JITGetByValGenerator>::create(
|
|
jit.codeBlock(), node->origin.semantic, callSiteIndex, AccessType::GetPrivateName,
|
|
params.unavailableRegisters(), JSValueRegs(baseGPR), JSValueRegs(propertyGPR), JSValueRegs(resultGPR));
|
|
|
|
CCallHelpers::Jump notCell;
|
|
if (!baseIsCell)
|
|
notCell = jit.branchIfNotCell(baseGPR);
|
|
|
|
generator->generateFastPath(jit);
|
|
CCallHelpers::Label done = jit.label();
|
|
|
|
params.addLatePath([=] (CCallHelpers& jit) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
if (notCell.isSet())
|
|
notCell.link(&jit);
|
|
generator->slowPathJump().link(&jit);
|
|
CCallHelpers::Label slowPathBegin = jit.label();
|
|
CCallHelpers::Call slowPathCall = callOperation(
|
|
*state, params.unavailableRegisters(), jit, node->origin.semantic,
|
|
exceptions.get(), operationGetPrivateNameOptimize, resultGPR,
|
|
jit.codeBlock()->globalObjectFor(node->origin.semantic),
|
|
CCallHelpers::TrustedImmPtr(generator->stubInfo()), baseGPR, propertyGPR).call();
|
|
jit.jump().linkTo(done, &jit);
|
|
|
|
generator->reportSlowPathCall(slowPathBegin, slowPathCall);
|
|
|
|
jit.addLinkTask([=] (LinkBuffer& linkBuffer) {
|
|
generator->finalize(linkBuffer, linkBuffer);
|
|
});
|
|
});
|
|
});
|
|
|
|
return patchpoint;
|
|
}
|
|
|
|
void compileGetPrivateName()
|
|
{
|
|
if (m_node->child1().useKind() == CellUse)
|
|
setJSValue(getPrivateName(lowCell(m_node->child1()), lowSymbol(m_node->child2())));
|
|
else {
|
|
LValue base = lowJSValue(m_node->child1());
|
|
LValue property = lowSymbol(m_node->child2());
|
|
|
|
LBasicBlock baseCellCase = m_out.newBlock();
|
|
LBasicBlock notCellCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
isCell(base, provenType(m_node->child1())), unsure(baseCellCase), unsure(notCellCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(baseCellCase, notCellCase);
|
|
|
|
ValueFromBlock cellResult = m_out.anchor(getPrivateName(base, property));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(notCellCase, continuation);
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
ValueFromBlock notCellResult = m_out.anchor(vmCall(
|
|
Int64, operationGetPrivateName,
|
|
weakPointer(globalObject), m_out.constIntPtr(0), base,
|
|
property));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, cellResult, notCellResult));
|
|
}
|
|
}
|
|
|
|
void compileGetPrivateNameById()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_node->origin.semantic);
|
|
if (m_node->child1().useKind() == CellUse)
|
|
setJSValue(getById(lowCell(m_node->child1()), AccessType::GetPrivateName));
|
|
else {
|
|
LValue base = lowJSValue(m_node->child1());
|
|
|
|
LBasicBlock baseCellCase = m_out.newBlock();
|
|
LBasicBlock notCellCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
isCell(base, provenType(m_node->child1())), unsure(baseCellCase), unsure(notCellCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(baseCellCase, notCellCase);
|
|
|
|
ValueFromBlock cellResult = m_out.anchor(getById(base, AccessType::GetPrivateName));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(notCellCase, continuation);
|
|
ValueFromBlock notCellResult = m_out.anchor(vmCall(
|
|
Int64, operationGetPrivateNameByIdGeneric,
|
|
weakPointer(globalObject), base,
|
|
m_out.constIntPtr(m_node->cacheableIdentifier().rawBits())));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, cellResult, notCellResult));
|
|
}
|
|
}
|
|
|
|
void compilePutByIdWithThis()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue base = lowJSValue(m_node->child1());
|
|
LValue thisValue = lowJSValue(m_node->child2());
|
|
LValue value = lowJSValue(m_node->child3());
|
|
|
|
vmCall(Void, m_node->ecmaMode().isStrict() ? operationPutByIdWithThisStrict : operationPutByIdWithThis,
|
|
weakPointer(globalObject), base, thisValue, value, m_out.constIntPtr(m_node->cacheableIdentifier().rawBits()));
|
|
}
|
|
|
|
void compilePutByValWithThis()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue base = lowJSValue(m_graph.varArgChild(m_node, 0));
|
|
LValue thisValue = lowJSValue(m_graph.varArgChild(m_node, 1));
|
|
LValue property = lowJSValue(m_graph.varArgChild(m_node, 2));
|
|
LValue value = lowJSValue(m_graph.varArgChild(m_node, 3));
|
|
|
|
vmCall(Void, m_node->ecmaMode().isStrict() ? operationPutByValWithThisStrict : operationPutByValWithThis,
|
|
weakPointer(globalObject), base, thisValue, property, value);
|
|
}
|
|
|
|
void compilePutPrivateNameById()
|
|
{
|
|
DFG_ASSERT(m_graph, m_node, m_node->child1().useKind() == CellUse, m_node->child1().useKind());
|
|
|
|
LValue base = lowCell(m_node->child1());
|
|
LValue value = lowJSValue(m_node->child2());
|
|
|
|
// We emit property check during DFG generation, so we don't need
|
|
// to check it here.
|
|
auto putKind = m_node->privateFieldPutKind().isDefine() ? PutKind::DirectPrivateFieldDefine : PutKind::DirectPrivateFieldSet;
|
|
cachedPutById(m_node, base, value, ECMAMode::strict(), putKind);
|
|
}
|
|
|
|
void compilePutPrivateName()
|
|
{
|
|
DFG_ASSERT(m_graph, m_node, m_node->child1().useKind() == UntypedUse, m_node->child1().useKind());
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_node->origin.semantic);
|
|
|
|
LValue base = lowJSValue(m_node->child1());
|
|
LValue property = lowSymbol(m_node->child2());
|
|
LValue value = lowJSValue(m_node->child3());
|
|
|
|
vmCall(Void, operationPutPrivateNameGeneric,
|
|
weakPointer(globalObject), base, property, value, m_out.constIntPtr(0), m_out.constInt32(m_node->privateFieldPutKind().value()));
|
|
}
|
|
|
|
void compileAtomicsReadModifyWrite()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
TypedArrayType type = m_node->arrayMode().typedArrayType();
|
|
unsigned numExtraArgs = numExtraAtomicsArgs(m_node->op());
|
|
Edge baseEdge = m_graph.child(m_node, 0);
|
|
Edge indexEdge = m_graph.child(m_node, 1);
|
|
Edge argEdges[maxNumExtraAtomicsArgs];
|
|
for (unsigned i = numExtraArgs; i--;)
|
|
argEdges[i] = m_graph.child(m_node, 2 + i);
|
|
Edge storageEdge = m_graph.child(m_node, 2 + numExtraArgs);
|
|
|
|
if (!storageEdge) {
|
|
auto callWith0 = [&] (auto* operation) {
|
|
ASSERT(numExtraArgs == 0);
|
|
return vmCall(Int64, operation, weakPointer(globalObject), lowJSValue(baseEdge), lowJSValue(indexEdge));
|
|
};
|
|
|
|
auto callWith1 = [&] (auto* operation) {
|
|
ASSERT(numExtraArgs == 1);
|
|
return vmCall(Int64, operation, weakPointer(globalObject), lowJSValue(baseEdge), lowJSValue(indexEdge), lowJSValue(argEdges[0]));
|
|
};
|
|
|
|
auto callWith2 = [&] (auto* operation) {
|
|
ASSERT(numExtraArgs == 2);
|
|
return vmCall(Int64, operation, weakPointer(globalObject), lowJSValue(baseEdge), lowJSValue(indexEdge), lowJSValue(argEdges[0]), lowJSValue(argEdges[1]));
|
|
};
|
|
|
|
LValue result;
|
|
switch (m_node->op()) {
|
|
case AtomicsAdd:
|
|
result = callWith1(operationAtomicsAdd);
|
|
break;
|
|
case AtomicsAnd:
|
|
result = callWith1(operationAtomicsAnd);
|
|
break;
|
|
case AtomicsCompareExchange:
|
|
result = callWith2(operationAtomicsCompareExchange);
|
|
break;
|
|
case AtomicsExchange:
|
|
result = callWith1(operationAtomicsExchange);
|
|
break;
|
|
case AtomicsLoad:
|
|
result = callWith0(operationAtomicsLoad);
|
|
break;
|
|
case AtomicsOr:
|
|
result = callWith1(operationAtomicsOr);
|
|
break;
|
|
case AtomicsStore:
|
|
result = callWith1(operationAtomicsStore);
|
|
break;
|
|
case AtomicsSub:
|
|
result = callWith1(operationAtomicsSub);
|
|
break;
|
|
case AtomicsXor:
|
|
result = callWith1(operationAtomicsXor);
|
|
break;
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
}
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
|
|
LValue index = lowInt32(indexEdge);
|
|
LValue args[2];
|
|
for (unsigned i = numExtraArgs; i--;)
|
|
args[i] = getIntTypedArrayStoreOperand(argEdges[i]);
|
|
LValue storage = lowStorage(storageEdge);
|
|
|
|
TypedPointer pointer = pointerIntoTypedArray(storage, index, type);
|
|
Width width = widthForBytes(elementSize(type));
|
|
|
|
LValue atomicValue;
|
|
LValue result;
|
|
|
|
auto sanitizeResult = [&] (LValue value) -> LValue {
|
|
if (isSigned(type)) {
|
|
switch (elementSize(type)) {
|
|
case 1:
|
|
value = m_out.bitAnd(value, m_out.constInt32(0xff));
|
|
break;
|
|
case 2:
|
|
value = m_out.bitAnd(value, m_out.constInt32(0xffff));
|
|
break;
|
|
case 4:
|
|
break;
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
break;
|
|
}
|
|
}
|
|
return value;
|
|
};
|
|
|
|
switch (m_node->op()) {
|
|
case AtomicsAdd:
|
|
atomicValue = m_out.atomicXchgAdd(args[0], pointer, width);
|
|
result = sanitizeResult(atomicValue);
|
|
break;
|
|
case AtomicsAnd:
|
|
atomicValue = m_out.atomicXchgAnd(args[0], pointer, width);
|
|
result = sanitizeResult(atomicValue);
|
|
break;
|
|
case AtomicsCompareExchange:
|
|
atomicValue = m_out.atomicStrongCAS(args[0], args[1], pointer, width);
|
|
result = sanitizeResult(atomicValue);
|
|
break;
|
|
case AtomicsExchange:
|
|
atomicValue = m_out.atomicXchg(args[0], pointer, width);
|
|
result = sanitizeResult(atomicValue);
|
|
break;
|
|
case AtomicsLoad:
|
|
atomicValue = m_out.atomicXchgAdd(m_out.int32Zero, pointer, width);
|
|
result = sanitizeResult(atomicValue);
|
|
break;
|
|
case AtomicsOr:
|
|
atomicValue = m_out.atomicXchgOr(args[0], pointer, width);
|
|
result = sanitizeResult(atomicValue);
|
|
break;
|
|
case AtomicsStore:
|
|
atomicValue = m_out.atomicXchg(args[0], pointer, width);
|
|
result = args[0];
|
|
break;
|
|
case AtomicsSub:
|
|
atomicValue = m_out.atomicXchgSub(args[0], pointer, width);
|
|
result = sanitizeResult(atomicValue);
|
|
break;
|
|
case AtomicsXor:
|
|
atomicValue = m_out.atomicXchgXor(args[0], pointer, width);
|
|
result = sanitizeResult(atomicValue);
|
|
break;
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
break;
|
|
}
|
|
// Signify that the state against which the atomic operations are serialized is confined to just
|
|
// the typed array storage, since that's as precise of an abstraction as we can have of shared
|
|
// array buffer storage.
|
|
m_heaps.decorateFencedAccess(&m_heaps.typedArrayProperties, atomicValue);
|
|
|
|
// We have to keep base alive since that keeps storage alive.
|
|
ensureStillAliveHere(lowCell(baseEdge));
|
|
setIntTypedArrayLoadResult(result, type);
|
|
}
|
|
|
|
void compileAtomicsIsLockFree()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
Edge child1 = m_graph.child(m_node, 0);
|
|
if (child1.useKind() != Int32Use) {
|
|
setJSValue(vmCall(Int64, operationAtomicsIsLockFree, weakPointer(globalObject), lowJSValue(child1)));
|
|
return;
|
|
}
|
|
|
|
LValue bytes = lowInt32(child1);
|
|
|
|
LBasicBlock trueCase = m_out.newBlock();
|
|
LBasicBlock falseCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(trueCase);
|
|
|
|
Vector<SwitchCase> cases;
|
|
cases.append(SwitchCase(m_out.constInt32(1), trueCase, Weight()));
|
|
cases.append(SwitchCase(m_out.constInt32(2), trueCase, Weight()));
|
|
cases.append(SwitchCase(m_out.constInt32(4), trueCase, Weight()));
|
|
m_out.switchInstruction(bytes, cases, falseCase, Weight());
|
|
|
|
m_out.appendTo(trueCase, falseCase);
|
|
ValueFromBlock trueValue = m_out.anchor(m_out.booleanTrue);
|
|
m_out.jump(continuation);
|
|
m_out.appendTo(falseCase, continuation);
|
|
ValueFromBlock falseValue = m_out.anchor(m_out.booleanFalse);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setBoolean(m_out.phi(Int32, trueValue, falseValue));
|
|
}
|
|
|
|
void compileDefineDataProperty()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue base = lowCell(m_graph.varArgChild(m_node, 0));
|
|
LValue value = lowJSValue(m_graph.varArgChild(m_node, 2));
|
|
LValue attributes = lowInt32(m_graph.varArgChild(m_node, 3));
|
|
Edge& propertyEdge = m_graph.varArgChild(m_node, 1);
|
|
switch (propertyEdge.useKind()) {
|
|
case StringUse: {
|
|
LValue property = lowString(propertyEdge);
|
|
vmCall(Void, operationDefineDataPropertyString, weakPointer(globalObject), base, property, value, attributes);
|
|
break;
|
|
}
|
|
case StringIdentUse: {
|
|
LValue property = lowStringIdent(propertyEdge);
|
|
vmCall(Void, operationDefineDataPropertyStringIdent, weakPointer(globalObject), base, property, value, attributes);
|
|
break;
|
|
}
|
|
case SymbolUse: {
|
|
LValue property = lowSymbol(propertyEdge);
|
|
vmCall(Void, operationDefineDataPropertySymbol, weakPointer(globalObject), base, property, value, attributes);
|
|
break;
|
|
}
|
|
case UntypedUse: {
|
|
LValue property = lowJSValue(propertyEdge);
|
|
vmCall(Void, operationDefineDataProperty, weakPointer(globalObject), base, property, value, attributes);
|
|
break;
|
|
}
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
}
|
|
}
|
|
|
|
void compileDefineAccessorProperty()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue base = lowCell(m_graph.varArgChild(m_node, 0));
|
|
LValue getter = lowCell(m_graph.varArgChild(m_node, 2));
|
|
LValue setter = lowCell(m_graph.varArgChild(m_node, 3));
|
|
LValue attributes = lowInt32(m_graph.varArgChild(m_node, 4));
|
|
Edge& propertyEdge = m_graph.varArgChild(m_node, 1);
|
|
switch (propertyEdge.useKind()) {
|
|
case StringUse: {
|
|
LValue property = lowString(propertyEdge);
|
|
vmCall(Void, operationDefineAccessorPropertyString, weakPointer(globalObject), base, property, getter, setter, attributes);
|
|
break;
|
|
}
|
|
case StringIdentUse: {
|
|
LValue property = lowStringIdent(propertyEdge);
|
|
vmCall(Void, operationDefineAccessorPropertyStringIdent, weakPointer(globalObject), base, property, getter, setter, attributes);
|
|
break;
|
|
}
|
|
case SymbolUse: {
|
|
LValue property = lowSymbol(propertyEdge);
|
|
vmCall(Void, operationDefineAccessorPropertySymbol, weakPointer(globalObject), base, property, getter, setter, attributes);
|
|
break;
|
|
}
|
|
case UntypedUse: {
|
|
LValue property = lowJSValue(propertyEdge);
|
|
vmCall(Void, operationDefineAccessorProperty, weakPointer(globalObject), base, property, getter, setter, attributes);
|
|
break;
|
|
}
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
}
|
|
}
|
|
|
|
void cachedPutById(Node* node, LValue base, LValue value, ECMAMode ecmaMode, PutKind putKind)
|
|
{
|
|
CacheableIdentifier identifier = node->cacheableIdentifier();
|
|
ASSERT(identifier);
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Void);
|
|
patchpoint->appendSomeRegister(base);
|
|
patchpoint->appendSomeRegister(value);
|
|
patchpoint->append(m_notCellMask, ValueRep::reg(GPRInfo::notCellMaskRegister));
|
|
patchpoint->append(m_numberTag, ValueRep::reg(GPRInfo::numberTagRegister));
|
|
patchpoint->clobber(RegisterSet::macroScratchRegisters());
|
|
|
|
// FIXME: If this is a PutByIdFlush, we might want to late-clobber volatile registers.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=152848
|
|
|
|
RefPtr<PatchpointExceptionHandle> exceptionHandle =
|
|
preparePatchpointForExceptions(patchpoint);
|
|
|
|
State* state = &m_ftlState;
|
|
|
|
patchpoint->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
CallSiteIndex callSiteIndex =
|
|
state->jitCode->common.codeOrigins->addUniqueCallSiteIndex(node->origin.semantic);
|
|
|
|
Box<CCallHelpers::JumpList> exceptions =
|
|
exceptionHandle->scheduleExitCreation(params)->jumps(jit);
|
|
|
|
// JS setter call ICs generated by the PutById IC will need this.
|
|
exceptionHandle->scheduleExitCreationForUnwind(params, callSiteIndex);
|
|
|
|
auto generator = Box<JITPutByIdGenerator>::create(
|
|
jit.codeBlock(), node->origin.semantic, callSiteIndex,
|
|
params.unavailableRegisters(), identifier, JSValueRegs(params[0].gpr()),
|
|
JSValueRegs(params[1].gpr()), GPRInfo::patchpointScratchRegister, ecmaMode,
|
|
putKind);
|
|
|
|
generator->generateFastPath(jit);
|
|
CCallHelpers::Label done = jit.label();
|
|
|
|
params.addLatePath(
|
|
[=] (CCallHelpers& jit) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
generator->slowPathJump().link(&jit);
|
|
CCallHelpers::Label slowPathBegin = jit.label();
|
|
CCallHelpers::Call slowPathCall = callOperation(
|
|
*state, params.unavailableRegisters(), jit, node->origin.semantic,
|
|
exceptions.get(), generator->slowPathFunction(), InvalidGPRReg,
|
|
jit.codeBlock()->globalObjectFor(node->origin.semantic),
|
|
CCallHelpers::TrustedImmPtr(generator->stubInfo()), params[1].gpr(),
|
|
params[0].gpr(), identifier.rawBits()).call();
|
|
jit.jump().linkTo(done, &jit);
|
|
|
|
generator->reportSlowPathCall(slowPathBegin, slowPathCall);
|
|
|
|
jit.addLinkTask(
|
|
[=] (LinkBuffer& linkBuffer) {
|
|
generator->finalize(linkBuffer, linkBuffer);
|
|
});
|
|
});
|
|
});
|
|
}
|
|
|
|
void compilePutById()
|
|
{
|
|
DFG_ASSERT(m_graph, m_node, m_node->child1().useKind() == CellUse, m_node->child1().useKind());
|
|
|
|
Node* node = m_node;
|
|
LValue base = lowCell(node->child1());
|
|
LValue value = lowJSValue(node->child2());
|
|
auto putKind = node->op() == PutByIdDirect ? PutKind::Direct : PutKind::NotDirect;
|
|
cachedPutById(node, base, value, node->ecmaMode(), putKind);
|
|
}
|
|
|
|
void compileGetButterfly()
|
|
{
|
|
LValue butterfly = m_out.loadPtr(lowCell(m_node->child1()), m_heaps.JSObject_butterfly);
|
|
setStorage(butterfly);
|
|
}
|
|
|
|
void compileConstantStoragePointer()
|
|
{
|
|
setStorage(m_out.constIntPtr(m_node->storagePointer()));
|
|
}
|
|
|
|
void compileGetIndexedPropertyStorage()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue cell = lowCell(m_node->child1());
|
|
|
|
if (m_node->arrayMode().type() == Array::String) {
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue fastResultValue = m_out.loadPtr(cell, m_heaps.JSString_value);
|
|
ValueFromBlock fastResult = m_out.anchor(fastResultValue);
|
|
|
|
m_out.branch(isRopeString(cell, m_node->child1()), rarely(slowPath), usually(continuation));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(slowPath, continuation);
|
|
|
|
ValueFromBlock slowResult = m_out.anchor(vmCall(pointerType(), operationResolveRope, weakPointer(globalObject), cell));
|
|
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
|
|
setStorage(m_out.loadPtr(m_out.phi(pointerType(), fastResult, slowResult), m_heaps.StringImpl_data));
|
|
return;
|
|
}
|
|
|
|
DFG_ASSERT(m_graph, m_node, isTypedView(m_node->arrayMode().typedArrayType()), m_node->arrayMode().typedArrayType());
|
|
LValue vector = m_out.loadPtr(cell, m_heaps.JSArrayBufferView_vector);
|
|
setStorage(caged(Gigacage::Primitive, vector, cell));
|
|
}
|
|
|
|
void compileCheckArray()
|
|
{
|
|
Edge edge = m_node->child1();
|
|
LValue cell = lowCell(edge);
|
|
|
|
if (m_node->arrayMode().alreadyChecked(m_graph, m_node, abstractValue(edge)))
|
|
return;
|
|
|
|
speculate(
|
|
BadIndexingType, jsValueValue(cell), nullptr,
|
|
m_out.logicalNot(isArrayTypeForCheckArray(cell, m_node->arrayMode())));
|
|
}
|
|
|
|
void compileCheckArrayOrEmpty()
|
|
{
|
|
Edge edge = m_node->child1();
|
|
LValue cell = lowCell(edge);
|
|
|
|
if (m_node->arrayMode().alreadyChecked(m_graph, m_node, abstractValue(edge))) {
|
|
// We can purge Empty check of CheckArrayOrEmpty completely in this case since CellUse only accepts SpecCell | SpecEmpty.
|
|
ASSERT(typeFilterFor(m_node->child1().useKind()) & SpecEmpty);
|
|
return;
|
|
}
|
|
|
|
bool maySeeEmptyValue = m_interpreter.forNode(m_node->child1()).m_type & SpecEmpty;
|
|
LBasicBlock continuation = nullptr;
|
|
LBasicBlock lastNext = nullptr;
|
|
if (maySeeEmptyValue) {
|
|
LBasicBlock notEmpty = m_out.newBlock();
|
|
continuation = m_out.newBlock();
|
|
m_out.branch(m_out.isZero64(cell), unsure(continuation), unsure(notEmpty));
|
|
lastNext = m_out.appendTo(notEmpty, continuation);
|
|
}
|
|
|
|
speculate(
|
|
BadIndexingType, jsValueValue(cell), nullptr,
|
|
m_out.logicalNot(isArrayTypeForCheckArray(cell, m_node->arrayMode())));
|
|
|
|
if (maySeeEmptyValue) {
|
|
m_out.jump(continuation);
|
|
m_out.appendTo(continuation, lastNext);
|
|
}
|
|
}
|
|
|
|
void compileCheckDetached()
|
|
{
|
|
Edge edge = m_node->child1();
|
|
LValue cell = lowCell(edge);
|
|
|
|
speculate(
|
|
BadIndexingType, jsValueValue(cell), edge.node(),
|
|
m_out.isNull(m_out.loadPtr(cell, m_heaps.JSArrayBufferView_vector)));
|
|
}
|
|
|
|
void compileGetTypedArrayByteOffset()
|
|
{
|
|
LValue basePtr = lowCell(m_node->child1());
|
|
|
|
LBasicBlock wastefulCase = m_out.newBlock();
|
|
LBasicBlock notNull = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
ValueFromBlock nullVectorOut = m_out.anchor(m_out.constIntPtr(0));
|
|
|
|
LValue mode = m_out.load32(basePtr, m_heaps.JSArrayBufferView_mode);
|
|
m_out.branch(
|
|
m_out.notEqual(mode, m_out.constInt32(WastefulTypedArray)),
|
|
unsure(continuation), unsure(wastefulCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(wastefulCase, notNull);
|
|
|
|
LValue vector = m_out.loadPtr(basePtr, m_heaps.JSArrayBufferView_vector);
|
|
m_out.branch(m_out.equal(vector, m_out.constIntPtr(JSArrayBufferView::nullVectorPtr())),
|
|
unsure(continuation), unsure(notNull));
|
|
|
|
m_out.appendTo(notNull, continuation);
|
|
|
|
LValue butterflyPtr = caged(Gigacage::JSValue, m_out.loadPtr(basePtr, m_heaps.JSObject_butterfly), basePtr);
|
|
LValue arrayBufferPtr = m_out.loadPtr(butterflyPtr, m_heaps.Butterfly_arrayBuffer);
|
|
|
|
LValue vectorPtr = caged(Gigacage::Primitive, vector, basePtr);
|
|
|
|
// FIXME: This needs caging.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=175515
|
|
LValue dataPtr = m_out.loadPtr(arrayBufferPtr, m_heaps.ArrayBuffer_data);
|
|
dataPtr = removeArrayPtrTag(dataPtr);
|
|
|
|
ValueFromBlock wastefulOut = m_out.anchor(m_out.sub(vectorPtr, dataPtr));
|
|
|
|
m_out.jump(continuation);
|
|
m_out.appendTo(continuation, lastNext);
|
|
|
|
setInt32(m_out.castToInt32(m_out.phi(pointerType(), nullVectorOut, wastefulOut)));
|
|
}
|
|
|
|
void compileGetPrototypeOf()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
|
|
LValue object = nullptr;
|
|
LValue structure = nullptr;
|
|
ValueFromBlock slowResult;
|
|
|
|
LBasicBlock loadPolyProto = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(continuation);
|
|
|
|
switch (m_node->child1().useKind()) {
|
|
case ArrayUse:
|
|
case FunctionUse:
|
|
case FinalObjectUse: {
|
|
object = lowCell(m_node->child1());
|
|
switch (m_node->child1().useKind()) {
|
|
case ArrayUse:
|
|
speculateArray(m_node->child1(), object);
|
|
break;
|
|
case FunctionUse:
|
|
speculateFunction(m_node->child1(), object);
|
|
break;
|
|
case FinalObjectUse:
|
|
speculateFinalObject(m_node->child1(), object);
|
|
break;
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
break;
|
|
}
|
|
|
|
structure = loadStructure(object);
|
|
|
|
AbstractValue& value = m_state.forNode(m_node->child1());
|
|
if ((value.m_type && !(value.m_type & ~SpecObject)) && value.m_structure.isFinite()) {
|
|
bool hasPolyProto = false;
|
|
bool hasMonoProto = false;
|
|
value.m_structure.forEach([&] (RegisteredStructure structure) {
|
|
if (structure->hasPolyProto())
|
|
hasPolyProto = true;
|
|
else
|
|
hasMonoProto = true;
|
|
});
|
|
|
|
if (hasMonoProto && !hasPolyProto) {
|
|
setJSValue(m_out.load64(structure, m_heaps.Structure_prototype));
|
|
return;
|
|
}
|
|
|
|
if (hasPolyProto && !hasMonoProto) {
|
|
setJSValue(m_out.load64(m_out.baseIndex(m_heaps.properties.atAnyNumber(), object, m_out.constInt64(knownPolyProtoOffset), ScaleEight, JSObject::offsetOfInlineStorage())));
|
|
return;
|
|
}
|
|
}
|
|
|
|
break;
|
|
}
|
|
case ObjectUse: {
|
|
object = lowObject(m_node->child1());
|
|
|
|
LBasicBlock fastPath = m_out.newBlock();
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
|
|
structure = loadStructure(object);
|
|
m_out.branch(
|
|
m_out.testIsZero32(
|
|
m_out.load16ZeroExt32(structure, m_heaps.Structure_outOfLineTypeFlags),
|
|
m_out.constInt32(OverridesGetPrototypeOutOfLine)),
|
|
usually(fastPath), rarely(slowPath));
|
|
|
|
m_out.appendTo(slowPath, fastPath);
|
|
slowResult = m_out.anchor(vmCall(Int64, operationGetPrototypeOfObject, weakPointer(globalObject), object));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(fastPath, loadPolyProto);
|
|
break;
|
|
}
|
|
default: {
|
|
object = lowJSValue(m_node->child1());
|
|
SpeculatedType valueType = provenType(m_node->child1());
|
|
|
|
LBasicBlock isCellPath = m_out.newBlock();
|
|
LBasicBlock isObjectPath = m_out.newBlock();
|
|
LBasicBlock fastPath = m_out.newBlock();
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
|
|
m_out.branch(isCell(object, valueType), usually(isCellPath), rarely(slowPath));
|
|
m_out.appendTo(isCellPath, isObjectPath);
|
|
m_out.branch(isObject(object, valueType), usually(isObjectPath), rarely(slowPath));
|
|
|
|
m_out.appendTo(isObjectPath, slowPath);
|
|
structure = loadStructure(object);
|
|
m_out.branch(
|
|
m_out.testIsZero32(
|
|
m_out.load16ZeroExt32(structure, m_heaps.Structure_outOfLineTypeFlags),
|
|
m_out.constInt32(OverridesGetPrototypeOutOfLine)),
|
|
usually(fastPath), rarely(slowPath));
|
|
|
|
m_out.appendTo(slowPath, fastPath);
|
|
slowResult = m_out.anchor(vmCall(Int64, operationGetPrototypeOf, weakPointer(globalObject), object));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(fastPath, loadPolyProto);
|
|
break;
|
|
}
|
|
}
|
|
|
|
ASSERT(object);
|
|
ASSERT(structure);
|
|
|
|
LValue prototypeBits = m_out.load64(structure, m_heaps.Structure_prototype);
|
|
ValueFromBlock monoProto = m_out.anchor(prototypeBits);
|
|
m_out.branch(m_out.isZero64(prototypeBits), unsure(loadPolyProto), unsure(continuation));
|
|
|
|
m_out.appendTo(loadPolyProto, continuation);
|
|
ValueFromBlock polyProto = m_out.anchor(
|
|
m_out.load64(m_out.baseIndex(m_heaps.properties.atAnyNumber(), object, m_out.constInt64(knownPolyProtoOffset), ScaleEight, JSObject::offsetOfInlineStorage())));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, monoProto, polyProto, slowResult));
|
|
}
|
|
|
|
void compileGetArrayLength()
|
|
{
|
|
switch (m_node->arrayMode().type()) {
|
|
case Array::Undecided:
|
|
case Array::Int32:
|
|
case Array::Double:
|
|
case Array::Contiguous: {
|
|
setInt32(m_out.load32NonNegative(lowStorage(m_node->child2()), m_heaps.Butterfly_publicLength));
|
|
return;
|
|
}
|
|
|
|
case Array::ArrayStorage:
|
|
case Array::SlowPutArrayStorage: {
|
|
LValue length = m_out.load32(lowStorage(m_node->child2()), m_heaps.ArrayStorage_publicLength);
|
|
speculate(Uncountable, noValue(), nullptr, m_out.lessThan(length, m_out.int32Zero));
|
|
setInt32(length);
|
|
return;
|
|
}
|
|
|
|
case Array::String: {
|
|
LValue string = lowCell(m_node->child1());
|
|
|
|
LBasicBlock ropePath = m_out.newBlock();
|
|
LBasicBlock nonRopePath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(isRopeString(string, m_node->child1()), rarely(ropePath), usually(nonRopePath));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(ropePath, nonRopePath);
|
|
ValueFromBlock ropeLength = m_out.anchor(m_out.load32NonNegative(string, m_heaps.JSRopeString_length));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(nonRopePath, continuation);
|
|
ValueFromBlock nonRopeLength = m_out.anchor(m_out.load32NonNegative(m_out.loadPtr(string, m_heaps.JSString_value), m_heaps.StringImpl_length));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setInt32(m_out.phi(Int32, ropeLength, nonRopeLength));
|
|
return;
|
|
}
|
|
|
|
case Array::DirectArguments: {
|
|
LValue arguments = lowCell(m_node->child1());
|
|
speculate(
|
|
ExoticObjectMode, noValue(), nullptr,
|
|
m_out.notNull(m_out.loadPtr(arguments, m_heaps.DirectArguments_mappedArguments)));
|
|
setInt32(m_out.load32NonNegative(arguments, m_heaps.DirectArguments_length));
|
|
return;
|
|
}
|
|
|
|
case Array::ScopedArguments: {
|
|
LValue arguments = lowCell(m_node->child1());
|
|
speculate(
|
|
ExoticObjectMode, noValue(), nullptr,
|
|
m_out.notZero32(m_out.load8ZeroExt32(arguments, m_heaps.ScopedArguments_overrodeThings)));
|
|
setInt32(m_out.load32NonNegative(arguments, m_heaps.ScopedArguments_totalLength));
|
|
return;
|
|
}
|
|
|
|
default:
|
|
if (m_node->arrayMode().isSomeTypedArrayView()) {
|
|
setInt32(
|
|
m_out.load32NonNegative(lowCell(m_node->child1()), m_heaps.JSArrayBufferView_length));
|
|
return;
|
|
}
|
|
|
|
DFG_CRASH(m_graph, m_node, "Bad array type");
|
|
return;
|
|
}
|
|
}
|
|
|
|
void compileGetVectorLength()
|
|
{
|
|
switch (m_node->arrayMode().type()) {
|
|
case Array::ArrayStorage:
|
|
case Array::SlowPutArrayStorage:
|
|
setInt32(m_out.load32NonNegative(lowStorage(m_node->child2()), m_heaps.ArrayStorage_vectorLength));
|
|
return;
|
|
default:
|
|
return;
|
|
}
|
|
}
|
|
|
|
void compileAssertInBounds()
|
|
{
|
|
ASSERT(Options::validateBoundsCheckElimination());
|
|
LValue index = lowInt32(m_node->child1());
|
|
LValue bounds = lowInt32(m_node->child2());
|
|
|
|
LBasicBlock outOfBoundsCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
m_out.branch(m_out.below(index, bounds), usually(continuation), rarely(outOfBoundsCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(outOfBoundsCase, continuation);
|
|
vmCall(Void, operationReportBoundsCheckEliminationErrorAndCrash,
|
|
m_out.constIntPtr(bitwise_cast<intptr_t>(codeBlock())),
|
|
m_out.constInt32(m_node->index()),
|
|
m_out.constInt32(m_node->child1()->index()),
|
|
m_out.constInt32(m_node->child2()->index()),
|
|
index, bounds);
|
|
m_out.unreachable();
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
}
|
|
|
|
void compileCheckInBounds()
|
|
{
|
|
speculate(
|
|
OutOfBounds, noValue(), nullptr,
|
|
m_out.aboveOrEqual(lowInt32(m_node->child1()), lowInt32(m_node->child2())));
|
|
|
|
// Even though we claim to have JSValue result, no user of us should
|
|
// depend on our value. Users of this node just need to maintain that
|
|
// we dominate them.
|
|
}
|
|
|
|
void compileGetByVal()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
switch (m_node->arrayMode().type()) {
|
|
case Array::Int32:
|
|
case Array::Contiguous: {
|
|
LValue index = lowInt32(m_graph.varArgChild(m_node, 1));
|
|
LValue storage = lowStorage(m_graph.varArgChild(m_node, 2));
|
|
|
|
IndexedAbstractHeap& heap = m_node->arrayMode().type() == Array::Int32 ?
|
|
m_heaps.indexedInt32Properties : m_heaps.indexedContiguousProperties;
|
|
|
|
LValue base = lowCell(m_graph.varArgChild(m_node, 0));
|
|
|
|
if (m_node->arrayMode().isInBounds()) {
|
|
LValue result = m_out.load64(baseIndex(heap, storage, index, m_graph.varArgChild(m_node, 1)));
|
|
LValue isHole = m_out.isZero64(result);
|
|
if (m_node->arrayMode().isInBoundsSaneChain()) {
|
|
DFG_ASSERT(
|
|
m_graph, m_node, m_node->arrayMode().type() == Array::Contiguous, m_node->arrayMode().type());
|
|
result = m_out.select(
|
|
isHole, m_out.constInt64(JSValue::encode(jsUndefined())), result);
|
|
} else
|
|
speculate(LoadFromHole, noValue(), nullptr, isHole);
|
|
// We have to keep base alive to keep content in storage alive.
|
|
if (m_node->arrayMode().type() == Array::Contiguous)
|
|
ensureStillAliveHere(base);
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
|
|
LBasicBlock fastCase = m_out.newBlock();
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
m_out.aboveOrEqual(
|
|
index, m_out.load32NonNegative(storage, m_heaps.Butterfly_publicLength)),
|
|
rarely(slowCase), usually(fastCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(fastCase, slowCase);
|
|
|
|
LValue fastResultValue = m_out.load64(baseIndex(heap, storage, index, m_graph.varArgChild(m_node, 1)));
|
|
ValueFromBlock fastResult = m_out.anchor(fastResultValue);
|
|
m_out.branch(
|
|
m_out.isZero64(fastResultValue), rarely(slowCase), usually(continuation));
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
ValueFromBlock slowResult;
|
|
if (m_node->arrayMode().isOutOfBoundsSaneChain()) {
|
|
speculate(NegativeIndex, noValue(), nullptr, m_out.lessThan(index, m_out.int32Zero));
|
|
slowResult = m_out.anchor(m_out.constInt64(JSValue::ValueUndefined));
|
|
} else
|
|
slowResult = m_out.anchor(vmCall(Int64, operationGetByValObjectInt, weakPointer(globalObject), base, index));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
// We have to keep base alive to keep content in storage alive.
|
|
if (m_node->arrayMode().type() == Array::Contiguous)
|
|
ensureStillAliveHere(base);
|
|
setJSValue(m_out.phi(Int64, fastResult, slowResult));
|
|
return;
|
|
}
|
|
|
|
case Array::Double: {
|
|
LValue base = lowCell(m_graph.varArgChild(m_node, 0));
|
|
LValue index = lowInt32(m_graph.varArgChild(m_node, 1));
|
|
LValue storage = lowStorage(m_graph.varArgChild(m_node, 2));
|
|
|
|
IndexedAbstractHeap& heap = m_heaps.indexedDoubleProperties;
|
|
|
|
if (m_node->arrayMode().isInBounds()) {
|
|
LValue result = m_out.loadDouble(
|
|
baseIndex(heap, storage, index, m_graph.varArgChild(m_node, 1)));
|
|
|
|
if (!m_node->arrayMode().isInBoundsSaneChain()) {
|
|
speculate(
|
|
LoadFromHole, noValue(), nullptr,
|
|
m_out.doubleNotEqualOrUnordered(result, result));
|
|
}
|
|
setDouble(result);
|
|
break;
|
|
}
|
|
|
|
bool resultIsUnboxed = m_node->arrayMode().isOutOfBoundsSaneChain() && !(m_node->flags() & NodeBytecodeUsesAsOther);
|
|
|
|
LBasicBlock inBounds = m_out.newBlock();
|
|
LBasicBlock boxPath = m_out.newBlock();
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
m_out.aboveOrEqual(
|
|
index, m_out.load32NonNegative(storage, m_heaps.Butterfly_publicLength)),
|
|
rarely(slowCase), usually(inBounds));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(inBounds, boxPath);
|
|
LValue doubleValue = m_out.loadDouble(
|
|
baseIndex(heap, storage, index, m_graph.varArgChild(m_node, 1)));
|
|
m_out.branch(
|
|
m_out.doubleNotEqualOrUnordered(doubleValue, doubleValue),
|
|
rarely(slowCase), usually(boxPath));
|
|
|
|
m_out.appendTo(boxPath, slowCase);
|
|
ValueFromBlock fastResult = m_out.anchor(resultIsUnboxed ? doubleValue : boxDouble(doubleValue));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
ValueFromBlock slowResult;
|
|
if (m_node->arrayMode().isOutOfBoundsSaneChain()) {
|
|
speculate(NegativeIndex, noValue(), nullptr, m_out.lessThan(index, m_out.int32Zero));
|
|
if (resultIsUnboxed)
|
|
slowResult = m_out.anchor(m_out.constDouble(PNaN));
|
|
else
|
|
slowResult = m_out.anchor(m_out.constInt64(JSValue::encode(jsUndefined())));
|
|
} else
|
|
slowResult = m_out.anchor(vmCall(Int64, operationGetByValObjectInt, weakPointer(globalObject), base, index));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
if (resultIsUnboxed)
|
|
setDouble(m_out.phi(Double, fastResult, slowResult));
|
|
else
|
|
setJSValue(m_out.phi(Int64, fastResult, slowResult));
|
|
return;
|
|
}
|
|
|
|
case Array::Undecided: {
|
|
LValue index = lowInt32(m_graph.varArgChild(m_node, 1));
|
|
|
|
speculate(OutOfBounds, noValue(), m_node, m_out.lessThan(index, m_out.int32Zero));
|
|
setJSValue(m_out.constInt64(JSValue::ValueUndefined));
|
|
return;
|
|
}
|
|
|
|
case Array::DirectArguments: {
|
|
LValue base = lowCell(m_graph.varArgChild(m_node, 0));
|
|
LValue index = lowInt32(m_graph.varArgChild(m_node, 1));
|
|
|
|
speculate(
|
|
ExoticObjectMode, noValue(), nullptr,
|
|
m_out.notNull(m_out.loadPtr(base, m_heaps.DirectArguments_mappedArguments)));
|
|
|
|
LValue length = m_out.load32NonNegative(base, m_heaps.DirectArguments_length);
|
|
auto isOutOfBounds = m_out.aboveOrEqual(index, length);
|
|
if (m_node->arrayMode().isInBounds()) {
|
|
speculate(OutOfBounds, noValue(), nullptr, isOutOfBounds);
|
|
TypedPointer address = m_out.baseIndex(
|
|
m_heaps.DirectArguments_storage, base, m_out.zeroExtPtr(index));
|
|
setJSValue(m_out.load64(address));
|
|
return;
|
|
}
|
|
|
|
LBasicBlock inBounds = m_out.newBlock();
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(isOutOfBounds, rarely(slowCase), usually(inBounds));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(inBounds, slowCase);
|
|
TypedPointer address = m_out.baseIndex(
|
|
m_heaps.DirectArguments_storage,
|
|
base,
|
|
m_out.zeroExt(index, pointerType()));
|
|
ValueFromBlock fastResult = m_out.anchor(m_out.load64(address));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
ValueFromBlock slowResult = m_out.anchor(vmCall(Int64, operationGetByValObjectInt, weakPointer(globalObject), base, index));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, fastResult, slowResult));
|
|
return;
|
|
}
|
|
|
|
case Array::ScopedArguments: {
|
|
LValue base = lowCell(m_graph.varArgChild(m_node, 0));
|
|
LValue index = lowInt32(m_graph.varArgChild(m_node, 1));
|
|
|
|
speculate(
|
|
ExoticObjectMode, noValue(), nullptr,
|
|
m_out.aboveOrEqual(
|
|
index,
|
|
m_out.load32NonNegative(base, m_heaps.ScopedArguments_totalLength)));
|
|
|
|
LValue table = m_out.loadPtr(base, m_heaps.ScopedArguments_table);
|
|
LValue namedLength = m_out.load32(table, m_heaps.ScopedArgumentsTable_length);
|
|
|
|
LBasicBlock namedCase = m_out.newBlock();
|
|
LBasicBlock overflowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
m_out.aboveOrEqual(index, namedLength), unsure(overflowCase), unsure(namedCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(namedCase, overflowCase);
|
|
|
|
LValue scope = m_out.loadPtr(base, m_heaps.ScopedArguments_scope);
|
|
LValue arguments = m_out.loadPtr(table, m_heaps.ScopedArgumentsTable_arguments);
|
|
|
|
TypedPointer address = m_out.baseIndex(
|
|
m_heaps.scopedArgumentsTableArguments, arguments, m_out.zeroExtPtr(index));
|
|
LValue scopeOffset = m_out.load32(address);
|
|
|
|
speculate(
|
|
ExoticObjectMode, noValue(), nullptr,
|
|
m_out.equal(scopeOffset, m_out.constInt32(ScopeOffset::invalidOffset)));
|
|
|
|
address = m_out.baseIndex(
|
|
m_heaps.JSLexicalEnvironment_variables, scope, m_out.zeroExtPtr(scopeOffset));
|
|
ValueFromBlock namedResult = m_out.anchor(m_out.load64(address));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(overflowCase, continuation);
|
|
|
|
LValue storage = m_out.loadPtr(base, m_heaps.ScopedArguments_storage);
|
|
address = m_out.baseIndex(
|
|
m_heaps.ScopedArguments_Storage_storage, storage,
|
|
m_out.zeroExtPtr(m_out.sub(index, namedLength)));
|
|
LValue overflowValue = m_out.load64(address);
|
|
speculate(ExoticObjectMode, noValue(), nullptr, m_out.isZero64(overflowValue));
|
|
ValueFromBlock overflowResult = m_out.anchor(overflowValue);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, namedResult, overflowResult));
|
|
return;
|
|
}
|
|
|
|
case Array::Generic: {
|
|
if (m_graph.m_slowGetByVal.contains(m_node)) {
|
|
if (m_graph.varArgChild(m_node, 0).useKind() == ObjectUse) {
|
|
if (m_graph.varArgChild(m_node, 1).useKind() == StringUse) {
|
|
setJSValue(vmCall(
|
|
Int64, operationGetByValObjectString, weakPointer(globalObject),
|
|
lowObject(m_graph.varArgChild(m_node, 0)), lowString(m_graph.varArgChild(m_node, 1))));
|
|
return;
|
|
}
|
|
|
|
if (m_graph.varArgChild(m_node, 1).useKind() == SymbolUse) {
|
|
setJSValue(vmCall(
|
|
Int64, operationGetByValObjectSymbol, weakPointer(globalObject),
|
|
lowObject(m_graph.varArgChild(m_node, 0)), lowSymbol(m_graph.varArgChild(m_node, 1))));
|
|
return;
|
|
}
|
|
}
|
|
|
|
setJSValue(vmCall(
|
|
Int64, operationGetByVal, weakPointer(globalObject),
|
|
lowJSValue(m_graph.varArgChild(m_node, 0)), lowJSValue(m_graph.varArgChild(m_node, 1))));
|
|
return;
|
|
}
|
|
|
|
Node* node = m_node;
|
|
|
|
LValue base = lowJSValue(m_graph.varArgChild(node, 0), ManualOperandSpeculation);
|
|
LValue property = lowJSValue(m_graph.varArgChild(node, 1), ManualOperandSpeculation);
|
|
|
|
speculate(m_graph.varArgChild(node, 0));
|
|
speculate(m_graph.varArgChild(node, 1));
|
|
bool baseIsCell = abstractValue(m_graph.varArgChild(node, 0)).isType(SpecCell);
|
|
bool propertyIsString = false;
|
|
bool propertyIsInt32 = false;
|
|
bool propertyIsSymbol = false;
|
|
if (abstractValue(m_graph.varArgChild(node, 1)).isType(SpecString))
|
|
propertyIsString = true;
|
|
else if (abstractValue(m_graph.varArgChild(node, 1)).isType(SpecInt32Only))
|
|
propertyIsInt32 = true;
|
|
else if (abstractValue(m_graph.varArgChild(node, 1)).isType(SpecSymbol))
|
|
propertyIsSymbol = true;
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int64);
|
|
patchpoint->appendSomeRegister(base);
|
|
patchpoint->appendSomeRegister(property);
|
|
patchpoint->append(m_notCellMask, ValueRep::lateReg(GPRInfo::notCellMaskRegister));
|
|
patchpoint->append(m_numberTag, ValueRep::lateReg(GPRInfo::numberTagRegister));
|
|
patchpoint->clobber(RegisterSet::macroScratchRegisters());
|
|
|
|
RefPtr<PatchpointExceptionHandle> exceptionHandle = preparePatchpointForExceptions(patchpoint);
|
|
|
|
State* state = &m_ftlState;
|
|
patchpoint->setGenerator([=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
CallSiteIndex callSiteIndex = state->jitCode->common.codeOrigins->addUniqueCallSiteIndex(node->origin.semantic);
|
|
|
|
// This is the direct exit target for operation calls.
|
|
Box<CCallHelpers::JumpList> exceptions = exceptionHandle->scheduleExitCreation(params)->jumps(jit);
|
|
|
|
// This is the exit for call IC's created by the IC for getters. We don't have
|
|
// to do anything weird other than call this, since it will associate the exit with
|
|
// the callsite index.
|
|
exceptionHandle->scheduleExitCreationForUnwind(params, callSiteIndex);
|
|
|
|
GPRReg resultGPR = params[0].gpr();
|
|
GPRReg baseGPR = params[1].gpr();
|
|
GPRReg propertyGPR = params[2].gpr();
|
|
|
|
auto generator = Box<JITGetByValGenerator>::create(
|
|
jit.codeBlock(), node->origin.semantic, callSiteIndex, AccessType::GetByVal,
|
|
params.unavailableRegisters(), JSValueRegs(baseGPR), JSValueRegs(propertyGPR), JSValueRegs(resultGPR));
|
|
|
|
generator->stubInfo()->propertyIsString = propertyIsString;
|
|
generator->stubInfo()->propertyIsInt32 = propertyIsInt32;
|
|
generator->stubInfo()->propertyIsSymbol = propertyIsSymbol;
|
|
|
|
CCallHelpers::Jump notCell;
|
|
if (!baseIsCell)
|
|
notCell = jit.branchIfNotCell(baseGPR);
|
|
|
|
generator->generateFastPath(jit);
|
|
CCallHelpers::Label done = jit.label();
|
|
|
|
params.addLatePath([=] (CCallHelpers& jit) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
if (notCell.isSet())
|
|
notCell.link(&jit);
|
|
generator->slowPathJump().link(&jit);
|
|
CCallHelpers::Label slowPathBegin = jit.label();
|
|
CCallHelpers::Call slowPathCall = callOperation(
|
|
*state, params.unavailableRegisters(), jit, node->origin.semantic,
|
|
exceptions.get(), operationGetByValOptimize, resultGPR,
|
|
jit.codeBlock()->globalObjectFor(node->origin.semantic),
|
|
CCallHelpers::TrustedImmPtr(generator->stubInfo()), CCallHelpers::TrustedImmPtr(nullptr), baseGPR, propertyGPR).call();
|
|
jit.jump().linkTo(done, &jit);
|
|
|
|
generator->reportSlowPathCall(slowPathBegin, slowPathCall);
|
|
|
|
jit.addLinkTask([=] (LinkBuffer& linkBuffer) {
|
|
generator->finalize(linkBuffer, linkBuffer);
|
|
});
|
|
});
|
|
});
|
|
|
|
setJSValue(patchpoint);
|
|
return;
|
|
}
|
|
|
|
case Array::ArrayStorage:
|
|
case Array::SlowPutArrayStorage: {
|
|
LValue base = lowCell(m_graph.varArgChild(m_node, 0));
|
|
LValue index = lowInt32(m_graph.varArgChild(m_node, 1));
|
|
LValue storage = lowStorage(m_graph.varArgChild(m_node, 2));
|
|
|
|
IndexedAbstractHeap& heap = m_heaps.ArrayStorage_vector;
|
|
|
|
if (m_node->arrayMode().isInBounds()) {
|
|
LValue result = m_out.load64(baseIndex(heap, storage, index, m_graph.varArgChild(m_node, 1)));
|
|
speculate(LoadFromHole, noValue(), nullptr, m_out.isZero64(result));
|
|
// We have to keep base alive to keep content in storage alive.
|
|
ensureStillAliveHere(base);
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
|
|
LBasicBlock inBounds = m_out.newBlock();
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
m_out.aboveOrEqual(index, m_out.load32NonNegative(storage, m_heaps.ArrayStorage_vectorLength)),
|
|
rarely(slowCase), usually(inBounds));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(inBounds, slowCase);
|
|
LValue result = m_out.load64(baseIndex(heap, storage, index, m_graph.varArgChild(m_node, 1)));
|
|
ValueFromBlock fastResult = m_out.anchor(result);
|
|
m_out.branch(
|
|
m_out.isZero64(result),
|
|
rarely(slowCase), usually(continuation));
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
ValueFromBlock slowResult = m_out.anchor(
|
|
vmCall(Int64, operationGetByValObjectInt, weakPointer(globalObject), base, index));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
// We have to keep base alive to keep content in storage alive.
|
|
ensureStillAliveHere(base);
|
|
setJSValue(m_out.phi(Int64, fastResult, slowResult));
|
|
return;
|
|
}
|
|
|
|
case Array::String: {
|
|
compileStringCharAt();
|
|
return;
|
|
}
|
|
|
|
case Array::Int8Array:
|
|
case Array::Int16Array:
|
|
case Array::Int32Array:
|
|
case Array::Uint8Array:
|
|
case Array::Uint8ClampedArray:
|
|
case Array::Uint16Array:
|
|
case Array::Uint32Array:
|
|
case Array::Float32Array:
|
|
case Array::Float64Array: {
|
|
LValue base = lowCell(m_graph.varArgChild(m_node, 0));
|
|
LValue index = lowInt32(m_graph.varArgChild(m_node, 1));
|
|
LValue storage = lowStorage(m_graph.varArgChild(m_node, 2));
|
|
|
|
TypedArrayType type = m_node->arrayMode().typedArrayType();
|
|
ASSERT(isTypedView(type));
|
|
{
|
|
TypedPointer pointer = pointerIntoTypedArray(storage, index, type);
|
|
|
|
if (isInt(type)) {
|
|
LValue result = loadFromIntTypedArray(pointer, type);
|
|
// We have to keep base alive since that keeps storage alive.
|
|
ensureStillAliveHere(base);
|
|
bool canSpeculate = true;
|
|
setIntTypedArrayLoadResult(result, type, canSpeculate);
|
|
return;
|
|
}
|
|
|
|
ASSERT(isFloat(type));
|
|
|
|
LValue result;
|
|
switch (type) {
|
|
case TypeFloat32:
|
|
result = m_out.floatToDouble(m_out.loadFloat(pointer));
|
|
break;
|
|
case TypeFloat64:
|
|
result = m_out.loadDouble(pointer);
|
|
break;
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad typed array type");
|
|
}
|
|
|
|
// We have to keep base alive since that keeps storage alive.
|
|
ensureStillAliveHere(base);
|
|
setDouble(result);
|
|
return;
|
|
}
|
|
}
|
|
|
|
case Array::AnyTypedArray:
|
|
case Array::ForceExit:
|
|
case Array::SelectUsingArguments:
|
|
case Array::SelectUsingPredictions:
|
|
case Array::Unprofiled:
|
|
DFG_CRASH(m_graph, m_node, "Bad array type");
|
|
return;
|
|
}
|
|
}
|
|
|
|
void compileGetMyArgumentByVal()
|
|
{
|
|
InlineCallFrame* inlineCallFrame = m_node->child1()->origin.semantic.inlineCallFrame();
|
|
|
|
LValue originalIndex = lowInt32(m_node->child2());
|
|
|
|
LValue numberOfArgsIncludingThis;
|
|
if (inlineCallFrame && !inlineCallFrame->isVarargs())
|
|
numberOfArgsIncludingThis = m_out.constInt32(inlineCallFrame->argumentCountIncludingThis);
|
|
else {
|
|
VirtualRegister argumentCountRegister = AssemblyHelpers::argumentCount(inlineCallFrame);
|
|
numberOfArgsIncludingThis = m_out.load32(payloadFor(argumentCountRegister));
|
|
}
|
|
|
|
LValue numberOfArgs = m_out.sub(numberOfArgsIncludingThis, m_out.int32One);
|
|
LValue indexToCheck = originalIndex;
|
|
LValue numberOfArgumentsToSkip = m_out.int32Zero;
|
|
if (m_node->numberOfArgumentsToSkip()) {
|
|
numberOfArgumentsToSkip = m_out.constInt32(m_node->numberOfArgumentsToSkip());
|
|
CheckValue* check = m_out.speculateAdd(indexToCheck, numberOfArgumentsToSkip);
|
|
blessSpeculation(check, Overflow, noValue(), nullptr, m_origin);
|
|
indexToCheck = check;
|
|
}
|
|
|
|
LValue isOutOfBounds = m_out.bitOr(m_out.aboveOrEqual(indexToCheck, numberOfArgs), m_out.below(indexToCheck, numberOfArgumentsToSkip));
|
|
LBasicBlock continuation = nullptr;
|
|
LBasicBlock lastNext = nullptr;
|
|
ValueFromBlock slowResult;
|
|
if (m_node->op() == GetMyArgumentByValOutOfBounds) {
|
|
LBasicBlock normalCase = m_out.newBlock();
|
|
continuation = m_out.newBlock();
|
|
|
|
slowResult = m_out.anchor(m_out.constInt64(JSValue::encode(jsUndefined())));
|
|
m_out.branch(isOutOfBounds, unsure(continuation), unsure(normalCase));
|
|
|
|
lastNext = m_out.appendTo(normalCase, continuation);
|
|
} else
|
|
speculate(OutOfBounds, noValue(), nullptr, isOutOfBounds);
|
|
|
|
LValue index = m_out.add(indexToCheck, m_out.int32One);
|
|
|
|
TypedPointer base;
|
|
if (inlineCallFrame) {
|
|
if (inlineCallFrame->argumentCountIncludingThis > 1)
|
|
base = addressFor(inlineCallFrame->argumentsWithFixup[0].virtualRegister());
|
|
} else
|
|
base = addressFor(virtualRegisterForArgumentIncludingThis(0));
|
|
|
|
LValue result;
|
|
if (base) {
|
|
LValue pointer = m_out.baseIndex(
|
|
base.value(), m_out.zeroExt(index, pointerType()), ScaleEight);
|
|
result = m_out.load64(TypedPointer(m_heaps.variables.atAnyIndex(), pointer));
|
|
} else
|
|
result = m_out.constInt64(JSValue::encode(jsUndefined()));
|
|
|
|
if (m_node->op() == GetMyArgumentByValOutOfBounds) {
|
|
ValueFromBlock normalResult = m_out.anchor(result);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
result = m_out.phi(Int64, slowResult, normalResult);
|
|
}
|
|
|
|
setJSValue(result);
|
|
}
|
|
|
|
void compilePutByVal()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
Edge child1 = m_graph.varArgChild(m_node, 0);
|
|
Edge child2 = m_graph.varArgChild(m_node, 1);
|
|
Edge child3 = m_graph.varArgChild(m_node, 2);
|
|
Edge child4 = m_graph.varArgChild(m_node, 3);
|
|
Edge child5 = m_graph.varArgChild(m_node, 4);
|
|
|
|
ArrayMode arrayMode = m_node->arrayMode().modeForPut();
|
|
switch (arrayMode.type()) {
|
|
case Array::Generic: {
|
|
if (child1.useKind() == CellUse) {
|
|
V_JITOperation_GCCJ operation = nullptr;
|
|
if (child2.useKind() == StringUse) {
|
|
if (m_node->op() == PutByValDirect) {
|
|
if (m_node->ecmaMode().isStrict())
|
|
operation = operationPutByValDirectCellStringStrict;
|
|
else
|
|
operation = operationPutByValDirectCellStringNonStrict;
|
|
} else {
|
|
if (m_node->ecmaMode().isStrict())
|
|
operation = operationPutByValCellStringStrict;
|
|
else
|
|
operation = operationPutByValCellStringNonStrict;
|
|
}
|
|
vmCall(Void, operation, weakPointer(globalObject), lowCell(child1), lowString(child2), lowJSValue(child3));
|
|
return;
|
|
}
|
|
|
|
if (child2.useKind() == SymbolUse) {
|
|
if (m_node->op() == PutByValDirect) {
|
|
if (m_node->ecmaMode().isStrict())
|
|
operation = operationPutByValDirectCellSymbolStrict;
|
|
else
|
|
operation = operationPutByValDirectCellSymbolNonStrict;
|
|
} else {
|
|
if (m_node->ecmaMode().isStrict())
|
|
operation = operationPutByValCellSymbolStrict;
|
|
else
|
|
operation = operationPutByValCellSymbolNonStrict;
|
|
}
|
|
vmCall(Void, operation, weakPointer(globalObject), lowCell(child1), lowSymbol(child2), lowJSValue(child3));
|
|
return;
|
|
}
|
|
}
|
|
|
|
V_JITOperation_GJJJ operation;
|
|
if (m_node->op() == PutByValDirect) {
|
|
if (m_node->ecmaMode().isStrict())
|
|
operation = operationPutByValDirectStrict;
|
|
else
|
|
operation = operationPutByValDirectNonStrict;
|
|
} else {
|
|
if (m_node->ecmaMode().isStrict())
|
|
operation = operationPutByValStrict;
|
|
else
|
|
operation = operationPutByValNonStrict;
|
|
}
|
|
|
|
vmCall(
|
|
Void, operation, weakPointer(globalObject),
|
|
lowJSValue(child1), lowJSValue(child2), lowJSValue(child3));
|
|
return;
|
|
}
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
LValue base = lowCell(child1);
|
|
LValue index = lowInt32(child2);
|
|
LValue storage = lowStorage(child4);
|
|
|
|
switch (arrayMode.type()) {
|
|
case Array::Int32:
|
|
case Array::Double:
|
|
case Array::Contiguous: {
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
LBasicBlock outerLastNext = m_out.appendTo(m_out.m_block, continuation);
|
|
|
|
switch (arrayMode.type()) {
|
|
case Array::Int32:
|
|
case Array::Contiguous: {
|
|
LValue value = lowJSValue(child3, ManualOperandSpeculation);
|
|
|
|
if (arrayMode.type() == Array::Int32)
|
|
FTL_TYPE_CHECK(jsValueValue(value), child3, SpecInt32Only, isNotInt32(value));
|
|
|
|
TypedPointer elementPointer = m_out.baseIndex(
|
|
arrayMode.type() == Array::Int32 ?
|
|
m_heaps.indexedInt32Properties : m_heaps.indexedContiguousProperties,
|
|
storage, m_out.zeroExtPtr(index), provenValue(child2));
|
|
|
|
if (m_node->op() == PutByValAlias) {
|
|
m_out.store64(value, elementPointer);
|
|
break;
|
|
}
|
|
|
|
contiguousPutByValOutOfBounds(
|
|
m_node->ecmaMode().isStrict()
|
|
? (m_node->op() == PutByValDirect ? operationPutByValDirectBeyondArrayBoundsStrict : operationPutByValBeyondArrayBoundsStrict)
|
|
: (m_node->op() == PutByValDirect ? operationPutByValDirectBeyondArrayBoundsNonStrict : operationPutByValBeyondArrayBoundsNonStrict),
|
|
base, storage, index, value, continuation);
|
|
|
|
m_out.store64(value, elementPointer);
|
|
break;
|
|
}
|
|
|
|
case Array::Double: {
|
|
LValue value = lowDouble(child3);
|
|
|
|
FTL_TYPE_CHECK(
|
|
doubleValue(value), child3, SpecDoubleReal,
|
|
m_out.doubleNotEqualOrUnordered(value, value));
|
|
|
|
TypedPointer elementPointer = m_out.baseIndex(
|
|
m_heaps.indexedDoubleProperties, storage, m_out.zeroExtPtr(index),
|
|
provenValue(child2));
|
|
|
|
if (m_node->op() == PutByValAlias) {
|
|
m_out.storeDouble(value, elementPointer);
|
|
break;
|
|
}
|
|
|
|
contiguousPutByValOutOfBounds(
|
|
m_node->ecmaMode().isStrict()
|
|
? (m_node->op() == PutByValDirect ? operationPutDoubleByValDirectBeyondArrayBoundsStrict : operationPutDoubleByValBeyondArrayBoundsStrict)
|
|
: (m_node->op() == PutByValDirect ? operationPutDoubleByValDirectBeyondArrayBoundsNonStrict : operationPutDoubleByValBeyondArrayBoundsNonStrict),
|
|
base, storage, index, value, continuation);
|
|
|
|
m_out.storeDouble(value, elementPointer);
|
|
break;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad array type");
|
|
}
|
|
|
|
m_out.jump(continuation);
|
|
m_out.appendTo(continuation, outerLastNext);
|
|
return;
|
|
}
|
|
|
|
case Array::ArrayStorage:
|
|
case Array::SlowPutArrayStorage: {
|
|
LValue value = lowJSValue(child3);
|
|
|
|
TypedPointer elementPointer = m_out.baseIndex(
|
|
m_heaps.ArrayStorage_vector, storage, m_out.zeroExtPtr(index),
|
|
provenValue(child2));
|
|
|
|
if (m_node->op() == PutByValAlias) {
|
|
m_out.store64(value, elementPointer);
|
|
return;
|
|
}
|
|
|
|
if (arrayMode.isInBounds()) {
|
|
speculate(StoreToHole, noValue(), nullptr, m_out.isZero64(m_out.load64(elementPointer)));
|
|
m_out.store64(value, elementPointer);
|
|
return;
|
|
}
|
|
|
|
LValue isOutOfBounds = m_out.aboveOrEqual(
|
|
index, m_out.load32NonNegative(storage, m_heaps.ArrayStorage_vectorLength));
|
|
|
|
auto slowPathFunction = m_node->ecmaMode().isStrict()
|
|
? (m_node->op() == PutByValDirect ? operationPutByValDirectBeyondArrayBoundsStrict : operationPutByValBeyondArrayBoundsStrict)
|
|
: (m_node->op() == PutByValDirect ? operationPutByValDirectBeyondArrayBoundsNonStrict : operationPutByValBeyondArrayBoundsNonStrict);
|
|
if (!arrayMode.isOutOfBounds()) {
|
|
speculate(OutOfBounds, noValue(), nullptr, isOutOfBounds);
|
|
isOutOfBounds = m_out.booleanFalse;
|
|
}
|
|
|
|
LBasicBlock inBoundCase = m_out.newBlock();
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock holeCase = m_out.newBlock();
|
|
LBasicBlock doStoreCase = m_out.newBlock();
|
|
LBasicBlock lengthUpdateCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(isOutOfBounds, rarely(slowCase), usually(inBoundCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(slowCase, inBoundCase);
|
|
vmCall(
|
|
Void, slowPathFunction,
|
|
weakPointer(globalObject), base, index, value);
|
|
m_out.jump(continuation);
|
|
|
|
|
|
if (arrayMode.isSlowPut()) {
|
|
m_out.appendTo(inBoundCase, doStoreCase);
|
|
m_out.branch(m_out.isZero64(m_out.load64(elementPointer)), rarely(slowCase), usually(doStoreCase));
|
|
} else {
|
|
m_out.appendTo(inBoundCase, holeCase);
|
|
m_out.branch(m_out.isZero64(m_out.load64(elementPointer)), rarely(holeCase), usually(doStoreCase));
|
|
|
|
m_out.appendTo(holeCase, lengthUpdateCase);
|
|
m_out.store32(
|
|
m_out.add(m_out.load32(storage, m_heaps.ArrayStorage_numValuesInVector), m_out.int32One),
|
|
storage, m_heaps.ArrayStorage_numValuesInVector);
|
|
m_out.branch(
|
|
m_out.below(
|
|
index, m_out.load32NonNegative(storage, m_heaps.ArrayStorage_publicLength)),
|
|
unsure(doStoreCase), unsure(lengthUpdateCase));
|
|
|
|
m_out.appendTo(lengthUpdateCase, doStoreCase);
|
|
m_out.store32(
|
|
m_out.add(index, m_out.int32One),
|
|
storage, m_heaps.ArrayStorage_publicLength);
|
|
m_out.jump(doStoreCase);
|
|
}
|
|
|
|
m_out.appendTo(doStoreCase, continuation);
|
|
m_out.store64(value, elementPointer);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
return;
|
|
}
|
|
|
|
case Array::Int8Array:
|
|
case Array::Int16Array:
|
|
case Array::Int32Array:
|
|
case Array::Uint8Array:
|
|
case Array::Uint8ClampedArray:
|
|
case Array::Uint16Array:
|
|
case Array::Uint32Array:
|
|
case Array::Float32Array:
|
|
case Array::Float64Array: {
|
|
TypedArrayType type = arrayMode.typedArrayType();
|
|
|
|
ASSERT(isTypedView(type));
|
|
{
|
|
TypedPointer pointer = TypedPointer(
|
|
m_heaps.typedArrayProperties,
|
|
m_out.add(
|
|
storage,
|
|
m_out.shl(
|
|
m_out.zeroExt(index, pointerType()),
|
|
m_out.constIntPtr(logElementSize(type)))));
|
|
|
|
LValue valueToStore;
|
|
|
|
if (isInt(type)) {
|
|
LValue intValue = getIntTypedArrayStoreOperand(child3, isClamped(type));
|
|
|
|
valueToStore = intValue;
|
|
} else /* !isInt(type) */ {
|
|
LValue value = lowDouble(child3);
|
|
switch (type) {
|
|
case TypeFloat32:
|
|
valueToStore = m_out.doubleToFloat(value);
|
|
break;
|
|
case TypeFloat64:
|
|
valueToStore = value;
|
|
break;
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad typed array type");
|
|
}
|
|
}
|
|
|
|
if (arrayMode.isInBounds() || m_node->op() == PutByValAlias)
|
|
m_out.store(valueToStore, pointer, storeType(type));
|
|
else {
|
|
LBasicBlock isInBounds = m_out.newBlock();
|
|
LBasicBlock isOutOfBounds = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
m_out.aboveOrEqual(index, lowInt32(child5)),
|
|
unsure(isOutOfBounds), unsure(isInBounds));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isInBounds, isOutOfBounds);
|
|
m_out.store(valueToStore, pointer, storeType(type));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(isOutOfBounds, continuation);
|
|
speculateTypedArrayIsNotDetached(base);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
}
|
|
|
|
// We have to keep base alive since that keeps storage alive.
|
|
ensureStillAliveHere(base);
|
|
return;
|
|
}
|
|
}
|
|
|
|
case Array::AnyTypedArray:
|
|
case Array::String:
|
|
case Array::DirectArguments:
|
|
case Array::ForceExit:
|
|
case Array::Generic:
|
|
case Array::ScopedArguments:
|
|
case Array::SelectUsingArguments:
|
|
case Array::SelectUsingPredictions:
|
|
case Array::Undecided:
|
|
case Array::Unprofiled:
|
|
DFG_CRASH(m_graph, m_node, "Bad array type");
|
|
break;
|
|
}
|
|
}
|
|
|
|
void compilePutAccessorById()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue base = lowCell(m_node->child1());
|
|
LValue accessor = lowCell(m_node->child2());
|
|
auto uid = m_graph.identifiers()[m_node->identifierNumber()];
|
|
vmCall(
|
|
Void,
|
|
m_node->op() == PutGetterById ? operationPutGetterById : operationPutSetterById,
|
|
weakPointer(globalObject), base, m_out.constIntPtr(uid), m_out.constInt32(m_node->accessorAttributes()), accessor);
|
|
}
|
|
|
|
void compilePutGetterSetterById()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue base = lowCell(m_node->child1());
|
|
LValue getter = lowJSValue(m_node->child2());
|
|
LValue setter = lowJSValue(m_node->child3());
|
|
auto uid = m_graph.identifiers()[m_node->identifierNumber()];
|
|
vmCall(
|
|
Void, operationPutGetterSetter,
|
|
weakPointer(globalObject), base, m_out.constIntPtr(uid), m_out.constInt32(m_node->accessorAttributes()), getter, setter);
|
|
|
|
}
|
|
|
|
void compilePutAccessorByVal()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue base = lowCell(m_node->child1());
|
|
LValue subscript = lowJSValue(m_node->child2());
|
|
LValue accessor = lowCell(m_node->child3());
|
|
vmCall(
|
|
Void,
|
|
m_node->op() == PutGetterByVal ? operationPutGetterByVal : operationPutSetterByVal,
|
|
weakPointer(globalObject), base, subscript, m_out.constInt32(m_node->accessorAttributes()), accessor);
|
|
}
|
|
|
|
template<DelByKind kind, typename SubscriptKind>
|
|
void compileDelBy(LValue base, SubscriptKind subscriptValue)
|
|
{
|
|
PatchpointValue* patchpoint;
|
|
if constexpr (kind == DelByKind::Normal) {
|
|
patchpoint = m_out.patchpoint(Int64);
|
|
patchpoint->append(ConstrainedValue(base, ValueRep::SomeLateRegister));
|
|
} else {
|
|
patchpoint = m_out.patchpoint(Int64);
|
|
patchpoint->append(ConstrainedValue(base, ValueRep::SomeLateRegister));
|
|
patchpoint->append(ConstrainedValue(subscriptValue, ValueRep::SomeLateRegister));
|
|
}
|
|
patchpoint->append(m_notCellMask, ValueRep::lateReg(GPRInfo::notCellMaskRegister));
|
|
patchpoint->append(m_numberTag, ValueRep::lateReg(GPRInfo::numberTagRegister));
|
|
patchpoint->clobber(RegisterSet::macroScratchRegisters());
|
|
patchpoint->numGPScratchRegisters = 1;
|
|
|
|
RefPtr<PatchpointExceptionHandle> exceptionHandle =
|
|
preparePatchpointForExceptions(patchpoint);
|
|
|
|
State* state = &m_ftlState;
|
|
Node* node = m_node;
|
|
patchpoint->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
CallSiteIndex callSiteIndex =
|
|
state->jitCode->common.codeOrigins->addUniqueCallSiteIndex(node->origin.semantic);
|
|
|
|
Box<CCallHelpers::JumpList> exceptions =
|
|
exceptionHandle->scheduleExitCreation(params)->jumps(jit);
|
|
CCallHelpers::JumpList slowCases;
|
|
|
|
auto base = JSValueRegs(params[1].gpr());
|
|
auto returnGPR = params[0].gpr();
|
|
ASSERT(base.gpr() != returnGPR);
|
|
ASSERT(base.gpr() != params.gpScratch(0));
|
|
ASSERT(returnGPR != params.gpScratch(0));
|
|
|
|
if (node->child1().useKind() == UntypedUse)
|
|
slowCases.append(jit.branchIfNotCell(base));
|
|
|
|
constexpr auto optimizationFunction = [&] () {
|
|
if constexpr (kind == DelByKind::Normal)
|
|
return operationDeleteByIdOptimize;
|
|
else
|
|
return operationDeleteByValOptimize;
|
|
}();
|
|
|
|
const auto subscript = [&] {
|
|
if constexpr (kind == DelByKind::Normal)
|
|
return CCallHelpers::TrustedImmPtr(subscriptValue.rawBits());
|
|
else {
|
|
ASSERT(params.gpScratch(0) != params[2].gpr());
|
|
if (node->child2().useKind() == UntypedUse)
|
|
slowCases.append(jit.branchIfNotCell(JSValueRegs(params[2].gpr())));
|
|
return JSValueRegs(params[2].gpr());
|
|
}
|
|
}();
|
|
|
|
const auto generator = [&] {
|
|
if constexpr (kind == DelByKind::Normal) {
|
|
return Box<JITDelByIdGenerator>::create(
|
|
jit.codeBlock(), node->origin.semantic, callSiteIndex,
|
|
params.unavailableRegisters(), subscriptValue, base,
|
|
JSValueRegs(returnGPR), params.gpScratch(0));
|
|
} else {
|
|
return Box<JITDelByValGenerator>::create(
|
|
jit.codeBlock(), node->origin.semantic, callSiteIndex,
|
|
params.unavailableRegisters(), base,
|
|
subscript, JSValueRegs(returnGPR), params.gpScratch(0));
|
|
}
|
|
}();
|
|
|
|
generator->generateFastPath(jit);
|
|
slowCases.append(generator->slowPathJump());
|
|
CCallHelpers::Label done = jit.label();
|
|
|
|
params.addLatePath(
|
|
[=] (CCallHelpers& jit) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
slowCases.link(&jit);
|
|
CCallHelpers::Label slowPathBegin = jit.label();
|
|
CCallHelpers::Call slowPathCall = callOperation(
|
|
*state, params.unavailableRegisters(), jit, node->origin.semantic,
|
|
exceptions.get(), optimizationFunction, returnGPR,
|
|
jit.codeBlock()->globalObjectFor(node->origin.semantic),
|
|
CCallHelpers::TrustedImmPtr(generator->stubInfo()), base,
|
|
subscript, CCallHelpers::TrustedImm32(node->ecmaMode().value())).call();
|
|
jit.jump().linkTo(done, &jit);
|
|
|
|
generator->reportSlowPathCall(slowPathBegin, slowPathCall);
|
|
|
|
jit.addLinkTask(
|
|
[=] (LinkBuffer& linkBuffer) {
|
|
generator->finalize(linkBuffer, linkBuffer);
|
|
});
|
|
});
|
|
});
|
|
|
|
setBoolean(m_out.notZero64(patchpoint));
|
|
}
|
|
|
|
void compileDeleteById()
|
|
{
|
|
switch (m_node->child1().useKind()) {
|
|
case CellUse: {
|
|
LValue base = lowCell(m_node->child1());
|
|
compileDelBy<DelByKind::Normal>(base, m_node->cacheableIdentifier());
|
|
break;
|
|
}
|
|
|
|
case UntypedUse: {
|
|
// FIXME: We should use IC even if child1 is UntypedUse. In that case, we should emit write-barrier after tha fast path of IC.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=209397
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue base = lowJSValue(m_node->child1());
|
|
setBoolean(m_out.notZero64(vmCall(Int64, operationDeleteByIdGeneric, weakPointer(globalObject), m_out.intPtrZero, base, m_out.constIntPtr(m_node->cacheableIdentifier().rawBits()), m_out.constInt32(m_node->ecmaMode().value()))));
|
|
break;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
return;
|
|
}
|
|
}
|
|
|
|
void compileDeleteByVal()
|
|
{
|
|
switch (m_node->child1().useKind()) {
|
|
case CellUse: {
|
|
LValue base = lowCell(m_node->child1());
|
|
LValue subscript;
|
|
switch (m_node->child2().useKind()) {
|
|
case CellUse: {
|
|
subscript = lowCell(m_node->child2());
|
|
break;
|
|
}
|
|
|
|
case UntypedUse: {
|
|
subscript = lowJSValue(m_node->child2());
|
|
break;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
return;
|
|
}
|
|
compileDelBy<DelByKind::NormalByVal>(base, subscript);
|
|
return;
|
|
}
|
|
|
|
case UntypedUse: {
|
|
// FIXME: We should use IC even if child1 is UntypedUse. In that case, we should emit write-barrier after tha fast path of IC.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=209397
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue base = lowJSValue(m_node->child1());
|
|
LValue subscript = lowJSValue(m_node->child2());
|
|
setBoolean(m_out.notZero64(vmCall(Int64, operationDeleteByValGeneric, weakPointer(globalObject), m_out.intPtrZero, base, subscript, m_out.constInt32(m_node->ecmaMode().value()))));
|
|
return;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
return;
|
|
}
|
|
}
|
|
|
|
void compileArrayPush()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue base = lowCell(m_graph.varArgChild(m_node, 1));
|
|
LValue storage = lowStorage(m_graph.varArgChild(m_node, 0));
|
|
unsigned elementOffset = 2;
|
|
unsigned elementCount = m_node->numChildren() - elementOffset;
|
|
|
|
switch (m_node->arrayMode().type()) {
|
|
case Array::Int32:
|
|
case Array::Contiguous:
|
|
case Array::Double: {
|
|
IndexedAbstractHeap& heap = m_heaps.forArrayType(m_node->arrayMode().type());
|
|
|
|
if (elementCount == 1) {
|
|
LValue value;
|
|
Output::StoreType storeType;
|
|
|
|
Edge& element = m_graph.varArgChild(m_node, elementOffset);
|
|
speculate(element);
|
|
if (m_node->arrayMode().type() != Array::Double) {
|
|
value = lowJSValue(element, ManualOperandSpeculation);
|
|
storeType = Output::Store64;
|
|
} else {
|
|
value = lowDouble(element);
|
|
storeType = Output::StoreDouble;
|
|
}
|
|
|
|
LValue prevLength = m_out.load32(storage, m_heaps.Butterfly_publicLength);
|
|
|
|
LBasicBlock fastPath = m_out.newBlock();
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
m_out.aboveOrEqual(
|
|
prevLength, m_out.load32(storage, m_heaps.Butterfly_vectorLength)),
|
|
unsure(slowPath), unsure(fastPath));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(fastPath, slowPath);
|
|
m_out.store(
|
|
value, m_out.baseIndex(heap, storage, m_out.zeroExtPtr(prevLength)), storeType);
|
|
LValue newLength = m_out.add(prevLength, m_out.int32One);
|
|
m_out.store32(newLength, storage, m_heaps.Butterfly_publicLength);
|
|
|
|
ValueFromBlock fastResult = m_out.anchor(boxInt32(newLength));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
LValue result;
|
|
if (m_node->arrayMode().type() != Array::Double)
|
|
result = vmCall(Int64, operationArrayPush, weakPointer(globalObject), value, base);
|
|
else
|
|
result = vmCall(Int64, operationArrayPushDouble, weakPointer(globalObject), value, base);
|
|
ValueFromBlock slowResult = m_out.anchor(result);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, fastResult, slowResult));
|
|
return;
|
|
}
|
|
|
|
for (unsigned elementIndex = 0; elementIndex < elementCount; ++elementIndex) {
|
|
Edge element = m_graph.varArgChild(m_node, elementIndex + elementOffset);
|
|
speculate(element);
|
|
}
|
|
|
|
LValue prevLength = m_out.load32(storage, m_heaps.Butterfly_publicLength);
|
|
LValue newLength = m_out.add(prevLength, m_out.constInt32(elementCount));
|
|
|
|
LBasicBlock fastPath = m_out.newBlock();
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock setup = m_out.newBlock();
|
|
LBasicBlock slowCallPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue beyondVectorLength = m_out.above(newLength, m_out.load32(storage, m_heaps.Butterfly_vectorLength));
|
|
|
|
m_out.branch(beyondVectorLength, unsure(slowPath), unsure(fastPath));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(fastPath, slowPath);
|
|
m_out.store32(newLength, storage, m_heaps.Butterfly_publicLength);
|
|
ValueFromBlock fastBufferResult = m_out.anchor(m_out.baseIndex(storage, m_out.zeroExtPtr(prevLength), ScaleEight));
|
|
m_out.jump(setup);
|
|
|
|
m_out.appendTo(slowPath, setup);
|
|
size_t scratchSize = sizeof(EncodedJSValue) * elementCount;
|
|
static_assert(sizeof(EncodedJSValue) == sizeof(double), "");
|
|
ASSERT(scratchSize);
|
|
ScratchBuffer* scratchBuffer = vm().scratchBufferForSize(scratchSize);
|
|
m_out.storePtr(m_out.constIntPtr(scratchSize), m_out.absolute(scratchBuffer->addressOfActiveLength()));
|
|
ValueFromBlock slowBufferResult = m_out.anchor(m_out.constIntPtr(static_cast<EncodedJSValue*>(scratchBuffer->dataBuffer())));
|
|
m_out.jump(setup);
|
|
|
|
m_out.appendTo(setup, slowCallPath);
|
|
LValue buffer = m_out.phi(pointerType(), fastBufferResult, slowBufferResult);
|
|
for (unsigned elementIndex = 0; elementIndex < elementCount; ++elementIndex) {
|
|
Edge& element = m_graph.varArgChild(m_node, elementIndex + elementOffset);
|
|
|
|
LValue value;
|
|
Output::StoreType storeType;
|
|
if (m_node->arrayMode().type() != Array::Double) {
|
|
value = lowJSValue(element, ManualOperandSpeculation);
|
|
storeType = Output::Store64;
|
|
} else {
|
|
value = lowDouble(element);
|
|
storeType = Output::StoreDouble;
|
|
}
|
|
|
|
m_out.store(value, m_out.baseIndex(heap, buffer, m_out.constInt32(elementIndex), jsNumber(elementIndex)), storeType);
|
|
}
|
|
ValueFromBlock fastResult = m_out.anchor(boxInt32(newLength));
|
|
|
|
m_out.branch(beyondVectorLength, unsure(slowCallPath), unsure(continuation));
|
|
|
|
m_out.appendTo(slowCallPath, continuation);
|
|
auto* operation = &operationArrayPushMultiple;
|
|
if (m_node->arrayMode().type() == Array::Double)
|
|
operation = &operationArrayPushDoubleMultiple;
|
|
ValueFromBlock slowResult = m_out.anchor(vmCall(Int64, operation, weakPointer(globalObject), base, buffer, m_out.constInt32(elementCount)));
|
|
m_out.storePtr(m_out.constIntPtr(0), m_out.absolute(scratchBuffer->addressOfActiveLength()));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, fastResult, slowResult));
|
|
return;
|
|
}
|
|
|
|
case Array::ArrayStorage: {
|
|
// This ensures that the result of ArrayPush is Int32 in AI.
|
|
int32_t largestPositiveInt32Length = 0x7fffffff - elementCount;
|
|
|
|
LValue prevLength = m_out.load32(storage, m_heaps.ArrayStorage_publicLength);
|
|
// Refuse to handle bizarre lengths.
|
|
speculate(Uncountable, noValue(), nullptr, m_out.above(prevLength, m_out.constInt32(largestPositiveInt32Length)));
|
|
|
|
if (elementCount == 1) {
|
|
Edge& element = m_graph.varArgChild(m_node, elementOffset);
|
|
|
|
LValue value = lowJSValue(element);
|
|
|
|
LBasicBlock fastPath = m_out.newBlock();
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
m_out.aboveOrEqual(
|
|
prevLength, m_out.load32(storage, m_heaps.ArrayStorage_vectorLength)),
|
|
rarely(slowPath), usually(fastPath));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(fastPath, slowPath);
|
|
m_out.store64(
|
|
value, m_out.baseIndex(m_heaps.ArrayStorage_vector, storage, m_out.zeroExtPtr(prevLength)));
|
|
LValue newLength = m_out.add(prevLength, m_out.int32One);
|
|
m_out.store32(newLength, storage, m_heaps.ArrayStorage_publicLength);
|
|
m_out.store32(
|
|
m_out.add(m_out.load32(storage, m_heaps.ArrayStorage_numValuesInVector), m_out.int32One),
|
|
storage, m_heaps.ArrayStorage_numValuesInVector);
|
|
|
|
ValueFromBlock fastResult = m_out.anchor(boxInt32(newLength));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
ValueFromBlock slowResult = m_out.anchor(
|
|
vmCall(Int64, operationArrayPush, weakPointer(globalObject), value, base));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, fastResult, slowResult));
|
|
return;
|
|
}
|
|
|
|
LValue newLength = m_out.add(prevLength, m_out.constInt32(elementCount));
|
|
|
|
LBasicBlock fastPath = m_out.newBlock();
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock setup = m_out.newBlock();
|
|
LBasicBlock slowCallPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue beyondVectorLength = m_out.above(newLength, m_out.load32(storage, m_heaps.ArrayStorage_vectorLength));
|
|
|
|
m_out.branch(beyondVectorLength, rarely(slowPath), usually(fastPath));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(fastPath, slowPath);
|
|
m_out.store32(newLength, storage, m_heaps.ArrayStorage_publicLength);
|
|
m_out.store32(
|
|
m_out.add(m_out.load32(storage, m_heaps.ArrayStorage_numValuesInVector), m_out.constInt32(elementCount)),
|
|
storage, m_heaps.ArrayStorage_numValuesInVector);
|
|
ValueFromBlock fastBufferResult = m_out.anchor(m_out.baseIndex(storage, m_out.zeroExtPtr(prevLength), ScaleEight, ArrayStorage::vectorOffset()));
|
|
m_out.jump(setup);
|
|
|
|
m_out.appendTo(slowPath, setup);
|
|
size_t scratchSize = sizeof(EncodedJSValue) * elementCount;
|
|
ASSERT(scratchSize);
|
|
ScratchBuffer* scratchBuffer = vm().scratchBufferForSize(scratchSize);
|
|
m_out.storePtr(m_out.constIntPtr(scratchSize), m_out.absolute(scratchBuffer->addressOfActiveLength()));
|
|
ValueFromBlock slowBufferResult = m_out.anchor(m_out.constIntPtr(static_cast<EncodedJSValue*>(scratchBuffer->dataBuffer())));
|
|
m_out.jump(setup);
|
|
|
|
m_out.appendTo(setup, slowCallPath);
|
|
LValue buffer = m_out.phi(pointerType(), fastBufferResult, slowBufferResult);
|
|
for (unsigned elementIndex = 0; elementIndex < elementCount; ++elementIndex) {
|
|
Edge& element = m_graph.varArgChild(m_node, elementIndex + elementOffset);
|
|
|
|
LValue value = lowJSValue(element);
|
|
m_out.store64(value, m_out.baseIndex(m_heaps.ArrayStorage_vector.atAnyIndex(), buffer, m_out.constIntPtr(elementIndex), ScaleEight));
|
|
}
|
|
ValueFromBlock fastResult = m_out.anchor(boxInt32(newLength));
|
|
|
|
m_out.branch(beyondVectorLength, rarely(slowCallPath), usually(continuation));
|
|
|
|
m_out.appendTo(slowCallPath, continuation);
|
|
ValueFromBlock slowResult = m_out.anchor(vmCall(Int64, operationArrayPushMultiple, weakPointer(globalObject), base, buffer, m_out.constInt32(elementCount)));
|
|
m_out.storePtr(m_out.constIntPtr(0), m_out.absolute(scratchBuffer->addressOfActiveLength()));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, fastResult, slowResult));
|
|
return;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad array type");
|
|
return;
|
|
}
|
|
}
|
|
|
|
std::pair<LValue, LValue> populateSliceRange(LValue start, LValue end, LValue length)
|
|
{
|
|
// end can be nullptr.
|
|
ASSERT(start);
|
|
ASSERT(length);
|
|
|
|
auto pickIndex = [&] (LValue index) {
|
|
return m_out.select(m_out.greaterThanOrEqual(index, m_out.int32Zero),
|
|
m_out.select(m_out.above(index, length), length, index),
|
|
m_out.select(m_out.lessThan(m_out.add(length, index), m_out.int32Zero), m_out.int32Zero, m_out.add(length, index)));
|
|
};
|
|
|
|
LValue endBoundary = length;
|
|
if (end)
|
|
endBoundary = pickIndex(end);
|
|
LValue startIndex = pickIndex(start);
|
|
return std::make_pair(startIndex, endBoundary);
|
|
}
|
|
|
|
void compileArraySlice()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
|
|
LValue sourceArray = lowCell(m_graph.varArgChild(m_node, 0));
|
|
LValue sourceStorage = lowStorage(m_graph.varArgChild(m_node, m_node->numChildren() - 1));
|
|
LValue inputLength = m_out.load32(sourceStorage, m_heaps.Butterfly_publicLength);
|
|
|
|
LValue startIndex = nullptr;
|
|
LValue resultLength = nullptr;
|
|
if (m_node->numChildren() == 2) {
|
|
startIndex = m_out.constInt32(0);
|
|
resultLength = inputLength;
|
|
} else {
|
|
LValue start = lowInt32(m_graph.varArgChild(m_node, 1));
|
|
LValue end = nullptr;
|
|
if (m_node->numChildren() != 3)
|
|
end = lowInt32(m_graph.varArgChild(m_node, 2));
|
|
|
|
auto range = populateSliceRange(start, end, inputLength);
|
|
startIndex = range.first;
|
|
LValue endBoundary = range.second;
|
|
|
|
resultLength = m_out.select(m_out.belowOrEqual(startIndex, endBoundary),
|
|
m_out.sub(endBoundary, startIndex),
|
|
m_out.constInt32(0));
|
|
}
|
|
|
|
ArrayValues arrayResult;
|
|
{
|
|
LValue indexingType = m_out.load8ZeroExt32(sourceArray, m_heaps.JSCell_indexingTypeAndMisc);
|
|
// We can ignore the writability of the cell since we won't write to the source.
|
|
indexingType = m_out.bitAnd(indexingType, m_out.constInt32(AllWritableArrayTypesAndHistory));
|
|
// When we emit an ArraySlice, we dominate the use of the array by a CheckStructure
|
|
// to ensure the incoming array is one to be one of the original array structures
|
|
// with one of the following indexing shapes: Int32, Contiguous, Double.
|
|
LValue structure = m_out.select(
|
|
m_out.equal(indexingType, m_out.constInt32(ArrayWithInt32)),
|
|
weakStructure(m_graph.registerStructure(globalObject->arrayStructureForIndexingTypeDuringAllocation(ArrayWithInt32))),
|
|
m_out.select(m_out.equal(indexingType, m_out.constInt32(ArrayWithContiguous)),
|
|
weakStructure(m_graph.registerStructure(globalObject->arrayStructureForIndexingTypeDuringAllocation(ArrayWithContiguous))),
|
|
weakStructure(m_graph.registerStructure(globalObject->arrayStructureForIndexingTypeDuringAllocation(ArrayWithDouble)))));
|
|
arrayResult = allocateJSArray(resultLength, resultLength, structure, indexingType, false, false);
|
|
}
|
|
|
|
// Keep the sourceArray alive at least until after anything that can GC.
|
|
ensureStillAliveHere(sourceArray);
|
|
|
|
LBasicBlock loop = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
resultLength = m_out.zeroExtPtr(resultLength);
|
|
ValueFromBlock startLoadIndex = m_out.anchor(m_out.zeroExtPtr(startIndex));
|
|
ValueFromBlock startStoreIndex = m_out.anchor(m_out.constIntPtr(0));
|
|
|
|
m_out.branch(
|
|
m_out.below(m_out.constIntPtr(0), resultLength), unsure(loop), unsure(continuation));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(loop, continuation);
|
|
LValue storeIndex = m_out.phi(pointerType(), startStoreIndex);
|
|
LValue loadIndex = m_out.phi(pointerType(), startLoadIndex);
|
|
LValue value = m_out.load64(m_out.baseIndex(m_heaps.root, sourceStorage, loadIndex, ScaleEight));
|
|
m_out.store64(value, m_out.baseIndex(m_heaps.root, arrayResult.butterfly, storeIndex, ScaleEight));
|
|
LValue nextStoreIndex = m_out.add(storeIndex, m_out.constIntPtr(1));
|
|
m_out.addIncomingToPhi(storeIndex, m_out.anchor(nextStoreIndex));
|
|
m_out.addIncomingToPhi(loadIndex, m_out.anchor(m_out.add(loadIndex, m_out.constIntPtr(1))));
|
|
m_out.branch(
|
|
m_out.below(nextStoreIndex, resultLength), unsure(loop), unsure(continuation));
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
|
|
mutatorFence();
|
|
setJSValue(arrayResult.array);
|
|
}
|
|
|
|
void compileArrayIndexOf()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue base = lowCell(m_graph.varArgChild(m_node, 0));
|
|
LValue storage = lowStorage(m_node->numChildren() == 3 ? m_graph.varArgChild(m_node, 2) : m_graph.varArgChild(m_node, 3));
|
|
LValue length = m_out.load32(storage, m_heaps.Butterfly_publicLength);
|
|
|
|
LValue startIndex;
|
|
if (m_node->numChildren() == 4) {
|
|
startIndex = lowInt32(m_graph.varArgChild(m_node, 2));
|
|
startIndex = m_out.select(m_out.greaterThanOrEqual(startIndex, m_out.int32Zero),
|
|
m_out.select(m_out.above(startIndex, length), length, startIndex),
|
|
m_out.select(m_out.lessThan(m_out.add(length, startIndex), m_out.int32Zero), m_out.int32Zero, m_out.add(length, startIndex)));
|
|
} else
|
|
startIndex = m_out.int32Zero;
|
|
|
|
Edge& searchElementEdge = m_graph.varArgChild(m_node, 1);
|
|
switch (searchElementEdge.useKind()) {
|
|
case Int32Use:
|
|
case ObjectUse:
|
|
case SymbolUse:
|
|
case OtherUse:
|
|
case DoubleRepUse: {
|
|
LBasicBlock loopHeader = m_out.newBlock();
|
|
LBasicBlock loopBody = m_out.newBlock();
|
|
LBasicBlock loopNext = m_out.newBlock();
|
|
LBasicBlock notFound = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue searchElement;
|
|
switch (searchElementEdge.useKind()) {
|
|
case Int32Use:
|
|
ASSERT(m_node->arrayMode().type() == Array::Int32);
|
|
speculate(searchElementEdge);
|
|
searchElement = lowJSValue(searchElementEdge, ManualOperandSpeculation);
|
|
break;
|
|
case ObjectUse:
|
|
ASSERT(m_node->arrayMode().type() == Array::Contiguous);
|
|
searchElement = lowObject(searchElementEdge);
|
|
break;
|
|
case SymbolUse:
|
|
ASSERT(m_node->arrayMode().type() == Array::Contiguous);
|
|
searchElement = lowSymbol(searchElementEdge);
|
|
break;
|
|
case OtherUse:
|
|
ASSERT(m_node->arrayMode().type() == Array::Contiguous);
|
|
speculate(searchElementEdge);
|
|
searchElement = lowJSValue(searchElementEdge, ManualOperandSpeculation);
|
|
break;
|
|
case DoubleRepUse:
|
|
ASSERT(m_node->arrayMode().type() == Array::Double);
|
|
searchElement = lowDouble(searchElementEdge);
|
|
break;
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
break;
|
|
}
|
|
|
|
startIndex = m_out.zeroExtPtr(startIndex);
|
|
length = m_out.zeroExtPtr(length);
|
|
|
|
ValueFromBlock initialStartIndex = m_out.anchor(startIndex);
|
|
m_out.jump(loopHeader);
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(loopHeader, loopBody);
|
|
LValue index = m_out.phi(pointerType(), initialStartIndex);
|
|
m_out.branch(m_out.notEqual(index, length), unsure(loopBody), unsure(notFound));
|
|
|
|
m_out.appendTo(loopBody, loopNext);
|
|
ValueFromBlock foundResult = m_out.anchor(index);
|
|
switch (searchElementEdge.useKind()) {
|
|
case Int32Use: {
|
|
// Empty value is ignored because of JSValue::NumberTag.
|
|
LValue value = m_out.load64(m_out.baseIndex(m_heaps.indexedInt32Properties, storage, index));
|
|
m_out.branch(m_out.equal(value, searchElement), unsure(continuation), unsure(loopNext));
|
|
break;
|
|
}
|
|
case ObjectUse:
|
|
case SymbolUse:
|
|
case OtherUse: {
|
|
// Empty value never matches against non-empty JS values.
|
|
LValue value = m_out.load64(m_out.baseIndex(m_heaps.indexedContiguousProperties, storage, index));
|
|
m_out.branch(m_out.equal(value, searchElement), unsure(continuation), unsure(loopNext));
|
|
break;
|
|
}
|
|
case DoubleRepUse: {
|
|
// Empty value is ignored because of NaN.
|
|
LValue value = m_out.loadDouble(m_out.baseIndex(m_heaps.indexedDoubleProperties, storage, index));
|
|
m_out.branch(m_out.doubleEqual(value, searchElement), unsure(continuation), unsure(loopNext));
|
|
break;
|
|
}
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
break;
|
|
}
|
|
|
|
m_out.appendTo(loopNext, notFound);
|
|
LValue nextIndex = m_out.add(index, m_out.intPtrOne);
|
|
m_out.addIncomingToPhi(index, m_out.anchor(nextIndex));
|
|
m_out.jump(loopHeader);
|
|
|
|
m_out.appendTo(notFound, continuation);
|
|
ValueFromBlock notFoundResult = m_out.anchor(m_out.constIntPtr(-1));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
// We have to keep base alive since that keeps content of storage alive.
|
|
ensureStillAliveHere(base);
|
|
setInt32(m_out.castToInt32(m_out.phi(pointerType(), notFoundResult, foundResult)));
|
|
return;
|
|
}
|
|
|
|
case StringUse:
|
|
ASSERT(m_node->arrayMode().type() == Array::Contiguous);
|
|
// We have to keep base alive since that keeps storage alive.
|
|
ensureStillAliveHere(base);
|
|
setInt32(m_out.castToInt32(vmCall(Int64, operationArrayIndexOfString, weakPointer(globalObject), storage, lowString(searchElementEdge), startIndex)));
|
|
return;
|
|
|
|
case UntypedUse:
|
|
switch (m_node->arrayMode().type()) {
|
|
case Array::Double:
|
|
setInt32(m_out.castToInt32(vmCall(Int64, operationArrayIndexOfValueDouble, weakPointer(globalObject), storage, lowJSValue(searchElementEdge), startIndex)));
|
|
return;
|
|
case Array::Contiguous:
|
|
// We have to keep base alive since that keeps content of storage alive.
|
|
ensureStillAliveHere(base);
|
|
FALLTHROUGH;
|
|
case Array::Int32:
|
|
setInt32(m_out.castToInt32(vmCall(Int64, operationArrayIndexOfValueInt32OrContiguous, weakPointer(globalObject), storage, lowJSValue(searchElementEdge), startIndex)));
|
|
return;
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
return;
|
|
}
|
|
return;
|
|
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
return;
|
|
}
|
|
}
|
|
|
|
|
|
void compileArrayPop()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue base = lowCell(m_node->child1());
|
|
LValue storage = lowStorage(m_node->child2());
|
|
|
|
switch (m_node->arrayMode().type()) {
|
|
case Array::Int32:
|
|
case Array::Double:
|
|
case Array::Contiguous: {
|
|
IndexedAbstractHeap& heap = m_heaps.forArrayType(m_node->arrayMode().type());
|
|
|
|
LBasicBlock fastCase = m_out.newBlock();
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue prevLength = m_out.load32(storage, m_heaps.Butterfly_publicLength);
|
|
|
|
Vector<ValueFromBlock, 3> results;
|
|
results.append(m_out.anchor(m_out.constInt64(JSValue::encode(jsUndefined()))));
|
|
m_out.branch(
|
|
m_out.isZero32(prevLength), rarely(continuation), usually(fastCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(fastCase, slowCase);
|
|
LValue newLength = m_out.sub(prevLength, m_out.int32One);
|
|
m_out.store32(newLength, storage, m_heaps.Butterfly_publicLength);
|
|
TypedPointer pointer = m_out.baseIndex(heap, storage, m_out.zeroExtPtr(newLength));
|
|
if (m_node->arrayMode().type() != Array::Double) {
|
|
LValue result = m_out.load64(pointer);
|
|
// We have to keep base alive to keep content in storage alive.
|
|
if (m_node->arrayMode().type() == Array::Contiguous)
|
|
ensureStillAliveHere(base);
|
|
m_out.store64(m_out.int64Zero, pointer);
|
|
results.append(m_out.anchor(result));
|
|
m_out.branch(
|
|
m_out.notZero64(result), usually(continuation), rarely(slowCase));
|
|
} else {
|
|
LValue result = m_out.loadDouble(pointer);
|
|
m_out.store64(m_out.constInt64(bitwise_cast<int64_t>(PNaN)), pointer);
|
|
results.append(m_out.anchor(boxDouble(result)));
|
|
m_out.branch(
|
|
m_out.doubleEqual(result, result),
|
|
usually(continuation), rarely(slowCase));
|
|
}
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
results.append(m_out.anchor(vmCall(
|
|
Int64, operationArrayPopAndRecoverLength, weakPointer(globalObject), base)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, results));
|
|
return;
|
|
}
|
|
|
|
case Array::ArrayStorage: {
|
|
LBasicBlock vectorLengthCheckCase = m_out.newBlock();
|
|
LBasicBlock popCheckCase = m_out.newBlock();
|
|
LBasicBlock fastCase = m_out.newBlock();
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue prevLength = m_out.load32(storage, m_heaps.ArrayStorage_publicLength);
|
|
|
|
Vector<ValueFromBlock, 3> results;
|
|
results.append(m_out.anchor(m_out.constInt64(JSValue::encode(jsUndefined()))));
|
|
m_out.branch(
|
|
m_out.isZero32(prevLength), rarely(continuation), usually(vectorLengthCheckCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(vectorLengthCheckCase, popCheckCase);
|
|
LValue newLength = m_out.sub(prevLength, m_out.int32One);
|
|
m_out.branch(
|
|
m_out.aboveOrEqual(newLength, m_out.load32(storage, m_heaps.ArrayStorage_vectorLength)), rarely(slowCase), usually(popCheckCase));
|
|
|
|
m_out.appendTo(popCheckCase, fastCase);
|
|
TypedPointer pointer = m_out.baseIndex(m_heaps.ArrayStorage_vector, storage, m_out.zeroExtPtr(newLength));
|
|
LValue result = m_out.load64(pointer);
|
|
// We have to keep base alive to keep content in storage alive.
|
|
ensureStillAliveHere(base);
|
|
m_out.branch(m_out.notZero64(result), usually(fastCase), rarely(slowCase));
|
|
|
|
m_out.appendTo(fastCase, slowCase);
|
|
m_out.store32(newLength, storage, m_heaps.ArrayStorage_publicLength);
|
|
m_out.store64(m_out.int64Zero, pointer);
|
|
m_out.store32(
|
|
m_out.sub(m_out.load32(storage, m_heaps.ArrayStorage_numValuesInVector), m_out.int32One),
|
|
storage, m_heaps.ArrayStorage_numValuesInVector);
|
|
results.append(m_out.anchor(result));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
results.append(m_out.anchor(vmCall(Int64, operationArrayPop, weakPointer(globalObject), base)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, results));
|
|
return;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad array type");
|
|
return;
|
|
}
|
|
}
|
|
|
|
void compilePushWithScope()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue parentScope = lowCell(m_node->child1());
|
|
auto objectEdge = m_node->child2();
|
|
if (objectEdge.useKind() == ObjectUse) {
|
|
LValue object = lowNonNullObject(objectEdge);
|
|
LValue result = vmCall(Int64, operationPushWithScopeObject, weakPointer(globalObject), parentScope, object);
|
|
setJSValue(result);
|
|
} else {
|
|
ASSERT(objectEdge.useKind() == UntypedUse);
|
|
LValue object = lowJSValue(m_node->child2());
|
|
LValue result = vmCall(Int64, operationPushWithScope, weakPointer(globalObject), parentScope, object);
|
|
setJSValue(result);
|
|
}
|
|
}
|
|
|
|
void compileCreateActivation()
|
|
{
|
|
LValue scope = lowCell(m_node->child1());
|
|
SymbolTable* table = m_node->castOperand<SymbolTable*>();
|
|
RegisteredStructure structure = m_graph.registerStructure(m_graph.globalObjectFor(m_origin.semantic)->activationStructure());
|
|
JSValue initializationValue = m_node->initializationValueForActivation();
|
|
ASSERT(initializationValue.isUndefined() || initializationValue == jsTDZValue());
|
|
if (table->singleton().isStillValid()) {
|
|
LValue callResult = vmCall(
|
|
Int64,
|
|
operationCreateActivationDirect, m_vmValue, weakStructure(structure),
|
|
scope, weakPointer(table), m_out.constInt64(JSValue::encode(initializationValue)));
|
|
setJSValue(callResult);
|
|
return;
|
|
}
|
|
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(slowPath);
|
|
|
|
LValue fastObject = allocateObject<JSLexicalEnvironment>(
|
|
JSLexicalEnvironment::allocationSize(table), structure, m_out.intPtrZero, slowPath);
|
|
|
|
// We don't need memory barriers since we just fast-created the activation, so the
|
|
// activation must be young.
|
|
m_out.storePtr(scope, fastObject, m_heaps.JSScope_next);
|
|
m_out.storePtr(weakPointer(table), fastObject, m_heaps.JSSymbolTableObject_symbolTable);
|
|
|
|
for (unsigned i = 0; i < table->scopeSize(); ++i) {
|
|
m_out.store64(
|
|
m_out.constInt64(JSValue::encode(initializationValue)),
|
|
fastObject, m_heaps.JSLexicalEnvironment_variables[i]);
|
|
}
|
|
|
|
mutatorFence();
|
|
|
|
ValueFromBlock fastResult = m_out.anchor(fastObject);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
VM& vm = this->vm();
|
|
LValue callResult = lazySlowPath(
|
|
[=, &vm] (const Vector<Location>& locations) -> RefPtr<LazySlowPath::Generator> {
|
|
return createLazyCallGenerator(vm,
|
|
operationCreateActivationDirect, locations[0].directGPR(), &vm,
|
|
CCallHelpers::TrustedImmPtr(structure.get()), locations[1].directGPR(),
|
|
CCallHelpers::TrustedImmPtr(table),
|
|
CCallHelpers::TrustedImm64(JSValue::encode(initializationValue)));
|
|
},
|
|
scope);
|
|
ValueFromBlock slowResult = m_out.anchor(callResult);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(pointerType(), fastResult, slowResult));
|
|
}
|
|
|
|
void compileNewFunction()
|
|
{
|
|
ASSERT(m_node->op() == NewFunction || m_node->op() == NewGeneratorFunction || m_node->op() == NewAsyncGeneratorFunction || m_node->op() == NewAsyncFunction);
|
|
bool isGeneratorFunction = m_node->op() == NewGeneratorFunction;
|
|
bool isAsyncFunction = m_node->op() == NewAsyncFunction;
|
|
bool isAsyncGeneratorFunction = m_node->op() == NewAsyncGeneratorFunction;
|
|
|
|
LValue scope = lowCell(m_node->child1());
|
|
|
|
FunctionExecutable* executable = m_node->castOperand<FunctionExecutable*>();
|
|
if (executable->singleton().isStillValid()) {
|
|
LValue callResult =
|
|
isGeneratorFunction ? vmCall(Int64, operationNewGeneratorFunction, m_vmValue, scope, weakPointer(executable)) :
|
|
isAsyncFunction ? vmCall(Int64, operationNewAsyncFunction, m_vmValue, scope, weakPointer(executable)) :
|
|
isAsyncGeneratorFunction ? vmCall(Int64, operationNewAsyncGeneratorFunction, m_vmValue, scope, weakPointer(executable)) :
|
|
vmCall(Int64, operationNewFunction, m_vmValue, scope, weakPointer(executable));
|
|
setJSValue(callResult);
|
|
return;
|
|
}
|
|
|
|
RegisteredStructure structure = m_graph.registerStructure(
|
|
[&] () {
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
switch (m_node->op()) {
|
|
case NewGeneratorFunction:
|
|
return globalObject->generatorFunctionStructure();
|
|
case NewAsyncFunction:
|
|
return globalObject->asyncFunctionStructure();
|
|
case NewAsyncGeneratorFunction:
|
|
return globalObject->asyncGeneratorFunctionStructure();
|
|
case NewFunction:
|
|
return JSFunction::selectStructureForNewFuncExp(globalObject, m_node->castOperand<FunctionExecutable*>());
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
}
|
|
}());
|
|
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(slowPath);
|
|
|
|
LValue fastObject =
|
|
isGeneratorFunction ? allocateObject<JSGeneratorFunction>(structure, m_out.intPtrZero, slowPath) :
|
|
isAsyncFunction ? allocateObject<JSAsyncFunction>(structure, m_out.intPtrZero, slowPath) :
|
|
isAsyncGeneratorFunction ? allocateObject<JSAsyncGeneratorFunction>(structure, m_out.intPtrZero, slowPath) :
|
|
allocateObject<JSFunction>(structure, m_out.intPtrZero, slowPath);
|
|
|
|
|
|
// We don't need memory barriers since we just fast-created the function, so it
|
|
// must be young.
|
|
m_out.storePtr(scope, fastObject, m_heaps.JSFunction_scope);
|
|
m_out.storePtr(weakPointer(executable), fastObject, m_heaps.JSFunction_executableOrRareData);
|
|
mutatorFence();
|
|
|
|
ValueFromBlock fastResult = m_out.anchor(fastObject);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
|
|
Vector<LValue> slowPathArguments;
|
|
slowPathArguments.append(scope);
|
|
VM& vm = this->vm();
|
|
LValue callResult = lazySlowPath(
|
|
[=, &vm] (const Vector<Location>& locations) -> RefPtr<LazySlowPath::Generator> {
|
|
auto* operation = operationNewFunctionWithInvalidatedReallocationWatchpoint;
|
|
if (isGeneratorFunction)
|
|
operation = operationNewGeneratorFunctionWithInvalidatedReallocationWatchpoint;
|
|
else if (isAsyncFunction)
|
|
operation = operationNewAsyncFunctionWithInvalidatedReallocationWatchpoint;
|
|
else if (isAsyncGeneratorFunction)
|
|
operation = operationNewAsyncGeneratorFunctionWithInvalidatedReallocationWatchpoint;
|
|
|
|
return createLazyCallGenerator(vm, operation,
|
|
locations[0].directGPR(), &vm, locations[1].directGPR(),
|
|
CCallHelpers::TrustedImmPtr(executable));
|
|
},
|
|
slowPathArguments);
|
|
ValueFromBlock slowResult = m_out.anchor(callResult);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(pointerType(), fastResult, slowResult));
|
|
}
|
|
|
|
void compileCreateDirectArguments()
|
|
{
|
|
// FIXME: A more effective way of dealing with the argument count and callee is to have
|
|
// them be explicit arguments to this node.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=142207
|
|
|
|
RegisteredStructure structure =
|
|
m_graph.registerStructure(m_graph.globalObjectFor(m_origin.semantic)->directArgumentsStructure());
|
|
|
|
unsigned minCapacity = m_graph.baselineCodeBlockFor(m_origin.semantic)->numParameters() - 1;
|
|
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(slowPath);
|
|
|
|
ArgumentsLength length = getArgumentsLength();
|
|
|
|
LValue fastObject;
|
|
if (length.isKnown) {
|
|
fastObject = allocateObject<DirectArguments>(
|
|
DirectArguments::allocationSize(std::max(length.known, minCapacity)), structure,
|
|
m_out.intPtrZero, slowPath);
|
|
} else {
|
|
LValue size = m_out.add(
|
|
m_out.shl(length.value, m_out.constInt32(3)),
|
|
m_out.constInt32(DirectArguments::storageOffset()));
|
|
|
|
size = m_out.select(
|
|
m_out.aboveOrEqual(length.value, m_out.constInt32(minCapacity)),
|
|
size, m_out.constInt32(DirectArguments::allocationSize(minCapacity)));
|
|
|
|
fastObject = allocateVariableSizedObject<DirectArguments>(
|
|
m_out.zeroExtPtr(size), structure, m_out.intPtrZero, slowPath);
|
|
}
|
|
|
|
m_out.store32(length.value, fastObject, m_heaps.DirectArguments_length);
|
|
m_out.store32(m_out.constInt32(minCapacity), fastObject, m_heaps.DirectArguments_minCapacity);
|
|
m_out.storePtr(m_out.intPtrZero, fastObject, m_heaps.DirectArguments_mappedArguments);
|
|
m_out.storePtr(m_out.intPtrZero, fastObject, m_heaps.DirectArguments_modifiedArgumentsDescriptor);
|
|
|
|
ValueFromBlock fastResult = m_out.anchor(fastObject);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
VM& vm = this->vm();
|
|
LValue callResult = lazySlowPath(
|
|
[=, &vm] (const Vector<Location>& locations) -> RefPtr<LazySlowPath::Generator> {
|
|
return createLazyCallGenerator(vm,
|
|
operationCreateDirectArguments, locations[0].directGPR(), &vm,
|
|
CCallHelpers::TrustedImmPtr(structure.get()), locations[1].directGPR(),
|
|
CCallHelpers::TrustedImm32(minCapacity));
|
|
}, length.value);
|
|
ValueFromBlock slowResult = m_out.anchor(callResult);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
LValue result = m_out.phi(pointerType(), fastResult, slowResult);
|
|
|
|
m_out.storePtr(getCurrentCallee(), result, m_heaps.DirectArguments_callee);
|
|
|
|
if (length.isKnown) {
|
|
VirtualRegister start = AssemblyHelpers::argumentsStart(m_origin.semantic);
|
|
for (unsigned i = 0; i < std::max(length.known, minCapacity); ++i) {
|
|
m_out.store64(
|
|
m_out.load64(addressFor(start + i)),
|
|
result, m_heaps.DirectArguments_storage[i]);
|
|
}
|
|
} else {
|
|
LValue stackBase = getArgumentsStart();
|
|
|
|
LBasicBlock loop = m_out.newBlock();
|
|
LBasicBlock end = m_out.newBlock();
|
|
|
|
ValueFromBlock originalLength;
|
|
if (minCapacity) {
|
|
LValue capacity = m_out.select(
|
|
m_out.aboveOrEqual(length.value, m_out.constInt32(minCapacity)),
|
|
length.value,
|
|
m_out.constInt32(minCapacity));
|
|
LValue originalLengthValue = m_out.zeroExtPtr(capacity);
|
|
originalLength = m_out.anchor(originalLengthValue);
|
|
m_out.jump(loop);
|
|
} else {
|
|
LValue originalLengthValue = m_out.zeroExtPtr(length.value);
|
|
originalLength = m_out.anchor(originalLengthValue);
|
|
m_out.branch(m_out.isNull(originalLengthValue), unsure(end), unsure(loop));
|
|
}
|
|
|
|
lastNext = m_out.appendTo(loop, end);
|
|
LValue previousIndex = m_out.phi(pointerType(), originalLength);
|
|
LValue index = m_out.sub(previousIndex, m_out.intPtrOne);
|
|
m_out.store64(
|
|
m_out.load64(m_out.baseIndex(m_heaps.variables, stackBase, index)),
|
|
m_out.baseIndex(m_heaps.DirectArguments_storage, result, index));
|
|
ValueFromBlock nextIndex = m_out.anchor(index);
|
|
m_out.addIncomingToPhi(previousIndex, nextIndex);
|
|
m_out.branch(m_out.isNull(index), unsure(end), unsure(loop));
|
|
|
|
m_out.appendTo(end, lastNext);
|
|
}
|
|
|
|
mutatorFence();
|
|
|
|
setJSValue(result);
|
|
}
|
|
|
|
void compileCreateScopedArguments()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue scope = lowCell(m_node->child1());
|
|
|
|
LValue result = vmCall(
|
|
Int64, operationCreateScopedArguments, weakPointer(globalObject),
|
|
weakPointer(
|
|
m_graph.globalObjectFor(m_origin.semantic)->scopedArgumentsStructure()),
|
|
getArgumentsStart(), getArgumentsLength().value, getCurrentCallee(), scope);
|
|
|
|
setJSValue(result);
|
|
}
|
|
|
|
void compileCreateClonedArguments()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue result = vmCall(
|
|
Int64, operationCreateClonedArguments, weakPointer(globalObject),
|
|
weakPointer(
|
|
m_graph.globalObjectFor(m_origin.semantic)->clonedArgumentsStructure()),
|
|
getArgumentsStart(), getArgumentsLength().value, getCurrentCallee());
|
|
|
|
setJSValue(result);
|
|
}
|
|
|
|
void compileCreateArgumentsButterfly()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue result = vmCall(Int64, operationCreateArgumentsButterfly, weakPointer(globalObject), getArgumentsStart(), getArgumentsLength().value);
|
|
setJSValue(result);
|
|
}
|
|
|
|
void compileCreateRest()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
if (m_graph.isWatchingHavingABadTimeWatchpoint(m_node)) {
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
LValue arrayLength = lowInt32(m_node->child1());
|
|
LBasicBlock loopStart = m_out.newBlock();
|
|
RegisteredStructure structure = m_graph.registerStructure(globalObject->originalRestParameterStructure());
|
|
ArrayValues arrayValues = allocateUninitializedContiguousJSArray(arrayLength, structure);
|
|
LValue array = arrayValues.array;
|
|
LValue butterfly = arrayValues.butterfly;
|
|
ValueFromBlock startLength = m_out.anchor(arrayLength);
|
|
LValue argumentRegion = m_out.add(getArgumentsStart(), m_out.constInt64(sizeof(Register) * m_node->numberOfArgumentsToSkip()));
|
|
m_out.branch(m_out.equal(arrayLength, m_out.constInt32(0)),
|
|
unsure(continuation), unsure(loopStart));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(loopStart, continuation);
|
|
LValue phiOffset = m_out.phi(Int32, startLength);
|
|
LValue currentOffset = m_out.sub(phiOffset, m_out.int32One);
|
|
m_out.addIncomingToPhi(phiOffset, m_out.anchor(currentOffset));
|
|
LValue loadedValue = m_out.load64(m_out.baseIndex(m_heaps.variables, argumentRegion, m_out.zeroExtPtr(currentOffset)));
|
|
IndexedAbstractHeap& heap = m_heaps.indexedContiguousProperties;
|
|
m_out.store64(loadedValue, m_out.baseIndex(heap, butterfly, m_out.zeroExtPtr(currentOffset)));
|
|
m_out.branch(m_out.equal(currentOffset, m_out.constInt32(0)), unsure(continuation), unsure(loopStart));
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
mutatorFence();
|
|
setJSValue(array);
|
|
return;
|
|
}
|
|
|
|
LValue arrayLength = lowInt32(m_node->child1());
|
|
LValue argumentStart = getArgumentsStart();
|
|
LValue numberOfArgumentsToSkip = m_out.constInt32(m_node->numberOfArgumentsToSkip());
|
|
setJSValue(vmCall(
|
|
Int64, operationCreateRest, weakPointer(globalObject), argumentStart, numberOfArgumentsToSkip, arrayLength));
|
|
}
|
|
|
|
void compileGetRestLength()
|
|
{
|
|
LBasicBlock nonZeroLength = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
ValueFromBlock zeroLengthResult = m_out.anchor(m_out.constInt32(0));
|
|
|
|
LValue numberOfArgumentsToSkip = m_out.constInt32(m_node->numberOfArgumentsToSkip());
|
|
LValue argumentsLength = getArgumentsLength().value;
|
|
m_out.branch(m_out.above(argumentsLength, numberOfArgumentsToSkip),
|
|
unsure(nonZeroLength), unsure(continuation));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(nonZeroLength, continuation);
|
|
ValueFromBlock nonZeroLengthResult = m_out.anchor(m_out.sub(argumentsLength, numberOfArgumentsToSkip));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setInt32(m_out.phi(Int32, zeroLengthResult, nonZeroLengthResult));
|
|
}
|
|
|
|
void compileObjectKeysOrObjectGetOwnPropertyNames()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
NodeType op = m_node->op();
|
|
switch (m_node->child1().useKind()) {
|
|
case ObjectUse: {
|
|
if (m_graph.isWatchingHavingABadTimeWatchpoint(m_node)) {
|
|
LBasicBlock notNullCase = m_out.newBlock();
|
|
LBasicBlock rareDataCase = m_out.newBlock();
|
|
LBasicBlock useCacheCase = m_out.newBlock();
|
|
LBasicBlock slowButArrayBufferCase = m_out.newBlock();
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue object = lowObject(m_node->child1());
|
|
LValue structure = loadStructure(object);
|
|
LValue previousOrRareData = m_out.loadPtr(structure, m_heaps.Structure_previousOrRareData);
|
|
m_out.branch(m_out.notNull(previousOrRareData), unsure(notNullCase), unsure(slowCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(notNullCase, rareDataCase);
|
|
m_out.branch(
|
|
m_out.notEqual(m_out.load32(previousOrRareData, m_heaps.JSCell_structureID), m_out.constInt32(m_graph.m_vm.structureStructure->structureID())),
|
|
unsure(rareDataCase), unsure(slowCase));
|
|
|
|
m_out.appendTo(rareDataCase, useCacheCase);
|
|
ASSERT(bitwise_cast<uintptr_t>(StructureRareData::cachedPropertyNamesSentinel()) == 1);
|
|
LValue cached = m_out.loadPtr(previousOrRareData, op == ObjectKeys ? m_heaps.StructureRareData_cachedKeys : m_heaps.StructureRareData_cachedGetOwnPropertyNames);
|
|
m_out.branch(m_out.belowOrEqual(cached, m_out.constIntPtr(bitwise_cast<void*>(StructureRareData::cachedPropertyNamesSentinel()))), unsure(slowCase), unsure(useCacheCase));
|
|
|
|
m_out.appendTo(useCacheCase, slowButArrayBufferCase);
|
|
RegisteredStructure arrayStructure = m_graph.registerStructure(globalObject->arrayStructureForIndexingTypeDuringAllocation(CopyOnWriteArrayWithContiguous));
|
|
LValue fastArray = allocateObject<JSArray>(arrayStructure, m_out.addPtr(cached, JSImmutableButterfly::offsetOfData()), slowButArrayBufferCase);
|
|
ValueFromBlock fastResult = m_out.anchor(fastArray);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowButArrayBufferCase, slowCase);
|
|
LValue slowArray = vmCall(Int64, operationNewArrayBuffer, m_vmValue, weakStructure(arrayStructure), cached);
|
|
ValueFromBlock slowButArrayBufferResult = m_out.anchor(slowArray);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
VM& vm = this->vm();
|
|
LValue slowResultValue = lazySlowPath(
|
|
[=, &vm] (const Vector<Location>& locations) -> RefPtr<LazySlowPath::Generator> {
|
|
return createLazyCallGenerator(vm,
|
|
op == ObjectKeys ? operationObjectKeysObject : operationObjectGetOwnPropertyNamesObject, locations[0].directGPR(), globalObject, locations[1].directGPR());
|
|
},
|
|
object);
|
|
ValueFromBlock slowResult = m_out.anchor(slowResultValue);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(pointerType(), fastResult, slowButArrayBufferResult, slowResult));
|
|
break;
|
|
}
|
|
setJSValue(vmCall(Int64, op == ObjectKeys ? operationObjectKeysObject : operationObjectGetOwnPropertyNamesObject, weakPointer(globalObject), lowObject(m_node->child1())));
|
|
break;
|
|
}
|
|
case UntypedUse:
|
|
setJSValue(vmCall(Int64, op == ObjectKeys ? operationObjectKeys : operationObjectGetOwnPropertyNames, weakPointer(globalObject), lowJSValue(m_node->child1())));
|
|
break;
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
break;
|
|
}
|
|
}
|
|
|
|
void compileObjectCreate()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
switch (m_node->child1().useKind()) {
|
|
case ObjectUse:
|
|
setJSValue(vmCall(Int64, operationObjectCreateObject, weakPointer(globalObject), lowObject(m_node->child1())));
|
|
break;
|
|
case UntypedUse:
|
|
setJSValue(vmCall(Int64, operationObjectCreate, weakPointer(globalObject), lowJSValue(m_node->child1())));
|
|
break;
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
break;
|
|
}
|
|
}
|
|
|
|
void compileNewObject()
|
|
{
|
|
setJSValue(allocateObject(m_node->structure()));
|
|
mutatorFence();
|
|
}
|
|
|
|
template<typename JSClass, typename Operation>
|
|
void compileNewInternalFieldObjectImpl(Operation operation)
|
|
{
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(slowCase);
|
|
|
|
LValue object = allocateObject<JSClass>(m_node->structure(), m_out.intPtrZero, slowCase);
|
|
auto initialValues = JSClass::initialValues();
|
|
static_assert(initialValues.size() == JSClass::numberOfInternalFields);
|
|
for (unsigned index = 0; index < initialValues.size(); ++index)
|
|
m_out.store64(m_out.constInt64(JSValue::encode(initialValues[index])), object, m_heaps.JSInternalFieldObjectImpl_internalFields[index]);
|
|
mutatorFence();
|
|
ValueFromBlock fastResult = m_out.anchor(object);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
ValueFromBlock slowResult = m_out.anchor(vmCall(pointerType(), operation, m_vmValue, frozenPointer(m_graph.freezeStrong(m_node->structure().get()))));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(pointerType(), fastResult, slowResult));
|
|
}
|
|
|
|
void compileNewGenerator()
|
|
{
|
|
compileNewInternalFieldObjectImpl<JSGenerator>(operationNewGenerator);
|
|
}
|
|
|
|
void compileNewAsyncGenerator()
|
|
{
|
|
compileNewInternalFieldObjectImpl<JSAsyncGenerator>(operationNewAsyncGenerator);
|
|
}
|
|
|
|
void compileNewInternalFieldObject()
|
|
{
|
|
switch (m_node->structure()->typeInfo().type()) {
|
|
case JSArrayIteratorType:
|
|
compileNewInternalFieldObjectImpl<JSArrayIterator>(operationNewArrayIterator);
|
|
break;
|
|
case JSMapIteratorType:
|
|
compileNewInternalFieldObjectImpl<JSMapIterator>(operationNewMapIterator);
|
|
break;
|
|
case JSSetIteratorType:
|
|
compileNewInternalFieldObjectImpl<JSSetIterator>(operationNewSetIterator);
|
|
break;
|
|
case JSPromiseType:
|
|
if (m_node->structure()->classInfo() == JSInternalPromise::info())
|
|
compileNewInternalFieldObjectImpl<JSInternalPromise>(operationNewInternalPromise);
|
|
else {
|
|
ASSERT(m_node->structure()->classInfo() == JSPromise::info());
|
|
compileNewInternalFieldObjectImpl<JSPromise>(operationNewPromise);
|
|
}
|
|
break;
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad structure");
|
|
}
|
|
}
|
|
|
|
void compileNewStringObject()
|
|
{
|
|
// FIXME: We should handle this as JSInternalFieldObject allocation.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=209453
|
|
RegisteredStructure structure = m_node->structure();
|
|
LValue string = lowString(m_node->child1());
|
|
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(slowCase);
|
|
|
|
LValue fastResultValue = allocateObject<StringObject>(structure, m_out.intPtrZero, slowCase);
|
|
m_out.store64(string, fastResultValue, m_heaps.JSWrapperObject_internalValue);
|
|
mutatorFence();
|
|
ValueFromBlock fastResult = m_out.anchor(fastResultValue);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
VM& vm = this->vm();
|
|
LValue slowResultValue = lazySlowPath(
|
|
[=, &vm] (const Vector<Location>& locations) -> RefPtr<LazySlowPath::Generator> {
|
|
return createLazyCallGenerator(vm,
|
|
operationNewStringObject, locations[0].directGPR(), &vm, locations[1].directGPR(),
|
|
CCallHelpers::TrustedImmPtr(structure.get()));
|
|
},
|
|
string);
|
|
ValueFromBlock slowResult = m_out.anchor(slowResultValue);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(pointerType(), fastResult, slowResult));
|
|
}
|
|
|
|
void compileNewSymbol()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
if (!m_node->child1()) {
|
|
setJSValue(vmCall(pointerType(), operationNewSymbol, m_vmValue));
|
|
return;
|
|
}
|
|
ASSERT(m_node->child1().useKind() == KnownStringUse);
|
|
setJSValue(vmCall(pointerType(), operationNewSymbolWithDescription, weakPointer(globalObject), lowString(m_node->child1())));
|
|
}
|
|
|
|
void compileNewArray()
|
|
{
|
|
// First speculate appropriately on all of the children. Do this unconditionally up here
|
|
// because some of the slow paths may otherwise forget to do it. It's sort of arguable
|
|
// that doing the speculations up here might be unprofitable for RA - so we can consider
|
|
// sinking this to below the allocation fast path if we find that this has a lot of
|
|
// register pressure.
|
|
// Because we first speculate on all of the children here, we can never exit after creating
|
|
// uninitialized contiguous JSArray, which ensures that we will never produce a half-baked JSArray.
|
|
for (unsigned operandIndex = 0; operandIndex < m_node->numChildren(); ++operandIndex)
|
|
speculate(m_graph.varArgChild(m_node, operandIndex));
|
|
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
RegisteredStructure structure = m_graph.registerStructure(globalObject->arrayStructureForIndexingTypeDuringAllocation(
|
|
m_node->indexingType()));
|
|
|
|
if (!globalObject->isHavingABadTime() && !hasAnyArrayStorage(m_node->indexingType())) {
|
|
unsigned numElements = m_node->numChildren();
|
|
unsigned vectorLengthHint = m_node->vectorLengthHint();
|
|
ASSERT(vectorLengthHint >= numElements);
|
|
|
|
ArrayValues arrayValues =
|
|
allocateUninitializedContiguousJSArray(numElements, vectorLengthHint, structure);
|
|
|
|
for (unsigned operandIndex = 0; operandIndex < m_node->numChildren(); ++operandIndex) {
|
|
Edge edge = m_graph.varArgChild(m_node, operandIndex);
|
|
|
|
switch (m_node->indexingType()) {
|
|
case ALL_BLANK_INDEXING_TYPES:
|
|
case ALL_UNDECIDED_INDEXING_TYPES:
|
|
DFG_CRASH(m_graph, m_node, "Bad indexing type");
|
|
break;
|
|
|
|
case ALL_DOUBLE_INDEXING_TYPES:
|
|
m_out.storeDouble(
|
|
lowDouble(edge),
|
|
arrayValues.butterfly, m_heaps.indexedDoubleProperties[operandIndex]);
|
|
break;
|
|
|
|
case ALL_INT32_INDEXING_TYPES:
|
|
case ALL_CONTIGUOUS_INDEXING_TYPES:
|
|
m_out.store64(
|
|
lowJSValue(edge, ManualOperandSpeculation),
|
|
arrayValues.butterfly,
|
|
m_heaps.forIndexingType(m_node->indexingType())->at(operandIndex));
|
|
break;
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Corrupt indexing type");
|
|
break;
|
|
}
|
|
}
|
|
|
|
setJSValue(arrayValues.array);
|
|
mutatorFence();
|
|
return;
|
|
}
|
|
|
|
if (!m_node->numChildren()) {
|
|
setJSValue(vmCall(
|
|
Int64, operationNewEmptyArray, m_vmValue,
|
|
weakStructure(structure)));
|
|
return;
|
|
}
|
|
|
|
size_t scratchSize = sizeof(EncodedJSValue) * m_node->numChildren();
|
|
ASSERT(scratchSize);
|
|
ScratchBuffer* scratchBuffer = vm().scratchBufferForSize(scratchSize);
|
|
EncodedJSValue* buffer = static_cast<EncodedJSValue*>(scratchBuffer->dataBuffer());
|
|
|
|
for (unsigned operandIndex = 0; operandIndex < m_node->numChildren(); ++operandIndex) {
|
|
Edge edge = m_graph.varArgChild(m_node, operandIndex);
|
|
LValue valueToStore;
|
|
switch (m_node->indexingType()) {
|
|
case ALL_DOUBLE_INDEXING_TYPES:
|
|
valueToStore = boxDouble(lowDouble(edge));
|
|
break;
|
|
default:
|
|
valueToStore = lowJSValue(edge, ManualOperandSpeculation);
|
|
break;
|
|
}
|
|
m_out.store64(valueToStore, m_out.absolute(buffer + operandIndex));
|
|
}
|
|
|
|
m_out.storePtr(
|
|
m_out.constIntPtr(scratchSize), m_out.absolute(scratchBuffer->addressOfActiveLength()));
|
|
|
|
LValue result = vmCall(
|
|
Int64, operationNewArray, weakPointer(globalObject),
|
|
weakStructure(structure), m_out.constIntPtr(buffer),
|
|
m_out.constIntPtr(m_node->numChildren()));
|
|
|
|
m_out.storePtr(m_out.intPtrZero, m_out.absolute(scratchBuffer->addressOfActiveLength()));
|
|
|
|
setJSValue(result);
|
|
}
|
|
|
|
void compileNewArrayWithSpread()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
if (m_graph.isWatchingHavingABadTimeWatchpoint(m_node)) {
|
|
CheckedInt32 startLength = 0;
|
|
BitVector* bitVector = m_node->bitVector();
|
|
HashMap<InlineCallFrame*, LValue, WTF::DefaultHash<InlineCallFrame*>, WTF::NullableHashTraits<InlineCallFrame*>> cachedSpreadLengths;
|
|
|
|
if (m_node->numChildren() == 1 && bitVector->get(0)) {
|
|
Edge use = m_graph.varArgChild(m_node, 0);
|
|
if (use->op() == PhantomSpread) {
|
|
if (use->child1()->op() == PhantomNewArrayBuffer) {
|
|
auto* immutableButterfly = use->child1()->castOperand<JSImmutableButterfly*>();
|
|
if (hasContiguous(immutableButterfly->indexingType())) {
|
|
RegisteredStructure structure = m_graph.registerStructure(globalObject->originalArrayStructureForIndexingType(CopyOnWriteArrayWithContiguous));
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue fastArray = allocateObject<JSArray>(structure, m_out.constIntPtr(immutableButterfly->toButterfly()), slowPath);
|
|
ValueFromBlock fastResult = m_out.anchor(fastArray);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
LValue slowArray = vmCall(Int64, operationNewArrayBuffer, m_vmValue, weakStructure(structure), frozenPointer(use->child1()->cellOperand()));
|
|
ValueFromBlock slowResult = m_out.anchor(slowArray);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation);
|
|
|
|
mutatorFence();
|
|
setJSValue(m_out.phi(pointerType(), slowResult, fastResult));
|
|
return;
|
|
}
|
|
}
|
|
} else {
|
|
// If a node is producing JSImmutableButterfly, it must be contiguous.
|
|
LValue immutableButterfly = lowCell(use);
|
|
|
|
RegisteredStructure structure = m_graph.registerStructure(globalObject->arrayStructureForIndexingTypeDuringAllocation(CopyOnWriteArrayWithContiguous));
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue fastArray = allocateObject<JSArray>(structure, toButterfly(immutableButterfly), slowPath);
|
|
ValueFromBlock fastResult = m_out.anchor(fastArray);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
LValue slowArray = vmCall(Int64, operationNewArrayBuffer, m_vmValue, weakStructure(structure), immutableButterfly);
|
|
ValueFromBlock slowResult = m_out.anchor(slowArray);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation);
|
|
|
|
mutatorFence();
|
|
setJSValue(m_out.phi(pointerType(), slowResult, fastResult));
|
|
return;
|
|
}
|
|
}
|
|
|
|
for (unsigned i = 0; i < m_node->numChildren(); ++i) {
|
|
if (!bitVector->get(i))
|
|
++startLength;
|
|
else {
|
|
Edge& child = m_graph.varArgChild(m_node, i);
|
|
if (child->op() == PhantomSpread && child->child1()->op() == PhantomNewArrayBuffer)
|
|
startLength += child->child1()->castOperand<JSImmutableButterfly*>()->length();
|
|
}
|
|
}
|
|
|
|
if (startLength.hasOverflowed()) {
|
|
terminate(Overflow);
|
|
return;
|
|
}
|
|
|
|
LValue length = m_out.constInt32(startLength.unsafeGet());
|
|
|
|
for (unsigned i = 0; i < m_node->numChildren(); ++i) {
|
|
if (bitVector->get(i)) {
|
|
Edge use = m_graph.varArgChild(m_node, i);
|
|
CheckValue* lengthCheck = nullptr;
|
|
if (use->op() == PhantomSpread) {
|
|
if (use->child1()->op() == PhantomCreateRest) {
|
|
InlineCallFrame* inlineCallFrame = use->child1()->origin.semantic.inlineCallFrame();
|
|
unsigned numberOfArgumentsToSkip = use->child1()->numberOfArgumentsToSkip();
|
|
LValue spreadLength = cachedSpreadLengths.ensure(inlineCallFrame, [&] () {
|
|
return getSpreadLengthFromInlineCallFrame(inlineCallFrame, numberOfArgumentsToSkip);
|
|
}).iterator->value;
|
|
lengthCheck = m_out.speculateAdd(length, spreadLength);
|
|
}
|
|
} else {
|
|
LValue immutableButterfly = lowCell(use);
|
|
lengthCheck = m_out.speculateAdd(length, m_out.load32(toButterfly(immutableButterfly), m_heaps.Butterfly_publicLength));
|
|
}
|
|
|
|
if (lengthCheck) {
|
|
blessSpeculation(lengthCheck, Overflow, noValue(), nullptr, m_origin);
|
|
length = lengthCheck;
|
|
}
|
|
}
|
|
}
|
|
|
|
LValue exceedsMaxAllowedLength = m_out.aboveOrEqual(length, m_out.constInt32(MIN_ARRAY_STORAGE_CONSTRUCTION_LENGTH));
|
|
blessSpeculation(m_out.speculate(exceedsMaxAllowedLength), Overflow, noValue(), nullptr, m_origin);
|
|
|
|
RegisteredStructure structure = m_graph.registerStructure(m_graph.globalObjectFor(m_origin.semantic)->originalArrayStructureForIndexingType(ArrayWithContiguous));
|
|
ArrayValues arrayValues = allocateUninitializedContiguousJSArray(length, structure);
|
|
LValue result = arrayValues.array;
|
|
LValue storage = arrayValues.butterfly;
|
|
LValue index = m_out.constIntPtr(0);
|
|
|
|
for (unsigned i = 0; i < m_node->numChildren(); ++i) {
|
|
Edge use = m_graph.varArgChild(m_node, i);
|
|
if (bitVector->get(i)) {
|
|
if (use->op() == PhantomSpread) {
|
|
if (use->child1()->op() == PhantomNewArrayBuffer) {
|
|
IndexedAbstractHeap& heap = m_heaps.indexedContiguousProperties;
|
|
auto* array = use->child1()->castOperand<JSImmutableButterfly*>();
|
|
for (unsigned i = 0; i < array->length(); ++i) {
|
|
// Because resulted array from NewArrayWithSpread is always contiguous, we should not generate value
|
|
// in Double form even if PhantomNewArrayBuffer's indexingType is ArrayWithDouble.
|
|
int64_t value = JSValue::encode(array->get(i));
|
|
m_out.store64(m_out.constInt64(value), m_out.baseIndex(heap, storage, index, JSValue(), (Checked<int32_t>(sizeof(JSValue)) * i).unsafeGet()));
|
|
}
|
|
index = m_out.add(index, m_out.constIntPtr(array->length()));
|
|
} else {
|
|
RELEASE_ASSERT(use->child1()->op() == PhantomCreateRest);
|
|
InlineCallFrame* inlineCallFrame = use->child1()->origin.semantic.inlineCallFrame();
|
|
unsigned numberOfArgumentsToSkip = use->child1()->numberOfArgumentsToSkip();
|
|
|
|
LValue length = m_out.zeroExtPtr(cachedSpreadLengths.get(inlineCallFrame));
|
|
LValue sourceStart = getArgumentsStart(inlineCallFrame, numberOfArgumentsToSkip);
|
|
|
|
LBasicBlock loopStart = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
ValueFromBlock loadIndexStart = m_out.anchor(m_out.constIntPtr(0));
|
|
ValueFromBlock arrayIndexStart = m_out.anchor(index);
|
|
ValueFromBlock arrayIndexStartForFinish = m_out.anchor(index);
|
|
|
|
m_out.branch(
|
|
m_out.isZero64(length),
|
|
unsure(continuation), unsure(loopStart));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(loopStart, continuation);
|
|
|
|
LValue arrayIndex = m_out.phi(pointerType(), arrayIndexStart);
|
|
LValue loadIndex = m_out.phi(pointerType(), loadIndexStart);
|
|
|
|
LValue item = m_out.load64(m_out.baseIndex(m_heaps.variables, sourceStart, loadIndex));
|
|
m_out.store64(item, m_out.baseIndex(m_heaps.indexedContiguousProperties, storage, arrayIndex));
|
|
|
|
LValue nextArrayIndex = m_out.add(arrayIndex, m_out.constIntPtr(1));
|
|
LValue nextLoadIndex = m_out.add(loadIndex, m_out.constIntPtr(1));
|
|
ValueFromBlock arrayIndexLoopForFinish = m_out.anchor(nextArrayIndex);
|
|
|
|
m_out.addIncomingToPhi(loadIndex, m_out.anchor(nextLoadIndex));
|
|
m_out.addIncomingToPhi(arrayIndex, m_out.anchor(nextArrayIndex));
|
|
|
|
m_out.branch(
|
|
m_out.below(nextLoadIndex, length),
|
|
unsure(loopStart), unsure(continuation));
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
index = m_out.phi(pointerType(), arrayIndexStartForFinish, arrayIndexLoopForFinish);
|
|
}
|
|
} else {
|
|
LBasicBlock loopStart = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue immutableButterfly = lowCell(use);
|
|
LValue immutableButterflyStorage = toButterfly(immutableButterfly);
|
|
|
|
ValueFromBlock immutableButterflyIndexStart = m_out.anchor(m_out.constIntPtr(0));
|
|
ValueFromBlock arrayIndexStart = m_out.anchor(index);
|
|
ValueFromBlock arrayIndexStartForFinish = m_out.anchor(index);
|
|
|
|
LValue immutableButterflySize = m_out.zeroExtPtr(m_out.load32(immutableButterflyStorage, m_heaps.Butterfly_publicLength));
|
|
|
|
m_out.branch(
|
|
m_out.isZero64(immutableButterflySize),
|
|
unsure(continuation), unsure(loopStart));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(loopStart, continuation);
|
|
|
|
LValue arrayIndex = m_out.phi(pointerType(), arrayIndexStart);
|
|
LValue immutableButterflyIndex = m_out.phi(pointerType(), immutableButterflyIndexStart);
|
|
|
|
LValue item = m_out.load64(m_out.baseIndex(m_heaps.indexedContiguousProperties, immutableButterflyStorage, immutableButterflyIndex));
|
|
m_out.store64(item, m_out.baseIndex(m_heaps.indexedContiguousProperties, storage, arrayIndex));
|
|
|
|
LValue nextArrayIndex = m_out.add(arrayIndex, m_out.constIntPtr(1));
|
|
LValue nextImmutableButterflyIndex = m_out.add(immutableButterflyIndex, m_out.constIntPtr(1));
|
|
ValueFromBlock arrayIndexLoopForFinish = m_out.anchor(nextArrayIndex);
|
|
|
|
m_out.addIncomingToPhi(immutableButterflyIndex, m_out.anchor(nextImmutableButterflyIndex));
|
|
m_out.addIncomingToPhi(arrayIndex, m_out.anchor(nextArrayIndex));
|
|
|
|
m_out.branch(
|
|
m_out.below(nextImmutableButterflyIndex, immutableButterflySize),
|
|
unsure(loopStart), unsure(continuation));
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
index = m_out.phi(pointerType(), arrayIndexStartForFinish, arrayIndexLoopForFinish);
|
|
}
|
|
} else {
|
|
IndexedAbstractHeap& heap = m_heaps.indexedContiguousProperties;
|
|
LValue item = lowJSValue(use);
|
|
m_out.store64(item, m_out.baseIndex(heap, storage, index));
|
|
index = m_out.add(index, m_out.constIntPtr(1));
|
|
}
|
|
}
|
|
|
|
mutatorFence();
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
|
|
ASSERT(m_node->numChildren());
|
|
size_t scratchSize = sizeof(EncodedJSValue) * m_node->numChildren();
|
|
ScratchBuffer* scratchBuffer = vm().scratchBufferForSize(scratchSize);
|
|
EncodedJSValue* buffer = static_cast<EncodedJSValue*>(scratchBuffer->dataBuffer());
|
|
BitVector* bitVector = m_node->bitVector();
|
|
for (unsigned i = 0; i < m_node->numChildren(); ++i) {
|
|
Edge use = m_graph.m_varArgChildren[m_node->firstChild() + i];
|
|
LValue value;
|
|
if (bitVector->get(i))
|
|
value = lowCell(use);
|
|
else
|
|
value = lowJSValue(use);
|
|
m_out.store64(value, m_out.absolute(&buffer[i]));
|
|
}
|
|
|
|
m_out.storePtr(m_out.constIntPtr(scratchSize), m_out.absolute(scratchBuffer->addressOfActiveLength()));
|
|
LValue result = vmCall(Int64, operationNewArrayWithSpreadSlow, weakPointer(globalObject), m_out.constIntPtr(buffer), m_out.constInt32(m_node->numChildren()));
|
|
m_out.storePtr(m_out.constIntPtr(0), m_out.absolute(scratchBuffer->addressOfActiveLength()));
|
|
|
|
setJSValue(result);
|
|
}
|
|
|
|
void compileCreateThis()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue callee = lowCell(m_node->child1());
|
|
|
|
LBasicBlock isFunctionBlock = m_out.newBlock();
|
|
LBasicBlock hasRareData = m_out.newBlock();
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(isFunction(callee, provenType(m_node->child1())), usually(isFunctionBlock), rarely(slowPath));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isFunctionBlock, hasRareData);
|
|
LValue rareDataTags = m_out.loadPtr(callee, m_heaps.JSFunction_executableOrRareData);
|
|
m_out.branch(m_out.testIsZeroPtr(rareDataTags, m_out.constIntPtr(JSFunction::rareDataTag)), rarely(slowPath), usually(hasRareData));
|
|
|
|
m_out.appendTo(hasRareData, slowPath);
|
|
LValue rareData = m_out.sub(rareDataTags, m_out.constIntPtr(JSFunction::rareDataTag));
|
|
LValue allocator = m_out.loadPtr(rareData, m_heaps.FunctionRareData_allocator);
|
|
LValue structure = m_out.loadPtr(rareData, m_heaps.FunctionRareData_structure);
|
|
LValue butterfly = m_out.constIntPtr(0);
|
|
ValueFromBlock fastResult = m_out.anchor(allocateObject(allocator, structure, butterfly, slowPath));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
ValueFromBlock slowResult = m_out.anchor(vmCall(
|
|
Int64, operationCreateThis, weakPointer(globalObject), callee, m_out.constInt32(m_node->inlineCapacity())));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
LValue result = m_out.phi(Int64, fastResult, slowResult);
|
|
|
|
mutatorFence();
|
|
setJSValue(result);
|
|
}
|
|
|
|
void compileCreatePromise()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
|
|
LValue callee = lowCell(m_node->child1());
|
|
|
|
LBasicBlock derivedCase = m_out.newBlock();
|
|
LBasicBlock isFunctionBlock = m_out.newBlock();
|
|
LBasicBlock hasRareData = m_out.newBlock();
|
|
LBasicBlock hasStructure = m_out.newBlock();
|
|
LBasicBlock checkGlobalObjectCase = m_out.newBlock();
|
|
LBasicBlock fastAllocationCase = m_out.newBlock();
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
ValueFromBlock promiseStructure = m_out.anchor(weakStructure(m_graph.registerStructure(m_node->isInternalPromise() ? globalObject->internalPromiseStructure() : globalObject->promiseStructure())));
|
|
m_out.branch(m_out.equal(callee, weakPointer(m_node->isInternalPromise() ? globalObject->internalPromiseConstructor() : globalObject->promiseConstructor())), unsure(fastAllocationCase), unsure(derivedCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(derivedCase, isFunctionBlock);
|
|
m_out.branch(isFunction(callee, provenType(m_node->child1())), usually(isFunctionBlock), rarely(slowCase));
|
|
|
|
m_out.appendTo(isFunctionBlock, hasRareData);
|
|
LValue rareDataTags = m_out.loadPtr(callee, m_heaps.JSFunction_executableOrRareData);
|
|
m_out.branch(m_out.testIsZeroPtr(rareDataTags, m_out.constIntPtr(JSFunction::rareDataTag)), rarely(slowCase), usually(hasRareData));
|
|
|
|
m_out.appendTo(hasRareData, hasStructure);
|
|
LValue rareData = m_out.sub(rareDataTags, m_out.constIntPtr(JSFunction::rareDataTag));
|
|
LValue structure = m_out.loadPtr(rareData, m_heaps.FunctionRareData_internalFunctionAllocationProfile_structure);
|
|
m_out.branch(m_out.isZero64(structure), rarely(slowCase), usually(hasStructure));
|
|
|
|
m_out.appendTo(hasStructure, checkGlobalObjectCase);
|
|
m_out.branch(m_out.equal(m_out.loadPtr(structure, m_heaps.Structure_classInfo), m_out.constIntPtr(m_node->isInternalPromise() ? JSInternalPromise::info() : JSPromise::info())), usually(checkGlobalObjectCase), rarely(slowCase));
|
|
|
|
m_out.appendTo(checkGlobalObjectCase, fastAllocationCase);
|
|
ValueFromBlock derivedStructure = m_out.anchor(structure);
|
|
m_out.branch(m_out.equal(m_out.loadPtr(structure, m_heaps.Structure_globalObject), weakPointer(globalObject)), usually(fastAllocationCase), rarely(slowCase));
|
|
|
|
m_out.appendTo(fastAllocationCase, slowCase);
|
|
LValue promise;
|
|
if (m_node->isInternalPromise())
|
|
promise = allocateObject<JSInternalPromise>(m_out.phi(pointerType(), promiseStructure, derivedStructure), m_out.intPtrZero, slowCase);
|
|
else
|
|
promise = allocateObject<JSPromise>(m_out.phi(pointerType(), promiseStructure, derivedStructure), m_out.intPtrZero, slowCase);
|
|
m_out.store64(m_out.constInt64(JSValue::encode(jsNumber(static_cast<unsigned>(JSPromise::Status::Pending)))), promise, m_heaps.JSInternalFieldObjectImpl_internalFields[static_cast<unsigned>(JSPromise::Field::Flags)]);
|
|
m_out.store64(m_out.constInt64(JSValue::encode(jsUndefined())), promise, m_heaps.JSInternalFieldObjectImpl_internalFields[static_cast<unsigned>(JSPromise::Field::ReactionsOrResult)]);
|
|
mutatorFence();
|
|
ValueFromBlock fastResult = m_out.anchor(promise);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
ValueFromBlock slowResult = m_out.anchor(vmCall(Int64, m_node->isInternalPromise() ? operationCreateInternalPromise : operationCreatePromise, weakPointer(globalObject), callee));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
LValue result = m_out.phi(Int64, fastResult, slowResult);
|
|
|
|
setJSValue(result);
|
|
}
|
|
|
|
template<typename JSClass, typename Operation>
|
|
void compileCreateInternalFieldObject(Operation operation)
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
|
|
LValue callee = lowCell(m_node->child1());
|
|
|
|
LBasicBlock isFunctionBlock = m_out.newBlock();
|
|
LBasicBlock hasRareData = m_out.newBlock();
|
|
LBasicBlock hasStructure = m_out.newBlock();
|
|
LBasicBlock checkGlobalObjectCase = m_out.newBlock();
|
|
LBasicBlock fastAllocationCase = m_out.newBlock();
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(isFunction(callee, provenType(m_node->child1())), usually(isFunctionBlock), rarely(slowCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isFunctionBlock, hasRareData);
|
|
LValue rareDataTags = m_out.loadPtr(callee, m_heaps.JSFunction_executableOrRareData);
|
|
m_out.branch(m_out.testIsZeroPtr(rareDataTags, m_out.constIntPtr(JSFunction::rareDataTag)), rarely(slowCase), usually(hasRareData));
|
|
|
|
m_out.appendTo(hasRareData, hasStructure);
|
|
LValue rareData = m_out.sub(rareDataTags, m_out.constIntPtr(JSFunction::rareDataTag));
|
|
LValue structure = m_out.loadPtr(rareData, m_heaps.FunctionRareData_internalFunctionAllocationProfile_structure);
|
|
m_out.branch(m_out.isZero64(structure), rarely(slowCase), usually(hasStructure));
|
|
|
|
m_out.appendTo(hasStructure, checkGlobalObjectCase);
|
|
m_out.branch(m_out.equal(m_out.loadPtr(structure, m_heaps.Structure_classInfo), m_out.constIntPtr(JSClass::info())), usually(checkGlobalObjectCase), rarely(slowCase));
|
|
|
|
m_out.appendTo(checkGlobalObjectCase, fastAllocationCase);
|
|
m_out.branch(m_out.equal(m_out.loadPtr(structure, m_heaps.Structure_globalObject), weakPointer(globalObject)), usually(fastAllocationCase), rarely(slowCase));
|
|
|
|
m_out.appendTo(fastAllocationCase, slowCase);
|
|
LValue object = allocateObject<JSClass>(structure, m_out.intPtrZero, slowCase);
|
|
auto initialValues = JSClass::initialValues();
|
|
static_assert(initialValues.size() == JSClass::numberOfInternalFields, "We don't support non-constant fields in create yet.");
|
|
for (unsigned index = 0; index < initialValues.size(); ++index)
|
|
m_out.store64(m_out.constInt64(JSValue::encode(initialValues[index])), object, m_heaps.JSInternalFieldObjectImpl_internalFields[index]);
|
|
mutatorFence();
|
|
ValueFromBlock fastResult = m_out.anchor(object);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
ValueFromBlock slowResult = m_out.anchor(vmCall(Int64, operation, weakPointer(globalObject), callee));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
LValue result = m_out.phi(Int64, fastResult, slowResult);
|
|
|
|
setJSValue(result);
|
|
}
|
|
|
|
void compileCreateGenerator()
|
|
{
|
|
compileCreateInternalFieldObject<JSGenerator>(operationCreateGenerator);
|
|
}
|
|
|
|
void compileCreateAsyncGenerator()
|
|
{
|
|
compileCreateInternalFieldObject<JSAsyncGenerator>(operationCreateAsyncGenerator);
|
|
}
|
|
|
|
void compileSpread()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
if (m_node->child1()->op() == PhantomNewArrayBuffer) {
|
|
ASSERT(m_graph.isWatchingHavingABadTimeWatchpoint(m_node->child1().node()));
|
|
|
|
// FIXME: JSImmutableButterfly::createFromArray should support re-using non contiguous indexing types as well.
|
|
auto* immutableButterfly = m_node->child1()->castOperand<JSImmutableButterfly*>();
|
|
if (hasContiguous(immutableButterfly->indexingType())) {
|
|
setJSValue(frozenPointer(m_node->child1()->cellOperand()));
|
|
return;
|
|
}
|
|
|
|
LBasicBlock slowAllocation = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
ASSERT(immutableButterfly->length() <= MAX_STORAGE_VECTOR_LENGTH);
|
|
|
|
LValue fastImmutableButterflyValue = allocateVariableSizedCell<JSImmutableButterfly>(
|
|
m_out.constIntPtr(JSImmutableButterfly::allocationSize(immutableButterfly->length()).unsafeGet()),
|
|
m_graph.m_vm.immutableButterflyStructures[arrayIndexFromIndexingType(CopyOnWriteArrayWithContiguous) - NumberOfIndexingShapes].get(), slowAllocation);
|
|
LValue fastImmutableButterflyStorage = toButterfly(fastImmutableButterflyValue);
|
|
m_out.store32(m_out.constInt32(immutableButterfly->length()), fastImmutableButterflyStorage, m_heaps.Butterfly_publicLength);
|
|
m_out.store32(m_out.constInt32(immutableButterfly->length()), fastImmutableButterflyStorage, m_heaps.Butterfly_vectorLength);
|
|
ValueFromBlock fastImmutableButterfly = m_out.anchor(fastImmutableButterflyValue);
|
|
m_out.jump(continuation);
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(slowAllocation, continuation);
|
|
ValueFromBlock slowImmutableButterfly = m_out.anchor(vmCall(pointerType(), operationCreateImmutableButterfly, weakPointer(globalObject), m_out.constInt32(immutableButterfly->length())));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
LValue immutableButterflyValue = m_out.phi(pointerType(), fastImmutableButterfly, slowImmutableButterfly);
|
|
LValue immutableButterflyStorage = toButterfly(immutableButterflyValue);
|
|
for (unsigned i = 0; i < immutableButterfly->length(); i++) {
|
|
// Because forwarded values are drained as JSValue, we should not generate value
|
|
// in Double form even if PhantomNewArrayBuffer's indexingType is ArrayWithDouble.
|
|
int64_t value = JSValue::encode(immutableButterfly->get(i));
|
|
m_out.store64(
|
|
m_out.constInt64(value),
|
|
m_out.baseIndex(m_heaps.indexedContiguousProperties, immutableButterflyStorage, m_out.constIntPtr(i), jsNumber(i)));
|
|
}
|
|
mutatorFence();
|
|
setJSValue(immutableButterflyValue);
|
|
return;
|
|
}
|
|
|
|
if (m_node->child1()->op() == PhantomCreateRest) {
|
|
// This IR is rare to generate since it requires escaping the Spread
|
|
// but not the CreateRest. In bytecode, we have only few operations that
|
|
// accept Spread's result as input. This usually leads to the Spread node not
|
|
// escaping. However, this can happen if for example we generate a PutStack on
|
|
// the Spread but nothing escapes the CreateRest.
|
|
|
|
ASSERT(m_graph.isWatchingHavingABadTimeWatchpoint(m_node->child1().node()));
|
|
|
|
LBasicBlock fastAllocation = m_out.newBlock();
|
|
LBasicBlock loopHeader = m_out.newBlock();
|
|
LBasicBlock loopBody = m_out.newBlock();
|
|
LBasicBlock slowAllocation = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
InlineCallFrame* inlineCallFrame = m_node->child1()->origin.semantic.inlineCallFrame();
|
|
unsigned numberOfArgumentsToSkip = m_node->child1()->numberOfArgumentsToSkip();
|
|
LValue sourceStart = getArgumentsStart(inlineCallFrame, numberOfArgumentsToSkip);
|
|
LValue length = getSpreadLengthFromInlineCallFrame(inlineCallFrame, numberOfArgumentsToSkip);
|
|
static_assert(sizeof(JSValue) == 8 && 1 << 3 == 8, "Assumed in the code below.");
|
|
LValue size = m_out.add(
|
|
m_out.shl(m_out.zeroExtPtr(length), m_out.constInt32(3)),
|
|
m_out.constIntPtr(JSImmutableButterfly::offsetOfData()));
|
|
m_out.branch(m_out.above(length, m_out.constInt32(MAX_STORAGE_VECTOR_LENGTH)), rarely(slowAllocation), usually(fastAllocation));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(fastAllocation, slowAllocation);
|
|
LValue fastArrayValue = allocateVariableSizedCell<JSImmutableButterfly>(size, m_graph.m_vm.immutableButterflyStructures[arrayIndexFromIndexingType(CopyOnWriteArrayWithContiguous) - NumberOfIndexingShapes].get(), slowAllocation);
|
|
LValue fastArrayStorage = toButterfly(fastArrayValue);
|
|
m_out.store32(length, fastArrayStorage, m_heaps.Butterfly_vectorLength);
|
|
m_out.store32(length, fastArrayStorage, m_heaps.Butterfly_publicLength);
|
|
ValueFromBlock fastArray = m_out.anchor(fastArrayValue);
|
|
m_out.jump(loopHeader);
|
|
|
|
m_out.appendTo(slowAllocation, loopHeader);
|
|
ValueFromBlock slowArray = m_out.anchor(vmCall(pointerType(), operationCreateImmutableButterfly, weakPointer(globalObject), length));
|
|
m_out.jump(loopHeader);
|
|
|
|
m_out.appendTo(loopHeader, loopBody);
|
|
LValue immutableButterfly = m_out.phi(pointerType(), fastArray, slowArray);
|
|
LValue immutableButterflyStorage = toButterfly(immutableButterfly);
|
|
ValueFromBlock startIndex = m_out.anchor(m_out.constIntPtr(0));
|
|
m_out.branch(m_out.isZero32(length), unsure(continuation), unsure(loopBody));
|
|
|
|
m_out.appendTo(loopBody, continuation);
|
|
LValue index = m_out.phi(pointerType(), startIndex);
|
|
LValue value = m_out.load64(
|
|
m_out.baseIndex(m_heaps.variables, sourceStart, index));
|
|
m_out.store64(value, m_out.baseIndex(m_heaps.indexedContiguousProperties, immutableButterflyStorage, index));
|
|
LValue nextIndex = m_out.add(m_out.constIntPtr(1), index);
|
|
m_out.addIncomingToPhi(index, m_out.anchor(nextIndex));
|
|
m_out.branch(m_out.below(nextIndex, m_out.zeroExtPtr(length)), unsure(loopBody), unsure(continuation));
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
mutatorFence();
|
|
setJSValue(immutableButterfly);
|
|
return;
|
|
}
|
|
|
|
LValue argument = lowCell(m_node->child1());
|
|
|
|
LValue result;
|
|
|
|
if (m_node->child1().useKind() == ArrayUse)
|
|
speculateArray(m_node->child1());
|
|
|
|
if (m_graph.canDoFastSpread(m_node, m_state.forNode(m_node->child1()))) {
|
|
LBasicBlock copyOnWriteContiguousCheck = m_out.newBlock();
|
|
LBasicBlock copyOnWritePropagation = m_out.newBlock();
|
|
LBasicBlock preLoop = m_out.newBlock();
|
|
LBasicBlock loopSelection = m_out.newBlock();
|
|
LBasicBlock contiguousLoopStart = m_out.newBlock();
|
|
LBasicBlock doubleLoopStart = m_out.newBlock();
|
|
LBasicBlock fastPath = m_out.newBlock();
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue indexingMode = m_out.load8ZeroExt32(argument, m_heaps.JSCell_indexingTypeAndMisc);
|
|
LValue indexingShape = m_out.bitAnd(indexingMode, m_out.constInt32(IndexingShapeMask));
|
|
LValue isOKIndexingType = m_out.belowOrEqual(
|
|
m_out.sub(indexingShape, m_out.constInt32(Int32Shape)),
|
|
m_out.constInt32(ContiguousShape - Int32Shape));
|
|
|
|
m_out.branch(isOKIndexingType, unsure(copyOnWriteContiguousCheck), unsure(slowPath));
|
|
LBasicBlock lastNext = m_out.appendTo(copyOnWriteContiguousCheck, copyOnWritePropagation);
|
|
LValue butterfly = m_out.loadPtr(argument, m_heaps.JSObject_butterfly);
|
|
m_out.branch(m_out.equal(m_out.bitAnd(indexingMode, m_out.constInt32(IndexingModeMask)), m_out.constInt32(CopyOnWriteArrayWithContiguous)), unsure(copyOnWritePropagation), unsure(preLoop));
|
|
|
|
m_out.appendTo(copyOnWritePropagation, preLoop);
|
|
ValueFromBlock sharedResult = m_out.anchor(m_out.add(butterfly, m_out.constIntPtr(-JSImmutableButterfly::offsetOfData())));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(preLoop, fastPath);
|
|
LValue length = m_out.load32NonNegative(butterfly, m_heaps.Butterfly_publicLength);
|
|
static_assert(sizeof(JSValue) == 8 && 1 << 3 == 8, "Assumed in the code below.");
|
|
LValue size = m_out.add(
|
|
m_out.shl(m_out.zeroExtPtr(length), m_out.constInt32(3)),
|
|
m_out.constIntPtr(JSImmutableButterfly::offsetOfData()));
|
|
m_out.branch(m_out.above(length, m_out.constInt32(MAX_STORAGE_VECTOR_LENGTH)), rarely(slowPath), usually(fastPath));
|
|
|
|
m_out.appendTo(fastPath, loopSelection);
|
|
LValue fastAllocation = allocateVariableSizedCell<JSImmutableButterfly>(size, m_graph.m_vm.immutableButterflyStructures[arrayIndexFromIndexingType(CopyOnWriteArrayWithContiguous) - NumberOfIndexingShapes].get(), slowPath);
|
|
LValue fastStorage = toButterfly(fastAllocation);
|
|
m_out.store32(length, fastStorage, m_heaps.Butterfly_vectorLength);
|
|
m_out.store32(length, fastStorage, m_heaps.Butterfly_publicLength);
|
|
ValueFromBlock fastResult = m_out.anchor(fastAllocation);
|
|
|
|
ValueFromBlock startIndexForContiguous = m_out.anchor(m_out.constIntPtr(0));
|
|
ValueFromBlock startIndexForDouble = m_out.anchor(m_out.constIntPtr(0));
|
|
|
|
m_out.branch(m_out.isZero32(length), unsure(continuation), unsure(loopSelection));
|
|
|
|
m_out.appendTo(loopSelection, contiguousLoopStart);
|
|
m_out.branch(m_out.equal(indexingShape, m_out.constInt32(DoubleShape)),
|
|
unsure(doubleLoopStart), unsure(contiguousLoopStart));
|
|
|
|
{
|
|
m_out.appendTo(contiguousLoopStart, doubleLoopStart);
|
|
LValue index = m_out.phi(pointerType(), startIndexForContiguous);
|
|
|
|
TypedPointer loadSite = m_out.baseIndex(m_heaps.root, butterfly, index, ScaleEight); // We read TOP here since we can be reading either int32 or contiguous properties.
|
|
LValue value = m_out.load64(loadSite);
|
|
value = m_out.select(m_out.isZero64(value), m_out.constInt64(JSValue::encode(jsUndefined())), value);
|
|
m_out.store64(value, m_out.baseIndex(m_heaps.indexedContiguousProperties, fastStorage, index));
|
|
|
|
LValue nextIndex = m_out.add(index, m_out.constIntPtr(1));
|
|
m_out.addIncomingToPhi(index, m_out.anchor(nextIndex));
|
|
|
|
m_out.branch(m_out.below(nextIndex, m_out.zeroExtPtr(length)),
|
|
unsure(contiguousLoopStart), unsure(continuation));
|
|
}
|
|
|
|
{
|
|
m_out.appendTo(doubleLoopStart, slowPath);
|
|
LValue index = m_out.phi(pointerType(), startIndexForDouble);
|
|
|
|
LValue value = m_out.loadDouble(m_out.baseIndex(m_heaps.indexedDoubleProperties, butterfly, index));
|
|
LValue isNaN = m_out.doubleNotEqualOrUnordered(value, value);
|
|
LValue holeResult = m_out.constInt64(JSValue::encode(jsUndefined()));
|
|
LValue normalResult = boxDouble(value);
|
|
value = m_out.select(isNaN, holeResult, normalResult);
|
|
m_out.store64(value, m_out.baseIndex(m_heaps.indexedContiguousProperties, fastStorage, index));
|
|
|
|
LValue nextIndex = m_out.add(index, m_out.constIntPtr(1));
|
|
m_out.addIncomingToPhi(index, m_out.anchor(nextIndex));
|
|
|
|
m_out.branch(m_out.below(nextIndex, m_out.zeroExtPtr(length)),
|
|
unsure(doubleLoopStart), unsure(continuation));
|
|
}
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
ValueFromBlock slowResult = m_out.anchor(vmCall(pointerType(), operationSpreadFastArray, weakPointer(globalObject), argument));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
result = m_out.phi(pointerType(), sharedResult, fastResult, slowResult);
|
|
mutatorFence();
|
|
} else
|
|
result = vmCall(pointerType(), operationSpreadGeneric, weakPointer(globalObject), argument);
|
|
|
|
setJSValue(result);
|
|
}
|
|
|
|
void compileNewArrayBuffer()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
RegisteredStructure structure = m_graph.registerStructure(globalObject->arrayStructureForIndexingTypeDuringAllocation(
|
|
m_node->indexingMode()));
|
|
auto* immutableButterfly = m_node->castOperand<JSImmutableButterfly*>();
|
|
|
|
if (!globalObject->isHavingABadTime() && !hasAnyArrayStorage(m_node->indexingMode())) {
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue fastArray = allocateObject<JSArray>(structure, m_out.constIntPtr(immutableButterfly->toButterfly()), slowPath);
|
|
ValueFromBlock fastResult = m_out.anchor(fastArray);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
LValue slowArray = vmCall(Int64, operationNewArrayBuffer, m_vmValue, weakStructure(structure), frozenPointer(m_node->cellOperand()));
|
|
ValueFromBlock slowResult = m_out.anchor(slowArray);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation);
|
|
|
|
mutatorFence();
|
|
setJSValue(m_out.phi(pointerType(), slowResult, fastResult));
|
|
return;
|
|
}
|
|
|
|
setJSValue(vmCall(
|
|
Int64, operationNewArrayBuffer, m_vmValue,
|
|
weakStructure(structure), frozenPointer(m_node->cellOperand())));
|
|
}
|
|
|
|
void compileNewArrayWithSize()
|
|
{
|
|
LValue publicLength = lowInt32(m_node->child1());
|
|
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
RegisteredStructure structure = m_graph.registerStructure(globalObject->arrayStructureForIndexingTypeDuringAllocation(
|
|
m_node->indexingType()));
|
|
|
|
if (!globalObject->isHavingABadTime() && !hasAnyArrayStorage(m_node->indexingType())) {
|
|
IndexingType indexingType = m_node->indexingType();
|
|
setJSValue(
|
|
allocateJSArray(
|
|
publicLength, publicLength, weakPointer(globalObject->arrayStructureForIndexingTypeDuringAllocation(indexingType)), m_out.constInt32(indexingType)).array);
|
|
mutatorFence();
|
|
return;
|
|
}
|
|
|
|
LValue structureValue = m_out.select(
|
|
m_out.aboveOrEqual(publicLength, m_out.constInt32(MIN_ARRAY_STORAGE_CONSTRUCTION_LENGTH)),
|
|
weakStructure(m_graph.registerStructure(globalObject->arrayStructureForIndexingTypeDuringAllocation(ArrayWithArrayStorage))),
|
|
weakStructure(structure));
|
|
setJSValue(vmCall(Int64, operationNewArrayWithSize, weakPointer(globalObject), structureValue, publicLength, m_out.intPtrZero));
|
|
}
|
|
|
|
void compileNewTypedArray()
|
|
{
|
|
TypedArrayType typedArrayType = m_node->typedArrayType();
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
|
|
switch (m_node->child1().useKind()) {
|
|
case Int32Use: {
|
|
RegisteredStructure structure = m_graph.registerStructure(globalObject->typedArrayStructureConcurrently(typedArrayType));
|
|
|
|
LValue size = lowInt32(m_node->child1());
|
|
|
|
LBasicBlock smallEnoughCase = m_out.newBlock();
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
ValueFromBlock noStorage = m_out.anchor(m_out.intPtrZero);
|
|
|
|
m_out.branch(
|
|
m_out.above(size, m_out.constInt32(JSArrayBufferView::fastSizeLimit)),
|
|
rarely(slowCase), usually(smallEnoughCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(smallEnoughCase, slowCase);
|
|
|
|
LValue byteSize =
|
|
m_out.shl(m_out.zeroExtPtr(size), m_out.constInt32(logElementSize(typedArrayType)));
|
|
if (elementSize(typedArrayType) < 8) {
|
|
byteSize = m_out.bitAnd(
|
|
m_out.add(byteSize, m_out.constIntPtr(7)),
|
|
m_out.constIntPtr(~static_cast<intptr_t>(7)));
|
|
}
|
|
|
|
LValue allocator = allocatorForSize(vm().primitiveGigacageAuxiliarySpace, byteSize, slowCase);
|
|
LValue storage = allocateHeapCell(allocator, slowCase);
|
|
|
|
splatWords(
|
|
storage,
|
|
m_out.int32Zero,
|
|
m_out.castToInt32(m_out.lShr(byteSize, m_out.constIntPtr(3))),
|
|
m_out.int64Zero,
|
|
m_heaps.typedArrayProperties);
|
|
|
|
#if CPU(ARM64E)
|
|
{
|
|
LValue sizePtr = m_out.zeroExtPtr(size);
|
|
PatchpointValue* authenticate = m_out.patchpoint(pointerType());
|
|
authenticate->appendSomeRegister(storage);
|
|
authenticate->append(sizePtr, B3::ValueRep(B3::ValueRep::SomeLateRegister));
|
|
authenticate->setGenerator([=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
jit.move(params[1].gpr(), params[0].gpr());
|
|
jit.tagArrayPtr(params[2].gpr(), params[0].gpr());
|
|
});
|
|
storage = authenticate;
|
|
}
|
|
#endif
|
|
|
|
ValueFromBlock haveStorage = m_out.anchor(storage);
|
|
|
|
LValue fastResultValue = nullptr;
|
|
switch (typedArrayType) {
|
|
#define TYPED_ARRAY_TYPE_CASE(name) \
|
|
case Type ## name: \
|
|
fastResultValue = allocateObject<JS##name##Array>(structure, m_out.intPtrZero, slowCase); \
|
|
break;
|
|
FOR_EACH_TYPED_ARRAY_TYPE_EXCLUDING_DATA_VIEW(TYPED_ARRAY_TYPE_CASE)
|
|
#undef TYPED_ARRAY_TYPE_CASE
|
|
case TypeDataView:
|
|
fastResultValue = allocateObject<JSDataView>(structure, m_out.intPtrZero, slowCase);
|
|
break;
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
break;
|
|
}
|
|
|
|
m_out.storePtr(storage, fastResultValue, m_heaps.JSArrayBufferView_vector);
|
|
m_out.store32(size, fastResultValue, m_heaps.JSArrayBufferView_length);
|
|
m_out.store32(m_out.constInt32(FastTypedArray), fastResultValue, m_heaps.JSArrayBufferView_mode);
|
|
|
|
mutatorFence();
|
|
ValueFromBlock fastResult = m_out.anchor(fastResultValue);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
LValue storageValue = m_out.phi(pointerType(), noStorage, haveStorage);
|
|
|
|
VM& vm = this->vm();
|
|
LValue slowResultValue = lazySlowPath(
|
|
[=, &vm] (const Vector<Location>& locations) -> RefPtr<LazySlowPath::Generator> {
|
|
return createLazyCallGenerator(vm,
|
|
operationNewTypedArrayWithSizeForType(typedArrayType), locations[0].directGPR(), globalObject,
|
|
CCallHelpers::TrustedImmPtr(structure.get()), locations[1].directGPR(),
|
|
locations[2].directGPR());
|
|
},
|
|
size, storageValue);
|
|
ValueFromBlock slowResult = m_out.anchor(slowResultValue);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(pointerType(), fastResult, slowResult));
|
|
return;
|
|
}
|
|
|
|
case UntypedUse: {
|
|
LValue argument = lowJSValue(m_node->child1());
|
|
|
|
LValue result = vmCall(
|
|
pointerType(), operationNewTypedArrayWithOneArgumentForType(typedArrayType),
|
|
weakPointer(globalObject), weakPointer(globalObject->typedArrayStructureConcurrently(typedArrayType)), argument);
|
|
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
return;
|
|
}
|
|
}
|
|
|
|
void compileAllocatePropertyStorage()
|
|
{
|
|
LValue object = lowCell(m_node->child1());
|
|
setStorage(allocatePropertyStorage(object, m_node->transition()->previous.get()));
|
|
}
|
|
|
|
void compileReallocatePropertyStorage()
|
|
{
|
|
Transition* transition = m_node->transition();
|
|
LValue object = lowCell(m_node->child1());
|
|
LValue oldStorage = lowStorage(m_node->child2());
|
|
|
|
setStorage(
|
|
reallocatePropertyStorage(
|
|
object, oldStorage, transition->previous.get(), transition->next.get()));
|
|
}
|
|
|
|
void compileNukeStructureAndSetButterfly()
|
|
{
|
|
nukeStructureAndSetButterfly(lowStorage(m_node->child2()), lowCell(m_node->child1()));
|
|
}
|
|
|
|
void compileToNumber()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue value = lowJSValue(m_node->child1());
|
|
|
|
if (!(abstractValue(m_node->child1()).m_type & SpecBytecodeNumber))
|
|
setJSValue(vmCall(Int64, operationToNumber, weakPointer(globalObject), value));
|
|
else {
|
|
LBasicBlock notNumber = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
ValueFromBlock fastResult = m_out.anchor(value);
|
|
m_out.branch(isNumber(value, provenType(m_node->child1())), unsure(continuation), unsure(notNumber));
|
|
|
|
// notNumber case.
|
|
LBasicBlock lastNext = m_out.appendTo(notNumber, continuation);
|
|
// We have several attempts to remove ToNumber. But ToNumber still exists.
|
|
// It means that converting non-numbers to numbers by this ToNumber is not rare.
|
|
// Instead of the lazy slow path generator, we call the operation here.
|
|
ValueFromBlock slowResult = m_out.anchor(vmCall(Int64, operationToNumber, weakPointer(globalObject), value));
|
|
m_out.jump(continuation);
|
|
|
|
// continuation case.
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, fastResult, slowResult));
|
|
}
|
|
}
|
|
|
|
void compileToNumeric()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue value = lowJSValue(m_node->child1());
|
|
|
|
if (abstractValue(m_node->child1()).m_type & (SpecBytecodeNumber | SpecBigInt)) {
|
|
LBasicBlock notNumber = m_out.newBlock();
|
|
#if USE(BIGINT32)
|
|
LBasicBlock notBigInt32 = m_out.newBlock();
|
|
#endif
|
|
LBasicBlock isCellPath = m_out.newBlock();
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
ValueFromBlock fastResult = m_out.anchor(value);
|
|
m_out.branch(isNumber(value, provenType(m_node->child1())), unsure(continuation), unsure(notNumber));
|
|
|
|
// notNumber case.
|
|
LBasicBlock lastNext = m_out.appendTo(notNumber, continuation);
|
|
#if USE(BIGINT32)
|
|
m_out.branch(isBigInt32(value, provenType(m_node->child1())), unsure(continuation), unsure(notBigInt32));
|
|
m_out.appendTo(notBigInt32);
|
|
#endif
|
|
m_out.branch(isCell(value, provenType(m_node->child1())), unsure(isCellPath), unsure(slowPath));
|
|
|
|
m_out.appendTo(isCellPath);
|
|
m_out.branch(isHeapBigInt(value, provenType(m_node->child1())), unsure(continuation), unsure(slowPath));
|
|
|
|
m_out.appendTo(slowPath);
|
|
// We have several attempts to remove ToNumeric. But ToNumeric still exists.
|
|
// It means that the slow path is not rare.
|
|
// Instead of the lazy slow path generator, we call the operation here.
|
|
ValueFromBlock slowResult = m_out.anchor(vmCall(Int64, operationToNumeric, weakPointer(globalObject), value));
|
|
m_out.jump(continuation);
|
|
|
|
// continuation case.
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, fastResult, slowResult));
|
|
} else
|
|
setJSValue(vmCall(Int64, operationToNumeric, weakPointer(globalObject), value));
|
|
}
|
|
|
|
void compileCallNumberConstructor()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
#if USE(BIGINT32)
|
|
if (m_node->child1().useKind() == BigInt32Use) {
|
|
LValue value = lowBigInt32(m_node->child1());
|
|
setInt32(unboxBigInt32(value));
|
|
return;
|
|
}
|
|
#endif
|
|
LValue value = lowJSValue(m_node->child1());
|
|
|
|
LBasicBlock notNumber = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
ValueFromBlock fastResult = m_out.anchor(value);
|
|
m_out.branch(isNumber(value, provenType(m_node->child1())), unsure(continuation), unsure(notNumber));
|
|
|
|
// notNumber case.
|
|
LBasicBlock lastNext = m_out.appendTo(notNumber, continuation);
|
|
ValueFromBlock slowResult = m_out.anchor(vmCall(Int64, operationCallNumberConstructor, weakPointer(globalObject), value));
|
|
m_out.jump(continuation);
|
|
|
|
// continuation case.
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, fastResult, slowResult));
|
|
}
|
|
|
|
|
|
void compileToStringOrCallStringConstructorOrStringValueOf()
|
|
{
|
|
ASSERT(m_node->op() != StringValueOf || m_node->child1().useKind() == UntypedUse);
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
switch (m_node->child1().useKind()) {
|
|
case StringObjectUse: {
|
|
// FIXME: We should convert this to GetInternalField(0).
|
|
// https://bugs.webkit.org/show_bug.cgi?id=209453
|
|
LValue cell = lowCell(m_node->child1());
|
|
speculateStringObjectForCell(m_node->child1(), cell);
|
|
setJSValue(m_out.loadPtr(cell, m_heaps.JSWrapperObject_internalValue));
|
|
return;
|
|
}
|
|
|
|
case StringOrStringObjectUse: {
|
|
LValue cell = lowCell(m_node->child1());
|
|
LValue type = m_out.load8ZeroExt32(cell, m_heaps.JSCell_typeInfoType);
|
|
|
|
LBasicBlock notString = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
ValueFromBlock simpleResult = m_out.anchor(cell);
|
|
m_out.branch(
|
|
m_out.equal(type, m_out.constInt32(StringType)),
|
|
unsure(continuation), unsure(notString));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(notString, continuation);
|
|
speculate(
|
|
BadType, jsValueValue(cell), m_node->child1().node(),
|
|
m_out.notEqual(type, m_out.constInt32(StringObjectType)));
|
|
ValueFromBlock unboxedResult = m_out.anchor(
|
|
m_out.loadPtr(cell, m_heaps.JSWrapperObject_internalValue));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, simpleResult, unboxedResult));
|
|
|
|
m_interpreter.filter(m_node->child1(), SpecString | SpecStringObject);
|
|
return;
|
|
}
|
|
|
|
case CellUse:
|
|
case NotCellUse:
|
|
case UntypedUse: {
|
|
LValue value;
|
|
if (m_node->child1().useKind() == CellUse)
|
|
value = lowCell(m_node->child1());
|
|
else if (m_node->child1().useKind() == NotCellUse)
|
|
value = lowNotCell(m_node->child1());
|
|
else
|
|
value = lowJSValue(m_node->child1());
|
|
|
|
LBasicBlock isCell = m_out.newBlock();
|
|
LBasicBlock notString = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue isCellPredicate;
|
|
if (m_node->child1().useKind() == CellUse)
|
|
isCellPredicate = m_out.booleanTrue;
|
|
else if (m_node->child1().useKind() == NotCellUse)
|
|
isCellPredicate = m_out.booleanFalse;
|
|
else
|
|
isCellPredicate = this->isCell(value, provenType(m_node->child1()));
|
|
m_out.branch(isCellPredicate, unsure(isCell), unsure(notString));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isCell, notString);
|
|
ValueFromBlock simpleResult = m_out.anchor(value);
|
|
LValue isStringPredicate;
|
|
if (m_node->child1()->prediction() & SpecString) {
|
|
isStringPredicate = isString(value, provenType(m_node->child1()));
|
|
} else
|
|
isStringPredicate = m_out.booleanFalse;
|
|
m_out.branch(isStringPredicate, unsure(continuation), unsure(notString));
|
|
|
|
m_out.appendTo(notString, continuation);
|
|
LValue result;
|
|
if (m_node->child1().useKind() == CellUse) {
|
|
ASSERT(m_node->op() != StringValueOf);
|
|
result = vmCall(Int64, m_node->op() == ToString ? operationToStringOnCell : operationCallStringConstructorOnCell, weakPointer(globalObject), value);
|
|
} else {
|
|
auto* operation = m_node->op() == ToString
|
|
? operationToString : m_node->op() == StringValueOf
|
|
? operationStringValueOf : operationCallStringConstructor;
|
|
result = vmCall(Int64, operation, weakPointer(globalObject), value);
|
|
}
|
|
ValueFromBlock convertedResult = m_out.anchor(result);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, simpleResult, convertedResult));
|
|
return;
|
|
}
|
|
|
|
case Int32Use:
|
|
setJSValue(vmCall(Int64, operationInt32ToStringWithValidRadix, weakPointer(globalObject), lowInt32(m_node->child1()), m_out.constInt32(10)));
|
|
return;
|
|
|
|
case Int52RepUse:
|
|
setJSValue(vmCall(Int64, operationInt52ToStringWithValidRadix, weakPointer(globalObject), lowStrictInt52(m_node->child1()), m_out.constInt32(10)));
|
|
return;
|
|
|
|
case DoubleRepUse:
|
|
setJSValue(vmCall(Int64, operationDoubleToStringWithValidRadix, weakPointer(globalObject), lowDouble(m_node->child1()), m_out.constInt32(10)));
|
|
return;
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
break;
|
|
}
|
|
}
|
|
|
|
void compileToPrimitive()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue value = lowJSValue(m_node->child1());
|
|
|
|
LBasicBlock isCellCase = m_out.newBlock();
|
|
LBasicBlock isObjectCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
Vector<ValueFromBlock, 3> results;
|
|
|
|
results.append(m_out.anchor(value));
|
|
m_out.branch(
|
|
isCell(value, provenType(m_node->child1())), unsure(isCellCase), unsure(continuation));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isCellCase, isObjectCase);
|
|
results.append(m_out.anchor(value));
|
|
m_out.branch(
|
|
isObject(value, provenType(m_node->child1())),
|
|
unsure(isObjectCase), unsure(continuation));
|
|
|
|
m_out.appendTo(isObjectCase, continuation);
|
|
results.append(m_out.anchor(vmCall(
|
|
Int64, operationToPrimitive, weakPointer(globalObject), value)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, results));
|
|
}
|
|
|
|
void compileToPropertyKey()
|
|
{
|
|
ASSERT(m_node->child1().useKind() == UntypedUse);
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue value = lowJSValue(m_node->child1());
|
|
|
|
LBasicBlock isCellCase = m_out.newBlock();
|
|
LBasicBlock notStringCase = m_out.newBlock();
|
|
LBasicBlock slowPathCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
Vector<ValueFromBlock, 3> results;
|
|
m_out.branch(
|
|
isCell(value, provenType(m_node->child1())), unsure(isCellCase), unsure(slowPathCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isCellCase, notStringCase);
|
|
results.append(m_out.anchor(value));
|
|
m_out.branch(isString(value, provenType(m_node->child1())), unsure(continuation), unsure(notStringCase));
|
|
|
|
m_out.appendTo(notStringCase, slowPathCase);
|
|
results.append(m_out.anchor(value));
|
|
m_out.branch(isSymbol(value, provenType(m_node->child1())), unsure(continuation), unsure(slowPathCase));
|
|
|
|
m_out.appendTo(slowPathCase, continuation);
|
|
results.append(m_out.anchor(vmCall(
|
|
Int64, operationToPropertyKey, weakPointer(globalObject), value)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, results));
|
|
}
|
|
|
|
void compileMakeRope()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
|
|
struct FlagsAndLength {
|
|
LValue flags;
|
|
LValue length;
|
|
};
|
|
|
|
Edge edges[3] = {
|
|
m_node->child1(),
|
|
m_node->child2(),
|
|
m_node->child3(),
|
|
};
|
|
LValue kids[3];
|
|
unsigned numKids;
|
|
kids[0] = lowCell(edges[0]);
|
|
kids[1] = lowCell(edges[1]);
|
|
if (edges[2]) {
|
|
kids[2] = lowCell(edges[2]);
|
|
numKids = 3;
|
|
} else {
|
|
kids[2] = nullptr;
|
|
numKids = 2;
|
|
}
|
|
|
|
LBasicBlock emptyCase = m_out.newBlock();
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
Allocator allocator = allocatorForNonVirtualConcurrently<JSRopeString>(vm(), sizeof(JSRopeString), AllocatorForMode::AllocatorIfExists);
|
|
|
|
LValue result = allocateCell(
|
|
m_out.constIntPtr(allocator.localAllocator()), vm().stringStructure.get(), slowPath);
|
|
|
|
// This puts nullptr for the first fiber. It makes visitChildren safe even if this JSRopeString is discarded due to the speculation failure in the following path.
|
|
m_out.storePtr(m_out.constIntPtr(JSString::isRopeInPointer), result, m_heaps.JSRopeString_fiber0);
|
|
|
|
auto getFlagsAndLength = [&] (Edge& edge, LValue child) {
|
|
if (JSString* string = edge->dynamicCastConstant<JSString*>(vm())) {
|
|
return FlagsAndLength {
|
|
m_out.constInt32(string->is8Bit() ? StringImpl::flagIs8Bit() : 0),
|
|
m_out.constInt32(string->length())
|
|
};
|
|
}
|
|
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
LBasicBlock ropeCase = m_out.newBlock();
|
|
LBasicBlock notRopeCase = m_out.newBlock();
|
|
|
|
m_out.branch(isRopeString(child, edge), unsure(ropeCase), unsure(notRopeCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(ropeCase, notRopeCase);
|
|
ValueFromBlock flagsForRope = m_out.anchor(m_out.load32NonNegative(child, m_heaps.JSRopeString_flags));
|
|
ValueFromBlock lengthForRope = m_out.anchor(m_out.load32NonNegative(child, m_heaps.JSRopeString_length));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(notRopeCase, continuation);
|
|
LValue stringImpl = m_out.loadPtr(child, m_heaps.JSString_value);
|
|
ValueFromBlock flagsForNonRope = m_out.anchor(m_out.load32NonNegative(stringImpl, m_heaps.StringImpl_hashAndFlags));
|
|
ValueFromBlock lengthForNonRope = m_out.anchor(m_out.load32NonNegative(stringImpl, m_heaps.StringImpl_length));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
return FlagsAndLength {
|
|
m_out.phi(Int32, flagsForRope, flagsForNonRope),
|
|
m_out.phi(Int32, lengthForRope, lengthForNonRope)
|
|
};
|
|
};
|
|
|
|
FlagsAndLength flagsAndLength = getFlagsAndLength(edges[0], kids[0]);
|
|
for (unsigned i = 1; i < numKids; ++i) {
|
|
auto mergeFlagsAndLength = [&] (Edge& edge, LValue child, FlagsAndLength previousFlagsAndLength) {
|
|
FlagsAndLength flagsAndLength = getFlagsAndLength(edge, child);
|
|
LValue flags = m_out.bitAnd(previousFlagsAndLength.flags, flagsAndLength.flags);
|
|
CheckValue* lengthCheck = m_out.speculateAdd(previousFlagsAndLength.length, flagsAndLength.length);
|
|
blessSpeculation(lengthCheck, Uncountable, noValue(), nullptr, m_origin);
|
|
return FlagsAndLength {
|
|
flags,
|
|
lengthCheck
|
|
};
|
|
};
|
|
flagsAndLength = mergeFlagsAndLength(edges[i], kids[i], flagsAndLength);
|
|
}
|
|
|
|
m_out.storePtr(
|
|
m_out.bitOr(
|
|
m_out.bitOr(kids[0], m_out.constIntPtr(JSString::isRopeInPointer)),
|
|
m_out.bitAnd(m_out.constIntPtr(JSRopeString::is8BitInPointer), m_out.zeroExtPtr(flagsAndLength.flags))),
|
|
result, m_heaps.JSRopeString_fiber0);
|
|
m_out.storePtr(
|
|
m_out.bitOr(m_out.zeroExtPtr(flagsAndLength.length), m_out.shl(kids[1], m_out.constInt32(32))),
|
|
result, m_heaps.JSRopeString_fiber1);
|
|
if (numKids == 2)
|
|
m_out.storePtr(m_out.lShr(kids[1], m_out.constInt32(32)), result, m_heaps.JSRopeString_fiber2);
|
|
else
|
|
m_out.storePtr(m_out.bitOr(m_out.lShr(kids[1], m_out.constInt32(32)), m_out.shl(kids[2], m_out.constInt32(16))), result, m_heaps.JSRopeString_fiber2);
|
|
|
|
mutatorFence();
|
|
ValueFromBlock fastResult = m_out.anchor(result);
|
|
m_out.branch(m_out.isZero32(flagsAndLength.length), rarely(emptyCase), usually(continuation));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(emptyCase, slowPath);
|
|
ValueFromBlock emptyResult = m_out.anchor(weakPointer(jsEmptyString(m_graph.m_vm)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
LValue slowResultValue;
|
|
VM& vm = this->vm();
|
|
switch (numKids) {
|
|
case 2:
|
|
slowResultValue = lazySlowPath(
|
|
[=, &vm] (const Vector<Location>& locations) -> RefPtr<LazySlowPath::Generator> {
|
|
return createLazyCallGenerator(vm,
|
|
operationMakeRope2, locations[0].directGPR(), globalObject, locations[1].directGPR(),
|
|
locations[2].directGPR());
|
|
}, kids[0], kids[1]);
|
|
break;
|
|
case 3:
|
|
slowResultValue = lazySlowPath(
|
|
[=, &vm] (const Vector<Location>& locations) -> RefPtr<LazySlowPath::Generator> {
|
|
return createLazyCallGenerator(vm,
|
|
operationMakeRope3, locations[0].directGPR(), globalObject, locations[1].directGPR(),
|
|
locations[2].directGPR(), locations[3].directGPR());
|
|
}, kids[0], kids[1], kids[2]);
|
|
break;
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad number of children");
|
|
break;
|
|
}
|
|
ValueFromBlock slowResult = m_out.anchor(slowResultValue);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, fastResult, emptyResult, slowResult));
|
|
}
|
|
|
|
void compileStringCharAt()
|
|
{
|
|
LValue base = lowString(m_graph.child(m_node, 0));
|
|
LValue index = lowInt32(m_graph.child(m_node, 1));
|
|
LValue storage = lowStorage(m_graph.child(m_node, 2));
|
|
|
|
LBasicBlock fastPath = m_out.newBlock();
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue stringImpl = m_out.loadPtr(base, m_heaps.JSString_value);
|
|
m_out.branch(
|
|
m_out.aboveOrEqual(
|
|
index, m_out.load32NonNegative(stringImpl, m_heaps.StringImpl_length)),
|
|
rarely(slowPath), usually(fastPath));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(fastPath, slowPath);
|
|
|
|
LBasicBlock is8Bit = m_out.newBlock();
|
|
LBasicBlock is16Bit = m_out.newBlock();
|
|
LBasicBlock bitsContinuation = m_out.newBlock();
|
|
LBasicBlock bigCharacter = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
m_out.testIsZero32(
|
|
m_out.load32(stringImpl, m_heaps.StringImpl_hashAndFlags),
|
|
m_out.constInt32(StringImpl::flagIs8Bit())),
|
|
unsure(is16Bit), unsure(is8Bit));
|
|
|
|
m_out.appendTo(is8Bit, is16Bit);
|
|
|
|
// FIXME: Need to cage strings!
|
|
// https://bugs.webkit.org/show_bug.cgi?id=174924
|
|
ValueFromBlock char8Bit = m_out.anchor(
|
|
m_out.load8ZeroExt32(m_out.baseIndex(
|
|
m_heaps.characters8, storage, m_out.zeroExtPtr(index),
|
|
provenValue(m_graph.child(m_node, 1)))));
|
|
m_out.jump(bitsContinuation);
|
|
|
|
m_out.appendTo(is16Bit, bigCharacter);
|
|
|
|
LValue char16BitValue = m_out.load16ZeroExt32(
|
|
m_out.baseIndex(
|
|
m_heaps.characters16, storage, m_out.zeroExtPtr(index),
|
|
provenValue(m_graph.child(m_node, 1))));
|
|
ValueFromBlock char16Bit = m_out.anchor(char16BitValue);
|
|
m_out.branch(
|
|
m_out.above(char16BitValue, m_out.constInt32(maxSingleCharacterString)),
|
|
rarely(bigCharacter), usually(bitsContinuation));
|
|
|
|
m_out.appendTo(bigCharacter, bitsContinuation);
|
|
|
|
Vector<ValueFromBlock, 4> results;
|
|
results.append(m_out.anchor(vmCall(
|
|
Int64, operationSingleCharacterString,
|
|
m_vmValue, char16BitValue)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(bitsContinuation, slowPath);
|
|
|
|
LValue character = m_out.phi(Int32, char8Bit, char16Bit);
|
|
|
|
LValue smallStrings = m_out.constIntPtr(vm().smallStrings.singleCharacterStrings());
|
|
|
|
results.append(m_out.anchor(m_out.loadPtr(m_out.baseIndex(
|
|
m_heaps.singleCharacterStrings, smallStrings, m_out.zeroExtPtr(character)))));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
|
|
if (m_node->arrayMode().isInBounds()) {
|
|
speculate(OutOfBounds, noValue(), nullptr, m_out.booleanTrue);
|
|
results.append(m_out.anchor(m_out.intPtrZero));
|
|
} else {
|
|
// FIXME: Revisit JSGlobalObject.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=203204
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
Structure* stringPrototypeStructure = globalObject->stringPrototype()->structure(vm());
|
|
Structure* objectPrototypeStructure = globalObject->objectPrototype()->structure(vm());
|
|
WTF::loadLoadFence();
|
|
|
|
if (globalObject->stringPrototypeChainIsSane()) {
|
|
// FIXME: This could be captured using a Speculation mode that means
|
|
// "out-of-bounds loads return a trivial value", something like
|
|
// OutOfBoundsSaneChain.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=144668
|
|
|
|
m_graph.registerAndWatchStructureTransition(stringPrototypeStructure);
|
|
m_graph.registerAndWatchStructureTransition(objectPrototypeStructure);
|
|
|
|
LBasicBlock negativeIndex = m_out.newBlock();
|
|
|
|
results.append(m_out.anchor(m_out.constInt64(JSValue::encode(jsUndefined()))));
|
|
m_out.branch(
|
|
m_out.lessThan(index, m_out.int32Zero),
|
|
rarely(negativeIndex), usually(continuation));
|
|
|
|
m_out.appendTo(negativeIndex, continuation);
|
|
}
|
|
|
|
results.append(m_out.anchor(vmCall(Int64, operationGetByValStringInt, weakPointer(globalObject), base, index)));
|
|
}
|
|
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
// We have to keep base alive since that keeps storage alive.
|
|
ensureStillAliveHere(base);
|
|
setJSValue(m_out.phi(Int64, results));
|
|
}
|
|
|
|
void compileStringCharCodeAt()
|
|
{
|
|
LBasicBlock is8Bit = m_out.newBlock();
|
|
LBasicBlock is16Bit = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue base = lowString(m_node->child1());
|
|
LValue index = lowInt32(m_node->child2());
|
|
LValue storage = lowStorage(m_node->child3());
|
|
|
|
LValue stringImpl = m_out.loadPtr(base, m_heaps.JSString_value);
|
|
|
|
speculate(
|
|
Uncountable, noValue(), nullptr,
|
|
m_out.aboveOrEqual(
|
|
index, m_out.load32NonNegative(stringImpl, m_heaps.StringImpl_length)));
|
|
|
|
m_out.branch(
|
|
m_out.testIsZero32(
|
|
m_out.load32(stringImpl, m_heaps.StringImpl_hashAndFlags),
|
|
m_out.constInt32(StringImpl::flagIs8Bit())),
|
|
unsure(is16Bit), unsure(is8Bit));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(is8Bit, is16Bit);
|
|
|
|
// FIXME: need to cage strings!
|
|
// https://bugs.webkit.org/show_bug.cgi?id=174924
|
|
ValueFromBlock char8Bit = m_out.anchor(
|
|
m_out.load8ZeroExt32(m_out.baseIndex(
|
|
m_heaps.characters8, storage, m_out.zeroExtPtr(index),
|
|
provenValue(m_node->child2()))));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(is16Bit, continuation);
|
|
|
|
ValueFromBlock char16Bit = m_out.anchor(
|
|
m_out.load16ZeroExt32(m_out.baseIndex(
|
|
m_heaps.characters16, storage, m_out.zeroExtPtr(index),
|
|
provenValue(m_node->child2()))));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
|
|
// We have to keep base alive since that keeps storage alive.
|
|
ensureStillAliveHere(base);
|
|
setInt32(m_out.phi(Int32, char8Bit, char16Bit));
|
|
}
|
|
|
|
void compileStringCodePointAt()
|
|
{
|
|
LBasicBlock is8Bit = m_out.newBlock();
|
|
LBasicBlock is16Bit = m_out.newBlock();
|
|
LBasicBlock isLeadSurrogate = m_out.newBlock();
|
|
LBasicBlock mayHaveTrailSurrogate = m_out.newBlock();
|
|
LBasicBlock hasTrailSurrogate = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue base = lowString(m_node->child1());
|
|
LValue index = lowInt32(m_node->child2());
|
|
LValue storage = lowStorage(m_node->child3());
|
|
|
|
LValue stringImpl = m_out.loadPtr(base, m_heaps.JSString_value);
|
|
LValue length = m_out.load32NonNegative(stringImpl, m_heaps.StringImpl_length);
|
|
|
|
speculate(Uncountable, noValue(), nullptr, m_out.aboveOrEqual(index, length));
|
|
|
|
m_out.branch(
|
|
m_out.testIsZero32(
|
|
m_out.load32(stringImpl, m_heaps.StringImpl_hashAndFlags),
|
|
m_out.constInt32(StringImpl::flagIs8Bit())),
|
|
unsure(is16Bit), unsure(is8Bit));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(is8Bit, is16Bit);
|
|
// FIXME: Need to cage strings!
|
|
// https://bugs.webkit.org/show_bug.cgi?id=174924
|
|
ValueFromBlock char8Bit = m_out.anchor(
|
|
m_out.load8ZeroExt32(m_out.baseIndex(
|
|
m_heaps.characters8, storage, m_out.zeroExtPtr(index),
|
|
provenValue(m_node->child2()))));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(is16Bit, isLeadSurrogate);
|
|
LValue leadCharacter = m_out.load16ZeroExt32(m_out.baseIndex(m_heaps.characters16, storage, m_out.zeroExtPtr(index), provenValue(m_node->child2())));
|
|
ValueFromBlock char16Bit = m_out.anchor(leadCharacter);
|
|
LValue nextIndex = m_out.add(index, m_out.int32One);
|
|
m_out.branch(m_out.aboveOrEqual(nextIndex, length), unsure(continuation), unsure(isLeadSurrogate));
|
|
|
|
m_out.appendTo(isLeadSurrogate, mayHaveTrailSurrogate);
|
|
m_out.branch(m_out.notEqual(m_out.bitAnd(leadCharacter, m_out.constInt32(0xfffffc00)), m_out.constInt32(0xd800)), unsure(continuation), unsure(mayHaveTrailSurrogate));
|
|
|
|
m_out.appendTo(mayHaveTrailSurrogate, hasTrailSurrogate);
|
|
JSValue indexValue = provenValue(m_node->child2());
|
|
JSValue nextIndexValue;
|
|
if (indexValue && indexValue.isInt32() && indexValue.asInt32() != INT32_MAX)
|
|
nextIndexValue = jsNumber(indexValue.asInt32() + 1);
|
|
LValue trailCharacter = m_out.load16ZeroExt32(m_out.baseIndex(m_heaps.characters16, storage, m_out.zeroExtPtr(nextIndex), nextIndexValue));
|
|
m_out.branch(m_out.notEqual(m_out.bitAnd(trailCharacter, m_out.constInt32(0xfffffc00)), m_out.constInt32(0xdc00)), unsure(continuation), unsure(hasTrailSurrogate));
|
|
|
|
m_out.appendTo(hasTrailSurrogate, continuation);
|
|
ValueFromBlock charSurrogatePair = m_out.anchor(m_out.sub(m_out.add(m_out.shl(leadCharacter, m_out.constInt32(10)), trailCharacter), m_out.constInt32(U16_SURROGATE_OFFSET)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
// We have to keep base alive since that keeps storage alive.
|
|
ensureStillAliveHere(base);
|
|
setInt32(m_out.phi(Int32, char8Bit, char16Bit, charSurrogatePair));
|
|
}
|
|
|
|
void compileStringFromCharCode()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
Edge childEdge = m_node->child1();
|
|
|
|
if (childEdge.useKind() == UntypedUse) {
|
|
LValue result = vmCall(
|
|
Int64, operationStringFromCharCodeUntyped, weakPointer(globalObject),
|
|
lowJSValue(childEdge));
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
|
|
DFG_ASSERT(m_graph, m_node, childEdge.useKind() == Int32Use, childEdge.useKind());
|
|
|
|
LValue value = lowInt32(childEdge);
|
|
|
|
LBasicBlock smallIntCase = m_out.newBlock();
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
m_out.above(value, m_out.constInt32(maxSingleCharacterString)),
|
|
rarely(slowCase), usually(smallIntCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(smallIntCase, slowCase);
|
|
|
|
LValue smallStrings = m_out.constIntPtr(vm().smallStrings.singleCharacterStrings());
|
|
LValue fastResultValue = m_out.loadPtr(
|
|
m_out.baseIndex(m_heaps.singleCharacterStrings, smallStrings, m_out.zeroExtPtr(value)));
|
|
ValueFromBlock fastResult = m_out.anchor(fastResultValue);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
|
|
LValue slowResultValue = vmCall(
|
|
pointerType(), operationStringFromCharCode, weakPointer(globalObject), value);
|
|
ValueFromBlock slowResult = m_out.anchor(slowResultValue);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
|
|
setJSValue(m_out.phi(Int64, fastResult, slowResult));
|
|
}
|
|
|
|
void compileGetByOffset()
|
|
{
|
|
StorageAccessData& data = m_node->storageAccessData();
|
|
|
|
LValue base = lowCell(m_node->child2());
|
|
LValue value = loadProperty(lowStorage(m_node->child1()), data.identifierNumber, data.offset);
|
|
// We have to keep base alive since that keeps content of storage alive.
|
|
ensureStillAliveHere(base);
|
|
setJSValue(value);
|
|
}
|
|
|
|
void compileGetGetter()
|
|
{
|
|
setJSValue(m_out.loadPtr(lowCell(m_node->child1()), m_heaps.GetterSetter_getter));
|
|
}
|
|
|
|
void compileGetSetter()
|
|
{
|
|
setJSValue(m_out.loadPtr(lowCell(m_node->child1()), m_heaps.GetterSetter_setter));
|
|
}
|
|
|
|
void compileMultiGetByOffset()
|
|
{
|
|
LValue base = lowCell(m_node->child1());
|
|
|
|
MultiGetByOffsetData& data = m_node->multiGetByOffsetData();
|
|
|
|
Vector<LBasicBlock, 2> blocks(data.cases.size());
|
|
for (unsigned i = data.cases.size(); i--;)
|
|
blocks[i] = m_out.newBlock();
|
|
LBasicBlock exit = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
Vector<SwitchCase, 2> cases;
|
|
RegisteredStructureSet baseSet;
|
|
for (unsigned i = data.cases.size(); i--;) {
|
|
MultiGetByOffsetCase getCase = data.cases[i];
|
|
for (unsigned j = getCase.set().size(); j--;) {
|
|
RegisteredStructure structure = getCase.set()[j];
|
|
baseSet.add(structure);
|
|
cases.append(SwitchCase(weakStructureID(structure), blocks[i], Weight(1)));
|
|
}
|
|
}
|
|
bool structuresChecked = m_interpreter.forNode(m_node->child1()).m_structure.isSubsetOf(baseSet);
|
|
emitSwitchForMultiByOffset(base, structuresChecked, cases, exit);
|
|
|
|
LBasicBlock lastNext = m_out.m_nextBlock;
|
|
|
|
Vector<ValueFromBlock, 2> results;
|
|
for (unsigned i = data.cases.size(); i--;) {
|
|
MultiGetByOffsetCase getCase = data.cases[i];
|
|
GetByOffsetMethod method = getCase.method();
|
|
|
|
m_out.appendTo(blocks[i], i + 1 < data.cases.size() ? blocks[i + 1] : exit);
|
|
|
|
LValue result;
|
|
|
|
switch (method.kind()) {
|
|
case GetByOffsetMethod::Invalid:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
break;
|
|
|
|
case GetByOffsetMethod::Constant:
|
|
result = m_out.constInt64(JSValue::encode(method.constant()->value()));
|
|
break;
|
|
|
|
case GetByOffsetMethod::Load:
|
|
case GetByOffsetMethod::LoadFromPrototype: {
|
|
LValue propertyBase;
|
|
if (method.kind() == GetByOffsetMethod::Load)
|
|
propertyBase = base;
|
|
else
|
|
propertyBase = weakPointer(method.prototype()->value().asCell());
|
|
if (!isInlineOffset(method.offset()))
|
|
propertyBase = m_out.loadPtr(propertyBase, m_heaps.JSObject_butterfly);
|
|
result = loadProperty(
|
|
propertyBase, data.identifierNumber, method.offset());
|
|
break;
|
|
} }
|
|
|
|
results.append(m_out.anchor(result));
|
|
m_out.jump(continuation);
|
|
}
|
|
|
|
m_out.appendTo(exit, continuation);
|
|
if (!structuresChecked)
|
|
speculate(BadCache, noValue(), nullptr, m_out.booleanTrue);
|
|
m_out.unreachable();
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
// We have to keep base alive since that keeps storage alive.
|
|
ensureStillAliveHere(base);
|
|
setJSValue(m_out.phi(Int64, results));
|
|
}
|
|
|
|
void compilePutByOffset()
|
|
{
|
|
StorageAccessData& data = m_node->storageAccessData();
|
|
|
|
storeProperty(
|
|
lowJSValue(m_node->child3()),
|
|
lowStorage(m_node->child1()), data.identifierNumber, data.offset);
|
|
}
|
|
|
|
void compileMultiPutByOffset()
|
|
{
|
|
LValue base = lowCell(m_node->child1());
|
|
LValue value = lowJSValue(m_node->child2());
|
|
|
|
MultiPutByOffsetData& data = m_node->multiPutByOffsetData();
|
|
|
|
Vector<LBasicBlock, 2> blocks(data.variants.size());
|
|
for (unsigned i = data.variants.size(); i--;)
|
|
blocks[i] = m_out.newBlock();
|
|
LBasicBlock exit = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
Vector<SwitchCase, 2> cases;
|
|
RegisteredStructureSet baseSet;
|
|
for (unsigned i = data.variants.size(); i--;) {
|
|
PutByIdVariant variant = data.variants[i];
|
|
for (unsigned j = variant.oldStructure().size(); j--;) {
|
|
RegisteredStructure structure = m_graph.registerStructure(variant.oldStructure()[j]);
|
|
baseSet.add(structure);
|
|
cases.append(SwitchCase(weakStructureID(structure), blocks[i], Weight(1)));
|
|
}
|
|
}
|
|
bool structuresChecked = m_interpreter.forNode(m_node->child1()).m_structure.isSubsetOf(baseSet);
|
|
emitSwitchForMultiByOffset(base, structuresChecked, cases, exit);
|
|
|
|
LBasicBlock lastNext = m_out.m_nextBlock;
|
|
|
|
for (unsigned i = data.variants.size(); i--;) {
|
|
m_out.appendTo(blocks[i], i + 1 < data.variants.size() ? blocks[i + 1] : exit);
|
|
|
|
PutByIdVariant variant = data.variants[i];
|
|
|
|
LValue storage;
|
|
if (variant.kind() == PutByIdVariant::Replace) {
|
|
if (isInlineOffset(variant.offset()))
|
|
storage = base;
|
|
else
|
|
storage = m_out.loadPtr(base, m_heaps.JSObject_butterfly);
|
|
} else {
|
|
DFG_ASSERT(m_graph, m_node, variant.kind() == PutByIdVariant::Transition, variant.kind());
|
|
m_graph.m_plan.transitions().addLazily(
|
|
codeBlock(), m_origin.semantic.codeOriginOwner(),
|
|
variant.oldStructureForTransition(), variant.newStructure());
|
|
|
|
storage = storageForTransition(
|
|
base, variant.offset(),
|
|
variant.oldStructureForTransition(), variant.newStructure());
|
|
}
|
|
|
|
storeProperty(value, storage, data.identifierNumber, variant.offset());
|
|
|
|
if (variant.kind() == PutByIdVariant::Transition) {
|
|
ASSERT(variant.oldStructureForTransition()->indexingType() == variant.newStructure()->indexingType());
|
|
ASSERT(variant.oldStructureForTransition()->typeInfo().inlineTypeFlags() == variant.newStructure()->typeInfo().inlineTypeFlags());
|
|
ASSERT(variant.oldStructureForTransition()->typeInfo().type() == variant.newStructure()->typeInfo().type());
|
|
m_out.store32(
|
|
weakStructureID(m_graph.registerStructure(variant.newStructure())), base, m_heaps.JSCell_structureID);
|
|
}
|
|
|
|
m_out.jump(continuation);
|
|
}
|
|
|
|
m_out.appendTo(exit, continuation);
|
|
if (!structuresChecked)
|
|
speculate(BadCache, noValue(), nullptr, m_out.booleanTrue);
|
|
m_out.unreachable();
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
}
|
|
|
|
void compileMultiDeleteByOffset()
|
|
{
|
|
LValue base = lowCell(m_node->child1());
|
|
MultiDeleteByOffsetData& data = m_node->multiDeleteByOffsetData();
|
|
|
|
unsigned missConfigurable = 0;
|
|
unsigned missNonconfigurable = 0;
|
|
|
|
for (unsigned i = data.variants.size(); i--;) {
|
|
DeleteByIdVariant variant = data.variants[i];
|
|
if (!variant.newStructure()) {
|
|
if (variant.result())
|
|
++missConfigurable;
|
|
else
|
|
++missNonconfigurable;
|
|
}
|
|
}
|
|
|
|
unsigned uniqueCaseCount = data.variants.size();
|
|
if (missConfigurable)
|
|
uniqueCaseCount -= missConfigurable - 1;
|
|
if (missNonconfigurable)
|
|
uniqueCaseCount -= missNonconfigurable - 1;
|
|
int trueBlock = missConfigurable ? uniqueCaseCount - 1 : -1;
|
|
int falseBlock = missNonconfigurable ? uniqueCaseCount - 1 - !!missConfigurable : -1;
|
|
|
|
Vector<LBasicBlock, 2> blocks(uniqueCaseCount);
|
|
for (unsigned i = blocks.size(); i--;)
|
|
blocks[i] = m_out.newBlock();
|
|
LBasicBlock exit = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
Vector<SwitchCase, 2> cases;
|
|
RegisteredStructureSet baseSet;
|
|
for (unsigned i = data.variants.size(), block = 0; i--;) {
|
|
DeleteByIdVariant variant = data.variants[i];
|
|
RegisteredStructure structure = m_graph.registerStructure(variant.oldStructure());
|
|
baseSet.add(structure);
|
|
if (variant.newStructure())
|
|
cases.append(SwitchCase(weakStructureID(structure), blocks[block++], Weight(1)));
|
|
else
|
|
cases.append(SwitchCase(weakStructureID(structure), blocks[variant.result() ? trueBlock : falseBlock], Weight(1)));
|
|
}
|
|
bool structuresChecked = m_interpreter.forNode(m_node->child1()).m_structure.isSubsetOf(baseSet);
|
|
emitSwitchForMultiByOffset(base, structuresChecked, cases, exit);
|
|
|
|
LBasicBlock lastNext = m_out.m_nextBlock;
|
|
|
|
Vector<ValueFromBlock, 2> results;
|
|
|
|
for (unsigned i = data.variants.size(), block = 0; i--;) {
|
|
DeleteByIdVariant variant = data.variants[i];
|
|
if (!variant.newStructure())
|
|
continue;
|
|
|
|
m_out.appendTo(blocks[block], block + 1 < blocks.size() ? blocks[block + 1] : exit);
|
|
|
|
LValue storage;
|
|
|
|
if (isInlineOffset(variant.offset()))
|
|
storage = base;
|
|
else
|
|
storage = m_out.loadPtr(base, m_heaps.JSObject_butterfly);
|
|
|
|
storeProperty(m_out.int64Zero, storage, data.identifierNumber, variant.offset());
|
|
|
|
ASSERT(variant.oldStructure()->indexingType() == variant.newStructure()->indexingType());
|
|
ASSERT(variant.oldStructure()->typeInfo().inlineTypeFlags() == variant.newStructure()->typeInfo().inlineTypeFlags());
|
|
ASSERT(variant.oldStructure()->typeInfo().type() == variant.newStructure()->typeInfo().type());
|
|
m_out.store32(
|
|
weakStructureID(m_graph.registerStructure(variant.newStructure())), base, m_heaps.JSCell_structureID);
|
|
|
|
results.append(m_out.anchor(variant.result() ? m_out.booleanTrue : m_out.booleanFalse));
|
|
m_out.jump(continuation);
|
|
++block;
|
|
}
|
|
|
|
if (missNonconfigurable) {
|
|
m_out.appendTo(blocks[falseBlock]);
|
|
results.append(m_out.anchor(m_out.booleanFalse));
|
|
m_out.jump(continuation);
|
|
}
|
|
|
|
if (missConfigurable) {
|
|
m_out.appendTo(blocks[trueBlock], exit);
|
|
results.append(m_out.anchor(m_out.booleanTrue));
|
|
m_out.jump(continuation);
|
|
}
|
|
|
|
m_out.appendTo(exit, continuation);
|
|
if (!structuresChecked)
|
|
speculate(BadCache, noValue(), nullptr, m_out.booleanTrue);
|
|
m_out.unreachable();
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setBoolean(m_out.phi(Int32, results));
|
|
|
|
if (data.writesStructures()) {
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Void);
|
|
patchpoint->setGenerator([] (CCallHelpers&, const StackmapGenerationParams&) { });
|
|
m_heaps.decoratePatchpointWrite(&m_heaps.JSCellHeaderAndNamedProperties, patchpoint);
|
|
}
|
|
}
|
|
|
|
void compileMatchStructure()
|
|
{
|
|
LValue base = lowCell(m_node->child1());
|
|
|
|
MatchStructureData& data = m_node->matchStructureData();
|
|
|
|
LBasicBlock trueBlock = m_out.newBlock();
|
|
LBasicBlock falseBlock = m_out.newBlock();
|
|
LBasicBlock exitBlock = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(trueBlock);
|
|
|
|
Vector<SwitchCase, 2> cases;
|
|
RegisteredStructureSet baseSet;
|
|
for (MatchStructureVariant& variant : data.variants) {
|
|
baseSet.add(variant.structure);
|
|
cases.append(SwitchCase(
|
|
weakStructureID(variant.structure),
|
|
variant.result ? trueBlock : falseBlock, Weight(1)));
|
|
}
|
|
bool structuresChecked = m_interpreter.forNode(m_node->child1()).m_structure.isSubsetOf(baseSet);
|
|
emitSwitchForMultiByOffset(base, structuresChecked, cases, exitBlock);
|
|
|
|
m_out.appendTo(trueBlock, falseBlock);
|
|
ValueFromBlock trueResult = m_out.anchor(m_out.booleanTrue);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(falseBlock, exitBlock);
|
|
ValueFromBlock falseResult = m_out.anchor(m_out.booleanFalse);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(exitBlock, continuation);
|
|
if (!structuresChecked)
|
|
speculate(BadCache, noValue(), nullptr, m_out.booleanTrue);
|
|
m_out.unreachable();
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setBoolean(m_out.phi(Int32, trueResult, falseResult));
|
|
}
|
|
|
|
void compileGetGlobalVariable()
|
|
{
|
|
setJSValue(m_out.load64(m_out.absolute(m_node->variablePointer())));
|
|
}
|
|
|
|
void compilePutGlobalVariable()
|
|
{
|
|
m_out.store64(
|
|
lowJSValue(m_node->child2()), m_out.absolute(m_node->variablePointer()));
|
|
}
|
|
|
|
void compileNotifyWrite()
|
|
{
|
|
WatchpointSet* set = m_node->watchpointSet();
|
|
|
|
LBasicBlock isNotInvalidated = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue state = m_out.load8ZeroExt32(m_out.absolute(set->addressOfState()));
|
|
m_out.branch(
|
|
m_out.equal(state, m_out.constInt32(IsInvalidated)),
|
|
usually(continuation), rarely(isNotInvalidated));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isNotInvalidated, continuation);
|
|
|
|
VM& vm = this->vm();
|
|
lazySlowPath(
|
|
[=, &vm] (const Vector<Location>&) -> RefPtr<LazySlowPath::Generator> {
|
|
return createLazyCallGenerator(vm,
|
|
operationNotifyWrite, InvalidGPRReg, &vm, CCallHelpers::TrustedImmPtr(set));
|
|
});
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
}
|
|
|
|
void compileGetCallee()
|
|
{
|
|
setJSValue(m_out.loadPtr(addressFor(VirtualRegister(CallFrameSlot::callee))));
|
|
}
|
|
|
|
void compileSetCallee()
|
|
{
|
|
auto callee = lowCell(m_node->child1());
|
|
m_out.storePtr(callee, payloadFor(VirtualRegister(CallFrameSlot::callee)));
|
|
}
|
|
|
|
void compileGetArgumentCountIncludingThis()
|
|
{
|
|
VirtualRegister argumentCountRegister;
|
|
if (InlineCallFrame* inlineCallFrame = m_node->argumentsInlineCallFrame())
|
|
argumentCountRegister = inlineCallFrame->argumentCountRegister;
|
|
else
|
|
argumentCountRegister = VirtualRegister(CallFrameSlot::argumentCountIncludingThis);
|
|
setInt32(m_out.load32(payloadFor(argumentCountRegister)));
|
|
}
|
|
|
|
void compileSetArgumentCountIncludingThis()
|
|
{
|
|
m_out.store32(m_out.constInt32(m_node->argumentCountIncludingThis()), payloadFor(VirtualRegister(CallFrameSlot::argumentCountIncludingThis)));
|
|
}
|
|
|
|
void compileGetScope()
|
|
{
|
|
setJSValue(m_out.loadPtr(lowCell(m_node->child1()), m_heaps.JSFunction_scope));
|
|
}
|
|
|
|
void compileSkipScope()
|
|
{
|
|
setJSValue(m_out.loadPtr(lowCell(m_node->child1()), m_heaps.JSScope_next));
|
|
}
|
|
|
|
void compileGetGlobalObject()
|
|
{
|
|
LValue structure = loadStructure(lowCell(m_node->child1()));
|
|
setJSValue(m_out.loadPtr(structure, m_heaps.Structure_globalObject));
|
|
}
|
|
|
|
void compileGetGlobalThis()
|
|
{
|
|
auto* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
setJSValue(m_out.loadPtr(m_out.absolute(globalObject->addressOfGlobalThis())));
|
|
}
|
|
|
|
void compileGetClosureVar()
|
|
{
|
|
setJSValue(
|
|
m_out.load64(
|
|
lowCell(m_node->child1()),
|
|
m_heaps.JSLexicalEnvironment_variables[m_node->scopeOffset().offset()]));
|
|
}
|
|
|
|
void compilePutClosureVar()
|
|
{
|
|
m_out.store64(
|
|
lowJSValue(m_node->child2()),
|
|
lowCell(m_node->child1()),
|
|
m_heaps.JSLexicalEnvironment_variables[m_node->scopeOffset().offset()]);
|
|
}
|
|
|
|
void compileGetInternalField()
|
|
{
|
|
setJSValue(
|
|
m_out.load64(
|
|
lowCell(m_node->child1()),
|
|
m_heaps.JSInternalFieldObjectImpl_internalFields[m_node->internalFieldIndex()]));
|
|
}
|
|
|
|
void compilePutInternalField()
|
|
{
|
|
m_out.store64(
|
|
lowJSValue(m_node->child2()),
|
|
lowCell(m_node->child1()),
|
|
m_heaps.JSInternalFieldObjectImpl_internalFields[m_node->internalFieldIndex()]);
|
|
}
|
|
|
|
void compileGetFromArguments()
|
|
{
|
|
setJSValue(
|
|
m_out.load64(
|
|
lowCell(m_node->child1()),
|
|
m_heaps.DirectArguments_storage[m_node->capturedArgumentsOffset().offset()]));
|
|
}
|
|
|
|
void compilePutToArguments()
|
|
{
|
|
m_out.store64(
|
|
lowJSValue(m_node->child2()),
|
|
lowCell(m_node->child1()),
|
|
m_heaps.DirectArguments_storage[m_node->capturedArgumentsOffset().offset()]);
|
|
}
|
|
|
|
void compileGetArgument()
|
|
{
|
|
LValue argumentCount = m_out.load32(payloadFor(AssemblyHelpers::argumentCount(m_origin.semantic)));
|
|
|
|
LBasicBlock inBounds = m_out.newBlock();
|
|
LBasicBlock outOfBounds = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(m_out.lessThanOrEqual(argumentCount, m_out.constInt32(m_node->argumentIndex())), unsure(outOfBounds), unsure(inBounds));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(inBounds, outOfBounds);
|
|
VirtualRegister arg = AssemblyHelpers::argumentsStart(m_origin.semantic) + m_node->argumentIndex() - 1;
|
|
ValueFromBlock inBoundsResult = m_out.anchor(m_out.load64(addressFor(arg)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(outOfBounds, continuation);
|
|
ValueFromBlock outOfBoundsResult = m_out.anchor(m_out.constInt64(JSValue::ValueUndefined));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, inBoundsResult, outOfBoundsResult));
|
|
}
|
|
|
|
void compileCompareEq()
|
|
{
|
|
if (m_node->isBinaryUseKind(Int32Use)
|
|
|| m_node->isBinaryUseKind(Int52RepUse)
|
|
|| m_node->isBinaryUseKind(DoubleRepUse)
|
|
|| m_node->isBinaryUseKind(ObjectUse)
|
|
|| m_node->isBinaryUseKind(BooleanUse)
|
|
|| m_node->isBinaryUseKind(SymbolUse)
|
|
|| m_node->isBinaryUseKind(StringIdentUse)
|
|
|| m_node->isBinaryUseKind(StringUse)
|
|
|| m_node->isBinaryUseKind(BigInt32Use)
|
|
|| m_node->isBinaryUseKind(HeapBigIntUse)
|
|
|| m_node->isBinaryUseKind(AnyBigIntUse)) {
|
|
compileCompareStrictEq();
|
|
return;
|
|
}
|
|
|
|
if (m_node->isBinaryUseKind(ObjectUse, ObjectOrOtherUse)) {
|
|
compareEqObjectOrOtherToObject(m_node->child2(), m_node->child1());
|
|
return;
|
|
}
|
|
|
|
if (m_node->isBinaryUseKind(ObjectOrOtherUse, ObjectUse)) {
|
|
compareEqObjectOrOtherToObject(m_node->child1(), m_node->child2());
|
|
return;
|
|
}
|
|
|
|
if (m_node->child1().useKind() == KnownOtherUse) {
|
|
ASSERT(!m_interpreter.needsTypeCheck(m_node->child1(), SpecOther));
|
|
setBoolean(equalNullOrUndefined(m_node->child2(), AllCellsAreFalse, EqualNullOrUndefined, ManualOperandSpeculation));
|
|
return;
|
|
}
|
|
|
|
if (m_node->child2().useKind() == KnownOtherUse) {
|
|
ASSERT(!m_interpreter.needsTypeCheck(m_node->child2(), SpecOther));
|
|
setBoolean(equalNullOrUndefined(m_node->child1(), AllCellsAreFalse, EqualNullOrUndefined, ManualOperandSpeculation));
|
|
return;
|
|
}
|
|
|
|
DFG_ASSERT(m_graph, m_node, m_node->isBinaryUseKind(UntypedUse), m_node->child1().useKind(), m_node->child2().useKind());
|
|
genericJSValueCompare(
|
|
[&] (LValue left, LValue right) {
|
|
return m_out.equal(left, right);
|
|
},
|
|
operationCompareEq);
|
|
}
|
|
|
|
void compileCompareStrictEq()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
if (m_node->isBinaryUseKind(Int32Use)) {
|
|
setBoolean(
|
|
m_out.equal(lowInt32(m_node->child1()), lowInt32(m_node->child2())));
|
|
return;
|
|
}
|
|
|
|
#if USE(BIGINT32)
|
|
if (m_node->isBinaryUseKind(BigInt32Use)) {
|
|
LValue left = lowBigInt32(m_node->child1());
|
|
LValue right = lowBigInt32(m_node->child2());
|
|
|
|
// No need to unbox since the tag bits are the same on both sides
|
|
LValue result = m_out.equal(left, right);
|
|
setBoolean(result);
|
|
return;
|
|
}
|
|
|
|
if (m_node->isBinaryUseKind(AnyBigIntUse)) {
|
|
LValue left = lowJSValue(m_node->child1(), ManualOperandSpeculation);
|
|
LValue right = lowJSValue(m_node->child2(), ManualOperandSpeculation);
|
|
|
|
// Note that we cannot start with if (left == right), because we must insert the right checks (see ManualOperandSpeculation above)
|
|
// The code that we generate looks like the following pseudo-code:
|
|
/*
|
|
if (isBigInt32(left)) {
|
|
if (isBigInt32(right))
|
|
return left == right;
|
|
CHECK(isHeapBigInt(right));
|
|
return call(JSBigInt::equalsToInt32(right, unboxed(left));
|
|
}
|
|
CHECK(isHeapBigInt(left))
|
|
if (left == right)
|
|
return true;
|
|
if (isBigInt32(right))
|
|
return call(JSBigInt::equalsToInt32(left, unboxed(right));
|
|
CHECK(isHeapBigInt(right));
|
|
return call(JSBigInt::equals(left, right));
|
|
*/
|
|
LBasicBlock leftIsBigInt32 = m_out.newBlock();
|
|
LBasicBlock bothAreBigInt32 = m_out.newBlock();
|
|
LBasicBlock onlyLeftIsBigInt32 = m_out.newBlock();
|
|
LBasicBlock leftIsNotBigInt32 = m_out.newBlock();
|
|
LBasicBlock leftEqualsRight = m_out.newBlock();
|
|
LBasicBlock leftIsHeapBigInt = m_out.newBlock();
|
|
LBasicBlock rightIsBigInt32 = m_out.newBlock();
|
|
LBasicBlock rightIsNotBigInt32 = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
// Inserts a check that a value is a HeapBigInt, assuming only that we know it is not a BigInt32
|
|
auto checkIsHeapBigInt = [&](LValue lowValue, Edge highValue) {
|
|
if (m_interpreter.needsTypeCheck(highValue, SpecHeapBigInt)) {
|
|
ASSERT(mayHaveTypeCheck(highValue.useKind()));
|
|
LValue checkFailed = isNotHeapBigIntUnknownWhetherCell(lowValue, provenType(highValue) & ~SpecBigInt32);
|
|
appendOSRExit(BadType, jsValueValue(lowValue), highValue.node(), checkFailed, m_origin);
|
|
}
|
|
};
|
|
|
|
m_out.branch(isBigInt32(left, provenType(m_node->child1())), unsure(leftIsBigInt32), unsure(leftIsNotBigInt32));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(leftIsBigInt32, bothAreBigInt32);
|
|
m_out.branch(isBigInt32(right, provenType(m_node->child2())), unsure(bothAreBigInt32), unsure(onlyLeftIsBigInt32));
|
|
|
|
m_out.appendTo(bothAreBigInt32, onlyLeftIsBigInt32);
|
|
ValueFromBlock resultBothAreBigInt32 = m_out.anchor(m_out.equal(left, right));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(onlyLeftIsBigInt32, leftIsNotBigInt32);
|
|
checkIsHeapBigInt(right, m_node->child2());
|
|
LValue unboxedLeft = unboxBigInt32(left);
|
|
ValueFromBlock resultLeftIsBigInt32 = m_out.anchor(m_out.notNull(vmCall(pointerType(), operationCompareEqHeapBigIntToInt32, weakPointer(globalObject), right, unboxedLeft)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(leftIsNotBigInt32, leftEqualsRight);
|
|
checkIsHeapBigInt(left, m_node->child1());
|
|
m_out.branch(m_out.equal(left, right), unsure(leftEqualsRight), unsure(leftIsHeapBigInt));
|
|
|
|
m_out.appendTo(leftEqualsRight, leftIsHeapBigInt);
|
|
ValueFromBlock resultLeftEqualsRight = m_out.anchor(m_out.booleanTrue);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(leftIsHeapBigInt, rightIsBigInt32);
|
|
m_out.branch(isBigInt32(right, provenType(m_node->child2())), unsure(rightIsBigInt32), unsure(rightIsNotBigInt32));
|
|
|
|
m_out.appendTo(rightIsBigInt32, rightIsNotBigInt32);
|
|
LValue unboxedRight = unboxBigInt32(right);
|
|
ValueFromBlock resultRightIsBigInt32 = m_out.anchor(m_out.notNull(vmCall(pointerType(), operationCompareEqHeapBigIntToInt32, weakPointer(globalObject), left, unboxedRight)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(rightIsNotBigInt32, continuation);
|
|
checkIsHeapBigInt(right, m_node->child2());
|
|
// FIXME: [ESNext][BigInt] Create specialized version of strict equals for big ints
|
|
// https://bugs.webkit.org/show_bug.cgi?id=182895
|
|
ValueFromBlock resultBothHeapBigInt = m_out.anchor(m_out.notNull(vmCall(pointerType(), operationCompareStrictEq, weakPointer(globalObject), left, right)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setBoolean(m_out.phi(Int32, resultBothAreBigInt32, resultLeftIsBigInt32, resultLeftEqualsRight, resultRightIsBigInt32, resultBothHeapBigInt));
|
|
|
|
m_interpreter.filter(m_node->child1(), SpecBigInt);
|
|
m_interpreter.filter(m_node->child2(), SpecBigInt);
|
|
return;
|
|
}
|
|
#endif // USE(BIGINT32)
|
|
|
|
if (m_node->isBinaryUseKind(Int52RepUse)) {
|
|
Int52Kind kind;
|
|
LValue left = lowWhicheverInt52(m_node->child1(), kind);
|
|
LValue right = lowInt52(m_node->child2(), kind);
|
|
setBoolean(m_out.equal(left, right));
|
|
return;
|
|
}
|
|
|
|
if (m_node->isBinaryUseKind(DoubleRepUse)) {
|
|
setBoolean(
|
|
m_out.doubleEqual(lowDouble(m_node->child1()), lowDouble(m_node->child2())));
|
|
return;
|
|
}
|
|
|
|
if (m_node->isBinaryUseKind(StringIdentUse)) {
|
|
setBoolean(
|
|
m_out.equal(lowStringIdent(m_node->child1()), lowStringIdent(m_node->child2())));
|
|
return;
|
|
}
|
|
|
|
if (m_node->isBinaryUseKind(StringUse)) {
|
|
LValue left = lowCell(m_node->child1());
|
|
LValue right = lowCell(m_node->child2());
|
|
|
|
LBasicBlock notTriviallyEqualCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
speculateString(m_node->child1(), left);
|
|
|
|
ValueFromBlock fastResult = m_out.anchor(m_out.booleanTrue);
|
|
m_out.branch(
|
|
m_out.equal(left, right), unsure(continuation), unsure(notTriviallyEqualCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(notTriviallyEqualCase, continuation);
|
|
|
|
speculateString(m_node->child2(), right);
|
|
|
|
ValueFromBlock slowResult = m_out.anchor(stringsEqual(left, right, m_node->child1(), m_node->child2()));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setBoolean(m_out.phi(Int32, fastResult, slowResult));
|
|
return;
|
|
}
|
|
|
|
if (m_node->isBinaryUseKind(ObjectUse, UntypedUse)) {
|
|
setBoolean(
|
|
m_out.equal(
|
|
lowNonNullObject(m_node->child1()),
|
|
lowJSValue(m_node->child2())));
|
|
return;
|
|
}
|
|
|
|
if (m_node->isBinaryUseKind(UntypedUse, ObjectUse)) {
|
|
setBoolean(
|
|
m_out.equal(
|
|
lowNonNullObject(m_node->child2()),
|
|
lowJSValue(m_node->child1())));
|
|
return;
|
|
}
|
|
|
|
if (m_node->isBinaryUseKind(ObjectUse)) {
|
|
setBoolean(
|
|
m_out.equal(
|
|
lowNonNullObject(m_node->child1()),
|
|
lowNonNullObject(m_node->child2())));
|
|
return;
|
|
}
|
|
|
|
if (m_node->isBinaryUseKind(BooleanUse)) {
|
|
setBoolean(
|
|
m_out.equal(lowBoolean(m_node->child1()), lowBoolean(m_node->child2())));
|
|
return;
|
|
}
|
|
|
|
if (m_node->isBinaryUseKind(SymbolUse)) {
|
|
LValue leftSymbol = lowSymbol(m_node->child1());
|
|
LValue rightSymbol = lowSymbol(m_node->child2());
|
|
setBoolean(m_out.equal(leftSymbol, rightSymbol));
|
|
return;
|
|
}
|
|
|
|
if (m_node->isBinaryUseKind(HeapBigIntUse)) {
|
|
// FIXME: [ESNext][BigInt] Create specialized version of strict equals for big ints
|
|
// https://bugs.webkit.org/show_bug.cgi?id=182895
|
|
LValue left = lowHeapBigInt(m_node->child1());
|
|
LValue right = lowHeapBigInt(m_node->child2());
|
|
|
|
LBasicBlock notTriviallyEqualCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
ValueFromBlock fastResult = m_out.anchor(m_out.booleanTrue);
|
|
m_out.branch(m_out.equal(left, right), rarely(continuation), usually(notTriviallyEqualCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(notTriviallyEqualCase, continuation);
|
|
|
|
ValueFromBlock slowResult = m_out.anchor(m_out.notNull(vmCall(
|
|
pointerType(), operationCompareStrictEq, weakPointer(globalObject), left, right)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setBoolean(m_out.phi(Int32, fastResult, slowResult));
|
|
return;
|
|
}
|
|
|
|
if (m_node->isBinaryUseKind(SymbolUse, UntypedUse)
|
|
|| m_node->isBinaryUseKind(UntypedUse, SymbolUse)) {
|
|
Edge symbolEdge = m_node->child1();
|
|
Edge untypedEdge = m_node->child2();
|
|
if (symbolEdge.useKind() != SymbolUse)
|
|
std::swap(symbolEdge, untypedEdge);
|
|
|
|
LValue leftSymbol = lowSymbol(symbolEdge);
|
|
LValue untypedValue = lowJSValue(untypedEdge);
|
|
|
|
setBoolean(m_out.equal(leftSymbol, untypedValue));
|
|
return;
|
|
}
|
|
|
|
if (m_node->isBinaryUseKind(MiscUse, UntypedUse)
|
|
|| m_node->isBinaryUseKind(UntypedUse, MiscUse)) {
|
|
speculate(m_node->child1());
|
|
speculate(m_node->child2());
|
|
LValue left = lowJSValue(m_node->child1(), ManualOperandSpeculation);
|
|
LValue right = lowJSValue(m_node->child2(), ManualOperandSpeculation);
|
|
setBoolean(m_out.equal(left, right));
|
|
return;
|
|
}
|
|
|
|
if (m_node->isBinaryUseKind(StringIdentUse, NotStringVarUse)
|
|
|| m_node->isBinaryUseKind(NotStringVarUse, StringIdentUse)) {
|
|
Edge leftEdge = m_node->childFor(StringIdentUse);
|
|
Edge rightEdge = m_node->childFor(NotStringVarUse);
|
|
|
|
LValue left = lowStringIdent(leftEdge);
|
|
LValue rightValue = lowJSValue(rightEdge, ManualOperandSpeculation);
|
|
|
|
LBasicBlock isCellCase = m_out.newBlock();
|
|
LBasicBlock isStringCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
ValueFromBlock notCellResult = m_out.anchor(m_out.booleanFalse);
|
|
m_out.branch(
|
|
isCell(rightValue, provenType(rightEdge)),
|
|
unsure(isCellCase), unsure(continuation));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isCellCase, isStringCase);
|
|
ValueFromBlock notStringResult = m_out.anchor(m_out.booleanFalse);
|
|
m_out.branch(
|
|
isString(rightValue, provenType(rightEdge)),
|
|
unsure(isStringCase), unsure(continuation));
|
|
|
|
m_out.appendTo(isStringCase, continuation);
|
|
LValue right = m_out.loadPtr(rightValue, m_heaps.JSString_value);
|
|
speculateStringIdent(rightEdge, rightValue, right);
|
|
ValueFromBlock isStringResult = m_out.anchor(m_out.equal(left, right));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setBoolean(m_out.phi(Int32, notCellResult, notStringResult, isStringResult));
|
|
return;
|
|
}
|
|
|
|
if (m_node->isBinaryUseKind(StringUse, UntypedUse)) {
|
|
compileStringToUntypedStrictEquality(m_node->child1(), m_node->child2());
|
|
return;
|
|
}
|
|
if (m_node->isBinaryUseKind(UntypedUse, StringUse)) {
|
|
compileStringToUntypedStrictEquality(m_node->child2(), m_node->child1());
|
|
return;
|
|
}
|
|
|
|
// FIXME: we can do something much smarter here, see the DFGSpeculativeJIT approach in e.g. SpeculativeJIT::nonSpeculativePeepholeStrictEq
|
|
DFG_ASSERT(m_graph, m_node, m_node->isBinaryUseKind(UntypedUse), m_node->child1().useKind(), m_node->child2().useKind());
|
|
genericJSValueCompare(
|
|
[&] (LValue left, LValue right) {
|
|
return m_out.equal(left, right);
|
|
},
|
|
operationCompareStrictEq);
|
|
}
|
|
|
|
void compileStringToUntypedStrictEquality(Edge stringEdge, Edge untypedEdge)
|
|
{
|
|
ASSERT(stringEdge.useKind() == StringUse);
|
|
ASSERT(untypedEdge.useKind() == UntypedUse);
|
|
|
|
LValue leftString = lowCell(stringEdge);
|
|
LValue rightValue = lowJSValue(untypedEdge);
|
|
SpeculatedType rightValueType = provenType(untypedEdge);
|
|
|
|
// Verify left is string.
|
|
speculateString(stringEdge, leftString);
|
|
|
|
LBasicBlock testUntypedEdgeIsCell = m_out.newBlock();
|
|
LBasicBlock testUntypedEdgeIsString = m_out.newBlock();
|
|
LBasicBlock testStringEquality = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
// Given left is string. If the value are strictly equal, rightValue has to be the same string.
|
|
ValueFromBlock fastTrue = m_out.anchor(m_out.booleanTrue);
|
|
m_out.branch(m_out.equal(leftString, rightValue), unsure(continuation), unsure(testUntypedEdgeIsCell));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(testUntypedEdgeIsCell, testUntypedEdgeIsString);
|
|
ValueFromBlock fastFalse = m_out.anchor(m_out.booleanFalse);
|
|
m_out.branch(isNotCell(rightValue, rightValueType), unsure(continuation), unsure(testUntypedEdgeIsString));
|
|
|
|
// Check if the untyped edge is a string.
|
|
m_out.appendTo(testUntypedEdgeIsString, testStringEquality);
|
|
m_out.branch(isNotString(rightValue, rightValueType), unsure(continuation), unsure(testStringEquality));
|
|
|
|
// Full String compare.
|
|
m_out.appendTo(testStringEquality, continuation);
|
|
ValueFromBlock slowResult = m_out.anchor(stringsEqual(leftString, rightValue, stringEdge, untypedEdge));
|
|
m_out.jump(continuation);
|
|
|
|
// Continuation.
|
|
m_out.appendTo(continuation, lastNext);
|
|
setBoolean(m_out.phi(Int32, fastTrue, fastFalse, slowResult));
|
|
}
|
|
|
|
void compileCompareEqPtr()
|
|
{
|
|
setBoolean(
|
|
m_out.equal(
|
|
lowJSValue(m_node->child1()),
|
|
weakPointer(m_node->cellOperand()->cell())));
|
|
}
|
|
|
|
void compileCompareLess()
|
|
{
|
|
compare(
|
|
[&] (LValue left, LValue right) {
|
|
return m_out.lessThan(left, right);
|
|
},
|
|
[&] (LValue left, LValue right) {
|
|
return m_out.doubleLessThan(left, right);
|
|
},
|
|
operationCompareStringImplLess,
|
|
operationCompareStringLess,
|
|
operationCompareLess);
|
|
}
|
|
|
|
void compileCompareLessEq()
|
|
{
|
|
compare(
|
|
[&] (LValue left, LValue right) {
|
|
return m_out.lessThanOrEqual(left, right);
|
|
},
|
|
[&] (LValue left, LValue right) {
|
|
return m_out.doubleLessThanOrEqual(left, right);
|
|
},
|
|
operationCompareStringImplLessEq,
|
|
operationCompareStringLessEq,
|
|
operationCompareLessEq);
|
|
}
|
|
|
|
void compileCompareGreater()
|
|
{
|
|
compare(
|
|
[&] (LValue left, LValue right) {
|
|
return m_out.greaterThan(left, right);
|
|
},
|
|
[&] (LValue left, LValue right) {
|
|
return m_out.doubleGreaterThan(left, right);
|
|
},
|
|
operationCompareStringImplGreater,
|
|
operationCompareStringGreater,
|
|
operationCompareGreater);
|
|
}
|
|
|
|
void compileCompareGreaterEq()
|
|
{
|
|
compare(
|
|
[&] (LValue left, LValue right) {
|
|
return m_out.greaterThanOrEqual(left, right);
|
|
},
|
|
[&] (LValue left, LValue right) {
|
|
return m_out.doubleGreaterThanOrEqual(left, right);
|
|
},
|
|
operationCompareStringImplGreaterEq,
|
|
operationCompareStringGreaterEq,
|
|
operationCompareGreaterEq);
|
|
}
|
|
|
|
void compileCompareBelow()
|
|
{
|
|
setBoolean(m_out.below(lowInt32(m_node->child1()), lowInt32(m_node->child2())));
|
|
}
|
|
|
|
void compileCompareBelowEq()
|
|
{
|
|
setBoolean(m_out.belowOrEqual(lowInt32(m_node->child1()), lowInt32(m_node->child2())));
|
|
}
|
|
|
|
void compileSameValue()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
if (m_node->isBinaryUseKind(DoubleRepUse)) {
|
|
LValue arg1 = lowDouble(m_node->child1());
|
|
LValue arg2 = lowDouble(m_node->child2());
|
|
|
|
LBasicBlock numberCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int32);
|
|
patchpoint->append(arg1, ValueRep::SomeRegister);
|
|
patchpoint->append(arg2, ValueRep::SomeRegister);
|
|
patchpoint->numGPScratchRegisters = 1;
|
|
patchpoint->setGenerator(
|
|
[] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
GPRReg scratchGPR = params.gpScratch(0);
|
|
jit.moveDoubleTo64(params[1].fpr(), scratchGPR);
|
|
jit.moveDoubleTo64(params[2].fpr(), params[0].gpr());
|
|
jit.compare64(CCallHelpers::Equal, scratchGPR, params[0].gpr(), params[0].gpr());
|
|
});
|
|
patchpoint->effects = Effects::none();
|
|
ValueFromBlock compareResult = m_out.anchor(patchpoint);
|
|
m_out.branch(patchpoint, unsure(continuation), unsure(numberCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(numberCase, continuation);
|
|
LValue isArg1NaN = m_out.doubleNotEqualOrUnordered(arg1, arg1);
|
|
LValue isArg2NaN = m_out.doubleNotEqualOrUnordered(arg2, arg2);
|
|
ValueFromBlock nanResult = m_out.anchor(m_out.bitAnd(isArg1NaN, isArg2NaN));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setBoolean(m_out.phi(Int32, compareResult, nanResult));
|
|
return;
|
|
}
|
|
|
|
ASSERT(m_node->isBinaryUseKind(UntypedUse));
|
|
setBoolean(vmCall(Int32, operationSameValue, weakPointer(globalObject), lowJSValue(m_node->child1()), lowJSValue(m_node->child2())));
|
|
}
|
|
|
|
void compileLogicalNot()
|
|
{
|
|
setBoolean(m_out.logicalNot(boolify(m_node->child1())));
|
|
}
|
|
|
|
void compileCallOrConstruct()
|
|
{
|
|
Node* node = m_node;
|
|
unsigned numArgs = node->numChildren() - 1;
|
|
|
|
LValue jsCallee = lowJSValue(m_graph.varArgChild(node, 0));
|
|
|
|
unsigned frameSize = (CallFrame::headerSizeInRegisters + numArgs) * sizeof(EncodedJSValue);
|
|
unsigned alignedFrameSize = WTF::roundUpToMultipleOf(stackAlignmentBytes(), frameSize);
|
|
|
|
// JS->JS calling convention requires that the caller allows this much space on top of stack to
|
|
// get trashed by the callee, even if not all of that space is used to pass arguments. We tell
|
|
// B3 this explicitly for two reasons:
|
|
//
|
|
// - We will only pass frameSize worth of stuff.
|
|
// - The trashed stack guarantee is logically separate from the act of passing arguments, so we
|
|
// shouldn't rely on Air to infer the trashed stack property based on the arguments it ends
|
|
// up seeing.
|
|
m_proc.requestCallArgAreaSizeInBytes(alignedFrameSize);
|
|
|
|
// Collect the arguments, since this can generate code and we want to generate it before we emit
|
|
// the call.
|
|
Vector<ConstrainedValue> arguments;
|
|
|
|
// Make sure that the callee goes into GPR0 because that's where the slow path thunks expect the
|
|
// callee to be.
|
|
arguments.append(ConstrainedValue(jsCallee, ValueRep::reg(GPRInfo::regT0)));
|
|
|
|
auto addArgument = [&] (LValue value, VirtualRegister reg, int offset) {
|
|
intptr_t offsetFromSP =
|
|
(reg.offset() - CallerFrameAndPC::sizeInRegisters) * sizeof(EncodedJSValue) + offset;
|
|
arguments.append(ConstrainedValue(value, ValueRep::stackArgument(offsetFromSP)));
|
|
};
|
|
|
|
addArgument(jsCallee, VirtualRegister(CallFrameSlot::callee), 0);
|
|
addArgument(m_out.constInt32(numArgs), VirtualRegister(CallFrameSlot::argumentCountIncludingThis), PayloadOffset);
|
|
for (unsigned i = 0; i < numArgs; ++i)
|
|
addArgument(lowJSValue(m_graph.varArgChild(node, 1 + i)), virtualRegisterForArgumentIncludingThis(i), 0);
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int64);
|
|
patchpoint->appendVector(arguments);
|
|
|
|
RefPtr<PatchpointExceptionHandle> exceptionHandle =
|
|
preparePatchpointForExceptions(patchpoint);
|
|
|
|
patchpoint->append(m_notCellMask, ValueRep::reg(GPRInfo::notCellMaskRegister));
|
|
patchpoint->append(m_numberTag, ValueRep::reg(GPRInfo::numberTagRegister));
|
|
patchpoint->clobber(RegisterSet::macroScratchRegisters());
|
|
patchpoint->clobberLate(RegisterSet::volatileRegistersForJSCall());
|
|
patchpoint->resultConstraints = { ValueRep::reg(GPRInfo::returnValueGPR) };
|
|
|
|
CodeOrigin codeOrigin = codeOriginDescriptionOfCallSite();
|
|
State* state = &m_ftlState;
|
|
VM* vm = &this->vm();
|
|
patchpoint->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
CallSiteIndex callSiteIndex = state->jitCode->common.codeOrigins->addUniqueCallSiteIndex(codeOrigin);
|
|
|
|
exceptionHandle->scheduleExitCreationForUnwind(params, callSiteIndex);
|
|
|
|
jit.store32(
|
|
CCallHelpers::TrustedImm32(callSiteIndex.bits()),
|
|
CCallHelpers::tagFor(VirtualRegister(CallFrameSlot::argumentCountIncludingThis)));
|
|
|
|
CallLinkInfo* callLinkInfo = jit.codeBlock()->addCallLinkInfo(node->origin.semantic);
|
|
|
|
CCallHelpers::DataLabelPtr targetToCheck;
|
|
CCallHelpers::Jump slowPath = jit.branchPtrWithPatch(
|
|
CCallHelpers::NotEqual, GPRInfo::regT0, targetToCheck,
|
|
CCallHelpers::TrustedImmPtr(nullptr));
|
|
|
|
CCallHelpers::Call fastCall = jit.nearCall();
|
|
CCallHelpers::Jump done = jit.jump();
|
|
|
|
slowPath.link(&jit);
|
|
|
|
jit.move(CCallHelpers::TrustedImmPtr(callLinkInfo), GPRInfo::regT2);
|
|
jit.move(CCallHelpers::TrustedImmPtr(jit.codeBlock()->globalObjectFor(node->origin.semantic)), GPRInfo::regT3);
|
|
CCallHelpers::Call slowCall = jit.nearCall();
|
|
done.link(&jit);
|
|
|
|
callLinkInfo->setUpCall(
|
|
node->op() == Construct ? CallLinkInfo::Construct : CallLinkInfo::Call, GPRInfo::regT0);
|
|
|
|
jit.addPtr(
|
|
CCallHelpers::TrustedImm32(-params.proc().frameSize()),
|
|
GPRInfo::callFrameRegister, CCallHelpers::stackPointerRegister);
|
|
|
|
jit.addLinkTask(
|
|
[=] (LinkBuffer& linkBuffer) {
|
|
MacroAssemblerCodePtr<JITThunkPtrTag> linkCall = vm->getCTIStub(linkCallThunkGenerator).code();
|
|
linkBuffer.link(slowCall, FunctionPtr<JITThunkPtrTag>(linkCall));
|
|
|
|
callLinkInfo->setCallLocations(
|
|
CodeLocationLabel<JSInternalPtrTag>(linkBuffer.locationOfNearCall<JSInternalPtrTag>(slowCall)),
|
|
CodeLocationLabel<JSInternalPtrTag>(linkBuffer.locationOf<JSInternalPtrTag>(targetToCheck)),
|
|
linkBuffer.locationOfNearCall<JSInternalPtrTag>(fastCall));
|
|
});
|
|
});
|
|
|
|
setJSValue(patchpoint);
|
|
}
|
|
|
|
void compileDirectCallOrConstruct()
|
|
{
|
|
Node* node = m_node;
|
|
bool isTail = node->op() == DirectTailCall;
|
|
bool isConstruct = node->op() == DirectConstruct;
|
|
|
|
ExecutableBase* executable = node->castOperand<ExecutableBase*>();
|
|
FunctionExecutable* functionExecutable = jsDynamicCast<FunctionExecutable*>(vm(), executable);
|
|
|
|
unsigned numPassedArgs = node->numChildren() - 1;
|
|
unsigned numAllocatedArgs = numPassedArgs;
|
|
|
|
if (functionExecutable) {
|
|
numAllocatedArgs = std::max(
|
|
numAllocatedArgs,
|
|
std::min(
|
|
static_cast<unsigned>(functionExecutable->parameterCount()) + 1,
|
|
Options::maximumDirectCallStackSize()));
|
|
}
|
|
|
|
LValue jsCallee = lowJSValue(m_graph.varArgChild(node, 0));
|
|
|
|
if (!isTail) {
|
|
unsigned frameSize = (CallFrame::headerSizeInRegisters + numAllocatedArgs) * sizeof(EncodedJSValue);
|
|
unsigned alignedFrameSize = WTF::roundUpToMultipleOf(stackAlignmentBytes(), frameSize);
|
|
|
|
m_proc.requestCallArgAreaSizeInBytes(alignedFrameSize);
|
|
}
|
|
|
|
Vector<ConstrainedValue> arguments;
|
|
|
|
arguments.append(ConstrainedValue(jsCallee, ValueRep::SomeRegister));
|
|
if (!isTail) {
|
|
auto addArgument = [&] (LValue value, VirtualRegister reg, int offset) {
|
|
intptr_t offsetFromSP =
|
|
(reg.offset() - CallerFrameAndPC::sizeInRegisters) * sizeof(EncodedJSValue) + offset;
|
|
arguments.append(ConstrainedValue(value, ValueRep::stackArgument(offsetFromSP)));
|
|
};
|
|
|
|
addArgument(jsCallee, VirtualRegister(CallFrameSlot::callee), 0);
|
|
addArgument(m_out.constInt32(numPassedArgs), VirtualRegister(CallFrameSlot::argumentCountIncludingThis), PayloadOffset);
|
|
for (unsigned i = 0; i < numPassedArgs; ++i)
|
|
addArgument(lowJSValue(m_graph.varArgChild(node, 1 + i)), virtualRegisterForArgumentIncludingThis(i), 0);
|
|
for (unsigned i = numPassedArgs; i < numAllocatedArgs; ++i)
|
|
addArgument(m_out.constInt64(JSValue::encode(jsUndefined())), virtualRegisterForArgumentIncludingThis(i), 0);
|
|
} else {
|
|
for (unsigned i = 0; i < numPassedArgs; ++i)
|
|
arguments.append(ConstrainedValue(lowJSValue(m_graph.varArgChild(node, 1 + i)), ValueRep::WarmAny));
|
|
}
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(isTail ? Void : Int64);
|
|
patchpoint->appendVector(arguments);
|
|
|
|
RefPtr<PatchpointExceptionHandle> exceptionHandle = preparePatchpointForExceptions(patchpoint);
|
|
|
|
if (isTail) {
|
|
// The shuffler needs tags.
|
|
patchpoint->append(m_notCellMask, ValueRep::reg(GPRInfo::notCellMaskRegister));
|
|
patchpoint->append(m_numberTag, ValueRep::reg(GPRInfo::numberTagRegister));
|
|
}
|
|
|
|
patchpoint->clobber(RegisterSet::macroScratchRegisters());
|
|
if (!isTail) {
|
|
patchpoint->clobberLate(RegisterSet::volatileRegistersForJSCall());
|
|
patchpoint->resultConstraints = { ValueRep::reg(GPRInfo::returnValueGPR) };
|
|
}
|
|
|
|
CodeOrigin codeOrigin = codeOriginDescriptionOfCallSite();
|
|
State* state = &m_ftlState;
|
|
patchpoint->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
CallSiteIndex callSiteIndex = state->jitCode->common.codeOrigins->addUniqueCallSiteIndex(codeOrigin);
|
|
|
|
GPRReg calleeGPR = params[!isTail].gpr();
|
|
|
|
exceptionHandle->scheduleExitCreationForUnwind(params, callSiteIndex);
|
|
|
|
Box<CCallHelpers::JumpList> exceptions =
|
|
exceptionHandle->scheduleExitCreation(params)->jumps(jit);
|
|
|
|
if (isTail) {
|
|
CallFrameShuffleData shuffleData;
|
|
shuffleData.numLocals = state->jitCode->common.frameRegisterCount;
|
|
|
|
RegisterSet toSave = params.unavailableRegisters();
|
|
shuffleData.callee = ValueRecovery::inGPR(calleeGPR, DataFormatCell);
|
|
toSave.set(calleeGPR);
|
|
for (unsigned i = 0; i < numPassedArgs; ++i) {
|
|
ValueRecovery recovery = params[1 + i].recoveryForJSValue();
|
|
shuffleData.args.append(recovery);
|
|
recovery.forEachReg(
|
|
[&] (Reg reg) {
|
|
toSave.set(reg);
|
|
});
|
|
}
|
|
for (unsigned i = numPassedArgs; i < numAllocatedArgs; ++i)
|
|
shuffleData.args.append(ValueRecovery::constant(jsUndefined()));
|
|
shuffleData.numPassedArgs = numPassedArgs;
|
|
shuffleData.setupCalleeSaveRegisters(jit.codeBlock());
|
|
|
|
CallLinkInfo* callLinkInfo = jit.codeBlock()->addCallLinkInfo(node->origin.semantic);
|
|
|
|
CCallHelpers::PatchableJump patchableJump = jit.patchableJump();
|
|
CCallHelpers::Label mainPath = jit.label();
|
|
|
|
jit.store32(
|
|
CCallHelpers::TrustedImm32(callSiteIndex.bits()),
|
|
CCallHelpers::tagFor(VirtualRegister(CallFrameSlot::argumentCountIncludingThis)));
|
|
|
|
callLinkInfo->setFrameShuffleData(shuffleData);
|
|
CallFrameShuffler(jit, shuffleData).prepareForTailCall();
|
|
|
|
CCallHelpers::Call call = jit.nearTailCall();
|
|
|
|
jit.abortWithReason(JITDidReturnFromTailCall);
|
|
|
|
CCallHelpers::Label slowPath = jit.label();
|
|
patchableJump.m_jump.linkTo(slowPath, &jit);
|
|
callOperation(
|
|
*state, toSave, jit,
|
|
node->origin.semantic, exceptions.get(), operationLinkDirectCall,
|
|
InvalidGPRReg, CCallHelpers::TrustedImmPtr(callLinkInfo), calleeGPR).call();
|
|
jit.jump().linkTo(mainPath, &jit);
|
|
|
|
callLinkInfo->setUpCall(CallLinkInfo::DirectTailCall, InvalidGPRReg);
|
|
callLinkInfo->setExecutableDuringCompilation(executable);
|
|
if (numAllocatedArgs > numPassedArgs)
|
|
callLinkInfo->setMaxArgumentCountIncludingThis(numAllocatedArgs);
|
|
|
|
jit.addLinkTask(
|
|
[=] (LinkBuffer& linkBuffer) {
|
|
CodeLocationLabel<JSInternalPtrTag> patchableJumpLocation = linkBuffer.locationOf<JSInternalPtrTag>(patchableJump);
|
|
CodeLocationNearCall<JSInternalPtrTag> callLocation = linkBuffer.locationOfNearCall<JSInternalPtrTag>(call);
|
|
CodeLocationLabel<JSInternalPtrTag> slowPathLocation = linkBuffer.locationOf<JSInternalPtrTag>(slowPath);
|
|
|
|
callLinkInfo->setCallLocations(
|
|
patchableJumpLocation,
|
|
slowPathLocation,
|
|
callLocation);
|
|
});
|
|
return;
|
|
}
|
|
|
|
CallLinkInfo* callLinkInfo = jit.codeBlock()->addCallLinkInfo(node->origin.semantic);
|
|
|
|
CCallHelpers::Label mainPath = jit.label();
|
|
|
|
jit.store32(
|
|
CCallHelpers::TrustedImm32(callSiteIndex.bits()),
|
|
CCallHelpers::tagFor(VirtualRegister(CallFrameSlot::argumentCountIncludingThis)));
|
|
|
|
CCallHelpers::Call call = jit.nearCall();
|
|
jit.addPtr(
|
|
CCallHelpers::TrustedImm32(-params.proc().frameSize()),
|
|
GPRInfo::callFrameRegister, CCallHelpers::stackPointerRegister);
|
|
|
|
callLinkInfo->setUpCall(
|
|
isConstruct ? CallLinkInfo::DirectConstruct : CallLinkInfo::DirectCall, InvalidGPRReg);
|
|
callLinkInfo->setExecutableDuringCompilation(executable);
|
|
if (numAllocatedArgs > numPassedArgs)
|
|
callLinkInfo->setMaxArgumentCountIncludingThis(numAllocatedArgs);
|
|
|
|
params.addLatePath(
|
|
[=] (CCallHelpers& jit) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
CCallHelpers::Label slowPath = jit.label();
|
|
if (isX86())
|
|
jit.pop(CCallHelpers::selectScratchGPR(calleeGPR));
|
|
|
|
callOperation(
|
|
*state, params.unavailableRegisters(), jit,
|
|
node->origin.semantic, exceptions.get(), operationLinkDirectCall,
|
|
InvalidGPRReg, CCallHelpers::TrustedImmPtr(callLinkInfo),
|
|
calleeGPR).call();
|
|
jit.jump().linkTo(mainPath, &jit);
|
|
|
|
jit.addLinkTask(
|
|
[=] (LinkBuffer& linkBuffer) {
|
|
CodeLocationNearCall<JSInternalPtrTag> callLocation = linkBuffer.locationOfNearCall<JSInternalPtrTag>(call);
|
|
CodeLocationLabel<JSInternalPtrTag> slowPathLocation = linkBuffer.locationOf<JSInternalPtrTag>(slowPath);
|
|
|
|
linkBuffer.link(call, slowPathLocation);
|
|
|
|
callLinkInfo->setCallLocations(
|
|
CodeLocationLabel<JSInternalPtrTag>(),
|
|
slowPathLocation,
|
|
callLocation);
|
|
});
|
|
});
|
|
});
|
|
|
|
if (isTail)
|
|
patchpoint->effects.terminal = true;
|
|
else
|
|
setJSValue(patchpoint);
|
|
}
|
|
|
|
void compileTailCall()
|
|
{
|
|
Node* node = m_node;
|
|
unsigned numArgs = node->numChildren() - 1;
|
|
|
|
// It seems counterintuitive that this is needed given that tail calls don't create a new frame
|
|
// on the stack. However, the tail call slow path builds the frame at SP instead of FP before
|
|
// calling into the slow path C code. This slow path may decide to throw an exception because
|
|
// the callee we're trying to call is not callable. Throwing an exception will cause us to walk
|
|
// the stack, which may read, for the sake of the correctness of this code, arbitrary slots on the
|
|
// stack to recover state. This call arg area ensures the call frame shuffler does not overwrite
|
|
// any of the slots the stack walking code requires when on the slow path.
|
|
m_proc.requestCallArgAreaSizeInBytes(
|
|
WTF::roundUpToMultipleOf(stackAlignmentBytes(), (CallFrame::headerSizeInRegisters + numArgs) * sizeof(EncodedJSValue)));
|
|
|
|
LValue jsCallee = lowJSValue(m_graph.varArgChild(node, 0));
|
|
|
|
// We want B3 to give us all of the arguments using whatever mechanism it thinks is
|
|
// convenient. The generator then shuffles those arguments into our own call frame,
|
|
// destroying our frame in the process.
|
|
|
|
// Note that we don't have to do anything special for exceptions. A tail call is only a
|
|
// tail call if it is not inside a try block.
|
|
|
|
Vector<ConstrainedValue> arguments;
|
|
|
|
arguments.append(ConstrainedValue(jsCallee, ValueRep::reg(GPRInfo::regT0)));
|
|
|
|
for (unsigned i = 0; i < numArgs; ++i) {
|
|
// Note: we could let the shuffler do boxing for us, but it's not super clear that this
|
|
// would be better. Also, if we wanted to do that, then we'd have to teach the shuffler
|
|
// that 32-bit values could land at 4-byte alignment but not 8-byte alignment.
|
|
|
|
ConstrainedValue constrainedValue(
|
|
lowJSValue(m_graph.varArgChild(node, 1 + i)),
|
|
ValueRep::WarmAny);
|
|
arguments.append(constrainedValue);
|
|
}
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Void);
|
|
patchpoint->appendVector(arguments);
|
|
|
|
patchpoint->append(m_notCellMask, ValueRep::reg(GPRInfo::notCellMaskRegister));
|
|
patchpoint->append(m_numberTag, ValueRep::reg(GPRInfo::numberTagRegister));
|
|
|
|
// Prevent any of the arguments from using the scratch register.
|
|
patchpoint->clobberEarly(RegisterSet::macroScratchRegisters());
|
|
|
|
patchpoint->effects.terminal = true;
|
|
|
|
// We don't have to tell the patchpoint that we will clobber registers, since we won't return
|
|
// anyway.
|
|
|
|
CodeOrigin codeOrigin = codeOriginDescriptionOfCallSite();
|
|
State* state = &m_ftlState;
|
|
VM* vm = &this->vm();
|
|
patchpoint->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
CallSiteIndex callSiteIndex = state->jitCode->common.codeOrigins->addUniqueCallSiteIndex(codeOrigin);
|
|
|
|
// Yes, this is really necessary. You could throw an exception in a host call on the
|
|
// slow path. That'll route us to operationLookupExceptionHandler(), which unwinds starting
|
|
// with the call site index of our frame. Bad things happen if it's not set.
|
|
jit.store32(
|
|
CCallHelpers::TrustedImm32(callSiteIndex.bits()),
|
|
CCallHelpers::tagFor(VirtualRegister(CallFrameSlot::argumentCountIncludingThis)));
|
|
|
|
CallFrameShuffleData shuffleData;
|
|
shuffleData.numLocals = state->jitCode->common.frameRegisterCount;
|
|
shuffleData.callee = ValueRecovery::inGPR(GPRInfo::regT0, DataFormatJS);
|
|
|
|
for (unsigned i = 0; i < numArgs; ++i)
|
|
shuffleData.args.append(params[1 + i].recoveryForJSValue());
|
|
|
|
shuffleData.numPassedArgs = numArgs;
|
|
|
|
shuffleData.setupCalleeSaveRegisters(jit.codeBlock());
|
|
|
|
CallLinkInfo* callLinkInfo = jit.codeBlock()->addCallLinkInfo(codeOrigin);
|
|
|
|
CCallHelpers::DataLabelPtr targetToCheck;
|
|
CCallHelpers::Jump slowPath = jit.branchPtrWithPatch(
|
|
CCallHelpers::NotEqual, GPRInfo::regT0, targetToCheck,
|
|
CCallHelpers::TrustedImmPtr(nullptr));
|
|
|
|
callLinkInfo->setFrameShuffleData(shuffleData);
|
|
CallFrameShuffler(jit, shuffleData).prepareForTailCall();
|
|
|
|
CCallHelpers::Call fastCall = jit.nearTailCall();
|
|
|
|
slowPath.link(&jit);
|
|
|
|
CallFrameShuffler slowPathShuffler(jit, shuffleData);
|
|
slowPathShuffler.setCalleeJSValueRegs(JSValueRegs(GPRInfo::regT0));
|
|
slowPathShuffler.prepareForSlowPath();
|
|
|
|
jit.move(CCallHelpers::TrustedImmPtr(callLinkInfo), GPRInfo::regT2);
|
|
jit.move(CCallHelpers::TrustedImmPtr(jit.codeBlock()->globalObjectFor(node->origin.semantic)), GPRInfo::regT3);
|
|
CCallHelpers::Call slowCall = jit.nearCall();
|
|
|
|
jit.abortWithReason(JITDidReturnFromTailCall);
|
|
|
|
callLinkInfo->setUpCall(CallLinkInfo::TailCall, GPRInfo::regT0);
|
|
|
|
jit.addLinkTask(
|
|
[=] (LinkBuffer& linkBuffer) {
|
|
MacroAssemblerCodePtr<JITThunkPtrTag> linkCall = vm->getCTIStub(linkCallThunkGenerator).code();
|
|
linkBuffer.link(slowCall, FunctionPtr<JITThunkPtrTag>(linkCall));
|
|
|
|
callLinkInfo->setCallLocations(
|
|
CodeLocationLabel<JSInternalPtrTag>(linkBuffer.locationOfNearCall<JSInternalPtrTag>(slowCall)),
|
|
CodeLocationLabel<JSInternalPtrTag>(linkBuffer.locationOf<JSInternalPtrTag>(targetToCheck)),
|
|
linkBuffer.locationOfNearCall<JSInternalPtrTag>(fastCall));
|
|
});
|
|
});
|
|
}
|
|
|
|
void compileCallOrConstructVarargsSpread()
|
|
{
|
|
Node* node = m_node;
|
|
Node* arguments = node->child3().node();
|
|
|
|
LValue jsCallee = lowJSValue(m_node->child1());
|
|
LValue thisArg = lowJSValue(m_node->child2());
|
|
|
|
RELEASE_ASSERT(arguments->op() == PhantomNewArrayWithSpread || arguments->op() == PhantomSpread || arguments->op() == PhantomNewArrayBuffer);
|
|
|
|
unsigned staticArgumentCount = 0;
|
|
Vector<LValue, 2> spreadLengths;
|
|
Vector<LValue, 8> patchpointArguments;
|
|
HashMap<InlineCallFrame*, LValue, WTF::DefaultHash<InlineCallFrame*>, WTF::NullableHashTraits<InlineCallFrame*>> cachedSpreadLengths;
|
|
auto pushAndCountArgumentsFromRightToLeft = recursableLambda([&](auto self, Node* target) -> void {
|
|
if (target->op() == PhantomSpread) {
|
|
self(target->child1().node());
|
|
return;
|
|
}
|
|
|
|
if (target->op() == PhantomNewArrayWithSpread) {
|
|
BitVector* bitVector = target->bitVector();
|
|
for (unsigned i = target->numChildren(); i--; ) {
|
|
if (bitVector->get(i))
|
|
self(m_graph.varArgChild(target, i).node());
|
|
else {
|
|
++staticArgumentCount;
|
|
LValue argument = this->lowJSValue(m_graph.varArgChild(target, i));
|
|
patchpointArguments.append(argument);
|
|
}
|
|
}
|
|
return;
|
|
}
|
|
|
|
if (target->op() == PhantomNewArrayBuffer) {
|
|
staticArgumentCount += target->castOperand<JSImmutableButterfly*>()->length();
|
|
return;
|
|
}
|
|
|
|
RELEASE_ASSERT(target->op() == PhantomCreateRest);
|
|
InlineCallFrame* inlineCallFrame = target->origin.semantic.inlineCallFrame();
|
|
unsigned numberOfArgumentsToSkip = target->numberOfArgumentsToSkip();
|
|
LValue length = cachedSpreadLengths.ensure(inlineCallFrame, [&] () {
|
|
return m_out.zeroExtPtr(this->getSpreadLengthFromInlineCallFrame(inlineCallFrame, numberOfArgumentsToSkip));
|
|
}).iterator->value;
|
|
patchpointArguments.append(length);
|
|
spreadLengths.append(length);
|
|
});
|
|
|
|
pushAndCountArgumentsFromRightToLeft(arguments);
|
|
LValue argumentCountIncludingThis = m_out.constIntPtr(staticArgumentCount + 1);
|
|
for (LValue length : spreadLengths)
|
|
argumentCountIncludingThis = m_out.add(length, argumentCountIncludingThis);
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int64);
|
|
|
|
patchpoint->append(jsCallee, ValueRep::reg(GPRInfo::regT0));
|
|
patchpoint->append(thisArg, ValueRep::WarmAny);
|
|
patchpoint->append(argumentCountIncludingThis, ValueRep::WarmAny);
|
|
patchpoint->appendVectorWithRep(patchpointArguments, ValueRep::WarmAny);
|
|
patchpoint->append(m_notCellMask, ValueRep::reg(GPRInfo::notCellMaskRegister));
|
|
patchpoint->append(m_numberTag, ValueRep::reg(GPRInfo::numberTagRegister));
|
|
|
|
RefPtr<PatchpointExceptionHandle> exceptionHandle = preparePatchpointForExceptions(patchpoint);
|
|
|
|
patchpoint->clobber(RegisterSet::macroScratchRegisters());
|
|
patchpoint->clobber(RegisterSet::volatileRegistersForJSCall()); // No inputs will be in a volatile register.
|
|
patchpoint->resultConstraints = { ValueRep::reg(GPRInfo::returnValueGPR) };
|
|
|
|
patchpoint->numGPScratchRegisters = 0;
|
|
|
|
// This is the minimum amount of call arg area stack space that all JS->JS calls always have.
|
|
unsigned minimumJSCallAreaSize =
|
|
sizeof(CallerFrameAndPC) +
|
|
WTF::roundUpToMultipleOf(stackAlignmentBytes(), 5 * sizeof(EncodedJSValue));
|
|
|
|
m_proc.requestCallArgAreaSizeInBytes(minimumJSCallAreaSize);
|
|
|
|
CodeOrigin codeOrigin = codeOriginDescriptionOfCallSite();
|
|
State* state = &m_ftlState;
|
|
VM* vm = &this->vm();
|
|
patchpoint->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
CallSiteIndex callSiteIndex =
|
|
state->jitCode->common.codeOrigins->addUniqueCallSiteIndex(codeOrigin);
|
|
|
|
Box<CCallHelpers::JumpList> exceptions =
|
|
exceptionHandle->scheduleExitCreation(params)->jumps(jit);
|
|
|
|
exceptionHandle->scheduleExitCreationForUnwind(params, callSiteIndex);
|
|
|
|
jit.store32(
|
|
CCallHelpers::TrustedImm32(callSiteIndex.bits()),
|
|
CCallHelpers::tagFor(VirtualRegister(CallFrameSlot::argumentCountIncludingThis)));
|
|
|
|
CallLinkInfo* callLinkInfo = jit.codeBlock()->addCallLinkInfo(node->origin.semantic);
|
|
|
|
RegisterSet usedRegisters = RegisterSet::allRegisters();
|
|
usedRegisters.exclude(RegisterSet::volatileRegistersForJSCall());
|
|
GPRReg calleeGPR = params[1].gpr();
|
|
usedRegisters.set(calleeGPR);
|
|
|
|
ScratchRegisterAllocator allocator(usedRegisters);
|
|
GPRReg scratchGPR1 = allocator.allocateScratchGPR();
|
|
GPRReg scratchGPR2 = allocator.allocateScratchGPR();
|
|
GPRReg scratchGPR3 = allocator.allocateScratchGPR();
|
|
GPRReg scratchGPR4 = allocator.allocateScratchGPR();
|
|
RELEASE_ASSERT(!allocator.numberOfReusedRegisters());
|
|
|
|
auto getValueFromRep = [&] (B3::ValueRep rep, GPRReg result) {
|
|
ASSERT(!usedRegisters.get(result));
|
|
|
|
if (rep.isConstant()) {
|
|
jit.move(CCallHelpers::Imm64(rep.value()), result);
|
|
return;
|
|
}
|
|
|
|
// Note: in this function, we only request 64 bit values.
|
|
if (rep.isStack()) {
|
|
jit.load64(
|
|
CCallHelpers::Address(GPRInfo::callFrameRegister, rep.offsetFromFP()),
|
|
result);
|
|
return;
|
|
}
|
|
|
|
RELEASE_ASSERT(rep.isGPR());
|
|
ASSERT(usedRegisters.get(rep.gpr()));
|
|
jit.move(rep.gpr(), result);
|
|
};
|
|
|
|
auto callWithExceptionCheck = [&] (void(*callee)(JSGlobalObject*)) {
|
|
jit.move(CCallHelpers::TrustedImmPtr(tagCFunction<OperationPtrTag>(callee)), GPRInfo::nonPreservedNonArgumentGPR0);
|
|
jit.call(GPRInfo::nonPreservedNonArgumentGPR0, OperationPtrTag);
|
|
exceptions->append(jit.emitExceptionCheck(*vm, AssemblyHelpers::NormalExceptionCheck, AssemblyHelpers::FarJumpWidth));
|
|
};
|
|
|
|
CCallHelpers::JumpList slowCase;
|
|
unsigned originalStackHeight = params.proc().frameSize();
|
|
|
|
{
|
|
unsigned numUsedSlots = WTF::roundUpToMultipleOf(stackAlignmentRegisters(), originalStackHeight / sizeof(EncodedJSValue));
|
|
B3::ValueRep argumentCountIncludingThisRep = params[3];
|
|
getValueFromRep(argumentCountIncludingThisRep, scratchGPR2);
|
|
slowCase.append(jit.branch32(CCallHelpers::Above, scratchGPR2, CCallHelpers::TrustedImm32(JSC::maxArguments + 1)));
|
|
|
|
jit.move(scratchGPR2, scratchGPR1);
|
|
jit.addPtr(CCallHelpers::TrustedImmPtr(static_cast<size_t>(numUsedSlots + CallFrame::headerSizeInRegisters)), scratchGPR1);
|
|
// scratchGPR1 now has the required frame size in Register units
|
|
// Round scratchGPR1 to next multiple of stackAlignmentRegisters()
|
|
jit.addPtr(CCallHelpers::TrustedImm32(stackAlignmentRegisters() - 1), scratchGPR1);
|
|
jit.andPtr(CCallHelpers::TrustedImm32(~(stackAlignmentRegisters() - 1)), scratchGPR1);
|
|
jit.negPtr(scratchGPR1);
|
|
jit.getEffectiveAddress(CCallHelpers::BaseIndex(GPRInfo::callFrameRegister, scratchGPR1, CCallHelpers::TimesEight), scratchGPR1);
|
|
|
|
// Before touching stack values, we should update the stack pointer to protect them from signal stack.
|
|
jit.addPtr(CCallHelpers::TrustedImm32(sizeof(CallerFrameAndPC)), scratchGPR1, CCallHelpers::stackPointerRegister);
|
|
|
|
jit.store32(scratchGPR2, CCallHelpers::Address(scratchGPR1, CallFrameSlot::argumentCountIncludingThis * static_cast<int>(sizeof(Register)) + PayloadOffset));
|
|
|
|
int storeOffset = CallFrame::thisArgumentOffset() * static_cast<int>(sizeof(Register));
|
|
|
|
unsigned paramsOffset = 4;
|
|
unsigned index = 0;
|
|
auto emitArgumentsFromRightToLeft = recursableLambda([&](auto self, Node* target) -> void {
|
|
if (target->op() == PhantomSpread) {
|
|
self(target->child1().node());
|
|
return;
|
|
}
|
|
|
|
if (target->op() == PhantomNewArrayWithSpread) {
|
|
BitVector* bitVector = target->bitVector();
|
|
for (unsigned i = target->numChildren(); i--; ) {
|
|
if (bitVector->get(i))
|
|
self(state->graph.varArgChild(target, i).node());
|
|
else {
|
|
jit.subPtr(CCallHelpers::TrustedImmPtr(static_cast<size_t>(1)), scratchGPR2);
|
|
getValueFromRep(params[paramsOffset + (index++)], scratchGPR3);
|
|
jit.store64(scratchGPR3,
|
|
CCallHelpers::BaseIndex(scratchGPR1, scratchGPR2, CCallHelpers::TimesEight, storeOffset));
|
|
}
|
|
}
|
|
return;
|
|
}
|
|
|
|
if (target->op() == PhantomNewArrayBuffer) {
|
|
auto* array = target->castOperand<JSImmutableButterfly*>();
|
|
Checked<int32_t> offsetCount { 1 };
|
|
for (unsigned i = array->length(); i--; ++offsetCount) {
|
|
// Because varargs values are drained as JSValue, we should not generate value
|
|
// in Double form even if PhantomNewArrayBuffer's indexingType is ArrayWithDouble.
|
|
int64_t value = JSValue::encode(array->get(i));
|
|
jit.move(CCallHelpers::TrustedImm64(value), scratchGPR3);
|
|
Checked<int32_t> currentStoreOffset { storeOffset };
|
|
currentStoreOffset -= (offsetCount * static_cast<int32_t>(sizeof(Register)));
|
|
jit.store64(scratchGPR3,
|
|
CCallHelpers::BaseIndex(scratchGPR1, scratchGPR2, CCallHelpers::TimesEight, currentStoreOffset.unsafeGet()));
|
|
}
|
|
jit.subPtr(CCallHelpers::TrustedImmPtr(static_cast<size_t>(array->length())), scratchGPR2);
|
|
return;
|
|
}
|
|
|
|
RELEASE_ASSERT(target->op() == PhantomCreateRest);
|
|
InlineCallFrame* inlineCallFrame = target->origin.semantic.inlineCallFrame();
|
|
|
|
unsigned numberOfArgumentsToSkip = target->numberOfArgumentsToSkip();
|
|
|
|
B3::ValueRep numArgumentsToCopy = params[paramsOffset + (index++)];
|
|
getValueFromRep(numArgumentsToCopy, scratchGPR3);
|
|
int loadOffset = (AssemblyHelpers::argumentsStart(inlineCallFrame).offset() + numberOfArgumentsToSkip) * static_cast<int>(sizeof(Register));
|
|
|
|
auto done = jit.branchTestPtr(MacroAssembler::Zero, scratchGPR3);
|
|
auto loopStart = jit.label();
|
|
jit.subPtr(CCallHelpers::TrustedImmPtr(static_cast<size_t>(1)), scratchGPR3);
|
|
jit.subPtr(CCallHelpers::TrustedImmPtr(static_cast<size_t>(1)), scratchGPR2);
|
|
jit.load64(CCallHelpers::BaseIndex(GPRInfo::callFrameRegister, scratchGPR3, CCallHelpers::TimesEight, loadOffset), scratchGPR4);
|
|
jit.store64(scratchGPR4,
|
|
CCallHelpers::BaseIndex(scratchGPR1, scratchGPR2, CCallHelpers::TimesEight, storeOffset));
|
|
jit.branchTestPtr(CCallHelpers::NonZero, scratchGPR3).linkTo(loopStart, &jit);
|
|
done.link(&jit);
|
|
});
|
|
emitArgumentsFromRightToLeft(arguments);
|
|
}
|
|
|
|
{
|
|
CCallHelpers::Jump dontThrow = jit.jump();
|
|
slowCase.link(&jit);
|
|
jit.setupArguments<decltype(operationThrowStackOverflowForVarargs)>(jit.codeBlock()->globalObjectFor(node->origin.semantic));
|
|
jit.prepareCallOperation(jit.vm());
|
|
callWithExceptionCheck(operationThrowStackOverflowForVarargs);
|
|
jit.abortWithReason(DFGVarargsThrowingPathDidNotThrow);
|
|
|
|
dontThrow.link(&jit);
|
|
}
|
|
|
|
ASSERT(calleeGPR == GPRInfo::regT0);
|
|
jit.store64(calleeGPR, CCallHelpers::calleeFrameSlot(CallFrameSlot::callee));
|
|
getValueFromRep(params[2], scratchGPR3);
|
|
jit.store64(scratchGPR3, CCallHelpers::calleeArgumentSlot(0));
|
|
|
|
CallLinkInfo::CallType callType;
|
|
if (node->op() == ConstructVarargs || node->op() == ConstructForwardVarargs)
|
|
callType = CallLinkInfo::ConstructVarargs;
|
|
else if (node->op() == TailCallVarargs || node->op() == TailCallForwardVarargs)
|
|
callType = CallLinkInfo::TailCallVarargs;
|
|
else
|
|
callType = CallLinkInfo::CallVarargs;
|
|
|
|
bool isTailCall = CallLinkInfo::callModeFor(callType) == CallMode::Tail;
|
|
|
|
CCallHelpers::DataLabelPtr targetToCheck;
|
|
CCallHelpers::Jump slowPath = jit.branchPtrWithPatch(
|
|
CCallHelpers::NotEqual, GPRInfo::regT0, targetToCheck,
|
|
CCallHelpers::TrustedImmPtr(nullptr));
|
|
|
|
CCallHelpers::Call fastCall;
|
|
CCallHelpers::Jump done;
|
|
|
|
if (isTailCall) {
|
|
jit.emitRestoreCalleeSaves();
|
|
jit.prepareForTailCallSlow();
|
|
fastCall = jit.nearTailCall();
|
|
} else {
|
|
fastCall = jit.nearCall();
|
|
done = jit.jump();
|
|
}
|
|
|
|
slowPath.link(&jit);
|
|
|
|
if (isTailCall)
|
|
jit.emitRestoreCalleeSaves();
|
|
ASSERT(!usedRegisters.get(GPRInfo::regT2));
|
|
jit.move(CCallHelpers::TrustedImmPtr(callLinkInfo), GPRInfo::regT2);
|
|
jit.move(CCallHelpers::TrustedImmPtr(jit.codeBlock()->globalObjectFor(node->origin.semantic)), GPRInfo::regT3);
|
|
CCallHelpers::Call slowCall = jit.nearCall();
|
|
|
|
if (isTailCall)
|
|
jit.abortWithReason(JITDidReturnFromTailCall);
|
|
else
|
|
done.link(&jit);
|
|
|
|
callLinkInfo->setUpCall(callType, GPRInfo::regT0);
|
|
|
|
jit.addPtr(
|
|
CCallHelpers::TrustedImm32(-originalStackHeight),
|
|
GPRInfo::callFrameRegister, CCallHelpers::stackPointerRegister);
|
|
|
|
jit.addLinkTask(
|
|
[=] (LinkBuffer& linkBuffer) {
|
|
MacroAssemblerCodePtr<JITThunkPtrTag> linkCall = vm->getCTIStub(linkCallThunkGenerator).code();
|
|
linkBuffer.link(slowCall, FunctionPtr<JITThunkPtrTag>(linkCall));
|
|
|
|
callLinkInfo->setCallLocations(
|
|
CodeLocationLabel<JSInternalPtrTag>(linkBuffer.locationOfNearCall<JSInternalPtrTag>(slowCall)),
|
|
CodeLocationLabel<JSInternalPtrTag>(linkBuffer.locationOf<JSInternalPtrTag>(targetToCheck)),
|
|
linkBuffer.locationOfNearCall<JSInternalPtrTag>(fastCall));
|
|
});
|
|
});
|
|
|
|
switch (node->op()) {
|
|
case TailCallForwardVarargs:
|
|
m_out.unreachable();
|
|
break;
|
|
|
|
default:
|
|
setJSValue(patchpoint);
|
|
break;
|
|
}
|
|
}
|
|
|
|
void compileCallOrConstructVarargs()
|
|
{
|
|
Node* node = m_node;
|
|
LValue jsCallee = lowJSValue(m_node->child1());
|
|
LValue thisArg = lowJSValue(m_node->child2());
|
|
|
|
LValue jsArguments = nullptr;
|
|
bool forwarding = false;
|
|
|
|
switch (node->op()) {
|
|
case CallVarargs:
|
|
case TailCallVarargs:
|
|
case TailCallVarargsInlinedCaller:
|
|
case ConstructVarargs:
|
|
jsArguments = lowJSValue(node->child3());
|
|
break;
|
|
case CallForwardVarargs:
|
|
case TailCallForwardVarargs:
|
|
case TailCallForwardVarargsInlinedCaller:
|
|
case ConstructForwardVarargs:
|
|
forwarding = true;
|
|
break;
|
|
default:
|
|
DFG_CRASH(m_graph, node, "bad node type");
|
|
break;
|
|
}
|
|
|
|
if (forwarding && m_node->child3()) {
|
|
Node* arguments = m_node->child3().node();
|
|
if (arguments->op() == PhantomNewArrayWithSpread || arguments->op() == PhantomNewArrayBuffer || arguments->op() == PhantomSpread) {
|
|
compileCallOrConstructVarargsSpread();
|
|
return;
|
|
}
|
|
}
|
|
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int64);
|
|
|
|
// Append the forms of the arguments that we will use before any clobbering happens.
|
|
patchpoint->append(jsCallee, ValueRep::reg(GPRInfo::regT0));
|
|
if (jsArguments)
|
|
patchpoint->appendSomeRegister(jsArguments);
|
|
patchpoint->appendSomeRegister(thisArg);
|
|
|
|
if (!forwarding) {
|
|
// Now append them again for after clobbering. Note that the compiler may ask us to use a
|
|
// different register for the late for the post-clobbering version of the value. This gives
|
|
// the compiler a chance to spill these values without having to burn any callee-saves.
|
|
patchpoint->append(jsCallee, ValueRep::LateColdAny);
|
|
patchpoint->append(jsArguments, ValueRep::LateColdAny);
|
|
patchpoint->append(thisArg, ValueRep::LateColdAny);
|
|
}
|
|
|
|
RefPtr<PatchpointExceptionHandle> exceptionHandle =
|
|
preparePatchpointForExceptions(patchpoint);
|
|
|
|
patchpoint->append(m_notCellMask, ValueRep::reg(GPRInfo::notCellMaskRegister));
|
|
patchpoint->append(m_numberTag, ValueRep::reg(GPRInfo::numberTagRegister));
|
|
|
|
patchpoint->clobber(RegisterSet::macroScratchRegisters());
|
|
patchpoint->clobberLate(RegisterSet::volatileRegistersForJSCall());
|
|
patchpoint->resultConstraints = { ValueRep::reg(GPRInfo::returnValueGPR) };
|
|
|
|
// This is the minimum amount of call arg area stack space that all JS->JS calls always have.
|
|
unsigned minimumJSCallAreaSize =
|
|
sizeof(CallerFrameAndPC) +
|
|
WTF::roundUpToMultipleOf(stackAlignmentBytes(), 5 * sizeof(EncodedJSValue));
|
|
|
|
m_proc.requestCallArgAreaSizeInBytes(minimumJSCallAreaSize);
|
|
|
|
CodeOrigin codeOrigin = codeOriginDescriptionOfCallSite();
|
|
State* state = &m_ftlState;
|
|
VM* vm = &this->vm();
|
|
patchpoint->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
CallSiteIndex callSiteIndex =
|
|
state->jitCode->common.codeOrigins->addUniqueCallSiteIndex(codeOrigin);
|
|
|
|
Box<CCallHelpers::JumpList> exceptions =
|
|
exceptionHandle->scheduleExitCreation(params)->jumps(jit);
|
|
|
|
exceptionHandle->scheduleExitCreationForUnwind(params, callSiteIndex);
|
|
|
|
jit.store32(
|
|
CCallHelpers::TrustedImm32(callSiteIndex.bits()),
|
|
CCallHelpers::tagFor(VirtualRegister(CallFrameSlot::argumentCountIncludingThis)));
|
|
|
|
CallLinkInfo* callLinkInfo = jit.codeBlock()->addCallLinkInfo(node->origin.semantic);
|
|
CallVarargsData* data = node->callVarargsData();
|
|
|
|
unsigned argIndex = 1;
|
|
GPRReg calleeGPR = params[argIndex++].gpr();
|
|
ASSERT(calleeGPR == GPRInfo::regT0);
|
|
GPRReg argumentsGPR = jsArguments ? params[argIndex++].gpr() : InvalidGPRReg;
|
|
GPRReg thisGPR = params[argIndex++].gpr();
|
|
|
|
B3::ValueRep calleeLateRep;
|
|
B3::ValueRep argumentsLateRep;
|
|
B3::ValueRep thisLateRep;
|
|
if (!forwarding) {
|
|
// If we're not forwarding then we'll need callee, arguments, and this after we
|
|
// have potentially clobbered calleeGPR, argumentsGPR, and thisGPR. Our technique
|
|
// for this is to supply all of those operands as late uses in addition to
|
|
// specifying them as early uses. It's possible that the late use uses a spill
|
|
// while the early use uses a register, and it's possible for the late and early
|
|
// uses to use different registers. We do know that the late uses interfere with
|
|
// all volatile registers and so won't use those, but the early uses may use
|
|
// volatile registers and in the case of calleeGPR, it's pinned to regT0 so it
|
|
// definitely will.
|
|
//
|
|
// Note that we have to be super careful with these. It's possible that these
|
|
// use a shuffling of the registers used for calleeGPR, argumentsGPR, and
|
|
// thisGPR. If that happens and we do for example:
|
|
//
|
|
// calleeLateRep.emitRestore(jit, calleeGPR);
|
|
// argumentsLateRep.emitRestore(jit, calleeGPR);
|
|
//
|
|
// Then we might end up with garbage if calleeLateRep.gpr() == argumentsGPR and
|
|
// argumentsLateRep.gpr() == calleeGPR.
|
|
//
|
|
// We do a variety of things to prevent this from happening. For example, we use
|
|
// argumentsLateRep before needing the other two and after we've already stopped
|
|
// using the *GPRs. Also, we pin calleeGPR to regT0, and rely on the fact that
|
|
// the *LateReps cannot use volatile registers (so they cannot be regT0, so
|
|
// calleeGPR != argumentsLateRep.gpr() and calleeGPR != thisLateRep.gpr()).
|
|
//
|
|
// An alternative would have been to just use early uses and early-clobber all
|
|
// volatile registers. But that would force callee, arguments, and this into
|
|
// callee-save registers even if we have to spill them. We don't want spilling to
|
|
// use up three callee-saves.
|
|
//
|
|
// TL;DR: The way we use LateReps here is dangerous and barely works but achieves
|
|
// some desirable performance properties, so don't mistake the cleverness for
|
|
// elegance.
|
|
calleeLateRep = params[argIndex++];
|
|
argumentsLateRep = params[argIndex++];
|
|
thisLateRep = params[argIndex++];
|
|
}
|
|
|
|
// Get some scratch registers.
|
|
RegisterSet usedRegisters;
|
|
usedRegisters.merge(RegisterSet::stackRegisters());
|
|
usedRegisters.merge(RegisterSet::reservedHardwareRegisters());
|
|
usedRegisters.merge(RegisterSet::calleeSaveRegisters());
|
|
usedRegisters.set(calleeGPR);
|
|
if (argumentsGPR != InvalidGPRReg)
|
|
usedRegisters.set(argumentsGPR);
|
|
usedRegisters.set(thisGPR);
|
|
if (calleeLateRep.isReg())
|
|
usedRegisters.set(calleeLateRep.reg());
|
|
if (argumentsLateRep.isReg())
|
|
usedRegisters.set(argumentsLateRep.reg());
|
|
if (thisLateRep.isReg())
|
|
usedRegisters.set(thisLateRep.reg());
|
|
ScratchRegisterAllocator allocator(usedRegisters);
|
|
GPRReg scratchGPR1 = allocator.allocateScratchGPR();
|
|
GPRReg scratchGPR2 = allocator.allocateScratchGPR();
|
|
GPRReg scratchGPR3 = forwarding ? allocator.allocateScratchGPR() : InvalidGPRReg;
|
|
RELEASE_ASSERT(!allocator.numberOfReusedRegisters());
|
|
|
|
auto callWithExceptionCheck = [&] (void(*callee)()) {
|
|
jit.move(CCallHelpers::TrustedImmPtr(tagCFunction<OperationPtrTag>(callee)), GPRInfo::nonPreservedNonArgumentGPR0);
|
|
jit.call(GPRInfo::nonPreservedNonArgumentGPR0, OperationPtrTag);
|
|
exceptions->append(jit.emitExceptionCheck(*vm, AssemblyHelpers::NormalExceptionCheck, AssemblyHelpers::FarJumpWidth));
|
|
};
|
|
|
|
unsigned originalStackHeight = params.proc().frameSize();
|
|
|
|
if (forwarding) {
|
|
jit.move(CCallHelpers::TrustedImm32(originalStackHeight / sizeof(EncodedJSValue)), scratchGPR2);
|
|
|
|
CCallHelpers::JumpList slowCase;
|
|
InlineCallFrame* inlineCallFrame;
|
|
if (node->child3())
|
|
inlineCallFrame = node->child3()->origin.semantic.inlineCallFrame();
|
|
else
|
|
inlineCallFrame = node->origin.semantic.inlineCallFrame();
|
|
|
|
// emitSetupVarargsFrameFastCase modifies the stack pointer if it succeeds.
|
|
emitSetupVarargsFrameFastCase(*vm, jit, scratchGPR2, scratchGPR1, scratchGPR2, scratchGPR3, inlineCallFrame, data->firstVarArgOffset, slowCase);
|
|
|
|
CCallHelpers::Jump done = jit.jump();
|
|
slowCase.link(&jit);
|
|
jit.setupArguments<decltype(operationThrowStackOverflowForVarargs)>(jit.codeBlock()->globalObjectFor(node->origin.semantic));
|
|
jit.prepareCallOperation(jit.vm());
|
|
callWithExceptionCheck(bitwise_cast<void(*)()>(operationThrowStackOverflowForVarargs));
|
|
jit.abortWithReason(DFGVarargsThrowingPathDidNotThrow);
|
|
|
|
done.link(&jit);
|
|
} else {
|
|
jit.move(CCallHelpers::TrustedImm32(originalStackHeight / sizeof(EncodedJSValue)), scratchGPR1);
|
|
jit.setupArguments<decltype(operationSizeFrameForVarargs)>(jit.codeBlock()->globalObjectFor(node->origin.semantic), argumentsGPR, scratchGPR1, CCallHelpers::TrustedImm32(data->firstVarArgOffset));
|
|
jit.prepareCallOperation(jit.vm());
|
|
callWithExceptionCheck(bitwise_cast<void(*)()>(operationSizeFrameForVarargs));
|
|
|
|
jit.move(GPRInfo::returnValueGPR, scratchGPR1);
|
|
jit.move(CCallHelpers::TrustedImm32(originalStackHeight / sizeof(EncodedJSValue)), scratchGPR2);
|
|
argumentsLateRep.emitRestore(jit, argumentsGPR);
|
|
emitSetVarargsFrame(jit, scratchGPR1, false, scratchGPR2, scratchGPR2);
|
|
jit.addPtr(CCallHelpers::TrustedImm32(-minimumJSCallAreaSize), scratchGPR2, CCallHelpers::stackPointerRegister);
|
|
jit.setupArguments<decltype(operationSetupVarargsFrame)>(jit.codeBlock()->globalObjectFor(node->origin.semantic), scratchGPR2, argumentsGPR, CCallHelpers::TrustedImm32(data->firstVarArgOffset), scratchGPR1);
|
|
jit.prepareCallOperation(jit.vm());
|
|
callWithExceptionCheck(bitwise_cast<void(*)()>(operationSetupVarargsFrame));
|
|
|
|
jit.addPtr(CCallHelpers::TrustedImm32(sizeof(CallerFrameAndPC)), GPRInfo::returnValueGPR, CCallHelpers::stackPointerRegister);
|
|
|
|
calleeLateRep.emitRestore(jit, GPRInfo::regT0);
|
|
|
|
// This may not emit code if thisGPR got a callee-save. Also, we're guaranteed
|
|
// that thisGPR != GPRInfo::regT0 because regT0 interferes with it.
|
|
thisLateRep.emitRestore(jit, thisGPR);
|
|
}
|
|
|
|
jit.store64(GPRInfo::regT0, CCallHelpers::calleeFrameSlot(CallFrameSlot::callee));
|
|
jit.store64(thisGPR, CCallHelpers::calleeArgumentSlot(0));
|
|
|
|
CallLinkInfo::CallType callType;
|
|
if (node->op() == ConstructVarargs || node->op() == ConstructForwardVarargs)
|
|
callType = CallLinkInfo::ConstructVarargs;
|
|
else if (node->op() == TailCallVarargs || node->op() == TailCallForwardVarargs)
|
|
callType = CallLinkInfo::TailCallVarargs;
|
|
else
|
|
callType = CallLinkInfo::CallVarargs;
|
|
|
|
bool isTailCall = CallLinkInfo::callModeFor(callType) == CallMode::Tail;
|
|
|
|
CCallHelpers::DataLabelPtr targetToCheck;
|
|
CCallHelpers::Jump slowPath = jit.branchPtrWithPatch(
|
|
CCallHelpers::NotEqual, GPRInfo::regT0, targetToCheck,
|
|
CCallHelpers::TrustedImmPtr(nullptr));
|
|
|
|
CCallHelpers::Call fastCall;
|
|
CCallHelpers::Jump done;
|
|
|
|
if (isTailCall) {
|
|
jit.emitRestoreCalleeSaves();
|
|
jit.prepareForTailCallSlow();
|
|
fastCall = jit.nearTailCall();
|
|
} else {
|
|
fastCall = jit.nearCall();
|
|
done = jit.jump();
|
|
}
|
|
|
|
slowPath.link(&jit);
|
|
|
|
if (isTailCall)
|
|
jit.emitRestoreCalleeSaves();
|
|
jit.move(CCallHelpers::TrustedImmPtr(callLinkInfo), GPRInfo::regT2);
|
|
jit.move(CCallHelpers::TrustedImmPtr(jit.codeBlock()->globalObjectFor(node->origin.semantic)), GPRInfo::regT3);
|
|
CCallHelpers::Call slowCall = jit.nearCall();
|
|
|
|
if (isTailCall)
|
|
jit.abortWithReason(JITDidReturnFromTailCall);
|
|
else
|
|
done.link(&jit);
|
|
|
|
callLinkInfo->setUpCall(callType, GPRInfo::regT0);
|
|
|
|
jit.addPtr(
|
|
CCallHelpers::TrustedImm32(-originalStackHeight),
|
|
GPRInfo::callFrameRegister, CCallHelpers::stackPointerRegister);
|
|
|
|
jit.addLinkTask(
|
|
[=] (LinkBuffer& linkBuffer) {
|
|
MacroAssemblerCodePtr<JITThunkPtrTag> linkCall = vm->getCTIStub(linkCallThunkGenerator).code();
|
|
linkBuffer.link(slowCall, FunctionPtr<JITThunkPtrTag>(linkCall));
|
|
|
|
callLinkInfo->setCallLocations(
|
|
CodeLocationLabel<JSInternalPtrTag>(linkBuffer.locationOfNearCall<JSInternalPtrTag>(slowCall)),
|
|
CodeLocationLabel<JSInternalPtrTag>(linkBuffer.locationOf<JSInternalPtrTag>(targetToCheck)),
|
|
linkBuffer.locationOfNearCall<JSInternalPtrTag>(fastCall));
|
|
});
|
|
});
|
|
|
|
switch (node->op()) {
|
|
case TailCallVarargs:
|
|
case TailCallForwardVarargs:
|
|
m_out.unreachable();
|
|
break;
|
|
|
|
default:
|
|
setJSValue(patchpoint);
|
|
break;
|
|
}
|
|
}
|
|
|
|
void compileCallEval()
|
|
{
|
|
Node* node = m_node;
|
|
unsigned numArgs = node->numChildren() - 1;
|
|
|
|
LValue jsCallee = lowJSValue(m_graph.varArgChild(node, 0));
|
|
|
|
unsigned frameSize = (CallFrame::headerSizeInRegisters + numArgs) * sizeof(EncodedJSValue);
|
|
unsigned alignedFrameSize = WTF::roundUpToMultipleOf(stackAlignmentBytes(), frameSize);
|
|
|
|
m_proc.requestCallArgAreaSizeInBytes(alignedFrameSize);
|
|
|
|
Vector<ConstrainedValue> arguments;
|
|
arguments.append(ConstrainedValue(jsCallee, ValueRep::reg(GPRInfo::regT0)));
|
|
|
|
auto addArgument = [&] (LValue value, VirtualRegister reg, int offset) {
|
|
intptr_t offsetFromSP =
|
|
(reg.offset() - CallerFrameAndPC::sizeInRegisters) * sizeof(EncodedJSValue) + offset;
|
|
arguments.append(ConstrainedValue(value, ValueRep::stackArgument(offsetFromSP)));
|
|
};
|
|
|
|
addArgument(jsCallee, VirtualRegister(CallFrameSlot::callee), 0);
|
|
addArgument(m_out.constInt32(numArgs), VirtualRegister(CallFrameSlot::argumentCountIncludingThis), PayloadOffset);
|
|
for (unsigned i = 0; i < numArgs; ++i)
|
|
addArgument(lowJSValue(m_graph.varArgChild(node, 1 + i)), virtualRegisterForArgumentIncludingThis(i), 0);
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int64);
|
|
patchpoint->appendVector(arguments);
|
|
|
|
RefPtr<PatchpointExceptionHandle> exceptionHandle = preparePatchpointForExceptions(patchpoint);
|
|
|
|
patchpoint->append(m_notCellMask, ValueRep::reg(GPRInfo::notCellMaskRegister));
|
|
patchpoint->append(m_numberTag, ValueRep::reg(GPRInfo::numberTagRegister));
|
|
patchpoint->clobber(RegisterSet::macroScratchRegisters());
|
|
patchpoint->clobberLate(RegisterSet::volatileRegistersForJSCall());
|
|
patchpoint->resultConstraints = { ValueRep::reg(GPRInfo::returnValueGPR) };
|
|
|
|
CodeOrigin codeOrigin = codeOriginDescriptionOfCallSite();
|
|
State* state = &m_ftlState;
|
|
VM& vm = this->vm();
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
patchpoint->setGenerator(
|
|
[=, &vm] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
CallSiteIndex callSiteIndex = state->jitCode->common.codeOrigins->addUniqueCallSiteIndex(codeOrigin);
|
|
|
|
Box<CCallHelpers::JumpList> exceptions = exceptionHandle->scheduleExitCreation(params)->jumps(jit);
|
|
|
|
exceptionHandle->scheduleExitCreationForUnwind(params, callSiteIndex);
|
|
|
|
jit.store32(
|
|
CCallHelpers::TrustedImm32(callSiteIndex.bits()),
|
|
CCallHelpers::tagFor(VirtualRegister(CallFrameSlot::argumentCountIncludingThis)));
|
|
|
|
CallLinkInfo* callLinkInfo = jit.codeBlock()->addCallLinkInfo(node->origin.semantic);
|
|
callLinkInfo->setUpCall(CallLinkInfo::Call, GPRInfo::regT0);
|
|
|
|
jit.addPtr(CCallHelpers::TrustedImm32(-static_cast<ptrdiff_t>(sizeof(CallerFrameAndPC))), CCallHelpers::stackPointerRegister, GPRInfo::regT1);
|
|
jit.storePtr(GPRInfo::callFrameRegister, CCallHelpers::Address(GPRInfo::regT1, CallFrame::callerFrameOffset()));
|
|
|
|
// Now we need to make room for:
|
|
// - The caller frame and PC for a call to operationCallEval.
|
|
// - Potentially two arguments on the stack.
|
|
unsigned requiredBytes = sizeof(CallerFrameAndPC) + sizeof(CallFrame*) * 2;
|
|
requiredBytes = WTF::roundUpToMultipleOf(stackAlignmentBytes(), requiredBytes);
|
|
jit.subPtr(CCallHelpers::TrustedImm32(requiredBytes), CCallHelpers::stackPointerRegister);
|
|
jit.move(CCallHelpers::TrustedImm32(node->ecmaMode().value()), GPRInfo::regT2);
|
|
jit.setupArguments<decltype(operationCallEval)>(globalObject, GPRInfo::regT1, GPRInfo::regT2);
|
|
jit.prepareCallOperation(vm);
|
|
jit.move(CCallHelpers::TrustedImmPtr(tagCFunction<OperationPtrTag>(operationCallEval)), GPRInfo::nonPreservedNonArgumentGPR0);
|
|
jit.call(GPRInfo::nonPreservedNonArgumentGPR0, OperationPtrTag);
|
|
exceptions->append(jit.emitExceptionCheck(state->vm(), AssemblyHelpers::NormalExceptionCheck, AssemblyHelpers::FarJumpWidth));
|
|
|
|
CCallHelpers::Jump done = jit.branchTest64(CCallHelpers::NonZero, GPRInfo::returnValueGPR);
|
|
|
|
jit.addPtr(CCallHelpers::TrustedImm32(requiredBytes), CCallHelpers::stackPointerRegister);
|
|
jit.load64(CCallHelpers::calleeFrameSlot(CallFrameSlot::callee), GPRInfo::regT0);
|
|
jit.emitVirtualCall(vm, globalObject, callLinkInfo);
|
|
|
|
done.link(&jit);
|
|
jit.addPtr(
|
|
CCallHelpers::TrustedImm32(-params.proc().frameSize()),
|
|
GPRInfo::callFrameRegister, CCallHelpers::stackPointerRegister);
|
|
});
|
|
|
|
setJSValue(patchpoint);
|
|
}
|
|
|
|
void compileVarargsLength()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LoadVarargsData* data = m_node->loadVarargsData();
|
|
LValue jsArguments = lowJSValue(m_node->argumentsChild());
|
|
|
|
LValue length = m_out.castToInt32(vmCall(Int64, operationSizeOfVarargs, weakPointer(globalObject), jsArguments, m_out.constInt32(data->offset)));
|
|
|
|
LValue lengthIncludingThis = m_out.add(length, m_out.int32One);
|
|
|
|
setInt32(lengthIncludingThis);
|
|
}
|
|
|
|
void compileLoadVarargs()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LoadVarargsData* data = m_node->loadVarargsData();
|
|
LValue jsArguments = lowJSValue(m_node->argumentsChild());
|
|
LValue lengthIncludingThis = lowInt32(m_node->child1());
|
|
|
|
speculate(
|
|
VarargsOverflow, noValue(), nullptr,
|
|
m_out.bitOr(m_out.isZero32(lengthIncludingThis), m_out.above(lengthIncludingThis, m_out.constInt32(data->limit))));
|
|
|
|
m_out.store32(lengthIncludingThis, payloadFor(data->machineCount));
|
|
|
|
// FIXME: This computation is rather silly. If operationLoadVarargs just took a pointer instead
|
|
// of a VirtualRegister, we wouldn't have to do this.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=141660
|
|
LValue machineStart = m_out.lShr(
|
|
m_out.sub(addressFor(data->machineStart).value(), m_callFrame),
|
|
m_out.constIntPtr(3));
|
|
|
|
vmCall(
|
|
Void, operationLoadVarargs, weakPointer(globalObject),
|
|
m_out.castToInt32(machineStart), jsArguments, m_out.constInt32(data->offset),
|
|
lengthIncludingThis, m_out.constInt32(data->mandatoryMinimum));
|
|
}
|
|
|
|
void compileForwardVarargs()
|
|
{
|
|
if (m_node->argumentsChild()) {
|
|
Node* arguments = m_node->argumentsChild().node();
|
|
if (arguments->op() == PhantomNewArrayWithSpread || arguments->op() == PhantomNewArrayBuffer || arguments->op() == PhantomSpread) {
|
|
compileForwardVarargsWithSpread();
|
|
return;
|
|
}
|
|
}
|
|
|
|
LoadVarargsData* data = m_node->loadVarargsData();
|
|
InlineCallFrame* inlineCallFrame;
|
|
if (m_node->argumentsChild())
|
|
inlineCallFrame = m_node->argumentsChild()->origin.semantic.inlineCallFrame();
|
|
else
|
|
inlineCallFrame = m_origin.semantic.inlineCallFrame();
|
|
|
|
unsigned numberOfArgumentsToSkip = data->offset;
|
|
LValue lengthIncludingThis = lowInt32(m_node->child1());
|
|
|
|
LValue length = m_out.sub(lengthIncludingThis, m_out.int32One);
|
|
speculate(
|
|
VarargsOverflow, noValue(), nullptr,
|
|
m_out.above(lengthIncludingThis, m_out.constInt32(data->limit)));
|
|
|
|
m_out.store32(lengthIncludingThis, payloadFor(data->machineCount));
|
|
|
|
LValue sourceStart = getArgumentsStart(inlineCallFrame, numberOfArgumentsToSkip);
|
|
LValue targetStart = addressFor(data->machineStart).value();
|
|
|
|
LBasicBlock undefinedLoop = m_out.newBlock();
|
|
LBasicBlock mainLoopEntry = m_out.newBlock();
|
|
LBasicBlock mainLoop = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue lengthAsPtr = m_out.zeroExtPtr(length);
|
|
LValue loopBoundValue = m_out.constIntPtr(data->mandatoryMinimum);
|
|
ValueFromBlock loopBound = m_out.anchor(loopBoundValue);
|
|
m_out.branch(
|
|
m_out.above(loopBoundValue, lengthAsPtr), unsure(undefinedLoop), unsure(mainLoopEntry));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(undefinedLoop, mainLoopEntry);
|
|
LValue previousIndex = m_out.phi(pointerType(), loopBound);
|
|
LValue currentIndex = m_out.sub(previousIndex, m_out.intPtrOne);
|
|
m_out.store64(
|
|
m_out.constInt64(JSValue::encode(jsUndefined())),
|
|
m_out.baseIndex(m_heaps.variables, targetStart, currentIndex));
|
|
ValueFromBlock nextIndex = m_out.anchor(currentIndex);
|
|
m_out.addIncomingToPhi(previousIndex, nextIndex);
|
|
m_out.branch(
|
|
m_out.above(currentIndex, lengthAsPtr), unsure(undefinedLoop), unsure(mainLoopEntry));
|
|
|
|
m_out.appendTo(mainLoopEntry, mainLoop);
|
|
loopBound = m_out.anchor(lengthAsPtr);
|
|
m_out.branch(m_out.notNull(lengthAsPtr), unsure(mainLoop), unsure(continuation));
|
|
|
|
m_out.appendTo(mainLoop, continuation);
|
|
previousIndex = m_out.phi(pointerType(), loopBound);
|
|
currentIndex = m_out.sub(previousIndex, m_out.intPtrOne);
|
|
LValue value = m_out.load64(
|
|
m_out.baseIndex(m_heaps.variables, sourceStart, currentIndex));
|
|
m_out.store64(value, m_out.baseIndex(m_heaps.variables, targetStart, currentIndex));
|
|
nextIndex = m_out.anchor(currentIndex);
|
|
m_out.addIncomingToPhi(previousIndex, nextIndex);
|
|
m_out.branch(m_out.isNull(currentIndex), unsure(continuation), unsure(mainLoop));
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
}
|
|
|
|
LValue getSpreadLengthFromInlineCallFrame(InlineCallFrame* inlineCallFrame, unsigned numberOfArgumentsToSkip)
|
|
{
|
|
ArgumentsLength argumentsLength = getArgumentsLength(inlineCallFrame);
|
|
if (argumentsLength.isKnown) {
|
|
unsigned knownLength = argumentsLength.known;
|
|
if (knownLength >= numberOfArgumentsToSkip)
|
|
knownLength = knownLength - numberOfArgumentsToSkip;
|
|
else
|
|
knownLength = 0;
|
|
return m_out.constInt32(knownLength);
|
|
}
|
|
|
|
|
|
// We need to perform the same logical operation as the code above, but through dynamic operations.
|
|
if (!numberOfArgumentsToSkip)
|
|
return argumentsLength.value;
|
|
|
|
RELEASE_ASSERT(numberOfArgumentsToSkip < static_cast<unsigned>(INT32_MIN));
|
|
|
|
LValue fixedLength = m_out.sub(argumentsLength.value, m_out.constInt32(numberOfArgumentsToSkip));
|
|
|
|
return m_out.select(m_out.greaterThanOrEqual(fixedLength, m_out.int32Zero), fixedLength, m_out.int32Zero, SelectPredictability::Predictable);
|
|
}
|
|
|
|
void compileForwardVarargsWithSpread()
|
|
{
|
|
Node* arguments = m_node->argumentsChild().node();
|
|
RELEASE_ASSERT(arguments->op() == PhantomNewArrayWithSpread || arguments->op() == PhantomNewArrayBuffer || arguments->op() == PhantomSpread);
|
|
|
|
LValue lengthIncludingThis = lowInt32(m_node->child1());
|
|
|
|
LoadVarargsData* data = m_node->loadVarargsData();
|
|
speculate(
|
|
VarargsOverflow, noValue(), nullptr,
|
|
m_out.above(lengthIncludingThis, m_out.constInt32(data->limit)));
|
|
|
|
m_out.store32(lengthIncludingThis, payloadFor(data->machineCount));
|
|
|
|
LValue targetStart = addressFor(data->machineStart).value();
|
|
|
|
auto forwardSpread = recursableLambda([this, &targetStart](auto self, Node* target, LValue storeIndex) -> LValue {
|
|
if (target->op() == PhantomSpread)
|
|
return self(target->child1().node(), storeIndex);
|
|
|
|
if (target->op() == PhantomNewArrayWithSpread) {
|
|
BitVector* bitVector = target->bitVector();
|
|
for (unsigned i = 0; i < target->numChildren(); i++) {
|
|
if (bitVector->get(i))
|
|
storeIndex = self(m_graph.varArgChild(target, i).node(), storeIndex);
|
|
else {
|
|
LValue value = this->lowJSValue(m_graph.varArgChild(target, i));
|
|
m_out.store64(value, m_out.baseIndex(m_heaps.variables, targetStart, storeIndex));
|
|
storeIndex = m_out.add(m_out.constIntPtr(1), storeIndex);
|
|
}
|
|
}
|
|
return storeIndex;
|
|
}
|
|
|
|
if (target->op() == PhantomNewArrayBuffer) {
|
|
auto* array = target->castOperand<JSImmutableButterfly*>();
|
|
for (unsigned i = 0; i < array->length(); i++) {
|
|
// Because forwarded values are drained as JSValue, we should not generate value
|
|
// in Double form even if PhantomNewArrayBuffer's indexingType is ArrayWithDouble.
|
|
int64_t value = JSValue::encode(array->get(i));
|
|
m_out.store64(m_out.constInt64(value), m_out.baseIndex(m_heaps.variables, targetStart, storeIndex, JSValue(), (Checked<int32_t>(sizeof(Register)) * i).unsafeGet()));
|
|
}
|
|
return m_out.add(m_out.constIntPtr(array->length()), storeIndex);
|
|
}
|
|
|
|
RELEASE_ASSERT(target->op() == PhantomCreateRest);
|
|
InlineCallFrame* inlineCallFrame = target->origin.semantic.inlineCallFrame();
|
|
|
|
auto numberOfArgumentsToSkip = target->numberOfArgumentsToSkip();
|
|
LValue sourceStart = this->getArgumentsStart(inlineCallFrame, numberOfArgumentsToSkip);
|
|
LValue spreadLength = m_out.zeroExtPtr(getSpreadLengthFromInlineCallFrame(inlineCallFrame, numberOfArgumentsToSkip));
|
|
|
|
LBasicBlock loop = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
ValueFromBlock startLoadIndex = m_out.anchor(m_out.constIntPtr(0));
|
|
ValueFromBlock startStoreIndex = m_out.anchor(storeIndex);
|
|
ValueFromBlock startStoreIndexForEnd = m_out.anchor(storeIndex);
|
|
|
|
m_out.branch(m_out.isZero64(spreadLength), unsure(continuation), unsure(loop));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(loop, continuation);
|
|
LValue loopStoreIndex = m_out.phi(Int64, startStoreIndex);
|
|
LValue loadIndex = m_out.phi(Int64, startLoadIndex);
|
|
LValue value = m_out.load64(
|
|
m_out.baseIndex(m_heaps.variables, sourceStart, loadIndex));
|
|
m_out.store64(value, m_out.baseIndex(m_heaps.variables, targetStart, loopStoreIndex));
|
|
LValue nextLoadIndex = m_out.add(m_out.constIntPtr(1), loadIndex);
|
|
m_out.addIncomingToPhi(loadIndex, m_out.anchor(nextLoadIndex));
|
|
LValue nextStoreIndex = m_out.add(m_out.constIntPtr(1), loopStoreIndex);
|
|
m_out.addIncomingToPhi(loopStoreIndex, m_out.anchor(nextStoreIndex));
|
|
ValueFromBlock loopStoreIndexForEnd = m_out.anchor(nextStoreIndex);
|
|
m_out.branch(m_out.below(nextLoadIndex, spreadLength), unsure(loop), unsure(continuation));
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
return m_out.phi(Int64, startStoreIndexForEnd, loopStoreIndexForEnd);
|
|
});
|
|
|
|
LValue storeIndex = forwardSpread(arguments, m_out.constIntPtr(0));
|
|
|
|
LBasicBlock undefinedLoop = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
ValueFromBlock startStoreIndex = m_out.anchor(storeIndex);
|
|
LValue loopBoundValue = m_out.constIntPtr(data->mandatoryMinimum);
|
|
m_out.branch(m_out.below(storeIndex, loopBoundValue),
|
|
unsure(undefinedLoop), unsure(continuation));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(undefinedLoop, continuation);
|
|
LValue loopStoreIndex = m_out.phi(Int64, startStoreIndex);
|
|
m_out.store64(
|
|
m_out.constInt64(JSValue::encode(jsUndefined())),
|
|
m_out.baseIndex(m_heaps.variables, targetStart, loopStoreIndex));
|
|
LValue nextIndex = m_out.add(loopStoreIndex, m_out.constIntPtr(1));
|
|
m_out.addIncomingToPhi(loopStoreIndex, m_out.anchor(nextIndex));
|
|
m_out.branch(
|
|
m_out.below(nextIndex, loopBoundValue), unsure(undefinedLoop), unsure(continuation));
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
}
|
|
|
|
void compileJump()
|
|
{
|
|
m_out.jump(lowBlock(m_node->targetBlock()));
|
|
}
|
|
|
|
void compileBranch()
|
|
{
|
|
m_out.branch(
|
|
boolify(m_node->child1()),
|
|
WeightedTarget(
|
|
lowBlock(m_node->branchData()->taken.block),
|
|
m_node->branchData()->taken.count),
|
|
WeightedTarget(
|
|
lowBlock(m_node->branchData()->notTaken.block),
|
|
m_node->branchData()->notTaken.count));
|
|
}
|
|
|
|
void compileSwitch()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
SwitchData* data = m_node->switchData();
|
|
switch (data->kind) {
|
|
case SwitchImm: {
|
|
Vector<ValueFromBlock, 2> intValues;
|
|
LBasicBlock switchOnInts = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(m_out.m_block, switchOnInts);
|
|
|
|
switch (m_node->child1().useKind()) {
|
|
case Int32Use: {
|
|
intValues.append(m_out.anchor(lowInt32(m_node->child1())));
|
|
m_out.jump(switchOnInts);
|
|
break;
|
|
}
|
|
|
|
case UntypedUse: {
|
|
LBasicBlock isInt = m_out.newBlock();
|
|
LBasicBlock isNotInt = m_out.newBlock();
|
|
LBasicBlock isDouble = m_out.newBlock();
|
|
|
|
LValue boxedValue = lowJSValue(m_node->child1());
|
|
m_out.branch(isNotInt32(boxedValue), unsure(isNotInt), unsure(isInt));
|
|
|
|
LBasicBlock innerLastNext = m_out.appendTo(isInt, isNotInt);
|
|
|
|
intValues.append(m_out.anchor(unboxInt32(boxedValue)));
|
|
m_out.jump(switchOnInts);
|
|
|
|
m_out.appendTo(isNotInt, isDouble);
|
|
m_out.branch(
|
|
isCellOrMiscOrBigInt32(boxedValue, provenType(m_node->child1())),
|
|
usually(lowBlock(data->fallThrough.block)), rarely(isDouble));
|
|
|
|
m_out.appendTo(isDouble, innerLastNext);
|
|
LValue doubleValue = unboxDouble(boxedValue);
|
|
LValue intInDouble = m_out.doubleToInt(doubleValue);
|
|
intValues.append(m_out.anchor(intInDouble));
|
|
m_out.branch(
|
|
m_out.doubleEqual(m_out.intToDouble(intInDouble), doubleValue),
|
|
unsure(switchOnInts), unsure(lowBlock(data->fallThrough.block)));
|
|
break;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
break;
|
|
}
|
|
|
|
m_out.appendTo(switchOnInts, lastNext);
|
|
buildSwitch(data, Int32, m_out.phi(Int32, intValues));
|
|
return;
|
|
}
|
|
|
|
case SwitchChar: {
|
|
LValue stringValue;
|
|
|
|
// FIXME: We should use something other than unsure() for the branch weight
|
|
// of the fallThrough block. The main challenge is just that we have multiple
|
|
// branches to fallThrough but a single count, so we would need to divvy it up
|
|
// among the different lowered branches.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=129082
|
|
|
|
switch (m_node->child1().useKind()) {
|
|
case StringUse: {
|
|
stringValue = lowString(m_node->child1());
|
|
break;
|
|
}
|
|
|
|
case UntypedUse: {
|
|
LValue unboxedValue = lowJSValue(m_node->child1());
|
|
|
|
LBasicBlock isCellCase = m_out.newBlock();
|
|
LBasicBlock isStringCase = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
isNotCell(unboxedValue, provenType(m_node->child1())),
|
|
unsure(lowBlock(data->fallThrough.block)), unsure(isCellCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isCellCase, isStringCase);
|
|
LValue cellValue = unboxedValue;
|
|
m_out.branch(
|
|
isNotString(cellValue, provenType(m_node->child1())),
|
|
unsure(lowBlock(data->fallThrough.block)), unsure(isStringCase));
|
|
|
|
m_out.appendTo(isStringCase, lastNext);
|
|
stringValue = cellValue;
|
|
break;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
break;
|
|
}
|
|
|
|
LBasicBlock lengthIs1 = m_out.newBlock();
|
|
LBasicBlock needResolution = m_out.newBlock();
|
|
LBasicBlock resolved = m_out.newBlock();
|
|
LBasicBlock is8Bit = m_out.newBlock();
|
|
LBasicBlock is16Bit = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
ValueFromBlock fastValue = m_out.anchor(m_out.loadPtr(stringValue, m_heaps.JSString_value));
|
|
m_out.branch(
|
|
isRopeString(stringValue, m_node->child1()),
|
|
rarely(needResolution), usually(resolved));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(needResolution, resolved);
|
|
ValueFromBlock slowValue = m_out.anchor(
|
|
vmCall(pointerType(), operationResolveRope, weakPointer(globalObject), stringValue));
|
|
m_out.jump(resolved);
|
|
|
|
m_out.appendTo(resolved, lengthIs1);
|
|
LValue value = m_out.phi(pointerType(), fastValue, slowValue);
|
|
m_out.branch(
|
|
m_out.notEqual(
|
|
m_out.load32NonNegative(value, m_heaps.StringImpl_length),
|
|
m_out.int32One),
|
|
unsure(lowBlock(data->fallThrough.block)), unsure(lengthIs1));
|
|
|
|
m_out.appendTo(lengthIs1, is8Bit);
|
|
LValue characterData = m_out.loadPtr(value, m_heaps.StringImpl_data);
|
|
m_out.branch(
|
|
m_out.testNonZero32(
|
|
m_out.load32(value, m_heaps.StringImpl_hashAndFlags),
|
|
m_out.constInt32(StringImpl::flagIs8Bit())),
|
|
unsure(is8Bit), unsure(is16Bit));
|
|
|
|
Vector<ValueFromBlock, 2> characters;
|
|
m_out.appendTo(is8Bit, is16Bit);
|
|
characters.append(m_out.anchor(m_out.load8ZeroExt32(characterData, m_heaps.characters8[0])));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(is16Bit, continuation);
|
|
characters.append(m_out.anchor(m_out.load16ZeroExt32(characterData, m_heaps.characters16[0])));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
buildSwitch(data, Int32, m_out.phi(Int32, characters));
|
|
return;
|
|
}
|
|
|
|
case SwitchString: {
|
|
switch (m_node->child1().useKind()) {
|
|
case StringIdentUse: {
|
|
LValue stringImpl = lowStringIdent(m_node->child1());
|
|
|
|
Vector<SwitchCase> cases;
|
|
for (unsigned i = 0; i < data->cases.size(); ++i) {
|
|
LValue value = m_out.constIntPtr(data->cases[i].value.stringImpl());
|
|
LBasicBlock block = lowBlock(data->cases[i].target.block);
|
|
Weight weight = Weight(data->cases[i].target.count);
|
|
cases.append(SwitchCase(value, block, weight));
|
|
}
|
|
|
|
m_out.switchInstruction(
|
|
stringImpl, cases, lowBlock(data->fallThrough.block),
|
|
Weight(data->fallThrough.count));
|
|
return;
|
|
}
|
|
|
|
case StringUse: {
|
|
switchString(data, lowString(m_node->child1()), m_node->child1());
|
|
return;
|
|
}
|
|
|
|
case UntypedUse: {
|
|
LValue value = lowJSValue(m_node->child1());
|
|
|
|
LBasicBlock isCellBlock = m_out.newBlock();
|
|
LBasicBlock isStringBlock = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
isCell(value, provenType(m_node->child1())),
|
|
unsure(isCellBlock), unsure(lowBlock(data->fallThrough.block)));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isCellBlock, isStringBlock);
|
|
|
|
m_out.branch(
|
|
isString(value, provenType(m_node->child1())),
|
|
unsure(isStringBlock), unsure(lowBlock(data->fallThrough.block)));
|
|
|
|
m_out.appendTo(isStringBlock, lastNext);
|
|
|
|
switchString(data, value, m_node->child1());
|
|
return;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
return;
|
|
}
|
|
return;
|
|
}
|
|
|
|
case SwitchCell: {
|
|
LValue cell;
|
|
switch (m_node->child1().useKind()) {
|
|
case CellUse: {
|
|
cell = lowCell(m_node->child1());
|
|
break;
|
|
}
|
|
|
|
case UntypedUse: {
|
|
LValue value = lowJSValue(m_node->child1());
|
|
LBasicBlock cellCase = m_out.newBlock();
|
|
m_out.branch(
|
|
isCell(value, provenType(m_node->child1())),
|
|
unsure(cellCase), unsure(lowBlock(data->fallThrough.block)));
|
|
m_out.appendTo(cellCase);
|
|
cell = value;
|
|
break;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
return;
|
|
}
|
|
|
|
buildSwitch(m_node->switchData(), pointerType(), cell);
|
|
return;
|
|
} }
|
|
|
|
DFG_CRASH(m_graph, m_node, "Bad switch kind");
|
|
}
|
|
|
|
void compileEntrySwitch()
|
|
{
|
|
Vector<LBasicBlock> successors;
|
|
for (DFG::BasicBlock* successor : m_node->entrySwitchData()->cases)
|
|
successors.append(lowBlock(successor));
|
|
m_out.entrySwitch(successors);
|
|
}
|
|
|
|
void compileReturn()
|
|
{
|
|
m_out.ret(lowJSValue(m_node->child1()));
|
|
}
|
|
|
|
void compileForceOSRExit()
|
|
{
|
|
terminate(InadequateCoverage);
|
|
}
|
|
|
|
void compileCPUIntrinsic()
|
|
{
|
|
#if CPU(X86_64)
|
|
Intrinsic intrinsic = m_node->intrinsic();
|
|
switch (intrinsic) {
|
|
case CPUMfenceIntrinsic:
|
|
case CPUCpuidIntrinsic:
|
|
case CPUPauseIntrinsic: {
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Void);
|
|
patchpoint->effects = Effects::forCall();
|
|
if (intrinsic == CPUCpuidIntrinsic)
|
|
patchpoint->clobber(RegisterSet { X86Registers::eax, X86Registers::ebx, X86Registers::ecx, X86Registers::edx });
|
|
|
|
patchpoint->setGenerator([=] (CCallHelpers& jit, const B3::StackmapGenerationParams&) {
|
|
switch (intrinsic) {
|
|
case CPUMfenceIntrinsic:
|
|
jit.mfence();
|
|
break;
|
|
case CPUCpuidIntrinsic:
|
|
jit.cpuid();
|
|
break;
|
|
case CPUPauseIntrinsic:
|
|
jit.pause();
|
|
break;
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
}
|
|
});
|
|
setJSValue(m_out.constInt64(JSValue::encode(jsUndefined())));
|
|
break;
|
|
}
|
|
case CPURdtscIntrinsic: {
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int32);
|
|
patchpoint->effects = Effects::forCall();
|
|
patchpoint->clobber(RegisterSet { X86Registers::eax, X86Registers::edx });
|
|
// The low 32-bits of rdtsc go into rax.
|
|
patchpoint->resultConstraints = { ValueRep::reg(X86Registers::eax) };
|
|
patchpoint->setGenerator( [=] (CCallHelpers& jit, const B3::StackmapGenerationParams&) {
|
|
jit.rdtsc();
|
|
});
|
|
setJSValue(boxInt32(patchpoint));
|
|
break;
|
|
}
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
|
|
}
|
|
#endif
|
|
}
|
|
|
|
void compileThrow()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue error = lowJSValue(m_node->child1());
|
|
vmCall(Void, operationThrowDFG, weakPointer(globalObject), error);
|
|
// vmCall() does an exception check so we should never reach this.
|
|
m_out.unreachable();
|
|
}
|
|
|
|
void compileThrowStaticError()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue errorMessage = lowString(m_node->child1());
|
|
LValue errorType = m_out.constInt32(m_node->errorType());
|
|
vmCall(Void, operationThrowStaticError, weakPointer(globalObject), errorMessage, errorType);
|
|
// vmCall() does an exception check so we should never reach this.
|
|
m_out.unreachable();
|
|
}
|
|
|
|
void compileInvalidationPoint()
|
|
{
|
|
if (verboseCompilationEnabled())
|
|
dataLog(" Invalidation point with availability: ", availabilityMap(), "\n");
|
|
|
|
DFG_ASSERT(m_graph, m_node, m_origin.exitOK);
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Void);
|
|
OSRExitDescriptor* descriptor = appendOSRExitDescriptor(noValue(), nullptr);
|
|
NodeOrigin origin = m_origin;
|
|
patchpoint->appendColdAnys(buildExitArguments(descriptor, origin.forExit, noValue()));
|
|
|
|
State* state = &m_ftlState;
|
|
|
|
auto nodeIndex = m_nodeIndexInGraph;
|
|
patchpoint->setGenerator(
|
|
[=] (CCallHelpers& jit, const B3::StackmapGenerationParams& params) {
|
|
// The MacroAssembler knows more about this than B3 does. The watchpointLabel() method
|
|
// will ensure that this is followed by a nop shadow but only when this is actually
|
|
// necessary.
|
|
CCallHelpers::Label label = jit.watchpointLabel();
|
|
|
|
RefPtr<OSRExitHandle> handle = descriptor->emitOSRExitLater(
|
|
*state, UncountableInvalidation, origin, params, nodeIndex, 0);
|
|
|
|
RefPtr<JITCode> jitCode = state->jitCode.get();
|
|
|
|
jit.addLinkTask(
|
|
[=] (LinkBuffer& linkBuffer) {
|
|
JumpReplacement jumpReplacement(
|
|
linkBuffer.locationOf<JSInternalPtrTag>(label),
|
|
linkBuffer.locationOf<OSRExitPtrTag>(handle->label));
|
|
jitCode->common.jumpReplacements.append(jumpReplacement);
|
|
});
|
|
});
|
|
|
|
// Set some obvious things.
|
|
patchpoint->effects.terminal = false;
|
|
patchpoint->effects.writesLocalState = false;
|
|
patchpoint->effects.readsLocalState = false;
|
|
|
|
// This is how we tell B3 about the possibility of jump replacement.
|
|
patchpoint->effects.exitsSideways = true;
|
|
|
|
// It's not possible for some prior branch to determine the safety of this operation. It's always
|
|
// fine to execute this on some path that wouldn't have originally executed it before
|
|
// optimization.
|
|
patchpoint->effects.controlDependent = false;
|
|
|
|
// If this falls through then it won't write anything.
|
|
patchpoint->effects.writes = HeapRange();
|
|
|
|
// When this abruptly terminates, it could read any heap location.
|
|
patchpoint->effects.reads = HeapRange::top();
|
|
}
|
|
|
|
void compileIsEmpty()
|
|
{
|
|
setBoolean(m_out.isZero64(lowJSValue(m_node->child1())));
|
|
}
|
|
|
|
void compileTypeOfIsUndefined()
|
|
{
|
|
setBoolean(equalNullOrUndefined(m_node->child1(), AllCellsAreFalse, EqualUndefined));
|
|
}
|
|
|
|
void compileIsUndefinedOrNull()
|
|
{
|
|
setBoolean(isOther(lowJSValue(m_node->child1()), provenType(m_node->child1())));
|
|
}
|
|
|
|
void compileIsBoolean()
|
|
{
|
|
setBoolean(isBoolean(lowJSValue(m_node->child1()), provenType(m_node->child1())));
|
|
}
|
|
|
|
void compileIsNumber()
|
|
{
|
|
setBoolean(isNumber(lowJSValue(m_node->child1()), provenType(m_node->child1())));
|
|
}
|
|
|
|
void compileNumberIsInteger()
|
|
{
|
|
LBasicBlock notInt32 = m_out.newBlock();
|
|
LBasicBlock doubleCase = m_out.newBlock();
|
|
LBasicBlock doubleNotNanOrInf = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue input = lowJSValue(m_node->child1());
|
|
|
|
ValueFromBlock trueResult = m_out.anchor(m_out.booleanTrue);
|
|
m_out.branch(
|
|
isInt32(input, provenType(m_node->child1())), unsure(continuation), unsure(notInt32));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(notInt32, doubleCase);
|
|
ValueFromBlock falseResult = m_out.anchor(m_out.booleanFalse);
|
|
m_out.branch(
|
|
isNotNumber(input, provenType(m_node->child1())), unsure(continuation), unsure(doubleCase));
|
|
|
|
m_out.appendTo(doubleCase, doubleNotNanOrInf);
|
|
LValue doubleAsInt;
|
|
LValue asDouble = unboxDouble(input, &doubleAsInt);
|
|
LValue expBits = m_out.bitAnd(m_out.lShr(doubleAsInt, m_out.constInt32(52)), m_out.constInt64(0x7ff));
|
|
m_out.branch(
|
|
m_out.equal(expBits, m_out.constInt64(0x7ff)),
|
|
unsure(continuation), unsure(doubleNotNanOrInf));
|
|
|
|
m_out.appendTo(doubleNotNanOrInf, continuation);
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int32);
|
|
patchpoint->appendSomeRegister(asDouble);
|
|
patchpoint->numFPScratchRegisters = 1;
|
|
patchpoint->effects = Effects::none();
|
|
patchpoint->setGenerator([=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
GPRReg result = params[0].gpr();
|
|
FPRReg input = params[1].fpr();
|
|
FPRReg temp = params.fpScratch(0);
|
|
jit.roundTowardZeroDouble(input, temp);
|
|
jit.compareDouble(MacroAssembler::DoubleEqualAndOrdered, input, temp, result);
|
|
});
|
|
ValueFromBlock patchpointResult = m_out.anchor(patchpoint);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setBoolean(m_out.phi(Int32, trueResult, falseResult, patchpointResult));
|
|
}
|
|
|
|
#if USE(BIGINT32)
|
|
void compileIsBigInt()
|
|
{
|
|
LValue value = lowJSValue(m_node->child1());
|
|
|
|
LBasicBlock isCellCase = m_out.newBlock();
|
|
LBasicBlock isNotCellCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(isCell(value, provenType(m_node->child1())), unsure(isCellCase), unsure(isNotCellCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isNotCellCase, isCellCase);
|
|
// FIXME: we should filter the provenType to include the fact that we know we are not dealing with a cell
|
|
ValueFromBlock notCellResult = m_out.anchor(isBigInt32(value, provenType(m_node->child1())));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(isCellCase, continuation);
|
|
ValueFromBlock cellResult = m_out.anchor(isCellWithType(value, m_node->queriedType(), m_node->speculatedTypeForQuery(), provenType(m_node->child1())));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setBoolean(m_out.phi(Int32, notCellResult, cellResult));
|
|
}
|
|
#else // if !USE(BIGINT32)
|
|
NO_RETURN_DUE_TO_CRASH ALWAYS_INLINE void compileIsBigInt()
|
|
{
|
|
// If we are not dealing with BigInt32, we should just emit IsCellWithType(HeapBigInt) instead.
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
}
|
|
#endif
|
|
void compileIsCellWithType()
|
|
{
|
|
if (m_node->child1().useKind() == UntypedUse) {
|
|
LValue value = lowJSValue(m_node->child1());
|
|
|
|
LBasicBlock isCellCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
ValueFromBlock notCellResult = m_out.anchor(m_out.booleanFalse);
|
|
m_out.branch(
|
|
isCell(value, provenType(m_node->child1())), unsure(isCellCase), unsure(continuation));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isCellCase, continuation);
|
|
ValueFromBlock cellResult = m_out.anchor(isCellWithType(value, m_node->queriedType(), m_node->speculatedTypeForQuery(), provenType(m_node->child1())));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setBoolean(m_out.phi(Int32, notCellResult, cellResult));
|
|
} else {
|
|
ASSERT(m_node->child1().useKind() == CellUse);
|
|
setBoolean(isCellWithType(lowCell(m_node->child1()), m_node->queriedType(), m_node->speculatedTypeForQuery(), provenType(m_node->child1())));
|
|
}
|
|
}
|
|
|
|
void compileIsObject()
|
|
{
|
|
LValue value = lowJSValue(m_node->child1());
|
|
|
|
LBasicBlock isCellCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
ValueFromBlock notCellResult = m_out.anchor(m_out.booleanFalse);
|
|
m_out.branch(
|
|
isCell(value, provenType(m_node->child1())), unsure(isCellCase), unsure(continuation));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isCellCase, continuation);
|
|
ValueFromBlock cellResult = m_out.anchor(isObject(value, provenType(m_node->child1())));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setBoolean(m_out.phi(Int32, notCellResult, cellResult));
|
|
}
|
|
|
|
LValue wangsInt64Hash(LValue input)
|
|
{
|
|
// key += ~(key << 32);
|
|
LValue key = input;
|
|
LValue temp = key;
|
|
temp = m_out.shl(temp, m_out.constInt32(32));
|
|
temp = m_out.bitNot(temp);
|
|
key = m_out.add(key, temp);
|
|
// key ^= (key >> 22);
|
|
temp = key;
|
|
temp = m_out.lShr(temp, m_out.constInt32(22));
|
|
key = m_out.bitXor(key, temp);
|
|
// key += ~(key << 13);
|
|
temp = key;
|
|
temp = m_out.shl(temp, m_out.constInt32(13));
|
|
temp = m_out.bitNot(temp);
|
|
key = m_out.add(key, temp);
|
|
// key ^= (key >> 8);
|
|
temp = key;
|
|
temp = m_out.lShr(temp, m_out.constInt32(8));
|
|
key = m_out.bitXor(key, temp);
|
|
// key += (key << 3);
|
|
temp = key;
|
|
temp = m_out.shl(temp, m_out.constInt32(3));
|
|
key = m_out.add(key, temp);
|
|
// key ^= (key >> 15);
|
|
temp = key;
|
|
temp = m_out.lShr(temp, m_out.constInt32(15));
|
|
key = m_out.bitXor(key, temp);
|
|
// key += ~(key << 27);
|
|
temp = key;
|
|
temp = m_out.shl(temp, m_out.constInt32(27));
|
|
temp = m_out.bitNot(temp);
|
|
key = m_out.add(key, temp);
|
|
// key ^= (key >> 31);
|
|
temp = key;
|
|
temp = m_out.lShr(temp, m_out.constInt32(31));
|
|
key = m_out.bitXor(key, temp);
|
|
key = m_out.castToInt32(key);
|
|
|
|
return key;
|
|
}
|
|
|
|
LValue mapHashString(LValue string, Edge& edge)
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LBasicBlock nonEmptyStringCase = m_out.newBlock();
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(isRopeString(string, edge), rarely(slowCase), usually(nonEmptyStringCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(nonEmptyStringCase, slowCase);
|
|
LValue stringImpl = m_out.loadPtr(string, m_heaps.JSString_value);
|
|
LValue hash = m_out.lShr(m_out.load32(stringImpl, m_heaps.StringImpl_hashAndFlags), m_out.constInt32(StringImpl::s_flagCount));
|
|
ValueFromBlock nonEmptyStringHashResult = m_out.anchor(hash);
|
|
m_out.branch(m_out.equal(hash, m_out.constInt32(0)),
|
|
unsure(slowCase), unsure(continuation));
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
ValueFromBlock slowResult = m_out.anchor(m_out.castToInt32(vmCall(Int64, operationMapHash, weakPointer(globalObject), string)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
return m_out.phi(Int32, slowResult, nonEmptyStringHashResult);
|
|
}
|
|
|
|
void compileMapHash()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
switch (m_node->child1().useKind()) {
|
|
#if USE(BIGINT32)
|
|
case BigInt32Use:
|
|
#endif
|
|
case BooleanUse:
|
|
case Int32Use:
|
|
case SymbolUse:
|
|
case ObjectUse: {
|
|
LValue key = lowJSValue(m_node->child1(), ManualOperandSpeculation);
|
|
speculate(m_node->child1());
|
|
setInt32(wangsInt64Hash(key));
|
|
return;
|
|
}
|
|
|
|
case HeapBigIntUse: {
|
|
LValue key = lowHeapBigInt(m_node->child1());
|
|
setInt32(m_out.castToInt32(vmCall(Int64, operationMapHashHeapBigInt, m_vmValue, key)));
|
|
return;
|
|
}
|
|
|
|
case CellUse: {
|
|
LBasicBlock isString = m_out.newBlock();
|
|
LBasicBlock notString = m_out.newBlock();
|
|
LBasicBlock isHeapBigIntCase = m_out.newBlock();
|
|
LBasicBlock notStringNorHeapBigIntCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue value = lowCell(m_node->child1());
|
|
LValue isStringValue = m_out.equal(m_out.load8ZeroExt32(value, m_heaps.JSCell_typeInfoType), m_out.constInt32(StringType));
|
|
m_out.branch(
|
|
isStringValue, unsure(isString), unsure(notString));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isString, notString);
|
|
ValueFromBlock stringResult = m_out.anchor(mapHashString(value, m_node->child1()));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(notString, isHeapBigIntCase);
|
|
m_out.branch(isHeapBigInt(value, (provenType(m_node->child1()) & ~SpecString)), unsure(isHeapBigIntCase), unsure(notStringNorHeapBigIntCase));
|
|
|
|
m_out.appendTo(isHeapBigIntCase, notStringNorHeapBigIntCase);
|
|
ValueFromBlock heapBigIntResult = m_out.anchor(m_out.castToInt32(vmCall(Int64, operationMapHashHeapBigInt, m_vmValue, value)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(notStringNorHeapBigIntCase, continuation);
|
|
ValueFromBlock notStringResult = m_out.anchor(wangsInt64Hash(value));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setInt32(m_out.phi(Int32, stringResult, heapBigIntResult, notStringResult));
|
|
return;
|
|
}
|
|
|
|
case StringUse: {
|
|
LValue string = lowString(m_node->child1());
|
|
setInt32(mapHashString(string, m_node->child1()));
|
|
return;
|
|
}
|
|
|
|
default:
|
|
RELEASE_ASSERT(m_node->child1().useKind() == UntypedUse);
|
|
break;
|
|
}
|
|
|
|
LValue value = lowJSValue(m_node->child1());
|
|
|
|
LBasicBlock isCellCase = m_out.newBlock();
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock straightHash = m_out.newBlock();
|
|
LBasicBlock isStringCase = m_out.newBlock();
|
|
LBasicBlock notStringCase = m_out.newBlock();
|
|
LBasicBlock nonEmptyStringCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
isCell(value, provenType(m_node->child1())), unsure(isCellCase), unsure(straightHash));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isCellCase, notStringCase);
|
|
LValue isString = m_out.equal(m_out.load8ZeroExt32(value, m_heaps.JSCell_typeInfoType), m_out.constInt32(StringType));
|
|
m_out.branch(
|
|
isString, unsure(isStringCase), unsure(notStringCase));
|
|
|
|
m_out.appendTo(notStringCase, isStringCase);
|
|
m_out.branch(isHeapBigInt(value, (provenType(m_node->child1()) & ~SpecString)), unsure(slowCase), unsure(straightHash));
|
|
|
|
m_out.appendTo(isStringCase, nonEmptyStringCase);
|
|
m_out.branch(isRopeString(value, m_node->child1()), rarely(slowCase), usually(nonEmptyStringCase));
|
|
|
|
m_out.appendTo(nonEmptyStringCase, straightHash);
|
|
LValue stringImpl = m_out.loadPtr(value, m_heaps.JSString_value);
|
|
LValue hash = m_out.lShr(m_out.load32(stringImpl, m_heaps.StringImpl_hashAndFlags), m_out.constInt32(StringImpl::s_flagCount));
|
|
ValueFromBlock nonEmptyStringHashResult = m_out.anchor(hash);
|
|
m_out.branch(m_out.equal(hash, m_out.constInt32(0)),
|
|
unsure(slowCase), unsure(continuation));
|
|
|
|
m_out.appendTo(straightHash, slowCase);
|
|
ValueFromBlock fastResult = m_out.anchor(wangsInt64Hash(value));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
ValueFromBlock slowResult = m_out.anchor(m_out.castToInt32(vmCall(Int64, operationMapHash, weakPointer(globalObject), value)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setInt32(m_out.phi(Int32, fastResult, slowResult, nonEmptyStringHashResult));
|
|
}
|
|
|
|
void compileNormalizeMapKey()
|
|
{
|
|
ASSERT(m_node->child1().useKind() == UntypedUse);
|
|
|
|
LBasicBlock isCellCase = m_out.newBlock();
|
|
LBasicBlock notCellCase = m_out.newBlock();
|
|
LBasicBlock isHeapBigIntCase = m_out.newBlock();
|
|
LBasicBlock isNumberCase = m_out.newBlock();
|
|
LBasicBlock notInt32NumberCase = m_out.newBlock();
|
|
LBasicBlock notNaNCase = m_out.newBlock();
|
|
LBasicBlock convertibleCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(isCellCase);
|
|
|
|
LValue key = lowJSValue(m_node->child1());
|
|
ValueFromBlock fastResult = m_out.anchor(key);
|
|
m_out.branch(isNotCell(key, provenType(m_node->child1())), unsure(notCellCase), unsure(isCellCase));
|
|
|
|
m_out.appendTo(isCellCase, isHeapBigIntCase);
|
|
m_out.branch(isNotHeapBigInt(key, (provenType(m_node->child1()) & SpecCellCheck)), unsure(continuation), unsure(isHeapBigIntCase));
|
|
|
|
m_out.appendTo(isHeapBigIntCase, notCellCase);
|
|
ValueFromBlock bigIntResult = m_out.anchor(vmCall(Int64, operationNormalizeMapKeyHeapBigInt, m_vmValue, key));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(notCellCase, isNumberCase);
|
|
m_out.branch(isNotNumber(key), unsure(continuation), unsure(isNumberCase));
|
|
|
|
m_out.appendTo(isNumberCase, notInt32NumberCase);
|
|
m_out.branch(isInt32(key), unsure(continuation), unsure(notInt32NumberCase));
|
|
|
|
m_out.appendTo(notInt32NumberCase, notNaNCase);
|
|
LValue doubleValue = unboxDouble(key);
|
|
ValueFromBlock normalizedNaNResult = m_out.anchor(m_out.constInt64(JSValue::encode(jsNaN())));
|
|
m_out.branch(m_out.doubleNotEqualOrUnordered(doubleValue, doubleValue), unsure(continuation), unsure(notNaNCase));
|
|
|
|
m_out.appendTo(notNaNCase, convertibleCase);
|
|
LValue integerValue = m_out.doubleToInt(doubleValue);
|
|
LValue integerValueConvertedToDouble = m_out.intToDouble(integerValue);
|
|
ValueFromBlock doubleResult = m_out.anchor(key);
|
|
m_out.branch(m_out.doubleNotEqualOrUnordered(doubleValue, integerValueConvertedToDouble), unsure(continuation), unsure(convertibleCase));
|
|
|
|
m_out.appendTo(convertibleCase, continuation);
|
|
ValueFromBlock boxedIntResult = m_out.anchor(boxInt32(integerValue));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, fastResult, bigIntResult, normalizedNaNResult, doubleResult, boxedIntResult));
|
|
}
|
|
|
|
void compileGetMapBucket()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LBasicBlock loopStart = m_out.newBlock();
|
|
LBasicBlock loopAround = m_out.newBlock();
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock notPresentInTable = m_out.newBlock();
|
|
LBasicBlock notEmptyValue = m_out.newBlock();
|
|
LBasicBlock notDeletedValue = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(loopStart);
|
|
|
|
LValue map;
|
|
if (m_node->child1().useKind() == MapObjectUse)
|
|
map = lowMapObject(m_node->child1());
|
|
else if (m_node->child1().useKind() == SetObjectUse)
|
|
map = lowSetObject(m_node->child1());
|
|
else
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
|
|
LValue key = lowJSValue(m_node->child2(), ManualOperandSpeculation);
|
|
if (m_node->child2().useKind() != UntypedUse)
|
|
speculate(m_node->child2());
|
|
|
|
LValue hash = lowInt32(m_node->child3());
|
|
|
|
LValue buffer = m_out.loadPtr(map, m_heaps.HashMapImpl_buffer);
|
|
LValue mask = m_out.sub(m_out.load32(map, m_heaps.HashMapImpl_capacity), m_out.int32One);
|
|
|
|
ValueFromBlock indexStart = m_out.anchor(hash);
|
|
m_out.jump(loopStart);
|
|
|
|
m_out.appendTo(loopStart, notEmptyValue);
|
|
LValue unmaskedIndex = m_out.phi(Int32, indexStart);
|
|
LValue index = m_out.bitAnd(mask, unmaskedIndex);
|
|
// FIXME: I think these buffers are caged?
|
|
// https://bugs.webkit.org/show_bug.cgi?id=174925
|
|
LValue hashMapBucket = m_out.load64(m_out.baseIndex(m_heaps.properties.atAnyNumber(), buffer, m_out.zeroExt(index, Int64), ScaleEight));
|
|
ValueFromBlock bucketResult = m_out.anchor(hashMapBucket);
|
|
m_out.branch(m_out.equal(hashMapBucket, m_out.constIntPtr(bitwise_cast<intptr_t>(HashMapImpl<HashMapBucket<HashMapBucketDataKey>>::emptyValue()))),
|
|
unsure(notPresentInTable), unsure(notEmptyValue));
|
|
|
|
m_out.appendTo(notEmptyValue, notDeletedValue);
|
|
m_out.branch(m_out.equal(hashMapBucket, m_out.constIntPtr(bitwise_cast<intptr_t>(HashMapImpl<HashMapBucket<HashMapBucketDataKey>>::deletedValue()))),
|
|
unsure(loopAround), unsure(notDeletedValue));
|
|
|
|
m_out.appendTo(notDeletedValue, loopAround);
|
|
LValue bucketKey = m_out.load64(hashMapBucket, m_heaps.HashMapBucket_key);
|
|
|
|
// Perform Object.is()
|
|
switch (m_node->child2().useKind()) {
|
|
case BooleanUse:
|
|
#if USE(BIGINT32)
|
|
case BigInt32Use:
|
|
#endif
|
|
case Int32Use:
|
|
case SymbolUse:
|
|
case ObjectUse: {
|
|
m_out.branch(m_out.equal(key, bucketKey),
|
|
unsure(continuation), unsure(loopAround));
|
|
break;
|
|
}
|
|
case StringUse: {
|
|
LBasicBlock notBitEqual = m_out.newBlock();
|
|
LBasicBlock bucketKeyIsCell = m_out.newBlock();
|
|
|
|
m_out.branch(m_out.equal(key, bucketKey),
|
|
unsure(continuation), unsure(notBitEqual));
|
|
|
|
m_out.appendTo(notBitEqual, bucketKeyIsCell);
|
|
m_out.branch(isCell(bucketKey),
|
|
unsure(bucketKeyIsCell), unsure(loopAround));
|
|
|
|
m_out.appendTo(bucketKeyIsCell, loopAround);
|
|
m_out.branch(isString(bucketKey),
|
|
unsure(slowPath), unsure(loopAround));
|
|
break;
|
|
}
|
|
case HeapBigIntUse: {
|
|
LBasicBlock notBitEqual = m_out.newBlock();
|
|
LBasicBlock bucketKeyIsCell = m_out.newBlock();
|
|
|
|
m_out.branch(m_out.equal(key, bucketKey),
|
|
unsure(continuation), unsure(notBitEqual));
|
|
|
|
m_out.appendTo(notBitEqual, bucketKeyIsCell);
|
|
m_out.branch(isCell(bucketKey),
|
|
unsure(bucketKeyIsCell), unsure(loopAround));
|
|
|
|
m_out.appendTo(bucketKeyIsCell, loopAround);
|
|
m_out.branch(isHeapBigInt(bucketKey),
|
|
unsure(slowPath), unsure(loopAround));
|
|
break;
|
|
}
|
|
case CellUse: {
|
|
LBasicBlock notBitEqual = m_out.newBlock();
|
|
LBasicBlock bucketKeyIsCell = m_out.newBlock();
|
|
LBasicBlock bucketKeyIsString = m_out.newBlock();
|
|
LBasicBlock bucketKeyIsNotString = m_out.newBlock();
|
|
LBasicBlock bucketKeyIsHeapBigInt = m_out.newBlock();
|
|
|
|
m_out.branch(m_out.equal(key, bucketKey),
|
|
unsure(continuation), unsure(notBitEqual));
|
|
|
|
m_out.appendTo(notBitEqual, bucketKeyIsCell);
|
|
m_out.branch(isCell(bucketKey),
|
|
unsure(bucketKeyIsCell), unsure(loopAround));
|
|
|
|
m_out.appendTo(bucketKeyIsCell, bucketKeyIsString);
|
|
m_out.branch(isString(bucketKey),
|
|
unsure(bucketKeyIsString), unsure(bucketKeyIsNotString));
|
|
|
|
m_out.appendTo(bucketKeyIsString, bucketKeyIsNotString);
|
|
m_out.branch(isString(key, provenType(m_node->child2())),
|
|
unsure(slowPath), unsure(loopAround));
|
|
|
|
m_out.appendTo(bucketKeyIsNotString, bucketKeyIsHeapBigInt);
|
|
m_out.branch(isHeapBigInt(bucketKey),
|
|
unsure(bucketKeyIsHeapBigInt), unsure(loopAround));
|
|
|
|
m_out.appendTo(bucketKeyIsHeapBigInt, loopAround);
|
|
m_out.branch(isHeapBigInt(key, provenType(m_node->child2())),
|
|
unsure(slowPath), unsure(loopAround));
|
|
break;
|
|
}
|
|
case UntypedUse: {
|
|
LBasicBlock notBitEqual = m_out.newBlock();
|
|
LBasicBlock bucketKeyIsCell = m_out.newBlock();
|
|
LBasicBlock bothAreCells = m_out.newBlock();
|
|
LBasicBlock bucketKeyIsString = m_out.newBlock();
|
|
LBasicBlock bucketKeyIsNotString = m_out.newBlock();
|
|
LBasicBlock bucketKeyIsHeapBigInt = m_out.newBlock();
|
|
|
|
m_out.branch(m_out.equal(key, bucketKey),
|
|
unsure(continuation), unsure(notBitEqual));
|
|
|
|
m_out.appendTo(notBitEqual, bucketKeyIsCell);
|
|
m_out.branch(isCell(bucketKey),
|
|
unsure(bucketKeyIsCell), unsure(loopAround));
|
|
|
|
m_out.appendTo(bucketKeyIsCell, bothAreCells);
|
|
m_out.branch(isCell(key),
|
|
unsure(bothAreCells), unsure(loopAround));
|
|
|
|
m_out.appendTo(bothAreCells, bucketKeyIsString);
|
|
m_out.branch(isString(bucketKey),
|
|
unsure(bucketKeyIsString), unsure(bucketKeyIsNotString));
|
|
|
|
m_out.appendTo(bucketKeyIsString, bucketKeyIsNotString);
|
|
m_out.branch(isString(key, provenType(m_node->child2())),
|
|
unsure(slowPath), unsure(loopAround));
|
|
|
|
m_out.appendTo(bucketKeyIsNotString, bucketKeyIsHeapBigInt);
|
|
m_out.branch(isHeapBigInt(bucketKey),
|
|
unsure(bucketKeyIsHeapBigInt), unsure(loopAround));
|
|
|
|
m_out.appendTo(bucketKeyIsHeapBigInt, loopAround);
|
|
m_out.branch(isHeapBigInt(key, provenType(m_node->child2())),
|
|
unsure(slowPath), unsure(loopAround));
|
|
break;
|
|
}
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
}
|
|
|
|
m_out.appendTo(loopAround, slowPath);
|
|
m_out.addIncomingToPhi(unmaskedIndex, m_out.anchor(m_out.add(index, m_out.int32One)));
|
|
m_out.jump(loopStart);
|
|
|
|
m_out.appendTo(slowPath, notPresentInTable);
|
|
ValueFromBlock slowPathResult = m_out.anchor(vmCall(pointerType(),
|
|
m_node->child1().useKind() == MapObjectUse ? operationJSMapFindBucket : operationJSSetFindBucket, weakPointer(globalObject), map, key, hash));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(notPresentInTable, continuation);
|
|
ValueFromBlock notPresentResult;
|
|
if (m_node->child1().useKind() == MapObjectUse)
|
|
notPresentResult = m_out.anchor(weakPointer(vm().sentinelMapBucket()));
|
|
else if (m_node->child1().useKind() == SetObjectUse)
|
|
notPresentResult = m_out.anchor(weakPointer(vm().sentinelSetBucket()));
|
|
else
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(pointerType(), bucketResult, slowPathResult, notPresentResult));
|
|
}
|
|
|
|
void compileGetMapBucketHead()
|
|
{
|
|
LValue map;
|
|
if (m_node->child1().useKind() == MapObjectUse)
|
|
map = lowMapObject(m_node->child1());
|
|
else if (m_node->child1().useKind() == SetObjectUse)
|
|
map = lowSetObject(m_node->child1());
|
|
else
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
|
|
ASSERT(HashMapImpl<HashMapBucket<HashMapBucketDataKey>>::offsetOfHead() == HashMapImpl<HashMapBucket<HashMapBucketDataKeyValue>>::offsetOfHead());
|
|
setJSValue(m_out.loadPtr(map, m_heaps.HashMapImpl_head));
|
|
}
|
|
|
|
void compileGetMapBucketNext()
|
|
{
|
|
LBasicBlock loopStart = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
LBasicBlock noBucket = m_out.newBlock();
|
|
LBasicBlock hasBucket = m_out.newBlock();
|
|
LBasicBlock nextBucket = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(loopStart);
|
|
|
|
ASSERT(HashMapBucket<HashMapBucketDataKey>::offsetOfNext() == HashMapBucket<HashMapBucketDataKeyValue>::offsetOfNext());
|
|
ASSERT(HashMapBucket<HashMapBucketDataKey>::offsetOfKey() == HashMapBucket<HashMapBucketDataKeyValue>::offsetOfKey());
|
|
LValue mapBucketPrev = lowCell(m_node->child1());
|
|
ValueFromBlock mapBucketStart = m_out.anchor(m_out.loadPtr(mapBucketPrev, m_heaps.HashMapBucket_next));
|
|
m_out.jump(loopStart);
|
|
|
|
m_out.appendTo(loopStart, noBucket);
|
|
LValue mapBucket = m_out.phi(pointerType(), mapBucketStart);
|
|
m_out.branch(m_out.isNull(mapBucket), unsure(noBucket), unsure(hasBucket));
|
|
|
|
m_out.appendTo(noBucket, hasBucket);
|
|
ValueFromBlock noBucketResult;
|
|
if (m_node->bucketOwnerType() == BucketOwnerType::Map)
|
|
noBucketResult = m_out.anchor(weakPointer(vm().sentinelMapBucket()));
|
|
else {
|
|
ASSERT(m_node->bucketOwnerType() == BucketOwnerType::Set);
|
|
noBucketResult = m_out.anchor(weakPointer(vm().sentinelSetBucket()));
|
|
}
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(hasBucket, nextBucket);
|
|
ValueFromBlock bucketResult = m_out.anchor(mapBucket);
|
|
m_out.branch(m_out.isZero64(m_out.load64(mapBucket, m_heaps.HashMapBucket_key)), unsure(nextBucket), unsure(continuation));
|
|
|
|
m_out.appendTo(nextBucket, continuation);
|
|
m_out.addIncomingToPhi(mapBucket, m_out.anchor(m_out.loadPtr(mapBucket, m_heaps.HashMapBucket_next)));
|
|
m_out.jump(loopStart);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(pointerType(), noBucketResult, bucketResult));
|
|
}
|
|
|
|
void compileLoadValueFromMapBucket()
|
|
{
|
|
LValue mapBucket = lowCell(m_node->child1());
|
|
setJSValue(m_out.load64(mapBucket, m_heaps.HashMapBucket_value));
|
|
}
|
|
|
|
void compileExtractValueFromWeakMapGet()
|
|
{
|
|
LValue value = lowJSValue(m_node->child1());
|
|
setJSValue(m_out.select(m_out.isZero64(value),
|
|
m_out.constInt64(JSValue::encode(jsUndefined())),
|
|
value));
|
|
}
|
|
|
|
void compileLoadKeyFromMapBucket()
|
|
{
|
|
LValue mapBucket = lowCell(m_node->child1());
|
|
setJSValue(m_out.load64(mapBucket, m_heaps.HashMapBucket_key));
|
|
}
|
|
|
|
void compileSetAdd()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue set = lowSetObject(m_node->child1());
|
|
LValue key = lowJSValue(m_node->child2());
|
|
LValue hash = lowInt32(m_node->child3());
|
|
|
|
setJSValue(vmCall(pointerType(), operationSetAdd, weakPointer(globalObject), set, key, hash));
|
|
}
|
|
|
|
void compileMapSet()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue map = lowMapObject(m_graph.varArgChild(m_node, 0));
|
|
LValue key = lowJSValue(m_graph.varArgChild(m_node, 1));
|
|
LValue value = lowJSValue(m_graph.varArgChild(m_node, 2));
|
|
LValue hash = lowInt32(m_graph.varArgChild(m_node, 3));
|
|
|
|
setJSValue(vmCall(pointerType(), operationMapSet, weakPointer(globalObject), map, key, value, hash));
|
|
}
|
|
|
|
void compileWeakMapGet()
|
|
{
|
|
LBasicBlock loopStart = m_out.newBlock();
|
|
LBasicBlock loopAround = m_out.newBlock();
|
|
LBasicBlock notEqualValue = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(loopStart);
|
|
|
|
LValue weakMap;
|
|
if (m_node->child1().useKind() == WeakMapObjectUse)
|
|
weakMap = lowWeakMapObject(m_node->child1());
|
|
else if (m_node->child1().useKind() == WeakSetObjectUse)
|
|
weakMap = lowWeakSetObject(m_node->child1());
|
|
else
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
LValue key = lowObject(m_node->child2());
|
|
LValue hash = lowInt32(m_node->child3());
|
|
|
|
LValue buffer = m_out.loadPtr(weakMap, m_heaps.WeakMapImpl_buffer);
|
|
LValue mask = m_out.sub(m_out.load32(weakMap, m_heaps.WeakMapImpl_capacity), m_out.int32One);
|
|
|
|
ValueFromBlock indexStart = m_out.anchor(hash);
|
|
m_out.jump(loopStart);
|
|
|
|
m_out.appendTo(loopStart, notEqualValue);
|
|
LValue unmaskedIndex = m_out.phi(Int32, indexStart);
|
|
LValue index = m_out.bitAnd(mask, unmaskedIndex);
|
|
|
|
LValue bucket;
|
|
|
|
if (m_node->child1().useKind() == WeakMapObjectUse) {
|
|
static_assert(hasOneBitSet(sizeof(WeakMapBucket<WeakMapBucketDataKeyValue>)), "Should be a power of 2");
|
|
bucket = m_out.add(buffer, m_out.shl(m_out.zeroExt(index, Int64), m_out.constInt32(getLSBSet(sizeof(WeakMapBucket<WeakMapBucketDataKeyValue>)))));
|
|
} else {
|
|
static_assert(hasOneBitSet(sizeof(WeakMapBucket<WeakMapBucketDataKey>)), "Should be a power of 2");
|
|
bucket = m_out.add(buffer, m_out.shl(m_out.zeroExt(index, Int64), m_out.constInt32(getLSBSet(sizeof(WeakMapBucket<WeakMapBucketDataKey>)))));
|
|
}
|
|
|
|
LValue bucketKey = m_out.load64(bucket, m_heaps.WeakMapBucket_key);
|
|
m_out.branch(m_out.equal(key, bucketKey), unsure(continuation), unsure(notEqualValue));
|
|
|
|
m_out.appendTo(notEqualValue, loopAround);
|
|
m_out.branch(m_out.isNull(bucketKey), unsure(continuation), unsure(loopAround));
|
|
|
|
m_out.appendTo(loopAround, continuation);
|
|
m_out.addIncomingToPhi(unmaskedIndex, m_out.anchor(m_out.add(index, m_out.int32One)));
|
|
m_out.jump(loopStart);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
LValue result;
|
|
if (m_node->child1().useKind() == WeakMapObjectUse)
|
|
result = m_out.load64(bucket, m_heaps.WeakMapBucket_value);
|
|
else
|
|
result = bucketKey;
|
|
setJSValue(result);
|
|
}
|
|
|
|
void compileWeakSetAdd()
|
|
{
|
|
LValue set = lowWeakSetObject(m_node->child1());
|
|
LValue key = lowObject(m_node->child2());
|
|
LValue hash = lowInt32(m_node->child3());
|
|
|
|
vmCall(Void, operationWeakSetAdd, m_vmValue, set, key, hash);
|
|
}
|
|
|
|
void compileWeakMapSet()
|
|
{
|
|
LValue map = lowWeakMapObject(m_graph.varArgChild(m_node, 0));
|
|
LValue key = lowObject(m_graph.varArgChild(m_node, 1));
|
|
LValue value = lowJSValue(m_graph.varArgChild(m_node, 2));
|
|
LValue hash = lowInt32(m_graph.varArgChild(m_node, 3));
|
|
|
|
vmCall(Void, operationWeakMapSet, m_vmValue, map, key, value, hash);
|
|
}
|
|
|
|
void compileTypeOfIsObject()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
|
|
Edge child = m_node->child1();
|
|
LValue value = lowJSValue(child);
|
|
|
|
LBasicBlock cellCase = m_out.newBlock();
|
|
LBasicBlock notFunctionCase = m_out.newBlock();
|
|
LBasicBlock objectCase = m_out.newBlock();
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock notCellCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(isCell(value, provenType(child)), unsure(cellCase), unsure(notCellCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(cellCase, notFunctionCase);
|
|
ValueFromBlock isFunctionResult = m_out.anchor(m_out.booleanFalse);
|
|
m_out.branch(
|
|
isFunction(value, provenType(child)),
|
|
unsure(continuation), unsure(notFunctionCase));
|
|
|
|
m_out.appendTo(notFunctionCase, objectCase);
|
|
ValueFromBlock notObjectResult = m_out.anchor(m_out.booleanFalse);
|
|
m_out.branch(
|
|
isObject(value, provenType(child)),
|
|
unsure(objectCase), unsure(continuation));
|
|
|
|
m_out.appendTo(objectCase, slowPath);
|
|
ValueFromBlock objectResult = m_out.anchor(m_out.booleanTrue);
|
|
m_out.branch(
|
|
isExoticForTypeof(value, provenType(child)),
|
|
rarely(slowPath), usually(continuation));
|
|
|
|
m_out.appendTo(slowPath, notCellCase);
|
|
VM& vm = this->vm();
|
|
LValue slowResultValue = lazySlowPath(
|
|
[=, &vm] (const Vector<Location>& locations) -> RefPtr<LazySlowPath::Generator> {
|
|
return createLazyCallGenerator(vm,
|
|
operationTypeOfIsObject, locations[0].directGPR(),
|
|
CCallHelpers::TrustedImmPtr(globalObject), locations[1].directGPR());
|
|
}, value);
|
|
ValueFromBlock slowResult = m_out.anchor(m_out.notZero64(slowResultValue));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(notCellCase, continuation);
|
|
LValue notCellResultValue = m_out.equal(value, m_out.constInt64(JSValue::encode(jsNull())));
|
|
ValueFromBlock notCellResult = m_out.anchor(notCellResultValue);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
LValue result = m_out.phi(
|
|
Int32,
|
|
isFunctionResult, notObjectResult, objectResult, slowResult, notCellResult);
|
|
setBoolean(result);
|
|
}
|
|
|
|
void compileIsCallable(S_JITOperation_GC slowPathOperation)
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
|
|
Edge child = m_node->child1();
|
|
LValue value = lowJSValue(child);
|
|
|
|
LBasicBlock cellCase = m_out.newBlock();
|
|
LBasicBlock notFunctionCase = m_out.newBlock();
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
ValueFromBlock notCellResult = m_out.anchor(m_out.booleanFalse);
|
|
m_out.branch(
|
|
isCell(value, provenType(child)), unsure(cellCase), unsure(continuation));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(cellCase, notFunctionCase);
|
|
ValueFromBlock functionResult = m_out.anchor(m_out.booleanTrue);
|
|
m_out.branch(
|
|
isFunction(value, provenType(child)),
|
|
unsure(continuation), unsure(notFunctionCase));
|
|
|
|
m_out.appendTo(notFunctionCase, slowPath);
|
|
ValueFromBlock objectResult = m_out.anchor(m_out.booleanFalse);
|
|
m_out.branch(
|
|
isExoticForTypeof(value, provenType(child)),
|
|
rarely(slowPath), usually(continuation));
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
VM& vm = this->vm();
|
|
LValue slowResultValue = lazySlowPath(
|
|
[=, &vm] (const Vector<Location>& locations) -> RefPtr<LazySlowPath::Generator> {
|
|
return createLazyCallGenerator(vm,
|
|
slowPathOperation, locations[0].directGPR(),
|
|
CCallHelpers::TrustedImmPtr(globalObject), locations[1].directGPR());
|
|
}, value);
|
|
ValueFromBlock slowResult = m_out.anchor(m_out.notNull(slowResultValue));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
LValue result = m_out.phi(
|
|
Int32, notCellResult, functionResult, objectResult, slowResult);
|
|
setBoolean(result);
|
|
}
|
|
|
|
void compileIsConstructor()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue value = lowJSValue(m_node->child1());
|
|
setBoolean(vmCall(Int32, operationIsConstructor, weakPointer(globalObject), value));
|
|
}
|
|
|
|
void compileIsTypedArrayView()
|
|
{
|
|
LValue value = lowJSValue(m_node->child1());
|
|
|
|
LBasicBlock isCellCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
ValueFromBlock notCellResult = m_out.anchor(m_out.booleanFalse);
|
|
m_out.branch(isCell(value, provenType(m_node->child1())), unsure(isCellCase), unsure(continuation));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isCellCase, continuation);
|
|
ValueFromBlock cellResult = m_out.anchor(isTypedArrayView(value, provenType(m_node->child1())));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setBoolean(m_out.phi(Int32, notCellResult, cellResult));
|
|
}
|
|
|
|
void compileTypeOf()
|
|
{
|
|
Edge child = m_node->child1();
|
|
LValue value = lowJSValue(child);
|
|
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(continuation);
|
|
|
|
Vector<ValueFromBlock> results;
|
|
|
|
buildTypeOf(
|
|
child, value,
|
|
[&] (TypeofType type) {
|
|
results.append(m_out.anchor(weakPointer(vm().smallStrings.typeString(type))));
|
|
m_out.jump(continuation);
|
|
});
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, results));
|
|
}
|
|
|
|
void compileInByVal()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
setJSValue(vmCall(Int64, operationInByVal, weakPointer(globalObject), lowCell(m_node->child1()), lowJSValue(m_node->child2())));
|
|
}
|
|
|
|
void compileInById()
|
|
{
|
|
Node* node = m_node;
|
|
CacheableIdentifier identifier = node->cacheableIdentifier();
|
|
LValue base = lowCell(m_node->child1());
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int64);
|
|
patchpoint->appendSomeRegister(base);
|
|
patchpoint->append(m_notCellMask, ValueRep::lateReg(GPRInfo::notCellMaskRegister));
|
|
patchpoint->append(m_numberTag, ValueRep::lateReg(GPRInfo::numberTagRegister));
|
|
|
|
patchpoint->clobber(RegisterSet::macroScratchRegisters());
|
|
|
|
RefPtr<PatchpointExceptionHandle> exceptionHandle =
|
|
preparePatchpointForExceptions(patchpoint);
|
|
|
|
State* state = &m_ftlState;
|
|
patchpoint->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
CallSiteIndex callSiteIndex =
|
|
state->jitCode->common.codeOrigins->addUniqueCallSiteIndex(node->origin.semantic);
|
|
|
|
// This is the direct exit target for operation calls.
|
|
Box<CCallHelpers::JumpList> exceptions =
|
|
exceptionHandle->scheduleExitCreation(params)->jumps(jit);
|
|
|
|
auto generator = Box<JITInByIdGenerator>::create(
|
|
jit.codeBlock(), node->origin.semantic, callSiteIndex,
|
|
params.unavailableRegisters(), identifier, JSValueRegs(params[1].gpr()),
|
|
JSValueRegs(params[0].gpr()));
|
|
|
|
generator->generateFastPath(jit);
|
|
CCallHelpers::Label done = jit.label();
|
|
|
|
params.addLatePath(
|
|
[=] (CCallHelpers& jit) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
generator->slowPathJump().link(&jit);
|
|
CCallHelpers::Label slowPathBegin = jit.label();
|
|
CCallHelpers::Call slowPathCall = callOperation(
|
|
*state, params.unavailableRegisters(), jit, node->origin.semantic,
|
|
exceptions.get(), operationInByIdOptimize, params[0].gpr(),
|
|
jit.codeBlock()->globalObjectFor(node->origin.semantic),
|
|
CCallHelpers::TrustedImmPtr(generator->stubInfo()), params[1].gpr(),
|
|
identifier.rawBits()).call();
|
|
jit.jump().linkTo(done, &jit);
|
|
|
|
generator->reportSlowPathCall(slowPathBegin, slowPathCall);
|
|
|
|
jit.addLinkTask(
|
|
[=] (LinkBuffer& linkBuffer) {
|
|
generator->finalize(linkBuffer, linkBuffer);
|
|
});
|
|
});
|
|
});
|
|
|
|
setJSValue(patchpoint);
|
|
}
|
|
|
|
void compileHasOwnProperty()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
LBasicBlock lastNext = nullptr;
|
|
|
|
LValue object = lowObject(m_node->child1());
|
|
LValue uniquedStringImpl;
|
|
LValue keyAsValue = nullptr;
|
|
switch (m_node->child2().useKind()) {
|
|
case StringUse: {
|
|
LBasicBlock isNonEmptyString = m_out.newBlock();
|
|
LBasicBlock isAtomString = m_out.newBlock();
|
|
|
|
keyAsValue = lowString(m_node->child2());
|
|
m_out.branch(isNotRopeString(keyAsValue, m_node->child2()), usually(isNonEmptyString), rarely(slowCase));
|
|
|
|
lastNext = m_out.appendTo(isNonEmptyString, isAtomString);
|
|
uniquedStringImpl = m_out.loadPtr(keyAsValue, m_heaps.JSString_value);
|
|
LValue isNotAtomic = m_out.testIsZero32(m_out.load32(uniquedStringImpl, m_heaps.StringImpl_hashAndFlags), m_out.constInt32(StringImpl::flagIsAtom()));
|
|
m_out.branch(isNotAtomic, rarely(slowCase), usually(isAtomString));
|
|
|
|
m_out.appendTo(isAtomString, slowCase);
|
|
break;
|
|
}
|
|
case SymbolUse: {
|
|
keyAsValue = lowSymbol(m_node->child2());
|
|
uniquedStringImpl = m_out.loadPtr(keyAsValue, m_heaps.Symbol_symbolImpl);
|
|
lastNext = m_out.insertNewBlocksBefore(slowCase);
|
|
break;
|
|
}
|
|
case UntypedUse: {
|
|
LBasicBlock isCellCase = m_out.newBlock();
|
|
LBasicBlock isStringCase = m_out.newBlock();
|
|
LBasicBlock notStringCase = m_out.newBlock();
|
|
LBasicBlock isNonEmptyString = m_out.newBlock();
|
|
LBasicBlock isSymbolCase = m_out.newBlock();
|
|
LBasicBlock hasUniquedStringImpl = m_out.newBlock();
|
|
|
|
keyAsValue = lowJSValue(m_node->child2());
|
|
m_out.branch(isCell(keyAsValue), usually(isCellCase), rarely(slowCase));
|
|
|
|
lastNext = m_out.appendTo(isCellCase, isStringCase);
|
|
m_out.branch(isString(keyAsValue), unsure(isStringCase), unsure(notStringCase));
|
|
|
|
m_out.appendTo(isStringCase, isNonEmptyString);
|
|
m_out.branch(isNotRopeString(keyAsValue, m_node->child2()), usually(isNonEmptyString), rarely(slowCase));
|
|
|
|
m_out.appendTo(isNonEmptyString, notStringCase);
|
|
LValue implFromString = m_out.loadPtr(keyAsValue, m_heaps.JSString_value);
|
|
ValueFromBlock stringResult = m_out.anchor(implFromString);
|
|
LValue isNotAtomic = m_out.testIsZero32(m_out.load32(implFromString, m_heaps.StringImpl_hashAndFlags), m_out.constInt32(StringImpl::flagIsAtom()));
|
|
m_out.branch(isNotAtomic, rarely(slowCase), usually(hasUniquedStringImpl));
|
|
|
|
m_out.appendTo(notStringCase, isSymbolCase);
|
|
m_out.branch(isSymbol(keyAsValue), unsure(isSymbolCase), unsure(slowCase));
|
|
|
|
m_out.appendTo(isSymbolCase, hasUniquedStringImpl);
|
|
ValueFromBlock symbolResult = m_out.anchor(m_out.loadPtr(keyAsValue, m_heaps.Symbol_symbolImpl));
|
|
m_out.jump(hasUniquedStringImpl);
|
|
|
|
m_out.appendTo(hasUniquedStringImpl, slowCase);
|
|
uniquedStringImpl = m_out.phi(pointerType(), stringResult, symbolResult);
|
|
break;
|
|
}
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
}
|
|
|
|
ASSERT(keyAsValue);
|
|
|
|
// Note that we don't test if the hash is zero here. AtomStringImpl's can't have a zero
|
|
// hash, however, a SymbolImpl may. But, because this is a cache, we don't care. We only
|
|
// ever load the result from the cache if the cache entry matches what we are querying for.
|
|
// So we either get super lucky and use zero for the hash and somehow collide with the entity
|
|
// we're looking for, or we realize we're comparing against another entity, and go to the
|
|
// slow path anyways.
|
|
LValue hash = m_out.lShr(m_out.load32(uniquedStringImpl, m_heaps.StringImpl_hashAndFlags), m_out.constInt32(StringImpl::s_flagCount));
|
|
|
|
LValue structureID = m_out.load32(object, m_heaps.JSCell_structureID);
|
|
LValue index = m_out.add(hash, structureID);
|
|
index = m_out.zeroExtPtr(m_out.bitAnd(index, m_out.constInt32(HasOwnPropertyCache::mask)));
|
|
ASSERT(vm().hasOwnPropertyCache());
|
|
LValue cache = m_out.constIntPtr(vm().hasOwnPropertyCache());
|
|
|
|
IndexedAbstractHeap& heap = m_heaps.HasOwnPropertyCache;
|
|
LValue sameStructureID = m_out.equal(structureID, m_out.load32(m_out.baseIndex(heap, cache, index, JSValue(), HasOwnPropertyCache::Entry::offsetOfStructureID())));
|
|
LValue sameImpl = m_out.equal(uniquedStringImpl, m_out.loadPtr(m_out.baseIndex(heap, cache, index, JSValue(), HasOwnPropertyCache::Entry::offsetOfImpl())));
|
|
ValueFromBlock fastResult = m_out.anchor(m_out.load8ZeroExt32(m_out.baseIndex(heap, cache, index, JSValue(), HasOwnPropertyCache::Entry::offsetOfResult())));
|
|
LValue cacheHit = m_out.bitAnd(sameStructureID, sameImpl);
|
|
|
|
m_out.branch(m_out.notZero32(cacheHit), usually(continuation), rarely(slowCase));
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
ValueFromBlock slowResult;
|
|
slowResult = m_out.anchor(m_out.notZero64(vmCall(Int64, operationHasOwnProperty, weakPointer(globalObject), object, keyAsValue)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setBoolean(m_out.phi(Int32, fastResult, slowResult));
|
|
}
|
|
|
|
void compileParseInt()
|
|
{
|
|
RELEASE_ASSERT(m_node->child1().useKind() == UntypedUse || m_node->child1().useKind() == StringUse);
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue result;
|
|
if (m_node->child2()) {
|
|
LValue radix = lowInt32(m_node->child2());
|
|
if (m_node->child1().useKind() == UntypedUse)
|
|
result = vmCall(Int64, operationParseIntGeneric, weakPointer(globalObject), lowJSValue(m_node->child1()), radix);
|
|
else
|
|
result = vmCall(Int64, operationParseIntString, weakPointer(globalObject), lowString(m_node->child1()), radix);
|
|
} else {
|
|
if (m_node->child1().useKind() == UntypedUse)
|
|
result = vmCall(Int64, operationParseIntNoRadixGeneric, weakPointer(globalObject), lowJSValue(m_node->child1()));
|
|
else
|
|
result = vmCall(Int64, operationParseIntStringNoRadix, weakPointer(globalObject), lowString(m_node->child1()));
|
|
}
|
|
setJSValue(result);
|
|
}
|
|
|
|
void compileOverridesHasInstance()
|
|
{
|
|
FrozenValue* defaultHasInstanceFunction = m_node->cellOperand();
|
|
ASSERT(defaultHasInstanceFunction->cell()->inherits<JSFunction>(vm()));
|
|
|
|
LValue constructor = lowCell(m_node->child1());
|
|
LValue hasInstance = lowJSValue(m_node->child2());
|
|
|
|
LBasicBlock defaultHasInstance = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
// Unlike in the DFG, we don't worry about cleaning this code up for the case where we have proven the hasInstanceValue is a constant as B3 should fix it for us.
|
|
|
|
ValueFromBlock notDefaultHasInstanceResult = m_out.anchor(m_out.booleanTrue);
|
|
m_out.branch(m_out.notEqual(hasInstance, frozenPointer(defaultHasInstanceFunction)), unsure(continuation), unsure(defaultHasInstance));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(defaultHasInstance, continuation);
|
|
ValueFromBlock implementsDefaultHasInstanceResult = m_out.anchor(m_out.testIsZero32(
|
|
m_out.load8ZeroExt32(constructor, m_heaps.JSCell_typeInfoFlags),
|
|
m_out.constInt32(ImplementsDefaultHasInstance)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setBoolean(m_out.phi(Int32, implementsDefaultHasInstanceResult, notDefaultHasInstanceResult));
|
|
}
|
|
|
|
void compileCheckTypeInfoFlags()
|
|
{
|
|
speculate(
|
|
BadTypeInfoFlags, noValue(), nullptr,
|
|
m_out.testIsZero32(
|
|
m_out.load8ZeroExt32(lowCell(m_node->child1()), m_heaps.JSCell_typeInfoFlags),
|
|
m_out.constInt32(m_node->typeInfoOperand())));
|
|
}
|
|
|
|
void compileInstanceOf()
|
|
{
|
|
Node* node = m_node;
|
|
State* state = &m_ftlState;
|
|
|
|
LValue value;
|
|
LValue prototype;
|
|
bool valueIsCell;
|
|
bool prototypeIsCell;
|
|
if (m_node->child1().useKind() == CellUse
|
|
&& m_node->child2().useKind() == CellUse) {
|
|
value = lowCell(m_node->child1());
|
|
prototype = lowCell(m_node->child2());
|
|
|
|
valueIsCell = true;
|
|
prototypeIsCell = true;
|
|
} else {
|
|
DFG_ASSERT(m_graph, m_node, m_node->child1().useKind() == UntypedUse);
|
|
DFG_ASSERT(m_graph, m_node, m_node->child2().useKind() == UntypedUse);
|
|
|
|
value = lowJSValue(m_node->child1());
|
|
prototype = lowJSValue(m_node->child2());
|
|
|
|
valueIsCell = abstractValue(m_node->child1()).isType(SpecCell);
|
|
prototypeIsCell = abstractValue(m_node->child2()).isType(SpecCell);
|
|
}
|
|
|
|
bool prototypeIsObject = abstractValue(m_node->child2()).isType(SpecObject | ~SpecCell);
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int64);
|
|
patchpoint->appendSomeRegister(value);
|
|
patchpoint->appendSomeRegister(prototype);
|
|
patchpoint->append(m_notCellMask, ValueRep::lateReg(GPRInfo::notCellMaskRegister));
|
|
patchpoint->append(m_numberTag, ValueRep::lateReg(GPRInfo::numberTagRegister));
|
|
patchpoint->numGPScratchRegisters = 2;
|
|
patchpoint->resultConstraints = { ValueRep::SomeEarlyRegister };
|
|
patchpoint->clobber(RegisterSet::macroScratchRegisters());
|
|
|
|
RefPtr<PatchpointExceptionHandle> exceptionHandle =
|
|
preparePatchpointForExceptions(patchpoint);
|
|
|
|
patchpoint->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
GPRReg resultGPR = params[0].gpr();
|
|
GPRReg valueGPR = params[1].gpr();
|
|
GPRReg prototypeGPR = params[2].gpr();
|
|
GPRReg scratchGPR = params.gpScratch(0);
|
|
GPRReg scratch2GPR = params.gpScratch(1);
|
|
|
|
CCallHelpers::Jump doneJump;
|
|
if (!valueIsCell) {
|
|
CCallHelpers::Jump isCell = jit.branchIfCell(valueGPR);
|
|
jit.boxBooleanPayload(false, resultGPR);
|
|
doneJump = jit.jump();
|
|
isCell.link(&jit);
|
|
}
|
|
|
|
CCallHelpers::JumpList slowCases;
|
|
if (!prototypeIsCell)
|
|
slowCases.append(jit.branchIfNotCell(prototypeGPR));
|
|
|
|
CallSiteIndex callSiteIndex =
|
|
state->jitCode->common.codeOrigins->addUniqueCallSiteIndex(node->origin.semantic);
|
|
|
|
// This is the direct exit target for operation calls.
|
|
Box<CCallHelpers::JumpList> exceptions =
|
|
exceptionHandle->scheduleExitCreation(params)->jumps(jit);
|
|
|
|
auto generator = Box<JITInstanceOfGenerator>::create(
|
|
jit.codeBlock(), node->origin.semantic, callSiteIndex,
|
|
params.unavailableRegisters(), resultGPR, valueGPR, prototypeGPR, scratchGPR,
|
|
scratch2GPR, prototypeIsObject);
|
|
generator->generateFastPath(jit);
|
|
CCallHelpers::Label done = jit.label();
|
|
|
|
params.addLatePath(
|
|
[=] (CCallHelpers& jit) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
J_JITOperation_GSsiJJ optimizationFunction = operationInstanceOfOptimize;
|
|
|
|
slowCases.link(&jit);
|
|
CCallHelpers::Label slowPathBegin = jit.label();
|
|
CCallHelpers::Call slowPathCall = callOperation(
|
|
*state, params.unavailableRegisters(), jit, node->origin.semantic,
|
|
exceptions.get(), optimizationFunction, resultGPR,
|
|
jit.codeBlock()->globalObjectFor(node->origin.semantic),
|
|
CCallHelpers::TrustedImmPtr(generator->stubInfo()), valueGPR,
|
|
prototypeGPR).call();
|
|
jit.jump().linkTo(done, &jit);
|
|
|
|
generator->reportSlowPathCall(slowPathBegin, slowPathCall);
|
|
|
|
jit.addLinkTask(
|
|
[=] (LinkBuffer& linkBuffer) {
|
|
generator->finalize(linkBuffer, linkBuffer);
|
|
});
|
|
});
|
|
|
|
if (doneJump.isSet())
|
|
doneJump.link(&jit);
|
|
});
|
|
|
|
// This returns a boxed boolean.
|
|
setJSValue(patchpoint);
|
|
}
|
|
|
|
void compileInstanceOfCustom()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue value = lowJSValue(m_node->child1());
|
|
LValue constructor = lowCell(m_node->child2());
|
|
LValue hasInstance = lowJSValue(m_node->child3());
|
|
|
|
setBoolean(m_out.notZero64(vmCall(Int64, operationInstanceOfCustom, weakPointer(globalObject), value, constructor, hasInstance)));
|
|
}
|
|
|
|
void compileCountExecution()
|
|
{
|
|
TypedPointer counter = m_out.absolute(m_node->executionCounter()->address());
|
|
m_out.store64(m_out.add(m_out.load64(counter), m_out.constInt64(1)), counter);
|
|
}
|
|
|
|
void compileSuperSamplerBegin()
|
|
{
|
|
TypedPointer counter = m_out.absolute(bitwise_cast<void*>(&g_superSamplerCount));
|
|
m_out.store32(m_out.add(m_out.load32(counter), m_out.constInt32(1)), counter);
|
|
}
|
|
|
|
void compileSuperSamplerEnd()
|
|
{
|
|
TypedPointer counter = m_out.absolute(bitwise_cast<void*>(&g_superSamplerCount));
|
|
m_out.store32(m_out.sub(m_out.load32(counter), m_out.constInt32(1)), counter);
|
|
}
|
|
|
|
void compileStoreBarrier()
|
|
{
|
|
emitStoreBarrier(lowCell(m_node->child1()), m_node->op() == FencedStoreBarrier);
|
|
}
|
|
|
|
void compileHasIndexedProperty(S_JITOperation_GCZ slowPathOperation)
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue base = lowCell(m_graph.varArgChild(m_node, 0));
|
|
LValue index = lowInt32(m_graph.varArgChild(m_node, 1));
|
|
ArrayMode mode = m_node->arrayMode();
|
|
|
|
switch (m_node->arrayMode().type()) {
|
|
case Array::Int32:
|
|
case Array::Contiguous: {
|
|
LValue storage = lowStorage(m_graph.varArgChild(m_node, 2));
|
|
|
|
IndexedAbstractHeap& heap = mode.type() == Array::Int32 ?
|
|
m_heaps.indexedInt32Properties : m_heaps.indexedContiguousProperties;
|
|
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
LBasicBlock lastNext = nullptr;
|
|
|
|
if (!mode.isInBounds()) {
|
|
LBasicBlock checkHole = m_out.newBlock();
|
|
m_out.branch(
|
|
m_out.aboveOrEqual(
|
|
index, m_out.load32NonNegative(storage, m_heaps.Butterfly_publicLength)),
|
|
rarely(slowCase), usually(checkHole));
|
|
lastNext = m_out.appendTo(checkHole, slowCase);
|
|
} else
|
|
lastNext = m_out.insertNewBlocksBefore(slowCase);
|
|
|
|
LValue checkHoleResultValue =
|
|
m_out.notZero64(m_out.load64(baseIndex(heap, storage, index, m_graph.varArgChild(m_node, 1))));
|
|
ValueFromBlock checkHoleResult = m_out.anchor(checkHoleResultValue);
|
|
if (mode.isInBoundsSaneChain())
|
|
m_out.jump(continuation);
|
|
else if (!mode.isInBounds())
|
|
m_out.branch(checkHoleResultValue, usually(continuation), rarely(slowCase));
|
|
else
|
|
speculateAndJump(continuation, LoadFromHole, noValue(), nullptr, checkHoleResultValue);
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
ValueFromBlock slowResult = m_out.anchor(
|
|
m_out.notZero64(vmCall(Int64, slowPathOperation, weakPointer(globalObject), base, index)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setBoolean(m_out.phi(Int32, checkHoleResult, slowResult));
|
|
return;
|
|
}
|
|
case Array::Double: {
|
|
LValue storage = lowStorage(m_graph.varArgChild(m_node, 2));
|
|
|
|
IndexedAbstractHeap& heap = m_heaps.indexedDoubleProperties;
|
|
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
LBasicBlock lastNext = nullptr;
|
|
|
|
if (!m_node->arrayMode().isInBounds()) {
|
|
LBasicBlock checkHole = m_out.newBlock();
|
|
m_out.branch(
|
|
m_out.aboveOrEqual(
|
|
index, m_out.load32NonNegative(storage, m_heaps.Butterfly_publicLength)),
|
|
rarely(slowCase), usually(checkHole));
|
|
lastNext = m_out.appendTo(checkHole, slowCase);
|
|
} else
|
|
lastNext = m_out.insertNewBlocksBefore(slowCase);
|
|
|
|
LValue doubleValue = m_out.loadDouble(baseIndex(heap, storage, index, m_graph.varArgChild(m_node, 1)));
|
|
LValue checkHoleResultValue = m_out.doubleEqual(doubleValue, doubleValue);
|
|
ValueFromBlock checkHoleResult = m_out.anchor(checkHoleResultValue);
|
|
if (mode.isInBoundsSaneChain())
|
|
m_out.jump(continuation);
|
|
else if (!mode.isInBounds())
|
|
m_out.branch(checkHoleResultValue, usually(continuation), rarely(slowCase));
|
|
else
|
|
speculateAndJump(continuation, LoadFromHole, noValue(), nullptr, checkHoleResultValue);
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
ValueFromBlock slowResult = m_out.anchor(
|
|
m_out.notZero64(vmCall(Int64, slowPathOperation, weakPointer(globalObject), base, index)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setBoolean(m_out.phi(Int32, checkHoleResult, slowResult));
|
|
return;
|
|
}
|
|
|
|
case Array::ArrayStorage: {
|
|
LValue storage = lowStorage(m_graph.varArgChild(m_node, 2));
|
|
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
LBasicBlock lastNext = nullptr;
|
|
|
|
if (!m_node->arrayMode().isInBounds()) {
|
|
LBasicBlock checkHole = m_out.newBlock();
|
|
m_out.branch(
|
|
m_out.aboveOrEqual(
|
|
index, m_out.load32NonNegative(storage, m_heaps.ArrayStorage_vectorLength)),
|
|
rarely(slowCase), usually(checkHole));
|
|
lastNext = m_out.appendTo(checkHole, slowCase);
|
|
} else
|
|
lastNext = m_out.insertNewBlocksBefore(slowCase);
|
|
|
|
LValue checkHoleResultValue =
|
|
m_out.notZero64(m_out.load64(baseIndex(m_heaps.ArrayStorage_vector, storage, index, m_graph.varArgChild(m_node, 1))));
|
|
ValueFromBlock checkHoleResult = m_out.anchor(checkHoleResultValue);
|
|
if (mode.isInBoundsSaneChain())
|
|
m_out.jump(continuation);
|
|
else if (!mode.isInBounds())
|
|
m_out.branch(checkHoleResultValue, usually(continuation), rarely(slowCase));
|
|
else
|
|
speculateAndJump(continuation, LoadFromHole, noValue(), nullptr, checkHoleResultValue);
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
ValueFromBlock slowResult = m_out.anchor(
|
|
m_out.notZero64(vmCall(Int64, slowPathOperation, weakPointer(globalObject), base, index)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setBoolean(m_out.phi(Int32, checkHoleResult, slowResult));
|
|
break;
|
|
}
|
|
|
|
default: {
|
|
setBoolean(m_out.notZero64(vmCall(Int64, slowPathOperation, weakPointer(globalObject), base, index)));
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
void compileHasEnumerableProperty()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue base = lowJSValue(m_node->child1());
|
|
LValue property = lowCell(m_node->child2());
|
|
setJSValue(vmCall(Int64, operationHasEnumerableProperty, weakPointer(globalObject), base, property));
|
|
}
|
|
|
|
template <typename SlowPathCall>
|
|
void compileHasStructurePropertyImpl(LValue base, SlowPathCall slowPathCall)
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue property = lowString(m_node->child2());
|
|
LValue enumerator = lowCell(m_node->child3());
|
|
|
|
LBasicBlock isCellCase = m_out.newBlock();
|
|
LBasicBlock correctStructure = m_out.newBlock();
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(isCell(base, provenType(m_node->child1())),
|
|
usually(isCellCase), rarely(slowPath));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isCellCase, correctStructure);
|
|
|
|
m_out.branch(m_out.notEqual(
|
|
m_out.load32(base, m_heaps.JSCell_structureID),
|
|
m_out.load32(enumerator, m_heaps.JSPropertyNameEnumerator_cachedStructureID)),
|
|
rarely(slowPath), usually(correctStructure));
|
|
|
|
m_out.appendTo(correctStructure, slowPath);
|
|
ValueFromBlock correctStructureResult = m_out.anchor(m_out.booleanTrue);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
ValueFromBlock slowPathResult = m_out.anchor(
|
|
m_out.equal(
|
|
m_out.constInt64(JSValue::encode(jsBoolean(true))),
|
|
vmCall(Int64, slowPathCall, weakPointer(globalObject), base, property)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setBoolean(m_out.phi(Int32, correctStructureResult, slowPathResult));
|
|
}
|
|
|
|
void compileHasEnumerableStructureProperty()
|
|
{
|
|
compileHasStructurePropertyImpl(lowJSValue(m_node->child1()), operationHasEnumerableProperty);
|
|
}
|
|
|
|
void compileHasOwnStructureProperty()
|
|
{
|
|
compileHasStructurePropertyImpl(lowCell(m_node->child1()), operationHasOwnStructureProperty);
|
|
}
|
|
|
|
void compileInStructureProperty()
|
|
{
|
|
compileHasStructurePropertyImpl(lowCell(m_node->child1()), operationInStructureProperty);
|
|
}
|
|
|
|
void compileGetDirectPname()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue base = lowCell(m_graph.varArgChild(m_node, 0));
|
|
LValue property = lowCell(m_graph.varArgChild(m_node, 1));
|
|
LValue index = lowInt32(m_graph.varArgChild(m_node, 2));
|
|
LValue enumerator = lowCell(m_graph.varArgChild(m_node, 3));
|
|
|
|
LBasicBlock checkOffset = m_out.newBlock();
|
|
LBasicBlock inlineLoad = m_out.newBlock();
|
|
LBasicBlock outOfLineLoad = m_out.newBlock();
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(m_out.notEqual(
|
|
m_out.load32(base, m_heaps.JSCell_structureID),
|
|
m_out.load32(enumerator, m_heaps.JSPropertyNameEnumerator_cachedStructureID)),
|
|
rarely(slowCase), usually(checkOffset));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(checkOffset, inlineLoad);
|
|
m_out.branch(m_out.aboveOrEqual(index, m_out.load32(enumerator, m_heaps.JSPropertyNameEnumerator_cachedInlineCapacity)),
|
|
unsure(outOfLineLoad), unsure(inlineLoad));
|
|
|
|
m_out.appendTo(inlineLoad, outOfLineLoad);
|
|
ValueFromBlock inlineResult = m_out.anchor(
|
|
m_out.load64(m_out.baseIndex(m_heaps.properties.atAnyNumber(),
|
|
base, m_out.zeroExt(index, Int64), ScaleEight, JSObject::offsetOfInlineStorage())));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(outOfLineLoad, slowCase);
|
|
LValue storage = m_out.loadPtr(base, m_heaps.JSObject_butterfly);
|
|
LValue realIndex = m_out.signExt32To64(
|
|
m_out.neg(m_out.sub(index, m_out.load32(enumerator, m_heaps.JSPropertyNameEnumerator_cachedInlineCapacity))));
|
|
int32_t offsetOfFirstProperty = static_cast<int32_t>(offsetInButterfly(firstOutOfLineOffset)) * sizeof(EncodedJSValue);
|
|
ValueFromBlock outOfLineResult = m_out.anchor(
|
|
m_out.load64(m_out.baseIndex(m_heaps.properties.atAnyNumber(), storage, realIndex, ScaleEight, offsetOfFirstProperty)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
ValueFromBlock slowCaseResult = m_out.anchor(
|
|
vmCall(Int64, operationGetByVal, weakPointer(globalObject), base, property));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, inlineResult, outOfLineResult, slowCaseResult));
|
|
}
|
|
|
|
void compileGetEnumerableLength()
|
|
{
|
|
LValue enumerator = lowCell(m_node->child1());
|
|
setInt32(m_out.load32(enumerator, m_heaps.JSPropertyNameEnumerator_indexLength));
|
|
}
|
|
|
|
void compileGetPropertyEnumerator()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
if (m_node->child1().useKind() == CellUse)
|
|
setJSValue(vmCall(Int64, operationGetPropertyEnumeratorCell, weakPointer(globalObject), lowCell(m_node->child1())));
|
|
else
|
|
setJSValue(vmCall(Int64, operationGetPropertyEnumerator, weakPointer(globalObject), lowJSValue(m_node->child1())));
|
|
}
|
|
|
|
void compileGetEnumeratorStructurePname()
|
|
{
|
|
LValue enumerator = lowCell(m_node->child1());
|
|
LValue index = lowInt32(m_node->child2());
|
|
|
|
LBasicBlock inBounds = m_out.newBlock();
|
|
LBasicBlock outOfBounds = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(m_out.below(index, m_out.load32(enumerator, m_heaps.JSPropertyNameEnumerator_endStructurePropertyIndex)),
|
|
usually(inBounds), rarely(outOfBounds));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(inBounds, outOfBounds);
|
|
LValue storage = m_out.loadPtr(enumerator, m_heaps.JSPropertyNameEnumerator_cachedPropertyNamesVector);
|
|
ValueFromBlock inBoundsResult = m_out.anchor(
|
|
m_out.loadPtr(m_out.baseIndex(m_heaps.JSPropertyNameEnumerator_cachedPropertyNamesVectorContents, storage, m_out.zeroExtPtr(index))));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(outOfBounds, continuation);
|
|
ValueFromBlock outOfBoundsResult = m_out.anchor(m_out.constInt64(JSValue::ValueNull));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, inBoundsResult, outOfBoundsResult));
|
|
}
|
|
|
|
void compileGetEnumeratorGenericPname()
|
|
{
|
|
LValue enumerator = lowCell(m_node->child1());
|
|
LValue index = lowInt32(m_node->child2());
|
|
|
|
LBasicBlock inBounds = m_out.newBlock();
|
|
LBasicBlock outOfBounds = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(m_out.below(index, m_out.load32(enumerator, m_heaps.JSPropertyNameEnumerator_endGenericPropertyIndex)),
|
|
usually(inBounds), rarely(outOfBounds));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(inBounds, outOfBounds);
|
|
LValue storage = m_out.loadPtr(enumerator, m_heaps.JSPropertyNameEnumerator_cachedPropertyNamesVector);
|
|
ValueFromBlock inBoundsResult = m_out.anchor(
|
|
m_out.loadPtr(m_out.baseIndex(m_heaps.JSPropertyNameEnumerator_cachedPropertyNamesVectorContents, storage, m_out.zeroExtPtr(index))));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(outOfBounds, continuation);
|
|
ValueFromBlock outOfBoundsResult = m_out.anchor(m_out.constInt64(JSValue::ValueNull));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, inBoundsResult, outOfBoundsResult));
|
|
}
|
|
|
|
void compileToIndexString()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue index = lowInt32(m_node->child1());
|
|
setJSValue(vmCall(Int64, operationToIndexString, weakPointer(globalObject), index));
|
|
}
|
|
|
|
void compileCheckStructureImmediate()
|
|
{
|
|
LValue structure = lowCell(m_node->child1());
|
|
checkStructure(
|
|
structure, noValue(), BadCache, m_node->structureSet(),
|
|
[this] (RegisteredStructure structure) {
|
|
return weakStructure(structure);
|
|
});
|
|
}
|
|
|
|
void compileMaterializeNewObject()
|
|
{
|
|
ObjectMaterializationData& data = m_node->objectMaterializationData();
|
|
|
|
// Lower the values first, to avoid creating values inside a control flow diamond.
|
|
|
|
Vector<LValue, 8> values;
|
|
for (unsigned i = 0; i < data.m_properties.size(); ++i) {
|
|
Edge edge = m_graph.varArgChild(m_node, 1 + i);
|
|
switch (data.m_properties[i].kind()) {
|
|
case PublicLengthPLoc:
|
|
case VectorLengthPLoc:
|
|
values.append(lowInt32(edge));
|
|
break;
|
|
default:
|
|
values.append(lowJSValue(edge));
|
|
break;
|
|
}
|
|
}
|
|
|
|
RegisteredStructureSet set = m_node->structureSet();
|
|
|
|
Vector<LBasicBlock, 1> blocks(set.size());
|
|
for (unsigned i = set.size(); i--;)
|
|
blocks[i] = m_out.newBlock();
|
|
LBasicBlock dummyDefault = m_out.newBlock();
|
|
LBasicBlock outerContinuation = m_out.newBlock();
|
|
|
|
Vector<SwitchCase, 1> cases(set.size());
|
|
for (unsigned i = set.size(); i--;)
|
|
cases[i] = SwitchCase(weakStructure(set.at(i)), blocks[i], Weight(1));
|
|
m_out.switchInstruction(
|
|
lowCell(m_graph.varArgChild(m_node, 0)), cases, dummyDefault, Weight(0));
|
|
|
|
LBasicBlock outerLastNext = m_out.m_nextBlock;
|
|
|
|
Vector<ValueFromBlock, 1> results;
|
|
|
|
for (unsigned i = set.size(); i--;) {
|
|
m_out.appendTo(blocks[i], i + 1 < set.size() ? blocks[i + 1] : dummyDefault);
|
|
|
|
RegisteredStructure structure = set.at(i);
|
|
|
|
LValue object;
|
|
LValue butterfly;
|
|
|
|
if (structure->outOfLineCapacity() || hasIndexedProperties(structure->indexingType())) {
|
|
Allocator cellAllocator;
|
|
if (structure->type() == JSType::ArrayType)
|
|
cellAllocator = allocatorForNonVirtualConcurrently<JSArray>(vm(), JSArray::allocationSize(structure->inlineCapacity()), AllocatorForMode::AllocatorIfExists);
|
|
else
|
|
cellAllocator = allocatorForNonVirtualConcurrently<JSFinalObject>(vm(), JSFinalObject::allocationSize(structure->inlineCapacity()), AllocatorForMode::AllocatorIfExists);
|
|
|
|
bool hasIndexingHeader = hasIndexedProperties(structure->indexingType());
|
|
unsigned indexingHeaderSize = 0;
|
|
LValue indexingPayloadSizeInBytes = m_out.intPtrZero;
|
|
LValue vectorLength = m_out.int32Zero;
|
|
LValue publicLength = m_out.int32Zero;
|
|
if (hasIndexingHeader) {
|
|
indexingHeaderSize = sizeof(IndexingHeader);
|
|
for (unsigned i = data.m_properties.size(); i--;) {
|
|
PromotedLocationDescriptor descriptor = data.m_properties[i];
|
|
switch (descriptor.kind()) {
|
|
case PublicLengthPLoc:
|
|
publicLength = values[i];
|
|
break;
|
|
case VectorLengthPLoc:
|
|
vectorLength = values[i];
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
indexingPayloadSizeInBytes =
|
|
m_out.mul(m_out.zeroExtPtr(vectorLength), m_out.intPtrEight);
|
|
}
|
|
|
|
LValue butterflySize = m_out.add(
|
|
m_out.constIntPtr(
|
|
structure->outOfLineCapacity() * sizeof(JSValue) + indexingHeaderSize),
|
|
indexingPayloadSizeInBytes);
|
|
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(slowPath);
|
|
|
|
ValueFromBlock noButterfly = m_out.anchor(m_out.intPtrZero);
|
|
|
|
LValue startOfStorage = allocateHeapCell(
|
|
allocatorForSize(vm().jsValueGigacageAuxiliarySpace, butterflySize, slowPath),
|
|
slowPath);
|
|
|
|
LValue fastButterflyValue = m_out.add(
|
|
startOfStorage,
|
|
m_out.constIntPtr(
|
|
structure->outOfLineCapacity() * sizeof(JSValue) + sizeof(IndexingHeader)));
|
|
|
|
ValueFromBlock haveButterfly = m_out.anchor(fastButterflyValue);
|
|
|
|
splatWords(
|
|
fastButterflyValue,
|
|
m_out.constInt32(-structure->outOfLineCapacity() - 1),
|
|
m_out.constInt32(-1),
|
|
m_out.int64Zero, m_heaps.properties.atAnyNumber());
|
|
|
|
m_out.store32(vectorLength, fastButterflyValue, m_heaps.Butterfly_vectorLength);
|
|
|
|
LValue fastObjectValue = allocateObject(
|
|
m_out.constIntPtr(cellAllocator.localAllocator()), structure, fastButterflyValue,
|
|
slowPath);
|
|
|
|
ValueFromBlock fastObject = m_out.anchor(fastObjectValue);
|
|
ValueFromBlock fastButterfly = m_out.anchor(fastButterflyValue);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
|
|
LValue butterflyValue = m_out.phi(pointerType(), noButterfly, haveButterfly);
|
|
|
|
VM& vm = this->vm();
|
|
LValue slowObjectValue;
|
|
if (hasIndexingHeader) {
|
|
slowObjectValue = lazySlowPath(
|
|
[=, &vm] (const Vector<Location>& locations) -> RefPtr<LazySlowPath::Generator> {
|
|
return createLazyCallGenerator(vm,
|
|
operationNewObjectWithButterflyWithIndexingHeaderAndVectorLength,
|
|
locations[0].directGPR(), &vm, CCallHelpers::TrustedImmPtr(structure.get()),
|
|
locations[1].directGPR(), locations[2].directGPR());
|
|
},
|
|
vectorLength, butterflyValue);
|
|
} else {
|
|
slowObjectValue = lazySlowPath(
|
|
[=, &vm] (const Vector<Location>& locations) -> RefPtr<LazySlowPath::Generator> {
|
|
return createLazyCallGenerator(vm,
|
|
operationNewObjectWithButterfly, locations[0].directGPR(), &vm,
|
|
CCallHelpers::TrustedImmPtr(structure.get()), locations[1].directGPR());
|
|
},
|
|
butterflyValue);
|
|
}
|
|
ValueFromBlock slowObject = m_out.anchor(slowObjectValue);
|
|
ValueFromBlock slowButterfly = m_out.anchor(
|
|
m_out.loadPtr(slowObjectValue, m_heaps.JSObject_butterfly));
|
|
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
|
|
object = m_out.phi(pointerType(), fastObject, slowObject);
|
|
butterfly = m_out.phi(pointerType(), fastButterfly, slowButterfly);
|
|
|
|
m_out.store32(publicLength, butterfly, m_heaps.Butterfly_publicLength);
|
|
|
|
initializeArrayElements(m_out.constInt32(structure->indexingType()), m_out.int32Zero, vectorLength, butterfly);
|
|
|
|
HashMap<int32_t, LValue, DefaultHash<int32_t>, WTF::UnsignedWithZeroKeyHashTraits<int32_t>> indexMap;
|
|
Vector<int32_t> indices;
|
|
for (unsigned i = data.m_properties.size(); i--;) {
|
|
PromotedLocationDescriptor descriptor = data.m_properties[i];
|
|
if (descriptor.kind() != IndexedPropertyPLoc)
|
|
continue;
|
|
int32_t index = static_cast<int32_t>(descriptor.info());
|
|
|
|
auto result = indexMap.add(index, values[i]);
|
|
DFG_ASSERT(m_graph, m_node, result); // Duplicates are illegal.
|
|
|
|
indices.append(index);
|
|
}
|
|
|
|
if (!indices.isEmpty()) {
|
|
std::sort(indices.begin(), indices.end());
|
|
|
|
Vector<LBasicBlock> blocksWithStores(indices.size());
|
|
Vector<LBasicBlock> blocksWithChecks(indices.size());
|
|
|
|
for (unsigned i = indices.size(); i--;) {
|
|
blocksWithStores[i] = m_out.newBlock();
|
|
blocksWithChecks[i] = m_out.newBlock(); // blocksWithChecks[0] is the continuation.
|
|
}
|
|
|
|
LBasicBlock indexLastNext = m_out.m_nextBlock;
|
|
|
|
for (unsigned i = indices.size(); i--;) {
|
|
int32_t index = indices[i];
|
|
LValue value = indexMap.get(index);
|
|
|
|
m_out.branch(
|
|
m_out.below(m_out.constInt32(index), publicLength),
|
|
unsure(blocksWithStores[i]), unsure(blocksWithChecks[i]));
|
|
|
|
m_out.appendTo(blocksWithStores[i], blocksWithChecks[i]);
|
|
|
|
// This has to type-check and convert its inputs, but it cannot do so in a
|
|
// way that updates AI. That's a bit annoying, but if you think about how
|
|
// sinking works, it's actually not a bad thing. We are virtually guaranteed
|
|
// that these type checks will not fail, since the type checks that guarded
|
|
// the original stores to the array are still somewhere above this point.
|
|
Output::StoreType storeType;
|
|
IndexedAbstractHeap* heap;
|
|
switch (structure->indexingType()) {
|
|
case ALL_INT32_INDEXING_TYPES:
|
|
// FIXME: This could use the proven type if we had the Edge for the
|
|
// value. https://bugs.webkit.org/show_bug.cgi?id=155311
|
|
speculate(BadType, noValue(), nullptr, isNotInt32(value));
|
|
storeType = Output::Store64;
|
|
heap = &m_heaps.indexedInt32Properties;
|
|
break;
|
|
|
|
case ALL_DOUBLE_INDEXING_TYPES: {
|
|
// FIXME: If the source is ValueRep, we should avoid emitting any
|
|
// checks. We could also avoid emitting checks if we had the Edge of
|
|
// this value. https://bugs.webkit.org/show_bug.cgi?id=155311
|
|
|
|
LBasicBlock intCase = m_out.newBlock();
|
|
LBasicBlock doubleCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(isInt32(value), unsure(intCase), unsure(doubleCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(intCase, doubleCase);
|
|
|
|
ValueFromBlock intResult =
|
|
m_out.anchor(m_out.intToDouble(unboxInt32(value)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(doubleCase, continuation);
|
|
|
|
speculate(BadType, noValue(), nullptr, isNumber(value));
|
|
ValueFromBlock doubleResult = m_out.anchor(unboxDouble(value));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
value = m_out.phi(Double, intResult, doubleResult);
|
|
storeType = Output::StoreDouble;
|
|
heap = &m_heaps.indexedDoubleProperties;
|
|
break;
|
|
}
|
|
|
|
case ALL_CONTIGUOUS_INDEXING_TYPES:
|
|
storeType = Output::Store64;
|
|
heap = &m_heaps.indexedContiguousProperties;
|
|
break;
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Invalid indexing type");
|
|
break;
|
|
}
|
|
|
|
m_out.store(value, m_out.address(butterfly, heap->at(index)), storeType);
|
|
|
|
m_out.jump(blocksWithChecks[i]);
|
|
m_out.appendTo(
|
|
blocksWithChecks[i], i ? blocksWithStores[i - 1] : indexLastNext);
|
|
}
|
|
}
|
|
} else {
|
|
// In the easy case where we can do a one-shot allocation, we simply allocate the
|
|
// object to directly have the desired structure.
|
|
object = allocateObject(structure);
|
|
butterfly = nullptr; // Don't have one, don't need one.
|
|
}
|
|
|
|
BitVector setInlineOffsets;
|
|
for (PropertyMapEntry entry : structure->getPropertiesConcurrently()) {
|
|
for (unsigned i = data.m_properties.size(); i--;) {
|
|
PromotedLocationDescriptor descriptor = data.m_properties[i];
|
|
if (descriptor.kind() != NamedPropertyPLoc)
|
|
continue;
|
|
if (m_graph.identifiers()[descriptor.info()] != entry.key)
|
|
continue;
|
|
|
|
LValue base;
|
|
if (isInlineOffset(entry.offset)) {
|
|
setInlineOffsets.set(entry.offset);
|
|
base = object;
|
|
} else
|
|
base = butterfly;
|
|
storeProperty(values[i], base, descriptor.info(), entry.offset);
|
|
break;
|
|
}
|
|
}
|
|
for (unsigned i = structure->inlineCapacity(); i--;) {
|
|
if (!setInlineOffsets.get(i))
|
|
m_out.store64(m_out.int64Zero, m_out.address(m_heaps.properties.atAnyNumber(), object, offsetRelativeToBase(i)));
|
|
}
|
|
|
|
results.append(m_out.anchor(object));
|
|
m_out.jump(outerContinuation);
|
|
}
|
|
|
|
m_out.appendTo(dummyDefault, outerContinuation);
|
|
m_out.unreachable();
|
|
|
|
m_out.appendTo(outerContinuation, outerLastNext);
|
|
setJSValue(m_out.phi(pointerType(), results));
|
|
mutatorFence();
|
|
}
|
|
|
|
void compileMaterializeCreateActivation()
|
|
{
|
|
ObjectMaterializationData& data = m_node->objectMaterializationData();
|
|
|
|
Vector<LValue, 8> values;
|
|
for (unsigned i = 0; i < data.m_properties.size(); ++i)
|
|
values.append(lowJSValue(m_graph.varArgChild(m_node, 2 + i)));
|
|
|
|
LValue scope = lowCell(m_graph.varArgChild(m_node, 1));
|
|
SymbolTable* table = m_node->castOperand<SymbolTable*>();
|
|
RegisteredStructure structure = m_graph.registerStructure(m_graph.globalObjectFor(m_origin.semantic)->activationStructure());
|
|
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(slowPath);
|
|
|
|
LValue fastObject = allocateObject<JSLexicalEnvironment>(
|
|
JSLexicalEnvironment::allocationSize(table), structure, m_out.intPtrZero, slowPath);
|
|
|
|
m_out.storePtr(scope, fastObject, m_heaps.JSScope_next);
|
|
m_out.storePtr(weakPointer(table), fastObject, m_heaps.JSSymbolTableObject_symbolTable);
|
|
|
|
|
|
ValueFromBlock fastResult = m_out.anchor(fastObject);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
// We ensure allocation sinking explictly sets bottom values for all field members.
|
|
// Therefore, it doesn't matter what JSValue we pass in as the initialization value
|
|
// because all fields will be overwritten.
|
|
// FIXME: It may be worth creating an operation that calls a constructor on JSLexicalEnvironment that
|
|
// doesn't initialize every slot because we are guaranteed to do that here.
|
|
VM& vm = this->vm();
|
|
LValue callResult = lazySlowPath(
|
|
[=, &vm] (const Vector<Location>& locations) -> RefPtr<LazySlowPath::Generator> {
|
|
return createLazyCallGenerator(vm,
|
|
operationCreateActivationDirect, locations[0].directGPR(), &vm,
|
|
CCallHelpers::TrustedImmPtr(structure.get()), locations[1].directGPR(),
|
|
CCallHelpers::TrustedImmPtr(table),
|
|
CCallHelpers::TrustedImm64(JSValue::encode(jsUndefined())));
|
|
}, scope);
|
|
ValueFromBlock slowResult = m_out.anchor(callResult);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
LValue activation = m_out.phi(pointerType(), fastResult, slowResult);
|
|
RELEASE_ASSERT(data.m_properties.size() == table->scopeSize());
|
|
for (unsigned i = 0; i < data.m_properties.size(); ++i) {
|
|
PromotedLocationDescriptor descriptor = data.m_properties[i];
|
|
ASSERT(descriptor.kind() == ClosureVarPLoc);
|
|
m_out.store64(
|
|
values[i], activation,
|
|
m_heaps.JSLexicalEnvironment_variables[descriptor.info()]);
|
|
}
|
|
|
|
if (validationEnabled()) {
|
|
// Validate to make sure every slot in the scope has one value.
|
|
ConcurrentJSLocker locker(table->m_lock);
|
|
for (auto iter = table->begin(locker), end = table->end(locker); iter != end; ++iter) {
|
|
bool found = false;
|
|
for (unsigned i = 0; i < data.m_properties.size(); ++i) {
|
|
PromotedLocationDescriptor descriptor = data.m_properties[i];
|
|
ASSERT(descriptor.kind() == ClosureVarPLoc);
|
|
if (iter->value.scopeOffset().offset() == descriptor.info()) {
|
|
found = true;
|
|
break;
|
|
}
|
|
}
|
|
ASSERT_UNUSED(found, found);
|
|
}
|
|
}
|
|
|
|
mutatorFence();
|
|
setJSValue(activation);
|
|
}
|
|
|
|
template<typename JSClass, typename Operation>
|
|
void compileMaterializeNewInternalFieldObjectImpl(Operation operation)
|
|
{
|
|
ObjectMaterializationData& data = m_node->objectMaterializationData();
|
|
|
|
Vector<LValue, JSClass::numberOfInternalFields> values;
|
|
ASSERT(data.m_properties.size() == JSClass::numberOfInternalFields);
|
|
for (unsigned i = 0; i < data.m_properties.size(); ++i)
|
|
values.append(lowJSValue(m_graph.varArgChild(m_node, 1 + i)));
|
|
|
|
RegisteredStructure structure = m_node->structure();
|
|
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(slowPath);
|
|
|
|
RELEASE_ASSERT(data.m_properties.size() == JSClass::numberOfInternalFields);
|
|
LValue fastObject = allocateObject<JSClass>(structure, m_out.intPtrZero, slowPath);
|
|
ValueFromBlock fastResult = m_out.anchor(fastObject);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
VM& vm = this->vm();
|
|
LValue callResult = lazySlowPath(
|
|
[=, &vm] (const Vector<Location>& locations) -> RefPtr<LazySlowPath::Generator> {
|
|
return createLazyCallGenerator(vm,
|
|
operation, locations[0].directGPR(), &vm,
|
|
CCallHelpers::TrustedImmPtr(structure.get()));
|
|
});
|
|
ValueFromBlock slowResult = m_out.anchor(callResult);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
LValue object = m_out.phi(pointerType(), fastResult, slowResult);
|
|
for (unsigned i = 0; i < data.m_properties.size(); ++i) {
|
|
PromotedLocationDescriptor descriptor = data.m_properties[i];
|
|
ASSERT(descriptor.kind() == InternalFieldObjectPLoc);
|
|
RELEASE_ASSERT(descriptor.info() < JSClass::numberOfInternalFields);
|
|
m_out.store64(values[i], object, m_heaps.JSInternalFieldObjectImpl_internalFields[descriptor.info()]);
|
|
}
|
|
|
|
mutatorFence();
|
|
setJSValue(object);
|
|
}
|
|
|
|
void compileMaterializeNewInternalFieldObject()
|
|
{
|
|
switch (m_node->structure()->typeInfo().type()) {
|
|
case JSArrayIteratorType:
|
|
compileMaterializeNewInternalFieldObjectImpl<JSArrayIterator>(operationNewArrayIterator);
|
|
break;
|
|
case JSMapIteratorType:
|
|
compileMaterializeNewInternalFieldObjectImpl<JSMapIterator>(operationNewMapIterator);
|
|
break;
|
|
case JSSetIteratorType:
|
|
compileMaterializeNewInternalFieldObjectImpl<JSSetIterator>(operationNewSetIterator);
|
|
break;
|
|
case JSPromiseType:
|
|
if (m_node->structure()->classInfo() == JSInternalPromise::info())
|
|
compileMaterializeNewInternalFieldObjectImpl<JSInternalPromise>(operationNewInternalPromise);
|
|
else {
|
|
ASSERT(m_node->structure()->classInfo() == JSPromise::info());
|
|
compileMaterializeNewInternalFieldObjectImpl<JSPromise>(operationNewPromise);
|
|
}
|
|
break;
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad structure");
|
|
}
|
|
}
|
|
|
|
void compileCheckTraps()
|
|
{
|
|
ASSERT(Options::usePollingTraps());
|
|
LBasicBlock needTrapHandling = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue state = m_out.load8ZeroExt32(m_out.absolute(vm().needTrapHandlingAddress()));
|
|
m_out.branch(m_out.isZero32(state),
|
|
usually(continuation), rarely(needTrapHandling));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(needTrapHandling, continuation);
|
|
|
|
VM& vm = this->vm();
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
lazySlowPath(
|
|
[=, &vm] (const Vector<Location>&) -> RefPtr<LazySlowPath::Generator> {
|
|
return createLazyCallGenerator(vm, operationHandleTraps, InvalidGPRReg, globalObject);
|
|
});
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
}
|
|
|
|
void compileRegExpExec()
|
|
{
|
|
LValue globalObject = lowCell(m_node->child1());
|
|
|
|
if (m_node->child2().useKind() == RegExpObjectUse) {
|
|
LValue base = lowRegExpObject(m_node->child2());
|
|
|
|
if (m_node->child3().useKind() == StringUse) {
|
|
LValue argument = lowString(m_node->child3());
|
|
LValue result = vmCall(Int64, operationRegExpExecString, globalObject, base, argument);
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
|
|
LValue argument = lowJSValue(m_node->child3());
|
|
LValue result = vmCall(Int64, operationRegExpExec, globalObject, base, argument);
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
|
|
LValue base = lowJSValue(m_node->child2());
|
|
LValue argument = lowJSValue(m_node->child3());
|
|
LValue result = vmCall(Int64, operationRegExpExecGeneric, globalObject, base, argument);
|
|
setJSValue(result);
|
|
}
|
|
|
|
void compileRegExpExecNonGlobalOrSticky()
|
|
{
|
|
LValue globalObject = lowCell(m_node->child1());
|
|
LValue argument = lowString(m_node->child2());
|
|
LValue result = vmCall(Int64, operationRegExpExecNonGlobalOrSticky, globalObject, frozenPointer(m_node->cellOperand()), argument);
|
|
setJSValue(result);
|
|
}
|
|
|
|
void compileRegExpMatchFastGlobal()
|
|
{
|
|
LValue globalObject = lowCell(m_node->child1());
|
|
LValue argument = lowString(m_node->child2());
|
|
LValue result = vmCall(Int64, operationRegExpMatchFastGlobalString, globalObject, frozenPointer(m_node->cellOperand()), argument);
|
|
setJSValue(result);
|
|
}
|
|
|
|
void compileRegExpTest()
|
|
{
|
|
LValue globalObject = lowCell(m_node->child1());
|
|
|
|
if (m_node->child2().useKind() == RegExpObjectUse) {
|
|
LValue base = lowRegExpObject(m_node->child2());
|
|
|
|
if (m_node->child3().useKind() == StringUse) {
|
|
LValue argument = lowString(m_node->child3());
|
|
LValue result = vmCall(Int32, operationRegExpTestString, globalObject, base, argument);
|
|
setBoolean(result);
|
|
return;
|
|
}
|
|
|
|
LValue argument = lowJSValue(m_node->child3());
|
|
LValue result = vmCall(Int32, operationRegExpTest, globalObject, base, argument);
|
|
setBoolean(result);
|
|
return;
|
|
}
|
|
|
|
LValue base = lowJSValue(m_node->child2());
|
|
LValue argument = lowJSValue(m_node->child3());
|
|
LValue result = vmCall(Int32, operationRegExpTestGeneric, globalObject, base, argument);
|
|
setBoolean(result);
|
|
}
|
|
|
|
void compileRegExpMatchFast()
|
|
{
|
|
LValue globalObject = lowCell(m_node->child1());
|
|
LValue base = lowRegExpObject(m_node->child2());
|
|
LValue argument = lowString(m_node->child3());
|
|
LValue result = vmCall(Int64, operationRegExpMatchFastString, globalObject, base, argument);
|
|
setJSValue(result);
|
|
}
|
|
|
|
void compileNewRegexp()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
FrozenValue* regexp = m_node->cellOperand();
|
|
LValue lastIndex = lowJSValue(m_node->child1());
|
|
ASSERT(regexp->cell()->inherits<RegExp>(vm()));
|
|
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(slowCase);
|
|
|
|
auto structure = m_graph.registerStructure(globalObject->regExpStructure());
|
|
LValue fastResultValue = allocateObject<RegExpObject>(structure, m_out.intPtrZero, slowCase);
|
|
m_out.storePtr(frozenPointer(regexp), fastResultValue, m_heaps.RegExpObject_regExpAndLastIndexIsNotWritableFlag);
|
|
m_out.store64(lastIndex, fastResultValue, m_heaps.RegExpObject_lastIndex);
|
|
mutatorFence();
|
|
ValueFromBlock fastResult = m_out.anchor(fastResultValue);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
VM& vm = this->vm();
|
|
RegExp* regexpCell = regexp->cast<RegExp*>();
|
|
LValue slowResultValue = lazySlowPath(
|
|
[=, &vm] (const Vector<Location>& locations) -> RefPtr<LazySlowPath::Generator> {
|
|
return createLazyCallGenerator(vm,
|
|
operationNewRegexpWithLastIndex, locations[0].directGPR(), globalObject,
|
|
CCallHelpers::TrustedImmPtr(regexpCell), locations[1].directGPR());
|
|
}, lastIndex);
|
|
ValueFromBlock slowResult = m_out.anchor(slowResultValue);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(pointerType(), fastResult, slowResult));
|
|
}
|
|
|
|
void compileSetFunctionName()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
vmCall(Void, operationSetFunctionName, weakPointer(globalObject),
|
|
lowCell(m_node->child1()), lowJSValue(m_node->child2()));
|
|
}
|
|
|
|
void compileStringReplace()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
if (m_node->child1().useKind() == StringUse
|
|
&& m_node->child2().useKind() == RegExpObjectUse
|
|
&& m_node->child3().useKind() == StringUse) {
|
|
|
|
if (JSString* replace = m_node->child3()->dynamicCastConstant<JSString*>(vm())) {
|
|
if (!replace->length()) {
|
|
LValue string = lowString(m_node->child1());
|
|
LValue regExp = lowRegExpObject(m_node->child2());
|
|
|
|
LValue result = vmCall(pointerType(), operationStringProtoFuncReplaceRegExpEmptyStr, weakPointer(globalObject), string, regExp);
|
|
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
}
|
|
|
|
LValue string = lowString(m_node->child1());
|
|
LValue regExp = lowRegExpObject(m_node->child2());
|
|
LValue replace = lowString(m_node->child3());
|
|
|
|
LValue result = vmCall(pointerType(), operationStringProtoFuncReplaceRegExpString, weakPointer(globalObject), string, regExp, replace);
|
|
|
|
setJSValue(result);
|
|
return;
|
|
}
|
|
|
|
LValue search;
|
|
if (m_node->child2().useKind() == StringUse)
|
|
search = lowString(m_node->child2());
|
|
else
|
|
search = lowJSValue(m_node->child2());
|
|
|
|
LValue result = vmCall(
|
|
pointerType(), operationStringProtoFuncReplaceGeneric,
|
|
weakPointer(globalObject),
|
|
lowJSValue(m_node->child1()), search,
|
|
lowJSValue(m_node->child3()));
|
|
|
|
setJSValue(result);
|
|
}
|
|
|
|
void compileGetRegExpObjectLastIndex()
|
|
{
|
|
setJSValue(m_out.load64(lowRegExpObject(m_node->child1()), m_heaps.RegExpObject_lastIndex));
|
|
}
|
|
|
|
void compileSetRegExpObjectLastIndex()
|
|
{
|
|
if (!m_node->ignoreLastIndexIsWritable()) {
|
|
LValue regExp = lowRegExpObject(m_node->child1());
|
|
LValue value = lowJSValue(m_node->child2());
|
|
|
|
speculate(
|
|
ExoticObjectMode, noValue(), nullptr,
|
|
m_out.testNonZeroPtr(
|
|
m_out.loadPtr(regExp, m_heaps.RegExpObject_regExpAndLastIndexIsNotWritableFlag),
|
|
m_out.constIntPtr(RegExpObject::lastIndexIsNotWritableFlag)));
|
|
|
|
m_out.store64(value, regExp, m_heaps.RegExpObject_lastIndex);
|
|
return;
|
|
}
|
|
|
|
m_out.store64(lowJSValue(m_node->child2()), lowCell(m_node->child1()), m_heaps.RegExpObject_lastIndex);
|
|
}
|
|
|
|
void compileLogShadowChickenPrologue()
|
|
{
|
|
LValue packet = ensureShadowChickenPacket();
|
|
LValue scope = lowCell(m_node->child1());
|
|
|
|
m_out.storePtr(m_callFrame, packet, m_heaps.ShadowChicken_Packet_frame);
|
|
m_out.storePtr(m_out.loadPtr(addressFor(VirtualRegister(0))), packet, m_heaps.ShadowChicken_Packet_callerFrame);
|
|
m_out.storePtr(m_out.loadPtr(payloadFor(VirtualRegister(CallFrameSlot::callee))), packet, m_heaps.ShadowChicken_Packet_callee);
|
|
m_out.storePtr(scope, packet, m_heaps.ShadowChicken_Packet_scope);
|
|
}
|
|
|
|
void compileLogShadowChickenTail()
|
|
{
|
|
LValue packet = ensureShadowChickenPacket();
|
|
LValue thisValue = lowJSValue(m_node->child1());
|
|
LValue scope = lowCell(m_node->child2());
|
|
CallSiteIndex callSiteIndex = m_ftlState.jitCode->common.codeOrigins->addCodeOrigin(m_origin.semantic);
|
|
|
|
m_out.storePtr(m_callFrame, packet, m_heaps.ShadowChicken_Packet_frame);
|
|
m_out.storePtr(m_out.constIntPtr(bitwise_cast<intptr_t>(ShadowChicken::Packet::tailMarker())), packet, m_heaps.ShadowChicken_Packet_callee);
|
|
m_out.store64(thisValue, packet, m_heaps.ShadowChicken_Packet_thisValue);
|
|
m_out.storePtr(scope, packet, m_heaps.ShadowChicken_Packet_scope);
|
|
// We don't want the CodeBlock to have a weak pointer to itself because
|
|
// that would cause it to always get collected.
|
|
m_out.storePtr(m_out.constIntPtr(bitwise_cast<intptr_t>(codeBlock())), packet, m_heaps.ShadowChicken_Packet_codeBlock);
|
|
m_out.store32(m_out.constInt32(callSiteIndex.bits()), packet, m_heaps.ShadowChicken_Packet_callSiteIndex);
|
|
}
|
|
|
|
void compileRecordRegExpCachedResult()
|
|
{
|
|
Edge globalObjectEdge = m_graph.varArgChild(m_node, 0);
|
|
Edge regExpEdge = m_graph.varArgChild(m_node, 1);
|
|
Edge stringEdge = m_graph.varArgChild(m_node, 2);
|
|
Edge startEdge = m_graph.varArgChild(m_node, 3);
|
|
Edge endEdge = m_graph.varArgChild(m_node, 4);
|
|
|
|
LValue globalObject = lowCell(globalObjectEdge);
|
|
LValue regExp = lowCell(regExpEdge);
|
|
LValue string = lowCell(stringEdge);
|
|
LValue start = lowInt32(startEdge);
|
|
LValue end = lowInt32(endEdge);
|
|
|
|
m_out.storePtr(regExp, globalObject, m_heaps.JSGlobalObject_regExpGlobalData_cachedResult_lastRegExp);
|
|
m_out.storePtr(string, globalObject, m_heaps.JSGlobalObject_regExpGlobalData_cachedResult_lastInput);
|
|
m_out.store32(start, globalObject, m_heaps.JSGlobalObject_regExpGlobalData_cachedResult_result_start);
|
|
m_out.store32(end, globalObject, m_heaps.JSGlobalObject_regExpGlobalData_cachedResult_result_end);
|
|
m_out.store32As8(
|
|
m_out.constInt32(0),
|
|
m_out.address(globalObject, m_heaps.JSGlobalObject_regExpGlobalData_cachedResult_reified));
|
|
}
|
|
|
|
struct ArgumentsLength {
|
|
ArgumentsLength()
|
|
: isKnown(false)
|
|
, known(UINT_MAX)
|
|
, value(nullptr)
|
|
{
|
|
}
|
|
|
|
bool isKnown;
|
|
unsigned known;
|
|
LValue value;
|
|
};
|
|
ArgumentsLength getArgumentsLength(InlineCallFrame* inlineCallFrame)
|
|
{
|
|
ArgumentsLength length;
|
|
|
|
if (inlineCallFrame && !inlineCallFrame->isVarargs()) {
|
|
length.known = static_cast<unsigned>(inlineCallFrame->argumentCountIncludingThis - 1);
|
|
length.isKnown = true;
|
|
length.value = m_out.constInt32(length.known);
|
|
} else {
|
|
length.known = UINT_MAX;
|
|
length.isKnown = false;
|
|
|
|
VirtualRegister argumentCountRegister;
|
|
if (!inlineCallFrame)
|
|
argumentCountRegister = VirtualRegister(CallFrameSlot::argumentCountIncludingThis);
|
|
else
|
|
argumentCountRegister = inlineCallFrame->argumentCountRegister;
|
|
length.value = m_out.sub(m_out.load32(payloadFor(argumentCountRegister)), m_out.int32One);
|
|
}
|
|
|
|
return length;
|
|
}
|
|
|
|
ArgumentsLength getArgumentsLength()
|
|
{
|
|
return getArgumentsLength(m_origin.semantic.inlineCallFrame());
|
|
}
|
|
|
|
LValue getCurrentCallee()
|
|
{
|
|
if (InlineCallFrame* frame = m_origin.semantic.inlineCallFrame()) {
|
|
if (frame->isClosureCall)
|
|
return m_out.loadPtr(addressFor(frame->calleeRecovery.virtualRegister()));
|
|
return weakPointer(frame->calleeRecovery.constant().asCell());
|
|
}
|
|
return m_out.loadPtr(addressFor(VirtualRegister(CallFrameSlot::callee)));
|
|
}
|
|
|
|
LValue getArgumentsStart(InlineCallFrame* inlineCallFrame, unsigned offset = 0)
|
|
{
|
|
VirtualRegister start = AssemblyHelpers::argumentsStart(inlineCallFrame) + offset;
|
|
return addressFor(start).value();
|
|
}
|
|
|
|
LValue getArgumentsStart()
|
|
{
|
|
return getArgumentsStart(m_origin.semantic.inlineCallFrame());
|
|
}
|
|
|
|
template<typename Functor>
|
|
void checkStructure(
|
|
LValue structureDiscriminant, const FormattedValue& formattedValue, ExitKind exitKind,
|
|
const RegisteredStructureSet& set, const Functor& weakStructureDiscriminant)
|
|
{
|
|
if (set.isEmpty()) {
|
|
terminate(exitKind);
|
|
return;
|
|
}
|
|
|
|
if (set.size() == 1) {
|
|
speculate(
|
|
exitKind, formattedValue, nullptr,
|
|
m_out.notEqual(structureDiscriminant, weakStructureDiscriminant(set[0])));
|
|
return;
|
|
}
|
|
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(continuation);
|
|
for (unsigned i = 0; i < set.size() - 1; ++i) {
|
|
LBasicBlock nextStructure = m_out.newBlock();
|
|
m_out.branch(
|
|
m_out.equal(structureDiscriminant, weakStructureDiscriminant(set[i])),
|
|
unsure(continuation), unsure(nextStructure));
|
|
m_out.appendTo(nextStructure);
|
|
}
|
|
|
|
speculate(
|
|
exitKind, formattedValue, nullptr,
|
|
m_out.notEqual(structureDiscriminant, weakStructureDiscriminant(set.last())));
|
|
|
|
m_out.jump(continuation);
|
|
m_out.appendTo(continuation, lastNext);
|
|
}
|
|
|
|
LValue numberOrNotCellNorBigIntToInt32(Edge edge, LValue value)
|
|
{
|
|
LBasicBlock intCase = m_out.newBlock();
|
|
LBasicBlock notIntCase = m_out.newBlock();
|
|
LBasicBlock doubleCase = nullptr;
|
|
LBasicBlock notNumberCase = nullptr;
|
|
if (edge.useKind() == NotCellNorBigIntUse) {
|
|
doubleCase = m_out.newBlock();
|
|
notNumberCase = m_out.newBlock();
|
|
}
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
Vector<ValueFromBlock> results;
|
|
|
|
m_out.branch(isNotInt32(value), unsure(notIntCase), unsure(intCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(intCase, notIntCase);
|
|
results.append(m_out.anchor(unboxInt32(value)));
|
|
m_out.jump(continuation);
|
|
|
|
if (edge.useKind() == NumberUse) {
|
|
m_out.appendTo(notIntCase, continuation);
|
|
FTL_TYPE_CHECK(jsValueValue(value), edge, SpecBytecodeNumber, isCellOrMiscOrBigInt32(value));
|
|
results.append(m_out.anchor(doubleToInt32(unboxDouble(value))));
|
|
m_out.jump(continuation);
|
|
} else {
|
|
m_out.appendTo(notIntCase, doubleCase);
|
|
m_out.branch(isCellOrMiscOrBigInt32(value, provenType(edge)), unsure(notNumberCase), unsure(doubleCase));
|
|
|
|
m_out.appendTo(doubleCase, notNumberCase);
|
|
results.append(m_out.anchor(doubleToInt32(unboxDouble(value))));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(notNumberCase, continuation);
|
|
|
|
FTL_TYPE_CHECK(jsValueValue(value), edge, ~SpecCellCheck, isCell(value));
|
|
#if USE(BIGINT32)
|
|
FTL_TYPE_CHECK(jsValueValue(value), edge, ~SpecCellCheck & ~SpecBigInt, isBigInt32(value));
|
|
#endif
|
|
|
|
LValue specialResult = m_out.select(
|
|
m_out.equal(value, m_out.constInt64(JSValue::encode(jsBoolean(true)))),
|
|
m_out.int32One, m_out.int32Zero);
|
|
results.append(m_out.anchor(specialResult));
|
|
m_out.jump(continuation);
|
|
}
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
return m_out.phi(Int32, results);
|
|
}
|
|
|
|
LValue loadProperty(LValue storage, unsigned identifierNumber, PropertyOffset offset)
|
|
{
|
|
return m_out.load64(addressOfProperty(storage, identifierNumber, offset));
|
|
}
|
|
|
|
void storeProperty(
|
|
LValue value, LValue storage, unsigned identifierNumber, PropertyOffset offset)
|
|
{
|
|
m_out.store64(value, addressOfProperty(storage, identifierNumber, offset));
|
|
}
|
|
|
|
TypedPointer addressOfProperty(
|
|
LValue storage, unsigned identifierNumber, PropertyOffset offset)
|
|
{
|
|
return m_out.address(
|
|
m_heaps.properties[identifierNumber], storage, offsetRelativeToBase(offset));
|
|
}
|
|
|
|
LValue storageForTransition(
|
|
LValue object, PropertyOffset offset,
|
|
Structure* previousStructure, Structure* nextStructure)
|
|
{
|
|
if (isInlineOffset(offset))
|
|
return object;
|
|
|
|
if (previousStructure->outOfLineCapacity() == nextStructure->outOfLineCapacity())
|
|
return m_out.loadPtr(object, m_heaps.JSObject_butterfly);
|
|
|
|
LValue result;
|
|
if (!previousStructure->outOfLineCapacity())
|
|
result = allocatePropertyStorage(object, previousStructure);
|
|
else {
|
|
result = reallocatePropertyStorage(
|
|
object, m_out.loadPtr(object, m_heaps.JSObject_butterfly),
|
|
previousStructure, nextStructure);
|
|
}
|
|
|
|
nukeStructureAndSetButterfly(result, object);
|
|
return result;
|
|
}
|
|
|
|
void initializeArrayElements(LValue indexingType, LValue begin, LValue end, LValue butterfly)
|
|
{
|
|
|
|
if (begin == end)
|
|
return;
|
|
|
|
if (indexingType->hasInt32()) {
|
|
IndexingType rawIndexingType = static_cast<IndexingType>(indexingType->asInt32());
|
|
if (hasUndecided(rawIndexingType))
|
|
return;
|
|
IndexedAbstractHeap* heap = m_heaps.forIndexingType(rawIndexingType);
|
|
DFG_ASSERT(m_graph, m_node, heap);
|
|
|
|
LValue hole;
|
|
if (hasDouble(rawIndexingType))
|
|
hole = m_out.constInt64(bitwise_cast<int64_t>(PNaN));
|
|
else
|
|
hole = m_out.constInt64(JSValue::encode(JSValue()));
|
|
|
|
splatWords(butterfly, begin, end, hole, heap->atAnyIndex());
|
|
} else {
|
|
LValue hole = m_out.select(
|
|
m_out.equal(m_out.bitAnd(indexingType, m_out.constInt32(IndexingShapeMask)), m_out.constInt32(DoubleShape)),
|
|
m_out.constInt64(bitwise_cast<int64_t>(PNaN)),
|
|
m_out.constInt64(JSValue::encode(JSValue())));
|
|
splatWords(butterfly, begin, end, hole, m_heaps.root);
|
|
}
|
|
}
|
|
|
|
void splatWords(LValue base, LValue begin, LValue end, LValue value, const AbstractHeap& heap)
|
|
{
|
|
const uint64_t unrollingLimit = 10;
|
|
if (begin->hasInt() && end->hasInt()) {
|
|
uint64_t beginConst = static_cast<uint64_t>(begin->asInt());
|
|
uint64_t endConst = static_cast<uint64_t>(end->asInt());
|
|
|
|
if (endConst - beginConst <= unrollingLimit) {
|
|
for (uint64_t i = beginConst; i < endConst; ++i) {
|
|
LValue pointer = m_out.add(base, m_out.constIntPtr(i * sizeof(uint64_t)));
|
|
m_out.store64(value, TypedPointer(heap, pointer));
|
|
}
|
|
return;
|
|
}
|
|
}
|
|
|
|
LBasicBlock initLoop = m_out.newBlock();
|
|
LBasicBlock initDone = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(initLoop);
|
|
|
|
ValueFromBlock originalIndex = m_out.anchor(end);
|
|
ValueFromBlock originalPointer = m_out.anchor(
|
|
m_out.add(base, m_out.shl(m_out.signExt32ToPtr(begin), m_out.constInt32(3))));
|
|
m_out.branch(m_out.notEqual(end, begin), unsure(initLoop), unsure(initDone));
|
|
|
|
m_out.appendTo(initLoop, initDone);
|
|
LValue index = m_out.phi(Int32, originalIndex);
|
|
LValue pointer = m_out.phi(pointerType(), originalPointer);
|
|
|
|
m_out.store64(value, TypedPointer(heap, pointer));
|
|
|
|
LValue nextIndex = m_out.sub(index, m_out.int32One);
|
|
m_out.addIncomingToPhi(index, m_out.anchor(nextIndex));
|
|
m_out.addIncomingToPhi(pointer, m_out.anchor(m_out.add(pointer, m_out.intPtrEight)));
|
|
m_out.branch(
|
|
m_out.notEqual(nextIndex, begin), unsure(initLoop), unsure(initDone));
|
|
|
|
m_out.appendTo(initDone, lastNext);
|
|
}
|
|
|
|
LValue allocatePropertyStorage(LValue object, Structure* previousStructure)
|
|
{
|
|
if (previousStructure->couldHaveIndexingHeader()) {
|
|
return vmCall(
|
|
pointerType(),
|
|
operationAllocateComplexPropertyStorageWithInitialCapacity,
|
|
m_vmValue, object);
|
|
}
|
|
|
|
LValue result = allocatePropertyStorageWithSizeImpl(initialOutOfLineCapacity);
|
|
|
|
splatWords(
|
|
result,
|
|
m_out.constInt32(-initialOutOfLineCapacity - 1), m_out.constInt32(-1),
|
|
m_out.int64Zero, m_heaps.properties.atAnyNumber());
|
|
|
|
return result;
|
|
}
|
|
|
|
LValue reallocatePropertyStorage(
|
|
LValue object, LValue oldStorage, Structure* previous, Structure* next)
|
|
{
|
|
size_t oldSize = previous->outOfLineCapacity();
|
|
size_t newSize = oldSize * outOfLineGrowthFactor;
|
|
|
|
ASSERT_UNUSED(next, newSize == next->outOfLineCapacity());
|
|
|
|
if (previous->couldHaveIndexingHeader()) {
|
|
LValue newAllocSize = m_out.constIntPtr(newSize);
|
|
return vmCall(pointerType(), operationAllocateComplexPropertyStorage, m_vmValue, object, newAllocSize);
|
|
}
|
|
|
|
LValue result = allocatePropertyStorageWithSizeImpl(newSize);
|
|
|
|
ptrdiff_t headerSize = -sizeof(IndexingHeader) - sizeof(void*);
|
|
ptrdiff_t endStorage = headerSize - static_cast<ptrdiff_t>(oldSize * sizeof(JSValue));
|
|
|
|
for (ptrdiff_t offset = headerSize; offset > endStorage; offset -= sizeof(void*)) {
|
|
LValue loaded =
|
|
m_out.loadPtr(m_out.address(m_heaps.properties.atAnyNumber(), oldStorage, offset));
|
|
m_out.storePtr(loaded, m_out.address(m_heaps.properties.atAnyNumber(), result, offset));
|
|
}
|
|
|
|
splatWords(
|
|
result,
|
|
m_out.constInt32(-newSize - 1), m_out.constInt32(-oldSize - 1),
|
|
m_out.int64Zero, m_heaps.properties.atAnyNumber());
|
|
|
|
return result;
|
|
}
|
|
|
|
LValue allocatePropertyStorageWithSizeImpl(size_t sizeInValues)
|
|
{
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(slowPath);
|
|
|
|
size_t sizeInBytes = sizeInValues * sizeof(JSValue);
|
|
Allocator allocator = vm().jsValueGigacageAuxiliarySpace.allocatorForNonVirtual(sizeInBytes, AllocatorForMode::AllocatorIfExists);
|
|
LValue startOfStorage = allocateHeapCell(
|
|
m_out.constIntPtr(allocator.localAllocator()), slowPath);
|
|
ValueFromBlock fastButterfly = m_out.anchor(
|
|
m_out.add(m_out.constIntPtr(sizeInBytes + sizeof(IndexingHeader)), startOfStorage));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
|
|
LValue slowButterflyValue;
|
|
VM& vm = this->vm();
|
|
if (sizeInValues == initialOutOfLineCapacity) {
|
|
slowButterflyValue = lazySlowPath(
|
|
[=, &vm] (const Vector<Location>& locations) -> RefPtr<LazySlowPath::Generator> {
|
|
return createLazyCallGenerator(vm,
|
|
operationAllocateSimplePropertyStorageWithInitialCapacity,
|
|
locations[0].directGPR(), &vm);
|
|
});
|
|
} else {
|
|
slowButterflyValue = lazySlowPath(
|
|
[=, &vm] (const Vector<Location>& locations) -> RefPtr<LazySlowPath::Generator> {
|
|
return createLazyCallGenerator(vm,
|
|
operationAllocateSimplePropertyStorage, locations[0].directGPR(), &vm,
|
|
CCallHelpers::TrustedImmPtr(sizeInValues));
|
|
});
|
|
}
|
|
ValueFromBlock slowButterfly = m_out.anchor(slowButterflyValue);
|
|
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
|
|
return m_out.phi(pointerType(), fastButterfly, slowButterfly);
|
|
}
|
|
|
|
LValue getById(LValue base, AccessType type)
|
|
{
|
|
Node* node = m_node;
|
|
CacheableIdentifier identifier = node->cacheableIdentifier();
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int64);
|
|
patchpoint->appendSomeRegister(base);
|
|
patchpoint->append(m_notCellMask, ValueRep::lateReg(GPRInfo::notCellMaskRegister));
|
|
patchpoint->append(m_numberTag, ValueRep::lateReg(GPRInfo::numberTagRegister));
|
|
|
|
// FIXME: If this is a GetByIdFlush/GetByIdDirectFlush, we might get some performance boost if we claim that it
|
|
// clobbers volatile registers late. It's not necessary for correctness, though, since the
|
|
// IC code is super smart about saving registers.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=152848
|
|
|
|
patchpoint->clobber(RegisterSet::macroScratchRegisters());
|
|
|
|
RefPtr<PatchpointExceptionHandle> exceptionHandle =
|
|
preparePatchpointForExceptions(patchpoint);
|
|
|
|
State* state = &m_ftlState;
|
|
patchpoint->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
CallSiteIndex callSiteIndex =
|
|
state->jitCode->common.codeOrigins->addUniqueCallSiteIndex(node->origin.semantic);
|
|
|
|
// This is the direct exit target for operation calls.
|
|
Box<CCallHelpers::JumpList> exceptions =
|
|
exceptionHandle->scheduleExitCreation(params)->jumps(jit);
|
|
|
|
// This is the exit for call IC's created by the getById for getters. We don't have
|
|
// to do anything weird other than call this, since it will associate the exit with
|
|
// the callsite index.
|
|
exceptionHandle->scheduleExitCreationForUnwind(params, callSiteIndex);
|
|
|
|
auto generator = Box<JITGetByIdGenerator>::create(
|
|
jit.codeBlock(), node->origin.semantic, callSiteIndex,
|
|
params.unavailableRegisters(), identifier, JSValueRegs(params[1].gpr()),
|
|
JSValueRegs(params[0].gpr()), type);
|
|
|
|
generator->generateFastPath(jit);
|
|
CCallHelpers::Label done = jit.label();
|
|
|
|
params.addLatePath(
|
|
[=] (CCallHelpers& jit) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
auto optimizationFunction = appropriateOptimizingGetByIdFunction(type);
|
|
|
|
generator->slowPathJump().link(&jit);
|
|
CCallHelpers::Label slowPathBegin = jit.label();
|
|
CCallHelpers::Call slowPathCall = callOperation(
|
|
*state, params.unavailableRegisters(), jit, node->origin.semantic,
|
|
exceptions.get(), optimizationFunction, params[0].gpr(),
|
|
jit.codeBlock()->globalObjectFor(node->origin.semantic),
|
|
CCallHelpers::TrustedImmPtr(generator->stubInfo()), params[1].gpr(),
|
|
CCallHelpers::TrustedImmPtr(identifier.rawBits())).call();
|
|
jit.jump().linkTo(done, &jit);
|
|
|
|
generator->reportSlowPathCall(slowPathBegin, slowPathCall);
|
|
|
|
jit.addLinkTask(
|
|
[=] (LinkBuffer& linkBuffer) {
|
|
generator->finalize(linkBuffer, linkBuffer);
|
|
});
|
|
});
|
|
});
|
|
|
|
return patchpoint;
|
|
}
|
|
|
|
LValue getByIdWithThis(LValue base, LValue thisValue)
|
|
{
|
|
Node* node = m_node;
|
|
CacheableIdentifier identifier = node->cacheableIdentifier();
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int64);
|
|
patchpoint->appendSomeRegister(base);
|
|
patchpoint->appendSomeRegister(thisValue);
|
|
patchpoint->append(m_notCellMask, ValueRep::lateReg(GPRInfo::notCellMaskRegister));
|
|
patchpoint->append(m_numberTag, ValueRep::lateReg(GPRInfo::numberTagRegister));
|
|
|
|
patchpoint->clobber(RegisterSet::macroScratchRegisters());
|
|
|
|
RefPtr<PatchpointExceptionHandle> exceptionHandle =
|
|
preparePatchpointForExceptions(patchpoint);
|
|
|
|
State* state = &m_ftlState;
|
|
patchpoint->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
CallSiteIndex callSiteIndex =
|
|
state->jitCode->common.codeOrigins->addUniqueCallSiteIndex(node->origin.semantic);
|
|
|
|
// This is the direct exit target for operation calls.
|
|
Box<CCallHelpers::JumpList> exceptions =
|
|
exceptionHandle->scheduleExitCreation(params)->jumps(jit);
|
|
|
|
// This is the exit for call IC's created by the getById for getters. We don't have
|
|
// to do anything weird other than call this, since it will associate the exit with
|
|
// the callsite index.
|
|
exceptionHandle->scheduleExitCreationForUnwind(params, callSiteIndex);
|
|
|
|
auto generator = Box<JITGetByIdWithThisGenerator>::create(
|
|
jit.codeBlock(), node->origin.semantic, callSiteIndex,
|
|
params.unavailableRegisters(), identifier, JSValueRegs(params[0].gpr()),
|
|
JSValueRegs(params[1].gpr()), JSValueRegs(params[2].gpr()));
|
|
|
|
generator->generateFastPath(jit);
|
|
CCallHelpers::Label done = jit.label();
|
|
|
|
params.addLatePath(
|
|
[=] (CCallHelpers& jit) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
auto optimizationFunction = operationGetByIdWithThisOptimize;
|
|
|
|
generator->slowPathJump().link(&jit);
|
|
CCallHelpers::Label slowPathBegin = jit.label();
|
|
CCallHelpers::Call slowPathCall = callOperation(
|
|
*state, params.unavailableRegisters(), jit, node->origin.semantic,
|
|
exceptions.get(), optimizationFunction, params[0].gpr(),
|
|
jit.codeBlock()->globalObjectFor(node->origin.semantic),
|
|
CCallHelpers::TrustedImmPtr(generator->stubInfo()), params[1].gpr(),
|
|
params[2].gpr(), CCallHelpers::TrustedImmPtr(identifier.rawBits())).call();
|
|
jit.jump().linkTo(done, &jit);
|
|
|
|
generator->reportSlowPathCall(slowPathBegin, slowPathCall);
|
|
|
|
jit.addLinkTask(
|
|
[=] (LinkBuffer& linkBuffer) {
|
|
generator->finalize(linkBuffer, linkBuffer);
|
|
});
|
|
});
|
|
});
|
|
|
|
return patchpoint;
|
|
}
|
|
|
|
LValue isFastTypedArray(LValue object)
|
|
{
|
|
return m_out.equal(
|
|
m_out.load32(object, m_heaps.JSArrayBufferView_mode),
|
|
m_out.constInt32(FastTypedArray));
|
|
}
|
|
|
|
TypedPointer baseIndex(IndexedAbstractHeap& heap, LValue storage, LValue index, Edge edge, ptrdiff_t offset = 0)
|
|
{
|
|
return m_out.baseIndex(
|
|
heap, storage, m_out.zeroExtPtr(index), provenValue(edge), offset);
|
|
}
|
|
|
|
template<typename IntFunctor, typename DoubleFunctor>
|
|
void compare(
|
|
const IntFunctor& intFunctor, const DoubleFunctor& doubleFunctor,
|
|
C_JITOperation_TT stringIdentFunction,
|
|
C_JITOperation_B_GJssJss stringFunction,
|
|
S_JITOperation_GJJ fallbackFunction)
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
if (m_node->isBinaryUseKind(Int32Use)) {
|
|
LValue left = lowInt32(m_node->child1());
|
|
LValue right = lowInt32(m_node->child2());
|
|
setBoolean(intFunctor(left, right));
|
|
return;
|
|
}
|
|
|
|
if (m_node->isBinaryUseKind(Int52RepUse)) {
|
|
Int52Kind kind;
|
|
LValue left = lowWhicheverInt52(m_node->child1(), kind);
|
|
LValue right = lowInt52(m_node->child2(), kind);
|
|
setBoolean(intFunctor(left, right));
|
|
return;
|
|
}
|
|
|
|
if (m_node->isBinaryUseKind(DoubleRepUse)) {
|
|
LValue left = lowDouble(m_node->child1());
|
|
LValue right = lowDouble(m_node->child2());
|
|
setBoolean(doubleFunctor(left, right));
|
|
return;
|
|
}
|
|
|
|
#if USE(BIGINT32)
|
|
if (m_node->isBinaryUseKind(BigInt32Use)) {
|
|
LValue left = lowBigInt32(m_node->child1());
|
|
LValue right = lowBigInt32(m_node->child2());
|
|
setBoolean(intFunctor(unboxBigInt32(left), unboxBigInt32(right)));
|
|
return;
|
|
}
|
|
#endif
|
|
|
|
if (m_node->isBinaryUseKind(StringIdentUse)) {
|
|
LValue left = lowStringIdent(m_node->child1());
|
|
LValue right = lowStringIdent(m_node->child2());
|
|
setBoolean(m_out.callWithoutSideEffects(Int32, stringIdentFunction, left, right));
|
|
return;
|
|
}
|
|
|
|
if (m_node->isBinaryUseKind(StringUse)) {
|
|
LValue left = lowCell(m_node->child1());
|
|
LValue right = lowCell(m_node->child2());
|
|
speculateString(m_node->child1(), left);
|
|
speculateString(m_node->child2(), right);
|
|
|
|
LValue result = vmCall(
|
|
Int32, stringFunction,
|
|
weakPointer(globalObject), left, right);
|
|
setBoolean(result);
|
|
return;
|
|
}
|
|
|
|
DFG_ASSERT(m_graph, m_node, m_node->isBinaryUseKind(UntypedUse) || m_node->isBinaryUseKind(HeapBigIntUse) || m_node->isBinaryUseKind(AnyBigIntUse), m_node->child1().useKind(), m_node->child2().useKind());
|
|
genericJSValueCompare(intFunctor, fallbackFunction);
|
|
}
|
|
|
|
void compileStringSlice()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LBasicBlock lengthCheckCase = m_out.newBlock();
|
|
LBasicBlock emptyCase = m_out.newBlock();
|
|
LBasicBlock notEmptyCase = m_out.newBlock();
|
|
LBasicBlock oneCharCase = m_out.newBlock();
|
|
LBasicBlock is8Bit = m_out.newBlock();
|
|
LBasicBlock is16Bit = m_out.newBlock();
|
|
LBasicBlock bitsContinuation = m_out.newBlock();
|
|
LBasicBlock bigCharacter = m_out.newBlock();
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock ropeSlowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue string = lowString(m_node->child1());
|
|
LValue start = lowInt32(m_node->child2());
|
|
LValue end = nullptr;
|
|
if (m_node->child3())
|
|
end = lowInt32(m_node->child3());
|
|
else
|
|
end = m_out.constInt32(std::numeric_limits<int32_t>::max());
|
|
m_out.branch(isRopeString(string, m_node->child1()), rarely(ropeSlowCase), usually(lengthCheckCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(lengthCheckCase, emptyCase);
|
|
LValue stringImpl = m_out.loadPtr(string, m_heaps.JSString_value);
|
|
LValue length = m_out.load32NonNegative(stringImpl, m_heaps.StringImpl_length);
|
|
auto range = populateSliceRange(start, end, length);
|
|
LValue from = range.first;
|
|
LValue to = range.second;
|
|
LValue span = m_out.sub(to, from);
|
|
m_out.branch(m_out.lessThanOrEqual(span, m_out.int32Zero), unsure(emptyCase), unsure(notEmptyCase));
|
|
|
|
Vector<ValueFromBlock, 5> results;
|
|
|
|
m_out.appendTo(emptyCase, notEmptyCase);
|
|
results.append(m_out.anchor(weakPointer(jsEmptyString(vm()))));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(notEmptyCase, oneCharCase);
|
|
m_out.branch(m_out.equal(span, m_out.int32One), unsure(oneCharCase), unsure(slowCase));
|
|
|
|
m_out.appendTo(oneCharCase, is8Bit);
|
|
LValue storage = m_out.loadPtr(stringImpl, m_heaps.StringImpl_data);
|
|
m_out.branch(
|
|
m_out.testIsZero32(
|
|
m_out.load32(stringImpl, m_heaps.StringImpl_hashAndFlags),
|
|
m_out.constInt32(StringImpl::flagIs8Bit())),
|
|
unsure(is16Bit), unsure(is8Bit));
|
|
|
|
m_out.appendTo(is8Bit, is16Bit);
|
|
ValueFromBlock char8Bit = m_out.anchor(m_out.load8ZeroExt32(m_out.baseIndex(m_heaps.characters8, storage, m_out.zeroExtPtr(from))));
|
|
m_out.jump(bitsContinuation);
|
|
|
|
m_out.appendTo(is16Bit, bigCharacter);
|
|
LValue char16BitValue = m_out.load16ZeroExt32(m_out.baseIndex(m_heaps.characters16, storage, m_out.zeroExtPtr(from)));
|
|
ValueFromBlock char16Bit = m_out.anchor(char16BitValue);
|
|
m_out.branch(
|
|
m_out.above(char16BitValue, m_out.constInt32(maxSingleCharacterString)),
|
|
rarely(bigCharacter), usually(bitsContinuation));
|
|
|
|
m_out.appendTo(bigCharacter, bitsContinuation);
|
|
results.append(m_out.anchor(vmCall(
|
|
Int64, operationSingleCharacterString,
|
|
m_vmValue, char16BitValue)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(bitsContinuation, slowCase);
|
|
LValue character = m_out.phi(Int32, char8Bit, char16Bit);
|
|
LValue smallStrings = m_out.constIntPtr(vm().smallStrings.singleCharacterStrings());
|
|
results.append(m_out.anchor(m_out.loadPtr(m_out.baseIndex(
|
|
m_heaps.singleCharacterStrings, smallStrings, m_out.zeroExtPtr(character)))));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowCase, ropeSlowCase);
|
|
results.append(m_out.anchor(vmCall(pointerType(), operationStringSubstr, weakPointer(globalObject), string, from, span)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(ropeSlowCase, continuation);
|
|
results.append(m_out.anchor(vmCall(pointerType(), operationStringSlice, weakPointer(globalObject), string, start, end)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(pointerType(), results));
|
|
}
|
|
|
|
void compileToLowerCase()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LBasicBlock notRope = m_out.newBlock();
|
|
LBasicBlock is8Bit = m_out.newBlock();
|
|
LBasicBlock loopTop = m_out.newBlock();
|
|
LBasicBlock loopBody = m_out.newBlock();
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue string = lowString(m_node->child1());
|
|
ValueFromBlock startIndex = m_out.anchor(m_out.constInt32(0));
|
|
ValueFromBlock startIndexForCall = m_out.anchor(m_out.constInt32(0));
|
|
m_out.branch(isRopeString(string, m_node->child1()),
|
|
unsure(slowPath), unsure(notRope));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(notRope, is8Bit);
|
|
LValue impl = m_out.loadPtr(string, m_heaps.JSString_value);
|
|
m_out.branch(
|
|
m_out.testIsZero32(
|
|
m_out.load32(impl, m_heaps.StringImpl_hashAndFlags),
|
|
m_out.constInt32(StringImpl::flagIs8Bit())),
|
|
unsure(slowPath), unsure(is8Bit));
|
|
|
|
m_out.appendTo(is8Bit, loopTop);
|
|
LValue length = m_out.load32(impl, m_heaps.StringImpl_length);
|
|
LValue buffer = m_out.loadPtr(impl, m_heaps.StringImpl_data);
|
|
ValueFromBlock fastResult = m_out.anchor(string);
|
|
m_out.jump(loopTop);
|
|
|
|
m_out.appendTo(loopTop, loopBody);
|
|
LValue index = m_out.phi(Int32, startIndex);
|
|
ValueFromBlock indexFromBlock = m_out.anchor(index);
|
|
m_out.branch(m_out.below(index, length),
|
|
unsure(loopBody), unsure(continuation));
|
|
|
|
m_out.appendTo(loopBody, slowPath);
|
|
|
|
// FIXME: Strings needs to be caged.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=174924
|
|
LValue byte = m_out.load8ZeroExt32(m_out.baseIndex(m_heaps.characters8, buffer, m_out.zeroExtPtr(index)));
|
|
LValue isInvalidAsciiRange = m_out.bitAnd(byte, m_out.constInt32(~0x7F));
|
|
LValue isUpperCase = m_out.belowOrEqual(m_out.sub(byte, m_out.constInt32('A')), m_out.constInt32('Z' - 'A'));
|
|
LValue isBadCharacter = m_out.bitOr(isInvalidAsciiRange, isUpperCase);
|
|
m_out.addIncomingToPhi(index, m_out.anchor(m_out.add(index, m_out.int32One)));
|
|
m_out.branch(isBadCharacter, unsure(slowPath), unsure(loopTop));
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
LValue slowPathIndex = m_out.phi(Int32, startIndexForCall, indexFromBlock);
|
|
ValueFromBlock slowResult = m_out.anchor(vmCall(pointerType(), operationToLowerCase, weakPointer(globalObject), string, slowPathIndex));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(pointerType(), fastResult, slowResult));
|
|
}
|
|
|
|
void compileNumberToStringWithRadix()
|
|
{
|
|
bool validRadixIsGuaranteed = false;
|
|
if (m_node->child2()->isInt32Constant()) {
|
|
int32_t radix = m_node->child2()->asInt32();
|
|
if (radix >= 2 && radix <= 36)
|
|
validRadixIsGuaranteed = true;
|
|
}
|
|
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
switch (m_node->child1().useKind()) {
|
|
case Int32Use:
|
|
setJSValue(vmCall(pointerType(), validRadixIsGuaranteed ? operationInt32ToStringWithValidRadix : operationInt32ToString, weakPointer(globalObject), lowInt32(m_node->child1()), lowInt32(m_node->child2())));
|
|
break;
|
|
case Int52RepUse:
|
|
setJSValue(vmCall(pointerType(), validRadixIsGuaranteed ? operationInt52ToStringWithValidRadix : operationInt52ToString, weakPointer(globalObject), lowStrictInt52(m_node->child1()), lowInt32(m_node->child2())));
|
|
break;
|
|
case DoubleRepUse:
|
|
setJSValue(vmCall(pointerType(), validRadixIsGuaranteed ? operationDoubleToStringWithValidRadix : operationDoubleToString, weakPointer(globalObject), lowDouble(m_node->child1()), lowInt32(m_node->child2())));
|
|
break;
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
}
|
|
}
|
|
|
|
void compileNumberToStringWithValidRadixConstant()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
switch (m_node->child1().useKind()) {
|
|
case Int32Use:
|
|
setJSValue(vmCall(pointerType(), operationInt32ToStringWithValidRadix, weakPointer(globalObject), lowInt32(m_node->child1()), m_out.constInt32(m_node->validRadixConstant())));
|
|
break;
|
|
case Int52RepUse:
|
|
setJSValue(vmCall(pointerType(), operationInt52ToStringWithValidRadix, weakPointer(globalObject), lowStrictInt52(m_node->child1()), m_out.constInt32(m_node->validRadixConstant())));
|
|
break;
|
|
case DoubleRepUse:
|
|
setJSValue(vmCall(pointerType(), operationDoubleToStringWithValidRadix, weakPointer(globalObject), lowDouble(m_node->child1()), m_out.constInt32(m_node->validRadixConstant())));
|
|
break;
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
}
|
|
}
|
|
|
|
void compileResolveScopeForHoistingFuncDeclInEval()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
UniquedStringImpl* uid = m_graph.identifiers()[m_node->identifierNumber()];
|
|
setJSValue(vmCall(pointerType(), operationResolveScopeForHoistingFuncDeclInEval, weakPointer(globalObject), lowCell(m_node->child1()), m_out.constIntPtr(uid)));
|
|
}
|
|
|
|
void compileResolveScope()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
UniquedStringImpl* uid = m_graph.identifiers()[m_node->identifierNumber()];
|
|
setJSValue(vmCall(pointerType(), operationResolveScope,
|
|
weakPointer(globalObject), lowCell(m_node->child1()), m_out.constIntPtr(uid)));
|
|
}
|
|
|
|
void compileGetDynamicVar()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
UniquedStringImpl* uid = m_graph.identifiers()[m_node->identifierNumber()];
|
|
setJSValue(vmCall(Int64, operationGetDynamicVar,
|
|
weakPointer(globalObject), lowCell(m_node->child1()), m_out.constIntPtr(uid), m_out.constInt32(m_node->getPutInfo())));
|
|
}
|
|
|
|
void compilePutDynamicVar()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
UniquedStringImpl* uid = m_graph.identifiers()[m_node->identifierNumber()];
|
|
setJSValue(vmCall(Void, m_node->ecmaMode().isStrict() ? operationPutDynamicVarStrict : operationPutDynamicVarNonStrict,
|
|
weakPointer(globalObject), lowCell(m_node->child1()), lowJSValue(m_node->child2()), m_out.constIntPtr(uid), m_out.constInt32(m_node->getPutInfo())));
|
|
}
|
|
|
|
void compileUnreachable()
|
|
{
|
|
// It's so tempting to assert that AI has proved that this is unreachable. But that's
|
|
// simply not a requirement of the Unreachable opcode at all. If you emit an opcode that
|
|
// *you* know will not return, then it's fine to end the basic block with Unreachable
|
|
// after that opcode. You don't have to also prove to AI that your opcode does not return.
|
|
// Hence, there is nothing to do here but emit code that will crash, so that we catch
|
|
// cases where you said Unreachable but you lied.
|
|
//
|
|
// It's also also worth noting that some clients emit this opcode because they're not 100% sure
|
|
// if the code is unreachable, but they would really prefer if we crashed rather than kept going
|
|
// if it did turn out to be reachable. Hence, this needs to deterministically crash.
|
|
|
|
crash();
|
|
}
|
|
|
|
void compileCheckJSCast()
|
|
{
|
|
DFG_ASSERT(m_graph, m_node, m_node->op() == CheckJSCast || m_node->op() == CheckNotJSCast);
|
|
LValue cell = lowCell(m_node->child1());
|
|
|
|
const ClassInfo* classInfo = m_node->classInfo();
|
|
|
|
if (classInfo->inheritsJSTypeRange) {
|
|
LValue hasType = isCellWithType(cell, classInfo->inheritsJSTypeRange.value(), speculationFromClassInfoInheritance(classInfo));
|
|
LValue condition = nullptr;
|
|
if (m_node->op() == CheckJSCast)
|
|
condition = m_out.bitNot(hasType);
|
|
else
|
|
condition = hasType;
|
|
speculate(BadType, jsValueValue(cell), m_node->child1().node(), condition);
|
|
return;
|
|
}
|
|
|
|
if (!classInfo->checkSubClassSnippet) {
|
|
LBasicBlock loop = m_out.newBlock();
|
|
LBasicBlock parentClass = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue structure = loadStructure(cell);
|
|
LValue classInfo = m_out.loadPtr(structure, m_heaps.Structure_classInfo);
|
|
ValueFromBlock otherAtStart = m_out.anchor(classInfo);
|
|
m_out.jump(loop);
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(loop, parentClass);
|
|
LValue other = m_out.phi(pointerType(), otherAtStart);
|
|
LValue foundCondition = m_out.equal(other, m_out.constIntPtr(classInfo));
|
|
if (m_node->op() == CheckNotJSCast) {
|
|
speculate(BadType, jsValueValue(cell), m_node->child1().node(), foundCondition);
|
|
m_out.jump(parentClass);
|
|
} else
|
|
m_out.branch(foundCondition, unsure(continuation), unsure(parentClass));
|
|
|
|
m_out.appendTo(parentClass, continuation);
|
|
LValue parent = m_out.loadPtr(other, m_heaps.ClassInfo_parentClass);
|
|
LValue parentIsNull = m_out.isNull(parent);
|
|
m_out.addIncomingToPhi(other, m_out.anchor(parent));
|
|
if (m_node->op() == CheckNotJSCast)
|
|
m_out.branch(parentIsNull, unsure(continuation), unsure(loop));
|
|
else {
|
|
speculate(BadType, jsValueValue(cell), m_node->child1().node(), parentIsNull);
|
|
m_out.jump(loop);
|
|
}
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
return;
|
|
}
|
|
|
|
RefPtr<Snippet> domJIT = classInfo->checkSubClassSnippet();
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Void);
|
|
patchpoint->appendSomeRegister(cell);
|
|
patchpoint->append(m_notCellMask, ValueRep::reg(GPRInfo::notCellMaskRegister));
|
|
patchpoint->append(m_numberTag, ValueRep::reg(GPRInfo::numberTagRegister));
|
|
|
|
NodeOrigin origin = m_origin;
|
|
unsigned osrExitArgumentOffset = patchpoint->numChildren();
|
|
OSRExitDescriptor* exitDescriptor = appendOSRExitDescriptor(jsValueValue(cell), m_node->child1().node());
|
|
patchpoint->appendColdAnys(buildExitArguments(exitDescriptor, origin.forExit, jsValueValue(cell)));
|
|
|
|
patchpoint->numGPScratchRegisters = domJIT->numGPScratchRegisters;
|
|
patchpoint->numFPScratchRegisters = domJIT->numFPScratchRegisters;
|
|
patchpoint->clobber(RegisterSet::macroScratchRegisters());
|
|
|
|
State* state = &m_ftlState;
|
|
Node* node = m_node;
|
|
NodeType op = m_node->op();
|
|
JSValue child1Constant = m_state.forNode(m_node->child1()).value();
|
|
|
|
auto nodeIndex = m_nodeIndexInGraph;
|
|
patchpoint->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
Vector<GPRReg> gpScratch;
|
|
Vector<FPRReg> fpScratch;
|
|
Vector<SnippetParams::Value> regs;
|
|
|
|
regs.append(SnippetParams::Value(params[0].gpr(), child1Constant));
|
|
|
|
for (unsigned i = 0; i < domJIT->numGPScratchRegisters; ++i)
|
|
gpScratch.append(params.gpScratch(i));
|
|
|
|
for (unsigned i = 0; i < domJIT->numFPScratchRegisters; ++i)
|
|
fpScratch.append(params.fpScratch(i));
|
|
|
|
RefPtr<OSRExitHandle> handle = exitDescriptor->emitOSRExitLater(*state, BadType, origin, params, nodeIndex, osrExitArgumentOffset);
|
|
|
|
SnippetParams domJITParams(*state, params, node, nullptr, WTFMove(regs), WTFMove(gpScratch), WTFMove(fpScratch));
|
|
CCallHelpers::JumpList failureCases = domJIT->generator()->run(jit, domJITParams);
|
|
CCallHelpers::JumpList notJSCastFailureCases;
|
|
if (op == CheckNotJSCast) {
|
|
notJSCastFailureCases.append(jit.jump());
|
|
failureCases.link(&jit);
|
|
}
|
|
|
|
jit.addLinkTask([=] (LinkBuffer& linkBuffer) {
|
|
if (op == CheckJSCast)
|
|
linkBuffer.link(failureCases, linkBuffer.locationOf<NoPtrTag>(handle->label));
|
|
else
|
|
linkBuffer.link(notJSCastFailureCases, linkBuffer.locationOf<NoPtrTag>(handle->label));
|
|
});
|
|
});
|
|
patchpoint->effects = Effects::forCheck();
|
|
}
|
|
|
|
void compileCallDOM()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
const DOMJIT::Signature* signature = m_node->signature();
|
|
|
|
// FIXME: We should have a way to call functions with the vector of registers.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=163099
|
|
Vector<LValue, JSC_DOMJIT_SIGNATURE_MAX_ARGUMENTS_INCLUDING_THIS> operands;
|
|
|
|
unsigned index = 0;
|
|
DFG_NODE_DO_TO_CHILDREN(m_graph, m_node, [&](Node*, Edge edge) {
|
|
if (!index)
|
|
operands.append(lowCell(edge));
|
|
else {
|
|
switch (signature->arguments[index - 1]) {
|
|
case SpecString:
|
|
operands.append(lowString(edge));
|
|
break;
|
|
case SpecInt32Only:
|
|
operands.append(lowInt32(edge));
|
|
break;
|
|
case SpecBoolean:
|
|
operands.append(lowBoolean(edge));
|
|
break;
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
break;
|
|
}
|
|
}
|
|
++index;
|
|
});
|
|
|
|
unsigned argumentCountIncludingThis = signature->argumentCount + 1;
|
|
LValue result;
|
|
// FIXME: Revisit JSGlobalObject.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=203204
|
|
auto function = CFunctionPtr(signature->functionWithoutTypeCheck);
|
|
switch (argumentCountIncludingThis) {
|
|
case 1:
|
|
result = vmCall(Int64, reinterpret_cast<J_JITOperation_GP>(function.get()), weakPointer(globalObject), operands[0]);
|
|
break;
|
|
case 2:
|
|
result = vmCall(Int64, reinterpret_cast<J_JITOperation_GPP>(function.get()), weakPointer(globalObject), operands[0], operands[1]);
|
|
break;
|
|
case 3:
|
|
result = vmCall(Int64, reinterpret_cast<J_JITOperation_GPPP>(function.get()), weakPointer(globalObject), operands[0], operands[1], operands[2]);
|
|
break;
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
break;
|
|
}
|
|
|
|
setJSValue(result);
|
|
}
|
|
|
|
void compileCallDOMGetter()
|
|
{
|
|
DOMJIT::CallDOMGetterSnippet* domJIT = m_node->callDOMGetterData()->snippet;
|
|
if (!domJIT) {
|
|
// The following function is not an operation: we directly call a custom accessor getter.
|
|
// Since the getter does not have code setting topCallFrame, As is the same to IC, we should set topCallFrame in caller side.
|
|
// FIXME: Revisit JSGlobalObject.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=203204
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
m_out.storePtr(m_callFrame, m_out.absolute(&vm().topCallFrame));
|
|
if (Options::useJITCage()) {
|
|
setJSValue(
|
|
vmCall(Int64, vmEntryCustomAccessor, weakPointer(globalObject), lowCell(m_node->child1()), m_out.constIntPtr(m_graph.identifiers()[m_node->callDOMGetterData()->identifierNumber]), m_out.constIntPtr(m_node->callDOMGetterData()->customAccessorGetter.executableAddress())));
|
|
} else {
|
|
FunctionPtr<CustomAccessorPtrTag> getter = m_node->callDOMGetterData()->customAccessorGetter;
|
|
FunctionPtr<OperationPtrTag> bypassedFunction = FunctionPtr<OperationPtrTag>(MacroAssemblerCodePtr<OperationPtrTag>(WTF::tagNativeCodePtrImpl<OperationPtrTag>(WTF::untagNativeCodePtrImpl<CustomAccessorPtrTag>(getter.executableAddress()))));
|
|
setJSValue(vmCall(Int64, bypassedFunction, weakPointer(globalObject), lowCell(m_node->child1()), m_out.constIntPtr(m_graph.identifiers()[m_node->callDOMGetterData()->identifierNumber])));
|
|
}
|
|
return;
|
|
}
|
|
|
|
Edge& baseEdge = m_node->child1();
|
|
LValue base = lowCell(baseEdge);
|
|
JSValue baseConstant = m_state.forNode(baseEdge).value();
|
|
|
|
LValue globalObject = nullptr;
|
|
JSValue globalObjectConstant;
|
|
if (domJIT->requireGlobalObject) {
|
|
Edge& globalObjectEdge = m_node->child2();
|
|
globalObject = lowCell(globalObjectEdge);
|
|
globalObjectConstant = m_state.forNode(globalObjectEdge).value();
|
|
}
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int64);
|
|
patchpoint->appendSomeRegister(base);
|
|
if (domJIT->requireGlobalObject)
|
|
patchpoint->appendSomeRegister(globalObject);
|
|
patchpoint->append(m_notCellMask, ValueRep::reg(GPRInfo::notCellMaskRegister));
|
|
patchpoint->append(m_numberTag, ValueRep::reg(GPRInfo::numberTagRegister));
|
|
RefPtr<PatchpointExceptionHandle> exceptionHandle = preparePatchpointForExceptions(patchpoint);
|
|
patchpoint->clobber(RegisterSet::macroScratchRegisters());
|
|
patchpoint->numGPScratchRegisters = domJIT->numGPScratchRegisters;
|
|
patchpoint->numFPScratchRegisters = domJIT->numFPScratchRegisters;
|
|
patchpoint->resultConstraints = { ValueRep::SomeEarlyRegister };
|
|
|
|
State* state = &m_ftlState;
|
|
Node* node = m_node;
|
|
patchpoint->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
Vector<GPRReg> gpScratch;
|
|
Vector<FPRReg> fpScratch;
|
|
Vector<SnippetParams::Value> regs;
|
|
|
|
regs.append(JSValueRegs(params[0].gpr()));
|
|
regs.append(SnippetParams::Value(params[1].gpr(), baseConstant));
|
|
if (domJIT->requireGlobalObject)
|
|
regs.append(SnippetParams::Value(params[2].gpr(), globalObjectConstant));
|
|
|
|
for (unsigned i = 0; i < domJIT->numGPScratchRegisters; ++i)
|
|
gpScratch.append(params.gpScratch(i));
|
|
|
|
for (unsigned i = 0; i < domJIT->numFPScratchRegisters; ++i)
|
|
fpScratch.append(params.fpScratch(i));
|
|
|
|
Box<CCallHelpers::JumpList> exceptions = exceptionHandle->scheduleExitCreation(params)->jumps(jit);
|
|
|
|
SnippetParams domJITParams(*state, params, node, exceptions, WTFMove(regs), WTFMove(gpScratch), WTFMove(fpScratch));
|
|
domJIT->generator()->run(jit, domJITParams);
|
|
});
|
|
patchpoint->effects = Effects::forCall();
|
|
setJSValue(patchpoint);
|
|
}
|
|
|
|
void compileFilterICStatus()
|
|
{
|
|
m_interpreter.filterICStatus(m_node);
|
|
}
|
|
|
|
LValue byteSwap32(LValue value)
|
|
{
|
|
// FIXME: teach B3 byteswap
|
|
// https://bugs.webkit.org/show_bug.cgi?id=188759
|
|
|
|
RELEASE_ASSERT(value->type() == Int32);
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int32);
|
|
patchpoint->appendSomeRegister(value);
|
|
patchpoint->setGenerator([=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
jit.move(params[1].gpr(), params[0].gpr());
|
|
jit.byteSwap32(params[0].gpr());
|
|
});
|
|
patchpoint->effects = Effects::none();
|
|
return patchpoint;
|
|
}
|
|
|
|
LValue byteSwap64(LValue value)
|
|
{
|
|
// FIXME: teach B3 byteswap
|
|
// https://bugs.webkit.org/show_bug.cgi?id=188759
|
|
|
|
RELEASE_ASSERT(value->type() == Int64);
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int64);
|
|
patchpoint->appendSomeRegister(value);
|
|
patchpoint->setGenerator([=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
jit.move(params[1].gpr(), params[0].gpr());
|
|
jit.byteSwap64(params[0].gpr());
|
|
});
|
|
patchpoint->effects = Effects::none();
|
|
return patchpoint;
|
|
}
|
|
|
|
template <typename F1, typename F2>
|
|
LValue emitCodeBasedOnEndiannessBranch(LValue isLittleEndian, const F1& emitLittleEndianCode, const F2& emitBigEndianCode)
|
|
{
|
|
LType type;
|
|
|
|
LBasicBlock bigEndianCase = m_out.newBlock();
|
|
LBasicBlock littleEndianCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(m_out.testIsZero32(isLittleEndian, m_out.constInt32(1)),
|
|
unsure(bigEndianCase), unsure(littleEndianCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(bigEndianCase, littleEndianCase);
|
|
LValue bigEndianValue = emitBigEndianCode();
|
|
type = bigEndianValue ? bigEndianValue->type() : Void;
|
|
ValueFromBlock bigEndianResult = bigEndianValue ? m_out.anchor(bigEndianValue) : ValueFromBlock();
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(littleEndianCase, continuation);
|
|
LValue littleEndianValue = emitLittleEndianCode();
|
|
ValueFromBlock littleEndianResult = littleEndianValue ? m_out.anchor(littleEndianValue) : ValueFromBlock();
|
|
RELEASE_ASSERT((!littleEndianValue && !bigEndianValue) || type == littleEndianValue->type());
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
RELEASE_ASSERT(!!bigEndianResult == !!littleEndianResult);
|
|
if (bigEndianResult)
|
|
return m_out.phi(type, bigEndianResult, littleEndianResult);
|
|
return nullptr;
|
|
}
|
|
|
|
void compileDataViewGet()
|
|
{
|
|
LValue dataView = lowDataViewObject(m_node->child1());
|
|
LValue index = lowInt32(m_node->child2());
|
|
LValue isLittleEndian = nullptr;
|
|
if (m_node->child3())
|
|
isLittleEndian = lowBoolean(m_node->child3());
|
|
|
|
DataViewData data = m_node->dataViewData();
|
|
|
|
LValue length = m_out.zeroExtPtr(m_out.load32NonNegative(dataView, m_heaps.JSArrayBufferView_length));
|
|
LValue indexToCheck = m_out.zeroExtPtr(index);
|
|
if (data.byteSize > 1)
|
|
indexToCheck = m_out.add(indexToCheck, m_out.constInt64(data.byteSize - 1));
|
|
speculate(OutOfBounds, noValue(), nullptr, m_out.aboveOrEqual(indexToCheck, length));
|
|
|
|
LValue vector = caged(Gigacage::Primitive, m_out.loadPtr(dataView, m_heaps.JSArrayBufferView_vector), dataView);
|
|
|
|
TypedPointer pointer(m_heaps.typedArrayProperties, m_out.add(vector, m_out.zeroExtPtr(index)));
|
|
|
|
if (m_node->op() == DataViewGetInt) {
|
|
switch (data.byteSize) {
|
|
case 1:
|
|
if (data.isSigned)
|
|
setInt32(m_out.load8SignExt32(pointer));
|
|
else
|
|
setInt32(m_out.load8ZeroExt32(pointer));
|
|
break;
|
|
case 2: {
|
|
auto emitLittleEndianLoad = [&] {
|
|
if (data.isSigned)
|
|
return m_out.load16SignExt32(pointer);
|
|
return m_out.load16ZeroExt32(pointer);
|
|
};
|
|
|
|
auto emitBigEndianLoad = [&] {
|
|
LValue val = m_out.load16ZeroExt32(pointer);
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int32);
|
|
patchpoint->appendSomeRegister(val);
|
|
patchpoint->setGenerator([=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
jit.move(params[1].gpr(), params[0].gpr());
|
|
jit.byteSwap16(params[0].gpr());
|
|
if (data.isSigned)
|
|
jit.signExtend16To32(params[0].gpr(), params[0].gpr());
|
|
});
|
|
patchpoint->effects = Effects::none();
|
|
|
|
return patchpoint;
|
|
};
|
|
|
|
if (data.isLittleEndian == TriState::False)
|
|
setInt32(emitBigEndianLoad());
|
|
else if (data.isLittleEndian == TriState::True)
|
|
setInt32(emitLittleEndianLoad());
|
|
else
|
|
setInt32(emitCodeBasedOnEndiannessBranch(isLittleEndian, emitLittleEndianLoad, emitBigEndianLoad));
|
|
|
|
break;
|
|
}
|
|
case 4: {
|
|
LValue loadedValue = m_out.load32(pointer);
|
|
|
|
if (data.isLittleEndian == TriState::False)
|
|
loadedValue = byteSwap32(loadedValue);
|
|
else if (data.isLittleEndian == TriState::Indeterminate) {
|
|
auto emitLittleEndianCode = [&] {
|
|
return loadedValue;
|
|
};
|
|
auto emitBigEndianCode = [&] {
|
|
return byteSwap32(loadedValue);
|
|
};
|
|
|
|
loadedValue = emitCodeBasedOnEndiannessBranch(isLittleEndian, emitLittleEndianCode, emitBigEndianCode);
|
|
}
|
|
|
|
if (data.isSigned)
|
|
setInt32(loadedValue);
|
|
else
|
|
setStrictInt52(m_out.zeroExt(loadedValue, Int64));
|
|
|
|
break;
|
|
}
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
}
|
|
} else {
|
|
switch (data.byteSize) {
|
|
case 4: {
|
|
auto emitLittleEndianCode = [&] {
|
|
return m_out.floatToDouble(m_out.loadFloat(pointer));
|
|
};
|
|
|
|
auto emitBigEndianCode = [&] {
|
|
LValue loadedValue = m_out.load32(pointer);
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Double);
|
|
patchpoint->appendSomeRegister(loadedValue);
|
|
patchpoint->numGPScratchRegisters = 1;
|
|
patchpoint->setGenerator([=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
jit.move(params[1].gpr(), params.gpScratch(0));
|
|
jit.byteSwap32(params.gpScratch(0));
|
|
jit.move32ToFloat(params.gpScratch(0), params[0].fpr());
|
|
jit.convertFloatToDouble(params[0].fpr(), params[0].fpr());
|
|
});
|
|
patchpoint->effects = Effects::none();
|
|
return patchpoint;
|
|
};
|
|
|
|
if (data.isLittleEndian == TriState::True)
|
|
setDouble(emitLittleEndianCode());
|
|
else if (data.isLittleEndian == TriState::False)
|
|
setDouble(emitBigEndianCode());
|
|
else
|
|
setDouble(emitCodeBasedOnEndiannessBranch(isLittleEndian, emitLittleEndianCode, emitBigEndianCode));
|
|
|
|
break;
|
|
}
|
|
case 8: {
|
|
auto emitLittleEndianCode = [&] {
|
|
return m_out.loadDouble(pointer);
|
|
};
|
|
|
|
auto emitBigEndianCode = [&] {
|
|
LValue loadedValue = m_out.load64(pointer);
|
|
loadedValue = byteSwap64(loadedValue);
|
|
return m_out.bitCast(loadedValue, Double);
|
|
};
|
|
|
|
if (data.isLittleEndian == TriState::True)
|
|
setDouble(emitLittleEndianCode());
|
|
else if (data.isLittleEndian == TriState::False)
|
|
setDouble(emitBigEndianCode());
|
|
else
|
|
setDouble(emitCodeBasedOnEndiannessBranch(isLittleEndian, emitLittleEndianCode, emitBigEndianCode));
|
|
|
|
break;
|
|
}
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
}
|
|
}
|
|
}
|
|
|
|
void compileDataViewSet()
|
|
{
|
|
LValue dataView = lowDataViewObject(m_graph.varArgChild(m_node, 0));
|
|
LValue index = lowInt32(m_graph.varArgChild(m_node, 1));
|
|
LValue isLittleEndian = nullptr;
|
|
if (m_graph.varArgChild(m_node, 3))
|
|
isLittleEndian = lowBoolean(m_graph.varArgChild(m_node, 3));
|
|
|
|
DataViewData data = m_node->dataViewData();
|
|
|
|
LValue length = m_out.zeroExtPtr(m_out.load32NonNegative(dataView, m_heaps.JSArrayBufferView_length));
|
|
LValue indexToCheck = m_out.zeroExtPtr(index);
|
|
if (data.byteSize > 1)
|
|
indexToCheck = m_out.add(indexToCheck, m_out.constInt64(data.byteSize - 1));
|
|
speculate(OutOfBounds, noValue(), nullptr, m_out.aboveOrEqual(indexToCheck, length));
|
|
|
|
Edge& valueEdge = m_graph.varArgChild(m_node, 2);
|
|
LValue valueToStore;
|
|
switch (valueEdge.useKind()) {
|
|
case Int32Use:
|
|
valueToStore = lowInt32(valueEdge);
|
|
break;
|
|
case DoubleRepUse:
|
|
valueToStore = lowDouble(valueEdge);
|
|
break;
|
|
case Int52RepUse:
|
|
valueToStore = lowStrictInt52(valueEdge);
|
|
break;
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
}
|
|
|
|
LValue vector = caged(Gigacage::Primitive, m_out.loadPtr(dataView, m_heaps.JSArrayBufferView_vector), dataView);
|
|
TypedPointer pointer(m_heaps.typedArrayProperties, m_out.add(vector, m_out.zeroExtPtr(index)));
|
|
|
|
if (data.isFloatingPoint) {
|
|
if (data.byteSize == 4) {
|
|
valueToStore = m_out.doubleToFloat(valueToStore);
|
|
|
|
auto emitLittleEndianCode = [&] () -> LValue {
|
|
m_out.storeFloat(valueToStore, pointer);
|
|
return nullptr;
|
|
};
|
|
|
|
auto emitBigEndianCode = [&] () -> LValue {
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int32);
|
|
patchpoint->appendSomeRegister(valueToStore);
|
|
patchpoint->setGenerator([=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
jit.moveFloatTo32(params[1].fpr(), params[0].gpr());
|
|
jit.byteSwap32(params[0].gpr());
|
|
});
|
|
patchpoint->effects = Effects::none();
|
|
m_out.store32(patchpoint, pointer);
|
|
return nullptr;
|
|
};
|
|
|
|
if (data.isLittleEndian == TriState::False)
|
|
emitBigEndianCode();
|
|
else if (data.isLittleEndian == TriState::True)
|
|
emitLittleEndianCode();
|
|
else
|
|
emitCodeBasedOnEndiannessBranch(isLittleEndian, emitLittleEndianCode, emitBigEndianCode);
|
|
|
|
} else {
|
|
RELEASE_ASSERT(data.byteSize == 8);
|
|
auto emitLittleEndianCode = [&] () -> LValue {
|
|
m_out.storeDouble(valueToStore, pointer);
|
|
return nullptr;
|
|
};
|
|
auto emitBigEndianCode = [&] () -> LValue {
|
|
m_out.store64(byteSwap64(m_out.bitCast(valueToStore, Int64)), pointer);
|
|
return nullptr;
|
|
};
|
|
|
|
if (data.isLittleEndian == TriState::False)
|
|
emitBigEndianCode();
|
|
else if (data.isLittleEndian == TriState::True)
|
|
emitLittleEndianCode();
|
|
else
|
|
emitCodeBasedOnEndiannessBranch(isLittleEndian, emitLittleEndianCode, emitBigEndianCode);
|
|
}
|
|
} else {
|
|
switch (data.byteSize) {
|
|
case 1:
|
|
RELEASE_ASSERT(valueEdge.useKind() == Int32Use);
|
|
m_out.store32As8(valueToStore, pointer);
|
|
break;
|
|
case 2: {
|
|
RELEASE_ASSERT(valueEdge.useKind() == Int32Use);
|
|
|
|
auto emitLittleEndianCode = [&] () -> LValue {
|
|
m_out.store32As16(valueToStore, pointer);
|
|
return nullptr;
|
|
};
|
|
auto emitBigEndianCode = [&] () -> LValue {
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int32);
|
|
patchpoint->appendSomeRegister(valueToStore);
|
|
patchpoint->setGenerator([=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
jit.move(params[1].gpr(), params[0].gpr());
|
|
jit.byteSwap16(params[0].gpr());
|
|
});
|
|
patchpoint->effects = Effects::none();
|
|
|
|
m_out.store32As16(patchpoint, pointer);
|
|
return nullptr;
|
|
};
|
|
|
|
if (data.isLittleEndian == TriState::False)
|
|
emitBigEndianCode();
|
|
else if (data.isLittleEndian == TriState::True)
|
|
emitLittleEndianCode();
|
|
else
|
|
emitCodeBasedOnEndiannessBranch(isLittleEndian, emitLittleEndianCode, emitBigEndianCode);
|
|
break;
|
|
}
|
|
case 4: {
|
|
RELEASE_ASSERT(valueEdge.useKind() == Int32Use || valueEdge.useKind() == Int52RepUse);
|
|
|
|
if (valueEdge.useKind() == Int52RepUse)
|
|
valueToStore = m_out.castToInt32(valueToStore);
|
|
|
|
auto emitLittleEndianCode = [&] () -> LValue {
|
|
m_out.store32(valueToStore, pointer);
|
|
return nullptr;
|
|
};
|
|
auto emitBigEndianCode = [&] () -> LValue {
|
|
m_out.store32(byteSwap32(valueToStore), pointer);
|
|
return nullptr;
|
|
};
|
|
|
|
if (data.isLittleEndian == TriState::False)
|
|
emitBigEndianCode();
|
|
else if (data.isLittleEndian == TriState::True)
|
|
emitLittleEndianCode();
|
|
else
|
|
emitCodeBasedOnEndiannessBranch(isLittleEndian, emitLittleEndianCode, emitBigEndianCode);
|
|
|
|
break;
|
|
}
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
}
|
|
}
|
|
}
|
|
|
|
void compileDateGet()
|
|
{
|
|
LValue base = lowDateObject(m_node->child1());
|
|
|
|
auto emitGetCodeWithCallback = [&] (const AbstractHeap& cachedDoubleOffset, const AbstractHeap& cachedDataOffset, auto* operation, auto callback) {
|
|
LBasicBlock dataExistsCase = m_out.newBlock();
|
|
LBasicBlock fastCase = m_out.newBlock();
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue data = m_out.loadPtr(base, m_heaps.DateInstance_data);
|
|
m_out.branch(m_out.notZero64(data), unsure(dataExistsCase), unsure(slowCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(dataExistsCase, fastCase);
|
|
LValue milliseconds = m_out.loadDouble(base, m_heaps.DateInstance_internalNumber);
|
|
LValue cachedMilliseconds = m_out.loadDouble(data, cachedDoubleOffset);
|
|
m_out.branch(m_out.doubleNotEqualOrUnordered(milliseconds, cachedMilliseconds), unsure(slowCase), unsure(fastCase));
|
|
|
|
m_out.appendTo(fastCase, slowCase);
|
|
ValueFromBlock fastResult = m_out.anchor(boxInt32(callback(m_out.load32(data, cachedDataOffset))));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
ValueFromBlock slowResult = m_out.anchor(vmCall(Int64, operation, m_vmValue, base));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setJSValue(m_out.phi(Int64, fastResult, slowResult));
|
|
};
|
|
|
|
auto emitGetCode = [&] (const AbstractHeap& cachedDoubleOffset, const AbstractHeap& cachedDataOffset, auto* operation) {
|
|
emitGetCodeWithCallback(cachedDoubleOffset, cachedDataOffset, operation, [] (LValue value) { return value; });
|
|
};
|
|
|
|
switch (m_node->intrinsic()) {
|
|
case DatePrototypeGetTimeIntrinsic:
|
|
setDouble(m_out.loadDouble(base, m_heaps.DateInstance_internalNumber));
|
|
break;
|
|
|
|
case DatePrototypeGetMillisecondsIntrinsic:
|
|
case DatePrototypeGetUTCMillisecondsIntrinsic: {
|
|
LValue milliseconds = m_out.loadDouble(base, m_heaps.DateInstance_internalNumber);
|
|
LValue msPerSecondConstant = m_out.constDouble(msPerSecond);
|
|
LValue seconds = m_out.doubleFloor(m_out.doubleDiv(milliseconds, msPerSecondConstant));
|
|
LValue result = m_out.doubleToInt(m_out.doubleSub(milliseconds, m_out.doubleMul(seconds, msPerSecondConstant)));
|
|
setJSValue(m_out.select(m_out.doubleNotEqualOrUnordered(milliseconds, milliseconds), m_out.constInt64(JSValue::encode(jsNaN())), boxInt32(result)));
|
|
break;
|
|
}
|
|
|
|
case DatePrototypeGetFullYearIntrinsic:
|
|
emitGetCode(m_heaps.DateInstanceData_gregorianDateTimeCachedForMS, m_heaps.DateInstanceData_cachedGregorianDateTime_year, operationDateGetFullYear);
|
|
break;
|
|
case DatePrototypeGetUTCFullYearIntrinsic:
|
|
emitGetCode(m_heaps.DateInstanceData_gregorianDateTimeUTCCachedForMS, m_heaps.DateInstanceData_cachedGregorianDateTimeUTC_year, operationDateGetUTCFullYear);
|
|
break;
|
|
case DatePrototypeGetMonthIntrinsic:
|
|
emitGetCode(m_heaps.DateInstanceData_gregorianDateTimeCachedForMS, m_heaps.DateInstanceData_cachedGregorianDateTime_month, operationDateGetMonth);
|
|
break;
|
|
case DatePrototypeGetUTCMonthIntrinsic:
|
|
emitGetCode(m_heaps.DateInstanceData_gregorianDateTimeUTCCachedForMS, m_heaps.DateInstanceData_cachedGregorianDateTimeUTC_month, operationDateGetUTCMonth);
|
|
break;
|
|
case DatePrototypeGetDateIntrinsic:
|
|
emitGetCode(m_heaps.DateInstanceData_gregorianDateTimeCachedForMS, m_heaps.DateInstanceData_cachedGregorianDateTime_monthDay, operationDateGetDate);
|
|
break;
|
|
case DatePrototypeGetUTCDateIntrinsic:
|
|
emitGetCode(m_heaps.DateInstanceData_gregorianDateTimeUTCCachedForMS, m_heaps.DateInstanceData_cachedGregorianDateTimeUTC_monthDay, operationDateGetUTCDate);
|
|
break;
|
|
case DatePrototypeGetDayIntrinsic:
|
|
emitGetCode(m_heaps.DateInstanceData_gregorianDateTimeCachedForMS, m_heaps.DateInstanceData_cachedGregorianDateTime_weekDay, operationDateGetDay);
|
|
break;
|
|
case DatePrototypeGetUTCDayIntrinsic:
|
|
emitGetCode(m_heaps.DateInstanceData_gregorianDateTimeUTCCachedForMS, m_heaps.DateInstanceData_cachedGregorianDateTimeUTC_weekDay, operationDateGetUTCDay);
|
|
break;
|
|
case DatePrototypeGetHoursIntrinsic:
|
|
emitGetCode(m_heaps.DateInstanceData_gregorianDateTimeCachedForMS, m_heaps.DateInstanceData_cachedGregorianDateTime_hour, operationDateGetHours);
|
|
break;
|
|
case DatePrototypeGetUTCHoursIntrinsic:
|
|
emitGetCode(m_heaps.DateInstanceData_gregorianDateTimeUTCCachedForMS, m_heaps.DateInstanceData_cachedGregorianDateTimeUTC_hour, operationDateGetUTCHours);
|
|
break;
|
|
case DatePrototypeGetMinutesIntrinsic:
|
|
emitGetCode(m_heaps.DateInstanceData_gregorianDateTimeCachedForMS, m_heaps.DateInstanceData_cachedGregorianDateTime_minute, operationDateGetMinutes);
|
|
break;
|
|
case DatePrototypeGetUTCMinutesIntrinsic:
|
|
emitGetCode(m_heaps.DateInstanceData_gregorianDateTimeUTCCachedForMS, m_heaps.DateInstanceData_cachedGregorianDateTimeUTC_minute, operationDateGetUTCMinutes);
|
|
break;
|
|
case DatePrototypeGetSecondsIntrinsic:
|
|
emitGetCode(m_heaps.DateInstanceData_gregorianDateTimeCachedForMS, m_heaps.DateInstanceData_cachedGregorianDateTime_second, operationDateGetSeconds);
|
|
break;
|
|
case DatePrototypeGetUTCSecondsIntrinsic:
|
|
emitGetCode(m_heaps.DateInstanceData_gregorianDateTimeUTCCachedForMS, m_heaps.DateInstanceData_cachedGregorianDateTimeUTC_second, operationDateGetUTCSeconds);
|
|
break;
|
|
|
|
case DatePrototypeGetTimezoneOffsetIntrinsic:
|
|
emitGetCodeWithCallback(m_heaps.DateInstanceData_gregorianDateTimeCachedForMS, m_heaps.DateInstanceData_cachedGregorianDateTime_utcOffsetInMinute, operationDateGetTimezoneOffset, [&] (LValue offset) {
|
|
return m_out.neg(offset);
|
|
});
|
|
break;
|
|
|
|
case DatePrototypeGetYearIntrinsic:
|
|
emitGetCodeWithCallback(m_heaps.DateInstanceData_gregorianDateTimeCachedForMS, m_heaps.DateInstanceData_cachedGregorianDateTime_year, operationDateGetYear, [&] (LValue year) {
|
|
return m_out.sub(year, m_out.constInt32(1900));
|
|
});
|
|
break;
|
|
|
|
default:
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
}
|
|
}
|
|
|
|
void compileLoopHint()
|
|
{
|
|
if (LIKELY(!Options::returnEarlyFromInfiniteLoopsForFuzzing()))
|
|
return;
|
|
|
|
CodeBlock* baselineCodeBlock = m_graph.baselineCodeBlockFor(m_origin.semantic);
|
|
if (!baselineCodeBlock->loopHintsAreEligibleForFuzzingEarlyReturn())
|
|
return;
|
|
|
|
BytecodeIndex bytecodeIndex = m_origin.semantic.bytecodeIndex();
|
|
const Instruction* instruction = baselineCodeBlock->instructions().at(bytecodeIndex.offset()).ptr();
|
|
VM* vm = &this->vm();
|
|
uint64_t* ptr = vm->getLoopHintExecutionCounter(instruction);
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Void);
|
|
patchpoint->effects = Effects::none();
|
|
patchpoint->effects.exitsSideways = true;
|
|
patchpoint->effects.writesLocalState = true;
|
|
patchpoint->setGenerator([=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
auto restore = [&] {
|
|
jit.popToRestore(GPRInfo::regT2);
|
|
jit.popToRestore(GPRInfo::regT1);
|
|
jit.popToRestore(GPRInfo::regT0);
|
|
};
|
|
|
|
jit.pushToSave(GPRInfo::regT0);
|
|
jit.pushToSave(GPRInfo::regT1);
|
|
jit.pushToSave(GPRInfo::regT2);
|
|
|
|
jit.move(CCallHelpers::TrustedImm64(Options::earlyReturnFromInfiniteLoopsLimit()), GPRInfo::regT2);
|
|
jit.move(CCallHelpers::TrustedImmPtr(ptr), GPRInfo::regT0);
|
|
jit.load64(CCallHelpers::Address(GPRInfo::regT0), GPRInfo::regT1);
|
|
auto skipEarlyReturn = jit.branch64(CCallHelpers::Below, GPRInfo::regT1, GPRInfo::regT2);
|
|
|
|
if constexpr (validateDFGDoesGC) {
|
|
if (Options::validateDoesGC()) {
|
|
// We need to mock what a Return does: claims to GC.
|
|
jit.move(CCallHelpers::TrustedImmPtr(vm->heap.addressOfDoesGC()), GPRInfo::regT0);
|
|
jit.move(CCallHelpers::TrustedImm32(DoesGCCheck::encode(true, DoesGCCheck::Special::Uninitialized)), GPRInfo::regT1);
|
|
jit.store32(GPRInfo::regT1, CCallHelpers::Address(GPRInfo::regT0));
|
|
}
|
|
}
|
|
restore();
|
|
jit.moveValue(baselineCodeBlock->globalObject(), JSValueRegs { GPRInfo::returnValueGPR });
|
|
params.code().emitEpilogue(jit);
|
|
|
|
skipEarlyReturn.link(&jit);
|
|
jit.add64(CCallHelpers::TrustedImm32(1), GPRInfo::regT1);
|
|
jit.store64(GPRInfo::regT1, CCallHelpers::Address(GPRInfo::regT0));
|
|
restore();
|
|
});
|
|
}
|
|
|
|
void emitSwitchForMultiByOffset(LValue base, bool structuresChecked, Vector<SwitchCase, 2>& cases, LBasicBlock exit)
|
|
{
|
|
if (cases.isEmpty()) {
|
|
m_out.jump(exit);
|
|
return;
|
|
}
|
|
|
|
if (structuresChecked) {
|
|
std::sort(
|
|
cases.begin(), cases.end(),
|
|
[&] (const SwitchCase& a, const SwitchCase& b) -> bool {
|
|
return a.value()->asInt() < b.value()->asInt();
|
|
});
|
|
SwitchCase last = cases.takeLast();
|
|
m_out.switchInstruction(
|
|
m_out.load32(base, m_heaps.JSCell_structureID), cases, last.target(), Weight(0));
|
|
return;
|
|
}
|
|
|
|
m_out.switchInstruction(
|
|
m_out.load32(base, m_heaps.JSCell_structureID), cases, exit, Weight(0));
|
|
}
|
|
|
|
void compareEqObjectOrOtherToObject(Edge leftChild, Edge rightChild)
|
|
{
|
|
LValue rightCell = lowCell(rightChild);
|
|
LValue leftValue = lowJSValue(leftChild, ManualOperandSpeculation);
|
|
|
|
speculateTruthyObject(rightChild, rightCell, SpecObject);
|
|
|
|
LBasicBlock leftCellCase = m_out.newBlock();
|
|
LBasicBlock leftNotCellCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
isCell(leftValue, provenType(leftChild)),
|
|
unsure(leftCellCase), unsure(leftNotCellCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(leftCellCase, leftNotCellCase);
|
|
speculateTruthyObject(leftChild, leftValue, SpecObject | (~SpecCellCheck));
|
|
ValueFromBlock cellResult = m_out.anchor(m_out.equal(rightCell, leftValue));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(leftNotCellCase, continuation);
|
|
FTL_TYPE_CHECK(
|
|
jsValueValue(leftValue), leftChild, SpecOther | SpecCellCheck, isNotOther(leftValue));
|
|
ValueFromBlock notCellResult = m_out.anchor(m_out.booleanFalse);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setBoolean(m_out.phi(Int32, cellResult, notCellResult));
|
|
}
|
|
|
|
void speculateTruthyObject(Edge edge, LValue cell, SpeculatedType filter)
|
|
{
|
|
if (masqueradesAsUndefinedWatchpointIsStillValid()) {
|
|
FTL_TYPE_CHECK(jsValueValue(cell), edge, filter, isNotObject(cell));
|
|
return;
|
|
}
|
|
|
|
FTL_TYPE_CHECK(jsValueValue(cell), edge, filter, isNotObject(cell));
|
|
speculate(
|
|
BadType, jsValueValue(cell), edge.node(),
|
|
m_out.testNonZero32(
|
|
m_out.load8ZeroExt32(cell, m_heaps.JSCell_typeInfoFlags),
|
|
m_out.constInt32(MasqueradesAsUndefined)));
|
|
}
|
|
|
|
template<typename IntFunctor>
|
|
void genericJSValueCompare(const IntFunctor& intFunctor, S_JITOperation_GJJ helperFunction)
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LValue left = lowJSValue(m_node->child1(), ManualOperandSpeculation);
|
|
LValue right = lowJSValue(m_node->child2(), ManualOperandSpeculation);
|
|
speculate(m_node->child1());
|
|
speculate(m_node->child2());
|
|
|
|
LBasicBlock leftIsInt = m_out.newBlock();
|
|
LBasicBlock fastPath = m_out.newBlock();
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(isNotInt32(left, provenType(m_node->child1())), rarely(slowPath), usually(leftIsInt));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(leftIsInt, fastPath);
|
|
m_out.branch(isNotInt32(right, provenType(m_node->child2())), rarely(slowPath), usually(fastPath));
|
|
|
|
m_out.appendTo(fastPath, slowPath);
|
|
ValueFromBlock fastResult = m_out.anchor(intFunctor(unboxInt32(left), unboxInt32(right)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
ValueFromBlock slowResult = m_out.anchor(m_out.notNull(vmCall(
|
|
pointerType(), helperFunction, weakPointer(globalObject), left, right)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
setBoolean(m_out.phi(Int32, fastResult, slowResult));
|
|
}
|
|
|
|
LValue stringsEqual(LValue leftJSString, LValue rightJSString, Edge leftJSStringEdge = Edge(), Edge rightJSStringEdge = Edge())
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LBasicBlock notTriviallyUnequalCase = m_out.newBlock();
|
|
LBasicBlock notEmptyCase = m_out.newBlock();
|
|
LBasicBlock leftReadyCase = m_out.newBlock();
|
|
LBasicBlock rightReadyCase = m_out.newBlock();
|
|
LBasicBlock left8BitCase = m_out.newBlock();
|
|
LBasicBlock right8BitCase = m_out.newBlock();
|
|
LBasicBlock loop = m_out.newBlock();
|
|
LBasicBlock bytesEqual = m_out.newBlock();
|
|
LBasicBlock trueCase = m_out.newBlock();
|
|
LBasicBlock falseCase = m_out.newBlock();
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(isRopeString(leftJSString, leftJSStringEdge), rarely(slowCase), usually(leftReadyCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(leftReadyCase, rightReadyCase);
|
|
m_out.branch(isRopeString(rightJSString, rightJSStringEdge), rarely(slowCase), usually(rightReadyCase));
|
|
|
|
m_out.appendTo(rightReadyCase, notTriviallyUnequalCase);
|
|
LValue left = m_out.loadPtr(leftJSString, m_heaps.JSString_value);
|
|
LValue right = m_out.loadPtr(rightJSString, m_heaps.JSString_value);
|
|
LValue length = m_out.load32(left, m_heaps.StringImpl_length);
|
|
m_out.branch(
|
|
m_out.notEqual(length, m_out.load32(right, m_heaps.StringImpl_length)),
|
|
unsure(falseCase), unsure(notTriviallyUnequalCase));
|
|
|
|
m_out.appendTo(notTriviallyUnequalCase, notEmptyCase);
|
|
m_out.branch(m_out.isZero32(length), unsure(trueCase), unsure(notEmptyCase));
|
|
|
|
m_out.appendTo(notEmptyCase, left8BitCase);
|
|
m_out.branch(
|
|
m_out.testIsZero32(
|
|
m_out.load32(left, m_heaps.StringImpl_hashAndFlags),
|
|
m_out.constInt32(StringImpl::flagIs8Bit())),
|
|
unsure(slowCase), unsure(left8BitCase));
|
|
|
|
m_out.appendTo(left8BitCase, right8BitCase);
|
|
m_out.branch(
|
|
m_out.testIsZero32(
|
|
m_out.load32(right, m_heaps.StringImpl_hashAndFlags),
|
|
m_out.constInt32(StringImpl::flagIs8Bit())),
|
|
unsure(slowCase), unsure(right8BitCase));
|
|
|
|
m_out.appendTo(right8BitCase, loop);
|
|
|
|
LValue leftData = m_out.loadPtr(left, m_heaps.StringImpl_data);
|
|
LValue rightData = m_out.loadPtr(right, m_heaps.StringImpl_data);
|
|
|
|
ValueFromBlock indexAtStart = m_out.anchor(length);
|
|
|
|
m_out.jump(loop);
|
|
|
|
m_out.appendTo(loop, bytesEqual);
|
|
|
|
LValue indexAtLoopTop = m_out.phi(Int32, indexAtStart);
|
|
LValue indexInLoop = m_out.sub(indexAtLoopTop, m_out.int32One);
|
|
|
|
LValue leftByte = m_out.load8ZeroExt32(
|
|
m_out.baseIndex(m_heaps.characters8, leftData, m_out.zeroExtPtr(indexInLoop)));
|
|
LValue rightByte = m_out.load8ZeroExt32(
|
|
m_out.baseIndex(m_heaps.characters8, rightData, m_out.zeroExtPtr(indexInLoop)));
|
|
|
|
m_out.branch(m_out.notEqual(leftByte, rightByte), unsure(falseCase), unsure(bytesEqual));
|
|
|
|
m_out.appendTo(bytesEqual, trueCase);
|
|
|
|
ValueFromBlock indexForNextIteration = m_out.anchor(indexInLoop);
|
|
m_out.addIncomingToPhi(indexAtLoopTop, indexForNextIteration);
|
|
m_out.branch(m_out.notZero32(indexInLoop), unsure(loop), unsure(trueCase));
|
|
|
|
m_out.appendTo(trueCase, falseCase);
|
|
|
|
ValueFromBlock trueResult = m_out.anchor(m_out.booleanTrue);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(falseCase, slowCase);
|
|
|
|
ValueFromBlock falseResult = m_out.anchor(m_out.booleanFalse);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
|
|
LValue slowResultValue = vmCall(
|
|
Int64, operationCompareStringEq, weakPointer(globalObject),
|
|
leftJSString, rightJSString);
|
|
ValueFromBlock slowResult = m_out.anchor(unboxBoolean(slowResultValue));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
return m_out.phi(Int32, trueResult, falseResult, slowResult);
|
|
}
|
|
|
|
enum ScratchFPRUsage {
|
|
DontNeedScratchFPR,
|
|
NeedScratchFPR
|
|
};
|
|
template<typename BinaryArithOpGenerator, ScratchFPRUsage scratchFPRUsage = DontNeedScratchFPR>
|
|
void emitBinarySnippet(J_JITOperation_GJJ slowPathFunction)
|
|
{
|
|
Node* node = m_node;
|
|
|
|
LValue left = lowJSValue(node->child1(), ManualOperandSpeculation);
|
|
LValue right = lowJSValue(node->child2(), ManualOperandSpeculation);
|
|
speculate(node->child1());
|
|
speculate(node->child2());
|
|
|
|
SnippetOperand leftOperand(m_state.forNode(node->child1()).resultType());
|
|
SnippetOperand rightOperand(m_state.forNode(node->child2()).resultType());
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int64);
|
|
patchpoint->appendSomeRegister(left);
|
|
patchpoint->appendSomeRegister(right);
|
|
patchpoint->append(m_notCellMask, ValueRep::lateReg(GPRInfo::notCellMaskRegister));
|
|
patchpoint->append(m_numberTag, ValueRep::lateReg(GPRInfo::numberTagRegister));
|
|
RefPtr<PatchpointExceptionHandle> exceptionHandle =
|
|
preparePatchpointForExceptions(patchpoint);
|
|
patchpoint->numGPScratchRegisters = 1;
|
|
patchpoint->numFPScratchRegisters = 2;
|
|
if (scratchFPRUsage == NeedScratchFPR)
|
|
patchpoint->numFPScratchRegisters++;
|
|
patchpoint->clobber(RegisterSet::macroScratchRegisters());
|
|
patchpoint->resultConstraints = { ValueRep::SomeEarlyRegister };
|
|
State* state = &m_ftlState;
|
|
patchpoint->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
Box<CCallHelpers::JumpList> exceptions =
|
|
exceptionHandle->scheduleExitCreation(params)->jumps(jit);
|
|
|
|
auto generator = Box<BinaryArithOpGenerator>::create(
|
|
leftOperand, rightOperand, JSValueRegs(params[0].gpr()),
|
|
JSValueRegs(params[1].gpr()), JSValueRegs(params[2].gpr()),
|
|
params.fpScratch(0), params.fpScratch(1), params.gpScratch(0),
|
|
scratchFPRUsage == NeedScratchFPR ? params.fpScratch(2) : InvalidFPRReg);
|
|
|
|
generator->generateFastPath(jit);
|
|
|
|
if (generator->didEmitFastPath()) {
|
|
generator->endJumpList().link(&jit);
|
|
CCallHelpers::Label done = jit.label();
|
|
|
|
params.addLatePath(
|
|
[=] (CCallHelpers& jit) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
generator->slowPathJumpList().link(&jit);
|
|
callOperation(
|
|
*state, params.unavailableRegisters(), jit, node->origin.semantic,
|
|
exceptions.get(), slowPathFunction, params[0].gpr(),
|
|
jit.codeBlock()->globalObjectFor(node->origin.semantic),
|
|
params[1].gpr(), params[2].gpr());
|
|
jit.jump().linkTo(done, &jit);
|
|
});
|
|
} else {
|
|
callOperation(
|
|
*state, params.unavailableRegisters(), jit, node->origin.semantic,
|
|
exceptions.get(), slowPathFunction, params[0].gpr(), jit.codeBlock()->globalObjectFor(node->origin.semantic), params[1].gpr(),
|
|
params[2].gpr());
|
|
}
|
|
});
|
|
|
|
setJSValue(patchpoint);
|
|
}
|
|
|
|
template<typename BinaryBitOpGenerator>
|
|
void emitBinaryBitOpSnippet(J_JITOperation_GJJ slowPathFunction)
|
|
{
|
|
Node* node = m_node;
|
|
|
|
DFG_ASSERT(m_graph, node, node->isBinaryUseKind(UntypedUse) || node->isBinaryUseKind(AnyBigIntUse) || node->isBinaryUseKind(BigInt32Use));
|
|
LValue left = lowJSValue(node->child1(), ManualOperandSpeculation);
|
|
LValue right = lowJSValue(node->child2(), ManualOperandSpeculation);
|
|
speculate(node, node->child1());
|
|
speculate(node, node->child2());
|
|
|
|
SnippetOperand leftOperand(m_state.forNode(node->child1()).resultType());
|
|
SnippetOperand rightOperand(m_state.forNode(node->child2()).resultType());
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int64);
|
|
patchpoint->appendSomeRegister(left);
|
|
patchpoint->appendSomeRegister(right);
|
|
patchpoint->append(m_notCellMask, ValueRep::lateReg(GPRInfo::notCellMaskRegister));
|
|
patchpoint->append(m_numberTag, ValueRep::lateReg(GPRInfo::numberTagRegister));
|
|
RefPtr<PatchpointExceptionHandle> exceptionHandle =
|
|
preparePatchpointForExceptions(patchpoint);
|
|
patchpoint->numGPScratchRegisters = 1;
|
|
patchpoint->clobber(RegisterSet::macroScratchRegisters());
|
|
patchpoint->resultConstraints = { ValueRep::SomeEarlyRegister };
|
|
State* state = &m_ftlState;
|
|
patchpoint->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
Box<CCallHelpers::JumpList> exceptions =
|
|
exceptionHandle->scheduleExitCreation(params)->jumps(jit);
|
|
|
|
auto generator = Box<BinaryBitOpGenerator>::create(
|
|
leftOperand, rightOperand, JSValueRegs(params[0].gpr()),
|
|
JSValueRegs(params[1].gpr()), JSValueRegs(params[2].gpr()), params.gpScratch(0));
|
|
|
|
generator->generateFastPath(jit);
|
|
generator->endJumpList().link(&jit);
|
|
CCallHelpers::Label done = jit.label();
|
|
|
|
params.addLatePath(
|
|
[=] (CCallHelpers& jit) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
generator->slowPathJumpList().link(&jit);
|
|
callOperation(
|
|
*state, params.unavailableRegisters(), jit, node->origin.semantic,
|
|
exceptions.get(), slowPathFunction, params[0].gpr(),
|
|
jit.codeBlock()->globalObjectFor(node->origin.semantic),
|
|
params[1].gpr(), params[2].gpr());
|
|
jit.jump().linkTo(done, &jit);
|
|
});
|
|
});
|
|
|
|
setJSValue(patchpoint);
|
|
}
|
|
|
|
void emitRightShiftSnippet(JITRightShiftGenerator::ShiftType shiftType)
|
|
{
|
|
Node* node = m_node;
|
|
|
|
// FIXME: Make this do exceptions.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=151686
|
|
|
|
LValue left = lowJSValue(node->child1());
|
|
LValue right = lowJSValue(node->child2());
|
|
|
|
SnippetOperand leftOperand(m_state.forNode(node->child1()).resultType());
|
|
SnippetOperand rightOperand(m_state.forNode(node->child2()).resultType());
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int64);
|
|
patchpoint->appendSomeRegister(left);
|
|
patchpoint->appendSomeRegister(right);
|
|
patchpoint->append(m_notCellMask, ValueRep::lateReg(GPRInfo::notCellMaskRegister));
|
|
patchpoint->append(m_numberTag, ValueRep::lateReg(GPRInfo::numberTagRegister));
|
|
RefPtr<PatchpointExceptionHandle> exceptionHandle =
|
|
preparePatchpointForExceptions(patchpoint);
|
|
patchpoint->numGPScratchRegisters = 1;
|
|
patchpoint->numFPScratchRegisters = 1;
|
|
patchpoint->clobber(RegisterSet::macroScratchRegisters());
|
|
patchpoint->resultConstraints = { ValueRep::SomeEarlyRegister };
|
|
State* state = &m_ftlState;
|
|
patchpoint->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
Box<CCallHelpers::JumpList> exceptions =
|
|
exceptionHandle->scheduleExitCreation(params)->jumps(jit);
|
|
|
|
auto generator = Box<JITRightShiftGenerator>::create(
|
|
leftOperand, rightOperand, JSValueRegs(params[0].gpr()),
|
|
JSValueRegs(params[1].gpr()), JSValueRegs(params[2].gpr()),
|
|
params.fpScratch(0), params.gpScratch(0), InvalidFPRReg, shiftType);
|
|
|
|
generator->generateFastPath(jit);
|
|
generator->endJumpList().link(&jit);
|
|
CCallHelpers::Label done = jit.label();
|
|
|
|
params.addLatePath(
|
|
[=] (CCallHelpers& jit) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
generator->slowPathJumpList().link(&jit);
|
|
|
|
J_JITOperation_GJJ slowPathFunction =
|
|
shiftType == JITRightShiftGenerator::SignedShift
|
|
? operationValueBitRShift : operationValueBitURShift;
|
|
|
|
callOperation(
|
|
*state, params.unavailableRegisters(), jit, node->origin.semantic,
|
|
exceptions.get(), slowPathFunction, params[0].gpr(),
|
|
jit.codeBlock()->globalObjectFor(node->origin.semantic),
|
|
params[1].gpr(), params[2].gpr());
|
|
jit.jump().linkTo(done, &jit);
|
|
});
|
|
});
|
|
|
|
setJSValue(patchpoint);
|
|
}
|
|
|
|
LValue allocateHeapCell(LValue allocator, LBasicBlock slowPath)
|
|
{
|
|
JITAllocator actualAllocator;
|
|
if (allocator->hasIntPtr())
|
|
actualAllocator = JITAllocator::constant(Allocator(bitwise_cast<LocalAllocator*>(allocator->asIntPtr())));
|
|
else
|
|
actualAllocator = JITAllocator::variable();
|
|
|
|
if (actualAllocator.isConstant()) {
|
|
if (!actualAllocator.allocator()) {
|
|
LBasicBlock haveAllocator = m_out.newBlock();
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(haveAllocator);
|
|
m_out.jump(slowPath);
|
|
m_out.appendTo(haveAllocator, lastNext);
|
|
return m_out.intPtrZero;
|
|
}
|
|
} else {
|
|
// This means that either we know that the allocator is null or we don't know what the
|
|
// allocator is. In either case, we need the null check.
|
|
LBasicBlock haveAllocator = m_out.newBlock();
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(haveAllocator);
|
|
m_out.branch(
|
|
m_out.notEqual(allocator, m_out.intPtrZero),
|
|
usually(haveAllocator), rarely(slowPath));
|
|
m_out.appendTo(haveAllocator, lastNext);
|
|
}
|
|
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(continuation);
|
|
|
|
PatchpointValue* patchpoint = m_out.patchpoint(pointerType());
|
|
if (isARM64()) {
|
|
// emitAllocateWithNonNullAllocator uses the scratch registers on ARM.
|
|
patchpoint->clobber(RegisterSet::macroScratchRegisters());
|
|
}
|
|
patchpoint->effects.terminal = true;
|
|
if (actualAllocator.isConstant())
|
|
patchpoint->numGPScratchRegisters++;
|
|
else
|
|
patchpoint->appendSomeRegisterWithClobber(allocator);
|
|
patchpoint->numGPScratchRegisters++;
|
|
patchpoint->resultConstraints = { ValueRep::SomeEarlyRegister };
|
|
|
|
m_out.appendSuccessor(usually(continuation));
|
|
m_out.appendSuccessor(rarely(slowPath));
|
|
|
|
patchpoint->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
AllowMacroScratchRegisterUsageIf allowScratchIf(jit, isARM64());
|
|
CCallHelpers::JumpList jumpToSlowPath;
|
|
|
|
GPRReg allocatorGPR;
|
|
if (actualAllocator.isConstant())
|
|
allocatorGPR = params.gpScratch(1);
|
|
else
|
|
allocatorGPR = params[1].gpr();
|
|
|
|
// We use a patchpoint to emit the allocation path because whenever we mess with
|
|
// allocation paths, we already reason about them at the machine code level. We know
|
|
// exactly what instruction sequence we want. We're confident that no compiler
|
|
// optimization could make this code better. So, it's best to have the code in
|
|
// AssemblyHelpers::emitAllocate(). That way, the same optimized path is shared by
|
|
// all of the compiler tiers.
|
|
jit.emitAllocateWithNonNullAllocator(
|
|
params[0].gpr(), actualAllocator, allocatorGPR, params.gpScratch(0),
|
|
jumpToSlowPath);
|
|
|
|
CCallHelpers::Jump jumpToSuccess;
|
|
if (!params.fallsThroughToSuccessor(0))
|
|
jumpToSuccess = jit.jump();
|
|
|
|
Vector<Box<CCallHelpers::Label>> labels = params.successorLabels();
|
|
|
|
params.addLatePath(
|
|
[=] (CCallHelpers& jit) {
|
|
jumpToSlowPath.linkTo(*labels[1], &jit);
|
|
if (jumpToSuccess.isSet())
|
|
jumpToSuccess.linkTo(*labels[0], &jit);
|
|
});
|
|
});
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
return patchpoint;
|
|
}
|
|
|
|
void storeStructure(LValue object, Structure* structure)
|
|
{
|
|
m_out.store32(m_out.constInt32(structure->id()), object, m_heaps.JSCell_structureID);
|
|
m_out.store32(
|
|
m_out.constInt32(structure->objectInitializationBlob()),
|
|
object, m_heaps.JSCell_usefulBytes);
|
|
}
|
|
|
|
void storeStructure(LValue object, LValue structure)
|
|
{
|
|
if (structure->hasIntPtr()) {
|
|
storeStructure(object, bitwise_cast<Structure*>(structure->asIntPtr()));
|
|
return;
|
|
}
|
|
|
|
LValue id = m_out.load32(structure, m_heaps.Structure_structureID);
|
|
m_out.store32(id, object, m_heaps.JSCell_structureID);
|
|
|
|
LValue blob = m_out.load32(structure, m_heaps.Structure_indexingModeIncludingHistory);
|
|
m_out.store32(blob, object, m_heaps.JSCell_usefulBytes);
|
|
}
|
|
|
|
template <typename StructureType>
|
|
LValue allocateCell(LValue allocator, StructureType structure, LBasicBlock slowPath)
|
|
{
|
|
LValue result = allocateHeapCell(allocator, slowPath);
|
|
storeStructure(result, structure);
|
|
return result;
|
|
}
|
|
|
|
LValue allocateObject(LValue allocator, RegisteredStructure structure, LValue butterfly, LBasicBlock slowPath)
|
|
{
|
|
return allocateObject(allocator, weakStructure(structure), butterfly, slowPath);
|
|
}
|
|
|
|
LValue allocateObject(LValue allocator, LValue structure, LValue butterfly, LBasicBlock slowPath)
|
|
{
|
|
LValue result = allocateCell(allocator, structure, slowPath);
|
|
if (structure->hasIntPtr()) {
|
|
splatWords(
|
|
result,
|
|
m_out.constInt32(JSFinalObject::offsetOfInlineStorage() / 8),
|
|
m_out.constInt32(JSFinalObject::offsetOfInlineStorage() / 8 + bitwise_cast<Structure*>(structure->asIntPtr())->inlineCapacity()),
|
|
m_out.int64Zero,
|
|
m_heaps.properties.atAnyNumber());
|
|
} else {
|
|
LValue end = m_out.add(
|
|
m_out.constInt32(JSFinalObject::offsetOfInlineStorage() / 8),
|
|
m_out.load8ZeroExt32(structure, m_heaps.Structure_inlineCapacity));
|
|
splatWords(
|
|
result,
|
|
m_out.constInt32(JSFinalObject::offsetOfInlineStorage() / 8),
|
|
end,
|
|
m_out.int64Zero,
|
|
m_heaps.properties.atAnyNumber());
|
|
}
|
|
|
|
m_out.storePtr(butterfly, result, m_heaps.JSObject_butterfly);
|
|
return result;
|
|
}
|
|
|
|
template<typename ClassType, typename StructureType>
|
|
LValue allocateObject(
|
|
size_t size, StructureType structure, LValue butterfly, LBasicBlock slowPath)
|
|
{
|
|
Allocator allocator = allocatorForNonVirtualConcurrently<ClassType>(vm(), size, AllocatorForMode::AllocatorIfExists);
|
|
return allocateObject(
|
|
m_out.constIntPtr(allocator.localAllocator()), structure, butterfly, slowPath);
|
|
}
|
|
|
|
template<typename ClassType, typename StructureType>
|
|
LValue allocateObject(StructureType structure, LValue butterfly, LBasicBlock slowPath)
|
|
{
|
|
return allocateObject<ClassType>(
|
|
ClassType::allocationSize(0), structure, butterfly, slowPath);
|
|
}
|
|
|
|
LValue allocatorForSize(LValue subspace, LValue size, LBasicBlock slowPath)
|
|
{
|
|
static_assert(!(MarkedSpace::sizeStep & (MarkedSpace::sizeStep - 1)), "MarkedSpace::sizeStep must be a power of two.");
|
|
|
|
// Try to do some constant-folding here.
|
|
if (subspace->hasIntPtr() && size->hasIntPtr()) {
|
|
CompleteSubspace* actualSubspace = bitwise_cast<CompleteSubspace*>(subspace->asIntPtr());
|
|
size_t actualSize = size->asIntPtr();
|
|
|
|
Allocator actualAllocator = actualSubspace->allocatorForNonVirtual(actualSize, AllocatorForMode::AllocatorIfExists);
|
|
if (!actualAllocator) {
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(continuation);
|
|
m_out.jump(slowPath);
|
|
m_out.appendTo(continuation, lastNext);
|
|
return m_out.intPtrZero;
|
|
}
|
|
|
|
return m_out.constIntPtr(actualAllocator.localAllocator());
|
|
}
|
|
|
|
unsigned stepShift = getLSBSet(MarkedSpace::sizeStep);
|
|
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(continuation);
|
|
|
|
LValue sizeClassIndex = m_out.lShr(
|
|
m_out.add(size, m_out.constIntPtr(MarkedSpace::sizeStep - 1)),
|
|
m_out.constInt32(stepShift));
|
|
|
|
m_out.branch(
|
|
m_out.above(sizeClassIndex, m_out.constIntPtr(MarkedSpace::largeCutoff >> stepShift)),
|
|
rarely(slowPath), usually(continuation));
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
|
|
return m_out.loadPtr(
|
|
m_out.baseIndex(
|
|
m_heaps.CompleteSubspace_allocatorForSizeStep,
|
|
subspace, sizeClassIndex));
|
|
}
|
|
|
|
LValue allocatorForSize(CompleteSubspace& subspace, LValue size, LBasicBlock slowPath)
|
|
{
|
|
return allocatorForSize(m_out.constIntPtr(&subspace), size, slowPath);
|
|
}
|
|
|
|
template<typename ClassType>
|
|
LValue allocateVariableSizedObject(
|
|
LValue size, RegisteredStructure structure, LValue butterfly, LBasicBlock slowPath)
|
|
{
|
|
CompleteSubspace* subspace = subspaceForConcurrently<ClassType>(vm());
|
|
RELEASE_ASSERT_WITH_MESSAGE(subspace, "CompleteSubspace is always allocated");
|
|
LValue allocator = allocatorForSize(*subspace, size, slowPath);
|
|
return allocateObject(allocator, structure, butterfly, slowPath);
|
|
}
|
|
|
|
template<typename ClassType>
|
|
LValue allocateVariableSizedCell(
|
|
LValue size, Structure* structure, LBasicBlock slowPath)
|
|
{
|
|
CompleteSubspace* subspace = subspaceForConcurrently<ClassType>(vm());
|
|
RELEASE_ASSERT_WITH_MESSAGE(subspace, "CompleteSubspace is always allocated");
|
|
LValue allocator = allocatorForSize(*subspace, size, slowPath);
|
|
return allocateCell(allocator, structure, slowPath);
|
|
}
|
|
|
|
LValue allocateObject(RegisteredStructure structure)
|
|
{
|
|
size_t allocationSize = JSFinalObject::allocationSize(structure.get()->inlineCapacity());
|
|
Allocator allocator = allocatorForNonVirtualConcurrently<JSFinalObject>(vm(), allocationSize, AllocatorForMode::AllocatorIfExists);
|
|
|
|
// FIXME: If the allocator is null, we could simply emit a normal C call to the allocator
|
|
// instead of putting it on the slow path.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=161062
|
|
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(slowPath);
|
|
|
|
ValueFromBlock fastResult = m_out.anchor(allocateObject(
|
|
m_out.constIntPtr(allocator.localAllocator()), structure, m_out.intPtrZero, slowPath));
|
|
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
|
|
VM& vm = this->vm();
|
|
LValue slowResultValue = lazySlowPath(
|
|
[=, &vm] (const Vector<Location>& locations) -> RefPtr<LazySlowPath::Generator> {
|
|
return createLazyCallGenerator(vm,
|
|
operationNewObject, locations[0].directGPR(), &vm,
|
|
CCallHelpers::TrustedImmPtr(structure.get()));
|
|
});
|
|
ValueFromBlock slowResult = m_out.anchor(slowResultValue);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
return m_out.phi(pointerType(), fastResult, slowResult);
|
|
}
|
|
|
|
struct ArrayValues {
|
|
ArrayValues()
|
|
: array(nullptr)
|
|
, butterfly(nullptr)
|
|
{
|
|
}
|
|
|
|
ArrayValues(LValue array, LValue butterfly)
|
|
: array(array)
|
|
, butterfly(butterfly)
|
|
{
|
|
}
|
|
|
|
LValue array;
|
|
LValue butterfly;
|
|
};
|
|
|
|
ArrayValues allocateJSArray(LValue publicLength, LValue vectorLength, LValue structure, LValue indexingType, bool shouldInitializeElements = true, bool shouldLargeArraySizeCreateArrayStorage = true)
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
LBasicBlock fastCase = m_out.newBlock();
|
|
LBasicBlock largeCase = m_out.newBlock();
|
|
LBasicBlock failCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(fastCase);
|
|
|
|
Optional<unsigned> staticVectorLength;
|
|
Optional<unsigned> staticVectorLengthFromPublicLength;
|
|
if (structure->hasIntPtr()) {
|
|
if (publicLength->hasInt32()) {
|
|
unsigned publicLengthConst = static_cast<unsigned>(publicLength->asInt32());
|
|
if (publicLengthConst <= MAX_STORAGE_VECTOR_LENGTH) {
|
|
publicLengthConst = Butterfly::optimalContiguousVectorLength(
|
|
bitwise_cast<Structure*>(structure->asIntPtr())->outOfLineCapacity(), publicLengthConst);
|
|
staticVectorLengthFromPublicLength = publicLengthConst;
|
|
}
|
|
|
|
}
|
|
if (vectorLength->hasInt32()) {
|
|
unsigned vectorLengthConst = static_cast<unsigned>(vectorLength->asInt32());
|
|
if (vectorLengthConst <= MAX_STORAGE_VECTOR_LENGTH) {
|
|
vectorLengthConst = Butterfly::optimalContiguousVectorLength(
|
|
bitwise_cast<Structure*>(structure->asIntPtr())->outOfLineCapacity(), vectorLengthConst);
|
|
vectorLength = m_out.constInt32(vectorLengthConst);
|
|
staticVectorLength = vectorLengthConst;
|
|
}
|
|
}
|
|
} else {
|
|
// We don't compute the optimal vector length for new Array(blah) where blah is not
|
|
// statically known, since the compute effort of doing it here is probably not worth it.
|
|
}
|
|
|
|
ValueFromBlock noButterfly = m_out.anchor(m_out.intPtrZero);
|
|
|
|
LValue predicate;
|
|
if (shouldLargeArraySizeCreateArrayStorage)
|
|
predicate = m_out.aboveOrEqual(publicLength, m_out.constInt32(MIN_ARRAY_STORAGE_CONSTRUCTION_LENGTH));
|
|
else
|
|
predicate = m_out.booleanFalse;
|
|
|
|
m_out.branch(predicate, rarely(largeCase), usually(fastCase));
|
|
|
|
m_out.appendTo(fastCase, largeCase);
|
|
|
|
LValue payloadSize =
|
|
m_out.shl(m_out.zeroExt(vectorLength, pointerType()), m_out.constIntPtr(3));
|
|
|
|
LValue butterflySize = m_out.add(
|
|
payloadSize, m_out.constIntPtr(sizeof(IndexingHeader)));
|
|
|
|
LValue allocator = allocatorForSize(vm().jsValueGigacageAuxiliarySpace, butterflySize, failCase);
|
|
LValue startOfStorage = allocateHeapCell(allocator, failCase);
|
|
|
|
LValue butterfly = m_out.add(startOfStorage, m_out.constIntPtr(sizeof(IndexingHeader)));
|
|
|
|
m_out.store32(publicLength, butterfly, m_heaps.Butterfly_publicLength);
|
|
m_out.store32(vectorLength, butterfly, m_heaps.Butterfly_vectorLength);
|
|
|
|
initializeArrayElements(
|
|
indexingType,
|
|
shouldInitializeElements ? m_out.int32Zero : publicLength, vectorLength,
|
|
butterfly);
|
|
|
|
ValueFromBlock haveButterfly = m_out.anchor(butterfly);
|
|
|
|
LValue object = allocateObject<JSArray>(structure, butterfly, failCase);
|
|
|
|
ValueFromBlock fastResult = m_out.anchor(object);
|
|
ValueFromBlock fastButterfly = m_out.anchor(butterfly);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(largeCase, failCase);
|
|
ValueFromBlock largeStructure = m_out.anchor(
|
|
weakStructure(m_graph.registerStructure(globalObject->arrayStructureForIndexingTypeDuringAllocation(ArrayWithArrayStorage))));
|
|
m_out.jump(slowCase);
|
|
|
|
m_out.appendTo(failCase, slowCase);
|
|
ValueFromBlock failStructure = m_out.anchor(structure);
|
|
m_out.jump(slowCase);
|
|
|
|
m_out.appendTo(slowCase, continuation);
|
|
LValue structureValue = m_out.phi(pointerType(), largeStructure, failStructure);
|
|
LValue butterflyValue = m_out.phi(pointerType(), noButterfly, haveButterfly);
|
|
|
|
VM& vm = this->vm();
|
|
LValue slowResultValue = nullptr;
|
|
if (vectorLength == publicLength
|
|
|| (staticVectorLengthFromPublicLength && staticVectorLength && staticVectorLength.value() == staticVectorLengthFromPublicLength.value())) {
|
|
slowResultValue = lazySlowPath(
|
|
[=, &vm] (const Vector<Location>& locations) -> RefPtr<LazySlowPath::Generator> {
|
|
return createLazyCallGenerator(vm,
|
|
operationNewArrayWithSize, locations[0].directGPR(), globalObject,
|
|
locations[1].directGPR(), locations[2].directGPR(), locations[3].directGPR());
|
|
},
|
|
structureValue, publicLength, butterflyValue);
|
|
} else {
|
|
slowResultValue = lazySlowPath(
|
|
[=, &vm] (const Vector<Location>& locations) -> RefPtr<LazySlowPath::Generator> {
|
|
return createLazyCallGenerator(vm,
|
|
operationNewArrayWithSizeAndHint, locations[0].directGPR(), globalObject,
|
|
locations[1].directGPR(), locations[2].directGPR(), locations[3].directGPR(), locations[4].directGPR());
|
|
},
|
|
structureValue, publicLength, vectorLength, butterflyValue);
|
|
}
|
|
|
|
ValueFromBlock slowResult = m_out.anchor(slowResultValue);
|
|
ValueFromBlock slowButterfly = m_out.anchor(
|
|
m_out.loadPtr(slowResultValue, m_heaps.JSObject_butterfly));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
return ArrayValues(
|
|
m_out.phi(pointerType(), fastResult, slowResult),
|
|
m_out.phi(pointerType(), fastButterfly, slowButterfly));
|
|
}
|
|
|
|
ArrayValues allocateUninitializedContiguousJSArrayInternal(LValue publicLength, LValue vectorLength, RegisteredStructure structure)
|
|
{
|
|
bool shouldInitializeElements = false;
|
|
bool shouldLargeArraySizeCreateArrayStorage = false;
|
|
return allocateJSArray(
|
|
publicLength, vectorLength, weakStructure(structure), m_out.constInt32(structure->indexingType()), shouldInitializeElements,
|
|
shouldLargeArraySizeCreateArrayStorage);
|
|
}
|
|
|
|
ArrayValues allocateUninitializedContiguousJSArray(LValue publicLength, RegisteredStructure structure)
|
|
{
|
|
return allocateUninitializedContiguousJSArrayInternal(publicLength, publicLength, structure);
|
|
}
|
|
|
|
ArrayValues allocateUninitializedContiguousJSArray(unsigned publicLength, unsigned vectorLength, RegisteredStructure structure)
|
|
{
|
|
ASSERT(vectorLength >= publicLength);
|
|
return allocateUninitializedContiguousJSArrayInternal(m_out.constInt32(publicLength), m_out.constInt32(vectorLength), structure);
|
|
}
|
|
|
|
LValue ensureShadowChickenPacket()
|
|
{
|
|
ShadowChicken* shadowChicken = vm().shadowChicken();
|
|
RELEASE_ASSERT(shadowChicken);
|
|
LBasicBlock slowCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
TypedPointer addressOfLogCursor = m_out.absolute(shadowChicken->addressOfLogCursor());
|
|
LValue logCursor = m_out.loadPtr(addressOfLogCursor);
|
|
|
|
ValueFromBlock fastResult = m_out.anchor(logCursor);
|
|
|
|
m_out.branch(
|
|
m_out.below(logCursor, m_out.constIntPtr(shadowChicken->logEnd())),
|
|
usually(continuation), rarely(slowCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(slowCase, continuation);
|
|
|
|
vmCall(Void, operationProcessShadowChickenLog, m_vmValue);
|
|
|
|
ValueFromBlock slowResult = m_out.anchor(m_out.loadPtr(addressOfLogCursor));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
LValue result = m_out.phi(pointerType(), fastResult, slowResult);
|
|
|
|
m_out.storePtr(
|
|
m_out.add(result, m_out.constIntPtr(sizeof(ShadowChicken::Packet))),
|
|
addressOfLogCursor);
|
|
|
|
return result;
|
|
}
|
|
|
|
LValue boolify(Edge edge)
|
|
{
|
|
switch (edge.useKind()) {
|
|
case BooleanUse:
|
|
case KnownBooleanUse:
|
|
return lowBoolean(edge);
|
|
case Int32Use:
|
|
return m_out.notZero32(lowInt32(edge));
|
|
case DoubleRepUse:
|
|
return m_out.doubleNotEqualAndOrdered(lowDouble(edge), m_out.doubleZero);
|
|
case ObjectOrOtherUse:
|
|
return m_out.logicalNot(
|
|
equalNullOrUndefined(
|
|
edge, CellCaseSpeculatesObject, SpeculateNullOrUndefined,
|
|
ManualOperandSpeculation));
|
|
case StringUse:
|
|
return m_out.notEqual(lowString(edge), weakPointer(jsEmptyString(m_graph.m_vm)));
|
|
case StringOrOtherUse: {
|
|
LValue value = lowJSValue(edge, ManualOperandSpeculation);
|
|
|
|
LBasicBlock cellCase = m_out.newBlock();
|
|
LBasicBlock notCellCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(isCell(value, provenType(edge)), unsure(cellCase), unsure(notCellCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(cellCase, notCellCase);
|
|
FTL_TYPE_CHECK(jsValueValue(value), edge, (~SpecCellCheck) | SpecString, isNotString(value));
|
|
ValueFromBlock stringResult = m_out.anchor(m_out.notEqual(value, weakPointer(jsEmptyString(m_graph.m_vm))));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(notCellCase, continuation);
|
|
FTL_TYPE_CHECK(jsValueValue(value), edge, SpecCellCheck | SpecOther, isNotOther(value));
|
|
ValueFromBlock notCellResult = m_out.anchor(m_out.booleanFalse);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
return m_out.phi(Int32, stringResult, notCellResult);
|
|
}
|
|
case UntypedUse: {
|
|
LValue value = lowJSValue(edge);
|
|
|
|
// Implements the following control flow structure:
|
|
// if (value is cell) {
|
|
// if (value is string or value is HeapBigInt)
|
|
// result = !!value->length
|
|
// else {
|
|
// do evil things for masquerades-as-undefined
|
|
// result = true
|
|
// }
|
|
// } else if (value is int32) {
|
|
// result = !!unboxInt32(value)
|
|
// } else if (value is number) {
|
|
// result = !!unboxDouble(value)
|
|
// } else if (value is BigInt32) {
|
|
// result = (value != BigInt32Tag)
|
|
// } else {
|
|
// result = value == jsTrue
|
|
// }
|
|
|
|
LBasicBlock cellCase = m_out.newBlock();
|
|
LBasicBlock notStringCase = m_out.newBlock();
|
|
LBasicBlock stringCase = m_out.newBlock();
|
|
LBasicBlock heapBigIntCase = m_out.newBlock();
|
|
LBasicBlock notStringNorHeapBigIntCase = m_out.newBlock();
|
|
LBasicBlock notCellCase = m_out.newBlock();
|
|
LBasicBlock int32Case = m_out.newBlock();
|
|
LBasicBlock notInt32Case = m_out.newBlock();
|
|
LBasicBlock doubleCase = m_out.newBlock();
|
|
LBasicBlock notDoubleCase = m_out.newBlock();
|
|
#if USE(BIGINT32)
|
|
LBasicBlock bigInt32Case = m_out.newBlock();
|
|
LBasicBlock notBigInt32Case = m_out.newBlock();
|
|
#endif
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
Vector<ValueFromBlock> results;
|
|
|
|
m_out.branch(isCell(value, provenType(edge)), unsure(cellCase), unsure(notCellCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(cellCase, notStringCase);
|
|
m_out.branch(
|
|
isString(value, provenType(edge) & SpecCell),
|
|
unsure(stringCase), unsure(notStringCase));
|
|
|
|
m_out.appendTo(notStringCase, stringCase);
|
|
m_out.branch(
|
|
isHeapBigInt(value, provenType(edge) & (SpecCell - SpecString)),
|
|
unsure(heapBigIntCase), unsure(notStringNorHeapBigIntCase));
|
|
|
|
m_out.appendTo(stringCase, heapBigIntCase);
|
|
results.append(m_out.anchor(m_out.notEqual(value, weakPointer(jsEmptyString(m_graph.m_vm)))));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(heapBigIntCase, notStringNorHeapBigIntCase);
|
|
LValue nonZeroBigInt = m_out.notZero32(
|
|
m_out.load32NonNegative(value, m_heaps.JSBigInt_length));
|
|
results.append(m_out.anchor(nonZeroBigInt));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(notStringNorHeapBigIntCase, notCellCase);
|
|
LValue isTruthyObject;
|
|
if (masqueradesAsUndefinedWatchpointIsStillValid())
|
|
isTruthyObject = m_out.booleanTrue;
|
|
else {
|
|
LBasicBlock masqueradesCase = m_out.newBlock();
|
|
|
|
results.append(m_out.anchor(m_out.booleanTrue));
|
|
|
|
m_out.branch(
|
|
m_out.testIsZero32(
|
|
m_out.load8ZeroExt32(value, m_heaps.JSCell_typeInfoFlags),
|
|
m_out.constInt32(MasqueradesAsUndefined)),
|
|
usually(continuation), rarely(masqueradesCase));
|
|
|
|
m_out.appendTo(masqueradesCase);
|
|
|
|
isTruthyObject = m_out.notEqual(
|
|
weakPointer(m_graph.globalObjectFor(m_origin.semantic)),
|
|
m_out.loadPtr(loadStructure(value), m_heaps.Structure_globalObject));
|
|
}
|
|
results.append(m_out.anchor(isTruthyObject));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(notCellCase, int32Case);
|
|
m_out.branch(
|
|
isInt32(value, provenType(edge) & ~SpecCell),
|
|
unsure(int32Case), unsure(notInt32Case));
|
|
|
|
m_out.appendTo(int32Case, notInt32Case);
|
|
results.append(m_out.anchor(m_out.notZero32(unboxInt32(value))));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(notInt32Case, doubleCase);
|
|
m_out.branch(
|
|
isNumber(value, provenType(edge) & ~SpecCell),
|
|
unsure(doubleCase), unsure(notDoubleCase));
|
|
|
|
m_out.appendTo(doubleCase, notDoubleCase);
|
|
LValue doubleIsTruthy = m_out.doubleNotEqualAndOrdered(
|
|
unboxDouble(value), m_out.constDouble(0));
|
|
results.append(m_out.anchor(doubleIsTruthy));
|
|
m_out.jump(continuation);
|
|
|
|
#if USE(BIGINT32)
|
|
m_out.appendTo(notDoubleCase, bigInt32Case);
|
|
m_out.branch(
|
|
isBigInt32(value, provenType(edge) & ~SpecCell),
|
|
unsure(bigInt32Case), unsure(notBigInt32Case));
|
|
|
|
m_out.appendTo(bigInt32Case, notBigInt32Case);
|
|
LValue bigInt32NotZero = m_out.notEqual(value, m_out.constInt64(JSValue::BigInt32Tag));
|
|
results.append(m_out.anchor(bigInt32NotZero));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(notBigInt32Case, continuation);
|
|
#else
|
|
m_out.appendTo(notDoubleCase, continuation);
|
|
#endif
|
|
LValue miscIsTruthy = m_out.equal(
|
|
value, m_out.constInt64(JSValue::encode(jsBoolean(true))));
|
|
results.append(m_out.anchor(miscIsTruthy));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
return m_out.phi(Int32, results);
|
|
}
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
return nullptr;
|
|
}
|
|
}
|
|
|
|
enum StringOrObjectMode {
|
|
AllCellsAreFalse,
|
|
CellCaseSpeculatesObject
|
|
};
|
|
enum EqualNullOrUndefinedMode {
|
|
EqualNull,
|
|
EqualUndefined,
|
|
EqualNullOrUndefined,
|
|
SpeculateNullOrUndefined
|
|
};
|
|
LValue equalNullOrUndefined(
|
|
Edge edge, StringOrObjectMode cellMode, EqualNullOrUndefinedMode primitiveMode,
|
|
OperandSpeculationMode operandMode = AutomaticOperandSpeculation)
|
|
{
|
|
bool validWatchpoint = masqueradesAsUndefinedWatchpointIsStillValid();
|
|
|
|
LValue value = lowJSValue(edge, operandMode);
|
|
|
|
LBasicBlock cellCase = m_out.newBlock();
|
|
LBasicBlock primitiveCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(isNotCell(value, provenType(edge)), unsure(primitiveCase), unsure(cellCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(cellCase, primitiveCase);
|
|
|
|
Vector<ValueFromBlock, 3> results;
|
|
|
|
switch (cellMode) {
|
|
case AllCellsAreFalse:
|
|
break;
|
|
case CellCaseSpeculatesObject:
|
|
FTL_TYPE_CHECK(
|
|
jsValueValue(value), edge, (~SpecCellCheck) | SpecObject, isNotObject(value));
|
|
break;
|
|
}
|
|
|
|
if (validWatchpoint) {
|
|
results.append(m_out.anchor(m_out.booleanFalse));
|
|
m_out.jump(continuation);
|
|
} else {
|
|
LBasicBlock masqueradesCase =
|
|
m_out.newBlock();
|
|
|
|
results.append(m_out.anchor(m_out.booleanFalse));
|
|
|
|
m_out.branch(
|
|
m_out.testNonZero32(
|
|
m_out.load8ZeroExt32(value, m_heaps.JSCell_typeInfoFlags),
|
|
m_out.constInt32(MasqueradesAsUndefined)),
|
|
rarely(masqueradesCase), usually(continuation));
|
|
|
|
m_out.appendTo(masqueradesCase, primitiveCase);
|
|
|
|
LValue structure = loadStructure(value);
|
|
|
|
results.append(m_out.anchor(
|
|
m_out.equal(
|
|
weakPointer(m_graph.globalObjectFor(m_origin.semantic)),
|
|
m_out.loadPtr(structure, m_heaps.Structure_globalObject))));
|
|
m_out.jump(continuation);
|
|
}
|
|
|
|
m_out.appendTo(primitiveCase, continuation);
|
|
|
|
LValue primitiveResult;
|
|
switch (primitiveMode) {
|
|
case EqualNull:
|
|
primitiveResult = m_out.equal(value, m_out.constInt64(JSValue::ValueNull));
|
|
break;
|
|
case EqualUndefined:
|
|
primitiveResult = m_out.equal(value, m_out.constInt64(JSValue::ValueUndefined));
|
|
break;
|
|
case EqualNullOrUndefined:
|
|
primitiveResult = isOther(value, provenType(edge));
|
|
break;
|
|
case SpeculateNullOrUndefined:
|
|
FTL_TYPE_CHECK(
|
|
jsValueValue(value), edge, SpecCellCheck | SpecOther, isNotOther(value));
|
|
primitiveResult = m_out.booleanTrue;
|
|
break;
|
|
}
|
|
results.append(m_out.anchor(primitiveResult));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
|
|
return m_out.phi(Int32, results);
|
|
}
|
|
|
|
template<typename FunctionType>
|
|
void contiguousPutByValOutOfBounds(
|
|
FunctionType slowPathFunction, LValue base, LValue storage, LValue index, LValue value,
|
|
LBasicBlock continuation)
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
if (!m_node->arrayMode().isInBounds()) {
|
|
LBasicBlock notInBoundsCase =
|
|
m_out.newBlock();
|
|
LBasicBlock performStore =
|
|
m_out.newBlock();
|
|
|
|
LValue isNotInBounds = m_out.aboveOrEqual(
|
|
index, m_out.load32NonNegative(storage, m_heaps.Butterfly_publicLength));
|
|
m_out.branch(isNotInBounds, unsure(notInBoundsCase), unsure(performStore));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(notInBoundsCase, performStore);
|
|
|
|
LValue isOutOfBounds = m_out.aboveOrEqual(
|
|
index, m_out.load32NonNegative(storage, m_heaps.Butterfly_vectorLength));
|
|
|
|
if (!m_node->arrayMode().isOutOfBounds())
|
|
speculate(OutOfBounds, noValue(), nullptr, isOutOfBounds);
|
|
else {
|
|
LBasicBlock outOfBoundsCase =
|
|
m_out.newBlock();
|
|
LBasicBlock holeCase =
|
|
m_out.newBlock();
|
|
|
|
m_out.branch(isOutOfBounds, rarely(outOfBoundsCase), usually(holeCase));
|
|
|
|
LBasicBlock innerLastNext = m_out.appendTo(outOfBoundsCase, holeCase);
|
|
|
|
vmCall(
|
|
Void, slowPathFunction,
|
|
weakPointer(globalObject), base, index, value);
|
|
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(holeCase, innerLastNext);
|
|
}
|
|
|
|
m_out.store32(
|
|
m_out.add(index, m_out.int32One),
|
|
storage, m_heaps.Butterfly_publicLength);
|
|
|
|
m_out.jump(performStore);
|
|
m_out.appendTo(performStore, lastNext);
|
|
}
|
|
}
|
|
|
|
LValue untagArrayPtr(LValue ptr, LValue size)
|
|
{
|
|
#if CPU(ARM64E)
|
|
PatchpointValue* authenticate = m_out.patchpoint(pointerType());
|
|
authenticate->appendSomeRegister(ptr);
|
|
authenticate->append(size, B3::ValueRep(B3::ValueRep::SomeLateRegister));
|
|
authenticate->setGenerator([=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
jit.move(params[1].gpr(), params[0].gpr());
|
|
jit.untagArrayPtr(params[2].gpr(), params[0].gpr());
|
|
});
|
|
return authenticate;
|
|
#else
|
|
UNUSED_PARAM(size);
|
|
return ptr;
|
|
#endif
|
|
}
|
|
|
|
LValue removeArrayPtrTag(LValue ptr)
|
|
{
|
|
#if CPU(ARM64E)
|
|
PatchpointValue* authenticate = m_out.patchpoint(pointerType());
|
|
authenticate->appendSomeRegister(ptr);
|
|
authenticate->setGenerator([=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
jit.move(params[1].gpr(), params[0].gpr());
|
|
jit.removeArrayPtrTag(params[0].gpr());
|
|
});
|
|
return authenticate;
|
|
#endif
|
|
return ptr;
|
|
}
|
|
|
|
LValue caged(Gigacage::Kind kind, LValue ptr, LValue base)
|
|
{
|
|
auto doUntagArrayPtr = [&](LValue taggedPtr) {
|
|
#if CPU(ARM64E)
|
|
if (kind == Gigacage::Primitive) {
|
|
LValue size = m_out.load32(base, m_heaps.JSArrayBufferView_length);
|
|
return untagArrayPtr(taggedPtr, size);
|
|
}
|
|
return ptr;
|
|
#else
|
|
UNUSED_PARAM(taggedPtr);
|
|
return ptr;
|
|
#endif
|
|
};
|
|
|
|
#if GIGACAGE_ENABLED
|
|
if (!Gigacage::isEnabled(kind))
|
|
return doUntagArrayPtr(ptr);
|
|
|
|
if (kind == Gigacage::Primitive && !Gigacage::disablingPrimitiveGigacageIsForbidden()) {
|
|
if (vm().primitiveGigacageEnabled().isStillValid())
|
|
m_graph.watchpoints().addLazily(vm().primitiveGigacageEnabled());
|
|
else
|
|
return doUntagArrayPtr(ptr);
|
|
}
|
|
|
|
LValue basePtr = m_out.constIntPtr(Gigacage::basePtr(kind));
|
|
LValue mask = m_out.constIntPtr(Gigacage::mask(kind));
|
|
|
|
LValue masked = m_out.bitAnd(ptr, mask);
|
|
LValue result = m_out.add(masked, basePtr);
|
|
|
|
#if CPU(ARM64E)
|
|
if (kind == Gigacage::Primitive) {
|
|
PatchpointValue* merge = m_out.patchpoint(pointerType());
|
|
merge->append(result, B3::ValueRep(B3::ValueRep::SomeLateRegister));
|
|
merge->appendSomeRegister(ptr);
|
|
merge->setGenerator([=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
jit.move(params[2].gpr(), params[0].gpr());
|
|
jit.bitFieldInsert64(params[1].gpr(), 0, 64 - MacroAssembler::numberOfPACBits, params[0].gpr());
|
|
});
|
|
|
|
result = doUntagArrayPtr(merge);
|
|
}
|
|
#endif // CPU(ARM64E)
|
|
|
|
// Make sure that B3 doesn't try to do smart reassociation of these pointer bits.
|
|
// FIXME: In an ideal world, B3 would not do harmful reassociations, and if it did, it would be able
|
|
// to undo them during constant hoisting and regalloc. As it stands, if you remove this then Octane
|
|
// gets 1.6% slower and Kraken gets 5% slower. It's all because the basePtr, which is a constant,
|
|
// gets reassociated out of the add above and into the address arithmetic. This disables hoisting of
|
|
// the basePtr constant. Hoisting that constant is worth a lot more perf than the reassociation. One
|
|
// way to make this all work happily is to combine offset legalization with constant hoisting, and
|
|
// then teach it reassociation. So, Add(Add(a, b), const) where a is loop-invariant while b isn't
|
|
// will turn into Add(Add(a, const), b) by the constant hoister. We would have to teach B3 to do this
|
|
// and possibly other smart things if we want to be able to remove this opaque.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=175493
|
|
return m_out.opaque(result);
|
|
#endif
|
|
|
|
UNUSED_PARAM(kind);
|
|
UNUSED_PARAM(base);
|
|
return doUntagArrayPtr(ptr);
|
|
}
|
|
|
|
void buildSwitch(SwitchData* data, LType type, LValue switchValue)
|
|
{
|
|
ASSERT(type == pointerType() || type == Int32);
|
|
|
|
Vector<SwitchCase> cases;
|
|
for (unsigned i = 0; i < data->cases.size(); ++i) {
|
|
SwitchCase newCase;
|
|
|
|
if (type == pointerType()) {
|
|
newCase = SwitchCase(m_out.constIntPtr(data->cases[i].value.switchLookupValue(data->kind)),
|
|
lowBlock(data->cases[i].target.block), Weight(data->cases[i].target.count));
|
|
} else if (type == Int32) {
|
|
newCase = SwitchCase(m_out.constInt32(data->cases[i].value.switchLookupValue(data->kind)),
|
|
lowBlock(data->cases[i].target.block), Weight(data->cases[i].target.count));
|
|
} else
|
|
CRASH();
|
|
|
|
cases.append(newCase);
|
|
}
|
|
|
|
m_out.switchInstruction(
|
|
switchValue, cases,
|
|
lowBlock(data->fallThrough.block), Weight(data->fallThrough.count));
|
|
}
|
|
|
|
void switchString(SwitchData* data, LValue string, Edge& edge)
|
|
{
|
|
bool canDoBinarySwitch = true;
|
|
unsigned totalLength = 0;
|
|
|
|
for (DFG::SwitchCase myCase : data->cases) {
|
|
StringImpl* string = myCase.value.stringImpl();
|
|
if (!string->is8Bit()) {
|
|
canDoBinarySwitch = false;
|
|
break;
|
|
}
|
|
if (string->length() > Options::maximumBinaryStringSwitchCaseLength()) {
|
|
canDoBinarySwitch = false;
|
|
break;
|
|
}
|
|
totalLength += string->length();
|
|
}
|
|
|
|
if (!canDoBinarySwitch || totalLength > Options::maximumBinaryStringSwitchTotalLength()) {
|
|
switchStringSlow(data, string);
|
|
return;
|
|
}
|
|
|
|
LBasicBlock hasImplBlock = m_out.newBlock();
|
|
LBasicBlock is8BitBlock = m_out.newBlock();
|
|
LBasicBlock slowBlock = m_out.newBlock();
|
|
|
|
m_out.branch(isRopeString(string, edge), unsure(slowBlock), unsure(hasImplBlock));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(hasImplBlock, is8BitBlock);
|
|
|
|
LValue stringImpl = m_out.loadPtr(string, m_heaps.JSString_value);
|
|
LValue length = m_out.load32(stringImpl, m_heaps.StringImpl_length);
|
|
|
|
m_out.branch(
|
|
m_out.testIsZero32(
|
|
m_out.load32(stringImpl, m_heaps.StringImpl_hashAndFlags),
|
|
m_out.constInt32(StringImpl::flagIs8Bit())),
|
|
unsure(slowBlock), unsure(is8BitBlock));
|
|
|
|
m_out.appendTo(is8BitBlock, slowBlock);
|
|
|
|
LValue buffer = m_out.loadPtr(stringImpl, m_heaps.StringImpl_data);
|
|
|
|
// FIXME: We should propagate branch weight data to the cases of this switch.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=144368
|
|
|
|
Vector<StringSwitchCase> cases;
|
|
for (DFG::SwitchCase myCase : data->cases)
|
|
cases.append(StringSwitchCase(myCase.value.stringImpl(), lowBlock(myCase.target.block)));
|
|
std::sort(cases.begin(), cases.end());
|
|
switchStringRecurse(data, buffer, length, cases, 0, 0, cases.size(), 0, false);
|
|
|
|
m_out.appendTo(slowBlock, lastNext);
|
|
switchStringSlow(data, string);
|
|
}
|
|
|
|
// The code for string switching is based closely on the same code in the DFG backend. While it
|
|
// would be nice to reduce the amount of similar-looking code, it seems like this is one of
|
|
// those algorithms where factoring out the common bits would result in more code than just
|
|
// duplicating.
|
|
|
|
struct StringSwitchCase {
|
|
StringSwitchCase() { }
|
|
|
|
StringSwitchCase(StringImpl* string, LBasicBlock target)
|
|
: string(string)
|
|
, target(target)
|
|
{
|
|
}
|
|
|
|
bool operator<(const StringSwitchCase& other) const
|
|
{
|
|
return stringLessThan(*string, *other.string);
|
|
}
|
|
|
|
StringImpl* string;
|
|
LBasicBlock target;
|
|
};
|
|
|
|
struct CharacterCase {
|
|
CharacterCase()
|
|
: character(0)
|
|
, begin(0)
|
|
, end(0)
|
|
{
|
|
}
|
|
|
|
CharacterCase(LChar character, unsigned begin, unsigned end)
|
|
: character(character)
|
|
, begin(begin)
|
|
, end(end)
|
|
{
|
|
}
|
|
|
|
bool operator<(const CharacterCase& other) const
|
|
{
|
|
return character < other.character;
|
|
}
|
|
|
|
LChar character;
|
|
unsigned begin;
|
|
unsigned end;
|
|
};
|
|
|
|
void switchStringRecurse(
|
|
SwitchData* data, LValue buffer, LValue length, const Vector<StringSwitchCase>& cases,
|
|
unsigned numChecked, unsigned begin, unsigned end, unsigned alreadyCheckedLength,
|
|
unsigned checkedExactLength)
|
|
{
|
|
LBasicBlock fallThrough = lowBlock(data->fallThrough.block);
|
|
|
|
if (begin == end) {
|
|
m_out.jump(fallThrough);
|
|
return;
|
|
}
|
|
|
|
unsigned minLength = cases[begin].string->length();
|
|
unsigned commonChars = minLength;
|
|
bool allLengthsEqual = true;
|
|
for (unsigned i = begin + 1; i < end; ++i) {
|
|
unsigned myCommonChars = numChecked;
|
|
unsigned limit = std::min(cases[begin].string->length(), cases[i].string->length());
|
|
for (unsigned j = numChecked; j < limit; ++j) {
|
|
if (cases[begin].string->at(j) != cases[i].string->at(j))
|
|
break;
|
|
myCommonChars++;
|
|
}
|
|
commonChars = std::min(commonChars, myCommonChars);
|
|
if (minLength != cases[i].string->length())
|
|
allLengthsEqual = false;
|
|
minLength = std::min(minLength, cases[i].string->length());
|
|
}
|
|
|
|
if (checkedExactLength) {
|
|
DFG_ASSERT(m_graph, m_node, alreadyCheckedLength == minLength, alreadyCheckedLength, minLength);
|
|
DFG_ASSERT(m_graph, m_node, allLengthsEqual);
|
|
}
|
|
|
|
DFG_ASSERT(m_graph, m_node, minLength >= commonChars, minLength, commonChars);
|
|
|
|
if (!allLengthsEqual && alreadyCheckedLength < minLength)
|
|
m_out.check(m_out.below(length, m_out.constInt32(minLength)), unsure(fallThrough));
|
|
if (allLengthsEqual && (alreadyCheckedLength < minLength || !checkedExactLength))
|
|
m_out.check(m_out.notEqual(length, m_out.constInt32(minLength)), unsure(fallThrough));
|
|
|
|
for (unsigned i = numChecked; i < commonChars; ++i) {
|
|
m_out.check(
|
|
m_out.notEqual(
|
|
m_out.load8ZeroExt32(buffer, m_heaps.characters8[i]),
|
|
m_out.constInt32(static_cast<uint16_t>(cases[begin].string->at(i)))),
|
|
unsure(fallThrough));
|
|
}
|
|
|
|
if (minLength == commonChars) {
|
|
// This is the case where one of the cases is a prefix of all of the other cases.
|
|
// We've already checked that the input string is a prefix of all of the cases,
|
|
// so we just check length to jump to that case.
|
|
|
|
DFG_ASSERT(m_graph, m_node, cases[begin].string->length() == commonChars, cases[begin].string->length(), commonChars);
|
|
for (unsigned i = begin + 1; i < end; ++i)
|
|
DFG_ASSERT(m_graph, m_node, cases[i].string->length() > commonChars, cases[i].string->length(), commonChars);
|
|
|
|
if (allLengthsEqual) {
|
|
DFG_ASSERT(m_graph, m_node, end == begin + 1, end, begin);
|
|
m_out.jump(cases[begin].target);
|
|
return;
|
|
}
|
|
|
|
m_out.check(
|
|
m_out.equal(length, m_out.constInt32(commonChars)),
|
|
unsure(cases[begin].target));
|
|
|
|
// We've checked if the length is >= minLength, and then we checked if the length is
|
|
// == commonChars. We get to this point if it is >= minLength but not == commonChars.
|
|
// Hence we know that it now must be > minLength, i.e. that it's >= minLength + 1.
|
|
switchStringRecurse(
|
|
data, buffer, length, cases, commonChars, begin + 1, end, minLength + 1, false);
|
|
return;
|
|
}
|
|
|
|
// At this point we know that the string is longer than commonChars, and we've only verified
|
|
// commonChars. Use a binary switch on the next unchecked character, i.e.
|
|
// string[commonChars].
|
|
|
|
DFG_ASSERT(m_graph, m_node, end >= begin + 2, end, begin);
|
|
|
|
LValue uncheckedChar = m_out.load8ZeroExt32(buffer, m_heaps.characters8[commonChars]);
|
|
|
|
Vector<CharacterCase> characterCases;
|
|
CharacterCase currentCase(cases[begin].string->at(commonChars), begin, begin + 1);
|
|
for (unsigned i = begin + 1; i < end; ++i) {
|
|
LChar currentChar = cases[i].string->at(commonChars);
|
|
if (currentChar != currentCase.character) {
|
|
currentCase.end = i;
|
|
characterCases.append(currentCase);
|
|
currentCase = CharacterCase(currentChar, i, i + 1);
|
|
} else
|
|
currentCase.end = i + 1;
|
|
}
|
|
characterCases.append(currentCase);
|
|
|
|
Vector<LBasicBlock> characterBlocks;
|
|
for (unsigned i = characterCases.size(); i--;)
|
|
characterBlocks.append(m_out.newBlock());
|
|
|
|
Vector<SwitchCase> switchCases;
|
|
for (unsigned i = 0; i < characterCases.size(); ++i) {
|
|
if (i)
|
|
DFG_ASSERT(m_graph, m_node, characterCases[i - 1].character < characterCases[i].character);
|
|
switchCases.append(SwitchCase(
|
|
m_out.constInt32(characterCases[i].character), characterBlocks[i], Weight()));
|
|
}
|
|
m_out.switchInstruction(uncheckedChar, switchCases, fallThrough, Weight());
|
|
|
|
LBasicBlock lastNext = m_out.m_nextBlock;
|
|
characterBlocks.append(lastNext); // Makes it convenient to set nextBlock.
|
|
for (unsigned i = 0; i < characterCases.size(); ++i) {
|
|
m_out.appendTo(characterBlocks[i], characterBlocks[i + 1]);
|
|
switchStringRecurse(
|
|
data, buffer, length, cases, commonChars + 1,
|
|
characterCases[i].begin, characterCases[i].end, minLength, allLengthsEqual);
|
|
}
|
|
|
|
DFG_ASSERT(m_graph, m_node, m_out.m_nextBlock == lastNext);
|
|
}
|
|
|
|
void switchStringSlow(SwitchData* data, LValue string)
|
|
{
|
|
// FIXME: We ought to be able to use computed gotos here. We would save the labels of the
|
|
// blocks we want to jump to, and then request their addresses after compilation completes.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=144369
|
|
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
|
|
LValue branchOffset = vmCall(
|
|
Int32, operationSwitchStringAndGetBranchOffset,
|
|
weakPointer(globalObject), m_out.constIntPtr(data->switchTableIndex), string);
|
|
|
|
StringJumpTable& table = codeBlock()->stringSwitchJumpTable(data->switchTableIndex);
|
|
|
|
Vector<SwitchCase> cases;
|
|
// These may be negative, or zero, or probably other stuff, too. We don't want to mess with HashSet's corner cases and we don't really care about throughput here.
|
|
StdUnorderedSet<int32_t> alreadyHandled;
|
|
for (unsigned i = 0; i < data->cases.size(); ++i) {
|
|
// FIXME: The fact that we're using the bytecode's switch table means that the
|
|
// following DFG IR transformation would be invalid.
|
|
//
|
|
// Original code:
|
|
// switch (v) {
|
|
// case "foo":
|
|
// case "bar":
|
|
// things();
|
|
// break;
|
|
// default:
|
|
// break;
|
|
// }
|
|
//
|
|
// New code:
|
|
// switch (v) {
|
|
// case "foo":
|
|
// instrumentFoo();
|
|
// goto _things;
|
|
// case "bar":
|
|
// instrumentBar();
|
|
// _things:
|
|
// things();
|
|
// break;
|
|
// default:
|
|
// break;
|
|
// }
|
|
//
|
|
// Luckily, we don't currently do any such transformation. But it's kind of silly that
|
|
// this is an issue.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=144635
|
|
|
|
DFG::SwitchCase myCase = data->cases[i];
|
|
StringJumpTable::StringOffsetTable::iterator iter =
|
|
table.offsetTable.find(myCase.value.stringImpl());
|
|
DFG_ASSERT(m_graph, m_node, iter != table.offsetTable.end());
|
|
|
|
if (!alreadyHandled.insert(iter->value.branchOffset).second)
|
|
continue;
|
|
|
|
cases.append(SwitchCase(
|
|
m_out.constInt32(iter->value.branchOffset),
|
|
lowBlock(myCase.target.block), Weight(myCase.target.count)));
|
|
}
|
|
|
|
m_out.switchInstruction(
|
|
branchOffset, cases, lowBlock(data->fallThrough.block),
|
|
Weight(data->fallThrough.count));
|
|
}
|
|
|
|
// Calls the functor at the point of code generation where we know what the result type is.
|
|
// You can emit whatever code you like at that point. Expects you to terminate the basic block.
|
|
// When buildTypeOf() returns, it will have terminated all basic blocks that it created. So, if
|
|
// you aren't using this as the terminator of a high-level block, you should create your own
|
|
// contination and set it as the nextBlock (m_out.insertNewBlocksBefore(continuation)) before
|
|
// calling this. For example:
|
|
//
|
|
// LBasicBlock continuation = m_out.newBlock();
|
|
// LBasicBlock lastNext = m_out.insertNewBlocksBefore(continuation);
|
|
// buildTypeOf(
|
|
// child, value,
|
|
// [&] (TypeofType type) {
|
|
// do things;
|
|
// m_out.jump(continuation);
|
|
// });
|
|
// m_out.appendTo(continuation, lastNext);
|
|
template<typename Functor>
|
|
void buildTypeOf(Edge child, LValue value, const Functor& functor)
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
|
|
// Implements the following branching structure:
|
|
//
|
|
// if (is cell) {
|
|
// if (is object) {
|
|
// if (is function) {
|
|
// return function;
|
|
// } else if (doesn't have call trap and doesn't masquerade as undefined) {
|
|
// return object
|
|
// } else {
|
|
// return slowPath();
|
|
// }
|
|
// } else if (is string) {
|
|
// return string
|
|
// } else if (is heapbigint) {
|
|
// return bigint
|
|
// } else {
|
|
// return symbol
|
|
// }
|
|
// } else if (is number) {
|
|
// return number
|
|
// } else if (is bigint32) {
|
|
// return bigint
|
|
// } else if (is null) {
|
|
// return object
|
|
// } else if (is boolean) {
|
|
// return boolean
|
|
// } else {
|
|
// return undefined
|
|
// }
|
|
//
|
|
// FIXME: typeof Symbol should be more frequently seen than BigInt.
|
|
// We should change the order of type detection based on this frequency.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=192650
|
|
|
|
LBasicBlock cellCase = m_out.newBlock();
|
|
LBasicBlock objectCase = m_out.newBlock();
|
|
LBasicBlock functionCase = m_out.newBlock();
|
|
LBasicBlock notFunctionCase = m_out.newBlock();
|
|
LBasicBlock reallyObjectCase = m_out.newBlock();
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock unreachable = m_out.newBlock();
|
|
LBasicBlock notObjectCase = m_out.newBlock();
|
|
LBasicBlock stringCase = m_out.newBlock();
|
|
LBasicBlock notStringCase = m_out.newBlock();
|
|
LBasicBlock bigIntCase = m_out.newBlock();
|
|
LBasicBlock symbolCase = m_out.newBlock();
|
|
LBasicBlock notCellCase = m_out.newBlock();
|
|
LBasicBlock numberCase = m_out.newBlock();
|
|
LBasicBlock notNumberCase = m_out.newBlock();
|
|
#if USE(BIGINT32)
|
|
LBasicBlock notBigInt32Case = m_out.newBlock();
|
|
#endif
|
|
LBasicBlock notNullCase = m_out.newBlock();
|
|
LBasicBlock booleanCase = m_out.newBlock();
|
|
LBasicBlock undefinedCase = m_out.newBlock();
|
|
|
|
m_out.branch(isCell(value, provenType(child)), unsure(cellCase), unsure(notCellCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(cellCase, objectCase);
|
|
m_out.branch(isObject(value, provenType(child)), unsure(objectCase), unsure(notObjectCase));
|
|
|
|
m_out.appendTo(objectCase, functionCase);
|
|
m_out.branch(
|
|
isFunction(value, provenType(child) & SpecObject),
|
|
unsure(functionCase), unsure(notFunctionCase));
|
|
|
|
m_out.appendTo(functionCase, notFunctionCase);
|
|
functor(TypeofType::Function);
|
|
|
|
m_out.appendTo(notFunctionCase, reallyObjectCase);
|
|
m_out.branch(
|
|
isExoticForTypeof(value, provenType(child) & (SpecObject - SpecFunction)),
|
|
rarely(slowPath), usually(reallyObjectCase));
|
|
|
|
m_out.appendTo(reallyObjectCase, slowPath);
|
|
functor(TypeofType::Object);
|
|
|
|
m_out.appendTo(slowPath, unreachable);
|
|
VM& vm = this->vm();
|
|
LValue result = lazySlowPath(
|
|
[=, &vm] (const Vector<Location>& locations) -> RefPtr<LazySlowPath::Generator> {
|
|
return createLazyCallGenerator(vm,
|
|
operationTypeOfObjectAsTypeofType, locations[0].directGPR(),
|
|
CCallHelpers::TrustedImmPtr(globalObject), locations[1].directGPR());
|
|
}, value);
|
|
Vector<SwitchCase, 3> cases;
|
|
cases.append(SwitchCase(m_out.constInt32(static_cast<int32_t>(TypeofType::Undefined)), undefinedCase));
|
|
cases.append(SwitchCase(m_out.constInt32(static_cast<int32_t>(TypeofType::Object)), reallyObjectCase));
|
|
cases.append(SwitchCase(m_out.constInt32(static_cast<int32_t>(TypeofType::Function)), functionCase));
|
|
m_out.switchInstruction(m_out.castToInt32(result), cases, unreachable, Weight());
|
|
|
|
m_out.appendTo(unreachable, notObjectCase);
|
|
m_out.unreachable();
|
|
|
|
m_out.appendTo(notObjectCase, stringCase);
|
|
m_out.branch(
|
|
isString(value, provenType(child) & (SpecCell - SpecObject)),
|
|
unsure(stringCase), unsure(notStringCase));
|
|
|
|
m_out.appendTo(stringCase, notStringCase);
|
|
functor(TypeofType::String);
|
|
|
|
m_out.appendTo(notStringCase, bigIntCase);
|
|
m_out.branch(
|
|
isHeapBigInt(value, provenType(child) & (SpecCell - SpecObject - SpecString)),
|
|
unsure(bigIntCase), unsure(symbolCase));
|
|
|
|
m_out.appendTo(bigIntCase, symbolCase);
|
|
functor(TypeofType::BigInt);
|
|
|
|
m_out.appendTo(symbolCase, notCellCase);
|
|
functor(TypeofType::Symbol);
|
|
|
|
m_out.appendTo(notCellCase, numberCase);
|
|
m_out.branch(
|
|
isNumber(value, provenType(child) & ~SpecCell),
|
|
unsure(numberCase), unsure(notNumberCase));
|
|
|
|
m_out.appendTo(numberCase, notNumberCase);
|
|
functor(TypeofType::Number);
|
|
|
|
#if USE(BIGINT32)
|
|
m_out.appendTo(notNumberCase, notBigInt32Case);
|
|
m_out.branch(isBigInt32(value, provenType(child) & ~SpecCell), unsure(bigIntCase), unsure(notBigInt32Case));
|
|
|
|
m_out.appendTo(notBigInt32Case, notNullCase);
|
|
#else
|
|
m_out.appendTo(notNumberCase, notNullCase);
|
|
#endif
|
|
LValue isNull;
|
|
if (provenType(child) & SpecOther)
|
|
isNull = m_out.equal(value, m_out.constInt64(JSValue::ValueNull));
|
|
else
|
|
isNull = m_out.booleanFalse;
|
|
m_out.branch(isNull, unsure(reallyObjectCase), unsure(notNullCase));
|
|
|
|
m_out.appendTo(notNullCase, booleanCase);
|
|
m_out.branch(
|
|
isBoolean(value, provenType(child) & ~(SpecCell | SpecFullNumber)),
|
|
unsure(booleanCase), unsure(undefinedCase));
|
|
|
|
m_out.appendTo(booleanCase, undefinedCase);
|
|
functor(TypeofType::Boolean);
|
|
|
|
m_out.appendTo(undefinedCase, lastNext);
|
|
functor(TypeofType::Undefined);
|
|
}
|
|
|
|
TypedPointer pointerIntoTypedArray(LValue storage, LValue index, TypedArrayType type)
|
|
{
|
|
LValue offset = m_out.shl(m_out.zeroExtPtr(index), m_out.constIntPtr(logElementSize(type)));
|
|
|
|
return TypedPointer(
|
|
m_heaps.typedArrayProperties,
|
|
m_out.add(
|
|
storage,
|
|
offset
|
|
));
|
|
}
|
|
|
|
LValue loadFromIntTypedArray(TypedPointer pointer, TypedArrayType type)
|
|
{
|
|
switch (elementSize(type)) {
|
|
case 1:
|
|
return isSigned(type) ? m_out.load8SignExt32(pointer) : m_out.load8ZeroExt32(pointer);
|
|
case 2:
|
|
return isSigned(type) ? m_out.load16SignExt32(pointer) : m_out.load16ZeroExt32(pointer);
|
|
case 4:
|
|
return m_out.load32(pointer);
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad element size");
|
|
}
|
|
}
|
|
|
|
Output::StoreType storeType(TypedArrayType type)
|
|
{
|
|
if (isInt(type)) {
|
|
switch (elementSize(type)) {
|
|
case 1:
|
|
return Output::Store32As8;
|
|
case 2:
|
|
return Output::Store32As16;
|
|
case 4:
|
|
return Output::Store32;
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad element size");
|
|
return Output::Store32;
|
|
}
|
|
}
|
|
switch (type) {
|
|
case TypeFloat32:
|
|
return Output::StoreFloat;
|
|
case TypeFloat64:
|
|
return Output::StoreDouble;
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad typed array type");
|
|
}
|
|
}
|
|
|
|
void setIntTypedArrayLoadResult(LValue result, TypedArrayType type, bool canSpeculate = false)
|
|
{
|
|
if (elementSize(type) < 4 || isSigned(type)) {
|
|
setInt32(result);
|
|
return;
|
|
}
|
|
|
|
if (m_node->shouldSpeculateInt32() && canSpeculate) {
|
|
speculate(
|
|
Overflow, noValue(), nullptr, m_out.lessThan(result, m_out.int32Zero));
|
|
setInt32(result);
|
|
return;
|
|
}
|
|
|
|
if (m_node->shouldSpeculateInt52()) {
|
|
setStrictInt52(m_out.zeroExt(result, Int64));
|
|
return;
|
|
}
|
|
|
|
setDouble(m_out.unsignedToDouble(result));
|
|
}
|
|
|
|
LValue getIntTypedArrayStoreOperand(Edge edge, bool isClamped = false)
|
|
{
|
|
LValue intValue;
|
|
switch (edge.useKind()) {
|
|
case Int52RepUse:
|
|
case Int32Use: {
|
|
if (edge.useKind() == Int32Use)
|
|
intValue = lowInt32(edge);
|
|
else
|
|
intValue = m_out.castToInt32(lowStrictInt52(edge));
|
|
|
|
if (isClamped) {
|
|
LBasicBlock atLeastZero = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
Vector<ValueFromBlock, 2> intValues;
|
|
intValues.append(m_out.anchor(m_out.int32Zero));
|
|
m_out.branch(
|
|
m_out.lessThan(intValue, m_out.int32Zero),
|
|
unsure(continuation), unsure(atLeastZero));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(atLeastZero, continuation);
|
|
|
|
intValues.append(m_out.anchor(m_out.select(
|
|
m_out.greaterThan(intValue, m_out.constInt32(255)),
|
|
m_out.constInt32(255),
|
|
intValue)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
intValue = m_out.phi(Int32, intValues);
|
|
}
|
|
break;
|
|
}
|
|
|
|
case DoubleRepUse: {
|
|
LValue doubleValue = lowDouble(edge);
|
|
|
|
if (isClamped) {
|
|
LBasicBlock atLeastZero = m_out.newBlock();
|
|
LBasicBlock withinRange = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
Vector<ValueFromBlock, 3> intValues;
|
|
intValues.append(m_out.anchor(m_out.int32Zero));
|
|
m_out.branch(
|
|
m_out.doubleLessThanOrUnordered(doubleValue, m_out.doubleZero),
|
|
unsure(continuation), unsure(atLeastZero));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(atLeastZero, withinRange);
|
|
intValues.append(m_out.anchor(m_out.constInt32(255)));
|
|
m_out.branch(
|
|
m_out.doubleGreaterThan(doubleValue, m_out.constDouble(255)),
|
|
unsure(continuation), unsure(withinRange));
|
|
|
|
m_out.appendTo(withinRange, continuation);
|
|
intValues.append(m_out.anchor(m_out.doubleToInt(doubleValue)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
intValue = m_out.phi(Int32, intValues);
|
|
} else
|
|
intValue = doubleToInt32(doubleValue);
|
|
break;
|
|
}
|
|
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
}
|
|
|
|
return intValue;
|
|
}
|
|
|
|
LValue doubleToInt32(LValue doubleValue, double low, double high, bool isSigned = true)
|
|
{
|
|
LBasicBlock greatEnough = m_out.newBlock();
|
|
LBasicBlock withinRange = m_out.newBlock();
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
Vector<ValueFromBlock, 2> results;
|
|
|
|
m_out.branch(
|
|
m_out.doubleGreaterThanOrEqual(doubleValue, m_out.constDouble(low)),
|
|
unsure(greatEnough), unsure(slowPath));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(greatEnough, withinRange);
|
|
m_out.branch(
|
|
m_out.doubleLessThanOrEqual(doubleValue, m_out.constDouble(high)),
|
|
unsure(withinRange), unsure(slowPath));
|
|
|
|
m_out.appendTo(withinRange, slowPath);
|
|
LValue fastResult;
|
|
if (isSigned)
|
|
fastResult = m_out.doubleToInt(doubleValue);
|
|
else
|
|
fastResult = m_out.doubleToUInt(doubleValue);
|
|
results.append(m_out.anchor(fastResult));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
results.append(m_out.anchor(m_out.castToInt32(m_out.callWithoutSideEffects(Int64, operationToInt32, doubleValue))));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
return m_out.phi(Int32, results);
|
|
}
|
|
|
|
LValue doubleToInt32(LValue doubleValue)
|
|
{
|
|
#if CPU(ARM64)
|
|
if (MacroAssemblerARM64::supportsDoubleToInt32ConversionUsingJavaScriptSemantics()) {
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Int32);
|
|
patchpoint->append(ConstrainedValue(doubleValue, B3::ValueRep::SomeRegister));
|
|
patchpoint->setGenerator([=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
jit.convertDoubleToInt32UsingJavaScriptSemantics(params[1].fpr(), params[0].gpr());
|
|
});
|
|
patchpoint->effects = Effects::none();
|
|
return patchpoint;
|
|
}
|
|
#endif
|
|
|
|
if (hasSensibleDoubleToInt())
|
|
return sensibleDoubleToInt32(doubleValue);
|
|
|
|
double limit = pow(2, 31) - 1;
|
|
return doubleToInt32(doubleValue, -limit, limit);
|
|
}
|
|
|
|
LValue sensibleDoubleToInt32(LValue doubleValue)
|
|
{
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue fastResultValue = m_out.doubleToInt(doubleValue);
|
|
ValueFromBlock fastResult = m_out.anchor(fastResultValue);
|
|
m_out.branch(
|
|
m_out.equal(fastResultValue, m_out.constInt32(0x80000000)),
|
|
rarely(slowPath), usually(continuation));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(slowPath, continuation);
|
|
ValueFromBlock slowResult = m_out.anchor(m_out.castToInt32(m_out.callWithoutSideEffects(Int64, operationToInt32SensibleSlow, doubleValue)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
return m_out.phi(Int32, fastResult, slowResult);
|
|
}
|
|
|
|
// This is a mechanism for creating a code generator that fills in a gap in the code using our
|
|
// own MacroAssembler. This is useful for slow paths that involve a lot of code and we don't want
|
|
// to pay the price of B3 optimizing it. A lazy slow path will only be generated if it actually
|
|
// executes. On the other hand, a lazy slow path always incurs the cost of two additional jumps.
|
|
// Also, the lazy slow path's register allocation state is slaved to whatever B3 did, so you
|
|
// have to use a ScratchRegisterAllocator to try to use some unused registers and you may have
|
|
// to spill to top of stack if there aren't enough registers available.
|
|
//
|
|
// Lazy slow paths involve three different stages of execution. Each stage has unique
|
|
// capabilities and knowledge. The stages are:
|
|
//
|
|
// 1) DFG->B3 lowering, i.e. code that runs in this phase. Lowering is the last time you will
|
|
// have access to LValues. If there is an LValue that needs to be fed as input to a lazy slow
|
|
// path, then you must pass it as an argument here (as one of the varargs arguments after the
|
|
// functor). But, lowering doesn't know which registers will be used for those LValues. Hence
|
|
// you pass a lambda to lazySlowPath() and that lambda will run during stage (2):
|
|
//
|
|
// 2) FTLCompile.cpp's fixFunctionBasedOnStackMaps. This code is the only stage at which we know
|
|
// the mapping from arguments passed to this method in (1) and the registers that B3
|
|
// selected for those arguments. You don't actually want to generate any code here, since then
|
|
// the slow path wouldn't actually be lazily generated. Instead, you want to save the
|
|
// registers being used for the arguments and defer code generation to stage (3) by creating
|
|
// and returning a LazySlowPath::Generator:
|
|
//
|
|
// 3) LazySlowPath's generate() method. This code runs in response to the lazy slow path
|
|
// executing for the first time. It will call the generator you created in stage (2).
|
|
//
|
|
// Note that each time you invoke stage (1), stage (2) may be invoked zero, one, or many times.
|
|
// Stage (2) will usually be invoked once for stage (1). But, B3 may kill the code, in which
|
|
// case stage (2) won't run. B3 may duplicate the code (for example via tail duplication),
|
|
// leading to many calls to your stage (2) lambda. Stage (3) may be called zero or once for each
|
|
// stage (2). It will be called zero times if the slow path never runs. This is what you hope for
|
|
// whenever you use the lazySlowPath() mechanism.
|
|
//
|
|
// A typical use of lazySlowPath() will look like the example below, which just creates a slow
|
|
// path that adds some value to the input and returns it.
|
|
//
|
|
// // Stage (1) is here. This is your last chance to figure out which LValues to use as inputs.
|
|
// // Notice how we pass "input" as an argument to lazySlowPath().
|
|
// LValue input = ...;
|
|
// int addend = ...;
|
|
// LValue output = lazySlowPath(
|
|
// [=] (const Vector<Location>& locations) -> RefPtr<LazySlowPath::Generator> {
|
|
// // Stage (2) is here. This is your last chance to figure out which registers are used
|
|
// // for which values. Location zero is always the return value. You can ignore it if
|
|
// // you don't want to return anything. Location 1 is the register for the first
|
|
// // argument to the lazySlowPath(), i.e. "input". Note that the Location object could
|
|
// // also hold an FPR, if you are passing a double.
|
|
// GPRReg outputGPR = locations[0].directGPR();
|
|
// GPRReg inputGPR = locations[1].directGPR();
|
|
// return LazySlowPath::createGenerator(
|
|
// [=] (CCallHelpers& jit, LazySlowPath::GenerationParams& params) {
|
|
// // Stage (3) is here. This is when you generate code. You have access to the
|
|
// // registers you collected in stage (2) because this lambda closes over those
|
|
// // variables (outputGPR and inputGPR). You also have access to whatever extra
|
|
// // data you collected in stage (1), such as the addend in this case.
|
|
// jit.add32(TrustedImm32(addend), inputGPR, outputGPR);
|
|
// // You have to end by jumping to done. There is nothing to fall through to.
|
|
// // You can also jump to the exception handler (see LazySlowPath.h for more
|
|
// // info). Note that currently you cannot OSR exit.
|
|
// params.doneJumps.append(jit.jump());
|
|
// });
|
|
// },
|
|
// input);
|
|
//
|
|
// You can basically pass as many inputs as you like, either using this varargs form, or by
|
|
// passing a Vector of LValues.
|
|
//
|
|
// Note that if your slow path is only doing a call, you can use the createLazyCallGenerator()
|
|
// helper. For example:
|
|
//
|
|
// LValue input = ...;
|
|
// LValue output = lazySlowPath(
|
|
// [=] (const Vector<Location>& locations) -> RefPtr<LazySlowPath::Generator> {
|
|
// return createLazyCallGenerator(
|
|
// operationDoThings, locations[0].directGPR(), locations[1].directGPR());
|
|
// }, input);
|
|
//
|
|
// Finally, note that all of the lambdas - both the stage (2) lambda and the stage (3) lambda -
|
|
// run after the function that created them returns. Hence, you should not use by-reference
|
|
// capture (i.e. [&]) in any of these lambdas.
|
|
template<typename Functor, typename... ArgumentTypes>
|
|
PatchpointValue* lazySlowPath(const Functor& functor, ArgumentTypes... arguments)
|
|
{
|
|
return lazySlowPath(functor, Vector<LValue>{ arguments... });
|
|
}
|
|
|
|
template<typename Functor>
|
|
PatchpointValue* lazySlowPath(const Functor& functor, const Vector<LValue>& userArguments)
|
|
{
|
|
CodeOrigin origin = m_origin.semantic;
|
|
|
|
PatchpointValue* result = m_out.patchpoint(B3::Int64);
|
|
for (LValue arg : userArguments)
|
|
result->append(ConstrainedValue(arg, B3::ValueRep::SomeRegister));
|
|
|
|
RefPtr<PatchpointExceptionHandle> exceptionHandle =
|
|
preparePatchpointForExceptions(result);
|
|
|
|
result->clobber(RegisterSet::macroScratchRegisters());
|
|
State* state = &m_ftlState;
|
|
|
|
result->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams& params) {
|
|
Vector<Location> locations;
|
|
for (const B3::ValueRep& rep : params)
|
|
locations.append(Location::forValueRep(rep));
|
|
|
|
RefPtr<LazySlowPath::Generator> generator = functor(locations);
|
|
|
|
CCallHelpers::PatchableJump patchableJump = jit.patchableJump();
|
|
CCallHelpers::Label done = jit.label();
|
|
|
|
RegisterSet usedRegisters = params.unavailableRegisters();
|
|
|
|
RefPtr<ExceptionTarget> exceptionTarget =
|
|
exceptionHandle->scheduleExitCreation(params);
|
|
|
|
// FIXME: As part of handling exceptions, we need to create a concrete OSRExit here.
|
|
// Doing so should automagically register late paths that emit exit thunks.
|
|
|
|
params.addLatePath(
|
|
[=] (CCallHelpers& jit) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
patchableJump.m_jump.link(&jit);
|
|
unsigned index = state->jitCode->lazySlowPaths.size();
|
|
state->jitCode->lazySlowPaths.append(nullptr);
|
|
jit.pushToSaveImmediateWithoutTouchingRegisters(
|
|
CCallHelpers::TrustedImm32(index));
|
|
CCallHelpers::Jump generatorJump = jit.jump();
|
|
|
|
// Note that so long as we're here, we don't really know if our late path
|
|
// runs before or after any other late paths that we might depend on, like
|
|
// the exception thunk.
|
|
|
|
RefPtr<JITCode> jitCode = state->jitCode;
|
|
VM* vm = &state->graph.m_vm;
|
|
|
|
jit.addLinkTask(
|
|
[=] (LinkBuffer& linkBuffer) {
|
|
linkBuffer.link(generatorJump,
|
|
CodeLocationLabel<JITThunkPtrTag>(vm->getCTIStub(lazySlowPathGenerationThunkGenerator).code()));
|
|
|
|
std::unique_ptr<LazySlowPath> lazySlowPath = makeUnique<LazySlowPath>();
|
|
|
|
auto linkedPatchableJump = CodeLocationJump<JSInternalPtrTag>(linkBuffer.locationOf<JSInternalPtrTag>(patchableJump));
|
|
|
|
CodeLocationLabel<JSInternalPtrTag> linkedDone = linkBuffer.locationOf<JSInternalPtrTag>(done);
|
|
|
|
CallSiteIndex callSiteIndex =
|
|
jitCode->common.codeOrigins->addUniqueCallSiteIndex(origin);
|
|
|
|
lazySlowPath->initialize(
|
|
linkedPatchableJump, linkedDone,
|
|
exceptionTarget->label(linkBuffer), usedRegisters,
|
|
callSiteIndex, generator);
|
|
|
|
jitCode->lazySlowPaths[index] = WTFMove(lazySlowPath);
|
|
});
|
|
});
|
|
});
|
|
return result;
|
|
}
|
|
|
|
void speculate(
|
|
ExitKind kind, FormattedValue lowValue, Node* highValue, LValue failCondition)
|
|
{
|
|
appendOSRExit(kind, lowValue, highValue, failCondition, m_origin);
|
|
}
|
|
|
|
void speculate(
|
|
ExitKind kind, FormattedValue lowValue, const MethodOfGettingAValueProfile& profile, LValue failCondition)
|
|
{
|
|
appendOSRExit(kind, lowValue, profile, failCondition, m_origin);
|
|
}
|
|
|
|
template<typename... Args>
|
|
void speculateAndJump(B3::BasicBlock* target, Args... args)
|
|
{
|
|
speculate(args...);
|
|
m_out.jump(target);
|
|
}
|
|
|
|
void terminate(ExitKind kind)
|
|
{
|
|
speculate(kind, noValue(), nullptr, m_out.booleanTrue);
|
|
didAlreadyTerminate();
|
|
}
|
|
|
|
void didAlreadyTerminate()
|
|
{
|
|
m_state.setIsValid(false);
|
|
}
|
|
|
|
void simulatedTypeCheck(Edge highValue, SpeculatedType typesPassedThrough)
|
|
{
|
|
m_interpreter.filter(highValue, typesPassedThrough);
|
|
}
|
|
|
|
void typeCheck(
|
|
FormattedValue lowValue, Edge highValue, SpeculatedType typesPassedThrough,
|
|
LValue failCondition, ExitKind exitKind = BadType)
|
|
{
|
|
appendTypeCheck(lowValue, highValue, typesPassedThrough, failCondition, exitKind);
|
|
}
|
|
|
|
void appendTypeCheck(
|
|
FormattedValue lowValue, Edge highValue, SpeculatedType typesPassedThrough,
|
|
LValue failCondition, ExitKind exitKind)
|
|
{
|
|
if (!m_interpreter.needsTypeCheck(highValue, typesPassedThrough))
|
|
return;
|
|
ASSERT(mayHaveTypeCheck(highValue.useKind()));
|
|
appendOSRExit(exitKind, lowValue, highValue.node(), failCondition, m_origin);
|
|
m_interpreter.filter(highValue, typesPassedThrough);
|
|
}
|
|
|
|
LValue lowInt32(Edge edge, OperandSpeculationMode mode = AutomaticOperandSpeculation)
|
|
{
|
|
ASSERT_UNUSED(mode, mode == ManualOperandSpeculation || (edge.useKind() == Int32Use || edge.useKind() == KnownInt32Use));
|
|
|
|
if (edge->hasConstant()) {
|
|
JSValue value = edge->asJSValue();
|
|
simulatedTypeCheck(edge, SpecInt32Only);
|
|
if (!value.isInt32()) {
|
|
if (mayHaveTypeCheck(edge.useKind()))
|
|
terminate(Uncountable);
|
|
return m_out.int32Zero;
|
|
}
|
|
LValue result = m_out.constInt32(value.asInt32());
|
|
result->setOrigin(B3::Origin(edge.node()));
|
|
return result;
|
|
}
|
|
|
|
LoweredNodeValue value = m_int32Values.get(edge.node());
|
|
if (isValid(value)) {
|
|
simulatedTypeCheck(edge, SpecInt32Only);
|
|
return value.value();
|
|
}
|
|
|
|
value = m_strictInt52Values.get(edge.node());
|
|
if (isValid(value))
|
|
return strictInt52ToInt32(edge, value.value());
|
|
|
|
value = m_int52Values.get(edge.node());
|
|
if (isValid(value))
|
|
return strictInt52ToInt32(edge, int52ToStrictInt52(value.value()));
|
|
|
|
value = m_jsValueValues.get(edge.node());
|
|
if (isValid(value)) {
|
|
LValue boxedResult = value.value();
|
|
FTL_TYPE_CHECK(
|
|
jsValueValue(boxedResult), edge, SpecInt32Only, isNotInt32(boxedResult));
|
|
LValue result = unboxInt32(boxedResult);
|
|
setInt32(edge.node(), result);
|
|
return result;
|
|
}
|
|
|
|
if (mayHaveTypeCheck(edge.useKind()))
|
|
terminate(Uncountable);
|
|
return m_out.int32Zero;
|
|
}
|
|
|
|
enum Int52Kind { StrictInt52, Int52 };
|
|
LValue lowInt52(Edge edge, Int52Kind kind)
|
|
{
|
|
DFG_ASSERT(m_graph, m_node, edge.useKind() == Int52RepUse, edge.useKind());
|
|
|
|
LoweredNodeValue value;
|
|
|
|
switch (kind) {
|
|
case Int52:
|
|
value = m_int52Values.get(edge.node());
|
|
if (isValid(value))
|
|
return value.value();
|
|
|
|
value = m_strictInt52Values.get(edge.node());
|
|
if (isValid(value))
|
|
return strictInt52ToInt52(value.value());
|
|
break;
|
|
|
|
case StrictInt52:
|
|
value = m_strictInt52Values.get(edge.node());
|
|
if (isValid(value))
|
|
return value.value();
|
|
|
|
value = m_int52Values.get(edge.node());
|
|
if (isValid(value))
|
|
return int52ToStrictInt52(value.value());
|
|
break;
|
|
}
|
|
|
|
if (mayHaveTypeCheck(edge.useKind()))
|
|
terminate(Uncountable);
|
|
return m_out.int64Zero;
|
|
}
|
|
|
|
LValue lowInt52(Edge edge)
|
|
{
|
|
return lowInt52(edge, Int52);
|
|
}
|
|
|
|
LValue lowStrictInt52(Edge edge)
|
|
{
|
|
return lowInt52(edge, StrictInt52);
|
|
}
|
|
|
|
bool betterUseStrictInt52(Node* node)
|
|
{
|
|
return !isValid(m_int52Values.get(node));
|
|
}
|
|
bool betterUseStrictInt52(Edge edge)
|
|
{
|
|
return betterUseStrictInt52(edge.node());
|
|
}
|
|
template<typename T>
|
|
Int52Kind bestInt52Kind(T node)
|
|
{
|
|
return betterUseStrictInt52(node) ? StrictInt52 : Int52;
|
|
}
|
|
Int52Kind opposite(Int52Kind kind)
|
|
{
|
|
switch (kind) {
|
|
case Int52:
|
|
return StrictInt52;
|
|
case StrictInt52:
|
|
return Int52;
|
|
}
|
|
DFG_CRASH(m_graph, m_node, "Bad use kind");
|
|
return Int52;
|
|
}
|
|
|
|
LValue lowWhicheverInt52(Edge edge, Int52Kind& kind)
|
|
{
|
|
kind = bestInt52Kind(edge);
|
|
return lowInt52(edge, kind);
|
|
}
|
|
|
|
LValue lowCell(Edge edge, OperandSpeculationMode mode = AutomaticOperandSpeculation)
|
|
{
|
|
DFG_ASSERT(m_graph, m_node, mode == ManualOperandSpeculation || DFG::isCell(edge.useKind()), edge.useKind());
|
|
|
|
if (edge->op() == JSConstant) {
|
|
FrozenValue* value = edge->constant();
|
|
simulatedTypeCheck(edge, SpecCellCheck);
|
|
if (!value->value().isCell()) {
|
|
if (mayHaveTypeCheck(edge.useKind()))
|
|
terminate(Uncountable);
|
|
return m_out.intPtrZero;
|
|
}
|
|
LValue result = frozenPointer(value);
|
|
result->setOrigin(B3::Origin(edge.node()));
|
|
return result;
|
|
}
|
|
|
|
LoweredNodeValue value = m_jsValueValues.get(edge.node());
|
|
if (isValid(value)) {
|
|
LValue uncheckedValue = value.value();
|
|
FTL_TYPE_CHECK(
|
|
jsValueValue(uncheckedValue), edge, SpecCellCheck, isNotCell(uncheckedValue));
|
|
return uncheckedValue;
|
|
}
|
|
|
|
if (mayHaveTypeCheck(edge.useKind()))
|
|
terminate(Uncountable);
|
|
return m_out.intPtrZero;
|
|
}
|
|
|
|
LValue lowObject(Edge edge, OperandSpeculationMode mode = AutomaticOperandSpeculation)
|
|
{
|
|
ASSERT_UNUSED(mode, mode == ManualOperandSpeculation || edge.useKind() == ObjectUse);
|
|
|
|
LValue result = lowCell(edge, mode);
|
|
speculateObject(edge, result);
|
|
return result;
|
|
}
|
|
|
|
LValue lowRegExpObject(Edge edge)
|
|
{
|
|
LValue result = lowCell(edge);
|
|
speculateRegExpObject(edge, result);
|
|
return result;
|
|
}
|
|
|
|
LValue lowMapObject(Edge edge)
|
|
{
|
|
LValue result = lowCell(edge);
|
|
speculateMapObject(edge, result);
|
|
return result;
|
|
}
|
|
|
|
LValue lowSetObject(Edge edge)
|
|
{
|
|
LValue result = lowCell(edge);
|
|
speculateSetObject(edge, result);
|
|
return result;
|
|
}
|
|
|
|
LValue lowWeakMapObject(Edge edge)
|
|
{
|
|
LValue result = lowCell(edge);
|
|
speculateWeakMapObject(edge, result);
|
|
return result;
|
|
}
|
|
|
|
LValue lowWeakSetObject(Edge edge)
|
|
{
|
|
LValue result = lowCell(edge);
|
|
speculateWeakSetObject(edge, result);
|
|
return result;
|
|
}
|
|
|
|
LValue lowDataViewObject(Edge edge)
|
|
{
|
|
LValue result = lowCell(edge);
|
|
speculateDataViewObject(edge, result);
|
|
return result;
|
|
}
|
|
|
|
LValue lowDateObject(Edge edge)
|
|
{
|
|
LValue result = lowCell(edge);
|
|
speculateDateObject(edge, result);
|
|
return result;
|
|
}
|
|
|
|
LValue lowString(Edge edge, OperandSpeculationMode mode = AutomaticOperandSpeculation)
|
|
{
|
|
ASSERT_UNUSED(mode, mode == ManualOperandSpeculation || edge.useKind() == StringUse || edge.useKind() == KnownStringUse || edge.useKind() == StringIdentUse);
|
|
|
|
LValue result = lowCell(edge, mode);
|
|
speculateString(edge, result);
|
|
return result;
|
|
}
|
|
|
|
LValue lowStringIdent(Edge edge, OperandSpeculationMode mode = AutomaticOperandSpeculation)
|
|
{
|
|
ASSERT_UNUSED(mode, mode == ManualOperandSpeculation || edge.useKind() == StringIdentUse);
|
|
|
|
LValue string = lowString(edge, mode);
|
|
LValue stringImpl = m_out.loadPtr(string, m_heaps.JSString_value);
|
|
speculateStringIdent(edge, string, stringImpl);
|
|
return stringImpl;
|
|
}
|
|
|
|
LValue lowSymbol(Edge edge, OperandSpeculationMode mode = AutomaticOperandSpeculation)
|
|
{
|
|
ASSERT_UNUSED(mode, mode == ManualOperandSpeculation || edge.useKind() == SymbolUse);
|
|
|
|
LValue result = lowCell(edge, mode);
|
|
speculateSymbol(edge, result);
|
|
return result;
|
|
}
|
|
|
|
LValue lowHeapBigInt(Edge edge, OperandSpeculationMode mode = AutomaticOperandSpeculation)
|
|
{
|
|
ASSERT_UNUSED(mode, mode == ManualOperandSpeculation || edge.useKind() == HeapBigIntUse);
|
|
|
|
LValue result = lowCell(edge, mode);
|
|
speculateHeapBigInt(edge, result);
|
|
return result;
|
|
}
|
|
|
|
#if USE(BIGINT32)
|
|
LValue lowBigInt32(Edge edge, OperandSpeculationMode mode = AutomaticOperandSpeculation)
|
|
{
|
|
ASSERT_UNUSED(mode, mode == ManualOperandSpeculation || edge.useKind() == BigInt32Use);
|
|
|
|
LoweredNodeValue value = m_jsValueValues.get(edge.node());
|
|
if (isValid(value)) {
|
|
LValue result = value.value();
|
|
FTL_TYPE_CHECK(jsValueValue(result), edge, SpecBigInt32, isNotBigInt32(result));
|
|
return result;
|
|
}
|
|
|
|
if (mayHaveTypeCheck(edge.useKind()))
|
|
terminate(Uncountable);
|
|
return m_out.bigInt32Zero;
|
|
}
|
|
#endif
|
|
|
|
LValue lowNonNullObject(Edge edge, OperandSpeculationMode mode = AutomaticOperandSpeculation)
|
|
{
|
|
ASSERT_UNUSED(mode, mode == ManualOperandSpeculation || edge.useKind() == ObjectUse);
|
|
|
|
LValue result = lowCell(edge, mode);
|
|
speculateNonNullObject(edge, result);
|
|
return result;
|
|
}
|
|
|
|
LValue lowBoolean(Edge edge, OperandSpeculationMode mode = AutomaticOperandSpeculation)
|
|
{
|
|
ASSERT_UNUSED(mode, mode == ManualOperandSpeculation || edge.useKind() == BooleanUse || edge.useKind() == KnownBooleanUse);
|
|
|
|
if (edge->hasConstant()) {
|
|
JSValue value = edge->asJSValue();
|
|
simulatedTypeCheck(edge, SpecBoolean);
|
|
if (!value.isBoolean()) {
|
|
if (mayHaveTypeCheck(edge.useKind()))
|
|
terminate(Uncountable);
|
|
return m_out.booleanFalse;
|
|
}
|
|
LValue result = m_out.constBool(value.asBoolean());
|
|
result->setOrigin(B3::Origin(edge.node()));
|
|
return result;
|
|
}
|
|
|
|
LoweredNodeValue value = m_booleanValues.get(edge.node());
|
|
if (isValid(value)) {
|
|
simulatedTypeCheck(edge, SpecBoolean);
|
|
return value.value();
|
|
}
|
|
|
|
value = m_jsValueValues.get(edge.node());
|
|
if (isValid(value)) {
|
|
LValue unboxedResult = value.value();
|
|
FTL_TYPE_CHECK(
|
|
jsValueValue(unboxedResult), edge, SpecBoolean, isNotBoolean(unboxedResult));
|
|
LValue result = unboxBoolean(unboxedResult);
|
|
setBoolean(edge.node(), result);
|
|
return result;
|
|
}
|
|
|
|
if (mayHaveTypeCheck(edge.useKind()))
|
|
terminate(Uncountable);
|
|
return m_out.booleanFalse;
|
|
}
|
|
|
|
LValue lowDouble(Edge edge)
|
|
{
|
|
DFG_ASSERT(m_graph, m_node, isDouble(edge.useKind()), edge.useKind());
|
|
|
|
LoweredNodeValue value = m_doubleValues.get(edge.node());
|
|
if (isValid(value))
|
|
return value.value();
|
|
if (mayHaveTypeCheck(edge.useKind()))
|
|
terminate(Uncountable);
|
|
return m_out.doubleZero;
|
|
}
|
|
|
|
LValue lowJSValue(Edge edge, OperandSpeculationMode mode = AutomaticOperandSpeculation)
|
|
{
|
|
DFG_ASSERT(m_graph, m_node, mode == ManualOperandSpeculation || edge.useKind() == UntypedUse, m_node->op(), edge.useKind());
|
|
DFG_ASSERT(m_graph, m_node, !isDouble(edge.useKind()), m_node->op(), edge.useKind());
|
|
DFG_ASSERT(m_graph, m_node, edge.useKind() != Int52RepUse, m_node->op(), edge.useKind());
|
|
|
|
if (edge->hasConstant()) {
|
|
LValue result = m_out.constInt64(JSValue::encode(edge->asJSValue()));
|
|
result->setOrigin(B3::Origin(edge.node()));
|
|
return result;
|
|
}
|
|
|
|
LoweredNodeValue value = m_jsValueValues.get(edge.node());
|
|
if (isValid(value))
|
|
return value.value();
|
|
|
|
value = m_int32Values.get(edge.node());
|
|
if (isValid(value)) {
|
|
LValue result = boxInt32(value.value());
|
|
setJSValue(edge.node(), result);
|
|
return result;
|
|
}
|
|
|
|
value = m_booleanValues.get(edge.node());
|
|
if (isValid(value)) {
|
|
LValue result = boxBoolean(value.value());
|
|
setJSValue(edge.node(), result);
|
|
return result;
|
|
}
|
|
|
|
DFG_CRASH(m_graph, m_node, makeString("Value not defined: ", String::number(edge.node()->index())).ascii().data());
|
|
return nullptr;
|
|
}
|
|
|
|
LValue lowNotCell(Edge edge)
|
|
{
|
|
LValue result = lowJSValue(edge, ManualOperandSpeculation);
|
|
FTL_TYPE_CHECK(jsValueValue(result), edge, ~SpecCellCheck, isCell(result));
|
|
return result;
|
|
}
|
|
|
|
LValue lowStorage(Edge edge)
|
|
{
|
|
LoweredNodeValue value = m_storageValues.get(edge.node());
|
|
if (isValid(value))
|
|
return value.value();
|
|
|
|
LValue result = lowCell(edge);
|
|
setStorage(edge.node(), result);
|
|
return result;
|
|
}
|
|
|
|
LValue strictInt52ToInt32(Edge edge, LValue value)
|
|
{
|
|
LValue result = m_out.castToInt32(value);
|
|
FTL_TYPE_CHECK(
|
|
noValue(), edge, SpecInt32Only,
|
|
m_out.notEqual(m_out.signExt32To64(result), value));
|
|
setInt32(edge.node(), result);
|
|
return result;
|
|
}
|
|
|
|
LValue strictInt52ToDouble(LValue value)
|
|
{
|
|
return m_out.intToDouble(value);
|
|
}
|
|
|
|
LValue strictInt52ToJSValue(LValue value)
|
|
{
|
|
LBasicBlock isInt32 = m_out.newBlock();
|
|
LBasicBlock isDouble = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
Vector<ValueFromBlock, 2> results;
|
|
|
|
LValue int32Value = m_out.castToInt32(value);
|
|
m_out.branch(
|
|
m_out.equal(m_out.signExt32To64(int32Value), value),
|
|
unsure(isInt32), unsure(isDouble));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isInt32, isDouble);
|
|
|
|
results.append(m_out.anchor(boxInt32(int32Value)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(isDouble, continuation);
|
|
|
|
results.append(m_out.anchor(boxDouble(m_out.intToDouble(value))));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
return m_out.phi(Int64, results);
|
|
}
|
|
|
|
LValue strictInt52ToInt52(LValue value)
|
|
{
|
|
return m_out.shl(value, m_out.constInt64(JSValue::int52ShiftAmount));
|
|
}
|
|
|
|
LValue int52ToStrictInt52(LValue value)
|
|
{
|
|
return m_out.aShr(value, m_out.constInt64(JSValue::int52ShiftAmount));
|
|
}
|
|
|
|
LValue isInt32(LValue jsValue, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type, SpecInt32Only))
|
|
return proven;
|
|
return m_out.aboveOrEqual(jsValue, m_numberTag);
|
|
}
|
|
LValue isNotInt32(LValue jsValue, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type, ~SpecInt32Only))
|
|
return proven;
|
|
return m_out.below(jsValue, m_numberTag);
|
|
}
|
|
LValue unboxInt32(LValue jsValue)
|
|
{
|
|
return m_out.castToInt32(jsValue);
|
|
}
|
|
LValue boxInt32(LValue value)
|
|
{
|
|
return m_out.add(m_out.zeroExt(value, Int64), m_numberTag);
|
|
}
|
|
|
|
#if USE(BIGINT32)
|
|
LValue isBigInt32(LValue jsValue, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type, SpecBigInt32))
|
|
return proven;
|
|
return m_out.equal(m_out.bitAnd(jsValue, m_out.constInt64(JSValue::BigInt32Mask)), m_out.constInt64(JSValue::BigInt32Tag));
|
|
}
|
|
LValue isNotBigInt32(LValue jsValue, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type, ~SpecBigInt32))
|
|
return proven;
|
|
return m_out.notEqual(m_out.bitAnd(jsValue, m_out.constInt64(JSValue::BigInt32Mask)), m_out.constInt64(JSValue::BigInt32Tag));
|
|
}
|
|
LValue unboxBigInt32(LValue jsValue)
|
|
{
|
|
return m_out.castToInt32(m_out.lShr(jsValue, m_out.constInt64(16)));
|
|
}
|
|
LValue boxBigInt32(LValue int32Value)
|
|
{
|
|
return m_out.bitOr(
|
|
m_out.shl(m_out.zeroExt(int32Value, B3::Int64), m_out.constInt64(16)),
|
|
m_out.constInt64(JSValue::BigInt32Tag));
|
|
}
|
|
LValue isNotAnyBigInt(LValue jsValue, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type, ~SpecBigInt))
|
|
return proven;
|
|
|
|
// if (isBigInt32)
|
|
// return false
|
|
// if (!isCell)
|
|
// return true;
|
|
// return !isHeapBigInt
|
|
LBasicBlock isBigInt32Case = m_out.newBlock();
|
|
LBasicBlock isNotBigInt32Case = m_out.newBlock();
|
|
LBasicBlock isNotCellCase = m_out.newBlock();
|
|
LBasicBlock isCellCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(isBigInt32(jsValue, type), unsure(isBigInt32Case), unsure(isNotBigInt32Case));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isBigInt32Case, isNotBigInt32Case);
|
|
ValueFromBlock returnFalse = m_out.anchor(m_out.booleanFalse);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(isNotBigInt32Case, isNotCellCase);
|
|
// FIXME: we should filter the type passed to isCell to account for the previous test that told us we are definitely not a BigInt32.
|
|
m_out.branch(isCell(jsValue, type), unsure(isCellCase), unsure(isNotCellCase));
|
|
|
|
m_out.appendTo(isNotCellCase, isCellCase);
|
|
ValueFromBlock returnTrue = m_out.anchor(m_out.booleanTrue);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(isCellCase, continuation);
|
|
ValueFromBlock returnIsNotHeapBigInt = m_out.anchor(isNotHeapBigInt(jsValue));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
return m_out.phi(Int32, returnFalse, returnTrue, returnIsNotHeapBigInt);
|
|
}
|
|
#endif // USE(BIGINT32)
|
|
|
|
LValue isCellOrMiscOrBigInt32(LValue jsValue, SpeculatedType type = SpecFullTop)
|
|
{
|
|
SpeculatedType filter = SpecCellCheck | SpecMisc;
|
|
#if USE(BIGINT32)
|
|
filter |= SpecBigInt32;
|
|
#endif // USE(BIGINT32)
|
|
if (LValue proven = isProvenValue(type, filter))
|
|
return proven;
|
|
return m_out.testIsZero64(jsValue, m_numberTag);
|
|
}
|
|
LValue isNotCellOrMiscOrBigInt32(LValue jsValue, SpeculatedType type = SpecFullTop)
|
|
{
|
|
SpeculatedType filter = SpecCellCheck | SpecMisc;
|
|
#if USE(BIGINT32)
|
|
filter |= SpecBigInt32;
|
|
#endif // USE(BIGINT32)
|
|
if (LValue proven = isProvenValue(type, ~filter))
|
|
return proven;
|
|
return m_out.testNonZero64(jsValue, m_numberTag);
|
|
}
|
|
|
|
LValue unboxDouble(LValue jsValue, LValue* unboxedAsInt = nullptr)
|
|
{
|
|
LValue asInt = m_out.add(jsValue, m_numberTag);
|
|
if (unboxedAsInt)
|
|
*unboxedAsInt = asInt;
|
|
return m_out.bitCast(asInt, Double);
|
|
}
|
|
LValue boxDouble(LValue doubleValue)
|
|
{
|
|
return m_out.sub(m_out.bitCast(doubleValue, Int64), m_numberTag);
|
|
}
|
|
|
|
LValue jsValueToStrictInt52(Edge edge, LValue boxedValue)
|
|
{
|
|
LBasicBlock intCase = m_out.newBlock();
|
|
LBasicBlock doubleCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue isNotInt32;
|
|
if (!m_interpreter.needsTypeCheck(edge, SpecInt32Only))
|
|
isNotInt32 = m_out.booleanFalse;
|
|
else if (!m_interpreter.needsTypeCheck(edge, ~SpecInt32Only))
|
|
isNotInt32 = m_out.booleanTrue;
|
|
else
|
|
isNotInt32 = this->isNotInt32(boxedValue);
|
|
m_out.branch(isNotInt32, unsure(doubleCase), unsure(intCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(intCase, doubleCase);
|
|
|
|
ValueFromBlock intToInt52 = m_out.anchor(
|
|
m_out.signExt32To64(unboxInt32(boxedValue)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(doubleCase, continuation);
|
|
|
|
LValue possibleResult = m_out.callWithoutSideEffects(Int64, operationConvertBoxedDoubleToInt52, boxedValue);
|
|
FTL_TYPE_CHECK(
|
|
jsValueValue(boxedValue), edge, SpecInt32Only | SpecAnyIntAsDouble,
|
|
m_out.equal(possibleResult, m_out.constInt64(JSValue::notInt52)));
|
|
|
|
ValueFromBlock doubleToInt52 = m_out.anchor(possibleResult);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
|
|
return m_out.phi(Int64, intToInt52, doubleToInt52);
|
|
}
|
|
|
|
LValue doubleToStrictInt52(Edge edge, LValue value)
|
|
{
|
|
LValue integerValue = m_out.doubleToInt64(value);
|
|
LValue integerValueConvertedToDouble = m_out.intToDouble(integerValue);
|
|
LValue valueNotConvertibleToInteger = m_out.doubleNotEqualOrUnordered(value, integerValueConvertedToDouble);
|
|
speculate(Int52Overflow, doubleValue(value), edge.node(), valueNotConvertibleToInteger);
|
|
|
|
LBasicBlock valueIsZero = m_out.newBlock();
|
|
LBasicBlock valueIsNotZero = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
m_out.branch(m_out.isZero64(integerValue), unsure(valueIsZero), unsure(valueIsNotZero));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(valueIsZero, valueIsNotZero);
|
|
LValue doubleBitcastToInt64 = m_out.bitCast(value, Int64);
|
|
LValue signBitSet = m_out.lessThan(doubleBitcastToInt64, m_out.constInt64(0));
|
|
speculate(Int52Overflow, doubleValue(value), edge.node(), signBitSet);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(valueIsNotZero, continuation);
|
|
speculate(Int52Overflow, doubleValue(value), edge.node(), m_out.greaterThanOrEqual(integerValue, m_out.constInt64(static_cast<int64_t>(1) << (JSValue::numberOfInt52Bits - 1))));
|
|
speculate(Int52Overflow, doubleValue(value), edge.node(), m_out.lessThan(integerValue, m_out.constInt64(-(static_cast<int64_t>(1) << (JSValue::numberOfInt52Bits - 1)))));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
m_interpreter.filter(edge, SpecAnyIntAsDouble);
|
|
return integerValue;
|
|
}
|
|
|
|
LValue convertDoubleToInt32(LValue value, bool shouldCheckNegativeZero)
|
|
{
|
|
LValue integerValue = m_out.doubleToInt(value);
|
|
LValue integerValueConvertedToDouble = m_out.intToDouble(integerValue);
|
|
LValue valueNotConvertibleToInteger = m_out.doubleNotEqualOrUnordered(value, integerValueConvertedToDouble);
|
|
speculate(Overflow, FormattedValue(DataFormatDouble, value), m_node, valueNotConvertibleToInteger);
|
|
|
|
if (shouldCheckNegativeZero) {
|
|
LBasicBlock valueIsZero = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
m_out.branch(m_out.isZero32(integerValue), unsure(valueIsZero), unsure(continuation));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(valueIsZero, continuation);
|
|
|
|
LValue doubleBitcastToInt64 = m_out.bitCast(value, Int64);
|
|
LValue signBitSet = m_out.lessThan(doubleBitcastToInt64, m_out.constInt64(0));
|
|
|
|
speculate(NegativeZero, FormattedValue(DataFormatDouble, value), m_node, signBitSet);
|
|
m_out.jump(continuation);
|
|
m_out.appendTo(continuation, lastNext);
|
|
}
|
|
return integerValue;
|
|
}
|
|
|
|
LValue isNumber(LValue jsValue, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type, SpecFullNumber))
|
|
return proven;
|
|
return isNotCellOrMiscOrBigInt32(jsValue);
|
|
}
|
|
LValue isNotNumber(LValue jsValue, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type, ~SpecFullNumber))
|
|
return proven;
|
|
return isCellOrMiscOrBigInt32(jsValue);
|
|
}
|
|
|
|
LValue isNotCell(LValue jsValue, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type, ~SpecCellCheck))
|
|
return proven;
|
|
return m_out.testNonZero64(jsValue, m_notCellMask);
|
|
}
|
|
|
|
LValue isCell(LValue jsValue, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type, SpecCellCheck))
|
|
return proven;
|
|
return m_out.testIsZero64(jsValue, m_notCellMask);
|
|
}
|
|
|
|
LValue isNotMisc(LValue value, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type, ~SpecMisc))
|
|
return proven;
|
|
return m_out.above(value, m_out.constInt64(JSValue::MiscTag));
|
|
}
|
|
|
|
LValue isMisc(LValue value, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type, SpecMisc))
|
|
return proven;
|
|
return m_out.logicalNot(isNotMisc(value));
|
|
}
|
|
|
|
LValue isNotBoolean(LValue jsValue, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type, ~SpecBoolean))
|
|
return proven;
|
|
return m_out.testNonZero64(
|
|
m_out.bitXor(jsValue, m_out.constInt64(JSValue::ValueFalse)),
|
|
m_out.constInt64(~1));
|
|
}
|
|
LValue isBoolean(LValue jsValue, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type, SpecBoolean))
|
|
return proven;
|
|
return m_out.logicalNot(isNotBoolean(jsValue));
|
|
}
|
|
LValue unboxBoolean(LValue jsValue)
|
|
{
|
|
// We want to use a cast that guarantees that B3 knows that even the integer
|
|
// value is just 0 or 1. But for now we do it the direct way.
|
|
return m_out.notZero64(m_out.bitAnd(jsValue, m_out.constInt64(1)));
|
|
}
|
|
LValue boxBoolean(LValue value)
|
|
{
|
|
return m_out.select(
|
|
value, m_out.constInt64(JSValue::ValueTrue), m_out.constInt64(JSValue::ValueFalse));
|
|
}
|
|
|
|
LValue isNotOther(LValue value, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type, ~SpecOther))
|
|
return proven;
|
|
return m_out.notEqual(
|
|
m_out.bitAnd(value, m_out.constInt64(~JSValue::UndefinedTag)),
|
|
m_out.constInt64(JSValue::ValueNull));
|
|
}
|
|
LValue isOther(LValue value, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type, SpecOther))
|
|
return proven;
|
|
return m_out.equal(
|
|
m_out.bitAnd(value, m_out.constInt64(~JSValue::UndefinedTag)),
|
|
m_out.constInt64(JSValue::ValueNull));
|
|
}
|
|
|
|
LValue isProvenValue(SpeculatedType provenType, SpeculatedType wantedType)
|
|
{
|
|
if (!(provenType & ~wantedType))
|
|
return m_out.booleanTrue;
|
|
if (!(provenType & wantedType))
|
|
return m_out.booleanFalse;
|
|
return nullptr;
|
|
}
|
|
|
|
void speculate(Edge edge)
|
|
{
|
|
switch (edge.useKind()) {
|
|
case UntypedUse:
|
|
break;
|
|
case KnownInt32Use:
|
|
case KnownStringUse:
|
|
case KnownPrimitiveUse:
|
|
case KnownOtherUse:
|
|
case DoubleRepUse:
|
|
case Int52RepUse:
|
|
case KnownCellUse:
|
|
case KnownBooleanUse:
|
|
ASSERT(!m_interpreter.needsTypeCheck(edge));
|
|
break;
|
|
case Int32Use:
|
|
speculateInt32(edge);
|
|
break;
|
|
case CellUse:
|
|
speculateCell(edge);
|
|
break;
|
|
case CellOrOtherUse:
|
|
speculateCellOrOther(edge);
|
|
break;
|
|
case AnyIntUse:
|
|
speculateAnyInt(edge);
|
|
break;
|
|
case ObjectUse:
|
|
speculateObject(edge);
|
|
break;
|
|
case ArrayUse:
|
|
speculateArray(edge);
|
|
break;
|
|
case FunctionUse:
|
|
speculateFunction(edge);
|
|
break;
|
|
case ObjectOrOtherUse:
|
|
speculateObjectOrOther(edge);
|
|
break;
|
|
case FinalObjectUse:
|
|
speculateFinalObject(edge);
|
|
break;
|
|
case RegExpObjectUse:
|
|
speculateRegExpObject(edge);
|
|
break;
|
|
case PromiseObjectUse:
|
|
speculatePromiseObject(edge);
|
|
break;
|
|
case ProxyObjectUse:
|
|
speculateProxyObject(edge);
|
|
break;
|
|
case DerivedArrayUse:
|
|
speculateDerivedArray(edge);
|
|
break;
|
|
case DateObjectUse:
|
|
speculateDateObject(edge);
|
|
break;
|
|
case MapObjectUse:
|
|
speculateMapObject(edge);
|
|
break;
|
|
case SetObjectUse:
|
|
speculateSetObject(edge);
|
|
break;
|
|
case WeakMapObjectUse:
|
|
speculateWeakMapObject(edge);
|
|
break;
|
|
case WeakSetObjectUse:
|
|
speculateWeakSetObject(edge);
|
|
break;
|
|
case DataViewObjectUse:
|
|
speculateDataViewObject(edge);
|
|
break;
|
|
case StringUse:
|
|
speculateString(edge);
|
|
break;
|
|
case StringOrOtherUse:
|
|
speculateStringOrOther(edge);
|
|
break;
|
|
case StringIdentUse:
|
|
speculateStringIdent(edge);
|
|
break;
|
|
case SymbolUse:
|
|
speculateSymbol(edge);
|
|
break;
|
|
case StringObjectUse:
|
|
speculateStringObject(edge);
|
|
break;
|
|
case StringOrStringObjectUse:
|
|
speculateStringOrStringObject(edge);
|
|
break;
|
|
case NumberUse:
|
|
speculateNumber(edge);
|
|
break;
|
|
case RealNumberUse:
|
|
speculateRealNumber(edge);
|
|
break;
|
|
case DoubleRepRealUse:
|
|
speculateDoubleRepReal(edge);
|
|
break;
|
|
case DoubleRepAnyIntUse:
|
|
speculateDoubleRepAnyInt(edge);
|
|
break;
|
|
case BooleanUse:
|
|
speculateBoolean(edge);
|
|
break;
|
|
#if USE(BIGINT32)
|
|
case BigInt32Use:
|
|
speculateBigInt32(edge);
|
|
break;
|
|
case AnyBigIntUse:
|
|
speculateAnyBigInt(edge);
|
|
break;
|
|
#endif // USE(BIGINT32)
|
|
case HeapBigIntUse:
|
|
speculateHeapBigInt(edge);
|
|
break;
|
|
case NotStringVarUse:
|
|
speculateNotStringVar(edge);
|
|
break;
|
|
case NotSymbolUse:
|
|
speculateNotSymbol(edge);
|
|
break;
|
|
case NotCellUse:
|
|
speculateNotCell(edge);
|
|
break;
|
|
case NotCellNorBigIntUse:
|
|
speculateNotCellNorBigInt(edge);
|
|
break;
|
|
case OtherUse:
|
|
speculateOther(edge);
|
|
break;
|
|
case MiscUse:
|
|
speculateMisc(edge);
|
|
break;
|
|
default:
|
|
DFG_CRASH(m_graph, m_node, "Unsupported speculation use kind");
|
|
}
|
|
}
|
|
|
|
void speculate(Node*, Edge edge)
|
|
{
|
|
speculate(edge);
|
|
}
|
|
|
|
void speculateInt32(Edge edge)
|
|
{
|
|
lowInt32(edge);
|
|
}
|
|
|
|
void speculateCell(Edge edge)
|
|
{
|
|
lowCell(edge);
|
|
}
|
|
|
|
void speculateNotCell(Edge edge)
|
|
{
|
|
if (!m_interpreter.needsTypeCheck(edge))
|
|
return;
|
|
lowNotCell(edge);
|
|
}
|
|
|
|
void speculateNotCellNorBigInt(Edge edge)
|
|
{
|
|
#if USE(BIGINT32)
|
|
if (!m_interpreter.needsTypeCheck(edge))
|
|
return;
|
|
LValue nonCell = lowNotCell(edge);
|
|
FTL_TYPE_CHECK(jsValueValue(nonCell), edge, ~SpecCellCheck & ~SpecBigInt, isBigInt32(nonCell));
|
|
#else
|
|
speculateNotCell(edge);
|
|
#endif
|
|
}
|
|
|
|
void speculateCellOrOther(Edge edge)
|
|
{
|
|
if (shouldNotHaveTypeCheck(edge.useKind()))
|
|
return;
|
|
|
|
LValue value = lowJSValue(edge, ManualOperandSpeculation);
|
|
|
|
LBasicBlock isNotCell = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(isCell(value, provenType(edge)), unsure(continuation), unsure(isNotCell));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isNotCell, continuation);
|
|
FTL_TYPE_CHECK(jsValueValue(value), edge, SpecCellCheck | SpecOther, isNotOther(value));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
}
|
|
|
|
void speculateAnyInt(Edge edge)
|
|
{
|
|
if (!m_interpreter.needsTypeCheck(edge))
|
|
return;
|
|
|
|
jsValueToStrictInt52(edge, lowJSValue(edge, ManualOperandSpeculation));
|
|
}
|
|
|
|
LValue isCellWithType(LValue cell, JSTypeRange queriedTypeRange, Optional<SpeculatedType> speculatedTypeForQuery, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (speculatedTypeForQuery) {
|
|
if (LValue proven = isProvenValue(type & SpecCell, speculatedTypeForQuery.value()))
|
|
return proven;
|
|
}
|
|
if (queriedTypeRange.first == queriedTypeRange.last) {
|
|
return m_out.equal(
|
|
m_out.load8ZeroExt32(cell, m_heaps.JSCell_typeInfoType),
|
|
m_out.constInt32(queriedTypeRange.first));
|
|
}
|
|
|
|
ASSERT(queriedTypeRange.last > queriedTypeRange.first);
|
|
LValue first = m_out.sub(
|
|
m_out.load8ZeroExt32(cell, m_heaps.JSCell_typeInfoType),
|
|
m_out.constInt32(queriedTypeRange.first));
|
|
return m_out.belowOrEqual(first, m_out.constInt32(queriedTypeRange.last - queriedTypeRange.first));
|
|
}
|
|
|
|
LValue isCellWithType(LValue cell, JSType queriedType, Optional<SpeculatedType> speculatedTypeForQuery, SpeculatedType type = SpecFullTop)
|
|
{
|
|
return isCellWithType(cell, JSTypeRange { queriedType, queriedType }, speculatedTypeForQuery, type);
|
|
}
|
|
|
|
LValue isTypedArrayView(LValue cell, SpeculatedType type = SpecFullTop)
|
|
{
|
|
return isCellWithType(cell, JSTypeRange { static_cast<JSType>(FirstTypedArrayType), static_cast<JSType>(LastTypedArrayTypeExcludingDataView) }, SpecTypedArrayView, type);
|
|
}
|
|
|
|
LValue isObject(LValue cell, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type & SpecCell, SpecObject))
|
|
return proven;
|
|
return m_out.aboveOrEqual(
|
|
m_out.load8ZeroExt32(cell, m_heaps.JSCell_typeInfoType),
|
|
m_out.constInt32(ObjectType));
|
|
}
|
|
|
|
LValue isNotObject(LValue cell, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type & SpecCell, ~SpecObject))
|
|
return proven;
|
|
return m_out.below(
|
|
m_out.load8ZeroExt32(cell, m_heaps.JSCell_typeInfoType),
|
|
m_out.constInt32(ObjectType));
|
|
}
|
|
|
|
LValue isNotString(LValue cell, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type & SpecCell, ~SpecString))
|
|
return proven;
|
|
return m_out.notEqual(
|
|
m_out.load32(cell, m_heaps.JSCell_structureID),
|
|
m_out.constInt32(vm().stringStructure->id()));
|
|
}
|
|
|
|
LValue isString(LValue cell, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type & SpecCell, SpecString))
|
|
return proven;
|
|
return m_out.equal(
|
|
m_out.load32(cell, m_heaps.JSCell_structureID),
|
|
m_out.constInt32(vm().stringStructure->id()));
|
|
}
|
|
|
|
LValue isRopeString(LValue string, Edge edge = Edge())
|
|
{
|
|
if (edge) {
|
|
if (!((provenType(edge) & SpecString) & ~SpecStringIdent))
|
|
return m_out.booleanFalse;
|
|
if (JSValue value = provenValue(edge)) {
|
|
if (value.isCell() && value.asCell()->type() == StringType && !asString(value)->isRope())
|
|
return m_out.booleanFalse;
|
|
}
|
|
String value = edge->tryGetString(m_graph);
|
|
if (!value.isNull()) {
|
|
// If this value is LazyValue, it will be converted to JSString, and the result must be non-rope string.
|
|
return m_out.booleanFalse;
|
|
}
|
|
}
|
|
|
|
return m_out.testNonZeroPtr(m_out.loadPtr(string, m_heaps.JSString_value), m_out.constIntPtr(JSString::isRopeInPointer));
|
|
}
|
|
|
|
LValue isNotRopeString(LValue string, Edge edge = Edge())
|
|
{
|
|
if (edge) {
|
|
if (!((provenType(edge) & SpecString) & ~SpecStringIdent))
|
|
return m_out.booleanTrue;
|
|
if (JSValue value = provenValue(edge)) {
|
|
if (value.isCell() && value.asCell()->type() == StringType && !asString(value)->isRope())
|
|
return m_out.booleanTrue;
|
|
}
|
|
String value = edge->tryGetString(m_graph);
|
|
if (!value.isNull()) {
|
|
// If this value is LazyValue, it will be converted to JSString, and the result must be non-rope string.
|
|
return m_out.booleanTrue;
|
|
}
|
|
}
|
|
|
|
return m_out.testIsZeroPtr(m_out.loadPtr(string, m_heaps.JSString_value), m_out.constIntPtr(JSString::isRopeInPointer));
|
|
}
|
|
|
|
LValue isNotSymbol(LValue cell, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type & SpecCell, ~SpecSymbol))
|
|
return proven;
|
|
return m_out.notEqual(
|
|
m_out.load32(cell, m_heaps.JSCell_structureID),
|
|
m_out.constInt32(vm().symbolStructure->id()));
|
|
}
|
|
|
|
LValue isSymbol(LValue cell, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type & SpecCell, SpecSymbol))
|
|
return proven;
|
|
return m_out.equal(
|
|
m_out.load32(cell, m_heaps.JSCell_structureID),
|
|
m_out.constInt32(vm().symbolStructure->id()));
|
|
}
|
|
|
|
LValue isNotHeapBigIntUnknownWhetherCell(LValue value, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type, ~SpecHeapBigInt))
|
|
return proven;
|
|
|
|
LBasicBlock isCellCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
ValueFromBlock defaultToTrue = m_out.anchor(m_out.booleanTrue);
|
|
m_out.branch(isCell(value, type), unsure(isCellCase), unsure(continuation));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isCellCase, continuation);
|
|
ValueFromBlock returnForCell = m_out.anchor(isNotHeapBigInt(value, type));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
LValue result = m_out.phi(Int32, defaultToTrue, returnForCell);
|
|
return result;
|
|
}
|
|
|
|
LValue isNotHeapBigInt(LValue cell, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type & SpecCell, ~SpecHeapBigInt))
|
|
return proven;
|
|
return m_out.notEqual(
|
|
m_out.load32(cell, m_heaps.JSCell_structureID),
|
|
m_out.constInt32(vm().bigIntStructure->id()));
|
|
}
|
|
|
|
LValue isHeapBigInt(LValue cell, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type & SpecCell, SpecHeapBigInt))
|
|
return proven;
|
|
return m_out.equal(
|
|
m_out.load32(cell, m_heaps.JSCell_structureID),
|
|
m_out.constInt32(vm().bigIntStructure->id()));
|
|
}
|
|
|
|
LValue isArrayTypeForArrayify(LValue cell, ArrayMode arrayMode)
|
|
{
|
|
switch (arrayMode.type()) {
|
|
case Array::Int32:
|
|
case Array::Double:
|
|
case Array::Contiguous:
|
|
case Array::Undecided:
|
|
case Array::ArrayStorage: {
|
|
IndexingType indexingModeMask = IsArray | IndexingShapeMask;
|
|
if (arrayMode.action() == Array::Write)
|
|
indexingModeMask |= CopyOnWrite;
|
|
|
|
IndexingType shape = arrayMode.shapeMask();
|
|
LValue indexingType = m_out.load8ZeroExt32(cell, m_heaps.JSCell_indexingTypeAndMisc);
|
|
|
|
switch (arrayMode.arrayClass()) {
|
|
case Array::OriginalArray:
|
|
case Array::OriginalCopyOnWriteArray:
|
|
DFG_CRASH(m_graph, m_node, "Unexpected original array");
|
|
return nullptr;
|
|
|
|
case Array::Array:
|
|
return m_out.equal(
|
|
m_out.bitAnd(indexingType, m_out.constInt32(indexingModeMask)),
|
|
m_out.constInt32(IsArray | shape));
|
|
|
|
case Array::NonArray:
|
|
case Array::OriginalNonArray:
|
|
return m_out.equal(
|
|
m_out.bitAnd(indexingType, m_out.constInt32(indexingModeMask)),
|
|
m_out.constInt32(shape));
|
|
|
|
case Array::PossiblyArray:
|
|
return m_out.equal(
|
|
m_out.bitAnd(indexingType, m_out.constInt32(indexingModeMask & ~IsArray)),
|
|
m_out.constInt32(shape));
|
|
}
|
|
break;
|
|
}
|
|
|
|
case Array::SlowPutArrayStorage: {
|
|
ASSERT(!arrayMode.isJSArrayWithOriginalStructure());
|
|
LValue indexingType = m_out.load8ZeroExt32(cell, m_heaps.JSCell_indexingTypeAndMisc);
|
|
|
|
LBasicBlock trueCase = m_out.newBlock();
|
|
LBasicBlock checkCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
ValueFromBlock falseValue = m_out.anchor(m_out.booleanFalse);
|
|
LValue isAnArrayStorageShape = m_out.belowOrEqual(
|
|
m_out.sub(
|
|
m_out.bitAnd(indexingType, m_out.constInt32(IndexingShapeMask)),
|
|
m_out.constInt32(ArrayStorageShape)),
|
|
m_out.constInt32(SlowPutArrayStorageShape - ArrayStorageShape));
|
|
m_out.branch(isAnArrayStorageShape, unsure(checkCase), unsure(continuation));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(checkCase, trueCase);
|
|
switch (arrayMode.arrayClass()) {
|
|
case Array::OriginalArray:
|
|
case Array::OriginalCopyOnWriteArray:
|
|
DFG_CRASH(m_graph, m_node, "Unexpected original array");
|
|
return nullptr;
|
|
|
|
case Array::Array:
|
|
m_out.branch(
|
|
m_out.testNonZero32(indexingType, m_out.constInt32(IsArray)),
|
|
unsure(trueCase), unsure(continuation));
|
|
break;
|
|
|
|
case Array::NonArray:
|
|
case Array::OriginalNonArray:
|
|
m_out.branch(
|
|
m_out.testIsZero32(indexingType, m_out.constInt32(IsArray)),
|
|
unsure(trueCase), unsure(continuation));
|
|
break;
|
|
|
|
case Array::PossiblyArray:
|
|
m_out.jump(trueCase);
|
|
break;
|
|
}
|
|
|
|
m_out.appendTo(trueCase, continuation);
|
|
ValueFromBlock trueValue = m_out.anchor(m_out.booleanTrue);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
return m_out.phi(Int32, falseValue, trueValue);
|
|
}
|
|
|
|
default:
|
|
break;
|
|
}
|
|
DFG_CRASH(m_graph, m_node, "Corrupt array class");
|
|
}
|
|
|
|
LValue isArrayTypeForCheckArray(LValue cell, ArrayMode arrayMode)
|
|
{
|
|
switch (arrayMode.type()) {
|
|
case Array::Int32:
|
|
case Array::Double:
|
|
case Array::Contiguous:
|
|
case Array::Undecided:
|
|
case Array::ArrayStorage:
|
|
case Array::SlowPutArrayStorage:
|
|
return isArrayTypeForArrayify(cell, arrayMode);
|
|
|
|
case Array::DirectArguments:
|
|
return m_out.equal(
|
|
m_out.load8ZeroExt32(cell, m_heaps.JSCell_typeInfoType),
|
|
m_out.constInt32(DirectArgumentsType));
|
|
|
|
case Array::ScopedArguments:
|
|
return m_out.equal(
|
|
m_out.load8ZeroExt32(cell, m_heaps.JSCell_typeInfoType),
|
|
m_out.constInt32(ScopedArgumentsType));
|
|
|
|
default: {
|
|
DFG_ASSERT(m_graph, m_node, arrayMode.isSomeTypedArrayView());
|
|
if (arrayMode.type() == Array::AnyTypedArray)
|
|
return isTypedArrayView(cell);
|
|
return m_out.equal(
|
|
m_out.load8ZeroExt32(cell, m_heaps.JSCell_typeInfoType),
|
|
m_out.constInt32(typeForTypedArrayType(arrayMode.typedArrayType())));
|
|
}
|
|
}
|
|
}
|
|
|
|
LValue isFunction(LValue cell, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type & SpecCell, SpecFunction))
|
|
return proven;
|
|
return isType(cell, JSFunctionType);
|
|
}
|
|
LValue isNotFunction(LValue cell, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (LValue proven = isProvenValue(type & SpecCell, ~SpecFunction))
|
|
return proven;
|
|
return isNotType(cell, JSFunctionType);
|
|
}
|
|
|
|
LValue isExoticForTypeof(LValue cell, SpeculatedType type = SpecFullTop)
|
|
{
|
|
if (!(type & SpecObjectOther))
|
|
return m_out.booleanFalse;
|
|
return m_out.testNonZero32(
|
|
m_out.load8ZeroExt32(cell, m_heaps.JSCell_typeInfoFlags),
|
|
m_out.constInt32(MasqueradesAsUndefined | OverridesGetCallData));
|
|
}
|
|
|
|
LValue isType(LValue cell, JSType type)
|
|
{
|
|
return m_out.equal(
|
|
m_out.load8ZeroExt32(cell, m_heaps.JSCell_typeInfoType),
|
|
m_out.constInt32(type));
|
|
}
|
|
|
|
LValue isNotType(LValue cell, JSType type)
|
|
{
|
|
return m_out.logicalNot(isType(cell, type));
|
|
}
|
|
|
|
void speculateObject(Edge edge, LValue cell)
|
|
{
|
|
FTL_TYPE_CHECK(jsValueValue(cell), edge, SpecObject, isNotObject(cell));
|
|
}
|
|
|
|
void speculateObject(Edge edge)
|
|
{
|
|
speculateObject(edge, lowCell(edge));
|
|
}
|
|
|
|
void speculateArray(Edge edge, LValue cell)
|
|
{
|
|
FTL_TYPE_CHECK(
|
|
jsValueValue(cell), edge, SpecArray, isNotType(cell, ArrayType));
|
|
}
|
|
|
|
void speculateArray(Edge edge)
|
|
{
|
|
speculateArray(edge, lowCell(edge));
|
|
}
|
|
|
|
void speculateFunction(Edge edge, LValue cell)
|
|
{
|
|
FTL_TYPE_CHECK(jsValueValue(cell), edge, SpecFunction, isNotFunction(cell));
|
|
}
|
|
|
|
void speculateFunction(Edge edge)
|
|
{
|
|
speculateFunction(edge, lowCell(edge));
|
|
}
|
|
|
|
void speculateObjectOrOther(Edge edge)
|
|
{
|
|
if (!m_interpreter.needsTypeCheck(edge))
|
|
return;
|
|
|
|
LValue value = lowJSValue(edge, ManualOperandSpeculation);
|
|
|
|
LBasicBlock cellCase = m_out.newBlock();
|
|
LBasicBlock primitiveCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(isNotCell(value, provenType(edge)), unsure(primitiveCase), unsure(cellCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(cellCase, primitiveCase);
|
|
|
|
FTL_TYPE_CHECK(
|
|
jsValueValue(value), edge, (~SpecCellCheck) | SpecObject, isNotObject(value));
|
|
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(primitiveCase, continuation);
|
|
|
|
FTL_TYPE_CHECK(
|
|
jsValueValue(value), edge, SpecCellCheck | SpecOther, isNotOther(value));
|
|
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
}
|
|
|
|
void speculateFinalObject(Edge edge, LValue cell)
|
|
{
|
|
FTL_TYPE_CHECK(
|
|
jsValueValue(cell), edge, SpecFinalObject, isNotType(cell, FinalObjectType));
|
|
}
|
|
|
|
void speculateFinalObject(Edge edge)
|
|
{
|
|
speculateFinalObject(edge, lowCell(edge));
|
|
}
|
|
|
|
void speculateRegExpObject(Edge edge, LValue cell)
|
|
{
|
|
FTL_TYPE_CHECK(
|
|
jsValueValue(cell), edge, SpecRegExpObject, isNotType(cell, RegExpObjectType));
|
|
}
|
|
|
|
void speculateRegExpObject(Edge edge)
|
|
{
|
|
speculateRegExpObject(edge, lowCell(edge));
|
|
}
|
|
|
|
void speculateProxyObject(Edge edge, LValue cell)
|
|
{
|
|
FTL_TYPE_CHECK(
|
|
jsValueValue(cell), edge, SpecProxyObject, isNotType(cell, ProxyObjectType));
|
|
}
|
|
|
|
void speculateProxyObject(Edge edge)
|
|
{
|
|
speculateProxyObject(edge, lowCell(edge));
|
|
}
|
|
|
|
void speculateDerivedArray(Edge edge, LValue cell)
|
|
{
|
|
FTL_TYPE_CHECK(
|
|
jsValueValue(cell), edge, SpecDerivedArray, isNotType(cell, DerivedArrayType));
|
|
}
|
|
|
|
void speculateDerivedArray(Edge edge)
|
|
{
|
|
speculateDerivedArray(edge, lowCell(edge));
|
|
}
|
|
|
|
void speculatePromiseObject(Edge edge, LValue cell)
|
|
{
|
|
FTL_TYPE_CHECK(
|
|
jsValueValue(cell), edge, SpecPromiseObject, isNotType(cell, JSPromiseType));
|
|
}
|
|
|
|
void speculatePromiseObject(Edge edge)
|
|
{
|
|
speculatePromiseObject(edge, lowCell(edge));
|
|
}
|
|
|
|
void speculateDateObject(Edge edge, LValue cell)
|
|
{
|
|
FTL_TYPE_CHECK(
|
|
jsValueValue(cell), edge, SpecDateObject, isNotType(cell, JSDateType));
|
|
}
|
|
|
|
void speculateDateObject(Edge edge)
|
|
{
|
|
speculateDateObject(edge, lowCell(edge));
|
|
}
|
|
|
|
void speculateMapObject(Edge edge, LValue cell)
|
|
{
|
|
FTL_TYPE_CHECK(
|
|
jsValueValue(cell), edge, SpecMapObject, isNotType(cell, JSMapType));
|
|
}
|
|
|
|
void speculateMapObject(Edge edge)
|
|
{
|
|
speculateMapObject(edge, lowCell(edge));
|
|
}
|
|
|
|
void speculateSetObject(Edge edge, LValue cell)
|
|
{
|
|
FTL_TYPE_CHECK(
|
|
jsValueValue(cell), edge, SpecSetObject, isNotType(cell, JSSetType));
|
|
}
|
|
|
|
void speculateSetObject(Edge edge)
|
|
{
|
|
speculateSetObject(edge, lowCell(edge));
|
|
}
|
|
|
|
void speculateWeakMapObject(Edge edge, LValue cell)
|
|
{
|
|
FTL_TYPE_CHECK(
|
|
jsValueValue(cell), edge, SpecWeakMapObject, isNotType(cell, JSWeakMapType));
|
|
}
|
|
|
|
void speculateWeakMapObject(Edge edge)
|
|
{
|
|
speculateWeakMapObject(edge, lowCell(edge));
|
|
}
|
|
|
|
void speculateWeakSetObject(Edge edge, LValue cell)
|
|
{
|
|
FTL_TYPE_CHECK(
|
|
jsValueValue(cell), edge, SpecWeakSetObject, isNotType(cell, JSWeakSetType));
|
|
}
|
|
|
|
void speculateWeakSetObject(Edge edge)
|
|
{
|
|
speculateWeakSetObject(edge, lowCell(edge));
|
|
}
|
|
|
|
void speculateDataViewObject(Edge edge, LValue cell)
|
|
{
|
|
FTL_TYPE_CHECK(
|
|
jsValueValue(cell), edge, SpecDataViewObject, isNotType(cell, DataViewType));
|
|
}
|
|
|
|
void speculateDataViewObject(Edge edge)
|
|
{
|
|
speculateDataViewObject(edge, lowCell(edge));
|
|
}
|
|
|
|
void speculateString(Edge edge, LValue cell)
|
|
{
|
|
FTL_TYPE_CHECK(jsValueValue(cell), edge, SpecString, isNotString(cell));
|
|
}
|
|
|
|
void speculateString(Edge edge)
|
|
{
|
|
speculateString(edge, lowCell(edge));
|
|
}
|
|
|
|
void speculateStringOrOther(Edge edge, LValue value)
|
|
{
|
|
if (!m_interpreter.needsTypeCheck(edge))
|
|
return;
|
|
|
|
LBasicBlock cellCase = m_out.newBlock();
|
|
LBasicBlock notCellCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(isCell(value, provenType(edge)), unsure(cellCase), unsure(notCellCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(cellCase, notCellCase);
|
|
|
|
FTL_TYPE_CHECK(jsValueValue(value), edge, (~SpecCellCheck) | SpecString, isNotString(value));
|
|
|
|
m_out.jump(continuation);
|
|
m_out.appendTo(notCellCase, continuation);
|
|
|
|
FTL_TYPE_CHECK(jsValueValue(value), edge, SpecCellCheck | SpecOther, isNotOther(value));
|
|
|
|
m_out.jump(continuation);
|
|
m_out.appendTo(continuation, lastNext);
|
|
}
|
|
|
|
void speculateStringOrOther(Edge edge)
|
|
{
|
|
speculateStringOrOther(edge, lowJSValue(edge, ManualOperandSpeculation));
|
|
}
|
|
|
|
void speculateStringIdent(Edge edge, LValue string, LValue stringImpl)
|
|
{
|
|
if (!m_interpreter.needsTypeCheck(edge, SpecStringIdent | ~SpecString))
|
|
return;
|
|
|
|
speculate(BadType, jsValueValue(string), edge.node(), isRopeString(string));
|
|
speculate(
|
|
BadType, jsValueValue(string), edge.node(),
|
|
m_out.testIsZero32(
|
|
m_out.load32(stringImpl, m_heaps.StringImpl_hashAndFlags),
|
|
m_out.constInt32(StringImpl::flagIsAtom())));
|
|
m_interpreter.filter(edge, SpecStringIdent | ~SpecString);
|
|
}
|
|
|
|
void speculateStringIdent(Edge edge)
|
|
{
|
|
lowStringIdent(edge);
|
|
}
|
|
|
|
void speculateStringObject(Edge edge)
|
|
{
|
|
if (!m_interpreter.needsTypeCheck(edge, SpecStringObject))
|
|
return;
|
|
|
|
speculateStringObjectForCell(edge, lowCell(edge));
|
|
}
|
|
|
|
void speculateStringOrStringObject(Edge edge)
|
|
{
|
|
if (!m_interpreter.needsTypeCheck(edge, SpecString | SpecStringObject))
|
|
return;
|
|
|
|
LValue cellBase = lowCell(edge);
|
|
if (!m_interpreter.needsTypeCheck(edge, SpecString | SpecStringObject))
|
|
return;
|
|
|
|
LBasicBlock notString = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue type = m_out.load8ZeroExt32(cellBase, m_heaps.JSCell_typeInfoType);
|
|
m_out.branch(
|
|
m_out.equal(type, m_out.constInt32(StringType)),
|
|
unsure(continuation), unsure(notString));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(notString, continuation);
|
|
speculate(
|
|
BadType, jsValueValue(cellBase), edge.node(),
|
|
m_out.notEqual(type, m_out.constInt32(StringObjectType)));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
m_interpreter.filter(edge, SpecString | SpecStringObject);
|
|
}
|
|
|
|
void speculateStringObjectForCell(Edge edge, LValue cell)
|
|
{
|
|
if (!m_interpreter.needsTypeCheck(edge, SpecStringObject))
|
|
return;
|
|
|
|
LValue type = m_out.load8ZeroExt32(cell, m_heaps.JSCell_typeInfoType);
|
|
FTL_TYPE_CHECK(jsValueValue(cell), edge, SpecStringObject, m_out.notEqual(type, m_out.constInt32(StringObjectType)));
|
|
}
|
|
|
|
void speculateSymbol(Edge edge, LValue cell)
|
|
{
|
|
FTL_TYPE_CHECK(jsValueValue(cell), edge, SpecSymbol, isNotSymbol(cell));
|
|
}
|
|
|
|
void speculateSymbol(Edge edge)
|
|
{
|
|
speculateSymbol(edge, lowCell(edge));
|
|
}
|
|
|
|
void speculateHeapBigInt(Edge edge, LValue cell)
|
|
{
|
|
FTL_TYPE_CHECK(jsValueValue(cell), edge, SpecHeapBigInt, isNotHeapBigInt(cell));
|
|
}
|
|
void speculateHeapBigInt(Edge edge)
|
|
{
|
|
speculateHeapBigInt(edge, lowCell(edge));
|
|
}
|
|
|
|
#if USE(BIGINT32)
|
|
void speculateBigInt32(Edge edge)
|
|
{
|
|
LValue value = lowJSValue(edge, ManualOperandSpeculation);
|
|
FTL_TYPE_CHECK(jsValueValue(value), edge, SpecBigInt32, isNotBigInt32(value));
|
|
}
|
|
|
|
void speculateAnyBigInt(Edge edge)
|
|
{
|
|
LValue value = lowJSValue(edge, ManualOperandSpeculation);
|
|
FTL_TYPE_CHECK(jsValueValue(value), edge, SpecBigInt, isNotAnyBigInt(value));
|
|
}
|
|
#endif
|
|
|
|
void speculateNonNullObject(Edge edge, LValue cell)
|
|
{
|
|
FTL_TYPE_CHECK(jsValueValue(cell), edge, SpecObject, isNotObject(cell));
|
|
if (masqueradesAsUndefinedWatchpointIsStillValid())
|
|
return;
|
|
|
|
speculate(
|
|
BadType, jsValueValue(cell), edge.node(),
|
|
m_out.testNonZero32(
|
|
m_out.load8ZeroExt32(cell, m_heaps.JSCell_typeInfoFlags),
|
|
m_out.constInt32(MasqueradesAsUndefined)));
|
|
}
|
|
|
|
void speculateNumber(Edge edge)
|
|
{
|
|
LValue value = lowJSValue(edge, ManualOperandSpeculation);
|
|
FTL_TYPE_CHECK(jsValueValue(value), edge, SpecBytecodeNumber, isNotNumber(value));
|
|
}
|
|
|
|
void speculateRealNumber(Edge edge)
|
|
{
|
|
// Do an early return here because lowDouble() can create a lot of control flow.
|
|
if (!m_interpreter.needsTypeCheck(edge))
|
|
return;
|
|
|
|
LValue value = lowJSValue(edge, ManualOperandSpeculation);
|
|
LValue doubleValue = unboxDouble(value);
|
|
|
|
LBasicBlock intCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
m_out.doubleEqual(doubleValue, doubleValue),
|
|
usually(continuation), rarely(intCase));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(intCase, continuation);
|
|
|
|
typeCheck(
|
|
jsValueValue(value), m_node->child1(), SpecBytecodeRealNumber,
|
|
isNotInt32(value, provenType(m_node->child1()) & ~SpecFullDouble));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
}
|
|
|
|
void speculateDoubleRepReal(Edge edge)
|
|
{
|
|
// Do an early return here because lowDouble() can create a lot of control flow.
|
|
if (!m_interpreter.needsTypeCheck(edge))
|
|
return;
|
|
|
|
LValue value = lowDouble(edge);
|
|
FTL_TYPE_CHECK(
|
|
doubleValue(value), edge, SpecDoubleReal,
|
|
m_out.doubleNotEqualOrUnordered(value, value));
|
|
}
|
|
|
|
void speculateDoubleRepAnyInt(Edge edge)
|
|
{
|
|
if (!m_interpreter.needsTypeCheck(edge))
|
|
return;
|
|
|
|
doubleToStrictInt52(edge, lowDouble(edge));
|
|
}
|
|
|
|
void speculateBoolean(Edge edge)
|
|
{
|
|
lowBoolean(edge);
|
|
}
|
|
|
|
void speculateNotStringVar(Edge edge)
|
|
{
|
|
if (!m_interpreter.needsTypeCheck(edge, ~SpecStringVar))
|
|
return;
|
|
|
|
LValue value = lowJSValue(edge, ManualOperandSpeculation);
|
|
|
|
LBasicBlock isCellCase = m_out.newBlock();
|
|
LBasicBlock isStringCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(isCell(value, provenType(edge)), unsure(isCellCase), unsure(continuation));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isCellCase, isStringCase);
|
|
m_out.branch(isString(value, provenType(edge)), unsure(isStringCase), unsure(continuation));
|
|
|
|
m_out.appendTo(isStringCase, continuation);
|
|
speculateStringIdent(edge, value, m_out.loadPtr(value, m_heaps.JSString_value));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
}
|
|
|
|
void speculateNotSymbol(Edge edge)
|
|
{
|
|
if (!m_interpreter.needsTypeCheck(edge, ~SpecSymbol))
|
|
return;
|
|
|
|
ASSERT(mayHaveTypeCheck(edge.useKind()));
|
|
LValue value = lowJSValue(edge, ManualOperandSpeculation);
|
|
|
|
LBasicBlock isCellCase = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(isCell(value, provenType(edge)), unsure(isCellCase), unsure(continuation));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isCellCase, continuation);
|
|
speculate(BadType, jsValueValue(value), edge.node(), isSymbol(value));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
|
|
m_interpreter.filter(edge, ~SpecSymbol);
|
|
}
|
|
|
|
void speculateOther(Edge edge)
|
|
{
|
|
if (!m_interpreter.needsTypeCheck(edge))
|
|
return;
|
|
|
|
LValue value = lowJSValue(edge, ManualOperandSpeculation);
|
|
typeCheck(jsValueValue(value), edge, SpecOther, isNotOther(value));
|
|
}
|
|
|
|
void speculateMisc(Edge edge)
|
|
{
|
|
if (!m_interpreter.needsTypeCheck(edge))
|
|
return;
|
|
|
|
LValue value = lowJSValue(edge, ManualOperandSpeculation);
|
|
typeCheck(jsValueValue(value), edge, SpecMisc, isNotMisc(value));
|
|
}
|
|
|
|
void speculateTypedArrayIsNotDetached(LValue base)
|
|
{
|
|
LBasicBlock isWasteful = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue mode = m_out.load32(base, m_heaps.JSArrayBufferView_mode);
|
|
m_out.branch(m_out.equal(mode, m_out.constInt32(WastefulTypedArray)),
|
|
unsure(isWasteful), unsure(continuation));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(isWasteful, continuation);
|
|
LValue vector = m_out.loadPtr(base, m_heaps.JSArrayBufferView_vector);
|
|
// FIXME: We could probably make this a mask.
|
|
// https://bugs.webkit.org/show_bug.cgi?id=197701
|
|
vector = removeArrayPtrTag(vector);
|
|
speculate(Uncountable, jsValueValue(vector), m_node, m_out.isZero64(vector));
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
}
|
|
|
|
bool masqueradesAsUndefinedWatchpointIsStillValid()
|
|
{
|
|
return m_graph.masqueradesAsUndefinedWatchpointIsStillValid(m_origin.semantic);
|
|
}
|
|
|
|
LValue loadCellState(LValue base)
|
|
{
|
|
return m_out.load8ZeroExt32(base, m_heaps.JSCell_cellState);
|
|
}
|
|
|
|
void emitStoreBarrier(LValue base, bool isFenced)
|
|
{
|
|
LBasicBlock recheckPath = nullptr;
|
|
if (isFenced)
|
|
recheckPath = m_out.newBlock();
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(isFenced ? recheckPath : slowPath);
|
|
|
|
LValue threshold;
|
|
if (isFenced)
|
|
threshold = m_out.load32(m_out.absolute(vm().heap.addressOfBarrierThreshold()));
|
|
else
|
|
threshold = m_out.constInt32(blackThreshold);
|
|
|
|
m_out.branch(
|
|
m_out.above(loadCellState(base), threshold),
|
|
usually(continuation), rarely(isFenced ? recheckPath : slowPath));
|
|
|
|
if (isFenced) {
|
|
m_out.appendTo(recheckPath, slowPath);
|
|
|
|
m_out.fence(&m_heaps.root, &m_heaps.JSCell_cellState);
|
|
|
|
m_out.branch(
|
|
m_out.above(loadCellState(base), m_out.constInt32(blackThreshold)),
|
|
usually(continuation), rarely(slowPath));
|
|
}
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
|
|
LValue call = vmCall(Void, operationWriteBarrierSlowPath, m_vmValue, base);
|
|
m_heaps.decorateCCallRead(&m_heaps.root, call);
|
|
m_heaps.decorateCCallWrite(&m_heaps.JSCell_cellState, call);
|
|
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
}
|
|
|
|
void mutatorFence()
|
|
{
|
|
if (isX86()) {
|
|
m_out.fence(&m_heaps.root, nullptr);
|
|
return;
|
|
}
|
|
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(slowPath);
|
|
|
|
m_out.branch(
|
|
m_out.load8ZeroExt32(m_out.absolute(vm().heap.addressOfMutatorShouldBeFenced())),
|
|
rarely(slowPath), usually(continuation));
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
|
|
m_out.fence(&m_heaps.root, nullptr);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
}
|
|
|
|
void nukeStructureAndSetButterfly(LValue butterfly, LValue object)
|
|
{
|
|
if (isX86()) {
|
|
m_out.store32(
|
|
m_out.bitOr(
|
|
m_out.load32(object, m_heaps.JSCell_structureID),
|
|
m_out.constInt32(nukedStructureIDBit())),
|
|
object, m_heaps.JSCell_structureID);
|
|
m_out.fence(&m_heaps.root, nullptr);
|
|
m_out.storePtr(butterfly, object, m_heaps.JSObject_butterfly);
|
|
m_out.fence(&m_heaps.root, nullptr);
|
|
return;
|
|
}
|
|
|
|
LBasicBlock fastPath = m_out.newBlock();
|
|
LBasicBlock slowPath = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LBasicBlock lastNext = m_out.insertNewBlocksBefore(fastPath);
|
|
|
|
m_out.branch(
|
|
m_out.load8ZeroExt32(m_out.absolute(vm().heap.addressOfMutatorShouldBeFenced())),
|
|
rarely(slowPath), usually(fastPath));
|
|
|
|
m_out.appendTo(fastPath, slowPath);
|
|
|
|
m_out.storePtr(butterfly, object, m_heaps.JSObject_butterfly);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(slowPath, continuation);
|
|
|
|
m_out.store32(
|
|
m_out.bitOr(
|
|
m_out.load32(object, m_heaps.JSCell_structureID),
|
|
m_out.constInt32(nukedStructureIDBit())),
|
|
object, m_heaps.JSCell_structureID);
|
|
m_out.fence(&m_heaps.root, nullptr);
|
|
m_out.storePtr(butterfly, object, m_heaps.JSObject_butterfly);
|
|
m_out.fence(&m_heaps.root, nullptr);
|
|
m_out.jump(continuation);
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
}
|
|
|
|
template<typename OperationType, typename... Args>
|
|
LValue vmCall(LType type, OperationType function, Args&&... args)
|
|
{
|
|
static_assert(!std::is_same<OperationType, LValue>::value);
|
|
if constexpr (!std::is_same_v<FunctionPtr<OperationPtrTag>, OperationType>)
|
|
static_assert(FunctionTraits<OperationType>::cCallArity() == sizeof...(Args), "Sanity check");
|
|
callPreflight();
|
|
LValue result = m_out.call(type, m_out.operation(function), std::forward<Args>(args)...);
|
|
if (mayExit(m_graph, m_node))
|
|
callCheck();
|
|
else {
|
|
// We can't exit due to an exception, so we also can't throw an exception.
|
|
#ifndef NDEBUG
|
|
LBasicBlock crash = m_out.newBlock();
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
LValue exception = m_out.load64(m_out.absolute(vm().addressOfException()));
|
|
LValue hadException = m_out.notZero64(exception);
|
|
|
|
m_out.branch(
|
|
hadException, rarely(crash), usually(continuation));
|
|
|
|
LBasicBlock lastNext = m_out.appendTo(crash, continuation);
|
|
m_out.unreachable();
|
|
|
|
m_out.appendTo(continuation, lastNext);
|
|
#endif
|
|
}
|
|
return result;
|
|
}
|
|
|
|
void callPreflight(CodeOrigin codeOrigin)
|
|
{
|
|
CallSiteIndex callSiteIndex = m_ftlState.jitCode->common.codeOrigins->addCodeOrigin(codeOrigin);
|
|
m_out.store32(
|
|
m_out.constInt32(callSiteIndex.bits()),
|
|
tagFor(VirtualRegister(CallFrameSlot::argumentCountIncludingThis)));
|
|
#if !USE(BUILTIN_FRAME_ADDRESS) || ASSERT_ENABLED
|
|
m_out.storePtr(m_callFrame, m_out.absolute(&vm().topCallFrame));
|
|
#endif
|
|
}
|
|
|
|
void callPreflight()
|
|
{
|
|
callPreflight(codeOriginDescriptionOfCallSite());
|
|
}
|
|
|
|
CodeOrigin codeOriginDescriptionOfCallSite() const
|
|
{
|
|
CodeOrigin codeOrigin = m_origin.semantic;
|
|
if (m_node->op() == TailCallInlinedCaller
|
|
|| m_node->op() == TailCallVarargsInlinedCaller
|
|
|| m_node->op() == TailCallForwardVarargsInlinedCaller
|
|
|| m_node->op() == DirectTailCallInlinedCaller) {
|
|
// This case arises when you have a situation like this:
|
|
// foo makes a call to bar, bar is inlined in foo. bar makes a call
|
|
// to baz and baz is inlined in bar. And then baz makes a tail-call to jaz,
|
|
// and jaz is inlined in baz. We want the callframe for jaz to appear to
|
|
// have caller be bar.
|
|
codeOrigin = *codeOrigin.inlineCallFrame()->getCallerSkippingTailCalls();
|
|
}
|
|
|
|
return codeOrigin;
|
|
}
|
|
|
|
void callCheck()
|
|
{
|
|
JSGlobalObject* globalObject = m_graph.globalObjectFor(m_origin.semantic);
|
|
if (Options::useExceptionFuzz()) {
|
|
#if !USE(BUILTIN_FRAME_ADDRESS) || ASSERT_ENABLED
|
|
m_out.storePtr(m_callFrame, m_out.absolute(&vm().topCallFrame));
|
|
#endif
|
|
m_out.call(Void, m_out.operation(operationExceptionFuzz), weakPointer(globalObject));
|
|
}
|
|
|
|
LValue exception = m_out.load64(m_out.absolute(vm().addressOfException()));
|
|
LValue hadException = m_out.notZero64(exception);
|
|
|
|
CodeOrigin opCatchOrigin;
|
|
HandlerInfo* exceptionHandler;
|
|
if (m_graph.willCatchExceptionInMachineFrame(m_origin.forExit, opCatchOrigin, exceptionHandler)) {
|
|
bool exitOK = true;
|
|
bool isExceptionHandler = true;
|
|
appendOSRExit(
|
|
ExceptionCheck, noValue(), nullptr, hadException,
|
|
m_origin.withForExitAndExitOK(opCatchOrigin, exitOK), isExceptionHandler);
|
|
return;
|
|
}
|
|
|
|
LBasicBlock continuation = m_out.newBlock();
|
|
|
|
m_out.branch(
|
|
hadException, rarely(m_handleExceptions), usually(continuation));
|
|
|
|
m_out.appendTo(continuation);
|
|
}
|
|
|
|
RefPtr<PatchpointExceptionHandle> preparePatchpointForExceptions(PatchpointValue* value)
|
|
{
|
|
CodeOrigin opCatchOrigin;
|
|
HandlerInfo* exceptionHandler;
|
|
bool willCatchException = m_graph.willCatchExceptionInMachineFrame(m_origin.forExit, opCatchOrigin, exceptionHandler);
|
|
if (!willCatchException)
|
|
return PatchpointExceptionHandle::defaultHandle(m_ftlState, m_nodeIndexInGraph);
|
|
|
|
dataLogLnIf(verboseCompilationEnabled(), " Patchpoint exception OSR exit #", m_ftlState.jitCode->osrExitDescriptors.size(), " with availability: ", availabilityMap());
|
|
|
|
bool exitOK = true;
|
|
NodeOrigin origin = m_origin.withForExitAndExitOK(opCatchOrigin, exitOK);
|
|
|
|
OSRExitDescriptor* exitDescriptor = appendOSRExitDescriptor(noValue(), nullptr);
|
|
|
|
// Compute the offset into the StackmapGenerationParams where we will find the exit arguments
|
|
// we are about to append. We need to account for both the children we've already added, and
|
|
// for the possibility of a result value if the patchpoint is not void.
|
|
unsigned offset = value->numChildren();
|
|
if (value->type() != Void)
|
|
offset++;
|
|
|
|
// Use LateColdAny to ensure that the stackmap arguments interfere with the patchpoint's
|
|
// result and with any late-clobbered registers.
|
|
value->appendVectorWithRep(
|
|
buildExitArguments(exitDescriptor, opCatchOrigin, noValue()),
|
|
ValueRep::LateColdAny);
|
|
|
|
return PatchpointExceptionHandle::create(
|
|
m_ftlState, exitDescriptor, origin, m_nodeIndexInGraph, offset, *exceptionHandler);
|
|
}
|
|
|
|
LBasicBlock lowBlock(DFG::BasicBlock* block)
|
|
{
|
|
return m_blocks.get(block);
|
|
}
|
|
|
|
OSRExitDescriptor* appendOSRExitDescriptor(FormattedValue lowValue, Node* highValue)
|
|
{
|
|
return appendOSRExitDescriptor(lowValue, m_graph.methodOfGettingAValueProfileFor(m_node, highValue));
|
|
}
|
|
|
|
OSRExitDescriptor* appendOSRExitDescriptor(FormattedValue lowValue, const MethodOfGettingAValueProfile& profile)
|
|
{
|
|
return &m_ftlState.jitCode->osrExitDescriptors.alloc(
|
|
lowValue.format(), profile,
|
|
availabilityMap().m_locals.numberOfArguments(),
|
|
availabilityMap().m_locals.numberOfLocals(),
|
|
availabilityMap().m_locals.numberOfTmps());
|
|
}
|
|
|
|
void appendOSRExit(
|
|
ExitKind kind, FormattedValue lowValue, Node* highValue, LValue failCondition,
|
|
NodeOrigin origin, bool isExceptionHandler = false)
|
|
{
|
|
return appendOSRExit(kind, lowValue, m_graph.methodOfGettingAValueProfileFor(m_node, highValue),
|
|
failCondition, origin, isExceptionHandler);
|
|
}
|
|
|
|
void appendOSRExit(
|
|
ExitKind kind, FormattedValue lowValue, const MethodOfGettingAValueProfile& profile, LValue failCondition,
|
|
NodeOrigin origin, bool isExceptionHandler = false)
|
|
{
|
|
dataLogLnIf(verboseCompilationEnabled(), " OSR exit #", m_ftlState.jitCode->osrExitDescriptors.size(), " with availability: ", availabilityMap());
|
|
|
|
DFG_ASSERT(m_graph, m_node, origin.exitOK);
|
|
|
|
if (!isExceptionHandler
|
|
&& Options::useOSRExitFuzz()
|
|
&& canUseOSRExitFuzzing(m_graph.baselineCodeBlockFor(m_origin.semantic))
|
|
&& doOSRExitFuzzing()) {
|
|
LValue numberOfFuzzChecks = m_out.add(
|
|
m_out.load32(m_out.absolute(&g_numberOfOSRExitFuzzChecks)),
|
|
m_out.int32One);
|
|
|
|
m_out.store32(numberOfFuzzChecks, m_out.absolute(&g_numberOfOSRExitFuzzChecks));
|
|
|
|
if (unsigned atOrAfter = Options::fireOSRExitFuzzAtOrAfter()) {
|
|
failCondition = m_out.bitOr(
|
|
failCondition,
|
|
m_out.aboveOrEqual(numberOfFuzzChecks, m_out.constInt32(atOrAfter)));
|
|
}
|
|
if (unsigned at = Options::fireOSRExitFuzzAt()) {
|
|
failCondition = m_out.bitOr(
|
|
failCondition,
|
|
m_out.equal(numberOfFuzzChecks, m_out.constInt32(at)));
|
|
}
|
|
}
|
|
|
|
if (failCondition == m_out.booleanFalse)
|
|
return;
|
|
|
|
blessSpeculation(
|
|
m_out.speculate(failCondition), kind, lowValue, profile, origin);
|
|
}
|
|
|
|
void blessSpeculation(CheckValue* value, ExitKind kind, FormattedValue lowValue, Node* highValue, NodeOrigin origin)
|
|
{
|
|
blessSpeculation(value, kind, lowValue, m_graph.methodOfGettingAValueProfileFor(m_node, highValue), origin);
|
|
}
|
|
|
|
void blessSpeculation(CheckValue* value, ExitKind kind, FormattedValue lowValue, const MethodOfGettingAValueProfile& profile, NodeOrigin origin)
|
|
{
|
|
OSRExitDescriptor* exitDescriptor = appendOSRExitDescriptor(lowValue, profile);
|
|
|
|
value->appendColdAnys(buildExitArguments(exitDescriptor, origin.forExit, lowValue));
|
|
|
|
State* state = &m_ftlState;
|
|
auto nodeIndex = m_nodeIndexInGraph;
|
|
value->setGenerator(
|
|
[=] (CCallHelpers& jit, const B3::StackmapGenerationParams& params) {
|
|
exitDescriptor->emitOSRExit(
|
|
*state, kind, origin, jit, params, nodeIndex, 0);
|
|
});
|
|
}
|
|
|
|
StackmapArgumentList buildExitArguments(
|
|
OSRExitDescriptor* exitDescriptor, CodeOrigin exitOrigin, FormattedValue lowValue,
|
|
unsigned offsetOfExitArgumentsInStackmapLocations = 0)
|
|
{
|
|
StackmapArgumentList result;
|
|
buildExitArguments(
|
|
exitDescriptor, exitOrigin, result, lowValue, offsetOfExitArgumentsInStackmapLocations);
|
|
return result;
|
|
}
|
|
|
|
void buildExitArguments(
|
|
OSRExitDescriptor* exitDescriptor, CodeOrigin exitOrigin, StackmapArgumentList& arguments, FormattedValue lowValue,
|
|
unsigned offsetOfExitArgumentsInStackmapLocations = 0)
|
|
{
|
|
if (!!lowValue)
|
|
arguments.append(lowValue.value());
|
|
|
|
AvailabilityMap availabilityMap = this->availabilityMap();
|
|
availabilityMap.pruneByLiveness(m_graph, exitOrigin);
|
|
|
|
HashMap<Node*, ExitTimeObjectMaterialization*> map;
|
|
availabilityMap.forEachAvailability(
|
|
[&] (Availability availability) {
|
|
if (!availability.shouldUseNode())
|
|
return;
|
|
|
|
Node* node = availability.node();
|
|
if (!node->isPhantomAllocation())
|
|
return;
|
|
|
|
auto result = map.add(node, nullptr);
|
|
if (result.isNewEntry) {
|
|
result.iterator->value =
|
|
exitDescriptor->m_materializations.add(node->op(), node->origin.semantic);
|
|
}
|
|
});
|
|
|
|
for (unsigned i = 0; i < exitDescriptor->m_values.size(); ++i) {
|
|
Operand operand = exitDescriptor->m_values.operandForIndex(i);
|
|
|
|
Availability availability = availabilityMap.m_locals[i];
|
|
|
|
// FIXME: It seems like we should be able to do at least some validation when OSR entering. https://bugs.webkit.org/show_bug.cgi?id=215511
|
|
if (Options::validateFTLOSRExitLiveness()
|
|
&& m_graph.m_plan.mode() != FTLForOSREntryMode) {
|
|
|
|
if (availability.isDead() && m_graph.isLiveInBytecode(operand, exitOrigin))
|
|
DFG_CRASH(m_graph, m_node, toCString("Live bytecode local not available: operand = ", operand, ", availability = ", availability, ", origin = ", exitOrigin).data());
|
|
}
|
|
ExitValue exitValue = exitValueForAvailability(arguments, map, availability);
|
|
if (exitValue.hasIndexInStackmapLocations())
|
|
exitValue.adjustStackmapLocationsIndexByOffset(offsetOfExitArgumentsInStackmapLocations);
|
|
exitDescriptor->m_values[i] = exitValue;
|
|
}
|
|
|
|
for (auto heapPair : availabilityMap.m_heap) {
|
|
Node* node = heapPair.key.base();
|
|
ExitTimeObjectMaterialization* materialization = map.get(node);
|
|
if (!materialization)
|
|
DFG_CRASH(m_graph, m_node, toCString("Could not find materialization for ", node, " in ", availabilityMap).data());
|
|
ExitValue exitValue = exitValueForAvailability(arguments, map, heapPair.value);
|
|
if (exitValue.hasIndexInStackmapLocations())
|
|
exitValue.adjustStackmapLocationsIndexByOffset(offsetOfExitArgumentsInStackmapLocations);
|
|
materialization->add(
|
|
heapPair.key.descriptor(),
|
|
exitValue);
|
|
}
|
|
|
|
if (verboseCompilationEnabled()) {
|
|
dataLog(" Exit values: ", exitDescriptor->m_values, "\n");
|
|
if (!exitDescriptor->m_materializations.isEmpty()) {
|
|
dataLog(" Materializations: \n");
|
|
for (ExitTimeObjectMaterialization* materialization : exitDescriptor->m_materializations)
|
|
dataLog(" ", pointerDump(materialization), "\n");
|
|
}
|
|
}
|
|
}
|
|
|
|
ExitValue exitValueForAvailability(
|
|
StackmapArgumentList& arguments, const HashMap<Node*, ExitTimeObjectMaterialization*>& map,
|
|
Availability availability)
|
|
{
|
|
FlushedAt flush = availability.flushedAt();
|
|
switch (flush.format()) {
|
|
case DeadFlush:
|
|
case ConflictingFlush:
|
|
if (availability.hasNode())
|
|
return exitValueForNode(arguments, map, availability.node());
|
|
|
|
// This means that the value is dead. It could be dead in bytecode or it could have
|
|
// been killed by our DCE, which can sometimes kill things even if they were live in
|
|
// bytecode.
|
|
return ExitValue::dead();
|
|
|
|
case FlushedJSValue:
|
|
case FlushedCell:
|
|
case FlushedBoolean:
|
|
return ExitValue::inJSStack(flush.virtualRegister());
|
|
|
|
case FlushedInt32:
|
|
return ExitValue::inJSStackAsInt32(flush.virtualRegister());
|
|
|
|
case FlushedInt52:
|
|
return ExitValue::inJSStackAsInt52(flush.virtualRegister());
|
|
|
|
case FlushedDouble:
|
|
return ExitValue::inJSStackAsDouble(flush.virtualRegister());
|
|
}
|
|
|
|
DFG_CRASH(m_graph, m_node, "Invalid flush format");
|
|
return ExitValue::dead();
|
|
}
|
|
|
|
ExitValue exitValueForNode(
|
|
StackmapArgumentList& arguments, const HashMap<Node*, ExitTimeObjectMaterialization*>& map,
|
|
Node* node)
|
|
{
|
|
// NOTE: In FTL->B3, we cannot generate code here, because m_output is positioned after the
|
|
// stackmap value. Like all values, the stackmap value cannot use a child that is defined after
|
|
// it.
|
|
|
|
ASSERT(node->shouldGenerate());
|
|
ASSERT(node->hasResult());
|
|
|
|
if (node) {
|
|
switch (node->op()) {
|
|
case BottomValue:
|
|
// This might arise in object materializations. I actually doubt that it would,
|
|
// but it seems worthwhile to be conservative.
|
|
return ExitValue::dead();
|
|
|
|
case JSConstant:
|
|
case Int52Constant:
|
|
case DoubleConstant:
|
|
return ExitValue::constant(node->asJSValue());
|
|
|
|
default:
|
|
if (node->isPhantomAllocation())
|
|
return ExitValue::materializeNewObject(map.get(node));
|
|
break;
|
|
}
|
|
}
|
|
|
|
LoweredNodeValue value = m_int32Values.get(node);
|
|
if (isValid(value))
|
|
return exitArgument(arguments, DataFormatInt32, value.value());
|
|
|
|
value = m_int52Values.get(node);
|
|
if (isValid(value))
|
|
return exitArgument(arguments, DataFormatInt52, value.value());
|
|
|
|
value = m_strictInt52Values.get(node);
|
|
if (isValid(value))
|
|
return exitArgument(arguments, DataFormatStrictInt52, value.value());
|
|
|
|
value = m_booleanValues.get(node);
|
|
if (isValid(value))
|
|
return exitArgument(arguments, DataFormatBoolean, value.value());
|
|
|
|
value = m_jsValueValues.get(node);
|
|
if (isValid(value))
|
|
return exitArgument(arguments, DataFormatJS, value.value());
|
|
|
|
value = m_doubleValues.get(node);
|
|
if (isValid(value))
|
|
return exitArgument(arguments, DataFormatDouble, value.value());
|
|
|
|
DFG_CRASH(m_graph, m_node, toCString("Cannot find value for node: ", node).data());
|
|
return ExitValue::dead();
|
|
}
|
|
|
|
ExitValue exitArgument(StackmapArgumentList& arguments, DataFormat format, LValue value)
|
|
{
|
|
ExitValue result = ExitValue::exitArgument(ExitArgument(format, arguments.size()));
|
|
arguments.append(value);
|
|
return result;
|
|
}
|
|
|
|
ExitValue exitValueForTailCall(StackmapArgumentList& arguments, Node* node)
|
|
{
|
|
ASSERT(node->shouldGenerate());
|
|
ASSERT(node->hasResult());
|
|
|
|
switch (node->op()) {
|
|
case JSConstant:
|
|
case Int52Constant:
|
|
case DoubleConstant:
|
|
return ExitValue::constant(node->asJSValue());
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
LoweredNodeValue value = m_jsValueValues.get(node);
|
|
if (isValid(value))
|
|
return exitArgument(arguments, DataFormatJS, value.value());
|
|
|
|
value = m_int32Values.get(node);
|
|
if (isValid(value))
|
|
return exitArgument(arguments, DataFormatJS, boxInt32(value.value()));
|
|
|
|
value = m_booleanValues.get(node);
|
|
if (isValid(value))
|
|
return exitArgument(arguments, DataFormatJS, boxBoolean(value.value()));
|
|
|
|
// Doubles and Int52 have been converted by ValueRep()
|
|
DFG_CRASH(m_graph, m_node, toCString("Cannot find value for node: ", node).data());
|
|
}
|
|
|
|
void setInt32(Node* node, LValue value)
|
|
{
|
|
m_int32Values.set(node, LoweredNodeValue(value, m_highBlock));
|
|
}
|
|
void setInt52(Node* node, LValue value)
|
|
{
|
|
m_int52Values.set(node, LoweredNodeValue(value, m_highBlock));
|
|
}
|
|
void setStrictInt52(Node* node, LValue value)
|
|
{
|
|
m_strictInt52Values.set(node, LoweredNodeValue(value, m_highBlock));
|
|
}
|
|
void setInt52(Node* node, LValue value, Int52Kind kind)
|
|
{
|
|
switch (kind) {
|
|
case Int52:
|
|
setInt52(node, value);
|
|
return;
|
|
|
|
case StrictInt52:
|
|
setStrictInt52(node, value);
|
|
return;
|
|
}
|
|
|
|
DFG_CRASH(m_graph, m_node, "Corrupt int52 kind");
|
|
}
|
|
void setJSValue(Node* node, LValue value)
|
|
{
|
|
m_jsValueValues.set(node, LoweredNodeValue(value, m_highBlock));
|
|
}
|
|
void setBoolean(Node* node, LValue value)
|
|
{
|
|
m_booleanValues.set(node, LoweredNodeValue(value, m_highBlock));
|
|
}
|
|
void setStorage(Node* node, LValue value)
|
|
{
|
|
m_storageValues.set(node, LoweredNodeValue(value, m_highBlock));
|
|
}
|
|
void setDouble(Node* node, LValue value)
|
|
{
|
|
m_doubleValues.set(node, LoweredNodeValue(value, m_highBlock));
|
|
}
|
|
|
|
void setInt32(LValue value)
|
|
{
|
|
setInt32(m_node, value);
|
|
}
|
|
void setInt52(LValue value)
|
|
{
|
|
setInt52(m_node, value);
|
|
}
|
|
void setStrictInt52(LValue value)
|
|
{
|
|
setStrictInt52(m_node, value);
|
|
}
|
|
void setInt52(LValue value, Int52Kind kind)
|
|
{
|
|
setInt52(m_node, value, kind);
|
|
}
|
|
void setJSValue(LValue value)
|
|
{
|
|
setJSValue(m_node, value);
|
|
}
|
|
void setBoolean(LValue value)
|
|
{
|
|
setBoolean(m_node, value);
|
|
}
|
|
void setStorage(LValue value)
|
|
{
|
|
setStorage(m_node, value);
|
|
}
|
|
void setDouble(LValue value)
|
|
{
|
|
setDouble(m_node, value);
|
|
}
|
|
|
|
bool isValid(const LoweredNodeValue& value)
|
|
{
|
|
if (!value)
|
|
return false;
|
|
if (!m_graph.m_ssaDominators->dominates(value.block(), m_highBlock))
|
|
return false;
|
|
return true;
|
|
}
|
|
|
|
void ensureStillAliveHere(LValue value)
|
|
{
|
|
PatchpointValue* patchpoint = m_out.patchpoint(Void);
|
|
patchpoint->effects = Effects::none();
|
|
patchpoint->effects.writesLocalState = true;
|
|
patchpoint->effects.reads = HeapRange::top();
|
|
patchpoint->append(value, ValueRep::ColdAny);
|
|
patchpoint->setGenerator([=] (CCallHelpers&, const StackmapGenerationParams&) { });
|
|
}
|
|
|
|
LValue toButterfly(LValue immutableButterfly)
|
|
{
|
|
return m_out.addPtr(immutableButterfly, JSImmutableButterfly::offsetOfData());
|
|
}
|
|
|
|
void addWeakReference(JSCell* target)
|
|
{
|
|
m_graph.m_plan.weakReferences().addLazily(target);
|
|
}
|
|
|
|
LValue loadStructure(LValue value)
|
|
{
|
|
LValue structureID = m_out.load32(value, m_heaps.JSCell_structureID);
|
|
LValue tableBase = m_out.loadPtr(m_out.absolute(vm().heap.structureIDTable().base()));
|
|
LValue tableIndex = m_out.aShr(structureID, m_out.constInt32(StructureIDTable::s_numberOfEntropyBits));
|
|
LValue entropyBits = m_out.shl(m_out.zeroExtPtr(structureID), m_out.constInt32(StructureIDTable::s_entropyBitsShiftForStructurePointer));
|
|
TypedPointer address = m_out.baseIndex(m_heaps.structureTable, tableBase, m_out.zeroExtPtr(tableIndex));
|
|
LValue encodedStructureBits = m_out.loadPtr(address);
|
|
return m_out.bitXor(encodedStructureBits, entropyBits);
|
|
}
|
|
|
|
LValue weakPointer(JSCell* pointer)
|
|
{
|
|
addWeakReference(pointer);
|
|
return m_out.alreadyRegisteredWeakPointer(m_graph, pointer);
|
|
}
|
|
|
|
LValue frozenPointer(FrozenValue* value)
|
|
{
|
|
return m_out.alreadyRegisteredFrozenPointer(value);
|
|
}
|
|
|
|
LValue weakStructureID(RegisteredStructure structure)
|
|
{
|
|
return m_out.constInt32(structure->id());
|
|
}
|
|
|
|
LValue weakStructure(RegisteredStructure structure)
|
|
{
|
|
ASSERT(!!structure.get());
|
|
return m_out.alreadyRegisteredWeakPointer(m_graph, structure.get());
|
|
}
|
|
|
|
TypedPointer addressFor(LValue base, Operand operand, ptrdiff_t offset = 0)
|
|
{
|
|
return m_out.address(base, m_heaps.variables[operand.virtualRegister().offset()], offset);
|
|
}
|
|
TypedPointer payloadFor(LValue base, Operand operand)
|
|
{
|
|
return addressFor(base, operand, PayloadOffset);
|
|
}
|
|
TypedPointer tagFor(LValue base, Operand operand)
|
|
{
|
|
return addressFor(base, operand, TagOffset);
|
|
}
|
|
TypedPointer addressFor(Operand operand, ptrdiff_t offset = 0)
|
|
{
|
|
return addressFor(operand.virtualRegister(), offset);
|
|
}
|
|
TypedPointer addressFor(VirtualRegister operand, ptrdiff_t offset = 0)
|
|
{
|
|
if (operand.isLocal())
|
|
return addressFor(m_captured, operand, offset);
|
|
return addressFor(m_callFrame, operand, offset);
|
|
}
|
|
TypedPointer payloadFor(Operand operand)
|
|
{
|
|
return payloadFor(operand.virtualRegister());
|
|
}
|
|
TypedPointer payloadFor(VirtualRegister operand)
|
|
{
|
|
return addressFor(operand, PayloadOffset);
|
|
}
|
|
TypedPointer tagFor(Operand operand)
|
|
{
|
|
return tagFor(operand.virtualRegister());
|
|
}
|
|
TypedPointer tagFor(VirtualRegister operand)
|
|
{
|
|
return addressFor(operand, TagOffset);
|
|
}
|
|
|
|
AbstractValue abstractValue(Node* node)
|
|
{
|
|
return m_state.forNode(node);
|
|
}
|
|
AbstractValue abstractValue(Edge edge)
|
|
{
|
|
return abstractValue(edge.node());
|
|
}
|
|
|
|
SpeculatedType provenType(Node* node)
|
|
{
|
|
return abstractValue(node).m_type;
|
|
}
|
|
SpeculatedType provenType(Edge edge)
|
|
{
|
|
return provenType(edge.node());
|
|
}
|
|
|
|
JSValue provenValue(Node* node)
|
|
{
|
|
return abstractValue(node).m_value;
|
|
}
|
|
JSValue provenValue(Edge edge)
|
|
{
|
|
return provenValue(edge.node());
|
|
}
|
|
|
|
StructureAbstractValue abstractStructure(Node* node)
|
|
{
|
|
return abstractValue(node).m_structure;
|
|
}
|
|
StructureAbstractValue abstractStructure(Edge edge)
|
|
{
|
|
return abstractStructure(edge.node());
|
|
}
|
|
|
|
void crash()
|
|
{
|
|
crash(m_highBlock, m_node);
|
|
}
|
|
void crash(DFG::BasicBlock* block, Node* node)
|
|
{
|
|
BlockIndex blockIndex = block->index;
|
|
unsigned nodeIndex = node ? node->index() : UINT_MAX;
|
|
#if !ASSERT_ENABLED
|
|
m_out.patchpoint(Void)->setGenerator(
|
|
[=] (CCallHelpers& jit, const StackmapGenerationParams&) {
|
|
AllowMacroScratchRegisterUsage allowScratch(jit);
|
|
|
|
jit.move(CCallHelpers::TrustedImm32(blockIndex), GPRInfo::regT0);
|
|
jit.move(CCallHelpers::TrustedImm32(nodeIndex), GPRInfo::regT1);
|
|
if (node)
|
|
jit.move(CCallHelpers::TrustedImm32(node->op()), GPRInfo::regT2);
|
|
jit.abortWithReason(FTLCrash);
|
|
});
|
|
#else // ASSERT_ENABLED
|
|
m_out.call(
|
|
Void,
|
|
m_out.constIntPtr(ftlUnreachable),
|
|
// We don't want the CodeBlock to have a weak pointer to itself because
|
|
// that would cause it to always get collected.
|
|
m_out.constIntPtr(bitwise_cast<intptr_t>(codeBlock())), m_out.constInt32(blockIndex),
|
|
m_out.constInt32(nodeIndex));
|
|
#endif // ASSERT_ENABLED
|
|
m_out.unreachable();
|
|
}
|
|
|
|
AvailabilityMap& availabilityMap() { return m_availabilityCalculator.m_availability; }
|
|
|
|
VM& vm() { return m_graph.m_vm; }
|
|
CodeBlock* codeBlock() { return m_graph.m_codeBlock; }
|
|
|
|
Graph& m_graph;
|
|
State& m_ftlState;
|
|
AbstractHeapRepository m_heaps;
|
|
Output m_out;
|
|
Procedure& m_proc;
|
|
|
|
LBasicBlock m_handleExceptions;
|
|
HashMap<DFG::BasicBlock*, LBasicBlock> m_blocks;
|
|
|
|
LValue m_callFrame;
|
|
LValue m_vmValue;
|
|
LValue m_captured;
|
|
LValue m_numberTag;
|
|
LValue m_notCellMask;
|
|
|
|
HashMap<Node*, LoweredNodeValue> m_int32Values;
|
|
HashMap<Node*, LoweredNodeValue> m_strictInt52Values;
|
|
HashMap<Node*, LoweredNodeValue> m_int52Values;
|
|
HashMap<Node*, LoweredNodeValue> m_jsValueValues;
|
|
HashMap<Node*, LoweredNodeValue> m_booleanValues;
|
|
HashMap<Node*, LoweredNodeValue> m_storageValues;
|
|
HashMap<Node*, LoweredNodeValue> m_doubleValues;
|
|
|
|
HashMap<Node*, LValue> m_phis;
|
|
|
|
LocalOSRAvailabilityCalculator m_availabilityCalculator;
|
|
|
|
InPlaceAbstractState m_state;
|
|
AbstractInterpreter<InPlaceAbstractState> m_interpreter;
|
|
DFG::BasicBlock* m_highBlock;
|
|
DFG::BasicBlock* m_nextHighBlock;
|
|
LBasicBlock m_nextLowBlock;
|
|
|
|
NodeOrigin m_origin;
|
|
unsigned m_nodeIndexInGraph { 0 };
|
|
Node* m_node;
|
|
|
|
// These are used for validating AI state.
|
|
HashMap<Node*, NodeSet> m_liveInToNode;
|
|
HashMap<Node*, AbstractValue> m_aiCheckedNodes;
|
|
String m_graphDump;
|
|
};
|
|
|
|
} // anonymous namespace
|
|
|
|
void lowerDFGToB3(State& state)
|
|
{
|
|
LowerDFGToB3 lowering(state);
|
|
lowering.lower();
|
|
}
|
|
|
|
} } // namespace JSC::FTL
|
|
|
|
#endif // ENABLE(FTL_JIT)
|
|
|