mirror of
https://github.com/darlinghq/darling-JavaScriptCore.git
synced 2024-11-23 04:09:40 +00:00
314 lines
9.5 KiB
C++
314 lines
9.5 KiB
C++
/*
|
|
* Copyright (C) 2011-2018 Apple Inc. All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY APPLE INC. ``AS IS'' AND ANY
|
|
* EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
|
|
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR
|
|
* CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
|
|
* EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
|
|
* PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
|
|
* PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY
|
|
* OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#pragma once
|
|
|
|
#include "CodeBlock.h"
|
|
#include "CodeBlockHash.h"
|
|
#include "CodeOrigin.h"
|
|
#include "ValueRecovery.h"
|
|
#include "WriteBarrier.h"
|
|
#include <wtf/PrintStream.h>
|
|
#include <wtf/StdLibExtras.h>
|
|
#include <wtf/Vector.h>
|
|
|
|
namespace JSC {
|
|
|
|
struct InlineCallFrame;
|
|
class CallFrame;
|
|
class JSFunction;
|
|
|
|
struct InlineCallFrame {
|
|
enum Kind {
|
|
Call,
|
|
Construct,
|
|
TailCall,
|
|
CallVarargs,
|
|
ConstructVarargs,
|
|
TailCallVarargs,
|
|
|
|
// For these, the stackOffset incorporates the argument count plus the true return PC
|
|
// slot.
|
|
GetterCall,
|
|
SetterCall
|
|
};
|
|
|
|
static CallMode callModeFor(Kind kind)
|
|
{
|
|
switch (kind) {
|
|
case Call:
|
|
case CallVarargs:
|
|
case GetterCall:
|
|
case SetterCall:
|
|
return CallMode::Regular;
|
|
case TailCall:
|
|
case TailCallVarargs:
|
|
return CallMode::Tail;
|
|
case Construct:
|
|
case ConstructVarargs:
|
|
return CallMode::Construct;
|
|
}
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
}
|
|
|
|
static Kind kindFor(CallMode callMode)
|
|
{
|
|
switch (callMode) {
|
|
case CallMode::Regular:
|
|
return Call;
|
|
case CallMode::Construct:
|
|
return Construct;
|
|
case CallMode::Tail:
|
|
return TailCall;
|
|
}
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
}
|
|
|
|
static Kind varargsKindFor(CallMode callMode)
|
|
{
|
|
switch (callMode) {
|
|
case CallMode::Regular:
|
|
return CallVarargs;
|
|
case CallMode::Construct:
|
|
return ConstructVarargs;
|
|
case CallMode::Tail:
|
|
return TailCallVarargs;
|
|
}
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
}
|
|
|
|
static CodeSpecializationKind specializationKindFor(Kind kind)
|
|
{
|
|
switch (kind) {
|
|
case Call:
|
|
case CallVarargs:
|
|
case TailCall:
|
|
case TailCallVarargs:
|
|
case GetterCall:
|
|
case SetterCall:
|
|
return CodeForCall;
|
|
case Construct:
|
|
case ConstructVarargs:
|
|
return CodeForConstruct;
|
|
}
|
|
RELEASE_ASSERT_NOT_REACHED();
|
|
}
|
|
|
|
static bool isVarargs(Kind kind)
|
|
{
|
|
switch (kind) {
|
|
case CallVarargs:
|
|
case TailCallVarargs:
|
|
case ConstructVarargs:
|
|
return true;
|
|
default:
|
|
return false;
|
|
}
|
|
}
|
|
|
|
static bool isTail(Kind kind)
|
|
{
|
|
switch (kind) {
|
|
case TailCall:
|
|
case TailCallVarargs:
|
|
return true;
|
|
default:
|
|
return false;
|
|
}
|
|
}
|
|
bool isTail() const
|
|
{
|
|
return isTail(static_cast<Kind>(kind));
|
|
}
|
|
|
|
static CodeOrigin* computeCallerSkippingTailCalls(InlineCallFrame* inlineCallFrame, Kind* callerCallKind = nullptr)
|
|
{
|
|
CodeOrigin* codeOrigin;
|
|
bool tailCallee;
|
|
int callKind;
|
|
do {
|
|
tailCallee = inlineCallFrame->isTail();
|
|
callKind = inlineCallFrame->kind;
|
|
codeOrigin = &inlineCallFrame->directCaller;
|
|
inlineCallFrame = codeOrigin->inlineCallFrame();
|
|
} while (inlineCallFrame && tailCallee);
|
|
|
|
if (tailCallee)
|
|
return nullptr;
|
|
|
|
if (callerCallKind)
|
|
*callerCallKind = static_cast<Kind>(callKind);
|
|
|
|
return codeOrigin;
|
|
}
|
|
|
|
CodeOrigin* getCallerSkippingTailCalls(Kind* callerCallKind = nullptr)
|
|
{
|
|
return computeCallerSkippingTailCalls(this, callerCallKind);
|
|
}
|
|
|
|
InlineCallFrame* getCallerInlineFrameSkippingTailCalls()
|
|
{
|
|
CodeOrigin* caller = getCallerSkippingTailCalls();
|
|
return caller ? caller->inlineCallFrame() : nullptr;
|
|
}
|
|
|
|
Vector<ValueRecovery> argumentsWithFixup; // Includes 'this' and arity fixups.
|
|
WriteBarrier<CodeBlock> baselineCodeBlock;
|
|
CodeOrigin directCaller;
|
|
|
|
unsigned argumentCountIncludingThis : 22; // Do not include fixups.
|
|
unsigned tmpOffset : 10;
|
|
signed stackOffset : 28;
|
|
unsigned kind : 3; // real type is Kind
|
|
bool isClosureCall : 1; // If false then we know that callee/scope are constants and the DFG won't treat them as variables, i.e. they have to be recovered manually.
|
|
VirtualRegister argumentCountRegister; // Only set when we inline a varargs call.
|
|
|
|
ValueRecovery calleeRecovery;
|
|
|
|
// There is really no good notion of a "default" set of values for
|
|
// InlineCallFrame's fields. This constructor is here just to reduce confusion if
|
|
// we forgot to initialize explicitly.
|
|
InlineCallFrame()
|
|
: argumentCountIncludingThis(0)
|
|
, tmpOffset(0)
|
|
, stackOffset(0)
|
|
, kind(Call)
|
|
, isClosureCall(false)
|
|
{
|
|
}
|
|
|
|
bool isVarargs() const
|
|
{
|
|
return isVarargs(static_cast<Kind>(kind));
|
|
}
|
|
|
|
CodeSpecializationKind specializationKind() const { return specializationKindFor(static_cast<Kind>(kind)); }
|
|
|
|
JSFunction* calleeConstant() const;
|
|
|
|
// Get the callee given a machine call frame to which this InlineCallFrame belongs.
|
|
JSFunction* calleeForCallFrame(CallFrame*) const;
|
|
|
|
CString inferredName() const;
|
|
CodeBlockHash hash() const;
|
|
CString hashAsStringIfPossible() const;
|
|
|
|
void setStackOffset(signed offset)
|
|
{
|
|
stackOffset = offset;
|
|
RELEASE_ASSERT(static_cast<signed>(stackOffset) == offset);
|
|
}
|
|
|
|
void setTmpOffset(unsigned offset)
|
|
{
|
|
tmpOffset = offset;
|
|
RELEASE_ASSERT(static_cast<unsigned>(tmpOffset) == offset);
|
|
}
|
|
|
|
ptrdiff_t callerFrameOffset() const { return stackOffset * sizeof(Register) + CallFrame::callerFrameOffset(); }
|
|
ptrdiff_t returnPCOffset() const { return stackOffset * sizeof(Register) + CallFrame::returnPCOffset(); }
|
|
|
|
bool isInStrictContext() const { return baselineCodeBlock->ownerExecutable()->isInStrictContext(); }
|
|
|
|
void dumpBriefFunctionInformation(PrintStream&) const;
|
|
void dump(PrintStream&) const;
|
|
void dumpInContext(PrintStream&, DumpContext*) const;
|
|
|
|
MAKE_PRINT_METHOD(InlineCallFrame, dumpBriefFunctionInformation, briefFunctionInformation);
|
|
|
|
};
|
|
|
|
inline CodeBlock* baselineCodeBlockForInlineCallFrame(InlineCallFrame* inlineCallFrame)
|
|
{
|
|
RELEASE_ASSERT(inlineCallFrame);
|
|
return inlineCallFrame->baselineCodeBlock.get();
|
|
}
|
|
|
|
inline CodeBlock* baselineCodeBlockForOriginAndBaselineCodeBlock(const CodeOrigin& codeOrigin, CodeBlock* baselineCodeBlock)
|
|
{
|
|
ASSERT(JITCode::isBaselineCode(baselineCodeBlock->jitType()));
|
|
auto* inlineCallFrame = codeOrigin.inlineCallFrame();
|
|
if (inlineCallFrame)
|
|
return baselineCodeBlockForInlineCallFrame(inlineCallFrame);
|
|
return baselineCodeBlock;
|
|
}
|
|
|
|
// These function is defined here and not in CodeOrigin because it needs access to the directCaller field in InlineCallFrame
|
|
template <typename Function>
|
|
inline void CodeOrigin::walkUpInlineStack(const Function& function) const
|
|
{
|
|
CodeOrigin codeOrigin = *this;
|
|
while (true) {
|
|
function(codeOrigin);
|
|
auto* inlineCallFrame = codeOrigin.inlineCallFrame();
|
|
if (!inlineCallFrame)
|
|
break;
|
|
codeOrigin = inlineCallFrame->directCaller;
|
|
}
|
|
}
|
|
|
|
inline bool CodeOrigin::inlineStackContainsActiveCheckpoint() const
|
|
{
|
|
bool result = false;
|
|
walkUpInlineStack([&] (CodeOrigin origin) {
|
|
if (origin.bytecodeIndex().checkpoint())
|
|
result = true;
|
|
});
|
|
return result;
|
|
}
|
|
|
|
ALWAYS_INLINE Operand remapOperand(InlineCallFrame* inlineCallFrame, Operand operand)
|
|
{
|
|
if (inlineCallFrame)
|
|
return operand.isTmp() ? Operand::tmp(operand.value() + inlineCallFrame->tmpOffset) : operand.virtualRegister() + inlineCallFrame->stackOffset;
|
|
return operand;
|
|
}
|
|
|
|
ALWAYS_INLINE Operand remapOperand(InlineCallFrame* inlineCallFrame, VirtualRegister reg)
|
|
{
|
|
return remapOperand(inlineCallFrame, Operand(reg));
|
|
}
|
|
|
|
ALWAYS_INLINE Operand unmapOperand(InlineCallFrame* inlineCallFrame, Operand operand)
|
|
{
|
|
if (inlineCallFrame)
|
|
return operand.isTmp() ? Operand::tmp(operand.value() - inlineCallFrame->tmpOffset) : Operand(operand.virtualRegister() - inlineCallFrame->stackOffset);
|
|
return operand;
|
|
}
|
|
|
|
ALWAYS_INLINE Operand unmapOperand(InlineCallFrame* inlineCallFrame, VirtualRegister reg)
|
|
{
|
|
return unmapOperand(inlineCallFrame, Operand(reg));
|
|
}
|
|
|
|
} // namespace JSC
|
|
|
|
namespace WTF {
|
|
|
|
void printInternal(PrintStream&, JSC::InlineCallFrame::Kind);
|
|
|
|
} // namespace WTF
|