/* * Copyright (C) 2010-2020 Apple Inc. All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY APPLE INC. AND ITS CONTRIBUTORS ``AS IS'' * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, * THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR ITS CONTRIBUTORS * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF * THE POSSIBILITY OF SUCH DAMAGE. */ #include "config.h" #include "ThunkGenerators.h" #include "JITOperations.h" #include "JITThunks.h" #include "JSBoundFunction.h" #include "LLIntThunks.h" #include "MaxFrameExtentForSlowPathCall.h" #include "SpecializedThunkJIT.h" #include #include #include #if ENABLE(JIT) namespace JSC { template inline void emitPointerValidation(CCallHelpers& jit, GPRReg pointerGPR, TagType tag) { if (!ASSERT_ENABLED) return; if (!Options::useJITCage()) { CCallHelpers::Jump isNonZero = jit.branchTestPtr(CCallHelpers::NonZero, pointerGPR); jit.abortWithReason(TGInvalidPointer); isNonZero.link(&jit); jit.pushToSave(pointerGPR); jit.untagPtr(tag, pointerGPR); jit.validateUntaggedPtr(pointerGPR); jit.popToRestore(pointerGPR); } } // We will jump here if the JIT code tries to make a call, but the // linking helper (C++ code) decides to throw an exception instead. MacroAssemblerCodeRef throwExceptionFromCallSlowPathGenerator(VM& vm) { CCallHelpers jit; // The call pushed a return address, so we need to pop it back off to re-align the stack, // even though we won't use it. jit.preserveReturnAddressAfterCall(GPRInfo::nonPreservedNonReturnGPR); jit.copyCalleeSavesToEntryFrameCalleeSavesBuffer(vm.topEntryFrame); jit.setupArguments(CCallHelpers::TrustedImmPtr(&vm)); jit.prepareCallOperation(vm); jit.move(CCallHelpers::TrustedImmPtr(tagCFunction(operationLookupExceptionHandler)), GPRInfo::nonArgGPR0); emitPointerValidation(jit, GPRInfo::nonArgGPR0, OperationPtrTag); jit.call(GPRInfo::nonArgGPR0, OperationPtrTag); jit.jumpToExceptionHandler(vm); LinkBuffer patchBuffer(jit, GLOBAL_THUNK_ID); return FINALIZE_CODE(patchBuffer, JITThunkPtrTag, "Throw exception from call slow path thunk"); } static void slowPathFor(CCallHelpers& jit, VM& vm, Sprt_JITOperation_EGCli slowPathFunction) { jit.sanitizeStackInline(vm, GPRInfo::nonArgGPR0); jit.emitFunctionPrologue(); jit.storePtr(GPRInfo::callFrameRegister, &vm.topCallFrame); #if OS(WINDOWS) && CPU(X86_64) // Windows X86_64 needs some space pointed to by arg0 for return types larger than 64 bits. // Other argument values are shift by 1. Use space on the stack for our two return values. // Moving the stack down maxFrameExtentForSlowPathCall bytes gives us room for our 3 arguments // and space for the 16 byte return area. jit.addPtr(CCallHelpers::TrustedImm32(-static_cast(maxFrameExtentForSlowPathCall)), CCallHelpers::stackPointerRegister); static_assert(GPRInfo::regT2 != GPRInfo::argumentGPR0); static_assert(GPRInfo::regT3 != GPRInfo::argumentGPR0); jit.move(GPRInfo::regT2, GPRInfo::argumentGPR0); jit.move(GPRInfo::regT3, GPRInfo::argumentGPR2); jit.move(GPRInfo::argumentGPR0, GPRInfo::argumentGPR3); jit.addPtr(CCallHelpers::TrustedImm32(32), CCallHelpers::stackPointerRegister, GPRInfo::argumentGPR0); jit.move(GPRInfo::callFrameRegister, GPRInfo::argumentGPR1); jit.move(CCallHelpers::TrustedImmPtr(tagCFunction(slowPathFunction)), GPRInfo::nonArgGPR0); emitPointerValidation(jit, GPRInfo::nonArgGPR0, OperationPtrTag); jit.call(GPRInfo::nonArgGPR0, OperationPtrTag); jit.loadPtr(CCallHelpers::Address(GPRInfo::returnValueGPR, 8), GPRInfo::returnValueGPR2); jit.loadPtr(CCallHelpers::Address(GPRInfo::returnValueGPR), GPRInfo::returnValueGPR); jit.addPtr(CCallHelpers::TrustedImm32(maxFrameExtentForSlowPathCall), CCallHelpers::stackPointerRegister); #else if (maxFrameExtentForSlowPathCall) jit.addPtr(CCallHelpers::TrustedImm32(-maxFrameExtentForSlowPathCall), CCallHelpers::stackPointerRegister); jit.setupArguments(GPRInfo::regT3, GPRInfo::regT2); jit.move(CCallHelpers::TrustedImmPtr(tagCFunction(slowPathFunction)), GPRInfo::nonArgGPR0); emitPointerValidation(jit, GPRInfo::nonArgGPR0, OperationPtrTag); jit.call(GPRInfo::nonArgGPR0, OperationPtrTag); if (maxFrameExtentForSlowPathCall) jit.addPtr(CCallHelpers::TrustedImm32(maxFrameExtentForSlowPathCall), CCallHelpers::stackPointerRegister); #endif // This slow call will return the address of one of the following: // 1) Exception throwing thunk. // 2) Host call return value returner thingy. // 3) The function to call. // The second return value GPR will hold a non-zero value for tail calls. emitPointerValidation(jit, GPRInfo::returnValueGPR, JSEntryPtrTag); jit.emitFunctionEpilogue(); jit.untagReturnAddress(); RELEASE_ASSERT(reinterpret_cast(KeepTheFrame) == reinterpret_cast(0)); CCallHelpers::Jump doNotTrash = jit.branchTestPtr(CCallHelpers::Zero, GPRInfo::returnValueGPR2); jit.preserveReturnAddressAfterCall(GPRInfo::nonPreservedNonReturnGPR); jit.prepareForTailCallSlow(GPRInfo::returnValueGPR); doNotTrash.link(&jit); jit.farJump(GPRInfo::returnValueGPR, JSEntryPtrTag); } MacroAssemblerCodeRef linkCallThunkGenerator(VM& vm) { // The return address is on the stack or in the link register. We will hence // save the return address to the call frame while we make a C++ function call // to perform linking and lazy compilation if necessary. We expect the callee // to be in regT0/regT1 (payload/tag), the CallFrame to have already // been adjusted, and all other registers to be available for use. CCallHelpers jit; slowPathFor(jit, vm, operationLinkCall); LinkBuffer patchBuffer(jit, GLOBAL_THUNK_ID); return FINALIZE_CODE(patchBuffer, JITThunkPtrTag, "Link call slow path thunk"); } // For closure optimizations, we only include calls, since if you're using closures for // object construction then you're going to lose big time anyway. MacroAssemblerCodeRef linkPolymorphicCallThunkGenerator(VM& vm) { CCallHelpers jit; slowPathFor(jit, vm, operationLinkPolymorphicCall); LinkBuffer patchBuffer(jit, GLOBAL_THUNK_ID); return FINALIZE_CODE(patchBuffer, JITThunkPtrTag, "Link polymorphic call slow path thunk"); } // FIXME: We should distinguish between a megamorphic virtual call vs. a slow // path virtual call so that we can enable fast tail calls for megamorphic // virtual calls by using the shuffler. // https://bugs.webkit.org/show_bug.cgi?id=148831 MacroAssemblerCodeRef virtualThunkFor(VM& vm, CallLinkInfo& callLinkInfo) { // The callee is in regT0 (for JSVALUE32_64, the tag is in regT1). // The return address is on the stack, or in the link register. We will hence // jump to the callee, or save the return address to the call frame while we // make a C++ function call to the appropriate JIT operation. CCallHelpers jit; CCallHelpers::JumpList slowCase; // This is a slow path execution, and regT2 contains the CallLinkInfo. Count the // slow path execution for the profiler. jit.add32( CCallHelpers::TrustedImm32(1), CCallHelpers::Address(GPRInfo::regT2, CallLinkInfo::offsetOfSlowPathCount())); // FIXME: we should have a story for eliminating these checks. In many cases, // the DFG knows that the value is definitely a cell, or definitely a function. #if USE(JSVALUE64) if (callLinkInfo.isTailCall()) { // Tail calls could have clobbered the GPRInfo::notCellMaskRegister because they // restore callee saved registers before getthing here. So, let's materialize // the NotCellMask in a temp register and use the temp instead. slowCase.append(jit.branchIfNotCell(GPRInfo::regT0, DoNotHaveTagRegisters)); } else slowCase.append(jit.branchIfNotCell(GPRInfo::regT0)); #else slowCase.append(jit.branchIfNotCell(GPRInfo::regT1)); #endif auto notJSFunction = jit.branchIfNotFunction(GPRInfo::regT0); // Now we know we have a JSFunction. jit.loadPtr(CCallHelpers::Address(GPRInfo::regT0, JSFunction::offsetOfExecutableOrRareData()), GPRInfo::regT4); auto hasExecutable = jit.branchTestPtr(CCallHelpers::Zero, GPRInfo::regT4, CCallHelpers::TrustedImm32(JSFunction::rareDataTag)); jit.loadPtr(CCallHelpers::Address(GPRInfo::regT4, FunctionRareData::offsetOfExecutable() - JSFunction::rareDataTag), GPRInfo::regT4); hasExecutable.link(&jit); jit.loadPtr( CCallHelpers::Address( GPRInfo::regT4, ExecutableBase::offsetOfJITCodeWithArityCheckFor( callLinkInfo.specializationKind())), GPRInfo::regT4); slowCase.append(jit.branchTestPtr(CCallHelpers::Zero, GPRInfo::regT4)); // Now we know that we have a CodeBlock, and we're committed to making a fast // call. // Make a tail call. This will return back to JIT code. JSInterfaceJIT::Label callCode(jit.label()); emitPointerValidation(jit, GPRInfo::regT4, JSEntryPtrTag); if (callLinkInfo.isTailCall()) { jit.preserveReturnAddressAfterCall(GPRInfo::regT0); jit.prepareForTailCallSlow(GPRInfo::regT4); } jit.farJump(GPRInfo::regT4, JSEntryPtrTag); // NullSetterFunctionType does not get the fast path support. But it is OK since using NullSetterFunctionType is extremely rare. notJSFunction.link(&jit); slowCase.append(jit.branchIfNotType(GPRInfo::regT0, InternalFunctionType)); void* executableAddress = vm.getCTIInternalFunctionTrampolineFor(callLinkInfo.specializationKind()).executableAddress(); jit.move(CCallHelpers::TrustedImmPtr(executableAddress), GPRInfo::regT4); jit.jump().linkTo(callCode, &jit); slowCase.link(&jit); // Here we don't know anything, so revert to the full slow path. slowPathFor(jit, vm, operationVirtualCall); LinkBuffer patchBuffer(jit, GLOBAL_THUNK_ID); return FINALIZE_CODE( patchBuffer, JITStubRoutinePtrTag, "Virtual %s slow path thunk", callLinkInfo.callMode() == CallMode::Regular ? "call" : callLinkInfo.callMode() == CallMode::Tail ? "tail call" : "construct"); } enum ThunkEntryType { EnterViaCall, EnterViaJumpWithSavedTags, EnterViaJumpWithoutSavedTags }; enum class ThunkFunctionType { JSFunction, InternalFunction }; static MacroAssemblerCodeRef nativeForGenerator(VM& vm, ThunkFunctionType thunkFunctionType, CodeSpecializationKind kind, ThunkEntryType entryType = EnterViaCall) { // FIXME: This should be able to log ShadowChicken prologue packets. // https://bugs.webkit.org/show_bug.cgi?id=155689 int executableOffsetToFunction = NativeExecutable::offsetOfNativeFunctionFor(kind); JSInterfaceJIT jit(&vm); switch (entryType) { case EnterViaCall: jit.emitFunctionPrologue(); break; case EnterViaJumpWithSavedTags: #if USE(JSVALUE64) // We're coming from a specialized thunk that has saved the prior tag registers' contents. // Restore them now. jit.popPair(JSInterfaceJIT::numberTagRegister, JSInterfaceJIT::notCellMaskRegister); #endif break; case EnterViaJumpWithoutSavedTags: jit.move(JSInterfaceJIT::framePointerRegister, JSInterfaceJIT::stackPointerRegister); break; } jit.emitPutToCallFrameHeader(nullptr, CallFrameSlot::codeBlock); jit.storePtr(GPRInfo::callFrameRegister, &vm.topCallFrame); // Host function signature: f(JSGlobalObject*, CallFrame*); #if CPU(X86_64) && OS(WINDOWS) // Leave space for the callee parameter home addresses. // At this point the stack is aligned to 16 bytes, but if this changes at some point, we need to emit code to align it. jit.subPtr(CCallHelpers::TrustedImm32(4 * sizeof(int64_t)), CCallHelpers::stackPointerRegister); #elif CPU(MIPS) // Allocate stack space for (unused) 16 bytes (8-byte aligned) for 4 arguments. jit.subPtr(CCallHelpers::TrustedImm32(16), CCallHelpers::stackPointerRegister); #endif jit.move(GPRInfo::callFrameRegister, GPRInfo::argumentGPR1); jit.emitGetFromCallFrameHeaderPtr(CallFrameSlot::callee, GPRInfo::argumentGPR2); if (thunkFunctionType == ThunkFunctionType::JSFunction) { jit.loadPtr(CCallHelpers::Address(GPRInfo::argumentGPR2, JSFunction::offsetOfScopeChain()), GPRInfo::argumentGPR0); jit.loadPtr(CCallHelpers::Address(GPRInfo::argumentGPR2, JSFunction::offsetOfExecutableOrRareData()), GPRInfo::argumentGPR2); auto hasExecutable = jit.branchTestPtr(CCallHelpers::Zero, GPRInfo::argumentGPR2, CCallHelpers::TrustedImm32(JSFunction::rareDataTag)); jit.loadPtr(CCallHelpers::Address(GPRInfo::argumentGPR2, FunctionRareData::offsetOfExecutable() - JSFunction::rareDataTag), GPRInfo::argumentGPR2); hasExecutable.link(&jit); if (Options::useJITCage()) { jit.loadPtr(CCallHelpers::Address(GPRInfo::argumentGPR2, executableOffsetToFunction), GPRInfo::argumentGPR2); auto operationCall = jit.call(OperationPtrTag); jit.addLinkTask([=] (LinkBuffer& linkBuffer) { linkBuffer.link(operationCall, FunctionPtr(vmEntryHostFunction)); }); } else jit.call(CCallHelpers::Address(GPRInfo::argumentGPR2, executableOffsetToFunction), HostFunctionPtrTag); } else { ASSERT(thunkFunctionType == ThunkFunctionType::InternalFunction); jit.loadPtr(CCallHelpers::Address(GPRInfo::argumentGPR2, InternalFunction::offsetOfGlobalObject()), GPRInfo::argumentGPR0); if (Options::useJITCage()) { jit.loadPtr(CCallHelpers::Address(GPRInfo::argumentGPR2, InternalFunction::offsetOfNativeFunctionFor(kind)), GPRInfo::argumentGPR2); auto operationCall = jit.call(OperationPtrTag); jit.addLinkTask([=] (LinkBuffer& linkBuffer) { linkBuffer.link(operationCall, FunctionPtr(vmEntryHostFunction)); }); } else jit.call(CCallHelpers::Address(GPRInfo::argumentGPR2, InternalFunction::offsetOfNativeFunctionFor(kind)), HostFunctionPtrTag); } #if CPU(X86_64) && OS(WINDOWS) jit.addPtr(CCallHelpers::TrustedImm32(4 * sizeof(int64_t)), CCallHelpers::stackPointerRegister); #elif CPU(MIPS) jit.addPtr(CCallHelpers::TrustedImm32(16), CCallHelpers::stackPointerRegister); #endif // Check for an exception #if USE(JSVALUE64) jit.loadPtr(vm.addressOfException(), JSInterfaceJIT::regT2); JSInterfaceJIT::Jump exceptionHandler = jit.branchTestPtr(JSInterfaceJIT::NonZero, JSInterfaceJIT::regT2); #else JSInterfaceJIT::Jump exceptionHandler = jit.branch32( JSInterfaceJIT::NotEqual, JSInterfaceJIT::AbsoluteAddress(vm.addressOfException()), JSInterfaceJIT::TrustedImm32(0)); #endif jit.emitFunctionEpilogue(); // Return. jit.ret(); // Handle an exception exceptionHandler.link(&jit); jit.copyCalleeSavesToEntryFrameCalleeSavesBuffer(vm.topEntryFrame); jit.storePtr(JSInterfaceJIT::callFrameRegister, &vm.topCallFrame); #if OS(WINDOWS) // Allocate space on stack for the 4 parameter registers. jit.subPtr(JSInterfaceJIT::TrustedImm32(4 * sizeof(int64_t)), JSInterfaceJIT::stackPointerRegister); #elif CPU(MIPS) // Allocate stack space for (unused) 16 bytes (8-byte aligned) for 4 arguments. jit.subPtr(CCallHelpers::TrustedImm32(16), CCallHelpers::stackPointerRegister); #endif jit.move(CCallHelpers::TrustedImmPtr(&vm), JSInterfaceJIT::argumentGPR0); jit.move(JSInterfaceJIT::TrustedImmPtr(tagCFunction(operationVMHandleException)), JSInterfaceJIT::regT3); jit.call(JSInterfaceJIT::regT3, OperationPtrTag); #if OS(WINDOWS) jit.addPtr(JSInterfaceJIT::TrustedImm32(4 * sizeof(int64_t)), JSInterfaceJIT::stackPointerRegister); #elif CPU(MIPS) jit.addPtr(CCallHelpers::TrustedImm32(16), CCallHelpers::stackPointerRegister); #endif jit.jumpToExceptionHandler(vm); LinkBuffer patchBuffer(jit, GLOBAL_THUNK_ID); return FINALIZE_CODE(patchBuffer, JITThunkPtrTag, "%s %s%s trampoline", thunkFunctionType == ThunkFunctionType::JSFunction ? "native" : "internal", entryType == EnterViaJumpWithSavedTags ? "Tail With Saved Tags " : entryType == EnterViaJumpWithoutSavedTags ? "Tail Without Saved Tags " : "", toCString(kind).data()); } MacroAssemblerCodeRef nativeCallGenerator(VM& vm) { return nativeForGenerator(vm, ThunkFunctionType::JSFunction, CodeForCall); } MacroAssemblerCodeRef nativeTailCallGenerator(VM& vm) { return nativeForGenerator(vm, ThunkFunctionType::JSFunction, CodeForCall, EnterViaJumpWithSavedTags); } MacroAssemblerCodeRef nativeTailCallWithoutSavedTagsGenerator(VM& vm) { return nativeForGenerator(vm, ThunkFunctionType::JSFunction, CodeForCall, EnterViaJumpWithoutSavedTags); } MacroAssemblerCodeRef nativeConstructGenerator(VM& vm) { return nativeForGenerator(vm, ThunkFunctionType::JSFunction, CodeForConstruct); } MacroAssemblerCodeRef internalFunctionCallGenerator(VM& vm) { return nativeForGenerator(vm, ThunkFunctionType::InternalFunction, CodeForCall); } MacroAssemblerCodeRef internalFunctionConstructGenerator(VM& vm) { return nativeForGenerator(vm, ThunkFunctionType::InternalFunction, CodeForConstruct); } MacroAssemblerCodeRef arityFixupGenerator(VM& vm) { JSInterfaceJIT jit(&vm); // We enter with fixup count in argumentGPR0 // We have the guarantee that a0, a1, a2, t3, t4 and t5 (or t0 for Windows) are all distinct :-) #if USE(JSVALUE64) #if OS(WINDOWS) const GPRReg extraTemp = JSInterfaceJIT::regT0; #else const GPRReg extraTemp = JSInterfaceJIT::regT5; #endif # if CPU(X86_64) jit.pop(JSInterfaceJIT::regT4); # endif jit.tagReturnAddress(); #if CPU(ARM64E) jit.loadPtr(JSInterfaceJIT::Address(GPRInfo::callFrameRegister, CallFrame::returnPCOffset()), GPRInfo::regT3); jit.addPtr(JSInterfaceJIT::TrustedImm32(sizeof(CallerFrameAndPC)), GPRInfo::callFrameRegister, extraTemp); jit.untagPtr(extraTemp, GPRInfo::regT3); jit.validateUntaggedPtr(GPRInfo::regT3, extraTemp); PtrTag tempReturnPCTag = static_cast(random()); jit.move(JSInterfaceJIT::TrustedImmPtr(tempReturnPCTag), extraTemp); jit.tagPtr(extraTemp, GPRInfo::regT3); jit.storePtr(GPRInfo::regT3, JSInterfaceJIT::Address(GPRInfo::callFrameRegister, CallFrame::returnPCOffset())); #endif jit.move(JSInterfaceJIT::callFrameRegister, JSInterfaceJIT::regT3); jit.load32(JSInterfaceJIT::addressFor(CallFrameSlot::argumentCountIncludingThis), JSInterfaceJIT::argumentGPR2); jit.add32(JSInterfaceJIT::TrustedImm32(CallFrame::headerSizeInRegisters), JSInterfaceJIT::argumentGPR2); // Check to see if we have extra slots we can use jit.move(JSInterfaceJIT::argumentGPR0, JSInterfaceJIT::argumentGPR1); jit.and32(JSInterfaceJIT::TrustedImm32(stackAlignmentRegisters() - 1), JSInterfaceJIT::argumentGPR1); JSInterfaceJIT::Jump noExtraSlot = jit.branchTest32(MacroAssembler::Zero, JSInterfaceJIT::argumentGPR1); jit.move(JSInterfaceJIT::TrustedImm64(JSValue::ValueUndefined), extraTemp); JSInterfaceJIT::Label fillExtraSlots(jit.label()); jit.store64(extraTemp, MacroAssembler::BaseIndex(JSInterfaceJIT::callFrameRegister, JSInterfaceJIT::argumentGPR2, JSInterfaceJIT::TimesEight)); jit.add32(JSInterfaceJIT::TrustedImm32(1), JSInterfaceJIT::argumentGPR2); jit.branchSub32(JSInterfaceJIT::NonZero, JSInterfaceJIT::TrustedImm32(1), JSInterfaceJIT::argumentGPR1).linkTo(fillExtraSlots, &jit); jit.and32(JSInterfaceJIT::TrustedImm32(-stackAlignmentRegisters()), JSInterfaceJIT::argumentGPR0); JSInterfaceJIT::Jump done = jit.branchTest32(MacroAssembler::Zero, JSInterfaceJIT::argumentGPR0); noExtraSlot.link(&jit); jit.neg64(JSInterfaceJIT::argumentGPR0); // Adjust call frame register and stack pointer to account for missing args. // We need to change the stack pointer first before performing copy/fill loops. // This stack space below the stack pointer is considered unused by OS. Therefore, // OS may corrupt this space when constructing a signal stack. jit.move(JSInterfaceJIT::argumentGPR0, extraTemp); jit.lshift64(JSInterfaceJIT::TrustedImm32(3), extraTemp); jit.addPtr(extraTemp, JSInterfaceJIT::callFrameRegister); jit.untagReturnAddress(); jit.addPtr(extraTemp, JSInterfaceJIT::stackPointerRegister); jit.tagReturnAddress(); // Move current frame down argumentGPR0 number of slots JSInterfaceJIT::Label copyLoop(jit.label()); jit.load64(JSInterfaceJIT::regT3, extraTemp); jit.store64(extraTemp, MacroAssembler::BaseIndex(JSInterfaceJIT::regT3, JSInterfaceJIT::argumentGPR0, JSInterfaceJIT::TimesEight)); jit.addPtr(JSInterfaceJIT::TrustedImm32(8), JSInterfaceJIT::regT3); jit.branchSub32(MacroAssembler::NonZero, JSInterfaceJIT::TrustedImm32(1), JSInterfaceJIT::argumentGPR2).linkTo(copyLoop, &jit); // Fill in argumentGPR0 missing arg slots with undefined jit.move(JSInterfaceJIT::argumentGPR0, JSInterfaceJIT::argumentGPR2); jit.move(JSInterfaceJIT::TrustedImm64(JSValue::ValueUndefined), extraTemp); JSInterfaceJIT::Label fillUndefinedLoop(jit.label()); jit.store64(extraTemp, MacroAssembler::BaseIndex(JSInterfaceJIT::regT3, JSInterfaceJIT::argumentGPR0, JSInterfaceJIT::TimesEight)); jit.addPtr(JSInterfaceJIT::TrustedImm32(8), JSInterfaceJIT::regT3); jit.branchAdd32(MacroAssembler::NonZero, JSInterfaceJIT::TrustedImm32(1), JSInterfaceJIT::argumentGPR2).linkTo(fillUndefinedLoop, &jit); done.link(&jit); #if CPU(ARM64E) jit.loadPtr(JSInterfaceJIT::Address(GPRInfo::callFrameRegister, CallFrame::returnPCOffset()), GPRInfo::regT3); jit.move(JSInterfaceJIT::TrustedImmPtr(tempReturnPCTag), extraTemp); jit.untagPtr(extraTemp, GPRInfo::regT3); jit.validateUntaggedPtr(GPRInfo::regT3, extraTemp); jit.addPtr(JSInterfaceJIT::TrustedImm32(sizeof(CallerFrameAndPC)), GPRInfo::callFrameRegister, extraTemp); jit.tagPtr(extraTemp, GPRInfo::regT3); jit.storePtr(GPRInfo::regT3, JSInterfaceJIT::Address(GPRInfo::callFrameRegister, CallFrame::returnPCOffset())); #endif # if CPU(X86_64) jit.push(JSInterfaceJIT::regT4); # endif jit.ret(); #else // USE(JSVALUE64) section above, USE(JSVALUE32_64) section below. jit.move(JSInterfaceJIT::callFrameRegister, JSInterfaceJIT::regT3); jit.load32(JSInterfaceJIT::addressFor(CallFrameSlot::argumentCountIncludingThis), JSInterfaceJIT::argumentGPR2); jit.add32(JSInterfaceJIT::TrustedImm32(CallFrame::headerSizeInRegisters), JSInterfaceJIT::argumentGPR2); // Check to see if we have extra slots we can use jit.move(JSInterfaceJIT::argumentGPR0, JSInterfaceJIT::argumentGPR1); jit.and32(JSInterfaceJIT::TrustedImm32(stackAlignmentRegisters() - 1), JSInterfaceJIT::argumentGPR1); JSInterfaceJIT::Jump noExtraSlot = jit.branchTest32(MacroAssembler::Zero, JSInterfaceJIT::argumentGPR1); JSInterfaceJIT::Label fillExtraSlots(jit.label()); jit.move(JSInterfaceJIT::TrustedImm32(0), JSInterfaceJIT::regT5); jit.store32(JSInterfaceJIT::regT5, MacroAssembler::BaseIndex(JSInterfaceJIT::callFrameRegister, JSInterfaceJIT::argumentGPR2, JSInterfaceJIT::TimesEight, PayloadOffset)); jit.move(JSInterfaceJIT::TrustedImm32(JSValue::UndefinedTag), JSInterfaceJIT::regT5); jit.store32(JSInterfaceJIT::regT5, MacroAssembler::BaseIndex(JSInterfaceJIT::callFrameRegister, JSInterfaceJIT::argumentGPR2, JSInterfaceJIT::TimesEight, TagOffset)); jit.add32(JSInterfaceJIT::TrustedImm32(1), JSInterfaceJIT::argumentGPR2); jit.branchSub32(JSInterfaceJIT::NonZero, JSInterfaceJIT::TrustedImm32(1), JSInterfaceJIT::argumentGPR1).linkTo(fillExtraSlots, &jit); jit.and32(JSInterfaceJIT::TrustedImm32(-stackAlignmentRegisters()), JSInterfaceJIT::argumentGPR0); JSInterfaceJIT::Jump done = jit.branchTest32(MacroAssembler::Zero, JSInterfaceJIT::argumentGPR0); noExtraSlot.link(&jit); jit.neg32(JSInterfaceJIT::argumentGPR0); // Adjust call frame register and stack pointer to account for missing args. // We need to change the stack pointer first before performing copy/fill loops. // This stack space below the stack pointer is considered unused by OS. Therefore, // OS may corrupt this space when constructing a signal stack. jit.move(JSInterfaceJIT::argumentGPR0, JSInterfaceJIT::regT5); jit.lshift32(JSInterfaceJIT::TrustedImm32(3), JSInterfaceJIT::regT5); jit.addPtr(JSInterfaceJIT::regT5, JSInterfaceJIT::callFrameRegister); jit.untagReturnAddress(); jit.addPtr(JSInterfaceJIT::regT5, JSInterfaceJIT::stackPointerRegister); jit.tagReturnAddress(); // Move current frame down argumentGPR0 number of slots JSInterfaceJIT::Label copyLoop(jit.label()); jit.load32(MacroAssembler::Address(JSInterfaceJIT::regT3, PayloadOffset), JSInterfaceJIT::regT5); jit.store32(JSInterfaceJIT::regT5, MacroAssembler::BaseIndex(JSInterfaceJIT::regT3, JSInterfaceJIT::argumentGPR0, JSInterfaceJIT::TimesEight, PayloadOffset)); jit.load32(MacroAssembler::Address(JSInterfaceJIT::regT3, TagOffset), JSInterfaceJIT::regT5); jit.store32(JSInterfaceJIT::regT5, MacroAssembler::BaseIndex(JSInterfaceJIT::regT3, JSInterfaceJIT::argumentGPR0, JSInterfaceJIT::TimesEight, TagOffset)); jit.addPtr(JSInterfaceJIT::TrustedImm32(8), JSInterfaceJIT::regT3); jit.branchSub32(MacroAssembler::NonZero, JSInterfaceJIT::TrustedImm32(1), JSInterfaceJIT::argumentGPR2).linkTo(copyLoop, &jit); // Fill in argumentGPR0 missing arg slots with undefined jit.move(JSInterfaceJIT::argumentGPR0, JSInterfaceJIT::argumentGPR2); JSInterfaceJIT::Label fillUndefinedLoop(jit.label()); jit.move(JSInterfaceJIT::TrustedImm32(0), JSInterfaceJIT::regT5); jit.store32(JSInterfaceJIT::regT5, MacroAssembler::BaseIndex(JSInterfaceJIT::regT3, JSInterfaceJIT::argumentGPR0, JSInterfaceJIT::TimesEight, PayloadOffset)); jit.move(JSInterfaceJIT::TrustedImm32(JSValue::UndefinedTag), JSInterfaceJIT::regT5); jit.store32(JSInterfaceJIT::regT5, MacroAssembler::BaseIndex(JSInterfaceJIT::regT3, JSInterfaceJIT::argumentGPR0, JSInterfaceJIT::TimesEight, TagOffset)); jit.addPtr(JSInterfaceJIT::TrustedImm32(8), JSInterfaceJIT::regT3); jit.branchAdd32(MacroAssembler::NonZero, JSInterfaceJIT::TrustedImm32(1), JSInterfaceJIT::argumentGPR2).linkTo(fillUndefinedLoop, &jit); done.link(&jit); jit.ret(); #endif // End of USE(JSVALUE32_64) section. LinkBuffer patchBuffer(jit, GLOBAL_THUNK_ID); return FINALIZE_CODE(patchBuffer, JITThunkPtrTag, "fixup arity"); } MacroAssemblerCodeRef unreachableGenerator(VM& vm) { JSInterfaceJIT jit(&vm); jit.breakpoint(); LinkBuffer patchBuffer(jit, GLOBAL_THUNK_ID); return FINALIZE_CODE(patchBuffer, JITThunkPtrTag, "unreachable thunk"); } MacroAssemblerCodeRef stringGetByValGenerator(VM& vm) { // regT0 is JSString*, and regT1 (64bit) or regT2 (32bit) is int index. // Return regT0 = result JSString* if succeeds. Otherwise, return regT0 = 0. #if USE(JSVALUE64) GPRReg stringGPR = GPRInfo::regT0; GPRReg indexGPR = GPRInfo::regT1; GPRReg scratchGPR = GPRInfo::regT2; #else GPRReg stringGPR = GPRInfo::regT0; GPRReg indexGPR = GPRInfo::regT2; GPRReg scratchGPR = GPRInfo::regT1; #endif JSInterfaceJIT jit(&vm); JSInterfaceJIT::JumpList failures; jit.tagReturnAddress(); // Load string length to regT2, and start the process of loading the data pointer into regT0 jit.loadPtr(JSInterfaceJIT::Address(stringGPR, JSString::offsetOfValue()), stringGPR); failures.append(jit.branchIfRopeStringImpl(stringGPR)); jit.load32(JSInterfaceJIT::Address(stringGPR, StringImpl::lengthMemoryOffset()), scratchGPR); // Do an unsigned compare to simultaneously filter negative indices as well as indices that are too large failures.append(jit.branch32(JSInterfaceJIT::AboveOrEqual, indexGPR, scratchGPR)); // Load the character JSInterfaceJIT::JumpList cont8Bit; // Load the string flags jit.load32(JSInterfaceJIT::Address(stringGPR, StringImpl::flagsOffset()), scratchGPR); jit.loadPtr(JSInterfaceJIT::Address(stringGPR, StringImpl::dataOffset()), stringGPR); auto is16Bit = jit.branchTest32(JSInterfaceJIT::Zero, scratchGPR, JSInterfaceJIT::TrustedImm32(StringImpl::flagIs8Bit())); jit.load8(JSInterfaceJIT::BaseIndex(stringGPR, indexGPR, JSInterfaceJIT::TimesOne, 0), stringGPR); cont8Bit.append(jit.jump()); is16Bit.link(&jit); jit.load16(JSInterfaceJIT::BaseIndex(stringGPR, indexGPR, JSInterfaceJIT::TimesTwo, 0), stringGPR); cont8Bit.link(&jit); failures.append(jit.branch32(JSInterfaceJIT::Above, stringGPR, JSInterfaceJIT::TrustedImm32(maxSingleCharacterString))); jit.move(JSInterfaceJIT::TrustedImmPtr(vm.smallStrings.singleCharacterStrings()), indexGPR); jit.loadPtr(JSInterfaceJIT::BaseIndex(indexGPR, stringGPR, JSInterfaceJIT::ScalePtr, 0), stringGPR); jit.ret(); failures.link(&jit); jit.move(JSInterfaceJIT::TrustedImm32(0), stringGPR); jit.ret(); LinkBuffer patchBuffer(jit, GLOBAL_THUNK_ID); return FINALIZE_CODE(patchBuffer, JITThunkPtrTag, "String get_by_val stub"); } static void stringCharLoad(SpecializedThunkJIT& jit) { // load string jit.loadJSStringArgument(SpecializedThunkJIT::ThisArgument, SpecializedThunkJIT::regT0); // Load string length to regT2, and start the process of loading the data pointer into regT0 jit.loadPtr(MacroAssembler::Address(SpecializedThunkJIT::regT0, JSString::offsetOfValue()), SpecializedThunkJIT::regT0); jit.appendFailure(jit.branchIfRopeStringImpl(SpecializedThunkJIT::regT0)); jit.load32(MacroAssembler::Address(SpecializedThunkJIT::regT0, StringImpl::lengthMemoryOffset()), SpecializedThunkJIT::regT2); // load index jit.loadInt32Argument(0, SpecializedThunkJIT::regT1); // regT1 contains the index // Do an unsigned compare to simultaneously filter negative indices as well as indices that are too large jit.appendFailure(jit.branch32(MacroAssembler::AboveOrEqual, SpecializedThunkJIT::regT1, SpecializedThunkJIT::regT2)); // Load the character SpecializedThunkJIT::JumpList is16Bit; SpecializedThunkJIT::JumpList cont8Bit; // Load the string flags jit.load32(MacroAssembler::Address(SpecializedThunkJIT::regT0, StringImpl::flagsOffset()), SpecializedThunkJIT::regT2); jit.loadPtr(MacroAssembler::Address(SpecializedThunkJIT::regT0, StringImpl::dataOffset()), SpecializedThunkJIT::regT0); is16Bit.append(jit.branchTest32(MacroAssembler::Zero, SpecializedThunkJIT::regT2, MacroAssembler::TrustedImm32(StringImpl::flagIs8Bit()))); jit.load8(MacroAssembler::BaseIndex(SpecializedThunkJIT::regT0, SpecializedThunkJIT::regT1, MacroAssembler::TimesOne, 0), SpecializedThunkJIT::regT0); cont8Bit.append(jit.jump()); is16Bit.link(&jit); jit.load16(MacroAssembler::BaseIndex(SpecializedThunkJIT::regT0, SpecializedThunkJIT::regT1, MacroAssembler::TimesTwo, 0), SpecializedThunkJIT::regT0); cont8Bit.link(&jit); } static void charToString(SpecializedThunkJIT& jit, VM& vm, MacroAssembler::RegisterID src, MacroAssembler::RegisterID dst, MacroAssembler::RegisterID scratch) { jit.appendFailure(jit.branch32(MacroAssembler::Above, src, MacroAssembler::TrustedImm32(maxSingleCharacterString))); jit.move(MacroAssembler::TrustedImmPtr(vm.smallStrings.singleCharacterStrings()), scratch); jit.loadPtr(MacroAssembler::BaseIndex(scratch, src, MacroAssembler::ScalePtr, 0), dst); jit.appendFailure(jit.branchTestPtr(MacroAssembler::Zero, dst)); } MacroAssemblerCodeRef charCodeAtThunkGenerator(VM& vm) { SpecializedThunkJIT jit(vm, 1); stringCharLoad(jit); jit.returnInt32(SpecializedThunkJIT::regT0); return jit.finalize(vm.jitStubs->ctiNativeTailCall(vm), "charCodeAt"); } MacroAssemblerCodeRef charAtThunkGenerator(VM& vm) { SpecializedThunkJIT jit(vm, 1); stringCharLoad(jit); charToString(jit, vm, SpecializedThunkJIT::regT0, SpecializedThunkJIT::regT0, SpecializedThunkJIT::regT1); jit.returnJSCell(SpecializedThunkJIT::regT0); return jit.finalize(vm.jitStubs->ctiNativeTailCall(vm), "charAt"); } MacroAssemblerCodeRef fromCharCodeThunkGenerator(VM& vm) { SpecializedThunkJIT jit(vm, 1); // load char code jit.loadInt32Argument(0, SpecializedThunkJIT::regT0); charToString(jit, vm, SpecializedThunkJIT::regT0, SpecializedThunkJIT::regT0, SpecializedThunkJIT::regT1); jit.returnJSCell(SpecializedThunkJIT::regT0); return jit.finalize(vm.jitStubs->ctiNativeTailCall(vm), "fromCharCode"); } MacroAssemblerCodeRef stringPrototypeCodePointAtThunkGenerator(VM& vm) { SpecializedThunkJIT jit(vm, 1); // load string jit.loadJSStringArgument(SpecializedThunkJIT::ThisArgument, GPRInfo::regT0); // Load string length to regT3, and start the process of loading the data pointer into regT2 jit.loadPtr(CCallHelpers::Address(GPRInfo::regT0, JSString::offsetOfValue()), GPRInfo::regT0); jit.appendFailure(jit.branchIfRopeStringImpl(GPRInfo::regT0)); jit.load32(CCallHelpers::Address(GPRInfo::regT0, StringImpl::lengthMemoryOffset()), GPRInfo::regT3); // load index jit.loadInt32Argument(0, GPRInfo::regT1); // regT1 contains the index // Do an unsigned compare to simultaneously filter negative indices as well as indices that are too large jit.appendFailure(jit.branch32(CCallHelpers::AboveOrEqual, GPRInfo::regT1, GPRInfo::regT3)); // Load the character CCallHelpers::JumpList done; // Load the string flags jit.loadPtr(CCallHelpers::Address(GPRInfo::regT0, StringImpl::dataOffset()), GPRInfo::regT2); auto is16Bit = jit.branchTest32(CCallHelpers::Zero, CCallHelpers::Address(GPRInfo::regT0, StringImpl::flagsOffset()), CCallHelpers::TrustedImm32(StringImpl::flagIs8Bit())); jit.load8(CCallHelpers::BaseIndex(GPRInfo::regT2, GPRInfo::regT1, CCallHelpers::TimesOne, 0), GPRInfo::regT0); done.append(jit.jump()); is16Bit.link(&jit); jit.load16(CCallHelpers::BaseIndex(GPRInfo::regT2, GPRInfo::regT1, CCallHelpers::TimesTwo, 0), GPRInfo::regT0); // Original index is int32_t, and here, we ensure that it is positive. If we interpret it as uint32_t, adding 1 never overflows. jit.add32(CCallHelpers::TrustedImm32(1), GPRInfo::regT1); done.append(jit.branch32(CCallHelpers::AboveOrEqual, GPRInfo::regT1, GPRInfo::regT3)); jit.and32(CCallHelpers::TrustedImm32(0xfffffc00), GPRInfo::regT0, GPRInfo::regT3); done.append(jit.branch32(CCallHelpers::NotEqual, GPRInfo::regT3, CCallHelpers::TrustedImm32(0xd800))); jit.load16(CCallHelpers::BaseIndex(GPRInfo::regT2, GPRInfo::regT1, CCallHelpers::TimesTwo, 0), GPRInfo::regT2); jit.and32(CCallHelpers::TrustedImm32(0xfffffc00), GPRInfo::regT2, GPRInfo::regT3); done.append(jit.branch32(CCallHelpers::NotEqual, GPRInfo::regT3, CCallHelpers::TrustedImm32(0xdc00))); jit.lshift32(CCallHelpers::TrustedImm32(10), GPRInfo::regT0); jit.getEffectiveAddress(CCallHelpers::BaseIndex(GPRInfo::regT0, GPRInfo::regT2, CCallHelpers::TimesOne, -U16_SURROGATE_OFFSET), GPRInfo::regT0); done.link(&jit); jit.returnInt32(GPRInfo::regT0); return jit.finalize(vm.jitStubs->ctiNativeTailCall(vm), "codePointAt"); } MacroAssemblerCodeRef clz32ThunkGenerator(VM& vm) { SpecializedThunkJIT jit(vm, 1); MacroAssembler::Jump nonIntArgJump; jit.loadInt32Argument(0, SpecializedThunkJIT::regT0, nonIntArgJump); SpecializedThunkJIT::Label convertedArgumentReentry(&jit); jit.countLeadingZeros32(SpecializedThunkJIT::regT0, SpecializedThunkJIT::regT1); jit.returnInt32(SpecializedThunkJIT::regT1); if (jit.supportsFloatingPointTruncate()) { nonIntArgJump.link(&jit); jit.loadDoubleArgument(0, SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::regT0); jit.branchTruncateDoubleToInt32(SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::regT0, SpecializedThunkJIT::BranchIfTruncateSuccessful).linkTo(convertedArgumentReentry, &jit); jit.appendFailure(jit.jump()); } else jit.appendFailure(nonIntArgJump); return jit.finalize(vm.jitStubs->ctiNativeTailCall(vm), "clz32"); } MacroAssemblerCodeRef sqrtThunkGenerator(VM& vm) { SpecializedThunkJIT jit(vm, 1); if (!jit.supportsFloatingPointSqrt()) return MacroAssemblerCodeRef::createSelfManagedCodeRef(vm.jitStubs->ctiNativeCall(vm)); jit.loadDoubleArgument(0, SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::regT0); jit.sqrtDouble(SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::fpRegT0); jit.returnDouble(SpecializedThunkJIT::fpRegT0); return jit.finalize(vm.jitStubs->ctiNativeTailCall(vm), "sqrt"); } #define UnaryDoubleOpWrapper(function) function##Wrapper enum MathThunkCallingConvention { }; typedef MathThunkCallingConvention(*MathThunk)(MathThunkCallingConvention); #if CPU(X86_64) && COMPILER(GCC_COMPATIBLE) && (OS(DARWIN) || OS(LINUX)) #define defineUnaryDoubleOpWrapper(function) \ asm( \ ".text\n" \ ".globl " SYMBOL_STRING(function##Thunk) "\n" \ HIDE_SYMBOL(function##Thunk) "\n" \ SYMBOL_STRING(function##Thunk) ":" "\n" \ "pushq %rax\n" \ "call " GLOBAL_REFERENCE(function) "\n" \ "popq %rcx\n" \ "ret\n" \ );\ extern "C" { \ MathThunkCallingConvention function##Thunk(MathThunkCallingConvention); \ JSC_ANNOTATE_JIT_OPERATION(function##ThunkId, function##Thunk); \ } \ static MathThunk UnaryDoubleOpWrapper(function) = &function##Thunk; #elif CPU(X86) && COMPILER(GCC_COMPATIBLE) && OS(LINUX) && defined(__PIC__) #define defineUnaryDoubleOpWrapper(function) \ asm( \ ".text\n" \ ".globl " SYMBOL_STRING(function##Thunk) "\n" \ HIDE_SYMBOL(function##Thunk) "\n" \ SYMBOL_STRING(function##Thunk) ":" "\n" \ "pushl %ebx\n" \ "subl $20, %esp\n" \ "movsd %xmm0, (%esp) \n" \ "call __x86.get_pc_thunk.bx\n" \ "addl $_GLOBAL_OFFSET_TABLE_, %ebx\n" \ "call " GLOBAL_REFERENCE(function) "\n" \ "fstpl (%esp) \n" \ "movsd (%esp), %xmm0 \n" \ "addl $20, %esp\n" \ "popl %ebx\n" \ "ret\n" \ );\ extern "C" { \ MathThunkCallingConvention function##Thunk(MathThunkCallingConvention); \ JSC_ANNOTATE_JIT_OPERATION(function##ThunkId, function##Thunk); \ } \ static MathThunk UnaryDoubleOpWrapper(function) = &function##Thunk; #elif CPU(X86) && COMPILER(GCC_COMPATIBLE) && (OS(DARWIN) || OS(LINUX)) #define defineUnaryDoubleOpWrapper(function) \ asm( \ ".text\n" \ ".globl " SYMBOL_STRING(function##Thunk) "\n" \ HIDE_SYMBOL(function##Thunk) "\n" \ SYMBOL_STRING(function##Thunk) ":" "\n" \ "subl $20, %esp\n" \ "movsd %xmm0, (%esp) \n" \ "call " GLOBAL_REFERENCE(function) "\n" \ "fstpl (%esp) \n" \ "movsd (%esp), %xmm0 \n" \ "addl $20, %esp\n" \ "ret\n" \ );\ extern "C" { \ MathThunkCallingConvention function##Thunk(MathThunkCallingConvention); \ JSC_ANNOTATE_JIT_OPERATION(function##ThunkId, function##Thunk); \ } \ static MathThunk UnaryDoubleOpWrapper(function) = &function##Thunk; #elif CPU(ARM_THUMB2) && COMPILER(GCC_COMPATIBLE) && OS(DARWIN) #define defineUnaryDoubleOpWrapper(function) \ asm( \ ".text\n" \ ".align 2\n" \ ".globl " SYMBOL_STRING(function##Thunk) "\n" \ HIDE_SYMBOL(function##Thunk) "\n" \ ".thumb\n" \ ".thumb_func " THUMB_FUNC_PARAM(function##Thunk) "\n" \ SYMBOL_STRING(function##Thunk) ":" "\n" \ "push {lr}\n" \ "vmov r0, r1, d0\n" \ "blx " GLOBAL_REFERENCE(function) "\n" \ "vmov d0, r0, r1\n" \ "pop {lr}\n" \ "bx lr\n" \ ); \ extern "C" { \ MathThunkCallingConvention function##Thunk(MathThunkCallingConvention); \ JSC_ANNOTATE_JIT_OPERATION(function##ThunkId, function##Thunk); \ } \ static MathThunk UnaryDoubleOpWrapper(function) = &function##Thunk; #elif CPU(ARM64) #define defineUnaryDoubleOpWrapper(function) \ asm( \ ".text\n" \ ".align 2\n" \ ".globl " SYMBOL_STRING(function##Thunk) "\n" \ HIDE_SYMBOL(function##Thunk) "\n" \ SYMBOL_STRING(function##Thunk) ":" "\n" \ "b " GLOBAL_REFERENCE(function) "\n" \ ".previous" \ ); \ extern "C" { \ MathThunkCallingConvention function##Thunk(MathThunkCallingConvention); \ JSC_ANNOTATE_JIT_OPERATION(function##ThunkId, function##Thunk); \ } \ static MathThunk UnaryDoubleOpWrapper(function) = &function##Thunk; #elif CPU(X86) && COMPILER(MSVC) && OS(WINDOWS) // MSVC does not accept floor, etc, to be called directly from inline assembly, so we need to wrap these functions. static double (_cdecl *floorFunction)(double) = floor; static double (_cdecl *ceilFunction)(double) = ceil; static double (_cdecl *truncFunction)(double) = trunc; static double (_cdecl *expFunction)(double) = exp; static double (_cdecl *logFunction)(double) = log; static double (_cdecl *jsRoundFunction)(double) = jsRound; #define defineUnaryDoubleOpWrapper(function) \ extern "C" __declspec(naked) MathThunkCallingConvention function##Thunk(MathThunkCallingConvention) \ { \ __asm \ { \ __asm sub esp, 20 \ __asm movsd mmword ptr [esp], xmm0 \ __asm call function##Function \ __asm fstp qword ptr [esp] \ __asm movsd xmm0, mmword ptr [esp] \ __asm add esp, 20 \ __asm ret \ } \ } \ JSC_ANNOTATE_JIT_OPERATION(function##ThunkId, function##Thunk); \ static MathThunk UnaryDoubleOpWrapper(function) = &function##Thunk; #else #define defineUnaryDoubleOpWrapper(function) \ static MathThunk UnaryDoubleOpWrapper(function) = 0 #endif defineUnaryDoubleOpWrapper(jsRound); defineUnaryDoubleOpWrapper(exp); defineUnaryDoubleOpWrapper(log); defineUnaryDoubleOpWrapper(floor); defineUnaryDoubleOpWrapper(ceil); defineUnaryDoubleOpWrapper(trunc); MacroAssemblerCodeRef floorThunkGenerator(VM& vm) { SpecializedThunkJIT jit(vm, 1); MacroAssembler::Jump nonIntJump; if (!UnaryDoubleOpWrapper(floor) || !jit.supportsFloatingPoint()) return MacroAssemblerCodeRef::createSelfManagedCodeRef(vm.jitStubs->ctiNativeCall(vm)); jit.loadInt32Argument(0, SpecializedThunkJIT::regT0, nonIntJump); jit.returnInt32(SpecializedThunkJIT::regT0); nonIntJump.link(&jit); jit.loadDoubleArgument(0, SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::regT0); if (jit.supportsFloatingPointRounding()) { SpecializedThunkJIT::JumpList doubleResult; jit.floorDouble(SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::fpRegT0); jit.branchConvertDoubleToInt32(SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::regT0, doubleResult, SpecializedThunkJIT::fpRegT1); jit.returnInt32(SpecializedThunkJIT::regT0); doubleResult.link(&jit); jit.returnDouble(SpecializedThunkJIT::fpRegT0); return jit.finalize(vm.jitStubs->ctiNativeTailCall(vm), "floor"); } SpecializedThunkJIT::Jump intResult; SpecializedThunkJIT::JumpList doubleResult; if (jit.supportsFloatingPointTruncate()) { jit.moveZeroToDouble(SpecializedThunkJIT::fpRegT1); doubleResult.append(jit.branchDouble(MacroAssembler::DoubleEqualAndOrdered, SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::fpRegT1)); SpecializedThunkJIT::JumpList slowPath; // Handle the negative doubles in the slow path for now. slowPath.append(jit.branchDouble(MacroAssembler::DoubleLessThanOrUnordered, SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::fpRegT1)); slowPath.append(jit.branchTruncateDoubleToInt32(SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::regT0)); intResult = jit.jump(); slowPath.link(&jit); } jit.callDoubleToDoublePreservingReturn(UnaryDoubleOpWrapper(floor)); jit.branchConvertDoubleToInt32(SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::regT0, doubleResult, SpecializedThunkJIT::fpRegT1); if (jit.supportsFloatingPointTruncate()) intResult.link(&jit); jit.returnInt32(SpecializedThunkJIT::regT0); doubleResult.link(&jit); jit.returnDouble(SpecializedThunkJIT::fpRegT0); return jit.finalize(vm.jitStubs->ctiNativeTailCall(vm), "floor"); } MacroAssemblerCodeRef ceilThunkGenerator(VM& vm) { SpecializedThunkJIT jit(vm, 1); if (!UnaryDoubleOpWrapper(ceil) || !jit.supportsFloatingPoint()) return MacroAssemblerCodeRef::createSelfManagedCodeRef(vm.jitStubs->ctiNativeCall(vm)); MacroAssembler::Jump nonIntJump; jit.loadInt32Argument(0, SpecializedThunkJIT::regT0, nonIntJump); jit.returnInt32(SpecializedThunkJIT::regT0); nonIntJump.link(&jit); jit.loadDoubleArgument(0, SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::regT0); if (jit.supportsFloatingPointRounding()) jit.ceilDouble(SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::fpRegT0); else jit.callDoubleToDoublePreservingReturn(UnaryDoubleOpWrapper(ceil)); SpecializedThunkJIT::JumpList doubleResult; jit.branchConvertDoubleToInt32(SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::regT0, doubleResult, SpecializedThunkJIT::fpRegT1); jit.returnInt32(SpecializedThunkJIT::regT0); doubleResult.link(&jit); jit.returnDouble(SpecializedThunkJIT::fpRegT0); return jit.finalize(vm.jitStubs->ctiNativeTailCall(vm), "ceil"); } MacroAssemblerCodeRef truncThunkGenerator(VM& vm) { SpecializedThunkJIT jit(vm, 1); if (!UnaryDoubleOpWrapper(trunc) || !jit.supportsFloatingPoint()) return MacroAssemblerCodeRef::createSelfManagedCodeRef(vm.jitStubs->ctiNativeCall(vm)); MacroAssembler::Jump nonIntJump; jit.loadInt32Argument(0, SpecializedThunkJIT::regT0, nonIntJump); jit.returnInt32(SpecializedThunkJIT::regT0); nonIntJump.link(&jit); jit.loadDoubleArgument(0, SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::regT0); if (jit.supportsFloatingPointRounding()) jit.roundTowardZeroDouble(SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::fpRegT0); else jit.callDoubleToDoublePreservingReturn(UnaryDoubleOpWrapper(trunc)); SpecializedThunkJIT::JumpList doubleResult; jit.branchConvertDoubleToInt32(SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::regT0, doubleResult, SpecializedThunkJIT::fpRegT1); jit.returnInt32(SpecializedThunkJIT::regT0); doubleResult.link(&jit); jit.returnDouble(SpecializedThunkJIT::fpRegT0); return jit.finalize(vm.jitStubs->ctiNativeTailCall(vm), "trunc"); } MacroAssemblerCodeRef roundThunkGenerator(VM& vm) { SpecializedThunkJIT jit(vm, 1); if (!UnaryDoubleOpWrapper(jsRound) || !jit.supportsFloatingPoint()) return MacroAssemblerCodeRef::createSelfManagedCodeRef(vm.jitStubs->ctiNativeCall(vm)); MacroAssembler::Jump nonIntJump; jit.loadInt32Argument(0, SpecializedThunkJIT::regT0, nonIntJump); jit.returnInt32(SpecializedThunkJIT::regT0); nonIntJump.link(&jit); jit.loadDoubleArgument(0, SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::regT0); SpecializedThunkJIT::JumpList doubleResult; if (jit.supportsFloatingPointRounding()) { jit.moveZeroToDouble(SpecializedThunkJIT::fpRegT1); doubleResult.append(jit.branchDouble(MacroAssembler::DoubleEqualAndOrdered, SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::fpRegT1)); jit.ceilDouble(SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::fpRegT1); static constexpr double halfConstant = -0.5; jit.loadDouble(MacroAssembler::TrustedImmPtr(&halfConstant), SpecializedThunkJIT::fpRegT2); jit.addDouble(SpecializedThunkJIT::fpRegT1, SpecializedThunkJIT::fpRegT2); MacroAssembler::Jump shouldRoundDown = jit.branchDouble(MacroAssembler::DoubleGreaterThanAndOrdered, SpecializedThunkJIT::fpRegT2, SpecializedThunkJIT::fpRegT0); jit.moveDouble(SpecializedThunkJIT::fpRegT1, SpecializedThunkJIT::fpRegT0); MacroAssembler::Jump continuation = jit.jump(); shouldRoundDown.link(&jit); static constexpr double oneConstant = 1.0; jit.loadDouble(MacroAssembler::TrustedImmPtr(&oneConstant), SpecializedThunkJIT::fpRegT2); jit.subDouble(SpecializedThunkJIT::fpRegT1, SpecializedThunkJIT::fpRegT2, SpecializedThunkJIT::fpRegT0); continuation.link(&jit); } else jit.callDoubleToDoublePreservingReturn(UnaryDoubleOpWrapper(jsRound)); jit.branchConvertDoubleToInt32(SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::regT0, doubleResult, SpecializedThunkJIT::fpRegT1); jit.returnInt32(SpecializedThunkJIT::regT0); doubleResult.link(&jit); jit.returnDouble(SpecializedThunkJIT::fpRegT0); return jit.finalize(vm.jitStubs->ctiNativeTailCall(vm), "round"); } MacroAssemblerCodeRef expThunkGenerator(VM& vm) { if (!UnaryDoubleOpWrapper(exp)) return MacroAssemblerCodeRef::createSelfManagedCodeRef(vm.jitStubs->ctiNativeCall(vm)); SpecializedThunkJIT jit(vm, 1); if (!jit.supportsFloatingPoint()) return MacroAssemblerCodeRef::createSelfManagedCodeRef(vm.jitStubs->ctiNativeCall(vm)); jit.loadDoubleArgument(0, SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::regT0); jit.callDoubleToDoublePreservingReturn(UnaryDoubleOpWrapper(exp)); jit.returnDouble(SpecializedThunkJIT::fpRegT0); return jit.finalize(vm.jitStubs->ctiNativeTailCall(vm), "exp"); } MacroAssemblerCodeRef logThunkGenerator(VM& vm) { if (!UnaryDoubleOpWrapper(log)) return MacroAssemblerCodeRef::createSelfManagedCodeRef(vm.jitStubs->ctiNativeCall(vm)); SpecializedThunkJIT jit(vm, 1); if (!jit.supportsFloatingPoint()) return MacroAssemblerCodeRef::createSelfManagedCodeRef(vm.jitStubs->ctiNativeCall(vm)); jit.loadDoubleArgument(0, SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::regT0); jit.callDoubleToDoublePreservingReturn(UnaryDoubleOpWrapper(log)); jit.returnDouble(SpecializedThunkJIT::fpRegT0); return jit.finalize(vm.jitStubs->ctiNativeTailCall(vm), "log"); } MacroAssemblerCodeRef absThunkGenerator(VM& vm) { SpecializedThunkJIT jit(vm, 1); if (!jit.supportsFloatingPointAbs()) return MacroAssemblerCodeRef::createSelfManagedCodeRef(vm.jitStubs->ctiNativeCall(vm)); #if USE(JSVALUE64) VirtualRegister virtualRegister = CallFrameSlot::firstArgument; jit.load64(AssemblyHelpers::addressFor(virtualRegister), GPRInfo::regT0); auto notInteger = jit.branchIfNotInt32(GPRInfo::regT0); // Abs Int32. jit.rshift32(GPRInfo::regT0, MacroAssembler::TrustedImm32(31), GPRInfo::regT1); jit.add32(GPRInfo::regT1, GPRInfo::regT0); jit.xor32(GPRInfo::regT1, GPRInfo::regT0); // IntMin cannot be inverted. MacroAssembler::Jump integerIsIntMin = jit.branchTest32(MacroAssembler::Signed, GPRInfo::regT0); // Box and finish. jit.or64(GPRInfo::numberTagRegister, GPRInfo::regT0); MacroAssembler::Jump doneWithIntegers = jit.jump(); // Handle Doubles. notInteger.link(&jit); jit.appendFailure(jit.branchIfNotNumber(GPRInfo::regT0)); jit.unboxDoubleWithoutAssertions(GPRInfo::regT0, GPRInfo::regT0, FPRInfo::fpRegT0); MacroAssembler::Label absFPR0Label = jit.label(); jit.absDouble(FPRInfo::fpRegT0, FPRInfo::fpRegT1); jit.boxDouble(FPRInfo::fpRegT1, GPRInfo::regT0); // Tail. doneWithIntegers.link(&jit); jit.returnJSValue(GPRInfo::regT0); // We know the value of regT0 is IntMin. We could load that value from memory but // it is simpler to just convert it. integerIsIntMin.link(&jit); jit.convertInt32ToDouble(GPRInfo::regT0, FPRInfo::fpRegT0); jit.jump().linkTo(absFPR0Label, &jit); #else MacroAssembler::Jump nonIntJump; jit.loadInt32Argument(0, SpecializedThunkJIT::regT0, nonIntJump); jit.rshift32(SpecializedThunkJIT::regT0, MacroAssembler::TrustedImm32(31), SpecializedThunkJIT::regT1); jit.add32(SpecializedThunkJIT::regT1, SpecializedThunkJIT::regT0); jit.xor32(SpecializedThunkJIT::regT1, SpecializedThunkJIT::regT0); jit.appendFailure(jit.branchTest32(MacroAssembler::Signed, SpecializedThunkJIT::regT0)); jit.returnInt32(SpecializedThunkJIT::regT0); nonIntJump.link(&jit); // Shame about the double int conversion here. jit.loadDoubleArgument(0, SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::regT0); jit.absDouble(SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::fpRegT1); jit.returnDouble(SpecializedThunkJIT::fpRegT1); #endif return jit.finalize(vm.jitStubs->ctiNativeTailCall(vm), "abs"); } MacroAssemblerCodeRef imulThunkGenerator(VM& vm) { SpecializedThunkJIT jit(vm, 2); MacroAssembler::Jump nonIntArg0Jump; jit.loadInt32Argument(0, SpecializedThunkJIT::regT0, nonIntArg0Jump); SpecializedThunkJIT::Label doneLoadingArg0(&jit); MacroAssembler::Jump nonIntArg1Jump; jit.loadInt32Argument(1, SpecializedThunkJIT::regT1, nonIntArg1Jump); SpecializedThunkJIT::Label doneLoadingArg1(&jit); jit.mul32(SpecializedThunkJIT::regT1, SpecializedThunkJIT::regT0); jit.returnInt32(SpecializedThunkJIT::regT0); if (jit.supportsFloatingPointTruncate()) { nonIntArg0Jump.link(&jit); jit.loadDoubleArgument(0, SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::regT0); jit.branchTruncateDoubleToInt32(SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::regT0, SpecializedThunkJIT::BranchIfTruncateSuccessful).linkTo(doneLoadingArg0, &jit); jit.appendFailure(jit.jump()); } else jit.appendFailure(nonIntArg0Jump); if (jit.supportsFloatingPointTruncate()) { nonIntArg1Jump.link(&jit); jit.loadDoubleArgument(1, SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::regT1); jit.branchTruncateDoubleToInt32(SpecializedThunkJIT::fpRegT0, SpecializedThunkJIT::regT1, SpecializedThunkJIT::BranchIfTruncateSuccessful).linkTo(doneLoadingArg1, &jit); jit.appendFailure(jit.jump()); } else jit.appendFailure(nonIntArg1Jump); return jit.finalize(vm.jitStubs->ctiNativeTailCall(vm), "imul"); } MacroAssemblerCodeRef randomThunkGenerator(VM& vm) { SpecializedThunkJIT jit(vm, 0); if (!jit.supportsFloatingPoint()) return MacroAssemblerCodeRef::createSelfManagedCodeRef(vm.jitStubs->ctiNativeCall(vm)); #if USE(JSVALUE64) jit.emitRandomThunk(vm, SpecializedThunkJIT::regT0, SpecializedThunkJIT::regT1, SpecializedThunkJIT::regT2, SpecializedThunkJIT::regT3, SpecializedThunkJIT::fpRegT0); jit.returnDouble(SpecializedThunkJIT::fpRegT0); return jit.finalize(vm.jitStubs->ctiNativeTailCall(vm), "random"); #else return MacroAssemblerCodeRef::createSelfManagedCodeRef(vm.jitStubs->ctiNativeCall(vm)); #endif } MacroAssemblerCodeRef boundFunctionCallGenerator(VM& vm) { CCallHelpers jit; jit.emitFunctionPrologue(); // Set up our call frame. jit.storePtr(CCallHelpers::TrustedImmPtr(nullptr), CCallHelpers::addressFor(CallFrameSlot::codeBlock)); jit.store32(CCallHelpers::TrustedImm32(0), CCallHelpers::tagFor(CallFrameSlot::argumentCountIncludingThis)); unsigned extraStackNeeded = 0; if (unsigned stackMisalignment = sizeof(CallerFrameAndPC) % stackAlignmentBytes()) extraStackNeeded = stackAlignmentBytes() - stackMisalignment; // We need to forward all of the arguments that we were passed. We aren't allowed to do a tail // call here as far as I can tell. At least not so long as the generic path doesn't do a tail // call, since that would be way too weird. // The formula for the number of stack bytes needed given some number of parameters (including // this) is: // // stackAlign((numParams + CallFrameHeaderSize) * sizeof(Register) - sizeof(CallerFrameAndPC)) // // Probably we want to write this as: // // stackAlign((numParams + (CallFrameHeaderSize - CallerFrameAndPCSize)) * sizeof(Register)) // // That's really all there is to this. We have all the registers we need to do it. jit.loadCell(CCallHelpers::addressFor(CallFrameSlot::callee), GPRInfo::regT0); jit.loadPtr(CCallHelpers::Address(GPRInfo::regT0, JSBoundFunction::offsetOfBoundArgs()), GPRInfo::regT2); jit.load32(CCallHelpers::payloadFor(CallFrameSlot::argumentCountIncludingThis), GPRInfo::regT1); jit.move(GPRInfo::regT1, GPRInfo::regT3); auto noArgs = jit.branchTestPtr(CCallHelpers::Zero, GPRInfo::regT2); jit.load32(CCallHelpers::Address(GPRInfo::regT2, JSImmutableButterfly::offsetOfPublicLength()), GPRInfo::regT2); jit.add32(GPRInfo::regT2, GPRInfo::regT1); jit.sub32(CCallHelpers::TrustedImm32(1), GPRInfo::regT1); noArgs.link(&jit); jit.add32(CCallHelpers::TrustedImm32(CallFrame::headerSizeInRegisters - CallerFrameAndPC::sizeInRegisters), GPRInfo::regT1, GPRInfo::regT2); jit.lshift32(CCallHelpers::TrustedImm32(3), GPRInfo::regT2); jit.add32(CCallHelpers::TrustedImm32(stackAlignmentBytes() - 1), GPRInfo::regT2); jit.and32(CCallHelpers::TrustedImm32(-stackAlignmentBytes()), GPRInfo::regT2); if (extraStackNeeded) jit.add32(CCallHelpers::TrustedImm32(extraStackNeeded), GPRInfo::regT2); // At this point regT1 has the actual argument count, regT2 has the amount of stack we will need, and regT3 has the passed argument count. // Check to see if we have enough stack space. jit.negPtr(GPRInfo::regT2); jit.addPtr(CCallHelpers::stackPointerRegister, GPRInfo::regT2); CCallHelpers::Jump haveStackSpace = jit.branchPtr(CCallHelpers::BelowOrEqual, CCallHelpers::AbsoluteAddress(vm.addressOfSoftStackLimit()), GPRInfo::regT2); // Throw Stack Overflow exception jit.copyCalleeSavesToEntryFrameCalleeSavesBuffer(vm.topEntryFrame); jit.loadPtr(CCallHelpers::Address(GPRInfo::regT0, JSBoundFunction::offsetOfScopeChain()), GPRInfo::regT3); jit.setupArguments(GPRInfo::regT3); jit.prepareCallOperation(vm); jit.move(CCallHelpers::TrustedImmPtr(tagCFunction(operationThrowStackOverflowErrorFromThunk)), GPRInfo::nonArgGPR0); emitPointerValidation(jit, GPRInfo::nonArgGPR0, OperationPtrTag); jit.call(GPRInfo::nonArgGPR0, OperationPtrTag); jit.jumpToExceptionHandler(vm); haveStackSpace.link(&jit); jit.move(GPRInfo::regT2, CCallHelpers::stackPointerRegister); // Do basic callee frame setup, including 'this'. jit.store32(GPRInfo::regT1, CCallHelpers::calleeFramePayloadSlot(CallFrameSlot::argumentCountIncludingThis)); JSValueRegs valueRegs = JSValueRegs::withTwoAvailableRegs(GPRInfo::regT4, GPRInfo::regT2); jit.loadValue(CCallHelpers::Address(GPRInfo::regT0, JSBoundFunction::offsetOfBoundThis()), valueRegs); jit.storeValue(valueRegs, CCallHelpers::calleeArgumentSlot(0)); // OK, now we can start copying. This is a simple matter of copying parameters from the caller's // frame to the callee's frame. Note that we know that regT3 (the argument count) must be at // least 1. jit.sub32(CCallHelpers::TrustedImm32(1), GPRInfo::regT3); jit.sub32(CCallHelpers::TrustedImm32(1), GPRInfo::regT1); CCallHelpers::Jump done = jit.branchTest32(CCallHelpers::Zero, GPRInfo::regT3); CCallHelpers::Label loop = jit.label(); jit.sub32(CCallHelpers::TrustedImm32(1), GPRInfo::regT3); jit.sub32(CCallHelpers::TrustedImm32(1), GPRInfo::regT1); jit.loadValue(CCallHelpers::addressFor(virtualRegisterForArgumentIncludingThis(1)).indexedBy(GPRInfo::regT3, CCallHelpers::TimesEight), valueRegs); jit.storeValue(valueRegs, CCallHelpers::calleeArgumentSlot(1).indexedBy(GPRInfo::regT1, CCallHelpers::TimesEight)); jit.branchTest32(CCallHelpers::NonZero, GPRInfo::regT3).linkTo(loop, &jit); done.link(&jit); auto noArgs2 = jit.branchTest32(CCallHelpers::Zero, GPRInfo::regT1); jit.loadPtr(CCallHelpers::Address(GPRInfo::regT0, JSBoundFunction::offsetOfBoundArgs()), GPRInfo::regT3); CCallHelpers::Label loopBound = jit.label(); jit.sub32(CCallHelpers::TrustedImm32(1), GPRInfo::regT1); jit.loadValue(CCallHelpers::BaseIndex(GPRInfo::regT3, GPRInfo::regT1, CCallHelpers::TimesEight, JSImmutableButterfly::offsetOfData() + sizeof(WriteBarrier)), valueRegs); jit.storeValue(valueRegs, CCallHelpers::calleeArgumentSlot(1).indexedBy(GPRInfo::regT1, CCallHelpers::TimesEight)); jit.branchTest32(CCallHelpers::NonZero, GPRInfo::regT1).linkTo(loopBound, &jit); noArgs2.link(&jit); jit.loadPtr(CCallHelpers::Address(GPRInfo::regT0, JSBoundFunction::offsetOfTargetFunction()), GPRInfo::regT2); jit.storeCell(GPRInfo::regT2, CCallHelpers::calleeFrameSlot(CallFrameSlot::callee)); jit.loadPtr(CCallHelpers::Address(GPRInfo::regT2, JSFunction::offsetOfExecutableOrRareData()), GPRInfo::regT0); auto hasExecutable = jit.branchTestPtr(CCallHelpers::Zero, GPRInfo::regT0, CCallHelpers::TrustedImm32(JSFunction::rareDataTag)); jit.loadPtr(CCallHelpers::Address(GPRInfo::regT0, FunctionRareData::offsetOfExecutable() - JSFunction::rareDataTag), GPRInfo::regT0); hasExecutable.link(&jit); jit.loadPtr( CCallHelpers::Address( GPRInfo::regT0, ExecutableBase::offsetOfJITCodeWithArityCheckFor(CodeForCall)), GPRInfo::regT0); CCallHelpers::Jump noCode = jit.branchTestPtr(CCallHelpers::Zero, GPRInfo::regT0); emitPointerValidation(jit, GPRInfo::regT0, JSEntryPtrTag); jit.call(GPRInfo::regT0, JSEntryPtrTag); jit.emitFunctionEpilogue(); jit.ret(); LinkBuffer linkBuffer(jit, GLOBAL_THUNK_ID); linkBuffer.link(noCode, CodeLocationLabel(vm.jitStubs->ctiNativeTailCallWithoutSavedTags(vm))); return FINALIZE_CODE( linkBuffer, JITThunkPtrTag, "Specialized thunk for bound function calls with no arguments"); } } // namespace JSC #endif // ENABLE(JIT)