| Index: src/arm64/builtins-arm64.cc
|
| diff --git a/src/arm64/builtins-arm64.cc b/src/arm64/builtins-arm64.cc
|
| index 18de7f8a0955d0e964233e80c3a71a61a031af71..1da38d018bd22991678dcf12d5c57bbf19fcf619 100644
|
| --- a/src/arm64/builtins-arm64.cc
|
| +++ b/src/arm64/builtins-arm64.cc
|
| @@ -905,6 +905,94 @@ void Builtins::Generate_InterpreterExitTrampoline(MacroAssembler* masm) {
|
| }
|
|
|
|
|
| +static void Generate_InterpreterNotifyDeoptimizedHelper(
|
| + MacroAssembler* masm, Deoptimizer::BailoutType type) {
|
| + // Enter an internal frame.
|
| + {
|
| + FrameScope scope(masm, StackFrame::INTERNAL);
|
| + __ Push(kInterpreterAccumulatorRegister); // Save accumulator register.
|
| +
|
| + // Pass the deoptimization type to the runtime system.
|
| + __ Mov(x1, Operand(Smi::FromInt(static_cast<int>(type))));
|
| + __ Push(x1);
|
| + __ CallRuntime(Runtime::kNotifyDeoptimized, 1);
|
| +
|
| + __ Pop(kInterpreterAccumulatorRegister); // Restore accumulator register.
|
| + // Tear down internal frame.
|
| + }
|
| +
|
| + // Drop state (we don't use these for interpreter deopts) and push PC at top
|
| + // of stack (to simulate initial call to bytecode handler in interpreter entry
|
| + // trampoline).
|
| + __ Pop(x1);
|
| + __ Drop(1);
|
| + __ Push(x1);
|
| +
|
| + // Initialize register file register and dispatch table register.
|
| + __ Add(kInterpreterRegisterFileRegister, fp,
|
| + Operand(InterpreterFrameConstants::kRegisterFilePointerFromFp));
|
| + __ LoadRoot(kInterpreterDispatchTableRegister,
|
| + Heap::kInterpreterTableRootIndex);
|
| + __ Add(kInterpreterDispatchTableRegister, kInterpreterDispatchTableRegister,
|
| + Operand(FixedArray::kHeaderSize - kHeapObjectTag));
|
| +
|
| +
|
| + // Get the context from the frame.
|
| + // TODO(rmcilroy): Update interpreter frame to expect current context at the
|
| + // context slot instead of the function context.
|
| + __ Ldr(kContextRegister,
|
| + MemOperand(kInterpreterRegisterFileRegister,
|
| + InterpreterFrameConstants::kContextFromRegisterPointer));
|
| +
|
| + // Get the bytecode array pointer from the frame.
|
| + __ Ldr(x1,
|
| + MemOperand(kInterpreterRegisterFileRegister,
|
| + InterpreterFrameConstants::kFunctionFromRegisterPointer));
|
| + __ Ldr(x1, FieldMemOperand(x1, JSFunction::kSharedFunctionInfoOffset));
|
| + __ Ldr(kInterpreterBytecodeArrayRegister,
|
| + FieldMemOperand(x1, SharedFunctionInfo::kFunctionDataOffset));
|
| +
|
| + if (FLAG_debug_code) {
|
| + // Check function data field is actually a BytecodeArray object.
|
| + __ AssertNotSmi(kInterpreterBytecodeArrayRegister,
|
| + kFunctionDataShouldBeBytecodeArrayOnInterpreterEntry);
|
| + __ CompareObjectType(kInterpreterBytecodeArrayRegister, x1, x1,
|
| + BYTECODE_ARRAY_TYPE);
|
| + __ Assert(eq, kFunctionDataShouldBeBytecodeArrayOnInterpreterEntry);
|
| + }
|
| +
|
| + // Get the target bytecode offset from the frame.
|
| + __ Ldr(kInterpreterBytecodeOffsetRegister,
|
| + MemOperand(
|
| + kInterpreterRegisterFileRegister,
|
| + InterpreterFrameConstants::kBytecodeOffsetFromRegisterPointer));
|
| + __ SmiUntag(kInterpreterBytecodeOffsetRegister);
|
| +
|
| + // Dispatch to the target bytecode.
|
| + __ Ldrb(x1, MemOperand(kInterpreterBytecodeArrayRegister,
|
| + kInterpreterBytecodeOffsetRegister));
|
| + __ Mov(x1, Operand(x1, LSL, kPointerSizeLog2));
|
| + __ Ldr(ip0, MemOperand(kInterpreterDispatchTableRegister, x1));
|
| + __ Add(ip0, ip0, Operand(Code::kHeaderSize - kHeapObjectTag));
|
| + __ Jump(ip0);
|
| +}
|
| +
|
| +
|
| +void Builtins::Generate_InterpreterNotifyDeoptimized(MacroAssembler* masm) {
|
| + Generate_InterpreterNotifyDeoptimizedHelper(masm, Deoptimizer::EAGER);
|
| +}
|
| +
|
| +
|
| +void Builtins::Generate_InterpreterNotifySoftDeoptimized(MacroAssembler* masm) {
|
| + Generate_InterpreterNotifyDeoptimizedHelper(masm, Deoptimizer::SOFT);
|
| +}
|
| +
|
| +
|
| +void Builtins::Generate_InterpreterNotifyLazyDeoptimized(MacroAssembler* masm) {
|
| + Generate_InterpreterNotifyDeoptimizedHelper(masm, Deoptimizer::LAZY);
|
| +}
|
| +
|
| +
|
| void Builtins::Generate_CompileLazy(MacroAssembler* masm) {
|
| CallRuntimePassFunction(masm, Runtime::kCompileLazy);
|
| GenerateTailCallToReturnedCode(masm);
|
|
|