| Index: src/arm/builtins-arm.cc
|
| diff --git a/src/arm/builtins-arm.cc b/src/arm/builtins-arm.cc
|
| index bef4bc3c4657c751a85b899da22307ad4a6bdbb8..5a47ef44c0d93b40935528d5097cdef83511b180 100644
|
| --- a/src/arm/builtins-arm.cc
|
| +++ b/src/arm/builtins-arm.cc
|
| @@ -289,8 +289,8 @@ void Builtins::Generate_StringConstructCode(MacroAssembler* masm) {
|
| }
|
|
|
|
|
| -static void CallRuntimePassFunction(MacroAssembler* masm,
|
| - Runtime::FunctionId function_id) {
|
| +static void CallRuntimePassFunction(
|
| + MacroAssembler* masm, Runtime::FunctionId function_id) {
|
| FrameScope scope(masm, StackFrame::INTERNAL);
|
| // Push a copy of the function onto the stack.
|
| __ push(r1);
|
| @@ -313,7 +313,13 @@ static void GenerateTailCallToSharedCode(MacroAssembler* masm) {
|
| }
|
|
|
|
|
| -void Builtins::Generate_InRecompileQueue(MacroAssembler* masm) {
|
| +static void GenerateTailCallToReturnedCode(MacroAssembler* masm) {
|
| + __ add(r0, r0, Operand(Code::kHeaderSize - kHeapObjectTag));
|
| + __ Jump(r0);
|
| +}
|
| +
|
| +
|
| +void Builtins::Generate_InOptimizationQueue(MacroAssembler* masm) {
|
| // Checking whether the queued function is ready for install is optional,
|
| // since we come across interrupts and stack checks elsewhere. However,
|
| // not checking may delay installing ready functions, and always checking
|
| @@ -324,22 +330,14 @@ void Builtins::Generate_InRecompileQueue(MacroAssembler* masm) {
|
| __ cmp(sp, Operand(ip));
|
| __ b(hs, &ok);
|
|
|
| - CallRuntimePassFunction(masm, Runtime::kTryInstallRecompiledCode);
|
| - // Tail call to returned code.
|
| - __ add(r0, r0, Operand(Code::kHeaderSize - kHeapObjectTag));
|
| - __ Jump(r0);
|
| + CallRuntimePassFunction(masm, Runtime::kTryInstallOptimizedCode);
|
| + GenerateTailCallToReturnedCode(masm);
|
|
|
| __ bind(&ok);
|
| GenerateTailCallToSharedCode(masm);
|
| }
|
|
|
|
|
| -void Builtins::Generate_ConcurrentRecompile(MacroAssembler* masm) {
|
| - CallRuntimePassFunction(masm, Runtime::kConcurrentRecompile);
|
| - GenerateTailCallToSharedCode(masm);
|
| -}
|
| -
|
| -
|
| static void Generate_JSConstructStubHelper(MacroAssembler* masm,
|
| bool is_api_function,
|
| bool count_constructions) {
|
| @@ -774,19 +772,38 @@ void Builtins::Generate_JSConstructEntryTrampoline(MacroAssembler* masm) {
|
| }
|
|
|
|
|
| -void Builtins::Generate_LazyCompile(MacroAssembler* masm) {
|
| - CallRuntimePassFunction(masm, Runtime::kLazyCompile);
|
| - // Do a tail-call of the compiled function.
|
| - __ add(r2, r0, Operand(Code::kHeaderSize - kHeapObjectTag));
|
| - __ Jump(r2);
|
| +void Builtins::Generate_CompileUnoptimized(MacroAssembler* masm) {
|
| + CallRuntimePassFunction(masm, Runtime::kCompileUnoptimized);
|
| + GenerateTailCallToReturnedCode(masm);
|
| }
|
|
|
|
|
| -void Builtins::Generate_LazyRecompile(MacroAssembler* masm) {
|
| - CallRuntimePassFunction(masm, Runtime::kLazyRecompile);
|
| - // Do a tail-call of the compiled function.
|
| - __ add(r2, r0, Operand(Code::kHeaderSize - kHeapObjectTag));
|
| - __ Jump(r2);
|
| +static void CallCompileOptimized(MacroAssembler* masm, bool concurrent) {
|
| + FrameScope scope(masm, StackFrame::INTERNAL);
|
| + // Push a copy of the function onto the stack.
|
| + __ push(r1);
|
| + // Push call kind information and function as parameter to the runtime call.
|
| + __ Push(r5, r1);
|
| + // Whether to compile in a background thread.
|
| + __ Push(masm->isolate()->factory()->ToBoolean(concurrent));
|
| +
|
| + __ CallRuntime(Runtime::kCompileOptimized, 2);
|
| + // Restore call kind information.
|
| + __ pop(r5);
|
| + // Restore receiver.
|
| + __ pop(r1);
|
| +}
|
| +
|
| +
|
| +void Builtins::Generate_CompileOptimized(MacroAssembler* masm) {
|
| + CallCompileOptimized(masm, false);
|
| + GenerateTailCallToReturnedCode(masm);
|
| +}
|
| +
|
| +
|
| +void Builtins::Generate_CompileOptimizedConcurrent(MacroAssembler* masm) {
|
| + CallCompileOptimized(masm, true);
|
| + GenerateTailCallToReturnedCode(masm);
|
| }
|
|
|
|
|
| @@ -803,7 +820,7 @@ static void GenerateMakeCodeYoungAgainCommon(MacroAssembler* masm) {
|
| // r1 - isolate
|
| FrameScope scope(masm, StackFrame::MANUAL);
|
| __ stm(db_w, sp, r0.bit() | r1.bit() | fp.bit() | lr.bit());
|
| - __ PrepareCallCFunction(1, 0, r2);
|
| + __ PrepareCallCFunction(2, 0, r2);
|
| __ mov(r1, Operand(ExternalReference::isolate_address(masm->isolate())));
|
| __ CallCFunction(
|
| ExternalReference::get_make_code_young_function(masm->isolate()), 2);
|
| @@ -836,14 +853,14 @@ void Builtins::Generate_MarkCodeAsExecutedOnce(MacroAssembler* masm) {
|
| // r1 - isolate
|
| FrameScope scope(masm, StackFrame::MANUAL);
|
| __ stm(db_w, sp, r0.bit() | r1.bit() | fp.bit() | lr.bit());
|
| - __ PrepareCallCFunction(1, 0, r2);
|
| + __ PrepareCallCFunction(2, 0, r2);
|
| __ mov(r1, Operand(ExternalReference::isolate_address(masm->isolate())));
|
| __ CallCFunction(ExternalReference::get_mark_code_as_executed_function(
|
| masm->isolate()), 2);
|
| __ ldm(ia_w, sp, r0.bit() | r1.bit() | fp.bit() | lr.bit());
|
|
|
| // Perform prologue operations usually performed by the young code stub.
|
| - __ stm(db_w, sp, r1.bit() | cp.bit() | fp.bit() | lr.bit());
|
| + __ PushFixedFrame(r1);
|
| __ add(fp, sp, Operand(StandardFrameConstants::kFixedFrameSizeFromFp));
|
|
|
| // Jump to point after the code-age stub.
|
| @@ -938,18 +955,9 @@ void Builtins::Generate_OnStackReplacement(MacroAssembler* masm) {
|
| __ ldr(r0, MemOperand(fp, JavaScriptFrameConstants::kFunctionOffset));
|
| {
|
| FrameScope scope(masm, StackFrame::INTERNAL);
|
| - // Lookup and calculate pc offset.
|
| - __ ldr(r1, MemOperand(fp, StandardFrameConstants::kCallerPCOffset));
|
| - __ ldr(r2, FieldMemOperand(r0, JSFunction::kSharedFunctionInfoOffset));
|
| - __ ldr(r2, FieldMemOperand(r2, SharedFunctionInfo::kCodeOffset));
|
| - __ sub(r1, r1, Operand(Code::kHeaderSize - kHeapObjectTag));
|
| - __ sub(r1, r1, r2);
|
| - __ SmiTag(r1);
|
| -
|
| - // Pass both function and pc offset as arguments.
|
| + // Pass function as argument.
|
| __ push(r0);
|
| - __ push(r1);
|
| - __ CallRuntime(Runtime::kCompileForOnStackReplacement, 2);
|
| + __ CallRuntime(Runtime::kCompileForOnStackReplacement, 1);
|
| }
|
|
|
| // If the code object is null, just return to the unoptimized code.
|
| @@ -1353,7 +1361,9 @@ void Builtins::Generate_FunctionApply(MacroAssembler* masm) {
|
| static void EnterArgumentsAdaptorFrame(MacroAssembler* masm) {
|
| __ SmiTag(r0);
|
| __ mov(r4, Operand(Smi::FromInt(StackFrame::ARGUMENTS_ADAPTOR)));
|
| - __ stm(db_w, sp, r0.bit() | r1.bit() | r4.bit() | fp.bit() | lr.bit());
|
| + __ stm(db_w, sp, r0.bit() | r1.bit() | r4.bit() |
|
| + (FLAG_enable_ool_constant_pool ? pp.bit() : 0) |
|
| + fp.bit() | lr.bit());
|
| __ add(fp, sp,
|
| Operand(StandardFrameConstants::kFixedFrameSizeFromFp + kPointerSize));
|
| }
|
|
|