| OLD | NEW | 
|---|
| 1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. | 
| 2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without | 
| 3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are | 
| 4 // met: | 4 // met: | 
| 5 // | 5 // | 
| 6 //     * Redistributions of source code must retain the above copyright | 6 //     * Redistributions of source code must retain the above copyright | 
| 7 //       notice, this list of conditions and the following disclaimer. | 7 //       notice, this list of conditions and the following disclaimer. | 
| 8 //     * Redistributions in binary form must reproduce the above | 8 //     * Redistributions in binary form must reproduce the above | 
| 9 //       copyright notice, this list of conditions and the following | 9 //       copyright notice, this list of conditions and the following | 
| 10 //       disclaimer in the documentation and/or other materials provided | 10 //       disclaimer in the documentation and/or other materials provided | 
| (...skipping 96 matching lines...) Expand 10 before | Expand all | Expand 10 after  Loading... | 
| 107 void Builtins::Generate_InOptimizationQueue(MacroAssembler* masm) { | 107 void Builtins::Generate_InOptimizationQueue(MacroAssembler* masm) { | 
| 108   // Checking whether the queued function is ready for install is optional, | 108   // Checking whether the queued function is ready for install is optional, | 
| 109   // since we come across interrupts and stack checks elsewhere.  However, | 109   // since we come across interrupts and stack checks elsewhere.  However, | 
| 110   // not checking may delay installing ready functions, and always checking | 110   // not checking may delay installing ready functions, and always checking | 
| 111   // would be quite expensive.  A good compromise is to first check against | 111   // would be quite expensive.  A good compromise is to first check against | 
| 112   // stack limit as a cue for an interrupt signal. | 112   // stack limit as a cue for an interrupt signal. | 
| 113   Label ok; | 113   Label ok; | 
| 114   __ CompareRoot(rsp, Heap::kStackLimitRootIndex); | 114   __ CompareRoot(rsp, Heap::kStackLimitRootIndex); | 
| 115   __ j(above_equal, &ok); | 115   __ j(above_equal, &ok); | 
| 116 | 116 | 
| 117   CallRuntimePassFunction(masm, Runtime::kTryInstallOptimizedCode); | 117   CallRuntimePassFunction(masm, Runtime::kHiddenTryInstallOptimizedCode); | 
| 118   GenerateTailCallToReturnedCode(masm); | 118   GenerateTailCallToReturnedCode(masm); | 
| 119 | 119 | 
| 120   __ bind(&ok); | 120   __ bind(&ok); | 
| 121   GenerateTailCallToSharedCode(masm); | 121   GenerateTailCallToSharedCode(masm); | 
| 122 } | 122 } | 
| 123 | 123 | 
| 124 | 124 | 
| 125 static void Generate_JSConstructStubHelper(MacroAssembler* masm, | 125 static void Generate_JSConstructStubHelper(MacroAssembler* masm, | 
| 126                                            bool is_api_function, | 126                                            bool is_api_function, | 
| 127                                            bool count_constructions, | 127                                            bool count_constructions, | 
| (...skipping 69 matching lines...) Expand 10 before | Expand all | Expand 10 after  Loading... | 
| 197         __ movp(rcx, FieldOperand(rdi, JSFunction::kSharedFunctionInfoOffset)); | 197         __ movp(rcx, FieldOperand(rdi, JSFunction::kSharedFunctionInfoOffset)); | 
| 198         __ decb(FieldOperand(rcx, | 198         __ decb(FieldOperand(rcx, | 
| 199                              SharedFunctionInfo::kConstructionCountOffset)); | 199                              SharedFunctionInfo::kConstructionCountOffset)); | 
| 200         __ j(not_zero, &allocate); | 200         __ j(not_zero, &allocate); | 
| 201 | 201 | 
| 202         __ Push(rax); | 202         __ Push(rax); | 
| 203         __ Push(rdi); | 203         __ Push(rdi); | 
| 204 | 204 | 
| 205         __ Push(rdi);  // constructor | 205         __ Push(rdi);  // constructor | 
| 206         // The call will replace the stub, so the countdown is only done once. | 206         // The call will replace the stub, so the countdown is only done once. | 
| 207         __ CallRuntime(Runtime::kFinalizeInstanceSize, 1); | 207         __ CallRuntime(Runtime::kHiddenFinalizeInstanceSize, 1); | 
| 208 | 208 | 
| 209         __ Pop(rdi); | 209         __ Pop(rdi); | 
| 210         __ Pop(rax); | 210         __ Pop(rax); | 
| 211 | 211 | 
| 212         __ bind(&allocate); | 212         __ bind(&allocate); | 
| 213       } | 213       } | 
| 214 | 214 | 
| 215       // Now allocate the JSObject on the heap. | 215       // Now allocate the JSObject on the heap. | 
| 216       __ movzxbq(rdi, FieldOperand(rax, Map::kInstanceSizeOffset)); | 216       __ movzxbq(rdi, FieldOperand(rax, Map::kInstanceSizeOffset)); | 
| 217       __ shl(rdi, Immediate(kPointerSizeLog2)); | 217       __ shl(rdi, Immediate(kPointerSizeLog2)); | 
| (...skipping 148 matching lines...) Expand 10 before | Expand all | Expand 10 after  Loading... | 
| 366       // Get the cell or allocation site. | 366       // Get the cell or allocation site. | 
| 367       __ movp(rdi, Operand(rsp, kPointerSize*2)); | 367       __ movp(rdi, Operand(rsp, kPointerSize*2)); | 
| 368       __ Push(rdi); | 368       __ Push(rdi); | 
| 369       offset = kPointerSize; | 369       offset = kPointerSize; | 
| 370     } | 370     } | 
| 371 | 371 | 
| 372     // Must restore rdi (constructor) before calling runtime. | 372     // Must restore rdi (constructor) before calling runtime. | 
| 373     __ movp(rdi, Operand(rsp, offset)); | 373     __ movp(rdi, Operand(rsp, offset)); | 
| 374     __ Push(rdi); | 374     __ Push(rdi); | 
| 375     if (create_memento) { | 375     if (create_memento) { | 
| 376       __ CallRuntime(Runtime::kNewObjectWithAllocationSite, 2); | 376       __ CallRuntime(Runtime::kHiddenNewObjectWithAllocationSite, 2); | 
| 377     } else { | 377     } else { | 
| 378       __ CallRuntime(Runtime::kNewObject, 1); | 378       __ CallRuntime(Runtime::kHiddenNewObject, 1); | 
| 379     } | 379     } | 
| 380     __ movp(rbx, rax);  // store result in rbx | 380     __ movp(rbx, rax);  // store result in rbx | 
| 381 | 381 | 
| 382     // If we ended up using the runtime, and we want a memento, then the | 382     // If we ended up using the runtime, and we want a memento, then the | 
| 383     // runtime call made it for us, and we shouldn't do create count | 383     // runtime call made it for us, and we shouldn't do create count | 
| 384     // increment. | 384     // increment. | 
| 385     Label count_incremented; | 385     Label count_incremented; | 
| 386     if (create_memento) { | 386     if (create_memento) { | 
| 387       __ jmp(&count_incremented); | 387       __ jmp(&count_incremented); | 
| 388     } | 388     } | 
| (...skipping 232 matching lines...) Expand 10 before | Expand all | Expand 10 after  Loading... | 
| 621   Generate_JSEntryTrampolineHelper(masm, false); | 621   Generate_JSEntryTrampolineHelper(masm, false); | 
| 622 } | 622 } | 
| 623 | 623 | 
| 624 | 624 | 
| 625 void Builtins::Generate_JSConstructEntryTrampoline(MacroAssembler* masm) { | 625 void Builtins::Generate_JSConstructEntryTrampoline(MacroAssembler* masm) { | 
| 626   Generate_JSEntryTrampolineHelper(masm, true); | 626   Generate_JSEntryTrampolineHelper(masm, true); | 
| 627 } | 627 } | 
| 628 | 628 | 
| 629 | 629 | 
| 630 void Builtins::Generate_CompileUnoptimized(MacroAssembler* masm) { | 630 void Builtins::Generate_CompileUnoptimized(MacroAssembler* masm) { | 
| 631   CallRuntimePassFunction(masm, Runtime::kCompileUnoptimized); | 631   CallRuntimePassFunction(masm, Runtime::kHiddenCompileUnoptimized); | 
| 632   GenerateTailCallToReturnedCode(masm); | 632   GenerateTailCallToReturnedCode(masm); | 
| 633 } | 633 } | 
| 634 | 634 | 
| 635 | 635 | 
| 636 static void CallCompileOptimized(MacroAssembler* masm, | 636 static void CallCompileOptimized(MacroAssembler* masm, | 
| 637                                             bool concurrent) { | 637                                             bool concurrent) { | 
| 638   FrameScope scope(masm, StackFrame::INTERNAL); | 638   FrameScope scope(masm, StackFrame::INTERNAL); | 
| 639   // Push a copy of the function onto the stack. | 639   // Push a copy of the function onto the stack. | 
| 640   __ Push(rdi); | 640   __ Push(rdi); | 
| 641   // Function is also the parameter to the runtime call. | 641   // Function is also the parameter to the runtime call. | 
| 642   __ Push(rdi); | 642   __ Push(rdi); | 
| 643   // Whether to compile in a background thread. | 643   // Whether to compile in a background thread. | 
| 644   __ Push(masm->isolate()->factory()->ToBoolean(concurrent)); | 644   __ Push(masm->isolate()->factory()->ToBoolean(concurrent)); | 
| 645 | 645 | 
| 646   __ CallRuntime(Runtime::kCompileOptimized, 2); | 646   __ CallRuntime(Runtime::kHiddenCompileOptimized, 2); | 
| 647   // Restore receiver. | 647   // Restore receiver. | 
| 648   __ Pop(rdi); | 648   __ Pop(rdi); | 
| 649 } | 649 } | 
| 650 | 650 | 
| 651 | 651 | 
| 652 void Builtins::Generate_CompileOptimized(MacroAssembler* masm) { | 652 void Builtins::Generate_CompileOptimized(MacroAssembler* masm) { | 
| 653   CallCompileOptimized(masm, false); | 653   CallCompileOptimized(masm, false); | 
| 654   GenerateTailCallToReturnedCode(masm); | 654   GenerateTailCallToReturnedCode(masm); | 
| 655 } | 655 } | 
| 656 | 656 | 
| (...skipping 80 matching lines...) Expand 10 before | Expand all | Expand 10 after  Loading... | 
| 737 static void Generate_NotifyStubFailureHelper(MacroAssembler* masm, | 737 static void Generate_NotifyStubFailureHelper(MacroAssembler* masm, | 
| 738                                              SaveFPRegsMode save_doubles) { | 738                                              SaveFPRegsMode save_doubles) { | 
| 739   // Enter an internal frame. | 739   // Enter an internal frame. | 
| 740   { | 740   { | 
| 741     FrameScope scope(masm, StackFrame::INTERNAL); | 741     FrameScope scope(masm, StackFrame::INTERNAL); | 
| 742 | 742 | 
| 743     // Preserve registers across notification, this is important for compiled | 743     // Preserve registers across notification, this is important for compiled | 
| 744     // stubs that tail call the runtime on deopts passing their parameters in | 744     // stubs that tail call the runtime on deopts passing their parameters in | 
| 745     // registers. | 745     // registers. | 
| 746     __ Pushad(); | 746     __ Pushad(); | 
| 747     __ CallRuntime(Runtime::kNotifyStubFailure, 0, save_doubles); | 747     __ CallRuntime(Runtime::kHiddenNotifyStubFailure, 0, save_doubles); | 
| 748     __ Popad(); | 748     __ Popad(); | 
| 749     // Tear down internal frame. | 749     // Tear down internal frame. | 
| 750   } | 750   } | 
| 751 | 751 | 
| 752   __ Pop(MemOperand(rsp, 0));  // Ignore state offset | 752   __ Pop(MemOperand(rsp, 0));  // Ignore state offset | 
| 753   __ ret(0);  // Return to IC Miss stub, continuation still on stack. | 753   __ ret(0);  // Return to IC Miss stub, continuation still on stack. | 
| 754 } | 754 } | 
| 755 | 755 | 
| 756 | 756 | 
| 757 void Builtins::Generate_NotifyStubFailure(MacroAssembler* masm) { | 757 void Builtins::Generate_NotifyStubFailure(MacroAssembler* masm) { | 
| 758   Generate_NotifyStubFailureHelper(masm, kDontSaveFPRegs); | 758   Generate_NotifyStubFailureHelper(masm, kDontSaveFPRegs); | 
| 759 } | 759 } | 
| 760 | 760 | 
| 761 | 761 | 
| 762 void Builtins::Generate_NotifyStubFailureSaveDoubles(MacroAssembler* masm) { | 762 void Builtins::Generate_NotifyStubFailureSaveDoubles(MacroAssembler* masm) { | 
| 763   Generate_NotifyStubFailureHelper(masm, kSaveFPRegs); | 763   Generate_NotifyStubFailureHelper(masm, kSaveFPRegs); | 
| 764 } | 764 } | 
| 765 | 765 | 
| 766 | 766 | 
| 767 static void Generate_NotifyDeoptimizedHelper(MacroAssembler* masm, | 767 static void Generate_NotifyDeoptimizedHelper(MacroAssembler* masm, | 
| 768                                              Deoptimizer::BailoutType type) { | 768                                              Deoptimizer::BailoutType type) { | 
| 769   // Enter an internal frame. | 769   // Enter an internal frame. | 
| 770   { | 770   { | 
| 771     FrameScope scope(masm, StackFrame::INTERNAL); | 771     FrameScope scope(masm, StackFrame::INTERNAL); | 
| 772 | 772 | 
| 773     // Pass the deoptimization type to the runtime system. | 773     // Pass the deoptimization type to the runtime system. | 
| 774     __ Push(Smi::FromInt(static_cast<int>(type))); | 774     __ Push(Smi::FromInt(static_cast<int>(type))); | 
| 775 | 775 | 
| 776     __ CallRuntime(Runtime::kNotifyDeoptimized, 1); | 776     __ CallRuntime(Runtime::kHiddenNotifyDeoptimized, 1); | 
| 777     // Tear down internal frame. | 777     // Tear down internal frame. | 
| 778   } | 778   } | 
| 779 | 779 | 
| 780   // Get the full codegen state from the stack and untag it. | 780   // Get the full codegen state from the stack and untag it. | 
| 781   __ SmiToInteger32(kScratchRegister, Operand(rsp, kPCOnStackSize)); | 781   __ SmiToInteger32(kScratchRegister, Operand(rsp, kPCOnStackSize)); | 
| 782 | 782 | 
| 783   // Switch on the state. | 783   // Switch on the state. | 
| 784   Label not_no_registers, not_tos_rax; | 784   Label not_no_registers, not_tos_rax; | 
| 785   __ cmpp(kScratchRegister, Immediate(FullCodeGenerator::NO_REGISTERS)); | 785   __ cmpp(kScratchRegister, Immediate(FullCodeGenerator::NO_REGISTERS)); | 
| 786   __ j(not_equal, ¬_no_registers, Label::kNear); | 786   __ j(not_equal, ¬_no_registers, Label::kNear); | 
| (...skipping 692 matching lines...) Expand 10 before | Expand all | Expand 10 after  Loading... | 
| 1479 } | 1479 } | 
| 1480 | 1480 | 
| 1481 | 1481 | 
| 1482 void Builtins::Generate_OsrAfterStackCheck(MacroAssembler* masm) { | 1482 void Builtins::Generate_OsrAfterStackCheck(MacroAssembler* masm) { | 
| 1483   // We check the stack limit as indicator that recompilation might be done. | 1483   // We check the stack limit as indicator that recompilation might be done. | 
| 1484   Label ok; | 1484   Label ok; | 
| 1485   __ CompareRoot(rsp, Heap::kStackLimitRootIndex); | 1485   __ CompareRoot(rsp, Heap::kStackLimitRootIndex); | 
| 1486   __ j(above_equal, &ok); | 1486   __ j(above_equal, &ok); | 
| 1487   { | 1487   { | 
| 1488     FrameScope scope(masm, StackFrame::INTERNAL); | 1488     FrameScope scope(masm, StackFrame::INTERNAL); | 
| 1489     __ CallRuntime(Runtime::kStackGuard, 0); | 1489     __ CallRuntime(Runtime::kHiddenStackGuard, 0); | 
| 1490   } | 1490   } | 
| 1491   __ jmp(masm->isolate()->builtins()->OnStackReplacement(), | 1491   __ jmp(masm->isolate()->builtins()->OnStackReplacement(), | 
| 1492          RelocInfo::CODE_TARGET); | 1492          RelocInfo::CODE_TARGET); | 
| 1493 | 1493 | 
| 1494   __ bind(&ok); | 1494   __ bind(&ok); | 
| 1495   __ ret(0); | 1495   __ ret(0); | 
| 1496 } | 1496 } | 
| 1497 | 1497 | 
| 1498 | 1498 | 
| 1499 #undef __ | 1499 #undef __ | 
| 1500 | 1500 | 
| 1501 } }  // namespace v8::internal | 1501 } }  // namespace v8::internal | 
| 1502 | 1502 | 
| 1503 #endif  // V8_TARGET_ARCH_X64 | 1503 #endif  // V8_TARGET_ARCH_X64 | 
| OLD | NEW | 
|---|