OLD | NEW |
1 // Copyright 2014 the V8 project authors. All rights reserved. | 1 // Copyright 2014 the V8 project authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 #include "src/compiler/code-generator.h" | 5 #include "src/compiler/code-generator.h" |
6 | 6 |
7 #include "src/arm64/frames-arm64.h" | 7 #include "src/arm64/frames-arm64.h" |
8 #include "src/arm64/macro-assembler-arm64.h" | 8 #include "src/arm64/macro-assembler-arm64.h" |
9 #include "src/ast/scopes.h" | 9 #include "src/ast/scopes.h" |
10 #include "src/compiler/code-generator-impl.h" | 10 #include "src/compiler/code-generator-impl.h" |
(...skipping 275 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
286 __ JumpIfSmi(value_, exit()); | 286 __ JumpIfSmi(value_, exit()); |
287 } | 287 } |
288 __ CheckPageFlagClear(value_, scratch0_, | 288 __ CheckPageFlagClear(value_, scratch0_, |
289 MemoryChunk::kPointersToHereAreInterestingMask, | 289 MemoryChunk::kPointersToHereAreInterestingMask, |
290 exit()); | 290 exit()); |
291 RememberedSetAction const remembered_set_action = | 291 RememberedSetAction const remembered_set_action = |
292 mode_ > RecordWriteMode::kValueIsMap ? EMIT_REMEMBERED_SET | 292 mode_ > RecordWriteMode::kValueIsMap ? EMIT_REMEMBERED_SET |
293 : OMIT_REMEMBERED_SET; | 293 : OMIT_REMEMBERED_SET; |
294 SaveFPRegsMode const save_fp_mode = | 294 SaveFPRegsMode const save_fp_mode = |
295 frame()->DidAllocateDoubleRegisters() ? kSaveFPRegs : kDontSaveFPRegs; | 295 frame()->DidAllocateDoubleRegisters() ? kSaveFPRegs : kDontSaveFPRegs; |
296 if (!frame()->needs_frame()) { | 296 if (!frame_access_state()->access_frame_with_fp()) { |
297 // We need to save and restore lr if the frame was elided. | 297 // We need to save and restore lr if the frame was elided. |
298 __ Push(lr); | 298 __ Push(lr); |
299 } | 299 } |
300 RecordWriteStub stub(isolate(), object_, scratch0_, scratch1_, | 300 RecordWriteStub stub(isolate(), object_, scratch0_, scratch1_, |
301 remembered_set_action, save_fp_mode); | 301 remembered_set_action, save_fp_mode); |
302 __ Add(scratch1_, object_, index_); | 302 __ Add(scratch1_, object_, index_); |
303 __ CallStub(&stub); | 303 __ CallStub(&stub); |
304 if (!frame()->needs_frame()) { | 304 if (!frame_access_state()->access_frame_with_fp()) { |
305 __ Pop(lr); | 305 __ Pop(lr); |
306 } | 306 } |
307 } | 307 } |
308 | 308 |
309 private: | 309 private: |
310 Register const object_; | 310 Register const object_; |
311 Operand const index_; | 311 Operand const index_; |
312 Register const value_; | 312 Register const value_; |
313 Register const scratch0_; | 313 Register const scratch0_; |
314 Register const scratch1_; | 314 Register const scratch1_; |
(...skipping 144 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
459 __ asm_instr(i.OutputRegister##width(), i.InputRegister##width(0), \ | 459 __ asm_instr(i.OutputRegister##width(), i.InputRegister##width(0), \ |
460 i.InputRegister##width(1)); \ | 460 i.InputRegister##width(1)); \ |
461 } else { \ | 461 } else { \ |
462 uint32_t imm = \ | 462 uint32_t imm = \ |
463 static_cast<uint32_t>(i.InputOperand##width(1).ImmediateValue()); \ | 463 static_cast<uint32_t>(i.InputOperand##width(1).ImmediateValue()); \ |
464 __ asm_instr(i.OutputRegister##width(), i.InputRegister##width(0), \ | 464 __ asm_instr(i.OutputRegister##width(), i.InputRegister##width(0), \ |
465 imm % (width)); \ | 465 imm % (width)); \ |
466 } \ | 466 } \ |
467 } while (0) | 467 } while (0) |
468 | 468 |
| 469 void CodeGenerator::AssembleDeconstructFrame() { |
| 470 const CallDescriptor* descriptor = linkage()->GetIncomingDescriptor(); |
| 471 if (descriptor->IsCFunctionCall() || descriptor->UseNativeStack()) { |
| 472 __ Mov(csp, fp); |
| 473 } else { |
| 474 __ Mov(jssp, fp); |
| 475 } |
| 476 __ Pop(fp, lr); |
| 477 } |
469 | 478 |
470 void CodeGenerator::AssembleDeconstructActivationRecord(int stack_param_delta) { | 479 void CodeGenerator::AssembleDeconstructActivationRecord(int stack_param_delta) { |
471 int sp_slot_delta = TailCallFrameStackSlotDelta(stack_param_delta); | 480 int sp_slot_delta = TailCallFrameStackSlotDelta(stack_param_delta); |
472 if (sp_slot_delta > 0) { | 481 if (sp_slot_delta > 0) { |
473 __ Drop(sp_slot_delta); | 482 __ Drop(sp_slot_delta); |
474 } | 483 } |
475 frame_access_state()->SetFrameAccessToDefault(); | 484 frame_access_state()->SetFrameAccessToDefault(); |
476 } | 485 } |
477 | 486 |
478 | 487 |
479 void CodeGenerator::AssemblePrepareTailCall(int stack_param_delta) { | 488 void CodeGenerator::AssemblePrepareTailCall(int stack_param_delta) { |
480 int sp_slot_delta = TailCallFrameStackSlotDelta(stack_param_delta); | 489 int sp_slot_delta = TailCallFrameStackSlotDelta(stack_param_delta); |
481 if (sp_slot_delta < 0) { | 490 if (sp_slot_delta < 0) { |
482 __ Claim(-sp_slot_delta); | 491 __ Claim(-sp_slot_delta); |
483 frame_access_state()->IncreaseSPDelta(-sp_slot_delta); | 492 frame_access_state()->IncreaseSPDelta(-sp_slot_delta); |
484 } | 493 } |
485 if (frame()->needs_frame()) { | 494 if (frame_access_state()->access_frame_with_fp()) { |
486 __ Ldr(lr, MemOperand(fp, StandardFrameConstants::kCallerPCOffset)); | 495 __ Ldr(lr, MemOperand(fp, StandardFrameConstants::kCallerPCOffset)); |
487 __ Ldr(fp, MemOperand(fp, StandardFrameConstants::kCallerFPOffset)); | 496 __ Ldr(fp, MemOperand(fp, StandardFrameConstants::kCallerFPOffset)); |
488 } | 497 } |
489 frame_access_state()->SetFrameAccessToSP(); | 498 frame_access_state()->SetFrameAccessToSP(); |
490 } | 499 } |
491 | 500 |
492 void CodeGenerator::AssemblePopArgumentsAdaptorFrame(Register args_reg, | 501 void CodeGenerator::AssemblePopArgumentsAdaptorFrame(Register args_reg, |
493 Register scratch1, | 502 Register scratch1, |
494 Register scratch2, | 503 Register scratch2, |
495 Register scratch3) { | 504 Register scratch3) { |
(...skipping 161 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
657 case kArchRet: | 666 case kArchRet: |
658 AssembleReturn(); | 667 AssembleReturn(); |
659 break; | 668 break; |
660 case kArchStackPointer: | 669 case kArchStackPointer: |
661 __ mov(i.OutputRegister(), masm()->StackPointer()); | 670 __ mov(i.OutputRegister(), masm()->StackPointer()); |
662 break; | 671 break; |
663 case kArchFramePointer: | 672 case kArchFramePointer: |
664 __ mov(i.OutputRegister(), fp); | 673 __ mov(i.OutputRegister(), fp); |
665 break; | 674 break; |
666 case kArchParentFramePointer: | 675 case kArchParentFramePointer: |
667 if (frame_access_state()->frame()->needs_frame()) { | 676 if (frame_access_state()->access_frame_with_fp()) { |
668 __ ldr(i.OutputRegister(), MemOperand(fp, 0)); | 677 __ ldr(i.OutputRegister(), MemOperand(fp, 0)); |
669 } else { | 678 } else { |
670 __ mov(i.OutputRegister(), fp); | 679 __ mov(i.OutputRegister(), fp); |
671 } | 680 } |
672 break; | 681 break; |
673 case kArchTruncateDoubleToI: | 682 case kArchTruncateDoubleToI: |
674 __ TruncateDoubleToI(i.OutputRegister(), i.InputDoubleRegister(0)); | 683 __ TruncateDoubleToI(i.OutputRegister(), i.InputDoubleRegister(0)); |
675 break; | 684 break; |
676 case kArchStoreWithWriteBarrier: { | 685 case kArchStoreWithWriteBarrier: { |
677 RecordWriteMode mode = | 686 RecordWriteMode mode = |
(...skipping 722 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1400 } | 1409 } |
1401 } else { | 1410 } else { |
1402 Condition cc = FlagsConditionToCondition(condition); | 1411 Condition cc = FlagsConditionToCondition(condition); |
1403 __ B(cc, tlabel); | 1412 __ B(cc, tlabel); |
1404 } | 1413 } |
1405 if (!branch->fallthru) __ B(flabel); // no fallthru to flabel. | 1414 if (!branch->fallthru) __ B(flabel); // no fallthru to flabel. |
1406 } | 1415 } |
1407 | 1416 |
1408 | 1417 |
1409 void CodeGenerator::AssembleArchJump(RpoNumber target) { | 1418 void CodeGenerator::AssembleArchJump(RpoNumber target) { |
| 1419 AssembleDeconstructFrameBetweenBlocks(); |
1410 if (!IsNextInAssemblyOrder(target)) __ B(GetLabel(target)); | 1420 if (!IsNextInAssemblyOrder(target)) __ B(GetLabel(target)); |
1411 } | 1421 } |
1412 | 1422 |
1413 | 1423 |
1414 // Assemble boolean materializations after this instruction. | 1424 // Assemble boolean materializations after this instruction. |
1415 void CodeGenerator::AssembleArchBoolean(Instruction* instr, | 1425 void CodeGenerator::AssembleArchBoolean(Instruction* instr, |
1416 FlagsCondition condition) { | 1426 FlagsCondition condition) { |
1417 Arm64OperandConverter i(this, instr); | 1427 Arm64OperandConverter i(this, instr); |
1418 | 1428 |
1419 // Materialize a full 64-bit 1 or 0 value. The result register is always the | 1429 // Materialize a full 64-bit 1 or 0 value. The result register is always the |
(...skipping 37 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1457 } | 1467 } |
1458 | 1468 |
1459 | 1469 |
1460 void CodeGenerator::AssembleDeoptimizerCall( | 1470 void CodeGenerator::AssembleDeoptimizerCall( |
1461 int deoptimization_id, Deoptimizer::BailoutType bailout_type) { | 1471 int deoptimization_id, Deoptimizer::BailoutType bailout_type) { |
1462 Address deopt_entry = Deoptimizer::GetDeoptimizationEntry( | 1472 Address deopt_entry = Deoptimizer::GetDeoptimizationEntry( |
1463 isolate(), deoptimization_id, bailout_type); | 1473 isolate(), deoptimization_id, bailout_type); |
1464 __ Call(deopt_entry, RelocInfo::RUNTIME_ENTRY); | 1474 __ Call(deopt_entry, RelocInfo::RUNTIME_ENTRY); |
1465 } | 1475 } |
1466 | 1476 |
| 1477 void CodeGenerator::SetupStackPointer() { |
| 1478 if (linkage()->GetIncomingDescriptor()->UseNativeStack()) { |
| 1479 __ SetStackPointer(csp); |
| 1480 } else { |
| 1481 __ SetStackPointer(jssp); |
| 1482 } |
| 1483 } |
1467 | 1484 |
1468 void CodeGenerator::AssemblePrologue() { | 1485 void CodeGenerator::AssemblePrologue() { |
1469 CallDescriptor* descriptor = linkage()->GetIncomingDescriptor(); | 1486 CallDescriptor* descriptor = linkage()->GetIncomingDescriptor(); |
1470 frame()->AlignFrame(16); | |
1471 int stack_shrink_slots = frame()->GetSpillSlotCount(); | 1487 int stack_shrink_slots = frame()->GetSpillSlotCount(); |
1472 if (frame()->needs_frame()) { | 1488 if (frame_access_state()->access_frame_with_fp()) { |
1473 if (descriptor->IsJSFunctionCall()) { | 1489 if (descriptor->IsJSFunctionCall()) { |
1474 DCHECK(!descriptor->UseNativeStack()); | 1490 DCHECK(!descriptor->UseNativeStack()); |
1475 __ SetStackPointer(jssp); | 1491 __ SetStackPointer(jssp); |
1476 __ Prologue(this->info()->GeneratePreagedPrologue()); | 1492 __ Prologue(this->info()->GeneratePreagedPrologue()); |
1477 } else { | 1493 } else { |
1478 if (descriptor->UseNativeStack() || descriptor->IsCFunctionCall()) { | 1494 if (descriptor->UseNativeStack() || descriptor->IsCFunctionCall()) { |
1479 __ SetStackPointer(csp); | 1495 __ SetStackPointer(csp); |
1480 } else { | 1496 } else { |
1481 __ SetStackPointer(jssp); | 1497 __ SetStackPointer(jssp); |
1482 } | 1498 } |
1483 if (descriptor->IsCFunctionCall()) { | 1499 if (descriptor->IsCFunctionCall()) { |
1484 __ Push(lr, fp); | 1500 __ Push(lr, fp); |
1485 __ Mov(fp, masm_.StackPointer()); | 1501 __ Mov(fp, masm_.StackPointer()); |
1486 __ Claim(stack_shrink_slots); | 1502 __ Claim(stack_shrink_slots); |
1487 } else { | 1503 } else { |
1488 __ StubPrologue(info()->GetOutputStackFrameType(), | 1504 __ StubPrologue(info()->GetOutputStackFrameType(), |
1489 frame()->GetTotalFrameSlotCount()); | 1505 frame()->GetTotalFrameSlotCount()); |
1490 } | 1506 } |
1491 } | 1507 } |
1492 } else { | |
1493 if (descriptor->UseNativeStack()) { | |
1494 __ SetStackPointer(csp); | |
1495 } else { | |
1496 __ SetStackPointer(jssp); | |
1497 } | |
1498 frame()->SetElidedFrameSizeInSlots(0); | |
1499 } | 1508 } |
1500 frame_access_state()->SetFrameAccessToDefault(); | 1509 |
1501 if (info()->is_osr()) { | 1510 if (info()->is_osr()) { |
1502 // TurboFan OSR-compiled functions cannot be entered directly. | 1511 // TurboFan OSR-compiled functions cannot be entered directly. |
1503 __ Abort(kShouldNotDirectlyEnterOsrFunction); | 1512 __ Abort(kShouldNotDirectlyEnterOsrFunction); |
1504 | 1513 |
1505 // Unoptimized code jumps directly to this entrypoint while the unoptimized | 1514 // Unoptimized code jumps directly to this entrypoint while the unoptimized |
1506 // frame is still on the stack. Optimized code uses OSR values directly from | 1515 // frame is still on the stack. Optimized code uses OSR values directly from |
1507 // the unoptimized frame. Thus, all that needs to be done is to allocate the | 1516 // the unoptimized frame. Thus, all that needs to be done is to allocate the |
1508 // remaining stack slots. | 1517 // remaining stack slots. |
1509 if (FLAG_code_comments) __ RecordComment("-- OSR entrypoint --"); | 1518 if (FLAG_code_comments) __ RecordComment("-- OSR entrypoint --"); |
1510 osr_pc_offset_ = __ pc_offset(); | 1519 osr_pc_offset_ = __ pc_offset(); |
(...skipping 40 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1551 | 1560 |
1552 // Restore fp registers. | 1561 // Restore fp registers. |
1553 CPURegList saves_fp = CPURegList(CPURegister::kFPRegister, kDRegSizeInBits, | 1562 CPURegList saves_fp = CPURegList(CPURegister::kFPRegister, kDRegSizeInBits, |
1554 descriptor->CalleeSavedFPRegisters()); | 1563 descriptor->CalleeSavedFPRegisters()); |
1555 if (saves_fp.Count() != 0) { | 1564 if (saves_fp.Count() != 0) { |
1556 __ PopCPURegList(saves_fp); | 1565 __ PopCPURegList(saves_fp); |
1557 } | 1566 } |
1558 | 1567 |
1559 int pop_count = static_cast<int>(descriptor->StackParameterCount()); | 1568 int pop_count = static_cast<int>(descriptor->StackParameterCount()); |
1560 if (descriptor->IsCFunctionCall()) { | 1569 if (descriptor->IsCFunctionCall()) { |
1561 __ Mov(csp, fp); | 1570 AssembleDeconstructFrameWhenLeaving(); |
1562 __ Pop(fp, lr); | 1571 } else if (frame_access_state()->access_frame_with_fp()) { |
1563 } else if (frame()->needs_frame()) { | |
1564 // Canonicalize JSFunction return sites for now. | 1572 // Canonicalize JSFunction return sites for now. |
1565 if (return_label_.is_bound()) { | 1573 if (return_label_.is_bound()) { |
| 1574 AssembleDeconstructFrameBetweenBlocks(); |
1566 __ B(&return_label_); | 1575 __ B(&return_label_); |
1567 return; | 1576 return; |
1568 } else { | 1577 } else { |
1569 __ Bind(&return_label_); | 1578 __ Bind(&return_label_); |
1570 if (descriptor->UseNativeStack()) { | 1579 AssembleDeconstructFrameWhenLeaving(); |
1571 __ Mov(csp, fp); | |
1572 pop_count += (pop_count & 1); // align | |
1573 } else { | |
1574 __ Mov(jssp, fp); | |
1575 } | |
1576 __ Pop(fp, lr); | |
1577 } | 1580 } |
1578 } else if (descriptor->UseNativeStack()) { | 1581 } else if (descriptor->UseNativeStack()) { |
1579 pop_count += (pop_count & 1); // align | 1582 pop_count += (pop_count & 1); // align |
1580 } | 1583 } |
1581 __ Drop(pop_count); | 1584 __ Drop(pop_count); |
1582 __ Ret(); | 1585 __ Ret(); |
1583 } | 1586 } |
1584 | 1587 |
1585 | 1588 |
1586 void CodeGenerator::AssembleMove(InstructionOperand* source, | 1589 void CodeGenerator::AssembleMove(InstructionOperand* source, |
(...skipping 178 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1765 padding_size -= kInstructionSize; | 1768 padding_size -= kInstructionSize; |
1766 } | 1769 } |
1767 } | 1770 } |
1768 } | 1771 } |
1769 | 1772 |
1770 #undef __ | 1773 #undef __ |
1771 | 1774 |
1772 } // namespace compiler | 1775 } // namespace compiler |
1773 } // namespace internal | 1776 } // namespace internal |
1774 } // namespace v8 | 1777 } // namespace v8 |
OLD | NEW |