OLD | NEW |
1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 17 matching lines...) Expand all Loading... |
28 #include "v8.h" | 28 #include "v8.h" |
29 | 29 |
30 #include "codegen.h" | 30 #include "codegen.h" |
31 #include "deoptimizer.h" | 31 #include "deoptimizer.h" |
32 #include "full-codegen.h" | 32 #include "full-codegen.h" |
33 #include "safepoint-table.h" | 33 #include "safepoint-table.h" |
34 | 34 |
35 namespace v8 { | 35 namespace v8 { |
36 namespace internal { | 36 namespace internal { |
37 | 37 |
38 const int Deoptimizer::table_entry_size_ = 16; | 38 const int Deoptimizer::table_entry_size_ = 12; |
39 | 39 |
40 | 40 |
41 int Deoptimizer::patch_size() { | 41 int Deoptimizer::patch_size() { |
42 const int kCallInstructionSizeInWords = 3; | 42 const int kCallInstructionSizeInWords = 3; |
43 return kCallInstructionSizeInWords * Assembler::kInstrSize; | 43 return kCallInstructionSizeInWords * Assembler::kInstrSize; |
44 } | 44 } |
45 | 45 |
46 | 46 |
47 void Deoptimizer::DeoptimizeFunctionWithPreparedFunctionList( | 47 void Deoptimizer::DeoptimizeFunctionWithPreparedFunctionList( |
48 JSFunction* function) { | 48 JSFunction* function) { |
(...skipping 409 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
458 // TODO(1588) Note that using pc with stm is deprecated, so we should perhaps | 458 // TODO(1588) Note that using pc with stm is deprecated, so we should perhaps |
459 // handle this a bit differently. | 459 // handle this a bit differently. |
460 __ stm(db_w, sp, restored_regs | sp.bit() | lr.bit() | pc.bit()); | 460 __ stm(db_w, sp, restored_regs | sp.bit() | lr.bit() | pc.bit()); |
461 | 461 |
462 const int kSavedRegistersAreaSize = | 462 const int kSavedRegistersAreaSize = |
463 (kNumberOfRegisters * kPointerSize) + kDoubleRegsSize; | 463 (kNumberOfRegisters * kPointerSize) + kDoubleRegsSize; |
464 | 464 |
465 // Get the bailout id from the stack. | 465 // Get the bailout id from the stack. |
466 __ ldr(r2, MemOperand(sp, kSavedRegistersAreaSize)); | 466 __ ldr(r2, MemOperand(sp, kSavedRegistersAreaSize)); |
467 | 467 |
468 // Get the address of the location in the code object if possible (r3) (return | 468 // Get the address of the location in the code object (r3) (return |
469 // address for lazy deoptimization) and compute the fp-to-sp delta in | 469 // address for lazy deoptimization) and compute the fp-to-sp delta in |
470 // register r4. | 470 // register r4. |
471 if (type() == EAGER || type() == SOFT) { | 471 __ mov(r3, lr); |
472 __ mov(r3, Operand::Zero()); | 472 // Correct one word for bailout id. |
473 // Correct one word for bailout id. | 473 __ add(r4, sp, Operand(kSavedRegistersAreaSize + (1 * kPointerSize))); |
474 __ add(r4, sp, Operand(kSavedRegistersAreaSize + (1 * kPointerSize))); | |
475 } else if (type() == OSR) { | |
476 __ mov(r3, lr); | |
477 // Correct one word for bailout id. | |
478 __ add(r4, sp, Operand(kSavedRegistersAreaSize + (1 * kPointerSize))); | |
479 } else { | |
480 __ mov(r3, lr); | |
481 // Correct two words for bailout id and return address. | |
482 __ add(r4, sp, Operand(kSavedRegistersAreaSize + (2 * kPointerSize))); | |
483 } | |
484 __ sub(r4, fp, r4); | 474 __ sub(r4, fp, r4); |
485 | 475 |
486 // Allocate a new deoptimizer object. | 476 // Allocate a new deoptimizer object. |
487 // Pass four arguments in r0 to r3 and fifth argument on stack. | 477 // Pass four arguments in r0 to r3 and fifth argument on stack. |
488 __ PrepareCallCFunction(6, r5); | 478 __ PrepareCallCFunction(6, r5); |
489 __ ldr(r0, MemOperand(fp, JavaScriptFrameConstants::kFunctionOffset)); | 479 __ ldr(r0, MemOperand(fp, JavaScriptFrameConstants::kFunctionOffset)); |
490 __ mov(r1, Operand(type())); // bailout type, | 480 __ mov(r1, Operand(type())); // bailout type, |
491 // r2: bailout id already loaded. | 481 // r2: bailout id already loaded. |
492 // r3: code address or 0 already loaded. | 482 // r3: code address or 0 already loaded. |
493 __ str(r4, MemOperand(sp, 0 * kPointerSize)); // Fp-to-sp delta. | 483 __ str(r4, MemOperand(sp, 0 * kPointerSize)); // Fp-to-sp delta. |
(...skipping 20 matching lines...) Expand all Loading... |
514 // Copy VFP registers to | 504 // Copy VFP registers to |
515 // double_registers_[DoubleRegister::kMaxNumAllocatableRegisters] | 505 // double_registers_[DoubleRegister::kMaxNumAllocatableRegisters] |
516 int double_regs_offset = FrameDescription::double_registers_offset(); | 506 int double_regs_offset = FrameDescription::double_registers_offset(); |
517 for (int i = 0; i < DwVfpRegister::kMaxNumAllocatableRegisters; ++i) { | 507 for (int i = 0; i < DwVfpRegister::kMaxNumAllocatableRegisters; ++i) { |
518 int dst_offset = i * kDoubleSize + double_regs_offset; | 508 int dst_offset = i * kDoubleSize + double_regs_offset; |
519 int src_offset = i * kDoubleSize + kNumberOfRegisters * kPointerSize; | 509 int src_offset = i * kDoubleSize + kNumberOfRegisters * kPointerSize; |
520 __ vldr(d0, sp, src_offset); | 510 __ vldr(d0, sp, src_offset); |
521 __ vstr(d0, r1, dst_offset); | 511 __ vstr(d0, r1, dst_offset); |
522 } | 512 } |
523 | 513 |
524 // Remove the bailout id, eventually return address, and the saved registers | 514 // Remove the bailout id and the saved registers from the stack. |
525 // from the stack. | 515 __ add(sp, sp, Operand(kSavedRegistersAreaSize + (1 * kPointerSize))); |
526 if (type() == EAGER || type() == SOFT || type() == OSR) { | |
527 __ add(sp, sp, Operand(kSavedRegistersAreaSize + (1 * kPointerSize))); | |
528 } else { | |
529 __ add(sp, sp, Operand(kSavedRegistersAreaSize + (2 * kPointerSize))); | |
530 } | |
531 | 516 |
532 // Compute a pointer to the unwinding limit in register r2; that is | 517 // Compute a pointer to the unwinding limit in register r2; that is |
533 // the first stack slot not part of the input frame. | 518 // the first stack slot not part of the input frame. |
534 __ ldr(r2, MemOperand(r1, FrameDescription::frame_size_offset())); | 519 __ ldr(r2, MemOperand(r1, FrameDescription::frame_size_offset())); |
535 __ add(r2, r2, sp); | 520 __ add(r2, r2, sp); |
536 | 521 |
537 // Unwind the stack down to - but not including - the unwinding | 522 // Unwind the stack down to - but not including - the unwinding |
538 // limit and copy the contents of the activation frame to the input | 523 // limit and copy the contents of the activation frame to the input |
539 // frame description. | 524 // frame description. |
540 __ add(r3, r1, Operand(FrameDescription::frame_content_offset())); | 525 __ add(r3, r1, Operand(FrameDescription::frame_content_offset())); |
(...skipping 88 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
629 | 614 |
630 __ pop(ip); // remove pc | 615 __ pop(ip); // remove pc |
631 __ pop(r7); // get continuation, leave pc on stack | 616 __ pop(r7); // get continuation, leave pc on stack |
632 __ pop(lr); | 617 __ pop(lr); |
633 __ Jump(r7); | 618 __ Jump(r7); |
634 __ stop("Unreachable."); | 619 __ stop("Unreachable."); |
635 } | 620 } |
636 | 621 |
637 | 622 |
638 void Deoptimizer::TableEntryGenerator::GeneratePrologue() { | 623 void Deoptimizer::TableEntryGenerator::GeneratePrologue() { |
639 // Create a sequence of deoptimization entries. Note that any | 624 // Create a sequence of deoptimization entries. |
640 // registers may be still live. | 625 // Note that registers are still live when jumping to an entry. |
641 Label done; | 626 Label done; |
642 for (int i = 0; i < count(); i++) { | 627 for (int i = 0; i < count(); i++) { |
643 int start = masm()->pc_offset(); | 628 int start = masm()->pc_offset(); |
644 USE(start); | 629 USE(start); |
645 if (type() == EAGER || type() == SOFT) { | |
646 __ nop(); | |
647 } else { | |
648 // Emulate ia32 like call by pushing return address to stack. | |
649 __ push(lr); | |
650 } | |
651 __ mov(ip, Operand(i)); | 630 __ mov(ip, Operand(i)); |
652 __ push(ip); | 631 __ push(ip); |
653 __ b(&done); | 632 __ b(&done); |
654 ASSERT(masm()->pc_offset() - start == table_entry_size_); | 633 ASSERT(masm()->pc_offset() - start == table_entry_size_); |
655 } | 634 } |
656 __ bind(&done); | 635 __ bind(&done); |
657 } | 636 } |
658 | 637 |
659 #undef __ | 638 #undef __ |
660 | 639 |
661 } } // namespace v8::internal | 640 } } // namespace v8::internal |
OLD | NEW |