| Index: src/ia32/deoptimizer-ia32.cc
|
| ===================================================================
|
| --- src/ia32/deoptimizer-ia32.cc (revision 6800)
|
| +++ src/ia32/deoptimizer-ia32.cc (working copy)
|
| @@ -1,4 +1,4 @@
|
| -// Copyright 2010 the V8 project authors. All rights reserved.
|
| +// Copyright 2011 the V8 project authors. All rights reserved.
|
| // Redistribution and use in source and binary forms, with or without
|
| // modification, are permitted provided that the following conditions are
|
| // met:
|
| @@ -27,6 +27,8 @@
|
|
|
| #include "v8.h"
|
|
|
| +#if defined(V8_TARGET_ARCH_IA32)
|
| +
|
| #include "codegen.h"
|
| #include "deoptimizer.h"
|
| #include "full-codegen.h"
|
| @@ -35,9 +37,24 @@
|
| namespace v8 {
|
| namespace internal {
|
|
|
| -
|
| int Deoptimizer::table_entry_size_ = 10;
|
|
|
| +
|
| +int Deoptimizer::patch_size() {
|
| + return Assembler::kCallInstructionLength;
|
| +}
|
| +
|
| +
|
| +static void ZapCodeRange(Address start, Address end) {
|
| +#ifdef DEBUG
|
| + ASSERT(start <= end);
|
| + int size = end - start;
|
| + CodePatcher destroyer(start, size);
|
| + while (size-- > 0) destroyer.masm()->int3();
|
| +#endif
|
| +}
|
| +
|
| +
|
| void Deoptimizer::DeoptimizeFunction(JSFunction* function) {
|
| AssertNoAllocation no_allocation;
|
|
|
| @@ -45,48 +62,64 @@
|
|
|
| // Get the optimized code.
|
| Code* code = function->code();
|
| + Address code_start_address = code->instruction_start();
|
|
|
| - // Invalidate the relocation information, as it will become invalid by the
|
| - // code patching below, and is not needed any more.
|
| - code->InvalidateRelocation();
|
| + // We will overwrite the code's relocation info in-place. Relocation info
|
| + // is written backward. The relocation info is the payload of a byte
|
| + // array. Later on we will slide this to the start of the byte array and
|
| + // create a filler object in the remaining space.
|
| + ByteArray* reloc_info = code->relocation_info();
|
| + Address reloc_end_address = reloc_info->address() + reloc_info->Size();
|
| + RelocInfoWriter reloc_info_writer(reloc_end_address, code_start_address);
|
|
|
| - // For each return after a safepoint insert a absolute call to the
|
| - // corresponding deoptimization entry.
|
| - unsigned last_pc_offset = 0;
|
| - SafepointTable table(function->code());
|
| - for (unsigned i = 0; i < table.length(); i++) {
|
| - unsigned pc_offset = table.GetPcOffset(i);
|
| - int deoptimization_index = table.GetDeoptimizationIndex(i);
|
| - int gap_code_size = table.GetGapCodeSize(i);
|
| -#ifdef DEBUG
|
| - // Destroy the code which is not supposed to run again.
|
| - unsigned instructions = pc_offset - last_pc_offset;
|
| - CodePatcher destroyer(code->instruction_start() + last_pc_offset,
|
| - instructions);
|
| - for (unsigned i = 0; i < instructions; i++) {
|
| - destroyer.masm()->int3();
|
| - }
|
| -#endif
|
| - last_pc_offset = pc_offset;
|
| + // For each return after a safepoint insert a call to the corresponding
|
| + // deoptimization entry. Since the call is a relative encoding, write new
|
| + // reloc info. We do not need any of the existing reloc info because the
|
| + // existing code will not be used again (we zap it in debug builds).
|
| + SafepointTable table(code);
|
| + Address prev_address = code_start_address;
|
| + for (unsigned i = 0; i < table.length(); ++i) {
|
| + Address curr_address = code_start_address + table.GetPcOffset(i);
|
| + ASSERT_GE(curr_address, prev_address);
|
| + ZapCodeRange(prev_address, curr_address);
|
| +
|
| + SafepointEntry safepoint_entry = table.GetEntry(i);
|
| + int deoptimization_index = safepoint_entry.deoptimization_index();
|
| if (deoptimization_index != Safepoint::kNoDeoptimizationIndex) {
|
| - CodePatcher patcher(
|
| - code->instruction_start() + pc_offset + gap_code_size,
|
| - Assembler::kCallInstructionLength);
|
| - patcher.masm()->call(GetDeoptimizationEntry(deoptimization_index, LAZY),
|
| - RelocInfo::NONE);
|
| - last_pc_offset += gap_code_size + Assembler::kCallInstructionLength;
|
| + // The gap code is needed to get to the state expected at the bailout.
|
| + curr_address += safepoint_entry.gap_code_size();
|
| +
|
| + CodePatcher patcher(curr_address, patch_size());
|
| + Address deopt_entry = GetDeoptimizationEntry(deoptimization_index, LAZY);
|
| + patcher.masm()->call(deopt_entry, RelocInfo::NONE);
|
| +
|
| + // We use RUNTIME_ENTRY for deoptimization bailouts.
|
| + RelocInfo rinfo(curr_address + 1, // 1 after the call opcode.
|
| + RelocInfo::RUNTIME_ENTRY,
|
| + reinterpret_cast<intptr_t>(deopt_entry));
|
| + reloc_info_writer.Write(&rinfo);
|
| + ASSERT_GE(reloc_info_writer.pos(),
|
| + reloc_info->address() + ByteArray::kHeaderSize);
|
| + curr_address += patch_size();
|
| }
|
| + prev_address = curr_address;
|
| }
|
| -#ifdef DEBUG
|
| - // Destroy the code which is not supposed to run again.
|
| - unsigned instructions = code->safepoint_table_start() - last_pc_offset;
|
| - CodePatcher destroyer(code->instruction_start() + last_pc_offset,
|
| - instructions);
|
| - for (unsigned i = 0; i < instructions; i++) {
|
| - destroyer.masm()->int3();
|
| - }
|
| -#endif
|
| + ZapCodeRange(prev_address,
|
| + code_start_address + code->safepoint_table_offset());
|
|
|
| + // Move the relocation info to the beginning of the byte array.
|
| + int new_reloc_size = reloc_end_address - reloc_info_writer.pos();
|
| + memmove(code->relocation_start(), reloc_info_writer.pos(), new_reloc_size);
|
| +
|
| + // The relocation info is in place, update the size.
|
| + reloc_info->set_length(new_reloc_size);
|
| +
|
| + // Handle the junk part after the new relocation info. We will create
|
| + // a non-live object in the extra space at the end of the former reloc info.
|
| + Address junk_address = reloc_info->address() + reloc_info->Size();
|
| + ASSERT(junk_address <= reloc_end_address);
|
| + Heap::CreateFillerObjectAt(junk_address, reloc_end_address - junk_address);
|
| +
|
| // Add the deoptimizing code to the list.
|
| DeoptimizingCodeListNode* node = new DeoptimizingCodeListNode(code);
|
| node->set_next(deoptimizing_code_list_);
|
| @@ -103,40 +136,53 @@
|
| }
|
|
|
|
|
| -void Deoptimizer::PatchStackCheckCode(RelocInfo* rinfo,
|
| - Code* replacement_code) {
|
| - // The stack check code matches the pattern (on ia32, for example):
|
| +void Deoptimizer::PatchStackCheckCodeAt(Address pc_after,
|
| + Code* check_code,
|
| + Code* replacement_code) {
|
| + Address call_target_address = pc_after - kIntSize;
|
| + ASSERT(check_code->entry() ==
|
| + Assembler::target_address_at(call_target_address));
|
| + // The stack check code matches the pattern:
|
| //
|
| // cmp esp, <limit>
|
| // jae ok
|
| // call <stack guard>
|
| + // test eax, <loop nesting depth>
|
| // ok: ...
|
| //
|
| - // We will patch the code to:
|
| + // We will patch away the branch so the code is:
|
| //
|
| // cmp esp, <limit> ;; Not changed
|
| // nop
|
| // nop
|
| // call <on-stack replacment>
|
| + // test eax, <loop nesting depth>
|
| // ok:
|
| - Address call_target_address = rinfo->pc();
|
| ASSERT(*(call_target_address - 3) == 0x73 && // jae
|
| - *(call_target_address - 2) == 0x05 && // offset
|
| + *(call_target_address - 2) == 0x07 && // offset
|
| *(call_target_address - 1) == 0xe8); // call
|
| *(call_target_address - 3) = 0x90; // nop
|
| *(call_target_address - 2) = 0x90; // nop
|
| - rinfo->set_target_address(replacement_code->entry());
|
| + Assembler::set_target_address_at(call_target_address,
|
| + replacement_code->entry());
|
| }
|
|
|
|
|
| -void Deoptimizer::RevertStackCheckCode(RelocInfo* rinfo, Code* check_code) {
|
| - Address call_target_address = rinfo->pc();
|
| +void Deoptimizer::RevertStackCheckCodeAt(Address pc_after,
|
| + Code* check_code,
|
| + Code* replacement_code) {
|
| + Address call_target_address = pc_after - kIntSize;
|
| + ASSERT(replacement_code->entry() ==
|
| + Assembler::target_address_at(call_target_address));
|
| + // Replace the nops from patching (Deoptimizer::PatchStackCheckCode) to
|
| + // restore the conditional branch.
|
| ASSERT(*(call_target_address - 3) == 0x90 && // nop
|
| *(call_target_address - 2) == 0x90 && // nop
|
| *(call_target_address - 1) == 0xe8); // call
|
| *(call_target_address - 3) = 0x73; // jae
|
| - *(call_target_address - 2) = 0x05; // offset
|
| - rinfo->set_target_address(check_code->entry());
|
| + *(call_target_address - 2) = 0x07; // offset
|
| + Assembler::set_target_address_at(call_target_address,
|
| + check_code->entry());
|
| }
|
|
|
|
|
| @@ -500,26 +546,25 @@
|
| __ mov(ebx, Operand(eax, Deoptimizer::input_offset()));
|
|
|
| // Fill in the input registers.
|
| - for (int i = 0; i < kNumberOfRegisters; i++) {
|
| - int offset = (i * kIntSize) + FrameDescription::registers_offset();
|
| - __ mov(ecx, Operand(esp, (kNumberOfRegisters - 1 - i) * kPointerSize));
|
| - __ mov(Operand(ebx, offset), ecx);
|
| + for (int i = kNumberOfRegisters - 1; i >= 0; i--) {
|
| + int offset = (i * kPointerSize) + FrameDescription::registers_offset();
|
| + __ pop(Operand(ebx, offset));
|
| }
|
|
|
| // Fill in the double input registers.
|
| int double_regs_offset = FrameDescription::double_registers_offset();
|
| for (int i = 0; i < XMMRegister::kNumAllocatableRegisters; ++i) {
|
| int dst_offset = i * kDoubleSize + double_regs_offset;
|
| - int src_offset = i * kDoubleSize + kNumberOfRegisters * kPointerSize;
|
| + int src_offset = i * kDoubleSize;
|
| __ movdbl(xmm0, Operand(esp, src_offset));
|
| __ movdbl(Operand(ebx, dst_offset), xmm0);
|
| }
|
|
|
| - // Remove the bailout id and the general purpose registers from the stack.
|
| + // Remove the bailout id and the double registers from the stack.
|
| if (type() == EAGER) {
|
| - __ add(Operand(esp), Immediate(kSavedRegistersAreaSize + kPointerSize));
|
| + __ add(Operand(esp), Immediate(kDoubleRegsSize + kPointerSize));
|
| } else {
|
| - __ add(Operand(esp), Immediate(kSavedRegistersAreaSize + 2 * kPointerSize));
|
| + __ add(Operand(esp), Immediate(kDoubleRegsSize + 2 * kPointerSize));
|
| }
|
|
|
| // Compute a pointer to the unwinding limit in register ecx; that is
|
| @@ -584,7 +629,7 @@
|
|
|
| // Push the registers from the last output frame.
|
| for (int i = 0; i < kNumberOfRegisters; i++) {
|
| - int offset = (i * kIntSize) + FrameDescription::registers_offset();
|
| + int offset = (i * kPointerSize) + FrameDescription::registers_offset();
|
| __ push(Operand(ebx, offset));
|
| }
|
|
|
| @@ -613,3 +658,5 @@
|
|
|
|
|
| } } // namespace v8::internal
|
| +
|
| +#endif // V8_TARGET_ARCH_IA32
|
|
|