| Index: src/arm/virtual-frame-arm.cc
|
| diff --git a/src/arm/virtual-frame-arm.cc b/src/arm/virtual-frame-arm.cc
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..b794d8b3e151b9f2b40c0470422828bd9779a9f2
|
| --- /dev/null
|
| +++ b/src/arm/virtual-frame-arm.cc
|
| @@ -0,0 +1,438 @@
|
| +// Copyright 2009 the V8 project authors. All rights reserved.
|
| +// Redistribution and use in source and binary forms, with or without
|
| +// modification, are permitted provided that the following conditions are
|
| +// met:
|
| +//
|
| +// * Redistributions of source code must retain the above copyright
|
| +// notice, this list of conditions and the following disclaimer.
|
| +// * Redistributions in binary form must reproduce the above
|
| +// copyright notice, this list of conditions and the following
|
| +// disclaimer in the documentation and/or other materials provided
|
| +// with the distribution.
|
| +// * Neither the name of Google Inc. nor the names of its
|
| +// contributors may be used to endorse or promote products derived
|
| +// from this software without specific prior written permission.
|
| +//
|
| +// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
| +// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
| +// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
| +// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
| +// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
| +// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
| +// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
| +// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
| +// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
| +// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
| +// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
| +
|
| +#include "v8.h"
|
| +
|
| +#include "codegen-inl.h"
|
| +#include "register-allocator-inl.h"
|
| +#include "scopes.h"
|
| +
|
| +namespace v8 { namespace internal {
|
| +
|
| +// -------------------------------------------------------------------------
|
| +// VirtualFrame implementation.
|
| +
|
| +#define __ ACCESS_MASM(masm_)
|
| +
|
| +
|
| +// On entry to a function, the virtual frame already contains the
|
| +// receiver and the parameters. All initial frame elements are in
|
| +// memory.
|
| +VirtualFrame::VirtualFrame(CodeGenerator* cgen)
|
| + : cgen_(cgen),
|
| + masm_(cgen->masm()),
|
| + elements_(cgen->scope()->num_parameters()
|
| + + cgen->scope()->num_stack_slots()
|
| + + kPreallocatedElements),
|
| + parameter_count_(cgen->scope()->num_parameters()),
|
| + local_count_(0),
|
| + stack_pointer_(parameter_count_), // 0-based index of TOS.
|
| + frame_pointer_(kIllegalIndex) {
|
| + for (int i = 0; i < parameter_count_ + 1; i++) {
|
| + elements_.Add(FrameElement::MemoryElement());
|
| + }
|
| + for (int i = 0; i < kNumRegisters; i++) {
|
| + register_locations_[i] = kIllegalIndex;
|
| + }
|
| +}
|
| +
|
| +
|
| +void VirtualFrame::SyncElementBelowStackPointer(int index) {
|
| + UNREACHABLE();
|
| +}
|
| +
|
| +
|
| +void VirtualFrame::SyncElementByPushing(int index) {
|
| + UNREACHABLE();
|
| +}
|
| +
|
| +
|
| +void VirtualFrame::MergeTo(VirtualFrame* expected) {
|
| + Comment cmnt(masm_, "[ Merge frame");
|
| + // We should always be merging the code generator's current frame to an
|
| + // expected frame.
|
| + ASSERT(cgen_->frame() == this);
|
| +
|
| + // Adjust the stack pointer upward (toward the top of the virtual
|
| + // frame) if necessary.
|
| + if (stack_pointer_ < expected->stack_pointer_) {
|
| + int difference = expected->stack_pointer_ - stack_pointer_;
|
| + stack_pointer_ = expected->stack_pointer_;
|
| + __ sub(sp, sp, Operand(difference * kPointerSize));
|
| + }
|
| +
|
| + MergeMoveRegistersToMemory(expected);
|
| + MergeMoveRegistersToRegisters(expected);
|
| + MergeMoveMemoryToRegisters(expected);
|
| +
|
| + // Fix any sync bit problems from the bottom-up, stopping when we
|
| + // hit the stack pointer or the top of the frame if the stack
|
| + // pointer is floating above the frame.
|
| + int limit = Min(stack_pointer_, elements_.length() - 1);
|
| + for (int i = 0; i <= limit; i++) {
|
| + FrameElement source = elements_[i];
|
| + FrameElement target = expected->elements_[i];
|
| + if (source.is_synced() && !target.is_synced()) {
|
| + elements_[i].clear_sync();
|
| + } else if (!source.is_synced() && target.is_synced()) {
|
| + SyncElementAt(i);
|
| + }
|
| + }
|
| +
|
| + // Adjust the stack point downard if necessary.
|
| + if (stack_pointer_ > expected->stack_pointer_) {
|
| + int difference = stack_pointer_ - expected->stack_pointer_;
|
| + stack_pointer_ = expected->stack_pointer_;
|
| + __ add(sp, sp, Operand(difference * kPointerSize));
|
| + }
|
| +
|
| + // At this point, the frames should be identical.
|
| + ASSERT(Equals(expected));
|
| +}
|
| +
|
| +
|
| +void VirtualFrame::MergeMoveRegistersToMemory(VirtualFrame* expected) {
|
| + ASSERT(stack_pointer_ >= expected->stack_pointer_);
|
| +
|
| + // Move registers, constants, and copies to memory. Perform moves
|
| + // from the top downward in the frame in order to leave the backing
|
| + // stores of copies in registers.
|
| + // On ARM, all elements are in memory.
|
| +
|
| +#ifdef DEBUG
|
| + int start = Min(stack_pointer_, elements_.length() - 1);
|
| + for (int i = start; i >= 0; i--) {
|
| + ASSERT(elements_[i].is_memory());
|
| + ASSERT(expected->elements_[i].is_memory());
|
| + }
|
| +#endif
|
| +}
|
| +
|
| +
|
| +void VirtualFrame::MergeMoveRegistersToRegisters(VirtualFrame* expected) {
|
| +}
|
| +
|
| +
|
| +void VirtualFrame::MergeMoveMemoryToRegisters(VirtualFrame *expected) {
|
| +}
|
| +
|
| +
|
| +void VirtualFrame::Enter() {
|
| + Comment cmnt(masm_, "[ Enter JS frame");
|
| +
|
| +#ifdef DEBUG
|
| + // Verify that r1 contains a JS function. The following code relies
|
| + // on r2 being available for use.
|
| + { Label map_check, done;
|
| + __ tst(r1, Operand(kSmiTagMask));
|
| + __ b(ne, &map_check);
|
| + __ stop("VirtualFrame::Enter - r1 is not a function (smi check).");
|
| + __ bind(&map_check);
|
| + __ ldr(r2, FieldMemOperand(r1, HeapObject::kMapOffset));
|
| + __ ldrb(r2, FieldMemOperand(r2, Map::kInstanceTypeOffset));
|
| + __ cmp(r2, Operand(JS_FUNCTION_TYPE));
|
| + __ b(eq, &done);
|
| + __ stop("VirtualFrame::Enter - r1 is not a function (map check).");
|
| + __ bind(&done);
|
| + }
|
| +#endif // DEBUG
|
| +
|
| + // We are about to push four values to the frame.
|
| + Adjust(4);
|
| + __ stm(db_w, sp, r1.bit() | cp.bit() | fp.bit() | lr.bit());
|
| + // Adjust FP to point to saved FP.
|
| + frame_pointer_ = elements_.length() - 2;
|
| + __ add(fp, sp, Operand(2 * kPointerSize));
|
| + cgen_->allocator()->Unuse(r1);
|
| + cgen_->allocator()->Unuse(lr);
|
| +}
|
| +
|
| +
|
| +void VirtualFrame::Exit() {
|
| + Comment cmnt(masm_, "[ Exit JS frame");
|
| + // Drop the execution stack down to the frame pointer and restore the caller
|
| + // frame pointer and return address.
|
| + __ mov(sp, fp);
|
| + __ ldm(ia_w, sp, fp.bit() | lr.bit());
|
| +}
|
| +
|
| +
|
| +void VirtualFrame::AllocateStackSlots(int count) {
|
| + ASSERT(height() == 0);
|
| + local_count_ = count;
|
| + Adjust(count);
|
| + if (count > 0) {
|
| + Comment cmnt(masm_, "[ Allocate space for locals");
|
| + // Initialize stack slots with 'undefined' value.
|
| + __ mov(ip, Operand(Factory::undefined_value()));
|
| + for (int i = 0; i < count; i++) {
|
| + __ push(ip);
|
| + }
|
| + }
|
| +}
|
| +
|
| +
|
| +void VirtualFrame::SaveContextRegister() {
|
| + UNIMPLEMENTED();
|
| +}
|
| +
|
| +
|
| +void VirtualFrame::RestoreContextRegister() {
|
| + UNIMPLEMENTED();
|
| +}
|
| +
|
| +
|
| +void VirtualFrame::PushReceiverSlotAddress() {
|
| + UNIMPLEMENTED();
|
| +}
|
| +
|
| +
|
| +int VirtualFrame::InvalidateFrameSlotAt(int index) {
|
| + UNIMPLEMENTED();
|
| + return kIllegalIndex;
|
| +}
|
| +
|
| +
|
| +void VirtualFrame::TakeFrameSlotAt(int index) {
|
| + UNIMPLEMENTED();
|
| +}
|
| +
|
| +
|
| +void VirtualFrame::StoreToFrameSlotAt(int index) {
|
| + UNIMPLEMENTED();
|
| +}
|
| +
|
| +
|
| +void VirtualFrame::PushTryHandler(HandlerType type) {
|
| + // Grow the expression stack by handler size less one (the return address
|
| + // is already pushed by a call instruction).
|
| + Adjust(kHandlerSize - 1);
|
| + __ PushTryHandler(IN_JAVASCRIPT, type);
|
| +}
|
| +
|
| +
|
| +Result VirtualFrame::RawCallStub(CodeStub* stub) {
|
| + ASSERT(cgen_->HasValidEntryRegisters());
|
| + __ CallStub(stub);
|
| + Result result = cgen_->allocator()->Allocate(r0);
|
| + ASSERT(result.is_valid());
|
| + return result;
|
| +}
|
| +
|
| +
|
| +Result VirtualFrame::CallStub(CodeStub* stub, Result* arg) {
|
| + PrepareForCall(0, 0);
|
| + arg->Unuse();
|
| + return RawCallStub(stub);
|
| +}
|
| +
|
| +
|
| +Result VirtualFrame::CallStub(CodeStub* stub, Result* arg0, Result* arg1) {
|
| + PrepareForCall(0, 0);
|
| + arg0->Unuse();
|
| + arg1->Unuse();
|
| + return RawCallStub(stub);
|
| +}
|
| +
|
| +
|
| +Result VirtualFrame::CallRuntime(Runtime::Function* f, int arg_count) {
|
| + PrepareForCall(arg_count, arg_count);
|
| + ASSERT(cgen_->HasValidEntryRegisters());
|
| + __ CallRuntime(f, arg_count);
|
| + Result result = cgen_->allocator()->Allocate(r0);
|
| + ASSERT(result.is_valid());
|
| + return result;
|
| +}
|
| +
|
| +
|
| +Result VirtualFrame::CallRuntime(Runtime::FunctionId id, int arg_count) {
|
| + PrepareForCall(arg_count, arg_count);
|
| + ASSERT(cgen_->HasValidEntryRegisters());
|
| + __ CallRuntime(id, arg_count);
|
| + Result result = cgen_->allocator()->Allocate(r0);
|
| + ASSERT(result.is_valid());
|
| + return result;
|
| +}
|
| +
|
| +
|
| +Result VirtualFrame::InvokeBuiltin(Builtins::JavaScript id,
|
| + InvokeJSFlags flags,
|
| + Result* arg_count_register,
|
| + int arg_count) {
|
| + ASSERT(arg_count_register->reg().is(r0));
|
| + PrepareForCall(arg_count, arg_count);
|
| + arg_count_register->Unuse();
|
| + __ InvokeBuiltin(id, flags);
|
| + Result result = cgen_->allocator()->Allocate(r0);
|
| + return result;
|
| +}
|
| +
|
| +
|
| +Result VirtualFrame::RawCallCodeObject(Handle<Code> code,
|
| + RelocInfo::Mode rmode) {
|
| + ASSERT(cgen_->HasValidEntryRegisters());
|
| + __ Call(code, rmode);
|
| + Result result = cgen_->allocator()->Allocate(r0);
|
| + ASSERT(result.is_valid());
|
| + return result;
|
| +}
|
| +
|
| +
|
| +Result VirtualFrame::CallCodeObject(Handle<Code> code,
|
| + RelocInfo::Mode rmode,
|
| + int dropped_args) {
|
| + int spilled_args = 0;
|
| + switch (code->kind()) {
|
| + case Code::CALL_IC:
|
| + spilled_args = dropped_args + 1;
|
| + break;
|
| + case Code::FUNCTION:
|
| + spilled_args = dropped_args + 1;
|
| + break;
|
| + case Code::KEYED_LOAD_IC:
|
| + ASSERT(dropped_args == 0);
|
| + spilled_args = 2;
|
| + break;
|
| + default:
|
| + // The other types of code objects are called with values
|
| + // in specific registers, and are handled in functions with
|
| + // a different signature.
|
| + UNREACHABLE();
|
| + break;
|
| + }
|
| + PrepareForCall(spilled_args, dropped_args);
|
| + return RawCallCodeObject(code, rmode);
|
| +}
|
| +
|
| +
|
| +Result VirtualFrame::CallCodeObject(Handle<Code> code,
|
| + RelocInfo::Mode rmode,
|
| + Result* arg,
|
| + int dropped_args) {
|
| + int spilled_args = 0;
|
| + switch (code->kind()) {
|
| + case Code::LOAD_IC:
|
| + ASSERT(arg->reg().is(r2));
|
| + ASSERT(dropped_args == 0);
|
| + spilled_args = 1;
|
| + break;
|
| + case Code::KEYED_STORE_IC:
|
| + ASSERT(arg->reg().is(r0));
|
| + ASSERT(dropped_args == 0);
|
| + spilled_args = 2;
|
| + break;
|
| + default:
|
| + // No other types of code objects are called with values
|
| + // in exactly one register.
|
| + UNREACHABLE();
|
| + break;
|
| + }
|
| + PrepareForCall(spilled_args, dropped_args);
|
| + arg->Unuse();
|
| + return RawCallCodeObject(code, rmode);
|
| +}
|
| +
|
| +
|
| +Result VirtualFrame::CallCodeObject(Handle<Code> code,
|
| + RelocInfo::Mode rmode,
|
| + Result* arg0,
|
| + Result* arg1,
|
| + int dropped_args) {
|
| + int spilled_args = 1;
|
| + switch (code->kind()) {
|
| + case Code::STORE_IC:
|
| + ASSERT(arg0->reg().is(r0));
|
| + ASSERT(arg1->reg().is(r2));
|
| + ASSERT(dropped_args == 0);
|
| + spilled_args = 1;
|
| + break;
|
| + case Code::BUILTIN:
|
| + ASSERT(*code == Builtins::builtin(Builtins::JSConstructCall));
|
| + ASSERT(arg0->reg().is(r0));
|
| + ASSERT(arg1->reg().is(r1));
|
| + spilled_args = dropped_args + 1;
|
| + break;
|
| + default:
|
| + // No other types of code objects are called with values
|
| + // in exactly two registers.
|
| + UNREACHABLE();
|
| + break;
|
| + }
|
| + PrepareForCall(spilled_args, dropped_args);
|
| + arg0->Unuse();
|
| + arg1->Unuse();
|
| + return RawCallCodeObject(code, rmode);
|
| +}
|
| +
|
| +
|
| +void VirtualFrame::Drop(int count) {
|
| + ASSERT(height() >= count);
|
| + int num_virtual_elements = (elements_.length() - 1) - stack_pointer_;
|
| +
|
| + // Emit code to lower the stack pointer if necessary.
|
| + if (num_virtual_elements < count) {
|
| + int num_dropped = count - num_virtual_elements;
|
| + stack_pointer_ -= num_dropped;
|
| + __ add(sp, sp, Operand(num_dropped * kPointerSize));
|
| + }
|
| +
|
| + // Discard elements from the virtual frame and free any registers.
|
| + for (int i = 0; i < count; i++) {
|
| + FrameElement dropped = elements_.RemoveLast();
|
| + if (dropped.is_register()) {
|
| + Unuse(dropped.reg());
|
| + }
|
| + }
|
| +}
|
| +
|
| +
|
| +Result VirtualFrame::Pop() {
|
| + UNIMPLEMENTED();
|
| + Result invalid(cgen_);
|
| + return invalid;
|
| +}
|
| +
|
| +
|
| +void VirtualFrame::EmitPop(Register reg) {
|
| + ASSERT(stack_pointer_ == elements_.length() - 1);
|
| + stack_pointer_--;
|
| + elements_.RemoveLast();
|
| + __ pop(reg);
|
| +}
|
| +
|
| +
|
| +void VirtualFrame::EmitPush(Register reg) {
|
| + ASSERT(stack_pointer_ == elements_.length() - 1);
|
| + elements_.Add(FrameElement::MemoryElement());
|
| + stack_pointer_++;
|
| + __ push(reg);
|
| +}
|
| +
|
| +
|
| +#undef __
|
| +
|
| +} } // namespace v8::internal
|
|
|