| Index: src/compiler/code-assembler.cc
|
| diff --git a/src/compiler/code-assembler.cc b/src/compiler/code-assembler.cc
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..f75ed852dd68fe9c15542f04ebce6a3af9b7015c
|
| --- /dev/null
|
| +++ b/src/compiler/code-assembler.cc
|
| @@ -0,0 +1,814 @@
|
| +// Copyright 2015 the V8 project authors. All rights reserved.
|
| +// Use of this source code is governed by a BSD-style license that can be
|
| +// found in the LICENSE file.
|
| +
|
| +#include "src/compiler/code-assembler.h"
|
| +
|
| +#include <ostream>
|
| +
|
| +#include "src/code-factory.h"
|
| +#include "src/compiler/graph.h"
|
| +#include "src/compiler/instruction-selector.h"
|
| +#include "src/compiler/linkage.h"
|
| +#include "src/compiler/pipeline.h"
|
| +#include "src/compiler/raw-machine-assembler.h"
|
| +#include "src/compiler/schedule.h"
|
| +#include "src/frames.h"
|
| +#include "src/interface-descriptors.h"
|
| +#include "src/interpreter/bytecodes.h"
|
| +#include "src/machine-type.h"
|
| +#include "src/macro-assembler.h"
|
| +#include "src/zone.h"
|
| +
|
| +namespace v8 {
|
| +namespace internal {
|
| +namespace compiler {
|
| +
|
| +CodeAssembler::CodeAssembler(Isolate* isolate, Zone* zone,
|
| + const CallInterfaceDescriptor& descriptor,
|
| + Code::Flags flags, const char* name,
|
| + size_t result_size)
|
| + : CodeAssembler(
|
| + isolate, zone,
|
| + Linkage::GetStubCallDescriptor(
|
| + isolate, zone, descriptor, descriptor.GetStackParameterCount(),
|
| + CallDescriptor::kNoFlags, Operator::kNoProperties,
|
| + MachineType::AnyTagged(), result_size),
|
| + flags, name) {}
|
| +
|
| +CodeAssembler::CodeAssembler(Isolate* isolate, Zone* zone, int parameter_count,
|
| + Code::Flags flags, const char* name)
|
| + : CodeAssembler(isolate, zone,
|
| + Linkage::GetJSCallDescriptor(zone, false, parameter_count,
|
| + CallDescriptor::kNoFlags),
|
| + flags, name) {}
|
| +
|
| +CodeAssembler::CodeAssembler(Isolate* isolate, Zone* zone,
|
| + CallDescriptor* call_descriptor, Code::Flags flags,
|
| + const char* name)
|
| + : raw_assembler_(new RawMachineAssembler(
|
| + isolate, new (zone) Graph(zone), call_descriptor,
|
| + MachineType::PointerRepresentation(),
|
| + InstructionSelector::SupportedMachineOperatorFlags())),
|
| + flags_(flags),
|
| + name_(name),
|
| + code_generated_(false),
|
| + variables_(zone) {}
|
| +
|
| +CodeAssembler::~CodeAssembler() {}
|
| +
|
| +void CodeAssembler::CallPrologue() {}
|
| +
|
| +void CodeAssembler::CallEpilogue() {}
|
| +
|
| +Handle<Code> CodeAssembler::GenerateCode() {
|
| + DCHECK(!code_generated_);
|
| +
|
| + Schedule* schedule = raw_assembler_->Export();
|
| + Handle<Code> code = Pipeline::GenerateCodeForCodeStub(
|
| + isolate(), raw_assembler_->call_descriptor(), graph(), schedule, flags_,
|
| + name_);
|
| +
|
| + code_generated_ = true;
|
| + return code;
|
| +}
|
| +
|
| +bool CodeAssembler::Is64() const { return raw_assembler_->machine()->Is64(); }
|
| +
|
| +bool CodeAssembler::IsFloat64RoundUpSupported() const {
|
| + return raw_assembler_->machine()->Float64RoundUp().IsSupported();
|
| +}
|
| +
|
| +bool CodeAssembler::IsFloat64RoundDownSupported() const {
|
| + return raw_assembler_->machine()->Float64RoundDown().IsSupported();
|
| +}
|
| +
|
| +bool CodeAssembler::IsFloat64RoundTruncateSupported() const {
|
| + return raw_assembler_->machine()->Float64RoundTruncate().IsSupported();
|
| +}
|
| +
|
| +Node* CodeAssembler::Int32Constant(int value) {
|
| + return raw_assembler_->Int32Constant(value);
|
| +}
|
| +
|
| +Node* CodeAssembler::IntPtrConstant(intptr_t value) {
|
| + return raw_assembler_->IntPtrConstant(value);
|
| +}
|
| +
|
| +Node* CodeAssembler::NumberConstant(double value) {
|
| + return raw_assembler_->NumberConstant(value);
|
| +}
|
| +
|
| +Node* CodeAssembler::SmiConstant(Smi* value) {
|
| + return IntPtrConstant(bit_cast<intptr_t>(value));
|
| +}
|
| +
|
| +Node* CodeAssembler::HeapConstant(Handle<HeapObject> object) {
|
| + return raw_assembler_->HeapConstant(object);
|
| +}
|
| +
|
| +Node* CodeAssembler::BooleanConstant(bool value) {
|
| + return raw_assembler_->BooleanConstant(value);
|
| +}
|
| +
|
| +Node* CodeAssembler::ExternalConstant(ExternalReference address) {
|
| + return raw_assembler_->ExternalConstant(address);
|
| +}
|
| +
|
| +Node* CodeAssembler::Float64Constant(double value) {
|
| + return raw_assembler_->Float64Constant(value);
|
| +}
|
| +
|
| +Node* CodeAssembler::BooleanMapConstant() {
|
| + return HeapConstant(isolate()->factory()->boolean_map());
|
| +}
|
| +
|
| +Node* CodeAssembler::EmptyStringConstant() {
|
| + return LoadRoot(Heap::kempty_stringRootIndex);
|
| +}
|
| +
|
| +Node* CodeAssembler::HeapNumberMapConstant() {
|
| + return HeapConstant(isolate()->factory()->heap_number_map());
|
| +}
|
| +
|
| +Node* CodeAssembler::NaNConstant() {
|
| + return LoadRoot(Heap::kNanValueRootIndex);
|
| +}
|
| +
|
| +Node* CodeAssembler::NoContextConstant() {
|
| + return SmiConstant(Smi::FromInt(0));
|
| +}
|
| +
|
| +Node* CodeAssembler::NullConstant() {
|
| + return LoadRoot(Heap::kNullValueRootIndex);
|
| +}
|
| +
|
| +Node* CodeAssembler::UndefinedConstant() {
|
| + return LoadRoot(Heap::kUndefinedValueRootIndex);
|
| +}
|
| +
|
| +Node* CodeAssembler::Parameter(int value) {
|
| + return raw_assembler_->Parameter(value);
|
| +}
|
| +
|
| +void CodeAssembler::Return(Node* value) {
|
| + return raw_assembler_->Return(value);
|
| +}
|
| +
|
| +void CodeAssembler::Bind(CodeAssembler::Label* label) { return label->Bind(); }
|
| +
|
| +Node* CodeAssembler::LoadFramePointer() {
|
| + return raw_assembler_->LoadFramePointer();
|
| +}
|
| +
|
| +Node* CodeAssembler::LoadParentFramePointer() {
|
| + return raw_assembler_->LoadParentFramePointer();
|
| +}
|
| +
|
| +Node* CodeAssembler::LoadStackPointer() {
|
| + return raw_assembler_->LoadStackPointer();
|
| +}
|
| +
|
| +Node* CodeAssembler::SmiShiftBitsConstant() {
|
| + return IntPtrConstant(kSmiShiftSize + kSmiTagSize);
|
| +}
|
| +
|
| +Node* CodeAssembler::SmiTag(Node* value) {
|
| + return raw_assembler_->WordShl(value, SmiShiftBitsConstant());
|
| +}
|
| +
|
| +Node* CodeAssembler::SmiUntag(Node* value) {
|
| + return raw_assembler_->WordSar(value, SmiShiftBitsConstant());
|
| +}
|
| +
|
| +#define DEFINE_CODE_ASSEMBLER_BINARY_OP(name) \
|
| + Node* CodeAssembler::name(Node* a, Node* b) { \
|
| + return raw_assembler_->name(a, b); \
|
| + }
|
| +CODE_ASSEMBLER_BINARY_OP_LIST(DEFINE_CODE_ASSEMBLER_BINARY_OP)
|
| +#undef DEFINE_CODE_ASSEMBLER_BINARY_OP
|
| +
|
| +Node* CodeAssembler::WordShl(Node* value, int shift) {
|
| + return raw_assembler_->WordShl(value, IntPtrConstant(shift));
|
| +}
|
| +
|
| +Node* CodeAssembler::TruncateFloat64ToInt32RoundToZero(Node* a) {
|
| + return raw_assembler_->TruncateFloat64ToInt32(TruncationMode::kRoundToZero,
|
| + a);
|
| +}
|
| +
|
| +Node* CodeAssembler::TruncateFloat64ToInt32JavaScript(Node* a) {
|
| + return raw_assembler_->TruncateFloat64ToInt32(TruncationMode::kJavaScript, a);
|
| +}
|
| +
|
| +#define DEFINE_CODE_ASSEMBLER_UNARY_OP(name) \
|
| + Node* CodeAssembler::name(Node* a) { return raw_assembler_->name(a); }
|
| +CODE_ASSEMBLER_UNARY_OP_LIST(DEFINE_CODE_ASSEMBLER_UNARY_OP)
|
| +#undef DEFINE_CODE_ASSEMBLER_UNARY_OP
|
| +
|
| +Node* CodeAssembler::LoadRoot(Heap::RootListIndex root_index) {
|
| + if (isolate()->heap()->RootCanBeTreatedAsConstant(root_index)) {
|
| + Handle<Object> root = isolate()->heap()->root_handle(root_index);
|
| + if (root->IsSmi()) {
|
| + return SmiConstant(Smi::cast(*root));
|
| + } else {
|
| + return HeapConstant(Handle<HeapObject>::cast(root));
|
| + }
|
| + }
|
| +
|
| + compiler::Node* roots_array_start =
|
| + ExternalConstant(ExternalReference::roots_array_start(isolate()));
|
| + USE(roots_array_start);
|
| +
|
| + // TODO(danno): Implement thee root-access case where the root is not constant
|
| + // and must be loaded from the root array.
|
| + UNIMPLEMENTED();
|
| + return nullptr;
|
| +}
|
| +
|
| +Node* CodeAssembler::AllocateRawUnaligned(Node* size_in_bytes,
|
| + AllocationFlags flags,
|
| + Node* top_address,
|
| + Node* limit_address) {
|
| + Node* top = Load(MachineType::Pointer(), top_address);
|
| + Node* limit = Load(MachineType::Pointer(), limit_address);
|
| +
|
| + // If there's not enough space, call the runtime.
|
| + RawMachineLabel runtime_call(RawMachineLabel::kDeferred), no_runtime_call,
|
| + merge_runtime;
|
| + raw_assembler_->Branch(
|
| + raw_assembler_->IntPtrLessThan(IntPtrSub(limit, top), size_in_bytes),
|
| + &runtime_call, &no_runtime_call);
|
| +
|
| + raw_assembler_->Bind(&runtime_call);
|
| + // AllocateInTargetSpace does not use the context.
|
| + Node* context = IntPtrConstant(0);
|
| + Node* runtime_flags = SmiTag(Int32Constant(
|
| + AllocateDoubleAlignFlag::encode(false) |
|
| + AllocateTargetSpace::encode(flags & kPretenured
|
| + ? AllocationSpace::OLD_SPACE
|
| + : AllocationSpace::NEW_SPACE)));
|
| + Node* runtime_result = CallRuntime(Runtime::kAllocateInTargetSpace, context,
|
| + SmiTag(size_in_bytes), runtime_flags);
|
| + raw_assembler_->Goto(&merge_runtime);
|
| +
|
| + // When there is enough space, return `top' and bump it up.
|
| + raw_assembler_->Bind(&no_runtime_call);
|
| + Node* no_runtime_result = top;
|
| + StoreNoWriteBarrier(MachineType::PointerRepresentation(), top_address,
|
| + IntPtrAdd(top, size_in_bytes));
|
| + no_runtime_result =
|
| + IntPtrAdd(no_runtime_result, IntPtrConstant(kHeapObjectTag));
|
| + raw_assembler_->Goto(&merge_runtime);
|
| +
|
| + raw_assembler_->Bind(&merge_runtime);
|
| + return raw_assembler_->Phi(MachineType::PointerRepresentation(),
|
| + runtime_result, no_runtime_result);
|
| +}
|
| +
|
| +Node* CodeAssembler::AllocateRawAligned(Node* size_in_bytes,
|
| + AllocationFlags flags,
|
| + Node* top_address,
|
| + Node* limit_address) {
|
| + Node* top = Load(MachineType::Pointer(), top_address);
|
| + Node* limit = Load(MachineType::Pointer(), limit_address);
|
| + Node* adjusted_size = size_in_bytes;
|
| + if (flags & kDoubleAlignment) {
|
| + // TODO(epertoso): Simd128 alignment.
|
| + RawMachineLabel aligned, not_aligned, merge;
|
| + raw_assembler_->Branch(WordAnd(top, IntPtrConstant(kDoubleAlignmentMask)),
|
| + ¬_aligned, &aligned);
|
| +
|
| + raw_assembler_->Bind(¬_aligned);
|
| + Node* not_aligned_size =
|
| + IntPtrAdd(size_in_bytes, IntPtrConstant(kPointerSize));
|
| + raw_assembler_->Goto(&merge);
|
| +
|
| + raw_assembler_->Bind(&aligned);
|
| + raw_assembler_->Goto(&merge);
|
| +
|
| + raw_assembler_->Bind(&merge);
|
| + adjusted_size = raw_assembler_->Phi(MachineType::PointerRepresentation(),
|
| + not_aligned_size, adjusted_size);
|
| + }
|
| +
|
| + Node* address = AllocateRawUnaligned(adjusted_size, kNone, top, limit);
|
| +
|
| + RawMachineLabel needs_filler, doesnt_need_filler, merge_address;
|
| + raw_assembler_->Branch(
|
| + raw_assembler_->IntPtrEqual(adjusted_size, size_in_bytes),
|
| + &doesnt_need_filler, &needs_filler);
|
| +
|
| + raw_assembler_->Bind(&needs_filler);
|
| + // Store a filler and increase the address by kPointerSize.
|
| + // TODO(epertoso): this code assumes that we only align to kDoubleSize. Change
|
| + // it when Simd128 alignment is supported.
|
| + StoreNoWriteBarrier(MachineType::PointerRepresentation(), top,
|
| + LoadRoot(Heap::kOnePointerFillerMapRootIndex));
|
| + Node* address_with_filler = IntPtrAdd(address, IntPtrConstant(kPointerSize));
|
| + raw_assembler_->Goto(&merge_address);
|
| +
|
| + raw_assembler_->Bind(&doesnt_need_filler);
|
| + Node* address_without_filler = address;
|
| + raw_assembler_->Goto(&merge_address);
|
| +
|
| + raw_assembler_->Bind(&merge_address);
|
| + address = raw_assembler_->Phi(MachineType::PointerRepresentation(),
|
| + address_with_filler, address_without_filler);
|
| + // Update the top.
|
| + StoreNoWriteBarrier(MachineType::PointerRepresentation(), top_address,
|
| + IntPtrAdd(top, adjusted_size));
|
| + return address;
|
| +}
|
| +
|
| +Node* CodeAssembler::Allocate(int size_in_bytes, AllocationFlags flags) {
|
| + bool const new_space = !(flags & kPretenured);
|
| + Node* top_address = ExternalConstant(
|
| + new_space
|
| + ? ExternalReference::new_space_allocation_top_address(isolate())
|
| + : ExternalReference::old_space_allocation_top_address(isolate()));
|
| + Node* limit_address = ExternalConstant(
|
| + new_space
|
| + ? ExternalReference::new_space_allocation_limit_address(isolate())
|
| + : ExternalReference::old_space_allocation_limit_address(isolate()));
|
| +
|
| +#ifdef V8_HOST_ARCH_32_BIT
|
| + if (flags & kDoubleAlignment) {
|
| + return AllocateRawAligned(IntPtrConstant(size_in_bytes), flags, top_address,
|
| + limit_address);
|
| + }
|
| +#endif
|
| +
|
| + return AllocateRawUnaligned(IntPtrConstant(size_in_bytes), flags, top_address,
|
| + limit_address);
|
| +}
|
| +
|
| +Node* CodeAssembler::InnerAllocate(Node* previous, int offset) {
|
| + return IntPtrAdd(previous, IntPtrConstant(offset));
|
| +}
|
| +
|
| +Node* CodeAssembler::Load(MachineType rep, Node* base) {
|
| + return raw_assembler_->Load(rep, base);
|
| +}
|
| +
|
| +Node* CodeAssembler::Load(MachineType rep, Node* base, Node* index) {
|
| + return raw_assembler_->Load(rep, base, index);
|
| +}
|
| +
|
| +Node* CodeAssembler::Store(MachineRepresentation rep, Node* base, Node* value) {
|
| + return raw_assembler_->Store(rep, base, value, kFullWriteBarrier);
|
| +}
|
| +
|
| +Node* CodeAssembler::Store(MachineRepresentation rep, Node* base, Node* index,
|
| + Node* value) {
|
| + return raw_assembler_->Store(rep, base, index, value, kFullWriteBarrier);
|
| +}
|
| +
|
| +Node* CodeAssembler::StoreNoWriteBarrier(MachineRepresentation rep, Node* base,
|
| + Node* value) {
|
| + return raw_assembler_->Store(rep, base, value, kNoWriteBarrier);
|
| +}
|
| +
|
| +Node* CodeAssembler::StoreNoWriteBarrier(MachineRepresentation rep, Node* base,
|
| + Node* index, Node* value) {
|
| + return raw_assembler_->Store(rep, base, index, value, kNoWriteBarrier);
|
| +}
|
| +
|
| +Node* CodeAssembler::Projection(int index, Node* value) {
|
| + return raw_assembler_->Projection(index, value);
|
| +}
|
| +
|
| +void CodeAssembler::BranchIf(Node* condition, Label* if_true, Label* if_false) {
|
| + Label if_condition_is_true(this), if_condition_is_false(this);
|
| + Branch(condition, &if_condition_is_true, &if_condition_is_false);
|
| + Bind(&if_condition_is_true);
|
| + Goto(if_true);
|
| + Bind(&if_condition_is_false);
|
| + Goto(if_false);
|
| +}
|
| +
|
| +Node* CodeAssembler::CallN(CallDescriptor* descriptor, Node* code_target,
|
| + Node** args) {
|
| + CallPrologue();
|
| + Node* return_value = raw_assembler_->CallN(descriptor, code_target, args);
|
| + CallEpilogue();
|
| + return return_value;
|
| +}
|
| +
|
| +Node* CodeAssembler::TailCallN(CallDescriptor* descriptor, Node* code_target,
|
| + Node** args) {
|
| + return raw_assembler_->TailCallN(descriptor, code_target, args);
|
| +}
|
| +
|
| +Node* CodeAssembler::CallRuntime(Runtime::FunctionId function_id,
|
| + Node* context) {
|
| + CallPrologue();
|
| + Node* return_value = raw_assembler_->CallRuntime0(function_id, context);
|
| + CallEpilogue();
|
| + return return_value;
|
| +}
|
| +
|
| +Node* CodeAssembler::CallRuntime(Runtime::FunctionId function_id, Node* context,
|
| + Node* arg1) {
|
| + CallPrologue();
|
| + Node* return_value = raw_assembler_->CallRuntime1(function_id, arg1, context);
|
| + CallEpilogue();
|
| + return return_value;
|
| +}
|
| +
|
| +Node* CodeAssembler::CallRuntime(Runtime::FunctionId function_id, Node* context,
|
| + Node* arg1, Node* arg2) {
|
| + CallPrologue();
|
| + Node* return_value =
|
| + raw_assembler_->CallRuntime2(function_id, arg1, arg2, context);
|
| + CallEpilogue();
|
| + return return_value;
|
| +}
|
| +
|
| +Node* CodeAssembler::CallRuntime(Runtime::FunctionId function_id, Node* context,
|
| + Node* arg1, Node* arg2, Node* arg3) {
|
| + CallPrologue();
|
| + Node* return_value =
|
| + raw_assembler_->CallRuntime3(function_id, arg1, arg2, arg3, context);
|
| + CallEpilogue();
|
| + return return_value;
|
| +}
|
| +
|
| +Node* CodeAssembler::CallRuntime(Runtime::FunctionId function_id, Node* context,
|
| + Node* arg1, Node* arg2, Node* arg3,
|
| + Node* arg4) {
|
| + CallPrologue();
|
| + Node* return_value = raw_assembler_->CallRuntime4(function_id, arg1, arg2,
|
| + arg3, arg4, context);
|
| + CallEpilogue();
|
| + return return_value;
|
| +}
|
| +
|
| +Node* CodeAssembler::TailCallRuntime(Runtime::FunctionId function_id,
|
| + Node* context) {
|
| + return raw_assembler_->TailCallRuntime0(function_id, context);
|
| +}
|
| +
|
| +Node* CodeAssembler::TailCallRuntime(Runtime::FunctionId function_id,
|
| + Node* context, Node* arg1) {
|
| + return raw_assembler_->TailCallRuntime1(function_id, arg1, context);
|
| +}
|
| +
|
| +Node* CodeAssembler::TailCallRuntime(Runtime::FunctionId function_id,
|
| + Node* context, Node* arg1, Node* arg2) {
|
| + return raw_assembler_->TailCallRuntime2(function_id, arg1, arg2, context);
|
| +}
|
| +
|
| +Node* CodeAssembler::TailCallRuntime(Runtime::FunctionId function_id,
|
| + Node* context, Node* arg1, Node* arg2,
|
| + Node* arg3) {
|
| + return raw_assembler_->TailCallRuntime3(function_id, arg1, arg2, arg3,
|
| + context);
|
| +}
|
| +
|
| +Node* CodeAssembler::TailCallRuntime(Runtime::FunctionId function_id,
|
| + Node* context, Node* arg1, Node* arg2,
|
| + Node* arg3, Node* arg4) {
|
| + return raw_assembler_->TailCallRuntime4(function_id, arg1, arg2, arg3, arg4,
|
| + context);
|
| +}
|
| +
|
| +Node* CodeAssembler::CallStub(Callable const& callable, Node* context,
|
| + Node* arg1, size_t result_size) {
|
| + Node* target = HeapConstant(callable.code());
|
| + return CallStub(callable.descriptor(), target, context, arg1, result_size);
|
| +}
|
| +
|
| +Node* CodeAssembler::CallStub(Callable const& callable, Node* context,
|
| + Node* arg1, Node* arg2, size_t result_size) {
|
| + Node* target = HeapConstant(callable.code());
|
| + return CallStub(callable.descriptor(), target, context, arg1, arg2,
|
| + result_size);
|
| +}
|
| +
|
| +Node* CodeAssembler::CallStub(Callable const& callable, Node* context,
|
| + Node* arg1, Node* arg2, Node* arg3,
|
| + size_t result_size) {
|
| + Node* target = HeapConstant(callable.code());
|
| + return CallStub(callable.descriptor(), target, context, arg1, arg2, arg3,
|
| + result_size);
|
| +}
|
| +
|
| +Node* CodeAssembler::CallStub(const CallInterfaceDescriptor& descriptor,
|
| + Node* target, Node* context, Node* arg1,
|
| + size_t result_size) {
|
| + CallDescriptor* call_descriptor = Linkage::GetStubCallDescriptor(
|
| + isolate(), zone(), descriptor, descriptor.GetStackParameterCount(),
|
| + CallDescriptor::kNoFlags, Operator::kNoProperties,
|
| + MachineType::AnyTagged(), result_size);
|
| +
|
| + Node** args = zone()->NewArray<Node*>(2);
|
| + args[0] = arg1;
|
| + args[1] = context;
|
| +
|
| + return CallN(call_descriptor, target, args);
|
| +}
|
| +
|
| +Node* CodeAssembler::CallStub(const CallInterfaceDescriptor& descriptor,
|
| + Node* target, Node* context, Node* arg1,
|
| + Node* arg2, size_t result_size) {
|
| + CallDescriptor* call_descriptor = Linkage::GetStubCallDescriptor(
|
| + isolate(), zone(), descriptor, descriptor.GetStackParameterCount(),
|
| + CallDescriptor::kNoFlags, Operator::kNoProperties,
|
| + MachineType::AnyTagged(), result_size);
|
| +
|
| + Node** args = zone()->NewArray<Node*>(3);
|
| + args[0] = arg1;
|
| + args[1] = arg2;
|
| + args[2] = context;
|
| +
|
| + return CallN(call_descriptor, target, args);
|
| +}
|
| +
|
| +Node* CodeAssembler::CallStub(const CallInterfaceDescriptor& descriptor,
|
| + Node* target, Node* context, Node* arg1,
|
| + Node* arg2, Node* arg3, size_t result_size) {
|
| + CallDescriptor* call_descriptor = Linkage::GetStubCallDescriptor(
|
| + isolate(), zone(), descriptor, descriptor.GetStackParameterCount(),
|
| + CallDescriptor::kNoFlags, Operator::kNoProperties,
|
| + MachineType::AnyTagged(), result_size);
|
| +
|
| + Node** args = zone()->NewArray<Node*>(4);
|
| + args[0] = arg1;
|
| + args[1] = arg2;
|
| + args[2] = arg3;
|
| + args[3] = context;
|
| +
|
| + return CallN(call_descriptor, target, args);
|
| +}
|
| +
|
| +Node* CodeAssembler::CallStub(const CallInterfaceDescriptor& descriptor,
|
| + Node* target, Node* context, Node* arg1,
|
| + Node* arg2, Node* arg3, Node* arg4,
|
| + size_t result_size) {
|
| + CallDescriptor* call_descriptor = Linkage::GetStubCallDescriptor(
|
| + isolate(), zone(), descriptor, descriptor.GetStackParameterCount(),
|
| + CallDescriptor::kNoFlags, Operator::kNoProperties,
|
| + MachineType::AnyTagged(), result_size);
|
| +
|
| + Node** args = zone()->NewArray<Node*>(5);
|
| + args[0] = arg1;
|
| + args[1] = arg2;
|
| + args[2] = arg3;
|
| + args[3] = arg4;
|
| + args[4] = context;
|
| +
|
| + return CallN(call_descriptor, target, args);
|
| +}
|
| +
|
| +Node* CodeAssembler::CallStub(const CallInterfaceDescriptor& descriptor,
|
| + Node* target, Node* context, Node* arg1,
|
| + Node* arg2, Node* arg3, Node* arg4, Node* arg5,
|
| + size_t result_size) {
|
| + CallDescriptor* call_descriptor = Linkage::GetStubCallDescriptor(
|
| + isolate(), zone(), descriptor, descriptor.GetStackParameterCount(),
|
| + CallDescriptor::kNoFlags, Operator::kNoProperties,
|
| + MachineType::AnyTagged(), result_size);
|
| +
|
| + Node** args = zone()->NewArray<Node*>(6);
|
| + args[0] = arg1;
|
| + args[1] = arg2;
|
| + args[2] = arg3;
|
| + args[3] = arg4;
|
| + args[4] = arg5;
|
| + args[5] = context;
|
| +
|
| + return CallN(call_descriptor, target, args);
|
| +}
|
| +
|
| +Node* CodeAssembler::TailCallStub(Callable const& callable, Node* context,
|
| + Node* arg1, Node* arg2, size_t result_size) {
|
| + Node* target = HeapConstant(callable.code());
|
| + return TailCallStub(callable.descriptor(), target, context, arg1, arg2,
|
| + result_size);
|
| +}
|
| +
|
| +Node* CodeAssembler::TailCallStub(const CallInterfaceDescriptor& descriptor,
|
| + Node* target, Node* context, Node* arg1,
|
| + Node* arg2, size_t result_size) {
|
| + CallDescriptor* call_descriptor = Linkage::GetStubCallDescriptor(
|
| + isolate(), zone(), descriptor, descriptor.GetStackParameterCount(),
|
| + CallDescriptor::kSupportsTailCalls, Operator::kNoProperties,
|
| + MachineType::AnyTagged(), result_size);
|
| +
|
| + Node** args = zone()->NewArray<Node*>(3);
|
| + args[0] = arg1;
|
| + args[1] = arg2;
|
| + args[2] = context;
|
| +
|
| + return raw_assembler_->TailCallN(call_descriptor, target, args);
|
| +}
|
| +
|
| +Node* CodeAssembler::TailCallBytecodeDispatch(
|
| + const CallInterfaceDescriptor& interface_descriptor,
|
| + Node* code_target_address, Node** args) {
|
| + CallDescriptor* descriptor = Linkage::GetBytecodeDispatchCallDescriptor(
|
| + isolate(), zone(), interface_descriptor,
|
| + interface_descriptor.GetStackParameterCount());
|
| + return raw_assembler_->TailCallN(descriptor, code_target_address, args);
|
| +}
|
| +
|
| +void CodeAssembler::Goto(CodeAssembler::Label* label) {
|
| + label->MergeVariables();
|
| + raw_assembler_->Goto(label->label_);
|
| +}
|
| +
|
| +void CodeAssembler::GotoIf(Node* condition, Label* true_label) {
|
| + Label false_label(this);
|
| + Branch(condition, true_label, &false_label);
|
| + Bind(&false_label);
|
| +}
|
| +
|
| +void CodeAssembler::GotoUnless(Node* condition, Label* false_label) {
|
| + Label true_label(this);
|
| + Branch(condition, &true_label, false_label);
|
| + Bind(&true_label);
|
| +}
|
| +
|
| +void CodeAssembler::Branch(Node* condition, CodeAssembler::Label* true_label,
|
| + CodeAssembler::Label* false_label) {
|
| + true_label->MergeVariables();
|
| + false_label->MergeVariables();
|
| + return raw_assembler_->Branch(condition, true_label->label_,
|
| + false_label->label_);
|
| +}
|
| +
|
| +void CodeAssembler::Switch(Node* index, Label* default_label,
|
| + int32_t* case_values, Label** case_labels,
|
| + size_t case_count) {
|
| + RawMachineLabel** labels =
|
| + new (zone()->New(sizeof(RawMachineLabel*) * case_count))
|
| + RawMachineLabel*[case_count];
|
| + for (size_t i = 0; i < case_count; ++i) {
|
| + labels[i] = case_labels[i]->label_;
|
| + case_labels[i]->MergeVariables();
|
| + default_label->MergeVariables();
|
| + }
|
| + return raw_assembler_->Switch(index, default_label->label_, case_values,
|
| + labels, case_count);
|
| +}
|
| +
|
| +// RawMachineAssembler delegate helpers:
|
| +Isolate* CodeAssembler::isolate() const { return raw_assembler_->isolate(); }
|
| +
|
| +Factory* CodeAssembler::factory() const { return isolate()->factory(); }
|
| +
|
| +Graph* CodeAssembler::graph() const { return raw_assembler_->graph(); }
|
| +
|
| +Zone* CodeAssembler::zone() const { return raw_assembler_->zone(); }
|
| +
|
| +// The core implementation of Variable is stored through an indirection so
|
| +// that it can outlive the often block-scoped Variable declarations. This is
|
| +// needed to ensure that variable binding and merging through phis can
|
| +// properly be verified.
|
| +class CodeAssembler::Variable::Impl : public ZoneObject {
|
| + public:
|
| + explicit Impl(MachineRepresentation rep) : value_(nullptr), rep_(rep) {}
|
| + Node* value_;
|
| + MachineRepresentation rep_;
|
| +};
|
| +
|
| +CodeAssembler::Variable::Variable(CodeAssembler* assembler,
|
| + MachineRepresentation rep)
|
| + : impl_(new (assembler->zone()) Impl(rep)) {
|
| + assembler->variables_.push_back(impl_);
|
| +}
|
| +
|
| +void CodeAssembler::Variable::Bind(Node* value) { impl_->value_ = value; }
|
| +
|
| +Node* CodeAssembler::Variable::value() const {
|
| + DCHECK_NOT_NULL(impl_->value_);
|
| + return impl_->value_;
|
| +}
|
| +
|
| +MachineRepresentation CodeAssembler::Variable::rep() const {
|
| + return impl_->rep_;
|
| +}
|
| +
|
| +bool CodeAssembler::Variable::IsBound() const {
|
| + return impl_->value_ != nullptr;
|
| +}
|
| +
|
| +CodeAssembler::Label::Label(CodeAssembler* assembler, int merged_value_count,
|
| + CodeAssembler::Variable** merged_variables,
|
| + CodeAssembler::Label::Type type)
|
| + : bound_(false), merge_count_(0), assembler_(assembler), label_(nullptr) {
|
| + void* buffer = assembler->zone()->New(sizeof(RawMachineLabel));
|
| + label_ = new (buffer)
|
| + RawMachineLabel(type == kDeferred ? RawMachineLabel::kDeferred
|
| + : RawMachineLabel::kNonDeferred);
|
| + for (int i = 0; i < merged_value_count; ++i) {
|
| + variable_phis_[merged_variables[i]->impl_] = nullptr;
|
| + }
|
| +}
|
| +
|
| +void CodeAssembler::Label::MergeVariables() {
|
| + ++merge_count_;
|
| + for (auto var : assembler_->variables_) {
|
| + size_t count = 0;
|
| + Node* node = var->value_;
|
| + if (node != nullptr) {
|
| + auto i = variable_merges_.find(var);
|
| + if (i != variable_merges_.end()) {
|
| + i->second.push_back(node);
|
| + count = i->second.size();
|
| + } else {
|
| + count = 1;
|
| + variable_merges_[var] = std::vector<Node*>(1, node);
|
| + }
|
| + }
|
| + // If the following asserts, then you've jumped to a label without a bound
|
| + // variable along that path that expects to merge its value into a phi.
|
| + DCHECK(variable_phis_.find(var) == variable_phis_.end() ||
|
| + count == merge_count_);
|
| + USE(count);
|
| +
|
| + // If the label is already bound, we already know the set of variables to
|
| + // merge and phi nodes have already been created.
|
| + if (bound_) {
|
| + auto phi = variable_phis_.find(var);
|
| + if (phi != variable_phis_.end()) {
|
| + DCHECK_NOT_NULL(phi->second);
|
| + assembler_->raw_assembler_->AppendPhiInput(phi->second, node);
|
| + } else {
|
| + auto i = variable_merges_.find(var);
|
| + if (i != variable_merges_.end()) {
|
| + // If the following assert fires, then you've declared a variable that
|
| + // has the same bound value along all paths up until the point you
|
| + // bound this label, but then later merged a path with a new value for
|
| + // the variable after the label bind (it's not possible to add phis to
|
| + // the bound label after the fact, just make sure to list the variable
|
| + // in the label's constructor's list of merged variables).
|
| + DCHECK(find_if(i->second.begin(), i->second.end(),
|
| + [node](Node* e) -> bool { return node != e; }) ==
|
| + i->second.end());
|
| + }
|
| + }
|
| + }
|
| + }
|
| +}
|
| +
|
| +void CodeAssembler::Label::Bind() {
|
| + DCHECK(!bound_);
|
| + assembler_->raw_assembler_->Bind(label_);
|
| +
|
| + // Make sure that all variables that have changed along any path up to this
|
| + // point are marked as merge variables.
|
| + for (auto var : assembler_->variables_) {
|
| + Node* shared_value = nullptr;
|
| + auto i = variable_merges_.find(var);
|
| + if (i != variable_merges_.end()) {
|
| + for (auto value : i->second) {
|
| + DCHECK(value != nullptr);
|
| + if (value != shared_value) {
|
| + if (shared_value == nullptr) {
|
| + shared_value = value;
|
| + } else {
|
| + variable_phis_[var] = nullptr;
|
| + }
|
| + }
|
| + }
|
| + }
|
| + }
|
| +
|
| + for (auto var : variable_phis_) {
|
| + CodeAssembler::Variable::Impl* var_impl = var.first;
|
| + auto i = variable_merges_.find(var_impl);
|
| + // If the following assert fires, then a variable that has been marked as
|
| + // being merged at the label--either by explicitly marking it so in the
|
| + // label constructor or by having seen different bound values at branches
|
| + // into the label--doesn't have a bound value along all of the paths that
|
| + // have been merged into the label up to this point.
|
| + DCHECK(i != variable_merges_.end() && i->second.size() == merge_count_);
|
| + Node* phi = assembler_->raw_assembler_->Phi(
|
| + var.first->rep_, static_cast<int>(merge_count_), &(i->second[0]));
|
| + variable_phis_[var_impl] = phi;
|
| + }
|
| +
|
| + // Bind all variables to a merge phi, the common value along all paths or
|
| + // null.
|
| + for (auto var : assembler_->variables_) {
|
| + auto i = variable_phis_.find(var);
|
| + if (i != variable_phis_.end()) {
|
| + var->value_ = i->second;
|
| + } else {
|
| + auto j = variable_merges_.find(var);
|
| + if (j != variable_merges_.end() && j->second.size() == merge_count_) {
|
| + var->value_ = j->second.back();
|
| + } else {
|
| + var->value_ = nullptr;
|
| + }
|
| + }
|
| + }
|
| +
|
| + bound_ = true;
|
| +}
|
| +
|
| +} // namespace compiler
|
| +} // namespace internal
|
| +} // namespace v8
|
|
|