OLD | NEW |
1 // Copyright 2013 the V8 project authors. All rights reserved. | 1 // Copyright 2013 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 426 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
437 | 437 |
438 void LCodeGen::DoCallNew(LCallNew* instr) { | 438 void LCodeGen::DoCallNew(LCallNew* instr) { |
439 ASSERT(ToRegister(instr->context()).is(cp)); | 439 ASSERT(ToRegister(instr->context()).is(cp)); |
440 ASSERT(instr->IsMarkedAsCall()); | 440 ASSERT(instr->IsMarkedAsCall()); |
441 ASSERT(ToRegister(instr->constructor()).is(x1)); | 441 ASSERT(ToRegister(instr->constructor()).is(x1)); |
442 | 442 |
443 __ Mov(x0, instr->arity()); | 443 __ Mov(x0, instr->arity()); |
444 // No cell in x2 for construct type feedback in optimized code. | 444 // No cell in x2 for construct type feedback in optimized code. |
445 __ LoadRoot(x2, Heap::kUndefinedValueRootIndex); | 445 __ LoadRoot(x2, Heap::kUndefinedValueRootIndex); |
446 | 446 |
447 CallConstructStub stub(NO_CALL_CONSTRUCTOR_FLAGS); | 447 CallConstructStub stub(NO_CALL_FUNCTION_FLAGS); |
448 CallCode(stub.GetCode(isolate()), RelocInfo::CONSTRUCT_CALL, instr); | 448 CallCode(stub.GetCode(isolate()), RelocInfo::CONSTRUCT_CALL, instr); |
449 | 449 |
450 ASSERT(ToRegister(instr->result()).is(x0)); | 450 ASSERT(ToRegister(instr->result()).is(x0)); |
451 } | 451 } |
452 | 452 |
453 | 453 |
454 void LCodeGen::DoCallNewArray(LCallNewArray* instr) { | 454 void LCodeGen::DoCallNewArray(LCallNewArray* instr) { |
455 ASSERT(instr->IsMarkedAsCall()); | 455 ASSERT(instr->IsMarkedAsCall()); |
456 ASSERT(ToRegister(instr->context()).is(cp)); | 456 ASSERT(ToRegister(instr->context()).is(cp)); |
457 ASSERT(ToRegister(instr->constructor()).is(x1)); | 457 ASSERT(ToRegister(instr->constructor()).is(x1)); |
(...skipping 5492 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
5950 __ Ldr(result, FieldMemOperand(object, JSObject::kPropertiesOffset)); | 5950 __ Ldr(result, FieldMemOperand(object, JSObject::kPropertiesOffset)); |
5951 // Index is equal to negated out of object property index plus 1. | 5951 // Index is equal to negated out of object property index plus 1. |
5952 __ Sub(result, result, Operand::UntagSmiAndScale(index, kPointerSizeLog2)); | 5952 __ Sub(result, result, Operand::UntagSmiAndScale(index, kPointerSizeLog2)); |
5953 __ Ldr(result, FieldMemOperand(result, | 5953 __ Ldr(result, FieldMemOperand(result, |
5954 FixedArray::kHeaderSize - kPointerSize)); | 5954 FixedArray::kHeaderSize - kPointerSize)); |
5955 __ Bind(deferred->exit()); | 5955 __ Bind(deferred->exit()); |
5956 __ Bind(&done); | 5956 __ Bind(&done); |
5957 } | 5957 } |
5958 | 5958 |
5959 } } // namespace v8::internal | 5959 } } // namespace v8::internal |
OLD | NEW |