Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(412)

Side by Side Diff: src/ia32/code-stubs-ia32.cc

Issue 12521011: Compile FastCloneShallowArrayStub using Crankshaft. (Closed) Base URL: https://v8.googlecode.com/svn/branches/bleeding_edge
Patch Set: Rebased. Created 7 years, 8 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View unified diff | Download patch | Annotate | Revision Log
« no previous file with comments | « src/hydrogen.cc ('k') | src/ia32/full-codegen-ia32.cc » ('j') | no next file with comments »
Toggle Intra-line Diffs ('i') | Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
OLDNEW
1 // Copyright 2012 the V8 project authors. All rights reserved. 1 // Copyright 2012 the V8 project authors. All rights reserved.
2 // Redistribution and use in source and binary forms, with or without 2 // Redistribution and use in source and binary forms, with or without
3 // modification, are permitted provided that the following conditions are 3 // modification, are permitted provided that the following conditions are
4 // met: 4 // met:
5 // 5 //
6 // * Redistributions of source code must retain the above copyright 6 // * Redistributions of source code must retain the above copyright
7 // notice, this list of conditions and the following disclaimer. 7 // notice, this list of conditions and the following disclaimer.
8 // * Redistributions in binary form must reproduce the above 8 // * Redistributions in binary form must reproduce the above
9 // copyright notice, this list of conditions and the following 9 // copyright notice, this list of conditions and the following
10 // disclaimer in the documentation and/or other materials provided 10 // disclaimer in the documentation and/or other materials provided
(...skipping 25 matching lines...) Expand all
36 #include "regexp-macro-assembler.h" 36 #include "regexp-macro-assembler.h"
37 #include "runtime.h" 37 #include "runtime.h"
38 #include "stub-cache.h" 38 #include "stub-cache.h"
39 #include "codegen.h" 39 #include "codegen.h"
40 #include "runtime.h" 40 #include "runtime.h"
41 41
42 namespace v8 { 42 namespace v8 {
43 namespace internal { 43 namespace internal {
44 44
45 45
46 void FastCloneShallowArrayStub::InitializeInterfaceDescriptor(
47 Isolate* isolate,
48 CodeStubInterfaceDescriptor* descriptor) {
49 static Register registers[] = { eax, ebx, ecx };
50 descriptor->register_param_count_ = 3;
51 descriptor->register_params_ = registers;
52 descriptor->stack_parameter_count_ = NULL;
53 descriptor->deoptimization_handler_ =
54 Runtime::FunctionForId(Runtime::kCreateArrayLiteralShallow)->entry;
55 }
56
57
46 void FastCloneShallowObjectStub::InitializeInterfaceDescriptor( 58 void FastCloneShallowObjectStub::InitializeInterfaceDescriptor(
47 Isolate* isolate, 59 Isolate* isolate,
48 CodeStubInterfaceDescriptor* descriptor) { 60 CodeStubInterfaceDescriptor* descriptor) {
49 static Register registers[] = { eax, ebx, ecx, edx }; 61 static Register registers[] = { eax, ebx, ecx, edx };
50 descriptor->register_param_count_ = 4; 62 descriptor->register_param_count_ = 4;
51 descriptor->register_params_ = registers; 63 descriptor->register_params_ = registers;
52 descriptor->stack_parameter_count_ = NULL; 64 descriptor->stack_parameter_count_ = NULL;
53 descriptor->deoptimization_handler_ = 65 descriptor->deoptimization_handler_ =
54 Runtime::FunctionForId(Runtime::kCreateObjectLiteralShallow)->entry; 66 Runtime::FunctionForId(Runtime::kCreateObjectLiteralShallow)->entry;
55 } 67 }
(...skipping 330 matching lines...) Expand 10 before | Expand all | Expand 10 after
386 // Return and remove the on-stack parameters. 398 // Return and remove the on-stack parameters.
387 __ mov(esi, eax); 399 __ mov(esi, eax);
388 __ ret(2 * kPointerSize); 400 __ ret(2 * kPointerSize);
389 401
390 // Need to collect. Call into runtime system. 402 // Need to collect. Call into runtime system.
391 __ bind(&gc); 403 __ bind(&gc);
392 __ TailCallRuntime(Runtime::kPushBlockContext, 2, 1); 404 __ TailCallRuntime(Runtime::kPushBlockContext, 2, 1);
393 } 405 }
394 406
395 407
396 static void GenerateFastCloneShallowArrayCommon(
397 MacroAssembler* masm,
398 int length,
399 FastCloneShallowArrayStub::Mode mode,
400 AllocationSiteMode allocation_site_mode,
401 Label* fail) {
402 // Registers on entry:
403 //
404 // ecx: boilerplate literal array.
405 ASSERT(mode != FastCloneShallowArrayStub::CLONE_ANY_ELEMENTS);
406
407 // All sizes here are multiples of kPointerSize.
408 int elements_size = 0;
409 if (length > 0) {
410 elements_size = mode == FastCloneShallowArrayStub::CLONE_DOUBLE_ELEMENTS
411 ? FixedDoubleArray::SizeFor(length)
412 : FixedArray::SizeFor(length);
413 }
414 int size = JSArray::kSize;
415 int allocation_info_start = size;
416 if (allocation_site_mode == TRACK_ALLOCATION_SITE) {
417 size += AllocationSiteInfo::kSize;
418 }
419 size += elements_size;
420
421 // Allocate both the JS array and the elements array in one big
422 // allocation. This avoids multiple limit checks.
423 AllocationFlags flags = TAG_OBJECT;
424 if (mode == FastCloneShallowArrayStub::CLONE_DOUBLE_ELEMENTS) {
425 flags = static_cast<AllocationFlags>(DOUBLE_ALIGNMENT | flags);
426 }
427 __ Allocate(size, eax, ebx, edx, fail, flags);
428
429 if (allocation_site_mode == TRACK_ALLOCATION_SITE) {
430 __ mov(FieldOperand(eax, allocation_info_start),
431 Immediate(Handle<Map>(masm->isolate()->heap()->
432 allocation_site_info_map())));
433 __ mov(FieldOperand(eax, allocation_info_start + kPointerSize), ecx);
434 }
435
436 // Copy the JS array part.
437 for (int i = 0; i < JSArray::kSize; i += kPointerSize) {
438 if ((i != JSArray::kElementsOffset) || (length == 0)) {
439 __ mov(ebx, FieldOperand(ecx, i));
440 __ mov(FieldOperand(eax, i), ebx);
441 }
442 }
443
444 if (length > 0) {
445 // Get hold of the elements array of the boilerplate and setup the
446 // elements pointer in the resulting object.
447 __ mov(ecx, FieldOperand(ecx, JSArray::kElementsOffset));
448 if (allocation_site_mode == TRACK_ALLOCATION_SITE) {
449 __ lea(edx, Operand(eax, JSArray::kSize + AllocationSiteInfo::kSize));
450 } else {
451 __ lea(edx, Operand(eax, JSArray::kSize));
452 }
453 __ mov(FieldOperand(eax, JSArray::kElementsOffset), edx);
454
455 // Copy the elements array.
456 if (mode == FastCloneShallowArrayStub::CLONE_ELEMENTS) {
457 for (int i = 0; i < elements_size; i += kPointerSize) {
458 __ mov(ebx, FieldOperand(ecx, i));
459 __ mov(FieldOperand(edx, i), ebx);
460 }
461 } else {
462 ASSERT(mode == FastCloneShallowArrayStub::CLONE_DOUBLE_ELEMENTS);
463 int i;
464 for (i = 0; i < FixedDoubleArray::kHeaderSize; i += kPointerSize) {
465 __ mov(ebx, FieldOperand(ecx, i));
466 __ mov(FieldOperand(edx, i), ebx);
467 }
468 while (i < elements_size) {
469 __ fld_d(FieldOperand(ecx, i));
470 __ fstp_d(FieldOperand(edx, i));
471 i += kDoubleSize;
472 }
473 ASSERT(i == elements_size);
474 }
475 }
476 }
477
478
479 void FastCloneShallowArrayStub::Generate(MacroAssembler* masm) {
480 // Stack layout on entry:
481 //
482 // [esp + kPointerSize]: constant elements.
483 // [esp + (2 * kPointerSize)]: literal index.
484 // [esp + (3 * kPointerSize)]: literals array.
485
486 // Load boilerplate object into ecx and check if we need to create a
487 // boilerplate.
488 __ mov(ecx, Operand(esp, 3 * kPointerSize));
489 __ mov(eax, Operand(esp, 2 * kPointerSize));
490 STATIC_ASSERT(kPointerSize == 4);
491 STATIC_ASSERT(kSmiTagSize == 1);
492 STATIC_ASSERT(kSmiTag == 0);
493 __ mov(ecx, FieldOperand(ecx, eax, times_half_pointer_size,
494 FixedArray::kHeaderSize));
495 Factory* factory = masm->isolate()->factory();
496 __ cmp(ecx, factory->undefined_value());
497 Label slow_case;
498 __ j(equal, &slow_case);
499
500 FastCloneShallowArrayStub::Mode mode = mode_;
501 // ecx is boilerplate object.
502 if (mode == CLONE_ANY_ELEMENTS) {
503 Label double_elements, check_fast_elements;
504 __ mov(ebx, FieldOperand(ecx, JSArray::kElementsOffset));
505 __ CheckMap(ebx, factory->fixed_cow_array_map(),
506 &check_fast_elements, DONT_DO_SMI_CHECK);
507 GenerateFastCloneShallowArrayCommon(masm, 0, COPY_ON_WRITE_ELEMENTS,
508 allocation_site_mode_,
509 &slow_case);
510 __ ret(3 * kPointerSize);
511
512 __ bind(&check_fast_elements);
513 __ CheckMap(ebx, factory->fixed_array_map(),
514 &double_elements, DONT_DO_SMI_CHECK);
515 GenerateFastCloneShallowArrayCommon(masm, length_, CLONE_ELEMENTS,
516 allocation_site_mode_,
517 &slow_case);
518 __ ret(3 * kPointerSize);
519
520 __ bind(&double_elements);
521 mode = CLONE_DOUBLE_ELEMENTS;
522 // Fall through to generate the code to handle double elements.
523 }
524
525 if (FLAG_debug_code) {
526 const char* message;
527 Handle<Map> expected_map;
528 if (mode == CLONE_ELEMENTS) {
529 message = "Expected (writable) fixed array";
530 expected_map = factory->fixed_array_map();
531 } else if (mode == CLONE_DOUBLE_ELEMENTS) {
532 message = "Expected (writable) fixed double array";
533 expected_map = factory->fixed_double_array_map();
534 } else {
535 ASSERT(mode == COPY_ON_WRITE_ELEMENTS);
536 message = "Expected copy-on-write fixed array";
537 expected_map = factory->fixed_cow_array_map();
538 }
539 __ push(ecx);
540 __ mov(ecx, FieldOperand(ecx, JSArray::kElementsOffset));
541 __ cmp(FieldOperand(ecx, HeapObject::kMapOffset), expected_map);
542 __ Assert(equal, message);
543 __ pop(ecx);
544 }
545
546 GenerateFastCloneShallowArrayCommon(masm, length_, mode,
547 allocation_site_mode_,
548 &slow_case);
549
550 // Return and remove the on-stack parameters.
551 __ ret(3 * kPointerSize);
552
553 __ bind(&slow_case);
554 __ TailCallRuntime(Runtime::kCreateArrayLiteralShallow, 3, 1);
555 }
556
557
558 // The stub expects its argument on the stack and returns its result in tos_: 408 // The stub expects its argument on the stack and returns its result in tos_:
559 // zero for false, and a non-zero value for true. 409 // zero for false, and a non-zero value for true.
560 void ToBooleanStub::Generate(MacroAssembler* masm) { 410 void ToBooleanStub::Generate(MacroAssembler* masm) {
561 // This stub overrides SometimesSetsUpAFrame() to return false. That means 411 // This stub overrides SometimesSetsUpAFrame() to return false. That means
562 // we cannot call anything that could cause a GC from this stub. 412 // we cannot call anything that could cause a GC from this stub.
563 Label patch; 413 Label patch;
564 Factory* factory = masm->isolate()->factory(); 414 Factory* factory = masm->isolate()->factory();
565 const Register argument = eax; 415 const Register argument = eax;
566 const Register map = edx; 416 const Register map = edx;
567 417
(...skipping 4501 matching lines...) Expand 10 before | Expand all | Expand 10 after
5069 4919
5070 bool CEntryStub::IsPregenerated() { 4920 bool CEntryStub::IsPregenerated() {
5071 return (!save_doubles_ || ISOLATE->fp_stubs_generated()) && 4921 return (!save_doubles_ || ISOLATE->fp_stubs_generated()) &&
5072 result_size_ == 1; 4922 result_size_ == 1;
5073 } 4923 }
5074 4924
5075 4925
5076 void CodeStub::GenerateStubsAheadOfTime(Isolate* isolate) { 4926 void CodeStub::GenerateStubsAheadOfTime(Isolate* isolate) {
5077 CEntryStub::GenerateAheadOfTime(isolate); 4927 CEntryStub::GenerateAheadOfTime(isolate);
5078 StoreBufferOverflowStub::GenerateFixedRegStubsAheadOfTime(isolate); 4928 StoreBufferOverflowStub::GenerateFixedRegStubsAheadOfTime(isolate);
4929 StubFailureTrampolineStub::GenerateAheadOfTime(isolate);
5079 // It is important that the store buffer overflow stubs are generated first. 4930 // It is important that the store buffer overflow stubs are generated first.
5080 RecordWriteStub::GenerateFixedRegStubsAheadOfTime(isolate); 4931 RecordWriteStub::GenerateFixedRegStubsAheadOfTime(isolate);
5081 } 4932 }
5082 4933
5083 4934
5084 void CodeStub::GenerateFPStubs(Isolate* isolate) { 4935 void CodeStub::GenerateFPStubs(Isolate* isolate) {
5085 if (CpuFeatures::IsSupported(SSE2)) { 4936 if (CpuFeatures::IsSupported(SSE2)) {
5086 CEntryStub save_doubles(1, kSaveFPRegs); 4937 CEntryStub save_doubles(1, kSaveFPRegs);
5087 // Stubs might already be in the snapshot, detect that and don't regenerate, 4938 // Stubs might already be in the snapshot, detect that and don't regenerate,
5088 // which would lead to code stub initialization state being messed up. 4939 // which would lead to code stub initialization state being messed up.
(...skipping 2690 matching lines...) Expand 10 before | Expand all | Expand 10 after
7779 edi, 7630 edi,
7780 xmm0, 7631 xmm0,
7781 &slow_elements_from_double, 7632 &slow_elements_from_double,
7782 false); 7633 false);
7783 __ pop(edx); 7634 __ pop(edx);
7784 __ ret(0); 7635 __ ret(0);
7785 } 7636 }
7786 7637
7787 7638
7788 void StubFailureTrampolineStub::Generate(MacroAssembler* masm) { 7639 void StubFailureTrampolineStub::Generate(MacroAssembler* masm) {
7789 ASSERT(!Serializer::enabled()); 7640 CEntryStub ces(1, fp_registers_ ? kSaveFPRegs : kDontSaveFPRegs);
7790 bool save_fp_regs = CpuFeatures::IsSupported(SSE2);
7791 CEntryStub ces(1, save_fp_regs ? kSaveFPRegs : kDontSaveFPRegs);
7792 __ call(ces.GetCode(masm->isolate()), RelocInfo::CODE_TARGET); 7641 __ call(ces.GetCode(masm->isolate()), RelocInfo::CODE_TARGET);
7793 int parameter_count_offset = 7642 int parameter_count_offset =
7794 StubFailureTrampolineFrame::kCallerStackParameterCountFrameOffset; 7643 StubFailureTrampolineFrame::kCallerStackParameterCountFrameOffset;
7795 __ mov(ebx, MemOperand(ebp, parameter_count_offset)); 7644 __ mov(ebx, MemOperand(ebp, parameter_count_offset));
7796 masm->LeaveFrame(StackFrame::STUB_FAILURE_TRAMPOLINE); 7645 masm->LeaveFrame(StackFrame::STUB_FAILURE_TRAMPOLINE);
7797 __ pop(ecx); 7646 __ pop(ecx);
7798 int additional_offset = function_mode_ == JS_FUNCTION_STUB_MODE 7647 int additional_offset = function_mode_ == JS_FUNCTION_STUB_MODE
7799 ? kPointerSize 7648 ? kPointerSize
7800 : 0; 7649 : 0;
7801 __ lea(esp, MemOperand(esp, ebx, times_pointer_size, additional_offset)); 7650 __ lea(esp, MemOperand(esp, ebx, times_pointer_size, additional_offset));
(...skipping 30 matching lines...) Expand all
7832 // Restore ecx. 7681 // Restore ecx.
7833 __ pop(ecx); 7682 __ pop(ecx);
7834 __ ret(0); 7683 __ ret(0);
7835 } 7684 }
7836 7685
7837 #undef __ 7686 #undef __
7838 7687
7839 } } // namespace v8::internal 7688 } } // namespace v8::internal
7840 7689
7841 #endif // V8_TARGET_ARCH_IA32 7690 #endif // V8_TARGET_ARCH_IA32
OLDNEW
« no previous file with comments | « src/hydrogen.cc ('k') | src/ia32/full-codegen-ia32.cc » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698