OLD | NEW |
---|---|
1 // Copyright 2013 the V8 project authors. All rights reserved. | 1 // Copyright 2013 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 9503 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
9514 set_code_no_write_barrier( | 9514 set_code_no_write_barrier( |
9515 GetIsolate()->builtins()->builtin(Builtins::kInOptimizationQueue)); | 9515 GetIsolate()->builtins()->builtin(Builtins::kInOptimizationQueue)); |
9516 // No write barrier required, since the builtin is part of the root set. | 9516 // No write barrier required, since the builtin is part of the root set. |
9517 } | 9517 } |
9518 | 9518 |
9519 | 9519 |
9520 void SharedFunctionInfo::AddToOptimizedCodeMap( | 9520 void SharedFunctionInfo::AddToOptimizedCodeMap( |
9521 Handle<SharedFunctionInfo> shared, | 9521 Handle<SharedFunctionInfo> shared, |
9522 Handle<Context> native_context, | 9522 Handle<Context> native_context, |
9523 Handle<Code> code, | 9523 Handle<Code> code, |
9524 Handle<FixedArray> literals) { | 9524 Handle<FixedArray> literals, |
9525 BailoutId osr_ast_id) { | |
9525 CALL_HEAP_FUNCTION_VOID( | 9526 CALL_HEAP_FUNCTION_VOID( |
9526 shared->GetIsolate(), | 9527 shared->GetIsolate(), |
9527 shared->AddToOptimizedCodeMap(*native_context, *code, *literals)); | 9528 shared->AddToOptimizedCodeMap( |
9529 *native_context, *code, *literals, osr_ast_id)); | |
9528 } | 9530 } |
9529 | 9531 |
9530 | 9532 |
9531 MaybeObject* SharedFunctionInfo::AddToOptimizedCodeMap(Context* native_context, | 9533 MaybeObject* SharedFunctionInfo::AddToOptimizedCodeMap(Context* native_context, |
9532 Code* code, | 9534 Code* code, |
9533 FixedArray* literals) { | 9535 FixedArray* literals, |
9536 BailoutId osr_ast_id) { | |
9534 ASSERT(code->kind() == Code::OPTIMIZED_FUNCTION); | 9537 ASSERT(code->kind() == Code::OPTIMIZED_FUNCTION); |
9535 ASSERT(native_context->IsNativeContext()); | 9538 ASSERT(native_context->IsNativeContext()); |
9536 STATIC_ASSERT(kEntryLength == 3); | 9539 STATIC_ASSERT(kEntryLength == 4); |
9537 Heap* heap = GetHeap(); | 9540 Heap* heap = GetHeap(); |
9538 FixedArray* new_code_map; | 9541 FixedArray* new_code_map; |
9539 Object* value = optimized_code_map(); | 9542 Object* value = optimized_code_map(); |
9543 Smi* osr_ast_id_smi = Smi::FromInt(osr_ast_id.ToInt()); | |
9540 if (value->IsSmi()) { | 9544 if (value->IsSmi()) { |
9541 // No optimized code map. | 9545 // No optimized code map. |
9542 ASSERT_EQ(0, Smi::cast(value)->value()); | 9546 ASSERT_EQ(0, Smi::cast(value)->value()); |
9543 // Create 3 entries per context {context, code, literals}. | 9547 // Create 3 entries per context {context, code, literals}. |
9544 MaybeObject* maybe = heap->AllocateFixedArray(kInitialLength); | 9548 MaybeObject* maybe = heap->AllocateFixedArray(kInitialLength); |
9545 if (!maybe->To(&new_code_map)) return maybe; | 9549 if (!maybe->To(&new_code_map)) return maybe; |
9546 new_code_map->set(kEntriesStart + 0, native_context); | 9550 new_code_map->set(kEntriesStart + kContextOffset, native_context); |
9547 new_code_map->set(kEntriesStart + 1, code); | 9551 new_code_map->set(kEntriesStart + kCachedCodeOffset, code); |
9548 new_code_map->set(kEntriesStart + 2, literals); | 9552 new_code_map->set(kEntriesStart + kLiteralsOffset, literals); |
9553 new_code_map->set(kEntriesStart + kOsrAstIdOffset, osr_ast_id_smi); | |
9549 } else { | 9554 } else { |
9550 // Copy old map and append one new entry. | 9555 // Copy old map and append one new entry. |
9551 FixedArray* old_code_map = FixedArray::cast(value); | 9556 FixedArray* old_code_map = FixedArray::cast(value); |
9552 ASSERT_EQ(-1, SearchOptimizedCodeMap(native_context)); | 9557 ASSERT_EQ(-1, SearchOptimizedCodeMap(native_context, osr_ast_id)); |
9553 int old_length = old_code_map->length(); | 9558 int old_length = old_code_map->length(); |
9554 int new_length = old_length + kEntryLength; | 9559 int new_length = old_length + kEntryLength; |
9555 MaybeObject* maybe = old_code_map->CopySize(new_length); | 9560 MaybeObject* maybe = old_code_map->CopySize(new_length); |
9556 if (!maybe->To(&new_code_map)) return maybe; | 9561 if (!maybe->To(&new_code_map)) return maybe; |
9557 new_code_map->set(old_length + 0, native_context); | 9562 new_code_map->set(old_length + kContextOffset, native_context); |
9558 new_code_map->set(old_length + 1, code); | 9563 new_code_map->set(old_length + kCachedCodeOffset, code); |
9559 new_code_map->set(old_length + 2, literals); | 9564 new_code_map->set(old_length + kLiteralsOffset, literals); |
9565 new_code_map->set(old_length + kOsrAstIdOffset, osr_ast_id_smi); | |
9560 // Zap the old map for the sake of the heap verifier. | 9566 // Zap the old map for the sake of the heap verifier. |
9561 if (Heap::ShouldZapGarbage()) { | 9567 if (Heap::ShouldZapGarbage()) { |
9562 Object** data = old_code_map->data_start(); | 9568 Object** data = old_code_map->data_start(); |
9563 MemsetPointer(data, heap->the_hole_value(), old_length); | 9569 MemsetPointer(data, heap->the_hole_value(), old_length); |
9564 } | 9570 } |
9565 } | 9571 } |
9566 #ifdef DEBUG | 9572 #ifdef DEBUG |
9567 for (int i = kEntriesStart; i < new_code_map->length(); i += kEntryLength) { | 9573 for (int i = kEntriesStart; i < new_code_map->length(); i += kEntryLength) { |
9568 ASSERT(new_code_map->get(i)->IsNativeContext()); | 9574 ASSERT(new_code_map->get(i + kContextOffset)->IsNativeContext()); |
9569 ASSERT(new_code_map->get(i + 1)->IsCode()); | 9575 ASSERT(new_code_map->get(i + kCachedCodeOffset)->IsCode()); |
9570 ASSERT(Code::cast(new_code_map->get(i + 1))->kind() == | 9576 ASSERT(Code::cast(new_code_map->get(i + kCachedCodeOffset))->kind() == |
9571 Code::OPTIMIZED_FUNCTION); | 9577 Code::OPTIMIZED_FUNCTION); |
9572 ASSERT(new_code_map->get(i + 2)->IsFixedArray()); | 9578 ASSERT(new_code_map->get(i + kLiteralsOffset)->IsFixedArray()); |
9579 ASSERT(new_code_map->get(i + kOsrAstIdOffset)->IsSmi()); | |
9573 } | 9580 } |
9574 #endif | 9581 #endif |
9575 set_optimized_code_map(new_code_map); | 9582 set_optimized_code_map(new_code_map); |
9576 return new_code_map; | 9583 return new_code_map; |
9577 } | 9584 } |
9578 | 9585 |
9579 | 9586 |
9580 FixedArray* SharedFunctionInfo::GetLiteralsFromOptimizedCodeMap(int index) { | 9587 FixedArray* SharedFunctionInfo::GetLiteralsFromOptimizedCodeMap(int index) { |
9581 ASSERT(index > kEntriesStart); | 9588 ASSERT(index > kEntriesStart); |
9582 FixedArray* code_map = FixedArray::cast(optimized_code_map()); | 9589 FixedArray* code_map = FixedArray::cast(optimized_code_map()); |
9583 if (!bound()) { | 9590 if (!bound()) { |
9584 FixedArray* cached_literals = FixedArray::cast(code_map->get(index + 1)); | 9591 FixedArray* cached_literals = FixedArray::cast(code_map->get(index + 1)); |
9585 ASSERT_NE(NULL, cached_literals); | 9592 ASSERT_NE(NULL, cached_literals); |
9586 return cached_literals; | 9593 return cached_literals; |
9587 } | 9594 } |
9588 return NULL; | 9595 return NULL; |
9589 } | 9596 } |
9590 | 9597 |
9591 | 9598 |
9592 | |
9593 Code* SharedFunctionInfo::GetCodeFromOptimizedCodeMap(int index) { | 9599 Code* SharedFunctionInfo::GetCodeFromOptimizedCodeMap(int index) { |
9594 ASSERT(index > kEntriesStart); | 9600 ASSERT(index > kEntriesStart); |
9595 FixedArray* code_map = FixedArray::cast(optimized_code_map()); | 9601 FixedArray* code_map = FixedArray::cast(optimized_code_map()); |
9596 Code* code = Code::cast(code_map->get(index)); | 9602 Code* code = Code::cast(code_map->get(index)); |
9597 ASSERT_NE(NULL, code); | 9603 ASSERT_NE(NULL, code); |
9598 return code; | 9604 return code; |
9599 } | 9605 } |
9600 | 9606 |
9601 | 9607 |
9602 void SharedFunctionInfo::ClearOptimizedCodeMap() { | 9608 void SharedFunctionInfo::ClearOptimizedCodeMap() { |
(...skipping 24 matching lines...) Expand all Loading... | |
9627 if (FLAG_trace_opt) { | 9633 if (FLAG_trace_opt) { |
9628 PrintF("[evicting entry from optimizing code map (%s) for ", reason); | 9634 PrintF("[evicting entry from optimizing code map (%s) for ", reason); |
9629 ShortPrint(); | 9635 ShortPrint(); |
9630 PrintF("]\n"); | 9636 PrintF("]\n"); |
9631 } | 9637 } |
9632 removed_entry = true; | 9638 removed_entry = true; |
9633 break; | 9639 break; |
9634 } | 9640 } |
9635 } | 9641 } |
9636 while (i < (code_map->length() - kEntryLength)) { | 9642 while (i < (code_map->length() - kEntryLength)) { |
9637 code_map->set(i, code_map->get(i + kEntryLength)); | 9643 code_map->set(i + kContextOffset, |
titzer
2013/12/16 11:59:34
Why not an inner loop over entry length somehow?
| |
9638 code_map->set(i + 1, code_map->get(i + 1 + kEntryLength)); | 9644 code_map->get(i + kContextOffset + kEntryLength)); |
9639 code_map->set(i + 2, code_map->get(i + 2 + kEntryLength)); | 9645 code_map->set(i + kCodeOffset, |
9646 code_map->get(i + kCodeOffset + kEntryLength)); | |
9647 code_map->set(i + kLiteralsOffset, | |
9648 code_map->get(i + kLiteralsOffset + kEntryLength)); | |
9649 code_map->set(i + kOsrAstIdOffset, | |
9650 code_map->get(i + kOsrAstIdOffset + kEntryLength)); | |
9640 i += kEntryLength; | 9651 i += kEntryLength; |
9641 } | 9652 } |
9642 if (removed_entry) { | 9653 if (removed_entry) { |
9643 // Always trim even when array is cleared because of heap verifier. | 9654 // Always trim even when array is cleared because of heap verifier. |
9644 RightTrimFixedArray<FROM_MUTATOR>(GetHeap(), code_map, kEntryLength); | 9655 RightTrimFixedArray<FROM_MUTATOR>(GetHeap(), code_map, kEntryLength); |
9645 if (code_map->length() == kEntriesStart) { | 9656 if (code_map->length() == kEntriesStart) { |
9646 ClearOptimizedCodeMap(); | 9657 ClearOptimizedCodeMap(); |
9647 } | 9658 } |
9648 } | 9659 } |
9649 } | 9660 } |
(...skipping 541 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
10191 // Resize the initial map and all maps in its transition tree. | 10202 // Resize the initial map and all maps in its transition tree. |
10192 map->TraverseTransitionTree(&ShrinkInstanceSize, &slack); | 10203 map->TraverseTransitionTree(&ShrinkInstanceSize, &slack); |
10193 | 10204 |
10194 // Give the correct expected_nof_properties to initial maps created later. | 10205 // Give the correct expected_nof_properties to initial maps created later. |
10195 ASSERT(expected_nof_properties() >= slack); | 10206 ASSERT(expected_nof_properties() >= slack); |
10196 set_expected_nof_properties(expected_nof_properties() - slack); | 10207 set_expected_nof_properties(expected_nof_properties() - slack); |
10197 } | 10208 } |
10198 } | 10209 } |
10199 | 10210 |
10200 | 10211 |
10201 int SharedFunctionInfo::SearchOptimizedCodeMap(Context* native_context) { | 10212 int SharedFunctionInfo::SearchOptimizedCodeMap(Context* native_context, |
10213 BailoutId osr_ast_id) { | |
10202 ASSERT(native_context->IsNativeContext()); | 10214 ASSERT(native_context->IsNativeContext()); |
10203 if (!FLAG_cache_optimized_code) return -1; | 10215 if (!FLAG_cache_optimized_code) return -1; |
10204 Object* value = optimized_code_map(); | 10216 Object* value = optimized_code_map(); |
10205 if (!value->IsSmi()) { | 10217 if (!value->IsSmi()) { |
10206 FixedArray* optimized_code_map = FixedArray::cast(value); | 10218 FixedArray* optimized_code_map = FixedArray::cast(value); |
10207 int length = optimized_code_map->length(); | 10219 int length = optimized_code_map->length(); |
10220 Smi* osr_ast_id_smi = Smi::FromInt(osr_ast_id.ToInt()); | |
10208 for (int i = kEntriesStart; i < length; i += kEntryLength) { | 10221 for (int i = kEntriesStart; i < length; i += kEntryLength) { |
10209 if (optimized_code_map->get(i) == native_context) { | 10222 if (optimized_code_map->get(i + kContextOffset) == native_context && |
10210 return i + 1; | 10223 optimized_code_map->get(i + kOsrAstIdOffset) == osr_ast_id_smi) { |
10224 return i + kCachedCodeOffset; | |
10211 } | 10225 } |
10212 } | 10226 } |
10213 if (FLAG_trace_opt) { | 10227 if (FLAG_trace_opt) { |
10214 PrintF("[didn't find optimized code in optimized code map for "); | 10228 PrintF("[didn't find optimized code in optimized code map for "); |
10215 ShortPrint(); | 10229 ShortPrint(); |
10216 PrintF("]\n"); | 10230 PrintF("]\n"); |
10217 } | 10231 } |
10218 } | 10232 } |
10219 return -1; | 10233 return -1; |
10220 } | 10234 } |
(...skipping 6370 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
16591 #define ERROR_MESSAGES_TEXTS(C, T) T, | 16605 #define ERROR_MESSAGES_TEXTS(C, T) T, |
16592 static const char* error_messages_[] = { | 16606 static const char* error_messages_[] = { |
16593 ERROR_MESSAGES_LIST(ERROR_MESSAGES_TEXTS) | 16607 ERROR_MESSAGES_LIST(ERROR_MESSAGES_TEXTS) |
16594 }; | 16608 }; |
16595 #undef ERROR_MESSAGES_TEXTS | 16609 #undef ERROR_MESSAGES_TEXTS |
16596 return error_messages_[reason]; | 16610 return error_messages_[reason]; |
16597 } | 16611 } |
16598 | 16612 |
16599 | 16613 |
16600 } } // namespace v8::internal | 16614 } } // namespace v8::internal |
OLD | NEW |