OLD | NEW |
1 // Copyright 2013 the V8 project authors. All rights reserved. | 1 // Copyright 2013 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 2701 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
2712 for (; descriptor < descriptors; descriptor++) { | 2712 for (; descriptor < descriptors; descriptor++) { |
2713 MaybeObject* maybe_map = new_map->CopyInstallDescriptors( | 2713 MaybeObject* maybe_map = new_map->CopyInstallDescriptors( |
2714 descriptor, new_descriptors); | 2714 descriptor, new_descriptors); |
2715 if (!maybe_map->To(&new_map)) { | 2715 if (!maybe_map->To(&new_map)) { |
2716 // Create a handle for the last created map to ensure it stays alive | 2716 // Create a handle for the last created map to ensure it stays alive |
2717 // during GC. Its descriptor array is too large, but it will be | 2717 // during GC. Its descriptor array is too large, but it will be |
2718 // overwritten during retry anyway. | 2718 // overwritten during retry anyway. |
2719 Handle<Map>(new_map); | 2719 Handle<Map>(new_map); |
2720 return maybe_map; | 2720 return maybe_map; |
2721 } | 2721 } |
2722 new_map->set_migration_target(true); | |
2723 } | 2722 } |
2724 | 2723 |
2725 new_map->set_owns_descriptors(true); | 2724 new_map->set_owns_descriptors(true); |
2726 return new_map; | 2725 return new_map; |
2727 } | 2726 } |
2728 | 2727 |
2729 | 2728 |
2730 Map* Map::CurrentMapForDeprecated() { | 2729 Map* Map::CurrentMapForDeprecated() { |
2731 DisallowHeapAllocation no_allocation; | 2730 DisallowHeapAllocation no_allocation; |
2732 if (!is_deprecated()) return this; | 2731 if (!is_deprecated()) return this; |
(...skipping 3778 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
6511 Map* result; | 6510 Map* result; |
6512 MaybeObject* maybe_result = RawCopy(new_instance_size); | 6511 MaybeObject* maybe_result = RawCopy(new_instance_size); |
6513 if (!maybe_result->To(&result)) return maybe_result; | 6512 if (!maybe_result->To(&result)) return maybe_result; |
6514 | 6513 |
6515 if (mode != CLEAR_INOBJECT_PROPERTIES) { | 6514 if (mode != CLEAR_INOBJECT_PROPERTIES) { |
6516 result->set_inobject_properties(inobject_properties()); | 6515 result->set_inobject_properties(inobject_properties()); |
6517 } | 6516 } |
6518 | 6517 |
6519 result->set_is_shared(sharing == SHARED_NORMALIZED_MAP); | 6518 result->set_is_shared(sharing == SHARED_NORMALIZED_MAP); |
6520 result->set_dictionary_map(true); | 6519 result->set_dictionary_map(true); |
6521 result->set_migration_target(false); | |
6522 | 6520 |
6523 #ifdef VERIFY_HEAP | 6521 #ifdef VERIFY_HEAP |
6524 if (FLAG_verify_heap && result->is_shared()) { | 6522 if (FLAG_verify_heap && result->is_shared()) { |
6525 result->SharedMapVerify(); | 6523 result->SharedMapVerify(); |
6526 } | 6524 } |
6527 #endif | 6525 #endif |
6528 | 6526 |
6529 return result; | 6527 return result; |
6530 } | 6528 } |
6531 | 6529 |
(...skipping 3267 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
9799 // TODO(3025757): In case the recompiled isn't equivalent to the | 9797 // TODO(3025757): In case the recompiled isn't equivalent to the |
9800 // old code, we have to replace it. We should try to avoid this | 9798 // old code, we have to replace it. We should try to avoid this |
9801 // altogether because it flushes valuable type feedback by | 9799 // altogether because it flushes valuable type feedback by |
9802 // effectively resetting all IC state. | 9800 // effectively resetting all IC state. |
9803 ReplaceCode(recompiled); | 9801 ReplaceCode(recompiled); |
9804 } | 9802 } |
9805 ASSERT(has_deoptimization_support()); | 9803 ASSERT(has_deoptimization_support()); |
9806 } | 9804 } |
9807 | 9805 |
9808 | 9806 |
9809 void SharedFunctionInfo::DisableOptimization(BailoutReason reason) { | 9807 void SharedFunctionInfo::DisableOptimization(const char* reason) { |
9810 // Disable optimization for the shared function info and mark the | 9808 // Disable optimization for the shared function info and mark the |
9811 // code as non-optimizable. The marker on the shared function info | 9809 // code as non-optimizable. The marker on the shared function info |
9812 // is there because we flush non-optimized code thereby loosing the | 9810 // is there because we flush non-optimized code thereby loosing the |
9813 // non-optimizable information for the code. When the code is | 9811 // non-optimizable information for the code. When the code is |
9814 // regenerated and set on the shared function info it is marked as | 9812 // regenerated and set on the shared function info it is marked as |
9815 // non-optimizable if optimization is disabled for the shared | 9813 // non-optimizable if optimization is disabled for the shared |
9816 // function info. | 9814 // function info. |
9817 set_optimization_disabled(true); | 9815 set_optimization_disabled(true); |
9818 // Code should be the lazy compilation stub or else unoptimized. If the | 9816 // Code should be the lazy compilation stub or else unoptimized. If the |
9819 // latter, disable optimization for the code too. | 9817 // latter, disable optimization for the code too. |
9820 ASSERT(code()->kind() == Code::FUNCTION || code()->kind() == Code::BUILTIN); | 9818 ASSERT(code()->kind() == Code::FUNCTION || code()->kind() == Code::BUILTIN); |
9821 if (code()->kind() == Code::FUNCTION) { | 9819 if (code()->kind() == Code::FUNCTION) { |
9822 code()->set_optimizable(false); | 9820 code()->set_optimizable(false); |
9823 } | 9821 } |
9824 if (FLAG_trace_opt) { | 9822 if (FLAG_trace_opt) { |
9825 PrintF("[disabled optimization for "); | 9823 PrintF("[disabled optimization for "); |
9826 ShortPrint(); | 9824 ShortPrint(); |
9827 PrintF(", reason: %s]\n", GetBailoutReason(reason)); | 9825 PrintF(", reason: %s]\n", reason); |
9828 } | 9826 } |
9829 } | 9827 } |
9830 | 9828 |
9831 | 9829 |
9832 bool SharedFunctionInfo::VerifyBailoutId(BailoutId id) { | 9830 bool SharedFunctionInfo::VerifyBailoutId(BailoutId id) { |
9833 ASSERT(!id.IsNone()); | 9831 ASSERT(!id.IsNone()); |
9834 Code* unoptimized = code(); | 9832 Code* unoptimized = code(); |
9835 DeoptimizationOutputData* data = | 9833 DeoptimizationOutputData* data = |
9836 DeoptimizationOutputData::cast(unoptimized->deoptimization_data()); | 9834 DeoptimizationOutputData::cast(unoptimized->deoptimization_data()); |
9837 unsigned ignore = Deoptimizer::GetOutputInfo(data, id, this); | 9835 unsigned ignore = Deoptimizer::GetOutputInfo(data, id, this); |
(...skipping 950 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
10788 PrintF(out, " argc: %d", entry.argument_count()); | 10786 PrintF(out, " argc: %d", entry.argument_count()); |
10789 } | 10787 } |
10790 PrintF(out, "\n"); | 10788 PrintF(out, "\n"); |
10791 } | 10789 } |
10792 PrintF(out, "\n"); | 10790 PrintF(out, "\n"); |
10793 } else if (kind() == FUNCTION) { | 10791 } else if (kind() == FUNCTION) { |
10794 unsigned offset = back_edge_table_offset(); | 10792 unsigned offset = back_edge_table_offset(); |
10795 // If there is no back edge table, the "table start" will be at or after | 10793 // If there is no back edge table, the "table start" will be at or after |
10796 // (due to alignment) the end of the instruction stream. | 10794 // (due to alignment) the end of the instruction stream. |
10797 if (static_cast<int>(offset) < instruction_size()) { | 10795 if (static_cast<int>(offset) < instruction_size()) { |
10798 FullCodeGenerator::BackEdgeTableIterator back_edges(this); | 10796 Address back_edge_cursor = instruction_start() + offset; |
10799 | 10797 uint32_t table_length = Memory::uint32_at(back_edge_cursor); |
10800 PrintF(out, "Back edges (size = %u)\n", back_edges.table_length()); | 10798 PrintF(out, "Back edges (size = %u)\n", table_length); |
10801 PrintF(out, "ast_id pc_offset loop_depth\n"); | 10799 PrintF(out, "ast_id pc_offset loop_depth\n"); |
10802 | 10800 for (uint32_t i = 0; i < table_length; ++i) { |
10803 for ( ; !back_edges.Done(); back_edges.Next()) { | 10801 uint32_t ast_id = Memory::uint32_at(back_edge_cursor); |
10804 PrintF(out, "%6d %9u %10u\n", back_edges.ast_id().ToInt(), | 10802 uint32_t pc_offset = Memory::uint32_at(back_edge_cursor + kIntSize); |
10805 back_edges.pc_offset(), | 10803 uint32_t loop_depth = Memory::uint32_at(back_edge_cursor + |
10806 back_edges.loop_depth()); | 10804 2 * kIntSize); |
| 10805 PrintF(out, "%6u %9u %10u\n", ast_id, pc_offset, loop_depth); |
| 10806 back_edge_cursor += FullCodeGenerator::kBackEdgeEntrySize; |
10807 } | 10807 } |
10808 | |
10809 PrintF(out, "\n"); | 10808 PrintF(out, "\n"); |
10810 } | 10809 } |
10811 #ifdef OBJECT_PRINT | 10810 #ifdef OBJECT_PRINT |
10812 if (!type_feedback_info()->IsUndefined()) { | 10811 if (!type_feedback_info()->IsUndefined()) { |
10813 TypeFeedbackInfo::cast(type_feedback_info())->TypeFeedbackInfoPrint(out); | 10812 TypeFeedbackInfo::cast(type_feedback_info())->TypeFeedbackInfoPrint(out); |
10814 PrintF(out, "\n"); | 10813 PrintF(out, "\n"); |
10815 } | 10814 } |
10816 #endif | 10815 #endif |
10817 } | 10816 } |
10818 | 10817 |
(...skipping 5139 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
15958 | 15957 |
15959 | 15958 |
15960 void PropertyCell::AddDependentCode(Handle<Code> code) { | 15959 void PropertyCell::AddDependentCode(Handle<Code> code) { |
15961 Handle<DependentCode> codes = DependentCode::Insert( | 15960 Handle<DependentCode> codes = DependentCode::Insert( |
15962 Handle<DependentCode>(dependent_code()), | 15961 Handle<DependentCode>(dependent_code()), |
15963 DependentCode::kPropertyCellChangedGroup, code); | 15962 DependentCode::kPropertyCellChangedGroup, code); |
15964 if (*codes != dependent_code()) set_dependent_code(*codes); | 15963 if (*codes != dependent_code()) set_dependent_code(*codes); |
15965 } | 15964 } |
15966 | 15965 |
15967 | 15966 |
15968 const char* GetBailoutReason(BailoutReason reason) { | |
15969 ASSERT(reason < kLastErrorMessage); | |
15970 #define ERROR_MESSAGES_TEXTS(C, T) T, | |
15971 static const char* error_messages_[] = { | |
15972 ERROR_MESSAGES_LIST(ERROR_MESSAGES_TEXTS) | |
15973 }; | |
15974 #undef ERROR_MESSAGES_TEXTS | |
15975 return error_messages_[reason]; | |
15976 } | |
15977 | |
15978 | |
15979 } } // namespace v8::internal | 15967 } } // namespace v8::internal |
OLD | NEW |