OLD | NEW |
1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 2122 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
2133 if (alignment != kObjectAlignment) { | 2133 if (alignment != kObjectAlignment) { |
2134 ASSERT(alignment == kDoubleAlignment); | 2134 ASSERT(alignment == kDoubleAlignment); |
2135 allocation_size += kPointerSize; | 2135 allocation_size += kPointerSize; |
2136 } | 2136 } |
2137 | 2137 |
2138 Heap* heap = map->GetHeap(); | 2138 Heap* heap = map->GetHeap(); |
2139 if (heap->ShouldBePromoted(object->address(), object_size)) { | 2139 if (heap->ShouldBePromoted(object->address(), object_size)) { |
2140 MaybeObject* maybe_result; | 2140 MaybeObject* maybe_result; |
2141 | 2141 |
2142 if (object_contents == DATA_OBJECT) { | 2142 if (object_contents == DATA_OBJECT) { |
2143 // TODO(mstarzinger): Turn this check into a regular assert soon! | 2143 ASSERT(heap->AllowedToBeMigrated(object, OLD_DATA_SPACE)); |
2144 CHECK(heap->AllowedToBeMigrated(object, OLD_DATA_SPACE)); | |
2145 maybe_result = heap->old_data_space()->AllocateRaw(allocation_size); | 2144 maybe_result = heap->old_data_space()->AllocateRaw(allocation_size); |
2146 } else { | 2145 } else { |
2147 // TODO(mstarzinger): Turn this check into a regular assert soon! | 2146 ASSERT(heap->AllowedToBeMigrated(object, OLD_POINTER_SPACE)); |
2148 CHECK(heap->AllowedToBeMigrated(object, OLD_POINTER_SPACE)); | |
2149 maybe_result = heap->old_pointer_space()->AllocateRaw(allocation_size); | 2147 maybe_result = heap->old_pointer_space()->AllocateRaw(allocation_size); |
2150 } | 2148 } |
2151 | 2149 |
2152 Object* result = NULL; // Initialization to please compiler. | 2150 Object* result = NULL; // Initialization to please compiler. |
2153 if (maybe_result->ToObject(&result)) { | 2151 if (maybe_result->ToObject(&result)) { |
2154 HeapObject* target = HeapObject::cast(result); | 2152 HeapObject* target = HeapObject::cast(result); |
2155 | 2153 |
2156 if (alignment != kObjectAlignment) { | 2154 if (alignment != kObjectAlignment) { |
2157 target = EnsureDoubleAligned(heap, target, allocation_size); | 2155 target = EnsureDoubleAligned(heap, target, allocation_size); |
2158 } | 2156 } |
(...skipping 10 matching lines...) Expand all Loading... |
2169 target, JSFunction::kNonWeakFieldsEndOffset); | 2167 target, JSFunction::kNonWeakFieldsEndOffset); |
2170 } else { | 2168 } else { |
2171 heap->promotion_queue()->insert(target, object_size); | 2169 heap->promotion_queue()->insert(target, object_size); |
2172 } | 2170 } |
2173 } | 2171 } |
2174 | 2172 |
2175 heap->tracer()->increment_promoted_objects_size(object_size); | 2173 heap->tracer()->increment_promoted_objects_size(object_size); |
2176 return; | 2174 return; |
2177 } | 2175 } |
2178 } | 2176 } |
2179 // TODO(mstarzinger): Turn this check into a regular assert soon! | 2177 ASSERT(heap->AllowedToBeMigrated(object, NEW_SPACE)); |
2180 CHECK(heap->AllowedToBeMigrated(object, NEW_SPACE)); | |
2181 MaybeObject* allocation = heap->new_space()->AllocateRaw(allocation_size); | 2178 MaybeObject* allocation = heap->new_space()->AllocateRaw(allocation_size); |
2182 heap->promotion_queue()->SetNewLimit(heap->new_space()->top()); | 2179 heap->promotion_queue()->SetNewLimit(heap->new_space()->top()); |
2183 Object* result = allocation->ToObjectUnchecked(); | 2180 Object* result = allocation->ToObjectUnchecked(); |
2184 HeapObject* target = HeapObject::cast(result); | 2181 HeapObject* target = HeapObject::cast(result); |
2185 | 2182 |
2186 if (alignment != kObjectAlignment) { | 2183 if (alignment != kObjectAlignment) { |
2187 target = EnsureDoubleAligned(heap, target, allocation_size); | 2184 target = EnsureDoubleAligned(heap, target, allocation_size); |
2188 } | 2185 } |
2189 | 2186 |
2190 // Order is important: slot might be inside of the target if target | 2187 // Order is important: slot might be inside of the target if target |
(...skipping 5683 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
7874 if (FLAG_concurrent_recompilation) { | 7871 if (FLAG_concurrent_recompilation) { |
7875 heap_->relocation_mutex_->Lock(); | 7872 heap_->relocation_mutex_->Lock(); |
7876 #ifdef DEBUG | 7873 #ifdef DEBUG |
7877 heap_->relocation_mutex_locked_by_optimizer_thread_ = | 7874 heap_->relocation_mutex_locked_by_optimizer_thread_ = |
7878 heap_->isolate()->optimizing_compiler_thread()->IsOptimizerThread(); | 7875 heap_->isolate()->optimizing_compiler_thread()->IsOptimizerThread(); |
7879 #endif // DEBUG | 7876 #endif // DEBUG |
7880 } | 7877 } |
7881 } | 7878 } |
7882 | 7879 |
7883 } } // namespace v8::internal | 7880 } } // namespace v8::internal |
OLD | NEW |