| OLD | NEW |
| 1 // Copyright 2006-2008 the V8 project authors. All rights reserved. | 1 // Copyright 2006-2008 the V8 project authors. All rights reserved. |
| 2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
| 3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
| 4 // met: | 4 // met: |
| 5 // | 5 // |
| 6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
| 7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
| 8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
| 9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
| 10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
| (...skipping 22 matching lines...) Expand all Loading... |
| 33 #include "v8-counters.h" | 33 #include "v8-counters.h" |
| 34 | 34 |
| 35 namespace v8 { | 35 namespace v8 { |
| 36 namespace internal { | 36 namespace internal { |
| 37 | 37 |
| 38 int Heap::MaxObjectSizeInPagedSpace() { | 38 int Heap::MaxObjectSizeInPagedSpace() { |
| 39 return Page::kMaxHeapObjectSize; | 39 return Page::kMaxHeapObjectSize; |
| 40 } | 40 } |
| 41 | 41 |
| 42 | 42 |
| 43 Object* Heap::AllocateSymbol(Vector<const char> str, | 43 MaybeObject* Heap::AllocateSymbol(Vector<const char> str, |
| 44 int chars, | 44 int chars, |
| 45 uint32_t hash_field) { | 45 uint32_t hash_field) { |
| 46 unibrow::Utf8InputBuffer<> buffer(str.start(), | 46 unibrow::Utf8InputBuffer<> buffer(str.start(), |
| 47 static_cast<unsigned>(str.length())); | 47 static_cast<unsigned>(str.length())); |
| 48 return AllocateInternalSymbol(&buffer, chars, hash_field); | 48 return AllocateInternalSymbol(&buffer, chars, hash_field); |
| 49 } | 49 } |
| 50 | 50 |
| 51 | 51 |
| 52 Object* Heap::CopyFixedArray(FixedArray* src) { | 52 MaybeObject* Heap::CopyFixedArray(FixedArray* src) { |
| 53 return CopyFixedArrayWithMap(src, src->map()); | 53 return CopyFixedArrayWithMap(src, src->map()); |
| 54 } | 54 } |
| 55 | 55 |
| 56 | 56 |
| 57 Object* Heap::AllocateRaw(int size_in_bytes, | 57 MaybeObject* Heap::AllocateRaw(int size_in_bytes, |
| 58 AllocationSpace space, | 58 AllocationSpace space, |
| 59 AllocationSpace retry_space) { | 59 AllocationSpace retry_space) { |
| 60 ASSERT(allocation_allowed_ && gc_state_ == NOT_IN_GC); | 60 ASSERT(allocation_allowed_ && gc_state_ == NOT_IN_GC); |
| 61 ASSERT(space != NEW_SPACE || | 61 ASSERT(space != NEW_SPACE || |
| 62 retry_space == OLD_POINTER_SPACE || | 62 retry_space == OLD_POINTER_SPACE || |
| 63 retry_space == OLD_DATA_SPACE || | 63 retry_space == OLD_DATA_SPACE || |
| 64 retry_space == LO_SPACE); | 64 retry_space == LO_SPACE); |
| 65 #ifdef DEBUG | 65 #ifdef DEBUG |
| 66 if (FLAG_gc_interval >= 0 && | 66 if (FLAG_gc_interval >= 0 && |
| 67 !disallow_allocation_failure_ && | 67 !disallow_allocation_failure_ && |
| 68 Heap::allocation_timeout_-- <= 0) { | 68 Heap::allocation_timeout_-- <= 0) { |
| 69 return Failure::RetryAfterGC(space); | 69 return Failure::RetryAfterGC(space); |
| 70 } | 70 } |
| 71 Counters::objs_since_last_full.Increment(); | 71 Counters::objs_since_last_full.Increment(); |
| 72 Counters::objs_since_last_young.Increment(); | 72 Counters::objs_since_last_young.Increment(); |
| 73 #endif | 73 #endif |
| 74 Object* result; | 74 MaybeObject* result; |
| 75 if (NEW_SPACE == space) { | 75 if (NEW_SPACE == space) { |
| 76 result = new_space_.AllocateRaw(size_in_bytes); | 76 result = new_space_.AllocateRaw(size_in_bytes); |
| 77 if (always_allocate() && result->IsFailure()) { | 77 if (always_allocate() && result->IsFailure()) { |
| 78 space = retry_space; | 78 space = retry_space; |
| 79 } else { | 79 } else { |
| 80 return result; | 80 return result; |
| 81 } | 81 } |
| 82 } | 82 } |
| 83 | 83 |
| 84 if (OLD_POINTER_SPACE == space) { | 84 if (OLD_POINTER_SPACE == space) { |
| 85 result = old_pointer_space_->AllocateRaw(size_in_bytes); | 85 result = old_pointer_space_->AllocateRaw(size_in_bytes); |
| 86 } else if (OLD_DATA_SPACE == space) { | 86 } else if (OLD_DATA_SPACE == space) { |
| 87 result = old_data_space_->AllocateRaw(size_in_bytes); | 87 result = old_data_space_->AllocateRaw(size_in_bytes); |
| 88 } else if (CODE_SPACE == space) { | 88 } else if (CODE_SPACE == space) { |
| 89 result = code_space_->AllocateRaw(size_in_bytes); | 89 result = code_space_->AllocateRaw(size_in_bytes); |
| 90 } else if (LO_SPACE == space) { | 90 } else if (LO_SPACE == space) { |
| 91 result = lo_space_->AllocateRaw(size_in_bytes); | 91 result = lo_space_->AllocateRaw(size_in_bytes); |
| 92 } else if (CELL_SPACE == space) { | 92 } else if (CELL_SPACE == space) { |
| 93 result = cell_space_->AllocateRaw(size_in_bytes); | 93 result = cell_space_->AllocateRaw(size_in_bytes); |
| 94 } else { | 94 } else { |
| 95 ASSERT(MAP_SPACE == space); | 95 ASSERT(MAP_SPACE == space); |
| 96 result = map_space_->AllocateRaw(size_in_bytes); | 96 result = map_space_->AllocateRaw(size_in_bytes); |
| 97 } | 97 } |
| 98 if (result->IsFailure()) old_gen_exhausted_ = true; | 98 if (result->IsFailure()) old_gen_exhausted_ = true; |
| 99 return result; | 99 return result; |
| 100 } | 100 } |
| 101 | 101 |
| 102 | 102 |
| 103 Object* Heap::NumberFromInt32(int32_t value) { | 103 MaybeObject* Heap::NumberFromInt32(int32_t value) { |
| 104 if (Smi::IsValid(value)) return Smi::FromInt(value); | 104 if (Smi::IsValid(value)) return Smi::FromInt(value); |
| 105 // Bypass NumberFromDouble to avoid various redundant checks. | 105 // Bypass NumberFromDouble to avoid various redundant checks. |
| 106 return AllocateHeapNumber(FastI2D(value)); | 106 return AllocateHeapNumber(FastI2D(value)); |
| 107 } | 107 } |
| 108 | 108 |
| 109 | 109 |
| 110 Object* Heap::NumberFromUint32(uint32_t value) { | 110 MaybeObject* Heap::NumberFromUint32(uint32_t value) { |
| 111 if ((int32_t)value >= 0 && Smi::IsValid((int32_t)value)) { | 111 if ((int32_t)value >= 0 && Smi::IsValid((int32_t)value)) { |
| 112 return Smi::FromInt((int32_t)value); | 112 return Smi::FromInt((int32_t)value); |
| 113 } | 113 } |
| 114 // Bypass NumberFromDouble to avoid various redundant checks. | 114 // Bypass NumberFromDouble to avoid various redundant checks. |
| 115 return AllocateHeapNumber(FastUI2D(value)); | 115 return AllocateHeapNumber(FastUI2D(value)); |
| 116 } | 116 } |
| 117 | 117 |
| 118 | 118 |
| 119 void Heap::FinalizeExternalString(String* string) { | 119 void Heap::FinalizeExternalString(String* string) { |
| 120 ASSERT(string->IsExternalString()); | 120 ASSERT(string->IsExternalString()); |
| 121 v8::String::ExternalStringResourceBase** resource_addr = | 121 v8::String::ExternalStringResourceBase** resource_addr = |
| 122 reinterpret_cast<v8::String::ExternalStringResourceBase**>( | 122 reinterpret_cast<v8::String::ExternalStringResourceBase**>( |
| 123 reinterpret_cast<byte*>(string) + | 123 reinterpret_cast<byte*>(string) + |
| 124 ExternalString::kResourceOffset - | 124 ExternalString::kResourceOffset - |
| 125 kHeapObjectTag); | 125 kHeapObjectTag); |
| 126 | 126 |
| 127 // Dispose of the C++ object if it has not already been disposed. | 127 // Dispose of the C++ object if it has not already been disposed. |
| 128 if (*resource_addr != NULL) { | 128 if (*resource_addr != NULL) { |
| 129 (*resource_addr)->Dispose(); | 129 (*resource_addr)->Dispose(); |
| 130 } | 130 } |
| 131 | 131 |
| 132 // Clear the resource pointer in the string. | 132 // Clear the resource pointer in the string. |
| 133 *resource_addr = NULL; | 133 *resource_addr = NULL; |
| 134 } | 134 } |
| 135 | 135 |
| 136 | 136 |
| 137 Object* Heap::AllocateRawMap() { | 137 MaybeObject* Heap::AllocateRawMap() { |
| 138 #ifdef DEBUG | 138 #ifdef DEBUG |
| 139 Counters::objs_since_last_full.Increment(); | 139 Counters::objs_since_last_full.Increment(); |
| 140 Counters::objs_since_last_young.Increment(); | 140 Counters::objs_since_last_young.Increment(); |
| 141 #endif | 141 #endif |
| 142 Object* result = map_space_->AllocateRaw(Map::kSize); | 142 MaybeObject* result = map_space_->AllocateRaw(Map::kSize); |
| 143 if (result->IsFailure()) old_gen_exhausted_ = true; | 143 if (result->IsFailure()) old_gen_exhausted_ = true; |
| 144 #ifdef DEBUG | 144 #ifdef DEBUG |
| 145 if (!result->IsFailure()) { | 145 if (!result->IsFailure()) { |
| 146 // Maps have their own alignment. | 146 // Maps have their own alignment. |
| 147 CHECK((reinterpret_cast<intptr_t>(result) & kMapAlignmentMask) == | 147 CHECK((reinterpret_cast<intptr_t>(result) & kMapAlignmentMask) == |
| 148 static_cast<intptr_t>(kHeapObjectTag)); | 148 static_cast<intptr_t>(kHeapObjectTag)); |
| 149 } | 149 } |
| 150 #endif | 150 #endif |
| 151 return result; | 151 return result; |
| 152 } | 152 } |
| 153 | 153 |
| 154 | 154 |
| 155 Object* Heap::AllocateRawCell() { | 155 MaybeObject* Heap::AllocateRawCell() { |
| 156 #ifdef DEBUG | 156 #ifdef DEBUG |
| 157 Counters::objs_since_last_full.Increment(); | 157 Counters::objs_since_last_full.Increment(); |
| 158 Counters::objs_since_last_young.Increment(); | 158 Counters::objs_since_last_young.Increment(); |
| 159 #endif | 159 #endif |
| 160 Object* result = cell_space_->AllocateRaw(JSGlobalPropertyCell::kSize); | 160 MaybeObject* result = cell_space_->AllocateRaw(JSGlobalPropertyCell::kSize); |
| 161 if (result->IsFailure()) old_gen_exhausted_ = true; | 161 if (result->IsFailure()) old_gen_exhausted_ = true; |
| 162 return result; | 162 return result; |
| 163 } | 163 } |
| 164 | 164 |
| 165 | 165 |
| 166 bool Heap::InNewSpace(Object* object) { | 166 bool Heap::InNewSpace(Object* object) { |
| 167 bool result = new_space_.Contains(object); | 167 bool result = new_space_.Contains(object); |
| 168 ASSERT(!result || // Either not in new space | 168 ASSERT(!result || // Either not in new space |
| 169 gc_state_ != NOT_IN_GC || // ... or in the middle of GC | 169 gc_state_ != NOT_IN_GC || // ... or in the middle of GC |
| 170 InToSpace(object)); // ... or in to-space (where we allocate). | 170 InToSpace(object)); // ... or in to-space (where we allocate). |
| (...skipping 152 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 323 if (first_word.IsForwardingAddress()) { | 323 if (first_word.IsForwardingAddress()) { |
| 324 *p = first_word.ToForwardingAddress(); | 324 *p = first_word.ToForwardingAddress(); |
| 325 return; | 325 return; |
| 326 } | 326 } |
| 327 | 327 |
| 328 // Call the slow part of scavenge object. | 328 // Call the slow part of scavenge object. |
| 329 return ScavengeObjectSlow(p, object); | 329 return ScavengeObjectSlow(p, object); |
| 330 } | 330 } |
| 331 | 331 |
| 332 | 332 |
| 333 Object* Heap::PrepareForCompare(String* str) { | 333 MaybeObject* Heap::PrepareForCompare(String* str) { |
| 334 // Always flatten small strings and force flattening of long strings | 334 // Always flatten small strings and force flattening of long strings |
| 335 // after we have accumulated a certain amount we failed to flatten. | 335 // after we have accumulated a certain amount we failed to flatten. |
| 336 static const int kMaxAlwaysFlattenLength = 32; | 336 static const int kMaxAlwaysFlattenLength = 32; |
| 337 static const int kFlattenLongThreshold = 16*KB; | 337 static const int kFlattenLongThreshold = 16*KB; |
| 338 | 338 |
| 339 const int length = str->length(); | 339 const int length = str->length(); |
| 340 Object* obj = str->TryFlatten(); | 340 MaybeObject* obj = str->TryFlatten(); |
| 341 if (length <= kMaxAlwaysFlattenLength || | 341 if (length <= kMaxAlwaysFlattenLength || |
| 342 unflattened_strings_length_ >= kFlattenLongThreshold) { | 342 unflattened_strings_length_ >= kFlattenLongThreshold) { |
| 343 return obj; | 343 return obj; |
| 344 } | 344 } |
| 345 if (obj->IsFailure()) { | 345 if (obj->IsFailure()) { |
| 346 unflattened_strings_length_ += length; | 346 unflattened_strings_length_ += length; |
| 347 } | 347 } |
| 348 return str; | 348 return str; |
| 349 } | 349 } |
| 350 | 350 |
| (...skipping 33 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 384 if (FLAG_gc_greedy) v8::internal::Heap::GarbageCollectionGreedyCheck() | 384 if (FLAG_gc_greedy) v8::internal::Heap::GarbageCollectionGreedyCheck() |
| 385 #else | 385 #else |
| 386 #define GC_GREEDY_CHECK() { } | 386 #define GC_GREEDY_CHECK() { } |
| 387 #endif | 387 #endif |
| 388 | 388 |
| 389 | 389 |
| 390 // Calls the FUNCTION_CALL function and retries it up to three times | 390 // Calls the FUNCTION_CALL function and retries it up to three times |
| 391 // to guarantee that any allocations performed during the call will | 391 // to guarantee that any allocations performed during the call will |
| 392 // succeed if there's enough memory. | 392 // succeed if there's enough memory. |
| 393 | 393 |
| 394 // Warning: Do not use the identifiers __object__ or __scope__ in a | 394 // Warning: Do not use the identifiers __object__, __maybe_object__ or |
| 395 // call to this macro. | 395 // __scope__ in a call to this macro. |
| 396 | 396 |
| 397 #define CALL_AND_RETRY(FUNCTION_CALL, RETURN_VALUE, RETURN_EMPTY) \ | 397 #define CALL_AND_RETRY(FUNCTION_CALL, RETURN_VALUE, RETURN_EMPTY) \ |
| 398 do { \ | 398 do { \ |
| 399 GC_GREEDY_CHECK(); \ | 399 GC_GREEDY_CHECK(); \ |
| 400 Object* __object__ = FUNCTION_CALL; \ | 400 MaybeObject* __maybe_object__ = FUNCTION_CALL; \ |
| 401 if (!__object__->IsFailure()) RETURN_VALUE; \ | 401 Object* __object__ = NULL; \ |
| 402 if (__object__->IsOutOfMemoryFailure()) { \ | 402 if (__maybe_object__->ToObject(&__object__)) RETURN_VALUE; \ |
| 403 if (__maybe_object__->IsOutOfMemory()) { \ |
| 403 v8::internal::V8::FatalProcessOutOfMemory("CALL_AND_RETRY_0", true);\ | 404 v8::internal::V8::FatalProcessOutOfMemory("CALL_AND_RETRY_0", true);\ |
| 404 } \ | 405 } \ |
| 405 if (!__object__->IsRetryAfterGC()) RETURN_EMPTY; \ | 406 if (!__maybe_object__->IsRetryAfterGC()) RETURN_EMPTY; \ |
| 406 Heap::CollectGarbage(Failure::cast(__object__)->allocation_space()); \ | 407 Heap::CollectGarbage(Failure::cast(__maybe_object__)-> \ |
| 407 __object__ = FUNCTION_CALL; \ | 408 allocation_space()); \ |
| 408 if (!__object__->IsFailure()) RETURN_VALUE; \ | 409 __maybe_object__ = FUNCTION_CALL; \ |
| 409 if (__object__->IsOutOfMemoryFailure()) { \ | 410 if (__maybe_object__->ToObject(&__object__)) RETURN_VALUE; \ |
| 411 if (__maybe_object__->IsOutOfMemory()) { \ |
| 410 v8::internal::V8::FatalProcessOutOfMemory("CALL_AND_RETRY_1", true);\ | 412 v8::internal::V8::FatalProcessOutOfMemory("CALL_AND_RETRY_1", true);\ |
| 411 } \ | 413 } \ |
| 412 if (!__object__->IsRetryAfterGC()) RETURN_EMPTY; \ | 414 if (!__maybe_object__->IsRetryAfterGC()) RETURN_EMPTY; \ |
| 413 Counters::gc_last_resort_from_handles.Increment(); \ | 415 Counters::gc_last_resort_from_handles.Increment(); \ |
| 414 Heap::CollectAllGarbage(false); \ | 416 Heap::CollectAllGarbage(false); \ |
| 415 { \ | 417 { \ |
| 416 AlwaysAllocateScope __scope__; \ | 418 AlwaysAllocateScope __scope__; \ |
| 417 __object__ = FUNCTION_CALL; \ | 419 __maybe_object__ = FUNCTION_CALL; \ |
| 418 } \ | 420 } \ |
| 419 if (!__object__->IsFailure()) RETURN_VALUE; \ | 421 if (__maybe_object__->ToObject(&__object__)) RETURN_VALUE; \ |
| 420 if (__object__->IsOutOfMemoryFailure() || \ | 422 if (__maybe_object__->IsOutOfMemory() || \ |
| 421 __object__->IsRetryAfterGC()) { \ | 423 __maybe_object__->IsRetryAfterGC()) { \ |
| 422 /* TODO(1181417): Fix this. */ \ | 424 /* TODO(1181417): Fix this. */ \ |
| 423 v8::internal::V8::FatalProcessOutOfMemory("CALL_AND_RETRY_2", true);\ | 425 v8::internal::V8::FatalProcessOutOfMemory("CALL_AND_RETRY_2", true);\ |
| 424 } \ | 426 } \ |
| 425 RETURN_EMPTY; \ | 427 RETURN_EMPTY; \ |
| 426 } while (false) | 428 } while (false) |
| 427 | 429 |
| 428 | 430 |
| 429 #define CALL_HEAP_FUNCTION(FUNCTION_CALL, TYPE) \ | 431 #define CALL_HEAP_FUNCTION(FUNCTION_CALL, TYPE) \ |
| 430 CALL_AND_RETRY(FUNCTION_CALL, \ | 432 CALL_AND_RETRY(FUNCTION_CALL, \ |
| 431 return Handle<TYPE>(TYPE::cast(__object__)), \ | 433 return Handle<TYPE>(TYPE::cast(__object__)), \ |
| (...skipping 61 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 493 | 495 |
| 494 | 496 |
| 495 void ExternalStringTable::ShrinkNewStrings(int position) { | 497 void ExternalStringTable::ShrinkNewStrings(int position) { |
| 496 new_space_strings_.Rewind(position); | 498 new_space_strings_.Rewind(position); |
| 497 Verify(); | 499 Verify(); |
| 498 } | 500 } |
| 499 | 501 |
| 500 } } // namespace v8::internal | 502 } } // namespace v8::internal |
| 501 | 503 |
| 502 #endif // V8_HEAP_INL_H_ | 504 #endif // V8_HEAP_INL_H_ |
| OLD | NEW |