OLD | NEW |
1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 193 matching lines...) Loading... |
204 | 204 |
205 | 205 |
206 MaybeObject* Heap::CopyFixedDoubleArray(FixedDoubleArray* src) { | 206 MaybeObject* Heap::CopyFixedDoubleArray(FixedDoubleArray* src) { |
207 return CopyFixedDoubleArrayWithMap(src, src->map()); | 207 return CopyFixedDoubleArrayWithMap(src, src->map()); |
208 } | 208 } |
209 | 209 |
210 | 210 |
211 MaybeObject* Heap::AllocateRaw(int size_in_bytes, | 211 MaybeObject* Heap::AllocateRaw(int size_in_bytes, |
212 AllocationSpace space, | 212 AllocationSpace space, |
213 AllocationSpace retry_space) { | 213 AllocationSpace retry_space) { |
214 SLOW_ASSERT(!isolate_->optimizing_compiler_thread()->IsOptimizerThread()); | 214 ASSERT(AllowHandleAllocation::IsAllowed() && gc_state_ == NOT_IN_GC); |
215 ASSERT(allocation_allowed_ && gc_state_ == NOT_IN_GC); | |
216 ASSERT(space != NEW_SPACE || | 215 ASSERT(space != NEW_SPACE || |
217 retry_space == OLD_POINTER_SPACE || | 216 retry_space == OLD_POINTER_SPACE || |
218 retry_space == OLD_DATA_SPACE || | 217 retry_space == OLD_DATA_SPACE || |
219 retry_space == LO_SPACE); | 218 retry_space == LO_SPACE); |
220 #ifdef DEBUG | 219 #ifdef DEBUG |
221 if (FLAG_gc_interval >= 0 && | 220 if (FLAG_gc_interval >= 0 && |
222 !disallow_allocation_failure_ && | 221 !disallow_allocation_failure_ && |
223 Heap::allocation_timeout_-- <= 0) { | 222 Heap::allocation_timeout_-- <= 0) { |
224 return Failure::RetryAfterGC(space); | 223 return Failure::RetryAfterGC(space); |
225 } | 224 } |
(...skipping 409 matching lines...) Loading... |
635 | 634 |
636 | 635 |
637 #define CALL_HEAP_FUNCTION_PASS_EXCEPTION(ISOLATE, FUNCTION_CALL) \ | 636 #define CALL_HEAP_FUNCTION_PASS_EXCEPTION(ISOLATE, FUNCTION_CALL) \ |
638 CALL_AND_RETRY(ISOLATE, \ | 637 CALL_AND_RETRY(ISOLATE, \ |
639 FUNCTION_CALL, \ | 638 FUNCTION_CALL, \ |
640 return __object__, \ | 639 return __object__, \ |
641 return __maybe_object__, \ | 640 return __maybe_object__, \ |
642 return __maybe_object__) | 641 return __maybe_object__) |
643 | 642 |
644 | 643 |
645 #ifdef DEBUG | |
646 | |
647 inline bool Heap::allow_allocation(bool new_state) { | |
648 bool old = allocation_allowed_; | |
649 allocation_allowed_ = new_state; | |
650 return old; | |
651 } | |
652 | |
653 inline void Heap::set_allow_allocation(bool allocation_allowed) { | |
654 allocation_allowed_ = allocation_allowed; | |
655 } | |
656 | |
657 #endif | |
658 | |
659 | |
660 void ExternalStringTable::AddString(String* string) { | 644 void ExternalStringTable::AddString(String* string) { |
661 ASSERT(string->IsExternalString()); | 645 ASSERT(string->IsExternalString()); |
662 if (heap_->InNewSpace(string)) { | 646 if (heap_->InNewSpace(string)) { |
663 new_space_strings_.Add(string); | 647 new_space_strings_.Add(string); |
664 } else { | 648 } else { |
665 old_space_strings_.Add(string); | 649 old_space_strings_.Add(string); |
666 } | 650 } |
667 } | 651 } |
668 | 652 |
669 | 653 |
(...skipping 190 matching lines...) Loading... |
860 } | 844 } |
861 | 845 |
862 | 846 |
863 DisallowAllocationFailure::~DisallowAllocationFailure() { | 847 DisallowAllocationFailure::~DisallowAllocationFailure() { |
864 #ifdef DEBUG | 848 #ifdef DEBUG |
865 HEAP->disallow_allocation_failure_ = old_state_; | 849 HEAP->disallow_allocation_failure_ = old_state_; |
866 #endif | 850 #endif |
867 } | 851 } |
868 | 852 |
869 | 853 |
870 #ifdef DEBUG | |
871 bool EnterAllocationScope(Isolate* isolate, bool allow_allocation) { | |
872 bool active = !isolate->optimizing_compiler_thread()->IsOptimizerThread(); | |
873 bool last_state = isolate->heap()->IsAllocationAllowed(); | |
874 if (active) { | |
875 // TODO(yangguo): Make HandleDereferenceGuard avoid isolate mutation in the | |
876 // same way if running on the optimizer thread. | |
877 isolate->heap()->set_allow_allocation(allow_allocation); | |
878 } | |
879 return last_state; | |
880 } | |
881 | |
882 | |
883 void ExitAllocationScope(Isolate* isolate, bool last_state) { | |
884 bool active = !isolate->optimizing_compiler_thread()->IsOptimizerThread(); | |
885 if (active) { | |
886 isolate->heap()->set_allow_allocation(last_state); | |
887 } | |
888 } | |
889 | |
890 | |
891 AssertNoAllocation::AssertNoAllocation() | |
892 : last_state_(EnterAllocationScope(ISOLATE, false)) { | |
893 } | |
894 | |
895 AssertNoAllocation::~AssertNoAllocation() { | |
896 ExitAllocationScope(ISOLATE, last_state_); | |
897 } | |
898 | |
899 DisableAssertNoAllocation::DisableAssertNoAllocation() | |
900 : last_state_(EnterAllocationScope(ISOLATE, true)) { | |
901 } | |
902 | |
903 DisableAssertNoAllocation::~DisableAssertNoAllocation() { | |
904 ExitAllocationScope(ISOLATE, last_state_); | |
905 } | |
906 #else | |
907 | |
908 AssertNoAllocation::AssertNoAllocation() { } | |
909 AssertNoAllocation::~AssertNoAllocation() { } | |
910 DisableAssertNoAllocation::DisableAssertNoAllocation() { } | |
911 DisableAssertNoAllocation::~DisableAssertNoAllocation() { } | |
912 | |
913 #endif | |
914 | |
915 | |
916 } } // namespace v8::internal | 854 } } // namespace v8::internal |
917 | 855 |
918 #endif // V8_HEAP_INL_H_ | 856 #endif // V8_HEAP_INL_H_ |
OLD | NEW |