OLD | NEW |
1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 822 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
833 Heap* heap, | 833 Heap* heap, |
834 AllocationSpace space, | 834 AllocationSpace space, |
835 const char* gc_reason = NULL) { | 835 const char* gc_reason = NULL) { |
836 heap->mark_compact_collector()->SetFlags(Heap::kAbortIncrementalMarkingMask); | 836 heap->mark_compact_collector()->SetFlags(Heap::kAbortIncrementalMarkingMask); |
837 bool result = heap->CollectGarbage(space, gc_reason); | 837 bool result = heap->CollectGarbage(space, gc_reason); |
838 heap->mark_compact_collector()->SetFlags(Heap::kNoGCFlags); | 838 heap->mark_compact_collector()->SetFlags(Heap::kNoGCFlags); |
839 return result; | 839 return result; |
840 } | 840 } |
841 | 841 |
842 | 842 |
843 void Heap::ReserveSpace( | 843 void Heap::ReserveSpace(int *sizes, Address *locations_out) { |
844 int *sizes, | |
845 Address *locations_out) { | |
846 bool gc_performed = true; | 844 bool gc_performed = true; |
847 int counter = 0; | 845 int counter = 0; |
848 static const int kThreshold = 20; | 846 static const int kThreshold = 20; |
849 while (gc_performed && counter++ < kThreshold) { | 847 while (gc_performed && counter++ < kThreshold) { |
850 gc_performed = false; | 848 gc_performed = false; |
851 ASSERT(NEW_SPACE == FIRST_PAGED_SPACE - 1); | 849 ASSERT(NEW_SPACE == FIRST_PAGED_SPACE - 1); |
852 for (int space = NEW_SPACE; space <= LAST_PAGED_SPACE; space++) { | 850 for (int space = NEW_SPACE; space <= LAST_PAGED_SPACE; space++) { |
853 if (sizes[space] != 0) { | 851 if (sizes[space] != 0) { |
854 MaybeObject* allocation; | 852 MaybeObject* allocation; |
855 if (space == NEW_SPACE) { | 853 if (space == NEW_SPACE) { |
(...skipping 7057 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
7913 if (FLAG_concurrent_recompilation) { | 7911 if (FLAG_concurrent_recompilation) { |
7914 heap_->relocation_mutex_->Lock(); | 7912 heap_->relocation_mutex_->Lock(); |
7915 #ifdef DEBUG | 7913 #ifdef DEBUG |
7916 heap_->relocation_mutex_locked_by_optimizer_thread_ = | 7914 heap_->relocation_mutex_locked_by_optimizer_thread_ = |
7917 heap_->isolate()->optimizing_compiler_thread()->IsOptimizerThread(); | 7915 heap_->isolate()->optimizing_compiler_thread()->IsOptimizerThread(); |
7918 #endif // DEBUG | 7916 #endif // DEBUG |
7919 } | 7917 } |
7920 } | 7918 } |
7921 | 7919 |
7922 } } // namespace v8::internal | 7920 } } // namespace v8::internal |
OLD | NEW |