| OLD | NEW |
| 1 // Copyright 2011 the V8 project authors. All rights reserved. | 1 // Copyright 2011 the V8 project authors. All rights reserved. |
| 2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
| 3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
| 4 // met: | 4 // met: |
| 5 // | 5 // |
| 6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
| 7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
| 8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
| 9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
| 10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
| (...skipping 51 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 62 reinterpret_cast<Address*>(RoundUp(start_as_int, kStoreBufferSize * 2)); | 62 reinterpret_cast<Address*>(RoundUp(start_as_int, kStoreBufferSize * 2)); |
| 63 limit_ = start_ + (kStoreBufferSize / kPointerSize); | 63 limit_ = start_ + (kStoreBufferSize / kPointerSize); |
| 64 | 64 |
| 65 old_virtual_memory_ = | 65 old_virtual_memory_ = |
| 66 new VirtualMemory(kOldStoreBufferLength * kPointerSize); | 66 new VirtualMemory(kOldStoreBufferLength * kPointerSize); |
| 67 old_top_ = old_start_ = | 67 old_top_ = old_start_ = |
| 68 reinterpret_cast<Address*>(old_virtual_memory_->address()); | 68 reinterpret_cast<Address*>(old_virtual_memory_->address()); |
| 69 // Don't know the alignment requirements of the OS, but it is certainly not | 69 // Don't know the alignment requirements of the OS, but it is certainly not |
| 70 // less than 0xfff. | 70 // less than 0xfff. |
| 71 ASSERT((reinterpret_cast<uintptr_t>(old_start_) & 0xfff) == 0); | 71 ASSERT((reinterpret_cast<uintptr_t>(old_start_) & 0xfff) == 0); |
| 72 int initial_length = OS::CommitPageSize() / kPointerSize; | 72 int initial_length = static_cast<int>(OS::CommitPageSize() / kPointerSize); |
| 73 ASSERT(initial_length > 0); | 73 ASSERT(initial_length > 0); |
| 74 ASSERT(initial_length <= kOldStoreBufferLength); | 74 ASSERT(initial_length <= kOldStoreBufferLength); |
| 75 old_limit_ = old_start_ + initial_length; | 75 old_limit_ = old_start_ + initial_length; |
| 76 old_reserved_limit_ = old_start_ + kOldStoreBufferLength; | 76 old_reserved_limit_ = old_start_ + kOldStoreBufferLength; |
| 77 | 77 |
| 78 CHECK(old_virtual_memory_->Commit( | 78 CHECK(old_virtual_memory_->Commit( |
| 79 reinterpret_cast<void*>(old_start_), | 79 reinterpret_cast<void*>(old_start_), |
| 80 (old_limit_ - old_start_) * kPointerSize, | 80 (old_limit_ - old_start_) * kPointerSize, |
| 81 false)); | 81 false)); |
| 82 | 82 |
| (...skipping 627 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 710 heap_->isolate()->counters()->store_buffer_compactions()->Increment(); | 710 heap_->isolate()->counters()->store_buffer_compactions()->Increment(); |
| 711 CheckForFullBuffer(); | 711 CheckForFullBuffer(); |
| 712 } | 712 } |
| 713 | 713 |
| 714 | 714 |
| 715 void StoreBuffer::CheckForFullBuffer() { | 715 void StoreBuffer::CheckForFullBuffer() { |
| 716 EnsureSpace(kStoreBufferSize * 2); | 716 EnsureSpace(kStoreBufferSize * 2); |
| 717 } | 717 } |
| 718 | 718 |
| 719 } } // namespace v8::internal | 719 } } // namespace v8::internal |
| OLD | NEW |