| Index: base/memory/discardable_memory_ashmem_allocator_unittest.cc
|
| diff --git a/base/memory/discardable_memory_allocator_android_unittest.cc b/base/memory/discardable_memory_ashmem_allocator_unittest.cc
|
| similarity index 62%
|
| rename from base/memory/discardable_memory_allocator_android_unittest.cc
|
| rename to base/memory/discardable_memory_ashmem_allocator_unittest.cc
|
| index b2e8d558608fb03c6f5a5142e63f1d30ef30009b..a20a418159b6755a414cfda69da802d948c73ada 100644
|
| --- a/base/memory/discardable_memory_allocator_android_unittest.cc
|
| +++ b/base/memory/discardable_memory_ashmem_allocator_unittest.cc
|
| @@ -2,7 +2,7 @@
|
| // Use of this source code is governed by a BSD-style license that can be
|
| // found in the LICENSE file.
|
|
|
| -#include "base/memory/discardable_memory_allocator_android.h"
|
| +#include "base/memory/discardable_memory_ashmem_allocator.h"
|
|
|
| #include <sys/types.h>
|
| #include <unistd.h>
|
| @@ -26,43 +26,44 @@ const size_t kPageSize = 4096;
|
| const size_t kMaxAllowedAllocationSize =
|
| std::numeric_limits<size_t>::max() - kPageSize + 1;
|
|
|
| -class DiscardableMemoryAllocatorTest : public testing::Test {
|
| +class DiscardableMemoryAshmemAllocatorTest : public testing::Test {
|
| protected:
|
| - DiscardableMemoryAllocatorTest()
|
| + DiscardableMemoryAshmemAllocatorTest()
|
| : allocator_(kAllocatorName, kAshmemRegionSizeForTesting) {
|
| }
|
|
|
| - DiscardableMemoryAllocator allocator_;
|
| + DiscardableMemoryAshmemAllocator allocator_;
|
| };
|
|
|
| -void WriteToDiscardableMemory(DiscardableMemory* memory, size_t size) {
|
| +void WriteToDiscardableAshmemChunk(DiscardableAshmemChunk* memory,
|
| + size_t size) {
|
| // Write to the first and the last pages only to avoid paging in up to 64
|
| // MBytes.
|
| static_cast<char*>(memory->Memory())[0] = 'a';
|
| static_cast<char*>(memory->Memory())[size - 1] = 'a';
|
| }
|
|
|
| -TEST_F(DiscardableMemoryAllocatorTest, Basic) {
|
| +TEST_F(DiscardableMemoryAshmemAllocatorTest, Basic) {
|
| const size_t size = 128;
|
| - scoped_ptr<DiscardableMemory> memory(allocator_.Allocate(size));
|
| + scoped_ptr<DiscardableAshmemChunk> memory(allocator_.Allocate(size));
|
| ASSERT_TRUE(memory);
|
| - WriteToDiscardableMemory(memory.get(), size);
|
| + WriteToDiscardableAshmemChunk(memory.get(), size);
|
| }
|
|
|
| -TEST_F(DiscardableMemoryAllocatorTest, ZeroAllocationIsNotSupported) {
|
| - scoped_ptr<DiscardableMemory> memory(allocator_.Allocate(0));
|
| +TEST_F(DiscardableMemoryAshmemAllocatorTest, ZeroAllocationIsNotSupported) {
|
| + scoped_ptr<DiscardableAshmemChunk> memory(allocator_.Allocate(0));
|
| ASSERT_FALSE(memory);
|
| }
|
|
|
| -TEST_F(DiscardableMemoryAllocatorTest, TooLargeAllocationFails) {
|
| - scoped_ptr<DiscardableMemory> memory(
|
| +TEST_F(DiscardableMemoryAshmemAllocatorTest, TooLargeAllocationFails) {
|
| + scoped_ptr<DiscardableAshmemChunk> memory(
|
| allocator_.Allocate(kMaxAllowedAllocationSize + 1));
|
| // Page-alignment would have caused an overflow resulting in a small
|
| // allocation if the input size wasn't checked correctly.
|
| ASSERT_FALSE(memory);
|
| }
|
|
|
| -TEST_F(DiscardableMemoryAllocatorTest,
|
| +TEST_F(DiscardableMemoryAshmemAllocatorTest,
|
| AshmemRegionsAreNotSmallerThanRequestedSize) {
|
| // The creation of the underlying ashmem region is expected to fail since
|
| // there should not be enough room in the address space. When ashmem creation
|
| @@ -70,43 +71,43 @@ TEST_F(DiscardableMemoryAllocatorTest,
|
| // size should not be smaller than the size the user requested so the
|
| // allocation here should just fail (and not succeed with the minimum ashmem
|
| // region size).
|
| - scoped_ptr<DiscardableMemory> memory(
|
| + scoped_ptr<DiscardableAshmemChunk> memory(
|
| allocator_.Allocate(kMaxAllowedAllocationSize));
|
| ASSERT_FALSE(memory);
|
| }
|
|
|
| -TEST_F(DiscardableMemoryAllocatorTest, AshmemRegionsAreAlwaysPageAligned) {
|
| +TEST_F(DiscardableMemoryAshmemAllocatorTest,
|
| + AshmemRegionsAreAlwaysPageAligned) {
|
| // Use a separate allocator here so that we can override the ashmem region
|
| // size.
|
| - DiscardableMemoryAllocator allocator(
|
| + DiscardableMemoryAshmemAllocator allocator(
|
| kAllocatorName, kMaxAllowedAllocationSize);
|
| - scoped_ptr<DiscardableMemory> memory(allocator.Allocate(kPageSize));
|
| + scoped_ptr<DiscardableAshmemChunk> memory(allocator.Allocate(kPageSize));
|
| ASSERT_TRUE(memory);
|
| EXPECT_GT(kMaxAllowedAllocationSize, allocator.last_ashmem_region_size());
|
| ASSERT_TRUE(allocator.last_ashmem_region_size() % kPageSize == 0);
|
| }
|
|
|
| -TEST_F(DiscardableMemoryAllocatorTest, LargeAllocation) {
|
| - // Note that large allocations should just use DiscardableMemoryAndroidSimple
|
| - // instead.
|
| +TEST_F(DiscardableMemoryAshmemAllocatorTest, LargeAllocation) {
|
| const size_t size = 64 * 1024 * 1024;
|
| - scoped_ptr<DiscardableMemory> memory(allocator_.Allocate(size));
|
| + scoped_ptr<DiscardableAshmemChunk> memory(allocator_.Allocate(size));
|
| ASSERT_TRUE(memory);
|
| - WriteToDiscardableMemory(memory.get(), size);
|
| + WriteToDiscardableAshmemChunk(memory.get(), size);
|
| }
|
|
|
| -TEST_F(DiscardableMemoryAllocatorTest, ChunksArePageAligned) {
|
| - scoped_ptr<DiscardableMemory> memory(allocator_.Allocate(kPageSize));
|
| +TEST_F(DiscardableMemoryAshmemAllocatorTest, ChunksArePageAligned) {
|
| + scoped_ptr<DiscardableAshmemChunk> memory(allocator_.Allocate(kPageSize));
|
| ASSERT_TRUE(memory);
|
| EXPECT_EQ(0U, reinterpret_cast<uint64_t>(memory->Memory()) % kPageSize);
|
| - WriteToDiscardableMemory(memory.get(), kPageSize);
|
| + WriteToDiscardableAshmemChunk(memory.get(), kPageSize);
|
| }
|
|
|
| -TEST_F(DiscardableMemoryAllocatorTest, AllocateFreeAllocate) {
|
| - scoped_ptr<DiscardableMemory> memory(allocator_.Allocate(kPageSize));
|
| +TEST_F(DiscardableMemoryAshmemAllocatorTest, AllocateFreeAllocate) {
|
| + scoped_ptr<DiscardableAshmemChunk> memory(allocator_.Allocate(kPageSize));
|
| // Extra allocation that prevents the region from being deleted when |memory|
|
| // gets deleted.
|
| - scoped_ptr<DiscardableMemory> memory_lock(allocator_.Allocate(kPageSize));
|
| + scoped_ptr<DiscardableAshmemChunk> memory_lock(
|
| + allocator_.Allocate(kPageSize));
|
| ASSERT_TRUE(memory);
|
| void* const address = memory->Memory();
|
| memory->Unlock(); // Tests that the reused chunk is being locked correctly.
|
| @@ -115,11 +116,12 @@ TEST_F(DiscardableMemoryAllocatorTest, AllocateFreeAllocate) {
|
| ASSERT_TRUE(memory);
|
| // The previously freed chunk should be reused.
|
| EXPECT_EQ(address, memory->Memory());
|
| - WriteToDiscardableMemory(memory.get(), kPageSize);
|
| + WriteToDiscardableAshmemChunk(memory.get(), kPageSize);
|
| }
|
|
|
| -TEST_F(DiscardableMemoryAllocatorTest, FreeingWholeAshmemRegionClosesAshmem) {
|
| - scoped_ptr<DiscardableMemory> memory(allocator_.Allocate(kPageSize));
|
| +TEST_F(DiscardableMemoryAshmemAllocatorTest,
|
| + FreeingWholeAshmemRegionClosesAshmem) {
|
| + scoped_ptr<DiscardableAshmemChunk> memory(allocator_.Allocate(kPageSize));
|
| ASSERT_TRUE(memory);
|
| const int kMagic = 0xdeadbeef;
|
| *static_cast<int*>(memory->Memory()) = kMagic;
|
| @@ -131,12 +133,15 @@ TEST_F(DiscardableMemoryAllocatorTest, FreeingWholeAshmemRegionClosesAshmem) {
|
| EXPECT_NE(kMagic, *static_cast<const int*>(memory->Memory()));
|
| }
|
|
|
| -TEST_F(DiscardableMemoryAllocatorTest, AllocateUsesBestFitAlgorithm) {
|
| - scoped_ptr<DiscardableMemory> memory1(allocator_.Allocate(3 * kPageSize));
|
| +TEST_F(DiscardableMemoryAshmemAllocatorTest, AllocateUsesBestFitAlgorithm) {
|
| + scoped_ptr<DiscardableAshmemChunk> memory1(
|
| + allocator_.Allocate(3 * kPageSize));
|
| ASSERT_TRUE(memory1);
|
| - scoped_ptr<DiscardableMemory> memory2(allocator_.Allocate(2 * kPageSize));
|
| + scoped_ptr<DiscardableAshmemChunk> memory2(
|
| + allocator_.Allocate(2 * kPageSize));
|
| ASSERT_TRUE(memory2);
|
| - scoped_ptr<DiscardableMemory> memory3(allocator_.Allocate(1 * kPageSize));
|
| + scoped_ptr<DiscardableAshmemChunk> memory3(
|
| + allocator_.Allocate(1 * kPageSize));
|
| ASSERT_TRUE(memory3);
|
| void* const address_3 = memory3->Memory();
|
| memory1.reset();
|
| @@ -146,17 +151,17 @@ TEST_F(DiscardableMemoryAllocatorTest, AllocateUsesBestFitAlgorithm) {
|
| ASSERT_TRUE(memory1);
|
| // The chunk whose size is closest to the requested size should be reused.
|
| EXPECT_EQ(address_3, memory1->Memory());
|
| - WriteToDiscardableMemory(memory1.get(), kPageSize);
|
| + WriteToDiscardableAshmemChunk(memory1.get(), kPageSize);
|
| }
|
|
|
| -TEST_F(DiscardableMemoryAllocatorTest, MergeFreeChunks) {
|
| - scoped_ptr<DiscardableMemory> memory1(allocator_.Allocate(kPageSize));
|
| +TEST_F(DiscardableMemoryAshmemAllocatorTest, MergeFreeChunks) {
|
| + scoped_ptr<DiscardableAshmemChunk> memory1(allocator_.Allocate(kPageSize));
|
| ASSERT_TRUE(memory1);
|
| - scoped_ptr<DiscardableMemory> memory2(allocator_.Allocate(kPageSize));
|
| + scoped_ptr<DiscardableAshmemChunk> memory2(allocator_.Allocate(kPageSize));
|
| ASSERT_TRUE(memory2);
|
| - scoped_ptr<DiscardableMemory> memory3(allocator_.Allocate(kPageSize));
|
| + scoped_ptr<DiscardableAshmemChunk> memory3(allocator_.Allocate(kPageSize));
|
| ASSERT_TRUE(memory3);
|
| - scoped_ptr<DiscardableMemory> memory4(allocator_.Allocate(kPageSize));
|
| + scoped_ptr<DiscardableAshmemChunk> memory4(allocator_.Allocate(kPageSize));
|
| ASSERT_TRUE(memory4);
|
| void* const memory1_address = memory1->Memory();
|
| memory1.reset();
|
| @@ -168,10 +173,12 @@ TEST_F(DiscardableMemoryAllocatorTest, MergeFreeChunks) {
|
| EXPECT_EQ(memory1_address, memory1->Memory());
|
| }
|
|
|
| -TEST_F(DiscardableMemoryAllocatorTest, MergeFreeChunksAdvanced) {
|
| - scoped_ptr<DiscardableMemory> memory1(allocator_.Allocate(4 * kPageSize));
|
| +TEST_F(DiscardableMemoryAshmemAllocatorTest, MergeFreeChunksAdvanced) {
|
| + scoped_ptr<DiscardableAshmemChunk> memory1(
|
| + allocator_.Allocate(4 * kPageSize));
|
| ASSERT_TRUE(memory1);
|
| - scoped_ptr<DiscardableMemory> memory2(allocator_.Allocate(4 * kPageSize));
|
| + scoped_ptr<DiscardableAshmemChunk> memory2(
|
| + allocator_.Allocate(4 * kPageSize));
|
| ASSERT_TRUE(memory2);
|
| void* const memory1_address = memory1->Memory();
|
| memory1.reset();
|
| @@ -185,15 +192,18 @@ TEST_F(DiscardableMemoryAllocatorTest, MergeFreeChunksAdvanced) {
|
| static_cast<const char*>(memory1_address) + 2 * kPageSize);
|
| }
|
|
|
| -TEST_F(DiscardableMemoryAllocatorTest, MergeFreeChunksAdvanced2) {
|
| - scoped_ptr<DiscardableMemory> memory1(allocator_.Allocate(4 * kPageSize));
|
| +TEST_F(DiscardableMemoryAshmemAllocatorTest, MergeFreeChunksAdvanced2) {
|
| + scoped_ptr<DiscardableAshmemChunk> memory1(
|
| + allocator_.Allocate(4 * kPageSize));
|
| ASSERT_TRUE(memory1);
|
| - scoped_ptr<DiscardableMemory> memory2(allocator_.Allocate(4 * kPageSize));
|
| + scoped_ptr<DiscardableAshmemChunk> memory2(
|
| + allocator_.Allocate(4 * kPageSize));
|
| ASSERT_TRUE(memory2);
|
| void* const memory1_address = memory1->Memory();
|
| memory1.reset();
|
| memory1 = allocator_.Allocate(2 * kPageSize);
|
| - scoped_ptr<DiscardableMemory> memory3(allocator_.Allocate(2 * kPageSize));
|
| + scoped_ptr<DiscardableAshmemChunk> memory3(
|
| + allocator_.Allocate(2 * kPageSize));
|
| // At this point, the region should be in this state:
|
| // 8 KBytes (used), 8 KBytes (used), 16 KBytes (used).
|
| memory3.reset();
|
| @@ -206,14 +216,18 @@ TEST_F(DiscardableMemoryAllocatorTest, MergeFreeChunksAdvanced2) {
|
| static_cast<const char*>(memory1_address) + 2 * kPageSize);
|
| }
|
|
|
| -TEST_F(DiscardableMemoryAllocatorTest, MergeFreeChunksAndDeleteAshmemRegion) {
|
| - scoped_ptr<DiscardableMemory> memory1(allocator_.Allocate(4 * kPageSize));
|
| +TEST_F(DiscardableMemoryAshmemAllocatorTest,
|
| + MergeFreeChunksAndDeleteAshmemRegion) {
|
| + scoped_ptr<DiscardableAshmemChunk> memory1(
|
| + allocator_.Allocate(4 * kPageSize));
|
| ASSERT_TRUE(memory1);
|
| - scoped_ptr<DiscardableMemory> memory2(allocator_.Allocate(4 * kPageSize));
|
| + scoped_ptr<DiscardableAshmemChunk> memory2(
|
| + allocator_.Allocate(4 * kPageSize));
|
| ASSERT_TRUE(memory2);
|
| memory1.reset();
|
| memory1 = allocator_.Allocate(2 * kPageSize);
|
| - scoped_ptr<DiscardableMemory> memory3(allocator_.Allocate(2 * kPageSize));
|
| + scoped_ptr<DiscardableAshmemChunk> memory3(
|
| + allocator_.Allocate(2 * kPageSize));
|
| // At this point, the region should be in this state:
|
| // 8 KBytes (used), 8 KBytes (used), 16 KBytes (used).
|
| memory1.reset();
|
| @@ -228,13 +242,13 @@ TEST_F(DiscardableMemoryAllocatorTest, MergeFreeChunksAndDeleteAshmemRegion) {
|
| EXPECT_NE(kMagic, *static_cast<int*>(memory2->Memory()));
|
| }
|
|
|
| -TEST_F(DiscardableMemoryAllocatorTest,
|
| +TEST_F(DiscardableMemoryAshmemAllocatorTest,
|
| TooLargeFreeChunksDontCauseTooMuchFragmentationWhenRecycled) {
|
| // Keep |memory_1| below allocated so that the ashmem region doesn't get
|
| // closed when |memory_2| is deleted.
|
| - scoped_ptr<DiscardableMemory> memory_1(allocator_.Allocate(64 * 1024));
|
| + scoped_ptr<DiscardableAshmemChunk> memory_1(allocator_.Allocate(64 * 1024));
|
| ASSERT_TRUE(memory_1);
|
| - scoped_ptr<DiscardableMemory> memory_2(allocator_.Allocate(32 * 1024));
|
| + scoped_ptr<DiscardableAshmemChunk> memory_2(allocator_.Allocate(32 * 1024));
|
| ASSERT_TRUE(memory_2);
|
| void* const address = memory_2->Memory();
|
| memory_2.reset();
|
| @@ -242,40 +256,41 @@ TEST_F(DiscardableMemoryAllocatorTest,
|
| memory_2 = allocator_.Allocate(size);
|
| ASSERT_TRUE(memory_2);
|
| EXPECT_EQ(address, memory_2->Memory());
|
| - WriteToDiscardableMemory(memory_2.get(), size);
|
| - scoped_ptr<DiscardableMemory> memory_3(allocator_.Allocate(size));
|
| + WriteToDiscardableAshmemChunk(memory_2.get(), size);
|
| + scoped_ptr<DiscardableAshmemChunk> memory_3(allocator_.Allocate(size));
|
| // The unused tail (16 KBytes large) of the previously freed chunk should be
|
| // reused.
|
| EXPECT_EQ(static_cast<char*>(address) + size, memory_3->Memory());
|
| - WriteToDiscardableMemory(memory_3.get(), size);
|
| + WriteToDiscardableAshmemChunk(memory_3.get(), size);
|
| }
|
|
|
| -TEST_F(DiscardableMemoryAllocatorTest, UseMultipleAshmemRegions) {
|
| +TEST_F(DiscardableMemoryAshmemAllocatorTest, UseMultipleAshmemRegions) {
|
| // Leave one page untouched at the end of the ashmem region.
|
| const size_t size = kAshmemRegionSizeForTesting - kPageSize;
|
| - scoped_ptr<DiscardableMemory> memory1(allocator_.Allocate(size));
|
| + scoped_ptr<DiscardableAshmemChunk> memory1(allocator_.Allocate(size));
|
| ASSERT_TRUE(memory1);
|
| - WriteToDiscardableMemory(memory1.get(), size);
|
| + WriteToDiscardableAshmemChunk(memory1.get(), size);
|
|
|
| - scoped_ptr<DiscardableMemory> memory2(
|
| + scoped_ptr<DiscardableAshmemChunk> memory2(
|
| allocator_.Allocate(kAshmemRegionSizeForTesting));
|
| ASSERT_TRUE(memory2);
|
| - WriteToDiscardableMemory(memory2.get(), kAshmemRegionSizeForTesting);
|
| + WriteToDiscardableAshmemChunk(memory2.get(), kAshmemRegionSizeForTesting);
|
| // The last page of the first ashmem region should be used for this
|
| // allocation.
|
| - scoped_ptr<DiscardableMemory> memory3(allocator_.Allocate(kPageSize));
|
| + scoped_ptr<DiscardableAshmemChunk> memory3(allocator_.Allocate(kPageSize));
|
| ASSERT_TRUE(memory3);
|
| - WriteToDiscardableMemory(memory3.get(), kPageSize);
|
| + WriteToDiscardableAshmemChunk(memory3.get(), kPageSize);
|
| EXPECT_EQ(memory3->Memory(), static_cast<char*>(memory1->Memory()) + size);
|
| }
|
|
|
| -TEST_F(DiscardableMemoryAllocatorTest,
|
| +TEST_F(DiscardableMemoryAshmemAllocatorTest,
|
| HighestAllocatedChunkPointerIsUpdatedWhenHighestChunkGetsSplit) {
|
| // Prevents the ashmem region from getting closed when |memory2| gets freed.
|
| - scoped_ptr<DiscardableMemory> memory1(allocator_.Allocate(kPageSize));
|
| + scoped_ptr<DiscardableAshmemChunk> memory1(allocator_.Allocate(kPageSize));
|
| ASSERT_TRUE(memory1);
|
|
|
| - scoped_ptr<DiscardableMemory> memory2(allocator_.Allocate(4 * kPageSize));
|
| + scoped_ptr<DiscardableAshmemChunk> memory2(
|
| + allocator_.Allocate(4 * kPageSize));
|
| ASSERT_TRUE(memory2);
|
|
|
| memory2.reset();
|
| @@ -288,7 +303,8 @@ TEST_F(DiscardableMemoryAllocatorTest,
|
|
|
| // Allocate more than 3 * |kPageSize| so that the free chunk of size 3 *
|
| // |kPageSize| is not reused and |highest_allocated_chunk_| gets used instead.
|
| - scoped_ptr<DiscardableMemory> memory3(allocator_.Allocate(4 * kPageSize));
|
| + scoped_ptr<DiscardableAshmemChunk> memory3(
|
| + allocator_.Allocate(4 * kPageSize));
|
| ASSERT_TRUE(memory3);
|
|
|
| // Deleting |memory3| (whose size is 4 * |kPageSize|) should result in a merge
|
|
|