Index: base/memory/discardable_memory_allocator_android_unittest.cc |
diff --git a/base/memory/discardable_memory_allocator_android_unittest.cc b/base/memory/discardable_memory_allocator_android_unittest.cc |
index b2e8d558608fb03c6f5a5142e63f1d30ef30009b..e5cc7e34f68906ef7d5b6934f8f1be186071118b 100644 |
--- a/base/memory/discardable_memory_allocator_android_unittest.cc |
+++ b/base/memory/discardable_memory_allocator_android_unittest.cc |
@@ -35,7 +35,8 @@ class DiscardableMemoryAllocatorTest : public testing::Test { |
DiscardableMemoryAllocator allocator_; |
}; |
-void WriteToDiscardableMemory(DiscardableMemory* memory, size_t size) { |
+void WriteToDiscardableAshmemChunk(DiscardableAshmemChunk* memory, |
+ size_t size) { |
// Write to the first and the last pages only to avoid paging in up to 64 |
// MBytes. |
static_cast<char*>(memory->Memory())[0] = 'a'; |
@@ -44,18 +45,18 @@ void WriteToDiscardableMemory(DiscardableMemory* memory, size_t size) { |
TEST_F(DiscardableMemoryAllocatorTest, Basic) { |
const size_t size = 128; |
- scoped_ptr<DiscardableMemory> memory(allocator_.Allocate(size)); |
+ scoped_ptr<DiscardableAshmemChunk> memory(allocator_.Allocate(size)); |
ASSERT_TRUE(memory); |
- WriteToDiscardableMemory(memory.get(), size); |
+ WriteToDiscardableAshmemChunk(memory.get(), size); |
} |
TEST_F(DiscardableMemoryAllocatorTest, ZeroAllocationIsNotSupported) { |
- scoped_ptr<DiscardableMemory> memory(allocator_.Allocate(0)); |
+ scoped_ptr<DiscardableAshmemChunk> memory(allocator_.Allocate(0)); |
ASSERT_FALSE(memory); |
} |
TEST_F(DiscardableMemoryAllocatorTest, TooLargeAllocationFails) { |
- scoped_ptr<DiscardableMemory> memory( |
+ scoped_ptr<DiscardableAshmemChunk> memory( |
allocator_.Allocate(kMaxAllowedAllocationSize + 1)); |
// Page-alignment would have caused an overflow resulting in a small |
// allocation if the input size wasn't checked correctly. |
@@ -70,7 +71,7 @@ TEST_F(DiscardableMemoryAllocatorTest, |
// size should not be smaller than the size the user requested so the |
// allocation here should just fail (and not succeed with the minimum ashmem |
// region size). |
- scoped_ptr<DiscardableMemory> memory( |
+ scoped_ptr<DiscardableAshmemChunk> memory( |
allocator_.Allocate(kMaxAllowedAllocationSize)); |
ASSERT_FALSE(memory); |
} |
@@ -80,33 +81,32 @@ TEST_F(DiscardableMemoryAllocatorTest, AshmemRegionsAreAlwaysPageAligned) { |
// size. |
DiscardableMemoryAllocator allocator( |
kAllocatorName, kMaxAllowedAllocationSize); |
- scoped_ptr<DiscardableMemory> memory(allocator.Allocate(kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory(allocator.Allocate(kPageSize)); |
ASSERT_TRUE(memory); |
EXPECT_GT(kMaxAllowedAllocationSize, allocator.last_ashmem_region_size()); |
ASSERT_TRUE(allocator.last_ashmem_region_size() % kPageSize == 0); |
} |
TEST_F(DiscardableMemoryAllocatorTest, LargeAllocation) { |
- // Note that large allocations should just use DiscardableMemoryAndroidSimple |
- // instead. |
const size_t size = 64 * 1024 * 1024; |
- scoped_ptr<DiscardableMemory> memory(allocator_.Allocate(size)); |
+ scoped_ptr<DiscardableAshmemChunk> memory(allocator_.Allocate(size)); |
ASSERT_TRUE(memory); |
- WriteToDiscardableMemory(memory.get(), size); |
+ WriteToDiscardableAshmemChunk(memory.get(), size); |
} |
TEST_F(DiscardableMemoryAllocatorTest, ChunksArePageAligned) { |
- scoped_ptr<DiscardableMemory> memory(allocator_.Allocate(kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory(allocator_.Allocate(kPageSize)); |
ASSERT_TRUE(memory); |
EXPECT_EQ(0U, reinterpret_cast<uint64_t>(memory->Memory()) % kPageSize); |
- WriteToDiscardableMemory(memory.get(), kPageSize); |
+ WriteToDiscardableAshmemChunk(memory.get(), kPageSize); |
} |
TEST_F(DiscardableMemoryAllocatorTest, AllocateFreeAllocate) { |
- scoped_ptr<DiscardableMemory> memory(allocator_.Allocate(kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory(allocator_.Allocate(kPageSize)); |
// Extra allocation that prevents the region from being deleted when |memory| |
// gets deleted. |
- scoped_ptr<DiscardableMemory> memory_lock(allocator_.Allocate(kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory_lock( |
+ allocator_.Allocate(kPageSize)); |
ASSERT_TRUE(memory); |
void* const address = memory->Memory(); |
memory->Unlock(); // Tests that the reused chunk is being locked correctly. |
@@ -115,11 +115,11 @@ TEST_F(DiscardableMemoryAllocatorTest, AllocateFreeAllocate) { |
ASSERT_TRUE(memory); |
// The previously freed chunk should be reused. |
EXPECT_EQ(address, memory->Memory()); |
- WriteToDiscardableMemory(memory.get(), kPageSize); |
+ WriteToDiscardableAshmemChunk(memory.get(), kPageSize); |
} |
TEST_F(DiscardableMemoryAllocatorTest, FreeingWholeAshmemRegionClosesAshmem) { |
- scoped_ptr<DiscardableMemory> memory(allocator_.Allocate(kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory(allocator_.Allocate(kPageSize)); |
ASSERT_TRUE(memory); |
const int kMagic = 0xdeadbeef; |
*static_cast<int*>(memory->Memory()) = kMagic; |
@@ -132,11 +132,14 @@ TEST_F(DiscardableMemoryAllocatorTest, FreeingWholeAshmemRegionClosesAshmem) { |
} |
TEST_F(DiscardableMemoryAllocatorTest, AllocateUsesBestFitAlgorithm) { |
- scoped_ptr<DiscardableMemory> memory1(allocator_.Allocate(3 * kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory1( |
+ allocator_.Allocate(3 * kPageSize)); |
ASSERT_TRUE(memory1); |
- scoped_ptr<DiscardableMemory> memory2(allocator_.Allocate(2 * kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory2( |
+ allocator_.Allocate(2 * kPageSize)); |
ASSERT_TRUE(memory2); |
- scoped_ptr<DiscardableMemory> memory3(allocator_.Allocate(1 * kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory3( |
+ allocator_.Allocate(1 * kPageSize)); |
ASSERT_TRUE(memory3); |
void* const address_3 = memory3->Memory(); |
memory1.reset(); |
@@ -146,17 +149,17 @@ TEST_F(DiscardableMemoryAllocatorTest, AllocateUsesBestFitAlgorithm) { |
ASSERT_TRUE(memory1); |
// The chunk whose size is closest to the requested size should be reused. |
EXPECT_EQ(address_3, memory1->Memory()); |
- WriteToDiscardableMemory(memory1.get(), kPageSize); |
+ WriteToDiscardableAshmemChunk(memory1.get(), kPageSize); |
} |
TEST_F(DiscardableMemoryAllocatorTest, MergeFreeChunks) { |
- scoped_ptr<DiscardableMemory> memory1(allocator_.Allocate(kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory1(allocator_.Allocate(kPageSize)); |
ASSERT_TRUE(memory1); |
- scoped_ptr<DiscardableMemory> memory2(allocator_.Allocate(kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory2(allocator_.Allocate(kPageSize)); |
ASSERT_TRUE(memory2); |
- scoped_ptr<DiscardableMemory> memory3(allocator_.Allocate(kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory3(allocator_.Allocate(kPageSize)); |
ASSERT_TRUE(memory3); |
- scoped_ptr<DiscardableMemory> memory4(allocator_.Allocate(kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory4(allocator_.Allocate(kPageSize)); |
ASSERT_TRUE(memory4); |
void* const memory1_address = memory1->Memory(); |
memory1.reset(); |
@@ -169,9 +172,11 @@ TEST_F(DiscardableMemoryAllocatorTest, MergeFreeChunks) { |
} |
TEST_F(DiscardableMemoryAllocatorTest, MergeFreeChunksAdvanced) { |
- scoped_ptr<DiscardableMemory> memory1(allocator_.Allocate(4 * kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory1( |
+ allocator_.Allocate(4 * kPageSize)); |
ASSERT_TRUE(memory1); |
- scoped_ptr<DiscardableMemory> memory2(allocator_.Allocate(4 * kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory2( |
+ allocator_.Allocate(4 * kPageSize)); |
ASSERT_TRUE(memory2); |
void* const memory1_address = memory1->Memory(); |
memory1.reset(); |
@@ -186,14 +191,17 @@ TEST_F(DiscardableMemoryAllocatorTest, MergeFreeChunksAdvanced) { |
} |
TEST_F(DiscardableMemoryAllocatorTest, MergeFreeChunksAdvanced2) { |
- scoped_ptr<DiscardableMemory> memory1(allocator_.Allocate(4 * kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory1( |
+ allocator_.Allocate(4 * kPageSize)); |
ASSERT_TRUE(memory1); |
- scoped_ptr<DiscardableMemory> memory2(allocator_.Allocate(4 * kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory2( |
+ allocator_.Allocate(4 * kPageSize)); |
ASSERT_TRUE(memory2); |
void* const memory1_address = memory1->Memory(); |
memory1.reset(); |
memory1 = allocator_.Allocate(2 * kPageSize); |
- scoped_ptr<DiscardableMemory> memory3(allocator_.Allocate(2 * kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory3( |
+ allocator_.Allocate(2 * kPageSize)); |
// At this point, the region should be in this state: |
// 8 KBytes (used), 8 KBytes (used), 16 KBytes (used). |
memory3.reset(); |
@@ -207,13 +215,16 @@ TEST_F(DiscardableMemoryAllocatorTest, MergeFreeChunksAdvanced2) { |
} |
TEST_F(DiscardableMemoryAllocatorTest, MergeFreeChunksAndDeleteAshmemRegion) { |
- scoped_ptr<DiscardableMemory> memory1(allocator_.Allocate(4 * kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory1( |
+ allocator_.Allocate(4 * kPageSize)); |
ASSERT_TRUE(memory1); |
- scoped_ptr<DiscardableMemory> memory2(allocator_.Allocate(4 * kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory2( |
+ allocator_.Allocate(4 * kPageSize)); |
ASSERT_TRUE(memory2); |
memory1.reset(); |
memory1 = allocator_.Allocate(2 * kPageSize); |
- scoped_ptr<DiscardableMemory> memory3(allocator_.Allocate(2 * kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory3( |
+ allocator_.Allocate(2 * kPageSize)); |
// At this point, the region should be in this state: |
// 8 KBytes (used), 8 KBytes (used), 16 KBytes (used). |
memory1.reset(); |
@@ -232,9 +243,9 @@ TEST_F(DiscardableMemoryAllocatorTest, |
TooLargeFreeChunksDontCauseTooMuchFragmentationWhenRecycled) { |
// Keep |memory_1| below allocated so that the ashmem region doesn't get |
// closed when |memory_2| is deleted. |
- scoped_ptr<DiscardableMemory> memory_1(allocator_.Allocate(64 * 1024)); |
+ scoped_ptr<DiscardableAshmemChunk> memory_1(allocator_.Allocate(64 * 1024)); |
ASSERT_TRUE(memory_1); |
- scoped_ptr<DiscardableMemory> memory_2(allocator_.Allocate(32 * 1024)); |
+ scoped_ptr<DiscardableAshmemChunk> memory_2(allocator_.Allocate(32 * 1024)); |
ASSERT_TRUE(memory_2); |
void* const address = memory_2->Memory(); |
memory_2.reset(); |
@@ -242,40 +253,41 @@ TEST_F(DiscardableMemoryAllocatorTest, |
memory_2 = allocator_.Allocate(size); |
ASSERT_TRUE(memory_2); |
EXPECT_EQ(address, memory_2->Memory()); |
- WriteToDiscardableMemory(memory_2.get(), size); |
- scoped_ptr<DiscardableMemory> memory_3(allocator_.Allocate(size)); |
+ WriteToDiscardableAshmemChunk(memory_2.get(), size); |
+ scoped_ptr<DiscardableAshmemChunk> memory_3(allocator_.Allocate(size)); |
// The unused tail (16 KBytes large) of the previously freed chunk should be |
// reused. |
EXPECT_EQ(static_cast<char*>(address) + size, memory_3->Memory()); |
- WriteToDiscardableMemory(memory_3.get(), size); |
+ WriteToDiscardableAshmemChunk(memory_3.get(), size); |
} |
TEST_F(DiscardableMemoryAllocatorTest, UseMultipleAshmemRegions) { |
// Leave one page untouched at the end of the ashmem region. |
const size_t size = kAshmemRegionSizeForTesting - kPageSize; |
- scoped_ptr<DiscardableMemory> memory1(allocator_.Allocate(size)); |
+ scoped_ptr<DiscardableAshmemChunk> memory1(allocator_.Allocate(size)); |
ASSERT_TRUE(memory1); |
- WriteToDiscardableMemory(memory1.get(), size); |
+ WriteToDiscardableAshmemChunk(memory1.get(), size); |
- scoped_ptr<DiscardableMemory> memory2( |
+ scoped_ptr<DiscardableAshmemChunk> memory2( |
allocator_.Allocate(kAshmemRegionSizeForTesting)); |
ASSERT_TRUE(memory2); |
- WriteToDiscardableMemory(memory2.get(), kAshmemRegionSizeForTesting); |
+ WriteToDiscardableAshmemChunk(memory2.get(), kAshmemRegionSizeForTesting); |
// The last page of the first ashmem region should be used for this |
// allocation. |
- scoped_ptr<DiscardableMemory> memory3(allocator_.Allocate(kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory3(allocator_.Allocate(kPageSize)); |
ASSERT_TRUE(memory3); |
- WriteToDiscardableMemory(memory3.get(), kPageSize); |
+ WriteToDiscardableAshmemChunk(memory3.get(), kPageSize); |
EXPECT_EQ(memory3->Memory(), static_cast<char*>(memory1->Memory()) + size); |
} |
TEST_F(DiscardableMemoryAllocatorTest, |
HighestAllocatedChunkPointerIsUpdatedWhenHighestChunkGetsSplit) { |
// Prevents the ashmem region from getting closed when |memory2| gets freed. |
- scoped_ptr<DiscardableMemory> memory1(allocator_.Allocate(kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory1(allocator_.Allocate(kPageSize)); |
ASSERT_TRUE(memory1); |
- scoped_ptr<DiscardableMemory> memory2(allocator_.Allocate(4 * kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory2( |
+ allocator_.Allocate(4 * kPageSize)); |
ASSERT_TRUE(memory2); |
memory2.reset(); |
@@ -288,7 +300,8 @@ TEST_F(DiscardableMemoryAllocatorTest, |
// Allocate more than 3 * |kPageSize| so that the free chunk of size 3 * |
// |kPageSize| is not reused and |highest_allocated_chunk_| gets used instead. |
- scoped_ptr<DiscardableMemory> memory3(allocator_.Allocate(4 * kPageSize)); |
+ scoped_ptr<DiscardableAshmemChunk> memory3( |
+ allocator_.Allocate(4 * kPageSize)); |
ASSERT_TRUE(memory3); |
// Deleting |memory3| (whose size is 4 * |kPageSize|) should result in a merge |