Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(388)

Unified Diff: third_party/WebKit/Source/platform/heap/ThreadState.cpp

Issue 2619493003: Replace ASSERTs in platform/heap/ with DCHECKs
Patch Set: temp Created 3 years, 11 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
Index: third_party/WebKit/Source/platform/heap/ThreadState.cpp
diff --git a/third_party/WebKit/Source/platform/heap/ThreadState.cpp b/third_party/WebKit/Source/platform/heap/ThreadState.cpp
index 6ae1096dacf00a01271040cb4cd6b5fd044a5a1e..3636110dd37d812cbf12c4af86df45e864c40c0e 100644
--- a/third_party/WebKit/Source/platform/heap/ThreadState.cpp
+++ b/third_party/WebKit/Source/platform/heap/ThreadState.cpp
@@ -177,8 +177,8 @@ ThreadState::ThreadState(BlinkGC::ThreadHeapMode threadHeapMode)
m_allocatedObjectSize(0),
m_markedObjectSize(0),
m_reportedMemoryToV8(0) {
- ASSERT(checkThread());
- ASSERT(!**s_threadSpecific);
+ DCHECK(checkThread());
+ DCHECK(!**s_threadSpecific);
**s_threadSpecific = this;
switch (m_threadHeapMode) {
@@ -200,7 +200,7 @@ ThreadState::ThreadState(BlinkGC::ThreadHeapMode threadHeapMode)
m_heap = new ThreadHeap();
break;
}
- ASSERT(m_heap);
+ DCHECK(m_heap);
m_heap->attach(this);
for (int arenaIndex = 0; arenaIndex < BlinkGC::LargeObjectArenaIndex;
@@ -215,7 +215,7 @@ ThreadState::ThreadState(BlinkGC::ThreadHeapMode threadHeapMode)
}
ThreadState::~ThreadState() {
- ASSERT(checkThread());
+ DCHECK(checkThread());
for (int i = 0; i < BlinkGC::NumberOfArenas; ++i)
delete m_arenas[i];
@@ -243,7 +243,7 @@ size_t ThreadState::threadStackSize() {
Address stackStart =
reinterpret_cast<Address>(StackFrameDepth::getStackStart());
- RELEASE_ASSERT(stackStart && stackStart > stackEnd);
+ CHECK(stackStart && stackStart > stackEnd);
m_threadStackSize = static_cast<size_t>(stackStart - stackEnd);
// When the third last page of the reserved stack is accessed as a
// guard page, the second last page will be committed (along with removing
@@ -256,25 +256,25 @@ size_t ThreadState::threadStackSize() {
//
// http://blogs.msdn.com/b/satyem/archive/2012/08/13/thread-s-stack-memory-management.aspx
// explains the details.
- RELEASE_ASSERT(m_threadStackSize > 4 * 0x1000);
+ CHECK(m_threadStackSize > 4 * 0x1000);
m_threadStackSize -= 4 * 0x1000;
return m_threadStackSize;
}
#endif
void ThreadState::attachMainThread() {
- RELEASE_ASSERT(!ProcessHeap::s_shutdownComplete);
+ CHECK(!ProcessHeap::s_shutdownComplete);
s_threadSpecific = new WTF::ThreadSpecific<ThreadState*>();
new (s_mainThreadStateStorage) ThreadState(BlinkGC::MainThreadHeapMode);
}
void ThreadState::attachCurrentThread(BlinkGC::ThreadHeapMode threadHeapMode) {
- RELEASE_ASSERT(!ProcessHeap::s_shutdownComplete);
+ CHECK(!ProcessHeap::s_shutdownComplete);
new ThreadState(threadHeapMode);
}
void ThreadState::cleanupPages() {
- ASSERT(checkThread());
+ DCHECK(checkThread());
for (int i = 0; i < BlinkGC::NumberOfArenas; ++i)
m_arenas[i]->cleanupPages();
}
@@ -284,7 +284,7 @@ void ThreadState::runTerminationGC() {
cleanupPages();
return;
}
- ASSERT(checkThread());
+ DCHECK(checkThread());
// Finish sweeping.
completeSweep();
@@ -307,7 +307,7 @@ void ThreadState::runTerminationGC() {
// changes and is above zero.
int oldCount = -1;
int currentCount = getPersistentRegion()->numberOfPersistents();
- ASSERT(currentCount >= 0);
+ DCHECK_GE(currentCount, 0);
while (currentCount != oldCount) {
collectGarbageForTerminatingThread();
// Release the thread-local static persistents that were
@@ -317,11 +317,11 @@ void ThreadState::runTerminationGC() {
currentCount = getPersistentRegion()->numberOfPersistents();
}
// We should not have any persistents left when getting to this point,
- // if we have it is probably a bug so adding a debug ASSERT to catch this.
- ASSERT(!currentCount);
+ // if we have it is probably a bug so adding a debug DCHECK to catch this.
+ DCHECK(!currentCount);
// All of pre-finalizers should be consumed.
- ASSERT(m_orderedPreFinalizers.isEmpty());
- RELEASE_ASSERT(gcState() == NoGCScheduled);
+ DCHECK(m_orderedPreFinalizers.isEmpty());
+ CHECK(gcState() == NoGCScheduled);
// Add pages to the orphaned page pool to ensure any global GCs from this
// point on will not trace objects on this thread's arenas.
@@ -329,7 +329,7 @@ void ThreadState::runTerminationGC() {
}
void ThreadState::cleanupMainThread() {
- ASSERT(isMainThread());
+ DCHECK(isMainThread());
releaseStaticPersistentNodes();
@@ -349,7 +349,7 @@ void ThreadState::detachMainThread() {
// threadAttachMutex and waiting for other threads to pause or reach a
// safepoint.
ThreadState* state = mainThreadState();
- ASSERT(!state->isSweepingInProgress());
+ DCHECK(!state->isSweepingInProgress());
state->heap().detach(state);
state->~ThreadState();
@@ -358,7 +358,7 @@ void ThreadState::detachMainThread() {
void ThreadState::detachCurrentThread() {
ThreadState* state = current();
state->heap().detach(state);
- RELEASE_ASSERT(state->gcState() == ThreadState::NoGCScheduled);
+ CHECK(state->gcState() == ThreadState::NoGCScheduled);
delete state;
}
@@ -456,7 +456,7 @@ void ThreadState::pushThreadLocalWeakCallback(void* object,
}
bool ThreadState::popAndInvokeThreadLocalWeakCallback(Visitor* visitor) {
- ASSERT(checkThread());
+ DCHECK(checkThread());
// For weak processing we should never reach orphaned pages since orphaned
// pages are not traced and thus objects on those pages are never be
// registered as objects on orphaned pages. We cannot assert this here since
@@ -470,8 +470,8 @@ bool ThreadState::popAndInvokeThreadLocalWeakCallback(Visitor* visitor) {
}
void ThreadState::threadLocalWeakProcessing() {
- ASSERT(checkThread());
- ASSERT(!sweepForbidden());
+ DCHECK(checkThread());
+ DCHECK(!sweepForbidden());
TRACE_EVENT0("blink_gc", "ThreadState::threadLocalWeakProcessing");
double startTime = WTF::currentTimeMS();
@@ -621,7 +621,7 @@ bool ThreadState::shouldForceMemoryPressureGC() {
}
void ThreadState::scheduleV8FollowupGCIfNeeded(BlinkGC::V8GCType gcType) {
- ASSERT(checkThread());
+ DCHECK(checkThread());
ThreadHeap::reportMemoryUsageForTracing();
#if PRINT_HEAP_STATS
@@ -635,8 +635,8 @@ void ThreadState::scheduleV8FollowupGCIfNeeded(BlinkGC::V8GCType gcType) {
// This completeSweep() will do nothing in common cases since we've
// called completeSweep() before V8 starts minor/major GCs.
completeSweep();
- ASSERT(!isSweepingInProgress());
- ASSERT(!sweepForbidden());
+ DCHECK(!isSweepingInProgress());
+ DCHECK(!sweepForbidden());
if ((gcType == BlinkGC::V8MajorGC && shouldForceMemoryPressureGC()) ||
shouldScheduleV8FollowupGC()) {
@@ -669,7 +669,7 @@ void ThreadState::willStartV8GC(BlinkGC::V8GCType gcType) {
void ThreadState::schedulePageNavigationGCIfNeeded(
float estimatedRemovalRatio) {
- ASSERT(checkThread());
+ DCHECK(checkThread());
ThreadHeap::reportMemoryUsageForTracing();
#if PRINT_HEAP_STATS
@@ -686,8 +686,8 @@ void ThreadState::schedulePageNavigationGCIfNeeded(
// TODO(haraken): It might not make sense to force completeSweep() for all
// page navigations.
completeSweep();
- ASSERT(!isSweepingInProgress());
- ASSERT(!sweepForbidden());
+ DCHECK(!isSweepingInProgress());
+ DCHECK(!sweepForbidden());
if (shouldForceMemoryPressureGC()) {
#if PRINT_HEAP_STATS
@@ -706,13 +706,13 @@ void ThreadState::schedulePageNavigationGCIfNeeded(
}
void ThreadState::schedulePageNavigationGC() {
- ASSERT(checkThread());
- ASSERT(!isSweepingInProgress());
+ DCHECK(checkThread());
+ DCHECK(!isSweepingInProgress());
setGCState(PageNavigationGCScheduled);
}
void ThreadState::scheduleGCIfNeeded() {
- ASSERT(checkThread());
+ DCHECK(checkThread());
ThreadHeap::reportMemoryUsageForTracing();
#if PRINT_HEAP_STATS
@@ -726,7 +726,7 @@ void ThreadState::scheduleGCIfNeeded() {
if (isSweepingInProgress())
return;
- ASSERT(!sweepForbidden());
+ DCHECK(!sweepForbidden());
reportMemoryToV8();
@@ -763,17 +763,17 @@ void ThreadState::scheduleGCIfNeeded() {
}
ThreadState* ThreadState::fromObject(const void* object) {
- ASSERT(object);
+ DCHECK(object);
BasePage* page = pageFromObject(object);
- ASSERT(page);
- ASSERT(page->arena());
+ DCHECK(page);
+ DCHECK(page->arena());
return page->arena()->getThreadState();
}
void ThreadState::performIdleGC(double deadlineSeconds) {
- ASSERT(checkThread());
- ASSERT(isMainThread());
- ASSERT(Platform::current()->currentThread()->scheduler());
+ DCHECK(checkThread());
+ DCHECK(isMainThread());
+ DCHECK(Platform::current()->currentThread()->scheduler());
if (gcState() != IdleGCScheduled)
return;
@@ -804,8 +804,8 @@ void ThreadState::performIdleGC(double deadlineSeconds) {
}
void ThreadState::performIdleLazySweep(double deadlineSeconds) {
- ASSERT(checkThread());
- ASSERT(isMainThread());
+ DCHECK(checkThread());
+ DCHECK(isMainThread());
// If we are not in a sweeping phase, there is nothing to do here.
if (!isSweepingInProgress())
@@ -885,7 +885,7 @@ void ThreadState::scheduleIdleLazySweep() {
}
void ThreadState::schedulePreciseGC() {
- ASSERT(checkThread());
+ DCHECK(checkThread());
if (isSweepingInProgress()) {
setGCState(SweepingAndPreciseGCScheduled);
return;
@@ -914,7 +914,7 @@ void unexpectedGCState(ThreadState::GCState gcState) {
UNEXPECTED_GCSTATE(SweepingAndIdleGCScheduled);
UNEXPECTED_GCSTATE(SweepingAndPreciseGCScheduled);
default:
- ASSERT_NOT_REACHED();
+ NOTREACHED();
return;
}
}
@@ -924,13 +924,14 @@ void unexpectedGCState(ThreadState::GCState gcState) {
} // namespace
#define VERIFY_STATE_TRANSITION(condition) \
- if (UNLIKELY(!(condition))) \
- unexpectedGCState(m_gcState)
+ if (UNLIKELY(!(condition))) { \
+ unexpectedGCState(m_gcState); \
+ }
void ThreadState::setGCState(GCState gcState) {
switch (gcState) {
case NoGCScheduled:
- ASSERT(checkThread());
+ DCHECK(checkThread());
VERIFY_STATE_TRANSITION(m_gcState == Sweeping ||
m_gcState == SweepingAndIdleGCScheduled);
break;
@@ -938,7 +939,7 @@ void ThreadState::setGCState(GCState gcState) {
case PreciseGCScheduled:
case FullGCScheduled:
case PageNavigationGCScheduled:
- ASSERT(checkThread());
+ DCHECK(checkThread());
VERIFY_STATE_TRANSITION(
m_gcState == NoGCScheduled || m_gcState == IdleGCScheduled ||
m_gcState == PreciseGCScheduled || m_gcState == FullGCScheduled ||
@@ -948,28 +949,28 @@ void ThreadState::setGCState(GCState gcState) {
completeSweep();
break;
case GCRunning:
- ASSERT(!isInGC());
+ DCHECK(!isInGC());
VERIFY_STATE_TRANSITION(m_gcState != GCRunning);
break;
case EagerSweepScheduled:
case LazySweepScheduled:
- ASSERT(isInGC());
+ DCHECK(isInGC());
VERIFY_STATE_TRANSITION(m_gcState == GCRunning);
break;
case Sweeping:
- ASSERT(checkThread());
+ DCHECK(checkThread());
VERIFY_STATE_TRANSITION(m_gcState == EagerSweepScheduled ||
m_gcState == LazySweepScheduled);
break;
case SweepingAndIdleGCScheduled:
case SweepingAndPreciseGCScheduled:
- ASSERT(checkThread());
+ DCHECK(checkThread());
VERIFY_STATE_TRANSITION(m_gcState == Sweeping ||
m_gcState == SweepingAndIdleGCScheduled ||
m_gcState == SweepingAndPreciseGCScheduled);
break;
default:
- ASSERT_NOT_REACHED();
+ NOTREACHED();
}
m_gcState = gcState;
}
@@ -977,7 +978,7 @@ void ThreadState::setGCState(GCState gcState) {
#undef VERIFY_STATE_TRANSITION
void ThreadState::runScheduledGC(BlinkGC::StackState stackState) {
- ASSERT(checkThread());
+ DCHECK(checkThread());
if (stackState != BlinkGC::NoHeapPointersOnStack)
return;
@@ -1017,7 +1018,7 @@ void ThreadState::flushHeapDoesNotContainCacheIfNeeded() {
}
void ThreadState::makeConsistentForGC() {
- ASSERT(isInGC());
+ DCHECK(isInGC());
TRACE_EVENT0("blink_gc", "ThreadState::makeConsistentForGC");
for (int i = 0; i < BlinkGC::NumberOfArenas; ++i)
m_arenas[i]->makeConsistentForGC();
@@ -1052,7 +1053,7 @@ void ThreadState::compact() {
}
void ThreadState::makeConsistentForMutator() {
- ASSERT(isInGC());
+ DCHECK(isInGC());
for (int i = 0; i < BlinkGC::NumberOfArenas; ++i)
m_arenas[i]->makeConsistentForMutator();
}
@@ -1062,7 +1063,7 @@ void ThreadState::preGC() {
m_performCleanup)
m_performCleanup(m_isolate);
- ASSERT(!isInGC());
+ DCHECK(!isInGC());
setGCState(GCRunning);
makeConsistentForGC();
flushHeapDoesNotContainCacheIfNeeded();
@@ -1091,7 +1092,7 @@ void ThreadState::postGC(BlinkGC::GCType gcType) {
m_invalidateDeadObjectsInWrappersMarkingDeque(m_isolate);
}
- ASSERT(isInGC());
+ DCHECK(isInGC());
for (int i = 0; i < BlinkGC::NumberOfArenas; i++)
m_arenas[i]->prepareForSweep();
@@ -1114,7 +1115,7 @@ void ThreadState::postGC(BlinkGC::GCType gcType) {
}
void ThreadState::preSweep() {
- ASSERT(checkThread());
+ DCHECK(checkThread());
if (gcState() != EagerSweepScheduled && gcState() != LazySweepScheduled)
return;
@@ -1184,11 +1185,11 @@ void ThreadState::eagerSweep() {
#if defined(ADDRESS_SANITIZER)
poisonEagerArena();
#endif
- ASSERT(checkThread());
+ DCHECK(checkThread());
// Some objects need to be finalized promptly and cannot be handled
// by lazy sweeping. Keep those in a designated heap and sweep it
// eagerly.
- ASSERT(isSweepingInProgress());
+ DCHECK(isSweepingInProgress());
// Mirroring the completeSweep() condition; see its comment.
if (sweepForbidden())
@@ -1203,7 +1204,7 @@ void ThreadState::eagerSweep() {
}
void ThreadState::completeSweep() {
- ASSERT(checkThread());
+ DCHECK(checkThread());
// If we are not in a sweeping phase, there is nothing to do here.
if (!isSweepingInProgress())
return;
@@ -1240,7 +1241,7 @@ void ThreadState::completeSweep() {
}
void ThreadState::postSweep() {
- ASSERT(checkThread());
+ DCHECK(checkThread());
ThreadHeap::reportMemoryUsageForTracing();
if (isMainThread()) {
@@ -1311,17 +1312,17 @@ void ThreadState::postSweep() {
scheduleIdleGC();
break;
default:
- ASSERT_NOT_REACHED();
+ NOTREACHED();
}
}
void ThreadState::prepareForThreadStateTermination() {
- ASSERT(checkThread());
+ DCHECK(checkThread());
for (int i = 0; i < BlinkGC::NumberOfArenas; ++i)
m_arenas[i]->prepareHeapForTermination();
}
-#if ENABLE(ASSERT)
+#if DCHECK_IS_ON()
BasePage* ThreadState::findPageFromAddress(Address address) {
for (int i = 0; i < BlinkGC::NumberOfArenas; ++i) {
if (BasePage* page = m_arenas[i]->findPageFromAddress(address))
@@ -1339,11 +1340,11 @@ size_t ThreadState::objectPayloadSizeForTesting() {
}
void ThreadState::safePoint(BlinkGC::StackState stackState) {
- ASSERT(checkThread());
+ DCHECK(checkThread());
ThreadHeap::reportMemoryUsageForTracing();
runScheduledGC(stackState);
- ASSERT(!m_atSafePoint);
+ DCHECK(!m_atSafePoint);
m_stackState = stackState;
m_atSafePoint = true;
m_heap->checkAndPark(this, nullptr);
@@ -1363,7 +1364,7 @@ NO_SANITIZE_ADDRESS static void* adjustScopeMarkerForAdressSanitizer(
void* scopeMarker) {
Address start = reinterpret_cast<Address>(StackFrameDepth::getStackStart());
Address end = reinterpret_cast<Address>(&start);
- RELEASE_ASSERT(end < start);
+ CHECK(end < start);
if (end <= scopeMarker && scopeMarker < start)
return scopeMarker;
@@ -1379,14 +1380,14 @@ NO_SANITIZE_ADDRESS static void* adjustScopeMarkerForAdressSanitizer(
void ThreadState::enterSafePoint(BlinkGC::StackState stackState,
void* scopeMarker) {
- ASSERT(checkThread());
+ DCHECK(checkThread());
#ifdef ADDRESS_SANITIZER
if (stackState == BlinkGC::HeapPointersOnStack)
scopeMarker = adjustScopeMarkerForAdressSanitizer(scopeMarker);
#endif
- ASSERT(stackState == BlinkGC::NoHeapPointersOnStack || scopeMarker);
+ DCHECK(stackState == BlinkGC::NoHeapPointersOnStack || scopeMarker);
runScheduledGC(stackState);
- ASSERT(!m_atSafePoint);
+ DCHECK(!m_atSafePoint);
m_atSafePoint = true;
m_stackState = stackState;
m_safePointScopeMarker = scopeMarker;
@@ -1394,8 +1395,8 @@ void ThreadState::enterSafePoint(BlinkGC::StackState stackState,
}
void ThreadState::leaveSafePoint(SafePointAwareMutexLocker* locker) {
- ASSERT(checkThread());
- ASSERT(m_atSafePoint);
+ DCHECK(checkThread());
+ DCHECK(m_atSafePoint);
m_heap->leaveSafePoint(this, locker);
m_atSafePoint = false;
m_stackState = BlinkGC::HeapPointersOnStack;
@@ -1440,19 +1441,19 @@ void ThreadState::copyStackUntilSafePointScope() {
Address* to = reinterpret_cast<Address*>(m_safePointScopeMarker);
Address* from = reinterpret_cast<Address*>(m_endOfStack);
- RELEASE_ASSERT(from < to);
- RELEASE_ASSERT(to <= reinterpret_cast<Address*>(m_startOfStack));
+ CHECK(from < to);
+ CHECK(to <= reinterpret_cast<Address*>(m_startOfStack));
size_t slotCount = static_cast<size_t>(to - from);
// Catch potential performance issues.
#if defined(LEAK_SANITIZER) || defined(ADDRESS_SANITIZER)
// ASan/LSan use more space on the stack and we therefore
// increase the allowed stack copying for those builds.
- ASSERT(slotCount < 2048);
+ DCHECK_LT(slotCount, 2048UL);
#else
- ASSERT(slotCount < 1024);
+ DCHECK_LT(slotCount, 1024UL);
#endif
- ASSERT(!m_safePointStackCopy.size());
+ DCHECK(!m_safePointStackCopy.size());
m_safePointStackCopy.resize(slotCount);
for (size_t i = 0; i < slotCount; ++i) {
m_safePointStackCopy[i] = from[i];
@@ -1461,7 +1462,7 @@ void ThreadState::copyStackUntilSafePointScope() {
void ThreadState::addInterruptor(
std::unique_ptr<BlinkGCInterruptor> interruptor) {
- ASSERT(checkThread());
+ DCHECK(checkThread());
SafePointScope scope(BlinkGC::HeapPointersOnStack);
{
MutexLocker locker(m_heap->threadAttachMutex());
@@ -1477,7 +1478,7 @@ void ThreadState::registerStaticPersistentNode(
return;
#endif
- ASSERT(!m_staticPersistents.contains(node));
+ DCHECK(!m_staticPersistents.contains(node));
m_staticPersistents.add(node, callback);
}
@@ -1499,7 +1500,7 @@ void ThreadState::freePersistentNode(PersistentNode* persistentNode) {
//
// There's no fundamental reason why this couldn't be supported,
// but no known use for it.
- ASSERT(!m_staticPersistents.contains(persistentNode));
+ DCHECK(!m_staticPersistents.contains(persistentNode));
}
#if defined(LEAK_SANITIZER)
@@ -1508,7 +1509,7 @@ void ThreadState::enterStaticReferenceRegistrationDisabledScope() {
}
void ThreadState::leaveStaticReferenceRegistrationDisabledScope() {
- ASSERT(m_disabledStaticPersistentsRegistration);
+ DCHECK(m_disabledStaticPersistentsRegistration);
m_disabledStaticPersistentsRegistration--;
}
#endif
@@ -1522,8 +1523,8 @@ void ThreadState::unlockThreadAttachMutex() {
}
void ThreadState::invokePreFinalizers() {
- ASSERT(checkThread());
- ASSERT(!sweepForbidden());
+ DCHECK(checkThread());
+ DCHECK(!sweepForbidden());
TRACE_EVENT0("blink_gc", "ThreadState::invokePreFinalizers");
double startTime = WTF::currentTimeMS();
@@ -1576,12 +1577,12 @@ int ThreadState::arenaIndexOfVectorArenaLeastRecentlyExpanded(
arenaIndexWithMinArenaAge = arenaIndex;
}
}
- ASSERT(isVectorArenaIndex(arenaIndexWithMinArenaAge));
+ DCHECK(isVectorArenaIndex(arenaIndexWithMinArenaAge));
return arenaIndexWithMinArenaAge;
}
BaseArena* ThreadState::expandedVectorBackingArena(size_t gcInfoIndex) {
- ASSERT(checkThread());
+ DCHECK(checkThread());
size_t entryIndex = gcInfoIndex & likelyToBePromptlyFreedArrayMask;
--m_likelyToBePromptlyFreed[entryIndex];
int arenaIndex = m_vectorBackingArenaIndex;
@@ -1599,14 +1600,14 @@ void ThreadState::allocationPointAdjusted(int arenaIndex) {
}
void ThreadState::promptlyFreed(size_t gcInfoIndex) {
- ASSERT(checkThread());
+ DCHECK(checkThread());
size_t entryIndex = gcInfoIndex & likelyToBePromptlyFreedArrayMask;
// See the comment in vectorBackingArena() for why this is +3.
m_likelyToBePromptlyFreed[entryIndex] += 3;
}
void ThreadState::takeSnapshot(SnapshotType type) {
- ASSERT(isInGC());
+ DCHECK(isInGC());
// 0 is used as index for freelist entries. Objects are indexed 1 to
// gcInfoIndex.
@@ -1630,7 +1631,7 @@ void ThreadState::takeSnapshot(SnapshotType type) {
heapsDumpName + "/" #ArenaType); \
break; \
default: \
- ASSERT_NOT_REACHED(); \
+ NOTREACHED(); \
} \
}
@@ -1648,7 +1649,7 @@ void ThreadState::takeSnapshot(SnapshotType type) {
SNAPSHOT_HEAP(LargeObject);
FOR_EACH_TYPED_ARENA(SNAPSHOT_HEAP);
- ASSERT(numberOfHeapsReported == BlinkGC::NumberOfArenas);
+ DCHECK_EQ(numberOfHeapsReported, BlinkGC::NumberOfArenas);
#undef SNAPSHOT_HEAP
@@ -1690,7 +1691,7 @@ void ThreadState::collectGarbage(BlinkGC::StackState stackState,
BlinkGC::GCType gcType,
BlinkGC::GCReason reason) {
// Nested collectGarbage() invocations aren't supported.
- RELEASE_ASSERT(!isGCForbidden());
+ CHECK(!isGCForbidden());
completeSweep();
GCForbiddenScope gcForbiddenScope(this);

Powered by Google App Engine
This is Rietveld 408576698