Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(1349)

Unified Diff: base/debug/trace_event_impl.cc

Issue 614103004: replace 'virtual ... OVERRIDE' with '... override' (Closed) Base URL: https://chromium.googlesource.com/chromium/src.git@master
Patch Set: process base/ Created 6 years, 3 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
Index: base/debug/trace_event_impl.cc
diff --git a/base/debug/trace_event_impl.cc b/base/debug/trace_event_impl.cc
index db321e2bfa42c893e34acce70c37d802459c0e6f..28c0c0d962fdbdd2d5de5f4650bca61880ff7ee2 100644
--- a/base/debug/trace_event_impl.cc
+++ b/base/debug/trace_event_impl.cc
@@ -140,7 +140,7 @@ class TraceBufferRingBuffer : public TraceBuffer {
recyclable_chunks_queue_[i] = i;
}
- virtual scoped_ptr<TraceBufferChunk> GetChunk(size_t* index) OVERRIDE {
+ scoped_ptr<TraceBufferChunk> GetChunk(size_t* index) override {
// Because the number of threads is much less than the number of chunks,
// the queue should never be empty.
DCHECK(!QueueIsEmpty());
@@ -162,8 +162,7 @@ class TraceBufferRingBuffer : public TraceBuffer {
return scoped_ptr<TraceBufferChunk>(chunk);
}
- virtual void ReturnChunk(size_t index,
- scoped_ptr<TraceBufferChunk> chunk) OVERRIDE {
+ void ReturnChunk(size_t index, scoped_ptr<TraceBufferChunk> chunk) override {
// When this method is called, the queue should not be full because it
// can contain all chunks including the one to be returned.
DCHECK(!QueueIsFull());
@@ -175,20 +174,18 @@ class TraceBufferRingBuffer : public TraceBuffer {
queue_tail_ = NextQueueIndex(queue_tail_);
}
- virtual bool IsFull() const OVERRIDE {
- return false;
- }
+ bool IsFull() const override { return false; }
- virtual size_t Size() const OVERRIDE {
+ size_t Size() const override {
// This is approximate because not all of the chunks are full.
return chunks_.size() * kTraceBufferChunkSize;
}
- virtual size_t Capacity() const OVERRIDE {
+ size_t Capacity() const override {
return max_chunks_ * kTraceBufferChunkSize;
}
- virtual TraceEvent* GetEventByHandle(TraceEventHandle handle) OVERRIDE {
+ TraceEvent* GetEventByHandle(TraceEventHandle handle) override {
if (handle.chunk_index >= chunks_.size())
return NULL;
TraceBufferChunk* chunk = chunks_[handle.chunk_index];
@@ -197,7 +194,7 @@ class TraceBufferRingBuffer : public TraceBuffer {
return chunk->GetEventAt(handle.event_index);
}
- virtual const TraceBufferChunk* NextChunk() OVERRIDE {
+ const TraceBufferChunk* NextChunk() override {
if (chunks_.empty())
return NULL;
@@ -212,7 +209,7 @@ class TraceBufferRingBuffer : public TraceBuffer {
return NULL;
}
- virtual scoped_ptr<TraceBuffer> CloneForIteration() const OVERRIDE {
+ scoped_ptr<TraceBuffer> CloneForIteration() const override {
scoped_ptr<ClonedTraceBuffer> cloned_buffer(new ClonedTraceBuffer());
for (size_t queue_index = queue_head_; queue_index != queue_tail_;
queue_index = NextQueueIndex(queue_index)) {
@@ -231,26 +228,25 @@ class TraceBufferRingBuffer : public TraceBuffer {
ClonedTraceBuffer() : current_iteration_index_(0) {}
// The only implemented method.
- virtual const TraceBufferChunk* NextChunk() OVERRIDE {
+ const TraceBufferChunk* NextChunk() override {
return current_iteration_index_ < chunks_.size() ?
chunks_[current_iteration_index_++] : NULL;
}
- virtual scoped_ptr<TraceBufferChunk> GetChunk(size_t* index) OVERRIDE {
+ scoped_ptr<TraceBufferChunk> GetChunk(size_t* index) override {
NOTIMPLEMENTED();
return scoped_ptr<TraceBufferChunk>();
}
- virtual void ReturnChunk(size_t index,
- scoped_ptr<TraceBufferChunk>) OVERRIDE {
+ void ReturnChunk(size_t index, scoped_ptr<TraceBufferChunk>) override {
NOTIMPLEMENTED();
}
- virtual bool IsFull() const OVERRIDE { return false; }
- virtual size_t Size() const OVERRIDE { return 0; }
- virtual size_t Capacity() const OVERRIDE { return 0; }
- virtual TraceEvent* GetEventByHandle(TraceEventHandle handle) OVERRIDE {
+ bool IsFull() const override { return false; }
+ size_t Size() const override { return 0; }
+ size_t Capacity() const override { return 0; }
+ TraceEvent* GetEventByHandle(TraceEventHandle handle) override {
return NULL;
}
- virtual scoped_ptr<TraceBuffer> CloneForIteration() const OVERRIDE {
+ scoped_ptr<TraceBuffer> CloneForIteration() const override {
NOTIMPLEMENTED();
return scoped_ptr<TraceBuffer>();
}
@@ -306,7 +302,7 @@ class TraceBufferVector : public TraceBuffer {
chunks_.reserve(max_chunks_);
}
- virtual scoped_ptr<TraceBufferChunk> GetChunk(size_t* index) OVERRIDE {
+ scoped_ptr<TraceBufferChunk> GetChunk(size_t* index) override {
// This function may be called when adding normal events or indirectly from
// AddMetadataEventsWhileLocked(). We can not DECHECK(!IsFull()) because we
// have to add the metadata events and flush thread-local buffers even if
@@ -319,8 +315,7 @@ class TraceBufferVector : public TraceBuffer {
new TraceBufferChunk(static_cast<uint32>(*index) + 1));
}
- virtual void ReturnChunk(size_t index,
- scoped_ptr<TraceBufferChunk> chunk) OVERRIDE {
+ void ReturnChunk(size_t index, scoped_ptr<TraceBufferChunk> chunk) override {
DCHECK_GT(in_flight_chunk_count_, 0u);
DCHECK_LT(index, chunks_.size());
DCHECK(!chunks_[index]);
@@ -328,20 +323,18 @@ class TraceBufferVector : public TraceBuffer {
chunks_[index] = chunk.release();
}
- virtual bool IsFull() const OVERRIDE {
- return chunks_.size() >= max_chunks_;
- }
+ bool IsFull() const override { return chunks_.size() >= max_chunks_; }
- virtual size_t Size() const OVERRIDE {
+ size_t Size() const override {
// This is approximate because not all of the chunks are full.
return chunks_.size() * kTraceBufferChunkSize;
}
- virtual size_t Capacity() const OVERRIDE {
+ size_t Capacity() const override {
return max_chunks_ * kTraceBufferChunkSize;
}
- virtual TraceEvent* GetEventByHandle(TraceEventHandle handle) OVERRIDE {
+ TraceEvent* GetEventByHandle(TraceEventHandle handle) override {
if (handle.chunk_index >= chunks_.size())
return NULL;
TraceBufferChunk* chunk = chunks_[handle.chunk_index];
@@ -350,7 +343,7 @@ class TraceBufferVector : public TraceBuffer {
return chunk->GetEventAt(handle.event_index);
}
- virtual const TraceBufferChunk* NextChunk() OVERRIDE {
+ const TraceBufferChunk* NextChunk() override {
while (current_iteration_index_ < chunks_.size()) {
// Skip in-flight chunks.
const TraceBufferChunk* chunk = chunks_[current_iteration_index_++];
@@ -360,7 +353,7 @@ class TraceBufferVector : public TraceBuffer {
return NULL;
}
- virtual scoped_ptr<TraceBuffer> CloneForIteration() const OVERRIDE {
+ scoped_ptr<TraceBuffer> CloneForIteration() const override {
NOTIMPLEMENTED();
return scoped_ptr<TraceBuffer>();
}
@@ -869,7 +862,7 @@ class TraceSamplingThread : public PlatformThread::Delegate {
virtual ~TraceSamplingThread();
// Implementation of PlatformThread::Delegate:
- virtual void ThreadMain() OVERRIDE;
+ void ThreadMain() override;
static void DefaultSamplingCallback(TraceBucketData* bucekt_data);
@@ -1066,7 +1059,7 @@ class TraceLog::ThreadLocalEventBuffer
private:
// MessageLoop::DestructionObserver
- virtual void WillDestroyCurrentMessageLoop() OVERRIDE;
+ void WillDestroyCurrentMessageLoop() override;
void FlushWhileLocked();

Powered by Google App Engine
This is Rietveld 408576698