Index: cc/resources/one_copy_tile_task_worker_pool.cc |
diff --git a/cc/resources/one_copy_tile_task_worker_pool.cc b/cc/resources/one_copy_tile_task_worker_pool.cc |
deleted file mode 100644 |
index 23f372c15e2b13add87aacaa046b6a62528ac335..0000000000000000000000000000000000000000 |
--- a/cc/resources/one_copy_tile_task_worker_pool.cc |
+++ /dev/null |
@@ -1,511 +0,0 @@ |
-// Copyright 2014 The Chromium Authors. All rights reserved. |
-// Use of this source code is governed by a BSD-style license that can be |
-// found in the LICENSE file. |
- |
-#include "cc/resources/one_copy_tile_task_worker_pool.h" |
- |
-#include <algorithm> |
-#include <limits> |
- |
-#include "base/strings/stringprintf.h" |
-#include "base/trace_event/trace_event.h" |
-#include "base/trace_event/trace_event_argument.h" |
-#include "cc/debug/traced_value.h" |
-#include "cc/resources/raster_buffer.h" |
-#include "cc/resources/resource_pool.h" |
-#include "cc/resources/scoped_resource.h" |
-#include "gpu/command_buffer/client/gles2_interface.h" |
-#include "ui/gfx/gpu_memory_buffer.h" |
- |
-namespace cc { |
-namespace { |
- |
-class RasterBufferImpl : public RasterBuffer { |
- public: |
- RasterBufferImpl(OneCopyTileTaskWorkerPool* worker_pool, |
- ResourceProvider* resource_provider, |
- ResourcePool* resource_pool, |
- ResourceFormat resource_format, |
- const Resource* resource) |
- : worker_pool_(worker_pool), |
- resource_provider_(resource_provider), |
- resource_pool_(resource_pool), |
- resource_(resource), |
- raster_resource_( |
- resource_pool->AcquireResource(resource->size(), resource_format)), |
- lock_(new ResourceProvider::ScopedWriteLockGpuMemoryBuffer( |
- resource_provider_, |
- raster_resource_->id())), |
- sequence_(0) {} |
- |
- ~RasterBufferImpl() override { |
- // Release write lock in case a copy was never scheduled. |
- lock_.reset(); |
- |
- // Make sure any scheduled copy operations are issued before we release the |
- // raster resource. |
- if (sequence_) |
- worker_pool_->AdvanceLastIssuedCopyTo(sequence_); |
- |
- // Return raster resource to pool so it can be used by another RasterBuffer |
- // instance. |
- if (raster_resource_) |
- resource_pool_->ReleaseResource(raster_resource_.Pass()); |
- } |
- |
- // Overridden from RasterBuffer: |
- void Playback(const RasterSource* raster_source, |
- const gfx::Rect& rect, |
- float scale) override { |
- sequence_ = worker_pool_->PlaybackAndScheduleCopyOnWorkerThread( |
- lock_.Pass(), raster_resource_.Pass(), resource_, raster_source, rect, |
- scale); |
- } |
- |
- private: |
- OneCopyTileTaskWorkerPool* worker_pool_; |
- ResourceProvider* resource_provider_; |
- ResourcePool* resource_pool_; |
- const Resource* resource_; |
- scoped_ptr<ScopedResource> raster_resource_; |
- scoped_ptr<ResourceProvider::ScopedWriteLockGpuMemoryBuffer> lock_; |
- CopySequenceNumber sequence_; |
- |
- DISALLOW_COPY_AND_ASSIGN(RasterBufferImpl); |
-}; |
- |
-// Flush interval when performing copy operations. |
-const int kCopyFlushPeriod = 4; |
- |
-// Number of in-flight copy operations to allow. |
-const int kMaxCopyOperations = 32; |
- |
-// Delay been checking for copy operations to complete. |
-const int kCheckForCompletedCopyOperationsTickRateMs = 1; |
- |
-// Number of failed attempts to allow before we perform a check that will |
-// wait for copy operations to complete if needed. |
-const int kFailedAttemptsBeforeWaitIfNeeded = 256; |
- |
-} // namespace |
- |
-OneCopyTileTaskWorkerPool::CopyOperation::CopyOperation( |
- scoped_ptr<ResourceProvider::ScopedWriteLockGpuMemoryBuffer> write_lock, |
- scoped_ptr<ScopedResource> src, |
- const Resource* dst) |
- : write_lock(write_lock.Pass()), src(src.Pass()), dst(dst) { |
-} |
- |
-OneCopyTileTaskWorkerPool::CopyOperation::~CopyOperation() { |
-} |
- |
-// static |
-scoped_ptr<TileTaskWorkerPool> OneCopyTileTaskWorkerPool::Create( |
- base::SequencedTaskRunner* task_runner, |
- TaskGraphRunner* task_graph_runner, |
- ContextProvider* context_provider, |
- ResourceProvider* resource_provider, |
- ResourcePool* resource_pool) { |
- return make_scoped_ptr<TileTaskWorkerPool>(new OneCopyTileTaskWorkerPool( |
- task_runner, task_graph_runner, context_provider, resource_provider, |
- resource_pool)); |
-} |
- |
-OneCopyTileTaskWorkerPool::OneCopyTileTaskWorkerPool( |
- base::SequencedTaskRunner* task_runner, |
- TaskGraphRunner* task_graph_runner, |
- ContextProvider* context_provider, |
- ResourceProvider* resource_provider, |
- ResourcePool* resource_pool) |
- : task_runner_(task_runner), |
- task_graph_runner_(task_graph_runner), |
- namespace_token_(task_graph_runner->GetNamespaceToken()), |
- context_provider_(context_provider), |
- resource_provider_(resource_provider), |
- resource_pool_(resource_pool), |
- last_issued_copy_operation_(0), |
- last_flushed_copy_operation_(0), |
- lock_(), |
- copy_operation_count_cv_(&lock_), |
- scheduled_copy_operation_count_(0), |
- issued_copy_operation_count_(0), |
- next_copy_operation_sequence_(1), |
- check_for_completed_copy_operations_pending_(false), |
- shutdown_(false), |
- weak_ptr_factory_(this), |
- task_set_finished_weak_ptr_factory_(this) { |
- DCHECK(context_provider_); |
-} |
- |
-OneCopyTileTaskWorkerPool::~OneCopyTileTaskWorkerPool() { |
- DCHECK_EQ(scheduled_copy_operation_count_, 0u); |
-} |
- |
-TileTaskRunner* OneCopyTileTaskWorkerPool::AsTileTaskRunner() { |
- return this; |
-} |
- |
-void OneCopyTileTaskWorkerPool::SetClient(TileTaskRunnerClient* client) { |
- client_ = client; |
-} |
- |
-void OneCopyTileTaskWorkerPool::Shutdown() { |
- TRACE_EVENT0("cc", "OneCopyTileTaskWorkerPool::Shutdown"); |
- |
- { |
- base::AutoLock lock(lock_); |
- |
- shutdown_ = true; |
- copy_operation_count_cv_.Signal(); |
- } |
- |
- TaskGraph empty; |
- task_graph_runner_->ScheduleTasks(namespace_token_, &empty); |
- task_graph_runner_->WaitForTasksToFinishRunning(namespace_token_); |
-} |
- |
-void OneCopyTileTaskWorkerPool::ScheduleTasks(TileTaskQueue* queue) { |
- TRACE_EVENT0("cc", "OneCopyTileTaskWorkerPool::ScheduleTasks"); |
- |
-#if DCHECK_IS_ON() |
- { |
- base::AutoLock lock(lock_); |
- DCHECK(!shutdown_); |
- } |
-#endif |
- |
- if (tasks_pending_.none()) |
- TRACE_EVENT_ASYNC_BEGIN0("cc", "ScheduledTasks", this); |
- |
- // Mark all task sets as pending. |
- tasks_pending_.set(); |
- |
- unsigned priority = kTileTaskPriorityBase; |
- |
- graph_.Reset(); |
- |
- // Cancel existing OnTaskSetFinished callbacks. |
- task_set_finished_weak_ptr_factory_.InvalidateWeakPtrs(); |
- |
- scoped_refptr<TileTask> new_task_set_finished_tasks[kNumberOfTaskSets]; |
- |
- size_t task_count[kNumberOfTaskSets] = {0}; |
- |
- for (TaskSet task_set = 0; task_set < kNumberOfTaskSets; ++task_set) { |
- new_task_set_finished_tasks[task_set] = CreateTaskSetFinishedTask( |
- task_runner_.get(), |
- base::Bind(&OneCopyTileTaskWorkerPool::OnTaskSetFinished, |
- task_set_finished_weak_ptr_factory_.GetWeakPtr(), task_set)); |
- } |
- |
- resource_pool_->CheckBusyResources(false); |
- |
- for (TileTaskQueue::Item::Vector::const_iterator it = queue->items.begin(); |
- it != queue->items.end(); ++it) { |
- const TileTaskQueue::Item& item = *it; |
- RasterTask* task = item.task; |
- DCHECK(!task->HasCompleted()); |
- |
- for (TaskSet task_set = 0; task_set < kNumberOfTaskSets; ++task_set) { |
- if (!item.task_sets[task_set]) |
- continue; |
- |
- ++task_count[task_set]; |
- |
- graph_.edges.push_back( |
- TaskGraph::Edge(task, new_task_set_finished_tasks[task_set].get())); |
- } |
- |
- InsertNodesForRasterTask(&graph_, task, task->dependencies(), priority++); |
- } |
- |
- for (TaskSet task_set = 0; task_set < kNumberOfTaskSets; ++task_set) { |
- InsertNodeForTask(&graph_, new_task_set_finished_tasks[task_set].get(), |
- kTaskSetFinishedTaskPriorityBase + task_set, |
- task_count[task_set]); |
- } |
- |
- ScheduleTasksOnOriginThread(this, &graph_); |
- task_graph_runner_->ScheduleTasks(namespace_token_, &graph_); |
- |
- std::copy(new_task_set_finished_tasks, |
- new_task_set_finished_tasks + kNumberOfTaskSets, |
- task_set_finished_tasks_); |
- |
- resource_pool_->ReduceResourceUsage(); |
- |
- TRACE_EVENT_ASYNC_STEP_INTO1("cc", "ScheduledTasks", this, "running", "state", |
- StateAsValue()); |
-} |
- |
-void OneCopyTileTaskWorkerPool::CheckForCompletedTasks() { |
- TRACE_EVENT0("cc", "OneCopyTileTaskWorkerPool::CheckForCompletedTasks"); |
- |
- task_graph_runner_->CollectCompletedTasks(namespace_token_, |
- &completed_tasks_); |
- |
- for (Task::Vector::const_iterator it = completed_tasks_.begin(); |
- it != completed_tasks_.end(); ++it) { |
- TileTask* task = static_cast<TileTask*>(it->get()); |
- |
- task->WillComplete(); |
- task->CompleteOnOriginThread(this); |
- task->DidComplete(); |
- |
- task->RunReplyOnOriginThread(); |
- } |
- completed_tasks_.clear(); |
-} |
- |
-ResourceFormat OneCopyTileTaskWorkerPool::GetResourceFormat() { |
- return resource_provider_->best_texture_format(); |
-} |
- |
-scoped_ptr<RasterBuffer> OneCopyTileTaskWorkerPool::AcquireBufferForRaster( |
- const Resource* resource) { |
- DCHECK_EQ(resource->format(), resource_provider_->best_texture_format()); |
- return make_scoped_ptr<RasterBuffer>( |
- new RasterBufferImpl(this, resource_provider_, resource_pool_, |
- resource_provider_->best_texture_format(), |
- resource)); |
-} |
- |
-void OneCopyTileTaskWorkerPool::ReleaseBufferForRaster( |
- scoped_ptr<RasterBuffer> buffer) { |
- // Nothing to do here. RasterBufferImpl destructor cleans up after itself. |
-} |
- |
-CopySequenceNumber |
-OneCopyTileTaskWorkerPool::PlaybackAndScheduleCopyOnWorkerThread( |
- scoped_ptr<ResourceProvider::ScopedWriteLockGpuMemoryBuffer> write_lock, |
- scoped_ptr<ScopedResource> src, |
- const Resource* dst, |
- const RasterSource* raster_source, |
- const gfx::Rect& rect, |
- float scale) { |
- base::AutoLock lock(lock_); |
- |
- int failed_attempts = 0; |
- while ((scheduled_copy_operation_count_ + issued_copy_operation_count_) >= |
- kMaxCopyOperations) { |
- // Ignore limit when shutdown is set. |
- if (shutdown_) |
- break; |
- |
- ++failed_attempts; |
- |
- // Schedule a check that will also wait for operations to complete |
- // after too many failed attempts. |
- bool wait_if_needed = failed_attempts > kFailedAttemptsBeforeWaitIfNeeded; |
- |
- // Schedule a check for completed copy operations if too many operations |
- // are currently in-flight. |
- ScheduleCheckForCompletedCopyOperationsWithLockAcquired(wait_if_needed); |
- |
- { |
- TRACE_EVENT0("cc", "WaitingForCopyOperationsToComplete"); |
- |
- // Wait for in-flight copy operations to drop below limit. |
- copy_operation_count_cv_.Wait(); |
- } |
- } |
- |
- // Increment |scheduled_copy_operation_count_| before releasing |lock_|. |
- ++scheduled_copy_operation_count_; |
- |
- // There may be more work available, so wake up another worker thread. |
- copy_operation_count_cv_.Signal(); |
- |
- { |
- base::AutoUnlock unlock(lock_); |
- |
- gfx::GpuMemoryBuffer* gpu_memory_buffer = write_lock->GetGpuMemoryBuffer(); |
- if (gpu_memory_buffer) { |
- void* data = NULL; |
- bool rv = gpu_memory_buffer->Map(&data); |
- DCHECK(rv); |
- int stride; |
- gpu_memory_buffer->GetStride(&stride); |
- TileTaskWorkerPool::PlaybackToMemory(data, src->format(), src->size(), |
- stride, raster_source, rect, scale); |
- gpu_memory_buffer->Unmap(); |
- } |
- } |
- |
- pending_copy_operations_.push_back( |
- make_scoped_ptr(new CopyOperation(write_lock.Pass(), src.Pass(), dst))); |
- |
- // Acquire a sequence number for this copy operation. |
- CopySequenceNumber sequence = next_copy_operation_sequence_++; |
- |
- // Post task that will advance last flushed copy operation to |sequence| |
- // if we have reached the flush period. |
- if ((sequence % kCopyFlushPeriod) == 0) { |
- task_runner_->PostTask( |
- FROM_HERE, |
- base::Bind(&OneCopyTileTaskWorkerPool::AdvanceLastFlushedCopyTo, |
- weak_ptr_factory_.GetWeakPtr(), sequence)); |
- } |
- |
- return sequence; |
-} |
- |
-void OneCopyTileTaskWorkerPool::AdvanceLastIssuedCopyTo( |
- CopySequenceNumber sequence) { |
- if (last_issued_copy_operation_ >= sequence) |
- return; |
- |
- IssueCopyOperations(sequence - last_issued_copy_operation_); |
- last_issued_copy_operation_ = sequence; |
-} |
- |
-void OneCopyTileTaskWorkerPool::AdvanceLastFlushedCopyTo( |
- CopySequenceNumber sequence) { |
- if (last_flushed_copy_operation_ >= sequence) |
- return; |
- |
- AdvanceLastIssuedCopyTo(sequence); |
- |
- // Flush all issued copy operations. |
- context_provider_->ContextGL()->ShallowFlushCHROMIUM(); |
- last_flushed_copy_operation_ = last_issued_copy_operation_; |
-} |
- |
-void OneCopyTileTaskWorkerPool::OnTaskSetFinished(TaskSet task_set) { |
- TRACE_EVENT1("cc", "OneCopyTileTaskWorkerPool::OnTaskSetFinished", "task_set", |
- task_set); |
- |
- DCHECK(tasks_pending_[task_set]); |
- tasks_pending_[task_set] = false; |
- if (tasks_pending_.any()) { |
- TRACE_EVENT_ASYNC_STEP_INTO1("cc", "ScheduledTasks", this, "running", |
- "state", StateAsValue()); |
- } else { |
- TRACE_EVENT_ASYNC_END0("cc", "ScheduledTasks", this); |
- } |
- client_->DidFinishRunningTileTasks(task_set); |
-} |
- |
-void OneCopyTileTaskWorkerPool::IssueCopyOperations(int64 count) { |
- TRACE_EVENT1("cc", "OneCopyTileTaskWorkerPool::IssueCopyOperations", "count", |
- count); |
- |
- CopyOperation::Deque copy_operations; |
- |
- { |
- base::AutoLock lock(lock_); |
- |
- for (int64 i = 0; i < count; ++i) { |
- DCHECK(!pending_copy_operations_.empty()); |
- copy_operations.push_back(pending_copy_operations_.take_front()); |
- } |
- |
- // Decrement |scheduled_copy_operation_count_| and increment |
- // |issued_copy_operation_count_| to reflect the transition of copy |
- // operations from "pending" to "issued" state. |
- DCHECK_GE(scheduled_copy_operation_count_, copy_operations.size()); |
- scheduled_copy_operation_count_ -= copy_operations.size(); |
- issued_copy_operation_count_ += copy_operations.size(); |
- } |
- |
- while (!copy_operations.empty()) { |
- scoped_ptr<CopyOperation> copy_operation = copy_operations.take_front(); |
- |
- // Remove the write lock. |
- copy_operation->write_lock.reset(); |
- |
- // Copy contents of source resource to destination resource. |
- resource_provider_->CopyResource(copy_operation->src->id(), |
- copy_operation->dst->id()); |
- |
- // Return source resource to pool where it can be reused once copy |
- // operation has completed and resource is no longer busy. |
- resource_pool_->ReleaseResource(copy_operation->src.Pass()); |
- } |
-} |
- |
-void OneCopyTileTaskWorkerPool:: |
- ScheduleCheckForCompletedCopyOperationsWithLockAcquired( |
- bool wait_if_needed) { |
- lock_.AssertAcquired(); |
- |
- if (check_for_completed_copy_operations_pending_) |
- return; |
- |
- base::TimeTicks now = base::TimeTicks::Now(); |
- |
- // Schedule a check for completed copy operations as soon as possible but |
- // don't allow two consecutive checks to be scheduled to run less than the |
- // tick rate apart. |
- base::TimeTicks next_check_for_completed_copy_operations_time = |
- std::max(last_check_for_completed_copy_operations_time_ + |
- base::TimeDelta::FromMilliseconds( |
- kCheckForCompletedCopyOperationsTickRateMs), |
- now); |
- |
- task_runner_->PostDelayedTask( |
- FROM_HERE, |
- base::Bind(&OneCopyTileTaskWorkerPool::CheckForCompletedCopyOperations, |
- weak_ptr_factory_.GetWeakPtr(), wait_if_needed), |
- next_check_for_completed_copy_operations_time - now); |
- |
- last_check_for_completed_copy_operations_time_ = |
- next_check_for_completed_copy_operations_time; |
- check_for_completed_copy_operations_pending_ = true; |
-} |
- |
-void OneCopyTileTaskWorkerPool::CheckForCompletedCopyOperations( |
- bool wait_if_needed) { |
- TRACE_EVENT1("cc", |
- "OneCopyTileTaskWorkerPool::CheckForCompletedCopyOperations", |
- "wait_if_needed", wait_if_needed); |
- |
- resource_pool_->CheckBusyResources(wait_if_needed); |
- |
- { |
- base::AutoLock lock(lock_); |
- |
- DCHECK(check_for_completed_copy_operations_pending_); |
- check_for_completed_copy_operations_pending_ = false; |
- |
- // The number of busy resources in the pool reflects the number of issued |
- // copy operations that have not yet completed. |
- issued_copy_operation_count_ = resource_pool_->busy_resource_count(); |
- |
- // There may be work blocked on too many in-flight copy operations, so wake |
- // up a worker thread. |
- copy_operation_count_cv_.Signal(); |
- } |
-} |
- |
-scoped_refptr<base::trace_event::ConvertableToTraceFormat> |
-OneCopyTileTaskWorkerPool::StateAsValue() const { |
- scoped_refptr<base::trace_event::TracedValue> state = |
- new base::trace_event::TracedValue(); |
- |
- state->BeginArray("tasks_pending"); |
- for (TaskSet task_set = 0; task_set < kNumberOfTaskSets; ++task_set) |
- state->AppendBoolean(tasks_pending_[task_set]); |
- state->EndArray(); |
- state->BeginDictionary("staging_state"); |
- StagingStateAsValueInto(state.get()); |
- state->EndDictionary(); |
- |
- return state; |
-} |
- |
-void OneCopyTileTaskWorkerPool::StagingStateAsValueInto( |
- base::trace_event::TracedValue* staging_state) const { |
- staging_state->SetInteger("staging_resource_count", |
- resource_pool_->total_resource_count()); |
- staging_state->SetInteger("bytes_used_for_staging_resources", |
- resource_pool_->total_memory_usage_bytes()); |
- staging_state->SetInteger("pending_copy_count", |
- resource_pool_->total_resource_count() - |
- resource_pool_->acquired_resource_count()); |
- staging_state->SetInteger("bytes_pending_copy", |
- resource_pool_->total_memory_usage_bytes() - |
- resource_pool_->acquired_memory_usage_bytes()); |
-} |
- |
-} // namespace cc |