| Index: services/ui/gles2/command_buffer_driver.cc
|
| diff --git a/services/ui/gles2/command_buffer_driver.cc b/services/ui/gles2/command_buffer_driver.cc
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..559439690fbb02ee0934866c9b80a97e1ce876c2
|
| --- /dev/null
|
| +++ b/services/ui/gles2/command_buffer_driver.cc
|
| @@ -0,0 +1,568 @@
|
| +// Copyright 2013 The Chromium Authors. All rights reserved.
|
| +// Use of this source code is governed by a BSD-style license that can be
|
| +// found in the LICENSE file.
|
| +
|
| +#include "services/ui/gles2/command_buffer_driver.h"
|
| +
|
| +#include <stddef.h>
|
| +#include <utility>
|
| +
|
| +#include "base/bind.h"
|
| +#include "base/memory/shared_memory.h"
|
| +#include "base/threading/thread_task_runner_handle.h"
|
| +#include "build/build_config.h"
|
| +#include "gpu/command_buffer/common/gpu_memory_buffer_support.h"
|
| +#include "gpu/command_buffer/service/command_buffer_service.h"
|
| +#include "gpu/command_buffer/service/command_executor.h"
|
| +#include "gpu/command_buffer/service/context_group.h"
|
| +#include "gpu/command_buffer/service/gles2_cmd_decoder.h"
|
| +#include "gpu/command_buffer/service/image_manager.h"
|
| +#include "gpu/command_buffer/service/mailbox_manager.h"
|
| +#include "gpu/command_buffer/service/query_manager.h"
|
| +#include "gpu/command_buffer/service/sync_point_manager.h"
|
| +#include "gpu/command_buffer/service/transfer_buffer_manager.h"
|
| +#include "mojo/public/cpp/system/platform_handle.h"
|
| +#include "services/ui/common/mojo_buffer_backing.h"
|
| +#include "services/ui/gles2/gl_surface_adapter.h"
|
| +#include "services/ui/gles2/gpu_memory_tracker.h"
|
| +#include "services/ui/gles2/gpu_state.h"
|
| +#include "ui/gfx/buffer_format_util.h"
|
| +#include "ui/gfx/gpu_memory_buffer.h"
|
| +#include "ui/gfx/vsync_provider.h"
|
| +#include "ui/gl/gl_context.h"
|
| +#include "ui/gl/gl_image_shared_memory.h"
|
| +#include "ui/gl/gl_surface.h"
|
| +#include "ui/gl/init/gl_factory.h"
|
| +
|
| +#if defined(USE_OZONE)
|
| +#include "ui/ozone/gl/gl_image_ozone_native_pixmap.h"
|
| +#endif
|
| +
|
| +namespace ui {
|
| +
|
| +namespace {
|
| +
|
| +// The first time polling a fence, delay some extra time to allow other
|
| +// stubs to process some work, or else the timing of the fences could
|
| +// allow a pattern of alternating fast and slow frames to occur.
|
| +const int64_t kHandleMoreWorkPeriodMs = 2;
|
| +const int64_t kHandleMoreWorkPeriodBusyMs = 1;
|
| +
|
| +// Prevents idle work from being starved.
|
| +const int64_t kMaxTimeSinceIdleMs = 10;
|
| +
|
| +} // namespace
|
| +
|
| +CommandBufferDriver::Client::~Client() {}
|
| +
|
| +CommandBufferDriver::CommandBufferDriver(
|
| + gpu::CommandBufferNamespace command_buffer_namespace,
|
| + gpu::CommandBufferId command_buffer_id,
|
| + gfx::AcceleratedWidget widget,
|
| + scoped_refptr<GpuState> gpu_state)
|
| + : command_buffer_namespace_(command_buffer_namespace),
|
| + command_buffer_id_(command_buffer_id),
|
| + widget_(widget),
|
| + client_(nullptr),
|
| + gpu_state_(gpu_state),
|
| + previous_processed_num_(0),
|
| + weak_factory_(this) {
|
| + DCHECK_EQ(base::ThreadTaskRunnerHandle::Get(),
|
| + gpu_state_->command_buffer_task_runner()->task_runner());
|
| +}
|
| +
|
| +CommandBufferDriver::~CommandBufferDriver() {
|
| + DCHECK(CalledOnValidThread());
|
| + DestroyDecoder();
|
| +}
|
| +
|
| +bool CommandBufferDriver::Initialize(
|
| + mojo::ScopedSharedBufferHandle shared_state,
|
| + mojo::Array<int32_t> attribs) {
|
| + DCHECK(CalledOnValidThread());
|
| + gpu::gles2::ContextCreationAttribHelper attrib_helper;
|
| + if (!attrib_helper.Parse(attribs.storage()))
|
| + return false;
|
| + // TODO(piman): attribs can't currently represent gpu_preference.
|
| +
|
| + const bool offscreen = widget_ == gfx::kNullAcceleratedWidget;
|
| + if (offscreen) {
|
| + surface_ = gl::init::CreateOffscreenGLSurface(gfx::Size(0, 0));
|
| + } else {
|
| +#if defined(USE_OZONE)
|
| + scoped_refptr<gl::GLSurface> underlying_surface =
|
| + gl::init::CreateSurfacelessViewGLSurface(widget_);
|
| + if (!underlying_surface)
|
| + underlying_surface = gl::init::CreateViewGLSurface(widget_);
|
| +#else
|
| + scoped_refptr<gl::GLSurface> underlying_surface =
|
| + gl::init::CreateViewGLSurface(widget_);
|
| +#endif
|
| + scoped_refptr<GLSurfaceAdapterMus> surface_adapter =
|
| + new GLSurfaceAdapterMus(underlying_surface);
|
| + surface_adapter->SetGpuCompletedSwapBuffersCallback(
|
| + base::Bind(&CommandBufferDriver::OnGpuCompletedSwapBuffers,
|
| + weak_factory_.GetWeakPtr()));
|
| + surface_ = surface_adapter;
|
| +
|
| + gfx::VSyncProvider* vsync_provider =
|
| + surface_ ? surface_->GetVSyncProvider() : nullptr;
|
| + if (vsync_provider) {
|
| + vsync_provider->GetVSyncParameters(
|
| + base::Bind(&CommandBufferDriver::OnUpdateVSyncParameters,
|
| + weak_factory_.GetWeakPtr()));
|
| + }
|
| + }
|
| +
|
| + if (!surface_.get())
|
| + return false;
|
| +
|
| + // TODO(piman): virtual contexts.
|
| + context_ = gl::init::CreateGLContext(
|
| + gpu_state_->share_group(), surface_.get(), attrib_helper.gpu_preference);
|
| + if (!context_.get())
|
| + return false;
|
| +
|
| + if (!context_->MakeCurrent(surface_.get()))
|
| + return false;
|
| +
|
| + // TODO(piman): ShaderTranslatorCache is currently per-ContextGroup but
|
| + // only needs to be per-thread.
|
| + const bool bind_generates_resource = attrib_helper.bind_generates_resource;
|
| + scoped_refptr<gpu::gles2::FeatureInfo> feature_info =
|
| + new gpu::gles2::FeatureInfo(gpu_state_->gpu_driver_bug_workarounds());
|
| + // TODO(erikchen): The ContextGroup needs a reference to the
|
| + // GpuMemoryBufferManager.
|
| + scoped_refptr<gpu::gles2::ContextGroup> context_group =
|
| + new gpu::gles2::ContextGroup(
|
| + gpu_state_->gpu_preferences(), gpu_state_->mailbox_manager(),
|
| + new GpuMemoryTracker,
|
| + new gpu::gles2::ShaderTranslatorCache(gpu_state_->gpu_preferences()),
|
| + new gpu::gles2::FramebufferCompletenessCache, feature_info,
|
| + bind_generates_resource, nullptr);
|
| +
|
| + command_buffer_.reset(
|
| + new gpu::CommandBufferService(context_group->transfer_buffer_manager()));
|
| +
|
| + decoder_.reset(::gpu::gles2::GLES2Decoder::Create(context_group.get()));
|
| + executor_.reset(new gpu::CommandExecutor(command_buffer_.get(),
|
| + decoder_.get(), decoder_.get()));
|
| + sync_point_order_data_ = gpu::SyncPointOrderData::Create();
|
| + sync_point_client_ = gpu_state_->sync_point_manager()->CreateSyncPointClient(
|
| + sync_point_order_data_, GetNamespaceID(), command_buffer_id_);
|
| + decoder_->set_engine(executor_.get());
|
| + decoder_->SetFenceSyncReleaseCallback(base::Bind(
|
| + &CommandBufferDriver::OnFenceSyncRelease, base::Unretained(this)));
|
| + decoder_->SetWaitFenceSyncCallback(base::Bind(
|
| + &CommandBufferDriver::OnWaitFenceSync, base::Unretained(this)));
|
| + decoder_->SetDescheduleUntilFinishedCallback(base::Bind(
|
| + &CommandBufferDriver::OnDescheduleUntilFinished, base::Unretained(this)));
|
| + decoder_->SetRescheduleAfterFinishedCallback(base::Bind(
|
| + &CommandBufferDriver::OnRescheduleAfterFinished, base::Unretained(this)));
|
| +
|
| + gpu::gles2::DisallowedFeatures disallowed_features;
|
| +
|
| + if (!decoder_->Initialize(surface_, context_, offscreen, disallowed_features,
|
| + attrib_helper))
|
| + return false;
|
| +
|
| + command_buffer_->SetPutOffsetChangeCallback(base::Bind(
|
| + &gpu::CommandExecutor::PutChanged, base::Unretained(executor_.get())));
|
| + command_buffer_->SetGetBufferChangeCallback(base::Bind(
|
| + &gpu::CommandExecutor::SetGetBuffer, base::Unretained(executor_.get())));
|
| + command_buffer_->SetParseErrorCallback(
|
| + base::Bind(&CommandBufferDriver::OnParseError, base::Unretained(this)));
|
| +
|
| + // TODO(piman): other callbacks
|
| +
|
| + const size_t kSize = sizeof(gpu::CommandBufferSharedState);
|
| + std::unique_ptr<gpu::BufferBacking> backing(
|
| + MojoBufferBacking::Create(std::move(shared_state), kSize));
|
| + if (!backing)
|
| + return false;
|
| +
|
| + command_buffer_->SetSharedStateBuffer(std::move(backing));
|
| + gpu_state_->driver_manager()->AddDriver(this);
|
| + return true;
|
| +}
|
| +
|
| +void CommandBufferDriver::SetGetBuffer(int32_t buffer) {
|
| + DCHECK(CalledOnValidThread());
|
| + command_buffer_->SetGetBuffer(buffer);
|
| +}
|
| +
|
| +void CommandBufferDriver::Flush(int32_t put_offset) {
|
| + DCHECK(CalledOnValidThread());
|
| + if (!MakeCurrent())
|
| + return;
|
| +
|
| + command_buffer_->Flush(put_offset);
|
| + ProcessPendingAndIdleWork();
|
| +}
|
| +
|
| +void CommandBufferDriver::RegisterTransferBuffer(
|
| + int32_t id,
|
| + mojo::ScopedSharedBufferHandle transfer_buffer,
|
| + uint32_t size) {
|
| + DCHECK(CalledOnValidThread());
|
| + // Take ownership of the memory and map it into this process.
|
| + // This validates the size.
|
| + std::unique_ptr<gpu::BufferBacking> backing(
|
| + MojoBufferBacking::Create(std::move(transfer_buffer), size));
|
| + if (!backing) {
|
| + DVLOG(0) << "Failed to map shared memory.";
|
| + return;
|
| + }
|
| + command_buffer_->RegisterTransferBuffer(id, std::move(backing));
|
| +}
|
| +
|
| +void CommandBufferDriver::DestroyTransferBuffer(int32_t id) {
|
| + DCHECK(CalledOnValidThread());
|
| + command_buffer_->DestroyTransferBuffer(id);
|
| +}
|
| +
|
| +void CommandBufferDriver::CreateImage(int32_t id,
|
| + mojo::ScopedHandle memory_handle,
|
| + int32_t type,
|
| + const gfx::Size& size,
|
| + int32_t format,
|
| + int32_t internal_format) {
|
| + DCHECK(CalledOnValidThread());
|
| + if (!MakeCurrent())
|
| + return;
|
| +
|
| + gpu::gles2::ImageManager* image_manager = decoder_->GetImageManager();
|
| + if (image_manager->LookupImage(id)) {
|
| + LOG(ERROR) << "Image already exists with same ID.";
|
| + return;
|
| + }
|
| +
|
| + gfx::BufferFormat gpu_format = static_cast<gfx::BufferFormat>(format);
|
| + if (!gpu::IsGpuMemoryBufferFormatSupported(gpu_format,
|
| + decoder_->GetCapabilities())) {
|
| + LOG(ERROR) << "Format is not supported.";
|
| + return;
|
| + }
|
| +
|
| + if (!gpu::IsImageSizeValidForGpuMemoryBufferFormat(size, gpu_format)) {
|
| + LOG(ERROR) << "Invalid image size for format.";
|
| + return;
|
| + }
|
| +
|
| + if (!gpu::IsImageFormatCompatibleWithGpuMemoryBufferFormat(internal_format,
|
| + gpu_format)) {
|
| + LOG(ERROR) << "Incompatible image format.";
|
| + return;
|
| + }
|
| +
|
| + if (type != gfx::SHARED_MEMORY_BUFFER) {
|
| + NOTIMPLEMENTED();
|
| + return;
|
| + }
|
| +
|
| + base::PlatformFile platform_file;
|
| + MojoResult unwrap_result = mojo::UnwrapPlatformFile(std::move(memory_handle),
|
| + &platform_file);
|
| + if (unwrap_result != MOJO_RESULT_OK) {
|
| + NOTREACHED();
|
| + return;
|
| + }
|
| +
|
| +#if defined(OS_WIN)
|
| + base::SharedMemoryHandle handle(platform_file, base::GetCurrentProcId());
|
| +#else
|
| + base::FileDescriptor handle(platform_file, false);
|
| +#endif
|
| +
|
| + scoped_refptr<gl::GLImageSharedMemory> image =
|
| + new gl::GLImageSharedMemory(size, internal_format);
|
| + // TODO(jam): also need a mojo enum for this enum
|
| + if (!image->Initialize(
|
| + handle, gfx::GpuMemoryBufferId(id), gpu_format, 0,
|
| + gfx::RowSizeForBufferFormat(size.width(), gpu_format, 0))) {
|
| + NOTREACHED();
|
| + return;
|
| + }
|
| +
|
| + image_manager->AddImage(image.get(), id);
|
| +}
|
| +
|
| +// TODO(rjkroege): It is conceivable that this code belongs in
|
| +// ozone_gpu_memory_buffer.cc
|
| +void CommandBufferDriver::CreateImageNativeOzone(int32_t id,
|
| + int32_t type,
|
| + gfx::Size size,
|
| + gfx::BufferFormat format,
|
| + uint32_t internal_format,
|
| + ui::NativePixmap* pixmap) {
|
| +#if defined(USE_OZONE)
|
| + gpu::gles2::ImageManager* image_manager = decoder_->GetImageManager();
|
| + if (image_manager->LookupImage(id)) {
|
| + LOG(ERROR) << "Image already exists with same ID.";
|
| + return;
|
| + }
|
| +
|
| + scoped_refptr<ui::GLImageOzoneNativePixmap> image =
|
| + new ui::GLImageOzoneNativePixmap(size, internal_format);
|
| + if (!image->Initialize(pixmap, format)) {
|
| + NOTREACHED();
|
| + return;
|
| + }
|
| +
|
| + image_manager->AddImage(image.get(), id);
|
| +#endif
|
| +}
|
| +
|
| +void CommandBufferDriver::DestroyImage(int32_t id) {
|
| + DCHECK(CalledOnValidThread());
|
| + gpu::gles2::ImageManager* image_manager = decoder_->GetImageManager();
|
| + if (!image_manager->LookupImage(id)) {
|
| + LOG(ERROR) << "Image with ID doesn't exist.";
|
| + return;
|
| + }
|
| + if (!MakeCurrent())
|
| + return;
|
| + image_manager->RemoveImage(id);
|
| +}
|
| +
|
| +bool CommandBufferDriver::IsScheduled() const {
|
| + DCHECK(CalledOnValidThread());
|
| + DCHECK(executor_);
|
| + return executor_->scheduled();
|
| +}
|
| +
|
| +bool CommandBufferDriver::HasUnprocessedCommands() const {
|
| + DCHECK(CalledOnValidThread());
|
| + if (command_buffer_) {
|
| + gpu::CommandBuffer::State state = GetLastState();
|
| + return command_buffer_->GetPutOffset() != state.get_offset &&
|
| + !gpu::error::IsError(state.error);
|
| + }
|
| + return false;
|
| +}
|
| +
|
| +gpu::Capabilities CommandBufferDriver::GetCapabilities() const {
|
| + DCHECK(CalledOnValidThread());
|
| + return decoder_->GetCapabilities();
|
| +}
|
| +
|
| +gpu::CommandBuffer::State CommandBufferDriver::GetLastState() const {
|
| + DCHECK(CalledOnValidThread());
|
| + return command_buffer_->GetLastState();
|
| +}
|
| +
|
| +uint32_t CommandBufferDriver::GetUnprocessedOrderNum() const {
|
| + DCHECK(CalledOnValidThread());
|
| + return sync_point_order_data_->unprocessed_order_num();
|
| +}
|
| +
|
| +uint32_t CommandBufferDriver::GetProcessedOrderNum() const {
|
| + DCHECK(CalledOnValidThread());
|
| + return sync_point_order_data_->processed_order_num();
|
| +}
|
| +
|
| +bool CommandBufferDriver::MakeCurrent() {
|
| + DCHECK(CalledOnValidThread());
|
| + if (!decoder_)
|
| + return false;
|
| + if (decoder_->MakeCurrent())
|
| + return true;
|
| + DLOG(ERROR) << "Context lost because MakeCurrent failed.";
|
| + gpu::error::ContextLostReason reason =
|
| + static_cast<gpu::error::ContextLostReason>(
|
| + decoder_->GetContextLostReason());
|
| + command_buffer_->SetContextLostReason(reason);
|
| + command_buffer_->SetParseError(gpu::error::kLostContext);
|
| + OnContextLost(reason);
|
| + return false;
|
| +}
|
| +
|
| +void CommandBufferDriver::ProcessPendingAndIdleWork() {
|
| + DCHECK(CalledOnValidThread());
|
| + executor_->ProcessPendingQueries();
|
| + ScheduleDelayedWork(
|
| + base::TimeDelta::FromMilliseconds(kHandleMoreWorkPeriodMs));
|
| +}
|
| +
|
| +void CommandBufferDriver::ScheduleDelayedWork(base::TimeDelta delay) {
|
| + DCHECK(CalledOnValidThread());
|
| + const bool has_more_work =
|
| + executor_->HasPendingQueries() || executor_->HasMoreIdleWork();
|
| + if (!has_more_work) {
|
| + last_idle_time_ = base::TimeTicks();
|
| + return;
|
| + }
|
| +
|
| + const base::TimeTicks current_time = base::TimeTicks::Now();
|
| + // |process_delayed_work_time_| is set if processing of delayed work is
|
| + // already scheduled. Just update the time if already scheduled.
|
| + if (!process_delayed_work_time_.is_null()) {
|
| + process_delayed_work_time_ = current_time + delay;
|
| + return;
|
| + }
|
| +
|
| + // Idle when no messages are processed between now and when PollWork is
|
| + // called.
|
| + previous_processed_num_ =
|
| + gpu_state_->driver_manager()->GetProcessedOrderNum();
|
| +
|
| + if (last_idle_time_.is_null())
|
| + last_idle_time_ = current_time;
|
| +
|
| + // scheduled() returns true after passing all unschedule fences and this is
|
| + // when we can start performing idle work. Idle work is done synchronously
|
| + // so we can set delay to 0 and instead poll for more work at the rate idle
|
| + // work is performed. This also ensures that idle work is done as
|
| + // efficiently as possible without any unnecessary delays.
|
| + if (executor_->scheduled() && executor_->HasMoreIdleWork())
|
| + delay = base::TimeDelta();
|
| +
|
| + process_delayed_work_time_ = current_time + delay;
|
| + gpu_state_->command_buffer_task_runner()->task_runner()->PostDelayedTask(
|
| + FROM_HERE,
|
| + base::Bind(&CommandBufferDriver::PollWork, weak_factory_.GetWeakPtr()),
|
| + delay);
|
| +}
|
| +
|
| +void CommandBufferDriver::PollWork() {
|
| + DCHECK(CalledOnValidThread());
|
| + // Post another delayed task if we have not yet reached the time at which
|
| + // we should process delayed work.
|
| + base::TimeTicks current_time = base::TimeTicks::Now();
|
| + DCHECK(!process_delayed_work_time_.is_null());
|
| + if (process_delayed_work_time_ > current_time) {
|
| + gpu_state_->command_buffer_task_runner()->task_runner()->PostDelayedTask(
|
| + FROM_HERE,
|
| + base::Bind(&CommandBufferDriver::PollWork, weak_factory_.GetWeakPtr()),
|
| + process_delayed_work_time_ - current_time);
|
| + return;
|
| + }
|
| + process_delayed_work_time_ = base::TimeTicks();
|
| + PerformWork();
|
| +}
|
| +
|
| +void CommandBufferDriver::PerformWork() {
|
| + DCHECK(CalledOnValidThread());
|
| + if (!MakeCurrent())
|
| + return;
|
| +
|
| + if (executor_) {
|
| + const uint32_t current_unprocessed_num =
|
| + gpu_state_->driver_manager()->GetUnprocessedOrderNum();
|
| + // We're idle when no messages were processed or scheduled.
|
| + bool is_idle = (previous_processed_num_ == current_unprocessed_num);
|
| + if (!is_idle && !last_idle_time_.is_null()) {
|
| + base::TimeDelta time_since_idle =
|
| + base::TimeTicks::Now() - last_idle_time_;
|
| + base::TimeDelta max_time_since_idle =
|
| + base::TimeDelta::FromMilliseconds(kMaxTimeSinceIdleMs);
|
| + // Force idle when it's been too long since last time we were idle.
|
| + if (time_since_idle > max_time_since_idle)
|
| + is_idle = true;
|
| + }
|
| +
|
| + if (is_idle) {
|
| + last_idle_time_ = base::TimeTicks::Now();
|
| + executor_->PerformIdleWork();
|
| + }
|
| + executor_->ProcessPendingQueries();
|
| + }
|
| +
|
| + ScheduleDelayedWork(
|
| + base::TimeDelta::FromMilliseconds(kHandleMoreWorkPeriodBusyMs));
|
| +}
|
| +
|
| +void CommandBufferDriver::DestroyDecoder() {
|
| + DCHECK(CalledOnValidThread());
|
| + if (decoder_) {
|
| + gpu_state_->driver_manager()->RemoveDriver(this);
|
| + bool have_context = decoder_->MakeCurrent();
|
| + decoder_->Destroy(have_context);
|
| + decoder_.reset();
|
| + }
|
| +}
|
| +
|
| +void CommandBufferDriver::OnUpdateVSyncParameters(
|
| + const base::TimeTicks timebase,
|
| + const base::TimeDelta interval) {
|
| + DCHECK(CalledOnValidThread());
|
| + if (client_)
|
| + client_->UpdateVSyncParameters(timebase, interval);
|
| +}
|
| +
|
| +void CommandBufferDriver::OnFenceSyncRelease(uint64_t release) {
|
| + DCHECK(CalledOnValidThread());
|
| + if (!sync_point_client_->client_state()->IsFenceSyncReleased(release))
|
| + sync_point_client_->ReleaseFenceSync(release);
|
| +}
|
| +
|
| +bool CommandBufferDriver::OnWaitFenceSync(
|
| + gpu::CommandBufferNamespace namespace_id,
|
| + gpu::CommandBufferId command_buffer_id,
|
| + uint64_t release) {
|
| + DCHECK(CalledOnValidThread());
|
| + DCHECK(IsScheduled());
|
| + gpu::SyncPointManager* sync_point_manager = gpu_state_->sync_point_manager();
|
| + DCHECK(sync_point_manager);
|
| +
|
| + scoped_refptr<gpu::SyncPointClientState> release_state =
|
| + sync_point_manager->GetSyncPointClientState(namespace_id,
|
| + command_buffer_id);
|
| +
|
| + if (!release_state)
|
| + return true;
|
| +
|
| + executor_->SetScheduled(false);
|
| + sync_point_client_->Wait(release_state.get(), release,
|
| + base::Bind(&gpu::CommandExecutor::SetScheduled,
|
| + executor_->AsWeakPtr(), true));
|
| + return executor_->scheduled();
|
| +}
|
| +
|
| +void CommandBufferDriver::OnDescheduleUntilFinished() {
|
| + DCHECK(CalledOnValidThread());
|
| + DCHECK(IsScheduled());
|
| + DCHECK(executor_->HasMoreIdleWork());
|
| +
|
| + executor_->SetScheduled(false);
|
| +}
|
| +
|
| +void CommandBufferDriver::OnRescheduleAfterFinished() {
|
| + DCHECK(CalledOnValidThread());
|
| + DCHECK(!executor_->scheduled());
|
| +
|
| + executor_->SetScheduled(true);
|
| +}
|
| +
|
| +void CommandBufferDriver::OnParseError() {
|
| + DCHECK(CalledOnValidThread());
|
| + gpu::CommandBuffer::State state = GetLastState();
|
| + OnContextLost(state.context_lost_reason);
|
| +}
|
| +
|
| +void CommandBufferDriver::OnContextLost(uint32_t reason) {
|
| + DCHECK(CalledOnValidThread());
|
| + if (client_)
|
| + client_->DidLoseContext(reason);
|
| +}
|
| +
|
| +void CommandBufferDriver::SignalQuery(uint32_t query_id,
|
| + const base::Closure& callback) {
|
| + DCHECK(CalledOnValidThread());
|
| +
|
| + gpu::gles2::QueryManager* query_manager = decoder_->GetQueryManager();
|
| + gpu::gles2::QueryManager::Query* query = query_manager->GetQuery(query_id);
|
| + if (query)
|
| + query->AddCallback(callback);
|
| + else
|
| + callback.Run();
|
| +}
|
| +
|
| +void CommandBufferDriver::OnGpuCompletedSwapBuffers(gfx::SwapResult result) {
|
| + DCHECK(CalledOnValidThread());
|
| + if (client_) {
|
| + client_->OnGpuCompletedSwapBuffers(result);
|
| + }
|
| +}
|
| +
|
| +} // namespace ui
|
|
|