| Index: gpu/ipc/service/gpu_channel.cc
|
| diff --git a/gpu/ipc/service/gpu_channel.cc b/gpu/ipc/service/gpu_channel.cc
|
| index a7f07ef6ffdc8477949b754e26046df0befcc663..7a56a62805d9262175af3138be8a0daa929067e3 100644
|
| --- a/gpu/ipc/service/gpu_channel.cc
|
| +++ b/gpu/ipc/service/gpu_channel.cc
|
| @@ -194,7 +194,7 @@ bool GpuChannelMessageQueue::PushBackMessage(const IPC::Message& message) {
|
|
|
| uint32_t order_num = sync_point_order_data_->GenerateUnprocessedOrderNumber(
|
| sync_point_manager_);
|
| - scoped_ptr<GpuChannelMessage> msg(
|
| + std::unique_ptr<GpuChannelMessage> msg(
|
| new GpuChannelMessage(message, order_num, base::TimeTicks::Now()));
|
|
|
| if (channel_messages_.empty()) {
|
| @@ -958,7 +958,7 @@ void GpuChannel::OnCreateCommandBuffer(
|
| return;
|
| }
|
|
|
| - scoped_ptr<GpuCommandBufferStub> stub(new GpuCommandBufferStub(
|
| + std::unique_ptr<GpuCommandBufferStub> stub(new GpuCommandBufferStub(
|
| this, sync_point_manager_, task_runner_.get(), share_group,
|
| surface_handle, mailbox_manager_.get(), preempted_flag_.get(),
|
| subscription_ref_set_.get(), pending_valuebuffer_state_.get(), size,
|
| @@ -983,7 +983,7 @@ void GpuChannel::OnDestroyCommandBuffer(int32_t route_id) {
|
| TRACE_EVENT1("gpu", "GpuChannel::OnDestroyCommandBuffer",
|
| "route_id", route_id);
|
|
|
| - scoped_ptr<GpuCommandBufferStub> stub = stubs_.take_and_erase(route_id);
|
| + std::unique_ptr<GpuCommandBufferStub> stub = stubs_.take_and_erase(route_id);
|
| // In case the renderer is currently blocked waiting for a sync reply from the
|
| // stub, we need to make sure to reschedule the correct stream here.
|
| if (stub && !stub->IsScheduled()) {
|
|
|