| Index: content/common/gpu/client/command_buffer_proxy_impl.cc
|
| diff --git a/content/common/gpu/client/command_buffer_proxy_impl.cc b/content/common/gpu/client/command_buffer_proxy_impl.cc
|
| index 768b427039343a30643dc432bc8e297294c05cf9..70e070f1a4984387ed056944586c0329006c59b1 100644
|
| --- a/content/common/gpu/client/command_buffer_proxy_impl.cc
|
| +++ b/content/common/gpu/client/command_buffer_proxy_impl.cc
|
| @@ -212,10 +212,16 @@ void CommandBufferProxyImpl::Flush(int32 put_offset) {
|
| }
|
|
|
| void CommandBufferProxyImpl::SetLatencyInfo(
|
| - const ui::LatencyInfo& latency_info) {
|
| + const std::vector<ui::LatencyInfo>& latency_info) {
|
| if (last_state_.error != gpu::error::kNoError)
|
| return;
|
| - Send(new GpuCommandBufferMsg_SetLatencyInfo(route_id_, latency_info));
|
| + // TODO(miletus) : Pass the std::vector<ui::LatencyInfo> latency_info
|
| + // directly without merging once GpuCommandBufferMsg_SetLatencyInfo
|
| + // is converted to contain std::vector<ui::LatencyInfo>.
|
| + ui::LatencyInfo merged_latency;
|
| + for (size_t i = 0; i < latency_info.size(); i++)
|
| + merged_latency.MergeWith(latency_info[i]);
|
| + Send(new GpuCommandBufferMsg_SetLatencyInfo(route_id_, merged_latency));
|
| }
|
|
|
| gpu::CommandBuffer::State CommandBufferProxyImpl::FlushSync(
|
|
|