Chromium Code Reviews| OLD | NEW |
|---|---|
| 1 // Copyright (c) 2012 The Chromium Authors. All rights reserved. | 1 // Copyright (c) 2012 The Chromium Authors. All rights reserved. |
| 2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
| 3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
| 4 | 4 |
| 5 #include "content/common/gpu/texture_image_transport_surface.h" | 5 #include "content/common/gpu/texture_image_transport_surface.h" |
| 6 | 6 |
| 7 #include <string> | 7 #include <string> |
| 8 #include <vector> | 8 #include <vector> |
| 9 | 9 |
| 10 #include "base/command_line.h" | 10 #include "base/command_line.h" |
| 11 #include "content/common/gpu/gl_scoped_binders.h" | 11 #include "content/common/gpu/gl_scoped_binders.h" |
| 12 #include "content/common/gpu/gpu_channel.h" | 12 #include "content/common/gpu/gpu_channel.h" |
| 13 #include "content/common/gpu/gpu_channel_manager.h" | 13 #include "content/common/gpu/gpu_channel_manager.h" |
| 14 #include "content/common/gpu/gpu_messages.h" | 14 #include "content/common/gpu/gpu_messages.h" |
| 15 #include "content/common/gpu/sync_point_manager.h" | 15 #include "content/common/gpu/sync_point_manager.h" |
| 16 #include "content/public/common/content_switches.h" | 16 #include "content/public/common/content_switches.h" |
| 17 #include "gpu/command_buffer/service/context_group.h" | 17 #include "gpu/command_buffer/service/context_group.h" |
| 18 #include "gpu/command_buffer/service/gpu_scheduler.h" | 18 #include "gpu/command_buffer/service/gpu_scheduler.h" |
| 19 #include "gpu/command_buffer/service/texture_manager.h" | 19 #include "gpu/command_buffer/service/texture_definition.h" |
| 20 | 20 |
| 21 using gpu::gles2::ContextGroup; | 21 using gpu::gles2::ContextGroup; |
| 22 using gpu::gles2::MailboxManager; | |
| 23 using gpu::gles2::MailboxName; | |
| 24 using gpu::gles2::TextureDefinition; | |
| 22 using gpu::gles2::TextureManager; | 25 using gpu::gles2::TextureManager; |
| 23 typedef TextureManager::TextureInfo TextureInfo; | |
| 24 | 26 |
| 25 namespace content { | 27 namespace content { |
| 26 | 28 |
| 27 TextureImageTransportSurface::Texture::Texture() | 29 TextureImageTransportSurface::Texture::Texture() |
| 28 : client_id(0), | 30 : service_id(0), |
| 29 sent_to_client(false) { | 31 identifier(0) { |
| 30 } | 32 } |
| 31 | 33 |
| 32 TextureImageTransportSurface::Texture::~Texture() { | 34 TextureImageTransportSurface::Texture::~Texture() { |
| 33 } | 35 } |
| 34 | 36 |
| 35 TextureImageTransportSurface::TextureImageTransportSurface( | 37 TextureImageTransportSurface::TextureImageTransportSurface( |
| 36 GpuChannelManager* manager, | 38 GpuChannelManager* manager, |
| 37 GpuCommandBufferStub* stub, | 39 GpuCommandBufferStub* stub, |
| 38 const gfx::GLSurfaceHandle& handle) | 40 const gfx::GLSurfaceHandle& handle) |
| 39 : fbo_id_(0), | 41 : fbo_id_(0), |
| 40 front_(0), | |
| 41 stub_destroyed_(false), | 42 stub_destroyed_(false), |
| 42 backbuffer_suggested_allocation_(true), | 43 backbuffer_suggested_allocation_(true), |
| 43 frontbuffer_suggested_allocation_(true), | 44 frontbuffer_suggested_allocation_(true), |
| 44 frontbuffer_is_protected_(true), | |
| 45 protection_state_id_(0), | |
| 46 handle_(handle), | 45 handle_(handle), |
| 47 parent_stub_(NULL), | |
| 48 is_swap_buffers_pending_(false), | 46 is_swap_buffers_pending_(false), |
| 49 did_unschedule_(false), | 47 did_unschedule_(false) { |
| 50 did_flip_(false) { | |
| 51 helper_.reset(new ImageTransportHelper(this, | 48 helper_.reset(new ImageTransportHelper(this, |
| 52 manager, | 49 manager, |
| 53 stub, | 50 stub, |
| 54 gfx::kNullPluginWindow)); | 51 gfx::kNullPluginWindow)); |
| 55 } | 52 } |
| 56 | 53 |
| 57 TextureImageTransportSurface::~TextureImageTransportSurface() { | 54 TextureImageTransportSurface::~TextureImageTransportSurface() { |
| 58 DCHECK(stub_destroyed_); | 55 DCHECK(stub_destroyed_); |
| 59 Destroy(); | 56 Destroy(); |
| 60 } | 57 } |
| 61 | 58 |
| 62 bool TextureImageTransportSurface::Initialize() { | 59 bool TextureImageTransportSurface::Initialize() { |
| 60 mailbox_manager_ = | |
| 61 helper_->stub()->decoder()->GetContextGroup()->mailbox_manager(); | |
| 62 | |
| 63 backbuffer_.identifier = 1; | |
| 64 | |
| 63 GpuChannelManager* manager = helper_->manager(); | 65 GpuChannelManager* manager = helper_->manager(); |
| 64 GpuChannel* parent_channel = manager->LookupChannel(handle_.parent_client_id); | |
| 65 if (!parent_channel) | |
| 66 return false; | |
| 67 | |
| 68 parent_stub_ = parent_channel->LookupCommandBuffer(handle_.parent_context_id); | |
| 69 if (!parent_stub_) | |
| 70 return false; | |
| 71 | |
| 72 parent_stub_->AddDestructionObserver(this); | |
| 73 TextureManager* texture_manager = | |
| 74 parent_stub_->decoder()->GetContextGroup()->texture_manager(); | |
| 75 DCHECK(texture_manager); | |
| 76 | |
| 77 for (int i = 0; i < 2; ++i) { | |
| 78 Texture& texture = textures_[i]; | |
| 79 texture.client_id = handle_.parent_texture_id[i]; | |
| 80 texture.info = texture_manager->GetTextureInfo(texture.client_id); | |
| 81 if (!texture.info) | |
| 82 return false; | |
| 83 | |
| 84 if (!texture.info->target()) | |
| 85 texture_manager->SetInfoTarget(texture.info, GL_TEXTURE_2D); | |
| 86 texture_manager->SetParameter( | |
| 87 texture.info, GL_TEXTURE_MIN_FILTER, GL_LINEAR); | |
| 88 texture_manager->SetParameter( | |
| 89 texture.info, GL_TEXTURE_MAG_FILTER, GL_LINEAR); | |
| 90 texture_manager->SetParameter( | |
| 91 texture.info, GL_TEXTURE_WRAP_S, GL_CLAMP_TO_EDGE); | |
| 92 texture_manager->SetParameter( | |
| 93 texture.info, GL_TEXTURE_WRAP_T, GL_CLAMP_TO_EDGE); | |
| 94 } | |
| 95 | |
| 96 surface_ = manager->GetDefaultOffscreenSurface(); | 66 surface_ = manager->GetDefaultOffscreenSurface(); |
| 97 if (!surface_.get()) | 67 if (!surface_.get()) |
| 98 return false; | 68 return false; |
| 99 | 69 |
| 100 if (!helper_->Initialize()) | 70 if (!helper_->Initialize()) |
| 101 return false; | 71 return false; |
| 102 | 72 |
| 103 const CommandLine* command_line = CommandLine::ForCurrentProcess(); | 73 GpuChannel* parent_channel = manager->LookupChannel(handle_.parent_client_id); |
| 104 if (command_line->HasSwitch(switches::kUIPrioritizeInGpuProcess)) | 74 if (parent_channel) { |
| 105 helper_->SetPreemptByCounter(parent_channel->MessagesPendingCount()); | 75 const CommandLine* command_line = CommandLine::ForCurrentProcess(); |
| 76 if (command_line->HasSwitch(switches::kUIPrioritizeInGpuProcess)) | |
| 77 helper_->SetPreemptByCounter(parent_channel->MessagesPendingCount()); | |
| 78 } | |
| 106 | 79 |
| 107 return true; | 80 return true; |
| 108 } | 81 } |
| 109 | 82 |
| 110 void TextureImageTransportSurface::Destroy() { | 83 void TextureImageTransportSurface::Destroy() { |
| 111 if (parent_stub_) { | 84 if (!stub_destroyed_) { |
| 112 parent_stub_->decoder()->MakeCurrent(); | 85 GpuHostMsg_AcceleratedSurfaceRelease_Params params; |
| 113 ReleaseParentStub(); | 86 helper_->SendAcceleratedSurfaceRelease(params); |
| 114 } | 87 } |
| 115 | 88 |
| 116 if (surface_.get()) | 89 if (surface_.get()) |
| 117 surface_ = NULL; | 90 surface_ = NULL; |
| 118 | 91 |
| 119 helper_->Destroy(); | 92 helper_->Destroy(); |
| 120 } | 93 } |
| 121 | 94 |
| 122 bool TextureImageTransportSurface::DeferDraws() { | 95 bool TextureImageTransportSurface::DeferDraws() { |
| 123 // The command buffer hit a draw/clear command that could clobber the | 96 // The command buffer hit a draw/clear command that could clobber the |
| (...skipping 18 matching lines...) Expand all Loading... | |
| 142 } | 115 } |
| 143 | 116 |
| 144 bool TextureImageTransportSurface::OnMakeCurrent(gfx::GLContext* context) { | 117 bool TextureImageTransportSurface::OnMakeCurrent(gfx::GLContext* context) { |
| 145 if (stub_destroyed_) { | 118 if (stub_destroyed_) { |
| 146 // Early-exit so that we don't recreate the fbo. We still want to return | 119 // Early-exit so that we don't recreate the fbo. We still want to return |
| 147 // true, so that the context is made current and the GLES2DecoderImpl can | 120 // true, so that the context is made current and the GLES2DecoderImpl can |
| 148 // release its own resources. | 121 // release its own resources. |
| 149 return true; | 122 return true; |
| 150 } | 123 } |
| 151 | 124 |
| 125 if (!context_.get()) { | |
| 126 DCHECK(helper_->stub()); | |
| 127 context_ = helper_->stub()->decoder()->GetGLContext(); | |
| 128 } | |
| 129 | |
| 152 if (!fbo_id_) { | 130 if (!fbo_id_) { |
| 153 glGenFramebuffersEXT(1, &fbo_id_); | 131 glGenFramebuffersEXT(1, &fbo_id_); |
| 154 glBindFramebufferEXT(GL_FRAMEBUFFER, fbo_id_); | 132 glBindFramebufferEXT(GL_FRAMEBUFFER, fbo_id_); |
| 155 CreateBackTexture(gfx::Size(1, 1)); | 133 current_size_ = gfx::Size(1, 1); |
| 134 helper_->stub()->AddDestructionObserver(this); | |
| 135 } | |
| 136 | |
| 137 // We could be receiving non-deferred GL commands, that is anything that does | |
| 138 // not need a framebuffer. | |
| 139 if (!backbuffer_.service_id && !is_swap_buffers_pending_) { | |
| 140 CreateBackTexture(); | |
| 156 | 141 |
| 157 #ifndef NDEBUG | 142 #ifndef NDEBUG |
| 158 GLenum status = glCheckFramebufferStatusEXT(GL_FRAMEBUFFER); | 143 GLenum status = glCheckFramebufferStatusEXT(GL_FRAMEBUFFER); |
| 159 if (status != GL_FRAMEBUFFER_COMPLETE) { | 144 if (status != GL_FRAMEBUFFER_COMPLETE) { |
| 160 DLOG(ERROR) << "Framebuffer incomplete."; | 145 DLOG(ERROR) << "Framebuffer incomplete."; |
| 161 glDeleteFramebuffersEXT(1, &fbo_id_); | 146 glDeleteFramebuffersEXT(1, &fbo_id_); |
| 162 fbo_id_ = 0; | 147 fbo_id_ = 0; |
| 163 return false; | 148 return false; |
| 164 } | 149 } |
| 165 #endif | 150 #endif |
| 166 DCHECK(helper_->stub()); | |
| 167 helper_->stub()->AddDestructionObserver(this); | |
| 168 } | 151 } |
| 169 | |
| 170 return true; | 152 return true; |
| 171 } | 153 } |
| 172 | 154 |
| 173 unsigned int TextureImageTransportSurface::GetBackingFrameBufferObject() { | 155 unsigned int TextureImageTransportSurface::GetBackingFrameBufferObject() { |
| 174 return fbo_id_; | 156 return fbo_id_; |
| 175 } | 157 } |
| 176 | 158 |
| 177 void TextureImageTransportSurface::SetBackbufferAllocation(bool allocation) { | 159 void TextureImageTransportSurface::SetBackbufferAllocation(bool allocation) { |
| 178 DCHECK(!is_swap_buffers_pending_); | 160 DCHECK(!is_swap_buffers_pending_); |
| 179 if (backbuffer_suggested_allocation_ == allocation) | 161 if (backbuffer_suggested_allocation_ == allocation) |
| 180 return; | 162 return; |
| 181 backbuffer_suggested_allocation_ = allocation; | 163 backbuffer_suggested_allocation_ = allocation; |
| 182 | 164 |
| 183 if (!helper_->MakeCurrent()) | 165 if (!backbuffer_suggested_allocation_ && helper_->MakeCurrent()) |
|
piman
2012/12/03 21:08:43
I think the MakeCurrent is superfluous (as was bef
no sievers
2012/12/03 23:20:38
Done.
| |
| 184 return; | 166 ReleaseBackBuffer(); |
|
piman
2012/12/03 21:08:43
How are we going to recreate the backbuffer when a
no sievers
2012/12/03 23:20:38
It's probably harmless to call CreateBackTexture()
piman
2012/12/03 23:33:56
It thought it wouldn't have, because at the time o
no sievers
2012/12/04 00:05:36
Good point. I added a check for backbuffer_suggest
| |
| 185 | |
| 186 if (backbuffer_suggested_allocation_) { | |
| 187 DCHECK(!textures_[back()].info->service_id() || | |
| 188 !textures_[back()].sent_to_client); | |
| 189 CreateBackTexture(textures_[back()].size); | |
| 190 } else { | |
| 191 ReleaseTexture(back()); | |
| 192 } | |
| 193 } | 167 } |
| 194 | 168 |
| 195 void TextureImageTransportSurface::SetFrontbufferAllocation(bool allocation) { | 169 void TextureImageTransportSurface::SetFrontbufferAllocation(bool allocation) { |
| 196 if (frontbuffer_suggested_allocation_ == allocation) | 170 if (frontbuffer_suggested_allocation_ == allocation) |
| 197 return; | 171 return; |
| 198 frontbuffer_suggested_allocation_ = allocation; | 172 frontbuffer_suggested_allocation_ = allocation; |
| 199 AdjustFrontBufferAllocation(); | 173 AdjustFrontBufferAllocation(); |
| 200 } | 174 } |
| 201 | 175 |
| 202 void TextureImageTransportSurface::AdjustFrontBufferAllocation() { | 176 void TextureImageTransportSurface::AdjustFrontBufferAllocation() { |
| 203 if (!helper_->MakeCurrent()) | 177 if (!frontbuffer_suggested_allocation_) { |
| 204 return; | 178 GpuHostMsg_AcceleratedSurfaceRelease_Params params; |
| 205 | 179 helper_->SendAcceleratedSurfaceRelease(params); |
| 206 if (!frontbuffer_suggested_allocation_ && !frontbuffer_is_protected_ && | |
| 207 textures_[front()].info->service_id()) { | |
| 208 ReleaseTexture(front()); | |
| 209 if (textures_[front()].sent_to_client) { | |
| 210 GpuHostMsg_AcceleratedSurfaceRelease_Params params; | |
| 211 params.identifier = textures_[front()].client_id; | |
| 212 helper_->SendAcceleratedSurfaceRelease(params); | |
| 213 textures_[front()].sent_to_client = false; | |
| 214 } | |
| 215 } | 180 } |
| 216 } | 181 } |
| 217 | 182 |
| 218 void* TextureImageTransportSurface::GetShareHandle() { | 183 void* TextureImageTransportSurface::GetShareHandle() { |
| 219 return GetHandle(); | 184 return GetHandle(); |
| 220 } | 185 } |
| 221 | 186 |
| 222 void* TextureImageTransportSurface::GetDisplay() { | 187 void* TextureImageTransportSurface::GetDisplay() { |
| 223 return surface_.get() ? surface_->GetDisplay() : NULL; | 188 return surface_.get() ? surface_->GetDisplay() : NULL; |
| 224 } | 189 } |
| 225 | 190 |
| 226 void* TextureImageTransportSurface::GetConfig() { | 191 void* TextureImageTransportSurface::GetConfig() { |
| 227 return surface_.get() ? surface_->GetConfig() : NULL; | 192 return surface_.get() ? surface_->GetConfig() : NULL; |
| 228 } | 193 } |
| 229 | 194 |
| 230 void TextureImageTransportSurface::OnResize(gfx::Size size) { | 195 void TextureImageTransportSurface::OnResize(gfx::Size size) { |
| 231 CreateBackTexture(size); | 196 current_size_ = size; |
| 197 CreateBackTexture(); | |
| 232 } | 198 } |
| 233 | 199 |
| 234 void TextureImageTransportSurface::OnWillDestroyStub( | 200 void TextureImageTransportSurface::OnWillDestroyStub( |
| 235 GpuCommandBufferStub* stub) { | 201 GpuCommandBufferStub* stub) { |
| 236 if (stub == parent_stub_) { | 202 DCHECK(stub == helper_->stub()); |
| 237 ReleaseParentStub(); | 203 stub->RemoveDestructionObserver(this); |
| 238 helper_->SetPreemptByCounter(NULL); | |
| 239 } else { | |
| 240 DCHECK(stub == helper_->stub()); | |
| 241 stub->RemoveDestructionObserver(this); | |
| 242 | 204 |
| 243 // We are losing the stub owning us, this is our last chance to clean up the | 205 // We are losing the stub owning us, this is our last chance to clean up the |
| 244 // resources we allocated in the stub's context. | 206 // resources we allocated in the stub's context. |
| 245 if (fbo_id_) { | 207 if (fbo_id_) { |
| 246 glDeleteFramebuffersEXT(1, &fbo_id_); | 208 glDeleteFramebuffersEXT(1, &fbo_id_); |
| 247 CHECK_GL_ERROR(); | 209 CHECK_GL_ERROR(); |
| 248 fbo_id_ = 0; | 210 fbo_id_ = 0; |
| 249 } | 211 } |
| 250 | 212 |
| 251 stub_destroyed_ = true; | 213 stub_destroyed_ = true; |
| 252 } | |
| 253 } | 214 } |
| 254 | 215 |
| 255 bool TextureImageTransportSurface::SwapBuffers() { | 216 bool TextureImageTransportSurface::SwapBuffers() { |
| 256 DCHECK(backbuffer_suggested_allocation_); | 217 DCHECK(backbuffer_suggested_allocation_); |
| 257 if (!frontbuffer_suggested_allocation_ || !frontbuffer_is_protected_) | 218 if (!frontbuffer_suggested_allocation_) |
| 258 return true; | 219 return true; |
| 259 if (!parent_stub_) { | |
| 260 LOG(ERROR) << "SwapBuffers failed because no parent stub."; | |
| 261 return false; | |
| 262 } | |
| 263 | 220 |
| 264 glFlush(); | 221 glFlush(); |
| 265 front_ = back(); | 222 const uint64 identifier = backbuffer_.identifier; |
| 266 previous_damage_rect_ = gfx::Rect(textures_[front()].size); | 223 ProduceTexture(backbuffer_); |
| 267 | 224 |
| 268 DCHECK(textures_[front()].client_id != 0); | 225 // Do not allow destruction while we are still waiting for a swap ACK, |
| 226 // so we do not leak a texture in the mailbox. | |
| 227 AddRef(); | |
| 269 | 228 |
| 270 GpuHostMsg_AcceleratedSurfaceBuffersSwapped_Params params; | 229 GpuHostMsg_AcceleratedSurfaceBuffersSwapped_Params params; |
| 271 params.surface_handle = textures_[front()].client_id; | 230 params.surface_handle = identifier; |
| 272 params.size = textures_[front()].size; | 231 params.size = current_size_; |
| 273 params.protection_state_id = protection_state_id_; | |
| 274 params.skip_ack = false; | |
| 275 helper_->SendAcceleratedSurfaceBuffersSwapped(params); | 232 helper_->SendAcceleratedSurfaceBuffersSwapped(params); |
| 276 | 233 |
| 277 DCHECK(!is_swap_buffers_pending_); | 234 DCHECK(!is_swap_buffers_pending_); |
| 278 is_swap_buffers_pending_ = true; | 235 is_swap_buffers_pending_ = true; |
| 279 return true; | 236 return true; |
| 280 } | 237 } |
| 281 | 238 |
| 282 bool TextureImageTransportSurface::PostSubBuffer( | 239 bool TextureImageTransportSurface::PostSubBuffer( |
| 283 int x, int y, int width, int height) { | 240 int x, int y, int width, int height) { |
| 284 DCHECK(backbuffer_suggested_allocation_); | 241 DCHECK(backbuffer_suggested_allocation_); |
| 285 DCHECK(textures_[back()].info->service_id()); | 242 if (!frontbuffer_suggested_allocation_) |
|
no sievers
2012/12/01 00:58:38
I'm removing the DCHECK(textures_[back()].info->se
| |
| 286 if (!frontbuffer_suggested_allocation_ || !frontbuffer_is_protected_) | |
| 287 return true; | 243 return true; |
| 288 // If we are recreating the frontbuffer with this swap, make sure we are | |
| 289 // drawing a full frame. | |
| 290 DCHECK(textures_[front()].info->service_id() || | |
| 291 (!x && !y && gfx::Size(width, height) == textures_[back()].size)); | |
| 292 if (!parent_stub_) { | |
| 293 LOG(ERROR) << "PostSubBuffer failed because no parent stub."; | |
| 294 return false; | |
| 295 } | |
| 296 | |
| 297 const gfx::Rect new_damage_rect(x, y, width, height); | 244 const gfx::Rect new_damage_rect(x, y, width, height); |
| 245 DCHECK(gfx::Rect(gfx::Point(), current_size_).Contains(new_damage_rect)); | |
| 298 | 246 |
| 299 // An empty damage rect is a successful no-op. | 247 // An empty damage rect is a successful no-op. |
| 300 if (new_damage_rect.IsEmpty()) | 248 if (new_damage_rect.IsEmpty()) |
| 301 return true; | 249 return true; |
| 302 | 250 |
| 303 int back_texture_service_id = textures_[back()].info->service_id(); | 251 glFlush(); |
| 304 int front_texture_service_id = textures_[front()].info->service_id(); | 252 const uint64 identifier = backbuffer_.identifier; |
| 253 ProduceTexture(backbuffer_); | |
| 305 | 254 |
| 306 gfx::Size expected_size = textures_[back()].size; | 255 // Do not allow destruction while we are still waiting for a swap ACK, |
| 307 bool surfaces_same_size = textures_[front()].size == expected_size; | 256 // so we do not leak a texture in the mailbox. |
| 308 | 257 AddRef(); |
| 309 if (surfaces_same_size) { | |
| 310 std::vector<gfx::Rect> regions_to_copy; | |
| 311 GetRegionsToCopy(previous_damage_rect_, new_damage_rect, ®ions_to_copy); | |
| 312 | |
| 313 ScopedFrameBufferBinder fbo_binder(fbo_id_); | |
| 314 glFramebufferTexture2DEXT(GL_FRAMEBUFFER, | |
| 315 GL_COLOR_ATTACHMENT0, | |
| 316 GL_TEXTURE_2D, | |
| 317 front_texture_service_id, | |
| 318 0); | |
| 319 ScopedTextureBinder texture_binder(back_texture_service_id); | |
| 320 | |
| 321 for (size_t i = 0; i < regions_to_copy.size(); ++i) { | |
| 322 const gfx::Rect& region_to_copy = regions_to_copy[i]; | |
| 323 if (!region_to_copy.IsEmpty()) { | |
| 324 glCopyTexSubImage2D(GL_TEXTURE_2D, 0, region_to_copy.x(), | |
| 325 region_to_copy.y(), region_to_copy.x(), region_to_copy.y(), | |
| 326 region_to_copy.width(), region_to_copy.height()); | |
| 327 } | |
| 328 } | |
| 329 } else if (!surfaces_same_size && did_flip_) { | |
| 330 DCHECK(new_damage_rect == gfx::Rect(expected_size)); | |
| 331 } | |
| 332 | |
| 333 glFlush(); | |
| 334 front_ = back(); | |
| 335 previous_damage_rect_ = new_damage_rect; | |
| 336 | |
| 337 DCHECK(textures_[front()].client_id); | |
| 338 | 258 |
| 339 GpuHostMsg_AcceleratedSurfacePostSubBuffer_Params params; | 259 GpuHostMsg_AcceleratedSurfacePostSubBuffer_Params params; |
| 340 params.surface_handle = textures_[front()].client_id; | 260 params.surface_handle = identifier; |
| 341 params.surface_size = textures_[front()].size; | 261 params.surface_size = current_size_; |
| 342 params.x = x; | 262 params.x = x; |
| 343 params.y = y; | 263 params.y = y; |
| 344 params.width = width; | 264 params.width = width; |
| 345 params.height = height; | 265 params.height = height; |
| 346 params.protection_state_id = protection_state_id_; | |
| 347 helper_->SendAcceleratedSurfacePostSubBuffer(params); | 266 helper_->SendAcceleratedSurfacePostSubBuffer(params); |
| 348 | 267 |
| 349 DCHECK(!is_swap_buffers_pending_); | 268 DCHECK(!is_swap_buffers_pending_); |
| 350 is_swap_buffers_pending_ = true; | 269 is_swap_buffers_pending_ = true; |
| 351 return true; | 270 return true; |
| 352 } | 271 } |
| 353 | 272 |
| 354 std::string TextureImageTransportSurface::GetExtensions() { | 273 std::string TextureImageTransportSurface::GetExtensions() { |
| 355 std::string extensions = gfx::GLSurface::GetExtensions(); | 274 std::string extensions = gfx::GLSurface::GetExtensions(); |
| 356 extensions += extensions.empty() ? "" : " "; | 275 extensions += extensions.empty() ? "" : " "; |
| 357 extensions += "GL_CHROMIUM_front_buffer_cached "; | 276 extensions += "GL_CHROMIUM_front_buffer_cached "; |
| 358 extensions += "GL_CHROMIUM_post_sub_buffer"; | 277 extensions += "GL_CHROMIUM_post_sub_buffer"; |
| 359 return extensions; | 278 return extensions; |
| 360 } | 279 } |
| 361 | 280 |
| 362 gfx::Size TextureImageTransportSurface::GetSize() { | 281 gfx::Size TextureImageTransportSurface::GetSize() { |
| 363 gfx::Size size = textures_[back()].size; | 282 gfx::Size size = current_size_; |
| 364 | 283 |
| 365 // OSMesa expects a non-zero size. | 284 // OSMesa expects a non-zero size. |
| 366 return gfx::Size(size.width() == 0 ? 1 : size.width(), | 285 return gfx::Size(size.width() == 0 ? 1 : size.width(), |
| 367 size.height() == 0 ? 1 : size.height()); | 286 size.height() == 0 ? 1 : size.height()); |
| 368 } | 287 } |
| 369 | 288 |
| 370 void* TextureImageTransportSurface::GetHandle() { | 289 void* TextureImageTransportSurface::GetHandle() { |
| 371 return surface_.get() ? surface_->GetHandle() : NULL; | 290 return surface_.get() ? surface_->GetHandle() : NULL; |
| 372 } | 291 } |
| 373 | 292 |
| 374 unsigned TextureImageTransportSurface::GetFormat() { | 293 unsigned TextureImageTransportSurface::GetFormat() { |
| 375 return surface_.get() ? surface_->GetFormat() : 0; | 294 return surface_.get() ? surface_->GetFormat() : 0; |
| 376 } | 295 } |
| 377 | 296 |
| 378 void TextureImageTransportSurface::OnSetFrontSurfaceIsProtected( | 297 void TextureImageTransportSurface::OnBufferPresented(uint64 surface_handle, |
| 379 bool is_protected, uint32 protection_state_id) { | |
| 380 protection_state_id_ = protection_state_id; | |
| 381 if (frontbuffer_is_protected_ == is_protected) | |
| 382 return; | |
| 383 frontbuffer_is_protected_ = is_protected; | |
| 384 AdjustFrontBufferAllocation(); | |
| 385 | |
| 386 // If surface is set to protected, and we haven't actually released it yet, | |
| 387 // we can set the ui surface handle now just by sending a swap message. | |
| 388 if (is_protected && textures_[front()].info->service_id() && | |
| 389 textures_[front()].sent_to_client) { | |
| 390 GpuHostMsg_AcceleratedSurfaceBuffersSwapped_Params params; | |
| 391 params.surface_handle = textures_[front()].client_id; | |
| 392 params.size = textures_[front()].size; | |
| 393 params.protection_state_id = protection_state_id_; | |
| 394 params.skip_ack = true; | |
| 395 helper_->SendAcceleratedSurfaceBuffersSwapped(params); | |
| 396 } | |
| 397 } | |
| 398 | |
| 399 void TextureImageTransportSurface::OnBufferPresented(bool presented, | |
| 400 uint32 sync_point) { | 298 uint32 sync_point) { |
| 401 if (sync_point == 0) { | 299 if (sync_point == 0) { |
| 402 BufferPresentedImpl(presented); | 300 BufferPresentedImpl(surface_handle); |
| 403 } else { | 301 } else { |
| 404 helper_->manager()->sync_point_manager()->AddSyncPointCallback( | 302 helper_->manager()->sync_point_manager()->AddSyncPointCallback( |
| 405 sync_point, | 303 sync_point, |
| 406 base::Bind(&TextureImageTransportSurface::BufferPresentedImpl, | 304 base::Bind(&TextureImageTransportSurface::BufferPresentedImpl, |
| 407 this->AsWeakPtr(), | 305 this, |
| 408 presented)); | 306 surface_handle)); |
| 409 } | 307 } |
| 308 | |
| 309 // Careful, we might get deleted now if we were only waiting for | |
| 310 // a final swap ACK. | |
| 311 Release(); | |
| 410 } | 312 } |
| 411 | 313 |
| 412 void TextureImageTransportSurface::BufferPresentedImpl(bool presented) { | 314 void TextureImageTransportSurface::BufferPresentedImpl(uint64 surface_handle) { |
| 315 DCHECK(!backbuffer_.service_id); | |
| 316 if (surface_handle) { | |
| 317 DCHECK(surface_handle == 1 || surface_handle == 2); | |
| 318 backbuffer_.identifier = surface_handle; | |
| 319 ConsumeTexture(backbuffer_); | |
| 320 } else { | |
| 321 // We didn't get back a texture, so allocate 'the other' buffer. | |
| 322 backbuffer_.identifier = (backbuffer_.identifier == 1) ? 2 : 1; | |
| 323 } | |
| 324 | |
| 325 if (stub_destroyed_ && backbuffer_.service_id) { | |
| 326 // TODO(sievers): Remove this after changes to the mailbox to take ownership | |
| 327 // of the service ids. | |
| 328 DCHECK(context_.get() && surface_.get()); | |
| 329 if (context_->MakeCurrent(surface_.get())) | |
| 330 glDeleteTextures(1, &backbuffer_.service_id); | |
| 331 | |
| 332 return; | |
| 333 } | |
| 334 | |
| 413 DCHECK(is_swap_buffers_pending_); | 335 DCHECK(is_swap_buffers_pending_); |
| 414 is_swap_buffers_pending_ = false; | 336 is_swap_buffers_pending_ = false; |
| 415 | 337 |
| 416 if (presented) { | |
| 417 // If we had not flipped, the two frame damage tracking is inconsistent. | |
| 418 // So conservatively take the whole frame. | |
| 419 if (!did_flip_) | |
| 420 previous_damage_rect_ = gfx::Rect(textures_[front()].size); | |
| 421 } else { | |
| 422 front_ = back(); | |
| 423 previous_damage_rect_ = gfx::Rect(0, 0, 0, 0); | |
| 424 } | |
| 425 | |
| 426 did_flip_ = presented; | |
| 427 | |
| 428 // We're relying on the fact that the parent context is | 338 // We're relying on the fact that the parent context is |
| 429 // finished with it's context when it inserts the sync point that | 339 // finished with it's context when it inserts the sync point that |
| 430 // triggers this callback. | 340 // triggers this callback. |
| 431 if (helper_->MakeCurrent()) { | 341 if (helper_->MakeCurrent()) { |
| 432 if (textures_[front()].size != textures_[back()].size || | 342 if (backbuffer_.size != current_size_ || !backbuffer_.service_id) |
| 433 !textures_[back()].info->service_id() || | 343 CreateBackTexture(); |
| 434 !textures_[back()].sent_to_client) { | 344 else |
| 435 // We may get an ACK from a stale swap just to reschedule. In that case, | |
| 436 // we may not have a backbuffer suggestion and should not recreate one. | |
| 437 if (backbuffer_suggested_allocation_) | |
| 438 CreateBackTexture(textures_[front()].size); | |
| 439 } else { | |
| 440 AttachBackTextureToFBO(); | 345 AttachBackTextureToFBO(); |
| 441 } | |
| 442 } | 346 } |
| 443 | 347 |
| 444 // Even if MakeCurrent fails, schedule anyway, to trigger the lost context | 348 // Even if MakeCurrent fails, schedule anyway, to trigger the lost context |
| 445 // logic. | 349 // logic. |
| 446 if (did_unschedule_) { | 350 if (did_unschedule_) { |
| 447 did_unschedule_ = false; | 351 did_unschedule_ = false; |
| 448 helper_->SetScheduled(true); | 352 helper_->SetScheduled(true); |
| 449 } | 353 } |
| 450 } | 354 } |
| 451 | 355 |
| 452 void TextureImageTransportSurface::OnResizeViewACK() { | 356 void TextureImageTransportSurface::OnResizeViewACK() { |
| 453 NOTREACHED(); | 357 NOTREACHED(); |
| 454 } | 358 } |
| 455 | 359 |
| 456 void TextureImageTransportSurface::ReleaseTexture(int id) { | 360 void TextureImageTransportSurface::ReleaseBackBuffer() { |
| 457 if (!parent_stub_) | 361 if (!backbuffer_.service_id) |
| 458 return; | 362 return; |
| 459 Texture& texture = textures_[id]; | |
| 460 TextureInfo* info = texture.info; | |
| 461 DCHECK(info); | |
| 462 | 363 |
| 463 GLuint service_id = info->service_id(); | 364 glDeleteTextures(1, &backbuffer_.service_id); |
| 464 if (!service_id) | 365 backbuffer_.service_id = 0; |
| 465 return; | |
| 466 info->SetServiceId(0); | |
| 467 | |
| 468 { | |
| 469 ScopedFrameBufferBinder fbo_binder(fbo_id_); | |
| 470 glDeleteTextures(1, &service_id); | |
| 471 } | |
| 472 glFlush(); | 366 glFlush(); |
| 473 CHECK_GL_ERROR(); | 367 CHECK_GL_ERROR(); |
| 474 } | 368 } |
| 475 | 369 |
| 476 void TextureImageTransportSurface::CreateBackTexture(const gfx::Size& size) { | 370 void TextureImageTransportSurface::CreateBackTexture() { |
| 477 if (!parent_stub_) | 371 // If |is_swap_buffers_pending| we are waiting for our backbuffer |
| 478 return; | 372 // in the mailbox, so we shouldn't be reallocating it now. |
| 479 Texture& texture = textures_[back()]; | 373 DCHECK(!is_swap_buffers_pending_); |
| 480 TextureInfo* info = texture.info; | |
| 481 DCHECK(info); | |
| 482 | 374 |
| 483 GLuint service_id = info->service_id(); | 375 if (backbuffer_.service_id && backbuffer_.size == current_size_) |
| 484 | |
| 485 if (service_id && texture.size == size && texture.sent_to_client) | |
| 486 return; | 376 return; |
| 487 | 377 |
| 488 if (!service_id) { | 378 if (!backbuffer_.service_id) { |
| 489 glGenTextures(1, &service_id); | 379 MailboxName new_mailbox_name; |
| 490 info->SetServiceId(service_id); | 380 mailbox_manager_->GenerateMailboxName(&new_mailbox_name); |
| 381 mailbox_name(backbuffer_.identifier) = new_mailbox_name; | |
| 382 glGenTextures(1, &backbuffer_.service_id); | |
| 491 } | 383 } |
| 492 | 384 |
| 493 if (size != texture.size) { | 385 backbuffer_.size = current_size_; |
| 494 texture.size = size; | |
| 495 TextureManager* texture_manager = | |
| 496 parent_stub_->decoder()->GetContextGroup()->texture_manager(); | |
| 497 texture_manager->SetLevelInfo( | |
| 498 info, | |
| 499 GL_TEXTURE_2D, | |
| 500 0, | |
| 501 GL_RGBA, | |
| 502 size.width(), | |
| 503 size.height(), | |
| 504 1, | |
| 505 0, | |
| 506 GL_RGBA, | |
| 507 GL_UNSIGNED_BYTE, | |
| 508 true); | |
| 509 } | |
| 510 | 386 |
| 511 { | 387 { |
| 512 ScopedTextureBinder texture_binder(service_id); | 388 ScopedTextureBinder texture_binder(backbuffer_.service_id); |
| 513 glTexParameterf(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_LINEAR); | 389 glTexParameterf(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_LINEAR); |
| 514 glTexParameterf(GL_TEXTURE_2D, GL_TEXTURE_MAG_FILTER, GL_LINEAR); | 390 glTexParameterf(GL_TEXTURE_2D, GL_TEXTURE_MAG_FILTER, GL_LINEAR); |
| 515 glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_S, GL_CLAMP_TO_EDGE); | 391 glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_S, GL_CLAMP_TO_EDGE); |
| 516 glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_T, GL_CLAMP_TO_EDGE); | 392 glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_T, GL_CLAMP_TO_EDGE); |
| 517 glTexImage2D(GL_TEXTURE_2D, 0, GL_RGBA, | 393 glTexImage2D(GL_TEXTURE_2D, 0, GL_RGBA, |
| 518 size.width(), size.height(), 0, | 394 current_size_.width(), current_size_.height(), 0, |
| 519 GL_RGBA, GL_UNSIGNED_BYTE, NULL); | 395 GL_RGBA, GL_UNSIGNED_BYTE, NULL); |
| 520 CHECK_GL_ERROR(); | 396 CHECK_GL_ERROR(); |
| 521 } | 397 } |
| 522 | 398 |
| 523 AttachBackTextureToFBO(); | 399 AttachBackTextureToFBO(); |
| 524 | 400 |
| 401 const MailboxName& name = mailbox_name(backbuffer_.identifier); | |
| 402 | |
| 525 GpuHostMsg_AcceleratedSurfaceNew_Params params; | 403 GpuHostMsg_AcceleratedSurfaceNew_Params params; |
| 526 params.width = size.width(); | 404 params.width = current_size_.width(); |
| 527 params.height = size.height(); | 405 params.height = current_size_.height(); |
| 528 params.surface_handle = texture.client_id; | 406 params.surface_handle = backbuffer_.identifier; |
| 407 params.mailbox_name.append( | |
| 408 reinterpret_cast<const char*>(&name), sizeof(name)); | |
| 529 helper_->SendAcceleratedSurfaceNew(params); | 409 helper_->SendAcceleratedSurfaceNew(params); |
| 530 texture.sent_to_client = true; | |
| 531 } | 410 } |
| 532 | 411 |
| 533 void TextureImageTransportSurface::AttachBackTextureToFBO() { | 412 void TextureImageTransportSurface::AttachBackTextureToFBO() { |
| 534 if (!parent_stub_) | 413 DCHECK(backbuffer_.service_id); |
| 535 return; | |
| 536 TextureInfo* info = textures_[back()].info; | |
| 537 DCHECK(info); | |
| 538 | |
| 539 ScopedFrameBufferBinder fbo_binder(fbo_id_); | 414 ScopedFrameBufferBinder fbo_binder(fbo_id_); |
| 540 glFramebufferTexture2DEXT(GL_FRAMEBUFFER, | 415 glFramebufferTexture2DEXT(GL_FRAMEBUFFER, |
| 541 GL_COLOR_ATTACHMENT0, | 416 GL_COLOR_ATTACHMENT0, |
| 542 GL_TEXTURE_2D, | 417 GL_TEXTURE_2D, |
| 543 info->service_id(), | 418 backbuffer_.service_id, |
| 544 0); | 419 0); |
| 545 glFlush(); | 420 glFlush(); |
| 546 CHECK_GL_ERROR(); | 421 CHECK_GL_ERROR(); |
| 547 | 422 |
| 548 #ifndef NDEBUG | 423 #ifndef NDEBUG |
| 549 GLenum status = glCheckFramebufferStatusEXT(GL_FRAMEBUFFER); | 424 GLenum status = glCheckFramebufferStatusEXT(GL_FRAMEBUFFER); |
| 550 if (status != GL_FRAMEBUFFER_COMPLETE) { | 425 if (status != GL_FRAMEBUFFER_COMPLETE) { |
| 551 DLOG(ERROR) << "Framebuffer incomplete."; | 426 DLOG(FATAL) << "Framebuffer incomplete: " << status; |
| 552 } | 427 } |
| 553 #endif | 428 #endif |
| 554 } | 429 } |
| 555 | 430 |
| 556 void TextureImageTransportSurface::ReleaseParentStub() { | 431 void TextureImageTransportSurface::ConsumeTexture(Texture& texture) { |
| 557 DCHECK(parent_stub_); | 432 DCHECK(!texture.service_id); |
| 558 parent_stub_->RemoveDestructionObserver(this); | 433 DCHECK(texture.identifier == 1 || texture.identifier == 2); |
| 559 for (int i = 0; i < 2; ++i) { | 434 |
| 560 Texture& texture = textures_[i]; | 435 scoped_ptr<TextureDefinition> definition(mailbox_manager_->ConsumeTexture( |
| 561 texture.info = NULL; | 436 GL_TEXTURE_2D, mailbox_name(texture.identifier))); |
| 562 if (!texture.sent_to_client) | 437 if (definition.get()) { |
| 563 continue; | 438 texture.service_id = definition->ReleaseServiceId(); |
| 564 GpuHostMsg_AcceleratedSurfaceRelease_Params params; | 439 texture.size = gfx::Size(definition->level_infos()[0][0].width, |
| 565 params.identifier = texture.client_id; | 440 definition->level_infos()[0][0].height); |
| 566 helper_->SendAcceleratedSurfaceRelease(params); | |
| 567 } | 441 } |
| 568 parent_stub_ = NULL; | 442 } |
| 443 | |
| 444 void TextureImageTransportSurface::ProduceTexture(Texture& texture) { | |
| 445 DCHECK(texture.service_id); | |
| 446 DCHECK(texture.identifier == 1 || texture.identifier == 2); | |
| 447 TextureManager* texture_manager = | |
| 448 helper_->stub()->decoder()->GetContextGroup()->texture_manager(); | |
| 449 DCHECK(texture.size.width() > 0 && texture.size.height() > 0); | |
| 450 TextureDefinition::LevelInfo info( | |
| 451 GL_TEXTURE_2D, GL_RGBA, texture.size.width(), texture.size.height(), 1, | |
| 452 0, GL_RGBA, GL_UNSIGNED_BYTE, true); | |
| 453 | |
| 454 TextureDefinition::LevelInfos level_infos; | |
| 455 level_infos.resize(1); | |
| 456 level_infos[0].resize(texture_manager->MaxLevelsForTarget(GL_TEXTURE_2D)); | |
| 457 level_infos[0][0] = info; | |
| 458 scoped_ptr<TextureDefinition> definition(new TextureDefinition( | |
| 459 GL_TEXTURE_2D, | |
| 460 texture.service_id, | |
| 461 true, | |
| 462 level_infos)); | |
| 463 // Pass NULL as |owner| here to avoid errors from glConsumeTextureCHROMIUM() | |
| 464 // when the renderer context group goes away before the RWHV handles a pending | |
| 465 // ACK. We avoid leaking a texture in the mailbox by waiting for the final ACK | |
| 466 // at which point we consume the correct texture back. | |
| 467 mailbox_manager_->ProduceTexture( | |
| 468 GL_TEXTURE_2D, | |
| 469 mailbox_name(texture.identifier), | |
| 470 definition.release(), | |
| 471 NULL); | |
| 472 texture.service_id = 0; | |
| 569 } | 473 } |
| 570 | 474 |
| 571 } // namespace content | 475 } // namespace content |
| OLD | NEW |