| OLD | NEW |
| 1 // Copyright 2012 The Chromium Authors. All rights reserved. | 1 // Copyright 2012 The Chromium Authors. All rights reserved. |
| 2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
| 3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
| 4 | 4 |
| 5 #include "cc/layers/delegated_renderer_layer_impl.h" | 5 #include "cc/layers/delegated_renderer_layer_impl.h" |
| 6 | 6 |
| 7 #include <algorithm> | 7 #include <algorithm> |
| 8 #include <utility> | 8 #include <utility> |
| 9 | 9 |
| 10 #include "base/bind.h" | 10 #include "base/bind.h" |
| 11 #include "base/containers/hash_tables.h" | 11 #include "base/containers/hash_tables.h" |
| 12 #include "cc/base/math_util.h" | 12 #include "cc/base/math_util.h" |
| 13 #include "cc/layers/append_quads_data.h" | 13 #include "cc/layers/append_quads_data.h" |
| 14 #include "cc/layers/quad_sink.h" | 14 #include "cc/layers/quad_sink.h" |
| 15 #include "cc/layers/render_pass_sink.h" | 15 #include "cc/layers/render_pass_sink.h" |
| 16 #include "cc/output/delegated_frame_data.h" | 16 #include "cc/output/delegated_frame_data.h" |
| 17 #include "cc/quads/render_pass_draw_quad.h" | 17 #include "cc/quads/render_pass_draw_quad.h" |
| 18 #include "cc/quads/solid_color_draw_quad.h" | 18 #include "cc/quads/solid_color_draw_quad.h" |
| 19 #include "cc/trees/layer_tree_impl.h" | 19 #include "cc/trees/layer_tree_impl.h" |
| 20 | 20 |
| 21 namespace cc { | 21 namespace cc { |
| 22 | 22 |
| 23 DelegatedRendererLayerImpl::DelegatedRendererLayerImpl(LayerTreeImpl* tree_impl, | 23 DelegatedRendererLayerImpl::DelegatedRendererLayerImpl( |
| 24 int id) | 24 LayerTreeImpl* tree_impl, int id) |
| 25 : LayerImpl(tree_impl, id), | 25 : LayerImpl(tree_impl, id), |
| 26 have_render_passes_to_push_(false), | 26 have_render_passes_to_push_(false), |
| 27 inverse_device_scale_factor_(1.0f), | |
| 28 child_id_(0), | 27 child_id_(0), |
| 29 own_child_id_(false) { | 28 own_child_id_(false) { |
| 30 } | 29 } |
| 31 | 30 |
| 32 DelegatedRendererLayerImpl::~DelegatedRendererLayerImpl() { | 31 DelegatedRendererLayerImpl::~DelegatedRendererLayerImpl() { |
| 33 ClearRenderPasses(); | 32 ClearRenderPasses(); |
| 34 ClearChildId(); | 33 ClearChildId(); |
| 35 } | 34 } |
| 36 | 35 |
| 37 bool DelegatedRendererLayerImpl::HasDelegatedContent() const { return true; } | 36 bool DelegatedRendererLayerImpl::HasDelegatedContent() const { return true; } |
| (...skipping 97 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 135 if (invalid_frame) { | 134 if (invalid_frame) { |
| 136 // Declare we are still using the last frame's resources. | 135 // Declare we are still using the last frame's resources. |
| 137 resource_provider->DeclareUsedResourcesFromChild(child_id_, resources_); | 136 resource_provider->DeclareUsedResourcesFromChild(child_id_, resources_); |
| 138 return; | 137 return; |
| 139 } | 138 } |
| 140 | 139 |
| 141 // Declare we are using the new frame's resources. | 140 // Declare we are using the new frame's resources. |
| 142 resources_.swap(resources_in_frame); | 141 resources_.swap(resources_in_frame); |
| 143 resource_provider->DeclareUsedResourcesFromChild(child_id_, resources_); | 142 resource_provider->DeclareUsedResourcesFromChild(child_id_, resources_); |
| 144 | 143 |
| 145 inverse_device_scale_factor_ = 1.0f / frame_data->device_scale_factor; | |
| 146 // Display size is already set so we can compute what the damage rect | 144 // Display size is already set so we can compute what the damage rect |
| 147 // will be in layer space. The damage may exceed the visible portion of | 145 // will be in layer space. The damage may exceed the visible portion of |
| 148 // the frame, so intersect the damage to the layer's bounds. | 146 // the frame, so intersect the damage to the layer's bounds. |
| 149 RenderPass* new_root_pass = render_pass_list.back(); | 147 RenderPass* new_root_pass = render_pass_list.back(); |
| 150 gfx::Size frame_size = new_root_pass->output_rect.size(); | 148 gfx::Size frame_size = new_root_pass->output_rect.size(); |
| 151 gfx::RectF damage_in_layer = damage_in_frame; | 149 gfx::RectF damage_in_layer = MathUtil::MapClippedRect( |
| 152 damage_in_layer.Scale(inverse_device_scale_factor_); | 150 DelegatedFrameToLayerSpaceTransform(frame_size), damage_in_frame); |
| 153 SetUpdateRect(gfx::IntersectRects( | 151 SetUpdateRect(gfx::IntersectRects( |
| 154 gfx::UnionRects(update_rect(), damage_in_layer), gfx::Rect(bounds()))); | 152 gfx::UnionRects(update_rect(), damage_in_layer), gfx::Rect(bounds()))); |
| 155 | 153 |
| 156 SetRenderPasses(&render_pass_list); | 154 SetRenderPasses(&render_pass_list); |
| 157 have_render_passes_to_push_ = true; | 155 have_render_passes_to_push_ = true; |
| 158 } | 156 } |
| 159 | 157 |
| 160 void DelegatedRendererLayerImpl::SetDisplaySize(const gfx::Size& size) { | 158 void DelegatedRendererLayerImpl::SetDisplaySize(const gfx::Size& size) { |
| 161 if (display_size_ == size) | 159 if (display_size_ == size) |
| 162 return; | 160 return; |
| (...skipping 28 matching lines...) Expand all Loading... |
| 191 LayerTreeImpl* tree_impl) { | 189 LayerTreeImpl* tree_impl) { |
| 192 return DelegatedRendererLayerImpl::Create( | 190 return DelegatedRendererLayerImpl::Create( |
| 193 tree_impl, id()).PassAs<LayerImpl>(); | 191 tree_impl, id()).PassAs<LayerImpl>(); |
| 194 } | 192 } |
| 195 | 193 |
| 196 void DelegatedRendererLayerImpl::ReleaseResources() { | 194 void DelegatedRendererLayerImpl::ReleaseResources() { |
| 197 ClearRenderPasses(); | 195 ClearRenderPasses(); |
| 198 ClearChildId(); | 196 ClearChildId(); |
| 199 } | 197 } |
| 200 | 198 |
| 199 gfx::Transform DelegatedRendererLayerImpl::DelegatedFrameToLayerSpaceTransform( |
| 200 const gfx::Size& frame_size) const { |
| 201 gfx::Size display_size = display_size_.IsEmpty() ? bounds() : display_size_; |
| 202 |
| 203 gfx::Transform delegated_frame_to_layer_space_transform; |
| 204 delegated_frame_to_layer_space_transform.Scale( |
| 205 static_cast<double>(display_size.width()) / frame_size.width(), |
| 206 static_cast<double>(display_size.height()) / frame_size.height()); |
| 207 return delegated_frame_to_layer_space_transform; |
| 208 } |
| 209 |
| 201 static inline int IndexToId(int index) { return index + 1; } | 210 static inline int IndexToId(int index) { return index + 1; } |
| 202 static inline int IdToIndex(int id) { return id - 1; } | 211 static inline int IdToIndex(int id) { return id - 1; } |
| 203 | 212 |
| 204 RenderPass::Id DelegatedRendererLayerImpl::FirstContributingRenderPassId() | 213 RenderPass::Id DelegatedRendererLayerImpl::FirstContributingRenderPassId() |
| 205 const { | 214 const { |
| 206 return RenderPass::Id(id(), IndexToId(0)); | 215 return RenderPass::Id(id(), IndexToId(0)); |
| 207 } | 216 } |
| 208 | 217 |
| 209 RenderPass::Id DelegatedRendererLayerImpl::NextContributingRenderPassId( | 218 RenderPass::Id DelegatedRendererLayerImpl::NextContributingRenderPassId( |
| 210 RenderPass::Id previous) const { | 219 RenderPass::Id previous) const { |
| (...skipping 15 matching lines...) Expand all Loading... |
| 226 return true; | 235 return true; |
| 227 } | 236 } |
| 228 | 237 |
| 229 void DelegatedRendererLayerImpl::AppendContributingRenderPasses( | 238 void DelegatedRendererLayerImpl::AppendContributingRenderPasses( |
| 230 RenderPassSink* render_pass_sink) { | 239 RenderPassSink* render_pass_sink) { |
| 231 DCHECK(HasContributingDelegatedRenderPasses()); | 240 DCHECK(HasContributingDelegatedRenderPasses()); |
| 232 | 241 |
| 233 const RenderPass* root_delegated_render_pass = | 242 const RenderPass* root_delegated_render_pass = |
| 234 render_passes_in_draw_order_.back(); | 243 render_passes_in_draw_order_.back(); |
| 235 gfx::Size frame_size = root_delegated_render_pass->output_rect.size(); | 244 gfx::Size frame_size = root_delegated_render_pass->output_rect.size(); |
| 236 gfx::Transform delegated_frame_to_root_transform = screen_space_transform(); | 245 gfx::Transform delegated_frame_to_root_transform = |
| 237 delegated_frame_to_root_transform.Scale(inverse_device_scale_factor_, | 246 screen_space_transform() * |
| 238 inverse_device_scale_factor_); | 247 DelegatedFrameToLayerSpaceTransform(frame_size); |
| 239 | 248 |
| 240 for (size_t i = 0; i < render_passes_in_draw_order_.size() - 1; ++i) { | 249 for (size_t i = 0; i < render_passes_in_draw_order_.size() - 1; ++i) { |
| 241 RenderPass::Id output_render_pass_id(-1, -1); | 250 RenderPass::Id output_render_pass_id(-1, -1); |
| 242 bool present = | 251 bool present = |
| 243 ConvertDelegatedRenderPassId(render_passes_in_draw_order_[i]->id, | 252 ConvertDelegatedRenderPassId(render_passes_in_draw_order_[i]->id, |
| 244 &output_render_pass_id); | 253 &output_render_pass_id); |
| 245 | 254 |
| 246 // Don't clash with the RenderPass we generate if we own a RenderSurface. | 255 // Don't clash with the RenderPass we generate if we own a RenderSurface. |
| 247 DCHECK(present) << render_passes_in_draw_order_[i]->id.layer_id << ", " | 256 DCHECK(present) << render_passes_in_draw_order_[i]->id.layer_id << ", " |
| 248 << render_passes_in_draw_order_[i]->id.index; | 257 << render_passes_in_draw_order_[i]->id.index; |
| (...skipping 151 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 400 delegated_shared_quad_state = delegated_quad->shared_quad_state; | 409 delegated_shared_quad_state = delegated_quad->shared_quad_state; |
| 401 output_shared_quad_state = quad_sink->CreateSharedQuadState(); | 410 output_shared_quad_state = quad_sink->CreateSharedQuadState(); |
| 402 output_shared_quad_state->CopyFrom(delegated_shared_quad_state); | 411 output_shared_quad_state->CopyFrom(delegated_shared_quad_state); |
| 403 | 412 |
| 404 bool is_root_delegated_render_pass = | 413 bool is_root_delegated_render_pass = |
| 405 delegated_render_pass == render_passes_in_draw_order_.back(); | 414 delegated_render_pass == render_passes_in_draw_order_.back(); |
| 406 if (is_root_delegated_render_pass) { | 415 if (is_root_delegated_render_pass) { |
| 407 // Don't allow areas inside the bounds that are empty. | 416 // Don't allow areas inside the bounds that are empty. |
| 408 DCHECK(display_size_.IsEmpty() || | 417 DCHECK(display_size_.IsEmpty() || |
| 409 gfx::Rect(display_size_).Contains(gfx::Rect(bounds()))); | 418 gfx::Rect(display_size_).Contains(gfx::Rect(bounds()))); |
| 410 gfx::Transform delegated_frame_to_target_transform = draw_transform(); | 419 gfx::Transform delegated_frame_to_target_transform = |
| 411 delegated_frame_to_target_transform.Scale(inverse_device_scale_factor_, | 420 draw_transform() * DelegatedFrameToLayerSpaceTransform(frame_size); |
| 412 inverse_device_scale_factor_); | |
| 413 | 421 |
| 414 output_shared_quad_state->content_to_target_transform.ConcatTransform( | 422 output_shared_quad_state->content_to_target_transform.ConcatTransform( |
| 415 delegated_frame_to_target_transform); | 423 delegated_frame_to_target_transform); |
| 416 | 424 |
| 417 if (render_target() == this) { | 425 if (render_target() == this) { |
| 418 DCHECK(!is_clipped()); | 426 DCHECK(!is_clipped()); |
| 419 DCHECK(render_surface()); | 427 DCHECK(render_surface()); |
| 420 DCHECK_EQ(0, num_unclipped_descendants()); | 428 DCHECK_EQ(0, num_unclipped_descendants()); |
| 421 output_shared_quad_state->clip_rect = | 429 output_shared_quad_state->clip_rect = |
| 422 MathUtil::MapEnclosingClippedRect( | 430 MathUtil::MapEnclosingClippedRect( |
| (...skipping 63 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 486 if (own_child_id_) { | 494 if (own_child_id_) { |
| 487 ResourceProvider* provider = layer_tree_impl()->resource_provider(); | 495 ResourceProvider* provider = layer_tree_impl()->resource_provider(); |
| 488 provider->DestroyChild(child_id_); | 496 provider->DestroyChild(child_id_); |
| 489 } | 497 } |
| 490 | 498 |
| 491 resources_.clear(); | 499 resources_.clear(); |
| 492 child_id_ = 0; | 500 child_id_ = 0; |
| 493 } | 501 } |
| 494 | 502 |
| 495 } // namespace cc | 503 } // namespace cc |
| OLD | NEW |