Chromium Code Reviews| OLD | NEW |
|---|---|
| 1 // Copyright 2012 The Chromium Authors. All rights reserved. | 1 // Copyright 2012 The Chromium Authors. All rights reserved. |
| 2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
| 3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
| 4 | 4 |
| 5 #include "cc/delegated_renderer_layer_impl.h" | 5 #include "cc/delegated_renderer_layer_impl.h" |
| 6 | 6 |
| 7 #include "cc/append_quads_data.h" | 7 #include "cc/append_quads_data.h" |
| 8 #include "cc/math_util.h" | 8 #include "cc/math_util.h" |
| 9 #include "cc/quad_sink.h" | 9 #include "cc/quad_sink.h" |
| 10 #include "cc/render_pass_draw_quad.h" | 10 #include "cc/render_pass_draw_quad.h" |
| 11 #include "cc/render_pass_sink.h" | 11 #include "cc/render_pass_sink.h" |
| 12 | 12 |
| 13 namespace cc { | 13 namespace cc { |
| 14 | 14 |
| 15 DelegatedRendererLayerImpl::DelegatedRendererLayerImpl(LayerTreeImpl* treeImpl, int id) | 15 DelegatedRendererLayerImpl::DelegatedRendererLayerImpl( |
| 16 : LayerImpl(treeImpl, id) | 16 LayerTreeImpl* tree_impl, int id) |
| 17 { | 17 : LayerImpl(tree_impl, id) { |
| 18 } | 18 } |
| 19 | 19 |
| 20 DelegatedRendererLayerImpl::~DelegatedRendererLayerImpl() | 20 DelegatedRendererLayerImpl::~DelegatedRendererLayerImpl() { |
| 21 { | 21 ClearRenderPasses(); |
| 22 clearRenderPasses(); | 22 } |
| 23 } | 23 |
| 24 | 24 bool DelegatedRendererLayerImpl::hasDelegatedContent() const { |
| 25 bool DelegatedRendererLayerImpl::hasDelegatedContent() const | 25 return !render_passes_in_draw_order_.empty(); |
| 26 { | 26 } |
| 27 return !m_renderPassesInDrawOrder.empty(); | 27 |
| 28 } | 28 bool DelegatedRendererLayerImpl::hasContributingDelegatedRenderPasses() const { |
| 29 | 29 // The root RenderPass for the layer is merged with its target |
| 30 bool DelegatedRendererLayerImpl::hasContributingDelegatedRenderPasses() const | 30 // RenderPass in each frame. So we only have extra RenderPasses |
| 31 { | 31 // to merge when we have a non-root RenderPass present. |
| 32 // The root RenderPass for the layer is merged with its target | 32 return render_passes_in_draw_order_.size() > 1; |
| 33 // RenderPass in each frame. So we only have extra RenderPasses | 33 } |
| 34 // to merge when we have a non-root RenderPass present. | 34 |
| 35 return m_renderPassesInDrawOrder.size() > 1; | 35 void DelegatedRendererLayerImpl::SetRenderPasses( |
| 36 } | 36 ScopedPtrVector<RenderPass>& render_passes_in_draw_order) { |
| 37 | 37 gfx::RectF old_root_damage; |
| 38 void DelegatedRendererLayerImpl::setRenderPasses(ScopedPtrVector<RenderPass>& re nderPassesInDrawOrder) | 38 if (!render_passes_in_draw_order_.empty()) |
| 39 { | 39 old_root_damage = render_passes_in_draw_order_.back()->damage_rect; |
| 40 gfx::RectF oldRootDamage; | 40 |
| 41 if (!m_renderPassesInDrawOrder.empty()) | 41 ClearRenderPasses(); |
| 42 oldRootDamage = m_renderPassesInDrawOrder.back()->damage_rect; | 42 |
| 43 | 43 for (size_t i = 0; i < render_passes_in_draw_order.size(); ++i) { |
| 44 clearRenderPasses(); | 44 render_passes_index_by_id_.insert( |
| 45 | 45 std::pair<RenderPass::Id, int>(render_passes_in_draw_order[i]->id, i)); |
| 46 for (size_t i = 0; i < renderPassesInDrawOrder.size(); ++i) { | 46 scoped_ptr<RenderPass> passed_render_pass = |
| 47 m_renderPassesIndexById.insert(std::pair<RenderPass::Id, int>(renderPass esInDrawOrder[i]->id, i)); | 47 render_passes_in_draw_order.take( |
| 48 m_renderPassesInDrawOrder.push_back(renderPassesInDrawOrder.take(renderP assesInDrawOrder.begin() + i)); | 48 render_passes_in_draw_order.begin() + i); |
| 49 render_passes_in_draw_order_.push_back(passed_render_pass.Pass()); | |
| 50 } | |
| 51 render_passes_in_draw_order.clear(); | |
| 52 | |
| 53 if (!render_passes_in_draw_order_.empty()) | |
| 54 render_passes_in_draw_order_.back()->damage_rect.Union(old_root_damage); | |
| 55 } | |
| 56 | |
| 57 void DelegatedRendererLayerImpl::ClearRenderPasses() { | |
| 58 // FIXME: Release the resources back to the nested compositor. | |
| 59 render_passes_index_by_id_.clear(); | |
| 60 render_passes_in_draw_order_.clear(); | |
| 61 } | |
| 62 | |
| 63 void DelegatedRendererLayerImpl::didLoseOutputSurface() { | |
| 64 ClearRenderPasses(); | |
| 65 } | |
| 66 | |
| 67 static inline int IndexToId(int index) { return index + 1; } | |
| 68 static inline int IdToIndex(int id) { return id - 1; } | |
| 69 | |
| 70 RenderPass::Id DelegatedRendererLayerImpl::firstContributingRenderPassId() | |
| 71 const { | |
| 72 return RenderPass::Id(id(), IndexToId(0)); | |
| 73 } | |
| 74 | |
| 75 RenderPass::Id DelegatedRendererLayerImpl::nextContributingRenderPassId( | |
| 76 RenderPass::Id previous) const { | |
| 77 return RenderPass::Id(previous.layer_id, previous.index + 1); | |
| 78 } | |
| 79 | |
| 80 RenderPass::Id DelegatedRendererLayerImpl::ConvertDelegatedRenderPassId( | |
| 81 RenderPass::Id delegated_render_pass_id) const { | |
| 82 base::hash_map<RenderPass::Id, int>::const_iterator found = | |
| 83 render_passes_index_by_id_.find(delegated_render_pass_id); | |
| 84 DCHECK(found != render_passes_index_by_id_.end()); | |
| 85 unsigned delegatedRenderPassIndex = found->second; | |
| 86 return RenderPass::Id(id(), IndexToId(delegatedRenderPassIndex)); | |
| 87 } | |
| 88 | |
| 89 void DelegatedRendererLayerImpl::AppendContributingRenderPasses( | |
| 90 RenderPassSink* render_pass_sink) { | |
| 91 DCHECK(hasContributingDelegatedRenderPasses()); | |
| 92 | |
| 93 for (size_t i = 0; i < render_passes_in_draw_order_.size() - 1; ++i) { | |
| 94 RenderPass::Id output_render_pass_id = | |
| 95 ConvertDelegatedRenderPassId(render_passes_in_draw_order_[i]->id); | |
| 96 | |
| 97 // Don't clash with the RenderPass we generate if we own a RenderSurface. | |
| 98 DCHECK(output_render_pass_id.index > 0); | |
| 99 | |
| 100 render_pass_sink->appendRenderPass( | |
| 101 render_passes_in_draw_order_[i]->Copy(output_render_pass_id)); | |
| 102 } | |
| 103 } | |
| 104 | |
| 105 void DelegatedRendererLayerImpl::appendQuads( | |
| 106 QuadSink& quad_sink, AppendQuadsData& append_quads_data) { | |
| 107 if (render_passes_in_draw_order_.empty()) | |
| 108 return; | |
| 109 | |
| 110 RenderPass::Id target_render_pass_id = append_quads_data.renderPassId; | |
| 111 | |
| 112 const RenderPass* root_delegated_render_pass = | |
| 113 render_passes_in_draw_order_.back(); | |
| 114 | |
| 115 DCHECK(root_delegated_render_pass->output_rect.origin().IsOrigin()); | |
| 116 gfx::Size frame_size = root_delegated_render_pass->output_rect.size(); | |
| 117 | |
| 118 // If the index of the renderPassId is 0, then it is a renderPass generated | |
| 119 // for a layer in this compositor, not the delegated renderer. Then we want to | |
| 120 // merge our root renderPass with the target renderPass. Otherwise, it is some | |
| 121 // renderPass which we added from the delegated renderer. | |
| 122 bool should_merge_root_render_pass_with_target = !target_render_pass_id.index; | |
| 123 if (should_merge_root_render_pass_with_target) { | |
| 124 // Verify that the renderPass we are appending to is created our | |
| 125 // renderTarget. | |
| 126 DCHECK(target_render_pass_id.layer_id == renderTarget()->id()); | |
| 127 | |
| 128 AppendRenderPassQuads( | |
| 129 &quad_sink, &append_quads_data, root_delegated_render_pass, frame_size); | |
| 130 } else { | |
| 131 // Verify that the renderPass we are appending to was created by us. | |
| 132 DCHECK(target_render_pass_id.layer_id == id()); | |
| 133 | |
| 134 int render_pass_index = IdToIndex(target_render_pass_id.index); | |
| 135 const RenderPass* delegated_render_pass = | |
| 136 render_passes_in_draw_order_[render_pass_index]; | |
| 137 AppendRenderPassQuads( | |
| 138 &quad_sink, &append_quads_data, delegated_render_pass, frame_size); | |
| 139 } | |
| 140 } | |
| 141 | |
| 142 void DelegatedRendererLayerImpl::AppendRenderPassQuads( | |
| 143 QuadSink* quad_sink, | |
| 144 AppendQuadsData* append_quads_data, | |
| 145 const RenderPass* delegated_render_pass, | |
| 146 gfx::Size frame_size) const { | |
| 147 | |
| 148 const SharedQuadState* delegated_shared_quad_state = NULL; | |
| 149 SharedQuadState* output_shared_quad_state = NULL; | |
| 150 | |
| 151 for (size_t i = 0; i < delegated_render_pass->quad_list.size(); ++i) { | |
| 152 const DrawQuad* delegated_quad = delegated_render_pass->quad_list[i]; | |
| 153 | |
| 154 if (delegated_quad->shared_quad_state != delegated_shared_quad_state) { | |
| 155 delegated_shared_quad_state = delegated_quad->shared_quad_state; | |
| 156 output_shared_quad_state = quad_sink->useSharedQuadState( | |
| 157 delegated_shared_quad_state->Copy()); | |
| 158 | |
| 159 bool is_root_delegated_render_pass = | |
| 160 delegated_render_pass == render_passes_in_draw_order_.back(); | |
| 161 if (is_root_delegated_render_pass) { | |
| 162 // Don't allow areas inside the bounds that are empty. | |
| 163 DCHECK(display_size_.IsEmpty() || | |
| 164 gfx::Rect(display_size_).Contains(gfx::Rect(bounds()))); | |
| 165 gfx::Size display_size = | |
| 166 display_size_.IsEmpty() ? bounds() : display_size_; | |
| 167 | |
| 168 gfx::Transform delegated_frame_to_layer_space_transform; | |
| 169 delegated_frame_to_layer_space_transform.Scale( | |
| 170 static_cast<double>(display_size.width()) / frame_size.width(), | |
| 171 static_cast<double>(display_size.height()) / frame_size.height()); | |
| 172 | |
| 173 gfx::Transform delegated_frame_to_target_transform = | |
| 174 drawTransform() * delegated_frame_to_layer_space_transform; | |
| 175 | |
| 176 output_shared_quad_state->content_to_target_transform.ConcatTransform( | |
| 177 delegated_frame_to_target_transform); | |
| 178 | |
| 179 if (renderTarget() == this) { | |
| 180 DCHECK(!isClipped()); | |
| 181 DCHECK(renderSurface()); | |
| 182 output_shared_quad_state->clip_rect = MathUtil::mapClippedRect( | |
| 183 delegated_frame_to_target_transform, | |
| 184 output_shared_quad_state->clip_rect); | |
| 185 } else { | |
| 186 gfx::Rect clip_rect = drawableContentRect(); | |
| 187 if (output_shared_quad_state->is_clipped) { | |
| 188 clip_rect.Intersect(MathUtil::mapClippedRect( | |
| 189 delegated_frame_to_target_transform, | |
| 190 output_shared_quad_state->clip_rect)); | |
| 191 } | |
| 192 output_shared_quad_state->clip_rect = clip_rect; | |
| 193 output_shared_quad_state->is_clipped = true; | |
| 194 } | |
| 195 | |
| 196 output_shared_quad_state->opacity *= drawOpacity(); | |
| 197 } | |
| 49 } | 198 } |
| 50 renderPassesInDrawOrder.clear(); | 199 DCHECK(output_shared_quad_state); |
| 51 | 200 |
| 52 if (!m_renderPassesInDrawOrder.empty()) | 201 scoped_ptr<DrawQuad> output_quad; |
| 53 m_renderPassesInDrawOrder.back()->damage_rect.Union(oldRootDamage); | 202 if (delegated_quad->material != DrawQuad::RENDER_PASS) |
| 54 } | 203 output_quad = delegated_quad->Copy(output_shared_quad_state); |
| 55 | 204 else { |
|
piman
2013/01/22 18:52:23
nit: need brackets around first clause if second o
danakj
2013/01/22 19:00:09
Ah, thanks. Done.
| |
| 56 void DelegatedRendererLayerImpl::clearRenderPasses() | 205 RenderPass::Id delegated_contributing_render_pass_id = |
| 57 { | 206 RenderPassDrawQuad::MaterialCast(delegated_quad)->render_pass_id; |
| 58 // FIXME: Release the resources back to the nested compositor. | 207 RenderPass::Id output_contributing_render_pass_id = |
| 59 m_renderPassesIndexById.clear(); | 208 ConvertDelegatedRenderPassId(delegated_contributing_render_pass_id); |
| 60 m_renderPassesInDrawOrder.clear(); | 209 DCHECK(output_contributing_render_pass_id != |
| 61 } | 210 append_quads_data->renderPassId); |
| 62 | 211 |
| 63 void DelegatedRendererLayerImpl::didLoseOutputSurface() | 212 output_quad = RenderPassDrawQuad::MaterialCast(delegated_quad)->Copy( |
| 64 { | 213 output_shared_quad_state, |
| 65 clearRenderPasses(); | 214 output_contributing_render_pass_id).PassAs<DrawQuad>(); |
| 66 } | |
| 67 | |
| 68 static inline int indexToId(int index) { return index + 1; } | |
| 69 static inline int idToIndex(int id) { return id - 1; } | |
| 70 | |
| 71 RenderPass::Id DelegatedRendererLayerImpl::firstContributingRenderPassId() const | |
| 72 { | |
| 73 return RenderPass::Id(id(), indexToId(0)); | |
| 74 } | |
| 75 | |
| 76 RenderPass::Id DelegatedRendererLayerImpl::nextContributingRenderPassId(RenderPa ss::Id previous) const | |
| 77 { | |
| 78 return RenderPass::Id(previous.layer_id, previous.index + 1); | |
| 79 } | |
| 80 | |
| 81 RenderPass::Id DelegatedRendererLayerImpl::convertDelegatedRenderPassId(RenderPa ss::Id delegatedRenderPassId) const | |
| 82 { | |
| 83 base::hash_map<RenderPass::Id, int>::const_iterator it = m_renderPassesIndex ById.find(delegatedRenderPassId); | |
| 84 DCHECK(it != m_renderPassesIndexById.end()); | |
| 85 unsigned delegatedRenderPassIndex = it->second; | |
| 86 return RenderPass::Id(id(), indexToId(delegatedRenderPassIndex)); | |
| 87 } | |
| 88 | |
| 89 void DelegatedRendererLayerImpl::appendContributingRenderPasses(RenderPassSink& renderPassSink) | |
| 90 { | |
| 91 DCHECK(hasContributingDelegatedRenderPasses()); | |
| 92 | |
| 93 for (size_t i = 0; i < m_renderPassesInDrawOrder.size() - 1; ++i) { | |
| 94 RenderPass::Id outputRenderPassId = convertDelegatedRenderPassId(m_rende rPassesInDrawOrder[i]->id); | |
| 95 | |
| 96 // Don't clash with the RenderPass we generate if we own a RenderSurface Impl. | |
| 97 DCHECK(outputRenderPassId.index > 0); | |
| 98 | |
| 99 renderPassSink.appendRenderPass(m_renderPassesInDrawOrder[i]->Copy(outpu tRenderPassId)); | |
| 100 } | 215 } |
| 101 } | 216 DCHECK(output_quad.get()); |
| 102 | 217 |
| 103 void DelegatedRendererLayerImpl::appendQuads(QuadSink& quadSink, AppendQuadsData & appendQuadsData) | 218 quad_sink->append(output_quad.Pass(), *append_quads_data); |
| 104 { | 219 } |
| 105 if (m_renderPassesInDrawOrder.empty()) | 220 } |
| 106 return; | 221 |
| 107 | 222 const char* DelegatedRendererLayerImpl::layerTypeAsString() const { |
| 108 RenderPass::Id targetRenderPassId = appendQuadsData.renderPassId; | 223 return "DelegatedRendererLayer"; |
| 109 | |
| 110 const RenderPass* rootDelegatedRenderPass = m_renderPassesInDrawOrder.back() ; | |
| 111 | |
| 112 DCHECK(rootDelegatedRenderPass->output_rect.origin().IsOrigin()); | |
| 113 gfx::Size frameSize = rootDelegatedRenderPass->output_rect.size(); | |
| 114 | |
| 115 // If the index of the renderPassId is 0, then it is a renderPass generated for a layer | |
| 116 // in this compositor, not the delegated renderer. Then we want to merge our root renderPass with | |
| 117 // the target renderPass. Otherwise, it is some renderPass which we added fr om the delegated | |
| 118 // renderer. | |
| 119 bool shouldMergeRootRenderPassWithTarget = !targetRenderPassId.index; | |
| 120 if (shouldMergeRootRenderPassWithTarget) { | |
| 121 // Verify that the renderPass we are appending to is created our renderT arget. | |
| 122 DCHECK(targetRenderPassId.layer_id == renderTarget()->id()); | |
| 123 | |
| 124 appendRenderPassQuads(quadSink, appendQuadsData, rootDelegatedRenderPass , frameSize); | |
| 125 } else { | |
| 126 // Verify that the renderPass we are appending to was created by us. | |
| 127 DCHECK(targetRenderPassId.layer_id == id()); | |
| 128 | |
| 129 int renderPassIndex = idToIndex(targetRenderPassId.index); | |
| 130 const RenderPass* delegatedRenderPass = m_renderPassesInDrawOrder[render PassIndex]; | |
| 131 appendRenderPassQuads(quadSink, appendQuadsData, delegatedRenderPass, fr ameSize); | |
| 132 } | |
| 133 } | |
| 134 | |
| 135 void DelegatedRendererLayerImpl::appendRenderPassQuads(QuadSink& quadSink, Appen dQuadsData& appendQuadsData, const RenderPass* delegatedRenderPass, gfx::Size fr ameSize) const | |
| 136 { | |
| 137 const SharedQuadState* currentSharedQuadState = 0; | |
| 138 SharedQuadState* copiedSharedQuadState = 0; | |
| 139 for (size_t i = 0; i < delegatedRenderPass->quad_list.size(); ++i) { | |
| 140 const DrawQuad* quad = delegatedRenderPass->quad_list[i]; | |
| 141 | |
| 142 if (quad->shared_quad_state != currentSharedQuadState) { | |
| 143 currentSharedQuadState = quad->shared_quad_state; | |
| 144 copiedSharedQuadState = quadSink.useSharedQuadState(currentSharedQua dState->Copy()); | |
| 145 bool isRootDelegatedRenderPass = delegatedRenderPass == m_renderPass esInDrawOrder.back(); | |
| 146 if (isRootDelegatedRenderPass) { | |
| 147 // Don't allow areas inside the bounds that are empty. | |
| 148 DCHECK(m_displaySize.IsEmpty() || gfx::Rect(m_displaySize).Contain s(gfx::Rect(bounds()))); | |
| 149 gfx::Size displaySize = m_displaySize.IsEmpty() ? bounds() : m_dis playSize; | |
| 150 | |
| 151 gfx::Transform delegatedFrameToLayerSpaceTransform; | |
| 152 delegatedFrameToLayerSpaceTransform.Scale( | |
| 153 static_cast<double>(displaySize.width()) / frameSize.width(), | |
| 154 static_cast<double>(displaySize.height()) / frameSize.height() ); | |
| 155 | |
| 156 copiedSharedQuadState->content_to_target_transform = drawTransform () * delegatedFrameToLayerSpaceTransform * copiedSharedQuadState->content_to_tar get_transform; | |
| 157 | |
| 158 if (renderTarget() == this) { | |
| 159 DCHECK(!isClipped()); | |
| 160 DCHECK(renderSurface()); | |
| 161 copiedSharedQuadState->clip_rect = MathUtil::mapClippedRect( | |
| 162 drawTransform() * delegatedFrameToLayerSpaceTransform, | |
| 163 copiedSharedQuadState->clip_rect); | |
| 164 } else { | |
| 165 gfx::Rect clip_rect = drawableContentRect(); | |
| 166 if (copiedSharedQuadState->is_clipped) { | |
| 167 clip_rect.Intersect(MathUtil::mapClippedRect( | |
| 168 drawTransform() * delegatedFrameToLayerSpaceTransform, | |
| 169 copiedSharedQuadState->clip_rect)); | |
| 170 } | |
| 171 copiedSharedQuadState->clip_rect = clip_rect; | |
| 172 copiedSharedQuadState->is_clipped = true; | |
| 173 } | |
| 174 | |
| 175 copiedSharedQuadState->opacity *= drawOpacity(); | |
| 176 } | |
| 177 } | |
| 178 DCHECK(copiedSharedQuadState); | |
| 179 | |
| 180 scoped_ptr<DrawQuad> copyQuad; | |
| 181 if (quad->material != DrawQuad::RENDER_PASS) | |
| 182 copyQuad = quad->Copy(copiedSharedQuadState); | |
| 183 else { | |
| 184 RenderPass::Id contributingDelegatedRenderPassId = RenderPassDrawQua d::MaterialCast(quad)->render_pass_id; | |
| 185 RenderPass::Id contributingRenderPassId = convertDelegatedRenderPass Id(contributingDelegatedRenderPassId); | |
| 186 DCHECK(contributingRenderPassId != appendQuadsData.renderPassId); | |
| 187 | |
| 188 copyQuad = RenderPassDrawQuad::MaterialCast(quad)->Copy(copiedShared QuadState, contributingRenderPassId).PassAs<DrawQuad>(); | |
| 189 } | |
| 190 DCHECK(copyQuad.get()); | |
| 191 | |
| 192 quadSink.append(copyQuad.Pass(), appendQuadsData); | |
| 193 } | |
| 194 } | |
| 195 | |
| 196 const char* DelegatedRendererLayerImpl::layerTypeAsString() const | |
| 197 { | |
| 198 return "DelegatedRendererLayer"; | |
| 199 } | 224 } |
| 200 | 225 |
| 201 } // namespace cc | 226 } // namespace cc |
| OLD | NEW |