| OLD | NEW |
| 1 /* | 1 /* |
| 2 * Copyright 2015 Google Inc. | 2 * Copyright 2015 Google Inc. |
| 3 * | 3 * |
| 4 * Use of this source code is governed by a BSD-style license that can be | 4 * Use of this source code is governed by a BSD-style license that can be |
| 5 * found in the LICENSE file. | 5 * found in the LICENSE file. |
| 6 */ | 6 */ |
| 7 | 7 |
| 8 #include "GrVkCommandBuffer.h" | 8 #include "GrVkCommandBuffer.h" |
| 9 | 9 |
| 10 #include "GrVkFramebuffer.h" | 10 #include "GrVkFramebuffer.h" |
| 11 #include "GrVkImageView.h" | 11 #include "GrVkImageView.h" |
| 12 #include "GrVkPipeline.h" | 12 #include "GrVkPipeline.h" |
| 13 #include "GrVkRenderPass.h" | 13 #include "GrVkRenderPass.h" |
| 14 #include "GrVkRenderTarget.h" | 14 #include "GrVkRenderTarget.h" |
| 15 #include "GrVkPipelineState.h" | 15 #include "GrVkPipelineState.h" |
| 16 #include "GrVkTransferBuffer.h" | 16 #include "GrVkTransferBuffer.h" |
| 17 #include "GrVkUtil.h" | 17 #include "GrVkUtil.h" |
| 18 | 18 |
| 19 GrVkCommandBuffer* GrVkCommandBuffer::Create(const GrVkGpu* gpu, VkCommandPool c
mdPool) { | |
| 20 const VkCommandBufferAllocateInfo cmdInfo = { | |
| 21 VK_STRUCTURE_TYPE_COMMAND_BUFFER_ALLOCATE_INFO, // sType | |
| 22 NULL, // pNext | |
| 23 cmdPool, // commandPool | |
| 24 VK_COMMAND_BUFFER_LEVEL_PRIMARY, // level | |
| 25 1 // bufferCount | |
| 26 }; | |
| 27 | |
| 28 VkCommandBuffer cmdBuffer; | |
| 29 VkResult err = GR_VK_CALL(gpu->vkInterface(), AllocateCommandBuffers(gpu->de
vice(), | |
| 30 &cmdInf
o, | |
| 31 &cmdBuf
fer)); | |
| 32 if (err) { | |
| 33 return nullptr; | |
| 34 } | |
| 35 return new GrVkCommandBuffer(cmdBuffer); | |
| 36 } | |
| 37 | 19 |
| 38 GrVkCommandBuffer::~GrVkCommandBuffer() { | 20 GrVkCommandBuffer::~GrVkCommandBuffer() { |
| 39 // Should have ended any render pass we're in the middle of | 21 // Should have ended any render pass we're in the middle of |
| 40 SkASSERT(!fActiveRenderPass); | 22 SkASSERT(!fActiveRenderPass); |
| 41 } | 23 } |
| 42 | 24 |
| 43 void GrVkCommandBuffer::invalidateState() { | 25 void GrVkCommandBuffer::invalidateState() { |
| 44 fBoundVertexBuffer = VK_NULL_HANDLE; | 26 fBoundVertexBuffer = VK_NULL_HANDLE; |
| 45 fBoundVertexBufferIsValid = false; | 27 fBoundVertexBufferIsValid = false; |
| 46 fBoundIndexBuffer = VK_NULL_HANDLE; | 28 fBoundIndexBuffer = VK_NULL_HANDLE; |
| (...skipping 25 matching lines...) Expand all Loading... |
| 72 GR_VK_CALL(gpu->vkInterface(), FreeCommandBuffers(gpu->device(), gpu->cmdPoo
l(), | 54 GR_VK_CALL(gpu->vkInterface(), FreeCommandBuffers(gpu->device(), gpu->cmdPoo
l(), |
| 73 1, &fCmdBuffer)); | 55 1, &fCmdBuffer)); |
| 74 } | 56 } |
| 75 | 57 |
| 76 void GrVkCommandBuffer::abandonSubResources() const { | 58 void GrVkCommandBuffer::abandonSubResources() const { |
| 77 for (int i = 0; i < fTrackedResources.count(); ++i) { | 59 for (int i = 0; i < fTrackedResources.count(); ++i) { |
| 78 fTrackedResources[i]->unrefAndAbandon(); | 60 fTrackedResources[i]->unrefAndAbandon(); |
| 79 } | 61 } |
| 80 } | 62 } |
| 81 | 63 |
| 82 void GrVkCommandBuffer::begin(const GrVkGpu* gpu) { | 64 //////////////////////////////////////////////////////////////////////////////// |
| 65 // CommandBuffer commands |
| 66 //////////////////////////////////////////////////////////////////////////////// |
| 67 |
| 68 void GrVkCommandBuffer::pipelineBarrier(const GrVkGpu* gpu, |
| 69 VkPipelineStageFlags srcStageMask, |
| 70 VkPipelineStageFlags dstStageMask, |
| 71 bool byRegion, |
| 72 BarrierType barrierType, |
| 73 void* barrier) const { |
| 74 SkASSERT(fIsActive); |
| 75 // For images we can have barriers inside of render passes but they require
us to add more |
| 76 // support in subpasses which need self dependencies to have barriers inside
them. Also, we can |
| 77 // never have buffer barriers inside of a render pass. For now we will just
assert that we are |
| 78 // not in a render pass. |
| 79 SkASSERT(!fActiveRenderPass); |
| 80 VkDependencyFlags dependencyFlags = byRegion ? VK_DEPENDENCY_BY_REGION_BIT :
0; |
| 81 |
| 82 switch (barrierType) { |
| 83 case kMemory_BarrierType: { |
| 84 const VkMemoryBarrier* barrierPtr = reinterpret_cast<VkMemoryBarrier
*>(barrier); |
| 85 GR_VK_CALL(gpu->vkInterface(), CmdPipelineBarrier(fCmdBuffer, srcSta
geMask, |
| 86 dstStageMask, depe
ndencyFlags, |
| 87 1, barrierPtr, |
| 88 0, nullptr, |
| 89 0, nullptr)); |
| 90 break; |
| 91 } |
| 92 |
| 93 case kBufferMemory_BarrierType: { |
| 94 const VkBufferMemoryBarrier* barrierPtr = |
| 95 reinterpret_cast<VkBufferMemory
Barrier*>(barrier); |
| 96 GR_VK_CALL(gpu->vkInterface(), CmdPipelineBarrier(fCmdBuffer, srcSta
geMask, |
| 97 dstStageMask, depe
ndencyFlags, |
| 98 0, nullptr, |
| 99 1, barrierPtr, |
| 100 0, nullptr)); |
| 101 break; |
| 102 } |
| 103 |
| 104 case kImageMemory_BarrierType: { |
| 105 const VkImageMemoryBarrier* barrierPtr = |
| 106 reinterpret_cast<VkImageMemory
Barrier*>(barrier); |
| 107 GR_VK_CALL(gpu->vkInterface(), CmdPipelineBarrier(fCmdBuffer, srcSta
geMask, |
| 108 dstStageMask, depe
ndencyFlags, |
| 109 0, nullptr, |
| 110 0, nullptr, |
| 111 1, barrierPtr)); |
| 112 break; |
| 113 } |
| 114 } |
| 115 |
| 116 } |
| 117 |
| 118 void GrVkCommandBuffer::clearAttachments(const GrVkGpu* gpu, |
| 119 int numAttachments, |
| 120 const VkClearAttachment* attachments, |
| 121 int numRects, |
| 122 const VkClearRect* clearRects) const { |
| 123 SkASSERT(fIsActive); |
| 124 SkASSERT(fActiveRenderPass); |
| 125 SkASSERT(numAttachments > 0); |
| 126 SkASSERT(numRects > 0); |
| 127 #ifdef SK_DEBUG |
| 128 for (int i = 0; i < numAttachments; ++i) { |
| 129 if (attachments[i].aspectMask == VK_IMAGE_ASPECT_COLOR_BIT) { |
| 130 uint32_t testIndex; |
| 131 SkAssertResult(fActiveRenderPass->colorAttachmentIndex(&testIndex)); |
| 132 SkASSERT(testIndex == attachments[i].colorAttachment); |
| 133 } |
| 134 } |
| 135 #endif |
| 136 GR_VK_CALL(gpu->vkInterface(), CmdClearAttachments(fCmdBuffer, |
| 137 numAttachments, |
| 138 attachments, |
| 139 numRects, |
| 140 clearRects)); |
| 141 } |
| 142 |
| 143 void GrVkCommandBuffer::bindDescriptorSets(const GrVkGpu* gpu, |
| 144 GrVkPipelineState* pipelineState, |
| 145 VkPipelineLayout layout, |
| 146 uint32_t firstSet, |
| 147 uint32_t setCount, |
| 148 const VkDescriptorSet* descriptorSets
, |
| 149 uint32_t dynamicOffsetCount, |
| 150 const uint32_t* dynamicOffsets) { |
| 151 SkASSERT(fIsActive); |
| 152 GR_VK_CALL(gpu->vkInterface(), CmdBindDescriptorSets(fCmdBuffer, |
| 153 VK_PIPELINE_BIND_POINT_
GRAPHICS, |
| 154 layout, |
| 155 firstSet, |
| 156 setCount, |
| 157 descriptorSets, |
| 158 dynamicOffsetCount, |
| 159 dynamicOffsets)); |
| 160 pipelineState->addUniformResources(*this); |
| 161 } |
| 162 |
| 163 void GrVkCommandBuffer::bindPipeline(const GrVkGpu* gpu, const GrVkPipeline* pip
eline) { |
| 164 SkASSERT(fIsActive); |
| 165 GR_VK_CALL(gpu->vkInterface(), CmdBindPipeline(fCmdBuffer, |
| 166 VK_PIPELINE_BIND_POINT_GRAPHI
CS, |
| 167 pipeline->pipeline())); |
| 168 addResource(pipeline); |
| 169 } |
| 170 |
| 171 void GrVkCommandBuffer::drawIndexed(const GrVkGpu* gpu, |
| 172 uint32_t indexCount, |
| 173 uint32_t instanceCount, |
| 174 uint32_t firstIndex, |
| 175 int32_t vertexOffset, |
| 176 uint32_t firstInstance) const { |
| 177 SkASSERT(fIsActive); |
| 178 SkASSERT(fActiveRenderPass); |
| 179 GR_VK_CALL(gpu->vkInterface(), CmdDrawIndexed(fCmdBuffer, |
| 180 indexCount, |
| 181 instanceCount, |
| 182 firstIndex, |
| 183 vertexOffset, |
| 184 firstInstance)); |
| 185 } |
| 186 |
| 187 void GrVkCommandBuffer::draw(const GrVkGpu* gpu, |
| 188 uint32_t vertexCount, |
| 189 uint32_t instanceCount, |
| 190 uint32_t firstVertex, |
| 191 uint32_t firstInstance) const { |
| 192 SkASSERT(fIsActive); |
| 193 SkASSERT(fActiveRenderPass); |
| 194 GR_VK_CALL(gpu->vkInterface(), CmdDraw(fCmdBuffer, |
| 195 vertexCount, |
| 196 instanceCount, |
| 197 firstVertex, |
| 198 firstInstance)); |
| 199 } |
| 200 |
| 201 void GrVkCommandBuffer::setViewport(const GrVkGpu* gpu, |
| 202 uint32_t firstViewport, |
| 203 uint32_t viewportCount, |
| 204 const VkViewport* viewports) { |
| 205 SkASSERT(fIsActive); |
| 206 SkASSERT(1 == viewportCount); |
| 207 if (memcmp(viewports, &fCachedViewport, sizeof(VkViewport))) { |
| 208 GR_VK_CALL(gpu->vkInterface(), CmdSetViewport(fCmdBuffer, |
| 209 firstViewport, |
| 210 viewportCount, |
| 211 viewports)); |
| 212 fCachedViewport = viewports[0]; |
| 213 } |
| 214 } |
| 215 |
| 216 void GrVkCommandBuffer::setScissor(const GrVkGpu* gpu, |
| 217 uint32_t firstScissor, |
| 218 uint32_t scissorCount, |
| 219 const VkRect2D* scissors) { |
| 220 SkASSERT(fIsActive); |
| 221 SkASSERT(1 == scissorCount); |
| 222 if (memcmp(scissors, &fCachedScissor, sizeof(VkRect2D))) { |
| 223 GR_VK_CALL(gpu->vkInterface(), CmdSetScissor(fCmdBuffer, |
| 224 firstScissor, |
| 225 scissorCount, |
| 226 scissors)); |
| 227 fCachedScissor = scissors[0]; |
| 228 } |
| 229 } |
| 230 |
| 231 void GrVkCommandBuffer::setBlendConstants(const GrVkGpu* gpu, |
| 232 const float blendConstants[4]) { |
| 233 SkASSERT(fIsActive); |
| 234 if (memcmp(blendConstants, fCachedBlendConstant, 4 * sizeof(float))) { |
| 235 GR_VK_CALL(gpu->vkInterface(), CmdSetBlendConstants(fCmdBuffer, blendCon
stants)); |
| 236 memcpy(fCachedBlendConstant, blendConstants, 4 * sizeof(float)); |
| 237 } |
| 238 } |
| 239 |
| 240 /////////////////////////////////////////////////////////////////////////////// |
| 241 // PrimaryCommandBuffer |
| 242 //////////////////////////////////////////////////////////////////////////////// |
| 243 GrVkPrimaryCommandBuffer* GrVkPrimaryCommandBuffer::Create(const GrVkGpu* gpu, |
| 244 VkCommandPool cmdPool
) { |
| 245 const VkCommandBufferAllocateInfo cmdInfo = { |
| 246 VK_STRUCTURE_TYPE_COMMAND_BUFFER_ALLOCATE_INFO, // sType |
| 247 NULL, // pNext |
| 248 cmdPool, // commandPool |
| 249 VK_COMMAND_BUFFER_LEVEL_PRIMARY, // level |
| 250 1 // bufferCount |
| 251 }; |
| 252 |
| 253 VkCommandBuffer cmdBuffer; |
| 254 VkResult err = GR_VK_CALL(gpu->vkInterface(), AllocateCommandBuffers(gpu->de
vice(), |
| 255 &cmdInf
o, |
| 256 &cmdBuf
fer)); |
| 257 if (err) { |
| 258 return nullptr; |
| 259 } |
| 260 return new GrVkPrimaryCommandBuffer(cmdBuffer); |
| 261 } |
| 262 |
| 263 void GrVkPrimaryCommandBuffer::begin(const GrVkGpu* gpu) { |
| 83 SkASSERT(!fIsActive); | 264 SkASSERT(!fIsActive); |
| 84 VkCommandBufferBeginInfo cmdBufferBeginInfo; | 265 VkCommandBufferBeginInfo cmdBufferBeginInfo; |
| 85 memset(&cmdBufferBeginInfo, 0, sizeof(VkCommandBufferBeginInfo)); | 266 memset(&cmdBufferBeginInfo, 0, sizeof(VkCommandBufferBeginInfo)); |
| 86 cmdBufferBeginInfo.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_BEGIN_INFO; | 267 cmdBufferBeginInfo.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_BEGIN_INFO; |
| 87 cmdBufferBeginInfo.pNext = nullptr; | 268 cmdBufferBeginInfo.pNext = nullptr; |
| 88 cmdBufferBeginInfo.flags = VK_COMMAND_BUFFER_USAGE_ONE_TIME_SUBMIT_BIT; | 269 cmdBufferBeginInfo.flags = VK_COMMAND_BUFFER_USAGE_ONE_TIME_SUBMIT_BIT; |
| 89 cmdBufferBeginInfo.pInheritanceInfo = nullptr; | 270 cmdBufferBeginInfo.pInheritanceInfo = nullptr; |
| 90 | 271 |
| 91 GR_VK_CALL_ERRCHECK(gpu->vkInterface(), BeginCommandBuffer(fCmdBuffer, | 272 GR_VK_CALL_ERRCHECK(gpu->vkInterface(), BeginCommandBuffer(fCmdBuffer, |
| 92 &cmdBufferBeginIn
fo)); | 273 &cmdBufferBeginIn
fo)); |
| 93 fIsActive = true; | 274 fIsActive = true; |
| 94 } | 275 } |
| 95 | 276 |
| 96 void GrVkCommandBuffer::end(const GrVkGpu* gpu) { | 277 void GrVkPrimaryCommandBuffer::end(const GrVkGpu* gpu) { |
| 97 SkASSERT(fIsActive); | 278 SkASSERT(fIsActive); |
| 98 SkASSERT(!fActiveRenderPass); | 279 SkASSERT(!fActiveRenderPass); |
| 99 GR_VK_CALL_ERRCHECK(gpu->vkInterface(), EndCommandBuffer(fCmdBuffer)); | 280 GR_VK_CALL_ERRCHECK(gpu->vkInterface(), EndCommandBuffer(fCmdBuffer)); |
| 100 this->invalidateState(); | 281 this->invalidateState(); |
| 101 fIsActive = false; | 282 fIsActive = false; |
| 102 } | 283 } |
| 103 | 284 |
| 104 /////////////////////////////////////////////////////////////////////////////// | 285 void GrVkPrimaryCommandBuffer::beginRenderPass(const GrVkGpu* gpu, |
| 105 | |
| 106 void GrVkCommandBuffer::beginRenderPass(const GrVkGpu* gpu, | |
| 107 const GrVkRenderPass* renderPass, | 286 const GrVkRenderPass* renderPass, |
| 108 const GrVkRenderTarget& target) { | 287 const GrVkRenderTarget& target) { |
| 109 SkASSERT(fIsActive); | 288 SkASSERT(fIsActive); |
| 110 SkASSERT(!fActiveRenderPass); | 289 SkASSERT(!fActiveRenderPass); |
| 111 VkRenderPassBeginInfo beginInfo; | 290 VkRenderPassBeginInfo beginInfo; |
| 112 VkSubpassContents contents; | 291 VkSubpassContents contents; |
| 113 renderPass->getBeginInfo(target, &beginInfo, &contents); | 292 renderPass->getBeginInfo(target, &beginInfo, &contents); |
| 114 GR_VK_CALL(gpu->vkInterface(), CmdBeginRenderPass(fCmdBuffer, &beginInfo, co
ntents)); | 293 GR_VK_CALL(gpu->vkInterface(), CmdBeginRenderPass(fCmdBuffer, &beginInfo, co
ntents)); |
| 115 fActiveRenderPass = renderPass; | 294 fActiveRenderPass = renderPass; |
| 116 this->addResource(renderPass); | 295 this->addResource(renderPass); |
| 117 target.addResources(*this); | 296 target.addResources(*this); |
| 118 } | 297 } |
| 119 | 298 |
| 120 void GrVkCommandBuffer::endRenderPass(const GrVkGpu* gpu) { | 299 void GrVkPrimaryCommandBuffer::endRenderPass(const GrVkGpu* gpu) { |
| 121 SkASSERT(fIsActive); | 300 SkASSERT(fIsActive); |
| 122 SkASSERT(fActiveRenderPass); | 301 SkASSERT(fActiveRenderPass); |
| 123 GR_VK_CALL(gpu->vkInterface(), CmdEndRenderPass(fCmdBuffer)); | 302 GR_VK_CALL(gpu->vkInterface(), CmdEndRenderPass(fCmdBuffer)); |
| 124 fActiveRenderPass = nullptr; | 303 fActiveRenderPass = nullptr; |
| 125 } | 304 } |
| 126 | 305 |
| 127 void GrVkCommandBuffer::submitToQueue(const GrVkGpu* gpu, VkQueue queue, GrVkGpu
::SyncQueue sync) { | 306 void GrVkPrimaryCommandBuffer::executeCommands(const GrVkGpu* gpu, |
| 307 const GrVkSecondaryCommandBuffer*
buffer) { |
| 308 SkASSERT(fIsActive); |
| 309 SkASSERT(fActiveRenderPass); |
| 310 SkASSERT(fActiveRenderPass->isCompatible(*buffer->fActiveRenderPass)); |
| 311 |
| 312 GR_VK_CALL(gpu->vkInterface(), CmdExecuteCommands(fCmdBuffer, 1, &buffer->fC
mdBuffer)); |
| 313 this->addResource(buffer); |
| 314 } |
| 315 |
| 316 void GrVkPrimaryCommandBuffer::submitToQueue(const GrVkGpu* gpu, |
| 317 VkQueue queue, |
| 318 GrVkGpu::SyncQueue sync) { |
| 128 SkASSERT(!fIsActive); | 319 SkASSERT(!fIsActive); |
| 129 | 320 |
| 130 VkResult err; | 321 VkResult err; |
| 131 VkFenceCreateInfo fenceInfo; | 322 VkFenceCreateInfo fenceInfo; |
| 132 memset(&fenceInfo, 0, sizeof(VkFenceCreateInfo)); | 323 memset(&fenceInfo, 0, sizeof(VkFenceCreateInfo)); |
| 133 fenceInfo.sType = VK_STRUCTURE_TYPE_FENCE_CREATE_INFO; | 324 fenceInfo.sType = VK_STRUCTURE_TYPE_FENCE_CREATE_INFO; |
| 134 err = GR_VK_CALL(gpu->vkInterface(), CreateFence(gpu->device(), &fenceInfo,
nullptr, | 325 err = GR_VK_CALL(gpu->vkInterface(), CreateFence(gpu->device(), &fenceInfo,
nullptr, |
| 135 &fSubmitFence)); | 326 &fSubmitFence)); |
| 136 SkASSERT(!err); | 327 SkASSERT(!err); |
| 137 | 328 |
| (...skipping 18 matching lines...) Expand all Loading... |
| 156 SkFAIL("failing"); | 347 SkFAIL("failing"); |
| 157 } | 348 } |
| 158 SkASSERT(!err); | 349 SkASSERT(!err); |
| 159 | 350 |
| 160 // Destroy the fence | 351 // Destroy the fence |
| 161 GR_VK_CALL(gpu->vkInterface(), DestroyFence(gpu->device(), fSubmitFence,
nullptr)); | 352 GR_VK_CALL(gpu->vkInterface(), DestroyFence(gpu->device(), fSubmitFence,
nullptr)); |
| 162 fSubmitFence = VK_NULL_HANDLE; | 353 fSubmitFence = VK_NULL_HANDLE; |
| 163 } | 354 } |
| 164 } | 355 } |
| 165 | 356 |
| 166 bool GrVkCommandBuffer::finished(const GrVkGpu* gpu) const { | 357 bool GrVkPrimaryCommandBuffer::finished(const GrVkGpu* gpu) const { |
| 167 if (VK_NULL_HANDLE == fSubmitFence) { | 358 if (VK_NULL_HANDLE == fSubmitFence) { |
| 168 return true; | 359 return true; |
| 169 } | 360 } |
| 170 | 361 |
| 171 VkResult err = GR_VK_CALL(gpu->vkInterface(), GetFenceStatus(gpu->device(),
fSubmitFence)); | 362 VkResult err = GR_VK_CALL(gpu->vkInterface(), GetFenceStatus(gpu->device(),
fSubmitFence)); |
| 172 switch (err) { | 363 switch (err) { |
| 173 case VK_SUCCESS: | 364 case VK_SUCCESS: |
| 174 return true; | 365 return true; |
| 175 | 366 |
| 176 case VK_NOT_READY: | 367 case VK_NOT_READY: |
| 177 return false; | 368 return false; |
| 178 | 369 |
| 179 default: | 370 default: |
| 180 SkDebugf("Error getting fence status: %d\n", err); | 371 SkDebugf("Error getting fence status: %d\n", err); |
| 181 SkFAIL("failing"); | 372 SkFAIL("failing"); |
| 182 break; | 373 break; |
| 183 } | 374 } |
| 184 | 375 |
| 185 return false; | 376 return false; |
| 186 } | 377 } |
| 187 | 378 |
| 188 //////////////////////////////////////////////////////////////////////////////// | 379 void GrVkPrimaryCommandBuffer::copyImage(const GrVkGpu* gpu, |
| 189 // CommandBuffer commands | 380 GrVkImage* srcImage, |
| 190 //////////////////////////////////////////////////////////////////////////////// | 381 VkImageLayout srcLayout, |
| 191 | 382 GrVkImage* dstImage, |
| 192 void GrVkCommandBuffer::pipelineBarrier(const GrVkGpu* gpu, | 383 VkImageLayout dstLayout, |
| 193 VkPipelineStageFlags srcStageMask, | 384 uint32_t copyRegionCount, |
| 194 VkPipelineStageFlags dstStageMask, | 385 const VkImageCopy* copyRegions) { |
| 195 bool byRegion, | |
| 196 BarrierType barrierType, | |
| 197 void* barrier) const { | |
| 198 SkASSERT(fIsActive); | |
| 199 // For images we can have barriers inside of render passes but they require
us to add more | |
| 200 // support in subpasses which need self dependencies to have barriers inside
them. Also, we can | |
| 201 // never have buffer barriers inside of a render pass. For now we will just
assert that we are | |
| 202 // not in a render pass. | |
| 203 SkASSERT(!fActiveRenderPass); | |
| 204 VkDependencyFlags dependencyFlags = byRegion ? VK_DEPENDENCY_BY_REGION_BIT :
0; | |
| 205 | |
| 206 switch (barrierType) { | |
| 207 case kMemory_BarrierType: { | |
| 208 const VkMemoryBarrier* barrierPtr = reinterpret_cast<VkMemoryBarrier
*>(barrier); | |
| 209 GR_VK_CALL(gpu->vkInterface(), CmdPipelineBarrier(fCmdBuffer, srcSta
geMask, | |
| 210 dstStageMask, depe
ndencyFlags, | |
| 211 1, barrierPtr, | |
| 212 0, nullptr, | |
| 213 0, nullptr)); | |
| 214 break; | |
| 215 } | |
| 216 | |
| 217 case kBufferMemory_BarrierType: { | |
| 218 const VkBufferMemoryBarrier* barrierPtr = | |
| 219 reinterpret_cast<VkBufferMemory
Barrier*>(barrier); | |
| 220 GR_VK_CALL(gpu->vkInterface(), CmdPipelineBarrier(fCmdBuffer, srcSta
geMask, | |
| 221 dstStageMask, depe
ndencyFlags, | |
| 222 0, nullptr, | |
| 223 1, barrierPtr, | |
| 224 0, nullptr)); | |
| 225 break; | |
| 226 } | |
| 227 | |
| 228 case kImageMemory_BarrierType: { | |
| 229 const VkImageMemoryBarrier* barrierPtr = | |
| 230 reinterpret_cast<VkImageMemory
Barrier*>(barrier); | |
| 231 GR_VK_CALL(gpu->vkInterface(), CmdPipelineBarrier(fCmdBuffer, srcSta
geMask, | |
| 232 dstStageMask, depe
ndencyFlags, | |
| 233 0, nullptr, | |
| 234 0, nullptr, | |
| 235 1, barrierPtr)); | |
| 236 break; | |
| 237 } | |
| 238 } | |
| 239 | |
| 240 } | |
| 241 | |
| 242 void GrVkCommandBuffer::copyImage(const GrVkGpu* gpu, | |
| 243 GrVkImage* srcImage, | |
| 244 VkImageLayout srcLayout, | |
| 245 GrVkImage* dstImage, | |
| 246 VkImageLayout dstLayout, | |
| 247 uint32_t copyRegionCount, | |
| 248 const VkImageCopy* copyRegions) { | |
| 249 SkASSERT(fIsActive); | 386 SkASSERT(fIsActive); |
| 250 SkASSERT(!fActiveRenderPass); | 387 SkASSERT(!fActiveRenderPass); |
| 251 this->addResource(srcImage->resource()); | 388 this->addResource(srcImage->resource()); |
| 252 this->addResource(dstImage->resource()); | 389 this->addResource(dstImage->resource()); |
| 253 GR_VK_CALL(gpu->vkInterface(), CmdCopyImage(fCmdBuffer, | 390 GR_VK_CALL(gpu->vkInterface(), CmdCopyImage(fCmdBuffer, |
| 254 srcImage->image(), | 391 srcImage->image(), |
| 255 srcLayout, | 392 srcLayout, |
| 256 dstImage->image(), | 393 dstImage->image(), |
| 257 dstLayout, | 394 dstLayout, |
| 258 copyRegionCount, | 395 copyRegionCount, |
| 259 copyRegions)); | 396 copyRegions)); |
| 260 } | 397 } |
| 261 | 398 |
| 262 void GrVkCommandBuffer::blitImage(const GrVkGpu* gpu, | 399 void GrVkPrimaryCommandBuffer::blitImage(const GrVkGpu* gpu, |
| 263 const GrVkResource* srcResource, | 400 const GrVkResource* srcResource, |
| 264 VkImage srcImage, | 401 VkImage srcImage, |
| 265 VkImageLayout srcLayout, | 402 VkImageLayout srcLayout, |
| 266 const GrVkResource* dstResource, | 403 const GrVkResource* dstResource, |
| 267 VkImage dstImage, | 404 VkImage dstImage, |
| 268 VkImageLayout dstLayout, | 405 VkImageLayout dstLayout, |
| 269 uint32_t blitRegionCount, | 406 uint32_t blitRegionCount, |
| 270 const VkImageBlit* blitRegions, | 407 const VkImageBlit* blitRegions, |
| 271 VkFilter filter) { | 408 VkFilter filter) { |
| 272 SkASSERT(fIsActive); | 409 SkASSERT(fIsActive); |
| 273 SkASSERT(!fActiveRenderPass); | 410 SkASSERT(!fActiveRenderPass); |
| 274 this->addResource(srcResource); | 411 this->addResource(srcResource); |
| 275 this->addResource(dstResource); | 412 this->addResource(dstResource); |
| 276 GR_VK_CALL(gpu->vkInterface(), CmdBlitImage(fCmdBuffer, | 413 GR_VK_CALL(gpu->vkInterface(), CmdBlitImage(fCmdBuffer, |
| 277 srcImage, | 414 srcImage, |
| 278 srcLayout, | 415 srcLayout, |
| 279 dstImage, | 416 dstImage, |
| 280 dstLayout, | 417 dstLayout, |
| 281 blitRegionCount, | 418 blitRegionCount, |
| 282 blitRegions, | 419 blitRegions, |
| 283 filter)); | 420 filter)); |
| 284 } | 421 } |
| 285 | 422 |
| 286 void GrVkCommandBuffer::copyImageToBuffer(const GrVkGpu* gpu, | 423 void GrVkPrimaryCommandBuffer::copyImageToBuffer(const GrVkGpu* gpu, |
| 287 GrVkImage* srcImage, | 424 GrVkImage* srcImage, |
| 288 VkImageLayout srcLayout, | 425 VkImageLayout srcLayout, |
| 289 GrVkTransferBuffer* dstBuffer, | 426 GrVkTransferBuffer* dstBuffer, |
| 290 uint32_t copyRegionCount, | 427 uint32_t copyRegionCount, |
| 291 const VkBufferImageCopy* copyRegions)
{ | 428 const VkBufferImageCopy* copyRe
gions) { |
| 292 SkASSERT(fIsActive); | 429 SkASSERT(fIsActive); |
| 293 SkASSERT(!fActiveRenderPass); | 430 SkASSERT(!fActiveRenderPass); |
| 294 this->addResource(srcImage->resource()); | 431 this->addResource(srcImage->resource()); |
| 295 this->addResource(dstBuffer->resource()); | 432 this->addResource(dstBuffer->resource()); |
| 296 GR_VK_CALL(gpu->vkInterface(), CmdCopyImageToBuffer(fCmdBuffer, | 433 GR_VK_CALL(gpu->vkInterface(), CmdCopyImageToBuffer(fCmdBuffer, |
| 297 srcImage->image(), | 434 srcImage->image(), |
| 298 srcLayout, | 435 srcLayout, |
| 299 dstBuffer->buffer(), | 436 dstBuffer->buffer(), |
| 300 copyRegionCount, | 437 copyRegionCount, |
| 301 copyRegions)); | 438 copyRegions)); |
| 302 } | 439 } |
| 303 | 440 |
| 304 void GrVkCommandBuffer::copyBufferToImage(const GrVkGpu* gpu, | 441 void GrVkPrimaryCommandBuffer::copyBufferToImage(const GrVkGpu* gpu, |
| 305 GrVkTransferBuffer* srcBuffer, | 442 GrVkTransferBuffer* srcBuffer, |
| 306 GrVkImage* dstImage, | 443 GrVkImage* dstImage, |
| 307 VkImageLayout dstLayout, | 444 VkImageLayout dstLayout, |
| 308 uint32_t copyRegionCount, | 445 uint32_t copyRegionCount, |
| 309 const VkBufferImageCopy* copyRegions)
{ | 446 const VkBufferImageCopy* copyRe
gions) { |
| 310 SkASSERT(fIsActive); | 447 SkASSERT(fIsActive); |
| 311 SkASSERT(!fActiveRenderPass); | 448 SkASSERT(!fActiveRenderPass); |
| 312 this->addResource(srcBuffer->resource()); | 449 this->addResource(srcBuffer->resource()); |
| 313 this->addResource(dstImage->resource()); | 450 this->addResource(dstImage->resource()); |
| 314 GR_VK_CALL(gpu->vkInterface(), CmdCopyBufferToImage(fCmdBuffer, | 451 GR_VK_CALL(gpu->vkInterface(), CmdCopyBufferToImage(fCmdBuffer, |
| 315 srcBuffer->buffer(), | 452 srcBuffer->buffer(), |
| 316 dstImage->image(), | 453 dstImage->image(), |
| 317 dstLayout, | 454 dstLayout, |
| 318 copyRegionCount, | 455 copyRegionCount, |
| 319 copyRegions)); | 456 copyRegions)); |
| 320 } | 457 } |
| 321 | 458 |
| 322 void GrVkCommandBuffer::clearColorImage(const GrVkGpu* gpu, | 459 void GrVkPrimaryCommandBuffer::clearColorImage(const GrVkGpu* gpu, |
| 323 GrVkImage* image, | 460 GrVkImage* image, |
| 324 const VkClearColorValue* color, | 461 const VkClearColorValue* color, |
| 325 uint32_t subRangeCount, | 462 uint32_t subRangeCount, |
| 326 const VkImageSubresourceRange* subRanges
) { | 463 const VkImageSubresourceRange* su
bRanges) { |
| 327 SkASSERT(fIsActive); | 464 SkASSERT(fIsActive); |
| 328 SkASSERT(!fActiveRenderPass); | 465 SkASSERT(!fActiveRenderPass); |
| 329 this->addResource(image->resource()); | 466 this->addResource(image->resource()); |
| 330 GR_VK_CALL(gpu->vkInterface(), CmdClearColorImage(fCmdBuffer, | 467 GR_VK_CALL(gpu->vkInterface(), CmdClearColorImage(fCmdBuffer, |
| 331 image->image(), | 468 image->image(), |
| 332 image->currentLayout(), | 469 image->currentLayout(), |
| 333 color, | 470 color, |
| 334 subRangeCount, | 471 subRangeCount, |
| 335 subRanges)); | 472 subRanges)); |
| 336 } | 473 } |
| 337 | 474 |
| 338 void GrVkCommandBuffer::clearDepthStencilImage(const GrVkGpu* gpu, | 475 void GrVkPrimaryCommandBuffer::clearDepthStencilImage(const GrVkGpu* gpu, |
| 339 GrVkImage* image, | 476 GrVkImage* image, |
| 340 const VkClearDepthStencilValue* c
olor, | 477 const VkClearDepthStencilV
alue* color, |
| 341 uint32_t subRangeCount, | 478 uint32_t subRangeCount, |
| 342 const VkImageSubresourceRange* su
bRanges) { | 479 const VkImageSubresourceRa
nge* subRanges) { |
| 343 SkASSERT(fIsActive); | 480 SkASSERT(fIsActive); |
| 344 SkASSERT(!fActiveRenderPass); | 481 SkASSERT(!fActiveRenderPass); |
| 345 this->addResource(image->resource()); | 482 this->addResource(image->resource()); |
| 346 GR_VK_CALL(gpu->vkInterface(), CmdClearDepthStencilImage(fCmdBuffer, | 483 GR_VK_CALL(gpu->vkInterface(), CmdClearDepthStencilImage(fCmdBuffer, |
| 347 image->image(), | 484 image->image(), |
| 348 image->currentLayou
t(), | 485 image->currentLayou
t(), |
| 349 color, | 486 color, |
| 350 subRangeCount, | 487 subRangeCount, |
| 351 subRanges)); | 488 subRanges)); |
| 352 } | 489 } |
| 353 | 490 |
| 354 void GrVkCommandBuffer::clearAttachments(const GrVkGpu* gpu, | 491 /////////////////////////////////////////////////////////////////////////////// |
| 355 int numAttachments, | 492 // SecondaryCommandBuffer |
| 356 const VkClearAttachment* attachments, | 493 //////////////////////////////////////////////////////////////////////////////// |
| 357 int numRects, | 494 |
| 358 const VkClearRect* clearRects) const { | 495 GrVkSecondaryCommandBuffer* GrVkSecondaryCommandBuffer::Create( |
| 359 SkASSERT(fIsActive); | 496 const GrVkGpu* gpu, |
| 360 SkASSERT(fActiveRenderPass); | 497 VkCommandPool cmdPool, |
| 361 SkASSERT(numAttachments > 0); | 498 const GrVkRenderPass* com
patibleRenderPass) { |
| 362 SkASSERT(numRects > 0); | 499 const VkCommandBufferAllocateInfo cmdInfo = { |
| 363 #ifdef SK_DEBUG | 500 VK_STRUCTURE_TYPE_COMMAND_BUFFER_ALLOCATE_INFO, // sType |
| 364 for (int i = 0; i < numAttachments; ++i) { | 501 NULL, // pNext |
| 365 if (attachments[i].aspectMask == VK_IMAGE_ASPECT_COLOR_BIT) { | 502 cmdPool, // commandPool |
| 366 uint32_t testIndex; | 503 VK_COMMAND_BUFFER_LEVEL_SECONDARY, // level |
| 367 SkAssertResult(fActiveRenderPass->colorAttachmentIndex(&testIndex)); | 504 1 // bufferCount |
| 368 SkASSERT(testIndex == attachments[i].colorAttachment); | 505 }; |
| 369 } | 506 |
| 507 VkCommandBuffer cmdBuffer; |
| 508 VkResult err = GR_VK_CALL(gpu->vkInterface(), AllocateCommandBuffers(gpu->de
vice(), |
| 509 &cmdInf
o, |
| 510 &cmdBuf
fer)); |
| 511 if (err) { |
| 512 return nullptr; |
| 370 } | 513 } |
| 371 #endif | 514 return new GrVkSecondaryCommandBuffer(cmdBuffer, compatibleRenderPass); |
| 372 GR_VK_CALL(gpu->vkInterface(), CmdClearAttachments(fCmdBuffer, | |
| 373 numAttachments, | |
| 374 attachments, | |
| 375 numRects, | |
| 376 clearRects)); | |
| 377 } | 515 } |
| 378 | 516 |
| 379 void GrVkCommandBuffer::bindDescriptorSets(const GrVkGpu* gpu, | 517 |
| 380 GrVkPipelineState* pipelineState, | 518 void GrVkSecondaryCommandBuffer::begin(const GrVkGpu* gpu, const GrVkFramebuffer
* framebuffer) { |
| 381 VkPipelineLayout layout, | 519 SkASSERT(!fIsActive); |
| 382 uint32_t firstSet, | 520 SkASSERT(fActiveRenderPass); |
| 383 uint32_t setCount, | 521 |
| 384 const VkDescriptorSet* descriptorSets
, | 522 VkCommandBufferInheritanceInfo inheritanceInfo; |
| 385 uint32_t dynamicOffsetCount, | 523 memset(&inheritanceInfo, 0, sizeof(VkCommandBufferInheritanceInfo)); |
| 386 const uint32_t* dynamicOffsets) { | 524 inheritanceInfo.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_INHERITANCE_INFO; |
| 387 SkASSERT(fIsActive); | 525 inheritanceInfo.pNext = nullptr; |
| 388 GR_VK_CALL(gpu->vkInterface(), CmdBindDescriptorSets(fCmdBuffer, | 526 inheritanceInfo.renderPass = fActiveRenderPass->vkRenderPass(); |
| 389 VK_PIPELINE_BIND_POINT_
GRAPHICS, | 527 inheritanceInfo.subpass = 0; // Currently only using 1 subpass for each rend
er pass |
| 390 layout, | 528 inheritanceInfo.framebuffer = framebuffer ? framebuffer->framebuffer() : VK_
NULL_HANDLE; |
| 391 firstSet, | 529 inheritanceInfo.occlusionQueryEnable = false; |
| 392 setCount, | 530 inheritanceInfo.queryFlags = 0; |
| 393 descriptorSets, | 531 inheritanceInfo.pipelineStatistics = 0; |
| 394 dynamicOffsetCount, | 532 |
| 395 dynamicOffsets)); | 533 VkCommandBufferBeginInfo cmdBufferBeginInfo; |
| 396 pipelineState->addUniformResources(*this); | 534 memset(&cmdBufferBeginInfo, 0, sizeof(VkCommandBufferBeginInfo)); |
| 535 cmdBufferBeginInfo.sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_BEGIN_INFO; |
| 536 cmdBufferBeginInfo.pNext = nullptr; |
| 537 cmdBufferBeginInfo.flags = VK_COMMAND_BUFFER_USAGE_RENDER_PASS_CONTINUE_BIT
| |
| 538 VK_COMMAND_BUFFER_USAGE_ONE_TIME_SUBMIT_BIT; |
| 539 cmdBufferBeginInfo.pInheritanceInfo = &inheritanceInfo; |
| 540 |
| 541 GR_VK_CALL_ERRCHECK(gpu->vkInterface(), BeginCommandBuffer(fCmdBuffer, |
| 542 &cmdBufferBeginIn
fo)); |
| 543 fIsActive = true; |
| 397 } | 544 } |
| 398 | 545 |
| 399 void GrVkCommandBuffer::bindPipeline(const GrVkGpu* gpu, const GrVkPipeline* pip
eline) { | 546 void GrVkSecondaryCommandBuffer::end(const GrVkGpu* gpu) { |
| 400 SkASSERT(fIsActive); | 547 SkASSERT(fIsActive); |
| 401 GR_VK_CALL(gpu->vkInterface(), CmdBindPipeline(fCmdBuffer, | 548 GR_VK_CALL_ERRCHECK(gpu->vkInterface(), EndCommandBuffer(fCmdBuffer)); |
| 402 VK_PIPELINE_BIND_POINT_GRAPHI
CS, | 549 this->invalidateState(); |
| 403 pipeline->pipeline())); | 550 fIsActive = false; |
| 404 addResource(pipeline); | |
| 405 } | 551 } |
| 406 | 552 |
| 407 void GrVkCommandBuffer::drawIndexed(const GrVkGpu* gpu, | |
| 408 uint32_t indexCount, | |
| 409 uint32_t instanceCount, | |
| 410 uint32_t firstIndex, | |
| 411 int32_t vertexOffset, | |
| 412 uint32_t firstInstance) const { | |
| 413 SkASSERT(fIsActive); | |
| 414 SkASSERT(fActiveRenderPass); | |
| 415 GR_VK_CALL(gpu->vkInterface(), CmdDrawIndexed(fCmdBuffer, | |
| 416 indexCount, | |
| 417 instanceCount, | |
| 418 firstIndex, | |
| 419 vertexOffset, | |
| 420 firstInstance)); | |
| 421 } | |
| 422 | |
| 423 void GrVkCommandBuffer::draw(const GrVkGpu* gpu, | |
| 424 uint32_t vertexCount, | |
| 425 uint32_t instanceCount, | |
| 426 uint32_t firstVertex, | |
| 427 uint32_t firstInstance) const { | |
| 428 SkASSERT(fIsActive); | |
| 429 SkASSERT(fActiveRenderPass); | |
| 430 GR_VK_CALL(gpu->vkInterface(), CmdDraw(fCmdBuffer, | |
| 431 vertexCount, | |
| 432 instanceCount, | |
| 433 firstVertex, | |
| 434 firstInstance)); | |
| 435 } | |
| 436 | |
| 437 void GrVkCommandBuffer::setViewport(const GrVkGpu* gpu, | |
| 438 uint32_t firstViewport, | |
| 439 uint32_t viewportCount, | |
| 440 const VkViewport* viewports) { | |
| 441 SkASSERT(fIsActive); | |
| 442 SkASSERT(1 == viewportCount); | |
| 443 if (memcmp(viewports, &fCachedViewport, sizeof(VkViewport))) { | |
| 444 GR_VK_CALL(gpu->vkInterface(), CmdSetViewport(fCmdBuffer, | |
| 445 firstViewport, | |
| 446 viewportCount, | |
| 447 viewports)); | |
| 448 fCachedViewport = viewports[0]; | |
| 449 } | |
| 450 } | |
| 451 | |
| 452 void GrVkCommandBuffer::setScissor(const GrVkGpu* gpu, | |
| 453 uint32_t firstScissor, | |
| 454 uint32_t scissorCount, | |
| 455 const VkRect2D* scissors) { | |
| 456 SkASSERT(fIsActive); | |
| 457 SkASSERT(1 == scissorCount); | |
| 458 if (memcmp(scissors, &fCachedScissor, sizeof(VkRect2D))) { | |
| 459 GR_VK_CALL(gpu->vkInterface(), CmdSetScissor(fCmdBuffer, | |
| 460 firstScissor, | |
| 461 scissorCount, | |
| 462 scissors)); | |
| 463 fCachedScissor = scissors[0]; | |
| 464 } | |
| 465 } | |
| 466 | |
| 467 void GrVkCommandBuffer::setBlendConstants(const GrVkGpu* gpu, | |
| 468 const float blendConstants[4]) { | |
| 469 SkASSERT(fIsActive); | |
| 470 if (memcmp(blendConstants, fCachedBlendConstant, 4 * sizeof(float))) { | |
| 471 GR_VK_CALL(gpu->vkInterface(), CmdSetBlendConstants(fCmdBuffer, blendCon
stants)); | |
| 472 memcpy(fCachedBlendConstant, blendConstants, 4 * sizeof(float)); | |
| 473 } | |
| 474 } | |
| OLD | NEW |