Index: src/gpu/GrBuffer.h |
diff --git a/src/gpu/GrBuffer.h b/src/gpu/GrBuffer.h |
new file mode 100644 |
index 0000000000000000000000000000000000000000..4fadba6aa72f7bc3d4252a38b63259053a80a472 |
--- /dev/null |
+++ b/src/gpu/GrBuffer.h |
@@ -0,0 +1,145 @@ |
+/* |
+ * Copyright 2016 Google Inc. |
+ * |
+ * Use of this source code is governed by a BSD-style license that can be |
+ * found in the LICENSE file. |
+ */ |
+ |
+#ifndef GrBuffer_DEFINED |
+#define GrBuffer_DEFINED |
+ |
+#include "GrGpuResource.h" |
+ |
+class GrGpu; |
+ |
+class GrBuffer : public GrGpuResource { |
+public: |
+ /** |
+ * Computes a scratch key for a buffer with a "dynamic" access pattern. (Buffers with "static" |
+ * and "stream" access patterns are disqualified by nature from being cached and reused.) |
+ */ |
+ static void ComputeScratchKeyForDynamicBuffer(GrBufferType type, size_t size, |
+ GrScratchKey* key) { |
+ static const GrScratchKey::ResourceType kType = GrScratchKey::GenerateResourceType(); |
+ GrScratchKey::Builder builder(key, kType, 1 + (sizeof(size_t) + 3) / 4); |
+ // TODO: There's not always reason to cache a buffer by type. In some (all?) APIs it's just |
+ // a chunk of memory we can use/reuse for any type of data. We really only need to |
+ // differentiate between the "read" types (e.g. kGpuToCpu_BufferType) and "draw" types. |
+ builder[0] = type; |
+ builder[1] = (uint32_t)size; |
+ if (sizeof(size_t) > 4) { |
+ builder[2] = (uint32_t)((uint64_t)size >> 32); |
+ } |
+ } |
+ |
+ GrBufferType type() const { return fType; } |
+ |
+ GrAccessPattern accessPattern() const { return fAccessPattern; } |
+ |
+ /** |
+ * Returns true if the buffer is a wrapper around a CPU array. If true it |
+ * indicates that map will always succeed and will be free. |
+ */ |
+ bool isCPUBacked() const { return fCPUBacked; } |
+ |
+ /** |
+ * Maps the buffer to be written by the CPU. |
+ * |
+ * The previous content of the buffer is invalidated. It is an error |
+ * to draw from the buffer while it is mapped. It may fail if the backend |
+ * doesn't support mapping the buffer. If the buffer is CPU backed then |
+ * it will always succeed and is a free operation. Once a buffer is mapped, |
+ * subsequent calls to map() are ignored. |
+ * |
+ * Note that buffer mapping does not go through GrContext and therefore is |
+ * not serialized with other operations. |
+ * |
+ * @return a pointer to the data or nullptr if the map fails. |
+ */ |
+ void* map() { |
+ if (!fMapPtr) { |
+ this->onMap(); |
+ } |
+ return fMapPtr; |
+ } |
+ |
+ /** |
+ * Unmaps the buffer. |
+ * |
+ * The pointer returned by the previous map call will no longer be valid. |
+ */ |
+ void unmap() { |
+ SkASSERT(fMapPtr); |
+ this->onUnmap(); |
+ fMapPtr = nullptr; |
+ } |
+ |
+ /** |
+ * Returns the same ptr that map() returned at time of map or nullptr if the |
+ * is not mapped. |
+ * |
+ * @return ptr to mapped buffer data or nullptr if buffer is not mapped. |
+ */ |
+ void* mapPtr() const { return fMapPtr; } |
+ |
+ /** |
+ Queries whether the buffer has been mapped. |
+ |
+ @return true if the buffer is mapped, false otherwise. |
+ */ |
+ bool isMapped() const { return SkToBool(fMapPtr); } |
+ |
+ /** |
+ * Updates the buffer data. |
+ * |
+ * The size of the buffer will be preserved. The src data will be |
+ * placed at the beginning of the buffer and any remaining contents will |
+ * be undefined. srcSizeInBytes must be <= to the buffer size. |
+ * |
+ * The buffer must not be mapped. |
+ * |
+ * Note that buffer updates do not go through GrContext and therefore are |
+ * not serialized with other operations. |
+ * |
+ * @return returns true if the update succeeds, false otherwise. |
+ */ |
+ bool updateData(const void* src, size_t srcSizeInBytes) { |
+ SkASSERT(!this->isMapped()); |
+ SkASSERT(srcSizeInBytes <= fGpuMemorySize); |
+ return this->onUpdateData(src, srcSizeInBytes); |
+ } |
+ |
+protected: |
+ GrBuffer(GrGpu* gpu, GrBufferType type, size_t gpuMemorySize, GrAccessPattern accessPattern, |
+ bool cpuBacked) |
+ : INHERITED(gpu, kCached_LifeCycle), |
+ fMapPtr(nullptr), |
+ fType(type), |
+ fGpuMemorySize(gpuMemorySize), // TODO: Zero for cpu backed buffers? |
+ fAccessPattern(accessPattern), |
+ fCPUBacked(cpuBacked) { |
+ if (!fCPUBacked && SkIsPow2(fGpuMemorySize) && kDynamic_GrAccessPattern == fAccessPattern) { |
+ GrScratchKey key; |
+ ComputeScratchKeyForDynamicBuffer(fType, fGpuMemorySize, &key); |
+ this->setScratchKey(key); |
+ } |
+ } |
+ |
+ void* fMapPtr; |
+ |
+private: |
+ virtual size_t onGpuMemorySize() const { return fGpuMemorySize; } |
+ |
+ virtual void onMap() = 0; |
+ virtual void onUnmap() = 0; |
+ virtual bool onUpdateData(const void* src, size_t srcSizeInBytes) = 0; |
+ |
+ GrBufferType fType; |
+ size_t fGpuMemorySize; |
+ GrAccessPattern fAccessPattern; |
+ bool fCPUBacked; |
+ |
+ typedef GrGpuResource INHERITED; |
+}; |
+ |
+#endif |