Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(538)

Side by Side Diff: src/gpu/GrBatch.h

Issue 1277233002: Make folder for batches (Closed) Base URL: https://skia.googlesource.com/skia.git@master
Patch Set: rebase Created 5 years, 4 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View unified diff | Download patch
« no previous file with comments | « src/gpu/GrAtlasTextContext.cpp ('k') | src/gpu/GrBatch.cpp » ('j') | no next file with comments »
Toggle Intra-line Diffs ('i') | Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
OLDNEW
(Empty)
1 /*
2 * Copyright 2015 Google Inc.
3 *
4 * Use of this source code is governed by a BSD-style license that can be
5 * found in the LICENSE file.
6 */
7
8 #ifndef GrBatch_DEFINED
9 #define GrBatch_DEFINED
10
11 #include <new>
12 #include "GrBatchTarget.h"
13 #include "GrGeometryProcessor.h"
14 #include "GrNonAtomicRef.h"
15 #include "GrVertices.h"
16 #include "SkAtomics.h"
17 #include "SkTypes.h"
18
19 class GrGpu;
20 class GrPipeline;
21
22 struct GrInitInvariantOutput;
23
24 /*
25 * GrBatch is the base class for all Ganesh deferred geometry generators. To fa cilitate
26 * reorderable batching, Ganesh does not generate geometry inline with draw call s. Instead, it
27 * captures the arguments to the draw and then generates the geometry on demand. This gives GrBatch
28 * subclasses complete freedom to decide how / what they can batch.
29 *
30 * Batches are created when GrContext processes a draw call. Batches of the same subclass may be
31 * merged using combineIfPossible. When two batches merge, one takes on the unio n of the data
32 * and the other is left empty. The merged batch becomes responsible for drawing the data from both
33 * the original batches.
34 *
35 * If there are any possible optimizations which might require knowing more abou t the full state of
36 * the draw, ie whether or not the GrBatch is allowed to tweak alpha for coverag e, then this
37 * information will be communicated to the GrBatch prior to geometry generation.
38 */
39 #define GR_BATCH_SPEW 0
40 #if GR_BATCH_SPEW
41 #define GrBATCH_INFO(...) SkDebugf(__VA_ARGS__)
42 #define GrBATCH_SPEW(code) code
43 #else
44 #define GrBATCH_SPEW(code)
45 #define GrBATCH_INFO(...)
46 #endif
47
48 class GrBatch : public GrNonAtomicRef {
49 public:
50 GrBatch()
51 : fClassID(kIllegalBatchID)
52 , fNumberOfDraws(0)
53 #if GR_BATCH_SPEW
54 , fUniqueID(GenID(&gCurrBatchUniqueID))
55 #endif
56 { SkDEBUGCODE(fUsed = false;) }
57 virtual ~GrBatch() {}
58
59 virtual const char* name() const = 0;
60 virtual void getInvariantOutputColor(GrInitInvariantOutput* out) const = 0;
61 virtual void getInvariantOutputCoverage(GrInitInvariantOutput* out) const = 0;
62
63 /*
64 * initBatchTracker is a hook for the some additional overrides / optimizati on possibilities
65 * from the GrXferProcessor.
66 */
67 virtual void initBatchTracker(const GrPipelineInfo& init) = 0;
68
69 bool combineIfPossible(GrBatch* that) {
70 if (this->classID() != that->classID()) {
71 return false;
72 }
73
74 return this->onCombineIfPossible(that);
75 }
76
77 virtual bool onCombineIfPossible(GrBatch*) = 0;
78
79 virtual void generateGeometry(GrBatchTarget*) = 0;
80
81 const SkRect& bounds() const { return fBounds; }
82
83 // TODO this goes away when batches are everywhere
84 void setNumberOfDraws(int numberOfDraws) { fNumberOfDraws = numberOfDraws; }
85 int numberOfDraws() const { return fNumberOfDraws; }
86
87 void* operator new(size_t size);
88 void operator delete(void* target);
89
90 void* operator new(size_t size, void* placement) {
91 return ::operator new(size, placement);
92 }
93 void operator delete(void* target, void* placement) {
94 ::operator delete(target, placement);
95 }
96
97 /**
98 * Helper for down-casting to a GrBatch subclass
99 */
100 template <typename T> const T& cast() const { return *static_cast<const T*>( this); }
101 template <typename T> T* cast() { return static_cast<T*>(this); }
102
103 uint32_t classID() const { SkASSERT(kIllegalBatchID != fClassID); return fCl assID; }
104
105 // TODO no GrPrimitiveProcessors yet read fragment position
106 bool willReadFragmentPosition() const { return false; }
107
108 SkDEBUGCODE(bool isUsed() const { return fUsed; })
109
110 const GrPipeline* pipeline() const { return fPipeline; }
111 void setPipeline(const GrPipeline* pipeline) { fPipeline.reset(SkRef(pipelin e)); }
112
113 #if GR_BATCH_SPEW
114 uint32_t uniqueID() const { return fUniqueID; }
115 #endif
116
117 protected:
118 template <typename PROC_SUBCLASS> void initClassID() {
119 static uint32_t kClassID = GenID(&gCurrBatchClassID);
120 fClassID = kClassID;
121 }
122
123 // NOTE, compute some bounds, even if extremely conservative. Do *NOT* setL argest on the bounds
124 // rect because we outset it for dst copy textures
125 void setBounds(const SkRect& newBounds) { fBounds = newBounds; }
126
127 void joinBounds(const SkRect& otherBounds) {
128 return fBounds.joinPossiblyEmptyRect(otherBounds);
129 }
130
131 /** Helper for rendering instances using an instanced index index buffer. Th is class creates the
132 space for the vertices and flushes the draws to the batch target.*/
133 class InstancedHelper {
134 public:
135 InstancedHelper() {}
136 /** Returns the allocated storage for the vertices. The caller should po pulate the before
137 vertices before calling issueDraws(). */
138 void* init(GrBatchTarget* batchTarget, GrPrimitiveType, size_t vertexStr ide,
139 const GrIndexBuffer*, int verticesPerInstance, int indicesPer Instance,
140 int instancesToDraw);
141
142 /** Call after init() to issue draws to the batch target.*/
143 void issueDraw(GrBatchTarget* batchTarget) {
144 SkASSERT(fVertices.instanceCount());
145 batchTarget->draw(fVertices);
146 }
147 private:
148 GrVertices fVertices;
149 };
150
151 static const int kVerticesPerQuad = 4;
152 static const int kIndicesPerQuad = 6;
153
154 /** A specialization of InstanceHelper for quad rendering. */
155 class QuadHelper : private InstancedHelper {
156 public:
157 QuadHelper() : INHERITED() {}
158 /** Finds the cached quad index buffer and reserves vertex space. Return s NULL on failure
159 and on sucess a pointer to the vertex data that the caller should po pulate before
160 calling issueDraws(). */
161 void* init(GrBatchTarget* batchTarget, size_t vertexStride, int quadsToD raw);
162
163 using InstancedHelper::issueDraw;
164
165 private:
166 typedef InstancedHelper INHERITED;
167 };
168
169 uint32_t fClassID;
170 SkRect fBounds;
171
172 private:
173 static uint32_t GenID(int32_t* idCounter) {
174 // fCurrProcessorClassID has been initialized to kIllegalProcessorClassI D. The
175 // atomic inc returns the old value not the incremented value. So we add
176 // 1 to the returned value.
177 uint32_t id = static_cast<uint32_t>(sk_atomic_inc(idCounter)) + 1;
178 if (!id) {
179 SkFAIL("This should never wrap as it should only be called once for each GrBatch "
180 "subclass.");
181 }
182 return id;
183 }
184
185 enum {
186 kIllegalBatchID = 0,
187 };
188 SkAutoTUnref<const GrPipeline> fPipeline;
189 static int32_t gCurrBatchClassID;
190 int fNumberOfDraws;
191 SkDEBUGCODE(bool fUsed;)
192 #if GR_BATCH_SPEW
193 static int32_t gCurrBatchUniqueID;
194 uint32_t fUniqueID;
195 #endif
196
197 typedef SkRefCnt INHERITED;
198 };
199
200 #endif
OLDNEW
« no previous file with comments | « src/gpu/GrAtlasTextContext.cpp ('k') | src/gpu/GrBatch.cpp » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698