OLD | NEW |
1 | 1 |
2 /* | 2 /* |
3 * Copyright 2010 Google Inc. | 3 * Copyright 2010 Google Inc. |
4 * | 4 * |
5 * Use of this source code is governed by a BSD-style license that can be | 5 * Use of this source code is governed by a BSD-style license that can be |
6 * found in the LICENSE file. | 6 * found in the LICENSE file. |
7 */ | 7 */ |
8 | 8 |
9 | 9 |
10 #include "GrBufferAllocPool.h" | 10 #include "GrBufferAllocPool.h" |
(...skipping 39 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
50 fBytesInUse = 0; | 50 fBytesInUse = 0; |
51 | 51 |
52 fPreallocBuffersInUse = 0; | 52 fPreallocBuffersInUse = 0; |
53 fPreallocBufferStartIdx = 0; | 53 fPreallocBufferStartIdx = 0; |
54 for (int i = 0; i < preallocBufferCnt; ++i) { | 54 for (int i = 0; i < preallocBufferCnt; ++i) { |
55 GrGeometryBuffer* buffer = this->createBuffer(fMinBlockSize); | 55 GrGeometryBuffer* buffer = this->createBuffer(fMinBlockSize); |
56 if (buffer) { | 56 if (buffer) { |
57 *fPreallocBuffers.append() = buffer; | 57 *fPreallocBuffers.append() = buffer; |
58 } | 58 } |
59 } | 59 } |
| 60 fGeometryBufferMapThreshold = gpu->caps()->geometryBufferMapThreshold(); |
60 } | 61 } |
61 | 62 |
62 GrBufferAllocPool::~GrBufferAllocPool() { | 63 GrBufferAllocPool::~GrBufferAllocPool() { |
63 VALIDATE(); | 64 VALIDATE(); |
64 if (fBlocks.count()) { | 65 if (fBlocks.count()) { |
65 GrGeometryBuffer* buffer = fBlocks.back().fBuffer; | 66 GrGeometryBuffer* buffer = fBlocks.back().fBuffer; |
66 if (buffer->isMapped()) { | 67 if (buffer->isMapped()) { |
67 UNMAP_BUFFER(fBlocks.back()); | 68 UNMAP_BUFFER(fBlocks.back()); |
68 } | 69 } |
69 } | 70 } |
(...skipping 211 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
281 fBufferPtr = NULL; | 282 fBufferPtr = NULL; |
282 } | 283 } |
283 | 284 |
284 SkASSERT(NULL == fBufferPtr); | 285 SkASSERT(NULL == fBufferPtr); |
285 | 286 |
286 // If the buffer is CPU-backed we map it because it is free to do so and sav
es a copy. | 287 // If the buffer is CPU-backed we map it because it is free to do so and sav
es a copy. |
287 // Otherwise when buffer mapping is supported we map if the buffer size is g
reater than the | 288 // Otherwise when buffer mapping is supported we map if the buffer size is g
reater than the |
288 // threshold. | 289 // threshold. |
289 bool attemptMap = block.fBuffer->isCPUBacked(); | 290 bool attemptMap = block.fBuffer->isCPUBacked(); |
290 if (!attemptMap && GrCaps::kNone_MapFlags != fGpu->caps()->mapBufferFlags())
{ | 291 if (!attemptMap && GrCaps::kNone_MapFlags != fGpu->caps()->mapBufferFlags())
{ |
291 attemptMap = size > GR_GEOM_BUFFER_MAP_THRESHOLD; | 292 attemptMap = size > fGeometryBufferMapThreshold; |
292 } | 293 } |
293 | 294 |
294 if (attemptMap) { | 295 if (attemptMap) { |
295 fBufferPtr = block.fBuffer->map(); | 296 fBufferPtr = block.fBuffer->map(); |
296 } | 297 } |
297 | 298 |
298 if (NULL == fBufferPtr) { | 299 if (NULL == fBufferPtr) { |
299 fBufferPtr = fCpuData.reset(size); | 300 fBufferPtr = fCpuData.reset(size); |
300 } | 301 } |
301 | 302 |
(...skipping 23 matching lines...) Expand all Loading... |
325 | 326 |
326 void GrBufferAllocPool::flushCpuData(const BufferBlock& block, size_t flushSize)
{ | 327 void GrBufferAllocPool::flushCpuData(const BufferBlock& block, size_t flushSize)
{ |
327 GrGeometryBuffer* buffer = block.fBuffer; | 328 GrGeometryBuffer* buffer = block.fBuffer; |
328 SkASSERT(buffer); | 329 SkASSERT(buffer); |
329 SkASSERT(!buffer->isMapped()); | 330 SkASSERT(!buffer->isMapped()); |
330 SkASSERT(fCpuData.get() == fBufferPtr); | 331 SkASSERT(fCpuData.get() == fBufferPtr); |
331 SkASSERT(flushSize <= buffer->gpuMemorySize()); | 332 SkASSERT(flushSize <= buffer->gpuMemorySize()); |
332 VALIDATE(true); | 333 VALIDATE(true); |
333 | 334 |
334 if (GrCaps::kNone_MapFlags != fGpu->caps()->mapBufferFlags() && | 335 if (GrCaps::kNone_MapFlags != fGpu->caps()->mapBufferFlags() && |
335 flushSize > GR_GEOM_BUFFER_MAP_THRESHOLD) { | 336 flushSize > fGeometryBufferMapThreshold) { |
336 void* data = buffer->map(); | 337 void* data = buffer->map(); |
337 if (data) { | 338 if (data) { |
338 memcpy(data, fBufferPtr, flushSize); | 339 memcpy(data, fBufferPtr, flushSize); |
339 UNMAP_BUFFER(block); | 340 UNMAP_BUFFER(block); |
340 return; | 341 return; |
341 } | 342 } |
342 } | 343 } |
343 buffer->updateData(fBufferPtr, flushSize); | 344 buffer->updateData(fBufferPtr, flushSize); |
344 VALIDATE(true); | 345 VALIDATE(true); |
345 } | 346 } |
(...skipping 66 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
412 &geomBuffer, | 413 &geomBuffer, |
413 &offset); | 414 &offset); |
414 | 415 |
415 *buffer = (const GrIndexBuffer*) geomBuffer; | 416 *buffer = (const GrIndexBuffer*) geomBuffer; |
416 SkASSERT(0 == offset % sizeof(uint16_t)); | 417 SkASSERT(0 == offset % sizeof(uint16_t)); |
417 *startIndex = static_cast<int>(offset / sizeof(uint16_t)); | 418 *startIndex = static_cast<int>(offset / sizeof(uint16_t)); |
418 return ptr; | 419 return ptr; |
419 } | 420 } |
420 | 421 |
421 | 422 |
OLD | NEW |