| OLD | NEW |
| 1 | 1 |
| 2 /* | 2 /* |
| 3 * Copyright 2011 Google Inc. | 3 * Copyright 2011 Google Inc. |
| 4 * | 4 * |
| 5 * Use of this source code is governed by a BSD-style license that can be | 5 * Use of this source code is governed by a BSD-style license that can be |
| 6 * found in the LICENSE file. | 6 * found in the LICENSE file. |
| 7 */ | 7 */ |
| 8 #ifndef SkPictureFlat_DEFINED | 8 #ifndef SkPictureFlat_DEFINED |
| 9 #define SkPictureFlat_DEFINED | 9 #define SkPictureFlat_DEFINED |
| 10 | 10 |
| 11 //#define SK_DEBUG_SIZE | 11 //#define SK_DEBUG_SIZE |
| 12 | 12 |
| 13 #include "SkChunkAlloc.h" | 13 #include "SkChunkAlloc.h" |
| 14 #include "SkBitmap.h" | 14 #include "SkBitmap.h" |
| 15 #include "SkBitmapHeap.h" | 15 #include "SkBitmapHeap.h" |
| 16 #include "SkOrderedReadBuffer.h" | 16 #include "SkOrderedReadBuffer.h" |
| 17 #include "SkOrderedWriteBuffer.h" | 17 #include "SkOrderedWriteBuffer.h" |
| 18 #include "SkPicture.h" | 18 #include "SkPicture.h" |
| 19 #include "SkPtrRecorder.h" | 19 #include "SkPtrRecorder.h" |
| 20 #include "SkMatrix.h" | 20 #include "SkMatrix.h" |
| 21 #include "SkPaint.h" | 21 #include "SkPaint.h" |
| 22 #include "SkPath.h" | 22 #include "SkPath.h" |
| 23 #include "SkRegion.h" | 23 #include "SkRegion.h" |
| 24 #include "SkTRefArray.h" | 24 #include "SkTRefArray.h" |
| 25 #include "SkTSearch.h" | 25 #include "SkTSearch.h" |
| 26 | 26 |
| 27 #include "SkChecksum.h" |
| 28 #include "SkTDynamicHash.h" |
| 29 |
| 27 enum DrawType { | 30 enum DrawType { |
| 28 UNUSED, | 31 UNUSED, |
| 29 CLIP_PATH, | 32 CLIP_PATH, |
| 30 CLIP_REGION, | 33 CLIP_REGION, |
| 31 CLIP_RECT, | 34 CLIP_RECT, |
| 32 CLIP_RRECT, | 35 CLIP_RRECT, |
| 33 CONCAT, | 36 CONCAT, |
| 34 DRAW_BITMAP, | 37 DRAW_BITMAP, |
| 35 DRAW_BITMAP_MATRIX, | 38 DRAW_BITMAP_MATRIX, |
| 36 DRAW_BITMAP_NINE, | 39 DRAW_BITMAP_NINE, |
| (...skipping 219 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 256 SkRefCntSet* fTypefaceSet; | 259 SkRefCntSet* fTypefaceSet; |
| 257 SkTypefacePlayback* fTypefacePlayback; | 260 SkTypefacePlayback* fTypefacePlayback; |
| 258 SkNamedFactorySet* fFactorySet; | 261 SkNamedFactorySet* fFactorySet; |
| 259 uint32_t fWriteBufferFlags; | 262 uint32_t fWriteBufferFlags; |
| 260 | 263 |
| 261 typedef SkRefCnt INHERITED; | 264 typedef SkRefCnt INHERITED; |
| 262 }; | 265 }; |
| 263 | 266 |
| 264 class SkFlatData { | 267 class SkFlatData { |
| 265 public: | 268 public: |
| 266 /** | 269 // Flatten obj into an SkFlatData with this index. controller owns the SkFl
atData*. |
| 267 * Compare two SkFlatData ptrs, returning -1, 0, 1 to allow them to be | 270 static SkFlatData* Create(SkFlatController* controller, |
| 268 * sorted. | 271 const void* obj, |
| 269 * | 272 int index, |
| 270 * Note: this assumes that a and b have different sentinel values, either | |
| 271 * InCache or AsCandidate, otherwise the loop will go beyond the end of | |
| 272 * the buffers. | |
| 273 * | |
| 274 * dataToCompare() returns 2 fields before the flattened data: | |
| 275 * - checksum | |
| 276 * - size | |
| 277 * This ensures that if we see two blocks of different length, we will | |
| 278 * notice that right away, and not read any further. It also ensures that | |
| 279 * we see the checksum right away, so that most of the time it is enough | |
| 280 * to short-circuit our comparison. | |
| 281 */ | |
| 282 static int Compare(const SkFlatData& a, const SkFlatData& b) { | |
| 283 const uint32_t* stop = a.dataStop(); | |
| 284 const uint32_t* a_ptr = a.dataToCompare() - 1; | |
| 285 const uint32_t* b_ptr = b.dataToCompare() - 1; | |
| 286 // We use -1 above, so we can pre-increment our pointers in the loop | |
| 287 while (*++a_ptr == *++b_ptr) {} | |
| 288 | |
| 289 if (a_ptr == stop) { // sentinel | |
| 290 SkASSERT(b.dataStop() == b_ptr); | |
| 291 return 0; | |
| 292 } | |
| 293 SkASSERT(a_ptr < a.dataStop()); | |
| 294 SkASSERT(b_ptr < b.dataStop()); | |
| 295 return (*a_ptr < *b_ptr) ? -1 : 1; | |
| 296 } | |
| 297 | |
| 298 // Adapts Compare to be used with SkTSearch | |
| 299 static bool Less(const SkFlatData& a, const SkFlatData& b) { | |
| 300 return Compare(a, b) < 0; | |
| 301 } | |
| 302 | |
| 303 int index() const { return fIndex; } | |
| 304 const void* data() const { return (const char*)this + sizeof(*this); } | |
| 305 void* data() { return (char*)this + sizeof(*this); } | |
| 306 // Our data is always 32bit aligned, so we can offer this accessor | |
| 307 uint32_t* data32() { return (uint32_t*)this->data(); } | |
| 308 // Returns the size of the flattened data. | |
| 309 size_t flatSize() const { return fFlatSize; } | |
| 310 | |
| 311 void setSentinelInCache() { | |
| 312 this->setSentinel(kInCache_Sentinel); | |
| 313 } | |
| 314 void setSentinelAsCandidate() { | |
| 315 this->setSentinel(kCandidate_Sentinel); | |
| 316 } | |
| 317 | |
| 318 uint32_t checksum() const { return fChecksum; } | |
| 319 | |
| 320 #ifdef SK_DEBUG_SIZE | |
| 321 // returns the logical size of our data. Does not return any sentinel or | |
| 322 // padding we might have. | |
| 323 size_t size() const { | |
| 324 return sizeof(SkFlatData) + fFlatSize; | |
| 325 } | |
| 326 #endif | |
| 327 | |
| 328 static SkFlatData* Create(SkFlatController* controller, const void* obj, int
index, | |
| 329 void (*flattenProc)(SkOrderedWriteBuffer&, const v
oid*)); | 273 void (*flattenProc)(SkOrderedWriteBuffer&, const v
oid*)); |
| 330 | 274 |
| 275 // Unflatten this into result, using bitmapHeap and facePlayback for bitmaps
and fonts if given. |
| 331 void unflatten(void* result, | 276 void unflatten(void* result, |
| 332 void (*unflattenProc)(SkOrderedReadBuffer&, void*), | 277 void (*unflattenProc)(SkOrderedReadBuffer&, void*), |
| 333 SkBitmapHeap* bitmapHeap = NULL, | 278 SkBitmapHeap* bitmapHeap = NULL, |
| 334 SkTypefacePlayback* facePlayback = NULL) const; | 279 SkTypefacePlayback* facePlayback = NULL) const; |
| 335 | 280 |
| 336 // When we purge an entry, we want to reuse an old index for the new entry, | 281 // Do these contain the same data? Ignores index() and topBot(). |
| 337 // so we expose this setter. | 282 bool operator==(const SkFlatData& that) const { |
| 338 void setIndex(int index) { fIndex = index; } | 283 if (this->checksum() != that.checksum()) return false; |
| 339 | 284 if (this->flatSize() != that.flatSize()) return false; |
| 340 // for unittesting | 285 return memcmp(this->data(), that.data(), this->flatSize()) == 0; |
| 341 friend bool operator==(const SkFlatData& a, const SkFlatData& b) { | |
| 342 size_t N = (const char*)a.dataStop() - (const char*)a.dataToCompare(); | |
| 343 return !memcmp(a.dataToCompare(), b.dataToCompare(), N); | |
| 344 } | 286 } |
| 345 | 287 |
| 346 // returns true if fTopBot[] has been recorded | 288 int index() const { return fIndex; } |
| 289 const uint8_t* data() const { return (const uint8_t*)this + sizeof(*this); } |
| 290 size_t flatSize() const { return fFlatSize; } |
| 291 uint32_t checksum() const { return fChecksum; } |
| 292 |
| 293 // Returns true if fTopBot[] has been recorded. |
| 347 bool isTopBotWritten() const { | 294 bool isTopBotWritten() const { |
| 348 return !SkScalarIsNaN(fTopBot[0]); | 295 return !SkScalarIsNaN(fTopBot[0]); |
| 349 } | 296 } |
| 350 | 297 |
| 351 // Returns fTopBot array, so it can be passed to a routine to compute them. | 298 // Returns fTopBot array, so it can be passed to a routine to compute them. |
| 352 // For efficiency, we assert that fTopBot have not been recorded yet. | 299 // For efficiency, we assert that fTopBot have not been recorded yet. |
| 353 SkScalar* writableTopBot() const { | 300 SkScalar* writableTopBot() const { |
| 354 SkASSERT(!this->isTopBotWritten()); | 301 SkASSERT(!this->isTopBotWritten()); |
| 355 return fTopBot; | 302 return fTopBot; |
| 356 } | 303 } |
| 357 | 304 |
| 358 // return the topbot[] after it has been recorded | 305 // Return the topbot[] after it has been recorded. |
| 359 const SkScalar* topBot() const { | 306 const SkScalar* topBot() const { |
| 360 SkASSERT(this->isTopBotWritten()); | 307 SkASSERT(this->isTopBotWritten()); |
| 361 return fTopBot; | 308 return fTopBot; |
| 362 } | 309 } |
| 363 | 310 |
| 364 private: | 311 private: |
| 365 // This is *not* part of the key for search/sort | 312 // For SkTDynamicHash |
| 366 int fIndex; | 313 static const SkFlatData& Identity(const SkFlatData& flat) { return flat; } |
| 314 static uint32_t Hash(const SkFlatData& flat) { return flat.checksum(); } |
| 315 static bool Equal(const SkFlatData& a, const SkFlatData& b) { return a == b;
} |
| 367 | 316 |
| 368 // Cache of paint's FontMetrics fTop,fBottom | 317 void setIndex(int index) { fIndex = index; } |
| 369 // initialied to [NaN,NaN] as a sentinel that they have not been recorded ye
t | 318 uint8_t* data() { return (uint8_t*)this + sizeof(*this); } |
| 370 // | |
| 371 // This is *not* part of the key for search/sort | |
| 372 mutable SkScalar fTopBot[2]; | |
| 373 | 319 |
| 374 // marks fTopBot[] as unrecorded | 320 // This assumes the payload flat data has already been written and does not
modify it. |
| 375 void setTopBotUnwritten() { | 321 void stampHeader(int index, int32_t size) { |
| 376 this->fTopBot[0] = SK_ScalarNaN; // initial to sentinel values | 322 SkASSERT(SkIsAlign4(size)); |
| 323 fIndex = index; |
| 324 fFlatSize = size; |
| 325 fTopBot[0] = SK_ScalarNaN; // Mark as unwritten. |
| 326 fChecksum = SkChecksum::Compute((uint32_t*)this->data(), size); |
| 377 } | 327 } |
| 378 | 328 |
| 379 // From here down is the data we look at in the search/sort. We always begin | 329 template <class T> friend class SkFlatDictionary; |
| 380 // with the checksum and then length. | 330 |
| 331 int fIndex; |
| 332 int32_t fFlatSize; |
| 381 uint32_t fChecksum; | 333 uint32_t fChecksum; |
| 382 int32_t fFlatSize; // size of flattened data | 334 mutable SkScalar fTopBot[2]; // Cache of FontMetrics fTop, fBottom. Starts
as [NaN,?]. |
| 383 // uint32_t flattenedData[] | 335 // uint32_t flattenedData[] implicitly hangs off the end. |
| 384 // uint32_t sentinelValue | |
| 385 | |
| 386 const uint32_t* dataToCompare() const { | |
| 387 return (const uint32_t*)&fChecksum; | |
| 388 } | |
| 389 const uint32_t* dataStop() const { | |
| 390 SkASSERT(SkIsAlign4(fFlatSize)); | |
| 391 return (const uint32_t*)((const char*)this->data() + fFlatSize); | |
| 392 } | |
| 393 | |
| 394 enum { | |
| 395 kInCache_Sentinel = 0, | |
| 396 kCandidate_Sentinel = ~0U, | |
| 397 }; | |
| 398 void setSentinel(uint32_t value) { | |
| 399 SkASSERT(SkIsAlign4(fFlatSize)); | |
| 400 this->data32()[fFlatSize >> 2] = value; | |
| 401 } | |
| 402 | |
| 403 // This does not modify the payload flat data, in case it's already been wri
tten. | |
| 404 void stampHeaderAndSentinel(int index, int32_t size); | |
| 405 template <class T> friend class SkFlatDictionary; // For stampHeaderAndSent
inel(). | |
| 406 }; | 336 }; |
| 407 | 337 |
| 408 template <class T> | 338 template <class T> |
| 409 class SkFlatDictionary { | 339 class SkFlatDictionary { |
| 410 static const size_t kWriteBufferGrowthBytes = 1024; | 340 static const size_t kWriteBufferGrowthBytes = 1024; |
| 411 | 341 |
| 412 public: | 342 public: |
| 413 SkFlatDictionary(SkFlatController* controller, size_t scratchSizeGuess = 0) | 343 SkFlatDictionary(SkFlatController* controller, size_t scratchSizeGuess = 0) |
| 414 : fFlattenProc(NULL) | 344 : fFlattenProc(NULL) |
| 415 , fUnflattenProc(NULL) | 345 , fUnflattenProc(NULL) |
| 416 , fController(SkRef(controller)) | 346 , fController(SkRef(controller)) |
| 417 , fScratchSize(scratchSizeGuess) | 347 , fScratchSize(scratchSizeGuess) |
| 418 , fScratch(AllocScratch(fScratchSize)) | 348 , fScratch(AllocScratch(fScratchSize)) |
| 419 , fWriteBuffer(kWriteBufferGrowthBytes) | 349 , fWriteBuffer(kWriteBufferGrowthBytes) |
| 420 , fWriteBufferReady(false) | 350 , fWriteBufferReady(false) { |
| 421 , fNextIndex(1) { // set to 1 since returning a zero from find() indicates
failure | 351 this->reset(); |
| 422 sk_bzero(fHash, sizeof(fHash)); | 352 } |
| 353 |
| 354 /** |
| 355 * Clears the dictionary of all entries. However, it does NOT free the |
| 356 * memory that was allocated for each entry (that's owned by controller). |
| 357 */ |
| 358 void reset() { |
| 359 fIndexed.rewind(); |
| 360 // TODO(mtklein): There's no reason to have the index start from 1. Cle
an this up. |
| 423 // index 0 is always empty since it is used as a signal that find failed | 361 // index 0 is always empty since it is used as a signal that find failed |
| 424 fIndexedData.push(NULL); | 362 fIndexed.push(NULL); |
| 363 fNextIndex = 1; |
| 425 } | 364 } |
| 426 | 365 |
| 427 ~SkFlatDictionary() { | 366 ~SkFlatDictionary() { |
| 428 sk_free(fScratch); | 367 sk_free(fScratch); |
| 429 } | 368 } |
| 430 | 369 |
| 431 int count() const { | 370 int count() const { |
| 432 SkASSERT(fIndexedData.count() == fSortedData.count()+1); | 371 return fIndexed.count() - 1; |
| 433 return fSortedData.count(); | |
| 434 } | 372 } |
| 435 | 373 |
| 436 const SkFlatData* operator[](int index) const { | 374 // For testing only. Index is zero-based. |
| 437 SkASSERT(index >= 0 && index < fSortedData.count()); | 375 const SkFlatData* operator[](int index) { |
| 438 return fSortedData[index]; | 376 return fIndexed[index+1]; |
| 439 } | 377 } |
| 440 | 378 |
| 441 /** | 379 /** |
| 442 * Clears the dictionary of all entries. However, it does NOT free the | 380 * Given an element of type T return its 1-based index in the dictionary. If |
| 443 * memory that was allocated for each entry. | 381 * the element wasn't previously in the dictionary it is automatically |
| 382 * added. |
| 383 * |
| 444 */ | 384 */ |
| 445 void reset() { | 385 int find(const T& element) { |
| 446 fSortedData.reset(); | 386 return this->findAndReturnFlat(element)->index(); |
| 447 fIndexedData.rewind(); | |
| 448 // index 0 is always empty since it is used as a signal that find failed | |
| 449 fIndexedData.push(NULL); | |
| 450 fNextIndex = 1; | |
| 451 sk_bzero(fHash, sizeof(fHash)); | |
| 452 } | 387 } |
| 453 | 388 |
| 454 /** | 389 /** |
| 455 * Similar to find. Allows the caller to specify an SkFlatData to replace in | 390 * Similar to find. Allows the caller to specify an SkFlatData to replace in |
| 456 * the case of an add. Also tells the caller whether a new SkFlatData was | 391 * the case of an add. Also tells the caller whether a new SkFlatData was |
| 457 * added and whether the old one was replaced. The parameters added and | 392 * added and whether the old one was replaced. The parameters added and |
| 458 * replaced are required to be non-NULL. Rather than returning the index of | 393 * replaced are required to be non-NULL. Rather than returning the index of |
| 459 * the entry in the dictionary, it returns the actual SkFlatData. | 394 * the entry in the dictionary, it returns the actual SkFlatData. |
| 460 */ | 395 */ |
| 461 const SkFlatData* findAndReplace(const T& element, | 396 const SkFlatData* findAndReplace(const T& element, |
| 462 const SkFlatData* toReplace, bool* added, | 397 const SkFlatData* toReplace, |
| 398 bool* added, |
| 463 bool* replaced) { | 399 bool* replaced) { |
| 464 SkASSERT(added != NULL && replaced != NULL); | 400 SkASSERT(added != NULL && replaced != NULL); |
| 465 int oldCount = fSortedData.count(); | 401 |
| 466 const SkFlatData* flat = this->findAndReturnFlat(element); | 402 const int oldCount = this->count(); |
| 467 *added = fSortedData.count() == oldCount + 1; | 403 SkFlatData* flat = this->findAndReturnMutableFlat(element); |
| 468 *replaced = false; | 404 *added = this->count() > oldCount; |
| 469 if (*added && toReplace != NULL) { | 405 |
| 470 // First, find the index of the one to replace | 406 // If we don't want to replace anything, we're done. |
| 471 int indexToReplace = fSortedData.find(toReplace); | 407 if (!*added || toReplace == NULL) { |
| 472 if (indexToReplace >= 0) { | 408 *replaced = false; |
| 473 // findAndReturnFlat set the index to fNextIndex and increased | 409 return flat; |
| 474 // fNextIndex by one. Reuse the index from the one being | |
| 475 // replaced and reset fNextIndex to the proper value. | |
| 476 int oldIndex = flat->index(); | |
| 477 const_cast<SkFlatData*>(flat)->setIndex(toReplace->index()); | |
| 478 fIndexedData[toReplace->index()] = flat; | |
| 479 fNextIndex--; | |
| 480 // Remove from the arrays. | |
| 481 fSortedData.remove(indexToReplace); | |
| 482 fIndexedData.remove(oldIndex); | |
| 483 // Remove from the hash table. | |
| 484 int oldHash = ChecksumToHashIndex(toReplace->checksum()); | |
| 485 if (fHash[oldHash] == toReplace) { | |
| 486 fHash[oldHash] = NULL; | |
| 487 } | |
| 488 // Delete the actual object. | |
| 489 fController->unalloc((void*)toReplace); | |
| 490 *replaced = true; | |
| 491 SkASSERT(fIndexedData.count() == fSortedData.count()+1); | |
| 492 } | |
| 493 } | 410 } |
| 411 |
| 412 // If we don't have the thing to replace, we're done. |
| 413 const SkFlatData* found = fHash.find(*toReplace); |
| 414 if (found == NULL) { |
| 415 *replaced = false; |
| 416 return flat; |
| 417 } |
| 418 |
| 419 // We can replace found with flat. First we have to set fNextIndex and
fIndexed back |
| 420 // to where they were before findAndReturnMutableFlat changed them. |
| 421 fIndexed.remove(flat->index()); |
| 422 fNextIndex--; |
| 423 |
| 424 // Do the replacement. |
| 425 flat->setIndex(found->index()); |
| 426 fIndexed[flat->index()] = flat; |
| 427 fHash.remove(*found); |
| 428 fController->unalloc((void*)found); |
| 429 SkASSERT(this->count() == oldCount); |
| 430 |
| 431 *replaced = true; |
| 494 return flat; | 432 return flat; |
| 495 } | 433 } |
| 496 | 434 |
| 497 /** | 435 /** |
| 498 * Given an element of type T return its 1-based index in the dictionary. If | |
| 499 * the element wasn't previously in the dictionary it is automatically | |
| 500 * added. | |
| 501 * | |
| 502 * To make the Compare function fast, we write a sentinel value at the end | |
| 503 * of each block. The blocks in our fSortedData[] all have a 0 sentinel. The | |
| 504 * newly created block we're comparing against has a -1 in the sentinel. | |
| 505 * | |
| 506 * This trick allows Compare to always loop until failure. If it fails on | |
| 507 * the sentinal value, we know the blocks are equal. | |
| 508 */ | |
| 509 int find(const T& element) { | |
| 510 return this->findAndReturnFlat(element)->index(); | |
| 511 } | |
| 512 | |
| 513 /** | |
| 514 * Unflatten the objects and return them in SkTRefArray, or return NULL | 436 * Unflatten the objects and return them in SkTRefArray, or return NULL |
| 515 * if there no objects (instead of an empty array). | 437 * if there no objects. Caller takes ownership of result. |
| 516 */ | 438 */ |
| 517 SkTRefArray<T>* unflattenToArray() const { | 439 SkTRefArray<T>* unflattenToArray() const { |
| 518 int count = fSortedData.count(); | 440 const int count = this->count(); |
| 519 SkTRefArray<T>* array = NULL; | 441 if (count == 0) { |
| 520 if (count > 0) { | 442 return NULL; |
| 521 array = SkTRefArray<T>::Create(count); | 443 } |
| 522 this->unflattenIntoArray(&array->writableAt(0)); | 444 SkTRefArray<T>* array = SkTRefArray<T>::Create(count); |
| 445 for (int i = 0; i < count; i++) { |
| 446 this->unflatten(&array->writableAt(i), fIndexed[i+1]); |
| 523 } | 447 } |
| 524 return array; | 448 return array; |
| 525 } | 449 } |
| 526 | 450 |
| 527 /** | 451 /** |
| 528 * Unflatten the specific object at the given index | 452 * Unflatten the specific object at the given index. |
| 453 * Caller takes ownership of the result. |
| 529 */ | 454 */ |
| 530 T* unflatten(int index) const { | 455 T* unflatten(int index) const { |
| 531 SkASSERT(fIndexedData.count() == fSortedData.count()+1); | 456 const SkFlatData* element = fIndexed[index]; |
| 532 const SkFlatData* element = fIndexedData[index]; | |
| 533 SkASSERT(index == element->index()); | 457 SkASSERT(index == element->index()); |
| 534 | 458 |
| 535 T* dst = new T; | 459 T* dst = new T; |
| 536 this->unflatten(dst, element); | 460 this->unflatten(dst, element); |
| 537 return dst; | 461 return dst; |
| 538 } | 462 } |
| 539 | 463 |
| 464 /** |
| 465 * Find or insert a flattened version of element into the dictionary. |
| 466 * Caller does not take ownership of the result. |
| 467 */ |
| 540 const SkFlatData* findAndReturnFlat(const T& element) { | 468 const SkFlatData* findAndReturnFlat(const T& element) { |
| 541 // Only valid until the next call to resetScratch(). | 469 return this->findAndReturnMutableFlat(element); |
| 542 const SkFlatData& scratch = this->resetScratch(element, fNextIndex); | |
| 543 | |
| 544 // See if we have it in the hash? | |
| 545 const int hashIndex = ChecksumToHashIndex(scratch.checksum()); | |
| 546 const SkFlatData* candidate = fHash[hashIndex]; | |
| 547 if (candidate != NULL && SkFlatData::Compare(scratch, *candidate) == 0)
{ | |
| 548 return candidate; | |
| 549 } | |
| 550 | |
| 551 // See if we have it at all? | |
| 552 const int index = SkTSearch<const SkFlatData, SkFlatData::Less>(fSortedD
ata.begin(), | |
| 553 fSortedD
ata.count(), | |
| 554 &scratch
, | |
| 555 sizeof(&
scratch)); | |
| 556 if (index >= 0) { | |
| 557 // Found. Update hash before we return. | |
| 558 fHash[hashIndex] = fSortedData[index]; | |
| 559 return fSortedData[index]; | |
| 560 } | |
| 561 | |
| 562 // We don't have it. Add it. | |
| 563 SkFlatData* detached = this->detachScratch(); | |
| 564 // detached will live beyond the next call to resetScratch(), but is own
ed by fController. | |
| 565 *fSortedData.insert(~index) = detached; // SkTSearch returned bit-not o
f where to insert. | |
| 566 *fIndexedData.insert(detached->index()) = detached; | |
| 567 fHash[hashIndex] = detached; | |
| 568 | |
| 569 SkASSERT(detached->index() == fNextIndex); | |
| 570 SkASSERT(fSortedData.count() == fNextIndex); | |
| 571 SkASSERT(fIndexedData.count() == fNextIndex+1); | |
| 572 fNextIndex++; | |
| 573 | |
| 574 return detached; | |
| 575 } | 470 } |
| 576 | 471 |
| 577 protected: | 472 protected: |
| 578 void (*fFlattenProc)(SkOrderedWriteBuffer&, const void*); | 473 void (*fFlattenProc)(SkOrderedWriteBuffer&, const void*); |
| 579 void (*fUnflattenProc)(SkOrderedReadBuffer&, void*); | 474 void (*fUnflattenProc)(SkOrderedReadBuffer&, void*); |
| 580 | 475 |
| 581 private: | 476 private: |
| 582 // Layout: [ SkFlatData header, 20 bytes ] [ data ..., 4-byte aligned ] [ se
ntinel, 4 bytes] | 477 // Layout: [ SkFlatData header, 20 bytes ] [ data ..., 4-byte aligned ] |
| 583 static size_t SizeWithPadding(size_t flatDataSize) { | 478 static size_t SizeWithPadding(size_t flatDataSize) { |
| 584 SkASSERT(SkIsAlign4(flatDataSize)); | 479 SkASSERT(SkIsAlign4(flatDataSize)); |
| 585 return sizeof(SkFlatData) + flatDataSize + sizeof(uint32_t); | 480 return sizeof(SkFlatData) + flatDataSize; |
| 586 } | 481 } |
| 587 | 482 |
| 588 // Allocate a new scratch SkFlatData. Must be sk_freed. | 483 // Allocate a new scratch SkFlatData. Must be sk_freed. |
| 589 static SkFlatData* AllocScratch(size_t scratchSize) { | 484 static SkFlatData* AllocScratch(size_t scratchSize) { |
| 590 return (SkFlatData*) sk_malloc_throw(SizeWithPadding(scratchSize)); | 485 return (SkFlatData*) sk_malloc_throw(SizeWithPadding(scratchSize)); |
| 591 } | 486 } |
| 592 | 487 |
| 593 // We have to delay fWriteBuffer's initialization until its first use; fCont
roller might not | 488 // We have to delay fWriteBuffer's initialization until its first use; fCont
roller might not |
| 594 // be fully set up by the time we get it in the constructor. | 489 // be fully set up by the time we get it in the constructor. |
| 595 void lazyWriteBufferInit() { | 490 void lazyWriteBufferInit() { |
| 596 if (fWriteBufferReady) { | 491 if (fWriteBufferReady) { |
| 597 return; | 492 return; |
| 598 } | 493 } |
| 599 // Without a bitmap heap, we'll flatten bitmaps into paints. That's nev
er what you want. | 494 // Without a bitmap heap, we'll flatten bitmaps into paints. That's nev
er what you want. |
| 600 SkASSERT(fController->getBitmapHeap() != NULL); | 495 SkASSERT(fController->getBitmapHeap() != NULL); |
| 601 fWriteBuffer.setBitmapHeap(fController->getBitmapHeap()); | 496 fWriteBuffer.setBitmapHeap(fController->getBitmapHeap()); |
| 602 fWriteBuffer.setTypefaceRecorder(fController->getTypefaceSet()); | 497 fWriteBuffer.setTypefaceRecorder(fController->getTypefaceSet()); |
| 603 fWriteBuffer.setNamedFactoryRecorder(fController->getNamedFactorySet()); | 498 fWriteBuffer.setNamedFactoryRecorder(fController->getNamedFactorySet()); |
| 604 fWriteBuffer.setFlags(fController->getWriteBufferFlags()); | 499 fWriteBuffer.setFlags(fController->getWriteBufferFlags()); |
| 605 fWriteBufferReady = true; | 500 fWriteBufferReady = true; |
| 606 } | 501 } |
| 607 | 502 |
| 503 SkFlatData* findAndReturnMutableFlat(const T& element) { |
| 504 // Only valid until the next call to resetScratch(). |
| 505 const SkFlatData& scratch = this->resetScratch(element, fNextIndex); |
| 506 |
| 507 SkFlatData* candidate = fHash.find(scratch); |
| 508 if (candidate != NULL) return candidate; |
| 509 |
| 510 SkFlatData* detached = this->detachScratch(); |
| 511 fHash.add(detached); |
| 512 *fIndexed.insert(fNextIndex) = detached; |
| 513 fNextIndex++; |
| 514 return detached; |
| 515 } |
| 516 |
| 608 // This reference is valid only until the next call to resetScratch() or det
achScratch(). | 517 // This reference is valid only until the next call to resetScratch() or det
achScratch(). |
| 609 const SkFlatData& resetScratch(const T& element, int index) { | 518 const SkFlatData& resetScratch(const T& element, int index) { |
| 610 this->lazyWriteBufferInit(); | 519 this->lazyWriteBufferInit(); |
| 611 | 520 |
| 612 // Flatten element into fWriteBuffer (using fScratch as storage). | 521 // Flatten element into fWriteBuffer (using fScratch as storage). |
| 613 fWriteBuffer.reset(fScratch->data(), fScratchSize); | 522 fWriteBuffer.reset(fScratch->data(), fScratchSize); |
| 614 fFlattenProc(fWriteBuffer, &element); | 523 fFlattenProc(fWriteBuffer, &element); |
| 615 const size_t bytesWritten = fWriteBuffer.bytesWritten(); | 524 const size_t bytesWritten = fWriteBuffer.bytesWritten(); |
| 616 | 525 |
| 617 // If all the flattened bytes fit into fScratch, we can skip a call to w
riteToMemory. | 526 // If all the flattened bytes fit into fScratch, we can skip a call to w
riteToMemory. |
| 618 if (!fWriteBuffer.wroteOnlyToStorage()) { | 527 if (!fWriteBuffer.wroteOnlyToStorage()) { |
| 619 SkASSERT(bytesWritten > fScratchSize); | 528 SkASSERT(bytesWritten > fScratchSize); |
| 620 // It didn't all fit. Copy into a larger replacement SkFlatData. | 529 // It didn't all fit. Copy into a larger replacement SkFlatData. |
| 621 // We can't just realloc because it might move the pointer and confu
se writeToMemory. | 530 // We can't just realloc because it might move the pointer and confu
se writeToMemory. |
| 622 SkFlatData* larger = AllocScratch(bytesWritten); | 531 SkFlatData* larger = AllocScratch(bytesWritten); |
| 623 fWriteBuffer.writeToMemory(larger->data()); | 532 fWriteBuffer.writeToMemory(larger->data()); |
| 624 | 533 |
| 625 // Carry on with this larger scratch to minimize the likelihood of f
uture resizing. | 534 // Carry on with this larger scratch to minimize the likelihood of f
uture resizing. |
| 626 sk_free(fScratch); | 535 sk_free(fScratch); |
| 627 fScratchSize = bytesWritten; | 536 fScratchSize = bytesWritten; |
| 628 fScratch = larger; | 537 fScratch = larger; |
| 629 } | 538 } |
| 630 | 539 |
| 631 // The data is in fScratch now, but we need to stamp its header and trai
ling sentinel. | 540 // The data is in fScratch now but we need to stamp its header. |
| 632 fScratch->stampHeaderAndSentinel(index, bytesWritten); | 541 fScratch->stampHeader(index, bytesWritten); |
| 633 return *fScratch; | 542 return *fScratch; |
| 634 } | 543 } |
| 635 | 544 |
| 636 // This result is owned by fController and lives as long as it does (unless
unalloc'd). | 545 // This result is owned by fController and lives as long as it does (unless
unalloc'd). |
| 637 SkFlatData* detachScratch() { | 546 SkFlatData* detachScratch() { |
| 638 // Allocate a new SkFlatData exactly big enough to hold our current scra
tch. | 547 // Allocate a new SkFlatData exactly big enough to hold our current scra
tch. |
| 639 // We use the controller for this allocation to extend the allocation's
lifetime and allow | 548 // We use the controller for this allocation to extend the allocation's
lifetime and allow |
| 640 // the controller to do whatever memory management it wants. | 549 // the controller to do whatever memory management it wants. |
| 641 const size_t paddedSize = SizeWithPadding(fScratch->flatSize()); | 550 const size_t paddedSize = SizeWithPadding(fScratch->flatSize()); |
| 642 SkFlatData* detached = (SkFlatData*)fController->allocThrow(paddedSize); | 551 SkFlatData* detached = (SkFlatData*)fController->allocThrow(paddedSize); |
| 643 | 552 |
| 644 // Copy scratch into the new SkFlatData, setting the sentinel for cache
storage. | 553 // Copy scratch into the new SkFlatData. |
| 645 memcpy(detached, fScratch, paddedSize); | 554 memcpy(detached, fScratch, paddedSize); |
| 646 detached->setSentinelInCache(); | |
| 647 | 555 |
| 648 // We can now reuse fScratch, and detached will live until fController d
ies. | 556 // We can now reuse fScratch, and detached will live until fController d
ies. |
| 649 return detached; | 557 return detached; |
| 650 } | 558 } |
| 651 | 559 |
| 652 void unflatten(T* dst, const SkFlatData* element) const { | 560 void unflatten(T* dst, const SkFlatData* element) const { |
| 653 element->unflatten(dst, fUnflattenProc, | 561 element->unflatten(dst, |
| 562 fUnflattenProc, |
| 654 fController->getBitmapHeap(), | 563 fController->getBitmapHeap(), |
| 655 fController->getTypefacePlayback()); | 564 fController->getTypefacePlayback()); |
| 656 } | 565 } |
| 657 | 566 |
| 658 void unflattenIntoArray(T* array) const { | 567 // All SkFlatData* stored in fIndexed and fHash are owned by the controller. |
| 659 const int count = fSortedData.count(); | |
| 660 SkASSERT(fIndexedData.count() == fSortedData.count()+1); | |
| 661 const SkFlatData* const* iter = fSortedData.begin(); | |
| 662 for (int i = 0; i < count; ++i) { | |
| 663 const SkFlatData* element = iter[i]; | |
| 664 int index = element->index() - 1; | |
| 665 SkASSERT((unsigned)index < (unsigned)count); | |
| 666 unflatten(&array[index], element); | |
| 667 } | |
| 668 } | |
| 669 | |
| 670 SkAutoTUnref<SkFlatController> fController; | 568 SkAutoTUnref<SkFlatController> fController; |
| 671 size_t fScratchSize; // How many bytes fScratch has allocated for data itse
lf. | 569 size_t fScratchSize; // How many bytes fScratch has allocated for data itse
lf. |
| 672 SkFlatData* fScratch; // Owned, must be freed with sk_free. | 570 SkFlatData* fScratch; // Owned, must be freed with sk_free. |
| 673 SkOrderedWriteBuffer fWriteBuffer; | 571 SkOrderedWriteBuffer fWriteBuffer; |
| 674 bool fWriteBufferReady; | 572 bool fWriteBufferReady; |
| 675 | 573 |
| 676 // SkFlatDictionary has two copies of the data one indexed by the | 574 // We map between SkFlatData and a 1-based integer index. |
| 677 // SkFlatData's index and the other sorted. The sorted data is used | |
| 678 // for finding and uniquification while the indexed copy is used | |
| 679 // for standard array-style lookups based on the SkFlatData's index | |
| 680 // (as in 'unflatten'). | |
| 681 int fNextIndex; | 575 int fNextIndex; |
| 682 SkTDArray<const SkFlatData*> fIndexedData; | |
| 683 // fSortedData is sorted by checksum/size/data. | |
| 684 SkTDArray<const SkFlatData*> fSortedData; | |
| 685 | 576 |
| 686 enum { | 577 // For index -> SkFlatData. fIndexed[0] is always NULL. |
| 687 // Determined by trying diff values on picture-recording benchmarks | 578 SkTDArray<const SkFlatData*> fIndexed; |
| 688 // (e.g. PictureRecordBench.cpp), choosing the smallest value that | |
| 689 // showed a big improvement. Even better would be to benchmark diff | |
| 690 // values on recording representative web-pages or other "real" content. | |
| 691 HASH_BITS = 7, | |
| 692 HASH_MASK = (1 << HASH_BITS) - 1, | |
| 693 HASH_COUNT = 1 << HASH_BITS | |
| 694 }; | |
| 695 const SkFlatData* fHash[HASH_COUNT]; | |
| 696 | 579 |
| 697 static int ChecksumToHashIndex(uint32_t checksum) { | 580 // For SkFlatData -> cached SkFlatData, which has index(). |
| 698 int n = checksum; | 581 SkTDynamicHash<SkFlatData, SkFlatData, |
| 699 if (HASH_BITS < 32) { | 582 SkFlatData::Identity, SkFlatData::Hash, SkFlatData::Equal> fH
ash; |
| 700 n ^= n >> 16; | |
| 701 } | |
| 702 if (HASH_BITS < 16) { | |
| 703 n ^= n >> 8; | |
| 704 } | |
| 705 if (HASH_BITS < 8) { | |
| 706 n ^= n >> 4; | |
| 707 } | |
| 708 return n & HASH_MASK; | |
| 709 } | |
| 710 }; | 583 }; |
| 711 | 584 |
| 712 /////////////////////////////////////////////////////////////////////////////// | 585 /////////////////////////////////////////////////////////////////////////////// |
| 713 // Some common dictionaries are defined here for both reference and convenience | 586 // Some common dictionaries are defined here for both reference and convenience |
| 714 /////////////////////////////////////////////////////////////////////////////// | 587 /////////////////////////////////////////////////////////////////////////////// |
| 715 | 588 |
| 716 template <class T> | 589 template <class T> |
| 717 static void SkFlattenObjectProc(SkOrderedWriteBuffer& buffer, const void* obj) { | 590 static void SkFlattenObjectProc(SkOrderedWriteBuffer& buffer, const void* obj) { |
| 718 ((T*)obj)->flatten(buffer); | 591 ((T*)obj)->flatten(buffer); |
| 719 } | 592 } |
| (...skipping 78 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 798 static void flattenRegion(SkOrderedWriteBuffer& buffer, const void* obj) { | 671 static void flattenRegion(SkOrderedWriteBuffer& buffer, const void* obj) { |
| 799 buffer.getWriter32()->writeRegion(*(SkRegion*)obj); | 672 buffer.getWriter32()->writeRegion(*(SkRegion*)obj); |
| 800 } | 673 } |
| 801 | 674 |
| 802 static void unflattenRegion(SkOrderedReadBuffer& buffer, void* obj) { | 675 static void unflattenRegion(SkOrderedReadBuffer& buffer, void* obj) { |
| 803 buffer.getReader32()->readRegion((SkRegion*)obj); | 676 buffer.getReader32()->readRegion((SkRegion*)obj); |
| 804 } | 677 } |
| 805 }; | 678 }; |
| 806 | 679 |
| 807 #endif | 680 #endif |
| OLD | NEW |