OLD | NEW |
---|---|
1 // Copyright 2015 The Chromium Authors. All rights reserved. | 1 // Copyright 2015 The Chromium Authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 #include "cc/tiles/image_decode_controller.h" | 5 #include "cc/tiles/image_decode_controller.h" |
6 | 6 |
7 #include "base/memory/discardable_memory.h" | |
7 #include "cc/debug/devtools_instrumentation.h" | 8 #include "cc/debug/devtools_instrumentation.h" |
9 #include "third_party/skia/include/core/SkCanvas.h" | |
8 | 10 |
9 namespace cc { | 11 namespace cc { |
10 namespace { | 12 namespace { |
11 | 13 |
14 // The amount of memory we can lock ahead of time (100MB). This limit is only | |
15 // used to inform the caller of the amount of space available in the cache. The | |
16 // caller can still request tasks which can cause this limit to be breached. | |
17 const size_t kLockedMemoryLimitBytes = 100 * 1024 * 1024; | |
18 | |
19 // The number of entries to keep around in the cache. This limit can be breached | |
20 // if more items are locked. That is, locked items ignore this limit. | |
21 const size_t kMaxItemsInCache = 100; | |
22 | |
12 class ImageDecodeTaskImpl : public ImageDecodeTask { | 23 class ImageDecodeTaskImpl : public ImageDecodeTask { |
13 public: | 24 public: |
14 ImageDecodeTaskImpl(ImageDecodeController* controller, | 25 ImageDecodeTaskImpl(ImageDecodeController* controller, |
15 const SkImage* image, | 26 const ImageDecodeController::ImageKey& image_key, |
16 int layer_id, | 27 const DrawImage& image, |
17 uint64_t source_prepare_tiles_id) | 28 uint64_t source_prepare_tiles_id) |
18 : controller_(controller), | 29 : controller_(controller), |
19 image_(skia::SharePtr(image)), | 30 image_key_(image_key), |
20 layer_id_(layer_id), | 31 image_(image), |
32 image_ref_(skia::SharePtr(image.image())), | |
21 source_prepare_tiles_id_(source_prepare_tiles_id) {} | 33 source_prepare_tiles_id_(source_prepare_tiles_id) {} |
22 | 34 |
23 // Overridden from Task: | 35 // Overridden from Task: |
24 void RunOnWorkerThread() override { | 36 void RunOnWorkerThread() override { |
25 TRACE_EVENT1("cc", "ImageDecodeTaskImpl::RunOnWorkerThread", | 37 TRACE_EVENT1("cc", "ImageDecodeTaskImpl::RunOnWorkerThread", |
26 "source_prepare_tiles_id", source_prepare_tiles_id_); | 38 "source_prepare_tiles_id", source_prepare_tiles_id_); |
27 devtools_instrumentation::ScopedImageDecodeTask image_decode_task( | 39 devtools_instrumentation::ScopedImageDecodeTask image_decode_task( |
28 image_.get()); | 40 image_ref_.get()); |
29 controller_->DecodeImage(image_.get()); | 41 controller_->DecodeImage(image_key_, image_); |
30 | |
31 // Release the reference after decoding image to ensure that it is not kept | |
32 // alive unless needed. | |
33 image_.clear(); | |
34 } | 42 } |
35 | 43 |
36 // Overridden from TileTask: | 44 // Overridden from TileTask: |
37 void ScheduleOnOriginThread(TileTaskClient* client) override {} | 45 void ScheduleOnOriginThread(TileTaskClient* client) override {} |
38 void CompleteOnOriginThread(TileTaskClient* client) override { | 46 void CompleteOnOriginThread(TileTaskClient* client) override { |
39 controller_->OnImageDecodeTaskCompleted(layer_id_, image_.get(), | 47 controller_->OnImageDecodeTaskCompleted(image_key_, !HasFinishedRunning()); |
40 !HasFinishedRunning()); | |
41 } | 48 } |
42 | 49 |
43 protected: | 50 protected: |
44 ~ImageDecodeTaskImpl() override {} | 51 ~ImageDecodeTaskImpl() override {} |
45 | 52 |
46 private: | 53 private: |
47 ImageDecodeController* controller_; | 54 ImageDecodeController* controller_; |
48 skia::RefPtr<const SkImage> image_; | 55 ImageDecodeController::ImageKey image_key_; |
49 int layer_id_; | 56 DrawImage image_; |
57 skia::RefPtr<const SkImage> image_ref_; | |
50 uint64_t source_prepare_tiles_id_; | 58 uint64_t source_prepare_tiles_id_; |
51 | 59 |
52 DISALLOW_COPY_AND_ASSIGN(ImageDecodeTaskImpl); | 60 DISALLOW_COPY_AND_ASSIGN(ImageDecodeTaskImpl); |
53 }; | 61 }; |
54 | 62 |
55 } // namespace | 63 } // namespace |
56 | 64 |
57 ImageDecodeController::ImageDecodeController() {} | 65 ImageDecodeController::ImageDecodeController() |
66 : locked_images_budget_(kLockedMemoryLimitBytes) {} | |
58 | 67 |
59 ImageDecodeController::~ImageDecodeController() {} | 68 ImageDecodeController::~ImageDecodeController() {} |
60 | 69 |
61 scoped_refptr<ImageDecodeTask> ImageDecodeController::GetTaskForImage( | 70 scoped_refptr<ImageDecodeTask> ImageDecodeController::GetTaskForImageAndRef( |
62 const DrawImage& image, | 71 const DrawImage& image, |
63 int layer_id, | |
64 uint64_t prepare_tiles_id) { | 72 uint64_t prepare_tiles_id) { |
65 uint32_t generation_id = image.image()->uniqueID(); | 73 // If the image already exist or if we're going to create a task for it, then |
66 scoped_refptr<ImageDecodeTask>& decode_task = | 74 // we need to ref this image. That means the image is or will be in the cache. |
67 image_decode_tasks_[layer_id][generation_id]; | 75 // When the ref goes to 0, it will be unpinned but will remain in the cache. |
68 if (!decode_task) | 76 // If the image does not fit into the budget, then we don't ref this image, |
69 decode_task = CreateTaskForImage(image.image(), layer_id, prepare_tiles_id); | 77 // since it will be decoded at raster time which is when it will be |
70 return decode_task; | 78 // temporarily put in the cache. |
79 ImageKey key = ImageKey::FromDrawImage(image); | |
80 | |
81 // Image already exists. | |
82 if (LockDecodedImageIfPossibleAndRef(key)) { | |
83 SanityCheckState(__LINE__, false); | |
84 return nullptr; | |
85 } | |
86 | |
87 base::AutoLock lock(lock_); | |
88 | |
89 scoped_refptr<ImageDecodeTask>& task = pending_image_tasks_[key]; | |
90 if (task) { | |
91 ++decoded_images_ref_counts_[key]; | |
92 SanityCheckState(__LINE__, true); | |
93 return task; | |
94 } | |
95 | |
96 // Image won't fit into the cache, we'll scale it at raster time. | |
97 if (locked_images_budget_.AvailableMemoryBytes() < key.target_bytes()) { | |
98 SanityCheckState(__LINE__, true); | |
99 return nullptr; | |
enne (OOO)
2015/10/29 22:59:45
I kind of wish there was a way to differentiate th
| |
100 } | |
101 | |
102 // We have account for memory usage when we create the task, not when the task | |
103 // puts an image in the cache, because we need to ensure that we don't create | |
104 // more tasks than we would have memory. That is, since tasks are created up | |
105 // front, it will be too late to count memory when they run. | |
106 locked_images_budget_.AddUsage(key.target_bytes()); | |
107 | |
108 task = CreateTaskForImage(key, image, prepare_tiles_id); | |
109 | |
110 ++decoded_images_ref_counts_[key]; | |
111 SanityCheckState(__LINE__, true); | |
112 return task; | |
113 } | |
114 | |
115 void ImageDecodeController::UnrefImage(const DrawImage& image) { | |
116 // When we unref the image, there are several situations we need to consider: | |
117 // 1. This image was not even scheduled to be in the locked cache, which means | |
118 // that the ref count doesn't even exist. | |
119 // 2. The ref did not reach 0, which means we have to keep the image locked. | |
120 // 3. The ref reached 0, we should unlock it. | |
121 // 3a. The image isn't in the locked cache because we didn't get to decode | |
122 // it yet (this is different from (1.) since it was actually scheduled | |
123 // to be decoded). | |
124 // 3b. Unlock the image but keep it in list. | |
125 const ImageKey& key = ImageKey::FromDrawImage(image); | |
126 | |
127 base::AutoLock lock(lock_); | |
128 auto ref_count_it = decoded_images_ref_counts_.find(key); | |
129 if (ref_count_it == decoded_images_ref_counts_.end()) { | |
130 SanityCheckState(__LINE__, true); | |
131 return; | |
132 } | |
133 | |
134 --ref_count_it->second; | |
135 if (ref_count_it->second == 0) { | |
136 decoded_images_ref_counts_.erase(key); | |
137 locked_images_budget_.SubtractUsage(key.target_bytes()); | |
138 | |
139 auto decoded_image_it = | |
140 std::find_if(decoded_images_.begin(), decoded_images_.end(), | |
141 [key](const AnnotatedDecodedImage& decoded_image) { | |
142 return key == decoded_image.first; | |
143 }); | |
144 // If we've never decoded the image before ref reached 0, then we wouldn't | |
145 // have it in our cache. This would happen if we canceled tasks. | |
146 if (decoded_image_it == decoded_images_.end()) { | |
147 SanityCheckState(__LINE__, true); | |
148 return; | |
149 } | |
150 DCHECK(decoded_image_it->second->is_locked()); | |
151 decoded_image_it->second->Unlock(); | |
152 } | |
153 SanityCheckState(__LINE__, true); | |
154 } | |
155 | |
156 bool ImageDecodeController::LockDecodedImageIfPossibleAndRef( | |
157 const ImageKey& key) { | |
158 base::AutoLock lock(lock_); | |
159 auto decoded_images_iterator = | |
160 std::find_if(decoded_images_.begin(), decoded_images_.end(), | |
161 [key](const AnnotatedDecodedImage& annotated_image) { | |
162 return key == annotated_image.first; | |
163 }); | |
164 if (decoded_images_iterator == decoded_images_.end()) { | |
165 SanityCheckState(__LINE__, true); | |
166 return false; | |
167 } | |
168 | |
169 AnnotatedDecodedImage decoded_image = *decoded_images_iterator; | |
170 decoded_images_.erase(decoded_images_iterator); | |
171 | |
172 // Figure out if the image is locked or try to lock it. | |
173 bool locked = decoded_image.second->is_locked(); | |
174 if (!locked) { | |
175 locked = decoded_image.second->Lock(); | |
176 if (!locked) { | |
177 SanityCheckState(__LINE__, true); | |
178 return false; | |
179 } | |
180 | |
181 locked_images_budget_.AddUsage(key.target_bytes()); | |
182 } | |
183 | |
184 // If the image is locked or we locked it, then it has to be in the cache. | |
185 DCHECK(std::find_if(decoded_images_.begin(), decoded_images_.end(), | |
186 [key](const AnnotatedDecodedImage& image) { | |
187 return image.first == key; | |
188 }) == decoded_images_.end()); | |
189 decoded_images_.push_back(decoded_image); | |
190 ++decoded_images_ref_counts_[key]; | |
191 SanityCheckState(__LINE__, true); | |
192 return true; | |
71 } | 193 } |
72 | 194 |
73 scoped_refptr<ImageDecodeTask> ImageDecodeController::CreateTaskForImage( | 195 scoped_refptr<ImageDecodeTask> ImageDecodeController::CreateTaskForImage( |
74 const SkImage* image, | 196 const ImageKey& key, |
75 int layer_id, | 197 const DrawImage& image, |
76 uint64_t prepare_tiles_id) { | 198 uint64_t prepare_tiles_id) { |
77 return make_scoped_refptr( | 199 return make_scoped_refptr( |
78 new ImageDecodeTaskImpl(this, image, layer_id, prepare_tiles_id)); | 200 new ImageDecodeTaskImpl(this, key, image, prepare_tiles_id)); |
79 } | 201 } |
80 | 202 |
81 void ImageDecodeController::DecodeImage(const SkImage* image) { | 203 void ImageDecodeController::DecodeImage(const ImageKey& key, |
82 image->preroll(); | 204 const DrawImage& image) { |
83 } | 205 if (!CanHandleFilterQuality(image.filter_quality())) |
84 | |
85 void ImageDecodeController::AddLayerUsedCount(int layer_id) { | |
86 ++used_layer_counts_[layer_id]; | |
87 } | |
88 | |
89 void ImageDecodeController::SubtractLayerUsedCount(int layer_id) { | |
90 if (--used_layer_counts_[layer_id]) | |
91 return; | 206 return; |
92 | 207 |
93 // Clean up decode tasks once a layer is no longer used. | 208 scoped_refptr<DecodedImage> decoded_image = |
94 used_layer_counts_.erase(layer_id); | 209 DecodeImageInternal(key, image.image()); |
95 image_decode_tasks_.erase(layer_id); | 210 |
96 } | 211 // Add the image to the cache. Don't add the budget usage, since it was |
97 | 212 // already handled by the code that created the task for this decode. |
98 void ImageDecodeController::OnImageDecodeTaskCompleted(int layer_id, | 213 base::AutoLock lock(lock_); |
99 const SkImage* image, | 214 |
215 // We could have finished all of the raster tasks (cancelled) while this image | |
216 // decode task was running, which means that we now have a locked image but no | |
217 // ref counts. Unlock it immediately in this case. | |
218 if (decoded_images_ref_counts_.find(key) == decoded_images_ref_counts_.end()) | |
219 decoded_image->Unlock(); | |
220 | |
221 DCHECK(std::find_if(decoded_images_.begin(), decoded_images_.end(), | |
222 [key](const AnnotatedDecodedImage& image) { | |
223 return image.first == key; | |
224 }) == decoded_images_.end()); | |
225 decoded_images_.push_back(AnnotatedDecodedImage(key, decoded_image)); | |
226 } | |
227 | |
228 scoped_refptr<ImageDecodeController::DecodedImage> | |
229 ImageDecodeController::DecodeImageInternal(const ImageKey& key, | |
230 const SkImage* image) { | |
231 // TODO(vmpstr, reed): Scale the image here without caching it in skia. | |
232 SkImageInfo info = SkImageInfo::MakeN32Premul(key.target_size().width(), | |
233 key.target_size().height()); | |
234 scoped_ptr<base::DiscardableMemory> locked_memory = | |
235 base::DiscardableMemoryAllocator::GetInstance() | |
236 ->AllocateLockedDiscardableMemory(info.minRowBytes() * info.height()); | |
237 | |
238 skia::RefPtr<SkCanvas> canvas = skia::AdoptRef(SkCanvas::NewRasterDirect( | |
239 info, locked_memory->data(), info.minRowBytes())); | |
240 canvas->setMatrix(SkMatrix::MakeRectToRect( | |
241 SkRect::MakeWH(image->width(), image->height()), | |
242 SkRect::MakeWH(key.target_size().width(), key.target_size().height()), | |
243 SkMatrix::kFill_ScaleToFit)); | |
244 canvas->clear(SK_ColorTRANSPARENT); | |
245 SkPaint paint; | |
246 paint.setFilterQuality(kHigh_SkFilterQuality); | |
247 canvas->drawImage(image, 0, 0, &paint); | |
248 canvas->flush(); | |
249 | |
250 return make_scoped_refptr(new DecodedImage(info, locked_memory.Pass())); | |
251 } | |
252 | |
253 void ImageDecodeController::OnImageDecodeTaskCompleted(const ImageKey& key, | |
100 bool was_canceled) { | 254 bool was_canceled) { |
101 // If the task has successfully finished, then keep the task until the layer | 255 pending_image_tasks_.erase(key); |
102 // is no longer in use. This ensures that we only decode a image once. | 256 SanityCheckState(__LINE__, false); |
103 // TODO(vmpstr): Remove this when decode lifetime is controlled by cc. | 257 } |
104 if (!was_canceled) | 258 |
259 DecodedDrawImage ImageDecodeController::GetDecodedImageAndRef( | |
260 const DrawImage& image) { | |
261 if (!CanHandleFilterQuality(image.filter_quality())) { | |
262 return DecodedDrawImage(image.image(), SkSize::Make(1.f, 1.f), | |
263 image.filter_quality()); | |
264 } | |
265 | |
266 ImageKey key = ImageKey::FromDrawImage(image); | |
267 | |
268 base::AutoLock lock(lock_); | |
269 auto decoded_images_iterator = | |
270 std::find_if(decoded_images_.begin(), decoded_images_.end(), | |
271 [key](const AnnotatedDecodedImage& annotated_image) { | |
272 return key == annotated_image.first; | |
273 }); | |
274 scoped_refptr<DecodedImage> decoded_image; | |
275 if (decoded_images_iterator != decoded_images_.end()) { | |
276 decoded_image = decoded_images_iterator->second; | |
277 if (!decoded_image->is_locked()) { | |
278 if (decoded_image->Lock()) { | |
279 locked_images_budget_.AddUsage(key.target_bytes()); | |
280 } else { | |
281 decoded_images_.erase(decoded_images_iterator); | |
282 decoded_image = nullptr; | |
283 } | |
284 } | |
285 } | |
286 | |
287 if (!decoded_image) { | |
288 // This means that we didn't have an image task to decode this (otherwise it | |
289 // would have run and locked the image already). So, we need to decode/scale | |
290 // in place. | |
291 decoded_image = DecodeImageInternal(key, image.image()); | |
292 DCHECK(std::find_if(decoded_images_.begin(), decoded_images_.end(), | |
293 [key](const AnnotatedDecodedImage& image) { | |
294 return image.first == key; | |
295 }) == decoded_images_.end()); | |
296 decoded_images_.push_back(AnnotatedDecodedImage(key, decoded_image)); | |
297 locked_images_budget_.AddUsage(key.target_bytes()); | |
298 } | |
299 | |
300 DCHECK(decoded_image->is_locked()) << key.ToString(); | |
301 float x_scale = | |
302 key.target_size().width() / static_cast<float>(image.image()->width()); | |
303 float y_scale = | |
304 key.target_size().height() / static_cast<float>(image.image()->height()); | |
305 | |
306 ++decoded_images_ref_counts_[key]; | |
307 SanityCheckState(__LINE__, true); | |
308 return DecodedDrawImage(decoded_image->image(), | |
309 SkSize::Make(x_scale, y_scale), kLow_SkFilterQuality); | |
310 } | |
311 | |
312 void ImageDecodeController::DrawWithImageFinished(const DrawImage& image) { | |
313 if (!CanHandleFilterQuality(image.filter_quality())) | |
105 return; | 314 return; |
106 | 315 UnrefImage(image); |
107 // Otherwise, we have to clean up the task so that a new one can be created if | 316 SanityCheckState(__LINE__, false); |
108 // we need to decode the image again. | 317 } |
109 LayerImageTaskMap::iterator layer_it = image_decode_tasks_.find(layer_id); | 318 |
110 if (layer_it == image_decode_tasks_.end()) | 319 bool ImageDecodeController::CanHandleFilterQuality( |
320 SkFilterQuality filter_quality) { | |
321 DCHECK(filter_quality != kNone_SkFilterQuality); | |
322 // We don't need to handle low quality filters. | |
323 if (filter_quality == kLow_SkFilterQuality) | |
324 return false; | |
325 // TODO(vmpstr): We need to start caching mipmaps for medium quality and | |
326 // caching the interpolated values from those. For now, we don't have this. | |
327 if (filter_quality == kMedium_SkFilterQuality) | |
328 return false; | |
329 DCHECK(filter_quality == kHigh_SkFilterQuality); | |
330 return true; | |
331 } | |
332 | |
333 void ImageDecodeController::ReduceCacheUsage() { | |
334 base::AutoLock lock(lock_); | |
335 size_t num_to_remove = (decoded_images_.size() > kMaxItemsInCache) | |
336 ? (decoded_images_.size() - kMaxItemsInCache) | |
337 : 0; | |
338 for (auto it = decoded_images_.begin(); | |
339 num_to_remove != 0 && it != decoded_images_.end();) { | |
340 if (it->second->is_locked()) { | |
341 ++it; | |
342 continue; | |
343 } | |
344 | |
345 it = decoded_images_.erase(it); | |
346 --num_to_remove; | |
347 } | |
348 } | |
349 | |
350 void ImageDecodeController::SanityCheckState(int line, bool lock_acquired) { | |
351 #if DCHECK_IS_ON() | |
352 if (!lock_acquired) { | |
353 base::AutoLock lock(lock_); | |
354 SanityCheckState(line, true); | |
111 return; | 355 return; |
112 | 356 } |
113 ImageTaskMap& image_tasks = layer_it->second; | 357 |
114 ImageTaskMap::iterator task_it = image_tasks.find(image->uniqueID()); | 358 MemoryBudget budget(kLockedMemoryLimitBytes); |
115 if (task_it == image_tasks.end()) | 359 for (const auto& annotated_image : decoded_images_) { |
116 return; | 360 auto ref_it = decoded_images_ref_counts_.find(annotated_image.first); |
117 image_tasks.erase(task_it); | 361 if (annotated_image.second->is_locked()) { |
362 budget.AddUsage(annotated_image.first.target_bytes()); | |
363 DCHECK(ref_it != decoded_images_ref_counts_.end()) << line; | |
364 } else { | |
365 DCHECK(ref_it == decoded_images_ref_counts_.end() || | |
366 pending_image_tasks_.find(annotated_image.first) != | |
367 pending_image_tasks_.end()) | |
368 << line; | |
369 } | |
370 } | |
371 DCHECK_GE(budget.AvailableMemoryBytes(), | |
372 locked_images_budget_.AvailableMemoryBytes()) | |
373 << line; | |
374 #endif // DCHECK_IS_ON() | |
375 } | |
376 | |
377 // ImageDecodeControllerKey | |
378 ImageDecodeControllerKey ImageDecodeControllerKey::FromDrawImage( | |
379 const DrawImage& image) { | |
380 gfx::Size target_size(std::abs(SkScalarRoundToInt(image.image()->width() * | |
381 image.scale().width())), | |
382 std::abs(SkScalarRoundToInt(image.image()->height() * | |
383 image.scale().height()))); | |
384 return ImageDecodeControllerKey(image.image()->uniqueID(), target_size, | |
385 image.filter_quality()); | |
386 } | |
387 | |
388 ImageDecodeControllerKey::ImageDecodeControllerKey( | |
389 uint32_t image_id, | |
390 const gfx::Size& size, | |
391 SkFilterQuality filter_quality) | |
392 : image_id_(image_id), size_(size), filter_quality_(filter_quality) {} | |
393 | |
394 std::string ImageDecodeControllerKey::ToString() const { | |
395 std::ostringstream str; | |
396 str << "id[" << image_id_ << "] size[" << size_.width() << "x" | |
397 << size_.height() << "] filter_quality[" << filter_quality_ << "]"; | |
398 return str.str(); | |
399 } | |
400 | |
401 // TODO(vmpstr): Not sure if this needs to do something. It could have some | |
402 // extra state to DCHECK if we pass it DecodedImage object as the context. | |
403 static void noop(const void* pixels, void* context) {} | |
404 | |
405 // DecodedImage | |
406 ImageDecodeController::DecodedImage::DecodedImage( | |
407 const SkImageInfo& info, | |
408 scoped_ptr<base::DiscardableMemory> memory) | |
409 : locked_(true), image_info_(info), memory_(memory.Pass()) { | |
410 CreateImageFromLockedMemory(); | |
411 } | |
412 | |
413 ImageDecodeController::DecodedImage::~DecodedImage() {} | |
414 | |
415 bool ImageDecodeController::DecodedImage::Lock() { | |
416 DCHECK(!locked_); | |
417 bool success = memory_->Lock(); | |
418 if (!success) | |
419 return false; | |
420 locked_ = true; | |
421 CreateImageFromLockedMemory(); | |
422 return true; | |
423 } | |
424 | |
425 void ImageDecodeController::DecodedImage::Unlock() { | |
426 DCHECK(locked_); | |
427 image_.clear(); | |
428 memory_->Unlock(); | |
429 locked_ = false; | |
430 } | |
431 | |
432 void ImageDecodeController::DecodedImage::CreateImageFromLockedMemory() { | |
433 image_ = skia::AdoptRef(SkImage::NewFromRaster( | |
434 image_info_, memory_->data(), image_info_.minRowBytes(), &noop, nullptr)); | |
435 } | |
436 | |
437 // MemoryBudget | |
438 ImageDecodeController::MemoryBudget::MemoryBudget(size_t limit_bytes) | |
439 : limit_bytes_(limit_bytes), current_usage_bytes_(0u) {} | |
440 | |
441 size_t ImageDecodeController::MemoryBudget::AvailableMemoryBytes() const { | |
442 size_t usage = GetCurrentUsageSafe(); | |
443 return usage >= limit_bytes_ ? 0u : (limit_bytes_ - usage); | |
444 } | |
445 | |
446 void ImageDecodeController::MemoryBudget::AddUsage(size_t usage) { | |
447 current_usage_bytes_ += usage; | |
448 } | |
449 | |
450 void ImageDecodeController::MemoryBudget::SubtractUsage(size_t usage) { | |
451 DCHECK_GE(current_usage_bytes_.ValueOrDefault(0u), usage); | |
452 current_usage_bytes_ -= usage; | |
453 } | |
454 | |
455 void ImageDecodeController::MemoryBudget::ResetUsage() { | |
456 current_usage_bytes_ = 0; | |
457 } | |
458 | |
459 bool ImageDecodeController::MemoryBudget::NeedsReset() const { | |
460 return !current_usage_bytes_.IsValid(); | |
461 } | |
462 | |
463 size_t ImageDecodeController::MemoryBudget::GetCurrentUsageSafe() const { | |
464 return current_usage_bytes_.ValueOrDefault( | |
465 std::numeric_limits<size_t>::max()); | |
118 } | 466 } |
119 | 467 |
120 } // namespace cc | 468 } // namespace cc |
OLD | NEW |