OLD | NEW |
1 /* | 1 /* |
2 * Copyright (C) 2010 Google Inc. All rights reserved. | 2 * Copyright (C) 2010 Google Inc. All rights reserved. |
3 * | 3 * |
4 * Redistribution and use in source and binary forms, with or without | 4 * Redistribution and use in source and binary forms, with or without |
5 * modification, are permitted provided that the following conditions | 5 * modification, are permitted provided that the following conditions |
6 * are met: | 6 * are met: |
7 * | 7 * |
8 * 1. Redistributions of source code must retain the above copyright | 8 * 1. Redistributions of source code must retain the above copyright |
9 * notice, this list of conditions and the following disclaimer. | 9 * notice, this list of conditions and the following disclaimer. |
10 * 2. Redistributions in binary form must reproduce the above copyright | 10 * 2. Redistributions in binary form must reproduce the above copyright |
(...skipping 15 matching lines...) Expand all Loading... |
26 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | 26 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
27 */ | 27 */ |
28 | 28 |
29 #include "config.h" | 29 #include "config.h" |
30 #include "core/platform/image-decoders/webp/WEBPImageDecoder.h" | 30 #include "core/platform/image-decoders/webp/WEBPImageDecoder.h" |
31 | 31 |
32 #include "core/platform/PlatformInstrumentation.h" | 32 #include "core/platform/PlatformInstrumentation.h" |
33 | 33 |
34 #ifdef QCMS_WEBP_COLOR_CORRECTION | 34 #ifdef QCMS_WEBP_COLOR_CORRECTION |
35 #include "qcms.h" | 35 #include "qcms.h" |
36 #include "webp/demux.h" | |
37 #else | |
38 #undef ICCP_FLAG | |
39 #define ICCP_FLAG 0 | |
40 #endif | 36 #endif |
41 | 37 |
42 // Backward emulation for earlier versions than 0.1.99. | 38 #ifdef WEBP_ICC_ANIMATION_SUPPORT |
| 39 #include "RuntimeEnabledFeatures.h" |
| 40 #include "webp/format_constants.h" |
| 41 #endif |
| 42 |
43 #if (WEBP_DECODER_ABI_VERSION < 0x0163) | 43 #if (WEBP_DECODER_ABI_VERSION < 0x0163) |
| 44 // Backward emulation for versions earlier than 0.1.99. |
44 #define MODE_rgbA MODE_RGBA | 45 #define MODE_rgbA MODE_RGBA |
45 #define MODE_bgrA MODE_BGRA | 46 #define MODE_bgrA MODE_BGRA |
| 47 #define ALPHA_FLAG 0 |
| 48 #elif (WEBP_DECODER_ABI_VERSION <= 0x0200) |
| 49 // Backward emulation for versions earlier than 0.3.0. |
| 50 #define ALPHA_FLAG 0x000010 |
46 #endif | 51 #endif |
47 | 52 |
48 #if CPU(BIG_ENDIAN) || CPU(MIDDLE_ENDIAN) | 53 #if CPU(BIG_ENDIAN) || CPU(MIDDLE_ENDIAN) |
49 inline WEBP_CSP_MODE outputMode(bool hasAlpha) { return hasAlpha ? MODE_rgbA : M
ODE_RGBA; } | 54 inline WEBP_CSP_MODE outputMode(bool hasAlpha) { return hasAlpha ? MODE_rgbA : M
ODE_RGBA; } |
50 #elif SK_B32_SHIFT | 55 #elif SK_B32_SHIFT |
51 inline WEBP_CSP_MODE outputMode(bool hasAlpha) { return hasAlpha ? MODE_rgbA : M
ODE_RGBA; } | 56 inline WEBP_CSP_MODE outputMode(bool hasAlpha) { return hasAlpha ? MODE_rgbA : M
ODE_RGBA; } |
52 #else // LITTLE_ENDIAN, output BGRA pixels. | 57 #else // LITTLE_ENDIAN, output BGRA pixels. |
53 inline WEBP_CSP_MODE outputMode(bool hasAlpha) { return hasAlpha ? MODE_bgrA : M
ODE_BGRA; } | 58 inline WEBP_CSP_MODE outputMode(bool hasAlpha) { return hasAlpha ? MODE_bgrA : M
ODE_BGRA; } |
54 #endif | 59 #endif |
55 | 60 |
56 namespace WebCore { | 61 namespace WebCore { |
57 | 62 |
58 WEBPImageDecoder::WEBPImageDecoder(ImageSource::AlphaOption alphaOption, | 63 WEBPImageDecoder::WEBPImageDecoder(ImageSource::AlphaOption alphaOption, |
59 ImageSource::GammaAndColorProfileOption gamma
AndColorProfileOption) | 64 ImageSource::GammaAndColorProfileOption gamma
AndColorProfileOption) |
60 : ImageDecoder(alphaOption, gammaAndColorProfileOption) | 65 : ImageDecoder(alphaOption, gammaAndColorProfileOption) |
61 , m_decoder(0) | 66 , m_decoder(0) |
62 , m_hasAlpha(false) | |
63 , m_formatFlags(0) | 67 , m_formatFlags(0) |
64 #ifdef QCMS_WEBP_COLOR_CORRECTION | 68 #ifdef QCMS_WEBP_COLOR_CORRECTION |
65 , m_haveReadProfile(false) | 69 , m_haveReadProfile(false) |
66 , m_transform(0) | 70 , m_transform(0) |
| 71 #endif |
| 72 #ifdef WEBP_ICC_ANIMATION_SUPPORT |
| 73 , m_demux(0) |
| 74 , m_demuxState(WEBP_DEMUX_PARSING_HEADER) |
| 75 , m_haveAlreadyParsedThisData(false) |
| 76 , m_haveReadAnimationParameters(false) |
| 77 , m_repetitionCount(cAnimationLoopOnce) |
67 , m_decodedHeight(0) | 78 , m_decodedHeight(0) |
68 #endif | 79 #endif |
69 { | 80 { |
70 WebPInitDecBuffer(&m_decoderBuffer); | |
71 } | 81 } |
72 | 82 |
73 WEBPImageDecoder::~WEBPImageDecoder() | 83 WEBPImageDecoder::~WEBPImageDecoder() |
74 { | 84 { |
75 clear(); | 85 clear(); |
76 } | 86 } |
77 | 87 |
78 void WEBPImageDecoder::clear() | 88 void WEBPImageDecoder::clear() |
79 { | 89 { |
80 #ifdef QCMS_WEBP_COLOR_CORRECTION | 90 #ifdef QCMS_WEBP_COLOR_CORRECTION |
81 if (m_transform) | 91 if (m_transform) |
82 qcms_transform_release(m_transform); | 92 qcms_transform_release(m_transform); |
83 m_transform = 0; | 93 m_transform = 0; |
84 #endif | 94 #endif |
85 WebPFreeDecBuffer(&m_decoderBuffer); | 95 #ifdef WEBP_ICC_ANIMATION_SUPPORT |
86 if (m_decoder) | 96 WebPDemuxDelete(m_demux); |
87 WebPIDelete(m_decoder); | 97 m_demux = 0; |
| 98 #endif |
| 99 clearDecoder(); |
| 100 } |
| 101 |
| 102 void WEBPImageDecoder::clearDecoder() |
| 103 { |
| 104 WebPIDelete(m_decoder); |
88 m_decoder = 0; | 105 m_decoder = 0; |
| 106 #ifdef WEBP_ICC_ANIMATION_SUPPORT |
| 107 m_decodedHeight = 0; |
| 108 #endif |
89 } | 109 } |
90 | 110 |
91 bool WEBPImageDecoder::isSizeAvailable() | 111 bool WEBPImageDecoder::isSizeAvailable() |
92 { | 112 { |
93 if (!ImageDecoder::isSizeAvailable()) | 113 if (!ImageDecoder::isSizeAvailable()) { |
94 decode(true); | 114 #ifdef WEBP_ICC_ANIMATION_SUPPORT |
95 | 115 if (!updateDemuxer()) |
| 116 return 0; |
| 117 #else |
| 118 decode(reinterpret_cast<const uint8_t*>(m_data->data()), m_data->size(),
true, 0); |
| 119 #endif |
| 120 } |
96 return ImageDecoder::isSizeAvailable(); | 121 return ImageDecoder::isSizeAvailable(); |
97 } | 122 } |
98 | 123 |
99 ImageFrame* WEBPImageDecoder::frameBufferAtIndex(size_t index) | 124 size_t WEBPImageDecoder::frameCount() |
100 { | 125 { |
101 if (index) | 126 #ifdef WEBP_ICC_ANIMATION_SUPPORT |
| 127 if (!updateDemuxer()) |
102 return 0; | 128 return 0; |
103 | 129 #else |
104 if (m_frameBufferCache.isEmpty()) { | 130 if (m_frameBufferCache.isEmpty()) { |
105 m_frameBufferCache.resize(1); | 131 m_frameBufferCache.resize(1); |
106 m_frameBufferCache[0].setPremultiplyAlpha(m_premultiplyAlpha); | 132 m_frameBufferCache[0].setPremultiplyAlpha(m_premultiplyAlpha); |
107 } | 133 } |
108 | 134 #endif |
109 ImageFrame& frame = m_frameBufferCache[0]; | 135 return m_frameBufferCache.size(); |
110 if (frame.status() != ImageFrame::FrameComplete) { | 136 } |
111 PlatformInstrumentation::willDecodeImage("WEBP"); | 137 |
112 decode(false); | 138 ImageFrame* WEBPImageDecoder::frameBufferAtIndex(size_t index) |
113 PlatformInstrumentation::didDecodeImage(); | 139 { |
114 } | 140 if (index >= frameCount()) |
| 141 return 0; |
| 142 |
| 143 ImageFrame& frame = m_frameBufferCache[index]; |
| 144 if (frame.status() == ImageFrame::FrameComplete) |
| 145 return &frame; |
| 146 |
| 147 #ifdef WEBP_ICC_ANIMATION_SUPPORT |
| 148 if (RuntimeEnabledFeatures::animatedWebPEnabled()) { |
| 149 Vector<size_t> framesToDecode; |
| 150 size_t frameToDecode = index; |
| 151 do { |
| 152 framesToDecode.append(frameToDecode); |
| 153 frameToDecode = m_frameBufferCache[frameToDecode].requiredPreviousFr
ameIndex(); |
| 154 } while (frameToDecode != notFound && m_frameBufferCache[frameToDecode].
status() != ImageFrame::FrameComplete); |
| 155 |
| 156 ASSERT(m_demux); |
| 157 // The |rend| variable is needed by some compilers that can't correctly |
| 158 // select from const and non-const versions of overloaded functions. |
| 159 // Can remove the variable if Android compiler can compile |
| 160 // 'iter != framesToDecode.rend()'. |
| 161 Vector<size_t>::const_reverse_iterator rend = framesToDecode.rend(); |
| 162 for (Vector<size_t>::const_reverse_iterator iter = framesToDecode.rbegin
(); iter != rend; ++iter) { |
| 163 size_t frameIndex = *iter; |
| 164 WebPIterator webpFrame; |
| 165 if (!WebPDemuxGetFrame(m_demux, frameIndex + 1, &webpFrame)) |
| 166 return 0; |
| 167 if ((m_formatFlags & ANIMATION_FLAG) && !initFrameBuffer(webpFrame,
frameIndex)) { |
| 168 WebPDemuxReleaseIterator(&webpFrame); |
| 169 return 0; |
| 170 } |
| 171 PlatformInstrumentation::willDecodeImage("WEBP"); |
| 172 decode(webpFrame.fragment.bytes, webpFrame.fragment.size, false, fra
meIndex); |
| 173 PlatformInstrumentation::didDecodeImage(); |
| 174 WebPDemuxReleaseIterator(&webpFrame); |
| 175 |
| 176 // We need more data to continue decoding. |
| 177 if (m_frameBufferCache[frameIndex].status() != ImageFrame::FrameComp
lete) |
| 178 break; |
| 179 } |
| 180 |
| 181 // It is also a fatal error if all data is received and we have decoded
all |
| 182 // frames available but the file is truncated. |
| 183 if (index >= m_frameBufferCache.size() - 1 && isAllDataReceived() && m_d
emux && m_demuxState != WEBP_DEMUX_DONE) |
| 184 setFailed(); |
| 185 |
| 186 return &frame; |
| 187 } |
| 188 #endif |
| 189 |
| 190 ASSERT(!index); |
| 191 PlatformInstrumentation::willDecodeImage("WEBP"); |
| 192 decode(reinterpret_cast<const uint8_t*>(m_data->data()), m_data->size(), fal
se, index); |
| 193 PlatformInstrumentation::didDecodeImage(); |
115 return &frame; | 194 return &frame; |
116 } | 195 } |
117 | 196 |
| 197 #ifdef WEBP_ICC_ANIMATION_SUPPORT |
| 198 |
| 199 void WEBPImageDecoder::setData(SharedBuffer* data, bool allDataReceived) |
| 200 { |
| 201 if (failed()) |
| 202 return; |
| 203 |
| 204 ImageDecoder::setData(data, allDataReceived); |
| 205 |
| 206 // Mark that we have new data. |
| 207 if (m_demuxState != WEBP_DEMUX_DONE) |
| 208 m_haveAlreadyParsedThisData = false; |
| 209 } |
| 210 |
| 211 int WEBPImageDecoder::repetitionCount() const |
| 212 { |
| 213 return failed() ? cAnimationLoopOnce : m_repetitionCount; |
| 214 } |
| 215 |
| 216 bool WEBPImageDecoder::frameIsCompleteAtIndex(size_t index) const |
| 217 { |
| 218 if (!m_demux || m_demuxState <= WEBP_DEMUX_PARSING_HEADER) |
| 219 return false; |
| 220 if (!(m_formatFlags & ANIMATION_FLAG)) |
| 221 return ImageDecoder::frameIsCompleteAtIndex(index); |
| 222 // In case of animation, number of complete frames is same as m_frameBufferC
ache.size() (see updateDemuxer()). |
| 223 return index < m_frameBufferCache.size(); |
| 224 } |
| 225 |
| 226 float WEBPImageDecoder::frameDurationAtIndex(size_t index) const |
| 227 { |
| 228 return index < m_frameBufferCache.size() ? m_frameBufferCache[index].duratio
n() : 0; |
| 229 } |
| 230 |
| 231 bool WEBPImageDecoder::updateDemuxer() |
| 232 { |
| 233 if (m_haveAlreadyParsedThisData) |
| 234 return true; |
| 235 |
| 236 m_haveAlreadyParsedThisData = true; |
| 237 |
| 238 static const size_t minSizeForDemux = RIFF_HEADER_SIZE + CHUNK_HEADER_SIZE; |
| 239 if (m_data->size() < minSizeForDemux) |
| 240 return false; // Wait for headers so that WebPDemuxPartial doesn't retur
n null. |
| 241 |
| 242 WebPDemuxDelete(m_demux); |
| 243 WebPData inputData = { reinterpret_cast<const uint8_t*>(m_data->data()), m_d
ata->size() }; |
| 244 m_demux = WebPDemuxPartial(&inputData, &m_demuxState); |
| 245 if (!m_demux) |
| 246 return setFailed(); // Must be a failure as we have at least 'minSizeFor
Demux' bytes. |
| 247 |
| 248 if (m_demuxState <= WEBP_DEMUX_PARSING_HEADER) |
| 249 return false; // Not enough data for parsing canvas width/height yet. |
| 250 |
| 251 bool hasAnimation = (m_formatFlags & ANIMATION_FLAG); |
| 252 if (!ImageDecoder::isSizeAvailable()) { |
| 253 m_formatFlags = WebPDemuxGetI(m_demux, WEBP_FF_FORMAT_FLAGS); |
| 254 hasAnimation = (m_formatFlags & ANIMATION_FLAG); |
| 255 if (hasAnimation && !RuntimeEnabledFeatures::animatedWebPEnabled()) |
| 256 return setFailed(); |
| 257 if (!setSize(WebPDemuxGetI(m_demux, WEBP_FF_CANVAS_WIDTH), WebPDemuxGetI
(m_demux, WEBP_FF_CANVAS_HEIGHT))) |
| 258 return setFailed(); |
| 259 } |
| 260 ASSERT(ImageDecoder::isSizeAvailable()); |
| 261 const size_t newFrameCount = WebPDemuxGetI(m_demux, WEBP_FF_FRAME_COUNT); |
| 262 if (hasAnimation && !m_haveReadAnimationParameters && (newFrameCount >= 1))
{ |
| 263 // As we have parsed at least one frame (even if partially), |
| 264 // we must already have parsed the animation properties. |
| 265 // This is because ANIM chunk always precedes ANMF chunks. |
| 266 m_repetitionCount = WebPDemuxGetI(m_demux, WEBP_FF_LOOP_COUNT); |
| 267 ASSERT(m_repetitionCount == (m_repetitionCount & 0xffff)); // Loop count
is always <= 16 bits. |
| 268 if (!m_repetitionCount) |
| 269 m_repetitionCount = cAnimationLoopInfinite; |
| 270 m_haveReadAnimationParameters = true; |
| 271 } |
| 272 const size_t oldFrameCount = m_frameBufferCache.size(); |
| 273 if (newFrameCount > oldFrameCount) { |
| 274 m_frameBufferCache.resize(newFrameCount); |
| 275 for (size_t i = oldFrameCount; i < newFrameCount; ++i) { |
| 276 m_frameBufferCache[i].setPremultiplyAlpha(m_premultiplyAlpha); |
| 277 if (!hasAnimation) |
| 278 continue; |
| 279 WebPIterator animatedFrame; |
| 280 int getSuccess = WebPDemuxGetFrame(m_demux, i + 1, &animatedFrame); |
| 281 ASSERT(getSuccess == 1 && animatedFrame.complete == 1); // Always tr
ue for animated case. |
| 282 m_frameBufferCache[i].setDuration(animatedFrame.duration); |
| 283 m_frameBufferCache[i].setDisposalMethod(animatedFrame.dispose_method
== WEBP_MUX_DISPOSE_BACKGROUND ? ImageFrame::DisposeOverwriteBgcolor : ImageFra
me::DisposeKeep); |
| 284 WebPDemuxReleaseIterator(&animatedFrame); |
| 285 m_frameBufferCache[i].setRequiredPreviousFrameIndex(findRequiredPrev
iousFrame(i)); |
| 286 } |
| 287 } |
| 288 return true; |
| 289 } |
| 290 |
| 291 bool WEBPImageDecoder::initFrameBuffer(const WebPIterator& frame, size_t frameIn
dex) |
| 292 { |
| 293 ImageFrame& buffer = m_frameBufferCache[frameIndex]; |
| 294 if (buffer.status() != ImageFrame::FrameEmpty) // Already initialized. |
| 295 return true; |
| 296 |
| 297 // Initialize the frame rect in our buffer. |
| 298 IntRect frameRect(frame.x_offset, frame.y_offset, frame.width, frame.height)
; |
| 299 |
| 300 // Make sure the frameRect doesn't extend outside the buffer. |
| 301 if (frameRect.maxX() > size().width()) |
| 302 frameRect.setWidth(size().width() - frame.x_offset); |
| 303 if (frameRect.maxY() > size().height()) |
| 304 frameRect.setHeight(size().height() - frame.y_offset); |
| 305 buffer.setOriginalFrameRect(frameRect); |
| 306 |
| 307 const size_t requiredPreviousFrameIndex = findRequiredPreviousFrame(frameInd
ex); |
| 308 if (requiredPreviousFrameIndex == notFound) { |
| 309 // This frame doesn't rely on any previous data. |
| 310 if (!buffer.setSize(size().width(), size().height())) |
| 311 return setFailed(); |
| 312 } else { |
| 313 const ImageFrame& prevBuffer = m_frameBufferCache[requiredPreviousFrameI
ndex]; |
| 314 ASSERT(prevBuffer.status() == ImageFrame::FrameComplete); |
| 315 |
| 316 // Preserve the last frame as the starting state for this frame. |
| 317 if (!buffer.copyBitmapData(prevBuffer)) |
| 318 return setFailed(); |
| 319 } |
| 320 |
| 321 buffer.setStatus(ImageFrame::FramePartial); |
| 322 // The buffer is transparent outside the decoded area while the image is loa
ding. |
| 323 // The correct value of 'hasAlpha' for the frame will be set when it is full
y decoded. |
| 324 buffer.setHasAlpha(true); |
| 325 return true; |
| 326 } |
| 327 |
| 328 size_t WEBPImageDecoder::clearCacheExceptFrame(size_t clearExceptFrame) |
| 329 { |
| 330 // Don't clear if there are no frames or only one frame. |
| 331 if (m_frameBufferCache.size() <= 1) |
| 332 return 0; |
| 333 |
| 334 // We need to preserve frames such that: |
| 335 // 1. We don't clear |clearExceptFrame|; |
| 336 // 2. We don't clear any frame from which a future initFrameBuffer() or |
| 337 // applyPostProcessing() call could copy bitmap data. |
| 338 // All other frames can be cleared. |
| 339 while ((clearExceptFrame < m_frameBufferCache.size()) && (m_frameBufferCache
[clearExceptFrame].status() != ImageFrame::FrameComplete)) |
| 340 clearExceptFrame = m_frameBufferCache[clearExceptFrame].requiredPrevious
FrameIndex(); |
| 341 |
| 342 size_t frameBytesCleared = 0; |
| 343 for (size_t i = 0; i < m_frameBufferCache.size(); ++i) { |
| 344 if (i != clearExceptFrame) { |
| 345 frameBytesCleared += frameBytesAtIndex(i); |
| 346 clearFrameBuffer(i); |
| 347 } |
| 348 } |
| 349 return frameBytesCleared; |
| 350 } |
| 351 |
| 352 void WEBPImageDecoder::clearFrameBuffer(size_t frameIndex) |
| 353 { |
| 354 if (m_demux && m_demuxState >= WEBP_DEMUX_PARSED_HEADER && m_frameBufferCach
e[frameIndex].status() == ImageFrame::FramePartial) { |
| 355 // Clear the decoder state so that image can be decoded again when reque
sted. |
| 356 clearDecoder(); |
| 357 } |
| 358 ImageDecoder::clearFrameBuffer(frameIndex); |
| 359 } |
| 360 |
| 361 #endif // WEBP_ICC_ANIMATION_SUPPORT |
| 362 |
118 #ifdef QCMS_WEBP_COLOR_CORRECTION | 363 #ifdef QCMS_WEBP_COLOR_CORRECTION |
119 | 364 |
120 void WEBPImageDecoder::createColorTransform(const char* data, size_t size) | 365 void WEBPImageDecoder::createColorTransform(const char* data, size_t size) |
121 { | 366 { |
122 if (m_transform) | 367 if (m_transform) |
123 qcms_transform_release(m_transform); | 368 qcms_transform_release(m_transform); |
124 m_transform = 0; | 369 m_transform = 0; |
125 | 370 |
126 qcms_profile* deviceProfile = ImageDecoder::qcmsOutputDeviceProfile(); | 371 qcms_profile* deviceProfile = ImageDecoder::qcmsOutputDeviceProfile(); |
127 if (!deviceProfile) | 372 if (!deviceProfile) |
128 return; | 373 return; |
129 qcms_profile* inputProfile = qcms_profile_from_memory(data, size); | 374 qcms_profile* inputProfile = qcms_profile_from_memory(data, size); |
130 if (!inputProfile) | 375 if (!inputProfile) |
131 return; | 376 return; |
132 | 377 |
133 // We currently only support color profiles for RGB profiled images. | 378 // We currently only support color profiles for RGB profiled images. |
134 ASSERT(icSigRgbData == qcms_profile_get_color_space(inputProfile)); | 379 ASSERT(icSigRgbData == qcms_profile_get_color_space(inputProfile)); |
135 // The input image pixels are RGBA format. | 380 // The input image pixels are RGBA format. |
136 qcms_data_type format = QCMS_DATA_RGBA_8; | 381 qcms_data_type format = QCMS_DATA_RGBA_8; |
137 // FIXME: Don't force perceptual intent if the image profile contains an int
ent. | 382 // FIXME: Don't force perceptual intent if the image profile contains an int
ent. |
138 m_transform = qcms_transform_create(inputProfile, format, deviceProfile, QCM
S_DATA_RGBA_8, QCMS_INTENT_PERCEPTUAL); | 383 m_transform = qcms_transform_create(inputProfile, format, deviceProfile, QCM
S_DATA_RGBA_8, QCMS_INTENT_PERCEPTUAL); |
139 | 384 |
140 qcms_profile_release(inputProfile); | 385 qcms_profile_release(inputProfile); |
141 } | 386 } |
142 | 387 |
143 void WEBPImageDecoder::readColorProfile(const uint8_t* data, size_t size) | 388 void WEBPImageDecoder::readColorProfile() |
144 { | 389 { |
145 WebPChunkIterator chunkIterator; | 390 WebPChunkIterator chunkIterator; |
146 WebPData inputData = { data, size }; | 391 if (!WebPDemuxGetChunk(m_demux, "ICCP", 1, &chunkIterator)) { |
147 WebPDemuxState state; | |
148 | |
149 WebPDemuxer* demuxer = WebPDemuxPartial(&inputData, &state); | |
150 if (!WebPDemuxGetChunk(demuxer, "ICCP", 1, &chunkIterator)) { | |
151 WebPDemuxReleaseChunkIterator(&chunkIterator); | 392 WebPDemuxReleaseChunkIterator(&chunkIterator); |
152 WebPDemuxDelete(demuxer); | |
153 return; | 393 return; |
154 } | 394 } |
155 | 395 |
156 const char* profileData = reinterpret_cast<const char*>(chunkIterator.chunk.
bytes); | 396 const char* profileData = reinterpret_cast<const char*>(chunkIterator.chunk.
bytes); |
157 size_t profileSize = chunkIterator.chunk.size; | 397 size_t profileSize = chunkIterator.chunk.size; |
158 | 398 |
159 // Only accept RGB color profiles from input class devices. | 399 // Only accept RGB color profiles from input class devices. |
160 bool ignoreProfile = false; | 400 bool ignoreProfile = false; |
161 if (profileSize < ImageDecoder::iccColorProfileHeaderLength) | 401 if (profileSize < ImageDecoder::iccColorProfileHeaderLength) |
162 ignoreProfile = true; | 402 ignoreProfile = true; |
163 else if (!ImageDecoder::rgbColorProfile(profileData, profileSize)) | 403 else if (!ImageDecoder::rgbColorProfile(profileData, profileSize)) |
164 ignoreProfile = true; | 404 ignoreProfile = true; |
165 else if (!ImageDecoder::inputDeviceColorProfile(profileData, profileSize)) | 405 else if (!ImageDecoder::inputDeviceColorProfile(profileData, profileSize)) |
166 ignoreProfile = true; | 406 ignoreProfile = true; |
167 | 407 |
168 if (!ignoreProfile) | 408 if (!ignoreProfile) |
169 createColorTransform(profileData, profileSize); | 409 createColorTransform(profileData, profileSize); |
170 | 410 |
171 WebPDemuxReleaseChunkIterator(&chunkIterator); | 411 WebPDemuxReleaseChunkIterator(&chunkIterator); |
172 WebPDemuxDelete(demuxer); | |
173 } | 412 } |
174 | 413 |
175 void WEBPImageDecoder::applyColorProfile(const uint8_t* data, size_t dataSize, I
mageFrame& buffer) | 414 #endif // QCMS_WEBP_COLOR_CORRECTION |
| 415 |
| 416 #ifdef WEBP_ICC_ANIMATION_SUPPORT |
| 417 void WEBPImageDecoder::applyPostProcessing(size_t frameIndex) |
176 { | 418 { |
| 419 ImageFrame& buffer = m_frameBufferCache[frameIndex]; |
177 int width; | 420 int width; |
178 int decodedHeight; | 421 int decodedHeight; |
179 if (!WebPIDecGetRGB(m_decoder, &decodedHeight, &width, 0, 0)) | 422 if (!WebPIDecGetRGB(m_decoder, &decodedHeight, &width, 0, 0)) |
180 return; // See also https://bugs.webkit.org/show_bug.cgi?id=74062 | 423 return; // See also https://bugs.webkit.org/show_bug.cgi?id=74062 |
181 if (decodedHeight <= 0) | 424 if (decodedHeight <= 0) |
182 return; | 425 return; |
183 | 426 |
184 if (!m_haveReadProfile) { | 427 const IntRect& frameRect = buffer.originalFrameRect(); |
185 readColorProfile(data, dataSize); | 428 ASSERT_WITH_SECURITY_IMPLICATION(width == frameRect.width()); |
186 m_haveReadProfile = true; | 429 ASSERT_WITH_SECURITY_IMPLICATION(decodedHeight <= frameRect.height()); |
| 430 const int left = frameRect.x(); |
| 431 const int top = frameRect.y(); |
| 432 |
| 433 #ifdef QCMS_WEBP_COLOR_CORRECTION |
| 434 if ((m_formatFlags & ICCP_FLAG) && !ignoresGammaAndColorProfile()) { |
| 435 if (!m_haveReadProfile) { |
| 436 readColorProfile(); |
| 437 m_haveReadProfile = true; |
| 438 } |
| 439 for (int y = m_decodedHeight; y < decodedHeight; ++y) { |
| 440 const int canvasY = top + y; |
| 441 uint8_t* row = reinterpret_cast<uint8_t*>(buffer.getAddr(left, canva
sY)); |
| 442 if (qcms_transform* transform = colorTransform()) |
| 443 qcms_transform_data_type(transform, row, row, width, QCMS_OUTPUT
_RGBX); |
| 444 uint8_t* pixel = row; |
| 445 for (int x = 0; x < width; ++x, pixel += 4) { |
| 446 const int canvasX = left + x; |
| 447 buffer.setRGBA(canvasX, canvasY, pixel[0], pixel[1], pixel[2], p
ixel[3]); |
| 448 } |
| 449 } |
187 } | 450 } |
| 451 #endif // QCMS_WEBP_COLOR_CORRECTION |
188 | 452 |
189 ASSERT(width == size().width()); | 453 // During the decoding of current frame, we may have set some pixels to be t
ransparent (i.e. alpha < 255). |
190 ASSERT(decodedHeight <= size().height()); | 454 // However, the value of each of these pixels should have been determined by
blending it against the value |
191 | 455 // of that pixel in the previous frame. So, we correct these pixels based on
disposal method of the previous |
192 for (int y = m_decodedHeight; y < decodedHeight; ++y) { | 456 // frame and the previous frame buffer. |
193 uint8_t* row = reinterpret_cast<uint8_t*>(buffer.getAddr(0, y)); | 457 if ((m_formatFlags & ANIMATION_FLAG) && frameIndex) { |
194 if (qcms_transform* transform = colorTransform()) | 458 ImageFrame& prevBuffer = m_frameBufferCache[frameIndex - 1]; |
195 qcms_transform_data_type(transform, row, row, width, QCMS_OUTPUT_RGB
X); | 459 ImageFrame::FrameDisposalMethod prevMethod = prevBuffer.disposalMethod()
; |
196 uint8_t* pixel = row; | 460 if (prevMethod == ImageFrame::DisposeKeep) { // Restore transparent pixe
ls to pixels in previous canvas. |
197 for (int x = 0; x < width; ++x, pixel += 4) | 461 ASSERT(prevBuffer.status() == ImageFrame::FrameComplete); // Because
required previous frames are always decoded first. |
198 buffer.setRGBA(x, y, pixel[0], pixel[1], pixel[2], pixel[3]); | 462 for (int y = m_decodedHeight; y < decodedHeight; ++y) { |
| 463 const int canvasY = top + y; |
| 464 for (int x = 0; x < width; ++x) { |
| 465 const int canvasX = left + x; |
| 466 ImageFrame::PixelData& pixel = *buffer.getAddr(canvasX, canv
asY); |
| 467 // FIXME: Use alpha-blending when alpha is between 0 and 255
. |
| 468 // Alpha-blending is being implemented in: https://bugs.webk
it.org/show_bug.cgi?id=17022 |
| 469 if (!((pixel >> SK_A32_SHIFT) & 0xff)) { // Need to restore. |
| 470 ImageFrame::PixelData prevPixel = *prevBuffer.getAddr(ca
nvasX, canvasY); |
| 471 pixel = prevPixel; |
| 472 } |
| 473 } |
| 474 } |
| 475 } else if (prevMethod == ImageFrame::DisposeOverwriteBgcolor && buffer.r
equiredPreviousFrameIndex() != notFound) { |
| 476 // Note: if the requiredPreviousFrameIndex is |notFound|, there's no
thing to do. |
| 477 const IntRect& prevRect = prevBuffer.originalFrameRect(); |
| 478 // We need to restore transparent pixels to as they were just after
initFrame() call. That is: |
| 479 // * Transparent if it belongs to prevRect <-- This is a no-op. |
| 480 // * Pixel in the previous canvas otherwise <-- Need to restore. |
| 481 for (int y = m_decodedHeight; y < decodedHeight; ++y) { |
| 482 const int canvasY = top + y; |
| 483 for (int x = 0; x < width; ++x) { |
| 484 const int canvasX = left + x; |
| 485 ImageFrame::PixelData& pixel = *buffer.getAddr(canvasX, canv
asY); |
| 486 // FIXME: Use alpha-blending when alpha is between 0 and 255
. |
| 487 if (!((pixel >> SK_A32_SHIFT) & 0xff) && !prevRect.contains(
IntPoint(canvasX, canvasY))) { // Need to restore. |
| 488 ImageFrame::PixelData prevPixel = *prevBuffer.getAddr(ca
nvasX, canvasY); |
| 489 pixel = prevPixel; |
| 490 } |
| 491 } |
| 492 } |
| 493 } |
199 } | 494 } |
200 | 495 |
201 m_decodedHeight = decodedHeight; | 496 m_decodedHeight = decodedHeight; |
202 } | 497 } |
| 498 #endif // WEBP_ICC_ANIMATION_SUPPORT |
203 | 499 |
204 #endif // QCMS_WEBP_COLOR_CORRECTION | 500 bool WEBPImageDecoder::decode(const uint8_t* dataBytes, size_t dataSize, bool on
lySize, size_t frameIndex) |
205 | |
206 bool WEBPImageDecoder::decode(bool onlySize) | |
207 { | 501 { |
208 if (failed()) | 502 if (failed()) |
209 return false; | 503 return false; |
210 | 504 |
211 const uint8_t* dataBytes = reinterpret_cast<const uint8_t*>(m_data->data()); | |
212 const size_t dataSize = m_data->size(); | |
213 | |
214 if (!ImageDecoder::isSizeAvailable()) { | 505 if (!ImageDecoder::isSizeAvailable()) { |
215 static const size_t imageHeaderSize = 30; | 506 static const size_t imageHeaderSize = 30; |
216 if (dataSize < imageHeaderSize) | 507 if (dataSize < imageHeaderSize) |
217 return false; | 508 return false; |
218 int width, height; | 509 int width, height; |
219 #ifdef QCMS_WEBP_COLOR_CORRECTION | 510 #if (WEBP_DECODER_ABI_VERSION >= 0x0163) |
220 WebPData inputData = { dataBytes, dataSize }; | |
221 WebPDemuxState state; | |
222 WebPDemuxer* demuxer = WebPDemuxPartial(&inputData, &state); | |
223 if (!demuxer) | |
224 return setFailed(); | |
225 | |
226 width = WebPDemuxGetI(demuxer, WEBP_FF_CANVAS_WIDTH); | |
227 height = WebPDemuxGetI(demuxer, WEBP_FF_CANVAS_HEIGHT); | |
228 m_formatFlags = WebPDemuxGetI(demuxer, WEBP_FF_FORMAT_FLAGS); | |
229 m_hasAlpha = !!(m_formatFlags & ALPHA_FLAG); | |
230 | |
231 WebPDemuxDelete(demuxer); | |
232 if (state <= WEBP_DEMUX_PARSING_HEADER) | |
233 return false; | |
234 #elif (WEBP_DECODER_ABI_VERSION >= 0x0163) | |
235 WebPBitstreamFeatures features; | 511 WebPBitstreamFeatures features; |
236 if (WebPGetFeatures(dataBytes, dataSize, &features) != VP8_STATUS_OK) | 512 if (WebPGetFeatures(dataBytes, dataSize, &features) != VP8_STATUS_OK) |
237 return setFailed(); | 513 return setFailed(); |
238 width = features.width; | 514 width = features.width; |
239 height = features.height; | 515 height = features.height; |
240 m_hasAlpha = features.has_alpha; | 516 m_formatFlags = features.has_alpha ? ALPHA_FLAG : 0; |
241 #else | 517 #else |
242 // Earlier version won't be able to display WebP files with alpha. | 518 // Earlier version won't be able to display WebP files with alpha. |
243 if (!WebPGetInfo(dataBytes, dataSize, &width, &height)) | 519 if (!WebPGetInfo(dataBytes, dataSize, &width, &height)) |
244 return setFailed(); | 520 return setFailed(); |
245 m_hasAlpha = false; | |
246 #endif | 521 #endif |
247 if (!setSize(width, height)) | 522 if (!setSize(width, height)) |
248 return setFailed(); | 523 return setFailed(); |
249 } | 524 } |
250 | 525 |
251 ASSERT(ImageDecoder::isSizeAvailable()); | 526 ASSERT(ImageDecoder::isSizeAvailable()); |
252 if (onlySize) | 527 if (onlySize) |
253 return true; | 528 return true; |
254 | 529 |
255 ASSERT(!m_frameBufferCache.isEmpty()); | 530 ASSERT(m_frameBufferCache.size() > frameIndex); |
256 ImageFrame& buffer = m_frameBufferCache[0]; | 531 ImageFrame& buffer = m_frameBufferCache[frameIndex]; |
257 ASSERT(buffer.status() != ImageFrame::FrameComplete); | 532 ASSERT(buffer.status() != ImageFrame::FrameComplete); |
258 | 533 |
259 if (buffer.status() == ImageFrame::FrameEmpty) { | 534 if (buffer.status() == ImageFrame::FrameEmpty) { |
260 if (!buffer.setSize(size().width(), size().height())) | 535 if (!buffer.setSize(size().width(), size().height())) |
261 return setFailed(); | 536 return setFailed(); |
262 buffer.setStatus(ImageFrame::FramePartial); | 537 buffer.setStatus(ImageFrame::FramePartial); |
263 buffer.setHasAlpha(m_hasAlpha); | 538 // The buffer is transparent outside the decoded area while the image is
loading. |
| 539 // The correct value of 'hasAlpha' for the frame will be set when it is
fully decoded. |
| 540 buffer.setHasAlpha(true); |
264 buffer.setOriginalFrameRect(IntRect(IntPoint(), size())); | 541 buffer.setOriginalFrameRect(IntRect(IntPoint(), size())); |
265 } | 542 } |
266 | 543 |
| 544 const IntRect& frameRect = buffer.originalFrameRect(); |
267 if (!m_decoder) { | 545 if (!m_decoder) { |
268 WEBP_CSP_MODE mode = outputMode(m_hasAlpha); | 546 WEBP_CSP_MODE mode = outputMode(m_formatFlags & ALPHA_FLAG); |
269 if (!m_premultiplyAlpha) | 547 if (!m_premultiplyAlpha) |
270 mode = outputMode(false); | 548 mode = outputMode(false); |
| 549 #ifdef QCMS_WEBP_COLOR_CORRECTION |
271 if ((m_formatFlags & ICCP_FLAG) && !ignoresGammaAndColorProfile()) | 550 if ((m_formatFlags & ICCP_FLAG) && !ignoresGammaAndColorProfile()) |
272 mode = MODE_RGBA; // Decode to RGBA for input to libqcms. | 551 mode = MODE_RGBA; // Decode to RGBA for input to libqcms. |
| 552 #endif |
| 553 WebPInitDecBuffer(&m_decoderBuffer); |
273 m_decoderBuffer.colorspace = mode; | 554 m_decoderBuffer.colorspace = mode; |
274 m_decoderBuffer.u.RGBA.stride = size().width() * sizeof(ImageFrame::Pixe
lData); | 555 m_decoderBuffer.u.RGBA.stride = size().width() * sizeof(ImageFrame::Pixe
lData); |
275 m_decoderBuffer.u.RGBA.size = m_decoderBuffer.u.RGBA.stride * size().hei
ght(); | 556 m_decoderBuffer.u.RGBA.size = m_decoderBuffer.u.RGBA.stride * frameRect.
height(); |
276 m_decoderBuffer.is_external_memory = 1; | 557 m_decoderBuffer.is_external_memory = 1; |
277 m_decoder = WebPINewDecoder(&m_decoderBuffer); | 558 m_decoder = WebPINewDecoder(&m_decoderBuffer); |
278 if (!m_decoder) | 559 if (!m_decoder) |
279 return setFailed(); | 560 return setFailed(); |
280 } | 561 } |
281 | 562 |
282 m_decoderBuffer.u.RGBA.rgba = reinterpret_cast<uint8_t*>(buffer.getAddr(0, 0
)); | 563 m_decoderBuffer.u.RGBA.rgba = reinterpret_cast<uint8_t*>(buffer.getAddr(fram
eRect.x(), frameRect.y())); |
283 | 564 |
284 switch (WebPIUpdate(m_decoder, dataBytes, dataSize)) { | 565 switch (WebPIUpdate(m_decoder, dataBytes, dataSize)) { |
285 case VP8_STATUS_OK: | 566 case VP8_STATUS_OK: |
286 if ((m_formatFlags & ICCP_FLAG) && !ignoresGammaAndColorProfile()) | 567 applyPostProcessing(frameIndex); |
287 applyColorProfile(dataBytes, dataSize, buffer); | 568 buffer.setHasAlpha(m_formatFlags & ALPHA_FLAG); |
288 buffer.setStatus(ImageFrame::FrameComplete); | 569 buffer.setStatus(ImageFrame::FrameComplete); |
289 clear(); | 570 clearDecoder(); |
290 return true; | 571 return true; |
291 case VP8_STATUS_SUSPENDED: | 572 case VP8_STATUS_SUSPENDED: |
292 if ((m_formatFlags & ICCP_FLAG) && !ignoresGammaAndColorProfile()) | 573 applyPostProcessing(frameIndex); |
293 applyColorProfile(dataBytes, dataSize, buffer); | |
294 return false; | 574 return false; |
295 default: | 575 default: |
296 clear(); | 576 clear(); |
297 return setFailed(); | 577 return setFailed(); |
298 } | 578 } |
299 } | 579 } |
300 | 580 |
301 } // namespace WebCore | 581 } // namespace WebCore |
OLD | NEW |