Chromium Code Reviews| OLD | NEW |
|---|---|
| 1 /* | 1 /* |
| 2 * Copyright (C) 2010 Google Inc. All rights reserved. | 2 * Copyright (C) 2010 Google Inc. All rights reserved. |
| 3 * | 3 * |
| 4 * Redistribution and use in source and binary forms, with or without | 4 * Redistribution and use in source and binary forms, with or without |
| 5 * modification, are permitted provided that the following conditions | 5 * modification, are permitted provided that the following conditions |
| 6 * are met: | 6 * are met: |
| 7 * | 7 * |
| 8 * 1. Redistributions of source code must retain the above copyright | 8 * 1. Redistributions of source code must retain the above copyright |
| 9 * notice, this list of conditions and the following disclaimer. | 9 * notice, this list of conditions and the following disclaimer. |
| 10 * 2. Redistributions in binary form must reproduce the above copyright | 10 * 2. Redistributions in binary form must reproduce the above copyright |
| (...skipping 15 matching lines...) Expand all Loading... | |
| 26 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | 26 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| 27 */ | 27 */ |
| 28 | 28 |
| 29 #include "config.h" | 29 #include "config.h" |
| 30 #include "WEBPImageDecoder.h" | 30 #include "WEBPImageDecoder.h" |
| 31 | 31 |
| 32 #include "PlatformInstrumentation.h" | 32 #include "PlatformInstrumentation.h" |
| 33 | 33 |
| 34 #ifdef QCMS_WEBP_COLOR_CORRECTION | 34 #ifdef QCMS_WEBP_COLOR_CORRECTION |
| 35 #include "qcms.h" | 35 #include "qcms.h" |
| 36 #include "webp/demux.h" | 36 #endif |
| 37 | |
| 38 #ifdef WEBP_ICC_ANIM_SUPPORT | |
| 39 #include "RuntimeEnabledFeatures.h" | |
| 40 #include "webp/format_constants.h" | |
| 37 #else | 41 #else |
| 38 #undef ICCP_FLAG | 42 #undef ICCP_FLAG |
| 39 #define ICCP_FLAG 0 | 43 #define ICCP_FLAG 0 |
| 40 #endif | 44 #undef ALPHA_FLAG |
| 45 #if (WEBP_DECODER_ABI_VERSION >= 0x0163) // Alpha supported, but need to define flag. | |
| 46 #define ALPHA_FLAG 0x00000010 | |
| 47 #else // Versions earlier than 0.1.99 don't support alpha. | |
| 48 #define ALPHA_FLAG 0 | |
| 49 #endif // WEBP_DECODER_ABI_VERSION >= 0x0163 | |
| 50 #endif // WEBP_ICC_ANIM_SUPPORT | |
| 41 | 51 |
| 42 // Backward emulation for earlier versions than 0.1.99. | 52 // Backward emulation for earlier versions than 0.1.99. |
| 43 #if (WEBP_DECODER_ABI_VERSION < 0x0163) | 53 #if (WEBP_DECODER_ABI_VERSION < 0x0163) |
| 44 #define MODE_rgbA MODE_RGBA | 54 #define MODE_rgbA MODE_RGBA |
| 45 #define MODE_bgrA MODE_BGRA | 55 #define MODE_bgrA MODE_BGRA |
| 46 #endif | 56 #endif |
| 47 | 57 |
| 48 #if CPU(BIG_ENDIAN) || CPU(MIDDLE_ENDIAN) | 58 #if CPU(BIG_ENDIAN) || CPU(MIDDLE_ENDIAN) |
| 49 inline WEBP_CSP_MODE outputMode(bool hasAlpha) { return hasAlpha ? MODE_rgbA : M ODE_RGBA; } | 59 inline WEBP_CSP_MODE outputMode(bool hasAlpha) { return hasAlpha ? MODE_rgbA : M ODE_RGBA; } |
| 50 #elif SK_B32_SHIFT | 60 #elif SK_B32_SHIFT |
| 51 inline WEBP_CSP_MODE outputMode(bool hasAlpha) { return hasAlpha ? MODE_rgbA : M ODE_RGBA; } | 61 inline WEBP_CSP_MODE outputMode(bool hasAlpha) { return hasAlpha ? MODE_rgbA : M ODE_RGBA; } |
| 52 #else // LITTLE_ENDIAN, output BGRA pixels. | 62 #else // LITTLE_ENDIAN, output BGRA pixels. |
| 53 inline WEBP_CSP_MODE outputMode(bool hasAlpha) { return hasAlpha ? MODE_bgrA : M ODE_BGRA; } | 63 inline WEBP_CSP_MODE outputMode(bool hasAlpha) { return hasAlpha ? MODE_bgrA : M ODE_BGRA; } |
| 54 #endif | 64 #endif |
| 55 | 65 |
| 56 namespace WebCore { | 66 namespace WebCore { |
| 57 | 67 |
| 58 WEBPImageDecoder::WEBPImageDecoder(ImageSource::AlphaOption alphaOption, | 68 WEBPImageDecoder::WEBPImageDecoder(ImageSource::AlphaOption alphaOption, |
| 59 ImageSource::GammaAndColorProfileOption gamma AndColorProfileOption) | 69 ImageSource::GammaAndColorProfileOption gamma AndColorProfileOption) |
| 60 : ImageDecoder(alphaOption, gammaAndColorProfileOption) | 70 : ImageDecoder(alphaOption, gammaAndColorProfileOption) |
| 61 , m_decoder(0) | 71 , m_decoder(0) |
| 62 , m_hasAlpha(false) | |
| 63 , m_formatFlags(0) | 72 , m_formatFlags(0) |
| 64 #ifdef QCMS_WEBP_COLOR_CORRECTION | 73 #ifdef QCMS_WEBP_COLOR_CORRECTION |
| 65 , m_haveReadProfile(false) | 74 , m_haveReadProfile(false) |
| 66 , m_transform(0) | 75 , m_transform(0) |
| 76 #endif | |
| 77 #ifdef WEBP_ICC_ANIM_SUPPORT | |
| 67 , m_decodedHeight(0) | 78 , m_decodedHeight(0) |
| 68 #endif | 79 , m_haveAlreadyParsedThisData(false) |
| 69 { | 80 , m_demux(0) |
| 70 WebPInitDecBuffer(&m_decoderBuffer); | 81 , m_demuxState(WEBP_DEMUX_PARSING_HEADER) |
| 82 , m_haveReadAnimParams(false) | |
| 83 , m_repetitionCount(cAnimationLoopOnce) | |
| 84 #endif | |
| 85 { | |
| 71 } | 86 } |
| 72 | 87 |
| 73 WEBPImageDecoder::~WEBPImageDecoder() | 88 WEBPImageDecoder::~WEBPImageDecoder() |
| 74 { | 89 { |
| 75 clear(); | 90 clearAll(); |
| 76 } | 91 } |
| 77 | 92 |
| 78 void WEBPImageDecoder::clear() | 93 void WEBPImageDecoder::clearAll() |
| 79 { | 94 { |
| 80 #ifdef QCMS_WEBP_COLOR_CORRECTION | 95 #ifdef QCMS_WEBP_COLOR_CORRECTION |
| 81 if (m_transform) | 96 if (m_transform) |
| 82 qcms_transform_release(m_transform); | 97 qcms_transform_release(m_transform); |
| 83 m_transform = 0; | 98 m_transform = 0; |
| 84 #endif | 99 #endif |
| 85 WebPFreeDecBuffer(&m_decoderBuffer); | 100 #ifdef WEBP_ICC_ANIM_SUPPORT |
| 86 if (m_decoder) | 101 WebPDemuxDelete(m_demux); |
| 87 WebPIDelete(m_decoder); | 102 m_demux = 0; |
| 103 #endif | |
| 104 clearDecoder(); | |
| 105 } | |
| 106 | |
| 107 void WEBPImageDecoder::clearDecoder() | |
| 108 { | |
| 109 WebPIDelete(m_decoder); | |
| 88 m_decoder = 0; | 110 m_decoder = 0; |
| 111 #ifdef WEBP_ICC_ANIM_SUPPORT | |
| 112 m_decodedHeight = 0; | |
| 113 #endif | |
| 89 } | 114 } |
| 90 | 115 |
| 91 bool WEBPImageDecoder::isSizeAvailable() | 116 bool WEBPImageDecoder::isSizeAvailable() |
| 92 { | 117 { |
| 93 if (!ImageDecoder::isSizeAvailable()) | 118 if (!ImageDecoder::isSizeAvailable()) { |
| 94 decode(true); | 119 #ifdef WEBP_ICC_ANIM_SUPPORT |
| 95 | 120 if (!updateDemuxer()) |
| 121 return 0; | |
| 122 #else | |
| 123 decode(reinterpret_cast<const uint8_t*>(m_data->data()), m_data->size(), true, 0); | |
| 124 #endif | |
| 125 } | |
| 96 return ImageDecoder::isSizeAvailable(); | 126 return ImageDecoder::isSizeAvailable(); |
| 97 } | 127 } |
| 98 | 128 |
| 99 ImageFrame* WEBPImageDecoder::frameBufferAtIndex(size_t index) | 129 size_t WEBPImageDecoder::frameCount() |
| 100 { | 130 { |
| 101 if (index) | 131 #ifdef WEBP_ICC_ANIM_SUPPORT |
| 132 if (!updateDemuxer()) | |
| 102 return 0; | 133 return 0; |
| 103 | 134 #else |
| 104 if (m_frameBufferCache.isEmpty()) { | 135 if (m_frameBufferCache.isEmpty()) { |
| 105 m_frameBufferCache.resize(1); | 136 m_frameBufferCache.resize(1); |
| 106 m_frameBufferCache[0].setPremultiplyAlpha(m_premultiplyAlpha); | 137 m_frameBufferCache[0].setPremultiplyAlpha(m_premultiplyAlpha); |
| 107 } | 138 } |
| 108 | 139 #endif |
| 109 ImageFrame& frame = m_frameBufferCache[0]; | 140 return m_frameBufferCache.size(); |
| 141 } | |
| 142 | |
| 143 ImageFrame* WEBPImageDecoder::frameBufferAtIndex(size_t index) | |
| 144 { | |
| 145 if (index >= frameCount()) | |
| 146 return 0; | |
| 147 | |
| 148 ImageFrame& frame = m_frameBufferCache[index]; | |
| 110 if (frame.status() != ImageFrame::FrameComplete) { | 149 if (frame.status() != ImageFrame::FrameComplete) { |
| 111 PlatformInstrumentation::willDecodeImage("WEBP"); | 150 #ifdef WEBP_ICC_ANIM_SUPPORT |
| 112 decode(false); | 151 if (RuntimeEnabledFeatures::animatedWebPEnabled()) { |
| 113 PlatformInstrumentation::didDecodeImage(); | 152 if (index && (m_frameBufferCache[index - 1].status() != ImageFrame:: FrameComplete)) |
| 153 return 0; // We haven't fully decoded the previous frame yet. | |
| 154 ASSERT(m_demux); | |
| 155 WebPIterator fIter; | |
| 156 if (!WebPDemuxGetFrame(m_demux, index + 1, &fIter)) | |
| 157 return 0; | |
| 158 if (m_formatFlags & ANIMATION_FLAG) { | |
| 159 if (!initFrameBuffer(fIter, index)) | |
| 160 return 0; | |
| 161 } | |
| 162 PlatformInstrumentation::willDecodeImage("WEBP"); | |
| 163 decode(fIter.fragment.bytes, fIter.fragment.size, false, index); | |
| 164 PlatformInstrumentation::didDecodeImage(); | |
| 165 WebPDemuxReleaseIterator(&fIter); | |
| 166 } else { | |
| 167 #endif | |
| 168 ASSERT(!index); | |
| 169 PlatformInstrumentation::willDecodeImage("WEBP"); | |
| 170 decode(reinterpret_cast<const uint8_t*>(m_data->data()), m_data->siz e(), false, index); | |
| 171 PlatformInstrumentation::didDecodeImage(); | |
| 172 #ifdef WEBP_ICC_ANIM_SUPPORT | |
| 173 } | |
| 174 #endif | |
| 114 } | 175 } |
| 115 return &frame; | 176 return &frame; |
| 116 } | 177 } |
| 117 | 178 |
| 179 #ifdef WEBP_ICC_ANIM_SUPPORT | |
| 180 | |
| 181 void WEBPImageDecoder::setData(SharedBuffer* data, bool allDataReceived) | |
| 182 { | |
| 183 if (failed()) | |
| 184 return; | |
| 185 | |
| 186 ImageDecoder::setData(data, allDataReceived); | |
| 187 | |
| 188 // Mark that we have new data. | |
| 189 if (m_demuxState != WEBP_DEMUX_DONE) | |
| 190 m_haveAlreadyParsedThisData = false; | |
| 191 } | |
| 192 | |
| 193 bool WEBPImageDecoder::updateDemuxer() | |
| 194 { | |
| 195 if (!m_haveAlreadyParsedThisData) { | |
| 196 WebPDemuxDelete(m_demux); | |
| 197 const uint8_t* dataBytes = reinterpret_cast<const uint8_t*>(m_data->data ()); | |
| 198 const size_t dataSize = m_data->size(); | |
| 199 | |
| 200 static const size_t minSizeForDemux = RIFF_HEADER_SIZE + CHUNK_HEADER_SI ZE; | |
| 201 if (dataSize < minSizeForDemux) | |
| 202 return 0; // Wait for headers so that WebPDemuxPartial doesn't retur n null. | |
| 203 | |
| 204 WebPData inputData = { dataBytes, dataSize }; | |
| 205 m_demux = WebPDemuxPartial(&inputData, &m_demuxState); | |
| 206 if (!m_demux) | |
| 207 return setFailed(); // Must be a failure as we have at least 'minSiz eForDemux' bytes. | |
| 208 if (m_demuxState >= WEBP_DEMUX_PARSED_HEADER) { | |
| 209 if (!ImageDecoder::isSizeAvailable()) { | |
| 210 if (!setSize(WebPDemuxGetI(m_demux, WEBP_FF_CANVAS_WIDTH), WebPD emuxGetI(m_demux, WEBP_FF_CANVAS_HEIGHT))) | |
| 211 setFailed(); | |
| 212 m_formatFlags = WebPDemuxGetI(m_demux, WEBP_FF_FORMAT_FLAGS); | |
| 213 } | |
| 214 ASSERT(ImageDecoder::isSizeAvailable()); | |
| 215 const bool hasAnimation = (m_formatFlags & ANIMATION_FLAG); | |
| 216 const size_t newFrameCount = WebPDemuxGetI(m_demux, WEBP_FF_FRAME_CO UNT); | |
| 217 if (RuntimeEnabledFeatures::animatedWebPEnabled() && hasAnimation && !m_haveReadAnimParams && (newFrameCount >= 1)) { | |
| 218 // As we have parsed at least one frame (even if partially), | |
| 219 // we must already have parsed the animation properties. | |
| 220 // This is because ANIM chunk always precedes ANMF chunks. | |
| 221 const uint32_t loopCount = WebPDemuxGetI(m_demux, WEBP_FF_LOOP_C OUNT); | |
| 222 // Note: The following casts an 'unsigned int' to 'int'. But tha t is fine, because loop count is always <= 16 bits. | |
| 223 m_repetitionCount = (!loopCount) ? cAnimationLoopInfinite : loop Count; | |
| 224 m_haveReadAnimParams = true; | |
| 225 } | |
| 226 if (newFrameCount > m_frameBufferCache.size()) { | |
| 227 m_frameBufferCache.resize(newFrameCount); | |
| 228 for (size_t i = 0; i < newFrameCount; ++i) | |
| 229 m_frameBufferCache[i].setPremultiplyAlpha(m_premultiplyAlpha ); | |
| 230 } | |
| 231 } | |
| 232 m_haveAlreadyParsedThisData = true; | |
| 233 } | |
| 234 return true; | |
| 235 } | |
| 236 | |
| 237 bool WEBPImageDecoder::initFrameBuffer(const WebPIterator& fIter, size_t frameIn dex) | |
| 238 { | |
| 239 ImageFrame& buffer = m_frameBufferCache[frameIndex]; | |
| 240 if (buffer.status() != ImageFrame::FrameEmpty) // Already initialized. | |
| 241 return true; | |
| 242 | |
| 243 // Initialize the frame rect in our buffer. | |
| 244 IntRect frameRect(fIter.x_offset, fIter.y_offset, fIter.width, fIter.height) ; | |
| 245 | |
| 246 // Make sure the frameRect doesn't extend outside the buffer. | |
| 247 if (frameRect.maxX() > size().width()) | |
| 248 frameRect.setWidth(size().width() - fIter.x_offset); | |
|
Alpha Left Google
2013/04/25 22:07:12
What if frameRect.x() is always greater then size(
urvang (Google)
2013/04/26 01:46:23
Note that I kept the initFrameBuffer() method here
| |
| 249 if (frameRect.maxY() > size().height()) | |
| 250 frameRect.setHeight(size().height() - fIter.y_offset); | |
|
Alpha Left Google
2013/04/25 22:07:12
Same here.
urvang (Google)
2013/04/26 01:46:23
same commment.
| |
| 251 | |
| 252 const int left = upperBoundScaledX(frameRect.x()); | |
| 253 const int right = lowerBoundScaledX(frameRect.maxX(), left); | |
| 254 const int top = upperBoundScaledY(frameRect.y()); | |
| 255 const int bottom = lowerBoundScaledY(frameRect.maxY(), top); | |
| 256 buffer.setOriginalFrameRect(IntRect(left, top, right - left, bottom - top)); | |
| 257 | |
| 258 buffer.setDisposalMethod(fIter.dispose_method == WEBP_MUX_DISPOSE_BACKGROUND ? ImageFrame::DisposeOverwriteBgcolor : ImageFrame::DisposeKeep); | |
| 259 buffer.setDuration(fIter.duration); | |
| 260 buffer.setHasAlpha(m_formatFlags & ALPHA_FLAG); | |
| 261 | |
| 262 if (!frameIndex) { | |
| 263 // This is the first frame, so we're not relying on any previous data. | |
| 264 if (!buffer.setSize(scaledSize().width(), scaledSize().height())) | |
| 265 return setFailed(); | |
| 266 } else { | |
| 267 // The starting state for this frame depends on the previous frame's | |
| 268 // disposal method. | |
| 269 const ImageFrame& prevBuffer = m_frameBufferCache[frameIndex - 1]; | |
| 270 ASSERT(prevBuffer.status() == ImageFrame::FrameComplete); | |
| 271 const IntRect& prevRect = prevBuffer.originalFrameRect(); | |
| 272 const ImageFrame::FrameDisposalMethod prevMethod = prevBuffer.disposalMe thod(); | |
| 273 if ((prevMethod == ImageFrame::DisposeKeep) || (prevMethod == ImageFrame ::DisposeNotSpecified)) { | |
| 274 // Preserve the last frame as the starting state for this frame. | |
| 275 if (!buffer.copyBitmapData(prevBuffer)) | |
| 276 return setFailed(); | |
| 277 } else { // prevMethod == ImageFrame::DisposeOverwriteBgcolor | |
| 278 // We want to clear the previous frame to transparent, without | |
| 279 // affecting pixels in the image outside of the frame. | |
| 280 // So, we copy the whole previous buffer, then clear just its frame. | |
| 281 if (!frameIndex || prevRect.contains(IntRect(IntPoint(), scaledSize( )))) { | |
| 282 // Clearing the first frame, or a frame the size of the whole | |
| 283 // image, results in a completely empty image. | |
| 284 if (!buffer.setSize(scaledSize().width(), scaledSize().height()) ) | |
| 285 return setFailed(); | |
| 286 } else { | |
| 287 // Copy the whole previous buffer, then clear just its frame. | |
| 288 if (!buffer.copyBitmapData(prevBuffer)) | |
| 289 return setFailed(); | |
| 290 for (int y = prevRect.y(); y < prevRect.maxY(); ++y) { | |
| 291 for (int x = prevRect.x(); x < prevRect.maxX(); ++x) | |
| 292 buffer.setRGBA(x, y, 0, 0, 0, 0); | |
| 293 } | |
| 294 } | |
| 295 } | |
| 296 } | |
| 297 // Update frame status to be partially complete. | |
| 298 buffer.setStatus(ImageFrame::FramePartial); | |
| 299 return true; | |
| 300 } | |
| 301 void WEBPImageDecoder::clearFrameBufferCache(size_t clearBeforeFrame) | |
| 302 { | |
| 303 // We always preserve at least one frame. | |
| 304 if (m_frameBufferCache.size() <= 1) | |
| 305 return; | |
| 306 | |
| 307 // Find the last frame we need to preserve in the cache to facilitate | |
| 308 // the construction of next frames (needed by initFrame() and | |
| 309 // applyPostProcessing()) . This frame is either: | |
| 310 // * The last decoded frame in cache, OR | |
| 311 // * The first frame (if cache doesn't contain any decoded frames). | |
| 312 const int lastFrame = std::min(clearBeforeFrame, m_frameBufferCache.size() - 1); | |
| 313 Vector<ImageFrame>::iterator i(m_frameBufferCache.begin() + lastFrame); | |
| 314 while ((i != m_frameBufferCache.begin()) && (i->status() != ImageFrame::Fram eComplete)) | |
| 315 --i; | |
| 316 | |
| 317 // Now |i| holds the last frame we need to preserve; clear prior frames. | |
| 318 for (Vector<ImageFrame>::iterator j(m_frameBufferCache.begin()); j != i; ++j ) { | |
| 319 ASSERT(j->status() != ImageFrame::FramePartial); | |
| 320 if (j->status() != ImageFrame::FrameEmpty) | |
| 321 j->clearPixelData(); | |
| 322 } | |
| 323 } | |
| 324 | |
| 325 #endif // WEBP_ICC_ANIM_SUPPORT | |
| 326 | |
| 118 #ifdef QCMS_WEBP_COLOR_CORRECTION | 327 #ifdef QCMS_WEBP_COLOR_CORRECTION |
| 119 | 328 |
| 120 void WEBPImageDecoder::createColorTransform(const char* data, size_t size) | 329 void WEBPImageDecoder::createColorTransform(const char* data, size_t size) |
| 121 { | 330 { |
| 122 if (m_transform) | 331 if (m_transform) |
| 123 qcms_transform_release(m_transform); | 332 qcms_transform_release(m_transform); |
| 124 m_transform = 0; | 333 m_transform = 0; |
| 125 | 334 |
| 126 qcms_profile* deviceProfile = ImageDecoder::qcmsOutputDeviceProfile(); | 335 qcms_profile* deviceProfile = ImageDecoder::qcmsOutputDeviceProfile(); |
| 127 if (!deviceProfile) | 336 if (!deviceProfile) |
| 128 return; | 337 return; |
| 129 qcms_profile* inputProfile = qcms_profile_from_memory(data, size); | 338 qcms_profile* inputProfile = qcms_profile_from_memory(data, size); |
| 130 if (!inputProfile) | 339 if (!inputProfile) |
| 131 return; | 340 return; |
| 132 | 341 |
| 133 // We currently only support color profiles for RGB profiled images. | 342 // We currently only support color profiles for RGB profiled images. |
| 134 ASSERT(icSigRgbData == qcms_profile_get_color_space(inputProfile)); | 343 ASSERT(icSigRgbData == qcms_profile_get_color_space(inputProfile)); |
| 135 // The input image pixels are RGBA format. | 344 // The input image pixels are RGBA format. |
| 136 qcms_data_type format = QCMS_DATA_RGBA_8; | 345 qcms_data_type format = QCMS_DATA_RGBA_8; |
| 137 // FIXME: Don't force perceptual intent if the image profile contains an int ent. | 346 // FIXME: Don't force perceptual intent if the image profile contains an int ent. |
| 138 m_transform = qcms_transform_create(inputProfile, format, deviceProfile, QCM S_DATA_RGBA_8, QCMS_INTENT_PERCEPTUAL); | 347 m_transform = qcms_transform_create(inputProfile, format, deviceProfile, QCM S_DATA_RGBA_8, QCMS_INTENT_PERCEPTUAL); |
| 139 | 348 |
| 140 qcms_profile_release(inputProfile); | 349 qcms_profile_release(inputProfile); |
| 141 } | 350 } |
| 142 | 351 |
| 143 void WEBPImageDecoder::readColorProfile(const uint8_t* data, size_t size) | 352 void WEBPImageDecoder::readColorProfile() |
| 144 { | 353 { |
| 145 WebPChunkIterator chunkIterator; | 354 WebPChunkIterator chunkIterator; |
| 146 WebPData inputData = { data, size }; | 355 if (!WebPDemuxGetChunk(m_demux, "ICCP", 1, &chunkIterator)) { |
| 147 WebPDemuxState state; | |
| 148 | |
| 149 WebPDemuxer* demuxer = WebPDemuxPartial(&inputData, &state); | |
| 150 if (!WebPDemuxGetChunk(demuxer, "ICCP", 1, &chunkIterator)) { | |
| 151 WebPDemuxReleaseChunkIterator(&chunkIterator); | 356 WebPDemuxReleaseChunkIterator(&chunkIterator); |
| 152 WebPDemuxDelete(demuxer); | |
| 153 return; | 357 return; |
| 154 } | 358 } |
| 155 | 359 |
| 156 const char* profileData = reinterpret_cast<const char*>(chunkIterator.chunk. bytes); | 360 const char* profileData = reinterpret_cast<const char*>(chunkIterator.chunk. bytes); |
| 157 size_t profileSize = chunkIterator.chunk.size; | 361 size_t profileSize = chunkIterator.chunk.size; |
| 158 | 362 |
| 159 // Only accept RGB color profiles from input class devices. | 363 // Only accept RGB color profiles from input class devices. |
| 160 bool ignoreProfile = false; | 364 bool ignoreProfile = false; |
| 161 if (profileSize < ImageDecoder::iccColorProfileHeaderLength) | 365 if (profileSize < ImageDecoder::iccColorProfileHeaderLength) |
| 162 ignoreProfile = true; | 366 ignoreProfile = true; |
| 163 else if (!ImageDecoder::rgbColorProfile(profileData, profileSize)) | 367 else if (!ImageDecoder::rgbColorProfile(profileData, profileSize)) |
| 164 ignoreProfile = true; | 368 ignoreProfile = true; |
| 165 else if (!ImageDecoder::inputDeviceColorProfile(profileData, profileSize)) | 369 else if (!ImageDecoder::inputDeviceColorProfile(profileData, profileSize)) |
| 166 ignoreProfile = true; | 370 ignoreProfile = true; |
| 167 | 371 |
| 168 if (!ignoreProfile) | 372 if (!ignoreProfile) |
| 169 createColorTransform(profileData, profileSize); | 373 createColorTransform(profileData, profileSize); |
| 170 | 374 |
| 171 WebPDemuxReleaseChunkIterator(&chunkIterator); | 375 WebPDemuxReleaseChunkIterator(&chunkIterator); |
| 172 WebPDemuxDelete(demuxer); | |
| 173 } | 376 } |
| 174 | 377 |
| 175 void WEBPImageDecoder::applyColorProfile(const uint8_t* data, size_t size, Image Frame& buffer) | 378 #endif // QCMS_WEBP_COLOR_CORRECTION |
| 379 | |
| 380 #ifdef WEBP_ICC_ANIM_SUPPORT | |
| 381 void WEBPImageDecoder::applyPostProcessing(size_t frameIndex) | |
| 176 { | 382 { |
| 383 ImageFrame& buffer = m_frameBufferCache[frameIndex]; | |
| 177 int width; | 384 int width; |
| 385 int stride; | |
| 178 int decodedHeight; | 386 int decodedHeight; |
| 179 if (!WebPIDecGetRGB(m_decoder, &decodedHeight, &width, 0, 0)) | 387 if (!WebPIDecGetRGB(m_decoder, &decodedHeight, &width, 0, &stride)) |
| 180 return; // See also https://bugs.webkit.org/show_bug.cgi?id=74062 | 388 return; // See also https://bugs.webkit.org/show_bug.cgi?id=74062 |
| 181 if (decodedHeight <= 0) | 389 if (decodedHeight <= 0) |
| 182 return; | 390 return; |
| 391 ASSERT(width == scaledSize().width()); | |
| 392 ASSERT(decodedHeight <= scaledSize().height()); | |
| 393 const int left = buffer.originalFrameRect().x(); | |
| 394 const int top = buffer.originalFrameRect().y(); | |
| 183 | 395 |
| 184 if (!m_haveReadProfile) { | 396 // Color Profile. |
| 185 readColorProfile(data, size); | 397 if ((m_formatFlags & ICCP_FLAG) && !ignoresGammaAndColorProfile()) { |
| 186 m_haveReadProfile = true; | 398 #ifdef QCMS_WEBP_COLOR_CORRECTION |
| 399 if (!m_haveReadProfile) { | |
| 400 readColorProfile(); | |
| 401 m_haveReadProfile = true; | |
| 402 } | |
| 403 #endif // QCMS_WEBP_COLOR_CORRECTION | |
| 404 for (int y = m_decodedHeight; y < decodedHeight; ++y) { | |
| 405 const int canvasY = top + y; | |
| 406 uint8_t* row = reinterpret_cast<uint8_t*>(buffer.getAddr(left, canva sY)); | |
| 407 #ifdef QCMS_WEBP_COLOR_CORRECTION | |
| 408 if (qcms_transform* transform = colorTransform()) | |
| 409 qcms_transform_data_type(transform, row, row, width, QCMS_OUTPUT _RGBX); | |
| 410 #endif // QCMS_WEBP_COLOR_CORRECTION | |
| 411 uint8_t* pixel = row; | |
| 412 for (int x = 0; x < width; ++x, pixel += 4) { | |
| 413 const int canvasX = left + x; | |
| 414 buffer.setRGBA(canvasX, canvasY, pixel[0], pixel[1], pixel[2], p ixel[3]); | |
| 415 } | |
| 416 } | |
| 187 } | 417 } |
| 188 | 418 |
| 189 ASSERT(width == scaledSize().width()); | 419 // Frame disposal. |
| 190 ASSERT(decodedHeight <= scaledSize().height()); | 420 if ((m_formatFlags & ANIMATION_FLAG) && frameIndex) { |
| 191 | 421 ImageFrame& prevBuffer = m_frameBufferCache[frameIndex - 1]; |
| 192 for (int y = m_decodedHeight; y < decodedHeight; ++y) { | 422 ImageFrame::FrameDisposalMethod prevMethod = prevBuffer.disposalMethod() ; |
| 193 uint8_t* row = reinterpret_cast<uint8_t*>(buffer.getAddr(0, y)); | 423 ASSERT(prevBuffer.status() == ImageFrame::FrameComplete); |
| 194 if (qcms_transform* transform = colorTransform()) | 424 if (prevMethod == ImageFrame::DisposeKeep) { // Restore transparent pixe ls to pixels in previous canvas. |
| 195 qcms_transform_data_type(transform, row, row, width, QCMS_OUTPUT_RGB X); | 425 for (int y = m_decodedHeight; y < decodedHeight; ++y) { |
| 196 uint8_t* pixel = row; | 426 const int canvasY = top + y; |
| 197 for (int x = 0; x < width; ++x, pixel += 4) | 427 for (int x = 0; x < width; ++x) { |
| 198 buffer.setRGBA(x, y, pixel[0], pixel[1], pixel[2], pixel[3]); | 428 const int canvasX = left + x; |
| 429 ImageFrame::PixelData& pixel = *buffer.getAddr(canvasX, canv asY); | |
| 430 // FIXME: Use alpha-blending when alpha is between 0 and 255 . | |
| 431 // Alpha-blending is being implemented in: https://bugs.webk it.org/show_bug.cgi?id=17022 | |
| 432 if (!((pixel >> 24) & 0xff)) { // Need to restore. | |
| 433 const ImageFrame::PixelData prevPixel = *prevBuffer.getA ddr(canvasX, canvasY); | |
| 434 pixel = prevPixel; | |
| 435 } | |
| 436 } | |
| 437 } | |
| 438 } else if (prevMethod == ImageFrame::DisposeOverwriteBgcolor) { | |
| 439 const IntRect& prevRect = prevBuffer.originalFrameRect(); | |
| 440 // We need to restore transparent pixels to as they were just after initFrame() call. That is: | |
|
Alpha Left Google
2013/04/25 22:07:12
This comment is confusing. Can you clarify? For ex
urvang (Google)
2013/04/26 01:46:23
You are right. I added a comment above to explain
| |
| 441 // * Transparent if it belongs to prevRect <-- This is a no-op. | |
| 442 // * Pixel in the previous canvas otherwise <-- Need to restore. | |
| 443 for (int y = m_decodedHeight; y < decodedHeight; ++y) { | |
| 444 const int canvasY = top + y; | |
| 445 for (int x = 0; x < width; ++x) { | |
| 446 const int canvasX = left + x; | |
| 447 ImageFrame::PixelData& pixel = *buffer.getAddr(canvasX, canv asY); | |
| 448 const ImageFrame::PixelData prevPixel = *prevBuffer.getAddr( canvasX, canvasY); | |
| 449 if (!((pixel >> 24) & 0xff) && !prevRect.contains(IntPoint(c anvasX, canvasY))) // Need to restore. | |
|
Alpha Left Google
2013/04/25 22:07:12
Does this have alpha blending issue as well? Pleas
urvang (Google)
2013/04/26 01:46:23
Done.
| |
| 450 pixel = prevPixel; | |
| 451 } | |
| 452 } | |
| 453 } | |
| 199 } | 454 } |
| 200 | 455 |
| 201 m_decodedHeight = decodedHeight; | 456 m_decodedHeight = decodedHeight; |
| 202 } | 457 } |
| 458 #endif // WEBP_ICC_ANIM_SUPPORT | |
| 203 | 459 |
| 204 #endif // QCMS_WEBP_COLOR_CORRECTION | 460 bool WEBPImageDecoder::decode(const uint8_t* dataBytes, size_t dataSize, bool on lySize, size_t frameIndex) |
| 205 | |
| 206 bool WEBPImageDecoder::decode(bool onlySize) | |
| 207 { | 461 { |
| 208 if (failed()) | 462 if (failed()) |
| 209 return false; | 463 return false; |
| 210 | 464 |
| 211 const uint8_t* dataBytes = reinterpret_cast<const uint8_t*>(m_data->data()); | |
| 212 const size_t dataSize = m_data->size(); | |
| 213 | |
| 214 if (!ImageDecoder::isSizeAvailable()) { | 465 if (!ImageDecoder::isSizeAvailable()) { |
| 215 static const size_t imageHeaderSize = 30; | 466 static const size_t imageHeaderSize = 30; |
| 216 if (dataSize < imageHeaderSize) | 467 if (dataSize < imageHeaderSize) |
| 217 return false; | 468 return false; |
| 218 int width, height; | 469 int width, height; |
| 219 #ifdef QCMS_WEBP_COLOR_CORRECTION | 470 #if (WEBP_DECODER_ABI_VERSION >= 0x0163) |
| 220 WebPData inputData = { dataBytes, dataSize }; | |
| 221 WebPDemuxState state; | |
| 222 WebPDemuxer* demuxer = WebPDemuxPartial(&inputData, &state); | |
| 223 if (!demuxer) | |
| 224 return setFailed(); | |
| 225 | |
| 226 width = WebPDemuxGetI(demuxer, WEBP_FF_CANVAS_WIDTH); | |
| 227 height = WebPDemuxGetI(demuxer, WEBP_FF_CANVAS_HEIGHT); | |
| 228 m_formatFlags = WebPDemuxGetI(demuxer, WEBP_FF_FORMAT_FLAGS); | |
| 229 m_hasAlpha = !!(m_formatFlags & ALPHA_FLAG); | |
| 230 | |
| 231 WebPDemuxDelete(demuxer); | |
| 232 if (state <= WEBP_DEMUX_PARSING_HEADER) | |
| 233 return false; | |
| 234 #elif (WEBP_DECODER_ABI_VERSION >= 0x0163) | |
| 235 WebPBitstreamFeatures features; | 471 WebPBitstreamFeatures features; |
| 236 if (WebPGetFeatures(dataBytes, dataSize, &features) != VP8_STATUS_OK) | 472 if (WebPGetFeatures(dataBytes, dataSize, &features) != VP8_STATUS_OK) |
| 237 return setFailed(); | 473 return setFailed(); |
| 238 width = features.width; | 474 width = features.width; |
| 239 height = features.height; | 475 height = features.height; |
| 240 m_hasAlpha = features.has_alpha; | 476 m_formatFlags = features.has_alpha ? ALPHA_FLAG : 0; |
| 241 #else | 477 #else |
| 242 // Earlier version won't be able to display WebP files with alpha. | 478 // Earlier version won't be able to display WebP files with alpha. |
| 243 if (!WebPGetInfo(dataBytes, dataSize, &width, &height)) | 479 if (!WebPGetInfo(dataBytes, dataSize, &width, &height)) |
| 244 return setFailed(); | 480 return setFailed(); |
| 245 m_hasAlpha = false; | |
| 246 #endif | 481 #endif |
| 247 if (!setSize(width, height)) | 482 if (!setSize(width, height)) |
| 248 return setFailed(); | 483 return setFailed(); |
| 249 } | 484 } |
| 250 | 485 |
| 251 ASSERT(ImageDecoder::isSizeAvailable()); | 486 ASSERT(ImageDecoder::isSizeAvailable()); |
| 252 if (onlySize) | 487 if (onlySize) |
| 253 return true; | 488 return true; |
| 254 | 489 |
| 255 ASSERT(!m_frameBufferCache.isEmpty()); | 490 ASSERT(m_frameBufferCache.size() > frameIndex); |
| 256 ImageFrame& buffer = m_frameBufferCache[0]; | 491 ImageFrame& buffer = m_frameBufferCache[frameIndex]; |
| 257 ASSERT(buffer.status() != ImageFrame::FrameComplete); | 492 ASSERT(buffer.status() != ImageFrame::FrameComplete); |
| 258 | 493 |
| 259 if (buffer.status() == ImageFrame::FrameEmpty) { | 494 if (buffer.status() == ImageFrame::FrameEmpty) { |
| 260 if (!buffer.setSize(size().width(), size().height())) | 495 if (!buffer.setSize(scaledSize().width(), scaledSize().height())) |
| 261 return setFailed(); | 496 return setFailed(); |
| 262 buffer.setStatus(ImageFrame::FramePartial); | 497 buffer.setStatus(ImageFrame::FramePartial); |
| 263 buffer.setHasAlpha(m_hasAlpha); | 498 buffer.setHasAlpha(m_formatFlags & ALPHA_FLAG); |
| 264 buffer.setOriginalFrameRect(IntRect(IntPoint(), size())); | 499 buffer.setOriginalFrameRect(IntRect(IntPoint(), size())); |
| 265 } | 500 } |
| 266 | 501 |
| 502 const IntRect& frameRect = buffer.originalFrameRect(); | |
| 267 if (!m_decoder) { | 503 if (!m_decoder) { |
| 268 WEBP_CSP_MODE mode = outputMode(m_hasAlpha); | 504 WEBP_CSP_MODE mode = outputMode(m_formatFlags & ALPHA_FLAG); |
| 269 if (!m_premultiplyAlpha) | 505 if (!m_premultiplyAlpha) |
| 270 mode = outputMode(false); | 506 mode = outputMode(false); |
| 271 if ((m_formatFlags & ICCP_FLAG) && !ignoresGammaAndColorProfile()) | 507 if ((m_formatFlags & ICCP_FLAG) && !ignoresGammaAndColorProfile()) |
| 272 mode = MODE_RGBA; // Decode to RGBA for input to libqcms. | 508 mode = MODE_RGBA; // Decode to RGBA for input to libqcms. |
| 509 WebPInitDecBuffer(&m_decoderBuffer); | |
| 273 m_decoderBuffer.colorspace = mode; | 510 m_decoderBuffer.colorspace = mode; |
| 274 m_decoderBuffer.u.RGBA.stride = size().width() * sizeof(ImageFrame::Pixe lData); | 511 m_decoderBuffer.u.RGBA.stride = size().width() * sizeof(ImageFrame::Pixe lData); |
| 275 m_decoderBuffer.u.RGBA.size = m_decoderBuffer.u.RGBA.stride * size().hei ght(); | 512 m_decoderBuffer.u.RGBA.size = m_decoderBuffer.u.RGBA.stride * frameRect. height(); |
| 276 m_decoderBuffer.is_external_memory = 1; | 513 m_decoderBuffer.is_external_memory = 1; |
| 277 m_decoder = WebPINewDecoder(&m_decoderBuffer); | 514 m_decoder = WebPINewDecoder(&m_decoderBuffer); |
| 278 if (!m_decoder) | 515 if (!m_decoder) |
| 279 return setFailed(); | 516 return setFailed(); |
| 280 } | 517 } |
| 281 | 518 |
| 282 m_decoderBuffer.u.RGBA.rgba = reinterpret_cast<uint8_t*>(buffer.getAddr(0, 0 )); | 519 m_decoderBuffer.u.RGBA.rgba = reinterpret_cast<uint8_t*>(buffer.getAddr(fram eRect.x(), frameRect.y())); |
| 283 | 520 |
| 284 switch (WebPIUpdate(m_decoder, dataBytes, dataSize)) { | 521 switch (WebPIUpdate(m_decoder, dataBytes, dataSize)) { |
| 285 case VP8_STATUS_OK: | 522 case VP8_STATUS_OK: |
| 286 if ((m_formatFlags & ICCP_FLAG) && !ignoresGammaAndColorProfile()) | 523 applyPostProcessing(frameIndex); |
| 287 applyColorProfile(dataBytes, dataSize, buffer); | |
| 288 buffer.setStatus(ImageFrame::FrameComplete); | 524 buffer.setStatus(ImageFrame::FrameComplete); |
| 289 clear(); | 525 clearDecoder(); |
| 290 return true; | 526 return true; |
| 291 case VP8_STATUS_SUSPENDED: | 527 case VP8_STATUS_SUSPENDED: |
| 292 if ((m_formatFlags & ICCP_FLAG) && !ignoresGammaAndColorProfile()) | 528 applyPostProcessing(frameIndex); |
| 293 applyColorProfile(dataBytes, dataSize, buffer); | |
| 294 return false; | 529 return false; |
| 295 default: | 530 default: |
| 296 clear(); | 531 clearAll(); |
| 297 return setFailed(); | 532 return setFailed(); |
| 298 } | 533 } |
| 299 } | 534 } |
| 300 | 535 |
| 301 } // namespace WebCore | 536 } // namespace WebCore |
| OLD | NEW |