Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(407)

Unified Diff: Source/core/platform/image-decoders/webp/WEBPImageDecoder.cpp

Issue 13980003: Add animation support for WebP images (Closed) Base URL: https://chromium.googlesource.com/chromium/blink.git@master
Patch Set: Created 7 years, 6 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
Index: Source/core/platform/image-decoders/webp/WEBPImageDecoder.cpp
diff --git a/Source/core/platform/image-decoders/webp/WEBPImageDecoder.cpp b/Source/core/platform/image-decoders/webp/WEBPImageDecoder.cpp
index 26171f55dae281d37ebcb21ce8f08c8b957750ed..92bce94d5bfef5ee8874e3566734d38e6211bb68 100644
--- a/Source/core/platform/image-decoders/webp/WEBPImageDecoder.cpp
+++ b/Source/core/platform/image-decoders/webp/WEBPImageDecoder.cpp
@@ -33,16 +33,21 @@
#ifdef QCMS_WEBP_COLOR_CORRECTION
#include "qcms.h"
-#include "webp/demux.h"
-#else
-#undef ICCP_FLAG
-#define ICCP_FLAG 0
#endif
-// Backward emulation for earlier versions than 0.1.99.
+#ifdef WEBP_ICC_ANIMATION_SUPPORT
+#include "RuntimeEnabledFeatures.h"
+#include "webp/format_constants.h"
+#endif
+
#if (WEBP_DECODER_ABI_VERSION < 0x0163)
+// Backward emulation for versions earlier than 0.1.99.
#define MODE_rgbA MODE_RGBA
#define MODE_bgrA MODE_BGRA
+#define ALPHA_FLAG 0
+#elif (WEBP_DECODER_ABI_VERSION <= 0x0200)
+// Backward emulation for versions earlier than 0.3.0.
+#define ALPHA_FLAG 0x000010
#endif
#if CPU(BIG_ENDIAN) || CPU(MIDDLE_ENDIAN)
@@ -59,15 +64,20 @@ WEBPImageDecoder::WEBPImageDecoder(ImageSource::AlphaOption alphaOption,
ImageSource::GammaAndColorProfileOption gammaAndColorProfileOption)
: ImageDecoder(alphaOption, gammaAndColorProfileOption)
, m_decoder(0)
- , m_hasAlpha(false)
, m_formatFlags(0)
#ifdef QCMS_WEBP_COLOR_CORRECTION
, m_haveReadProfile(false)
, m_transform(0)
+#endif
+#ifdef WEBP_ICC_ANIMATION_SUPPORT
+ , m_demux(0)
+ , m_demuxState(WEBP_DEMUX_PARSING_HEADER)
+ , m_haveAlreadyParsedThisData(false)
+ , m_haveReadAnimationParameters(false)
+ , m_repetitionCount(cAnimationLoopOnce)
, m_decodedHeight(0)
#endif
{
- WebPInitDecBuffer(&m_decoderBuffer);
}
WEBPImageDecoder::~WEBPImageDecoder()
@@ -82,39 +92,275 @@ void WEBPImageDecoder::clear()
qcms_transform_release(m_transform);
m_transform = 0;
#endif
- WebPFreeDecBuffer(&m_decoderBuffer);
- if (m_decoder)
- WebPIDelete(m_decoder);
+#ifdef WEBP_ICC_ANIMATION_SUPPORT
+ WebPDemuxDelete(m_demux);
+ m_demux = 0;
+#endif
+ clearDecoder();
+}
+
+void WEBPImageDecoder::clearDecoder()
+{
+ WebPIDelete(m_decoder);
m_decoder = 0;
+#ifdef WEBP_ICC_ANIMATION_SUPPORT
+ m_decodedHeight = 0;
+#endif
}
bool WEBPImageDecoder::isSizeAvailable()
{
- if (!ImageDecoder::isSizeAvailable())
- decode(true);
-
+ if (!ImageDecoder::isSizeAvailable()) {
+#ifdef WEBP_ICC_ANIMATION_SUPPORT
+ updateDemuxer();
+#else
+ decode(reinterpret_cast<const uint8_t*>(m_data->data()), m_data->size(), true, 0);
+#endif
+ }
return ImageDecoder::isSizeAvailable();
}
-ImageFrame* WEBPImageDecoder::frameBufferAtIndex(size_t index)
+size_t WEBPImageDecoder::frameCount()
{
- if (index)
+#ifdef WEBP_ICC_ANIMATION_SUPPORT
+ if (!updateDemuxer())
return 0;
-
+#else
if (m_frameBufferCache.isEmpty()) {
m_frameBufferCache.resize(1);
m_frameBufferCache[0].setPremultiplyAlpha(m_premultiplyAlpha);
}
+#endif
+ return m_frameBufferCache.size();
+}
- ImageFrame& frame = m_frameBufferCache[0];
- if (frame.status() != ImageFrame::FrameComplete) {
- PlatformInstrumentation::willDecodeImage("WEBP");
- decode(false);
- PlatformInstrumentation::didDecodeImage();
+ImageFrame* WEBPImageDecoder::frameBufferAtIndex(size_t index)
+{
+ if (index >= frameCount())
+ return 0;
+
+ ImageFrame& frame = m_frameBufferCache[index];
+ if (frame.status() == ImageFrame::FrameComplete)
+ return &frame;
+
+#ifdef WEBP_ICC_ANIMATION_SUPPORT
+ if (RuntimeEnabledFeatures::animatedWebPEnabled()) {
+ Vector<size_t> framesToDecode;
+ size_t frameToDecode = index;
+ do {
+ framesToDecode.append(frameToDecode);
+ frameToDecode = m_frameBufferCache[frameToDecode].requiredPreviousFrameIndex();
Noel Gordon 2013/06/13 15:50:17 In general, you should have qinmin to look over al
urvang (Google) 2013/06/14 01:13:13 I requested him to review the same (he has just co
+ } while (frameToDecode != notFound && m_frameBufferCache[frameToDecode].status() != ImageFrame::FrameComplete);
+
+ ASSERT(m_demux);
+ // The |rend| variable is needed by some compilers that can't correctly
+ // select from const and non-const versions of overloaded functions.
+ // Can remove the variable if Android compiler can compile
+ // 'iter != framesToDecode.rend()'.
+ Vector<size_t>::const_reverse_iterator rend = framesToDecode.rend();
+ for (Vector<size_t>::const_reverse_iterator iter = framesToDecode.rbegin(); iter != rend; ++iter) {
Noel Gordon 2013/06/13 15:50:17 156-161: I wonder if a reverse index over the vect
Xianzhu 2013/06/13 20:04:19 Urvang, when you change this, please also change t
urvang (Google) 2013/06/14 01:13:13 Oh ya, I was wondering the same (had just copied t
urvang (Google) 2013/06/14 01:13:13 Done.
+ size_t frameIndex = *iter;
+ WebPIterator webpFrame;
+ if (!WebPDemuxGetFrame(m_demux, frameIndex + 1, &webpFrame))
+ return 0;
+ if ((m_formatFlags & ANIMATION_FLAG) && !initFrameBuffer(webpFrame, frameIndex)) {
+ WebPDemuxReleaseIterator(&webpFrame);
+ return 0;
+ }
+ PlatformInstrumentation::willDecodeImage("WEBP");
+ decode(webpFrame.fragment.bytes, webpFrame.fragment.size, false, frameIndex);
+ PlatformInstrumentation::didDecodeImage();
+ WebPDemuxReleaseIterator(&webpFrame);
+
+ // We need more data to continue decoding.
+ if (m_frameBufferCache[frameIndex].status() != ImageFrame::FrameComplete)
+ break;
+ }
+
+ // It is also a fatal error if all data is received and we have decoded all
+ // frames available but the file is truncated.
+ if (index >= m_frameBufferCache.size() - 1 && isAllDataReceived() && m_demux && m_demuxState != WEBP_DEMUX_DONE)
+ setFailed();
+
+ return &frame;
}
+#endif
+
+ ASSERT(!index);
+ PlatformInstrumentation::willDecodeImage("WEBP");
+ decode(reinterpret_cast<const uint8_t*>(m_data->data()), m_data->size(), false, index);
+ PlatformInstrumentation::didDecodeImage();
return &frame;
}
+#ifdef WEBP_ICC_ANIMATION_SUPPORT
+
+void WEBPImageDecoder::setData(SharedBuffer* data, bool allDataReceived)
+{
+ if (failed())
+ return;
+
+ ImageDecoder::setData(data, allDataReceived);
+
+ // Mark that we have new data.
+ if (m_demuxState != WEBP_DEMUX_DONE)
+ m_haveAlreadyParsedThisData = false;
+}
+
+int WEBPImageDecoder::repetitionCount() const
+{
+ return failed() ? cAnimationLoopOnce : m_repetitionCount;
+}
+
+bool WEBPImageDecoder::frameIsCompleteAtIndex(size_t index) const
+{
+ if (!RuntimeEnabledFeatures::animatedWebPEnabled())
+ return ImageDecoder::frameIsCompleteAtIndex(index);
+ if (!m_demux || m_demuxState <= WEBP_DEMUX_PARSING_HEADER)
+ return false;
+ if (!(m_formatFlags & ANIMATION_FLAG))
+ return ImageDecoder::frameIsCompleteAtIndex(index);
+ // In case of animation, number of complete frames is same as m_frameBufferCache.size() (see updateDemuxer()).
+ return index < m_frameBufferCache.size();
+}
+
+float WEBPImageDecoder::frameDurationAtIndex(size_t index) const
+{
+ return index < m_frameBufferCache.size() ? m_frameBufferCache[index].duration() : 0;
+}
+
+bool WEBPImageDecoder::updateDemuxer()
+{
+ if (m_haveAlreadyParsedThisData)
+ return true;
+
+ m_haveAlreadyParsedThisData = true;
+
+ static const size_t minSizeForDemux = RIFF_HEADER_SIZE + CHUNK_HEADER_SIZE;
+ if (m_data->size() < minSizeForDemux)
+ return false; // Wait for headers so that WebPDemuxPartial doesn't return null.
+
+ WebPDemuxDelete(m_demux);
+ WebPData inputData = { reinterpret_cast<const uint8_t*>(m_data->data()), m_data->size() };
+ m_demux = WebPDemuxPartial(&inputData, &m_demuxState);
+ if (!m_demux)
+ return setFailed(); // Must be a failure as we have at least 'minSizeForDemux' bytes.
+
+ if (m_demuxState <= WEBP_DEMUX_PARSING_HEADER)
+ return false; // Not enough data for parsing canvas width/height yet.
+
+ bool hasAnimation = (m_formatFlags & ANIMATION_FLAG);
+ if (!ImageDecoder::isSizeAvailable()) {
+ m_formatFlags = WebPDemuxGetI(m_demux, WEBP_FF_FORMAT_FLAGS);
+ hasAnimation = (m_formatFlags & ANIMATION_FLAG);
+ if (hasAnimation && !RuntimeEnabledFeatures::animatedWebPEnabled())
+ return setFailed();
+ if (!setSize(WebPDemuxGetI(m_demux, WEBP_FF_CANVAS_WIDTH), WebPDemuxGetI(m_demux, WEBP_FF_CANVAS_HEIGHT)))
+ return setFailed();
+ }
+ ASSERT(ImageDecoder::isSizeAvailable());
+ const size_t newFrameCount = WebPDemuxGetI(m_demux, WEBP_FF_FRAME_COUNT);
+ if (hasAnimation && !m_haveReadAnimationParameters && (newFrameCount >= 1)) {
+ // As we have parsed at least one frame (even if partially),
+ // we must already have parsed the animation properties.
+ // This is because ANIM chunk always precedes ANMF chunks.
+ m_repetitionCount = WebPDemuxGetI(m_demux, WEBP_FF_LOOP_COUNT);
+ ASSERT(m_repetitionCount == (m_repetitionCount & 0xffff)); // Loop count is always <= 16 bits.
+ if (!m_repetitionCount)
+ m_repetitionCount = cAnimationLoopInfinite;
+ m_haveReadAnimationParameters = true;
+ }
+ const size_t oldFrameCount = m_frameBufferCache.size();
+ if (newFrameCount > oldFrameCount) {
+ m_frameBufferCache.resize(newFrameCount);
+ for (size_t i = oldFrameCount; i < newFrameCount; ++i) {
+ m_frameBufferCache[i].setPremultiplyAlpha(m_premultiplyAlpha);
+ if (!hasAnimation)
+ continue;
+ WebPIterator animatedFrame;
+ WebPDemuxGetFrame(m_demux, i + 1, &animatedFrame);
+ ASSERT(animatedFrame.complete == 1); // Always true for animated case.
+ m_frameBufferCache[i].setDuration(animatedFrame.duration);
+ m_frameBufferCache[i].setDisposalMethod(animatedFrame.dispose_method == WEBP_MUX_DISPOSE_BACKGROUND ? ImageFrame::DisposeOverwriteBgcolor : ImageFrame::DisposeKeep);
+ WebPDemuxReleaseIterator(&animatedFrame);
+ m_frameBufferCache[i].setRequiredPreviousFrameIndex(findRequiredPreviousFrame(i));
+ }
+ }
+ return true;
+}
+
+bool WEBPImageDecoder::initFrameBuffer(const WebPIterator& frame, size_t frameIndex)
+{
+ ImageFrame& buffer = m_frameBufferCache[frameIndex];
+ if (buffer.status() != ImageFrame::FrameEmpty) // Already initialized.
+ return true;
+
+ // Initialize the frame rect in our buffer.
+ IntRect frameRect(frame.x_offset, frame.y_offset, frame.width, frame.height);
+
+ // Make sure the frameRect doesn't extend outside the buffer.
+ if (frameRect.maxX() > size().width())
+ frameRect.setWidth(size().width() - frame.x_offset);
+ if (frameRect.maxY() > size().height())
+ frameRect.setHeight(size().height() - frame.y_offset);
+ buffer.setOriginalFrameRect(frameRect);
+
+ const size_t requiredPreviousFrameIndex = findRequiredPreviousFrame(frameIndex);
Xianzhu 2013/06/13 20:04:19 Why not using buffer.requiredPreviousFrameIndex()
urvang (Google) 2013/06/14 01:13:13 Oh yes, thanks!
+ if (requiredPreviousFrameIndex == notFound) {
+ // This frame doesn't rely on any previous data.
+ if (!buffer.setSize(size().width(), size().height()))
+ return setFailed();
+ } else {
+ const ImageFrame& prevBuffer = m_frameBufferCache[requiredPreviousFrameIndex];
+ ASSERT(prevBuffer.status() == ImageFrame::FrameComplete);
+
+ // Preserve the last frame as the starting state for this frame.
+ if (!buffer.copyBitmapData(prevBuffer))
+ return setFailed();
+ }
+
Xianzhu 2013/06/13 20:04:19 For GIF, if the previous frame has DisposeOverwrit
urvang (Google) 2013/06/14 01:13:13 Oh yes, that's required here too (I did have the s
+ buffer.setStatus(ImageFrame::FramePartial);
+ // The buffer is transparent outside the decoded area while the image is loading.
+ // The correct value of 'hasAlpha' for the frame will be set when it is fully decoded.
+ buffer.setHasAlpha(true);
+ return true;
+}
+
+size_t WEBPImageDecoder::clearCacheExceptFrame(size_t clearExceptFrame)
Xianzhu 2013/06/13 20:04:19 This method seems the same as the ImageDecoder::cl
urvang (Google) 2013/06/14 01:13:13 There is one subtle difference, actually, due to a
Xianzhu 2013/06/14 17:50:27 Thanks for the explanation. I put clearCacheExcept
+{
+ // Don't clear if there are no frames or only one frame.
+ if (m_frameBufferCache.size() <= 1)
+ return 0;
+
+ // We need to preserve frames such that:
+ // 1. We don't clear |clearExceptFrame|;
+ // 2. We don't clear any frame from which a future initFrameBuffer() or
+ // applyPostProcessing() call could copy bitmap data.
+ // All other frames can be cleared.
+ while ((clearExceptFrame < m_frameBufferCache.size()) && (m_frameBufferCache[clearExceptFrame].status() != ImageFrame::FrameComplete))
+ clearExceptFrame = m_frameBufferCache[clearExceptFrame].requiredPreviousFrameIndex();
+
+ size_t frameBytesCleared = 0;
+ for (size_t i = 0; i < m_frameBufferCache.size(); ++i) {
+ if (i != clearExceptFrame) {
+ frameBytesCleared += frameBytesAtIndex(i);
+ clearFrameBuffer(i);
+ }
+ }
+ return frameBytesCleared;
+}
+
+void WEBPImageDecoder::clearFrameBuffer(size_t frameIndex)
+{
+ if (m_demux && m_demuxState >= WEBP_DEMUX_PARSED_HEADER && m_frameBufferCache[frameIndex].status() == ImageFrame::FramePartial) {
+ // Clear the decoder state so that image can be decoded again when requested.
Noel Gordon 2013/06/13 15:50:17 Sentences: "that image"? Did you mean to say "the
urvang (Google) 2013/06/14 01:13:13 "so that" is meant to be read together here. Tweak
+ clearDecoder();
+ }
+ ImageDecoder::clearFrameBuffer(frameIndex);
+}
+
+#endif // WEBP_ICC_ANIMATION_SUPPORT
+
#ifdef QCMS_WEBP_COLOR_CORRECTION
void WEBPImageDecoder::createColorTransform(const char* data, size_t size)
@@ -140,16 +386,11 @@ void WEBPImageDecoder::createColorTransform(const char* data, size_t size)
qcms_profile_release(inputProfile);
}
-void WEBPImageDecoder::readColorProfile(const uint8_t* data, size_t size)
+void WEBPImageDecoder::readColorProfile()
{
WebPChunkIterator chunkIterator;
- WebPData inputData = { data, size };
- WebPDemuxState state;
-
- WebPDemuxer* demuxer = WebPDemuxPartial(&inputData, &state);
- if (!WebPDemuxGetChunk(demuxer, "ICCP", 1, &chunkIterator)) {
+ if (!WebPDemuxGetChunk(m_demux, "ICCP", 1, &chunkIterator)) {
WebPDemuxReleaseChunkIterator(&chunkIterator);
- WebPDemuxDelete(demuxer);
return;
}
@@ -169,11 +410,14 @@ void WEBPImageDecoder::readColorProfile(const uint8_t* data, size_t size)
createColorTransform(profileData, profileSize);
WebPDemuxReleaseChunkIterator(&chunkIterator);
- WebPDemuxDelete(demuxer);
}
-void WEBPImageDecoder::applyColorProfile(const uint8_t* data, size_t dataSize, ImageFrame& buffer)
+#endif // QCMS_WEBP_COLOR_CORRECTION
+
+#ifdef WEBP_ICC_ANIMATION_SUPPORT
+void WEBPImageDecoder::applyPostProcessing(size_t frameIndex)
{
+ ImageFrame& buffer = m_frameBufferCache[frameIndex];
int width;
int decodedHeight;
if (!WebPIDecGetRGB(m_decoder, &decodedHeight, &width, 0, 0))
@@ -181,68 +425,100 @@ void WEBPImageDecoder::applyColorProfile(const uint8_t* data, size_t dataSize, I
if (decodedHeight <= 0)
return;
- if (!m_haveReadProfile) {
- readColorProfile(data, dataSize);
- m_haveReadProfile = true;
- }
+ const IntRect& frameRect = buffer.originalFrameRect();
+ ASSERT_WITH_SECURITY_IMPLICATION(width == frameRect.width());
+ ASSERT_WITH_SECURITY_IMPLICATION(decodedHeight <= frameRect.height());
+ const int left = frameRect.x();
+ const int top = frameRect.y();
- ASSERT(width == size().width());
- ASSERT(decodedHeight <= size().height());
+#ifdef QCMS_WEBP_COLOR_CORRECTION
+ if ((m_formatFlags & ICCP_FLAG) && !ignoresGammaAndColorProfile()) {
+ if (!m_haveReadProfile) {
+ readColorProfile();
+ m_haveReadProfile = true;
+ }
+ for (int y = m_decodedHeight; y < decodedHeight; ++y) {
+ const int canvasY = top + y;
+ uint8_t* row = reinterpret_cast<uint8_t*>(buffer.getAddr(left, canvasY));
+ if (qcms_transform* transform = colorTransform())
+ qcms_transform_data_type(transform, row, row, width, QCMS_OUTPUT_RGBX);
+ uint8_t* pixel = row;
+ for (int x = 0; x < width; ++x, pixel += 4) {
+ const int canvasX = left + x;
+ buffer.setRGBA(canvasX, canvasY, pixel[0], pixel[1], pixel[2], pixel[3]);
+ }
+ }
+ }
+#endif // QCMS_WEBP_COLOR_CORRECTION
- for (int y = m_decodedHeight; y < decodedHeight; ++y) {
- uint8_t* row = reinterpret_cast<uint8_t*>(buffer.getAddr(0, y));
- if (qcms_transform* transform = colorTransform())
- qcms_transform_data_type(transform, row, row, width, QCMS_OUTPUT_RGBX);
- uint8_t* pixel = row;
- for (int x = 0; x < width; ++x, pixel += 4)
- buffer.setRGBA(x, y, pixel[0], pixel[1], pixel[2], pixel[3]);
+ // During the decoding of current frame, we may have set some pixels to be transparent (i.e. alpha < 255).
+ // However, the value of each of these pixels should have been determined by blending it against the value
+ // of that pixel in the previous frame. So, we correct these pixels based on disposal method of the previous
+ // frame and the previous frame buffer.
+ if ((m_formatFlags & ANIMATION_FLAG) && frameIndex) {
+ ImageFrame& prevBuffer = m_frameBufferCache[frameIndex - 1];
+ ImageFrame::FrameDisposalMethod prevMethod = prevBuffer.disposalMethod();
+ if (prevMethod == ImageFrame::DisposeKeep) { // Restore transparent pixels to pixels in previous canvas.
+ ASSERT(prevBuffer.status() == ImageFrame::FrameComplete); // Because required previous frames are always decoded first.
+ for (int y = m_decodedHeight; y < decodedHeight; ++y) {
+ const int canvasY = top + y;
+ for (int x = 0; x < width; ++x) {
+ const int canvasX = left + x;
+ ImageFrame::PixelData& pixel = *buffer.getAddr(canvasX, canvasY);
+ // FIXME: Use alpha-blending when alpha is between 0 and 255.
+ // Alpha-blending is being implemented in: https://bugs.webkit.org/show_bug.cgi?id=17022
+ if (!((pixel >> SK_A32_SHIFT) & 0xff)) { // Need to restore.
+ ImageFrame::PixelData prevPixel = *prevBuffer.getAddr(canvasX, canvasY);
+ pixel = prevPixel;
+ }
+ }
+ }
+ } else if (prevMethod == ImageFrame::DisposeOverwriteBgcolor && buffer.requiredPreviousFrameIndex() != notFound) {
+ // Note: if the requiredPreviousFrameIndex is |notFound|, there's nothing to do.
Noel Gordon 2013/06/13 15:50:17 The previous frame buffer will be combined with th
urvang (Google) 2013/06/14 01:13:13 Indeed, added the ASSERT.
+ const IntRect& prevRect = prevBuffer.originalFrameRect();
+ // We need to restore transparent pixels to as they were just after initFrame() call. That is:
+ // * Transparent if it belongs to prevRect <-- This is a no-op.
+ // * Pixel in the previous canvas otherwise <-- Need to restore.
+ for (int y = m_decodedHeight; y < decodedHeight; ++y) {
+ const int canvasY = top + y;
+ for (int x = 0; x < width; ++x) {
+ const int canvasX = left + x;
+ ImageFrame::PixelData& pixel = *buffer.getAddr(canvasX, canvasY);
+ // FIXME: Use alpha-blending when alpha is between 0 and 255.
+ if (!((pixel >> SK_A32_SHIFT) & 0xff) && !prevRect.contains(IntPoint(canvasX, canvasY))) { // Need to restore.
+ ImageFrame::PixelData prevPixel = *prevBuffer.getAddr(canvasX, canvasY);
+ pixel = prevPixel;
+ }
+ }
+ }
+ }
}
m_decodedHeight = decodedHeight;
}
+#endif // WEBP_ICC_ANIMATION_SUPPORT
-#endif // QCMS_WEBP_COLOR_CORRECTION
-
-bool WEBPImageDecoder::decode(bool onlySize)
+bool WEBPImageDecoder::decode(const uint8_t* dataBytes, size_t dataSize, bool onlySize, size_t frameIndex)
{
if (failed())
return false;
- const uint8_t* dataBytes = reinterpret_cast<const uint8_t*>(m_data->data());
- const size_t dataSize = m_data->size();
-
if (!ImageDecoder::isSizeAvailable()) {
static const size_t imageHeaderSize = 30;
if (dataSize < imageHeaderSize)
return false;
int width, height;
-#ifdef QCMS_WEBP_COLOR_CORRECTION
- WebPData inputData = { dataBytes, dataSize };
- WebPDemuxState state;
- WebPDemuxer* demuxer = WebPDemuxPartial(&inputData, &state);
- if (!demuxer)
- return setFailed();
-
- width = WebPDemuxGetI(demuxer, WEBP_FF_CANVAS_WIDTH);
- height = WebPDemuxGetI(demuxer, WEBP_FF_CANVAS_HEIGHT);
- m_formatFlags = WebPDemuxGetI(demuxer, WEBP_FF_FORMAT_FLAGS);
- m_hasAlpha = !!(m_formatFlags & ALPHA_FLAG);
-
- WebPDemuxDelete(demuxer);
- if (state <= WEBP_DEMUX_PARSING_HEADER)
- return false;
-#elif (WEBP_DECODER_ABI_VERSION >= 0x0163)
+#if (WEBP_DECODER_ABI_VERSION >= 0x0163)
WebPBitstreamFeatures features;
if (WebPGetFeatures(dataBytes, dataSize, &features) != VP8_STATUS_OK)
return setFailed();
width = features.width;
height = features.height;
- m_hasAlpha = features.has_alpha;
+ m_formatFlags = features.has_alpha ? ALPHA_FLAG : 0;
#else
// Earlier version won't be able to display WebP files with alpha.
if (!WebPGetInfo(dataBytes, dataSize, &width, &height))
return setFailed();
- m_hasAlpha = false;
#endif
if (!setSize(width, height))
return setFailed();
@@ -252,48 +528,53 @@ bool WEBPImageDecoder::decode(bool onlySize)
if (onlySize)
return true;
- ASSERT(!m_frameBufferCache.isEmpty());
- ImageFrame& buffer = m_frameBufferCache[0];
+ ASSERT(m_frameBufferCache.size() > frameIndex);
+ ImageFrame& buffer = m_frameBufferCache[frameIndex];
ASSERT(buffer.status() != ImageFrame::FrameComplete);
if (buffer.status() == ImageFrame::FrameEmpty) {
if (!buffer.setSize(size().width(), size().height()))
return setFailed();
buffer.setStatus(ImageFrame::FramePartial);
- buffer.setHasAlpha(m_hasAlpha);
+ // The buffer is transparent outside the decoded area while the image is loading.
+ // The correct value of 'hasAlpha' for the frame will be set when it is fully decoded.
+ buffer.setHasAlpha(true);
buffer.setOriginalFrameRect(IntRect(IntPoint(), size()));
}
+ const IntRect& frameRect = buffer.originalFrameRect();
if (!m_decoder) {
- WEBP_CSP_MODE mode = outputMode(m_hasAlpha);
+ WEBP_CSP_MODE mode = outputMode(m_formatFlags & ALPHA_FLAG);
if (!m_premultiplyAlpha)
mode = outputMode(false);
+#ifdef QCMS_WEBP_COLOR_CORRECTION
if ((m_formatFlags & ICCP_FLAG) && !ignoresGammaAndColorProfile())
mode = MODE_RGBA; // Decode to RGBA for input to libqcms.
+#endif
+ WebPInitDecBuffer(&m_decoderBuffer);
m_decoderBuffer.colorspace = mode;
m_decoderBuffer.u.RGBA.stride = size().width() * sizeof(ImageFrame::PixelData);
- m_decoderBuffer.u.RGBA.size = m_decoderBuffer.u.RGBA.stride * size().height();
+ m_decoderBuffer.u.RGBA.size = m_decoderBuffer.u.RGBA.stride * frameRect.height();
m_decoderBuffer.is_external_memory = 1;
m_decoder = WebPINewDecoder(&m_decoderBuffer);
if (!m_decoder)
return setFailed();
}
- m_decoderBuffer.u.RGBA.rgba = reinterpret_cast<uint8_t*>(buffer.getAddr(0, 0));
+ m_decoderBuffer.u.RGBA.rgba = reinterpret_cast<uint8_t*>(buffer.getAddr(frameRect.x(), frameRect.y()));
switch (WebPIUpdate(m_decoder, dataBytes, dataSize)) {
case VP8_STATUS_OK:
- if ((m_formatFlags & ICCP_FLAG) && !ignoresGammaAndColorProfile())
- applyColorProfile(dataBytes, dataSize, buffer);
+ applyPostProcessing(frameIndex);
+ buffer.setHasAlpha(m_formatFlags & ALPHA_FLAG);
buffer.setStatus(ImageFrame::FrameComplete);
- clear();
+ clearDecoder();
return true;
case VP8_STATUS_SUSPENDED:
- if ((m_formatFlags & ICCP_FLAG) && !ignoresGammaAndColorProfile())
- applyColorProfile(dataBytes, dataSize, buffer);
+ applyPostProcessing(frameIndex);
return false;
default:
- clear();
+ clear();
return setFailed();
}
}

Powered by Google App Engine
This is Rietveld 408576698