| Index: cc/CCVideoLayerImpl.cpp
|
| diff --git a/cc/CCVideoLayerImpl.cpp b/cc/CCVideoLayerImpl.cpp
|
| deleted file mode 100644
|
| index d1fb41277cb5a4927b0c83acf0b137467aa0b845..0000000000000000000000000000000000000000
|
| --- a/cc/CCVideoLayerImpl.cpp
|
| +++ /dev/null
|
| @@ -1,391 +0,0 @@
|
| -// Copyright 2011 The Chromium Authors. All rights reserved.
|
| -// Use of this source code is governed by a BSD-style license that can be
|
| -// found in the LICENSE file.
|
| -
|
| -#include "config.h"
|
| -
|
| -#if USE(ACCELERATED_COMPOSITING)
|
| -
|
| -#include "CCVideoLayerImpl.h"
|
| -
|
| -#include "CCIOSurfaceDrawQuad.h"
|
| -#include "CCLayerTreeHostImpl.h"
|
| -#include "CCProxy.h"
|
| -#include "CCQuadSink.h"
|
| -#include "CCResourceProvider.h"
|
| -#include "CCStreamVideoDrawQuad.h"
|
| -#include "CCTextureDrawQuad.h"
|
| -#include "CCYUVVideoDrawQuad.h"
|
| -#include "Extensions3DChromium.h"
|
| -#include "GraphicsContext3D.h"
|
| -#include "NotImplemented.h"
|
| -#include <public/WebVideoFrame.h>
|
| -
|
| -namespace cc {
|
| -
|
| -CCVideoLayerImpl::CCVideoLayerImpl(int id, WebKit::WebVideoFrameProvider* provider)
|
| - : CCLayerImpl(id)
|
| - , m_provider(provider)
|
| - , m_frame(0)
|
| - , m_externalTextureResource(0)
|
| -{
|
| - // This matrix is the default transformation for stream textures, and flips on the Y axis.
|
| - m_streamTextureMatrix = WebKit::WebTransformationMatrix(
|
| - 1, 0, 0, 0,
|
| - 0, -1, 0, 0,
|
| - 0, 0, 1, 0,
|
| - 0, 1, 0, 1);
|
| -
|
| - // This only happens during a commit on the compositor thread while the main
|
| - // thread is blocked. That makes this a thread-safe call to set the video
|
| - // frame provider client that does not require a lock. The same is true of
|
| - // the call in the destructor.
|
| - ASSERT(CCProxy::isMainThreadBlocked());
|
| - m_provider->setVideoFrameProviderClient(this);
|
| -}
|
| -
|
| -CCVideoLayerImpl::~CCVideoLayerImpl()
|
| -{
|
| - // See comment in constructor for why this doesn't need a lock.
|
| - ASSERT(CCProxy::isMainThreadBlocked());
|
| - if (m_provider) {
|
| - m_provider->setVideoFrameProviderClient(0);
|
| - m_provider = 0;
|
| - }
|
| - freePlaneData(layerTreeHostImpl()->resourceProvider());
|
| -
|
| -#if !ASSERT_DISABLED
|
| - for (unsigned i = 0; i < WebKit::WebVideoFrame::maxPlanes; ++i)
|
| - ASSERT(!m_framePlanes[i].resourceId);
|
| - ASSERT(!m_externalTextureResource);
|
| -#endif
|
| -}
|
| -
|
| -void CCVideoLayerImpl::stopUsingProvider()
|
| -{
|
| - // Block the provider from shutting down until this client is done
|
| - // using the frame.
|
| - MutexLocker locker(m_providerMutex);
|
| - ASSERT(!m_frame);
|
| - m_provider = 0;
|
| -}
|
| -
|
| -// Convert WebKit::WebVideoFrame::Format to GraphicsContext3D's format enum values.
|
| -static GC3Denum convertVFCFormatToGC3DFormat(const WebKit::WebVideoFrame& frame)
|
| -{
|
| - switch (frame.format()) {
|
| - case WebKit::WebVideoFrame::FormatYV12:
|
| - case WebKit::WebVideoFrame::FormatYV16:
|
| - return GraphicsContext3D::LUMINANCE;
|
| - case WebKit::WebVideoFrame::FormatNativeTexture:
|
| - return frame.textureTarget();
|
| - case WebKit::WebVideoFrame::FormatInvalid:
|
| - case WebKit::WebVideoFrame::FormatRGB32:
|
| - case WebKit::WebVideoFrame::FormatEmpty:
|
| - case WebKit::WebVideoFrame::FormatI420:
|
| - notImplemented();
|
| - }
|
| - return GraphicsContext3D::INVALID_VALUE;
|
| -}
|
| -
|
| -void CCVideoLayerImpl::willDraw(CCResourceProvider* resourceProvider)
|
| -{
|
| - ASSERT(CCProxy::isImplThread());
|
| - CCLayerImpl::willDraw(resourceProvider);
|
| -
|
| - // Explicitly lock and unlock the provider mutex so it can be held from
|
| - // willDraw to didDraw. Since the compositor thread is in the middle of
|
| - // drawing, the layer will not be destroyed before didDraw is called.
|
| - // Therefore, the only thing that will prevent this lock from being released
|
| - // is the GPU process locking it. As the GPU process can't cause the
|
| - // destruction of the provider (calling stopUsingProvider), holding this
|
| - // lock should not cause a deadlock.
|
| - m_providerMutex.lock();
|
| -
|
| - willDrawInternal(resourceProvider);
|
| - freeUnusedPlaneData(resourceProvider);
|
| -
|
| - if (!m_frame)
|
| - m_providerMutex.unlock();
|
| -}
|
| -
|
| -void CCVideoLayerImpl::willDrawInternal(CCResourceProvider* resourceProvider)
|
| -{
|
| - ASSERT(CCProxy::isImplThread());
|
| - ASSERT(!m_externalTextureResource);
|
| -
|
| - if (!m_provider) {
|
| - m_frame = 0;
|
| - return;
|
| - }
|
| -
|
| - m_frame = m_provider->getCurrentFrame();
|
| -
|
| - if (!m_frame)
|
| - return;
|
| -
|
| - m_format = convertVFCFormatToGC3DFormat(*m_frame);
|
| -
|
| - if (m_format == GraphicsContext3D::INVALID_VALUE) {
|
| - m_provider->putCurrentFrame(m_frame);
|
| - m_frame = 0;
|
| - return;
|
| - }
|
| -
|
| - if (m_frame->planes() > WebKit::WebVideoFrame::maxPlanes) {
|
| - m_provider->putCurrentFrame(m_frame);
|
| - m_frame = 0;
|
| - return;
|
| - }
|
| -
|
| - if (!allocatePlaneData(resourceProvider)) {
|
| - m_provider->putCurrentFrame(m_frame);
|
| - m_frame = 0;
|
| - return;
|
| - }
|
| -
|
| - if (!copyPlaneData(resourceProvider)) {
|
| - m_provider->putCurrentFrame(m_frame);
|
| - m_frame = 0;
|
| - return;
|
| - }
|
| -
|
| - if (m_format == GraphicsContext3D::TEXTURE_2D)
|
| - m_externalTextureResource = resourceProvider->createResourceFromExternalTexture(m_frame->textureId());
|
| -}
|
| -
|
| -void CCVideoLayerImpl::appendQuads(CCQuadSink& quadSink, CCAppendQuadsData& appendQuadsData)
|
| -{
|
| - ASSERT(CCProxy::isImplThread());
|
| -
|
| - if (!m_frame)
|
| - return;
|
| -
|
| - CCSharedQuadState* sharedQuadState = quadSink.useSharedQuadState(createSharedQuadState());
|
| - appendDebugBorderQuad(quadSink, sharedQuadState, appendQuadsData);
|
| -
|
| - // FIXME: When we pass quads out of process, we need to double-buffer, or
|
| - // otherwise synchonize use of all textures in the quad.
|
| -
|
| - IntRect quadRect(IntPoint(), contentBounds());
|
| -
|
| - switch (m_format) {
|
| - case GraphicsContext3D::LUMINANCE: {
|
| - // YUV software decoder.
|
| - const FramePlane& yPlane = m_framePlanes[WebKit::WebVideoFrame::yPlane];
|
| - const FramePlane& uPlane = m_framePlanes[WebKit::WebVideoFrame::uPlane];
|
| - const FramePlane& vPlane = m_framePlanes[WebKit::WebVideoFrame::vPlane];
|
| - scoped_ptr<CCYUVVideoDrawQuad> yuvVideoQuad = CCYUVVideoDrawQuad::create(sharedQuadState, quadRect, yPlane, uPlane, vPlane);
|
| - quadSink.append(yuvVideoQuad.PassAs<CCDrawQuad>(), appendQuadsData);
|
| - break;
|
| - }
|
| - case GraphicsContext3D::RGBA: {
|
| - // RGBA software decoder.
|
| - const FramePlane& plane = m_framePlanes[WebKit::WebVideoFrame::rgbPlane];
|
| - float widthScaleFactor = static_cast<float>(plane.visibleSize.width()) / plane.size.width();
|
| -
|
| - bool premultipliedAlpha = true;
|
| - FloatRect uvRect(0, 0, widthScaleFactor, 1);
|
| - bool flipped = false;
|
| - scoped_ptr<CCTextureDrawQuad> textureQuad = CCTextureDrawQuad::create(sharedQuadState, quadRect, plane.resourceId, premultipliedAlpha, uvRect, flipped);
|
| - quadSink.append(textureQuad.PassAs<CCDrawQuad>(), appendQuadsData);
|
| - break;
|
| - }
|
| - case GraphicsContext3D::TEXTURE_2D: {
|
| - // NativeTexture hardware decoder.
|
| - bool premultipliedAlpha = true;
|
| - FloatRect uvRect(0, 0, 1, 1);
|
| - bool flipped = false;
|
| - scoped_ptr<CCTextureDrawQuad> textureQuad = CCTextureDrawQuad::create(sharedQuadState, quadRect, m_externalTextureResource, premultipliedAlpha, uvRect, flipped);
|
| - quadSink.append(textureQuad.PassAs<CCDrawQuad>(), appendQuadsData);
|
| - break;
|
| - }
|
| - case Extensions3D::TEXTURE_RECTANGLE_ARB: {
|
| - IntSize textureSize(m_frame->width(), m_frame->height());
|
| - scoped_ptr<CCIOSurfaceDrawQuad> ioSurfaceQuad = CCIOSurfaceDrawQuad::create(sharedQuadState, quadRect, textureSize, m_frame->textureId(), CCIOSurfaceDrawQuad::Unflipped);
|
| - quadSink.append(ioSurfaceQuad.PassAs<CCDrawQuad>(), appendQuadsData);
|
| - break;
|
| - }
|
| - case Extensions3DChromium::GL_TEXTURE_EXTERNAL_OES: {
|
| - // StreamTexture hardware decoder.
|
| - scoped_ptr<CCStreamVideoDrawQuad> streamVideoQuad = CCStreamVideoDrawQuad::create(sharedQuadState, quadRect, m_frame->textureId(), m_streamTextureMatrix);
|
| - quadSink.append(streamVideoQuad.PassAs<CCDrawQuad>(), appendQuadsData);
|
| - break;
|
| - }
|
| - default:
|
| - CRASH(); // Someone updated convertVFCFormatToGC3DFormat above but update this!
|
| - }
|
| -}
|
| -
|
| -void CCVideoLayerImpl::didDraw(CCResourceProvider* resourceProvider)
|
| -{
|
| - ASSERT(CCProxy::isImplThread());
|
| - CCLayerImpl::didDraw(resourceProvider);
|
| -
|
| - if (!m_frame)
|
| - return;
|
| -
|
| - if (m_format == GraphicsContext3D::TEXTURE_2D) {
|
| - ASSERT(m_externalTextureResource);
|
| - // FIXME: the following assert will not be true when sending resources to a
|
| - // parent compositor. We will probably need to hold on to m_frame for
|
| - // longer, and have several "current frames" in the pipeline.
|
| - ASSERT(!resourceProvider->inUseByConsumer(m_externalTextureResource));
|
| - resourceProvider->deleteResource(m_externalTextureResource);
|
| - m_externalTextureResource = 0;
|
| - }
|
| -
|
| - m_provider->putCurrentFrame(m_frame);
|
| - m_frame = 0;
|
| -
|
| - m_providerMutex.unlock();
|
| -}
|
| -
|
| -static int videoFrameDimension(int originalDimension, unsigned plane, int format)
|
| -{
|
| - if (format == WebKit::WebVideoFrame::FormatYV12 && plane != WebKit::WebVideoFrame::yPlane)
|
| - return originalDimension / 2;
|
| - return originalDimension;
|
| -}
|
| -
|
| -static bool hasPaddingBytes(const WebKit::WebVideoFrame& frame, unsigned plane)
|
| -{
|
| - return frame.stride(plane) > videoFrameDimension(frame.width(), plane, frame.format());
|
| -}
|
| -
|
| -IntSize CCVideoLayerImpl::computeVisibleSize(const WebKit::WebVideoFrame& frame, unsigned plane)
|
| -{
|
| - int visibleWidth = videoFrameDimension(frame.width(), plane, frame.format());
|
| - int originalWidth = visibleWidth;
|
| - int visibleHeight = videoFrameDimension(frame.height(), plane, frame.format());
|
| -
|
| - // When there are dead pixels at the edge of the texture, decrease
|
| - // the frame width by 1 to prevent the rightmost pixels from
|
| - // interpolating with the dead pixels.
|
| - if (hasPaddingBytes(frame, plane))
|
| - --visibleWidth;
|
| -
|
| - // In YV12, every 2x2 square of Y values corresponds to one U and
|
| - // one V value. If we decrease the width of the UV plane, we must decrease the
|
| - // width of the Y texture by 2 for proper alignment. This must happen
|
| - // always, even if Y's texture does not have padding bytes.
|
| - if (plane == WebKit::WebVideoFrame::yPlane && frame.format() == WebKit::WebVideoFrame::FormatYV12) {
|
| - if (hasPaddingBytes(frame, WebKit::WebVideoFrame::uPlane))
|
| - visibleWidth = originalWidth - 2;
|
| - }
|
| -
|
| - return IntSize(visibleWidth, visibleHeight);
|
| -}
|
| -
|
| -bool CCVideoLayerImpl::FramePlane::allocateData(CCResourceProvider* resourceProvider)
|
| -{
|
| - if (resourceId)
|
| - return true;
|
| -
|
| - resourceId = resourceProvider->createResource(CCRenderer::ImplPool, size, format, CCResourceProvider::TextureUsageAny);
|
| - return resourceId;
|
| -}
|
| -
|
| -void CCVideoLayerImpl::FramePlane::freeData(CCResourceProvider* resourceProvider)
|
| -{
|
| - if (!resourceId)
|
| - return;
|
| -
|
| - resourceProvider->deleteResource(resourceId);
|
| - resourceId = 0;
|
| -}
|
| -
|
| -bool CCVideoLayerImpl::allocatePlaneData(CCResourceProvider* resourceProvider)
|
| -{
|
| - int maxTextureSize = resourceProvider->maxTextureSize();
|
| - for (unsigned planeIndex = 0; planeIndex < m_frame->planes(); ++planeIndex) {
|
| - CCVideoLayerImpl::FramePlane& plane = m_framePlanes[planeIndex];
|
| -
|
| - IntSize requiredTextureSize(m_frame->stride(planeIndex), videoFrameDimension(m_frame->height(), planeIndex, m_frame->format()));
|
| - // FIXME: Remove the test against maxTextureSize when tiled layers are implemented.
|
| - if (requiredTextureSize.isZero() || requiredTextureSize.width() > maxTextureSize || requiredTextureSize.height() > maxTextureSize)
|
| - return false;
|
| -
|
| - if (plane.size != requiredTextureSize || plane.format != m_format) {
|
| - plane.freeData(resourceProvider);
|
| - plane.size = requiredTextureSize;
|
| - plane.format = m_format;
|
| - }
|
| -
|
| - if (!plane.resourceId) {
|
| - if (!plane.allocateData(resourceProvider))
|
| - return false;
|
| - plane.visibleSize = computeVisibleSize(*m_frame, planeIndex);
|
| - }
|
| - }
|
| - return true;
|
| -}
|
| -
|
| -bool CCVideoLayerImpl::copyPlaneData(CCResourceProvider* resourceProvider)
|
| -{
|
| - size_t softwarePlaneCount = m_frame->planes();
|
| - if (!softwarePlaneCount)
|
| - return true;
|
| -
|
| - for (size_t softwarePlaneIndex = 0; softwarePlaneIndex < softwarePlaneCount; ++softwarePlaneIndex) {
|
| - CCVideoLayerImpl::FramePlane& plane = m_framePlanes[softwarePlaneIndex];
|
| - const uint8_t* softwarePlanePixels = static_cast<const uint8_t*>(m_frame->data(softwarePlaneIndex));
|
| - IntRect planeRect(IntPoint(), plane.size);
|
| - resourceProvider->upload(plane.resourceId, softwarePlanePixels, planeRect, planeRect, IntSize());
|
| - }
|
| - return true;
|
| -}
|
| -
|
| -void CCVideoLayerImpl::freePlaneData(CCResourceProvider* resourceProvider)
|
| -{
|
| - for (unsigned i = 0; i < WebKit::WebVideoFrame::maxPlanes; ++i)
|
| - m_framePlanes[i].freeData(resourceProvider);
|
| -}
|
| -
|
| -void CCVideoLayerImpl::freeUnusedPlaneData(CCResourceProvider* resourceProvider)
|
| -{
|
| - unsigned firstUnusedPlane = m_frame ? m_frame->planes() : 0;
|
| - for (unsigned i = firstUnusedPlane; i < WebKit::WebVideoFrame::maxPlanes; ++i)
|
| - m_framePlanes[i].freeData(resourceProvider);
|
| -}
|
| -
|
| -void CCVideoLayerImpl::didReceiveFrame()
|
| -{
|
| - setNeedsRedraw();
|
| -}
|
| -
|
| -void CCVideoLayerImpl::didUpdateMatrix(const float matrix[16])
|
| -{
|
| - m_streamTextureMatrix = WebKit::WebTransformationMatrix(
|
| - matrix[0], matrix[1], matrix[2], matrix[3],
|
| - matrix[4], matrix[5], matrix[6], matrix[7],
|
| - matrix[8], matrix[9], matrix[10], matrix[11],
|
| - matrix[12], matrix[13], matrix[14], matrix[15]);
|
| - setNeedsRedraw();
|
| -}
|
| -
|
| -void CCVideoLayerImpl::didLoseContext()
|
| -{
|
| - freePlaneData(layerTreeHostImpl()->resourceProvider());
|
| -}
|
| -
|
| -void CCVideoLayerImpl::setNeedsRedraw()
|
| -{
|
| - layerTreeHostImpl()->setNeedsRedraw();
|
| -}
|
| -
|
| -void CCVideoLayerImpl::dumpLayerProperties(std::string* str, int indent) const
|
| -{
|
| - str->append(indentString(indent));
|
| - str->append("video layer\n");
|
| - CCLayerImpl::dumpLayerProperties(str, indent);
|
| -}
|
| -
|
| -const char* CCVideoLayerImpl::layerTypeAsString() const
|
| -{
|
| - return "VideoLayer";
|
| -}
|
| -
|
| -}
|
| -
|
| -#endif // USE(ACCELERATED_COMPOSITING)
|
|
|