Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(1579)

Unified Diff: content/common/gpu/media/android_video_decode_accelerator.cc

Issue 1313913003: Begin refactor of AVDA to support zero-copy. (Closed) Base URL: https://chromium.googlesource.com/chromium/src.git@master
Patch Set: minor updates. Created 5 years, 4 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
Index: content/common/gpu/media/android_video_decode_accelerator.cc
diff --git a/content/common/gpu/media/android_video_decode_accelerator.cc b/content/common/gpu/media/android_video_decode_accelerator.cc
index ce5c20cc3af754a5d08db5d7e766e6b78ec10723..283bc20dd83c35cc10e8d2b7ebafbb90d28567ff 100644
--- a/content/common/gpu/media/android_video_decode_accelerator.cc
+++ b/content/common/gpu/media/android_video_decode_accelerator.cc
@@ -20,608 +20,94 @@
namespace content {
-// Helper macros for dealing with failure. If |result| evaluates false, emit
-// |log| to ERROR, register |error| with the decoder, and return.
-#define RETURN_ON_FAILURE(result, log, error) \
- do { \
- if (!(result)) { \
- DLOG(ERROR) << log; \
- base::MessageLoop::current()->PostTask( \
- FROM_HERE, \
- base::Bind(&AndroidVideoDecodeAccelerator::NotifyError, \
- weak_this_factory_.GetWeakPtr(), \
- error)); \
- state_ = ERROR; \
- return; \
- } \
- } while (0)
-
// TODO(dwkang): We only need kMaxVideoFrames to pass media stack's prerolling
// phase, but 1 is added due to crbug.com/176036. This should be tuned when we
// have actual use case.
enum { kNumPictureBuffers = media::limits::kMaxVideoFrames + 1 };
-// Max number of bitstreams notified to the client with
-// NotifyEndOfBitstreamBuffer() before getting output from the bitstream.
-enum { kMaxBitstreamsNotifiedInAdvance = 32 };
-
-#if defined(ENABLE_MEDIA_PIPELINE_ON_ANDROID)
-// MediaCodec is only guaranteed to support baseline, but some devices may
-// support others. Advertise support for all H264 profiles and let the
-// MediaCodec fail when decoding if it's not actually supported. It's assumed
-// that consumers won't have software fallback for H264 on Android anyway.
-static const media::VideoCodecProfile kSupportedH264Profiles[] = {
- media::H264PROFILE_BASELINE,
- media::H264PROFILE_MAIN,
- media::H264PROFILE_EXTENDED,
- media::H264PROFILE_HIGH,
- media::H264PROFILE_HIGH10PROFILE,
- media::H264PROFILE_HIGH422PROFILE,
- media::H264PROFILE_HIGH444PREDICTIVEPROFILE,
- media::H264PROFILE_SCALABLEBASELINE,
- media::H264PROFILE_SCALABLEHIGH,
- media::H264PROFILE_STEREOHIGH,
- media::H264PROFILE_MULTIVIEWHIGH
-};
-#endif
-
-// Because MediaCodec is thread-hostile (must be poked on a single thread) and
-// has no callback mechanism (b/11990118), we must drive it by polling for
-// complete frames (and available input buffers, when the codec is fully
-// saturated). This function defines the polling delay. The value used is an
-// arbitrary choice that trades off CPU utilization (spinning) against latency.
-// Mirrors android_video_encode_accelerator.cc:EncodePollDelay().
-static inline const base::TimeDelta DecodePollDelay() {
- // An alternative to this polling scheme could be to dedicate a new thread
- // (instead of using the ChildThread) to run the MediaCodec, and make that
- // thread use the timeout-based flavor of MediaCodec's dequeue methods when it
- // believes the codec should complete "soon" (e.g. waiting for an input
- // buffer, or waiting for a picture when it knows enough complete input
- // pictures have been fed to saturate any internal buffering). This is
- // speculative and it's unclear that this would be a win (nor that there's a
- // reasonably device-agnostic way to fill in the "believes" above).
- return base::TimeDelta::FromMilliseconds(10);
-}
-
-static inline const base::TimeDelta NoWaitTimeOut() {
- return base::TimeDelta::FromMicroseconds(0);
-}
-
AndroidVideoDecodeAccelerator::AndroidVideoDecodeAccelerator(
const base::WeakPtr<gpu::gles2::GLES2Decoder> decoder,
const base::Callback<bool(void)>& make_context_current)
- : client_(NULL),
- make_context_current_(make_context_current),
- codec_(media::kCodecH264),
- state_(NO_ERROR),
- surface_texture_id_(0),
- picturebuffers_requested_(false),
- gl_decoder_(decoder),
- weak_this_factory_(this) {}
-
-AndroidVideoDecodeAccelerator::~AndroidVideoDecodeAccelerator() {
- DCHECK(thread_checker_.CalledOnValidThread());
-}
-
-bool AndroidVideoDecodeAccelerator::Initialize(media::VideoCodecProfile profile,
- Client* client) {
- DCHECK(!media_codec_);
- DCHECK(thread_checker_.CalledOnValidThread());
-
- client_ = client;
- codec_ = VideoCodecProfileToVideoCodec(profile);
-
- bool profile_supported = codec_ == media::kCodecVP8;
-#if defined(ENABLE_MEDIA_PIPELINE_ON_ANDROID)
- profile_supported |=
- (codec_ == media::kCodecVP9 || codec_ == media::kCodecH264);
-#endif
-
- if (!profile_supported) {
- LOG(ERROR) << "Unsupported profile: " << profile;
- return false;
- }
-
- // Only use MediaCodec for VP8/9 if it's likely backed by hardware.
- if ((codec_ == media::kCodecVP8 || codec_ == media::kCodecVP9) &&
- media::VideoCodecBridge::IsKnownUnaccelerated(
- codec_, media::MEDIA_CODEC_DECODER)) {
- DVLOG(1) << "Initialization failed: "
- << (codec_ == media::kCodecVP8 ? "vp8" : "vp9")
- << " is not hardware accelerated";
- return false;
- }
-
- if (!make_context_current_.Run()) {
- LOG(ERROR) << "Failed to make this decoder's GL context current.";
- return false;
- }
-
- if (!gl_decoder_) {
- LOG(ERROR) << "Failed to get gles2 decoder instance.";
- return false;
- }
- glGenTextures(1, &surface_texture_id_);
- glActiveTexture(GL_TEXTURE0);
- glBindTexture(GL_TEXTURE_EXTERNAL_OES, surface_texture_id_);
-
- glTexParameteri(GL_TEXTURE_EXTERNAL_OES, GL_TEXTURE_MAG_FILTER, GL_NEAREST);
- glTexParameteri(GL_TEXTURE_EXTERNAL_OES, GL_TEXTURE_MIN_FILTER, GL_NEAREST);
- glTexParameteri(GL_TEXTURE_EXTERNAL_OES,
- GL_TEXTURE_WRAP_S, GL_CLAMP_TO_EDGE);
- glTexParameteri(GL_TEXTURE_EXTERNAL_OES,
- GL_TEXTURE_WRAP_T, GL_CLAMP_TO_EDGE);
- gl_decoder_->RestoreTextureUnitBindings(0);
- gl_decoder_->RestoreActiveTexture();
-
- surface_texture_ = gfx::SurfaceTexture::Create(surface_texture_id_);
-
- if (!ConfigureMediaCodec()) {
- LOG(ERROR) << "Failed to create MediaCodec instance.";
- return false;
- }
-
- return true;
-}
-
-void AndroidVideoDecodeAccelerator::DoIOTask() {
- DCHECK(thread_checker_.CalledOnValidThread());
- if (state_ == ERROR) {
- return;
- }
-
- QueueInput();
- DequeueOutput();
-}
-
-void AndroidVideoDecodeAccelerator::QueueInput() {
- DCHECK(thread_checker_.CalledOnValidThread());
- if (bitstreams_notified_in_advance_.size() > kMaxBitstreamsNotifiedInAdvance)
- return;
- if (pending_bitstream_buffers_.empty())
- return;
-
- int input_buf_index = 0;
- media::MediaCodecStatus status = media_codec_->DequeueInputBuffer(
- NoWaitTimeOut(), &input_buf_index);
- if (status != media::MEDIA_CODEC_OK) {
- DCHECK(status == media::MEDIA_CODEC_DEQUEUE_INPUT_AGAIN_LATER ||
- status == media::MEDIA_CODEC_ERROR);
- return;
- }
-
- base::Time queued_time = pending_bitstream_buffers_.front().second;
- UMA_HISTOGRAM_TIMES("Media.AVDA.InputQueueTime",
- base::Time::Now() - queued_time);
- media::BitstreamBuffer bitstream_buffer =
- pending_bitstream_buffers_.front().first;
- pending_bitstream_buffers_.pop();
-
- if (bitstream_buffer.id() == -1) {
- media_codec_->QueueEOS(input_buf_index);
- return;
- }
-
- // Abuse the presentation time argument to propagate the bitstream
- // buffer ID to the output, so we can report it back to the client in
- // PictureReady().
- base::TimeDelta timestamp =
- base::TimeDelta::FromMicroseconds(bitstream_buffer.id());
-
- scoped_ptr<base::SharedMemory> shm(
- new base::SharedMemory(bitstream_buffer.handle(), true));
-
- RETURN_ON_FAILURE(shm->Map(bitstream_buffer.size()),
- "Failed to SharedMemory::Map()",
- UNREADABLE_INPUT);
-
- status =
- media_codec_->QueueInputBuffer(input_buf_index,
- static_cast<const uint8*>(shm->memory()),
- bitstream_buffer.size(),
- timestamp);
- RETURN_ON_FAILURE(status == media::MEDIA_CODEC_OK,
- "Failed to QueueInputBuffer: " << status,
- PLATFORM_FAILURE);
-
- // We should call NotifyEndOfBitstreamBuffer(), when no more decoded output
- // will be returned from the bitstream buffer. However, MediaCodec API is
- // not enough to guarantee it.
- // So, here, we calls NotifyEndOfBitstreamBuffer() in advance in order to
- // keep getting more bitstreams from the client, and throttle them by using
- // |bitstreams_notified_in_advance_|.
- // TODO(dwkang): check if there is a way to remove this workaround.
- base::MessageLoop::current()->PostTask(
- FROM_HERE,
- base::Bind(&AndroidVideoDecodeAccelerator::NotifyEndOfBitstreamBuffer,
- weak_this_factory_.GetWeakPtr(),
- bitstream_buffer.id()));
- bitstreams_notified_in_advance_.push_back(bitstream_buffer.id());
-}
-
-void AndroidVideoDecodeAccelerator::DequeueOutput() {
- DCHECK(thread_checker_.CalledOnValidThread());
- if (picturebuffers_requested_ && output_picture_buffers_.empty())
- return;
-
- if (!output_picture_buffers_.empty() && free_picture_ids_.empty()) {
- // Don't have any picture buffer to send. Need to wait more.
- return;
- }
-
- bool eos = false;
- base::TimeDelta timestamp;
- int32 buf_index = 0;
- do {
- size_t offset = 0;
- size_t size = 0;
-
- media::MediaCodecStatus status = media_codec_->DequeueOutputBuffer(
- NoWaitTimeOut(), &buf_index, &offset, &size, &timestamp, &eos, NULL);
- switch (status) {
- case media::MEDIA_CODEC_DEQUEUE_OUTPUT_AGAIN_LATER:
- case media::MEDIA_CODEC_ERROR:
- return;
-
- case media::MEDIA_CODEC_OUTPUT_FORMAT_CHANGED: {
- int32 width, height;
- media_codec_->GetOutputFormat(&width, &height);
-
- if (!picturebuffers_requested_) {
- picturebuffers_requested_ = true;
- size_ = gfx::Size(width, height);
- base::MessageLoop::current()->PostTask(
- FROM_HERE,
- base::Bind(&AndroidVideoDecodeAccelerator::RequestPictureBuffers,
- weak_this_factory_.GetWeakPtr()));
- } else {
- // Dynamic resolution change support is not specified by the Android
- // platform at and before JB-MR1, so it's not possible to smoothly
- // continue playback at this point. Instead, error out immediately,
- // expecting clients to Reset() as appropriate to avoid this.
- // b/7093648
- RETURN_ON_FAILURE(size_ == gfx::Size(width, height),
- "Dynamic resolution change is not supported.",
- PLATFORM_FAILURE);
- }
- return;
- }
-
- case media::MEDIA_CODEC_OUTPUT_BUFFERS_CHANGED:
- break;
-
- case media::MEDIA_CODEC_OK:
- DCHECK_GE(buf_index, 0);
- break;
-
- default:
- NOTREACHED();
- break;
- }
- } while (buf_index < 0);
-
- // This ignores the emitted ByteBuffer and instead relies on rendering to the
- // codec's SurfaceTexture and then copying from that texture to the client's
- // PictureBuffer's texture. This means that each picture's data is written
- // three times: once to the ByteBuffer, once to the SurfaceTexture, and once
- // to the client's texture. It would be nicer to either:
- // 1) Render directly to the client's texture from MediaCodec (one write); or
- // 2) Upload the ByteBuffer to the client's texture (two writes).
- // Unfortunately neither is possible:
- // 1) MediaCodec's use of SurfaceTexture is a singleton, and the texture
- // written to can't change during the codec's lifetime. b/11990461
- // 2) The ByteBuffer is likely to contain the pixels in a vendor-specific,
- // opaque/non-standard format. It's not possible to negotiate the decoder
- // to emit a specific colorspace, even using HW CSC. b/10706245
- // So, we live with these two extra copies per picture :(
- media_codec_->ReleaseOutputBuffer(buf_index, true);
-
- if (eos) {
- base::MessageLoop::current()->PostTask(
- FROM_HERE,
- base::Bind(&AndroidVideoDecodeAccelerator::NotifyFlushDone,
- weak_this_factory_.GetWeakPtr()));
- } else {
- int64 bitstream_buffer_id = timestamp.InMicroseconds();
- SendCurrentSurfaceToClient(static_cast<int32>(bitstream_buffer_id));
-
- // Removes ids former or equal than the id from decoder. Note that
- // |bitstreams_notified_in_advance_| does not mean bitstream ids in decoder
- // because of frame reordering issue. We just maintain this roughly and use
- // for the throttling purpose.
- std::list<int32>::iterator it;
- for (it = bitstreams_notified_in_advance_.begin();
- it != bitstreams_notified_in_advance_.end();
- ++it) {
- if (*it == bitstream_buffer_id) {
- bitstreams_notified_in_advance_.erase(
- bitstreams_notified_in_advance_.begin(), ++it);
- break;
- }
- }
- }
-}
-
-void AndroidVideoDecodeAccelerator::SendCurrentSurfaceToClient(
- int32 bitstream_id) {
- DCHECK(thread_checker_.CalledOnValidThread());
- DCHECK_NE(bitstream_id, -1);
- DCHECK(!free_picture_ids_.empty());
-
- RETURN_ON_FAILURE(make_context_current_.Run(),
- "Failed to make this decoder's GL context current.",
- PLATFORM_FAILURE);
-
- int32 picture_buffer_id = free_picture_ids_.front();
- free_picture_ids_.pop();
-
- float transfrom_matrix[16];
- surface_texture_->UpdateTexImage();
- surface_texture_->GetTransformMatrix(transfrom_matrix);
-
- OutputBufferMap::const_iterator i =
- output_picture_buffers_.find(picture_buffer_id);
- RETURN_ON_FAILURE(i != output_picture_buffers_.end(),
- "Can't find a PictureBuffer for " << picture_buffer_id,
- PLATFORM_FAILURE);
- uint32 picture_buffer_texture_id = i->second.texture_id();
-
- RETURN_ON_FAILURE(gl_decoder_.get(),
- "Failed to get gles2 decoder instance.",
- ILLEGAL_STATE);
- // Defer initializing the CopyTextureCHROMIUMResourceManager until it is
- // needed because it takes 10s of milliseconds to initialize.
- if (!copier_) {
- copier_.reset(new gpu::CopyTextureCHROMIUMResourceManager());
- copier_->Initialize(gl_decoder_.get());
- }
-
- // Here, we copy |surface_texture_id_| to the picture buffer instead of
- // setting new texture to |surface_texture_| by calling attachToGLContext()
- // because:
- // 1. Once we call detachFrameGLContext(), it deletes the texture previous
- // attached.
- // 2. SurfaceTexture requires us to apply a transform matrix when we show
- // the texture.
- // TODO(hkuang): get the StreamTexture transform matrix in GPU process
- // instead of using default matrix crbug.com/226218.
- const static GLfloat default_matrix[16] = {1.0f, 0.0f, 0.0f, 0.0f,
- 0.0f, 1.0f, 0.0f, 0.0f,
- 0.0f, 0.0f, 1.0f, 0.0f,
- 0.0f, 0.0f, 0.0f, 1.0f};
- copier_->DoCopyTextureWithTransform(gl_decoder_.get(),
- GL_TEXTURE_EXTERNAL_OES,
- surface_texture_id_,
- picture_buffer_texture_id,
- size_.width(),
- size_.height(),
- false,
- false,
- false,
- default_matrix);
-
- // TODO(henryhsu): Pass (0, 0) as visible size will cause several test
- // cases failed. We should make sure |size_| is coded size or visible size.
- base::MessageLoop::current()->PostTask(
- FROM_HERE, base::Bind(&AndroidVideoDecodeAccelerator::NotifyPictureReady,
- weak_this_factory_.GetWeakPtr(),
- media::Picture(picture_buffer_id, bitstream_id,
- gfx::Rect(size_), false)));
-}
-
-void AndroidVideoDecodeAccelerator::Decode(
- const media::BitstreamBuffer& bitstream_buffer) {
- DCHECK(thread_checker_.CalledOnValidThread());
- if (bitstream_buffer.id() != -1 && bitstream_buffer.size() == 0) {
- base::MessageLoop::current()->PostTask(
- FROM_HERE,
- base::Bind(&AndroidVideoDecodeAccelerator::NotifyEndOfBitstreamBuffer,
- weak_this_factory_.GetWeakPtr(),
- bitstream_buffer.id()));
- return;
- }
-
- pending_bitstream_buffers_.push(
- std::make_pair(bitstream_buffer, base::Time::Now()));
-
- DoIOTask();
+ : AndroidVideoDecodeAcceleratorBase(decoder, make_context_current) {
}
-void AndroidVideoDecodeAccelerator::AssignPictureBuffers(
- const std::vector<media::PictureBuffer>& buffers) {
- DCHECK(thread_checker_.CalledOnValidThread());
- DCHECK(output_picture_buffers_.empty());
- DCHECK(free_picture_ids_.empty());
-
- for (size_t i = 0; i < buffers.size(); ++i) {
- RETURN_ON_FAILURE(buffers[i].size() == size_,
- "Invalid picture buffer size was passed.",
- INVALID_ARGUMENT);
- int32 id = buffers[i].id();
- output_picture_buffers_.insert(std::make_pair(id, buffers[i]));
- free_picture_ids_.push(id);
- // Since the client might be re-using |picture_buffer_id| values, forget
- // about previously-dismissed IDs now. See ReusePictureBuffer() comment
- // about "zombies" for why we maintain this set in the first place.
- dismissed_picture_ids_.erase(id);
- }
-
- RETURN_ON_FAILURE(output_picture_buffers_.size() >= kNumPictureBuffers,
- "Invalid picture buffers were passed.",
- INVALID_ARGUMENT);
-
- DoIOTask();
-}
-
-void AndroidVideoDecodeAccelerator::ReusePictureBuffer(
- int32 picture_buffer_id) {
- DCHECK(thread_checker_.CalledOnValidThread());
-
- // This ReusePictureBuffer() might have been in a pipe somewhere (queued in
- // IPC, or in a PostTask either at the sender or receiver) when we sent a
- // DismissPictureBuffer() for this |picture_buffer_id|. Account for such
- // potential "zombie" IDs here.
- if (dismissed_picture_ids_.erase(picture_buffer_id))
- return;
-
- free_picture_ids_.push(picture_buffer_id);
-
- DoIOTask();
-}
-
-void AndroidVideoDecodeAccelerator::Flush() {
- DCHECK(thread_checker_.CalledOnValidThread());
-
- Decode(media::BitstreamBuffer(-1, base::SharedMemoryHandle(), 0));
-}
-
-bool AndroidVideoDecodeAccelerator::ConfigureMediaCodec() {
- DCHECK(thread_checker_.CalledOnValidThread());
- DCHECK(surface_texture_.get());
-
- gfx::ScopedJavaSurface surface(surface_texture_.get());
-
- // Pass a dummy 320x240 canvas size and let the codec signal the real size
- // when it's known from the bitstream.
- media_codec_.reset(media::VideoCodecBridge::CreateDecoder(
- codec_, false, gfx::Size(320, 240), surface.j_surface().obj(), NULL));
- if (!media_codec_)
- return false;
-
- io_timer_.Start(FROM_HERE,
- DecodePollDelay(),
- this,
- &AndroidVideoDecodeAccelerator::DoIOTask);
- return true;
-}
-
-void AndroidVideoDecodeAccelerator::Reset() {
- DCHECK(thread_checker_.CalledOnValidThread());
-
- while (!pending_bitstream_buffers_.empty()) {
- int32 bitstream_buffer_id = pending_bitstream_buffers_.front().first.id();
- pending_bitstream_buffers_.pop();
-
- if (bitstream_buffer_id != -1) {
- base::MessageLoop::current()->PostTask(
- FROM_HERE,
- base::Bind(&AndroidVideoDecodeAccelerator::NotifyEndOfBitstreamBuffer,
- weak_this_factory_.GetWeakPtr(),
- bitstream_buffer_id));
- }
- }
- bitstreams_notified_in_advance_.clear();
-
- for (OutputBufferMap::iterator it = output_picture_buffers_.begin();
- it != output_picture_buffers_.end();
- ++it) {
- client_->DismissPictureBuffer(it->first);
- dismissed_picture_ids_.insert(it->first);
- }
- output_picture_buffers_.clear();
- std::queue<int32> empty;
- std::swap(free_picture_ids_, empty);
- CHECK(free_picture_ids_.empty());
- picturebuffers_requested_ = false;
-
- // On some devices, and up to at least JB-MR1,
- // - flush() can fail after EOS (b/8125974); and
- // - mid-stream resolution change is unsupported (b/7093648).
- // To cope with these facts, we always stop & restart the codec on Reset().
- io_timer_.Stop();
- media_codec_->Stop();
- ConfigureMediaCodec();
- state_ = NO_ERROR;
-
- base::MessageLoop::current()->PostTask(
- FROM_HERE,
- base::Bind(&AndroidVideoDecodeAccelerator::NotifyResetDone,
- weak_this_factory_.GetWeakPtr()));
+AndroidVideoDecodeAccelerator::~AndroidVideoDecodeAccelerator() {
+ DCHECK(ThreadChecker().CalledOnValidThread());
}
void AndroidVideoDecodeAccelerator::Destroy() {
- DCHECK(thread_checker_.CalledOnValidThread());
-
- weak_this_factory_.InvalidateWeakPtrs();
- if (media_codec_) {
- io_timer_.Stop();
- media_codec_->Stop();
- }
- if (surface_texture_id_)
- glDeleteTextures(1, &surface_texture_id_);
- if (copier_)
- copier_->Destroy();
- delete this;
-}
-
-bool AndroidVideoDecodeAccelerator::CanDecodeOnIOThread() {
- return false;
-}
-
-void AndroidVideoDecodeAccelerator::RequestPictureBuffers() {
- client_->ProvidePictureBuffers(kNumPictureBuffers, size_, GL_TEXTURE_2D);
-}
-
-void AndroidVideoDecodeAccelerator::NotifyPictureReady(
- const media::Picture& picture) {
- client_->PictureReady(picture);
-}
-
-void AndroidVideoDecodeAccelerator::NotifyEndOfBitstreamBuffer(
- int input_buffer_id) {
- client_->NotifyEndOfBitstreamBuffer(input_buffer_id);
-}
-
-void AndroidVideoDecodeAccelerator::NotifyFlushDone() {
- client_->NotifyFlushDone();
-}
-
-void AndroidVideoDecodeAccelerator::NotifyResetDone() {
- client_->NotifyResetDone();
-}
-
-void AndroidVideoDecodeAccelerator::NotifyError(
- media::VideoDecodeAccelerator::Error error) {
- client_->NotifyError(error);
-}
-
-// static
-media::VideoDecodeAccelerator::SupportedProfiles
-AndroidVideoDecodeAccelerator::GetSupportedProfiles() {
- SupportedProfiles profiles;
-
- if (!media::VideoCodecBridge::IsKnownUnaccelerated(
- media::kCodecVP8, media::MEDIA_CODEC_DECODER)) {
- SupportedProfile profile;
- profile.profile = media::VP8PROFILE_ANY;
- profile.min_resolution.SetSize(0, 0);
- profile.max_resolution.SetSize(1920, 1088);
- profiles.push_back(profile);
- }
-
-#if defined(ENABLE_MEDIA_PIPELINE_ON_ANDROID)
- if (!media::VideoCodecBridge::IsKnownUnaccelerated(
- media::kCodecVP9, media::MEDIA_CODEC_DECODER)) {
- SupportedProfile profile;
- profile.profile = media::VP9PROFILE_ANY;
- profile.min_resolution.SetSize(0, 0);
- profile.max_resolution.SetSize(1920, 1088);
- profiles.push_back(profile);
- }
-
- for (const auto& supported_profile : kSupportedH264Profiles) {
- SupportedProfile profile;
- profile.profile = supported_profile;
- profile.min_resolution.SetSize(0, 0);
- // Advertise support for 4k and let the MediaCodec fail when decoding if it
- // doesn't support the resolution. It's assumed that consumers won't have
- // software fallback for H264 on Android anyway.
- profile.max_resolution.SetSize(3840, 2160);
- profiles.push_back(profile);
- }
-#endif
+ if (copier_)
+ copier_->Destroy();
+ AndroidVideoDecodeAcceleratorBase::Destroy();
+}
+
+uint32 AndroidVideoDecodeAccelerator::GetNumPictureBuffers() const {
+ return kNumPictureBuffers;
+}
+
+uint32 AndroidVideoDecodeAccelerator::GetTextureTarget() const {
+ return GL_TEXTURE_2D;
+}
+
+void AndroidVideoDecodeAccelerator::AssignCurrentSurfaceToPictureBuffer(
+ int32 codec_buf_index, const media::PictureBuffer& picture_buffer) {
+
+ // Render the codec buffer into |surface_texture_|, and switch it to be
+ // the front buffer.
+ // This ignores the emitted ByteBuffer and instead relies on rendering to
+ // the codec's SurfaceTexture and then copying from that texture to the
+ // client's PictureBuffer's texture. This means that each picture's data
+ // is written three times: once to the ByteBuffer, once to the
+ // SurfaceTexture, and once to the client's texture. It would be nicer to
+ // either:
+ // 1) Render directly to the client's texture from MediaCodec (one write);
+ // or
+ // 2) Upload the ByteBuffer to the client's texture (two writes).
+ // Unfortunately neither is possible:
+ // 1) MediaCodec's use of SurfaceTexture is a singleton, and the texture
+ // written to can't change during the codec's lifetime. b/11990461
+ // 2) The ByteBuffer is likely to contain the pixels in a vendor-specific,
+ // opaque/non-standard format. It's not possible to negotiate the
+ // decoder to emit a specific colorspace, even using HW CSC. b/10706245
+ // So, we live with these two extra copies per picture :(
+ GetMediaCodec()->ReleaseOutputBuffer(codec_buf_index, true);
+ GetSurfaceTexture()->UpdateTexImage();
+
+ float transfrom_matrix[16];
+ GetSurfaceTexture()->GetTransformMatrix(transfrom_matrix);
+
+ uint32 picture_buffer_texture_id = picture_buffer.texture_id();
+
+ // Defer initializing the CopyTextureCHROMIUMResourceManager until it is
+ // needed because it takes 10s of milliseconds to initialize.
+ if (!copier_) {
+ copier_.reset(new gpu::CopyTextureCHROMIUMResourceManager());
+ copier_->Initialize(GetGlDecoder());
+ }
- return profiles;
+ // Here, we copy |surface_texture_id_| to the picture buffer instead of
+ // setting new texture to |surface_texture_| by calling attachToGLContext()
+ // because:
+ // 1. Once we call detachFrameGLContext(), it deletes the texture previous
+ // attached.
+ // 2. SurfaceTexture requires us to apply a transform matrix when we show
+ // the texture.
+ // TODO(hkuang): get the StreamTexture transform matrix in GPU process
+ // instead of using default matrix crbug.com/226218.
+ const static GLfloat default_matrix[16] = {1.0f, 0.0f, 0.0f, 0.0f,
+ 0.0f, 1.0f, 0.0f, 0.0f,
+ 0.0f, 0.0f, 1.0f, 0.0f,
+ 0.0f, 0.0f, 0.0f, 1.0f};
+ copier_->DoCopyTextureWithTransform(GetGlDecoder(),
+ GL_TEXTURE_EXTERNAL_OES,
+ GetSurfaceTextureId(),
+ picture_buffer_texture_id,
+ GetSize().width(),
+ GetSize().height(),
+ false,
+ false,
+ false,
+ default_matrix);
}
} // namespace content

Powered by Google App Engine
This is Rietveld 408576698