Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(1209)

Unified Diff: content/common/gpu/media/gpu_jpeg_decode_accelerator.cc

Issue 1124423008: MJPEG acceleration for video capture using VAAPI, the GPU and IPC part (Closed) Base URL: https://chromium.googlesource.com/chromium/src.git@mjpeg-1-media
Patch Set: gpu_jpeg_decode_accelerator.cc filter to dispatch decode task Created 5 years, 7 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
Index: content/common/gpu/media/gpu_jpeg_decode_accelerator.cc
diff --git a/content/common/gpu/media/gpu_jpeg_decode_accelerator.cc b/content/common/gpu/media/gpu_jpeg_decode_accelerator.cc
new file mode 100644
index 0000000000000000000000000000000000000000..ba3976f5198ea03f5f6187f8533d70fb1ce5a97d
--- /dev/null
+++ b/content/common/gpu/media/gpu_jpeg_decode_accelerator.cc
@@ -0,0 +1,273 @@
+// Copyright 2015 The Chromium Authors. All rights reserved.
+// Use of this source code is governed by a BSD-style license that can be
+// found in the LICENSE file.
+
+#include "content/common/gpu/media/gpu_jpeg_decode_accelerator.h"
+
+#include <stdint.h>
+
+#include "base/bind.h"
+#include "base/logging.h"
+#include "base/memory/shared_memory.h"
+#include "base/single_thread_task_runner.h"
+#include "base/trace_event/trace_event.h"
+#include "content/common/gpu/gpu_channel.h"
+#include "content/common/gpu/gpu_messages.h"
+#include "ipc/ipc_message_macros.h"
+#include "ipc/message_filter.h"
+#include "media/filters/jpeg_parser.h"
+#include "ui/gfx/geometry/size.h"
+
+#if defined(OS_CHROMEOS) && defined(ARCH_CPU_X86_FAMILY)
+#include "content/common/gpu/media/vaapi_jpeg_decode_accelerator.h"
+#endif
+
+namespace {
+
+void DecodeFinished(scoped_ptr<base::SharedMemory> shm) {
+ // Do nothing. Because VideoFrame is backed by |shm|, the purpose of this
+ // function is to just keep reference of |shm| to make sure it lives util
+ // decode finishes.
+}
+
+} // namespace
+
+namespace content {
+
+class GpuJpegDecodeAccelerator::Client
+ : public media::JpegDecodeAccelerator::Client {
+ public:
+ Client(content::GpuJpegDecodeAccelerator* owner, int32 route_id)
+ : owner_(owner), route_id_(route_id) {}
+
+ // media::JpegDecodeAccelerator::Client implementation.
+ void VideoFrameReady(int32_t bitstream_buffer_id) override {
+ owner_->NotifyDecodeStatus(route_id_, bitstream_buffer_id,
+ media::JpegDecodeAccelerator::NO_ERROR);
+ }
+
+ void NotifyError(int32_t bitstream_buffer_id,
+ media::JpegDecodeAccelerator::Error error) override {
+ owner_->NotifyDecodeStatus(route_id_, bitstream_buffer_id, error);
+ }
+
+ void set_accelerator(scoped_ptr<media::JpegDecodeAccelerator> accelerator) {
+ accelerator_ = accelerator.Pass();
+ }
+
+ private:
+ content::GpuJpegDecodeAccelerator* owner_;
+ int32 route_id_;
+ scoped_ptr<media::JpegDecodeAccelerator> accelerator_;
+};
+
+class GpuJpegDecodeAccelerator::MessageFilter : public IPC::MessageFilter {
+ public:
+ MessageFilter(GpuJpegDecodeAccelerator* owner) : owner_(owner) {}
piman 2015/05/26 23:31:53 nit: explicit.
kcwu 2015/05/27 14:13:23 Done.
+
+ void OnChannelError() override { sender_ = NULL; }
+
+ void OnChannelClosing() override { sender_ = NULL; }
+
+ void OnFilterAdded(IPC::Sender* sender) override { sender_ = sender; }
+
+ bool OnMessageReceived(const IPC::Message& msg) override {
+ const int32 route_id = msg.routing_id();
+ if (!accelerator_map_.Lookup(route_id))
+ return false;
+
+ bool handled = true;
+ IPC_BEGIN_MESSAGE_MAP_WITH_PARAM(MessageFilter, msg, &route_id)
+ IPC_MESSAGE_HANDLER(AcceleratedJpegDecoderMsg_Decode, OnDecodeOnIOThread)
+ IPC_MESSAGE_HANDLER(AcceleratedJpegDecoderMsg_Destroy,
+ OnDestroyOnIOThread)
+ IPC_MESSAGE_UNHANDLED(handled = false)
+ IPC_END_MESSAGE_MAP()
+ return handled;
+ }
+
+ bool SendOnIOThread(IPC::Message* message) {
+ DCHECK(!message->is_sync());
+ if (!sender_) {
+ delete message;
+ return false;
+ }
+ return sender_->Send(message);
+ }
+
+ void AddClientOnIOThread(int32 route_id,
+ media::JpegDecodeAccelerator* accelerator,
+ IPC::Message* reply_msg) {
+ DCHECK(owner_->io_task_runner_->BelongsToCurrentThread());
+ DCHECK(!accelerator_map_.Lookup(route_id));
+
+ accelerator_map_.AddWithID(accelerator, route_id);
+ GpuMsg_CreateJpegDecoder::WriteReplyParams(reply_msg, true);
+ SendOnIOThread(reply_msg);
+ }
+
+ void OnDestroyOnIOThread(const int32* route_id) {
+ DCHECK(owner_->io_task_runner_->BelongsToCurrentThread());
+ accelerator_map_.Remove(*route_id);
+
+ owner_->child_task_runner_->PostTask(
+ FROM_HERE, base::Bind(&GpuJpegDecodeAccelerator::RemoveClient,
+ base::Unretained(owner_), *route_id));
piman 2015/05/26 23:31:53 The GpuJpegDecodeAccelerator could be destroyed be
kcwu 2015/05/27 14:13:23 Done.Thanks for your detail suggestion. However a
piman 2015/05/27 21:08:40 I would prefer if you didn't wait there. I don't t
wuchengli 2015/05/28 03:41:39 GpuJpegDecodeAccelerator owns the filter. When GJD
kcwu 2015/05/28 12:10:26 Done. I made ~GpuJpegDecodeAccelerator waiting for
+ }
+
+ void NotifyDecodeStatusOnIOThread(int32 route_id,
+ int32_t buffer_id,
+ media::JpegDecodeAccelerator::Error error) {
+ DCHECK(owner_->io_task_runner_->BelongsToCurrentThread());
+ SendOnIOThread(new AcceleratedJpegDecoderHostMsg_DecodeAck(
+ route_id, buffer_id, error));
+ }
+
+ void OnDecodeOnIOThread(
+ const int32* route_id,
+ const AcceleratedJpegDecoderMsg_Decode_Params& params) {
+ DCHECK(owner_->io_task_runner_->BelongsToCurrentThread());
+ DCHECK(route_id);
+ TRACE_EVENT0("jpeg", "GpuJpegDecodeAccelerator::MessageFilter::OnDecode");
+
+ if (params.input_buffer_id < 0) {
+ LOG(ERROR) << "BitstreamBuffer id " << params.input_buffer_id
+ << " out of range";
+ NotifyDecodeStatusOnIOThread(
+ *route_id, params.input_buffer_id,
+ media::JpegDecodeAccelerator::INVALID_ARGUMENT);
+ return;
+ }
+
+ media::BitstreamBuffer input_buffer(params.input_buffer_id,
+ params.input_buffer_handle,
+ params.input_buffer_size);
+
+ scoped_ptr<base::SharedMemory> output_shm(
+ new base::SharedMemory(params.output_video_frame_handle, false));
+ if (!output_shm->Map(params.output_buffer_size)) {
+ LOG(ERROR) << "Could not map output shared memory for input buffer id "
+ << params.input_buffer_id;
+ NotifyDecodeStatusOnIOThread(
+ *route_id, params.input_buffer_id,
+ media::JpegDecodeAccelerator::PLATFORM_FAILURE);
+ return;
+ }
+
+ uint8* shm_memory = reinterpret_cast<uint8*>(output_shm->memory());
+ scoped_refptr<media::VideoFrame> frame =
+ media::VideoFrame::WrapExternalPackedMemory(
+ media::VideoFrame::I420,
+ params.coded_size,
+ gfx::Rect(params.coded_size),
+ params.coded_size,
+ shm_memory,
+ params.output_buffer_size,
+ params.output_video_frame_handle,
+ 0,
+ base::TimeDelta(),
+ base::Bind(DecodeFinished, base::Passed(&output_shm)));
+
+ if (!frame.get()) {
+ LOG(ERROR) << "Could not create VideoFrame for input buffer id "
+ << params.input_buffer_id;
+ NotifyDecodeStatusOnIOThread(
+ *route_id, params.input_buffer_id,
+ media::JpegDecodeAccelerator::PLATFORM_FAILURE);
+ return;
+ }
+
+ media::JpegDecodeAccelerator* accelerator =
+ accelerator_map_.Lookup(*route_id);
+ DCHECK(accelerator);
+ accelerator->Decode(input_buffer, frame);
+ }
+
+ protected:
+ virtual ~MessageFilter() {}
+
+ private:
+ GpuJpegDecodeAccelerator* owner_;
+
+ // The sender to which this filter was added.
+ IPC::Sender* sender_;
+
+ // A map from route id to JpegDecodeAccelerator.
+ // This doesn't take ownership. The ownership is still maintained by
+ // GpuJpegDecodeAccelerator on child thread.
+ IDMap<media::JpegDecodeAccelerator, IDMapExternalPointer> accelerator_map_;
+};
+
+GpuJpegDecodeAccelerator::GpuJpegDecodeAccelerator(
+ GpuChannel* channel,
+ const scoped_refptr<base::SingleThreadTaskRunner>& io_task_runner)
+ : channel_(channel),
+ filter_(new MessageFilter(this)),
+ child_task_runner_(base::ThreadTaskRunnerHandle::Get()),
+ io_task_runner_(io_task_runner) {
+}
+
+GpuJpegDecodeAccelerator::~GpuJpegDecodeAccelerator() {
+ DCHECK(CalledOnValidThread());
+ DCHECK(client_map_.IsEmpty());
+}
+
+bool GpuJpegDecodeAccelerator::OnMessageReceived(const IPC::Message& msg) {
+ // Messages are actually handled in filter on IO thread.
+ return false;
+}
+
+void GpuJpegDecodeAccelerator::AddClient(int32 route_id,
+ IPC::Message* reply_msg) {
+ DCHECK(CalledOnValidThread());
+ scoped_ptr<media::JpegDecodeAccelerator> accelerator;
+
+// When adding more platforms, GpuJpegDecoder::Supported need
+// update as well.
+#if defined(OS_CHROMEOS) && defined(ARCH_CPU_X86_FAMILY)
+ accelerator.reset(new VaapiJpegDecodeAccelerator(io_task_runner_));
+#else
+ DVLOG(1) << "HW JPEG decode acceleration not available.";
+#endif
+
+ scoped_ptr<Client> client(new Client(this, route_id));
+ if (!accelerator.get() || !accelerator->Initialize(client.get())) {
+ DLOG(ERROR) << "JPEG accelerator Initialize failed";
+ GpuMsg_CreateJpegDecoder::WriteReplyParams(reply_msg, false);
+ Send(reply_msg);
+ return;
+ }
+
+ if (client_map_.IsEmpty())
+ channel_->AddFilter(filter_.get());
+
+ io_task_runner_->PostTask(
+ FROM_HERE, base::Bind(&MessageFilter::AddClientOnIOThread, filter_,
+ route_id, accelerator.get(), reply_msg));
+
+ client->set_accelerator(accelerator.Pass());
+ client_map_.AddWithID(client.release(), route_id);
+}
+
+void GpuJpegDecodeAccelerator::NotifyDecodeStatus(
+ int32 route_id,
+ int32_t buffer_id,
+ media::JpegDecodeAccelerator::Error error) {
+ DCHECK(CalledOnValidThread());
+ Send(new AcceleratedJpegDecoderHostMsg_DecodeAck(route_id, buffer_id, error));
+}
+
+void GpuJpegDecodeAccelerator::RemoveClient(int32 route_id) {
+ DCHECK(CalledOnValidThread());
+
+ client_map_.Remove(route_id);
+ if (client_map_.IsEmpty())
+ channel_->RemoveFilter(filter_.get());
+}
+
+bool GpuJpegDecodeAccelerator::Send(IPC::Message* message) {
+ DCHECK(CalledOnValidThread());
+ return channel_->Send(message);
+}
+
+} // namespace content

Powered by Google App Engine
This is Rietveld 408576698