Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(653)

Unified Diff: content/common/gpu/media/v4l2_video_encode_accelerator.cc

Issue 1882373004: Migrate content/common/gpu/media code to media/gpu (Closed) Base URL: https://chromium.googlesource.com/chromium/src.git@master
Patch Set: Fix several more bot-identified build issues Created 4 years, 8 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
Index: content/common/gpu/media/v4l2_video_encode_accelerator.cc
diff --git a/content/common/gpu/media/v4l2_video_encode_accelerator.cc b/content/common/gpu/media/v4l2_video_encode_accelerator.cc
deleted file mode 100644
index f8b56ab89faba4e1f1a74d78939dbdeb55c4bdc1..0000000000000000000000000000000000000000
--- a/content/common/gpu/media/v4l2_video_encode_accelerator.cc
+++ /dev/null
@@ -1,1249 +0,0 @@
-// Copyright 2014 The Chromium Authors. All rights reserved.
-// Use of this source code is governed by a BSD-style license that can be
-// found in the LICENSE file.
-
-#include <fcntl.h>
-#include <linux/videodev2.h>
-#include <poll.h>
-#include <string.h>
-#include <sys/eventfd.h>
-#include <sys/ioctl.h>
-#include <sys/mman.h>
-#include <utility>
-
-#include "base/callback.h"
-#include "base/command_line.h"
-#include "base/macros.h"
-#include "base/numerics/safe_conversions.h"
-#include "base/thread_task_runner_handle.h"
-#include "base/trace_event/trace_event.h"
-#include "content/common/gpu/media/shared_memory_region.h"
-#include "content/common/gpu/media/v4l2_video_encode_accelerator.h"
-#include "media/base/bind_to_current_loop.h"
-#include "media/base/bitstream_buffer.h"
-
-#define NOTIFY_ERROR(x) \
- do { \
- LOG(ERROR) << "Setting error state:" << x; \
- SetErrorState(x); \
- } while (0)
-
-#define IOCTL_OR_ERROR_RETURN_VALUE(type, arg, value, type_str) \
- do { \
- if (device_->Ioctl(type, arg) != 0) { \
- PLOG(ERROR) << __func__ << "(): ioctl() failed: " << type_str; \
- NOTIFY_ERROR(kPlatformFailureError); \
- return value; \
- } \
- } while (0)
-
-#define IOCTL_OR_ERROR_RETURN(type, arg) \
- IOCTL_OR_ERROR_RETURN_VALUE(type, arg, ((void)0), #type)
-
-#define IOCTL_OR_ERROR_RETURN_FALSE(type, arg) \
- IOCTL_OR_ERROR_RETURN_VALUE(type, arg, false, #type)
-
-#define IOCTL_OR_LOG_ERROR(type, arg) \
- do { \
- if (device_->Ioctl(type, arg) != 0) \
- PLOG(ERROR) << __func__ << "(): ioctl() failed: " << #type; \
- } while (0)
-
-namespace content {
-
-struct V4L2VideoEncodeAccelerator::BitstreamBufferRef {
- BitstreamBufferRef(int32_t id, std::unique_ptr<SharedMemoryRegion> shm)
- : id(id), shm(std::move(shm)) {}
- const int32_t id;
- const std::unique_ptr<SharedMemoryRegion> shm;
-};
-
-V4L2VideoEncodeAccelerator::InputRecord::InputRecord() : at_device(false) {
-}
-
-V4L2VideoEncodeAccelerator::InputRecord::~InputRecord() {
-}
-
-V4L2VideoEncodeAccelerator::OutputRecord::OutputRecord()
- : at_device(false), address(NULL), length(0) {
-}
-
-V4L2VideoEncodeAccelerator::OutputRecord::~OutputRecord() {
-}
-
-V4L2VideoEncodeAccelerator::ImageProcessorInputRecord::
- ImageProcessorInputRecord()
- : force_keyframe(false) {}
-
-V4L2VideoEncodeAccelerator::ImageProcessorInputRecord::
- ~ImageProcessorInputRecord() {}
-
-V4L2VideoEncodeAccelerator::V4L2VideoEncodeAccelerator(
- const scoped_refptr<V4L2Device>& device)
- : child_task_runner_(base::ThreadTaskRunnerHandle::Get()),
- output_buffer_byte_size_(0),
- device_input_format_(media::PIXEL_FORMAT_UNKNOWN),
- input_planes_count_(0),
- output_format_fourcc_(0),
- encoder_state_(kUninitialized),
- stream_header_size_(0),
- device_(device),
- input_streamon_(false),
- input_buffer_queued_count_(0),
- input_memory_type_(V4L2_MEMORY_USERPTR),
- output_streamon_(false),
- output_buffer_queued_count_(0),
- encoder_thread_("V4L2EncoderThread"),
- device_poll_thread_("V4L2EncoderDevicePollThread"),
- weak_this_ptr_factory_(this) {
- weak_this_ = weak_this_ptr_factory_.GetWeakPtr();
-}
-
-V4L2VideoEncodeAccelerator::~V4L2VideoEncodeAccelerator() {
- DCHECK(!encoder_thread_.IsRunning());
- DCHECK(!device_poll_thread_.IsRunning());
- DVLOG(4) << __func__;
-
- DestroyInputBuffers();
- DestroyOutputBuffers();
-}
-
-bool V4L2VideoEncodeAccelerator::Initialize(
- media::VideoPixelFormat input_format,
- const gfx::Size& input_visible_size,
- media::VideoCodecProfile output_profile,
- uint32_t initial_bitrate,
- Client* client) {
- DVLOG(3) << __func__
- << ": input_format=" << media::VideoPixelFormatToString(input_format)
- << ", input_visible_size=" << input_visible_size.ToString()
- << ", output_profile=" << output_profile
- << ", initial_bitrate=" << initial_bitrate;
-
- visible_size_ = input_visible_size;
-
- client_ptr_factory_.reset(new base::WeakPtrFactory<Client>(client));
- client_ = client_ptr_factory_->GetWeakPtr();
-
- DCHECK(child_task_runner_->BelongsToCurrentThread());
- DCHECK_EQ(encoder_state_, kUninitialized);
-
- struct v4l2_capability caps;
- memset(&caps, 0, sizeof(caps));
- const __u32 kCapsRequired = V4L2_CAP_VIDEO_M2M_MPLANE | V4L2_CAP_STREAMING;
- IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_QUERYCAP, &caps);
- if ((caps.capabilities & kCapsRequired) != kCapsRequired) {
- LOG(ERROR) << "Initialize(): ioctl() failed: VIDIOC_QUERYCAP: "
- "caps check failed: 0x" << std::hex << caps.capabilities;
- return false;
- }
-
- if (!SetFormats(input_format, output_profile)) {
- DLOG(ERROR) << "Failed setting up formats";
- return false;
- }
-
- if (input_format != device_input_format_) {
- DVLOG(1) << "Input format not supported by the HW, will convert to "
- << media::VideoPixelFormatToString(device_input_format_);
-
- scoped_refptr<V4L2Device> device =
- V4L2Device::Create(V4L2Device::kImageProcessor);
- image_processor_.reset(new V4L2ImageProcessor(device));
-
- // Convert from input_format to device_input_format_, keeping the size
- // at visible_size_ and requiring the output buffers to be of at least
- // input_allocated_size_.
- if (!image_processor_->Initialize(
- input_format, device_input_format_, visible_size_, visible_size_,
- input_allocated_size_, kImageProcBufferCount,
- base::Bind(&V4L2VideoEncodeAccelerator::ImageProcessorError,
- weak_this_))) {
- LOG(ERROR) << "Failed initializing image processor";
- return false;
- }
-
- for (int i = 0; i < kImageProcBufferCount; i++) {
- std::vector<base::ScopedFD> fds =
- image_processor_->GetDmabufsForOutputBuffer(i);
- if (fds.size() == 0) {
- LOG(ERROR) << __func__ << ": failed to get fds of image processor.";
- return false;
- }
- image_processor_output_buffer_map_.push_back(std::move(fds));
- free_image_processor_output_buffers_.push_back(i);
- }
- }
-
- if (!InitControls())
- return false;
-
- if (!CreateOutputBuffers())
- return false;
-
- if (!encoder_thread_.Start()) {
- LOG(ERROR) << "Initialize(): encoder thread failed to start";
- return false;
- }
-
- RequestEncodingParametersChange(initial_bitrate, kInitialFramerate);
-
- encoder_state_ = kInitialized;
-
- child_task_runner_->PostTask(
- FROM_HERE,
- base::Bind(&Client::RequireBitstreamBuffers, client_, kInputBufferCount,
- image_processor_.get()
- ? image_processor_->input_allocated_size()
- : input_allocated_size_,
- output_buffer_byte_size_));
- return true;
-}
-
-void V4L2VideoEncodeAccelerator::ImageProcessorError() {
- LOG(ERROR) << "Image processor error";
- NOTIFY_ERROR(kPlatformFailureError);
-}
-
-void V4L2VideoEncodeAccelerator::Encode(
- const scoped_refptr<media::VideoFrame>& frame,
- bool force_keyframe) {
- DVLOG(3) << "Encode(): force_keyframe=" << force_keyframe;
- DCHECK(child_task_runner_->BelongsToCurrentThread());
-
- if (image_processor_) {
- if (free_image_processor_output_buffers_.size() > 0) {
- int output_buffer_index = free_image_processor_output_buffers_.back();
- free_image_processor_output_buffers_.pop_back();
- image_processor_->Process(
- frame, output_buffer_index,
- base::Bind(&V4L2VideoEncodeAccelerator::FrameProcessed, weak_this_,
- force_keyframe, frame->timestamp()));
- } else {
- ImageProcessorInputRecord record;
- record.frame = frame;
- record.force_keyframe = force_keyframe;
- image_processor_input_queue_.push(record);
- }
- } else {
- encoder_thread_.message_loop()->PostTask(
- FROM_HERE,
- base::Bind(&V4L2VideoEncodeAccelerator::EncodeTask,
- base::Unretained(this),
- frame,
- force_keyframe));
- }
-}
-
-void V4L2VideoEncodeAccelerator::UseOutputBitstreamBuffer(
- const media::BitstreamBuffer& buffer) {
- DVLOG(3) << "UseOutputBitstreamBuffer(): id=" << buffer.id();
- DCHECK(child_task_runner_->BelongsToCurrentThread());
-
- if (buffer.size() < output_buffer_byte_size_) {
- NOTIFY_ERROR(kInvalidArgumentError);
- return;
- }
-
- std::unique_ptr<SharedMemoryRegion> shm(
- new SharedMemoryRegion(buffer, false));
- if (!shm->Map()) {
- NOTIFY_ERROR(kPlatformFailureError);
- return;
- }
-
- std::unique_ptr<BitstreamBufferRef> buffer_ref(
- new BitstreamBufferRef(buffer.id(), std::move(shm)));
- encoder_thread_.message_loop()->PostTask(
- FROM_HERE,
- base::Bind(&V4L2VideoEncodeAccelerator::UseOutputBitstreamBufferTask,
- base::Unretained(this),
- base::Passed(&buffer_ref)));
-}
-
-void V4L2VideoEncodeAccelerator::RequestEncodingParametersChange(
- uint32_t bitrate,
- uint32_t framerate) {
- DVLOG(3) << "RequestEncodingParametersChange(): bitrate=" << bitrate
- << ", framerate=" << framerate;
- DCHECK(child_task_runner_->BelongsToCurrentThread());
-
- encoder_thread_.message_loop()->PostTask(
- FROM_HERE,
- base::Bind(
- &V4L2VideoEncodeAccelerator::RequestEncodingParametersChangeTask,
- base::Unretained(this),
- bitrate,
- framerate));
-}
-
-void V4L2VideoEncodeAccelerator::Destroy() {
- DVLOG(3) << "Destroy()";
- DCHECK(child_task_runner_->BelongsToCurrentThread());
-
- // We're destroying; cancel all callbacks.
- client_ptr_factory_.reset();
- weak_this_ptr_factory_.InvalidateWeakPtrs();
-
- if (image_processor_.get())
- image_processor_.release()->Destroy();
-
- // If the encoder thread is running, destroy using posted task.
- if (encoder_thread_.IsRunning()) {
- encoder_thread_.message_loop()->PostTask(
- FROM_HERE,
- base::Bind(&V4L2VideoEncodeAccelerator::DestroyTask,
- base::Unretained(this)));
- // DestroyTask() will put the encoder into kError state and cause all tasks
- // to no-op.
- encoder_thread_.Stop();
- } else {
- // Otherwise, call the destroy task directly.
- DestroyTask();
- }
-
- // Set to kError state just in case.
- encoder_state_ = kError;
-
- delete this;
-}
-
-media::VideoEncodeAccelerator::SupportedProfiles
-V4L2VideoEncodeAccelerator::GetSupportedProfiles() {
- SupportedProfiles profiles;
- SupportedProfile profile;
- profile.max_framerate_numerator = 30;
- profile.max_framerate_denominator = 1;
-
- gfx::Size min_resolution;
- v4l2_fmtdesc fmtdesc;
- memset(&fmtdesc, 0, sizeof(fmtdesc));
- fmtdesc.type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE;
- for (; device_->Ioctl(VIDIOC_ENUM_FMT, &fmtdesc) == 0; ++fmtdesc.index) {
- device_->GetSupportedResolution(fmtdesc.pixelformat,
- &min_resolution, &profile.max_resolution);
- switch (fmtdesc.pixelformat) {
- case V4L2_PIX_FMT_H264:
- profile.profile = media::H264PROFILE_MAIN;
- profiles.push_back(profile);
- break;
- case V4L2_PIX_FMT_VP8:
- profile.profile = media::VP8PROFILE_ANY;
- profiles.push_back(profile);
- break;
- case V4L2_PIX_FMT_VP9:
- profile.profile = media::VP9PROFILE_PROFILE0;
- profiles.push_back(profile);
- profile.profile = media::VP9PROFILE_PROFILE1;
- profiles.push_back(profile);
- profile.profile = media::VP9PROFILE_PROFILE2;
- profiles.push_back(profile);
- profile.profile = media::VP9PROFILE_PROFILE3;
- profiles.push_back(profile);
- break;
- }
- }
-
- return profiles;
-}
-
-void V4L2VideoEncodeAccelerator::FrameProcessed(bool force_keyframe,
- base::TimeDelta timestamp,
- int output_buffer_index) {
- DCHECK(child_task_runner_->BelongsToCurrentThread());
- DVLOG(3) << "FrameProcessed(): force_keyframe=" << force_keyframe
- << ", output_buffer_index=" << output_buffer_index;
- DCHECK_GE(output_buffer_index, 0);
- DCHECK_LT(output_buffer_index, image_processor_output_buffer_map_.size());
-
- std::vector<base::ScopedFD>& scoped_fds =
- image_processor_output_buffer_map_[output_buffer_index];
- std::vector<int> fds;
- for (auto& fd : scoped_fds) {
- fds.push_back(fd.get());
- }
- scoped_refptr<media::VideoFrame> output_frame =
- media::VideoFrame::WrapExternalDmabufs(
- device_input_format_, image_processor_->output_allocated_size(),
- gfx::Rect(visible_size_), visible_size_, fds, timestamp);
- if (!output_frame) {
- NOTIFY_ERROR(kPlatformFailureError);
- return;
- }
- output_frame->AddDestructionObserver(media::BindToCurrentLoop(
- base::Bind(&V4L2VideoEncodeAccelerator::ReuseImageProcessorOutputBuffer,
- weak_this_, output_buffer_index)));
-
- encoder_thread_.message_loop()->PostTask(
- FROM_HERE,
- base::Bind(&V4L2VideoEncodeAccelerator::EncodeTask,
- base::Unretained(this), output_frame, force_keyframe));
-}
-
-void V4L2VideoEncodeAccelerator::ReuseImageProcessorOutputBuffer(
- int output_buffer_index) {
- DCHECK(child_task_runner_->BelongsToCurrentThread());
- DVLOG(3) << __func__ << ": output_buffer_index=" << output_buffer_index;
- free_image_processor_output_buffers_.push_back(output_buffer_index);
- if (!image_processor_input_queue_.empty()) {
- ImageProcessorInputRecord record = image_processor_input_queue_.front();
- image_processor_input_queue_.pop();
- Encode(record.frame, record.force_keyframe);
- }
-}
-
-void V4L2VideoEncodeAccelerator::EncodeTask(
- const scoped_refptr<media::VideoFrame>& frame,
- bool force_keyframe) {
- DVLOG(3) << "EncodeTask(): force_keyframe=" << force_keyframe;
- DCHECK_EQ(encoder_thread_.message_loop(), base::MessageLoop::current());
- DCHECK_NE(encoder_state_, kUninitialized);
-
- if (encoder_state_ == kError) {
- DVLOG(2) << "EncodeTask(): early out: kError state";
- return;
- }
-
- encoder_input_queue_.push(frame);
- Enqueue();
-
- if (force_keyframe) {
- // TODO(posciak): this presently makes for slightly imprecise encoding
- // parameters updates. To precisely align the parameter updates with the
- // incoming input frame, we should queue the parameters together with the
- // frame onto encoder_input_queue_ and apply them when the input is about
- // to be queued to the codec.
- std::vector<struct v4l2_ext_control> ctrls;
- struct v4l2_ext_control ctrl;
- memset(&ctrl, 0, sizeof(ctrl));
- ctrl.id = V4L2_CID_MPEG_VIDEO_FORCE_KEY_FRAME;
- ctrls.push_back(ctrl);
- if (!SetExtCtrls(ctrls)) {
- // Some platforms still use the old control. Fallback before they are
- // updated.
- ctrls.clear();
- memset(&ctrl, 0, sizeof(ctrl));
- ctrl.id = V4L2_CID_MPEG_MFC51_VIDEO_FORCE_FRAME_TYPE;
- ctrl.value = V4L2_MPEG_MFC51_VIDEO_FORCE_FRAME_TYPE_I_FRAME;
- ctrls.push_back(ctrl);
- if (!SetExtCtrls(ctrls)) {
- LOG(ERROR) << "Failed requesting keyframe";
- NOTIFY_ERROR(kPlatformFailureError);
- return;
- }
- }
- }
-}
-
-void V4L2VideoEncodeAccelerator::UseOutputBitstreamBufferTask(
- std::unique_ptr<BitstreamBufferRef> buffer_ref) {
- DVLOG(3) << "UseOutputBitstreamBufferTask(): id=" << buffer_ref->id;
- DCHECK_EQ(encoder_thread_.message_loop(), base::MessageLoop::current());
-
- encoder_output_queue_.push_back(
- linked_ptr<BitstreamBufferRef>(buffer_ref.release()));
- Enqueue();
-
- if (encoder_state_ == kInitialized) {
- // Finish setting up our OUTPUT queue. See: Initialize().
- // VIDIOC_REQBUFS on OUTPUT queue.
- if (!CreateInputBuffers())
- return;
- if (!StartDevicePoll())
- return;
- encoder_state_ = kEncoding;
- }
-}
-
-void V4L2VideoEncodeAccelerator::DestroyTask() {
- DVLOG(3) << "DestroyTask()";
-
- // DestroyTask() should run regardless of encoder_state_.
-
- // Stop streaming and the device_poll_thread_.
- StopDevicePoll();
-
- // Set our state to kError, and early-out all tasks.
- encoder_state_ = kError;
-}
-
-void V4L2VideoEncodeAccelerator::ServiceDeviceTask() {
- DVLOG(3) << "ServiceDeviceTask()";
- DCHECK_EQ(encoder_thread_.message_loop(), base::MessageLoop::current());
- DCHECK_NE(encoder_state_, kUninitialized);
- DCHECK_NE(encoder_state_, kInitialized);
-
- if (encoder_state_ == kError) {
- DVLOG(2) << "ServiceDeviceTask(): early out: kError state";
- return;
- }
-
- Dequeue();
- Enqueue();
-
- // Clear the interrupt fd.
- if (!device_->ClearDevicePollInterrupt())
- return;
-
- // Device can be polled as soon as either input or output buffers are queued.
- bool poll_device =
- (input_buffer_queued_count_ + output_buffer_queued_count_ > 0);
-
- // ServiceDeviceTask() should only ever be scheduled from DevicePollTask(),
- // so either:
- // * device_poll_thread_ is running normally
- // * device_poll_thread_ scheduled us, but then a DestroyTask() shut it down,
- // in which case we're in kError state, and we should have early-outed
- // already.
- DCHECK(device_poll_thread_.message_loop());
- // Queue the DevicePollTask() now.
- device_poll_thread_.message_loop()->PostTask(
- FROM_HERE,
- base::Bind(&V4L2VideoEncodeAccelerator::DevicePollTask,
- base::Unretained(this),
- poll_device));
-
- DVLOG(2) << __func__ << ": buffer counts: ENC["
- << encoder_input_queue_.size() << "] => DEVICE["
- << free_input_buffers_.size() << "+"
- << input_buffer_queued_count_ << "/"
- << input_buffer_map_.size() << "->"
- << free_output_buffers_.size() << "+"
- << output_buffer_queued_count_ << "/"
- << output_buffer_map_.size() << "] => OUT["
- << encoder_output_queue_.size() << "]";
-}
-
-void V4L2VideoEncodeAccelerator::Enqueue() {
- DCHECK_EQ(encoder_thread_.message_loop(), base::MessageLoop::current());
-
- DVLOG(3) << "Enqueue() "
- << "free_input_buffers: " << free_input_buffers_.size()
- << "input_queue: " << encoder_input_queue_.size();
-
- // Enqueue all the inputs we can.
- const int old_inputs_queued = input_buffer_queued_count_;
- // while (!ready_input_buffers_.empty()) {
- while (!encoder_input_queue_.empty() && !free_input_buffers_.empty()) {
- if (!EnqueueInputRecord())
- return;
- }
- if (old_inputs_queued == 0 && input_buffer_queued_count_ != 0) {
- // We just started up a previously empty queue.
- // Queue state changed; signal interrupt.
- if (!device_->SetDevicePollInterrupt())
- return;
- // Start VIDIOC_STREAMON if we haven't yet.
- if (!input_streamon_) {
- __u32 type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE;
- IOCTL_OR_ERROR_RETURN(VIDIOC_STREAMON, &type);
- input_streamon_ = true;
- }
- }
-
- // Enqueue all the outputs we can.
- const int old_outputs_queued = output_buffer_queued_count_;
- while (!free_output_buffers_.empty() && !encoder_output_queue_.empty()) {
- if (!EnqueueOutputRecord())
- return;
- }
- if (old_outputs_queued == 0 && output_buffer_queued_count_ != 0) {
- // We just started up a previously empty queue.
- // Queue state changed; signal interrupt.
- if (!device_->SetDevicePollInterrupt())
- return;
- // Start VIDIOC_STREAMON if we haven't yet.
- if (!output_streamon_) {
- __u32 type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE;
- IOCTL_OR_ERROR_RETURN(VIDIOC_STREAMON, &type);
- output_streamon_ = true;
- }
- }
-}
-
-void V4L2VideoEncodeAccelerator::Dequeue() {
- DVLOG(3) << "Dequeue()";
- DCHECK_EQ(encoder_thread_.message_loop(), base::MessageLoop::current());
-
- // Dequeue completed input (VIDEO_OUTPUT) buffers, and recycle to the free
- // list.
- struct v4l2_buffer dqbuf;
- struct v4l2_plane planes[VIDEO_MAX_PLANES];
- while (input_buffer_queued_count_ > 0) {
- DVLOG(4) << "inputs queued: " << input_buffer_queued_count_;
- DCHECK(input_streamon_);
- memset(&dqbuf, 0, sizeof(dqbuf));
- memset(&planes, 0, sizeof(planes));
- dqbuf.type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE;
- dqbuf.memory = input_memory_type_;
- dqbuf.m.planes = planes;
- dqbuf.length = input_planes_count_;
- if (device_->Ioctl(VIDIOC_DQBUF, &dqbuf) != 0) {
- if (errno == EAGAIN) {
- // EAGAIN if we're just out of buffers to dequeue.
- break;
- }
- PLOG(ERROR) << "Dequeue(): ioctl() failed: VIDIOC_DQBUF";
- NOTIFY_ERROR(kPlatformFailureError);
- return;
- }
- InputRecord& input_record = input_buffer_map_[dqbuf.index];
- DCHECK(input_record.at_device);
- input_record.at_device = false;
-
- input_record.frame = NULL;
- free_input_buffers_.push_back(dqbuf.index);
- input_buffer_queued_count_--;
- }
-
- // Dequeue completed output (VIDEO_CAPTURE) buffers, and recycle to the
- // free list. Notify the client that an output buffer is complete.
- while (output_buffer_queued_count_ > 0) {
- DCHECK(output_streamon_);
- memset(&dqbuf, 0, sizeof(dqbuf));
- memset(planes, 0, sizeof(planes));
- dqbuf.type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE;
- dqbuf.memory = V4L2_MEMORY_MMAP;
- dqbuf.m.planes = planes;
- dqbuf.length = 1;
- if (device_->Ioctl(VIDIOC_DQBUF, &dqbuf) != 0) {
- if (errno == EAGAIN) {
- // EAGAIN if we're just out of buffers to dequeue.
- break;
- }
- PLOG(ERROR) << "Dequeue(): ioctl() failed: VIDIOC_DQBUF";
- NOTIFY_ERROR(kPlatformFailureError);
- return;
- }
- const bool key_frame = ((dqbuf.flags & V4L2_BUF_FLAG_KEYFRAME) != 0);
- OutputRecord& output_record = output_buffer_map_[dqbuf.index];
- DCHECK(output_record.at_device);
- DCHECK(output_record.buffer_ref.get());
-
- void* output_data = output_record.address;
- size_t output_size = dqbuf.m.planes[0].bytesused;
- // This shouldn't happen, but just in case. We should be able to recover
- // after next keyframe after showing some corruption.
- DCHECK_LE(output_size, output_buffer_byte_size_);
- if (output_size > output_buffer_byte_size_)
- output_size = output_buffer_byte_size_;
- uint8_t* target_data =
- reinterpret_cast<uint8_t*>(output_record.buffer_ref->shm->memory());
- if (output_format_fourcc_ == V4L2_PIX_FMT_H264) {
- if (stream_header_size_ == 0) {
- // Assume that the first buffer dequeued is the stream header.
- stream_header_size_ = output_size;
- stream_header_.reset(new uint8_t[stream_header_size_]);
- memcpy(stream_header_.get(), output_data, stream_header_size_);
- }
- if (key_frame &&
- output_buffer_byte_size_ - stream_header_size_ >= output_size) {
- // Insert stream header before every keyframe.
- memcpy(target_data, stream_header_.get(), stream_header_size_);
- memcpy(target_data + stream_header_size_, output_data, output_size);
- output_size += stream_header_size_;
- } else {
- memcpy(target_data, output_data, output_size);
- }
- } else {
- memcpy(target_data, output_data, output_size);
- }
-
- DVLOG(3) << "Dequeue(): returning "
- "bitstream_buffer_id=" << output_record.buffer_ref->id
- << ", size=" << output_size << ", key_frame=" << key_frame;
- child_task_runner_->PostTask(
- FROM_HERE,
- base::Bind(&Client::BitstreamBufferReady, client_,
- output_record.buffer_ref->id, output_size, key_frame));
- output_record.at_device = false;
- output_record.buffer_ref.reset();
- free_output_buffers_.push_back(dqbuf.index);
- output_buffer_queued_count_--;
- }
-}
-
-bool V4L2VideoEncodeAccelerator::EnqueueInputRecord() {
- DVLOG(3) << "EnqueueInputRecord()";
- DCHECK(!free_input_buffers_.empty());
- DCHECK(!encoder_input_queue_.empty());
-
- // Enqueue an input (VIDEO_OUTPUT) buffer.
- scoped_refptr<media::VideoFrame> frame = encoder_input_queue_.front();
- const int index = free_input_buffers_.back();
- InputRecord& input_record = input_buffer_map_[index];
- DCHECK(!input_record.at_device);
- struct v4l2_buffer qbuf;
- struct v4l2_plane qbuf_planes[VIDEO_MAX_PLANES];
- memset(&qbuf, 0, sizeof(qbuf));
- memset(qbuf_planes, 0, sizeof(qbuf_planes));
- qbuf.index = index;
- qbuf.type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE;
- qbuf.m.planes = qbuf_planes;
-
- DCHECK_EQ(device_input_format_, frame->format());
- for (size_t i = 0; i < input_planes_count_; ++i) {
- qbuf.m.planes[i].bytesused =
- base::checked_cast<__u32>(media::VideoFrame::PlaneSize(
- frame->format(), i, input_allocated_size_).GetArea());
-
- switch (input_memory_type_) {
- case V4L2_MEMORY_USERPTR:
- qbuf.m.planes[i].length = qbuf.m.planes[i].bytesused;
- qbuf.m.planes[i].m.userptr =
- reinterpret_cast<unsigned long>(frame->data(i));
- DCHECK(qbuf.m.planes[i].m.userptr);
- break;
-
- case V4L2_MEMORY_DMABUF:
- qbuf.m.planes[i].m.fd = frame->dmabuf_fd(i);
- DCHECK_NE(qbuf.m.planes[i].m.fd, -1);
- break;
-
- default:
- NOTREACHED();
- return false;
- }
- }
-
- qbuf.memory = input_memory_type_;
- qbuf.length = input_planes_count_;
-
- IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_QBUF, &qbuf);
- input_record.at_device = true;
- input_record.frame = frame;
- encoder_input_queue_.pop();
- free_input_buffers_.pop_back();
- input_buffer_queued_count_++;
- return true;
-}
-
-bool V4L2VideoEncodeAccelerator::EnqueueOutputRecord() {
- DVLOG(3) << "EnqueueOutputRecord()";
- DCHECK(!free_output_buffers_.empty());
- DCHECK(!encoder_output_queue_.empty());
-
- // Enqueue an output (VIDEO_CAPTURE) buffer.
- linked_ptr<BitstreamBufferRef> output_buffer = encoder_output_queue_.back();
- const int index = free_output_buffers_.back();
- OutputRecord& output_record = output_buffer_map_[index];
- DCHECK(!output_record.at_device);
- DCHECK(!output_record.buffer_ref.get());
- struct v4l2_buffer qbuf;
- struct v4l2_plane qbuf_planes[1];
- memset(&qbuf, 0, sizeof(qbuf));
- memset(qbuf_planes, 0, sizeof(qbuf_planes));
- qbuf.index = index;
- qbuf.type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE;
- qbuf.memory = V4L2_MEMORY_MMAP;
- qbuf.m.planes = qbuf_planes;
- qbuf.length = 1;
- IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_QBUF, &qbuf);
- output_record.at_device = true;
- output_record.buffer_ref = output_buffer;
- encoder_output_queue_.pop_back();
- free_output_buffers_.pop_back();
- output_buffer_queued_count_++;
- return true;
-}
-
-bool V4L2VideoEncodeAccelerator::StartDevicePoll() {
- DVLOG(3) << "StartDevicePoll()";
- DCHECK_EQ(encoder_thread_.message_loop(), base::MessageLoop::current());
- DCHECK(!device_poll_thread_.IsRunning());
-
- // Start up the device poll thread and schedule its first DevicePollTask().
- if (!device_poll_thread_.Start()) {
- LOG(ERROR) << "StartDevicePoll(): Device thread failed to start";
- NOTIFY_ERROR(kPlatformFailureError);
- return false;
- }
- // Enqueue a poll task with no devices to poll on -- it will wait only on the
- // interrupt fd.
- device_poll_thread_.message_loop()->PostTask(
- FROM_HERE,
- base::Bind(&V4L2VideoEncodeAccelerator::DevicePollTask,
- base::Unretained(this),
- false));
-
- return true;
-}
-
-bool V4L2VideoEncodeAccelerator::StopDevicePoll() {
- DVLOG(3) << "StopDevicePoll()";
-
- // Signal the DevicePollTask() to stop, and stop the device poll thread.
- if (!device_->SetDevicePollInterrupt())
- return false;
- device_poll_thread_.Stop();
- // Clear the interrupt now, to be sure.
- if (!device_->ClearDevicePollInterrupt())
- return false;
-
- if (input_streamon_) {
- __u32 type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE;
- IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_STREAMOFF, &type);
- }
- input_streamon_ = false;
-
- if (output_streamon_) {
- __u32 type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE;
- IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_STREAMOFF, &type);
- }
- output_streamon_ = false;
-
- // Reset all our accounting info.
- while (!encoder_input_queue_.empty())
- encoder_input_queue_.pop();
- free_input_buffers_.clear();
- for (size_t i = 0; i < input_buffer_map_.size(); ++i) {
- InputRecord& input_record = input_buffer_map_[i];
- input_record.at_device = false;
- input_record.frame = NULL;
- free_input_buffers_.push_back(i);
- }
- input_buffer_queued_count_ = 0;
-
- free_output_buffers_.clear();
- for (size_t i = 0; i < output_buffer_map_.size(); ++i) {
- OutputRecord& output_record = output_buffer_map_[i];
- output_record.at_device = false;
- output_record.buffer_ref.reset();
- free_output_buffers_.push_back(i);
- }
- output_buffer_queued_count_ = 0;
-
- encoder_output_queue_.clear();
-
- DVLOG(3) << "StopDevicePoll(): device poll stopped";
- return true;
-}
-
-void V4L2VideoEncodeAccelerator::DevicePollTask(bool poll_device) {
- DVLOG(3) << "DevicePollTask()";
- DCHECK_EQ(device_poll_thread_.message_loop(), base::MessageLoop::current());
-
- bool event_pending;
- if (!device_->Poll(poll_device, &event_pending)) {
- NOTIFY_ERROR(kPlatformFailureError);
- return;
- }
-
- // All processing should happen on ServiceDeviceTask(), since we shouldn't
- // touch encoder state from this thread.
- encoder_thread_.message_loop()->PostTask(
- FROM_HERE,
- base::Bind(&V4L2VideoEncodeAccelerator::ServiceDeviceTask,
- base::Unretained(this)));
-}
-
-void V4L2VideoEncodeAccelerator::NotifyError(Error error) {
- DVLOG(1) << "NotifyError(): error=" << error;
-
- if (!child_task_runner_->BelongsToCurrentThread()) {
- child_task_runner_->PostTask(
- FROM_HERE, base::Bind(&V4L2VideoEncodeAccelerator::NotifyError,
- weak_this_, error));
- return;
- }
-
- if (client_) {
- client_->NotifyError(error);
- client_ptr_factory_.reset();
- }
-}
-
-void V4L2VideoEncodeAccelerator::SetErrorState(Error error) {
- // We can touch encoder_state_ only if this is the encoder thread or the
- // encoder thread isn't running.
- if (encoder_thread_.message_loop() != NULL &&
- encoder_thread_.message_loop() != base::MessageLoop::current()) {
- encoder_thread_.message_loop()->PostTask(
- FROM_HERE, base::Bind(&V4L2VideoEncodeAccelerator::SetErrorState,
- base::Unretained(this), error));
- return;
- }
-
- // Post NotifyError only if we are already initialized, as the API does
- // not allow doing so before that.
- if (encoder_state_ != kError && encoder_state_ != kUninitialized)
- NotifyError(error);
-
- encoder_state_ = kError;
-}
-
-void V4L2VideoEncodeAccelerator::RequestEncodingParametersChangeTask(
- uint32_t bitrate,
- uint32_t framerate) {
- DVLOG(3) << "RequestEncodingParametersChangeTask(): bitrate=" << bitrate
- << ", framerate=" << framerate;
- DCHECK_EQ(encoder_thread_.message_loop(), base::MessageLoop::current());
-
- if (bitrate < 1)
- bitrate = 1;
- if (framerate < 1)
- framerate = 1;
-
- std::vector<struct v4l2_ext_control> ctrls;
- struct v4l2_ext_control ctrl;
- memset(&ctrl, 0, sizeof(ctrl));
- ctrl.id = V4L2_CID_MPEG_VIDEO_BITRATE;
- ctrl.value = bitrate;
- ctrls.push_back(ctrl);
- if (!SetExtCtrls(ctrls)) {
- LOG(ERROR) << "Failed changing bitrate";
- NOTIFY_ERROR(kPlatformFailureError);
- return;
- }
-
- struct v4l2_streamparm parms;
- memset(&parms, 0, sizeof(parms));
- parms.type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE;
- // Note that we are provided "frames per second" but V4L2 expects "time per
- // frame"; hence we provide the reciprocal of the framerate here.
- parms.parm.output.timeperframe.numerator = 1;
- parms.parm.output.timeperframe.denominator = framerate;
- IOCTL_OR_ERROR_RETURN(VIDIOC_S_PARM, &parms);
-}
-
-bool V4L2VideoEncodeAccelerator::SetOutputFormat(
- media::VideoCodecProfile output_profile) {
- DCHECK(child_task_runner_->BelongsToCurrentThread());
- DCHECK(!input_streamon_);
- DCHECK(!output_streamon_);
-
- output_format_fourcc_ =
- V4L2Device::VideoCodecProfileToV4L2PixFmt(output_profile, false);
- if (!output_format_fourcc_) {
- LOG(ERROR) << "Initialize(): invalid output_profile=" << output_profile;
- return false;
- }
-
- output_buffer_byte_size_ = kOutputBufferSize;
-
- struct v4l2_format format;
- memset(&format, 0, sizeof(format));
- format.type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE;
- format.fmt.pix_mp.width = visible_size_.width();
- format.fmt.pix_mp.height = visible_size_.height();
- format.fmt.pix_mp.pixelformat = output_format_fourcc_;
- format.fmt.pix_mp.plane_fmt[0].sizeimage =
- base::checked_cast<__u32>(output_buffer_byte_size_);
- format.fmt.pix_mp.num_planes = 1;
- IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_S_FMT, &format);
-
- // Device might have adjusted the required output size.
- size_t adjusted_output_buffer_size =
- base::checked_cast<size_t>(format.fmt.pix_mp.plane_fmt[0].sizeimage);
- output_buffer_byte_size_ = adjusted_output_buffer_size;
-
- return true;
-}
-
-bool V4L2VideoEncodeAccelerator::NegotiateInputFormat(
- media::VideoPixelFormat input_format) {
- DVLOG(3) << "NegotiateInputFormat()";
- DCHECK(child_task_runner_->BelongsToCurrentThread());
- DCHECK(!input_streamon_);
- DCHECK(!output_streamon_);
-
- device_input_format_ = media::PIXEL_FORMAT_UNKNOWN;
- input_planes_count_ = 0;
-
- uint32_t input_format_fourcc =
- V4L2Device::VideoPixelFormatToV4L2PixFmt(input_format);
- if (!input_format_fourcc) {
- LOG(ERROR) << "Unsupported input format" << input_format_fourcc;
- return false;
- }
-
- size_t input_planes_count = media::VideoFrame::NumPlanes(input_format);
- DCHECK_LE(input_planes_count, static_cast<size_t>(VIDEO_MAX_PLANES));
-
- // First see if we the device can use the provided input_format directly.
- struct v4l2_format format;
- memset(&format, 0, sizeof(format));
- format.type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE;
- format.fmt.pix_mp.width = visible_size_.width();
- format.fmt.pix_mp.height = visible_size_.height();
- format.fmt.pix_mp.pixelformat = input_format_fourcc;
- format.fmt.pix_mp.num_planes = input_planes_count;
- if (device_->Ioctl(VIDIOC_S_FMT, &format) != 0) {
- // Error or format unsupported by device, try to negotiate a fallback.
- input_format_fourcc = device_->PreferredInputFormat();
- input_format =
- V4L2Device::V4L2PixFmtToVideoPixelFormat(input_format_fourcc);
- if (input_format == media::PIXEL_FORMAT_UNKNOWN) {
- LOG(ERROR) << "Unsupported input format" << input_format_fourcc;
- return false;
- }
-
- input_planes_count = media::VideoFrame::NumPlanes(input_format);
- DCHECK_LE(input_planes_count, static_cast<size_t>(VIDEO_MAX_PLANES));
-
- // Device might have adjusted parameters, reset them along with the format.
- memset(&format, 0, sizeof(format));
- format.type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE;
- format.fmt.pix_mp.width = visible_size_.width();
- format.fmt.pix_mp.height = visible_size_.height();
- format.fmt.pix_mp.pixelformat = input_format_fourcc;
- format.fmt.pix_mp.num_planes = input_planes_count;
- IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_S_FMT, &format);
- DCHECK_EQ(format.fmt.pix_mp.num_planes, input_planes_count);
- }
-
- // Take device-adjusted sizes for allocated size. If the size is adjusted
- // down, it means the input is too big and the hardware does not support it.
- input_allocated_size_ = V4L2Device::CodedSizeFromV4L2Format(format);
- if (!gfx::Rect(input_allocated_size_).Contains(gfx::Rect(visible_size_))) {
- DVLOG(1) << "Input size too big " << visible_size_.ToString()
- << ", adjusted to " << input_allocated_size_.ToString();
- return false;
- }
-
- device_input_format_ = input_format;
- input_planes_count_ = input_planes_count;
- return true;
-}
-
-bool V4L2VideoEncodeAccelerator::SetFormats(
- media::VideoPixelFormat input_format,
- media::VideoCodecProfile output_profile) {
- DVLOG(3) << "SetFormats()";
- DCHECK(child_task_runner_->BelongsToCurrentThread());
- DCHECK(!input_streamon_);
- DCHECK(!output_streamon_);
-
- if (!SetOutputFormat(output_profile))
- return false;
-
- if (!NegotiateInputFormat(input_format))
- return false;
-
- struct v4l2_crop crop;
- memset(&crop, 0, sizeof(crop));
- crop.type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE;
- crop.c.left = 0;
- crop.c.top = 0;
- crop.c.width = visible_size_.width();
- crop.c.height = visible_size_.height();
- IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_S_CROP, &crop);
-
- // The width and height might be adjusted by driver.
- // Need to read it back and set to visible_size_.
- if (device_->Ioctl(VIDIOC_G_CROP, &crop) != 0) {
- // Some devices haven't supported G_CROP yet, so treat the failure
- // non-fatal for now.
- // TODO(kcwu): NOTIFY_ERROR and return false after all devices support it.
- PLOG(WARNING) << "SetFormats(): ioctl() VIDIOC_G_CROP failed";
- return true;
- }
- visible_size_.SetSize(crop.c.width, crop.c.height);
- DVLOG(3) << "After adjusted by driver, visible_size_="
- << visible_size_.ToString();
-
- return true;
-}
-
-bool V4L2VideoEncodeAccelerator::SetExtCtrls(
- std::vector<struct v4l2_ext_control> ctrls) {
- struct v4l2_ext_controls ext_ctrls;
- memset(&ext_ctrls, 0, sizeof(ext_ctrls));
- ext_ctrls.ctrl_class = V4L2_CTRL_CLASS_MPEG;
- ext_ctrls.count = ctrls.size();
- ext_ctrls.controls = &ctrls[0];
- return device_->Ioctl(VIDIOC_S_EXT_CTRLS, &ext_ctrls) == 0;
-}
-
-bool V4L2VideoEncodeAccelerator::InitControls() {
- std::vector<struct v4l2_ext_control> ctrls;
- struct v4l2_ext_control ctrl;
-
- // Enable frame-level bitrate control. This is the only mandatory control.
- memset(&ctrl, 0, sizeof(ctrl));
- ctrl.id = V4L2_CID_MPEG_VIDEO_FRAME_RC_ENABLE;
- ctrl.value = 1;
- ctrls.push_back(ctrl);
- if (!SetExtCtrls(ctrls)) {
- LOG(ERROR) << "Failed enabling bitrate control";
- NOTIFY_ERROR(kPlatformFailureError);
- return false;
- }
-
- // Optional controls.
- ctrls.clear();
- if (output_format_fourcc_ == V4L2_PIX_FMT_H264) {
- // No B-frames, for lowest decoding latency.
- memset(&ctrl, 0, sizeof(ctrl));
- ctrl.id = V4L2_CID_MPEG_VIDEO_B_FRAMES;
- ctrl.value = 0;
- ctrls.push_back(ctrl);
-
- // Quantization parameter maximum value (for variable bitrate control).
- memset(&ctrl, 0, sizeof(ctrl));
- ctrl.id = V4L2_CID_MPEG_VIDEO_H264_MAX_QP;
- ctrl.value = 51;
- ctrls.push_back(ctrl);
-
- // Use H.264 level 4.0 to match the supported max resolution.
- memset(&ctrl, 0, sizeof(ctrl));
- ctrl.id = V4L2_CID_MPEG_VIDEO_H264_LEVEL;
- ctrl.value = V4L2_MPEG_VIDEO_H264_LEVEL_4_0;
- ctrls.push_back(ctrl);
-
- // Separate stream header so we can cache it and insert into the stream.
- memset(&ctrl, 0, sizeof(ctrl));
- ctrl.id = V4L2_CID_MPEG_VIDEO_HEADER_MODE;
- ctrl.value = V4L2_MPEG_VIDEO_HEADER_MODE_SEPARATE;
- ctrls.push_back(ctrl);
- }
-
- // Enable macroblock-level bitrate control.
- memset(&ctrl, 0, sizeof(ctrl));
- ctrl.id = V4L2_CID_MPEG_VIDEO_MB_RC_ENABLE;
- ctrl.value = 1;
- ctrls.push_back(ctrl);
-
- // Disable periodic key frames.
- memset(&ctrl, 0, sizeof(ctrl));
- ctrl.id = V4L2_CID_MPEG_VIDEO_GOP_SIZE;
- ctrl.value = 0;
- ctrls.push_back(ctrl);
-
- // Ignore return value as these controls are optional.
- SetExtCtrls(ctrls);
-
- // Optional Exynos specific controls.
- ctrls.clear();
- // Enable "tight" bitrate mode. For this to work properly, frame- and mb-level
- // bitrate controls have to be enabled as well.
- memset(&ctrl, 0, sizeof(ctrl));
- ctrl.id = V4L2_CID_MPEG_MFC51_VIDEO_RC_REACTION_COEFF;
- ctrl.value = 1;
- ctrls.push_back(ctrl);
-
- // Force bitrate control to average over a GOP (for tight bitrate
- // tolerance).
- memset(&ctrl, 0, sizeof(ctrl));
- ctrl.id = V4L2_CID_MPEG_MFC51_VIDEO_RC_FIXED_TARGET_BIT;
- ctrl.value = 1;
- ctrls.push_back(ctrl);
-
- // Ignore return value as these controls are optional.
- SetExtCtrls(ctrls);
-
- return true;
-}
-
-bool V4L2VideoEncodeAccelerator::CreateInputBuffers() {
- DVLOG(3) << "CreateInputBuffers()";
- // This function runs on encoder_thread_ after output buffers have been
- // provided by the client.
- DCHECK_EQ(encoder_thread_.message_loop(), base::MessageLoop::current());
- DCHECK(!input_streamon_);
-
- struct v4l2_requestbuffers reqbufs;
- memset(&reqbufs, 0, sizeof(reqbufs));
- // Driver will modify to the appropriate number of buffers.
- reqbufs.count = 1;
- reqbufs.type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE;
- // TODO(posciak): Once we start doing zero-copy, we should decide based on
- // the current pipeline setup which memory type to use. This should probably
- // be decided based on an argument to Initialize().
- if (image_processor_.get())
- input_memory_type_ = V4L2_MEMORY_DMABUF;
- else
- input_memory_type_ = V4L2_MEMORY_USERPTR;
-
- reqbufs.memory = input_memory_type_;
- IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_REQBUFS, &reqbufs);
-
- DCHECK(input_buffer_map_.empty());
- input_buffer_map_.resize(reqbufs.count);
- for (size_t i = 0; i < input_buffer_map_.size(); ++i)
- free_input_buffers_.push_back(i);
-
- return true;
-}
-
-bool V4L2VideoEncodeAccelerator::CreateOutputBuffers() {
- DVLOG(3) << "CreateOutputBuffers()";
- DCHECK(child_task_runner_->BelongsToCurrentThread());
- DCHECK(!output_streamon_);
-
- struct v4l2_requestbuffers reqbufs;
- memset(&reqbufs, 0, sizeof(reqbufs));
- reqbufs.count = kOutputBufferCount;
- reqbufs.type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE;
- reqbufs.memory = V4L2_MEMORY_MMAP;
- IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_REQBUFS, &reqbufs);
-
- DCHECK(output_buffer_map_.empty());
- output_buffer_map_.resize(reqbufs.count);
- for (size_t i = 0; i < output_buffer_map_.size(); ++i) {
- struct v4l2_plane planes[1];
- struct v4l2_buffer buffer;
- memset(&buffer, 0, sizeof(buffer));
- memset(planes, 0, sizeof(planes));
- buffer.index = i;
- buffer.type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE;
- buffer.memory = V4L2_MEMORY_MMAP;
- buffer.m.planes = planes;
- buffer.length = arraysize(planes);
- IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_QUERYBUF, &buffer);
- void* address = device_->Mmap(NULL,
- buffer.m.planes[0].length,
- PROT_READ | PROT_WRITE,
- MAP_SHARED,
- buffer.m.planes[0].m.mem_offset);
- if (address == MAP_FAILED) {
- PLOG(ERROR) << "CreateOutputBuffers(): mmap() failed";
- return false;
- }
- output_buffer_map_[i].address = address;
- output_buffer_map_[i].length = buffer.m.planes[0].length;
- free_output_buffers_.push_back(i);
- }
-
- return true;
-}
-
-void V4L2VideoEncodeAccelerator::DestroyInputBuffers() {
- DVLOG(3) << "DestroyInputBuffers()";
- DCHECK(child_task_runner_->BelongsToCurrentThread());
- DCHECK(!input_streamon_);
-
- struct v4l2_requestbuffers reqbufs;
- memset(&reqbufs, 0, sizeof(reqbufs));
- reqbufs.count = 0;
- reqbufs.type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE;
- reqbufs.memory = input_memory_type_;
- IOCTL_OR_LOG_ERROR(VIDIOC_REQBUFS, &reqbufs);
-
- input_buffer_map_.clear();
- free_input_buffers_.clear();
-}
-
-void V4L2VideoEncodeAccelerator::DestroyOutputBuffers() {
- DVLOG(3) << "DestroyOutputBuffers()";
- DCHECK(child_task_runner_->BelongsToCurrentThread());
- DCHECK(!output_streamon_);
-
- for (size_t i = 0; i < output_buffer_map_.size(); ++i) {
- if (output_buffer_map_[i].address != NULL)
- device_->Munmap(output_buffer_map_[i].address,
- output_buffer_map_[i].length);
- }
-
- struct v4l2_requestbuffers reqbufs;
- memset(&reqbufs, 0, sizeof(reqbufs));
- reqbufs.count = 0;
- reqbufs.type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE;
- reqbufs.memory = V4L2_MEMORY_MMAP;
- IOCTL_OR_LOG_ERROR(VIDIOC_REQBUFS, &reqbufs);
-
- output_buffer_map_.clear();
- free_output_buffers_.clear();
-}
-
-} // namespace content

Powered by Google App Engine
This is Rietveld 408576698