| OLD | NEW |
| (Empty) |
| 1 // Copyright 2014 The Chromium Authors. All rights reserved. | |
| 2 // Use of this source code is governed by a BSD-style license that can be | |
| 3 // found in the LICENSE file. | |
| 4 | |
| 5 #include "content/common/gpu/media/v4l2_video_encode_accelerator.h" | |
| 6 | |
| 7 #include <fcntl.h> | |
| 8 #include <linux/videodev2.h> | |
| 9 #include <poll.h> | |
| 10 #include <string.h> | |
| 11 #include <sys/eventfd.h> | |
| 12 #include <sys/ioctl.h> | |
| 13 #include <sys/mman.h> | |
| 14 | |
| 15 #include <utility> | |
| 16 | |
| 17 #include "base/callback.h" | |
| 18 #include "base/command_line.h" | |
| 19 #include "base/macros.h" | |
| 20 #include "base/numerics/safe_conversions.h" | |
| 21 #include "base/thread_task_runner_handle.h" | |
| 22 #include "base/trace_event/trace_event.h" | |
| 23 #include "content/common/gpu/media/shared_memory_region.h" | |
| 24 #include "media/base/bind_to_current_loop.h" | |
| 25 #include "media/base/bitstream_buffer.h" | |
| 26 | |
| 27 #define NOTIFY_ERROR(x) \ | |
| 28 do { \ | |
| 29 LOG(ERROR) << "Setting error state:" << x; \ | |
| 30 SetErrorState(x); \ | |
| 31 } while (0) | |
| 32 | |
| 33 #define IOCTL_OR_ERROR_RETURN_VALUE(type, arg, value, type_str) \ | |
| 34 do { \ | |
| 35 if (device_->Ioctl(type, arg) != 0) { \ | |
| 36 PLOG(ERROR) << __func__ << "(): ioctl() failed: " << type_str; \ | |
| 37 NOTIFY_ERROR(kPlatformFailureError); \ | |
| 38 return value; \ | |
| 39 } \ | |
| 40 } while (0) | |
| 41 | |
| 42 #define IOCTL_OR_ERROR_RETURN(type, arg) \ | |
| 43 IOCTL_OR_ERROR_RETURN_VALUE(type, arg, ((void)0), #type) | |
| 44 | |
| 45 #define IOCTL_OR_ERROR_RETURN_FALSE(type, arg) \ | |
| 46 IOCTL_OR_ERROR_RETURN_VALUE(type, arg, false, #type) | |
| 47 | |
| 48 #define IOCTL_OR_LOG_ERROR(type, arg) \ | |
| 49 do { \ | |
| 50 if (device_->Ioctl(type, arg) != 0) \ | |
| 51 PLOG(ERROR) << __func__ << "(): ioctl() failed: " << #type; \ | |
| 52 } while (0) | |
| 53 | |
| 54 namespace content { | |
| 55 | |
| 56 struct V4L2VideoEncodeAccelerator::BitstreamBufferRef { | |
| 57 BitstreamBufferRef(int32_t id, std::unique_ptr<SharedMemoryRegion> shm) | |
| 58 : id(id), shm(std::move(shm)) {} | |
| 59 const int32_t id; | |
| 60 const std::unique_ptr<SharedMemoryRegion> shm; | |
| 61 }; | |
| 62 | |
| 63 V4L2VideoEncodeAccelerator::InputRecord::InputRecord() : at_device(false) { | |
| 64 } | |
| 65 | |
| 66 V4L2VideoEncodeAccelerator::InputRecord::~InputRecord() { | |
| 67 } | |
| 68 | |
| 69 V4L2VideoEncodeAccelerator::OutputRecord::OutputRecord() | |
| 70 : at_device(false), address(NULL), length(0) { | |
| 71 } | |
| 72 | |
| 73 V4L2VideoEncodeAccelerator::OutputRecord::~OutputRecord() { | |
| 74 } | |
| 75 | |
| 76 V4L2VideoEncodeAccelerator::ImageProcessorInputRecord:: | |
| 77 ImageProcessorInputRecord() | |
| 78 : force_keyframe(false) {} | |
| 79 | |
| 80 V4L2VideoEncodeAccelerator::ImageProcessorInputRecord:: | |
| 81 ~ImageProcessorInputRecord() {} | |
| 82 | |
| 83 V4L2VideoEncodeAccelerator::V4L2VideoEncodeAccelerator( | |
| 84 const scoped_refptr<V4L2Device>& device) | |
| 85 : child_task_runner_(base::ThreadTaskRunnerHandle::Get()), | |
| 86 output_buffer_byte_size_(0), | |
| 87 device_input_format_(media::PIXEL_FORMAT_UNKNOWN), | |
| 88 input_planes_count_(0), | |
| 89 output_format_fourcc_(0), | |
| 90 encoder_state_(kUninitialized), | |
| 91 stream_header_size_(0), | |
| 92 device_(device), | |
| 93 input_streamon_(false), | |
| 94 input_buffer_queued_count_(0), | |
| 95 input_memory_type_(V4L2_MEMORY_USERPTR), | |
| 96 output_streamon_(false), | |
| 97 output_buffer_queued_count_(0), | |
| 98 encoder_thread_("V4L2EncoderThread"), | |
| 99 device_poll_thread_("V4L2EncoderDevicePollThread"), | |
| 100 weak_this_ptr_factory_(this) { | |
| 101 weak_this_ = weak_this_ptr_factory_.GetWeakPtr(); | |
| 102 } | |
| 103 | |
| 104 V4L2VideoEncodeAccelerator::~V4L2VideoEncodeAccelerator() { | |
| 105 DCHECK(!encoder_thread_.IsRunning()); | |
| 106 DCHECK(!device_poll_thread_.IsRunning()); | |
| 107 DVLOG(4) << __func__; | |
| 108 | |
| 109 DestroyInputBuffers(); | |
| 110 DestroyOutputBuffers(); | |
| 111 } | |
| 112 | |
| 113 bool V4L2VideoEncodeAccelerator::Initialize( | |
| 114 media::VideoPixelFormat input_format, | |
| 115 const gfx::Size& input_visible_size, | |
| 116 media::VideoCodecProfile output_profile, | |
| 117 uint32_t initial_bitrate, | |
| 118 Client* client) { | |
| 119 DVLOG(3) << __func__ | |
| 120 << ": input_format=" << media::VideoPixelFormatToString(input_format) | |
| 121 << ", input_visible_size=" << input_visible_size.ToString() | |
| 122 << ", output_profile=" << output_profile | |
| 123 << ", initial_bitrate=" << initial_bitrate; | |
| 124 | |
| 125 visible_size_ = input_visible_size; | |
| 126 | |
| 127 client_ptr_factory_.reset(new base::WeakPtrFactory<Client>(client)); | |
| 128 client_ = client_ptr_factory_->GetWeakPtr(); | |
| 129 | |
| 130 DCHECK(child_task_runner_->BelongsToCurrentThread()); | |
| 131 DCHECK_EQ(encoder_state_, kUninitialized); | |
| 132 | |
| 133 struct v4l2_capability caps; | |
| 134 memset(&caps, 0, sizeof(caps)); | |
| 135 const __u32 kCapsRequired = V4L2_CAP_VIDEO_M2M_MPLANE | V4L2_CAP_STREAMING; | |
| 136 IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_QUERYCAP, &caps); | |
| 137 if ((caps.capabilities & kCapsRequired) != kCapsRequired) { | |
| 138 LOG(ERROR) << "Initialize(): ioctl() failed: VIDIOC_QUERYCAP: " | |
| 139 "caps check failed: 0x" << std::hex << caps.capabilities; | |
| 140 return false; | |
| 141 } | |
| 142 | |
| 143 if (!SetFormats(input_format, output_profile)) { | |
| 144 DLOG(ERROR) << "Failed setting up formats"; | |
| 145 return false; | |
| 146 } | |
| 147 | |
| 148 if (input_format != device_input_format_) { | |
| 149 DVLOG(1) << "Input format not supported by the HW, will convert to " | |
| 150 << media::VideoPixelFormatToString(device_input_format_); | |
| 151 | |
| 152 scoped_refptr<V4L2Device> device = | |
| 153 V4L2Device::Create(V4L2Device::kImageProcessor); | |
| 154 image_processor_.reset(new V4L2ImageProcessor(device)); | |
| 155 | |
| 156 // Convert from input_format to device_input_format_, keeping the size | |
| 157 // at visible_size_ and requiring the output buffers to be of at least | |
| 158 // input_allocated_size_. Unretained is safe because |this| owns image | |
| 159 // processor and there will be no callbacks after processor destroys. | |
| 160 if (!image_processor_->Initialize( | |
| 161 input_format, device_input_format_, V4L2_MEMORY_USERPTR, | |
| 162 visible_size_, visible_size_, visible_size_, input_allocated_size_, | |
| 163 kImageProcBufferCount, | |
| 164 base::Bind(&V4L2VideoEncodeAccelerator::ImageProcessorError, | |
| 165 base::Unretained(this)))) { | |
| 166 LOG(ERROR) << "Failed initializing image processor"; | |
| 167 return false; | |
| 168 } | |
| 169 // The output of image processor is the input of encoder. Output coded | |
| 170 // width of processor must be the same as input coded width of encoder. | |
| 171 // Output coded height of processor can be larger but not smaller than the | |
| 172 // input coded height of encoder. For example, suppose input size of encoder | |
| 173 // is 320x193. It is OK if the output of processor is 320x208. | |
| 174 if (image_processor_->output_allocated_size().width() != | |
| 175 input_allocated_size_.width() || | |
| 176 image_processor_->output_allocated_size().height() < | |
| 177 input_allocated_size_.height()) { | |
| 178 LOG(ERROR) << "Invalid image processor output coded size " | |
| 179 << image_processor_->output_allocated_size().ToString() | |
| 180 << ", encode input coded size is " | |
| 181 << input_allocated_size_.ToString(); | |
| 182 return false; | |
| 183 } | |
| 184 | |
| 185 for (int i = 0; i < kImageProcBufferCount; i++) { | |
| 186 std::vector<base::ScopedFD> fds = | |
| 187 image_processor_->GetDmabufsForOutputBuffer(i); | |
| 188 if (fds.size() == 0) { | |
| 189 LOG(ERROR) << __func__ << ": failed to get fds of image processor."; | |
| 190 return false; | |
| 191 } | |
| 192 image_processor_output_buffer_map_.push_back(std::move(fds)); | |
| 193 free_image_processor_output_buffers_.push_back(i); | |
| 194 } | |
| 195 } | |
| 196 | |
| 197 if (!InitControls()) | |
| 198 return false; | |
| 199 | |
| 200 if (!CreateOutputBuffers()) | |
| 201 return false; | |
| 202 | |
| 203 if (!encoder_thread_.Start()) { | |
| 204 LOG(ERROR) << "Initialize(): encoder thread failed to start"; | |
| 205 return false; | |
| 206 } | |
| 207 | |
| 208 RequestEncodingParametersChange(initial_bitrate, kInitialFramerate); | |
| 209 | |
| 210 encoder_state_ = kInitialized; | |
| 211 | |
| 212 child_task_runner_->PostTask( | |
| 213 FROM_HERE, | |
| 214 base::Bind(&Client::RequireBitstreamBuffers, client_, kInputBufferCount, | |
| 215 image_processor_.get() | |
| 216 ? image_processor_->input_allocated_size() | |
| 217 : input_allocated_size_, | |
| 218 output_buffer_byte_size_)); | |
| 219 return true; | |
| 220 } | |
| 221 | |
| 222 void V4L2VideoEncodeAccelerator::ImageProcessorError() { | |
| 223 LOG(ERROR) << "Image processor error"; | |
| 224 NOTIFY_ERROR(kPlatformFailureError); | |
| 225 } | |
| 226 | |
| 227 void V4L2VideoEncodeAccelerator::Encode( | |
| 228 const scoped_refptr<media::VideoFrame>& frame, | |
| 229 bool force_keyframe) { | |
| 230 DVLOG(3) << "Encode(): force_keyframe=" << force_keyframe; | |
| 231 DCHECK(child_task_runner_->BelongsToCurrentThread()); | |
| 232 | |
| 233 if (image_processor_) { | |
| 234 if (free_image_processor_output_buffers_.size() > 0) { | |
| 235 int output_buffer_index = free_image_processor_output_buffers_.back(); | |
| 236 free_image_processor_output_buffers_.pop_back(); | |
| 237 // Unretained is safe because |this| owns image processor and there will | |
| 238 // be no callbacks after processor destroys. | |
| 239 image_processor_->Process( | |
| 240 frame, output_buffer_index, | |
| 241 base::Bind(&V4L2VideoEncodeAccelerator::FrameProcessed, | |
| 242 base::Unretained(this), force_keyframe, | |
| 243 frame->timestamp())); | |
| 244 } else { | |
| 245 ImageProcessorInputRecord record; | |
| 246 record.frame = frame; | |
| 247 record.force_keyframe = force_keyframe; | |
| 248 image_processor_input_queue_.push(record); | |
| 249 } | |
| 250 } else { | |
| 251 encoder_thread_.message_loop()->PostTask( | |
| 252 FROM_HERE, | |
| 253 base::Bind(&V4L2VideoEncodeAccelerator::EncodeTask, | |
| 254 base::Unretained(this), | |
| 255 frame, | |
| 256 force_keyframe)); | |
| 257 } | |
| 258 } | |
| 259 | |
| 260 void V4L2VideoEncodeAccelerator::UseOutputBitstreamBuffer( | |
| 261 const media::BitstreamBuffer& buffer) { | |
| 262 DVLOG(3) << "UseOutputBitstreamBuffer(): id=" << buffer.id(); | |
| 263 DCHECK(child_task_runner_->BelongsToCurrentThread()); | |
| 264 | |
| 265 if (buffer.size() < output_buffer_byte_size_) { | |
| 266 NOTIFY_ERROR(kInvalidArgumentError); | |
| 267 return; | |
| 268 } | |
| 269 | |
| 270 std::unique_ptr<SharedMemoryRegion> shm( | |
| 271 new SharedMemoryRegion(buffer, false)); | |
| 272 if (!shm->Map()) { | |
| 273 NOTIFY_ERROR(kPlatformFailureError); | |
| 274 return; | |
| 275 } | |
| 276 | |
| 277 std::unique_ptr<BitstreamBufferRef> buffer_ref( | |
| 278 new BitstreamBufferRef(buffer.id(), std::move(shm))); | |
| 279 encoder_thread_.message_loop()->PostTask( | |
| 280 FROM_HERE, | |
| 281 base::Bind(&V4L2VideoEncodeAccelerator::UseOutputBitstreamBufferTask, | |
| 282 base::Unretained(this), | |
| 283 base::Passed(&buffer_ref))); | |
| 284 } | |
| 285 | |
| 286 void V4L2VideoEncodeAccelerator::RequestEncodingParametersChange( | |
| 287 uint32_t bitrate, | |
| 288 uint32_t framerate) { | |
| 289 DVLOG(3) << "RequestEncodingParametersChange(): bitrate=" << bitrate | |
| 290 << ", framerate=" << framerate; | |
| 291 DCHECK(child_task_runner_->BelongsToCurrentThread()); | |
| 292 | |
| 293 encoder_thread_.message_loop()->PostTask( | |
| 294 FROM_HERE, | |
| 295 base::Bind( | |
| 296 &V4L2VideoEncodeAccelerator::RequestEncodingParametersChangeTask, | |
| 297 base::Unretained(this), | |
| 298 bitrate, | |
| 299 framerate)); | |
| 300 } | |
| 301 | |
| 302 void V4L2VideoEncodeAccelerator::Destroy() { | |
| 303 DVLOG(3) << "Destroy()"; | |
| 304 DCHECK(child_task_runner_->BelongsToCurrentThread()); | |
| 305 | |
| 306 // We're destroying; cancel all callbacks. | |
| 307 client_ptr_factory_.reset(); | |
| 308 weak_this_ptr_factory_.InvalidateWeakPtrs(); | |
| 309 | |
| 310 if (image_processor_.get()) | |
| 311 image_processor_.release()->Destroy(); | |
| 312 | |
| 313 // If the encoder thread is running, destroy using posted task. | |
| 314 if (encoder_thread_.IsRunning()) { | |
| 315 encoder_thread_.message_loop()->PostTask( | |
| 316 FROM_HERE, | |
| 317 base::Bind(&V4L2VideoEncodeAccelerator::DestroyTask, | |
| 318 base::Unretained(this))); | |
| 319 // DestroyTask() will put the encoder into kError state and cause all tasks | |
| 320 // to no-op. | |
| 321 encoder_thread_.Stop(); | |
| 322 } else { | |
| 323 // Otherwise, call the destroy task directly. | |
| 324 DestroyTask(); | |
| 325 } | |
| 326 | |
| 327 // Set to kError state just in case. | |
| 328 encoder_state_ = kError; | |
| 329 | |
| 330 delete this; | |
| 331 } | |
| 332 | |
| 333 media::VideoEncodeAccelerator::SupportedProfiles | |
| 334 V4L2VideoEncodeAccelerator::GetSupportedProfiles() { | |
| 335 SupportedProfiles profiles; | |
| 336 SupportedProfile profile; | |
| 337 profile.max_framerate_numerator = 30; | |
| 338 profile.max_framerate_denominator = 1; | |
| 339 | |
| 340 gfx::Size min_resolution; | |
| 341 v4l2_fmtdesc fmtdesc; | |
| 342 memset(&fmtdesc, 0, sizeof(fmtdesc)); | |
| 343 fmtdesc.type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE; | |
| 344 for (; device_->Ioctl(VIDIOC_ENUM_FMT, &fmtdesc) == 0; ++fmtdesc.index) { | |
| 345 device_->GetSupportedResolution(fmtdesc.pixelformat, | |
| 346 &min_resolution, &profile.max_resolution); | |
| 347 switch (fmtdesc.pixelformat) { | |
| 348 case V4L2_PIX_FMT_H264: | |
| 349 profile.profile = media::H264PROFILE_MAIN; | |
| 350 profiles.push_back(profile); | |
| 351 break; | |
| 352 case V4L2_PIX_FMT_VP8: | |
| 353 profile.profile = media::VP8PROFILE_ANY; | |
| 354 profiles.push_back(profile); | |
| 355 break; | |
| 356 case V4L2_PIX_FMT_VP9: | |
| 357 profile.profile = media::VP9PROFILE_PROFILE0; | |
| 358 profiles.push_back(profile); | |
| 359 profile.profile = media::VP9PROFILE_PROFILE1; | |
| 360 profiles.push_back(profile); | |
| 361 profile.profile = media::VP9PROFILE_PROFILE2; | |
| 362 profiles.push_back(profile); | |
| 363 profile.profile = media::VP9PROFILE_PROFILE3; | |
| 364 profiles.push_back(profile); | |
| 365 break; | |
| 366 } | |
| 367 } | |
| 368 | |
| 369 return profiles; | |
| 370 } | |
| 371 | |
| 372 void V4L2VideoEncodeAccelerator::FrameProcessed(bool force_keyframe, | |
| 373 base::TimeDelta timestamp, | |
| 374 int output_buffer_index) { | |
| 375 DCHECK(child_task_runner_->BelongsToCurrentThread()); | |
| 376 DVLOG(3) << "FrameProcessed(): force_keyframe=" << force_keyframe | |
| 377 << ", output_buffer_index=" << output_buffer_index; | |
| 378 DCHECK_GE(output_buffer_index, 0); | |
| 379 DCHECK_LT(static_cast<size_t>(output_buffer_index), | |
| 380 image_processor_output_buffer_map_.size()); | |
| 381 | |
| 382 std::vector<base::ScopedFD>& scoped_fds = | |
| 383 image_processor_output_buffer_map_[output_buffer_index]; | |
| 384 std::vector<int> fds; | |
| 385 for (auto& fd : scoped_fds) { | |
| 386 fds.push_back(fd.get()); | |
| 387 } | |
| 388 scoped_refptr<media::VideoFrame> output_frame = | |
| 389 media::VideoFrame::WrapExternalDmabufs( | |
| 390 device_input_format_, image_processor_->output_allocated_size(), | |
| 391 gfx::Rect(visible_size_), visible_size_, fds, timestamp); | |
| 392 if (!output_frame) { | |
| 393 NOTIFY_ERROR(kPlatformFailureError); | |
| 394 return; | |
| 395 } | |
| 396 output_frame->AddDestructionObserver(media::BindToCurrentLoop( | |
| 397 base::Bind(&V4L2VideoEncodeAccelerator::ReuseImageProcessorOutputBuffer, | |
| 398 weak_this_, output_buffer_index))); | |
| 399 | |
| 400 encoder_thread_.message_loop()->PostTask( | |
| 401 FROM_HERE, | |
| 402 base::Bind(&V4L2VideoEncodeAccelerator::EncodeTask, | |
| 403 base::Unretained(this), output_frame, force_keyframe)); | |
| 404 } | |
| 405 | |
| 406 void V4L2VideoEncodeAccelerator::ReuseImageProcessorOutputBuffer( | |
| 407 int output_buffer_index) { | |
| 408 DCHECK(child_task_runner_->BelongsToCurrentThread()); | |
| 409 DVLOG(3) << __func__ << ": output_buffer_index=" << output_buffer_index; | |
| 410 free_image_processor_output_buffers_.push_back(output_buffer_index); | |
| 411 if (!image_processor_input_queue_.empty()) { | |
| 412 ImageProcessorInputRecord record = image_processor_input_queue_.front(); | |
| 413 image_processor_input_queue_.pop(); | |
| 414 Encode(record.frame, record.force_keyframe); | |
| 415 } | |
| 416 } | |
| 417 | |
| 418 void V4L2VideoEncodeAccelerator::EncodeTask( | |
| 419 const scoped_refptr<media::VideoFrame>& frame, | |
| 420 bool force_keyframe) { | |
| 421 DVLOG(3) << "EncodeTask(): force_keyframe=" << force_keyframe; | |
| 422 DCHECK_EQ(encoder_thread_.message_loop(), base::MessageLoop::current()); | |
| 423 DCHECK_NE(encoder_state_, kUninitialized); | |
| 424 | |
| 425 if (encoder_state_ == kError) { | |
| 426 DVLOG(2) << "EncodeTask(): early out: kError state"; | |
| 427 return; | |
| 428 } | |
| 429 | |
| 430 encoder_input_queue_.push(frame); | |
| 431 Enqueue(); | |
| 432 | |
| 433 if (force_keyframe) { | |
| 434 // TODO(posciak): this presently makes for slightly imprecise encoding | |
| 435 // parameters updates. To precisely align the parameter updates with the | |
| 436 // incoming input frame, we should queue the parameters together with the | |
| 437 // frame onto encoder_input_queue_ and apply them when the input is about | |
| 438 // to be queued to the codec. | |
| 439 std::vector<struct v4l2_ext_control> ctrls; | |
| 440 struct v4l2_ext_control ctrl; | |
| 441 memset(&ctrl, 0, sizeof(ctrl)); | |
| 442 ctrl.id = V4L2_CID_MPEG_VIDEO_FORCE_KEY_FRAME; | |
| 443 ctrls.push_back(ctrl); | |
| 444 if (!SetExtCtrls(ctrls)) { | |
| 445 // Some platforms still use the old control. Fallback before they are | |
| 446 // updated. | |
| 447 ctrls.clear(); | |
| 448 memset(&ctrl, 0, sizeof(ctrl)); | |
| 449 ctrl.id = V4L2_CID_MPEG_MFC51_VIDEO_FORCE_FRAME_TYPE; | |
| 450 ctrl.value = V4L2_MPEG_MFC51_VIDEO_FORCE_FRAME_TYPE_I_FRAME; | |
| 451 ctrls.push_back(ctrl); | |
| 452 if (!SetExtCtrls(ctrls)) { | |
| 453 LOG(ERROR) << "Failed requesting keyframe"; | |
| 454 NOTIFY_ERROR(kPlatformFailureError); | |
| 455 return; | |
| 456 } | |
| 457 } | |
| 458 } | |
| 459 } | |
| 460 | |
| 461 void V4L2VideoEncodeAccelerator::UseOutputBitstreamBufferTask( | |
| 462 std::unique_ptr<BitstreamBufferRef> buffer_ref) { | |
| 463 DVLOG(3) << "UseOutputBitstreamBufferTask(): id=" << buffer_ref->id; | |
| 464 DCHECK_EQ(encoder_thread_.message_loop(), base::MessageLoop::current()); | |
| 465 | |
| 466 encoder_output_queue_.push_back( | |
| 467 linked_ptr<BitstreamBufferRef>(buffer_ref.release())); | |
| 468 Enqueue(); | |
| 469 | |
| 470 if (encoder_state_ == kInitialized) { | |
| 471 // Finish setting up our OUTPUT queue. See: Initialize(). | |
| 472 // VIDIOC_REQBUFS on OUTPUT queue. | |
| 473 if (!CreateInputBuffers()) | |
| 474 return; | |
| 475 if (!StartDevicePoll()) | |
| 476 return; | |
| 477 encoder_state_ = kEncoding; | |
| 478 } | |
| 479 } | |
| 480 | |
| 481 void V4L2VideoEncodeAccelerator::DestroyTask() { | |
| 482 DVLOG(3) << "DestroyTask()"; | |
| 483 | |
| 484 // DestroyTask() should run regardless of encoder_state_. | |
| 485 | |
| 486 // Stop streaming and the device_poll_thread_. | |
| 487 StopDevicePoll(); | |
| 488 | |
| 489 // Set our state to kError, and early-out all tasks. | |
| 490 encoder_state_ = kError; | |
| 491 } | |
| 492 | |
| 493 void V4L2VideoEncodeAccelerator::ServiceDeviceTask() { | |
| 494 DVLOG(3) << "ServiceDeviceTask()"; | |
| 495 DCHECK_EQ(encoder_thread_.message_loop(), base::MessageLoop::current()); | |
| 496 DCHECK_NE(encoder_state_, kUninitialized); | |
| 497 DCHECK_NE(encoder_state_, kInitialized); | |
| 498 | |
| 499 if (encoder_state_ == kError) { | |
| 500 DVLOG(2) << "ServiceDeviceTask(): early out: kError state"; | |
| 501 return; | |
| 502 } | |
| 503 | |
| 504 Dequeue(); | |
| 505 Enqueue(); | |
| 506 | |
| 507 // Clear the interrupt fd. | |
| 508 if (!device_->ClearDevicePollInterrupt()) | |
| 509 return; | |
| 510 | |
| 511 // Device can be polled as soon as either input or output buffers are queued. | |
| 512 bool poll_device = | |
| 513 (input_buffer_queued_count_ + output_buffer_queued_count_ > 0); | |
| 514 | |
| 515 // ServiceDeviceTask() should only ever be scheduled from DevicePollTask(), | |
| 516 // so either: | |
| 517 // * device_poll_thread_ is running normally | |
| 518 // * device_poll_thread_ scheduled us, but then a DestroyTask() shut it down, | |
| 519 // in which case we're in kError state, and we should have early-outed | |
| 520 // already. | |
| 521 DCHECK(device_poll_thread_.message_loop()); | |
| 522 // Queue the DevicePollTask() now. | |
| 523 device_poll_thread_.message_loop()->PostTask( | |
| 524 FROM_HERE, | |
| 525 base::Bind(&V4L2VideoEncodeAccelerator::DevicePollTask, | |
| 526 base::Unretained(this), | |
| 527 poll_device)); | |
| 528 | |
| 529 DVLOG(2) << __func__ << ": buffer counts: ENC[" | |
| 530 << encoder_input_queue_.size() << "] => DEVICE[" | |
| 531 << free_input_buffers_.size() << "+" | |
| 532 << input_buffer_queued_count_ << "/" | |
| 533 << input_buffer_map_.size() << "->" | |
| 534 << free_output_buffers_.size() << "+" | |
| 535 << output_buffer_queued_count_ << "/" | |
| 536 << output_buffer_map_.size() << "] => OUT[" | |
| 537 << encoder_output_queue_.size() << "]"; | |
| 538 } | |
| 539 | |
| 540 void V4L2VideoEncodeAccelerator::Enqueue() { | |
| 541 DCHECK_EQ(encoder_thread_.message_loop(), base::MessageLoop::current()); | |
| 542 | |
| 543 DVLOG(3) << "Enqueue() " | |
| 544 << "free_input_buffers: " << free_input_buffers_.size() | |
| 545 << "input_queue: " << encoder_input_queue_.size(); | |
| 546 | |
| 547 // Enqueue all the inputs we can. | |
| 548 const int old_inputs_queued = input_buffer_queued_count_; | |
| 549 // while (!ready_input_buffers_.empty()) { | |
| 550 while (!encoder_input_queue_.empty() && !free_input_buffers_.empty()) { | |
| 551 if (!EnqueueInputRecord()) | |
| 552 return; | |
| 553 } | |
| 554 if (old_inputs_queued == 0 && input_buffer_queued_count_ != 0) { | |
| 555 // We just started up a previously empty queue. | |
| 556 // Queue state changed; signal interrupt. | |
| 557 if (!device_->SetDevicePollInterrupt()) | |
| 558 return; | |
| 559 // Start VIDIOC_STREAMON if we haven't yet. | |
| 560 if (!input_streamon_) { | |
| 561 __u32 type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE; | |
| 562 IOCTL_OR_ERROR_RETURN(VIDIOC_STREAMON, &type); | |
| 563 input_streamon_ = true; | |
| 564 } | |
| 565 } | |
| 566 | |
| 567 // Enqueue all the outputs we can. | |
| 568 const int old_outputs_queued = output_buffer_queued_count_; | |
| 569 while (!free_output_buffers_.empty() && !encoder_output_queue_.empty()) { | |
| 570 if (!EnqueueOutputRecord()) | |
| 571 return; | |
| 572 } | |
| 573 if (old_outputs_queued == 0 && output_buffer_queued_count_ != 0) { | |
| 574 // We just started up a previously empty queue. | |
| 575 // Queue state changed; signal interrupt. | |
| 576 if (!device_->SetDevicePollInterrupt()) | |
| 577 return; | |
| 578 // Start VIDIOC_STREAMON if we haven't yet. | |
| 579 if (!output_streamon_) { | |
| 580 __u32 type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE; | |
| 581 IOCTL_OR_ERROR_RETURN(VIDIOC_STREAMON, &type); | |
| 582 output_streamon_ = true; | |
| 583 } | |
| 584 } | |
| 585 } | |
| 586 | |
| 587 void V4L2VideoEncodeAccelerator::Dequeue() { | |
| 588 DVLOG(3) << "Dequeue()"; | |
| 589 DCHECK_EQ(encoder_thread_.message_loop(), base::MessageLoop::current()); | |
| 590 | |
| 591 // Dequeue completed input (VIDEO_OUTPUT) buffers, and recycle to the free | |
| 592 // list. | |
| 593 struct v4l2_buffer dqbuf; | |
| 594 struct v4l2_plane planes[VIDEO_MAX_PLANES]; | |
| 595 while (input_buffer_queued_count_ > 0) { | |
| 596 DVLOG(4) << "inputs queued: " << input_buffer_queued_count_; | |
| 597 DCHECK(input_streamon_); | |
| 598 memset(&dqbuf, 0, sizeof(dqbuf)); | |
| 599 memset(&planes, 0, sizeof(planes)); | |
| 600 dqbuf.type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE; | |
| 601 dqbuf.memory = input_memory_type_; | |
| 602 dqbuf.m.planes = planes; | |
| 603 dqbuf.length = input_planes_count_; | |
| 604 if (device_->Ioctl(VIDIOC_DQBUF, &dqbuf) != 0) { | |
| 605 if (errno == EAGAIN) { | |
| 606 // EAGAIN if we're just out of buffers to dequeue. | |
| 607 break; | |
| 608 } | |
| 609 PLOG(ERROR) << "Dequeue(): ioctl() failed: VIDIOC_DQBUF"; | |
| 610 NOTIFY_ERROR(kPlatformFailureError); | |
| 611 return; | |
| 612 } | |
| 613 InputRecord& input_record = input_buffer_map_[dqbuf.index]; | |
| 614 DCHECK(input_record.at_device); | |
| 615 input_record.at_device = false; | |
| 616 | |
| 617 input_record.frame = NULL; | |
| 618 free_input_buffers_.push_back(dqbuf.index); | |
| 619 input_buffer_queued_count_--; | |
| 620 } | |
| 621 | |
| 622 // Dequeue completed output (VIDEO_CAPTURE) buffers, and recycle to the | |
| 623 // free list. Notify the client that an output buffer is complete. | |
| 624 while (output_buffer_queued_count_ > 0) { | |
| 625 DCHECK(output_streamon_); | |
| 626 memset(&dqbuf, 0, sizeof(dqbuf)); | |
| 627 memset(planes, 0, sizeof(planes)); | |
| 628 dqbuf.type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE; | |
| 629 dqbuf.memory = V4L2_MEMORY_MMAP; | |
| 630 dqbuf.m.planes = planes; | |
| 631 dqbuf.length = 1; | |
| 632 if (device_->Ioctl(VIDIOC_DQBUF, &dqbuf) != 0) { | |
| 633 if (errno == EAGAIN) { | |
| 634 // EAGAIN if we're just out of buffers to dequeue. | |
| 635 break; | |
| 636 } | |
| 637 PLOG(ERROR) << "Dequeue(): ioctl() failed: VIDIOC_DQBUF"; | |
| 638 NOTIFY_ERROR(kPlatformFailureError); | |
| 639 return; | |
| 640 } | |
| 641 const bool key_frame = ((dqbuf.flags & V4L2_BUF_FLAG_KEYFRAME) != 0); | |
| 642 OutputRecord& output_record = output_buffer_map_[dqbuf.index]; | |
| 643 DCHECK(output_record.at_device); | |
| 644 DCHECK(output_record.buffer_ref.get()); | |
| 645 | |
| 646 void* output_data = output_record.address; | |
| 647 size_t output_size = dqbuf.m.planes[0].bytesused; | |
| 648 // This shouldn't happen, but just in case. We should be able to recover | |
| 649 // after next keyframe after showing some corruption. | |
| 650 DCHECK_LE(output_size, output_buffer_byte_size_); | |
| 651 if (output_size > output_buffer_byte_size_) | |
| 652 output_size = output_buffer_byte_size_; | |
| 653 uint8_t* target_data = | |
| 654 reinterpret_cast<uint8_t*>(output_record.buffer_ref->shm->memory()); | |
| 655 if (output_format_fourcc_ == V4L2_PIX_FMT_H264) { | |
| 656 if (stream_header_size_ == 0) { | |
| 657 // Assume that the first buffer dequeued is the stream header. | |
| 658 stream_header_size_ = output_size; | |
| 659 stream_header_.reset(new uint8_t[stream_header_size_]); | |
| 660 memcpy(stream_header_.get(), output_data, stream_header_size_); | |
| 661 } | |
| 662 if (key_frame && | |
| 663 output_buffer_byte_size_ - stream_header_size_ >= output_size) { | |
| 664 // Insert stream header before every keyframe. | |
| 665 memcpy(target_data, stream_header_.get(), stream_header_size_); | |
| 666 memcpy(target_data + stream_header_size_, output_data, output_size); | |
| 667 output_size += stream_header_size_; | |
| 668 } else { | |
| 669 memcpy(target_data, output_data, output_size); | |
| 670 } | |
| 671 } else { | |
| 672 memcpy(target_data, output_data, output_size); | |
| 673 } | |
| 674 | |
| 675 DVLOG(3) << "Dequeue(): returning " | |
| 676 "bitstream_buffer_id=" << output_record.buffer_ref->id | |
| 677 << ", size=" << output_size << ", key_frame=" << key_frame; | |
| 678 child_task_runner_->PostTask( | |
| 679 FROM_HERE, | |
| 680 base::Bind(&Client::BitstreamBufferReady, client_, | |
| 681 output_record.buffer_ref->id, output_size, key_frame)); | |
| 682 output_record.at_device = false; | |
| 683 output_record.buffer_ref.reset(); | |
| 684 free_output_buffers_.push_back(dqbuf.index); | |
| 685 output_buffer_queued_count_--; | |
| 686 } | |
| 687 } | |
| 688 | |
| 689 bool V4L2VideoEncodeAccelerator::EnqueueInputRecord() { | |
| 690 DVLOG(3) << "EnqueueInputRecord()"; | |
| 691 DCHECK(!free_input_buffers_.empty()); | |
| 692 DCHECK(!encoder_input_queue_.empty()); | |
| 693 | |
| 694 // Enqueue an input (VIDEO_OUTPUT) buffer. | |
| 695 scoped_refptr<media::VideoFrame> frame = encoder_input_queue_.front(); | |
| 696 const int index = free_input_buffers_.back(); | |
| 697 InputRecord& input_record = input_buffer_map_[index]; | |
| 698 DCHECK(!input_record.at_device); | |
| 699 struct v4l2_buffer qbuf; | |
| 700 struct v4l2_plane qbuf_planes[VIDEO_MAX_PLANES]; | |
| 701 memset(&qbuf, 0, sizeof(qbuf)); | |
| 702 memset(qbuf_planes, 0, sizeof(qbuf_planes)); | |
| 703 qbuf.index = index; | |
| 704 qbuf.type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE; | |
| 705 qbuf.m.planes = qbuf_planes; | |
| 706 | |
| 707 DCHECK_EQ(device_input_format_, frame->format()); | |
| 708 for (size_t i = 0; i < input_planes_count_; ++i) { | |
| 709 qbuf.m.planes[i].bytesused = | |
| 710 base::checked_cast<__u32>(media::VideoFrame::PlaneSize( | |
| 711 frame->format(), i, input_allocated_size_).GetArea()); | |
| 712 | |
| 713 switch (input_memory_type_) { | |
| 714 case V4L2_MEMORY_USERPTR: | |
| 715 qbuf.m.planes[i].length = qbuf.m.planes[i].bytesused; | |
| 716 qbuf.m.planes[i].m.userptr = | |
| 717 reinterpret_cast<unsigned long>(frame->data(i)); | |
| 718 DCHECK(qbuf.m.planes[i].m.userptr); | |
| 719 break; | |
| 720 | |
| 721 case V4L2_MEMORY_DMABUF: | |
| 722 qbuf.m.planes[i].m.fd = frame->dmabuf_fd(i); | |
| 723 DCHECK_NE(qbuf.m.planes[i].m.fd, -1); | |
| 724 break; | |
| 725 | |
| 726 default: | |
| 727 NOTREACHED(); | |
| 728 return false; | |
| 729 } | |
| 730 } | |
| 731 | |
| 732 qbuf.memory = input_memory_type_; | |
| 733 qbuf.length = input_planes_count_; | |
| 734 | |
| 735 IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_QBUF, &qbuf); | |
| 736 input_record.at_device = true; | |
| 737 input_record.frame = frame; | |
| 738 encoder_input_queue_.pop(); | |
| 739 free_input_buffers_.pop_back(); | |
| 740 input_buffer_queued_count_++; | |
| 741 return true; | |
| 742 } | |
| 743 | |
| 744 bool V4L2VideoEncodeAccelerator::EnqueueOutputRecord() { | |
| 745 DVLOG(3) << "EnqueueOutputRecord()"; | |
| 746 DCHECK(!free_output_buffers_.empty()); | |
| 747 DCHECK(!encoder_output_queue_.empty()); | |
| 748 | |
| 749 // Enqueue an output (VIDEO_CAPTURE) buffer. | |
| 750 linked_ptr<BitstreamBufferRef> output_buffer = encoder_output_queue_.back(); | |
| 751 const int index = free_output_buffers_.back(); | |
| 752 OutputRecord& output_record = output_buffer_map_[index]; | |
| 753 DCHECK(!output_record.at_device); | |
| 754 DCHECK(!output_record.buffer_ref.get()); | |
| 755 struct v4l2_buffer qbuf; | |
| 756 struct v4l2_plane qbuf_planes[1]; | |
| 757 memset(&qbuf, 0, sizeof(qbuf)); | |
| 758 memset(qbuf_planes, 0, sizeof(qbuf_planes)); | |
| 759 qbuf.index = index; | |
| 760 qbuf.type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE; | |
| 761 qbuf.memory = V4L2_MEMORY_MMAP; | |
| 762 qbuf.m.planes = qbuf_planes; | |
| 763 qbuf.length = 1; | |
| 764 IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_QBUF, &qbuf); | |
| 765 output_record.at_device = true; | |
| 766 output_record.buffer_ref = output_buffer; | |
| 767 encoder_output_queue_.pop_back(); | |
| 768 free_output_buffers_.pop_back(); | |
| 769 output_buffer_queued_count_++; | |
| 770 return true; | |
| 771 } | |
| 772 | |
| 773 bool V4L2VideoEncodeAccelerator::StartDevicePoll() { | |
| 774 DVLOG(3) << "StartDevicePoll()"; | |
| 775 DCHECK_EQ(encoder_thread_.message_loop(), base::MessageLoop::current()); | |
| 776 DCHECK(!device_poll_thread_.IsRunning()); | |
| 777 | |
| 778 // Start up the device poll thread and schedule its first DevicePollTask(). | |
| 779 if (!device_poll_thread_.Start()) { | |
| 780 LOG(ERROR) << "StartDevicePoll(): Device thread failed to start"; | |
| 781 NOTIFY_ERROR(kPlatformFailureError); | |
| 782 return false; | |
| 783 } | |
| 784 // Enqueue a poll task with no devices to poll on -- it will wait only on the | |
| 785 // interrupt fd. | |
| 786 device_poll_thread_.message_loop()->PostTask( | |
| 787 FROM_HERE, | |
| 788 base::Bind(&V4L2VideoEncodeAccelerator::DevicePollTask, | |
| 789 base::Unretained(this), | |
| 790 false)); | |
| 791 | |
| 792 return true; | |
| 793 } | |
| 794 | |
| 795 bool V4L2VideoEncodeAccelerator::StopDevicePoll() { | |
| 796 DVLOG(3) << "StopDevicePoll()"; | |
| 797 | |
| 798 // Signal the DevicePollTask() to stop, and stop the device poll thread. | |
| 799 if (!device_->SetDevicePollInterrupt()) | |
| 800 return false; | |
| 801 device_poll_thread_.Stop(); | |
| 802 // Clear the interrupt now, to be sure. | |
| 803 if (!device_->ClearDevicePollInterrupt()) | |
| 804 return false; | |
| 805 | |
| 806 if (input_streamon_) { | |
| 807 __u32 type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE; | |
| 808 IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_STREAMOFF, &type); | |
| 809 } | |
| 810 input_streamon_ = false; | |
| 811 | |
| 812 if (output_streamon_) { | |
| 813 __u32 type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE; | |
| 814 IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_STREAMOFF, &type); | |
| 815 } | |
| 816 output_streamon_ = false; | |
| 817 | |
| 818 // Reset all our accounting info. | |
| 819 while (!encoder_input_queue_.empty()) | |
| 820 encoder_input_queue_.pop(); | |
| 821 free_input_buffers_.clear(); | |
| 822 for (size_t i = 0; i < input_buffer_map_.size(); ++i) { | |
| 823 InputRecord& input_record = input_buffer_map_[i]; | |
| 824 input_record.at_device = false; | |
| 825 input_record.frame = NULL; | |
| 826 free_input_buffers_.push_back(i); | |
| 827 } | |
| 828 input_buffer_queued_count_ = 0; | |
| 829 | |
| 830 free_output_buffers_.clear(); | |
| 831 for (size_t i = 0; i < output_buffer_map_.size(); ++i) { | |
| 832 OutputRecord& output_record = output_buffer_map_[i]; | |
| 833 output_record.at_device = false; | |
| 834 output_record.buffer_ref.reset(); | |
| 835 free_output_buffers_.push_back(i); | |
| 836 } | |
| 837 output_buffer_queued_count_ = 0; | |
| 838 | |
| 839 encoder_output_queue_.clear(); | |
| 840 | |
| 841 DVLOG(3) << "StopDevicePoll(): device poll stopped"; | |
| 842 return true; | |
| 843 } | |
| 844 | |
| 845 void V4L2VideoEncodeAccelerator::DevicePollTask(bool poll_device) { | |
| 846 DVLOG(3) << "DevicePollTask()"; | |
| 847 DCHECK_EQ(device_poll_thread_.message_loop(), base::MessageLoop::current()); | |
| 848 | |
| 849 bool event_pending; | |
| 850 if (!device_->Poll(poll_device, &event_pending)) { | |
| 851 NOTIFY_ERROR(kPlatformFailureError); | |
| 852 return; | |
| 853 } | |
| 854 | |
| 855 // All processing should happen on ServiceDeviceTask(), since we shouldn't | |
| 856 // touch encoder state from this thread. | |
| 857 encoder_thread_.message_loop()->PostTask( | |
| 858 FROM_HERE, | |
| 859 base::Bind(&V4L2VideoEncodeAccelerator::ServiceDeviceTask, | |
| 860 base::Unretained(this))); | |
| 861 } | |
| 862 | |
| 863 void V4L2VideoEncodeAccelerator::NotifyError(Error error) { | |
| 864 DVLOG(1) << "NotifyError(): error=" << error; | |
| 865 | |
| 866 if (!child_task_runner_->BelongsToCurrentThread()) { | |
| 867 child_task_runner_->PostTask( | |
| 868 FROM_HERE, base::Bind(&V4L2VideoEncodeAccelerator::NotifyError, | |
| 869 weak_this_, error)); | |
| 870 return; | |
| 871 } | |
| 872 | |
| 873 if (client_) { | |
| 874 client_->NotifyError(error); | |
| 875 client_ptr_factory_.reset(); | |
| 876 } | |
| 877 } | |
| 878 | |
| 879 void V4L2VideoEncodeAccelerator::SetErrorState(Error error) { | |
| 880 // We can touch encoder_state_ only if this is the encoder thread or the | |
| 881 // encoder thread isn't running. | |
| 882 if (encoder_thread_.message_loop() != NULL && | |
| 883 encoder_thread_.message_loop() != base::MessageLoop::current()) { | |
| 884 encoder_thread_.message_loop()->PostTask( | |
| 885 FROM_HERE, base::Bind(&V4L2VideoEncodeAccelerator::SetErrorState, | |
| 886 base::Unretained(this), error)); | |
| 887 return; | |
| 888 } | |
| 889 | |
| 890 // Post NotifyError only if we are already initialized, as the API does | |
| 891 // not allow doing so before that. | |
| 892 if (encoder_state_ != kError && encoder_state_ != kUninitialized) | |
| 893 NotifyError(error); | |
| 894 | |
| 895 encoder_state_ = kError; | |
| 896 } | |
| 897 | |
| 898 void V4L2VideoEncodeAccelerator::RequestEncodingParametersChangeTask( | |
| 899 uint32_t bitrate, | |
| 900 uint32_t framerate) { | |
| 901 DVLOG(3) << "RequestEncodingParametersChangeTask(): bitrate=" << bitrate | |
| 902 << ", framerate=" << framerate; | |
| 903 DCHECK_EQ(encoder_thread_.message_loop(), base::MessageLoop::current()); | |
| 904 | |
| 905 if (bitrate < 1) | |
| 906 bitrate = 1; | |
| 907 if (framerate < 1) | |
| 908 framerate = 1; | |
| 909 | |
| 910 std::vector<struct v4l2_ext_control> ctrls; | |
| 911 struct v4l2_ext_control ctrl; | |
| 912 memset(&ctrl, 0, sizeof(ctrl)); | |
| 913 ctrl.id = V4L2_CID_MPEG_VIDEO_BITRATE; | |
| 914 ctrl.value = bitrate; | |
| 915 ctrls.push_back(ctrl); | |
| 916 if (!SetExtCtrls(ctrls)) { | |
| 917 LOG(ERROR) << "Failed changing bitrate"; | |
| 918 NOTIFY_ERROR(kPlatformFailureError); | |
| 919 return; | |
| 920 } | |
| 921 | |
| 922 struct v4l2_streamparm parms; | |
| 923 memset(&parms, 0, sizeof(parms)); | |
| 924 parms.type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE; | |
| 925 // Note that we are provided "frames per second" but V4L2 expects "time per | |
| 926 // frame"; hence we provide the reciprocal of the framerate here. | |
| 927 parms.parm.output.timeperframe.numerator = 1; | |
| 928 parms.parm.output.timeperframe.denominator = framerate; | |
| 929 IOCTL_OR_ERROR_RETURN(VIDIOC_S_PARM, &parms); | |
| 930 } | |
| 931 | |
| 932 bool V4L2VideoEncodeAccelerator::SetOutputFormat( | |
| 933 media::VideoCodecProfile output_profile) { | |
| 934 DCHECK(child_task_runner_->BelongsToCurrentThread()); | |
| 935 DCHECK(!input_streamon_); | |
| 936 DCHECK(!output_streamon_); | |
| 937 | |
| 938 output_format_fourcc_ = | |
| 939 V4L2Device::VideoCodecProfileToV4L2PixFmt(output_profile, false); | |
| 940 if (!output_format_fourcc_) { | |
| 941 LOG(ERROR) << "Initialize(): invalid output_profile=" << output_profile; | |
| 942 return false; | |
| 943 } | |
| 944 | |
| 945 output_buffer_byte_size_ = kOutputBufferSize; | |
| 946 | |
| 947 struct v4l2_format format; | |
| 948 memset(&format, 0, sizeof(format)); | |
| 949 format.type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE; | |
| 950 format.fmt.pix_mp.width = visible_size_.width(); | |
| 951 format.fmt.pix_mp.height = visible_size_.height(); | |
| 952 format.fmt.pix_mp.pixelformat = output_format_fourcc_; | |
| 953 format.fmt.pix_mp.plane_fmt[0].sizeimage = | |
| 954 base::checked_cast<__u32>(output_buffer_byte_size_); | |
| 955 format.fmt.pix_mp.num_planes = 1; | |
| 956 IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_S_FMT, &format); | |
| 957 | |
| 958 // Device might have adjusted the required output size. | |
| 959 size_t adjusted_output_buffer_size = | |
| 960 base::checked_cast<size_t>(format.fmt.pix_mp.plane_fmt[0].sizeimage); | |
| 961 output_buffer_byte_size_ = adjusted_output_buffer_size; | |
| 962 | |
| 963 return true; | |
| 964 } | |
| 965 | |
| 966 bool V4L2VideoEncodeAccelerator::NegotiateInputFormat( | |
| 967 media::VideoPixelFormat input_format) { | |
| 968 DVLOG(3) << "NegotiateInputFormat()"; | |
| 969 DCHECK(child_task_runner_->BelongsToCurrentThread()); | |
| 970 DCHECK(!input_streamon_); | |
| 971 DCHECK(!output_streamon_); | |
| 972 | |
| 973 device_input_format_ = media::PIXEL_FORMAT_UNKNOWN; | |
| 974 input_planes_count_ = 0; | |
| 975 | |
| 976 uint32_t input_format_fourcc = | |
| 977 V4L2Device::VideoPixelFormatToV4L2PixFmt(input_format); | |
| 978 if (!input_format_fourcc) { | |
| 979 LOG(ERROR) << "Unsupported input format" << input_format_fourcc; | |
| 980 return false; | |
| 981 } | |
| 982 | |
| 983 size_t input_planes_count = media::VideoFrame::NumPlanes(input_format); | |
| 984 DCHECK_LE(input_planes_count, static_cast<size_t>(VIDEO_MAX_PLANES)); | |
| 985 | |
| 986 // First see if we the device can use the provided input_format directly. | |
| 987 struct v4l2_format format; | |
| 988 memset(&format, 0, sizeof(format)); | |
| 989 format.type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE; | |
| 990 format.fmt.pix_mp.width = visible_size_.width(); | |
| 991 format.fmt.pix_mp.height = visible_size_.height(); | |
| 992 format.fmt.pix_mp.pixelformat = input_format_fourcc; | |
| 993 format.fmt.pix_mp.num_planes = input_planes_count; | |
| 994 if (device_->Ioctl(VIDIOC_S_FMT, &format) != 0) { | |
| 995 // Error or format unsupported by device, try to negotiate a fallback. | |
| 996 input_format_fourcc = device_->PreferredInputFormat(); | |
| 997 input_format = | |
| 998 V4L2Device::V4L2PixFmtToVideoPixelFormat(input_format_fourcc); | |
| 999 if (input_format == media::PIXEL_FORMAT_UNKNOWN) { | |
| 1000 LOG(ERROR) << "Unsupported input format" << input_format_fourcc; | |
| 1001 return false; | |
| 1002 } | |
| 1003 | |
| 1004 input_planes_count = media::VideoFrame::NumPlanes(input_format); | |
| 1005 DCHECK_LE(input_planes_count, static_cast<size_t>(VIDEO_MAX_PLANES)); | |
| 1006 | |
| 1007 // Device might have adjusted parameters, reset them along with the format. | |
| 1008 memset(&format, 0, sizeof(format)); | |
| 1009 format.type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE; | |
| 1010 format.fmt.pix_mp.width = visible_size_.width(); | |
| 1011 format.fmt.pix_mp.height = visible_size_.height(); | |
| 1012 format.fmt.pix_mp.pixelformat = input_format_fourcc; | |
| 1013 format.fmt.pix_mp.num_planes = input_planes_count; | |
| 1014 IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_S_FMT, &format); | |
| 1015 DCHECK_EQ(format.fmt.pix_mp.num_planes, input_planes_count); | |
| 1016 } | |
| 1017 | |
| 1018 // Take device-adjusted sizes for allocated size. If the size is adjusted | |
| 1019 // down, it means the input is too big and the hardware does not support it. | |
| 1020 input_allocated_size_ = V4L2Device::CodedSizeFromV4L2Format(format); | |
| 1021 if (!gfx::Rect(input_allocated_size_).Contains(gfx::Rect(visible_size_))) { | |
| 1022 DVLOG(1) << "Input size too big " << visible_size_.ToString() | |
| 1023 << ", adjusted to " << input_allocated_size_.ToString(); | |
| 1024 return false; | |
| 1025 } | |
| 1026 | |
| 1027 device_input_format_ = input_format; | |
| 1028 input_planes_count_ = input_planes_count; | |
| 1029 return true; | |
| 1030 } | |
| 1031 | |
| 1032 bool V4L2VideoEncodeAccelerator::SetFormats( | |
| 1033 media::VideoPixelFormat input_format, | |
| 1034 media::VideoCodecProfile output_profile) { | |
| 1035 DVLOG(3) << "SetFormats()"; | |
| 1036 DCHECK(child_task_runner_->BelongsToCurrentThread()); | |
| 1037 DCHECK(!input_streamon_); | |
| 1038 DCHECK(!output_streamon_); | |
| 1039 | |
| 1040 if (!SetOutputFormat(output_profile)) | |
| 1041 return false; | |
| 1042 | |
| 1043 if (!NegotiateInputFormat(input_format)) | |
| 1044 return false; | |
| 1045 | |
| 1046 struct v4l2_crop crop; | |
| 1047 memset(&crop, 0, sizeof(crop)); | |
| 1048 crop.type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE; | |
| 1049 crop.c.left = 0; | |
| 1050 crop.c.top = 0; | |
| 1051 crop.c.width = visible_size_.width(); | |
| 1052 crop.c.height = visible_size_.height(); | |
| 1053 IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_S_CROP, &crop); | |
| 1054 | |
| 1055 // The width and height might be adjusted by driver. | |
| 1056 // Need to read it back and set to visible_size_. | |
| 1057 if (device_->Ioctl(VIDIOC_G_CROP, &crop) != 0) { | |
| 1058 // Some devices haven't supported G_CROP yet, so treat the failure | |
| 1059 // non-fatal for now. | |
| 1060 // TODO(kcwu): NOTIFY_ERROR and return false after all devices support it. | |
| 1061 PLOG(WARNING) << "SetFormats(): ioctl() VIDIOC_G_CROP failed"; | |
| 1062 return true; | |
| 1063 } | |
| 1064 visible_size_.SetSize(crop.c.width, crop.c.height); | |
| 1065 DVLOG(3) << "After adjusted by driver, visible_size_=" | |
| 1066 << visible_size_.ToString(); | |
| 1067 | |
| 1068 return true; | |
| 1069 } | |
| 1070 | |
| 1071 bool V4L2VideoEncodeAccelerator::SetExtCtrls( | |
| 1072 std::vector<struct v4l2_ext_control> ctrls) { | |
| 1073 struct v4l2_ext_controls ext_ctrls; | |
| 1074 memset(&ext_ctrls, 0, sizeof(ext_ctrls)); | |
| 1075 ext_ctrls.ctrl_class = V4L2_CTRL_CLASS_MPEG; | |
| 1076 ext_ctrls.count = ctrls.size(); | |
| 1077 ext_ctrls.controls = &ctrls[0]; | |
| 1078 return device_->Ioctl(VIDIOC_S_EXT_CTRLS, &ext_ctrls) == 0; | |
| 1079 } | |
| 1080 | |
| 1081 bool V4L2VideoEncodeAccelerator::InitControls() { | |
| 1082 std::vector<struct v4l2_ext_control> ctrls; | |
| 1083 struct v4l2_ext_control ctrl; | |
| 1084 | |
| 1085 // Enable frame-level bitrate control. This is the only mandatory control. | |
| 1086 memset(&ctrl, 0, sizeof(ctrl)); | |
| 1087 ctrl.id = V4L2_CID_MPEG_VIDEO_FRAME_RC_ENABLE; | |
| 1088 ctrl.value = 1; | |
| 1089 ctrls.push_back(ctrl); | |
| 1090 if (!SetExtCtrls(ctrls)) { | |
| 1091 LOG(ERROR) << "Failed enabling bitrate control"; | |
| 1092 NOTIFY_ERROR(kPlatformFailureError); | |
| 1093 return false; | |
| 1094 } | |
| 1095 | |
| 1096 // Optional controls. | |
| 1097 ctrls.clear(); | |
| 1098 if (output_format_fourcc_ == V4L2_PIX_FMT_H264) { | |
| 1099 // No B-frames, for lowest decoding latency. | |
| 1100 memset(&ctrl, 0, sizeof(ctrl)); | |
| 1101 ctrl.id = V4L2_CID_MPEG_VIDEO_B_FRAMES; | |
| 1102 ctrl.value = 0; | |
| 1103 ctrls.push_back(ctrl); | |
| 1104 | |
| 1105 // Quantization parameter maximum value (for variable bitrate control). | |
| 1106 memset(&ctrl, 0, sizeof(ctrl)); | |
| 1107 ctrl.id = V4L2_CID_MPEG_VIDEO_H264_MAX_QP; | |
| 1108 ctrl.value = 51; | |
| 1109 ctrls.push_back(ctrl); | |
| 1110 | |
| 1111 // Use H.264 level 4.0 to match the supported max resolution. | |
| 1112 memset(&ctrl, 0, sizeof(ctrl)); | |
| 1113 ctrl.id = V4L2_CID_MPEG_VIDEO_H264_LEVEL; | |
| 1114 ctrl.value = V4L2_MPEG_VIDEO_H264_LEVEL_4_0; | |
| 1115 ctrls.push_back(ctrl); | |
| 1116 | |
| 1117 // Separate stream header so we can cache it and insert into the stream. | |
| 1118 memset(&ctrl, 0, sizeof(ctrl)); | |
| 1119 ctrl.id = V4L2_CID_MPEG_VIDEO_HEADER_MODE; | |
| 1120 ctrl.value = V4L2_MPEG_VIDEO_HEADER_MODE_SEPARATE; | |
| 1121 ctrls.push_back(ctrl); | |
| 1122 } | |
| 1123 | |
| 1124 // Enable macroblock-level bitrate control. | |
| 1125 memset(&ctrl, 0, sizeof(ctrl)); | |
| 1126 ctrl.id = V4L2_CID_MPEG_VIDEO_MB_RC_ENABLE; | |
| 1127 ctrl.value = 1; | |
| 1128 ctrls.push_back(ctrl); | |
| 1129 | |
| 1130 // Disable periodic key frames. | |
| 1131 memset(&ctrl, 0, sizeof(ctrl)); | |
| 1132 ctrl.id = V4L2_CID_MPEG_VIDEO_GOP_SIZE; | |
| 1133 ctrl.value = 0; | |
| 1134 ctrls.push_back(ctrl); | |
| 1135 | |
| 1136 // Ignore return value as these controls are optional. | |
| 1137 SetExtCtrls(ctrls); | |
| 1138 | |
| 1139 // Optional Exynos specific controls. | |
| 1140 ctrls.clear(); | |
| 1141 // Enable "tight" bitrate mode. For this to work properly, frame- and mb-level | |
| 1142 // bitrate controls have to be enabled as well. | |
| 1143 memset(&ctrl, 0, sizeof(ctrl)); | |
| 1144 ctrl.id = V4L2_CID_MPEG_MFC51_VIDEO_RC_REACTION_COEFF; | |
| 1145 ctrl.value = 1; | |
| 1146 ctrls.push_back(ctrl); | |
| 1147 | |
| 1148 // Force bitrate control to average over a GOP (for tight bitrate | |
| 1149 // tolerance). | |
| 1150 memset(&ctrl, 0, sizeof(ctrl)); | |
| 1151 ctrl.id = V4L2_CID_MPEG_MFC51_VIDEO_RC_FIXED_TARGET_BIT; | |
| 1152 ctrl.value = 1; | |
| 1153 ctrls.push_back(ctrl); | |
| 1154 | |
| 1155 // Ignore return value as these controls are optional. | |
| 1156 SetExtCtrls(ctrls); | |
| 1157 | |
| 1158 return true; | |
| 1159 } | |
| 1160 | |
| 1161 bool V4L2VideoEncodeAccelerator::CreateInputBuffers() { | |
| 1162 DVLOG(3) << "CreateInputBuffers()"; | |
| 1163 // This function runs on encoder_thread_ after output buffers have been | |
| 1164 // provided by the client. | |
| 1165 DCHECK_EQ(encoder_thread_.message_loop(), base::MessageLoop::current()); | |
| 1166 DCHECK(!input_streamon_); | |
| 1167 | |
| 1168 struct v4l2_requestbuffers reqbufs; | |
| 1169 memset(&reqbufs, 0, sizeof(reqbufs)); | |
| 1170 // Driver will modify to the appropriate number of buffers. | |
| 1171 reqbufs.count = 1; | |
| 1172 reqbufs.type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE; | |
| 1173 // TODO(posciak): Once we start doing zero-copy, we should decide based on | |
| 1174 // the current pipeline setup which memory type to use. This should probably | |
| 1175 // be decided based on an argument to Initialize(). | |
| 1176 if (image_processor_.get()) | |
| 1177 input_memory_type_ = V4L2_MEMORY_DMABUF; | |
| 1178 else | |
| 1179 input_memory_type_ = V4L2_MEMORY_USERPTR; | |
| 1180 | |
| 1181 reqbufs.memory = input_memory_type_; | |
| 1182 IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_REQBUFS, &reqbufs); | |
| 1183 | |
| 1184 DCHECK(input_buffer_map_.empty()); | |
| 1185 input_buffer_map_.resize(reqbufs.count); | |
| 1186 for (size_t i = 0; i < input_buffer_map_.size(); ++i) | |
| 1187 free_input_buffers_.push_back(i); | |
| 1188 | |
| 1189 return true; | |
| 1190 } | |
| 1191 | |
| 1192 bool V4L2VideoEncodeAccelerator::CreateOutputBuffers() { | |
| 1193 DVLOG(3) << "CreateOutputBuffers()"; | |
| 1194 DCHECK(child_task_runner_->BelongsToCurrentThread()); | |
| 1195 DCHECK(!output_streamon_); | |
| 1196 | |
| 1197 struct v4l2_requestbuffers reqbufs; | |
| 1198 memset(&reqbufs, 0, sizeof(reqbufs)); | |
| 1199 reqbufs.count = kOutputBufferCount; | |
| 1200 reqbufs.type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE; | |
| 1201 reqbufs.memory = V4L2_MEMORY_MMAP; | |
| 1202 IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_REQBUFS, &reqbufs); | |
| 1203 | |
| 1204 DCHECK(output_buffer_map_.empty()); | |
| 1205 output_buffer_map_.resize(reqbufs.count); | |
| 1206 for (size_t i = 0; i < output_buffer_map_.size(); ++i) { | |
| 1207 struct v4l2_plane planes[1]; | |
| 1208 struct v4l2_buffer buffer; | |
| 1209 memset(&buffer, 0, sizeof(buffer)); | |
| 1210 memset(planes, 0, sizeof(planes)); | |
| 1211 buffer.index = i; | |
| 1212 buffer.type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE; | |
| 1213 buffer.memory = V4L2_MEMORY_MMAP; | |
| 1214 buffer.m.planes = planes; | |
| 1215 buffer.length = arraysize(planes); | |
| 1216 IOCTL_OR_ERROR_RETURN_FALSE(VIDIOC_QUERYBUF, &buffer); | |
| 1217 void* address = device_->Mmap(NULL, | |
| 1218 buffer.m.planes[0].length, | |
| 1219 PROT_READ | PROT_WRITE, | |
| 1220 MAP_SHARED, | |
| 1221 buffer.m.planes[0].m.mem_offset); | |
| 1222 if (address == MAP_FAILED) { | |
| 1223 PLOG(ERROR) << "CreateOutputBuffers(): mmap() failed"; | |
| 1224 return false; | |
| 1225 } | |
| 1226 output_buffer_map_[i].address = address; | |
| 1227 output_buffer_map_[i].length = buffer.m.planes[0].length; | |
| 1228 free_output_buffers_.push_back(i); | |
| 1229 } | |
| 1230 | |
| 1231 return true; | |
| 1232 } | |
| 1233 | |
| 1234 void V4L2VideoEncodeAccelerator::DestroyInputBuffers() { | |
| 1235 DVLOG(3) << "DestroyInputBuffers()"; | |
| 1236 DCHECK(child_task_runner_->BelongsToCurrentThread()); | |
| 1237 DCHECK(!input_streamon_); | |
| 1238 | |
| 1239 struct v4l2_requestbuffers reqbufs; | |
| 1240 memset(&reqbufs, 0, sizeof(reqbufs)); | |
| 1241 reqbufs.count = 0; | |
| 1242 reqbufs.type = V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE; | |
| 1243 reqbufs.memory = input_memory_type_; | |
| 1244 IOCTL_OR_LOG_ERROR(VIDIOC_REQBUFS, &reqbufs); | |
| 1245 | |
| 1246 input_buffer_map_.clear(); | |
| 1247 free_input_buffers_.clear(); | |
| 1248 } | |
| 1249 | |
| 1250 void V4L2VideoEncodeAccelerator::DestroyOutputBuffers() { | |
| 1251 DVLOG(3) << "DestroyOutputBuffers()"; | |
| 1252 DCHECK(child_task_runner_->BelongsToCurrentThread()); | |
| 1253 DCHECK(!output_streamon_); | |
| 1254 | |
| 1255 for (size_t i = 0; i < output_buffer_map_.size(); ++i) { | |
| 1256 if (output_buffer_map_[i].address != NULL) | |
| 1257 device_->Munmap(output_buffer_map_[i].address, | |
| 1258 output_buffer_map_[i].length); | |
| 1259 } | |
| 1260 | |
| 1261 struct v4l2_requestbuffers reqbufs; | |
| 1262 memset(&reqbufs, 0, sizeof(reqbufs)); | |
| 1263 reqbufs.count = 0; | |
| 1264 reqbufs.type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE; | |
| 1265 reqbufs.memory = V4L2_MEMORY_MMAP; | |
| 1266 IOCTL_OR_LOG_ERROR(VIDIOC_REQBUFS, &reqbufs); | |
| 1267 | |
| 1268 output_buffer_map_.clear(); | |
| 1269 free_output_buffers_.clear(); | |
| 1270 } | |
| 1271 | |
| 1272 } // namespace content | |
| OLD | NEW |