Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(619)

Unified Diff: content/common/gpu/media/h264_decoder.cc

Issue 813693006: Add accelerated video decoder interface, VP8 and H.264 implementations and hook up to V4L2SVDA (Closed) Base URL: https://chromium.googlesource.com/chromium/src.git@master
Patch Set: Created 5 years, 11 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « content/common/gpu/media/h264_decoder.h ('k') | content/common/gpu/media/h264_dpb.h » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: content/common/gpu/media/h264_decoder.cc
diff --git a/content/common/gpu/media/vaapi_h264_decoder.cc b/content/common/gpu/media/h264_decoder.cc
similarity index 56%
copy from content/common/gpu/media/vaapi_h264_decoder.cc
copy to content/common/gpu/media/h264_decoder.cc
index 55b0d007cda566936c3ed0c26043f3a5dfd8f16e..af36b03e7cdb8f5c55c0a3dd03180b6c36ee3dfa 100644
--- a/content/common/gpu/media/vaapi_h264_decoder.cc
+++ b/content/common/gpu/media/h264_decoder.cc
@@ -7,58 +7,20 @@
#include "base/bind.h"
#include "base/bind_helpers.h"
+#include "base/callback_helpers.h"
#include "base/numerics/safe_conversions.h"
#include "base/stl_util.h"
-#include "content/common/gpu/media/vaapi_h264_decoder.h"
+#include "content/common/gpu/media/h264_decoder.h"
namespace content {
-// Decode surface, used for decoding and reference. input_id comes from client
-// and is associated with the surface that was produced as the result
-// of decoding a bitstream buffer with that id.
-class VaapiH264Decoder::DecodeSurface {
- public:
- DecodeSurface(int poc,
- int32 input_id,
- const scoped_refptr<VASurface>& va_surface);
- DecodeSurface(int poc, const scoped_refptr<DecodeSurface>& dec_surface);
- ~DecodeSurface();
-
- int poc() {
- return poc_;
- }
-
- scoped_refptr<VASurface> va_surface() {
- return va_surface_;
- }
-
- int32 input_id() {
- return input_id_;
- }
-
- private:
- int poc_;
- int32 input_id_;
- scoped_refptr<VASurface> va_surface_;
-};
-
-VaapiH264Decoder::DecodeSurface::DecodeSurface(
- int poc,
- int32 input_id,
- const scoped_refptr<VASurface>& va_surface)
- : poc_(poc),
- input_id_(input_id),
- va_surface_(va_surface) {
- DCHECK(va_surface_.get());
+H264Decoder::H264Accelerator::H264Accelerator() {
}
-VaapiH264Decoder::DecodeSurface::~DecodeSurface() {
+H264Decoder::H264Accelerator::~H264Accelerator() {
}
-VaapiH264Decoder::VaapiH264Decoder(
- VaapiWrapper* vaapi_wrapper,
- const OutputPicCB& output_pic_cb,
- const ReportErrorToUmaCB& report_error_to_uma_cb)
+H264Decoder::H264Decoder(H264Accelerator* accelerator)
: max_pic_order_cnt_lsb_(0),
max_frame_num_(0),
max_pic_num_(0),
@@ -66,20 +28,20 @@ VaapiH264Decoder::VaapiH264Decoder(
max_num_reorder_frames_(0),
curr_sps_id_(-1),
curr_pps_id_(-1),
- vaapi_wrapper_(vaapi_wrapper),
- output_pic_cb_(output_pic_cb),
- report_error_to_uma_cb_(report_error_to_uma_cb) {
+ accelerator_(accelerator) {
+ DCHECK(accelerator_);
Reset();
state_ = kNeedStreamMetadata;
}
-VaapiH264Decoder::~VaapiH264Decoder() {
+H264Decoder::~H264Decoder() {
}
-void VaapiH264Decoder::Reset() {
- curr_pic_.reset();
+void H264Decoder::Reset() {
+ curr_pic_ = nullptr;
+ curr_nalu_ = nullptr;
+ curr_slice_hdr_ = nullptr;
- curr_input_id_ = -1;
frame_num_ = 0;
prev_frame_num_ = -1;
prev_frame_num_offset_ = -1;
@@ -90,21 +52,9 @@ void VaapiH264Decoder::Reset() {
prev_ref_pic_order_cnt_lsb_ = -1;
prev_ref_field_ = H264Picture::FIELD_NONE;
- vaapi_wrapper_->DestroyPendingBuffers();
-
- ref_pic_list0_.clear();
- ref_pic_list1_.clear();
-
- for (DecSurfacesInUse::iterator it = decode_surfaces_in_use_.begin();
- it != decode_surfaces_in_use_.end(); ) {
- int poc = it->second->poc();
- // Must be incremented before UnassignSurfaceFromPoC as this call
- // invalidates |it|.
- ++it;
- UnassignSurfaceFromPoC(poc);
- }
- DCHECK(decode_surfaces_in_use_.empty());
-
+ ref_pic_list_p0_.clear();
+ ref_pic_list_b0_.clear();
+ ref_pic_list_b1_.clear();
dpb_.Clear();
parser_.Reset();
last_output_poc_ = std::numeric_limits<int>::min();
@@ -114,394 +64,41 @@ void VaapiH264Decoder::Reset() {
state_ = kAfterReset;
}
-void VaapiH264Decoder::ReuseSurface(
- const scoped_refptr<VASurface>& va_surface) {
- available_va_surfaces_.push_back(va_surface);
-}
-
-// Fill |va_pic| with default/neutral values.
-static void InitVAPicture(VAPictureH264* va_pic) {
- memset(va_pic, 0, sizeof(*va_pic));
- va_pic->picture_id = VA_INVALID_ID;
- va_pic->flags = VA_PICTURE_H264_INVALID;
-}
-
-void VaapiH264Decoder::FillVAPicture(VAPictureH264 *va_pic, H264Picture* pic) {
- DCHECK(pic);
-
- DecodeSurface* dec_surface = DecodeSurfaceByPoC(pic->pic_order_cnt);
- if (!dec_surface) {
- // Cannot provide a ref picture, will corrupt output, but may be able
- // to recover.
- InitVAPicture(va_pic);
- return;
- }
-
- va_pic->picture_id = dec_surface->va_surface()->id();
- va_pic->frame_idx = pic->frame_num;
- va_pic->flags = 0;
-
- switch (pic->field) {
- case H264Picture::FIELD_NONE:
- break;
- case H264Picture::FIELD_TOP:
- va_pic->flags |= VA_PICTURE_H264_TOP_FIELD;
- break;
- case H264Picture::FIELD_BOTTOM:
- va_pic->flags |= VA_PICTURE_H264_BOTTOM_FIELD;
- break;
- }
-
- if (pic->ref) {
- va_pic->flags |= pic->long_term ? VA_PICTURE_H264_LONG_TERM_REFERENCE
- : VA_PICTURE_H264_SHORT_TERM_REFERENCE;
- }
-
- va_pic->TopFieldOrderCnt = pic->top_field_order_cnt;
- va_pic->BottomFieldOrderCnt = pic->bottom_field_order_cnt;
-}
-
-int VaapiH264Decoder::FillVARefFramesFromDPB(VAPictureH264 *va_pics,
- int num_pics) {
- H264DPB::Pictures::reverse_iterator rit;
- int i;
-
- // Return reference frames in reverse order of insertion.
- // Libva does not document this, but other implementations (e.g. mplayer)
- // do it this way as well.
- for (rit = dpb_.rbegin(), i = 0; rit != dpb_.rend() && i < num_pics; ++rit) {
- if ((*rit)->ref)
- FillVAPicture(&va_pics[i++], *rit);
- }
-
- return i;
+void H264Decoder::PrepareRefPicLists(media::H264SliceHeader* slice_hdr) {
+ ConstructReferencePicListsP(slice_hdr);
+ ConstructReferencePicListsB(slice_hdr);
}
-VaapiH264Decoder::DecodeSurface* VaapiH264Decoder::DecodeSurfaceByPoC(int poc) {
- DecSurfacesInUse::iterator iter = decode_surfaces_in_use_.find(poc);
- if (iter == decode_surfaces_in_use_.end()) {
- DVLOG(1) << "Could not find surface assigned to POC: " << poc;
- return NULL;
- }
-
- return iter->second.get();
-}
-
-bool VaapiH264Decoder::AssignSurfaceToPoC(int32 input_id, int poc) {
- if (available_va_surfaces_.empty()) {
- DVLOG(1) << "No VA Surfaces available";
- return false;
- }
-
- linked_ptr<DecodeSurface> dec_surface(new DecodeSurface(
- poc, input_id, available_va_surfaces_.back()));
- available_va_surfaces_.pop_back();
-
- DVLOG(4) << "POC " << poc
- << " will use surface " << dec_surface->va_surface()->id();
-
- bool inserted = decode_surfaces_in_use_.insert(
- std::make_pair(poc, dec_surface)).second;
- DCHECK(inserted);
-
- return true;
-}
-
-void VaapiH264Decoder::UnassignSurfaceFromPoC(int poc) {
- DecSurfacesInUse::iterator it = decode_surfaces_in_use_.find(poc);
- if (it == decode_surfaces_in_use_.end()) {
- DVLOG(1) << "Asked to unassign an unassigned POC " << poc;
- return;
- }
-
- DVLOG(4) << "POC " << poc << " no longer using VA surface "
- << it->second->va_surface()->id();
-
- decode_surfaces_in_use_.erase(it);
-}
-
-bool VaapiH264Decoder::SendPPS() {
- const media::H264PPS* pps = parser_.GetPPS(curr_pps_id_);
- DCHECK(pps);
-
- const media::H264SPS* sps = parser_.GetSPS(pps->seq_parameter_set_id);
- DCHECK(sps);
-
- DCHECK(curr_pic_.get());
-
- VAPictureParameterBufferH264 pic_param;
- memset(&pic_param, 0, sizeof(VAPictureParameterBufferH264));
-
-#define FROM_SPS_TO_PP(a) pic_param.a = sps->a;
-#define FROM_SPS_TO_PP2(a, b) pic_param.b = sps->a;
- FROM_SPS_TO_PP2(pic_width_in_mbs_minus1, picture_width_in_mbs_minus1);
- // This assumes non-interlaced video
- FROM_SPS_TO_PP2(pic_height_in_map_units_minus1,
- picture_height_in_mbs_minus1);
- FROM_SPS_TO_PP(bit_depth_luma_minus8);
- FROM_SPS_TO_PP(bit_depth_chroma_minus8);
-#undef FROM_SPS_TO_PP
-#undef FROM_SPS_TO_PP2
-
-#define FROM_SPS_TO_PP_SF(a) pic_param.seq_fields.bits.a = sps->a;
-#define FROM_SPS_TO_PP_SF2(a, b) pic_param.seq_fields.bits.b = sps->a;
- FROM_SPS_TO_PP_SF(chroma_format_idc);
- FROM_SPS_TO_PP_SF2(separate_colour_plane_flag,
- residual_colour_transform_flag);
- FROM_SPS_TO_PP_SF(gaps_in_frame_num_value_allowed_flag);
- FROM_SPS_TO_PP_SF(frame_mbs_only_flag);
- FROM_SPS_TO_PP_SF(mb_adaptive_frame_field_flag);
- FROM_SPS_TO_PP_SF(direct_8x8_inference_flag);
- pic_param.seq_fields.bits.MinLumaBiPredSize8x8 = (sps->level_idc >= 31);
- FROM_SPS_TO_PP_SF(log2_max_frame_num_minus4);
- FROM_SPS_TO_PP_SF(pic_order_cnt_type);
- FROM_SPS_TO_PP_SF(log2_max_pic_order_cnt_lsb_minus4);
- FROM_SPS_TO_PP_SF(delta_pic_order_always_zero_flag);
-#undef FROM_SPS_TO_PP_SF
-#undef FROM_SPS_TO_PP_SF2
-
-#define FROM_PPS_TO_PP(a) pic_param.a = pps->a;
- FROM_PPS_TO_PP(num_slice_groups_minus1);
- pic_param.slice_group_map_type = 0;
- pic_param.slice_group_change_rate_minus1 = 0;
- FROM_PPS_TO_PP(pic_init_qp_minus26);
- FROM_PPS_TO_PP(pic_init_qs_minus26);
- FROM_PPS_TO_PP(chroma_qp_index_offset);
- FROM_PPS_TO_PP(second_chroma_qp_index_offset);
-#undef FROM_PPS_TO_PP
-
-#define FROM_PPS_TO_PP_PF(a) pic_param.pic_fields.bits.a = pps->a;
-#define FROM_PPS_TO_PP_PF2(a, b) pic_param.pic_fields.bits.b = pps->a;
- FROM_PPS_TO_PP_PF(entropy_coding_mode_flag);
- FROM_PPS_TO_PP_PF(weighted_pred_flag);
- FROM_PPS_TO_PP_PF(weighted_bipred_idc);
- FROM_PPS_TO_PP_PF(transform_8x8_mode_flag);
-
- pic_param.pic_fields.bits.field_pic_flag = 0;
- FROM_PPS_TO_PP_PF(constrained_intra_pred_flag);
- FROM_PPS_TO_PP_PF2(bottom_field_pic_order_in_frame_present_flag,
- pic_order_present_flag);
- FROM_PPS_TO_PP_PF(deblocking_filter_control_present_flag);
- FROM_PPS_TO_PP_PF(redundant_pic_cnt_present_flag);
- pic_param.pic_fields.bits.reference_pic_flag = curr_pic_->ref;
-#undef FROM_PPS_TO_PP_PF
-#undef FROM_PPS_TO_PP_PF2
-
- pic_param.frame_num = curr_pic_->frame_num;
-
- InitVAPicture(&pic_param.CurrPic);
- FillVAPicture(&pic_param.CurrPic, curr_pic_.get());
-
- // Init reference pictures' array.
- for (int i = 0; i < 16; ++i)
- InitVAPicture(&pic_param.ReferenceFrames[i]);
-
- // And fill it with picture info from DPB.
- FillVARefFramesFromDPB(pic_param.ReferenceFrames,
- arraysize(pic_param.ReferenceFrames));
-
- pic_param.num_ref_frames = sps->max_num_ref_frames;
-
- return vaapi_wrapper_->SubmitBuffer(VAPictureParameterBufferType,
- sizeof(VAPictureParameterBufferH264),
- &pic_param);
-}
-
-bool VaapiH264Decoder::SendIQMatrix() {
- const media::H264PPS* pps = parser_.GetPPS(curr_pps_id_);
- DCHECK(pps);
-
- VAIQMatrixBufferH264 iq_matrix_buf;
- memset(&iq_matrix_buf, 0, sizeof(VAIQMatrixBufferH264));
-
- if (pps->pic_scaling_matrix_present_flag) {
- for (int i = 0; i < 6; ++i) {
- for (int j = 0; j < 16; ++j)
- iq_matrix_buf.ScalingList4x4[i][j] = pps->scaling_list4x4[i][j];
- }
-
- for (int i = 0; i < 2; ++i) {
- for (int j = 0; j < 64; ++j)
- iq_matrix_buf.ScalingList8x8[i][j] = pps->scaling_list8x8[i][j];
- }
- } else {
- const media::H264SPS* sps = parser_.GetSPS(pps->seq_parameter_set_id);
- DCHECK(sps);
- for (int i = 0; i < 6; ++i) {
- for (int j = 0; j < 16; ++j)
- iq_matrix_buf.ScalingList4x4[i][j] = sps->scaling_list4x4[i][j];
- }
-
- for (int i = 0; i < 2; ++i) {
- for (int j = 0; j < 64; ++j)
- iq_matrix_buf.ScalingList8x8[i][j] = sps->scaling_list8x8[i][j];
- }
- }
-
- return vaapi_wrapper_->SubmitBuffer(VAIQMatrixBufferType,
- sizeof(VAIQMatrixBufferH264),
- &iq_matrix_buf);
-}
-
-bool VaapiH264Decoder::SendVASliceParam(media::H264SliceHeader* slice_hdr) {
- const media::H264PPS* pps = parser_.GetPPS(slice_hdr->pic_parameter_set_id);
- DCHECK(pps);
-
- const media::H264SPS* sps = parser_.GetSPS(pps->seq_parameter_set_id);
- DCHECK(sps);
-
- VASliceParameterBufferH264 slice_param;
- memset(&slice_param, 0, sizeof(VASliceParameterBufferH264));
-
- slice_param.slice_data_size = slice_hdr->nalu_size;
- slice_param.slice_data_offset = 0;
- slice_param.slice_data_flag = VA_SLICE_DATA_FLAG_ALL;
- slice_param.slice_data_bit_offset = slice_hdr->header_bit_size;
-
-#define SHDRToSP(a) slice_param.a = slice_hdr->a;
- SHDRToSP(first_mb_in_slice);
- slice_param.slice_type = slice_hdr->slice_type % 5;
- SHDRToSP(direct_spatial_mv_pred_flag);
-
- // TODO posciak: make sure parser sets those even when override flags
- // in slice header is off.
- SHDRToSP(num_ref_idx_l0_active_minus1);
- SHDRToSP(num_ref_idx_l1_active_minus1);
- SHDRToSP(cabac_init_idc);
- SHDRToSP(slice_qp_delta);
- SHDRToSP(disable_deblocking_filter_idc);
- SHDRToSP(slice_alpha_c0_offset_div2);
- SHDRToSP(slice_beta_offset_div2);
-
- if (((slice_hdr->IsPSlice() || slice_hdr->IsSPSlice()) &&
- pps->weighted_pred_flag) ||
- (slice_hdr->IsBSlice() && pps->weighted_bipred_idc == 1)) {
- SHDRToSP(luma_log2_weight_denom);
- SHDRToSP(chroma_log2_weight_denom);
-
- SHDRToSP(luma_weight_l0_flag);
- SHDRToSP(luma_weight_l1_flag);
-
- SHDRToSP(chroma_weight_l0_flag);
- SHDRToSP(chroma_weight_l1_flag);
-
- for (int i = 0; i <= slice_param.num_ref_idx_l0_active_minus1; ++i) {
- slice_param.luma_weight_l0[i] =
- slice_hdr->pred_weight_table_l0.luma_weight[i];
- slice_param.luma_offset_l0[i] =
- slice_hdr->pred_weight_table_l0.luma_offset[i];
-
- for (int j = 0; j < 2; ++j) {
- slice_param.chroma_weight_l0[i][j] =
- slice_hdr->pred_weight_table_l0.chroma_weight[i][j];
- slice_param.chroma_offset_l0[i][j] =
- slice_hdr->pred_weight_table_l0.chroma_offset[i][j];
- }
- }
-
- if (slice_hdr->IsBSlice()) {
- for (int i = 0; i <= slice_param.num_ref_idx_l1_active_minus1; ++i) {
- slice_param.luma_weight_l1[i] =
- slice_hdr->pred_weight_table_l1.luma_weight[i];
- slice_param.luma_offset_l1[i] =
- slice_hdr->pred_weight_table_l1.luma_offset[i];
-
- for (int j = 0; j < 2; ++j) {
- slice_param.chroma_weight_l1[i][j] =
- slice_hdr->pred_weight_table_l1.chroma_weight[i][j];
- slice_param.chroma_offset_l1[i][j] =
- slice_hdr->pred_weight_table_l1.chroma_offset[i][j];
- }
- }
- }
- }
-
- for (int i = 0; i < 32; ++i) {
- InitVAPicture(&slice_param.RefPicList0[i]);
- InitVAPicture(&slice_param.RefPicList1[i]);
- }
-
- int i;
- H264Picture::PtrVector::iterator it;
- for (it = ref_pic_list0_.begin(), i = 0; it != ref_pic_list0_.end() && *it;
- ++it, ++i)
- FillVAPicture(&slice_param.RefPicList0[i], *it);
- for (it = ref_pic_list1_.begin(), i = 0; it != ref_pic_list1_.end() && *it;
- ++it, ++i)
- FillVAPicture(&slice_param.RefPicList1[i], *it);
-
- return vaapi_wrapper_->SubmitBuffer(VASliceParameterBufferType,
- sizeof(VASliceParameterBufferH264),
- &slice_param);
-}
-
-bool VaapiH264Decoder::SendSliceData(const uint8* ptr, size_t size) {
- // Can't help it, blame libva...
- void* non_const_ptr = const_cast<uint8*>(ptr);
- return vaapi_wrapper_->SubmitBuffer(VASliceDataBufferType, size,
- non_const_ptr);
-}
-
-bool VaapiH264Decoder::PrepareRefPicLists(media::H264SliceHeader* slice_hdr) {
- ref_pic_list0_.clear();
- ref_pic_list1_.clear();
+bool H264Decoder::ModifyReferencePicLists(media::H264SliceHeader* slice_hdr,
+ H264Picture::Vector* ref_pic_list0,
+ H264Picture::Vector* ref_pic_list1) {
+ ref_pic_list0->clear();
+ ref_pic_list1->clear();
// Fill reference picture lists for B and S/SP slices.
if (slice_hdr->IsPSlice() || slice_hdr->IsSPSlice()) {
- ConstructReferencePicListsP(slice_hdr);
- return ModifyReferencePicList(slice_hdr, 0);
- }
-
- if (slice_hdr->IsBSlice()) {
- ConstructReferencePicListsB(slice_hdr);
- return ModifyReferencePicList(slice_hdr, 0) &&
- ModifyReferencePicList(slice_hdr, 1);
+ *ref_pic_list0 = ref_pic_list_p0_;
+ return ModifyReferencePicList(slice_hdr, 0, ref_pic_list0);
+ } else if (slice_hdr->IsBSlice()) {
+ *ref_pic_list0 = ref_pic_list_b0_;
+ *ref_pic_list1 = ref_pic_list_b1_;
+ return ModifyReferencePicList(slice_hdr, 0, ref_pic_list0) &&
+ ModifyReferencePicList(slice_hdr, 1, ref_pic_list1);
}
return true;
}
-bool VaapiH264Decoder::QueueSlice(media::H264SliceHeader* slice_hdr) {
- DCHECK(curr_pic_.get());
-
- if (!PrepareRefPicLists(slice_hdr))
- return false;
-
- if (!SendVASliceParam(slice_hdr))
- return false;
-
- if (!SendSliceData(slice_hdr->nalu_data, slice_hdr->nalu_size))
- return false;
-
- return true;
-}
-
-// TODO(posciak) start using vaMapBuffer instead of vaCreateBuffer wherever
-// possible.
-bool VaapiH264Decoder::DecodePicture() {
+bool H264Decoder::DecodePicture() {
DCHECK(curr_pic_.get());
DVLOG(4) << "Decoding POC " << curr_pic_->pic_order_cnt;
- DecodeSurface* dec_surface = DecodeSurfaceByPoC(curr_pic_->pic_order_cnt);
- if (!dec_surface) {
- DVLOG(1) << "Asked to decode an invalid POC " << curr_pic_->pic_order_cnt;
- return false;
- }
-
- if (!vaapi_wrapper_->ExecuteAndDestroyPendingBuffers(
- dec_surface->va_surface()->id())) {
- DVLOG(1) << "Failed decoding picture";
- return false;
- }
-
- return true;
+ return accelerator_->SubmitDecode(curr_pic_);
}
-bool VaapiH264Decoder::InitCurrPicture(media::H264SliceHeader* slice_hdr) {
+bool H264Decoder::InitCurrPicture(media::H264SliceHeader* slice_hdr) {
DCHECK(curr_pic_.get());
- memset(curr_pic_.get(), 0, sizeof(H264Picture));
-
curr_pic_->idr = slice_hdr->idr_pic_flag;
if (slice_hdr->field_pic_flag) {
@@ -518,12 +115,6 @@ bool VaapiH264Decoder::InitCurrPicture(media::H264SliceHeader* slice_hdr) {
if (!CalculatePicOrderCounts(slice_hdr))
return false;
- // Try to get an empty surface to decode this picture to.
- if (!AssignSurfaceToPoC(curr_input_id_, curr_pic_->pic_order_cnt)) {
- DVLOG(1) << "Failed getting a free surface for a picture";
- return false;
- }
-
curr_pic_->long_term_reference_flag = slice_hdr->long_term_reference_flag;
curr_pic_->adaptive_ref_pic_marking_mode_flag =
slice_hdr->adaptive_ref_pic_marking_mode_flag;
@@ -532,9 +123,9 @@ bool VaapiH264Decoder::InitCurrPicture(media::H264SliceHeader* slice_hdr) {
// process after this picture is decoded, store required data for that
// purpose.
if (slice_hdr->adaptive_ref_pic_marking_mode_flag) {
- static_assert(sizeof(curr_pic_->ref_pic_marking) ==
- sizeof(slice_hdr->ref_pic_marking),
- "ref_pic_marking array sizes do not match");
+ COMPILE_ASSERT(sizeof(curr_pic_->ref_pic_marking) ==
+ sizeof(slice_hdr->ref_pic_marking),
+ ref_pic_marking_array_sizes_do_not_match);
memcpy(curr_pic_->ref_pic_marking, slice_hdr->ref_pic_marking,
sizeof(curr_pic_->ref_pic_marking));
}
@@ -542,8 +133,7 @@ bool VaapiH264Decoder::InitCurrPicture(media::H264SliceHeader* slice_hdr) {
return true;
}
-bool VaapiH264Decoder::CalculatePicOrderCounts(
- media::H264SliceHeader* slice_hdr) {
+bool H264Decoder::CalculatePicOrderCounts(media::H264SliceHeader* slice_hdr) {
DCHECK_NE(curr_sps_id_, -1);
const media::H264SPS* sps = parser_.GetSPS(curr_sps_id_);
@@ -649,17 +239,17 @@ bool VaapiH264Decoder::CalculatePicOrderCounts(
if (!slice_hdr->field_pic_flag) {
curr_pic_->top_field_order_cnt = expected_pic_order_cnt +
- slice_hdr->delta_pic_order_cnt[0];
+ slice_hdr->delta_pic_order_cnt0;
curr_pic_->bottom_field_order_cnt = curr_pic_->top_field_order_cnt +
sps->offset_for_top_to_bottom_field +
- slice_hdr->delta_pic_order_cnt[1];
+ slice_hdr->delta_pic_order_cnt1;
} else if (!slice_hdr->bottom_field_flag) {
curr_pic_->top_field_order_cnt = expected_pic_order_cnt +
- slice_hdr->delta_pic_order_cnt[0];
+ slice_hdr->delta_pic_order_cnt0;
} else {
curr_pic_->bottom_field_order_cnt = expected_pic_order_cnt +
sps->offset_for_top_to_bottom_field +
- slice_hdr->delta_pic_order_cnt[0];
+ slice_hdr->delta_pic_order_cnt0;
}
break;
}
@@ -718,10 +308,8 @@ bool VaapiH264Decoder::CalculatePicOrderCounts(
return true;
}
-void VaapiH264Decoder::UpdatePicNums() {
- for (H264DPB::Pictures::iterator it = dpb_.begin(); it != dpb_.end(); ++it) {
- H264Picture* pic = *it;
- DCHECK(pic);
+void H264Decoder::UpdatePicNums() {
+ for (auto& pic : dpb_) {
if (!pic->ref)
continue;
@@ -741,75 +329,82 @@ void VaapiH264Decoder::UpdatePicNums() {
}
struct PicNumDescCompare {
- bool operator()(const H264Picture* a, const H264Picture* b) const {
+ bool operator()(const scoped_refptr<H264Picture>& a,
+ const scoped_refptr<H264Picture>& b) const {
return a->pic_num > b->pic_num;
}
};
struct LongTermPicNumAscCompare {
- bool operator()(const H264Picture* a, const H264Picture* b) const {
+ bool operator()(const scoped_refptr<H264Picture>& a,
+ const scoped_refptr<H264Picture>& b) const {
return a->long_term_pic_num < b->long_term_pic_num;
}
};
-void VaapiH264Decoder::ConstructReferencePicListsP(
+void H264Decoder::ConstructReferencePicListsP(
media::H264SliceHeader* slice_hdr) {
// RefPicList0 (8.2.4.2.1) [[1] [2]], where:
// [1] shortterm ref pics sorted by descending pic_num,
// [2] longterm ref pics by ascending long_term_pic_num.
- DCHECK(ref_pic_list0_.empty() && ref_pic_list1_.empty());
+ ref_pic_list_p0_.clear();
+
// First get the short ref pics...
- dpb_.GetShortTermRefPicsAppending(ref_pic_list0_);
- size_t num_short_refs = ref_pic_list0_.size();
+ dpb_.GetShortTermRefPicsAppending(&ref_pic_list_p0_);
+ size_t num_short_refs = ref_pic_list_p0_.size();
// and sort them to get [1].
- std::sort(ref_pic_list0_.begin(), ref_pic_list0_.end(), PicNumDescCompare());
+ std::sort(ref_pic_list_p0_.begin(), ref_pic_list_p0_.end(),
+ PicNumDescCompare());
// Now get long term pics and sort them by long_term_pic_num to get [2].
- dpb_.GetLongTermRefPicsAppending(ref_pic_list0_);
- std::sort(ref_pic_list0_.begin() + num_short_refs, ref_pic_list0_.end(),
+ dpb_.GetLongTermRefPicsAppending(&ref_pic_list_p0_);
+ std::sort(ref_pic_list_p0_.begin() + num_short_refs, ref_pic_list_p0_.end(),
LongTermPicNumAscCompare());
// Cut off if we have more than requested in slice header.
- ref_pic_list0_.resize(slice_hdr->num_ref_idx_l0_active_minus1 + 1);
+ ref_pic_list_p0_.resize(slice_hdr->num_ref_idx_l0_active_minus1 + 1);
}
struct POCAscCompare {
- bool operator()(const H264Picture* a, const H264Picture* b) const {
+ bool operator()(const scoped_refptr<H264Picture>& a,
+ const scoped_refptr<H264Picture>& b) const {
return a->pic_order_cnt < b->pic_order_cnt;
}
};
struct POCDescCompare {
- bool operator()(const H264Picture* a, const H264Picture* b) const {
+ bool operator()(const scoped_refptr<H264Picture>& a,
+ const scoped_refptr<H264Picture>& b) const {
return a->pic_order_cnt > b->pic_order_cnt;
}
};
-void VaapiH264Decoder::ConstructReferencePicListsB(
+void H264Decoder::ConstructReferencePicListsB(
media::H264SliceHeader* slice_hdr) {
// RefPicList0 (8.2.4.2.3) [[1] [2] [3]], where:
// [1] shortterm ref pics with POC < curr_pic's POC sorted by descending POC,
// [2] shortterm ref pics with POC > curr_pic's POC by ascending POC,
// [3] longterm ref pics by ascending long_term_pic_num.
- DCHECK(ref_pic_list0_.empty() && ref_pic_list1_.empty());
- dpb_.GetShortTermRefPicsAppending(ref_pic_list0_);
- size_t num_short_refs = ref_pic_list0_.size();
+ ref_pic_list_b0_.clear();
+ ref_pic_list_b1_.clear();
+ dpb_.GetShortTermRefPicsAppending(&ref_pic_list_b0_);
+ size_t num_short_refs = ref_pic_list_b0_.size();
// First sort ascending, this will put [1] in right place and finish [2].
- std::sort(ref_pic_list0_.begin(), ref_pic_list0_.end(), POCAscCompare());
+ std::sort(ref_pic_list_b0_.begin(), ref_pic_list_b0_.end(), POCAscCompare());
// Find first with POC > curr_pic's POC to get first element in [2]...
- H264Picture::PtrVector::iterator iter;
- iter = std::upper_bound(ref_pic_list0_.begin(), ref_pic_list0_.end(),
+ H264Picture::Vector::iterator iter;
+ iter = std::upper_bound(ref_pic_list_b0_.begin(), ref_pic_list_b0_.end(),
curr_pic_.get(), POCAscCompare());
// and sort [1] descending, thus finishing sequence [1] [2].
- std::sort(ref_pic_list0_.begin(), iter, POCDescCompare());
+ std::sort(ref_pic_list_b0_.begin(), iter, POCDescCompare());
// Now add [3] and sort by ascending long_term_pic_num.
- dpb_.GetLongTermRefPicsAppending(ref_pic_list0_);
- std::sort(ref_pic_list0_.begin() + num_short_refs, ref_pic_list0_.end(),
+ dpb_.GetLongTermRefPicsAppending(&ref_pic_list_b0_);
+ std::sort(ref_pic_list_b0_.begin() + num_short_refs, ref_pic_list_b0_.end(),
LongTermPicNumAscCompare());
// RefPicList1 (8.2.4.2.4) [[1] [2] [3]], where:
@@ -817,50 +412,50 @@ void VaapiH264Decoder::ConstructReferencePicListsB(
// [2] shortterm ref pics with POC < curr_pic's POC by descending POC,
// [3] longterm ref pics by ascending long_term_pic_num.
- dpb_.GetShortTermRefPicsAppending(ref_pic_list1_);
- num_short_refs = ref_pic_list1_.size();
+ dpb_.GetShortTermRefPicsAppending(&ref_pic_list_b1_);
+ num_short_refs = ref_pic_list_b1_.size();
// First sort by descending POC.
- std::sort(ref_pic_list1_.begin(), ref_pic_list1_.end(), POCDescCompare());
+ std::sort(ref_pic_list_b1_.begin(), ref_pic_list_b1_.end(), POCDescCompare());
// Find first with POC < curr_pic's POC to get first element in [2]...
- iter = std::upper_bound(ref_pic_list1_.begin(), ref_pic_list1_.end(),
+ iter = std::upper_bound(ref_pic_list_b1_.begin(), ref_pic_list_b1_.end(),
curr_pic_.get(), POCDescCompare());
// and sort [1] ascending.
- std::sort(ref_pic_list1_.begin(), iter, POCAscCompare());
+ std::sort(ref_pic_list_b1_.begin(), iter, POCAscCompare());
// Now add [3] and sort by ascending long_term_pic_num
- dpb_.GetShortTermRefPicsAppending(ref_pic_list1_);
- std::sort(ref_pic_list1_.begin() + num_short_refs, ref_pic_list1_.end(),
+ dpb_.GetShortTermRefPicsAppending(&ref_pic_list_b1_);
+ std::sort(ref_pic_list_b1_.begin() + num_short_refs, ref_pic_list_b1_.end(),
LongTermPicNumAscCompare());
// If lists identical, swap first two entries in RefPicList1 (spec 8.2.4.2.3)
- if (ref_pic_list1_.size() > 1 &&
- std::equal(ref_pic_list0_.begin(), ref_pic_list0_.end(),
- ref_pic_list1_.begin()))
- std::swap(ref_pic_list1_[0], ref_pic_list1_[1]);
+ if (ref_pic_list_b1_.size() > 1 &&
+ std::equal(ref_pic_list_b0_.begin(), ref_pic_list_b0_.end(),
+ ref_pic_list_b1_.begin()))
+ std::swap(ref_pic_list_b1_[0], ref_pic_list_b1_[1]);
// Per 8.2.4.2 it's possible for num_ref_idx_lX_active_minus1 to indicate
// there should be more ref pics on list than we constructed.
// Those superfluous ones should be treated as non-reference.
- ref_pic_list0_.resize(slice_hdr->num_ref_idx_l0_active_minus1 + 1);
- ref_pic_list1_.resize(slice_hdr->num_ref_idx_l1_active_minus1 + 1);
+ ref_pic_list_b0_.resize(slice_hdr->num_ref_idx_l0_active_minus1 + 1);
+ ref_pic_list_b1_.resize(slice_hdr->num_ref_idx_l1_active_minus1 + 1);
}
// See 8.2.4
-int VaapiH264Decoder::PicNumF(H264Picture *pic) {
+int H264Decoder::PicNumF(const scoped_refptr<H264Picture>& pic) {
if (!pic)
- return -1;
+ return -1;
if (!pic->long_term)
- return pic->pic_num;
+ return pic->pic_num;
else
- return max_pic_num_;
+ return max_pic_num_;
}
// See 8.2.4
-int VaapiH264Decoder::LongTermPicNumF(H264Picture *pic) {
+int H264Decoder::LongTermPicNumF(const scoped_refptr<H264Picture>& pic) {
if (pic->ref && pic->long_term)
return pic->long_term_pic_num;
else
@@ -869,10 +464,10 @@ int VaapiH264Decoder::LongTermPicNumF(H264Picture *pic) {
// Shift elements on the |v| starting from |from| to |to|, inclusive,
// one position to the right and insert pic at |from|.
-static void ShiftRightAndInsert(H264Picture::PtrVector *v,
+static void ShiftRightAndInsert(H264Picture::Vector* v,
int from,
int to,
- H264Picture* pic) {
+ const scoped_refptr<H264Picture>& pic) {
// Security checks, do not disable in Debug mode.
CHECK(from <= to);
CHECK(to <= std::numeric_limits<int>::max() - 2);
@@ -890,10 +485,10 @@ static void ShiftRightAndInsert(H264Picture::PtrVector *v,
(*v)[from] = pic;
}
-bool VaapiH264Decoder::ModifyReferencePicList(media::H264SliceHeader* slice_hdr,
- int list) {
+bool H264Decoder::ModifyReferencePicList(media::H264SliceHeader* slice_hdr,
+ int list,
+ H264Picture::Vector* ref_pic_listx) {
int num_ref_idx_lX_active_minus1;
- H264Picture::PtrVector* ref_pic_listx;
media::H264ModificationOfPicNum* list_mod;
// This can process either ref_pic_list0 or ref_pic_list1, depending on
@@ -903,19 +498,14 @@ bool VaapiH264Decoder::ModifyReferencePicList(media::H264SliceHeader* slice_hdr,
return true;
list_mod = slice_hdr->ref_list_l0_modifications;
- num_ref_idx_lX_active_minus1 = ref_pic_list0_.size() - 1;
-
- ref_pic_listx = &ref_pic_list0_;
} else {
if (!slice_hdr->ref_pic_list_modification_flag_l1)
return true;
list_mod = slice_hdr->ref_list_l1_modifications;
- num_ref_idx_lX_active_minus1 = ref_pic_list1_.size() - 1;
-
- ref_pic_listx = &ref_pic_list1_;
}
+ num_ref_idx_lX_active_minus1 = ref_pic_listx->size() - 1;
DCHECK_GE(num_ref_idx_lX_active_minus1, 0);
// Spec 8.2.4.3:
@@ -925,7 +515,7 @@ bool VaapiH264Decoder::ModifyReferencePicList(media::H264SliceHeader* slice_hdr,
int pic_num_lx_no_wrap;
int pic_num_lx;
bool done = false;
- H264Picture* pic;
+ scoped_refptr<H264Picture> pic;
for (int i = 0; i < media::H264SliceHeader::kRefListModSize && !done; ++i) {
switch (list_mod->modification_of_pic_nums_idc) {
case 0:
@@ -991,8 +581,8 @@ bool VaapiH264Decoder::ModifyReferencePicList(media::H264SliceHeader* slice_hdr,
for (int src = ref_idx_lx, dst = ref_idx_lx;
src <= num_ref_idx_lX_active_minus1 + 1; ++src) {
- if (LongTermPicNumF((*ref_pic_listx)[src])
- != static_cast<int>(list_mod->long_term_pic_num))
+ if (LongTermPicNumF((*ref_pic_listx)[src]) !=
+ static_cast<int>(list_mod->long_term_pic_num))
(*ref_pic_listx)[dst++] = (*ref_pic_listx)[src];
}
break;
@@ -1021,83 +611,48 @@ bool VaapiH264Decoder::ModifyReferencePicList(media::H264SliceHeader* slice_hdr,
return true;
}
-bool VaapiH264Decoder::OutputPic(H264Picture* pic) {
+void H264Decoder::OutputPic(scoped_refptr<H264Picture> pic) {
DCHECK(!pic->outputted);
pic->outputted = true;
last_output_poc_ = pic->pic_order_cnt;
- DecodeSurface* dec_surface = DecodeSurfaceByPoC(pic->pic_order_cnt);
- if (!dec_surface)
- return false;
-
- DCHECK_GE(dec_surface->input_id(), 0);
- DVLOG(4) << "Posting output task for POC: " << pic->pic_order_cnt
- << " input_id: " << dec_surface->input_id();
- output_pic_cb_.Run(dec_surface->input_id(), dec_surface->va_surface());
-
- return true;
+ DVLOG(4) << "Posting output task for POC: " << pic->pic_order_cnt;
+ accelerator_->OutputPicture(pic);
}
-void VaapiH264Decoder::ClearDPB() {
+void H264Decoder::ClearDPB() {
// Clear DPB contents, marking the pictures as unused first.
- for (H264DPB::Pictures::iterator it = dpb_.begin(); it != dpb_.end(); ++it)
- UnassignSurfaceFromPoC((*it)->pic_order_cnt);
-
dpb_.Clear();
last_output_poc_ = std::numeric_limits<int>::min();
}
-bool VaapiH264Decoder::OutputAllRemainingPics() {
+bool H264Decoder::OutputAllRemainingPics() {
// Output all pictures that are waiting to be outputted.
FinishPrevFrameIfPresent();
- H264Picture::PtrVector to_output;
- dpb_.GetNotOutputtedPicsAppending(to_output);
+ H264Picture::Vector to_output;
+ dpb_.GetNotOutputtedPicsAppending(&to_output);
// Sort them by ascending POC to output in order.
std::sort(to_output.begin(), to_output.end(), POCAscCompare());
- H264Picture::PtrVector::iterator it;
- for (it = to_output.begin(); it != to_output.end(); ++it) {
- if (!OutputPic(*it)) {
- DVLOG(1) << "Failed to output pic POC: " << (*it)->pic_order_cnt;
- return false;
- }
- }
+ for (auto& pic : to_output)
+ OutputPic(pic);
return true;
}
-bool VaapiH264Decoder::Flush() {
+bool H264Decoder::Flush() {
DVLOG(2) << "Decoder flush";
if (!OutputAllRemainingPics())
return false;
ClearDPB();
-
- DCHECK(decode_surfaces_in_use_.empty());
+ DVLOG(2) << "Decoder flush finished";
return true;
}
-bool VaapiH264Decoder::StartNewFrame(media::H264SliceHeader* slice_hdr) {
+bool H264Decoder::StartNewFrame(media::H264SliceHeader* slice_hdr) {
// TODO posciak: add handling of max_num_ref_frames per spec.
-
- // If the new frame is an IDR, output what's left to output and clear DPB
- if (slice_hdr->idr_pic_flag) {
- // (unless we are explicitly instructed not to do so).
- if (!slice_hdr->no_output_of_prior_pics_flag) {
- // Output DPB contents.
- if (!Flush())
- return false;
- }
- dpb_.Clear();
- last_output_poc_ = std::numeric_limits<int>::min();
- }
-
- // curr_pic_ should have either been added to DPB or discarded when finishing
- // the last frame. DPB is responsible for releasing that memory once it's
- // not needed anymore.
- DCHECK(!curr_pic_.get());
- curr_pic_.reset(new H264Picture);
CHECK(curr_pic_.get());
if (!InitCurrPicture(slice_hdr))
@@ -1106,27 +661,29 @@ bool VaapiH264Decoder::StartNewFrame(media::H264SliceHeader* slice_hdr) {
DCHECK_GT(max_frame_num_, 0);
UpdatePicNums();
+ DCHECK(slice_hdr);
+ PrepareRefPicLists(slice_hdr);
- // Send parameter buffers before each new picture, before the first slice.
- if (!SendPPS())
- return false;
-
- if (!SendIQMatrix())
- return false;
+ const media::H264PPS* pps = parser_.GetPPS(curr_pps_id_);
+ DCHECK(pps);
+ const media::H264SPS* sps = parser_.GetSPS(pps->seq_parameter_set_id);
+ DCHECK(sps);
- if (!QueueSlice(slice_hdr))
+ if (!accelerator_->SubmitFrameMetadata(sps, pps, dpb_, ref_pic_list_p0_,
+ ref_pic_list_b0_, ref_pic_list_b1_,
+ curr_pic_.get()))
return false;
return true;
}
-bool VaapiH264Decoder::HandleMemoryManagementOps() {
+bool H264Decoder::HandleMemoryManagementOps() {
// 8.2.5.4
for (unsigned int i = 0; i < arraysize(curr_pic_->ref_pic_marking); ++i) {
// Code below does not support interlaced stream (per-field pictures).
media::H264DecRefPicMarking* ref_pic_marking =
&curr_pic_->ref_pic_marking[i];
- H264Picture* to_mark;
+ scoped_refptr<H264Picture> to_mark;
int pic_num_x;
switch (ref_pic_marking->memory_mgmnt_control_operation) {
@@ -1138,7 +695,7 @@ bool VaapiH264Decoder::HandleMemoryManagementOps() {
// Mark a short term reference picture as unused so it can be removed
// if outputted.
pic_num_x = curr_pic_->pic_num -
- (ref_pic_marking->difference_of_pic_nums_minus1 + 1);
+ (ref_pic_marking->difference_of_pic_nums_minus1 + 1);
to_mark = dpb_.GetShortRefPicByPicNum(pic_num_x);
if (to_mark) {
to_mark->ref = false;
@@ -1164,7 +721,7 @@ bool VaapiH264Decoder::HandleMemoryManagementOps() {
case 3:
// Mark a short term reference picture as long term reference.
pic_num_x = curr_pic_->pic_num -
- (ref_pic_marking->difference_of_pic_nums_minus1 + 1);
+ (ref_pic_marking->difference_of_pic_nums_minus1 + 1);
to_mark = dpb_.GetShortRefPicByPicNum(pic_num_x);
if (to_mark) {
DCHECK(to_mark->ref && !to_mark->long_term);
@@ -1178,12 +735,12 @@ bool VaapiH264Decoder::HandleMemoryManagementOps() {
case 4: {
// Unmark all reference pictures with long_term_frame_idx over new max.
- max_long_term_frame_idx_
- = ref_pic_marking->max_long_term_frame_idx_plus1 - 1;
- H264Picture::PtrVector long_terms;
- dpb_.GetLongTermRefPicsAppending(long_terms);
+ max_long_term_frame_idx_ =
+ ref_pic_marking->max_long_term_frame_idx_plus1 - 1;
+ H264Picture::Vector long_terms;
+ dpb_.GetLongTermRefPicsAppending(&long_terms);
for (size_t i = 0; i < long_terms.size(); ++i) {
- H264Picture* pic = long_terms[i];
+ scoped_refptr<H264Picture>& pic = long_terms[i];
DCHECK(pic->ref && pic->long_term);
// Ok to cast, max_long_term_frame_idx is much smaller than 16bit.
if (pic->long_term_frame_idx >
@@ -1203,10 +760,10 @@ bool VaapiH264Decoder::HandleMemoryManagementOps() {
case 6: {
// Replace long term reference pictures with current picture.
// First unmark if any existing with this long_term_frame_idx...
- H264Picture::PtrVector long_terms;
- dpb_.GetLongTermRefPicsAppending(long_terms);
+ H264Picture::Vector long_terms;
+ dpb_.GetLongTermRefPicsAppending(&long_terms);
for (size_t i = 0; i < long_terms.size(); ++i) {
- H264Picture* pic = long_terms[i];
+ scoped_refptr<H264Picture>& pic = long_terms[i];
DCHECK(pic->ref && pic->long_term);
// Ok to cast, long_term_frame_idx is much smaller than 16bit.
if (pic->long_term_frame_idx ==
@@ -1235,7 +792,7 @@ bool VaapiH264Decoder::HandleMemoryManagementOps() {
// procedure to remove the oldest one.
// It also performs marking and unmarking pictures as reference.
// See spac 8.2.5.1.
-void VaapiH264Decoder::ReferencePictureMarking() {
+void H264Decoder::ReferencePictureMarking() {
if (curr_pic_->idr) {
// If current picture is an IDR, all reference pictures are unmarked.
dpb_.MarkAllUnusedForRef();
@@ -1254,9 +811,9 @@ void VaapiH264Decoder::ReferencePictureMarking() {
// ensure DPB doesn't overflow, discard oldest picture.
// See spec 8.2.5.3.
if (curr_pic_->field == H264Picture::FIELD_NONE) {
- DCHECK_LE(dpb_.CountRefPics(),
- std::max<int>(parser_.GetSPS(curr_sps_id_)->max_num_ref_frames,
- 1));
+ DCHECK_LE(
+ dpb_.CountRefPics(),
+ std::max<int>(parser_.GetSPS(curr_sps_id_)->max_num_ref_frames, 1));
if (dpb_.CountRefPics() ==
std::max<int>(parser_.GetSPS(curr_sps_id_)->max_num_ref_frames,
1)) {
@@ -1264,7 +821,8 @@ void VaapiH264Decoder::ReferencePictureMarking() {
// need to remove one of the short term ones.
// Find smallest frame_num_wrap short reference picture and mark
// it as unused.
- H264Picture* to_unmark = dpb_.GetLowestFrameNumWrapShortRefPic();
+ scoped_refptr<H264Picture> to_unmark =
+ dpb_.GetLowestFrameNumWrapShortRefPic();
if (to_unmark == NULL) {
DVLOG(1) << "Couldn't find a short ref picture to unmark";
return;
@@ -1274,7 +832,6 @@ void VaapiH264Decoder::ReferencePictureMarking() {
} else {
// Shouldn't get here.
DVLOG(1) << "Interlaced video not supported.";
- report_error_to_uma_cb_.Run(INTERLACED_STREAM);
}
} else {
// Stream has instructions how to discard pictures from DPB and how
@@ -1285,13 +842,12 @@ void VaapiH264Decoder::ReferencePictureMarking() {
} else {
// Shouldn't get here.
DVLOG(1) << "Interlaced video not supported.";
- report_error_to_uma_cb_.Run(INTERLACED_STREAM);
}
}
}
}
-bool VaapiH264Decoder::FinishPicture() {
+bool H264Decoder::FinishPicture() {
DCHECK(curr_pic_.get());
// Finish processing previous picture.
@@ -1310,27 +866,23 @@ bool VaapiH264Decoder::FinishPicture() {
// Remove unused (for reference or later output) pictures from DPB, marking
// them as such.
- for (H264DPB::Pictures::iterator it = dpb_.begin(); it != dpb_.end(); ++it) {
- if ((*it)->outputted && !(*it)->ref)
- UnassignSurfaceFromPoC((*it)->pic_order_cnt);
- }
dpb_.DeleteUnused();
DVLOG(4) << "Finishing picture, entries in DPB: " << dpb_.size();
// Whatever happens below, curr_pic_ will stop managing the pointer to the
- // picture after this function returns. The ownership will either be
- // transferred to DPB, if the image is still needed (for output and/or
- // reference), or the memory will be released if we manage to output it here
- // without having to store it for future reference.
- scoped_ptr<H264Picture> pic(curr_pic_.release());
+ // picture after this. The ownership will either be transferred to DPB, if
+ // the image is still needed (for output and/or reference), or the memory
+ // will be released if we manage to output it here without having to store
+ // it for future reference.
+ scoped_refptr<H264Picture> pic = curr_pic_;
+ curr_pic_ = nullptr;
// Get all pictures that haven't been outputted yet.
- H264Picture::PtrVector not_outputted;
- // TODO(posciak): pass as pointer, not reference (violates coding style).
- dpb_.GetNotOutputtedPicsAppending(not_outputted);
+ H264Picture::Vector not_outputted;
+ dpb_.GetNotOutputtedPicsAppending(&not_outputted);
// Include the one we've just decoded.
- not_outputted.push_back(pic.get());
+ not_outputted.push_back(pic);
// Sort in output order.
std::sort(not_outputted.begin(), not_outputted.end(), POCAscCompare());
@@ -1340,21 +892,18 @@ bool VaapiH264Decoder::FinishPicture() {
// in DPB afterwards would at least be equal to max_num_reorder_frames.
// If the outputted picture is not a reference picture, it doesn't have
// to remain in the DPB and can be removed.
- H264Picture::PtrVector::iterator output_candidate = not_outputted.begin();
+ H264Picture::Vector::iterator output_candidate = not_outputted.begin();
size_t num_remaining = not_outputted.size();
while (num_remaining > max_num_reorder_frames_) {
int poc = (*output_candidate)->pic_order_cnt;
DCHECK_GE(poc, last_output_poc_);
- if (!OutputPic(*output_candidate))
- return false;
+ OutputPic(*output_candidate);
if (!(*output_candidate)->ref) {
// Current picture hasn't been inserted into DPB yet, so don't remove it
// if we managed to output it immediately.
- if (*output_candidate != pic)
+ if ((*output_candidate)->pic_order_cnt != pic->pic_order_cnt)
dpb_.DeleteByPOC(poc);
- // Mark as unused.
- UnassignSurfaceFromPoC(poc);
}
++output_candidate;
@@ -1371,7 +920,7 @@ bool VaapiH264Decoder::FinishPicture() {
return false;
}
- dpb_.StorePic(pic.release());
+ dpb_.StorePic(pic);
}
return true;
@@ -1402,7 +951,7 @@ static int LevelToMaxDpbMbs(int level) {
}
}
-bool VaapiH264Decoder::UpdateMaxNumReorderFrames(const media::H264SPS* sps) {
+bool H264Decoder::UpdateMaxNumReorderFrames(const media::H264SPS* sps) {
if (sps->vui_parameters_present_flag && sps->bitstream_restriction_flag) {
max_num_reorder_frames_ =
base::checked_cast<size_t>(sps->max_num_reorder_frames);
@@ -1439,7 +988,7 @@ bool VaapiH264Decoder::UpdateMaxNumReorderFrames(const media::H264SPS* sps) {
return true;
}
-bool VaapiH264Decoder::ProcessSPS(int sps_id, bool* need_new_buffers) {
+bool H264Decoder::ProcessSPS(int sps_id, bool* need_new_buffers) {
const media::H264SPS* sps = parser_.GetSPS(sps_id);
DCHECK(sps);
DVLOG(4) << "Processing SPS";
@@ -1448,13 +997,11 @@ bool VaapiH264Decoder::ProcessSPS(int sps_id, bool* need_new_buffers) {
if (sps->frame_mbs_only_flag == 0) {
DVLOG(1) << "frame_mbs_only_flag != 1 not supported";
- report_error_to_uma_cb_.Run(FRAME_MBS_ONLY_FLAG_NOT_ONE);
return false;
}
if (sps->gaps_in_frame_num_value_allowed_flag) {
DVLOG(1) << "Gaps in frame numbers not supported";
- report_error_to_uma_cb_.Run(GAPS_IN_FRAME_NUM);
return false;
}
@@ -1464,7 +1011,7 @@ bool VaapiH264Decoder::ProcessSPS(int sps_id, bool* need_new_buffers) {
// (spec 7.4.2.1.1, 7.4.3).
int width_mb = sps->pic_width_in_mbs_minus1 + 1;
int height_mb = (2 - sps->frame_mbs_only_flag) *
- (sps->pic_height_in_map_units_minus1 + 1);
+ (sps->pic_height_in_map_units_minus1 + 1);
gfx::Size new_pic_size(16 * width_mb, 16 * height_mb);
if (new_pic_size.IsEmpty()) {
@@ -1507,7 +1054,7 @@ bool VaapiH264Decoder::ProcessSPS(int sps_id, bool* need_new_buffers) {
return true;
}
-bool VaapiH264Decoder::ProcessPPS(int pps_id) {
+bool H264Decoder::ProcessPPS(int pps_id) {
const media::H264PPS* pps = parser_.GetPPS(pps_id);
DCHECK(pps);
@@ -1516,7 +1063,7 @@ bool VaapiH264Decoder::ProcessPPS(int pps_id) {
return true;
}
-bool VaapiH264Decoder::FinishPrevFrameIfPresent() {
+bool H264Decoder::FinishPrevFrameIfPresent() {
// If we already have a frame waiting to be decoded, decode it and finish.
if (curr_pic_ != NULL) {
if (!DecodePicture())
@@ -1527,13 +1074,12 @@ bool VaapiH264Decoder::FinishPrevFrameIfPresent() {
return true;
}
-bool VaapiH264Decoder::ProcessSlice(media::H264SliceHeader* slice_hdr) {
+bool H264Decoder::PreprocessSlice(media::H264SliceHeader* slice_hdr) {
prev_frame_num_ = frame_num_;
frame_num_ = slice_hdr->frame_num;
if (prev_frame_num_ > 0 && prev_frame_num_ < frame_num_ - 1) {
DVLOG(1) << "Gap in frame_num!";
- report_error_to_uma_cb_.Run(GAPS_IN_FRAME_NUM);
return false;
}
@@ -1544,66 +1090,81 @@ bool VaapiH264Decoder::ProcessSlice(media::H264SliceHeader* slice_hdr) {
// TODO posciak: switch to new picture detection per 7.4.1.2.4.
if (curr_pic_ != NULL && slice_hdr->first_mb_in_slice != 0) {
- // This is just some more slice data of the current picture, so
- // just queue it and return.
- QueueSlice(slice_hdr);
+ // More slice data of the current picture.
return true;
} else {
// A new frame, so first finish the previous one before processing it...
if (!FinishPrevFrameIfPresent())
return false;
+ }
- // and then start a new one.
- return StartNewFrame(slice_hdr);
+ // If the new frame is an IDR, output what's left to output and clear DPB
+ if (slice_hdr->idr_pic_flag) {
+ // (unless we are explicitly instructed not to do so).
+ if (!slice_hdr->no_output_of_prior_pics_flag) {
+ // Output DPB contents.
+ if (!Flush())
+ return false;
+ }
+ dpb_.Clear();
+ last_output_poc_ = std::numeric_limits<int>::min();
}
+
+ return true;
}
-#define SET_ERROR_AND_RETURN() \
- do { \
- DVLOG(1) << "Error during decode"; \
- state_ = kError; \
- return VaapiH264Decoder::kDecodeError; \
+bool H264Decoder::ProcessSlice(media::H264SliceHeader* slice_hdr) {
+ DCHECK(curr_pic_.get());
+ H264Picture::Vector ref_pic_list0, ref_pic_list1;
+
+ if (!ModifyReferencePicLists(slice_hdr, &ref_pic_list0, &ref_pic_list1))
+ return false;
+
+ const media::H264PPS* pps = parser_.GetPPS(slice_hdr->pic_parameter_set_id);
+ DCHECK(pps);
+
+ if (!accelerator_->SubmitSlice(pps, slice_hdr, ref_pic_list0, ref_pic_list1,
+ curr_pic_.get(), slice_hdr->nalu_data,
+ slice_hdr->nalu_size))
+ return false;
+
+ curr_slice_hdr_.reset();
+ return true;
+}
+
+#define SET_ERROR_AND_RETURN() \
+ do { \
+ DVLOG(1) << "Error during decode"; \
+ state_ = kError; \
+ return H264Decoder::kDecodeError; \
} while (0)
-void VaapiH264Decoder::SetStream(const uint8* ptr,
- size_t size,
- int32 input_id) {
+void H264Decoder::SetStream(const uint8_t* ptr, size_t size) {
DCHECK(ptr);
DCHECK(size);
- // Got new input stream data from the client.
- DVLOG(4) << "New input stream id: " << input_id << " at: " << (void*) ptr
- << " size: " << size;
+ DVLOG(4) << "New input stream at: " << (void*)ptr << " size: " << size;
parser_.SetStream(ptr, size);
- curr_input_id_ = input_id;
}
-VaapiH264Decoder::DecResult VaapiH264Decoder::Decode() {
- media::H264Parser::Result par_res;
- media::H264NALU nalu;
+H264Decoder::DecodeResult H264Decoder::Decode() {
DCHECK_NE(state_, kError);
while (1) {
- // If we've already decoded some of the stream (after reset, i.e. we are
- // not in kNeedStreamMetadata state), we may be able to go back into
- // decoding state not only starting at/resuming from an SPS, but also from
- // other resume points, such as IDRs. In the latter case we need an output
- // surface, because we will end up decoding that IDR in the process.
- // Otherwise we just look for an SPS and don't produce any output frames.
- if (state_ != kNeedStreamMetadata && available_va_surfaces_.empty()) {
- DVLOG(4) << "No output surfaces available";
- return kRanOutOfSurfaces;
+ media::H264Parser::Result par_res;
+
+ if (!curr_nalu_) {
+ curr_nalu_.reset(new media::H264NALU());
+ par_res = parser_.AdvanceToNextNALU(curr_nalu_.get());
+ if (par_res == media::H264Parser::kEOStream)
+ return kRanOutOfStreamData;
+ else if (par_res != media::H264Parser::kOk)
+ SET_ERROR_AND_RETURN();
}
- par_res = parser_.AdvanceToNextNALU(&nalu);
- if (par_res == media::H264Parser::kEOStream)
- return kRanOutOfStreamData;
- else if (par_res != media::H264Parser::kOk)
- SET_ERROR_AND_RETURN();
+ DVLOG(4) << "NALU found: " << static_cast<int>(curr_nalu_->nal_unit_type);
- DVLOG(4) << "NALU found: " << static_cast<int>(nalu.nal_unit_type);
-
- switch (nalu.nal_unit_type) {
+ switch (curr_nalu_->nal_unit_type) {
case media::H264NALU::kNonIDRSlice:
// We can't resume from a non-IDR slice.
if (state_ != kDecoding)
@@ -1619,13 +1180,29 @@ VaapiH264Decoder::DecResult VaapiH264Decoder::Decode() {
}
// If after reset, we should be able to recover from an IDR.
- media::H264SliceHeader slice_hdr;
+ if (!curr_slice_hdr_) {
+ curr_slice_hdr_.reset(new media::H264SliceHeader());
+ par_res =
+ parser_.ParseSliceHeader(*curr_nalu_, curr_slice_hdr_.get());
+ if (par_res != media::H264Parser::kOk)
+ SET_ERROR_AND_RETURN();
+
+ if (!PreprocessSlice(curr_slice_hdr_.get()))
+ SET_ERROR_AND_RETURN();
+ }
- par_res = parser_.ParseSliceHeader(nalu, &slice_hdr);
- if (par_res != media::H264Parser::kOk)
- SET_ERROR_AND_RETURN();
+ if (!curr_pic_) {
+ // New picture/finished previous one, try to start a new one
+ // or tell the client we need more surfaces.
+ curr_pic_ = accelerator_->CreateH264Picture();
+ if (!curr_pic_)
+ return kRanOutOfSurfaces;
+
+ if (!StartNewFrame(curr_slice_hdr_.get()))
+ SET_ERROR_AND_RETURN();
+ }
- if (!ProcessSlice(&slice_hdr))
+ if (!ProcessSlice(curr_slice_hdr_.get()))
SET_ERROR_AND_RETURN();
state_ = kDecoding;
@@ -1652,7 +1229,12 @@ VaapiH264Decoder::DecResult VaapiH264Decoder::Decode() {
if (!Flush())
return kDecodeError;
- available_va_surfaces_.clear();
+ curr_pic_ = nullptr;
+ curr_nalu_ = nullptr;
+ ref_pic_list_p0_.clear();
+ ref_pic_list_b0_.clear();
+ ref_pic_list_b1_.clear();
+
return kAllocateNewSurfaces;
}
break;
@@ -1676,24 +1258,17 @@ VaapiH264Decoder::DecResult VaapiH264Decoder::Decode() {
break;
}
- case media::H264NALU::kAUD:
- case media::H264NALU::kEOSeq:
- case media::H264NALU::kEOStream:
- if (state_ != kDecoding)
- break;
- if (!FinishPrevFrameIfPresent())
- SET_ERROR_AND_RETURN();
-
- break;
-
default:
- DVLOG(4) << "Skipping NALU type: " << nalu.nal_unit_type;
+ DVLOG(4) << "Skipping NALU type: " << curr_nalu_->nal_unit_type;
break;
}
+
+ DVLOG(4) << "Dropping nalu";
+ curr_nalu_.reset();
}
}
-size_t VaapiH264Decoder::GetRequiredNumOfPictures() {
+size_t H264Decoder::GetRequiredNumOfPictures() const {
return dpb_.max_num_pics() + kPicsInPipeline;
}
« no previous file with comments | « content/common/gpu/media/h264_decoder.h ('k') | content/common/gpu/media/h264_dpb.h » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698