Index: content/common/gpu/media/h264_decoder.cc |
diff --git a/content/common/gpu/media/vaapi_h264_decoder.cc b/content/common/gpu/media/h264_decoder.cc |
similarity index 56% |
copy from content/common/gpu/media/vaapi_h264_decoder.cc |
copy to content/common/gpu/media/h264_decoder.cc |
index 55b0d007cda566936c3ed0c26043f3a5dfd8f16e..af36b03e7cdb8f5c55c0a3dd03180b6c36ee3dfa 100644 |
--- a/content/common/gpu/media/vaapi_h264_decoder.cc |
+++ b/content/common/gpu/media/h264_decoder.cc |
@@ -7,58 +7,20 @@ |
#include "base/bind.h" |
#include "base/bind_helpers.h" |
+#include "base/callback_helpers.h" |
#include "base/numerics/safe_conversions.h" |
#include "base/stl_util.h" |
-#include "content/common/gpu/media/vaapi_h264_decoder.h" |
+#include "content/common/gpu/media/h264_decoder.h" |
namespace content { |
-// Decode surface, used for decoding and reference. input_id comes from client |
-// and is associated with the surface that was produced as the result |
-// of decoding a bitstream buffer with that id. |
-class VaapiH264Decoder::DecodeSurface { |
- public: |
- DecodeSurface(int poc, |
- int32 input_id, |
- const scoped_refptr<VASurface>& va_surface); |
- DecodeSurface(int poc, const scoped_refptr<DecodeSurface>& dec_surface); |
- ~DecodeSurface(); |
- |
- int poc() { |
- return poc_; |
- } |
- |
- scoped_refptr<VASurface> va_surface() { |
- return va_surface_; |
- } |
- |
- int32 input_id() { |
- return input_id_; |
- } |
- |
- private: |
- int poc_; |
- int32 input_id_; |
- scoped_refptr<VASurface> va_surface_; |
-}; |
- |
-VaapiH264Decoder::DecodeSurface::DecodeSurface( |
- int poc, |
- int32 input_id, |
- const scoped_refptr<VASurface>& va_surface) |
- : poc_(poc), |
- input_id_(input_id), |
- va_surface_(va_surface) { |
- DCHECK(va_surface_.get()); |
+H264Decoder::H264Accelerator::H264Accelerator() { |
} |
-VaapiH264Decoder::DecodeSurface::~DecodeSurface() { |
+H264Decoder::H264Accelerator::~H264Accelerator() { |
} |
-VaapiH264Decoder::VaapiH264Decoder( |
- VaapiWrapper* vaapi_wrapper, |
- const OutputPicCB& output_pic_cb, |
- const ReportErrorToUmaCB& report_error_to_uma_cb) |
+H264Decoder::H264Decoder(H264Accelerator* accelerator) |
: max_pic_order_cnt_lsb_(0), |
max_frame_num_(0), |
max_pic_num_(0), |
@@ -66,20 +28,20 @@ VaapiH264Decoder::VaapiH264Decoder( |
max_num_reorder_frames_(0), |
curr_sps_id_(-1), |
curr_pps_id_(-1), |
- vaapi_wrapper_(vaapi_wrapper), |
- output_pic_cb_(output_pic_cb), |
- report_error_to_uma_cb_(report_error_to_uma_cb) { |
+ accelerator_(accelerator) { |
+ DCHECK(accelerator_); |
Reset(); |
state_ = kNeedStreamMetadata; |
} |
-VaapiH264Decoder::~VaapiH264Decoder() { |
+H264Decoder::~H264Decoder() { |
} |
-void VaapiH264Decoder::Reset() { |
- curr_pic_.reset(); |
+void H264Decoder::Reset() { |
+ curr_pic_ = nullptr; |
+ curr_nalu_ = nullptr; |
+ curr_slice_hdr_ = nullptr; |
- curr_input_id_ = -1; |
frame_num_ = 0; |
prev_frame_num_ = -1; |
prev_frame_num_offset_ = -1; |
@@ -90,21 +52,9 @@ void VaapiH264Decoder::Reset() { |
prev_ref_pic_order_cnt_lsb_ = -1; |
prev_ref_field_ = H264Picture::FIELD_NONE; |
- vaapi_wrapper_->DestroyPendingBuffers(); |
- |
- ref_pic_list0_.clear(); |
- ref_pic_list1_.clear(); |
- |
- for (DecSurfacesInUse::iterator it = decode_surfaces_in_use_.begin(); |
- it != decode_surfaces_in_use_.end(); ) { |
- int poc = it->second->poc(); |
- // Must be incremented before UnassignSurfaceFromPoC as this call |
- // invalidates |it|. |
- ++it; |
- UnassignSurfaceFromPoC(poc); |
- } |
- DCHECK(decode_surfaces_in_use_.empty()); |
- |
+ ref_pic_list_p0_.clear(); |
+ ref_pic_list_b0_.clear(); |
+ ref_pic_list_b1_.clear(); |
dpb_.Clear(); |
parser_.Reset(); |
last_output_poc_ = std::numeric_limits<int>::min(); |
@@ -114,394 +64,41 @@ void VaapiH264Decoder::Reset() { |
state_ = kAfterReset; |
} |
-void VaapiH264Decoder::ReuseSurface( |
- const scoped_refptr<VASurface>& va_surface) { |
- available_va_surfaces_.push_back(va_surface); |
-} |
- |
-// Fill |va_pic| with default/neutral values. |
-static void InitVAPicture(VAPictureH264* va_pic) { |
- memset(va_pic, 0, sizeof(*va_pic)); |
- va_pic->picture_id = VA_INVALID_ID; |
- va_pic->flags = VA_PICTURE_H264_INVALID; |
-} |
- |
-void VaapiH264Decoder::FillVAPicture(VAPictureH264 *va_pic, H264Picture* pic) { |
- DCHECK(pic); |
- |
- DecodeSurface* dec_surface = DecodeSurfaceByPoC(pic->pic_order_cnt); |
- if (!dec_surface) { |
- // Cannot provide a ref picture, will corrupt output, but may be able |
- // to recover. |
- InitVAPicture(va_pic); |
- return; |
- } |
- |
- va_pic->picture_id = dec_surface->va_surface()->id(); |
- va_pic->frame_idx = pic->frame_num; |
- va_pic->flags = 0; |
- |
- switch (pic->field) { |
- case H264Picture::FIELD_NONE: |
- break; |
- case H264Picture::FIELD_TOP: |
- va_pic->flags |= VA_PICTURE_H264_TOP_FIELD; |
- break; |
- case H264Picture::FIELD_BOTTOM: |
- va_pic->flags |= VA_PICTURE_H264_BOTTOM_FIELD; |
- break; |
- } |
- |
- if (pic->ref) { |
- va_pic->flags |= pic->long_term ? VA_PICTURE_H264_LONG_TERM_REFERENCE |
- : VA_PICTURE_H264_SHORT_TERM_REFERENCE; |
- } |
- |
- va_pic->TopFieldOrderCnt = pic->top_field_order_cnt; |
- va_pic->BottomFieldOrderCnt = pic->bottom_field_order_cnt; |
-} |
- |
-int VaapiH264Decoder::FillVARefFramesFromDPB(VAPictureH264 *va_pics, |
- int num_pics) { |
- H264DPB::Pictures::reverse_iterator rit; |
- int i; |
- |
- // Return reference frames in reverse order of insertion. |
- // Libva does not document this, but other implementations (e.g. mplayer) |
- // do it this way as well. |
- for (rit = dpb_.rbegin(), i = 0; rit != dpb_.rend() && i < num_pics; ++rit) { |
- if ((*rit)->ref) |
- FillVAPicture(&va_pics[i++], *rit); |
- } |
- |
- return i; |
+void H264Decoder::PrepareRefPicLists(media::H264SliceHeader* slice_hdr) { |
+ ConstructReferencePicListsP(slice_hdr); |
+ ConstructReferencePicListsB(slice_hdr); |
} |
-VaapiH264Decoder::DecodeSurface* VaapiH264Decoder::DecodeSurfaceByPoC(int poc) { |
- DecSurfacesInUse::iterator iter = decode_surfaces_in_use_.find(poc); |
- if (iter == decode_surfaces_in_use_.end()) { |
- DVLOG(1) << "Could not find surface assigned to POC: " << poc; |
- return NULL; |
- } |
- |
- return iter->second.get(); |
-} |
- |
-bool VaapiH264Decoder::AssignSurfaceToPoC(int32 input_id, int poc) { |
- if (available_va_surfaces_.empty()) { |
- DVLOG(1) << "No VA Surfaces available"; |
- return false; |
- } |
- |
- linked_ptr<DecodeSurface> dec_surface(new DecodeSurface( |
- poc, input_id, available_va_surfaces_.back())); |
- available_va_surfaces_.pop_back(); |
- |
- DVLOG(4) << "POC " << poc |
- << " will use surface " << dec_surface->va_surface()->id(); |
- |
- bool inserted = decode_surfaces_in_use_.insert( |
- std::make_pair(poc, dec_surface)).second; |
- DCHECK(inserted); |
- |
- return true; |
-} |
- |
-void VaapiH264Decoder::UnassignSurfaceFromPoC(int poc) { |
- DecSurfacesInUse::iterator it = decode_surfaces_in_use_.find(poc); |
- if (it == decode_surfaces_in_use_.end()) { |
- DVLOG(1) << "Asked to unassign an unassigned POC " << poc; |
- return; |
- } |
- |
- DVLOG(4) << "POC " << poc << " no longer using VA surface " |
- << it->second->va_surface()->id(); |
- |
- decode_surfaces_in_use_.erase(it); |
-} |
- |
-bool VaapiH264Decoder::SendPPS() { |
- const media::H264PPS* pps = parser_.GetPPS(curr_pps_id_); |
- DCHECK(pps); |
- |
- const media::H264SPS* sps = parser_.GetSPS(pps->seq_parameter_set_id); |
- DCHECK(sps); |
- |
- DCHECK(curr_pic_.get()); |
- |
- VAPictureParameterBufferH264 pic_param; |
- memset(&pic_param, 0, sizeof(VAPictureParameterBufferH264)); |
- |
-#define FROM_SPS_TO_PP(a) pic_param.a = sps->a; |
-#define FROM_SPS_TO_PP2(a, b) pic_param.b = sps->a; |
- FROM_SPS_TO_PP2(pic_width_in_mbs_minus1, picture_width_in_mbs_minus1); |
- // This assumes non-interlaced video |
- FROM_SPS_TO_PP2(pic_height_in_map_units_minus1, |
- picture_height_in_mbs_minus1); |
- FROM_SPS_TO_PP(bit_depth_luma_minus8); |
- FROM_SPS_TO_PP(bit_depth_chroma_minus8); |
-#undef FROM_SPS_TO_PP |
-#undef FROM_SPS_TO_PP2 |
- |
-#define FROM_SPS_TO_PP_SF(a) pic_param.seq_fields.bits.a = sps->a; |
-#define FROM_SPS_TO_PP_SF2(a, b) pic_param.seq_fields.bits.b = sps->a; |
- FROM_SPS_TO_PP_SF(chroma_format_idc); |
- FROM_SPS_TO_PP_SF2(separate_colour_plane_flag, |
- residual_colour_transform_flag); |
- FROM_SPS_TO_PP_SF(gaps_in_frame_num_value_allowed_flag); |
- FROM_SPS_TO_PP_SF(frame_mbs_only_flag); |
- FROM_SPS_TO_PP_SF(mb_adaptive_frame_field_flag); |
- FROM_SPS_TO_PP_SF(direct_8x8_inference_flag); |
- pic_param.seq_fields.bits.MinLumaBiPredSize8x8 = (sps->level_idc >= 31); |
- FROM_SPS_TO_PP_SF(log2_max_frame_num_minus4); |
- FROM_SPS_TO_PP_SF(pic_order_cnt_type); |
- FROM_SPS_TO_PP_SF(log2_max_pic_order_cnt_lsb_minus4); |
- FROM_SPS_TO_PP_SF(delta_pic_order_always_zero_flag); |
-#undef FROM_SPS_TO_PP_SF |
-#undef FROM_SPS_TO_PP_SF2 |
- |
-#define FROM_PPS_TO_PP(a) pic_param.a = pps->a; |
- FROM_PPS_TO_PP(num_slice_groups_minus1); |
- pic_param.slice_group_map_type = 0; |
- pic_param.slice_group_change_rate_minus1 = 0; |
- FROM_PPS_TO_PP(pic_init_qp_minus26); |
- FROM_PPS_TO_PP(pic_init_qs_minus26); |
- FROM_PPS_TO_PP(chroma_qp_index_offset); |
- FROM_PPS_TO_PP(second_chroma_qp_index_offset); |
-#undef FROM_PPS_TO_PP |
- |
-#define FROM_PPS_TO_PP_PF(a) pic_param.pic_fields.bits.a = pps->a; |
-#define FROM_PPS_TO_PP_PF2(a, b) pic_param.pic_fields.bits.b = pps->a; |
- FROM_PPS_TO_PP_PF(entropy_coding_mode_flag); |
- FROM_PPS_TO_PP_PF(weighted_pred_flag); |
- FROM_PPS_TO_PP_PF(weighted_bipred_idc); |
- FROM_PPS_TO_PP_PF(transform_8x8_mode_flag); |
- |
- pic_param.pic_fields.bits.field_pic_flag = 0; |
- FROM_PPS_TO_PP_PF(constrained_intra_pred_flag); |
- FROM_PPS_TO_PP_PF2(bottom_field_pic_order_in_frame_present_flag, |
- pic_order_present_flag); |
- FROM_PPS_TO_PP_PF(deblocking_filter_control_present_flag); |
- FROM_PPS_TO_PP_PF(redundant_pic_cnt_present_flag); |
- pic_param.pic_fields.bits.reference_pic_flag = curr_pic_->ref; |
-#undef FROM_PPS_TO_PP_PF |
-#undef FROM_PPS_TO_PP_PF2 |
- |
- pic_param.frame_num = curr_pic_->frame_num; |
- |
- InitVAPicture(&pic_param.CurrPic); |
- FillVAPicture(&pic_param.CurrPic, curr_pic_.get()); |
- |
- // Init reference pictures' array. |
- for (int i = 0; i < 16; ++i) |
- InitVAPicture(&pic_param.ReferenceFrames[i]); |
- |
- // And fill it with picture info from DPB. |
- FillVARefFramesFromDPB(pic_param.ReferenceFrames, |
- arraysize(pic_param.ReferenceFrames)); |
- |
- pic_param.num_ref_frames = sps->max_num_ref_frames; |
- |
- return vaapi_wrapper_->SubmitBuffer(VAPictureParameterBufferType, |
- sizeof(VAPictureParameterBufferH264), |
- &pic_param); |
-} |
- |
-bool VaapiH264Decoder::SendIQMatrix() { |
- const media::H264PPS* pps = parser_.GetPPS(curr_pps_id_); |
- DCHECK(pps); |
- |
- VAIQMatrixBufferH264 iq_matrix_buf; |
- memset(&iq_matrix_buf, 0, sizeof(VAIQMatrixBufferH264)); |
- |
- if (pps->pic_scaling_matrix_present_flag) { |
- for (int i = 0; i < 6; ++i) { |
- for (int j = 0; j < 16; ++j) |
- iq_matrix_buf.ScalingList4x4[i][j] = pps->scaling_list4x4[i][j]; |
- } |
- |
- for (int i = 0; i < 2; ++i) { |
- for (int j = 0; j < 64; ++j) |
- iq_matrix_buf.ScalingList8x8[i][j] = pps->scaling_list8x8[i][j]; |
- } |
- } else { |
- const media::H264SPS* sps = parser_.GetSPS(pps->seq_parameter_set_id); |
- DCHECK(sps); |
- for (int i = 0; i < 6; ++i) { |
- for (int j = 0; j < 16; ++j) |
- iq_matrix_buf.ScalingList4x4[i][j] = sps->scaling_list4x4[i][j]; |
- } |
- |
- for (int i = 0; i < 2; ++i) { |
- for (int j = 0; j < 64; ++j) |
- iq_matrix_buf.ScalingList8x8[i][j] = sps->scaling_list8x8[i][j]; |
- } |
- } |
- |
- return vaapi_wrapper_->SubmitBuffer(VAIQMatrixBufferType, |
- sizeof(VAIQMatrixBufferH264), |
- &iq_matrix_buf); |
-} |
- |
-bool VaapiH264Decoder::SendVASliceParam(media::H264SliceHeader* slice_hdr) { |
- const media::H264PPS* pps = parser_.GetPPS(slice_hdr->pic_parameter_set_id); |
- DCHECK(pps); |
- |
- const media::H264SPS* sps = parser_.GetSPS(pps->seq_parameter_set_id); |
- DCHECK(sps); |
- |
- VASliceParameterBufferH264 slice_param; |
- memset(&slice_param, 0, sizeof(VASliceParameterBufferH264)); |
- |
- slice_param.slice_data_size = slice_hdr->nalu_size; |
- slice_param.slice_data_offset = 0; |
- slice_param.slice_data_flag = VA_SLICE_DATA_FLAG_ALL; |
- slice_param.slice_data_bit_offset = slice_hdr->header_bit_size; |
- |
-#define SHDRToSP(a) slice_param.a = slice_hdr->a; |
- SHDRToSP(first_mb_in_slice); |
- slice_param.slice_type = slice_hdr->slice_type % 5; |
- SHDRToSP(direct_spatial_mv_pred_flag); |
- |
- // TODO posciak: make sure parser sets those even when override flags |
- // in slice header is off. |
- SHDRToSP(num_ref_idx_l0_active_minus1); |
- SHDRToSP(num_ref_idx_l1_active_minus1); |
- SHDRToSP(cabac_init_idc); |
- SHDRToSP(slice_qp_delta); |
- SHDRToSP(disable_deblocking_filter_idc); |
- SHDRToSP(slice_alpha_c0_offset_div2); |
- SHDRToSP(slice_beta_offset_div2); |
- |
- if (((slice_hdr->IsPSlice() || slice_hdr->IsSPSlice()) && |
- pps->weighted_pred_flag) || |
- (slice_hdr->IsBSlice() && pps->weighted_bipred_idc == 1)) { |
- SHDRToSP(luma_log2_weight_denom); |
- SHDRToSP(chroma_log2_weight_denom); |
- |
- SHDRToSP(luma_weight_l0_flag); |
- SHDRToSP(luma_weight_l1_flag); |
- |
- SHDRToSP(chroma_weight_l0_flag); |
- SHDRToSP(chroma_weight_l1_flag); |
- |
- for (int i = 0; i <= slice_param.num_ref_idx_l0_active_minus1; ++i) { |
- slice_param.luma_weight_l0[i] = |
- slice_hdr->pred_weight_table_l0.luma_weight[i]; |
- slice_param.luma_offset_l0[i] = |
- slice_hdr->pred_weight_table_l0.luma_offset[i]; |
- |
- for (int j = 0; j < 2; ++j) { |
- slice_param.chroma_weight_l0[i][j] = |
- slice_hdr->pred_weight_table_l0.chroma_weight[i][j]; |
- slice_param.chroma_offset_l0[i][j] = |
- slice_hdr->pred_weight_table_l0.chroma_offset[i][j]; |
- } |
- } |
- |
- if (slice_hdr->IsBSlice()) { |
- for (int i = 0; i <= slice_param.num_ref_idx_l1_active_minus1; ++i) { |
- slice_param.luma_weight_l1[i] = |
- slice_hdr->pred_weight_table_l1.luma_weight[i]; |
- slice_param.luma_offset_l1[i] = |
- slice_hdr->pred_weight_table_l1.luma_offset[i]; |
- |
- for (int j = 0; j < 2; ++j) { |
- slice_param.chroma_weight_l1[i][j] = |
- slice_hdr->pred_weight_table_l1.chroma_weight[i][j]; |
- slice_param.chroma_offset_l1[i][j] = |
- slice_hdr->pred_weight_table_l1.chroma_offset[i][j]; |
- } |
- } |
- } |
- } |
- |
- for (int i = 0; i < 32; ++i) { |
- InitVAPicture(&slice_param.RefPicList0[i]); |
- InitVAPicture(&slice_param.RefPicList1[i]); |
- } |
- |
- int i; |
- H264Picture::PtrVector::iterator it; |
- for (it = ref_pic_list0_.begin(), i = 0; it != ref_pic_list0_.end() && *it; |
- ++it, ++i) |
- FillVAPicture(&slice_param.RefPicList0[i], *it); |
- for (it = ref_pic_list1_.begin(), i = 0; it != ref_pic_list1_.end() && *it; |
- ++it, ++i) |
- FillVAPicture(&slice_param.RefPicList1[i], *it); |
- |
- return vaapi_wrapper_->SubmitBuffer(VASliceParameterBufferType, |
- sizeof(VASliceParameterBufferH264), |
- &slice_param); |
-} |
- |
-bool VaapiH264Decoder::SendSliceData(const uint8* ptr, size_t size) { |
- // Can't help it, blame libva... |
- void* non_const_ptr = const_cast<uint8*>(ptr); |
- return vaapi_wrapper_->SubmitBuffer(VASliceDataBufferType, size, |
- non_const_ptr); |
-} |
- |
-bool VaapiH264Decoder::PrepareRefPicLists(media::H264SliceHeader* slice_hdr) { |
- ref_pic_list0_.clear(); |
- ref_pic_list1_.clear(); |
+bool H264Decoder::ModifyReferencePicLists(media::H264SliceHeader* slice_hdr, |
+ H264Picture::Vector* ref_pic_list0, |
+ H264Picture::Vector* ref_pic_list1) { |
+ ref_pic_list0->clear(); |
+ ref_pic_list1->clear(); |
// Fill reference picture lists for B and S/SP slices. |
if (slice_hdr->IsPSlice() || slice_hdr->IsSPSlice()) { |
- ConstructReferencePicListsP(slice_hdr); |
- return ModifyReferencePicList(slice_hdr, 0); |
- } |
- |
- if (slice_hdr->IsBSlice()) { |
- ConstructReferencePicListsB(slice_hdr); |
- return ModifyReferencePicList(slice_hdr, 0) && |
- ModifyReferencePicList(slice_hdr, 1); |
+ *ref_pic_list0 = ref_pic_list_p0_; |
+ return ModifyReferencePicList(slice_hdr, 0, ref_pic_list0); |
+ } else if (slice_hdr->IsBSlice()) { |
+ *ref_pic_list0 = ref_pic_list_b0_; |
+ *ref_pic_list1 = ref_pic_list_b1_; |
+ return ModifyReferencePicList(slice_hdr, 0, ref_pic_list0) && |
+ ModifyReferencePicList(slice_hdr, 1, ref_pic_list1); |
} |
return true; |
} |
-bool VaapiH264Decoder::QueueSlice(media::H264SliceHeader* slice_hdr) { |
- DCHECK(curr_pic_.get()); |
- |
- if (!PrepareRefPicLists(slice_hdr)) |
- return false; |
- |
- if (!SendVASliceParam(slice_hdr)) |
- return false; |
- |
- if (!SendSliceData(slice_hdr->nalu_data, slice_hdr->nalu_size)) |
- return false; |
- |
- return true; |
-} |
- |
-// TODO(posciak) start using vaMapBuffer instead of vaCreateBuffer wherever |
-// possible. |
-bool VaapiH264Decoder::DecodePicture() { |
+bool H264Decoder::DecodePicture() { |
DCHECK(curr_pic_.get()); |
DVLOG(4) << "Decoding POC " << curr_pic_->pic_order_cnt; |
- DecodeSurface* dec_surface = DecodeSurfaceByPoC(curr_pic_->pic_order_cnt); |
- if (!dec_surface) { |
- DVLOG(1) << "Asked to decode an invalid POC " << curr_pic_->pic_order_cnt; |
- return false; |
- } |
- |
- if (!vaapi_wrapper_->ExecuteAndDestroyPendingBuffers( |
- dec_surface->va_surface()->id())) { |
- DVLOG(1) << "Failed decoding picture"; |
- return false; |
- } |
- |
- return true; |
+ return accelerator_->SubmitDecode(curr_pic_); |
} |
-bool VaapiH264Decoder::InitCurrPicture(media::H264SliceHeader* slice_hdr) { |
+bool H264Decoder::InitCurrPicture(media::H264SliceHeader* slice_hdr) { |
DCHECK(curr_pic_.get()); |
- memset(curr_pic_.get(), 0, sizeof(H264Picture)); |
- |
curr_pic_->idr = slice_hdr->idr_pic_flag; |
if (slice_hdr->field_pic_flag) { |
@@ -518,12 +115,6 @@ bool VaapiH264Decoder::InitCurrPicture(media::H264SliceHeader* slice_hdr) { |
if (!CalculatePicOrderCounts(slice_hdr)) |
return false; |
- // Try to get an empty surface to decode this picture to. |
- if (!AssignSurfaceToPoC(curr_input_id_, curr_pic_->pic_order_cnt)) { |
- DVLOG(1) << "Failed getting a free surface for a picture"; |
- return false; |
- } |
- |
curr_pic_->long_term_reference_flag = slice_hdr->long_term_reference_flag; |
curr_pic_->adaptive_ref_pic_marking_mode_flag = |
slice_hdr->adaptive_ref_pic_marking_mode_flag; |
@@ -532,9 +123,9 @@ bool VaapiH264Decoder::InitCurrPicture(media::H264SliceHeader* slice_hdr) { |
// process after this picture is decoded, store required data for that |
// purpose. |
if (slice_hdr->adaptive_ref_pic_marking_mode_flag) { |
- static_assert(sizeof(curr_pic_->ref_pic_marking) == |
- sizeof(slice_hdr->ref_pic_marking), |
- "ref_pic_marking array sizes do not match"); |
+ COMPILE_ASSERT(sizeof(curr_pic_->ref_pic_marking) == |
+ sizeof(slice_hdr->ref_pic_marking), |
+ ref_pic_marking_array_sizes_do_not_match); |
memcpy(curr_pic_->ref_pic_marking, slice_hdr->ref_pic_marking, |
sizeof(curr_pic_->ref_pic_marking)); |
} |
@@ -542,8 +133,7 @@ bool VaapiH264Decoder::InitCurrPicture(media::H264SliceHeader* slice_hdr) { |
return true; |
} |
-bool VaapiH264Decoder::CalculatePicOrderCounts( |
- media::H264SliceHeader* slice_hdr) { |
+bool H264Decoder::CalculatePicOrderCounts(media::H264SliceHeader* slice_hdr) { |
DCHECK_NE(curr_sps_id_, -1); |
const media::H264SPS* sps = parser_.GetSPS(curr_sps_id_); |
@@ -649,17 +239,17 @@ bool VaapiH264Decoder::CalculatePicOrderCounts( |
if (!slice_hdr->field_pic_flag) { |
curr_pic_->top_field_order_cnt = expected_pic_order_cnt + |
- slice_hdr->delta_pic_order_cnt[0]; |
+ slice_hdr->delta_pic_order_cnt0; |
curr_pic_->bottom_field_order_cnt = curr_pic_->top_field_order_cnt + |
sps->offset_for_top_to_bottom_field + |
- slice_hdr->delta_pic_order_cnt[1]; |
+ slice_hdr->delta_pic_order_cnt1; |
} else if (!slice_hdr->bottom_field_flag) { |
curr_pic_->top_field_order_cnt = expected_pic_order_cnt + |
- slice_hdr->delta_pic_order_cnt[0]; |
+ slice_hdr->delta_pic_order_cnt0; |
} else { |
curr_pic_->bottom_field_order_cnt = expected_pic_order_cnt + |
sps->offset_for_top_to_bottom_field + |
- slice_hdr->delta_pic_order_cnt[0]; |
+ slice_hdr->delta_pic_order_cnt0; |
} |
break; |
} |
@@ -718,10 +308,8 @@ bool VaapiH264Decoder::CalculatePicOrderCounts( |
return true; |
} |
-void VaapiH264Decoder::UpdatePicNums() { |
- for (H264DPB::Pictures::iterator it = dpb_.begin(); it != dpb_.end(); ++it) { |
- H264Picture* pic = *it; |
- DCHECK(pic); |
+void H264Decoder::UpdatePicNums() { |
+ for (auto& pic : dpb_) { |
if (!pic->ref) |
continue; |
@@ -741,75 +329,82 @@ void VaapiH264Decoder::UpdatePicNums() { |
} |
struct PicNumDescCompare { |
- bool operator()(const H264Picture* a, const H264Picture* b) const { |
+ bool operator()(const scoped_refptr<H264Picture>& a, |
+ const scoped_refptr<H264Picture>& b) const { |
return a->pic_num > b->pic_num; |
} |
}; |
struct LongTermPicNumAscCompare { |
- bool operator()(const H264Picture* a, const H264Picture* b) const { |
+ bool operator()(const scoped_refptr<H264Picture>& a, |
+ const scoped_refptr<H264Picture>& b) const { |
return a->long_term_pic_num < b->long_term_pic_num; |
} |
}; |
-void VaapiH264Decoder::ConstructReferencePicListsP( |
+void H264Decoder::ConstructReferencePicListsP( |
media::H264SliceHeader* slice_hdr) { |
// RefPicList0 (8.2.4.2.1) [[1] [2]], where: |
// [1] shortterm ref pics sorted by descending pic_num, |
// [2] longterm ref pics by ascending long_term_pic_num. |
- DCHECK(ref_pic_list0_.empty() && ref_pic_list1_.empty()); |
+ ref_pic_list_p0_.clear(); |
+ |
// First get the short ref pics... |
- dpb_.GetShortTermRefPicsAppending(ref_pic_list0_); |
- size_t num_short_refs = ref_pic_list0_.size(); |
+ dpb_.GetShortTermRefPicsAppending(&ref_pic_list_p0_); |
+ size_t num_short_refs = ref_pic_list_p0_.size(); |
// and sort them to get [1]. |
- std::sort(ref_pic_list0_.begin(), ref_pic_list0_.end(), PicNumDescCompare()); |
+ std::sort(ref_pic_list_p0_.begin(), ref_pic_list_p0_.end(), |
+ PicNumDescCompare()); |
// Now get long term pics and sort them by long_term_pic_num to get [2]. |
- dpb_.GetLongTermRefPicsAppending(ref_pic_list0_); |
- std::sort(ref_pic_list0_.begin() + num_short_refs, ref_pic_list0_.end(), |
+ dpb_.GetLongTermRefPicsAppending(&ref_pic_list_p0_); |
+ std::sort(ref_pic_list_p0_.begin() + num_short_refs, ref_pic_list_p0_.end(), |
LongTermPicNumAscCompare()); |
// Cut off if we have more than requested in slice header. |
- ref_pic_list0_.resize(slice_hdr->num_ref_idx_l0_active_minus1 + 1); |
+ ref_pic_list_p0_.resize(slice_hdr->num_ref_idx_l0_active_minus1 + 1); |
} |
struct POCAscCompare { |
- bool operator()(const H264Picture* a, const H264Picture* b) const { |
+ bool operator()(const scoped_refptr<H264Picture>& a, |
+ const scoped_refptr<H264Picture>& b) const { |
return a->pic_order_cnt < b->pic_order_cnt; |
} |
}; |
struct POCDescCompare { |
- bool operator()(const H264Picture* a, const H264Picture* b) const { |
+ bool operator()(const scoped_refptr<H264Picture>& a, |
+ const scoped_refptr<H264Picture>& b) const { |
return a->pic_order_cnt > b->pic_order_cnt; |
} |
}; |
-void VaapiH264Decoder::ConstructReferencePicListsB( |
+void H264Decoder::ConstructReferencePicListsB( |
media::H264SliceHeader* slice_hdr) { |
// RefPicList0 (8.2.4.2.3) [[1] [2] [3]], where: |
// [1] shortterm ref pics with POC < curr_pic's POC sorted by descending POC, |
// [2] shortterm ref pics with POC > curr_pic's POC by ascending POC, |
// [3] longterm ref pics by ascending long_term_pic_num. |
- DCHECK(ref_pic_list0_.empty() && ref_pic_list1_.empty()); |
- dpb_.GetShortTermRefPicsAppending(ref_pic_list0_); |
- size_t num_short_refs = ref_pic_list0_.size(); |
+ ref_pic_list_b0_.clear(); |
+ ref_pic_list_b1_.clear(); |
+ dpb_.GetShortTermRefPicsAppending(&ref_pic_list_b0_); |
+ size_t num_short_refs = ref_pic_list_b0_.size(); |
// First sort ascending, this will put [1] in right place and finish [2]. |
- std::sort(ref_pic_list0_.begin(), ref_pic_list0_.end(), POCAscCompare()); |
+ std::sort(ref_pic_list_b0_.begin(), ref_pic_list_b0_.end(), POCAscCompare()); |
// Find first with POC > curr_pic's POC to get first element in [2]... |
- H264Picture::PtrVector::iterator iter; |
- iter = std::upper_bound(ref_pic_list0_.begin(), ref_pic_list0_.end(), |
+ H264Picture::Vector::iterator iter; |
+ iter = std::upper_bound(ref_pic_list_b0_.begin(), ref_pic_list_b0_.end(), |
curr_pic_.get(), POCAscCompare()); |
// and sort [1] descending, thus finishing sequence [1] [2]. |
- std::sort(ref_pic_list0_.begin(), iter, POCDescCompare()); |
+ std::sort(ref_pic_list_b0_.begin(), iter, POCDescCompare()); |
// Now add [3] and sort by ascending long_term_pic_num. |
- dpb_.GetLongTermRefPicsAppending(ref_pic_list0_); |
- std::sort(ref_pic_list0_.begin() + num_short_refs, ref_pic_list0_.end(), |
+ dpb_.GetLongTermRefPicsAppending(&ref_pic_list_b0_); |
+ std::sort(ref_pic_list_b0_.begin() + num_short_refs, ref_pic_list_b0_.end(), |
LongTermPicNumAscCompare()); |
// RefPicList1 (8.2.4.2.4) [[1] [2] [3]], where: |
@@ -817,50 +412,50 @@ void VaapiH264Decoder::ConstructReferencePicListsB( |
// [2] shortterm ref pics with POC < curr_pic's POC by descending POC, |
// [3] longterm ref pics by ascending long_term_pic_num. |
- dpb_.GetShortTermRefPicsAppending(ref_pic_list1_); |
- num_short_refs = ref_pic_list1_.size(); |
+ dpb_.GetShortTermRefPicsAppending(&ref_pic_list_b1_); |
+ num_short_refs = ref_pic_list_b1_.size(); |
// First sort by descending POC. |
- std::sort(ref_pic_list1_.begin(), ref_pic_list1_.end(), POCDescCompare()); |
+ std::sort(ref_pic_list_b1_.begin(), ref_pic_list_b1_.end(), POCDescCompare()); |
// Find first with POC < curr_pic's POC to get first element in [2]... |
- iter = std::upper_bound(ref_pic_list1_.begin(), ref_pic_list1_.end(), |
+ iter = std::upper_bound(ref_pic_list_b1_.begin(), ref_pic_list_b1_.end(), |
curr_pic_.get(), POCDescCompare()); |
// and sort [1] ascending. |
- std::sort(ref_pic_list1_.begin(), iter, POCAscCompare()); |
+ std::sort(ref_pic_list_b1_.begin(), iter, POCAscCompare()); |
// Now add [3] and sort by ascending long_term_pic_num |
- dpb_.GetShortTermRefPicsAppending(ref_pic_list1_); |
- std::sort(ref_pic_list1_.begin() + num_short_refs, ref_pic_list1_.end(), |
+ dpb_.GetShortTermRefPicsAppending(&ref_pic_list_b1_); |
+ std::sort(ref_pic_list_b1_.begin() + num_short_refs, ref_pic_list_b1_.end(), |
LongTermPicNumAscCompare()); |
// If lists identical, swap first two entries in RefPicList1 (spec 8.2.4.2.3) |
- if (ref_pic_list1_.size() > 1 && |
- std::equal(ref_pic_list0_.begin(), ref_pic_list0_.end(), |
- ref_pic_list1_.begin())) |
- std::swap(ref_pic_list1_[0], ref_pic_list1_[1]); |
+ if (ref_pic_list_b1_.size() > 1 && |
+ std::equal(ref_pic_list_b0_.begin(), ref_pic_list_b0_.end(), |
+ ref_pic_list_b1_.begin())) |
+ std::swap(ref_pic_list_b1_[0], ref_pic_list_b1_[1]); |
// Per 8.2.4.2 it's possible for num_ref_idx_lX_active_minus1 to indicate |
// there should be more ref pics on list than we constructed. |
// Those superfluous ones should be treated as non-reference. |
- ref_pic_list0_.resize(slice_hdr->num_ref_idx_l0_active_minus1 + 1); |
- ref_pic_list1_.resize(slice_hdr->num_ref_idx_l1_active_minus1 + 1); |
+ ref_pic_list_b0_.resize(slice_hdr->num_ref_idx_l0_active_minus1 + 1); |
+ ref_pic_list_b1_.resize(slice_hdr->num_ref_idx_l1_active_minus1 + 1); |
} |
// See 8.2.4 |
-int VaapiH264Decoder::PicNumF(H264Picture *pic) { |
+int H264Decoder::PicNumF(const scoped_refptr<H264Picture>& pic) { |
if (!pic) |
- return -1; |
+ return -1; |
if (!pic->long_term) |
- return pic->pic_num; |
+ return pic->pic_num; |
else |
- return max_pic_num_; |
+ return max_pic_num_; |
} |
// See 8.2.4 |
-int VaapiH264Decoder::LongTermPicNumF(H264Picture *pic) { |
+int H264Decoder::LongTermPicNumF(const scoped_refptr<H264Picture>& pic) { |
if (pic->ref && pic->long_term) |
return pic->long_term_pic_num; |
else |
@@ -869,10 +464,10 @@ int VaapiH264Decoder::LongTermPicNumF(H264Picture *pic) { |
// Shift elements on the |v| starting from |from| to |to|, inclusive, |
// one position to the right and insert pic at |from|. |
-static void ShiftRightAndInsert(H264Picture::PtrVector *v, |
+static void ShiftRightAndInsert(H264Picture::Vector* v, |
int from, |
int to, |
- H264Picture* pic) { |
+ const scoped_refptr<H264Picture>& pic) { |
// Security checks, do not disable in Debug mode. |
CHECK(from <= to); |
CHECK(to <= std::numeric_limits<int>::max() - 2); |
@@ -890,10 +485,10 @@ static void ShiftRightAndInsert(H264Picture::PtrVector *v, |
(*v)[from] = pic; |
} |
-bool VaapiH264Decoder::ModifyReferencePicList(media::H264SliceHeader* slice_hdr, |
- int list) { |
+bool H264Decoder::ModifyReferencePicList(media::H264SliceHeader* slice_hdr, |
+ int list, |
+ H264Picture::Vector* ref_pic_listx) { |
int num_ref_idx_lX_active_minus1; |
- H264Picture::PtrVector* ref_pic_listx; |
media::H264ModificationOfPicNum* list_mod; |
// This can process either ref_pic_list0 or ref_pic_list1, depending on |
@@ -903,19 +498,14 @@ bool VaapiH264Decoder::ModifyReferencePicList(media::H264SliceHeader* slice_hdr, |
return true; |
list_mod = slice_hdr->ref_list_l0_modifications; |
- num_ref_idx_lX_active_minus1 = ref_pic_list0_.size() - 1; |
- |
- ref_pic_listx = &ref_pic_list0_; |
} else { |
if (!slice_hdr->ref_pic_list_modification_flag_l1) |
return true; |
list_mod = slice_hdr->ref_list_l1_modifications; |
- num_ref_idx_lX_active_minus1 = ref_pic_list1_.size() - 1; |
- |
- ref_pic_listx = &ref_pic_list1_; |
} |
+ num_ref_idx_lX_active_minus1 = ref_pic_listx->size() - 1; |
DCHECK_GE(num_ref_idx_lX_active_minus1, 0); |
// Spec 8.2.4.3: |
@@ -925,7 +515,7 @@ bool VaapiH264Decoder::ModifyReferencePicList(media::H264SliceHeader* slice_hdr, |
int pic_num_lx_no_wrap; |
int pic_num_lx; |
bool done = false; |
- H264Picture* pic; |
+ scoped_refptr<H264Picture> pic; |
for (int i = 0; i < media::H264SliceHeader::kRefListModSize && !done; ++i) { |
switch (list_mod->modification_of_pic_nums_idc) { |
case 0: |
@@ -991,8 +581,8 @@ bool VaapiH264Decoder::ModifyReferencePicList(media::H264SliceHeader* slice_hdr, |
for (int src = ref_idx_lx, dst = ref_idx_lx; |
src <= num_ref_idx_lX_active_minus1 + 1; ++src) { |
- if (LongTermPicNumF((*ref_pic_listx)[src]) |
- != static_cast<int>(list_mod->long_term_pic_num)) |
+ if (LongTermPicNumF((*ref_pic_listx)[src]) != |
+ static_cast<int>(list_mod->long_term_pic_num)) |
(*ref_pic_listx)[dst++] = (*ref_pic_listx)[src]; |
} |
break; |
@@ -1021,83 +611,48 @@ bool VaapiH264Decoder::ModifyReferencePicList(media::H264SliceHeader* slice_hdr, |
return true; |
} |
-bool VaapiH264Decoder::OutputPic(H264Picture* pic) { |
+void H264Decoder::OutputPic(scoped_refptr<H264Picture> pic) { |
DCHECK(!pic->outputted); |
pic->outputted = true; |
last_output_poc_ = pic->pic_order_cnt; |
- DecodeSurface* dec_surface = DecodeSurfaceByPoC(pic->pic_order_cnt); |
- if (!dec_surface) |
- return false; |
- |
- DCHECK_GE(dec_surface->input_id(), 0); |
- DVLOG(4) << "Posting output task for POC: " << pic->pic_order_cnt |
- << " input_id: " << dec_surface->input_id(); |
- output_pic_cb_.Run(dec_surface->input_id(), dec_surface->va_surface()); |
- |
- return true; |
+ DVLOG(4) << "Posting output task for POC: " << pic->pic_order_cnt; |
+ accelerator_->OutputPicture(pic); |
} |
-void VaapiH264Decoder::ClearDPB() { |
+void H264Decoder::ClearDPB() { |
// Clear DPB contents, marking the pictures as unused first. |
- for (H264DPB::Pictures::iterator it = dpb_.begin(); it != dpb_.end(); ++it) |
- UnassignSurfaceFromPoC((*it)->pic_order_cnt); |
- |
dpb_.Clear(); |
last_output_poc_ = std::numeric_limits<int>::min(); |
} |
-bool VaapiH264Decoder::OutputAllRemainingPics() { |
+bool H264Decoder::OutputAllRemainingPics() { |
// Output all pictures that are waiting to be outputted. |
FinishPrevFrameIfPresent(); |
- H264Picture::PtrVector to_output; |
- dpb_.GetNotOutputtedPicsAppending(to_output); |
+ H264Picture::Vector to_output; |
+ dpb_.GetNotOutputtedPicsAppending(&to_output); |
// Sort them by ascending POC to output in order. |
std::sort(to_output.begin(), to_output.end(), POCAscCompare()); |
- H264Picture::PtrVector::iterator it; |
- for (it = to_output.begin(); it != to_output.end(); ++it) { |
- if (!OutputPic(*it)) { |
- DVLOG(1) << "Failed to output pic POC: " << (*it)->pic_order_cnt; |
- return false; |
- } |
- } |
+ for (auto& pic : to_output) |
+ OutputPic(pic); |
return true; |
} |
-bool VaapiH264Decoder::Flush() { |
+bool H264Decoder::Flush() { |
DVLOG(2) << "Decoder flush"; |
if (!OutputAllRemainingPics()) |
return false; |
ClearDPB(); |
- |
- DCHECK(decode_surfaces_in_use_.empty()); |
+ DVLOG(2) << "Decoder flush finished"; |
return true; |
} |
-bool VaapiH264Decoder::StartNewFrame(media::H264SliceHeader* slice_hdr) { |
+bool H264Decoder::StartNewFrame(media::H264SliceHeader* slice_hdr) { |
// TODO posciak: add handling of max_num_ref_frames per spec. |
- |
- // If the new frame is an IDR, output what's left to output and clear DPB |
- if (slice_hdr->idr_pic_flag) { |
- // (unless we are explicitly instructed not to do so). |
- if (!slice_hdr->no_output_of_prior_pics_flag) { |
- // Output DPB contents. |
- if (!Flush()) |
- return false; |
- } |
- dpb_.Clear(); |
- last_output_poc_ = std::numeric_limits<int>::min(); |
- } |
- |
- // curr_pic_ should have either been added to DPB or discarded when finishing |
- // the last frame. DPB is responsible for releasing that memory once it's |
- // not needed anymore. |
- DCHECK(!curr_pic_.get()); |
- curr_pic_.reset(new H264Picture); |
CHECK(curr_pic_.get()); |
if (!InitCurrPicture(slice_hdr)) |
@@ -1106,27 +661,29 @@ bool VaapiH264Decoder::StartNewFrame(media::H264SliceHeader* slice_hdr) { |
DCHECK_GT(max_frame_num_, 0); |
UpdatePicNums(); |
+ DCHECK(slice_hdr); |
+ PrepareRefPicLists(slice_hdr); |
- // Send parameter buffers before each new picture, before the first slice. |
- if (!SendPPS()) |
- return false; |
- |
- if (!SendIQMatrix()) |
- return false; |
+ const media::H264PPS* pps = parser_.GetPPS(curr_pps_id_); |
+ DCHECK(pps); |
+ const media::H264SPS* sps = parser_.GetSPS(pps->seq_parameter_set_id); |
+ DCHECK(sps); |
- if (!QueueSlice(slice_hdr)) |
+ if (!accelerator_->SubmitFrameMetadata(sps, pps, dpb_, ref_pic_list_p0_, |
+ ref_pic_list_b0_, ref_pic_list_b1_, |
+ curr_pic_.get())) |
return false; |
return true; |
} |
-bool VaapiH264Decoder::HandleMemoryManagementOps() { |
+bool H264Decoder::HandleMemoryManagementOps() { |
// 8.2.5.4 |
for (unsigned int i = 0; i < arraysize(curr_pic_->ref_pic_marking); ++i) { |
// Code below does not support interlaced stream (per-field pictures). |
media::H264DecRefPicMarking* ref_pic_marking = |
&curr_pic_->ref_pic_marking[i]; |
- H264Picture* to_mark; |
+ scoped_refptr<H264Picture> to_mark; |
int pic_num_x; |
switch (ref_pic_marking->memory_mgmnt_control_operation) { |
@@ -1138,7 +695,7 @@ bool VaapiH264Decoder::HandleMemoryManagementOps() { |
// Mark a short term reference picture as unused so it can be removed |
// if outputted. |
pic_num_x = curr_pic_->pic_num - |
- (ref_pic_marking->difference_of_pic_nums_minus1 + 1); |
+ (ref_pic_marking->difference_of_pic_nums_minus1 + 1); |
to_mark = dpb_.GetShortRefPicByPicNum(pic_num_x); |
if (to_mark) { |
to_mark->ref = false; |
@@ -1164,7 +721,7 @@ bool VaapiH264Decoder::HandleMemoryManagementOps() { |
case 3: |
// Mark a short term reference picture as long term reference. |
pic_num_x = curr_pic_->pic_num - |
- (ref_pic_marking->difference_of_pic_nums_minus1 + 1); |
+ (ref_pic_marking->difference_of_pic_nums_minus1 + 1); |
to_mark = dpb_.GetShortRefPicByPicNum(pic_num_x); |
if (to_mark) { |
DCHECK(to_mark->ref && !to_mark->long_term); |
@@ -1178,12 +735,12 @@ bool VaapiH264Decoder::HandleMemoryManagementOps() { |
case 4: { |
// Unmark all reference pictures with long_term_frame_idx over new max. |
- max_long_term_frame_idx_ |
- = ref_pic_marking->max_long_term_frame_idx_plus1 - 1; |
- H264Picture::PtrVector long_terms; |
- dpb_.GetLongTermRefPicsAppending(long_terms); |
+ max_long_term_frame_idx_ = |
+ ref_pic_marking->max_long_term_frame_idx_plus1 - 1; |
+ H264Picture::Vector long_terms; |
+ dpb_.GetLongTermRefPicsAppending(&long_terms); |
for (size_t i = 0; i < long_terms.size(); ++i) { |
- H264Picture* pic = long_terms[i]; |
+ scoped_refptr<H264Picture>& pic = long_terms[i]; |
DCHECK(pic->ref && pic->long_term); |
// Ok to cast, max_long_term_frame_idx is much smaller than 16bit. |
if (pic->long_term_frame_idx > |
@@ -1203,10 +760,10 @@ bool VaapiH264Decoder::HandleMemoryManagementOps() { |
case 6: { |
// Replace long term reference pictures with current picture. |
// First unmark if any existing with this long_term_frame_idx... |
- H264Picture::PtrVector long_terms; |
- dpb_.GetLongTermRefPicsAppending(long_terms); |
+ H264Picture::Vector long_terms; |
+ dpb_.GetLongTermRefPicsAppending(&long_terms); |
for (size_t i = 0; i < long_terms.size(); ++i) { |
- H264Picture* pic = long_terms[i]; |
+ scoped_refptr<H264Picture>& pic = long_terms[i]; |
DCHECK(pic->ref && pic->long_term); |
// Ok to cast, long_term_frame_idx is much smaller than 16bit. |
if (pic->long_term_frame_idx == |
@@ -1235,7 +792,7 @@ bool VaapiH264Decoder::HandleMemoryManagementOps() { |
// procedure to remove the oldest one. |
// It also performs marking and unmarking pictures as reference. |
// See spac 8.2.5.1. |
-void VaapiH264Decoder::ReferencePictureMarking() { |
+void H264Decoder::ReferencePictureMarking() { |
if (curr_pic_->idr) { |
// If current picture is an IDR, all reference pictures are unmarked. |
dpb_.MarkAllUnusedForRef(); |
@@ -1254,9 +811,9 @@ void VaapiH264Decoder::ReferencePictureMarking() { |
// ensure DPB doesn't overflow, discard oldest picture. |
// See spec 8.2.5.3. |
if (curr_pic_->field == H264Picture::FIELD_NONE) { |
- DCHECK_LE(dpb_.CountRefPics(), |
- std::max<int>(parser_.GetSPS(curr_sps_id_)->max_num_ref_frames, |
- 1)); |
+ DCHECK_LE( |
+ dpb_.CountRefPics(), |
+ std::max<int>(parser_.GetSPS(curr_sps_id_)->max_num_ref_frames, 1)); |
if (dpb_.CountRefPics() == |
std::max<int>(parser_.GetSPS(curr_sps_id_)->max_num_ref_frames, |
1)) { |
@@ -1264,7 +821,8 @@ void VaapiH264Decoder::ReferencePictureMarking() { |
// need to remove one of the short term ones. |
// Find smallest frame_num_wrap short reference picture and mark |
// it as unused. |
- H264Picture* to_unmark = dpb_.GetLowestFrameNumWrapShortRefPic(); |
+ scoped_refptr<H264Picture> to_unmark = |
+ dpb_.GetLowestFrameNumWrapShortRefPic(); |
if (to_unmark == NULL) { |
DVLOG(1) << "Couldn't find a short ref picture to unmark"; |
return; |
@@ -1274,7 +832,6 @@ void VaapiH264Decoder::ReferencePictureMarking() { |
} else { |
// Shouldn't get here. |
DVLOG(1) << "Interlaced video not supported."; |
- report_error_to_uma_cb_.Run(INTERLACED_STREAM); |
} |
} else { |
// Stream has instructions how to discard pictures from DPB and how |
@@ -1285,13 +842,12 @@ void VaapiH264Decoder::ReferencePictureMarking() { |
} else { |
// Shouldn't get here. |
DVLOG(1) << "Interlaced video not supported."; |
- report_error_to_uma_cb_.Run(INTERLACED_STREAM); |
} |
} |
} |
} |
-bool VaapiH264Decoder::FinishPicture() { |
+bool H264Decoder::FinishPicture() { |
DCHECK(curr_pic_.get()); |
// Finish processing previous picture. |
@@ -1310,27 +866,23 @@ bool VaapiH264Decoder::FinishPicture() { |
// Remove unused (for reference or later output) pictures from DPB, marking |
// them as such. |
- for (H264DPB::Pictures::iterator it = dpb_.begin(); it != dpb_.end(); ++it) { |
- if ((*it)->outputted && !(*it)->ref) |
- UnassignSurfaceFromPoC((*it)->pic_order_cnt); |
- } |
dpb_.DeleteUnused(); |
DVLOG(4) << "Finishing picture, entries in DPB: " << dpb_.size(); |
// Whatever happens below, curr_pic_ will stop managing the pointer to the |
- // picture after this function returns. The ownership will either be |
- // transferred to DPB, if the image is still needed (for output and/or |
- // reference), or the memory will be released if we manage to output it here |
- // without having to store it for future reference. |
- scoped_ptr<H264Picture> pic(curr_pic_.release()); |
+ // picture after this. The ownership will either be transferred to DPB, if |
+ // the image is still needed (for output and/or reference), or the memory |
+ // will be released if we manage to output it here without having to store |
+ // it for future reference. |
+ scoped_refptr<H264Picture> pic = curr_pic_; |
+ curr_pic_ = nullptr; |
// Get all pictures that haven't been outputted yet. |
- H264Picture::PtrVector not_outputted; |
- // TODO(posciak): pass as pointer, not reference (violates coding style). |
- dpb_.GetNotOutputtedPicsAppending(not_outputted); |
+ H264Picture::Vector not_outputted; |
+ dpb_.GetNotOutputtedPicsAppending(¬_outputted); |
// Include the one we've just decoded. |
- not_outputted.push_back(pic.get()); |
+ not_outputted.push_back(pic); |
// Sort in output order. |
std::sort(not_outputted.begin(), not_outputted.end(), POCAscCompare()); |
@@ -1340,21 +892,18 @@ bool VaapiH264Decoder::FinishPicture() { |
// in DPB afterwards would at least be equal to max_num_reorder_frames. |
// If the outputted picture is not a reference picture, it doesn't have |
// to remain in the DPB and can be removed. |
- H264Picture::PtrVector::iterator output_candidate = not_outputted.begin(); |
+ H264Picture::Vector::iterator output_candidate = not_outputted.begin(); |
size_t num_remaining = not_outputted.size(); |
while (num_remaining > max_num_reorder_frames_) { |
int poc = (*output_candidate)->pic_order_cnt; |
DCHECK_GE(poc, last_output_poc_); |
- if (!OutputPic(*output_candidate)) |
- return false; |
+ OutputPic(*output_candidate); |
if (!(*output_candidate)->ref) { |
// Current picture hasn't been inserted into DPB yet, so don't remove it |
// if we managed to output it immediately. |
- if (*output_candidate != pic) |
+ if ((*output_candidate)->pic_order_cnt != pic->pic_order_cnt) |
dpb_.DeleteByPOC(poc); |
- // Mark as unused. |
- UnassignSurfaceFromPoC(poc); |
} |
++output_candidate; |
@@ -1371,7 +920,7 @@ bool VaapiH264Decoder::FinishPicture() { |
return false; |
} |
- dpb_.StorePic(pic.release()); |
+ dpb_.StorePic(pic); |
} |
return true; |
@@ -1402,7 +951,7 @@ static int LevelToMaxDpbMbs(int level) { |
} |
} |
-bool VaapiH264Decoder::UpdateMaxNumReorderFrames(const media::H264SPS* sps) { |
+bool H264Decoder::UpdateMaxNumReorderFrames(const media::H264SPS* sps) { |
if (sps->vui_parameters_present_flag && sps->bitstream_restriction_flag) { |
max_num_reorder_frames_ = |
base::checked_cast<size_t>(sps->max_num_reorder_frames); |
@@ -1439,7 +988,7 @@ bool VaapiH264Decoder::UpdateMaxNumReorderFrames(const media::H264SPS* sps) { |
return true; |
} |
-bool VaapiH264Decoder::ProcessSPS(int sps_id, bool* need_new_buffers) { |
+bool H264Decoder::ProcessSPS(int sps_id, bool* need_new_buffers) { |
const media::H264SPS* sps = parser_.GetSPS(sps_id); |
DCHECK(sps); |
DVLOG(4) << "Processing SPS"; |
@@ -1448,13 +997,11 @@ bool VaapiH264Decoder::ProcessSPS(int sps_id, bool* need_new_buffers) { |
if (sps->frame_mbs_only_flag == 0) { |
DVLOG(1) << "frame_mbs_only_flag != 1 not supported"; |
- report_error_to_uma_cb_.Run(FRAME_MBS_ONLY_FLAG_NOT_ONE); |
return false; |
} |
if (sps->gaps_in_frame_num_value_allowed_flag) { |
DVLOG(1) << "Gaps in frame numbers not supported"; |
- report_error_to_uma_cb_.Run(GAPS_IN_FRAME_NUM); |
return false; |
} |
@@ -1464,7 +1011,7 @@ bool VaapiH264Decoder::ProcessSPS(int sps_id, bool* need_new_buffers) { |
// (spec 7.4.2.1.1, 7.4.3). |
int width_mb = sps->pic_width_in_mbs_minus1 + 1; |
int height_mb = (2 - sps->frame_mbs_only_flag) * |
- (sps->pic_height_in_map_units_minus1 + 1); |
+ (sps->pic_height_in_map_units_minus1 + 1); |
gfx::Size new_pic_size(16 * width_mb, 16 * height_mb); |
if (new_pic_size.IsEmpty()) { |
@@ -1507,7 +1054,7 @@ bool VaapiH264Decoder::ProcessSPS(int sps_id, bool* need_new_buffers) { |
return true; |
} |
-bool VaapiH264Decoder::ProcessPPS(int pps_id) { |
+bool H264Decoder::ProcessPPS(int pps_id) { |
const media::H264PPS* pps = parser_.GetPPS(pps_id); |
DCHECK(pps); |
@@ -1516,7 +1063,7 @@ bool VaapiH264Decoder::ProcessPPS(int pps_id) { |
return true; |
} |
-bool VaapiH264Decoder::FinishPrevFrameIfPresent() { |
+bool H264Decoder::FinishPrevFrameIfPresent() { |
// If we already have a frame waiting to be decoded, decode it and finish. |
if (curr_pic_ != NULL) { |
if (!DecodePicture()) |
@@ -1527,13 +1074,12 @@ bool VaapiH264Decoder::FinishPrevFrameIfPresent() { |
return true; |
} |
-bool VaapiH264Decoder::ProcessSlice(media::H264SliceHeader* slice_hdr) { |
+bool H264Decoder::PreprocessSlice(media::H264SliceHeader* slice_hdr) { |
prev_frame_num_ = frame_num_; |
frame_num_ = slice_hdr->frame_num; |
if (prev_frame_num_ > 0 && prev_frame_num_ < frame_num_ - 1) { |
DVLOG(1) << "Gap in frame_num!"; |
- report_error_to_uma_cb_.Run(GAPS_IN_FRAME_NUM); |
return false; |
} |
@@ -1544,66 +1090,81 @@ bool VaapiH264Decoder::ProcessSlice(media::H264SliceHeader* slice_hdr) { |
// TODO posciak: switch to new picture detection per 7.4.1.2.4. |
if (curr_pic_ != NULL && slice_hdr->first_mb_in_slice != 0) { |
- // This is just some more slice data of the current picture, so |
- // just queue it and return. |
- QueueSlice(slice_hdr); |
+ // More slice data of the current picture. |
return true; |
} else { |
// A new frame, so first finish the previous one before processing it... |
if (!FinishPrevFrameIfPresent()) |
return false; |
+ } |
- // and then start a new one. |
- return StartNewFrame(slice_hdr); |
+ // If the new frame is an IDR, output what's left to output and clear DPB |
+ if (slice_hdr->idr_pic_flag) { |
+ // (unless we are explicitly instructed not to do so). |
+ if (!slice_hdr->no_output_of_prior_pics_flag) { |
+ // Output DPB contents. |
+ if (!Flush()) |
+ return false; |
+ } |
+ dpb_.Clear(); |
+ last_output_poc_ = std::numeric_limits<int>::min(); |
} |
+ |
+ return true; |
} |
-#define SET_ERROR_AND_RETURN() \ |
- do { \ |
- DVLOG(1) << "Error during decode"; \ |
- state_ = kError; \ |
- return VaapiH264Decoder::kDecodeError; \ |
+bool H264Decoder::ProcessSlice(media::H264SliceHeader* slice_hdr) { |
+ DCHECK(curr_pic_.get()); |
+ H264Picture::Vector ref_pic_list0, ref_pic_list1; |
+ |
+ if (!ModifyReferencePicLists(slice_hdr, &ref_pic_list0, &ref_pic_list1)) |
+ return false; |
+ |
+ const media::H264PPS* pps = parser_.GetPPS(slice_hdr->pic_parameter_set_id); |
+ DCHECK(pps); |
+ |
+ if (!accelerator_->SubmitSlice(pps, slice_hdr, ref_pic_list0, ref_pic_list1, |
+ curr_pic_.get(), slice_hdr->nalu_data, |
+ slice_hdr->nalu_size)) |
+ return false; |
+ |
+ curr_slice_hdr_.reset(); |
+ return true; |
+} |
+ |
+#define SET_ERROR_AND_RETURN() \ |
+ do { \ |
+ DVLOG(1) << "Error during decode"; \ |
+ state_ = kError; \ |
+ return H264Decoder::kDecodeError; \ |
} while (0) |
-void VaapiH264Decoder::SetStream(const uint8* ptr, |
- size_t size, |
- int32 input_id) { |
+void H264Decoder::SetStream(const uint8_t* ptr, size_t size) { |
DCHECK(ptr); |
DCHECK(size); |
- // Got new input stream data from the client. |
- DVLOG(4) << "New input stream id: " << input_id << " at: " << (void*) ptr |
- << " size: " << size; |
+ DVLOG(4) << "New input stream at: " << (void*)ptr << " size: " << size; |
parser_.SetStream(ptr, size); |
- curr_input_id_ = input_id; |
} |
-VaapiH264Decoder::DecResult VaapiH264Decoder::Decode() { |
- media::H264Parser::Result par_res; |
- media::H264NALU nalu; |
+H264Decoder::DecodeResult H264Decoder::Decode() { |
DCHECK_NE(state_, kError); |
while (1) { |
- // If we've already decoded some of the stream (after reset, i.e. we are |
- // not in kNeedStreamMetadata state), we may be able to go back into |
- // decoding state not only starting at/resuming from an SPS, but also from |
- // other resume points, such as IDRs. In the latter case we need an output |
- // surface, because we will end up decoding that IDR in the process. |
- // Otherwise we just look for an SPS and don't produce any output frames. |
- if (state_ != kNeedStreamMetadata && available_va_surfaces_.empty()) { |
- DVLOG(4) << "No output surfaces available"; |
- return kRanOutOfSurfaces; |
+ media::H264Parser::Result par_res; |
+ |
+ if (!curr_nalu_) { |
+ curr_nalu_.reset(new media::H264NALU()); |
+ par_res = parser_.AdvanceToNextNALU(curr_nalu_.get()); |
+ if (par_res == media::H264Parser::kEOStream) |
+ return kRanOutOfStreamData; |
+ else if (par_res != media::H264Parser::kOk) |
+ SET_ERROR_AND_RETURN(); |
} |
- par_res = parser_.AdvanceToNextNALU(&nalu); |
- if (par_res == media::H264Parser::kEOStream) |
- return kRanOutOfStreamData; |
- else if (par_res != media::H264Parser::kOk) |
- SET_ERROR_AND_RETURN(); |
+ DVLOG(4) << "NALU found: " << static_cast<int>(curr_nalu_->nal_unit_type); |
- DVLOG(4) << "NALU found: " << static_cast<int>(nalu.nal_unit_type); |
- |
- switch (nalu.nal_unit_type) { |
+ switch (curr_nalu_->nal_unit_type) { |
case media::H264NALU::kNonIDRSlice: |
// We can't resume from a non-IDR slice. |
if (state_ != kDecoding) |
@@ -1619,13 +1180,29 @@ VaapiH264Decoder::DecResult VaapiH264Decoder::Decode() { |
} |
// If after reset, we should be able to recover from an IDR. |
- media::H264SliceHeader slice_hdr; |
+ if (!curr_slice_hdr_) { |
+ curr_slice_hdr_.reset(new media::H264SliceHeader()); |
+ par_res = |
+ parser_.ParseSliceHeader(*curr_nalu_, curr_slice_hdr_.get()); |
+ if (par_res != media::H264Parser::kOk) |
+ SET_ERROR_AND_RETURN(); |
+ |
+ if (!PreprocessSlice(curr_slice_hdr_.get())) |
+ SET_ERROR_AND_RETURN(); |
+ } |
- par_res = parser_.ParseSliceHeader(nalu, &slice_hdr); |
- if (par_res != media::H264Parser::kOk) |
- SET_ERROR_AND_RETURN(); |
+ if (!curr_pic_) { |
+ // New picture/finished previous one, try to start a new one |
+ // or tell the client we need more surfaces. |
+ curr_pic_ = accelerator_->CreateH264Picture(); |
+ if (!curr_pic_) |
+ return kRanOutOfSurfaces; |
+ |
+ if (!StartNewFrame(curr_slice_hdr_.get())) |
+ SET_ERROR_AND_RETURN(); |
+ } |
- if (!ProcessSlice(&slice_hdr)) |
+ if (!ProcessSlice(curr_slice_hdr_.get())) |
SET_ERROR_AND_RETURN(); |
state_ = kDecoding; |
@@ -1652,7 +1229,12 @@ VaapiH264Decoder::DecResult VaapiH264Decoder::Decode() { |
if (!Flush()) |
return kDecodeError; |
- available_va_surfaces_.clear(); |
+ curr_pic_ = nullptr; |
+ curr_nalu_ = nullptr; |
+ ref_pic_list_p0_.clear(); |
+ ref_pic_list_b0_.clear(); |
+ ref_pic_list_b1_.clear(); |
+ |
return kAllocateNewSurfaces; |
} |
break; |
@@ -1676,24 +1258,17 @@ VaapiH264Decoder::DecResult VaapiH264Decoder::Decode() { |
break; |
} |
- case media::H264NALU::kAUD: |
- case media::H264NALU::kEOSeq: |
- case media::H264NALU::kEOStream: |
- if (state_ != kDecoding) |
- break; |
- if (!FinishPrevFrameIfPresent()) |
- SET_ERROR_AND_RETURN(); |
- |
- break; |
- |
default: |
- DVLOG(4) << "Skipping NALU type: " << nalu.nal_unit_type; |
+ DVLOG(4) << "Skipping NALU type: " << curr_nalu_->nal_unit_type; |
break; |
} |
+ |
+ DVLOG(4) << "Dropping nalu"; |
+ curr_nalu_.reset(); |
} |
} |
-size_t VaapiH264Decoder::GetRequiredNumOfPictures() { |
+size_t H264Decoder::GetRequiredNumOfPictures() const { |
return dpb_.max_num_pics() + kPicsInPipeline; |
} |