OLD | NEW |
1 // Copyright (c) 2012 The Chromium Authors. All rights reserved. | 1 // Copyright (c) 2012 The Chromium Authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 #include "media/filters/ffmpeg_video_decoder.h" | 5 #include "media/filters/ffmpeg_video_decoder.h" |
6 | 6 |
7 #include <algorithm> | 7 #include <algorithm> |
8 #include <string> | 8 #include <string> |
9 | 9 |
10 #include "base/bind.h" | 10 #include "base/bind.h" |
(...skipping 122 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
133 av_buffer_create(frame->data[0], | 133 av_buffer_create(frame->data[0], |
134 VideoFrame::AllocationSize(format, coded_size), | 134 VideoFrame::AllocationSize(format, coded_size), |
135 ReleaseVideoBufferImpl, | 135 ReleaseVideoBufferImpl, |
136 opaque, | 136 opaque, |
137 0); | 137 0); |
138 return 0; | 138 return 0; |
139 } | 139 } |
140 | 140 |
141 void FFmpegVideoDecoder::Initialize(const VideoDecoderConfig& config, | 141 void FFmpegVideoDecoder::Initialize(const VideoDecoderConfig& config, |
142 bool low_delay, | 142 bool low_delay, |
143 const PipelineStatusCB& status_cb) { | 143 const PipelineStatusCB& status_cb, |
| 144 const OutputCB& output_cb) { |
144 DCHECK(task_runner_->BelongsToCurrentThread()); | 145 DCHECK(task_runner_->BelongsToCurrentThread()); |
145 DCHECK(decode_cb_.is_null()); | |
146 DCHECK(!config.is_encrypted()); | 146 DCHECK(!config.is_encrypted()); |
| 147 DCHECK(!output_cb.is_null()); |
147 | 148 |
148 FFmpegGlue::InitializeFFmpeg(); | 149 FFmpegGlue::InitializeFFmpeg(); |
149 | 150 |
150 config_ = config; | 151 config_ = config; |
151 PipelineStatusCB initialize_cb = BindToCurrentLoop(status_cb); | 152 PipelineStatusCB initialize_cb = BindToCurrentLoop(status_cb); |
152 | 153 |
153 if (!config.IsValidConfig() || !ConfigureDecoder(low_delay)) { | 154 if (!config.IsValidConfig() || !ConfigureDecoder(low_delay)) { |
154 initialize_cb.Run(DECODER_ERROR_NOT_SUPPORTED); | 155 initialize_cb.Run(DECODER_ERROR_NOT_SUPPORTED); |
155 return; | 156 return; |
156 } | 157 } |
157 | 158 |
| 159 output_cb_ = BindToCurrentLoop(output_cb); |
| 160 |
158 // Success! | 161 // Success! |
159 state_ = kNormal; | 162 state_ = kNormal; |
160 initialize_cb.Run(PIPELINE_OK); | 163 initialize_cb.Run(PIPELINE_OK); |
161 } | 164 } |
162 | 165 |
163 void FFmpegVideoDecoder::Decode(const scoped_refptr<DecoderBuffer>& buffer, | 166 void FFmpegVideoDecoder::Decode(const scoped_refptr<DecoderBuffer>& buffer, |
164 const DecodeCB& decode_cb) { | 167 const DecodeCB& decode_cb) { |
165 DCHECK(task_runner_->BelongsToCurrentThread()); | 168 DCHECK(task_runner_->BelongsToCurrentThread()); |
| 169 DCHECK(buffer); |
166 DCHECK(!decode_cb.is_null()); | 170 DCHECK(!decode_cb.is_null()); |
167 CHECK_NE(state_, kUninitialized); | 171 CHECK_NE(state_, kUninitialized); |
168 CHECK(decode_cb_.is_null()) << "Overlapping decodes are not supported."; | 172 |
169 decode_cb_ = BindToCurrentLoop(decode_cb); | 173 DecodeCB decode_cb_bound = BindToCurrentLoop(decode_cb); |
170 | 174 |
171 if (state_ == kError) { | 175 if (state_ == kError) { |
172 base::ResetAndReturn(&decode_cb_).Run(kDecodeError, NULL); | 176 decode_cb_bound.Run(kDecodeError); |
173 return; | 177 return; |
174 } | 178 } |
175 | 179 |
176 // Return empty frames if decoding has finished. | |
177 if (state_ == kDecodeFinished) { | 180 if (state_ == kDecodeFinished) { |
178 base::ResetAndReturn(&decode_cb_).Run(kOk, VideoFrame::CreateEOSFrame()); | 181 output_cb_.Run(VideoFrame::CreateEOSFrame()); |
| 182 decode_cb_bound.Run(kOk); |
179 return; | 183 return; |
180 } | 184 } |
181 | 185 |
182 DecodeBuffer(buffer); | 186 DCHECK_EQ(state_, kNormal); |
183 } | |
184 | |
185 void FFmpegVideoDecoder::Reset(const base::Closure& closure) { | |
186 DCHECK(task_runner_->BelongsToCurrentThread()); | |
187 DCHECK(decode_cb_.is_null()); | |
188 | |
189 avcodec_flush_buffers(codec_context_.get()); | |
190 state_ = kNormal; | |
191 task_runner_->PostTask(FROM_HERE, closure); | |
192 } | |
193 | |
194 void FFmpegVideoDecoder::Stop() { | |
195 DCHECK(task_runner_->BelongsToCurrentThread()); | |
196 | |
197 if (state_ == kUninitialized) | |
198 return; | |
199 | |
200 ReleaseFFmpegResources(); | |
201 state_ = kUninitialized; | |
202 } | |
203 | |
204 FFmpegVideoDecoder::~FFmpegVideoDecoder() { | |
205 DCHECK_EQ(kUninitialized, state_); | |
206 DCHECK(!codec_context_); | |
207 DCHECK(!av_frame_); | |
208 } | |
209 | |
210 void FFmpegVideoDecoder::DecodeBuffer( | |
211 const scoped_refptr<DecoderBuffer>& buffer) { | |
212 DCHECK(task_runner_->BelongsToCurrentThread()); | |
213 DCHECK_NE(state_, kUninitialized); | |
214 DCHECK_NE(state_, kDecodeFinished); | |
215 DCHECK_NE(state_, kError); | |
216 DCHECK(!decode_cb_.is_null()); | |
217 DCHECK(buffer); | |
218 | 187 |
219 // During decode, because reads are issued asynchronously, it is possible to | 188 // During decode, because reads are issued asynchronously, it is possible to |
220 // receive multiple end of stream buffers since each decode is acked. When the | 189 // receive multiple end of stream buffers since each decode is acked. When the |
221 // first end of stream buffer is read, FFmpeg may still have frames queued | 190 // first end of stream buffer is read, FFmpeg may still have frames queued |
222 // up in the decoder so we need to go through the decode loop until it stops | 191 // up in the decoder so we need to go through the decode loop until it stops |
223 // giving sensible data. After that, the decoder should output empty | 192 // giving sensible data. After that, the decoder should output empty |
224 // frames. There are three states the decoder can be in: | 193 // frames. There are three states the decoder can be in: |
225 // | 194 // |
226 // kNormal: This is the starting state. Buffers are decoded. Decode errors | 195 // kNormal: This is the starting state. Buffers are decoded. Decode errors |
227 // are discarded. | 196 // are discarded. |
228 // kFlushCodec: There isn't any more input data. Call avcodec_decode_video2 | |
229 // until no more data is returned to flush out remaining | |
230 // frames. The input buffer is ignored at this point. | |
231 // kDecodeFinished: All calls return empty frames. | 197 // kDecodeFinished: All calls return empty frames. |
232 // kError: Unexpected error happened. | 198 // kError: Unexpected error happened. |
233 // | 199 // |
234 // These are the possible state transitions. | 200 // These are the possible state transitions. |
235 // | 201 // |
236 // kNormal -> kFlushCodec: | 202 // kNormal -> kDecodeFinished: |
237 // When buffer->end_of_stream() is first true. | 203 // When EOS buffer is received and the codec has been flushed. |
238 // kNormal -> kError: | 204 // kNormal -> kError: |
239 // A decoding error occurs and decoding needs to stop. | 205 // A decoding error occurs and decoding needs to stop. |
240 // kFlushCodec -> kDecodeFinished: | |
241 // When avcodec_decode_video2() returns 0 data. | |
242 // kFlushCodec -> kError: | |
243 // When avcodec_decode_video2() errors out. | |
244 // (any state) -> kNormal: | 206 // (any state) -> kNormal: |
245 // Any time Reset() is called. | 207 // Any time Reset() is called. |
246 | 208 |
247 // Transition to kFlushCodec on the first end of stream buffer. | 209 bool has_produced_frame; |
248 if (state_ == kNormal && buffer->end_of_stream()) { | 210 do { |
249 state_ = kFlushCodec; | 211 has_produced_frame = false; |
| 212 if (!FFmpegDecode(buffer, &has_produced_frame)) { |
| 213 state_ = kError; |
| 214 decode_cb_bound.Run(kDecodeError); |
| 215 return; |
| 216 } |
| 217 // Repeat to flush the decoder after receiving EOS buffer. |
| 218 } while (buffer->end_of_stream() && has_produced_frame); |
| 219 |
| 220 if (buffer->end_of_stream()) { |
| 221 output_cb_.Run(VideoFrame::CreateEOSFrame()); |
| 222 state_ = kDecodeFinished; |
250 } | 223 } |
251 | 224 |
252 scoped_refptr<VideoFrame> video_frame; | 225 decode_cb_bound.Run(kOk); |
253 if (!FFmpegDecode(buffer, &video_frame)) { | 226 } |
254 state_ = kError; | 227 |
255 base::ResetAndReturn(&decode_cb_).Run(kDecodeError, NULL); | 228 void FFmpegVideoDecoder::Reset(const base::Closure& closure) { |
| 229 DCHECK(task_runner_->BelongsToCurrentThread()); |
| 230 |
| 231 avcodec_flush_buffers(codec_context_.get()); |
| 232 state_ = kNormal; |
| 233 task_runner_->PostTask(FROM_HERE, closure); |
| 234 } |
| 235 |
| 236 void FFmpegVideoDecoder::Stop() { |
| 237 DCHECK(task_runner_->BelongsToCurrentThread()); |
| 238 |
| 239 if (state_ == kUninitialized) |
256 return; | 240 return; |
257 } | |
258 | 241 |
259 if (!video_frame.get()) { | 242 ReleaseFFmpegResources(); |
260 if (state_ == kFlushCodec) { | 243 state_ = kUninitialized; |
261 DCHECK(buffer->end_of_stream()); | 244 } |
262 state_ = kDecodeFinished; | |
263 base::ResetAndReturn(&decode_cb_) | |
264 .Run(kOk, VideoFrame::CreateEOSFrame()); | |
265 return; | |
266 } | |
267 | 245 |
268 base::ResetAndReturn(&decode_cb_).Run(kNotEnoughData, NULL); | 246 FFmpegVideoDecoder::~FFmpegVideoDecoder() { |
269 return; | 247 DCHECK_EQ(kUninitialized, state_); |
270 } | 248 DCHECK(!codec_context_); |
271 | 249 DCHECK(!av_frame_); |
272 base::ResetAndReturn(&decode_cb_).Run(kOk, video_frame); | |
273 } | 250 } |
274 | 251 |
275 bool FFmpegVideoDecoder::FFmpegDecode( | 252 bool FFmpegVideoDecoder::FFmpegDecode( |
276 const scoped_refptr<DecoderBuffer>& buffer, | 253 const scoped_refptr<DecoderBuffer>& buffer, |
277 scoped_refptr<VideoFrame>* video_frame) { | 254 bool* has_produced_frame) { |
278 DCHECK(video_frame); | 255 DCHECK(!*has_produced_frame); |
279 | 256 |
280 // Create a packet for input data. | 257 // Create a packet for input data. |
281 // Due to FFmpeg API changes we no longer have const read-only pointers. | 258 // Due to FFmpeg API changes we no longer have const read-only pointers. |
282 AVPacket packet; | 259 AVPacket packet; |
283 av_init_packet(&packet); | 260 av_init_packet(&packet); |
284 if (buffer->end_of_stream()) { | 261 if (buffer->end_of_stream()) { |
285 packet.data = NULL; | 262 packet.data = NULL; |
286 packet.size = 0; | 263 packet.size = 0; |
287 } else { | 264 } else { |
288 packet.data = const_cast<uint8*>(buffer->data()); | 265 packet.data = const_cast<uint8*>(buffer->data()); |
289 packet.size = buffer->data_size(); | 266 packet.size = buffer->data_size(); |
290 | 267 |
291 // Let FFmpeg handle presentation timestamp reordering. | 268 // Let FFmpeg handle presentation timestamp reordering. |
292 codec_context_->reordered_opaque = buffer->timestamp().InMicroseconds(); | 269 codec_context_->reordered_opaque = buffer->timestamp().InMicroseconds(); |
293 } | 270 } |
294 | 271 |
295 int frame_decoded = 0; | 272 int frame_decoded = 0; |
296 int result = avcodec_decode_video2(codec_context_.get(), | 273 int result = avcodec_decode_video2(codec_context_.get(), |
297 av_frame_.get(), | 274 av_frame_.get(), |
298 &frame_decoded, | 275 &frame_decoded, |
299 &packet); | 276 &packet); |
300 // Log the problem if we can't decode a video frame and exit early. | 277 // Log the problem if we can't decode a video frame and exit early. |
301 if (result < 0) { | 278 if (result < 0) { |
302 LOG(ERROR) << "Error decoding video: " << buffer->AsHumanReadableString(); | 279 LOG(ERROR) << "Error decoding video: " << buffer->AsHumanReadableString(); |
303 *video_frame = NULL; | |
304 return false; | 280 return false; |
305 } | 281 } |
306 | 282 |
307 // FFmpeg says some codecs might have multiple frames per packet. Previous | 283 // FFmpeg says some codecs might have multiple frames per packet. Previous |
308 // discussions with rbultje@ indicate this shouldn't be true for the codecs | 284 // discussions with rbultje@ indicate this shouldn't be true for the codecs |
309 // we use. | 285 // we use. |
310 DCHECK_EQ(result, packet.size); | 286 DCHECK_EQ(result, packet.size); |
311 | 287 |
312 // If no frame was produced then signal that more data is required to | 288 // If no frame was produced then signal that more data is required to |
313 // produce more frames. This can happen under two circumstances: | 289 // produce more frames. This can happen under two circumstances: |
314 // 1) Decoder was recently initialized/flushed | 290 // 1) Decoder was recently initialized/flushed |
315 // 2) End of stream was reached and all internal frames have been output | 291 // 2) End of stream was reached and all internal frames have been output |
316 if (frame_decoded == 0) { | 292 if (frame_decoded == 0) { |
317 *video_frame = NULL; | |
318 return true; | 293 return true; |
319 } | 294 } |
320 | 295 |
321 // TODO(fbarchard): Work around for FFmpeg http://crbug.com/27675 | 296 // TODO(fbarchard): Work around for FFmpeg http://crbug.com/27675 |
322 // The decoder is in a bad state and not decoding correctly. | 297 // The decoder is in a bad state and not decoding correctly. |
323 // Checking for NULL avoids a crash in CopyPlane(). | 298 // Checking for NULL avoids a crash in CopyPlane(). |
324 if (!av_frame_->data[VideoFrame::kYPlane] || | 299 if (!av_frame_->data[VideoFrame::kYPlane] || |
325 !av_frame_->data[VideoFrame::kUPlane] || | 300 !av_frame_->data[VideoFrame::kUPlane] || |
326 !av_frame_->data[VideoFrame::kVPlane]) { | 301 !av_frame_->data[VideoFrame::kVPlane]) { |
327 LOG(ERROR) << "Video frame was produced yet has invalid frame data."; | 302 LOG(ERROR) << "Video frame was produced yet has invalid frame data."; |
328 *video_frame = NULL; | |
329 av_frame_unref(av_frame_.get()); | 303 av_frame_unref(av_frame_.get()); |
330 return false; | 304 return false; |
331 } | 305 } |
332 | 306 |
333 *video_frame = | 307 scoped_refptr<VideoFrame> frame = |
334 reinterpret_cast<VideoFrame*>(av_buffer_get_opaque(av_frame_->buf[0])); | 308 reinterpret_cast<VideoFrame*>(av_buffer_get_opaque(av_frame_->buf[0])); |
335 | 309 frame->set_timestamp( |
336 (*video_frame)->set_timestamp( | |
337 base::TimeDelta::FromMicroseconds(av_frame_->reordered_opaque)); | 310 base::TimeDelta::FromMicroseconds(av_frame_->reordered_opaque)); |
| 311 *has_produced_frame = true; |
| 312 output_cb_.Run(frame); |
338 | 313 |
339 av_frame_unref(av_frame_.get()); | 314 av_frame_unref(av_frame_.get()); |
340 return true; | 315 return true; |
341 } | 316 } |
342 | 317 |
343 void FFmpegVideoDecoder::ReleaseFFmpegResources() { | 318 void FFmpegVideoDecoder::ReleaseFFmpegResources() { |
344 codec_context_.reset(); | 319 codec_context_.reset(); |
345 av_frame_.reset(); | 320 av_frame_.reset(); |
346 } | 321 } |
347 | 322 |
(...skipping 16 matching lines...) Expand all Loading... |
364 if (!codec || avcodec_open2(codec_context_.get(), codec, NULL) < 0) { | 339 if (!codec || avcodec_open2(codec_context_.get(), codec, NULL) < 0) { |
365 ReleaseFFmpegResources(); | 340 ReleaseFFmpegResources(); |
366 return false; | 341 return false; |
367 } | 342 } |
368 | 343 |
369 av_frame_.reset(av_frame_alloc()); | 344 av_frame_.reset(av_frame_alloc()); |
370 return true; | 345 return true; |
371 } | 346 } |
372 | 347 |
373 } // namespace media | 348 } // namespace media |
OLD | NEW |