| Index: media/filters/ffmpeg_audio_decoder.cc
|
| diff --git a/media/filters/ffmpeg_audio_decoder.cc b/media/filters/ffmpeg_audio_decoder.cc
|
| index 5384d63c48874a70c01b8c00aece2eebe5392d5b..3f1e81a37f7ddba6950a455336580c8c36eacc0f 100644
|
| --- a/media/filters/ffmpeg_audio_decoder.cc
|
| +++ b/media/filters/ffmpeg_audio_decoder.cc
|
| @@ -8,6 +8,7 @@
|
| #include "base/callback_helpers.h"
|
| #include "base/location.h"
|
| #include "base/message_loop_proxy.h"
|
| +#include "media/base/audio_bus.h"
|
| #include "media/base/audio_decoder_config.h"
|
| #include "media/base/audio_timestamp_helper.h"
|
| #include "media/base/data_buffer.h"
|
| @@ -283,6 +284,10 @@ bool FFmpegAudioDecoder::ConfigureDecoder() {
|
| codec_context_ = avcodec_alloc_context3(NULL);
|
| AudioDecoderConfigToAVCodecContext(config, codec_context_);
|
|
|
| + // MP3 decodes to S16P which we don't support, tell it to use S16 instead.
|
| + if (codec_context_->sample_fmt == AV_SAMPLE_FMT_S16P)
|
| + codec_context_->request_sample_fmt = AV_SAMPLE_FMT_S16;
|
| +
|
| AVCodec* codec = avcodec_find_decoder(codec_context_->codec_id);
|
| if (!codec || avcodec_open2(codec_context_, codec, NULL) < 0) {
|
| DLOG(ERROR) << "Could not initialize audio decoder: "
|
| @@ -290,6 +295,26 @@ bool FFmpegAudioDecoder::ConfigureDecoder() {
|
| return false;
|
| }
|
|
|
| + // Ensure avcodec_open2() respected our format request.
|
| + if (codec_context_->sample_fmt == AV_SAMPLE_FMT_S16P) {
|
| + DLOG(ERROR) << "Unable to configure a supported sample format: "
|
| + << codec_context_->sample_fmt;
|
| + return false;
|
| + }
|
| +
|
| + // Some codecs will only output float data, so we need to convert to integer
|
| + // before returning the decoded buffer.
|
| + if (codec_context_->sample_fmt == AV_SAMPLE_FMT_FLTP ||
|
| + codec_context_->sample_fmt == AV_SAMPLE_FMT_FLT) {
|
| + // Preallocate the AudioBus for float conversions. We can treat interleaved
|
| + // float data as a single planar channel since our output is expected in an
|
| + // interleaved format anyways.
|
| + int channels = codec_context_->channels;
|
| + if (codec_context_->sample_fmt == AV_SAMPLE_FMT_FLT)
|
| + channels = 1;
|
| + converter_bus_ = AudioBus::CreateWrapper(channels);
|
| + }
|
| +
|
| // Success!
|
| av_frame_ = avcodec_alloc_frame();
|
| bits_per_channel_ = config.bits_per_channel();
|
| @@ -297,6 +322,7 @@ bool FFmpegAudioDecoder::ConfigureDecoder() {
|
| samples_per_second_ = config.samples_per_second();
|
| output_timestamp_helper_.reset(new AudioTimestampHelper(
|
| config.bytes_per_frame(), config.samples_per_second()));
|
| + bytes_per_frame_ = config.bytes_per_frame();
|
| return true;
|
| }
|
|
|
| @@ -374,7 +400,6 @@ void FFmpegAudioDecoder::RunDecodeLoop(
|
| }
|
| }
|
|
|
| - const uint8* decoded_audio_data = NULL;
|
| int decoded_audio_size = 0;
|
| if (frame_decoded) {
|
| int output_sample_rate = av_frame_->sample_rate;
|
| @@ -388,24 +413,64 @@ void FFmpegAudioDecoder::RunDecodeLoop(
|
| break;
|
| }
|
|
|
| - decoded_audio_data = av_frame_->data[0];
|
| decoded_audio_size = av_samples_get_buffer_size(
|
| NULL, codec_context_->channels, av_frame_->nb_samples,
|
| codec_context_->sample_fmt, 1);
|
| + // If we're decoding into float, adjust audio size.
|
| + if (converter_bus_ && bits_per_channel_ / 8 != sizeof(float)) {
|
| + DCHECK(codec_context_->sample_fmt == AV_SAMPLE_FMT_FLT ||
|
| + codec_context_->sample_fmt == AV_SAMPLE_FMT_FLTP);
|
| + decoded_audio_size *=
|
| + static_cast<float>(bits_per_channel_ / 8) / sizeof(float);
|
| + }
|
| }
|
|
|
| - scoped_refptr<DataBuffer> output;
|
| -
|
| + int start_sample = 0;
|
| if (decoded_audio_size > 0 && output_bytes_to_drop_ > 0) {
|
| + DCHECK_EQ(decoded_audio_size % bytes_per_frame_, 0)
|
| + << "Decoder didn't output full frames";
|
| +
|
| int dropped_size = std::min(decoded_audio_size, output_bytes_to_drop_);
|
| - decoded_audio_data += dropped_size;
|
| + start_sample = dropped_size / bytes_per_frame_;
|
| decoded_audio_size -= dropped_size;
|
| output_bytes_to_drop_ -= dropped_size;
|
| }
|
|
|
| + scoped_refptr<DataBuffer> output;
|
| if (decoded_audio_size > 0) {
|
| - // Copy the audio samples into an output buffer.
|
| - output = new DataBuffer(decoded_audio_data, decoded_audio_size);
|
| + DCHECK_EQ(decoded_audio_size % bytes_per_frame_, 0)
|
| + << "Decoder didn't output full frames";
|
| +
|
| + // Convert float data using an AudioBus.
|
| + if (converter_bus_) {
|
| + // Setup the AudioBus as a wrapper of the AVFrame data and then use
|
| + // AudioBus::ToInterleaved() to convert the data as necessary.
|
| + int skip_frames = start_sample;
|
| + int total_frames = av_frame_->nb_samples - start_sample;
|
| + if (codec_context_->sample_fmt == AV_SAMPLE_FMT_FLT) {
|
| + DCHECK_EQ(converter_bus_->channels(), 1);
|
| + total_frames *= codec_context_->channels;
|
| + skip_frames *= codec_context_->channels;
|
| + }
|
| + converter_bus_->set_frames(total_frames);
|
| + DCHECK_EQ(decoded_audio_size,
|
| + converter_bus_->frames() * bytes_per_frame_);
|
| +
|
| + for (int i = 0; i < converter_bus_->channels(); ++i) {
|
| + converter_bus_->SetChannelData(i, reinterpret_cast<float*>(
|
| + av_frame_->extended_data[i]) + skip_frames);
|
| + }
|
| +
|
| + output = new DataBuffer(decoded_audio_size);
|
| + output->SetDataSize(decoded_audio_size);
|
| + converter_bus_->ToInterleaved(
|
| + converter_bus_->frames(), bits_per_channel_ / 8,
|
| + output->GetWritableData());
|
| + } else {
|
| + output = new DataBuffer(
|
| + av_frame_->extended_data[0] + start_sample * bytes_per_frame_,
|
| + decoded_audio_size);
|
| + }
|
| output->SetTimestamp(output_timestamp_helper_->GetTimestamp());
|
| output->SetDuration(
|
| output_timestamp_helper_->GetDuration(decoded_audio_size));
|
|
|