Chromium Code Reviews| Index: media/cast/audio_receiver/audio_decoder.cc |
| diff --git a/media/cast/audio_receiver/audio_decoder.cc b/media/cast/audio_receiver/audio_decoder.cc |
| index b1a8256f2e9d7dd3dbdfcfc957d012987440d788..e46800f663d42a664e67bf234dc70d1fe6d985f8 100644 |
| --- a/media/cast/audio_receiver/audio_decoder.cc |
| +++ b/media/cast/audio_receiver/audio_decoder.cc |
| @@ -2,165 +2,257 @@ |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| -#include "base/logging.h" |
| #include "media/cast/audio_receiver/audio_decoder.h" |
| -#include "third_party/webrtc/modules/audio_coding/main/interface/audio_coding_module.h" |
| -#include "third_party/webrtc/modules/interface/module_common_types.h" |
| +#include "base/bind.h" |
| +#include "base/bind_helpers.h" |
| +#include "base/location.h" |
| +#include "base/logging.h" |
| +#include "base/memory/ref_counted.h" |
| +#include "base/stl_util.h" |
| +#include "base/sys_byteorder.h" |
| +#include "media/cast/cast_defines.h" |
| +#include "third_party/opus/src/include/opus.h" |
| namespace media { |
| namespace cast { |
| -AudioDecoder::AudioDecoder(scoped_refptr<CastEnvironment> cast_environment, |
| - const AudioReceiverConfig& audio_config, |
| - RtpPayloadFeedback* incoming_payload_feedback) |
| - : cast_environment_(cast_environment), |
| - audio_decoder_(webrtc::AudioCodingModule::Create(0)), |
| - cast_message_builder_(cast_environment->Clock(), |
| - incoming_payload_feedback, |
| - &frame_id_map_, |
| - audio_config.incoming_ssrc, |
| - true, |
| - 0), |
| - have_received_packets_(false), |
| - last_played_out_timestamp_(0) { |
| - audio_decoder_->InitializeReceiver(); |
| - |
| - webrtc::CodecInst receive_codec; |
| - switch (audio_config.codec) { |
| - case transport::kPcm16: |
| - receive_codec.pltype = audio_config.rtp_payload_type; |
| - strncpy(receive_codec.plname, "L16", 4); |
| - receive_codec.plfreq = audio_config.frequency; |
| - receive_codec.pacsize = -1; |
| - receive_codec.channels = audio_config.channels; |
| - receive_codec.rate = -1; |
| - break; |
| - case transport::kOpus: |
| - receive_codec.pltype = audio_config.rtp_payload_type; |
| - strncpy(receive_codec.plname, "opus", 5); |
| - receive_codec.plfreq = audio_config.frequency; |
| - receive_codec.pacsize = -1; |
| - receive_codec.channels = audio_config.channels; |
| - receive_codec.rate = -1; |
| - break; |
| - case transport::kExternalAudio: |
| - NOTREACHED() << "Codec must be specified for audio decoder"; |
| - break; |
| - } |
| - if (audio_decoder_->RegisterReceiveCodec(receive_codec) != 0) { |
| - NOTREACHED() << "Failed to register receive codec"; |
| +// Base class that handles the common problem of detecting dropped packets, and |
| +// then invoking the Decode() method implemented by the subclasses to convert |
| +// the packet payload into usable audio data. |
| +class AudioDecoder::ImplBase |
| + : public base::RefCountedThreadSafe<AudioDecoder::ImplBase> { |
| + public: |
| + ImplBase(const scoped_refptr<CastEnvironment>& cast_environment, |
| + transport::AudioCodec codec, |
| + int num_channels, |
| + int sampling_rate) |
| + : cast_environment_(cast_environment), |
| + codec_(codec), |
| + num_channels_(num_channels), |
| + cast_initialization_status_(STATUS_AUDIO_UNINITIALIZED), |
| + seen_first_packet_(false) { |
| + if (num_channels_ <= 0 || sampling_rate <= 0 || sampling_rate % 100 != 0) |
| + cast_initialization_status_ = STATUS_INVALID_AUDIO_CONFIGURATION; |
| } |
| - audio_decoder_->SetMaximumPlayoutDelay(audio_config.rtp_max_delay_ms); |
| - audio_decoder_->SetPlayoutMode(webrtc::streaming); |
| -} |
| + CastInitializationStatus InitializationResult() const { |
| + return cast_initialization_status_; |
| + } |
| -AudioDecoder::~AudioDecoder() {} |
| + void DecodeFrame(scoped_ptr<transport::EncodedAudioFrame> encoded_frame, |
| + const DecodeFrameCallback& callback) { |
| + DCHECK_EQ(cast_initialization_status_, STATUS_AUDIO_INITIALIZED); |
| -bool AudioDecoder::GetRawAudioFrame(int number_of_10ms_blocks, |
| - int desired_frequency, |
| - PcmAudioFrame* audio_frame, |
| - uint32* rtp_timestamp) { |
| - DCHECK(cast_environment_->CurrentlyOn(CastEnvironment::AUDIO)); |
| - // We don't care about the race case where a packet arrives at the same time |
| - // as this function in called. The data will be there the next time this |
| - // function is called. |
| - lock_.Acquire(); |
| - // Get a local copy under lock. |
| - bool have_received_packets = have_received_packets_; |
| - lock_.Release(); |
| - |
| - if (!have_received_packets) |
| - return false; |
| - |
| - audio_frame->samples.clear(); |
| - |
| - for (int i = 0; i < number_of_10ms_blocks; ++i) { |
| - webrtc::AudioFrame webrtc_audio_frame; |
| - if (0 != audio_decoder_->PlayoutData10Ms(desired_frequency, |
| - &webrtc_audio_frame)) { |
| - return false; |
| + scoped_ptr<AudioBus> decoded_audio; |
| + if (encoded_frame->codec != codec_) { |
| + NOTREACHED(); |
| + cast_environment_->PostTask(CastEnvironment::MAIN, |
| + FROM_HERE, |
| + base::Bind(callback, |
| + base::Passed(&decoded_audio), |
| + false)); |
| } |
| - if (webrtc_audio_frame.speech_type_ == webrtc::AudioFrame::kPLCCNG || |
| - webrtc_audio_frame.speech_type_ == webrtc::AudioFrame::kUndefined) { |
| - // We are only interested in real decoded audio. |
| - return false; |
| - } |
| - audio_frame->frequency = webrtc_audio_frame.sample_rate_hz_; |
| - audio_frame->channels = webrtc_audio_frame.num_channels_; |
| - if (i == 0) { |
| - // Use the timestamp from the first 10ms block. |
| - if (0 != audio_decoder_->PlayoutTimestamp(rtp_timestamp)) { |
| - return false; |
| + COMPILE_ASSERT(sizeof(encoded_frame->frame_id) == sizeof(last_frame_id_), |
| + size_of_frame_id_types_do_not_match); |
| + bool is_continuous = true; |
| + if (seen_first_packet_) { |
| + const uint32 packets_ahead = encoded_frame->frame_id - last_frame_id_; |
| + if (packets_ahead > 1) { |
| + RecoverBecausePacketsWereDropped(); |
| + is_continuous = false; |
| } |
| - lock_.Acquire(); |
| - last_played_out_timestamp_ = *rtp_timestamp; |
| - lock_.Release(); |
| + } else { |
| + seen_first_packet_ = true; |
| } |
| - int samples_per_10ms = webrtc_audio_frame.samples_per_channel_; |
| + last_frame_id_ = encoded_frame->frame_id; |
| - audio_frame->samples.insert( |
| - audio_frame->samples.end(), |
| - &webrtc_audio_frame.data_[0], |
| - &webrtc_audio_frame.data_[samples_per_10ms * audio_frame->channels]); |
| + decoded_audio = Decode( |
| + reinterpret_cast<uint8*>(string_as_array(&encoded_frame->data)), |
| + encoded_frame->data.size()); |
| + cast_environment_->PostTask(CastEnvironment::MAIN, |
| + FROM_HERE, |
| + base::Bind(callback, |
| + base::Passed(&decoded_audio), |
| + is_continuous)); |
| } |
| - return true; |
| -} |
| -void AudioDecoder::IncomingParsedRtpPacket(const uint8* payload_data, |
| - size_t payload_size, |
| - const RtpCastHeader& rtp_header) { |
| - DCHECK(cast_environment_->CurrentlyOn(CastEnvironment::MAIN)); |
| - DCHECK_LE(payload_size, kMaxIpPacketSize); |
| - audio_decoder_->IncomingPacket( |
| - payload_data, static_cast<int32>(payload_size), rtp_header.webrtc); |
| - lock_.Acquire(); |
| - have_received_packets_ = true; |
| - uint32 last_played_out_timestamp = last_played_out_timestamp_; |
| - lock_.Release(); |
| - |
| - PacketType packet_type = frame_id_map_.InsertPacket(rtp_header); |
| - if (packet_type != kNewPacketCompletingFrame) |
| - return; |
| + protected: |
| + friend class base::RefCountedThreadSafe<ImplBase>; |
| + virtual ~ImplBase() {} |
| - cast_message_builder_.CompleteFrameReceived(rtp_header.frame_id, |
| - rtp_header.is_key_frame); |
| + virtual void RecoverBecausePacketsWereDropped() {} |
| - frame_id_rtp_timestamp_map_[rtp_header.frame_id] = |
| - rtp_header.webrtc.header.timestamp; |
| + // Note: Implementation of Decode() is allowed to mutate |data|. |
| + virtual scoped_ptr<AudioBus> Decode(uint8* data, size_t size) = 0; |
| - if (last_played_out_timestamp == 0) |
| - return; // Nothing is played out yet. |
| + const scoped_refptr<CastEnvironment> cast_environment_; |
| + const transport::AudioCodec codec_; |
| + const int num_channels_; |
| - uint32 latest_frame_id_to_remove = 0; |
| - bool frame_to_remove = false; |
| + // Subclass' ctor is expected to set this to STATUS_AUDIO_INITIALIZED. |
| + CastInitializationStatus cast_initialization_status_; |
| - FrameIdRtpTimestampMap::iterator it = frame_id_rtp_timestamp_map_.begin(); |
| - while (it != frame_id_rtp_timestamp_map_.end()) { |
| - if (IsNewerRtpTimestamp(it->second, last_played_out_timestamp)) { |
| - break; |
| + private: |
| + bool seen_first_packet_; |
| + uint32 last_frame_id_; |
| + |
| + DISALLOW_COPY_AND_ASSIGN(ImplBase); |
| +}; |
| + |
| +class AudioDecoder::OpusImpl : public AudioDecoder::ImplBase { |
| + public: |
| + OpusImpl(const scoped_refptr<CastEnvironment>& cast_environment, |
| + int num_channels, |
| + int sampling_rate) |
| + : ImplBase(cast_environment, |
| + transport::kOpus, |
| + num_channels, |
| + sampling_rate), |
| + decoder_memory_(new uint8[opus_decoder_get_size(num_channels)]), |
| + opus_decoder_(reinterpret_cast<OpusDecoder*>(decoder_memory_.get())), |
| + max_samples_per_frame_( |
| + kOpusMaxPacketDurationMillis * sampling_rate / 1000), |
| + buffer_(new float[max_samples_per_frame_ * num_channels]) { |
| + if (ImplBase::cast_initialization_status_ != STATUS_AUDIO_UNINITIALIZED) |
|
hubbe
2014/03/28 19:50:59
DCHECK?
miu
2014/03/28 23:51:58
No, it needs to be run-time code in normal builds.
|
| + return; |
| + if (opus_decoder_init(opus_decoder_, sampling_rate, num_channels) != |
| + OPUS_OK) { |
| + ImplBase::cast_initialization_status_ = |
| + STATUS_INVALID_AUDIO_CONFIGURATION; |
| + return; |
| } |
| - frame_to_remove = true; |
| - latest_frame_id_to_remove = it->first; |
| - frame_id_rtp_timestamp_map_.erase(it); |
| - it = frame_id_rtp_timestamp_map_.begin(); |
| + ImplBase::cast_initialization_status_ = STATUS_AUDIO_INITIALIZED; |
| } |
| - if (!frame_to_remove) |
| - return; |
| - frame_id_map_.RemoveOldFrames(latest_frame_id_to_remove); |
| + private: |
| + virtual ~OpusImpl() {} |
| + |
| + virtual void RecoverBecausePacketsWereDropped() OVERRIDE { |
| + // Passing NULL for the input data notifies the decoder of packet loss. |
| + const opus_int32 result = |
| + opus_decode_float( |
| + opus_decoder_, NULL, 0, buffer_.get(), max_samples_per_frame_, 0); |
| + DCHECK_GE(result, 0); |
| + } |
| + |
| + virtual scoped_ptr<AudioBus> Decode(uint8* data, size_t size) OVERRIDE { |
| + scoped_ptr<AudioBus> audio_bus; |
| + const opus_int32 num_samples_decoded = opus_decode_float( |
| + opus_decoder_, data, size, buffer_.get(), max_samples_per_frame_, 0); |
| + if (num_samples_decoded <= 0) |
| + return audio_bus.Pass(); // Decode error. |
| + |
| + // Copy interleaved samples from |buffer_| into a new AudioBus (where |
| + // samples are stored in planar format, for each channel). |
| + audio_bus = AudioBus::Create(num_channels_, num_samples_decoded).Pass(); |
| + // TODO(miu): This should be moved into AudioBus::FromInterleaved(). |
| + for (int ch = 0; ch < num_channels_; ++ch) { |
| + const float* src = buffer_.get() + ch; |
| + const float* const src_end = src + num_samples_decoded * num_channels_; |
| + float* dest = audio_bus->channel(ch); |
| + for (; src < src_end; src += num_channels_, ++dest) |
| + *dest = *src; |
| + } |
| + return audio_bus.Pass(); |
| + } |
| + |
| + const scoped_ptr<uint8[]> decoder_memory_; |
| + OpusDecoder* const opus_decoder_; |
| + const int max_samples_per_frame_; |
| + const scoped_ptr<float[]> buffer_; |
| + |
| + // According to documentation in third_party/opus/src/include/opus.h, we must |
| + // provide enough space in |buffer_| to contain 120ms of samples. At 48 kHz, |
| + // then, that means 5760 samples times the number of channels. |
| + static const int kOpusMaxPacketDurationMillis = 120; |
| + |
| + DISALLOW_COPY_AND_ASSIGN(OpusImpl); |
| +}; |
| + |
| +class AudioDecoder::Pcm16Impl : public AudioDecoder::ImplBase { |
| + public: |
| + Pcm16Impl(const scoped_refptr<CastEnvironment>& cast_environment, |
| + int num_channels, |
| + int sampling_rate) |
| + : ImplBase(cast_environment, |
| + transport::kPcm16, |
| + num_channels, |
| + sampling_rate) { |
| + if (ImplBase::cast_initialization_status_ != STATUS_AUDIO_UNINITIALIZED) |
| + return; |
| + ImplBase::cast_initialization_status_ = STATUS_AUDIO_INITIALIZED; |
| + } |
| + |
| + private: |
| + virtual ~Pcm16Impl() {} |
| + |
| + virtual scoped_ptr<AudioBus> Decode(uint8* data, size_t size) OVERRIDE { |
| + scoped_ptr<AudioBus> audio_bus; |
| + const int num_samples = size / sizeof(int16) / num_channels_; |
| + if (num_samples <= 0) |
| + return audio_bus.Pass(); |
| + |
| + int16* const pcm_data = reinterpret_cast<int16*>(data); |
| +#if defined(ARCH_CPU_LITTLE_ENDIAN) |
| + // Convert endianness. |
| + const int num_elements = num_samples * num_channels_; |
| + for (int i = 0; i < num_elements; ++i) |
| + pcm_data[i] = static_cast<int16>(base::NetToHost16(pcm_data[i])); |
| +#endif |
| + audio_bus = AudioBus::Create(num_channels_, num_samples).Pass(); |
| + audio_bus->FromInterleaved(pcm_data, num_samples, sizeof(int16)); |
| + return audio_bus.Pass(); |
| + } |
| + |
| + DISALLOW_COPY_AND_ASSIGN(Pcm16Impl); |
| +}; |
| + |
| +AudioDecoder::AudioDecoder( |
| + const scoped_refptr<CastEnvironment>& cast_environment, |
| + const AudioReceiverConfig& audio_config) |
| + : cast_environment_(cast_environment) { |
| + switch (audio_config.codec) { |
| + case transport::kOpus: |
| + impl_ = new OpusImpl(cast_environment, |
| + audio_config.channels, |
| + audio_config.frequency); |
| + break; |
| + case transport::kPcm16: |
| + impl_ = new Pcm16Impl(cast_environment, |
| + audio_config.channels, |
| + audio_config.frequency); |
| + break; |
| + default: |
| + NOTREACHED() << "Unknown or unspecified codec."; |
| + break; |
| + } |
| } |
| -bool AudioDecoder::TimeToSendNextCastMessage(base::TimeTicks* time_to_send) { |
| - DCHECK(cast_environment_->CurrentlyOn(CastEnvironment::MAIN)); |
| - return cast_message_builder_.TimeToSendNextCastMessage(time_to_send); |
| +AudioDecoder::~AudioDecoder() {} |
| + |
| +CastInitializationStatus AudioDecoder::InitializationResult() const { |
| + if (impl_) |
| + return impl_->InitializationResult(); |
| + return STATUS_UNSUPPORTED_AUDIO_CODEC; |
| } |
| -void AudioDecoder::SendCastMessage() { |
| - DCHECK(cast_environment_->CurrentlyOn(CastEnvironment::MAIN)); |
| - cast_message_builder_.UpdateCastMessage(); |
| +void AudioDecoder::DecodeFrame( |
| + scoped_ptr<transport::EncodedAudioFrame> encoded_frame, |
| + const DecodeFrameCallback& callback) { |
| + DCHECK(encoded_frame.get()); |
| + DCHECK(!callback.is_null()); |
| + if (!impl_) { |
| + callback.Run(make_scoped_ptr<AudioBus>(NULL), false); |
| + return; |
| + } |
| + cast_environment_->PostTask(CastEnvironment::AUDIO, |
| + FROM_HERE, |
| + base::Bind(&AudioDecoder::ImplBase::DecodeFrame, |
| + impl_, |
| + base::Passed(&encoded_frame), |
| + callback)); |
| } |
| } // namespace cast |