Index: content/renderer/media/media_stream_audio_track.cc |
diff --git a/content/renderer/media/media_stream_audio_track.cc b/content/renderer/media/media_stream_audio_track.cc |
index 575c277e8e30a9db5b34730b7844d3b3197a8946..9eee57d61428d6146bda28f8c930294a8f84857e 100644 |
--- a/content/renderer/media/media_stream_audio_track.cc |
+++ b/content/renderer/media/media_stream_audio_track.cc |
@@ -6,22 +6,22 @@ |
#include "base/callback_helpers.h" |
#include "base/logging.h" |
-#include "content/public/renderer/media_stream_audio_sink.h" |
-#include "media/base/audio_bus.h" |
#include "third_party/WebKit/public/platform/WebMediaStreamSource.h" |
+#include "third_party/webrtc/api/mediastreaminterface.h" |
namespace content { |
MediaStreamAudioTrack::MediaStreamAudioTrack(bool is_local_track) |
- : MediaStreamTrack(is_local_track), is_enabled_(1), weak_factory_(this) { |
- DVLOG(1) << "MediaStreamAudioTrack@" << this << "::MediaStreamAudioTrack(" |
+ : MediaStreamTrack(is_local_track) { |
+ DVLOG(1) << "MediaStreamAudioTrack::MediaStreamAudioTrack(is a " |
<< (is_local_track ? "local" : "remote") << " track)"; |
} |
MediaStreamAudioTrack::~MediaStreamAudioTrack() { |
- DCHECK(thread_checker_.CalledOnValidThread()); |
- DVLOG(1) << "MediaStreamAudioTrack@" << this << " is being destroyed."; |
- Stop(); |
+ DCHECK(main_render_thread_checker_.CalledOnValidThread()); |
+ DVLOG(1) << "MediaStreamAudioTrack::~MediaStreamAudioTrack()"; |
+ DCHECK(stop_callback_.is_null()) |
+ << "BUG: Subclass must ensure Stop() is called."; |
} |
// static |
@@ -34,101 +34,27 @@ |
return static_cast<MediaStreamAudioTrack*>(track.getExtraData()); |
} |
-void MediaStreamAudioTrack::AddSink(MediaStreamAudioSink* sink) { |
- DCHECK(thread_checker_.CalledOnValidThread()); |
- |
- DVLOG(1) << "Adding MediaStreamAudioSink@" << sink |
- << " to MediaStreamAudioTrack@" << this << '.'; |
- |
- // If the track has already stopped, just notify the sink of this fact without |
- // adding it. |
- if (stop_callback_.is_null()) { |
- sink->OnReadyStateChanged(blink::WebMediaStreamSource::ReadyStateEnded); |
- return; |
- } |
- |
- deliverer_.AddConsumer(sink); |
- sink->OnEnabledChanged(!!base::subtle::NoBarrier_Load(&is_enabled_)); |
-} |
- |
-void MediaStreamAudioTrack::RemoveSink(MediaStreamAudioSink* sink) { |
- DCHECK(thread_checker_.CalledOnValidThread()); |
- deliverer_.RemoveConsumer(sink); |
- DVLOG(1) << "Removed MediaStreamAudioSink@" << sink |
- << " from MediaStreamAudioTrack@" << this << '.'; |
-} |
- |
-media::AudioParameters MediaStreamAudioTrack::GetOutputFormat() const { |
- return deliverer_.GetAudioParameters(); |
-} |
- |
-void MediaStreamAudioTrack::SetEnabled(bool enabled) { |
- DCHECK(thread_checker_.CalledOnValidThread()); |
- DVLOG(1) << "MediaStreamAudioTrack@" << this << "::SetEnabled(" |
- << (enabled ? 'Y' : 'N') << ')'; |
- |
- const bool previously_enabled = |
- !!base::subtle::NoBarrier_AtomicExchange(&is_enabled_, enabled ? 1 : 0); |
- if (enabled == previously_enabled) |
- return; |
- |
- std::vector<MediaStreamAudioSink*> sinks_to_notify; |
- deliverer_.GetConsumerList(&sinks_to_notify); |
- for (MediaStreamAudioSink* sink : sinks_to_notify) |
- sink->OnEnabledChanged(enabled); |
-} |
- |
-void* MediaStreamAudioTrack::GetClassIdentifier() const { |
- return nullptr; |
-} |
- |
void MediaStreamAudioTrack::Start(const base::Closure& stop_callback) { |
- DCHECK(thread_checker_.CalledOnValidThread()); |
+ DCHECK(main_render_thread_checker_.CalledOnValidThread()); |
DCHECK(!stop_callback.is_null()); |
DCHECK(stop_callback_.is_null()); |
- DVLOG(1) << "Starting MediaStreamAudioTrack@" << this << '.'; |
+ DVLOG(1) << "MediaStreamAudioTrack::Start()"; |
stop_callback_ = stop_callback; |
} |
void MediaStreamAudioTrack::Stop() { |
- DCHECK(thread_checker_.CalledOnValidThread()); |
- DVLOG(1) << "Stopping MediaStreamAudioTrack@" << this << '.'; |
- |
+ DCHECK(main_render_thread_checker_.CalledOnValidThread()); |
+ DVLOG(1) << "MediaStreamAudioTrack::Stop()"; |
if (!stop_callback_.is_null()) |
base::ResetAndReturn(&stop_callback_).Run(); |
- |
- std::vector<MediaStreamAudioSink*> sinks_to_end; |
- deliverer_.GetConsumerList(&sinks_to_end); |
- for (MediaStreamAudioSink* sink : sinks_to_end) { |
- deliverer_.RemoveConsumer(sink); |
- sink->OnReadyStateChanged(blink::WebMediaStreamSource::ReadyStateEnded); |
- } |
- |
- weak_factory_.InvalidateWeakPtrs(); |
+ OnStop(); |
} |
-void MediaStreamAudioTrack::OnSetFormat(const media::AudioParameters& params) { |
- deliverer_.OnSetFormat(params); |
-} |
+void MediaStreamAudioTrack::OnStop() {} |
-void MediaStreamAudioTrack::OnData(const media::AudioBus& audio_bus, |
- base::TimeTicks reference_time) { |
- // Note: Using NoBarrier_Load because the timing of when the audio thread sees |
- // a changed |is_enabled_| value can be relaxed. |
- const bool deliver_data = !!base::subtle::NoBarrier_Load(&is_enabled_); |
- |
- if (deliver_data) { |
- deliverer_.OnData(audio_bus, reference_time); |
- } else { |
- // The W3C spec requires silent audio to flow while a track is disabled. |
- if (!silent_bus_ || silent_bus_->channels() != audio_bus.channels() || |
- silent_bus_->frames() != audio_bus.frames()) { |
- silent_bus_ = media::AudioBus::Create(audio_bus.channels(), |
- audio_bus.frames()); |
- silent_bus_->Zero(); |
- } |
- deliverer_.OnData(*silent_bus_, reference_time); |
- } |
+webrtc::AudioTrackInterface* MediaStreamAudioTrack::GetAudioAdapter() { |
+ NOTREACHED(); |
+ return nullptr; |
} |
} // namespace content |