Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(371)

Unified Diff: content/renderer/media/webrtc/media_stream_remote_audio_track.cc

Issue 1966043006: Revert of MediaStream audio: Refactor 3 separate "glue" implementations into one. (Closed) Base URL: https://chromium.googlesource.com/chromium/src.git@master
Patch Set: Created 4 years, 7 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
Index: content/renderer/media/webrtc/media_stream_remote_audio_track.cc
diff --git a/content/renderer/media/webrtc/media_stream_remote_audio_track.cc b/content/renderer/media/webrtc/media_stream_remote_audio_track.cc
new file mode 100644
index 0000000000000000000000000000000000000000..e3940ab72b3bce7149897913d4af49745b673da0
--- /dev/null
+++ b/content/renderer/media/webrtc/media_stream_remote_audio_track.cc
@@ -0,0 +1,237 @@
+// Copyright 2015 The Chromium Authors. All rights reserved.
+// Use of this source code is governed by a BSD-style license that can be
+// found in the LICENSE file.
+
+#include "content/renderer/media/webrtc/media_stream_remote_audio_track.h"
+
+#include <stddef.h>
+
+#include <list>
+
+#include "base/logging.h"
+#include "content/public/renderer/media_stream_audio_sink.h"
+#include "third_party/webrtc/api/mediastreaminterface.h"
+
+namespace content {
+
+class MediaStreamRemoteAudioSource::AudioSink
+ : public webrtc::AudioTrackSinkInterface {
+ public:
+ AudioSink() {
+ }
+ ~AudioSink() override {
+ DCHECK(sinks_.empty());
+ }
+
+ void Add(MediaStreamAudioSink* sink, MediaStreamAudioTrack* track,
+ bool enabled) {
+ DCHECK(thread_checker_.CalledOnValidThread());
+ SinkInfo info(sink, track, enabled);
+ base::AutoLock lock(lock_);
+ sinks_.push_back(info);
+ }
+
+ void Remove(MediaStreamAudioSink* sink, MediaStreamAudioTrack* track) {
+ DCHECK(thread_checker_.CalledOnValidThread());
+ base::AutoLock lock(lock_);
+ sinks_.remove_if([&sink, &track](const SinkInfo& info) {
+ return info.sink == sink && info.track == track;
+ });
+ }
+
+ void SetEnabled(MediaStreamAudioTrack* track, bool enabled) {
+ DCHECK(thread_checker_.CalledOnValidThread());
+ base::AutoLock lock(lock_);
+ for (SinkInfo& info : sinks_) {
+ if (info.track == track)
+ info.enabled = enabled;
+ }
+ }
+
+ void RemoveAll(MediaStreamAudioTrack* track) {
+ base::AutoLock lock(lock_);
+ sinks_.remove_if([&track](const SinkInfo& info) {
+ return info.track == track;
+ });
+ }
+
+ bool IsNeeded() const {
+ DCHECK(thread_checker_.CalledOnValidThread());
+ return !sinks_.empty();
+ }
+
+ private:
+ void OnData(const void* audio_data, int bits_per_sample, int sample_rate,
+ size_t number_of_channels, size_t number_of_frames) override {
+ if (!audio_bus_ ||
+ static_cast<size_t>(audio_bus_->channels()) != number_of_channels ||
+ static_cast<size_t>(audio_bus_->frames()) != number_of_frames) {
+ audio_bus_ = media::AudioBus::Create(number_of_channels,
+ number_of_frames);
+ }
+
+ audio_bus_->FromInterleaved(audio_data, number_of_frames,
+ bits_per_sample / 8);
+
+ bool format_changed = false;
+ if (params_.format() != media::AudioParameters::AUDIO_PCM_LOW_LATENCY ||
+ static_cast<size_t>(params_.channels()) != number_of_channels ||
+ params_.sample_rate() != sample_rate ||
+ static_cast<size_t>(params_.frames_per_buffer()) != number_of_frames) {
+ params_ = media::AudioParameters(
+ media::AudioParameters::AUDIO_PCM_LOW_LATENCY,
+ media::GuessChannelLayout(number_of_channels),
+ sample_rate, 16, number_of_frames);
+ format_changed = true;
+ }
+
+ // TODO(tommi): We should get the timestamp from WebRTC.
+ base::TimeTicks estimated_capture_time(base::TimeTicks::Now());
+
+ base::AutoLock lock(lock_);
+ for (const SinkInfo& info : sinks_) {
+ if (info.enabled) {
+ if (format_changed)
+ info.sink->OnSetFormat(params_);
+ info.sink->OnData(*audio_bus_.get(), estimated_capture_time);
+ }
+ }
+ }
+
+ mutable base::Lock lock_;
+ struct SinkInfo {
+ SinkInfo(MediaStreamAudioSink* sink, MediaStreamAudioTrack* track,
+ bool enabled) : sink(sink), track(track), enabled(enabled) {}
+ MediaStreamAudioSink* sink;
+ MediaStreamAudioTrack* track;
+ bool enabled;
+ };
+ std::list<SinkInfo> sinks_;
+ base::ThreadChecker thread_checker_;
+ media::AudioParameters params_; // Only used on the callback thread.
+ std::unique_ptr<media::AudioBus>
+ audio_bus_; // Only used on the callback thread.
+};
+
+MediaStreamRemoteAudioTrack::MediaStreamRemoteAudioTrack(
+ const blink::WebMediaStreamSource& source, bool enabled)
+ : MediaStreamAudioTrack(false), source_(source), enabled_(enabled) {
+ DCHECK(source.getExtraData()); // Make sure the source has a native source.
+}
+
+MediaStreamRemoteAudioTrack::~MediaStreamRemoteAudioTrack() {
+ DCHECK(main_render_thread_checker_.CalledOnValidThread());
+ // Ensure the track is stopped.
+ MediaStreamAudioTrack::Stop();
+}
+
+void MediaStreamRemoteAudioTrack::SetEnabled(bool enabled) {
+ DCHECK(main_render_thread_checker_.CalledOnValidThread());
+
+ // This affects the shared state of the source for whether or not it's a part
+ // of the mixed audio that's rendered for remote tracks from WebRTC.
+ // All tracks from the same source will share this state and thus can step
+ // on each other's toes.
+ // This is also why we can't check the |enabled_| state for equality with
+ // |enabled| before setting the mixing enabled state. |enabled_| and the
+ // shared state might not be the same.
+ source()->SetEnabledForMixing(enabled);
+
+ enabled_ = enabled;
+ source()->SetSinksEnabled(this, enabled);
+}
+
+void MediaStreamRemoteAudioTrack::OnStop() {
+ DCHECK(main_render_thread_checker_.CalledOnValidThread());
+ DVLOG(1) << "MediaStreamRemoteAudioTrack::OnStop()";
+
+ source()->RemoveAll(this);
+
+ // Stop means that a track should be stopped permanently. But
+ // since there is no proper way of doing that on a remote track, we can
+ // at least disable the track. Blink will not call down to the content layer
+ // after a track has been stopped.
+ SetEnabled(false);
+}
+
+void MediaStreamRemoteAudioTrack::AddSink(MediaStreamAudioSink* sink) {
+ DCHECK(main_render_thread_checker_.CalledOnValidThread());
+ return source()->AddSink(sink, this, enabled_);
+}
+
+void MediaStreamRemoteAudioTrack::RemoveSink(MediaStreamAudioSink* sink) {
+ DCHECK(main_render_thread_checker_.CalledOnValidThread());
+ return source()->RemoveSink(sink, this);
+}
+
+media::AudioParameters MediaStreamRemoteAudioTrack::GetOutputFormat() const {
+ DCHECK(main_render_thread_checker_.CalledOnValidThread());
+ // This method is not implemented on purpose and should be removed.
+ // TODO(tommi): See comment for GetOutputFormat in MediaStreamAudioTrack.
+ NOTIMPLEMENTED();
+ return media::AudioParameters();
+}
+
+webrtc::AudioTrackInterface* MediaStreamRemoteAudioTrack::GetAudioAdapter() {
+ DCHECK(main_render_thread_checker_.CalledOnValidThread());
+ return source()->GetAudioAdapter();
+}
+
+MediaStreamRemoteAudioSource* MediaStreamRemoteAudioTrack::source() const {
+ return static_cast<MediaStreamRemoteAudioSource*>(source_.getExtraData());
+}
+
+MediaStreamRemoteAudioSource::MediaStreamRemoteAudioSource(
+ const scoped_refptr<webrtc::AudioTrackInterface>& track) : track_(track) {}
+
+MediaStreamRemoteAudioSource::~MediaStreamRemoteAudioSource() {
+ DCHECK(thread_checker_.CalledOnValidThread());
+}
+
+void MediaStreamRemoteAudioSource::SetEnabledForMixing(bool enabled) {
+ DCHECK(thread_checker_.CalledOnValidThread());
+ track_->set_enabled(enabled);
+}
+
+void MediaStreamRemoteAudioSource::AddSink(MediaStreamAudioSink* sink,
+ MediaStreamAudioTrack* track,
+ bool enabled) {
+ DCHECK(thread_checker_.CalledOnValidThread());
+ if (!sink_) {
+ sink_.reset(new AudioSink());
+ track_->AddSink(sink_.get());
+ }
+
+ sink_->Add(sink, track, enabled);
+}
+
+void MediaStreamRemoteAudioSource::RemoveSink(MediaStreamAudioSink* sink,
+ MediaStreamAudioTrack* track) {
+ DCHECK(thread_checker_.CalledOnValidThread());
+ DCHECK(sink_);
+
+ sink_->Remove(sink, track);
+
+ if (!sink_->IsNeeded()) {
+ track_->RemoveSink(sink_.get());
+ sink_.reset();
+ }
+}
+
+void MediaStreamRemoteAudioSource::SetSinksEnabled(MediaStreamAudioTrack* track,
+ bool enabled) {
+ if (sink_)
+ sink_->SetEnabled(track, enabled);
+}
+
+void MediaStreamRemoteAudioSource::RemoveAll(MediaStreamAudioTrack* track) {
+ if (sink_)
+ sink_->RemoveAll(track);
+}
+
+webrtc::AudioTrackInterface* MediaStreamRemoteAudioSource::GetAudioAdapter() {
+ DCHECK(thread_checker_.CalledOnValidThread());
+ return track_.get();
+}
+
+} // namespace content

Powered by Google App Engine
This is Rietveld 408576698