Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(959)

Unified Diff: content/renderer/media/webrtc/peer_connection_remote_audio_source.cc

Issue 1834323002: MediaStream audio: Refactor 3 separate "glue" implementations into one. (Closed) Base URL: https://chromium.googlesource.com/chromium/src.git@master
Patch Set: Addressed comments from PS2: AudioInputDevice --> AudioCapturerSource, and refptr foo in WebRtcMedi… Created 4 years, 9 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
Index: content/renderer/media/webrtc/peer_connection_remote_audio_source.cc
diff --git a/content/renderer/media/webrtc/media_stream_remote_audio_track.cc b/content/renderer/media/webrtc/peer_connection_remote_audio_source.cc
similarity index 14%
rename from content/renderer/media/webrtc/media_stream_remote_audio_track.cc
rename to content/renderer/media/webrtc/peer_connection_remote_audio_source.cc
index d03c80676ea49e8453b7d402c61dab2fdf98b927..fa490548d2cdc357270a191e0036006741624ba4 100644
--- a/content/renderer/media/webrtc/media_stream_remote_audio_track.cc
+++ b/content/renderer/media/webrtc/peer_connection_remote_audio_source.cc
@@ -2,149 +2,64 @@
// Use of this source code is governed by a BSD-style license that can be
// found in the LICENSE file.
-#include "content/renderer/media/webrtc/media_stream_remote_audio_track.h"
-
-#include <stddef.h>
-
-#include <list>
+#include "content/renderer/media/webrtc/peer_connection_remote_audio_source.h"
#include "base/logging.h"
-#include "content/public/renderer/media_stream_audio_sink.h"
-#include "third_party/webrtc/api/mediastreaminterface.h"
+#include "base/time/time.h"
+#include "media/base/audio_bus.h"
namespace content {
-class MediaStreamRemoteAudioSource::AudioSink
- : public webrtc::AudioTrackSinkInterface {
- public:
- AudioSink() {
- }
- ~AudioSink() override {
- DCHECK(sinks_.empty());
- }
-
- void Add(MediaStreamAudioSink* sink, MediaStreamAudioTrack* track,
- bool enabled) {
- DCHECK(thread_checker_.CalledOnValidThread());
- SinkInfo info(sink, track, enabled);
- base::AutoLock lock(lock_);
- sinks_.push_back(info);
- }
-
- void Remove(MediaStreamAudioSink* sink, MediaStreamAudioTrack* track) {
- DCHECK(thread_checker_.CalledOnValidThread());
- base::AutoLock lock(lock_);
- sinks_.remove_if([&sink, &track](const SinkInfo& info) {
- return info.sink == sink && info.track == track;
- });
- }
-
- void SetEnabled(MediaStreamAudioTrack* track, bool enabled) {
- DCHECK(thread_checker_.CalledOnValidThread());
- base::AutoLock lock(lock_);
- for (SinkInfo& info : sinks_) {
- if (info.track == track)
- info.enabled = enabled;
- }
- }
+namespace {
+// Used as an identifier for the down-casters.
+void* const kClassIdentifier = const_cast<void**>(&kClassIdentifier);
+} // namespace
- void RemoveAll(MediaStreamAudioTrack* track) {
- base::AutoLock lock(lock_);
- sinks_.remove_if([&track](const SinkInfo& info) {
- return info.track == track;
- });
- }
-
- bool IsNeeded() const {
- DCHECK(thread_checker_.CalledOnValidThread());
- return !sinks_.empty();
- }
-
- private:
- void OnData(const void* audio_data, int bits_per_sample, int sample_rate,
- size_t number_of_channels, size_t number_of_frames) override {
- if (!audio_bus_ ||
- static_cast<size_t>(audio_bus_->channels()) != number_of_channels ||
- static_cast<size_t>(audio_bus_->frames()) != number_of_frames) {
- audio_bus_ = media::AudioBus::Create(number_of_channels,
- number_of_frames);
- }
-
- audio_bus_->FromInterleaved(audio_data, number_of_frames,
- bits_per_sample / 8);
-
- bool format_changed = false;
- if (params_.format() != media::AudioParameters::AUDIO_PCM_LOW_LATENCY ||
- static_cast<size_t>(params_.channels()) != number_of_channels ||
- params_.sample_rate() != sample_rate ||
- static_cast<size_t>(params_.frames_per_buffer()) != number_of_frames) {
- params_ = media::AudioParameters(
- media::AudioParameters::AUDIO_PCM_LOW_LATENCY,
- media::GuessChannelLayout(number_of_channels),
- sample_rate, 16, number_of_frames);
- format_changed = true;
- }
-
- // TODO(tommi): We should get the timestamp from WebRTC.
- base::TimeTicks estimated_capture_time(base::TimeTicks::Now());
-
- base::AutoLock lock(lock_);
- for (const SinkInfo& info : sinks_) {
- if (info.enabled) {
- if (format_changed)
- info.sink->OnSetFormat(params_);
- info.sink->OnData(*audio_bus_.get(), estimated_capture_time);
- }
- }
- }
-
- mutable base::Lock lock_;
- struct SinkInfo {
- SinkInfo(MediaStreamAudioSink* sink, MediaStreamAudioTrack* track,
- bool enabled) : sink(sink), track(track), enabled(enabled) {}
- MediaStreamAudioSink* sink;
- MediaStreamAudioTrack* track;
- bool enabled;
- };
- std::list<SinkInfo> sinks_;
- base::ThreadChecker thread_checker_;
- media::AudioParameters params_; // Only used on the callback thread.
- scoped_ptr<media::AudioBus> audio_bus_; // Only used on the callback thread.
-};
-
-MediaStreamRemoteAudioTrack::MediaStreamRemoteAudioTrack(
- const blink::WebMediaStreamSource& source, bool enabled)
- : MediaStreamAudioTrack(false), source_(source), enabled_(enabled) {
- DCHECK(source.getExtraData()); // Make sure the source has a native source.
+PeerConnectionRemoteAudioTrack::PeerConnectionRemoteAudioTrack(
+ scoped_refptr<webrtc::AudioTrackInterface> track_interface)
+ : MediaStreamAudioTrack(false /* is_local_track */),
+ track_interface_(std::move(track_interface)) {
+ DVLOG(1)
+ << "PeerConnectionRemoteAudioTrack::PeerConnectionRemoteAudioTrack()";
}
-MediaStreamRemoteAudioTrack::~MediaStreamRemoteAudioTrack() {
- DCHECK(main_render_thread_checker_.CalledOnValidThread());
+PeerConnectionRemoteAudioTrack::~PeerConnectionRemoteAudioTrack() {
+ DVLOG(1)
+ << "PeerConnectionRemoteAudioTrack::~PeerConnectionRemoteAudioTrack()";
// Ensure the track is stopped.
MediaStreamAudioTrack::Stop();
}
-void MediaStreamRemoteAudioTrack::SetEnabled(bool enabled) {
- DCHECK(main_render_thread_checker_.CalledOnValidThread());
+// static
+PeerConnectionRemoteAudioTrack* PeerConnectionRemoteAudioTrack::From(
+ MediaStreamAudioTrack* track) {
+ if (track && track->GetClassIdentifier() == kClassIdentifier)
+ return static_cast<PeerConnectionRemoteAudioTrack*>(track);
+ return nullptr;
+}
+
+void PeerConnectionRemoteAudioTrack::SetEnabled(bool enabled) {
+ DCHECK(thread_checker_.CalledOnValidThread());
// This affects the shared state of the source for whether or not it's a part
// of the mixed audio that's rendered for remote tracks from WebRTC.
// All tracks from the same source will share this state and thus can step
// on each other's toes.
- // This is also why we can't check the |enabled_| state for equality with
- // |enabled| before setting the mixing enabled state. |enabled_| and the
- // shared state might not be the same.
- source()->SetEnabledForMixing(enabled);
+ // This is also why we can't check the enabled state for equality with
+ // |enabled| before setting the mixing enabled state. This track's enabled
+ // state and the shared state might not be the same.
+ track_interface_->set_enabled(enabled);
- enabled_ = enabled;
- source()->SetSinksEnabled(this, enabled);
+ MediaStreamAudioTrack::SetEnabled(enabled);
}
-void MediaStreamRemoteAudioTrack::OnStop() {
- DCHECK(main_render_thread_checker_.CalledOnValidThread());
- DVLOG(1) << "MediaStreamRemoteAudioTrack::OnStop()";
+void* PeerConnectionRemoteAudioTrack::GetClassIdentifier() const {
+ return kClassIdentifier;
+}
- source()->RemoveAll(this);
+void PeerConnectionRemoteAudioTrack::OnStop() {
+ DCHECK(thread_checker_.CalledOnValidThread());
+ DVLOG(1) << "PeerConnectionRemoteAudioTrack::OnStop()";
// Stop means that a track should be stopped permanently. But
// since there is no proper way of doing that on a remote track, we can
@@ -153,84 +68,81 @@ void MediaStreamRemoteAudioTrack::OnStop() {
SetEnabled(false);
}
-void MediaStreamRemoteAudioTrack::AddSink(MediaStreamAudioSink* sink) {
- DCHECK(main_render_thread_checker_.CalledOnValidThread());
- return source()->AddSink(sink, this, enabled_);
+PeerConnectionRemoteAudioSource::PeerConnectionRemoteAudioSource(
+ scoped_refptr<webrtc::AudioTrackInterface> track_interface)
+ : MediaStreamAudioSource(false /* is_local_source */),
+ track_interface_(std::move(track_interface)),
+ is_sink_of_peer_connection_(false) {
+ DCHECK(track_interface_);
+ DVLOG(1)
+ << "PeerConnectionRemoteAudioSource::PeerConnectionRemoteAudioSource()";
}
-void MediaStreamRemoteAudioTrack::RemoveSink(MediaStreamAudioSink* sink) {
- DCHECK(main_render_thread_checker_.CalledOnValidThread());
- return source()->RemoveSink(sink, this);
+PeerConnectionRemoteAudioSource::~PeerConnectionRemoteAudioSource() {
+ DVLOG(1)
+ << "PeerConnectionRemoteAudioSource::~PeerConnectionRemoteAudioSource()";
+ EnsureSourceIsStopped();
}
-media::AudioParameters MediaStreamRemoteAudioTrack::GetOutputFormat() const {
- DCHECK(main_render_thread_checker_.CalledOnValidThread());
- // This method is not implemented on purpose and should be removed.
- // TODO(tommi): See comment for GetOutputFormat in MediaStreamAudioTrack.
- NOTIMPLEMENTED();
- return media::AudioParameters();
-}
-
-webrtc::AudioTrackInterface* MediaStreamRemoteAudioTrack::GetAudioAdapter() {
- DCHECK(main_render_thread_checker_.CalledOnValidThread());
- return source()->GetAudioAdapter();
-}
-
-MediaStreamRemoteAudioSource* MediaStreamRemoteAudioTrack::source() const {
- return static_cast<MediaStreamRemoteAudioSource*>(source_.getExtraData());
-}
-
-MediaStreamRemoteAudioSource::MediaStreamRemoteAudioSource(
- const scoped_refptr<webrtc::AudioTrackInterface>& track) : track_(track) {}
-
-MediaStreamRemoteAudioSource::~MediaStreamRemoteAudioSource() {
+scoped_ptr<MediaStreamAudioTrack>
+PeerConnectionRemoteAudioSource::CreateMediaStreamAudioTrack(
+ const std::string& id) {
DCHECK(thread_checker_.CalledOnValidThread());
+ return scoped_ptr<MediaStreamAudioTrack>(
+ new PeerConnectionRemoteAudioTrack(track_interface_));
}
-void MediaStreamRemoteAudioSource::SetEnabledForMixing(bool enabled) {
+bool PeerConnectionRemoteAudioSource::EnsureSourceIsStarted() {
DCHECK(thread_checker_.CalledOnValidThread());
- track_->set_enabled(enabled);
+ if (is_sink_of_peer_connection_)
+ return true;
+ VLOG(1) << "Starting PeerConnection remote audio source with id="
+ << track_interface_->id();
+ track_interface_->AddSink(this);
+ is_sink_of_peer_connection_ = true;
+ return true;
}
-void MediaStreamRemoteAudioSource::AddSink(MediaStreamAudioSink* sink,
- MediaStreamAudioTrack* track,
- bool enabled) {
+void PeerConnectionRemoteAudioSource::EnsureSourceIsStopped() {
DCHECK(thread_checker_.CalledOnValidThread());
- if (!sink_) {
- sink_.reset(new AudioSink());
- track_->AddSink(sink_.get());
+ if (is_sink_of_peer_connection_) {
+ track_interface_->RemoveSink(this);
+ is_sink_of_peer_connection_ = false;
+ VLOG(1) << "Stopped PeerConnection remote audio source with id="
+ << track_interface_->id();
}
-
- sink_->Add(sink, track, enabled);
}
-void MediaStreamRemoteAudioSource::RemoveSink(MediaStreamAudioSink* sink,
- MediaStreamAudioTrack* track) {
- DCHECK(thread_checker_.CalledOnValidThread());
- DCHECK(sink_);
-
- sink_->Remove(sink, track);
+void PeerConnectionRemoteAudioSource::OnData(const void* audio_data,
+ int bits_per_sample,
+ int sample_rate,
+ size_t number_of_channels,
+ size_t number_of_frames) {
+ // TODO(tommi): We should get the timestamp from WebRTC.
+ base::TimeTicks playout_time(base::TimeTicks::Now());
o1ka 2016/04/01 15:11:41 Probably have paranoid thread checks in all those
miu 2016/04/19 00:40:22 Done. But, it's not that they should be coming fr
o1ka 2016/04/21 18:51:22 Acknowledged.
- if (!sink_->IsNeeded()) {
- track_->RemoveSink(sink_.get());
- sink_.reset();
+ if (!audio_bus_ ||
+ static_cast<size_t>(audio_bus_->channels()) != number_of_channels ||
+ static_cast<size_t>(audio_bus_->frames()) != number_of_frames) {
+ audio_bus_ = media::AudioBus::Create(number_of_channels, number_of_frames);
}
-}
-void MediaStreamRemoteAudioSource::SetSinksEnabled(MediaStreamAudioTrack* track,
- bool enabled) {
- if (sink_)
- sink_->SetEnabled(track, enabled);
-}
-
-void MediaStreamRemoteAudioSource::RemoveAll(MediaStreamAudioTrack* track) {
- if (sink_)
- sink_->RemoveAll(track);
-}
+ audio_bus_->FromInterleaved(audio_data, number_of_frames,
+ bits_per_sample / 8);
+
+ media::AudioParameters params = MediaStreamAudioSource::GetAudioParameters();
+ if (!params.IsValid() ||
+ params.format() != media::AudioParameters::AUDIO_PCM_LOW_LATENCY ||
+ static_cast<size_t>(params.channels()) != number_of_channels ||
+ params.sample_rate() != sample_rate ||
+ static_cast<size_t>(params.frames_per_buffer()) != number_of_frames) {
+ MediaStreamAudioSource::SetFormat(
+ media::AudioParameters(media::AudioParameters::AUDIO_PCM_LOW_LATENCY,
+ media::GuessChannelLayout(number_of_channels),
+ sample_rate, bits_per_sample, number_of_frames));
+ }
-webrtc::AudioTrackInterface* MediaStreamRemoteAudioSource::GetAudioAdapter() {
- DCHECK(thread_checker_.CalledOnValidThread());
- return track_.get();
+ MediaStreamAudioSource::DeliverDataToTracks(*audio_bus_, playout_time);
}
} // namespace content

Powered by Google App Engine
This is Rietveld 408576698