Index: content/renderer/media/media_stream_dependency_factory.cc |
diff --git a/content/renderer/media/media_stream_dependency_factory.cc b/content/renderer/media/media_stream_dependency_factory.cc |
deleted file mode 100644 |
index 9cd339adfcaee0530f625bfa83cb819807f202d6..0000000000000000000000000000000000000000 |
--- a/content/renderer/media/media_stream_dependency_factory.cc |
+++ /dev/null |
@@ -1,669 +0,0 @@ |
-// Copyright (c) 2012 The Chromium Authors. All rights reserved. |
-// Use of this source code is governed by a BSD-style license that can be |
-// found in the LICENSE file. |
- |
-#include "content/renderer/media/media_stream_dependency_factory.h" |
- |
-#include <vector> |
- |
-#include "base/command_line.h" |
-#include "base/strings/utf_string_conversions.h" |
-#include "base/synchronization/waitable_event.h" |
-#include "content/common/media/media_stream_messages.h" |
-#include "content/public/common/content_switches.h" |
-#include "content/renderer/media/media_stream.h" |
-#include "content/renderer/media/media_stream_audio_processor_options.h" |
-#include "content/renderer/media/media_stream_audio_source.h" |
-#include "content/renderer/media/media_stream_video_source.h" |
-#include "content/renderer/media/media_stream_video_track.h" |
-#include "content/renderer/media/peer_connection_identity_service.h" |
-#include "content/renderer/media/rtc_media_constraints.h" |
-#include "content/renderer/media/rtc_peer_connection_handler.h" |
-#include "content/renderer/media/rtc_video_decoder_factory.h" |
-#include "content/renderer/media/rtc_video_encoder_factory.h" |
-#include "content/renderer/media/webaudio_capturer_source.h" |
-#include "content/renderer/media/webrtc/webrtc_local_audio_track_adapter.h" |
-#include "content/renderer/media/webrtc/webrtc_video_capturer_adapter.h" |
-#include "content/renderer/media/webrtc_audio_device_impl.h" |
-#include "content/renderer/media/webrtc_local_audio_track.h" |
-#include "content/renderer/media/webrtc_uma_histograms.h" |
-#include "content/renderer/p2p/ipc_network_manager.h" |
-#include "content/renderer/p2p/ipc_socket_factory.h" |
-#include "content/renderer/p2p/port_allocator.h" |
-#include "content/renderer/render_thread_impl.h" |
-#include "jingle/glue/thread_wrapper.h" |
-#include "media/filters/gpu_video_accelerator_factories.h" |
-#include "third_party/WebKit/public/platform/WebMediaConstraints.h" |
-#include "third_party/WebKit/public/platform/WebMediaStream.h" |
-#include "third_party/WebKit/public/platform/WebMediaStreamSource.h" |
-#include "third_party/WebKit/public/platform/WebMediaStreamTrack.h" |
-#include "third_party/WebKit/public/platform/WebURL.h" |
-#include "third_party/WebKit/public/web/WebDocument.h" |
-#include "third_party/WebKit/public/web/WebFrame.h" |
-#include "third_party/libjingle/source/talk/app/webrtc/mediaconstraintsinterface.h" |
- |
-#if defined(USE_OPENSSL) |
-#include "third_party/libjingle/source/talk/base/ssladapter.h" |
-#else |
-#include "net/socket/nss_ssl_util.h" |
-#endif |
- |
-#if defined(OS_ANDROID) |
-#include "media/base/android/media_codec_bridge.h" |
-#endif |
- |
-namespace content { |
- |
-// Map of corresponding media constraints and platform effects. |
-struct { |
- const char* constraint; |
- const media::AudioParameters::PlatformEffectsMask effect; |
-} const kConstraintEffectMap[] = { |
- { content::kMediaStreamAudioDucking, |
- media::AudioParameters::DUCKING }, |
- { webrtc::MediaConstraintsInterface::kEchoCancellation, |
- media::AudioParameters::ECHO_CANCELLER }, |
-}; |
- |
-// If any platform effects are available, check them against the constraints. |
-// Disable effects to match false constraints, but if a constraint is true, set |
-// the constraint to false to later disable the software effect. |
-// |
-// This function may modify both |constraints| and |effects|. |
-void HarmonizeConstraintsAndEffects(RTCMediaConstraints* constraints, |
- int* effects) { |
- if (*effects != media::AudioParameters::NO_EFFECTS) { |
- for (size_t i = 0; i < ARRAYSIZE_UNSAFE(kConstraintEffectMap); ++i) { |
- bool value; |
- size_t is_mandatory = 0; |
- if (!webrtc::FindConstraint(constraints, |
- kConstraintEffectMap[i].constraint, |
- &value, |
- &is_mandatory) || !value) { |
- // If the constraint is false, or does not exist, disable the platform |
- // effect. |
- *effects &= ~kConstraintEffectMap[i].effect; |
- DVLOG(1) << "Disabling platform effect: " |
- << kConstraintEffectMap[i].effect; |
- } else if (*effects & kConstraintEffectMap[i].effect) { |
- // If the constraint is true, leave the platform effect enabled, and |
- // set the constraint to false to later disable the software effect. |
- if (is_mandatory) { |
- constraints->AddMandatory(kConstraintEffectMap[i].constraint, |
- webrtc::MediaConstraintsInterface::kValueFalse, true); |
- } else { |
- constraints->AddOptional(kConstraintEffectMap[i].constraint, |
- webrtc::MediaConstraintsInterface::kValueFalse, true); |
- } |
- DVLOG(1) << "Disabling constraint: " |
- << kConstraintEffectMap[i].constraint; |
- } |
- } |
- } |
-} |
- |
-class P2PPortAllocatorFactory : public webrtc::PortAllocatorFactoryInterface { |
- public: |
- P2PPortAllocatorFactory( |
- P2PSocketDispatcher* socket_dispatcher, |
- talk_base::NetworkManager* network_manager, |
- talk_base::PacketSocketFactory* socket_factory, |
- blink::WebFrame* web_frame) |
- : socket_dispatcher_(socket_dispatcher), |
- network_manager_(network_manager), |
- socket_factory_(socket_factory), |
- web_frame_(web_frame) { |
- } |
- |
- virtual cricket::PortAllocator* CreatePortAllocator( |
- const std::vector<StunConfiguration>& stun_servers, |
- const std::vector<TurnConfiguration>& turn_configurations) OVERRIDE { |
- CHECK(web_frame_); |
- P2PPortAllocator::Config config; |
- if (stun_servers.size() > 0) { |
- config.stun_server = stun_servers[0].server.hostname(); |
- config.stun_server_port = stun_servers[0].server.port(); |
- } |
- config.legacy_relay = false; |
- for (size_t i = 0; i < turn_configurations.size(); ++i) { |
- P2PPortAllocator::Config::RelayServerConfig relay_config; |
- relay_config.server_address = turn_configurations[i].server.hostname(); |
- relay_config.port = turn_configurations[i].server.port(); |
- relay_config.username = turn_configurations[i].username; |
- relay_config.password = turn_configurations[i].password; |
- relay_config.transport_type = turn_configurations[i].transport_type; |
- relay_config.secure = turn_configurations[i].secure; |
- config.relays.push_back(relay_config); |
- } |
- |
- // Use first turn server as the stun server. |
- if (turn_configurations.size() > 0) { |
- config.stun_server = config.relays[0].server_address; |
- config.stun_server_port = config.relays[0].port; |
- } |
- |
- return new P2PPortAllocator( |
- web_frame_, socket_dispatcher_.get(), network_manager_, |
- socket_factory_, config); |
- } |
- |
- protected: |
- virtual ~P2PPortAllocatorFactory() {} |
- |
- private: |
- scoped_refptr<P2PSocketDispatcher> socket_dispatcher_; |
- // |network_manager_| and |socket_factory_| are a weak references, owned by |
- // MediaStreamDependencyFactory. |
- talk_base::NetworkManager* network_manager_; |
- talk_base::PacketSocketFactory* socket_factory_; |
- // Raw ptr to the WebFrame that created the P2PPortAllocatorFactory. |
- blink::WebFrame* web_frame_; |
-}; |
- |
-MediaStreamDependencyFactory::MediaStreamDependencyFactory( |
- P2PSocketDispatcher* p2p_socket_dispatcher) |
- : network_manager_(NULL), |
- p2p_socket_dispatcher_(p2p_socket_dispatcher), |
- signaling_thread_(NULL), |
- worker_thread_(NULL), |
- chrome_worker_thread_("Chrome_libJingle_WorkerThread") { |
-} |
- |
-MediaStreamDependencyFactory::~MediaStreamDependencyFactory() { |
- CleanupPeerConnectionFactory(); |
-} |
- |
-blink::WebRTCPeerConnectionHandler* |
-MediaStreamDependencyFactory::CreateRTCPeerConnectionHandler( |
- blink::WebRTCPeerConnectionHandlerClient* client) { |
- // Save histogram data so we can see how much PeerConnetion is used. |
- // The histogram counts the number of calls to the JS API |
- // webKitRTCPeerConnection. |
- UpdateWebRTCMethodCount(WEBKIT_RTC_PEER_CONNECTION); |
- |
- return new RTCPeerConnectionHandler(client, this); |
-} |
- |
-bool MediaStreamDependencyFactory::InitializeMediaStreamAudioSource( |
- int render_view_id, |
- const blink::WebMediaConstraints& audio_constraints, |
- MediaStreamAudioSource* source_data) { |
- DVLOG(1) << "InitializeMediaStreamAudioSources()"; |
- |
- // Do additional source initialization if the audio source is a valid |
- // microphone or tab audio. |
- RTCMediaConstraints native_audio_constraints(audio_constraints); |
- MediaAudioConstraints::ApplyFixedAudioConstraints(&native_audio_constraints); |
- |
- StreamDeviceInfo device_info = source_data->device_info(); |
- RTCMediaConstraints constraints = native_audio_constraints; |
- // May modify both |constraints| and |effects|. |
- HarmonizeConstraintsAndEffects(&constraints, |
- &device_info.device.input.effects); |
- |
- scoped_refptr<WebRtcAudioCapturer> capturer( |
- CreateAudioCapturer(render_view_id, device_info, audio_constraints, |
- source_data)); |
- if (!capturer.get()) { |
- DLOG(WARNING) << "Failed to create the capturer for device " |
- << device_info.device.id; |
- // TODO(xians): Don't we need to check if source_observer is observing |
- // something? If not, then it looks like we have a leak here. |
- // OTOH, if it _is_ observing something, then the callback might |
- // be called multiple times which is likely also a bug. |
- return false; |
- } |
- source_data->SetAudioCapturer(capturer); |
- |
- // Creates a LocalAudioSource object which holds audio options. |
- // TODO(xians): The option should apply to the track instead of the source. |
- // TODO(perkj): Move audio constraints parsing to Chrome. |
- // Currently there are a few constraints that are parsed by libjingle and |
- // the state is set to ended if parsing fails. |
- scoped_refptr<webrtc::AudioSourceInterface> rtc_source( |
- CreateLocalAudioSource(&constraints).get()); |
- if (rtc_source->state() != webrtc::MediaSourceInterface::kLive) { |
- DLOG(WARNING) << "Failed to create rtc LocalAudioSource."; |
- return false; |
- } |
- source_data->SetLocalAudioSource(rtc_source); |
- return true; |
-} |
- |
-WebRtcVideoCapturerAdapter* MediaStreamDependencyFactory::CreateVideoCapturer( |
- bool is_screeencast) { |
- // We need to make sure the libjingle thread wrappers have been created |
- // before we can use an instance of a WebRtcVideoCapturerAdapter. This is |
- // since the base class of WebRtcVideoCapturerAdapter is a |
- // cricket::VideoCapturer and it uses the libjingle thread wrappers. |
- if (!GetPcFactory()) |
- return NULL; |
- return new WebRtcVideoCapturerAdapter(is_screeencast); |
-} |
- |
-scoped_refptr<webrtc::VideoSourceInterface> |
-MediaStreamDependencyFactory::CreateVideoSource( |
- cricket::VideoCapturer* capturer, |
- const blink::WebMediaConstraints& constraints) { |
- RTCMediaConstraints webrtc_constraints(constraints); |
- scoped_refptr<webrtc::VideoSourceInterface> source = |
- GetPcFactory()->CreateVideoSource(capturer, &webrtc_constraints).get(); |
- return source; |
-} |
- |
-const scoped_refptr<webrtc::PeerConnectionFactoryInterface>& |
-MediaStreamDependencyFactory::GetPcFactory() { |
- if (!pc_factory_) |
- CreatePeerConnectionFactory(); |
- CHECK(pc_factory_); |
- return pc_factory_; |
-} |
- |
-void MediaStreamDependencyFactory::CreatePeerConnectionFactory() { |
- DCHECK(!pc_factory_.get()); |
- DCHECK(!signaling_thread_); |
- DCHECK(!worker_thread_); |
- DCHECK(!network_manager_); |
- DCHECK(!socket_factory_); |
- DCHECK(!chrome_worker_thread_.IsRunning()); |
- |
- DVLOG(1) << "MediaStreamDependencyFactory::CreatePeerConnectionFactory()"; |
- |
- jingle_glue::JingleThreadWrapper::EnsureForCurrentMessageLoop(); |
- jingle_glue::JingleThreadWrapper::current()->set_send_allowed(true); |
- signaling_thread_ = jingle_glue::JingleThreadWrapper::current(); |
- CHECK(signaling_thread_); |
- |
- CHECK(chrome_worker_thread_.Start()); |
- |
- base::WaitableEvent start_worker_event(true, false); |
- chrome_worker_thread_.message_loop()->PostTask(FROM_HERE, base::Bind( |
- &MediaStreamDependencyFactory::InitializeWorkerThread, |
- base::Unretained(this), |
- &worker_thread_, |
- &start_worker_event)); |
- start_worker_event.Wait(); |
- CHECK(worker_thread_); |
- |
- base::WaitableEvent create_network_manager_event(true, false); |
- chrome_worker_thread_.message_loop()->PostTask(FROM_HERE, base::Bind( |
- &MediaStreamDependencyFactory::CreateIpcNetworkManagerOnWorkerThread, |
- base::Unretained(this), |
- &create_network_manager_event)); |
- create_network_manager_event.Wait(); |
- |
- socket_factory_.reset( |
- new IpcPacketSocketFactory(p2p_socket_dispatcher_.get())); |
- |
- // Init SSL, which will be needed by PeerConnection. |
-#if defined(USE_OPENSSL) |
- if (!talk_base::InitializeSSL()) { |
- LOG(ERROR) << "Failed on InitializeSSL."; |
- NOTREACHED(); |
- return; |
- } |
-#else |
- // TODO(ronghuawu): Replace this call with InitializeSSL. |
- net::EnsureNSSSSLInit(); |
-#endif |
- |
- scoped_ptr<cricket::WebRtcVideoDecoderFactory> decoder_factory; |
- scoped_ptr<cricket::WebRtcVideoEncoderFactory> encoder_factory; |
- |
- const CommandLine* cmd_line = CommandLine::ForCurrentProcess(); |
- scoped_refptr<media::GpuVideoAcceleratorFactories> gpu_factories = |
- RenderThreadImpl::current()->GetGpuFactories(); |
- if (!cmd_line->HasSwitch(switches::kDisableWebRtcHWDecoding)) { |
- if (gpu_factories) |
- decoder_factory.reset(new RTCVideoDecoderFactory(gpu_factories)); |
- } |
- |
- if (!cmd_line->HasSwitch(switches::kDisableWebRtcHWEncoding)) { |
- if (gpu_factories) |
- encoder_factory.reset(new RTCVideoEncoderFactory(gpu_factories)); |
- } |
- |
-#if defined(OS_ANDROID) |
- if (!media::MediaCodecBridge::SupportsSetParameters()) |
- encoder_factory.reset(); |
-#endif |
- |
- EnsureWebRtcAudioDeviceImpl(); |
- |
- scoped_refptr<webrtc::PeerConnectionFactoryInterface> factory( |
- webrtc::CreatePeerConnectionFactory(worker_thread_, |
- signaling_thread_, |
- audio_device_.get(), |
- encoder_factory.release(), |
- decoder_factory.release())); |
- CHECK(factory); |
- |
- pc_factory_ = factory; |
- webrtc::PeerConnectionFactoryInterface::Options factory_options; |
- factory_options.disable_sctp_data_channels = false; |
- factory_options.disable_encryption = |
- cmd_line->HasSwitch(switches::kDisableWebRtcEncryption); |
- pc_factory_->SetOptions(factory_options); |
- |
- // |aec_dump_file| will be invalid when dump is not enabled. |
- if (aec_dump_file_.IsValid()) |
- StartAecDump(aec_dump_file_.Pass()); |
-} |
- |
-bool MediaStreamDependencyFactory::PeerConnectionFactoryCreated() { |
- return pc_factory_.get() != NULL; |
-} |
- |
-scoped_refptr<webrtc::PeerConnectionInterface> |
-MediaStreamDependencyFactory::CreatePeerConnection( |
- const webrtc::PeerConnectionInterface::IceServers& ice_servers, |
- const webrtc::MediaConstraintsInterface* constraints, |
- blink::WebFrame* web_frame, |
- webrtc::PeerConnectionObserver* observer) { |
- CHECK(web_frame); |
- CHECK(observer); |
- if (!GetPcFactory()) |
- return NULL; |
- |
- scoped_refptr<P2PPortAllocatorFactory> pa_factory = |
- new talk_base::RefCountedObject<P2PPortAllocatorFactory>( |
- p2p_socket_dispatcher_.get(), |
- network_manager_, |
- socket_factory_.get(), |
- web_frame); |
- |
- PeerConnectionIdentityService* identity_service = |
- new PeerConnectionIdentityService( |
- GURL(web_frame->document().url().spec()).GetOrigin()); |
- |
- return GetPcFactory()->CreatePeerConnection(ice_servers, |
- constraints, |
- pa_factory.get(), |
- identity_service, |
- observer).get(); |
-} |
- |
-scoped_refptr<webrtc::MediaStreamInterface> |
-MediaStreamDependencyFactory::CreateLocalMediaStream( |
- const std::string& label) { |
- return GetPcFactory()->CreateLocalMediaStream(label).get(); |
-} |
- |
-scoped_refptr<webrtc::AudioSourceInterface> |
-MediaStreamDependencyFactory::CreateLocalAudioSource( |
- const webrtc::MediaConstraintsInterface* constraints) { |
- scoped_refptr<webrtc::AudioSourceInterface> source = |
- GetPcFactory()->CreateAudioSource(constraints).get(); |
- return source; |
-} |
- |
-void MediaStreamDependencyFactory::CreateLocalAudioTrack( |
- const blink::WebMediaStreamTrack& track) { |
- blink::WebMediaStreamSource source = track.source(); |
- DCHECK_EQ(source.type(), blink::WebMediaStreamSource::TypeAudio); |
- MediaStreamAudioSource* source_data = |
- static_cast<MediaStreamAudioSource*>(source.extraData()); |
- |
- scoped_refptr<WebAudioCapturerSource> webaudio_source; |
- if (!source_data) { |
- if (source.requiresAudioConsumer()) { |
- // We're adding a WebAudio MediaStream. |
- // Create a specific capturer for each WebAudio consumer. |
- webaudio_source = CreateWebAudioSource(&source); |
- source_data = |
- static_cast<MediaStreamAudioSource*>(source.extraData()); |
- } else { |
- // TODO(perkj): Implement support for sources from |
- // remote MediaStreams. |
- NOTIMPLEMENTED(); |
- return; |
- } |
- } |
- |
- // Creates an adapter to hold all the libjingle objects. |
- scoped_refptr<WebRtcLocalAudioTrackAdapter> adapter( |
- WebRtcLocalAudioTrackAdapter::Create(track.id().utf8(), |
- source_data->local_audio_source())); |
- static_cast<webrtc::AudioTrackInterface*>(adapter.get())->set_enabled( |
- track.isEnabled()); |
- |
- // TODO(xians): Merge |source| to the capturer(). We can't do this today |
- // because only one capturer() is supported while one |source| is created |
- // for each audio track. |
- scoped_ptr<WebRtcLocalAudioTrack> audio_track( |
- new WebRtcLocalAudioTrack(adapter, |
- source_data->GetAudioCapturer(), |
- webaudio_source)); |
- |
- StartLocalAudioTrack(audio_track.get()); |
- |
- // Pass the ownership of the native local audio track to the blink track. |
- blink::WebMediaStreamTrack writable_track = track; |
- writable_track.setExtraData(audio_track.release()); |
-} |
- |
-void MediaStreamDependencyFactory::StartLocalAudioTrack( |
- WebRtcLocalAudioTrack* audio_track) { |
- // Add the WebRtcAudioDevice as the sink to the local audio track. |
- // TODO(xians): Implement a PeerConnection sink adapter and remove this |
- // AddSink() call. |
- audio_track->AddSink(GetWebRtcAudioDevice()); |
- // Start the audio track. This will hook the |audio_track| to the capturer |
- // as the sink of the audio, and only start the source of the capturer if |
- // it is the first audio track connecting to the capturer. |
- audio_track->Start(); |
-} |
- |
-scoped_refptr<WebAudioCapturerSource> |
-MediaStreamDependencyFactory::CreateWebAudioSource( |
- blink::WebMediaStreamSource* source) { |
- DVLOG(1) << "MediaStreamDependencyFactory::CreateWebAudioSource()"; |
- |
- scoped_refptr<WebAudioCapturerSource> |
- webaudio_capturer_source(new WebAudioCapturerSource()); |
- MediaStreamAudioSource* source_data = new MediaStreamAudioSource(); |
- |
- // Use the current default capturer for the WebAudio track so that the |
- // WebAudio track can pass a valid delay value and |need_audio_processing| |
- // flag to PeerConnection. |
- // TODO(xians): Remove this after moving APM to Chrome. |
- if (GetWebRtcAudioDevice()) { |
- source_data->SetAudioCapturer( |
- GetWebRtcAudioDevice()->GetDefaultCapturer()); |
- } |
- |
- // Create a LocalAudioSource object which holds audio options. |
- // SetLocalAudioSource() affects core audio parts in third_party/Libjingle. |
- source_data->SetLocalAudioSource(CreateLocalAudioSource(NULL).get()); |
- source->setExtraData(source_data); |
- |
- // Replace the default source with WebAudio as source instead. |
- source->addAudioConsumer(webaudio_capturer_source.get()); |
- |
- return webaudio_capturer_source; |
-} |
- |
-scoped_refptr<webrtc::VideoTrackInterface> |
-MediaStreamDependencyFactory::CreateLocalVideoTrack( |
- const std::string& id, |
- webrtc::VideoSourceInterface* source) { |
- return GetPcFactory()->CreateVideoTrack(id, source).get(); |
-} |
- |
-scoped_refptr<webrtc::VideoTrackInterface> |
-MediaStreamDependencyFactory::CreateLocalVideoTrack( |
- const std::string& id, cricket::VideoCapturer* capturer) { |
- if (!capturer) { |
- LOG(ERROR) << "CreateLocalVideoTrack called with null VideoCapturer."; |
- return NULL; |
- } |
- |
- // Create video source from the |capturer|. |
- scoped_refptr<webrtc::VideoSourceInterface> source = |
- GetPcFactory()->CreateVideoSource(capturer, NULL).get(); |
- |
- // Create native track from the source. |
- return GetPcFactory()->CreateVideoTrack(id, source.get()).get(); |
-} |
- |
-webrtc::SessionDescriptionInterface* |
-MediaStreamDependencyFactory::CreateSessionDescription( |
- const std::string& type, |
- const std::string& sdp, |
- webrtc::SdpParseError* error) { |
- return webrtc::CreateSessionDescription(type, sdp, error); |
-} |
- |
-webrtc::IceCandidateInterface* MediaStreamDependencyFactory::CreateIceCandidate( |
- const std::string& sdp_mid, |
- int sdp_mline_index, |
- const std::string& sdp) { |
- return webrtc::CreateIceCandidate(sdp_mid, sdp_mline_index, sdp); |
-} |
- |
-WebRtcAudioDeviceImpl* |
-MediaStreamDependencyFactory::GetWebRtcAudioDevice() { |
- return audio_device_.get(); |
-} |
- |
-void MediaStreamDependencyFactory::InitializeWorkerThread( |
- talk_base::Thread** thread, |
- base::WaitableEvent* event) { |
- jingle_glue::JingleThreadWrapper::EnsureForCurrentMessageLoop(); |
- jingle_glue::JingleThreadWrapper::current()->set_send_allowed(true); |
- *thread = jingle_glue::JingleThreadWrapper::current(); |
- event->Signal(); |
-} |
- |
-void MediaStreamDependencyFactory::CreateIpcNetworkManagerOnWorkerThread( |
- base::WaitableEvent* event) { |
- DCHECK_EQ(base::MessageLoop::current(), chrome_worker_thread_.message_loop()); |
- network_manager_ = new IpcNetworkManager(p2p_socket_dispatcher_.get()); |
- event->Signal(); |
-} |
- |
-void MediaStreamDependencyFactory::DeleteIpcNetworkManager() { |
- DCHECK_EQ(base::MessageLoop::current(), chrome_worker_thread_.message_loop()); |
- delete network_manager_; |
- network_manager_ = NULL; |
-} |
- |
-void MediaStreamDependencyFactory::CleanupPeerConnectionFactory() { |
- pc_factory_ = NULL; |
- if (network_manager_) { |
- // The network manager needs to free its resources on the thread they were |
- // created, which is the worked thread. |
- if (chrome_worker_thread_.IsRunning()) { |
- chrome_worker_thread_.message_loop()->PostTask(FROM_HERE, base::Bind( |
- &MediaStreamDependencyFactory::DeleteIpcNetworkManager, |
- base::Unretained(this))); |
- // Stopping the thread will wait until all tasks have been |
- // processed before returning. We wait for the above task to finish before |
- // letting the the function continue to avoid any potential race issues. |
- chrome_worker_thread_.Stop(); |
- } else { |
- NOTREACHED() << "Worker thread not running."; |
- } |
- } |
-} |
- |
-scoped_refptr<WebRtcAudioCapturer> |
-MediaStreamDependencyFactory::CreateAudioCapturer( |
- int render_view_id, |
- const StreamDeviceInfo& device_info, |
- const blink::WebMediaConstraints& constraints, |
- MediaStreamAudioSource* audio_source) { |
- // TODO(xians): Handle the cases when gUM is called without a proper render |
- // view, for example, by an extension. |
- DCHECK_GE(render_view_id, 0); |
- |
- EnsureWebRtcAudioDeviceImpl(); |
- DCHECK(GetWebRtcAudioDevice()); |
- return WebRtcAudioCapturer::CreateCapturer(render_view_id, device_info, |
- constraints, |
- GetWebRtcAudioDevice(), |
- audio_source); |
-} |
- |
-void MediaStreamDependencyFactory::AddNativeAudioTrackToBlinkTrack( |
- webrtc::MediaStreamTrackInterface* native_track, |
- const blink::WebMediaStreamTrack& webkit_track, |
- bool is_local_track) { |
- DCHECK(!webkit_track.isNull() && !webkit_track.extraData()); |
- DCHECK_EQ(blink::WebMediaStreamSource::TypeAudio, |
- webkit_track.source().type()); |
- blink::WebMediaStreamTrack track = webkit_track; |
- |
- DVLOG(1) << "AddNativeTrackToBlinkTrack() audio"; |
- track.setExtraData( |
- new MediaStreamTrack( |
- static_cast<webrtc::AudioTrackInterface*>(native_track), |
- is_local_track)); |
-} |
- |
-scoped_refptr<base::MessageLoopProxy> |
-MediaStreamDependencyFactory::GetWebRtcWorkerThread() const { |
- DCHECK(CalledOnValidThread()); |
- return chrome_worker_thread_.message_loop_proxy(); |
-} |
- |
-bool MediaStreamDependencyFactory::OnControlMessageReceived( |
- const IPC::Message& message) { |
- bool handled = true; |
- IPC_BEGIN_MESSAGE_MAP(MediaStreamDependencyFactory, message) |
- IPC_MESSAGE_HANDLER(MediaStreamMsg_EnableAecDump, OnAecDumpFile) |
- IPC_MESSAGE_HANDLER(MediaStreamMsg_DisableAecDump, OnDisableAecDump) |
- IPC_MESSAGE_UNHANDLED(handled = false) |
- IPC_END_MESSAGE_MAP() |
- return handled; |
-} |
- |
-void MediaStreamDependencyFactory::OnAecDumpFile( |
- IPC::PlatformFileForTransit file_handle) { |
- DCHECK(!aec_dump_file_.IsValid()); |
- base::File file = IPC::PlatformFileForTransitToFile(file_handle); |
- DCHECK(file.IsValid()); |
- |
- if (CommandLine::ForCurrentProcess()->HasSwitch( |
- switches::kEnableAudioTrackProcessing)) { |
- EnsureWebRtcAudioDeviceImpl(); |
- GetWebRtcAudioDevice()->EnableAecDump(file.Pass()); |
- return; |
- } |
- |
- // TODO(xians): Remove the following code after kEnableAudioTrackProcessing |
- // is removed. |
- if (PeerConnectionFactoryCreated()) |
- StartAecDump(file.Pass()); |
- else |
- aec_dump_file_ = file.Pass(); |
-} |
- |
-void MediaStreamDependencyFactory::OnDisableAecDump() { |
- if (CommandLine::ForCurrentProcess()->HasSwitch( |
- switches::kEnableAudioTrackProcessing)) { |
- GetWebRtcAudioDevice()->DisableAecDump(); |
- return; |
- } |
- |
- // TODO(xians): Remove the following code after kEnableAudioTrackProcessing |
- // is removed. |
- if (aec_dump_file_.IsValid()) |
- aec_dump_file_.Close(); |
-} |
- |
-void MediaStreamDependencyFactory::StartAecDump(base::File aec_dump_file) { |
- // |pc_factory_| always takes ownership of |aec_dump_file|. If StartAecDump() |
- // fails, |aec_dump_file| will be closed. |
- if (!GetPcFactory()->StartAecDump(aec_dump_file.TakePlatformFile())) |
- VLOG(1) << "Could not start AEC dump."; |
-} |
- |
-void MediaStreamDependencyFactory::EnsureWebRtcAudioDeviceImpl() { |
- if (audio_device_) |
- return; |
- |
- audio_device_ = new WebRtcAudioDeviceImpl(); |
-} |
- |
-} // namespace content |