Index: content/renderer/media/media_stream_audio_processor.h |
diff --git a/content/renderer/media/media_stream_audio_processor.h b/content/renderer/media/media_stream_audio_processor.h |
index 8211fccf88210c57c2fe0c70e64e4cd4b650173f..95512faf1642c0481a8d7ff1719cd233794566f5 100644 |
--- a/content/renderer/media/media_stream_audio_processor.h |
+++ b/content/renderer/media/media_stream_audio_processor.h |
@@ -35,6 +35,7 @@ class TypingDetection; |
namespace content { |
+class MediaStreamAudioFifo; |
class RTCMediaConstraints; |
using webrtc::AudioProcessorInterface; |
@@ -59,30 +60,32 @@ class CONTENT_EXPORT MediaStreamAudioProcessor : |
int effects, |
WebRtcPlayoutDataSource* playout_data_source); |
- // Called when format of the capture data has changed. |
- // Called on the main render thread. The caller is responsible for stopping |
+ // Called when the format of the capture data has changed. |
+ // Called on the main render thread. The caller is responsible for stopping |
// the capture thread before calling this method. |
// After this method, the capture thread will be changed to a new capture |
// thread. |
void OnCaptureFormatChanged(const media::AudioParameters& source_params); |
- // Pushes capture data in |audio_source| to the internal FIFO. |
+ // Pushes capture data in |audio_source| to the internal FIFO. Each call to |
+ // this method should be followed by calls to ProcessAndConsumeData() while |
+ // it returns false, to pull out all available data. |
// Called on the capture audio thread. |
void PushCaptureData(const media::AudioBus* audio_source); |
// Processes a block of 10 ms data from the internal FIFO and outputs it via |
// |out|. |out| is the address of the pointer that will be pointed to |
// the post-processed data if the method is returning a true. The lifetime |
- // of the data represeted by |out| is guaranteed to outlive the method call. |
- // That also says *|out| won't change until this method is called again. |
+ // of the data represeted by |out| is guaranteed until this method is called |
+ // again. |
// |new_volume| receives the new microphone volume from the AGC. |
- // The new microphoen volume range is [0, 255], and the value will be 0 if |
+ // The new microphone volume range is [0, 255], and the value will be 0 if |
// the microphone volume should not be adjusted. |
// Returns true if the internal FIFO has at least 10 ms data for processing, |
// otherwise false. |
- // |capture_delay|, |volume| and |key_pressed| will be passed to |
- // webrtc::AudioProcessing to help processing the data. |
// Called on the capture audio thread. |
+ // |
+ // TODO(ajm): Don't we want this to output float? |
bool ProcessAndConsumeData(base::TimeDelta capture_delay, |
int volume, |
bool key_pressed, |
@@ -118,8 +121,6 @@ class CONTENT_EXPORT MediaStreamAudioProcessor : |
FRIEND_TEST_ALL_PREFIXES(MediaStreamAudioProcessorTest, |
GetAecDumpMessageFilter); |
- class MediaStreamAudioConverter; |
- |
// WebRtcPlayoutDataSource::Sink implementation. |
virtual void OnPlayoutData(media::AudioBus* audio_bus, |
int sample_rate, |
@@ -135,44 +136,41 @@ class CONTENT_EXPORT MediaStreamAudioProcessor : |
const blink::WebMediaConstraints& constraints, int effects); |
// Helper to initialize the capture converter. |
- void InitializeCaptureConverter(const media::AudioParameters& source_params); |
+ void InitializeCaptureFifo(const media::AudioParameters& input_format); |
// Helper to initialize the render converter. |
- void InitializeRenderConverterIfNeeded(int sample_rate, |
- int number_of_channels, |
- int frames_per_buffer); |
+ void InitializeRenderFifoIfNeeded(int sample_rate, |
+ int number_of_channels, |
+ int frames_per_buffer); |
// Called by ProcessAndConsumeData(). |
// Returns the new microphone volume in the range of |0, 255]. |
// When the volume does not need to be updated, it returns 0. |
- int ProcessData(webrtc::AudioFrame* audio_frame, |
+ int ProcessData(const media::AudioBus* input, |
base::TimeDelta capture_delay, |
int volume, |
- bool key_pressed); |
+ bool key_pressed, |
+ media::AudioBus* output); |
// Cached value for the render delay latency. This member is accessed by |
// both the capture audio thread and the render audio thread. |
base::subtle::Atomic32 render_delay_ms_; |
- // webrtc::AudioProcessing module which does AEC, AGC, NS, HighPass filter, |
- // ..etc. |
+ // Module to handle processing and format conversion. |
scoped_ptr<webrtc::AudioProcessing> audio_processing_; |
- // Converter used for the down-mixing and resampling of the capture data. |
- scoped_ptr<MediaStreamAudioConverter> capture_converter_; |
- |
- // AudioFrame used to hold the output of |capture_converter_|. |
- webrtc::AudioFrame capture_frame_; |
- |
- // Converter used for the down-mixing and resampling of the render data when |
- // the AEC is enabled. |
- scoped_ptr<MediaStreamAudioConverter> render_converter_; |
+ // FIFO to provide 10 ms capture chunks. |
+ scoped_ptr<MediaStreamAudioFifo> capture_fifo_; |
+ // Receives processing output. |
+ scoped_ptr<media::AudioBus> output_bus_; |
+ // Receives interleaved int16 data for output. |
+ scoped_ptr<int16[]> output_data_; |
- // AudioFrame used to hold the output of |render_converter_|. |
- webrtc::AudioFrame render_frame_; |
+ // FIFO to provide 10 ms render chunks when the AEC is enabled. |
+ scoped_ptr<MediaStreamAudioFifo> render_fifo_; |
- // Data bus to help converting interleaved data to an AudioBus. |
- scoped_ptr<media::AudioBus> render_data_bus_; |
+ media::AudioParameters input_format_; |
+ media::AudioParameters output_format_; |
// Raw pointer to the WebRtcPlayoutDataSource, which is valid for the |
// lifetime of RenderThread. |
@@ -180,19 +178,15 @@ class CONTENT_EXPORT MediaStreamAudioProcessor : |
// Used to DCHECK that the destructor is called on the main render thread. |
base::ThreadChecker main_thread_checker_; |
- |
// Used to DCHECK that some methods are called on the capture audio thread. |
base::ThreadChecker capture_thread_checker_; |
- |
// Used to DCHECK that PushRenderData() is called on the render audio thread. |
base::ThreadChecker render_thread_checker_; |
- // Flag to enable the stereo channels mirroring. |
+ // Flag to enable stereo channel mirroring. |
bool audio_mirroring_; |
- // Used by the typing detection. |
scoped_ptr<webrtc::TypingDetection> typing_detector_; |
- |
// This flag is used to show the result of typing detection. |
// It can be accessed by the capture audio thread and by the libjingle thread |
// which calls GetStats(). |