| Index: content/renderer/media/media_stream_audio_processor.h
|
| diff --git a/content/renderer/media/media_stream_audio_processor.h b/content/renderer/media/media_stream_audio_processor.h
|
| index 8211fccf88210c57c2fe0c70e64e4cd4b650173f..3c44f808aedc0d869b75528aec734cb52485aead 100644
|
| --- a/content/renderer/media/media_stream_audio_processor.h
|
| +++ b/content/renderer/media/media_stream_audio_processor.h
|
| @@ -35,6 +35,8 @@ class TypingDetection;
|
|
|
| namespace content {
|
|
|
| +class MediaStreamAudioBus;
|
| +class MediaStreamAudioFifo;
|
| class RTCMediaConstraints;
|
|
|
| using webrtc::AudioProcessorInterface;
|
| @@ -59,30 +61,32 @@ class CONTENT_EXPORT MediaStreamAudioProcessor :
|
| int effects,
|
| WebRtcPlayoutDataSource* playout_data_source);
|
|
|
| - // Called when format of the capture data has changed.
|
| - // Called on the main render thread. The caller is responsible for stopping
|
| + // Called when the format of the capture data has changed.
|
| + // Called on the main render thread. The caller is responsible for stopping
|
| // the capture thread before calling this method.
|
| // After this method, the capture thread will be changed to a new capture
|
| // thread.
|
| void OnCaptureFormatChanged(const media::AudioParameters& source_params);
|
|
|
| - // Pushes capture data in |audio_source| to the internal FIFO.
|
| + // Pushes capture data in |audio_source| to the internal FIFO. Each call to
|
| + // this method should be followed by calls to ProcessAndConsumeData() while
|
| + // it returns false, to pull out all available data.
|
| // Called on the capture audio thread.
|
| void PushCaptureData(const media::AudioBus* audio_source);
|
|
|
| // Processes a block of 10 ms data from the internal FIFO and outputs it via
|
| // |out|. |out| is the address of the pointer that will be pointed to
|
| // the post-processed data if the method is returning a true. The lifetime
|
| - // of the data represeted by |out| is guaranteed to outlive the method call.
|
| - // That also says *|out| won't change until this method is called again.
|
| + // of the data represeted by |out| is guaranteed until this method is called
|
| + // again.
|
| // |new_volume| receives the new microphone volume from the AGC.
|
| - // The new microphoen volume range is [0, 255], and the value will be 0 if
|
| + // The new microphone volume range is [0, 255], and the value will be 0 if
|
| // the microphone volume should not be adjusted.
|
| // Returns true if the internal FIFO has at least 10 ms data for processing,
|
| // otherwise false.
|
| - // |capture_delay|, |volume| and |key_pressed| will be passed to
|
| - // webrtc::AudioProcessing to help processing the data.
|
| // Called on the capture audio thread.
|
| + //
|
| + // TODO(ajm): Don't we want this to output float?
|
| bool ProcessAndConsumeData(base::TimeDelta capture_delay,
|
| int volume,
|
| bool key_pressed,
|
| @@ -93,10 +97,9 @@ class CONTENT_EXPORT MediaStreamAudioProcessor :
|
| // this method.
|
| void Stop();
|
|
|
| - // The audio format of the input to the processor.
|
| + // The audio formats of the capture input to and output from the processor.
|
| + // Must only be called on the main render or audio capture threads.
|
| const media::AudioParameters& InputFormat() const;
|
| -
|
| - // The audio format of the output from the processor.
|
| const media::AudioParameters& OutputFormat() const;
|
|
|
| // Accessor to check if the audio processing is enabled or not.
|
| @@ -118,8 +121,6 @@ class CONTENT_EXPORT MediaStreamAudioProcessor :
|
| FRIEND_TEST_ALL_PREFIXES(MediaStreamAudioProcessorTest,
|
| GetAecDumpMessageFilter);
|
|
|
| - class MediaStreamAudioConverter;
|
| -
|
| // WebRtcPlayoutDataSource::Sink implementation.
|
| virtual void OnPlayoutData(media::AudioBus* audio_bus,
|
| int sample_rate,
|
| @@ -135,64 +136,63 @@ class CONTENT_EXPORT MediaStreamAudioProcessor :
|
| const blink::WebMediaConstraints& constraints, int effects);
|
|
|
| // Helper to initialize the capture converter.
|
| - void InitializeCaptureConverter(const media::AudioParameters& source_params);
|
| + void InitializeCaptureFifo(const media::AudioParameters& input_format);
|
|
|
| // Helper to initialize the render converter.
|
| - void InitializeRenderConverterIfNeeded(int sample_rate,
|
| - int number_of_channels,
|
| - int frames_per_buffer);
|
| + void InitializeRenderFifoIfNeeded(int sample_rate,
|
| + int number_of_channels,
|
| + int frames_per_buffer);
|
|
|
| // Called by ProcessAndConsumeData().
|
| // Returns the new microphone volume in the range of |0, 255].
|
| // When the volume does not need to be updated, it returns 0.
|
| - int ProcessData(webrtc::AudioFrame* audio_frame,
|
| + int ProcessData(const float* const* process_ptrs,
|
| + int process_frames,
|
| base::TimeDelta capture_delay,
|
| int volume,
|
| - bool key_pressed);
|
| + bool key_pressed,
|
| + float* const* output_ptrs);
|
|
|
| // Cached value for the render delay latency. This member is accessed by
|
| // both the capture audio thread and the render audio thread.
|
| base::subtle::Atomic32 render_delay_ms_;
|
|
|
| - // webrtc::AudioProcessing module which does AEC, AGC, NS, HighPass filter,
|
| - // ..etc.
|
| + // Module to handle processing and format conversion.
|
| scoped_ptr<webrtc::AudioProcessing> audio_processing_;
|
|
|
| - // Converter used for the down-mixing and resampling of the capture data.
|
| - scoped_ptr<MediaStreamAudioConverter> capture_converter_;
|
| -
|
| - // AudioFrame used to hold the output of |capture_converter_|.
|
| - webrtc::AudioFrame capture_frame_;
|
| -
|
| - // Converter used for the down-mixing and resampling of the render data when
|
| - // the AEC is enabled.
|
| - scoped_ptr<MediaStreamAudioConverter> render_converter_;
|
| + // FIFO to provide 10 ms capture chunks.
|
| + scoped_ptr<MediaStreamAudioFifo> capture_fifo_;
|
| + // Receives processing output.
|
| + scoped_ptr<MediaStreamAudioBus> output_bus_;
|
| + // Receives interleaved int16 data for output.
|
| + scoped_ptr<int16[]> output_data_;
|
|
|
| - // AudioFrame used to hold the output of |render_converter_|.
|
| - webrtc::AudioFrame render_frame_;
|
| + // FIFO to provide 10 ms render chunks when the AEC is enabled.
|
| + scoped_ptr<MediaStreamAudioFifo> render_fifo_;
|
|
|
| - // Data bus to help converting interleaved data to an AudioBus.
|
| - scoped_ptr<media::AudioBus> render_data_bus_;
|
| + // These are mutated on the main render thread in OnCaptureFormatChanged().
|
| + // The caller guarantees this does not run concurrently with accesses on the
|
| + // capture audio thread.
|
| + media::AudioParameters input_format_;
|
| + media::AudioParameters output_format_;
|
| + // Only used on the render audio thread.
|
| + media::AudioParameters render_format_;
|
|
|
| // Raw pointer to the WebRtcPlayoutDataSource, which is valid for the
|
| // lifetime of RenderThread.
|
| WebRtcPlayoutDataSource* playout_data_source_;
|
|
|
| - // Used to DCHECK that the destructor is called on the main render thread.
|
| + // Used to DCHECK that some methods are called on the main render thread.
|
| base::ThreadChecker main_thread_checker_;
|
| -
|
| // Used to DCHECK that some methods are called on the capture audio thread.
|
| base::ThreadChecker capture_thread_checker_;
|
| -
|
| - // Used to DCHECK that PushRenderData() is called on the render audio thread.
|
| + // Used to DCHECK that some methods are called on the render audio thread.
|
| base::ThreadChecker render_thread_checker_;
|
|
|
| - // Flag to enable the stereo channels mirroring.
|
| + // Flag to enable stereo channel mirroring.
|
| bool audio_mirroring_;
|
|
|
| - // Used by the typing detection.
|
| scoped_ptr<webrtc::TypingDetection> typing_detector_;
|
| -
|
| // This flag is used to show the result of typing detection.
|
| // It can be accessed by the capture audio thread and by the libjingle thread
|
| // which calls GetStats().
|
|
|