Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(1059)

Unified Diff: content/renderer/media/media_stream_audio_processor.h

Issue 420603004: Use the AudioProcessing float interface in MediaStreamAudioProcessor. (Closed) Base URL: https://chromium.googlesource.com/chromium/src.git@master
Patch Set: OnDataCallback can be called more than once. Created 6 years, 4 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
Index: content/renderer/media/media_stream_audio_processor.h
diff --git a/content/renderer/media/media_stream_audio_processor.h b/content/renderer/media/media_stream_audio_processor.h
index 8211fccf88210c57c2fe0c70e64e4cd4b650173f..3c44f808aedc0d869b75528aec734cb52485aead 100644
--- a/content/renderer/media/media_stream_audio_processor.h
+++ b/content/renderer/media/media_stream_audio_processor.h
@@ -35,6 +35,8 @@ class TypingDetection;
namespace content {
+class MediaStreamAudioBus;
+class MediaStreamAudioFifo;
class RTCMediaConstraints;
using webrtc::AudioProcessorInterface;
@@ -59,30 +61,32 @@ class CONTENT_EXPORT MediaStreamAudioProcessor :
int effects,
WebRtcPlayoutDataSource* playout_data_source);
- // Called when format of the capture data has changed.
- // Called on the main render thread. The caller is responsible for stopping
+ // Called when the format of the capture data has changed.
+ // Called on the main render thread. The caller is responsible for stopping
// the capture thread before calling this method.
// After this method, the capture thread will be changed to a new capture
// thread.
void OnCaptureFormatChanged(const media::AudioParameters& source_params);
- // Pushes capture data in |audio_source| to the internal FIFO.
+ // Pushes capture data in |audio_source| to the internal FIFO. Each call to
+ // this method should be followed by calls to ProcessAndConsumeData() while
+ // it returns false, to pull out all available data.
// Called on the capture audio thread.
void PushCaptureData(const media::AudioBus* audio_source);
// Processes a block of 10 ms data from the internal FIFO and outputs it via
// |out|. |out| is the address of the pointer that will be pointed to
// the post-processed data if the method is returning a true. The lifetime
- // of the data represeted by |out| is guaranteed to outlive the method call.
- // That also says *|out| won't change until this method is called again.
+ // of the data represeted by |out| is guaranteed until this method is called
+ // again.
// |new_volume| receives the new microphone volume from the AGC.
- // The new microphoen volume range is [0, 255], and the value will be 0 if
+ // The new microphone volume range is [0, 255], and the value will be 0 if
// the microphone volume should not be adjusted.
// Returns true if the internal FIFO has at least 10 ms data for processing,
// otherwise false.
- // |capture_delay|, |volume| and |key_pressed| will be passed to
- // webrtc::AudioProcessing to help processing the data.
// Called on the capture audio thread.
+ //
+ // TODO(ajm): Don't we want this to output float?
bool ProcessAndConsumeData(base::TimeDelta capture_delay,
int volume,
bool key_pressed,
@@ -93,10 +97,9 @@ class CONTENT_EXPORT MediaStreamAudioProcessor :
// this method.
void Stop();
- // The audio format of the input to the processor.
+ // The audio formats of the capture input to and output from the processor.
+ // Must only be called on the main render or audio capture threads.
const media::AudioParameters& InputFormat() const;
-
- // The audio format of the output from the processor.
const media::AudioParameters& OutputFormat() const;
// Accessor to check if the audio processing is enabled or not.
@@ -118,8 +121,6 @@ class CONTENT_EXPORT MediaStreamAudioProcessor :
FRIEND_TEST_ALL_PREFIXES(MediaStreamAudioProcessorTest,
GetAecDumpMessageFilter);
- class MediaStreamAudioConverter;
-
// WebRtcPlayoutDataSource::Sink implementation.
virtual void OnPlayoutData(media::AudioBus* audio_bus,
int sample_rate,
@@ -135,64 +136,63 @@ class CONTENT_EXPORT MediaStreamAudioProcessor :
const blink::WebMediaConstraints& constraints, int effects);
// Helper to initialize the capture converter.
- void InitializeCaptureConverter(const media::AudioParameters& source_params);
+ void InitializeCaptureFifo(const media::AudioParameters& input_format);
// Helper to initialize the render converter.
- void InitializeRenderConverterIfNeeded(int sample_rate,
- int number_of_channels,
- int frames_per_buffer);
+ void InitializeRenderFifoIfNeeded(int sample_rate,
+ int number_of_channels,
+ int frames_per_buffer);
// Called by ProcessAndConsumeData().
// Returns the new microphone volume in the range of |0, 255].
// When the volume does not need to be updated, it returns 0.
- int ProcessData(webrtc::AudioFrame* audio_frame,
+ int ProcessData(const float* const* process_ptrs,
+ int process_frames,
base::TimeDelta capture_delay,
int volume,
- bool key_pressed);
+ bool key_pressed,
+ float* const* output_ptrs);
// Cached value for the render delay latency. This member is accessed by
// both the capture audio thread and the render audio thread.
base::subtle::Atomic32 render_delay_ms_;
- // webrtc::AudioProcessing module which does AEC, AGC, NS, HighPass filter,
- // ..etc.
+ // Module to handle processing and format conversion.
scoped_ptr<webrtc::AudioProcessing> audio_processing_;
- // Converter used for the down-mixing and resampling of the capture data.
- scoped_ptr<MediaStreamAudioConverter> capture_converter_;
-
- // AudioFrame used to hold the output of |capture_converter_|.
- webrtc::AudioFrame capture_frame_;
-
- // Converter used for the down-mixing and resampling of the render data when
- // the AEC is enabled.
- scoped_ptr<MediaStreamAudioConverter> render_converter_;
+ // FIFO to provide 10 ms capture chunks.
+ scoped_ptr<MediaStreamAudioFifo> capture_fifo_;
+ // Receives processing output.
+ scoped_ptr<MediaStreamAudioBus> output_bus_;
+ // Receives interleaved int16 data for output.
+ scoped_ptr<int16[]> output_data_;
- // AudioFrame used to hold the output of |render_converter_|.
- webrtc::AudioFrame render_frame_;
+ // FIFO to provide 10 ms render chunks when the AEC is enabled.
+ scoped_ptr<MediaStreamAudioFifo> render_fifo_;
- // Data bus to help converting interleaved data to an AudioBus.
- scoped_ptr<media::AudioBus> render_data_bus_;
+ // These are mutated on the main render thread in OnCaptureFormatChanged().
+ // The caller guarantees this does not run concurrently with accesses on the
+ // capture audio thread.
+ media::AudioParameters input_format_;
+ media::AudioParameters output_format_;
+ // Only used on the render audio thread.
+ media::AudioParameters render_format_;
// Raw pointer to the WebRtcPlayoutDataSource, which is valid for the
// lifetime of RenderThread.
WebRtcPlayoutDataSource* playout_data_source_;
- // Used to DCHECK that the destructor is called on the main render thread.
+ // Used to DCHECK that some methods are called on the main render thread.
base::ThreadChecker main_thread_checker_;
-
// Used to DCHECK that some methods are called on the capture audio thread.
base::ThreadChecker capture_thread_checker_;
-
- // Used to DCHECK that PushRenderData() is called on the render audio thread.
+ // Used to DCHECK that some methods are called on the render audio thread.
base::ThreadChecker render_thread_checker_;
- // Flag to enable the stereo channels mirroring.
+ // Flag to enable stereo channel mirroring.
bool audio_mirroring_;
- // Used by the typing detection.
scoped_ptr<webrtc::TypingDetection> typing_detector_;
-
// This flag is used to show the result of typing detection.
// It can be accessed by the capture audio thread and by the libjingle thread
// which calls GetStats().

Powered by Google App Engine
This is Rietveld 408576698