| Index: content/browser/speech/speech_recognizer_impl.cc
|
| diff --git a/content/browser/speech/speech_recognizer_impl.cc b/content/browser/speech/speech_recognizer_impl.cc
|
| index b38963dc91e90e7be68b8493f2bfc0884ce2e277..4a96cef50072eb9be5405643a9668df1bc88b55b 100644
|
| --- a/content/browser/speech/speech_recognizer_impl.cc
|
| +++ b/content/browser/speech/speech_recognizer_impl.cc
|
| @@ -239,7 +239,7 @@ bool SpeechRecognizerImpl::IsActive() const {
|
| }
|
|
|
| bool SpeechRecognizerImpl::IsCapturingAudio() const {
|
| - DCHECK_CURRENTLY_ON(BrowserThread::IO); // See IsActive().
|
| + DCHECK_CURRENTLY_ON(BrowserThread::IO); // See IsActive().
|
| const bool is_capturing_audio = state_ >= STATE_STARTING &&
|
| state_ <= STATE_RECOGNIZING;
|
| DCHECK((is_capturing_audio && (audio_controller_.get() != NULL)) ||
|
| @@ -489,7 +489,7 @@ void SpeechRecognizerImpl::ProcessAudioPipeline(const AudioChunk& raw_audio) {
|
| endpointer_.ProcessAudio(raw_audio, &rms);
|
|
|
| if (route_to_vumeter) {
|
| - DCHECK(route_to_endpointer); // Depends on endpointer due to |rms|.
|
| + DCHECK(route_to_endpointer); // Depends on endpointer due to |rms|.
|
| UpdateSignalAndNoiseLevels(rms, clip_detected);
|
| }
|
| if (route_to_sr_engine) {
|
| @@ -563,10 +563,10 @@ SpeechRecognizerImpl::StartRecording(const FSMEventArgs&) {
|
| ((in_params.sample_rate() * chunk_duration_ms) / 1000.0) + 0.5;
|
| input_parameters.Reset(in_params.format(),
|
| in_params.channel_layout(),
|
| - in_params.channels(),
|
| in_params.sample_rate(),
|
| in_params.bits_per_sample(),
|
| frames_per_buffer);
|
| + input_parameters.set_channels_for_discrete(in_params.channels());
|
| DVLOG(1) << "SRI::input_parameters: "
|
| << input_parameters.AsHumanReadableString();
|
| }
|
|
|