| Index: content/browser/speech/speech_recognizer_impl.cc
|
| diff --git a/content/browser/speech/speech_recognizer_impl.cc b/content/browser/speech/speech_recognizer_impl.cc
|
| index b38963dc91e90e7be68b8493f2bfc0884ce2e277..238317d7cc8cda6b0e7527c609afbd99d205a666 100644
|
| --- a/content/browser/speech/speech_recognizer_impl.cc
|
| +++ b/content/browser/speech/speech_recognizer_impl.cc
|
| @@ -239,7 +239,7 @@ bool SpeechRecognizerImpl::IsActive() const {
|
| }
|
|
|
| bool SpeechRecognizerImpl::IsCapturingAudio() const {
|
| - DCHECK_CURRENTLY_ON(BrowserThread::IO); // See IsActive().
|
| + DCHECK_CURRENTLY_ON(BrowserThread::IO); // See IsActive().
|
| const bool is_capturing_audio = state_ >= STATE_STARTING &&
|
| state_ <= STATE_RECOGNIZING;
|
| DCHECK((is_capturing_audio && (audio_controller_.get() != NULL)) ||
|
| @@ -489,7 +489,7 @@ void SpeechRecognizerImpl::ProcessAudioPipeline(const AudioChunk& raw_audio) {
|
| endpointer_.ProcessAudio(raw_audio, &rms);
|
|
|
| if (route_to_vumeter) {
|
| - DCHECK(route_to_endpointer); // Depends on endpointer due to |rms|.
|
| + DCHECK(route_to_endpointer); // Depends on endpointer due to |rms|.
|
| UpdateSignalAndNoiseLevels(rms, clip_detected);
|
| }
|
| if (route_to_sr_engine) {
|
| @@ -561,12 +561,7 @@ SpeechRecognizerImpl::StartRecording(const FSMEventArgs&) {
|
| // in_params.sample_rate()
|
| frames_per_buffer =
|
| ((in_params.sample_rate() * chunk_duration_ms) / 1000.0) + 0.5;
|
| - input_parameters.Reset(in_params.format(),
|
| - in_params.channel_layout(),
|
| - in_params.channels(),
|
| - in_params.sample_rate(),
|
| - in_params.bits_per_sample(),
|
| - frames_per_buffer);
|
| + input_parameters.set_frames_per_buffer(frames_per_buffer);
|
| DVLOG(1) << "SRI::input_parameters: "
|
| << input_parameters.AsHumanReadableString();
|
| }
|
|
|