| Index: content/renderer/speech_recognition_dispatcher.cc
|
| diff --git a/content/renderer/speech_recognition_dispatcher.cc b/content/renderer/speech_recognition_dispatcher.cc
|
| index dea16bc5cfe07c1f668b1c89f87d54307e9aa873..433e9c66c4649e82cd43d2fc14ea3db7e49feb15 100644
|
| --- a/content/renderer/speech_recognition_dispatcher.cc
|
| +++ b/content/renderer/speech_recognition_dispatcher.cc
|
| @@ -7,7 +7,7 @@
|
| #include "base/basictypes.h"
|
| #include "base/strings/utf_string_conversions.h"
|
| #include "content/common/speech_recognition_messages.h"
|
| -#include "content/renderer/render_view_impl.h"
|
| +#include "content/renderer/render_frame_impl.h"
|
| #include "third_party/WebKit/public/platform/WebString.h"
|
| #include "third_party/WebKit/public/platform/WebVector.h"
|
| #include "third_party/WebKit/public/web/WebSpeechGrammar.h"
|
| @@ -30,19 +30,13 @@ using blink::WebSpeechRecognizerClient;
|
| namespace content {
|
|
|
| SpeechRecognitionDispatcher::SpeechRecognitionDispatcher(
|
| - RenderViewImpl* render_view)
|
| - : RenderViewObserver(render_view),
|
| + RenderFrame* render_frame)
|
| + : RenderFrameObserver(render_frame),
|
| recognizer_client_(NULL),
|
| next_id_(1) {}
|
|
|
| SpeechRecognitionDispatcher::~SpeechRecognitionDispatcher() {}
|
|
|
| -void SpeechRecognitionDispatcher::AbortAllRecognitions() {
|
| - ResetAudioSink();
|
| - Send(new SpeechRecognitionHostMsg_AbortAllRequests(
|
| - routing_id()));
|
| -}
|
| -
|
| bool SpeechRecognitionDispatcher::OnMessageReceived(
|
| const IPC::Message& message) {
|
| bool handled = true;
|
| @@ -63,6 +57,13 @@ bool SpeechRecognitionDispatcher::OnMessageReceived(
|
| return handled;
|
| }
|
|
|
| +void SpeechRecognitionDispatcher::WasHidden() {
|
| +#if defined(OS_ANDROID) && defined(ENABLE_WEBRTC)
|
| + ResetAudioSink();
|
| + Send(new SpeechRecognitionHostMsg_AbortAllRequests(routing_id()));
|
| +#endif
|
| +}
|
| +
|
| void SpeechRecognitionDispatcher::start(
|
| const WebSpeechRecognitionHandle& handle,
|
| const WebSpeechRecognitionParams& params,
|
| @@ -104,7 +105,6 @@ void SpeechRecognitionDispatcher::start(
|
| msg_params.continuous = params.continuous();
|
| msg_params.interim_results = params.interimResults();
|
| msg_params.origin_url = params.origin().toString().utf8();
|
| - msg_params.render_view_id = routing_id();
|
| msg_params.request_id = GetOrCreateIDForHandle(handle);
|
| #if defined(ENABLE_WEBRTC)
|
| // Fall back to default input when the track is not allowed.
|
| @@ -112,6 +112,7 @@ void SpeechRecognitionDispatcher::start(
|
| #else
|
| msg_params.using_audio_track = false;
|
| #endif
|
| + msg_params.render_frame_id = routing_id();
|
| // The handle mapping will be removed in |OnRecognitionEnd|.
|
| Send(new SpeechRecognitionHostMsg_StartRequest(msg_params));
|
| }
|
|
|