OLD | NEW |
1 // Copyright (c) 2012 The Chromium Authors. All rights reserved. | 1 // Copyright (c) 2012 The Chromium Authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 #ifndef CONTENT_RENDERER_SPEECH_RECOGNITION_DISPATCHER_H_ | 5 #ifndef CONTENT_RENDERER_SPEECH_RECOGNITION_DISPATCHER_H_ |
6 #define CONTENT_RENDERER_SPEECH_RECOGNITION_DISPATCHER_H_ | 6 #define CONTENT_RENDERER_SPEECH_RECOGNITION_DISPATCHER_H_ |
7 | 7 |
8 #include <map> | 8 #include <map> |
9 #include <memory> | 9 #include <memory> |
10 | 10 |
11 #include "base/macros.h" | 11 #include "base/macros.h" |
12 #include "base/memory/shared_memory.h" | 12 #include "base/memory/shared_memory.h" |
13 #include "base/sync_socket.h" | 13 #include "base/sync_socket.h" |
14 #include "content/public/common/speech_recognition_result.h" | 14 #include "content/public/common/speech_recognition_result.h" |
15 #include "content/public/renderer/render_view_observer.h" | 15 #include "content/public/renderer/render_view_observer.h" |
| 16 #include "media/media_features.h" |
16 #include "third_party/WebKit/public/platform/WebMediaStreamTrack.h" | 17 #include "third_party/WebKit/public/platform/WebMediaStreamTrack.h" |
17 #include "third_party/WebKit/public/platform/WebVector.h" | 18 #include "third_party/WebKit/public/platform/WebVector.h" |
18 #include "third_party/WebKit/public/web/WebSpeechRecognitionHandle.h" | 19 #include "third_party/WebKit/public/web/WebSpeechRecognitionHandle.h" |
19 #include "third_party/WebKit/public/web/WebSpeechRecognizer.h" | 20 #include "third_party/WebKit/public/web/WebSpeechRecognizer.h" |
20 | 21 |
21 namespace media { | 22 namespace media { |
22 class AudioParameters; | 23 class AudioParameters; |
23 } | 24 } |
24 | 25 |
25 namespace content { | 26 namespace content { |
26 class RenderViewImpl; | 27 class RenderViewImpl; |
27 #if defined(ENABLE_WEBRTC) | 28 #if BUILDFLAG(ENABLE_WEBRTC) |
28 class SpeechRecognitionAudioSink; | 29 class SpeechRecognitionAudioSink; |
29 #endif | 30 #endif |
30 struct SpeechRecognitionError; | 31 struct SpeechRecognitionError; |
31 struct SpeechRecognitionResult; | 32 struct SpeechRecognitionResult; |
32 | 33 |
33 // SpeechRecognitionDispatcher is a delegate for methods used by WebKit for | 34 // SpeechRecognitionDispatcher is a delegate for methods used by WebKit for |
34 // scripted JS speech APIs. It's the complement of | 35 // scripted JS speech APIs. It's the complement of |
35 // SpeechRecognitionDispatcherHost (owned by RenderViewHost). | 36 // SpeechRecognitionDispatcherHost (owned by RenderViewHost). |
36 class SpeechRecognitionDispatcher : public RenderViewObserver, | 37 class SpeechRecognitionDispatcher : public RenderViewObserver, |
37 public blink::WebSpeechRecognizer { | 38 public blink::WebSpeechRecognizer { |
(...skipping 34 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
72 | 73 |
73 void ResetAudioSink(); | 74 void ResetAudioSink(); |
74 | 75 |
75 int GetOrCreateIDForHandle(const blink::WebSpeechRecognitionHandle& handle); | 76 int GetOrCreateIDForHandle(const blink::WebSpeechRecognitionHandle& handle); |
76 bool HandleExists(const blink::WebSpeechRecognitionHandle& handle); | 77 bool HandleExists(const blink::WebSpeechRecognitionHandle& handle); |
77 const blink::WebSpeechRecognitionHandle& GetHandleFromID(int handle_id); | 78 const blink::WebSpeechRecognitionHandle& GetHandleFromID(int handle_id); |
78 | 79 |
79 // The WebKit client class that we use to send events back to the JS world. | 80 // The WebKit client class that we use to send events back to the JS world. |
80 blink::WebSpeechRecognizerClient* recognizer_client_; | 81 blink::WebSpeechRecognizerClient* recognizer_client_; |
81 | 82 |
82 #if defined(ENABLE_WEBRTC) | 83 #if BUILDFLAG(ENABLE_WEBRTC) |
83 // Media stream audio track that the speech recognition connects to. | 84 // Media stream audio track that the speech recognition connects to. |
84 // Accessed on the render thread. | 85 // Accessed on the render thread. |
85 blink::WebMediaStreamTrack audio_track_; | 86 blink::WebMediaStreamTrack audio_track_; |
86 | 87 |
87 // Audio sink used to provide audio from the track. | 88 // Audio sink used to provide audio from the track. |
88 std::unique_ptr<SpeechRecognitionAudioSink> speech_audio_sink_; | 89 std::unique_ptr<SpeechRecognitionAudioSink> speech_audio_sink_; |
89 #endif | 90 #endif |
90 | 91 |
91 typedef std::map<int, blink::WebSpeechRecognitionHandle> HandleMap; | 92 typedef std::map<int, blink::WebSpeechRecognitionHandle> HandleMap; |
92 HandleMap handle_map_; | 93 HandleMap handle_map_; |
93 int next_id_; | 94 int next_id_; |
94 | 95 |
95 DISALLOW_COPY_AND_ASSIGN(SpeechRecognitionDispatcher); | 96 DISALLOW_COPY_AND_ASSIGN(SpeechRecognitionDispatcher); |
96 }; | 97 }; |
97 | 98 |
98 } // namespace content | 99 } // namespace content |
99 | 100 |
100 #endif // CONTENT_RENDERER_SPEECH_RECOGNITION_DISPATCHER_H_ | 101 #endif // CONTENT_RENDERER_SPEECH_RECOGNITION_DISPATCHER_H_ |
OLD | NEW |