OLD | NEW |
1 // Copyright (c) 2012 The Chromium Authors. All rights reserved. | 1 // Copyright (c) 2012 The Chromium Authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 #include "remoting/host/video_scheduler.h" | 5 #include "remoting/host/video_scheduler.h" |
6 | 6 |
7 #include <algorithm> | 7 #include <algorithm> |
8 | 8 |
9 #include "base/bind.h" | 9 #include "base/bind.h" |
10 #include "base/callback.h" | 10 #include "base/callback.h" |
11 #include "base/logging.h" | 11 #include "base/logging.h" |
12 #include "base/memory/scoped_ptr.h" | 12 #include "base/memory/scoped_ptr.h" |
13 #include "base/message_loop/message_loop_proxy.h" | 13 #include "base/message_loop/message_loop_proxy.h" |
14 #include "base/stl_util.h" | 14 #include "base/task_runner_util.h" |
15 #include "base/sys_info.h" | |
16 #include "base/time/time.h" | 15 #include "base/time/time.h" |
| 16 #include "remoting/host/capture_scheduler.h" |
17 #include "remoting/proto/control.pb.h" | 17 #include "remoting/proto/control.pb.h" |
18 #include "remoting/proto/internal.pb.h" | 18 #include "remoting/proto/internal.pb.h" |
19 #include "remoting/proto/video.pb.h" | 19 #include "remoting/proto/video.pb.h" |
20 #include "remoting/protocol/cursor_shape_stub.h" | 20 #include "remoting/protocol/cursor_shape_stub.h" |
21 #include "remoting/protocol/message_decoder.h" | |
22 #include "remoting/protocol/video_stub.h" | 21 #include "remoting/protocol/video_stub.h" |
23 #include "third_party/webrtc/modules/desktop_capture/desktop_capturer.h" | 22 #include "third_party/webrtc/modules/desktop_capture/desktop_capturer.h" |
24 #include "third_party/webrtc/modules/desktop_capture/desktop_frame.h" | 23 #include "third_party/webrtc/modules/desktop_capture/desktop_frame.h" |
25 #include "third_party/webrtc/modules/desktop_capture/mouse_cursor.h" | 24 #include "third_party/webrtc/modules/desktop_capture/mouse_cursor.h" |
26 | 25 |
27 namespace remoting { | 26 namespace remoting { |
28 | 27 |
29 // Maximum number of frames that can be processed simultaneously. | 28 namespace { |
30 // TODO(hclam): Move this value to CaptureScheduler. | 29 |
31 static const int kMaxPendingFrames = 2; | 30 // Helper used to encode frames on the encode thread. |
| 31 // |
| 32 // TODO(sergeyu): This functions doesn't do much beside calling |
| 33 // VideoEncoder::Encode(). It's only needed to handle empty frames properly and |
| 34 // that logic can be moved to VideoEncoder implementations. |
| 35 scoped_ptr<VideoPacket> EncodeFrame(VideoEncoder* encoder, |
| 36 scoped_ptr<webrtc::DesktopFrame> frame) { |
| 37 // If there is nothing to encode then send an empty packet. |
| 38 if (!frame || frame->updated_region().is_empty()) |
| 39 return make_scoped_ptr(new VideoPacket()); |
| 40 |
| 41 return encoder->Encode(*frame); |
| 42 } |
| 43 |
| 44 } // namespace |
32 | 45 |
33 // Interval between empty keep-alive frames. These frames are sent only when the | 46 // Interval between empty keep-alive frames. These frames are sent only when the |
34 // stream is paused or inactive for some other reason (e.g. when blocked on | 47 // stream is paused or inactive for some other reason (e.g. when blocked on |
35 // capturer). To prevent PseudoTCP from resetting congestion window this value | 48 // capturer). To prevent PseudoTCP from resetting congestion window this value |
36 // must be smaller than the minimum RTO used in PseudoTCP, which is 250ms. | 49 // must be smaller than the minimum RTO used in PseudoTCP, which is 250ms. |
37 static const int kKeepAlivePacketIntervalMs = 200; | 50 static const int kKeepAlivePacketIntervalMs = 200; |
38 | 51 |
39 static bool g_enable_timestamps = false; | 52 static bool g_enable_timestamps = false; |
40 | 53 |
41 // static | 54 // static |
(...skipping 11 matching lines...) Expand all Loading... |
53 protocol::CursorShapeStub* cursor_stub, | 66 protocol::CursorShapeStub* cursor_stub, |
54 protocol::VideoStub* video_stub) | 67 protocol::VideoStub* video_stub) |
55 : capture_task_runner_(capture_task_runner), | 68 : capture_task_runner_(capture_task_runner), |
56 encode_task_runner_(encode_task_runner), | 69 encode_task_runner_(encode_task_runner), |
57 network_task_runner_(network_task_runner), | 70 network_task_runner_(network_task_runner), |
58 capturer_(capturer.Pass()), | 71 capturer_(capturer.Pass()), |
59 mouse_cursor_monitor_(mouse_cursor_monitor.Pass()), | 72 mouse_cursor_monitor_(mouse_cursor_monitor.Pass()), |
60 encoder_(encoder.Pass()), | 73 encoder_(encoder.Pass()), |
61 cursor_stub_(cursor_stub), | 74 cursor_stub_(cursor_stub), |
62 video_stub_(video_stub), | 75 video_stub_(video_stub), |
63 pending_frames_(0), | |
64 capture_pending_(false), | |
65 did_skip_frame_(false), | |
66 is_paused_(false), | |
67 latest_event_timestamp_(0) { | 76 latest_event_timestamp_(0) { |
68 DCHECK(network_task_runner_->BelongsToCurrentThread()); | 77 DCHECK(network_task_runner_->BelongsToCurrentThread()); |
69 DCHECK(capturer_); | 78 DCHECK(capturer_); |
70 DCHECK(mouse_cursor_monitor_); | 79 DCHECK(mouse_cursor_monitor_); |
71 DCHECK(encoder_); | 80 DCHECK(encoder_); |
72 DCHECK(cursor_stub_); | 81 DCHECK(cursor_stub_); |
73 DCHECK(video_stub_); | 82 DCHECK(video_stub_); |
74 } | 83 } |
75 | 84 |
76 // Public methods -------------------------------------------------------------- | 85 // Public methods -------------------------------------------------------------- |
77 | 86 |
78 webrtc::SharedMemory* VideoScheduler::CreateSharedMemory(size_t size) { | |
79 return nullptr; | |
80 } | |
81 | |
82 void VideoScheduler::OnCaptureCompleted(webrtc::DesktopFrame* frame) { | |
83 DCHECK(capture_task_runner_->BelongsToCurrentThread()); | |
84 | |
85 capture_pending_ = false; | |
86 | |
87 scoped_ptr<webrtc::DesktopFrame> owned_frame(frame); | |
88 | |
89 if (owned_frame) { | |
90 scheduler_.RecordCaptureTime( | |
91 base::TimeDelta::FromMilliseconds(owned_frame->capture_time_ms())); | |
92 } | |
93 | |
94 // Even when |frame| is nullptr we still need to post it to the encode thread | |
95 // to make sure frames are freed in the same order they are received and | |
96 // that we don't start capturing frame n+2 before frame n is freed. | |
97 encode_task_runner_->PostTask( | |
98 FROM_HERE, base::Bind(&VideoScheduler::EncodeFrame, this, | |
99 base::Passed(&owned_frame), latest_event_timestamp_, | |
100 base::TimeTicks::Now())); | |
101 | |
102 // If a frame was skipped, try to capture it again. | |
103 if (did_skip_frame_) { | |
104 capture_task_runner_->PostTask( | |
105 FROM_HERE, base::Bind(&VideoScheduler::CaptureNextFrame, this)); | |
106 } | |
107 } | |
108 | |
109 void VideoScheduler::OnMouseCursor(webrtc::MouseCursor* cursor) { | |
110 DCHECK(capture_task_runner_->BelongsToCurrentThread()); | |
111 | |
112 scoped_ptr<webrtc::MouseCursor> owned_cursor(cursor); | |
113 | |
114 // Do nothing if the scheduler is being stopped. | |
115 if (!capturer_) | |
116 return; | |
117 | |
118 scoped_ptr<protocol::CursorShapeInfo> cursor_proto( | |
119 new protocol::CursorShapeInfo()); | |
120 cursor_proto->set_width(cursor->image()->size().width()); | |
121 cursor_proto->set_height(cursor->image()->size().height()); | |
122 cursor_proto->set_hotspot_x(cursor->hotspot().x()); | |
123 cursor_proto->set_hotspot_y(cursor->hotspot().y()); | |
124 | |
125 cursor_proto->set_data(std::string()); | |
126 uint8_t* current_row = cursor->image()->data(); | |
127 for (int y = 0; y < cursor->image()->size().height(); ++y) { | |
128 cursor_proto->mutable_data()->append( | |
129 current_row, | |
130 current_row + cursor->image()->size().width() * | |
131 webrtc::DesktopFrame::kBytesPerPixel); | |
132 current_row += cursor->image()->stride(); | |
133 } | |
134 | |
135 network_task_runner_->PostTask( | |
136 FROM_HERE, base::Bind(&VideoScheduler::SendCursorShape, this, | |
137 base::Passed(&cursor_proto))); | |
138 } | |
139 | |
140 void VideoScheduler::OnMouseCursorPosition( | |
141 webrtc::MouseCursorMonitor::CursorState state, | |
142 const webrtc::DesktopVector& position) { | |
143 // We're not subscribing to mouse position changes. | |
144 NOTREACHED(); | |
145 } | |
146 | |
147 void VideoScheduler::Start() { | 87 void VideoScheduler::Start() { |
148 DCHECK(network_task_runner_->BelongsToCurrentThread()); | 88 DCHECK(network_task_runner_->BelongsToCurrentThread()); |
149 | 89 |
| 90 keep_alive_timer_.reset(new base::DelayTimer<VideoScheduler>( |
| 91 FROM_HERE, base::TimeDelta::FromMilliseconds(kKeepAlivePacketIntervalMs), |
| 92 this, &VideoScheduler::SendKeepAlivePacket)); |
| 93 |
| 94 capture_scheduler_.reset(new CaptureScheduler( |
| 95 base::Bind(&VideoScheduler::CaptureNextFrame, this))); |
| 96 capture_scheduler_->Start(); |
| 97 |
150 capture_task_runner_->PostTask( | 98 capture_task_runner_->PostTask( |
151 FROM_HERE, base::Bind(&VideoScheduler::StartOnCaptureThread, this)); | 99 FROM_HERE, base::Bind(&VideoScheduler::StartOnCaptureThread, this)); |
152 } | 100 } |
153 | 101 |
154 void VideoScheduler::Stop() { | 102 void VideoScheduler::Stop() { |
155 DCHECK(network_task_runner_->BelongsToCurrentThread()); | 103 DCHECK(network_task_runner_->BelongsToCurrentThread()); |
156 | 104 |
157 // Clear stubs to prevent further updates reaching the client. | 105 // Clear stubs to prevent further updates reaching the client. |
158 cursor_stub_ = nullptr; | 106 cursor_stub_ = nullptr; |
159 video_stub_ = nullptr; | 107 video_stub_ = nullptr; |
160 | 108 |
| 109 capture_scheduler_.reset(); |
161 keep_alive_timer_.reset(); | 110 keep_alive_timer_.reset(); |
162 | 111 |
163 capture_task_runner_->PostTask( | 112 capture_task_runner_->PostTask( |
164 FROM_HERE, base::Bind(&VideoScheduler::StopOnCaptureThread, this)); | 113 FROM_HERE, base::Bind(&VideoScheduler::StopOnCaptureThread, this)); |
165 } | 114 } |
166 | 115 |
167 void VideoScheduler::Pause(bool pause) { | 116 void VideoScheduler::Pause(bool pause) { |
168 if (!capture_task_runner_->BelongsToCurrentThread()) { | 117 DCHECK(network_task_runner_->BelongsToCurrentThread()); |
169 DCHECK(network_task_runner_->BelongsToCurrentThread()); | |
170 capture_task_runner_->PostTask( | |
171 FROM_HERE, base::Bind(&VideoScheduler::Pause, this, pause)); | |
172 return; | |
173 } | |
174 | 118 |
175 if (is_paused_ != pause) { | 119 capture_scheduler_->Pause(pause); |
176 is_paused_ = pause; | |
177 | |
178 // Restart captures if we're resuming and there are none scheduled. | |
179 if (!is_paused_ && capture_timer_ && !capture_timer_->IsRunning()) | |
180 CaptureNextFrame(); | |
181 } | |
182 } | 120 } |
183 | 121 |
184 void VideoScheduler::SetLatestEventTimestamp(int64 latest_event_timestamp) { | 122 void VideoScheduler::SetLatestEventTimestamp(int64 latest_event_timestamp) { |
185 if (!capture_task_runner_->BelongsToCurrentThread()) { | 123 DCHECK(network_task_runner_->BelongsToCurrentThread()); |
186 DCHECK(network_task_runner_->BelongsToCurrentThread()); | |
187 capture_task_runner_->PostTask( | |
188 FROM_HERE, base::Bind(&VideoScheduler::SetLatestEventTimestamp, | |
189 this, latest_event_timestamp)); | |
190 return; | |
191 } | |
192 | 124 |
193 latest_event_timestamp_ = latest_event_timestamp; | 125 latest_event_timestamp_ = latest_event_timestamp; |
194 } | 126 } |
195 | 127 |
196 void VideoScheduler::SetLosslessEncode(bool want_lossless) { | 128 void VideoScheduler::SetLosslessEncode(bool want_lossless) { |
197 if (!encode_task_runner_->BelongsToCurrentThread()) { | 129 DCHECK(network_task_runner_->BelongsToCurrentThread()); |
198 DCHECK(network_task_runner_->BelongsToCurrentThread()); | |
199 encode_task_runner_->PostTask( | |
200 FROM_HERE, base::Bind(&VideoScheduler::SetLosslessEncode, | |
201 this, want_lossless)); | |
202 return; | |
203 } | |
204 | 130 |
205 encoder_->SetLosslessEncode(want_lossless); | 131 encode_task_runner_->PostTask( |
| 132 FROM_HERE, base::Bind(&VideoEncoder::SetLosslessEncode, |
| 133 base::Unretained(encoder_.get()), want_lossless)); |
206 } | 134 } |
207 | 135 |
208 void VideoScheduler::SetLosslessColor(bool want_lossless) { | 136 void VideoScheduler::SetLosslessColor(bool want_lossless) { |
209 if (!encode_task_runner_->BelongsToCurrentThread()) { | 137 DCHECK(network_task_runner_->BelongsToCurrentThread()); |
210 DCHECK(network_task_runner_->BelongsToCurrentThread()); | |
211 encode_task_runner_->PostTask( | |
212 FROM_HERE, base::Bind(&VideoScheduler::SetLosslessColor, | |
213 this, want_lossless)); | |
214 return; | |
215 } | |
216 | 138 |
217 encoder_->SetLosslessColor(want_lossless); | 139 encode_task_runner_->PostTask( |
| 140 FROM_HERE, base::Bind(&VideoEncoder::SetLosslessColor, |
| 141 base::Unretained(encoder_.get()), want_lossless)); |
218 } | 142 } |
219 | 143 |
220 // Private methods ----------------------------------------------------------- | 144 // Private methods ----------------------------------------------------------- |
221 | 145 |
222 VideoScheduler::~VideoScheduler() { | 146 VideoScheduler::~VideoScheduler() { |
223 // Destroy the capturer and encoder on their respective threads. | 147 // Destroy the capturer and encoder on their respective threads. |
224 capture_task_runner_->DeleteSoon(FROM_HERE, capturer_.release()); | 148 capture_task_runner_->DeleteSoon(FROM_HERE, capturer_.release()); |
225 capture_task_runner_->DeleteSoon(FROM_HERE, mouse_cursor_monitor_.release()); | 149 capture_task_runner_->DeleteSoon(FROM_HERE, mouse_cursor_monitor_.release()); |
226 encode_task_runner_->DeleteSoon(FROM_HERE, encoder_.release()); | 150 encode_task_runner_->DeleteSoon(FROM_HERE, encoder_.release()); |
227 } | 151 } |
228 | 152 |
229 // Capturer thread ------------------------------------------------------------- | 153 // Capturer thread ------------------------------------------------------------- |
230 | 154 |
| 155 webrtc::SharedMemory* VideoScheduler::CreateSharedMemory(size_t size) { |
| 156 return nullptr; |
| 157 } |
| 158 |
| 159 void VideoScheduler::OnCaptureCompleted(webrtc::DesktopFrame* frame) { |
| 160 DCHECK(capture_task_runner_->BelongsToCurrentThread()); |
| 161 |
| 162 network_task_runner_->PostTask( |
| 163 FROM_HERE, base::Bind(&VideoScheduler::EncodeAndSendFrame, this, |
| 164 base::Passed(make_scoped_ptr(frame)))); |
| 165 } |
| 166 |
| 167 void VideoScheduler::OnMouseCursor(webrtc::MouseCursor* cursor) { |
| 168 DCHECK(capture_task_runner_->BelongsToCurrentThread()); |
| 169 |
| 170 scoped_ptr<webrtc::MouseCursor> owned_cursor(cursor); |
| 171 |
| 172 scoped_ptr<protocol::CursorShapeInfo> cursor_proto( |
| 173 new protocol::CursorShapeInfo()); |
| 174 cursor_proto->set_width(cursor->image()->size().width()); |
| 175 cursor_proto->set_height(cursor->image()->size().height()); |
| 176 cursor_proto->set_hotspot_x(cursor->hotspot().x()); |
| 177 cursor_proto->set_hotspot_y(cursor->hotspot().y()); |
| 178 |
| 179 cursor_proto->set_data(std::string()); |
| 180 uint8_t* current_row = cursor->image()->data(); |
| 181 for (int y = 0; y < cursor->image()->size().height(); ++y) { |
| 182 cursor_proto->mutable_data()->append( |
| 183 current_row, |
| 184 current_row + cursor->image()->size().width() * |
| 185 webrtc::DesktopFrame::kBytesPerPixel); |
| 186 current_row += cursor->image()->stride(); |
| 187 } |
| 188 |
| 189 network_task_runner_->PostTask( |
| 190 FROM_HERE, base::Bind(&VideoScheduler::SendCursorShape, this, |
| 191 base::Passed(&cursor_proto))); |
| 192 } |
| 193 |
| 194 void VideoScheduler::OnMouseCursorPosition( |
| 195 webrtc::MouseCursorMonitor::CursorState state, |
| 196 const webrtc::DesktopVector& position) { |
| 197 // We're not subscribing to mouse position changes. |
| 198 NOTREACHED(); |
| 199 } |
| 200 |
231 void VideoScheduler::StartOnCaptureThread() { | 201 void VideoScheduler::StartOnCaptureThread() { |
232 DCHECK(capture_task_runner_->BelongsToCurrentThread()); | 202 DCHECK(capture_task_runner_->BelongsToCurrentThread()); |
233 DCHECK(!capture_timer_); | |
234 | 203 |
235 // Start mouse cursor monitor. | |
236 mouse_cursor_monitor_->Init(this, webrtc::MouseCursorMonitor::SHAPE_ONLY); | 204 mouse_cursor_monitor_->Init(this, webrtc::MouseCursorMonitor::SHAPE_ONLY); |
237 | |
238 // Start the capturer. | |
239 capturer_->Start(this); | 205 capturer_->Start(this); |
240 | |
241 capture_timer_.reset(new base::OneShotTimer<VideoScheduler>()); | |
242 keep_alive_timer_.reset(new base::DelayTimer<VideoScheduler>( | |
243 FROM_HERE, base::TimeDelta::FromMilliseconds(kKeepAlivePacketIntervalMs), | |
244 this, &VideoScheduler::SendKeepAlivePacket)); | |
245 | |
246 // Capture first frame immediately. | |
247 CaptureNextFrame(); | |
248 } | 206 } |
249 | 207 |
250 void VideoScheduler::StopOnCaptureThread() { | 208 void VideoScheduler::StopOnCaptureThread() { |
251 DCHECK(capture_task_runner_->BelongsToCurrentThread()); | 209 DCHECK(capture_task_runner_->BelongsToCurrentThread()); |
252 | 210 |
253 // This doesn't deleted already captured frames, so encoder can keep using the | 211 // This doesn't deleted already captured frames, so encoder can keep using the |
254 // frames that were captured previously. | 212 // frames that were captured previously. |
255 capturer_.reset(); | 213 capturer_.reset(); |
256 | 214 |
257 // |capture_timer_| must be destroyed on the thread on which it is used. | 215 mouse_cursor_monitor_.reset(); |
258 capture_timer_.reset(); | |
259 } | 216 } |
260 | 217 |
261 void VideoScheduler::ScheduleNextCapture() { | 218 void VideoScheduler::CaptureNextFrameOnCaptureThread() { |
262 DCHECK(capture_task_runner_->BelongsToCurrentThread()); | 219 DCHECK(capture_task_runner_->BelongsToCurrentThread()); |
263 | 220 |
264 capture_timer_->Start(FROM_HERE, | 221 // Capture mouse shape first and then screen content. |
265 scheduler_.NextCaptureDelay(), | |
266 this, | |
267 &VideoScheduler::CaptureNextFrame); | |
268 } | |
269 | |
270 void VideoScheduler::CaptureNextFrame() { | |
271 DCHECK(capture_task_runner_->BelongsToCurrentThread()); | |
272 | |
273 // If we are stopping (|capturer_| is nullptr), or paused, then don't capture. | |
274 if (!capturer_ || is_paused_) | |
275 return; | |
276 | |
277 // Make sure we have at most two outstanding recordings. We can simply return | |
278 // if we can't make a capture now, the next capture will be started by the | |
279 // end of an encode operation. | |
280 if (pending_frames_ >= kMaxPendingFrames || capture_pending_) { | |
281 did_skip_frame_ = true; | |
282 return; | |
283 } | |
284 | |
285 did_skip_frame_ = false; | |
286 | |
287 // At this point we are going to perform one capture so save the current time. | |
288 pending_frames_++; | |
289 DCHECK_LE(pending_frames_, kMaxPendingFrames); | |
290 | |
291 // Before doing a capture schedule for the next one. | |
292 ScheduleNextCapture(); | |
293 | |
294 capture_pending_ = true; | |
295 | |
296 // Capture the mouse shape. | |
297 mouse_cursor_monitor_->Capture(); | 222 mouse_cursor_monitor_->Capture(); |
298 | |
299 // And finally perform one capture. | |
300 capturer_->Capture(webrtc::DesktopRegion()); | 223 capturer_->Capture(webrtc::DesktopRegion()); |
301 } | 224 } |
302 | 225 |
303 void VideoScheduler::FrameCaptureCompleted() { | |
304 DCHECK(capture_task_runner_->BelongsToCurrentThread()); | |
305 | |
306 // Decrement the pending capture count. | |
307 pending_frames_--; | |
308 DCHECK_GE(pending_frames_, 0); | |
309 | |
310 // If we've skipped a frame capture because too we had too many captures | |
311 // pending then schedule one now. | |
312 if (did_skip_frame_) | |
313 CaptureNextFrame(); | |
314 } | |
315 | |
316 // Network thread -------------------------------------------------------------- | 226 // Network thread -------------------------------------------------------------- |
317 | 227 |
318 void VideoScheduler::SendVideoPacket(scoped_ptr<VideoPacket> packet) { | 228 void VideoScheduler::CaptureNextFrame() { |
| 229 DCHECK(network_task_runner_->BelongsToCurrentThread()); |
| 230 |
| 231 capture_task_runner_->PostTask( |
| 232 FROM_HERE, |
| 233 base::Bind(&VideoScheduler::CaptureNextFrameOnCaptureThread, this)); |
| 234 } |
| 235 |
| 236 void VideoScheduler::EncodeAndSendFrame( |
| 237 scoped_ptr<webrtc::DesktopFrame> frame) { |
319 DCHECK(network_task_runner_->BelongsToCurrentThread()); | 238 DCHECK(network_task_runner_->BelongsToCurrentThread()); |
320 | 239 |
321 if (!video_stub_) | 240 if (!video_stub_) |
322 return; | 241 return; |
323 | 242 |
| 243 capture_scheduler_->OnCaptureCompleted(); |
| 244 |
| 245 // Even when |frame| is nullptr we still need to post it to the encode thread |
| 246 // to make sure frames are freed in the same order they are received and |
| 247 // that we don't start capturing frame n+2 before frame n is freed. |
| 248 base::PostTaskAndReplyWithResult( |
| 249 encode_task_runner_.get(), FROM_HERE, |
| 250 base::Bind(&EncodeFrame, encoder_.get(), base::Passed(&frame)), |
| 251 base::Bind(&VideoScheduler::SendEncodedFrame, this, |
| 252 latest_event_timestamp_, base::TimeTicks::Now())); |
| 253 } |
| 254 |
| 255 void VideoScheduler::SendEncodedFrame(int64 latest_event_timestamp, |
| 256 base::TimeTicks timestamp, |
| 257 scoped_ptr<VideoPacket> packet) { |
| 258 DCHECK(network_task_runner_->BelongsToCurrentThread()); |
| 259 |
| 260 if (!video_stub_) |
| 261 return; |
| 262 |
| 263 if (g_enable_timestamps) |
| 264 packet->set_timestamp(timestamp.ToInternalValue()); |
| 265 |
| 266 packet->set_latest_event_timestamp(latest_event_timestamp); |
| 267 |
| 268 capture_scheduler_->OnFrameEncoded( |
| 269 base::TimeDelta::FromMilliseconds(packet->encode_time_ms())); |
| 270 |
324 video_stub_->ProcessVideoPacket( | 271 video_stub_->ProcessVideoPacket( |
325 packet.Pass(), base::Bind(&VideoScheduler::OnVideoPacketSent, this)); | 272 packet.Pass(), base::Bind(&VideoScheduler::OnVideoPacketSent, this)); |
326 } | 273 } |
327 | 274 |
328 void VideoScheduler::OnVideoPacketSent() { | 275 void VideoScheduler::OnVideoPacketSent() { |
329 DCHECK(network_task_runner_->BelongsToCurrentThread()); | 276 DCHECK(network_task_runner_->BelongsToCurrentThread()); |
330 | 277 |
331 if (!video_stub_) | 278 if (!video_stub_) |
332 return; | 279 return; |
333 | 280 |
| 281 capture_scheduler_->OnFrameSent(); |
334 keep_alive_timer_->Reset(); | 282 keep_alive_timer_->Reset(); |
335 | |
336 capture_task_runner_->PostTask( | |
337 FROM_HERE, base::Bind(&VideoScheduler::FrameCaptureCompleted, this)); | |
338 } | 283 } |
339 | 284 |
340 void VideoScheduler::SendKeepAlivePacket() { | 285 void VideoScheduler::SendKeepAlivePacket() { |
341 DCHECK(network_task_runner_->BelongsToCurrentThread()); | 286 DCHECK(network_task_runner_->BelongsToCurrentThread()); |
342 | 287 |
343 if (!video_stub_) | |
344 return; | |
345 | |
346 video_stub_->ProcessVideoPacket( | 288 video_stub_->ProcessVideoPacket( |
347 make_scoped_ptr(new VideoPacket()), | 289 make_scoped_ptr(new VideoPacket()), |
348 base::Bind(&VideoScheduler::OnKeepAlivePacketSent, this)); | 290 base::Bind(&VideoScheduler::OnKeepAlivePacketSent, this)); |
349 } | 291 } |
350 | 292 |
351 void VideoScheduler::OnKeepAlivePacketSent() { | 293 void VideoScheduler::OnKeepAlivePacketSent() { |
352 DCHECK(network_task_runner_->BelongsToCurrentThread()); | 294 DCHECK(network_task_runner_->BelongsToCurrentThread()); |
353 | 295 |
354 if (keep_alive_timer_) | 296 if (keep_alive_timer_) |
355 keep_alive_timer_->Reset(); | 297 keep_alive_timer_->Reset(); |
356 } | 298 } |
357 | 299 |
358 void VideoScheduler::SendCursorShape( | 300 void VideoScheduler::SendCursorShape( |
359 scoped_ptr<protocol::CursorShapeInfo> cursor_shape) { | 301 scoped_ptr<protocol::CursorShapeInfo> cursor_shape) { |
360 DCHECK(network_task_runner_->BelongsToCurrentThread()); | 302 DCHECK(network_task_runner_->BelongsToCurrentThread()); |
361 | 303 |
362 if (!cursor_stub_) | 304 if (!cursor_stub_) |
363 return; | 305 return; |
364 | 306 |
365 cursor_stub_->SetCursorShape(*cursor_shape); | 307 cursor_stub_->SetCursorShape(*cursor_shape); |
366 } | 308 } |
367 | 309 |
368 // Encoder thread -------------------------------------------------------------- | |
369 | |
370 void VideoScheduler::EncodeFrame( | |
371 scoped_ptr<webrtc::DesktopFrame> frame, | |
372 int64 latest_event_timestamp, | |
373 base::TimeTicks timestamp) { | |
374 DCHECK(encode_task_runner_->BelongsToCurrentThread()); | |
375 | |
376 // If there is nothing to encode then send an empty packet. | |
377 if (!frame || frame->updated_region().is_empty()) { | |
378 capture_task_runner_->DeleteSoon(FROM_HERE, frame.release()); | |
379 scoped_ptr<VideoPacket> packet(new VideoPacket()); | |
380 packet->set_latest_event_timestamp(latest_event_timestamp); | |
381 network_task_runner_->PostTask( | |
382 FROM_HERE, | |
383 base::Bind( | |
384 &VideoScheduler::SendVideoPacket, this, base::Passed(&packet))); | |
385 return; | |
386 } | |
387 | |
388 scoped_ptr<VideoPacket> packet = encoder_->Encode(*frame); | |
389 packet->set_latest_event_timestamp(latest_event_timestamp); | |
390 | |
391 if (g_enable_timestamps) { | |
392 packet->set_timestamp(timestamp.ToInternalValue()); | |
393 } | |
394 | |
395 // Destroy the frame before sending |packet| because SendVideoPacket() may | |
396 // trigger another frame to be captured, and the screen capturer expects the | |
397 // old frame to be freed by then. | |
398 frame.reset(); | |
399 | |
400 scheduler_.RecordEncodeTime( | |
401 base::TimeDelta::FromMilliseconds(packet->encode_time_ms())); | |
402 network_task_runner_->PostTask( | |
403 FROM_HERE, base::Bind(&VideoScheduler::SendVideoPacket, this, | |
404 base::Passed(&packet))); | |
405 } | |
406 | |
407 } // namespace remoting | 310 } // namespace remoting |
OLD | NEW |