OLD | NEW |
| (Empty) |
1 // Copyright 2014 The Chromium Authors. All rights reserved. | |
2 // Use of this source code is governed by a BSD-style license that can be | |
3 // found in the LICENSE file. | |
4 | |
5 #include "mojo/edk/system/raw_channel.h" | |
6 | |
7 #include <string.h> | |
8 | |
9 #include <algorithm> | |
10 | |
11 #include "base/bind.h" | |
12 #include "base/location.h" | |
13 #include "base/logging.h" | |
14 #include "base/message_loop/message_loop.h" | |
15 #include "base/stl_util.h" | |
16 #include "mojo/edk/system/message_in_transit.h" | |
17 #include "mojo/edk/system/transport_data.h" | |
18 | |
19 namespace mojo { | |
20 namespace system { | |
21 | |
22 const size_t kReadSize = 4096; | |
23 | |
24 // RawChannel::ReadBuffer ------------------------------------------------------ | |
25 | |
26 RawChannel::ReadBuffer::ReadBuffer() : buffer_(kReadSize), num_valid_bytes_(0) { | |
27 } | |
28 | |
29 RawChannel::ReadBuffer::~ReadBuffer() { | |
30 } | |
31 | |
32 void RawChannel::ReadBuffer::GetBuffer(char** addr, size_t* size) { | |
33 DCHECK_GE(buffer_.size(), num_valid_bytes_ + kReadSize); | |
34 *addr = &buffer_[0] + num_valid_bytes_; | |
35 *size = kReadSize; | |
36 } | |
37 | |
38 // RawChannel::WriteBuffer ----------------------------------------------------- | |
39 | |
40 RawChannel::WriteBuffer::WriteBuffer(size_t serialized_platform_handle_size) | |
41 : serialized_platform_handle_size_(serialized_platform_handle_size), | |
42 platform_handles_offset_(0), | |
43 data_offset_(0) { | |
44 } | |
45 | |
46 RawChannel::WriteBuffer::~WriteBuffer() { | |
47 STLDeleteElements(&message_queue_); | |
48 } | |
49 | |
50 bool RawChannel::WriteBuffer::HavePlatformHandlesToSend() const { | |
51 if (message_queue_.empty()) | |
52 return false; | |
53 | |
54 const TransportData* transport_data = | |
55 message_queue_.front()->transport_data(); | |
56 if (!transport_data) | |
57 return false; | |
58 | |
59 const embedder::PlatformHandleVector* all_platform_handles = | |
60 transport_data->platform_handles(); | |
61 if (!all_platform_handles) { | |
62 DCHECK_EQ(platform_handles_offset_, 0u); | |
63 return false; | |
64 } | |
65 if (platform_handles_offset_ >= all_platform_handles->size()) { | |
66 DCHECK_EQ(platform_handles_offset_, all_platform_handles->size()); | |
67 return false; | |
68 } | |
69 | |
70 return true; | |
71 } | |
72 | |
73 void RawChannel::WriteBuffer::GetPlatformHandlesToSend( | |
74 size_t* num_platform_handles, | |
75 embedder::PlatformHandle** platform_handles, | |
76 void** serialization_data) { | |
77 DCHECK(HavePlatformHandlesToSend()); | |
78 | |
79 TransportData* transport_data = message_queue_.front()->transport_data(); | |
80 embedder::PlatformHandleVector* all_platform_handles = | |
81 transport_data->platform_handles(); | |
82 *num_platform_handles = | |
83 all_platform_handles->size() - platform_handles_offset_; | |
84 *platform_handles = &(*all_platform_handles)[platform_handles_offset_]; | |
85 size_t serialization_data_offset = | |
86 transport_data->platform_handle_table_offset(); | |
87 DCHECK_GT(serialization_data_offset, 0u); | |
88 serialization_data_offset += | |
89 platform_handles_offset_ * serialized_platform_handle_size_; | |
90 *serialization_data = | |
91 static_cast<char*>(transport_data->buffer()) + serialization_data_offset; | |
92 } | |
93 | |
94 void RawChannel::WriteBuffer::GetBuffers(std::vector<Buffer>* buffers) const { | |
95 buffers->clear(); | |
96 | |
97 if (message_queue_.empty()) | |
98 return; | |
99 | |
100 MessageInTransit* message = message_queue_.front(); | |
101 DCHECK_LT(data_offset_, message->total_size()); | |
102 size_t bytes_to_write = message->total_size() - data_offset_; | |
103 | |
104 size_t transport_data_buffer_size = | |
105 message->transport_data() ? message->transport_data()->buffer_size() : 0; | |
106 | |
107 if (!transport_data_buffer_size) { | |
108 // Only write from the main buffer. | |
109 DCHECK_LT(data_offset_, message->main_buffer_size()); | |
110 DCHECK_LE(bytes_to_write, message->main_buffer_size()); | |
111 Buffer buffer = { | |
112 static_cast<const char*>(message->main_buffer()) + data_offset_, | |
113 bytes_to_write}; | |
114 buffers->push_back(buffer); | |
115 return; | |
116 } | |
117 | |
118 if (data_offset_ >= message->main_buffer_size()) { | |
119 // Only write from the transport data buffer. | |
120 DCHECK_LT(data_offset_ - message->main_buffer_size(), | |
121 transport_data_buffer_size); | |
122 DCHECK_LE(bytes_to_write, transport_data_buffer_size); | |
123 Buffer buffer = { | |
124 static_cast<const char*>(message->transport_data()->buffer()) + | |
125 (data_offset_ - message->main_buffer_size()), | |
126 bytes_to_write}; | |
127 buffers->push_back(buffer); | |
128 return; | |
129 } | |
130 | |
131 // TODO(vtl): We could actually send out buffers from multiple messages, with | |
132 // the "stopping" condition being reaching a message with platform handles | |
133 // attached. | |
134 | |
135 // Write from both buffers. | |
136 DCHECK_EQ( | |
137 bytes_to_write, | |
138 message->main_buffer_size() - data_offset_ + transport_data_buffer_size); | |
139 Buffer buffer1 = { | |
140 static_cast<const char*>(message->main_buffer()) + data_offset_, | |
141 message->main_buffer_size() - data_offset_}; | |
142 buffers->push_back(buffer1); | |
143 Buffer buffer2 = { | |
144 static_cast<const char*>(message->transport_data()->buffer()), | |
145 transport_data_buffer_size}; | |
146 buffers->push_back(buffer2); | |
147 } | |
148 | |
149 // RawChannel ------------------------------------------------------------------ | |
150 | |
151 RawChannel::RawChannel() | |
152 : message_loop_for_io_(nullptr), | |
153 delegate_(nullptr), | |
154 read_stopped_(false), | |
155 write_stopped_(false), | |
156 weak_ptr_factory_(this) { | |
157 } | |
158 | |
159 RawChannel::~RawChannel() { | |
160 DCHECK(!read_buffer_); | |
161 DCHECK(!write_buffer_); | |
162 | |
163 // No need to take the |write_lock_| here -- if there are still weak pointers | |
164 // outstanding, then we're hosed anyway (since we wouldn't be able to | |
165 // invalidate them cleanly, since we might not be on the I/O thread). | |
166 DCHECK(!weak_ptr_factory_.HasWeakPtrs()); | |
167 } | |
168 | |
169 bool RawChannel::Init(Delegate* delegate) { | |
170 DCHECK(delegate); | |
171 | |
172 DCHECK(!delegate_); | |
173 delegate_ = delegate; | |
174 | |
175 CHECK_EQ(base::MessageLoop::current()->type(), base::MessageLoop::TYPE_IO); | |
176 DCHECK(!message_loop_for_io_); | |
177 message_loop_for_io_ = | |
178 static_cast<base::MessageLoopForIO*>(base::MessageLoop::current()); | |
179 | |
180 // No need to take the lock. No one should be using us yet. | |
181 DCHECK(!read_buffer_); | |
182 read_buffer_.reset(new ReadBuffer); | |
183 DCHECK(!write_buffer_); | |
184 write_buffer_.reset(new WriteBuffer(GetSerializedPlatformHandleSize())); | |
185 | |
186 if (!OnInit()) { | |
187 delegate_ = nullptr; | |
188 message_loop_for_io_ = nullptr; | |
189 read_buffer_.reset(); | |
190 write_buffer_.reset(); | |
191 return false; | |
192 } | |
193 | |
194 IOResult io_result = ScheduleRead(); | |
195 if (io_result != IO_PENDING) { | |
196 // This will notify the delegate about the read failure. Although we're on | |
197 // the I/O thread, don't call it in the nested context. | |
198 message_loop_for_io_->PostTask(FROM_HERE, | |
199 base::Bind(&RawChannel::OnReadCompleted, | |
200 weak_ptr_factory_.GetWeakPtr(), | |
201 io_result, | |
202 0)); | |
203 } | |
204 | |
205 // ScheduleRead() failure is treated as a read failure (by notifying the | |
206 // delegate), not as an init failure. | |
207 return true; | |
208 } | |
209 | |
210 void RawChannel::Shutdown() { | |
211 DCHECK_EQ(base::MessageLoop::current(), message_loop_for_io_); | |
212 | |
213 base::AutoLock locker(write_lock_); | |
214 | |
215 LOG_IF(WARNING, !write_buffer_->message_queue_.empty()) | |
216 << "Shutting down RawChannel with write buffer nonempty"; | |
217 | |
218 // Reset the delegate so that it won't receive further calls. | |
219 delegate_ = nullptr; | |
220 read_stopped_ = true; | |
221 write_stopped_ = true; | |
222 weak_ptr_factory_.InvalidateWeakPtrs(); | |
223 | |
224 OnShutdownNoLock(read_buffer_.Pass(), write_buffer_.Pass()); | |
225 } | |
226 | |
227 // Reminder: This must be thread-safe. | |
228 bool RawChannel::WriteMessage(scoped_ptr<MessageInTransit> message) { | |
229 DCHECK(message); | |
230 | |
231 base::AutoLock locker(write_lock_); | |
232 if (write_stopped_) | |
233 return false; | |
234 | |
235 if (!write_buffer_->message_queue_.empty()) { | |
236 EnqueueMessageNoLock(message.Pass()); | |
237 return true; | |
238 } | |
239 | |
240 EnqueueMessageNoLock(message.Pass()); | |
241 DCHECK_EQ(write_buffer_->data_offset_, 0u); | |
242 | |
243 size_t platform_handles_written = 0; | |
244 size_t bytes_written = 0; | |
245 IOResult io_result = WriteNoLock(&platform_handles_written, &bytes_written); | |
246 if (io_result == IO_PENDING) | |
247 return true; | |
248 | |
249 bool result = OnWriteCompletedNoLock( | |
250 io_result, platform_handles_written, bytes_written); | |
251 if (!result) { | |
252 // Even if we're on the I/O thread, don't call |OnError()| in the nested | |
253 // context. | |
254 message_loop_for_io_->PostTask(FROM_HERE, | |
255 base::Bind(&RawChannel::CallOnError, | |
256 weak_ptr_factory_.GetWeakPtr(), | |
257 Delegate::ERROR_WRITE)); | |
258 } | |
259 | |
260 return result; | |
261 } | |
262 | |
263 // Reminder: This must be thread-safe. | |
264 bool RawChannel::IsWriteBufferEmpty() { | |
265 base::AutoLock locker(write_lock_); | |
266 return write_buffer_->message_queue_.empty(); | |
267 } | |
268 | |
269 void RawChannel::OnReadCompleted(IOResult io_result, size_t bytes_read) { | |
270 DCHECK_EQ(base::MessageLoop::current(), message_loop_for_io_); | |
271 | |
272 if (read_stopped_) { | |
273 NOTREACHED(); | |
274 return; | |
275 } | |
276 | |
277 // Keep reading data in a loop, and dispatch messages if enough data is | |
278 // received. Exit the loop if any of the following happens: | |
279 // - one or more messages were dispatched; | |
280 // - the last read failed, was a partial read or would block; | |
281 // - |Shutdown()| was called. | |
282 do { | |
283 switch (io_result) { | |
284 case IO_SUCCEEDED: | |
285 break; | |
286 case IO_FAILED_SHUTDOWN: | |
287 case IO_FAILED_BROKEN: | |
288 case IO_FAILED_UNKNOWN: | |
289 read_stopped_ = true; | |
290 CallOnError(ReadIOResultToError(io_result)); | |
291 return; | |
292 case IO_PENDING: | |
293 NOTREACHED(); | |
294 return; | |
295 } | |
296 | |
297 read_buffer_->num_valid_bytes_ += bytes_read; | |
298 | |
299 // Dispatch all the messages that we can. | |
300 bool did_dispatch_message = false; | |
301 // Tracks the offset of the first undispatched message in |read_buffer_|. | |
302 // Currently, we copy data to ensure that this is zero at the beginning. | |
303 size_t read_buffer_start = 0; | |
304 size_t remaining_bytes = read_buffer_->num_valid_bytes_; | |
305 size_t message_size; | |
306 // Note that we rely on short-circuit evaluation here: | |
307 // - |read_buffer_start| may be an invalid index into | |
308 // |read_buffer_->buffer_| if |remaining_bytes| is zero. | |
309 // - |message_size| is only valid if |GetNextMessageSize()| returns true. | |
310 // TODO(vtl): Use |message_size| more intelligently (e.g., to request the | |
311 // next read). | |
312 // TODO(vtl): Validate that |message_size| is sane. | |
313 while (remaining_bytes > 0 && MessageInTransit::GetNextMessageSize( | |
314 &read_buffer_->buffer_[read_buffer_start], | |
315 remaining_bytes, | |
316 &message_size) && | |
317 remaining_bytes >= message_size) { | |
318 MessageInTransit::View message_view( | |
319 message_size, &read_buffer_->buffer_[read_buffer_start]); | |
320 DCHECK_EQ(message_view.total_size(), message_size); | |
321 | |
322 const char* error_message = nullptr; | |
323 if (!message_view.IsValid(GetSerializedPlatformHandleSize(), | |
324 &error_message)) { | |
325 DCHECK(error_message); | |
326 LOG(ERROR) << "Received invalid message: " << error_message; | |
327 read_stopped_ = true; | |
328 CallOnError(Delegate::ERROR_READ_BAD_MESSAGE); | |
329 return; | |
330 } | |
331 | |
332 if (message_view.type() == MessageInTransit::kTypeRawChannel) { | |
333 if (!OnReadMessageForRawChannel(message_view)) { | |
334 read_stopped_ = true; | |
335 CallOnError(Delegate::ERROR_READ_BAD_MESSAGE); | |
336 return; | |
337 } | |
338 } else { | |
339 embedder::ScopedPlatformHandleVectorPtr platform_handles; | |
340 if (message_view.transport_data_buffer()) { | |
341 size_t num_platform_handles; | |
342 const void* platform_handle_table; | |
343 TransportData::GetPlatformHandleTable( | |
344 message_view.transport_data_buffer(), | |
345 &num_platform_handles, | |
346 &platform_handle_table); | |
347 | |
348 if (num_platform_handles > 0) { | |
349 platform_handles = | |
350 GetReadPlatformHandles(num_platform_handles, | |
351 platform_handle_table).Pass(); | |
352 if (!platform_handles) { | |
353 LOG(ERROR) << "Invalid number of platform handles received"; | |
354 read_stopped_ = true; | |
355 CallOnError(Delegate::ERROR_READ_BAD_MESSAGE); | |
356 return; | |
357 } | |
358 } | |
359 } | |
360 | |
361 // TODO(vtl): In the case that we aren't expecting any platform handles, | |
362 // for the POSIX implementation, we should confirm that none are stored. | |
363 | |
364 // Dispatch the message. | |
365 DCHECK(delegate_); | |
366 delegate_->OnReadMessage(message_view, platform_handles.Pass()); | |
367 if (read_stopped_) { | |
368 // |Shutdown()| was called in |OnReadMessage()|. | |
369 // TODO(vtl): Add test for this case. | |
370 return; | |
371 } | |
372 } | |
373 | |
374 did_dispatch_message = true; | |
375 | |
376 // Update our state. | |
377 read_buffer_start += message_size; | |
378 remaining_bytes -= message_size; | |
379 } | |
380 | |
381 if (read_buffer_start > 0) { | |
382 // Move data back to start. | |
383 read_buffer_->num_valid_bytes_ = remaining_bytes; | |
384 if (read_buffer_->num_valid_bytes_ > 0) { | |
385 memmove(&read_buffer_->buffer_[0], | |
386 &read_buffer_->buffer_[read_buffer_start], | |
387 remaining_bytes); | |
388 } | |
389 read_buffer_start = 0; | |
390 } | |
391 | |
392 if (read_buffer_->buffer_.size() - read_buffer_->num_valid_bytes_ < | |
393 kReadSize) { | |
394 // Use power-of-2 buffer sizes. | |
395 // TODO(vtl): Make sure the buffer doesn't get too large (and enforce the | |
396 // maximum message size to whatever extent necessary). | |
397 // TODO(vtl): We may often be able to peek at the header and get the real | |
398 // required extra space (which may be much bigger than |kReadSize|). | |
399 size_t new_size = std::max(read_buffer_->buffer_.size(), kReadSize); | |
400 while (new_size < read_buffer_->num_valid_bytes_ + kReadSize) | |
401 new_size *= 2; | |
402 | |
403 // TODO(vtl): It's suboptimal to zero out the fresh memory. | |
404 read_buffer_->buffer_.resize(new_size, 0); | |
405 } | |
406 | |
407 // (1) If we dispatched any messages, stop reading for now (and let the | |
408 // message loop do its thing for another round). | |
409 // TODO(vtl): Is this the behavior we want? (Alternatives: i. Dispatch only | |
410 // a single message. Risks: slower, more complex if we want to avoid lots of | |
411 // copying. ii. Keep reading until there's no more data and dispatch all the | |
412 // messages we can. Risks: starvation of other users of the message loop.) | |
413 // (2) If we didn't max out |kReadSize|, stop reading for now. | |
414 bool schedule_for_later = did_dispatch_message || bytes_read < kReadSize; | |
415 bytes_read = 0; | |
416 io_result = schedule_for_later ? ScheduleRead() : Read(&bytes_read); | |
417 } while (io_result != IO_PENDING); | |
418 } | |
419 | |
420 void RawChannel::OnWriteCompleted(IOResult io_result, | |
421 size_t platform_handles_written, | |
422 size_t bytes_written) { | |
423 DCHECK_EQ(base::MessageLoop::current(), message_loop_for_io_); | |
424 DCHECK_NE(io_result, IO_PENDING); | |
425 | |
426 bool did_fail = false; | |
427 { | |
428 base::AutoLock locker(write_lock_); | |
429 DCHECK_EQ(write_stopped_, write_buffer_->message_queue_.empty()); | |
430 | |
431 if (write_stopped_) { | |
432 NOTREACHED(); | |
433 return; | |
434 } | |
435 | |
436 did_fail = !OnWriteCompletedNoLock( | |
437 io_result, platform_handles_written, bytes_written); | |
438 } | |
439 | |
440 if (did_fail) | |
441 CallOnError(Delegate::ERROR_WRITE); | |
442 } | |
443 | |
444 void RawChannel::EnqueueMessageNoLock(scoped_ptr<MessageInTransit> message) { | |
445 write_lock_.AssertAcquired(); | |
446 write_buffer_->message_queue_.push_back(message.release()); | |
447 } | |
448 | |
449 bool RawChannel::OnReadMessageForRawChannel( | |
450 const MessageInTransit::View& message_view) { | |
451 // No non-implementation specific |RawChannel| control messages. | |
452 LOG(ERROR) << "Invalid control message (subtype " << message_view.subtype() | |
453 << ")"; | |
454 return false; | |
455 } | |
456 | |
457 // static | |
458 RawChannel::Delegate::Error RawChannel::ReadIOResultToError( | |
459 IOResult io_result) { | |
460 switch (io_result) { | |
461 case IO_FAILED_SHUTDOWN: | |
462 return Delegate::ERROR_READ_SHUTDOWN; | |
463 case IO_FAILED_BROKEN: | |
464 return Delegate::ERROR_READ_BROKEN; | |
465 case IO_FAILED_UNKNOWN: | |
466 return Delegate::ERROR_READ_UNKNOWN; | |
467 case IO_SUCCEEDED: | |
468 case IO_PENDING: | |
469 NOTREACHED(); | |
470 break; | |
471 } | |
472 return Delegate::ERROR_READ_UNKNOWN; | |
473 } | |
474 | |
475 void RawChannel::CallOnError(Delegate::Error error) { | |
476 DCHECK_EQ(base::MessageLoop::current(), message_loop_for_io_); | |
477 // TODO(vtl): Add a "write_lock_.AssertNotAcquired()"? | |
478 if (delegate_) | |
479 delegate_->OnError(error); | |
480 } | |
481 | |
482 bool RawChannel::OnWriteCompletedNoLock(IOResult io_result, | |
483 size_t platform_handles_written, | |
484 size_t bytes_written) { | |
485 write_lock_.AssertAcquired(); | |
486 | |
487 DCHECK(!write_stopped_); | |
488 DCHECK(!write_buffer_->message_queue_.empty()); | |
489 | |
490 if (io_result == IO_SUCCEEDED) { | |
491 write_buffer_->platform_handles_offset_ += platform_handles_written; | |
492 write_buffer_->data_offset_ += bytes_written; | |
493 | |
494 MessageInTransit* message = write_buffer_->message_queue_.front(); | |
495 if (write_buffer_->data_offset_ >= message->total_size()) { | |
496 // Complete write. | |
497 CHECK_EQ(write_buffer_->data_offset_, message->total_size()); | |
498 write_buffer_->message_queue_.pop_front(); | |
499 delete message; | |
500 write_buffer_->platform_handles_offset_ = 0; | |
501 write_buffer_->data_offset_ = 0; | |
502 | |
503 if (write_buffer_->message_queue_.empty()) | |
504 return true; | |
505 } | |
506 | |
507 // Schedule the next write. | |
508 io_result = ScheduleWriteNoLock(); | |
509 if (io_result == IO_PENDING) | |
510 return true; | |
511 DCHECK_NE(io_result, IO_SUCCEEDED); | |
512 } | |
513 | |
514 write_stopped_ = true; | |
515 STLDeleteElements(&write_buffer_->message_queue_); | |
516 write_buffer_->platform_handles_offset_ = 0; | |
517 write_buffer_->data_offset_ = 0; | |
518 return false; | |
519 } | |
520 | |
521 } // namespace system | |
522 } // namespace mojo | |
OLD | NEW |