Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(110)

Side by Side Diff: mojo/edk/system/raw_channel.cc

Issue 623883002: Revert "Move mojo edk into mojo/edk" (Closed) Base URL: https://chromium.googlesource.com/chromium/src.git@master
Patch Set: Created 6 years, 2 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View unified diff | Download patch
« no previous file with comments | « mojo/edk/system/raw_channel.h ('k') | mojo/edk/system/raw_channel_posix.cc » ('j') | no next file with comments »
Toggle Intra-line Diffs ('i') | Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
OLDNEW
(Empty)
1 // Copyright 2014 The Chromium Authors. All rights reserved.
2 // Use of this source code is governed by a BSD-style license that can be
3 // found in the LICENSE file.
4
5 #include "mojo/edk/system/raw_channel.h"
6
7 #include <string.h>
8
9 #include <algorithm>
10
11 #include "base/bind.h"
12 #include "base/location.h"
13 #include "base/logging.h"
14 #include "base/message_loop/message_loop.h"
15 #include "base/stl_util.h"
16 #include "mojo/edk/system/message_in_transit.h"
17 #include "mojo/edk/system/transport_data.h"
18
19 namespace mojo {
20 namespace system {
21
22 const size_t kReadSize = 4096;
23
24 // RawChannel::ReadBuffer ------------------------------------------------------
25
26 RawChannel::ReadBuffer::ReadBuffer() : buffer_(kReadSize), num_valid_bytes_(0) {
27 }
28
29 RawChannel::ReadBuffer::~ReadBuffer() {
30 }
31
32 void RawChannel::ReadBuffer::GetBuffer(char** addr, size_t* size) {
33 DCHECK_GE(buffer_.size(), num_valid_bytes_ + kReadSize);
34 *addr = &buffer_[0] + num_valid_bytes_;
35 *size = kReadSize;
36 }
37
38 // RawChannel::WriteBuffer -----------------------------------------------------
39
40 RawChannel::WriteBuffer::WriteBuffer(size_t serialized_platform_handle_size)
41 : serialized_platform_handle_size_(serialized_platform_handle_size),
42 platform_handles_offset_(0),
43 data_offset_(0) {
44 }
45
46 RawChannel::WriteBuffer::~WriteBuffer() {
47 STLDeleteElements(&message_queue_);
48 }
49
50 bool RawChannel::WriteBuffer::HavePlatformHandlesToSend() const {
51 if (message_queue_.empty())
52 return false;
53
54 const TransportData* transport_data =
55 message_queue_.front()->transport_data();
56 if (!transport_data)
57 return false;
58
59 const embedder::PlatformHandleVector* all_platform_handles =
60 transport_data->platform_handles();
61 if (!all_platform_handles) {
62 DCHECK_EQ(platform_handles_offset_, 0u);
63 return false;
64 }
65 if (platform_handles_offset_ >= all_platform_handles->size()) {
66 DCHECK_EQ(platform_handles_offset_, all_platform_handles->size());
67 return false;
68 }
69
70 return true;
71 }
72
73 void RawChannel::WriteBuffer::GetPlatformHandlesToSend(
74 size_t* num_platform_handles,
75 embedder::PlatformHandle** platform_handles,
76 void** serialization_data) {
77 DCHECK(HavePlatformHandlesToSend());
78
79 TransportData* transport_data = message_queue_.front()->transport_data();
80 embedder::PlatformHandleVector* all_platform_handles =
81 transport_data->platform_handles();
82 *num_platform_handles =
83 all_platform_handles->size() - platform_handles_offset_;
84 *platform_handles = &(*all_platform_handles)[platform_handles_offset_];
85 size_t serialization_data_offset =
86 transport_data->platform_handle_table_offset();
87 DCHECK_GT(serialization_data_offset, 0u);
88 serialization_data_offset +=
89 platform_handles_offset_ * serialized_platform_handle_size_;
90 *serialization_data =
91 static_cast<char*>(transport_data->buffer()) + serialization_data_offset;
92 }
93
94 void RawChannel::WriteBuffer::GetBuffers(std::vector<Buffer>* buffers) const {
95 buffers->clear();
96
97 if (message_queue_.empty())
98 return;
99
100 MessageInTransit* message = message_queue_.front();
101 DCHECK_LT(data_offset_, message->total_size());
102 size_t bytes_to_write = message->total_size() - data_offset_;
103
104 size_t transport_data_buffer_size =
105 message->transport_data() ? message->transport_data()->buffer_size() : 0;
106
107 if (!transport_data_buffer_size) {
108 // Only write from the main buffer.
109 DCHECK_LT(data_offset_, message->main_buffer_size());
110 DCHECK_LE(bytes_to_write, message->main_buffer_size());
111 Buffer buffer = {
112 static_cast<const char*>(message->main_buffer()) + data_offset_,
113 bytes_to_write};
114 buffers->push_back(buffer);
115 return;
116 }
117
118 if (data_offset_ >= message->main_buffer_size()) {
119 // Only write from the transport data buffer.
120 DCHECK_LT(data_offset_ - message->main_buffer_size(),
121 transport_data_buffer_size);
122 DCHECK_LE(bytes_to_write, transport_data_buffer_size);
123 Buffer buffer = {
124 static_cast<const char*>(message->transport_data()->buffer()) +
125 (data_offset_ - message->main_buffer_size()),
126 bytes_to_write};
127 buffers->push_back(buffer);
128 return;
129 }
130
131 // TODO(vtl): We could actually send out buffers from multiple messages, with
132 // the "stopping" condition being reaching a message with platform handles
133 // attached.
134
135 // Write from both buffers.
136 DCHECK_EQ(
137 bytes_to_write,
138 message->main_buffer_size() - data_offset_ + transport_data_buffer_size);
139 Buffer buffer1 = {
140 static_cast<const char*>(message->main_buffer()) + data_offset_,
141 message->main_buffer_size() - data_offset_};
142 buffers->push_back(buffer1);
143 Buffer buffer2 = {
144 static_cast<const char*>(message->transport_data()->buffer()),
145 transport_data_buffer_size};
146 buffers->push_back(buffer2);
147 }
148
149 // RawChannel ------------------------------------------------------------------
150
151 RawChannel::RawChannel()
152 : message_loop_for_io_(nullptr),
153 delegate_(nullptr),
154 read_stopped_(false),
155 write_stopped_(false),
156 weak_ptr_factory_(this) {
157 }
158
159 RawChannel::~RawChannel() {
160 DCHECK(!read_buffer_);
161 DCHECK(!write_buffer_);
162
163 // No need to take the |write_lock_| here -- if there are still weak pointers
164 // outstanding, then we're hosed anyway (since we wouldn't be able to
165 // invalidate them cleanly, since we might not be on the I/O thread).
166 DCHECK(!weak_ptr_factory_.HasWeakPtrs());
167 }
168
169 bool RawChannel::Init(Delegate* delegate) {
170 DCHECK(delegate);
171
172 DCHECK(!delegate_);
173 delegate_ = delegate;
174
175 CHECK_EQ(base::MessageLoop::current()->type(), base::MessageLoop::TYPE_IO);
176 DCHECK(!message_loop_for_io_);
177 message_loop_for_io_ =
178 static_cast<base::MessageLoopForIO*>(base::MessageLoop::current());
179
180 // No need to take the lock. No one should be using us yet.
181 DCHECK(!read_buffer_);
182 read_buffer_.reset(new ReadBuffer);
183 DCHECK(!write_buffer_);
184 write_buffer_.reset(new WriteBuffer(GetSerializedPlatformHandleSize()));
185
186 if (!OnInit()) {
187 delegate_ = nullptr;
188 message_loop_for_io_ = nullptr;
189 read_buffer_.reset();
190 write_buffer_.reset();
191 return false;
192 }
193
194 IOResult io_result = ScheduleRead();
195 if (io_result != IO_PENDING) {
196 // This will notify the delegate about the read failure. Although we're on
197 // the I/O thread, don't call it in the nested context.
198 message_loop_for_io_->PostTask(FROM_HERE,
199 base::Bind(&RawChannel::OnReadCompleted,
200 weak_ptr_factory_.GetWeakPtr(),
201 io_result,
202 0));
203 }
204
205 // ScheduleRead() failure is treated as a read failure (by notifying the
206 // delegate), not as an init failure.
207 return true;
208 }
209
210 void RawChannel::Shutdown() {
211 DCHECK_EQ(base::MessageLoop::current(), message_loop_for_io_);
212
213 base::AutoLock locker(write_lock_);
214
215 LOG_IF(WARNING, !write_buffer_->message_queue_.empty())
216 << "Shutting down RawChannel with write buffer nonempty";
217
218 // Reset the delegate so that it won't receive further calls.
219 delegate_ = nullptr;
220 read_stopped_ = true;
221 write_stopped_ = true;
222 weak_ptr_factory_.InvalidateWeakPtrs();
223
224 OnShutdownNoLock(read_buffer_.Pass(), write_buffer_.Pass());
225 }
226
227 // Reminder: This must be thread-safe.
228 bool RawChannel::WriteMessage(scoped_ptr<MessageInTransit> message) {
229 DCHECK(message);
230
231 base::AutoLock locker(write_lock_);
232 if (write_stopped_)
233 return false;
234
235 if (!write_buffer_->message_queue_.empty()) {
236 EnqueueMessageNoLock(message.Pass());
237 return true;
238 }
239
240 EnqueueMessageNoLock(message.Pass());
241 DCHECK_EQ(write_buffer_->data_offset_, 0u);
242
243 size_t platform_handles_written = 0;
244 size_t bytes_written = 0;
245 IOResult io_result = WriteNoLock(&platform_handles_written, &bytes_written);
246 if (io_result == IO_PENDING)
247 return true;
248
249 bool result = OnWriteCompletedNoLock(
250 io_result, platform_handles_written, bytes_written);
251 if (!result) {
252 // Even if we're on the I/O thread, don't call |OnError()| in the nested
253 // context.
254 message_loop_for_io_->PostTask(FROM_HERE,
255 base::Bind(&RawChannel::CallOnError,
256 weak_ptr_factory_.GetWeakPtr(),
257 Delegate::ERROR_WRITE));
258 }
259
260 return result;
261 }
262
263 // Reminder: This must be thread-safe.
264 bool RawChannel::IsWriteBufferEmpty() {
265 base::AutoLock locker(write_lock_);
266 return write_buffer_->message_queue_.empty();
267 }
268
269 void RawChannel::OnReadCompleted(IOResult io_result, size_t bytes_read) {
270 DCHECK_EQ(base::MessageLoop::current(), message_loop_for_io_);
271
272 if (read_stopped_) {
273 NOTREACHED();
274 return;
275 }
276
277 // Keep reading data in a loop, and dispatch messages if enough data is
278 // received. Exit the loop if any of the following happens:
279 // - one or more messages were dispatched;
280 // - the last read failed, was a partial read or would block;
281 // - |Shutdown()| was called.
282 do {
283 switch (io_result) {
284 case IO_SUCCEEDED:
285 break;
286 case IO_FAILED_SHUTDOWN:
287 case IO_FAILED_BROKEN:
288 case IO_FAILED_UNKNOWN:
289 read_stopped_ = true;
290 CallOnError(ReadIOResultToError(io_result));
291 return;
292 case IO_PENDING:
293 NOTREACHED();
294 return;
295 }
296
297 read_buffer_->num_valid_bytes_ += bytes_read;
298
299 // Dispatch all the messages that we can.
300 bool did_dispatch_message = false;
301 // Tracks the offset of the first undispatched message in |read_buffer_|.
302 // Currently, we copy data to ensure that this is zero at the beginning.
303 size_t read_buffer_start = 0;
304 size_t remaining_bytes = read_buffer_->num_valid_bytes_;
305 size_t message_size;
306 // Note that we rely on short-circuit evaluation here:
307 // - |read_buffer_start| may be an invalid index into
308 // |read_buffer_->buffer_| if |remaining_bytes| is zero.
309 // - |message_size| is only valid if |GetNextMessageSize()| returns true.
310 // TODO(vtl): Use |message_size| more intelligently (e.g., to request the
311 // next read).
312 // TODO(vtl): Validate that |message_size| is sane.
313 while (remaining_bytes > 0 && MessageInTransit::GetNextMessageSize(
314 &read_buffer_->buffer_[read_buffer_start],
315 remaining_bytes,
316 &message_size) &&
317 remaining_bytes >= message_size) {
318 MessageInTransit::View message_view(
319 message_size, &read_buffer_->buffer_[read_buffer_start]);
320 DCHECK_EQ(message_view.total_size(), message_size);
321
322 const char* error_message = nullptr;
323 if (!message_view.IsValid(GetSerializedPlatformHandleSize(),
324 &error_message)) {
325 DCHECK(error_message);
326 LOG(ERROR) << "Received invalid message: " << error_message;
327 read_stopped_ = true;
328 CallOnError(Delegate::ERROR_READ_BAD_MESSAGE);
329 return;
330 }
331
332 if (message_view.type() == MessageInTransit::kTypeRawChannel) {
333 if (!OnReadMessageForRawChannel(message_view)) {
334 read_stopped_ = true;
335 CallOnError(Delegate::ERROR_READ_BAD_MESSAGE);
336 return;
337 }
338 } else {
339 embedder::ScopedPlatformHandleVectorPtr platform_handles;
340 if (message_view.transport_data_buffer()) {
341 size_t num_platform_handles;
342 const void* platform_handle_table;
343 TransportData::GetPlatformHandleTable(
344 message_view.transport_data_buffer(),
345 &num_platform_handles,
346 &platform_handle_table);
347
348 if (num_platform_handles > 0) {
349 platform_handles =
350 GetReadPlatformHandles(num_platform_handles,
351 platform_handle_table).Pass();
352 if (!platform_handles) {
353 LOG(ERROR) << "Invalid number of platform handles received";
354 read_stopped_ = true;
355 CallOnError(Delegate::ERROR_READ_BAD_MESSAGE);
356 return;
357 }
358 }
359 }
360
361 // TODO(vtl): In the case that we aren't expecting any platform handles,
362 // for the POSIX implementation, we should confirm that none are stored.
363
364 // Dispatch the message.
365 DCHECK(delegate_);
366 delegate_->OnReadMessage(message_view, platform_handles.Pass());
367 if (read_stopped_) {
368 // |Shutdown()| was called in |OnReadMessage()|.
369 // TODO(vtl): Add test for this case.
370 return;
371 }
372 }
373
374 did_dispatch_message = true;
375
376 // Update our state.
377 read_buffer_start += message_size;
378 remaining_bytes -= message_size;
379 }
380
381 if (read_buffer_start > 0) {
382 // Move data back to start.
383 read_buffer_->num_valid_bytes_ = remaining_bytes;
384 if (read_buffer_->num_valid_bytes_ > 0) {
385 memmove(&read_buffer_->buffer_[0],
386 &read_buffer_->buffer_[read_buffer_start],
387 remaining_bytes);
388 }
389 read_buffer_start = 0;
390 }
391
392 if (read_buffer_->buffer_.size() - read_buffer_->num_valid_bytes_ <
393 kReadSize) {
394 // Use power-of-2 buffer sizes.
395 // TODO(vtl): Make sure the buffer doesn't get too large (and enforce the
396 // maximum message size to whatever extent necessary).
397 // TODO(vtl): We may often be able to peek at the header and get the real
398 // required extra space (which may be much bigger than |kReadSize|).
399 size_t new_size = std::max(read_buffer_->buffer_.size(), kReadSize);
400 while (new_size < read_buffer_->num_valid_bytes_ + kReadSize)
401 new_size *= 2;
402
403 // TODO(vtl): It's suboptimal to zero out the fresh memory.
404 read_buffer_->buffer_.resize(new_size, 0);
405 }
406
407 // (1) If we dispatched any messages, stop reading for now (and let the
408 // message loop do its thing for another round).
409 // TODO(vtl): Is this the behavior we want? (Alternatives: i. Dispatch only
410 // a single message. Risks: slower, more complex if we want to avoid lots of
411 // copying. ii. Keep reading until there's no more data and dispatch all the
412 // messages we can. Risks: starvation of other users of the message loop.)
413 // (2) If we didn't max out |kReadSize|, stop reading for now.
414 bool schedule_for_later = did_dispatch_message || bytes_read < kReadSize;
415 bytes_read = 0;
416 io_result = schedule_for_later ? ScheduleRead() : Read(&bytes_read);
417 } while (io_result != IO_PENDING);
418 }
419
420 void RawChannel::OnWriteCompleted(IOResult io_result,
421 size_t platform_handles_written,
422 size_t bytes_written) {
423 DCHECK_EQ(base::MessageLoop::current(), message_loop_for_io_);
424 DCHECK_NE(io_result, IO_PENDING);
425
426 bool did_fail = false;
427 {
428 base::AutoLock locker(write_lock_);
429 DCHECK_EQ(write_stopped_, write_buffer_->message_queue_.empty());
430
431 if (write_stopped_) {
432 NOTREACHED();
433 return;
434 }
435
436 did_fail = !OnWriteCompletedNoLock(
437 io_result, platform_handles_written, bytes_written);
438 }
439
440 if (did_fail)
441 CallOnError(Delegate::ERROR_WRITE);
442 }
443
444 void RawChannel::EnqueueMessageNoLock(scoped_ptr<MessageInTransit> message) {
445 write_lock_.AssertAcquired();
446 write_buffer_->message_queue_.push_back(message.release());
447 }
448
449 bool RawChannel::OnReadMessageForRawChannel(
450 const MessageInTransit::View& message_view) {
451 // No non-implementation specific |RawChannel| control messages.
452 LOG(ERROR) << "Invalid control message (subtype " << message_view.subtype()
453 << ")";
454 return false;
455 }
456
457 // static
458 RawChannel::Delegate::Error RawChannel::ReadIOResultToError(
459 IOResult io_result) {
460 switch (io_result) {
461 case IO_FAILED_SHUTDOWN:
462 return Delegate::ERROR_READ_SHUTDOWN;
463 case IO_FAILED_BROKEN:
464 return Delegate::ERROR_READ_BROKEN;
465 case IO_FAILED_UNKNOWN:
466 return Delegate::ERROR_READ_UNKNOWN;
467 case IO_SUCCEEDED:
468 case IO_PENDING:
469 NOTREACHED();
470 break;
471 }
472 return Delegate::ERROR_READ_UNKNOWN;
473 }
474
475 void RawChannel::CallOnError(Delegate::Error error) {
476 DCHECK_EQ(base::MessageLoop::current(), message_loop_for_io_);
477 // TODO(vtl): Add a "write_lock_.AssertNotAcquired()"?
478 if (delegate_)
479 delegate_->OnError(error);
480 }
481
482 bool RawChannel::OnWriteCompletedNoLock(IOResult io_result,
483 size_t platform_handles_written,
484 size_t bytes_written) {
485 write_lock_.AssertAcquired();
486
487 DCHECK(!write_stopped_);
488 DCHECK(!write_buffer_->message_queue_.empty());
489
490 if (io_result == IO_SUCCEEDED) {
491 write_buffer_->platform_handles_offset_ += platform_handles_written;
492 write_buffer_->data_offset_ += bytes_written;
493
494 MessageInTransit* message = write_buffer_->message_queue_.front();
495 if (write_buffer_->data_offset_ >= message->total_size()) {
496 // Complete write.
497 CHECK_EQ(write_buffer_->data_offset_, message->total_size());
498 write_buffer_->message_queue_.pop_front();
499 delete message;
500 write_buffer_->platform_handles_offset_ = 0;
501 write_buffer_->data_offset_ = 0;
502
503 if (write_buffer_->message_queue_.empty())
504 return true;
505 }
506
507 // Schedule the next write.
508 io_result = ScheduleWriteNoLock();
509 if (io_result == IO_PENDING)
510 return true;
511 DCHECK_NE(io_result, IO_SUCCEEDED);
512 }
513
514 write_stopped_ = true;
515 STLDeleteElements(&write_buffer_->message_queue_);
516 write_buffer_->platform_handles_offset_ = 0;
517 write_buffer_->data_offset_ = 0;
518 return false;
519 }
520
521 } // namespace system
522 } // namespace mojo
OLDNEW
« no previous file with comments | « mojo/edk/system/raw_channel.h ('k') | mojo/edk/system/raw_channel_posix.cc » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698