Index: mojo/edk/system/watcher_dispatcher.cc |
diff --git a/mojo/edk/system/watcher_dispatcher.cc b/mojo/edk/system/watcher_dispatcher.cc |
new file mode 100644 |
index 0000000000000000000000000000000000000000..e4cc875580ad3f82322cf6559dc3f4ac677d1e71 |
--- /dev/null |
+++ b/mojo/edk/system/watcher_dispatcher.cc |
@@ -0,0 +1,219 @@ |
+// Copyright 2017 The Chromium Authors. All rights reserved. |
+// Use of this source code is governed by a BSD-style license that can be |
+// found in the LICENSE file. |
+ |
+#include "mojo/edk/system/watcher_dispatcher.h" |
+ |
+#include <limits> |
+#include <map> |
+ |
+#include "base/macros.h" |
+#include "base/memory/ptr_util.h" |
+#include "mojo/edk/system/watch.h" |
+ |
+namespace mojo { |
+namespace edk { |
+ |
+WatcherDispatcher::WatcherDispatcher(MojoWatcherCallback callback) |
+ : callback_(callback) {} |
+ |
+void WatcherDispatcher::NotifyHandleState(Dispatcher* dispatcher, |
+ const HandleSignalsState& state) { |
+ base::AutoLock lock(lock_); |
+ auto it = watched_handles_.find(dispatcher); |
+ if (it == watched_handles_.end()) |
+ return; |
+ |
+ // Maybe fire a notification to the watch assoicated with this dispatcher, |
+ // provided we're armed it cares about the new state. |
+ if (it->second->NotifyState(state, armed_)) { |
+ ready_watches_.insert(it->second.get()); |
+ |
+ // If we were armed and got here, we notified the watch. Disarm. |
+ armed_ = false; |
+ } else { |
+ ready_watches_.erase(it->second.get()); |
+ } |
+} |
+ |
+void WatcherDispatcher::NotifyHandleClosed(Dispatcher* dispatcher) { |
+ scoped_refptr<Watch> watch; |
+ { |
+ base::AutoLock lock(lock_); |
+ auto it = watched_handles_.find(dispatcher); |
+ if (it == watched_handles_.end()) |
+ return; |
+ |
+ watch = std::move(it->second); |
+ |
+ // Wipe out all state associated with the closed dispatcher. |
+ watches_.erase(watch->context()); |
+ ready_watches_.erase(watch.get()); |
+ watched_handles_.erase(it); |
+ } |
+ |
+ // NOTE: It's important that this is called outside of |lock_| since it |
+ // acquires internal Watch locks. |
+ watch->Cancel(); |
+} |
+ |
+void WatcherDispatcher::InvokeWatchCallback( |
+ uintptr_t context, |
+ MojoResult result, |
+ const HandleSignalsState& state, |
+ MojoWatcherNotificationFlags flags) { |
+ { |
+ // We avoid holding the lock during dispatch. It's OK for notification |
+ // callbacks to close this watcher, and it's OK for notifications to race |
+ // with closure, if for example the watcher is closed from another thread |
+ // between this test and the invocation of |callback_| below. |
+ // |
+ // Because cancellation synchronously blocks all future notifications, and |
+ // because notifications themselves are mutually exclusive for any given |
+ // context, we still guarantee that a single MOJO_RESULT_CANCELLED result |
+ // is the last notification received for any given context. |
+ // |
+ // This guarantee is sufficient to make safe, synchronized, per-context |
+ // state management possible in user code. |
+ base::AutoLock lock(lock_); |
+ if (closed_ && result != MOJO_RESULT_CANCELLED) |
+ return; |
+ } |
+ |
+ callback_(context, result, static_cast<MojoHandleSignalsState>(state), flags); |
+} |
+ |
+Dispatcher::Type WatcherDispatcher::GetType() const { |
+ return Type::WATCHER; |
+} |
+ |
+MojoResult WatcherDispatcher::Close() { |
+ // We swap out all the watched handle information onto the stack so we can |
+ // call into their dispatchers without our own lock held. |
+ std::map<uintptr_t, scoped_refptr<Watch>> watches; |
+ { |
+ base::AutoLock lock(lock_); |
+ DCHECK(!closed_); |
+ closed_ = true; |
+ std::swap(watches, watches_); |
+ watched_handles_.clear(); |
+ } |
+ |
+ // Remove all refs from our watched dispatchers and fire cancellations. |
+ for (auto& entry : watches) { |
+ entry.second->dispatcher()->RemoveWatcherRef(this, entry.first); |
+ entry.second->Cancel(); |
+ } |
+ |
+ return MOJO_RESULT_OK; |
+} |
+ |
+MojoResult WatcherDispatcher::WatchDispatcher( |
+ scoped_refptr<Dispatcher> dispatcher, |
+ MojoHandleSignals signals, |
+ uintptr_t context) { |
+ // NOTE: Because it's critical to avoid acquiring any other dispatcher locks |
+ // while |lock_| is held, we defer adding oursevles to the dispatcher until |
+ // after we've updated all our own relevant state and released |lock_|. |
+ { |
+ base::AutoLock lock(lock_); |
+ if (watches_.count(context) || watched_handles_.count(dispatcher.get())) |
yzshen1
2017/03/13 20:21:18
[just to double check] IIUC, we want |context| to
Ken Rockot(use gerrit already)
2017/03/13 22:00:10
Correct. Unlike the WaitSet API, I didn't want wat
|
+ return MOJO_RESULT_ALREADY_EXISTS; |
+ |
+ scoped_refptr<Watch> watch = new Watch(this, dispatcher, context, signals); |
+ watches_.insert({context, watch}); |
+ auto result = |
+ watched_handles_.insert(std::make_pair(dispatcher.get(), watch)); |
+ DCHECK(result.second); |
+ } |
+ |
+ MojoResult rv = dispatcher->AddWatcherRef(this, context); |
+ if (rv != MOJO_RESULT_OK) { |
+ // Oops. This was not a valid handle to watch. Undo the above work and |
+ // fail gracefully. |
+ base::AutoLock lock(lock_); |
+ watches_.erase(context); |
+ watched_handles_.erase(dispatcher.get()); |
+ return rv; |
+ } |
+ |
+ return MOJO_RESULT_OK; |
+} |
+ |
+MojoResult WatcherDispatcher::CancelWatch(uintptr_t context) { |
+ // We may remove the last stored ref to the Watch below, so we retain |
+ // a reference on the stack. |
+ scoped_refptr<Watch> watch; |
+ { |
+ base::AutoLock lock(lock_); |
+ auto it = watches_.find(context); |
+ if (it == watches_.end()) |
+ return MOJO_RESULT_NOT_FOUND; |
+ watch = it->second; |
+ watches_.erase(it); |
+ } |
+ |
+ // Mark the watch as cancelled so no further notifications get through. |
+ watch->Cancel(); |
+ |
+ // We remove the watcher ref for this context before updating any more |
+ // internal watcher state, ensuring that we don't receiving further |
+ // notifications for this context. |
+ watch->dispatcher()->RemoveWatcherRef(this, context); |
+ |
+ { |
+ base::AutoLock lock(lock_); |
+ auto handle_it = watched_handles_.find(watch->dispatcher().get()); |
+ DCHECK(handle_it != watched_handles_.end()); |
+ ready_watches_.erase(handle_it->second.get()); |
+ watched_handles_.erase(handle_it); |
+ } |
+ |
+ return MOJO_RESULT_OK; |
+} |
+ |
+MojoResult WatcherDispatcher::Arm( |
+ uint32_t* num_ready_contexts, |
+ uintptr_t* ready_contexts, |
+ MojoResult* ready_results, |
+ MojoHandleSignalsState* ready_signals_states) { |
+ base::AutoLock lock(lock_); |
+ if (num_ready_contexts && |
+ (!ready_contexts || !ready_results || !ready_signals_states)) { |
+ return MOJO_RESULT_INVALID_ARGUMENT; |
+ } |
+ |
+ if (watched_handles_.empty()) |
+ return MOJO_RESULT_NOT_FOUND; |
+ |
+ if (ready_watches_.empty()) { |
+ // Fast path: No watches are ready to notify, so we're done. |
+ armed_ = true; |
+ return MOJO_RESULT_OK; |
+ } |
+ |
+ if (num_ready_contexts) { |
+ uint32_t max_ready_contexts = *num_ready_contexts; |
+ *num_ready_contexts = 0; |
+ for (auto& entry : watched_handles_) { |
+ if (max_ready_contexts == *num_ready_contexts) |
+ break; |
+ |
+ const Watch* watch = entry.second.get(); |
+ if (!watch->ready()) |
+ continue; |
+ |
+ *(ready_contexts++) = watch->context(); |
+ *(ready_results++) = watch->last_known_result(); |
+ *(ready_signals_states++) = watch->last_known_signals_state(); |
+ ++(*num_ready_contexts); |
+ } |
+ } |
+ |
+ return MOJO_RESULT_FAILED_PRECONDITION; |
+} |
+ |
+WatcherDispatcher::~WatcherDispatcher() {} |
+ |
+} // namespace edk |
+} // namespace mojo |