Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(230)

Side by Side Diff: base/task_scheduler/scheduler_single_thread_task_runner_manager.cc

Issue 2726073002: WILL BE MERGED Change Ownership of Sequence to the Single Thread SchedulerWorker Delegate (Closed)
Patch Set: Created 3 years, 9 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View unified diff | Download patch
OLDNEW
1 // Copyright 2017 The Chromium Authors. All rights reserved. 1 // Copyright 2017 The Chromium Authors. All rights reserved.
2 // Use of this source code is governed by a BSD-style license that can be 2 // Use of this source code is governed by a BSD-style license that can be
3 // found in the LICENSE file. 3 // found in the LICENSE file.
4 4
5 #include "base/task_scheduler/scheduler_single_thread_task_runner_manager.h" 5 #include "base/task_scheduler/scheduler_single_thread_task_runner_manager.h"
6 6
7 #include <algorithm> 7 #include <algorithm>
8 #include <memory> 8 #include <memory>
9 #include <string> 9 #include <string>
10 10
(...skipping 46 matching lines...) Expand 10 before | Expand all | Expand 10 after
57 : thread_name_(thread_name) {} 57 : thread_name_(thread_name) {}
58 58
59 // SchedulerWorker::Delegate: 59 // SchedulerWorker::Delegate:
60 void OnMainEntry(SchedulerWorker* worker) override { 60 void OnMainEntry(SchedulerWorker* worker) override {
61 thread_ref_checker_.Set(); 61 thread_ref_checker_.Set();
62 PlatformThread::SetName(thread_name_); 62 PlatformThread::SetName(thread_name_);
63 } 63 }
64 64
65 scoped_refptr<Sequence> GetWork(SchedulerWorker* worker) override { 65 scoped_refptr<Sequence> GetWork(SchedulerWorker* worker) override {
66 AutoSchedulerLock auto_lock(sequence_lock_); 66 AutoSchedulerLock auto_lock(sequence_lock_);
67 return std::move(sequence_); 67 bool has_work = has_work_;
68 has_work_ = false;
69 return has_work ? sequence_ : nullptr;
68 } 70 }
69 71
70 void DidRunTask() override {} 72 void DidRunTask() override {}
71 73
72 void ReEnqueueSequence(scoped_refptr<Sequence> sequence) override { 74 void ReEnqueueSequence(scoped_refptr<Sequence> sequence) override {
73 AutoSchedulerLock auto_lock(sequence_lock_); 75 AutoSchedulerLock auto_lock(sequence_lock_);
74 DCHECK(!sequence_); 76 DCHECK_EQ(sequence, sequence_);
75 sequence_ = std::move(sequence); 77 has_work_ = true;
76 } 78 }
77 79
78 TimeDelta GetSleepTimeout() override { return TimeDelta::Max(); } 80 TimeDelta GetSleepTimeout() override { return TimeDelta::Max(); }
79 81
80 bool CanDetach(SchedulerWorker* worker) override { return false; } 82 bool CanDetach(SchedulerWorker* worker) override { return false; }
81 83
82 void OnDetach() override { NOTREACHED(); } 84 void OnDetach() override { NOTREACHED(); }
83 85
84 bool RunsTasksOnCurrentThread() { 86 bool RunsTasksOnCurrentThread() {
85 // We check the thread ref instead of the sequence for the benefit of COM 87 // We check the thread ref instead of the sequence for the benefit of COM
86 // callbacks which may execute without a sequence context. 88 // callbacks which may execute without a sequence context.
87 return thread_ref_checker_.IsCurrentThreadSameAsSetThread(); 89 return thread_ref_checker_.IsCurrentThreadSameAsSetThread();
88 } 90 }
89 91
92 void OnMainExit() override {
93 // To reclaim skipped tasks on shutdown, we null out the sequence to allow
94 // the tasks to destroy themselves.
fdoray 2017/03/02 14:37:48 DCHECK(sequence->HasOneRef());
robliao 2017/03/03 04:24:17 In general for refcounted things, it's a dangerous
95 sequence_ = nullptr;
96 }
97
98 // SchedulerWorkerDelegate:
99 Sequence* sequence() { return sequence_.get(); }
100
90 private: 101 private:
91 const std::string thread_name_; 102 const std::string thread_name_;
92 103
93 // Synchronizes access to |sequence_| and handles the fact that 104 // Synchronizes access to |sequence_| and |has_work_|.
94 // ReEnqueueSequence() is called on both the worker thread for reenqueuing
95 // the sequence and off of the worker thread to seed the sequence for
96 // GetWork().
97 SchedulerLock sequence_lock_; 105 SchedulerLock sequence_lock_;
98 scoped_refptr<Sequence> sequence_; 106 scoped_refptr<Sequence> sequence_ = new Sequence;
107 bool has_work_ = false;
99 108
100 AtomicThreadRefChecker thread_ref_checker_; 109 AtomicThreadRefChecker thread_ref_checker_;
101 110
102 DISALLOW_COPY_AND_ASSIGN(SchedulerWorkerDelegate); 111 DISALLOW_COPY_AND_ASSIGN(SchedulerWorkerDelegate);
103 }; 112 };
104 113
105 } // namespace 114 } // namespace
106 115
107 class SchedulerSingleThreadTaskRunnerManager::SchedulerSingleThreadTaskRunner 116 class SchedulerSingleThreadTaskRunnerManager::SchedulerSingleThreadTaskRunner
108 : public SingleThreadTaskRunner { 117 : public SingleThreadTaskRunner {
109 public: 118 public:
110 // Constructs a SchedulerSingleThreadTaskRunner that indirectly controls the 119 // Constructs a SchedulerSingleThreadTaskRunner that indirectly controls the
111 // lifetime of a dedicated |worker| for |traits|. 120 // lifetime of a dedicated |worker| for |traits|.
112 SchedulerSingleThreadTaskRunner( 121 SchedulerSingleThreadTaskRunner(
113 SchedulerSingleThreadTaskRunnerManager* const outer, 122 SchedulerSingleThreadTaskRunnerManager* const outer,
114 const TaskTraits& traits, 123 const TaskTraits& traits,
115 SchedulerWorker* worker) 124 SchedulerWorker* worker)
116 : outer_(outer), traits_(traits), worker_(worker) { 125 : outer_(outer), traits_(traits), worker_(worker) {
117 DCHECK(outer_); 126 DCHECK(outer_);
118 DCHECK(worker_); 127 DCHECK(worker_);
119 } 128 }
120 129
121 // SingleThreadTaskRunner: 130 // SingleThreadTaskRunner:
122 bool PostDelayedTask(const tracked_objects::Location& from_here, 131 bool PostDelayedTask(const tracked_objects::Location& from_here,
123 const Closure& closure, 132 const Closure& closure,
124 TimeDelta delay) override; 133 TimeDelta delay) override {
134 auto task = MakeUnique<Task>(from_here, closure, traits_, delay);
135 task->single_thread_task_runner_ref = this;
136
137 if (!outer_->task_tracker_->WillPostTask(task.get()))
138 return false;
139
140 if (task->delayed_run_time.is_null()) {
141 PostTaskNow(std::move(task));
142 } else {
143 outer_->delayed_task_manager_->AddDelayedTask(
144 std::move(task), Bind(&SchedulerSingleThreadTaskRunner::PostTaskNow,
145 Unretained(this)));
146 }
147 return true;
148 }
125 149
126 bool PostNonNestableDelayedTask(const tracked_objects::Location& from_here, 150 bool PostNonNestableDelayedTask(const tracked_objects::Location& from_here,
127 const Closure& closure, 151 const Closure& closure,
128 base::TimeDelta delay) override { 152 base::TimeDelta delay) override {
129 // Tasks are never nested within the task scheduler. 153 // Tasks are never nested within the task scheduler.
130 return PostDelayedTask(from_here, closure, delay); 154 return PostDelayedTask(from_here, closure, delay);
131 } 155 }
132 156
133 bool RunsTasksOnCurrentThread() const override { 157 bool RunsTasksOnCurrentThread() const override {
134 auto* delegate = static_cast<SchedulerWorkerDelegate*>(worker_->delegate()); 158 return GetDelegate()->RunsTasksOnCurrentThread();
135 return delegate->RunsTasksOnCurrentThread();
136 } 159 }
137 160
138 private: 161 private:
139 ~SchedulerSingleThreadTaskRunner() override { 162 ~SchedulerSingleThreadTaskRunner() override {
140 outer_->UnregisterSchedulerWorker(worker_); 163 outer_->UnregisterSchedulerWorker(worker_);
141 } 164 }
142 165
143 void PostTaskNow(std::unique_ptr<Task> task); 166 void PostTaskNow(std::unique_ptr<Task> task) {
167 Sequence* sequence = GetDelegate()->sequence();
fdoray 2017/03/02 14:37:48 When shutdown completes, the thread exits and the
robliao 2017/03/03 04:24:17 Handled by making this a scoped_refptr. At this po
168 const bool sequence_was_empty = sequence->PushTask(std::move(task));
169 if (sequence_was_empty) {
170 GetDelegate()->ReEnqueueSequence(sequence);
171 worker_->WakeUp();
172 }
173 }
144 174
145 // Sequence for all Tasks posted through this TaskRunner. 175 SchedulerWorkerDelegate* GetDelegate() const {
146 const scoped_refptr<Sequence> sequence_ = new Sequence; 176 return static_cast<SchedulerWorkerDelegate*>(worker_->delegate());
177 }
147 178
148 SchedulerSingleThreadTaskRunnerManager* const outer_; 179 SchedulerSingleThreadTaskRunnerManager* const outer_;
149 const TaskTraits traits_; 180 const TaskTraits traits_;
150 SchedulerWorker* const worker_; 181 SchedulerWorker* const worker_;
151 182
152 DISALLOW_COPY_AND_ASSIGN(SchedulerSingleThreadTaskRunner); 183 DISALLOW_COPY_AND_ASSIGN(SchedulerSingleThreadTaskRunner);
153 }; 184 };
154 185
155 bool SchedulerSingleThreadTaskRunnerManager::SchedulerSingleThreadTaskRunner::
156 PostDelayedTask(const tracked_objects::Location& from_here,
157 const Closure& closure,
158 TimeDelta delay) {
159 auto task = MakeUnique<Task>(from_here, closure, traits_, delay);
160 task->single_thread_task_runner_ref = this;
161
162 if (!outer_->task_tracker_->WillPostTask(task.get()))
163 return false;
164
165 if (task->delayed_run_time.is_null()) {
166 PostTaskNow(std::move(task));
167 } else {
168 outer_->delayed_task_manager_->AddDelayedTask(
169 std::move(task),
170 Bind(&SchedulerSingleThreadTaskRunner::PostTaskNow, Unretained(this)));
171 }
172 return true;
173 }
174
175 void SchedulerSingleThreadTaskRunnerManager::SchedulerSingleThreadTaskRunner::
176 PostTaskNow(std::unique_ptr<Task> task) {
177 const bool sequence_was_empty = sequence_->PushTask(std::move(task));
178 if (sequence_was_empty) {
179 auto* delegate = static_cast<SchedulerWorkerDelegate*>(worker_->delegate());
180 delegate->ReEnqueueSequence(sequence_);
181 worker_->WakeUp();
182 }
183 }
184
185 SchedulerSingleThreadTaskRunnerManager::SchedulerSingleThreadTaskRunnerManager( 186 SchedulerSingleThreadTaskRunnerManager::SchedulerSingleThreadTaskRunnerManager(
186 const std::vector<SchedulerWorkerPoolParams>& worker_pool_params_vector, 187 const std::vector<SchedulerWorkerPoolParams>& worker_pool_params_vector,
187 const TaskScheduler::WorkerPoolIndexForTraitsCallback& 188 const TaskScheduler::WorkerPoolIndexForTraitsCallback&
188 worker_pool_index_for_traits_callback, 189 worker_pool_index_for_traits_callback,
189 TaskTracker* task_tracker, 190 TaskTracker* task_tracker,
190 DelayedTaskManager* delayed_task_manager) 191 DelayedTaskManager* delayed_task_manager)
191 : worker_pool_params_vector_(worker_pool_params_vector), 192 : worker_pool_params_vector_(worker_pool_params_vector),
192 worker_pool_index_for_traits_callback_( 193 worker_pool_index_for_traits_callback_(
193 worker_pool_index_for_traits_callback), 194 worker_pool_index_for_traits_callback),
194 task_tracker_(task_tracker), 195 task_tracker_(task_tracker),
195 delayed_task_manager_(delayed_task_manager) { 196 delayed_task_manager_(delayed_task_manager) {
196 DCHECK_GT(worker_pool_params_vector_.size(), 0U); 197 DCHECK_GT(worker_pool_params_vector_.size(), 0U);
197 DCHECK(worker_pool_index_for_traits_callback_); 198 DCHECK(worker_pool_index_for_traits_callback_);
198 DCHECK(task_tracker_); 199 DCHECK(task_tracker_);
199 DCHECK(delayed_task_manager_); 200 DCHECK(delayed_task_manager_);
200 } 201 }
201 202
202 SchedulerSingleThreadTaskRunnerManager:: 203 SchedulerSingleThreadTaskRunnerManager::
203 ~SchedulerSingleThreadTaskRunnerManager() { 204 ~SchedulerSingleThreadTaskRunnerManager() {
204 DCHECK(workers_.empty()) << "SchedulerSingleThreadTaskRunners must outlive " 205 size_t workers_unregistered_during_join =
205 "SchedulerSingleThreadTaskRunnerManager"; 206 subtle::NoBarrier_Load(&workers_unregistered_during_join_);
207 DCHECK_EQ(workers_unregistered_during_join, workers_.size())
208 << "SchedulerSingleThreadTaskRunners must outlive "
fdoray 2017/03/02 14:37:48 I think this would be clearer from a user point of
robliao 2017/03/03 04:24:17 While that's true, it doesn't make sense in the co
209 "SchedulerSingleThreadTaskRunnerManager";
206 } 210 }
207 211
208 scoped_refptr<SingleThreadTaskRunner> 212 scoped_refptr<SingleThreadTaskRunner>
209 SchedulerSingleThreadTaskRunnerManager::CreateSingleThreadTaskRunnerWithTraits( 213 SchedulerSingleThreadTaskRunnerManager::CreateSingleThreadTaskRunnerWithTraits(
210 const TaskTraits& traits) { 214 const TaskTraits& traits) {
211 size_t index = worker_pool_index_for_traits_callback_.Run(traits); 215 size_t index = worker_pool_index_for_traits_callback_.Run(traits);
212 DCHECK_LT(index, worker_pool_params_vector_.size()); 216 DCHECK_LT(index, worker_pool_params_vector_.size());
213 return new SchedulerSingleThreadTaskRunner( 217 return new SchedulerSingleThreadTaskRunner(
214 this, traits, 218 this, traits,
215 CreateAndRegisterSchedulerWorker(worker_pool_params_vector_[index])); 219 CreateAndRegisterSchedulerWorker(worker_pool_params_vector_[index]));
(...skipping 31 matching lines...) Expand 10 before | Expand all | Expand 10 after
247 } 251 }
248 252
249 void SchedulerSingleThreadTaskRunnerManager::UnregisterSchedulerWorker( 253 void SchedulerSingleThreadTaskRunnerManager::UnregisterSchedulerWorker(
250 SchedulerWorker* worker) { 254 SchedulerWorker* worker) {
251 // Cleanup uses a SchedulerLock, so call Cleanup() after releasing 255 // Cleanup uses a SchedulerLock, so call Cleanup() after releasing
252 // |workers_lock_|. 256 // |workers_lock_|.
253 scoped_refptr<SchedulerWorker> worker_to_destroy; 257 scoped_refptr<SchedulerWorker> worker_to_destroy;
254 { 258 {
255 AutoSchedulerLock auto_lock(workers_lock_); 259 AutoSchedulerLock auto_lock(workers_lock_);
256 260
257 // We might be joining, so no-op this if |workers_| is empty. 261 // We might be joining, so record that a worker was unregistered for
258 if (workers_.empty()) 262 // verification at destruction.
263 if (workers_.empty()) {
264 subtle::NoBarrier_AtomicIncrement(&workers_unregistered_during_join_, 1);
fdoray 2017/03/02 14:37:48 Use base::AtomicRefCountInc and base::AtomicRefCou
robliao 2017/03/03 04:24:17 Those functions are intended for reference countin
fdoray 2017/03/03 05:31:13 They want people to use base/memory/ref_counted.h
fdoray 2017/03/03 07:22:39 ping
robliao 2017/03/03 07:50:05 The intent of atomic_ref_count.h as written is to
259 return; 265 return;
266 }
260 267
261 auto worker_iter = 268 auto worker_iter =
262 std::find_if(workers_.begin(), workers_.end(), 269 std::find_if(workers_.begin(), workers_.end(),
263 [worker](const scoped_refptr<SchedulerWorker>& candidate) { 270 [worker](const scoped_refptr<SchedulerWorker>& candidate) {
264 return candidate.get() == worker; 271 return candidate.get() == worker;
265 }); 272 });
266 DCHECK(worker_iter != workers_.end()); 273 DCHECK(worker_iter != workers_.end());
267 worker_to_destroy = std::move(*worker_iter); 274 worker_to_destroy = std::move(*worker_iter);
268 workers_.erase(worker_iter); 275 workers_.erase(worker_iter);
269 } 276 }
270 worker_to_destroy->Cleanup(); 277 worker_to_destroy->Cleanup();
271 } 278 }
272 279
273 } // namespace internal 280 } // namespace internal
274 } // namespace base 281 } // namespace base
OLDNEW

Powered by Google App Engine
This is Rietveld 408576698