Chromium Code Reviews| OLD | NEW |
|---|---|
| (Empty) | |
| 1 // Copyright (c) 2011 The Chromium Authors. All rights reserved. | |
| 2 // Use of this source code is governed by a BSD-style license that can be | |
| 3 // found in the LICENSE file. | |
| 4 | |
| 5 #include "content/common/sequenced_worker_pool.h" | |
| 6 | |
| 7 #include <deque> | |
| 8 #include <set> | |
| 9 | |
| 10 #include "base/atomicops.h" | |
| 11 #include "base/bind.h" | |
| 12 #include "base/memory/scoped_ptr.h" | |
| 13 #include "base/metrics/histogram.h" | |
| 14 #include "base/threading/thread.h" | |
| 15 #include "base/stringprintf.h" | |
| 16 #include "base/synchronization/waitable_event.h" | |
| 17 | |
| 18 namespace { | |
| 19 | |
| 20 struct SequencedTask { | |
| 21 int sequence_token_id; | |
| 22 SequencedWorkerPool::WorkerShutdown shutdown_behavior; | |
| 23 tracked_objects::Location location; | |
| 24 base::Closure task; | |
| 25 }; | |
| 26 | |
| 27 class Worker { | |
| 28 public: | |
| 29 Worker(int thread_number) | |
| 30 : thread_(StringPrintf("Browser worker %d", thread_number).c_str()), | |
| 31 current_shutdown_mode_(SequencedWorkerPool::CONTINUE_ON_SHUTDOWN) { | |
| 32 thread_.Start(); | |
| 33 } | |
| 34 ~Worker() { | |
| 35 } | |
| 36 | |
| 37 // Posts a task to the worker's message loop for running. The actual task | |
| 38 // to run should be Inner's RunTask. The SequencedTask is passed in only | |
| 39 // to get statistics out of. | |
| 40 // | |
| 41 // This should only be called from within Inner's lock. | |
|
jar (doing other things)
2011/11/23 20:08:16
It is surprising to require a lock (when posting a
| |
| 42 void PostTask(const SequencedTask& sequenced_info, | |
| 43 const base::Closure& task) { | |
| 44 // Use the original task birthplace as the source for this call so we can | |
| 45 // trace back who made us do this work. | |
| 46 thread_.message_loop()->PostTask(sequenced_info.location, task); | |
|
jar (doing other things)
2011/11/23 20:08:16
It is more surprising to see a local lock held as
| |
| 47 current_shutdown_mode_ = sequenced_info.shutdown_behavior; | |
| 48 } | |
| 49 | |
| 50 // Cleans up after a task is complete. This should be called from within | |
| 51 // Inner's lock as soon as the task is complete. | |
| 52 void WorkComplete() { | |
| 53 current_shutdown_mode_ = SequencedWorkerPool::CONTINUE_ON_SHUTDOWN; | |
| 54 } | |
| 55 | |
| 56 // When running a task, the shutdown mode is stored on the worker. It will | |
| 57 // be INTERRUPT_ON_SHUTDOWN if there is no running task. | |
| 58 // | |
| 59 // This must only be called when holding the Inner class' lock. | |
| 60 const SequencedWorkerPool::WorkerShutdown current_shutdown_mode() const { | |
| 61 return current_shutdown_mode_; | |
| 62 } | |
| 63 | |
| 64 private: | |
| 65 base::Thread thread_; | |
| 66 | |
| 67 SequencedWorkerPool::WorkerShutdown current_shutdown_mode_; | |
| 68 | |
| 69 DISALLOW_COPY_AND_ASSIGN(Worker); | |
| 70 }; | |
| 71 | |
| 72 } // namespace | |
| 73 | |
| 74 // Inner ---------------------------------------------------------------------- | |
| 75 | |
| 76 class SequencedWorkerPool::Inner | |
|
jar (doing other things)
2011/11/23 20:08:16
Suggest you add a comment to clarify that this is
| |
| 77 : public base::RefCountedThreadSafe<SequencedWorkerPool::Inner> { | |
| 78 public: | |
| 79 Inner(size_t max_threads); | |
| 80 | |
| 81 // Returns a unique token that can be used to sequence tasks posted to | |
| 82 // PostSequencedWorkerTask(). Valid tokens are alwys nonzero. | |
| 83 SequenceToken GetSequenceToken(); | |
| 84 | |
| 85 // Posts a task. See PostSequencedWorkerTask. The token ID will be 0 for | |
| 86 // unsequenced tasks. | |
| 87 bool PostTask(int sequence_token_id, | |
| 88 SequencedWorkerPool::WorkerShutdown shutdown_behavior, | |
| 89 const tracked_objects::Location& from_here, | |
| 90 const base::Closure& task); | |
| 91 | |
| 92 void Shutdown(); | |
| 93 | |
| 94 void SetTestingObserver(SequencedWorkerPool::TestingObserver* observer); | |
| 95 | |
| 96 private: | |
| 97 // Attempts to schedule work on an available thread. Call this when a | |
| 98 // task is added to the pending_tasks_ list or a thread is added to the | |
| 99 // idle_threads_ list. The lock must be already held for this function. | |
| 100 void ScheduleWork(); | |
| 101 | |
| 102 // Worker task that actually runs on a worker thread to execute the given | |
| 103 // task. The worker it's running on is passed as the first argument. | |
| 104 void RunTask(Worker* this_worker, const SequencedTask& task); | |
| 105 | |
| 106 // Returns true if any worker is running a shutdown-blocking task. This | |
| 107 // includes BLOCK_SHUTDOWN and also SKIP_ON_SHUTDOWN tasks that are already | |
| 108 // running (since they must be run to completion). It must be called with | |
| 109 // the lock held. | |
| 110 bool IsRunningBlockingTask() const; | |
| 111 | |
| 112 // The last sequence number used. Managed by GetSequenceToken, since this | |
| 113 // only does threadsafe increment operations, you do not need to hold the | |
| 114 // lock. | |
| 115 volatile base::subtle::Atomic32 last_sequence_number_; | |
| 116 | |
| 117 // This lock protects |everything in this class|. Do not read or modify | |
| 118 // anything without holding this lock. Do not block while holding this | |
| 119 // lock. | |
| 120 base::Lock lock_; | |
| 121 | |
| 122 // The maximum number of worker threads we'll create. | |
| 123 size_t max_threads_; | |
| 124 | |
| 125 // Owning pointers to all threads we've created so far. Since we lazily | |
| 126 // create threads, this may be less than max_threads_ and will be initially | |
| 127 // empty. Each of these pointers will also appear either in idle_threads_ | |
| 128 // or running_threads_. | |
| 129 std::vector<linked_ptr<Worker> > threads_; | |
| 130 | |
| 131 // Lists all currently idle worker threads. These pointers are non-owning, | |
| 132 // the threads_ array manages their lifetimes. | |
| 133 std::deque<Worker*> idle_threads_; | |
|
jar (doing other things)
2011/11/23 20:08:16
You should use a stack, as I don't think there is
| |
| 134 | |
| 135 // The opposite of idle_threads_, this contains non-owning pointers to all | |
| 136 // currently running or scheduled threads. | |
| 137 std::set<Worker*> running_threads_; | |
| 138 | |
| 139 // In-order list of all pending tasks. These are tasks waiting for a thread | |
| 140 // to run on or that are blocked on a previous task in their sequence. | |
| 141 // | |
| 142 // We maintain the pending_task_count_ separately for metrics because | |
| 143 // list.size() can be linear time. | |
| 144 std::list<SequencedTask> pending_tasks_; | |
| 145 size_t pending_task_count_; | |
| 146 | |
| 147 // Lists all sequence tokens currently executing. | |
| 148 std::set<int> current_sequences_; | |
| 149 | |
| 150 // Set when the app is terminating and no further tasks should be allowed. | |
| 151 bool terminated_; | |
| 152 | |
| 153 SequencedWorkerPool::TestingObserver* testing_observer_; | |
| 154 | |
| 155 // Created lazily when terminated_ is set and there are pending tasks, this | |
| 156 // is signaled by ScheduleWork whel all blocking tasks have completed. | |
| 157 scoped_ptr<base::WaitableEvent> shutdown_complete_; | |
| 158 }; | |
| 159 | |
| 160 SequencedWorkerPool::Inner::Inner(size_t max_threads) | |
| 161 : last_sequence_number_(0), | |
| 162 max_threads_(max_threads), | |
| 163 pending_task_count_(0), | |
| 164 terminated_(false) { | |
| 165 } | |
| 166 | |
| 167 SequencedWorkerPool::SequenceToken | |
| 168 SequencedWorkerPool::Inner::GetSequenceToken() { | |
|
jar (doing other things)
2011/11/23 20:08:16
Cool as this is... I'm suspicious that we need an
brettw
2011/11/23 21:01:13
I think an enum destroys the whole point of this p
| |
| 169 base::subtle::Atomic32 result = | |
| 170 base::subtle::NoBarrier_AtomicIncrement(&last_sequence_number_, 1); | |
| 171 return SequenceToken(static_cast<int>(result)); | |
| 172 } | |
| 173 | |
| 174 bool SequencedWorkerPool::Inner::PostTask( | |
| 175 int sequence_token_id, | |
| 176 SequencedWorkerPool::WorkerShutdown shutdown_behavior, | |
| 177 const tracked_objects::Location& from_here, | |
| 178 const base::Closure& task) { | |
| 179 base::AutoLock lock(lock_); | |
| 180 | |
| 181 if (terminated_) | |
| 182 return false; | |
| 183 | |
| 184 SequencedTask sequenced; | |
| 185 sequenced.sequence_token_id = sequence_token_id; | |
| 186 sequenced.shutdown_behavior = shutdown_behavior; | |
| 187 sequenced.location = from_here; | |
| 188 sequenced.task = task; | |
| 189 pending_tasks_.push_back(sequenced); | |
| 190 pending_task_count_++; | |
| 191 | |
| 192 ScheduleWork(); | |
| 193 return true; | |
| 194 } | |
| 195 | |
| 196 void SequencedWorkerPool::Inner::Shutdown() { | |
|
jar (doing other things)
2011/11/23 20:08:16
This function is written as a blocking call. I mu
brettw
2011/11/23 21:01:13
This has to be the main thread. It has to block si
| |
| 197 // Mark us as terminated and go through and drop all tasks that aren't | |
| 198 // required to run on shutdown. Since no new tasks will get posted once the | |
| 199 // terminated flag is set, this ensures that all remaining tasks are required | |
| 200 // for shutdown whenever the termianted_ flag is set. | |
| 201 { | |
| 202 base::AutoLock lock(lock_); | |
| 203 DCHECK(!terminated_); | |
| 204 terminated_ = true; | |
| 205 | |
| 206 std::list<SequencedTask>::iterator i = pending_tasks_.begin(); | |
| 207 while (i != pending_tasks_.end()) { | |
| 208 if (i->shutdown_behavior == BLOCK_SHUTDOWN) { | |
| 209 i++; | |
| 210 } else { | |
| 211 i = pending_tasks_.erase(i); | |
|
jar (doing other things)
2011/11/23 20:08:16
Does this potentially induce a task destructor? D
| |
| 212 pending_task_count_--; | |
| 213 } | |
| 214 } | |
| 215 DCHECK_EQ(pending_tasks_.size(), pending_task_count_); | |
| 216 | |
| 217 if (pending_tasks_.empty() && !IsRunningBlockingTask()) { | |
| 218 // There are no pending or running tasks blocking shutdown, we're done. | |
| 219 return; | |
| 220 } | |
| 221 | |
| 222 // Need to wait for some tasks, create the event. | |
| 223 DCHECK(!shutdown_complete_.get()); | |
| 224 shutdown_complete_.reset(new base::WaitableEvent(false, false)); | |
| 225 } | |
| 226 | |
| 227 // If we get here, we know we're either waiting on a blocking task that's | |
| 228 // currently running, waiting on a blocking task that hasn't been scheduled | |
| 229 // yet, or both. Block on the "queue empty" event to know when all tasks are | |
| 230 // complete. This must be done outside the lock. | |
| 231 if (testing_observer_) | |
| 232 testing_observer_->WillWaitForShutdown(); | |
| 233 | |
| 234 base::TimeTicks shutdown_wait_begin = base::TimeTicks::Now(); | |
| 235 shutdown_complete_->Wait(); | |
| 236 UMA_HISTOGRAM_TIMES("SequencedWorkerPool.ShutdownDelayTime", | |
| 237 base::TimeTicks::Now() - shutdown_wait_begin); | |
| 238 } | |
| 239 | |
| 240 void SequencedWorkerPool::Inner::SetTestingObserver( | |
| 241 SequencedWorkerPool::TestingObserver* observer) { | |
| 242 base::AutoLock lock(lock_); | |
| 243 testing_observer_ = observer; | |
| 244 } | |
| 245 | |
| 246 void SequencedWorkerPool::Inner::ScheduleWork() { | |
|
jar (doing other things)
2011/11/23 20:08:16
You might consider adding a comment that this func
| |
| 247 lock_.AssertAcquired(); | |
| 248 | |
| 249 DCHECK_EQ(pending_tasks_.size(), pending_task_count_); | |
| 250 UMA_HISTOGRAM_COUNTS_100("SequencedWorkerPool.TaskCount", | |
| 251 pending_task_count_); | |
| 252 | |
| 253 if (terminated_ && shutdown_complete_.get()) { | |
| 254 // When the app is terminating, check for "no more blocking work" and | |
| 255 // signal if shutdown tasks are complete. | |
| 256 if (pending_tasks_.empty() && !IsRunningBlockingTask()) { | |
| 257 shutdown_complete_->Signal(); | |
| 258 return; | |
| 259 } | |
| 260 } | |
| 261 | |
| 262 if (pending_tasks_.empty() || | |
| 263 (idle_threads_.empty() && threads_.size() == max_threads_)) | |
| 264 return; // No work to schedule or no threads to schedule them on. | |
| 265 | |
| 266 // Find the next task with a sequence token that's not currently in use. | |
| 267 // If the token is in use, that means another thread is running something | |
| 268 // in that sequence, and we can't run it without going out-of-order. | |
| 269 // | |
| 270 // This algorithm is simple and fair, but inefficient in some cases. For | |
| 271 // example, say somebody schedules 1000 slow tasks with the same sequence | |
| 272 // number. We'll have to go through all those tasks each time we feel like | |
| 273 // there might be work to schedule. If this proves to be a problem, we | |
| 274 // should make this more efficient. | |
| 275 // | |
| 276 // One possible enhancement would be to keep a map from sequence ID to a | |
| 277 // list of pending but currently blocked SequencedTasks for that ID. | |
| 278 // When a worker finishes a task of one sequence token, it can pick up the | |
| 279 // next one from that token right away. | |
|
jar (doing other things)
2011/11/23 20:08:16
That is a good optimization for efficiency... but
| |
| 280 // | |
| 281 // This may lead to starvation if there are sufficient numbers of sequences | |
| 282 // in use. To alleviate this, we could add an incrementing priority counter | |
| 283 // to each SequencedTask. Then maintain a priority_queue of all runnable | |
| 284 // tasks, sorted by priority counter. When a sequenced task is completed | |
| 285 // we would pop the head element off of that tasks pending list and add it | |
| 286 // to the priority queue. Then we would run the first item in the priority | |
| 287 // queue. | |
| 288 int unrunnable_tasks = 0; | |
| 289 for (std::list<SequencedTask>::iterator i = pending_tasks_.begin(); | |
| 290 i != pending_tasks_.end(); ++i) { | |
| 291 if (!i->sequence_token_id || | |
| 292 current_sequences_.find(i->sequence_token_id) == | |
| 293 current_sequences_.end()) { | |
| 294 // This token is free, run this task on the first available worker, | |
| 295 // creating one if necessary. | |
| 296 Worker* worker; | |
|
jar (doing other things)
2011/11/23 20:08:16
Should this be a linked_ptr?
| |
| 297 if (idle_threads_.empty()) { | |
| 298 // We should have early exited above if we're out of threads to | |
| 299 // schedule, so there should always be a free slot. | |
| 300 DCHECK(threads_.size() < max_threads_); | |
| 301 worker = new Worker(threads_.size()); | |
| 302 threads_.push_back(linked_ptr<Worker>(worker)); | |
| 303 } else { | |
| 304 worker = idle_threads_.front(); | |
| 305 idle_threads_.pop_front(); | |
| 306 } | |
| 307 running_threads_.insert(worker); | |
| 308 | |
| 309 // Mark the task's sequence number as in use. | |
| 310 if (i->sequence_token_id) | |
| 311 current_sequences_.insert(i->sequence_token_id); | |
| 312 | |
| 313 worker->PostTask(*i, base::Bind(&Inner::RunTask, this, worker, *i)); | |
|
jar (doing other things)
2011/11/23 20:08:16
Now that we have the worker thread out of the idle
| |
| 314 pending_tasks_.erase(i); | |
| 315 pending_task_count_--; | |
| 316 break; | |
| 317 } | |
| 318 unrunnable_tasks++; | |
| 319 } | |
| 320 | |
| 321 // Track the number of tasks we had to skip over to see if we should be | |
| 322 // making this more efficient. If this number ever becomes large or is | |
| 323 // frequently "some", we should consider the optimization above. | |
| 324 UMA_HISTOGRAM_COUNTS_100("SequencedWorkerPool.UnrunnableTaskCount", | |
| 325 unrunnable_tasks); | |
| 326 } | |
| 327 | |
| 328 void SequencedWorkerPool::Inner::RunTask(Worker* this_worker, | |
| 329 const SequencedTask& task) { | |
| 330 task.task.Run(); | |
| 331 | |
| 332 // Now that this thread is free, mark ourselves and try to schedule more. | |
| 333 { | |
| 334 base::AutoLock lock(lock_); | |
| 335 this_worker->WorkComplete(); | |
| 336 | |
| 337 if (task.sequence_token_id) | |
| 338 current_sequences_.erase(task.sequence_token_id); | |
| 339 running_threads_.erase(this_worker); | |
| 340 idle_threads_.push_front(this_worker); | |
| 341 | |
| 342 ScheduleWork(); | |
| 343 } | |
| 344 } | |
| 345 | |
| 346 bool SequencedWorkerPool::Inner::IsRunningBlockingTask() const { | |
| 347 lock_.AssertAcquired(); | |
| 348 | |
| 349 for (std::set<Worker*>::const_iterator i = running_threads_.begin(); | |
| 350 i != running_threads_.end(); ++i) { | |
| 351 if ((*i)->current_shutdown_mode() == SequencedWorkerPool::BLOCK_SHUTDOWN || | |
| 352 (*i)->current_shutdown_mode() == SequencedWorkerPool::SKIP_ON_SHUTDOWN) | |
| 353 return true; | |
| 354 } | |
| 355 return false; | |
| 356 } | |
| 357 | |
| 358 // SequencedWorkerPool -------------------------------------------------------- | |
| 359 | |
| 360 SequencedWorkerPool::SequencedWorkerPool(size_t max_threads) | |
| 361 : inner_(new Inner(max_threads)) { | |
| 362 } | |
| 363 | |
| 364 SequencedWorkerPool::~SequencedWorkerPool() { | |
| 365 } | |
| 366 | |
| 367 SequencedWorkerPool::SequenceToken SequencedWorkerPool::GetSequenceToken() { | |
| 368 return inner_->GetSequenceToken(); | |
| 369 } | |
| 370 | |
| 371 bool SequencedWorkerPool::PostWorkerTask( | |
| 372 WorkerShutdown shutdown_behavior, | |
| 373 const tracked_objects::Location& from_here, | |
| 374 const base::Closure& task) { | |
| 375 return inner_->PostTask(0, shutdown_behavior, from_here, task); | |
| 376 } | |
| 377 | |
| 378 bool SequencedWorkerPool::PostSequencedWorkerTask( | |
| 379 SequenceToken sequence_token, | |
| 380 WorkerShutdown shutdown_behavior, | |
| 381 const tracked_objects::Location& from_here, | |
| 382 const base::Closure& task) { | |
| 383 return inner_->PostTask(sequence_token.id_, shutdown_behavior, | |
| 384 from_here, task); | |
| 385 } | |
| 386 | |
| 387 void SequencedWorkerPool::Shutdown() { | |
| 388 inner_->Shutdown(); | |
| 389 } | |
| 390 | |
| 391 void SequencedWorkerPool::SetTestingObserver(TestingObserver* observer) { | |
| 392 inner_->SetTestingObserver(observer); | |
| 393 } | |
| OLD | NEW |