| OLD | NEW |
| 1 // Copyright 2014 The Chromium Authors. All rights reserved. | 1 // Copyright 2014 The Chromium Authors. All rights reserved. |
| 2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
| 3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
| 4 | 4 |
| 5 #include "components/scheduler/base/task_queue_manager.h" | 5 #include "components/scheduler/base/task_queue_manager.h" |
| 6 | 6 |
| 7 #include <queue> | 7 #include <queue> |
| 8 #include <set> | 8 #include <set> |
| 9 | 9 |
| 10 #include "base/bind.h" | 10 #include "base/bind.h" |
| 11 #include "components/scheduler/base/lazy_now.h" | 11 #include "components/scheduler/base/real_time_domain.h" |
| 12 #include "components/scheduler/base/task_queue_impl.h" | 12 #include "components/scheduler/base/task_queue_impl.h" |
| 13 #include "components/scheduler/base/task_queue_manager_delegate.h" | 13 #include "components/scheduler/base/task_queue_manager_delegate.h" |
| 14 #include "components/scheduler/base/task_queue_selector.h" | 14 #include "components/scheduler/base/task_queue_selector.h" |
| 15 #include "components/scheduler/base/task_queue_sets.h" | 15 #include "components/scheduler/base/task_queue_sets.h" |
| 16 | 16 |
| 17 namespace { | 17 namespace { |
| 18 const int64_t kMaxTimeTicks = std::numeric_limits<int64>::max(); | 18 const int64_t kMaxTimeTicks = std::numeric_limits<int64>::max(); |
| 19 } | 19 } |
| 20 | 20 |
| 21 namespace scheduler { | 21 namespace scheduler { |
| (...skipping 17 matching lines...) Expand all Loading... |
| 39 weak_factory_(this) { | 39 weak_factory_(this) { |
| 40 DCHECK(delegate->RunsTasksOnCurrentThread()); | 40 DCHECK(delegate->RunsTasksOnCurrentThread()); |
| 41 TRACE_EVENT_OBJECT_CREATED_WITH_ID(disabled_by_default_tracing_category, | 41 TRACE_EVENT_OBJECT_CREATED_WITH_ID(disabled_by_default_tracing_category, |
| 42 "TaskQueueManager", this); | 42 "TaskQueueManager", this); |
| 43 selector_.SetTaskQueueSelectorObserver(this); | 43 selector_.SetTaskQueueSelectorObserver(this); |
| 44 | 44 |
| 45 decrement_pending_and_do_work_closure_ = | 45 decrement_pending_and_do_work_closure_ = |
| 46 base::Bind(&TaskQueueManager::DoWork, weak_factory_.GetWeakPtr(), true); | 46 base::Bind(&TaskQueueManager::DoWork, weak_factory_.GetWeakPtr(), true); |
| 47 do_work_closure_ = | 47 do_work_closure_ = |
| 48 base::Bind(&TaskQueueManager::DoWork, weak_factory_.GetWeakPtr(), false); | 48 base::Bind(&TaskQueueManager::DoWork, weak_factory_.GetWeakPtr(), false); |
| 49 |
| 50 default_time_domain_ = |
| 51 make_scoped_refptr(new RealTimeDomain(delegate.get(), do_work_closure_)); |
| 52 time_domains_.insert(default_time_domain_); |
| 49 } | 53 } |
| 50 | 54 |
| 51 TaskQueueManager::~TaskQueueManager() { | 55 TaskQueueManager::~TaskQueueManager() { |
| 52 TRACE_EVENT_OBJECT_DELETED_WITH_ID(disabled_by_default_tracing_category_, | 56 TRACE_EVENT_OBJECT_DELETED_WITH_ID(disabled_by_default_tracing_category_, |
| 53 "TaskQueueManager", this); | 57 "TaskQueueManager", this); |
| 54 | 58 |
| 55 while (!queues_.empty()) | 59 while (!queues_.empty()) |
| 56 (*queues_.begin())->UnregisterTaskQueue(); | 60 (*queues_.begin())->UnregisterTaskQueue(); |
| 57 | 61 |
| 58 selector_.SetTaskQueueSelectorObserver(nullptr); | 62 selector_.SetTaskQueueSelectorObserver(nullptr); |
| 59 } | 63 } |
| 60 | 64 |
| 61 scoped_refptr<internal::TaskQueueImpl> TaskQueueManager::NewTaskQueue( | 65 scoped_refptr<internal::TaskQueueImpl> TaskQueueManager::NewTaskQueue( |
| 62 const TaskQueue::Spec& spec) { | 66 const TaskQueue::Spec& spec) { |
| 63 TRACE_EVENT1(tracing_category_, | 67 TRACE_EVENT1(tracing_category_, |
| 64 "TaskQueueManager::NewTaskQueue", "queue_name", spec.name); | 68 "TaskQueueManager::NewTaskQueue", "queue_name", spec.name); |
| 65 DCHECK(main_thread_checker_.CalledOnValidThread()); | 69 DCHECK(main_thread_checker_.CalledOnValidThread()); |
| 66 scoped_refptr<internal::TaskQueueImpl> queue( | 70 scoped_refptr<internal::TaskQueueImpl> queue( |
| 67 make_scoped_refptr(new internal::TaskQueueImpl( | 71 make_scoped_refptr(new internal::TaskQueueImpl( |
| 68 this, spec, disabled_by_default_tracing_category_, | 72 this, |
| 73 spec.time_domain ? spec.time_domain : default_time_domain_.get(), |
| 74 spec, disabled_by_default_tracing_category_, |
| 69 disabled_by_default_verbose_tracing_category_))); | 75 disabled_by_default_verbose_tracing_category_))); |
| 70 queues_.insert(queue); | 76 queues_.insert(queue); |
| 71 selector_.AddQueue(queue.get()); | 77 selector_.AddQueue(queue.get()); |
| 72 return queue; | 78 return queue; |
| 73 } | 79 } |
| 74 | 80 |
| 75 void TaskQueueManager::SetObserver(Observer* observer) { | 81 void TaskQueueManager::SetObserver(Observer* observer) { |
| 76 DCHECK(main_thread_checker_.CalledOnValidThread()); | 82 DCHECK(main_thread_checker_.CalledOnValidThread()); |
| 77 observer_ = observer; | 83 observer_ = observer; |
| 78 } | 84 } |
| 79 | 85 |
| 80 void TaskQueueManager::UnregisterTaskQueue( | 86 void TaskQueueManager::UnregisterTaskQueue( |
| 81 scoped_refptr<internal::TaskQueueImpl> task_queue) { | 87 scoped_refptr<internal::TaskQueueImpl> task_queue) { |
| 82 TRACE_EVENT1(tracing_category_, | 88 TRACE_EVENT1(tracing_category_, |
| 83 "TaskQueueManager::UnregisterTaskQueue", "queue_name", | 89 "TaskQueueManager::UnregisterTaskQueue", "queue_name", |
| 84 task_queue->GetName()); | 90 task_queue->GetName()); |
| 85 DCHECK(main_thread_checker_.CalledOnValidThread()); | 91 DCHECK(main_thread_checker_.CalledOnValidThread()); |
| 86 if (observer_) | 92 if (observer_) |
| 87 observer_->OnUnregisterTaskQueue(task_queue); | 93 observer_->OnUnregisterTaskQueue(task_queue); |
| 88 | 94 |
| 89 // Add |task_queue| to |queues_to_delete_| so we can prevent it from being | 95 // Add |task_queue| to |queues_to_delete_| so we can prevent it from being |
| 90 // freed while any of our structures hold hold a raw pointer to it. | 96 // freed while any of our structures hold hold a raw pointer to it. |
| 91 queues_to_delete_.insert(task_queue); | 97 queues_to_delete_.insert(task_queue); |
| 92 queues_.erase(task_queue); | 98 queues_.erase(task_queue); |
| 93 selector_.RemoveQueue(task_queue.get()); | 99 selector_.RemoveQueue(task_queue.get()); |
| 94 | |
| 95 // We need to remove |task_queue| from delayed_wakeup_multimap_ which is a | |
| 96 // little awkward since it's keyed by time. O(n) running time. | |
| 97 for (DelayedWakeupMultimap::iterator iter = delayed_wakeup_multimap_.begin(); | |
| 98 iter != delayed_wakeup_multimap_.end();) { | |
| 99 if (iter->second == task_queue.get()) { | |
| 100 DelayedWakeupMultimap::iterator temp = iter; | |
| 101 iter++; | |
| 102 // O(1) amortized. | |
| 103 delayed_wakeup_multimap_.erase(temp); | |
| 104 } else { | |
| 105 iter++; | |
| 106 } | |
| 107 } | |
| 108 | |
| 109 // |newly_updatable_| might contain |task_queue|, we use | |
| 110 // MoveNewlyUpdatableQueuesIntoUpdatableQueueSet to flush it out. | |
| 111 MoveNewlyUpdatableQueuesIntoUpdatableQueueSet(); | |
| 112 updatable_queue_set_.erase(task_queue.get()); | |
| 113 } | 100 } |
| 114 | 101 |
| 115 base::TimeTicks TaskQueueManager::NextPendingDelayedTaskRunTime() { | 102 base::TimeTicks TaskQueueManager::NextPendingDelayedTaskRunTime() { |
| 116 DCHECK(main_thread_checker_.CalledOnValidThread()); | 103 DCHECK(main_thread_checker_.CalledOnValidThread()); |
| 117 bool found_pending_task = false; | 104 bool found_pending_task = false; |
| 118 base::TimeTicks next_pending_delayed_task( | 105 base::TimeTicks next_pending_delayed_task( |
| 119 base::TimeTicks::FromInternalValue(kMaxTimeTicks)); | 106 base::TimeTicks::FromInternalValue(kMaxTimeTicks)); |
| 120 for (auto& queue : queues_) { | 107 for (auto& queue : queues_) { |
| 121 base::TimeTicks queues_next_pending_delayed_task; | 108 base::TimeTicks queues_next_pending_delayed_task; |
| 122 if (queue->NextPendingDelayedTaskRunTime( | 109 if (queue->NextPendingDelayedTaskRunTime( |
| 123 &queues_next_pending_delayed_task)) { | 110 &queues_next_pending_delayed_task)) { |
| 124 found_pending_task = true; | 111 found_pending_task = true; |
| 125 next_pending_delayed_task = | 112 next_pending_delayed_task = |
| 126 std::min(next_pending_delayed_task, queues_next_pending_delayed_task); | 113 std::min(next_pending_delayed_task, queues_next_pending_delayed_task); |
| 127 } | 114 } |
| 128 } | 115 } |
| 129 | 116 |
| 130 if (!found_pending_task) | 117 if (!found_pending_task) |
| 131 return base::TimeTicks(); | 118 return base::TimeTicks(); |
| 132 | 119 |
| 133 DCHECK_NE(next_pending_delayed_task, | 120 DCHECK_NE(next_pending_delayed_task, |
| 134 base::TimeTicks::FromInternalValue(kMaxTimeTicks)); | 121 base::TimeTicks::FromInternalValue(kMaxTimeTicks)); |
| 135 return next_pending_delayed_task; | 122 return next_pending_delayed_task; |
| 136 } | 123 } |
| 137 | 124 |
| 138 void TaskQueueManager::RegisterAsUpdatableTaskQueue( | |
| 139 internal::TaskQueueImpl* queue) { | |
| 140 base::AutoLock lock(newly_updatable_lock_); | |
| 141 newly_updatable_.push_back(queue); | |
| 142 } | |
| 143 | |
| 144 void TaskQueueManager::UnregisterAsUpdatableTaskQueue( | |
| 145 internal::TaskQueueImpl* queue) { | |
| 146 DCHECK(main_thread_checker_.CalledOnValidThread()); | |
| 147 MoveNewlyUpdatableQueuesIntoUpdatableQueueSet(); | |
| 148 #ifndef NDEBUG | |
| 149 { | |
| 150 base::AutoLock lock(newly_updatable_lock_); | |
| 151 DCHECK(!(updatable_queue_set_.find(queue) == updatable_queue_set_.end() && | |
| 152 std::find(newly_updatable_.begin(), newly_updatable_.end(), | |
| 153 queue) != newly_updatable_.end())); | |
| 154 } | |
| 155 #endif | |
| 156 updatable_queue_set_.erase(queue); | |
| 157 } | |
| 158 | |
| 159 void TaskQueueManager::MoveNewlyUpdatableQueuesIntoUpdatableQueueSet() { | |
| 160 DCHECK(main_thread_checker_.CalledOnValidThread()); | |
| 161 base::AutoLock lock(newly_updatable_lock_); | |
| 162 while (!newly_updatable_.empty()) { | |
| 163 updatable_queue_set_.insert(newly_updatable_.back()); | |
| 164 newly_updatable_.pop_back(); | |
| 165 } | |
| 166 } | |
| 167 | |
| 168 void TaskQueueManager::UpdateWorkQueues( | 125 void TaskQueueManager::UpdateWorkQueues( |
| 169 bool should_trigger_wakeup, | 126 bool should_trigger_wakeup, |
| 170 const internal::TaskQueueImpl::Task* previous_task) { | 127 const internal::TaskQueueImpl::Task* previous_task) { |
| 171 DCHECK(main_thread_checker_.CalledOnValidThread()); | |
| 172 TRACE_EVENT0(disabled_by_default_tracing_category_, | 128 TRACE_EVENT0(disabled_by_default_tracing_category_, |
| 173 "TaskQueueManager::UpdateWorkQueues"); | 129 "TaskQueueManager::UpdateWorkQueues"); |
| 174 internal::LazyNow lazy_now(delegate().get()); | |
| 175 | 130 |
| 176 // Move any ready delayed tasks into the incomming queues. | 131 for (const scoped_refptr<TimeDomain>& time_domain : time_domains_) { |
| 177 WakeupReadyDelayedQueues(&lazy_now); | 132 time_domain->UpdateWorkQueues(should_trigger_wakeup, previous_task); |
| 178 | |
| 179 MoveNewlyUpdatableQueuesIntoUpdatableQueueSet(); | |
| 180 | |
| 181 auto iter = updatable_queue_set_.begin(); | |
| 182 while (iter != updatable_queue_set_.end()) { | |
| 183 internal::TaskQueueImpl* queue = *iter++; | |
| 184 // NOTE Update work queue may erase itself from |updatable_queue_set_|. | |
| 185 // This is fine, erasing an element won't invalidate any interator, as long | |
| 186 // as the iterator isn't the element being delated. | |
| 187 if (queue->work_queue().empty()) | |
| 188 queue->UpdateWorkQueue(&lazy_now, should_trigger_wakeup, previous_task); | |
| 189 } | 133 } |
| 190 } | 134 } |
| 191 | 135 |
| 192 void TaskQueueManager::ScheduleDelayedWorkTask( | |
| 193 scoped_refptr<internal::TaskQueueImpl> queue, | |
| 194 base::TimeTicks delayed_run_time) { | |
| 195 internal::LazyNow lazy_now(delegate().get()); | |
| 196 ScheduleDelayedWork(queue.get(), delayed_run_time, &lazy_now); | |
| 197 } | |
| 198 | |
| 199 void TaskQueueManager::ScheduleDelayedWork(internal::TaskQueueImpl* queue, | |
| 200 base::TimeTicks delayed_run_time, | |
| 201 internal::LazyNow* lazy_now) { | |
| 202 if (!delegate_->BelongsToCurrentThread()) { | |
| 203 // NOTE posting a delayed task from a different thread is not expected to be | |
| 204 // common. This pathway is less optimal than perhaps it could be because | |
| 205 // it causes two main thread tasks to be run. Should this assumption prove | |
| 206 // to be false in future, we may need to revisit this. | |
| 207 delegate_->PostTask( | |
| 208 FROM_HERE, base::Bind(&TaskQueueManager::ScheduleDelayedWorkTask, | |
| 209 weak_factory_.GetWeakPtr(), | |
| 210 scoped_refptr<internal::TaskQueueImpl>(queue), | |
| 211 delayed_run_time)); | |
| 212 return; | |
| 213 } | |
| 214 | |
| 215 // Make sure there's one (and only one) task posted to |delegate_| | |
| 216 // to call |DelayedDoWork| at |delayed_run_time|. | |
| 217 if (delayed_wakeup_multimap_.find(delayed_run_time) == | |
| 218 delayed_wakeup_multimap_.end()) { | |
| 219 base::TimeDelta delay = | |
| 220 std::max(base::TimeDelta(), delayed_run_time - lazy_now->Now()); | |
| 221 delegate_->PostDelayedTask(FROM_HERE, do_work_closure_, delay); | |
| 222 } | |
| 223 delayed_wakeup_multimap_.insert(std::make_pair(delayed_run_time, queue)); | |
| 224 } | |
| 225 | |
| 226 void TaskQueueManager::WakeupReadyDelayedQueues(internal::LazyNow* lazy_now) { | |
| 227 // Wake up any queues with pending delayed work. Note std::multipmap stores | |
| 228 // the elements sorted by key, so the begin() iterator points to the earliest | |
| 229 // queue to wakeup. | |
| 230 std::set<internal::TaskQueueImpl*> dedup_set; | |
| 231 while (!delayed_wakeup_multimap_.empty()) { | |
| 232 DelayedWakeupMultimap::iterator next_wakeup = | |
| 233 delayed_wakeup_multimap_.begin(); | |
| 234 if (next_wakeup->first > lazy_now->Now()) | |
| 235 break; | |
| 236 // A queue could have any number of delayed tasks pending so it's worthwhile | |
| 237 // deduping calls to MoveReadyDelayedTasksToIncomingQueue since it takes a | |
| 238 // lock. NOTE the order in which these are called matters since the order | |
| 239 // in which EnqueueTaskLocks is called is respected when choosing which | |
| 240 // queue to execute a task from. | |
| 241 if (dedup_set.insert(next_wakeup->second).second) | |
| 242 next_wakeup->second->MoveReadyDelayedTasksToIncomingQueue(lazy_now); | |
| 243 delayed_wakeup_multimap_.erase(next_wakeup); | |
| 244 } | |
| 245 } | |
| 246 | |
| 247 void TaskQueueManager::MaybePostDoWorkOnMainRunner() { | 136 void TaskQueueManager::MaybePostDoWorkOnMainRunner() { |
| 248 bool on_main_thread = delegate_->BelongsToCurrentThread(); | 137 bool on_main_thread = delegate_->BelongsToCurrentThread(); |
| 249 if (on_main_thread) { | 138 if (on_main_thread) { |
| 250 // We only want one pending DoWork posted from the main thread, or we risk | 139 // We only want one pending DoWork posted from the main thread, or we risk |
| 251 // an explosion of pending DoWorks which could starve out everything else. | 140 // an explosion of pending DoWorks which could starve out everything else. |
| 252 if (pending_dowork_count_ > 0) { | 141 if (pending_dowork_count_ > 0) { |
| 253 return; | 142 return; |
| 254 } | 143 } |
| 255 pending_dowork_count_++; | 144 pending_dowork_count_++; |
| 256 delegate_->PostTask(FROM_HERE, decrement_pending_and_do_work_closure_); | 145 delegate_->PostTask(FROM_HERE, decrement_pending_and_do_work_closure_); |
| (...skipping 39 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 296 | 185 |
| 297 // Only run a single task per batch in nested run loops so that we can | 186 // Only run a single task per batch in nested run loops so that we can |
| 298 // properly exit the nested loop when someone calls RunLoop::Quit(). | 187 // properly exit the nested loop when someone calls RunLoop::Quit(). |
| 299 if (delegate_->IsNested()) | 188 if (delegate_->IsNested()) |
| 300 break; | 189 break; |
| 301 } | 190 } |
| 302 | 191 |
| 303 // TODO(alexclarke): Consider refactoring the above loop to terminate only | 192 // TODO(alexclarke): Consider refactoring the above loop to terminate only |
| 304 // when there's no more work left to be done, rather than posting a | 193 // when there's no more work left to be done, rather than posting a |
| 305 // continuation task. | 194 // continuation task. |
| 306 if (!selector_.EnabledWorkQueuesEmpty()) { | 195 if (!selector_.EnabledWorkQueuesEmpty() || ATimeDomainCanAdvance()) { |
| 307 MaybePostDoWorkOnMainRunner(); | 196 MaybePostDoWorkOnMainRunner(); |
| 308 } else { | 197 } else { |
| 309 // Tell the task runner we have no more work. | 198 // Tell the task runner we have no more work. |
| 310 delegate_->OnNoMoreImmediateWork(); | 199 delegate_->OnNoMoreImmediateWork(); |
| 311 } | 200 } |
| 312 } | 201 } |
| 313 | 202 |
| 203 bool TaskQueueManager::ATimeDomainCanAdvance() { |
| 204 bool can_advance = false; |
| 205 for (const scoped_refptr<TimeDomain>& time_domain : time_domains_) { |
| 206 can_advance |= time_domain->MaybeAdvanceTime(); |
| 207 } |
| 208 return can_advance; |
| 209 } |
| 210 |
| 314 bool TaskQueueManager::SelectQueueToService( | 211 bool TaskQueueManager::SelectQueueToService( |
| 315 internal::TaskQueueImpl** out_queue) { | 212 internal::TaskQueueImpl** out_queue) { |
| 316 bool should_run = selector_.SelectQueueToService(out_queue); | 213 bool should_run = selector_.SelectQueueToService(out_queue); |
| 317 TRACE_EVENT_OBJECT_SNAPSHOT_WITH_ID( | 214 TRACE_EVENT_OBJECT_SNAPSHOT_WITH_ID( |
| 318 disabled_by_default_tracing_category_, "TaskQueueManager", this, | 215 disabled_by_default_tracing_category_, "TaskQueueManager", this, |
| 319 AsValueWithSelectorResult(should_run, *out_queue)); | 216 AsValueWithSelectorResult(should_run, *out_queue)); |
| 320 return should_run; | 217 return should_run; |
| 321 } | 218 } |
| 322 | 219 |
| 323 void TaskQueueManager::DidQueueTask( | 220 void TaskQueueManager::DidQueueTask( |
| (...skipping 95 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 419 state->BeginArray("queues"); | 316 state->BeginArray("queues"); |
| 420 for (auto& queue : queues_) | 317 for (auto& queue : queues_) |
| 421 queue->AsValueInto(state.get()); | 318 queue->AsValueInto(state.get()); |
| 422 state->EndArray(); | 319 state->EndArray(); |
| 423 state->BeginDictionary("selector"); | 320 state->BeginDictionary("selector"); |
| 424 selector_.AsValueInto(state.get()); | 321 selector_.AsValueInto(state.get()); |
| 425 state->EndDictionary(); | 322 state->EndDictionary(); |
| 426 if (should_run) | 323 if (should_run) |
| 427 state->SetString("selected_queue", selected_queue->GetName()); | 324 state->SetString("selected_queue", selected_queue->GetName()); |
| 428 | 325 |
| 429 state->BeginArray("updatable_queue_set"); | 326 state->BeginArray("time_domains"); |
| 430 for (auto& queue : updatable_queue_set_) | 327 for (auto& time_domain : time_domains_) |
| 431 state->AppendString(queue->GetName()); | 328 time_domain->AsValueInto(state.get()); |
| 432 state->EndArray(); | 329 state->EndArray(); |
| 433 return state; | 330 return state; |
| 434 } | 331 } |
| 435 | 332 |
| 436 void TaskQueueManager::OnTaskQueueEnabled(internal::TaskQueueImpl* queue) { | 333 void TaskQueueManager::OnTaskQueueEnabled(internal::TaskQueueImpl* queue) { |
| 437 DCHECK(main_thread_checker_.CalledOnValidThread()); | 334 DCHECK(main_thread_checker_.CalledOnValidThread()); |
| 438 // Only schedule DoWork if there's something to do. | 335 // Only schedule DoWork if there's something to do. |
| 439 if (!queue->work_queue().empty()) | 336 if (!queue->work_queue().empty()) |
| 440 MaybePostDoWorkOnMainRunner(); | 337 MaybePostDoWorkOnMainRunner(); |
| 441 } | 338 } |
| 442 | 339 |
| 443 } // namespace scheduler | 340 } // namespace scheduler |
| OLD | NEW |