OLD | NEW |
(Empty) | |
| 1 // Copyright (c) 2017 The Chromium Authors. All rights reserved. |
| 2 // Use of this source code is governed by a BSD-style license that can be |
| 3 // found in the LICENSE file. |
| 4 |
| 5 #include "gpu/command_buffer/service/scheduler.h" |
| 6 |
| 7 #include <algorithm> |
| 8 |
| 9 #include "base/callback.h" |
| 10 #include "base/memory/ptr_util.h" |
| 11 #include "base/stl_util.h" |
| 12 #include "base/trace_event/trace_event.h" |
| 13 #include "base/trace_event/trace_event_argument.h" |
| 14 #include "gpu/command_buffer/service/sync_point_manager.h" |
| 15 |
| 16 namespace gpu { |
| 17 |
| 18 class Scheduler::Sequence { |
| 19 public: |
| 20 Sequence(SequenceId sequence_id, |
| 21 GpuStreamPriority priority, |
| 22 scoped_refptr<SyncPointOrderData> order_data); |
| 23 |
| 24 ~Sequence(); |
| 25 |
| 26 void Destroy(); |
| 27 |
| 28 bool destroyed() const { return destroyed_; } |
| 29 |
| 30 SequenceId sequence_id() const { return sequence_id_; } |
| 31 |
| 32 const SchedulingState& scheduling_state() const { return scheduling_state_; } |
| 33 |
| 34 bool enabled() const { return enabled_; } |
| 35 |
| 36 bool scheduled() const { return running_state_ == SCHEDULED; } |
| 37 |
| 38 bool running() const { return running_state_ == RUNNING; } |
| 39 |
| 40 // The sequence is runnable if its enabled and has tasks which are not blocked |
| 41 // by wait fences. |
| 42 bool IsRunnable() const; |
| 43 |
| 44 bool NeedsRescheduling() const; |
| 45 |
| 46 void UpdateSchedulingState(); |
| 47 |
| 48 // If this sequence runs before the other sequence. |
| 49 bool RunsBefore(const Sequence* other) const; |
| 50 |
| 51 void SetEnabled(bool enabled); |
| 52 |
| 53 // Sets running state to SCHEDULED. |
| 54 void SetScheduled(); |
| 55 |
| 56 // Called before running the next task on the sequence. Returns the closure |
| 57 // for the task. Sets running state to RUNNING. |
| 58 base::OnceClosure BeginTask(); |
| 59 |
| 60 // Called after running the closure returned by BeginTask. Sets running state |
| 61 // to IDLE. |
| 62 void FinishTask(); |
| 63 |
| 64 // Enqueues a task in the sequence and returns the generated order number. |
| 65 uint32_t ScheduleTask(base::OnceClosure closure); |
| 66 |
| 67 // Continue running the current task with the given closure. Must be called in |
| 68 // between |BeginTask| and |FinishTask|. |
| 69 void ContinueTask(base::OnceClosure closure); |
| 70 |
| 71 // Add a sync token fence that this sequence should wait on. |
| 72 void AddWaitFence(const SyncToken& sync_token, uint32_t order_num); |
| 73 |
| 74 // Remove a waiting sync token fence. |
| 75 void RemoveWaitFence(const SyncToken& sync_token, uint32_t order_num); |
| 76 |
| 77 // Add a sync token fence that this sequence is expected to release. |
| 78 void AddReleaseFence(const SyncToken& sync_token, uint32_t order_num); |
| 79 |
| 80 // Remove a release sync token fence. |
| 81 void RemoveReleaseFence(const SyncToken& sync_token, uint32_t order_num); |
| 82 |
| 83 private: |
| 84 enum RunningState { IDLE, SCHEDULED, RUNNING }; |
| 85 |
| 86 struct Fence { |
| 87 SyncToken sync_token; |
| 88 uint32_t order_num; |
| 89 |
| 90 bool operator==(const Fence& other) const { |
| 91 return std::tie(sync_token, order_num) == |
| 92 std::tie(other.sync_token, other.order_num); |
| 93 } |
| 94 }; |
| 95 |
| 96 struct Task { |
| 97 base::OnceClosure closure; |
| 98 uint32_t order_num; |
| 99 }; |
| 100 |
| 101 GpuStreamPriority GetSchedulingPriority() const; |
| 102 |
| 103 bool destroyed_ = false; |
| 104 |
| 105 // If the sequence is enabled. Sequences are disabled/enabled based on when |
| 106 // the command buffer is descheduled/sc |
| 107 bool enabled_ = true; |
| 108 |
| 109 RunningState running_state_; |
| 110 |
| 111 // Cached scheduling state used for comparison with other sequences using |
| 112 // |RunsBefore|. Updated in |UpdateSchedulingState|. |
| 113 SchedulingState scheduling_state_; |
| 114 |
| 115 const SequenceId sequence_id_; |
| 116 |
| 117 const GpuStreamPriority priority_; |
| 118 |
| 119 scoped_refptr<SyncPointOrderData> order_data_; |
| 120 |
| 121 // Deque of tasks. Tasks are inserted at the back with increasing order number |
| 122 // generated from SyncPointOrderData. If a running task needs to be continued, |
| 123 // it is inserted at the front with the same order number. |
| 124 std::deque<Task> tasks_; |
| 125 |
| 126 // List of fences that this sequence is waiting on. Fences are inserted in |
| 127 // increasing order number but may be removed out of order. Tasks are blocked |
| 128 // if there's a wait fence with order number less than or equal to the task's |
| 129 // order number. |
| 130 std::vector<Fence> wait_fences_; |
| 131 |
| 132 // List of fences that this sequence is expected to release. If this list is |
| 133 // non-empty, the priority of the sequence is raised. |
| 134 std::vector<Fence> release_fences_; |
| 135 |
| 136 DISALLOW_COPY_AND_ASSIGN(Sequence); |
| 137 }; |
| 138 |
| 139 Scheduler::SchedulingState::SchedulingState() = default; |
| 140 Scheduler::SchedulingState::SchedulingState(const SchedulingState& other) = |
| 141 default; |
| 142 Scheduler::SchedulingState::~SchedulingState() = default; |
| 143 |
| 144 std::unique_ptr<base::trace_event::ConvertableToTraceFormat> |
| 145 Scheduler::SchedulingState::AsValue() const { |
| 146 std::unique_ptr<base::trace_event::TracedValue> state( |
| 147 new base::trace_event::TracedValue()); |
| 148 state->SetInteger("sequence_id", sequence_id.GetUnsafeValue()); |
| 149 state->SetString("priority", GpuStreamPriorityToString(priority)); |
| 150 state->SetInteger("order_num", order_num); |
| 151 return std::move(state); |
| 152 } |
| 153 |
| 154 Scheduler::Sequence::Sequence(SequenceId sequence_id, |
| 155 GpuStreamPriority priority, |
| 156 scoped_refptr<SyncPointOrderData> order_data) |
| 157 : sequence_id_(sequence_id), priority_(priority), order_data_(order_data) {} |
| 158 |
| 159 Scheduler::Sequence::~Sequence() = default; |
| 160 |
| 161 void Scheduler::Sequence::Destroy() { |
| 162 DCHECK(!destroyed_); |
| 163 destroyed_ = true; |
| 164 } |
| 165 |
| 166 bool Scheduler::Sequence::NeedsRescheduling() const { |
| 167 return running_state_ != IDLE && |
| 168 scheduling_state_.priority != GetSchedulingPriority(); |
| 169 } |
| 170 |
| 171 bool Scheduler::Sequence::IsRunnable() const { |
| 172 return enabled_ && !tasks_.empty() && |
| 173 (wait_fences_.empty() || |
| 174 wait_fences_.front().order_num > tasks_.front().order_num); |
| 175 } |
| 176 |
| 177 GpuStreamPriority Scheduler::Sequence::GetSchedulingPriority() const { |
| 178 if (!release_fences_.empty()) |
| 179 return std::min(priority_, GpuStreamPriority::HIGH); |
| 180 return priority_; |
| 181 } |
| 182 |
| 183 bool Scheduler::Sequence::RunsBefore(const Scheduler::Sequence* other) const { |
| 184 return other->scheduling_state() < scheduling_state_; |
| 185 } |
| 186 |
| 187 void Scheduler::Sequence::SetEnabled(bool enabled) { |
| 188 if (enabled_ == enabled) |
| 189 return; |
| 190 DCHECK_EQ(running_state_, enabled ? IDLE : RUNNING); |
| 191 enabled_ = enabled; |
| 192 } |
| 193 |
| 194 void Scheduler::Sequence::SetScheduled() { |
| 195 DCHECK_NE(running_state_, RUNNING); |
| 196 running_state_ = SCHEDULED; |
| 197 UpdateSchedulingState(); |
| 198 } |
| 199 |
| 200 void Scheduler::Sequence::UpdateSchedulingState() { |
| 201 scheduling_state_.sequence_id = sequence_id_; |
| 202 scheduling_state_.priority = GetSchedulingPriority(); |
| 203 |
| 204 uint32_t order_num = UINT32_MAX; // IDLE |
| 205 if (running_state_ == SCHEDULED) { |
| 206 DCHECK(!tasks_.empty()); |
| 207 order_num = tasks_.front().order_num; |
| 208 } else if (running_state_ == RUNNING) { |
| 209 order_num = order_data_->current_order_num(); |
| 210 } |
| 211 scheduling_state_.order_num = order_num; |
| 212 } |
| 213 |
| 214 void Scheduler::Sequence::ContinueTask(base::OnceClosure closure) { |
| 215 DCHECK_EQ(running_state_, RUNNING); |
| 216 tasks_.push_front({std::move(closure), order_data_->current_order_num()}); |
| 217 } |
| 218 |
| 219 uint32_t Scheduler::Sequence::ScheduleTask(base::OnceClosure closure) { |
| 220 uint32_t order_num = order_data_->GenerateUnprocessedOrderNumber(); |
| 221 tasks_.push_back({std::move(closure), order_num}); |
| 222 return order_num; |
| 223 } |
| 224 |
| 225 base::OnceClosure Scheduler::Sequence::BeginTask() { |
| 226 DCHECK(!tasks_.empty()); |
| 227 |
| 228 DCHECK_EQ(running_state_, SCHEDULED); |
| 229 running_state_ = RUNNING; |
| 230 |
| 231 base::OnceClosure closure = std::move(tasks_.front().closure); |
| 232 uint32_t order_num = tasks_.front().order_num; |
| 233 tasks_.pop_front(); |
| 234 |
| 235 order_data_->BeginProcessingOrderNumber(order_num); |
| 236 |
| 237 UpdateSchedulingState(); |
| 238 |
| 239 return closure; |
| 240 } |
| 241 |
| 242 void Scheduler::Sequence::FinishTask() { |
| 243 DCHECK_EQ(running_state_, RUNNING); |
| 244 running_state_ = IDLE; |
| 245 uint32_t order_num = order_data_->current_order_num(); |
| 246 if (!tasks_.empty() && tasks_.front().order_num == order_num) { |
| 247 order_data_->PauseProcessingOrderNumber(order_num); |
| 248 } else { |
| 249 order_data_->FinishProcessingOrderNumber(order_num); |
| 250 } |
| 251 UpdateSchedulingState(); |
| 252 } |
| 253 |
| 254 void Scheduler::Sequence::AddWaitFence(const SyncToken& sync_token, |
| 255 uint32_t order_num) { |
| 256 wait_fences_.push_back({sync_token, order_num}); |
| 257 } |
| 258 |
| 259 void Scheduler::Sequence::RemoveWaitFence(const SyncToken& sync_token, |
| 260 uint32_t order_num) { |
| 261 base::Erase(wait_fences_, Fence{sync_token, order_num}); |
| 262 } |
| 263 |
| 264 void Scheduler::Sequence::AddReleaseFence(const SyncToken& sync_token, |
| 265 uint32_t order_num) { |
| 266 release_fences_.push_back({sync_token, order_num}); |
| 267 } |
| 268 |
| 269 void Scheduler::Sequence::RemoveReleaseFence(const SyncToken& sync_token, |
| 270 uint32_t order_num) { |
| 271 base::Erase(release_fences_, Fence{sync_token, order_num}); |
| 272 } |
| 273 |
| 274 Scheduler::Scheduler(scoped_refptr<base::SingleThreadTaskRunner> task_runner, |
| 275 SyncPointManager* sync_point_manager) |
| 276 : task_runner_(std::move(task_runner)), |
| 277 sync_point_manager_(sync_point_manager), |
| 278 weak_factory_(this) { |
| 279 DCHECK(thread_checker_.CalledOnValidThread()); |
| 280 } |
| 281 |
| 282 Scheduler::~Scheduler() { |
| 283 DCHECK(thread_checker_.CalledOnValidThread()); |
| 284 } |
| 285 |
| 286 SequenceId Scheduler::CreateSequence(GpuStreamPriority priority) { |
| 287 DCHECK(thread_checker_.CalledOnValidThread()); |
| 288 base::AutoLock auto_lock(lock_); |
| 289 scoped_refptr<SyncPointOrderData> order_data = |
| 290 sync_point_manager_->CreateSyncPointOrderData(); |
| 291 SequenceId sequence_id = order_data->sequence_id(); |
| 292 std::unique_ptr<Sequence> sequence = |
| 293 base::MakeUnique<Sequence>(sequence_id, priority, std::move(order_data)); |
| 294 sequences_.emplace(sequence_id, std::move(sequence)); |
| 295 return sequence_id; |
| 296 } |
| 297 |
| 298 void Scheduler::DestroySequence(SequenceId sequence_id) { |
| 299 DCHECK(thread_checker_.CalledOnValidThread()); |
| 300 base::AutoLock auto_lock(lock_); |
| 301 |
| 302 Sequence* sequence = GetSequence(sequence_id); |
| 303 DCHECK(sequence); |
| 304 sequence->Destroy(); |
| 305 |
| 306 if (sequence->running()) |
| 307 return; |
| 308 |
| 309 if (sequence->scheduled()) |
| 310 rebuild_scheduling_queue_ = true; |
| 311 sequences_.erase(sequence_id); |
| 312 } |
| 313 |
| 314 Scheduler::Sequence* Scheduler::GetSequence(SequenceId sequence_id) { |
| 315 lock_.AssertAcquired(); |
| 316 auto it = sequences_.find(sequence_id); |
| 317 if (it != sequences_.end()) |
| 318 return it->second.get(); |
| 319 return nullptr; |
| 320 } |
| 321 |
| 322 void Scheduler::EnableSequence(SequenceId sequence_id) { |
| 323 DCHECK(thread_checker_.CalledOnValidThread()); |
| 324 base::AutoLock auto_lock(lock_); |
| 325 Sequence* sequence = GetSequence(sequence_id); |
| 326 DCHECK(sequence); |
| 327 sequence->SetEnabled(true); |
| 328 TryScheduleSequence(sequence); |
| 329 } |
| 330 |
| 331 void Scheduler::DisableSequence(SequenceId sequence_id) { |
| 332 DCHECK(thread_checker_.CalledOnValidThread()); |
| 333 base::AutoLock auto_lock(lock_); |
| 334 Sequence* sequence = GetSequence(sequence_id); |
| 335 DCHECK(sequence); |
| 336 sequence->SetEnabled(false); |
| 337 } |
| 338 |
| 339 void Scheduler::ScheduleTask(SequenceId sequence_id, |
| 340 base::OnceClosure closure, |
| 341 const std::vector<SyncToken>& sync_token_fences) { |
| 342 base::AutoLock auto_lock(lock_); |
| 343 Sequence* sequence = GetSequence(sequence_id); |
| 344 DCHECK(sequence); |
| 345 |
| 346 uint32_t order_num = sequence->ScheduleTask(std::move(closure)); |
| 347 |
| 348 for (const SyncToken& sync_token : sync_token_fences) { |
| 349 SequenceId release_id = |
| 350 sync_point_manager_->GetSyncTokenReleaseSequenceId(sync_token); |
| 351 Sequence* release_sequence = GetSequence(release_id); |
| 352 if (!release_sequence) |
| 353 continue; |
| 354 if (sync_point_manager_->Wait( |
| 355 sync_token, order_num, |
| 356 base::Bind(&Scheduler::SyncTokenFenceReleased, |
| 357 weak_factory_.GetWeakPtr(), sync_token, order_num, |
| 358 release_id, sequence_id))) { |
| 359 sequence->AddWaitFence(sync_token, order_num); |
| 360 release_sequence->AddReleaseFence(sync_token, order_num); |
| 361 TryScheduleSequence(release_sequence); |
| 362 } |
| 363 } |
| 364 |
| 365 TryScheduleSequence(sequence); |
| 366 } |
| 367 |
| 368 void Scheduler::ContinueTask(SequenceId sequence_id, |
| 369 base::OnceClosure closure) { |
| 370 DCHECK(thread_checker_.CalledOnValidThread()); |
| 371 base::AutoLock auto_lock(lock_); |
| 372 Sequence* sequence = GetSequence(sequence_id); |
| 373 DCHECK(sequence); |
| 374 sequence->ContinueTask(std::move(closure)); |
| 375 } |
| 376 |
| 377 bool Scheduler::ShouldYield(SequenceId sequence_id) { |
| 378 DCHECK(thread_checker_.CalledOnValidThread()); |
| 379 base::AutoLock auto_lock(lock_); |
| 380 |
| 381 Sequence* sequence = GetSequence(sequence_id); |
| 382 DCHECK(sequence); |
| 383 DCHECK(sequence->running()); |
| 384 |
| 385 if (should_yield_) |
| 386 return true; |
| 387 |
| 388 RebuildSchedulingQueue(); |
| 389 |
| 390 sequence->UpdateSchedulingState(); |
| 391 |
| 392 if (!scheduling_queue_.empty()) { |
| 393 Sequence* next_sequence = GetSequence(scheduling_queue_.top().sequence_id); |
| 394 DCHECK(next_sequence); |
| 395 if (next_sequence->RunsBefore(sequence)) |
| 396 should_yield_ = true; |
| 397 } |
| 398 |
| 399 return should_yield_; |
| 400 } |
| 401 |
| 402 void Scheduler::SyncTokenFenceReleased(const SyncToken& sync_token, |
| 403 uint32_t order_num, |
| 404 SequenceId release_sequence_id, |
| 405 SequenceId waiting_sequence_id) { |
| 406 base::AutoLock auto_lock(lock_); |
| 407 Sequence* sequence = GetSequence(waiting_sequence_id); |
| 408 if (sequence) { |
| 409 sequence->RemoveWaitFence(sync_token, order_num); |
| 410 TryScheduleSequence(sequence); |
| 411 } |
| 412 Sequence* release_sequence = GetSequence(release_sequence_id); |
| 413 if (release_sequence) { |
| 414 release_sequence->RemoveReleaseFence(sync_token, order_num); |
| 415 TryScheduleSequence(release_sequence); |
| 416 } |
| 417 } |
| 418 |
| 419 void Scheduler::TryScheduleSequence(Sequence* sequence) { |
| 420 lock_.AssertAcquired(); |
| 421 |
| 422 if (sequence->running()) |
| 423 return; |
| 424 |
| 425 if (sequence->NeedsRescheduling()) { |
| 426 DCHECK(sequence->IsRunnable()); |
| 427 rebuild_scheduling_queue_ = true; |
| 428 } else if (!sequence->scheduled() && sequence->IsRunnable()) { |
| 429 sequence->SetScheduled(); |
| 430 scheduling_queue_.push(sequence->scheduling_state()); |
| 431 } |
| 432 |
| 433 if (!running_) { |
| 434 TRACE_EVENT_ASYNC_BEGIN0("gpu", "Scheduler::Running", this); |
| 435 running_ = true; |
| 436 task_runner_->PostTask(FROM_HERE, base::Bind(&Scheduler::RunNextTask, |
| 437 weak_factory_.GetWeakPtr())); |
| 438 } |
| 439 } |
| 440 |
| 441 void Scheduler::RebuildSchedulingQueue() { |
| 442 DCHECK(thread_checker_.CalledOnValidThread()); |
| 443 lock_.AssertAcquired(); |
| 444 |
| 445 if (!rebuild_scheduling_queue_) |
| 446 return; |
| 447 rebuild_scheduling_queue_ = false; |
| 448 |
| 449 std::vector<SchedulingState> states; |
| 450 for (const auto& kv : sequences_) { |
| 451 Sequence* sequence = kv.second.get(); |
| 452 if (!sequence->IsRunnable() || sequence->running()) |
| 453 continue; |
| 454 sequence->SetScheduled(); |
| 455 states.push_back(sequence->scheduling_state()); |
| 456 } |
| 457 |
| 458 scheduling_queue_ = SchedulingQueue(states.begin(), states.end()); |
| 459 } |
| 460 |
| 461 void Scheduler::RunNextTask() { |
| 462 DCHECK(thread_checker_.CalledOnValidThread()); |
| 463 base::AutoLock auto_lock(lock_); |
| 464 |
| 465 should_yield_ = false; |
| 466 |
| 467 RebuildSchedulingQueue(); |
| 468 |
| 469 if (scheduling_queue_.empty()) { |
| 470 TRACE_EVENT_ASYNC_END0("gpu", "Scheduler::Running", this); |
| 471 running_ = false; |
| 472 return; |
| 473 } |
| 474 |
| 475 SchedulingState state = scheduling_queue_.top(); |
| 476 scheduling_queue_.pop(); |
| 477 |
| 478 TRACE_EVENT1("gpu", "Scheduler::RunNextTask", "state", state.AsValue()); |
| 479 |
| 480 Sequence* sequence = GetSequence(state.sequence_id); |
| 481 DCHECK(sequence); |
| 482 |
| 483 base::OnceClosure closure = sequence->BeginTask(); |
| 484 |
| 485 { |
| 486 base::AutoUnlock auto_unlock(lock_); |
| 487 std::move(closure).Run(); |
| 488 } |
| 489 |
| 490 sequence->FinishTask(); |
| 491 |
| 492 if (sequence->destroyed()) { |
| 493 sequences_.erase(sequence->sequence_id()); |
| 494 } else if (sequence->IsRunnable()) { |
| 495 sequence->SetScheduled(); |
| 496 scheduling_queue_.push(sequence->scheduling_state()); |
| 497 } |
| 498 |
| 499 task_runner_->PostTask(FROM_HERE, base::Bind(&Scheduler::RunNextTask, |
| 500 weak_factory_.GetWeakPtr())); |
| 501 } |
| 502 |
| 503 } // namespace gpu |
OLD | NEW |