Chromium Code Reviews| OLD | NEW |
|---|---|
| (Empty) | |
| 1 // Copyright (c) 2017 The Chromium Authors. All rights reserved. | |
| 2 // Use of this source code is governed by a BSD-style license that can be | |
| 3 // found in the LICENSE file. | |
| 4 | |
| 5 #include "gpu/command_buffer/service/scheduler.h" | |
| 6 | |
| 7 #include <algorithm> | |
| 8 | |
| 9 #include "base/callback.h" | |
| 10 #include "base/memory/ptr_util.h" | |
| 11 #include "base/stl_util.h" | |
| 12 #include "base/trace_event/trace_event.h" | |
| 13 #include "base/trace_event/trace_event_argument.h" | |
| 14 #include "gpu/command_buffer/service/sync_point_manager.h" | |
| 15 | |
| 16 namespace gpu { | |
| 17 | |
| 18 class Scheduler::Sequence { | |
| 19 public: | |
| 20 Sequence(SequenceId sequence_id, | |
| 21 GpuStreamPriority priority, | |
| 22 scoped_refptr<SyncPointOrderData> order_data); | |
| 23 | |
| 24 ~Sequence(); | |
| 25 | |
| 26 void Destroy(); | |
| 27 | |
| 28 bool destroyed() const { return destroyed_; } | |
| 29 | |
| 30 SequenceId sequence_id() const { return sequence_id_; } | |
| 31 | |
| 32 const SchedulingState& scheduling_state() const { return scheduling_state_; } | |
| 33 | |
| 34 bool enabled() const { return enabled_; } | |
| 35 | |
| 36 bool scheduled() const { return running_state_ == SCHEDULED; } | |
| 37 | |
| 38 bool running() const { return running_state_ == RUNNING; } | |
| 39 | |
| 40 // The sequence is runnable if its enabled and has tasks which are not blocked | |
| 41 // by wait fences. | |
| 42 bool IsRunnable() const; | |
| 43 | |
| 44 bool NeedsRescheduling() const; | |
| 45 | |
| 46 void UpdateSchedulingState(); | |
| 47 | |
| 48 // If this sequence runs before the other sequence. | |
| 49 bool RunsBefore(const Sequence* other) const; | |
| 50 | |
| 51 void SetEnabled(bool enabled); | |
| 52 | |
| 53 // Sets running state to SCHEDULED. | |
| 54 void SetScheduled(); | |
| 55 | |
| 56 // Called before running the next task on the sequence. Returns the closure | |
| 57 // for the task. Sets running state to RUNNING. | |
| 58 base::OnceClosure BeginTask(); | |
| 59 | |
| 60 // Called after running the closure returned by BeginTask. Sets running state | |
| 61 // to IDLE. | |
| 62 void FinishTask(); | |
| 63 | |
| 64 // Enqueues a task in the sequence and returns the generated order number. | |
| 65 uint32_t ScheduleTask(base::OnceClosure closure); | |
| 66 | |
| 67 // Continue running the current task with the given closure. Must be called in | |
| 68 // between |BeginTask| and |FinishTask|. | |
| 69 void ContinueTask(base::OnceClosure closure); | |
| 70 | |
| 71 // Add a sync token fence that this sequence should wait on. | |
| 72 void AddWaitFence(const SyncToken& sync_token, uint32_t order_num); | |
| 73 | |
| 74 // Remove a waiting sync token fence. | |
| 75 void RemoveWaitFence(const SyncToken& sync_token, uint32_t order_num); | |
| 76 | |
| 77 // Add a sync token fence that this sequence is expected to release. | |
| 78 void AddReleaseFence(const SyncToken& sync_token, uint32_t order_num); | |
| 79 | |
| 80 // Remove a release sync token fence. | |
| 81 void RemoveReleaseFence(const SyncToken& sync_token, uint32_t order_num); | |
| 82 | |
| 83 private: | |
| 84 enum RunningState { IDLE, SCHEDULED, RUNNING }; | |
| 85 | |
| 86 struct Fence { | |
| 87 SyncToken sync_token; | |
| 88 uint32_t order_num; | |
| 89 | |
| 90 bool operator==(const Fence& other) const { | |
| 91 return std::tie(sync_token, order_num) == | |
| 92 std::tie(other.sync_token, other.order_num); | |
| 93 } | |
| 94 }; | |
| 95 | |
| 96 struct Task { | |
| 97 base::OnceClosure closure; | |
| 98 uint32_t order_num; | |
| 99 }; | |
| 100 | |
| 101 GpuStreamPriority GetSchedulingPriority() const; | |
| 102 | |
| 103 bool destroyed_ = false; | |
| 104 | |
| 105 // If the sequence is enabled. Sequences are disabled/enabled based on when | |
| 106 // the command buffer is descheduled/sc | |
|
piman
2017/05/10 00:38:50
nit: "/sc" typo or missing end of comment?
sunnyps
2017/05/10 23:15:15
Done.
| |
| 107 bool enabled_ = true; | |
| 108 | |
| 109 RunningState running_state_; | |
| 110 | |
| 111 // Cached scheduling state used for comparison with other sequences using | |
| 112 // |RunsBefore|. Updated in |UpdateSchedulingState|. | |
| 113 SchedulingState scheduling_state_; | |
| 114 | |
| 115 const SequenceId sequence_id_; | |
| 116 | |
| 117 const GpuStreamPriority priority_; | |
| 118 | |
| 119 scoped_refptr<SyncPointOrderData> order_data_; | |
| 120 | |
| 121 // Deque of tasks. Tasks are inserted at the back with increasing order number | |
| 122 // generated from SyncPointOrderData. If a running task needs to be continued, | |
| 123 // it is inserted at the front with the same order number. | |
| 124 std::deque<Task> tasks_; | |
| 125 | |
| 126 // List of fences that this sequence is waiting on. Fences are inserted in | |
| 127 // increasing order number but may be removed out of order. Tasks are blocked | |
| 128 // if there's a wait fence with order number less than or equal to the task's | |
| 129 // order number. | |
| 130 std::vector<Fence> wait_fences_; | |
| 131 | |
| 132 // List of fences that this sequence is expected to release. If this list is | |
| 133 // non-empty, the priority of the sequence is raised. | |
| 134 std::vector<Fence> release_fences_; | |
| 135 | |
| 136 DISALLOW_COPY_AND_ASSIGN(Sequence); | |
| 137 }; | |
| 138 | |
| 139 Scheduler::SchedulingState::SchedulingState() = default; | |
| 140 Scheduler::SchedulingState::SchedulingState(const SchedulingState& other) = | |
| 141 default; | |
| 142 Scheduler::SchedulingState::~SchedulingState() = default; | |
| 143 | |
| 144 std::unique_ptr<base::trace_event::ConvertableToTraceFormat> | |
| 145 Scheduler::SchedulingState::AsValue() const { | |
| 146 std::unique_ptr<base::trace_event::TracedValue> state( | |
| 147 new base::trace_event::TracedValue()); | |
| 148 state->SetInteger("sequence_id", sequence_id.GetUnsafeValue()); | |
| 149 state->SetString("priority", GpuStreamPriorityToString(priority)); | |
| 150 state->SetInteger("order_num", order_num); | |
| 151 return std::move(state); | |
| 152 } | |
| 153 | |
| 154 Scheduler::Sequence::Sequence(SequenceId sequence_id, | |
| 155 GpuStreamPriority priority, | |
| 156 scoped_refptr<SyncPointOrderData> order_data) | |
| 157 : sequence_id_(sequence_id), priority_(priority), order_data_(order_data) {} | |
| 158 | |
| 159 Scheduler::Sequence::~Sequence() = default; | |
| 160 | |
| 161 void Scheduler::Sequence::Destroy() { | |
| 162 DCHECK(!destroyed_); | |
| 163 destroyed_ = true; | |
| 164 } | |
| 165 | |
| 166 bool Scheduler::Sequence::NeedsRescheduling() const { | |
| 167 return running_state_ != IDLE && | |
| 168 scheduling_state_.priority != GetSchedulingPriority(); | |
| 169 } | |
| 170 | |
| 171 bool Scheduler::Sequence::IsRunnable() const { | |
| 172 return enabled_ && !tasks_.empty() && | |
| 173 (wait_fences_.empty() || | |
| 174 wait_fences_.front().order_num > tasks_.front().order_num); | |
| 175 } | |
| 176 | |
| 177 GpuStreamPriority Scheduler::Sequence::GetSchedulingPriority() const { | |
| 178 if (!release_fences_.empty()) | |
| 179 return std::min(priority_, GpuStreamPriority::HIGH); | |
| 180 return priority_; | |
| 181 } | |
| 182 | |
| 183 bool Scheduler::Sequence::RunsBefore(const Scheduler::Sequence* other) const { | |
| 184 return other->scheduling_state() < scheduling_state_; | |
| 185 } | |
| 186 | |
| 187 void Scheduler::Sequence::SetEnabled(bool enabled) { | |
| 188 if (enabled_ == enabled) | |
| 189 return; | |
| 190 DCHECK_EQ(running_state_, enabled ? IDLE : RUNNING); | |
| 191 enabled_ = enabled; | |
| 192 } | |
| 193 | |
| 194 void Scheduler::Sequence::SetScheduled() { | |
| 195 DCHECK_NE(running_state_, RUNNING); | |
| 196 running_state_ = SCHEDULED; | |
| 197 UpdateSchedulingState(); | |
| 198 } | |
| 199 | |
| 200 void Scheduler::Sequence::UpdateSchedulingState() { | |
| 201 scheduling_state_.sequence_id = sequence_id_; | |
| 202 scheduling_state_.priority = GetSchedulingPriority(); | |
| 203 | |
| 204 uint32_t order_num = UINT32_MAX; // IDLE | |
| 205 if (running_state_ == SCHEDULED) { | |
| 206 DCHECK(!tasks_.empty()); | |
| 207 order_num = tasks_.front().order_num; | |
| 208 } else if (running_state_ == RUNNING) { | |
| 209 order_num = order_data_->current_order_num(); | |
| 210 } | |
| 211 scheduling_state_.order_num = order_num; | |
| 212 } | |
| 213 | |
| 214 void Scheduler::Sequence::ContinueTask(base::OnceClosure closure) { | |
| 215 DCHECK_EQ(running_state_, RUNNING); | |
| 216 tasks_.push_front({std::move(closure), order_data_->current_order_num()}); | |
| 217 } | |
| 218 | |
| 219 uint32_t Scheduler::Sequence::ScheduleTask(base::OnceClosure closure) { | |
| 220 uint32_t order_num = order_data_->GenerateUnprocessedOrderNumber(); | |
| 221 tasks_.push_back({std::move(closure), order_num}); | |
| 222 return order_num; | |
| 223 } | |
| 224 | |
| 225 base::OnceClosure Scheduler::Sequence::BeginTask() { | |
| 226 DCHECK(!tasks_.empty()); | |
| 227 | |
| 228 DCHECK_EQ(running_state_, SCHEDULED); | |
| 229 running_state_ = RUNNING; | |
| 230 | |
| 231 base::OnceClosure closure = std::move(tasks_.front().closure); | |
| 232 uint32_t order_num = tasks_.front().order_num; | |
| 233 tasks_.pop_front(); | |
| 234 | |
| 235 order_data_->BeginProcessingOrderNumber(order_num); | |
| 236 | |
| 237 UpdateSchedulingState(); | |
| 238 | |
| 239 return closure; | |
| 240 } | |
| 241 | |
| 242 void Scheduler::Sequence::FinishTask() { | |
| 243 DCHECK_EQ(running_state_, RUNNING); | |
| 244 running_state_ = IDLE; | |
|
piman
2017/05/10 00:38:50
nit: could be premature optimization, but in Sched
sunnyps
2017/05/10 23:15:15
I kept it this way because I wanted to be explicit
| |
| 245 uint32_t order_num = order_data_->current_order_num(); | |
| 246 if (!tasks_.empty() && tasks_.front().order_num == order_num) { | |
| 247 order_data_->PauseProcessingOrderNumber(order_num); | |
| 248 } else { | |
| 249 order_data_->FinishProcessingOrderNumber(order_num); | |
| 250 } | |
| 251 UpdateSchedulingState(); | |
| 252 } | |
| 253 | |
| 254 void Scheduler::Sequence::AddWaitFence(const SyncToken& sync_token, | |
| 255 uint32_t order_num) { | |
| 256 wait_fences_.push_back({sync_token, order_num}); | |
| 257 } | |
| 258 | |
| 259 void Scheduler::Sequence::RemoveWaitFence(const SyncToken& sync_token, | |
| 260 uint32_t order_num) { | |
| 261 base::Erase(wait_fences_, Fence{sync_token, order_num}); | |
| 262 } | |
| 263 | |
| 264 void Scheduler::Sequence::AddReleaseFence(const SyncToken& sync_token, | |
| 265 uint32_t order_num) { | |
| 266 release_fences_.push_back({sync_token, order_num}); | |
| 267 } | |
| 268 | |
| 269 void Scheduler::Sequence::RemoveReleaseFence(const SyncToken& sync_token, | |
| 270 uint32_t order_num) { | |
| 271 base::Erase(release_fences_, Fence{sync_token, order_num}); | |
| 272 } | |
| 273 | |
| 274 Scheduler::Scheduler(scoped_refptr<base::SingleThreadTaskRunner> task_runner, | |
| 275 SyncPointManager* sync_point_manager) | |
| 276 : task_runner_(std::move(task_runner)), | |
| 277 sync_point_manager_(sync_point_manager), | |
| 278 weak_factory_(this) { | |
| 279 DCHECK(thread_checker_.CalledOnValidThread()); | |
| 280 } | |
| 281 | |
| 282 Scheduler::~Scheduler() { | |
| 283 DCHECK(thread_checker_.CalledOnValidThread()); | |
| 284 } | |
| 285 | |
| 286 SequenceId Scheduler::CreateSequence(GpuStreamPriority priority) { | |
| 287 DCHECK(thread_checker_.CalledOnValidThread()); | |
| 288 base::AutoLock auto_lock(lock_); | |
| 289 scoped_refptr<SyncPointOrderData> order_data = | |
| 290 sync_point_manager_->CreateSyncPointOrderData(); | |
| 291 SequenceId sequence_id = order_data->sequence_id(); | |
| 292 std::unique_ptr<Sequence> sequence = | |
|
piman
2017/05/10 00:38:50
nit: it's ok to use auto when using MakeUnique (th
sunnyps
2017/05/10 23:15:15
Done.
| |
| 293 base::MakeUnique<Sequence>(sequence_id, priority, std::move(order_data)); | |
| 294 sequences_.emplace(sequence_id, std::move(sequence)); | |
| 295 return sequence_id; | |
| 296 } | |
| 297 | |
| 298 void Scheduler::DestroySequence(SequenceId sequence_id) { | |
| 299 DCHECK(thread_checker_.CalledOnValidThread()); | |
| 300 base::AutoLock auto_lock(lock_); | |
| 301 | |
| 302 Sequence* sequence = GetSequence(sequence_id); | |
| 303 DCHECK(sequence); | |
| 304 sequence->Destroy(); | |
|
vmiura
2017/05/04 21:27:37
Is it necessary to defer the Destroy? Is it due t
sunnyps
2017/05/10 23:15:15
This isn't necessary as long as destroying the str
| |
| 305 | |
| 306 if (sequence->running()) | |
| 307 return; | |
| 308 | |
| 309 if (sequence->scheduled()) | |
| 310 rebuild_scheduling_queue_ = true; | |
| 311 sequences_.erase(sequence_id); | |
| 312 } | |
| 313 | |
| 314 Scheduler::Sequence* Scheduler::GetSequence(SequenceId sequence_id) { | |
| 315 lock_.AssertAcquired(); | |
| 316 auto it = sequences_.find(sequence_id); | |
| 317 if (it != sequences_.end()) | |
| 318 return it->second.get(); | |
| 319 return nullptr; | |
| 320 } | |
| 321 | |
| 322 void Scheduler::EnableSequence(SequenceId sequence_id) { | |
| 323 DCHECK(thread_checker_.CalledOnValidThread()); | |
| 324 base::AutoLock auto_lock(lock_); | |
| 325 Sequence* sequence = GetSequence(sequence_id); | |
| 326 DCHECK(sequence); | |
| 327 sequence->SetEnabled(true); | |
| 328 TryScheduleSequence(sequence); | |
| 329 } | |
| 330 | |
| 331 void Scheduler::DisableSequence(SequenceId sequence_id) { | |
| 332 DCHECK(thread_checker_.CalledOnValidThread()); | |
| 333 base::AutoLock auto_lock(lock_); | |
| 334 Sequence* sequence = GetSequence(sequence_id); | |
| 335 DCHECK(sequence); | |
| 336 sequence->SetEnabled(false); | |
| 337 } | |
| 338 | |
| 339 void Scheduler::ScheduleTask(SequenceId sequence_id, | |
| 340 base::OnceClosure closure, | |
| 341 const std::vector<SyncToken>& sync_token_fences) { | |
| 342 base::AutoLock auto_lock(lock_); | |
| 343 Sequence* sequence = GetSequence(sequence_id); | |
| 344 DCHECK(sequence); | |
| 345 | |
| 346 uint32_t order_num = sequence->ScheduleTask(std::move(closure)); | |
| 347 | |
| 348 for (const SyncToken& sync_token : sync_token_fences) { | |
| 349 SequenceId release_id = | |
| 350 sync_point_manager_->GetSyncTokenReleaseSequenceId(sync_token); | |
| 351 Sequence* release_sequence = GetSequence(release_id); | |
| 352 if (!release_sequence) | |
| 353 continue; | |
| 354 if (sync_point_manager_->Wait( | |
| 355 sync_token, order_num, | |
| 356 base::Bind(&Scheduler::SyncTokenFenceReleased, | |
| 357 weak_factory_.GetWeakPtr(), sync_token, order_num, | |
| 358 release_id, sequence_id))) { | |
| 359 sequence->AddWaitFence(sync_token, order_num); | |
| 360 release_sequence->AddReleaseFence(sync_token, order_num); | |
| 361 TryScheduleSequence(release_sequence); | |
| 362 } | |
| 363 } | |
| 364 | |
| 365 TryScheduleSequence(sequence); | |
| 366 } | |
| 367 | |
| 368 void Scheduler::ContinueTask(SequenceId sequence_id, | |
| 369 base::OnceClosure closure) { | |
| 370 DCHECK(thread_checker_.CalledOnValidThread()); | |
| 371 base::AutoLock auto_lock(lock_); | |
| 372 Sequence* sequence = GetSequence(sequence_id); | |
| 373 DCHECK(sequence); | |
| 374 sequence->ContinueTask(std::move(closure)); | |
| 375 } | |
| 376 | |
| 377 bool Scheduler::ShouldYield(SequenceId sequence_id) { | |
| 378 DCHECK(thread_checker_.CalledOnValidThread()); | |
| 379 base::AutoLock auto_lock(lock_); | |
| 380 | |
| 381 Sequence* sequence = GetSequence(sequence_id); | |
| 382 DCHECK(sequence); | |
| 383 DCHECK(sequence->running()); | |
| 384 | |
| 385 if (should_yield_) | |
| 386 return true; | |
| 387 | |
| 388 RebuildSchedulingQueue(); | |
| 389 | |
| 390 sequence->UpdateSchedulingState(); | |
| 391 | |
| 392 if (!scheduling_queue_.empty()) { | |
| 393 Sequence* next_sequence = GetSequence(scheduling_queue_.top().sequence_id); | |
| 394 DCHECK(next_sequence); | |
| 395 if (next_sequence->RunsBefore(sequence)) | |
| 396 should_yield_ = true; | |
| 397 } | |
| 398 | |
| 399 return should_yield_; | |
| 400 } | |
| 401 | |
| 402 void Scheduler::SyncTokenFenceReleased(const SyncToken& sync_token, | |
| 403 uint32_t order_num, | |
| 404 SequenceId release_sequence_id, | |
| 405 SequenceId waiting_sequence_id) { | |
| 406 base::AutoLock auto_lock(lock_); | |
| 407 Sequence* sequence = GetSequence(waiting_sequence_id); | |
| 408 if (sequence) { | |
| 409 sequence->RemoveWaitFence(sync_token, order_num); | |
| 410 TryScheduleSequence(sequence); | |
| 411 } | |
| 412 Sequence* release_sequence = GetSequence(release_sequence_id); | |
| 413 if (release_sequence) { | |
| 414 release_sequence->RemoveReleaseFence(sync_token, order_num); | |
| 415 TryScheduleSequence(release_sequence); | |
| 416 } | |
| 417 } | |
| 418 | |
| 419 void Scheduler::TryScheduleSequence(Sequence* sequence) { | |
| 420 lock_.AssertAcquired(); | |
| 421 | |
| 422 if (sequence->running()) | |
| 423 return; | |
| 424 | |
| 425 if (sequence->NeedsRescheduling()) { | |
| 426 DCHECK(sequence->IsRunnable()); | |
| 427 rebuild_scheduling_queue_ = true; | |
| 428 } else if (!sequence->scheduled() && sequence->IsRunnable()) { | |
| 429 sequence->SetScheduled(); | |
| 430 scheduling_queue_.push(sequence->scheduling_state()); | |
| 431 } | |
| 432 | |
| 433 if (!running_) { | |
| 434 TRACE_EVENT_ASYNC_BEGIN0("gpu", "Scheduler::Running", this); | |
| 435 running_ = true; | |
| 436 task_runner_->PostTask(FROM_HERE, base::Bind(&Scheduler::RunNextTask, | |
| 437 weak_factory_.GetWeakPtr())); | |
| 438 } | |
| 439 } | |
| 440 | |
| 441 void Scheduler::RebuildSchedulingQueue() { | |
| 442 DCHECK(thread_checker_.CalledOnValidThread()); | |
| 443 lock_.AssertAcquired(); | |
| 444 | |
| 445 if (!rebuild_scheduling_queue_) | |
| 446 return; | |
| 447 rebuild_scheduling_queue_ = false; | |
| 448 | |
| 449 std::vector<SchedulingState> states; | |
|
piman
2017/05/10 00:38:50
nit: states.reserve(sequences_.count()); to avoid
sunnyps
2017/05/10 23:15:15
Not needed any more (see below).
| |
| 450 for (const auto& kv : sequences_) { | |
| 451 Sequence* sequence = kv.second.get(); | |
| 452 if (!sequence->IsRunnable() || sequence->running()) | |
| 453 continue; | |
| 454 sequence->SetScheduled(); | |
| 455 states.push_back(sequence->scheduling_state()); | |
| 456 } | |
| 457 | |
| 458 scheduling_queue_ = SchedulingQueue(states.begin(), states.end()); | |
|
vmiura
2017/05/04 21:27:37
For efficiency, it would be better to use the std:
piman
2017/05/10 00:38:50
I'd even argue, if we access underlying container
sunnyps
2017/05/10 23:15:15
Changed to use make_heap, etc.
| |
| 459 } | |
| 460 | |
| 461 void Scheduler::RunNextTask() { | |
| 462 DCHECK(thread_checker_.CalledOnValidThread()); | |
| 463 base::AutoLock auto_lock(lock_); | |
| 464 | |
| 465 should_yield_ = false; | |
| 466 | |
| 467 RebuildSchedulingQueue(); | |
| 468 | |
| 469 if (scheduling_queue_.empty()) { | |
| 470 TRACE_EVENT_ASYNC_END0("gpu", "Scheduler::Running", this); | |
| 471 running_ = false; | |
| 472 return; | |
| 473 } | |
| 474 | |
| 475 SchedulingState state = scheduling_queue_.top(); | |
| 476 scheduling_queue_.pop(); | |
| 477 | |
| 478 TRACE_EVENT1("gpu", "Scheduler::RunNextTask", "state", state.AsValue()); | |
| 479 | |
| 480 Sequence* sequence = GetSequence(state.sequence_id); | |
| 481 DCHECK(sequence); | |
| 482 | |
| 483 base::OnceClosure closure = sequence->BeginTask(); | |
| 484 | |
| 485 { | |
| 486 base::AutoUnlock auto_unlock(lock_); | |
| 487 std::move(closure).Run(); | |
| 488 } | |
| 489 | |
| 490 sequence->FinishTask(); | |
| 491 | |
| 492 if (sequence->destroyed()) { | |
| 493 sequences_.erase(sequence->sequence_id()); | |
| 494 } else if (sequence->IsRunnable()) { | |
| 495 sequence->SetScheduled(); | |
| 496 scheduling_queue_.push(sequence->scheduling_state()); | |
| 497 } | |
| 498 | |
| 499 task_runner_->PostTask(FROM_HERE, base::Bind(&Scheduler::RunNextTask, | |
| 500 weak_factory_.GetWeakPtr())); | |
| 501 } | |
| 502 | |
| 503 } // namespace gpu | |
| OLD | NEW |