Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(47)

Side by Side Diff: gpu/command_buffer/service/scheduler.cc

Issue 2814843002: gpu: GPU service scheduler. (Closed)
Patch Set: rebase Created 3 years, 7 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View unified diff | Download patch
OLDNEW
(Empty)
1 // Copyright (c) 2017 The Chromium Authors. All rights reserved.
2 // Use of this source code is governed by a BSD-style license that can be
3 // found in the LICENSE file.
4
5 #include "gpu/command_buffer/service/scheduler.h"
6
7 #include <algorithm>
8
9 #include "base/callback.h"
10 #include "base/memory/ptr_util.h"
11 #include "base/stl_util.h"
12 #include "base/trace_event/trace_event.h"
13 #include "base/trace_event/trace_event_argument.h"
14 #include "gpu/command_buffer/service/sync_point_manager.h"
15
16 namespace gpu {
17
18 class Scheduler::Sequence {
19 public:
20 Sequence(SequenceId sequence_id,
21 GpuStreamPriority priority,
22 scoped_refptr<SyncPointOrderData> order_data);
23
24 ~Sequence();
25
26 void Destroy();
27
28 bool destroyed() const { return destroyed_; }
29
30 SequenceId sequence_id() const { return sequence_id_; }
31
32 const SchedulingState& scheduling_state() const { return scheduling_state_; }
33
34 bool enabled() const { return enabled_; }
35
36 bool scheduled() const { return running_state_ == SCHEDULED; }
37
38 bool running() const { return running_state_ == RUNNING; }
39
40 // The sequence is runnable if its enabled and has tasks which are not blocked
41 // by wait fences.
42 bool IsRunnable() const;
43
44 bool NeedsRescheduling() const;
45
46 void UpdateSchedulingState();
47
48 // If this sequence runs before the other sequence.
49 bool RunsBefore(const Sequence* other) const;
50
51 void SetEnabled(bool enabled);
52
53 // Sets running state to SCHEDULED.
54 void SetScheduled();
55
56 // Called before running the next task on the sequence. Returns the closure
57 // for the task. Sets running state to RUNNING.
58 base::OnceClosure BeginTask();
59
60 // Called after running the closure returned by BeginTask. Sets running state
61 // to IDLE.
62 void FinishTask();
63
64 // Enqueues a task in the sequence and returns the generated order number.
65 uint32_t ScheduleTask(base::OnceClosure closure);
66
67 // Continue running the current task with the given closure. Must be called in
68 // between |BeginTask| and |FinishTask|.
69 void ContinueTask(base::OnceClosure closure);
70
71 // Add a sync token fence that this sequence should wait on.
72 void AddWaitFence(const SyncToken& sync_token, uint32_t order_num);
73
74 // Remove a waiting sync token fence.
75 void RemoveWaitFence(const SyncToken& sync_token, uint32_t order_num);
76
77 // Add a sync token fence that this sequence is expected to release.
78 void AddReleaseFence(const SyncToken& sync_token, uint32_t order_num);
79
80 // Remove a release sync token fence.
81 void RemoveReleaseFence(const SyncToken& sync_token, uint32_t order_num);
82
83 private:
84 enum RunningState { IDLE, SCHEDULED, RUNNING };
85
86 struct Fence {
87 SyncToken sync_token;
88 uint32_t order_num;
89
90 bool operator==(const Fence& other) const {
91 return std::tie(sync_token, order_num) ==
92 std::tie(other.sync_token, other.order_num);
93 }
94 };
95
96 struct Task {
97 base::OnceClosure closure;
98 uint32_t order_num;
99 };
100
101 GpuStreamPriority GetSchedulingPriority() const;
102
103 bool destroyed_ = false;
104
105 // If the sequence is enabled. Sequences are disabled/enabled based on when
106 // the command buffer is descheduled/sc
piman 2017/05/10 00:38:50 nit: "/sc" typo or missing end of comment?
sunnyps 2017/05/10 23:15:15 Done.
107 bool enabled_ = true;
108
109 RunningState running_state_;
110
111 // Cached scheduling state used for comparison with other sequences using
112 // |RunsBefore|. Updated in |UpdateSchedulingState|.
113 SchedulingState scheduling_state_;
114
115 const SequenceId sequence_id_;
116
117 const GpuStreamPriority priority_;
118
119 scoped_refptr<SyncPointOrderData> order_data_;
120
121 // Deque of tasks. Tasks are inserted at the back with increasing order number
122 // generated from SyncPointOrderData. If a running task needs to be continued,
123 // it is inserted at the front with the same order number.
124 std::deque<Task> tasks_;
125
126 // List of fences that this sequence is waiting on. Fences are inserted in
127 // increasing order number but may be removed out of order. Tasks are blocked
128 // if there's a wait fence with order number less than or equal to the task's
129 // order number.
130 std::vector<Fence> wait_fences_;
131
132 // List of fences that this sequence is expected to release. If this list is
133 // non-empty, the priority of the sequence is raised.
134 std::vector<Fence> release_fences_;
135
136 DISALLOW_COPY_AND_ASSIGN(Sequence);
137 };
138
139 Scheduler::SchedulingState::SchedulingState() = default;
140 Scheduler::SchedulingState::SchedulingState(const SchedulingState& other) =
141 default;
142 Scheduler::SchedulingState::~SchedulingState() = default;
143
144 std::unique_ptr<base::trace_event::ConvertableToTraceFormat>
145 Scheduler::SchedulingState::AsValue() const {
146 std::unique_ptr<base::trace_event::TracedValue> state(
147 new base::trace_event::TracedValue());
148 state->SetInteger("sequence_id", sequence_id.GetUnsafeValue());
149 state->SetString("priority", GpuStreamPriorityToString(priority));
150 state->SetInteger("order_num", order_num);
151 return std::move(state);
152 }
153
154 Scheduler::Sequence::Sequence(SequenceId sequence_id,
155 GpuStreamPriority priority,
156 scoped_refptr<SyncPointOrderData> order_data)
157 : sequence_id_(sequence_id), priority_(priority), order_data_(order_data) {}
158
159 Scheduler::Sequence::~Sequence() = default;
160
161 void Scheduler::Sequence::Destroy() {
162 DCHECK(!destroyed_);
163 destroyed_ = true;
164 }
165
166 bool Scheduler::Sequence::NeedsRescheduling() const {
167 return running_state_ != IDLE &&
168 scheduling_state_.priority != GetSchedulingPriority();
169 }
170
171 bool Scheduler::Sequence::IsRunnable() const {
172 return enabled_ && !tasks_.empty() &&
173 (wait_fences_.empty() ||
174 wait_fences_.front().order_num > tasks_.front().order_num);
175 }
176
177 GpuStreamPriority Scheduler::Sequence::GetSchedulingPriority() const {
178 if (!release_fences_.empty())
179 return std::min(priority_, GpuStreamPriority::HIGH);
180 return priority_;
181 }
182
183 bool Scheduler::Sequence::RunsBefore(const Scheduler::Sequence* other) const {
184 return other->scheduling_state() < scheduling_state_;
185 }
186
187 void Scheduler::Sequence::SetEnabled(bool enabled) {
188 if (enabled_ == enabled)
189 return;
190 DCHECK_EQ(running_state_, enabled ? IDLE : RUNNING);
191 enabled_ = enabled;
192 }
193
194 void Scheduler::Sequence::SetScheduled() {
195 DCHECK_NE(running_state_, RUNNING);
196 running_state_ = SCHEDULED;
197 UpdateSchedulingState();
198 }
199
200 void Scheduler::Sequence::UpdateSchedulingState() {
201 scheduling_state_.sequence_id = sequence_id_;
202 scheduling_state_.priority = GetSchedulingPriority();
203
204 uint32_t order_num = UINT32_MAX; // IDLE
205 if (running_state_ == SCHEDULED) {
206 DCHECK(!tasks_.empty());
207 order_num = tasks_.front().order_num;
208 } else if (running_state_ == RUNNING) {
209 order_num = order_data_->current_order_num();
210 }
211 scheduling_state_.order_num = order_num;
212 }
213
214 void Scheduler::Sequence::ContinueTask(base::OnceClosure closure) {
215 DCHECK_EQ(running_state_, RUNNING);
216 tasks_.push_front({std::move(closure), order_data_->current_order_num()});
217 }
218
219 uint32_t Scheduler::Sequence::ScheduleTask(base::OnceClosure closure) {
220 uint32_t order_num = order_data_->GenerateUnprocessedOrderNumber();
221 tasks_.push_back({std::move(closure), order_num});
222 return order_num;
223 }
224
225 base::OnceClosure Scheduler::Sequence::BeginTask() {
226 DCHECK(!tasks_.empty());
227
228 DCHECK_EQ(running_state_, SCHEDULED);
229 running_state_ = RUNNING;
230
231 base::OnceClosure closure = std::move(tasks_.front().closure);
232 uint32_t order_num = tasks_.front().order_num;
233 tasks_.pop_front();
234
235 order_data_->BeginProcessingOrderNumber(order_num);
236
237 UpdateSchedulingState();
238
239 return closure;
240 }
241
242 void Scheduler::Sequence::FinishTask() {
243 DCHECK_EQ(running_state_, RUNNING);
244 running_state_ = IDLE;
piman 2017/05/10 00:38:50 nit: could be premature optimization, but in Sched
sunnyps 2017/05/10 23:15:15 I kept it this way because I wanted to be explicit
245 uint32_t order_num = order_data_->current_order_num();
246 if (!tasks_.empty() && tasks_.front().order_num == order_num) {
247 order_data_->PauseProcessingOrderNumber(order_num);
248 } else {
249 order_data_->FinishProcessingOrderNumber(order_num);
250 }
251 UpdateSchedulingState();
252 }
253
254 void Scheduler::Sequence::AddWaitFence(const SyncToken& sync_token,
255 uint32_t order_num) {
256 wait_fences_.push_back({sync_token, order_num});
257 }
258
259 void Scheduler::Sequence::RemoveWaitFence(const SyncToken& sync_token,
260 uint32_t order_num) {
261 base::Erase(wait_fences_, Fence{sync_token, order_num});
262 }
263
264 void Scheduler::Sequence::AddReleaseFence(const SyncToken& sync_token,
265 uint32_t order_num) {
266 release_fences_.push_back({sync_token, order_num});
267 }
268
269 void Scheduler::Sequence::RemoveReleaseFence(const SyncToken& sync_token,
270 uint32_t order_num) {
271 base::Erase(release_fences_, Fence{sync_token, order_num});
272 }
273
274 Scheduler::Scheduler(scoped_refptr<base::SingleThreadTaskRunner> task_runner,
275 SyncPointManager* sync_point_manager)
276 : task_runner_(std::move(task_runner)),
277 sync_point_manager_(sync_point_manager),
278 weak_factory_(this) {
279 DCHECK(thread_checker_.CalledOnValidThread());
280 }
281
282 Scheduler::~Scheduler() {
283 DCHECK(thread_checker_.CalledOnValidThread());
284 }
285
286 SequenceId Scheduler::CreateSequence(GpuStreamPriority priority) {
287 DCHECK(thread_checker_.CalledOnValidThread());
288 base::AutoLock auto_lock(lock_);
289 scoped_refptr<SyncPointOrderData> order_data =
290 sync_point_manager_->CreateSyncPointOrderData();
291 SequenceId sequence_id = order_data->sequence_id();
292 std::unique_ptr<Sequence> sequence =
piman 2017/05/10 00:38:50 nit: it's ok to use auto when using MakeUnique (th
sunnyps 2017/05/10 23:15:15 Done.
293 base::MakeUnique<Sequence>(sequence_id, priority, std::move(order_data));
294 sequences_.emplace(sequence_id, std::move(sequence));
295 return sequence_id;
296 }
297
298 void Scheduler::DestroySequence(SequenceId sequence_id) {
299 DCHECK(thread_checker_.CalledOnValidThread());
300 base::AutoLock auto_lock(lock_);
301
302 Sequence* sequence = GetSequence(sequence_id);
303 DCHECK(sequence);
304 sequence->Destroy();
vmiura 2017/05/04 21:27:37 Is it necessary to defer the Destroy? Is it due t
sunnyps 2017/05/10 23:15:15 This isn't necessary as long as destroying the str
305
306 if (sequence->running())
307 return;
308
309 if (sequence->scheduled())
310 rebuild_scheduling_queue_ = true;
311 sequences_.erase(sequence_id);
312 }
313
314 Scheduler::Sequence* Scheduler::GetSequence(SequenceId sequence_id) {
315 lock_.AssertAcquired();
316 auto it = sequences_.find(sequence_id);
317 if (it != sequences_.end())
318 return it->second.get();
319 return nullptr;
320 }
321
322 void Scheduler::EnableSequence(SequenceId sequence_id) {
323 DCHECK(thread_checker_.CalledOnValidThread());
324 base::AutoLock auto_lock(lock_);
325 Sequence* sequence = GetSequence(sequence_id);
326 DCHECK(sequence);
327 sequence->SetEnabled(true);
328 TryScheduleSequence(sequence);
329 }
330
331 void Scheduler::DisableSequence(SequenceId sequence_id) {
332 DCHECK(thread_checker_.CalledOnValidThread());
333 base::AutoLock auto_lock(lock_);
334 Sequence* sequence = GetSequence(sequence_id);
335 DCHECK(sequence);
336 sequence->SetEnabled(false);
337 }
338
339 void Scheduler::ScheduleTask(SequenceId sequence_id,
340 base::OnceClosure closure,
341 const std::vector<SyncToken>& sync_token_fences) {
342 base::AutoLock auto_lock(lock_);
343 Sequence* sequence = GetSequence(sequence_id);
344 DCHECK(sequence);
345
346 uint32_t order_num = sequence->ScheduleTask(std::move(closure));
347
348 for (const SyncToken& sync_token : sync_token_fences) {
349 SequenceId release_id =
350 sync_point_manager_->GetSyncTokenReleaseSequenceId(sync_token);
351 Sequence* release_sequence = GetSequence(release_id);
352 if (!release_sequence)
353 continue;
354 if (sync_point_manager_->Wait(
355 sync_token, order_num,
356 base::Bind(&Scheduler::SyncTokenFenceReleased,
357 weak_factory_.GetWeakPtr(), sync_token, order_num,
358 release_id, sequence_id))) {
359 sequence->AddWaitFence(sync_token, order_num);
360 release_sequence->AddReleaseFence(sync_token, order_num);
361 TryScheduleSequence(release_sequence);
362 }
363 }
364
365 TryScheduleSequence(sequence);
366 }
367
368 void Scheduler::ContinueTask(SequenceId sequence_id,
369 base::OnceClosure closure) {
370 DCHECK(thread_checker_.CalledOnValidThread());
371 base::AutoLock auto_lock(lock_);
372 Sequence* sequence = GetSequence(sequence_id);
373 DCHECK(sequence);
374 sequence->ContinueTask(std::move(closure));
375 }
376
377 bool Scheduler::ShouldYield(SequenceId sequence_id) {
378 DCHECK(thread_checker_.CalledOnValidThread());
379 base::AutoLock auto_lock(lock_);
380
381 Sequence* sequence = GetSequence(sequence_id);
382 DCHECK(sequence);
383 DCHECK(sequence->running());
384
385 if (should_yield_)
386 return true;
387
388 RebuildSchedulingQueue();
389
390 sequence->UpdateSchedulingState();
391
392 if (!scheduling_queue_.empty()) {
393 Sequence* next_sequence = GetSequence(scheduling_queue_.top().sequence_id);
394 DCHECK(next_sequence);
395 if (next_sequence->RunsBefore(sequence))
396 should_yield_ = true;
397 }
398
399 return should_yield_;
400 }
401
402 void Scheduler::SyncTokenFenceReleased(const SyncToken& sync_token,
403 uint32_t order_num,
404 SequenceId release_sequence_id,
405 SequenceId waiting_sequence_id) {
406 base::AutoLock auto_lock(lock_);
407 Sequence* sequence = GetSequence(waiting_sequence_id);
408 if (sequence) {
409 sequence->RemoveWaitFence(sync_token, order_num);
410 TryScheduleSequence(sequence);
411 }
412 Sequence* release_sequence = GetSequence(release_sequence_id);
413 if (release_sequence) {
414 release_sequence->RemoveReleaseFence(sync_token, order_num);
415 TryScheduleSequence(release_sequence);
416 }
417 }
418
419 void Scheduler::TryScheduleSequence(Sequence* sequence) {
420 lock_.AssertAcquired();
421
422 if (sequence->running())
423 return;
424
425 if (sequence->NeedsRescheduling()) {
426 DCHECK(sequence->IsRunnable());
427 rebuild_scheduling_queue_ = true;
428 } else if (!sequence->scheduled() && sequence->IsRunnable()) {
429 sequence->SetScheduled();
430 scheduling_queue_.push(sequence->scheduling_state());
431 }
432
433 if (!running_) {
434 TRACE_EVENT_ASYNC_BEGIN0("gpu", "Scheduler::Running", this);
435 running_ = true;
436 task_runner_->PostTask(FROM_HERE, base::Bind(&Scheduler::RunNextTask,
437 weak_factory_.GetWeakPtr()));
438 }
439 }
440
441 void Scheduler::RebuildSchedulingQueue() {
442 DCHECK(thread_checker_.CalledOnValidThread());
443 lock_.AssertAcquired();
444
445 if (!rebuild_scheduling_queue_)
446 return;
447 rebuild_scheduling_queue_ = false;
448
449 std::vector<SchedulingState> states;
piman 2017/05/10 00:38:50 nit: states.reserve(sequences_.count()); to avoid
sunnyps 2017/05/10 23:15:15 Not needed any more (see below).
450 for (const auto& kv : sequences_) {
451 Sequence* sequence = kv.second.get();
452 if (!sequence->IsRunnable() || sequence->running())
453 continue;
454 sequence->SetScheduled();
455 states.push_back(sequence->scheduling_state());
456 }
457
458 scheduling_queue_ = SchedulingQueue(states.begin(), states.end());
vmiura 2017/05/04 21:27:37 For efficiency, it would be better to use the std:
piman 2017/05/10 00:38:50 I'd even argue, if we access underlying container
sunnyps 2017/05/10 23:15:15 Changed to use make_heap, etc.
459 }
460
461 void Scheduler::RunNextTask() {
462 DCHECK(thread_checker_.CalledOnValidThread());
463 base::AutoLock auto_lock(lock_);
464
465 should_yield_ = false;
466
467 RebuildSchedulingQueue();
468
469 if (scheduling_queue_.empty()) {
470 TRACE_EVENT_ASYNC_END0("gpu", "Scheduler::Running", this);
471 running_ = false;
472 return;
473 }
474
475 SchedulingState state = scheduling_queue_.top();
476 scheduling_queue_.pop();
477
478 TRACE_EVENT1("gpu", "Scheduler::RunNextTask", "state", state.AsValue());
479
480 Sequence* sequence = GetSequence(state.sequence_id);
481 DCHECK(sequence);
482
483 base::OnceClosure closure = sequence->BeginTask();
484
485 {
486 base::AutoUnlock auto_unlock(lock_);
487 std::move(closure).Run();
488 }
489
490 sequence->FinishTask();
491
492 if (sequence->destroyed()) {
493 sequences_.erase(sequence->sequence_id());
494 } else if (sequence->IsRunnable()) {
495 sequence->SetScheduled();
496 scheduling_queue_.push(sequence->scheduling_state());
497 }
498
499 task_runner_->PostTask(FROM_HERE, base::Bind(&Scheduler::RunNextTask,
500 weak_factory_.GetWeakPtr()));
501 }
502
503 } // namespace gpu
OLDNEW

Powered by Google App Engine
This is Rietveld 408576698