OLD | NEW |
---|---|
(Empty) | |
1 // Copyright (c) 2017 The Chromium Authors. All rights reserved. | |
2 // Use of this source code is governed by a BSD-style license that can be | |
3 // found in the LICENSE file. | |
4 | |
5 #include "gpu/command_buffer/service/scheduler.h" | |
6 | |
7 #include <algorithm> | |
8 | |
9 #include "base/callback.h" | |
10 #include "base/memory/ptr_util.h" | |
11 #include "base/stl_util.h" | |
12 #include "base/trace_event/trace_event.h" | |
13 #include "base/trace_event/trace_event_argument.h" | |
14 #include "gpu/command_buffer/service/sync_point_manager.h" | |
15 | |
16 namespace gpu { | |
17 | |
18 class Scheduler::Sequence { | |
19 public: | |
20 Sequence(SequenceId sequence_id, | |
21 GpuStreamPriority priority, | |
22 scoped_refptr<SyncPointOrderData> order_data); | |
23 | |
24 ~Sequence(); | |
25 | |
26 void Destroy(); | |
27 | |
28 bool destroyed() const { return destroyed_; } | |
29 | |
30 SequenceId sequence_id() const { return sequence_id_; } | |
31 | |
32 const SchedulingState& scheduling_state() const { return scheduling_state_; } | |
33 | |
34 bool enabled() const { return enabled_; } | |
35 | |
36 bool scheduled() const { return running_state_ == SCHEDULED; } | |
37 | |
38 bool running() const { return running_state_ == RUNNING; } | |
39 | |
40 // The sequence is runnable if its enabled and has tasks which are not blocked | |
41 // by wait fences. | |
42 bool IsRunnable() const; | |
43 | |
44 bool NeedsRescheduling() const; | |
45 | |
46 void UpdateSchedulingState(); | |
47 | |
48 // If this sequence runs before the other sequence. | |
49 bool RunsBefore(const Sequence* other) const; | |
50 | |
51 void SetEnabled(bool enabled); | |
52 | |
53 // Sets running state to SCHEDULED. | |
54 void SetScheduled(); | |
55 | |
56 // Called before running the next task on the sequence. Returns the closure | |
57 // for the task. Sets running state to RUNNING. | |
58 base::OnceClosure BeginTask(); | |
59 | |
60 // Called after running the closure returned by BeginTask. Sets running state | |
61 // to IDLE. | |
62 void FinishTask(); | |
63 | |
64 // Enqueues a task in the sequence and returns the generated order number. | |
65 uint32_t ScheduleTask(base::OnceClosure closure); | |
66 | |
67 // Continue running the current task with the given closure. Must be called in | |
68 // between |BeginTask| and |FinishTask|. | |
69 void ContinueTask(base::OnceClosure closure); | |
70 | |
71 // Add a sync token fence that this sequence should wait on. | |
72 void AddWaitFence(const SyncToken& sync_token, uint32_t order_num); | |
73 | |
74 // Remove a waiting sync token fence. | |
75 void RemoveWaitFence(const SyncToken& sync_token, uint32_t order_num); | |
76 | |
77 // Add a sync token fence that this sequence is expected to release. | |
78 void AddReleaseFence(const SyncToken& sync_token, uint32_t order_num); | |
79 | |
80 // Remove a release sync token fence. | |
81 void RemoveReleaseFence(const SyncToken& sync_token, uint32_t order_num); | |
82 | |
83 private: | |
84 enum RunningState { IDLE, SCHEDULED, RUNNING }; | |
85 | |
86 struct Fence { | |
87 SyncToken sync_token; | |
88 uint32_t order_num; | |
89 | |
90 bool operator==(const Fence& other) const { | |
91 return std::tie(sync_token, order_num) == | |
92 std::tie(other.sync_token, other.order_num); | |
93 } | |
94 }; | |
95 | |
96 struct Task { | |
97 base::OnceClosure closure; | |
98 uint32_t order_num; | |
99 }; | |
100 | |
101 GpuStreamPriority GetSchedulingPriority() const; | |
102 | |
103 bool destroyed_ = false; | |
104 | |
105 // If the sequence is enabled. Sequences are disabled/enabled based on when | |
106 // the command buffer is descheduled/sc | |
piman
2017/05/10 00:38:50
nit: "/sc" typo or missing end of comment?
sunnyps
2017/05/10 23:15:15
Done.
| |
107 bool enabled_ = true; | |
108 | |
109 RunningState running_state_; | |
110 | |
111 // Cached scheduling state used for comparison with other sequences using | |
112 // |RunsBefore|. Updated in |UpdateSchedulingState|. | |
113 SchedulingState scheduling_state_; | |
114 | |
115 const SequenceId sequence_id_; | |
116 | |
117 const GpuStreamPriority priority_; | |
118 | |
119 scoped_refptr<SyncPointOrderData> order_data_; | |
120 | |
121 // Deque of tasks. Tasks are inserted at the back with increasing order number | |
122 // generated from SyncPointOrderData. If a running task needs to be continued, | |
123 // it is inserted at the front with the same order number. | |
124 std::deque<Task> tasks_; | |
125 | |
126 // List of fences that this sequence is waiting on. Fences are inserted in | |
127 // increasing order number but may be removed out of order. Tasks are blocked | |
128 // if there's a wait fence with order number less than or equal to the task's | |
129 // order number. | |
130 std::vector<Fence> wait_fences_; | |
131 | |
132 // List of fences that this sequence is expected to release. If this list is | |
133 // non-empty, the priority of the sequence is raised. | |
134 std::vector<Fence> release_fences_; | |
135 | |
136 DISALLOW_COPY_AND_ASSIGN(Sequence); | |
137 }; | |
138 | |
139 Scheduler::SchedulingState::SchedulingState() = default; | |
140 Scheduler::SchedulingState::SchedulingState(const SchedulingState& other) = | |
141 default; | |
142 Scheduler::SchedulingState::~SchedulingState() = default; | |
143 | |
144 std::unique_ptr<base::trace_event::ConvertableToTraceFormat> | |
145 Scheduler::SchedulingState::AsValue() const { | |
146 std::unique_ptr<base::trace_event::TracedValue> state( | |
147 new base::trace_event::TracedValue()); | |
148 state->SetInteger("sequence_id", sequence_id.GetUnsafeValue()); | |
149 state->SetString("priority", GpuStreamPriorityToString(priority)); | |
150 state->SetInteger("order_num", order_num); | |
151 return std::move(state); | |
152 } | |
153 | |
154 Scheduler::Sequence::Sequence(SequenceId sequence_id, | |
155 GpuStreamPriority priority, | |
156 scoped_refptr<SyncPointOrderData> order_data) | |
157 : sequence_id_(sequence_id), priority_(priority), order_data_(order_data) {} | |
158 | |
159 Scheduler::Sequence::~Sequence() = default; | |
160 | |
161 void Scheduler::Sequence::Destroy() { | |
162 DCHECK(!destroyed_); | |
163 destroyed_ = true; | |
164 } | |
165 | |
166 bool Scheduler::Sequence::NeedsRescheduling() const { | |
167 return running_state_ != IDLE && | |
168 scheduling_state_.priority != GetSchedulingPriority(); | |
169 } | |
170 | |
171 bool Scheduler::Sequence::IsRunnable() const { | |
172 return enabled_ && !tasks_.empty() && | |
173 (wait_fences_.empty() || | |
174 wait_fences_.front().order_num > tasks_.front().order_num); | |
175 } | |
176 | |
177 GpuStreamPriority Scheduler::Sequence::GetSchedulingPriority() const { | |
178 if (!release_fences_.empty()) | |
179 return std::min(priority_, GpuStreamPriority::HIGH); | |
180 return priority_; | |
181 } | |
182 | |
183 bool Scheduler::Sequence::RunsBefore(const Scheduler::Sequence* other) const { | |
184 return other->scheduling_state() < scheduling_state_; | |
185 } | |
186 | |
187 void Scheduler::Sequence::SetEnabled(bool enabled) { | |
188 if (enabled_ == enabled) | |
189 return; | |
190 DCHECK_EQ(running_state_, enabled ? IDLE : RUNNING); | |
191 enabled_ = enabled; | |
192 } | |
193 | |
194 void Scheduler::Sequence::SetScheduled() { | |
195 DCHECK_NE(running_state_, RUNNING); | |
196 running_state_ = SCHEDULED; | |
197 UpdateSchedulingState(); | |
198 } | |
199 | |
200 void Scheduler::Sequence::UpdateSchedulingState() { | |
201 scheduling_state_.sequence_id = sequence_id_; | |
202 scheduling_state_.priority = GetSchedulingPriority(); | |
203 | |
204 uint32_t order_num = UINT32_MAX; // IDLE | |
205 if (running_state_ == SCHEDULED) { | |
206 DCHECK(!tasks_.empty()); | |
207 order_num = tasks_.front().order_num; | |
208 } else if (running_state_ == RUNNING) { | |
209 order_num = order_data_->current_order_num(); | |
210 } | |
211 scheduling_state_.order_num = order_num; | |
212 } | |
213 | |
214 void Scheduler::Sequence::ContinueTask(base::OnceClosure closure) { | |
215 DCHECK_EQ(running_state_, RUNNING); | |
216 tasks_.push_front({std::move(closure), order_data_->current_order_num()}); | |
217 } | |
218 | |
219 uint32_t Scheduler::Sequence::ScheduleTask(base::OnceClosure closure) { | |
220 uint32_t order_num = order_data_->GenerateUnprocessedOrderNumber(); | |
221 tasks_.push_back({std::move(closure), order_num}); | |
222 return order_num; | |
223 } | |
224 | |
225 base::OnceClosure Scheduler::Sequence::BeginTask() { | |
226 DCHECK(!tasks_.empty()); | |
227 | |
228 DCHECK_EQ(running_state_, SCHEDULED); | |
229 running_state_ = RUNNING; | |
230 | |
231 base::OnceClosure closure = std::move(tasks_.front().closure); | |
232 uint32_t order_num = tasks_.front().order_num; | |
233 tasks_.pop_front(); | |
234 | |
235 order_data_->BeginProcessingOrderNumber(order_num); | |
236 | |
237 UpdateSchedulingState(); | |
238 | |
239 return closure; | |
240 } | |
241 | |
242 void Scheduler::Sequence::FinishTask() { | |
243 DCHECK_EQ(running_state_, RUNNING); | |
244 running_state_ = IDLE; | |
piman
2017/05/10 00:38:50
nit: could be premature optimization, but in Sched
sunnyps
2017/05/10 23:15:15
I kept it this way because I wanted to be explicit
| |
245 uint32_t order_num = order_data_->current_order_num(); | |
246 if (!tasks_.empty() && tasks_.front().order_num == order_num) { | |
247 order_data_->PauseProcessingOrderNumber(order_num); | |
248 } else { | |
249 order_data_->FinishProcessingOrderNumber(order_num); | |
250 } | |
251 UpdateSchedulingState(); | |
252 } | |
253 | |
254 void Scheduler::Sequence::AddWaitFence(const SyncToken& sync_token, | |
255 uint32_t order_num) { | |
256 wait_fences_.push_back({sync_token, order_num}); | |
257 } | |
258 | |
259 void Scheduler::Sequence::RemoveWaitFence(const SyncToken& sync_token, | |
260 uint32_t order_num) { | |
261 base::Erase(wait_fences_, Fence{sync_token, order_num}); | |
262 } | |
263 | |
264 void Scheduler::Sequence::AddReleaseFence(const SyncToken& sync_token, | |
265 uint32_t order_num) { | |
266 release_fences_.push_back({sync_token, order_num}); | |
267 } | |
268 | |
269 void Scheduler::Sequence::RemoveReleaseFence(const SyncToken& sync_token, | |
270 uint32_t order_num) { | |
271 base::Erase(release_fences_, Fence{sync_token, order_num}); | |
272 } | |
273 | |
274 Scheduler::Scheduler(scoped_refptr<base::SingleThreadTaskRunner> task_runner, | |
275 SyncPointManager* sync_point_manager) | |
276 : task_runner_(std::move(task_runner)), | |
277 sync_point_manager_(sync_point_manager), | |
278 weak_factory_(this) { | |
279 DCHECK(thread_checker_.CalledOnValidThread()); | |
280 } | |
281 | |
282 Scheduler::~Scheduler() { | |
283 DCHECK(thread_checker_.CalledOnValidThread()); | |
284 } | |
285 | |
286 SequenceId Scheduler::CreateSequence(GpuStreamPriority priority) { | |
287 DCHECK(thread_checker_.CalledOnValidThread()); | |
288 base::AutoLock auto_lock(lock_); | |
289 scoped_refptr<SyncPointOrderData> order_data = | |
290 sync_point_manager_->CreateSyncPointOrderData(); | |
291 SequenceId sequence_id = order_data->sequence_id(); | |
292 std::unique_ptr<Sequence> sequence = | |
piman
2017/05/10 00:38:50
nit: it's ok to use auto when using MakeUnique (th
sunnyps
2017/05/10 23:15:15
Done.
| |
293 base::MakeUnique<Sequence>(sequence_id, priority, std::move(order_data)); | |
294 sequences_.emplace(sequence_id, std::move(sequence)); | |
295 return sequence_id; | |
296 } | |
297 | |
298 void Scheduler::DestroySequence(SequenceId sequence_id) { | |
299 DCHECK(thread_checker_.CalledOnValidThread()); | |
300 base::AutoLock auto_lock(lock_); | |
301 | |
302 Sequence* sequence = GetSequence(sequence_id); | |
303 DCHECK(sequence); | |
304 sequence->Destroy(); | |
vmiura
2017/05/04 21:27:37
Is it necessary to defer the Destroy? Is it due t
sunnyps
2017/05/10 23:15:15
This isn't necessary as long as destroying the str
| |
305 | |
306 if (sequence->running()) | |
307 return; | |
308 | |
309 if (sequence->scheduled()) | |
310 rebuild_scheduling_queue_ = true; | |
311 sequences_.erase(sequence_id); | |
312 } | |
313 | |
314 Scheduler::Sequence* Scheduler::GetSequence(SequenceId sequence_id) { | |
315 lock_.AssertAcquired(); | |
316 auto it = sequences_.find(sequence_id); | |
317 if (it != sequences_.end()) | |
318 return it->second.get(); | |
319 return nullptr; | |
320 } | |
321 | |
322 void Scheduler::EnableSequence(SequenceId sequence_id) { | |
323 DCHECK(thread_checker_.CalledOnValidThread()); | |
324 base::AutoLock auto_lock(lock_); | |
325 Sequence* sequence = GetSequence(sequence_id); | |
326 DCHECK(sequence); | |
327 sequence->SetEnabled(true); | |
328 TryScheduleSequence(sequence); | |
329 } | |
330 | |
331 void Scheduler::DisableSequence(SequenceId sequence_id) { | |
332 DCHECK(thread_checker_.CalledOnValidThread()); | |
333 base::AutoLock auto_lock(lock_); | |
334 Sequence* sequence = GetSequence(sequence_id); | |
335 DCHECK(sequence); | |
336 sequence->SetEnabled(false); | |
337 } | |
338 | |
339 void Scheduler::ScheduleTask(SequenceId sequence_id, | |
340 base::OnceClosure closure, | |
341 const std::vector<SyncToken>& sync_token_fences) { | |
342 base::AutoLock auto_lock(lock_); | |
343 Sequence* sequence = GetSequence(sequence_id); | |
344 DCHECK(sequence); | |
345 | |
346 uint32_t order_num = sequence->ScheduleTask(std::move(closure)); | |
347 | |
348 for (const SyncToken& sync_token : sync_token_fences) { | |
349 SequenceId release_id = | |
350 sync_point_manager_->GetSyncTokenReleaseSequenceId(sync_token); | |
351 Sequence* release_sequence = GetSequence(release_id); | |
352 if (!release_sequence) | |
353 continue; | |
354 if (sync_point_manager_->Wait( | |
355 sync_token, order_num, | |
356 base::Bind(&Scheduler::SyncTokenFenceReleased, | |
357 weak_factory_.GetWeakPtr(), sync_token, order_num, | |
358 release_id, sequence_id))) { | |
359 sequence->AddWaitFence(sync_token, order_num); | |
360 release_sequence->AddReleaseFence(sync_token, order_num); | |
361 TryScheduleSequence(release_sequence); | |
362 } | |
363 } | |
364 | |
365 TryScheduleSequence(sequence); | |
366 } | |
367 | |
368 void Scheduler::ContinueTask(SequenceId sequence_id, | |
369 base::OnceClosure closure) { | |
370 DCHECK(thread_checker_.CalledOnValidThread()); | |
371 base::AutoLock auto_lock(lock_); | |
372 Sequence* sequence = GetSequence(sequence_id); | |
373 DCHECK(sequence); | |
374 sequence->ContinueTask(std::move(closure)); | |
375 } | |
376 | |
377 bool Scheduler::ShouldYield(SequenceId sequence_id) { | |
378 DCHECK(thread_checker_.CalledOnValidThread()); | |
379 base::AutoLock auto_lock(lock_); | |
380 | |
381 Sequence* sequence = GetSequence(sequence_id); | |
382 DCHECK(sequence); | |
383 DCHECK(sequence->running()); | |
384 | |
385 if (should_yield_) | |
386 return true; | |
387 | |
388 RebuildSchedulingQueue(); | |
389 | |
390 sequence->UpdateSchedulingState(); | |
391 | |
392 if (!scheduling_queue_.empty()) { | |
393 Sequence* next_sequence = GetSequence(scheduling_queue_.top().sequence_id); | |
394 DCHECK(next_sequence); | |
395 if (next_sequence->RunsBefore(sequence)) | |
396 should_yield_ = true; | |
397 } | |
398 | |
399 return should_yield_; | |
400 } | |
401 | |
402 void Scheduler::SyncTokenFenceReleased(const SyncToken& sync_token, | |
403 uint32_t order_num, | |
404 SequenceId release_sequence_id, | |
405 SequenceId waiting_sequence_id) { | |
406 base::AutoLock auto_lock(lock_); | |
407 Sequence* sequence = GetSequence(waiting_sequence_id); | |
408 if (sequence) { | |
409 sequence->RemoveWaitFence(sync_token, order_num); | |
410 TryScheduleSequence(sequence); | |
411 } | |
412 Sequence* release_sequence = GetSequence(release_sequence_id); | |
413 if (release_sequence) { | |
414 release_sequence->RemoveReleaseFence(sync_token, order_num); | |
415 TryScheduleSequence(release_sequence); | |
416 } | |
417 } | |
418 | |
419 void Scheduler::TryScheduleSequence(Sequence* sequence) { | |
420 lock_.AssertAcquired(); | |
421 | |
422 if (sequence->running()) | |
423 return; | |
424 | |
425 if (sequence->NeedsRescheduling()) { | |
426 DCHECK(sequence->IsRunnable()); | |
427 rebuild_scheduling_queue_ = true; | |
428 } else if (!sequence->scheduled() && sequence->IsRunnable()) { | |
429 sequence->SetScheduled(); | |
430 scheduling_queue_.push(sequence->scheduling_state()); | |
431 } | |
432 | |
433 if (!running_) { | |
434 TRACE_EVENT_ASYNC_BEGIN0("gpu", "Scheduler::Running", this); | |
435 running_ = true; | |
436 task_runner_->PostTask(FROM_HERE, base::Bind(&Scheduler::RunNextTask, | |
437 weak_factory_.GetWeakPtr())); | |
438 } | |
439 } | |
440 | |
441 void Scheduler::RebuildSchedulingQueue() { | |
442 DCHECK(thread_checker_.CalledOnValidThread()); | |
443 lock_.AssertAcquired(); | |
444 | |
445 if (!rebuild_scheduling_queue_) | |
446 return; | |
447 rebuild_scheduling_queue_ = false; | |
448 | |
449 std::vector<SchedulingState> states; | |
piman
2017/05/10 00:38:50
nit: states.reserve(sequences_.count()); to avoid
sunnyps
2017/05/10 23:15:15
Not needed any more (see below).
| |
450 for (const auto& kv : sequences_) { | |
451 Sequence* sequence = kv.second.get(); | |
452 if (!sequence->IsRunnable() || sequence->running()) | |
453 continue; | |
454 sequence->SetScheduled(); | |
455 states.push_back(sequence->scheduling_state()); | |
456 } | |
457 | |
458 scheduling_queue_ = SchedulingQueue(states.begin(), states.end()); | |
vmiura
2017/05/04 21:27:37
For efficiency, it would be better to use the std:
piman
2017/05/10 00:38:50
I'd even argue, if we access underlying container
sunnyps
2017/05/10 23:15:15
Changed to use make_heap, etc.
| |
459 } | |
460 | |
461 void Scheduler::RunNextTask() { | |
462 DCHECK(thread_checker_.CalledOnValidThread()); | |
463 base::AutoLock auto_lock(lock_); | |
464 | |
465 should_yield_ = false; | |
466 | |
467 RebuildSchedulingQueue(); | |
468 | |
469 if (scheduling_queue_.empty()) { | |
470 TRACE_EVENT_ASYNC_END0("gpu", "Scheduler::Running", this); | |
471 running_ = false; | |
472 return; | |
473 } | |
474 | |
475 SchedulingState state = scheduling_queue_.top(); | |
476 scheduling_queue_.pop(); | |
477 | |
478 TRACE_EVENT1("gpu", "Scheduler::RunNextTask", "state", state.AsValue()); | |
479 | |
480 Sequence* sequence = GetSequence(state.sequence_id); | |
481 DCHECK(sequence); | |
482 | |
483 base::OnceClosure closure = sequence->BeginTask(); | |
484 | |
485 { | |
486 base::AutoUnlock auto_unlock(lock_); | |
487 std::move(closure).Run(); | |
488 } | |
489 | |
490 sequence->FinishTask(); | |
491 | |
492 if (sequence->destroyed()) { | |
493 sequences_.erase(sequence->sequence_id()); | |
494 } else if (sequence->IsRunnable()) { | |
495 sequence->SetScheduled(); | |
496 scheduling_queue_.push(sequence->scheduling_state()); | |
497 } | |
498 | |
499 task_runner_->PostTask(FROM_HERE, base::Bind(&Scheduler::RunNextTask, | |
500 weak_factory_.GetWeakPtr())); | |
501 } | |
502 | |
503 } // namespace gpu | |
OLD | NEW |