OLD | NEW |
---|---|
1 // Copyright 2015 the V8 project authors. All rights reserved. | 1 // Copyright 2015 the V8 project authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 #ifndef V8_GREEDY_ALLOCATOR_H_ | 5 #ifndef V8_GREEDY_ALLOCATOR_H_ |
6 #define V8_GREEDY_ALLOCATOR_H_ | 6 #define V8_GREEDY_ALLOCATOR_H_ |
7 | 7 |
8 #include "src/compiler/coalesced-live-ranges.h" | |
8 #include "src/compiler/register-allocator.h" | 9 #include "src/compiler/register-allocator.h" |
9 #include "src/zone-containers.h" | 10 #include "src/zone-containers.h" |
10 | 11 |
11 namespace v8 { | 12 namespace v8 { |
12 namespace internal { | 13 namespace internal { |
13 namespace compiler { | 14 namespace compiler { |
14 | 15 |
15 class CoalescedLiveRanges; | 16 |
17 // The object of allocation scheduling. At minimum, this is a LiveRange, but | |
18 // we may extend this to groups of LiveRanges. It has to be comparable. | |
19 // TODO(mtrofin): size calculation may happen redundantly for a range, consider | |
20 // making size a member of LiveRange. | |
21 struct AllocationCandidate { | |
22 explicit AllocationCandidate(LiveRange* range); | |
23 bool operator<(const AllocationCandidate& other) const { | |
24 return size_ < other.size_; | |
25 } | |
Benedikt Meurer
2015/06/25 04:41:06
Nit: Please add an appropriate operator> if you de
Mircea Trofin
2015/06/25 23:06:22
Done.
| |
26 LiveRange* live_range() const { return range_; } | |
27 unsigned size() const { return size_; } | |
28 | |
29 private: | |
30 LiveRange* range_; | |
31 unsigned size_; | |
32 }; | |
33 | |
34 | |
35 // Schedule processing (allocating) of AllocationCandidates. | |
36 class AllocationScheduler final : ZoneObject { | |
37 public: | |
38 explicit AllocationScheduler(Zone* zone) : storage_(zone) {} | |
39 void Schedule(LiveRange* range) { storage_.push(AllocationCandidate(range)); } | |
40 AllocationCandidate GetNext(); | |
41 bool empty() const { return storage_.empty(); } | |
42 | |
43 private: | |
44 typedef ZonePriorityQueue<AllocationCandidate> ScheduleStorage; | |
45 ScheduleStorage storage_; | |
46 | |
47 DISALLOW_COPY_AND_ASSIGN(AllocationScheduler); | |
48 }; | |
16 | 49 |
17 | 50 |
18 // A variant of the LLVM Greedy Register Allocator. See | 51 // A variant of the LLVM Greedy Register Allocator. See |
19 // http://blog.llvm.org/2011/09/greedy-register-allocation-in-llvm-30.html | 52 // http://blog.llvm.org/2011/09/greedy-register-allocation-in-llvm-30.html |
20 class GreedyAllocator final : public RegisterAllocator { | 53 class GreedyAllocator final : public RegisterAllocator { |
21 public: | 54 public: |
22 explicit GreedyAllocator(RegisterAllocationData* data, RegisterKind kind, | 55 explicit GreedyAllocator(RegisterAllocationData* data, RegisterKind kind, |
23 Zone* local_zone); | 56 Zone* local_zone); |
24 | 57 |
25 void AllocateRegisters(); | 58 void AllocateRegisters(); |
26 | 59 |
27 private: | 60 private: |
28 LifetimePosition GetSplittablePos(LifetimePosition pos); | 61 AllocationScheduler& scheduler() { return scheduler_; } |
29 const RegisterConfiguration* config() const { return data()->config(); } | 62 CoalescedLiveRanges* current_allocations(unsigned i) { |
63 return allocations_[i]; | |
64 } | |
30 Zone* local_zone() const { return local_zone_; } | 65 Zone* local_zone() const { return local_zone_; } |
31 | 66 |
32 int GetHintedRegister(LiveRange* range); | 67 // Insert fixed ranges. |
68 void InitializeAllocations(); | |
titzer
2015/06/24 23:05:44
Why not just name the method after what it does? I
Mircea Trofin
2015/06/25 23:06:23
Good suggestion. I'm proposing PreallocateFixedRan
| |
33 | 69 |
34 typedef ZonePriorityQueue<std::pair<unsigned, LiveRange*>> PQueue; | 70 // Schedule unassigned live ranges for allocation. |
71 // TODO(mtrofin): groups. | |
72 void InitializeScheduler(); | |
titzer
2015/06/24 23:05:44
Same here.
Mircea Trofin
2015/06/25 23:06:23
Done.
| |
35 | 73 |
36 unsigned GetLiveRangeSize(LiveRange* range); | 74 // Find the optimal split for ranges defined by a memory operand, e.g. |
37 void Enqueue(LiveRange* range); | 75 // constants or function parameters passed on the stack. |
76 void HandleRangesDefinedByMemoryOperand(); | |
titzer
2015/06/24 23:05:44
"handle" is too general. Please name the method af
| |
38 | 77 |
39 void Evict(LiveRange* range); | 78 void ProcessCandidate(const AllocationCandidate& candidate); |
titzer
2015/06/24 23:05:44
What does Process mean?
Mircea Trofin
2015/06/25 23:06:23
Done.
| |
40 float CalculateSpillWeight(LiveRange* range); | 79 void ProcessLiveRange(LiveRange* range, unsigned size); |
41 float CalculateMaxSpillWeight(const ZoneSet<LiveRange*>& ranges); | |
42 | 80 |
81 bool CanProcessRange(LiveRange* range) const { | |
82 return range != nullptr && !range->IsEmpty() && range->kind() == mode(); | |
83 } | |
43 | 84 |
44 bool TryAllocate(LiveRange* current, ZoneSet<LiveRange*>* conflicting); | 85 // Calculate the weight of a candidate for allocation. |
45 bool TryAllocatePhysicalRegister(unsigned reg_id, LiveRange* range, | 86 float GetCandidateRangeWeight(LiveRange* range, unsigned size); |
titzer
2015/06/24 23:05:44
Even though this is camel case (and thus CouldPerf
Mircea Trofin
2015/06/25 23:06:23
Done by virtue of having had changed a bit more on
| |
46 ZoneSet<LiveRange*>* conflicting); | 87 |
88 // Calculate the new weight of a range that is about to be allocated. | |
89 float GetAllocatedRangeWeight(float candidate_weight); | |
90 | |
91 // TODO(mtrofin): HandleSpillOperands feel like belonging to | |
92 // a preceding phase, rather than here. | |
47 bool HandleSpillOperands(LiveRange* range); | 93 bool HandleSpillOperands(LiveRange* range); |
titzer
2015/06/24 23:05:44
Again, no idea what "Handle" means here.
Mircea Trofin
2015/06/25 23:06:23
Done.
| |
48 void AllocateBlockedRange(LiveRange* current, LifetimePosition pos, | |
49 bool spill); | |
50 | 94 |
51 LiveRange* SpillBetweenUntil(LiveRange* range, LifetimePosition start, | 95 // This is the extension point for splitting heuristics. |
52 LifetimePosition until, LifetimePosition end); | 96 void HandleBlockedRange(LiveRange* range); |
53 void AssignRangeToRegister(int reg_id, LiveRange* range); | |
54 | 97 |
55 LifetimePosition FindProgressingSplitPosition(LiveRange* range, | 98 // Necessary heuristic: spill when all else failed. |
56 bool* is_spill_pos); | 99 void SpillRangeAsLastResort(LiveRange* range); |
100 | |
101 void AssignRangeToRegister(int reg_id, LiveRange* range, float weight); | |
57 | 102 |
58 Zone* local_zone_; | 103 Zone* local_zone_; |
59 ZoneVector<CoalescedLiveRanges*> allocations_; | 104 ZoneVector<CoalescedLiveRanges*> allocations_; |
60 PQueue queue_; | 105 AllocationScheduler scheduler_; |
61 DISALLOW_COPY_AND_ASSIGN(GreedyAllocator); | 106 DISALLOW_COPY_AND_ASSIGN(GreedyAllocator); |
62 }; | 107 }; |
63 } // namespace compiler | 108 } // namespace compiler |
64 } // namespace internal | 109 } // namespace internal |
65 } // namespace v8 | 110 } // namespace v8 |
66 #endif // V8_GREEDY_ALLOCATOR_H_ | 111 #endif // V8_GREEDY_ALLOCATOR_H_ |
OLD | NEW |