OLD | NEW |
1 //===- subzero/src/IceTargetLowering.h - Lowering interface -----*- C++ -*-===// | 1 //===- subzero/src/IceTargetLowering.h - Lowering interface -----*- C++ -*-===// |
2 // | 2 // |
3 // The Subzero Code Generator | 3 // The Subzero Code Generator |
4 // | 4 // |
5 // This file is distributed under the University of Illinois Open Source | 5 // This file is distributed under the University of Illinois Open Source |
6 // License. See LICENSE.TXT for details. | 6 // License. See LICENSE.TXT for details. |
7 // | 7 // |
8 //===----------------------------------------------------------------------===// | 8 //===----------------------------------------------------------------------===// |
9 /// | 9 /// |
10 /// \file | 10 /// \file |
11 /// This file declares the TargetLowering, LoweringContext, and | 11 /// This file declares the TargetLowering, LoweringContext, and |
12 /// TargetDataLowering classes. TargetLowering is an abstract class | 12 /// TargetDataLowering classes. TargetLowering is an abstract class used to |
13 /// used to drive the translation/lowering process. LoweringContext | 13 /// drive the translation/lowering process. LoweringContext maintains a context |
14 /// maintains a context for lowering each instruction, offering | 14 /// for lowering each instruction, offering conveniences such as iterating over |
15 /// conveniences such as iterating over non-deleted instructions. | 15 /// non-deleted instructions. TargetDataLowering is an abstract class used to |
16 /// TargetDataLowering is an abstract class used to drive the | 16 /// drive the lowering/emission of global initializers, external global |
17 /// lowering/emission of global initializers, external global | |
18 /// declarations, and internal constant pools. | 17 /// declarations, and internal constant pools. |
19 /// | 18 /// |
20 //===----------------------------------------------------------------------===// | 19 //===----------------------------------------------------------------------===// |
21 | 20 |
22 #ifndef SUBZERO_SRC_ICETARGETLOWERING_H | 21 #ifndef SUBZERO_SRC_ICETARGETLOWERING_H |
23 #define SUBZERO_SRC_ICETARGETLOWERING_H | 22 #define SUBZERO_SRC_ICETARGETLOWERING_H |
24 | 23 |
25 #include "IceDefs.h" | 24 #include "IceDefs.h" |
26 #include "IceInst.h" // for the names of the Inst subtypes | 25 #include "IceInst.h" // for the names of the Inst subtypes |
27 #include "IceOperand.h" | 26 #include "IceOperand.h" |
28 #include "IceTypes.h" | 27 #include "IceTypes.h" |
29 | 28 |
30 namespace Ice { | 29 namespace Ice { |
31 | 30 |
32 /// LoweringContext makes it easy to iterate through non-deleted | 31 /// LoweringContext makes it easy to iterate through non-deleted instructions in |
33 /// instructions in a node, and insert new (lowered) instructions at | 32 /// a node, and insert new (lowered) instructions at the current point. Along |
34 /// the current point. Along with the instruction list container and | 33 /// with the instruction list container and associated iterators, it holds the |
35 /// associated iterators, it holds the current node, which is needed | 34 /// current node, which is needed when inserting new instructions in order to |
36 /// when inserting new instructions in order to track whether variables | 35 /// track whether variables are used as single-block or multi-block. |
37 /// are used as single-block or multi-block. | |
38 class LoweringContext { | 36 class LoweringContext { |
39 LoweringContext(const LoweringContext &) = delete; | 37 LoweringContext(const LoweringContext &) = delete; |
40 LoweringContext &operator=(const LoweringContext &) = delete; | 38 LoweringContext &operator=(const LoweringContext &) = delete; |
41 | 39 |
42 public: | 40 public: |
43 LoweringContext() = default; | 41 LoweringContext() = default; |
44 ~LoweringContext() = default; | 42 ~LoweringContext() = default; |
45 void init(CfgNode *Node); | 43 void init(CfgNode *Node); |
46 Inst *getNextInst() const { | 44 Inst *getNextInst() const { |
47 if (Next == End) | 45 if (Next == End) |
(...skipping 17 matching lines...) Expand all Loading... |
65 void advanceNext() { advanceForward(Next); } | 63 void advanceNext() { advanceForward(Next); } |
66 void setCur(InstList::iterator C) { Cur = C; } | 64 void setCur(InstList::iterator C) { Cur = C; } |
67 void setNext(InstList::iterator N) { Next = N; } | 65 void setNext(InstList::iterator N) { Next = N; } |
68 void rewind(); | 66 void rewind(); |
69 void setInsertPoint(const InstList::iterator &Position) { Next = Position; } | 67 void setInsertPoint(const InstList::iterator &Position) { Next = Position; } |
70 | 68 |
71 private: | 69 private: |
72 /// Node is the argument to Inst::updateVars(). | 70 /// Node is the argument to Inst::updateVars(). |
73 CfgNode *Node = nullptr; | 71 CfgNode *Node = nullptr; |
74 Inst *LastInserted = nullptr; | 72 Inst *LastInserted = nullptr; |
75 /// Cur points to the current instruction being considered. It is | 73 /// Cur points to the current instruction being considered. It is guaranteed |
76 /// guaranteed to point to a non-deleted instruction, or to be End. | 74 /// to point to a non-deleted instruction, or to be End. |
77 InstList::iterator Cur; | 75 InstList::iterator Cur; |
78 /// Next doubles as a pointer to the next valid instruction (if any), | 76 /// Next doubles as a pointer to the next valid instruction (if any), and the |
79 /// and the new-instruction insertion point. It is also updated for | 77 /// new-instruction insertion point. It is also updated for the caller in case |
80 /// the caller in case the lowering consumes more than one high-level | 78 /// the lowering consumes more than one high-level instruction. It is |
81 /// instruction. It is guaranteed to point to a non-deleted | 79 /// guaranteed to point to a non-deleted instruction after Cur, or to be End. |
82 /// instruction after Cur, or to be End. TODO: Consider separating | 80 // TODO: Consider separating the notion of "next valid instruction" and "new |
83 /// the notion of "next valid instruction" and "new instruction | 81 // instruction insertion point", to avoid confusion when previously-deleted |
84 /// insertion point", to avoid confusion when previously-deleted | 82 // instructions come between the two points. |
85 /// instructions come between the two points. | |
86 InstList::iterator Next; | 83 InstList::iterator Next; |
87 /// Begin is a copy of Insts.begin(), used if iterators are moved backward. | 84 /// Begin is a copy of Insts.begin(), used if iterators are moved backward. |
88 InstList::iterator Begin; | 85 InstList::iterator Begin; |
89 /// End is a copy of Insts.end(), used if Next needs to be advanced. | 86 /// End is a copy of Insts.end(), used if Next needs to be advanced. |
90 InstList::iterator End; | 87 InstList::iterator End; |
91 | 88 |
92 void skipDeleted(InstList::iterator &I) const; | 89 void skipDeleted(InstList::iterator &I) const; |
93 void advanceForward(InstList::iterator &I) const; | 90 void advanceForward(InstList::iterator &I) const; |
94 }; | 91 }; |
95 | 92 |
(...skipping 56 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
152 | 149 |
153 /// Tries to do address mode optimization on a single instruction. | 150 /// Tries to do address mode optimization on a single instruction. |
154 void doAddressOpt(); | 151 void doAddressOpt(); |
155 /// Randomly insert NOPs. | 152 /// Randomly insert NOPs. |
156 void doNopInsertion(RandomNumberGenerator &RNG); | 153 void doNopInsertion(RandomNumberGenerator &RNG); |
157 /// Lowers a single non-Phi instruction. | 154 /// Lowers a single non-Phi instruction. |
158 void lower(); | 155 void lower(); |
159 /// Inserts and lowers a single high-level instruction at a specific insertion | 156 /// Inserts and lowers a single high-level instruction at a specific insertion |
160 /// point. | 157 /// point. |
161 void lowerInst(CfgNode *Node, InstList::iterator Next, InstHighLevel *Instr); | 158 void lowerInst(CfgNode *Node, InstList::iterator Next, InstHighLevel *Instr); |
162 /// Does preliminary lowering of the set of Phi instructions in the | 159 /// Does preliminary lowering of the set of Phi instructions in the current |
163 /// current node. The main intention is to do what's needed to keep | 160 /// node. The main intention is to do what's needed to keep the unlowered Phi |
164 /// the unlowered Phi instructions consistent with the lowered | 161 /// instructions consistent with the lowered non-Phi instructions, e.g. to |
165 /// non-Phi instructions, e.g. to lower 64-bit operands on a 32-bit | 162 /// lower 64-bit operands on a 32-bit target. |
166 /// target. | |
167 virtual void prelowerPhis() {} | 163 virtual void prelowerPhis() {} |
168 /// Tries to do branch optimization on a single instruction. Returns | 164 /// Tries to do branch optimization on a single instruction. Returns true if |
169 /// true if some optimization was done. | 165 /// some optimization was done. |
170 virtual bool doBranchOpt(Inst * /*I*/, const CfgNode * /*NextNode*/) { | 166 virtual bool doBranchOpt(Inst * /*I*/, const CfgNode * /*NextNode*/) { |
171 return false; | 167 return false; |
172 } | 168 } |
173 | 169 |
174 virtual SizeT getNumRegisters() const = 0; | 170 virtual SizeT getNumRegisters() const = 0; |
175 /// Returns a variable pre-colored to the specified physical | 171 /// Returns a variable pre-colored to the specified physical register. This is |
176 /// register. This is generally used to get very direct access to | 172 /// generally used to get very direct access to the register such as in the |
177 /// the register such as in the prolog or epilog or for marking | 173 /// prolog or epilog or for marking scratch registers as killed by a call. If |
178 /// scratch registers as killed by a call. If a Type is not | 174 /// a Type is not provided, a target-specific default type is used. |
179 /// provided, a target-specific default type is used. | |
180 virtual Variable *getPhysicalRegister(SizeT RegNum, | 175 virtual Variable *getPhysicalRegister(SizeT RegNum, |
181 Type Ty = IceType_void) = 0; | 176 Type Ty = IceType_void) = 0; |
182 /// Returns a printable name for the register. | 177 /// Returns a printable name for the register. |
183 virtual IceString getRegName(SizeT RegNum, Type Ty) const = 0; | 178 virtual IceString getRegName(SizeT RegNum, Type Ty) const = 0; |
184 | 179 |
185 virtual bool hasFramePointer() const { return false; } | 180 virtual bool hasFramePointer() const { return false; } |
186 virtual SizeT getFrameOrStackReg() const = 0; | 181 virtual SizeT getFrameOrStackReg() const = 0; |
187 virtual size_t typeWidthInBytesOnStack(Type Ty) const = 0; | 182 virtual size_t typeWidthInBytesOnStack(Type Ty) const = 0; |
188 | 183 |
189 bool hasComputedFrame() const { return HasComputedFrame; } | 184 bool hasComputedFrame() const { return HasComputedFrame; } |
190 /// Returns true if this function calls a function that has the | 185 /// Returns true if this function calls a function that has the "returns |
191 /// "returns twice" attribute. | 186 /// twice" attribute. |
192 bool callsReturnsTwice() const { return CallsReturnsTwice; } | 187 bool callsReturnsTwice() const { return CallsReturnsTwice; } |
193 void setCallsReturnsTwice(bool RetTwice) { CallsReturnsTwice = RetTwice; } | 188 void setCallsReturnsTwice(bool RetTwice) { CallsReturnsTwice = RetTwice; } |
194 int32_t getStackAdjustment() const { return StackAdjustment; } | 189 int32_t getStackAdjustment() const { return StackAdjustment; } |
195 void updateStackAdjustment(int32_t Offset) { StackAdjustment += Offset; } | 190 void updateStackAdjustment(int32_t Offset) { StackAdjustment += Offset; } |
196 void resetStackAdjustment() { StackAdjustment = 0; } | 191 void resetStackAdjustment() { StackAdjustment = 0; } |
197 SizeT makeNextLabelNumber() { return NextLabelNumber++; } | 192 SizeT makeNextLabelNumber() { return NextLabelNumber++; } |
198 SizeT makeNextJumpTableNumber() { return NextJumpTableNumber++; } | 193 SizeT makeNextJumpTableNumber() { return NextJumpTableNumber++; } |
199 LoweringContext &getContext() { return Context; } | 194 LoweringContext &getContext() { return Context; } |
200 | 195 |
201 enum RegSet { | 196 enum RegSet { |
(...skipping 11 matching lines...) Expand all Loading... |
213 virtual const llvm::SmallBitVector &getRegisterSetForType(Type Ty) const = 0; | 208 virtual const llvm::SmallBitVector &getRegisterSetForType(Type Ty) const = 0; |
214 virtual const llvm::SmallBitVector &getAliasesForRegister(SizeT) const = 0; | 209 virtual const llvm::SmallBitVector &getAliasesForRegister(SizeT) const = 0; |
215 | 210 |
216 void regAlloc(RegAllocKind Kind); | 211 void regAlloc(RegAllocKind Kind); |
217 | 212 |
218 virtual void | 213 virtual void |
219 makeRandomRegisterPermutation(llvm::SmallVectorImpl<int32_t> &Permutation, | 214 makeRandomRegisterPermutation(llvm::SmallVectorImpl<int32_t> &Permutation, |
220 const llvm::SmallBitVector &ExcludeRegisters, | 215 const llvm::SmallBitVector &ExcludeRegisters, |
221 uint64_t Salt) const = 0; | 216 uint64_t Salt) const = 0; |
222 | 217 |
223 /// Save/restore any mutable state for the situation where code | 218 /// Save/restore any mutable state for the situation where code emission needs |
224 /// emission needs multiple passes, such as sandboxing or relaxation. | 219 /// multiple passes, such as sandboxing or relaxation. Subclasses may provide |
225 /// Subclasses may provide their own implementation, but should be | 220 /// their own implementation, but should be sure to also call the parent |
226 /// sure to also call the parent class's methods. | 221 /// class's methods. |
227 virtual void snapshotEmitState() { | 222 virtual void snapshotEmitState() { |
228 SnapshotStackAdjustment = StackAdjustment; | 223 SnapshotStackAdjustment = StackAdjustment; |
229 } | 224 } |
230 virtual void rollbackEmitState() { | 225 virtual void rollbackEmitState() { |
231 StackAdjustment = SnapshotStackAdjustment; | 226 StackAdjustment = SnapshotStackAdjustment; |
232 } | 227 } |
233 | 228 |
234 /// Get the minimum number of clusters required for a jump table to be | 229 /// Get the minimum number of clusters required for a jump table to be |
235 /// considered. | 230 /// considered. |
236 virtual SizeT getMinJumpTableSize() const = 0; | 231 virtual SizeT getMinJumpTableSize() const = 0; |
(...skipping 40 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
277 virtual void lowerSelect(const InstSelect *Inst) = 0; | 272 virtual void lowerSelect(const InstSelect *Inst) = 0; |
278 virtual void lowerStore(const InstStore *Inst) = 0; | 273 virtual void lowerStore(const InstStore *Inst) = 0; |
279 virtual void lowerSwitch(const InstSwitch *Inst) = 0; | 274 virtual void lowerSwitch(const InstSwitch *Inst) = 0; |
280 virtual void lowerUnreachable(const InstUnreachable *Inst) = 0; | 275 virtual void lowerUnreachable(const InstUnreachable *Inst) = 0; |
281 virtual void lowerOther(const Inst *Instr); | 276 virtual void lowerOther(const Inst *Instr); |
282 | 277 |
283 virtual void doAddressOptLoad() {} | 278 virtual void doAddressOptLoad() {} |
284 virtual void doAddressOptStore() {} | 279 virtual void doAddressOptStore() {} |
285 virtual void randomlyInsertNop(float Probability, | 280 virtual void randomlyInsertNop(float Probability, |
286 RandomNumberGenerator &RNG) = 0; | 281 RandomNumberGenerator &RNG) = 0; |
287 /// This gives the target an opportunity to post-process the lowered | 282 /// This gives the target an opportunity to post-process the lowered expansion |
288 /// expansion before returning. | 283 /// before returning. |
289 virtual void postLower() {} | 284 virtual void postLower() {} |
290 | 285 |
291 /// Find two-address non-SSA instructions and set the DestNonKillable flag | 286 /// Find two-address non-SSA instructions and set the DestNonKillable flag to |
292 /// to keep liveness analysis consistent. | 287 /// keep liveness analysis consistent. |
293 void inferTwoAddress(); | 288 void inferTwoAddress(); |
294 | 289 |
295 /// Make a pass over the Cfg to determine which variables need stack slots | 290 /// Make a pass over the Cfg to determine which variables need stack slots and |
296 /// and place them in a sorted list (SortedSpilledVariables). Among those, | 291 /// place them in a sorted list (SortedSpilledVariables). Among those, vars, |
297 /// vars, classify the spill variables as local to the basic block vs | 292 /// classify the spill variables as local to the basic block vs global |
298 /// global (multi-block) in order to compute the parameters GlobalsSize | 293 /// (multi-block) in order to compute the parameters GlobalsSize and |
299 /// and SpillAreaSizeBytes (represents locals or general vars if the | 294 /// SpillAreaSizeBytes (represents locals or general vars if the coalescing of |
300 /// coalescing of locals is disallowed) along with alignments required | 295 /// locals is disallowed) along with alignments required for variables in each |
301 /// for variables in each area. We rely on accurate VMetadata in order to | 296 /// area. We rely on accurate VMetadata in order to classify a variable as |
302 /// classify a variable as global vs local (otherwise the variable is | 297 /// global vs local (otherwise the variable is conservatively global). The |
303 /// conservatively global). The in-args should be initialized to 0. | 298 /// in-args should be initialized to 0. |
304 /// | 299 /// |
305 /// This is only a pre-pass and the actual stack slot assignment is | 300 /// This is only a pre-pass and the actual stack slot assignment is handled |
306 /// handled separately. | 301 /// separately. |
307 /// | 302 /// |
308 /// There may be target-specific Variable types, which will be handled | 303 /// There may be target-specific Variable types, which will be handled by |
309 /// by TargetVarHook. If the TargetVarHook returns true, then the variable | 304 /// TargetVarHook. If the TargetVarHook returns true, then the variable is |
310 /// is skipped and not considered with the rest of the spilled variables. | 305 /// skipped and not considered with the rest of the spilled variables. |
311 void getVarStackSlotParams(VarList &SortedSpilledVariables, | 306 void getVarStackSlotParams(VarList &SortedSpilledVariables, |
312 llvm::SmallBitVector &RegsUsed, | 307 llvm::SmallBitVector &RegsUsed, |
313 size_t *GlobalsSize, size_t *SpillAreaSizeBytes, | 308 size_t *GlobalsSize, size_t *SpillAreaSizeBytes, |
314 uint32_t *SpillAreaAlignmentBytes, | 309 uint32_t *SpillAreaAlignmentBytes, |
315 uint32_t *LocalsSlotsAlignmentBytes, | 310 uint32_t *LocalsSlotsAlignmentBytes, |
316 std::function<bool(Variable *)> TargetVarHook); | 311 std::function<bool(Variable *)> TargetVarHook); |
317 | 312 |
318 /// Calculate the amount of padding needed to align the local and global | 313 /// Calculate the amount of padding needed to align the local and global areas |
319 /// areas to the required alignment. This assumes the globals/locals layout | 314 /// to the required alignment. This assumes the globals/locals layout used by |
320 /// used by getVarStackSlotParams and assignVarStackSlots. | 315 /// getVarStackSlotParams and assignVarStackSlots. |
321 void alignStackSpillAreas(uint32_t SpillAreaStartOffset, | 316 void alignStackSpillAreas(uint32_t SpillAreaStartOffset, |
322 uint32_t SpillAreaAlignmentBytes, | 317 uint32_t SpillAreaAlignmentBytes, |
323 size_t GlobalsSize, | 318 size_t GlobalsSize, |
324 uint32_t LocalsSlotsAlignmentBytes, | 319 uint32_t LocalsSlotsAlignmentBytes, |
325 uint32_t *SpillAreaPaddingBytes, | 320 uint32_t *SpillAreaPaddingBytes, |
326 uint32_t *LocalsSlotsPaddingBytes); | 321 uint32_t *LocalsSlotsPaddingBytes); |
327 | 322 |
328 /// Make a pass through the SortedSpilledVariables and actually assign | 323 /// Make a pass through the SortedSpilledVariables and actually assign stack |
329 /// stack slots. SpillAreaPaddingBytes takes into account stack alignment | 324 /// slots. SpillAreaPaddingBytes takes into account stack alignment padding. |
330 /// padding. The SpillArea starts after that amount of padding. | 325 /// The SpillArea starts after that amount of padding. This matches the scheme |
331 /// This matches the scheme in getVarStackSlotParams, where there may | 326 /// in getVarStackSlotParams, where there may be a separate multi-block global |
332 /// be a separate multi-block global var spill area and a local var | 327 /// var spill area and a local var spill area. |
333 /// spill area. | |
334 void assignVarStackSlots(VarList &SortedSpilledVariables, | 328 void assignVarStackSlots(VarList &SortedSpilledVariables, |
335 size_t SpillAreaPaddingBytes, | 329 size_t SpillAreaPaddingBytes, |
336 size_t SpillAreaSizeBytes, | 330 size_t SpillAreaSizeBytes, |
337 size_t GlobalsAndSubsequentPaddingSize, | 331 size_t GlobalsAndSubsequentPaddingSize, |
338 bool UsesFramePointer); | 332 bool UsesFramePointer); |
339 | 333 |
340 /// Sort the variables in Source based on required alignment. | 334 /// Sort the variables in Source based on required alignment. The variables |
341 /// The variables with the largest alignment need are placed in the front | 335 /// with the largest alignment need are placed in the front of the Dest list. |
342 /// of the Dest list. | |
343 void sortVarsByAlignment(VarList &Dest, const VarList &Source) const; | 336 void sortVarsByAlignment(VarList &Dest, const VarList &Source) const; |
344 | 337 |
345 /// Make a call to an external helper function. | 338 /// Make a call to an external helper function. |
346 InstCall *makeHelperCall(const IceString &Name, Variable *Dest, | 339 InstCall *makeHelperCall(const IceString &Name, Variable *Dest, |
347 SizeT MaxSrcs); | 340 SizeT MaxSrcs); |
348 | 341 |
349 void | 342 void |
350 _bundle_lock(InstBundleLock::Option BundleOption = InstBundleLock::Opt_None) { | 343 _bundle_lock(InstBundleLock::Option BundleOption = InstBundleLock::Opt_None) { |
351 Context.insert(InstBundleLock::create(Func, BundleOption)); | 344 Context.insert(InstBundleLock::create(Func, BundleOption)); |
352 } | 345 } |
353 void _bundle_unlock() { Context.insert(InstBundleUnlock::create(Func)); } | 346 void _bundle_unlock() { Context.insert(InstBundleUnlock::create(Func)); } |
354 void _set_dest_nonkillable() { | 347 void _set_dest_nonkillable() { |
355 Context.getLastInserted()->setDestNonKillable(); | 348 Context.getLastInserted()->setDestNonKillable(); |
356 } | 349 } |
357 | 350 |
358 bool shouldOptimizeMemIntrins(); | 351 bool shouldOptimizeMemIntrins(); |
359 | 352 |
360 Cfg *Func; | 353 Cfg *Func; |
361 GlobalContext *Ctx; | 354 GlobalContext *Ctx; |
362 bool HasComputedFrame = false; | 355 bool HasComputedFrame = false; |
363 bool CallsReturnsTwice = false; | 356 bool CallsReturnsTwice = false; |
364 /// StackAdjustment keeps track of the current stack offset from its | 357 /// StackAdjustment keeps track of the current stack offset from its natural |
365 /// natural location, as arguments are pushed for a function call. | 358 /// location, as arguments are pushed for a function call. |
366 int32_t StackAdjustment = 0; | 359 int32_t StackAdjustment = 0; |
367 SizeT NextLabelNumber = 0; | 360 SizeT NextLabelNumber = 0; |
368 SizeT NextJumpTableNumber = 0; | 361 SizeT NextJumpTableNumber = 0; |
369 LoweringContext Context; | 362 LoweringContext Context; |
370 | 363 |
371 // Runtime helper function names | 364 // Runtime helper function names |
372 const static constexpr char *H_bitcast_16xi1_i16 = "__Sz_bitcast_16xi1_i16"; | 365 const static constexpr char *H_bitcast_16xi1_i16 = "__Sz_bitcast_16xi1_i16"; |
373 const static constexpr char *H_bitcast_8xi1_i8 = "__Sz_bitcast_8xi1_i8"; | 366 const static constexpr char *H_bitcast_8xi1_i8 = "__Sz_bitcast_8xi1_i8"; |
374 const static constexpr char *H_bitcast_i16_16xi1 = "__Sz_bitcast_i16_16xi1"; | 367 const static constexpr char *H_bitcast_i16_16xi1 = "__Sz_bitcast_i16_16xi1"; |
375 const static constexpr char *H_bitcast_i8_8xi1 = "__Sz_bitcast_i8_8xi1"; | 368 const static constexpr char *H_bitcast_i8_8xi1 = "__Sz_bitcast_i8_8xi1"; |
(...skipping 27 matching lines...) Expand all Loading... |
403 const static constexpr char *H_uitofp_i32_f64 = "__Sz_uitofp_i32_f64"; | 396 const static constexpr char *H_uitofp_i32_f64 = "__Sz_uitofp_i32_f64"; |
404 const static constexpr char *H_uitofp_i64_f32 = "__Sz_uitofp_i64_f32"; | 397 const static constexpr char *H_uitofp_i64_f32 = "__Sz_uitofp_i64_f32"; |
405 const static constexpr char *H_uitofp_i64_f64 = "__Sz_uitofp_i64_f64"; | 398 const static constexpr char *H_uitofp_i64_f64 = "__Sz_uitofp_i64_f64"; |
406 const static constexpr char *H_urem_i32 = "__umodsi3"; | 399 const static constexpr char *H_urem_i32 = "__umodsi3"; |
407 const static constexpr char *H_urem_i64 = "__umoddi3"; | 400 const static constexpr char *H_urem_i64 = "__umoddi3"; |
408 | 401 |
409 private: | 402 private: |
410 int32_t SnapshotStackAdjustment = 0; | 403 int32_t SnapshotStackAdjustment = 0; |
411 }; | 404 }; |
412 | 405 |
413 /// TargetDataLowering is used for "lowering" data including initializers | 406 /// TargetDataLowering is used for "lowering" data including initializers for |
414 /// for global variables, and the internal constant pools. It is separated | 407 /// global variables, and the internal constant pools. It is separated out from |
415 /// out from TargetLowering because it does not require a Cfg. | 408 /// TargetLowering because it does not require a Cfg. |
416 class TargetDataLowering { | 409 class TargetDataLowering { |
417 TargetDataLowering() = delete; | 410 TargetDataLowering() = delete; |
418 TargetDataLowering(const TargetDataLowering &) = delete; | 411 TargetDataLowering(const TargetDataLowering &) = delete; |
419 TargetDataLowering &operator=(const TargetDataLowering &) = delete; | 412 TargetDataLowering &operator=(const TargetDataLowering &) = delete; |
420 | 413 |
421 public: | 414 public: |
422 static std::unique_ptr<TargetDataLowering> createLowering(GlobalContext *Ctx); | 415 static std::unique_ptr<TargetDataLowering> createLowering(GlobalContext *Ctx); |
423 virtual ~TargetDataLowering(); | 416 virtual ~TargetDataLowering(); |
424 | 417 |
425 virtual void lowerGlobals(const VariableDeclarationList &Vars, | 418 virtual void lowerGlobals(const VariableDeclarationList &Vars, |
426 const IceString &SectionSuffix) = 0; | 419 const IceString &SectionSuffix) = 0; |
427 virtual void lowerConstants() = 0; | 420 virtual void lowerConstants() = 0; |
428 virtual void lowerJumpTables() = 0; | 421 virtual void lowerJumpTables() = 0; |
429 | 422 |
430 protected: | 423 protected: |
431 void emitGlobal(const VariableDeclaration &Var, | 424 void emitGlobal(const VariableDeclaration &Var, |
432 const IceString &SectionSuffix); | 425 const IceString &SectionSuffix); |
433 | 426 |
434 /// For now, we assume .long is the right directive for emitting 4 byte | 427 /// For now, we assume .long is the right directive for emitting 4 byte emit |
435 /// emit global relocations. However, LLVM MIPS usually uses .4byte instead. | 428 /// global relocations. However, LLVM MIPS usually uses .4byte instead. |
436 /// Perhaps there is some difference when the location is unaligned. | 429 /// Perhaps there is some difference when the location is unaligned. |
437 static const char *getEmit32Directive() { return ".long"; } | 430 static const char *getEmit32Directive() { return ".long"; } |
438 | 431 |
439 explicit TargetDataLowering(GlobalContext *Ctx) : Ctx(Ctx) {} | 432 explicit TargetDataLowering(GlobalContext *Ctx) : Ctx(Ctx) {} |
440 GlobalContext *Ctx; | 433 GlobalContext *Ctx; |
441 }; | 434 }; |
442 | 435 |
443 /// TargetHeaderLowering is used to "lower" the header of an output file. | 436 /// TargetHeaderLowering is used to "lower" the header of an output file. It |
444 /// It writes out the target-specific header attributes. E.g., for ARM | 437 /// writes out the target-specific header attributes. E.g., for ARM this writes |
445 /// this writes out the build attributes (float ABI, etc.). | 438 /// out the build attributes (float ABI, etc.). |
446 class TargetHeaderLowering { | 439 class TargetHeaderLowering { |
447 TargetHeaderLowering() = delete; | 440 TargetHeaderLowering() = delete; |
448 TargetHeaderLowering(const TargetHeaderLowering &) = delete; | 441 TargetHeaderLowering(const TargetHeaderLowering &) = delete; |
449 TargetHeaderLowering &operator=(const TargetHeaderLowering &) = delete; | 442 TargetHeaderLowering &operator=(const TargetHeaderLowering &) = delete; |
450 | 443 |
451 public: | 444 public: |
452 static std::unique_ptr<TargetHeaderLowering> | 445 static std::unique_ptr<TargetHeaderLowering> |
453 createLowering(GlobalContext *Ctx); | 446 createLowering(GlobalContext *Ctx); |
454 virtual ~TargetHeaderLowering(); | 447 virtual ~TargetHeaderLowering(); |
455 | 448 |
456 virtual void lower() {} | 449 virtual void lower() {} |
457 | 450 |
458 protected: | 451 protected: |
459 explicit TargetHeaderLowering(GlobalContext *Ctx) : Ctx(Ctx) {} | 452 explicit TargetHeaderLowering(GlobalContext *Ctx) : Ctx(Ctx) {} |
460 GlobalContext *Ctx; | 453 GlobalContext *Ctx; |
461 }; | 454 }; |
462 | 455 |
463 } // end of namespace Ice | 456 } // end of namespace Ice |
464 | 457 |
465 #endif // SUBZERO_SRC_ICETARGETLOWERING_H | 458 #endif // SUBZERO_SRC_ICETARGETLOWERING_H |
OLD | NEW |