OLD | NEW |
---|---|
1 // Copyright 2011 the V8 project authors. All rights reserved. | 1 // Copyright 2011 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 60 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
71 MacroAssembler* masm() const { return masm_; } | 71 MacroAssembler* masm() const { return masm_; } |
72 CompilationInfo* info() const { return info_; } | 72 CompilationInfo* info() const { return info_; } |
73 Isolate* isolate() const { return info_->isolate(); } | 73 Isolate* isolate() const { return info_->isolate(); } |
74 Factory* factory() const { return isolate()->factory(); } | 74 Factory* factory() const { return isolate()->factory(); } |
75 Heap* heap() const { return isolate()->heap(); } | 75 Heap* heap() const { return isolate()->heap(); } |
76 | 76 |
77 // Support for converting LOperands to assembler types. | 77 // Support for converting LOperands to assembler types. |
78 Operand ToOperand(LOperand* op) const; | 78 Operand ToOperand(LOperand* op) const; |
79 Register ToRegister(LOperand* op) const; | 79 Register ToRegister(LOperand* op) const; |
80 XMMRegister ToDoubleRegister(LOperand* op) const; | 80 XMMRegister ToDoubleRegister(LOperand* op) const; |
81 Immediate ToImmediate(LOperand* op); | 81 |
82 bool IsInteger32(LConstantOperand* op) const; | |
83 Immediate ToInteger32Immediate(LOperand* op) const { | |
84 return Immediate(ToInteger32(LConstantOperand::cast(op))); | |
danno
2011/12/23 09:09:23
::cast is redundant?
Vyacheslav Egorov (Chromium)
2011/12/23 10:36:37
I don't think so,
ToInteger takes LConstantOpera
| |
85 } | |
86 | |
87 Handle<Object> ToHandle(LConstantOperand* op) const; | |
82 | 88 |
83 // The operand denoting the second word (the one with a higher address) of | 89 // The operand denoting the second word (the one with a higher address) of |
84 // a double stack slot. | 90 // a double stack slot. |
85 Operand HighOperand(LOperand* op); | 91 Operand HighOperand(LOperand* op); |
86 | 92 |
87 // Try to generate code for the entire chunk, but it may fail if the | 93 // Try to generate code for the entire chunk, but it may fail if the |
88 // chunk contains constructs we cannot handle. Returns true if the | 94 // chunk contains constructs we cannot handle. Returns true if the |
89 // code generation attempt succeeded. | 95 // code generation attempt succeeded. |
90 bool GenerateCode(); | 96 bool GenerateCode(); |
91 | 97 |
(...skipping 126 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
218 LOperand* op, | 224 LOperand* op, |
219 bool is_tagged); | 225 bool is_tagged); |
220 void PopulateDeoptimizationData(Handle<Code> code); | 226 void PopulateDeoptimizationData(Handle<Code> code); |
221 int DefineDeoptimizationLiteral(Handle<Object> literal); | 227 int DefineDeoptimizationLiteral(Handle<Object> literal); |
222 | 228 |
223 void PopulateDeoptimizationLiteralsWithInlinedFunctions(); | 229 void PopulateDeoptimizationLiteralsWithInlinedFunctions(); |
224 | 230 |
225 Register ToRegister(int index) const; | 231 Register ToRegister(int index) const; |
226 XMMRegister ToDoubleRegister(int index) const; | 232 XMMRegister ToDoubleRegister(int index) const; |
227 int ToInteger32(LConstantOperand* op) const; | 233 int ToInteger32(LConstantOperand* op) const; |
228 Handle<Object> ToHandle(LConstantOperand* op) const; | 234 |
229 double ToDouble(LConstantOperand* op) const; | 235 double ToDouble(LConstantOperand* op) const; |
230 Operand BuildFastArrayOperand(LOperand* elements_pointer, | 236 Operand BuildFastArrayOperand(LOperand* elements_pointer, |
231 LOperand* key, | 237 LOperand* key, |
232 ElementsKind elements_kind, | 238 ElementsKind elements_kind, |
233 uint32_t offset); | 239 uint32_t offset); |
234 | 240 |
235 // Specific math operations - used from DoUnaryMathOperation. | 241 // Specific math operations - used from DoUnaryMathOperation. |
236 void EmitIntegerMathAbs(LUnaryMathOperation* instr); | 242 void EmitIntegerMathAbs(LUnaryMathOperation* instr); |
237 void DoMathAbs(LUnaryMathOperation* instr); | 243 void DoMathAbs(LUnaryMathOperation* instr); |
238 void DoMathFloor(LUnaryMathOperation* instr); | 244 void DoMathFloor(LUnaryMathOperation* instr); |
(...skipping 148 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
387 LCodeGen* codegen_; | 393 LCodeGen* codegen_; |
388 Label entry_; | 394 Label entry_; |
389 Label exit_; | 395 Label exit_; |
390 Label* external_exit_; | 396 Label* external_exit_; |
391 int instruction_index_; | 397 int instruction_index_; |
392 }; | 398 }; |
393 | 399 |
394 } } // namespace v8::internal | 400 } } // namespace v8::internal |
395 | 401 |
396 #endif // V8_IA32_LITHIUM_CODEGEN_IA32_H_ | 402 #endif // V8_IA32_LITHIUM_CODEGEN_IA32_H_ |
OLD | NEW |