OLD | NEW |
1 // Copyright 2011 the V8 project authors. All rights reserved. | 1 // Copyright 2011 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 12 matching lines...) Expand all Loading... |
23 // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY | 23 // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
24 // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT | 24 // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
25 // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE | 25 // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
26 // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | 26 // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
27 | 27 |
28 #ifndef V8_X64_CODEGEN_X64_H_ | 28 #ifndef V8_X64_CODEGEN_X64_H_ |
29 #define V8_X64_CODEGEN_X64_H_ | 29 #define V8_X64_CODEGEN_X64_H_ |
30 | 30 |
31 #include "ast.h" | 31 #include "ast.h" |
32 #include "ic-inl.h" | 32 #include "ic-inl.h" |
33 #include "jump-target-heavy.h" | |
34 | 33 |
35 namespace v8 { | 34 namespace v8 { |
36 namespace internal { | 35 namespace internal { |
37 | 36 |
38 // Forward declarations | 37 // Forward declarations |
39 class CompilationInfo; | 38 class CompilationInfo; |
40 class DeferredCode; | |
41 class RegisterAllocator; | |
42 class RegisterFile; | |
43 | 39 |
44 enum InitState { CONST_INIT, NOT_CONST_INIT }; | |
45 enum TypeofState { INSIDE_TYPEOF, NOT_INSIDE_TYPEOF }; | 40 enum TypeofState { INSIDE_TYPEOF, NOT_INSIDE_TYPEOF }; |
46 | 41 |
47 | 42 |
48 // ------------------------------------------------------------------------- | 43 // ------------------------------------------------------------------------- |
49 // Reference support | |
50 | |
51 // A reference is a C++ stack-allocated object that puts a | |
52 // reference on the virtual frame. The reference may be consumed | |
53 // by GetValue, TakeValue, SetValue, and Codegen::UnloadReference. | |
54 // When the lifetime (scope) of a valid reference ends, it must have | |
55 // been consumed, and be in state UNLOADED. | |
56 class Reference BASE_EMBEDDED { | |
57 public: | |
58 // The values of the types is important, see size(). | |
59 enum Type { UNLOADED = -2, ILLEGAL = -1, SLOT = 0, NAMED = 1, KEYED = 2 }; | |
60 | |
61 Reference(CodeGenerator* cgen, | |
62 Expression* expression, | |
63 bool persist_after_get = false); | |
64 ~Reference(); | |
65 | |
66 Expression* expression() const { return expression_; } | |
67 Type type() const { return type_; } | |
68 void set_type(Type value) { | |
69 ASSERT_EQ(ILLEGAL, type_); | |
70 type_ = value; | |
71 } | |
72 | |
73 void set_unloaded() { | |
74 ASSERT_NE(ILLEGAL, type_); | |
75 ASSERT_NE(UNLOADED, type_); | |
76 type_ = UNLOADED; | |
77 } | |
78 // The size the reference takes up on the stack. | |
79 int size() const { | |
80 return (type_ < SLOT) ? 0 : type_; | |
81 } | |
82 | |
83 bool is_illegal() const { return type_ == ILLEGAL; } | |
84 bool is_slot() const { return type_ == SLOT; } | |
85 bool is_property() const { return type_ == NAMED || type_ == KEYED; } | |
86 bool is_unloaded() const { return type_ == UNLOADED; } | |
87 | |
88 // Return the name. Only valid for named property references. | |
89 Handle<String> GetName(); | |
90 | |
91 // Generate code to push the value of the reference on top of the | |
92 // expression stack. The reference is expected to be already on top of | |
93 // the expression stack, and it is consumed by the call unless the | |
94 // reference is for a compound assignment. | |
95 // If the reference is not consumed, it is left in place under its value. | |
96 void GetValue(); | |
97 | |
98 // Like GetValue except that the slot is expected to be written to before | |
99 // being read from again. The value of the reference may be invalidated, | |
100 // causing subsequent attempts to read it to fail. | |
101 void TakeValue(); | |
102 | |
103 // Generate code to store the value on top of the expression stack in the | |
104 // reference. The reference is expected to be immediately below the value | |
105 // on the expression stack. The value is stored in the location specified | |
106 // by the reference, and is left on top of the stack, after the reference | |
107 // is popped from beneath it (unloaded). | |
108 void SetValue(InitState init_state); | |
109 | |
110 private: | |
111 CodeGenerator* cgen_; | |
112 Expression* expression_; | |
113 Type type_; | |
114 bool persist_after_get_; | |
115 }; | |
116 | |
117 | |
118 // ------------------------------------------------------------------------- | |
119 // Control destinations. | |
120 | |
121 // A control destination encapsulates a pair of jump targets and a | |
122 // flag indicating which one is the preferred fall-through. The | |
123 // preferred fall-through must be unbound, the other may be already | |
124 // bound (ie, a backward target). | |
125 // | |
126 // The true and false targets may be jumped to unconditionally or | |
127 // control may split conditionally. Unconditional jumping and | |
128 // splitting should be emitted in tail position (as the last thing | |
129 // when compiling an expression) because they can cause either label | |
130 // to be bound or the non-fall through to be jumped to leaving an | |
131 // invalid virtual frame. | |
132 // | |
133 // The labels in the control destination can be extracted and | |
134 // manipulated normally without affecting the state of the | |
135 // destination. | |
136 | |
137 class ControlDestination BASE_EMBEDDED { | |
138 public: | |
139 ControlDestination(JumpTarget* true_target, | |
140 JumpTarget* false_target, | |
141 bool true_is_fall_through) | |
142 : true_target_(true_target), | |
143 false_target_(false_target), | |
144 true_is_fall_through_(true_is_fall_through), | |
145 is_used_(false) { | |
146 ASSERT(true_is_fall_through ? !true_target->is_bound() | |
147 : !false_target->is_bound()); | |
148 } | |
149 | |
150 // Accessors for the jump targets. Directly jumping or branching to | |
151 // or binding the targets will not update the destination's state. | |
152 JumpTarget* true_target() const { return true_target_; } | |
153 JumpTarget* false_target() const { return false_target_; } | |
154 | |
155 // True if the the destination has been jumped to unconditionally or | |
156 // control has been split to both targets. This predicate does not | |
157 // test whether the targets have been extracted and manipulated as | |
158 // raw jump targets. | |
159 bool is_used() const { return is_used_; } | |
160 | |
161 // True if the destination is used and the true target (respectively | |
162 // false target) was the fall through. If the target is backward, | |
163 // "fall through" included jumping unconditionally to it. | |
164 bool true_was_fall_through() const { | |
165 return is_used_ && true_is_fall_through_; | |
166 } | |
167 | |
168 bool false_was_fall_through() const { | |
169 return is_used_ && !true_is_fall_through_; | |
170 } | |
171 | |
172 // Emit a branch to one of the true or false targets, and bind the | |
173 // other target. Because this binds the fall-through target, it | |
174 // should be emitted in tail position (as the last thing when | |
175 // compiling an expression). | |
176 void Split(Condition cc) { | |
177 ASSERT(!is_used_); | |
178 if (true_is_fall_through_) { | |
179 false_target_->Branch(NegateCondition(cc)); | |
180 true_target_->Bind(); | |
181 } else { | |
182 true_target_->Branch(cc); | |
183 false_target_->Bind(); | |
184 } | |
185 is_used_ = true; | |
186 } | |
187 | |
188 // Emit an unconditional jump in tail position, to the true target | |
189 // (if the argument is true) or the false target. The "jump" will | |
190 // actually bind the jump target if it is forward, jump to it if it | |
191 // is backward. | |
192 void Goto(bool where) { | |
193 ASSERT(!is_used_); | |
194 JumpTarget* target = where ? true_target_ : false_target_; | |
195 if (target->is_bound()) { | |
196 target->Jump(); | |
197 } else { | |
198 target->Bind(); | |
199 } | |
200 is_used_ = true; | |
201 true_is_fall_through_ = where; | |
202 } | |
203 | |
204 // Mark this jump target as used as if Goto had been called, but | |
205 // without generating a jump or binding a label (the control effect | |
206 // should have already happened). This is used when the left | |
207 // subexpression of the short-circuit boolean operators are | |
208 // compiled. | |
209 void Use(bool where) { | |
210 ASSERT(!is_used_); | |
211 ASSERT((where ? true_target_ : false_target_)->is_bound()); | |
212 is_used_ = true; | |
213 true_is_fall_through_ = where; | |
214 } | |
215 | |
216 // Swap the true and false targets but keep the same actual label as | |
217 // the fall through. This is used when compiling negated | |
218 // expressions, where we want to swap the targets but preserve the | |
219 // state. | |
220 void Invert() { | |
221 JumpTarget* temp_target = true_target_; | |
222 true_target_ = false_target_; | |
223 false_target_ = temp_target; | |
224 | |
225 true_is_fall_through_ = !true_is_fall_through_; | |
226 } | |
227 | |
228 private: | |
229 // True and false jump targets. | |
230 JumpTarget* true_target_; | |
231 JumpTarget* false_target_; | |
232 | |
233 // Before using the destination: true if the true target is the | |
234 // preferred fall through, false if the false target is. After | |
235 // using the destination: true if the true target was actually used | |
236 // as the fall through, false if the false target was. | |
237 bool true_is_fall_through_; | |
238 | |
239 // True if the Split or Goto functions have been called. | |
240 bool is_used_; | |
241 }; | |
242 | |
243 | |
244 // ------------------------------------------------------------------------- | |
245 // Code generation state | |
246 | |
247 // The state is passed down the AST by the code generator (and back up, in | |
248 // the form of the state of the jump target pair). It is threaded through | |
249 // the call stack. Constructing a state implicitly pushes it on the owning | |
250 // code generator's stack of states, and destroying one implicitly pops it. | |
251 // | |
252 // The code generator state is only used for expressions, so statements have | |
253 // the initial state. | |
254 | |
255 class CodeGenState BASE_EMBEDDED { | |
256 public: | |
257 // Create an initial code generator state. Destroying the initial state | |
258 // leaves the code generator with a NULL state. | |
259 explicit CodeGenState(CodeGenerator* owner); | |
260 | |
261 // Create a code generator state based on a code generator's current | |
262 // state. The new state has its own control destination. | |
263 CodeGenState(CodeGenerator* owner, ControlDestination* destination); | |
264 | |
265 // Destroy a code generator state and restore the owning code generator's | |
266 // previous state. | |
267 ~CodeGenState(); | |
268 | |
269 // Accessors for the state. | |
270 ControlDestination* destination() const { return destination_; } | |
271 | |
272 private: | |
273 // The owning code generator. | |
274 CodeGenerator* owner_; | |
275 | |
276 // A control destination in case the expression has a control-flow | |
277 // effect. | |
278 ControlDestination* destination_; | |
279 | |
280 // The previous state of the owning code generator, restored when | |
281 // this state is destroyed. | |
282 CodeGenState* previous_; | |
283 }; | |
284 | |
285 | |
286 // ------------------------------------------------------------------------- | |
287 // Arguments allocation mode | |
288 | |
289 enum ArgumentsAllocationMode { | |
290 NO_ARGUMENTS_ALLOCATION, | |
291 EAGER_ARGUMENTS_ALLOCATION, | |
292 LAZY_ARGUMENTS_ALLOCATION | |
293 }; | |
294 | |
295 | |
296 // ------------------------------------------------------------------------- | |
297 // CodeGenerator | 44 // CodeGenerator |
298 | 45 |
299 class CodeGenerator: public AstVisitor { | 46 class CodeGenerator: public AstVisitor { |
300 public: | 47 public: |
301 static bool MakeCode(CompilationInfo* info); | 48 static bool MakeCode(CompilationInfo* info); |
302 | 49 |
303 // Printing of AST, etc. as requested by flags. | 50 // Printing of AST, etc. as requested by flags. |
304 static void MakeCodePrologue(CompilationInfo* info); | 51 static void MakeCodePrologue(CompilationInfo* info); |
305 | 52 |
306 // Allocate and install the code. | 53 // Allocate and install the code. |
307 static Handle<Code> MakeCodeEpilogue(MacroAssembler* masm, | 54 static Handle<Code> MakeCodeEpilogue(MacroAssembler* masm, |
308 Code::Flags flags, | 55 Code::Flags flags, |
309 CompilationInfo* info); | 56 CompilationInfo* info); |
310 | 57 |
311 // Print the code after compiling it. | 58 // Print the code after compiling it. |
312 static void PrintCode(Handle<Code> code, CompilationInfo* info); | 59 static void PrintCode(Handle<Code> code, CompilationInfo* info); |
313 | 60 |
314 #ifdef ENABLE_LOGGING_AND_PROFILING | 61 #ifdef ENABLE_LOGGING_AND_PROFILING |
315 static bool ShouldGenerateLog(Expression* type); | 62 static bool ShouldGenerateLog(Expression* type); |
316 #endif | 63 #endif |
317 | 64 |
318 static bool RecordPositions(MacroAssembler* masm, | 65 static bool RecordPositions(MacroAssembler* masm, |
319 int pos, | 66 int pos, |
320 bool right_here = false); | 67 bool right_here = false); |
321 | 68 |
322 // Accessors | |
323 MacroAssembler* masm() { return masm_; } | |
324 VirtualFrame* frame() const { return frame_; } | |
325 inline Handle<Script> script(); | |
326 | |
327 bool has_valid_frame() const { return frame_ != NULL; } | |
328 | |
329 // Set the virtual frame to be new_frame, with non-frame register | |
330 // reference counts given by non_frame_registers. The non-frame | |
331 // register reference counts of the old frame are returned in | |
332 // non_frame_registers. | |
333 void SetFrame(VirtualFrame* new_frame, RegisterFile* non_frame_registers); | |
334 | |
335 void DeleteFrame(); | |
336 | |
337 RegisterAllocator* allocator() const { return allocator_; } | |
338 | |
339 CodeGenState* state() { return state_; } | |
340 void set_state(CodeGenState* state) { state_ = state; } | |
341 | |
342 void AddDeferred(DeferredCode* code) { deferred_.Add(code); } | |
343 | |
344 bool in_spilled_code() const { return in_spilled_code_; } | |
345 void set_in_spilled_code(bool flag) { in_spilled_code_ = flag; } | |
346 | |
347 private: | 69 private: |
348 // Type of a member function that generates inline code for a native function. | |
349 typedef void (CodeGenerator::*InlineFunctionGenerator) | |
350 (ZoneList<Expression*>*); | |
351 | |
352 static const InlineFunctionGenerator kInlineFunctionGenerators[]; | |
353 | |
354 // Construction/Destruction | |
355 explicit CodeGenerator(MacroAssembler* masm); | |
356 | |
357 // Accessors | |
358 inline bool is_eval(); | |
359 inline Scope* scope(); | |
360 inline bool is_strict_mode(); | |
361 inline StrictModeFlag strict_mode_flag(); | |
362 | |
363 // Generating deferred code. | |
364 void ProcessDeferred(); | |
365 | |
366 // State | |
367 ControlDestination* destination() const { return state_->destination(); } | |
368 | |
369 // Track loop nesting level. | |
370 int loop_nesting() const { return loop_nesting_; } | |
371 void IncrementLoopNesting() { loop_nesting_++; } | |
372 void DecrementLoopNesting() { loop_nesting_--; } | |
373 | |
374 | |
375 // Node visitors. | |
376 void VisitStatements(ZoneList<Statement*>* statements); | |
377 | |
378 virtual void VisitSlot(Slot* node); | |
379 #define DEF_VISIT(type) \ | |
380 virtual void Visit##type(type* node); | |
381 AST_NODE_LIST(DEF_VISIT) | |
382 #undef DEF_VISIT | |
383 | |
384 // Visit a statement and then spill the virtual frame if control flow can | |
385 // reach the end of the statement (ie, it does not exit via break, | |
386 // continue, return, or throw). This function is used temporarily while | |
387 // the code generator is being transformed. | |
388 void VisitAndSpill(Statement* statement); | |
389 | |
390 // Visit a list of statements and then spill the virtual frame if control | |
391 // flow can reach the end of the list. | |
392 void VisitStatementsAndSpill(ZoneList<Statement*>* statements); | |
393 | |
394 // Main code generation function | |
395 void Generate(CompilationInfo* info); | |
396 | |
397 // Generate the return sequence code. Should be called no more than | |
398 // once per compiled function, immediately after binding the return | |
399 // target (which can not be done more than once). | |
400 void GenerateReturnSequence(Result* return_value); | |
401 | |
402 // Generate code for a fast smi loop. | |
403 void GenerateFastSmiLoop(ForStatement* node); | |
404 | |
405 // Returns the arguments allocation mode. | |
406 ArgumentsAllocationMode ArgumentsMode(); | |
407 | |
408 // Store the arguments object and allocate it if necessary. | |
409 Result StoreArgumentsObject(bool initial); | |
410 | |
411 // The following are used by class Reference. | |
412 void LoadReference(Reference* ref); | |
413 void UnloadReference(Reference* ref); | |
414 | |
415 Operand SlotOperand(Slot* slot, Register tmp); | |
416 | |
417 Operand ContextSlotOperandCheckExtensions(Slot* slot, | |
418 Result tmp, | |
419 JumpTarget* slow); | |
420 | |
421 // Expressions | |
422 void LoadCondition(Expression* x, | |
423 ControlDestination* destination, | |
424 bool force_control); | |
425 void Load(Expression* expr); | |
426 void LoadGlobal(); | |
427 void LoadGlobalReceiver(); | |
428 | |
429 // Generate code to push the value of an expression on top of the frame | |
430 // and then spill the frame fully to memory. This function is used | |
431 // temporarily while the code generator is being transformed. | |
432 void LoadAndSpill(Expression* expression); | |
433 | |
434 // Read a value from a slot and leave it on top of the expression stack. | |
435 void LoadFromSlot(Slot* slot, TypeofState typeof_state); | |
436 void LoadFromSlotCheckForArguments(Slot* slot, TypeofState state); | |
437 Result LoadFromGlobalSlotCheckExtensions(Slot* slot, | |
438 TypeofState typeof_state, | |
439 JumpTarget* slow); | |
440 | |
441 // Support for loading from local/global variables and arguments | |
442 // whose location is known unless they are shadowed by | |
443 // eval-introduced bindings. Generates no code for unsupported slot | |
444 // types and therefore expects to fall through to the slow jump target. | |
445 void EmitDynamicLoadFromSlotFastCase(Slot* slot, | |
446 TypeofState typeof_state, | |
447 Result* result, | |
448 JumpTarget* slow, | |
449 JumpTarget* done); | |
450 | |
451 // Store the value on top of the expression stack into a slot, leaving the | |
452 // value in place. | |
453 void StoreToSlot(Slot* slot, InitState init_state); | |
454 | |
455 // Support for compiling assignment expressions. | |
456 void EmitSlotAssignment(Assignment* node); | |
457 void EmitNamedPropertyAssignment(Assignment* node); | |
458 void EmitKeyedPropertyAssignment(Assignment* node); | |
459 | |
460 // Receiver is passed on the frame and not consumed. | |
461 Result EmitNamedLoad(Handle<String> name, bool is_contextual); | |
462 | |
463 // If the store is contextual, value is passed on the frame and consumed. | |
464 // Otherwise, receiver and value are passed on the frame and consumed. | |
465 Result EmitNamedStore(Handle<String> name, bool is_contextual); | |
466 | |
467 // Load a property of an object, returning it in a Result. | |
468 // The object and the property name are passed on the stack, and | |
469 // not changed. | |
470 Result EmitKeyedLoad(); | |
471 | |
472 // Receiver, key, and value are passed on the frame and consumed. | |
473 Result EmitKeyedStore(StaticType* key_type); | |
474 | |
475 // Special code for typeof expressions: Unfortunately, we must | |
476 // be careful when loading the expression in 'typeof' | |
477 // expressions. We are not allowed to throw reference errors for | |
478 // non-existing properties of the global object, so we must make it | |
479 // look like an explicit property access, instead of an access | |
480 // through the context chain. | |
481 void LoadTypeofExpression(Expression* x); | |
482 | |
483 // Translate the value on top of the frame into control flow to the | |
484 // control destination. | |
485 void ToBoolean(ControlDestination* destination); | |
486 | |
487 // Generate code that computes a shortcutting logical operation. | |
488 void GenerateLogicalBooleanOperation(BinaryOperation* node); | |
489 | |
490 void GenericBinaryOperation(BinaryOperation* expr, | |
491 OverwriteMode overwrite_mode); | |
492 | |
493 // Generate a stub call from the virtual frame. | |
494 Result GenerateGenericBinaryOpStubCall(GenericBinaryOpStub* stub, | |
495 Result* left, | |
496 Result* right); | |
497 | |
498 // Emits code sequence that jumps to a JumpTarget if the inputs | |
499 // are both smis. Cannot be in MacroAssembler because it takes | |
500 // advantage of TypeInfo to skip unneeded checks. | |
501 void JumpIfBothSmiUsingTypeInfo(Result* left, | |
502 Result* right, | |
503 JumpTarget* both_smi); | |
504 | |
505 // Emits code sequence that jumps to deferred code if the input | |
506 // is not a smi. Cannot be in MacroAssembler because it takes | |
507 // advantage of TypeInfo to skip unneeded checks. | |
508 void JumpIfNotSmiUsingTypeInfo(Register reg, | |
509 TypeInfo type, | |
510 DeferredCode* deferred); | |
511 | |
512 // Emits code sequence that jumps to deferred code if the inputs | |
513 // are not both smis. Cannot be in MacroAssembler because it takes | |
514 // advantage of TypeInfo to skip unneeded checks. | |
515 void JumpIfNotBothSmiUsingTypeInfo(Register left, | |
516 Register right, | |
517 TypeInfo left_info, | |
518 TypeInfo right_info, | |
519 DeferredCode* deferred); | |
520 | |
521 // If possible, combine two constant smi values using op to produce | |
522 // a smi result, and push it on the virtual frame, all at compile time. | |
523 // Returns true if it succeeds. Otherwise it has no effect. | |
524 bool FoldConstantSmis(Token::Value op, int left, int right); | |
525 | |
526 // Emit code to perform a binary operation on a constant | |
527 // smi and a likely smi. Consumes the Result *operand. | |
528 Result ConstantSmiBinaryOperation(BinaryOperation* expr, | |
529 Result* operand, | |
530 Handle<Object> constant_operand, | |
531 bool reversed, | |
532 OverwriteMode overwrite_mode); | |
533 | |
534 // Emit code to perform a binary operation on two likely smis. | |
535 // The code to handle smi arguments is produced inline. | |
536 // Consumes the Results *left and *right. | |
537 Result LikelySmiBinaryOperation(BinaryOperation* expr, | |
538 Result* left, | |
539 Result* right, | |
540 OverwriteMode overwrite_mode); | |
541 | |
542 void Comparison(AstNode* node, | |
543 Condition cc, | |
544 bool strict, | |
545 ControlDestination* destination); | |
546 | |
547 // If at least one of the sides is a constant smi, generate optimized code. | |
548 void ConstantSmiComparison(Condition cc, | |
549 bool strict, | |
550 ControlDestination* destination, | |
551 Result* left_side, | |
552 Result* right_side, | |
553 bool left_side_constant_smi, | |
554 bool right_side_constant_smi, | |
555 bool is_loop_condition); | |
556 | |
557 void GenerateInlineNumberComparison(Result* left_side, | |
558 Result* right_side, | |
559 Condition cc, | |
560 ControlDestination* dest); | |
561 | |
562 // To prevent long attacker-controlled byte sequences, integer constants | |
563 // from the JavaScript source are loaded in two parts if they are larger | |
564 // than 16 bits. | |
565 static const int kMaxSmiInlinedBits = 16; | |
566 bool IsUnsafeSmi(Handle<Object> value); | |
567 // Load an integer constant x into a register target using | |
568 // at most 16 bits of user-controlled data per assembly operation. | |
569 void LoadUnsafeSmi(Register target, Handle<Object> value); | |
570 | |
571 void CallWithArguments(ZoneList<Expression*>* arguments, | |
572 CallFunctionFlags flags, | |
573 int position); | |
574 | |
575 // An optimized implementation of expressions of the form | |
576 // x.apply(y, arguments). We call x the applicand and y the receiver. | |
577 // The optimization avoids allocating an arguments object if possible. | |
578 void CallApplyLazy(Expression* applicand, | |
579 Expression* receiver, | |
580 VariableProxy* arguments, | |
581 int position); | |
582 | |
583 void CheckStack(); | |
584 | |
585 bool CheckForInlineRuntimeCall(CallRuntime* node); | |
586 | |
587 void ProcessDeclarations(ZoneList<Declaration*>* declarations); | |
588 | |
589 // Declare global variables and functions in the given array of | |
590 // name/value pairs. | |
591 void DeclareGlobals(Handle<FixedArray> pairs); | |
592 | |
593 // Instantiate the function based on the shared function info. | |
594 void InstantiateFunction(Handle<SharedFunctionInfo> function_info, | |
595 bool pretenure); | |
596 | |
597 // Support for type checks. | |
598 void GenerateIsSmi(ZoneList<Expression*>* args); | |
599 void GenerateIsNonNegativeSmi(ZoneList<Expression*>* args); | |
600 void GenerateIsArray(ZoneList<Expression*>* args); | |
601 void GenerateIsRegExp(ZoneList<Expression*>* args); | |
602 void GenerateIsObject(ZoneList<Expression*>* args); | |
603 void GenerateIsSpecObject(ZoneList<Expression*>* args); | |
604 void GenerateIsFunction(ZoneList<Expression*>* args); | |
605 void GenerateIsUndetectableObject(ZoneList<Expression*>* args); | |
606 void GenerateIsStringWrapperSafeForDefaultValueOf( | |
607 ZoneList<Expression*>* args); | |
608 | |
609 // Support for construct call checks. | |
610 void GenerateIsConstructCall(ZoneList<Expression*>* args); | |
611 | |
612 // Support for arguments.length and arguments[?]. | |
613 void GenerateArgumentsLength(ZoneList<Expression*>* args); | |
614 void GenerateArguments(ZoneList<Expression*>* args); | |
615 | |
616 // Support for accessing the class and value fields of an object. | |
617 void GenerateClassOf(ZoneList<Expression*>* args); | |
618 void GenerateValueOf(ZoneList<Expression*>* args); | |
619 void GenerateSetValueOf(ZoneList<Expression*>* args); | |
620 | |
621 // Fast support for charCodeAt(n). | |
622 void GenerateStringCharCodeAt(ZoneList<Expression*>* args); | |
623 | |
624 // Fast support for string.charAt(n) and string[n]. | |
625 void GenerateStringCharFromCode(ZoneList<Expression*>* args); | |
626 | |
627 // Fast support for string.charAt(n) and string[n]. | |
628 void GenerateStringCharAt(ZoneList<Expression*>* args); | |
629 | |
630 // Fast support for object equality testing. | |
631 void GenerateObjectEquals(ZoneList<Expression*>* args); | |
632 | |
633 void GenerateLog(ZoneList<Expression*>* args); | |
634 | |
635 void GenerateGetFramePointer(ZoneList<Expression*>* args); | |
636 | |
637 // Fast support for Math.random(). | |
638 void GenerateRandomHeapNumber(ZoneList<Expression*>* args); | |
639 | |
640 // Fast support for StringAdd. | |
641 void GenerateStringAdd(ZoneList<Expression*>* args); | |
642 | |
643 // Fast support for SubString. | |
644 void GenerateSubString(ZoneList<Expression*>* args); | |
645 | |
646 // Fast support for StringCompare. | |
647 void GenerateStringCompare(ZoneList<Expression*>* args); | |
648 | |
649 // Support for direct calls from JavaScript to native RegExp code. | |
650 void GenerateRegExpExec(ZoneList<Expression*>* args); | |
651 | |
652 void GenerateRegExpConstructResult(ZoneList<Expression*>* args); | |
653 | |
654 // Support for fast native caches. | |
655 void GenerateGetFromCache(ZoneList<Expression*>* args); | |
656 | |
657 // Fast support for number to string. | |
658 void GenerateNumberToString(ZoneList<Expression*>* args); | |
659 | |
660 // Fast swapping of elements. Takes three expressions, the object and two | |
661 // indices. This should only be used if the indices are known to be | |
662 // non-negative and within bounds of the elements array at the call site. | |
663 void GenerateSwapElements(ZoneList<Expression*>* args); | |
664 | |
665 // Fast call for custom callbacks. | |
666 void GenerateCallFunction(ZoneList<Expression*>* args); | |
667 | |
668 // Fast call to math functions. | |
669 void GenerateMathPow(ZoneList<Expression*>* args); | |
670 void GenerateMathSin(ZoneList<Expression*>* args); | |
671 void GenerateMathCos(ZoneList<Expression*>* args); | |
672 void GenerateMathSqrt(ZoneList<Expression*>* args); | |
673 void GenerateMathLog(ZoneList<Expression*>* args); | |
674 | |
675 // Check whether two RegExps are equivalent. | |
676 void GenerateIsRegExpEquivalent(ZoneList<Expression*>* args); | |
677 | |
678 void GenerateHasCachedArrayIndex(ZoneList<Expression*>* args); | |
679 void GenerateGetCachedArrayIndex(ZoneList<Expression*>* args); | |
680 void GenerateFastAsciiArrayJoin(ZoneList<Expression*>* args); | |
681 | |
682 // Simple condition analysis. | |
683 enum ConditionAnalysis { | |
684 ALWAYS_TRUE, | |
685 ALWAYS_FALSE, | |
686 DONT_KNOW | |
687 }; | |
688 ConditionAnalysis AnalyzeCondition(Expression* cond); | |
689 | |
690 // Methods used to indicate which source code is generated for. Source | |
691 // positions are collected by the assembler and emitted with the relocation | |
692 // information. | |
693 void CodeForFunctionPosition(FunctionLiteral* fun); | |
694 void CodeForReturnPosition(FunctionLiteral* fun); | |
695 void CodeForStatementPosition(Statement* node); | |
696 void CodeForDoWhileConditionPosition(DoWhileStatement* stmt); | |
697 void CodeForSourcePosition(int pos); | |
698 | |
699 void SetTypeForStackSlot(Slot* slot, TypeInfo info); | |
700 | |
701 #ifdef DEBUG | |
702 // True if the registers are valid for entry to a block. There should | |
703 // be no frame-external references to (non-reserved) registers. | |
704 bool HasValidEntryRegisters(); | |
705 #endif | |
706 | |
707 ZoneList<DeferredCode*> deferred_; | |
708 | |
709 // Assembler | |
710 MacroAssembler* masm_; // to generate code | |
711 | |
712 CompilationInfo* info_; | |
713 | |
714 // Code generation state | |
715 VirtualFrame* frame_; | |
716 RegisterAllocator* allocator_; | |
717 CodeGenState* state_; | |
718 int loop_nesting_; | |
719 | |
720 // Jump targets. | |
721 // The target of the return from the function. | |
722 BreakTarget function_return_; | |
723 | |
724 // True if the function return is shadowed (ie, jumping to the target | |
725 // function_return_ does not jump to the true function return, but rather | |
726 // to some unlinking code). | |
727 bool function_return_is_shadowed_; | |
728 | |
729 // True when we are in code that expects the virtual frame to be fully | |
730 // spilled. Some virtual frame function are disabled in DEBUG builds when | |
731 // called from spilled code, because they do not leave the virtual frame | |
732 // in a spilled state. | |
733 bool in_spilled_code_; | |
734 | |
735 friend class VirtualFrame; | |
736 friend class Isolate; | |
737 friend class JumpTarget; | |
738 friend class Reference; | |
739 friend class Result; | |
740 friend class FastCodeGenerator; | |
741 friend class FullCodeGenerator; | |
742 friend class FullCodeGenSyntaxChecker; | |
743 | |
744 friend class CodeGeneratorPatcher; // Used in test-log-stack-tracer.cc | |
745 friend class InlineRuntimeFunctionsTable; | |
746 | |
747 DISALLOW_COPY_AND_ASSIGN(CodeGenerator); | 70 DISALLOW_COPY_AND_ASSIGN(CodeGenerator); |
748 }; | 71 }; |
749 | 72 |
750 | 73 |
751 } } // namespace v8::internal | 74 } } // namespace v8::internal |
752 | 75 |
753 #endif // V8_X64_CODEGEN_X64_H_ | 76 #endif // V8_X64_CODEGEN_X64_H_ |
OLD | NEW |