| OLD | NEW |
| 1 // Copyright 2009 the V8 project authors. All rights reserved. | 1 // Copyright 2009 the V8 project authors. All rights reserved. |
| 2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
| 3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
| 4 // met: | 4 // met: |
| 5 // | 5 // |
| 6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
| 7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
| 8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
| 9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
| 10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
| (...skipping 406 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 417 void VirtualFrame::MergeMoveMemoryToRegisters(VirtualFrame* expected) { | 417 void VirtualFrame::MergeMoveMemoryToRegisters(VirtualFrame* expected) { |
| 418 // Move memory, constants, and copies to registers. This is the | 418 // Move memory, constants, and copies to registers. This is the |
| 419 // final step and is done from the bottom up so that the backing | 419 // final step and is done from the bottom up so that the backing |
| 420 // elements of copies are in their correct locations when we | 420 // elements of copies are in their correct locations when we |
| 421 // encounter the copies. | 421 // encounter the copies. |
| 422 for (int i = 0; i < kNumRegisters; i++) { | 422 for (int i = 0; i < kNumRegisters; i++) { |
| 423 int index = expected->register_locations_[i]; | 423 int index = expected->register_locations_[i]; |
| 424 if (index != kIllegalIndex) { | 424 if (index != kIllegalIndex) { |
| 425 FrameElement source = elements_[index]; | 425 FrameElement source = elements_[index]; |
| 426 FrameElement target = expected->elements_[index]; | 426 FrameElement target = expected->elements_[index]; |
| 427 Register target_reg = { i }; |
| 428 ASSERT(expected->elements_[index].reg().is(target_reg)); |
| 427 switch (source.type()) { | 429 switch (source.type()) { |
| 428 case FrameElement::INVALID: // Fall through. | 430 case FrameElement::INVALID: // Fall through. |
| 429 UNREACHABLE(); | 431 UNREACHABLE(); |
| 430 break; | 432 break; |
| 431 case FrameElement::REGISTER: | 433 case FrameElement::REGISTER: |
| 432 ASSERT(source.reg().is(target.reg())); | 434 ASSERT(source.reg().is(target_reg)); |
| 433 continue; // Go to next iteration. Skips Use(target.reg()) below. | 435 continue; // Go to next iteration. Skips Use(target_reg) below. |
| 434 break; | 436 break; |
| 435 case FrameElement::MEMORY: | 437 case FrameElement::MEMORY: |
| 436 ASSERT(index <= stack_pointer_); | 438 ASSERT(index <= stack_pointer_); |
| 437 __ mov(target.reg(), Operand(ebp, fp_relative(index))); | 439 __ mov(target_reg, Operand(ebp, fp_relative(index))); |
| 438 break; | 440 break; |
| 439 | 441 |
| 440 case FrameElement::CONSTANT: | 442 case FrameElement::CONSTANT: |
| 441 if (cgen_->IsUnsafeSmi(source.handle())) { | 443 if (cgen_->IsUnsafeSmi(source.handle())) { |
| 442 cgen_->LoadUnsafeSmi(target.reg(), source.handle()); | 444 cgen_->LoadUnsafeSmi(target_reg, source.handle()); |
| 443 } else { | 445 } else { |
| 444 __ Set(target.reg(), Immediate(source.handle())); | 446 __ Set(target_reg, Immediate(source.handle())); |
| 445 } | 447 } |
| 446 break; | 448 break; |
| 447 | 449 |
| 448 case FrameElement::COPY: { | 450 case FrameElement::COPY: { |
| 449 int backing_index = source.index(); | 451 int backing_index = source.index(); |
| 450 FrameElement backing = elements_[backing_index]; | 452 FrameElement backing = elements_[backing_index]; |
| 451 ASSERT(backing.is_memory() || backing.is_register()); | 453 ASSERT(backing.is_memory() || backing.is_register()); |
| 452 if (backing.is_memory()) { | 454 if (backing.is_memory()) { |
| 453 ASSERT(backing_index <= stack_pointer_); | 455 ASSERT(backing_index <= stack_pointer_); |
| 454 // Code optimization if backing store should also move | 456 // Code optimization if backing store should also move |
| 455 // to a register: move backing store to its register first. | 457 // to a register: move backing store to its register first. |
| 456 if (expected->elements_[backing_index].is_register()) { | 458 if (expected->elements_[backing_index].is_register()) { |
| 457 FrameElement new_backing = expected->elements_[backing_index]; | 459 FrameElement new_backing = expected->elements_[backing_index]; |
| 458 Register new_backing_reg = new_backing.reg(); | 460 Register new_backing_reg = new_backing.reg(); |
| 459 ASSERT(!is_used(new_backing_reg)); | 461 ASSERT(!is_used(new_backing_reg)); |
| 460 elements_[backing_index] = new_backing; | 462 elements_[backing_index] = new_backing; |
| 461 Use(new_backing_reg, backing_index); | 463 Use(new_backing_reg, backing_index); |
| 462 __ mov(new_backing_reg, | 464 __ mov(new_backing_reg, |
| 463 Operand(ebp, fp_relative(backing_index))); | 465 Operand(ebp, fp_relative(backing_index))); |
| 464 __ mov(target.reg(), new_backing_reg); | 466 __ mov(target_reg, new_backing_reg); |
| 465 } else { | 467 } else { |
| 466 __ mov(target.reg(), Operand(ebp, fp_relative(backing_index))); | 468 __ mov(target_reg, Operand(ebp, fp_relative(backing_index))); |
| 467 } | 469 } |
| 468 } else { | 470 } else { |
| 469 __ mov(target.reg(), backing.reg()); | 471 __ mov(target_reg, backing.reg()); |
| 470 } | 472 } |
| 471 } | 473 } |
| 472 } | 474 } |
| 473 // Ensure the proper sync state. If the source was memory no | 475 // Ensure the proper sync state. If the source was memory no |
| 474 // code needs to be emitted. | 476 // code needs to be emitted. |
| 475 if (target.is_synced() && !source.is_synced()) { | 477 if (target.is_synced() && !source.is_synced()) { |
| 476 __ mov(Operand(ebp, fp_relative(index)), target.reg()); | 478 __ mov(Operand(ebp, fp_relative(index)), target_reg); |
| 477 } | 479 } |
| 478 Use(target.reg(), index); | 480 Use(target_reg, index); |
| 479 elements_[index] = target; | 481 elements_[index] = target; |
| 480 } | 482 } |
| 481 } | 483 } |
| 482 } | 484 } |
| 483 | 485 |
| 484 | 486 |
| 485 void VirtualFrame::Enter() { | 487 void VirtualFrame::Enter() { |
| 486 // Registers live on entry: esp, ebp, esi, edi. | 488 // Registers live on entry: esp, ebp, esi, edi. |
| 487 Comment cmnt(masm_, "[ Enter JS frame"); | 489 Comment cmnt(masm_, "[ Enter JS frame"); |
| 488 | 490 |
| (...skipping 630 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 1119 ASSERT(stack_pointer_ == elements_.length() - 1); | 1121 ASSERT(stack_pointer_ == elements_.length() - 1); |
| 1120 elements_.Add(FrameElement::MemoryElement()); | 1122 elements_.Add(FrameElement::MemoryElement()); |
| 1121 stack_pointer_++; | 1123 stack_pointer_++; |
| 1122 __ push(immediate); | 1124 __ push(immediate); |
| 1123 } | 1125 } |
| 1124 | 1126 |
| 1125 | 1127 |
| 1126 #undef __ | 1128 #undef __ |
| 1127 | 1129 |
| 1128 } } // namespace v8::internal | 1130 } } // namespace v8::internal |
| OLD | NEW |