| OLD | NEW |
| 1 // Copyright 2011 the V8 project authors. All rights reserved. | 1 // Copyright 2011 the V8 project authors. All rights reserved. |
| 2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
| 3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
| 4 // met: | 4 // met: |
| 5 // | 5 // |
| 6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
| 7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
| 8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
| 9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
| 10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
| (...skipping 30 matching lines...) Expand all Loading... |
| 41 // ------------------------------------------------------------------------- | 41 // ------------------------------------------------------------------------- |
| 42 // MacroAssembler implementation. | 42 // MacroAssembler implementation. |
| 43 | 43 |
| 44 MacroAssembler::MacroAssembler(void* buffer, int size) | 44 MacroAssembler::MacroAssembler(void* buffer, int size) |
| 45 : Assembler(buffer, size), | 45 : Assembler(buffer, size), |
| 46 generating_stub_(false), | 46 generating_stub_(false), |
| 47 allow_stub_calls_(true), | 47 allow_stub_calls_(true), |
| 48 code_object_(Heap::undefined_value()) { | 48 code_object_(Heap::undefined_value()) { |
| 49 } | 49 } |
| 50 | 50 |
| 51 |
| 52 void MacroAssembler::IncrementalMarkingRecordWrite(Register object, |
| 53 Register value, |
| 54 Register scratch) { |
| 55 ASSERT(!object.is(scratch)); |
| 56 ASSERT(!value.is(scratch)); |
| 57 push(eax); |
| 58 push(ecx); |
| 59 push(edx); |
| 60 PrepareCallCFunction(2, scratch); |
| 61 mov(Operand(esp, 0 * kPointerSize), object); |
| 62 mov(Operand(esp, 1 * kPointerSize), value); |
| 63 CallCFunction( |
| 64 ExternalReference::incremental_marking_record_write_function(), 2); |
| 65 pop(edx); |
| 66 pop(ecx); |
| 67 pop(eax); |
| 68 } |
| 69 |
| 70 |
| 51 void MacroAssembler::RecordWriteHelper(Register object, | 71 void MacroAssembler::RecordWriteHelper(Register object, |
| 52 Register addr, | 72 Register addr, |
| 53 Register scratch, | 73 Register scratch, |
| 54 SaveFPRegsMode save_fp) { | 74 SaveFPRegsMode save_fp) { |
| 55 if (FLAG_debug_code) { | 75 if (FLAG_debug_code) { |
| 56 // Check that the object is not in new space. | 76 // Check that the object is not in new space. |
| 57 Label not_in_new_space; | 77 Label not_in_new_space; |
| 58 InNewSpace(object, scratch, not_equal, ¬_in_new_space); | 78 InNewSpace(object, scratch, not_equal, ¬_in_new_space); |
| 59 Abort("new-space object passed to RecordWriteHelper"); | 79 Abort("new-space object passed to RecordWriteHelper"); |
| 60 bind(¬_in_new_space); | 80 bind(¬_in_new_space); |
| (...skipping 20 matching lines...) Expand all Loading... |
| 81 } | 101 } |
| 82 | 102 |
| 83 | 103 |
| 84 void MacroAssembler::RecordWrite(Register object, | 104 void MacroAssembler::RecordWrite(Register object, |
| 85 int offset, | 105 int offset, |
| 86 Register value, | 106 Register value, |
| 87 Register scratch, | 107 Register scratch, |
| 88 SaveFPRegsMode save_fp) { | 108 SaveFPRegsMode save_fp) { |
| 89 // First, check if a write barrier is even needed. The tests below | 109 // First, check if a write barrier is even needed. The tests below |
| 90 // catch stores of Smis and stores into young gen. | 110 // catch stores of Smis and stores into young gen. |
| 91 NearLabel done; | 111 Label done; |
| 92 | 112 |
| 93 // Skip barrier if writing a smi. | 113 // Skip barrier if writing a smi. |
| 94 ASSERT_EQ(0, kSmiTag); | 114 ASSERT_EQ(0, kSmiTag); |
| 95 test(value, Immediate(kSmiTagMask)); | 115 test(value, Immediate(kSmiTagMask)); |
| 96 j(zero, &done); | 116 j(zero, &done); |
| 97 | 117 |
| 118 IncrementalMarkingRecordWrite(object, value, scratch); |
| 119 |
| 98 InNewSpace(object, value, equal, &done); | 120 InNewSpace(object, value, equal, &done); |
| 99 | 121 |
| 100 // The offset is relative to a tagged or untagged HeapObject pointer, | 122 // The offset is relative to a tagged or untagged HeapObject pointer, |
| 101 // so either offset or offset + kHeapObjectTag must be a | 123 // so either offset or offset + kHeapObjectTag must be a |
| 102 // multiple of kPointerSize. | 124 // multiple of kPointerSize. |
| 103 ASSERT(IsAligned(offset, kPointerSize) || | 125 ASSERT(IsAligned(offset, kPointerSize) || |
| 104 IsAligned(offset + kHeapObjectTag, kPointerSize)); | 126 IsAligned(offset + kHeapObjectTag, kPointerSize)); |
| 105 | 127 |
| 106 Register dst = scratch; | 128 Register dst = scratch; |
| 107 if (offset != 0) { | 129 if (offset != 0) { |
| (...skipping 1900 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 2008 | 2030 |
| 2009 // Check that the code was patched as expected. | 2031 // Check that the code was patched as expected. |
| 2010 ASSERT(masm_.pc_ == address_ + size_); | 2032 ASSERT(masm_.pc_ == address_ + size_); |
| 2011 ASSERT(masm_.reloc_info_writer.pos() == address_ + size_ + Assembler::kGap); | 2033 ASSERT(masm_.reloc_info_writer.pos() == address_ + size_ + Assembler::kGap); |
| 2012 } | 2034 } |
| 2013 | 2035 |
| 2014 | 2036 |
| 2015 } } // namespace v8::internal | 2037 } } // namespace v8::internal |
| 2016 | 2038 |
| 2017 #endif // V8_TARGET_ARCH_IA32 | 2039 #endif // V8_TARGET_ARCH_IA32 |
| OLD | NEW |