| OLD | NEW |
| 1 // Copyright (c) 1994-2006 Sun Microsystems Inc. | 1 // Copyright (c) 1994-2006 Sun Microsystems Inc. |
| 2 // All Rights Reserved. | 2 // All Rights Reserved. |
| 3 // | 3 // |
| 4 // Redistribution and use in source and binary forms, with or without | 4 // Redistribution and use in source and binary forms, with or without |
| 5 // modification, are permitted provided that the following conditions are | 5 // modification, are permitted provided that the following conditions are |
| 6 // met: | 6 // met: |
| 7 // | 7 // |
| 8 // - Redistributions of source code must retain the above copyright notice, | 8 // - Redistributions of source code must retain the above copyright notice, |
| 9 // this list of conditions and the following disclaimer. | 9 // this list of conditions and the following disclaimer. |
| 10 // | 10 // |
| (...skipping 120 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 131 uint64_t a; | 131 uint64_t a; |
| 132 uint64_t b; | 132 uint64_t b; |
| 133 } double_negate_constant = {V8_UINT64_C(0x8000000000000000), | 133 } double_negate_constant = {V8_UINT64_C(0x8000000000000000), |
| 134 V8_UINT64_C(0x8000000000000000)}; | 134 V8_UINT64_C(0x8000000000000000)}; |
| 135 | 135 |
| 136 const char* const RelocInfo::kFillerCommentString = "DEOPTIMIZATION PADDING"; | 136 const char* const RelocInfo::kFillerCommentString = "DEOPTIMIZATION PADDING"; |
| 137 | 137 |
| 138 // ----------------------------------------------------------------------------- | 138 // ----------------------------------------------------------------------------- |
| 139 // Implementation of AssemblerBase | 139 // Implementation of AssemblerBase |
| 140 | 140 |
| 141 AssemblerBase::AssemblerBase(Isolate* isolate, void* buffer, int buffer_size) | 141 AssemblerBase::IsolateData::IsolateData(Isolate* isolate) |
| 142 : isolate_(isolate), | 142 : serializer_enabled_(isolate->serializer_enabled()) |
| 143 jit_cookie_(0), | 143 #if V8_TARGET_ARCH_IA32 || V8_TARGET_ARCH_X64 |
| 144 , |
| 145 max_old_generation_size_(isolate->heap()->MaxOldGenerationSize()) |
| 146 #endif |
| 147 #if V8_TARGET_ARCH_X64 |
| 148 , |
| 149 code_range_start_( |
| 150 isolate->heap()->memory_allocator()->code_range()->start()) |
| 151 #endif |
| 152 { |
| 153 } |
| 154 |
| 155 AssemblerBase::AssemblerBase(IsolateData isolate_data, void* buffer, |
| 156 int buffer_size) |
| 157 : isolate_data_(isolate_data), |
| 144 enabled_cpu_features_(0), | 158 enabled_cpu_features_(0), |
| 145 emit_debug_code_(FLAG_debug_code), | 159 emit_debug_code_(FLAG_debug_code), |
| 146 predictable_code_size_(false), | 160 predictable_code_size_(false), |
| 147 // We may use the assembler without an isolate. | |
| 148 serializer_enabled_(isolate && isolate->serializer_enabled()), | |
| 149 constant_pool_available_(false) { | 161 constant_pool_available_(false) { |
| 150 DCHECK_NOT_NULL(isolate); | |
| 151 if (FLAG_mask_constants_with_cookie) { | |
| 152 jit_cookie_ = isolate->random_number_generator()->NextInt(); | |
| 153 } | |
| 154 own_buffer_ = buffer == NULL; | 162 own_buffer_ = buffer == NULL; |
| 155 if (buffer_size == 0) buffer_size = kMinimalBufferSize; | 163 if (buffer_size == 0) buffer_size = kMinimalBufferSize; |
| 156 DCHECK(buffer_size > 0); | 164 DCHECK(buffer_size > 0); |
| 157 if (own_buffer_) buffer = NewArray<byte>(buffer_size); | 165 if (own_buffer_) buffer = NewArray<byte>(buffer_size); |
| 158 buffer_ = static_cast<byte*>(buffer); | 166 buffer_ = static_cast<byte*>(buffer); |
| 159 buffer_size_ = buffer_size; | 167 buffer_size_ = buffer_size; |
| 160 | |
| 161 pc_ = buffer_; | 168 pc_ = buffer_; |
| 162 } | 169 } |
| 163 | 170 |
| 164 | |
| 165 AssemblerBase::~AssemblerBase() { | 171 AssemblerBase::~AssemblerBase() { |
| 166 if (own_buffer_) DeleteArray(buffer_); | 172 if (own_buffer_) DeleteArray(buffer_); |
| 167 } | 173 } |
| 168 | 174 |
| 169 | |
| 170 void AssemblerBase::FlushICache(Isolate* isolate, void* start, size_t size) { | 175 void AssemblerBase::FlushICache(Isolate* isolate, void* start, size_t size) { |
| 171 if (size == 0) return; | 176 if (size == 0) return; |
| 172 | 177 |
| 173 #if defined(USE_SIMULATOR) | 178 #if defined(USE_SIMULATOR) |
| 174 base::LockGuard<base::Mutex> lock_guard(isolate->simulator_i_cache_mutex()); | 179 base::LockGuard<base::Mutex> lock_guard(isolate->simulator_i_cache_mutex()); |
| 175 Simulator::FlushICache(isolate->simulator_i_cache(), start, size); | 180 Simulator::FlushICache(isolate->simulator_i_cache(), start, size); |
| 176 #else | 181 #else |
| 177 CpuFeatures::FlushICache(start, size); | 182 CpuFeatures::FlushICache(start, size); |
| 178 #endif // USE_SIMULATOR | 183 #endif // USE_SIMULATOR |
| 179 } | 184 } |
| 180 | 185 |
| 181 | 186 void AssemblerBase::Print(Isolate* isolate) { |
| 182 void AssemblerBase::Print() { | |
| 183 OFStream os(stdout); | 187 OFStream os(stdout); |
| 184 v8::internal::Disassembler::Decode(isolate(), &os, buffer_, pc_, nullptr); | 188 v8::internal::Disassembler::Decode(isolate, &os, buffer_, pc_, nullptr); |
| 185 } | 189 } |
| 186 | 190 |
| 187 | 191 |
| 188 // ----------------------------------------------------------------------------- | 192 // ----------------------------------------------------------------------------- |
| 189 // Implementation of PredictableCodeSizeScope | 193 // Implementation of PredictableCodeSizeScope |
| 190 | 194 |
| 191 PredictableCodeSizeScope::PredictableCodeSizeScope(AssemblerBase* assembler) | 195 PredictableCodeSizeScope::PredictableCodeSizeScope(AssemblerBase* assembler) |
| 192 : PredictableCodeSizeScope(assembler, -1) {} | 196 : PredictableCodeSizeScope(assembler, -1) {} |
| 193 | 197 |
| 194 | 198 |
| (...skipping 106 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 301 const int kChunkBits = 7; | 305 const int kChunkBits = 7; |
| 302 const int kChunkMask = (1 << kChunkBits) - 1; | 306 const int kChunkMask = (1 << kChunkBits) - 1; |
| 303 const int kLastChunkTagBits = 1; | 307 const int kLastChunkTagBits = 1; |
| 304 const int kLastChunkTagMask = 1; | 308 const int kLastChunkTagMask = 1; |
| 305 const int kLastChunkTag = 1; | 309 const int kLastChunkTag = 1; |
| 306 | 310 |
| 307 const int kCodeWithIdTag = 0; | 311 const int kCodeWithIdTag = 0; |
| 308 const int kDeoptReasonTag = 1; | 312 const int kDeoptReasonTag = 1; |
| 309 | 313 |
| 310 void RelocInfo::update_wasm_memory_reference( | 314 void RelocInfo::update_wasm_memory_reference( |
| 311 Address old_base, Address new_base, ICacheFlushMode icache_flush_mode) { | 315 Isolate* isolate, Address old_base, Address new_base, |
| 316 ICacheFlushMode icache_flush_mode) { |
| 312 DCHECK(IsWasmMemoryReference(rmode_)); | 317 DCHECK(IsWasmMemoryReference(rmode_)); |
| 313 DCHECK_GE(wasm_memory_reference(), old_base); | 318 DCHECK_GE(wasm_memory_reference(), old_base); |
| 314 Address updated_reference = new_base + (wasm_memory_reference() - old_base); | 319 Address updated_reference = new_base + (wasm_memory_reference() - old_base); |
| 315 // The reference is not checked here but at runtime. Validity of references | 320 // The reference is not checked here but at runtime. Validity of references |
| 316 // may change over time. | 321 // may change over time. |
| 317 unchecked_update_wasm_memory_reference(updated_reference, icache_flush_mode); | 322 unchecked_update_wasm_memory_reference(isolate, updated_reference, |
| 323 icache_flush_mode); |
| 318 } | 324 } |
| 319 | 325 |
| 320 void RelocInfo::update_wasm_memory_size(uint32_t old_size, uint32_t new_size, | 326 void RelocInfo::update_wasm_memory_size(Isolate* isolate, uint32_t old_size, |
| 327 uint32_t new_size, |
| 321 ICacheFlushMode icache_flush_mode) { | 328 ICacheFlushMode icache_flush_mode) { |
| 322 DCHECK(IsWasmMemorySizeReference(rmode_)); | 329 DCHECK(IsWasmMemorySizeReference(rmode_)); |
| 323 uint32_t current_size_reference = wasm_memory_size_reference(); | 330 uint32_t current_size_reference = wasm_memory_size_reference(); |
| 324 uint32_t updated_size_reference = | 331 uint32_t updated_size_reference = |
| 325 new_size + (current_size_reference - old_size); | 332 new_size + (current_size_reference - old_size); |
| 326 unchecked_update_wasm_size(updated_size_reference, icache_flush_mode); | 333 unchecked_update_wasm_size(isolate, updated_size_reference, |
| 334 icache_flush_mode); |
| 327 } | 335 } |
| 328 | 336 |
| 329 void RelocInfo::update_wasm_global_reference( | 337 void RelocInfo::update_wasm_global_reference( |
| 330 Address old_base, Address new_base, ICacheFlushMode icache_flush_mode) { | 338 Isolate* isolate, Address old_base, Address new_base, |
| 339 ICacheFlushMode icache_flush_mode) { |
| 331 DCHECK(IsWasmGlobalReference(rmode_)); | 340 DCHECK(IsWasmGlobalReference(rmode_)); |
| 332 Address updated_reference; | 341 Address updated_reference; |
| 333 DCHECK_LE(old_base, wasm_global_reference()); | 342 DCHECK_LE(old_base, wasm_global_reference()); |
| 334 updated_reference = new_base + (wasm_global_reference() - old_base); | 343 updated_reference = new_base + (wasm_global_reference() - old_base); |
| 335 DCHECK_LE(new_base, updated_reference); | 344 DCHECK_LE(new_base, updated_reference); |
| 336 unchecked_update_wasm_memory_reference(updated_reference, icache_flush_mode); | 345 unchecked_update_wasm_memory_reference(isolate, updated_reference, |
| 346 icache_flush_mode); |
| 337 } | 347 } |
| 338 | 348 |
| 339 void RelocInfo::update_wasm_function_table_size_reference( | 349 void RelocInfo::update_wasm_function_table_size_reference( |
| 340 uint32_t old_size, uint32_t new_size, ICacheFlushMode icache_flush_mode) { | 350 Isolate* isolate, uint32_t old_size, uint32_t new_size, |
| 351 ICacheFlushMode icache_flush_mode) { |
| 341 DCHECK(IsWasmFunctionTableSizeReference(rmode_)); | 352 DCHECK(IsWasmFunctionTableSizeReference(rmode_)); |
| 342 uint32_t current_size_reference = wasm_function_table_size_reference(); | 353 uint32_t current_size_reference = wasm_function_table_size_reference(); |
| 343 uint32_t updated_size_reference = | 354 uint32_t updated_size_reference = |
| 344 new_size + (current_size_reference - old_size); | 355 new_size + (current_size_reference - old_size); |
| 345 unchecked_update_wasm_size(updated_size_reference, icache_flush_mode); | 356 unchecked_update_wasm_size(isolate, updated_size_reference, |
| 357 icache_flush_mode); |
| 346 } | 358 } |
| 347 | 359 |
| 348 void RelocInfo::set_target_address(Address target, | 360 void RelocInfo::set_target_address(Isolate* isolate, Address target, |
| 349 WriteBarrierMode write_barrier_mode, | 361 WriteBarrierMode write_barrier_mode, |
| 350 ICacheFlushMode icache_flush_mode) { | 362 ICacheFlushMode icache_flush_mode) { |
| 351 DCHECK(IsCodeTarget(rmode_) || IsRuntimeEntry(rmode_)); | 363 DCHECK(IsCodeTarget(rmode_) || IsRuntimeEntry(rmode_)); |
| 352 Assembler::set_target_address_at(isolate_, pc_, host_, target, | 364 Assembler::set_target_address_at(isolate, pc_, host_, target, |
| 353 icache_flush_mode); | 365 icache_flush_mode); |
| 354 if (write_barrier_mode == UPDATE_WRITE_BARRIER && host() != NULL && | 366 if (write_barrier_mode == UPDATE_WRITE_BARRIER && host() != NULL && |
| 355 IsCodeTarget(rmode_)) { | 367 IsCodeTarget(rmode_)) { |
| 356 Object* target_code = Code::GetCodeFromTargetAddress(target); | 368 Code* target_code = Code::GetCodeFromTargetAddress(target); |
| 357 host()->GetHeap()->incremental_marking()->RecordWriteIntoCode( | 369 host()->GetHeap()->incremental_marking()->RecordWriteIntoCode(host(), this, |
| 358 host(), this, HeapObject::cast(target_code)); | 370 target_code); |
| 359 } | 371 } |
| 360 } | 372 } |
| 361 | 373 |
| 362 uint32_t RelocInfoWriter::WriteLongPCJump(uint32_t pc_delta) { | 374 uint32_t RelocInfoWriter::WriteLongPCJump(uint32_t pc_delta) { |
| 363 // Return if the pc_delta can fit in kSmallPCDeltaBits bits. | 375 // Return if the pc_delta can fit in kSmallPCDeltaBits bits. |
| 364 // Otherwise write a variable length PC jump for the bits that do | 376 // Otherwise write a variable length PC jump for the bits that do |
| 365 // not fit in the kSmallPCDeltaBits bits. | 377 // not fit in the kSmallPCDeltaBits bits. |
| 366 if (is_uintn(pc_delta, kSmallPCDeltaBits)) return pc_delta; | 378 if (is_uintn(pc_delta, kSmallPCDeltaBits)) return pc_delta; |
| 367 WriteMode(RelocInfo::PC_JUMP); | 379 WriteMode(RelocInfo::PC_JUMP); |
| 368 uint32_t pc_jump = pc_delta >> kSmallPCDeltaBits; | 380 uint32_t pc_jump = pc_delta >> kSmallPCDeltaBits; |
| (...skipping 262 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 631 code_age_sequence_ = NULL; | 643 code_age_sequence_ = NULL; |
| 632 if (SetMode(RelocInfo::CODE_AGE_SEQUENCE)) { | 644 if (SetMode(RelocInfo::CODE_AGE_SEQUENCE)) { |
| 633 rinfo_.data_ = 0; | 645 rinfo_.data_ = 0; |
| 634 rinfo_.pc_ = old_code_age_sequence; | 646 rinfo_.pc_ = old_code_age_sequence; |
| 635 return; | 647 return; |
| 636 } | 648 } |
| 637 } | 649 } |
| 638 done_ = true; | 650 done_ = true; |
| 639 } | 651 } |
| 640 | 652 |
| 641 | 653 RelocIterator::RelocIterator(Code* code, int mode_mask) { |
| 642 RelocIterator::RelocIterator(Code* code, int mode_mask) | |
| 643 : rinfo_(code->map()->GetIsolate()) { | |
| 644 rinfo_.host_ = code; | 654 rinfo_.host_ = code; |
| 645 rinfo_.pc_ = code->instruction_start(); | 655 rinfo_.pc_ = code->instruction_start(); |
| 646 rinfo_.data_ = 0; | 656 rinfo_.data_ = 0; |
| 647 // Relocation info is read backwards. | 657 // Relocation info is read backwards. |
| 648 pos_ = code->relocation_start() + code->relocation_size(); | 658 pos_ = code->relocation_start() + code->relocation_size(); |
| 649 end_ = code->relocation_start(); | 659 end_ = code->relocation_start(); |
| 650 done_ = false; | 660 done_ = false; |
| 651 mode_mask_ = mode_mask; | 661 mode_mask_ = mode_mask; |
| 652 last_id_ = 0; | 662 last_id_ = 0; |
| 653 byte* sequence = code->FindCodeAgeSequence(); | 663 byte* sequence = code->FindCodeAgeSequence(); |
| 654 // We get the isolate from the map, because at serialization time | 664 // We get the isolate from the map, because at serialization time |
| 655 // the code pointer has been cloned and isn't really in heap space. | 665 // the code pointer has been cloned and isn't really in heap space. |
| 656 Isolate* isolate = code->map()->GetIsolate(); | 666 Isolate* isolate = code->map()->GetIsolate(); |
| 657 if (sequence != NULL && !Code::IsYoungSequence(isolate, sequence)) { | 667 if (sequence != NULL && !Code::IsYoungSequence(isolate, sequence)) { |
| 658 code_age_sequence_ = sequence; | 668 code_age_sequence_ = sequence; |
| 659 } else { | 669 } else { |
| 660 code_age_sequence_ = NULL; | 670 code_age_sequence_ = NULL; |
| 661 } | 671 } |
| 662 if (mode_mask_ == 0) pos_ = end_; | 672 if (mode_mask_ == 0) pos_ = end_; |
| 663 next(); | 673 next(); |
| 664 } | 674 } |
| 665 | 675 |
| 666 | 676 RelocIterator::RelocIterator(const CodeDesc& desc, int mode_mask) { |
| 667 RelocIterator::RelocIterator(const CodeDesc& desc, int mode_mask) | |
| 668 : rinfo_(desc.origin->isolate()) { | |
| 669 rinfo_.pc_ = desc.buffer; | 677 rinfo_.pc_ = desc.buffer; |
| 670 rinfo_.data_ = 0; | 678 rinfo_.data_ = 0; |
| 671 // Relocation info is read backwards. | 679 // Relocation info is read backwards. |
| 672 pos_ = desc.buffer + desc.buffer_size; | 680 pos_ = desc.buffer + desc.buffer_size; |
| 673 end_ = pos_ - desc.reloc_size; | 681 end_ = pos_ - desc.reloc_size; |
| 674 done_ = false; | 682 done_ = false; |
| 675 mode_mask_ = mode_mask; | 683 mode_mask_ = mode_mask; |
| 676 last_id_ = 0; | 684 last_id_ = 0; |
| 677 code_age_sequence_ = NULL; | 685 code_age_sequence_ = NULL; |
| 678 if (mode_mask_ == 0) pos_ = end_; | 686 if (mode_mask_ == 0) pos_ = end_; |
| 679 next(); | 687 next(); |
| 680 } | 688 } |
| 681 | 689 |
| 682 | 690 |
| 683 // ----------------------------------------------------------------------------- | 691 // ----------------------------------------------------------------------------- |
| 684 // Implementation of RelocInfo | 692 // Implementation of RelocInfo |
| 685 | 693 |
| 686 bool RelocInfo::IsPatchedDebugBreakSlotSequence() { | 694 bool RelocInfo::IsPatchedDebugBreakSlotSequence() { |
| 687 return DebugCodegen::DebugBreakSlotIsPatched(pc_); | 695 return DebugCodegen::DebugBreakSlotIsPatched(pc_); |
| 688 } | 696 } |
| 689 | 697 |
| 690 #ifdef DEBUG | 698 #ifdef DEBUG |
| 691 bool RelocInfo::RequiresRelocation(const CodeDesc& desc) { | 699 bool RelocInfo::RequiresRelocation(Isolate* isolate, const CodeDesc& desc) { |
| 692 // Ensure there are no code targets or embedded objects present in the | 700 // Ensure there are no code targets or embedded objects present in the |
| 693 // deoptimization entries, they would require relocation after code | 701 // deoptimization entries, they would require relocation after code |
| 694 // generation. | 702 // generation. |
| 695 int mode_mask = RelocInfo::kCodeTargetMask | | 703 int mode_mask = RelocInfo::kCodeTargetMask | |
| 696 RelocInfo::ModeMask(RelocInfo::EMBEDDED_OBJECT) | | 704 RelocInfo::ModeMask(RelocInfo::EMBEDDED_OBJECT) | |
| 697 RelocInfo::ModeMask(RelocInfo::CELL) | | 705 RelocInfo::ModeMask(RelocInfo::CELL) | |
| 698 RelocInfo::kApplyMask; | 706 RelocInfo::kApplyMask; |
| 699 RelocIterator it(desc, mode_mask); | 707 RelocIterator it(desc, mode_mask); |
| 700 return !it.done(); | 708 return !it.done(); |
| 701 } | 709 } |
| (...skipping 1187 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 1889 } | 1897 } |
| 1890 | 1898 |
| 1891 return !empty ? emitted_label_.pos() : 0; | 1899 return !empty ? emitted_label_.pos() : 0; |
| 1892 } | 1900 } |
| 1893 | 1901 |
| 1894 | 1902 |
| 1895 // Platform specific but identical code for all the platforms. | 1903 // Platform specific but identical code for all the platforms. |
| 1896 | 1904 |
| 1897 void Assembler::RecordDeoptReason(DeoptimizeReason reason, | 1905 void Assembler::RecordDeoptReason(DeoptimizeReason reason, |
| 1898 SourcePosition position, int id) { | 1906 SourcePosition position, int id) { |
| 1899 if (FLAG_trace_deopt || isolate()->is_profiling()) { | 1907 EnsureSpace ensure_space(this); |
| 1900 EnsureSpace ensure_space(this); | 1908 RecordRelocInfo(RelocInfo::DEOPT_SCRIPT_OFFSET, position.ScriptOffset()); |
| 1901 RecordRelocInfo(RelocInfo::DEOPT_SCRIPT_OFFSET, position.ScriptOffset()); | 1909 RecordRelocInfo(RelocInfo::DEOPT_INLINING_ID, position.InliningId()); |
| 1902 RecordRelocInfo(RelocInfo::DEOPT_INLINING_ID, position.InliningId()); | 1910 RecordRelocInfo(RelocInfo::DEOPT_REASON, static_cast<int>(reason)); |
| 1903 RecordRelocInfo(RelocInfo::DEOPT_REASON, static_cast<int>(reason)); | 1911 RecordRelocInfo(RelocInfo::DEOPT_ID, id); |
| 1904 RecordRelocInfo(RelocInfo::DEOPT_ID, id); | |
| 1905 } | |
| 1906 } | 1912 } |
| 1907 | 1913 |
| 1908 | 1914 |
| 1909 void Assembler::RecordComment(const char* msg) { | 1915 void Assembler::RecordComment(const char* msg) { |
| 1910 if (FLAG_code_comments) { | 1916 if (FLAG_code_comments) { |
| 1911 EnsureSpace ensure_space(this); | 1917 EnsureSpace ensure_space(this); |
| 1912 RecordRelocInfo(RelocInfo::COMMENT, reinterpret_cast<intptr_t>(msg)); | 1918 RecordRelocInfo(RelocInfo::COMMENT, reinterpret_cast<intptr_t>(msg)); |
| 1913 } | 1919 } |
| 1914 } | 1920 } |
| 1915 | 1921 |
| 1916 | 1922 |
| 1917 void Assembler::RecordDebugBreakSlot(RelocInfo::Mode mode) { | 1923 void Assembler::RecordDebugBreakSlot(RelocInfo::Mode mode) { |
| 1918 EnsureSpace ensure_space(this); | 1924 EnsureSpace ensure_space(this); |
| 1919 DCHECK(RelocInfo::IsDebugBreakSlot(mode)); | 1925 DCHECK(RelocInfo::IsDebugBreakSlot(mode)); |
| 1920 RecordRelocInfo(mode); | 1926 RecordRelocInfo(mode); |
| 1921 } | 1927 } |
| 1922 | 1928 |
| 1923 | 1929 |
| 1924 void Assembler::DataAlign(int m) { | 1930 void Assembler::DataAlign(int m) { |
| 1925 DCHECK(m >= 2 && base::bits::IsPowerOfTwo32(m)); | 1931 DCHECK(m >= 2 && base::bits::IsPowerOfTwo32(m)); |
| 1926 while ((pc_offset() & (m - 1)) != 0) { | 1932 while ((pc_offset() & (m - 1)) != 0) { |
| 1927 db(0); | 1933 db(0); |
| 1928 } | 1934 } |
| 1929 } | 1935 } |
| 1930 } // namespace internal | 1936 } // namespace internal |
| 1931 } // namespace v8 | 1937 } // namespace v8 |
| OLD | NEW |