| OLD | NEW | 
|    1 // Copyright 2013 the V8 project authors. All rights reserved. |    1 // Copyright 2013 the V8 project authors. All rights reserved. | 
|    2 // Use of this source code is governed by a BSD-style license that can be |    2 // Use of this source code is governed by a BSD-style license that can be | 
|    3 // found in the LICENSE file. |    3 // found in the LICENSE file. | 
|    4  |    4  | 
|    5 #if V8_TARGET_ARCH_ARM64 |    5 #if V8_TARGET_ARCH_ARM64 | 
|    6  |    6  | 
|    7 #include "src/codegen.h" |    7 #include "src/codegen.h" | 
|    8 #include "src/ic/ic.h" |    8 #include "src/ic/ic.h" | 
|    9 #include "src/ic/ic-compiler.h" |    9 #include "src/ic/ic-compiler.h" | 
|   10 #include "src/ic/stub-cache.h" |   10 #include "src/ic/stub-cache.h" | 
| (...skipping 262 matching lines...) Expand 10 before | Expand all | Expand 10 after  Loading... | 
|  273   GenerateDictionaryLoad(masm, &slow, dictionary, |  273   GenerateDictionaryLoad(masm, &slow, dictionary, | 
|  274                          LoadDescriptor::NameRegister(), x0, x3, x4); |  274                          LoadDescriptor::NameRegister(), x0, x3, x4); | 
|  275   __ Ret(); |  275   __ Ret(); | 
|  276  |  276  | 
|  277   // Dictionary load failed, go slow (but don't miss). |  277   // Dictionary load failed, go slow (but don't miss). | 
|  278   __ Bind(&slow); |  278   __ Bind(&slow); | 
|  279   GenerateRuntimeGetProperty(masm, language_mode); |  279   GenerateRuntimeGetProperty(masm, language_mode); | 
|  280 } |  280 } | 
|  281  |  281  | 
|  282  |  282  | 
|  283 void LoadIC::GenerateMiss(MacroAssembler* masm, int stress) { |  283 void LoadIC::GenerateMiss(MacroAssembler* masm) { | 
|  284   // The return address is in lr. |  284   // The return address is in lr. | 
|  285   Isolate* isolate = masm->isolate(); |  285   Isolate* isolate = masm->isolate(); | 
|  286   ASM_LOCATION("LoadIC::GenerateMiss"); |  286   ASM_LOCATION("LoadIC::GenerateMiss"); | 
|  287  |  287  | 
|  288   DCHECK(!AreAliased(x4, x5, LoadWithVectorDescriptor::SlotRegister(), |  288   DCHECK(!AreAliased(x4, x5, LoadWithVectorDescriptor::SlotRegister(), | 
|  289                      LoadWithVectorDescriptor::VectorRegister())); |  289                      LoadWithVectorDescriptor::VectorRegister())); | 
|  290   __ IncrementCounter(isolate->counters()->load_miss(), 1, x4, x5); |  290   __ IncrementCounter(isolate->counters()->load_miss(), 1, x4, x5); | 
|  291  |  291  | 
|  292   // Perform tail call to the entry. |  292   // Perform tail call to the entry. | 
|  293   __ Push(LoadWithVectorDescriptor::ReceiverRegister(), |  293   __ Push(LoadWithVectorDescriptor::ReceiverRegister(), | 
| (...skipping 596 matching lines...) Expand 10 before | Expand all | Expand 10 after  Loading... | 
|  890   } else { |  890   } else { | 
|  891     DCHECK(to_patch->Mask(TestBranchMask) == TBNZ); |  891     DCHECK(to_patch->Mask(TestBranchMask) == TBNZ); | 
|  892     // This is JumpIfSmi(smi_reg, branch_imm). |  892     // This is JumpIfSmi(smi_reg, branch_imm). | 
|  893     patcher.tbz(smi_reg, 0, branch_imm); |  893     patcher.tbz(smi_reg, 0, branch_imm); | 
|  894   } |  894   } | 
|  895 } |  895 } | 
|  896 }  // namespace internal |  896 }  // namespace internal | 
|  897 }  // namespace v8 |  897 }  // namespace v8 | 
|  898  |  898  | 
|  899 #endif  // V8_TARGET_ARCH_ARM64 |  899 #endif  // V8_TARGET_ARCH_ARM64 | 
| OLD | NEW |