OLD | NEW |
1 // Copyright (c) 2014, the Dart project authors. Please see the AUTHORS file | 1 // Copyright (c) 2014, the Dart project authors. Please see the AUTHORS file |
2 // for details. All rights reserved. Use of this source code is governed by a | 2 // for details. All rights reserved. Use of this source code is governed by a |
3 // BSD-style license that can be found in the LICENSE file. | 3 // BSD-style license that can be found in the LICENSE file. |
4 | 4 |
5 #include "vm/globals.h" | 5 #include "vm/globals.h" |
6 #if defined(TARGET_ARCH_ARM64) | 6 #if defined(TARGET_ARCH_ARM64) |
7 | 7 |
8 #include "vm/assembler.h" | 8 #include "vm/assembler.h" |
9 #include "vm/code_generator.h" | 9 #include "vm/code_generator.h" |
10 #include "vm/compiler.h" | 10 #include "vm/compiler.h" |
(...skipping 34 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
45 | 45 |
46 __ SetPrologueOffset(); | 46 __ SetPrologueOffset(); |
47 __ Comment("CallToRuntimeStub"); | 47 __ Comment("CallToRuntimeStub"); |
48 __ EnterStubFrame(); | 48 __ EnterStubFrame(); |
49 | 49 |
50 COMPILE_ASSERT((kAbiPreservedCpuRegs & (1 << R28)) != 0); | 50 COMPILE_ASSERT((kAbiPreservedCpuRegs & (1 << R28)) != 0); |
51 __ LoadIsolate(R28); | 51 __ LoadIsolate(R28); |
52 | 52 |
53 // Save exit frame information to enable stack walking as we are about | 53 // Save exit frame information to enable stack walking as we are about |
54 // to transition to Dart VM C++ code. | 54 // to transition to Dart VM C++ code. |
55 __ StoreToOffset(FP, THR, Thread::top_exit_frame_info_offset(), kNoPP); | 55 __ StoreToOffset(FP, THR, Thread::top_exit_frame_info_offset()); |
56 | 56 |
57 #if defined(DEBUG) | 57 #if defined(DEBUG) |
58 { Label ok; | 58 { Label ok; |
59 // Check that we are always entering from Dart code. | 59 // Check that we are always entering from Dart code. |
60 __ LoadFromOffset(R8, R28, Isolate::vm_tag_offset(), kNoPP); | 60 __ LoadFromOffset(R8, R28, Isolate::vm_tag_offset()); |
61 __ CompareImmediate(R8, VMTag::kDartTagId, kNoPP); | 61 __ CompareImmediate(R8, VMTag::kDartTagId); |
62 __ b(&ok, EQ); | 62 __ b(&ok, EQ); |
63 __ Stop("Not coming from Dart code."); | 63 __ Stop("Not coming from Dart code."); |
64 __ Bind(&ok); | 64 __ Bind(&ok); |
65 } | 65 } |
66 #endif | 66 #endif |
67 | 67 |
68 // Mark that the isolate is executing VM code. | 68 // Mark that the isolate is executing VM code. |
69 __ StoreToOffset(R5, R28, Isolate::vm_tag_offset(), kNoPP); | 69 __ StoreToOffset(R5, R28, Isolate::vm_tag_offset()); |
70 | 70 |
71 // Reserve space for arguments and align frame before entering C++ world. | 71 // Reserve space for arguments and align frame before entering C++ world. |
72 // NativeArguments are passed in registers. | 72 // NativeArguments are passed in registers. |
73 __ Comment("align stack"); | 73 __ Comment("align stack"); |
74 // Reserve space for arguments. | 74 // Reserve space for arguments. |
75 ASSERT(sizeof(NativeArguments) == 4 * kWordSize); | 75 ASSERT(sizeof(NativeArguments) == 4 * kWordSize); |
76 __ ReserveAlignedFrameSpace(sizeof(NativeArguments)); | 76 __ ReserveAlignedFrameSpace(sizeof(NativeArguments)); |
77 | 77 |
78 // Pass NativeArguments structure by value and call runtime. | 78 // Pass NativeArguments structure by value and call runtime. |
79 // Registers R0, R1, R2, and R3 are used. | 79 // Registers R0, R1, R2, and R3 are used. |
80 | 80 |
81 ASSERT(thread_offset == 0 * kWordSize); | 81 ASSERT(thread_offset == 0 * kWordSize); |
82 // Set thread in NativeArgs. | 82 // Set thread in NativeArgs. |
83 __ mov(R0, THR); | 83 __ mov(R0, THR); |
84 | 84 |
85 // There are no runtime calls to closures, so we do not need to set the tag | 85 // There are no runtime calls to closures, so we do not need to set the tag |
86 // bits kClosureFunctionBit and kInstanceFunctionBit in argc_tag_. | 86 // bits kClosureFunctionBit and kInstanceFunctionBit in argc_tag_. |
87 ASSERT(argc_tag_offset == 1 * kWordSize); | 87 ASSERT(argc_tag_offset == 1 * kWordSize); |
88 __ mov(R1, R4); // Set argc in NativeArguments. | 88 __ mov(R1, R4); // Set argc in NativeArguments. |
89 | 89 |
90 ASSERT(argv_offset == 2 * kWordSize); | 90 ASSERT(argv_offset == 2 * kWordSize); |
91 __ add(R2, ZR, Operand(R4, LSL, 3)); | 91 __ add(R2, ZR, Operand(R4, LSL, 3)); |
92 __ add(R2, FP, Operand(R2)); // Compute argv. | 92 __ add(R2, FP, Operand(R2)); // Compute argv. |
93 // Set argv in NativeArguments. | 93 // Set argv in NativeArguments. |
94 __ AddImmediate(R2, R2, exitframe_last_param_slot_from_fp * kWordSize, kNoPP); | 94 __ AddImmediate(R2, R2, exitframe_last_param_slot_from_fp * kWordSize); |
95 | 95 |
96 ASSERT(retval_offset == 3 * kWordSize); | 96 ASSERT(retval_offset == 3 * kWordSize); |
97 __ AddImmediate(R3, R2, kWordSize, kNoPP); | 97 __ AddImmediate(R3, R2, kWordSize); |
98 | 98 |
99 __ StoreToOffset(R0, SP, thread_offset, kNoPP); | 99 __ StoreToOffset(R0, SP, thread_offset); |
100 __ StoreToOffset(R1, SP, argc_tag_offset, kNoPP); | 100 __ StoreToOffset(R1, SP, argc_tag_offset); |
101 __ StoreToOffset(R2, SP, argv_offset, kNoPP); | 101 __ StoreToOffset(R2, SP, argv_offset); |
102 __ StoreToOffset(R3, SP, retval_offset, kNoPP); | 102 __ StoreToOffset(R3, SP, retval_offset); |
103 __ mov(R0, SP); // Pass the pointer to the NativeArguments. | 103 __ mov(R0, SP); // Pass the pointer to the NativeArguments. |
104 | 104 |
105 // We are entering runtime code, so the C stack pointer must be restored from | 105 // We are entering runtime code, so the C stack pointer must be restored from |
106 // the stack limit to the top of the stack. We cache the stack limit address | 106 // the stack limit to the top of the stack. We cache the stack limit address |
107 // in a callee-saved register. | 107 // in a callee-saved register. |
108 __ mov(R26, CSP); | 108 __ mov(R26, CSP); |
109 __ mov(CSP, SP); | 109 __ mov(CSP, SP); |
110 | 110 |
111 __ blr(R5); | 111 __ blr(R5); |
112 __ Comment("CallToRuntimeStub return"); | 112 __ Comment("CallToRuntimeStub return"); |
113 | 113 |
114 // Restore SP and CSP. | 114 // Restore SP and CSP. |
115 __ mov(SP, CSP); | 115 __ mov(SP, CSP); |
116 __ mov(CSP, R26); | 116 __ mov(CSP, R26); |
117 | 117 |
118 // Retval is next to 1st argument. | 118 // Retval is next to 1st argument. |
119 // Mark that the isolate is executing Dart code. | 119 // Mark that the isolate is executing Dart code. |
120 __ LoadImmediate(R2, VMTag::kDartTagId, kNoPP); | 120 __ LoadImmediate(R2, VMTag::kDartTagId); |
121 __ StoreToOffset(R2, R28, Isolate::vm_tag_offset(), kNoPP); | 121 __ StoreToOffset(R2, R28, Isolate::vm_tag_offset()); |
122 | 122 |
123 // Reset exit frame information in Isolate structure. | 123 // Reset exit frame information in Isolate structure. |
124 __ StoreToOffset(ZR, THR, Thread::top_exit_frame_info_offset(), kNoPP); | 124 __ StoreToOffset(ZR, THR, Thread::top_exit_frame_info_offset()); |
125 | 125 |
126 __ LeaveStubFrame(); | 126 __ LeaveStubFrame(); |
127 __ ret(); | 127 __ ret(); |
128 } | 128 } |
129 | 129 |
130 | 130 |
131 void StubCode::GeneratePrintStopMessageStub(Assembler* assembler) { | 131 void StubCode::GeneratePrintStopMessageStub(Assembler* assembler) { |
132 __ Stop("GeneratePrintStopMessageStub"); | 132 __ Stop("GeneratePrintStopMessageStub"); |
133 } | 133 } |
134 | 134 |
(...skipping 10 matching lines...) Expand all Loading... |
145 const intptr_t argv_offset = NativeArguments::argv_offset(); | 145 const intptr_t argv_offset = NativeArguments::argv_offset(); |
146 const intptr_t retval_offset = NativeArguments::retval_offset(); | 146 const intptr_t retval_offset = NativeArguments::retval_offset(); |
147 | 147 |
148 __ EnterStubFrame(); | 148 __ EnterStubFrame(); |
149 | 149 |
150 COMPILE_ASSERT((kAbiPreservedCpuRegs & (1 << R28)) != 0); | 150 COMPILE_ASSERT((kAbiPreservedCpuRegs & (1 << R28)) != 0); |
151 __ LoadIsolate(R28); | 151 __ LoadIsolate(R28); |
152 | 152 |
153 // Save exit frame information to enable stack walking as we are about | 153 // Save exit frame information to enable stack walking as we are about |
154 // to transition to native code. | 154 // to transition to native code. |
155 __ StoreToOffset(FP, THR, Thread::top_exit_frame_info_offset(), kNoPP); | 155 __ StoreToOffset(FP, THR, Thread::top_exit_frame_info_offset()); |
156 | 156 |
157 #if defined(DEBUG) | 157 #if defined(DEBUG) |
158 { Label ok; | 158 { Label ok; |
159 // Check that we are always entering from Dart code. | 159 // Check that we are always entering from Dart code. |
160 __ LoadFromOffset(R6, R28, Isolate::vm_tag_offset(), kNoPP); | 160 __ LoadFromOffset(R6, R28, Isolate::vm_tag_offset()); |
161 __ CompareImmediate(R6, VMTag::kDartTagId, kNoPP); | 161 __ CompareImmediate(R6, VMTag::kDartTagId); |
162 __ b(&ok, EQ); | 162 __ b(&ok, EQ); |
163 __ Stop("Not coming from Dart code."); | 163 __ Stop("Not coming from Dart code."); |
164 __ Bind(&ok); | 164 __ Bind(&ok); |
165 } | 165 } |
166 #endif | 166 #endif |
167 | 167 |
168 // Mark that the isolate is executing Native code. | 168 // Mark that the isolate is executing Native code. |
169 __ StoreToOffset(R5, R28, Isolate::vm_tag_offset(), kNoPP); | 169 __ StoreToOffset(R5, R28, Isolate::vm_tag_offset()); |
170 | 170 |
171 // Reserve space for the native arguments structure passed on the stack (the | 171 // Reserve space for the native arguments structure passed on the stack (the |
172 // outgoing pointer parameter to the native arguments structure is passed in | 172 // outgoing pointer parameter to the native arguments structure is passed in |
173 // R0) and align frame before entering the C++ world. | 173 // R0) and align frame before entering the C++ world. |
174 __ ReserveAlignedFrameSpace(sizeof(NativeArguments)); | 174 __ ReserveAlignedFrameSpace(sizeof(NativeArguments)); |
175 | 175 |
176 // Initialize NativeArguments structure and call native function. | 176 // Initialize NativeArguments structure and call native function. |
177 // Registers R0, R1, R2, and R3 are used. | 177 // Registers R0, R1, R2, and R3 are used. |
178 | 178 |
179 ASSERT(thread_offset == 0 * kWordSize); | 179 ASSERT(thread_offset == 0 * kWordSize); |
180 // Set thread in NativeArgs. | 180 // Set thread in NativeArgs. |
181 __ mov(R0, THR); | 181 __ mov(R0, THR); |
182 | 182 |
183 // There are no native calls to closures, so we do not need to set the tag | 183 // There are no native calls to closures, so we do not need to set the tag |
184 // bits kClosureFunctionBit and kInstanceFunctionBit in argc_tag_. | 184 // bits kClosureFunctionBit and kInstanceFunctionBit in argc_tag_. |
185 ASSERT(argc_tag_offset == 1 * kWordSize); | 185 ASSERT(argc_tag_offset == 1 * kWordSize); |
186 // Set argc in NativeArguments: R1 already contains argc. | 186 // Set argc in NativeArguments: R1 already contains argc. |
187 | 187 |
188 ASSERT(argv_offset == 2 * kWordSize); | 188 ASSERT(argv_offset == 2 * kWordSize); |
189 // Set argv in NativeArguments: R2 already contains argv. | 189 // Set argv in NativeArguments: R2 already contains argv. |
190 | 190 |
191 // Set retval in NativeArgs. | 191 // Set retval in NativeArgs. |
192 ASSERT(retval_offset == 3 * kWordSize); | 192 ASSERT(retval_offset == 3 * kWordSize); |
193 __ AddImmediate(R3, FP, 2 * kWordSize, kNoPP); | 193 __ AddImmediate(R3, FP, 2 * kWordSize); |
194 | 194 |
195 // Passing the structure by value as in runtime calls would require changing | 195 // Passing the structure by value as in runtime calls would require changing |
196 // Dart API for native functions. | 196 // Dart API for native functions. |
197 // For now, space is reserved on the stack and we pass a pointer to it. | 197 // For now, space is reserved on the stack and we pass a pointer to it. |
198 __ StoreToOffset(R0, SP, thread_offset, kNoPP); | 198 __ StoreToOffset(R0, SP, thread_offset); |
199 __ StoreToOffset(R1, SP, argc_tag_offset, kNoPP); | 199 __ StoreToOffset(R1, SP, argc_tag_offset); |
200 __ StoreToOffset(R2, SP, argv_offset, kNoPP); | 200 __ StoreToOffset(R2, SP, argv_offset); |
201 __ StoreToOffset(R3, SP, retval_offset, kNoPP); | 201 __ StoreToOffset(R3, SP, retval_offset); |
202 __ mov(R0, SP); // Pass the pointer to the NativeArguments. | 202 __ mov(R0, SP); // Pass the pointer to the NativeArguments. |
203 | 203 |
204 // We are entering runtime code, so the C stack pointer must be restored from | 204 // We are entering runtime code, so the C stack pointer must be restored from |
205 // the stack limit to the top of the stack. We cache the stack limit address | 205 // the stack limit to the top of the stack. We cache the stack limit address |
206 // in the Dart SP register, which is callee-saved in the C ABI. | 206 // in the Dart SP register, which is callee-saved in the C ABI. |
207 __ mov(R26, CSP); | 207 __ mov(R26, CSP); |
208 __ mov(CSP, SP); | 208 __ mov(CSP, SP); |
209 | 209 |
210 __ mov(R1, R5); // Pass the function entrypoint to call. | 210 __ mov(R1, R5); // Pass the function entrypoint to call. |
211 // Call native function invocation wrapper or redirection via simulator. | 211 // Call native function invocation wrapper or redirection via simulator. |
212 #if defined(USING_SIMULATOR) | 212 #if defined(USING_SIMULATOR) |
213 uword entry = reinterpret_cast<uword>(NativeEntry::NativeCallWrapper); | 213 uword entry = reinterpret_cast<uword>(NativeEntry::NativeCallWrapper); |
214 entry = Simulator::RedirectExternalReference( | 214 entry = Simulator::RedirectExternalReference( |
215 entry, Simulator::kNativeCall, NativeEntry::kNumCallWrapperArguments); | 215 entry, Simulator::kNativeCall, NativeEntry::kNumCallWrapperArguments); |
216 __ LoadImmediate(R2, entry, kNoPP); | 216 __ LoadImmediate(R2, entry); |
217 __ blr(R2); | 217 __ blr(R2); |
218 #else | 218 #else |
219 __ BranchLink(&NativeEntry::NativeCallWrapperLabel(), kNoPP); | 219 __ BranchLink(&NativeEntry::NativeCallWrapperLabel()); |
220 #endif | 220 #endif |
221 | 221 |
222 // Restore SP and CSP. | 222 // Restore SP and CSP. |
223 __ mov(SP, CSP); | 223 __ mov(SP, CSP); |
224 __ mov(CSP, R26); | 224 __ mov(CSP, R26); |
225 | 225 |
226 // Mark that the isolate is executing Dart code. | 226 // Mark that the isolate is executing Dart code. |
227 __ LoadImmediate(R2, VMTag::kDartTagId, kNoPP); | 227 __ LoadImmediate(R2, VMTag::kDartTagId); |
228 __ StoreToOffset(R2, R28, Isolate::vm_tag_offset(), kNoPP); | 228 __ StoreToOffset(R2, R28, Isolate::vm_tag_offset()); |
229 | 229 |
230 // Reset exit frame information in Isolate structure. | 230 // Reset exit frame information in Isolate structure. |
231 __ StoreToOffset(ZR, THR, Thread::top_exit_frame_info_offset(), kNoPP); | 231 __ StoreToOffset(ZR, THR, Thread::top_exit_frame_info_offset()); |
232 | 232 |
233 __ LeaveStubFrame(); | 233 __ LeaveStubFrame(); |
234 __ ret(); | 234 __ ret(); |
235 } | 235 } |
236 | 236 |
237 | 237 |
238 // Input parameters: | 238 // Input parameters: |
239 // LR : return address. | 239 // LR : return address. |
240 // SP : address of return value. | 240 // SP : address of return value. |
241 // R5 : address of the native function to call. | 241 // R5 : address of the native function to call. |
242 // R2 : address of first argument in argument array. | 242 // R2 : address of first argument in argument array. |
243 // R1 : argc_tag including number of arguments and function kind. | 243 // R1 : argc_tag including number of arguments and function kind. |
244 void StubCode::GenerateCallBootstrapCFunctionStub(Assembler* assembler) { | 244 void StubCode::GenerateCallBootstrapCFunctionStub(Assembler* assembler) { |
245 const intptr_t thread_offset = NativeArguments::thread_offset(); | 245 const intptr_t thread_offset = NativeArguments::thread_offset(); |
246 const intptr_t argc_tag_offset = NativeArguments::argc_tag_offset(); | 246 const intptr_t argc_tag_offset = NativeArguments::argc_tag_offset(); |
247 const intptr_t argv_offset = NativeArguments::argv_offset(); | 247 const intptr_t argv_offset = NativeArguments::argv_offset(); |
248 const intptr_t retval_offset = NativeArguments::retval_offset(); | 248 const intptr_t retval_offset = NativeArguments::retval_offset(); |
249 | 249 |
250 __ EnterStubFrame(); | 250 __ EnterStubFrame(); |
251 | 251 |
252 COMPILE_ASSERT((kAbiPreservedCpuRegs & (1 << R28)) != 0); | 252 COMPILE_ASSERT((kAbiPreservedCpuRegs & (1 << R28)) != 0); |
253 __ LoadIsolate(R28); | 253 __ LoadIsolate(R28); |
254 | 254 |
255 // Save exit frame information to enable stack walking as we are about | 255 // Save exit frame information to enable stack walking as we are about |
256 // to transition to native code. | 256 // to transition to native code. |
257 __ StoreToOffset(FP, THR, Thread::top_exit_frame_info_offset(), kNoPP); | 257 __ StoreToOffset(FP, THR, Thread::top_exit_frame_info_offset()); |
258 | 258 |
259 #if defined(DEBUG) | 259 #if defined(DEBUG) |
260 { Label ok; | 260 { Label ok; |
261 // Check that we are always entering from Dart code. | 261 // Check that we are always entering from Dart code. |
262 __ LoadFromOffset(R6, R28, Isolate::vm_tag_offset(), kNoPP); | 262 __ LoadFromOffset(R6, R28, Isolate::vm_tag_offset()); |
263 __ CompareImmediate(R6, VMTag::kDartTagId, kNoPP); | 263 __ CompareImmediate(R6, VMTag::kDartTagId); |
264 __ b(&ok, EQ); | 264 __ b(&ok, EQ); |
265 __ Stop("Not coming from Dart code."); | 265 __ Stop("Not coming from Dart code."); |
266 __ Bind(&ok); | 266 __ Bind(&ok); |
267 } | 267 } |
268 #endif | 268 #endif |
269 | 269 |
270 // Mark that the isolate is executing Native code. | 270 // Mark that the isolate is executing Native code. |
271 __ StoreToOffset(R5, R28, Isolate::vm_tag_offset(), kNoPP); | 271 __ StoreToOffset(R5, R28, Isolate::vm_tag_offset()); |
272 | 272 |
273 // Reserve space for the native arguments structure passed on the stack (the | 273 // Reserve space for the native arguments structure passed on the stack (the |
274 // outgoing pointer parameter to the native arguments structure is passed in | 274 // outgoing pointer parameter to the native arguments structure is passed in |
275 // R0) and align frame before entering the C++ world. | 275 // R0) and align frame before entering the C++ world. |
276 __ ReserveAlignedFrameSpace(sizeof(NativeArguments)); | 276 __ ReserveAlignedFrameSpace(sizeof(NativeArguments)); |
277 | 277 |
278 // Initialize NativeArguments structure and call native function. | 278 // Initialize NativeArguments structure and call native function. |
279 // Registers R0, R1, R2, and R3 are used. | 279 // Registers R0, R1, R2, and R3 are used. |
280 | 280 |
281 ASSERT(thread_offset == 0 * kWordSize); | 281 ASSERT(thread_offset == 0 * kWordSize); |
282 // Set thread in NativeArgs. | 282 // Set thread in NativeArgs. |
283 __ mov(R0, THR); | 283 __ mov(R0, THR); |
284 | 284 |
285 // There are no native calls to closures, so we do not need to set the tag | 285 // There are no native calls to closures, so we do not need to set the tag |
286 // bits kClosureFunctionBit and kInstanceFunctionBit in argc_tag_. | 286 // bits kClosureFunctionBit and kInstanceFunctionBit in argc_tag_. |
287 ASSERT(argc_tag_offset == 1 * kWordSize); | 287 ASSERT(argc_tag_offset == 1 * kWordSize); |
288 // Set argc in NativeArguments: R1 already contains argc. | 288 // Set argc in NativeArguments: R1 already contains argc. |
289 | 289 |
290 ASSERT(argv_offset == 2 * kWordSize); | 290 ASSERT(argv_offset == 2 * kWordSize); |
291 // Set argv in NativeArguments: R2 already contains argv. | 291 // Set argv in NativeArguments: R2 already contains argv. |
292 | 292 |
293 // Set retval in NativeArgs. | 293 // Set retval in NativeArgs. |
294 ASSERT(retval_offset == 3 * kWordSize); | 294 ASSERT(retval_offset == 3 * kWordSize); |
295 __ AddImmediate(R3, FP, 2 * kWordSize, kNoPP); | 295 __ AddImmediate(R3, FP, 2 * kWordSize); |
296 | 296 |
297 // Passing the structure by value as in runtime calls would require changing | 297 // Passing the structure by value as in runtime calls would require changing |
298 // Dart API for native functions. | 298 // Dart API for native functions. |
299 // For now, space is reserved on the stack and we pass a pointer to it. | 299 // For now, space is reserved on the stack and we pass a pointer to it. |
300 __ StoreToOffset(R0, SP, thread_offset, kNoPP); | 300 __ StoreToOffset(R0, SP, thread_offset); |
301 __ StoreToOffset(R1, SP, argc_tag_offset, kNoPP); | 301 __ StoreToOffset(R1, SP, argc_tag_offset); |
302 __ StoreToOffset(R2, SP, argv_offset, kNoPP); | 302 __ StoreToOffset(R2, SP, argv_offset); |
303 __ StoreToOffset(R3, SP, retval_offset, kNoPP); | 303 __ StoreToOffset(R3, SP, retval_offset); |
304 __ mov(R0, SP); // Pass the pointer to the NativeArguments. | 304 __ mov(R0, SP); // Pass the pointer to the NativeArguments. |
305 | 305 |
306 // We are entering runtime code, so the C stack pointer must be restored from | 306 // We are entering runtime code, so the C stack pointer must be restored from |
307 // the stack limit to the top of the stack. We cache the stack limit address | 307 // the stack limit to the top of the stack. We cache the stack limit address |
308 // in the Dart SP register, which is callee-saved in the C ABI. | 308 // in the Dart SP register, which is callee-saved in the C ABI. |
309 __ mov(R26, CSP); | 309 __ mov(R26, CSP); |
310 __ mov(CSP, SP); | 310 __ mov(CSP, SP); |
311 | 311 |
312 // Call native function or redirection via simulator. | 312 // Call native function or redirection via simulator. |
313 __ blr(R5); | 313 __ blr(R5); |
314 | 314 |
315 // Restore SP and CSP. | 315 // Restore SP and CSP. |
316 __ mov(SP, CSP); | 316 __ mov(SP, CSP); |
317 __ mov(CSP, R26); | 317 __ mov(CSP, R26); |
318 | 318 |
319 // Mark that the isolate is executing Dart code. | 319 // Mark that the isolate is executing Dart code. |
320 __ LoadImmediate(R2, VMTag::kDartTagId, kNoPP); | 320 __ LoadImmediate(R2, VMTag::kDartTagId); |
321 __ StoreToOffset(R2, R28, Isolate::vm_tag_offset(), kNoPP); | 321 __ StoreToOffset(R2, R28, Isolate::vm_tag_offset()); |
322 | 322 |
323 // Reset exit frame information in Isolate structure. | 323 // Reset exit frame information in Isolate structure. |
324 __ StoreToOffset(ZR, THR, Thread::top_exit_frame_info_offset(), kNoPP); | 324 __ StoreToOffset(ZR, THR, Thread::top_exit_frame_info_offset()); |
325 | 325 |
326 __ LeaveStubFrame(); | 326 __ LeaveStubFrame(); |
327 __ ret(); | 327 __ ret(); |
328 } | 328 } |
329 | 329 |
330 | 330 |
331 // Input parameters: | 331 // Input parameters: |
332 // R4: arguments descriptor array. | 332 // R4: arguments descriptor array. |
333 void StubCode::GenerateCallStaticFunctionStub(Assembler* assembler) { | 333 void StubCode::GenerateCallStaticFunctionStub(Assembler* assembler) { |
334 // Create a stub frame as we are pushing some objects on the stack before | 334 // Create a stub frame as we are pushing some objects on the stack before |
335 // calling into the runtime. | 335 // calling into the runtime. |
336 __ EnterStubFrame(); | 336 __ EnterStubFrame(); |
337 // Setup space on stack for return value and preserve arguments descriptor. | 337 // Setup space on stack for return value and preserve arguments descriptor. |
338 __ Push(R4); | 338 __ Push(R4); |
339 __ PushObject(Object::null_object(), PP); | 339 __ PushObject(Object::null_object()); |
340 __ CallRuntime(kPatchStaticCallRuntimeEntry, 0); | 340 __ CallRuntime(kPatchStaticCallRuntimeEntry, 0); |
341 // Get Code object result and restore arguments descriptor array. | 341 // Get Code object result and restore arguments descriptor array. |
342 __ Pop(R0); | 342 __ Pop(R0); |
343 __ Pop(R4); | 343 __ Pop(R4); |
344 // Remove the stub frame. | 344 // Remove the stub frame. |
345 __ LeaveStubFrame(); | 345 __ LeaveStubFrame(); |
346 // Jump to the dart function. | 346 // Jump to the dart function. |
347 __ LoadFieldFromOffset(R0, R0, Code::instructions_offset(), kNoPP); | 347 __ LoadFieldFromOffset(R0, R0, Code::instructions_offset()); |
348 __ AddImmediate(R0, R0, Instructions::HeaderSize() - kHeapObjectTag, kNoPP); | 348 __ AddImmediate(R0, R0, Instructions::HeaderSize() - kHeapObjectTag); |
349 __ br(R0); | 349 __ br(R0); |
350 } | 350 } |
351 | 351 |
352 | 352 |
353 // Called from a static call only when an invalid code has been entered | 353 // Called from a static call only when an invalid code has been entered |
354 // (invalid because its function was optimized or deoptimized). | 354 // (invalid because its function was optimized or deoptimized). |
355 // R4: arguments descriptor array. | 355 // R4: arguments descriptor array. |
356 void StubCode::GenerateFixCallersTargetStub(Assembler* assembler) { | 356 void StubCode::GenerateFixCallersTargetStub(Assembler* assembler) { |
357 // Create a stub frame as we are pushing some objects on the stack before | 357 // Create a stub frame as we are pushing some objects on the stack before |
358 // calling into the runtime. | 358 // calling into the runtime. |
359 __ EnterStubFrame(); | 359 __ EnterStubFrame(); |
360 // Setup space on stack for return value and preserve arguments descriptor. | 360 // Setup space on stack for return value and preserve arguments descriptor. |
361 __ Push(R4); | 361 __ Push(R4); |
362 __ PushObject(Object::null_object(), PP); | 362 __ PushObject(Object::null_object()); |
363 __ CallRuntime(kFixCallersTargetRuntimeEntry, 0); | 363 __ CallRuntime(kFixCallersTargetRuntimeEntry, 0); |
364 // Get Code object result and restore arguments descriptor array. | 364 // Get Code object result and restore arguments descriptor array. |
365 __ Pop(R0); | 365 __ Pop(R0); |
366 __ Pop(R4); | 366 __ Pop(R4); |
367 // Remove the stub frame. | 367 // Remove the stub frame. |
368 __ LeaveStubFrame(); | 368 __ LeaveStubFrame(); |
369 // Jump to the dart function. | 369 // Jump to the dart function. |
370 __ LoadFieldFromOffset(R0, R0, Code::instructions_offset(), kNoPP); | 370 __ LoadFieldFromOffset(R0, R0, Code::instructions_offset()); |
371 __ AddImmediate(R0, R0, Instructions::HeaderSize() - kHeapObjectTag, kNoPP); | 371 __ AddImmediate(R0, R0, Instructions::HeaderSize() - kHeapObjectTag); |
372 __ br(R0); | 372 __ br(R0); |
373 } | 373 } |
374 | 374 |
375 | 375 |
376 // Called from object allocate instruction when the allocation stub has been | 376 // Called from object allocate instruction when the allocation stub has been |
377 // disabled. | 377 // disabled. |
378 void StubCode::GenerateFixAllocationStubTargetStub(Assembler* assembler) { | 378 void StubCode::GenerateFixAllocationStubTargetStub(Assembler* assembler) { |
379 __ EnterStubFrame(); | 379 __ EnterStubFrame(); |
380 // Setup space on stack for return value. | 380 // Setup space on stack for return value. |
381 __ PushObject(Object::null_object(), PP); | 381 __ PushObject(Object::null_object()); |
382 __ CallRuntime(kFixAllocationStubTargetRuntimeEntry, 0); | 382 __ CallRuntime(kFixAllocationStubTargetRuntimeEntry, 0); |
383 // Get Code object result. | 383 // Get Code object result. |
384 __ Pop(R0); | 384 __ Pop(R0); |
385 // Remove the stub frame. | 385 // Remove the stub frame. |
386 __ LeaveStubFrame(); | 386 __ LeaveStubFrame(); |
387 // Jump to the dart function. | 387 // Jump to the dart function. |
388 __ LoadFieldFromOffset(R0, R0, Code::instructions_offset(), kNoPP); | 388 __ LoadFieldFromOffset(R0, R0, Code::instructions_offset()); |
389 __ AddImmediate(R0, R0, Instructions::HeaderSize() - kHeapObjectTag, kNoPP); | 389 __ AddImmediate(R0, R0, Instructions::HeaderSize() - kHeapObjectTag); |
390 __ br(R0); | 390 __ br(R0); |
391 } | 391 } |
392 | 392 |
393 | 393 |
394 // Input parameters: | 394 // Input parameters: |
395 // R2: smi-tagged argument count, may be zero. | 395 // R2: smi-tagged argument count, may be zero. |
396 // FP[kParamEndSlotFromFp + 1]: last argument. | 396 // FP[kParamEndSlotFromFp + 1]: last argument. |
397 static void PushArgumentsArray(Assembler* assembler) { | 397 static void PushArgumentsArray(Assembler* assembler) { |
398 // Allocate array to store arguments of caller. | 398 // Allocate array to store arguments of caller. |
399 __ LoadObject(R1, Object::null_object(), PP); | 399 __ LoadObject(R1, Object::null_object()); |
400 // R1: null element type for raw Array. | 400 // R1: null element type for raw Array. |
401 // R2: smi-tagged argument count, may be zero. | 401 // R2: smi-tagged argument count, may be zero. |
402 const ExternalLabel array_label(StubCode::AllocateArrayEntryPoint()); | 402 const ExternalLabel array_label(StubCode::AllocateArrayEntryPoint()); |
403 __ BranchLink(&array_label, PP); | 403 __ BranchLink(&array_label); |
404 // R0: newly allocated array. | 404 // R0: newly allocated array. |
405 // R2: smi-tagged argument count, may be zero (was preserved by the stub). | 405 // R2: smi-tagged argument count, may be zero (was preserved by the stub). |
406 __ Push(R0); // Array is in R0 and on top of stack. | 406 __ Push(R0); // Array is in R0 and on top of stack. |
407 __ add(R1, FP, Operand(R2, LSL, 2)); | 407 __ add(R1, FP, Operand(R2, LSL, 2)); |
408 __ AddImmediate(R1, R1, kParamEndSlotFromFp * kWordSize, PP); | 408 __ AddImmediate(R1, R1, kParamEndSlotFromFp * kWordSize); |
409 __ AddImmediate(R3, R0, Array::data_offset() - kHeapObjectTag, PP); | 409 __ AddImmediate(R3, R0, Array::data_offset() - kHeapObjectTag); |
410 // R1: address of first argument on stack. | 410 // R1: address of first argument on stack. |
411 // R3: address of first argument in array. | 411 // R3: address of first argument in array. |
412 | 412 |
413 Label loop, loop_exit; | 413 Label loop, loop_exit; |
414 __ CompareRegisters(R2, ZR); | 414 __ CompareRegisters(R2, ZR); |
415 __ b(&loop_exit, LE); | 415 __ b(&loop_exit, LE); |
416 __ Bind(&loop); | 416 __ Bind(&loop); |
417 __ ldr(R7, Address(R1)); | 417 __ ldr(R7, Address(R1)); |
418 __ AddImmediate(R1, R1, -kWordSize, PP); | 418 __ AddImmediate(R1, R1, -kWordSize); |
419 __ AddImmediate(R3, R3, kWordSize, PP); | 419 __ AddImmediate(R3, R3, kWordSize); |
420 __ AddImmediateSetFlags(R2, R2, -Smi::RawValue(1), PP); | 420 __ AddImmediateSetFlags(R2, R2, -Smi::RawValue(1)); |
421 __ str(R7, Address(R3, -kWordSize)); | 421 __ str(R7, Address(R3, -kWordSize)); |
422 __ b(&loop, GE); | 422 __ b(&loop, GE); |
423 __ Bind(&loop_exit); | 423 __ Bind(&loop_exit); |
424 } | 424 } |
425 | 425 |
426 | 426 |
427 DECLARE_LEAF_RUNTIME_ENTRY(intptr_t, DeoptimizeCopyFrame, | 427 DECLARE_LEAF_RUNTIME_ENTRY(intptr_t, DeoptimizeCopyFrame, |
428 intptr_t deopt_reason, | 428 intptr_t deopt_reason, |
429 uword saved_registers_address); | 429 uword saved_registers_address); |
430 | 430 |
(...skipping 49 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
480 __ PushQuad(vreg); | 480 __ PushQuad(vreg); |
481 } | 481 } |
482 | 482 |
483 __ mov(R0, SP); // Pass address of saved registers block. | 483 __ mov(R0, SP); // Pass address of saved registers block. |
484 __ ReserveAlignedFrameSpace(0); | 484 __ ReserveAlignedFrameSpace(0); |
485 __ CallRuntime(kDeoptimizeCopyFrameRuntimeEntry, 1); | 485 __ CallRuntime(kDeoptimizeCopyFrameRuntimeEntry, 1); |
486 // Result (R0) is stack-size (FP - SP) in bytes. | 486 // Result (R0) is stack-size (FP - SP) in bytes. |
487 | 487 |
488 if (preserve_result) { | 488 if (preserve_result) { |
489 // Restore result into R1 temporarily. | 489 // Restore result into R1 temporarily. |
490 __ LoadFromOffset(R1, FP, saved_result_slot_from_fp * kWordSize, kNoPP); | 490 __ LoadFromOffset(R1, FP, saved_result_slot_from_fp * kWordSize); |
491 } | 491 } |
492 | 492 |
493 // There is a Dart Frame on the stack. We must restore PP and leave frame. | 493 // There is a Dart Frame on the stack. We must restore PP and leave frame. |
494 __ LeaveDartFrame(); | 494 __ LeaveDartFrame(); |
495 __ sub(SP, FP, Operand(R0)); | 495 __ sub(SP, FP, Operand(R0)); |
496 | 496 |
497 // DeoptimizeFillFrame expects a Dart frame, i.e. EnterDartFrame(0), but there | 497 // DeoptimizeFillFrame expects a Dart frame, i.e. EnterDartFrame(0), but there |
498 // is no need to set the correct PC marker or load PP, since they get patched. | 498 // is no need to set the correct PC marker or load PP, since they get patched. |
499 __ EnterFrame(0); | 499 __ EnterFrame(0); |
500 __ TagAndPushPPAndPcMarker(ZR); | 500 __ TagAndPushPPAndPcMarker(ZR); |
501 | 501 |
502 if (preserve_result) { | 502 if (preserve_result) { |
503 __ Push(R1); // Preserve result as first local. | 503 __ Push(R1); // Preserve result as first local. |
504 } | 504 } |
505 __ ReserveAlignedFrameSpace(0); | 505 __ ReserveAlignedFrameSpace(0); |
506 __ mov(R0, FP); // Pass last FP as parameter in R0. | 506 __ mov(R0, FP); // Pass last FP as parameter in R0. |
507 __ CallRuntime(kDeoptimizeFillFrameRuntimeEntry, 1); | 507 __ CallRuntime(kDeoptimizeFillFrameRuntimeEntry, 1); |
508 if (preserve_result) { | 508 if (preserve_result) { |
509 // Restore result into R1. | 509 // Restore result into R1. |
510 __ LoadFromOffset(R1, FP, kFirstLocalSlotFromFp * kWordSize, kNoPP); | 510 __ LoadFromOffset(R1, FP, kFirstLocalSlotFromFp * kWordSize); |
511 } | 511 } |
512 // Code above cannot cause GC. | 512 // Code above cannot cause GC. |
513 // There is a Dart Frame on the stack. We must restore PP and leave frame. | 513 // There is a Dart Frame on the stack. We must restore PP and leave frame. |
514 __ LeaveDartFrame(); | 514 __ LeaveDartFrame(); |
515 | 515 |
516 // Frame is fully rewritten at this point and it is safe to perform a GC. | 516 // Frame is fully rewritten at this point and it is safe to perform a GC. |
517 // Materialize any objects that were deferred by FillFrame because they | 517 // Materialize any objects that were deferred by FillFrame because they |
518 // require allocation. | 518 // require allocation. |
519 // Enter stub frame with loading PP. The caller's PP is not materialized yet. | 519 // Enter stub frame with loading PP. The caller's PP is not materialized yet. |
520 __ EnterStubFrame(); | 520 __ EnterStubFrame(); |
(...skipping 14 matching lines...) Expand all Loading... |
535 __ LeaveStubFrame(); | 535 __ LeaveStubFrame(); |
536 // Remove materialization arguments. | 536 // Remove materialization arguments. |
537 __ add(SP, SP, Operand(R1)); | 537 __ add(SP, SP, Operand(R1)); |
538 __ ret(); | 538 __ ret(); |
539 } | 539 } |
540 | 540 |
541 | 541 |
542 void StubCode::GenerateDeoptimizeLazyStub(Assembler* assembler) { | 542 void StubCode::GenerateDeoptimizeLazyStub(Assembler* assembler) { |
543 // Correct return address to point just after the call that is being | 543 // Correct return address to point just after the call that is being |
544 // deoptimized. | 544 // deoptimized. |
545 __ AddImmediate(LR, LR, -CallPattern::kLengthInBytes, kNoPP); | 545 __ AddImmediate(LR, LR, -CallPattern::kLengthInBytes); |
546 GenerateDeoptimizationSequence(assembler, true); // Preserve R0. | 546 GenerateDeoptimizationSequence(assembler, true); // Preserve R0. |
547 } | 547 } |
548 | 548 |
549 | 549 |
550 void StubCode::GenerateDeoptimizeStub(Assembler* assembler) { | 550 void StubCode::GenerateDeoptimizeStub(Assembler* assembler) { |
551 GenerateDeoptimizationSequence(assembler, false); // Don't preserve R0. | 551 GenerateDeoptimizationSequence(assembler, false); // Don't preserve R0. |
552 } | 552 } |
553 | 553 |
554 | 554 |
555 static void GenerateDispatcherCode(Assembler* assembler, | 555 static void GenerateDispatcherCode(Assembler* assembler, |
556 Label* call_target_function) { | 556 Label* call_target_function) { |
557 __ Comment("NoSuchMethodDispatch"); | 557 __ Comment("NoSuchMethodDispatch"); |
558 // When lazily generated invocation dispatchers are disabled, the | 558 // When lazily generated invocation dispatchers are disabled, the |
559 // miss-handler may return null. | 559 // miss-handler may return null. |
560 __ CompareObject(R0, Object::null_object(), PP); | 560 __ CompareObject(R0, Object::null_object()); |
561 __ b(call_target_function, NE); | 561 __ b(call_target_function, NE); |
562 __ EnterStubFrame(); | 562 __ EnterStubFrame(); |
563 | 563 |
564 // Load the receiver. | 564 // Load the receiver. |
565 __ LoadFieldFromOffset(R2, R4, ArgumentsDescriptor::count_offset(), kNoPP); | 565 __ LoadFieldFromOffset(R2, R4, ArgumentsDescriptor::count_offset()); |
566 __ add(TMP, FP, Operand(R2, LSL, 2)); // R2 is Smi. | 566 __ add(TMP, FP, Operand(R2, LSL, 2)); // R2 is Smi. |
567 __ LoadFromOffset(R6, TMP, kParamEndSlotFromFp * kWordSize, kNoPP); | 567 __ LoadFromOffset(R6, TMP, kParamEndSlotFromFp * kWordSize); |
568 __ PushObject(Object::null_object(), PP); | 568 __ PushObject(Object::null_object()); |
569 __ Push(R6); | 569 __ Push(R6); |
570 __ Push(R5); | 570 __ Push(R5); |
571 __ Push(R4); | 571 __ Push(R4); |
572 // R2: Smi-tagged arguments array length. | 572 // R2: Smi-tagged arguments array length. |
573 PushArgumentsArray(assembler); | 573 PushArgumentsArray(assembler); |
574 const intptr_t kNumArgs = 4; | 574 const intptr_t kNumArgs = 4; |
575 __ CallRuntime(kInvokeNoSuchMethodDispatcherRuntimeEntry, kNumArgs); | 575 __ CallRuntime(kInvokeNoSuchMethodDispatcherRuntimeEntry, kNumArgs); |
576 __ Drop(4); | 576 __ Drop(4); |
577 __ Pop(R0); // Return value. | 577 __ Pop(R0); // Return value. |
578 __ LeaveStubFrame(); | 578 __ LeaveStubFrame(); |
579 __ ret(); | 579 __ ret(); |
580 } | 580 } |
581 | 581 |
582 | 582 |
583 void StubCode::GenerateMegamorphicMissStub(Assembler* assembler) { | 583 void StubCode::GenerateMegamorphicMissStub(Assembler* assembler) { |
584 __ EnterStubFrame(); | 584 __ EnterStubFrame(); |
585 | 585 |
586 // Load the receiver. | 586 // Load the receiver. |
587 __ LoadFieldFromOffset(R2, R4, ArgumentsDescriptor::count_offset(), kNoPP); | 587 __ LoadFieldFromOffset(R2, R4, ArgumentsDescriptor::count_offset()); |
588 __ add(TMP, FP, Operand(R2, LSL, 2)); // R2 is Smi. | 588 __ add(TMP, FP, Operand(R2, LSL, 2)); // R2 is Smi. |
589 __ LoadFromOffset(R6, TMP, kParamEndSlotFromFp * kWordSize, kNoPP); | 589 __ LoadFromOffset(R6, TMP, kParamEndSlotFromFp * kWordSize); |
590 | 590 |
591 // Preserve IC data and arguments descriptor. | 591 // Preserve IC data and arguments descriptor. |
592 __ Push(R5); | 592 __ Push(R5); |
593 __ Push(R4); | 593 __ Push(R4); |
594 | 594 |
595 // Push space for the return value. | 595 // Push space for the return value. |
596 // Push the receiver. | 596 // Push the receiver. |
597 // Push IC data object. | 597 // Push IC data object. |
598 // Push arguments descriptor array. | 598 // Push arguments descriptor array. |
599 __ PushObject(Object::null_object(), PP); | 599 __ PushObject(Object::null_object()); |
600 __ Push(R6); | 600 __ Push(R6); |
601 __ Push(R5); | 601 __ Push(R5); |
602 __ Push(R4); | 602 __ Push(R4); |
603 __ CallRuntime(kMegamorphicCacheMissHandlerRuntimeEntry, 3); | 603 __ CallRuntime(kMegamorphicCacheMissHandlerRuntimeEntry, 3); |
604 // Remove arguments. | 604 // Remove arguments. |
605 __ Drop(3); | 605 __ Drop(3); |
606 __ Pop(R0); // Get result into R0 (target function). | 606 __ Pop(R0); // Get result into R0 (target function). |
607 | 607 |
608 // Restore IC data and arguments descriptor. | 608 // Restore IC data and arguments descriptor. |
609 __ Pop(R4); | 609 __ Pop(R4); |
610 __ Pop(R5); | 610 __ Pop(R5); |
611 | 611 |
612 __ LeaveStubFrame(); | 612 __ LeaveStubFrame(); |
613 | 613 |
614 if (!FLAG_lazy_dispatchers) { | 614 if (!FLAG_lazy_dispatchers) { |
615 Label call_target_function; | 615 Label call_target_function; |
616 GenerateDispatcherCode(assembler, &call_target_function); | 616 GenerateDispatcherCode(assembler, &call_target_function); |
617 __ Bind(&call_target_function); | 617 __ Bind(&call_target_function); |
618 } | 618 } |
619 | 619 |
620 // Tail-call to target function. | 620 // Tail-call to target function. |
621 __ LoadFieldFromOffset(R2, R0, Function::instructions_offset(), kNoPP); | 621 __ LoadFieldFromOffset(R2, R0, Function::instructions_offset()); |
622 __ AddImmediate(R2, R2, Instructions::HeaderSize() - kHeapObjectTag, PP); | 622 __ AddImmediate(R2, R2, Instructions::HeaderSize() - kHeapObjectTag); |
623 __ br(R2); | 623 __ br(R2); |
624 } | 624 } |
625 | 625 |
626 | 626 |
627 // Called for inline allocation of arrays. | 627 // Called for inline allocation of arrays. |
628 // Input parameters: | 628 // Input parameters: |
629 // LR: return address. | 629 // LR: return address. |
630 // R2: array length as Smi. | 630 // R2: array length as Smi. |
631 // R1: array element type (either NULL or an instantiated type). | 631 // R1: array element type (either NULL or an instantiated type). |
632 // NOTE: R2 cannot be clobbered here as the caller relies on it being saved. | 632 // NOTE: R2 cannot be clobbered here as the caller relies on it being saved. |
633 // The newly allocated object is returned in R0. | 633 // The newly allocated object is returned in R0. |
634 void StubCode::GenerateAllocateArrayStub(Assembler* assembler) { | 634 void StubCode::GenerateAllocateArrayStub(Assembler* assembler) { |
635 Label slow_case; | 635 Label slow_case; |
636 // Compute the size to be allocated, it is based on the array length | 636 // Compute the size to be allocated, it is based on the array length |
637 // and is computed as: | 637 // and is computed as: |
638 // RoundedAllocationSize((array_length * kwordSize) + sizeof(RawArray)). | 638 // RoundedAllocationSize((array_length * kwordSize) + sizeof(RawArray)). |
639 // Assert that length is a Smi. | 639 // Assert that length is a Smi. |
640 __ tsti(R2, Immediate(kSmiTagMask)); | 640 __ tsti(R2, Immediate(kSmiTagMask)); |
641 if (FLAG_use_slow_path) { | 641 if (FLAG_use_slow_path) { |
642 __ b(&slow_case); | 642 __ b(&slow_case); |
643 } else { | 643 } else { |
644 __ b(&slow_case, NE); | 644 __ b(&slow_case, NE); |
645 } | 645 } |
646 __ cmp(R2, Operand(0)); | 646 __ cmp(R2, Operand(0)); |
647 __ b(&slow_case, LT); | 647 __ b(&slow_case, LT); |
648 | 648 |
649 // Check for maximum allowed length. | 649 // Check for maximum allowed length. |
650 const intptr_t max_len = | 650 const intptr_t max_len = |
651 reinterpret_cast<intptr_t>(Smi::New(Array::kMaxElements)); | 651 reinterpret_cast<intptr_t>(Smi::New(Array::kMaxElements)); |
652 __ CompareImmediate(R2, max_len, kNoPP); | 652 __ CompareImmediate(R2, max_len); |
653 __ b(&slow_case, GT); | 653 __ b(&slow_case, GT); |
654 | 654 |
655 const intptr_t cid = kArrayCid; | 655 const intptr_t cid = kArrayCid; |
656 __ MaybeTraceAllocation(kArrayCid, R4, kNoPP, &slow_case, | 656 __ MaybeTraceAllocation(kArrayCid, R4, &slow_case, |
657 /* inline_isolate = */ false); | 657 /* inline_isolate = */ false); |
658 | 658 |
659 Heap::Space space = Heap::SpaceForAllocation(cid); | 659 Heap::Space space = Heap::SpaceForAllocation(cid); |
660 __ LoadIsolate(R8); | 660 __ LoadIsolate(R8); |
661 __ ldr(R8, Address(R8, Isolate::heap_offset())); | 661 __ ldr(R8, Address(R8, Isolate::heap_offset())); |
662 | 662 |
663 // Calculate and align allocation size. | 663 // Calculate and align allocation size. |
664 // Load new object start and calculate next object start. | 664 // Load new object start and calculate next object start. |
665 // R1: array element type. | 665 // R1: array element type. |
666 // R2: array length as Smi. | 666 // R2: array length as Smi. |
667 // R8: heap. | 667 // R8: heap. |
668 __ LoadFromOffset(R0, R8, Heap::TopOffset(space), kNoPP); | 668 __ LoadFromOffset(R0, R8, Heap::TopOffset(space)); |
669 intptr_t fixed_size = sizeof(RawArray) + kObjectAlignment - 1; | 669 intptr_t fixed_size = sizeof(RawArray) + kObjectAlignment - 1; |
670 __ LoadImmediate(R3, fixed_size, kNoPP); | 670 __ LoadImmediate(R3, fixed_size); |
671 __ add(R3, R3, Operand(R2, LSL, 2)); // R2 is Smi. | 671 __ add(R3, R3, Operand(R2, LSL, 2)); // R2 is Smi. |
672 ASSERT(kSmiTagShift == 1); | 672 ASSERT(kSmiTagShift == 1); |
673 __ andi(R3, R3, Immediate(~(kObjectAlignment - 1))); | 673 __ andi(R3, R3, Immediate(~(kObjectAlignment - 1))); |
674 // R0: potential new object start. | 674 // R0: potential new object start. |
675 // R3: object size in bytes. | 675 // R3: object size in bytes. |
676 __ adds(R7, R3, Operand(R0)); | 676 __ adds(R7, R3, Operand(R0)); |
677 __ b(&slow_case, CS); // Branch if unsigned overflow. | 677 __ b(&slow_case, CS); // Branch if unsigned overflow. |
678 | 678 |
679 // Check if the allocation fits into the remaining space. | 679 // Check if the allocation fits into the remaining space. |
680 // R0: potential new object start. | 680 // R0: potential new object start. |
681 // R1: array element type. | 681 // R1: array element type. |
682 // R2: array length as Smi. | 682 // R2: array length as Smi. |
683 // R3: array size. | 683 // R3: array size. |
684 // R7: potential next object start. | 684 // R7: potential next object start. |
685 // R8: heap. | 685 // R8: heap. |
686 __ LoadFromOffset(TMP, R8, Heap::EndOffset(space), kNoPP); | 686 __ LoadFromOffset(TMP, R8, Heap::EndOffset(space)); |
687 __ CompareRegisters(R7, TMP); | 687 __ CompareRegisters(R7, TMP); |
688 __ b(&slow_case, CS); // Branch if unsigned higher or equal. | 688 __ b(&slow_case, CS); // Branch if unsigned higher or equal. |
689 | 689 |
690 // Successfully allocated the object(s), now update top to point to | 690 // Successfully allocated the object(s), now update top to point to |
691 // next object start and initialize the object. | 691 // next object start and initialize the object. |
692 // R0: potential new object start. | 692 // R0: potential new object start. |
693 // R3: array size. | 693 // R3: array size. |
694 // R7: potential next object start. | 694 // R7: potential next object start. |
695 // R8: heap. | 695 // R8: heap. |
696 __ StoreToOffset(R7, R8, Heap::TopOffset(space), kNoPP); | 696 __ StoreToOffset(R7, R8, Heap::TopOffset(space)); |
697 __ add(R0, R0, Operand(kHeapObjectTag)); | 697 __ add(R0, R0, Operand(kHeapObjectTag)); |
698 __ UpdateAllocationStatsWithSize(cid, R3, kNoPP, space, | 698 __ UpdateAllocationStatsWithSize(cid, R3, space, |
699 /* inline_isolate = */ false); | 699 /* inline_isolate = */ false); |
700 | 700 |
701 // R0: new object start as a tagged pointer. | 701 // R0: new object start as a tagged pointer. |
702 // R1: array element type. | 702 // R1: array element type. |
703 // R2: array length as Smi. | 703 // R2: array length as Smi. |
704 // R3: array size. | 704 // R3: array size. |
705 // R7: new object end address. | 705 // R7: new object end address. |
706 | 706 |
707 // Store the type argument field. | 707 // Store the type argument field. |
708 __ StoreIntoObjectOffsetNoBarrier( | 708 __ StoreIntoObjectOffsetNoBarrier( |
709 R0, Array::type_arguments_offset(), R1, PP); | 709 R0, Array::type_arguments_offset(), R1); |
710 | 710 |
711 // Set the length field. | 711 // Set the length field. |
712 __ StoreIntoObjectOffsetNoBarrier(R0, Array::length_offset(), R2, PP); | 712 __ StoreIntoObjectOffsetNoBarrier(R0, Array::length_offset(), R2); |
713 | 713 |
714 // Calculate the size tag. | 714 // Calculate the size tag. |
715 // R0: new object start as a tagged pointer. | 715 // R0: new object start as a tagged pointer. |
716 // R2: array length as Smi. | 716 // R2: array length as Smi. |
717 // R3: array size. | 717 // R3: array size. |
718 // R7: new object end address. | 718 // R7: new object end address. |
719 const intptr_t shift = RawObject::kSizeTagPos - kObjectAlignmentLog2; | 719 const intptr_t shift = RawObject::kSizeTagPos - kObjectAlignmentLog2; |
720 __ CompareImmediate(R3, RawObject::SizeTag::kMaxSizeTag, kNoPP); | 720 __ CompareImmediate(R3, RawObject::SizeTag::kMaxSizeTag); |
721 // If no size tag overflow, shift R1 left, else set R1 to zero. | 721 // If no size tag overflow, shift R1 left, else set R1 to zero. |
722 __ LslImmediate(TMP, R3, shift); | 722 __ LslImmediate(TMP, R3, shift); |
723 __ csel(R1, TMP, R1, LS); | 723 __ csel(R1, TMP, R1, LS); |
724 __ csel(R1, ZR, R1, HI); | 724 __ csel(R1, ZR, R1, HI); |
725 | 725 |
726 // Get the class index and insert it into the tags. | 726 // Get the class index and insert it into the tags. |
727 __ LoadImmediate(TMP, RawObject::ClassIdTag::encode(cid), kNoPP); | 727 __ LoadImmediate(TMP, RawObject::ClassIdTag::encode(cid)); |
728 __ orr(R1, R1, Operand(TMP)); | 728 __ orr(R1, R1, Operand(TMP)); |
729 __ StoreFieldToOffset(R1, R0, Array::tags_offset(), kNoPP); | 729 __ StoreFieldToOffset(R1, R0, Array::tags_offset()); |
730 | 730 |
731 // Initialize all array elements to raw_null. | 731 // Initialize all array elements to raw_null. |
732 // R0: new object start as a tagged pointer. | 732 // R0: new object start as a tagged pointer. |
733 // R7: new object end address. | 733 // R7: new object end address. |
734 // R2: array length as Smi. | 734 // R2: array length as Smi. |
735 __ AddImmediate(R1, R0, Array::data_offset() - kHeapObjectTag, kNoPP); | 735 __ AddImmediate(R1, R0, Array::data_offset() - kHeapObjectTag); |
736 // R1: iterator which initially points to the start of the variable | 736 // R1: iterator which initially points to the start of the variable |
737 // data area to be initialized. | 737 // data area to be initialized. |
738 __ LoadObject(TMP, Object::null_object(), PP); | 738 __ LoadObject(TMP, Object::null_object()); |
739 Label loop, done; | 739 Label loop, done; |
740 __ Bind(&loop); | 740 __ Bind(&loop); |
741 // TODO(cshapiro): StoreIntoObjectNoBarrier | 741 // TODO(cshapiro): StoreIntoObjectNoBarrier |
742 __ CompareRegisters(R1, R7); | 742 __ CompareRegisters(R1, R7); |
743 __ b(&done, CS); | 743 __ b(&done, CS); |
744 __ str(TMP, Address(R1)); // Store if unsigned lower. | 744 __ str(TMP, Address(R1)); // Store if unsigned lower. |
745 __ AddImmediate(R1, R1, kWordSize, kNoPP); | 745 __ AddImmediate(R1, R1, kWordSize); |
746 __ b(&loop); // Loop until R1 == R7. | 746 __ b(&loop); // Loop until R1 == R7. |
747 __ Bind(&done); | 747 __ Bind(&done); |
748 | 748 |
749 // Done allocating and initializing the array. | 749 // Done allocating and initializing the array. |
750 // R0: new object. | 750 // R0: new object. |
751 // R2: array length as Smi (preserved for the caller.) | 751 // R2: array length as Smi (preserved for the caller.) |
752 __ ret(); | 752 __ ret(); |
753 | 753 |
754 // Unable to allocate the array using the fast inline code, just call | 754 // Unable to allocate the array using the fast inline code, just call |
755 // into the runtime. | 755 // into the runtime. |
756 __ Bind(&slow_case); | 756 __ Bind(&slow_case); |
757 // Create a stub frame as we are pushing some objects on the stack before | 757 // Create a stub frame as we are pushing some objects on the stack before |
758 // calling into the runtime. | 758 // calling into the runtime. |
759 __ EnterStubFrame(); | 759 __ EnterStubFrame(); |
760 // Setup space on stack for return value. | 760 // Setup space on stack for return value. |
761 // Push array length as Smi and element type. | 761 // Push array length as Smi and element type. |
762 __ PushObject(Object::null_object(), PP); | 762 __ PushObject(Object::null_object()); |
763 __ Push(R2); | 763 __ Push(R2); |
764 __ Push(R1); | 764 __ Push(R1); |
765 __ CallRuntime(kAllocateArrayRuntimeEntry, 2); | 765 __ CallRuntime(kAllocateArrayRuntimeEntry, 2); |
766 // Pop arguments; result is popped in IP. | 766 // Pop arguments; result is popped in IP. |
767 __ Pop(R1); | 767 __ Pop(R1); |
768 __ Pop(R2); | 768 __ Pop(R2); |
769 __ Pop(R0); | 769 __ Pop(R0); |
770 __ LeaveStubFrame(); | 770 __ LeaveStubFrame(); |
771 __ ret(); | 771 __ ret(); |
772 } | 772 } |
(...skipping 25 matching lines...) Expand all Loading... |
798 | 798 |
799 // Save the bottom 64-bits of callee-saved V registers. | 799 // Save the bottom 64-bits of callee-saved V registers. |
800 for (int i = kAbiFirstPreservedFpuReg; i <= kAbiLastPreservedFpuReg; i++) { | 800 for (int i = kAbiFirstPreservedFpuReg; i <= kAbiLastPreservedFpuReg; i++) { |
801 const VRegister r = static_cast<VRegister>(i); | 801 const VRegister r = static_cast<VRegister>(i); |
802 __ PushDouble(r); | 802 __ PushDouble(r); |
803 } | 803 } |
804 | 804 |
805 // We now load the pool pointer(PP) as we are about to invoke dart code and we | 805 // We now load the pool pointer(PP) as we are about to invoke dart code and we |
806 // could potentially invoke some intrinsic functions which need the PP to be | 806 // could potentially invoke some intrinsic functions which need the PP to be |
807 // set up. | 807 // set up. |
808 __ LoadPoolPointer(PP); | 808 __ LoadPoolPointer(); |
809 | 809 |
810 // Set up THR, which caches the current thread in Dart code. | 810 // Set up THR, which caches the current thread in Dart code. |
811 if (THR != R3) { | 811 if (THR != R3) { |
812 __ mov(THR, R3); | 812 __ mov(THR, R3); |
813 } | 813 } |
814 // Load Isolate pointer into temporary register R5. | 814 // Load Isolate pointer into temporary register R5. |
815 __ LoadIsolate(R5); | 815 __ LoadIsolate(R5); |
816 | 816 |
817 // Save the current VMTag on the stack. | 817 // Save the current VMTag on the stack. |
818 __ LoadFromOffset(R4, R5, Isolate::vm_tag_offset(), PP); | 818 __ LoadFromOffset(R4, R5, Isolate::vm_tag_offset()); |
819 __ Push(R4); | 819 __ Push(R4); |
820 | 820 |
821 // Mark that the isolate is executing Dart code. | 821 // Mark that the isolate is executing Dart code. |
822 __ LoadImmediate(R6, VMTag::kDartTagId, PP); | 822 __ LoadImmediate(R6, VMTag::kDartTagId); |
823 __ StoreToOffset(R6, R5, Isolate::vm_tag_offset(), PP); | 823 __ StoreToOffset(R6, R5, Isolate::vm_tag_offset()); |
824 | 824 |
825 // Save top resource and top exit frame info. Use R6 as a temporary register. | 825 // Save top resource and top exit frame info. Use R6 as a temporary register. |
826 // StackFrameIterator reads the top exit frame info saved in this frame. | 826 // StackFrameIterator reads the top exit frame info saved in this frame. |
827 __ LoadFromOffset(R6, THR, Thread::top_resource_offset(), PP); | 827 __ LoadFromOffset(R6, THR, Thread::top_resource_offset()); |
828 __ StoreToOffset(ZR, THR, Thread::top_resource_offset(), PP); | 828 __ StoreToOffset(ZR, THR, Thread::top_resource_offset()); |
829 __ Push(R6); | 829 __ Push(R6); |
830 __ LoadFromOffset(R6, THR, Thread::top_exit_frame_info_offset(), PP); | 830 __ LoadFromOffset(R6, THR, Thread::top_exit_frame_info_offset()); |
831 __ StoreToOffset(ZR, THR, Thread::top_exit_frame_info_offset(), PP); | 831 __ StoreToOffset(ZR, THR, Thread::top_exit_frame_info_offset()); |
832 // kExitLinkSlotFromEntryFp must be kept in sync with the code below. | 832 // kExitLinkSlotFromEntryFp must be kept in sync with the code below. |
833 ASSERT(kExitLinkSlotFromEntryFp == -21); | 833 ASSERT(kExitLinkSlotFromEntryFp == -21); |
834 __ Push(R6); | 834 __ Push(R6); |
835 | 835 |
836 // Load arguments descriptor array into R4, which is passed to Dart code. | 836 // Load arguments descriptor array into R4, which is passed to Dart code. |
837 __ LoadFromOffset(R4, R1, VMHandles::kOffsetOfRawPtrInHandle, PP); | 837 __ LoadFromOffset(R4, R1, VMHandles::kOffsetOfRawPtrInHandle); |
838 | 838 |
839 // Load number of arguments into S5. | 839 // Load number of arguments into S5. |
840 __ LoadFieldFromOffset(R5, R4, ArgumentsDescriptor::count_offset(), PP); | 840 __ LoadFieldFromOffset(R5, R4, ArgumentsDescriptor::count_offset()); |
841 __ SmiUntag(R5); | 841 __ SmiUntag(R5); |
842 | 842 |
843 // Compute address of 'arguments array' data area into R2. | 843 // Compute address of 'arguments array' data area into R2. |
844 __ LoadFromOffset(R2, R2, VMHandles::kOffsetOfRawPtrInHandle, PP); | 844 __ LoadFromOffset(R2, R2, VMHandles::kOffsetOfRawPtrInHandle); |
845 __ AddImmediate(R2, R2, Array::data_offset() - kHeapObjectTag, PP); | 845 __ AddImmediate(R2, R2, Array::data_offset() - kHeapObjectTag); |
846 | 846 |
847 // Set up arguments for the Dart call. | 847 // Set up arguments for the Dart call. |
848 Label push_arguments; | 848 Label push_arguments; |
849 Label done_push_arguments; | 849 Label done_push_arguments; |
850 __ cmp(R5, Operand(0)); | 850 __ cmp(R5, Operand(0)); |
851 __ b(&done_push_arguments, EQ); // check if there are arguments. | 851 __ b(&done_push_arguments, EQ); // check if there are arguments. |
852 __ LoadImmediate(R1, 0, PP); | 852 __ LoadImmediate(R1, 0); |
853 __ Bind(&push_arguments); | 853 __ Bind(&push_arguments); |
854 __ ldr(R3, Address(R2)); | 854 __ ldr(R3, Address(R2)); |
855 __ Push(R3); | 855 __ Push(R3); |
856 __ add(R1, R1, Operand(1)); | 856 __ add(R1, R1, Operand(1)); |
857 __ add(R2, R2, Operand(kWordSize)); | 857 __ add(R2, R2, Operand(kWordSize)); |
858 __ cmp(R1, Operand(R5)); | 858 __ cmp(R1, Operand(R5)); |
859 __ b(&push_arguments, LT); | 859 __ b(&push_arguments, LT); |
860 __ Bind(&done_push_arguments); | 860 __ Bind(&done_push_arguments); |
861 | 861 |
862 // Call the Dart code entrypoint. | 862 // Call the Dart code entrypoint. |
863 __ blr(R0); // R4 is the arguments descriptor array. | 863 __ blr(R0); // R4 is the arguments descriptor array. |
864 __ Comment("InvokeDartCodeStub return"); | 864 __ Comment("InvokeDartCodeStub return"); |
865 | 865 |
866 // Restore constant pool pointer after return. | 866 // Restore constant pool pointer after return. |
867 __ LoadPoolPointer(PP); | 867 __ LoadPoolPointer(); |
868 | 868 |
869 // Get rid of arguments pushed on the stack. | 869 // Get rid of arguments pushed on the stack. |
870 __ AddImmediate(SP, FP, kExitLinkSlotFromEntryFp * kWordSize, PP); | 870 __ AddImmediate(SP, FP, kExitLinkSlotFromEntryFp * kWordSize); |
871 | 871 |
872 __ LoadIsolate(R28); | 872 __ LoadIsolate(R28); |
873 | 873 |
874 // Restore the saved top exit frame info and top resource back into the | 874 // Restore the saved top exit frame info and top resource back into the |
875 // Isolate structure. Uses R6 as a temporary register for this. | 875 // Isolate structure. Uses R6 as a temporary register for this. |
876 __ Pop(R6); | 876 __ Pop(R6); |
877 __ StoreToOffset(R6, THR, Thread::top_exit_frame_info_offset(), PP); | 877 __ StoreToOffset(R6, THR, Thread::top_exit_frame_info_offset()); |
878 __ Pop(R6); | 878 __ Pop(R6); |
879 __ StoreToOffset(R6, THR, Thread::top_resource_offset(), PP); | 879 __ StoreToOffset(R6, THR, Thread::top_resource_offset()); |
880 | 880 |
881 // Restore the current VMTag from the stack. | 881 // Restore the current VMTag from the stack. |
882 __ Pop(R4); | 882 __ Pop(R4); |
883 __ StoreToOffset(R4, R28, Isolate::vm_tag_offset(), PP); | 883 __ StoreToOffset(R4, R28, Isolate::vm_tag_offset()); |
884 | 884 |
885 // Restore the bottom 64-bits of callee-saved V registers. | 885 // Restore the bottom 64-bits of callee-saved V registers. |
886 for (int i = kAbiLastPreservedFpuReg; i >= kAbiFirstPreservedFpuReg; i--) { | 886 for (int i = kAbiLastPreservedFpuReg; i >= kAbiFirstPreservedFpuReg; i--) { |
887 const VRegister r = static_cast<VRegister>(i); | 887 const VRegister r = static_cast<VRegister>(i); |
888 __ PopDouble(r); | 888 __ PopDouble(r); |
889 } | 889 } |
890 | 890 |
891 // Restore C++ ABI callee-saved registers. | 891 // Restore C++ ABI callee-saved registers. |
892 for (int i = kAbiLastPreservedCpuReg; i >= kAbiFirstPreservedCpuReg; i--) { | 892 for (int i = kAbiLastPreservedCpuReg; i >= kAbiFirstPreservedCpuReg; i--) { |
893 Register r = static_cast<Register>(i); | 893 Register r = static_cast<Register>(i); |
894 // We use ldr instead of the Pop macro because we will be popping the PP | 894 // We use ldr instead of the Pop macro because we will be popping the PP |
895 // register when it is not holding a pool-pointer since we are returning to | 895 // register when it is not holding a pool-pointer since we are returning to |
896 // C++ code. We also skip the dart stack pointer SP, since we are still | 896 // C++ code. We also skip the dart stack pointer SP, since we are still |
897 // using it as the stack pointer. | 897 // using it as the stack pointer. |
898 __ ldr(r, Address(SP, 1 * kWordSize, Address::PostIndex)); | 898 __ ldr(r, Address(SP, 1 * kWordSize, Address::PostIndex)); |
899 } | 899 } |
| 900 __ set_constant_pool_allowed(false); |
900 | 901 |
901 // Restore the frame pointer and C stack pointer and return. | 902 // Restore the frame pointer and C stack pointer and return. |
902 __ LeaveFrame(); | 903 __ LeaveFrame(); |
903 __ mov(CSP, SP); | 904 __ mov(CSP, SP); |
904 __ ret(); | 905 __ ret(); |
905 } | 906 } |
906 | 907 |
907 | 908 |
908 // Called for inline allocation of contexts. | 909 // Called for inline allocation of contexts. |
909 // Input: | 910 // Input: |
910 // R1: number of context variables. | 911 // R1: number of context variables. |
911 // Output: | 912 // Output: |
912 // R0: new allocated RawContext object. | 913 // R0: new allocated RawContext object. |
913 void StubCode::GenerateAllocateContextStub(Assembler* assembler) { | 914 void StubCode::GenerateAllocateContextStub(Assembler* assembler) { |
914 if (FLAG_inline_alloc) { | 915 if (FLAG_inline_alloc) { |
915 Label slow_case; | 916 Label slow_case; |
916 // First compute the rounded instance size. | 917 // First compute the rounded instance size. |
917 // R1: number of context variables. | 918 // R1: number of context variables. |
918 intptr_t fixed_size = sizeof(RawContext) + kObjectAlignment - 1; | 919 intptr_t fixed_size = sizeof(RawContext) + kObjectAlignment - 1; |
919 __ LoadImmediate(R2, fixed_size, kNoPP); | 920 __ LoadImmediate(R2, fixed_size); |
920 __ add(R2, R2, Operand(R1, LSL, 3)); | 921 __ add(R2, R2, Operand(R1, LSL, 3)); |
921 ASSERT(kSmiTagShift == 1); | 922 ASSERT(kSmiTagShift == 1); |
922 __ andi(R2, R2, Immediate(~(kObjectAlignment - 1))); | 923 __ andi(R2, R2, Immediate(~(kObjectAlignment - 1))); |
923 | 924 |
924 // Now allocate the object. | 925 // Now allocate the object. |
925 // R1: number of context variables. | 926 // R1: number of context variables. |
926 // R2: object size. | 927 // R2: object size. |
927 const intptr_t cid = kContextCid; | 928 const intptr_t cid = kContextCid; |
928 Heap::Space space = Heap::SpaceForAllocation(cid); | 929 Heap::Space space = Heap::SpaceForAllocation(cid); |
929 __ LoadIsolate(R5); | 930 __ LoadIsolate(R5); |
(...skipping 16 matching lines...) Expand all Loading... |
946 | 947 |
947 // Successfully allocated the object, now update top to point to | 948 // Successfully allocated the object, now update top to point to |
948 // next object start and initialize the object. | 949 // next object start and initialize the object. |
949 // R0: new object. | 950 // R0: new object. |
950 // R1: number of context variables. | 951 // R1: number of context variables. |
951 // R2: object size. | 952 // R2: object size. |
952 // R3: next object start. | 953 // R3: next object start. |
953 // R5: heap. | 954 // R5: heap. |
954 __ str(R3, Address(R5, Heap::TopOffset(space))); | 955 __ str(R3, Address(R5, Heap::TopOffset(space))); |
955 __ add(R0, R0, Operand(kHeapObjectTag)); | 956 __ add(R0, R0, Operand(kHeapObjectTag)); |
956 __ UpdateAllocationStatsWithSize(cid, R2, kNoPP, space, | 957 __ UpdateAllocationStatsWithSize(cid, R2, space, |
957 /* inline_isolate = */ false); | 958 /* inline_isolate = */ false); |
958 | 959 |
959 // Calculate the size tag. | 960 // Calculate the size tag. |
960 // R0: new object. | 961 // R0: new object. |
961 // R1: number of context variables. | 962 // R1: number of context variables. |
962 // R2: object size. | 963 // R2: object size. |
963 const intptr_t shift = RawObject::kSizeTagPos - kObjectAlignmentLog2; | 964 const intptr_t shift = RawObject::kSizeTagPos - kObjectAlignmentLog2; |
964 __ CompareImmediate(R2, RawObject::SizeTag::kMaxSizeTag, kNoPP); | 965 __ CompareImmediate(R2, RawObject::SizeTag::kMaxSizeTag); |
965 // If no size tag overflow, shift R2 left, else set R2 to zero. | 966 // If no size tag overflow, shift R2 left, else set R2 to zero. |
966 __ LslImmediate(TMP, R2, shift); | 967 __ LslImmediate(TMP, R2, shift); |
967 __ csel(R2, TMP, R2, LS); | 968 __ csel(R2, TMP, R2, LS); |
968 __ csel(R2, ZR, R2, HI); | 969 __ csel(R2, ZR, R2, HI); |
969 | 970 |
970 // Get the class index and insert it into the tags. | 971 // Get the class index and insert it into the tags. |
971 // R2: size and bit tags. | 972 // R2: size and bit tags. |
972 __ LoadImmediate(TMP, RawObject::ClassIdTag::encode(cid), kNoPP); | 973 __ LoadImmediate(TMP, RawObject::ClassIdTag::encode(cid)); |
973 __ orr(R2, R2, Operand(TMP)); | 974 __ orr(R2, R2, Operand(TMP)); |
974 __ StoreFieldToOffset(R2, R0, Context::tags_offset(), kNoPP); | 975 __ StoreFieldToOffset(R2, R0, Context::tags_offset()); |
975 | 976 |
976 // Setup up number of context variables field. | 977 // Setup up number of context variables field. |
977 // R0: new object. | 978 // R0: new object. |
978 // R1: number of context variables as integer value (not object). | 979 // R1: number of context variables as integer value (not object). |
979 __ StoreFieldToOffset(R1, R0, Context::num_variables_offset(), kNoPP); | 980 __ StoreFieldToOffset(R1, R0, Context::num_variables_offset()); |
980 | 981 |
981 // Setup the parent field. | 982 // Setup the parent field. |
982 // R0: new object. | 983 // R0: new object. |
983 // R1: number of context variables. | 984 // R1: number of context variables. |
984 __ LoadObject(R2, Object::null_object(), PP); | 985 __ LoadObject(R2, Object::null_object()); |
985 __ StoreFieldToOffset(R2, R0, Context::parent_offset(), kNoPP); | 986 __ StoreFieldToOffset(R2, R0, Context::parent_offset()); |
986 | 987 |
987 // Initialize the context variables. | 988 // Initialize the context variables. |
988 // R0: new object. | 989 // R0: new object. |
989 // R1: number of context variables. | 990 // R1: number of context variables. |
990 // R2: raw null. | 991 // R2: raw null. |
991 Label loop, done; | 992 Label loop, done; |
992 __ AddImmediate( | 993 __ AddImmediate( |
993 R3, R0, Context::variable_offset(0) - kHeapObjectTag, kNoPP); | 994 R3, R0, Context::variable_offset(0) - kHeapObjectTag); |
994 __ Bind(&loop); | 995 __ Bind(&loop); |
995 __ subs(R1, R1, Operand(1)); | 996 __ subs(R1, R1, Operand(1)); |
996 __ b(&done, MI); | 997 __ b(&done, MI); |
997 __ str(R2, Address(R3, R1, UXTX, Address::Scaled)); | 998 __ str(R2, Address(R3, R1, UXTX, Address::Scaled)); |
998 __ b(&loop, NE); // Loop if R1 not zero. | 999 __ b(&loop, NE); // Loop if R1 not zero. |
999 __ Bind(&done); | 1000 __ Bind(&done); |
1000 | 1001 |
1001 // Done allocating and initializing the context. | 1002 // Done allocating and initializing the context. |
1002 // R0: new object. | 1003 // R0: new object. |
1003 __ ret(); | 1004 __ ret(); |
1004 | 1005 |
1005 __ Bind(&slow_case); | 1006 __ Bind(&slow_case); |
1006 } | 1007 } |
1007 // Create a stub frame as we are pushing some objects on the stack before | 1008 // Create a stub frame as we are pushing some objects on the stack before |
1008 // calling into the runtime. | 1009 // calling into the runtime. |
1009 __ EnterStubFrame(); | 1010 __ EnterStubFrame(); |
1010 // Setup space on stack for return value. | 1011 // Setup space on stack for return value. |
1011 __ SmiTag(R1); | 1012 __ SmiTag(R1); |
1012 __ PushObject(Object::null_object(), PP); | 1013 __ PushObject(Object::null_object()); |
1013 __ Push(R1); | 1014 __ Push(R1); |
1014 __ CallRuntime(kAllocateContextRuntimeEntry, 1); // Allocate context. | 1015 __ CallRuntime(kAllocateContextRuntimeEntry, 1); // Allocate context. |
1015 __ Drop(1); // Pop number of context variables argument. | 1016 __ Drop(1); // Pop number of context variables argument. |
1016 __ Pop(R0); // Pop the new context object. | 1017 __ Pop(R0); // Pop the new context object. |
1017 // R0: new object | 1018 // R0: new object |
1018 // Restore the frame pointer. | 1019 // Restore the frame pointer. |
1019 __ LeaveStubFrame(); | 1020 __ LeaveStubFrame(); |
1020 __ ret(); | 1021 __ ret(); |
1021 } | 1022 } |
1022 | 1023 |
1023 | 1024 |
1024 DECLARE_LEAF_RUNTIME_ENTRY(void, StoreBufferBlockProcess, Isolate* isolate); | 1025 DECLARE_LEAF_RUNTIME_ENTRY(void, StoreBufferBlockProcess, Isolate* isolate); |
1025 | 1026 |
1026 // Helper stub to implement Assembler::StoreIntoObject. | 1027 // Helper stub to implement Assembler::StoreIntoObject. |
1027 // Input parameters: | 1028 // Input parameters: |
1028 // R0: Address being stored | 1029 // R0: Address being stored |
1029 void StubCode::GenerateUpdateStoreBufferStub(Assembler* assembler) { | 1030 void StubCode::GenerateUpdateStoreBufferStub(Assembler* assembler) { |
1030 Label add_to_buffer; | 1031 Label add_to_buffer; |
1031 // Check whether this object has already been remembered. Skip adding to the | 1032 // Check whether this object has already been remembered. Skip adding to the |
1032 // store buffer if the object is in the store buffer already. | 1033 // store buffer if the object is in the store buffer already. |
1033 __ LoadFieldFromOffset(TMP, R0, Object::tags_offset(), kNoPP); | 1034 __ LoadFieldFromOffset(TMP, R0, Object::tags_offset()); |
1034 __ tsti(TMP, Immediate(1 << RawObject::kRememberedBit)); | 1035 __ tsti(TMP, Immediate(1 << RawObject::kRememberedBit)); |
1035 __ b(&add_to_buffer, EQ); | 1036 __ b(&add_to_buffer, EQ); |
1036 __ ret(); | 1037 __ ret(); |
1037 | 1038 |
1038 __ Bind(&add_to_buffer); | 1039 __ Bind(&add_to_buffer); |
1039 // Save values being destroyed. | 1040 // Save values being destroyed. |
1040 __ Push(R1); | 1041 __ Push(R1); |
1041 __ Push(R2); | 1042 __ Push(R2); |
1042 __ Push(R3); | 1043 __ Push(R3); |
1043 | 1044 |
1044 __ orri(R2, TMP, Immediate(1 << RawObject::kRememberedBit)); | 1045 __ orri(R2, TMP, Immediate(1 << RawObject::kRememberedBit)); |
1045 __ StoreFieldToOffset(R2, R0, Object::tags_offset(), kNoPP); | 1046 __ StoreFieldToOffset(R2, R0, Object::tags_offset()); |
1046 | 1047 |
1047 // Load the StoreBuffer block out of the thread. Then load top_ out of the | 1048 // Load the StoreBuffer block out of the thread. Then load top_ out of the |
1048 // StoreBufferBlock and add the address to the pointers_. | 1049 // StoreBufferBlock and add the address to the pointers_. |
1049 __ LoadFromOffset(R1, THR, Thread::store_buffer_block_offset(), kNoPP); | 1050 __ LoadFromOffset(R1, THR, Thread::store_buffer_block_offset()); |
1050 __ LoadFromOffset(R2, R1, StoreBufferBlock::top_offset(), | 1051 __ LoadFromOffset(R2, R1, StoreBufferBlock::top_offset(), kUnsignedWord); |
1051 kNoPP, kUnsignedWord); | |
1052 __ add(R3, R1, Operand(R2, LSL, 3)); | 1052 __ add(R3, R1, Operand(R2, LSL, 3)); |
1053 __ StoreToOffset(R0, R3, StoreBufferBlock::pointers_offset(), kNoPP); | 1053 __ StoreToOffset(R0, R3, StoreBufferBlock::pointers_offset()); |
1054 | 1054 |
1055 // Increment top_ and check for overflow. | 1055 // Increment top_ and check for overflow. |
1056 // R2: top_. | 1056 // R2: top_. |
1057 // R1: StoreBufferBlock. | 1057 // R1: StoreBufferBlock. |
1058 Label L; | 1058 Label L; |
1059 __ add(R2, R2, Operand(1)); | 1059 __ add(R2, R2, Operand(1)); |
1060 __ StoreToOffset(R2, R1, StoreBufferBlock::top_offset(), | 1060 __ StoreToOffset(R2, R1, StoreBufferBlock::top_offset(), kUnsignedWord); |
1061 kNoPP, kUnsignedWord); | 1061 __ CompareImmediate(R2, StoreBufferBlock::kSize); |
1062 __ CompareImmediate(R2, StoreBufferBlock::kSize, kNoPP); | |
1063 // Restore values. | 1062 // Restore values. |
1064 __ Pop(R3); | 1063 __ Pop(R3); |
1065 __ Pop(R2); | 1064 __ Pop(R2); |
1066 __ Pop(R1); | 1065 __ Pop(R1); |
1067 __ b(&L, EQ); | 1066 __ b(&L, EQ); |
1068 __ ret(); | 1067 __ ret(); |
1069 | 1068 |
1070 // Handle overflow: Call the runtime leaf function. | 1069 // Handle overflow: Call the runtime leaf function. |
1071 __ Bind(&L); | 1070 __ Bind(&L); |
1072 // Setup frame, push callee-saved registers. | 1071 // Setup frame, push callee-saved registers. |
(...skipping 30 matching lines...) Expand all Loading... |
1103 // R1: instantiated type arguments. | 1102 // R1: instantiated type arguments. |
1104 } | 1103 } |
1105 if (FLAG_inline_alloc && Heap::IsAllocatableInNewSpace(instance_size) && | 1104 if (FLAG_inline_alloc && Heap::IsAllocatableInNewSpace(instance_size) && |
1106 !cls.trace_allocation()) { | 1105 !cls.trace_allocation()) { |
1107 Label slow_case; | 1106 Label slow_case; |
1108 // Allocate the object and update top to point to | 1107 // Allocate the object and update top to point to |
1109 // next object start and initialize the allocated object. | 1108 // next object start and initialize the allocated object. |
1110 // R1: instantiated type arguments (if is_cls_parameterized). | 1109 // R1: instantiated type arguments (if is_cls_parameterized). |
1111 Heap* heap = Isolate::Current()->heap(); | 1110 Heap* heap = Isolate::Current()->heap(); |
1112 Heap::Space space = Heap::SpaceForAllocation(cls.id()); | 1111 Heap::Space space = Heap::SpaceForAllocation(cls.id()); |
1113 __ LoadImmediate(R5, heap->TopAddress(space), kNoPP); | 1112 __ LoadImmediate(R5, heap->TopAddress(space)); |
1114 __ ldr(R2, Address(R5)); | 1113 __ ldr(R2, Address(R5)); |
1115 __ AddImmediate(R3, R2, instance_size, kNoPP); | 1114 __ AddImmediate(R3, R2, instance_size); |
1116 // Check if the allocation fits into the remaining space. | 1115 // Check if the allocation fits into the remaining space. |
1117 // R2: potential new object start. | 1116 // R2: potential new object start. |
1118 // R3: potential next object start. | 1117 // R3: potential next object start. |
1119 __ LoadImmediate(TMP, heap->EndAddress(space), kNoPP); | 1118 __ LoadImmediate(TMP, heap->EndAddress(space)); |
1120 __ ldr(TMP, Address(TMP)); | 1119 __ ldr(TMP, Address(TMP)); |
1121 __ CompareRegisters(R3, TMP); | 1120 __ CompareRegisters(R3, TMP); |
1122 if (FLAG_use_slow_path) { | 1121 if (FLAG_use_slow_path) { |
1123 __ b(&slow_case); | 1122 __ b(&slow_case); |
1124 } else { | 1123 } else { |
1125 __ b(&slow_case, CS); // Unsigned higher or equal. | 1124 __ b(&slow_case, CS); // Unsigned higher or equal. |
1126 } | 1125 } |
1127 __ str(R3, Address(R5)); | 1126 __ str(R3, Address(R5)); |
1128 __ UpdateAllocationStats(cls.id(), kNoPP, space); | 1127 __ UpdateAllocationStats(cls.id(), space); |
1129 | 1128 |
1130 // R2: new object start. | 1129 // R2: new object start. |
1131 // R3: next object start. | 1130 // R3: next object start. |
1132 // R1: new object type arguments (if is_cls_parameterized). | 1131 // R1: new object type arguments (if is_cls_parameterized). |
1133 // Set the tags. | 1132 // Set the tags. |
1134 uword tags = 0; | 1133 uword tags = 0; |
1135 tags = RawObject::SizeTag::update(instance_size, tags); | 1134 tags = RawObject::SizeTag::update(instance_size, tags); |
1136 ASSERT(cls.id() != kIllegalCid); | 1135 ASSERT(cls.id() != kIllegalCid); |
1137 tags = RawObject::ClassIdTag::update(cls.id(), tags); | 1136 tags = RawObject::ClassIdTag::update(cls.id(), tags); |
1138 __ LoadImmediate(R0, tags, kNoPP); | 1137 __ LoadImmediate(R0, tags); |
1139 __ StoreToOffset(R0, R2, Instance::tags_offset(), kNoPP); | 1138 __ StoreToOffset(R0, R2, Instance::tags_offset()); |
1140 | 1139 |
1141 // Initialize the remaining words of the object. | 1140 // Initialize the remaining words of the object. |
1142 __ LoadObject(R0, Object::null_object(), PP); | 1141 __ LoadObject(R0, Object::null_object()); |
1143 | 1142 |
1144 // R0: raw null. | 1143 // R0: raw null. |
1145 // R2: new object start. | 1144 // R2: new object start. |
1146 // R3: next object start. | 1145 // R3: next object start. |
1147 // R1: new object type arguments (if is_cls_parameterized). | 1146 // R1: new object type arguments (if is_cls_parameterized). |
1148 // First try inlining the initialization without a loop. | 1147 // First try inlining the initialization without a loop. |
1149 if (instance_size < (kInlineInstanceSize * kWordSize)) { | 1148 if (instance_size < (kInlineInstanceSize * kWordSize)) { |
1150 // Check if the object contains any non-header fields. | 1149 // Check if the object contains any non-header fields. |
1151 // Small objects are initialized using a consecutive set of writes. | 1150 // Small objects are initialized using a consecutive set of writes. |
1152 for (intptr_t current_offset = Instance::NextFieldOffset(); | 1151 for (intptr_t current_offset = Instance::NextFieldOffset(); |
1153 current_offset < instance_size; | 1152 current_offset < instance_size; |
1154 current_offset += kWordSize) { | 1153 current_offset += kWordSize) { |
1155 __ StoreToOffset(R0, R2, current_offset, kNoPP); | 1154 __ StoreToOffset(R0, R2, current_offset); |
1156 } | 1155 } |
1157 } else { | 1156 } else { |
1158 __ AddImmediate(R4, R2, Instance::NextFieldOffset(), kNoPP); | 1157 __ AddImmediate(R4, R2, Instance::NextFieldOffset()); |
1159 // Loop until the whole object is initialized. | 1158 // Loop until the whole object is initialized. |
1160 // R0: raw null. | 1159 // R0: raw null. |
1161 // R2: new object. | 1160 // R2: new object. |
1162 // R3: next object start. | 1161 // R3: next object start. |
1163 // R4: next word to be initialized. | 1162 // R4: next word to be initialized. |
1164 // R1: new object type arguments (if is_cls_parameterized). | 1163 // R1: new object type arguments (if is_cls_parameterized). |
1165 Label init_loop; | 1164 Label init_loop; |
1166 Label done; | 1165 Label done; |
1167 __ Bind(&init_loop); | 1166 __ Bind(&init_loop); |
1168 __ CompareRegisters(R4, R3); | 1167 __ CompareRegisters(R4, R3); |
1169 __ b(&done, CS); | 1168 __ b(&done, CS); |
1170 __ str(R0, Address(R4)); | 1169 __ str(R0, Address(R4)); |
1171 __ AddImmediate(R4, R4, kWordSize, kNoPP); | 1170 __ AddImmediate(R4, R4, kWordSize); |
1172 __ b(&init_loop); | 1171 __ b(&init_loop); |
1173 __ Bind(&done); | 1172 __ Bind(&done); |
1174 } | 1173 } |
1175 if (is_cls_parameterized) { | 1174 if (is_cls_parameterized) { |
1176 // R1: new object type arguments. | 1175 // R1: new object type arguments. |
1177 // Set the type arguments in the new object. | 1176 // Set the type arguments in the new object. |
1178 __ StoreToOffset(R1, R2, cls.type_arguments_field_offset(), kNoPP); | 1177 __ StoreToOffset(R1, R2, cls.type_arguments_field_offset()); |
1179 } | 1178 } |
1180 // Done allocating and initializing the instance. | 1179 // Done allocating and initializing the instance. |
1181 // R2: new object still missing its heap tag. | 1180 // R2: new object still missing its heap tag. |
1182 __ add(R0, R2, Operand(kHeapObjectTag)); | 1181 __ add(R0, R2, Operand(kHeapObjectTag)); |
1183 // R0: new object. | 1182 // R0: new object. |
1184 __ ret(); | 1183 __ ret(); |
1185 | 1184 |
1186 __ Bind(&slow_case); | 1185 __ Bind(&slow_case); |
1187 } | 1186 } |
1188 // If is_cls_parameterized: | 1187 // If is_cls_parameterized: |
1189 // R1: new object type arguments. | 1188 // R1: new object type arguments. |
1190 // Create a stub frame as we are pushing some objects on the stack before | 1189 // Create a stub frame as we are pushing some objects on the stack before |
1191 // calling into the runtime. | 1190 // calling into the runtime. |
1192 __ EnterStubFrame(); // Uses pool pointer to pass cls to runtime. | 1191 __ EnterStubFrame(); // Uses pool pointer to pass cls to runtime. |
1193 // Setup space on stack for return value. | 1192 // Setup space on stack for return value. |
1194 __ PushObject(Object::null_object(), PP); | 1193 __ PushObject(Object::null_object()); |
1195 __ PushObject(cls, PP); // Push class of object to be allocated. | 1194 __ PushObject(cls); // Push class of object to be allocated. |
1196 if (is_cls_parameterized) { | 1195 if (is_cls_parameterized) { |
1197 // Push type arguments. | 1196 // Push type arguments. |
1198 __ Push(R1); | 1197 __ Push(R1); |
1199 } else { | 1198 } else { |
1200 // Push null type arguments. | 1199 // Push null type arguments. |
1201 __ PushObject(Object::null_object(), PP); | 1200 __ PushObject(Object::null_object()); |
1202 } | 1201 } |
1203 __ CallRuntime(kAllocateObjectRuntimeEntry, 2); // Allocate object. | 1202 __ CallRuntime(kAllocateObjectRuntimeEntry, 2); // Allocate object. |
1204 __ Drop(2); // Pop arguments. | 1203 __ Drop(2); // Pop arguments. |
1205 __ Pop(R0); // Pop result (newly allocated object). | 1204 __ Pop(R0); // Pop result (newly allocated object). |
1206 // R0: new object | 1205 // R0: new object |
1207 // Restore the frame pointer. | 1206 // Restore the frame pointer. |
1208 __ LeaveStubFrame(); | 1207 __ LeaveStubFrame(); |
1209 __ ret(); | 1208 __ ret(); |
1210 *patch_code_pc_offset = assembler->CodeSize(); | 1209 *patch_code_pc_offset = assembler->CodeSize(); |
1211 __ BranchPatchable(&StubCode::FixAllocationStubTargetLabel()); | 1210 __ BranchPatchable(&StubCode::FixAllocationStubTargetLabel()); |
1212 } | 1211 } |
1213 | 1212 |
1214 | 1213 |
1215 // Called for invoking "dynamic noSuchMethod(Invocation invocation)" function | 1214 // Called for invoking "dynamic noSuchMethod(Invocation invocation)" function |
1216 // from the entry code of a dart function after an error in passed argument | 1215 // from the entry code of a dart function after an error in passed argument |
1217 // name or number is detected. | 1216 // name or number is detected. |
1218 // Input parameters: | 1217 // Input parameters: |
1219 // LR : return address. | 1218 // LR : return address. |
1220 // SP : address of last argument. | 1219 // SP : address of last argument. |
1221 // R4: arguments descriptor array. | 1220 // R4: arguments descriptor array. |
1222 void StubCode::GenerateCallClosureNoSuchMethodStub(Assembler* assembler) { | 1221 void StubCode::GenerateCallClosureNoSuchMethodStub(Assembler* assembler) { |
1223 __ EnterStubFrame(); | 1222 __ EnterStubFrame(); |
1224 | 1223 |
1225 // Load the receiver. | 1224 // Load the receiver. |
1226 __ LoadFieldFromOffset(R2, R4, ArgumentsDescriptor::count_offset(), kNoPP); | 1225 __ LoadFieldFromOffset(R2, R4, ArgumentsDescriptor::count_offset()); |
1227 __ add(TMP, FP, Operand(R2, LSL, 2)); // R2 is Smi. | 1226 __ add(TMP, FP, Operand(R2, LSL, 2)); // R2 is Smi. |
1228 __ LoadFromOffset(R6, TMP, kParamEndSlotFromFp * kWordSize, kNoPP); | 1227 __ LoadFromOffset(R6, TMP, kParamEndSlotFromFp * kWordSize); |
1229 | 1228 |
1230 // Push space for the return value. | 1229 // Push space for the return value. |
1231 // Push the receiver. | 1230 // Push the receiver. |
1232 // Push arguments descriptor array. | 1231 // Push arguments descriptor array. |
1233 __ PushObject(Object::null_object(), PP); | 1232 __ PushObject(Object::null_object()); |
1234 __ Push(R6); | 1233 __ Push(R6); |
1235 __ Push(R4); | 1234 __ Push(R4); |
1236 | 1235 |
1237 // R2: Smi-tagged arguments array length. | 1236 // R2: Smi-tagged arguments array length. |
1238 PushArgumentsArray(assembler); | 1237 PushArgumentsArray(assembler); |
1239 | 1238 |
1240 const intptr_t kNumArgs = 3; | 1239 const intptr_t kNumArgs = 3; |
1241 __ CallRuntime(kInvokeClosureNoSuchMethodRuntimeEntry, kNumArgs); | 1240 __ CallRuntime(kInvokeClosureNoSuchMethodRuntimeEntry, kNumArgs); |
1242 // noSuchMethod on closures always throws an error, so it will never return. | 1241 // noSuchMethod on closures always throws an error, so it will never return. |
1243 __ brk(0); | 1242 __ brk(0); |
(...skipping 13 matching lines...) Expand all Loading... |
1257 __ Push(R5); // Preserve. | 1256 __ Push(R5); // Preserve. |
1258 __ Push(ic_reg); // Argument. | 1257 __ Push(ic_reg); // Argument. |
1259 __ Push(func_reg); // Argument. | 1258 __ Push(func_reg); // Argument. |
1260 __ CallRuntime(kTraceICCallRuntimeEntry, 2); | 1259 __ CallRuntime(kTraceICCallRuntimeEntry, 2); |
1261 __ Drop(2); // Discard argument; | 1260 __ Drop(2); // Discard argument; |
1262 __ Pop(R5); // Restore. | 1261 __ Pop(R5); // Restore. |
1263 __ Pop(R6); // Restore. | 1262 __ Pop(R6); // Restore. |
1264 __ LeaveStubFrame(); | 1263 __ LeaveStubFrame(); |
1265 } | 1264 } |
1266 __ LoadFieldFromOffset( | 1265 __ LoadFieldFromOffset( |
1267 R7, func_reg, Function::usage_counter_offset(), kNoPP, kWord); | 1266 R7, func_reg, Function::usage_counter_offset(), kWord); |
1268 __ add(R7, R7, Operand(1)); | 1267 __ add(R7, R7, Operand(1)); |
1269 __ StoreFieldToOffset( | 1268 __ StoreFieldToOffset( |
1270 R7, func_reg, Function::usage_counter_offset(), kNoPP, kWord); | 1269 R7, func_reg, Function::usage_counter_offset(), kWord); |
1271 } | 1270 } |
1272 | 1271 |
1273 | 1272 |
1274 // Loads function into 'temp_reg'. | 1273 // Loads function into 'temp_reg'. |
1275 void StubCode::GenerateUsageCounterIncrement(Assembler* assembler, | 1274 void StubCode::GenerateUsageCounterIncrement(Assembler* assembler, |
1276 Register temp_reg) { | 1275 Register temp_reg) { |
1277 if (FLAG_optimization_counter_threshold >= 0) { | 1276 if (FLAG_optimization_counter_threshold >= 0) { |
1278 Register ic_reg = R5; | 1277 Register ic_reg = R5; |
1279 Register func_reg = temp_reg; | 1278 Register func_reg = temp_reg; |
1280 ASSERT(temp_reg == R6); | 1279 ASSERT(temp_reg == R6); |
1281 __ Comment("Increment function counter"); | 1280 __ Comment("Increment function counter"); |
1282 __ LoadFieldFromOffset(func_reg, ic_reg, ICData::owner_offset(), kNoPP); | 1281 __ LoadFieldFromOffset(func_reg, ic_reg, ICData::owner_offset()); |
1283 __ LoadFieldFromOffset( | 1282 __ LoadFieldFromOffset( |
1284 R7, func_reg, Function::usage_counter_offset(), kNoPP, kWord); | 1283 R7, func_reg, Function::usage_counter_offset(), kWord); |
1285 __ AddImmediate(R7, R7, 1, kNoPP); | 1284 __ AddImmediate(R7, R7, 1); |
1286 __ StoreFieldToOffset( | 1285 __ StoreFieldToOffset( |
1287 R7, func_reg, Function::usage_counter_offset(), kNoPP, kWord); | 1286 R7, func_reg, Function::usage_counter_offset(), kWord); |
1288 } | 1287 } |
1289 } | 1288 } |
1290 | 1289 |
1291 | 1290 |
1292 // Note: R5 must be preserved. | 1291 // Note: R5 must be preserved. |
1293 // Attempt a quick Smi operation for known operations ('kind'). The ICData | 1292 // Attempt a quick Smi operation for known operations ('kind'). The ICData |
1294 // must have been primed with a Smi/Smi check that will be used for counting | 1293 // must have been primed with a Smi/Smi check that will be used for counting |
1295 // the invocations. | 1294 // the invocations. |
1296 static void EmitFastSmiOp(Assembler* assembler, | 1295 static void EmitFastSmiOp(Assembler* assembler, |
1297 Token::Kind kind, | 1296 Token::Kind kind, |
(...skipping 16 matching lines...) Expand all Loading... |
1314 __ b(not_smi_or_overflow, VS); // Branch if overflow. | 1313 __ b(not_smi_or_overflow, VS); // Branch if overflow. |
1315 break; | 1314 break; |
1316 } | 1315 } |
1317 case Token::kSUB: { | 1316 case Token::kSUB: { |
1318 __ subs(R0, R1, Operand(R0)); // Subtract. | 1317 __ subs(R0, R1, Operand(R0)); // Subtract. |
1319 __ b(not_smi_or_overflow, VS); // Branch if overflow. | 1318 __ b(not_smi_or_overflow, VS); // Branch if overflow. |
1320 break; | 1319 break; |
1321 } | 1320 } |
1322 case Token::kEQ: { | 1321 case Token::kEQ: { |
1323 __ CompareRegisters(R0, R1); | 1322 __ CompareRegisters(R0, R1); |
1324 __ LoadObject(R0, Bool::True(), PP); | 1323 __ LoadObject(R0, Bool::True()); |
1325 __ LoadObject(R1, Bool::False(), PP); | 1324 __ LoadObject(R1, Bool::False()); |
1326 __ csel(R0, R1, R0, NE); | 1325 __ csel(R0, R1, R0, NE); |
1327 break; | 1326 break; |
1328 } | 1327 } |
1329 default: UNIMPLEMENTED(); | 1328 default: UNIMPLEMENTED(); |
1330 } | 1329 } |
1331 | 1330 |
1332 if (should_update_result_range) { | 1331 if (should_update_result_range) { |
1333 Label done; | 1332 Label done; |
1334 __ UpdateRangeFeedback(R0, 2, R5, R1, R6, &done); | 1333 __ UpdateRangeFeedback(R0, 2, R5, R1, R6, &done); |
1335 __ Bind(&done); | 1334 __ Bind(&done); |
1336 } | 1335 } |
1337 | 1336 |
1338 // R5: IC data object (preserved). | 1337 // R5: IC data object (preserved). |
1339 __ LoadFieldFromOffset(R6, R5, ICData::ic_data_offset(), kNoPP); | 1338 __ LoadFieldFromOffset(R6, R5, ICData::ic_data_offset()); |
1340 // R6: ic_data_array with check entries: classes and target functions. | 1339 // R6: ic_data_array with check entries: classes and target functions. |
1341 __ AddImmediate(R6, R6, Array::data_offset() - kHeapObjectTag, kNoPP); | 1340 __ AddImmediate(R6, R6, Array::data_offset() - kHeapObjectTag); |
1342 // R6: points directly to the first ic data array element. | 1341 // R6: points directly to the first ic data array element. |
1343 #if defined(DEBUG) | 1342 #if defined(DEBUG) |
1344 // Check that first entry is for Smi/Smi. | 1343 // Check that first entry is for Smi/Smi. |
1345 Label error, ok; | 1344 Label error, ok; |
1346 const intptr_t imm_smi_cid = reinterpret_cast<intptr_t>(Smi::New(kSmiCid)); | 1345 const intptr_t imm_smi_cid = reinterpret_cast<intptr_t>(Smi::New(kSmiCid)); |
1347 __ ldr(R1, Address(R6, 0)); | 1346 __ ldr(R1, Address(R6, 0)); |
1348 __ CompareImmediate(R1, imm_smi_cid, kNoPP); | 1347 __ CompareImmediate(R1, imm_smi_cid); |
1349 __ b(&error, NE); | 1348 __ b(&error, NE); |
1350 __ ldr(R1, Address(R6, kWordSize)); | 1349 __ ldr(R1, Address(R6, kWordSize)); |
1351 __ CompareImmediate(R1, imm_smi_cid, kNoPP); | 1350 __ CompareImmediate(R1, imm_smi_cid); |
1352 __ b(&ok, EQ); | 1351 __ b(&ok, EQ); |
1353 __ Bind(&error); | 1352 __ Bind(&error); |
1354 __ Stop("Incorrect IC data"); | 1353 __ Stop("Incorrect IC data"); |
1355 __ Bind(&ok); | 1354 __ Bind(&ok); |
1356 #endif | 1355 #endif |
1357 if (FLAG_optimization_counter_threshold >= 0) { | 1356 if (FLAG_optimization_counter_threshold >= 0) { |
1358 const intptr_t count_offset = ICData::CountIndexFor(num_args) * kWordSize; | 1357 const intptr_t count_offset = ICData::CountIndexFor(num_args) * kWordSize; |
1359 // Update counter. | 1358 // Update counter. |
1360 __ LoadFromOffset(R1, R6, count_offset, kNoPP); | 1359 __ LoadFromOffset(R1, R6, count_offset); |
1361 __ adds(R1, R1, Operand(Smi::RawValue(1))); | 1360 __ adds(R1, R1, Operand(Smi::RawValue(1))); |
1362 __ LoadImmediate(R2, Smi::RawValue(Smi::kMaxValue), kNoPP); | 1361 __ LoadImmediate(R2, Smi::RawValue(Smi::kMaxValue)); |
1363 __ csel(R1, R2, R1, VS); // Overflow. | 1362 __ csel(R1, R2, R1, VS); // Overflow. |
1364 __ StoreToOffset(R1, R6, count_offset, kNoPP); | 1363 __ StoreToOffset(R1, R6, count_offset); |
1365 } | 1364 } |
1366 | 1365 |
1367 __ ret(); | 1366 __ ret(); |
1368 } | 1367 } |
1369 | 1368 |
1370 | 1369 |
1371 // Generate inline cache check for 'num_args'. | 1370 // Generate inline cache check for 'num_args'. |
1372 // LR: return address. | 1371 // LR: return address. |
1373 // R5: inline cache data object. | 1372 // R5: inline cache data object. |
1374 // Control flow: | 1373 // Control flow: |
1375 // - If receiver is null -> jump to IC miss. | 1374 // - If receiver is null -> jump to IC miss. |
1376 // - If receiver is Smi -> load Smi class. | 1375 // - If receiver is Smi -> load Smi class. |
1377 // - If receiver is not-Smi -> load receiver's class. | 1376 // - If receiver is not-Smi -> load receiver's class. |
1378 // - Check if 'num_args' (including receiver) match any IC data group. | 1377 // - Check if 'num_args' (including receiver) match any IC data group. |
1379 // - Match found -> jump to target. | 1378 // - Match found -> jump to target. |
1380 // - Match not found -> jump to IC miss. | 1379 // - Match not found -> jump to IC miss. |
1381 void StubCode::GenerateNArgsCheckInlineCacheStub( | 1380 void StubCode::GenerateNArgsCheckInlineCacheStub( |
1382 Assembler* assembler, | 1381 Assembler* assembler, |
1383 intptr_t num_args, | 1382 intptr_t num_args, |
1384 const RuntimeEntry& handle_ic_miss, | 1383 const RuntimeEntry& handle_ic_miss, |
1385 Token::Kind kind, | 1384 Token::Kind kind, |
1386 RangeCollectionMode range_collection_mode, | 1385 RangeCollectionMode range_collection_mode, |
1387 bool optimized) { | 1386 bool optimized) { |
1388 ASSERT(num_args > 0); | 1387 ASSERT(num_args > 0); |
1389 #if defined(DEBUG) | 1388 #if defined(DEBUG) |
1390 { Label ok; | 1389 { Label ok; |
1391 // Check that the IC data array has NumArgsTested() == num_args. | 1390 // Check that the IC data array has NumArgsTested() == num_args. |
1392 // 'NumArgsTested' is stored in the least significant bits of 'state_bits'. | 1391 // 'NumArgsTested' is stored in the least significant bits of 'state_bits'. |
1393 __ LoadFromOffset(R6, R5, ICData::state_bits_offset() - kHeapObjectTag, | 1392 __ LoadFromOffset(R6, R5, ICData::state_bits_offset() - kHeapObjectTag, |
1394 kNoPP, kUnsignedWord); | 1393 kUnsignedWord); |
1395 ASSERT(ICData::NumArgsTestedShift() == 0); // No shift needed. | 1394 ASSERT(ICData::NumArgsTestedShift() == 0); // No shift needed. |
1396 __ andi(R6, R6, Immediate(ICData::NumArgsTestedMask())); | 1395 __ andi(R6, R6, Immediate(ICData::NumArgsTestedMask())); |
1397 __ CompareImmediate(R6, num_args, kNoPP); | 1396 __ CompareImmediate(R6, num_args); |
1398 __ b(&ok, EQ); | 1397 __ b(&ok, EQ); |
1399 __ Stop("Incorrect stub for IC data"); | 1398 __ Stop("Incorrect stub for IC data"); |
1400 __ Bind(&ok); | 1399 __ Bind(&ok); |
1401 } | 1400 } |
1402 #endif // DEBUG | 1401 #endif // DEBUG |
1403 | 1402 |
1404 Label stepping, done_stepping; | 1403 Label stepping, done_stepping; |
1405 if (FLAG_support_debugger && !optimized) { | 1404 if (FLAG_support_debugger && !optimized) { |
1406 __ Comment("Check single stepping"); | 1405 __ Comment("Check single stepping"); |
1407 __ LoadIsolate(R6); | 1406 __ LoadIsolate(R6); |
1408 __ LoadFromOffset( | 1407 __ LoadFromOffset( |
1409 R6, R6, Isolate::single_step_offset(), kNoPP, kUnsignedByte); | 1408 R6, R6, Isolate::single_step_offset(), kUnsignedByte); |
1410 __ CompareRegisters(R6, ZR); | 1409 __ CompareRegisters(R6, ZR); |
1411 __ b(&stepping, NE); | 1410 __ b(&stepping, NE); |
1412 __ Bind(&done_stepping); | 1411 __ Bind(&done_stepping); |
1413 } | 1412 } |
1414 | 1413 |
1415 __ Comment("Range feedback collection"); | 1414 __ Comment("Range feedback collection"); |
1416 Label not_smi_or_overflow; | 1415 Label not_smi_or_overflow; |
1417 if (range_collection_mode == kCollectRanges) { | 1416 if (range_collection_mode == kCollectRanges) { |
1418 ASSERT((num_args == 1) || (num_args == 2)); | 1417 ASSERT((num_args == 1) || (num_args == 2)); |
1419 if (num_args == 2) { | 1418 if (num_args == 2) { |
1420 __ ldr(R0, Address(SP, 1 * kWordSize)); | 1419 __ ldr(R0, Address(SP, 1 * kWordSize)); |
1421 __ UpdateRangeFeedback(R0, 0, R5, R1, R4, ¬_smi_or_overflow); | 1420 __ UpdateRangeFeedback(R0, 0, R5, R1, R4, ¬_smi_or_overflow); |
1422 } | 1421 } |
1423 | 1422 |
1424 __ ldr(R0, Address(SP, 0 * kWordSize)); | 1423 __ ldr(R0, Address(SP, 0 * kWordSize)); |
1425 __ UpdateRangeFeedback(R0, num_args - 1, R5, R1, R4, ¬_smi_or_overflow); | 1424 __ UpdateRangeFeedback(R0, num_args - 1, R5, R1, R4, ¬_smi_or_overflow); |
1426 } | 1425 } |
1427 if (kind != Token::kILLEGAL) { | 1426 if (kind != Token::kILLEGAL) { |
1428 EmitFastSmiOp(assembler, | 1427 EmitFastSmiOp(assembler, |
1429 kind, | 1428 kind, |
1430 num_args, | 1429 num_args, |
1431 ¬_smi_or_overflow, | 1430 ¬_smi_or_overflow, |
1432 (range_collection_mode == kCollectRanges)); | 1431 (range_collection_mode == kCollectRanges)); |
1433 } | 1432 } |
1434 __ Bind(¬_smi_or_overflow); | 1433 __ Bind(¬_smi_or_overflow); |
1435 | 1434 |
1436 __ Comment("Extract ICData initial values and receiver cid"); | 1435 __ Comment("Extract ICData initial values and receiver cid"); |
1437 // Load arguments descriptor into R4. | 1436 // Load arguments descriptor into R4. |
1438 __ LoadFieldFromOffset(R4, R5, ICData::arguments_descriptor_offset(), kNoPP); | 1437 __ LoadFieldFromOffset(R4, R5, ICData::arguments_descriptor_offset()); |
1439 // Loop that checks if there is an IC data match. | 1438 // Loop that checks if there is an IC data match. |
1440 Label loop, update, test, found; | 1439 Label loop, update, test, found; |
1441 // R5: IC data object (preserved). | 1440 // R5: IC data object (preserved). |
1442 __ LoadFieldFromOffset(R6, R5, ICData::ic_data_offset(), kNoPP); | 1441 __ LoadFieldFromOffset(R6, R5, ICData::ic_data_offset()); |
1443 // R6: ic_data_array with check entries: classes and target functions. | 1442 // R6: ic_data_array with check entries: classes and target functions. |
1444 __ AddImmediate(R6, R6, Array::data_offset() - kHeapObjectTag, kNoPP); | 1443 __ AddImmediate(R6, R6, Array::data_offset() - kHeapObjectTag); |
1445 // R6: points directly to the first ic data array element. | 1444 // R6: points directly to the first ic data array element. |
1446 | 1445 |
1447 // Get the receiver's class ID (first read number of arguments from | 1446 // Get the receiver's class ID (first read number of arguments from |
1448 // arguments descriptor array and then access the receiver from the stack). | 1447 // arguments descriptor array and then access the receiver from the stack). |
1449 __ LoadFieldFromOffset(R7, R4, ArgumentsDescriptor::count_offset(), kNoPP); | 1448 __ LoadFieldFromOffset(R7, R4, ArgumentsDescriptor::count_offset()); |
1450 __ SmiUntag(R7); // Untag so we can use the LSL 3 addressing mode. | 1449 __ SmiUntag(R7); // Untag so we can use the LSL 3 addressing mode. |
1451 __ sub(R7, R7, Operand(1)); | 1450 __ sub(R7, R7, Operand(1)); |
1452 | 1451 |
1453 // R0 <- [SP + (R7 << 3)] | 1452 // R0 <- [SP + (R7 << 3)] |
1454 __ ldr(R0, Address(SP, R7, UXTX, Address::Scaled)); | 1453 __ ldr(R0, Address(SP, R7, UXTX, Address::Scaled)); |
1455 __ LoadTaggedClassIdMayBeSmi(R0, R0); | 1454 __ LoadTaggedClassIdMayBeSmi(R0, R0); |
1456 | 1455 |
1457 // R7: argument_count - 1 (untagged). | 1456 // R7: argument_count - 1 (untagged). |
1458 // R0: receiver's class ID (smi). | 1457 // R0: receiver's class ID (smi). |
1459 __ ldr(R1, Address(R6)); // First class id (smi) to check. | 1458 __ ldr(R1, Address(R6)); // First class id (smi) to check. |
1460 __ b(&test); | 1459 __ b(&test); |
1461 | 1460 |
1462 __ Comment("ICData loop"); | 1461 __ Comment("ICData loop"); |
1463 __ Bind(&loop); | 1462 __ Bind(&loop); |
1464 for (int i = 0; i < num_args; i++) { | 1463 for (int i = 0; i < num_args; i++) { |
1465 if (i > 0) { | 1464 if (i > 0) { |
1466 // If not the first, load the next argument's class ID. | 1465 // If not the first, load the next argument's class ID. |
1467 __ AddImmediate(R0, R7, -i, kNoPP); | 1466 __ AddImmediate(R0, R7, -i); |
1468 // R0 <- [SP + (R0 << 3)] | 1467 // R0 <- [SP + (R0 << 3)] |
1469 __ ldr(R0, Address(SP, R0, UXTX, Address::Scaled)); | 1468 __ ldr(R0, Address(SP, R0, UXTX, Address::Scaled)); |
1470 __ LoadTaggedClassIdMayBeSmi(R0, R0); | 1469 __ LoadTaggedClassIdMayBeSmi(R0, R0); |
1471 // R0: next argument class ID (smi). | 1470 // R0: next argument class ID (smi). |
1472 __ LoadFromOffset(R1, R6, i * kWordSize, kNoPP); | 1471 __ LoadFromOffset(R1, R6, i * kWordSize); |
1473 // R1: next class ID to check (smi). | 1472 // R1: next class ID to check (smi). |
1474 } | 1473 } |
1475 __ CompareRegisters(R0, R1); // Class id match? | 1474 __ CompareRegisters(R0, R1); // Class id match? |
1476 if (i < (num_args - 1)) { | 1475 if (i < (num_args - 1)) { |
1477 __ b(&update, NE); // Continue. | 1476 __ b(&update, NE); // Continue. |
1478 } else { | 1477 } else { |
1479 // Last check, all checks before matched. | 1478 // Last check, all checks before matched. |
1480 __ b(&found, EQ); // Break. | 1479 __ b(&found, EQ); // Break. |
1481 } | 1480 } |
1482 } | 1481 } |
1483 __ Bind(&update); | 1482 __ Bind(&update); |
1484 // Reload receiver class ID. It has not been destroyed when num_args == 1. | 1483 // Reload receiver class ID. It has not been destroyed when num_args == 1. |
1485 if (num_args > 1) { | 1484 if (num_args > 1) { |
1486 __ ldr(R0, Address(SP, R7, UXTX, Address::Scaled)); | 1485 __ ldr(R0, Address(SP, R7, UXTX, Address::Scaled)); |
1487 __ LoadTaggedClassIdMayBeSmi(R0, R0); | 1486 __ LoadTaggedClassIdMayBeSmi(R0, R0); |
1488 } | 1487 } |
1489 | 1488 |
1490 const intptr_t entry_size = ICData::TestEntryLengthFor(num_args) * kWordSize; | 1489 const intptr_t entry_size = ICData::TestEntryLengthFor(num_args) * kWordSize; |
1491 __ AddImmediate(R6, R6, entry_size, kNoPP); // Next entry. | 1490 __ AddImmediate(R6, R6, entry_size); // Next entry. |
1492 __ ldr(R1, Address(R6)); // Next class ID. | 1491 __ ldr(R1, Address(R6)); // Next class ID. |
1493 | 1492 |
1494 __ Bind(&test); | 1493 __ Bind(&test); |
1495 __ CompareImmediate(R1, Smi::RawValue(kIllegalCid), kNoPP); // Done? | 1494 __ CompareImmediate(R1, Smi::RawValue(kIllegalCid)); // Done? |
1496 __ b(&loop, NE); | 1495 __ b(&loop, NE); |
1497 | 1496 |
1498 __ Comment("IC miss"); | 1497 __ Comment("IC miss"); |
1499 // Compute address of arguments. | 1498 // Compute address of arguments. |
1500 // R7: argument_count - 1 (untagged). | 1499 // R7: argument_count - 1 (untagged). |
1501 // R7 <- SP + (R7 << 3) | 1500 // R7 <- SP + (R7 << 3) |
1502 __ add(R7, SP, Operand(R7, UXTX, 3)); // R7 is Untagged. | 1501 __ add(R7, SP, Operand(R7, UXTX, 3)); // R7 is Untagged. |
1503 // R7: address of receiver. | 1502 // R7: address of receiver. |
1504 // Create a stub frame as we are pushing some objects on the stack before | 1503 // Create a stub frame as we are pushing some objects on the stack before |
1505 // calling into the runtime. | 1504 // calling into the runtime. |
1506 __ EnterStubFrame(); | 1505 __ EnterStubFrame(); |
1507 // Preserve IC data object and arguments descriptor array and | 1506 // Preserve IC data object and arguments descriptor array and |
1508 // setup space on stack for result (target code object). | 1507 // setup space on stack for result (target code object). |
1509 __ Push(R4); // Preserve arguments descriptor array. | 1508 __ Push(R4); // Preserve arguments descriptor array. |
1510 __ Push(R5); // Preserve IC Data. | 1509 __ Push(R5); // Preserve IC Data. |
1511 // Setup space on stack for the result (target code object). | 1510 // Setup space on stack for the result (target code object). |
1512 __ PushObject(Object::null_object(), PP); | 1511 __ PushObject(Object::null_object()); |
1513 // Push call arguments. | 1512 // Push call arguments. |
1514 for (intptr_t i = 0; i < num_args; i++) { | 1513 for (intptr_t i = 0; i < num_args; i++) { |
1515 __ LoadFromOffset(TMP, R7, -i * kWordSize, kNoPP); | 1514 __ LoadFromOffset(TMP, R7, -i * kWordSize); |
1516 __ Push(TMP); | 1515 __ Push(TMP); |
1517 } | 1516 } |
1518 // Pass IC data object. | 1517 // Pass IC data object. |
1519 __ Push(R5); | 1518 __ Push(R5); |
1520 __ CallRuntime(handle_ic_miss, num_args + 1); | 1519 __ CallRuntime(handle_ic_miss, num_args + 1); |
1521 // Remove the call arguments pushed earlier, including the IC data object. | 1520 // Remove the call arguments pushed earlier, including the IC data object. |
1522 __ Drop(num_args + 1); | 1521 __ Drop(num_args + 1); |
1523 // Pop returned function object into R0. | 1522 // Pop returned function object into R0. |
1524 // Restore arguments descriptor array and IC data array. | 1523 // Restore arguments descriptor array and IC data array. |
1525 __ Pop(R0); // Pop returned function object into R0. | 1524 __ Pop(R0); // Pop returned function object into R0. |
1526 __ Pop(R5); // Restore IC Data. | 1525 __ Pop(R5); // Restore IC Data. |
1527 __ Pop(R4); // Restore arguments descriptor array. | 1526 __ Pop(R4); // Restore arguments descriptor array. |
1528 __ LeaveStubFrame(); | 1527 __ LeaveStubFrame(); |
1529 Label call_target_function; | 1528 Label call_target_function; |
1530 if (!FLAG_lazy_dispatchers) { | 1529 if (!FLAG_lazy_dispatchers) { |
1531 GenerateDispatcherCode(assembler, &call_target_function); | 1530 GenerateDispatcherCode(assembler, &call_target_function); |
1532 } else { | 1531 } else { |
1533 __ b(&call_target_function); | 1532 __ b(&call_target_function); |
1534 } | 1533 } |
1535 | 1534 |
1536 __ Bind(&found); | 1535 __ Bind(&found); |
1537 __ Comment("Update caller's counter"); | 1536 __ Comment("Update caller's counter"); |
1538 // R6: pointer to an IC data check group. | 1537 // R6: pointer to an IC data check group. |
1539 const intptr_t target_offset = ICData::TargetIndexFor(num_args) * kWordSize; | 1538 const intptr_t target_offset = ICData::TargetIndexFor(num_args) * kWordSize; |
1540 const intptr_t count_offset = ICData::CountIndexFor(num_args) * kWordSize; | 1539 const intptr_t count_offset = ICData::CountIndexFor(num_args) * kWordSize; |
1541 __ LoadFromOffset(R0, R6, target_offset, kNoPP); | 1540 __ LoadFromOffset(R0, R6, target_offset); |
1542 | 1541 |
1543 if (FLAG_optimization_counter_threshold >= 0) { | 1542 if (FLAG_optimization_counter_threshold >= 0) { |
1544 // Update counter. | 1543 // Update counter. |
1545 __ LoadFromOffset(R1, R6, count_offset, kNoPP); | 1544 __ LoadFromOffset(R1, R6, count_offset); |
1546 __ adds(R1, R1, Operand(Smi::RawValue(1))); | 1545 __ adds(R1, R1, Operand(Smi::RawValue(1))); |
1547 __ LoadImmediate(R2, Smi::RawValue(Smi::kMaxValue), kNoPP); | 1546 __ LoadImmediate(R2, Smi::RawValue(Smi::kMaxValue)); |
1548 __ csel(R1, R2, R1, VS); // Overflow. | 1547 __ csel(R1, R2, R1, VS); // Overflow. |
1549 __ StoreToOffset(R1, R6, count_offset, kNoPP); | 1548 __ StoreToOffset(R1, R6, count_offset); |
1550 } | 1549 } |
1551 | 1550 |
1552 __ Comment("Call target"); | 1551 __ Comment("Call target"); |
1553 __ Bind(&call_target_function); | 1552 __ Bind(&call_target_function); |
1554 // R0: target function. | 1553 // R0: target function. |
1555 __ LoadFieldFromOffset(R2, R0, Function::instructions_offset(), kNoPP); | 1554 __ LoadFieldFromOffset(R2, R0, Function::instructions_offset()); |
1556 __ AddImmediate( | 1555 __ AddImmediate( |
1557 R2, R2, Instructions::HeaderSize() - kHeapObjectTag, kNoPP); | 1556 R2, R2, Instructions::HeaderSize() - kHeapObjectTag); |
1558 if (range_collection_mode == kCollectRanges) { | 1557 if (range_collection_mode == kCollectRanges) { |
1559 __ ldr(R1, Address(SP, 0 * kWordSize)); | 1558 __ ldr(R1, Address(SP, 0 * kWordSize)); |
1560 if (num_args == 2) { | 1559 if (num_args == 2) { |
1561 __ ldr(R3, Address(SP, 1 * kWordSize)); | 1560 __ ldr(R3, Address(SP, 1 * kWordSize)); |
1562 } | 1561 } |
1563 __ EnterStubFrame(); | 1562 __ EnterStubFrame(); |
1564 __ Push(R5); | 1563 __ Push(R5); |
1565 if (num_args == 2) { | 1564 if (num_args == 2) { |
1566 __ Push(R3); | 1565 __ Push(R3); |
1567 } | 1566 } |
(...skipping 110 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1678 } | 1677 } |
1679 | 1678 |
1680 | 1679 |
1681 void StubCode::GenerateZeroArgsUnoptimizedStaticCallStub(Assembler* assembler) { | 1680 void StubCode::GenerateZeroArgsUnoptimizedStaticCallStub(Assembler* assembler) { |
1682 GenerateUsageCounterIncrement(assembler, R6); | 1681 GenerateUsageCounterIncrement(assembler, R6); |
1683 #if defined(DEBUG) | 1682 #if defined(DEBUG) |
1684 { Label ok; | 1683 { Label ok; |
1685 // Check that the IC data array has NumArgsTested() == 0. | 1684 // Check that the IC data array has NumArgsTested() == 0. |
1686 // 'NumArgsTested' is stored in the least significant bits of 'state_bits'. | 1685 // 'NumArgsTested' is stored in the least significant bits of 'state_bits'. |
1687 __ LoadFromOffset(R6, R5, ICData::state_bits_offset() - kHeapObjectTag, | 1686 __ LoadFromOffset(R6, R5, ICData::state_bits_offset() - kHeapObjectTag, |
1688 kNoPP, kUnsignedWord); | 1687 kUnsignedWord); |
1689 ASSERT(ICData::NumArgsTestedShift() == 0); // No shift needed. | 1688 ASSERT(ICData::NumArgsTestedShift() == 0); // No shift needed. |
1690 __ andi(R6, R6, Immediate(ICData::NumArgsTestedMask())); | 1689 __ andi(R6, R6, Immediate(ICData::NumArgsTestedMask())); |
1691 __ CompareImmediate(R6, 0, kNoPP); | 1690 __ CompareImmediate(R6, 0); |
1692 __ b(&ok, EQ); | 1691 __ b(&ok, EQ); |
1693 __ Stop("Incorrect IC data for unoptimized static call"); | 1692 __ Stop("Incorrect IC data for unoptimized static call"); |
1694 __ Bind(&ok); | 1693 __ Bind(&ok); |
1695 } | 1694 } |
1696 #endif // DEBUG | 1695 #endif // DEBUG |
1697 | 1696 |
1698 // Check single stepping. | 1697 // Check single stepping. |
1699 Label stepping, done_stepping; | 1698 Label stepping, done_stepping; |
1700 if (FLAG_support_debugger) { | 1699 if (FLAG_support_debugger) { |
1701 __ LoadIsolate(R6); | 1700 __ LoadIsolate(R6); |
1702 __ LoadFromOffset( | 1701 __ LoadFromOffset( |
1703 R6, R6, Isolate::single_step_offset(), kNoPP, kUnsignedByte); | 1702 R6, R6, Isolate::single_step_offset(), kUnsignedByte); |
1704 __ CompareImmediate(R6, 0, kNoPP); | 1703 __ CompareImmediate(R6, 0); |
1705 __ b(&stepping, NE); | 1704 __ b(&stepping, NE); |
1706 __ Bind(&done_stepping); | 1705 __ Bind(&done_stepping); |
1707 } | 1706 } |
1708 | 1707 |
1709 // R5: IC data object (preserved). | 1708 // R5: IC data object (preserved). |
1710 __ LoadFieldFromOffset(R6, R5, ICData::ic_data_offset(), kNoPP); | 1709 __ LoadFieldFromOffset(R6, R5, ICData::ic_data_offset()); |
1711 // R6: ic_data_array with entries: target functions and count. | 1710 // R6: ic_data_array with entries: target functions and count. |
1712 __ AddImmediate(R6, R6, Array::data_offset() - kHeapObjectTag, kNoPP); | 1711 __ AddImmediate(R6, R6, Array::data_offset() - kHeapObjectTag); |
1713 // R6: points directly to the first ic data array element. | 1712 // R6: points directly to the first ic data array element. |
1714 const intptr_t target_offset = ICData::TargetIndexFor(0) * kWordSize; | 1713 const intptr_t target_offset = ICData::TargetIndexFor(0) * kWordSize; |
1715 const intptr_t count_offset = ICData::CountIndexFor(0) * kWordSize; | 1714 const intptr_t count_offset = ICData::CountIndexFor(0) * kWordSize; |
1716 | 1715 |
1717 if (FLAG_optimization_counter_threshold >= 0) { | 1716 if (FLAG_optimization_counter_threshold >= 0) { |
1718 // Increment count for this call. | 1717 // Increment count for this call. |
1719 __ LoadFromOffset(R1, R6, count_offset, kNoPP); | 1718 __ LoadFromOffset(R1, R6, count_offset); |
1720 __ adds(R1, R1, Operand(Smi::RawValue(1))); | 1719 __ adds(R1, R1, Operand(Smi::RawValue(1))); |
1721 __ LoadImmediate(R2, Smi::RawValue(Smi::kMaxValue), kNoPP); | 1720 __ LoadImmediate(R2, Smi::RawValue(Smi::kMaxValue)); |
1722 __ csel(R1, R2, R1, VS); // Overflow. | 1721 __ csel(R1, R2, R1, VS); // Overflow. |
1723 __ StoreToOffset(R1, R6, count_offset, kNoPP); | 1722 __ StoreToOffset(R1, R6, count_offset); |
1724 } | 1723 } |
1725 | 1724 |
1726 // Load arguments descriptor into R4. | 1725 // Load arguments descriptor into R4. |
1727 __ LoadFieldFromOffset(R4, R5, ICData::arguments_descriptor_offset(), kNoPP); | 1726 __ LoadFieldFromOffset(R4, R5, ICData::arguments_descriptor_offset()); |
1728 | 1727 |
1729 // Get function and call it, if possible. | 1728 // Get function and call it, if possible. |
1730 __ LoadFromOffset(R0, R6, target_offset, kNoPP); | 1729 __ LoadFromOffset(R0, R6, target_offset); |
1731 __ LoadFieldFromOffset(R2, R0, Function::instructions_offset(), kNoPP); | 1730 __ LoadFieldFromOffset(R2, R0, Function::instructions_offset()); |
1732 | 1731 |
1733 // R0: function. | 1732 // R0: function. |
1734 // R2: target instructons. | 1733 // R2: target instructons. |
1735 __ AddImmediate( | 1734 __ AddImmediate( |
1736 R2, R2, Instructions::HeaderSize() - kHeapObjectTag, kNoPP); | 1735 R2, R2, Instructions::HeaderSize() - kHeapObjectTag); |
1737 __ br(R2); | 1736 __ br(R2); |
1738 | 1737 |
1739 if (FLAG_support_debugger) { | 1738 if (FLAG_support_debugger) { |
1740 __ Bind(&stepping); | 1739 __ Bind(&stepping); |
1741 __ EnterStubFrame(); | 1740 __ EnterStubFrame(); |
1742 __ Push(R5); // Preserve IC data. | 1741 __ Push(R5); // Preserve IC data. |
1743 __ CallRuntime(kSingleStepHandlerRuntimeEntry, 0); | 1742 __ CallRuntime(kSingleStepHandlerRuntimeEntry, 0); |
1744 __ Pop(R5); | 1743 __ Pop(R5); |
1745 __ LeaveStubFrame(); | 1744 __ LeaveStubFrame(); |
1746 __ b(&done_stepping); | 1745 __ b(&done_stepping); |
(...skipping 26 matching lines...) Expand all Loading... |
1773 __ EnterStubFrame(); | 1772 __ EnterStubFrame(); |
1774 __ Push(R5); // Save IC Data. | 1773 __ Push(R5); // Save IC Data. |
1775 __ Push(R4); // Save arg. desc. | 1774 __ Push(R4); // Save arg. desc. |
1776 __ Push(R0); // Pass function. | 1775 __ Push(R0); // Pass function. |
1777 __ CallRuntime(kCompileFunctionRuntimeEntry, 1); | 1776 __ CallRuntime(kCompileFunctionRuntimeEntry, 1); |
1778 __ Pop(R0); // Restore argument. | 1777 __ Pop(R0); // Restore argument. |
1779 __ Pop(R4); // Restore arg desc. | 1778 __ Pop(R4); // Restore arg desc. |
1780 __ Pop(R5); // Restore IC Data. | 1779 __ Pop(R5); // Restore IC Data. |
1781 __ LeaveStubFrame(); | 1780 __ LeaveStubFrame(); |
1782 | 1781 |
1783 __ LoadFieldFromOffset(R2, R0, Function::instructions_offset(), kNoPP); | 1782 __ LoadFieldFromOffset(R2, R0, Function::instructions_offset()); |
1784 __ AddImmediate( | 1783 __ AddImmediate( |
1785 R2, R2, Instructions::HeaderSize() - kHeapObjectTag, kNoPP); | 1784 R2, R2, Instructions::HeaderSize() - kHeapObjectTag); |
1786 __ br(R2); | 1785 __ br(R2); |
1787 } | 1786 } |
1788 | 1787 |
1789 | 1788 |
1790 // R5: Contains an ICData. | 1789 // R5: Contains an ICData. |
1791 void StubCode::GenerateICCallBreakpointStub(Assembler* assembler) { | 1790 void StubCode::GenerateICCallBreakpointStub(Assembler* assembler) { |
1792 __ EnterStubFrame(); | 1791 __ EnterStubFrame(); |
1793 __ Push(R5); | 1792 __ Push(R5); |
1794 __ PushObject(Object::null_object(), PP); // Space for result. | 1793 __ PushObject(Object::null_object()); // Space for result. |
1795 __ CallRuntime(kBreakpointRuntimeHandlerRuntimeEntry, 0); | 1794 __ CallRuntime(kBreakpointRuntimeHandlerRuntimeEntry, 0); |
1796 __ Pop(R0); | 1795 __ Pop(R0); |
1797 __ Pop(R5); | 1796 __ Pop(R5); |
1798 __ LeaveStubFrame(); | 1797 __ LeaveStubFrame(); |
1799 __ br(R0); | 1798 __ br(R0); |
1800 } | 1799 } |
1801 | 1800 |
1802 | 1801 |
1803 void StubCode::GenerateRuntimeCallBreakpointStub(Assembler* assembler) { | 1802 void StubCode::GenerateRuntimeCallBreakpointStub(Assembler* assembler) { |
1804 __ EnterStubFrame(); | 1803 __ EnterStubFrame(); |
1805 __ PushObject(Object::null_object(), PP); // Space for result. | 1804 __ PushObject(Object::null_object()); // Space for result. |
1806 __ CallRuntime(kBreakpointRuntimeHandlerRuntimeEntry, 0); | 1805 __ CallRuntime(kBreakpointRuntimeHandlerRuntimeEntry, 0); |
1807 __ Pop(R0); | 1806 __ Pop(R0); |
1808 __ LeaveStubFrame(); | 1807 __ LeaveStubFrame(); |
1809 __ br(R0); | 1808 __ br(R0); |
1810 } | 1809 } |
1811 | 1810 |
1812 // Called only from unoptimized code. All relevant registers have been saved. | 1811 // Called only from unoptimized code. All relevant registers have been saved. |
1813 void StubCode::GenerateDebugStepCheckStub( | 1812 void StubCode::GenerateDebugStepCheckStub( |
1814 Assembler* assembler) { | 1813 Assembler* assembler) { |
1815 // Check single stepping. | 1814 // Check single stepping. |
1816 Label stepping, done_stepping; | 1815 Label stepping, done_stepping; |
1817 __ LoadIsolate(R1); | 1816 __ LoadIsolate(R1); |
1818 __ LoadFromOffset( | 1817 __ LoadFromOffset( |
1819 R1, R1, Isolate::single_step_offset(), kNoPP, kUnsignedByte); | 1818 R1, R1, Isolate::single_step_offset(), kUnsignedByte); |
1820 __ CompareImmediate(R1, 0, kNoPP); | 1819 __ CompareImmediate(R1, 0); |
1821 __ b(&stepping, NE); | 1820 __ b(&stepping, NE); |
1822 __ Bind(&done_stepping); | 1821 __ Bind(&done_stepping); |
1823 | 1822 |
1824 __ ret(); | 1823 __ ret(); |
1825 | 1824 |
1826 __ Bind(&stepping); | 1825 __ Bind(&stepping); |
1827 __ EnterStubFrame(); | 1826 __ EnterStubFrame(); |
1828 __ CallRuntime(kSingleStepHandlerRuntimeEntry, 0); | 1827 __ CallRuntime(kSingleStepHandlerRuntimeEntry, 0); |
1829 __ LeaveStubFrame(); | 1828 __ LeaveStubFrame(); |
1830 __ b(&done_stepping); | 1829 __ b(&done_stepping); |
1831 } | 1830 } |
1832 | 1831 |
1833 | 1832 |
1834 // Used to check class and type arguments. Arguments passed in registers: | 1833 // Used to check class and type arguments. Arguments passed in registers: |
1835 // LR: return address. | 1834 // LR: return address. |
1836 // R0: instance (must be preserved). | 1835 // R0: instance (must be preserved). |
1837 // R1: instantiator type arguments or NULL. | 1836 // R1: instantiator type arguments or NULL. |
1838 // R2: cache array. | 1837 // R2: cache array. |
1839 // Result in R1: null -> not found, otherwise result (true or false). | 1838 // Result in R1: null -> not found, otherwise result (true or false). |
1840 static void GenerateSubtypeNTestCacheStub(Assembler* assembler, int n) { | 1839 static void GenerateSubtypeNTestCacheStub(Assembler* assembler, int n) { |
1841 ASSERT((1 <= n) && (n <= 3)); | 1840 ASSERT((1 <= n) && (n <= 3)); |
1842 if (n > 1) { | 1841 if (n > 1) { |
1843 // Get instance type arguments. | 1842 // Get instance type arguments. |
1844 __ LoadClass(R3, R0, kNoPP); | 1843 __ LoadClass(R3, R0); |
1845 // Compute instance type arguments into R4. | 1844 // Compute instance type arguments into R4. |
1846 Label has_no_type_arguments; | 1845 Label has_no_type_arguments; |
1847 __ LoadObject(R4, Object::null_object(), PP); | 1846 __ LoadObject(R4, Object::null_object()); |
1848 __ LoadFieldFromOffset(R5, R3, | 1847 __ LoadFieldFromOffset(R5, R3, |
1849 Class::type_arguments_field_offset_in_words_offset(), kNoPP, kWord); | 1848 Class::type_arguments_field_offset_in_words_offset(), kWord); |
1850 __ CompareImmediate(R5, Class::kNoTypeArguments, kNoPP); | 1849 __ CompareImmediate(R5, Class::kNoTypeArguments); |
1851 __ b(&has_no_type_arguments, EQ); | 1850 __ b(&has_no_type_arguments, EQ); |
1852 __ add(R5, R0, Operand(R5, LSL, 3)); | 1851 __ add(R5, R0, Operand(R5, LSL, 3)); |
1853 __ LoadFieldFromOffset(R4, R5, 0, kNoPP); | 1852 __ LoadFieldFromOffset(R4, R5, 0); |
1854 __ Bind(&has_no_type_arguments); | 1853 __ Bind(&has_no_type_arguments); |
1855 } | 1854 } |
1856 __ LoadClassId(R3, R0, kNoPP); | 1855 __ LoadClassId(R3, R0); |
1857 // R0: instance. | 1856 // R0: instance. |
1858 // R1: instantiator type arguments or NULL. | 1857 // R1: instantiator type arguments or NULL. |
1859 // R2: SubtypeTestCache. | 1858 // R2: SubtypeTestCache. |
1860 // R3: instance class id. | 1859 // R3: instance class id. |
1861 // R4: instance type arguments (null if none), used only if n > 1. | 1860 // R4: instance type arguments (null if none), used only if n > 1. |
1862 __ LoadFieldFromOffset(R2, R2, SubtypeTestCache::cache_offset(), kNoPP); | 1861 __ LoadFieldFromOffset(R2, R2, SubtypeTestCache::cache_offset()); |
1863 __ AddImmediate(R2, R2, Array::data_offset() - kHeapObjectTag, kNoPP); | 1862 __ AddImmediate(R2, R2, Array::data_offset() - kHeapObjectTag); |
1864 | 1863 |
1865 Label loop, found, not_found, next_iteration; | 1864 Label loop, found, not_found, next_iteration; |
1866 // R2: entry start. | 1865 // R2: entry start. |
1867 // R3: instance class id. | 1866 // R3: instance class id. |
1868 // R4: instance type arguments. | 1867 // R4: instance type arguments. |
1869 __ SmiTag(R3); | 1868 __ SmiTag(R3); |
1870 __ Bind(&loop); | 1869 __ Bind(&loop); |
1871 __ LoadFromOffset( | 1870 __ LoadFromOffset( |
1872 R5, R2, kWordSize * SubtypeTestCache::kInstanceClassId, kNoPP); | 1871 R5, R2, kWordSize * SubtypeTestCache::kInstanceClassId); |
1873 __ CompareObject(R5, Object::null_object(), PP); | 1872 __ CompareObject(R5, Object::null_object()); |
1874 __ b(¬_found, EQ); | 1873 __ b(¬_found, EQ); |
1875 __ CompareRegisters(R5, R3); | 1874 __ CompareRegisters(R5, R3); |
1876 if (n == 1) { | 1875 if (n == 1) { |
1877 __ b(&found, EQ); | 1876 __ b(&found, EQ); |
1878 } else { | 1877 } else { |
1879 __ b(&next_iteration, NE); | 1878 __ b(&next_iteration, NE); |
1880 __ LoadFromOffset( | 1879 __ LoadFromOffset( |
1881 R5, R2, kWordSize * SubtypeTestCache::kInstanceTypeArguments, kNoPP); | 1880 R5, R2, kWordSize * SubtypeTestCache::kInstanceTypeArguments); |
1882 __ CompareRegisters(R5, R4); | 1881 __ CompareRegisters(R5, R4); |
1883 if (n == 2) { | 1882 if (n == 2) { |
1884 __ b(&found, EQ); | 1883 __ b(&found, EQ); |
1885 } else { | 1884 } else { |
1886 __ b(&next_iteration, NE); | 1885 __ b(&next_iteration, NE); |
1887 __ LoadFromOffset(R5, R2, | 1886 __ LoadFromOffset(R5, R2, |
1888 kWordSize * SubtypeTestCache::kInstantiatorTypeArguments, kNoPP); | 1887 kWordSize * SubtypeTestCache::kInstantiatorTypeArguments); |
1889 __ CompareRegisters(R5, R1); | 1888 __ CompareRegisters(R5, R1); |
1890 __ b(&found, EQ); | 1889 __ b(&found, EQ); |
1891 } | 1890 } |
1892 } | 1891 } |
1893 __ Bind(&next_iteration); | 1892 __ Bind(&next_iteration); |
1894 __ AddImmediate( | 1893 __ AddImmediate( |
1895 R2, R2, kWordSize * SubtypeTestCache::kTestEntryLength, kNoPP); | 1894 R2, R2, kWordSize * SubtypeTestCache::kTestEntryLength); |
1896 __ b(&loop); | 1895 __ b(&loop); |
1897 // Fall through to not found. | 1896 // Fall through to not found. |
1898 __ Bind(¬_found); | 1897 __ Bind(¬_found); |
1899 __ LoadObject(R1, Object::null_object(), PP); | 1898 __ LoadObject(R1, Object::null_object()); |
1900 __ ret(); | 1899 __ ret(); |
1901 | 1900 |
1902 __ Bind(&found); | 1901 __ Bind(&found); |
1903 __ LoadFromOffset(R1, R2, kWordSize * SubtypeTestCache::kTestResult, kNoPP); | 1902 __ LoadFromOffset(R1, R2, kWordSize * SubtypeTestCache::kTestResult); |
1904 __ ret(); | 1903 __ ret(); |
1905 } | 1904 } |
1906 | 1905 |
1907 | 1906 |
1908 // Used to check class and type arguments. Arguments passed on stack: | 1907 // Used to check class and type arguments. Arguments passed on stack: |
1909 // TOS + 0: return address. | 1908 // TOS + 0: return address. |
1910 // TOS + 1: instantiator type arguments or NULL. | 1909 // TOS + 1: instantiator type arguments or NULL. |
1911 // TOS + 2: instance. | 1910 // TOS + 2: instance. |
1912 // TOS + 3: cache array. | 1911 // TOS + 3: cache array. |
1913 // Result in RCX: null -> not found, otherwise result (true or false). | 1912 // Result in RCX: null -> not found, otherwise result (true or false). |
(...skipping 43 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1957 ASSERT(kExceptionObjectReg == R0); | 1956 ASSERT(kExceptionObjectReg == R0); |
1958 ASSERT(kStackTraceObjectReg == R1); | 1957 ASSERT(kStackTraceObjectReg == R1); |
1959 __ mov(LR, R0); // Program counter. | 1958 __ mov(LR, R0); // Program counter. |
1960 __ mov(SP, R1); // Stack pointer. | 1959 __ mov(SP, R1); // Stack pointer. |
1961 __ mov(FP, R2); // Frame_pointer. | 1960 __ mov(FP, R2); // Frame_pointer. |
1962 __ mov(R0, R3); // Exception object. | 1961 __ mov(R0, R3); // Exception object. |
1963 __ mov(R1, R4); // StackTrace object. | 1962 __ mov(R1, R4); // StackTrace object. |
1964 __ mov(THR, R5); | 1963 __ mov(THR, R5); |
1965 __ LoadIsolate(R5); | 1964 __ LoadIsolate(R5); |
1966 // Set the tag. | 1965 // Set the tag. |
1967 __ LoadImmediate(R2, VMTag::kDartTagId, kNoPP); | 1966 __ LoadImmediate(R2, VMTag::kDartTagId); |
1968 __ StoreToOffset(R2, R5, Isolate::vm_tag_offset(), kNoPP); | 1967 __ StoreToOffset(R2, R5, Isolate::vm_tag_offset()); |
1969 // Clear top exit frame. | 1968 // Clear top exit frame. |
1970 __ StoreToOffset(ZR, THR, Thread::top_exit_frame_info_offset(), kNoPP); | 1969 __ StoreToOffset(ZR, THR, Thread::top_exit_frame_info_offset()); |
1971 __ ret(); // Jump to the exception handler code. | 1970 __ ret(); // Jump to the exception handler code. |
1972 } | 1971 } |
1973 | 1972 |
1974 | 1973 |
1975 // Calls to the runtime to optimize the given function. | 1974 // Calls to the runtime to optimize the given function. |
1976 // R6: function to be re-optimized. | 1975 // R6: function to be re-optimized. |
1977 // R4: argument descriptor (preserved). | 1976 // R4: argument descriptor (preserved). |
1978 void StubCode::GenerateOptimizeFunctionStub(Assembler* assembler) { | 1977 void StubCode::GenerateOptimizeFunctionStub(Assembler* assembler) { |
1979 __ EnterStubFrame(); | 1978 __ EnterStubFrame(); |
1980 __ Push(R4); | 1979 __ Push(R4); |
1981 // Setup space on stack for the return value. | 1980 // Setup space on stack for the return value. |
1982 __ PushObject(Object::null_object(), PP); | 1981 __ PushObject(Object::null_object()); |
1983 __ Push(R6); | 1982 __ Push(R6); |
1984 __ CallRuntime(kOptimizeInvokedFunctionRuntimeEntry, 1); | 1983 __ CallRuntime(kOptimizeInvokedFunctionRuntimeEntry, 1); |
1985 __ Pop(R0); // Discard argument. | 1984 __ Pop(R0); // Discard argument. |
1986 __ Pop(R0); // Get Code object | 1985 __ Pop(R0); // Get Code object |
1987 __ Pop(R4); // Restore argument descriptor. | 1986 __ Pop(R4); // Restore argument descriptor. |
1988 __ LoadFieldFromOffset(R0, R0, Code::instructions_offset(), kNoPP); | 1987 __ LoadFieldFromOffset(R0, R0, Code::instructions_offset()); |
1989 __ AddImmediate(R0, R0, Instructions::HeaderSize() - kHeapObjectTag, PP); | 1988 __ AddImmediate(R0, R0, Instructions::HeaderSize() - kHeapObjectTag); |
1990 __ LeaveStubFrame(); | 1989 __ LeaveStubFrame(); |
1991 __ br(R0); | 1990 __ br(R0); |
1992 __ brk(0); | 1991 __ brk(0); |
1993 } | 1992 } |
1994 | 1993 |
1995 | 1994 |
1996 DECLARE_LEAF_RUNTIME_ENTRY(intptr_t, | 1995 DECLARE_LEAF_RUNTIME_ENTRY(intptr_t, |
1997 BigintCompare, | 1996 BigintCompare, |
1998 RawBigint* left, | 1997 RawBigint* left, |
1999 RawBigint* right); | 1998 RawBigint* right); |
2000 | 1999 |
2001 | 2000 |
2002 // Does identical check (object references are equal or not equal) with special | 2001 // Does identical check (object references are equal or not equal) with special |
2003 // checks for boxed numbers. | 2002 // checks for boxed numbers. |
2004 // Left and right are pushed on stack. | 2003 // Left and right are pushed on stack. |
2005 // Return Zero condition flag set if equal. | 2004 // Return Zero condition flag set if equal. |
2006 // Note: A Mint cannot contain a value that would fit in Smi, a Bigint | 2005 // Note: A Mint cannot contain a value that would fit in Smi, a Bigint |
2007 // cannot contain a value that fits in Mint or Smi. | 2006 // cannot contain a value that fits in Mint or Smi. |
2008 static void GenerateIdenticalWithNumberCheckStub(Assembler* assembler, | 2007 static void GenerateIdenticalWithNumberCheckStub(Assembler* assembler, |
2009 const Register left, | 2008 const Register left, |
2010 const Register right) { | 2009 const Register right) { |
2011 Label reference_compare, done, check_mint, check_bigint; | 2010 Label reference_compare, done, check_mint, check_bigint; |
2012 // If any of the arguments is Smi do reference compare. | 2011 // If any of the arguments is Smi do reference compare. |
2013 __ tsti(left, Immediate(kSmiTagMask)); | 2012 __ tsti(left, Immediate(kSmiTagMask)); |
2014 __ b(&reference_compare, EQ); | 2013 __ b(&reference_compare, EQ); |
2015 __ tsti(right, Immediate(kSmiTagMask)); | 2014 __ tsti(right, Immediate(kSmiTagMask)); |
2016 __ b(&reference_compare, EQ); | 2015 __ b(&reference_compare, EQ); |
2017 | 2016 |
2018 // Value compare for two doubles. | 2017 // Value compare for two doubles. |
2019 __ CompareClassId(left, kDoubleCid, kNoPP); | 2018 __ CompareClassId(left, kDoubleCid); |
2020 __ b(&check_mint, NE); | 2019 __ b(&check_mint, NE); |
2021 __ CompareClassId(right, kDoubleCid, kNoPP); | 2020 __ CompareClassId(right, kDoubleCid); |
2022 __ b(&done, NE); | 2021 __ b(&done, NE); |
2023 | 2022 |
2024 // Double values bitwise compare. | 2023 // Double values bitwise compare. |
2025 __ LoadFieldFromOffset(left, left, Double::value_offset(), kNoPP); | 2024 __ LoadFieldFromOffset(left, left, Double::value_offset()); |
2026 __ LoadFieldFromOffset(right, right, Double::value_offset(), kNoPP); | 2025 __ LoadFieldFromOffset(right, right, Double::value_offset()); |
2027 __ CompareRegisters(left, right); | 2026 __ CompareRegisters(left, right); |
2028 __ b(&done); | 2027 __ b(&done); |
2029 | 2028 |
2030 __ Bind(&check_mint); | 2029 __ Bind(&check_mint); |
2031 __ CompareClassId(left, kMintCid, kNoPP); | 2030 __ CompareClassId(left, kMintCid); |
2032 __ b(&check_bigint, NE); | 2031 __ b(&check_bigint, NE); |
2033 __ CompareClassId(right, kMintCid, kNoPP); | 2032 __ CompareClassId(right, kMintCid); |
2034 __ b(&done, NE); | 2033 __ b(&done, NE); |
2035 __ LoadFieldFromOffset(left, left, Mint::value_offset(), kNoPP); | 2034 __ LoadFieldFromOffset(left, left, Mint::value_offset()); |
2036 __ LoadFieldFromOffset(right, right, Mint::value_offset(), kNoPP); | 2035 __ LoadFieldFromOffset(right, right, Mint::value_offset()); |
2037 __ b(&done); | 2036 __ b(&done); |
2038 | 2037 |
2039 __ Bind(&check_bigint); | 2038 __ Bind(&check_bigint); |
2040 __ CompareClassId(left, kBigintCid, kNoPP); | 2039 __ CompareClassId(left, kBigintCid); |
2041 __ b(&reference_compare, NE); | 2040 __ b(&reference_compare, NE); |
2042 __ CompareClassId(right, kBigintCid, kNoPP); | 2041 __ CompareClassId(right, kBigintCid); |
2043 __ b(&done, NE); | 2042 __ b(&done, NE); |
2044 __ EnterStubFrame(); | 2043 __ EnterStubFrame(); |
2045 __ ReserveAlignedFrameSpace(2 * kWordSize); | 2044 __ ReserveAlignedFrameSpace(2 * kWordSize); |
2046 __ StoreToOffset(left, SP, 0 * kWordSize, kNoPP); | 2045 __ StoreToOffset(left, SP, 0 * kWordSize); |
2047 __ StoreToOffset(right, SP, 1 * kWordSize, kNoPP); | 2046 __ StoreToOffset(right, SP, 1 * kWordSize); |
2048 __ CallRuntime(kBigintCompareRuntimeEntry, 2); | 2047 __ CallRuntime(kBigintCompareRuntimeEntry, 2); |
2049 // Result in R0, 0 means equal. | 2048 // Result in R0, 0 means equal. |
2050 __ LeaveStubFrame(); | 2049 __ LeaveStubFrame(); |
2051 __ cmp(R0, Operand(0)); | 2050 __ cmp(R0, Operand(0)); |
2052 __ b(&done); | 2051 __ b(&done); |
2053 | 2052 |
2054 __ Bind(&reference_compare); | 2053 __ Bind(&reference_compare); |
2055 __ CompareRegisters(left, right); | 2054 __ CompareRegisters(left, right); |
2056 __ Bind(&done); | 2055 __ Bind(&done); |
2057 } | 2056 } |
2058 | 2057 |
2059 | 2058 |
2060 // Called only from unoptimized code. All relevant registers have been saved. | 2059 // Called only from unoptimized code. All relevant registers have been saved. |
2061 // LR: return address. | 2060 // LR: return address. |
2062 // SP + 4: left operand. | 2061 // SP + 4: left operand. |
2063 // SP + 0: right operand. | 2062 // SP + 0: right operand. |
2064 // Return Zero condition flag set if equal. | 2063 // Return Zero condition flag set if equal. |
2065 void StubCode::GenerateUnoptimizedIdenticalWithNumberCheckStub( | 2064 void StubCode::GenerateUnoptimizedIdenticalWithNumberCheckStub( |
2066 Assembler* assembler) { | 2065 Assembler* assembler) { |
2067 // Check single stepping. | 2066 // Check single stepping. |
2068 Label stepping, done_stepping; | 2067 Label stepping, done_stepping; |
2069 if (FLAG_support_debugger) { | 2068 if (FLAG_support_debugger) { |
2070 __ LoadIsolate(R1); | 2069 __ LoadIsolate(R1); |
2071 __ LoadFromOffset( | 2070 __ LoadFromOffset(R1, R1, Isolate::single_step_offset(), kUnsignedByte); |
2072 R1, R1, Isolate::single_step_offset(), kNoPP, kUnsignedByte); | 2071 __ CompareImmediate(R1, 0); |
2073 __ CompareImmediate(R1, 0, kNoPP); | |
2074 __ b(&stepping, NE); | 2072 __ b(&stepping, NE); |
2075 __ Bind(&done_stepping); | 2073 __ Bind(&done_stepping); |
2076 } | 2074 } |
2077 | 2075 |
2078 const Register left = R1; | 2076 const Register left = R1; |
2079 const Register right = R0; | 2077 const Register right = R0; |
2080 __ LoadFromOffset(left, SP, 1 * kWordSize, kNoPP); | 2078 __ LoadFromOffset(left, SP, 1 * kWordSize); |
2081 __ LoadFromOffset(right, SP, 0 * kWordSize, kNoPP); | 2079 __ LoadFromOffset(right, SP, 0 * kWordSize); |
2082 GenerateIdenticalWithNumberCheckStub(assembler, left, right); | 2080 GenerateIdenticalWithNumberCheckStub(assembler, left, right); |
2083 __ ret(); | 2081 __ ret(); |
2084 | 2082 |
2085 if (FLAG_support_debugger) { | 2083 if (FLAG_support_debugger) { |
2086 __ Bind(&stepping); | 2084 __ Bind(&stepping); |
2087 __ EnterStubFrame(); | 2085 __ EnterStubFrame(); |
2088 __ CallRuntime(kSingleStepHandlerRuntimeEntry, 0); | 2086 __ CallRuntime(kSingleStepHandlerRuntimeEntry, 0); |
2089 __ LeaveStubFrame(); | 2087 __ LeaveStubFrame(); |
2090 __ b(&done_stepping); | 2088 __ b(&done_stepping); |
2091 } | 2089 } |
2092 } | 2090 } |
2093 | 2091 |
2094 | 2092 |
2095 // Called from optimized code only. | 2093 // Called from optimized code only. |
2096 // LR: return address. | 2094 // LR: return address. |
2097 // SP + 4: left operand. | 2095 // SP + 4: left operand. |
2098 // SP + 0: right operand. | 2096 // SP + 0: right operand. |
2099 // Return Zero condition flag set if equal. | 2097 // Return Zero condition flag set if equal. |
2100 void StubCode::GenerateOptimizedIdenticalWithNumberCheckStub( | 2098 void StubCode::GenerateOptimizedIdenticalWithNumberCheckStub( |
2101 Assembler* assembler) { | 2099 Assembler* assembler) { |
2102 const Register left = R1; | 2100 const Register left = R1; |
2103 const Register right = R0; | 2101 const Register right = R0; |
2104 __ LoadFromOffset(left, SP, 1 * kWordSize, kNoPP); | 2102 __ LoadFromOffset(left, SP, 1 * kWordSize); |
2105 __ LoadFromOffset(right, SP, 0 * kWordSize, kNoPP); | 2103 __ LoadFromOffset(right, SP, 0 * kWordSize); |
2106 GenerateIdenticalWithNumberCheckStub(assembler, left, right); | 2104 GenerateIdenticalWithNumberCheckStub(assembler, left, right); |
2107 __ ret(); | 2105 __ ret(); |
2108 } | 2106 } |
2109 | 2107 |
2110 | 2108 |
2111 void StubCode::EmitMegamorphicLookup( | 2109 void StubCode::EmitMegamorphicLookup( |
2112 Assembler* assembler, Register receiver, Register cache, Register target) { | 2110 Assembler* assembler, Register receiver, Register cache, Register target) { |
2113 ASSERT((cache != R0) && (cache != R2)); | 2111 ASSERT((cache != R0) && (cache != R2)); |
2114 __ LoadTaggedClassIdMayBeSmi(R0, receiver); | 2112 __ LoadTaggedClassIdMayBeSmi(R0, receiver); |
2115 // R0: class ID of the receiver (smi). | 2113 // R0: class ID of the receiver (smi). |
2116 __ LoadFieldFromOffset(R2, cache, MegamorphicCache::buckets_offset(), PP); | 2114 __ LoadFieldFromOffset(R2, cache, MegamorphicCache::buckets_offset()); |
2117 __ LoadFieldFromOffset(R1, cache, MegamorphicCache::mask_offset(), PP); | 2115 __ LoadFieldFromOffset(R1, cache, MegamorphicCache::mask_offset()); |
2118 // R2: cache buckets array. | 2116 // R2: cache buckets array. |
2119 // R1: mask. | 2117 // R1: mask. |
2120 __ mov(R3, R0); | 2118 __ mov(R3, R0); |
2121 | 2119 |
2122 Label loop, update, call_target_function; | 2120 Label loop, update, call_target_function; |
2123 __ b(&loop); | 2121 __ b(&loop); |
2124 | 2122 |
2125 __ Bind(&update); | 2123 __ Bind(&update); |
2126 __ add(R3, R3, Operand(Smi::RawValue(1))); | 2124 __ add(R3, R3, Operand(Smi::RawValue(1))); |
2127 __ Bind(&loop); | 2125 __ Bind(&loop); |
2128 __ and_(R3, R3, Operand(R1)); | 2126 __ and_(R3, R3, Operand(R1)); |
2129 const intptr_t base = Array::data_offset(); | 2127 const intptr_t base = Array::data_offset(); |
2130 // R3 is smi tagged, but table entries are 16 bytes, so LSL 3. | 2128 // R3 is smi tagged, but table entries are 16 bytes, so LSL 3. |
2131 __ add(TMP, R2, Operand(R3, LSL, 3)); | 2129 __ add(TMP, R2, Operand(R3, LSL, 3)); |
2132 __ LoadFieldFromOffset(R4, TMP, base, PP); | 2130 __ LoadFieldFromOffset(R4, TMP, base); |
2133 | 2131 |
2134 ASSERT(kIllegalCid == 0); | 2132 ASSERT(kIllegalCid == 0); |
2135 __ tst(R4, Operand(R4)); | 2133 __ tst(R4, Operand(R4)); |
2136 __ b(&call_target_function, EQ); | 2134 __ b(&call_target_function, EQ); |
2137 __ CompareRegisters(R4, R0); | 2135 __ CompareRegisters(R4, R0); |
2138 __ b(&update, NE); | 2136 __ b(&update, NE); |
2139 | 2137 |
2140 __ Bind(&call_target_function); | 2138 __ Bind(&call_target_function); |
2141 // Call the target found in the cache. For a class id match, this is a | 2139 // Call the target found in the cache. For a class id match, this is a |
2142 // proper target for the given name and arguments descriptor. If the | 2140 // proper target for the given name and arguments descriptor. If the |
2143 // illegal class id was found, the target is a cache miss handler that can | 2141 // illegal class id was found, the target is a cache miss handler that can |
2144 // be invoked as a normal Dart function. | 2142 // be invoked as a normal Dart function. |
2145 __ add(TMP, R2, Operand(R3, LSL, 3)); | 2143 __ add(TMP, R2, Operand(R3, LSL, 3)); |
2146 __ LoadFieldFromOffset(R0, TMP, base + kWordSize, PP); | 2144 __ LoadFieldFromOffset(R0, TMP, base + kWordSize); |
2147 __ LoadFieldFromOffset(R1, R0, Function::instructions_offset(), PP); | 2145 __ LoadFieldFromOffset(R1, R0, Function::instructions_offset()); |
2148 // TODO(srdjan): Evaluate performance impact of moving the instruction below | 2146 // TODO(srdjan): Evaluate performance impact of moving the instruction below |
2149 // to the call site, instead of having it here. | 2147 // to the call site, instead of having it here. |
2150 __ AddImmediate(target, R1, Instructions::HeaderSize() - kHeapObjectTag, PP); | 2148 __ AddImmediate(target, R1, Instructions::HeaderSize() - kHeapObjectTag); |
2151 } | 2149 } |
2152 | 2150 |
2153 | 2151 |
2154 // Called from megamorphic calls. | 2152 // Called from megamorphic calls. |
2155 // R0: receiver. | 2153 // R0: receiver. |
2156 // R1: lookup cache. | 2154 // R1: lookup cache. |
2157 // Result: | 2155 // Result: |
2158 // R1: entry point. | 2156 // R1: entry point. |
2159 void StubCode::GenerateMegamorphicLookupStub(Assembler* assembler) { | 2157 void StubCode::GenerateMegamorphicLookupStub(Assembler* assembler) { |
2160 EmitMegamorphicLookup(assembler, R0, R1, R1); | 2158 EmitMegamorphicLookup(assembler, R0, R1, R1); |
2161 __ ret(); | 2159 __ ret(); |
2162 } | 2160 } |
2163 | 2161 |
2164 } // namespace dart | 2162 } // namespace dart |
2165 | 2163 |
2166 #endif // defined TARGET_ARCH_ARM64 | 2164 #endif // defined TARGET_ARCH_ARM64 |
OLD | NEW |