OLD | NEW |
1 // Copyright (c) 2013, the Dart project authors. Please see the AUTHORS file | 1 // Copyright (c) 2013, the Dart project authors. Please see the AUTHORS file |
2 // for details. All rights reserved. Use of this source code is governed by a | 2 // for details. All rights reserved. Use of this source code is governed by a |
3 // BSD-style license that can be found in the LICENSE file. | 3 // BSD-style license that can be found in the LICENSE file. |
4 | 4 |
5 #include "vm/globals.h" // Needed here to get TARGET_ARCH_ARM. | 5 #include "vm/globals.h" // Needed here to get TARGET_ARCH_ARM. |
6 #if defined(TARGET_ARCH_ARM) | 6 #if defined(TARGET_ARCH_ARM) |
7 | 7 |
8 #include "vm/intrinsifier.h" | 8 #include "vm/intrinsifier.h" |
9 | 9 |
10 #include "vm/assembler.h" | 10 #include "vm/assembler.h" |
(...skipping 162 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
173 __ SmiUntag(R2); \ | 173 __ SmiUntag(R2); \ |
174 /* Check for maximum allowed length. */ \ | 174 /* Check for maximum allowed length. */ \ |
175 /* R2: untagged array length. */ \ | 175 /* R2: untagged array length. */ \ |
176 __ CompareImmediate(R2, max_len); \ | 176 __ CompareImmediate(R2, max_len); \ |
177 __ b(&fall_through, GT); \ | 177 __ b(&fall_through, GT); \ |
178 __ mov(R2, Operand(R2, LSL, scale_shift)); \ | 178 __ mov(R2, Operand(R2, LSL, scale_shift)); \ |
179 const intptr_t fixed_size_plus_alignment_padding = \ | 179 const intptr_t fixed_size_plus_alignment_padding = \ |
180 sizeof(Raw##type_name) + kObjectAlignment - 1; \ | 180 sizeof(Raw##type_name) + kObjectAlignment - 1; \ |
181 __ AddImmediate(R2, fixed_size_plus_alignment_padding); \ | 181 __ AddImmediate(R2, fixed_size_plus_alignment_padding); \ |
182 __ bic(R2, R2, Operand(kObjectAlignment - 1)); \ | 182 __ bic(R2, R2, Operand(kObjectAlignment - 1)); \ |
183 Heap::Space space = Heap::kNew; \ | 183 NOT_IN_PRODUCT(Heap::Space space = Heap::kNew); \ |
184 __ ldr(R3, Address(THR, Thread::heap_offset())); \ | 184 __ ldr(R0, Address(THR, Thread::top_offset())); \ |
185 __ ldr(R0, Address(R3, Heap::TopOffset(space))); \ | |
186 \ | 185 \ |
187 /* R2: allocation size. */ \ | 186 /* R2: allocation size. */ \ |
188 __ adds(R1, R0, Operand(R2)); \ | 187 __ adds(R1, R0, Operand(R2)); \ |
189 __ b(&fall_through, CS); /* Fail on unsigned overflow. */ \ | 188 __ b(&fall_through, CS); /* Fail on unsigned overflow. */ \ |
190 \ | 189 \ |
191 /* Check if the allocation fits into the remaining space. */ \ | 190 /* Check if the allocation fits into the remaining space. */ \ |
192 /* R0: potential new object start. */ \ | 191 /* R0: potential new object start. */ \ |
193 /* R1: potential next object start. */ \ | 192 /* R1: potential next object start. */ \ |
194 /* R2: allocation size. */ \ | 193 /* R2: allocation size. */ \ |
195 /* R3: heap. */ \ | 194 __ ldr(IP, Address(THR, Thread::end_offset())); \ |
196 __ ldr(IP, Address(R3, Heap::EndOffset(space))); \ | |
197 __ cmp(R1, Operand(IP)); \ | 195 __ cmp(R1, Operand(IP)); \ |
198 __ b(&fall_through, CS); \ | 196 __ b(&fall_through, CS); \ |
199 \ | 197 \ |
200 /* Successfully allocated the object(s), now update top to point to */ \ | 198 /* Successfully allocated the object(s), now update top to point to */ \ |
201 /* next object start and initialize the object. */ \ | 199 /* next object start and initialize the object. */ \ |
202 NOT_IN_PRODUCT(__ LoadAllocationStatsAddress(R4, cid)); \ | 200 NOT_IN_PRODUCT(__ LoadAllocationStatsAddress(R4, cid)); \ |
203 __ str(R1, Address(R3, Heap::TopOffset(space))); \ | 201 __ str(R1, Address(THR, Thread::top_offset())); \ |
204 __ AddImmediate(R0, kHeapObjectTag); \ | 202 __ AddImmediate(R0, kHeapObjectTag); \ |
205 /* Initialize the tags. */ \ | 203 /* Initialize the tags. */ \ |
206 /* R0: new object start as a tagged pointer. */ \ | 204 /* R0: new object start as a tagged pointer. */ \ |
207 /* R1: new object end address. */ \ | 205 /* R1: new object end address. */ \ |
208 /* R2: allocation size. */ \ | 206 /* R2: allocation size. */ \ |
209 /* R4: allocation stats address */ \ | 207 /* R4: allocation stats address */ \ |
210 { \ | 208 { \ |
211 __ CompareImmediate(R2, RawObject::SizeTag::kMaxSizeTag); \ | 209 __ CompareImmediate(R2, RawObject::SizeTag::kMaxSizeTag); \ |
212 __ mov(R3, \ | 210 __ mov(R3, \ |
213 Operand(R2, LSL, RawObject::kSizeTagPos - kObjectAlignmentLog2), \ | 211 Operand(R2, LSL, RawObject::kSizeTagPos - kObjectAlignmentLog2), \ |
(...skipping 1793 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
2007 NOT_IN_PRODUCT(__ MaybeTraceAllocation(kOneByteStringCid, R0, failure)); | 2005 NOT_IN_PRODUCT(__ MaybeTraceAllocation(kOneByteStringCid, R0, failure)); |
2008 __ mov(R8, Operand(length_reg)); // Save the length register. | 2006 __ mov(R8, Operand(length_reg)); // Save the length register. |
2009 // TODO(koda): Protect against negative length and overflow here. | 2007 // TODO(koda): Protect against negative length and overflow here. |
2010 __ SmiUntag(length_reg); | 2008 __ SmiUntag(length_reg); |
2011 const intptr_t fixed_size_plus_alignment_padding = | 2009 const intptr_t fixed_size_plus_alignment_padding = |
2012 sizeof(RawString) + kObjectAlignment - 1; | 2010 sizeof(RawString) + kObjectAlignment - 1; |
2013 __ AddImmediate(length_reg, fixed_size_plus_alignment_padding); | 2011 __ AddImmediate(length_reg, fixed_size_plus_alignment_padding); |
2014 __ bic(length_reg, length_reg, Operand(kObjectAlignment - 1)); | 2012 __ bic(length_reg, length_reg, Operand(kObjectAlignment - 1)); |
2015 | 2013 |
2016 const intptr_t cid = kOneByteStringCid; | 2014 const intptr_t cid = kOneByteStringCid; |
2017 Heap::Space space = Heap::kNew; | 2015 NOT_IN_PRODUCT(Heap::Space space = Heap::kNew); |
2018 __ ldr(R3, Address(THR, Thread::heap_offset())); | 2016 __ ldr(R0, Address(THR, Thread::top_offset())); |
2019 __ ldr(R0, Address(R3, Heap::TopOffset(space))); | |
2020 | 2017 |
2021 // length_reg: allocation size. | 2018 // length_reg: allocation size. |
2022 __ adds(R1, R0, Operand(length_reg)); | 2019 __ adds(R1, R0, Operand(length_reg)); |
2023 __ b(&fail, CS); // Fail on unsigned overflow. | 2020 __ b(&fail, CS); // Fail on unsigned overflow. |
2024 | 2021 |
2025 // Check if the allocation fits into the remaining space. | 2022 // Check if the allocation fits into the remaining space. |
2026 // R0: potential new object start. | 2023 // R0: potential new object start. |
2027 // R1: potential next object start. | 2024 // R1: potential next object start. |
2028 // R2: allocation size. | 2025 // R2: allocation size. |
2029 // R3: heap. | 2026 __ ldr(NOTFP, Address(THR, Thread::end_offset())); |
2030 __ ldr(NOTFP, Address(R3, Heap::EndOffset(space))); | |
2031 __ cmp(R1, Operand(NOTFP)); | 2027 __ cmp(R1, Operand(NOTFP)); |
2032 __ b(&fail, CS); | 2028 __ b(&fail, CS); |
2033 | 2029 |
2034 // Successfully allocated the object(s), now update top to point to | 2030 // Successfully allocated the object(s), now update top to point to |
2035 // next object start and initialize the object. | 2031 // next object start and initialize the object. |
2036 NOT_IN_PRODUCT(__ LoadAllocationStatsAddress(R4, cid)); | 2032 NOT_IN_PRODUCT(__ LoadAllocationStatsAddress(R4, cid)); |
2037 __ str(R1, Address(R3, Heap::TopOffset(space))); | 2033 __ str(R1, Address(THR, Thread::top_offset())); |
2038 __ AddImmediate(R0, kHeapObjectTag); | 2034 __ AddImmediate(R0, kHeapObjectTag); |
2039 | 2035 |
2040 // Initialize the tags. | 2036 // Initialize the tags. |
2041 // R0: new object start as a tagged pointer. | 2037 // R0: new object start as a tagged pointer. |
2042 // R1: new object end address. | 2038 // R1: new object end address. |
2043 // R2: allocation size. | 2039 // R2: allocation size. |
2044 // R4: allocation stats address. | 2040 // R4: allocation stats address. |
2045 { | 2041 { |
2046 const intptr_t shift = RawObject::kSizeTagPos - kObjectAlignmentLog2; | 2042 const intptr_t shift = RawObject::kSizeTagPos - kObjectAlignmentLog2; |
2047 | 2043 |
(...skipping 274 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
2322 | 2318 |
2323 void Intrinsifier::SetAsyncThreadStackTrace(Assembler* assembler) { | 2319 void Intrinsifier::SetAsyncThreadStackTrace(Assembler* assembler) { |
2324 __ ldr(R0, Address(THR, Thread::async_stack_trace_offset())); | 2320 __ ldr(R0, Address(THR, Thread::async_stack_trace_offset())); |
2325 __ LoadObject(R0, Object::null_object()); | 2321 __ LoadObject(R0, Object::null_object()); |
2326 __ Ret(); | 2322 __ Ret(); |
2327 } | 2323 } |
2328 | 2324 |
2329 } // namespace dart | 2325 } // namespace dart |
2330 | 2326 |
2331 #endif // defined TARGET_ARCH_ARM | 2327 #endif // defined TARGET_ARCH_ARM |
OLD | NEW |