Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(53)

Side by Side Diff: runtime/vm/intrinsifier_arm.cc

Issue 2951333002: Moves the top_ and end_ words of the Scavenger into mutator thread. (Closed)
Patch Set: Removes the ZeroSizeScavenger test. Proper testing requires a second vm isolate. Created 3 years, 5 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View unified diff | Download patch
« no previous file with comments | « runtime/vm/heap.cc ('k') | runtime/vm/intrinsifier_arm64.cc » ('j') | no next file with comments »
Toggle Intra-line Diffs ('i') | Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
OLDNEW
1 // Copyright (c) 2013, the Dart project authors. Please see the AUTHORS file 1 // Copyright (c) 2013, the Dart project authors. Please see the AUTHORS file
2 // for details. All rights reserved. Use of this source code is governed by a 2 // for details. All rights reserved. Use of this source code is governed by a
3 // BSD-style license that can be found in the LICENSE file. 3 // BSD-style license that can be found in the LICENSE file.
4 4
5 #include "vm/globals.h" // Needed here to get TARGET_ARCH_ARM. 5 #include "vm/globals.h" // Needed here to get TARGET_ARCH_ARM.
6 #if defined(TARGET_ARCH_ARM) 6 #if defined(TARGET_ARCH_ARM)
7 7
8 #include "vm/intrinsifier.h" 8 #include "vm/intrinsifier.h"
9 9
10 #include "vm/assembler.h" 10 #include "vm/assembler.h"
(...skipping 162 matching lines...) Expand 10 before | Expand all | Expand 10 after
173 __ SmiUntag(R2); \ 173 __ SmiUntag(R2); \
174 /* Check for maximum allowed length. */ \ 174 /* Check for maximum allowed length. */ \
175 /* R2: untagged array length. */ \ 175 /* R2: untagged array length. */ \
176 __ CompareImmediate(R2, max_len); \ 176 __ CompareImmediate(R2, max_len); \
177 __ b(&fall_through, GT); \ 177 __ b(&fall_through, GT); \
178 __ mov(R2, Operand(R2, LSL, scale_shift)); \ 178 __ mov(R2, Operand(R2, LSL, scale_shift)); \
179 const intptr_t fixed_size_plus_alignment_padding = \ 179 const intptr_t fixed_size_plus_alignment_padding = \
180 sizeof(Raw##type_name) + kObjectAlignment - 1; \ 180 sizeof(Raw##type_name) + kObjectAlignment - 1; \
181 __ AddImmediate(R2, fixed_size_plus_alignment_padding); \ 181 __ AddImmediate(R2, fixed_size_plus_alignment_padding); \
182 __ bic(R2, R2, Operand(kObjectAlignment - 1)); \ 182 __ bic(R2, R2, Operand(kObjectAlignment - 1)); \
183 Heap::Space space = Heap::kNew; \ 183 NOT_IN_PRODUCT(Heap::Space space = Heap::kNew); \
184 __ ldr(R3, Address(THR, Thread::heap_offset())); \ 184 __ ldr(R0, Address(THR, Thread::top_offset())); \
185 __ ldr(R0, Address(R3, Heap::TopOffset(space))); \
186 \ 185 \
187 /* R2: allocation size. */ \ 186 /* R2: allocation size. */ \
188 __ adds(R1, R0, Operand(R2)); \ 187 __ adds(R1, R0, Operand(R2)); \
189 __ b(&fall_through, CS); /* Fail on unsigned overflow. */ \ 188 __ b(&fall_through, CS); /* Fail on unsigned overflow. */ \
190 \ 189 \
191 /* Check if the allocation fits into the remaining space. */ \ 190 /* Check if the allocation fits into the remaining space. */ \
192 /* R0: potential new object start. */ \ 191 /* R0: potential new object start. */ \
193 /* R1: potential next object start. */ \ 192 /* R1: potential next object start. */ \
194 /* R2: allocation size. */ \ 193 /* R2: allocation size. */ \
195 /* R3: heap. */ \ 194 __ ldr(IP, Address(THR, Thread::end_offset())); \
196 __ ldr(IP, Address(R3, Heap::EndOffset(space))); \
197 __ cmp(R1, Operand(IP)); \ 195 __ cmp(R1, Operand(IP)); \
198 __ b(&fall_through, CS); \ 196 __ b(&fall_through, CS); \
199 \ 197 \
200 /* Successfully allocated the object(s), now update top to point to */ \ 198 /* Successfully allocated the object(s), now update top to point to */ \
201 /* next object start and initialize the object. */ \ 199 /* next object start and initialize the object. */ \
202 NOT_IN_PRODUCT(__ LoadAllocationStatsAddress(R4, cid)); \ 200 NOT_IN_PRODUCT(__ LoadAllocationStatsAddress(R4, cid)); \
203 __ str(R1, Address(R3, Heap::TopOffset(space))); \ 201 __ str(R1, Address(THR, Thread::top_offset())); \
204 __ AddImmediate(R0, kHeapObjectTag); \ 202 __ AddImmediate(R0, kHeapObjectTag); \
205 /* Initialize the tags. */ \ 203 /* Initialize the tags. */ \
206 /* R0: new object start as a tagged pointer. */ \ 204 /* R0: new object start as a tagged pointer. */ \
207 /* R1: new object end address. */ \ 205 /* R1: new object end address. */ \
208 /* R2: allocation size. */ \ 206 /* R2: allocation size. */ \
209 /* R4: allocation stats address */ \ 207 /* R4: allocation stats address */ \
210 { \ 208 { \
211 __ CompareImmediate(R2, RawObject::SizeTag::kMaxSizeTag); \ 209 __ CompareImmediate(R2, RawObject::SizeTag::kMaxSizeTag); \
212 __ mov(R3, \ 210 __ mov(R3, \
213 Operand(R2, LSL, RawObject::kSizeTagPos - kObjectAlignmentLog2), \ 211 Operand(R2, LSL, RawObject::kSizeTagPos - kObjectAlignmentLog2), \
(...skipping 1793 matching lines...) Expand 10 before | Expand all | Expand 10 after
2007 NOT_IN_PRODUCT(__ MaybeTraceAllocation(kOneByteStringCid, R0, failure)); 2005 NOT_IN_PRODUCT(__ MaybeTraceAllocation(kOneByteStringCid, R0, failure));
2008 __ mov(R8, Operand(length_reg)); // Save the length register. 2006 __ mov(R8, Operand(length_reg)); // Save the length register.
2009 // TODO(koda): Protect against negative length and overflow here. 2007 // TODO(koda): Protect against negative length and overflow here.
2010 __ SmiUntag(length_reg); 2008 __ SmiUntag(length_reg);
2011 const intptr_t fixed_size_plus_alignment_padding = 2009 const intptr_t fixed_size_plus_alignment_padding =
2012 sizeof(RawString) + kObjectAlignment - 1; 2010 sizeof(RawString) + kObjectAlignment - 1;
2013 __ AddImmediate(length_reg, fixed_size_plus_alignment_padding); 2011 __ AddImmediate(length_reg, fixed_size_plus_alignment_padding);
2014 __ bic(length_reg, length_reg, Operand(kObjectAlignment - 1)); 2012 __ bic(length_reg, length_reg, Operand(kObjectAlignment - 1));
2015 2013
2016 const intptr_t cid = kOneByteStringCid; 2014 const intptr_t cid = kOneByteStringCid;
2017 Heap::Space space = Heap::kNew; 2015 NOT_IN_PRODUCT(Heap::Space space = Heap::kNew);
2018 __ ldr(R3, Address(THR, Thread::heap_offset())); 2016 __ ldr(R0, Address(THR, Thread::top_offset()));
2019 __ ldr(R0, Address(R3, Heap::TopOffset(space)));
2020 2017
2021 // length_reg: allocation size. 2018 // length_reg: allocation size.
2022 __ adds(R1, R0, Operand(length_reg)); 2019 __ adds(R1, R0, Operand(length_reg));
2023 __ b(&fail, CS); // Fail on unsigned overflow. 2020 __ b(&fail, CS); // Fail on unsigned overflow.
2024 2021
2025 // Check if the allocation fits into the remaining space. 2022 // Check if the allocation fits into the remaining space.
2026 // R0: potential new object start. 2023 // R0: potential new object start.
2027 // R1: potential next object start. 2024 // R1: potential next object start.
2028 // R2: allocation size. 2025 // R2: allocation size.
2029 // R3: heap. 2026 __ ldr(NOTFP, Address(THR, Thread::end_offset()));
2030 __ ldr(NOTFP, Address(R3, Heap::EndOffset(space)));
2031 __ cmp(R1, Operand(NOTFP)); 2027 __ cmp(R1, Operand(NOTFP));
2032 __ b(&fail, CS); 2028 __ b(&fail, CS);
2033 2029
2034 // Successfully allocated the object(s), now update top to point to 2030 // Successfully allocated the object(s), now update top to point to
2035 // next object start and initialize the object. 2031 // next object start and initialize the object.
2036 NOT_IN_PRODUCT(__ LoadAllocationStatsAddress(R4, cid)); 2032 NOT_IN_PRODUCT(__ LoadAllocationStatsAddress(R4, cid));
2037 __ str(R1, Address(R3, Heap::TopOffset(space))); 2033 __ str(R1, Address(THR, Thread::top_offset()));
2038 __ AddImmediate(R0, kHeapObjectTag); 2034 __ AddImmediate(R0, kHeapObjectTag);
2039 2035
2040 // Initialize the tags. 2036 // Initialize the tags.
2041 // R0: new object start as a tagged pointer. 2037 // R0: new object start as a tagged pointer.
2042 // R1: new object end address. 2038 // R1: new object end address.
2043 // R2: allocation size. 2039 // R2: allocation size.
2044 // R4: allocation stats address. 2040 // R4: allocation stats address.
2045 { 2041 {
2046 const intptr_t shift = RawObject::kSizeTagPos - kObjectAlignmentLog2; 2042 const intptr_t shift = RawObject::kSizeTagPos - kObjectAlignmentLog2;
2047 2043
(...skipping 274 matching lines...) Expand 10 before | Expand all | Expand 10 after
2322 2318
2323 void Intrinsifier::SetAsyncThreadStackTrace(Assembler* assembler) { 2319 void Intrinsifier::SetAsyncThreadStackTrace(Assembler* assembler) {
2324 __ ldr(R0, Address(THR, Thread::async_stack_trace_offset())); 2320 __ ldr(R0, Address(THR, Thread::async_stack_trace_offset()));
2325 __ LoadObject(R0, Object::null_object()); 2321 __ LoadObject(R0, Object::null_object());
2326 __ Ret(); 2322 __ Ret();
2327 } 2323 }
2328 2324
2329 } // namespace dart 2325 } // namespace dart
2330 2326
2331 #endif // defined TARGET_ARCH_ARM 2327 #endif // defined TARGET_ARCH_ARM
OLDNEW
« no previous file with comments | « runtime/vm/heap.cc ('k') | runtime/vm/intrinsifier_arm64.cc » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698