Index: runtime/vm/intrinsifier_mips.cc |
diff --git a/runtime/vm/intrinsifier_mips.cc b/runtime/vm/intrinsifier_mips.cc |
index 1dca1af5810a6cbcd0fb7948334d4004b51cc3c8..ca95bb57822e6288234867193e56d6d3a164f56f 100644 |
--- a/runtime/vm/intrinsifier_mips.cc |
+++ b/runtime/vm/intrinsifier_mips.cc |
@@ -167,9 +167,8 @@ void Intrinsifier::GrowableArray_add(Assembler* assembler) { |
/* T2: untagged array length. */ \ |
__ BranchSignedGreater(T2, Immediate(max_len), &fall_through); \ |
__ sll(T2, T2, scale_shift); \ |
- const intptr_t fixed_size_plus_alignment_padding = \ |
- sizeof(Raw##type_name) + kObjectAlignment - 1; \ |
- __ AddImmediate(T2, fixed_size_plus_alignment_padding); \ |
+ const intptr_t fixed_size = sizeof(Raw##type_name) + kObjectAlignment - 1; \ |
+ __ AddImmediate(T2, fixed_size); \ |
__ LoadImmediate(TMP, -kObjectAlignment); \ |
__ and_(T2, T2, TMP); \ |
Heap::Space space = Heap::kNew; \ |
@@ -2132,9 +2131,8 @@ static void TryAllocateOnebyteString(Assembler* assembler, |
__ mov(T6, length_reg); // Save the length register. |
// TODO(koda): Protect against negative length and overflow here. |
__ SmiUntag(length_reg); |
- const intptr_t fixed_size_plus_alignment_padding = |
- sizeof(RawString) + kObjectAlignment - 1; |
- __ AddImmediate(length_reg, fixed_size_plus_alignment_padding); |
+ const intptr_t fixed_size = sizeof(RawString) + kObjectAlignment - 1; |
+ __ AddImmediate(length_reg, fixed_size); |
__ LoadImmediate(TMP, ~(kObjectAlignment - 1)); |
__ and_(length_reg, length_reg, TMP); |