Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(315)

Unified Diff: src/arm64/builtins-arm64.cc

Issue 1641083003: [builtins] Make Math.max and Math.min fast by default. (Closed) Base URL: https://chromium.googlesource.com/v8/v8.git@master
Patch Set: SKIP unrelated ignition failures. Created 4 years, 11 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « src/arm/macro-assembler-arm.cc ('k') | src/bootstrapper.cc » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: src/arm64/builtins-arm64.cc
diff --git a/src/arm64/builtins-arm64.cc b/src/arm64/builtins-arm64.cc
index eb36c1bf4a258d09397f1d93b70286eabccba99a..cee5621ef855e8b85940db2709a23357f84cd69e 100644
--- a/src/arm64/builtins-arm64.cc
+++ b/src/arm64/builtins-arm64.cc
@@ -138,6 +138,110 @@ void Builtins::Generate_ArrayCode(MacroAssembler* masm) {
// static
+void Builtins::Generate_MathMaxMin(MacroAssembler* masm, MathMaxMinKind kind) {
+ // ----------- S t a t e -------------
+ // -- x0 : number of arguments
+ // -- lr : return address
+ // -- sp[(argc - n) * 8] : arg[n] (zero-based)
+ // -- sp[(argc + 1) * 8] : receiver
+ // -----------------------------------
+ ASM_LOCATION("Builtins::Generate_MathMaxMin");
+
+ Condition const cc_done = (kind == MathMaxMinKind::kMin) ? mi : gt;
+ Condition const cc_swap = (kind == MathMaxMinKind::kMin) ? gt : mi;
+ Heap::RootListIndex const root_index =
+ (kind == MathMaxMinKind::kMin) ? Heap::kInfinityValueRootIndex
+ : Heap::kMinusInfinityValueRootIndex;
+ DoubleRegister const reg = (kind == MathMaxMinKind::kMin) ? d2 : d1;
+
+ // Load the accumulator with the default return value (either -Infinity or
+ // +Infinity), with the tagged value in x1 and the double value in d1.
+ __ LoadRoot(x1, root_index);
+ __ Ldr(d1, FieldMemOperand(x1, HeapNumber::kValueOffset));
+ __ Mov(x4, x0);
jbramley 2016/01/28 18:15:35 x4 just seems to track how much we need to drop at
+
+ Label done_loop, loop;
+ __ Bind(&loop);
+ {
+ // Check if all parameters done.
+ __ Subs(x0, x0, 1);
+ __ B(lt, &done_loop);
+
+ // Load the next parameter tagged value into x2.
+ __ Peek(x2, Operand(x0, LSL, kPointerSizeLog2));
+
+ // Load the double value of the parameter into d2, maybe converting the
+ // parameter to a number first using the ToNumberStub if necessary.
+ Label convert, convert_smi, convert_number, done_convert;
+ __ Bind(&convert);
+ __ JumpIfSmi(x2, &convert_smi);
+ __ Ldr(x3, FieldMemOperand(x2, HeapObject::kMapOffset));
+ __ JumpIfRoot(x3, Heap::kHeapNumberMapRootIndex, &convert_number);
jbramley 2016/01/28 18:15:35 Ldr + JumpIfRoot can be JumpIfHeapNumber(x2, &conv
+ {
+ // Parameter is not a Number, use the ToNumberStub to convert it.
+ FrameScope scope(masm, StackFrame::INTERNAL);
+ __ SmiTag(x0);
+ __ SmiTag(x4);
+ __ Push(x0, x1, x4);
+ __ Mov(x0, x2);
+ ToNumberStub stub(masm->isolate());
+ __ CallStub(&stub);
+ __ Mov(x2, x0);
+ __ Pop(x4, x1, x0);
+ {
+ // Restore the double accumulator value (d1).
+ Label restore_smi, done_restore;
+ __ JumpIfSmi(x1, &restore_smi);
+ __ Ldr(d1, FieldMemOperand(x1, HeapNumber::kValueOffset));
+ __ B(&done_restore);
+ __ Bind(&restore_smi);
+ __ SmiUntagToDouble(d1, x1);
+ __ bind(&done_restore);
jbramley 2016/01/28 18:15:35 Bind
+ }
+ __ SmiUntag(x4);
+ __ SmiUntag(x0);
+ }
+ __ B(&convert);
+ __ Bind(&convert_number);
+ __ Ldr(d2, FieldMemOperand(x2, HeapNumber::kValueOffset));
+ __ B(&done_convert);
+ __ Bind(&convert_smi);
+ __ SmiUntagToDouble(d2, x2);
jbramley 2016/01/28 18:15:35 If a smi argument is the common case, we could unt
+ __ Bind(&done_convert);
+
+ // Perform the actual comparison with the accumulator value on the left hand
+ // side (d1) and the next parameter value on the right hand side (d2).
+ Label compare_nan, compare_swap;
+ __ Fcmp(d1, d2);
jbramley 2016/01/28 18:15:35 I think this whole operation could be replaced wit
+ __ B(cc_done, &loop);
+ __ B(cc_swap, &compare_swap);
+ __ B(vs, &compare_nan);
+
+ // Left and right hand side are equal, check for -0 vs. +0.
+ __ Fmov(x3, reg);
+ __ TestAndBranchIfAllClear(x3, V8_INT64_C(0x8000000000000000), &loop);
+
+ // Result is on the right hand side.
+ __ Bind(&compare_swap);
+ __ Fmov(d1, d2);
+ __ Mov(x1, x2);
+ __ B(&loop);
+
+ // At least one side is NaN, which means that the result will be NaN too.
+ __ Bind(&compare_nan);
+ __ LoadRoot(x1, Heap::kNanValueRootIndex);
+ __ Ldr(d1, FieldMemOperand(x1, HeapNumber::kValueOffset));
+ __ B(&loop);
+ }
+
+ __ Bind(&done_loop);
+ __ Mov(x0, x1);
+ __ Drop(x4);
+ __ Drop(1);
+ __ Ret();
+}
+
+// static
void Builtins::Generate_NumberConstructor(MacroAssembler* masm) {
// ----------- S t a t e -------------
// -- x0 : number of arguments
« no previous file with comments | « src/arm/macro-assembler-arm.cc ('k') | src/bootstrapper.cc » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698