Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(50)

Unified Diff: src/ppc/code-stubs-ppc.cc

Issue 1895643002: Revert of [Atomics] Remove Atomics code stubs; use TF ops (Closed) Base URL: https://chromium.googlesource.com/v8/v8.git@master
Patch Set: Created 4 years, 8 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « src/mips64/interface-descriptors-mips64.cc ('k') | src/ppc/interface-descriptors-ppc.cc » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: src/ppc/code-stubs-ppc.cc
diff --git a/src/ppc/code-stubs-ppc.cc b/src/ppc/code-stubs-ppc.cc
index 560fd6e5d54eaad2211e318f29fa229b0ee4f15c..cafc3ceb3ec8fdeefa4b401c9d1a0147fadbc194 100644
--- a/src/ppc/code-stubs-ppc.cc
+++ b/src/ppc/code-stubs-ppc.cc
@@ -5788,6 +5788,193 @@
kStackUnwindSpace, NULL, return_value_operand, NULL);
}
+namespace {
+
+void GetTypedArrayBackingStore(MacroAssembler* masm, Register backing_store,
+ Register object, Register scratch,
+ DoubleRegister double_scratch) {
+ Label offset_is_not_smi, done_offset;
+ __ LoadP(scratch, FieldMemOperand(object, JSTypedArray::kBufferOffset));
+ __ LoadP(backing_store,
+ FieldMemOperand(scratch, JSArrayBuffer::kBackingStoreOffset));
+ __ LoadP(scratch,
+ FieldMemOperand(object, JSArrayBufferView::kByteOffsetOffset));
+ __ JumpIfNotSmi(scratch, &offset_is_not_smi);
+ // offset is smi
+ __ SmiUntag(scratch);
+ __ b(&done_offset);
+
+ // offset is a heap number
+ __ bind(&offset_is_not_smi);
+ __ lfd(double_scratch, FieldMemOperand(scratch, HeapNumber::kValueOffset));
+ __ ConvertDoubleToInt64(double_scratch,
+#if !V8_TARGET_ARCH_PPC64
+ r0,
+#endif
+ scratch, double_scratch);
+ __ bind(&done_offset);
+ __ add(backing_store, backing_store, scratch);
+}
+
+void TypedArrayJumpTablePrologue(MacroAssembler* masm, Register object,
+ Register scratch, Register scratch2,
+ Label* table) {
+ __ LoadP(scratch, FieldMemOperand(object, JSObject::kElementsOffset));
+ __ LoadP(scratch, FieldMemOperand(scratch, HeapObject::kMapOffset));
+ __ lbz(scratch, FieldMemOperand(scratch, Map::kInstanceTypeOffset));
+ __ subi(scratch, scratch,
+ Operand(static_cast<uint8_t>(FIXED_INT8_ARRAY_TYPE)));
+ if (__ emit_debug_code()) {
+ __ cmpi(scratch, Operand::Zero());
+ __ Check(ge, kOffsetOutOfRange);
+ }
+ __ ShiftLeftImm(scratch, scratch, Operand(kPointerSizeLog2));
+ __ mov_label_addr(scratch2, table);
+ __ LoadPX(scratch, MemOperand(scratch2, scratch));
+ __ Jump(scratch);
+}
+
+void TypedArrayJumpTableEpilogue(MacroAssembler* masm, Label* table, Label* i8,
+ Label* u8, Label* i16, Label* u16, Label* i32,
+ Label* u32, Label* u8c) {
+ STATIC_ASSERT(FIXED_UINT8_ARRAY_TYPE == FIXED_INT8_ARRAY_TYPE + 1);
+ STATIC_ASSERT(FIXED_INT16_ARRAY_TYPE == FIXED_INT8_ARRAY_TYPE + 2);
+ STATIC_ASSERT(FIXED_UINT16_ARRAY_TYPE == FIXED_INT8_ARRAY_TYPE + 3);
+ STATIC_ASSERT(FIXED_INT32_ARRAY_TYPE == FIXED_INT8_ARRAY_TYPE + 4);
+ STATIC_ASSERT(FIXED_UINT32_ARRAY_TYPE == FIXED_INT8_ARRAY_TYPE + 5);
+ STATIC_ASSERT(FIXED_FLOAT32_ARRAY_TYPE == FIXED_INT8_ARRAY_TYPE + 6);
+ STATIC_ASSERT(FIXED_FLOAT64_ARRAY_TYPE == FIXED_INT8_ARRAY_TYPE + 7);
+ STATIC_ASSERT(FIXED_UINT8_CLAMPED_ARRAY_TYPE == FIXED_INT8_ARRAY_TYPE + 8);
+
+ Label abort;
+ __ bind(table);
+ __ emit_label_addr(i8); // Int8Array
+ __ emit_label_addr(u8); // Uint8Array
+ __ emit_label_addr(i16); // Int16Array
+ __ emit_label_addr(u16); // Uint16Array
+ __ emit_label_addr(i32); // Int32Array
+ __ emit_label_addr(u32); // Uint32Array
+ __ emit_label_addr(&abort); // Float32Array
+ __ emit_label_addr(&abort); // Float64Array
+ __ emit_label_addr(u8c); // Uint8ClampedArray
+
+ __ bind(&abort);
+ __ Abort(kNoReason);
+}
+
+#if !V8_TARGET_ARCH_PPC64
+void ReturnInteger32(MacroAssembler* masm, DoubleRegister dst, Register value,
+ Label* use_heap_number) {
+ Label not_smi;
+ __ JumpIfNotSmiCandidate(value, r0, &not_smi);
+ __ SmiTag(r3, value);
+ __ blr();
+
+ __ bind(&not_smi);
+ __ ConvertIntToDouble(value, dst);
+ __ b(use_heap_number);
+}
+#endif
+
+void ReturnUnsignedInteger32(MacroAssembler* masm, DoubleRegister dst,
+ Register value, Label* use_heap_number) {
+ Label not_smi;
+ __ JumpIfNotUnsignedSmiCandidate(value, r0, &not_smi);
+ __ SmiTag(r3, value);
+ __ blr();
+
+ __ bind(&not_smi);
+ __ ConvertUnsignedIntToDouble(value, dst);
+ __ b(use_heap_number);
+}
+
+void ReturnAllocatedHeapNumber(MacroAssembler* masm, DoubleRegister value,
+ Register scratch, Register scratch2,
+ Register scratch3) {
+ Label call_runtime;
+ __ LoadRoot(scratch3, Heap::kHeapNumberMapRootIndex);
+ __ AllocateHeapNumber(r3, scratch, scratch2, scratch3, &call_runtime);
+ __ stfd(value, FieldMemOperand(r3, HeapNumber::kValueOffset));
+ __ blr();
+
+ __ bind(&call_runtime);
+ {
+ FrameAndConstantPoolScope scope(masm, StackFrame::INTERNAL);
+ __ CallRuntimeSaveDoubles(Runtime::kAllocateHeapNumber);
+ __ stfd(value, FieldMemOperand(r3, HeapNumber::kValueOffset));
+ }
+ __ blr();
+}
+
+} // anonymous namespace
+
+#define ASSEMBLE_ATOMIC_LOAD(instr, dst, base, index) \
+ do { \
+ Label not_taken; \
+ __ sync(); \
+ __ instr(dst, MemOperand(base, index)); \
+ __ bind(&not_taken); \
+ __ cmp(dst, dst); \
+ __ bne(&not_taken); \
+ __ isync(); \
+ } while (0)
+
+void AtomicsLoadStub::Generate(MacroAssembler* masm) {
+ Register object = r4;
+ Register index = r3; // Index is an untagged word32.
+ Register backing_store = r5;
+ Label table, i8, u8, i16, u16, i32, u32;
+
+ GetTypedArrayBackingStore(masm, backing_store, object, r6, d0);
+ TypedArrayJumpTablePrologue(masm, object, r6, r7, &table);
+
+ __ bind(&i8);
+ ASSEMBLE_ATOMIC_LOAD(lbzx, r3, backing_store, index);
+ __ extsb(r3, r3);
+ __ SmiTag(r3);
+ __ blr();
+
+ __ bind(&u8);
+ ASSEMBLE_ATOMIC_LOAD(lbzx, r3, backing_store, index);
+ __ SmiTag(r3);
+ __ blr();
+
+ __ bind(&i16);
+ __ ShiftLeftImm(index, index, Operand(1));
+ ASSEMBLE_ATOMIC_LOAD(lhax, r3, backing_store, index);
+ __ SmiTag(r3);
+ __ blr();
+
+ __ bind(&u16);
+ __ ShiftLeftImm(index, index, Operand(1));
+ ASSEMBLE_ATOMIC_LOAD(lhzx, r3, backing_store, index);
+ __ SmiTag(r3);
+ __ blr();
+
+ Label use_heap_number;
+
+ __ bind(&i32);
+ __ ShiftLeftImm(index, index, Operand(2));
+ ASSEMBLE_ATOMIC_LOAD(lwax, r3, backing_store, index);
+#if V8_TARGET_ARCH_PPC64
+ __ SmiTag(r3);
+ __ blr();
+#else
+ ReturnInteger32(masm, d0, r3, &use_heap_number);
+#endif
+
+ __ bind(&u32);
+ __ ShiftLeftImm(index, index, Operand(2));
+ ASSEMBLE_ATOMIC_LOAD(lwzx, r3, backing_store, index);
+ ReturnUnsignedInteger32(masm, d0, r3, &use_heap_number);
+
+ __ bind(&use_heap_number);
+ ReturnAllocatedHeapNumber(masm, d0, r4, r5, r6);
+
+ TypedArrayJumpTableEpilogue(masm, &table, &i8, &u8, &i16, &u16, &i32, &u32,
+ &u8);
+}
+
#undef __
} // namespace internal
} // namespace v8
« no previous file with comments | « src/mips64/interface-descriptors-mips64.cc ('k') | src/ppc/interface-descriptors-ppc.cc » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698