| Index: src/runtime/runtime-atomics.cc
 | 
| diff --git a/src/runtime/runtime-atomics.cc b/src/runtime/runtime-atomics.cc
 | 
| index aece8d29bd15aaafd3e0817d6c73c4d699edc0d2..ff7ded9b090e365dced25ce8a3a6ce6f6ba4f0cb 100644
 | 
| --- a/src/runtime/runtime-atomics.cc
 | 
| +++ b/src/runtime/runtime-atomics.cc
 | 
| @@ -17,6 +17,312 @@
 | 
|  namespace v8 {
 | 
|  namespace internal {
 | 
|  
 | 
| +namespace {
 | 
| +
 | 
| +inline bool AtomicIsLockFree(uint32_t size) {
 | 
| +  return size == 1 || size == 2 || size == 4;
 | 
| +}
 | 
| +
 | 
| +#if V8_CC_GNU
 | 
| +
 | 
| +template <typename T>
 | 
| +inline T CompareExchangeSeqCst(T* p, T oldval, T newval) {
 | 
| +  (void)__atomic_compare_exchange_n(p, &oldval, newval, 0, __ATOMIC_SEQ_CST,
 | 
| +                                    __ATOMIC_SEQ_CST);
 | 
| +  return oldval;
 | 
| +}
 | 
| +
 | 
| +template <typename T>
 | 
| +inline T AddSeqCst(T* p, T value) {
 | 
| +  return __atomic_fetch_add(p, value, __ATOMIC_SEQ_CST);
 | 
| +}
 | 
| +
 | 
| +template <typename T>
 | 
| +inline T SubSeqCst(T* p, T value) {
 | 
| +  return __atomic_fetch_sub(p, value, __ATOMIC_SEQ_CST);
 | 
| +}
 | 
| +
 | 
| +template <typename T>
 | 
| +inline T AndSeqCst(T* p, T value) {
 | 
| +  return __atomic_fetch_and(p, value, __ATOMIC_SEQ_CST);
 | 
| +}
 | 
| +
 | 
| +template <typename T>
 | 
| +inline T OrSeqCst(T* p, T value) {
 | 
| +  return __atomic_fetch_or(p, value, __ATOMIC_SEQ_CST);
 | 
| +}
 | 
| +
 | 
| +template <typename T>
 | 
| +inline T XorSeqCst(T* p, T value) {
 | 
| +  return __atomic_fetch_xor(p, value, __ATOMIC_SEQ_CST);
 | 
| +}
 | 
| +
 | 
| +template <typename T>
 | 
| +inline T ExchangeSeqCst(T* p, T value) {
 | 
| +  return __atomic_exchange_n(p, value, __ATOMIC_SEQ_CST);
 | 
| +}
 | 
| +
 | 
| +#elif V8_CC_MSVC
 | 
| +
 | 
| +#define InterlockedCompareExchange32 _InterlockedCompareExchange
 | 
| +#define InterlockedExchange32 _InterlockedExchange
 | 
| +#define InterlockedExchangeAdd32 _InterlockedExchangeAdd
 | 
| +#define InterlockedAnd32 _InterlockedAnd
 | 
| +#define InterlockedOr32 _InterlockedOr
 | 
| +#define InterlockedXor32 _InterlockedXor
 | 
| +#define InterlockedExchangeAdd16 _InterlockedExchangeAdd16
 | 
| +#define InterlockedCompareExchange8 _InterlockedCompareExchange8
 | 
| +#define InterlockedExchangeAdd8 _InterlockedExchangeAdd8
 | 
| +
 | 
| +#define ATOMIC_OPS(type, suffix, vctype)                                    \
 | 
| +  inline type AddSeqCst(type* p, type value) {                              \
 | 
| +    return InterlockedExchangeAdd##suffix(reinterpret_cast<vctype*>(p),     \
 | 
| +                                          bit_cast<vctype>(value));         \
 | 
| +  }                                                                         \
 | 
| +  inline type SubSeqCst(type* p, type value) {                              \
 | 
| +    return InterlockedExchangeAdd##suffix(reinterpret_cast<vctype*>(p),     \
 | 
| +                                          -bit_cast<vctype>(value));        \
 | 
| +  }                                                                         \
 | 
| +  inline type AndSeqCst(type* p, type value) {                              \
 | 
| +    return InterlockedAnd##suffix(reinterpret_cast<vctype*>(p),             \
 | 
| +                                  bit_cast<vctype>(value));                 \
 | 
| +  }                                                                         \
 | 
| +  inline type OrSeqCst(type* p, type value) {                               \
 | 
| +    return InterlockedOr##suffix(reinterpret_cast<vctype*>(p),              \
 | 
| +                                 bit_cast<vctype>(value));                  \
 | 
| +  }                                                                         \
 | 
| +  inline type XorSeqCst(type* p, type value) {                              \
 | 
| +    return InterlockedXor##suffix(reinterpret_cast<vctype*>(p),             \
 | 
| +                                  bit_cast<vctype>(value));                 \
 | 
| +  }                                                                         \
 | 
| +  inline type ExchangeSeqCst(type* p, type value) {                         \
 | 
| +    return InterlockedExchange##suffix(reinterpret_cast<vctype*>(p),        \
 | 
| +                                       bit_cast<vctype>(value));            \
 | 
| +  }                                                                         \
 | 
| +                                                                            \
 | 
| +  inline type CompareExchangeSeqCst(type* p, type oldval, type newval) {    \
 | 
| +    return InterlockedCompareExchange##suffix(reinterpret_cast<vctype*>(p), \
 | 
| +                                              bit_cast<vctype>(newval),     \
 | 
| +                                              bit_cast<vctype>(oldval));    \
 | 
| +  }
 | 
| +
 | 
| +ATOMIC_OPS(int8_t, 8, char)
 | 
| +ATOMIC_OPS(uint8_t, 8, char)
 | 
| +ATOMIC_OPS(int16_t, 16, short)  /* NOLINT(runtime/int) */
 | 
| +ATOMIC_OPS(uint16_t, 16, short) /* NOLINT(runtime/int) */
 | 
| +ATOMIC_OPS(int32_t, 32, long)   /* NOLINT(runtime/int) */
 | 
| +ATOMIC_OPS(uint32_t, 32, long)  /* NOLINT(runtime/int) */
 | 
| +
 | 
| +#undef ATOMIC_OPS_INTEGER
 | 
| +#undef ATOMIC_OPS
 | 
| +
 | 
| +#undef InterlockedCompareExchange32
 | 
| +#undef InterlockedExchange32
 | 
| +#undef InterlockedExchangeAdd32
 | 
| +#undef InterlockedAnd32
 | 
| +#undef InterlockedOr32
 | 
| +#undef InterlockedXor32
 | 
| +#undef InterlockedExchangeAdd16
 | 
| +#undef InterlockedCompareExchange8
 | 
| +#undef InterlockedExchangeAdd8
 | 
| +
 | 
| +#else
 | 
| +
 | 
| +#error Unsupported platform!
 | 
| +
 | 
| +#endif
 | 
| +
 | 
| +template <typename T>
 | 
| +T FromObject(Handle<Object> number);
 | 
| +
 | 
| +template <>
 | 
| +inline uint8_t FromObject<uint8_t>(Handle<Object> number) {
 | 
| +  return NumberToUint32(*number);
 | 
| +}
 | 
| +
 | 
| +template <>
 | 
| +inline int8_t FromObject<int8_t>(Handle<Object> number) {
 | 
| +  return NumberToInt32(*number);
 | 
| +}
 | 
| +
 | 
| +template <>
 | 
| +inline uint16_t FromObject<uint16_t>(Handle<Object> number) {
 | 
| +  return NumberToUint32(*number);
 | 
| +}
 | 
| +
 | 
| +template <>
 | 
| +inline int16_t FromObject<int16_t>(Handle<Object> number) {
 | 
| +  return NumberToInt32(*number);
 | 
| +}
 | 
| +
 | 
| +template <>
 | 
| +inline uint32_t FromObject<uint32_t>(Handle<Object> number) {
 | 
| +  return NumberToUint32(*number);
 | 
| +}
 | 
| +
 | 
| +template <>
 | 
| +inline int32_t FromObject<int32_t>(Handle<Object> number) {
 | 
| +  return NumberToInt32(*number);
 | 
| +}
 | 
| +
 | 
| +
 | 
| +inline Object* ToObject(Isolate* isolate, int8_t t) { return Smi::FromInt(t); }
 | 
| +
 | 
| +inline Object* ToObject(Isolate* isolate, uint8_t t) { return Smi::FromInt(t); }
 | 
| +
 | 
| +inline Object* ToObject(Isolate* isolate, int16_t t) { return Smi::FromInt(t); }
 | 
| +
 | 
| +inline Object* ToObject(Isolate* isolate, uint16_t t) {
 | 
| +  return Smi::FromInt(t);
 | 
| +}
 | 
| +
 | 
| +
 | 
| +inline Object* ToObject(Isolate* isolate, int32_t t) {
 | 
| +  return *isolate->factory()->NewNumber(t);
 | 
| +}
 | 
| +
 | 
| +
 | 
| +inline Object* ToObject(Isolate* isolate, uint32_t t) {
 | 
| +  return *isolate->factory()->NewNumber(t);
 | 
| +}
 | 
| +
 | 
| +
 | 
| +template <typename T>
 | 
| +inline Object* DoCompareExchange(Isolate* isolate, void* buffer, size_t index,
 | 
| +                                 Handle<Object> oldobj, Handle<Object> newobj) {
 | 
| +  T oldval = FromObject<T>(oldobj);
 | 
| +  T newval = FromObject<T>(newobj);
 | 
| +  T result =
 | 
| +      CompareExchangeSeqCst(static_cast<T*>(buffer) + index, oldval, newval);
 | 
| +  return ToObject(isolate, result);
 | 
| +}
 | 
| +
 | 
| +
 | 
| +template <typename T>
 | 
| +inline Object* DoAdd(Isolate* isolate, void* buffer, size_t index,
 | 
| +                     Handle<Object> obj) {
 | 
| +  T value = FromObject<T>(obj);
 | 
| +  T result = AddSeqCst(static_cast<T*>(buffer) + index, value);
 | 
| +  return ToObject(isolate, result);
 | 
| +}
 | 
| +
 | 
| +
 | 
| +template <typename T>
 | 
| +inline Object* DoSub(Isolate* isolate, void* buffer, size_t index,
 | 
| +                     Handle<Object> obj) {
 | 
| +  T value = FromObject<T>(obj);
 | 
| +  T result = SubSeqCst(static_cast<T*>(buffer) + index, value);
 | 
| +  return ToObject(isolate, result);
 | 
| +}
 | 
| +
 | 
| +
 | 
| +template <typename T>
 | 
| +inline Object* DoAnd(Isolate* isolate, void* buffer, size_t index,
 | 
| +                     Handle<Object> obj) {
 | 
| +  T value = FromObject<T>(obj);
 | 
| +  T result = AndSeqCst(static_cast<T*>(buffer) + index, value);
 | 
| +  return ToObject(isolate, result);
 | 
| +}
 | 
| +
 | 
| +
 | 
| +template <typename T>
 | 
| +inline Object* DoOr(Isolate* isolate, void* buffer, size_t index,
 | 
| +                    Handle<Object> obj) {
 | 
| +  T value = FromObject<T>(obj);
 | 
| +  T result = OrSeqCst(static_cast<T*>(buffer) + index, value);
 | 
| +  return ToObject(isolate, result);
 | 
| +}
 | 
| +
 | 
| +
 | 
| +template <typename T>
 | 
| +inline Object* DoXor(Isolate* isolate, void* buffer, size_t index,
 | 
| +                     Handle<Object> obj) {
 | 
| +  T value = FromObject<T>(obj);
 | 
| +  T result = XorSeqCst(static_cast<T*>(buffer) + index, value);
 | 
| +  return ToObject(isolate, result);
 | 
| +}
 | 
| +
 | 
| +
 | 
| +template <typename T>
 | 
| +inline Object* DoExchange(Isolate* isolate, void* buffer, size_t index,
 | 
| +                          Handle<Object> obj) {
 | 
| +  T value = FromObject<T>(obj);
 | 
| +  T result = ExchangeSeqCst(static_cast<T*>(buffer) + index, value);
 | 
| +  return ToObject(isolate, result);
 | 
| +}
 | 
| +
 | 
| +
 | 
| +// Uint8Clamped functions
 | 
| +
 | 
| +uint8_t ClampToUint8(int32_t value) {
 | 
| +  if (value < 0) return 0;
 | 
| +  if (value > 255) return 255;
 | 
| +  return value;
 | 
| +}
 | 
| +
 | 
| +
 | 
| +inline Object* DoCompareExchangeUint8Clamped(Isolate* isolate, void* buffer,
 | 
| +                                             size_t index,
 | 
| +                                             Handle<Object> oldobj,
 | 
| +                                             Handle<Object> newobj) {
 | 
| +  typedef int32_t convert_type;
 | 
| +  uint8_t oldval = ClampToUint8(FromObject<convert_type>(oldobj));
 | 
| +  uint8_t newval = ClampToUint8(FromObject<convert_type>(newobj));
 | 
| +  uint8_t result = CompareExchangeSeqCst(static_cast<uint8_t*>(buffer) + index,
 | 
| +                                         oldval, newval);
 | 
| +  return ToObject(isolate, result);
 | 
| +}
 | 
| +
 | 
| +
 | 
| +#define DO_UINT8_CLAMPED_OP(name, op)                                        \
 | 
| +  inline Object* Do##name##Uint8Clamped(Isolate* isolate, void* buffer,      \
 | 
| +                                        size_t index, Handle<Object> obj) {  \
 | 
| +    typedef int32_t convert_type;                                            \
 | 
| +    uint8_t* p = static_cast<uint8_t*>(buffer) + index;                      \
 | 
| +    convert_type operand = FromObject<convert_type>(obj);                    \
 | 
| +    uint8_t expected;                                                        \
 | 
| +    uint8_t result;                                                          \
 | 
| +    do {                                                                     \
 | 
| +      expected = *p;                                                         \
 | 
| +      result = ClampToUint8(static_cast<convert_type>(expected) op operand); \
 | 
| +    } while (CompareExchangeSeqCst(p, expected, result) != expected);        \
 | 
| +    return ToObject(isolate, expected);                                      \
 | 
| +  }
 | 
| +
 | 
| +DO_UINT8_CLAMPED_OP(Add, +)
 | 
| +DO_UINT8_CLAMPED_OP(Sub, -)
 | 
| +DO_UINT8_CLAMPED_OP(And, &)
 | 
| +DO_UINT8_CLAMPED_OP(Or, | )
 | 
| +DO_UINT8_CLAMPED_OP(Xor, ^)
 | 
| +
 | 
| +#undef DO_UINT8_CLAMPED_OP
 | 
| +
 | 
| +
 | 
| +inline Object* DoExchangeUint8Clamped(Isolate* isolate, void* buffer,
 | 
| +                                      size_t index, Handle<Object> obj) {
 | 
| +  typedef int32_t convert_type;
 | 
| +  uint8_t* p = static_cast<uint8_t*>(buffer) + index;
 | 
| +  uint8_t result = ClampToUint8(FromObject<convert_type>(obj));
 | 
| +  uint8_t expected;
 | 
| +  do {
 | 
| +    expected = *p;
 | 
| +  } while (CompareExchangeSeqCst(p, expected, result) != expected);
 | 
| +  return ToObject(isolate, expected);
 | 
| +}
 | 
| +
 | 
| +
 | 
| +}  // anonymous namespace
 | 
| +
 | 
| +// Duplicated from objects.h
 | 
| +// V has parameters (Type, type, TYPE, C type, element_size)
 | 
| +#define INTEGER_TYPED_ARRAYS(V)          \
 | 
| +  V(Uint8, uint8, UINT8, uint8_t, 1)     \
 | 
| +  V(Int8, int8, INT8, int8_t, 1)         \
 | 
| +  V(Uint16, uint16, UINT16, uint16_t, 2) \
 | 
| +  V(Int16, int16, INT16, int16_t, 2)     \
 | 
| +  V(Uint32, uint32, UINT32, uint32_t, 4) \
 | 
| +  V(Int32, int32, INT32, int32_t, 4)
 | 
| +
 | 
|  RUNTIME_FUNCTION(Runtime_ThrowNotIntegerSharedTypedArrayError) {
 | 
|    HandleScope scope(isolate);
 | 
|    DCHECK_EQ(1, args.length());
 | 
| @@ -41,5 +347,238 @@ RUNTIME_FUNCTION(Runtime_ThrowInvalidAtomicAccessIndexError) {
 | 
|        isolate, NewRangeError(MessageTemplate::kInvalidAtomicAccessIndex));
 | 
|  }
 | 
|  
 | 
| +RUNTIME_FUNCTION(Runtime_AtomicsCompareExchange) {
 | 
| +  HandleScope scope(isolate);
 | 
| +  DCHECK_EQ(4, args.length());
 | 
| +  CONVERT_ARG_HANDLE_CHECKED(JSTypedArray, sta, 0);
 | 
| +  CONVERT_SIZE_ARG_CHECKED(index, 1);
 | 
| +  CONVERT_NUMBER_ARG_HANDLE_CHECKED(oldobj, 2);
 | 
| +  CONVERT_NUMBER_ARG_HANDLE_CHECKED(newobj, 3);
 | 
| +  CHECK(sta->GetBuffer()->is_shared());
 | 
| +  CHECK_LT(index, NumberToSize(sta->length()));
 | 
| +
 | 
| +  uint8_t* source = static_cast<uint8_t*>(sta->GetBuffer()->backing_store()) +
 | 
| +                    NumberToSize(sta->byte_offset());
 | 
| +
 | 
| +  switch (sta->type()) {
 | 
| +#define TYPED_ARRAY_CASE(Type, typeName, TYPE, ctype, size) \
 | 
| +  case kExternal##Type##Array:                              \
 | 
| +    return DoCompareExchange<ctype>(isolate, source, index, oldobj, newobj);
 | 
| +
 | 
| +    INTEGER_TYPED_ARRAYS(TYPED_ARRAY_CASE)
 | 
| +#undef TYPED_ARRAY_CASE
 | 
| +
 | 
| +    case kExternalUint8ClampedArray:
 | 
| +      return DoCompareExchangeUint8Clamped(isolate, source, index, oldobj,
 | 
| +                                           newobj);
 | 
| +
 | 
| +    default:
 | 
| +      break;
 | 
| +  }
 | 
| +
 | 
| +  UNREACHABLE();
 | 
| +  return isolate->heap()->undefined_value();
 | 
| +}
 | 
| +
 | 
| +
 | 
| +RUNTIME_FUNCTION(Runtime_AtomicsAdd) {
 | 
| +  HandleScope scope(isolate);
 | 
| +  DCHECK_EQ(3, args.length());
 | 
| +  CONVERT_ARG_HANDLE_CHECKED(JSTypedArray, sta, 0);
 | 
| +  CONVERT_SIZE_ARG_CHECKED(index, 1);
 | 
| +  CONVERT_NUMBER_ARG_HANDLE_CHECKED(value, 2);
 | 
| +  CHECK(sta->GetBuffer()->is_shared());
 | 
| +  CHECK_LT(index, NumberToSize(sta->length()));
 | 
| +
 | 
| +  uint8_t* source = static_cast<uint8_t*>(sta->GetBuffer()->backing_store()) +
 | 
| +                    NumberToSize(sta->byte_offset());
 | 
| +
 | 
| +  switch (sta->type()) {
 | 
| +#define TYPED_ARRAY_CASE(Type, typeName, TYPE, ctype, size) \
 | 
| +  case kExternal##Type##Array:                              \
 | 
| +    return DoAdd<ctype>(isolate, source, index, value);
 | 
| +
 | 
| +    INTEGER_TYPED_ARRAYS(TYPED_ARRAY_CASE)
 | 
| +#undef TYPED_ARRAY_CASE
 | 
| +
 | 
| +    case kExternalUint8ClampedArray:
 | 
| +      return DoAddUint8Clamped(isolate, source, index, value);
 | 
| +
 | 
| +    default:
 | 
| +      break;
 | 
| +  }
 | 
| +
 | 
| +  UNREACHABLE();
 | 
| +  return isolate->heap()->undefined_value();
 | 
| +}
 | 
| +
 | 
| +
 | 
| +RUNTIME_FUNCTION(Runtime_AtomicsSub) {
 | 
| +  HandleScope scope(isolate);
 | 
| +  DCHECK_EQ(3, args.length());
 | 
| +  CONVERT_ARG_HANDLE_CHECKED(JSTypedArray, sta, 0);
 | 
| +  CONVERT_SIZE_ARG_CHECKED(index, 1);
 | 
| +  CONVERT_NUMBER_ARG_HANDLE_CHECKED(value, 2);
 | 
| +  CHECK(sta->GetBuffer()->is_shared());
 | 
| +  CHECK_LT(index, NumberToSize(sta->length()));
 | 
| +
 | 
| +  uint8_t* source = static_cast<uint8_t*>(sta->GetBuffer()->backing_store()) +
 | 
| +                    NumberToSize(sta->byte_offset());
 | 
| +
 | 
| +  switch (sta->type()) {
 | 
| +#define TYPED_ARRAY_CASE(Type, typeName, TYPE, ctype, size) \
 | 
| +  case kExternal##Type##Array:                              \
 | 
| +    return DoSub<ctype>(isolate, source, index, value);
 | 
| +
 | 
| +    INTEGER_TYPED_ARRAYS(TYPED_ARRAY_CASE)
 | 
| +#undef TYPED_ARRAY_CASE
 | 
| +
 | 
| +    case kExternalUint8ClampedArray:
 | 
| +      return DoSubUint8Clamped(isolate, source, index, value);
 | 
| +
 | 
| +    default:
 | 
| +      break;
 | 
| +  }
 | 
| +
 | 
| +  UNREACHABLE();
 | 
| +  return isolate->heap()->undefined_value();
 | 
| +}
 | 
| +
 | 
| +
 | 
| +RUNTIME_FUNCTION(Runtime_AtomicsAnd) {
 | 
| +  HandleScope scope(isolate);
 | 
| +  DCHECK_EQ(3, args.length());
 | 
| +  CONVERT_ARG_HANDLE_CHECKED(JSTypedArray, sta, 0);
 | 
| +  CONVERT_SIZE_ARG_CHECKED(index, 1);
 | 
| +  CONVERT_NUMBER_ARG_HANDLE_CHECKED(value, 2);
 | 
| +  CHECK(sta->GetBuffer()->is_shared());
 | 
| +  CHECK_LT(index, NumberToSize(sta->length()));
 | 
| +
 | 
| +  uint8_t* source = static_cast<uint8_t*>(sta->GetBuffer()->backing_store()) +
 | 
| +                    NumberToSize(sta->byte_offset());
 | 
| +
 | 
| +  switch (sta->type()) {
 | 
| +#define TYPED_ARRAY_CASE(Type, typeName, TYPE, ctype, size) \
 | 
| +  case kExternal##Type##Array:                              \
 | 
| +    return DoAnd<ctype>(isolate, source, index, value);
 | 
| +
 | 
| +    INTEGER_TYPED_ARRAYS(TYPED_ARRAY_CASE)
 | 
| +#undef TYPED_ARRAY_CASE
 | 
| +
 | 
| +    case kExternalUint8ClampedArray:
 | 
| +      return DoAndUint8Clamped(isolate, source, index, value);
 | 
| +
 | 
| +    default:
 | 
| +      break;
 | 
| +  }
 | 
| +
 | 
| +  UNREACHABLE();
 | 
| +  return isolate->heap()->undefined_value();
 | 
| +}
 | 
| +
 | 
| +
 | 
| +RUNTIME_FUNCTION(Runtime_AtomicsOr) {
 | 
| +  HandleScope scope(isolate);
 | 
| +  DCHECK_EQ(3, args.length());
 | 
| +  CONVERT_ARG_HANDLE_CHECKED(JSTypedArray, sta, 0);
 | 
| +  CONVERT_SIZE_ARG_CHECKED(index, 1);
 | 
| +  CONVERT_NUMBER_ARG_HANDLE_CHECKED(value, 2);
 | 
| +  CHECK(sta->GetBuffer()->is_shared());
 | 
| +  CHECK_LT(index, NumberToSize(sta->length()));
 | 
| +
 | 
| +  uint8_t* source = static_cast<uint8_t*>(sta->GetBuffer()->backing_store()) +
 | 
| +                    NumberToSize(sta->byte_offset());
 | 
| +
 | 
| +  switch (sta->type()) {
 | 
| +#define TYPED_ARRAY_CASE(Type, typeName, TYPE, ctype, size) \
 | 
| +  case kExternal##Type##Array:                              \
 | 
| +    return DoOr<ctype>(isolate, source, index, value);
 | 
| +
 | 
| +    INTEGER_TYPED_ARRAYS(TYPED_ARRAY_CASE)
 | 
| +#undef TYPED_ARRAY_CASE
 | 
| +
 | 
| +    case kExternalUint8ClampedArray:
 | 
| +      return DoOrUint8Clamped(isolate, source, index, value);
 | 
| +
 | 
| +    default:
 | 
| +      break;
 | 
| +  }
 | 
| +
 | 
| +  UNREACHABLE();
 | 
| +  return isolate->heap()->undefined_value();
 | 
| +}
 | 
| +
 | 
| +
 | 
| +RUNTIME_FUNCTION(Runtime_AtomicsXor) {
 | 
| +  HandleScope scope(isolate);
 | 
| +  DCHECK_EQ(3, args.length());
 | 
| +  CONVERT_ARG_HANDLE_CHECKED(JSTypedArray, sta, 0);
 | 
| +  CONVERT_SIZE_ARG_CHECKED(index, 1);
 | 
| +  CONVERT_NUMBER_ARG_HANDLE_CHECKED(value, 2);
 | 
| +  CHECK(sta->GetBuffer()->is_shared());
 | 
| +  CHECK_LT(index, NumberToSize(sta->length()));
 | 
| +
 | 
| +  uint8_t* source = static_cast<uint8_t*>(sta->GetBuffer()->backing_store()) +
 | 
| +                    NumberToSize(sta->byte_offset());
 | 
| +
 | 
| +  switch (sta->type()) {
 | 
| +#define TYPED_ARRAY_CASE(Type, typeName, TYPE, ctype, size) \
 | 
| +  case kExternal##Type##Array:                              \
 | 
| +    return DoXor<ctype>(isolate, source, index, value);
 | 
| +
 | 
| +    INTEGER_TYPED_ARRAYS(TYPED_ARRAY_CASE)
 | 
| +#undef TYPED_ARRAY_CASE
 | 
| +
 | 
| +    case kExternalUint8ClampedArray:
 | 
| +      return DoXorUint8Clamped(isolate, source, index, value);
 | 
| +
 | 
| +    default:
 | 
| +      break;
 | 
| +  }
 | 
| +
 | 
| +  UNREACHABLE();
 | 
| +  return isolate->heap()->undefined_value();
 | 
| +}
 | 
| +
 | 
| +
 | 
| +RUNTIME_FUNCTION(Runtime_AtomicsExchange) {
 | 
| +  HandleScope scope(isolate);
 | 
| +  DCHECK_EQ(3, args.length());
 | 
| +  CONVERT_ARG_HANDLE_CHECKED(JSTypedArray, sta, 0);
 | 
| +  CONVERT_SIZE_ARG_CHECKED(index, 1);
 | 
| +  CONVERT_NUMBER_ARG_HANDLE_CHECKED(value, 2);
 | 
| +  CHECK(sta->GetBuffer()->is_shared());
 | 
| +  CHECK_LT(index, NumberToSize(sta->length()));
 | 
| +
 | 
| +  uint8_t* source = static_cast<uint8_t*>(sta->GetBuffer()->backing_store()) +
 | 
| +                    NumberToSize(sta->byte_offset());
 | 
| +
 | 
| +  switch (sta->type()) {
 | 
| +#define TYPED_ARRAY_CASE(Type, typeName, TYPE, ctype, size) \
 | 
| +  case kExternal##Type##Array:                              \
 | 
| +    return DoExchange<ctype>(isolate, source, index, value);
 | 
| +
 | 
| +    INTEGER_TYPED_ARRAYS(TYPED_ARRAY_CASE)
 | 
| +#undef TYPED_ARRAY_CASE
 | 
| +
 | 
| +    case kExternalUint8ClampedArray:
 | 
| +      return DoExchangeUint8Clamped(isolate, source, index, value);
 | 
| +
 | 
| +    default:
 | 
| +      break;
 | 
| +  }
 | 
| +
 | 
| +  UNREACHABLE();
 | 
| +  return isolate->heap()->undefined_value();
 | 
| +}
 | 
| +
 | 
| +
 | 
| +RUNTIME_FUNCTION(Runtime_AtomicsIsLockFree) {
 | 
| +  HandleScope scope(isolate);
 | 
| +  DCHECK_EQ(1, args.length());
 | 
| +  CONVERT_NUMBER_ARG_HANDLE_CHECKED(size, 0);
 | 
| +  uint32_t usize = NumberToUint32(*size);
 | 
| +  return isolate->heap()->ToBoolean(AtomicIsLockFree(usize));
 | 
| +}
 | 
|  }  // namespace internal
 | 
|  }  // namespace v8
 | 
| 
 |