| Index: src/runtime/runtime-atomics.cc
|
| diff --git a/src/runtime/runtime-atomics.cc b/src/runtime/runtime-atomics.cc
|
| index aece8d29bd15aaafd3e0817d6c73c4d699edc0d2..ff7ded9b090e365dced25ce8a3a6ce6f6ba4f0cb 100644
|
| --- a/src/runtime/runtime-atomics.cc
|
| +++ b/src/runtime/runtime-atomics.cc
|
| @@ -17,6 +17,312 @@
|
| namespace v8 {
|
| namespace internal {
|
|
|
| +namespace {
|
| +
|
| +inline bool AtomicIsLockFree(uint32_t size) {
|
| + return size == 1 || size == 2 || size == 4;
|
| +}
|
| +
|
| +#if V8_CC_GNU
|
| +
|
| +template <typename T>
|
| +inline T CompareExchangeSeqCst(T* p, T oldval, T newval) {
|
| + (void)__atomic_compare_exchange_n(p, &oldval, newval, 0, __ATOMIC_SEQ_CST,
|
| + __ATOMIC_SEQ_CST);
|
| + return oldval;
|
| +}
|
| +
|
| +template <typename T>
|
| +inline T AddSeqCst(T* p, T value) {
|
| + return __atomic_fetch_add(p, value, __ATOMIC_SEQ_CST);
|
| +}
|
| +
|
| +template <typename T>
|
| +inline T SubSeqCst(T* p, T value) {
|
| + return __atomic_fetch_sub(p, value, __ATOMIC_SEQ_CST);
|
| +}
|
| +
|
| +template <typename T>
|
| +inline T AndSeqCst(T* p, T value) {
|
| + return __atomic_fetch_and(p, value, __ATOMIC_SEQ_CST);
|
| +}
|
| +
|
| +template <typename T>
|
| +inline T OrSeqCst(T* p, T value) {
|
| + return __atomic_fetch_or(p, value, __ATOMIC_SEQ_CST);
|
| +}
|
| +
|
| +template <typename T>
|
| +inline T XorSeqCst(T* p, T value) {
|
| + return __atomic_fetch_xor(p, value, __ATOMIC_SEQ_CST);
|
| +}
|
| +
|
| +template <typename T>
|
| +inline T ExchangeSeqCst(T* p, T value) {
|
| + return __atomic_exchange_n(p, value, __ATOMIC_SEQ_CST);
|
| +}
|
| +
|
| +#elif V8_CC_MSVC
|
| +
|
| +#define InterlockedCompareExchange32 _InterlockedCompareExchange
|
| +#define InterlockedExchange32 _InterlockedExchange
|
| +#define InterlockedExchangeAdd32 _InterlockedExchangeAdd
|
| +#define InterlockedAnd32 _InterlockedAnd
|
| +#define InterlockedOr32 _InterlockedOr
|
| +#define InterlockedXor32 _InterlockedXor
|
| +#define InterlockedExchangeAdd16 _InterlockedExchangeAdd16
|
| +#define InterlockedCompareExchange8 _InterlockedCompareExchange8
|
| +#define InterlockedExchangeAdd8 _InterlockedExchangeAdd8
|
| +
|
| +#define ATOMIC_OPS(type, suffix, vctype) \
|
| + inline type AddSeqCst(type* p, type value) { \
|
| + return InterlockedExchangeAdd##suffix(reinterpret_cast<vctype*>(p), \
|
| + bit_cast<vctype>(value)); \
|
| + } \
|
| + inline type SubSeqCst(type* p, type value) { \
|
| + return InterlockedExchangeAdd##suffix(reinterpret_cast<vctype*>(p), \
|
| + -bit_cast<vctype>(value)); \
|
| + } \
|
| + inline type AndSeqCst(type* p, type value) { \
|
| + return InterlockedAnd##suffix(reinterpret_cast<vctype*>(p), \
|
| + bit_cast<vctype>(value)); \
|
| + } \
|
| + inline type OrSeqCst(type* p, type value) { \
|
| + return InterlockedOr##suffix(reinterpret_cast<vctype*>(p), \
|
| + bit_cast<vctype>(value)); \
|
| + } \
|
| + inline type XorSeqCst(type* p, type value) { \
|
| + return InterlockedXor##suffix(reinterpret_cast<vctype*>(p), \
|
| + bit_cast<vctype>(value)); \
|
| + } \
|
| + inline type ExchangeSeqCst(type* p, type value) { \
|
| + return InterlockedExchange##suffix(reinterpret_cast<vctype*>(p), \
|
| + bit_cast<vctype>(value)); \
|
| + } \
|
| + \
|
| + inline type CompareExchangeSeqCst(type* p, type oldval, type newval) { \
|
| + return InterlockedCompareExchange##suffix(reinterpret_cast<vctype*>(p), \
|
| + bit_cast<vctype>(newval), \
|
| + bit_cast<vctype>(oldval)); \
|
| + }
|
| +
|
| +ATOMIC_OPS(int8_t, 8, char)
|
| +ATOMIC_OPS(uint8_t, 8, char)
|
| +ATOMIC_OPS(int16_t, 16, short) /* NOLINT(runtime/int) */
|
| +ATOMIC_OPS(uint16_t, 16, short) /* NOLINT(runtime/int) */
|
| +ATOMIC_OPS(int32_t, 32, long) /* NOLINT(runtime/int) */
|
| +ATOMIC_OPS(uint32_t, 32, long) /* NOLINT(runtime/int) */
|
| +
|
| +#undef ATOMIC_OPS_INTEGER
|
| +#undef ATOMIC_OPS
|
| +
|
| +#undef InterlockedCompareExchange32
|
| +#undef InterlockedExchange32
|
| +#undef InterlockedExchangeAdd32
|
| +#undef InterlockedAnd32
|
| +#undef InterlockedOr32
|
| +#undef InterlockedXor32
|
| +#undef InterlockedExchangeAdd16
|
| +#undef InterlockedCompareExchange8
|
| +#undef InterlockedExchangeAdd8
|
| +
|
| +#else
|
| +
|
| +#error Unsupported platform!
|
| +
|
| +#endif
|
| +
|
| +template <typename T>
|
| +T FromObject(Handle<Object> number);
|
| +
|
| +template <>
|
| +inline uint8_t FromObject<uint8_t>(Handle<Object> number) {
|
| + return NumberToUint32(*number);
|
| +}
|
| +
|
| +template <>
|
| +inline int8_t FromObject<int8_t>(Handle<Object> number) {
|
| + return NumberToInt32(*number);
|
| +}
|
| +
|
| +template <>
|
| +inline uint16_t FromObject<uint16_t>(Handle<Object> number) {
|
| + return NumberToUint32(*number);
|
| +}
|
| +
|
| +template <>
|
| +inline int16_t FromObject<int16_t>(Handle<Object> number) {
|
| + return NumberToInt32(*number);
|
| +}
|
| +
|
| +template <>
|
| +inline uint32_t FromObject<uint32_t>(Handle<Object> number) {
|
| + return NumberToUint32(*number);
|
| +}
|
| +
|
| +template <>
|
| +inline int32_t FromObject<int32_t>(Handle<Object> number) {
|
| + return NumberToInt32(*number);
|
| +}
|
| +
|
| +
|
| +inline Object* ToObject(Isolate* isolate, int8_t t) { return Smi::FromInt(t); }
|
| +
|
| +inline Object* ToObject(Isolate* isolate, uint8_t t) { return Smi::FromInt(t); }
|
| +
|
| +inline Object* ToObject(Isolate* isolate, int16_t t) { return Smi::FromInt(t); }
|
| +
|
| +inline Object* ToObject(Isolate* isolate, uint16_t t) {
|
| + return Smi::FromInt(t);
|
| +}
|
| +
|
| +
|
| +inline Object* ToObject(Isolate* isolate, int32_t t) {
|
| + return *isolate->factory()->NewNumber(t);
|
| +}
|
| +
|
| +
|
| +inline Object* ToObject(Isolate* isolate, uint32_t t) {
|
| + return *isolate->factory()->NewNumber(t);
|
| +}
|
| +
|
| +
|
| +template <typename T>
|
| +inline Object* DoCompareExchange(Isolate* isolate, void* buffer, size_t index,
|
| + Handle<Object> oldobj, Handle<Object> newobj) {
|
| + T oldval = FromObject<T>(oldobj);
|
| + T newval = FromObject<T>(newobj);
|
| + T result =
|
| + CompareExchangeSeqCst(static_cast<T*>(buffer) + index, oldval, newval);
|
| + return ToObject(isolate, result);
|
| +}
|
| +
|
| +
|
| +template <typename T>
|
| +inline Object* DoAdd(Isolate* isolate, void* buffer, size_t index,
|
| + Handle<Object> obj) {
|
| + T value = FromObject<T>(obj);
|
| + T result = AddSeqCst(static_cast<T*>(buffer) + index, value);
|
| + return ToObject(isolate, result);
|
| +}
|
| +
|
| +
|
| +template <typename T>
|
| +inline Object* DoSub(Isolate* isolate, void* buffer, size_t index,
|
| + Handle<Object> obj) {
|
| + T value = FromObject<T>(obj);
|
| + T result = SubSeqCst(static_cast<T*>(buffer) + index, value);
|
| + return ToObject(isolate, result);
|
| +}
|
| +
|
| +
|
| +template <typename T>
|
| +inline Object* DoAnd(Isolate* isolate, void* buffer, size_t index,
|
| + Handle<Object> obj) {
|
| + T value = FromObject<T>(obj);
|
| + T result = AndSeqCst(static_cast<T*>(buffer) + index, value);
|
| + return ToObject(isolate, result);
|
| +}
|
| +
|
| +
|
| +template <typename T>
|
| +inline Object* DoOr(Isolate* isolate, void* buffer, size_t index,
|
| + Handle<Object> obj) {
|
| + T value = FromObject<T>(obj);
|
| + T result = OrSeqCst(static_cast<T*>(buffer) + index, value);
|
| + return ToObject(isolate, result);
|
| +}
|
| +
|
| +
|
| +template <typename T>
|
| +inline Object* DoXor(Isolate* isolate, void* buffer, size_t index,
|
| + Handle<Object> obj) {
|
| + T value = FromObject<T>(obj);
|
| + T result = XorSeqCst(static_cast<T*>(buffer) + index, value);
|
| + return ToObject(isolate, result);
|
| +}
|
| +
|
| +
|
| +template <typename T>
|
| +inline Object* DoExchange(Isolate* isolate, void* buffer, size_t index,
|
| + Handle<Object> obj) {
|
| + T value = FromObject<T>(obj);
|
| + T result = ExchangeSeqCst(static_cast<T*>(buffer) + index, value);
|
| + return ToObject(isolate, result);
|
| +}
|
| +
|
| +
|
| +// Uint8Clamped functions
|
| +
|
| +uint8_t ClampToUint8(int32_t value) {
|
| + if (value < 0) return 0;
|
| + if (value > 255) return 255;
|
| + return value;
|
| +}
|
| +
|
| +
|
| +inline Object* DoCompareExchangeUint8Clamped(Isolate* isolate, void* buffer,
|
| + size_t index,
|
| + Handle<Object> oldobj,
|
| + Handle<Object> newobj) {
|
| + typedef int32_t convert_type;
|
| + uint8_t oldval = ClampToUint8(FromObject<convert_type>(oldobj));
|
| + uint8_t newval = ClampToUint8(FromObject<convert_type>(newobj));
|
| + uint8_t result = CompareExchangeSeqCst(static_cast<uint8_t*>(buffer) + index,
|
| + oldval, newval);
|
| + return ToObject(isolate, result);
|
| +}
|
| +
|
| +
|
| +#define DO_UINT8_CLAMPED_OP(name, op) \
|
| + inline Object* Do##name##Uint8Clamped(Isolate* isolate, void* buffer, \
|
| + size_t index, Handle<Object> obj) { \
|
| + typedef int32_t convert_type; \
|
| + uint8_t* p = static_cast<uint8_t*>(buffer) + index; \
|
| + convert_type operand = FromObject<convert_type>(obj); \
|
| + uint8_t expected; \
|
| + uint8_t result; \
|
| + do { \
|
| + expected = *p; \
|
| + result = ClampToUint8(static_cast<convert_type>(expected) op operand); \
|
| + } while (CompareExchangeSeqCst(p, expected, result) != expected); \
|
| + return ToObject(isolate, expected); \
|
| + }
|
| +
|
| +DO_UINT8_CLAMPED_OP(Add, +)
|
| +DO_UINT8_CLAMPED_OP(Sub, -)
|
| +DO_UINT8_CLAMPED_OP(And, &)
|
| +DO_UINT8_CLAMPED_OP(Or, | )
|
| +DO_UINT8_CLAMPED_OP(Xor, ^)
|
| +
|
| +#undef DO_UINT8_CLAMPED_OP
|
| +
|
| +
|
| +inline Object* DoExchangeUint8Clamped(Isolate* isolate, void* buffer,
|
| + size_t index, Handle<Object> obj) {
|
| + typedef int32_t convert_type;
|
| + uint8_t* p = static_cast<uint8_t*>(buffer) + index;
|
| + uint8_t result = ClampToUint8(FromObject<convert_type>(obj));
|
| + uint8_t expected;
|
| + do {
|
| + expected = *p;
|
| + } while (CompareExchangeSeqCst(p, expected, result) != expected);
|
| + return ToObject(isolate, expected);
|
| +}
|
| +
|
| +
|
| +} // anonymous namespace
|
| +
|
| +// Duplicated from objects.h
|
| +// V has parameters (Type, type, TYPE, C type, element_size)
|
| +#define INTEGER_TYPED_ARRAYS(V) \
|
| + V(Uint8, uint8, UINT8, uint8_t, 1) \
|
| + V(Int8, int8, INT8, int8_t, 1) \
|
| + V(Uint16, uint16, UINT16, uint16_t, 2) \
|
| + V(Int16, int16, INT16, int16_t, 2) \
|
| + V(Uint32, uint32, UINT32, uint32_t, 4) \
|
| + V(Int32, int32, INT32, int32_t, 4)
|
| +
|
| RUNTIME_FUNCTION(Runtime_ThrowNotIntegerSharedTypedArrayError) {
|
| HandleScope scope(isolate);
|
| DCHECK_EQ(1, args.length());
|
| @@ -41,5 +347,238 @@ RUNTIME_FUNCTION(Runtime_ThrowInvalidAtomicAccessIndexError) {
|
| isolate, NewRangeError(MessageTemplate::kInvalidAtomicAccessIndex));
|
| }
|
|
|
| +RUNTIME_FUNCTION(Runtime_AtomicsCompareExchange) {
|
| + HandleScope scope(isolate);
|
| + DCHECK_EQ(4, args.length());
|
| + CONVERT_ARG_HANDLE_CHECKED(JSTypedArray, sta, 0);
|
| + CONVERT_SIZE_ARG_CHECKED(index, 1);
|
| + CONVERT_NUMBER_ARG_HANDLE_CHECKED(oldobj, 2);
|
| + CONVERT_NUMBER_ARG_HANDLE_CHECKED(newobj, 3);
|
| + CHECK(sta->GetBuffer()->is_shared());
|
| + CHECK_LT(index, NumberToSize(sta->length()));
|
| +
|
| + uint8_t* source = static_cast<uint8_t*>(sta->GetBuffer()->backing_store()) +
|
| + NumberToSize(sta->byte_offset());
|
| +
|
| + switch (sta->type()) {
|
| +#define TYPED_ARRAY_CASE(Type, typeName, TYPE, ctype, size) \
|
| + case kExternal##Type##Array: \
|
| + return DoCompareExchange<ctype>(isolate, source, index, oldobj, newobj);
|
| +
|
| + INTEGER_TYPED_ARRAYS(TYPED_ARRAY_CASE)
|
| +#undef TYPED_ARRAY_CASE
|
| +
|
| + case kExternalUint8ClampedArray:
|
| + return DoCompareExchangeUint8Clamped(isolate, source, index, oldobj,
|
| + newobj);
|
| +
|
| + default:
|
| + break;
|
| + }
|
| +
|
| + UNREACHABLE();
|
| + return isolate->heap()->undefined_value();
|
| +}
|
| +
|
| +
|
| +RUNTIME_FUNCTION(Runtime_AtomicsAdd) {
|
| + HandleScope scope(isolate);
|
| + DCHECK_EQ(3, args.length());
|
| + CONVERT_ARG_HANDLE_CHECKED(JSTypedArray, sta, 0);
|
| + CONVERT_SIZE_ARG_CHECKED(index, 1);
|
| + CONVERT_NUMBER_ARG_HANDLE_CHECKED(value, 2);
|
| + CHECK(sta->GetBuffer()->is_shared());
|
| + CHECK_LT(index, NumberToSize(sta->length()));
|
| +
|
| + uint8_t* source = static_cast<uint8_t*>(sta->GetBuffer()->backing_store()) +
|
| + NumberToSize(sta->byte_offset());
|
| +
|
| + switch (sta->type()) {
|
| +#define TYPED_ARRAY_CASE(Type, typeName, TYPE, ctype, size) \
|
| + case kExternal##Type##Array: \
|
| + return DoAdd<ctype>(isolate, source, index, value);
|
| +
|
| + INTEGER_TYPED_ARRAYS(TYPED_ARRAY_CASE)
|
| +#undef TYPED_ARRAY_CASE
|
| +
|
| + case kExternalUint8ClampedArray:
|
| + return DoAddUint8Clamped(isolate, source, index, value);
|
| +
|
| + default:
|
| + break;
|
| + }
|
| +
|
| + UNREACHABLE();
|
| + return isolate->heap()->undefined_value();
|
| +}
|
| +
|
| +
|
| +RUNTIME_FUNCTION(Runtime_AtomicsSub) {
|
| + HandleScope scope(isolate);
|
| + DCHECK_EQ(3, args.length());
|
| + CONVERT_ARG_HANDLE_CHECKED(JSTypedArray, sta, 0);
|
| + CONVERT_SIZE_ARG_CHECKED(index, 1);
|
| + CONVERT_NUMBER_ARG_HANDLE_CHECKED(value, 2);
|
| + CHECK(sta->GetBuffer()->is_shared());
|
| + CHECK_LT(index, NumberToSize(sta->length()));
|
| +
|
| + uint8_t* source = static_cast<uint8_t*>(sta->GetBuffer()->backing_store()) +
|
| + NumberToSize(sta->byte_offset());
|
| +
|
| + switch (sta->type()) {
|
| +#define TYPED_ARRAY_CASE(Type, typeName, TYPE, ctype, size) \
|
| + case kExternal##Type##Array: \
|
| + return DoSub<ctype>(isolate, source, index, value);
|
| +
|
| + INTEGER_TYPED_ARRAYS(TYPED_ARRAY_CASE)
|
| +#undef TYPED_ARRAY_CASE
|
| +
|
| + case kExternalUint8ClampedArray:
|
| + return DoSubUint8Clamped(isolate, source, index, value);
|
| +
|
| + default:
|
| + break;
|
| + }
|
| +
|
| + UNREACHABLE();
|
| + return isolate->heap()->undefined_value();
|
| +}
|
| +
|
| +
|
| +RUNTIME_FUNCTION(Runtime_AtomicsAnd) {
|
| + HandleScope scope(isolate);
|
| + DCHECK_EQ(3, args.length());
|
| + CONVERT_ARG_HANDLE_CHECKED(JSTypedArray, sta, 0);
|
| + CONVERT_SIZE_ARG_CHECKED(index, 1);
|
| + CONVERT_NUMBER_ARG_HANDLE_CHECKED(value, 2);
|
| + CHECK(sta->GetBuffer()->is_shared());
|
| + CHECK_LT(index, NumberToSize(sta->length()));
|
| +
|
| + uint8_t* source = static_cast<uint8_t*>(sta->GetBuffer()->backing_store()) +
|
| + NumberToSize(sta->byte_offset());
|
| +
|
| + switch (sta->type()) {
|
| +#define TYPED_ARRAY_CASE(Type, typeName, TYPE, ctype, size) \
|
| + case kExternal##Type##Array: \
|
| + return DoAnd<ctype>(isolate, source, index, value);
|
| +
|
| + INTEGER_TYPED_ARRAYS(TYPED_ARRAY_CASE)
|
| +#undef TYPED_ARRAY_CASE
|
| +
|
| + case kExternalUint8ClampedArray:
|
| + return DoAndUint8Clamped(isolate, source, index, value);
|
| +
|
| + default:
|
| + break;
|
| + }
|
| +
|
| + UNREACHABLE();
|
| + return isolate->heap()->undefined_value();
|
| +}
|
| +
|
| +
|
| +RUNTIME_FUNCTION(Runtime_AtomicsOr) {
|
| + HandleScope scope(isolate);
|
| + DCHECK_EQ(3, args.length());
|
| + CONVERT_ARG_HANDLE_CHECKED(JSTypedArray, sta, 0);
|
| + CONVERT_SIZE_ARG_CHECKED(index, 1);
|
| + CONVERT_NUMBER_ARG_HANDLE_CHECKED(value, 2);
|
| + CHECK(sta->GetBuffer()->is_shared());
|
| + CHECK_LT(index, NumberToSize(sta->length()));
|
| +
|
| + uint8_t* source = static_cast<uint8_t*>(sta->GetBuffer()->backing_store()) +
|
| + NumberToSize(sta->byte_offset());
|
| +
|
| + switch (sta->type()) {
|
| +#define TYPED_ARRAY_CASE(Type, typeName, TYPE, ctype, size) \
|
| + case kExternal##Type##Array: \
|
| + return DoOr<ctype>(isolate, source, index, value);
|
| +
|
| + INTEGER_TYPED_ARRAYS(TYPED_ARRAY_CASE)
|
| +#undef TYPED_ARRAY_CASE
|
| +
|
| + case kExternalUint8ClampedArray:
|
| + return DoOrUint8Clamped(isolate, source, index, value);
|
| +
|
| + default:
|
| + break;
|
| + }
|
| +
|
| + UNREACHABLE();
|
| + return isolate->heap()->undefined_value();
|
| +}
|
| +
|
| +
|
| +RUNTIME_FUNCTION(Runtime_AtomicsXor) {
|
| + HandleScope scope(isolate);
|
| + DCHECK_EQ(3, args.length());
|
| + CONVERT_ARG_HANDLE_CHECKED(JSTypedArray, sta, 0);
|
| + CONVERT_SIZE_ARG_CHECKED(index, 1);
|
| + CONVERT_NUMBER_ARG_HANDLE_CHECKED(value, 2);
|
| + CHECK(sta->GetBuffer()->is_shared());
|
| + CHECK_LT(index, NumberToSize(sta->length()));
|
| +
|
| + uint8_t* source = static_cast<uint8_t*>(sta->GetBuffer()->backing_store()) +
|
| + NumberToSize(sta->byte_offset());
|
| +
|
| + switch (sta->type()) {
|
| +#define TYPED_ARRAY_CASE(Type, typeName, TYPE, ctype, size) \
|
| + case kExternal##Type##Array: \
|
| + return DoXor<ctype>(isolate, source, index, value);
|
| +
|
| + INTEGER_TYPED_ARRAYS(TYPED_ARRAY_CASE)
|
| +#undef TYPED_ARRAY_CASE
|
| +
|
| + case kExternalUint8ClampedArray:
|
| + return DoXorUint8Clamped(isolate, source, index, value);
|
| +
|
| + default:
|
| + break;
|
| + }
|
| +
|
| + UNREACHABLE();
|
| + return isolate->heap()->undefined_value();
|
| +}
|
| +
|
| +
|
| +RUNTIME_FUNCTION(Runtime_AtomicsExchange) {
|
| + HandleScope scope(isolate);
|
| + DCHECK_EQ(3, args.length());
|
| + CONVERT_ARG_HANDLE_CHECKED(JSTypedArray, sta, 0);
|
| + CONVERT_SIZE_ARG_CHECKED(index, 1);
|
| + CONVERT_NUMBER_ARG_HANDLE_CHECKED(value, 2);
|
| + CHECK(sta->GetBuffer()->is_shared());
|
| + CHECK_LT(index, NumberToSize(sta->length()));
|
| +
|
| + uint8_t* source = static_cast<uint8_t*>(sta->GetBuffer()->backing_store()) +
|
| + NumberToSize(sta->byte_offset());
|
| +
|
| + switch (sta->type()) {
|
| +#define TYPED_ARRAY_CASE(Type, typeName, TYPE, ctype, size) \
|
| + case kExternal##Type##Array: \
|
| + return DoExchange<ctype>(isolate, source, index, value);
|
| +
|
| + INTEGER_TYPED_ARRAYS(TYPED_ARRAY_CASE)
|
| +#undef TYPED_ARRAY_CASE
|
| +
|
| + case kExternalUint8ClampedArray:
|
| + return DoExchangeUint8Clamped(isolate, source, index, value);
|
| +
|
| + default:
|
| + break;
|
| + }
|
| +
|
| + UNREACHABLE();
|
| + return isolate->heap()->undefined_value();
|
| +}
|
| +
|
| +
|
| +RUNTIME_FUNCTION(Runtime_AtomicsIsLockFree) {
|
| + HandleScope scope(isolate);
|
| + DCHECK_EQ(1, args.length());
|
| + CONVERT_NUMBER_ARG_HANDLE_CHECKED(size, 0);
|
| + uint32_t usize = NumberToUint32(*size);
|
| + return isolate->heap()->ToBoolean(AtomicIsLockFree(usize));
|
| +}
|
| } // namespace internal
|
| } // namespace v8
|
|
|