Index: tests_lit/reader_tests/nacl-atomic-intrinsics.ll |
diff --git a/tests_lit/reader_tests/nacl-atomic-intrinsics.ll b/tests_lit/reader_tests/nacl-atomic-intrinsics.ll |
index b59b380e7fc5b9956e5b460e8fe4f021e68534fe..d12c723891b0193a4e782a0c0a6218ad41bfb933 100644 |
--- a/tests_lit/reader_tests/nacl-atomic-intrinsics.ll |
+++ b/tests_lit/reader_tests/nacl-atomic-intrinsics.ll |
@@ -28,7 +28,7 @@ declare i1 @llvm.nacl.atomic.is.lock.free(i32, i8*) |
;;; Load |
-define i32 @test_atomic_load_8(i32 %iptr) { |
+define internal i32 @test_atomic_load_8(i32 %iptr) { |
entry: |
%ptr = inttoptr i32 %iptr to i8* |
; parameter value "6" is for the sequential consistency memory order. |
@@ -37,14 +37,14 @@ entry: |
ret i32 %r |
} |
-; CHECK: define i32 @test_atomic_load_8(i32 %iptr) { |
+; CHECK: define internal i32 @test_atomic_load_8(i32 %iptr) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %i = call i8 @llvm.nacl.atomic.load.i8(i32 %iptr, i32 6) |
; CHECK-NEXT: %r = zext i8 %i to i32 |
; CHECK-NEXT: ret i32 %r |
; CHECK-NEXT: } |
-define i32 @test_atomic_load_16(i32 %iptr) { |
+define internal i32 @test_atomic_load_16(i32 %iptr) { |
entry: |
%ptr = inttoptr i32 %iptr to i16* |
%i = call i16 @llvm.nacl.atomic.load.i16(i16* %ptr, i32 6) |
@@ -52,34 +52,34 @@ entry: |
ret i32 %r |
} |
-; CHECK-NEXT: define i32 @test_atomic_load_16(i32 %iptr) { |
+; CHECK-NEXT: define internal i32 @test_atomic_load_16(i32 %iptr) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %i = call i16 @llvm.nacl.atomic.load.i16(i32 %iptr, i32 6) |
; CHECK-NEXT: %r = zext i16 %i to i32 |
; CHECK-NEXT: ret i32 %r |
; CHECK-NEXT: } |
-define i32 @test_atomic_load_32(i32 %iptr) { |
+define internal i32 @test_atomic_load_32(i32 %iptr) { |
entry: |
%ptr = inttoptr i32 %iptr to i32* |
%r = call i32 @llvm.nacl.atomic.load.i32(i32* %ptr, i32 6) |
ret i32 %r |
} |
-; CHECK-NEXT: define i32 @test_atomic_load_32(i32 %iptr) { |
+; CHECK-NEXT: define internal i32 @test_atomic_load_32(i32 %iptr) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %r = call i32 @llvm.nacl.atomic.load.i32(i32 %iptr, i32 6) |
; CHECK-NEXT: ret i32 %r |
; CHECK-NEXT: } |
-define i64 @test_atomic_load_64(i32 %iptr) { |
+define internal i64 @test_atomic_load_64(i32 %iptr) { |
entry: |
%ptr = inttoptr i32 %iptr to i64* |
%r = call i64 @llvm.nacl.atomic.load.i64(i64* %ptr, i32 6) |
ret i64 %r |
} |
-; CHECK-NEXT: define i64 @test_atomic_load_64(i32 %iptr) { |
+; CHECK-NEXT: define internal i64 @test_atomic_load_64(i32 %iptr) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %r = call i64 @llvm.nacl.atomic.load.i64(i32 %iptr, i32 6) |
; CHECK-NEXT: ret i64 %r |
@@ -87,7 +87,7 @@ entry: |
;;; Store |
-define void @test_atomic_store_8(i32 %iptr, i32 %v) { |
+define internal void @test_atomic_store_8(i32 %iptr, i32 %v) { |
entry: |
%truncv = trunc i32 %v to i8 |
%ptr = inttoptr i32 %iptr to i8* |
@@ -95,14 +95,14 @@ entry: |
ret void |
} |
-; CHECK-NEXT: define void @test_atomic_store_8(i32 %iptr, i32 %v) { |
+; CHECK-NEXT: define internal void @test_atomic_store_8(i32 %iptr, i32 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %truncv = trunc i32 %v to i8 |
; CHECK-NEXT: call void @llvm.nacl.atomic.store.i8(i8 %truncv, i32 %iptr, i32 6) |
; CHECK-NEXT: ret void |
; CHECK-NEXT: } |
-define void @test_atomic_store_16(i32 %iptr, i32 %v) { |
+define internal void @test_atomic_store_16(i32 %iptr, i32 %v) { |
entry: |
%truncv = trunc i32 %v to i16 |
%ptr = inttoptr i32 %iptr to i16* |
@@ -110,47 +110,47 @@ entry: |
ret void |
} |
-; CHECK-NEXT: define void @test_atomic_store_16(i32 %iptr, i32 %v) { |
+; CHECK-NEXT: define internal void @test_atomic_store_16(i32 %iptr, i32 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %truncv = trunc i32 %v to i16 |
; CHECK-NEXT: call void @llvm.nacl.atomic.store.i16(i16 %truncv, i32 %iptr, i32 6) |
; CHECK-NEXT: ret void |
; CHECK-NEXT: } |
-define void @test_atomic_store_32(i32 %iptr, i32 %v) { |
+define internal void @test_atomic_store_32(i32 %iptr, i32 %v) { |
entry: |
%ptr = inttoptr i32 %iptr to i32* |
call void @llvm.nacl.atomic.store.i32(i32 %v, i32* %ptr, i32 6) |
ret void |
} |
-; CHECK-NEXT: define void @test_atomic_store_32(i32 %iptr, i32 %v) { |
+; CHECK-NEXT: define internal void @test_atomic_store_32(i32 %iptr, i32 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: call void @llvm.nacl.atomic.store.i32(i32 %v, i32 %iptr, i32 6) |
; CHECK-NEXT: ret void |
; CHECK-NEXT: } |
-define void @test_atomic_store_64(i32 %iptr, i64 %v) { |
+define internal void @test_atomic_store_64(i32 %iptr, i64 %v) { |
entry: |
%ptr = inttoptr i32 %iptr to i64* |
call void @llvm.nacl.atomic.store.i64(i64 %v, i64* %ptr, i32 6) |
ret void |
} |
-; CHECK-NEXT: define void @test_atomic_store_64(i32 %iptr, i64 %v) { |
+; CHECK-NEXT: define internal void @test_atomic_store_64(i32 %iptr, i64 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: call void @llvm.nacl.atomic.store.i64(i64 %v, i32 %iptr, i32 6) |
; CHECK-NEXT: ret void |
; CHECK-NEXT: } |
-define void @test_atomic_store_64_const(i32 %iptr) { |
+define internal void @test_atomic_store_64_const(i32 %iptr) { |
entry: |
%ptr = inttoptr i32 %iptr to i64* |
call void @llvm.nacl.atomic.store.i64(i64 12345678901234, i64* %ptr, i32 6) |
ret void |
} |
-; CHECK-NEXT: define void @test_atomic_store_64_const(i32 %iptr) { |
+; CHECK-NEXT: define internal void @test_atomic_store_64_const(i32 %iptr) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: call void @llvm.nacl.atomic.store.i64(i64 12345678901234, i32 %iptr, i32 6) |
; CHECK-NEXT: ret void |
@@ -160,7 +160,7 @@ entry: |
;; add |
-define i32 @test_atomic_rmw_add_8(i32 %iptr, i32 %v) { |
+define internal i32 @test_atomic_rmw_add_8(i32 %iptr, i32 %v) { |
entry: |
%trunc = trunc i32 %v to i8 |
%ptr = inttoptr i32 %iptr to i8* |
@@ -170,7 +170,7 @@ entry: |
ret i32 %a_ext |
} |
-; CHECK-NEXT: define i32 @test_atomic_rmw_add_8(i32 %iptr, i32 %v) { |
+; CHECK-NEXT: define internal i32 @test_atomic_rmw_add_8(i32 %iptr, i32 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %trunc = trunc i32 %v to i8 |
; CHECK-NEXT: %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 1, i32 %iptr, i8 %trunc, i32 6) |
@@ -178,7 +178,7 @@ entry: |
; CHECK-NEXT: ret i32 %a_ext |
; CHECK-NEXT: } |
-define i32 @test_atomic_rmw_add_16(i32 %iptr, i32 %v) { |
+define internal i32 @test_atomic_rmw_add_16(i32 %iptr, i32 %v) { |
entry: |
%trunc = trunc i32 %v to i16 |
%ptr = inttoptr i32 %iptr to i16* |
@@ -187,7 +187,7 @@ entry: |
ret i32 %a_ext |
} |
-; CHECK-NEXT: define i32 @test_atomic_rmw_add_16(i32 %iptr, i32 %v) { |
+; CHECK-NEXT: define internal i32 @test_atomic_rmw_add_16(i32 %iptr, i32 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %trunc = trunc i32 %v to i16 |
; CHECK-NEXT: %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 1, i32 %iptr, i16 %trunc, i32 6) |
@@ -195,27 +195,27 @@ entry: |
; CHECK-NEXT: ret i32 %a_ext |
; CHECK-NEXT: } |
-define i32 @test_atomic_rmw_add_32(i32 %iptr, i32 %v) { |
+define internal i32 @test_atomic_rmw_add_32(i32 %iptr, i32 %v) { |
entry: |
%ptr = inttoptr i32 %iptr to i32* |
%a = call i32 @llvm.nacl.atomic.rmw.i32(i32 1, i32* %ptr, i32 %v, i32 6) |
ret i32 %a |
} |
-; CHECK-NEXT: define i32 @test_atomic_rmw_add_32(i32 %iptr, i32 %v) { |
+; CHECK-NEXT: define internal i32 @test_atomic_rmw_add_32(i32 %iptr, i32 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 1, i32 %iptr, i32 %v, i32 6) |
; CHECK-NEXT: ret i32 %a |
; CHECK-NEXT: } |
-define i64 @test_atomic_rmw_add_64(i32 %iptr, i64 %v) { |
+define internal i64 @test_atomic_rmw_add_64(i32 %iptr, i64 %v) { |
entry: |
%ptr = inttoptr i32 %iptr to i64* |
%a = call i64 @llvm.nacl.atomic.rmw.i64(i32 1, i64* %ptr, i64 %v, i32 6) |
ret i64 %a |
} |
-; CHECK-NEXT: define i64 @test_atomic_rmw_add_64(i32 %iptr, i64 %v) { |
+; CHECK-NEXT: define internal i64 @test_atomic_rmw_add_64(i32 %iptr, i64 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 1, i32 %iptr, i64 %v, i32 6) |
; CHECK-NEXT: ret i64 %a |
@@ -223,7 +223,7 @@ entry: |
;; sub |
-define i32 @test_atomic_rmw_sub_8(i32 %iptr, i32 %v) { |
+define internal i32 @test_atomic_rmw_sub_8(i32 %iptr, i32 %v) { |
entry: |
%trunc = trunc i32 %v to i8 |
%ptr = inttoptr i32 %iptr to i8* |
@@ -232,7 +232,7 @@ entry: |
ret i32 %a_ext |
} |
-; CHECK-NEXT: define i32 @test_atomic_rmw_sub_8(i32 %iptr, i32 %v) { |
+; CHECK-NEXT: define internal i32 @test_atomic_rmw_sub_8(i32 %iptr, i32 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %trunc = trunc i32 %v to i8 |
; CHECK-NEXT: %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 2, i32 %iptr, i8 %trunc, i32 6) |
@@ -240,7 +240,7 @@ entry: |
; CHECK-NEXT: ret i32 %a_ext |
; CHECK-NEXT: } |
-define i32 @test_atomic_rmw_sub_16(i32 %iptr, i32 %v) { |
+define internal i32 @test_atomic_rmw_sub_16(i32 %iptr, i32 %v) { |
entry: |
%trunc = trunc i32 %v to i16 |
%ptr = inttoptr i32 %iptr to i16* |
@@ -249,7 +249,7 @@ entry: |
ret i32 %a_ext |
} |
-; CHECK-NEXT: define i32 @test_atomic_rmw_sub_16(i32 %iptr, i32 %v) { |
+; CHECK-NEXT: define internal i32 @test_atomic_rmw_sub_16(i32 %iptr, i32 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %trunc = trunc i32 %v to i16 |
; CHECK-NEXT: %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 2, i32 %iptr, i16 %trunc, i32 6) |
@@ -257,27 +257,27 @@ entry: |
; CHECK-NEXT: ret i32 %a_ext |
; CHECK-NEXT: } |
-define i32 @test_atomic_rmw_sub_32(i32 %iptr, i32 %v) { |
+define internal i32 @test_atomic_rmw_sub_32(i32 %iptr, i32 %v) { |
entry: |
%ptr = inttoptr i32 %iptr to i32* |
%a = call i32 @llvm.nacl.atomic.rmw.i32(i32 2, i32* %ptr, i32 %v, i32 6) |
ret i32 %a |
} |
-; CHECK-NEXT: define i32 @test_atomic_rmw_sub_32(i32 %iptr, i32 %v) { |
+; CHECK-NEXT: define internal i32 @test_atomic_rmw_sub_32(i32 %iptr, i32 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 2, i32 %iptr, i32 %v, i32 6) |
; CHECK-NEXT: ret i32 %a |
; CHECK-NEXT: } |
-define i64 @test_atomic_rmw_sub_64(i32 %iptr, i64 %v) { |
+define internal i64 @test_atomic_rmw_sub_64(i32 %iptr, i64 %v) { |
entry: |
%ptr = inttoptr i32 %iptr to i64* |
%a = call i64 @llvm.nacl.atomic.rmw.i64(i32 2, i64* %ptr, i64 %v, i32 6) |
ret i64 %a |
} |
-; CHECK-NEXT: define i64 @test_atomic_rmw_sub_64(i32 %iptr, i64 %v) { |
+; CHECK-NEXT: define internal i64 @test_atomic_rmw_sub_64(i32 %iptr, i64 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 2, i32 %iptr, i64 %v, i32 6) |
; CHECK-NEXT: ret i64 %a |
@@ -285,7 +285,7 @@ entry: |
;; or |
-define i32 @test_atomic_rmw_or_8(i32 %iptr, i32 %v) { |
+define internal i32 @test_atomic_rmw_or_8(i32 %iptr, i32 %v) { |
entry: |
%trunc = trunc i32 %v to i8 |
%ptr = inttoptr i32 %iptr to i8* |
@@ -294,7 +294,7 @@ entry: |
ret i32 %a_ext |
} |
-; CHECK-NEXT: define i32 @test_atomic_rmw_or_8(i32 %iptr, i32 %v) { |
+; CHECK-NEXT: define internal i32 @test_atomic_rmw_or_8(i32 %iptr, i32 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %trunc = trunc i32 %v to i8 |
; CHECK-NEXT: %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 3, i32 %iptr, i8 %trunc, i32 6) |
@@ -302,7 +302,7 @@ entry: |
; CHECK-NEXT: ret i32 %a_ext |
; CHECK-NEXT: } |
-define i32 @test_atomic_rmw_or_16(i32 %iptr, i32 %v) { |
+define internal i32 @test_atomic_rmw_or_16(i32 %iptr, i32 %v) { |
entry: |
%trunc = trunc i32 %v to i16 |
%ptr = inttoptr i32 %iptr to i16* |
@@ -311,7 +311,7 @@ entry: |
ret i32 %a_ext |
} |
-; CHECK-NEXT: define i32 @test_atomic_rmw_or_16(i32 %iptr, i32 %v) { |
+; CHECK-NEXT: define internal i32 @test_atomic_rmw_or_16(i32 %iptr, i32 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %trunc = trunc i32 %v to i16 |
; CHECK-NEXT: %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 3, i32 %iptr, i16 %trunc, i32 6) |
@@ -319,27 +319,27 @@ entry: |
; CHECK-NEXT: ret i32 %a_ext |
; CHECK-NEXT: } |
-define i32 @test_atomic_rmw_or_32(i32 %iptr, i32 %v) { |
+define internal i32 @test_atomic_rmw_or_32(i32 %iptr, i32 %v) { |
entry: |
%ptr = inttoptr i32 %iptr to i32* |
%a = call i32 @llvm.nacl.atomic.rmw.i32(i32 3, i32* %ptr, i32 %v, i32 6) |
ret i32 %a |
} |
-; CHECK-NEXT: define i32 @test_atomic_rmw_or_32(i32 %iptr, i32 %v) { |
+; CHECK-NEXT: define internal i32 @test_atomic_rmw_or_32(i32 %iptr, i32 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 3, i32 %iptr, i32 %v, i32 6) |
; CHECK-NEXT: ret i32 %a |
; CHECK-NEXT: } |
-define i64 @test_atomic_rmw_or_64(i32 %iptr, i64 %v) { |
+define internal i64 @test_atomic_rmw_or_64(i32 %iptr, i64 %v) { |
entry: |
%ptr = inttoptr i32 %iptr to i64* |
%a = call i64 @llvm.nacl.atomic.rmw.i64(i32 3, i64* %ptr, i64 %v, i32 6) |
ret i64 %a |
} |
-; CHECK-NEXT: define i64 @test_atomic_rmw_or_64(i32 %iptr, i64 %v) { |
+; CHECK-NEXT: define internal i64 @test_atomic_rmw_or_64(i32 %iptr, i64 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 3, i32 %iptr, i64 %v, i32 6) |
; CHECK-NEXT: ret i64 %a |
@@ -347,7 +347,7 @@ entry: |
;; and |
-define i32 @test_atomic_rmw_and_8(i32 %iptr, i32 %v) { |
+define internal i32 @test_atomic_rmw_and_8(i32 %iptr, i32 %v) { |
entry: |
%trunc = trunc i32 %v to i8 |
%ptr = inttoptr i32 %iptr to i8* |
@@ -356,7 +356,7 @@ entry: |
ret i32 %a_ext |
} |
-; CHECK-NEXT: define i32 @test_atomic_rmw_and_8(i32 %iptr, i32 %v) { |
+; CHECK-NEXT: define internal i32 @test_atomic_rmw_and_8(i32 %iptr, i32 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %trunc = trunc i32 %v to i8 |
; CHECK-NEXT: %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 4, i32 %iptr, i8 %trunc, i32 6) |
@@ -364,7 +364,7 @@ entry: |
; CHECK-NEXT: ret i32 %a_ext |
; CHECK-NEXT: } |
-define i32 @test_atomic_rmw_and_16(i32 %iptr, i32 %v) { |
+define internal i32 @test_atomic_rmw_and_16(i32 %iptr, i32 %v) { |
entry: |
%trunc = trunc i32 %v to i16 |
%ptr = inttoptr i32 %iptr to i16* |
@@ -373,7 +373,7 @@ entry: |
ret i32 %a_ext |
} |
-; CHECK-NEXT: define i32 @test_atomic_rmw_and_16(i32 %iptr, i32 %v) { |
+; CHECK-NEXT: define internal i32 @test_atomic_rmw_and_16(i32 %iptr, i32 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %trunc = trunc i32 %v to i16 |
; CHECK-NEXT: %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 4, i32 %iptr, i16 %trunc, i32 6) |
@@ -381,27 +381,27 @@ entry: |
; CHECK-NEXT: ret i32 %a_ext |
; CHECK-NEXT: } |
-define i32 @test_atomic_rmw_and_32(i32 %iptr, i32 %v) { |
+define internal i32 @test_atomic_rmw_and_32(i32 %iptr, i32 %v) { |
entry: |
%ptr = inttoptr i32 %iptr to i32* |
%a = call i32 @llvm.nacl.atomic.rmw.i32(i32 4, i32* %ptr, i32 %v, i32 6) |
ret i32 %a |
} |
-; CHECK-NEXT: define i32 @test_atomic_rmw_and_32(i32 %iptr, i32 %v) { |
+; CHECK-NEXT: define internal i32 @test_atomic_rmw_and_32(i32 %iptr, i32 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 4, i32 %iptr, i32 %v, i32 6) |
; CHECK-NEXT: ret i32 %a |
; CHECK-NEXT: } |
-define i64 @test_atomic_rmw_and_64(i32 %iptr, i64 %v) { |
+define internal i64 @test_atomic_rmw_and_64(i32 %iptr, i64 %v) { |
entry: |
%ptr = inttoptr i32 %iptr to i64* |
%a = call i64 @llvm.nacl.atomic.rmw.i64(i32 4, i64* %ptr, i64 %v, i32 6) |
ret i64 %a |
} |
-; CHECK-NEXT: define i64 @test_atomic_rmw_and_64(i32 %iptr, i64 %v) { |
+; CHECK-NEXT: define internal i64 @test_atomic_rmw_and_64(i32 %iptr, i64 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 4, i32 %iptr, i64 %v, i32 6) |
; CHECK-NEXT: ret i64 %a |
@@ -409,7 +409,7 @@ entry: |
;; xor |
-define i32 @test_atomic_rmw_xor_8(i32 %iptr, i32 %v) { |
+define internal i32 @test_atomic_rmw_xor_8(i32 %iptr, i32 %v) { |
entry: |
%trunc = trunc i32 %v to i8 |
%ptr = inttoptr i32 %iptr to i8* |
@@ -418,7 +418,7 @@ entry: |
ret i32 %a_ext |
} |
-; CHECK-NEXT: define i32 @test_atomic_rmw_xor_8(i32 %iptr, i32 %v) { |
+; CHECK-NEXT: define internal i32 @test_atomic_rmw_xor_8(i32 %iptr, i32 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %trunc = trunc i32 %v to i8 |
; CHECK-NEXT: %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 5, i32 %iptr, i8 %trunc, i32 6) |
@@ -426,7 +426,7 @@ entry: |
; CHECK-NEXT: ret i32 %a_ext |
; CHECK-NEXT: } |
-define i32 @test_atomic_rmw_xor_16(i32 %iptr, i32 %v) { |
+define internal i32 @test_atomic_rmw_xor_16(i32 %iptr, i32 %v) { |
entry: |
%trunc = trunc i32 %v to i16 |
%ptr = inttoptr i32 %iptr to i16* |
@@ -435,7 +435,7 @@ entry: |
ret i32 %a_ext |
} |
-; CHECK-NEXT: define i32 @test_atomic_rmw_xor_16(i32 %iptr, i32 %v) { |
+; CHECK-NEXT: define internal i32 @test_atomic_rmw_xor_16(i32 %iptr, i32 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %trunc = trunc i32 %v to i16 |
; CHECK-NEXT: %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 5, i32 %iptr, i16 %trunc, i32 6) |
@@ -443,27 +443,27 @@ entry: |
; CHECK-NEXT: ret i32 %a_ext |
; CHECK-NEXT: } |
-define i32 @test_atomic_rmw_xor_32(i32 %iptr, i32 %v) { |
+define internal i32 @test_atomic_rmw_xor_32(i32 %iptr, i32 %v) { |
entry: |
%ptr = inttoptr i32 %iptr to i32* |
%a = call i32 @llvm.nacl.atomic.rmw.i32(i32 5, i32* %ptr, i32 %v, i32 6) |
ret i32 %a |
} |
-; CHECK-NEXT: define i32 @test_atomic_rmw_xor_32(i32 %iptr, i32 %v) { |
+; CHECK-NEXT: define internal i32 @test_atomic_rmw_xor_32(i32 %iptr, i32 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 5, i32 %iptr, i32 %v, i32 6) |
; CHECK-NEXT: ret i32 %a |
; CHECK-NEXT: } |
-define i64 @test_atomic_rmw_xor_64(i32 %iptr, i64 %v) { |
+define internal i64 @test_atomic_rmw_xor_64(i32 %iptr, i64 %v) { |
entry: |
%ptr = inttoptr i32 %iptr to i64* |
%a = call i64 @llvm.nacl.atomic.rmw.i64(i32 5, i64* %ptr, i64 %v, i32 6) |
ret i64 %a |
} |
-; CHECK-NEXT: define i64 @test_atomic_rmw_xor_64(i32 %iptr, i64 %v) { |
+; CHECK-NEXT: define internal i64 @test_atomic_rmw_xor_64(i32 %iptr, i64 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 5, i32 %iptr, i64 %v, i32 6) |
; CHECK-NEXT: ret i64 %a |
@@ -471,7 +471,7 @@ entry: |
;; exchange |
-define i32 @test_atomic_rmw_xchg_8(i32 %iptr, i32 %v) { |
+define internal i32 @test_atomic_rmw_xchg_8(i32 %iptr, i32 %v) { |
entry: |
%trunc = trunc i32 %v to i8 |
%ptr = inttoptr i32 %iptr to i8* |
@@ -480,7 +480,7 @@ entry: |
ret i32 %a_ext |
} |
-; CHECK-NEXT: define i32 @test_atomic_rmw_xchg_8(i32 %iptr, i32 %v) { |
+; CHECK-NEXT: define internal i32 @test_atomic_rmw_xchg_8(i32 %iptr, i32 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %trunc = trunc i32 %v to i8 |
; CHECK-NEXT: %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 6, i32 %iptr, i8 %trunc, i32 6) |
@@ -488,7 +488,7 @@ entry: |
; CHECK-NEXT: ret i32 %a_ext |
; CHECK-NEXT: } |
-define i32 @test_atomic_rmw_xchg_16(i32 %iptr, i32 %v) { |
+define internal i32 @test_atomic_rmw_xchg_16(i32 %iptr, i32 %v) { |
entry: |
%trunc = trunc i32 %v to i16 |
%ptr = inttoptr i32 %iptr to i16* |
@@ -497,7 +497,7 @@ entry: |
ret i32 %a_ext |
} |
-; CHECK-NEXT: define i32 @test_atomic_rmw_xchg_16(i32 %iptr, i32 %v) { |
+; CHECK-NEXT: define internal i32 @test_atomic_rmw_xchg_16(i32 %iptr, i32 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %trunc = trunc i32 %v to i16 |
; CHECK-NEXT: %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 6, i32 %iptr, i16 %trunc, i32 6) |
@@ -505,27 +505,27 @@ entry: |
; CHECK-NEXT: ret i32 %a_ext |
; CHECK-NEXT: } |
-define i32 @test_atomic_rmw_xchg_32(i32 %iptr, i32 %v) { |
+define internal i32 @test_atomic_rmw_xchg_32(i32 %iptr, i32 %v) { |
entry: |
%ptr = inttoptr i32 %iptr to i32* |
%a = call i32 @llvm.nacl.atomic.rmw.i32(i32 6, i32* %ptr, i32 %v, i32 6) |
ret i32 %a |
} |
-; CHECK-NEXT: define i32 @test_atomic_rmw_xchg_32(i32 %iptr, i32 %v) { |
+; CHECK-NEXT: define internal i32 @test_atomic_rmw_xchg_32(i32 %iptr, i32 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 6, i32 %iptr, i32 %v, i32 6) |
; CHECK-NEXT: ret i32 %a |
; CHECK-NEXT: } |
-define i64 @test_atomic_rmw_xchg_64(i32 %iptr, i64 %v) { |
+define internal i64 @test_atomic_rmw_xchg_64(i32 %iptr, i64 %v) { |
entry: |
%ptr = inttoptr i32 %iptr to i64* |
%a = call i64 @llvm.nacl.atomic.rmw.i64(i32 6, i64* %ptr, i64 %v, i32 6) |
ret i64 %a |
} |
-; CHECK-NEXT: define i64 @test_atomic_rmw_xchg_64(i32 %iptr, i64 %v) { |
+; CHECK-NEXT: define internal i64 @test_atomic_rmw_xchg_64(i32 %iptr, i64 %v) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 6, i32 %iptr, i64 %v, i32 6) |
; CHECK-NEXT: ret i64 %a |
@@ -533,7 +533,7 @@ entry: |
;;;; Cmpxchg |
-define i32 @test_atomic_cmpxchg_8(i32 %iptr, i32 %expected, i32 %desired) { |
+define internal i32 @test_atomic_cmpxchg_8(i32 %iptr, i32 %expected, i32 %desired) { |
entry: |
%trunc_exp = trunc i32 %expected to i8 |
%trunc_des = trunc i32 %desired to i8 |
@@ -544,7 +544,7 @@ entry: |
ret i32 %old_ext |
} |
-; CHECK-NEXT: define i32 @test_atomic_cmpxchg_8(i32 %iptr, i32 %expected, i32 %desired) { |
+; CHECK-NEXT: define internal i32 @test_atomic_cmpxchg_8(i32 %iptr, i32 %expected, i32 %desired) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %trunc_exp = trunc i32 %expected to i8 |
; CHECK-NEXT: %trunc_des = trunc i32 %desired to i8 |
@@ -553,7 +553,7 @@ entry: |
; CHECK-NEXT: ret i32 %old_ext |
; CHECK-NEXT: } |
-define i32 @test_atomic_cmpxchg_16(i32 %iptr, i32 %expected, i32 %desired) { |
+define internal i32 @test_atomic_cmpxchg_16(i32 %iptr, i32 %expected, i32 %desired) { |
entry: |
%trunc_exp = trunc i32 %expected to i16 |
%trunc_des = trunc i32 %desired to i16 |
@@ -564,7 +564,7 @@ entry: |
ret i32 %old_ext |
} |
-; CHECK-NEXT: define i32 @test_atomic_cmpxchg_16(i32 %iptr, i32 %expected, i32 %desired) { |
+; CHECK-NEXT: define internal i32 @test_atomic_cmpxchg_16(i32 %iptr, i32 %expected, i32 %desired) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %trunc_exp = trunc i32 %expected to i16 |
; CHECK-NEXT: %trunc_des = trunc i32 %desired to i16 |
@@ -573,7 +573,7 @@ entry: |
; CHECK-NEXT: ret i32 %old_ext |
; CHECK-NEXT: } |
-define i32 @test_atomic_cmpxchg_32(i32 %iptr, i32 %expected, i32 %desired) { |
+define internal i32 @test_atomic_cmpxchg_32(i32 %iptr, i32 %expected, i32 %desired) { |
entry: |
%ptr = inttoptr i32 %iptr to i32* |
%old = call i32 @llvm.nacl.atomic.cmpxchg.i32(i32* %ptr, i32 %expected, |
@@ -581,13 +581,13 @@ entry: |
ret i32 %old |
} |
-; CHECK-NEXT: define i32 @test_atomic_cmpxchg_32(i32 %iptr, i32 %expected, i32 %desired) { |
+; CHECK-NEXT: define internal i32 @test_atomic_cmpxchg_32(i32 %iptr, i32 %expected, i32 %desired) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %old = call i32 @llvm.nacl.atomic.cmpxchg.i32(i32 %iptr, i32 %expected, i32 %desired, i32 6, i32 6) |
; CHECK-NEXT: ret i32 %old |
; CHECK-NEXT: } |
-define i64 @test_atomic_cmpxchg_64(i32 %iptr, i64 %expected, i64 %desired) { |
+define internal i64 @test_atomic_cmpxchg_64(i32 %iptr, i64 %expected, i64 %desired) { |
entry: |
%ptr = inttoptr i32 %iptr to i64* |
%old = call i64 @llvm.nacl.atomic.cmpxchg.i64(i64* %ptr, i64 %expected, |
@@ -595,7 +595,7 @@ entry: |
ret i64 %old |
} |
-; CHECK-NEXT: define i64 @test_atomic_cmpxchg_64(i32 %iptr, i64 %expected, i64 %desired) { |
+; CHECK-NEXT: define internal i64 @test_atomic_cmpxchg_64(i32 %iptr, i64 %expected, i64 %desired) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %old = call i64 @llvm.nacl.atomic.cmpxchg.i64(i32 %iptr, i64 %expected, i64 %desired, i32 6, i32 6) |
; CHECK-NEXT: ret i64 %old |
@@ -603,31 +603,31 @@ entry: |
;;;; Fence and is-lock-free. |
-define void @test_atomic_fence() { |
+define internal void @test_atomic_fence() { |
entry: |
call void @llvm.nacl.atomic.fence(i32 6) |
ret void |
} |
-; CHECK-NEXT: define void @test_atomic_fence() { |
+; CHECK-NEXT: define internal void @test_atomic_fence() { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: call void @llvm.nacl.atomic.fence(i32 6) |
; CHECK-NEXT: ret void |
; CHECK-NEXT: } |
-define void @test_atomic_fence_all() { |
+define internal void @test_atomic_fence_all() { |
entry: |
call void @llvm.nacl.atomic.fence.all() |
ret void |
} |
-; CHECK-NEXT: define void @test_atomic_fence_all() { |
+; CHECK-NEXT: define internal void @test_atomic_fence_all() { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: call void @llvm.nacl.atomic.fence.all() |
; CHECK-NEXT: ret void |
; CHECK-NEXT: } |
-define i32 @test_atomic_is_lock_free(i32 %iptr) { |
+define internal i32 @test_atomic_is_lock_free(i32 %iptr) { |
entry: |
%ptr = inttoptr i32 %iptr to i8* |
%i = call i1 @llvm.nacl.atomic.is.lock.free(i32 4, i8* %ptr) |
@@ -635,7 +635,7 @@ entry: |
ret i32 %r |
} |
-; CHECK-NEXT: define i32 @test_atomic_is_lock_free(i32 %iptr) { |
+; CHECK-NEXT: define internal i32 @test_atomic_is_lock_free(i32 %iptr) { |
; CHECK-NEXT: entry: |
; CHECK-NEXT: %i = call i1 @llvm.nacl.atomic.is.lock.free(i32 4, i32 %iptr) |
; CHECK-NEXT: %r = zext i1 %i to i32 |