Index: tests_lit/llvm2ice_tests/nacl-atomic-intrinsics.ll |
diff --git a/tests_lit/llvm2ice_tests/nacl-atomic-intrinsics.ll b/tests_lit/llvm2ice_tests/nacl-atomic-intrinsics.ll |
index 45ee8b3e134f94503313b0b75531f02c8c4e112c..12d42675ea54ca1df00dfca739943e0a8740f7c0 100644 |
--- a/tests_lit/llvm2ice_tests/nacl-atomic-intrinsics.ll |
+++ b/tests_lit/llvm2ice_tests/nacl-atomic-intrinsics.ll |
@@ -46,10 +46,10 @@ declare void @llvm.nacl.atomic.fence(i32) |
declare void @llvm.nacl.atomic.fence.all() |
declare i1 @llvm.nacl.atomic.is.lock.free(i32, i8*) |
-@Global8 = internal global [1 x i8] zeroinitializer, align 1 |
-@Global16 = internal global [2 x i8] zeroinitializer, align 2 |
-@Global32 = internal global [4 x i8] zeroinitializer, align 4 |
-@Global64 = internal global [8 x i8] zeroinitializer, align 8 |
+@SzGlobal8 = internal global [1 x i8] zeroinitializer, align 1 |
+@SzGlobal16 = internal global [2 x i8] zeroinitializer, align 2 |
+@SzGlobal32 = internal global [4 x i8] zeroinitializer, align 4 |
+@SzGlobal64 = internal global [8 x i8] zeroinitializer, align 8 |
; NOTE: The LLC equivalent for 16-bit atomic operations are expanded |
; as 32-bit operations. For Subzero, assume that real 16-bit operations |
@@ -350,7 +350,7 @@ entry: |
; Same test as above, but with a global address to test FakeUse issues. |
define internal i64 @test_atomic_rmw_add_64_global(i64 %v) { |
entry: |
- %ptr = bitcast [8 x i8]* @Global64 to i64* |
+ %ptr = bitcast [8 x i8]* @SzGlobal64 to i64* |
%a = call i64 @llvm.nacl.atomic.rmw.i64(i32 1, i64* %ptr, i64 %v, i32 6) |
ret i64 %a |
} |
@@ -595,15 +595,15 @@ entry: |
define internal i32 @test_atomic_rmw_or_8_global(i32 %v) { |
entry: |
%trunc = trunc i32 %v to i8 |
- %ptr = bitcast [1 x i8]* @Global8 to i8* |
+ %ptr = bitcast [1 x i8]* @SzGlobal8 to i8* |
%a = call i8 @llvm.nacl.atomic.rmw.i8(i32 3, i8* %ptr, i8 %trunc, i32 6) |
%a_ext = zext i8 %a to i32 |
ret i32 %a_ext |
} |
; CHECK-LABEL: test_atomic_rmw_or_8_global |
; ARM32-LABEL: test_atomic_rmw_or_8_global |
-; ARM32: movw [[PTR:r[0-9]+]], #:lower16:Global8 |
-; ARM32: movt [[PTR]], #:upper16:Global8 |
+; ARM32: movw [[PTR:r[0-9]+]], #:lower16:SzGlobal8 |
+; ARM32: movt [[PTR]], #:upper16:SzGlobal8 |
; ARM32: dmb |
; ARM32: ldrexb r{{[0-9]+}}, {{[[]}}[[PTR]]{{[]]}} |
; ARM32: orr |
@@ -636,15 +636,15 @@ entry: |
define internal i32 @test_atomic_rmw_or_16_global(i32 %v) { |
entry: |
%trunc = trunc i32 %v to i16 |
- %ptr = bitcast [2 x i8]* @Global16 to i16* |
+ %ptr = bitcast [2 x i8]* @SzGlobal16 to i16* |
%a = call i16 @llvm.nacl.atomic.rmw.i16(i32 3, i16* %ptr, i16 %trunc, i32 6) |
%a_ext = zext i16 %a to i32 |
ret i32 %a_ext |
} |
; CHECK-LABEL: test_atomic_rmw_or_16_global |
; ARM32-LABEL: test_atomic_rmw_or_16_global |
-; ARM32: movw [[PTR:r[0-9]+]], #:lower16:Global16 |
-; ARM32: movt [[PTR]], #:upper16:Global16 |
+; ARM32: movw [[PTR:r[0-9]+]], #:lower16:SzGlobal16 |
+; ARM32: movt [[PTR]], #:upper16:SzGlobal16 |
; ARM32: dmb |
; ARM32: ldrexh r{{[0-9]+}}, {{[[]}}[[PTR]]{{[]]}} |
; ARM32: orr |
@@ -674,14 +674,14 @@ entry: |
; Same test as above, but with a global address to test FakeUse issues. |
define internal i32 @test_atomic_rmw_or_32_global(i32 %v) { |
entry: |
- %ptr = bitcast [4 x i8]* @Global32 to i32* |
+ %ptr = bitcast [4 x i8]* @SzGlobal32 to i32* |
%a = call i32 @llvm.nacl.atomic.rmw.i32(i32 3, i32* %ptr, i32 %v, i32 6) |
ret i32 %a |
} |
; CHECK-LABEL: test_atomic_rmw_or_32_global |
; ARM32-LABEL: test_atomic_rmw_or_32_global |
-; ARM32: movw [[PTR:r[0-9]+]], #:lower16:Global32 |
-; ARM32: movt [[PTR]], #:upper16:Global32 |
+; ARM32: movw [[PTR:r[0-9]+]], #:lower16:SzGlobal32 |
+; ARM32: movt [[PTR]], #:upper16:SzGlobal32 |
; ARM32: dmb |
; ARM32: ldrex r{{[0-9]+}}, {{[[]}}[[PTR]]{{[]]}} |
; ARM32: orr |
@@ -1393,7 +1393,7 @@ entry: |
br label %body |
body: |
%i = phi i32 [ 1, %entry ], [ %i_plus_1, %body ] |
- %g = bitcast [4 x i8]* @Global32 to i32* |
+ %g = bitcast [4 x i8]* @SzGlobal32 to i32* |
%unused = call i32 @llvm.nacl.atomic.rmw.i32(i32 1, i32* %g, i32 %i, i32 6) |
%i_plus_1 = add i32 %i, 1 |
%cmp = icmp eq i32 %i_plus_1, 1001 |
@@ -1414,7 +1414,7 @@ entry: |
br label %body |
body: |
%i = phi i32 [ 1, %entry ], [ %i_plus_1, %body ] |
- %g = bitcast [4 x i8]* @Global32 to i32* |
+ %g = bitcast [4 x i8]* @SzGlobal32 to i32* |
%unused = call i32 @llvm.nacl.atomic.rmw.i32(i32 6, i32* %g, i32 %i, i32 6) |
%i_plus_1 = add i32 %i, 1 |
%cmp = icmp eq i32 %i_plus_1, 1001 |
@@ -1435,7 +1435,7 @@ entry: |
br label %body |
body: |
%i = phi i32 [ 1, %entry ], [ %i_plus_1, %body ] |
- %g = bitcast [4 x i8]* @Global32 to i32* |
+ %g = bitcast [4 x i8]* @SzGlobal32 to i32* |
%unused = call i32 @llvm.nacl.atomic.cmpxchg.i32(i32* %g, i32 %i, i32 %i, i32 6, i32 6) |
%i_plus_1 = add i32 %i, 1 |
%cmp = icmp eq i32 %i_plus_1, 1001 |
@@ -1456,7 +1456,7 @@ entry: |
br label %body |
body: |
%i = phi i32 [ 1, %entry ], [ %i_plus_1, %body ] |
- %g = bitcast [8 x i8]* @Global64 to i64* |
+ %g = bitcast [8 x i8]* @SzGlobal64 to i64* |
%i_64 = zext i32 %i to i64 |
%unused = call i64 @llvm.nacl.atomic.cmpxchg.i64(i64* %g, i64 %i_64, i64 %i_64, i32 6, i32 6) |
%i_plus_1 = add i32 %i, 1 |