OLD | NEW |
1 ; Test parsing NaCl atomic instructions. | 1 ; Test parsing NaCl atomic instructions. |
2 | 2 |
3 ; RUN: %p2i -i %s --insts | FileCheck %s | 3 ; RUN: %p2i -i %s --insts | FileCheck %s |
4 ; RUN: %if --need=allow_disable_ir_gen --command \ | 4 ; RUN: %if --need=allow_disable_ir_gen --command \ |
5 ; RUN: %p2i -i %s --args -notranslate -timing -no-ir-gen \ | 5 ; RUN: %p2i -i %s --args -notranslate -timing -no-ir-gen \ |
6 ; RUN: | %if --need=allow_disable_ir_gen --command \ | 6 ; RUN: | %if --need=allow_disable_ir_gen --command \ |
7 ; RUN: FileCheck --check-prefix=NOIR %s | 7 ; RUN: FileCheck --check-prefix=NOIR %s |
8 | 8 |
9 declare i8 @llvm.nacl.atomic.load.i8(i8*, i32) | 9 declare i8 @llvm.nacl.atomic.load.i8(i8*, i32) |
10 declare i16 @llvm.nacl.atomic.load.i16(i16*, i32) | 10 declare i16 @llvm.nacl.atomic.load.i16(i16*, i32) |
(...skipping 10 matching lines...) Expand all Loading... |
21 declare i8 @llvm.nacl.atomic.cmpxchg.i8(i8*, i8, i8, i32, i32) | 21 declare i8 @llvm.nacl.atomic.cmpxchg.i8(i8*, i8, i8, i32, i32) |
22 declare i16 @llvm.nacl.atomic.cmpxchg.i16(i16*, i16, i16, i32, i32) | 22 declare i16 @llvm.nacl.atomic.cmpxchg.i16(i16*, i16, i16, i32, i32) |
23 declare i32 @llvm.nacl.atomic.cmpxchg.i32(i32*, i32, i32, i32, i32) | 23 declare i32 @llvm.nacl.atomic.cmpxchg.i32(i32*, i32, i32, i32, i32) |
24 declare i64 @llvm.nacl.atomic.cmpxchg.i64(i64*, i64, i64, i32, i32) | 24 declare i64 @llvm.nacl.atomic.cmpxchg.i64(i64*, i64, i64, i32, i32) |
25 declare void @llvm.nacl.atomic.fence(i32) | 25 declare void @llvm.nacl.atomic.fence(i32) |
26 declare void @llvm.nacl.atomic.fence.all() | 26 declare void @llvm.nacl.atomic.fence.all() |
27 declare i1 @llvm.nacl.atomic.is.lock.free(i32, i8*) | 27 declare i1 @llvm.nacl.atomic.is.lock.free(i32, i8*) |
28 | 28 |
29 ;;; Load | 29 ;;; Load |
30 | 30 |
31 define i32 @test_atomic_load_8(i32 %iptr) { | 31 define internal i32 @test_atomic_load_8(i32 %iptr) { |
32 entry: | 32 entry: |
33 %ptr = inttoptr i32 %iptr to i8* | 33 %ptr = inttoptr i32 %iptr to i8* |
34 ; parameter value "6" is for the sequential consistency memory order. | 34 ; parameter value "6" is for the sequential consistency memory order. |
35 %i = call i8 @llvm.nacl.atomic.load.i8(i8* %ptr, i32 6) | 35 %i = call i8 @llvm.nacl.atomic.load.i8(i8* %ptr, i32 6) |
36 %r = zext i8 %i to i32 | 36 %r = zext i8 %i to i32 |
37 ret i32 %r | 37 ret i32 %r |
38 } | 38 } |
39 | 39 |
40 ; CHECK: define i32 @test_atomic_load_8(i32 %iptr) { | 40 ; CHECK: define internal i32 @test_atomic_load_8(i32 %iptr) { |
41 ; CHECK-NEXT: entry: | 41 ; CHECK-NEXT: entry: |
42 ; CHECK-NEXT: %i = call i8 @llvm.nacl.atomic.load.i8(i32 %iptr, i32 6) | 42 ; CHECK-NEXT: %i = call i8 @llvm.nacl.atomic.load.i8(i32 %iptr, i32 6) |
43 ; CHECK-NEXT: %r = zext i8 %i to i32 | 43 ; CHECK-NEXT: %r = zext i8 %i to i32 |
44 ; CHECK-NEXT: ret i32 %r | 44 ; CHECK-NEXT: ret i32 %r |
45 ; CHECK-NEXT: } | 45 ; CHECK-NEXT: } |
46 | 46 |
47 define i32 @test_atomic_load_16(i32 %iptr) { | 47 define internal i32 @test_atomic_load_16(i32 %iptr) { |
48 entry: | 48 entry: |
49 %ptr = inttoptr i32 %iptr to i16* | 49 %ptr = inttoptr i32 %iptr to i16* |
50 %i = call i16 @llvm.nacl.atomic.load.i16(i16* %ptr, i32 6) | 50 %i = call i16 @llvm.nacl.atomic.load.i16(i16* %ptr, i32 6) |
51 %r = zext i16 %i to i32 | 51 %r = zext i16 %i to i32 |
52 ret i32 %r | 52 ret i32 %r |
53 } | 53 } |
54 | 54 |
55 ; CHECK-NEXT: define i32 @test_atomic_load_16(i32 %iptr) { | 55 ; CHECK-NEXT: define internal i32 @test_atomic_load_16(i32 %iptr) { |
56 ; CHECK-NEXT: entry: | 56 ; CHECK-NEXT: entry: |
57 ; CHECK-NEXT: %i = call i16 @llvm.nacl.atomic.load.i16(i32 %iptr, i32 6) | 57 ; CHECK-NEXT: %i = call i16 @llvm.nacl.atomic.load.i16(i32 %iptr, i32 6) |
58 ; CHECK-NEXT: %r = zext i16 %i to i32 | 58 ; CHECK-NEXT: %r = zext i16 %i to i32 |
59 ; CHECK-NEXT: ret i32 %r | 59 ; CHECK-NEXT: ret i32 %r |
60 ; CHECK-NEXT: } | 60 ; CHECK-NEXT: } |
61 | 61 |
62 define i32 @test_atomic_load_32(i32 %iptr) { | 62 define internal i32 @test_atomic_load_32(i32 %iptr) { |
63 entry: | 63 entry: |
64 %ptr = inttoptr i32 %iptr to i32* | 64 %ptr = inttoptr i32 %iptr to i32* |
65 %r = call i32 @llvm.nacl.atomic.load.i32(i32* %ptr, i32 6) | 65 %r = call i32 @llvm.nacl.atomic.load.i32(i32* %ptr, i32 6) |
66 ret i32 %r | 66 ret i32 %r |
67 } | 67 } |
68 | 68 |
69 ; CHECK-NEXT: define i32 @test_atomic_load_32(i32 %iptr) { | 69 ; CHECK-NEXT: define internal i32 @test_atomic_load_32(i32 %iptr) { |
70 ; CHECK-NEXT: entry: | 70 ; CHECK-NEXT: entry: |
71 ; CHECK-NEXT: %r = call i32 @llvm.nacl.atomic.load.i32(i32 %iptr, i32 6) | 71 ; CHECK-NEXT: %r = call i32 @llvm.nacl.atomic.load.i32(i32 %iptr, i32 6) |
72 ; CHECK-NEXT: ret i32 %r | 72 ; CHECK-NEXT: ret i32 %r |
73 ; CHECK-NEXT: } | 73 ; CHECK-NEXT: } |
74 | 74 |
75 define i64 @test_atomic_load_64(i32 %iptr) { | 75 define internal i64 @test_atomic_load_64(i32 %iptr) { |
76 entry: | 76 entry: |
77 %ptr = inttoptr i32 %iptr to i64* | 77 %ptr = inttoptr i32 %iptr to i64* |
78 %r = call i64 @llvm.nacl.atomic.load.i64(i64* %ptr, i32 6) | 78 %r = call i64 @llvm.nacl.atomic.load.i64(i64* %ptr, i32 6) |
79 ret i64 %r | 79 ret i64 %r |
80 } | 80 } |
81 | 81 |
82 ; CHECK-NEXT: define i64 @test_atomic_load_64(i32 %iptr) { | 82 ; CHECK-NEXT: define internal i64 @test_atomic_load_64(i32 %iptr) { |
83 ; CHECK-NEXT: entry: | 83 ; CHECK-NEXT: entry: |
84 ; CHECK-NEXT: %r = call i64 @llvm.nacl.atomic.load.i64(i32 %iptr, i32 6) | 84 ; CHECK-NEXT: %r = call i64 @llvm.nacl.atomic.load.i64(i32 %iptr, i32 6) |
85 ; CHECK-NEXT: ret i64 %r | 85 ; CHECK-NEXT: ret i64 %r |
86 ; CHECK-NEXT: } | 86 ; CHECK-NEXT: } |
87 | 87 |
88 ;;; Store | 88 ;;; Store |
89 | 89 |
90 define void @test_atomic_store_8(i32 %iptr, i32 %v) { | 90 define internal void @test_atomic_store_8(i32 %iptr, i32 %v) { |
91 entry: | 91 entry: |
92 %truncv = trunc i32 %v to i8 | 92 %truncv = trunc i32 %v to i8 |
93 %ptr = inttoptr i32 %iptr to i8* | 93 %ptr = inttoptr i32 %iptr to i8* |
94 call void @llvm.nacl.atomic.store.i8(i8 %truncv, i8* %ptr, i32 6) | 94 call void @llvm.nacl.atomic.store.i8(i8 %truncv, i8* %ptr, i32 6) |
95 ret void | 95 ret void |
96 } | 96 } |
97 | 97 |
98 ; CHECK-NEXT: define void @test_atomic_store_8(i32 %iptr, i32 %v) { | 98 ; CHECK-NEXT: define internal void @test_atomic_store_8(i32 %iptr, i32 %v) { |
99 ; CHECK-NEXT: entry: | 99 ; CHECK-NEXT: entry: |
100 ; CHECK-NEXT: %truncv = trunc i32 %v to i8 | 100 ; CHECK-NEXT: %truncv = trunc i32 %v to i8 |
101 ; CHECK-NEXT: call void @llvm.nacl.atomic.store.i8(i8 %truncv, i32 %iptr, i32
6) | 101 ; CHECK-NEXT: call void @llvm.nacl.atomic.store.i8(i8 %truncv, i32 %iptr, i32
6) |
102 ; CHECK-NEXT: ret void | 102 ; CHECK-NEXT: ret void |
103 ; CHECK-NEXT: } | 103 ; CHECK-NEXT: } |
104 | 104 |
105 define void @test_atomic_store_16(i32 %iptr, i32 %v) { | 105 define internal void @test_atomic_store_16(i32 %iptr, i32 %v) { |
106 entry: | 106 entry: |
107 %truncv = trunc i32 %v to i16 | 107 %truncv = trunc i32 %v to i16 |
108 %ptr = inttoptr i32 %iptr to i16* | 108 %ptr = inttoptr i32 %iptr to i16* |
109 call void @llvm.nacl.atomic.store.i16(i16 %truncv, i16* %ptr, i32 6) | 109 call void @llvm.nacl.atomic.store.i16(i16 %truncv, i16* %ptr, i32 6) |
110 ret void | 110 ret void |
111 } | 111 } |
112 | 112 |
113 ; CHECK-NEXT: define void @test_atomic_store_16(i32 %iptr, i32 %v) { | 113 ; CHECK-NEXT: define internal void @test_atomic_store_16(i32 %iptr, i32 %v) { |
114 ; CHECK-NEXT: entry: | 114 ; CHECK-NEXT: entry: |
115 ; CHECK-NEXT: %truncv = trunc i32 %v to i16 | 115 ; CHECK-NEXT: %truncv = trunc i32 %v to i16 |
116 ; CHECK-NEXT: call void @llvm.nacl.atomic.store.i16(i16 %truncv, i32 %iptr, i3
2 6) | 116 ; CHECK-NEXT: call void @llvm.nacl.atomic.store.i16(i16 %truncv, i32 %iptr, i3
2 6) |
117 ; CHECK-NEXT: ret void | 117 ; CHECK-NEXT: ret void |
118 ; CHECK-NEXT: } | 118 ; CHECK-NEXT: } |
119 | 119 |
120 define void @test_atomic_store_32(i32 %iptr, i32 %v) { | 120 define internal void @test_atomic_store_32(i32 %iptr, i32 %v) { |
121 entry: | 121 entry: |
122 %ptr = inttoptr i32 %iptr to i32* | 122 %ptr = inttoptr i32 %iptr to i32* |
123 call void @llvm.nacl.atomic.store.i32(i32 %v, i32* %ptr, i32 6) | 123 call void @llvm.nacl.atomic.store.i32(i32 %v, i32* %ptr, i32 6) |
124 ret void | 124 ret void |
125 } | 125 } |
126 | 126 |
127 ; CHECK-NEXT: define void @test_atomic_store_32(i32 %iptr, i32 %v) { | 127 ; CHECK-NEXT: define internal void @test_atomic_store_32(i32 %iptr, i32 %v) { |
128 ; CHECK-NEXT: entry: | 128 ; CHECK-NEXT: entry: |
129 ; CHECK-NEXT: call void @llvm.nacl.atomic.store.i32(i32 %v, i32 %iptr, i32 6) | 129 ; CHECK-NEXT: call void @llvm.nacl.atomic.store.i32(i32 %v, i32 %iptr, i32 6) |
130 ; CHECK-NEXT: ret void | 130 ; CHECK-NEXT: ret void |
131 ; CHECK-NEXT: } | 131 ; CHECK-NEXT: } |
132 | 132 |
133 define void @test_atomic_store_64(i32 %iptr, i64 %v) { | 133 define internal void @test_atomic_store_64(i32 %iptr, i64 %v) { |
134 entry: | 134 entry: |
135 %ptr = inttoptr i32 %iptr to i64* | 135 %ptr = inttoptr i32 %iptr to i64* |
136 call void @llvm.nacl.atomic.store.i64(i64 %v, i64* %ptr, i32 6) | 136 call void @llvm.nacl.atomic.store.i64(i64 %v, i64* %ptr, i32 6) |
137 ret void | 137 ret void |
138 } | 138 } |
139 | 139 |
140 ; CHECK-NEXT: define void @test_atomic_store_64(i32 %iptr, i64 %v) { | 140 ; CHECK-NEXT: define internal void @test_atomic_store_64(i32 %iptr, i64 %v) { |
141 ; CHECK-NEXT: entry: | 141 ; CHECK-NEXT: entry: |
142 ; CHECK-NEXT: call void @llvm.nacl.atomic.store.i64(i64 %v, i32 %iptr, i32 6) | 142 ; CHECK-NEXT: call void @llvm.nacl.atomic.store.i64(i64 %v, i32 %iptr, i32 6) |
143 ; CHECK-NEXT: ret void | 143 ; CHECK-NEXT: ret void |
144 ; CHECK-NEXT: } | 144 ; CHECK-NEXT: } |
145 | 145 |
146 define void @test_atomic_store_64_const(i32 %iptr) { | 146 define internal void @test_atomic_store_64_const(i32 %iptr) { |
147 entry: | 147 entry: |
148 %ptr = inttoptr i32 %iptr to i64* | 148 %ptr = inttoptr i32 %iptr to i64* |
149 call void @llvm.nacl.atomic.store.i64(i64 12345678901234, i64* %ptr, i32 6) | 149 call void @llvm.nacl.atomic.store.i64(i64 12345678901234, i64* %ptr, i32 6) |
150 ret void | 150 ret void |
151 } | 151 } |
152 | 152 |
153 ; CHECK-NEXT: define void @test_atomic_store_64_const(i32 %iptr) { | 153 ; CHECK-NEXT: define internal void @test_atomic_store_64_const(i32 %iptr) { |
154 ; CHECK-NEXT: entry: | 154 ; CHECK-NEXT: entry: |
155 ; CHECK-NEXT: call void @llvm.nacl.atomic.store.i64(i64 12345678901234, i32 %i
ptr, i32 6) | 155 ; CHECK-NEXT: call void @llvm.nacl.atomic.store.i64(i64 12345678901234, i32 %i
ptr, i32 6) |
156 ; CHECK-NEXT: ret void | 156 ; CHECK-NEXT: ret void |
157 ; CHECK-NEXT: } | 157 ; CHECK-NEXT: } |
158 | 158 |
159 ;;; RMW | 159 ;;; RMW |
160 | 160 |
161 ;; add | 161 ;; add |
162 | 162 |
163 define i32 @test_atomic_rmw_add_8(i32 %iptr, i32 %v) { | 163 define internal i32 @test_atomic_rmw_add_8(i32 %iptr, i32 %v) { |
164 entry: | 164 entry: |
165 %trunc = trunc i32 %v to i8 | 165 %trunc = trunc i32 %v to i8 |
166 %ptr = inttoptr i32 %iptr to i8* | 166 %ptr = inttoptr i32 %iptr to i8* |
167 ; "1" is an atomic add, and "6" is sequential consistency. | 167 ; "1" is an atomic add, and "6" is sequential consistency. |
168 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 1, i8* %ptr, i8 %trunc, i32 6) | 168 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 1, i8* %ptr, i8 %trunc, i32 6) |
169 %a_ext = zext i8 %a to i32 | 169 %a_ext = zext i8 %a to i32 |
170 ret i32 %a_ext | 170 ret i32 %a_ext |
171 } | 171 } |
172 | 172 |
173 ; CHECK-NEXT: define i32 @test_atomic_rmw_add_8(i32 %iptr, i32 %v) { | 173 ; CHECK-NEXT: define internal i32 @test_atomic_rmw_add_8(i32 %iptr, i32 %v) { |
174 ; CHECK-NEXT: entry: | 174 ; CHECK-NEXT: entry: |
175 ; CHECK-NEXT: %trunc = trunc i32 %v to i8 | 175 ; CHECK-NEXT: %trunc = trunc i32 %v to i8 |
176 ; CHECK-NEXT: %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 1, i32 %iptr, i8 %trun
c, i32 6) | 176 ; CHECK-NEXT: %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 1, i32 %iptr, i8 %trun
c, i32 6) |
177 ; CHECK-NEXT: %a_ext = zext i8 %a to i32 | 177 ; CHECK-NEXT: %a_ext = zext i8 %a to i32 |
178 ; CHECK-NEXT: ret i32 %a_ext | 178 ; CHECK-NEXT: ret i32 %a_ext |
179 ; CHECK-NEXT: } | 179 ; CHECK-NEXT: } |
180 | 180 |
181 define i32 @test_atomic_rmw_add_16(i32 %iptr, i32 %v) { | 181 define internal i32 @test_atomic_rmw_add_16(i32 %iptr, i32 %v) { |
182 entry: | 182 entry: |
183 %trunc = trunc i32 %v to i16 | 183 %trunc = trunc i32 %v to i16 |
184 %ptr = inttoptr i32 %iptr to i16* | 184 %ptr = inttoptr i32 %iptr to i16* |
185 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 1, i16* %ptr, i16 %trunc, i32 6) | 185 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 1, i16* %ptr, i16 %trunc, i32 6) |
186 %a_ext = zext i16 %a to i32 | 186 %a_ext = zext i16 %a to i32 |
187 ret i32 %a_ext | 187 ret i32 %a_ext |
188 } | 188 } |
189 | 189 |
190 ; CHECK-NEXT: define i32 @test_atomic_rmw_add_16(i32 %iptr, i32 %v) { | 190 ; CHECK-NEXT: define internal i32 @test_atomic_rmw_add_16(i32 %iptr, i32 %v) { |
191 ; CHECK-NEXT: entry: | 191 ; CHECK-NEXT: entry: |
192 ; CHECK-NEXT: %trunc = trunc i32 %v to i16 | 192 ; CHECK-NEXT: %trunc = trunc i32 %v to i16 |
193 ; CHECK-NEXT: %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 1, i32 %iptr, i16 %t
runc, i32 6) | 193 ; CHECK-NEXT: %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 1, i32 %iptr, i16 %t
runc, i32 6) |
194 ; CHECK-NEXT: %a_ext = zext i16 %a to i32 | 194 ; CHECK-NEXT: %a_ext = zext i16 %a to i32 |
195 ; CHECK-NEXT: ret i32 %a_ext | 195 ; CHECK-NEXT: ret i32 %a_ext |
196 ; CHECK-NEXT: } | 196 ; CHECK-NEXT: } |
197 | 197 |
198 define i32 @test_atomic_rmw_add_32(i32 %iptr, i32 %v) { | 198 define internal i32 @test_atomic_rmw_add_32(i32 %iptr, i32 %v) { |
199 entry: | 199 entry: |
200 %ptr = inttoptr i32 %iptr to i32* | 200 %ptr = inttoptr i32 %iptr to i32* |
201 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 1, i32* %ptr, i32 %v, i32 6) | 201 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 1, i32* %ptr, i32 %v, i32 6) |
202 ret i32 %a | 202 ret i32 %a |
203 } | 203 } |
204 | 204 |
205 ; CHECK-NEXT: define i32 @test_atomic_rmw_add_32(i32 %iptr, i32 %v) { | 205 ; CHECK-NEXT: define internal i32 @test_atomic_rmw_add_32(i32 %iptr, i32 %v) { |
206 ; CHECK-NEXT: entry: | 206 ; CHECK-NEXT: entry: |
207 ; CHECK-NEXT: %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 1, i32 %iptr, i32 %v
, i32 6) | 207 ; CHECK-NEXT: %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 1, i32 %iptr, i32 %v
, i32 6) |
208 ; CHECK-NEXT: ret i32 %a | 208 ; CHECK-NEXT: ret i32 %a |
209 ; CHECK-NEXT: } | 209 ; CHECK-NEXT: } |
210 | 210 |
211 define i64 @test_atomic_rmw_add_64(i32 %iptr, i64 %v) { | 211 define internal i64 @test_atomic_rmw_add_64(i32 %iptr, i64 %v) { |
212 entry: | 212 entry: |
213 %ptr = inttoptr i32 %iptr to i64* | 213 %ptr = inttoptr i32 %iptr to i64* |
214 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 1, i64* %ptr, i64 %v, i32 6) | 214 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 1, i64* %ptr, i64 %v, i32 6) |
215 ret i64 %a | 215 ret i64 %a |
216 } | 216 } |
217 | 217 |
218 ; CHECK-NEXT: define i64 @test_atomic_rmw_add_64(i32 %iptr, i64 %v) { | 218 ; CHECK-NEXT: define internal i64 @test_atomic_rmw_add_64(i32 %iptr, i64 %v) { |
219 ; CHECK-NEXT: entry: | 219 ; CHECK-NEXT: entry: |
220 ; CHECK-NEXT: %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 1, i32 %iptr, i64 %v
, i32 6) | 220 ; CHECK-NEXT: %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 1, i32 %iptr, i64 %v
, i32 6) |
221 ; CHECK-NEXT: ret i64 %a | 221 ; CHECK-NEXT: ret i64 %a |
222 ; CHECK-NEXT: } | 222 ; CHECK-NEXT: } |
223 | 223 |
224 ;; sub | 224 ;; sub |
225 | 225 |
226 define i32 @test_atomic_rmw_sub_8(i32 %iptr, i32 %v) { | 226 define internal i32 @test_atomic_rmw_sub_8(i32 %iptr, i32 %v) { |
227 entry: | 227 entry: |
228 %trunc = trunc i32 %v to i8 | 228 %trunc = trunc i32 %v to i8 |
229 %ptr = inttoptr i32 %iptr to i8* | 229 %ptr = inttoptr i32 %iptr to i8* |
230 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 2, i8* %ptr, i8 %trunc, i32 6) | 230 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 2, i8* %ptr, i8 %trunc, i32 6) |
231 %a_ext = zext i8 %a to i32 | 231 %a_ext = zext i8 %a to i32 |
232 ret i32 %a_ext | 232 ret i32 %a_ext |
233 } | 233 } |
234 | 234 |
235 ; CHECK-NEXT: define i32 @test_atomic_rmw_sub_8(i32 %iptr, i32 %v) { | 235 ; CHECK-NEXT: define internal i32 @test_atomic_rmw_sub_8(i32 %iptr, i32 %v) { |
236 ; CHECK-NEXT: entry: | 236 ; CHECK-NEXT: entry: |
237 ; CHECK-NEXT: %trunc = trunc i32 %v to i8 | 237 ; CHECK-NEXT: %trunc = trunc i32 %v to i8 |
238 ; CHECK-NEXT: %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 2, i32 %iptr, i8 %trun
c, i32 6) | 238 ; CHECK-NEXT: %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 2, i32 %iptr, i8 %trun
c, i32 6) |
239 ; CHECK-NEXT: %a_ext = zext i8 %a to i32 | 239 ; CHECK-NEXT: %a_ext = zext i8 %a to i32 |
240 ; CHECK-NEXT: ret i32 %a_ext | 240 ; CHECK-NEXT: ret i32 %a_ext |
241 ; CHECK-NEXT: } | 241 ; CHECK-NEXT: } |
242 | 242 |
243 define i32 @test_atomic_rmw_sub_16(i32 %iptr, i32 %v) { | 243 define internal i32 @test_atomic_rmw_sub_16(i32 %iptr, i32 %v) { |
244 entry: | 244 entry: |
245 %trunc = trunc i32 %v to i16 | 245 %trunc = trunc i32 %v to i16 |
246 %ptr = inttoptr i32 %iptr to i16* | 246 %ptr = inttoptr i32 %iptr to i16* |
247 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 2, i16* %ptr, i16 %trunc, i32 6) | 247 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 2, i16* %ptr, i16 %trunc, i32 6) |
248 %a_ext = zext i16 %a to i32 | 248 %a_ext = zext i16 %a to i32 |
249 ret i32 %a_ext | 249 ret i32 %a_ext |
250 } | 250 } |
251 | 251 |
252 ; CHECK-NEXT: define i32 @test_atomic_rmw_sub_16(i32 %iptr, i32 %v) { | 252 ; CHECK-NEXT: define internal i32 @test_atomic_rmw_sub_16(i32 %iptr, i32 %v) { |
253 ; CHECK-NEXT: entry: | 253 ; CHECK-NEXT: entry: |
254 ; CHECK-NEXT: %trunc = trunc i32 %v to i16 | 254 ; CHECK-NEXT: %trunc = trunc i32 %v to i16 |
255 ; CHECK-NEXT: %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 2, i32 %iptr, i16 %t
runc, i32 6) | 255 ; CHECK-NEXT: %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 2, i32 %iptr, i16 %t
runc, i32 6) |
256 ; CHECK-NEXT: %a_ext = zext i16 %a to i32 | 256 ; CHECK-NEXT: %a_ext = zext i16 %a to i32 |
257 ; CHECK-NEXT: ret i32 %a_ext | 257 ; CHECK-NEXT: ret i32 %a_ext |
258 ; CHECK-NEXT: } | 258 ; CHECK-NEXT: } |
259 | 259 |
260 define i32 @test_atomic_rmw_sub_32(i32 %iptr, i32 %v) { | 260 define internal i32 @test_atomic_rmw_sub_32(i32 %iptr, i32 %v) { |
261 entry: | 261 entry: |
262 %ptr = inttoptr i32 %iptr to i32* | 262 %ptr = inttoptr i32 %iptr to i32* |
263 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 2, i32* %ptr, i32 %v, i32 6) | 263 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 2, i32* %ptr, i32 %v, i32 6) |
264 ret i32 %a | 264 ret i32 %a |
265 } | 265 } |
266 | 266 |
267 ; CHECK-NEXT: define i32 @test_atomic_rmw_sub_32(i32 %iptr, i32 %v) { | 267 ; CHECK-NEXT: define internal i32 @test_atomic_rmw_sub_32(i32 %iptr, i32 %v) { |
268 ; CHECK-NEXT: entry: | 268 ; CHECK-NEXT: entry: |
269 ; CHECK-NEXT: %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 2, i32 %iptr, i32 %v
, i32 6) | 269 ; CHECK-NEXT: %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 2, i32 %iptr, i32 %v
, i32 6) |
270 ; CHECK-NEXT: ret i32 %a | 270 ; CHECK-NEXT: ret i32 %a |
271 ; CHECK-NEXT: } | 271 ; CHECK-NEXT: } |
272 | 272 |
273 define i64 @test_atomic_rmw_sub_64(i32 %iptr, i64 %v) { | 273 define internal i64 @test_atomic_rmw_sub_64(i32 %iptr, i64 %v) { |
274 entry: | 274 entry: |
275 %ptr = inttoptr i32 %iptr to i64* | 275 %ptr = inttoptr i32 %iptr to i64* |
276 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 2, i64* %ptr, i64 %v, i32 6) | 276 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 2, i64* %ptr, i64 %v, i32 6) |
277 ret i64 %a | 277 ret i64 %a |
278 } | 278 } |
279 | 279 |
280 ; CHECK-NEXT: define i64 @test_atomic_rmw_sub_64(i32 %iptr, i64 %v) { | 280 ; CHECK-NEXT: define internal i64 @test_atomic_rmw_sub_64(i32 %iptr, i64 %v) { |
281 ; CHECK-NEXT: entry: | 281 ; CHECK-NEXT: entry: |
282 ; CHECK-NEXT: %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 2, i32 %iptr, i64 %v
, i32 6) | 282 ; CHECK-NEXT: %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 2, i32 %iptr, i64 %v
, i32 6) |
283 ; CHECK-NEXT: ret i64 %a | 283 ; CHECK-NEXT: ret i64 %a |
284 ; CHECK-NEXT: } | 284 ; CHECK-NEXT: } |
285 | 285 |
286 ;; or | 286 ;; or |
287 | 287 |
288 define i32 @test_atomic_rmw_or_8(i32 %iptr, i32 %v) { | 288 define internal i32 @test_atomic_rmw_or_8(i32 %iptr, i32 %v) { |
289 entry: | 289 entry: |
290 %trunc = trunc i32 %v to i8 | 290 %trunc = trunc i32 %v to i8 |
291 %ptr = inttoptr i32 %iptr to i8* | 291 %ptr = inttoptr i32 %iptr to i8* |
292 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 3, i8* %ptr, i8 %trunc, i32 6) | 292 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 3, i8* %ptr, i8 %trunc, i32 6) |
293 %a_ext = zext i8 %a to i32 | 293 %a_ext = zext i8 %a to i32 |
294 ret i32 %a_ext | 294 ret i32 %a_ext |
295 } | 295 } |
296 | 296 |
297 ; CHECK-NEXT: define i32 @test_atomic_rmw_or_8(i32 %iptr, i32 %v) { | 297 ; CHECK-NEXT: define internal i32 @test_atomic_rmw_or_8(i32 %iptr, i32 %v) { |
298 ; CHECK-NEXT: entry: | 298 ; CHECK-NEXT: entry: |
299 ; CHECK-NEXT: %trunc = trunc i32 %v to i8 | 299 ; CHECK-NEXT: %trunc = trunc i32 %v to i8 |
300 ; CHECK-NEXT: %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 3, i32 %iptr, i8 %trun
c, i32 6) | 300 ; CHECK-NEXT: %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 3, i32 %iptr, i8 %trun
c, i32 6) |
301 ; CHECK-NEXT: %a_ext = zext i8 %a to i32 | 301 ; CHECK-NEXT: %a_ext = zext i8 %a to i32 |
302 ; CHECK-NEXT: ret i32 %a_ext | 302 ; CHECK-NEXT: ret i32 %a_ext |
303 ; CHECK-NEXT: } | 303 ; CHECK-NEXT: } |
304 | 304 |
305 define i32 @test_atomic_rmw_or_16(i32 %iptr, i32 %v) { | 305 define internal i32 @test_atomic_rmw_or_16(i32 %iptr, i32 %v) { |
306 entry: | 306 entry: |
307 %trunc = trunc i32 %v to i16 | 307 %trunc = trunc i32 %v to i16 |
308 %ptr = inttoptr i32 %iptr to i16* | 308 %ptr = inttoptr i32 %iptr to i16* |
309 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 3, i16* %ptr, i16 %trunc, i32 6) | 309 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 3, i16* %ptr, i16 %trunc, i32 6) |
310 %a_ext = zext i16 %a to i32 | 310 %a_ext = zext i16 %a to i32 |
311 ret i32 %a_ext | 311 ret i32 %a_ext |
312 } | 312 } |
313 | 313 |
314 ; CHECK-NEXT: define i32 @test_atomic_rmw_or_16(i32 %iptr, i32 %v) { | 314 ; CHECK-NEXT: define internal i32 @test_atomic_rmw_or_16(i32 %iptr, i32 %v) { |
315 ; CHECK-NEXT: entry: | 315 ; CHECK-NEXT: entry: |
316 ; CHECK-NEXT: %trunc = trunc i32 %v to i16 | 316 ; CHECK-NEXT: %trunc = trunc i32 %v to i16 |
317 ; CHECK-NEXT: %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 3, i32 %iptr, i16 %t
runc, i32 6) | 317 ; CHECK-NEXT: %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 3, i32 %iptr, i16 %t
runc, i32 6) |
318 ; CHECK-NEXT: %a_ext = zext i16 %a to i32 | 318 ; CHECK-NEXT: %a_ext = zext i16 %a to i32 |
319 ; CHECK-NEXT: ret i32 %a_ext | 319 ; CHECK-NEXT: ret i32 %a_ext |
320 ; CHECK-NEXT: } | 320 ; CHECK-NEXT: } |
321 | 321 |
322 define i32 @test_atomic_rmw_or_32(i32 %iptr, i32 %v) { | 322 define internal i32 @test_atomic_rmw_or_32(i32 %iptr, i32 %v) { |
323 entry: | 323 entry: |
324 %ptr = inttoptr i32 %iptr to i32* | 324 %ptr = inttoptr i32 %iptr to i32* |
325 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 3, i32* %ptr, i32 %v, i32 6) | 325 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 3, i32* %ptr, i32 %v, i32 6) |
326 ret i32 %a | 326 ret i32 %a |
327 } | 327 } |
328 | 328 |
329 ; CHECK-NEXT: define i32 @test_atomic_rmw_or_32(i32 %iptr, i32 %v) { | 329 ; CHECK-NEXT: define internal i32 @test_atomic_rmw_or_32(i32 %iptr, i32 %v) { |
330 ; CHECK-NEXT: entry: | 330 ; CHECK-NEXT: entry: |
331 ; CHECK-NEXT: %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 3, i32 %iptr, i32 %v
, i32 6) | 331 ; CHECK-NEXT: %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 3, i32 %iptr, i32 %v
, i32 6) |
332 ; CHECK-NEXT: ret i32 %a | 332 ; CHECK-NEXT: ret i32 %a |
333 ; CHECK-NEXT: } | 333 ; CHECK-NEXT: } |
334 | 334 |
335 define i64 @test_atomic_rmw_or_64(i32 %iptr, i64 %v) { | 335 define internal i64 @test_atomic_rmw_or_64(i32 %iptr, i64 %v) { |
336 entry: | 336 entry: |
337 %ptr = inttoptr i32 %iptr to i64* | 337 %ptr = inttoptr i32 %iptr to i64* |
338 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 3, i64* %ptr, i64 %v, i32 6) | 338 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 3, i64* %ptr, i64 %v, i32 6) |
339 ret i64 %a | 339 ret i64 %a |
340 } | 340 } |
341 | 341 |
342 ; CHECK-NEXT: define i64 @test_atomic_rmw_or_64(i32 %iptr, i64 %v) { | 342 ; CHECK-NEXT: define internal i64 @test_atomic_rmw_or_64(i32 %iptr, i64 %v) { |
343 ; CHECK-NEXT: entry: | 343 ; CHECK-NEXT: entry: |
344 ; CHECK-NEXT: %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 3, i32 %iptr, i64 %v
, i32 6) | 344 ; CHECK-NEXT: %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 3, i32 %iptr, i64 %v
, i32 6) |
345 ; CHECK-NEXT: ret i64 %a | 345 ; CHECK-NEXT: ret i64 %a |
346 ; CHECK-NEXT: } | 346 ; CHECK-NEXT: } |
347 | 347 |
348 ;; and | 348 ;; and |
349 | 349 |
350 define i32 @test_atomic_rmw_and_8(i32 %iptr, i32 %v) { | 350 define internal i32 @test_atomic_rmw_and_8(i32 %iptr, i32 %v) { |
351 entry: | 351 entry: |
352 %trunc = trunc i32 %v to i8 | 352 %trunc = trunc i32 %v to i8 |
353 %ptr = inttoptr i32 %iptr to i8* | 353 %ptr = inttoptr i32 %iptr to i8* |
354 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 4, i8* %ptr, i8 %trunc, i32 6) | 354 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 4, i8* %ptr, i8 %trunc, i32 6) |
355 %a_ext = zext i8 %a to i32 | 355 %a_ext = zext i8 %a to i32 |
356 ret i32 %a_ext | 356 ret i32 %a_ext |
357 } | 357 } |
358 | 358 |
359 ; CHECK-NEXT: define i32 @test_atomic_rmw_and_8(i32 %iptr, i32 %v) { | 359 ; CHECK-NEXT: define internal i32 @test_atomic_rmw_and_8(i32 %iptr, i32 %v) { |
360 ; CHECK-NEXT: entry: | 360 ; CHECK-NEXT: entry: |
361 ; CHECK-NEXT: %trunc = trunc i32 %v to i8 | 361 ; CHECK-NEXT: %trunc = trunc i32 %v to i8 |
362 ; CHECK-NEXT: %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 4, i32 %iptr, i8 %trun
c, i32 6) | 362 ; CHECK-NEXT: %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 4, i32 %iptr, i8 %trun
c, i32 6) |
363 ; CHECK-NEXT: %a_ext = zext i8 %a to i32 | 363 ; CHECK-NEXT: %a_ext = zext i8 %a to i32 |
364 ; CHECK-NEXT: ret i32 %a_ext | 364 ; CHECK-NEXT: ret i32 %a_ext |
365 ; CHECK-NEXT: } | 365 ; CHECK-NEXT: } |
366 | 366 |
367 define i32 @test_atomic_rmw_and_16(i32 %iptr, i32 %v) { | 367 define internal i32 @test_atomic_rmw_and_16(i32 %iptr, i32 %v) { |
368 entry: | 368 entry: |
369 %trunc = trunc i32 %v to i16 | 369 %trunc = trunc i32 %v to i16 |
370 %ptr = inttoptr i32 %iptr to i16* | 370 %ptr = inttoptr i32 %iptr to i16* |
371 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 4, i16* %ptr, i16 %trunc, i32 6) | 371 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 4, i16* %ptr, i16 %trunc, i32 6) |
372 %a_ext = zext i16 %a to i32 | 372 %a_ext = zext i16 %a to i32 |
373 ret i32 %a_ext | 373 ret i32 %a_ext |
374 } | 374 } |
375 | 375 |
376 ; CHECK-NEXT: define i32 @test_atomic_rmw_and_16(i32 %iptr, i32 %v) { | 376 ; CHECK-NEXT: define internal i32 @test_atomic_rmw_and_16(i32 %iptr, i32 %v) { |
377 ; CHECK-NEXT: entry: | 377 ; CHECK-NEXT: entry: |
378 ; CHECK-NEXT: %trunc = trunc i32 %v to i16 | 378 ; CHECK-NEXT: %trunc = trunc i32 %v to i16 |
379 ; CHECK-NEXT: %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 4, i32 %iptr, i16 %t
runc, i32 6) | 379 ; CHECK-NEXT: %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 4, i32 %iptr, i16 %t
runc, i32 6) |
380 ; CHECK-NEXT: %a_ext = zext i16 %a to i32 | 380 ; CHECK-NEXT: %a_ext = zext i16 %a to i32 |
381 ; CHECK-NEXT: ret i32 %a_ext | 381 ; CHECK-NEXT: ret i32 %a_ext |
382 ; CHECK-NEXT: } | 382 ; CHECK-NEXT: } |
383 | 383 |
384 define i32 @test_atomic_rmw_and_32(i32 %iptr, i32 %v) { | 384 define internal i32 @test_atomic_rmw_and_32(i32 %iptr, i32 %v) { |
385 entry: | 385 entry: |
386 %ptr = inttoptr i32 %iptr to i32* | 386 %ptr = inttoptr i32 %iptr to i32* |
387 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 4, i32* %ptr, i32 %v, i32 6) | 387 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 4, i32* %ptr, i32 %v, i32 6) |
388 ret i32 %a | 388 ret i32 %a |
389 } | 389 } |
390 | 390 |
391 ; CHECK-NEXT: define i32 @test_atomic_rmw_and_32(i32 %iptr, i32 %v) { | 391 ; CHECK-NEXT: define internal i32 @test_atomic_rmw_and_32(i32 %iptr, i32 %v) { |
392 ; CHECK-NEXT: entry: | 392 ; CHECK-NEXT: entry: |
393 ; CHECK-NEXT: %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 4, i32 %iptr, i32 %v
, i32 6) | 393 ; CHECK-NEXT: %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 4, i32 %iptr, i32 %v
, i32 6) |
394 ; CHECK-NEXT: ret i32 %a | 394 ; CHECK-NEXT: ret i32 %a |
395 ; CHECK-NEXT: } | 395 ; CHECK-NEXT: } |
396 | 396 |
397 define i64 @test_atomic_rmw_and_64(i32 %iptr, i64 %v) { | 397 define internal i64 @test_atomic_rmw_and_64(i32 %iptr, i64 %v) { |
398 entry: | 398 entry: |
399 %ptr = inttoptr i32 %iptr to i64* | 399 %ptr = inttoptr i32 %iptr to i64* |
400 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 4, i64* %ptr, i64 %v, i32 6) | 400 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 4, i64* %ptr, i64 %v, i32 6) |
401 ret i64 %a | 401 ret i64 %a |
402 } | 402 } |
403 | 403 |
404 ; CHECK-NEXT: define i64 @test_atomic_rmw_and_64(i32 %iptr, i64 %v) { | 404 ; CHECK-NEXT: define internal i64 @test_atomic_rmw_and_64(i32 %iptr, i64 %v) { |
405 ; CHECK-NEXT: entry: | 405 ; CHECK-NEXT: entry: |
406 ; CHECK-NEXT: %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 4, i32 %iptr, i64 %v
, i32 6) | 406 ; CHECK-NEXT: %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 4, i32 %iptr, i64 %v
, i32 6) |
407 ; CHECK-NEXT: ret i64 %a | 407 ; CHECK-NEXT: ret i64 %a |
408 ; CHECK-NEXT: } | 408 ; CHECK-NEXT: } |
409 | 409 |
410 ;; xor | 410 ;; xor |
411 | 411 |
412 define i32 @test_atomic_rmw_xor_8(i32 %iptr, i32 %v) { | 412 define internal i32 @test_atomic_rmw_xor_8(i32 %iptr, i32 %v) { |
413 entry: | 413 entry: |
414 %trunc = trunc i32 %v to i8 | 414 %trunc = trunc i32 %v to i8 |
415 %ptr = inttoptr i32 %iptr to i8* | 415 %ptr = inttoptr i32 %iptr to i8* |
416 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 5, i8* %ptr, i8 %trunc, i32 6) | 416 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 5, i8* %ptr, i8 %trunc, i32 6) |
417 %a_ext = zext i8 %a to i32 | 417 %a_ext = zext i8 %a to i32 |
418 ret i32 %a_ext | 418 ret i32 %a_ext |
419 } | 419 } |
420 | 420 |
421 ; CHECK-NEXT: define i32 @test_atomic_rmw_xor_8(i32 %iptr, i32 %v) { | 421 ; CHECK-NEXT: define internal i32 @test_atomic_rmw_xor_8(i32 %iptr, i32 %v) { |
422 ; CHECK-NEXT: entry: | 422 ; CHECK-NEXT: entry: |
423 ; CHECK-NEXT: %trunc = trunc i32 %v to i8 | 423 ; CHECK-NEXT: %trunc = trunc i32 %v to i8 |
424 ; CHECK-NEXT: %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 5, i32 %iptr, i8 %trun
c, i32 6) | 424 ; CHECK-NEXT: %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 5, i32 %iptr, i8 %trun
c, i32 6) |
425 ; CHECK-NEXT: %a_ext = zext i8 %a to i32 | 425 ; CHECK-NEXT: %a_ext = zext i8 %a to i32 |
426 ; CHECK-NEXT: ret i32 %a_ext | 426 ; CHECK-NEXT: ret i32 %a_ext |
427 ; CHECK-NEXT: } | 427 ; CHECK-NEXT: } |
428 | 428 |
429 define i32 @test_atomic_rmw_xor_16(i32 %iptr, i32 %v) { | 429 define internal i32 @test_atomic_rmw_xor_16(i32 %iptr, i32 %v) { |
430 entry: | 430 entry: |
431 %trunc = trunc i32 %v to i16 | 431 %trunc = trunc i32 %v to i16 |
432 %ptr = inttoptr i32 %iptr to i16* | 432 %ptr = inttoptr i32 %iptr to i16* |
433 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 5, i16* %ptr, i16 %trunc, i32 6) | 433 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 5, i16* %ptr, i16 %trunc, i32 6) |
434 %a_ext = zext i16 %a to i32 | 434 %a_ext = zext i16 %a to i32 |
435 ret i32 %a_ext | 435 ret i32 %a_ext |
436 } | 436 } |
437 | 437 |
438 ; CHECK-NEXT: define i32 @test_atomic_rmw_xor_16(i32 %iptr, i32 %v) { | 438 ; CHECK-NEXT: define internal i32 @test_atomic_rmw_xor_16(i32 %iptr, i32 %v) { |
439 ; CHECK-NEXT: entry: | 439 ; CHECK-NEXT: entry: |
440 ; CHECK-NEXT: %trunc = trunc i32 %v to i16 | 440 ; CHECK-NEXT: %trunc = trunc i32 %v to i16 |
441 ; CHECK-NEXT: %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 5, i32 %iptr, i16 %t
runc, i32 6) | 441 ; CHECK-NEXT: %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 5, i32 %iptr, i16 %t
runc, i32 6) |
442 ; CHECK-NEXT: %a_ext = zext i16 %a to i32 | 442 ; CHECK-NEXT: %a_ext = zext i16 %a to i32 |
443 ; CHECK-NEXT: ret i32 %a_ext | 443 ; CHECK-NEXT: ret i32 %a_ext |
444 ; CHECK-NEXT: } | 444 ; CHECK-NEXT: } |
445 | 445 |
446 define i32 @test_atomic_rmw_xor_32(i32 %iptr, i32 %v) { | 446 define internal i32 @test_atomic_rmw_xor_32(i32 %iptr, i32 %v) { |
447 entry: | 447 entry: |
448 %ptr = inttoptr i32 %iptr to i32* | 448 %ptr = inttoptr i32 %iptr to i32* |
449 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 5, i32* %ptr, i32 %v, i32 6) | 449 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 5, i32* %ptr, i32 %v, i32 6) |
450 ret i32 %a | 450 ret i32 %a |
451 } | 451 } |
452 | 452 |
453 ; CHECK-NEXT: define i32 @test_atomic_rmw_xor_32(i32 %iptr, i32 %v) { | 453 ; CHECK-NEXT: define internal i32 @test_atomic_rmw_xor_32(i32 %iptr, i32 %v) { |
454 ; CHECK-NEXT: entry: | 454 ; CHECK-NEXT: entry: |
455 ; CHECK-NEXT: %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 5, i32 %iptr, i32 %v
, i32 6) | 455 ; CHECK-NEXT: %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 5, i32 %iptr, i32 %v
, i32 6) |
456 ; CHECK-NEXT: ret i32 %a | 456 ; CHECK-NEXT: ret i32 %a |
457 ; CHECK-NEXT: } | 457 ; CHECK-NEXT: } |
458 | 458 |
459 define i64 @test_atomic_rmw_xor_64(i32 %iptr, i64 %v) { | 459 define internal i64 @test_atomic_rmw_xor_64(i32 %iptr, i64 %v) { |
460 entry: | 460 entry: |
461 %ptr = inttoptr i32 %iptr to i64* | 461 %ptr = inttoptr i32 %iptr to i64* |
462 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 5, i64* %ptr, i64 %v, i32 6) | 462 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 5, i64* %ptr, i64 %v, i32 6) |
463 ret i64 %a | 463 ret i64 %a |
464 } | 464 } |
465 | 465 |
466 ; CHECK-NEXT: define i64 @test_atomic_rmw_xor_64(i32 %iptr, i64 %v) { | 466 ; CHECK-NEXT: define internal i64 @test_atomic_rmw_xor_64(i32 %iptr, i64 %v) { |
467 ; CHECK-NEXT: entry: | 467 ; CHECK-NEXT: entry: |
468 ; CHECK-NEXT: %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 5, i32 %iptr, i64 %v
, i32 6) | 468 ; CHECK-NEXT: %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 5, i32 %iptr, i64 %v
, i32 6) |
469 ; CHECK-NEXT: ret i64 %a | 469 ; CHECK-NEXT: ret i64 %a |
470 ; CHECK-NEXT: } | 470 ; CHECK-NEXT: } |
471 | 471 |
472 ;; exchange | 472 ;; exchange |
473 | 473 |
474 define i32 @test_atomic_rmw_xchg_8(i32 %iptr, i32 %v) { | 474 define internal i32 @test_atomic_rmw_xchg_8(i32 %iptr, i32 %v) { |
475 entry: | 475 entry: |
476 %trunc = trunc i32 %v to i8 | 476 %trunc = trunc i32 %v to i8 |
477 %ptr = inttoptr i32 %iptr to i8* | 477 %ptr = inttoptr i32 %iptr to i8* |
478 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 6, i8* %ptr, i8 %trunc, i32 6) | 478 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 6, i8* %ptr, i8 %trunc, i32 6) |
479 %a_ext = zext i8 %a to i32 | 479 %a_ext = zext i8 %a to i32 |
480 ret i32 %a_ext | 480 ret i32 %a_ext |
481 } | 481 } |
482 | 482 |
483 ; CHECK-NEXT: define i32 @test_atomic_rmw_xchg_8(i32 %iptr, i32 %v) { | 483 ; CHECK-NEXT: define internal i32 @test_atomic_rmw_xchg_8(i32 %iptr, i32 %v) { |
484 ; CHECK-NEXT: entry: | 484 ; CHECK-NEXT: entry: |
485 ; CHECK-NEXT: %trunc = trunc i32 %v to i8 | 485 ; CHECK-NEXT: %trunc = trunc i32 %v to i8 |
486 ; CHECK-NEXT: %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 6, i32 %iptr, i8 %trun
c, i32 6) | 486 ; CHECK-NEXT: %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 6, i32 %iptr, i8 %trun
c, i32 6) |
487 ; CHECK-NEXT: %a_ext = zext i8 %a to i32 | 487 ; CHECK-NEXT: %a_ext = zext i8 %a to i32 |
488 ; CHECK-NEXT: ret i32 %a_ext | 488 ; CHECK-NEXT: ret i32 %a_ext |
489 ; CHECK-NEXT: } | 489 ; CHECK-NEXT: } |
490 | 490 |
491 define i32 @test_atomic_rmw_xchg_16(i32 %iptr, i32 %v) { | 491 define internal i32 @test_atomic_rmw_xchg_16(i32 %iptr, i32 %v) { |
492 entry: | 492 entry: |
493 %trunc = trunc i32 %v to i16 | 493 %trunc = trunc i32 %v to i16 |
494 %ptr = inttoptr i32 %iptr to i16* | 494 %ptr = inttoptr i32 %iptr to i16* |
495 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 6, i16* %ptr, i16 %trunc, i32 6) | 495 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 6, i16* %ptr, i16 %trunc, i32 6) |
496 %a_ext = zext i16 %a to i32 | 496 %a_ext = zext i16 %a to i32 |
497 ret i32 %a_ext | 497 ret i32 %a_ext |
498 } | 498 } |
499 | 499 |
500 ; CHECK-NEXT: define i32 @test_atomic_rmw_xchg_16(i32 %iptr, i32 %v) { | 500 ; CHECK-NEXT: define internal i32 @test_atomic_rmw_xchg_16(i32 %iptr, i32 %v) { |
501 ; CHECK-NEXT: entry: | 501 ; CHECK-NEXT: entry: |
502 ; CHECK-NEXT: %trunc = trunc i32 %v to i16 | 502 ; CHECK-NEXT: %trunc = trunc i32 %v to i16 |
503 ; CHECK-NEXT: %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 6, i32 %iptr, i16 %t
runc, i32 6) | 503 ; CHECK-NEXT: %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 6, i32 %iptr, i16 %t
runc, i32 6) |
504 ; CHECK-NEXT: %a_ext = zext i16 %a to i32 | 504 ; CHECK-NEXT: %a_ext = zext i16 %a to i32 |
505 ; CHECK-NEXT: ret i32 %a_ext | 505 ; CHECK-NEXT: ret i32 %a_ext |
506 ; CHECK-NEXT: } | 506 ; CHECK-NEXT: } |
507 | 507 |
508 define i32 @test_atomic_rmw_xchg_32(i32 %iptr, i32 %v) { | 508 define internal i32 @test_atomic_rmw_xchg_32(i32 %iptr, i32 %v) { |
509 entry: | 509 entry: |
510 %ptr = inttoptr i32 %iptr to i32* | 510 %ptr = inttoptr i32 %iptr to i32* |
511 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 6, i32* %ptr, i32 %v, i32 6) | 511 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 6, i32* %ptr, i32 %v, i32 6) |
512 ret i32 %a | 512 ret i32 %a |
513 } | 513 } |
514 | 514 |
515 ; CHECK-NEXT: define i32 @test_atomic_rmw_xchg_32(i32 %iptr, i32 %v) { | 515 ; CHECK-NEXT: define internal i32 @test_atomic_rmw_xchg_32(i32 %iptr, i32 %v) { |
516 ; CHECK-NEXT: entry: | 516 ; CHECK-NEXT: entry: |
517 ; CHECK-NEXT: %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 6, i32 %iptr, i32 %v
, i32 6) | 517 ; CHECK-NEXT: %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 6, i32 %iptr, i32 %v
, i32 6) |
518 ; CHECK-NEXT: ret i32 %a | 518 ; CHECK-NEXT: ret i32 %a |
519 ; CHECK-NEXT: } | 519 ; CHECK-NEXT: } |
520 | 520 |
521 define i64 @test_atomic_rmw_xchg_64(i32 %iptr, i64 %v) { | 521 define internal i64 @test_atomic_rmw_xchg_64(i32 %iptr, i64 %v) { |
522 entry: | 522 entry: |
523 %ptr = inttoptr i32 %iptr to i64* | 523 %ptr = inttoptr i32 %iptr to i64* |
524 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 6, i64* %ptr, i64 %v, i32 6) | 524 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 6, i64* %ptr, i64 %v, i32 6) |
525 ret i64 %a | 525 ret i64 %a |
526 } | 526 } |
527 | 527 |
528 ; CHECK-NEXT: define i64 @test_atomic_rmw_xchg_64(i32 %iptr, i64 %v) { | 528 ; CHECK-NEXT: define internal i64 @test_atomic_rmw_xchg_64(i32 %iptr, i64 %v) { |
529 ; CHECK-NEXT: entry: | 529 ; CHECK-NEXT: entry: |
530 ; CHECK-NEXT: %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 6, i32 %iptr, i64 %v
, i32 6) | 530 ; CHECK-NEXT: %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 6, i32 %iptr, i64 %v
, i32 6) |
531 ; CHECK-NEXT: ret i64 %a | 531 ; CHECK-NEXT: ret i64 %a |
532 ; CHECK-NEXT: } | 532 ; CHECK-NEXT: } |
533 | 533 |
534 ;;;; Cmpxchg | 534 ;;;; Cmpxchg |
535 | 535 |
536 define i32 @test_atomic_cmpxchg_8(i32 %iptr, i32 %expected, i32 %desired) { | 536 define internal i32 @test_atomic_cmpxchg_8(i32 %iptr, i32 %expected, i32 %desire
d) { |
537 entry: | 537 entry: |
538 %trunc_exp = trunc i32 %expected to i8 | 538 %trunc_exp = trunc i32 %expected to i8 |
539 %trunc_des = trunc i32 %desired to i8 | 539 %trunc_des = trunc i32 %desired to i8 |
540 %ptr = inttoptr i32 %iptr to i8* | 540 %ptr = inttoptr i32 %iptr to i8* |
541 %old = call i8 @llvm.nacl.atomic.cmpxchg.i8(i8* %ptr, i8 %trunc_exp, | 541 %old = call i8 @llvm.nacl.atomic.cmpxchg.i8(i8* %ptr, i8 %trunc_exp, |
542 i8 %trunc_des, i32 6, i32 6) | 542 i8 %trunc_des, i32 6, i32 6) |
543 %old_ext = zext i8 %old to i32 | 543 %old_ext = zext i8 %old to i32 |
544 ret i32 %old_ext | 544 ret i32 %old_ext |
545 } | 545 } |
546 | 546 |
547 ; CHECK-NEXT: define i32 @test_atomic_cmpxchg_8(i32 %iptr, i32 %expected, i32 %d
esired) { | 547 ; CHECK-NEXT: define internal i32 @test_atomic_cmpxchg_8(i32 %iptr, i32 %expecte
d, i32 %desired) { |
548 ; CHECK-NEXT: entry: | 548 ; CHECK-NEXT: entry: |
549 ; CHECK-NEXT: %trunc_exp = trunc i32 %expected to i8 | 549 ; CHECK-NEXT: %trunc_exp = trunc i32 %expected to i8 |
550 ; CHECK-NEXT: %trunc_des = trunc i32 %desired to i8 | 550 ; CHECK-NEXT: %trunc_des = trunc i32 %desired to i8 |
551 ; CHECK-NEXT: %old = call i8 @llvm.nacl.atomic.cmpxchg.i8(i32 %iptr, i8 %trunc
_exp, i8 %trunc_des, i32 6, i32 6) | 551 ; CHECK-NEXT: %old = call i8 @llvm.nacl.atomic.cmpxchg.i8(i32 %iptr, i8 %trunc
_exp, i8 %trunc_des, i32 6, i32 6) |
552 ; CHECK-NEXT: %old_ext = zext i8 %old to i32 | 552 ; CHECK-NEXT: %old_ext = zext i8 %old to i32 |
553 ; CHECK-NEXT: ret i32 %old_ext | 553 ; CHECK-NEXT: ret i32 %old_ext |
554 ; CHECK-NEXT: } | 554 ; CHECK-NEXT: } |
555 | 555 |
556 define i32 @test_atomic_cmpxchg_16(i32 %iptr, i32 %expected, i32 %desired) { | 556 define internal i32 @test_atomic_cmpxchg_16(i32 %iptr, i32 %expected, i32 %desir
ed) { |
557 entry: | 557 entry: |
558 %trunc_exp = trunc i32 %expected to i16 | 558 %trunc_exp = trunc i32 %expected to i16 |
559 %trunc_des = trunc i32 %desired to i16 | 559 %trunc_des = trunc i32 %desired to i16 |
560 %ptr = inttoptr i32 %iptr to i16* | 560 %ptr = inttoptr i32 %iptr to i16* |
561 %old = call i16 @llvm.nacl.atomic.cmpxchg.i16(i16* %ptr, i16 %trunc_exp, | 561 %old = call i16 @llvm.nacl.atomic.cmpxchg.i16(i16* %ptr, i16 %trunc_exp, |
562 i16 %trunc_des, i32 6, i32 6) | 562 i16 %trunc_des, i32 6, i32 6) |
563 %old_ext = zext i16 %old to i32 | 563 %old_ext = zext i16 %old to i32 |
564 ret i32 %old_ext | 564 ret i32 %old_ext |
565 } | 565 } |
566 | 566 |
567 ; CHECK-NEXT: define i32 @test_atomic_cmpxchg_16(i32 %iptr, i32 %expected, i32 %
desired) { | 567 ; CHECK-NEXT: define internal i32 @test_atomic_cmpxchg_16(i32 %iptr, i32 %expect
ed, i32 %desired) { |
568 ; CHECK-NEXT: entry: | 568 ; CHECK-NEXT: entry: |
569 ; CHECK-NEXT: %trunc_exp = trunc i32 %expected to i16 | 569 ; CHECK-NEXT: %trunc_exp = trunc i32 %expected to i16 |
570 ; CHECK-NEXT: %trunc_des = trunc i32 %desired to i16 | 570 ; CHECK-NEXT: %trunc_des = trunc i32 %desired to i16 |
571 ; CHECK-NEXT: %old = call i16 @llvm.nacl.atomic.cmpxchg.i16(i32 %iptr, i16 %tr
unc_exp, i16 %trunc_des, i32 6, i32 6) | 571 ; CHECK-NEXT: %old = call i16 @llvm.nacl.atomic.cmpxchg.i16(i32 %iptr, i16 %tr
unc_exp, i16 %trunc_des, i32 6, i32 6) |
572 ; CHECK-NEXT: %old_ext = zext i16 %old to i32 | 572 ; CHECK-NEXT: %old_ext = zext i16 %old to i32 |
573 ; CHECK-NEXT: ret i32 %old_ext | 573 ; CHECK-NEXT: ret i32 %old_ext |
574 ; CHECK-NEXT: } | 574 ; CHECK-NEXT: } |
575 | 575 |
576 define i32 @test_atomic_cmpxchg_32(i32 %iptr, i32 %expected, i32 %desired) { | 576 define internal i32 @test_atomic_cmpxchg_32(i32 %iptr, i32 %expected, i32 %desir
ed) { |
577 entry: | 577 entry: |
578 %ptr = inttoptr i32 %iptr to i32* | 578 %ptr = inttoptr i32 %iptr to i32* |
579 %old = call i32 @llvm.nacl.atomic.cmpxchg.i32(i32* %ptr, i32 %expected, | 579 %old = call i32 @llvm.nacl.atomic.cmpxchg.i32(i32* %ptr, i32 %expected, |
580 i32 %desired, i32 6, i32 6) | 580 i32 %desired, i32 6, i32 6) |
581 ret i32 %old | 581 ret i32 %old |
582 } | 582 } |
583 | 583 |
584 ; CHECK-NEXT: define i32 @test_atomic_cmpxchg_32(i32 %iptr, i32 %expected, i32 %
desired) { | 584 ; CHECK-NEXT: define internal i32 @test_atomic_cmpxchg_32(i32 %iptr, i32 %expect
ed, i32 %desired) { |
585 ; CHECK-NEXT: entry: | 585 ; CHECK-NEXT: entry: |
586 ; CHECK-NEXT: %old = call i32 @llvm.nacl.atomic.cmpxchg.i32(i32 %iptr, i32 %ex
pected, i32 %desired, i32 6, i32 6) | 586 ; CHECK-NEXT: %old = call i32 @llvm.nacl.atomic.cmpxchg.i32(i32 %iptr, i32 %ex
pected, i32 %desired, i32 6, i32 6) |
587 ; CHECK-NEXT: ret i32 %old | 587 ; CHECK-NEXT: ret i32 %old |
588 ; CHECK-NEXT: } | 588 ; CHECK-NEXT: } |
589 | 589 |
590 define i64 @test_atomic_cmpxchg_64(i32 %iptr, i64 %expected, i64 %desired) { | 590 define internal i64 @test_atomic_cmpxchg_64(i32 %iptr, i64 %expected, i64 %desir
ed) { |
591 entry: | 591 entry: |
592 %ptr = inttoptr i32 %iptr to i64* | 592 %ptr = inttoptr i32 %iptr to i64* |
593 %old = call i64 @llvm.nacl.atomic.cmpxchg.i64(i64* %ptr, i64 %expected, | 593 %old = call i64 @llvm.nacl.atomic.cmpxchg.i64(i64* %ptr, i64 %expected, |
594 i64 %desired, i32 6, i32 6) | 594 i64 %desired, i32 6, i32 6) |
595 ret i64 %old | 595 ret i64 %old |
596 } | 596 } |
597 | 597 |
598 ; CHECK-NEXT: define i64 @test_atomic_cmpxchg_64(i32 %iptr, i64 %expected, i64 %
desired) { | 598 ; CHECK-NEXT: define internal i64 @test_atomic_cmpxchg_64(i32 %iptr, i64 %expect
ed, i64 %desired) { |
599 ; CHECK-NEXT: entry: | 599 ; CHECK-NEXT: entry: |
600 ; CHECK-NEXT: %old = call i64 @llvm.nacl.atomic.cmpxchg.i64(i32 %iptr, i64 %ex
pected, i64 %desired, i32 6, i32 6) | 600 ; CHECK-NEXT: %old = call i64 @llvm.nacl.atomic.cmpxchg.i64(i32 %iptr, i64 %ex
pected, i64 %desired, i32 6, i32 6) |
601 ; CHECK-NEXT: ret i64 %old | 601 ; CHECK-NEXT: ret i64 %old |
602 ; CHECK-NEXT: } | 602 ; CHECK-NEXT: } |
603 | 603 |
604 ;;;; Fence and is-lock-free. | 604 ;;;; Fence and is-lock-free. |
605 | 605 |
606 define void @test_atomic_fence() { | 606 define internal void @test_atomic_fence() { |
607 entry: | 607 entry: |
608 call void @llvm.nacl.atomic.fence(i32 6) | 608 call void @llvm.nacl.atomic.fence(i32 6) |
609 ret void | 609 ret void |
610 } | 610 } |
611 | 611 |
612 ; CHECK-NEXT: define void @test_atomic_fence() { | 612 ; CHECK-NEXT: define internal void @test_atomic_fence() { |
613 ; CHECK-NEXT: entry: | 613 ; CHECK-NEXT: entry: |
614 ; CHECK-NEXT: call void @llvm.nacl.atomic.fence(i32 6) | 614 ; CHECK-NEXT: call void @llvm.nacl.atomic.fence(i32 6) |
615 ; CHECK-NEXT: ret void | 615 ; CHECK-NEXT: ret void |
616 ; CHECK-NEXT: } | 616 ; CHECK-NEXT: } |
617 | 617 |
618 define void @test_atomic_fence_all() { | 618 define internal void @test_atomic_fence_all() { |
619 entry: | 619 entry: |
620 call void @llvm.nacl.atomic.fence.all() | 620 call void @llvm.nacl.atomic.fence.all() |
621 ret void | 621 ret void |
622 } | 622 } |
623 | 623 |
624 ; CHECK-NEXT: define void @test_atomic_fence_all() { | 624 ; CHECK-NEXT: define internal void @test_atomic_fence_all() { |
625 ; CHECK-NEXT: entry: | 625 ; CHECK-NEXT: entry: |
626 ; CHECK-NEXT: call void @llvm.nacl.atomic.fence.all() | 626 ; CHECK-NEXT: call void @llvm.nacl.atomic.fence.all() |
627 ; CHECK-NEXT: ret void | 627 ; CHECK-NEXT: ret void |
628 ; CHECK-NEXT: } | 628 ; CHECK-NEXT: } |
629 | 629 |
630 define i32 @test_atomic_is_lock_free(i32 %iptr) { | 630 define internal i32 @test_atomic_is_lock_free(i32 %iptr) { |
631 entry: | 631 entry: |
632 %ptr = inttoptr i32 %iptr to i8* | 632 %ptr = inttoptr i32 %iptr to i8* |
633 %i = call i1 @llvm.nacl.atomic.is.lock.free(i32 4, i8* %ptr) | 633 %i = call i1 @llvm.nacl.atomic.is.lock.free(i32 4, i8* %ptr) |
634 %r = zext i1 %i to i32 | 634 %r = zext i1 %i to i32 |
635 ret i32 %r | 635 ret i32 %r |
636 } | 636 } |
637 | 637 |
638 ; CHECK-NEXT: define i32 @test_atomic_is_lock_free(i32 %iptr) { | 638 ; CHECK-NEXT: define internal i32 @test_atomic_is_lock_free(i32 %iptr) { |
639 ; CHECK-NEXT: entry: | 639 ; CHECK-NEXT: entry: |
640 ; CHECK-NEXT: %i = call i1 @llvm.nacl.atomic.is.lock.free(i32 4, i32 %iptr) | 640 ; CHECK-NEXT: %i = call i1 @llvm.nacl.atomic.is.lock.free(i32 4, i32 %iptr) |
641 ; CHECK-NEXT: %r = zext i1 %i to i32 | 641 ; CHECK-NEXT: %r = zext i1 %i to i32 |
642 ; CHECK-NEXT: ret i32 %r | 642 ; CHECK-NEXT: ret i32 %r |
643 ; CHECK-NEXT: } | 643 ; CHECK-NEXT: } |
644 | 644 |
645 ; NOIR: Total across all functions | 645 ; NOIR: Total across all functions |
OLD | NEW |