| OLD | NEW |
| 1 ; This tests each of the supported NaCl atomic instructions for every | 1 ; This tests each of the supported NaCl atomic instructions for every |
| 2 ; size allowed. | 2 ; size allowed. |
| 3 | 3 |
| 4 ; RUN: %p2i -i %s --args -O2 --verbose none \ | 4 ; RUN: %p2i -i %s --assemble --disassemble --args -O2 --verbose none \ |
| 5 ; RUN: | llvm-mc -triple=i686-none-nacl -filetype=obj \ | |
| 6 ; RUN: | llvm-objdump -d --symbolize -x86-asm-syntax=intel - \ | |
| 7 ; RUN: | FileCheck %s | 5 ; RUN: | FileCheck %s |
| 8 ; RUN: %p2i -i %s --args -O2 --verbose none \ | 6 ; RUN: %p2i -i %s --assemble --disassemble --args -O2 --verbose none \ |
| 9 ; RUN: | llvm-mc -triple=i686-none-nacl -filetype=obj \ | |
| 10 ; RUN: | llvm-objdump -d --symbolize -x86-asm-syntax=intel - \ | |
| 11 ; RUN: | FileCheck --check-prefix=CHECKO2 %s | 7 ; RUN: | FileCheck --check-prefix=CHECKO2 %s |
| 12 ; RUN: %p2i -i %s --args -Om1 --verbose none \ | 8 ; RUN: %p2i -i %s --assemble --disassemble --args -Om1 --verbose none \ |
| 13 ; RUN: | llvm-mc -triple=i686-none-nacl -filetype=obj \ | |
| 14 ; RUN: | llvm-objdump -d --symbolize -x86-asm-syntax=intel - \ | |
| 15 ; RUN: | FileCheck %s | 9 ; RUN: | FileCheck %s |
| 16 | 10 |
| 17 declare i8 @llvm.nacl.atomic.load.i8(i8*, i32) | 11 declare i8 @llvm.nacl.atomic.load.i8(i8*, i32) |
| 18 declare i16 @llvm.nacl.atomic.load.i16(i16*, i32) | 12 declare i16 @llvm.nacl.atomic.load.i16(i16*, i32) |
| 19 declare i32 @llvm.nacl.atomic.load.i32(i32*, i32) | 13 declare i32 @llvm.nacl.atomic.load.i32(i32*, i32) |
| 20 declare i64 @llvm.nacl.atomic.load.i64(i64*, i32) | 14 declare i64 @llvm.nacl.atomic.load.i64(i64*, i32) |
| 21 declare void @llvm.nacl.atomic.store.i8(i8, i8*, i32) | 15 declare void @llvm.nacl.atomic.store.i8(i8, i8*, i32) |
| 22 declare void @llvm.nacl.atomic.store.i16(i16, i16*, i32) | 16 declare void @llvm.nacl.atomic.store.i16(i16, i16*, i32) |
| 23 declare void @llvm.nacl.atomic.store.i32(i32, i32*, i32) | 17 declare void @llvm.nacl.atomic.store.i32(i32, i32*, i32) |
| 24 declare void @llvm.nacl.atomic.store.i64(i64, i64*, i32) | 18 declare void @llvm.nacl.atomic.store.i64(i64, i64*, i32) |
| (...skipping 21 matching lines...) Expand all Loading... |
| 46 | 40 |
| 47 define i32 @test_atomic_load_8(i32 %iptr) { | 41 define i32 @test_atomic_load_8(i32 %iptr) { |
| 48 entry: | 42 entry: |
| 49 %ptr = inttoptr i32 %iptr to i8* | 43 %ptr = inttoptr i32 %iptr to i8* |
| 50 ; parameter value "6" is for the sequential consistency memory order. | 44 ; parameter value "6" is for the sequential consistency memory order. |
| 51 %i = call i8 @llvm.nacl.atomic.load.i8(i8* %ptr, i32 6) | 45 %i = call i8 @llvm.nacl.atomic.load.i8(i8* %ptr, i32 6) |
| 52 %r = zext i8 %i to i32 | 46 %r = zext i8 %i to i32 |
| 53 ret i32 %r | 47 ret i32 %r |
| 54 } | 48 } |
| 55 ; CHECK-LABEL: test_atomic_load_8 | 49 ; CHECK-LABEL: test_atomic_load_8 |
| 56 ; CHECK: mov {{.*}}, dword | 50 ; CHECK: mov {{.*}},DWORD |
| 57 ; CHECK: mov {{.*}}, byte | 51 ; CHECK: mov {{.*}},byte |
| 58 | 52 |
| 59 define i32 @test_atomic_load_16(i32 %iptr) { | 53 define i32 @test_atomic_load_16(i32 %iptr) { |
| 60 entry: | 54 entry: |
| 61 %ptr = inttoptr i32 %iptr to i16* | 55 %ptr = inttoptr i32 %iptr to i16* |
| 62 %i = call i16 @llvm.nacl.atomic.load.i16(i16* %ptr, i32 6) | 56 %i = call i16 @llvm.nacl.atomic.load.i16(i16* %ptr, i32 6) |
| 63 %r = zext i16 %i to i32 | 57 %r = zext i16 %i to i32 |
| 64 ret i32 %r | 58 ret i32 %r |
| 65 } | 59 } |
| 66 ; CHECK-LABEL: test_atomic_load_16 | 60 ; CHECK-LABEL: test_atomic_load_16 |
| 67 ; CHECK: mov {{.*}}, dword | 61 ; CHECK: mov {{.*}},DWORD |
| 68 ; CHECK: mov {{.*}}, word | 62 ; CHECK: mov {{.*}},word |
| 69 | 63 |
| 70 define i32 @test_atomic_load_32(i32 %iptr) { | 64 define i32 @test_atomic_load_32(i32 %iptr) { |
| 71 entry: | 65 entry: |
| 72 %ptr = inttoptr i32 %iptr to i32* | 66 %ptr = inttoptr i32 %iptr to i32* |
| 73 %r = call i32 @llvm.nacl.atomic.load.i32(i32* %ptr, i32 6) | 67 %r = call i32 @llvm.nacl.atomic.load.i32(i32* %ptr, i32 6) |
| 74 ret i32 %r | 68 ret i32 %r |
| 75 } | 69 } |
| 76 ; CHECK-LABEL: test_atomic_load_32 | 70 ; CHECK-LABEL: test_atomic_load_32 |
| 77 ; CHECK: mov {{.*}}, dword | 71 ; CHECK: mov {{.*}},DWORD |
| 78 ; CHECK: mov {{.*}}, dword | 72 ; CHECK: mov {{.*}},DWORD |
| 79 | 73 |
| 80 define i64 @test_atomic_load_64(i32 %iptr) { | 74 define i64 @test_atomic_load_64(i32 %iptr) { |
| 81 entry: | 75 entry: |
| 82 %ptr = inttoptr i32 %iptr to i64* | 76 %ptr = inttoptr i32 %iptr to i64* |
| 83 %r = call i64 @llvm.nacl.atomic.load.i64(i64* %ptr, i32 6) | 77 %r = call i64 @llvm.nacl.atomic.load.i64(i64* %ptr, i32 6) |
| 84 ret i64 %r | 78 ret i64 %r |
| 85 } | 79 } |
| 86 ; CHECK-LABEL: test_atomic_load_64 | 80 ; CHECK-LABEL: test_atomic_load_64 |
| 87 ; CHECK: movq x{{.*}}, qword | 81 ; CHECK: movq x{{.*}},QWORD |
| 88 ; CHECK: movq qword {{.*}}, x{{.*}} | 82 ; CHECK: movq QWORD {{.*}},x{{.*}} |
| 89 | 83 |
| 90 define i32 @test_atomic_load_32_with_arith(i32 %iptr) { | 84 define i32 @test_atomic_load_32_with_arith(i32 %iptr) { |
| 91 entry: | 85 entry: |
| 92 br label %next | 86 br label %next |
| 93 | 87 |
| 94 next: | 88 next: |
| 95 %ptr = inttoptr i32 %iptr to i32* | 89 %ptr = inttoptr i32 %iptr to i32* |
| 96 %r = call i32 @llvm.nacl.atomic.load.i32(i32* %ptr, i32 6) | 90 %r = call i32 @llvm.nacl.atomic.load.i32(i32* %ptr, i32 6) |
| 97 %r2 = add i32 %r, 32 | 91 %r2 = add i32 %r, 32 |
| 98 ret i32 %r2 | 92 ret i32 %r2 |
| 99 } | 93 } |
| 100 ; CHECK-LABEL: test_atomic_load_32_with_arith | 94 ; CHECK-LABEL: test_atomic_load_32_with_arith |
| 101 ; CHECK: mov {{.*}}, dword | 95 ; CHECK: mov {{.*}},DWORD |
| 102 ; The next instruction may be a separate load or folded into an add. | 96 ; The next instruction may be a separate load or folded into an add. |
| 103 ; | 97 ; |
| 104 ; In O2 mode, we know that the load and add are going to be fused. | 98 ; In O2 mode, we know that the load and add are going to be fused. |
| 105 ; CHECKO2-LABEL: test_atomic_load_32_with_arith | 99 ; CHECKO2-LABEL: test_atomic_load_32_with_arith |
| 106 ; CHECKO2: mov {{.*}}, dword | 100 ; CHECKO2: mov {{.*}}, DWORD |
| 107 ; CHECKO2: add {{.*}}, dword | 101 ; CHECKO2: add {{.*}}, DWORD |
| 108 | 102 |
| 109 define i32 @test_atomic_load_32_ignored(i32 %iptr) { | 103 define i32 @test_atomic_load_32_ignored(i32 %iptr) { |
| 110 entry: | 104 entry: |
| 111 %ptr = inttoptr i32 %iptr to i32* | 105 %ptr = inttoptr i32 %iptr to i32* |
| 112 %ignored = call i32 @llvm.nacl.atomic.load.i32(i32* %ptr, i32 6) | 106 %ignored = call i32 @llvm.nacl.atomic.load.i32(i32* %ptr, i32 6) |
| 113 ret i32 0 | 107 ret i32 0 |
| 114 } | 108 } |
| 115 ; CHECK-LABEL: test_atomic_load_32_ignored | 109 ; CHECK-LABEL: test_atomic_load_32_ignored |
| 116 ; CHECK: mov {{.*}}, dword | 110 ; CHECK: mov {{.*}},DWORD |
| 117 ; CHECK: mov {{.*}}, dword | 111 ; CHECK: mov {{.*}},DWORD |
| 118 ; CHECKO2-LABEL: test_atomic_load_32_ignored | 112 ; CHECKO2-LABEL: test_atomic_load_32_ignored |
| 119 ; CHECKO2: mov {{.*}}, dword | 113 ; CHECKO2: mov {{.*}}, DWORD |
| 120 ; CHECKO2: mov {{.*}}, dword | 114 ; CHECKO2: mov {{.*}}, DWORD |
| 121 | 115 |
| 122 define i64 @test_atomic_load_64_ignored(i32 %iptr) { | 116 define i64 @test_atomic_load_64_ignored(i32 %iptr) { |
| 123 entry: | 117 entry: |
| 124 %ptr = inttoptr i32 %iptr to i64* | 118 %ptr = inttoptr i32 %iptr to i64* |
| 125 %ignored = call i64 @llvm.nacl.atomic.load.i64(i64* %ptr, i32 6) | 119 %ignored = call i64 @llvm.nacl.atomic.load.i64(i64* %ptr, i32 6) |
| 126 ret i64 0 | 120 ret i64 0 |
| 127 } | 121 } |
| 128 ; CHECK-LABEL: test_atomic_load_64_ignored | 122 ; CHECK-LABEL: test_atomic_load_64_ignored |
| 129 ; CHECK: movq x{{.*}}, qword | 123 ; CHECK: movq x{{.*}},QWORD |
| 130 ; CHECK: movq qword {{.*}}, x{{.*}} | 124 ; CHECK: movq QWORD {{.*}},x{{.*}} |
| 131 | 125 |
| 132 ;;; Store | 126 ;;; Store |
| 133 | 127 |
| 134 define void @test_atomic_store_8(i32 %iptr, i32 %v) { | 128 define void @test_atomic_store_8(i32 %iptr, i32 %v) { |
| 135 entry: | 129 entry: |
| 136 %truncv = trunc i32 %v to i8 | 130 %truncv = trunc i32 %v to i8 |
| 137 %ptr = inttoptr i32 %iptr to i8* | 131 %ptr = inttoptr i32 %iptr to i8* |
| 138 call void @llvm.nacl.atomic.store.i8(i8 %truncv, i8* %ptr, i32 6) | 132 call void @llvm.nacl.atomic.store.i8(i8 %truncv, i8* %ptr, i32 6) |
| 139 ret void | 133 ret void |
| 140 } | 134 } |
| (...skipping 12 matching lines...) Expand all Loading... |
| 153 ; CHECK: mov word | 147 ; CHECK: mov word |
| 154 ; CHECK: mfence | 148 ; CHECK: mfence |
| 155 | 149 |
| 156 define void @test_atomic_store_32(i32 %iptr, i32 %v) { | 150 define void @test_atomic_store_32(i32 %iptr, i32 %v) { |
| 157 entry: | 151 entry: |
| 158 %ptr = inttoptr i32 %iptr to i32* | 152 %ptr = inttoptr i32 %iptr to i32* |
| 159 call void @llvm.nacl.atomic.store.i32(i32 %v, i32* %ptr, i32 6) | 153 call void @llvm.nacl.atomic.store.i32(i32 %v, i32* %ptr, i32 6) |
| 160 ret void | 154 ret void |
| 161 } | 155 } |
| 162 ; CHECK-LABEL: test_atomic_store_32 | 156 ; CHECK-LABEL: test_atomic_store_32 |
| 163 ; CHECK: mov dword | 157 ; CHECK: mov DWORD |
| 164 ; CHECK: mfence | 158 ; CHECK: mfence |
| 165 | 159 |
| 166 define void @test_atomic_store_64(i32 %iptr, i64 %v) { | 160 define void @test_atomic_store_64(i32 %iptr, i64 %v) { |
| 167 entry: | 161 entry: |
| 168 %ptr = inttoptr i32 %iptr to i64* | 162 %ptr = inttoptr i32 %iptr to i64* |
| 169 call void @llvm.nacl.atomic.store.i64(i64 %v, i64* %ptr, i32 6) | 163 call void @llvm.nacl.atomic.store.i64(i64 %v, i64* %ptr, i32 6) |
| 170 ret void | 164 ret void |
| 171 } | 165 } |
| 172 ; CHECK-LABEL: test_atomic_store_64 | 166 ; CHECK-LABEL: test_atomic_store_64 |
| 173 ; CHECK: movq x{{.*}}, qword | 167 ; CHECK: movq x{{.*}},QWORD |
| 174 ; CHECK: movq qword {{.*}}, x{{.*}} | 168 ; CHECK: movq QWORD {{.*}},x{{.*}} |
| 175 ; CHECK: mfence | 169 ; CHECK: mfence |
| 176 | 170 |
| 177 define void @test_atomic_store_64_const(i32 %iptr) { | 171 define void @test_atomic_store_64_const(i32 %iptr) { |
| 178 entry: | 172 entry: |
| 179 %ptr = inttoptr i32 %iptr to i64* | 173 %ptr = inttoptr i32 %iptr to i64* |
| 180 call void @llvm.nacl.atomic.store.i64(i64 12345678901234, i64* %ptr, i32 6) | 174 call void @llvm.nacl.atomic.store.i64(i64 12345678901234, i64* %ptr, i32 6) |
| 181 ret void | 175 ret void |
| 182 } | 176 } |
| 183 ; CHECK-LABEL: test_atomic_store_64_const | 177 ; CHECK-LABEL: test_atomic_store_64_const |
| 184 ; CHECK: mov {{.*}}, 1942892530 | 178 ; CHECK: mov {{.*}},1942892530 |
| 185 ; CHECK: mov {{.*}}, 2874 | 179 ; CHECK: mov {{.*}},2874 |
| 186 ; CHECK: movq x{{.*}}, qword | 180 ; CHECK: movq x{{.*}},QWORD |
| 187 ; CHECK: movq qword {{.*}}, x{{.*}} | 181 ; CHECK: movq QWORD {{.*}},x{{.*}} |
| 188 ; CHECK: mfence | 182 ; CHECK: mfence |
| 189 | 183 |
| 190 | 184 |
| 191 ;;; RMW | 185 ;;; RMW |
| 192 | 186 |
| 193 ;; add | 187 ;; add |
| 194 | 188 |
| 195 define i32 @test_atomic_rmw_add_8(i32 %iptr, i32 %v) { | 189 define i32 @test_atomic_rmw_add_8(i32 %iptr, i32 %v) { |
| 196 entry: | 190 entry: |
| 197 %trunc = trunc i32 %v to i8 | 191 %trunc = trunc i32 %v to i8 |
| 198 %ptr = inttoptr i32 %iptr to i8* | 192 %ptr = inttoptr i32 %iptr to i8* |
| 199 ; "1" is an atomic add, and "6" is sequential consistency. | 193 ; "1" is an atomic add, and "6" is sequential consistency. |
| 200 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 1, i8* %ptr, i8 %trunc, i32 6) | 194 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 1, i8* %ptr, i8 %trunc, i32 6) |
| 201 %a_ext = zext i8 %a to i32 | 195 %a_ext = zext i8 %a to i32 |
| 202 ret i32 %a_ext | 196 ret i32 %a_ext |
| 203 } | 197 } |
| 204 ; CHECK-LABEL: test_atomic_rmw_add_8 | 198 ; CHECK-LABEL: test_atomic_rmw_add_8 |
| 205 ; CHECK: lock | 199 ; CHECK: lock |
| 206 ; CHECK-NEXT: xadd byte {{.*}}, [[REG:.*]] | 200 ; CHECK-NEXT: xadd byte {{.*}}, [[REG:.*]] |
| 207 ; CHECK: {{mov|movzx}} {{.*}}, [[REG]] | 201 ; CHECK: {{mov|movzx}} {{.*}},[[REG]] |
| 208 | 202 |
| 209 define i32 @test_atomic_rmw_add_16(i32 %iptr, i32 %v) { | 203 define i32 @test_atomic_rmw_add_16(i32 %iptr, i32 %v) { |
| 210 entry: | 204 entry: |
| 211 %trunc = trunc i32 %v to i16 | 205 %trunc = trunc i32 %v to i16 |
| 212 %ptr = inttoptr i32 %iptr to i16* | 206 %ptr = inttoptr i32 %iptr to i16* |
| 213 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 1, i16* %ptr, i16 %trunc, i32 6) | 207 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 1, i16* %ptr, i16 %trunc, i32 6) |
| 214 %a_ext = zext i16 %a to i32 | 208 %a_ext = zext i16 %a to i32 |
| 215 ret i32 %a_ext | 209 ret i32 %a_ext |
| 216 } | 210 } |
| 217 ; CHECK-LABEL: test_atomic_rmw_add_16 | 211 ; CHECK-LABEL: test_atomic_rmw_add_16 |
| 218 ; CHECK: lock | 212 ; CHECK: lock |
| 219 ; CHECK-NEXT: xadd word {{.*}}, [[REG:.*]] | 213 ; CHECK-NEXT: xadd word {{.*}}, [[REG:.*]] |
| 220 ; CHECK: {{mov|movzx}} {{.*}}, [[REG]] | 214 ; CHECK: {{mov|movzx}} {{.*}},[[REG]] |
| 221 | 215 |
| 222 define i32 @test_atomic_rmw_add_32(i32 %iptr, i32 %v) { | 216 define i32 @test_atomic_rmw_add_32(i32 %iptr, i32 %v) { |
| 223 entry: | 217 entry: |
| 224 %ptr = inttoptr i32 %iptr to i32* | 218 %ptr = inttoptr i32 %iptr to i32* |
| 225 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 1, i32* %ptr, i32 %v, i32 6) | 219 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 1, i32* %ptr, i32 %v, i32 6) |
| 226 ret i32 %a | 220 ret i32 %a |
| 227 } | 221 } |
| 228 ; CHECK-LABEL: test_atomic_rmw_add_32 | 222 ; CHECK-LABEL: test_atomic_rmw_add_32 |
| 229 ; CHECK: lock | 223 ; CHECK: lock |
| 230 ; CHECK-NEXT: xadd dword {{.*}}, [[REG:.*]] | 224 ; CHECK-NEXT: xadd DWORD {{.*}}, [[REG:.*]] |
| 231 ; CHECK: mov {{.*}}, [[REG]] | 225 ; CHECK: mov {{.*}},[[REG]] |
| 232 | 226 |
| 233 define i64 @test_atomic_rmw_add_64(i32 %iptr, i64 %v) { | 227 define i64 @test_atomic_rmw_add_64(i32 %iptr, i64 %v) { |
| 234 entry: | 228 entry: |
| 235 %ptr = inttoptr i32 %iptr to i64* | 229 %ptr = inttoptr i32 %iptr to i64* |
| 236 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 1, i64* %ptr, i64 %v, i32 6) | 230 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 1, i64* %ptr, i64 %v, i32 6) |
| 237 ret i64 %a | 231 ret i64 %a |
| 238 } | 232 } |
| 239 ; CHECK-LABEL: test_atomic_rmw_add_64 | 233 ; CHECK-LABEL: test_atomic_rmw_add_64 |
| 240 ; CHECK: push ebx | 234 ; CHECK: push ebx |
| 241 ; CHECK: mov eax, dword ptr [{{.*}}] | 235 ; CHECK: mov eax,DWORD ptr [{{.*}}] |
| 242 ; CHECK: mov edx, dword ptr [{{.*}} + 4] | 236 ; CHECK: mov edx,DWORD ptr [{{.*}}+4] |
| 243 ; CHECK: mov ebx, eax | 237 ; CHECK: mov ebx,eax |
| 244 ; RHS of add cannot be any of the e[abcd]x regs because they are | 238 ; RHS of add cannot be any of the e[abcd]x regs because they are |
| 245 ; clobbered in the loop, and the RHS needs to be remain live. | 239 ; clobbered in the loop, and the RHS needs to be remain live. |
| 246 ; CHECK: add ebx, {{.*e.[^x]}} | 240 ; CHECK: add ebx, {{.*e.[^x]}} |
| 247 ; CHECK: mov ecx, edx | 241 ; CHECK: mov ecx,edx |
| 248 ; CHECK: adc ecx, {{.*e.[^x]}} | 242 ; CHECK: adc ecx, {{.*e.[^x]}} |
| 249 ; Ptr cannot be eax, ebx, ecx, or edx (used up for the expected and desired). | 243 ; Ptr cannot be eax, ebx, ecx, or edx (used up for the expected and desired). |
| 250 ; It can be esi, edi, or ebp though, for example (so we need to be careful | 244 ; It can be esi, edi, or ebp though, for example (so we need to be careful |
| 251 ; about rejecting eb* and ed*.) | 245 ; about rejecting eb* and ed*.) |
| 252 ; CHECK: lock | 246 ; CHECK: lock |
| 253 ; CHECK-NEXT: cmpxchg8b qword ptr [e{{.[^x]}}] | 247 ; CHECK-NEXT: cmpxchg8b QWORD PTR [e{{.[^x]}}] |
| 254 ; CHECK: jne -{{[0-9]}} | 248 ; CHECK: jne-{{[0-9]}} |
| 255 | 249 |
| 256 ; Test with some more register pressure. When we have an alloca, ebp is | 250 ; Test with some more register pressure. When we have an alloca, ebp is |
| 257 ; used to manage the stack frame, so it cannot be used as a register either. | 251 ; used to manage the stack frame, so it cannot be used as a register either. |
| 258 define void @use_ptr(i32 %iptr) { | 252 define void @use_ptr(i32 %iptr) { |
| 259 entry: | 253 entry: |
| 260 ret void | 254 ret void |
| 261 } | 255 } |
| 262 | 256 |
| 263 define i64 @test_atomic_rmw_add_64_alloca(i32 %iptr, i64 %v) { | 257 define i64 @test_atomic_rmw_add_64_alloca(i32 %iptr, i64 %v) { |
| 264 entry: | 258 entry: |
| (...skipping 13 matching lines...) Expand all Loading... |
| 278 ; CHECK-DAG: mov edx | 272 ; CHECK-DAG: mov edx |
| 279 ; CHECK-DAG: mov eax | 273 ; CHECK-DAG: mov eax |
| 280 ; CHECK-DAG: mov ecx | 274 ; CHECK-DAG: mov ecx |
| 281 ; CHECK-DAG: mov ebx | 275 ; CHECK-DAG: mov ebx |
| 282 ; Ptr cannot be eax, ebx, ecx, or edx (used up for the expected and desired). | 276 ; Ptr cannot be eax, ebx, ecx, or edx (used up for the expected and desired). |
| 283 ; It also cannot be ebp since we use that for alloca. Also make sure it's | 277 ; It also cannot be ebp since we use that for alloca. Also make sure it's |
| 284 ; not esp, since that's the stack pointer and mucking with it will break | 278 ; not esp, since that's the stack pointer and mucking with it will break |
| 285 ; the later use_ptr function call. | 279 ; the later use_ptr function call. |
| 286 ; That pretty much leaves esi, or edi as the only viable registers. | 280 ; That pretty much leaves esi, or edi as the only viable registers. |
| 287 ; CHECK: lock | 281 ; CHECK: lock |
| 288 ; CHECK-NEXT: cmpxchg8b qword ptr [e{{[ds]}}i] | 282 ; CHECK-NEXT: cmpxchg8b QWORD PTR [e{{[ds]}}i] |
| 289 ; CHECK: call use_ptr | 283 ; CHECK: call |
| 284 ; CHECK-NEXT: R_{{.*}} use_ptr |
| 290 | 285 |
| 291 define i32 @test_atomic_rmw_add_32_ignored(i32 %iptr, i32 %v) { | 286 define i32 @test_atomic_rmw_add_32_ignored(i32 %iptr, i32 %v) { |
| 292 entry: | 287 entry: |
| 293 %ptr = inttoptr i32 %iptr to i32* | 288 %ptr = inttoptr i32 %iptr to i32* |
| 294 %ignored = call i32 @llvm.nacl.atomic.rmw.i32(i32 1, i32* %ptr, i32 %v, i32 6) | 289 %ignored = call i32 @llvm.nacl.atomic.rmw.i32(i32 1, i32* %ptr, i32 %v, i32 6) |
| 295 ret i32 %v | 290 ret i32 %v |
| 296 } | 291 } |
| 297 ; Technically this could use "lock add" instead of "lock xadd", if liveness | 292 ; Technically this could use "lock add" instead of "lock xadd", if liveness |
| 298 ; tells us that the destination variable is dead. | 293 ; tells us that the destination variable is dead. |
| 299 ; CHECK-LABEL: test_atomic_rmw_add_32_ignored | 294 ; CHECK-LABEL: test_atomic_rmw_add_32_ignored |
| 300 ; CHECK: lock | 295 ; CHECK: lock |
| 301 ; CHECK-NEXT: xadd dword {{.*}}, [[REG:.*]] | 296 ; CHECK-NEXT: xadd DWORD {{.*}}, [[REG:.*]] |
| 302 | 297 |
| 303 ; Atomic RMW 64 needs to be expanded into its own loop. | 298 ; Atomic RMW 64 needs to be expanded into its own loop. |
| 304 ; Make sure that works w/ non-trivial function bodies. | 299 ; Make sure that works w/ non-trivial function bodies. |
| 305 define i64 @test_atomic_rmw_add_64_loop(i32 %iptr, i64 %v) { | 300 define i64 @test_atomic_rmw_add_64_loop(i32 %iptr, i64 %v) { |
| 306 entry: | 301 entry: |
| 307 %x = icmp ult i64 %v, 100 | 302 %x = icmp ult i64 %v, 100 |
| 308 br i1 %x, label %err, label %loop | 303 br i1 %x, label %err, label %loop |
| 309 | 304 |
| 310 loop: | 305 loop: |
| 311 %v_next = phi i64 [ %v, %entry ], [ %next, %loop ] | 306 %v_next = phi i64 [ %v, %entry ], [ %next, %loop ] |
| 312 %ptr = inttoptr i32 %iptr to i64* | 307 %ptr = inttoptr i32 %iptr to i64* |
| 313 %next = call i64 @llvm.nacl.atomic.rmw.i64(i32 1, i64* %ptr, i64 %v_next, i32
6) | 308 %next = call i64 @llvm.nacl.atomic.rmw.i64(i32 1, i64* %ptr, i64 %v_next, i32
6) |
| 314 %success = icmp eq i64 %next, 100 | 309 %success = icmp eq i64 %next, 100 |
| 315 br i1 %success, label %done, label %loop | 310 br i1 %success, label %done, label %loop |
| 316 | 311 |
| 317 done: | 312 done: |
| 318 ret i64 %next | 313 ret i64 %next |
| 319 | 314 |
| 320 err: | 315 err: |
| 321 ret i64 0 | 316 ret i64 0 |
| 322 } | 317 } |
| 323 ; CHECK-LABEL: test_atomic_rmw_add_64_loop | 318 ; CHECK-LABEL: test_atomic_rmw_add_64_loop |
| 324 ; CHECK: push ebx | 319 ; CHECK: push ebx |
| 325 ; CHECK: mov eax, dword ptr [{{.*}}] | 320 ; CHECK: mov eax,DWORD ptr [{{.*}}] |
| 326 ; CHECK: mov edx, dword ptr [{{.*}} + 4] | 321 ; CHECK: mov edx,DWORD ptr [{{.*}}+4] |
| 327 ; CHECK: mov ebx, eax | 322 ; CHECK: mov ebx,eax |
| 328 ; CHECK: add ebx, {{.*e.[^x]}} | 323 ; CHECK: add ebx, {{.*e.[^x]}} |
| 329 ; CHECK: mov ecx, edx | 324 ; CHECK: mov ecx,edx |
| 330 ; CHECK: adc ecx, {{.*e.[^x]}} | 325 ; CHECK: adc ecx, {{.*e.[^x]}} |
| 331 ; CHECK: lock | 326 ; CHECK: lock |
| 332 ; CHECK-NEXT: cmpxchg8b qword ptr [e{{.[^x]}}] | 327 ; CHECK-NEXT: cmpxchg8b QWORD PTR [e{{.[^x]}}] |
| 333 ; CHECK: jne -{{[0-9]}} | 328 ; CHECK: jne-{{[0-9]}} |
| 334 | 329 |
| 335 ;; sub | 330 ;; sub |
| 336 | 331 |
| 337 define i32 @test_atomic_rmw_sub_8(i32 %iptr, i32 %v) { | 332 define i32 @test_atomic_rmw_sub_8(i32 %iptr, i32 %v) { |
| 338 entry: | 333 entry: |
| 339 %trunc = trunc i32 %v to i8 | 334 %trunc = trunc i32 %v to i8 |
| 340 %ptr = inttoptr i32 %iptr to i8* | 335 %ptr = inttoptr i32 %iptr to i8* |
| 341 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 2, i8* %ptr, i8 %trunc, i32 6) | 336 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 2, i8* %ptr, i8 %trunc, i32 6) |
| 342 %a_ext = zext i8 %a to i32 | 337 %a_ext = zext i8 %a to i32 |
| 343 ret i32 %a_ext | 338 ret i32 %a_ext |
| 344 } | 339 } |
| 345 ; CHECK-LABEL: test_atomic_rmw_sub_8 | 340 ; CHECK-LABEL: test_atomic_rmw_sub_8 |
| 346 ; CHECK: neg [[REG:.*]] | 341 ; CHECK: neg [[REG:.*]] |
| 347 ; CHECK: lock | 342 ; CHECK: lock |
| 348 ; CHECK-NEXT: xadd byte {{.*}}, [[REG]] | 343 ; CHECK-NEXT: xadd byte {{.*}}, [[REG]] |
| 349 ; CHECK: {{mov|movzx}} {{.*}}, [[REG]] | 344 ; CHECK: {{mov|movzx}} {{.*}},[[REG]] |
| 350 | 345 |
| 351 define i32 @test_atomic_rmw_sub_16(i32 %iptr, i32 %v) { | 346 define i32 @test_atomic_rmw_sub_16(i32 %iptr, i32 %v) { |
| 352 entry: | 347 entry: |
| 353 %trunc = trunc i32 %v to i16 | 348 %trunc = trunc i32 %v to i16 |
| 354 %ptr = inttoptr i32 %iptr to i16* | 349 %ptr = inttoptr i32 %iptr to i16* |
| 355 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 2, i16* %ptr, i16 %trunc, i32 6) | 350 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 2, i16* %ptr, i16 %trunc, i32 6) |
| 356 %a_ext = zext i16 %a to i32 | 351 %a_ext = zext i16 %a to i32 |
| 357 ret i32 %a_ext | 352 ret i32 %a_ext |
| 358 } | 353 } |
| 359 ; CHECK-LABEL: test_atomic_rmw_sub_16 | 354 ; CHECK-LABEL: test_atomic_rmw_sub_16 |
| 360 ; CHECK: neg [[REG:.*]] | 355 ; CHECK: neg [[REG:.*]] |
| 361 ; CHECK: lock | 356 ; CHECK: lock |
| 362 ; CHECK-NEXT: xadd word {{.*}}, [[REG]] | 357 ; CHECK-NEXT: xadd word {{.*}}, [[REG]] |
| 363 ; CHECK: {{mov|movzx}} {{.*}}, [[REG]] | 358 ; CHECK: {{mov|movzx}} {{.*}},[[REG]] |
| 364 | 359 |
| 365 define i32 @test_atomic_rmw_sub_32(i32 %iptr, i32 %v) { | 360 define i32 @test_atomic_rmw_sub_32(i32 %iptr, i32 %v) { |
| 366 entry: | 361 entry: |
| 367 %ptr = inttoptr i32 %iptr to i32* | 362 %ptr = inttoptr i32 %iptr to i32* |
| 368 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 2, i32* %ptr, i32 %v, i32 6) | 363 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 2, i32* %ptr, i32 %v, i32 6) |
| 369 ret i32 %a | 364 ret i32 %a |
| 370 } | 365 } |
| 371 ; CHECK-LABEL: test_atomic_rmw_sub_32 | 366 ; CHECK-LABEL: test_atomic_rmw_sub_32 |
| 372 ; CHECK: neg [[REG:.*]] | 367 ; CHECK: neg [[REG:.*]] |
| 373 ; CHECK: lock | 368 ; CHECK: lock |
| 374 ; CHECK-NEXT: xadd dword {{.*}}, [[REG]] | 369 ; CHECK-NEXT: xadd DWORD {{.*}}, [[REG]] |
| 375 ; CHECK: mov {{.*}}, [[REG]] | 370 ; CHECK: mov {{.*}},[[REG]] |
| 376 | 371 |
| 377 define i64 @test_atomic_rmw_sub_64(i32 %iptr, i64 %v) { | 372 define i64 @test_atomic_rmw_sub_64(i32 %iptr, i64 %v) { |
| 378 entry: | 373 entry: |
| 379 %ptr = inttoptr i32 %iptr to i64* | 374 %ptr = inttoptr i32 %iptr to i64* |
| 380 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 2, i64* %ptr, i64 %v, i32 6) | 375 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 2, i64* %ptr, i64 %v, i32 6) |
| 381 ret i64 %a | 376 ret i64 %a |
| 382 } | 377 } |
| 383 ; CHECK-LABEL: test_atomic_rmw_sub_64 | 378 ; CHECK-LABEL: test_atomic_rmw_sub_64 |
| 384 ; CHECK: push ebx | 379 ; CHECK: push ebx |
| 385 ; CHECK: mov eax, dword ptr [{{.*}}] | 380 ; CHECK: mov eax,DWORD ptr [{{.*}}] |
| 386 ; CHECK: mov edx, dword ptr [{{.*}} + 4] | 381 ; CHECK: mov edx,DWORD ptr [{{.*}}+4] |
| 387 ; CHECK: mov ebx, eax | 382 ; CHECK: mov ebx,eax |
| 388 ; CHECK: sub ebx, {{.*e.[^x]}} | 383 ; CHECK: sub ebx, {{.*e.[^x]}} |
| 389 ; CHECK: mov ecx, edx | 384 ; CHECK: mov ecx,edx |
| 390 ; CHECK: sbb ecx, {{.*e.[^x]}} | 385 ; CHECK: sbb ecx, {{.*e.[^x]}} |
| 391 ; CHECK: lock | 386 ; CHECK: lock |
| 392 ; CHECK-NEXT: cmpxchg8b qword ptr [e{{.[^x]}}] | 387 ; CHECK-NEXT: cmpxchg8b QWORD PTR [e{{.[^x]}}] |
| 393 ; CHECK: jne -{{[0-9]}} | 388 ; CHECK: jne-{{[0-9]}} |
| 394 | 389 |
| 395 | 390 |
| 396 define i32 @test_atomic_rmw_sub_32_ignored(i32 %iptr, i32 %v) { | 391 define i32 @test_atomic_rmw_sub_32_ignored(i32 %iptr, i32 %v) { |
| 397 entry: | 392 entry: |
| 398 %ptr = inttoptr i32 %iptr to i32* | 393 %ptr = inttoptr i32 %iptr to i32* |
| 399 %ignored = call i32 @llvm.nacl.atomic.rmw.i32(i32 2, i32* %ptr, i32 %v, i32 6) | 394 %ignored = call i32 @llvm.nacl.atomic.rmw.i32(i32 2, i32* %ptr, i32 %v, i32 6) |
| 400 ret i32 %v | 395 ret i32 %v |
| 401 } | 396 } |
| 402 ; Could use "lock sub" instead of "neg; lock xadd" | 397 ; Could use "lock sub" instead of "neg; lock xadd" |
| 403 ; CHECK-LABEL: test_atomic_rmw_sub_32_ignored | 398 ; CHECK-LABEL: test_atomic_rmw_sub_32_ignored |
| 404 ; CHECK: neg [[REG:.*]] | 399 ; CHECK: neg [[REG:.*]] |
| 405 ; CHECK: lock | 400 ; CHECK: lock |
| 406 ; CHECK-NEXT: xadd dword {{.*}}, [[REG]] | 401 ; CHECK-NEXT: xadd DWORD {{.*}}, [[REG]] |
| 407 | 402 |
| 408 ;; or | 403 ;; or |
| 409 | 404 |
| 410 define i32 @test_atomic_rmw_or_8(i32 %iptr, i32 %v) { | 405 define i32 @test_atomic_rmw_or_8(i32 %iptr, i32 %v) { |
| 411 entry: | 406 entry: |
| 412 %trunc = trunc i32 %v to i8 | 407 %trunc = trunc i32 %v to i8 |
| 413 %ptr = inttoptr i32 %iptr to i8* | 408 %ptr = inttoptr i32 %iptr to i8* |
| 414 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 3, i8* %ptr, i8 %trunc, i32 6) | 409 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 3, i8* %ptr, i8 %trunc, i32 6) |
| 415 %a_ext = zext i8 %a to i32 | 410 %a_ext = zext i8 %a to i32 |
| 416 ret i32 %a_ext | 411 ret i32 %a_ext |
| 417 } | 412 } |
| 418 ; CHECK-LABEL: test_atomic_rmw_or_8 | 413 ; CHECK-LABEL: test_atomic_rmw_or_8 |
| 419 ; CHECK: mov al, byte ptr | 414 ; CHECK: mov al,BYTE PTR |
| 420 ; Dest cannot be eax here, because eax is used for the old value. Also want | 415 ; Dest cannot be eax here, because eax is used for the old value. Also want |
| 421 ; to make sure that cmpxchg's source is the same register. | 416 ; to make sure that cmpxchg's source is the same register. |
| 422 ; CHECK: or [[REG:[^a].]] | 417 ; CHECK: or [[REG:[^a].]] |
| 423 ; CHECK: lock | 418 ; CHECK: lock |
| 424 ; CHECK-NEXT: cmpxchg byte ptr [e{{[^a].}}], [[REG]] | 419 ; CHECK-NEXT: cmpxchg BYTE PTR [e{{[^a].}}], [[REG]] |
| 425 ; CHECK: jne -{{[0-9]}} | 420 ; CHECK: jne-{{[0-9]}} |
| 426 | 421 |
| 427 define i32 @test_atomic_rmw_or_16(i32 %iptr, i32 %v) { | 422 define i32 @test_atomic_rmw_or_16(i32 %iptr, i32 %v) { |
| 428 entry: | 423 entry: |
| 429 %trunc = trunc i32 %v to i16 | 424 %trunc = trunc i32 %v to i16 |
| 430 %ptr = inttoptr i32 %iptr to i16* | 425 %ptr = inttoptr i32 %iptr to i16* |
| 431 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 3, i16* %ptr, i16 %trunc, i32 6) | 426 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 3, i16* %ptr, i16 %trunc, i32 6) |
| 432 %a_ext = zext i16 %a to i32 | 427 %a_ext = zext i16 %a to i32 |
| 433 ret i32 %a_ext | 428 ret i32 %a_ext |
| 434 } | 429 } |
| 435 ; CHECK-LABEL: test_atomic_rmw_or_16 | 430 ; CHECK-LABEL: test_atomic_rmw_or_16 |
| 436 ; CHECK: mov ax, word ptr | 431 ; CHECK: mov ax,word ptr |
| 437 ; CHECK: or [[REG:[^a].]] | 432 ; CHECK: or [[REG:[^a].]] |
| 438 ; CHECK: lock | 433 ; CHECK: lock |
| 439 ; CHECK-NEXT: cmpxchg word ptr [e{{[^a].}}], [[REG]] | 434 ; CHECK-NEXT: cmpxchg word ptr [e{{[^a].}}], [[REG]] |
| 440 ; CHECK: jne -{{[0-9]}} | 435 ; CHECK: jne-{{[0-9]}} |
| 441 | 436 |
| 442 define i32 @test_atomic_rmw_or_32(i32 %iptr, i32 %v) { | 437 define i32 @test_atomic_rmw_or_32(i32 %iptr, i32 %v) { |
| 443 entry: | 438 entry: |
| 444 %ptr = inttoptr i32 %iptr to i32* | 439 %ptr = inttoptr i32 %iptr to i32* |
| 445 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 3, i32* %ptr, i32 %v, i32 6) | 440 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 3, i32* %ptr, i32 %v, i32 6) |
| 446 ret i32 %a | 441 ret i32 %a |
| 447 } | 442 } |
| 448 ; CHECK-LABEL: test_atomic_rmw_or_32 | 443 ; CHECK-LABEL: test_atomic_rmw_or_32 |
| 449 ; CHECK: mov eax, dword ptr | 444 ; CHECK: mov eax,DWORD ptr |
| 450 ; CHECK: or [[REG:e[^a].]] | 445 ; CHECK: or [[REG:e[^a].]] |
| 451 ; CHECK: lock | 446 ; CHECK: lock |
| 452 ; CHECK-NEXT: cmpxchg dword ptr [e{{[^a].}}], [[REG]] | 447 ; CHECK-NEXT: cmpxchg DWORD PTR [e{{[^a].}}], [[REG]] |
| 453 ; CHECK: jne -{{[0-9]}} | 448 ; CHECK: jne-{{[0-9]}} |
| 454 | 449 |
| 455 define i64 @test_atomic_rmw_or_64(i32 %iptr, i64 %v) { | 450 define i64 @test_atomic_rmw_or_64(i32 %iptr, i64 %v) { |
| 456 entry: | 451 entry: |
| 457 %ptr = inttoptr i32 %iptr to i64* | 452 %ptr = inttoptr i32 %iptr to i64* |
| 458 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 3, i64* %ptr, i64 %v, i32 6) | 453 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 3, i64* %ptr, i64 %v, i32 6) |
| 459 ret i64 %a | 454 ret i64 %a |
| 460 } | 455 } |
| 461 ; CHECK-LABEL: test_atomic_rmw_or_64 | 456 ; CHECK-LABEL: test_atomic_rmw_or_64 |
| 462 ; CHECK: push ebx | 457 ; CHECK: push ebx |
| 463 ; CHECK: mov eax, dword ptr [{{.*}}] | 458 ; CHECK: mov eax,DWORD ptr [{{.*}}] |
| 464 ; CHECK: mov edx, dword ptr [{{.*}} + 4] | 459 ; CHECK: mov edx,DWORD ptr [{{.*}}+4] |
| 465 ; CHECK: mov ebx, eax | 460 ; CHECK: mov ebx,eax |
| 466 ; CHECK: or ebx, {{.*e.[^x]}} | 461 ; CHECK: or ebx, {{.*e.[^x]}} |
| 467 ; CHECK: mov ecx, edx | 462 ; CHECK: mov ecx,edx |
| 468 ; CHECK: or ecx, {{.*e.[^x]}} | 463 ; CHECK: or ecx, {{.*e.[^x]}} |
| 469 ; CHECK: lock | 464 ; CHECK: lock |
| 470 ; CHECK-NEXT: cmpxchg8b qword ptr [e{{.[^x]}}] | 465 ; CHECK-NEXT: cmpxchg8b QWORD PTR [e{{.[^x]}}] |
| 471 ; CHECK: jne -{{[0-9]}} | 466 ; CHECK: jne-{{[0-9]}} |
| 472 | 467 |
| 473 define i32 @test_atomic_rmw_or_32_ignored(i32 %iptr, i32 %v) { | 468 define i32 @test_atomic_rmw_or_32_ignored(i32 %iptr, i32 %v) { |
| 474 entry: | 469 entry: |
| 475 %ptr = inttoptr i32 %iptr to i32* | 470 %ptr = inttoptr i32 %iptr to i32* |
| 476 %ignored = call i32 @llvm.nacl.atomic.rmw.i32(i32 3, i32* %ptr, i32 %v, i32 6) | 471 %ignored = call i32 @llvm.nacl.atomic.rmw.i32(i32 3, i32* %ptr, i32 %v, i32 6) |
| 477 ret i32 %v | 472 ret i32 %v |
| 478 } | 473 } |
| 479 ; CHECK-LABEL: test_atomic_rmw_or_32_ignored | 474 ; CHECK-LABEL: test_atomic_rmw_or_32_ignored |
| 480 ; Could just "lock or", if we inspect the liveness information first. | 475 ; Could just "lock or", if we inspect the liveness information first. |
| 481 ; Would also need a way to introduce "lock"'edness to binary | 476 ; Would also need a way to introduce "lock"'edness to binary |
| 482 ; operators without introducing overhead on the more common binary ops. | 477 ; operators without introducing overhead on the more common binary ops. |
| 483 ; CHECK: mov eax, dword ptr | 478 ; CHECK: mov eax,DWORD ptr |
| 484 ; CHECK: or [[REG:e[^a].]] | 479 ; CHECK: or [[REG:e[^a].]] |
| 485 ; CHECK: lock | 480 ; CHECK: lock |
| 486 ; CHECK-NEXT: cmpxchg dword ptr [e{{[^a].}}], [[REG]] | 481 ; CHECK-NEXT: cmpxchg DWORD PTR [e{{[^a].}}], [[REG]] |
| 487 ; CHECK: jne -{{[0-9]}} | 482 ; CHECK: jne-{{[0-9]}} |
| 488 | 483 |
| 489 ;; and | 484 ;; and |
| 490 | 485 |
| 491 define i32 @test_atomic_rmw_and_8(i32 %iptr, i32 %v) { | 486 define i32 @test_atomic_rmw_and_8(i32 %iptr, i32 %v) { |
| 492 entry: | 487 entry: |
| 493 %trunc = trunc i32 %v to i8 | 488 %trunc = trunc i32 %v to i8 |
| 494 %ptr = inttoptr i32 %iptr to i8* | 489 %ptr = inttoptr i32 %iptr to i8* |
| 495 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 4, i8* %ptr, i8 %trunc, i32 6) | 490 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 4, i8* %ptr, i8 %trunc, i32 6) |
| 496 %a_ext = zext i8 %a to i32 | 491 %a_ext = zext i8 %a to i32 |
| 497 ret i32 %a_ext | 492 ret i32 %a_ext |
| 498 } | 493 } |
| 499 ; CHECK-LABEL: test_atomic_rmw_and_8 | 494 ; CHECK-LABEL: test_atomic_rmw_and_8 |
| 500 ; CHECK: mov al, byte ptr | 495 ; CHECK: mov al,BYTE PTR |
| 501 ; CHECK: and [[REG:[^a].]] | 496 ; CHECK: and [[REG:[^a].]] |
| 502 ; CHECK: lock | 497 ; CHECK: lock |
| 503 ; CHECK-NEXT: cmpxchg byte ptr [e{{[^a].}}], [[REG]] | 498 ; CHECK-NEXT: cmpxchg BYTE PTR [e{{[^a].}}], [[REG]] |
| 504 ; CHECK: jne -{{[0-9]}} | 499 ; CHECK: jne-{{[0-9]}} |
| 505 | 500 |
| 506 define i32 @test_atomic_rmw_and_16(i32 %iptr, i32 %v) { | 501 define i32 @test_atomic_rmw_and_16(i32 %iptr, i32 %v) { |
| 507 entry: | 502 entry: |
| 508 %trunc = trunc i32 %v to i16 | 503 %trunc = trunc i32 %v to i16 |
| 509 %ptr = inttoptr i32 %iptr to i16* | 504 %ptr = inttoptr i32 %iptr to i16* |
| 510 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 4, i16* %ptr, i16 %trunc, i32 6) | 505 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 4, i16* %ptr, i16 %trunc, i32 6) |
| 511 %a_ext = zext i16 %a to i32 | 506 %a_ext = zext i16 %a to i32 |
| 512 ret i32 %a_ext | 507 ret i32 %a_ext |
| 513 } | 508 } |
| 514 ; CHECK-LABEL: test_atomic_rmw_and_16 | 509 ; CHECK-LABEL: test_atomic_rmw_and_16 |
| 515 ; CHECK: mov ax, word ptr | 510 ; CHECK: mov ax,word ptr |
| 516 ; CHECK: and | 511 ; CHECK: and |
| 517 ; CHECK: lock | 512 ; CHECK: lock |
| 518 ; CHECK-NEXT: cmpxchg word ptr [e{{[^a].}}] | 513 ; CHECK-NEXT: cmpxchg word ptr [e{{[^a].}}] |
| 519 ; CHECK: jne -{{[0-9]}} | 514 ; CHECK: jne-{{[0-9]}} |
| 520 | 515 |
| 521 define i32 @test_atomic_rmw_and_32(i32 %iptr, i32 %v) { | 516 define i32 @test_atomic_rmw_and_32(i32 %iptr, i32 %v) { |
| 522 entry: | 517 entry: |
| 523 %ptr = inttoptr i32 %iptr to i32* | 518 %ptr = inttoptr i32 %iptr to i32* |
| 524 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 4, i32* %ptr, i32 %v, i32 6) | 519 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 4, i32* %ptr, i32 %v, i32 6) |
| 525 ret i32 %a | 520 ret i32 %a |
| 526 } | 521 } |
| 527 ; CHECK-LABEL: test_atomic_rmw_and_32 | 522 ; CHECK-LABEL: test_atomic_rmw_and_32 |
| 528 ; CHECK: mov eax, dword ptr | 523 ; CHECK: mov eax,DWORD ptr |
| 529 ; CHECK: and | 524 ; CHECK: and |
| 530 ; CHECK: lock | 525 ; CHECK: lock |
| 531 ; CHECK-NEXT: cmpxchg dword ptr [e{{[^a].}}] | 526 ; CHECK-NEXT: cmpxchg DWORD PTR [e{{[^a].}}] |
| 532 ; CHECK: jne -{{[0-9]}} | 527 ; CHECK: jne-{{[0-9]}} |
| 533 | 528 |
| 534 define i64 @test_atomic_rmw_and_64(i32 %iptr, i64 %v) { | 529 define i64 @test_atomic_rmw_and_64(i32 %iptr, i64 %v) { |
| 535 entry: | 530 entry: |
| 536 %ptr = inttoptr i32 %iptr to i64* | 531 %ptr = inttoptr i32 %iptr to i64* |
| 537 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 4, i64* %ptr, i64 %v, i32 6) | 532 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 4, i64* %ptr, i64 %v, i32 6) |
| 538 ret i64 %a | 533 ret i64 %a |
| 539 } | 534 } |
| 540 ; CHECK-LABEL: test_atomic_rmw_and_64 | 535 ; CHECK-LABEL: test_atomic_rmw_and_64 |
| 541 ; CHECK: push ebx | 536 ; CHECK: push ebx |
| 542 ; CHECK: mov eax, dword ptr [{{.*}}] | 537 ; CHECK: mov eax,DWORD ptr [{{.*}}] |
| 543 ; CHECK: mov edx, dword ptr [{{.*}} + 4] | 538 ; CHECK: mov edx,DWORD ptr [{{.*}}+4] |
| 544 ; CHECK: mov ebx, eax | 539 ; CHECK: mov ebx,eax |
| 545 ; CHECK: and ebx, {{.*e.[^x]}} | 540 ; CHECK: and ebx, {{.*e.[^x]}} |
| 546 ; CHECK: mov ecx, edx | 541 ; CHECK: mov ecx,edx |
| 547 ; CHECK: and ecx, {{.*e.[^x]}} | 542 ; CHECK: and ecx, {{.*e.[^x]}} |
| 548 ; CHECK: lock | 543 ; CHECK: lock |
| 549 ; CHECK-NEXT: cmpxchg8b qword ptr [e{{.[^x]}}] | 544 ; CHECK-NEXT: cmpxchg8b QWORD PTR [e{{.[^x]}}] |
| 550 ; CHECK: jne -{{[0-9]}} | 545 ; CHECK: jne-{{[0-9]}} |
| 551 | 546 |
| 552 define i32 @test_atomic_rmw_and_32_ignored(i32 %iptr, i32 %v) { | 547 define i32 @test_atomic_rmw_and_32_ignored(i32 %iptr, i32 %v) { |
| 553 entry: | 548 entry: |
| 554 %ptr = inttoptr i32 %iptr to i32* | 549 %ptr = inttoptr i32 %iptr to i32* |
| 555 %ignored = call i32 @llvm.nacl.atomic.rmw.i32(i32 4, i32* %ptr, i32 %v, i32 6) | 550 %ignored = call i32 @llvm.nacl.atomic.rmw.i32(i32 4, i32* %ptr, i32 %v, i32 6) |
| 556 ret i32 %v | 551 ret i32 %v |
| 557 } | 552 } |
| 558 ; CHECK-LABEL: test_atomic_rmw_and_32_ignored | 553 ; CHECK-LABEL: test_atomic_rmw_and_32_ignored |
| 559 ; Could just "lock and" | 554 ; Could just "lock and" |
| 560 ; CHECK: mov eax, dword ptr | 555 ; CHECK: mov eax,DWORD ptr |
| 561 ; CHECK: and | 556 ; CHECK: and |
| 562 ; CHECK: lock | 557 ; CHECK: lock |
| 563 ; CHECK-NEXT: cmpxchg dword ptr [e{{[^a].}}] | 558 ; CHECK-NEXT: cmpxchg DWORD PTR [e{{[^a].}}] |
| 564 ; CHECK: jne -{{[0-9]}} | 559 ; CHECK: jne-{{[0-9]}} |
| 565 | 560 |
| 566 ;; xor | 561 ;; xor |
| 567 | 562 |
| 568 define i32 @test_atomic_rmw_xor_8(i32 %iptr, i32 %v) { | 563 define i32 @test_atomic_rmw_xor_8(i32 %iptr, i32 %v) { |
| 569 entry: | 564 entry: |
| 570 %trunc = trunc i32 %v to i8 | 565 %trunc = trunc i32 %v to i8 |
| 571 %ptr = inttoptr i32 %iptr to i8* | 566 %ptr = inttoptr i32 %iptr to i8* |
| 572 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 5, i8* %ptr, i8 %trunc, i32 6) | 567 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 5, i8* %ptr, i8 %trunc, i32 6) |
| 573 %a_ext = zext i8 %a to i32 | 568 %a_ext = zext i8 %a to i32 |
| 574 ret i32 %a_ext | 569 ret i32 %a_ext |
| 575 } | 570 } |
| 576 ; CHECK-LABEL: test_atomic_rmw_xor_8 | 571 ; CHECK-LABEL: test_atomic_rmw_xor_8 |
| 577 ; CHECK: mov al, byte ptr | 572 ; CHECK: mov al,BYTE PTR |
| 578 ; CHECK: xor [[REG:[^a].]] | 573 ; CHECK: xor [[REG:[^a].]] |
| 579 ; CHECK: lock | 574 ; CHECK: lock |
| 580 ; CHECK-NEXT: cmpxchg byte ptr [e{{[^a].}}], [[REG]] | 575 ; CHECK-NEXT: cmpxchg BYTE PTR [e{{[^a].}}], [[REG]] |
| 581 ; CHECK: jne -{{[0-9]}} | 576 ; CHECK: jne-{{[0-9]}} |
| 582 | 577 |
| 583 define i32 @test_atomic_rmw_xor_16(i32 %iptr, i32 %v) { | 578 define i32 @test_atomic_rmw_xor_16(i32 %iptr, i32 %v) { |
| 584 entry: | 579 entry: |
| 585 %trunc = trunc i32 %v to i16 | 580 %trunc = trunc i32 %v to i16 |
| 586 %ptr = inttoptr i32 %iptr to i16* | 581 %ptr = inttoptr i32 %iptr to i16* |
| 587 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 5, i16* %ptr, i16 %trunc, i32 6) | 582 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 5, i16* %ptr, i16 %trunc, i32 6) |
| 588 %a_ext = zext i16 %a to i32 | 583 %a_ext = zext i16 %a to i32 |
| 589 ret i32 %a_ext | 584 ret i32 %a_ext |
| 590 } | 585 } |
| 591 ; CHECK-LABEL: test_atomic_rmw_xor_16 | 586 ; CHECK-LABEL: test_atomic_rmw_xor_16 |
| 592 ; CHECK: mov ax, word ptr | 587 ; CHECK: mov ax,word ptr |
| 593 ; CHECK: xor | 588 ; CHECK: xor |
| 594 ; CHECK: lock | 589 ; CHECK: lock |
| 595 ; CHECK-NEXT: cmpxchg word ptr [e{{[^a].}}] | 590 ; CHECK-NEXT: cmpxchg word ptr [e{{[^a].}}] |
| 596 ; CHECK: jne -{{[0-9]}} | 591 ; CHECK: jne-{{[0-9]}} |
| 597 | 592 |
| 598 | 593 |
| 599 define i32 @test_atomic_rmw_xor_32(i32 %iptr, i32 %v) { | 594 define i32 @test_atomic_rmw_xor_32(i32 %iptr, i32 %v) { |
| 600 entry: | 595 entry: |
| 601 %ptr = inttoptr i32 %iptr to i32* | 596 %ptr = inttoptr i32 %iptr to i32* |
| 602 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 5, i32* %ptr, i32 %v, i32 6) | 597 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 5, i32* %ptr, i32 %v, i32 6) |
| 603 ret i32 %a | 598 ret i32 %a |
| 604 } | 599 } |
| 605 ; CHECK-LABEL: test_atomic_rmw_xor_32 | 600 ; CHECK-LABEL: test_atomic_rmw_xor_32 |
| 606 ; CHECK: mov eax, dword ptr | 601 ; CHECK: mov eax,DWORD ptr |
| 607 ; CHECK: xor | 602 ; CHECK: xor |
| 608 ; CHECK: lock | 603 ; CHECK: lock |
| 609 ; CHECK-NEXT: cmpxchg dword ptr [e{{[^a].}}] | 604 ; CHECK-NEXT: cmpxchg DWORD PTR [e{{[^a].}}] |
| 610 ; CHECK: jne -{{[0-9]}} | 605 ; CHECK: jne-{{[0-9]}} |
| 611 | 606 |
| 612 define i64 @test_atomic_rmw_xor_64(i32 %iptr, i64 %v) { | 607 define i64 @test_atomic_rmw_xor_64(i32 %iptr, i64 %v) { |
| 613 entry: | 608 entry: |
| 614 %ptr = inttoptr i32 %iptr to i64* | 609 %ptr = inttoptr i32 %iptr to i64* |
| 615 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 5, i64* %ptr, i64 %v, i32 6) | 610 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 5, i64* %ptr, i64 %v, i32 6) |
| 616 ret i64 %a | 611 ret i64 %a |
| 617 } | 612 } |
| 618 ; CHECK-LABEL: test_atomic_rmw_xor_64 | 613 ; CHECK-LABEL: test_atomic_rmw_xor_64 |
| 619 ; CHECK: push ebx | 614 ; CHECK: push ebx |
| 620 ; CHECK: mov eax, dword ptr [{{.*}}] | 615 ; CHECK: mov eax,DWORD ptr [{{.*}}] |
| 621 ; CHECK: mov edx, dword ptr [{{.*}} + 4] | 616 ; CHECK: mov edx,DWORD ptr [{{.*}}+4] |
| 622 ; CHECK: mov ebx, eax | 617 ; CHECK: mov ebx,eax |
| 623 ; CHECK: or ebx, {{.*e.[^x]}} | 618 ; CHECK: or ebx, {{.*e.[^x]}} |
| 624 ; CHECK: mov ecx, edx | 619 ; CHECK: mov ecx,edx |
| 625 ; CHECK: or ecx, {{.*e.[^x]}} | 620 ; CHECK: or ecx, {{.*e.[^x]}} |
| 626 ; CHECK: lock | 621 ; CHECK: lock |
| 627 ; CHECK-NEXT: cmpxchg8b qword ptr [e{{.[^x]}}] | 622 ; CHECK-NEXT: cmpxchg8b QWORD PTR [e{{.[^x]}}] |
| 628 ; CHECK: jne -{{[0-9]}} | 623 ; CHECK: jne-{{[0-9]}} |
| 629 | 624 |
| 630 define i32 @test_atomic_rmw_xor_32_ignored(i32 %iptr, i32 %v) { | 625 define i32 @test_atomic_rmw_xor_32_ignored(i32 %iptr, i32 %v) { |
| 631 entry: | 626 entry: |
| 632 %ptr = inttoptr i32 %iptr to i32* | 627 %ptr = inttoptr i32 %iptr to i32* |
| 633 %ignored = call i32 @llvm.nacl.atomic.rmw.i32(i32 5, i32* %ptr, i32 %v, i32 6) | 628 %ignored = call i32 @llvm.nacl.atomic.rmw.i32(i32 5, i32* %ptr, i32 %v, i32 6) |
| 634 ret i32 %v | 629 ret i32 %v |
| 635 } | 630 } |
| 636 ; CHECK-LABEL: test_atomic_rmw_xor_32_ignored | 631 ; CHECK-LABEL: test_atomic_rmw_xor_32_ignored |
| 637 ; CHECK: mov eax, dword ptr | 632 ; CHECK: mov eax,DWORD ptr |
| 638 ; CHECK: xor | 633 ; CHECK: xor |
| 639 ; CHECK: lock | 634 ; CHECK: lock |
| 640 ; CHECK-NEXT: cmpxchg dword ptr [e{{[^a].}}] | 635 ; CHECK-NEXT: cmpxchg DWORD PTR [e{{[^a].}}] |
| 641 ; CHECK: jne -{{[0-9]}} | 636 ; CHECK: jne-{{[0-9]}} |
| 642 | 637 |
| 643 ;; exchange | 638 ;; exchange |
| 644 | 639 |
| 645 define i32 @test_atomic_rmw_xchg_8(i32 %iptr, i32 %v) { | 640 define i32 @test_atomic_rmw_xchg_8(i32 %iptr, i32 %v) { |
| 646 entry: | 641 entry: |
| 647 %trunc = trunc i32 %v to i8 | 642 %trunc = trunc i32 %v to i8 |
| 648 %ptr = inttoptr i32 %iptr to i8* | 643 %ptr = inttoptr i32 %iptr to i8* |
| 649 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 6, i8* %ptr, i8 %trunc, i32 6) | 644 %a = call i8 @llvm.nacl.atomic.rmw.i8(i32 6, i8* %ptr, i8 %trunc, i32 6) |
| 650 %a_ext = zext i8 %a to i32 | 645 %a_ext = zext i8 %a to i32 |
| 651 ret i32 %a_ext | 646 ret i32 %a_ext |
| 652 } | 647 } |
| 653 ; CHECK-LABEL: test_atomic_rmw_xchg_8 | 648 ; CHECK-LABEL: test_atomic_rmw_xchg_8 |
| 654 ; CHECK: xchg byte ptr {{.*}}, [[REG:.*]] | 649 ; CHECK: xchg BYTE PTR {{.*}}, [[REG:.*]] |
| 655 | 650 |
| 656 define i32 @test_atomic_rmw_xchg_16(i32 %iptr, i32 %v) { | 651 define i32 @test_atomic_rmw_xchg_16(i32 %iptr, i32 %v) { |
| 657 entry: | 652 entry: |
| 658 %trunc = trunc i32 %v to i16 | 653 %trunc = trunc i32 %v to i16 |
| 659 %ptr = inttoptr i32 %iptr to i16* | 654 %ptr = inttoptr i32 %iptr to i16* |
| 660 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 6, i16* %ptr, i16 %trunc, i32 6) | 655 %a = call i16 @llvm.nacl.atomic.rmw.i16(i32 6, i16* %ptr, i16 %trunc, i32 6) |
| 661 %a_ext = zext i16 %a to i32 | 656 %a_ext = zext i16 %a to i32 |
| 662 ret i32 %a_ext | 657 ret i32 %a_ext |
| 663 } | 658 } |
| 664 ; CHECK-LABEL: test_atomic_rmw_xchg_16 | 659 ; CHECK-LABEL: test_atomic_rmw_xchg_16 |
| 665 ; CHECK: xchg word ptr {{.*}}, [[REG:.*]] | 660 ; CHECK: xchg word ptr {{.*}}, [[REG:.*]] |
| 666 | 661 |
| 667 define i32 @test_atomic_rmw_xchg_32(i32 %iptr, i32 %v) { | 662 define i32 @test_atomic_rmw_xchg_32(i32 %iptr, i32 %v) { |
| 668 entry: | 663 entry: |
| 669 %ptr = inttoptr i32 %iptr to i32* | 664 %ptr = inttoptr i32 %iptr to i32* |
| 670 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 6, i32* %ptr, i32 %v, i32 6) | 665 %a = call i32 @llvm.nacl.atomic.rmw.i32(i32 6, i32* %ptr, i32 %v, i32 6) |
| 671 ret i32 %a | 666 ret i32 %a |
| 672 } | 667 } |
| 673 ; CHECK-LABEL: test_atomic_rmw_xchg_32 | 668 ; CHECK-LABEL: test_atomic_rmw_xchg_32 |
| 674 ; CHECK: xchg dword ptr {{.*}}, [[REG:.*]] | 669 ; CHECK: xchg DWORD PTR {{.*}}, [[REG:.*]] |
| 675 | 670 |
| 676 define i64 @test_atomic_rmw_xchg_64(i32 %iptr, i64 %v) { | 671 define i64 @test_atomic_rmw_xchg_64(i32 %iptr, i64 %v) { |
| 677 entry: | 672 entry: |
| 678 %ptr = inttoptr i32 %iptr to i64* | 673 %ptr = inttoptr i32 %iptr to i64* |
| 679 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 6, i64* %ptr, i64 %v, i32 6) | 674 %a = call i64 @llvm.nacl.atomic.rmw.i64(i32 6, i64* %ptr, i64 %v, i32 6) |
| 680 ret i64 %a | 675 ret i64 %a |
| 681 } | 676 } |
| 682 ; CHECK-LABEL: test_atomic_rmw_xchg_64 | 677 ; CHECK-LABEL: test_atomic_rmw_xchg_64 |
| 683 ; CHECK: push ebx | 678 ; CHECK: push ebx |
| 684 ; CHECK-DAG: mov edx | 679 ; CHECK-DAG: mov edx |
| 685 ; CHECK-DAG: mov eax | 680 ; CHECK-DAG: mov eax |
| 686 ; CHECK-DAG: mov ecx | 681 ; CHECK-DAG: mov ecx |
| 687 ; CHECK-DAG: mov ebx | 682 ; CHECK-DAG: mov ebx |
| 688 ; CHECK: lock | 683 ; CHECK: lock |
| 689 ; CHECK-NEXT: cmpxchg8b qword ptr [{{e.[^x]}}] | 684 ; CHECK-NEXT: cmpxchg8b QWORD PTR [{{e.[^x]}}] |
| 690 ; CHECK: jne -{{[0-9]}} | 685 ; CHECK: jne-{{[0-9]}} |
| 691 | 686 |
| 692 define i32 @test_atomic_rmw_xchg_32_ignored(i32 %iptr, i32 %v) { | 687 define i32 @test_atomic_rmw_xchg_32_ignored(i32 %iptr, i32 %v) { |
| 693 entry: | 688 entry: |
| 694 %ptr = inttoptr i32 %iptr to i32* | 689 %ptr = inttoptr i32 %iptr to i32* |
| 695 %ignored = call i32 @llvm.nacl.atomic.rmw.i32(i32 6, i32* %ptr, i32 %v, i32 6) | 690 %ignored = call i32 @llvm.nacl.atomic.rmw.i32(i32 6, i32* %ptr, i32 %v, i32 6) |
| 696 ret i32 %v | 691 ret i32 %v |
| 697 } | 692 } |
| 698 ; In this case, ignoring the return value doesn't help. The xchg is | 693 ; In this case, ignoring the return value doesn't help. The xchg is |
| 699 ; used to do an atomic store. | 694 ; used to do an atomic store. |
| 700 ; CHECK-LABEL: test_atomic_rmw_xchg_32_ignored | 695 ; CHECK-LABEL: test_atomic_rmw_xchg_32_ignored |
| 701 ; CHECK: xchg dword ptr {{.*}}, [[REG:.*]] | 696 ; CHECK: xchg DWORD PTR {{.*}}, [[REG:.*]] |
| 702 | 697 |
| 703 ;;;; Cmpxchg | 698 ;;;; Cmpxchg |
| 704 | 699 |
| 705 define i32 @test_atomic_cmpxchg_8(i32 %iptr, i32 %expected, i32 %desired) { | 700 define i32 @test_atomic_cmpxchg_8(i32 %iptr, i32 %expected, i32 %desired) { |
| 706 entry: | 701 entry: |
| 707 %trunc_exp = trunc i32 %expected to i8 | 702 %trunc_exp = trunc i32 %expected to i8 |
| 708 %trunc_des = trunc i32 %desired to i8 | 703 %trunc_des = trunc i32 %desired to i8 |
| 709 %ptr = inttoptr i32 %iptr to i8* | 704 %ptr = inttoptr i32 %iptr to i8* |
| 710 %old = call i8 @llvm.nacl.atomic.cmpxchg.i8(i8* %ptr, i8 %trunc_exp, | 705 %old = call i8 @llvm.nacl.atomic.cmpxchg.i8(i8* %ptr, i8 %trunc_exp, |
| 711 i8 %trunc_des, i32 6, i32 6) | 706 i8 %trunc_des, i32 6, i32 6) |
| 712 %old_ext = zext i8 %old to i32 | 707 %old_ext = zext i8 %old to i32 |
| 713 ret i32 %old_ext | 708 ret i32 %old_ext |
| 714 } | 709 } |
| 715 ; CHECK-LABEL: test_atomic_cmpxchg_8 | 710 ; CHECK-LABEL: test_atomic_cmpxchg_8 |
| 716 ; CHECK: mov eax, {{.*}} | 711 ; CHECK: mov eax,{{.*}} |
| 717 ; Need to check that eax isn't used as the address register or the desired. | 712 ; Need to check that eax isn't used as the address register or the desired. |
| 718 ; since it is already used as the *expected* register. | 713 ; since it is already used as the *expected* register. |
| 719 ; CHECK: lock | 714 ; CHECK: lock |
| 720 ; CHECK-NEXT: cmpxchg byte ptr [e{{[^a].}}], {{[^a]}}l | 715 ; CHECK-NEXT: cmpxchg BYTE PTR [e{{[^a].}}], {{[^a]}}l |
| 721 | 716 |
| 722 define i32 @test_atomic_cmpxchg_16(i32 %iptr, i32 %expected, i32 %desired) { | 717 define i32 @test_atomic_cmpxchg_16(i32 %iptr, i32 %expected, i32 %desired) { |
| 723 entry: | 718 entry: |
| 724 %trunc_exp = trunc i32 %expected to i16 | 719 %trunc_exp = trunc i32 %expected to i16 |
| 725 %trunc_des = trunc i32 %desired to i16 | 720 %trunc_des = trunc i32 %desired to i16 |
| 726 %ptr = inttoptr i32 %iptr to i16* | 721 %ptr = inttoptr i32 %iptr to i16* |
| 727 %old = call i16 @llvm.nacl.atomic.cmpxchg.i16(i16* %ptr, i16 %trunc_exp, | 722 %old = call i16 @llvm.nacl.atomic.cmpxchg.i16(i16* %ptr, i16 %trunc_exp, |
| 728 i16 %trunc_des, i32 6, i32 6) | 723 i16 %trunc_des, i32 6, i32 6) |
| 729 %old_ext = zext i16 %old to i32 | 724 %old_ext = zext i16 %old to i32 |
| 730 ret i32 %old_ext | 725 ret i32 %old_ext |
| 731 } | 726 } |
| 732 ; CHECK-LABEL: test_atomic_cmpxchg_16 | 727 ; CHECK-LABEL: test_atomic_cmpxchg_16 |
| 733 ; CHECK: mov eax, {{.*}} | 728 ; CHECK: mov eax,{{.*}} |
| 734 ; CHECK: lock | 729 ; CHECK: lock |
| 735 ; CHECK-NEXT: cmpxchg word ptr [e{{[^a].}}], {{[^a]}}x | 730 ; CHECK-NEXT: cmpxchg word ptr [e{{[^a].}}], {{[^a]}}x |
| 736 | 731 |
| 737 define i32 @test_atomic_cmpxchg_32(i32 %iptr, i32 %expected, i32 %desired) { | 732 define i32 @test_atomic_cmpxchg_32(i32 %iptr, i32 %expected, i32 %desired) { |
| 738 entry: | 733 entry: |
| 739 %ptr = inttoptr i32 %iptr to i32* | 734 %ptr = inttoptr i32 %iptr to i32* |
| 740 %old = call i32 @llvm.nacl.atomic.cmpxchg.i32(i32* %ptr, i32 %expected, | 735 %old = call i32 @llvm.nacl.atomic.cmpxchg.i32(i32* %ptr, i32 %expected, |
| 741 i32 %desired, i32 6, i32 6) | 736 i32 %desired, i32 6, i32 6) |
| 742 ret i32 %old | 737 ret i32 %old |
| 743 } | 738 } |
| 744 ; CHECK-LABEL: test_atomic_cmpxchg_32 | 739 ; CHECK-LABEL: test_atomic_cmpxchg_32 |
| 745 ; CHECK: mov eax, {{.*}} | 740 ; CHECK: mov eax,{{.*}} |
| 746 ; CHECK: lock | 741 ; CHECK: lock |
| 747 ; CHECK-NEXT: cmpxchg dword ptr [e{{[^a].}}], e{{[^a]}} | 742 ; CHECK-NEXT: cmpxchg DWORD PTR [e{{[^a].}}], e{{[^a]}} |
| 748 | 743 |
| 749 define i64 @test_atomic_cmpxchg_64(i32 %iptr, i64 %expected, i64 %desired) { | 744 define i64 @test_atomic_cmpxchg_64(i32 %iptr, i64 %expected, i64 %desired) { |
| 750 entry: | 745 entry: |
| 751 %ptr = inttoptr i32 %iptr to i64* | 746 %ptr = inttoptr i32 %iptr to i64* |
| 752 %old = call i64 @llvm.nacl.atomic.cmpxchg.i64(i64* %ptr, i64 %expected, | 747 %old = call i64 @llvm.nacl.atomic.cmpxchg.i64(i64* %ptr, i64 %expected, |
| 753 i64 %desired, i32 6, i32 6) | 748 i64 %desired, i32 6, i32 6) |
| 754 ret i64 %old | 749 ret i64 %old |
| 755 } | 750 } |
| 756 ; CHECK-LABEL: test_atomic_cmpxchg_64 | 751 ; CHECK-LABEL: test_atomic_cmpxchg_64 |
| 757 ; CHECK: push ebx | 752 ; CHECK: push ebx |
| 758 ; CHECK-DAG: mov edx | 753 ; CHECK-DAG: mov edx |
| 759 ; CHECK-DAG: mov eax | 754 ; CHECK-DAG: mov eax |
| 760 ; CHECK-DAG: mov ecx | 755 ; CHECK-DAG: mov ecx |
| 761 ; CHECK-DAG: mov ebx | 756 ; CHECK-DAG: mov ebx |
| 762 ; CHECK: lock | 757 ; CHECK: lock |
| 763 ; CHECK-NEXT: cmpxchg8b qword ptr [e{{.[^x]}}] | 758 ; CHECK-NEXT: cmpxchg8b QWORD PTR [e{{.[^x]}}] |
| 764 ; edx and eax are already the return registers, so they don't actually | 759 ; edx and eax are already the return registers, so they don't actually |
| 765 ; need to be reshuffled via movs. The next test stores the result | 760 ; need to be reshuffled via movs. The next test stores the result |
| 766 ; somewhere, so in that case they do need to be mov'ed. | 761 ; somewhere, so in that case they do need to be mov'ed. |
| 767 | 762 |
| 768 ; Test a case where %old really does need to be copied out of edx:eax. | 763 ; Test a case where %old really does need to be copied out of edx:eax. |
| 769 define void @test_atomic_cmpxchg_64_store(i32 %ret_iptr, i32 %iptr, i64 %expecte
d, i64 %desired) { | 764 define void @test_atomic_cmpxchg_64_store(i32 %ret_iptr, i32 %iptr, i64 %expecte
d, i64 %desired) { |
| 770 entry: | 765 entry: |
| 771 %ptr = inttoptr i32 %iptr to i64* | 766 %ptr = inttoptr i32 %iptr to i64* |
| 772 %old = call i64 @llvm.nacl.atomic.cmpxchg.i64(i64* %ptr, i64 %expected, | 767 %old = call i64 @llvm.nacl.atomic.cmpxchg.i64(i64* %ptr, i64 %expected, |
| 773 i64 %desired, i32 6, i32 6) | 768 i64 %desired, i32 6, i32 6) |
| 774 %__6 = inttoptr i32 %ret_iptr to i64* | 769 %__6 = inttoptr i32 %ret_iptr to i64* |
| 775 store i64 %old, i64* %__6, align 1 | 770 store i64 %old, i64* %__6, align 1 |
| 776 ret void | 771 ret void |
| 777 } | 772 } |
| 778 ; CHECK-LABEL: test_atomic_cmpxchg_64_store | 773 ; CHECK-LABEL: test_atomic_cmpxchg_64_store |
| 779 ; CHECK: push ebx | 774 ; CHECK: push ebx |
| 780 ; CHECK-DAG: mov edx | 775 ; CHECK-DAG: mov edx |
| 781 ; CHECK-DAG: mov eax | 776 ; CHECK-DAG: mov eax |
| 782 ; CHECK-DAG: mov ecx | 777 ; CHECK-DAG: mov ecx |
| 783 ; CHECK-DAG: mov ebx | 778 ; CHECK-DAG: mov ebx |
| 784 ; CHECK: lock | 779 ; CHECK: lock |
| 785 ; CHECK-NEXT: cmpxchg8b qword ptr [e{{.[^x]}}] | 780 ; CHECK-NEXT: cmpxchg8b QWORD PTR [e{{.[^x]}}] |
| 786 ; CHECK-DAG: mov {{.*}}, edx | 781 ; CHECK-DAG: mov {{.*}}, edx |
| 787 ; CHECK-DAG: mov {{.*}}, eax | 782 ; CHECK-DAG: mov {{.*}}, eax |
| 788 | 783 |
| 789 ; Test with some more register pressure. When we have an alloca, ebp is | 784 ; Test with some more register pressure. When we have an alloca, ebp is |
| 790 ; used to manage the stack frame, so it cannot be used as a register either. | 785 ; used to manage the stack frame, so it cannot be used as a register either. |
| 791 define i64 @test_atomic_cmpxchg_64_alloca(i32 %iptr, i64 %expected, i64 %desired
) { | 786 define i64 @test_atomic_cmpxchg_64_alloca(i32 %iptr, i64 %expected, i64 %desired
) { |
| 792 entry: | 787 entry: |
| 793 %alloca_ptr = alloca i8, i32 16, align 16 | 788 %alloca_ptr = alloca i8, i32 16, align 16 |
| 794 %ptr = inttoptr i32 %iptr to i64* | 789 %ptr = inttoptr i32 %iptr to i64* |
| 795 %old = call i64 @llvm.nacl.atomic.cmpxchg.i64(i64* %ptr, i64 %expected, | 790 %old = call i64 @llvm.nacl.atomic.cmpxchg.i64(i64* %ptr, i64 %expected, |
| (...skipping 11 matching lines...) Expand all Loading... |
| 807 ; CHECK-DAG: mov edx | 802 ; CHECK-DAG: mov edx |
| 808 ; CHECK-DAG: mov eax | 803 ; CHECK-DAG: mov eax |
| 809 ; CHECK-DAG: mov ecx | 804 ; CHECK-DAG: mov ecx |
| 810 ; CHECK-DAG: mov ebx | 805 ; CHECK-DAG: mov ebx |
| 811 ; Ptr cannot be eax, ebx, ecx, or edx (used up for the expected and desired). | 806 ; Ptr cannot be eax, ebx, ecx, or edx (used up for the expected and desired). |
| 812 ; It also cannot be ebp since we use that for alloca. Also make sure it's | 807 ; It also cannot be ebp since we use that for alloca. Also make sure it's |
| 813 ; not esp, since that's the stack pointer and mucking with it will break | 808 ; not esp, since that's the stack pointer and mucking with it will break |
| 814 ; the later use_ptr function call. | 809 ; the later use_ptr function call. |
| 815 ; That pretty much leaves esi, or edi as the only viable registers. | 810 ; That pretty much leaves esi, or edi as the only viable registers. |
| 816 ; CHECK: lock | 811 ; CHECK: lock |
| 817 ; CHECK-NEXT: cmpxchg8b qword ptr [e{{[ds]}}i] | 812 ; CHECK-NEXT: cmpxchg8b QWORD PTR [e{{[ds]}}i] |
| 818 ; CHECK: call use_ptr | 813 ; CHECK: call |
| 814 ; CHECK-NEXT: R_{{.*}} use_ptr |
| 819 | 815 |
| 820 define i32 @test_atomic_cmpxchg_32_ignored(i32 %iptr, i32 %expected, i32 %desire
d) { | 816 define i32 @test_atomic_cmpxchg_32_ignored(i32 %iptr, i32 %expected, i32 %desire
d) { |
| 821 entry: | 817 entry: |
| 822 %ptr = inttoptr i32 %iptr to i32* | 818 %ptr = inttoptr i32 %iptr to i32* |
| 823 %ignored = call i32 @llvm.nacl.atomic.cmpxchg.i32(i32* %ptr, i32 %expected, | 819 %ignored = call i32 @llvm.nacl.atomic.cmpxchg.i32(i32* %ptr, i32 %expected, |
| 824 i32 %desired, i32 6, i32 6) | 820 i32 %desired, i32 6, i32 6) |
| 825 ret i32 0 | 821 ret i32 0 |
| 826 } | 822 } |
| 827 ; CHECK-LABEL: test_atomic_cmpxchg_32_ignored | 823 ; CHECK-LABEL: test_atomic_cmpxchg_32_ignored |
| 828 ; CHECK: mov eax, {{.*}} | 824 ; CHECK: mov eax,{{.*}} |
| 829 ; CHECK: lock | 825 ; CHECK: lock |
| 830 ; CHECK-NEXT: cmpxchg dword ptr [e{{[^a].}}] | 826 ; CHECK-NEXT: cmpxchg DWORD PTR [e{{[^a].}}] |
| 831 | 827 |
| 832 define i64 @test_atomic_cmpxchg_64_ignored(i32 %iptr, i64 %expected, i64 %desire
d) { | 828 define i64 @test_atomic_cmpxchg_64_ignored(i32 %iptr, i64 %expected, i64 %desire
d) { |
| 833 entry: | 829 entry: |
| 834 %ptr = inttoptr i32 %iptr to i64* | 830 %ptr = inttoptr i32 %iptr to i64* |
| 835 %ignored = call i64 @llvm.nacl.atomic.cmpxchg.i64(i64* %ptr, i64 %expected, | 831 %ignored = call i64 @llvm.nacl.atomic.cmpxchg.i64(i64* %ptr, i64 %expected, |
| 836 i64 %desired, i32 6, i32 6) | 832 i64 %desired, i32 6, i32 6) |
| 837 ret i64 0 | 833 ret i64 0 |
| 838 } | 834 } |
| 839 ; CHECK-LABEL: test_atomic_cmpxchg_64_ignored | 835 ; CHECK-LABEL: test_atomic_cmpxchg_64_ignored |
| 840 ; CHECK: push ebx | 836 ; CHECK: push ebx |
| 841 ; CHECK-DAG: mov edx | 837 ; CHECK-DAG: mov edx |
| 842 ; CHECK-DAG: mov eax | 838 ; CHECK-DAG: mov eax |
| 843 ; CHECK-DAG: mov ecx | 839 ; CHECK-DAG: mov ecx |
| 844 ; CHECK-DAG: mov ebx | 840 ; CHECK-DAG: mov ebx |
| 845 ; CHECK: lock | 841 ; CHECK: lock |
| 846 ; CHECK-NEXT: cmpxchg8b qword ptr [e{{.[^x]}}] | 842 ; CHECK-NEXT: cmpxchg8b QWORD PTR [e{{.[^x]}}] |
| 847 | 843 |
| 848 ;;;; Fence and is-lock-free. | 844 ;;;; Fence and is-lock-free. |
| 849 | 845 |
| 850 define void @test_atomic_fence() { | 846 define void @test_atomic_fence() { |
| 851 entry: | 847 entry: |
| 852 call void @llvm.nacl.atomic.fence(i32 6) | 848 call void @llvm.nacl.atomic.fence(i32 6) |
| 853 ret void | 849 ret void |
| 854 } | 850 } |
| 855 ; CHECK-LABEL: test_atomic_fence | 851 ; CHECK-LABEL: test_atomic_fence |
| 856 ; CHECK: mfence | 852 ; CHECK: mfence |
| 857 | 853 |
| 858 define void @test_atomic_fence_all() { | 854 define void @test_atomic_fence_all() { |
| 859 entry: | 855 entry: |
| 860 call void @llvm.nacl.atomic.fence.all() | 856 call void @llvm.nacl.atomic.fence.all() |
| 861 ret void | 857 ret void |
| 862 } | 858 } |
| 863 ; CHECK-LABEL: test_atomic_fence_all | 859 ; CHECK-LABEL: test_atomic_fence_all |
| 864 ; CHECK: mfence | 860 ; CHECK: mfence |
| 865 | 861 |
| 866 define i32 @test_atomic_is_lock_free(i32 %iptr) { | 862 define i32 @test_atomic_is_lock_free(i32 %iptr) { |
| 867 entry: | 863 entry: |
| 868 %ptr = inttoptr i32 %iptr to i8* | 864 %ptr = inttoptr i32 %iptr to i8* |
| 869 %i = call i1 @llvm.nacl.atomic.is.lock.free(i32 4, i8* %ptr) | 865 %i = call i1 @llvm.nacl.atomic.is.lock.free(i32 4, i8* %ptr) |
| 870 %r = zext i1 %i to i32 | 866 %r = zext i1 %i to i32 |
| 871 ret i32 %r | 867 ret i32 %r |
| 872 } | 868 } |
| 873 ; CHECK-LABEL: test_atomic_is_lock_free | 869 ; CHECK-LABEL: test_atomic_is_lock_free |
| 874 ; CHECK: mov {{.*}}, 1 | 870 ; CHECK: mov {{.*}},1 |
| 875 | 871 |
| 876 define i32 @test_not_lock_free(i32 %iptr) { | 872 define i32 @test_not_lock_free(i32 %iptr) { |
| 877 entry: | 873 entry: |
| 878 %ptr = inttoptr i32 %iptr to i8* | 874 %ptr = inttoptr i32 %iptr to i8* |
| 879 %i = call i1 @llvm.nacl.atomic.is.lock.free(i32 7, i8* %ptr) | 875 %i = call i1 @llvm.nacl.atomic.is.lock.free(i32 7, i8* %ptr) |
| 880 %r = zext i1 %i to i32 | 876 %r = zext i1 %i to i32 |
| 881 ret i32 %r | 877 ret i32 %r |
| 882 } | 878 } |
| 883 ; CHECK-LABEL: test_not_lock_free | 879 ; CHECK-LABEL: test_not_lock_free |
| 884 ; CHECK: mov {{.*}}, 0 | 880 ; CHECK: mov {{.*}},0 |
| 885 | 881 |
| 886 define i32 @test_atomic_is_lock_free_ignored(i32 %iptr) { | 882 define i32 @test_atomic_is_lock_free_ignored(i32 %iptr) { |
| 887 entry: | 883 entry: |
| 888 %ptr = inttoptr i32 %iptr to i8* | 884 %ptr = inttoptr i32 %iptr to i8* |
| 889 %ignored = call i1 @llvm.nacl.atomic.is.lock.free(i32 4, i8* %ptr) | 885 %ignored = call i1 @llvm.nacl.atomic.is.lock.free(i32 4, i8* %ptr) |
| 890 ret i32 0 | 886 ret i32 0 |
| 891 } | 887 } |
| 892 ; CHECK-LABEL: test_atomic_is_lock_free_ignored | 888 ; CHECK-LABEL: test_atomic_is_lock_free_ignored |
| 893 ; CHECK: mov {{.*}}, 0 | 889 ; CHECK: mov {{.*}},0 |
| 894 ; This can get optimized out, because it's side-effect-free. | 890 ; This can get optimized out, because it's side-effect-free. |
| 895 ; CHECKO2-LABEL: test_atomic_is_lock_free_ignored | 891 ; CHECKO2-LABEL: test_atomic_is_lock_free_ignored |
| 896 ; CHECKO2-NOT: mov {{.*}}, 1 | 892 ; CHECKO2-NOT: mov {{.*}}, 1 |
| 897 ; CHECKO2: mov {{.*}}, 0 | 893 ; CHECKO2: mov {{.*}}, 0 |
| 898 | 894 |
| 899 ; TODO(jvoung): at some point we can take advantage of the | 895 ; TODO(jvoung): at some point we can take advantage of the |
| 900 ; fact that nacl.atomic.is.lock.free will resolve to a constant | 896 ; fact that nacl.atomic.is.lock.free will resolve to a constant |
| 901 ; (which adds DCE opportunities). Once we optimize, the test expectations | 897 ; (which adds DCE opportunities). Once we optimize, the test expectations |
| 902 ; for this case should change. | 898 ; for this case should change. |
| 903 define i32 @test_atomic_is_lock_free_can_dce(i32 %iptr, i32 %x, i32 %y) { | 899 define i32 @test_atomic_is_lock_free_can_dce(i32 %iptr, i32 %x, i32 %y) { |
| 904 entry: | 900 entry: |
| 905 %ptr = inttoptr i32 %iptr to i8* | 901 %ptr = inttoptr i32 %iptr to i8* |
| 906 %i = call i1 @llvm.nacl.atomic.is.lock.free(i32 4, i8* %ptr) | 902 %i = call i1 @llvm.nacl.atomic.is.lock.free(i32 4, i8* %ptr) |
| 907 %i_ext = zext i1 %i to i32 | 903 %i_ext = zext i1 %i to i32 |
| 908 %cmp = icmp eq i32 %i_ext, 1 | 904 %cmp = icmp eq i32 %i_ext, 1 |
| 909 br i1 %cmp, label %lock_free, label %not_lock_free | 905 br i1 %cmp, label %lock_free, label %not_lock_free |
| 910 lock_free: | 906 lock_free: |
| 911 ret i32 %i_ext | 907 ret i32 %i_ext |
| 912 | 908 |
| 913 not_lock_free: | 909 not_lock_free: |
| 914 %z = add i32 %x, %y | 910 %z = add i32 %x, %y |
| 915 ret i32 %z | 911 ret i32 %z |
| 916 } | 912 } |
| 917 ; CHECK-LABEL: test_atomic_is_lock_free_can_dce | 913 ; CHECK-LABEL: test_atomic_is_lock_free_can_dce |
| 918 ; CHECK: mov {{.*}}, 1 | 914 ; CHECK: mov {{.*}},1 |
| 919 ; CHECK: ret | 915 ; CHECK: ret |
| 920 ; CHECK: add | 916 ; CHECK: add |
| 921 ; CHECK: ret | 917 ; CHECK: ret |
| OLD | NEW |