OLD | NEW |
1 ; RUN: pnacl-llc -mtriple=armv7-unknown-nacl -mattr=+neon -filetype=obj %s -o -
\ | 1 ; RUN: pnacl-llc -mtriple=armv7-unknown-nacl -mattr=+neon -filetype=obj %s -o -
\ |
2 ; RUN: | llvm-objdump -disassemble -triple armv7 - | FileCheck %s | 2 ; RUN: | llvm-objdump -disassemble -triple armv7 - | FileCheck %s |
3 | 3 |
4 define void @vst4i8(i8* %A, <8 x i8>* %B) nounwind { | 4 define void @vst4i8(i8* %A, <8 x i8>* %B) nounwind { |
5 %tmp1 = load <8 x i8>* %B | 5 %tmp1 = load <8 x i8>, <8 x i8>* %B |
6 call void @llvm.arm.neon.vst4.v8i8(i8* %A, <8 x i8> %tmp1, <8 x i8> %tmp1, <8
x i8> %tmp1, <8 x i8> %tmp1, i32 8) | 6 call void @llvm.arm.neon.vst4.v8i8(i8* %A, <8 x i8> %tmp1, <8 x i8> %tmp1, <8
x i8> %tmp1, <8 x i8> %tmp1, i32 8) |
7 ; CHECK: bic r0, r0, #3221225472 | 7 ; CHECK: bic r0, r0, #-1073741824 |
8 ; CHECK-NEXT: vst4.8 {d{{[0-9]+}}, d{{[0-9]+}}, d{{[0-9]+}}, d{{[0-9]+}}}, [r
0:64] | 8 ; CHECK-NEXT: vst4.{{[0-9]+}} {d{{[0-9]+}}, d{{[0-9]+}}, d{{[0-9]+}}, d{{[0-9
]+}}}, [r0:64] |
9 ret void | 9 ret void |
10 } | 10 } |
11 | 11 |
12 define void @vst4i16(i16* %A, <4 x i16>* %B) nounwind { | 12 define void @vst4i16(i16* %A, <4 x i16>* %B) nounwind { |
13 %tmp0 = bitcast i16* %A to i8* | 13 %tmp0 = bitcast i16* %A to i8* |
14 %tmp1 = load <4 x i16>* %B | 14 %tmp1 = load <4 x i16>, <4 x i16>* %B |
15 call void @llvm.arm.neon.vst4.v4i16(i8* %tmp0, <4 x i16> %tmp1, <4 x i16> %tmp
1, <4 x i16> %tmp1, <4 x i16> %tmp1, i32 16) | 15 call void @llvm.arm.neon.vst4.v4i16(i8* %tmp0, <4 x i16> %tmp1, <4 x i16> %tmp
1, <4 x i16> %tmp1, <4 x i16> %tmp1, i32 16) |
16 ; CHECK: bic r0, r0, #3221225472 | 16 ; CHECK: bic r0, r0, #-1073741824 |
17 ; CHECK-NEXT: vst4.16 {d{{[0-9]+}}, d{{[0-9]+}}, d{{[0-9]+}}, d{{[0-9]+}}}, [
r0:128] | 17 ; CHECK-NEXT: vst4.{{[0-9]+}} {d{{[0-9]+}}, d{{[0-9]+}}, d{{[0-9]+}}, d{{[0-9
]+}}}, [r0:128] |
18 ret void | 18 ret void |
19 } | 19 } |
20 | 20 |
21 define void @vst4i32(i32* %A, <2 x i32>* %B) nounwind { | 21 define void @vst4i32(i32* %A, <2 x i32>* %B) nounwind { |
22 %tmp0 = bitcast i32* %A to i8* | 22 %tmp0 = bitcast i32* %A to i8* |
23 %tmp1 = load <2 x i32>* %B | 23 %tmp1 = load <2 x i32>, <2 x i32>* %B |
24 call void @llvm.arm.neon.vst4.v2i32(i8* %tmp0, <2 x i32> %tmp1, <2 x i32> %tmp
1, <2 x i32> %tmp1, <2 x i32> %tmp1, i32 32) | 24 call void @llvm.arm.neon.vst4.v2i32(i8* %tmp0, <2 x i32> %tmp1, <2 x i32> %tmp
1, <2 x i32> %tmp1, <2 x i32> %tmp1, i32 32) |
25 ; CHECK: bic r0, r0, #3221225472 | 25 ; CHECK: bic r0, r0, #-1073741824 |
26 ; CHECK-NEXT: vst4.32 {d{{[0-9]+}}, d{{[0-9]+}}, d{{[0-9]+}}, d{{[0-9]+}}}, [
r0:256] | 26 ; CHECK-NEXT: vst4.{{[0-9]+}} {d{{[0-9]+}}, d{{[0-9]+}}, d{{[0-9]+}}, d{{[0-9
]+}}}, [r0:256] |
27 ret void | 27 ret void |
28 } | 28 } |
29 | 29 |
30 ;Check for a post-increment updating store. | 30 ;Check for a post-increment updating store. |
31 define void @vst4Qf_update(float** %ptr, <4 x float>* %B) nounwind { | 31 define void @vst4Qf_update(float** %ptr, <4 x float>* %B) nounwind { |
32 %A = load float** %ptr | 32 %A = load float*, float** %ptr |
33 %tmp0 = bitcast float* %A to i8* | 33 %tmp0 = bitcast float* %A to i8* |
34 %tmp1 = load <4 x float>* %B | 34 %tmp1 = load <4 x float>, <4 x float>* %B |
35 ; CHECK: bic r1, r1, #3221225472 | 35 ; CHECK: bic r1, r1, #-1073741824 |
36 ; CHECK-NEXT: vld1.32 {d{{[0-9]+}}, d{{[0-9]+}}}, [r1] | 36 ; CHECK-NEXT: vld1.{{[0-9]+}} {d{{[0-9]+}}, d{{[0-9]+}}}, [r1] |
37 call void @llvm.arm.neon.vst4.v4f32(i8* %tmp0, <4 x float> %tmp1, <4 x float>
%tmp1, <4 x float> %tmp1, <4 x float> %tmp1, i32 1) | 37 call void @llvm.arm.neon.vst4.v4f32(i8* %tmp0, <4 x float> %tmp1, <4 x float>
%tmp1, <4 x float> %tmp1, <4 x float> %tmp1, i32 1) |
38 ; CHECK: bic r1, r1, #3221225472 | 38 ; CHECK: bic r1, r1, #-1073741824 |
39 ; CHECK-NEXT: vst4.32 {d{{[0-9]+}}, d{{[0-9]+}}, d{{[0-9]+}}, d{{[0-9]+}}}, [
r1]! | 39 ; CHECK-NEXT: vst4.{{[0-9]+}} {d{{[0-9]+}}, d{{[0-9]+}}, d{{[0-9]+}}, d{{[0-9
]+}}}, [r1]! |
40 %tmp2 = getelementptr float* %A, i32 16 | 40 %tmp2 = getelementptr float, float* %A, i32 16 |
41 store float* %tmp2, float** %ptr | 41 store float* %tmp2, float** %ptr |
42 ret void | 42 ret void |
43 } | 43 } |
44 | 44 |
45 declare void @llvm.arm.neon.vst4.v8i8(i8*, <8 x i8>, <8 x i8>, <8 x i8>, <8 x i8
>, i32) nounwind | 45 declare void @llvm.arm.neon.vst4.v8i8(i8*, <8 x i8>, <8 x i8>, <8 x i8>, <8 x i8
>, i32) nounwind |
46 declare void @llvm.arm.neon.vst4.v4i16(i8*, <4 x i16>, <4 x i16>, <4 x i16>, <4
x i16>, i32) nounwind | 46 declare void @llvm.arm.neon.vst4.v4i16(i8*, <4 x i16>, <4 x i16>, <4 x i16>, <4
x i16>, i32) nounwind |
47 declare void @llvm.arm.neon.vst4.v2i32(i8*, <2 x i32>, <2 x i32>, <2 x i32>, <2
x i32>, i32) nounwind | 47 declare void @llvm.arm.neon.vst4.v2i32(i8*, <2 x i32>, <2 x i32>, <2 x i32>, <2
x i32>, i32) nounwind |
48 declare void @llvm.arm.neon.vst4.v2f32(i8*, <2 x float>, <2 x float>, <2 x float
>, <2 x float>, i32) nounwind | 48 declare void @llvm.arm.neon.vst4.v2f32(i8*, <2 x float>, <2 x float>, <2 x float
>, <2 x float>, i32) nounwind |
49 declare void @llvm.arm.neon.vst4.v1i64(i8*, <1 x i64>, <1 x i64>, <1 x i64>, <1
x i64>, i32) nounwind | 49 declare void @llvm.arm.neon.vst4.v1i64(i8*, <1 x i64>, <1 x i64>, <1 x i64>, <1
x i64>, i32) nounwind |
50 | 50 |
51 declare void @llvm.arm.neon.vst4.v16i8(i8*, <16 x i8>, <16 x i8>, <16 x i8>, <16
x i8>, i32) nounwind | 51 declare void @llvm.arm.neon.vst4.v16i8(i8*, <16 x i8>, <16 x i8>, <16 x i8>, <16
x i8>, i32) nounwind |
52 declare void @llvm.arm.neon.vst4.v8i16(i8*, <8 x i16>, <8 x i16>, <8 x i16>, <8
x i16>, i32) nounwind | 52 declare void @llvm.arm.neon.vst4.v8i16(i8*, <8 x i16>, <8 x i16>, <8 x i16>, <8
x i16>, i32) nounwind |
53 declare void @llvm.arm.neon.vst4.v4i32(i8*, <4 x i32>, <4 x i32>, <4 x i32>, <4
x i32>, i32) nounwind | 53 declare void @llvm.arm.neon.vst4.v4i32(i8*, <4 x i32>, <4 x i32>, <4 x i32>, <4
x i32>, i32) nounwind |
54 declare void @llvm.arm.neon.vst4.v4f32(i8*, <4 x float>, <4 x float>, <4 x float
>, <4 x float>, i32) nounwind | 54 declare void @llvm.arm.neon.vst4.v4f32(i8*, <4 x float>, <4 x float>, <4 x float
>, <4 x float>, i32) nounwind |
55 | 55 |
OLD | NEW |