Index: tests_lit/llvm2ice_tests/load_cast.ll |
diff --git a/tests_lit/llvm2ice_tests/load_cast.ll b/tests_lit/llvm2ice_tests/load_cast.ll |
new file mode 100644 |
index 0000000000000000000000000000000000000000..5395d04ff82f3ceef53a9494fc15d01d443411c6 |
--- /dev/null |
+++ b/tests_lit/llvm2ice_tests/load_cast.ll |
@@ -0,0 +1,266 @@ |
+; Tests desired and undesired folding of load instructions into cast |
+; instructions. The folding is only done when liveness analysis is performed, |
+; so only O2 is tested. |
+ |
+; RUN: %p2i --filetype=obj --disassemble -i %s --args -O2 | FileCheck %s |
+ |
+; Not testing trunc, or 32-bit bitcast, because the lowered code uses pretty |
+; much the same mov instructions regardless of whether folding is done. |
+ |
+define internal i32 @zext_fold(i32 %arg) { |
+entry: |
+ %ptr = add i32 %arg, 200 |
+ %addr = inttoptr i32 %ptr to i8* |
+ %load = load i8* %addr, align 1 |
+ %result = zext i8 %load to i32 |
+ ret i32 %result |
+} |
+; CHECK-LABEL: zext_fold |
+; CHECK: movzx {{.*}},BYTE PTR [{{.*}}+0xc8] |
+ |
+define internal i32 @zext_nofold(i32 %arg) { |
+entry: |
+ %ptr = add i32 %arg, 200 |
+ %addr = inttoptr i32 %ptr to i8* |
+ %load = load i8* %addr, align 1 |
+ %tmp1 = zext i8 %load to i32 |
+ %tmp2 = zext i8 %load to i32 |
+ %result = add i32 %tmp1, %tmp2 |
+ ret i32 %result |
+} |
+; Test that load folding does not happen. |
+; CHECK-LABEL: zext_nofold |
+; CHECK-NOT: movzx {{.*}},BYTE PTR [{{.*}}+0xc8] |
+ |
+define internal i32 @sext_fold(i32 %arg) { |
+entry: |
+ %ptr = add i32 %arg, 200 |
+ %addr = inttoptr i32 %ptr to i8* |
+ %load = load i8* %addr, align 1 |
+ %result = sext i8 %load to i32 |
+ ret i32 %result |
+} |
+; CHECK-LABEL: sext_fold |
+; CHECK: movsx {{.*}},BYTE PTR [{{.*}}+0xc8] |
+ |
+define internal i32 @sext_nofold(i32 %arg) { |
+entry: |
+ %ptr = add i32 %arg, 200 |
+ %addr = inttoptr i32 %ptr to i8* |
+ %load = load i8* %addr, align 1 |
+ %tmp1 = sext i8 %load to i32 |
+ %tmp2 = sext i8 %load to i32 |
+ %result = add i32 %tmp1, %tmp2 |
+ ret i32 %result |
+} |
+; Test that load folding does not happen. |
+; CHECK-LABEL: sext_nofold |
+; CHECK-NOT: movsx {{.*}},BYTE PTR [{{.*}}+0xc8] |
+ |
+define internal float @fptrunc_fold(i32 %arg) { |
+entry: |
+ %ptr = add i32 %arg, 200 |
+ %addr = inttoptr i32 %ptr to double* |
+ %load = load double* %addr, align 8 |
+ %result = fptrunc double %load to float |
+ ret float %result |
+} |
+; CHECK-LABEL: fptrunc_fold |
+; CHECK: cvtsd2ss {{.*}},QWORD PTR [{{.*}}+0xc8] |
+ |
+define internal float @fptrunc_nofold(i32 %arg) { |
+entry: |
+ %ptr = add i32 %arg, 200 |
+ %addr = inttoptr i32 %ptr to double* |
+ %load = load double* %addr, align 8 |
+ %tmp1 = fptrunc double %load to float |
+ %tmp2 = fptrunc double %load to float |
+ %result = fadd float %tmp1, %tmp2 |
+ ret float %result |
+} |
+; Test that load folding does not happen. |
+; CHECK-LABEL: fptrunc_nofold |
+; CHECK-NOT: cvtsd2ss {{.*}},QWORD PTR [{{.*}}+0xc8] |
+ |
+define internal double @fpext_fold(i32 %arg) { |
+entry: |
+ %ptr = add i32 %arg, 200 |
+ %addr = inttoptr i32 %ptr to float* |
+ %load = load float* %addr, align 4 |
+ %result = fpext float %load to double |
+ ret double %result |
+} |
+; CHECK-LABEL: fpext_fold |
+; CHECK: cvtss2sd {{.*}},DWORD PTR [{{.*}}+0xc8] |
+ |
+define internal double @fpext_nofold(i32 %arg) { |
+entry: |
+ %ptr = add i32 %arg, 200 |
+ %addr = inttoptr i32 %ptr to float* |
+ %load = load float* %addr, align 4 |
+ %tmp1 = fpext float %load to double |
+ %tmp2 = fpext float %load to double |
+ %result = fadd double %tmp1, %tmp2 |
+ ret double %result |
+} |
+; Test that load folding does not happen. |
+; CHECK-LABEL: fpext_nofold |
+; CHECK-NOT: cvtss2sd {{.*}},DWORD PTR [{{.*}}+0xc8] |
+ |
+define internal i32 @fptoui_fold(i32 %arg) { |
+entry: |
+ %ptr = add i32 %arg, 200 |
+ %addr = inttoptr i32 %ptr to double* |
+ %load = load double* %addr, align 8 |
+ %result = fptoui double %load to i16 |
+ %result2 = zext i16 %result to i32 |
+ ret i32 %result2 |
+} |
+; CHECK-LABEL: fptoui_fold |
+; CHECK: cvttsd2si {{.*}},QWORD PTR [{{.*}}+0xc8] |
+ |
+define internal i32 @fptoui_nofold(i32 %arg) { |
+entry: |
+ %ptr = add i32 %arg, 200 |
+ %addr = inttoptr i32 %ptr to double* |
+ %load = load double* %addr, align 8 |
+ %tmp1 = fptoui double %load to i16 |
+ %tmp2 = fptoui double %load to i16 |
+ %result = add i16 %tmp1, %tmp2 |
+ %result2 = zext i16 %result to i32 |
+ ret i32 %result2 |
+} |
+; Test that load folding does not happen. |
+; CHECK-LABEL: fptoui_nofold |
+; CHECK-NOT: cvttsd2si {{.*}},QWORD PTR [{{.*}}+0xc8] |
+ |
+define internal i32 @fptosi_fold(i32 %arg) { |
+entry: |
+ %ptr = add i32 %arg, 200 |
+ %addr = inttoptr i32 %ptr to double* |
+ %load = load double* %addr, align 8 |
+ %result = fptosi double %load to i16 |
+ %result2 = zext i16 %result to i32 |
+ ret i32 %result2 |
+} |
+; CHECK-LABEL: fptosi_fold |
+; CHECK: cvttsd2si {{.*}},QWORD PTR [{{.*}}+0xc8] |
+ |
+define internal i32 @fptosi_nofold(i32 %arg) { |
+entry: |
+ %ptr = add i32 %arg, 200 |
+ %addr = inttoptr i32 %ptr to double* |
+ %load = load double* %addr, align 8 |
+ %tmp1 = fptosi double %load to i16 |
+ %tmp2 = fptosi double %load to i16 |
+ %result = add i16 %tmp1, %tmp2 |
+ %result2 = zext i16 %result to i32 |
+ ret i32 %result2 |
+} |
+; Test that load folding does not happen. |
+; CHECK-LABEL: fptosi_nofold |
+; CHECK-NOT: cvttsd2si {{.*}},QWORD PTR [{{.*}}+0xc8] |
+ |
+define internal double @uitofp_fold(i32 %arg) { |
+entry: |
+ %ptr = add i32 %arg, 200 |
+ %addr = inttoptr i32 %ptr to i16* |
+ %load = load i16* %addr, align 1 |
+ %result = uitofp i16 %load to double |
+ ret double %result |
+} |
+; CHECK-LABEL: uitofp_fold |
+; CHECK: movzx {{.*}},WORD PTR [{{.*}}+0xc8] |
+ |
+define internal double @uitofp_nofold(i32 %arg) { |
+entry: |
+ %ptr = add i32 %arg, 200 |
+ %addr = inttoptr i32 %ptr to i16* |
+ %load = load i16* %addr, align 1 |
+ %tmp1 = uitofp i16 %load to double |
+ %tmp2 = uitofp i16 %load to double |
+ %result = fadd double %tmp1, %tmp2 |
+ ret double %result |
+} |
+; Test that load folding does not happen. |
+; CHECK-LABEL: uitofp_nofold |
+; CHECK-NOT: movzx {{.*}},WORD PTR [{{.*}}+0xc8] |
+ |
+define internal double @sitofp_fold(i32 %arg) { |
+entry: |
+ %ptr = add i32 %arg, 200 |
+ %addr = inttoptr i32 %ptr to i16* |
+ %load = load i16* %addr, align 1 |
+ %result = sitofp i16 %load to double |
+ ret double %result |
+} |
+; CHECK-LABEL: sitofp_fold |
+; CHECK: movsx {{.*}},WORD PTR [{{.*}}+0xc8] |
+ |
+define internal double @sitofp_nofold(i32 %arg) { |
+entry: |
+ %ptr = add i32 %arg, 200 |
+ %addr = inttoptr i32 %ptr to i16* |
+ %load = load i16* %addr, align 1 |
+ %tmp1 = sitofp i16 %load to double |
+ %tmp2 = sitofp i16 %load to double |
+ %result = fadd double %tmp1, %tmp2 |
+ ret double %result |
+} |
+; Test that load folding does not happen. |
+; CHECK-LABEL: sitofp_nofold |
+; CHECK-NOT: movsx {{.*}},WORD PTR [{{.*}}+0xc8] |
+ |
+define internal double @bitcast_i64_fold(i32 %arg) { |
+entry: |
+ %ptr = add i32 %arg, 200 |
+ %addr = inttoptr i32 %ptr to i64* |
+ %load = load i64* %addr, align 1 |
+ %result = bitcast i64 %load to double |
+ ret double %result |
+} |
+; CHECK-LABEL: bitcast_i64_fold |
+; CHECK: movq {{.*}},QWORD PTR [{{.*}}+0xc8] |
+ |
+define internal double @bitcast_i64_nofold(i32 %arg) { |
+entry: |
+ %ptr = add i32 %arg, 200 |
+ %addr = inttoptr i32 %ptr to i64* |
+ %load = load i64* %addr, align 1 |
+ %tmp1 = bitcast i64 %load to double |
+ %tmp2 = bitcast i64 %load to double |
+ %result = fadd double %tmp1, %tmp2 |
+ ret double %result |
+} |
+; Test that load folding does not happen. |
+; CHECK-LABEL: bitcast_i64_nofold |
+; CHECK-NOT: movq {{.*}},QWORD PTR [{{.*}}+0xc8] |
+ |
+define internal i64 @bitcast_double_fold(i32 %arg) { |
+entry: |
+ %ptr = add i32 %arg, 200 |
+ %addr = inttoptr i32 %ptr to double* |
+ %load = load double* %addr, align 8 |
+ %result = bitcast double %load to i64 |
+ ret i64 %result |
+} |
+; CHECK-LABEL: bitcast_double_fold |
+; CHECK-NOT: QWORD PTR |
+; CHECK: mov {{.*}},DWORD PTR [{{.*}}+0xc8] |
+; CHECK: mov {{.*}},DWORD PTR [{{.*}}+0xcc] |
+; CHECK-NOT: QWORD PTR |
+ |
+define internal i64 @bitcast_double_nofold(i32 %arg) { |
+entry: |
+ %ptr = add i32 %arg, 200 |
+ %addr = inttoptr i32 %ptr to double* |
+ %load = load double* %addr, align 8 |
+ %tmp1 = bitcast double %load to i64 |
+ %tmp2 = bitcast double %load to i64 |
+ %result = add i64 %tmp1, %tmp2 |
+ ret i64 %result |
+} |
+; Test that load folding does not happen. |
+; CHECK-LABEL: bitcast_double_nofold |
+; CHECK: QWORD PTR |
+; CHECK: QWORD PTR |