| OLD | NEW |
| 1 ; RUN: opt %s -expand-getelementptr -replace-ptrs-with-ints \ | 1 ; RUN: opt %s -expand-getelementptr -replace-ptrs-with-ints \ |
| 2 ; RUN: -minsfi-sandbox-memory-accesses -S \ | 2 ; RUN: -minsfi-sandbox-memory-accesses -S \ |
| 3 ; RUN: | FileCheck %s -check-prefix=CHECK-GEP | 3 ; RUN: | FileCheck %s -check-prefix=CHECK-GEP |
| 4 ; RUN: opt %s -expand-getelementptr -replace-ptrs-with-ints \ | 4 ; RUN: opt %s -expand-getelementptr -replace-ptrs-with-ints \ |
| 5 ; RUN: -minsfi-sandbox-memory-accesses -minsfi-ptrsize=20 -S \ | 5 ; RUN: -minsfi-sandbox-memory-accesses -minsfi-ptrsize=20 -S \ |
| 6 ; RUN: | FileCheck %s -check-prefix=CHECK-GEP-MASK | 6 ; RUN: | FileCheck %s -check-prefix=CHECK-GEP-MASK |
| 7 ; RUN: opt %s -expand-getelementptr -minsfi-sandbox-memory-accesses \ | 7 ; RUN: opt %s -expand-getelementptr -minsfi-sandbox-memory-accesses \ |
| 8 ; RUN: -minsfi-ptrsize=20 -S \ | 8 ; RUN: -minsfi-ptrsize=20 -S \ |
| 9 ; RUN: | FileCheck %s | 9 ; RUN: | FileCheck %s |
| 10 | 10 |
| 11 target datalayout = "p:32:32:32" | 11 target datalayout = "p:32:32:32" |
| 12 target triple = "le32-unknown-nacl" | 12 target triple = "le32-unknown-nacl" |
| 13 | 13 |
| 14 declare void @llvm.memcpy.p0i8.p0i8.i32(i8* nocapture, i8* nocapture readonly, i
32, i32, i1) | 14 declare void @llvm.memcpy.p0i8.p0i8.i32(i8* nocapture, i8* nocapture readonly, i
32, i32, i1) |
| 15 declare void @llvm.memmove.p0i8.p0i8.i32(i8* nocapture, i8* nocapture readonly,
i32, i32, i1) | 15 declare void @llvm.memmove.p0i8.p0i8.i32(i8* nocapture, i8* nocapture readonly,
i32, i32, i1) |
| 16 declare void @llvm.memset.p0i8.i32(i8* nocapture, i8, i32, i32, i1) | 16 declare void @llvm.memset.p0i8.i32(i8* nocapture, i8, i32, i32, i1) |
| 17 | 17 |
| 18 ; This test verifies that the pass recognizes the pointer arithmetic pattern | 18 ; This test verifies that the pass recognizes the pointer arithmetic pattern |
| 19 ; produced by the ExpandGetElementPtr pass and that it emits a more efficient | 19 ; produced by the ExpandGetElementPtr pass and that it emits a more efficient |
| 20 ; address sandboxing than in the general case. | 20 ; address sandboxing than in the general case. |
| 21 | 21 |
| 22 define i32 @test_load_elementptr([100 x i32]* %foo) { | 22 define i32 @test_load_elementptr([100 x i32]* %foo) { |
| 23 %elem = getelementptr inbounds [100 x i32]* %foo, i32 0, i32 97 | 23 %elem = getelementptr inbounds [100 x i32], [100 x i32]* %foo, i32 0, i32 97 |
| 24 %val = load i32* %elem | 24 %val = load i32, i32* %elem |
| 25 ret i32 %val | 25 ret i32 %val |
| 26 } | 26 } |
| 27 | 27 |
| 28 ; CHECK-GEP-LABEL: define i32 @test_load_elementptr(i32 %foo) { | 28 ; CHECK-GEP-LABEL: define i32 @test_load_elementptr(i32 %foo) { |
| 29 ; CHECK-GEP-NEXT: %mem_base = load i64* @__sfi_memory_base | 29 ; CHECK-GEP-NEXT: %mem_base = load i64, i64* @__sfi_memory_base |
| 30 ; CHECK-GEP-NEXT: %1 = zext i32 %foo to i64 | 30 ; CHECK-GEP-NEXT: %1 = zext i32 %foo to i64 |
| 31 ; CHECK-GEP-NEXT: %2 = add i64 %mem_base, %1 | 31 ; CHECK-GEP-NEXT: %2 = add i64 %mem_base, %1 |
| 32 ; CHECK-GEP-NEXT: %3 = add i64 %2, 388 | 32 ; CHECK-GEP-NEXT: %3 = add i64 %2, 388 |
| 33 ; CHECK-GEP-NEXT: %4 = inttoptr i64 %3 to i32* | 33 ; CHECK-GEP-NEXT: %4 = inttoptr i64 %3 to i32* |
| 34 ; CHECK-GEP-NEXT: %val = load i32* %4 | 34 ; CHECK-GEP-NEXT: %val = load i32, i32* %4 |
| 35 ; CHECK-GEP-NEXT: ret i32 %val | 35 ; CHECK-GEP-NEXT: ret i32 %val |
| 36 ; CHECK-GEP-NEXT: } | 36 ; CHECK-GEP-NEXT: } |
| 37 | 37 |
| 38 ; CHECK-GEP-MASK-LABEL: define i32 @test_load_elementptr(i32 %foo) { | 38 ; CHECK-GEP-MASK-LABEL: define i32 @test_load_elementptr(i32 %foo) { |
| 39 ; CHECK-GEP-MASK-NEXT: %mem_base = load i64* @__sfi_memory_base | 39 ; CHECK-GEP-MASK-NEXT: %mem_base = load i64, i64* @__sfi_memory_base |
| 40 ; CHECK-GEP-MASK-NEXT: %1 = and i32 %foo, 1048575 | 40 ; CHECK-GEP-MASK-NEXT: %1 = and i32 %foo, 1048575 |
| 41 ; CHECK-GEP-MASK-NEXT: %2 = zext i32 %1 to i64 | 41 ; CHECK-GEP-MASK-NEXT: %2 = zext i32 %1 to i64 |
| 42 ; CHECK-GEP-MASK-NEXT: %3 = add i64 %mem_base, %2 | 42 ; CHECK-GEP-MASK-NEXT: %3 = add i64 %mem_base, %2 |
| 43 ; CHECK-GEP-MASK-NEXT: %4 = add i64 %3, 388 | 43 ; CHECK-GEP-MASK-NEXT: %4 = add i64 %3, 388 |
| 44 ; CHECK-GEP-MASK-NEXT: %5 = inttoptr i64 %4 to i32* | 44 ; CHECK-GEP-MASK-NEXT: %5 = inttoptr i64 %4 to i32* |
| 45 ; CHECK-GEP-MASK-NEXT: %val = load i32* %5 | 45 ; CHECK-GEP-MASK-NEXT: %val = load i32, i32* %5 |
| 46 ; CHECK-GEP-MASK-NEXT: ret i32 %val | 46 ; CHECK-GEP-MASK-NEXT: ret i32 %val |
| 47 ; CHECK-GEP-MASK-NEXT: } | 47 ; CHECK-GEP-MASK-NEXT: } |
| 48 | 48 |
| 49 define <4 x float> @test_max_offset(i32 %x) { | 49 define <4 x float> @test_max_offset(i32 %x) { |
| 50 %1 = add i32 %x, 1048560 ; 1MB - 16B | 50 %1 = add i32 %x, 1048560 ; 1MB - 16B |
| 51 %ptr = inttoptr i32 %1 to <4 x float>* | 51 %ptr = inttoptr i32 %1 to <4 x float>* |
| 52 %val = load <4 x float>* %ptr | 52 %val = load <4 x float>, <4 x float>* %ptr |
| 53 ret <4 x float> %val | 53 ret <4 x float> %val |
| 54 } | 54 } |
| 55 | 55 |
| 56 ; CHECK-LABEL: define <4 x float> @test_max_offset(i32 %x) { | 56 ; CHECK-LABEL: define <4 x float> @test_max_offset(i32 %x) { |
| 57 ; CHECK-NEXT: %mem_base = load i64* @__sfi_memory_base | 57 ; CHECK-NEXT: %mem_base = load i64, i64* @__sfi_memory_base |
| 58 ; CHECK-NEXT: %1 = and i32 %x, 1048575 | 58 ; CHECK-NEXT: %1 = and i32 %x, 1048575 |
| 59 ; CHECK-NEXT: %2 = zext i32 %1 to i64 | 59 ; CHECK-NEXT: %2 = zext i32 %1 to i64 |
| 60 ; CHECK-NEXT: %3 = add i64 %mem_base, %2 | 60 ; CHECK-NEXT: %3 = add i64 %mem_base, %2 |
| 61 ; CHECK-NEXT: %4 = add i64 %3, 1048560 | 61 ; CHECK-NEXT: %4 = add i64 %3, 1048560 |
| 62 ; CHECK-NEXT: %5 = inttoptr i64 %4 to <4 x float>* | 62 ; CHECK-NEXT: %5 = inttoptr i64 %4 to <4 x float>* |
| 63 ; CHECK-NEXT: %val = load <4 x float>* %5 | 63 ; CHECK-NEXT: %val = load <4 x float>, <4 x float>* %5 |
| 64 ; CHECK-NEXT: ret <4 x float> %val | 64 ; CHECK-NEXT: ret <4 x float> %val |
| 65 ; CHECK-NEXT: } | 65 ; CHECK-NEXT: } |
| 66 | 66 |
| 67 ; This will not get optimized as it could access memory past the guard region. | 67 ; This will not get optimized as it could access memory past the guard region. |
| 68 define <4 x float> @test_offset_overflow(i32 %x) { | 68 define <4 x float> @test_offset_overflow(i32 %x) { |
| 69 %1 = add i32 %x, 1048561 | 69 %1 = add i32 %x, 1048561 |
| 70 %ptr = inttoptr i32 %1 to <4 x float>* | 70 %ptr = inttoptr i32 %1 to <4 x float>* |
| 71 %val = load <4 x float>* %ptr | 71 %val = load <4 x float>, <4 x float>* %ptr |
| 72 ret <4 x float> %val | 72 ret <4 x float> %val |
| 73 } | 73 } |
| 74 | 74 |
| 75 ; CHECK-LABEL: define <4 x float> @test_offset_overflow(i32 %x) { | 75 ; CHECK-LABEL: define <4 x float> @test_offset_overflow(i32 %x) { |
| 76 ; CHECK-NEXT: %mem_base = load i64* @__sfi_memory_base | 76 ; CHECK-NEXT: %mem_base = load i64, i64* @__sfi_memory_base |
| 77 ; CHECK-NEXT: %1 = add i32 %x, 1048561 | 77 ; CHECK-NEXT: %1 = add i32 %x, 1048561 |
| 78 ; CHECK-NEXT: %ptr = inttoptr i32 %1 to <4 x float>* | 78 ; CHECK-NEXT: %ptr = inttoptr i32 %1 to <4 x float>* |
| 79 ; CHECK-NEXT: %2 = ptrtoint <4 x float>* %ptr to i32 | 79 ; CHECK-NEXT: %2 = ptrtoint <4 x float>* %ptr to i32 |
| 80 ; CHECK-NEXT: %3 = and i32 %2, 1048575 | 80 ; CHECK-NEXT: %3 = and i32 %2, 1048575 |
| 81 ; CHECK-NEXT: %4 = zext i32 %3 to i64 | 81 ; CHECK-NEXT: %4 = zext i32 %3 to i64 |
| 82 ; CHECK-NEXT: %5 = add i64 %mem_base, %4 | 82 ; CHECK-NEXT: %5 = add i64 %mem_base, %4 |
| 83 ; CHECK-NEXT: %6 = inttoptr i64 %5 to <4 x float>* | 83 ; CHECK-NEXT: %6 = inttoptr i64 %5 to <4 x float>* |
| 84 ; CHECK-NEXT: %val = load <4 x float>* %6 | 84 ; CHECK-NEXT: %val = load <4 x float>, <4 x float>* %6 |
| 85 ; CHECK-NEXT: ret <4 x float> %val | 85 ; CHECK-NEXT: ret <4 x float> %val |
| 86 ; CHECK-NEXT: } | 86 ; CHECK-NEXT: } |
| 87 | 87 |
| 88 define void @test_not_applied_on_memcpy(i32 %x) { | 88 define void @test_not_applied_on_memcpy(i32 %x) { |
| 89 %1 = add i32 %x, 1024 | 89 %1 = add i32 %x, 1024 |
| 90 %ptr = inttoptr i32 %1 to i8* | 90 %ptr = inttoptr i32 %1 to i8* |
| 91 call void @llvm.memcpy.p0i8.p0i8.i32(i8* %ptr, i8* %ptr, i32 2048, i32 4, i1 f
alse); | 91 call void @llvm.memcpy.p0i8.p0i8.i32(i8* %ptr, i8* %ptr, i32 2048, i32 4, i1 f
alse); |
| 92 ret void | 92 ret void |
| 93 } | 93 } |
| 94 | 94 |
| (...skipping 37 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 132 ret void | 132 ret void |
| 133 } | 133 } |
| 134 | 134 |
| 135 ; CHECK-LABEL: define void @test_not_applied_on_memset(i32 %x) { | 135 ; CHECK-LABEL: define void @test_not_applied_on_memset(i32 %x) { |
| 136 ; CHECK: [[IPTR:%[0-9]+]] = ptrtoint i8* %ptr to i32 | 136 ; CHECK: [[IPTR:%[0-9]+]] = ptrtoint i8* %ptr to i32 |
| 137 ; CHECK-NEXT: [[AND:%[0-9]+]] = and i32 [[IPTR]], 1048575 | 137 ; CHECK-NEXT: [[AND:%[0-9]+]] = and i32 [[IPTR]], 1048575 |
| 138 ; CHECK-NEXT: [[ZEXT:%[0-9]+]] = zext i32 [[AND]] to i64 | 138 ; CHECK-NEXT: [[ZEXT:%[0-9]+]] = zext i32 [[AND]] to i64 |
| 139 ; CHECK-NEXT: [[BASE:%[0-9]+]] = add i64 %mem_base, [[ZEXT]] | 139 ; CHECK-NEXT: [[BASE:%[0-9]+]] = add i64 %mem_base, [[ZEXT]] |
| 140 ; CHECK-NEXT: inttoptr i64 [[BASE]] to i8* | 140 ; CHECK-NEXT: inttoptr i64 [[BASE]] to i8* |
| 141 ; CHECK: call void @llvm.memset.p0i8.i32 | 141 ; CHECK: call void @llvm.memset.p0i8.i32 |
| OLD | NEW |