OLD | NEW |
1 ; RUN: opt %s -expand-getelementptr -replace-ptrs-with-ints \ | 1 ; RUN: opt %s -expand-getelementptr -replace-ptrs-with-ints \ |
2 ; RUN: -minsfi-sandbox-memory-accesses -S \ | 2 ; RUN: -minsfi-sandbox-memory-accesses -S \ |
3 ; RUN: | FileCheck %s -check-prefix=CHECK-GEP | 3 ; RUN: | FileCheck %s -check-prefix=CHECK-GEP |
4 ; RUN: opt %s -expand-getelementptr -replace-ptrs-with-ints \ | 4 ; RUN: opt %s -expand-getelementptr -replace-ptrs-with-ints \ |
5 ; RUN: -minsfi-sandbox-memory-accesses -minsfi-ptrsize=20 -S \ | 5 ; RUN: -minsfi-sandbox-memory-accesses -minsfi-ptrsize=20 -S \ |
6 ; RUN: | FileCheck %s -check-prefix=CHECK-GEP-MASK | 6 ; RUN: | FileCheck %s -check-prefix=CHECK-GEP-MASK |
7 ; RUN: opt %s -expand-getelementptr -minsfi-sandbox-memory-accesses \ | 7 ; RUN: opt %s -expand-getelementptr -minsfi-sandbox-memory-accesses \ |
8 ; RUN: -minsfi-ptrsize=20 -S \ | 8 ; RUN: -minsfi-ptrsize=20 -S \ |
9 ; RUN: | FileCheck %s | 9 ; RUN: | FileCheck %s |
10 | 10 |
11 target datalayout = "p:32:32:32" | 11 target datalayout = "p:32:32:32" |
12 target triple = "le32-unknown-nacl" | 12 target triple = "le32-unknown-nacl" |
13 | 13 |
14 declare void @llvm.memcpy.p0i8.p0i8.i32(i8* nocapture, i8* nocapture readonly, i
32, i32, i1) | 14 declare void @llvm.memcpy.p0i8.p0i8.i32(i8* nocapture, i8* nocapture readonly, i
32, i32, i1) |
15 declare void @llvm.memmove.p0i8.p0i8.i32(i8* nocapture, i8* nocapture readonly,
i32, i32, i1) | 15 declare void @llvm.memmove.p0i8.p0i8.i32(i8* nocapture, i8* nocapture readonly,
i32, i32, i1) |
16 declare void @llvm.memset.p0i8.i32(i8* nocapture, i8, i32, i32, i1) | 16 declare void @llvm.memset.p0i8.i32(i8* nocapture, i8, i32, i32, i1) |
17 | 17 |
18 ; This test verifies that the pass recognizes the pointer arithmetic pattern | 18 ; This test verifies that the pass recognizes the pointer arithmetic pattern |
19 ; produced by the ExpandGetElementPtr pass and that it emits a more efficient | 19 ; produced by the ExpandGetElementPtr pass and that it emits a more efficient |
20 ; address sandboxing than in the general case. | 20 ; address sandboxing than in the general case. |
21 | 21 |
22 define i32 @test_load_elementptr([100 x i32]* %foo) { | 22 define i32 @test_load_elementptr([100 x i32]* %foo) { |
23 %elem = getelementptr inbounds [100 x i32]* %foo, i32 0, i32 97 | 23 %elem = getelementptr inbounds [100 x i32], [100 x i32]* %foo, i32 0, i32 97 |
24 %val = load i32* %elem | 24 %val = load i32, i32* %elem |
25 ret i32 %val | 25 ret i32 %val |
26 } | 26 } |
27 | 27 |
28 ; CHECK-GEP-LABEL: define i32 @test_load_elementptr(i32 %foo) { | 28 ; CHECK-GEP-LABEL: define i32 @test_load_elementptr(i32 %foo) { |
29 ; CHECK-GEP-NEXT: %mem_base = load i64* @__sfi_memory_base | 29 ; CHECK-GEP-NEXT: %mem_base = load i64, i64* @__sfi_memory_base |
30 ; CHECK-GEP-NEXT: %1 = zext i32 %foo to i64 | 30 ; CHECK-GEP-NEXT: %1 = zext i32 %foo to i64 |
31 ; CHECK-GEP-NEXT: %2 = add i64 %mem_base, %1 | 31 ; CHECK-GEP-NEXT: %2 = add i64 %mem_base, %1 |
32 ; CHECK-GEP-NEXT: %3 = add i64 %2, 388 | 32 ; CHECK-GEP-NEXT: %3 = add i64 %2, 388 |
33 ; CHECK-GEP-NEXT: %4 = inttoptr i64 %3 to i32* | 33 ; CHECK-GEP-NEXT: %4 = inttoptr i64 %3 to i32* |
34 ; CHECK-GEP-NEXT: %val = load i32* %4 | 34 ; CHECK-GEP-NEXT: %val = load i32, i32* %4 |
35 ; CHECK-GEP-NEXT: ret i32 %val | 35 ; CHECK-GEP-NEXT: ret i32 %val |
36 ; CHECK-GEP-NEXT: } | 36 ; CHECK-GEP-NEXT: } |
37 | 37 |
38 ; CHECK-GEP-MASK-LABEL: define i32 @test_load_elementptr(i32 %foo) { | 38 ; CHECK-GEP-MASK-LABEL: define i32 @test_load_elementptr(i32 %foo) { |
39 ; CHECK-GEP-MASK-NEXT: %mem_base = load i64* @__sfi_memory_base | 39 ; CHECK-GEP-MASK-NEXT: %mem_base = load i64, i64* @__sfi_memory_base |
40 ; CHECK-GEP-MASK-NEXT: %1 = and i32 %foo, 1048575 | 40 ; CHECK-GEP-MASK-NEXT: %1 = and i32 %foo, 1048575 |
41 ; CHECK-GEP-MASK-NEXT: %2 = zext i32 %1 to i64 | 41 ; CHECK-GEP-MASK-NEXT: %2 = zext i32 %1 to i64 |
42 ; CHECK-GEP-MASK-NEXT: %3 = add i64 %mem_base, %2 | 42 ; CHECK-GEP-MASK-NEXT: %3 = add i64 %mem_base, %2 |
43 ; CHECK-GEP-MASK-NEXT: %4 = add i64 %3, 388 | 43 ; CHECK-GEP-MASK-NEXT: %4 = add i64 %3, 388 |
44 ; CHECK-GEP-MASK-NEXT: %5 = inttoptr i64 %4 to i32* | 44 ; CHECK-GEP-MASK-NEXT: %5 = inttoptr i64 %4 to i32* |
45 ; CHECK-GEP-MASK-NEXT: %val = load i32* %5 | 45 ; CHECK-GEP-MASK-NEXT: %val = load i32, i32* %5 |
46 ; CHECK-GEP-MASK-NEXT: ret i32 %val | 46 ; CHECK-GEP-MASK-NEXT: ret i32 %val |
47 ; CHECK-GEP-MASK-NEXT: } | 47 ; CHECK-GEP-MASK-NEXT: } |
48 | 48 |
49 define <4 x float> @test_max_offset(i32 %x) { | 49 define <4 x float> @test_max_offset(i32 %x) { |
50 %1 = add i32 %x, 1048560 ; 1MB - 16B | 50 %1 = add i32 %x, 1048560 ; 1MB - 16B |
51 %ptr = inttoptr i32 %1 to <4 x float>* | 51 %ptr = inttoptr i32 %1 to <4 x float>* |
52 %val = load <4 x float>* %ptr | 52 %val = load <4 x float>, <4 x float>* %ptr |
53 ret <4 x float> %val | 53 ret <4 x float> %val |
54 } | 54 } |
55 | 55 |
56 ; CHECK-LABEL: define <4 x float> @test_max_offset(i32 %x) { | 56 ; CHECK-LABEL: define <4 x float> @test_max_offset(i32 %x) { |
57 ; CHECK-NEXT: %mem_base = load i64* @__sfi_memory_base | 57 ; CHECK-NEXT: %mem_base = load i64, i64* @__sfi_memory_base |
58 ; CHECK-NEXT: %1 = and i32 %x, 1048575 | 58 ; CHECK-NEXT: %1 = and i32 %x, 1048575 |
59 ; CHECK-NEXT: %2 = zext i32 %1 to i64 | 59 ; CHECK-NEXT: %2 = zext i32 %1 to i64 |
60 ; CHECK-NEXT: %3 = add i64 %mem_base, %2 | 60 ; CHECK-NEXT: %3 = add i64 %mem_base, %2 |
61 ; CHECK-NEXT: %4 = add i64 %3, 1048560 | 61 ; CHECK-NEXT: %4 = add i64 %3, 1048560 |
62 ; CHECK-NEXT: %5 = inttoptr i64 %4 to <4 x float>* | 62 ; CHECK-NEXT: %5 = inttoptr i64 %4 to <4 x float>* |
63 ; CHECK-NEXT: %val = load <4 x float>* %5 | 63 ; CHECK-NEXT: %val = load <4 x float>, <4 x float>* %5 |
64 ; CHECK-NEXT: ret <4 x float> %val | 64 ; CHECK-NEXT: ret <4 x float> %val |
65 ; CHECK-NEXT: } | 65 ; CHECK-NEXT: } |
66 | 66 |
67 ; This will not get optimized as it could access memory past the guard region. | 67 ; This will not get optimized as it could access memory past the guard region. |
68 define <4 x float> @test_offset_overflow(i32 %x) { | 68 define <4 x float> @test_offset_overflow(i32 %x) { |
69 %1 = add i32 %x, 1048561 | 69 %1 = add i32 %x, 1048561 |
70 %ptr = inttoptr i32 %1 to <4 x float>* | 70 %ptr = inttoptr i32 %1 to <4 x float>* |
71 %val = load <4 x float>* %ptr | 71 %val = load <4 x float>, <4 x float>* %ptr |
72 ret <4 x float> %val | 72 ret <4 x float> %val |
73 } | 73 } |
74 | 74 |
75 ; CHECK-LABEL: define <4 x float> @test_offset_overflow(i32 %x) { | 75 ; CHECK-LABEL: define <4 x float> @test_offset_overflow(i32 %x) { |
76 ; CHECK-NEXT: %mem_base = load i64* @__sfi_memory_base | 76 ; CHECK-NEXT: %mem_base = load i64, i64* @__sfi_memory_base |
77 ; CHECK-NEXT: %1 = add i32 %x, 1048561 | 77 ; CHECK-NEXT: %1 = add i32 %x, 1048561 |
78 ; CHECK-NEXT: %ptr = inttoptr i32 %1 to <4 x float>* | 78 ; CHECK-NEXT: %ptr = inttoptr i32 %1 to <4 x float>* |
79 ; CHECK-NEXT: %2 = ptrtoint <4 x float>* %ptr to i32 | 79 ; CHECK-NEXT: %2 = ptrtoint <4 x float>* %ptr to i32 |
80 ; CHECK-NEXT: %3 = and i32 %2, 1048575 | 80 ; CHECK-NEXT: %3 = and i32 %2, 1048575 |
81 ; CHECK-NEXT: %4 = zext i32 %3 to i64 | 81 ; CHECK-NEXT: %4 = zext i32 %3 to i64 |
82 ; CHECK-NEXT: %5 = add i64 %mem_base, %4 | 82 ; CHECK-NEXT: %5 = add i64 %mem_base, %4 |
83 ; CHECK-NEXT: %6 = inttoptr i64 %5 to <4 x float>* | 83 ; CHECK-NEXT: %6 = inttoptr i64 %5 to <4 x float>* |
84 ; CHECK-NEXT: %val = load <4 x float>* %6 | 84 ; CHECK-NEXT: %val = load <4 x float>, <4 x float>* %6 |
85 ; CHECK-NEXT: ret <4 x float> %val | 85 ; CHECK-NEXT: ret <4 x float> %val |
86 ; CHECK-NEXT: } | 86 ; CHECK-NEXT: } |
87 | 87 |
88 define void @test_not_applied_on_memcpy(i32 %x) { | 88 define void @test_not_applied_on_memcpy(i32 %x) { |
89 %1 = add i32 %x, 1024 | 89 %1 = add i32 %x, 1024 |
90 %ptr = inttoptr i32 %1 to i8* | 90 %ptr = inttoptr i32 %1 to i8* |
91 call void @llvm.memcpy.p0i8.p0i8.i32(i8* %ptr, i8* %ptr, i32 2048, i32 4, i1 f
alse); | 91 call void @llvm.memcpy.p0i8.p0i8.i32(i8* %ptr, i8* %ptr, i32 2048, i32 4, i1 f
alse); |
92 ret void | 92 ret void |
93 } | 93 } |
94 | 94 |
(...skipping 37 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
132 ret void | 132 ret void |
133 } | 133 } |
134 | 134 |
135 ; CHECK-LABEL: define void @test_not_applied_on_memset(i32 %x) { | 135 ; CHECK-LABEL: define void @test_not_applied_on_memset(i32 %x) { |
136 ; CHECK: [[IPTR:%[0-9]+]] = ptrtoint i8* %ptr to i32 | 136 ; CHECK: [[IPTR:%[0-9]+]] = ptrtoint i8* %ptr to i32 |
137 ; CHECK-NEXT: [[AND:%[0-9]+]] = and i32 [[IPTR]], 1048575 | 137 ; CHECK-NEXT: [[AND:%[0-9]+]] = and i32 [[IPTR]], 1048575 |
138 ; CHECK-NEXT: [[ZEXT:%[0-9]+]] = zext i32 [[AND]] to i64 | 138 ; CHECK-NEXT: [[ZEXT:%[0-9]+]] = zext i32 [[AND]] to i64 |
139 ; CHECK-NEXT: [[BASE:%[0-9]+]] = add i64 %mem_base, [[ZEXT]] | 139 ; CHECK-NEXT: [[BASE:%[0-9]+]] = add i64 %mem_base, [[ZEXT]] |
140 ; CHECK-NEXT: inttoptr i64 [[BASE]] to i8* | 140 ; CHECK-NEXT: inttoptr i64 [[BASE]] to i8* |
141 ; CHECK: call void @llvm.memset.p0i8.i32 | 141 ; CHECK: call void @llvm.memset.p0i8.i32 |
OLD | NEW |