OLD | NEW |
1 ; Test that direct loads and stores of local variables are not checked. | 1 ; Test that direct loads and stores of local variables are not checked. |
| 2 ; Also test that redundant checks of the same variable are elided. |
2 | 3 |
3 ; REQUIRES: allow_dump | 4 ; REQUIRES: allow_dump |
4 | 5 |
5 ; RUN: %p2i -i %s --args -verbose=inst -threads=0 -fsanitize-address \ | 6 ; RUN: %p2i -i %s --args -verbose=inst -threads=0 -fsanitize-address \ |
6 ; RUN: | FileCheck --check-prefix=DUMP %s | 7 ; RUN: | FileCheck --check-prefix=DUMP %s |
7 | 8 |
8 define internal void @foo() { | 9 define internal void @foo() { |
9 %ptr8 = alloca i8, i32 1, align 4 | 10 %ptr8 = alloca i8, i32 1, align 4 |
10 %ptr16 = alloca i8, i32 2, align 4 | 11 %ptr16 = alloca i8, i32 2, align 4 |
11 %ptr32 = alloca i8, i32 4, align 4 | 12 %ptr32 = alloca i8, i32 4, align 4 |
(...skipping 31 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
43 %off32 = add i32 %addr32, -1 | 44 %off32 = add i32 %addr32, -1 |
44 %off64 = add i32 %addr64, -1 | 45 %off64 = add i32 %addr64, -1 |
45 %off128 = add i32 %addr128, -1 | 46 %off128 = add i32 %addr128, -1 |
46 | 47 |
47 %offtarget8 = inttoptr i32 %off8 to i8* | 48 %offtarget8 = inttoptr i32 %off8 to i8* |
48 %offtarget16 = inttoptr i32 %off16 to i16* | 49 %offtarget16 = inttoptr i32 %off16 to i16* |
49 %offtarget32 = inttoptr i32 %off32 to i32* | 50 %offtarget32 = inttoptr i32 %off32 to i32* |
50 %offtarget64 = inttoptr i32 %off64 to i64* | 51 %offtarget64 = inttoptr i32 %off64 to i64* |
51 %offtarget128 = inttoptr i32 %off128 to <4 x i32>* | 52 %offtarget128 = inttoptr i32 %off128 to <4 x i32>* |
52 | 53 |
| 54 ; checked stores |
| 55 store i8 42, i8* %offtarget8, align 1 |
| 56 store i16 42, i16* %offtarget16, align 1 |
| 57 store i32 42, i32* %offtarget32, align 1 |
| 58 |
53 ; checked loads | 59 ; checked loads |
| 60 %offloaded64 = load i64, i64* %offtarget64, align 1 |
| 61 %offloaded128 = load <4 x i32>, <4 x i32>* %offtarget128, align 4 |
| 62 |
| 63 ; loads and stores with elided redundant checks |
54 %offloaded8 = load i8, i8* %offtarget8, align 1 | 64 %offloaded8 = load i8, i8* %offtarget8, align 1 |
55 %offloaded16 = load i16, i16* %offtarget16, align 1 | 65 %offloaded16 = load i16, i16* %offtarget16, align 1 |
56 %offloaded32 = load i32, i32* %offtarget32, align 1 | 66 %offloaded32 = load i32, i32* %offtarget32, align 1 |
57 %offloaded64 = load i64, i64* %offtarget64, align 1 | |
58 %offloaded128 = load <4 x i32>, <4 x i32>* %offtarget128, align 4 | |
59 | |
60 ; checked stores | |
61 store i8 %offloaded8, i8* %offtarget8, align 1 | |
62 store i16 %offloaded16, i16* %offtarget16, align 1 | |
63 store i32 %offloaded32, i32* %offtarget32, align 1 | |
64 store i64 %offloaded64, i64* %offtarget64, align 1 | 67 store i64 %offloaded64, i64* %offtarget64, align 1 |
65 store <4 x i32> %offloaded128, <4 x i32>* %offtarget128, align 4 | 68 store <4 x i32> %offloaded128, <4 x i32>* %offtarget128, align 4 |
66 | 69 |
67 ret void | 70 ret void |
68 } | 71 } |
69 | 72 |
70 ; DUMP-LABEL: ================ Instrumented CFG ================ | 73 ; DUMP-LABEL: ================ Instrumented CFG ================ |
71 ; DUMP-NEXT: define internal void @foo() { | 74 ; DUMP-NEXT: define internal void @foo() { |
72 | 75 |
73 ; Unchecked loads and stores | 76 ; Direct unchecked loads and stores |
74 ; DUMP: %loaded8 = load i8, i8* %ptr8, align 1 | 77 ; DUMP: %loaded8 = load i8, i8* %ptr8, align 1 |
75 ; DUMP-NEXT: %loaded16 = load i16, i16* %ptr16, align 1 | 78 ; DUMP-NEXT: %loaded16 = load i16, i16* %ptr16, align 1 |
76 ; DUMP-NEXT: %loaded32 = load i32, i32* %ptr32, align 1 | 79 ; DUMP-NEXT: %loaded32 = load i32, i32* %ptr32, align 1 |
77 ; DUMP-NEXT: %loaded64 = load i64, i64* %ptr64, align 1 | 80 ; DUMP-NEXT: %loaded64 = load i64, i64* %ptr64, align 1 |
78 ; DUMP-NEXT: %loaded128 = load <4 x i32>, <4 x i32>* %ptr128, align 4 | 81 ; DUMP-NEXT: %loaded128 = load <4 x i32>, <4 x i32>* %ptr128, align 4 |
79 ; DUMP-NEXT: store i8 %loaded8, i8* %ptr8, align 1 | 82 ; DUMP-NEXT: store i8 %loaded8, i8* %ptr8, align 1 |
80 ; DUMP-NEXT: store i16 %loaded16, i16* %ptr16, align 1 | 83 ; DUMP-NEXT: store i16 %loaded16, i16* %ptr16, align 1 |
81 ; DUMP-NEXT: store i32 %loaded32, i32* %ptr32, align 1 | 84 ; DUMP-NEXT: store i32 %loaded32, i32* %ptr32, align 1 |
82 ; DUMP-NEXT: store i64 %loaded64, i64* %ptr64, align 1 | 85 ; DUMP-NEXT: store i64 %loaded64, i64* %ptr64, align 1 |
83 ; DUMP-NEXT: store <4 x i32> %loaded128, <4 x i32>* %ptr128, align 4 | 86 ; DUMP-NEXT: store <4 x i32> %loaded128, <4 x i32>* %ptr128, align 4 |
84 | 87 |
85 ; Checked loads and stores | 88 ; Checked stores |
86 ; DUMP: call void @__asan_check_load(i32 %off8, i32 1) | 89 ; DUMP: call void @__asan_check_store(i32 %off8, i32 1) |
87 ; DUMP-NEXT: %offloaded8 = load i8, i8* %off8, align 1 | 90 ; DUMP-NEXT: store i8 42, i8* %off8, align 1 |
88 ; DUMP-NEXT: call void @__asan_check_load(i32 %off16, i32 2) | 91 ; DUMP-NEXT: call void @__asan_check_store(i32 %off16, i32 2) |
89 ; DUMP-NEXT: %offloaded16 = load i16, i16* %off16, align 1 | 92 ; DUMP-NEXT: store i16 42, i16* %off16, align 1 |
90 ; DUMP-NEXT: call void @__asan_check_load(i32 %off32, i32 4) | 93 ; DUMP-NEXT: call void @__asan_check_store(i32 %off32, i32 4) |
91 ; DUMP-NEXT: %offloaded32 = load i32, i32* %off32, align 1 | 94 ; DUMP-NEXT: store i32 42, i32* %off32, align 1 |
| 95 |
| 96 ; Checked loads |
92 ; DUMP-NEXT: call void @__asan_check_load(i32 %off64, i32 8) | 97 ; DUMP-NEXT: call void @__asan_check_load(i32 %off64, i32 8) |
93 ; DUMP-NEXT: %offloaded64 = load i64, i64* %off64, align 1 | 98 ; DUMP-NEXT: %offloaded64 = load i64, i64* %off64, align 1 |
94 ; DUMP-NEXT: call void @__asan_check_load(i32 %off128, i32 16) | 99 ; DUMP-NEXT: call void @__asan_check_load(i32 %off128, i32 16) |
95 ; DUMP-NEXT: %offloaded128 = load <4 x i32>, <4 x i32>* %off128, align 4 | 100 ; DUMP-NEXT: %offloaded128 = load <4 x i32>, <4 x i32>* %off128, align 4 |
96 ; DUMP-NEXT: call void @__asan_check_store(i32 %off8, i32 1) | 101 |
97 ; DUMP-NEXT: store i8 %offloaded8, i8* %off8, align 1 | 102 ; Loads and stores with elided redundant checks |
98 ; DUMP-NEXT: call void @__asan_check_store(i32 %off16, i32 2) | 103 ; DUMP-NEXT: %offloaded8 = load i8, i8* %off8, align 1 |
99 ; DUMP-NEXT: store i16 %offloaded16, i16* %off16, align 1 | 104 ; DUMP-NEXT: %offloaded16 = load i16, i16* %off16, align 1 |
100 ; DUMP-NEXT: call void @__asan_check_store(i32 %off32, i32 4) | 105 ; DUMP-NEXT: %offloaded32 = load i32, i32* %off32, align 1 |
101 ; DUMP-NEXT: store i32 %offloaded32, i32* %off32, align 1 | 106 ; DUMP-NEXT: store i64 %offloaded64, i64* %off64, align 1, beacon %offloaded64 |
102 ; DUMP-NEXT: call void @__asan_check_store(i32 %off64, i32 8) | 107 ; DUMP-NEXT: store <4 x i32> %offloaded128, <4 x i32>* %off128, align 4, beacon
%offloaded128 |
103 ; DUMP-NEXT: store i64 %offloaded64, i64* %off64, align 1 | |
104 ; DUMP-NEXT: call void @__asan_check_store(i32 %off128, i32 16) | |
105 ; DUMP-NEXT: store <4 x i32> %offloaded128, <4 x i32>* %off128, align 4 | |
OLD | NEW |