| Index: tests_lit/assembler/arm32/sandboxing.ll
|
| diff --git a/tests_lit/assembler/arm32/sandboxing.ll b/tests_lit/assembler/arm32/sandboxing.ll
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..bf99351293b0f531505d45fc33d63f70d95b2ee1
|
| --- /dev/null
|
| +++ b/tests_lit/assembler/arm32/sandboxing.ll
|
| @@ -0,0 +1,149 @@
|
| +; Tests basics and corner cases of x86-32 sandboxing, using -Om1 in
|
| +; the hope that the output will remain stable. When packing bundles,
|
| +; we try to limit to a few instructions with well known sizes and
|
| +; minimal use of registers and stack slots in the lowering sequence.
|
| +
|
| +; RUN: %p2i -i %s --sandbox --filetype=asm --target=arm32 --assemble \
|
| +; RUN: --disassemble --args -Om1 -allow-externally-defined-symbols \
|
| +; RUN: -ffunction-sections | FileCheck %s
|
| +
|
| +declare void @call_target()
|
| +@global_short = internal global [2 x i8] zeroinitializer
|
| +
|
| +; A direct call sequence uses the right mask and register-call sequence.
|
| +define internal void @test_direct_call() {
|
| +entry:
|
| + call void @call_target()
|
| + ret void
|
| +}
|
| +; CHECK-LABEL: test_direct_call
|
| +; CHECK: nop
|
| +; CHECK: c: {{.*}} bl {{.*}} call_target
|
| +; CHECK-NEXT: 10:
|
| +
|
| +; An indirect call sequence uses the right mask and register-call sequence.
|
| +define internal void @test_indirect_call(i32 %target) {
|
| +entry:
|
| + %__1 = inttoptr i32 %target to void ()*
|
| + call void %__1()
|
| + ret void
|
| +}
|
| +; CHECK-LABEL: test_indirect_call
|
| +; CHECK: ldr [[REG:.*]], [sp,
|
| +; CHECK-NEXT: nop
|
| +; CHECK-NEXT: nop
|
| +; CHECK: 18: {{.*}} bic [[REG]], [[REG]], {{.*}} 0xc000000f
|
| +; CHECK-NEXT: blx [[REG]]
|
| +; CHECk-NEXT: 20:
|
| +
|
| +; A return sequences uses the right pop / mask / jmp sequence.
|
| +define internal void @test_ret() {
|
| +entry:
|
| + ret void
|
| +}
|
| +; CHECK-LABEL: test_ret
|
| +; CHECK: 0: {{.*}} bic lr, lr, {{.*}} 0xc000000f
|
| +; CHECK-NEXT: bx lr
|
| +
|
| +; Bundle lock without padding.
|
| +define internal void @bundle_lock_without_padding() {
|
| +entry:
|
| + %addr_short = bitcast [2 x i8]* @global_short to i16*
|
| + store i16 0, i16* %addr_short, align 1
|
| + ret void
|
| +}
|
| +; CHECK-LABEL: bundle_lock_without_padding
|
| +; CHECK: 0: {{.*}} movw
|
| +; CHECK-NEXT: movt
|
| +; CHECK-NEXT: movw
|
| +; CHECK-NEXT: strh
|
| +; CHECK-NEXT: bic lr, lr, {{.*}} 0xc000000f
|
| +; CHECK-NEXT: {{.*}} bx lr
|
| +
|
| +; Bundle lock with padding.
|
| +define internal void @bundle_lock_with_padding() {
|
| +entry:
|
| + call void @call_target()
|
| + ; bundle boundary
|
| + store i16 0, i16* undef, align 1 ; 3 insts
|
| + store i16 0, i16* undef, align 1 ; 3 insts
|
| + store i16 0, i16* undef, align 1 ; 3 insts
|
| + ; SP adjustment + pop
|
| + ; nop
|
| + ; bundle boundary
|
| + ret void
|
| +}
|
| +; CHECK-LABEL: bundle_lock_with_padding
|
| +; CHECK: 38: {{.*}} pop
|
| +; CHECK-NEXT: nop
|
| +; CHECK-NEXT: bic lr, lr, {{.*}} 0xc000000f
|
| +; CHECK-NEXT: {{.*}} bx lr
|
| +
|
| +; Bundle lock align_to_end without any padding.
|
| +define internal void @bundle_lock_align_to_end_padding_0() {
|
| +entry:
|
| + call void @call_target()
|
| + ; bundle boundary
|
| + store i16 0, i16* undef, align 1
|
| + call void @call_target()
|
| + ; bundle boundary
|
| + ret void
|
| +}
|
| +; CHECK-LABEL: bundle_lock_align_to_end_padding_0
|
| +; CHECK: c: {{.*}} bl {{.*}} call_target
|
| +; CHECK-NEXT: movw
|
| +; CHECK-NEXT: movw
|
| +; CHECK-NEXT: strh
|
| +; CHECK-NEXT: bl {{.*}} call_target
|
| +; CHECK-NEXT: add
|
| +; CHECK-NEXT: pop
|
| +; CHECK-NEXT: bic lr, lr, {{.*}} 0xc000000f
|
| +; CHECK-NEXT: {{.*}} bx lr
|
| +
|
| +; Bundle lock align_to_end with one bunch of padding.
|
| +define internal void @bundle_lock_align_to_end_padding_1() {
|
| +entry:
|
| + call void @call_target()
|
| + ; bundle boundary
|
| + store i16 0, i16* undef, align 1
|
| + store i16 0, i16* undef, align 1
|
| + ; bundle boundary
|
| + call void @call_target()
|
| + ; bundle boundary
|
| + ret void
|
| +}
|
| +; CHECK-LABEL: bundle_lock_align_to_end_padding_1
|
| +; CHECK: c: {{.*}} bl {{.*}} call_target
|
| +; CHECK-NEXT: movw
|
| +; CHECK-NEXT: movw
|
| +; CHECK-NEXT: strh
|
| +; CHECK-NEXT: movw
|
| +; CHECK-NEXT: movw
|
| +; CHECK-NEXT: strh
|
| +; CHECK-NEXT: nop
|
| +; CHECK-NEXT: bl {{.*}} call_target
|
| +; CHECK-NEXT: add
|
| +; CHECK-NEXT: pop
|
| +; CHECK-NEXT: bic lr, lr, {{.*}} 0xc000000f
|
| +; CHECK-NEXT: {{.*}} bx lr
|
| +
|
| +; Bundle lock align_to_end with two bunches of padding.
|
| +define internal void @bundle_lock_align_to_end_padding_2(i32 %target) {
|
| +entry:
|
| + call void @call_target()
|
| + ; bundle boundary
|
| + %__1 = inttoptr i32 %target to void ()*
|
| + store i8 0, i8* undef, align 1
|
| + call void %__1()
|
| + ret void
|
| +}
|
| +; CHECK-LABEL: bundle_lock_align_to_end_padding_2
|
| +; CHECK: c: {{.*}} bl {{.*}} call_target
|
| +; CHECK-NEXT: movw
|
| +; CHECK-NEXT: movw
|
| +; CHECK-NEXT: strb
|
| +; CHECK: 20: {{.*}} nop
|
| +; CHECK-NEXT: nop
|
| +; CHECK-NEXT: bic [[REG:r[0-9]+]], [[REG]], {{.*}} 0xc000000f
|
| +; CHECK-NEXT: {{.*}} blx [[REG]]
|
| +
|
|
|