| Index: tests_lit/assembler/arm32/sandboxing.ll
|
| diff --git a/tests_lit/assembler/arm32/sandboxing.ll b/tests_lit/assembler/arm32/sandboxing.ll
|
| deleted file mode 100644
|
| index bf99351293b0f531505d45fc33d63f70d95b2ee1..0000000000000000000000000000000000000000
|
| --- a/tests_lit/assembler/arm32/sandboxing.ll
|
| +++ /dev/null
|
| @@ -1,149 +0,0 @@
|
| -; Tests basics and corner cases of x86-32 sandboxing, using -Om1 in
|
| -; the hope that the output will remain stable. When packing bundles,
|
| -; we try to limit to a few instructions with well known sizes and
|
| -; minimal use of registers and stack slots in the lowering sequence.
|
| -
|
| -; RUN: %p2i -i %s --sandbox --filetype=asm --target=arm32 --assemble \
|
| -; RUN: --disassemble --args -Om1 -allow-externally-defined-symbols \
|
| -; RUN: -ffunction-sections | FileCheck %s
|
| -
|
| -declare void @call_target()
|
| -@global_short = internal global [2 x i8] zeroinitializer
|
| -
|
| -; A direct call sequence uses the right mask and register-call sequence.
|
| -define internal void @test_direct_call() {
|
| -entry:
|
| - call void @call_target()
|
| - ret void
|
| -}
|
| -; CHECK-LABEL: test_direct_call
|
| -; CHECK: nop
|
| -; CHECK: c: {{.*}} bl {{.*}} call_target
|
| -; CHECK-NEXT: 10:
|
| -
|
| -; An indirect call sequence uses the right mask and register-call sequence.
|
| -define internal void @test_indirect_call(i32 %target) {
|
| -entry:
|
| - %__1 = inttoptr i32 %target to void ()*
|
| - call void %__1()
|
| - ret void
|
| -}
|
| -; CHECK-LABEL: test_indirect_call
|
| -; CHECK: ldr [[REG:.*]], [sp,
|
| -; CHECK-NEXT: nop
|
| -; CHECK-NEXT: nop
|
| -; CHECK: 18: {{.*}} bic [[REG]], [[REG]], {{.*}} 0xc000000f
|
| -; CHECK-NEXT: blx [[REG]]
|
| -; CHECk-NEXT: 20:
|
| -
|
| -; A return sequences uses the right pop / mask / jmp sequence.
|
| -define internal void @test_ret() {
|
| -entry:
|
| - ret void
|
| -}
|
| -; CHECK-LABEL: test_ret
|
| -; CHECK: 0: {{.*}} bic lr, lr, {{.*}} 0xc000000f
|
| -; CHECK-NEXT: bx lr
|
| -
|
| -; Bundle lock without padding.
|
| -define internal void @bundle_lock_without_padding() {
|
| -entry:
|
| - %addr_short = bitcast [2 x i8]* @global_short to i16*
|
| - store i16 0, i16* %addr_short, align 1
|
| - ret void
|
| -}
|
| -; CHECK-LABEL: bundle_lock_without_padding
|
| -; CHECK: 0: {{.*}} movw
|
| -; CHECK-NEXT: movt
|
| -; CHECK-NEXT: movw
|
| -; CHECK-NEXT: strh
|
| -; CHECK-NEXT: bic lr, lr, {{.*}} 0xc000000f
|
| -; CHECK-NEXT: {{.*}} bx lr
|
| -
|
| -; Bundle lock with padding.
|
| -define internal void @bundle_lock_with_padding() {
|
| -entry:
|
| - call void @call_target()
|
| - ; bundle boundary
|
| - store i16 0, i16* undef, align 1 ; 3 insts
|
| - store i16 0, i16* undef, align 1 ; 3 insts
|
| - store i16 0, i16* undef, align 1 ; 3 insts
|
| - ; SP adjustment + pop
|
| - ; nop
|
| - ; bundle boundary
|
| - ret void
|
| -}
|
| -; CHECK-LABEL: bundle_lock_with_padding
|
| -; CHECK: 38: {{.*}} pop
|
| -; CHECK-NEXT: nop
|
| -; CHECK-NEXT: bic lr, lr, {{.*}} 0xc000000f
|
| -; CHECK-NEXT: {{.*}} bx lr
|
| -
|
| -; Bundle lock align_to_end without any padding.
|
| -define internal void @bundle_lock_align_to_end_padding_0() {
|
| -entry:
|
| - call void @call_target()
|
| - ; bundle boundary
|
| - store i16 0, i16* undef, align 1
|
| - call void @call_target()
|
| - ; bundle boundary
|
| - ret void
|
| -}
|
| -; CHECK-LABEL: bundle_lock_align_to_end_padding_0
|
| -; CHECK: c: {{.*}} bl {{.*}} call_target
|
| -; CHECK-NEXT: movw
|
| -; CHECK-NEXT: movw
|
| -; CHECK-NEXT: strh
|
| -; CHECK-NEXT: bl {{.*}} call_target
|
| -; CHECK-NEXT: add
|
| -; CHECK-NEXT: pop
|
| -; CHECK-NEXT: bic lr, lr, {{.*}} 0xc000000f
|
| -; CHECK-NEXT: {{.*}} bx lr
|
| -
|
| -; Bundle lock align_to_end with one bunch of padding.
|
| -define internal void @bundle_lock_align_to_end_padding_1() {
|
| -entry:
|
| - call void @call_target()
|
| - ; bundle boundary
|
| - store i16 0, i16* undef, align 1
|
| - store i16 0, i16* undef, align 1
|
| - ; bundle boundary
|
| - call void @call_target()
|
| - ; bundle boundary
|
| - ret void
|
| -}
|
| -; CHECK-LABEL: bundle_lock_align_to_end_padding_1
|
| -; CHECK: c: {{.*}} bl {{.*}} call_target
|
| -; CHECK-NEXT: movw
|
| -; CHECK-NEXT: movw
|
| -; CHECK-NEXT: strh
|
| -; CHECK-NEXT: movw
|
| -; CHECK-NEXT: movw
|
| -; CHECK-NEXT: strh
|
| -; CHECK-NEXT: nop
|
| -; CHECK-NEXT: bl {{.*}} call_target
|
| -; CHECK-NEXT: add
|
| -; CHECK-NEXT: pop
|
| -; CHECK-NEXT: bic lr, lr, {{.*}} 0xc000000f
|
| -; CHECK-NEXT: {{.*}} bx lr
|
| -
|
| -; Bundle lock align_to_end with two bunches of padding.
|
| -define internal void @bundle_lock_align_to_end_padding_2(i32 %target) {
|
| -entry:
|
| - call void @call_target()
|
| - ; bundle boundary
|
| - %__1 = inttoptr i32 %target to void ()*
|
| - store i8 0, i8* undef, align 1
|
| - call void %__1()
|
| - ret void
|
| -}
|
| -; CHECK-LABEL: bundle_lock_align_to_end_padding_2
|
| -; CHECK: c: {{.*}} bl {{.*}} call_target
|
| -; CHECK-NEXT: movw
|
| -; CHECK-NEXT: movw
|
| -; CHECK-NEXT: strb
|
| -; CHECK: 20: {{.*}} nop
|
| -; CHECK-NEXT: nop
|
| -; CHECK-NEXT: bic [[REG:r[0-9]+]], [[REG]], {{.*}} 0xc000000f
|
| -; CHECK-NEXT: {{.*}} blx [[REG]]
|
| -
|
|
|