| Index: third_party/boringssl/linux-arm/crypto/sha/sha1-armv4-large.S
|
| diff --git a/third_party/boringssl/linux-arm/crypto/sha/sha1-armv4-large.S b/third_party/boringssl/linux-arm/crypto/sha/sha1-armv4-large.S
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..cc436446582a0c96b3c1fdb30a929bcc660c0d0c
|
| --- /dev/null
|
| +++ b/third_party/boringssl/linux-arm/crypto/sha/sha1-armv4-large.S
|
| @@ -0,0 +1,455 @@
|
| +#if defined(__arm__)
|
| +#include "arm_arch.h"
|
| +
|
| +.text
|
| +
|
| +.global sha1_block_data_order
|
| +.type sha1_block_data_order,%function
|
| +
|
| +.align 2
|
| +sha1_block_data_order:
|
| + stmdb sp!,{r4-r12,lr}
|
| + add r2,r1,r2,lsl#6 @ r2 to point at the end of r1
|
| + ldmia r0,{r3,r4,r5,r6,r7}
|
| +.Lloop:
|
| + ldr r8,.LK_00_19
|
| + mov r14,sp
|
| + sub sp,sp,#15*4
|
| + mov r5,r5,ror#30
|
| + mov r6,r6,ror#30
|
| + mov r7,r7,ror#30 @ [6]
|
| +.L_00_15:
|
| +#if __ARM_ARCH__<7
|
| + ldrb r10,[r1,#2]
|
| + ldrb r9,[r1,#3]
|
| + ldrb r11,[r1,#1]
|
| + add r7,r8,r7,ror#2 @ E+=K_00_19
|
| + ldrb r12,[r1],#4
|
| + orr r9,r9,r10,lsl#8
|
| + eor r10,r5,r6 @ F_xx_xx
|
| + orr r9,r9,r11,lsl#16
|
| + add r7,r7,r3,ror#27 @ E+=ROR(A,27)
|
| + orr r9,r9,r12,lsl#24
|
| +#else
|
| + ldr r9,[r1],#4 @ handles unaligned
|
| + add r7,r8,r7,ror#2 @ E+=K_00_19
|
| + eor r10,r5,r6 @ F_xx_xx
|
| + add r7,r7,r3,ror#27 @ E+=ROR(A,27)
|
| +#ifdef __ARMEL__
|
| + rev r9,r9 @ byte swap
|
| +#endif
|
| +#endif
|
| + and r10,r4,r10,ror#2
|
| + add r7,r7,r9 @ E+=X[i]
|
| + eor r10,r10,r6,ror#2 @ F_00_19(B,C,D)
|
| + str r9,[r14,#-4]!
|
| + add r7,r7,r10 @ E+=F_00_19(B,C,D)
|
| +#if __ARM_ARCH__<7
|
| + ldrb r10,[r1,#2]
|
| + ldrb r9,[r1,#3]
|
| + ldrb r11,[r1,#1]
|
| + add r6,r8,r6,ror#2 @ E+=K_00_19
|
| + ldrb r12,[r1],#4
|
| + orr r9,r9,r10,lsl#8
|
| + eor r10,r4,r5 @ F_xx_xx
|
| + orr r9,r9,r11,lsl#16
|
| + add r6,r6,r7,ror#27 @ E+=ROR(A,27)
|
| + orr r9,r9,r12,lsl#24
|
| +#else
|
| + ldr r9,[r1],#4 @ handles unaligned
|
| + add r6,r8,r6,ror#2 @ E+=K_00_19
|
| + eor r10,r4,r5 @ F_xx_xx
|
| + add r6,r6,r7,ror#27 @ E+=ROR(A,27)
|
| +#ifdef __ARMEL__
|
| + rev r9,r9 @ byte swap
|
| +#endif
|
| +#endif
|
| + and r10,r3,r10,ror#2
|
| + add r6,r6,r9 @ E+=X[i]
|
| + eor r10,r10,r5,ror#2 @ F_00_19(B,C,D)
|
| + str r9,[r14,#-4]!
|
| + add r6,r6,r10 @ E+=F_00_19(B,C,D)
|
| +#if __ARM_ARCH__<7
|
| + ldrb r10,[r1,#2]
|
| + ldrb r9,[r1,#3]
|
| + ldrb r11,[r1,#1]
|
| + add r5,r8,r5,ror#2 @ E+=K_00_19
|
| + ldrb r12,[r1],#4
|
| + orr r9,r9,r10,lsl#8
|
| + eor r10,r3,r4 @ F_xx_xx
|
| + orr r9,r9,r11,lsl#16
|
| + add r5,r5,r6,ror#27 @ E+=ROR(A,27)
|
| + orr r9,r9,r12,lsl#24
|
| +#else
|
| + ldr r9,[r1],#4 @ handles unaligned
|
| + add r5,r8,r5,ror#2 @ E+=K_00_19
|
| + eor r10,r3,r4 @ F_xx_xx
|
| + add r5,r5,r6,ror#27 @ E+=ROR(A,27)
|
| +#ifdef __ARMEL__
|
| + rev r9,r9 @ byte swap
|
| +#endif
|
| +#endif
|
| + and r10,r7,r10,ror#2
|
| + add r5,r5,r9 @ E+=X[i]
|
| + eor r10,r10,r4,ror#2 @ F_00_19(B,C,D)
|
| + str r9,[r14,#-4]!
|
| + add r5,r5,r10 @ E+=F_00_19(B,C,D)
|
| +#if __ARM_ARCH__<7
|
| + ldrb r10,[r1,#2]
|
| + ldrb r9,[r1,#3]
|
| + ldrb r11,[r1,#1]
|
| + add r4,r8,r4,ror#2 @ E+=K_00_19
|
| + ldrb r12,[r1],#4
|
| + orr r9,r9,r10,lsl#8
|
| + eor r10,r7,r3 @ F_xx_xx
|
| + orr r9,r9,r11,lsl#16
|
| + add r4,r4,r5,ror#27 @ E+=ROR(A,27)
|
| + orr r9,r9,r12,lsl#24
|
| +#else
|
| + ldr r9,[r1],#4 @ handles unaligned
|
| + add r4,r8,r4,ror#2 @ E+=K_00_19
|
| + eor r10,r7,r3 @ F_xx_xx
|
| + add r4,r4,r5,ror#27 @ E+=ROR(A,27)
|
| +#ifdef __ARMEL__
|
| + rev r9,r9 @ byte swap
|
| +#endif
|
| +#endif
|
| + and r10,r6,r10,ror#2
|
| + add r4,r4,r9 @ E+=X[i]
|
| + eor r10,r10,r3,ror#2 @ F_00_19(B,C,D)
|
| + str r9,[r14,#-4]!
|
| + add r4,r4,r10 @ E+=F_00_19(B,C,D)
|
| +#if __ARM_ARCH__<7
|
| + ldrb r10,[r1,#2]
|
| + ldrb r9,[r1,#3]
|
| + ldrb r11,[r1,#1]
|
| + add r3,r8,r3,ror#2 @ E+=K_00_19
|
| + ldrb r12,[r1],#4
|
| + orr r9,r9,r10,lsl#8
|
| + eor r10,r6,r7 @ F_xx_xx
|
| + orr r9,r9,r11,lsl#16
|
| + add r3,r3,r4,ror#27 @ E+=ROR(A,27)
|
| + orr r9,r9,r12,lsl#24
|
| +#else
|
| + ldr r9,[r1],#4 @ handles unaligned
|
| + add r3,r8,r3,ror#2 @ E+=K_00_19
|
| + eor r10,r6,r7 @ F_xx_xx
|
| + add r3,r3,r4,ror#27 @ E+=ROR(A,27)
|
| +#ifdef __ARMEL__
|
| + rev r9,r9 @ byte swap
|
| +#endif
|
| +#endif
|
| + and r10,r5,r10,ror#2
|
| + add r3,r3,r9 @ E+=X[i]
|
| + eor r10,r10,r7,ror#2 @ F_00_19(B,C,D)
|
| + str r9,[r14,#-4]!
|
| + add r3,r3,r10 @ E+=F_00_19(B,C,D)
|
| + teq r14,sp
|
| + bne .L_00_15 @ [((11+4)*5+2)*3]
|
| + sub sp,sp,#25*4
|
| +#if __ARM_ARCH__<7
|
| + ldrb r10,[r1,#2]
|
| + ldrb r9,[r1,#3]
|
| + ldrb r11,[r1,#1]
|
| + add r7,r8,r7,ror#2 @ E+=K_00_19
|
| + ldrb r12,[r1],#4
|
| + orr r9,r9,r10,lsl#8
|
| + eor r10,r5,r6 @ F_xx_xx
|
| + orr r9,r9,r11,lsl#16
|
| + add r7,r7,r3,ror#27 @ E+=ROR(A,27)
|
| + orr r9,r9,r12,lsl#24
|
| +#else
|
| + ldr r9,[r1],#4 @ handles unaligned
|
| + add r7,r8,r7,ror#2 @ E+=K_00_19
|
| + eor r10,r5,r6 @ F_xx_xx
|
| + add r7,r7,r3,ror#27 @ E+=ROR(A,27)
|
| +#ifdef __ARMEL__
|
| + rev r9,r9 @ byte swap
|
| +#endif
|
| +#endif
|
| + and r10,r4,r10,ror#2
|
| + add r7,r7,r9 @ E+=X[i]
|
| + eor r10,r10,r6,ror#2 @ F_00_19(B,C,D)
|
| + str r9,[r14,#-4]!
|
| + add r7,r7,r10 @ E+=F_00_19(B,C,D)
|
| + ldr r9,[r14,#15*4]
|
| + ldr r10,[r14,#13*4]
|
| + ldr r11,[r14,#7*4]
|
| + add r6,r8,r6,ror#2 @ E+=K_xx_xx
|
| + ldr r12,[r14,#2*4]
|
| + eor r9,r9,r10
|
| + eor r11,r11,r12 @ 1 cycle stall
|
| + eor r10,r4,r5 @ F_xx_xx
|
| + mov r9,r9,ror#31
|
| + add r6,r6,r7,ror#27 @ E+=ROR(A,27)
|
| + eor r9,r9,r11,ror#31
|
| + str r9,[r14,#-4]!
|
| + and r10,r3,r10,ror#2 @ F_xx_xx
|
| + @ F_xx_xx
|
| + add r6,r6,r9 @ E+=X[i]
|
| + eor r10,r10,r5,ror#2 @ F_00_19(B,C,D)
|
| + add r6,r6,r10 @ E+=F_00_19(B,C,D)
|
| + ldr r9,[r14,#15*4]
|
| + ldr r10,[r14,#13*4]
|
| + ldr r11,[r14,#7*4]
|
| + add r5,r8,r5,ror#2 @ E+=K_xx_xx
|
| + ldr r12,[r14,#2*4]
|
| + eor r9,r9,r10
|
| + eor r11,r11,r12 @ 1 cycle stall
|
| + eor r10,r3,r4 @ F_xx_xx
|
| + mov r9,r9,ror#31
|
| + add r5,r5,r6,ror#27 @ E+=ROR(A,27)
|
| + eor r9,r9,r11,ror#31
|
| + str r9,[r14,#-4]!
|
| + and r10,r7,r10,ror#2 @ F_xx_xx
|
| + @ F_xx_xx
|
| + add r5,r5,r9 @ E+=X[i]
|
| + eor r10,r10,r4,ror#2 @ F_00_19(B,C,D)
|
| + add r5,r5,r10 @ E+=F_00_19(B,C,D)
|
| + ldr r9,[r14,#15*4]
|
| + ldr r10,[r14,#13*4]
|
| + ldr r11,[r14,#7*4]
|
| + add r4,r8,r4,ror#2 @ E+=K_xx_xx
|
| + ldr r12,[r14,#2*4]
|
| + eor r9,r9,r10
|
| + eor r11,r11,r12 @ 1 cycle stall
|
| + eor r10,r7,r3 @ F_xx_xx
|
| + mov r9,r9,ror#31
|
| + add r4,r4,r5,ror#27 @ E+=ROR(A,27)
|
| + eor r9,r9,r11,ror#31
|
| + str r9,[r14,#-4]!
|
| + and r10,r6,r10,ror#2 @ F_xx_xx
|
| + @ F_xx_xx
|
| + add r4,r4,r9 @ E+=X[i]
|
| + eor r10,r10,r3,ror#2 @ F_00_19(B,C,D)
|
| + add r4,r4,r10 @ E+=F_00_19(B,C,D)
|
| + ldr r9,[r14,#15*4]
|
| + ldr r10,[r14,#13*4]
|
| + ldr r11,[r14,#7*4]
|
| + add r3,r8,r3,ror#2 @ E+=K_xx_xx
|
| + ldr r12,[r14,#2*4]
|
| + eor r9,r9,r10
|
| + eor r11,r11,r12 @ 1 cycle stall
|
| + eor r10,r6,r7 @ F_xx_xx
|
| + mov r9,r9,ror#31
|
| + add r3,r3,r4,ror#27 @ E+=ROR(A,27)
|
| + eor r9,r9,r11,ror#31
|
| + str r9,[r14,#-4]!
|
| + and r10,r5,r10,ror#2 @ F_xx_xx
|
| + @ F_xx_xx
|
| + add r3,r3,r9 @ E+=X[i]
|
| + eor r10,r10,r7,ror#2 @ F_00_19(B,C,D)
|
| + add r3,r3,r10 @ E+=F_00_19(B,C,D)
|
| +
|
| + ldr r8,.LK_20_39 @ [+15+16*4]
|
| + cmn sp,#0 @ [+3], clear carry to denote 20_39
|
| +.L_20_39_or_60_79:
|
| + ldr r9,[r14,#15*4]
|
| + ldr r10,[r14,#13*4]
|
| + ldr r11,[r14,#7*4]
|
| + add r7,r8,r7,ror#2 @ E+=K_xx_xx
|
| + ldr r12,[r14,#2*4]
|
| + eor r9,r9,r10
|
| + eor r11,r11,r12 @ 1 cycle stall
|
| + eor r10,r5,r6 @ F_xx_xx
|
| + mov r9,r9,ror#31
|
| + add r7,r7,r3,ror#27 @ E+=ROR(A,27)
|
| + eor r9,r9,r11,ror#31
|
| + str r9,[r14,#-4]!
|
| + eor r10,r4,r10,ror#2 @ F_xx_xx
|
| + @ F_xx_xx
|
| + add r7,r7,r9 @ E+=X[i]
|
| + add r7,r7,r10 @ E+=F_20_39(B,C,D)
|
| + ldr r9,[r14,#15*4]
|
| + ldr r10,[r14,#13*4]
|
| + ldr r11,[r14,#7*4]
|
| + add r6,r8,r6,ror#2 @ E+=K_xx_xx
|
| + ldr r12,[r14,#2*4]
|
| + eor r9,r9,r10
|
| + eor r11,r11,r12 @ 1 cycle stall
|
| + eor r10,r4,r5 @ F_xx_xx
|
| + mov r9,r9,ror#31
|
| + add r6,r6,r7,ror#27 @ E+=ROR(A,27)
|
| + eor r9,r9,r11,ror#31
|
| + str r9,[r14,#-4]!
|
| + eor r10,r3,r10,ror#2 @ F_xx_xx
|
| + @ F_xx_xx
|
| + add r6,r6,r9 @ E+=X[i]
|
| + add r6,r6,r10 @ E+=F_20_39(B,C,D)
|
| + ldr r9,[r14,#15*4]
|
| + ldr r10,[r14,#13*4]
|
| + ldr r11,[r14,#7*4]
|
| + add r5,r8,r5,ror#2 @ E+=K_xx_xx
|
| + ldr r12,[r14,#2*4]
|
| + eor r9,r9,r10
|
| + eor r11,r11,r12 @ 1 cycle stall
|
| + eor r10,r3,r4 @ F_xx_xx
|
| + mov r9,r9,ror#31
|
| + add r5,r5,r6,ror#27 @ E+=ROR(A,27)
|
| + eor r9,r9,r11,ror#31
|
| + str r9,[r14,#-4]!
|
| + eor r10,r7,r10,ror#2 @ F_xx_xx
|
| + @ F_xx_xx
|
| + add r5,r5,r9 @ E+=X[i]
|
| + add r5,r5,r10 @ E+=F_20_39(B,C,D)
|
| + ldr r9,[r14,#15*4]
|
| + ldr r10,[r14,#13*4]
|
| + ldr r11,[r14,#7*4]
|
| + add r4,r8,r4,ror#2 @ E+=K_xx_xx
|
| + ldr r12,[r14,#2*4]
|
| + eor r9,r9,r10
|
| + eor r11,r11,r12 @ 1 cycle stall
|
| + eor r10,r7,r3 @ F_xx_xx
|
| + mov r9,r9,ror#31
|
| + add r4,r4,r5,ror#27 @ E+=ROR(A,27)
|
| + eor r9,r9,r11,ror#31
|
| + str r9,[r14,#-4]!
|
| + eor r10,r6,r10,ror#2 @ F_xx_xx
|
| + @ F_xx_xx
|
| + add r4,r4,r9 @ E+=X[i]
|
| + add r4,r4,r10 @ E+=F_20_39(B,C,D)
|
| + ldr r9,[r14,#15*4]
|
| + ldr r10,[r14,#13*4]
|
| + ldr r11,[r14,#7*4]
|
| + add r3,r8,r3,ror#2 @ E+=K_xx_xx
|
| + ldr r12,[r14,#2*4]
|
| + eor r9,r9,r10
|
| + eor r11,r11,r12 @ 1 cycle stall
|
| + eor r10,r6,r7 @ F_xx_xx
|
| + mov r9,r9,ror#31
|
| + add r3,r3,r4,ror#27 @ E+=ROR(A,27)
|
| + eor r9,r9,r11,ror#31
|
| + str r9,[r14,#-4]!
|
| + eor r10,r5,r10,ror#2 @ F_xx_xx
|
| + @ F_xx_xx
|
| + add r3,r3,r9 @ E+=X[i]
|
| + add r3,r3,r10 @ E+=F_20_39(B,C,D)
|
| + teq r14,sp @ preserve carry
|
| + bne .L_20_39_or_60_79 @ [+((12+3)*5+2)*4]
|
| + bcs .L_done @ [+((12+3)*5+2)*4], spare 300 bytes
|
| +
|
| + ldr r8,.LK_40_59
|
| + sub sp,sp,#20*4 @ [+2]
|
| +.L_40_59:
|
| + ldr r9,[r14,#15*4]
|
| + ldr r10,[r14,#13*4]
|
| + ldr r11,[r14,#7*4]
|
| + add r7,r8,r7,ror#2 @ E+=K_xx_xx
|
| + ldr r12,[r14,#2*4]
|
| + eor r9,r9,r10
|
| + eor r11,r11,r12 @ 1 cycle stall
|
| + eor r10,r5,r6 @ F_xx_xx
|
| + mov r9,r9,ror#31
|
| + add r7,r7,r3,ror#27 @ E+=ROR(A,27)
|
| + eor r9,r9,r11,ror#31
|
| + str r9,[r14,#-4]!
|
| + and r10,r4,r10,ror#2 @ F_xx_xx
|
| + and r11,r5,r6 @ F_xx_xx
|
| + add r7,r7,r9 @ E+=X[i]
|
| + add r7,r7,r10 @ E+=F_40_59(B,C,D)
|
| + add r7,r7,r11,ror#2
|
| + ldr r9,[r14,#15*4]
|
| + ldr r10,[r14,#13*4]
|
| + ldr r11,[r14,#7*4]
|
| + add r6,r8,r6,ror#2 @ E+=K_xx_xx
|
| + ldr r12,[r14,#2*4]
|
| + eor r9,r9,r10
|
| + eor r11,r11,r12 @ 1 cycle stall
|
| + eor r10,r4,r5 @ F_xx_xx
|
| + mov r9,r9,ror#31
|
| + add r6,r6,r7,ror#27 @ E+=ROR(A,27)
|
| + eor r9,r9,r11,ror#31
|
| + str r9,[r14,#-4]!
|
| + and r10,r3,r10,ror#2 @ F_xx_xx
|
| + and r11,r4,r5 @ F_xx_xx
|
| + add r6,r6,r9 @ E+=X[i]
|
| + add r6,r6,r10 @ E+=F_40_59(B,C,D)
|
| + add r6,r6,r11,ror#2
|
| + ldr r9,[r14,#15*4]
|
| + ldr r10,[r14,#13*4]
|
| + ldr r11,[r14,#7*4]
|
| + add r5,r8,r5,ror#2 @ E+=K_xx_xx
|
| + ldr r12,[r14,#2*4]
|
| + eor r9,r9,r10
|
| + eor r11,r11,r12 @ 1 cycle stall
|
| + eor r10,r3,r4 @ F_xx_xx
|
| + mov r9,r9,ror#31
|
| + add r5,r5,r6,ror#27 @ E+=ROR(A,27)
|
| + eor r9,r9,r11,ror#31
|
| + str r9,[r14,#-4]!
|
| + and r10,r7,r10,ror#2 @ F_xx_xx
|
| + and r11,r3,r4 @ F_xx_xx
|
| + add r5,r5,r9 @ E+=X[i]
|
| + add r5,r5,r10 @ E+=F_40_59(B,C,D)
|
| + add r5,r5,r11,ror#2
|
| + ldr r9,[r14,#15*4]
|
| + ldr r10,[r14,#13*4]
|
| + ldr r11,[r14,#7*4]
|
| + add r4,r8,r4,ror#2 @ E+=K_xx_xx
|
| + ldr r12,[r14,#2*4]
|
| + eor r9,r9,r10
|
| + eor r11,r11,r12 @ 1 cycle stall
|
| + eor r10,r7,r3 @ F_xx_xx
|
| + mov r9,r9,ror#31
|
| + add r4,r4,r5,ror#27 @ E+=ROR(A,27)
|
| + eor r9,r9,r11,ror#31
|
| + str r9,[r14,#-4]!
|
| + and r10,r6,r10,ror#2 @ F_xx_xx
|
| + and r11,r7,r3 @ F_xx_xx
|
| + add r4,r4,r9 @ E+=X[i]
|
| + add r4,r4,r10 @ E+=F_40_59(B,C,D)
|
| + add r4,r4,r11,ror#2
|
| + ldr r9,[r14,#15*4]
|
| + ldr r10,[r14,#13*4]
|
| + ldr r11,[r14,#7*4]
|
| + add r3,r8,r3,ror#2 @ E+=K_xx_xx
|
| + ldr r12,[r14,#2*4]
|
| + eor r9,r9,r10
|
| + eor r11,r11,r12 @ 1 cycle stall
|
| + eor r10,r6,r7 @ F_xx_xx
|
| + mov r9,r9,ror#31
|
| + add r3,r3,r4,ror#27 @ E+=ROR(A,27)
|
| + eor r9,r9,r11,ror#31
|
| + str r9,[r14,#-4]!
|
| + and r10,r5,r10,ror#2 @ F_xx_xx
|
| + and r11,r6,r7 @ F_xx_xx
|
| + add r3,r3,r9 @ E+=X[i]
|
| + add r3,r3,r10 @ E+=F_40_59(B,C,D)
|
| + add r3,r3,r11,ror#2
|
| + teq r14,sp
|
| + bne .L_40_59 @ [+((12+5)*5+2)*4]
|
| +
|
| + ldr r8,.LK_60_79
|
| + sub sp,sp,#20*4
|
| + cmp sp,#0 @ set carry to denote 60_79
|
| + b .L_20_39_or_60_79 @ [+4], spare 300 bytes
|
| +.L_done:
|
| + add sp,sp,#80*4 @ "deallocate" stack frame
|
| + ldmia r0,{r8,r9,r10,r11,r12}
|
| + add r3,r8,r3
|
| + add r4,r9,r4
|
| + add r5,r10,r5,ror#2
|
| + add r6,r11,r6,ror#2
|
| + add r7,r12,r7,ror#2
|
| + stmia r0,{r3,r4,r5,r6,r7}
|
| + teq r1,r2
|
| + bne .Lloop @ [+18], total 1307
|
| +
|
| +#if __ARM_ARCH__>=5
|
| + ldmia sp!,{r4-r12,pc}
|
| +#else
|
| + ldmia sp!,{r4-r12,lr}
|
| + tst lr,#1
|
| + moveq pc,lr @ be binary compatible with V4, yet
|
| + .word 0xe12fff1e @ interoperable with Thumb ISA:-)
|
| +#endif
|
| +.align 2
|
| +.LK_00_19: .word 0x5a827999
|
| +.LK_20_39: .word 0x6ed9eba1
|
| +.LK_40_59: .word 0x8f1bbcdc
|
| +.LK_60_79: .word 0xca62c1d6
|
| +.size sha1_block_data_order,.-sha1_block_data_order
|
| +.asciz "SHA1 block transform for ARMv4, CRYPTOGAMS by <appro@openssl.org>"
|
| +.align 2
|
| +
|
| +#endif
|
|
|