Index: third_party/boringssl/linux-arm/crypto/sha/sha1-armv4-large.S |
diff --git a/third_party/boringssl/linux-arm/crypto/sha/sha1-armv4-large.S b/third_party/boringssl/linux-arm/crypto/sha/sha1-armv4-large.S |
new file mode 100644 |
index 0000000000000000000000000000000000000000..cc436446582a0c96b3c1fdb30a929bcc660c0d0c |
--- /dev/null |
+++ b/third_party/boringssl/linux-arm/crypto/sha/sha1-armv4-large.S |
@@ -0,0 +1,455 @@ |
+#if defined(__arm__) |
+#include "arm_arch.h" |
+ |
+.text |
+ |
+.global sha1_block_data_order |
+.type sha1_block_data_order,%function |
+ |
+.align 2 |
+sha1_block_data_order: |
+ stmdb sp!,{r4-r12,lr} |
+ add r2,r1,r2,lsl#6 @ r2 to point at the end of r1 |
+ ldmia r0,{r3,r4,r5,r6,r7} |
+.Lloop: |
+ ldr r8,.LK_00_19 |
+ mov r14,sp |
+ sub sp,sp,#15*4 |
+ mov r5,r5,ror#30 |
+ mov r6,r6,ror#30 |
+ mov r7,r7,ror#30 @ [6] |
+.L_00_15: |
+#if __ARM_ARCH__<7 |
+ ldrb r10,[r1,#2] |
+ ldrb r9,[r1,#3] |
+ ldrb r11,[r1,#1] |
+ add r7,r8,r7,ror#2 @ E+=K_00_19 |
+ ldrb r12,[r1],#4 |
+ orr r9,r9,r10,lsl#8 |
+ eor r10,r5,r6 @ F_xx_xx |
+ orr r9,r9,r11,lsl#16 |
+ add r7,r7,r3,ror#27 @ E+=ROR(A,27) |
+ orr r9,r9,r12,lsl#24 |
+#else |
+ ldr r9,[r1],#4 @ handles unaligned |
+ add r7,r8,r7,ror#2 @ E+=K_00_19 |
+ eor r10,r5,r6 @ F_xx_xx |
+ add r7,r7,r3,ror#27 @ E+=ROR(A,27) |
+#ifdef __ARMEL__ |
+ rev r9,r9 @ byte swap |
+#endif |
+#endif |
+ and r10,r4,r10,ror#2 |
+ add r7,r7,r9 @ E+=X[i] |
+ eor r10,r10,r6,ror#2 @ F_00_19(B,C,D) |
+ str r9,[r14,#-4]! |
+ add r7,r7,r10 @ E+=F_00_19(B,C,D) |
+#if __ARM_ARCH__<7 |
+ ldrb r10,[r1,#2] |
+ ldrb r9,[r1,#3] |
+ ldrb r11,[r1,#1] |
+ add r6,r8,r6,ror#2 @ E+=K_00_19 |
+ ldrb r12,[r1],#4 |
+ orr r9,r9,r10,lsl#8 |
+ eor r10,r4,r5 @ F_xx_xx |
+ orr r9,r9,r11,lsl#16 |
+ add r6,r6,r7,ror#27 @ E+=ROR(A,27) |
+ orr r9,r9,r12,lsl#24 |
+#else |
+ ldr r9,[r1],#4 @ handles unaligned |
+ add r6,r8,r6,ror#2 @ E+=K_00_19 |
+ eor r10,r4,r5 @ F_xx_xx |
+ add r6,r6,r7,ror#27 @ E+=ROR(A,27) |
+#ifdef __ARMEL__ |
+ rev r9,r9 @ byte swap |
+#endif |
+#endif |
+ and r10,r3,r10,ror#2 |
+ add r6,r6,r9 @ E+=X[i] |
+ eor r10,r10,r5,ror#2 @ F_00_19(B,C,D) |
+ str r9,[r14,#-4]! |
+ add r6,r6,r10 @ E+=F_00_19(B,C,D) |
+#if __ARM_ARCH__<7 |
+ ldrb r10,[r1,#2] |
+ ldrb r9,[r1,#3] |
+ ldrb r11,[r1,#1] |
+ add r5,r8,r5,ror#2 @ E+=K_00_19 |
+ ldrb r12,[r1],#4 |
+ orr r9,r9,r10,lsl#8 |
+ eor r10,r3,r4 @ F_xx_xx |
+ orr r9,r9,r11,lsl#16 |
+ add r5,r5,r6,ror#27 @ E+=ROR(A,27) |
+ orr r9,r9,r12,lsl#24 |
+#else |
+ ldr r9,[r1],#4 @ handles unaligned |
+ add r5,r8,r5,ror#2 @ E+=K_00_19 |
+ eor r10,r3,r4 @ F_xx_xx |
+ add r5,r5,r6,ror#27 @ E+=ROR(A,27) |
+#ifdef __ARMEL__ |
+ rev r9,r9 @ byte swap |
+#endif |
+#endif |
+ and r10,r7,r10,ror#2 |
+ add r5,r5,r9 @ E+=X[i] |
+ eor r10,r10,r4,ror#2 @ F_00_19(B,C,D) |
+ str r9,[r14,#-4]! |
+ add r5,r5,r10 @ E+=F_00_19(B,C,D) |
+#if __ARM_ARCH__<7 |
+ ldrb r10,[r1,#2] |
+ ldrb r9,[r1,#3] |
+ ldrb r11,[r1,#1] |
+ add r4,r8,r4,ror#2 @ E+=K_00_19 |
+ ldrb r12,[r1],#4 |
+ orr r9,r9,r10,lsl#8 |
+ eor r10,r7,r3 @ F_xx_xx |
+ orr r9,r9,r11,lsl#16 |
+ add r4,r4,r5,ror#27 @ E+=ROR(A,27) |
+ orr r9,r9,r12,lsl#24 |
+#else |
+ ldr r9,[r1],#4 @ handles unaligned |
+ add r4,r8,r4,ror#2 @ E+=K_00_19 |
+ eor r10,r7,r3 @ F_xx_xx |
+ add r4,r4,r5,ror#27 @ E+=ROR(A,27) |
+#ifdef __ARMEL__ |
+ rev r9,r9 @ byte swap |
+#endif |
+#endif |
+ and r10,r6,r10,ror#2 |
+ add r4,r4,r9 @ E+=X[i] |
+ eor r10,r10,r3,ror#2 @ F_00_19(B,C,D) |
+ str r9,[r14,#-4]! |
+ add r4,r4,r10 @ E+=F_00_19(B,C,D) |
+#if __ARM_ARCH__<7 |
+ ldrb r10,[r1,#2] |
+ ldrb r9,[r1,#3] |
+ ldrb r11,[r1,#1] |
+ add r3,r8,r3,ror#2 @ E+=K_00_19 |
+ ldrb r12,[r1],#4 |
+ orr r9,r9,r10,lsl#8 |
+ eor r10,r6,r7 @ F_xx_xx |
+ orr r9,r9,r11,lsl#16 |
+ add r3,r3,r4,ror#27 @ E+=ROR(A,27) |
+ orr r9,r9,r12,lsl#24 |
+#else |
+ ldr r9,[r1],#4 @ handles unaligned |
+ add r3,r8,r3,ror#2 @ E+=K_00_19 |
+ eor r10,r6,r7 @ F_xx_xx |
+ add r3,r3,r4,ror#27 @ E+=ROR(A,27) |
+#ifdef __ARMEL__ |
+ rev r9,r9 @ byte swap |
+#endif |
+#endif |
+ and r10,r5,r10,ror#2 |
+ add r3,r3,r9 @ E+=X[i] |
+ eor r10,r10,r7,ror#2 @ F_00_19(B,C,D) |
+ str r9,[r14,#-4]! |
+ add r3,r3,r10 @ E+=F_00_19(B,C,D) |
+ teq r14,sp |
+ bne .L_00_15 @ [((11+4)*5+2)*3] |
+ sub sp,sp,#25*4 |
+#if __ARM_ARCH__<7 |
+ ldrb r10,[r1,#2] |
+ ldrb r9,[r1,#3] |
+ ldrb r11,[r1,#1] |
+ add r7,r8,r7,ror#2 @ E+=K_00_19 |
+ ldrb r12,[r1],#4 |
+ orr r9,r9,r10,lsl#8 |
+ eor r10,r5,r6 @ F_xx_xx |
+ orr r9,r9,r11,lsl#16 |
+ add r7,r7,r3,ror#27 @ E+=ROR(A,27) |
+ orr r9,r9,r12,lsl#24 |
+#else |
+ ldr r9,[r1],#4 @ handles unaligned |
+ add r7,r8,r7,ror#2 @ E+=K_00_19 |
+ eor r10,r5,r6 @ F_xx_xx |
+ add r7,r7,r3,ror#27 @ E+=ROR(A,27) |
+#ifdef __ARMEL__ |
+ rev r9,r9 @ byte swap |
+#endif |
+#endif |
+ and r10,r4,r10,ror#2 |
+ add r7,r7,r9 @ E+=X[i] |
+ eor r10,r10,r6,ror#2 @ F_00_19(B,C,D) |
+ str r9,[r14,#-4]! |
+ add r7,r7,r10 @ E+=F_00_19(B,C,D) |
+ ldr r9,[r14,#15*4] |
+ ldr r10,[r14,#13*4] |
+ ldr r11,[r14,#7*4] |
+ add r6,r8,r6,ror#2 @ E+=K_xx_xx |
+ ldr r12,[r14,#2*4] |
+ eor r9,r9,r10 |
+ eor r11,r11,r12 @ 1 cycle stall |
+ eor r10,r4,r5 @ F_xx_xx |
+ mov r9,r9,ror#31 |
+ add r6,r6,r7,ror#27 @ E+=ROR(A,27) |
+ eor r9,r9,r11,ror#31 |
+ str r9,[r14,#-4]! |
+ and r10,r3,r10,ror#2 @ F_xx_xx |
+ @ F_xx_xx |
+ add r6,r6,r9 @ E+=X[i] |
+ eor r10,r10,r5,ror#2 @ F_00_19(B,C,D) |
+ add r6,r6,r10 @ E+=F_00_19(B,C,D) |
+ ldr r9,[r14,#15*4] |
+ ldr r10,[r14,#13*4] |
+ ldr r11,[r14,#7*4] |
+ add r5,r8,r5,ror#2 @ E+=K_xx_xx |
+ ldr r12,[r14,#2*4] |
+ eor r9,r9,r10 |
+ eor r11,r11,r12 @ 1 cycle stall |
+ eor r10,r3,r4 @ F_xx_xx |
+ mov r9,r9,ror#31 |
+ add r5,r5,r6,ror#27 @ E+=ROR(A,27) |
+ eor r9,r9,r11,ror#31 |
+ str r9,[r14,#-4]! |
+ and r10,r7,r10,ror#2 @ F_xx_xx |
+ @ F_xx_xx |
+ add r5,r5,r9 @ E+=X[i] |
+ eor r10,r10,r4,ror#2 @ F_00_19(B,C,D) |
+ add r5,r5,r10 @ E+=F_00_19(B,C,D) |
+ ldr r9,[r14,#15*4] |
+ ldr r10,[r14,#13*4] |
+ ldr r11,[r14,#7*4] |
+ add r4,r8,r4,ror#2 @ E+=K_xx_xx |
+ ldr r12,[r14,#2*4] |
+ eor r9,r9,r10 |
+ eor r11,r11,r12 @ 1 cycle stall |
+ eor r10,r7,r3 @ F_xx_xx |
+ mov r9,r9,ror#31 |
+ add r4,r4,r5,ror#27 @ E+=ROR(A,27) |
+ eor r9,r9,r11,ror#31 |
+ str r9,[r14,#-4]! |
+ and r10,r6,r10,ror#2 @ F_xx_xx |
+ @ F_xx_xx |
+ add r4,r4,r9 @ E+=X[i] |
+ eor r10,r10,r3,ror#2 @ F_00_19(B,C,D) |
+ add r4,r4,r10 @ E+=F_00_19(B,C,D) |
+ ldr r9,[r14,#15*4] |
+ ldr r10,[r14,#13*4] |
+ ldr r11,[r14,#7*4] |
+ add r3,r8,r3,ror#2 @ E+=K_xx_xx |
+ ldr r12,[r14,#2*4] |
+ eor r9,r9,r10 |
+ eor r11,r11,r12 @ 1 cycle stall |
+ eor r10,r6,r7 @ F_xx_xx |
+ mov r9,r9,ror#31 |
+ add r3,r3,r4,ror#27 @ E+=ROR(A,27) |
+ eor r9,r9,r11,ror#31 |
+ str r9,[r14,#-4]! |
+ and r10,r5,r10,ror#2 @ F_xx_xx |
+ @ F_xx_xx |
+ add r3,r3,r9 @ E+=X[i] |
+ eor r10,r10,r7,ror#2 @ F_00_19(B,C,D) |
+ add r3,r3,r10 @ E+=F_00_19(B,C,D) |
+ |
+ ldr r8,.LK_20_39 @ [+15+16*4] |
+ cmn sp,#0 @ [+3], clear carry to denote 20_39 |
+.L_20_39_or_60_79: |
+ ldr r9,[r14,#15*4] |
+ ldr r10,[r14,#13*4] |
+ ldr r11,[r14,#7*4] |
+ add r7,r8,r7,ror#2 @ E+=K_xx_xx |
+ ldr r12,[r14,#2*4] |
+ eor r9,r9,r10 |
+ eor r11,r11,r12 @ 1 cycle stall |
+ eor r10,r5,r6 @ F_xx_xx |
+ mov r9,r9,ror#31 |
+ add r7,r7,r3,ror#27 @ E+=ROR(A,27) |
+ eor r9,r9,r11,ror#31 |
+ str r9,[r14,#-4]! |
+ eor r10,r4,r10,ror#2 @ F_xx_xx |
+ @ F_xx_xx |
+ add r7,r7,r9 @ E+=X[i] |
+ add r7,r7,r10 @ E+=F_20_39(B,C,D) |
+ ldr r9,[r14,#15*4] |
+ ldr r10,[r14,#13*4] |
+ ldr r11,[r14,#7*4] |
+ add r6,r8,r6,ror#2 @ E+=K_xx_xx |
+ ldr r12,[r14,#2*4] |
+ eor r9,r9,r10 |
+ eor r11,r11,r12 @ 1 cycle stall |
+ eor r10,r4,r5 @ F_xx_xx |
+ mov r9,r9,ror#31 |
+ add r6,r6,r7,ror#27 @ E+=ROR(A,27) |
+ eor r9,r9,r11,ror#31 |
+ str r9,[r14,#-4]! |
+ eor r10,r3,r10,ror#2 @ F_xx_xx |
+ @ F_xx_xx |
+ add r6,r6,r9 @ E+=X[i] |
+ add r6,r6,r10 @ E+=F_20_39(B,C,D) |
+ ldr r9,[r14,#15*4] |
+ ldr r10,[r14,#13*4] |
+ ldr r11,[r14,#7*4] |
+ add r5,r8,r5,ror#2 @ E+=K_xx_xx |
+ ldr r12,[r14,#2*4] |
+ eor r9,r9,r10 |
+ eor r11,r11,r12 @ 1 cycle stall |
+ eor r10,r3,r4 @ F_xx_xx |
+ mov r9,r9,ror#31 |
+ add r5,r5,r6,ror#27 @ E+=ROR(A,27) |
+ eor r9,r9,r11,ror#31 |
+ str r9,[r14,#-4]! |
+ eor r10,r7,r10,ror#2 @ F_xx_xx |
+ @ F_xx_xx |
+ add r5,r5,r9 @ E+=X[i] |
+ add r5,r5,r10 @ E+=F_20_39(B,C,D) |
+ ldr r9,[r14,#15*4] |
+ ldr r10,[r14,#13*4] |
+ ldr r11,[r14,#7*4] |
+ add r4,r8,r4,ror#2 @ E+=K_xx_xx |
+ ldr r12,[r14,#2*4] |
+ eor r9,r9,r10 |
+ eor r11,r11,r12 @ 1 cycle stall |
+ eor r10,r7,r3 @ F_xx_xx |
+ mov r9,r9,ror#31 |
+ add r4,r4,r5,ror#27 @ E+=ROR(A,27) |
+ eor r9,r9,r11,ror#31 |
+ str r9,[r14,#-4]! |
+ eor r10,r6,r10,ror#2 @ F_xx_xx |
+ @ F_xx_xx |
+ add r4,r4,r9 @ E+=X[i] |
+ add r4,r4,r10 @ E+=F_20_39(B,C,D) |
+ ldr r9,[r14,#15*4] |
+ ldr r10,[r14,#13*4] |
+ ldr r11,[r14,#7*4] |
+ add r3,r8,r3,ror#2 @ E+=K_xx_xx |
+ ldr r12,[r14,#2*4] |
+ eor r9,r9,r10 |
+ eor r11,r11,r12 @ 1 cycle stall |
+ eor r10,r6,r7 @ F_xx_xx |
+ mov r9,r9,ror#31 |
+ add r3,r3,r4,ror#27 @ E+=ROR(A,27) |
+ eor r9,r9,r11,ror#31 |
+ str r9,[r14,#-4]! |
+ eor r10,r5,r10,ror#2 @ F_xx_xx |
+ @ F_xx_xx |
+ add r3,r3,r9 @ E+=X[i] |
+ add r3,r3,r10 @ E+=F_20_39(B,C,D) |
+ teq r14,sp @ preserve carry |
+ bne .L_20_39_or_60_79 @ [+((12+3)*5+2)*4] |
+ bcs .L_done @ [+((12+3)*5+2)*4], spare 300 bytes |
+ |
+ ldr r8,.LK_40_59 |
+ sub sp,sp,#20*4 @ [+2] |
+.L_40_59: |
+ ldr r9,[r14,#15*4] |
+ ldr r10,[r14,#13*4] |
+ ldr r11,[r14,#7*4] |
+ add r7,r8,r7,ror#2 @ E+=K_xx_xx |
+ ldr r12,[r14,#2*4] |
+ eor r9,r9,r10 |
+ eor r11,r11,r12 @ 1 cycle stall |
+ eor r10,r5,r6 @ F_xx_xx |
+ mov r9,r9,ror#31 |
+ add r7,r7,r3,ror#27 @ E+=ROR(A,27) |
+ eor r9,r9,r11,ror#31 |
+ str r9,[r14,#-4]! |
+ and r10,r4,r10,ror#2 @ F_xx_xx |
+ and r11,r5,r6 @ F_xx_xx |
+ add r7,r7,r9 @ E+=X[i] |
+ add r7,r7,r10 @ E+=F_40_59(B,C,D) |
+ add r7,r7,r11,ror#2 |
+ ldr r9,[r14,#15*4] |
+ ldr r10,[r14,#13*4] |
+ ldr r11,[r14,#7*4] |
+ add r6,r8,r6,ror#2 @ E+=K_xx_xx |
+ ldr r12,[r14,#2*4] |
+ eor r9,r9,r10 |
+ eor r11,r11,r12 @ 1 cycle stall |
+ eor r10,r4,r5 @ F_xx_xx |
+ mov r9,r9,ror#31 |
+ add r6,r6,r7,ror#27 @ E+=ROR(A,27) |
+ eor r9,r9,r11,ror#31 |
+ str r9,[r14,#-4]! |
+ and r10,r3,r10,ror#2 @ F_xx_xx |
+ and r11,r4,r5 @ F_xx_xx |
+ add r6,r6,r9 @ E+=X[i] |
+ add r6,r6,r10 @ E+=F_40_59(B,C,D) |
+ add r6,r6,r11,ror#2 |
+ ldr r9,[r14,#15*4] |
+ ldr r10,[r14,#13*4] |
+ ldr r11,[r14,#7*4] |
+ add r5,r8,r5,ror#2 @ E+=K_xx_xx |
+ ldr r12,[r14,#2*4] |
+ eor r9,r9,r10 |
+ eor r11,r11,r12 @ 1 cycle stall |
+ eor r10,r3,r4 @ F_xx_xx |
+ mov r9,r9,ror#31 |
+ add r5,r5,r6,ror#27 @ E+=ROR(A,27) |
+ eor r9,r9,r11,ror#31 |
+ str r9,[r14,#-4]! |
+ and r10,r7,r10,ror#2 @ F_xx_xx |
+ and r11,r3,r4 @ F_xx_xx |
+ add r5,r5,r9 @ E+=X[i] |
+ add r5,r5,r10 @ E+=F_40_59(B,C,D) |
+ add r5,r5,r11,ror#2 |
+ ldr r9,[r14,#15*4] |
+ ldr r10,[r14,#13*4] |
+ ldr r11,[r14,#7*4] |
+ add r4,r8,r4,ror#2 @ E+=K_xx_xx |
+ ldr r12,[r14,#2*4] |
+ eor r9,r9,r10 |
+ eor r11,r11,r12 @ 1 cycle stall |
+ eor r10,r7,r3 @ F_xx_xx |
+ mov r9,r9,ror#31 |
+ add r4,r4,r5,ror#27 @ E+=ROR(A,27) |
+ eor r9,r9,r11,ror#31 |
+ str r9,[r14,#-4]! |
+ and r10,r6,r10,ror#2 @ F_xx_xx |
+ and r11,r7,r3 @ F_xx_xx |
+ add r4,r4,r9 @ E+=X[i] |
+ add r4,r4,r10 @ E+=F_40_59(B,C,D) |
+ add r4,r4,r11,ror#2 |
+ ldr r9,[r14,#15*4] |
+ ldr r10,[r14,#13*4] |
+ ldr r11,[r14,#7*4] |
+ add r3,r8,r3,ror#2 @ E+=K_xx_xx |
+ ldr r12,[r14,#2*4] |
+ eor r9,r9,r10 |
+ eor r11,r11,r12 @ 1 cycle stall |
+ eor r10,r6,r7 @ F_xx_xx |
+ mov r9,r9,ror#31 |
+ add r3,r3,r4,ror#27 @ E+=ROR(A,27) |
+ eor r9,r9,r11,ror#31 |
+ str r9,[r14,#-4]! |
+ and r10,r5,r10,ror#2 @ F_xx_xx |
+ and r11,r6,r7 @ F_xx_xx |
+ add r3,r3,r9 @ E+=X[i] |
+ add r3,r3,r10 @ E+=F_40_59(B,C,D) |
+ add r3,r3,r11,ror#2 |
+ teq r14,sp |
+ bne .L_40_59 @ [+((12+5)*5+2)*4] |
+ |
+ ldr r8,.LK_60_79 |
+ sub sp,sp,#20*4 |
+ cmp sp,#0 @ set carry to denote 60_79 |
+ b .L_20_39_or_60_79 @ [+4], spare 300 bytes |
+.L_done: |
+ add sp,sp,#80*4 @ "deallocate" stack frame |
+ ldmia r0,{r8,r9,r10,r11,r12} |
+ add r3,r8,r3 |
+ add r4,r9,r4 |
+ add r5,r10,r5,ror#2 |
+ add r6,r11,r6,ror#2 |
+ add r7,r12,r7,ror#2 |
+ stmia r0,{r3,r4,r5,r6,r7} |
+ teq r1,r2 |
+ bne .Lloop @ [+18], total 1307 |
+ |
+#if __ARM_ARCH__>=5 |
+ ldmia sp!,{r4-r12,pc} |
+#else |
+ ldmia sp!,{r4-r12,lr} |
+ tst lr,#1 |
+ moveq pc,lr @ be binary compatible with V4, yet |
+ .word 0xe12fff1e @ interoperable with Thumb ISA:-) |
+#endif |
+.align 2 |
+.LK_00_19: .word 0x5a827999 |
+.LK_20_39: .word 0x6ed9eba1 |
+.LK_40_59: .word 0x8f1bbcdc |
+.LK_60_79: .word 0xca62c1d6 |
+.size sha1_block_data_order,.-sha1_block_data_order |
+.asciz "SHA1 block transform for ARMv4, CRYPTOGAMS by <appro@openssl.org>" |
+.align 2 |
+ |
+#endif |