| Index: openssl/crypto/sha/asm/sha256-armv4.pl
|
| diff --git a/openssl/crypto/sha/asm/sha256-armv4.pl b/openssl/crypto/sha/asm/sha256-armv4.pl
|
| deleted file mode 100644
|
| index 9c84e8d93c301a35b544b112cb87a90b7ddbd573..0000000000000000000000000000000000000000
|
| --- a/openssl/crypto/sha/asm/sha256-armv4.pl
|
| +++ /dev/null
|
| @@ -1,211 +0,0 @@
|
| -#!/usr/bin/env perl
|
| -
|
| -# ====================================================================
|
| -# Written by Andy Polyakov <appro@fy.chalmers.se> for the OpenSSL
|
| -# project. The module is, however, dual licensed under OpenSSL and
|
| -# CRYPTOGAMS licenses depending on where you obtain it. For further
|
| -# details see http://www.openssl.org/~appro/cryptogams/.
|
| -# ====================================================================
|
| -
|
| -# SHA256 block procedure for ARMv4. May 2007.
|
| -
|
| -# Performance is ~2x better than gcc 3.4 generated code and in "abso-
|
| -# lute" terms is ~2250 cycles per 64-byte block or ~35 cycles per
|
| -# byte [on single-issue Xscale PXA250 core].
|
| -
|
| -# July 2010.
|
| -#
|
| -# Rescheduling for dual-issue pipeline resulted in 22% improvement on
|
| -# Cortex A8 core and ~20 cycles per processed byte.
|
| -
|
| -# February 2011.
|
| -#
|
| -# Profiler-assisted and platform-specific optimization resulted in 16%
|
| -# improvement on Cortex A8 core and ~17 cycles per processed byte.
|
| -
|
| -while (($output=shift) && ($output!~/^\w[\w\-]*\.\w+$/)) {}
|
| -open STDOUT,">$output";
|
| -
|
| -$ctx="r0"; $t0="r0";
|
| -$inp="r1"; $t3="r1";
|
| -$len="r2"; $t1="r2";
|
| -$T1="r3";
|
| -$A="r4";
|
| -$B="r5";
|
| -$C="r6";
|
| -$D="r7";
|
| -$E="r8";
|
| -$F="r9";
|
| -$G="r10";
|
| -$H="r11";
|
| -@V=($A,$B,$C,$D,$E,$F,$G,$H);
|
| -$t2="r12";
|
| -$Ktbl="r14";
|
| -
|
| -@Sigma0=( 2,13,22);
|
| -@Sigma1=( 6,11,25);
|
| -@sigma0=( 7,18, 3);
|
| -@sigma1=(17,19,10);
|
| -
|
| -sub BODY_00_15 {
|
| -my ($i,$a,$b,$c,$d,$e,$f,$g,$h) = @_;
|
| -
|
| -$code.=<<___ if ($i<16);
|
| -#if __ARM_ARCH__>=7
|
| - ldr $T1,[$inp],#4
|
| -#else
|
| - ldrb $T1,[$inp,#3] @ $i
|
| - ldrb $t2,[$inp,#2]
|
| - ldrb $t1,[$inp,#1]
|
| - ldrb $t0,[$inp],#4
|
| - orr $T1,$T1,$t2,lsl#8
|
| - orr $T1,$T1,$t1,lsl#16
|
| - orr $T1,$T1,$t0,lsl#24
|
| -#endif
|
| -___
|
| -$code.=<<___;
|
| - mov $t0,$e,ror#$Sigma1[0]
|
| - ldr $t2,[$Ktbl],#4 @ *K256++
|
| - eor $t0,$t0,$e,ror#$Sigma1[1]
|
| - eor $t1,$f,$g
|
| -#if $i>=16
|
| - add $T1,$T1,$t3 @ from BODY_16_xx
|
| -#elif __ARM_ARCH__>=7 && defined(__ARMEL__)
|
| - rev $T1,$T1
|
| -#endif
|
| -#if $i==15
|
| - str $inp,[sp,#17*4] @ leave room for $t3
|
| -#endif
|
| - eor $t0,$t0,$e,ror#$Sigma1[2] @ Sigma1(e)
|
| - and $t1,$t1,$e
|
| - str $T1,[sp,#`$i%16`*4]
|
| - add $T1,$T1,$t0
|
| - eor $t1,$t1,$g @ Ch(e,f,g)
|
| - add $T1,$T1,$h
|
| - mov $h,$a,ror#$Sigma0[0]
|
| - add $T1,$T1,$t1
|
| - eor $h,$h,$a,ror#$Sigma0[1]
|
| - add $T1,$T1,$t2
|
| - eor $h,$h,$a,ror#$Sigma0[2] @ Sigma0(a)
|
| -#if $i>=15
|
| - ldr $t3,[sp,#`($i+2)%16`*4] @ from BODY_16_xx
|
| -#endif
|
| - orr $t0,$a,$b
|
| - and $t1,$a,$b
|
| - and $t0,$t0,$c
|
| - add $h,$h,$T1
|
| - orr $t0,$t0,$t1 @ Maj(a,b,c)
|
| - add $d,$d,$T1
|
| - add $h,$h,$t0
|
| -___
|
| -}
|
| -
|
| -sub BODY_16_XX {
|
| -my ($i,$a,$b,$c,$d,$e,$f,$g,$h) = @_;
|
| -
|
| -$code.=<<___;
|
| - @ ldr $t3,[sp,#`($i+1)%16`*4] @ $i
|
| - ldr $t2,[sp,#`($i+14)%16`*4]
|
| - mov $t0,$t3,ror#$sigma0[0]
|
| - ldr $T1,[sp,#`($i+0)%16`*4]
|
| - eor $t0,$t0,$t3,ror#$sigma0[1]
|
| - ldr $t1,[sp,#`($i+9)%16`*4]
|
| - eor $t0,$t0,$t3,lsr#$sigma0[2] @ sigma0(X[i+1])
|
| - mov $t3,$t2,ror#$sigma1[0]
|
| - add $T1,$T1,$t0
|
| - eor $t3,$t3,$t2,ror#$sigma1[1]
|
| - add $T1,$T1,$t1
|
| - eor $t3,$t3,$t2,lsr#$sigma1[2] @ sigma1(X[i+14])
|
| - @ add $T1,$T1,$t3
|
| -___
|
| - &BODY_00_15(@_);
|
| -}
|
| -
|
| -$code=<<___;
|
| -#include "arm_arch.h"
|
| -
|
| -.text
|
| -.code 32
|
| -
|
| -.type K256,%object
|
| -.align 5
|
| -K256:
|
| -.word 0x428a2f98,0x71374491,0xb5c0fbcf,0xe9b5dba5
|
| -.word 0x3956c25b,0x59f111f1,0x923f82a4,0xab1c5ed5
|
| -.word 0xd807aa98,0x12835b01,0x243185be,0x550c7dc3
|
| -.word 0x72be5d74,0x80deb1fe,0x9bdc06a7,0xc19bf174
|
| -.word 0xe49b69c1,0xefbe4786,0x0fc19dc6,0x240ca1cc
|
| -.word 0x2de92c6f,0x4a7484aa,0x5cb0a9dc,0x76f988da
|
| -.word 0x983e5152,0xa831c66d,0xb00327c8,0xbf597fc7
|
| -.word 0xc6e00bf3,0xd5a79147,0x06ca6351,0x14292967
|
| -.word 0x27b70a85,0x2e1b2138,0x4d2c6dfc,0x53380d13
|
| -.word 0x650a7354,0x766a0abb,0x81c2c92e,0x92722c85
|
| -.word 0xa2bfe8a1,0xa81a664b,0xc24b8b70,0xc76c51a3
|
| -.word 0xd192e819,0xd6990624,0xf40e3585,0x106aa070
|
| -.word 0x19a4c116,0x1e376c08,0x2748774c,0x34b0bcb5
|
| -.word 0x391c0cb3,0x4ed8aa4a,0x5b9cca4f,0x682e6ff3
|
| -.word 0x748f82ee,0x78a5636f,0x84c87814,0x8cc70208
|
| -.word 0x90befffa,0xa4506ceb,0xbef9a3f7,0xc67178f2
|
| -.size K256,.-K256
|
| -
|
| -.global sha256_block_data_order
|
| -.type sha256_block_data_order,%function
|
| -sha256_block_data_order:
|
| - sub r3,pc,#8 @ sha256_block_data_order
|
| - add $len,$inp,$len,lsl#6 @ len to point at the end of inp
|
| - stmdb sp!,{$ctx,$inp,$len,r4-r11,lr}
|
| - ldmia $ctx,{$A,$B,$C,$D,$E,$F,$G,$H}
|
| - sub $Ktbl,r3,#256 @ K256
|
| - sub sp,sp,#16*4 @ alloca(X[16])
|
| -.Loop:
|
| -___
|
| -for($i=0;$i<16;$i++) { &BODY_00_15($i,@V); unshift(@V,pop(@V)); }
|
| -$code.=".Lrounds_16_xx:\n";
|
| -for (;$i<32;$i++) { &BODY_16_XX($i,@V); unshift(@V,pop(@V)); }
|
| -$code.=<<___;
|
| - and $t2,$t2,#0xff
|
| - cmp $t2,#0xf2
|
| - bne .Lrounds_16_xx
|
| -
|
| - ldr $T1,[sp,#16*4] @ pull ctx
|
| - ldr $t0,[$T1,#0]
|
| - ldr $t1,[$T1,#4]
|
| - ldr $t2,[$T1,#8]
|
| - add $A,$A,$t0
|
| - ldr $t0,[$T1,#12]
|
| - add $B,$B,$t1
|
| - ldr $t1,[$T1,#16]
|
| - add $C,$C,$t2
|
| - ldr $t2,[$T1,#20]
|
| - add $D,$D,$t0
|
| - ldr $t0,[$T1,#24]
|
| - add $E,$E,$t1
|
| - ldr $t1,[$T1,#28]
|
| - add $F,$F,$t2
|
| - ldr $inp,[sp,#17*4] @ pull inp
|
| - ldr $t2,[sp,#18*4] @ pull inp+len
|
| - add $G,$G,$t0
|
| - add $H,$H,$t1
|
| - stmia $T1,{$A,$B,$C,$D,$E,$F,$G,$H}
|
| - cmp $inp,$t2
|
| - sub $Ktbl,$Ktbl,#256 @ rewind Ktbl
|
| - bne .Loop
|
| -
|
| - add sp,sp,#`16+3`*4 @ destroy frame
|
| -#if __ARM_ARCH__>=5
|
| - ldmia sp!,{r4-r11,pc}
|
| -#else
|
| - ldmia sp!,{r4-r11,lr}
|
| - tst lr,#1
|
| - moveq pc,lr @ be binary compatible with V4, yet
|
| - bx lr @ interoperable with Thumb ISA:-)
|
| -#endif
|
| -.size sha256_block_data_order,.-sha256_block_data_order
|
| -.asciz "SHA256 block transform for ARMv4, CRYPTOGAMS by <appro\@openssl.org>"
|
| -.align 2
|
| -___
|
| -
|
| -$code =~ s/\`([^\`]*)\`/eval $1/gem;
|
| -$code =~ s/\bbx\s+lr\b/.word\t0xe12fff1e/gm; # make it possible to compile with -march=armv4
|
| -print $code;
|
| -close STDOUT; # enforce flush
|
|
|