| Index: openssl/crypto/aes/asm/aes-x86_64.pl
|
| ===================================================================
|
| --- openssl/crypto/aes/asm/aes-x86_64.pl (revision 105093)
|
| +++ openssl/crypto/aes/asm/aes-x86_64.pl (working copy)
|
| @@ -2,11 +2,12 @@
|
| #
|
| # ====================================================================
|
| # Written by Andy Polyakov <appro@fy.chalmers.se> for the OpenSSL
|
| -# project. Rights for redistribution and usage in source and binary
|
| -# forms are granted according to the OpenSSL license.
|
| +# project. The module is, however, dual licensed under OpenSSL and
|
| +# CRYPTOGAMS licenses depending on where you obtain it. For further
|
| +# details see http://www.openssl.org/~appro/cryptogams/.
|
| # ====================================================================
|
| #
|
| -# Version 1.2.
|
| +# Version 2.1.
|
| #
|
| # aes-*-cbc benchmarks are improved by >70% [compared to gcc 3.3.2 on
|
| # Opteron 240 CPU] plus all the bells-n-whistles from 32-bit version
|
| @@ -17,17 +18,29 @@
|
| #
|
| # Performance in number of cycles per processed byte for 128-bit key:
|
| #
|
| -# ECB CBC encrypt
|
| -# AMD64 13.7 13.0(*)
|
| -# EM64T 20.2 18.6(*)
|
| +# ECB encrypt ECB decrypt CBC large chunk
|
| +# AMD64 33 41 13.0
|
| +# EM64T 38 59 18.6(*)
|
| +# Core 2 30 43 14.5(*)
|
| #
|
| -# (*) CBC benchmarks are better than ECB thanks to custom ABI used
|
| -# by the private block encryption function.
|
| +# (*) with hyper-threading off
|
|
|
| +$flavour = shift;
|
| +$output = shift;
|
| +if ($flavour =~ /\./) { $output = $flavour; undef $flavour; }
|
| +
|
| +$win64=0; $win64=1 if ($flavour =~ /[nm]asm|mingw64/ || $output =~ /\.asm$/);
|
| +
|
| +$0 =~ m/(.*[\/\\])[^\/\\]+$/; $dir=$1;
|
| +( $xlate="${dir}x86_64-xlate.pl" and -f $xlate ) or
|
| +( $xlate="${dir}../../perlasm/x86_64-xlate.pl" and -f $xlate) or
|
| +die "can't locate x86_64-xlate.pl";
|
| +
|
| +open STDOUT,"| $^X $xlate $flavour $output";
|
| +
|
| $verticalspin=1; # unlike 32-bit version $verticalspin performs
|
| # ~15% better on both AMD and Intel cores
|
| -$output=shift;
|
| -open STDOUT,"| $^X ../perlasm/x86_64-xlate.pl $output";
|
| +$speed_limit=512; # see aes-586.pl for details
|
|
|
| $code=".text\n";
|
|
|
| @@ -35,9 +48,9 @@
|
| $s1="%ebx";
|
| $s2="%ecx";
|
| $s3="%edx";
|
| -$acc0="%esi";
|
| -$acc1="%edi";
|
| -$acc2="%ebp";
|
| +$acc0="%esi"; $mask80="%rsi";
|
| +$acc1="%edi"; $maskfe="%rdi";
|
| +$acc2="%ebp"; $mask1b="%rbp";
|
| $inp="%r8";
|
| $out="%r9";
|
| $t0="%r10d";
|
| @@ -51,6 +64,8 @@
|
| sub lo() { my $r=shift; $r =~ s/%[er]([a-d])x/%\1l/;
|
| $r =~ s/%[er]([sd]i)/%\1l/;
|
| $r =~ s/%(r[0-9]+)[d]?/%\1b/; $r; }
|
| +sub LO() { my $r=shift; $r =~ s/%r([a-z]+)/%e\1/;
|
| + $r =~ s/%r([0-9]+)/%r\1d/; $r; }
|
| sub _data_word()
|
| { my $i;
|
| while(defined($i=shift)) { $code.=sprintf".long\t0x%08x,0x%08x\n",$i,$i; }
|
| @@ -138,22 +153,17 @@
|
| movzb `&lo("$s0")`,$acc0
|
| movzb `&lo("$s1")`,$acc1
|
| movzb `&lo("$s2")`,$acc2
|
| - mov 2($sbox,$acc0,8),$t0
|
| - mov 2($sbox,$acc1,8),$t1
|
| - mov 2($sbox,$acc2,8),$t2
|
| + movzb 2($sbox,$acc0,8),$t0
|
| + movzb 2($sbox,$acc1,8),$t1
|
| + movzb 2($sbox,$acc2,8),$t2
|
|
|
| - and \$0x000000ff,$t0
|
| - and \$0x000000ff,$t1
|
| - and \$0x000000ff,$t2
|
| -
|
| movzb `&lo("$s3")`,$acc0
|
| movzb `&hi("$s1")`,$acc1
|
| movzb `&hi("$s2")`,$acc2
|
| - mov 2($sbox,$acc0,8),$t3
|
| + movzb 2($sbox,$acc0,8),$t3
|
| mov 0($sbox,$acc1,8),$acc1 #$t0
|
| mov 0($sbox,$acc2,8),$acc2 #$t1
|
|
|
| - and \$0x000000ff,$t3
|
| and \$0x0000ff00,$acc1
|
| and \$0x0000ff00,$acc2
|
|
|
| @@ -345,6 +355,234 @@
|
| .size _x86_64_AES_encrypt,.-_x86_64_AES_encrypt
|
| ___
|
|
|
| +# it's possible to implement this by shifting tN by 8, filling least
|
| +# significant byte with byte load and finally bswap-ing at the end,
|
| +# but such partial register load kills Core 2...
|
| +sub enccompactvert()
|
| +{ my ($t3,$t4,$t5)=("%r8d","%r9d","%r13d");
|
| +
|
| +$code.=<<___;
|
| + movzb `&lo("$s0")`,$t0
|
| + movzb `&lo("$s1")`,$t1
|
| + movzb `&lo("$s2")`,$t2
|
| + movzb ($sbox,$t0,1),$t0
|
| + movzb ($sbox,$t1,1),$t1
|
| + movzb ($sbox,$t2,1),$t2
|
| +
|
| + movzb `&lo("$s3")`,$t3
|
| + movzb `&hi("$s1")`,$acc0
|
| + movzb `&hi("$s2")`,$acc1
|
| + movzb ($sbox,$t3,1),$t3
|
| + movzb ($sbox,$acc0,1),$t4 #$t0
|
| + movzb ($sbox,$acc1,1),$t5 #$t1
|
| +
|
| + movzb `&hi("$s3")`,$acc2
|
| + movzb `&hi("$s0")`,$acc0
|
| + shr \$16,$s2
|
| + movzb ($sbox,$acc2,1),$acc2 #$t2
|
| + movzb ($sbox,$acc0,1),$acc0 #$t3
|
| + shr \$16,$s3
|
| +
|
| + movzb `&lo("$s2")`,$acc1
|
| + shl \$8,$t4
|
| + shl \$8,$t5
|
| + movzb ($sbox,$acc1,1),$acc1 #$t0
|
| + xor $t4,$t0
|
| + xor $t5,$t1
|
| +
|
| + movzb `&lo("$s3")`,$t4
|
| + shr \$16,$s0
|
| + shr \$16,$s1
|
| + movzb `&lo("$s0")`,$t5
|
| + shl \$8,$acc2
|
| + shl \$8,$acc0
|
| + movzb ($sbox,$t4,1),$t4 #$t1
|
| + movzb ($sbox,$t5,1),$t5 #$t2
|
| + xor $acc2,$t2
|
| + xor $acc0,$t3
|
| +
|
| + movzb `&lo("$s1")`,$acc2
|
| + movzb `&hi("$s3")`,$acc0
|
| + shl \$16,$acc1
|
| + movzb ($sbox,$acc2,1),$acc2 #$t3
|
| + movzb ($sbox,$acc0,1),$acc0 #$t0
|
| + xor $acc1,$t0
|
| +
|
| + movzb `&hi("$s0")`,$acc1
|
| + shr \$8,$s2
|
| + shr \$8,$s1
|
| + movzb ($sbox,$acc1,1),$acc1 #$t1
|
| + movzb ($sbox,$s2,1),$s3 #$t3
|
| + movzb ($sbox,$s1,1),$s2 #$t2
|
| + shl \$16,$t4
|
| + shl \$16,$t5
|
| + shl \$16,$acc2
|
| + xor $t4,$t1
|
| + xor $t5,$t2
|
| + xor $acc2,$t3
|
| +
|
| + shl \$24,$acc0
|
| + shl \$24,$acc1
|
| + shl \$24,$s3
|
| + xor $acc0,$t0
|
| + shl \$24,$s2
|
| + xor $acc1,$t1
|
| + mov $t0,$s0
|
| + mov $t1,$s1
|
| + xor $t2,$s2
|
| + xor $t3,$s3
|
| +___
|
| +}
|
| +
|
| +sub enctransform_ref()
|
| +{ my $sn = shift;
|
| + my ($acc,$r2,$tmp)=("%r8d","%r9d","%r13d");
|
| +
|
| +$code.=<<___;
|
| + mov $sn,$acc
|
| + and \$0x80808080,$acc
|
| + mov $acc,$tmp
|
| + shr \$7,$tmp
|
| + lea ($sn,$sn),$r2
|
| + sub $tmp,$acc
|
| + and \$0xfefefefe,$r2
|
| + and \$0x1b1b1b1b,$acc
|
| + mov $sn,$tmp
|
| + xor $acc,$r2
|
| +
|
| + xor $r2,$sn
|
| + rol \$24,$sn
|
| + xor $r2,$sn
|
| + ror \$16,$tmp
|
| + xor $tmp,$sn
|
| + ror \$8,$tmp
|
| + xor $tmp,$sn
|
| +___
|
| +}
|
| +
|
| +# unlike decrypt case it does not pay off to parallelize enctransform
|
| +sub enctransform()
|
| +{ my ($t3,$r20,$r21)=($acc2,"%r8d","%r9d");
|
| +
|
| +$code.=<<___;
|
| + mov $s0,$acc0
|
| + mov $s1,$acc1
|
| + and \$0x80808080,$acc0
|
| + and \$0x80808080,$acc1
|
| + mov $acc0,$t0
|
| + mov $acc1,$t1
|
| + shr \$7,$t0
|
| + lea ($s0,$s0),$r20
|
| + shr \$7,$t1
|
| + lea ($s1,$s1),$r21
|
| + sub $t0,$acc0
|
| + sub $t1,$acc1
|
| + and \$0xfefefefe,$r20
|
| + and \$0xfefefefe,$r21
|
| + and \$0x1b1b1b1b,$acc0
|
| + and \$0x1b1b1b1b,$acc1
|
| + mov $s0,$t0
|
| + mov $s1,$t1
|
| + xor $acc0,$r20
|
| + xor $acc1,$r21
|
| +
|
| + xor $r20,$s0
|
| + xor $r21,$s1
|
| + mov $s2,$acc0
|
| + mov $s3,$acc1
|
| + rol \$24,$s0
|
| + rol \$24,$s1
|
| + and \$0x80808080,$acc0
|
| + and \$0x80808080,$acc1
|
| + xor $r20,$s0
|
| + xor $r21,$s1
|
| + mov $acc0,$t2
|
| + mov $acc1,$t3
|
| + ror \$16,$t0
|
| + ror \$16,$t1
|
| + shr \$7,$t2
|
| + lea ($s2,$s2),$r20
|
| + xor $t0,$s0
|
| + xor $t1,$s1
|
| + shr \$7,$t3
|
| + lea ($s3,$s3),$r21
|
| + ror \$8,$t0
|
| + ror \$8,$t1
|
| + sub $t2,$acc0
|
| + sub $t3,$acc1
|
| + xor $t0,$s0
|
| + xor $t1,$s1
|
| +
|
| + and \$0xfefefefe,$r20
|
| + and \$0xfefefefe,$r21
|
| + and \$0x1b1b1b1b,$acc0
|
| + and \$0x1b1b1b1b,$acc1
|
| + mov $s2,$t2
|
| + mov $s3,$t3
|
| + xor $acc0,$r20
|
| + xor $acc1,$r21
|
| +
|
| + xor $r20,$s2
|
| + xor $r21,$s3
|
| + rol \$24,$s2
|
| + rol \$24,$s3
|
| + xor $r20,$s2
|
| + xor $r21,$s3
|
| + mov 0($sbox),$acc0 # prefetch Te4
|
| + ror \$16,$t2
|
| + ror \$16,$t3
|
| + mov 64($sbox),$acc1
|
| + xor $t2,$s2
|
| + xor $t3,$s3
|
| + mov 128($sbox),$r20
|
| + ror \$8,$t2
|
| + ror \$8,$t3
|
| + mov 192($sbox),$r21
|
| + xor $t2,$s2
|
| + xor $t3,$s3
|
| +___
|
| +}
|
| +
|
| +$code.=<<___;
|
| +.type _x86_64_AES_encrypt_compact,\@abi-omnipotent
|
| +.align 16
|
| +_x86_64_AES_encrypt_compact:
|
| + lea 128($sbox),$inp # size optimization
|
| + mov 0-128($inp),$acc1 # prefetch Te4
|
| + mov 32-128($inp),$acc2
|
| + mov 64-128($inp),$t0
|
| + mov 96-128($inp),$t1
|
| + mov 128-128($inp),$acc1
|
| + mov 160-128($inp),$acc2
|
| + mov 192-128($inp),$t0
|
| + mov 224-128($inp),$t1
|
| + jmp .Lenc_loop_compact
|
| +.align 16
|
| +.Lenc_loop_compact:
|
| + xor 0($key),$s0 # xor with key
|
| + xor 4($key),$s1
|
| + xor 8($key),$s2
|
| + xor 12($key),$s3
|
| + lea 16($key),$key
|
| +___
|
| + &enccompactvert();
|
| +$code.=<<___;
|
| + cmp 16(%rsp),$key
|
| + je .Lenc_compact_done
|
| +___
|
| + &enctransform();
|
| +$code.=<<___;
|
| + jmp .Lenc_loop_compact
|
| +.align 16
|
| +.Lenc_compact_done:
|
| + xor 0($key),$s0
|
| + xor 4($key),$s1
|
| + xor 8($key),$s2
|
| + xor 12($key),$s3
|
| + .byte 0xf3,0xc3 # rep ret
|
| +.size _x86_64_AES_encrypt_compact,.-_x86_64_AES_encrypt_compact
|
| +___
|
| +
|
| # void AES_encrypt (const void *inp,void *out,const AES_KEY *key);
|
| $code.=<<___;
|
| .globl AES_encrypt
|
| @@ -358,31 +596,57 @@
|
| push %r14
|
| push %r15
|
|
|
| + # allocate frame "above" key schedule
|
| + mov %rsp,%r10
|
| + lea -63(%rdx),%rcx # %rdx is key argument
|
| + and \$-64,%rsp
|
| + sub %rsp,%rcx
|
| + neg %rcx
|
| + and \$0x3c0,%rcx
|
| + sub %rcx,%rsp
|
| + sub \$32,%rsp
|
| +
|
| + mov %rsi,16(%rsp) # save out
|
| + mov %r10,24(%rsp) # save real stack pointer
|
| +.Lenc_prologue:
|
| +
|
| mov %rdx,$key
|
| - mov %rdi,$inp
|
| - mov %rsi,$out
|
| + mov 240($key),$rnds # load rounds
|
|
|
| - .picmeup $sbox
|
| - lea AES_Te-.($sbox),$sbox
|
| + mov 0(%rdi),$s0 # load input vector
|
| + mov 4(%rdi),$s1
|
| + mov 8(%rdi),$s2
|
| + mov 12(%rdi),$s3
|
|
|
| - mov 0($inp),$s0
|
| - mov 4($inp),$s1
|
| - mov 8($inp),$s2
|
| - mov 12($inp),$s3
|
| + shl \$4,$rnds
|
| + lea ($key,$rnds),%rbp
|
| + mov $key,(%rsp) # key schedule
|
| + mov %rbp,8(%rsp) # end of key schedule
|
|
|
| - call _x86_64_AES_encrypt
|
| + # pick Te4 copy which can't "overlap" with stack frame or key schedule
|
| + lea .LAES_Te+2048(%rip),$sbox
|
| + lea 768(%rsp),%rbp
|
| + sub $sbox,%rbp
|
| + and \$0x300,%rbp
|
| + lea ($sbox,%rbp),$sbox
|
|
|
| - mov $s0,0($out)
|
| + call _x86_64_AES_encrypt_compact
|
| +
|
| + mov 16(%rsp),$out # restore out
|
| + mov 24(%rsp),%rsi # restore saved stack pointer
|
| + mov $s0,0($out) # write output vector
|
| mov $s1,4($out)
|
| mov $s2,8($out)
|
| mov $s3,12($out)
|
|
|
| - pop %r15
|
| - pop %r14
|
| - pop %r13
|
| - pop %r12
|
| - pop %rbp
|
| - pop %rbx
|
| + mov (%rsi),%r15
|
| + mov 8(%rsi),%r14
|
| + mov 16(%rsi),%r13
|
| + mov 24(%rsi),%r12
|
| + mov 32(%rsi),%rbp
|
| + mov 40(%rsi),%rbx
|
| + lea 48(%rsi),%rsp
|
| +.Lenc_epilogue:
|
| ret
|
| .size AES_encrypt,.-AES_encrypt
|
| ___
|
| @@ -453,19 +717,20 @@
|
| { my $t3="%r8d"; # zaps $inp!
|
|
|
| $code.=<<___;
|
| + lea 2048($sbox),$sbox # size optimization
|
| movzb `&lo("$s0")`,$acc0
|
| movzb `&lo("$s1")`,$acc1
|
| movzb `&lo("$s2")`,$acc2
|
| - movzb 2048($sbox,$acc0,1),$t0
|
| - movzb 2048($sbox,$acc1,1),$t1
|
| - movzb 2048($sbox,$acc2,1),$t2
|
| + movzb ($sbox,$acc0,1),$t0
|
| + movzb ($sbox,$acc1,1),$t1
|
| + movzb ($sbox,$acc2,1),$t2
|
|
|
| movzb `&lo("$s3")`,$acc0
|
| movzb `&hi("$s3")`,$acc1
|
| movzb `&hi("$s0")`,$acc2
|
| - movzb 2048($sbox,$acc0,1),$t3
|
| - movzb 2048($sbox,$acc1,1),$acc1 #$t0
|
| - movzb 2048($sbox,$acc2,1),$acc2 #$t1
|
| + movzb ($sbox,$acc0,1),$t3
|
| + movzb ($sbox,$acc1,1),$acc1 #$t0
|
| + movzb ($sbox,$acc2,1),$acc2 #$t1
|
|
|
| shl \$8,$acc1
|
| shl \$8,$acc2
|
| @@ -477,8 +742,8 @@
|
| movzb `&hi("$s1")`,$acc0
|
| movzb `&hi("$s2")`,$acc1
|
| shr \$16,$s0
|
| - movzb 2048($sbox,$acc0,1),$acc0 #$t2
|
| - movzb 2048($sbox,$acc1,1),$acc1 #$t3
|
| + movzb ($sbox,$acc0,1),$acc0 #$t2
|
| + movzb ($sbox,$acc1,1),$acc1 #$t3
|
|
|
| shl \$8,$acc0
|
| shl \$8,$acc1
|
| @@ -490,9 +755,9 @@
|
| movzb `&lo("$s2")`,$acc0
|
| movzb `&lo("$s3")`,$acc1
|
| movzb `&lo("$s0")`,$acc2
|
| - movzb 2048($sbox,$acc0,1),$acc0 #$t0
|
| - movzb 2048($sbox,$acc1,1),$acc1 #$t1
|
| - movzb 2048($sbox,$acc2,1),$acc2 #$t2
|
| + movzb ($sbox,$acc0,1),$acc0 #$t0
|
| + movzb ($sbox,$acc1,1),$acc1 #$t1
|
| + movzb ($sbox,$acc2,1),$acc2 #$t2
|
|
|
| shl \$16,$acc0
|
| shl \$16,$acc1
|
| @@ -505,9 +770,9 @@
|
| movzb `&lo("$s1")`,$acc0
|
| movzb `&hi("$s1")`,$acc1
|
| movzb `&hi("$s2")`,$acc2
|
| - movzb 2048($sbox,$acc0,1),$acc0 #$t3
|
| - movzb 2048($sbox,$acc1,1),$acc1 #$t0
|
| - movzb 2048($sbox,$acc2,1),$acc2 #$t1
|
| + movzb ($sbox,$acc0,1),$acc0 #$t3
|
| + movzb ($sbox,$acc1,1),$acc1 #$t0
|
| + movzb ($sbox,$acc2,1),$acc2 #$t1
|
|
|
| shl \$16,$acc0
|
| shl \$24,$acc1
|
| @@ -520,8 +785,8 @@
|
| movzb `&hi("$s3")`,$acc0
|
| movzb `&hi("$s0")`,$acc1
|
| mov 16+12($key),$s3
|
| - movzb 2048($sbox,$acc0,1),$acc0 #$t2
|
| - movzb 2048($sbox,$acc1,1),$acc1 #$t3
|
| + movzb ($sbox,$acc0,1),$acc0 #$t2
|
| + movzb ($sbox,$acc1,1),$acc1 #$t3
|
| mov 16+0($key),$s0
|
|
|
| shl \$24,$acc0
|
| @@ -532,6 +797,7 @@
|
|
|
| mov 16+4($key),$s1
|
| mov 16+8($key),$s2
|
| + lea -2048($sbox),$sbox
|
| xor $t0,$s0
|
| xor $t1,$s1
|
| xor $t2,$s2
|
| @@ -659,6 +925,260 @@
|
| .size _x86_64_AES_decrypt,.-_x86_64_AES_decrypt
|
| ___
|
|
|
| +sub deccompactvert()
|
| +{ my ($t3,$t4,$t5)=("%r8d","%r9d","%r13d");
|
| +
|
| +$code.=<<___;
|
| + movzb `&lo("$s0")`,$t0
|
| + movzb `&lo("$s1")`,$t1
|
| + movzb `&lo("$s2")`,$t2
|
| + movzb ($sbox,$t0,1),$t0
|
| + movzb ($sbox,$t1,1),$t1
|
| + movzb ($sbox,$t2,1),$t2
|
| +
|
| + movzb `&lo("$s3")`,$t3
|
| + movzb `&hi("$s3")`,$acc0
|
| + movzb `&hi("$s0")`,$acc1
|
| + movzb ($sbox,$t3,1),$t3
|
| + movzb ($sbox,$acc0,1),$t4 #$t0
|
| + movzb ($sbox,$acc1,1),$t5 #$t1
|
| +
|
| + movzb `&hi("$s1")`,$acc2
|
| + movzb `&hi("$s2")`,$acc0
|
| + shr \$16,$s2
|
| + movzb ($sbox,$acc2,1),$acc2 #$t2
|
| + movzb ($sbox,$acc0,1),$acc0 #$t3
|
| + shr \$16,$s3
|
| +
|
| + movzb `&lo("$s2")`,$acc1
|
| + shl \$8,$t4
|
| + shl \$8,$t5
|
| + movzb ($sbox,$acc1,1),$acc1 #$t0
|
| + xor $t4,$t0
|
| + xor $t5,$t1
|
| +
|
| + movzb `&lo("$s3")`,$t4
|
| + shr \$16,$s0
|
| + shr \$16,$s1
|
| + movzb `&lo("$s0")`,$t5
|
| + shl \$8,$acc2
|
| + shl \$8,$acc0
|
| + movzb ($sbox,$t4,1),$t4 #$t1
|
| + movzb ($sbox,$t5,1),$t5 #$t2
|
| + xor $acc2,$t2
|
| + xor $acc0,$t3
|
| +
|
| + movzb `&lo("$s1")`,$acc2
|
| + movzb `&hi("$s1")`,$acc0
|
| + shl \$16,$acc1
|
| + movzb ($sbox,$acc2,1),$acc2 #$t3
|
| + movzb ($sbox,$acc0,1),$acc0 #$t0
|
| + xor $acc1,$t0
|
| +
|
| + movzb `&hi("$s2")`,$acc1
|
| + shl \$16,$t4
|
| + shl \$16,$t5
|
| + movzb ($sbox,$acc1,1),$s1 #$t1
|
| + xor $t4,$t1
|
| + xor $t5,$t2
|
| +
|
| + movzb `&hi("$s3")`,$acc1
|
| + shr \$8,$s0
|
| + shl \$16,$acc2
|
| + movzb ($sbox,$acc1,1),$s2 #$t2
|
| + movzb ($sbox,$s0,1),$s3 #$t3
|
| + xor $acc2,$t3
|
| +
|
| + shl \$24,$acc0
|
| + shl \$24,$s1
|
| + shl \$24,$s2
|
| + xor $acc0,$t0
|
| + shl \$24,$s3
|
| + xor $t1,$s1
|
| + mov $t0,$s0
|
| + xor $t2,$s2
|
| + xor $t3,$s3
|
| +___
|
| +}
|
| +
|
| +# parallelized version! input is pair of 64-bit values: %rax=s1.s0
|
| +# and %rcx=s3.s2, output is four 32-bit values in %eax=s0, %ebx=s1,
|
| +# %ecx=s2 and %edx=s3.
|
| +sub dectransform()
|
| +{ my ($tp10,$tp20,$tp40,$tp80,$acc0)=("%rax","%r8", "%r9", "%r10","%rbx");
|
| + my ($tp18,$tp28,$tp48,$tp88,$acc8)=("%rcx","%r11","%r12","%r13","%rdx");
|
| + my $prefetch = shift;
|
| +
|
| +$code.=<<___;
|
| + mov $tp10,$acc0
|
| + mov $tp18,$acc8
|
| + and $mask80,$acc0
|
| + and $mask80,$acc8
|
| + mov $acc0,$tp40
|
| + mov $acc8,$tp48
|
| + shr \$7,$tp40
|
| + lea ($tp10,$tp10),$tp20
|
| + shr \$7,$tp48
|
| + lea ($tp18,$tp18),$tp28
|
| + sub $tp40,$acc0
|
| + sub $tp48,$acc8
|
| + and $maskfe,$tp20
|
| + and $maskfe,$tp28
|
| + and $mask1b,$acc0
|
| + and $mask1b,$acc8
|
| + xor $tp20,$acc0
|
| + xor $tp28,$acc8
|
| + mov $acc0,$tp20
|
| + mov $acc8,$tp28
|
| +
|
| + and $mask80,$acc0
|
| + and $mask80,$acc8
|
| + mov $acc0,$tp80
|
| + mov $acc8,$tp88
|
| + shr \$7,$tp80
|
| + lea ($tp20,$tp20),$tp40
|
| + shr \$7,$tp88
|
| + lea ($tp28,$tp28),$tp48
|
| + sub $tp80,$acc0
|
| + sub $tp88,$acc8
|
| + and $maskfe,$tp40
|
| + and $maskfe,$tp48
|
| + and $mask1b,$acc0
|
| + and $mask1b,$acc8
|
| + xor $tp40,$acc0
|
| + xor $tp48,$acc8
|
| + mov $acc0,$tp40
|
| + mov $acc8,$tp48
|
| +
|
| + and $mask80,$acc0
|
| + and $mask80,$acc8
|
| + mov $acc0,$tp80
|
| + mov $acc8,$tp88
|
| + shr \$7,$tp80
|
| + xor $tp10,$tp20 # tp2^=tp1
|
| + shr \$7,$tp88
|
| + xor $tp18,$tp28 # tp2^=tp1
|
| + sub $tp80,$acc0
|
| + sub $tp88,$acc8
|
| + lea ($tp40,$tp40),$tp80
|
| + lea ($tp48,$tp48),$tp88
|
| + xor $tp10,$tp40 # tp4^=tp1
|
| + xor $tp18,$tp48 # tp4^=tp1
|
| + and $maskfe,$tp80
|
| + and $maskfe,$tp88
|
| + and $mask1b,$acc0
|
| + and $mask1b,$acc8
|
| + xor $acc0,$tp80
|
| + xor $acc8,$tp88
|
| +
|
| + xor $tp80,$tp10 # tp1^=tp8
|
| + xor $tp88,$tp18 # tp1^=tp8
|
| + xor $tp80,$tp20 # tp2^tp1^=tp8
|
| + xor $tp88,$tp28 # tp2^tp1^=tp8
|
| + mov $tp10,$acc0
|
| + mov $tp18,$acc8
|
| + xor $tp80,$tp40 # tp4^tp1^=tp8
|
| + xor $tp88,$tp48 # tp4^tp1^=tp8
|
| + shr \$32,$acc0
|
| + shr \$32,$acc8
|
| + xor $tp20,$tp80 # tp8^=tp8^tp2^tp1=tp2^tp1
|
| + xor $tp28,$tp88 # tp8^=tp8^tp2^tp1=tp2^tp1
|
| + rol \$8,`&LO("$tp10")` # ROTATE(tp1^tp8,8)
|
| + rol \$8,`&LO("$tp18")` # ROTATE(tp1^tp8,8)
|
| + xor $tp40,$tp80 # tp2^tp1^=tp8^tp4^tp1=tp8^tp4^tp2
|
| + xor $tp48,$tp88 # tp2^tp1^=tp8^tp4^tp1=tp8^tp4^tp2
|
| +
|
| + rol \$8,`&LO("$acc0")` # ROTATE(tp1^tp8,8)
|
| + rol \$8,`&LO("$acc8")` # ROTATE(tp1^tp8,8)
|
| + xor `&LO("$tp80")`,`&LO("$tp10")`
|
| + xor `&LO("$tp88")`,`&LO("$tp18")`
|
| + shr \$32,$tp80
|
| + shr \$32,$tp88
|
| + xor `&LO("$tp80")`,`&LO("$acc0")`
|
| + xor `&LO("$tp88")`,`&LO("$acc8")`
|
| +
|
| + mov $tp20,$tp80
|
| + mov $tp28,$tp88
|
| + shr \$32,$tp80
|
| + shr \$32,$tp88
|
| + rol \$24,`&LO("$tp20")` # ROTATE(tp2^tp1^tp8,24)
|
| + rol \$24,`&LO("$tp28")` # ROTATE(tp2^tp1^tp8,24)
|
| + rol \$24,`&LO("$tp80")` # ROTATE(tp2^tp1^tp8,24)
|
| + rol \$24,`&LO("$tp88")` # ROTATE(tp2^tp1^tp8,24)
|
| + xor `&LO("$tp20")`,`&LO("$tp10")`
|
| + xor `&LO("$tp28")`,`&LO("$tp18")`
|
| + mov $tp40,$tp20
|
| + mov $tp48,$tp28
|
| + xor `&LO("$tp80")`,`&LO("$acc0")`
|
| + xor `&LO("$tp88")`,`&LO("$acc8")`
|
| +
|
| + `"mov 0($sbox),$mask80" if ($prefetch)`
|
| + shr \$32,$tp20
|
| + shr \$32,$tp28
|
| + `"mov 64($sbox),$maskfe" if ($prefetch)`
|
| + rol \$16,`&LO("$tp40")` # ROTATE(tp4^tp1^tp8,16)
|
| + rol \$16,`&LO("$tp48")` # ROTATE(tp4^tp1^tp8,16)
|
| + `"mov 128($sbox),$mask1b" if ($prefetch)`
|
| + rol \$16,`&LO("$tp20")` # ROTATE(tp4^tp1^tp8,16)
|
| + rol \$16,`&LO("$tp28")` # ROTATE(tp4^tp1^tp8,16)
|
| + `"mov 192($sbox),$tp80" if ($prefetch)`
|
| + xor `&LO("$tp40")`,`&LO("$tp10")`
|
| + xor `&LO("$tp48")`,`&LO("$tp18")`
|
| + `"mov 256($sbox),$tp88" if ($prefetch)`
|
| + xor `&LO("$tp20")`,`&LO("$acc0")`
|
| + xor `&LO("$tp28")`,`&LO("$acc8")`
|
| +___
|
| +}
|
| +
|
| +$code.=<<___;
|
| +.type _x86_64_AES_decrypt_compact,\@abi-omnipotent
|
| +.align 16
|
| +_x86_64_AES_decrypt_compact:
|
| + lea 128($sbox),$inp # size optimization
|
| + mov 0-128($inp),$acc1 # prefetch Td4
|
| + mov 32-128($inp),$acc2
|
| + mov 64-128($inp),$t0
|
| + mov 96-128($inp),$t1
|
| + mov 128-128($inp),$acc1
|
| + mov 160-128($inp),$acc2
|
| + mov 192-128($inp),$t0
|
| + mov 224-128($inp),$t1
|
| + jmp .Ldec_loop_compact
|
| +
|
| +.align 16
|
| +.Ldec_loop_compact:
|
| + xor 0($key),$s0 # xor with key
|
| + xor 4($key),$s1
|
| + xor 8($key),$s2
|
| + xor 12($key),$s3
|
| + lea 16($key),$key
|
| +___
|
| + &deccompactvert();
|
| +$code.=<<___;
|
| + cmp 16(%rsp),$key
|
| + je .Ldec_compact_done
|
| +
|
| + mov 256+0($sbox),$mask80
|
| + shl \$32,%rbx
|
| + shl \$32,%rdx
|
| + mov 256+8($sbox),$maskfe
|
| + or %rbx,%rax
|
| + or %rdx,%rcx
|
| + mov 256+16($sbox),$mask1b
|
| +___
|
| + &dectransform(1);
|
| +$code.=<<___;
|
| + jmp .Ldec_loop_compact
|
| +.align 16
|
| +.Ldec_compact_done:
|
| + xor 0($key),$s0
|
| + xor 4($key),$s1
|
| + xor 8($key),$s2
|
| + xor 12($key),$s3
|
| + .byte 0xf3,0xc3 # rep ret
|
| +.size _x86_64_AES_decrypt_compact,.-_x86_64_AES_decrypt_compact
|
| +___
|
| +
|
| # void AES_decrypt (const void *inp,void *out,const AES_KEY *key);
|
| $code.=<<___;
|
| .globl AES_decrypt
|
| @@ -672,43 +1192,59 @@
|
| push %r14
|
| push %r15
|
|
|
| + # allocate frame "above" key schedule
|
| + mov %rsp,%r10
|
| + lea -63(%rdx),%rcx # %rdx is key argument
|
| + and \$-64,%rsp
|
| + sub %rsp,%rcx
|
| + neg %rcx
|
| + and \$0x3c0,%rcx
|
| + sub %rcx,%rsp
|
| + sub \$32,%rsp
|
| +
|
| + mov %rsi,16(%rsp) # save out
|
| + mov %r10,24(%rsp) # save real stack pointer
|
| +.Ldec_prologue:
|
| +
|
| mov %rdx,$key
|
| - mov %rdi,$inp
|
| - mov %rsi,$out
|
| + mov 240($key),$rnds # load rounds
|
|
|
| - .picmeup $sbox
|
| - lea AES_Td-.($sbox),$sbox
|
| + mov 0(%rdi),$s0 # load input vector
|
| + mov 4(%rdi),$s1
|
| + mov 8(%rdi),$s2
|
| + mov 12(%rdi),$s3
|
|
|
| - # prefetch Td4
|
| - lea 2048+128($sbox),$sbox;
|
| - mov 0-128($sbox),$s0
|
| - mov 32-128($sbox),$s1
|
| - mov 64-128($sbox),$s2
|
| - mov 96-128($sbox),$s3
|
| - mov 128-128($sbox),$s0
|
| - mov 160-128($sbox),$s1
|
| - mov 192-128($sbox),$s2
|
| - mov 224-128($sbox),$s3
|
| - lea -2048-128($sbox),$sbox;
|
| + shl \$4,$rnds
|
| + lea ($key,$rnds),%rbp
|
| + mov $key,(%rsp) # key schedule
|
| + mov %rbp,8(%rsp) # end of key schedule
|
|
|
| - mov 0($inp),$s0
|
| - mov 4($inp),$s1
|
| - mov 8($inp),$s2
|
| - mov 12($inp),$s3
|
| + # pick Td4 copy which can't "overlap" with stack frame or key schedule
|
| + lea .LAES_Td+2048(%rip),$sbox
|
| + lea 768(%rsp),%rbp
|
| + sub $sbox,%rbp
|
| + and \$0x300,%rbp
|
| + lea ($sbox,%rbp),$sbox
|
| + shr \$3,%rbp # recall "magic" constants!
|
| + add %rbp,$sbox
|
|
|
| - call _x86_64_AES_decrypt
|
| + call _x86_64_AES_decrypt_compact
|
|
|
| - mov $s0,0($out)
|
| + mov 16(%rsp),$out # restore out
|
| + mov 24(%rsp),%rsi # restore saved stack pointer
|
| + mov $s0,0($out) # write output vector
|
| mov $s1,4($out)
|
| mov $s2,8($out)
|
| mov $s3,12($out)
|
|
|
| - pop %r15
|
| - pop %r14
|
| - pop %r13
|
| - pop %r12
|
| - pop %rbp
|
| - pop %rbx
|
| + mov (%rsi),%r15
|
| + mov 8(%rsi),%r14
|
| + mov 16(%rsi),%r13
|
| + mov 24(%rsi),%r12
|
| + mov 32(%rsi),%rbp
|
| + mov 40(%rsi),%rbx
|
| + lea 48(%rsi),%rsp
|
| +.Ldec_epilogue:
|
| ret
|
| .size AES_decrypt,.-AES_decrypt
|
| ___
|
| @@ -718,27 +1254,26 @@
|
| {
|
| $code.=<<___;
|
| movz %dl,%esi # rk[i]>>0
|
| - mov 2(%rbp,%rsi,8),%ebx
|
| + movzb -128(%rbp,%rsi),%ebx
|
| movz %dh,%esi # rk[i]>>8
|
| - and \$0xFF000000,%ebx
|
| + shl \$24,%ebx
|
| xor %ebx,%eax
|
|
|
| - mov 2(%rbp,%rsi,8),%ebx
|
| + movzb -128(%rbp,%rsi),%ebx
|
| shr \$16,%edx
|
| - and \$0x000000FF,%ebx
|
| movz %dl,%esi # rk[i]>>16
|
| xor %ebx,%eax
|
|
|
| - mov 0(%rbp,%rsi,8),%ebx
|
| + movzb -128(%rbp,%rsi),%ebx
|
| movz %dh,%esi # rk[i]>>24
|
| - and \$0x0000FF00,%ebx
|
| + shl \$8,%ebx
|
| xor %ebx,%eax
|
|
|
| - mov 0(%rbp,%rsi,8),%ebx
|
| - and \$0x00FF0000,%ebx
|
| + movzb -128(%rbp,%rsi),%ebx
|
| + shl \$16,%ebx
|
| xor %ebx,%eax
|
|
|
| - xor 2048(%rbp,%rcx,4),%eax # rcon
|
| + xor 1024-128(%rbp,%rcx,4),%eax # rcon
|
| ___
|
| }
|
|
|
| @@ -751,13 +1286,23 @@
|
| AES_set_encrypt_key:
|
| push %rbx
|
| push %rbp
|
| + push %r12 # redundant, but allows to share
|
| + push %r13 # exception handler...
|
| + push %r14
|
| + push %r15
|
| sub \$8,%rsp
|
| +.Lenc_key_prologue:
|
|
|
| call _x86_64_AES_set_encrypt_key
|
|
|
| - mov 8(%rsp),%rbp
|
| - mov 16(%rsp),%rbx
|
| - add \$24,%rsp
|
| + mov 8(%rsp),%r15
|
| + mov 16(%rsp),%r14
|
| + mov 24(%rsp),%r13
|
| + mov 32(%rsp),%r12
|
| + mov 40(%rsp),%rbp
|
| + mov 48(%rsp),%rbx
|
| + add \$56,%rsp
|
| +.Lenc_key_epilogue:
|
| ret
|
| .size AES_set_encrypt_key,.-AES_set_encrypt_key
|
|
|
| @@ -773,9 +1318,19 @@
|
| test \$-1,%rdi
|
| jz .Lbadpointer
|
|
|
| - .picmeup %rbp
|
| - lea AES_Te-.(%rbp),%rbp
|
| + lea .LAES_Te(%rip),%rbp
|
| + lea 2048+128(%rbp),%rbp
|
|
|
| + # prefetch Te4
|
| + mov 0-128(%rbp),%eax
|
| + mov 32-128(%rbp),%ebx
|
| + mov 64-128(%rbp),%r8d
|
| + mov 96-128(%rbp),%edx
|
| + mov 128-128(%rbp),%eax
|
| + mov 160-128(%rbp),%ebx
|
| + mov 192-128(%rbp),%r8d
|
| + mov 224-128(%rbp),%edx
|
| +
|
| cmp \$128,%ecx
|
| je .L10rounds
|
| cmp \$192,%ecx
|
| @@ -786,15 +1341,12 @@
|
| jmp .Lexit
|
|
|
| .L10rounds:
|
| - mov 0(%rsi),%eax # copy first 4 dwords
|
| - mov 4(%rsi),%ebx
|
| - mov 8(%rsi),%ecx
|
| - mov 12(%rsi),%edx
|
| - mov %eax,0(%rdi)
|
| - mov %ebx,4(%rdi)
|
| - mov %ecx,8(%rdi)
|
| - mov %edx,12(%rdi)
|
| + mov 0(%rsi),%rax # copy first 4 dwords
|
| + mov 8(%rsi),%rdx
|
| + mov %rax,0(%rdi)
|
| + mov %rdx,8(%rdi)
|
|
|
| + shr \$32,%rdx
|
| xor %ecx,%ecx
|
| jmp .L10shortcut
|
| .align 4
|
| @@ -822,19 +1374,14 @@
|
| jmp .Lexit
|
|
|
| .L12rounds:
|
| - mov 0(%rsi),%eax # copy first 6 dwords
|
| - mov 4(%rsi),%ebx
|
| - mov 8(%rsi),%ecx
|
| - mov 12(%rsi),%edx
|
| - mov %eax,0(%rdi)
|
| - mov %ebx,4(%rdi)
|
| - mov %ecx,8(%rdi)
|
| - mov %edx,12(%rdi)
|
| - mov 16(%rsi),%ecx
|
| - mov 20(%rsi),%edx
|
| - mov %ecx,16(%rdi)
|
| - mov %edx,20(%rdi)
|
| + mov 0(%rsi),%rax # copy first 6 dwords
|
| + mov 8(%rsi),%rbx
|
| + mov 16(%rsi),%rdx
|
| + mov %rax,0(%rdi)
|
| + mov %rbx,8(%rdi)
|
| + mov %rdx,16(%rdi)
|
|
|
| + shr \$32,%rdx
|
| xor %ecx,%ecx
|
| jmp .L12shortcut
|
| .align 4
|
| @@ -870,30 +1417,23 @@
|
| jmp .Lexit
|
|
|
| .L14rounds:
|
| - mov 0(%rsi),%eax # copy first 8 dwords
|
| - mov 4(%rsi),%ebx
|
| - mov 8(%rsi),%ecx
|
| - mov 12(%rsi),%edx
|
| - mov %eax,0(%rdi)
|
| - mov %ebx,4(%rdi)
|
| - mov %ecx,8(%rdi)
|
| - mov %edx,12(%rdi)
|
| - mov 16(%rsi),%eax
|
| - mov 20(%rsi),%ebx
|
| - mov 24(%rsi),%ecx
|
| - mov 28(%rsi),%edx
|
| - mov %eax,16(%rdi)
|
| - mov %ebx,20(%rdi)
|
| - mov %ecx,24(%rdi)
|
| - mov %edx,28(%rdi)
|
| + mov 0(%rsi),%rax # copy first 8 dwords
|
| + mov 8(%rsi),%rbx
|
| + mov 16(%rsi),%rcx
|
| + mov 24(%rsi),%rdx
|
| + mov %rax,0(%rdi)
|
| + mov %rbx,8(%rdi)
|
| + mov %rcx,16(%rdi)
|
| + mov %rdx,24(%rdi)
|
|
|
| + shr \$32,%rdx
|
| xor %ecx,%ecx
|
| jmp .L14shortcut
|
| .align 4
|
| .L14loop:
|
| + mov 0(%rdi),%eax # rk[0]
|
| mov 28(%rdi),%edx # rk[4]
|
| .L14shortcut:
|
| - mov 0(%rdi),%eax # rk[0]
|
| ___
|
| &enckey ();
|
| $code.=<<___;
|
| @@ -912,24 +1452,23 @@
|
| mov %eax,%edx
|
| mov 16(%rdi),%eax # rk[4]
|
| movz %dl,%esi # rk[11]>>0
|
| - mov 2(%rbp,%rsi,8),%ebx
|
| + movzb -128(%rbp,%rsi),%ebx
|
| movz %dh,%esi # rk[11]>>8
|
| - and \$0x000000FF,%ebx
|
| xor %ebx,%eax
|
|
|
| - mov 0(%rbp,%rsi,8),%ebx
|
| + movzb -128(%rbp,%rsi),%ebx
|
| shr \$16,%edx
|
| - and \$0x0000FF00,%ebx
|
| + shl \$8,%ebx
|
| movz %dl,%esi # rk[11]>>16
|
| xor %ebx,%eax
|
|
|
| - mov 0(%rbp,%rsi,8),%ebx
|
| + movzb -128(%rbp,%rsi),%ebx
|
| movz %dh,%esi # rk[11]>>24
|
| - and \$0x00FF0000,%ebx
|
| + shl \$16,%ebx
|
| xor %ebx,%eax
|
|
|
| - mov 2(%rbp,%rsi,8),%ebx
|
| - and \$0xFF000000,%ebx
|
| + movzb -128(%rbp,%rsi),%ebx
|
| + shl \$24,%ebx
|
| xor %ebx,%eax
|
|
|
| mov %eax,48(%rdi) # rk[12]
|
| @@ -950,29 +1489,61 @@
|
| .Lbadpointer:
|
| mov \$-1,%rax
|
| .Lexit:
|
| - .byte 0xf3,0xc3 # rep ret
|
| + .byte 0xf3,0xc3 # rep ret
|
| .size _x86_64_AES_set_encrypt_key,.-_x86_64_AES_set_encrypt_key
|
| ___
|
|
|
| -sub deckey()
|
| +sub deckey_ref()
|
| { my ($i,$ptr,$te,$td) = @_;
|
| + my ($tp1,$tp2,$tp4,$tp8,$acc)=("%eax","%ebx","%edi","%edx","%r8d");
|
| $code.=<<___;
|
| - mov $i($ptr),%eax
|
| - mov %eax,%edx
|
| - movz %ah,%ebx
|
| - shr \$16,%edx
|
| - and \$0xFF,%eax
|
| - movzb 2($te,%rax,8),%rax
|
| - movzb 2($te,%rbx,8),%rbx
|
| - mov 0($td,%rax,8),%eax
|
| - xor 3($td,%rbx,8),%eax
|
| - movzb %dh,%ebx
|
| - and \$0xFF,%edx
|
| - movzb 2($te,%rdx,8),%rdx
|
| - movzb 2($te,%rbx,8),%rbx
|
| - xor 2($td,%rdx,8),%eax
|
| - xor 1($td,%rbx,8),%eax
|
| - mov %eax,$i($ptr)
|
| + mov $i($ptr),$tp1
|
| + mov $tp1,$acc
|
| + and \$0x80808080,$acc
|
| + mov $acc,$tp4
|
| + shr \$7,$tp4
|
| + lea 0($tp1,$tp1),$tp2
|
| + sub $tp4,$acc
|
| + and \$0xfefefefe,$tp2
|
| + and \$0x1b1b1b1b,$acc
|
| + xor $tp2,$acc
|
| + mov $acc,$tp2
|
| +
|
| + and \$0x80808080,$acc
|
| + mov $acc,$tp8
|
| + shr \$7,$tp8
|
| + lea 0($tp2,$tp2),$tp4
|
| + sub $tp8,$acc
|
| + and \$0xfefefefe,$tp4
|
| + and \$0x1b1b1b1b,$acc
|
| + xor $tp1,$tp2 # tp2^tp1
|
| + xor $tp4,$acc
|
| + mov $acc,$tp4
|
| +
|
| + and \$0x80808080,$acc
|
| + mov $acc,$tp8
|
| + shr \$7,$tp8
|
| + sub $tp8,$acc
|
| + lea 0($tp4,$tp4),$tp8
|
| + xor $tp1,$tp4 # tp4^tp1
|
| + and \$0xfefefefe,$tp8
|
| + and \$0x1b1b1b1b,$acc
|
| + xor $acc,$tp8
|
| +
|
| + xor $tp8,$tp1 # tp1^tp8
|
| + rol \$8,$tp1 # ROTATE(tp1^tp8,8)
|
| + xor $tp8,$tp2 # tp2^tp1^tp8
|
| + xor $tp8,$tp4 # tp4^tp1^tp8
|
| + xor $tp2,$tp8
|
| + xor $tp4,$tp8 # tp8^(tp8^tp4^tp1)^(tp8^tp2^tp1)=tp8^tp4^tp2
|
| +
|
| + xor $tp8,$tp1
|
| + rol \$24,$tp2 # ROTATE(tp2^tp1^tp8,24)
|
| + xor $tp2,$tp1
|
| + rol \$16,$tp4 # ROTATE(tp4^tp1^tp8,16)
|
| + xor $tp4,$tp1
|
| +
|
| + mov $tp1,$i($ptr)
|
| ___
|
| }
|
|
|
| @@ -985,16 +1556,21 @@
|
| AES_set_decrypt_key:
|
| push %rbx
|
| push %rbp
|
| + push %r12
|
| + push %r13
|
| + push %r14
|
| + push %r15
|
| push %rdx # save key schedule
|
| +.Ldec_key_prologue:
|
|
|
| call _x86_64_AES_set_encrypt_key
|
| mov (%rsp),%r8 # restore key schedule
|
| cmp \$0,%eax
|
| jne .Labort
|
|
|
| - mov 240(%r8),%ecx # pull number of rounds
|
| + mov 240(%r8),%r14d # pull number of rounds
|
| xor %rdi,%rdi
|
| - lea (%rdi,%rcx,4),%rcx
|
| + lea (%rdi,%r14d,4),%rcx
|
| mov %r8,%rsi
|
| lea (%r8,%rcx,4),%rdi # pointer to last chunk
|
| .align 4
|
| @@ -1012,30 +1588,39 @@
|
| cmp %rsi,%rdi
|
| jne .Linvert
|
|
|
| - .picmeup %r9
|
| - lea AES_Td-.(%r9),%rdi
|
| - lea AES_Te-AES_Td(%rdi),%r9
|
| + lea .LAES_Te+2048+1024(%rip),%rax # rcon
|
|
|
| - mov %r8,%rsi
|
| - mov 240(%r8),%ecx # pull number of rounds
|
| - sub \$1,%ecx
|
| + mov 40(%rax),$mask80
|
| + mov 48(%rax),$maskfe
|
| + mov 56(%rax),$mask1b
|
| +
|
| + mov %r8,$key
|
| + sub \$1,%r14d
|
| .align 4
|
| .Lpermute:
|
| - lea 16(%rsi),%rsi
|
| + lea 16($key),$key
|
| + mov 0($key),%rax
|
| + mov 8($key),%rcx
|
| ___
|
| - &deckey (0,"%rsi","%r9","%rdi");
|
| - &deckey (4,"%rsi","%r9","%rdi");
|
| - &deckey (8,"%rsi","%r9","%rdi");
|
| - &deckey (12,"%rsi","%r9","%rdi");
|
| + &dectransform ();
|
| $code.=<<___;
|
| - sub \$1,%ecx
|
| + mov %eax,0($key)
|
| + mov %ebx,4($key)
|
| + mov %ecx,8($key)
|
| + mov %edx,12($key)
|
| + sub \$1,%r14d
|
| jnz .Lpermute
|
|
|
| xor %rax,%rax
|
| .Labort:
|
| - mov 8(%rsp),%rbp
|
| - mov 16(%rsp),%rbx
|
| - add \$24,%rsp
|
| + mov 8(%rsp),%r15
|
| + mov 16(%rsp),%r14
|
| + mov 24(%rsp),%r13
|
| + mov 32(%rsp),%r12
|
| + mov 40(%rsp),%rbp
|
| + mov 48(%rsp),%rbx
|
| + add \$56,%rsp
|
| +.Ldec_key_epilogue:
|
| ret
|
| .size AES_set_decrypt_key,.-AES_set_decrypt_key
|
| ___
|
| @@ -1046,47 +1631,59 @@
|
| {
|
| # stack frame layout
|
| # -8(%rsp) return address
|
| -my $_rsp="0(%rsp)"; # saved %rsp
|
| -my $_len="8(%rsp)"; # copy of 3rd parameter, length
|
| -my $_key="16(%rsp)"; # copy of 4th parameter, key
|
| -my $_ivp="24(%rsp)"; # copy of 5th parameter, ivp
|
| -my $keyp="32(%rsp)"; # one to pass as $key
|
| -my $ivec="40(%rsp)"; # ivec[16]
|
| -my $aes_key="56(%rsp)"; # copy of aes_key
|
| -my $mark="56+240(%rsp)"; # copy of aes_key->rounds
|
| +my $keyp="0(%rsp)"; # one to pass as $key
|
| +my $keyend="8(%rsp)"; # &(keyp->rd_key[4*keyp->rounds])
|
| +my $_rsp="16(%rsp)"; # saved %rsp
|
| +my $_inp="24(%rsp)"; # copy of 1st parameter, inp
|
| +my $_out="32(%rsp)"; # copy of 2nd parameter, out
|
| +my $_len="40(%rsp)"; # copy of 3rd parameter, length
|
| +my $_key="48(%rsp)"; # copy of 4th parameter, key
|
| +my $_ivp="56(%rsp)"; # copy of 5th parameter, ivp
|
| +my $ivec="64(%rsp)"; # ivec[16]
|
| +my $aes_key="80(%rsp)"; # copy of aes_key
|
| +my $mark="80+240(%rsp)"; # copy of aes_key->rounds
|
|
|
| $code.=<<___;
|
| .globl AES_cbc_encrypt
|
| .type AES_cbc_encrypt,\@function,6
|
| .align 16
|
| +.extern OPENSSL_ia32cap_P
|
| AES_cbc_encrypt:
|
| cmp \$0,%rdx # check length
|
| - je .Lcbc_just_ret
|
| + je .Lcbc_epilogue
|
| + pushfq
|
| push %rbx
|
| push %rbp
|
| push %r12
|
| push %r13
|
| push %r14
|
| push %r15
|
| - pushfq
|
| +.Lcbc_prologue:
|
| +
|
| cld
|
| mov %r9d,%r9d # clear upper half of enc
|
|
|
| - .picmeup $sbox
|
| -.Lcbc_pic_point:
|
| -
|
| + lea .LAES_Te(%rip),$sbox
|
| cmp \$0,%r9
|
| - je .LDECRYPT
|
| + jne .Lcbc_picked_te
|
| + lea .LAES_Td(%rip),$sbox
|
| +.Lcbc_picked_te:
|
|
|
| - lea AES_Te-.Lcbc_pic_point($sbox),$sbox
|
| + mov OPENSSL_ia32cap_P(%rip),%r10d
|
| + cmp \$$speed_limit,%rdx
|
| + jb .Lcbc_slow_prologue
|
| + test \$15,%rdx
|
| + jnz .Lcbc_slow_prologue
|
| + bt \$28,%r10d
|
| + jc .Lcbc_slow_prologue
|
|
|
| # allocate aligned stack frame...
|
| - lea -64-248(%rsp),$key
|
| + lea -88-248(%rsp),$key
|
| and \$-64,$key
|
|
|
| - # ... and make it doesn't alias with AES_Te modulo 4096
|
| + # ... and make sure it doesn't alias with AES_T[ed] modulo 4096
|
| mov $sbox,%r10
|
| - lea 2048($sbox),%r11
|
| + lea 2304($sbox),%r11
|
| mov $key,%r12
|
| and \$0xFFF,%r10 # s = $sbox&0xfff
|
| and \$0xFFF,%r11 # e = ($sbox+2048)&0xfff
|
| @@ -1106,22 +1703,27 @@
|
| .Lcbc_te_ok:
|
|
|
| xchg %rsp,$key
|
| - add \$8,%rsp # reserve for return address!
|
| + #add \$8,%rsp # reserve for return address!
|
| mov $key,$_rsp # save %rsp
|
| +.Lcbc_fast_body:
|
| + mov %rdi,$_inp # save copy of inp
|
| + mov %rsi,$_out # save copy of out
|
| mov %rdx,$_len # save copy of len
|
| mov %rcx,$_key # save copy of key
|
| mov %r8,$_ivp # save copy of ivp
|
| movl \$0,$mark # copy of aes_key->rounds = 0;
|
| mov %r8,%rbp # rearrange input arguments
|
| + mov %r9,%rbx
|
| mov %rsi,$out
|
| mov %rdi,$inp
|
| mov %rcx,$key
|
|
|
| + mov 240($key),%eax # key->rounds
|
| # do we copy key schedule to stack?
|
| mov $key,%r10
|
| sub $sbox,%r10
|
| and \$0xfff,%r10
|
| - cmp \$2048,%r10
|
| + cmp \$2304,%r10
|
| jb .Lcbc_do_ecopy
|
| cmp \$4096-248,%r10
|
| jb .Lcbc_skip_ecopy
|
| @@ -1132,12 +1734,11 @@
|
| lea $aes_key,$key
|
| mov \$240/8,%ecx
|
| .long 0x90A548F3 # rep movsq
|
| - mov (%rsi),%eax # copy aes_key->rounds
|
| - mov %eax,(%rdi)
|
| + mov %eax,(%rdi) # copy aes_key->rounds
|
| .Lcbc_skip_ecopy:
|
| mov $key,$keyp # save key pointer
|
|
|
| - mov \$16,%ecx
|
| + mov \$18,%ecx
|
| .align 4
|
| .Lcbc_prefetch_te:
|
| mov 0($sbox),%r10
|
| @@ -1147,184 +1748,77 @@
|
| lea 128($sbox),$sbox
|
| sub \$1,%ecx
|
| jnz .Lcbc_prefetch_te
|
| - sub \$2048,$sbox
|
| + lea -2304($sbox),$sbox
|
|
|
| - test \$-16,%rdx # check upon length
|
| - mov %rdx,%r10
|
| + cmp \$0,%rbx
|
| + je .LFAST_DECRYPT
|
| +
|
| +#----------------------------- ENCRYPT -----------------------------#
|
| mov 0(%rbp),$s0 # load iv
|
| mov 4(%rbp),$s1
|
| mov 8(%rbp),$s2
|
| mov 12(%rbp),$s3
|
| - jz .Lcbc_enc_tail # short input...
|
|
|
| .align 4
|
| -.Lcbc_enc_loop:
|
| +.Lcbc_fast_enc_loop:
|
| xor 0($inp),$s0
|
| xor 4($inp),$s1
|
| xor 8($inp),$s2
|
| xor 12($inp),$s3
|
| - mov $inp,$ivec # if ($verticalspin) save inp
|
| + mov $keyp,$key # restore key
|
| + mov $inp,$_inp # if ($verticalspin) save inp
|
|
|
| - mov $keyp,$key # restore key
|
| call _x86_64_AES_encrypt
|
|
|
| - mov $ivec,$inp # if ($verticalspin) restore inp
|
| + mov $_inp,$inp # if ($verticalspin) restore inp
|
| + mov $_len,%r10
|
| mov $s0,0($out)
|
| mov $s1,4($out)
|
| mov $s2,8($out)
|
| mov $s3,12($out)
|
|
|
| - mov $_len,%r10
|
| lea 16($inp),$inp
|
| lea 16($out),$out
|
| sub \$16,%r10
|
| test \$-16,%r10
|
| mov %r10,$_len
|
| - jnz .Lcbc_enc_loop
|
| - test \$15,%r10
|
| - jnz .Lcbc_enc_tail
|
| + jnz .Lcbc_fast_enc_loop
|
| mov $_ivp,%rbp # restore ivp
|
| mov $s0,0(%rbp) # save ivec
|
| mov $s1,4(%rbp)
|
| mov $s2,8(%rbp)
|
| mov $s3,12(%rbp)
|
|
|
| -.align 4
|
| -.Lcbc_cleanup:
|
| - cmpl \$0,$mark # was the key schedule copied?
|
| - lea $aes_key,%rdi
|
| - je .Lcbc_exit
|
| - mov \$240/8,%ecx
|
| - xor %rax,%rax
|
| - .long 0x90AB48F3 # rep stosq
|
| -.Lcbc_exit:
|
| - mov $_rsp,%rsp
|
| - popfq
|
| - pop %r15
|
| - pop %r14
|
| - pop %r13
|
| - pop %r12
|
| - pop %rbp
|
| - pop %rbx
|
| -.Lcbc_just_ret:
|
| - ret
|
| -.align 4
|
| -.Lcbc_enc_tail:
|
| - mov %rax,%r11
|
| - mov %rcx,%r12
|
| - mov %r10,%rcx
|
| - mov $inp,%rsi
|
| - mov $out,%rdi
|
| - .long 0xF689A4F3 # rep movsb
|
| - mov \$16,%rcx # zero tail
|
| - sub %r10,%rcx
|
| - xor %rax,%rax
|
| - .long 0xF689AAF3 # rep stosb
|
| - mov $out,$inp # this is not a mistake!
|
| - movq \$16,$_len # len=16
|
| - mov %r11,%rax
|
| - mov %r12,%rcx
|
| - jmp .Lcbc_enc_loop # one more spin...
|
| + jmp .Lcbc_fast_cleanup
|
| +
|
| #----------------------------- DECRYPT -----------------------------#
|
| .align 16
|
| -.LDECRYPT:
|
| - lea AES_Td-.Lcbc_pic_point($sbox),$sbox
|
| -
|
| - # allocate aligned stack frame...
|
| - lea -64-248(%rsp),$key
|
| - and \$-64,$key
|
| -
|
| - # ... and make it doesn't alias with AES_Td modulo 4096
|
| - mov $sbox,%r10
|
| - lea 2304($sbox),%r11
|
| - mov $key,%r12
|
| - and \$0xFFF,%r10 # s = $sbox&0xfff
|
| - and \$0xFFF,%r11 # e = ($sbox+2048+256)&0xfff
|
| - and \$0xFFF,%r12 # p = %rsp&0xfff
|
| -
|
| - cmp %r11,%r12 # if (p=>e) %rsp =- (p-e);
|
| - jb .Lcbc_td_break_out
|
| - sub %r11,%r12
|
| - sub %r12,$key
|
| - jmp .Lcbc_td_ok
|
| -.Lcbc_td_break_out: # else %rsp -= (p-s)&0xfff + framesz
|
| - sub %r10,%r12
|
| - and \$0xFFF,%r12
|
| - add \$320,%r12
|
| - sub %r12,$key
|
| -.align 4
|
| -.Lcbc_td_ok:
|
| -
|
| - xchg %rsp,$key
|
| - add \$8,%rsp # reserve for return address!
|
| - mov $key,$_rsp # save %rsp
|
| - mov %rdx,$_len # save copy of len
|
| - mov %rcx,$_key # save copy of key
|
| - mov %r8,$_ivp # save copy of ivp
|
| - movl \$0,$mark # copy of aes_key->rounds = 0;
|
| - mov %r8,%rbp # rearrange input arguments
|
| - mov %rsi,$out
|
| - mov %rdi,$inp
|
| - mov %rcx,$key
|
| -
|
| - # do we copy key schedule to stack?
|
| - mov $key,%r10
|
| - sub $sbox,%r10
|
| - and \$0xfff,%r10
|
| - cmp \$2304,%r10
|
| - jb .Lcbc_do_dcopy
|
| - cmp \$4096-248,%r10
|
| - jb .Lcbc_skip_dcopy
|
| -.align 4
|
| -.Lcbc_do_dcopy:
|
| - mov $key,%rsi
|
| - lea $aes_key,%rdi
|
| - lea $aes_key,$key
|
| - mov \$240/8,%ecx
|
| - .long 0x90A548F3 # rep movsq
|
| - mov (%rsi),%eax # copy aes_key->rounds
|
| - mov %eax,(%rdi)
|
| -.Lcbc_skip_dcopy:
|
| - mov $key,$keyp # save key pointer
|
| -
|
| - mov \$18,%ecx
|
| -.align 4
|
| -.Lcbc_prefetch_td:
|
| - mov 0($sbox),%r10
|
| - mov 32($sbox),%r11
|
| - mov 64($sbox),%r12
|
| - mov 96($sbox),%r13
|
| - lea 128($sbox),$sbox
|
| - sub \$1,%ecx
|
| - jnz .Lcbc_prefetch_td
|
| - sub \$2304,$sbox
|
| -
|
| +.LFAST_DECRYPT:
|
| cmp $inp,$out
|
| - je .Lcbc_dec_in_place
|
| + je .Lcbc_fast_dec_in_place
|
|
|
| mov %rbp,$ivec
|
| .align 4
|
| -.Lcbc_dec_loop:
|
| - mov 0($inp),$s0 # read input
|
| +.Lcbc_fast_dec_loop:
|
| + mov 0($inp),$s0 # read input
|
| mov 4($inp),$s1
|
| mov 8($inp),$s2
|
| mov 12($inp),$s3
|
| - mov $inp,8+$ivec # if ($verticalspin) save inp
|
| + mov $keyp,$key # restore key
|
| + mov $inp,$_inp # if ($verticalspin) save inp
|
|
|
| - mov $keyp,$key # restore key
|
| call _x86_64_AES_decrypt
|
|
|
| mov $ivec,%rbp # load ivp
|
| - mov 8+$ivec,$inp # if ($verticalspin) restore inp
|
| + mov $_inp,$inp # if ($verticalspin) restore inp
|
| + mov $_len,%r10 # load len
|
| xor 0(%rbp),$s0 # xor iv
|
| xor 4(%rbp),$s1
|
| xor 8(%rbp),$s2
|
| xor 12(%rbp),$s3
|
| mov $inp,%rbp # current input, next iv
|
|
|
| - mov $_len,%r10 # load len
|
| sub \$16,%r10
|
| - jc .Lcbc_dec_partial
|
| mov %r10,$_len # update len
|
| mov %rbp,$ivec # update ivp
|
|
|
| @@ -1335,81 +1829,281 @@
|
|
|
| lea 16($inp),$inp
|
| lea 16($out),$out
|
| - jnz .Lcbc_dec_loop
|
| -.Lcbc_dec_end:
|
| + jnz .Lcbc_fast_dec_loop
|
| mov $_ivp,%r12 # load user ivp
|
| mov 0(%rbp),%r10 # load iv
|
| mov 8(%rbp),%r11
|
| mov %r10,0(%r12) # copy back to user
|
| mov %r11,8(%r12)
|
| - jmp .Lcbc_cleanup
|
| + jmp .Lcbc_fast_cleanup
|
|
|
| +.align 16
|
| +.Lcbc_fast_dec_in_place:
|
| + mov 0(%rbp),%r10 # copy iv to stack
|
| + mov 8(%rbp),%r11
|
| + mov %r10,0+$ivec
|
| + mov %r11,8+$ivec
|
| .align 4
|
| -.Lcbc_dec_partial:
|
| - mov $s0,0+$ivec # dump output to stack
|
| - mov $s1,4+$ivec
|
| - mov $s2,8+$ivec
|
| - mov $s3,12+$ivec
|
| - mov $out,%rdi
|
| - lea $ivec,%rsi
|
| - mov \$16,%rcx
|
| - add %r10,%rcx # number of bytes to copy
|
| - .long 0xF689A4F3 # rep movsb
|
| - jmp .Lcbc_dec_end
|
| -
|
| -.align 16
|
| -.Lcbc_dec_in_place:
|
| +.Lcbc_fast_dec_in_place_loop:
|
| mov 0($inp),$s0 # load input
|
| mov 4($inp),$s1
|
| mov 8($inp),$s2
|
| mov 12($inp),$s3
|
| + mov $keyp,$key # restore key
|
| + mov $inp,$_inp # if ($verticalspin) save inp
|
|
|
| - mov $inp,$ivec # if ($verticalspin) save inp
|
| - mov $keyp,$key
|
| call _x86_64_AES_decrypt
|
|
|
| - mov $ivec,$inp # if ($verticalspin) restore inp
|
| - mov $_ivp,%rbp
|
| - xor 0(%rbp),$s0
|
| - xor 4(%rbp),$s1
|
| - xor 8(%rbp),$s2
|
| - xor 12(%rbp),$s3
|
| + mov $_inp,$inp # if ($verticalspin) restore inp
|
| + mov $_len,%r10
|
| + xor 0+$ivec,$s0
|
| + xor 4+$ivec,$s1
|
| + xor 8+$ivec,$s2
|
| + xor 12+$ivec,$s3
|
|
|
| - mov 0($inp),%r10 # copy input to iv
|
| - mov 8($inp),%r11
|
| - mov %r10,0(%rbp)
|
| - mov %r11,8(%rbp)
|
| + mov 0($inp),%r11 # load input
|
| + mov 8($inp),%r12
|
| + sub \$16,%r10
|
| + jz .Lcbc_fast_dec_in_place_done
|
|
|
| + mov %r11,0+$ivec # copy input to iv
|
| + mov %r12,8+$ivec
|
| +
|
| mov $s0,0($out) # save output [zaps input]
|
| mov $s1,4($out)
|
| mov $s2,8($out)
|
| mov $s3,12($out)
|
|
|
| - mov $_len,%rcx
|
| lea 16($inp),$inp
|
| lea 16($out),$out
|
| - sub \$16,%rcx
|
| - jc .Lcbc_dec_in_place_partial
|
| - mov %rcx,$_len
|
| - jnz .Lcbc_dec_in_place
|
| - jmp .Lcbc_cleanup
|
| + mov %r10,$_len
|
| + jmp .Lcbc_fast_dec_in_place_loop
|
| +.Lcbc_fast_dec_in_place_done:
|
| + mov $_ivp,%rdi
|
| + mov %r11,0(%rdi) # copy iv back to user
|
| + mov %r12,8(%rdi)
|
|
|
| + mov $s0,0($out) # save output [zaps input]
|
| + mov $s1,4($out)
|
| + mov $s2,8($out)
|
| + mov $s3,12($out)
|
| +
|
| .align 4
|
| -.Lcbc_dec_in_place_partial:
|
| - # one can argue if this is actually required
|
| - lea ($out,%rcx),%rdi
|
| - lea (%rbp,%rcx),%rsi
|
| - neg %rcx
|
| - .long 0xF689A4F3 # rep movsb # restore tail
|
| - jmp .Lcbc_cleanup
|
| +.Lcbc_fast_cleanup:
|
| + cmpl \$0,$mark # was the key schedule copied?
|
| + lea $aes_key,%rdi
|
| + je .Lcbc_exit
|
| + mov \$240/8,%ecx
|
| + xor %rax,%rax
|
| + .long 0x90AB48F3 # rep stosq
|
| +
|
| + jmp .Lcbc_exit
|
| +
|
| +#--------------------------- SLOW ROUTINE ---------------------------#
|
| +.align 16
|
| +.Lcbc_slow_prologue:
|
| + # allocate aligned stack frame...
|
| + lea -88(%rsp),%rbp
|
| + and \$-64,%rbp
|
| + # ... just "above" key schedule
|
| + lea -88-63(%rcx),%r10
|
| + sub %rbp,%r10
|
| + neg %r10
|
| + and \$0x3c0,%r10
|
| + sub %r10,%rbp
|
| +
|
| + xchg %rsp,%rbp
|
| + #add \$8,%rsp # reserve for return address!
|
| + mov %rbp,$_rsp # save %rsp
|
| +.Lcbc_slow_body:
|
| + #mov %rdi,$_inp # save copy of inp
|
| + #mov %rsi,$_out # save copy of out
|
| + #mov %rdx,$_len # save copy of len
|
| + #mov %rcx,$_key # save copy of key
|
| + mov %r8,$_ivp # save copy of ivp
|
| + mov %r8,%rbp # rearrange input arguments
|
| + mov %r9,%rbx
|
| + mov %rsi,$out
|
| + mov %rdi,$inp
|
| + mov %rcx,$key
|
| + mov %rdx,%r10
|
| +
|
| + mov 240($key),%eax
|
| + mov $key,$keyp # save key pointer
|
| + shl \$4,%eax
|
| + lea ($key,%rax),%rax
|
| + mov %rax,$keyend
|
| +
|
| + # pick Te4 copy which can't "overlap" with stack frame or key scdedule
|
| + lea 2048($sbox),$sbox
|
| + lea 768-8(%rsp),%rax
|
| + sub $sbox,%rax
|
| + and \$0x300,%rax
|
| + lea ($sbox,%rax),$sbox
|
| +
|
| + cmp \$0,%rbx
|
| + je .LSLOW_DECRYPT
|
| +
|
| +#--------------------------- SLOW ENCRYPT ---------------------------#
|
| + test \$-16,%r10 # check upon length
|
| + mov 0(%rbp),$s0 # load iv
|
| + mov 4(%rbp),$s1
|
| + mov 8(%rbp),$s2
|
| + mov 12(%rbp),$s3
|
| + jz .Lcbc_slow_enc_tail # short input...
|
| +
|
| +.align 4
|
| +.Lcbc_slow_enc_loop:
|
| + xor 0($inp),$s0
|
| + xor 4($inp),$s1
|
| + xor 8($inp),$s2
|
| + xor 12($inp),$s3
|
| + mov $keyp,$key # restore key
|
| + mov $inp,$_inp # save inp
|
| + mov $out,$_out # save out
|
| + mov %r10,$_len # save len
|
| +
|
| + call _x86_64_AES_encrypt_compact
|
| +
|
| + mov $_inp,$inp # restore inp
|
| + mov $_out,$out # restore out
|
| + mov $_len,%r10 # restore len
|
| + mov $s0,0($out)
|
| + mov $s1,4($out)
|
| + mov $s2,8($out)
|
| + mov $s3,12($out)
|
| +
|
| + lea 16($inp),$inp
|
| + lea 16($out),$out
|
| + sub \$16,%r10
|
| + test \$-16,%r10
|
| + jnz .Lcbc_slow_enc_loop
|
| + test \$15,%r10
|
| + jnz .Lcbc_slow_enc_tail
|
| + mov $_ivp,%rbp # restore ivp
|
| + mov $s0,0(%rbp) # save ivec
|
| + mov $s1,4(%rbp)
|
| + mov $s2,8(%rbp)
|
| + mov $s3,12(%rbp)
|
| +
|
| + jmp .Lcbc_exit
|
| +
|
| +.align 4
|
| +.Lcbc_slow_enc_tail:
|
| + mov %rax,%r11
|
| + mov %rcx,%r12
|
| + mov %r10,%rcx
|
| + mov $inp,%rsi
|
| + mov $out,%rdi
|
| + .long 0x9066A4F3 # rep movsb
|
| + mov \$16,%rcx # zero tail
|
| + sub %r10,%rcx
|
| + xor %rax,%rax
|
| + .long 0x9066AAF3 # rep stosb
|
| + mov $out,$inp # this is not a mistake!
|
| + mov \$16,%r10 # len=16
|
| + mov %r11,%rax
|
| + mov %r12,%rcx
|
| + jmp .Lcbc_slow_enc_loop # one more spin...
|
| +#--------------------------- SLOW DECRYPT ---------------------------#
|
| +.align 16
|
| +.LSLOW_DECRYPT:
|
| + shr \$3,%rax
|
| + add %rax,$sbox # recall "magic" constants!
|
| +
|
| + mov 0(%rbp),%r11 # copy iv to stack
|
| + mov 8(%rbp),%r12
|
| + mov %r11,0+$ivec
|
| + mov %r12,8+$ivec
|
| +
|
| +.align 4
|
| +.Lcbc_slow_dec_loop:
|
| + mov 0($inp),$s0 # load input
|
| + mov 4($inp),$s1
|
| + mov 8($inp),$s2
|
| + mov 12($inp),$s3
|
| + mov $keyp,$key # restore key
|
| + mov $inp,$_inp # save inp
|
| + mov $out,$_out # save out
|
| + mov %r10,$_len # save len
|
| +
|
| + call _x86_64_AES_decrypt_compact
|
| +
|
| + mov $_inp,$inp # restore inp
|
| + mov $_out,$out # restore out
|
| + mov $_len,%r10
|
| + xor 0+$ivec,$s0
|
| + xor 4+$ivec,$s1
|
| + xor 8+$ivec,$s2
|
| + xor 12+$ivec,$s3
|
| +
|
| + mov 0($inp),%r11 # load input
|
| + mov 8($inp),%r12
|
| + sub \$16,%r10
|
| + jc .Lcbc_slow_dec_partial
|
| + jz .Lcbc_slow_dec_done
|
| +
|
| + mov %r11,0+$ivec # copy input to iv
|
| + mov %r12,8+$ivec
|
| +
|
| + mov $s0,0($out) # save output [can zap input]
|
| + mov $s1,4($out)
|
| + mov $s2,8($out)
|
| + mov $s3,12($out)
|
| +
|
| + lea 16($inp),$inp
|
| + lea 16($out),$out
|
| + jmp .Lcbc_slow_dec_loop
|
| +.Lcbc_slow_dec_done:
|
| + mov $_ivp,%rdi
|
| + mov %r11,0(%rdi) # copy iv back to user
|
| + mov %r12,8(%rdi)
|
| +
|
| + mov $s0,0($out) # save output [can zap input]
|
| + mov $s1,4($out)
|
| + mov $s2,8($out)
|
| + mov $s3,12($out)
|
| +
|
| + jmp .Lcbc_exit
|
| +
|
| +.align 4
|
| +.Lcbc_slow_dec_partial:
|
| + mov $_ivp,%rdi
|
| + mov %r11,0(%rdi) # copy iv back to user
|
| + mov %r12,8(%rdi)
|
| +
|
| + mov $s0,0+$ivec # save output to stack
|
| + mov $s1,4+$ivec
|
| + mov $s2,8+$ivec
|
| + mov $s3,12+$ivec
|
| +
|
| + mov $out,%rdi
|
| + lea $ivec,%rsi
|
| + lea 16(%r10),%rcx
|
| + .long 0x9066A4F3 # rep movsb
|
| + jmp .Lcbc_exit
|
| +
|
| +.align 16
|
| +.Lcbc_exit:
|
| + mov $_rsp,%rsi
|
| + mov (%rsi),%r15
|
| + mov 8(%rsi),%r14
|
| + mov 16(%rsi),%r13
|
| + mov 24(%rsi),%r12
|
| + mov 32(%rsi),%rbp
|
| + mov 40(%rsi),%rbx
|
| + lea 48(%rsi),%rsp
|
| +.Lcbc_popfq:
|
| + popfq
|
| +.Lcbc_epilogue:
|
| + ret
|
| .size AES_cbc_encrypt,.-AES_cbc_encrypt
|
| ___
|
| }
|
|
|
| $code.=<<___;
|
| -.globl AES_Te
|
| .align 64
|
| -AES_Te:
|
| +.LAES_Te:
|
| ___
|
| &_data_word(0xa56363c6, 0x847c7cf8, 0x997777ee, 0x8d7b7bf6);
|
| &_data_word(0x0df2f2ff, 0xbd6b6bd6, 0xb16f6fde, 0x54c5c591);
|
| @@ -1475,16 +2169,149 @@
|
| &_data_word(0xdabfbf65, 0x31e6e6d7, 0xc6424284, 0xb86868d0);
|
| &_data_word(0xc3414182, 0xb0999929, 0x772d2d5a, 0x110f0f1e);
|
| &_data_word(0xcbb0b07b, 0xfc5454a8, 0xd6bbbb6d, 0x3a16162c);
|
| +
|
| +#Te4 # four copies of Te4 to choose from to avoid L1 aliasing
|
| + &data_byte(0x63, 0x7c, 0x77, 0x7b, 0xf2, 0x6b, 0x6f, 0xc5);
|
| + &data_byte(0x30, 0x01, 0x67, 0x2b, 0xfe, 0xd7, 0xab, 0x76);
|
| + &data_byte(0xca, 0x82, 0xc9, 0x7d, 0xfa, 0x59, 0x47, 0xf0);
|
| + &data_byte(0xad, 0xd4, 0xa2, 0xaf, 0x9c, 0xa4, 0x72, 0xc0);
|
| + &data_byte(0xb7, 0xfd, 0x93, 0x26, 0x36, 0x3f, 0xf7, 0xcc);
|
| + &data_byte(0x34, 0xa5, 0xe5, 0xf1, 0x71, 0xd8, 0x31, 0x15);
|
| + &data_byte(0x04, 0xc7, 0x23, 0xc3, 0x18, 0x96, 0x05, 0x9a);
|
| + &data_byte(0x07, 0x12, 0x80, 0xe2, 0xeb, 0x27, 0xb2, 0x75);
|
| + &data_byte(0x09, 0x83, 0x2c, 0x1a, 0x1b, 0x6e, 0x5a, 0xa0);
|
| + &data_byte(0x52, 0x3b, 0xd6, 0xb3, 0x29, 0xe3, 0x2f, 0x84);
|
| + &data_byte(0x53, 0xd1, 0x00, 0xed, 0x20, 0xfc, 0xb1, 0x5b);
|
| + &data_byte(0x6a, 0xcb, 0xbe, 0x39, 0x4a, 0x4c, 0x58, 0xcf);
|
| + &data_byte(0xd0, 0xef, 0xaa, 0xfb, 0x43, 0x4d, 0x33, 0x85);
|
| + &data_byte(0x45, 0xf9, 0x02, 0x7f, 0x50, 0x3c, 0x9f, 0xa8);
|
| + &data_byte(0x51, 0xa3, 0x40, 0x8f, 0x92, 0x9d, 0x38, 0xf5);
|
| + &data_byte(0xbc, 0xb6, 0xda, 0x21, 0x10, 0xff, 0xf3, 0xd2);
|
| + &data_byte(0xcd, 0x0c, 0x13, 0xec, 0x5f, 0x97, 0x44, 0x17);
|
| + &data_byte(0xc4, 0xa7, 0x7e, 0x3d, 0x64, 0x5d, 0x19, 0x73);
|
| + &data_byte(0x60, 0x81, 0x4f, 0xdc, 0x22, 0x2a, 0x90, 0x88);
|
| + &data_byte(0x46, 0xee, 0xb8, 0x14, 0xde, 0x5e, 0x0b, 0xdb);
|
| + &data_byte(0xe0, 0x32, 0x3a, 0x0a, 0x49, 0x06, 0x24, 0x5c);
|
| + &data_byte(0xc2, 0xd3, 0xac, 0x62, 0x91, 0x95, 0xe4, 0x79);
|
| + &data_byte(0xe7, 0xc8, 0x37, 0x6d, 0x8d, 0xd5, 0x4e, 0xa9);
|
| + &data_byte(0x6c, 0x56, 0xf4, 0xea, 0x65, 0x7a, 0xae, 0x08);
|
| + &data_byte(0xba, 0x78, 0x25, 0x2e, 0x1c, 0xa6, 0xb4, 0xc6);
|
| + &data_byte(0xe8, 0xdd, 0x74, 0x1f, 0x4b, 0xbd, 0x8b, 0x8a);
|
| + &data_byte(0x70, 0x3e, 0xb5, 0x66, 0x48, 0x03, 0xf6, 0x0e);
|
| + &data_byte(0x61, 0x35, 0x57, 0xb9, 0x86, 0xc1, 0x1d, 0x9e);
|
| + &data_byte(0xe1, 0xf8, 0x98, 0x11, 0x69, 0xd9, 0x8e, 0x94);
|
| + &data_byte(0x9b, 0x1e, 0x87, 0xe9, 0xce, 0x55, 0x28, 0xdf);
|
| + &data_byte(0x8c, 0xa1, 0x89, 0x0d, 0xbf, 0xe6, 0x42, 0x68);
|
| + &data_byte(0x41, 0x99, 0x2d, 0x0f, 0xb0, 0x54, 0xbb, 0x16);
|
| +
|
| + &data_byte(0x63, 0x7c, 0x77, 0x7b, 0xf2, 0x6b, 0x6f, 0xc5);
|
| + &data_byte(0x30, 0x01, 0x67, 0x2b, 0xfe, 0xd7, 0xab, 0x76);
|
| + &data_byte(0xca, 0x82, 0xc9, 0x7d, 0xfa, 0x59, 0x47, 0xf0);
|
| + &data_byte(0xad, 0xd4, 0xa2, 0xaf, 0x9c, 0xa4, 0x72, 0xc0);
|
| + &data_byte(0xb7, 0xfd, 0x93, 0x26, 0x36, 0x3f, 0xf7, 0xcc);
|
| + &data_byte(0x34, 0xa5, 0xe5, 0xf1, 0x71, 0xd8, 0x31, 0x15);
|
| + &data_byte(0x04, 0xc7, 0x23, 0xc3, 0x18, 0x96, 0x05, 0x9a);
|
| + &data_byte(0x07, 0x12, 0x80, 0xe2, 0xeb, 0x27, 0xb2, 0x75);
|
| + &data_byte(0x09, 0x83, 0x2c, 0x1a, 0x1b, 0x6e, 0x5a, 0xa0);
|
| + &data_byte(0x52, 0x3b, 0xd6, 0xb3, 0x29, 0xe3, 0x2f, 0x84);
|
| + &data_byte(0x53, 0xd1, 0x00, 0xed, 0x20, 0xfc, 0xb1, 0x5b);
|
| + &data_byte(0x6a, 0xcb, 0xbe, 0x39, 0x4a, 0x4c, 0x58, 0xcf);
|
| + &data_byte(0xd0, 0xef, 0xaa, 0xfb, 0x43, 0x4d, 0x33, 0x85);
|
| + &data_byte(0x45, 0xf9, 0x02, 0x7f, 0x50, 0x3c, 0x9f, 0xa8);
|
| + &data_byte(0x51, 0xa3, 0x40, 0x8f, 0x92, 0x9d, 0x38, 0xf5);
|
| + &data_byte(0xbc, 0xb6, 0xda, 0x21, 0x10, 0xff, 0xf3, 0xd2);
|
| + &data_byte(0xcd, 0x0c, 0x13, 0xec, 0x5f, 0x97, 0x44, 0x17);
|
| + &data_byte(0xc4, 0xa7, 0x7e, 0x3d, 0x64, 0x5d, 0x19, 0x73);
|
| + &data_byte(0x60, 0x81, 0x4f, 0xdc, 0x22, 0x2a, 0x90, 0x88);
|
| + &data_byte(0x46, 0xee, 0xb8, 0x14, 0xde, 0x5e, 0x0b, 0xdb);
|
| + &data_byte(0xe0, 0x32, 0x3a, 0x0a, 0x49, 0x06, 0x24, 0x5c);
|
| + &data_byte(0xc2, 0xd3, 0xac, 0x62, 0x91, 0x95, 0xe4, 0x79);
|
| + &data_byte(0xe7, 0xc8, 0x37, 0x6d, 0x8d, 0xd5, 0x4e, 0xa9);
|
| + &data_byte(0x6c, 0x56, 0xf4, 0xea, 0x65, 0x7a, 0xae, 0x08);
|
| + &data_byte(0xba, 0x78, 0x25, 0x2e, 0x1c, 0xa6, 0xb4, 0xc6);
|
| + &data_byte(0xe8, 0xdd, 0x74, 0x1f, 0x4b, 0xbd, 0x8b, 0x8a);
|
| + &data_byte(0x70, 0x3e, 0xb5, 0x66, 0x48, 0x03, 0xf6, 0x0e);
|
| + &data_byte(0x61, 0x35, 0x57, 0xb9, 0x86, 0xc1, 0x1d, 0x9e);
|
| + &data_byte(0xe1, 0xf8, 0x98, 0x11, 0x69, 0xd9, 0x8e, 0x94);
|
| + &data_byte(0x9b, 0x1e, 0x87, 0xe9, 0xce, 0x55, 0x28, 0xdf);
|
| + &data_byte(0x8c, 0xa1, 0x89, 0x0d, 0xbf, 0xe6, 0x42, 0x68);
|
| + &data_byte(0x41, 0x99, 0x2d, 0x0f, 0xb0, 0x54, 0xbb, 0x16);
|
| +
|
| + &data_byte(0x63, 0x7c, 0x77, 0x7b, 0xf2, 0x6b, 0x6f, 0xc5);
|
| + &data_byte(0x30, 0x01, 0x67, 0x2b, 0xfe, 0xd7, 0xab, 0x76);
|
| + &data_byte(0xca, 0x82, 0xc9, 0x7d, 0xfa, 0x59, 0x47, 0xf0);
|
| + &data_byte(0xad, 0xd4, 0xa2, 0xaf, 0x9c, 0xa4, 0x72, 0xc0);
|
| + &data_byte(0xb7, 0xfd, 0x93, 0x26, 0x36, 0x3f, 0xf7, 0xcc);
|
| + &data_byte(0x34, 0xa5, 0xe5, 0xf1, 0x71, 0xd8, 0x31, 0x15);
|
| + &data_byte(0x04, 0xc7, 0x23, 0xc3, 0x18, 0x96, 0x05, 0x9a);
|
| + &data_byte(0x07, 0x12, 0x80, 0xe2, 0xeb, 0x27, 0xb2, 0x75);
|
| + &data_byte(0x09, 0x83, 0x2c, 0x1a, 0x1b, 0x6e, 0x5a, 0xa0);
|
| + &data_byte(0x52, 0x3b, 0xd6, 0xb3, 0x29, 0xe3, 0x2f, 0x84);
|
| + &data_byte(0x53, 0xd1, 0x00, 0xed, 0x20, 0xfc, 0xb1, 0x5b);
|
| + &data_byte(0x6a, 0xcb, 0xbe, 0x39, 0x4a, 0x4c, 0x58, 0xcf);
|
| + &data_byte(0xd0, 0xef, 0xaa, 0xfb, 0x43, 0x4d, 0x33, 0x85);
|
| + &data_byte(0x45, 0xf9, 0x02, 0x7f, 0x50, 0x3c, 0x9f, 0xa8);
|
| + &data_byte(0x51, 0xa3, 0x40, 0x8f, 0x92, 0x9d, 0x38, 0xf5);
|
| + &data_byte(0xbc, 0xb6, 0xda, 0x21, 0x10, 0xff, 0xf3, 0xd2);
|
| + &data_byte(0xcd, 0x0c, 0x13, 0xec, 0x5f, 0x97, 0x44, 0x17);
|
| + &data_byte(0xc4, 0xa7, 0x7e, 0x3d, 0x64, 0x5d, 0x19, 0x73);
|
| + &data_byte(0x60, 0x81, 0x4f, 0xdc, 0x22, 0x2a, 0x90, 0x88);
|
| + &data_byte(0x46, 0xee, 0xb8, 0x14, 0xde, 0x5e, 0x0b, 0xdb);
|
| + &data_byte(0xe0, 0x32, 0x3a, 0x0a, 0x49, 0x06, 0x24, 0x5c);
|
| + &data_byte(0xc2, 0xd3, 0xac, 0x62, 0x91, 0x95, 0xe4, 0x79);
|
| + &data_byte(0xe7, 0xc8, 0x37, 0x6d, 0x8d, 0xd5, 0x4e, 0xa9);
|
| + &data_byte(0x6c, 0x56, 0xf4, 0xea, 0x65, 0x7a, 0xae, 0x08);
|
| + &data_byte(0xba, 0x78, 0x25, 0x2e, 0x1c, 0xa6, 0xb4, 0xc6);
|
| + &data_byte(0xe8, 0xdd, 0x74, 0x1f, 0x4b, 0xbd, 0x8b, 0x8a);
|
| + &data_byte(0x70, 0x3e, 0xb5, 0x66, 0x48, 0x03, 0xf6, 0x0e);
|
| + &data_byte(0x61, 0x35, 0x57, 0xb9, 0x86, 0xc1, 0x1d, 0x9e);
|
| + &data_byte(0xe1, 0xf8, 0x98, 0x11, 0x69, 0xd9, 0x8e, 0x94);
|
| + &data_byte(0x9b, 0x1e, 0x87, 0xe9, 0xce, 0x55, 0x28, 0xdf);
|
| + &data_byte(0x8c, 0xa1, 0x89, 0x0d, 0xbf, 0xe6, 0x42, 0x68);
|
| + &data_byte(0x41, 0x99, 0x2d, 0x0f, 0xb0, 0x54, 0xbb, 0x16);
|
| +
|
| + &data_byte(0x63, 0x7c, 0x77, 0x7b, 0xf2, 0x6b, 0x6f, 0xc5);
|
| + &data_byte(0x30, 0x01, 0x67, 0x2b, 0xfe, 0xd7, 0xab, 0x76);
|
| + &data_byte(0xca, 0x82, 0xc9, 0x7d, 0xfa, 0x59, 0x47, 0xf0);
|
| + &data_byte(0xad, 0xd4, 0xa2, 0xaf, 0x9c, 0xa4, 0x72, 0xc0);
|
| + &data_byte(0xb7, 0xfd, 0x93, 0x26, 0x36, 0x3f, 0xf7, 0xcc);
|
| + &data_byte(0x34, 0xa5, 0xe5, 0xf1, 0x71, 0xd8, 0x31, 0x15);
|
| + &data_byte(0x04, 0xc7, 0x23, 0xc3, 0x18, 0x96, 0x05, 0x9a);
|
| + &data_byte(0x07, 0x12, 0x80, 0xe2, 0xeb, 0x27, 0xb2, 0x75);
|
| + &data_byte(0x09, 0x83, 0x2c, 0x1a, 0x1b, 0x6e, 0x5a, 0xa0);
|
| + &data_byte(0x52, 0x3b, 0xd6, 0xb3, 0x29, 0xe3, 0x2f, 0x84);
|
| + &data_byte(0x53, 0xd1, 0x00, 0xed, 0x20, 0xfc, 0xb1, 0x5b);
|
| + &data_byte(0x6a, 0xcb, 0xbe, 0x39, 0x4a, 0x4c, 0x58, 0xcf);
|
| + &data_byte(0xd0, 0xef, 0xaa, 0xfb, 0x43, 0x4d, 0x33, 0x85);
|
| + &data_byte(0x45, 0xf9, 0x02, 0x7f, 0x50, 0x3c, 0x9f, 0xa8);
|
| + &data_byte(0x51, 0xa3, 0x40, 0x8f, 0x92, 0x9d, 0x38, 0xf5);
|
| + &data_byte(0xbc, 0xb6, 0xda, 0x21, 0x10, 0xff, 0xf3, 0xd2);
|
| + &data_byte(0xcd, 0x0c, 0x13, 0xec, 0x5f, 0x97, 0x44, 0x17);
|
| + &data_byte(0xc4, 0xa7, 0x7e, 0x3d, 0x64, 0x5d, 0x19, 0x73);
|
| + &data_byte(0x60, 0x81, 0x4f, 0xdc, 0x22, 0x2a, 0x90, 0x88);
|
| + &data_byte(0x46, 0xee, 0xb8, 0x14, 0xde, 0x5e, 0x0b, 0xdb);
|
| + &data_byte(0xe0, 0x32, 0x3a, 0x0a, 0x49, 0x06, 0x24, 0x5c);
|
| + &data_byte(0xc2, 0xd3, 0xac, 0x62, 0x91, 0x95, 0xe4, 0x79);
|
| + &data_byte(0xe7, 0xc8, 0x37, 0x6d, 0x8d, 0xd5, 0x4e, 0xa9);
|
| + &data_byte(0x6c, 0x56, 0xf4, 0xea, 0x65, 0x7a, 0xae, 0x08);
|
| + &data_byte(0xba, 0x78, 0x25, 0x2e, 0x1c, 0xa6, 0xb4, 0xc6);
|
| + &data_byte(0xe8, 0xdd, 0x74, 0x1f, 0x4b, 0xbd, 0x8b, 0x8a);
|
| + &data_byte(0x70, 0x3e, 0xb5, 0x66, 0x48, 0x03, 0xf6, 0x0e);
|
| + &data_byte(0x61, 0x35, 0x57, 0xb9, 0x86, 0xc1, 0x1d, 0x9e);
|
| + &data_byte(0xe1, 0xf8, 0x98, 0x11, 0x69, 0xd9, 0x8e, 0x94);
|
| + &data_byte(0x9b, 0x1e, 0x87, 0xe9, 0xce, 0x55, 0x28, 0xdf);
|
| + &data_byte(0x8c, 0xa1, 0x89, 0x0d, 0xbf, 0xe6, 0x42, 0x68);
|
| + &data_byte(0x41, 0x99, 0x2d, 0x0f, 0xb0, 0x54, 0xbb, 0x16);
|
| #rcon:
|
| $code.=<<___;
|
| .long 0x00000001, 0x00000002, 0x00000004, 0x00000008
|
| .long 0x00000010, 0x00000020, 0x00000040, 0x00000080
|
| - .long 0x0000001b, 0x00000036, 0, 0, 0, 0, 0, 0
|
| + .long 0x0000001b, 0x00000036, 0x80808080, 0x80808080
|
| + .long 0xfefefefe, 0xfefefefe, 0x1b1b1b1b, 0x1b1b1b1b
|
| ___
|
| $code.=<<___;
|
| -.globl AES_Td
|
| .align 64
|
| -AES_Td:
|
| +.LAES_Td:
|
| ___
|
| &_data_word(0x50a7f451, 0x5365417e, 0xc3a4171a, 0x965e273a);
|
| &_data_word(0xcb6bab3b, 0xf1459d1f, 0xab58faac, 0x9303e34b);
|
| @@ -1550,7 +2377,8 @@
|
| &_data_word(0x72c31d16, 0x0c25e2bc, 0x8b493c28, 0x41950dff);
|
| &_data_word(0x7101a839, 0xdeb30c08, 0x9ce4b4d8, 0x90c15664);
|
| &_data_word(0x6184cb7b, 0x70b632d5, 0x745c6c48, 0x4257b8d0);
|
| -#Td4:
|
| +
|
| +#Td4: # four copies of Td4 to choose from to avoid L1 aliasing
|
| &data_byte(0x52, 0x09, 0x6a, 0xd5, 0x30, 0x36, 0xa5, 0x38);
|
| &data_byte(0xbf, 0x40, 0xa3, 0x9e, 0x81, 0xf3, 0xd7, 0xfb);
|
| &data_byte(0x7c, 0xe3, 0x39, 0x82, 0x9b, 0x2f, 0xff, 0x87);
|
| @@ -1583,7 +2411,397 @@
|
| &data_byte(0xc8, 0xeb, 0xbb, 0x3c, 0x83, 0x53, 0x99, 0x61);
|
| &data_byte(0x17, 0x2b, 0x04, 0x7e, 0xba, 0x77, 0xd6, 0x26);
|
| &data_byte(0xe1, 0x69, 0x14, 0x63, 0x55, 0x21, 0x0c, 0x7d);
|
| +$code.=<<___;
|
| + .long 0x80808080, 0x80808080, 0xfefefefe, 0xfefefefe
|
| + .long 0x1b1b1b1b, 0x1b1b1b1b, 0, 0
|
| +___
|
| + &data_byte(0x52, 0x09, 0x6a, 0xd5, 0x30, 0x36, 0xa5, 0x38);
|
| + &data_byte(0xbf, 0x40, 0xa3, 0x9e, 0x81, 0xf3, 0xd7, 0xfb);
|
| + &data_byte(0x7c, 0xe3, 0x39, 0x82, 0x9b, 0x2f, 0xff, 0x87);
|
| + &data_byte(0x34, 0x8e, 0x43, 0x44, 0xc4, 0xde, 0xe9, 0xcb);
|
| + &data_byte(0x54, 0x7b, 0x94, 0x32, 0xa6, 0xc2, 0x23, 0x3d);
|
| + &data_byte(0xee, 0x4c, 0x95, 0x0b, 0x42, 0xfa, 0xc3, 0x4e);
|
| + &data_byte(0x08, 0x2e, 0xa1, 0x66, 0x28, 0xd9, 0x24, 0xb2);
|
| + &data_byte(0x76, 0x5b, 0xa2, 0x49, 0x6d, 0x8b, 0xd1, 0x25);
|
| + &data_byte(0x72, 0xf8, 0xf6, 0x64, 0x86, 0x68, 0x98, 0x16);
|
| + &data_byte(0xd4, 0xa4, 0x5c, 0xcc, 0x5d, 0x65, 0xb6, 0x92);
|
| + &data_byte(0x6c, 0x70, 0x48, 0x50, 0xfd, 0xed, 0xb9, 0xda);
|
| + &data_byte(0x5e, 0x15, 0x46, 0x57, 0xa7, 0x8d, 0x9d, 0x84);
|
| + &data_byte(0x90, 0xd8, 0xab, 0x00, 0x8c, 0xbc, 0xd3, 0x0a);
|
| + &data_byte(0xf7, 0xe4, 0x58, 0x05, 0xb8, 0xb3, 0x45, 0x06);
|
| + &data_byte(0xd0, 0x2c, 0x1e, 0x8f, 0xca, 0x3f, 0x0f, 0x02);
|
| + &data_byte(0xc1, 0xaf, 0xbd, 0x03, 0x01, 0x13, 0x8a, 0x6b);
|
| + &data_byte(0x3a, 0x91, 0x11, 0x41, 0x4f, 0x67, 0xdc, 0xea);
|
| + &data_byte(0x97, 0xf2, 0xcf, 0xce, 0xf0, 0xb4, 0xe6, 0x73);
|
| + &data_byte(0x96, 0xac, 0x74, 0x22, 0xe7, 0xad, 0x35, 0x85);
|
| + &data_byte(0xe2, 0xf9, 0x37, 0xe8, 0x1c, 0x75, 0xdf, 0x6e);
|
| + &data_byte(0x47, 0xf1, 0x1a, 0x71, 0x1d, 0x29, 0xc5, 0x89);
|
| + &data_byte(0x6f, 0xb7, 0x62, 0x0e, 0xaa, 0x18, 0xbe, 0x1b);
|
| + &data_byte(0xfc, 0x56, 0x3e, 0x4b, 0xc6, 0xd2, 0x79, 0x20);
|
| + &data_byte(0x9a, 0xdb, 0xc0, 0xfe, 0x78, 0xcd, 0x5a, 0xf4);
|
| + &data_byte(0x1f, 0xdd, 0xa8, 0x33, 0x88, 0x07, 0xc7, 0x31);
|
| + &data_byte(0xb1, 0x12, 0x10, 0x59, 0x27, 0x80, 0xec, 0x5f);
|
| + &data_byte(0x60, 0x51, 0x7f, 0xa9, 0x19, 0xb5, 0x4a, 0x0d);
|
| + &data_byte(0x2d, 0xe5, 0x7a, 0x9f, 0x93, 0xc9, 0x9c, 0xef);
|
| + &data_byte(0xa0, 0xe0, 0x3b, 0x4d, 0xae, 0x2a, 0xf5, 0xb0);
|
| + &data_byte(0xc8, 0xeb, 0xbb, 0x3c, 0x83, 0x53, 0x99, 0x61);
|
| + &data_byte(0x17, 0x2b, 0x04, 0x7e, 0xba, 0x77, 0xd6, 0x26);
|
| + &data_byte(0xe1, 0x69, 0x14, 0x63, 0x55, 0x21, 0x0c, 0x7d);
|
| +$code.=<<___;
|
| + .long 0x80808080, 0x80808080, 0xfefefefe, 0xfefefefe
|
| + .long 0x1b1b1b1b, 0x1b1b1b1b, 0, 0
|
| +___
|
| + &data_byte(0x52, 0x09, 0x6a, 0xd5, 0x30, 0x36, 0xa5, 0x38);
|
| + &data_byte(0xbf, 0x40, 0xa3, 0x9e, 0x81, 0xf3, 0xd7, 0xfb);
|
| + &data_byte(0x7c, 0xe3, 0x39, 0x82, 0x9b, 0x2f, 0xff, 0x87);
|
| + &data_byte(0x34, 0x8e, 0x43, 0x44, 0xc4, 0xde, 0xe9, 0xcb);
|
| + &data_byte(0x54, 0x7b, 0x94, 0x32, 0xa6, 0xc2, 0x23, 0x3d);
|
| + &data_byte(0xee, 0x4c, 0x95, 0x0b, 0x42, 0xfa, 0xc3, 0x4e);
|
| + &data_byte(0x08, 0x2e, 0xa1, 0x66, 0x28, 0xd9, 0x24, 0xb2);
|
| + &data_byte(0x76, 0x5b, 0xa2, 0x49, 0x6d, 0x8b, 0xd1, 0x25);
|
| + &data_byte(0x72, 0xf8, 0xf6, 0x64, 0x86, 0x68, 0x98, 0x16);
|
| + &data_byte(0xd4, 0xa4, 0x5c, 0xcc, 0x5d, 0x65, 0xb6, 0x92);
|
| + &data_byte(0x6c, 0x70, 0x48, 0x50, 0xfd, 0xed, 0xb9, 0xda);
|
| + &data_byte(0x5e, 0x15, 0x46, 0x57, 0xa7, 0x8d, 0x9d, 0x84);
|
| + &data_byte(0x90, 0xd8, 0xab, 0x00, 0x8c, 0xbc, 0xd3, 0x0a);
|
| + &data_byte(0xf7, 0xe4, 0x58, 0x05, 0xb8, 0xb3, 0x45, 0x06);
|
| + &data_byte(0xd0, 0x2c, 0x1e, 0x8f, 0xca, 0x3f, 0x0f, 0x02);
|
| + &data_byte(0xc1, 0xaf, 0xbd, 0x03, 0x01, 0x13, 0x8a, 0x6b);
|
| + &data_byte(0x3a, 0x91, 0x11, 0x41, 0x4f, 0x67, 0xdc, 0xea);
|
| + &data_byte(0x97, 0xf2, 0xcf, 0xce, 0xf0, 0xb4, 0xe6, 0x73);
|
| + &data_byte(0x96, 0xac, 0x74, 0x22, 0xe7, 0xad, 0x35, 0x85);
|
| + &data_byte(0xe2, 0xf9, 0x37, 0xe8, 0x1c, 0x75, 0xdf, 0x6e);
|
| + &data_byte(0x47, 0xf1, 0x1a, 0x71, 0x1d, 0x29, 0xc5, 0x89);
|
| + &data_byte(0x6f, 0xb7, 0x62, 0x0e, 0xaa, 0x18, 0xbe, 0x1b);
|
| + &data_byte(0xfc, 0x56, 0x3e, 0x4b, 0xc6, 0xd2, 0x79, 0x20);
|
| + &data_byte(0x9a, 0xdb, 0xc0, 0xfe, 0x78, 0xcd, 0x5a, 0xf4);
|
| + &data_byte(0x1f, 0xdd, 0xa8, 0x33, 0x88, 0x07, 0xc7, 0x31);
|
| + &data_byte(0xb1, 0x12, 0x10, 0x59, 0x27, 0x80, 0xec, 0x5f);
|
| + &data_byte(0x60, 0x51, 0x7f, 0xa9, 0x19, 0xb5, 0x4a, 0x0d);
|
| + &data_byte(0x2d, 0xe5, 0x7a, 0x9f, 0x93, 0xc9, 0x9c, 0xef);
|
| + &data_byte(0xa0, 0xe0, 0x3b, 0x4d, 0xae, 0x2a, 0xf5, 0xb0);
|
| + &data_byte(0xc8, 0xeb, 0xbb, 0x3c, 0x83, 0x53, 0x99, 0x61);
|
| + &data_byte(0x17, 0x2b, 0x04, 0x7e, 0xba, 0x77, 0xd6, 0x26);
|
| + &data_byte(0xe1, 0x69, 0x14, 0x63, 0x55, 0x21, 0x0c, 0x7d);
|
| +$code.=<<___;
|
| + .long 0x80808080, 0x80808080, 0xfefefefe, 0xfefefefe
|
| + .long 0x1b1b1b1b, 0x1b1b1b1b, 0, 0
|
| +___
|
| + &data_byte(0x52, 0x09, 0x6a, 0xd5, 0x30, 0x36, 0xa5, 0x38);
|
| + &data_byte(0xbf, 0x40, 0xa3, 0x9e, 0x81, 0xf3, 0xd7, 0xfb);
|
| + &data_byte(0x7c, 0xe3, 0x39, 0x82, 0x9b, 0x2f, 0xff, 0x87);
|
| + &data_byte(0x34, 0x8e, 0x43, 0x44, 0xc4, 0xde, 0xe9, 0xcb);
|
| + &data_byte(0x54, 0x7b, 0x94, 0x32, 0xa6, 0xc2, 0x23, 0x3d);
|
| + &data_byte(0xee, 0x4c, 0x95, 0x0b, 0x42, 0xfa, 0xc3, 0x4e);
|
| + &data_byte(0x08, 0x2e, 0xa1, 0x66, 0x28, 0xd9, 0x24, 0xb2);
|
| + &data_byte(0x76, 0x5b, 0xa2, 0x49, 0x6d, 0x8b, 0xd1, 0x25);
|
| + &data_byte(0x72, 0xf8, 0xf6, 0x64, 0x86, 0x68, 0x98, 0x16);
|
| + &data_byte(0xd4, 0xa4, 0x5c, 0xcc, 0x5d, 0x65, 0xb6, 0x92);
|
| + &data_byte(0x6c, 0x70, 0x48, 0x50, 0xfd, 0xed, 0xb9, 0xda);
|
| + &data_byte(0x5e, 0x15, 0x46, 0x57, 0xa7, 0x8d, 0x9d, 0x84);
|
| + &data_byte(0x90, 0xd8, 0xab, 0x00, 0x8c, 0xbc, 0xd3, 0x0a);
|
| + &data_byte(0xf7, 0xe4, 0x58, 0x05, 0xb8, 0xb3, 0x45, 0x06);
|
| + &data_byte(0xd0, 0x2c, 0x1e, 0x8f, 0xca, 0x3f, 0x0f, 0x02);
|
| + &data_byte(0xc1, 0xaf, 0xbd, 0x03, 0x01, 0x13, 0x8a, 0x6b);
|
| + &data_byte(0x3a, 0x91, 0x11, 0x41, 0x4f, 0x67, 0xdc, 0xea);
|
| + &data_byte(0x97, 0xf2, 0xcf, 0xce, 0xf0, 0xb4, 0xe6, 0x73);
|
| + &data_byte(0x96, 0xac, 0x74, 0x22, 0xe7, 0xad, 0x35, 0x85);
|
| + &data_byte(0xe2, 0xf9, 0x37, 0xe8, 0x1c, 0x75, 0xdf, 0x6e);
|
| + &data_byte(0x47, 0xf1, 0x1a, 0x71, 0x1d, 0x29, 0xc5, 0x89);
|
| + &data_byte(0x6f, 0xb7, 0x62, 0x0e, 0xaa, 0x18, 0xbe, 0x1b);
|
| + &data_byte(0xfc, 0x56, 0x3e, 0x4b, 0xc6, 0xd2, 0x79, 0x20);
|
| + &data_byte(0x9a, 0xdb, 0xc0, 0xfe, 0x78, 0xcd, 0x5a, 0xf4);
|
| + &data_byte(0x1f, 0xdd, 0xa8, 0x33, 0x88, 0x07, 0xc7, 0x31);
|
| + &data_byte(0xb1, 0x12, 0x10, 0x59, 0x27, 0x80, 0xec, 0x5f);
|
| + &data_byte(0x60, 0x51, 0x7f, 0xa9, 0x19, 0xb5, 0x4a, 0x0d);
|
| + &data_byte(0x2d, 0xe5, 0x7a, 0x9f, 0x93, 0xc9, 0x9c, 0xef);
|
| + &data_byte(0xa0, 0xe0, 0x3b, 0x4d, 0xae, 0x2a, 0xf5, 0xb0);
|
| + &data_byte(0xc8, 0xeb, 0xbb, 0x3c, 0x83, 0x53, 0x99, 0x61);
|
| + &data_byte(0x17, 0x2b, 0x04, 0x7e, 0xba, 0x77, 0xd6, 0x26);
|
| + &data_byte(0xe1, 0x69, 0x14, 0x63, 0x55, 0x21, 0x0c, 0x7d);
|
| +$code.=<<___;
|
| + .long 0x80808080, 0x80808080, 0xfefefefe, 0xfefefefe
|
| + .long 0x1b1b1b1b, 0x1b1b1b1b, 0, 0
|
| +.asciz "AES for x86_64, CRYPTOGAMS by <appro\@openssl.org>"
|
| +.align 64
|
| +___
|
|
|
| +# EXCEPTION_DISPOSITION handler (EXCEPTION_RECORD *rec,ULONG64 frame,
|
| +# CONTEXT *context,DISPATCHER_CONTEXT *disp)
|
| +if ($win64) {
|
| +$rec="%rcx";
|
| +$frame="%rdx";
|
| +$context="%r8";
|
| +$disp="%r9";
|
| +
|
| +$code.=<<___;
|
| +.extern __imp_RtlVirtualUnwind
|
| +.type block_se_handler,\@abi-omnipotent
|
| +.align 16
|
| +block_se_handler:
|
| + push %rsi
|
| + push %rdi
|
| + push %rbx
|
| + push %rbp
|
| + push %r12
|
| + push %r13
|
| + push %r14
|
| + push %r15
|
| + pushfq
|
| + sub \$64,%rsp
|
| +
|
| + mov 120($context),%rax # pull context->Rax
|
| + mov 248($context),%rbx # pull context->Rip
|
| +
|
| + mov 8($disp),%rsi # disp->ImageBase
|
| + mov 56($disp),%r11 # disp->HandlerData
|
| +
|
| + mov 0(%r11),%r10d # HandlerData[0]
|
| + lea (%rsi,%r10),%r10 # prologue label
|
| + cmp %r10,%rbx # context->Rip<prologue label
|
| + jb .Lin_block_prologue
|
| +
|
| + mov 152($context),%rax # pull context->Rsp
|
| +
|
| + mov 4(%r11),%r10d # HandlerData[1]
|
| + lea (%rsi,%r10),%r10 # epilogue label
|
| + cmp %r10,%rbx # context->Rip>=epilogue label
|
| + jae .Lin_block_prologue
|
| +
|
| + mov 24(%rax),%rax # pull saved real stack pointer
|
| + lea 48(%rax),%rax # adjust...
|
| +
|
| + mov -8(%rax),%rbx
|
| + mov -16(%rax),%rbp
|
| + mov -24(%rax),%r12
|
| + mov -32(%rax),%r13
|
| + mov -40(%rax),%r14
|
| + mov -48(%rax),%r15
|
| + mov %rbx,144($context) # restore context->Rbx
|
| + mov %rbp,160($context) # restore context->Rbp
|
| + mov %r12,216($context) # restore context->R12
|
| + mov %r13,224($context) # restore context->R13
|
| + mov %r14,232($context) # restore context->R14
|
| + mov %r15,240($context) # restore context->R15
|
| +
|
| +.Lin_block_prologue:
|
| + mov 8(%rax),%rdi
|
| + mov 16(%rax),%rsi
|
| + mov %rax,152($context) # restore context->Rsp
|
| + mov %rsi,168($context) # restore context->Rsi
|
| + mov %rdi,176($context) # restore context->Rdi
|
| +
|
| + jmp .Lcommon_seh_exit
|
| +.size block_se_handler,.-block_se_handler
|
| +
|
| +.type key_se_handler,\@abi-omnipotent
|
| +.align 16
|
| +key_se_handler:
|
| + push %rsi
|
| + push %rdi
|
| + push %rbx
|
| + push %rbp
|
| + push %r12
|
| + push %r13
|
| + push %r14
|
| + push %r15
|
| + pushfq
|
| + sub \$64,%rsp
|
| +
|
| + mov 120($context),%rax # pull context->Rax
|
| + mov 248($context),%rbx # pull context->Rip
|
| +
|
| + mov 8($disp),%rsi # disp->ImageBase
|
| + mov 56($disp),%r11 # disp->HandlerData
|
| +
|
| + mov 0(%r11),%r10d # HandlerData[0]
|
| + lea (%rsi,%r10),%r10 # prologue label
|
| + cmp %r10,%rbx # context->Rip<prologue label
|
| + jb .Lin_key_prologue
|
| +
|
| + mov 152($context),%rax # pull context->Rsp
|
| +
|
| + mov 4(%r11),%r10d # HandlerData[1]
|
| + lea (%rsi,%r10),%r10 # epilogue label
|
| + cmp %r10,%rbx # context->Rip>=epilogue label
|
| + jae .Lin_key_prologue
|
| +
|
| + lea 56(%rax),%rax
|
| +
|
| + mov -8(%rax),%rbx
|
| + mov -16(%rax),%rbp
|
| + mov -24(%rax),%r12
|
| + mov -32(%rax),%r13
|
| + mov -40(%rax),%r14
|
| + mov -48(%rax),%r15
|
| + mov %rbx,144($context) # restore context->Rbx
|
| + mov %rbp,160($context) # restore context->Rbp
|
| + mov %r12,216($context) # restore context->R12
|
| + mov %r13,224($context) # restore context->R13
|
| + mov %r14,232($context) # restore context->R14
|
| + mov %r15,240($context) # restore context->R15
|
| +
|
| +.Lin_key_prologue:
|
| + mov 8(%rax),%rdi
|
| + mov 16(%rax),%rsi
|
| + mov %rax,152($context) # restore context->Rsp
|
| + mov %rsi,168($context) # restore context->Rsi
|
| + mov %rdi,176($context) # restore context->Rdi
|
| +
|
| + jmp .Lcommon_seh_exit
|
| +.size key_se_handler,.-key_se_handler
|
| +
|
| +.type cbc_se_handler,\@abi-omnipotent
|
| +.align 16
|
| +cbc_se_handler:
|
| + push %rsi
|
| + push %rdi
|
| + push %rbx
|
| + push %rbp
|
| + push %r12
|
| + push %r13
|
| + push %r14
|
| + push %r15
|
| + pushfq
|
| + sub \$64,%rsp
|
| +
|
| + mov 120($context),%rax # pull context->Rax
|
| + mov 248($context),%rbx # pull context->Rip
|
| +
|
| + lea .Lcbc_prologue(%rip),%r10
|
| + cmp %r10,%rbx # context->Rip<.Lcbc_prologue
|
| + jb .Lin_cbc_prologue
|
| +
|
| + lea .Lcbc_fast_body(%rip),%r10
|
| + cmp %r10,%rbx # context->Rip<.Lcbc_fast_body
|
| + jb .Lin_cbc_frame_setup
|
| +
|
| + lea .Lcbc_slow_prologue(%rip),%r10
|
| + cmp %r10,%rbx # context->Rip<.Lcbc_slow_prologue
|
| + jb .Lin_cbc_body
|
| +
|
| + lea .Lcbc_slow_body(%rip),%r10
|
| + cmp %r10,%rbx # context->Rip<.Lcbc_slow_body
|
| + jb .Lin_cbc_frame_setup
|
| +
|
| +.Lin_cbc_body:
|
| + mov 152($context),%rax # pull context->Rsp
|
| +
|
| + lea .Lcbc_epilogue(%rip),%r10
|
| + cmp %r10,%rbx # context->Rip>=.Lcbc_epilogue
|
| + jae .Lin_cbc_prologue
|
| +
|
| + lea 8(%rax),%rax
|
| +
|
| + lea .Lcbc_popfq(%rip),%r10
|
| + cmp %r10,%rbx # context->Rip>=.Lcbc_popfq
|
| + jae .Lin_cbc_prologue
|
| +
|
| + mov `16-8`(%rax),%rax # biased $_rsp
|
| + lea 56(%rax),%rax
|
| +
|
| +.Lin_cbc_frame_setup:
|
| + mov -16(%rax),%rbx
|
| + mov -24(%rax),%rbp
|
| + mov -32(%rax),%r12
|
| + mov -40(%rax),%r13
|
| + mov -48(%rax),%r14
|
| + mov -56(%rax),%r15
|
| + mov %rbx,144($context) # restore context->Rbx
|
| + mov %rbp,160($context) # restore context->Rbp
|
| + mov %r12,216($context) # restore context->R12
|
| + mov %r13,224($context) # restore context->R13
|
| + mov %r14,232($context) # restore context->R14
|
| + mov %r15,240($context) # restore context->R15
|
| +
|
| +.Lin_cbc_prologue:
|
| + mov 8(%rax),%rdi
|
| + mov 16(%rax),%rsi
|
| + mov %rax,152($context) # restore context->Rsp
|
| + mov %rsi,168($context) # restore context->Rsi
|
| + mov %rdi,176($context) # restore context->Rdi
|
| +
|
| +.Lcommon_seh_exit:
|
| +
|
| + mov 40($disp),%rdi # disp->ContextRecord
|
| + mov $context,%rsi # context
|
| + mov \$`1232/8`,%ecx # sizeof(CONTEXT)
|
| + .long 0xa548f3fc # cld; rep movsq
|
| +
|
| + mov $disp,%rsi
|
| + xor %rcx,%rcx # arg1, UNW_FLAG_NHANDLER
|
| + mov 8(%rsi),%rdx # arg2, disp->ImageBase
|
| + mov 0(%rsi),%r8 # arg3, disp->ControlPc
|
| + mov 16(%rsi),%r9 # arg4, disp->FunctionEntry
|
| + mov 40(%rsi),%r10 # disp->ContextRecord
|
| + lea 56(%rsi),%r11 # &disp->HandlerData
|
| + lea 24(%rsi),%r12 # &disp->EstablisherFrame
|
| + mov %r10,32(%rsp) # arg5
|
| + mov %r11,40(%rsp) # arg6
|
| + mov %r12,48(%rsp) # arg7
|
| + mov %rcx,56(%rsp) # arg8, (NULL)
|
| + call *__imp_RtlVirtualUnwind(%rip)
|
| +
|
| + mov \$1,%eax # ExceptionContinueSearch
|
| + add \$64,%rsp
|
| + popfq
|
| + pop %r15
|
| + pop %r14
|
| + pop %r13
|
| + pop %r12
|
| + pop %rbp
|
| + pop %rbx
|
| + pop %rdi
|
| + pop %rsi
|
| + ret
|
| +.size cbc_se_handler,.-cbc_se_handler
|
| +
|
| +.section .pdata
|
| +.align 4
|
| + .rva .LSEH_begin_AES_encrypt
|
| + .rva .LSEH_end_AES_encrypt
|
| + .rva .LSEH_info_AES_encrypt
|
| +
|
| + .rva .LSEH_begin_AES_decrypt
|
| + .rva .LSEH_end_AES_decrypt
|
| + .rva .LSEH_info_AES_decrypt
|
| +
|
| + .rva .LSEH_begin_AES_set_encrypt_key
|
| + .rva .LSEH_end_AES_set_encrypt_key
|
| + .rva .LSEH_info_AES_set_encrypt_key
|
| +
|
| + .rva .LSEH_begin_AES_set_decrypt_key
|
| + .rva .LSEH_end_AES_set_decrypt_key
|
| + .rva .LSEH_info_AES_set_decrypt_key
|
| +
|
| + .rva .LSEH_begin_AES_cbc_encrypt
|
| + .rva .LSEH_end_AES_cbc_encrypt
|
| + .rva .LSEH_info_AES_cbc_encrypt
|
| +
|
| +.section .xdata
|
| +.align 8
|
| +.LSEH_info_AES_encrypt:
|
| + .byte 9,0,0,0
|
| + .rva block_se_handler
|
| + .rva .Lenc_prologue,.Lenc_epilogue # HandlerData[]
|
| +.LSEH_info_AES_decrypt:
|
| + .byte 9,0,0,0
|
| + .rva block_se_handler
|
| + .rva .Ldec_prologue,.Ldec_epilogue # HandlerData[]
|
| +.LSEH_info_AES_set_encrypt_key:
|
| + .byte 9,0,0,0
|
| + .rva key_se_handler
|
| + .rva .Lenc_key_prologue,.Lenc_key_epilogue # HandlerData[]
|
| +.LSEH_info_AES_set_decrypt_key:
|
| + .byte 9,0,0,0
|
| + .rva key_se_handler
|
| + .rva .Ldec_key_prologue,.Ldec_key_epilogue # HandlerData[]
|
| +.LSEH_info_AES_cbc_encrypt:
|
| + .byte 9,0,0,0
|
| + .rva cbc_se_handler
|
| +___
|
| +}
|
| +
|
| $code =~ s/\`([^\`]*)\`/eval($1)/gem;
|
|
|
| print $code;
|
|
|