| Index: openssl/crypto/modes/asm/ghash-sparcv9.pl
|
| diff --git a/openssl/crypto/modes/asm/ghash-sparcv9.pl b/openssl/crypto/modes/asm/ghash-sparcv9.pl
|
| deleted file mode 100644
|
| index 70e7b044a3ec1f6afb3234108b210514ec1a87ce..0000000000000000000000000000000000000000
|
| --- a/openssl/crypto/modes/asm/ghash-sparcv9.pl
|
| +++ /dev/null
|
| @@ -1,330 +0,0 @@
|
| -#!/usr/bin/env perl
|
| -
|
| -# ====================================================================
|
| -# Written by Andy Polyakov <appro@openssl.org> for the OpenSSL
|
| -# project. The module is, however, dual licensed under OpenSSL and
|
| -# CRYPTOGAMS licenses depending on where you obtain it. For further
|
| -# details see http://www.openssl.org/~appro/cryptogams/.
|
| -# ====================================================================
|
| -
|
| -# March 2010
|
| -#
|
| -# The module implements "4-bit" GCM GHASH function and underlying
|
| -# single multiplication operation in GF(2^128). "4-bit" means that it
|
| -# uses 256 bytes per-key table [+128 bytes shared table]. Performance
|
| -# results are for streamed GHASH subroutine on UltraSPARC pre-Tx CPU
|
| -# and are expressed in cycles per processed byte, less is better:
|
| -#
|
| -# gcc 3.3.x cc 5.2 this assembler
|
| -#
|
| -# 32-bit build 81.4 43.3 12.6 (+546%/+244%)
|
| -# 64-bit build 20.2 21.2 12.6 (+60%/+68%)
|
| -#
|
| -# Here is data collected on UltraSPARC T1 system running Linux:
|
| -#
|
| -# gcc 4.4.1 this assembler
|
| -#
|
| -# 32-bit build 566 50 (+1000%)
|
| -# 64-bit build 56 50 (+12%)
|
| -#
|
| -# I don't quite understand why difference between 32-bit and 64-bit
|
| -# compiler-generated code is so big. Compilers *were* instructed to
|
| -# generate code for UltraSPARC and should have used 64-bit registers
|
| -# for Z vector (see C code) even in 32-bit build... Oh well, it only
|
| -# means more impressive improvement coefficients for this assembler
|
| -# module;-) Loops are aggressively modulo-scheduled in respect to
|
| -# references to input data and Z.hi updates to achieve 12 cycles
|
| -# timing. To anchor to something else, sha1-sparcv9.pl spends 11.6
|
| -# cycles to process one byte on UltraSPARC pre-Tx CPU and ~24 on T1.
|
| -
|
| -$bits=32;
|
| -for (@ARGV) { $bits=64 if (/\-m64/ || /\-xarch\=v9/); }
|
| -if ($bits==64) { $bias=2047; $frame=192; }
|
| -else { $bias=0; $frame=112; }
|
| -
|
| -$output=shift;
|
| -open STDOUT,">$output";
|
| -
|
| -$Zhi="%o0"; # 64-bit values
|
| -$Zlo="%o1";
|
| -$Thi="%o2";
|
| -$Tlo="%o3";
|
| -$rem="%o4";
|
| -$tmp="%o5";
|
| -
|
| -$nhi="%l0"; # small values and pointers
|
| -$nlo="%l1";
|
| -$xi0="%l2";
|
| -$xi1="%l3";
|
| -$rem_4bit="%l4";
|
| -$remi="%l5";
|
| -$Htblo="%l6";
|
| -$cnt="%l7";
|
| -
|
| -$Xi="%i0"; # input argument block
|
| -$Htbl="%i1";
|
| -$inp="%i2";
|
| -$len="%i3";
|
| -
|
| -$code.=<<___;
|
| -.section ".text",#alloc,#execinstr
|
| -
|
| -.align 64
|
| -rem_4bit:
|
| - .long `0x0000<<16`,0,`0x1C20<<16`,0,`0x3840<<16`,0,`0x2460<<16`,0
|
| - .long `0x7080<<16`,0,`0x6CA0<<16`,0,`0x48C0<<16`,0,`0x54E0<<16`,0
|
| - .long `0xE100<<16`,0,`0xFD20<<16`,0,`0xD940<<16`,0,`0xC560<<16`,0
|
| - .long `0x9180<<16`,0,`0x8DA0<<16`,0,`0xA9C0<<16`,0,`0xB5E0<<16`,0
|
| -.type rem_4bit,#object
|
| -.size rem_4bit,(.-rem_4bit)
|
| -
|
| -.globl gcm_ghash_4bit
|
| -.align 32
|
| -gcm_ghash_4bit:
|
| - save %sp,-$frame,%sp
|
| - ldub [$inp+15],$nlo
|
| - ldub [$Xi+15],$xi0
|
| - ldub [$Xi+14],$xi1
|
| - add $len,$inp,$len
|
| - add $Htbl,8,$Htblo
|
| -
|
| -1: call .+8
|
| - add %o7,rem_4bit-1b,$rem_4bit
|
| -
|
| -.Louter:
|
| - xor $xi0,$nlo,$nlo
|
| - and $nlo,0xf0,$nhi
|
| - and $nlo,0x0f,$nlo
|
| - sll $nlo,4,$nlo
|
| - ldx [$Htblo+$nlo],$Zlo
|
| - ldx [$Htbl+$nlo],$Zhi
|
| -
|
| - ldub [$inp+14],$nlo
|
| -
|
| - ldx [$Htblo+$nhi],$Tlo
|
| - and $Zlo,0xf,$remi
|
| - ldx [$Htbl+$nhi],$Thi
|
| - sll $remi,3,$remi
|
| - ldx [$rem_4bit+$remi],$rem
|
| - srlx $Zlo,4,$Zlo
|
| - mov 13,$cnt
|
| - sllx $Zhi,60,$tmp
|
| - xor $Tlo,$Zlo,$Zlo
|
| - srlx $Zhi,4,$Zhi
|
| - xor $Zlo,$tmp,$Zlo
|
| -
|
| - xor $xi1,$nlo,$nlo
|
| - and $Zlo,0xf,$remi
|
| - and $nlo,0xf0,$nhi
|
| - and $nlo,0x0f,$nlo
|
| - ba .Lghash_inner
|
| - sll $nlo,4,$nlo
|
| -.align 32
|
| -.Lghash_inner:
|
| - ldx [$Htblo+$nlo],$Tlo
|
| - sll $remi,3,$remi
|
| - xor $Thi,$Zhi,$Zhi
|
| - ldx [$Htbl+$nlo],$Thi
|
| - srlx $Zlo,4,$Zlo
|
| - xor $rem,$Zhi,$Zhi
|
| - ldx [$rem_4bit+$remi],$rem
|
| - sllx $Zhi,60,$tmp
|
| - xor $Tlo,$Zlo,$Zlo
|
| - ldub [$inp+$cnt],$nlo
|
| - srlx $Zhi,4,$Zhi
|
| - xor $Zlo,$tmp,$Zlo
|
| - ldub [$Xi+$cnt],$xi1
|
| - xor $Thi,$Zhi,$Zhi
|
| - and $Zlo,0xf,$remi
|
| -
|
| - ldx [$Htblo+$nhi],$Tlo
|
| - sll $remi,3,$remi
|
| - xor $rem,$Zhi,$Zhi
|
| - ldx [$Htbl+$nhi],$Thi
|
| - srlx $Zlo,4,$Zlo
|
| - ldx [$rem_4bit+$remi],$rem
|
| - sllx $Zhi,60,$tmp
|
| - xor $xi1,$nlo,$nlo
|
| - srlx $Zhi,4,$Zhi
|
| - and $nlo,0xf0,$nhi
|
| - addcc $cnt,-1,$cnt
|
| - xor $Zlo,$tmp,$Zlo
|
| - and $nlo,0x0f,$nlo
|
| - xor $Tlo,$Zlo,$Zlo
|
| - sll $nlo,4,$nlo
|
| - blu .Lghash_inner
|
| - and $Zlo,0xf,$remi
|
| -
|
| - ldx [$Htblo+$nlo],$Tlo
|
| - sll $remi,3,$remi
|
| - xor $Thi,$Zhi,$Zhi
|
| - ldx [$Htbl+$nlo],$Thi
|
| - srlx $Zlo,4,$Zlo
|
| - xor $rem,$Zhi,$Zhi
|
| - ldx [$rem_4bit+$remi],$rem
|
| - sllx $Zhi,60,$tmp
|
| - xor $Tlo,$Zlo,$Zlo
|
| - srlx $Zhi,4,$Zhi
|
| - xor $Zlo,$tmp,$Zlo
|
| - xor $Thi,$Zhi,$Zhi
|
| -
|
| - add $inp,16,$inp
|
| - cmp $inp,$len
|
| - be,pn `$bits==64?"%xcc":"%icc"`,.Ldone
|
| - and $Zlo,0xf,$remi
|
| -
|
| - ldx [$Htblo+$nhi],$Tlo
|
| - sll $remi,3,$remi
|
| - xor $rem,$Zhi,$Zhi
|
| - ldx [$Htbl+$nhi],$Thi
|
| - srlx $Zlo,4,$Zlo
|
| - ldx [$rem_4bit+$remi],$rem
|
| - sllx $Zhi,60,$tmp
|
| - xor $Tlo,$Zlo,$Zlo
|
| - ldub [$inp+15],$nlo
|
| - srlx $Zhi,4,$Zhi
|
| - xor $Zlo,$tmp,$Zlo
|
| - xor $Thi,$Zhi,$Zhi
|
| - stx $Zlo,[$Xi+8]
|
| - xor $rem,$Zhi,$Zhi
|
| - stx $Zhi,[$Xi]
|
| - srl $Zlo,8,$xi1
|
| - and $Zlo,0xff,$xi0
|
| - ba .Louter
|
| - and $xi1,0xff,$xi1
|
| -.align 32
|
| -.Ldone:
|
| - ldx [$Htblo+$nhi],$Tlo
|
| - sll $remi,3,$remi
|
| - xor $rem,$Zhi,$Zhi
|
| - ldx [$Htbl+$nhi],$Thi
|
| - srlx $Zlo,4,$Zlo
|
| - ldx [$rem_4bit+$remi],$rem
|
| - sllx $Zhi,60,$tmp
|
| - xor $Tlo,$Zlo,$Zlo
|
| - srlx $Zhi,4,$Zhi
|
| - xor $Zlo,$tmp,$Zlo
|
| - xor $Thi,$Zhi,$Zhi
|
| - stx $Zlo,[$Xi+8]
|
| - xor $rem,$Zhi,$Zhi
|
| - stx $Zhi,[$Xi]
|
| -
|
| - ret
|
| - restore
|
| -.type gcm_ghash_4bit,#function
|
| -.size gcm_ghash_4bit,(.-gcm_ghash_4bit)
|
| -___
|
| -
|
| -undef $inp;
|
| -undef $len;
|
| -
|
| -$code.=<<___;
|
| -.globl gcm_gmult_4bit
|
| -.align 32
|
| -gcm_gmult_4bit:
|
| - save %sp,-$frame,%sp
|
| - ldub [$Xi+15],$nlo
|
| - add $Htbl,8,$Htblo
|
| -
|
| -1: call .+8
|
| - add %o7,rem_4bit-1b,$rem_4bit
|
| -
|
| - and $nlo,0xf0,$nhi
|
| - and $nlo,0x0f,$nlo
|
| - sll $nlo,4,$nlo
|
| - ldx [$Htblo+$nlo],$Zlo
|
| - ldx [$Htbl+$nlo],$Zhi
|
| -
|
| - ldub [$Xi+14],$nlo
|
| -
|
| - ldx [$Htblo+$nhi],$Tlo
|
| - and $Zlo,0xf,$remi
|
| - ldx [$Htbl+$nhi],$Thi
|
| - sll $remi,3,$remi
|
| - ldx [$rem_4bit+$remi],$rem
|
| - srlx $Zlo,4,$Zlo
|
| - mov 13,$cnt
|
| - sllx $Zhi,60,$tmp
|
| - xor $Tlo,$Zlo,$Zlo
|
| - srlx $Zhi,4,$Zhi
|
| - xor $Zlo,$tmp,$Zlo
|
| -
|
| - and $Zlo,0xf,$remi
|
| - and $nlo,0xf0,$nhi
|
| - and $nlo,0x0f,$nlo
|
| - ba .Lgmult_inner
|
| - sll $nlo,4,$nlo
|
| -.align 32
|
| -.Lgmult_inner:
|
| - ldx [$Htblo+$nlo],$Tlo
|
| - sll $remi,3,$remi
|
| - xor $Thi,$Zhi,$Zhi
|
| - ldx [$Htbl+$nlo],$Thi
|
| - srlx $Zlo,4,$Zlo
|
| - xor $rem,$Zhi,$Zhi
|
| - ldx [$rem_4bit+$remi],$rem
|
| - sllx $Zhi,60,$tmp
|
| - xor $Tlo,$Zlo,$Zlo
|
| - ldub [$Xi+$cnt],$nlo
|
| - srlx $Zhi,4,$Zhi
|
| - xor $Zlo,$tmp,$Zlo
|
| - xor $Thi,$Zhi,$Zhi
|
| - and $Zlo,0xf,$remi
|
| -
|
| - ldx [$Htblo+$nhi],$Tlo
|
| - sll $remi,3,$remi
|
| - xor $rem,$Zhi,$Zhi
|
| - ldx [$Htbl+$nhi],$Thi
|
| - srlx $Zlo,4,$Zlo
|
| - ldx [$rem_4bit+$remi],$rem
|
| - sllx $Zhi,60,$tmp
|
| - srlx $Zhi,4,$Zhi
|
| - and $nlo,0xf0,$nhi
|
| - addcc $cnt,-1,$cnt
|
| - xor $Zlo,$tmp,$Zlo
|
| - and $nlo,0x0f,$nlo
|
| - xor $Tlo,$Zlo,$Zlo
|
| - sll $nlo,4,$nlo
|
| - blu .Lgmult_inner
|
| - and $Zlo,0xf,$remi
|
| -
|
| - ldx [$Htblo+$nlo],$Tlo
|
| - sll $remi,3,$remi
|
| - xor $Thi,$Zhi,$Zhi
|
| - ldx [$Htbl+$nlo],$Thi
|
| - srlx $Zlo,4,$Zlo
|
| - xor $rem,$Zhi,$Zhi
|
| - ldx [$rem_4bit+$remi],$rem
|
| - sllx $Zhi,60,$tmp
|
| - xor $Tlo,$Zlo,$Zlo
|
| - srlx $Zhi,4,$Zhi
|
| - xor $Zlo,$tmp,$Zlo
|
| - xor $Thi,$Zhi,$Zhi
|
| - and $Zlo,0xf,$remi
|
| -
|
| - ldx [$Htblo+$nhi],$Tlo
|
| - sll $remi,3,$remi
|
| - xor $rem,$Zhi,$Zhi
|
| - ldx [$Htbl+$nhi],$Thi
|
| - srlx $Zlo,4,$Zlo
|
| - ldx [$rem_4bit+$remi],$rem
|
| - sllx $Zhi,60,$tmp
|
| - xor $Tlo,$Zlo,$Zlo
|
| - srlx $Zhi,4,$Zhi
|
| - xor $Zlo,$tmp,$Zlo
|
| - xor $Thi,$Zhi,$Zhi
|
| - stx $Zlo,[$Xi+8]
|
| - xor $rem,$Zhi,$Zhi
|
| - stx $Zhi,[$Xi]
|
| -
|
| - ret
|
| - restore
|
| -.type gcm_gmult_4bit,#function
|
| -.size gcm_gmult_4bit,(.-gcm_gmult_4bit)
|
| -.asciz "GHASH for SPARCv9, CRYPTOGAMS by <appro\@openssl.org>"
|
| -.align 4
|
| -___
|
| -
|
| -$code =~ s/\`([^\`]*)\`/eval $1/gem;
|
| -print $code;
|
| -close STDOUT;
|
|
|