| Index: third_party/boringssl/linux-aarch64/crypto/modes/ghashv8-armx64.S
|
| diff --git a/third_party/boringssl/linux-aarch64/crypto/modes/ghashv8-armx64.S b/third_party/boringssl/linux-aarch64/crypto/modes/ghashv8-armx64.S
|
| deleted file mode 100644
|
| index f39f3ba870b7d660d1f721c091ed486344bf3799..0000000000000000000000000000000000000000
|
| --- a/third_party/boringssl/linux-aarch64/crypto/modes/ghashv8-armx64.S
|
| +++ /dev/null
|
| @@ -1,235 +0,0 @@
|
| -#if defined(__aarch64__)
|
| -#include <openssl/arm_arch.h>
|
| -
|
| -.text
|
| -#if !defined(__clang__)
|
| -.arch armv8-a+crypto
|
| -#endif
|
| -.globl gcm_init_v8
|
| -.hidden gcm_init_v8
|
| -.type gcm_init_v8,%function
|
| -.align 4
|
| -gcm_init_v8:
|
| - ld1 {v17.2d},[x1] //load input H
|
| - movi v19.16b,#0xe1
|
| - shl v19.2d,v19.2d,#57 //0xc2.0
|
| - ext v3.16b,v17.16b,v17.16b,#8
|
| - ushr v18.2d,v19.2d,#63
|
| - dup v17.4s,v17.s[1]
|
| - ext v16.16b,v18.16b,v19.16b,#8 //t0=0xc2....01
|
| - ushr v18.2d,v3.2d,#63
|
| - sshr v17.4s,v17.4s,#31 //broadcast carry bit
|
| - and v18.16b,v18.16b,v16.16b
|
| - shl v3.2d,v3.2d,#1
|
| - ext v18.16b,v18.16b,v18.16b,#8
|
| - and v16.16b,v16.16b,v17.16b
|
| - orr v3.16b,v3.16b,v18.16b //H<<<=1
|
| - eor v20.16b,v3.16b,v16.16b //twisted H
|
| - st1 {v20.2d},[x0],#16 //store Htable[0]
|
| -
|
| - //calculate H^2
|
| - ext v16.16b,v20.16b,v20.16b,#8 //Karatsuba pre-processing
|
| - pmull v0.1q,v20.1d,v20.1d
|
| - eor v16.16b,v16.16b,v20.16b
|
| - pmull2 v2.1q,v20.2d,v20.2d
|
| - pmull v1.1q,v16.1d,v16.1d
|
| -
|
| - ext v17.16b,v0.16b,v2.16b,#8 //Karatsuba post-processing
|
| - eor v18.16b,v0.16b,v2.16b
|
| - eor v1.16b,v1.16b,v17.16b
|
| - eor v1.16b,v1.16b,v18.16b
|
| - pmull v18.1q,v0.1d,v19.1d //1st phase
|
| -
|
| - ins v2.d[0],v1.d[1]
|
| - ins v1.d[1],v0.d[0]
|
| - eor v0.16b,v1.16b,v18.16b
|
| -
|
| - ext v18.16b,v0.16b,v0.16b,#8 //2nd phase
|
| - pmull v0.1q,v0.1d,v19.1d
|
| - eor v18.16b,v18.16b,v2.16b
|
| - eor v22.16b,v0.16b,v18.16b
|
| -
|
| - ext v17.16b,v22.16b,v22.16b,#8 //Karatsuba pre-processing
|
| - eor v17.16b,v17.16b,v22.16b
|
| - ext v21.16b,v16.16b,v17.16b,#8 //pack Karatsuba pre-processed
|
| - st1 {v21.2d,v22.2d},[x0] //store Htable[1..2]
|
| -
|
| - ret
|
| -.size gcm_init_v8,.-gcm_init_v8
|
| -.globl gcm_gmult_v8
|
| -.hidden gcm_gmult_v8
|
| -.type gcm_gmult_v8,%function
|
| -.align 4
|
| -gcm_gmult_v8:
|
| - ld1 {v17.2d},[x0] //load Xi
|
| - movi v19.16b,#0xe1
|
| - ld1 {v20.2d,v21.2d},[x1] //load twisted H, ...
|
| - shl v19.2d,v19.2d,#57
|
| -#ifndef __ARMEB__
|
| - rev64 v17.16b,v17.16b
|
| -#endif
|
| - ext v3.16b,v17.16b,v17.16b,#8
|
| -
|
| - pmull v0.1q,v20.1d,v3.1d //H.lo·Xi.lo
|
| - eor v17.16b,v17.16b,v3.16b //Karatsuba pre-processing
|
| - pmull2 v2.1q,v20.2d,v3.2d //H.hi·Xi.hi
|
| - pmull v1.1q,v21.1d,v17.1d //(H.lo+H.hi)·(Xi.lo+Xi.hi)
|
| -
|
| - ext v17.16b,v0.16b,v2.16b,#8 //Karatsuba post-processing
|
| - eor v18.16b,v0.16b,v2.16b
|
| - eor v1.16b,v1.16b,v17.16b
|
| - eor v1.16b,v1.16b,v18.16b
|
| - pmull v18.1q,v0.1d,v19.1d //1st phase of reduction
|
| -
|
| - ins v2.d[0],v1.d[1]
|
| - ins v1.d[1],v0.d[0]
|
| - eor v0.16b,v1.16b,v18.16b
|
| -
|
| - ext v18.16b,v0.16b,v0.16b,#8 //2nd phase of reduction
|
| - pmull v0.1q,v0.1d,v19.1d
|
| - eor v18.16b,v18.16b,v2.16b
|
| - eor v0.16b,v0.16b,v18.16b
|
| -
|
| -#ifndef __ARMEB__
|
| - rev64 v0.16b,v0.16b
|
| -#endif
|
| - ext v0.16b,v0.16b,v0.16b,#8
|
| - st1 {v0.2d},[x0] //write out Xi
|
| -
|
| - ret
|
| -.size gcm_gmult_v8,.-gcm_gmult_v8
|
| -.globl gcm_ghash_v8
|
| -.hidden gcm_ghash_v8
|
| -.type gcm_ghash_v8,%function
|
| -.align 4
|
| -gcm_ghash_v8:
|
| - ld1 {v0.2d},[x0] //load [rotated] Xi
|
| - //"[rotated]" means that
|
| - //loaded value would have
|
| - //to be rotated in order to
|
| - //make it appear as in
|
| - //alorithm specification
|
| - subs x3,x3,#32 //see if x3 is 32 or larger
|
| - mov x12,#16 //x12 is used as post-
|
| - //increment for input pointer;
|
| - //as loop is modulo-scheduled
|
| - //x12 is zeroed just in time
|
| - //to preclude oversteping
|
| - //inp[len], which means that
|
| - //last block[s] are actually
|
| - //loaded twice, but last
|
| - //copy is not processed
|
| - ld1 {v20.2d,v21.2d},[x1],#32 //load twisted H, ..., H^2
|
| - movi v19.16b,#0xe1
|
| - ld1 {v22.2d},[x1]
|
| - csel x12,xzr,x12,eq //is it time to zero x12?
|
| - ext v0.16b,v0.16b,v0.16b,#8 //rotate Xi
|
| - ld1 {v16.2d},[x2],#16 //load [rotated] I[0]
|
| - shl v19.2d,v19.2d,#57 //compose 0xc2.0 constant
|
| -#ifndef __ARMEB__
|
| - rev64 v16.16b,v16.16b
|
| - rev64 v0.16b,v0.16b
|
| -#endif
|
| - ext v3.16b,v16.16b,v16.16b,#8 //rotate I[0]
|
| - b.lo .Lodd_tail_v8 //x3 was less than 32
|
| - ld1 {v17.2d},[x2],x12 //load [rotated] I[1]
|
| -#ifndef __ARMEB__
|
| - rev64 v17.16b,v17.16b
|
| -#endif
|
| - ext v7.16b,v17.16b,v17.16b,#8
|
| - eor v3.16b,v3.16b,v0.16b //I[i]^=Xi
|
| - pmull v4.1q,v20.1d,v7.1d //H·Ii+1
|
| - eor v17.16b,v17.16b,v7.16b //Karatsuba pre-processing
|
| - pmull2 v6.1q,v20.2d,v7.2d
|
| - b .Loop_mod2x_v8
|
| -
|
| -.align 4
|
| -.Loop_mod2x_v8:
|
| - ext v18.16b,v3.16b,v3.16b,#8
|
| - subs x3,x3,#32 //is there more data?
|
| - pmull v0.1q,v22.1d,v3.1d //H^2.lo·Xi.lo
|
| - csel x12,xzr,x12,lo //is it time to zero x12?
|
| -
|
| - pmull v5.1q,v21.1d,v17.1d
|
| - eor v18.16b,v18.16b,v3.16b //Karatsuba pre-processing
|
| - pmull2 v2.1q,v22.2d,v3.2d //H^2.hi·Xi.hi
|
| - eor v0.16b,v0.16b,v4.16b //accumulate
|
| - pmull2 v1.1q,v21.2d,v18.2d //(H^2.lo+H^2.hi)·(Xi.lo+Xi.hi)
|
| - ld1 {v16.2d},[x2],x12 //load [rotated] I[i+2]
|
| -
|
| - eor v2.16b,v2.16b,v6.16b
|
| - csel x12,xzr,x12,eq //is it time to zero x12?
|
| - eor v1.16b,v1.16b,v5.16b
|
| -
|
| - ext v17.16b,v0.16b,v2.16b,#8 //Karatsuba post-processing
|
| - eor v18.16b,v0.16b,v2.16b
|
| - eor v1.16b,v1.16b,v17.16b
|
| - ld1 {v17.2d},[x2],x12 //load [rotated] I[i+3]
|
| -#ifndef __ARMEB__
|
| - rev64 v16.16b,v16.16b
|
| -#endif
|
| - eor v1.16b,v1.16b,v18.16b
|
| - pmull v18.1q,v0.1d,v19.1d //1st phase of reduction
|
| -
|
| -#ifndef __ARMEB__
|
| - rev64 v17.16b,v17.16b
|
| -#endif
|
| - ins v2.d[0],v1.d[1]
|
| - ins v1.d[1],v0.d[0]
|
| - ext v7.16b,v17.16b,v17.16b,#8
|
| - ext v3.16b,v16.16b,v16.16b,#8
|
| - eor v0.16b,v1.16b,v18.16b
|
| - pmull v4.1q,v20.1d,v7.1d //H·Ii+1
|
| - eor v3.16b,v3.16b,v2.16b //accumulate v3.16b early
|
| -
|
| - ext v18.16b,v0.16b,v0.16b,#8 //2nd phase of reduction
|
| - pmull v0.1q,v0.1d,v19.1d
|
| - eor v3.16b,v3.16b,v18.16b
|
| - eor v17.16b,v17.16b,v7.16b //Karatsuba pre-processing
|
| - eor v3.16b,v3.16b,v0.16b
|
| - pmull2 v6.1q,v20.2d,v7.2d
|
| - b.hs .Loop_mod2x_v8 //there was at least 32 more bytes
|
| -
|
| - eor v2.16b,v2.16b,v18.16b
|
| - ext v3.16b,v16.16b,v16.16b,#8 //re-construct v3.16b
|
| - adds x3,x3,#32 //re-construct x3
|
| - eor v0.16b,v0.16b,v2.16b //re-construct v0.16b
|
| - b.eq .Ldone_v8 //is x3 zero?
|
| -.Lodd_tail_v8:
|
| - ext v18.16b,v0.16b,v0.16b,#8
|
| - eor v3.16b,v3.16b,v0.16b //inp^=Xi
|
| - eor v17.16b,v16.16b,v18.16b //v17.16b is rotated inp^Xi
|
| -
|
| - pmull v0.1q,v20.1d,v3.1d //H.lo·Xi.lo
|
| - eor v17.16b,v17.16b,v3.16b //Karatsuba pre-processing
|
| - pmull2 v2.1q,v20.2d,v3.2d //H.hi·Xi.hi
|
| - pmull v1.1q,v21.1d,v17.1d //(H.lo+H.hi)·(Xi.lo+Xi.hi)
|
| -
|
| - ext v17.16b,v0.16b,v2.16b,#8 //Karatsuba post-processing
|
| - eor v18.16b,v0.16b,v2.16b
|
| - eor v1.16b,v1.16b,v17.16b
|
| - eor v1.16b,v1.16b,v18.16b
|
| - pmull v18.1q,v0.1d,v19.1d //1st phase of reduction
|
| -
|
| - ins v2.d[0],v1.d[1]
|
| - ins v1.d[1],v0.d[0]
|
| - eor v0.16b,v1.16b,v18.16b
|
| -
|
| - ext v18.16b,v0.16b,v0.16b,#8 //2nd phase of reduction
|
| - pmull v0.1q,v0.1d,v19.1d
|
| - eor v18.16b,v18.16b,v2.16b
|
| - eor v0.16b,v0.16b,v18.16b
|
| -
|
| -.Ldone_v8:
|
| -#ifndef __ARMEB__
|
| - rev64 v0.16b,v0.16b
|
| -#endif
|
| - ext v0.16b,v0.16b,v0.16b,#8
|
| - st1 {v0.2d},[x0] //write out Xi
|
| -
|
| - ret
|
| -.size gcm_ghash_v8,.-gcm_ghash_v8
|
| -.byte 71,72,65,83,72,32,102,111,114,32,65,82,77,118,56,44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,103,62,0
|
| -.align 2
|
| -.align 2
|
| -#endif
|
|
|