| Index: openssl/crypto/bn/asm/x86_64-gf2m.S
|
| diff --git a/openssl/crypto/bn/asm/x86_64-gf2m.S b/openssl/crypto/bn/asm/x86_64-gf2m.S
|
| deleted file mode 100644
|
| index ccd2ed701c7cf0ac142e78c169b46abcfc27da8f..0000000000000000000000000000000000000000
|
| --- a/openssl/crypto/bn/asm/x86_64-gf2m.S
|
| +++ /dev/null
|
| @@ -1,291 +0,0 @@
|
| -.text
|
| -
|
| -.type _mul_1x1,@function
|
| -.align 16
|
| -_mul_1x1:
|
| - subq $128+8,%rsp
|
| - movq $-1,%r9
|
| - leaq (%rax,%rax,1),%rsi
|
| - shrq $3,%r9
|
| - leaq (,%rax,4),%rdi
|
| - andq %rax,%r9
|
| - leaq (,%rax,8),%r12
|
| - sarq $63,%rax
|
| - leaq (%r9,%r9,1),%r10
|
| - sarq $63,%rsi
|
| - leaq (,%r9,4),%r11
|
| - andq %rbp,%rax
|
| - sarq $63,%rdi
|
| - movq %rax,%rdx
|
| - shlq $63,%rax
|
| - andq %rbp,%rsi
|
| - shrq $1,%rdx
|
| - movq %rsi,%rcx
|
| - shlq $62,%rsi
|
| - andq %rbp,%rdi
|
| - shrq $2,%rcx
|
| - xorq %rsi,%rax
|
| - movq %rdi,%rbx
|
| - shlq $61,%rdi
|
| - xorq %rcx,%rdx
|
| - shrq $3,%rbx
|
| - xorq %rdi,%rax
|
| - xorq %rbx,%rdx
|
| -
|
| - movq %r9,%r13
|
| - movq $0,0(%rsp)
|
| - xorq %r10,%r13
|
| - movq %r9,8(%rsp)
|
| - movq %r11,%r14
|
| - movq %r10,16(%rsp)
|
| - xorq %r12,%r14
|
| - movq %r13,24(%rsp)
|
| -
|
| - xorq %r11,%r9
|
| - movq %r11,32(%rsp)
|
| - xorq %r11,%r10
|
| - movq %r9,40(%rsp)
|
| - xorq %r11,%r13
|
| - movq %r10,48(%rsp)
|
| - xorq %r14,%r9
|
| - movq %r13,56(%rsp)
|
| - xorq %r14,%r10
|
| -
|
| - movq %r12,64(%rsp)
|
| - xorq %r14,%r13
|
| - movq %r9,72(%rsp)
|
| - xorq %r11,%r9
|
| - movq %r10,80(%rsp)
|
| - xorq %r11,%r10
|
| - movq %r13,88(%rsp)
|
| -
|
| - xorq %r11,%r13
|
| - movq %r14,96(%rsp)
|
| - movq %r8,%rsi
|
| - movq %r9,104(%rsp)
|
| - andq %rbp,%rsi
|
| - movq %r10,112(%rsp)
|
| - shrq $4,%rbp
|
| - movq %r13,120(%rsp)
|
| - movq %r8,%rdi
|
| - andq %rbp,%rdi
|
| - shrq $4,%rbp
|
| -
|
| - movq (%rsp,%rsi,8),%xmm0
|
| - movq %r8,%rsi
|
| - andq %rbp,%rsi
|
| - shrq $4,%rbp
|
| - movq (%rsp,%rdi,8),%rcx
|
| - movq %r8,%rdi
|
| - movq %rcx,%rbx
|
| - shlq $4,%rcx
|
| - andq %rbp,%rdi
|
| - movq (%rsp,%rsi,8),%xmm1
|
| - shrq $60,%rbx
|
| - xorq %rcx,%rax
|
| - pslldq $1,%xmm1
|
| - movq %r8,%rsi
|
| - shrq $4,%rbp
|
| - xorq %rbx,%rdx
|
| - andq %rbp,%rsi
|
| - shrq $4,%rbp
|
| - pxor %xmm1,%xmm0
|
| - movq (%rsp,%rdi,8),%rcx
|
| - movq %r8,%rdi
|
| - movq %rcx,%rbx
|
| - shlq $12,%rcx
|
| - andq %rbp,%rdi
|
| - movq (%rsp,%rsi,8),%xmm1
|
| - shrq $52,%rbx
|
| - xorq %rcx,%rax
|
| - pslldq $2,%xmm1
|
| - movq %r8,%rsi
|
| - shrq $4,%rbp
|
| - xorq %rbx,%rdx
|
| - andq %rbp,%rsi
|
| - shrq $4,%rbp
|
| - pxor %xmm1,%xmm0
|
| - movq (%rsp,%rdi,8),%rcx
|
| - movq %r8,%rdi
|
| - movq %rcx,%rbx
|
| - shlq $20,%rcx
|
| - andq %rbp,%rdi
|
| - movq (%rsp,%rsi,8),%xmm1
|
| - shrq $44,%rbx
|
| - xorq %rcx,%rax
|
| - pslldq $3,%xmm1
|
| - movq %r8,%rsi
|
| - shrq $4,%rbp
|
| - xorq %rbx,%rdx
|
| - andq %rbp,%rsi
|
| - shrq $4,%rbp
|
| - pxor %xmm1,%xmm0
|
| - movq (%rsp,%rdi,8),%rcx
|
| - movq %r8,%rdi
|
| - movq %rcx,%rbx
|
| - shlq $28,%rcx
|
| - andq %rbp,%rdi
|
| - movq (%rsp,%rsi,8),%xmm1
|
| - shrq $36,%rbx
|
| - xorq %rcx,%rax
|
| - pslldq $4,%xmm1
|
| - movq %r8,%rsi
|
| - shrq $4,%rbp
|
| - xorq %rbx,%rdx
|
| - andq %rbp,%rsi
|
| - shrq $4,%rbp
|
| - pxor %xmm1,%xmm0
|
| - movq (%rsp,%rdi,8),%rcx
|
| - movq %r8,%rdi
|
| - movq %rcx,%rbx
|
| - shlq $36,%rcx
|
| - andq %rbp,%rdi
|
| - movq (%rsp,%rsi,8),%xmm1
|
| - shrq $28,%rbx
|
| - xorq %rcx,%rax
|
| - pslldq $5,%xmm1
|
| - movq %r8,%rsi
|
| - shrq $4,%rbp
|
| - xorq %rbx,%rdx
|
| - andq %rbp,%rsi
|
| - shrq $4,%rbp
|
| - pxor %xmm1,%xmm0
|
| - movq (%rsp,%rdi,8),%rcx
|
| - movq %r8,%rdi
|
| - movq %rcx,%rbx
|
| - shlq $44,%rcx
|
| - andq %rbp,%rdi
|
| - movq (%rsp,%rsi,8),%xmm1
|
| - shrq $20,%rbx
|
| - xorq %rcx,%rax
|
| - pslldq $6,%xmm1
|
| - movq %r8,%rsi
|
| - shrq $4,%rbp
|
| - xorq %rbx,%rdx
|
| - andq %rbp,%rsi
|
| - shrq $4,%rbp
|
| - pxor %xmm1,%xmm0
|
| - movq (%rsp,%rdi,8),%rcx
|
| - movq %r8,%rdi
|
| - movq %rcx,%rbx
|
| - shlq $52,%rcx
|
| - andq %rbp,%rdi
|
| - movq (%rsp,%rsi,8),%xmm1
|
| - shrq $12,%rbx
|
| - xorq %rcx,%rax
|
| - pslldq $7,%xmm1
|
| - movq %r8,%rsi
|
| - shrq $4,%rbp
|
| - xorq %rbx,%rdx
|
| - andq %rbp,%rsi
|
| - shrq $4,%rbp
|
| - pxor %xmm1,%xmm0
|
| - movq (%rsp,%rdi,8),%rcx
|
| - movq %rcx,%rbx
|
| - shlq $60,%rcx
|
| -.byte 102,72,15,126,198
|
| - shrq $4,%rbx
|
| - xorq %rcx,%rax
|
| - psrldq $8,%xmm0
|
| - xorq %rbx,%rdx
|
| -.byte 102,72,15,126,199
|
| - xorq %rsi,%rax
|
| - xorq %rdi,%rdx
|
| -
|
| - addq $128+8,%rsp
|
| - .byte 0xf3,0xc3
|
| -.Lend_mul_1x1:
|
| -.size _mul_1x1,.-_mul_1x1
|
| -
|
| -.globl bn_GF2m_mul_2x2
|
| -.type bn_GF2m_mul_2x2,@function
|
| -.align 16
|
| -bn_GF2m_mul_2x2:
|
| - movq OPENSSL_ia32cap_P(%rip),%rax
|
| - btq $33,%rax
|
| - jnc .Lvanilla_mul_2x2
|
| -
|
| -.byte 102,72,15,110,198
|
| -.byte 102,72,15,110,201
|
| -.byte 102,72,15,110,210
|
| -.byte 102,73,15,110,216
|
| - movdqa %xmm0,%xmm4
|
| - movdqa %xmm1,%xmm5
|
| -.byte 102,15,58,68,193,0
|
| - pxor %xmm2,%xmm4
|
| - pxor %xmm3,%xmm5
|
| -.byte 102,15,58,68,211,0
|
| -.byte 102,15,58,68,229,0
|
| - xorps %xmm0,%xmm4
|
| - xorps %xmm2,%xmm4
|
| - movdqa %xmm4,%xmm5
|
| - pslldq $8,%xmm4
|
| - psrldq $8,%xmm5
|
| - pxor %xmm4,%xmm2
|
| - pxor %xmm5,%xmm0
|
| - movdqu %xmm2,0(%rdi)
|
| - movdqu %xmm0,16(%rdi)
|
| - .byte 0xf3,0xc3
|
| -
|
| -.align 16
|
| -.Lvanilla_mul_2x2:
|
| - leaq -136(%rsp),%rsp
|
| - movq %r14,80(%rsp)
|
| - movq %r13,88(%rsp)
|
| - movq %r12,96(%rsp)
|
| - movq %rbp,104(%rsp)
|
| - movq %rbx,112(%rsp)
|
| -.Lbody_mul_2x2:
|
| - movq %rdi,32(%rsp)
|
| - movq %rsi,40(%rsp)
|
| - movq %rdx,48(%rsp)
|
| - movq %rcx,56(%rsp)
|
| - movq %r8,64(%rsp)
|
| -
|
| - movq $15,%r8
|
| - movq %rsi,%rax
|
| - movq %rcx,%rbp
|
| - call _mul_1x1
|
| - movq %rax,16(%rsp)
|
| - movq %rdx,24(%rsp)
|
| -
|
| - movq 48(%rsp),%rax
|
| - movq 64(%rsp),%rbp
|
| - call _mul_1x1
|
| - movq %rax,0(%rsp)
|
| - movq %rdx,8(%rsp)
|
| -
|
| - movq 40(%rsp),%rax
|
| - movq 56(%rsp),%rbp
|
| - xorq 48(%rsp),%rax
|
| - xorq 64(%rsp),%rbp
|
| - call _mul_1x1
|
| - movq 0(%rsp),%rbx
|
| - movq 8(%rsp),%rcx
|
| - movq 16(%rsp),%rdi
|
| - movq 24(%rsp),%rsi
|
| - movq 32(%rsp),%rbp
|
| -
|
| - xorq %rdx,%rax
|
| - xorq %rcx,%rdx
|
| - xorq %rbx,%rax
|
| - movq %rbx,0(%rbp)
|
| - xorq %rdi,%rdx
|
| - movq %rsi,24(%rbp)
|
| - xorq %rsi,%rax
|
| - xorq %rsi,%rdx
|
| - xorq %rdx,%rax
|
| - movq %rdx,16(%rbp)
|
| - movq %rax,8(%rbp)
|
| -
|
| - movq 80(%rsp),%r14
|
| - movq 88(%rsp),%r13
|
| - movq 96(%rsp),%r12
|
| - movq 104(%rsp),%rbp
|
| - movq 112(%rsp),%rbx
|
| - leaq 136(%rsp),%rsp
|
| - .byte 0xf3,0xc3
|
| -.Lend_mul_2x2:
|
| -.size bn_GF2m_mul_2x2,.-bn_GF2m_mul_2x2
|
| -.byte 71,70,40,50,94,109,41,32,77,117,108,116,105,112,108,105,99,97,116,105,111,110,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,103,62,0
|
| -.align 16
|
|
|