Index: openssl/crypto/bn/asm/x86_64-gf2m.S |
diff --git a/openssl/crypto/bn/asm/x86_64-gf2m.S b/openssl/crypto/bn/asm/x86_64-gf2m.S |
deleted file mode 100644 |
index ccd2ed701c7cf0ac142e78c169b46abcfc27da8f..0000000000000000000000000000000000000000 |
--- a/openssl/crypto/bn/asm/x86_64-gf2m.S |
+++ /dev/null |
@@ -1,291 +0,0 @@ |
-.text |
- |
-.type _mul_1x1,@function |
-.align 16 |
-_mul_1x1: |
- subq $128+8,%rsp |
- movq $-1,%r9 |
- leaq (%rax,%rax,1),%rsi |
- shrq $3,%r9 |
- leaq (,%rax,4),%rdi |
- andq %rax,%r9 |
- leaq (,%rax,8),%r12 |
- sarq $63,%rax |
- leaq (%r9,%r9,1),%r10 |
- sarq $63,%rsi |
- leaq (,%r9,4),%r11 |
- andq %rbp,%rax |
- sarq $63,%rdi |
- movq %rax,%rdx |
- shlq $63,%rax |
- andq %rbp,%rsi |
- shrq $1,%rdx |
- movq %rsi,%rcx |
- shlq $62,%rsi |
- andq %rbp,%rdi |
- shrq $2,%rcx |
- xorq %rsi,%rax |
- movq %rdi,%rbx |
- shlq $61,%rdi |
- xorq %rcx,%rdx |
- shrq $3,%rbx |
- xorq %rdi,%rax |
- xorq %rbx,%rdx |
- |
- movq %r9,%r13 |
- movq $0,0(%rsp) |
- xorq %r10,%r13 |
- movq %r9,8(%rsp) |
- movq %r11,%r14 |
- movq %r10,16(%rsp) |
- xorq %r12,%r14 |
- movq %r13,24(%rsp) |
- |
- xorq %r11,%r9 |
- movq %r11,32(%rsp) |
- xorq %r11,%r10 |
- movq %r9,40(%rsp) |
- xorq %r11,%r13 |
- movq %r10,48(%rsp) |
- xorq %r14,%r9 |
- movq %r13,56(%rsp) |
- xorq %r14,%r10 |
- |
- movq %r12,64(%rsp) |
- xorq %r14,%r13 |
- movq %r9,72(%rsp) |
- xorq %r11,%r9 |
- movq %r10,80(%rsp) |
- xorq %r11,%r10 |
- movq %r13,88(%rsp) |
- |
- xorq %r11,%r13 |
- movq %r14,96(%rsp) |
- movq %r8,%rsi |
- movq %r9,104(%rsp) |
- andq %rbp,%rsi |
- movq %r10,112(%rsp) |
- shrq $4,%rbp |
- movq %r13,120(%rsp) |
- movq %r8,%rdi |
- andq %rbp,%rdi |
- shrq $4,%rbp |
- |
- movq (%rsp,%rsi,8),%xmm0 |
- movq %r8,%rsi |
- andq %rbp,%rsi |
- shrq $4,%rbp |
- movq (%rsp,%rdi,8),%rcx |
- movq %r8,%rdi |
- movq %rcx,%rbx |
- shlq $4,%rcx |
- andq %rbp,%rdi |
- movq (%rsp,%rsi,8),%xmm1 |
- shrq $60,%rbx |
- xorq %rcx,%rax |
- pslldq $1,%xmm1 |
- movq %r8,%rsi |
- shrq $4,%rbp |
- xorq %rbx,%rdx |
- andq %rbp,%rsi |
- shrq $4,%rbp |
- pxor %xmm1,%xmm0 |
- movq (%rsp,%rdi,8),%rcx |
- movq %r8,%rdi |
- movq %rcx,%rbx |
- shlq $12,%rcx |
- andq %rbp,%rdi |
- movq (%rsp,%rsi,8),%xmm1 |
- shrq $52,%rbx |
- xorq %rcx,%rax |
- pslldq $2,%xmm1 |
- movq %r8,%rsi |
- shrq $4,%rbp |
- xorq %rbx,%rdx |
- andq %rbp,%rsi |
- shrq $4,%rbp |
- pxor %xmm1,%xmm0 |
- movq (%rsp,%rdi,8),%rcx |
- movq %r8,%rdi |
- movq %rcx,%rbx |
- shlq $20,%rcx |
- andq %rbp,%rdi |
- movq (%rsp,%rsi,8),%xmm1 |
- shrq $44,%rbx |
- xorq %rcx,%rax |
- pslldq $3,%xmm1 |
- movq %r8,%rsi |
- shrq $4,%rbp |
- xorq %rbx,%rdx |
- andq %rbp,%rsi |
- shrq $4,%rbp |
- pxor %xmm1,%xmm0 |
- movq (%rsp,%rdi,8),%rcx |
- movq %r8,%rdi |
- movq %rcx,%rbx |
- shlq $28,%rcx |
- andq %rbp,%rdi |
- movq (%rsp,%rsi,8),%xmm1 |
- shrq $36,%rbx |
- xorq %rcx,%rax |
- pslldq $4,%xmm1 |
- movq %r8,%rsi |
- shrq $4,%rbp |
- xorq %rbx,%rdx |
- andq %rbp,%rsi |
- shrq $4,%rbp |
- pxor %xmm1,%xmm0 |
- movq (%rsp,%rdi,8),%rcx |
- movq %r8,%rdi |
- movq %rcx,%rbx |
- shlq $36,%rcx |
- andq %rbp,%rdi |
- movq (%rsp,%rsi,8),%xmm1 |
- shrq $28,%rbx |
- xorq %rcx,%rax |
- pslldq $5,%xmm1 |
- movq %r8,%rsi |
- shrq $4,%rbp |
- xorq %rbx,%rdx |
- andq %rbp,%rsi |
- shrq $4,%rbp |
- pxor %xmm1,%xmm0 |
- movq (%rsp,%rdi,8),%rcx |
- movq %r8,%rdi |
- movq %rcx,%rbx |
- shlq $44,%rcx |
- andq %rbp,%rdi |
- movq (%rsp,%rsi,8),%xmm1 |
- shrq $20,%rbx |
- xorq %rcx,%rax |
- pslldq $6,%xmm1 |
- movq %r8,%rsi |
- shrq $4,%rbp |
- xorq %rbx,%rdx |
- andq %rbp,%rsi |
- shrq $4,%rbp |
- pxor %xmm1,%xmm0 |
- movq (%rsp,%rdi,8),%rcx |
- movq %r8,%rdi |
- movq %rcx,%rbx |
- shlq $52,%rcx |
- andq %rbp,%rdi |
- movq (%rsp,%rsi,8),%xmm1 |
- shrq $12,%rbx |
- xorq %rcx,%rax |
- pslldq $7,%xmm1 |
- movq %r8,%rsi |
- shrq $4,%rbp |
- xorq %rbx,%rdx |
- andq %rbp,%rsi |
- shrq $4,%rbp |
- pxor %xmm1,%xmm0 |
- movq (%rsp,%rdi,8),%rcx |
- movq %rcx,%rbx |
- shlq $60,%rcx |
-.byte 102,72,15,126,198 |
- shrq $4,%rbx |
- xorq %rcx,%rax |
- psrldq $8,%xmm0 |
- xorq %rbx,%rdx |
-.byte 102,72,15,126,199 |
- xorq %rsi,%rax |
- xorq %rdi,%rdx |
- |
- addq $128+8,%rsp |
- .byte 0xf3,0xc3 |
-.Lend_mul_1x1: |
-.size _mul_1x1,.-_mul_1x1 |
- |
-.globl bn_GF2m_mul_2x2 |
-.type bn_GF2m_mul_2x2,@function |
-.align 16 |
-bn_GF2m_mul_2x2: |
- movq OPENSSL_ia32cap_P(%rip),%rax |
- btq $33,%rax |
- jnc .Lvanilla_mul_2x2 |
- |
-.byte 102,72,15,110,198 |
-.byte 102,72,15,110,201 |
-.byte 102,72,15,110,210 |
-.byte 102,73,15,110,216 |
- movdqa %xmm0,%xmm4 |
- movdqa %xmm1,%xmm5 |
-.byte 102,15,58,68,193,0 |
- pxor %xmm2,%xmm4 |
- pxor %xmm3,%xmm5 |
-.byte 102,15,58,68,211,0 |
-.byte 102,15,58,68,229,0 |
- xorps %xmm0,%xmm4 |
- xorps %xmm2,%xmm4 |
- movdqa %xmm4,%xmm5 |
- pslldq $8,%xmm4 |
- psrldq $8,%xmm5 |
- pxor %xmm4,%xmm2 |
- pxor %xmm5,%xmm0 |
- movdqu %xmm2,0(%rdi) |
- movdqu %xmm0,16(%rdi) |
- .byte 0xf3,0xc3 |
- |
-.align 16 |
-.Lvanilla_mul_2x2: |
- leaq -136(%rsp),%rsp |
- movq %r14,80(%rsp) |
- movq %r13,88(%rsp) |
- movq %r12,96(%rsp) |
- movq %rbp,104(%rsp) |
- movq %rbx,112(%rsp) |
-.Lbody_mul_2x2: |
- movq %rdi,32(%rsp) |
- movq %rsi,40(%rsp) |
- movq %rdx,48(%rsp) |
- movq %rcx,56(%rsp) |
- movq %r8,64(%rsp) |
- |
- movq $15,%r8 |
- movq %rsi,%rax |
- movq %rcx,%rbp |
- call _mul_1x1 |
- movq %rax,16(%rsp) |
- movq %rdx,24(%rsp) |
- |
- movq 48(%rsp),%rax |
- movq 64(%rsp),%rbp |
- call _mul_1x1 |
- movq %rax,0(%rsp) |
- movq %rdx,8(%rsp) |
- |
- movq 40(%rsp),%rax |
- movq 56(%rsp),%rbp |
- xorq 48(%rsp),%rax |
- xorq 64(%rsp),%rbp |
- call _mul_1x1 |
- movq 0(%rsp),%rbx |
- movq 8(%rsp),%rcx |
- movq 16(%rsp),%rdi |
- movq 24(%rsp),%rsi |
- movq 32(%rsp),%rbp |
- |
- xorq %rdx,%rax |
- xorq %rcx,%rdx |
- xorq %rbx,%rax |
- movq %rbx,0(%rbp) |
- xorq %rdi,%rdx |
- movq %rsi,24(%rbp) |
- xorq %rsi,%rax |
- xorq %rsi,%rdx |
- xorq %rdx,%rax |
- movq %rdx,16(%rbp) |
- movq %rax,8(%rbp) |
- |
- movq 80(%rsp),%r14 |
- movq 88(%rsp),%r13 |
- movq 96(%rsp),%r12 |
- movq 104(%rsp),%rbp |
- movq 112(%rsp),%rbx |
- leaq 136(%rsp),%rsp |
- .byte 0xf3,0xc3 |
-.Lend_mul_2x2: |
-.size bn_GF2m_mul_2x2,.-bn_GF2m_mul_2x2 |
-.byte 71,70,40,50,94,109,41,32,77,117,108,116,105,112,108,105,99,97,116,105,111,110,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,103,62,0 |
-.align 16 |