Index: openssl/crypto/bn/asm/x86_64-mont5.S |
diff --git a/openssl/crypto/bn/asm/x86_64-mont5.S b/openssl/crypto/bn/asm/x86_64-mont5.S |
deleted file mode 100644 |
index 49ec6ac67b471db42b11128ab01a6b53160d4c76..0000000000000000000000000000000000000000 |
--- a/openssl/crypto/bn/asm/x86_64-mont5.S |
+++ /dev/null |
@@ -1,784 +0,0 @@ |
-.text |
- |
-.globl bn_mul_mont_gather5 |
-.type bn_mul_mont_gather5,@function |
-.align 64 |
-bn_mul_mont_gather5: |
- testl $3,%r9d |
- jnz .Lmul_enter |
- cmpl $8,%r9d |
- jb .Lmul_enter |
- jmp .Lmul4x_enter |
- |
-.align 16 |
-.Lmul_enter: |
- movl %r9d,%r9d |
- movl 8(%rsp),%r10d |
- pushq %rbx |
- pushq %rbp |
- pushq %r12 |
- pushq %r13 |
- pushq %r14 |
- pushq %r15 |
- movq %rsp,%rax |
- leaq 2(%r9),%r11 |
- negq %r11 |
- leaq (%rsp,%r11,8),%rsp |
- andq $-1024,%rsp |
- |
- movq %rax,8(%rsp,%r9,8) |
-.Lmul_body: |
- movq %rdx,%r12 |
- movq %r10,%r11 |
- shrq $3,%r10 |
- andq $7,%r11 |
- notq %r10 |
- leaq .Lmagic_masks(%rip),%rax |
- andq $3,%r10 |
- leaq 96(%r12,%r11,8),%r12 |
- movq 0(%rax,%r10,8),%xmm4 |
- movq 8(%rax,%r10,8),%xmm5 |
- movq 16(%rax,%r10,8),%xmm6 |
- movq 24(%rax,%r10,8),%xmm7 |
- |
- movq -96(%r12),%xmm0 |
- movq -32(%r12),%xmm1 |
- pand %xmm4,%xmm0 |
- movq 32(%r12),%xmm2 |
- pand %xmm5,%xmm1 |
- movq 96(%r12),%xmm3 |
- pand %xmm6,%xmm2 |
- por %xmm1,%xmm0 |
- pand %xmm7,%xmm3 |
- por %xmm2,%xmm0 |
- leaq 256(%r12),%r12 |
- por %xmm3,%xmm0 |
- |
-.byte 102,72,15,126,195 |
- |
- movq (%r8),%r8 |
- movq (%rsi),%rax |
- |
- xorq %r14,%r14 |
- xorq %r15,%r15 |
- |
- movq -96(%r12),%xmm0 |
- movq -32(%r12),%xmm1 |
- pand %xmm4,%xmm0 |
- movq 32(%r12),%xmm2 |
- pand %xmm5,%xmm1 |
- |
- movq %r8,%rbp |
- mulq %rbx |
- movq %rax,%r10 |
- movq (%rcx),%rax |
- |
- movq 96(%r12),%xmm3 |
- pand %xmm6,%xmm2 |
- por %xmm1,%xmm0 |
- pand %xmm7,%xmm3 |
- |
- imulq %r10,%rbp |
- movq %rdx,%r11 |
- |
- por %xmm2,%xmm0 |
- leaq 256(%r12),%r12 |
- por %xmm3,%xmm0 |
- |
- mulq %rbp |
- addq %rax,%r10 |
- movq 8(%rsi),%rax |
- adcq $0,%rdx |
- movq %rdx,%r13 |
- |
- leaq 1(%r15),%r15 |
- jmp .L1st_enter |
- |
-.align 16 |
-.L1st: |
- addq %rax,%r13 |
- movq (%rsi,%r15,8),%rax |
- adcq $0,%rdx |
- addq %r11,%r13 |
- movq %r10,%r11 |
- adcq $0,%rdx |
- movq %r13,-16(%rsp,%r15,8) |
- movq %rdx,%r13 |
- |
-.L1st_enter: |
- mulq %rbx |
- addq %rax,%r11 |
- movq (%rcx,%r15,8),%rax |
- adcq $0,%rdx |
- leaq 1(%r15),%r15 |
- movq %rdx,%r10 |
- |
- mulq %rbp |
- cmpq %r9,%r15 |
- jne .L1st |
- |
-.byte 102,72,15,126,195 |
- |
- addq %rax,%r13 |
- movq (%rsi),%rax |
- adcq $0,%rdx |
- addq %r11,%r13 |
- adcq $0,%rdx |
- movq %r13,-16(%rsp,%r15,8) |
- movq %rdx,%r13 |
- movq %r10,%r11 |
- |
- xorq %rdx,%rdx |
- addq %r11,%r13 |
- adcq $0,%rdx |
- movq %r13,-8(%rsp,%r9,8) |
- movq %rdx,(%rsp,%r9,8) |
- |
- leaq 1(%r14),%r14 |
- jmp .Louter |
-.align 16 |
-.Louter: |
- xorq %r15,%r15 |
- movq %r8,%rbp |
- movq (%rsp),%r10 |
- |
- movq -96(%r12),%xmm0 |
- movq -32(%r12),%xmm1 |
- pand %xmm4,%xmm0 |
- movq 32(%r12),%xmm2 |
- pand %xmm5,%xmm1 |
- |
- mulq %rbx |
- addq %rax,%r10 |
- movq (%rcx),%rax |
- adcq $0,%rdx |
- |
- movq 96(%r12),%xmm3 |
- pand %xmm6,%xmm2 |
- por %xmm1,%xmm0 |
- pand %xmm7,%xmm3 |
- |
- imulq %r10,%rbp |
- movq %rdx,%r11 |
- |
- por %xmm2,%xmm0 |
- leaq 256(%r12),%r12 |
- por %xmm3,%xmm0 |
- |
- mulq %rbp |
- addq %rax,%r10 |
- movq 8(%rsi),%rax |
- adcq $0,%rdx |
- movq 8(%rsp),%r10 |
- movq %rdx,%r13 |
- |
- leaq 1(%r15),%r15 |
- jmp .Linner_enter |
- |
-.align 16 |
-.Linner: |
- addq %rax,%r13 |
- movq (%rsi,%r15,8),%rax |
- adcq $0,%rdx |
- addq %r10,%r13 |
- movq (%rsp,%r15,8),%r10 |
- adcq $0,%rdx |
- movq %r13,-16(%rsp,%r15,8) |
- movq %rdx,%r13 |
- |
-.Linner_enter: |
- mulq %rbx |
- addq %rax,%r11 |
- movq (%rcx,%r15,8),%rax |
- adcq $0,%rdx |
- addq %r11,%r10 |
- movq %rdx,%r11 |
- adcq $0,%r11 |
- leaq 1(%r15),%r15 |
- |
- mulq %rbp |
- cmpq %r9,%r15 |
- jne .Linner |
- |
-.byte 102,72,15,126,195 |
- |
- addq %rax,%r13 |
- movq (%rsi),%rax |
- adcq $0,%rdx |
- addq %r10,%r13 |
- movq (%rsp,%r15,8),%r10 |
- adcq $0,%rdx |
- movq %r13,-16(%rsp,%r15,8) |
- movq %rdx,%r13 |
- |
- xorq %rdx,%rdx |
- addq %r11,%r13 |
- adcq $0,%rdx |
- addq %r10,%r13 |
- adcq $0,%rdx |
- movq %r13,-8(%rsp,%r9,8) |
- movq %rdx,(%rsp,%r9,8) |
- |
- leaq 1(%r14),%r14 |
- cmpq %r9,%r14 |
- jl .Louter |
- |
- xorq %r14,%r14 |
- movq (%rsp),%rax |
- leaq (%rsp),%rsi |
- movq %r9,%r15 |
- jmp .Lsub |
-.align 16 |
-.Lsub: sbbq (%rcx,%r14,8),%rax |
- movq %rax,(%rdi,%r14,8) |
- movq 8(%rsi,%r14,8),%rax |
- leaq 1(%r14),%r14 |
- decq %r15 |
- jnz .Lsub |
- |
- sbbq $0,%rax |
- xorq %r14,%r14 |
- andq %rax,%rsi |
- notq %rax |
- movq %rdi,%rcx |
- andq %rax,%rcx |
- movq %r9,%r15 |
- orq %rcx,%rsi |
-.align 16 |
-.Lcopy: |
- movq (%rsi,%r14,8),%rax |
- movq %r14,(%rsp,%r14,8) |
- movq %rax,(%rdi,%r14,8) |
- leaq 1(%r14),%r14 |
- subq $1,%r15 |
- jnz .Lcopy |
- |
- movq 8(%rsp,%r9,8),%rsi |
- movq $1,%rax |
- movq (%rsi),%r15 |
- movq 8(%rsi),%r14 |
- movq 16(%rsi),%r13 |
- movq 24(%rsi),%r12 |
- movq 32(%rsi),%rbp |
- movq 40(%rsi),%rbx |
- leaq 48(%rsi),%rsp |
-.Lmul_epilogue: |
- .byte 0xf3,0xc3 |
-.size bn_mul_mont_gather5,.-bn_mul_mont_gather5 |
-.type bn_mul4x_mont_gather5,@function |
-.align 16 |
-bn_mul4x_mont_gather5: |
-.Lmul4x_enter: |
- movl %r9d,%r9d |
- movl 8(%rsp),%r10d |
- pushq %rbx |
- pushq %rbp |
- pushq %r12 |
- pushq %r13 |
- pushq %r14 |
- pushq %r15 |
- movq %rsp,%rax |
- leaq 4(%r9),%r11 |
- negq %r11 |
- leaq (%rsp,%r11,8),%rsp |
- andq $-1024,%rsp |
- |
- movq %rax,8(%rsp,%r9,8) |
-.Lmul4x_body: |
- movq %rdi,16(%rsp,%r9,8) |
- movq %rdx,%r12 |
- movq %r10,%r11 |
- shrq $3,%r10 |
- andq $7,%r11 |
- notq %r10 |
- leaq .Lmagic_masks(%rip),%rax |
- andq $3,%r10 |
- leaq 96(%r12,%r11,8),%r12 |
- movq 0(%rax,%r10,8),%xmm4 |
- movq 8(%rax,%r10,8),%xmm5 |
- movq 16(%rax,%r10,8),%xmm6 |
- movq 24(%rax,%r10,8),%xmm7 |
- |
- movq -96(%r12),%xmm0 |
- movq -32(%r12),%xmm1 |
- pand %xmm4,%xmm0 |
- movq 32(%r12),%xmm2 |
- pand %xmm5,%xmm1 |
- movq 96(%r12),%xmm3 |
- pand %xmm6,%xmm2 |
- por %xmm1,%xmm0 |
- pand %xmm7,%xmm3 |
- por %xmm2,%xmm0 |
- leaq 256(%r12),%r12 |
- por %xmm3,%xmm0 |
- |
-.byte 102,72,15,126,195 |
- movq (%r8),%r8 |
- movq (%rsi),%rax |
- |
- xorq %r14,%r14 |
- xorq %r15,%r15 |
- |
- movq -96(%r12),%xmm0 |
- movq -32(%r12),%xmm1 |
- pand %xmm4,%xmm0 |
- movq 32(%r12),%xmm2 |
- pand %xmm5,%xmm1 |
- |
- movq %r8,%rbp |
- mulq %rbx |
- movq %rax,%r10 |
- movq (%rcx),%rax |
- |
- movq 96(%r12),%xmm3 |
- pand %xmm6,%xmm2 |
- por %xmm1,%xmm0 |
- pand %xmm7,%xmm3 |
- |
- imulq %r10,%rbp |
- movq %rdx,%r11 |
- |
- por %xmm2,%xmm0 |
- leaq 256(%r12),%r12 |
- por %xmm3,%xmm0 |
- |
- mulq %rbp |
- addq %rax,%r10 |
- movq 8(%rsi),%rax |
- adcq $0,%rdx |
- movq %rdx,%rdi |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq 8(%rcx),%rax |
- adcq $0,%rdx |
- movq %rdx,%r10 |
- |
- mulq %rbp |
- addq %rax,%rdi |
- movq 16(%rsi),%rax |
- adcq $0,%rdx |
- addq %r11,%rdi |
- leaq 4(%r15),%r15 |
- adcq $0,%rdx |
- movq %rdi,(%rsp) |
- movq %rdx,%r13 |
- jmp .L1st4x |
-.align 16 |
-.L1st4x: |
- mulq %rbx |
- addq %rax,%r10 |
- movq -16(%rcx,%r15,8),%rax |
- adcq $0,%rdx |
- movq %rdx,%r11 |
- |
- mulq %rbp |
- addq %rax,%r13 |
- movq -8(%rsi,%r15,8),%rax |
- adcq $0,%rdx |
- addq %r10,%r13 |
- adcq $0,%rdx |
- movq %r13,-24(%rsp,%r15,8) |
- movq %rdx,%rdi |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq -8(%rcx,%r15,8),%rax |
- adcq $0,%rdx |
- movq %rdx,%r10 |
- |
- mulq %rbp |
- addq %rax,%rdi |
- movq (%rsi,%r15,8),%rax |
- adcq $0,%rdx |
- addq %r11,%rdi |
- adcq $0,%rdx |
- movq %rdi,-16(%rsp,%r15,8) |
- movq %rdx,%r13 |
- |
- mulq %rbx |
- addq %rax,%r10 |
- movq (%rcx,%r15,8),%rax |
- adcq $0,%rdx |
- movq %rdx,%r11 |
- |
- mulq %rbp |
- addq %rax,%r13 |
- movq 8(%rsi,%r15,8),%rax |
- adcq $0,%rdx |
- addq %r10,%r13 |
- adcq $0,%rdx |
- movq %r13,-8(%rsp,%r15,8) |
- movq %rdx,%rdi |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq 8(%rcx,%r15,8),%rax |
- adcq $0,%rdx |
- leaq 4(%r15),%r15 |
- movq %rdx,%r10 |
- |
- mulq %rbp |
- addq %rax,%rdi |
- movq -16(%rsi,%r15,8),%rax |
- adcq $0,%rdx |
- addq %r11,%rdi |
- adcq $0,%rdx |
- movq %rdi,-32(%rsp,%r15,8) |
- movq %rdx,%r13 |
- cmpq %r9,%r15 |
- jl .L1st4x |
- |
- mulq %rbx |
- addq %rax,%r10 |
- movq -16(%rcx,%r15,8),%rax |
- adcq $0,%rdx |
- movq %rdx,%r11 |
- |
- mulq %rbp |
- addq %rax,%r13 |
- movq -8(%rsi,%r15,8),%rax |
- adcq $0,%rdx |
- addq %r10,%r13 |
- adcq $0,%rdx |
- movq %r13,-24(%rsp,%r15,8) |
- movq %rdx,%rdi |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq -8(%rcx,%r15,8),%rax |
- adcq $0,%rdx |
- movq %rdx,%r10 |
- |
- mulq %rbp |
- addq %rax,%rdi |
- movq (%rsi),%rax |
- adcq $0,%rdx |
- addq %r11,%rdi |
- adcq $0,%rdx |
- movq %rdi,-16(%rsp,%r15,8) |
- movq %rdx,%r13 |
- |
-.byte 102,72,15,126,195 |
- |
- xorq %rdi,%rdi |
- addq %r10,%r13 |
- adcq $0,%rdi |
- movq %r13,-8(%rsp,%r15,8) |
- movq %rdi,(%rsp,%r15,8) |
- |
- leaq 1(%r14),%r14 |
-.align 4 |
-.Louter4x: |
- xorq %r15,%r15 |
- movq -96(%r12),%xmm0 |
- movq -32(%r12),%xmm1 |
- pand %xmm4,%xmm0 |
- movq 32(%r12),%xmm2 |
- pand %xmm5,%xmm1 |
- |
- movq (%rsp),%r10 |
- movq %r8,%rbp |
- mulq %rbx |
- addq %rax,%r10 |
- movq (%rcx),%rax |
- adcq $0,%rdx |
- |
- movq 96(%r12),%xmm3 |
- pand %xmm6,%xmm2 |
- por %xmm1,%xmm0 |
- pand %xmm7,%xmm3 |
- |
- imulq %r10,%rbp |
- movq %rdx,%r11 |
- |
- por %xmm2,%xmm0 |
- leaq 256(%r12),%r12 |
- por %xmm3,%xmm0 |
- |
- mulq %rbp |
- addq %rax,%r10 |
- movq 8(%rsi),%rax |
- adcq $0,%rdx |
- movq %rdx,%rdi |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq 8(%rcx),%rax |
- adcq $0,%rdx |
- addq 8(%rsp),%r11 |
- adcq $0,%rdx |
- movq %rdx,%r10 |
- |
- mulq %rbp |
- addq %rax,%rdi |
- movq 16(%rsi),%rax |
- adcq $0,%rdx |
- addq %r11,%rdi |
- leaq 4(%r15),%r15 |
- adcq $0,%rdx |
- movq %rdx,%r13 |
- jmp .Linner4x |
-.align 16 |
-.Linner4x: |
- mulq %rbx |
- addq %rax,%r10 |
- movq -16(%rcx,%r15,8),%rax |
- adcq $0,%rdx |
- addq -16(%rsp,%r15,8),%r10 |
- adcq $0,%rdx |
- movq %rdx,%r11 |
- |
- mulq %rbp |
- addq %rax,%r13 |
- movq -8(%rsi,%r15,8),%rax |
- adcq $0,%rdx |
- addq %r10,%r13 |
- adcq $0,%rdx |
- movq %rdi,-32(%rsp,%r15,8) |
- movq %rdx,%rdi |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq -8(%rcx,%r15,8),%rax |
- adcq $0,%rdx |
- addq -8(%rsp,%r15,8),%r11 |
- adcq $0,%rdx |
- movq %rdx,%r10 |
- |
- mulq %rbp |
- addq %rax,%rdi |
- movq (%rsi,%r15,8),%rax |
- adcq $0,%rdx |
- addq %r11,%rdi |
- adcq $0,%rdx |
- movq %r13,-24(%rsp,%r15,8) |
- movq %rdx,%r13 |
- |
- mulq %rbx |
- addq %rax,%r10 |
- movq (%rcx,%r15,8),%rax |
- adcq $0,%rdx |
- addq (%rsp,%r15,8),%r10 |
- adcq $0,%rdx |
- movq %rdx,%r11 |
- |
- mulq %rbp |
- addq %rax,%r13 |
- movq 8(%rsi,%r15,8),%rax |
- adcq $0,%rdx |
- addq %r10,%r13 |
- adcq $0,%rdx |
- movq %rdi,-16(%rsp,%r15,8) |
- movq %rdx,%rdi |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq 8(%rcx,%r15,8),%rax |
- adcq $0,%rdx |
- addq 8(%rsp,%r15,8),%r11 |
- adcq $0,%rdx |
- leaq 4(%r15),%r15 |
- movq %rdx,%r10 |
- |
- mulq %rbp |
- addq %rax,%rdi |
- movq -16(%rsi,%r15,8),%rax |
- adcq $0,%rdx |
- addq %r11,%rdi |
- adcq $0,%rdx |
- movq %r13,-40(%rsp,%r15,8) |
- movq %rdx,%r13 |
- cmpq %r9,%r15 |
- jl .Linner4x |
- |
- mulq %rbx |
- addq %rax,%r10 |
- movq -16(%rcx,%r15,8),%rax |
- adcq $0,%rdx |
- addq -16(%rsp,%r15,8),%r10 |
- adcq $0,%rdx |
- movq %rdx,%r11 |
- |
- mulq %rbp |
- addq %rax,%r13 |
- movq -8(%rsi,%r15,8),%rax |
- adcq $0,%rdx |
- addq %r10,%r13 |
- adcq $0,%rdx |
- movq %rdi,-32(%rsp,%r15,8) |
- movq %rdx,%rdi |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq -8(%rcx,%r15,8),%rax |
- adcq $0,%rdx |
- addq -8(%rsp,%r15,8),%r11 |
- adcq $0,%rdx |
- leaq 1(%r14),%r14 |
- movq %rdx,%r10 |
- |
- mulq %rbp |
- addq %rax,%rdi |
- movq (%rsi),%rax |
- adcq $0,%rdx |
- addq %r11,%rdi |
- adcq $0,%rdx |
- movq %r13,-24(%rsp,%r15,8) |
- movq %rdx,%r13 |
- |
-.byte 102,72,15,126,195 |
- movq %rdi,-16(%rsp,%r15,8) |
- |
- xorq %rdi,%rdi |
- addq %r10,%r13 |
- adcq $0,%rdi |
- addq (%rsp,%r9,8),%r13 |
- adcq $0,%rdi |
- movq %r13,-8(%rsp,%r15,8) |
- movq %rdi,(%rsp,%r15,8) |
- |
- cmpq %r9,%r14 |
- jl .Louter4x |
- movq 16(%rsp,%r9,8),%rdi |
- movq 0(%rsp),%rax |
- pxor %xmm0,%xmm0 |
- movq 8(%rsp),%rdx |
- shrq $2,%r9 |
- leaq (%rsp),%rsi |
- xorq %r14,%r14 |
- |
- subq 0(%rcx),%rax |
- movq 16(%rsi),%rbx |
- movq 24(%rsi),%rbp |
- sbbq 8(%rcx),%rdx |
- leaq -1(%r9),%r15 |
- jmp .Lsub4x |
-.align 16 |
-.Lsub4x: |
- movq %rax,0(%rdi,%r14,8) |
- movq %rdx,8(%rdi,%r14,8) |
- sbbq 16(%rcx,%r14,8),%rbx |
- movq 32(%rsi,%r14,8),%rax |
- movq 40(%rsi,%r14,8),%rdx |
- sbbq 24(%rcx,%r14,8),%rbp |
- movq %rbx,16(%rdi,%r14,8) |
- movq %rbp,24(%rdi,%r14,8) |
- sbbq 32(%rcx,%r14,8),%rax |
- movq 48(%rsi,%r14,8),%rbx |
- movq 56(%rsi,%r14,8),%rbp |
- sbbq 40(%rcx,%r14,8),%rdx |
- leaq 4(%r14),%r14 |
- decq %r15 |
- jnz .Lsub4x |
- |
- movq %rax,0(%rdi,%r14,8) |
- movq 32(%rsi,%r14,8),%rax |
- sbbq 16(%rcx,%r14,8),%rbx |
- movq %rdx,8(%rdi,%r14,8) |
- sbbq 24(%rcx,%r14,8),%rbp |
- movq %rbx,16(%rdi,%r14,8) |
- |
- sbbq $0,%rax |
- movq %rbp,24(%rdi,%r14,8) |
- xorq %r14,%r14 |
- andq %rax,%rsi |
- notq %rax |
- movq %rdi,%rcx |
- andq %rax,%rcx |
- leaq -1(%r9),%r15 |
- orq %rcx,%rsi |
- |
- movdqu (%rsi),%xmm1 |
- movdqa %xmm0,(%rsp) |
- movdqu %xmm1,(%rdi) |
- jmp .Lcopy4x |
-.align 16 |
-.Lcopy4x: |
- movdqu 16(%rsi,%r14,1),%xmm2 |
- movdqu 32(%rsi,%r14,1),%xmm1 |
- movdqa %xmm0,16(%rsp,%r14,1) |
- movdqu %xmm2,16(%rdi,%r14,1) |
- movdqa %xmm0,32(%rsp,%r14,1) |
- movdqu %xmm1,32(%rdi,%r14,1) |
- leaq 32(%r14),%r14 |
- decq %r15 |
- jnz .Lcopy4x |
- |
- shlq $2,%r9 |
- movdqu 16(%rsi,%r14,1),%xmm2 |
- movdqa %xmm0,16(%rsp,%r14,1) |
- movdqu %xmm2,16(%rdi,%r14,1) |
- movq 8(%rsp,%r9,8),%rsi |
- movq $1,%rax |
- movq (%rsi),%r15 |
- movq 8(%rsi),%r14 |
- movq 16(%rsi),%r13 |
- movq 24(%rsi),%r12 |
- movq 32(%rsi),%rbp |
- movq 40(%rsi),%rbx |
- leaq 48(%rsi),%rsp |
-.Lmul4x_epilogue: |
- .byte 0xf3,0xc3 |
-.size bn_mul4x_mont_gather5,.-bn_mul4x_mont_gather5 |
-.globl bn_scatter5 |
-.type bn_scatter5,@function |
-.align 16 |
-bn_scatter5: |
- cmpq $0,%rsi |
- jz .Lscatter_epilogue |
- leaq (%rdx,%rcx,8),%rdx |
-.Lscatter: |
- movq (%rdi),%rax |
- leaq 8(%rdi),%rdi |
- movq %rax,(%rdx) |
- leaq 256(%rdx),%rdx |
- subq $1,%rsi |
- jnz .Lscatter |
-.Lscatter_epilogue: |
- .byte 0xf3,0xc3 |
-.size bn_scatter5,.-bn_scatter5 |
- |
-.globl bn_gather5 |
-.type bn_gather5,@function |
-.align 16 |
-bn_gather5: |
- movq %rcx,%r11 |
- shrq $3,%rcx |
- andq $7,%r11 |
- notq %rcx |
- leaq .Lmagic_masks(%rip),%rax |
- andq $3,%rcx |
- leaq 96(%rdx,%r11,8),%rdx |
- movq 0(%rax,%rcx,8),%xmm4 |
- movq 8(%rax,%rcx,8),%xmm5 |
- movq 16(%rax,%rcx,8),%xmm6 |
- movq 24(%rax,%rcx,8),%xmm7 |
- jmp .Lgather |
-.align 16 |
-.Lgather: |
- movq -96(%rdx),%xmm0 |
- movq -32(%rdx),%xmm1 |
- pand %xmm4,%xmm0 |
- movq 32(%rdx),%xmm2 |
- pand %xmm5,%xmm1 |
- movq 96(%rdx),%xmm3 |
- pand %xmm6,%xmm2 |
- por %xmm1,%xmm0 |
- pand %xmm7,%xmm3 |
- por %xmm2,%xmm0 |
- leaq 256(%rdx),%rdx |
- por %xmm3,%xmm0 |
- |
- movq %xmm0,(%rdi) |
- leaq 8(%rdi),%rdi |
- subq $1,%rsi |
- jnz .Lgather |
- .byte 0xf3,0xc3 |
-.LSEH_end_bn_gather5: |
-.size bn_gather5,.-bn_gather5 |
-.align 64 |
-.Lmagic_masks: |
-.long 0,0, 0,0, 0,0, -1,-1 |
-.long 0,0, 0,0, 0,0, 0,0 |
-.byte 77,111,110,116,103,111,109,101,114,121,32,77,117,108,116,105,112,108,105,99,97,116,105,111,110,32,119,105,116,104,32,115,99,97,116,116,101,114,47,103,97,116,104,101,114,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,103,62,0 |