| Index: openssl/crypto/bn/asm/x86_64-mont.S
|
| diff --git a/openssl/crypto/bn/asm/x86_64-mont.S b/openssl/crypto/bn/asm/x86_64-mont.S
|
| deleted file mode 100644
|
| index 95e29052873c014726849ac3cdf6c13068b15b4c..0000000000000000000000000000000000000000
|
| --- a/openssl/crypto/bn/asm/x86_64-mont.S
|
| +++ /dev/null
|
| @@ -1,1374 +0,0 @@
|
| -.text
|
| -
|
| -.globl bn_mul_mont
|
| -.type bn_mul_mont,@function
|
| -.align 16
|
| -bn_mul_mont:
|
| - testl $3,%r9d
|
| - jnz .Lmul_enter
|
| - cmpl $8,%r9d
|
| - jb .Lmul_enter
|
| - cmpq %rsi,%rdx
|
| - jne .Lmul4x_enter
|
| - jmp .Lsqr4x_enter
|
| -
|
| -.align 16
|
| -.Lmul_enter:
|
| - pushq %rbx
|
| - pushq %rbp
|
| - pushq %r12
|
| - pushq %r13
|
| - pushq %r14
|
| - pushq %r15
|
| -
|
| - movl %r9d,%r9d
|
| - leaq 2(%r9),%r10
|
| - movq %rsp,%r11
|
| - negq %r10
|
| - leaq (%rsp,%r10,8),%rsp
|
| - andq $-1024,%rsp
|
| -
|
| - movq %r11,8(%rsp,%r9,8)
|
| -.Lmul_body:
|
| - movq %rdx,%r12
|
| - movq (%r8),%r8
|
| - movq (%r12),%rbx
|
| - movq (%rsi),%rax
|
| -
|
| - xorq %r14,%r14
|
| - xorq %r15,%r15
|
| -
|
| - movq %r8,%rbp
|
| - mulq %rbx
|
| - movq %rax,%r10
|
| - movq (%rcx),%rax
|
| -
|
| - imulq %r10,%rbp
|
| - movq %rdx,%r11
|
| -
|
| - mulq %rbp
|
| - addq %rax,%r10
|
| - movq 8(%rsi),%rax
|
| - adcq $0,%rdx
|
| - movq %rdx,%r13
|
| -
|
| - leaq 1(%r15),%r15
|
| - jmp .L1st_enter
|
| -
|
| -.align 16
|
| -.L1st:
|
| - addq %rax,%r13
|
| - movq (%rsi,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - addq %r11,%r13
|
| - movq %r10,%r11
|
| - adcq $0,%rdx
|
| - movq %r13,-16(%rsp,%r15,8)
|
| - movq %rdx,%r13
|
| -
|
| -.L1st_enter:
|
| - mulq %rbx
|
| - addq %rax,%r11
|
| - movq (%rcx,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - leaq 1(%r15),%r15
|
| - movq %rdx,%r10
|
| -
|
| - mulq %rbp
|
| - cmpq %r9,%r15
|
| - jne .L1st
|
| -
|
| - addq %rax,%r13
|
| - movq (%rsi),%rax
|
| - adcq $0,%rdx
|
| - addq %r11,%r13
|
| - adcq $0,%rdx
|
| - movq %r13,-16(%rsp,%r15,8)
|
| - movq %rdx,%r13
|
| - movq %r10,%r11
|
| -
|
| - xorq %rdx,%rdx
|
| - addq %r11,%r13
|
| - adcq $0,%rdx
|
| - movq %r13,-8(%rsp,%r9,8)
|
| - movq %rdx,(%rsp,%r9,8)
|
| -
|
| - leaq 1(%r14),%r14
|
| - jmp .Louter
|
| -.align 16
|
| -.Louter:
|
| - movq (%r12,%r14,8),%rbx
|
| - xorq %r15,%r15
|
| - movq %r8,%rbp
|
| - movq (%rsp),%r10
|
| - mulq %rbx
|
| - addq %rax,%r10
|
| - movq (%rcx),%rax
|
| - adcq $0,%rdx
|
| -
|
| - imulq %r10,%rbp
|
| - movq %rdx,%r11
|
| -
|
| - mulq %rbp
|
| - addq %rax,%r10
|
| - movq 8(%rsi),%rax
|
| - adcq $0,%rdx
|
| - movq 8(%rsp),%r10
|
| - movq %rdx,%r13
|
| -
|
| - leaq 1(%r15),%r15
|
| - jmp .Linner_enter
|
| -
|
| -.align 16
|
| -.Linner:
|
| - addq %rax,%r13
|
| - movq (%rsi,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - addq %r10,%r13
|
| - movq (%rsp,%r15,8),%r10
|
| - adcq $0,%rdx
|
| - movq %r13,-16(%rsp,%r15,8)
|
| - movq %rdx,%r13
|
| -
|
| -.Linner_enter:
|
| - mulq %rbx
|
| - addq %rax,%r11
|
| - movq (%rcx,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - addq %r11,%r10
|
| - movq %rdx,%r11
|
| - adcq $0,%r11
|
| - leaq 1(%r15),%r15
|
| -
|
| - mulq %rbp
|
| - cmpq %r9,%r15
|
| - jne .Linner
|
| -
|
| - addq %rax,%r13
|
| - movq (%rsi),%rax
|
| - adcq $0,%rdx
|
| - addq %r10,%r13
|
| - movq (%rsp,%r15,8),%r10
|
| - adcq $0,%rdx
|
| - movq %r13,-16(%rsp,%r15,8)
|
| - movq %rdx,%r13
|
| -
|
| - xorq %rdx,%rdx
|
| - addq %r11,%r13
|
| - adcq $0,%rdx
|
| - addq %r10,%r13
|
| - adcq $0,%rdx
|
| - movq %r13,-8(%rsp,%r9,8)
|
| - movq %rdx,(%rsp,%r9,8)
|
| -
|
| - leaq 1(%r14),%r14
|
| - cmpq %r9,%r14
|
| - jl .Louter
|
| -
|
| - xorq %r14,%r14
|
| - movq (%rsp),%rax
|
| - leaq (%rsp),%rsi
|
| - movq %r9,%r15
|
| - jmp .Lsub
|
| -.align 16
|
| -.Lsub: sbbq (%rcx,%r14,8),%rax
|
| - movq %rax,(%rdi,%r14,8)
|
| - movq 8(%rsi,%r14,8),%rax
|
| - leaq 1(%r14),%r14
|
| - decq %r15
|
| - jnz .Lsub
|
| -
|
| - sbbq $0,%rax
|
| - xorq %r14,%r14
|
| - andq %rax,%rsi
|
| - notq %rax
|
| - movq %rdi,%rcx
|
| - andq %rax,%rcx
|
| - movq %r9,%r15
|
| - orq %rcx,%rsi
|
| -.align 16
|
| -.Lcopy:
|
| - movq (%rsi,%r14,8),%rax
|
| - movq %r14,(%rsp,%r14,8)
|
| - movq %rax,(%rdi,%r14,8)
|
| - leaq 1(%r14),%r14
|
| - subq $1,%r15
|
| - jnz .Lcopy
|
| -
|
| - movq 8(%rsp,%r9,8),%rsi
|
| - movq $1,%rax
|
| - movq (%rsi),%r15
|
| - movq 8(%rsi),%r14
|
| - movq 16(%rsi),%r13
|
| - movq 24(%rsi),%r12
|
| - movq 32(%rsi),%rbp
|
| - movq 40(%rsi),%rbx
|
| - leaq 48(%rsi),%rsp
|
| -.Lmul_epilogue:
|
| - .byte 0xf3,0xc3
|
| -.size bn_mul_mont,.-bn_mul_mont
|
| -.type bn_mul4x_mont,@function
|
| -.align 16
|
| -bn_mul4x_mont:
|
| -.Lmul4x_enter:
|
| - pushq %rbx
|
| - pushq %rbp
|
| - pushq %r12
|
| - pushq %r13
|
| - pushq %r14
|
| - pushq %r15
|
| -
|
| - movl %r9d,%r9d
|
| - leaq 4(%r9),%r10
|
| - movq %rsp,%r11
|
| - negq %r10
|
| - leaq (%rsp,%r10,8),%rsp
|
| - andq $-1024,%rsp
|
| -
|
| - movq %r11,8(%rsp,%r9,8)
|
| -.Lmul4x_body:
|
| - movq %rdi,16(%rsp,%r9,8)
|
| - movq %rdx,%r12
|
| - movq (%r8),%r8
|
| - movq (%r12),%rbx
|
| - movq (%rsi),%rax
|
| -
|
| - xorq %r14,%r14
|
| - xorq %r15,%r15
|
| -
|
| - movq %r8,%rbp
|
| - mulq %rbx
|
| - movq %rax,%r10
|
| - movq (%rcx),%rax
|
| -
|
| - imulq %r10,%rbp
|
| - movq %rdx,%r11
|
| -
|
| - mulq %rbp
|
| - addq %rax,%r10
|
| - movq 8(%rsi),%rax
|
| - adcq $0,%rdx
|
| - movq %rdx,%rdi
|
| -
|
| - mulq %rbx
|
| - addq %rax,%r11
|
| - movq 8(%rcx),%rax
|
| - adcq $0,%rdx
|
| - movq %rdx,%r10
|
| -
|
| - mulq %rbp
|
| - addq %rax,%rdi
|
| - movq 16(%rsi),%rax
|
| - adcq $0,%rdx
|
| - addq %r11,%rdi
|
| - leaq 4(%r15),%r15
|
| - adcq $0,%rdx
|
| - movq %rdi,(%rsp)
|
| - movq %rdx,%r13
|
| - jmp .L1st4x
|
| -.align 16
|
| -.L1st4x:
|
| - mulq %rbx
|
| - addq %rax,%r10
|
| - movq -16(%rcx,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - movq %rdx,%r11
|
| -
|
| - mulq %rbp
|
| - addq %rax,%r13
|
| - movq -8(%rsi,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - addq %r10,%r13
|
| - adcq $0,%rdx
|
| - movq %r13,-24(%rsp,%r15,8)
|
| - movq %rdx,%rdi
|
| -
|
| - mulq %rbx
|
| - addq %rax,%r11
|
| - movq -8(%rcx,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - movq %rdx,%r10
|
| -
|
| - mulq %rbp
|
| - addq %rax,%rdi
|
| - movq (%rsi,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - addq %r11,%rdi
|
| - adcq $0,%rdx
|
| - movq %rdi,-16(%rsp,%r15,8)
|
| - movq %rdx,%r13
|
| -
|
| - mulq %rbx
|
| - addq %rax,%r10
|
| - movq (%rcx,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - movq %rdx,%r11
|
| -
|
| - mulq %rbp
|
| - addq %rax,%r13
|
| - movq 8(%rsi,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - addq %r10,%r13
|
| - adcq $0,%rdx
|
| - movq %r13,-8(%rsp,%r15,8)
|
| - movq %rdx,%rdi
|
| -
|
| - mulq %rbx
|
| - addq %rax,%r11
|
| - movq 8(%rcx,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - leaq 4(%r15),%r15
|
| - movq %rdx,%r10
|
| -
|
| - mulq %rbp
|
| - addq %rax,%rdi
|
| - movq -16(%rsi,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - addq %r11,%rdi
|
| - adcq $0,%rdx
|
| - movq %rdi,-32(%rsp,%r15,8)
|
| - movq %rdx,%r13
|
| - cmpq %r9,%r15
|
| - jl .L1st4x
|
| -
|
| - mulq %rbx
|
| - addq %rax,%r10
|
| - movq -16(%rcx,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - movq %rdx,%r11
|
| -
|
| - mulq %rbp
|
| - addq %rax,%r13
|
| - movq -8(%rsi,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - addq %r10,%r13
|
| - adcq $0,%rdx
|
| - movq %r13,-24(%rsp,%r15,8)
|
| - movq %rdx,%rdi
|
| -
|
| - mulq %rbx
|
| - addq %rax,%r11
|
| - movq -8(%rcx,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - movq %rdx,%r10
|
| -
|
| - mulq %rbp
|
| - addq %rax,%rdi
|
| - movq (%rsi),%rax
|
| - adcq $0,%rdx
|
| - addq %r11,%rdi
|
| - adcq $0,%rdx
|
| - movq %rdi,-16(%rsp,%r15,8)
|
| - movq %rdx,%r13
|
| -
|
| - xorq %rdi,%rdi
|
| - addq %r10,%r13
|
| - adcq $0,%rdi
|
| - movq %r13,-8(%rsp,%r15,8)
|
| - movq %rdi,(%rsp,%r15,8)
|
| -
|
| - leaq 1(%r14),%r14
|
| -.align 4
|
| -.Louter4x:
|
| - movq (%r12,%r14,8),%rbx
|
| - xorq %r15,%r15
|
| - movq (%rsp),%r10
|
| - movq %r8,%rbp
|
| - mulq %rbx
|
| - addq %rax,%r10
|
| - movq (%rcx),%rax
|
| - adcq $0,%rdx
|
| -
|
| - imulq %r10,%rbp
|
| - movq %rdx,%r11
|
| -
|
| - mulq %rbp
|
| - addq %rax,%r10
|
| - movq 8(%rsi),%rax
|
| - adcq $0,%rdx
|
| - movq %rdx,%rdi
|
| -
|
| - mulq %rbx
|
| - addq %rax,%r11
|
| - movq 8(%rcx),%rax
|
| - adcq $0,%rdx
|
| - addq 8(%rsp),%r11
|
| - adcq $0,%rdx
|
| - movq %rdx,%r10
|
| -
|
| - mulq %rbp
|
| - addq %rax,%rdi
|
| - movq 16(%rsi),%rax
|
| - adcq $0,%rdx
|
| - addq %r11,%rdi
|
| - leaq 4(%r15),%r15
|
| - adcq $0,%rdx
|
| - movq %rdi,(%rsp)
|
| - movq %rdx,%r13
|
| - jmp .Linner4x
|
| -.align 16
|
| -.Linner4x:
|
| - mulq %rbx
|
| - addq %rax,%r10
|
| - movq -16(%rcx,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - addq -16(%rsp,%r15,8),%r10
|
| - adcq $0,%rdx
|
| - movq %rdx,%r11
|
| -
|
| - mulq %rbp
|
| - addq %rax,%r13
|
| - movq -8(%rsi,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - addq %r10,%r13
|
| - adcq $0,%rdx
|
| - movq %r13,-24(%rsp,%r15,8)
|
| - movq %rdx,%rdi
|
| -
|
| - mulq %rbx
|
| - addq %rax,%r11
|
| - movq -8(%rcx,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - addq -8(%rsp,%r15,8),%r11
|
| - adcq $0,%rdx
|
| - movq %rdx,%r10
|
| -
|
| - mulq %rbp
|
| - addq %rax,%rdi
|
| - movq (%rsi,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - addq %r11,%rdi
|
| - adcq $0,%rdx
|
| - movq %rdi,-16(%rsp,%r15,8)
|
| - movq %rdx,%r13
|
| -
|
| - mulq %rbx
|
| - addq %rax,%r10
|
| - movq (%rcx,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - addq (%rsp,%r15,8),%r10
|
| - adcq $0,%rdx
|
| - movq %rdx,%r11
|
| -
|
| - mulq %rbp
|
| - addq %rax,%r13
|
| - movq 8(%rsi,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - addq %r10,%r13
|
| - adcq $0,%rdx
|
| - movq %r13,-8(%rsp,%r15,8)
|
| - movq %rdx,%rdi
|
| -
|
| - mulq %rbx
|
| - addq %rax,%r11
|
| - movq 8(%rcx,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - addq 8(%rsp,%r15,8),%r11
|
| - adcq $0,%rdx
|
| - leaq 4(%r15),%r15
|
| - movq %rdx,%r10
|
| -
|
| - mulq %rbp
|
| - addq %rax,%rdi
|
| - movq -16(%rsi,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - addq %r11,%rdi
|
| - adcq $0,%rdx
|
| - movq %rdi,-32(%rsp,%r15,8)
|
| - movq %rdx,%r13
|
| - cmpq %r9,%r15
|
| - jl .Linner4x
|
| -
|
| - mulq %rbx
|
| - addq %rax,%r10
|
| - movq -16(%rcx,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - addq -16(%rsp,%r15,8),%r10
|
| - adcq $0,%rdx
|
| - movq %rdx,%r11
|
| -
|
| - mulq %rbp
|
| - addq %rax,%r13
|
| - movq -8(%rsi,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - addq %r10,%r13
|
| - adcq $0,%rdx
|
| - movq %r13,-24(%rsp,%r15,8)
|
| - movq %rdx,%rdi
|
| -
|
| - mulq %rbx
|
| - addq %rax,%r11
|
| - movq -8(%rcx,%r15,8),%rax
|
| - adcq $0,%rdx
|
| - addq -8(%rsp,%r15,8),%r11
|
| - adcq $0,%rdx
|
| - leaq 1(%r14),%r14
|
| - movq %rdx,%r10
|
| -
|
| - mulq %rbp
|
| - addq %rax,%rdi
|
| - movq (%rsi),%rax
|
| - adcq $0,%rdx
|
| - addq %r11,%rdi
|
| - adcq $0,%rdx
|
| - movq %rdi,-16(%rsp,%r15,8)
|
| - movq %rdx,%r13
|
| -
|
| - xorq %rdi,%rdi
|
| - addq %r10,%r13
|
| - adcq $0,%rdi
|
| - addq (%rsp,%r9,8),%r13
|
| - adcq $0,%rdi
|
| - movq %r13,-8(%rsp,%r15,8)
|
| - movq %rdi,(%rsp,%r15,8)
|
| -
|
| - cmpq %r9,%r14
|
| - jl .Louter4x
|
| - movq 16(%rsp,%r9,8),%rdi
|
| - movq 0(%rsp),%rax
|
| - pxor %xmm0,%xmm0
|
| - movq 8(%rsp),%rdx
|
| - shrq $2,%r9
|
| - leaq (%rsp),%rsi
|
| - xorq %r14,%r14
|
| -
|
| - subq 0(%rcx),%rax
|
| - movq 16(%rsi),%rbx
|
| - movq 24(%rsi),%rbp
|
| - sbbq 8(%rcx),%rdx
|
| - leaq -1(%r9),%r15
|
| - jmp .Lsub4x
|
| -.align 16
|
| -.Lsub4x:
|
| - movq %rax,0(%rdi,%r14,8)
|
| - movq %rdx,8(%rdi,%r14,8)
|
| - sbbq 16(%rcx,%r14,8),%rbx
|
| - movq 32(%rsi,%r14,8),%rax
|
| - movq 40(%rsi,%r14,8),%rdx
|
| - sbbq 24(%rcx,%r14,8),%rbp
|
| - movq %rbx,16(%rdi,%r14,8)
|
| - movq %rbp,24(%rdi,%r14,8)
|
| - sbbq 32(%rcx,%r14,8),%rax
|
| - movq 48(%rsi,%r14,8),%rbx
|
| - movq 56(%rsi,%r14,8),%rbp
|
| - sbbq 40(%rcx,%r14,8),%rdx
|
| - leaq 4(%r14),%r14
|
| - decq %r15
|
| - jnz .Lsub4x
|
| -
|
| - movq %rax,0(%rdi,%r14,8)
|
| - movq 32(%rsi,%r14,8),%rax
|
| - sbbq 16(%rcx,%r14,8),%rbx
|
| - movq %rdx,8(%rdi,%r14,8)
|
| - sbbq 24(%rcx,%r14,8),%rbp
|
| - movq %rbx,16(%rdi,%r14,8)
|
| -
|
| - sbbq $0,%rax
|
| - movq %rbp,24(%rdi,%r14,8)
|
| - xorq %r14,%r14
|
| - andq %rax,%rsi
|
| - notq %rax
|
| - movq %rdi,%rcx
|
| - andq %rax,%rcx
|
| - leaq -1(%r9),%r15
|
| - orq %rcx,%rsi
|
| -
|
| - movdqu (%rsi),%xmm1
|
| - movdqa %xmm0,(%rsp)
|
| - movdqu %xmm1,(%rdi)
|
| - jmp .Lcopy4x
|
| -.align 16
|
| -.Lcopy4x:
|
| - movdqu 16(%rsi,%r14,1),%xmm2
|
| - movdqu 32(%rsi,%r14,1),%xmm1
|
| - movdqa %xmm0,16(%rsp,%r14,1)
|
| - movdqu %xmm2,16(%rdi,%r14,1)
|
| - movdqa %xmm0,32(%rsp,%r14,1)
|
| - movdqu %xmm1,32(%rdi,%r14,1)
|
| - leaq 32(%r14),%r14
|
| - decq %r15
|
| - jnz .Lcopy4x
|
| -
|
| - shlq $2,%r9
|
| - movdqu 16(%rsi,%r14,1),%xmm2
|
| - movdqa %xmm0,16(%rsp,%r14,1)
|
| - movdqu %xmm2,16(%rdi,%r14,1)
|
| - movq 8(%rsp,%r9,8),%rsi
|
| - movq $1,%rax
|
| - movq (%rsi),%r15
|
| - movq 8(%rsi),%r14
|
| - movq 16(%rsi),%r13
|
| - movq 24(%rsi),%r12
|
| - movq 32(%rsi),%rbp
|
| - movq 40(%rsi),%rbx
|
| - leaq 48(%rsi),%rsp
|
| -.Lmul4x_epilogue:
|
| - .byte 0xf3,0xc3
|
| -.size bn_mul4x_mont,.-bn_mul4x_mont
|
| -.type bn_sqr4x_mont,@function
|
| -.align 16
|
| -bn_sqr4x_mont:
|
| -.Lsqr4x_enter:
|
| - pushq %rbx
|
| - pushq %rbp
|
| - pushq %r12
|
| - pushq %r13
|
| - pushq %r14
|
| - pushq %r15
|
| -
|
| - shll $3,%r9d
|
| - xorq %r10,%r10
|
| - movq %rsp,%r11
|
| - subq %r9,%r10
|
| - movq (%r8),%r8
|
| - leaq -72(%rsp,%r10,2),%rsp
|
| - andq $-1024,%rsp
|
| -
|
| -
|
| -
|
| -
|
| -
|
| -
|
| -
|
| -
|
| -
|
| -
|
| -
|
| - movq %rdi,32(%rsp)
|
| - movq %rcx,40(%rsp)
|
| - movq %r8,48(%rsp)
|
| - movq %r11,56(%rsp)
|
| -.Lsqr4x_body:
|
| -
|
| -
|
| -
|
| -
|
| -
|
| -
|
| -
|
| - leaq 32(%r10),%rbp
|
| - leaq (%rsi,%r9,1),%rsi
|
| -
|
| - movq %r9,%rcx
|
| -
|
| -
|
| - movq -32(%rsi,%rbp,1),%r14
|
| - leaq 64(%rsp,%r9,2),%rdi
|
| - movq -24(%rsi,%rbp,1),%rax
|
| - leaq -32(%rdi,%rbp,1),%rdi
|
| - movq -16(%rsi,%rbp,1),%rbx
|
| - movq %rax,%r15
|
| -
|
| - mulq %r14
|
| - movq %rax,%r10
|
| - movq %rbx,%rax
|
| - movq %rdx,%r11
|
| - movq %r10,-24(%rdi,%rbp,1)
|
| -
|
| - xorq %r10,%r10
|
| - mulq %r14
|
| - addq %rax,%r11
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r10
|
| - movq %r11,-16(%rdi,%rbp,1)
|
| -
|
| - leaq -16(%rbp),%rcx
|
| -
|
| -
|
| - movq 8(%rsi,%rcx,1),%rbx
|
| - mulq %r15
|
| - movq %rax,%r12
|
| - movq %rbx,%rax
|
| - movq %rdx,%r13
|
| -
|
| - xorq %r11,%r11
|
| - addq %r12,%r10
|
| - leaq 16(%rcx),%rcx
|
| - adcq $0,%r11
|
| - mulq %r14
|
| - addq %rax,%r10
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r11
|
| - movq %r10,-8(%rdi,%rcx,1)
|
| - jmp .Lsqr4x_1st
|
| -
|
| -.align 16
|
| -.Lsqr4x_1st:
|
| - movq (%rsi,%rcx,1),%rbx
|
| - xorq %r12,%r12
|
| - mulq %r15
|
| - addq %rax,%r13
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r12
|
| -
|
| - xorq %r10,%r10
|
| - addq %r13,%r11
|
| - adcq $0,%r10
|
| - mulq %r14
|
| - addq %rax,%r11
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r10
|
| - movq %r11,(%rdi,%rcx,1)
|
| -
|
| -
|
| - movq 8(%rsi,%rcx,1),%rbx
|
| - xorq %r13,%r13
|
| - mulq %r15
|
| - addq %rax,%r12
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r13
|
| -
|
| - xorq %r11,%r11
|
| - addq %r12,%r10
|
| - adcq $0,%r11
|
| - mulq %r14
|
| - addq %rax,%r10
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r11
|
| - movq %r10,8(%rdi,%rcx,1)
|
| -
|
| - movq 16(%rsi,%rcx,1),%rbx
|
| - xorq %r12,%r12
|
| - mulq %r15
|
| - addq %rax,%r13
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r12
|
| -
|
| - xorq %r10,%r10
|
| - addq %r13,%r11
|
| - adcq $0,%r10
|
| - mulq %r14
|
| - addq %rax,%r11
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r10
|
| - movq %r11,16(%rdi,%rcx,1)
|
| -
|
| -
|
| - movq 24(%rsi,%rcx,1),%rbx
|
| - xorq %r13,%r13
|
| - mulq %r15
|
| - addq %rax,%r12
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r13
|
| -
|
| - xorq %r11,%r11
|
| - addq %r12,%r10
|
| - leaq 32(%rcx),%rcx
|
| - adcq $0,%r11
|
| - mulq %r14
|
| - addq %rax,%r10
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r11
|
| - movq %r10,-8(%rdi,%rcx,1)
|
| -
|
| - cmpq $0,%rcx
|
| - jne .Lsqr4x_1st
|
| -
|
| - xorq %r12,%r12
|
| - addq %r11,%r13
|
| - adcq $0,%r12
|
| - mulq %r15
|
| - addq %rax,%r13
|
| - adcq %rdx,%r12
|
| -
|
| - movq %r13,(%rdi)
|
| - leaq 16(%rbp),%rbp
|
| - movq %r12,8(%rdi)
|
| - jmp .Lsqr4x_outer
|
| -
|
| -.align 16
|
| -.Lsqr4x_outer:
|
| - movq -32(%rsi,%rbp,1),%r14
|
| - leaq 64(%rsp,%r9,2),%rdi
|
| - movq -24(%rsi,%rbp,1),%rax
|
| - leaq -32(%rdi,%rbp,1),%rdi
|
| - movq -16(%rsi,%rbp,1),%rbx
|
| - movq %rax,%r15
|
| -
|
| - movq -24(%rdi,%rbp,1),%r10
|
| - xorq %r11,%r11
|
| - mulq %r14
|
| - addq %rax,%r10
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r11
|
| - movq %r10,-24(%rdi,%rbp,1)
|
| -
|
| - xorq %r10,%r10
|
| - addq -16(%rdi,%rbp,1),%r11
|
| - adcq $0,%r10
|
| - mulq %r14
|
| - addq %rax,%r11
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r10
|
| - movq %r11,-16(%rdi,%rbp,1)
|
| -
|
| - leaq -16(%rbp),%rcx
|
| - xorq %r12,%r12
|
| -
|
| -
|
| - movq 8(%rsi,%rcx,1),%rbx
|
| - xorq %r13,%r13
|
| - addq 8(%rdi,%rcx,1),%r12
|
| - adcq $0,%r13
|
| - mulq %r15
|
| - addq %rax,%r12
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r13
|
| -
|
| - xorq %r11,%r11
|
| - addq %r12,%r10
|
| - adcq $0,%r11
|
| - mulq %r14
|
| - addq %rax,%r10
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r11
|
| - movq %r10,8(%rdi,%rcx,1)
|
| -
|
| - leaq 16(%rcx),%rcx
|
| - jmp .Lsqr4x_inner
|
| -
|
| -.align 16
|
| -.Lsqr4x_inner:
|
| - movq (%rsi,%rcx,1),%rbx
|
| - xorq %r12,%r12
|
| - addq (%rdi,%rcx,1),%r13
|
| - adcq $0,%r12
|
| - mulq %r15
|
| - addq %rax,%r13
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r12
|
| -
|
| - xorq %r10,%r10
|
| - addq %r13,%r11
|
| - adcq $0,%r10
|
| - mulq %r14
|
| - addq %rax,%r11
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r10
|
| - movq %r11,(%rdi,%rcx,1)
|
| -
|
| - movq 8(%rsi,%rcx,1),%rbx
|
| - xorq %r13,%r13
|
| - addq 8(%rdi,%rcx,1),%r12
|
| - adcq $0,%r13
|
| - mulq %r15
|
| - addq %rax,%r12
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r13
|
| -
|
| - xorq %r11,%r11
|
| - addq %r12,%r10
|
| - leaq 16(%rcx),%rcx
|
| - adcq $0,%r11
|
| - mulq %r14
|
| - addq %rax,%r10
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r11
|
| - movq %r10,-8(%rdi,%rcx,1)
|
| -
|
| - cmpq $0,%rcx
|
| - jne .Lsqr4x_inner
|
| -
|
| - xorq %r12,%r12
|
| - addq %r11,%r13
|
| - adcq $0,%r12
|
| - mulq %r15
|
| - addq %rax,%r13
|
| - adcq %rdx,%r12
|
| -
|
| - movq %r13,(%rdi)
|
| - movq %r12,8(%rdi)
|
| -
|
| - addq $16,%rbp
|
| - jnz .Lsqr4x_outer
|
| -
|
| -
|
| - movq -32(%rsi),%r14
|
| - leaq 64(%rsp,%r9,2),%rdi
|
| - movq -24(%rsi),%rax
|
| - leaq -32(%rdi,%rbp,1),%rdi
|
| - movq -16(%rsi),%rbx
|
| - movq %rax,%r15
|
| -
|
| - xorq %r11,%r11
|
| - mulq %r14
|
| - addq %rax,%r10
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r11
|
| - movq %r10,-24(%rdi)
|
| -
|
| - xorq %r10,%r10
|
| - addq %r13,%r11
|
| - adcq $0,%r10
|
| - mulq %r14
|
| - addq %rax,%r11
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r10
|
| - movq %r11,-16(%rdi)
|
| -
|
| - movq -8(%rsi),%rbx
|
| - mulq %r15
|
| - addq %rax,%r12
|
| - movq %rbx,%rax
|
| - adcq $0,%rdx
|
| -
|
| - xorq %r11,%r11
|
| - addq %r12,%r10
|
| - movq %rdx,%r13
|
| - adcq $0,%r11
|
| - mulq %r14
|
| - addq %rax,%r10
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r11
|
| - movq %r10,-8(%rdi)
|
| -
|
| - xorq %r12,%r12
|
| - addq %r11,%r13
|
| - adcq $0,%r12
|
| - mulq %r15
|
| - addq %rax,%r13
|
| - movq -16(%rsi),%rax
|
| - adcq %rdx,%r12
|
| -
|
| - movq %r13,(%rdi)
|
| - movq %r12,8(%rdi)
|
| -
|
| - mulq %rbx
|
| - addq $16,%rbp
|
| - xorq %r14,%r14
|
| - subq %r9,%rbp
|
| - xorq %r15,%r15
|
| -
|
| - addq %r12,%rax
|
| - adcq $0,%rdx
|
| - movq %rax,8(%rdi)
|
| - movq %rdx,16(%rdi)
|
| - movq %r15,24(%rdi)
|
| -
|
| - movq -16(%rsi,%rbp,1),%rax
|
| - leaq 64(%rsp,%r9,2),%rdi
|
| - xorq %r10,%r10
|
| - movq -24(%rdi,%rbp,2),%r11
|
| -
|
| - leaq (%r14,%r10,2),%r12
|
| - shrq $63,%r10
|
| - leaq (%rcx,%r11,2),%r13
|
| - shrq $63,%r11
|
| - orq %r10,%r13
|
| - movq -16(%rdi,%rbp,2),%r10
|
| - movq %r11,%r14
|
| - mulq %rax
|
| - negq %r15
|
| - movq -8(%rdi,%rbp,2),%r11
|
| - adcq %rax,%r12
|
| - movq -8(%rsi,%rbp,1),%rax
|
| - movq %r12,-32(%rdi,%rbp,2)
|
| - adcq %rdx,%r13
|
| -
|
| - leaq (%r14,%r10,2),%rbx
|
| - movq %r13,-24(%rdi,%rbp,2)
|
| - sbbq %r15,%r15
|
| - shrq $63,%r10
|
| - leaq (%rcx,%r11,2),%r8
|
| - shrq $63,%r11
|
| - orq %r10,%r8
|
| - movq 0(%rdi,%rbp,2),%r10
|
| - movq %r11,%r14
|
| - mulq %rax
|
| - negq %r15
|
| - movq 8(%rdi,%rbp,2),%r11
|
| - adcq %rax,%rbx
|
| - movq 0(%rsi,%rbp,1),%rax
|
| - movq %rbx,-16(%rdi,%rbp,2)
|
| - adcq %rdx,%r8
|
| - leaq 16(%rbp),%rbp
|
| - movq %r8,-40(%rdi,%rbp,2)
|
| - sbbq %r15,%r15
|
| - jmp .Lsqr4x_shift_n_add
|
| -
|
| -.align 16
|
| -.Lsqr4x_shift_n_add:
|
| - leaq (%r14,%r10,2),%r12
|
| - shrq $63,%r10
|
| - leaq (%rcx,%r11,2),%r13
|
| - shrq $63,%r11
|
| - orq %r10,%r13
|
| - movq -16(%rdi,%rbp,2),%r10
|
| - movq %r11,%r14
|
| - mulq %rax
|
| - negq %r15
|
| - movq -8(%rdi,%rbp,2),%r11
|
| - adcq %rax,%r12
|
| - movq -8(%rsi,%rbp,1),%rax
|
| - movq %r12,-32(%rdi,%rbp,2)
|
| - adcq %rdx,%r13
|
| -
|
| - leaq (%r14,%r10,2),%rbx
|
| - movq %r13,-24(%rdi,%rbp,2)
|
| - sbbq %r15,%r15
|
| - shrq $63,%r10
|
| - leaq (%rcx,%r11,2),%r8
|
| - shrq $63,%r11
|
| - orq %r10,%r8
|
| - movq 0(%rdi,%rbp,2),%r10
|
| - movq %r11,%r14
|
| - mulq %rax
|
| - negq %r15
|
| - movq 8(%rdi,%rbp,2),%r11
|
| - adcq %rax,%rbx
|
| - movq 0(%rsi,%rbp,1),%rax
|
| - movq %rbx,-16(%rdi,%rbp,2)
|
| - adcq %rdx,%r8
|
| -
|
| - leaq (%r14,%r10,2),%r12
|
| - movq %r8,-8(%rdi,%rbp,2)
|
| - sbbq %r15,%r15
|
| - shrq $63,%r10
|
| - leaq (%rcx,%r11,2),%r13
|
| - shrq $63,%r11
|
| - orq %r10,%r13
|
| - movq 16(%rdi,%rbp,2),%r10
|
| - movq %r11,%r14
|
| - mulq %rax
|
| - negq %r15
|
| - movq 24(%rdi,%rbp,2),%r11
|
| - adcq %rax,%r12
|
| - movq 8(%rsi,%rbp,1),%rax
|
| - movq %r12,0(%rdi,%rbp,2)
|
| - adcq %rdx,%r13
|
| -
|
| - leaq (%r14,%r10,2),%rbx
|
| - movq %r13,8(%rdi,%rbp,2)
|
| - sbbq %r15,%r15
|
| - shrq $63,%r10
|
| - leaq (%rcx,%r11,2),%r8
|
| - shrq $63,%r11
|
| - orq %r10,%r8
|
| - movq 32(%rdi,%rbp,2),%r10
|
| - movq %r11,%r14
|
| - mulq %rax
|
| - negq %r15
|
| - movq 40(%rdi,%rbp,2),%r11
|
| - adcq %rax,%rbx
|
| - movq 16(%rsi,%rbp,1),%rax
|
| - movq %rbx,16(%rdi,%rbp,2)
|
| - adcq %rdx,%r8
|
| - movq %r8,24(%rdi,%rbp,2)
|
| - sbbq %r15,%r15
|
| - addq $32,%rbp
|
| - jnz .Lsqr4x_shift_n_add
|
| -
|
| - leaq (%r14,%r10,2),%r12
|
| - shrq $63,%r10
|
| - leaq (%rcx,%r11,2),%r13
|
| - shrq $63,%r11
|
| - orq %r10,%r13
|
| - movq -16(%rdi),%r10
|
| - movq %r11,%r14
|
| - mulq %rax
|
| - negq %r15
|
| - movq -8(%rdi),%r11
|
| - adcq %rax,%r12
|
| - movq -8(%rsi),%rax
|
| - movq %r12,-32(%rdi)
|
| - adcq %rdx,%r13
|
| -
|
| - leaq (%r14,%r10,2),%rbx
|
| - movq %r13,-24(%rdi)
|
| - sbbq %r15,%r15
|
| - shrq $63,%r10
|
| - leaq (%rcx,%r11,2),%r8
|
| - shrq $63,%r11
|
| - orq %r10,%r8
|
| - mulq %rax
|
| - negq %r15
|
| - adcq %rax,%rbx
|
| - adcq %rdx,%r8
|
| - movq %rbx,-16(%rdi)
|
| - movq %r8,-8(%rdi)
|
| - movq 40(%rsp),%rsi
|
| - movq 48(%rsp),%r8
|
| - xorq %rcx,%rcx
|
| - movq %r9,0(%rsp)
|
| - subq %r9,%rcx
|
| - movq 64(%rsp),%r10
|
| - movq %r8,%r14
|
| - leaq 64(%rsp,%r9,2),%rax
|
| - leaq 64(%rsp,%r9,1),%rdi
|
| - movq %rax,8(%rsp)
|
| - leaq (%rsi,%r9,1),%rsi
|
| - xorq %rbp,%rbp
|
| -
|
| - movq 0(%rsi,%rcx,1),%rax
|
| - movq 8(%rsi,%rcx,1),%r9
|
| - imulq %r10,%r14
|
| - movq %rax,%rbx
|
| - jmp .Lsqr4x_mont_outer
|
| -
|
| -.align 16
|
| -.Lsqr4x_mont_outer:
|
| - xorq %r11,%r11
|
| - mulq %r14
|
| - addq %rax,%r10
|
| - movq %r9,%rax
|
| - adcq %rdx,%r11
|
| - movq %r8,%r15
|
| -
|
| - xorq %r10,%r10
|
| - addq 8(%rdi,%rcx,1),%r11
|
| - adcq $0,%r10
|
| - mulq %r14
|
| - addq %rax,%r11
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r10
|
| -
|
| - imulq %r11,%r15
|
| -
|
| - movq 16(%rsi,%rcx,1),%rbx
|
| - xorq %r13,%r13
|
| - addq %r11,%r12
|
| - adcq $0,%r13
|
| - mulq %r15
|
| - addq %rax,%r12
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r13
|
| - movq %r12,8(%rdi,%rcx,1)
|
| -
|
| - xorq %r11,%r11
|
| - addq 16(%rdi,%rcx,1),%r10
|
| - adcq $0,%r11
|
| - mulq %r14
|
| - addq %rax,%r10
|
| - movq %r9,%rax
|
| - adcq %rdx,%r11
|
| -
|
| - movq 24(%rsi,%rcx,1),%r9
|
| - xorq %r12,%r12
|
| - addq %r10,%r13
|
| - adcq $0,%r12
|
| - mulq %r15
|
| - addq %rax,%r13
|
| - movq %r9,%rax
|
| - adcq %rdx,%r12
|
| - movq %r13,16(%rdi,%rcx,1)
|
| -
|
| - xorq %r10,%r10
|
| - addq 24(%rdi,%rcx,1),%r11
|
| - leaq 32(%rcx),%rcx
|
| - adcq $0,%r10
|
| - mulq %r14
|
| - addq %rax,%r11
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r10
|
| - jmp .Lsqr4x_mont_inner
|
| -
|
| -.align 16
|
| -.Lsqr4x_mont_inner:
|
| - movq (%rsi,%rcx,1),%rbx
|
| - xorq %r13,%r13
|
| - addq %r11,%r12
|
| - adcq $0,%r13
|
| - mulq %r15
|
| - addq %rax,%r12
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r13
|
| - movq %r12,-8(%rdi,%rcx,1)
|
| -
|
| - xorq %r11,%r11
|
| - addq (%rdi,%rcx,1),%r10
|
| - adcq $0,%r11
|
| - mulq %r14
|
| - addq %rax,%r10
|
| - movq %r9,%rax
|
| - adcq %rdx,%r11
|
| -
|
| - movq 8(%rsi,%rcx,1),%r9
|
| - xorq %r12,%r12
|
| - addq %r10,%r13
|
| - adcq $0,%r12
|
| - mulq %r15
|
| - addq %rax,%r13
|
| - movq %r9,%rax
|
| - adcq %rdx,%r12
|
| - movq %r13,(%rdi,%rcx,1)
|
| -
|
| - xorq %r10,%r10
|
| - addq 8(%rdi,%rcx,1),%r11
|
| - adcq $0,%r10
|
| - mulq %r14
|
| - addq %rax,%r11
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r10
|
| -
|
| -
|
| - movq 16(%rsi,%rcx,1),%rbx
|
| - xorq %r13,%r13
|
| - addq %r11,%r12
|
| - adcq $0,%r13
|
| - mulq %r15
|
| - addq %rax,%r12
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r13
|
| - movq %r12,8(%rdi,%rcx,1)
|
| -
|
| - xorq %r11,%r11
|
| - addq 16(%rdi,%rcx,1),%r10
|
| - adcq $0,%r11
|
| - mulq %r14
|
| - addq %rax,%r10
|
| - movq %r9,%rax
|
| - adcq %rdx,%r11
|
| -
|
| - movq 24(%rsi,%rcx,1),%r9
|
| - xorq %r12,%r12
|
| - addq %r10,%r13
|
| - adcq $0,%r12
|
| - mulq %r15
|
| - addq %rax,%r13
|
| - movq %r9,%rax
|
| - adcq %rdx,%r12
|
| - movq %r13,16(%rdi,%rcx,1)
|
| -
|
| - xorq %r10,%r10
|
| - addq 24(%rdi,%rcx,1),%r11
|
| - leaq 32(%rcx),%rcx
|
| - adcq $0,%r10
|
| - mulq %r14
|
| - addq %rax,%r11
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r10
|
| - cmpq $0,%rcx
|
| - jne .Lsqr4x_mont_inner
|
| -
|
| - subq 0(%rsp),%rcx
|
| - movq %r8,%r14
|
| -
|
| - xorq %r13,%r13
|
| - addq %r11,%r12
|
| - adcq $0,%r13
|
| - mulq %r15
|
| - addq %rax,%r12
|
| - movq %r9,%rax
|
| - adcq %rdx,%r13
|
| - movq %r12,-8(%rdi)
|
| -
|
| - xorq %r11,%r11
|
| - addq (%rdi),%r10
|
| - adcq $0,%r11
|
| - movq 0(%rsi,%rcx,1),%rbx
|
| - addq %rbp,%r10
|
| - adcq $0,%r11
|
| -
|
| - imulq 16(%rdi,%rcx,1),%r14
|
| - xorq %r12,%r12
|
| - movq 8(%rsi,%rcx,1),%r9
|
| - addq %r10,%r13
|
| - movq 16(%rdi,%rcx,1),%r10
|
| - adcq $0,%r12
|
| - mulq %r15
|
| - addq %rax,%r13
|
| - movq %rbx,%rax
|
| - adcq %rdx,%r12
|
| - movq %r13,(%rdi)
|
| -
|
| - xorq %rbp,%rbp
|
| - addq 8(%rdi),%r12
|
| - adcq %rbp,%rbp
|
| - addq %r11,%r12
|
| - leaq 16(%rdi),%rdi
|
| - adcq $0,%rbp
|
| - movq %r12,-8(%rdi)
|
| - cmpq 8(%rsp),%rdi
|
| - jb .Lsqr4x_mont_outer
|
| -
|
| - movq 0(%rsp),%r9
|
| - movq %rbp,(%rdi)
|
| - movq 64(%rsp,%r9,1),%rax
|
| - leaq 64(%rsp,%r9,1),%rbx
|
| - movq 40(%rsp),%rsi
|
| - shrq $5,%r9
|
| - movq 8(%rbx),%rdx
|
| - xorq %rbp,%rbp
|
| -
|
| - movq 32(%rsp),%rdi
|
| - subq 0(%rsi),%rax
|
| - movq 16(%rbx),%r10
|
| - movq 24(%rbx),%r11
|
| - sbbq 8(%rsi),%rdx
|
| - leaq -1(%r9),%rcx
|
| - jmp .Lsqr4x_sub
|
| -.align 16
|
| -.Lsqr4x_sub:
|
| - movq %rax,0(%rdi,%rbp,8)
|
| - movq %rdx,8(%rdi,%rbp,8)
|
| - sbbq 16(%rsi,%rbp,8),%r10
|
| - movq 32(%rbx,%rbp,8),%rax
|
| - movq 40(%rbx,%rbp,8),%rdx
|
| - sbbq 24(%rsi,%rbp,8),%r11
|
| - movq %r10,16(%rdi,%rbp,8)
|
| - movq %r11,24(%rdi,%rbp,8)
|
| - sbbq 32(%rsi,%rbp,8),%rax
|
| - movq 48(%rbx,%rbp,8),%r10
|
| - movq 56(%rbx,%rbp,8),%r11
|
| - sbbq 40(%rsi,%rbp,8),%rdx
|
| - leaq 4(%rbp),%rbp
|
| - decq %rcx
|
| - jnz .Lsqr4x_sub
|
| -
|
| - movq %rax,0(%rdi,%rbp,8)
|
| - movq 32(%rbx,%rbp,8),%rax
|
| - sbbq 16(%rsi,%rbp,8),%r10
|
| - movq %rdx,8(%rdi,%rbp,8)
|
| - sbbq 24(%rsi,%rbp,8),%r11
|
| - movq %r10,16(%rdi,%rbp,8)
|
| -
|
| - sbbq $0,%rax
|
| - movq %r11,24(%rdi,%rbp,8)
|
| - xorq %rbp,%rbp
|
| - andq %rax,%rbx
|
| - notq %rax
|
| - movq %rdi,%rsi
|
| - andq %rax,%rsi
|
| - leaq -1(%r9),%rcx
|
| - orq %rsi,%rbx
|
| -
|
| - pxor %xmm0,%xmm0
|
| - leaq 64(%rsp,%r9,8),%rsi
|
| - movdqu (%rbx),%xmm1
|
| - leaq (%rsi,%r9,8),%rsi
|
| - movdqa %xmm0,64(%rsp)
|
| - movdqa %xmm0,(%rsi)
|
| - movdqu %xmm1,(%rdi)
|
| - jmp .Lsqr4x_copy
|
| -.align 16
|
| -.Lsqr4x_copy:
|
| - movdqu 16(%rbx,%rbp,1),%xmm2
|
| - movdqu 32(%rbx,%rbp,1),%xmm1
|
| - movdqa %xmm0,80(%rsp,%rbp,1)
|
| - movdqa %xmm0,96(%rsp,%rbp,1)
|
| - movdqa %xmm0,16(%rsi,%rbp,1)
|
| - movdqa %xmm0,32(%rsi,%rbp,1)
|
| - movdqu %xmm2,16(%rdi,%rbp,1)
|
| - movdqu %xmm1,32(%rdi,%rbp,1)
|
| - leaq 32(%rbp),%rbp
|
| - decq %rcx
|
| - jnz .Lsqr4x_copy
|
| -
|
| - movdqu 16(%rbx,%rbp,1),%xmm2
|
| - movdqa %xmm0,80(%rsp,%rbp,1)
|
| - movdqa %xmm0,16(%rsi,%rbp,1)
|
| - movdqu %xmm2,16(%rdi,%rbp,1)
|
| - movq 56(%rsp),%rsi
|
| - movq $1,%rax
|
| - movq 0(%rsi),%r15
|
| - movq 8(%rsi),%r14
|
| - movq 16(%rsi),%r13
|
| - movq 24(%rsi),%r12
|
| - movq 32(%rsi),%rbp
|
| - movq 40(%rsi),%rbx
|
| - leaq 48(%rsi),%rsp
|
| -.Lsqr4x_epilogue:
|
| - .byte 0xf3,0xc3
|
| -.size bn_sqr4x_mont,.-bn_sqr4x_mont
|
| -.byte 77,111,110,116,103,111,109,101,114,121,32,77,117,108,116,105,112,108,105,99,97,116,105,111,110,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,103,62,0
|
| -.align 16
|
|
|