| Index: third_party/boringssl/linux-x86_64/crypto/bn/x86_64-mont.S
|
| diff --git a/third_party/boringssl/linux-x86_64/crypto/bn/x86_64-mont.S b/third_party/boringssl/linux-x86_64/crypto/bn/x86_64-mont.S
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..9616df7aa65dd0c6ef7a8ed562037c1691b30df0
|
| --- /dev/null
|
| +++ b/third_party/boringssl/linux-x86_64/crypto/bn/x86_64-mont.S
|
| @@ -0,0 +1,726 @@
|
| +#if defined(__x86_64__)
|
| +.text
|
| +
|
| +
|
| +
|
| +.globl bn_mul_mont
|
| +.type bn_mul_mont,@function
|
| +.align 16
|
| +bn_mul_mont:
|
| + testl $3,%r9d
|
| + jnz .Lmul_enter
|
| + cmpl $8,%r9d
|
| + jb .Lmul_enter
|
| + cmpq %rsi,%rdx
|
| + jne .Lmul4x_enter
|
| + testl $7,%r9d
|
| + jz .Lsqr8x_enter
|
| + jmp .Lmul4x_enter
|
| +
|
| +.align 16
|
| +.Lmul_enter:
|
| + pushq %rbx
|
| + pushq %rbp
|
| + pushq %r12
|
| + pushq %r13
|
| + pushq %r14
|
| + pushq %r15
|
| +
|
| + movl %r9d,%r9d
|
| + leaq 2(%r9),%r10
|
| + movq %rsp,%r11
|
| + negq %r10
|
| + leaq (%rsp,%r10,8),%rsp
|
| + andq $-1024,%rsp
|
| +
|
| + movq %r11,8(%rsp,%r9,8)
|
| +.Lmul_body:
|
| + movq %rdx,%r12
|
| + movq (%r8),%r8
|
| + movq (%r12),%rbx
|
| + movq (%rsi),%rax
|
| +
|
| + xorq %r14,%r14
|
| + xorq %r15,%r15
|
| +
|
| + movq %r8,%rbp
|
| + mulq %rbx
|
| + movq %rax,%r10
|
| + movq (%rcx),%rax
|
| +
|
| + imulq %r10,%rbp
|
| + movq %rdx,%r11
|
| +
|
| + mulq %rbp
|
| + addq %rax,%r10
|
| + movq 8(%rsi),%rax
|
| + adcq $0,%rdx
|
| + movq %rdx,%r13
|
| +
|
| + leaq 1(%r15),%r15
|
| + jmp .L1st_enter
|
| +
|
| +.align 16
|
| +.L1st:
|
| + addq %rax,%r13
|
| + movq (%rsi,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%r13
|
| + movq %r10,%r11
|
| + adcq $0,%rdx
|
| + movq %r13,-16(%rsp,%r15,8)
|
| + movq %rdx,%r13
|
| +
|
| +.L1st_enter:
|
| + mulq %rbx
|
| + addq %rax,%r11
|
| + movq (%rcx,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + leaq 1(%r15),%r15
|
| + movq %rdx,%r10
|
| +
|
| + mulq %rbp
|
| + cmpq %r9,%r15
|
| + jne .L1st
|
| +
|
| + addq %rax,%r13
|
| + movq (%rsi),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%r13
|
| + adcq $0,%rdx
|
| + movq %r13,-16(%rsp,%r15,8)
|
| + movq %rdx,%r13
|
| + movq %r10,%r11
|
| +
|
| + xorq %rdx,%rdx
|
| + addq %r11,%r13
|
| + adcq $0,%rdx
|
| + movq %r13,-8(%rsp,%r9,8)
|
| + movq %rdx,(%rsp,%r9,8)
|
| +
|
| + leaq 1(%r14),%r14
|
| + jmp .Louter
|
| +.align 16
|
| +.Louter:
|
| + movq (%r12,%r14,8),%rbx
|
| + xorq %r15,%r15
|
| + movq %r8,%rbp
|
| + movq (%rsp),%r10
|
| + mulq %rbx
|
| + addq %rax,%r10
|
| + movq (%rcx),%rax
|
| + adcq $0,%rdx
|
| +
|
| + imulq %r10,%rbp
|
| + movq %rdx,%r11
|
| +
|
| + mulq %rbp
|
| + addq %rax,%r10
|
| + movq 8(%rsi),%rax
|
| + adcq $0,%rdx
|
| + movq 8(%rsp),%r10
|
| + movq %rdx,%r13
|
| +
|
| + leaq 1(%r15),%r15
|
| + jmp .Linner_enter
|
| +
|
| +.align 16
|
| +.Linner:
|
| + addq %rax,%r13
|
| + movq (%rsi,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + addq %r10,%r13
|
| + movq (%rsp,%r15,8),%r10
|
| + adcq $0,%rdx
|
| + movq %r13,-16(%rsp,%r15,8)
|
| + movq %rdx,%r13
|
| +
|
| +.Linner_enter:
|
| + mulq %rbx
|
| + addq %rax,%r11
|
| + movq (%rcx,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%r10
|
| + movq %rdx,%r11
|
| + adcq $0,%r11
|
| + leaq 1(%r15),%r15
|
| +
|
| + mulq %rbp
|
| + cmpq %r9,%r15
|
| + jne .Linner
|
| +
|
| + addq %rax,%r13
|
| + movq (%rsi),%rax
|
| + adcq $0,%rdx
|
| + addq %r10,%r13
|
| + movq (%rsp,%r15,8),%r10
|
| + adcq $0,%rdx
|
| + movq %r13,-16(%rsp,%r15,8)
|
| + movq %rdx,%r13
|
| +
|
| + xorq %rdx,%rdx
|
| + addq %r11,%r13
|
| + adcq $0,%rdx
|
| + addq %r10,%r13
|
| + adcq $0,%rdx
|
| + movq %r13,-8(%rsp,%r9,8)
|
| + movq %rdx,(%rsp,%r9,8)
|
| +
|
| + leaq 1(%r14),%r14
|
| + cmpq %r9,%r14
|
| + jb .Louter
|
| +
|
| + xorq %r14,%r14
|
| + movq (%rsp),%rax
|
| + leaq (%rsp),%rsi
|
| + movq %r9,%r15
|
| + jmp .Lsub
|
| +.align 16
|
| +.Lsub: sbbq (%rcx,%r14,8),%rax
|
| + movq %rax,(%rdi,%r14,8)
|
| + movq 8(%rsi,%r14,8),%rax
|
| + leaq 1(%r14),%r14
|
| + decq %r15
|
| + jnz .Lsub
|
| +
|
| + sbbq $0,%rax
|
| + xorq %r14,%r14
|
| + movq %r9,%r15
|
| +.align 16
|
| +.Lcopy:
|
| + movq (%rsp,%r14,8),%rsi
|
| + movq (%rdi,%r14,8),%rcx
|
| + xorq %rcx,%rsi
|
| + andq %rax,%rsi
|
| + xorq %rcx,%rsi
|
| + movq %r14,(%rsp,%r14,8)
|
| + movq %rsi,(%rdi,%r14,8)
|
| + leaq 1(%r14),%r14
|
| + subq $1,%r15
|
| + jnz .Lcopy
|
| +
|
| + movq 8(%rsp,%r9,8),%rsi
|
| + movq $1,%rax
|
| + movq (%rsi),%r15
|
| + movq 8(%rsi),%r14
|
| + movq 16(%rsi),%r13
|
| + movq 24(%rsi),%r12
|
| + movq 32(%rsi),%rbp
|
| + movq 40(%rsi),%rbx
|
| + leaq 48(%rsi),%rsp
|
| +.Lmul_epilogue:
|
| + .byte 0xf3,0xc3
|
| +.size bn_mul_mont,.-bn_mul_mont
|
| +.type bn_mul4x_mont,@function
|
| +.align 16
|
| +bn_mul4x_mont:
|
| +.Lmul4x_enter:
|
| + pushq %rbx
|
| + pushq %rbp
|
| + pushq %r12
|
| + pushq %r13
|
| + pushq %r14
|
| + pushq %r15
|
| +
|
| + movl %r9d,%r9d
|
| + leaq 4(%r9),%r10
|
| + movq %rsp,%r11
|
| + negq %r10
|
| + leaq (%rsp,%r10,8),%rsp
|
| + andq $-1024,%rsp
|
| +
|
| + movq %r11,8(%rsp,%r9,8)
|
| +.Lmul4x_body:
|
| + movq %rdi,16(%rsp,%r9,8)
|
| + movq %rdx,%r12
|
| + movq (%r8),%r8
|
| + movq (%r12),%rbx
|
| + movq (%rsi),%rax
|
| +
|
| + xorq %r14,%r14
|
| + xorq %r15,%r15
|
| +
|
| + movq %r8,%rbp
|
| + mulq %rbx
|
| + movq %rax,%r10
|
| + movq (%rcx),%rax
|
| +
|
| + imulq %r10,%rbp
|
| + movq %rdx,%r11
|
| +
|
| + mulq %rbp
|
| + addq %rax,%r10
|
| + movq 8(%rsi),%rax
|
| + adcq $0,%rdx
|
| + movq %rdx,%rdi
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r11
|
| + movq 8(%rcx),%rax
|
| + adcq $0,%rdx
|
| + movq %rdx,%r10
|
| +
|
| + mulq %rbp
|
| + addq %rax,%rdi
|
| + movq 16(%rsi),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%rdi
|
| + leaq 4(%r15),%r15
|
| + adcq $0,%rdx
|
| + movq %rdi,(%rsp)
|
| + movq %rdx,%r13
|
| + jmp .L1st4x
|
| +.align 16
|
| +.L1st4x:
|
| + mulq %rbx
|
| + addq %rax,%r10
|
| + movq -16(%rcx,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + movq %rdx,%r11
|
| +
|
| + mulq %rbp
|
| + addq %rax,%r13
|
| + movq -8(%rsi,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + addq %r10,%r13
|
| + adcq $0,%rdx
|
| + movq %r13,-24(%rsp,%r15,8)
|
| + movq %rdx,%rdi
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r11
|
| + movq -8(%rcx,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + movq %rdx,%r10
|
| +
|
| + mulq %rbp
|
| + addq %rax,%rdi
|
| + movq (%rsi,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%rdi
|
| + adcq $0,%rdx
|
| + movq %rdi,-16(%rsp,%r15,8)
|
| + movq %rdx,%r13
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r10
|
| + movq (%rcx,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + movq %rdx,%r11
|
| +
|
| + mulq %rbp
|
| + addq %rax,%r13
|
| + movq 8(%rsi,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + addq %r10,%r13
|
| + adcq $0,%rdx
|
| + movq %r13,-8(%rsp,%r15,8)
|
| + movq %rdx,%rdi
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r11
|
| + movq 8(%rcx,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + leaq 4(%r15),%r15
|
| + movq %rdx,%r10
|
| +
|
| + mulq %rbp
|
| + addq %rax,%rdi
|
| + movq -16(%rsi,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%rdi
|
| + adcq $0,%rdx
|
| + movq %rdi,-32(%rsp,%r15,8)
|
| + movq %rdx,%r13
|
| + cmpq %r9,%r15
|
| + jb .L1st4x
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r10
|
| + movq -16(%rcx,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + movq %rdx,%r11
|
| +
|
| + mulq %rbp
|
| + addq %rax,%r13
|
| + movq -8(%rsi,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + addq %r10,%r13
|
| + adcq $0,%rdx
|
| + movq %r13,-24(%rsp,%r15,8)
|
| + movq %rdx,%rdi
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r11
|
| + movq -8(%rcx,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + movq %rdx,%r10
|
| +
|
| + mulq %rbp
|
| + addq %rax,%rdi
|
| + movq (%rsi),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%rdi
|
| + adcq $0,%rdx
|
| + movq %rdi,-16(%rsp,%r15,8)
|
| + movq %rdx,%r13
|
| +
|
| + xorq %rdi,%rdi
|
| + addq %r10,%r13
|
| + adcq $0,%rdi
|
| + movq %r13,-8(%rsp,%r15,8)
|
| + movq %rdi,(%rsp,%r15,8)
|
| +
|
| + leaq 1(%r14),%r14
|
| +.align 4
|
| +.Louter4x:
|
| + movq (%r12,%r14,8),%rbx
|
| + xorq %r15,%r15
|
| + movq (%rsp),%r10
|
| + movq %r8,%rbp
|
| + mulq %rbx
|
| + addq %rax,%r10
|
| + movq (%rcx),%rax
|
| + adcq $0,%rdx
|
| +
|
| + imulq %r10,%rbp
|
| + movq %rdx,%r11
|
| +
|
| + mulq %rbp
|
| + addq %rax,%r10
|
| + movq 8(%rsi),%rax
|
| + adcq $0,%rdx
|
| + movq %rdx,%rdi
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r11
|
| + movq 8(%rcx),%rax
|
| + adcq $0,%rdx
|
| + addq 8(%rsp),%r11
|
| + adcq $0,%rdx
|
| + movq %rdx,%r10
|
| +
|
| + mulq %rbp
|
| + addq %rax,%rdi
|
| + movq 16(%rsi),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%rdi
|
| + leaq 4(%r15),%r15
|
| + adcq $0,%rdx
|
| + movq %rdi,(%rsp)
|
| + movq %rdx,%r13
|
| + jmp .Linner4x
|
| +.align 16
|
| +.Linner4x:
|
| + mulq %rbx
|
| + addq %rax,%r10
|
| + movq -16(%rcx,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + addq -16(%rsp,%r15,8),%r10
|
| + adcq $0,%rdx
|
| + movq %rdx,%r11
|
| +
|
| + mulq %rbp
|
| + addq %rax,%r13
|
| + movq -8(%rsi,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + addq %r10,%r13
|
| + adcq $0,%rdx
|
| + movq %r13,-24(%rsp,%r15,8)
|
| + movq %rdx,%rdi
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r11
|
| + movq -8(%rcx,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + addq -8(%rsp,%r15,8),%r11
|
| + adcq $0,%rdx
|
| + movq %rdx,%r10
|
| +
|
| + mulq %rbp
|
| + addq %rax,%rdi
|
| + movq (%rsi,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%rdi
|
| + adcq $0,%rdx
|
| + movq %rdi,-16(%rsp,%r15,8)
|
| + movq %rdx,%r13
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r10
|
| + movq (%rcx,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + addq (%rsp,%r15,8),%r10
|
| + adcq $0,%rdx
|
| + movq %rdx,%r11
|
| +
|
| + mulq %rbp
|
| + addq %rax,%r13
|
| + movq 8(%rsi,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + addq %r10,%r13
|
| + adcq $0,%rdx
|
| + movq %r13,-8(%rsp,%r15,8)
|
| + movq %rdx,%rdi
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r11
|
| + movq 8(%rcx,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + addq 8(%rsp,%r15,8),%r11
|
| + adcq $0,%rdx
|
| + leaq 4(%r15),%r15
|
| + movq %rdx,%r10
|
| +
|
| + mulq %rbp
|
| + addq %rax,%rdi
|
| + movq -16(%rsi,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%rdi
|
| + adcq $0,%rdx
|
| + movq %rdi,-32(%rsp,%r15,8)
|
| + movq %rdx,%r13
|
| + cmpq %r9,%r15
|
| + jb .Linner4x
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r10
|
| + movq -16(%rcx,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + addq -16(%rsp,%r15,8),%r10
|
| + adcq $0,%rdx
|
| + movq %rdx,%r11
|
| +
|
| + mulq %rbp
|
| + addq %rax,%r13
|
| + movq -8(%rsi,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + addq %r10,%r13
|
| + adcq $0,%rdx
|
| + movq %r13,-24(%rsp,%r15,8)
|
| + movq %rdx,%rdi
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r11
|
| + movq -8(%rcx,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + addq -8(%rsp,%r15,8),%r11
|
| + adcq $0,%rdx
|
| + leaq 1(%r14),%r14
|
| + movq %rdx,%r10
|
| +
|
| + mulq %rbp
|
| + addq %rax,%rdi
|
| + movq (%rsi),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%rdi
|
| + adcq $0,%rdx
|
| + movq %rdi,-16(%rsp,%r15,8)
|
| + movq %rdx,%r13
|
| +
|
| + xorq %rdi,%rdi
|
| + addq %r10,%r13
|
| + adcq $0,%rdi
|
| + addq (%rsp,%r9,8),%r13
|
| + adcq $0,%rdi
|
| + movq %r13,-8(%rsp,%r15,8)
|
| + movq %rdi,(%rsp,%r15,8)
|
| +
|
| + cmpq %r9,%r14
|
| + jb .Louter4x
|
| + movq 16(%rsp,%r9,8),%rdi
|
| + movq 0(%rsp),%rax
|
| + movq 8(%rsp),%rdx
|
| + shrq $2,%r9
|
| + leaq (%rsp),%rsi
|
| + xorq %r14,%r14
|
| +
|
| + subq 0(%rcx),%rax
|
| + movq 16(%rsi),%rbx
|
| + movq 24(%rsi),%rbp
|
| + sbbq 8(%rcx),%rdx
|
| + leaq -1(%r9),%r15
|
| + jmp .Lsub4x
|
| +.align 16
|
| +.Lsub4x:
|
| + movq %rax,0(%rdi,%r14,8)
|
| + movq %rdx,8(%rdi,%r14,8)
|
| + sbbq 16(%rcx,%r14,8),%rbx
|
| + movq 32(%rsi,%r14,8),%rax
|
| + movq 40(%rsi,%r14,8),%rdx
|
| + sbbq 24(%rcx,%r14,8),%rbp
|
| + movq %rbx,16(%rdi,%r14,8)
|
| + movq %rbp,24(%rdi,%r14,8)
|
| + sbbq 32(%rcx,%r14,8),%rax
|
| + movq 48(%rsi,%r14,8),%rbx
|
| + movq 56(%rsi,%r14,8),%rbp
|
| + sbbq 40(%rcx,%r14,8),%rdx
|
| + leaq 4(%r14),%r14
|
| + decq %r15
|
| + jnz .Lsub4x
|
| +
|
| + movq %rax,0(%rdi,%r14,8)
|
| + movq 32(%rsi,%r14,8),%rax
|
| + sbbq 16(%rcx,%r14,8),%rbx
|
| + movq %rdx,8(%rdi,%r14,8)
|
| + sbbq 24(%rcx,%r14,8),%rbp
|
| + movq %rbx,16(%rdi,%r14,8)
|
| +
|
| + sbbq $0,%rax
|
| + movq %rax,%xmm0
|
| + punpcklqdq %xmm0,%xmm0
|
| + movq %rbp,24(%rdi,%r14,8)
|
| + xorq %r14,%r14
|
| +
|
| + movq %r9,%r15
|
| + pxor %xmm5,%xmm5
|
| + jmp .Lcopy4x
|
| +.align 16
|
| +.Lcopy4x:
|
| + movdqu (%rsp,%r14,1),%xmm2
|
| + movdqu 16(%rsp,%r14,1),%xmm4
|
| + movdqu (%rdi,%r14,1),%xmm1
|
| + movdqu 16(%rdi,%r14,1),%xmm3
|
| + pxor %xmm1,%xmm2
|
| + pxor %xmm3,%xmm4
|
| + pand %xmm0,%xmm2
|
| + pand %xmm0,%xmm4
|
| + pxor %xmm1,%xmm2
|
| + pxor %xmm3,%xmm4
|
| + movdqu %xmm2,(%rdi,%r14,1)
|
| + movdqu %xmm4,16(%rdi,%r14,1)
|
| + movdqa %xmm5,(%rsp,%r14,1)
|
| + movdqa %xmm5,16(%rsp,%r14,1)
|
| +
|
| + leaq 32(%r14),%r14
|
| + decq %r15
|
| + jnz .Lcopy4x
|
| +
|
| + shlq $2,%r9
|
| + movq 8(%rsp,%r9,8),%rsi
|
| + movq $1,%rax
|
| + movq (%rsi),%r15
|
| + movq 8(%rsi),%r14
|
| + movq 16(%rsi),%r13
|
| + movq 24(%rsi),%r12
|
| + movq 32(%rsi),%rbp
|
| + movq 40(%rsi),%rbx
|
| + leaq 48(%rsi),%rsp
|
| +.Lmul4x_epilogue:
|
| + .byte 0xf3,0xc3
|
| +.size bn_mul4x_mont,.-bn_mul4x_mont
|
| +
|
| +
|
| +
|
| +.type bn_sqr8x_mont,@function
|
| +.align 32
|
| +bn_sqr8x_mont:
|
| +.Lsqr8x_enter:
|
| + movq %rsp,%rax
|
| + pushq %rbx
|
| + pushq %rbp
|
| + pushq %r12
|
| + pushq %r13
|
| + pushq %r14
|
| + pushq %r15
|
| +
|
| + movl %r9d,%r10d
|
| + shll $3,%r9d
|
| + shlq $3+2,%r10
|
| + negq %r9
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| + leaq -64(%rsp,%r9,4),%r11
|
| + movq (%r8),%r8
|
| + subq %rsi,%r11
|
| + andq $4095,%r11
|
| + cmpq %r11,%r10
|
| + jb .Lsqr8x_sp_alt
|
| + subq %r11,%rsp
|
| + leaq -64(%rsp,%r9,4),%rsp
|
| + jmp .Lsqr8x_sp_done
|
| +
|
| +.align 32
|
| +.Lsqr8x_sp_alt:
|
| + leaq 4096-64(,%r9,4),%r10
|
| + leaq -64(%rsp,%r9,4),%rsp
|
| + subq %r10,%r11
|
| + movq $0,%r10
|
| + cmovcq %r10,%r11
|
| + subq %r11,%rsp
|
| +.Lsqr8x_sp_done:
|
| + andq $-64,%rsp
|
| + movq %r9,%r10
|
| + negq %r9
|
| +
|
| + leaq 64(%rsp,%r9,2),%r11
|
| + movq %r8,32(%rsp)
|
| + movq %rax,40(%rsp)
|
| +.Lsqr8x_body:
|
| +
|
| + movq %r9,%rbp
|
| +.byte 102,73,15,110,211
|
| + shrq $3+2,%rbp
|
| + movl OPENSSL_ia32cap_P+8(%rip),%eax
|
| + jmp .Lsqr8x_copy_n
|
| +
|
| +.align 32
|
| +.Lsqr8x_copy_n:
|
| + movq 0(%rcx),%xmm0
|
| + movq 8(%rcx),%xmm1
|
| + movq 16(%rcx),%xmm3
|
| + movq 24(%rcx),%xmm4
|
| + leaq 32(%rcx),%rcx
|
| + movdqa %xmm0,0(%r11)
|
| + movdqa %xmm1,16(%r11)
|
| + movdqa %xmm3,32(%r11)
|
| + movdqa %xmm4,48(%r11)
|
| + leaq 64(%r11),%r11
|
| + decq %rbp
|
| + jnz .Lsqr8x_copy_n
|
| +
|
| + pxor %xmm0,%xmm0
|
| +.byte 102,72,15,110,207
|
| +.byte 102,73,15,110,218
|
| + call bn_sqr8x_internal
|
| +
|
| + pxor %xmm0,%xmm0
|
| + leaq 48(%rsp),%rax
|
| + leaq 64(%rsp,%r9,2),%rdx
|
| + shrq $3+2,%r9
|
| + movq 40(%rsp),%rsi
|
| + jmp .Lsqr8x_zero
|
| +
|
| +.align 32
|
| +.Lsqr8x_zero:
|
| + movdqa %xmm0,0(%rax)
|
| + movdqa %xmm0,16(%rax)
|
| + movdqa %xmm0,32(%rax)
|
| + movdqa %xmm0,48(%rax)
|
| + leaq 64(%rax),%rax
|
| + movdqa %xmm0,0(%rdx)
|
| + movdqa %xmm0,16(%rdx)
|
| + movdqa %xmm0,32(%rdx)
|
| + movdqa %xmm0,48(%rdx)
|
| + leaq 64(%rdx),%rdx
|
| + decq %r9
|
| + jnz .Lsqr8x_zero
|
| +
|
| + movq $1,%rax
|
| + movq -48(%rsi),%r15
|
| + movq -40(%rsi),%r14
|
| + movq -32(%rsi),%r13
|
| + movq -24(%rsi),%r12
|
| + movq -16(%rsi),%rbp
|
| + movq -8(%rsi),%rbx
|
| + leaq (%rsi),%rsp
|
| +.Lsqr8x_epilogue:
|
| + .byte 0xf3,0xc3
|
| +.size bn_sqr8x_mont,.-bn_sqr8x_mont
|
| +.byte 77,111,110,116,103,111,109,101,114,121,32,77,117,108,116,105,112,108,105,99,97,116,105,111,110,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,103,62,0
|
| +.align 16
|
| +#endif
|
|
|