| Index: third_party/boringssl/linux-x86_64/crypto/bn/x86_64-mont5.S
|
| diff --git a/third_party/boringssl/linux-x86_64/crypto/bn/x86_64-mont5.S b/third_party/boringssl/linux-x86_64/crypto/bn/x86_64-mont5.S
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..fa01653eee98e023cfdb3bf67e19f5186f221e46
|
| --- /dev/null
|
| +++ b/third_party/boringssl/linux-x86_64/crypto/bn/x86_64-mont5.S
|
| @@ -0,0 +1,1830 @@
|
| +#if defined(__x86_64__)
|
| +.text
|
| +
|
| +
|
| +
|
| +.globl bn_mul_mont_gather5
|
| +.type bn_mul_mont_gather5,@function
|
| +.align 64
|
| +bn_mul_mont_gather5:
|
| + testl $7,%r9d
|
| + jnz .Lmul_enter
|
| + jmp .Lmul4x_enter
|
| +
|
| +.align 16
|
| +.Lmul_enter:
|
| + movl %r9d,%r9d
|
| + movq %rsp,%rax
|
| + movl 8(%rsp),%r10d
|
| + pushq %rbx
|
| + pushq %rbp
|
| + pushq %r12
|
| + pushq %r13
|
| + pushq %r14
|
| + pushq %r15
|
| + leaq 2(%r9),%r11
|
| + negq %r11
|
| + leaq (%rsp,%r11,8),%rsp
|
| + andq $-1024,%rsp
|
| +
|
| + movq %rax,8(%rsp,%r9,8)
|
| +.Lmul_body:
|
| + movq %rdx,%r12
|
| + movq %r10,%r11
|
| + shrq $3,%r10
|
| + andq $7,%r11
|
| + notq %r10
|
| + leaq .Lmagic_masks(%rip),%rax
|
| + andq $3,%r10
|
| + leaq 96(%r12,%r11,8),%r12
|
| + movq 0(%rax,%r10,8),%xmm4
|
| + movq 8(%rax,%r10,8),%xmm5
|
| + movq 16(%rax,%r10,8),%xmm6
|
| + movq 24(%rax,%r10,8),%xmm7
|
| +
|
| + movq -96(%r12),%xmm0
|
| + movq -32(%r12),%xmm1
|
| + pand %xmm4,%xmm0
|
| + movq 32(%r12),%xmm2
|
| + pand %xmm5,%xmm1
|
| + movq 96(%r12),%xmm3
|
| + pand %xmm6,%xmm2
|
| + por %xmm1,%xmm0
|
| + pand %xmm7,%xmm3
|
| + por %xmm2,%xmm0
|
| + leaq 256(%r12),%r12
|
| + por %xmm3,%xmm0
|
| +
|
| +.byte 102,72,15,126,195
|
| +
|
| + movq (%r8),%r8
|
| + movq (%rsi),%rax
|
| +
|
| + xorq %r14,%r14
|
| + xorq %r15,%r15
|
| +
|
| + movq -96(%r12),%xmm0
|
| + movq -32(%r12),%xmm1
|
| + pand %xmm4,%xmm0
|
| + movq 32(%r12),%xmm2
|
| + pand %xmm5,%xmm1
|
| +
|
| + movq %r8,%rbp
|
| + mulq %rbx
|
| + movq %rax,%r10
|
| + movq (%rcx),%rax
|
| +
|
| + movq 96(%r12),%xmm3
|
| + pand %xmm6,%xmm2
|
| + por %xmm1,%xmm0
|
| + pand %xmm7,%xmm3
|
| +
|
| + imulq %r10,%rbp
|
| + movq %rdx,%r11
|
| +
|
| + por %xmm2,%xmm0
|
| + leaq 256(%r12),%r12
|
| + por %xmm3,%xmm0
|
| +
|
| + mulq %rbp
|
| + addq %rax,%r10
|
| + movq 8(%rsi),%rax
|
| + adcq $0,%rdx
|
| + movq %rdx,%r13
|
| +
|
| + leaq 1(%r15),%r15
|
| + jmp .L1st_enter
|
| +
|
| +.align 16
|
| +.L1st:
|
| + addq %rax,%r13
|
| + movq (%rsi,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%r13
|
| + movq %r10,%r11
|
| + adcq $0,%rdx
|
| + movq %r13,-16(%rsp,%r15,8)
|
| + movq %rdx,%r13
|
| +
|
| +.L1st_enter:
|
| + mulq %rbx
|
| + addq %rax,%r11
|
| + movq (%rcx,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + leaq 1(%r15),%r15
|
| + movq %rdx,%r10
|
| +
|
| + mulq %rbp
|
| + cmpq %r9,%r15
|
| + jne .L1st
|
| +
|
| +.byte 102,72,15,126,195
|
| +
|
| + addq %rax,%r13
|
| + movq (%rsi),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%r13
|
| + adcq $0,%rdx
|
| + movq %r13,-16(%rsp,%r15,8)
|
| + movq %rdx,%r13
|
| + movq %r10,%r11
|
| +
|
| + xorq %rdx,%rdx
|
| + addq %r11,%r13
|
| + adcq $0,%rdx
|
| + movq %r13,-8(%rsp,%r9,8)
|
| + movq %rdx,(%rsp,%r9,8)
|
| +
|
| + leaq 1(%r14),%r14
|
| + jmp .Louter
|
| +.align 16
|
| +.Louter:
|
| + xorq %r15,%r15
|
| + movq %r8,%rbp
|
| + movq (%rsp),%r10
|
| +
|
| + movq -96(%r12),%xmm0
|
| + movq -32(%r12),%xmm1
|
| + pand %xmm4,%xmm0
|
| + movq 32(%r12),%xmm2
|
| + pand %xmm5,%xmm1
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r10
|
| + movq (%rcx),%rax
|
| + adcq $0,%rdx
|
| +
|
| + movq 96(%r12),%xmm3
|
| + pand %xmm6,%xmm2
|
| + por %xmm1,%xmm0
|
| + pand %xmm7,%xmm3
|
| +
|
| + imulq %r10,%rbp
|
| + movq %rdx,%r11
|
| +
|
| + por %xmm2,%xmm0
|
| + leaq 256(%r12),%r12
|
| + por %xmm3,%xmm0
|
| +
|
| + mulq %rbp
|
| + addq %rax,%r10
|
| + movq 8(%rsi),%rax
|
| + adcq $0,%rdx
|
| + movq 8(%rsp),%r10
|
| + movq %rdx,%r13
|
| +
|
| + leaq 1(%r15),%r15
|
| + jmp .Linner_enter
|
| +
|
| +.align 16
|
| +.Linner:
|
| + addq %rax,%r13
|
| + movq (%rsi,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + addq %r10,%r13
|
| + movq (%rsp,%r15,8),%r10
|
| + adcq $0,%rdx
|
| + movq %r13,-16(%rsp,%r15,8)
|
| + movq %rdx,%r13
|
| +
|
| +.Linner_enter:
|
| + mulq %rbx
|
| + addq %rax,%r11
|
| + movq (%rcx,%r15,8),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%r10
|
| + movq %rdx,%r11
|
| + adcq $0,%r11
|
| + leaq 1(%r15),%r15
|
| +
|
| + mulq %rbp
|
| + cmpq %r9,%r15
|
| + jne .Linner
|
| +
|
| +.byte 102,72,15,126,195
|
| +
|
| + addq %rax,%r13
|
| + movq (%rsi),%rax
|
| + adcq $0,%rdx
|
| + addq %r10,%r13
|
| + movq (%rsp,%r15,8),%r10
|
| + adcq $0,%rdx
|
| + movq %r13,-16(%rsp,%r15,8)
|
| + movq %rdx,%r13
|
| +
|
| + xorq %rdx,%rdx
|
| + addq %r11,%r13
|
| + adcq $0,%rdx
|
| + addq %r10,%r13
|
| + adcq $0,%rdx
|
| + movq %r13,-8(%rsp,%r9,8)
|
| + movq %rdx,(%rsp,%r9,8)
|
| +
|
| + leaq 1(%r14),%r14
|
| + cmpq %r9,%r14
|
| + jb .Louter
|
| +
|
| + xorq %r14,%r14
|
| + movq (%rsp),%rax
|
| + leaq (%rsp),%rsi
|
| + movq %r9,%r15
|
| + jmp .Lsub
|
| +.align 16
|
| +.Lsub: sbbq (%rcx,%r14,8),%rax
|
| + movq %rax,(%rdi,%r14,8)
|
| + movq 8(%rsi,%r14,8),%rax
|
| + leaq 1(%r14),%r14
|
| + decq %r15
|
| + jnz .Lsub
|
| +
|
| + sbbq $0,%rax
|
| + xorq %r14,%r14
|
| + movq %r9,%r15
|
| +.align 16
|
| +.Lcopy:
|
| + movq (%rsp,%r14,8),%rsi
|
| + movq (%rdi,%r14,8),%rcx
|
| + xorq %rcx,%rsi
|
| + andq %rax,%rsi
|
| + xorq %rcx,%rsi
|
| + movq %r14,(%rsp,%r14,8)
|
| + movq %rsi,(%rdi,%r14,8)
|
| + leaq 1(%r14),%r14
|
| + subq $1,%r15
|
| + jnz .Lcopy
|
| +
|
| + movq 8(%rsp,%r9,8),%rsi
|
| + movq $1,%rax
|
| + movq -48(%rsi),%r15
|
| + movq -40(%rsi),%r14
|
| + movq -32(%rsi),%r13
|
| + movq -24(%rsi),%r12
|
| + movq -16(%rsi),%rbp
|
| + movq -8(%rsi),%rbx
|
| + leaq (%rsi),%rsp
|
| +.Lmul_epilogue:
|
| + .byte 0xf3,0xc3
|
| +.size bn_mul_mont_gather5,.-bn_mul_mont_gather5
|
| +.type bn_mul4x_mont_gather5,@function
|
| +.align 32
|
| +bn_mul4x_mont_gather5:
|
| +.Lmul4x_enter:
|
| +.byte 0x67
|
| + movq %rsp,%rax
|
| + pushq %rbx
|
| + pushq %rbp
|
| + pushq %r12
|
| + pushq %r13
|
| + pushq %r14
|
| + pushq %r15
|
| +.byte 0x67
|
| + movl %r9d,%r10d
|
| + shll $3,%r9d
|
| + shll $3+2,%r10d
|
| + negq %r9
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| + leaq -64(%rsp,%r9,2),%r11
|
| + subq %rsi,%r11
|
| + andq $4095,%r11
|
| + cmpq %r11,%r10
|
| + jb .Lmul4xsp_alt
|
| + subq %r11,%rsp
|
| + leaq -64(%rsp,%r9,2),%rsp
|
| + jmp .Lmul4xsp_done
|
| +
|
| +.align 32
|
| +.Lmul4xsp_alt:
|
| + leaq 4096-64(,%r9,2),%r10
|
| + leaq -64(%rsp,%r9,2),%rsp
|
| + subq %r10,%r11
|
| + movq $0,%r10
|
| + cmovcq %r10,%r11
|
| + subq %r11,%rsp
|
| +.Lmul4xsp_done:
|
| + andq $-64,%rsp
|
| + negq %r9
|
| +
|
| + movq %rax,40(%rsp)
|
| +.Lmul4x_body:
|
| +
|
| + call mul4x_internal
|
| +
|
| + movq 40(%rsp),%rsi
|
| + movq $1,%rax
|
| + movq -48(%rsi),%r15
|
| + movq -40(%rsi),%r14
|
| + movq -32(%rsi),%r13
|
| + movq -24(%rsi),%r12
|
| + movq -16(%rsi),%rbp
|
| + movq -8(%rsi),%rbx
|
| + leaq (%rsi),%rsp
|
| +.Lmul4x_epilogue:
|
| + .byte 0xf3,0xc3
|
| +.size bn_mul4x_mont_gather5,.-bn_mul4x_mont_gather5
|
| +
|
| +.type mul4x_internal,@function
|
| +.align 32
|
| +mul4x_internal:
|
| + shlq $5,%r9
|
| + movl 8(%rax),%r10d
|
| + leaq 256(%rdx,%r9,1),%r13
|
| + shrq $5,%r9
|
| + movq %r10,%r11
|
| + shrq $3,%r10
|
| + andq $7,%r11
|
| + notq %r10
|
| + leaq .Lmagic_masks(%rip),%rax
|
| + andq $3,%r10
|
| + leaq 96(%rdx,%r11,8),%r12
|
| + movq 0(%rax,%r10,8),%xmm4
|
| + movq 8(%rax,%r10,8),%xmm5
|
| + addq $7,%r11
|
| + movq 16(%rax,%r10,8),%xmm6
|
| + movq 24(%rax,%r10,8),%xmm7
|
| + andq $7,%r11
|
| +
|
| + movq -96(%r12),%xmm0
|
| + leaq 256(%r12),%r14
|
| + movq -32(%r12),%xmm1
|
| + pand %xmm4,%xmm0
|
| + movq 32(%r12),%xmm2
|
| + pand %xmm5,%xmm1
|
| + movq 96(%r12),%xmm3
|
| + pand %xmm6,%xmm2
|
| +.byte 0x67
|
| + por %xmm1,%xmm0
|
| + movq -96(%r14),%xmm1
|
| +.byte 0x67
|
| + pand %xmm7,%xmm3
|
| +.byte 0x67
|
| + por %xmm2,%xmm0
|
| + movq -32(%r14),%xmm2
|
| +.byte 0x67
|
| + pand %xmm4,%xmm1
|
| +.byte 0x67
|
| + por %xmm3,%xmm0
|
| + movq 32(%r14),%xmm3
|
| +
|
| +.byte 102,72,15,126,195
|
| + movq 96(%r14),%xmm0
|
| + movq %r13,16+8(%rsp)
|
| + movq %rdi,56+8(%rsp)
|
| +
|
| + movq (%r8),%r8
|
| + movq (%rsi),%rax
|
| + leaq (%rsi,%r9,1),%rsi
|
| + negq %r9
|
| +
|
| + movq %r8,%rbp
|
| + mulq %rbx
|
| + movq %rax,%r10
|
| + movq (%rcx),%rax
|
| +
|
| + pand %xmm5,%xmm2
|
| + pand %xmm6,%xmm3
|
| + por %xmm2,%xmm1
|
| +
|
| + imulq %r10,%rbp
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| + leaq 64+8(%rsp,%r11,8),%r14
|
| + movq %rdx,%r11
|
| +
|
| + pand %xmm7,%xmm0
|
| + por %xmm3,%xmm1
|
| + leaq 512(%r12),%r12
|
| + por %xmm1,%xmm0
|
| +
|
| + mulq %rbp
|
| + addq %rax,%r10
|
| + movq 8(%rsi,%r9,1),%rax
|
| + adcq $0,%rdx
|
| + movq %rdx,%rdi
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r11
|
| + movq 16(%rcx),%rax
|
| + adcq $0,%rdx
|
| + movq %rdx,%r10
|
| +
|
| + mulq %rbp
|
| + addq %rax,%rdi
|
| + movq 16(%rsi,%r9,1),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%rdi
|
| + leaq 32(%r9),%r15
|
| + leaq 64(%rcx),%rcx
|
| + adcq $0,%rdx
|
| + movq %rdi,(%r14)
|
| + movq %rdx,%r13
|
| + jmp .L1st4x
|
| +
|
| +.align 32
|
| +.L1st4x:
|
| + mulq %rbx
|
| + addq %rax,%r10
|
| + movq -32(%rcx),%rax
|
| + leaq 32(%r14),%r14
|
| + adcq $0,%rdx
|
| + movq %rdx,%r11
|
| +
|
| + mulq %rbp
|
| + addq %rax,%r13
|
| + movq -8(%rsi,%r15,1),%rax
|
| + adcq $0,%rdx
|
| + addq %r10,%r13
|
| + adcq $0,%rdx
|
| + movq %r13,-24(%r14)
|
| + movq %rdx,%rdi
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r11
|
| + movq -16(%rcx),%rax
|
| + adcq $0,%rdx
|
| + movq %rdx,%r10
|
| +
|
| + mulq %rbp
|
| + addq %rax,%rdi
|
| + movq (%rsi,%r15,1),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%rdi
|
| + adcq $0,%rdx
|
| + movq %rdi,-16(%r14)
|
| + movq %rdx,%r13
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r10
|
| + movq 0(%rcx),%rax
|
| + adcq $0,%rdx
|
| + movq %rdx,%r11
|
| +
|
| + mulq %rbp
|
| + addq %rax,%r13
|
| + movq 8(%rsi,%r15,1),%rax
|
| + adcq $0,%rdx
|
| + addq %r10,%r13
|
| + adcq $0,%rdx
|
| + movq %r13,-8(%r14)
|
| + movq %rdx,%rdi
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r11
|
| + movq 16(%rcx),%rax
|
| + adcq $0,%rdx
|
| + movq %rdx,%r10
|
| +
|
| + mulq %rbp
|
| + addq %rax,%rdi
|
| + movq 16(%rsi,%r15,1),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%rdi
|
| + leaq 64(%rcx),%rcx
|
| + adcq $0,%rdx
|
| + movq %rdi,(%r14)
|
| + movq %rdx,%r13
|
| +
|
| + addq $32,%r15
|
| + jnz .L1st4x
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r10
|
| + movq -32(%rcx),%rax
|
| + leaq 32(%r14),%r14
|
| + adcq $0,%rdx
|
| + movq %rdx,%r11
|
| +
|
| + mulq %rbp
|
| + addq %rax,%r13
|
| + movq -8(%rsi),%rax
|
| + adcq $0,%rdx
|
| + addq %r10,%r13
|
| + adcq $0,%rdx
|
| + movq %r13,-24(%r14)
|
| + movq %rdx,%rdi
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r11
|
| + movq -16(%rcx),%rax
|
| + adcq $0,%rdx
|
| + movq %rdx,%r10
|
| +
|
| + mulq %rbp
|
| + addq %rax,%rdi
|
| + movq (%rsi,%r9,1),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%rdi
|
| + adcq $0,%rdx
|
| + movq %rdi,-16(%r14)
|
| + movq %rdx,%r13
|
| +
|
| +.byte 102,72,15,126,195
|
| + leaq (%rcx,%r9,2),%rcx
|
| +
|
| + xorq %rdi,%rdi
|
| + addq %r10,%r13
|
| + adcq $0,%rdi
|
| + movq %r13,-8(%r14)
|
| +
|
| + jmp .Louter4x
|
| +
|
| +.align 32
|
| +.Louter4x:
|
| + movq (%r14,%r9,1),%r10
|
| + movq %r8,%rbp
|
| + mulq %rbx
|
| + addq %rax,%r10
|
| + movq (%rcx),%rax
|
| + adcq $0,%rdx
|
| +
|
| + movq -96(%r12),%xmm0
|
| + movq -32(%r12),%xmm1
|
| + pand %xmm4,%xmm0
|
| + movq 32(%r12),%xmm2
|
| + pand %xmm5,%xmm1
|
| + movq 96(%r12),%xmm3
|
| +
|
| + imulq %r10,%rbp
|
| +.byte 0x67
|
| + movq %rdx,%r11
|
| + movq %rdi,(%r14)
|
| +
|
| + pand %xmm6,%xmm2
|
| + por %xmm1,%xmm0
|
| + pand %xmm7,%xmm3
|
| + por %xmm2,%xmm0
|
| + leaq (%r14,%r9,1),%r14
|
| + leaq 256(%r12),%r12
|
| + por %xmm3,%xmm0
|
| +
|
| + mulq %rbp
|
| + addq %rax,%r10
|
| + movq 8(%rsi,%r9,1),%rax
|
| + adcq $0,%rdx
|
| + movq %rdx,%rdi
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r11
|
| + movq 16(%rcx),%rax
|
| + adcq $0,%rdx
|
| + addq 8(%r14),%r11
|
| + adcq $0,%rdx
|
| + movq %rdx,%r10
|
| +
|
| + mulq %rbp
|
| + addq %rax,%rdi
|
| + movq 16(%rsi,%r9,1),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%rdi
|
| + leaq 32(%r9),%r15
|
| + leaq 64(%rcx),%rcx
|
| + adcq $0,%rdx
|
| + movq %rdx,%r13
|
| + jmp .Linner4x
|
| +
|
| +.align 32
|
| +.Linner4x:
|
| + mulq %rbx
|
| + addq %rax,%r10
|
| + movq -32(%rcx),%rax
|
| + adcq $0,%rdx
|
| + addq 16(%r14),%r10
|
| + leaq 32(%r14),%r14
|
| + adcq $0,%rdx
|
| + movq %rdx,%r11
|
| +
|
| + mulq %rbp
|
| + addq %rax,%r13
|
| + movq -8(%rsi,%r15,1),%rax
|
| + adcq $0,%rdx
|
| + addq %r10,%r13
|
| + adcq $0,%rdx
|
| + movq %rdi,-32(%r14)
|
| + movq %rdx,%rdi
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r11
|
| + movq -16(%rcx),%rax
|
| + adcq $0,%rdx
|
| + addq -8(%r14),%r11
|
| + adcq $0,%rdx
|
| + movq %rdx,%r10
|
| +
|
| + mulq %rbp
|
| + addq %rax,%rdi
|
| + movq (%rsi,%r15,1),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%rdi
|
| + adcq $0,%rdx
|
| + movq %r13,-24(%r14)
|
| + movq %rdx,%r13
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r10
|
| + movq 0(%rcx),%rax
|
| + adcq $0,%rdx
|
| + addq (%r14),%r10
|
| + adcq $0,%rdx
|
| + movq %rdx,%r11
|
| +
|
| + mulq %rbp
|
| + addq %rax,%r13
|
| + movq 8(%rsi,%r15,1),%rax
|
| + adcq $0,%rdx
|
| + addq %r10,%r13
|
| + adcq $0,%rdx
|
| + movq %rdi,-16(%r14)
|
| + movq %rdx,%rdi
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r11
|
| + movq 16(%rcx),%rax
|
| + adcq $0,%rdx
|
| + addq 8(%r14),%r11
|
| + adcq $0,%rdx
|
| + movq %rdx,%r10
|
| +
|
| + mulq %rbp
|
| + addq %rax,%rdi
|
| + movq 16(%rsi,%r15,1),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%rdi
|
| + leaq 64(%rcx),%rcx
|
| + adcq $0,%rdx
|
| + movq %r13,-8(%r14)
|
| + movq %rdx,%r13
|
| +
|
| + addq $32,%r15
|
| + jnz .Linner4x
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r10
|
| + movq -32(%rcx),%rax
|
| + adcq $0,%rdx
|
| + addq 16(%r14),%r10
|
| + leaq 32(%r14),%r14
|
| + adcq $0,%rdx
|
| + movq %rdx,%r11
|
| +
|
| + mulq %rbp
|
| + addq %rax,%r13
|
| + movq -8(%rsi),%rax
|
| + adcq $0,%rdx
|
| + addq %r10,%r13
|
| + adcq $0,%rdx
|
| + movq %rdi,-32(%r14)
|
| + movq %rdx,%rdi
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r11
|
| + movq %rbp,%rax
|
| + movq -16(%rcx),%rbp
|
| + adcq $0,%rdx
|
| + addq -8(%r14),%r11
|
| + adcq $0,%rdx
|
| + movq %rdx,%r10
|
| +
|
| + mulq %rbp
|
| + addq %rax,%rdi
|
| + movq (%rsi,%r9,1),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%rdi
|
| + adcq $0,%rdx
|
| + movq %r13,-24(%r14)
|
| + movq %rdx,%r13
|
| +
|
| +.byte 102,72,15,126,195
|
| + movq %rdi,-16(%r14)
|
| + leaq (%rcx,%r9,2),%rcx
|
| +
|
| + xorq %rdi,%rdi
|
| + addq %r10,%r13
|
| + adcq $0,%rdi
|
| + addq (%r14),%r13
|
| + adcq $0,%rdi
|
| + movq %r13,-8(%r14)
|
| +
|
| + cmpq 16+8(%rsp),%r12
|
| + jb .Louter4x
|
| + subq %r13,%rbp
|
| + adcq %r15,%r15
|
| + orq %r15,%rdi
|
| + xorq $1,%rdi
|
| + leaq (%r14,%r9,1),%rbx
|
| + leaq (%rcx,%rdi,8),%rbp
|
| + movq %r9,%rcx
|
| + sarq $3+2,%rcx
|
| + movq 56+8(%rsp),%rdi
|
| + jmp .Lsqr4x_sub
|
| +.size mul4x_internal,.-mul4x_internal
|
| +.globl bn_power5
|
| +.type bn_power5,@function
|
| +.align 32
|
| +bn_power5:
|
| + movq %rsp,%rax
|
| + pushq %rbx
|
| + pushq %rbp
|
| + pushq %r12
|
| + pushq %r13
|
| + pushq %r14
|
| + pushq %r15
|
| + movl %r9d,%r10d
|
| + shll $3,%r9d
|
| + shll $3+2,%r10d
|
| + negq %r9
|
| + movq (%r8),%r8
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| + leaq -64(%rsp,%r9,2),%r11
|
| + subq %rsi,%r11
|
| + andq $4095,%r11
|
| + cmpq %r11,%r10
|
| + jb .Lpwr_sp_alt
|
| + subq %r11,%rsp
|
| + leaq -64(%rsp,%r9,2),%rsp
|
| + jmp .Lpwr_sp_done
|
| +
|
| +.align 32
|
| +.Lpwr_sp_alt:
|
| + leaq 4096-64(,%r9,2),%r10
|
| + leaq -64(%rsp,%r9,2),%rsp
|
| + subq %r10,%r11
|
| + movq $0,%r10
|
| + cmovcq %r10,%r11
|
| + subq %r11,%rsp
|
| +.Lpwr_sp_done:
|
| + andq $-64,%rsp
|
| + movq %r9,%r10
|
| + negq %r9
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| + movq %r8,32(%rsp)
|
| + movq %rax,40(%rsp)
|
| +.Lpower5_body:
|
| +.byte 102,72,15,110,207
|
| +.byte 102,72,15,110,209
|
| +.byte 102,73,15,110,218
|
| +.byte 102,72,15,110,226
|
| +
|
| + call __bn_sqr8x_internal
|
| + call __bn_sqr8x_internal
|
| + call __bn_sqr8x_internal
|
| + call __bn_sqr8x_internal
|
| + call __bn_sqr8x_internal
|
| +
|
| +.byte 102,72,15,126,209
|
| +.byte 102,72,15,126,226
|
| + movq %rsi,%rdi
|
| + movq 40(%rsp),%rax
|
| + leaq 32(%rsp),%r8
|
| +
|
| + call mul4x_internal
|
| +
|
| + movq 40(%rsp),%rsi
|
| + movq $1,%rax
|
| + movq -48(%rsi),%r15
|
| + movq -40(%rsi),%r14
|
| + movq -32(%rsi),%r13
|
| + movq -24(%rsi),%r12
|
| + movq -16(%rsi),%rbp
|
| + movq -8(%rsi),%rbx
|
| + leaq (%rsi),%rsp
|
| +.Lpower5_epilogue:
|
| + .byte 0xf3,0xc3
|
| +.size bn_power5,.-bn_power5
|
| +
|
| +.globl bn_sqr8x_internal
|
| +.hidden bn_sqr8x_internal
|
| +.type bn_sqr8x_internal,@function
|
| +.align 32
|
| +bn_sqr8x_internal:
|
| +__bn_sqr8x_internal:
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| + leaq 32(%r10),%rbp
|
| + leaq (%rsi,%r9,1),%rsi
|
| +
|
| + movq %r9,%rcx
|
| +
|
| +
|
| + movq -32(%rsi,%rbp,1),%r14
|
| + leaq 48+8(%rsp,%r9,2),%rdi
|
| + movq -24(%rsi,%rbp,1),%rax
|
| + leaq -32(%rdi,%rbp,1),%rdi
|
| + movq -16(%rsi,%rbp,1),%rbx
|
| + movq %rax,%r15
|
| +
|
| + mulq %r14
|
| + movq %rax,%r10
|
| + movq %rbx,%rax
|
| + movq %rdx,%r11
|
| + movq %r10,-24(%rdi,%rbp,1)
|
| +
|
| + mulq %r14
|
| + addq %rax,%r11
|
| + movq %rbx,%rax
|
| + adcq $0,%rdx
|
| + movq %r11,-16(%rdi,%rbp,1)
|
| + movq %rdx,%r10
|
| +
|
| +
|
| + movq -8(%rsi,%rbp,1),%rbx
|
| + mulq %r15
|
| + movq %rax,%r12
|
| + movq %rbx,%rax
|
| + movq %rdx,%r13
|
| +
|
| + leaq (%rbp),%rcx
|
| + mulq %r14
|
| + addq %rax,%r10
|
| + movq %rbx,%rax
|
| + movq %rdx,%r11
|
| + adcq $0,%r11
|
| + addq %r12,%r10
|
| + adcq $0,%r11
|
| + movq %r10,-8(%rdi,%rcx,1)
|
| + jmp .Lsqr4x_1st
|
| +
|
| +.align 32
|
| +.Lsqr4x_1st:
|
| + movq (%rsi,%rcx,1),%rbx
|
| + mulq %r15
|
| + addq %rax,%r13
|
| + movq %rbx,%rax
|
| + movq %rdx,%r12
|
| + adcq $0,%r12
|
| +
|
| + mulq %r14
|
| + addq %rax,%r11
|
| + movq %rbx,%rax
|
| + movq 8(%rsi,%rcx,1),%rbx
|
| + movq %rdx,%r10
|
| + adcq $0,%r10
|
| + addq %r13,%r11
|
| + adcq $0,%r10
|
| +
|
| +
|
| + mulq %r15
|
| + addq %rax,%r12
|
| + movq %rbx,%rax
|
| + movq %r11,(%rdi,%rcx,1)
|
| + movq %rdx,%r13
|
| + adcq $0,%r13
|
| +
|
| + mulq %r14
|
| + addq %rax,%r10
|
| + movq %rbx,%rax
|
| + movq 16(%rsi,%rcx,1),%rbx
|
| + movq %rdx,%r11
|
| + adcq $0,%r11
|
| + addq %r12,%r10
|
| + adcq $0,%r11
|
| +
|
| + mulq %r15
|
| + addq %rax,%r13
|
| + movq %rbx,%rax
|
| + movq %r10,8(%rdi,%rcx,1)
|
| + movq %rdx,%r12
|
| + adcq $0,%r12
|
| +
|
| + mulq %r14
|
| + addq %rax,%r11
|
| + movq %rbx,%rax
|
| + movq 24(%rsi,%rcx,1),%rbx
|
| + movq %rdx,%r10
|
| + adcq $0,%r10
|
| + addq %r13,%r11
|
| + adcq $0,%r10
|
| +
|
| +
|
| + mulq %r15
|
| + addq %rax,%r12
|
| + movq %rbx,%rax
|
| + movq %r11,16(%rdi,%rcx,1)
|
| + movq %rdx,%r13
|
| + adcq $0,%r13
|
| + leaq 32(%rcx),%rcx
|
| +
|
| + mulq %r14
|
| + addq %rax,%r10
|
| + movq %rbx,%rax
|
| + movq %rdx,%r11
|
| + adcq $0,%r11
|
| + addq %r12,%r10
|
| + adcq $0,%r11
|
| + movq %r10,-8(%rdi,%rcx,1)
|
| +
|
| + cmpq $0,%rcx
|
| + jne .Lsqr4x_1st
|
| +
|
| + mulq %r15
|
| + addq %rax,%r13
|
| + leaq 16(%rbp),%rbp
|
| + adcq $0,%rdx
|
| + addq %r11,%r13
|
| + adcq $0,%rdx
|
| +
|
| + movq %r13,(%rdi)
|
| + movq %rdx,%r12
|
| + movq %rdx,8(%rdi)
|
| + jmp .Lsqr4x_outer
|
| +
|
| +.align 32
|
| +.Lsqr4x_outer:
|
| + movq -32(%rsi,%rbp,1),%r14
|
| + leaq 48+8(%rsp,%r9,2),%rdi
|
| + movq -24(%rsi,%rbp,1),%rax
|
| + leaq -32(%rdi,%rbp,1),%rdi
|
| + movq -16(%rsi,%rbp,1),%rbx
|
| + movq %rax,%r15
|
| +
|
| + mulq %r14
|
| + movq -24(%rdi,%rbp,1),%r10
|
| + addq %rax,%r10
|
| + movq %rbx,%rax
|
| + adcq $0,%rdx
|
| + movq %r10,-24(%rdi,%rbp,1)
|
| + movq %rdx,%r11
|
| +
|
| + mulq %r14
|
| + addq %rax,%r11
|
| + movq %rbx,%rax
|
| + adcq $0,%rdx
|
| + addq -16(%rdi,%rbp,1),%r11
|
| + movq %rdx,%r10
|
| + adcq $0,%r10
|
| + movq %r11,-16(%rdi,%rbp,1)
|
| +
|
| + xorq %r12,%r12
|
| +
|
| + movq -8(%rsi,%rbp,1),%rbx
|
| + mulq %r15
|
| + addq %rax,%r12
|
| + movq %rbx,%rax
|
| + adcq $0,%rdx
|
| + addq -8(%rdi,%rbp,1),%r12
|
| + movq %rdx,%r13
|
| + adcq $0,%r13
|
| +
|
| + mulq %r14
|
| + addq %rax,%r10
|
| + movq %rbx,%rax
|
| + adcq $0,%rdx
|
| + addq %r12,%r10
|
| + movq %rdx,%r11
|
| + adcq $0,%r11
|
| + movq %r10,-8(%rdi,%rbp,1)
|
| +
|
| + leaq (%rbp),%rcx
|
| + jmp .Lsqr4x_inner
|
| +
|
| +.align 32
|
| +.Lsqr4x_inner:
|
| + movq (%rsi,%rcx,1),%rbx
|
| + mulq %r15
|
| + addq %rax,%r13
|
| + movq %rbx,%rax
|
| + movq %rdx,%r12
|
| + adcq $0,%r12
|
| + addq (%rdi,%rcx,1),%r13
|
| + adcq $0,%r12
|
| +
|
| +.byte 0x67
|
| + mulq %r14
|
| + addq %rax,%r11
|
| + movq %rbx,%rax
|
| + movq 8(%rsi,%rcx,1),%rbx
|
| + movq %rdx,%r10
|
| + adcq $0,%r10
|
| + addq %r13,%r11
|
| + adcq $0,%r10
|
| +
|
| + mulq %r15
|
| + addq %rax,%r12
|
| + movq %r11,(%rdi,%rcx,1)
|
| + movq %rbx,%rax
|
| + movq %rdx,%r13
|
| + adcq $0,%r13
|
| + addq 8(%rdi,%rcx,1),%r12
|
| + leaq 16(%rcx),%rcx
|
| + adcq $0,%r13
|
| +
|
| + mulq %r14
|
| + addq %rax,%r10
|
| + movq %rbx,%rax
|
| + adcq $0,%rdx
|
| + addq %r12,%r10
|
| + movq %rdx,%r11
|
| + adcq $0,%r11
|
| + movq %r10,-8(%rdi,%rcx,1)
|
| +
|
| + cmpq $0,%rcx
|
| + jne .Lsqr4x_inner
|
| +
|
| +.byte 0x67
|
| + mulq %r15
|
| + addq %rax,%r13
|
| + adcq $0,%rdx
|
| + addq %r11,%r13
|
| + adcq $0,%rdx
|
| +
|
| + movq %r13,(%rdi)
|
| + movq %rdx,%r12
|
| + movq %rdx,8(%rdi)
|
| +
|
| + addq $16,%rbp
|
| + jnz .Lsqr4x_outer
|
| +
|
| +
|
| + movq -32(%rsi),%r14
|
| + leaq 48+8(%rsp,%r9,2),%rdi
|
| + movq -24(%rsi),%rax
|
| + leaq -32(%rdi,%rbp,1),%rdi
|
| + movq -16(%rsi),%rbx
|
| + movq %rax,%r15
|
| +
|
| + mulq %r14
|
| + addq %rax,%r10
|
| + movq %rbx,%rax
|
| + movq %rdx,%r11
|
| + adcq $0,%r11
|
| +
|
| + mulq %r14
|
| + addq %rax,%r11
|
| + movq %rbx,%rax
|
| + movq %r10,-24(%rdi)
|
| + movq %rdx,%r10
|
| + adcq $0,%r10
|
| + addq %r13,%r11
|
| + movq -8(%rsi),%rbx
|
| + adcq $0,%r10
|
| +
|
| + mulq %r15
|
| + addq %rax,%r12
|
| + movq %rbx,%rax
|
| + movq %r11,-16(%rdi)
|
| + movq %rdx,%r13
|
| + adcq $0,%r13
|
| +
|
| + mulq %r14
|
| + addq %rax,%r10
|
| + movq %rbx,%rax
|
| + movq %rdx,%r11
|
| + adcq $0,%r11
|
| + addq %r12,%r10
|
| + adcq $0,%r11
|
| + movq %r10,-8(%rdi)
|
| +
|
| + mulq %r15
|
| + addq %rax,%r13
|
| + movq -16(%rsi),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%r13
|
| + adcq $0,%rdx
|
| +
|
| + movq %r13,(%rdi)
|
| + movq %rdx,%r12
|
| + movq %rdx,8(%rdi)
|
| +
|
| + mulq %rbx
|
| + addq $16,%rbp
|
| + xorq %r14,%r14
|
| + subq %r9,%rbp
|
| + xorq %r15,%r15
|
| +
|
| + addq %r12,%rax
|
| + adcq $0,%rdx
|
| + movq %rax,8(%rdi)
|
| + movq %rdx,16(%rdi)
|
| + movq %r15,24(%rdi)
|
| +
|
| + movq -16(%rsi,%rbp,1),%rax
|
| + leaq 48+8(%rsp),%rdi
|
| + xorq %r10,%r10
|
| + movq 8(%rdi),%r11
|
| +
|
| + leaq (%r14,%r10,2),%r12
|
| + shrq $63,%r10
|
| + leaq (%rcx,%r11,2),%r13
|
| + shrq $63,%r11
|
| + orq %r10,%r13
|
| + movq 16(%rdi),%r10
|
| + movq %r11,%r14
|
| + mulq %rax
|
| + negq %r15
|
| + movq 24(%rdi),%r11
|
| + adcq %rax,%r12
|
| + movq -8(%rsi,%rbp,1),%rax
|
| + movq %r12,(%rdi)
|
| + adcq %rdx,%r13
|
| +
|
| + leaq (%r14,%r10,2),%rbx
|
| + movq %r13,8(%rdi)
|
| + sbbq %r15,%r15
|
| + shrq $63,%r10
|
| + leaq (%rcx,%r11,2),%r8
|
| + shrq $63,%r11
|
| + orq %r10,%r8
|
| + movq 32(%rdi),%r10
|
| + movq %r11,%r14
|
| + mulq %rax
|
| + negq %r15
|
| + movq 40(%rdi),%r11
|
| + adcq %rax,%rbx
|
| + movq 0(%rsi,%rbp,1),%rax
|
| + movq %rbx,16(%rdi)
|
| + adcq %rdx,%r8
|
| + leaq 16(%rbp),%rbp
|
| + movq %r8,24(%rdi)
|
| + sbbq %r15,%r15
|
| + leaq 64(%rdi),%rdi
|
| + jmp .Lsqr4x_shift_n_add
|
| +
|
| +.align 32
|
| +.Lsqr4x_shift_n_add:
|
| + leaq (%r14,%r10,2),%r12
|
| + shrq $63,%r10
|
| + leaq (%rcx,%r11,2),%r13
|
| + shrq $63,%r11
|
| + orq %r10,%r13
|
| + movq -16(%rdi),%r10
|
| + movq %r11,%r14
|
| + mulq %rax
|
| + negq %r15
|
| + movq -8(%rdi),%r11
|
| + adcq %rax,%r12
|
| + movq -8(%rsi,%rbp,1),%rax
|
| + movq %r12,-32(%rdi)
|
| + adcq %rdx,%r13
|
| +
|
| + leaq (%r14,%r10,2),%rbx
|
| + movq %r13,-24(%rdi)
|
| + sbbq %r15,%r15
|
| + shrq $63,%r10
|
| + leaq (%rcx,%r11,2),%r8
|
| + shrq $63,%r11
|
| + orq %r10,%r8
|
| + movq 0(%rdi),%r10
|
| + movq %r11,%r14
|
| + mulq %rax
|
| + negq %r15
|
| + movq 8(%rdi),%r11
|
| + adcq %rax,%rbx
|
| + movq 0(%rsi,%rbp,1),%rax
|
| + movq %rbx,-16(%rdi)
|
| + adcq %rdx,%r8
|
| +
|
| + leaq (%r14,%r10,2),%r12
|
| + movq %r8,-8(%rdi)
|
| + sbbq %r15,%r15
|
| + shrq $63,%r10
|
| + leaq (%rcx,%r11,2),%r13
|
| + shrq $63,%r11
|
| + orq %r10,%r13
|
| + movq 16(%rdi),%r10
|
| + movq %r11,%r14
|
| + mulq %rax
|
| + negq %r15
|
| + movq 24(%rdi),%r11
|
| + adcq %rax,%r12
|
| + movq 8(%rsi,%rbp,1),%rax
|
| + movq %r12,0(%rdi)
|
| + adcq %rdx,%r13
|
| +
|
| + leaq (%r14,%r10,2),%rbx
|
| + movq %r13,8(%rdi)
|
| + sbbq %r15,%r15
|
| + shrq $63,%r10
|
| + leaq (%rcx,%r11,2),%r8
|
| + shrq $63,%r11
|
| + orq %r10,%r8
|
| + movq 32(%rdi),%r10
|
| + movq %r11,%r14
|
| + mulq %rax
|
| + negq %r15
|
| + movq 40(%rdi),%r11
|
| + adcq %rax,%rbx
|
| + movq 16(%rsi,%rbp,1),%rax
|
| + movq %rbx,16(%rdi)
|
| + adcq %rdx,%r8
|
| + movq %r8,24(%rdi)
|
| + sbbq %r15,%r15
|
| + leaq 64(%rdi),%rdi
|
| + addq $32,%rbp
|
| + jnz .Lsqr4x_shift_n_add
|
| +
|
| + leaq (%r14,%r10,2),%r12
|
| +.byte 0x67
|
| + shrq $63,%r10
|
| + leaq (%rcx,%r11,2),%r13
|
| + shrq $63,%r11
|
| + orq %r10,%r13
|
| + movq -16(%rdi),%r10
|
| + movq %r11,%r14
|
| + mulq %rax
|
| + negq %r15
|
| + movq -8(%rdi),%r11
|
| + adcq %rax,%r12
|
| + movq -8(%rsi),%rax
|
| + movq %r12,-32(%rdi)
|
| + adcq %rdx,%r13
|
| +
|
| + leaq (%r14,%r10,2),%rbx
|
| + movq %r13,-24(%rdi)
|
| + sbbq %r15,%r15
|
| + shrq $63,%r10
|
| + leaq (%rcx,%r11,2),%r8
|
| + shrq $63,%r11
|
| + orq %r10,%r8
|
| + mulq %rax
|
| + negq %r15
|
| + adcq %rax,%rbx
|
| + adcq %rdx,%r8
|
| + movq %rbx,-16(%rdi)
|
| + movq %r8,-8(%rdi)
|
| +.byte 102,72,15,126,213
|
| +sqr8x_reduction:
|
| + xorq %rax,%rax
|
| + leaq (%rbp,%r9,2),%rcx
|
| + leaq 48+8(%rsp,%r9,2),%rdx
|
| + movq %rcx,0+8(%rsp)
|
| + leaq 48+8(%rsp,%r9,1),%rdi
|
| + movq %rdx,8+8(%rsp)
|
| + negq %r9
|
| + jmp .L8x_reduction_loop
|
| +
|
| +.align 32
|
| +.L8x_reduction_loop:
|
| + leaq (%rdi,%r9,1),%rdi
|
| +.byte 0x66
|
| + movq 0(%rdi),%rbx
|
| + movq 8(%rdi),%r9
|
| + movq 16(%rdi),%r10
|
| + movq 24(%rdi),%r11
|
| + movq 32(%rdi),%r12
|
| + movq 40(%rdi),%r13
|
| + movq 48(%rdi),%r14
|
| + movq 56(%rdi),%r15
|
| + movq %rax,(%rdx)
|
| + leaq 64(%rdi),%rdi
|
| +
|
| +.byte 0x67
|
| + movq %rbx,%r8
|
| + imulq 32+8(%rsp),%rbx
|
| + movq 0(%rbp),%rax
|
| + movl $8,%ecx
|
| + jmp .L8x_reduce
|
| +
|
| +.align 32
|
| +.L8x_reduce:
|
| + mulq %rbx
|
| + movq 16(%rbp),%rax
|
| + negq %r8
|
| + movq %rdx,%r8
|
| + adcq $0,%r8
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r9
|
| + movq 32(%rbp),%rax
|
| + adcq $0,%rdx
|
| + addq %r9,%r8
|
| + movq %rbx,48-8+8(%rsp,%rcx,8)
|
| + movq %rdx,%r9
|
| + adcq $0,%r9
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r10
|
| + movq 48(%rbp),%rax
|
| + adcq $0,%rdx
|
| + addq %r10,%r9
|
| + movq 32+8(%rsp),%rsi
|
| + movq %rdx,%r10
|
| + adcq $0,%r10
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r11
|
| + movq 64(%rbp),%rax
|
| + adcq $0,%rdx
|
| + imulq %r8,%rsi
|
| + addq %r11,%r10
|
| + movq %rdx,%r11
|
| + adcq $0,%r11
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r12
|
| + movq 80(%rbp),%rax
|
| + adcq $0,%rdx
|
| + addq %r12,%r11
|
| + movq %rdx,%r12
|
| + adcq $0,%r12
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r13
|
| + movq 96(%rbp),%rax
|
| + adcq $0,%rdx
|
| + addq %r13,%r12
|
| + movq %rdx,%r13
|
| + adcq $0,%r13
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r14
|
| + movq 112(%rbp),%rax
|
| + adcq $0,%rdx
|
| + addq %r14,%r13
|
| + movq %rdx,%r14
|
| + adcq $0,%r14
|
| +
|
| + mulq %rbx
|
| + movq %rsi,%rbx
|
| + addq %rax,%r15
|
| + movq 0(%rbp),%rax
|
| + adcq $0,%rdx
|
| + addq %r15,%r14
|
| + movq %rdx,%r15
|
| + adcq $0,%r15
|
| +
|
| + decl %ecx
|
| + jnz .L8x_reduce
|
| +
|
| + leaq 128(%rbp),%rbp
|
| + xorq %rax,%rax
|
| + movq 8+8(%rsp),%rdx
|
| + cmpq 0+8(%rsp),%rbp
|
| + jae .L8x_no_tail
|
| +
|
| +.byte 0x66
|
| + addq 0(%rdi),%r8
|
| + adcq 8(%rdi),%r9
|
| + adcq 16(%rdi),%r10
|
| + adcq 24(%rdi),%r11
|
| + adcq 32(%rdi),%r12
|
| + adcq 40(%rdi),%r13
|
| + adcq 48(%rdi),%r14
|
| + adcq 56(%rdi),%r15
|
| + sbbq %rsi,%rsi
|
| +
|
| + movq 48+56+8(%rsp),%rbx
|
| + movl $8,%ecx
|
| + movq 0(%rbp),%rax
|
| + jmp .L8x_tail
|
| +
|
| +.align 32
|
| +.L8x_tail:
|
| + mulq %rbx
|
| + addq %rax,%r8
|
| + movq 16(%rbp),%rax
|
| + movq %r8,(%rdi)
|
| + movq %rdx,%r8
|
| + adcq $0,%r8
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r9
|
| + movq 32(%rbp),%rax
|
| + adcq $0,%rdx
|
| + addq %r9,%r8
|
| + leaq 8(%rdi),%rdi
|
| + movq %rdx,%r9
|
| + adcq $0,%r9
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r10
|
| + movq 48(%rbp),%rax
|
| + adcq $0,%rdx
|
| + addq %r10,%r9
|
| + movq %rdx,%r10
|
| + adcq $0,%r10
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r11
|
| + movq 64(%rbp),%rax
|
| + adcq $0,%rdx
|
| + addq %r11,%r10
|
| + movq %rdx,%r11
|
| + adcq $0,%r11
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r12
|
| + movq 80(%rbp),%rax
|
| + adcq $0,%rdx
|
| + addq %r12,%r11
|
| + movq %rdx,%r12
|
| + adcq $0,%r12
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r13
|
| + movq 96(%rbp),%rax
|
| + adcq $0,%rdx
|
| + addq %r13,%r12
|
| + movq %rdx,%r13
|
| + adcq $0,%r13
|
| +
|
| + mulq %rbx
|
| + addq %rax,%r14
|
| + movq 112(%rbp),%rax
|
| + adcq $0,%rdx
|
| + addq %r14,%r13
|
| + movq %rdx,%r14
|
| + adcq $0,%r14
|
| +
|
| + mulq %rbx
|
| + movq 48-16+8(%rsp,%rcx,8),%rbx
|
| + addq %rax,%r15
|
| + adcq $0,%rdx
|
| + addq %r15,%r14
|
| + movq 0(%rbp),%rax
|
| + movq %rdx,%r15
|
| + adcq $0,%r15
|
| +
|
| + decl %ecx
|
| + jnz .L8x_tail
|
| +
|
| + leaq 128(%rbp),%rbp
|
| + movq 8+8(%rsp),%rdx
|
| + cmpq 0+8(%rsp),%rbp
|
| + jae .L8x_tail_done
|
| +
|
| + movq 48+56+8(%rsp),%rbx
|
| + negq %rsi
|
| + movq 0(%rbp),%rax
|
| + adcq 0(%rdi),%r8
|
| + adcq 8(%rdi),%r9
|
| + adcq 16(%rdi),%r10
|
| + adcq 24(%rdi),%r11
|
| + adcq 32(%rdi),%r12
|
| + adcq 40(%rdi),%r13
|
| + adcq 48(%rdi),%r14
|
| + adcq 56(%rdi),%r15
|
| + sbbq %rsi,%rsi
|
| +
|
| + movl $8,%ecx
|
| + jmp .L8x_tail
|
| +
|
| +.align 32
|
| +.L8x_tail_done:
|
| + addq (%rdx),%r8
|
| + xorq %rax,%rax
|
| +
|
| + negq %rsi
|
| +.L8x_no_tail:
|
| + adcq 0(%rdi),%r8
|
| + adcq 8(%rdi),%r9
|
| + adcq 16(%rdi),%r10
|
| + adcq 24(%rdi),%r11
|
| + adcq 32(%rdi),%r12
|
| + adcq 40(%rdi),%r13
|
| + adcq 48(%rdi),%r14
|
| + adcq 56(%rdi),%r15
|
| + adcq $0,%rax
|
| + movq -16(%rbp),%rcx
|
| + xorq %rsi,%rsi
|
| +
|
| +.byte 102,72,15,126,213
|
| +
|
| + movq %r8,0(%rdi)
|
| + movq %r9,8(%rdi)
|
| +.byte 102,73,15,126,217
|
| + movq %r10,16(%rdi)
|
| + movq %r11,24(%rdi)
|
| + movq %r12,32(%rdi)
|
| + movq %r13,40(%rdi)
|
| + movq %r14,48(%rdi)
|
| + movq %r15,56(%rdi)
|
| + leaq 64(%rdi),%rdi
|
| +
|
| + cmpq %rdx,%rdi
|
| + jb .L8x_reduction_loop
|
| +
|
| + subq %r15,%rcx
|
| + leaq (%rdi,%r9,1),%rbx
|
| + adcq %rsi,%rsi
|
| + movq %r9,%rcx
|
| + orq %rsi,%rax
|
| +.byte 102,72,15,126,207
|
| + xorq $1,%rax
|
| +.byte 102,72,15,126,206
|
| + leaq (%rbp,%rax,8),%rbp
|
| + sarq $3+2,%rcx
|
| + jmp .Lsqr4x_sub
|
| +
|
| +.align 32
|
| +.Lsqr4x_sub:
|
| +.byte 0x66
|
| + movq 0(%rbx),%r12
|
| + movq 8(%rbx),%r13
|
| + sbbq 0(%rbp),%r12
|
| + movq 16(%rbx),%r14
|
| + sbbq 16(%rbp),%r13
|
| + movq 24(%rbx),%r15
|
| + leaq 32(%rbx),%rbx
|
| + sbbq 32(%rbp),%r14
|
| + movq %r12,0(%rdi)
|
| + sbbq 48(%rbp),%r15
|
| + leaq 64(%rbp),%rbp
|
| + movq %r13,8(%rdi)
|
| + movq %r14,16(%rdi)
|
| + movq %r15,24(%rdi)
|
| + leaq 32(%rdi),%rdi
|
| +
|
| + incq %rcx
|
| + jnz .Lsqr4x_sub
|
| + movq %r9,%r10
|
| + negq %r9
|
| + .byte 0xf3,0xc3
|
| +.size bn_sqr8x_internal,.-bn_sqr8x_internal
|
| +.globl bn_from_montgomery
|
| +.type bn_from_montgomery,@function
|
| +.align 32
|
| +bn_from_montgomery:
|
| + testl $7,%r9d
|
| + jz bn_from_mont8x
|
| + xorl %eax,%eax
|
| + .byte 0xf3,0xc3
|
| +.size bn_from_montgomery,.-bn_from_montgomery
|
| +
|
| +.type bn_from_mont8x,@function
|
| +.align 32
|
| +bn_from_mont8x:
|
| +.byte 0x67
|
| + movq %rsp,%rax
|
| + pushq %rbx
|
| + pushq %rbp
|
| + pushq %r12
|
| + pushq %r13
|
| + pushq %r14
|
| + pushq %r15
|
| +.byte 0x67
|
| + movl %r9d,%r10d
|
| + shll $3,%r9d
|
| + shll $3+2,%r10d
|
| + negq %r9
|
| + movq (%r8),%r8
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| + leaq -64(%rsp,%r9,2),%r11
|
| + subq %rsi,%r11
|
| + andq $4095,%r11
|
| + cmpq %r11,%r10
|
| + jb .Lfrom_sp_alt
|
| + subq %r11,%rsp
|
| + leaq -64(%rsp,%r9,2),%rsp
|
| + jmp .Lfrom_sp_done
|
| +
|
| +.align 32
|
| +.Lfrom_sp_alt:
|
| + leaq 4096-64(,%r9,2),%r10
|
| + leaq -64(%rsp,%r9,2),%rsp
|
| + subq %r10,%r11
|
| + movq $0,%r10
|
| + cmovcq %r10,%r11
|
| + subq %r11,%rsp
|
| +.Lfrom_sp_done:
|
| + andq $-64,%rsp
|
| + movq %r9,%r10
|
| + negq %r9
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| +
|
| + movq %r8,32(%rsp)
|
| + movq %rax,40(%rsp)
|
| +.Lfrom_body:
|
| + movq %r9,%r11
|
| + leaq 48(%rsp),%rax
|
| + pxor %xmm0,%xmm0
|
| + jmp .Lmul_by_1
|
| +
|
| +.align 32
|
| +.Lmul_by_1:
|
| + movdqu (%rsi),%xmm1
|
| + movdqu 16(%rsi),%xmm2
|
| + movdqu 32(%rsi),%xmm3
|
| + movdqa %xmm0,(%rax,%r9,1)
|
| + movdqu 48(%rsi),%xmm4
|
| + movdqa %xmm0,16(%rax,%r9,1)
|
| +.byte 0x48,0x8d,0xb6,0x40,0x00,0x00,0x00
|
| + movdqa %xmm1,(%rax)
|
| + movdqa %xmm0,32(%rax,%r9,1)
|
| + movdqa %xmm2,16(%rax)
|
| + movdqa %xmm0,48(%rax,%r9,1)
|
| + movdqa %xmm3,32(%rax)
|
| + movdqa %xmm4,48(%rax)
|
| + leaq 64(%rax),%rax
|
| + subq $64,%r11
|
| + jnz .Lmul_by_1
|
| +
|
| +.byte 102,72,15,110,207
|
| +.byte 102,72,15,110,209
|
| +.byte 0x67
|
| + movq %rcx,%rbp
|
| +.byte 102,73,15,110,218
|
| + call sqr8x_reduction
|
| +
|
| + pxor %xmm0,%xmm0
|
| + leaq 48(%rsp),%rax
|
| + movq 40(%rsp),%rsi
|
| + jmp .Lfrom_mont_zero
|
| +
|
| +.align 32
|
| +.Lfrom_mont_zero:
|
| + movdqa %xmm0,0(%rax)
|
| + movdqa %xmm0,16(%rax)
|
| + movdqa %xmm0,32(%rax)
|
| + movdqa %xmm0,48(%rax)
|
| + leaq 64(%rax),%rax
|
| + subq $32,%r9
|
| + jnz .Lfrom_mont_zero
|
| +
|
| + movq $1,%rax
|
| + movq -48(%rsi),%r15
|
| + movq -40(%rsi),%r14
|
| + movq -32(%rsi),%r13
|
| + movq -24(%rsi),%r12
|
| + movq -16(%rsi),%rbp
|
| + movq -8(%rsi),%rbx
|
| + leaq (%rsi),%rsp
|
| +.Lfrom_epilogue:
|
| + .byte 0xf3,0xc3
|
| +.size bn_from_mont8x,.-bn_from_mont8x
|
| +.globl bn_get_bits5
|
| +.type bn_get_bits5,@function
|
| +.align 16
|
| +bn_get_bits5:
|
| + movq %rdi,%r10
|
| + movl %esi,%ecx
|
| + shrl $3,%esi
|
| + movzwl (%r10,%rsi,1),%eax
|
| + andl $7,%ecx
|
| + shrl %cl,%eax
|
| + andl $31,%eax
|
| + .byte 0xf3,0xc3
|
| +.size bn_get_bits5,.-bn_get_bits5
|
| +
|
| +.globl bn_scatter5
|
| +.type bn_scatter5,@function
|
| +.align 16
|
| +bn_scatter5:
|
| + cmpl $0,%esi
|
| + jz .Lscatter_epilogue
|
| + leaq (%rdx,%rcx,8),%rdx
|
| +.Lscatter:
|
| + movq (%rdi),%rax
|
| + leaq 8(%rdi),%rdi
|
| + movq %rax,(%rdx)
|
| + leaq 256(%rdx),%rdx
|
| + subl $1,%esi
|
| + jnz .Lscatter
|
| +.Lscatter_epilogue:
|
| + .byte 0xf3,0xc3
|
| +.size bn_scatter5,.-bn_scatter5
|
| +
|
| +.globl bn_gather5
|
| +.type bn_gather5,@function
|
| +.align 16
|
| +bn_gather5:
|
| + movl %ecx,%r11d
|
| + shrl $3,%ecx
|
| + andq $7,%r11
|
| + notl %ecx
|
| + leaq .Lmagic_masks(%rip),%rax
|
| + andl $3,%ecx
|
| + leaq 128(%rdx,%r11,8),%rdx
|
| + movq 0(%rax,%rcx,8),%xmm4
|
| + movq 8(%rax,%rcx,8),%xmm5
|
| + movq 16(%rax,%rcx,8),%xmm6
|
| + movq 24(%rax,%rcx,8),%xmm7
|
| + jmp .Lgather
|
| +.align 16
|
| +.Lgather:
|
| + movq -128(%rdx),%xmm0
|
| + movq -64(%rdx),%xmm1
|
| + pand %xmm4,%xmm0
|
| + movq 0(%rdx),%xmm2
|
| + pand %xmm5,%xmm1
|
| + movq 64(%rdx),%xmm3
|
| + pand %xmm6,%xmm2
|
| + por %xmm1,%xmm0
|
| + pand %xmm7,%xmm3
|
| +.byte 0x67,0x67
|
| + por %xmm2,%xmm0
|
| + leaq 256(%rdx),%rdx
|
| + por %xmm3,%xmm0
|
| +
|
| + movq %xmm0,(%rdi)
|
| + leaq 8(%rdi),%rdi
|
| + subl $1,%esi
|
| + jnz .Lgather
|
| + .byte 0xf3,0xc3
|
| +.LSEH_end_bn_gather5:
|
| +.size bn_gather5,.-bn_gather5
|
| +.align 64
|
| +.Lmagic_masks:
|
| +.long 0,0, 0,0, 0,0, -1,-1
|
| +.long 0,0, 0,0, 0,0, 0,0
|
| +.byte 77,111,110,116,103,111,109,101,114,121,32,77,117,108,116,105,112,108,105,99,97,116,105,111,110,32,119,105,116,104,32,115,99,97,116,116,101,114,47,103,97,116,104,101,114,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,103,62,0
|
| +#endif
|
|
|