Index: third_party/boringssl/linux-x86_64/crypto/bn/x86_64-mont5.S |
diff --git a/third_party/boringssl/linux-x86_64/crypto/bn/x86_64-mont5.S b/third_party/boringssl/linux-x86_64/crypto/bn/x86_64-mont5.S |
deleted file mode 100644 |
index 554df1ffac1cf0806561810285c16cb5229dd364..0000000000000000000000000000000000000000 |
--- a/third_party/boringssl/linux-x86_64/crypto/bn/x86_64-mont5.S |
+++ /dev/null |
@@ -1,2241 +0,0 @@ |
-#if defined(__x86_64__) |
-.text |
- |
-.extern OPENSSL_ia32cap_P |
-.hidden OPENSSL_ia32cap_P |
- |
-.globl bn_mul_mont_gather5 |
-.hidden bn_mul_mont_gather5 |
-.type bn_mul_mont_gather5,@function |
-.align 64 |
-bn_mul_mont_gather5: |
- testl $7,%r9d |
- jnz .Lmul_enter |
- jmp .Lmul4x_enter |
- |
-.align 16 |
-.Lmul_enter: |
- movl %r9d,%r9d |
- movq %rsp,%rax |
- movd 8(%rsp),%xmm5 |
- leaq .Linc(%rip),%r10 |
- pushq %rbx |
- pushq %rbp |
- pushq %r12 |
- pushq %r13 |
- pushq %r14 |
- pushq %r15 |
- |
- leaq 2(%r9),%r11 |
- negq %r11 |
- leaq -264(%rsp,%r11,8),%rsp |
- andq $-1024,%rsp |
- |
- movq %rax,8(%rsp,%r9,8) |
-.Lmul_body: |
- leaq 128(%rdx),%r12 |
- movdqa 0(%r10),%xmm0 |
- movdqa 16(%r10),%xmm1 |
- leaq 24-112(%rsp,%r9,8),%r10 |
- andq $-16,%r10 |
- |
- pshufd $0,%xmm5,%xmm5 |
- movdqa %xmm1,%xmm4 |
- movdqa %xmm1,%xmm2 |
- paddd %xmm0,%xmm1 |
- pcmpeqd %xmm5,%xmm0 |
-.byte 0x67 |
- movdqa %xmm4,%xmm3 |
- paddd %xmm1,%xmm2 |
- pcmpeqd %xmm5,%xmm1 |
- movdqa %xmm0,112(%r10) |
- movdqa %xmm4,%xmm0 |
- |
- paddd %xmm2,%xmm3 |
- pcmpeqd %xmm5,%xmm2 |
- movdqa %xmm1,128(%r10) |
- movdqa %xmm4,%xmm1 |
- |
- paddd %xmm3,%xmm0 |
- pcmpeqd %xmm5,%xmm3 |
- movdqa %xmm2,144(%r10) |
- movdqa %xmm4,%xmm2 |
- |
- paddd %xmm0,%xmm1 |
- pcmpeqd %xmm5,%xmm0 |
- movdqa %xmm3,160(%r10) |
- movdqa %xmm4,%xmm3 |
- paddd %xmm1,%xmm2 |
- pcmpeqd %xmm5,%xmm1 |
- movdqa %xmm0,176(%r10) |
- movdqa %xmm4,%xmm0 |
- |
- paddd %xmm2,%xmm3 |
- pcmpeqd %xmm5,%xmm2 |
- movdqa %xmm1,192(%r10) |
- movdqa %xmm4,%xmm1 |
- |
- paddd %xmm3,%xmm0 |
- pcmpeqd %xmm5,%xmm3 |
- movdqa %xmm2,208(%r10) |
- movdqa %xmm4,%xmm2 |
- |
- paddd %xmm0,%xmm1 |
- pcmpeqd %xmm5,%xmm0 |
- movdqa %xmm3,224(%r10) |
- movdqa %xmm4,%xmm3 |
- paddd %xmm1,%xmm2 |
- pcmpeqd %xmm5,%xmm1 |
- movdqa %xmm0,240(%r10) |
- movdqa %xmm4,%xmm0 |
- |
- paddd %xmm2,%xmm3 |
- pcmpeqd %xmm5,%xmm2 |
- movdqa %xmm1,256(%r10) |
- movdqa %xmm4,%xmm1 |
- |
- paddd %xmm3,%xmm0 |
- pcmpeqd %xmm5,%xmm3 |
- movdqa %xmm2,272(%r10) |
- movdqa %xmm4,%xmm2 |
- |
- paddd %xmm0,%xmm1 |
- pcmpeqd %xmm5,%xmm0 |
- movdqa %xmm3,288(%r10) |
- movdqa %xmm4,%xmm3 |
- paddd %xmm1,%xmm2 |
- pcmpeqd %xmm5,%xmm1 |
- movdqa %xmm0,304(%r10) |
- |
- paddd %xmm2,%xmm3 |
-.byte 0x67 |
- pcmpeqd %xmm5,%xmm2 |
- movdqa %xmm1,320(%r10) |
- |
- pcmpeqd %xmm5,%xmm3 |
- movdqa %xmm2,336(%r10) |
- pand 64(%r12),%xmm0 |
- |
- pand 80(%r12),%xmm1 |
- pand 96(%r12),%xmm2 |
- movdqa %xmm3,352(%r10) |
- pand 112(%r12),%xmm3 |
- por %xmm2,%xmm0 |
- por %xmm3,%xmm1 |
- movdqa -128(%r12),%xmm4 |
- movdqa -112(%r12),%xmm5 |
- movdqa -96(%r12),%xmm2 |
- pand 112(%r10),%xmm4 |
- movdqa -80(%r12),%xmm3 |
- pand 128(%r10),%xmm5 |
- por %xmm4,%xmm0 |
- pand 144(%r10),%xmm2 |
- por %xmm5,%xmm1 |
- pand 160(%r10),%xmm3 |
- por %xmm2,%xmm0 |
- por %xmm3,%xmm1 |
- movdqa -64(%r12),%xmm4 |
- movdqa -48(%r12),%xmm5 |
- movdqa -32(%r12),%xmm2 |
- pand 176(%r10),%xmm4 |
- movdqa -16(%r12),%xmm3 |
- pand 192(%r10),%xmm5 |
- por %xmm4,%xmm0 |
- pand 208(%r10),%xmm2 |
- por %xmm5,%xmm1 |
- pand 224(%r10),%xmm3 |
- por %xmm2,%xmm0 |
- por %xmm3,%xmm1 |
- movdqa 0(%r12),%xmm4 |
- movdqa 16(%r12),%xmm5 |
- movdqa 32(%r12),%xmm2 |
- pand 240(%r10),%xmm4 |
- movdqa 48(%r12),%xmm3 |
- pand 256(%r10),%xmm5 |
- por %xmm4,%xmm0 |
- pand 272(%r10),%xmm2 |
- por %xmm5,%xmm1 |
- pand 288(%r10),%xmm3 |
- por %xmm2,%xmm0 |
- por %xmm3,%xmm1 |
- por %xmm1,%xmm0 |
- pshufd $0x4e,%xmm0,%xmm1 |
- por %xmm1,%xmm0 |
- leaq 256(%r12),%r12 |
-.byte 102,72,15,126,195 |
- |
- movq (%r8),%r8 |
- movq (%rsi),%rax |
- |
- xorq %r14,%r14 |
- xorq %r15,%r15 |
- |
- movq %r8,%rbp |
- mulq %rbx |
- movq %rax,%r10 |
- movq (%rcx),%rax |
- |
- imulq %r10,%rbp |
- movq %rdx,%r11 |
- |
- mulq %rbp |
- addq %rax,%r10 |
- movq 8(%rsi),%rax |
- adcq $0,%rdx |
- movq %rdx,%r13 |
- |
- leaq 1(%r15),%r15 |
- jmp .L1st_enter |
- |
-.align 16 |
-.L1st: |
- addq %rax,%r13 |
- movq (%rsi,%r15,8),%rax |
- adcq $0,%rdx |
- addq %r11,%r13 |
- movq %r10,%r11 |
- adcq $0,%rdx |
- movq %r13,-16(%rsp,%r15,8) |
- movq %rdx,%r13 |
- |
-.L1st_enter: |
- mulq %rbx |
- addq %rax,%r11 |
- movq (%rcx,%r15,8),%rax |
- adcq $0,%rdx |
- leaq 1(%r15),%r15 |
- movq %rdx,%r10 |
- |
- mulq %rbp |
- cmpq %r9,%r15 |
- jne .L1st |
- |
- |
- addq %rax,%r13 |
- adcq $0,%rdx |
- addq %r11,%r13 |
- adcq $0,%rdx |
- movq %r13,-16(%rsp,%r9,8) |
- movq %rdx,%r13 |
- movq %r10,%r11 |
- |
- xorq %rdx,%rdx |
- addq %r11,%r13 |
- adcq $0,%rdx |
- movq %r13,-8(%rsp,%r9,8) |
- movq %rdx,(%rsp,%r9,8) |
- |
- leaq 1(%r14),%r14 |
- jmp .Louter |
-.align 16 |
-.Louter: |
- leaq 24+128(%rsp,%r9,8),%rdx |
- andq $-16,%rdx |
- pxor %xmm4,%xmm4 |
- pxor %xmm5,%xmm5 |
- movdqa -128(%r12),%xmm0 |
- movdqa -112(%r12),%xmm1 |
- movdqa -96(%r12),%xmm2 |
- movdqa -80(%r12),%xmm3 |
- pand -128(%rdx),%xmm0 |
- pand -112(%rdx),%xmm1 |
- por %xmm0,%xmm4 |
- pand -96(%rdx),%xmm2 |
- por %xmm1,%xmm5 |
- pand -80(%rdx),%xmm3 |
- por %xmm2,%xmm4 |
- por %xmm3,%xmm5 |
- movdqa -64(%r12),%xmm0 |
- movdqa -48(%r12),%xmm1 |
- movdqa -32(%r12),%xmm2 |
- movdqa -16(%r12),%xmm3 |
- pand -64(%rdx),%xmm0 |
- pand -48(%rdx),%xmm1 |
- por %xmm0,%xmm4 |
- pand -32(%rdx),%xmm2 |
- por %xmm1,%xmm5 |
- pand -16(%rdx),%xmm3 |
- por %xmm2,%xmm4 |
- por %xmm3,%xmm5 |
- movdqa 0(%r12),%xmm0 |
- movdqa 16(%r12),%xmm1 |
- movdqa 32(%r12),%xmm2 |
- movdqa 48(%r12),%xmm3 |
- pand 0(%rdx),%xmm0 |
- pand 16(%rdx),%xmm1 |
- por %xmm0,%xmm4 |
- pand 32(%rdx),%xmm2 |
- por %xmm1,%xmm5 |
- pand 48(%rdx),%xmm3 |
- por %xmm2,%xmm4 |
- por %xmm3,%xmm5 |
- movdqa 64(%r12),%xmm0 |
- movdqa 80(%r12),%xmm1 |
- movdqa 96(%r12),%xmm2 |
- movdqa 112(%r12),%xmm3 |
- pand 64(%rdx),%xmm0 |
- pand 80(%rdx),%xmm1 |
- por %xmm0,%xmm4 |
- pand 96(%rdx),%xmm2 |
- por %xmm1,%xmm5 |
- pand 112(%rdx),%xmm3 |
- por %xmm2,%xmm4 |
- por %xmm3,%xmm5 |
- por %xmm5,%xmm4 |
- pshufd $0x4e,%xmm4,%xmm0 |
- por %xmm4,%xmm0 |
- leaq 256(%r12),%r12 |
- |
- movq (%rsi),%rax |
-.byte 102,72,15,126,195 |
- |
- xorq %r15,%r15 |
- movq %r8,%rbp |
- movq (%rsp),%r10 |
- |
- mulq %rbx |
- addq %rax,%r10 |
- movq (%rcx),%rax |
- adcq $0,%rdx |
- |
- imulq %r10,%rbp |
- movq %rdx,%r11 |
- |
- mulq %rbp |
- addq %rax,%r10 |
- movq 8(%rsi),%rax |
- adcq $0,%rdx |
- movq 8(%rsp),%r10 |
- movq %rdx,%r13 |
- |
- leaq 1(%r15),%r15 |
- jmp .Linner_enter |
- |
-.align 16 |
-.Linner: |
- addq %rax,%r13 |
- movq (%rsi,%r15,8),%rax |
- adcq $0,%rdx |
- addq %r10,%r13 |
- movq (%rsp,%r15,8),%r10 |
- adcq $0,%rdx |
- movq %r13,-16(%rsp,%r15,8) |
- movq %rdx,%r13 |
- |
-.Linner_enter: |
- mulq %rbx |
- addq %rax,%r11 |
- movq (%rcx,%r15,8),%rax |
- adcq $0,%rdx |
- addq %r11,%r10 |
- movq %rdx,%r11 |
- adcq $0,%r11 |
- leaq 1(%r15),%r15 |
- |
- mulq %rbp |
- cmpq %r9,%r15 |
- jne .Linner |
- |
- addq %rax,%r13 |
- adcq $0,%rdx |
- addq %r10,%r13 |
- movq (%rsp,%r9,8),%r10 |
- adcq $0,%rdx |
- movq %r13,-16(%rsp,%r9,8) |
- movq %rdx,%r13 |
- |
- xorq %rdx,%rdx |
- addq %r11,%r13 |
- adcq $0,%rdx |
- addq %r10,%r13 |
- adcq $0,%rdx |
- movq %r13,-8(%rsp,%r9,8) |
- movq %rdx,(%rsp,%r9,8) |
- |
- leaq 1(%r14),%r14 |
- cmpq %r9,%r14 |
- jb .Louter |
- |
- xorq %r14,%r14 |
- movq (%rsp),%rax |
- leaq (%rsp),%rsi |
- movq %r9,%r15 |
- jmp .Lsub |
-.align 16 |
-.Lsub: sbbq (%rcx,%r14,8),%rax |
- movq %rax,(%rdi,%r14,8) |
- movq 8(%rsi,%r14,8),%rax |
- leaq 1(%r14),%r14 |
- decq %r15 |
- jnz .Lsub |
- |
- sbbq $0,%rax |
- xorq %r14,%r14 |
- movq %r9,%r15 |
-.align 16 |
-.Lcopy: |
- movq (%rsp,%r14,8),%rsi |
- movq (%rdi,%r14,8),%rcx |
- xorq %rcx,%rsi |
- andq %rax,%rsi |
- xorq %rcx,%rsi |
- movq %r14,(%rsp,%r14,8) |
- movq %rsi,(%rdi,%r14,8) |
- leaq 1(%r14),%r14 |
- subq $1,%r15 |
- jnz .Lcopy |
- |
- movq 8(%rsp,%r9,8),%rsi |
- movq $1,%rax |
- |
- movq -48(%rsi),%r15 |
- movq -40(%rsi),%r14 |
- movq -32(%rsi),%r13 |
- movq -24(%rsi),%r12 |
- movq -16(%rsi),%rbp |
- movq -8(%rsi),%rbx |
- leaq (%rsi),%rsp |
-.Lmul_epilogue: |
- .byte 0xf3,0xc3 |
-.size bn_mul_mont_gather5,.-bn_mul_mont_gather5 |
-.type bn_mul4x_mont_gather5,@function |
-.align 32 |
-bn_mul4x_mont_gather5: |
-.Lmul4x_enter: |
-.byte 0x67 |
- movq %rsp,%rax |
- pushq %rbx |
- pushq %rbp |
- pushq %r12 |
- pushq %r13 |
- pushq %r14 |
- pushq %r15 |
- |
-.byte 0x67 |
- shll $3,%r9d |
- leaq (%r9,%r9,2),%r10 |
- negq %r9 |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- leaq -320(%rsp,%r9,2),%r11 |
- subq %rdi,%r11 |
- andq $4095,%r11 |
- cmpq %r11,%r10 |
- jb .Lmul4xsp_alt |
- subq %r11,%rsp |
- leaq -320(%rsp,%r9,2),%rsp |
- jmp .Lmul4xsp_done |
- |
-.align 32 |
-.Lmul4xsp_alt: |
- leaq 4096-320(,%r9,2),%r10 |
- leaq -320(%rsp,%r9,2),%rsp |
- subq %r10,%r11 |
- movq $0,%r10 |
- cmovcq %r10,%r11 |
- subq %r11,%rsp |
-.Lmul4xsp_done: |
- andq $-64,%rsp |
- negq %r9 |
- |
- movq %rax,40(%rsp) |
-.Lmul4x_body: |
- |
- call mul4x_internal |
- |
- movq 40(%rsp),%rsi |
- movq $1,%rax |
- |
- movq -48(%rsi),%r15 |
- movq -40(%rsi),%r14 |
- movq -32(%rsi),%r13 |
- movq -24(%rsi),%r12 |
- movq -16(%rsi),%rbp |
- movq -8(%rsi),%rbx |
- leaq (%rsi),%rsp |
-.Lmul4x_epilogue: |
- .byte 0xf3,0xc3 |
-.size bn_mul4x_mont_gather5,.-bn_mul4x_mont_gather5 |
- |
-.type mul4x_internal,@function |
-.align 32 |
-mul4x_internal: |
- shlq $5,%r9 |
- movd 8(%rax),%xmm5 |
- leaq .Linc(%rip),%rax |
- leaq 128(%rdx,%r9,1),%r13 |
- shrq $5,%r9 |
- movdqa 0(%rax),%xmm0 |
- movdqa 16(%rax),%xmm1 |
- leaq 88-112(%rsp,%r9,1),%r10 |
- leaq 128(%rdx),%r12 |
- |
- pshufd $0,%xmm5,%xmm5 |
- movdqa %xmm1,%xmm4 |
-.byte 0x67,0x67 |
- movdqa %xmm1,%xmm2 |
- paddd %xmm0,%xmm1 |
- pcmpeqd %xmm5,%xmm0 |
-.byte 0x67 |
- movdqa %xmm4,%xmm3 |
- paddd %xmm1,%xmm2 |
- pcmpeqd %xmm5,%xmm1 |
- movdqa %xmm0,112(%r10) |
- movdqa %xmm4,%xmm0 |
- |
- paddd %xmm2,%xmm3 |
- pcmpeqd %xmm5,%xmm2 |
- movdqa %xmm1,128(%r10) |
- movdqa %xmm4,%xmm1 |
- |
- paddd %xmm3,%xmm0 |
- pcmpeqd %xmm5,%xmm3 |
- movdqa %xmm2,144(%r10) |
- movdqa %xmm4,%xmm2 |
- |
- paddd %xmm0,%xmm1 |
- pcmpeqd %xmm5,%xmm0 |
- movdqa %xmm3,160(%r10) |
- movdqa %xmm4,%xmm3 |
- paddd %xmm1,%xmm2 |
- pcmpeqd %xmm5,%xmm1 |
- movdqa %xmm0,176(%r10) |
- movdqa %xmm4,%xmm0 |
- |
- paddd %xmm2,%xmm3 |
- pcmpeqd %xmm5,%xmm2 |
- movdqa %xmm1,192(%r10) |
- movdqa %xmm4,%xmm1 |
- |
- paddd %xmm3,%xmm0 |
- pcmpeqd %xmm5,%xmm3 |
- movdqa %xmm2,208(%r10) |
- movdqa %xmm4,%xmm2 |
- |
- paddd %xmm0,%xmm1 |
- pcmpeqd %xmm5,%xmm0 |
- movdqa %xmm3,224(%r10) |
- movdqa %xmm4,%xmm3 |
- paddd %xmm1,%xmm2 |
- pcmpeqd %xmm5,%xmm1 |
- movdqa %xmm0,240(%r10) |
- movdqa %xmm4,%xmm0 |
- |
- paddd %xmm2,%xmm3 |
- pcmpeqd %xmm5,%xmm2 |
- movdqa %xmm1,256(%r10) |
- movdqa %xmm4,%xmm1 |
- |
- paddd %xmm3,%xmm0 |
- pcmpeqd %xmm5,%xmm3 |
- movdqa %xmm2,272(%r10) |
- movdqa %xmm4,%xmm2 |
- |
- paddd %xmm0,%xmm1 |
- pcmpeqd %xmm5,%xmm0 |
- movdqa %xmm3,288(%r10) |
- movdqa %xmm4,%xmm3 |
- paddd %xmm1,%xmm2 |
- pcmpeqd %xmm5,%xmm1 |
- movdqa %xmm0,304(%r10) |
- |
- paddd %xmm2,%xmm3 |
-.byte 0x67 |
- pcmpeqd %xmm5,%xmm2 |
- movdqa %xmm1,320(%r10) |
- |
- pcmpeqd %xmm5,%xmm3 |
- movdqa %xmm2,336(%r10) |
- pand 64(%r12),%xmm0 |
- |
- pand 80(%r12),%xmm1 |
- pand 96(%r12),%xmm2 |
- movdqa %xmm3,352(%r10) |
- pand 112(%r12),%xmm3 |
- por %xmm2,%xmm0 |
- por %xmm3,%xmm1 |
- movdqa -128(%r12),%xmm4 |
- movdqa -112(%r12),%xmm5 |
- movdqa -96(%r12),%xmm2 |
- pand 112(%r10),%xmm4 |
- movdqa -80(%r12),%xmm3 |
- pand 128(%r10),%xmm5 |
- por %xmm4,%xmm0 |
- pand 144(%r10),%xmm2 |
- por %xmm5,%xmm1 |
- pand 160(%r10),%xmm3 |
- por %xmm2,%xmm0 |
- por %xmm3,%xmm1 |
- movdqa -64(%r12),%xmm4 |
- movdqa -48(%r12),%xmm5 |
- movdqa -32(%r12),%xmm2 |
- pand 176(%r10),%xmm4 |
- movdqa -16(%r12),%xmm3 |
- pand 192(%r10),%xmm5 |
- por %xmm4,%xmm0 |
- pand 208(%r10),%xmm2 |
- por %xmm5,%xmm1 |
- pand 224(%r10),%xmm3 |
- por %xmm2,%xmm0 |
- por %xmm3,%xmm1 |
- movdqa 0(%r12),%xmm4 |
- movdqa 16(%r12),%xmm5 |
- movdqa 32(%r12),%xmm2 |
- pand 240(%r10),%xmm4 |
- movdqa 48(%r12),%xmm3 |
- pand 256(%r10),%xmm5 |
- por %xmm4,%xmm0 |
- pand 272(%r10),%xmm2 |
- por %xmm5,%xmm1 |
- pand 288(%r10),%xmm3 |
- por %xmm2,%xmm0 |
- por %xmm3,%xmm1 |
- por %xmm1,%xmm0 |
- pshufd $0x4e,%xmm0,%xmm1 |
- por %xmm1,%xmm0 |
- leaq 256(%r12),%r12 |
-.byte 102,72,15,126,195 |
- |
- movq %r13,16+8(%rsp) |
- movq %rdi,56+8(%rsp) |
- |
- movq (%r8),%r8 |
- movq (%rsi),%rax |
- leaq (%rsi,%r9,1),%rsi |
- negq %r9 |
- |
- movq %r8,%rbp |
- mulq %rbx |
- movq %rax,%r10 |
- movq (%rcx),%rax |
- |
- imulq %r10,%rbp |
- leaq 64+8(%rsp),%r14 |
- movq %rdx,%r11 |
- |
- mulq %rbp |
- addq %rax,%r10 |
- movq 8(%rsi,%r9,1),%rax |
- adcq $0,%rdx |
- movq %rdx,%rdi |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq 8(%rcx),%rax |
- adcq $0,%rdx |
- movq %rdx,%r10 |
- |
- mulq %rbp |
- addq %rax,%rdi |
- movq 16(%rsi,%r9,1),%rax |
- adcq $0,%rdx |
- addq %r11,%rdi |
- leaq 32(%r9),%r15 |
- leaq 32(%rcx),%rcx |
- adcq $0,%rdx |
- movq %rdi,(%r14) |
- movq %rdx,%r13 |
- jmp .L1st4x |
- |
-.align 32 |
-.L1st4x: |
- mulq %rbx |
- addq %rax,%r10 |
- movq -16(%rcx),%rax |
- leaq 32(%r14),%r14 |
- adcq $0,%rdx |
- movq %rdx,%r11 |
- |
- mulq %rbp |
- addq %rax,%r13 |
- movq -8(%rsi,%r15,1),%rax |
- adcq $0,%rdx |
- addq %r10,%r13 |
- adcq $0,%rdx |
- movq %r13,-24(%r14) |
- movq %rdx,%rdi |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq -8(%rcx),%rax |
- adcq $0,%rdx |
- movq %rdx,%r10 |
- |
- mulq %rbp |
- addq %rax,%rdi |
- movq (%rsi,%r15,1),%rax |
- adcq $0,%rdx |
- addq %r11,%rdi |
- adcq $0,%rdx |
- movq %rdi,-16(%r14) |
- movq %rdx,%r13 |
- |
- mulq %rbx |
- addq %rax,%r10 |
- movq 0(%rcx),%rax |
- adcq $0,%rdx |
- movq %rdx,%r11 |
- |
- mulq %rbp |
- addq %rax,%r13 |
- movq 8(%rsi,%r15,1),%rax |
- adcq $0,%rdx |
- addq %r10,%r13 |
- adcq $0,%rdx |
- movq %r13,-8(%r14) |
- movq %rdx,%rdi |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq 8(%rcx),%rax |
- adcq $0,%rdx |
- movq %rdx,%r10 |
- |
- mulq %rbp |
- addq %rax,%rdi |
- movq 16(%rsi,%r15,1),%rax |
- adcq $0,%rdx |
- addq %r11,%rdi |
- leaq 32(%rcx),%rcx |
- adcq $0,%rdx |
- movq %rdi,(%r14) |
- movq %rdx,%r13 |
- |
- addq $32,%r15 |
- jnz .L1st4x |
- |
- mulq %rbx |
- addq %rax,%r10 |
- movq -16(%rcx),%rax |
- leaq 32(%r14),%r14 |
- adcq $0,%rdx |
- movq %rdx,%r11 |
- |
- mulq %rbp |
- addq %rax,%r13 |
- movq -8(%rsi),%rax |
- adcq $0,%rdx |
- addq %r10,%r13 |
- adcq $0,%rdx |
- movq %r13,-24(%r14) |
- movq %rdx,%rdi |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq -8(%rcx),%rax |
- adcq $0,%rdx |
- movq %rdx,%r10 |
- |
- mulq %rbp |
- addq %rax,%rdi |
- movq (%rsi,%r9,1),%rax |
- adcq $0,%rdx |
- addq %r11,%rdi |
- adcq $0,%rdx |
- movq %rdi,-16(%r14) |
- movq %rdx,%r13 |
- |
- leaq (%rcx,%r9,1),%rcx |
- |
- xorq %rdi,%rdi |
- addq %r10,%r13 |
- adcq $0,%rdi |
- movq %r13,-8(%r14) |
- |
- jmp .Louter4x |
- |
-.align 32 |
-.Louter4x: |
- leaq 16+128(%r14),%rdx |
- pxor %xmm4,%xmm4 |
- pxor %xmm5,%xmm5 |
- movdqa -128(%r12),%xmm0 |
- movdqa -112(%r12),%xmm1 |
- movdqa -96(%r12),%xmm2 |
- movdqa -80(%r12),%xmm3 |
- pand -128(%rdx),%xmm0 |
- pand -112(%rdx),%xmm1 |
- por %xmm0,%xmm4 |
- pand -96(%rdx),%xmm2 |
- por %xmm1,%xmm5 |
- pand -80(%rdx),%xmm3 |
- por %xmm2,%xmm4 |
- por %xmm3,%xmm5 |
- movdqa -64(%r12),%xmm0 |
- movdqa -48(%r12),%xmm1 |
- movdqa -32(%r12),%xmm2 |
- movdqa -16(%r12),%xmm3 |
- pand -64(%rdx),%xmm0 |
- pand -48(%rdx),%xmm1 |
- por %xmm0,%xmm4 |
- pand -32(%rdx),%xmm2 |
- por %xmm1,%xmm5 |
- pand -16(%rdx),%xmm3 |
- por %xmm2,%xmm4 |
- por %xmm3,%xmm5 |
- movdqa 0(%r12),%xmm0 |
- movdqa 16(%r12),%xmm1 |
- movdqa 32(%r12),%xmm2 |
- movdqa 48(%r12),%xmm3 |
- pand 0(%rdx),%xmm0 |
- pand 16(%rdx),%xmm1 |
- por %xmm0,%xmm4 |
- pand 32(%rdx),%xmm2 |
- por %xmm1,%xmm5 |
- pand 48(%rdx),%xmm3 |
- por %xmm2,%xmm4 |
- por %xmm3,%xmm5 |
- movdqa 64(%r12),%xmm0 |
- movdqa 80(%r12),%xmm1 |
- movdqa 96(%r12),%xmm2 |
- movdqa 112(%r12),%xmm3 |
- pand 64(%rdx),%xmm0 |
- pand 80(%rdx),%xmm1 |
- por %xmm0,%xmm4 |
- pand 96(%rdx),%xmm2 |
- por %xmm1,%xmm5 |
- pand 112(%rdx),%xmm3 |
- por %xmm2,%xmm4 |
- por %xmm3,%xmm5 |
- por %xmm5,%xmm4 |
- pshufd $0x4e,%xmm4,%xmm0 |
- por %xmm4,%xmm0 |
- leaq 256(%r12),%r12 |
-.byte 102,72,15,126,195 |
- |
- movq (%r14,%r9,1),%r10 |
- movq %r8,%rbp |
- mulq %rbx |
- addq %rax,%r10 |
- movq (%rcx),%rax |
- adcq $0,%rdx |
- |
- imulq %r10,%rbp |
- movq %rdx,%r11 |
- movq %rdi,(%r14) |
- |
- leaq (%r14,%r9,1),%r14 |
- |
- mulq %rbp |
- addq %rax,%r10 |
- movq 8(%rsi,%r9,1),%rax |
- adcq $0,%rdx |
- movq %rdx,%rdi |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq 8(%rcx),%rax |
- adcq $0,%rdx |
- addq 8(%r14),%r11 |
- adcq $0,%rdx |
- movq %rdx,%r10 |
- |
- mulq %rbp |
- addq %rax,%rdi |
- movq 16(%rsi,%r9,1),%rax |
- adcq $0,%rdx |
- addq %r11,%rdi |
- leaq 32(%r9),%r15 |
- leaq 32(%rcx),%rcx |
- adcq $0,%rdx |
- movq %rdx,%r13 |
- jmp .Linner4x |
- |
-.align 32 |
-.Linner4x: |
- mulq %rbx |
- addq %rax,%r10 |
- movq -16(%rcx),%rax |
- adcq $0,%rdx |
- addq 16(%r14),%r10 |
- leaq 32(%r14),%r14 |
- adcq $0,%rdx |
- movq %rdx,%r11 |
- |
- mulq %rbp |
- addq %rax,%r13 |
- movq -8(%rsi,%r15,1),%rax |
- adcq $0,%rdx |
- addq %r10,%r13 |
- adcq $0,%rdx |
- movq %rdi,-32(%r14) |
- movq %rdx,%rdi |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq -8(%rcx),%rax |
- adcq $0,%rdx |
- addq -8(%r14),%r11 |
- adcq $0,%rdx |
- movq %rdx,%r10 |
- |
- mulq %rbp |
- addq %rax,%rdi |
- movq (%rsi,%r15,1),%rax |
- adcq $0,%rdx |
- addq %r11,%rdi |
- adcq $0,%rdx |
- movq %r13,-24(%r14) |
- movq %rdx,%r13 |
- |
- mulq %rbx |
- addq %rax,%r10 |
- movq 0(%rcx),%rax |
- adcq $0,%rdx |
- addq (%r14),%r10 |
- adcq $0,%rdx |
- movq %rdx,%r11 |
- |
- mulq %rbp |
- addq %rax,%r13 |
- movq 8(%rsi,%r15,1),%rax |
- adcq $0,%rdx |
- addq %r10,%r13 |
- adcq $0,%rdx |
- movq %rdi,-16(%r14) |
- movq %rdx,%rdi |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq 8(%rcx),%rax |
- adcq $0,%rdx |
- addq 8(%r14),%r11 |
- adcq $0,%rdx |
- movq %rdx,%r10 |
- |
- mulq %rbp |
- addq %rax,%rdi |
- movq 16(%rsi,%r15,1),%rax |
- adcq $0,%rdx |
- addq %r11,%rdi |
- leaq 32(%rcx),%rcx |
- adcq $0,%rdx |
- movq %r13,-8(%r14) |
- movq %rdx,%r13 |
- |
- addq $32,%r15 |
- jnz .Linner4x |
- |
- mulq %rbx |
- addq %rax,%r10 |
- movq -16(%rcx),%rax |
- adcq $0,%rdx |
- addq 16(%r14),%r10 |
- leaq 32(%r14),%r14 |
- adcq $0,%rdx |
- movq %rdx,%r11 |
- |
- mulq %rbp |
- addq %rax,%r13 |
- movq -8(%rsi),%rax |
- adcq $0,%rdx |
- addq %r10,%r13 |
- adcq $0,%rdx |
- movq %rdi,-32(%r14) |
- movq %rdx,%rdi |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq %rbp,%rax |
- movq -8(%rcx),%rbp |
- adcq $0,%rdx |
- addq -8(%r14),%r11 |
- adcq $0,%rdx |
- movq %rdx,%r10 |
- |
- mulq %rbp |
- addq %rax,%rdi |
- movq (%rsi,%r9,1),%rax |
- adcq $0,%rdx |
- addq %r11,%rdi |
- adcq $0,%rdx |
- movq %r13,-24(%r14) |
- movq %rdx,%r13 |
- |
- movq %rdi,-16(%r14) |
- leaq (%rcx,%r9,1),%rcx |
- |
- xorq %rdi,%rdi |
- addq %r10,%r13 |
- adcq $0,%rdi |
- addq (%r14),%r13 |
- adcq $0,%rdi |
- movq %r13,-8(%r14) |
- |
- cmpq 16+8(%rsp),%r12 |
- jb .Louter4x |
- xorq %rax,%rax |
- subq %r13,%rbp |
- adcq %r15,%r15 |
- orq %r15,%rdi |
- subq %rdi,%rax |
- leaq (%r14,%r9,1),%rbx |
- movq (%rcx),%r12 |
- leaq (%rcx),%rbp |
- movq %r9,%rcx |
- sarq $3+2,%rcx |
- movq 56+8(%rsp),%rdi |
- decq %r12 |
- xorq %r10,%r10 |
- movq 8(%rbp),%r13 |
- movq 16(%rbp),%r14 |
- movq 24(%rbp),%r15 |
- jmp .Lsqr4x_sub_entry |
-.size mul4x_internal,.-mul4x_internal |
-.globl bn_power5 |
-.hidden bn_power5 |
-.type bn_power5,@function |
-.align 32 |
-bn_power5: |
- movq %rsp,%rax |
- pushq %rbx |
- pushq %rbp |
- pushq %r12 |
- pushq %r13 |
- pushq %r14 |
- pushq %r15 |
- |
- shll $3,%r9d |
- leal (%r9,%r9,2),%r10d |
- negq %r9 |
- movq (%r8),%r8 |
- |
- |
- |
- |
- |
- |
- |
- |
- leaq -320(%rsp,%r9,2),%r11 |
- subq %rdi,%r11 |
- andq $4095,%r11 |
- cmpq %r11,%r10 |
- jb .Lpwr_sp_alt |
- subq %r11,%rsp |
- leaq -320(%rsp,%r9,2),%rsp |
- jmp .Lpwr_sp_done |
- |
-.align 32 |
-.Lpwr_sp_alt: |
- leaq 4096-320(,%r9,2),%r10 |
- leaq -320(%rsp,%r9,2),%rsp |
- subq %r10,%r11 |
- movq $0,%r10 |
- cmovcq %r10,%r11 |
- subq %r11,%rsp |
-.Lpwr_sp_done: |
- andq $-64,%rsp |
- movq %r9,%r10 |
- negq %r9 |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- movq %r8,32(%rsp) |
- movq %rax,40(%rsp) |
-.Lpower5_body: |
-.byte 102,72,15,110,207 |
-.byte 102,72,15,110,209 |
-.byte 102,73,15,110,218 |
-.byte 102,72,15,110,226 |
- |
- call __bn_sqr8x_internal |
- call __bn_post4x_internal |
- call __bn_sqr8x_internal |
- call __bn_post4x_internal |
- call __bn_sqr8x_internal |
- call __bn_post4x_internal |
- call __bn_sqr8x_internal |
- call __bn_post4x_internal |
- call __bn_sqr8x_internal |
- call __bn_post4x_internal |
- |
-.byte 102,72,15,126,209 |
-.byte 102,72,15,126,226 |
- movq %rsi,%rdi |
- movq 40(%rsp),%rax |
- leaq 32(%rsp),%r8 |
- |
- call mul4x_internal |
- |
- movq 40(%rsp),%rsi |
- movq $1,%rax |
- movq -48(%rsi),%r15 |
- movq -40(%rsi),%r14 |
- movq -32(%rsi),%r13 |
- movq -24(%rsi),%r12 |
- movq -16(%rsi),%rbp |
- movq -8(%rsi),%rbx |
- leaq (%rsi),%rsp |
-.Lpower5_epilogue: |
- .byte 0xf3,0xc3 |
-.size bn_power5,.-bn_power5 |
- |
-.globl bn_sqr8x_internal |
-.hidden bn_sqr8x_internal |
-.hidden bn_sqr8x_internal |
-.type bn_sqr8x_internal,@function |
-.align 32 |
-bn_sqr8x_internal: |
-__bn_sqr8x_internal: |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- leaq 32(%r10),%rbp |
- leaq (%rsi,%r9,1),%rsi |
- |
- movq %r9,%rcx |
- |
- |
- movq -32(%rsi,%rbp,1),%r14 |
- leaq 48+8(%rsp,%r9,2),%rdi |
- movq -24(%rsi,%rbp,1),%rax |
- leaq -32(%rdi,%rbp,1),%rdi |
- movq -16(%rsi,%rbp,1),%rbx |
- movq %rax,%r15 |
- |
- mulq %r14 |
- movq %rax,%r10 |
- movq %rbx,%rax |
- movq %rdx,%r11 |
- movq %r10,-24(%rdi,%rbp,1) |
- |
- mulq %r14 |
- addq %rax,%r11 |
- movq %rbx,%rax |
- adcq $0,%rdx |
- movq %r11,-16(%rdi,%rbp,1) |
- movq %rdx,%r10 |
- |
- |
- movq -8(%rsi,%rbp,1),%rbx |
- mulq %r15 |
- movq %rax,%r12 |
- movq %rbx,%rax |
- movq %rdx,%r13 |
- |
- leaq (%rbp),%rcx |
- mulq %r14 |
- addq %rax,%r10 |
- movq %rbx,%rax |
- movq %rdx,%r11 |
- adcq $0,%r11 |
- addq %r12,%r10 |
- adcq $0,%r11 |
- movq %r10,-8(%rdi,%rcx,1) |
- jmp .Lsqr4x_1st |
- |
-.align 32 |
-.Lsqr4x_1st: |
- movq (%rsi,%rcx,1),%rbx |
- mulq %r15 |
- addq %rax,%r13 |
- movq %rbx,%rax |
- movq %rdx,%r12 |
- adcq $0,%r12 |
- |
- mulq %r14 |
- addq %rax,%r11 |
- movq %rbx,%rax |
- movq 8(%rsi,%rcx,1),%rbx |
- movq %rdx,%r10 |
- adcq $0,%r10 |
- addq %r13,%r11 |
- adcq $0,%r10 |
- |
- |
- mulq %r15 |
- addq %rax,%r12 |
- movq %rbx,%rax |
- movq %r11,(%rdi,%rcx,1) |
- movq %rdx,%r13 |
- adcq $0,%r13 |
- |
- mulq %r14 |
- addq %rax,%r10 |
- movq %rbx,%rax |
- movq 16(%rsi,%rcx,1),%rbx |
- movq %rdx,%r11 |
- adcq $0,%r11 |
- addq %r12,%r10 |
- adcq $0,%r11 |
- |
- mulq %r15 |
- addq %rax,%r13 |
- movq %rbx,%rax |
- movq %r10,8(%rdi,%rcx,1) |
- movq %rdx,%r12 |
- adcq $0,%r12 |
- |
- mulq %r14 |
- addq %rax,%r11 |
- movq %rbx,%rax |
- movq 24(%rsi,%rcx,1),%rbx |
- movq %rdx,%r10 |
- adcq $0,%r10 |
- addq %r13,%r11 |
- adcq $0,%r10 |
- |
- |
- mulq %r15 |
- addq %rax,%r12 |
- movq %rbx,%rax |
- movq %r11,16(%rdi,%rcx,1) |
- movq %rdx,%r13 |
- adcq $0,%r13 |
- leaq 32(%rcx),%rcx |
- |
- mulq %r14 |
- addq %rax,%r10 |
- movq %rbx,%rax |
- movq %rdx,%r11 |
- adcq $0,%r11 |
- addq %r12,%r10 |
- adcq $0,%r11 |
- movq %r10,-8(%rdi,%rcx,1) |
- |
- cmpq $0,%rcx |
- jne .Lsqr4x_1st |
- |
- mulq %r15 |
- addq %rax,%r13 |
- leaq 16(%rbp),%rbp |
- adcq $0,%rdx |
- addq %r11,%r13 |
- adcq $0,%rdx |
- |
- movq %r13,(%rdi) |
- movq %rdx,%r12 |
- movq %rdx,8(%rdi) |
- jmp .Lsqr4x_outer |
- |
-.align 32 |
-.Lsqr4x_outer: |
- movq -32(%rsi,%rbp,1),%r14 |
- leaq 48+8(%rsp,%r9,2),%rdi |
- movq -24(%rsi,%rbp,1),%rax |
- leaq -32(%rdi,%rbp,1),%rdi |
- movq -16(%rsi,%rbp,1),%rbx |
- movq %rax,%r15 |
- |
- mulq %r14 |
- movq -24(%rdi,%rbp,1),%r10 |
- addq %rax,%r10 |
- movq %rbx,%rax |
- adcq $0,%rdx |
- movq %r10,-24(%rdi,%rbp,1) |
- movq %rdx,%r11 |
- |
- mulq %r14 |
- addq %rax,%r11 |
- movq %rbx,%rax |
- adcq $0,%rdx |
- addq -16(%rdi,%rbp,1),%r11 |
- movq %rdx,%r10 |
- adcq $0,%r10 |
- movq %r11,-16(%rdi,%rbp,1) |
- |
- xorq %r12,%r12 |
- |
- movq -8(%rsi,%rbp,1),%rbx |
- mulq %r15 |
- addq %rax,%r12 |
- movq %rbx,%rax |
- adcq $0,%rdx |
- addq -8(%rdi,%rbp,1),%r12 |
- movq %rdx,%r13 |
- adcq $0,%r13 |
- |
- mulq %r14 |
- addq %rax,%r10 |
- movq %rbx,%rax |
- adcq $0,%rdx |
- addq %r12,%r10 |
- movq %rdx,%r11 |
- adcq $0,%r11 |
- movq %r10,-8(%rdi,%rbp,1) |
- |
- leaq (%rbp),%rcx |
- jmp .Lsqr4x_inner |
- |
-.align 32 |
-.Lsqr4x_inner: |
- movq (%rsi,%rcx,1),%rbx |
- mulq %r15 |
- addq %rax,%r13 |
- movq %rbx,%rax |
- movq %rdx,%r12 |
- adcq $0,%r12 |
- addq (%rdi,%rcx,1),%r13 |
- adcq $0,%r12 |
- |
-.byte 0x67 |
- mulq %r14 |
- addq %rax,%r11 |
- movq %rbx,%rax |
- movq 8(%rsi,%rcx,1),%rbx |
- movq %rdx,%r10 |
- adcq $0,%r10 |
- addq %r13,%r11 |
- adcq $0,%r10 |
- |
- mulq %r15 |
- addq %rax,%r12 |
- movq %r11,(%rdi,%rcx,1) |
- movq %rbx,%rax |
- movq %rdx,%r13 |
- adcq $0,%r13 |
- addq 8(%rdi,%rcx,1),%r12 |
- leaq 16(%rcx),%rcx |
- adcq $0,%r13 |
- |
- mulq %r14 |
- addq %rax,%r10 |
- movq %rbx,%rax |
- adcq $0,%rdx |
- addq %r12,%r10 |
- movq %rdx,%r11 |
- adcq $0,%r11 |
- movq %r10,-8(%rdi,%rcx,1) |
- |
- cmpq $0,%rcx |
- jne .Lsqr4x_inner |
- |
-.byte 0x67 |
- mulq %r15 |
- addq %rax,%r13 |
- adcq $0,%rdx |
- addq %r11,%r13 |
- adcq $0,%rdx |
- |
- movq %r13,(%rdi) |
- movq %rdx,%r12 |
- movq %rdx,8(%rdi) |
- |
- addq $16,%rbp |
- jnz .Lsqr4x_outer |
- |
- |
- movq -32(%rsi),%r14 |
- leaq 48+8(%rsp,%r9,2),%rdi |
- movq -24(%rsi),%rax |
- leaq -32(%rdi,%rbp,1),%rdi |
- movq -16(%rsi),%rbx |
- movq %rax,%r15 |
- |
- mulq %r14 |
- addq %rax,%r10 |
- movq %rbx,%rax |
- movq %rdx,%r11 |
- adcq $0,%r11 |
- |
- mulq %r14 |
- addq %rax,%r11 |
- movq %rbx,%rax |
- movq %r10,-24(%rdi) |
- movq %rdx,%r10 |
- adcq $0,%r10 |
- addq %r13,%r11 |
- movq -8(%rsi),%rbx |
- adcq $0,%r10 |
- |
- mulq %r15 |
- addq %rax,%r12 |
- movq %rbx,%rax |
- movq %r11,-16(%rdi) |
- movq %rdx,%r13 |
- adcq $0,%r13 |
- |
- mulq %r14 |
- addq %rax,%r10 |
- movq %rbx,%rax |
- movq %rdx,%r11 |
- adcq $0,%r11 |
- addq %r12,%r10 |
- adcq $0,%r11 |
- movq %r10,-8(%rdi) |
- |
- mulq %r15 |
- addq %rax,%r13 |
- movq -16(%rsi),%rax |
- adcq $0,%rdx |
- addq %r11,%r13 |
- adcq $0,%rdx |
- |
- movq %r13,(%rdi) |
- movq %rdx,%r12 |
- movq %rdx,8(%rdi) |
- |
- mulq %rbx |
- addq $16,%rbp |
- xorq %r14,%r14 |
- subq %r9,%rbp |
- xorq %r15,%r15 |
- |
- addq %r12,%rax |
- adcq $0,%rdx |
- movq %rax,8(%rdi) |
- movq %rdx,16(%rdi) |
- movq %r15,24(%rdi) |
- |
- movq -16(%rsi,%rbp,1),%rax |
- leaq 48+8(%rsp),%rdi |
- xorq %r10,%r10 |
- movq 8(%rdi),%r11 |
- |
- leaq (%r14,%r10,2),%r12 |
- shrq $63,%r10 |
- leaq (%rcx,%r11,2),%r13 |
- shrq $63,%r11 |
- orq %r10,%r13 |
- movq 16(%rdi),%r10 |
- movq %r11,%r14 |
- mulq %rax |
- negq %r15 |
- movq 24(%rdi),%r11 |
- adcq %rax,%r12 |
- movq -8(%rsi,%rbp,1),%rax |
- movq %r12,(%rdi) |
- adcq %rdx,%r13 |
- |
- leaq (%r14,%r10,2),%rbx |
- movq %r13,8(%rdi) |
- sbbq %r15,%r15 |
- shrq $63,%r10 |
- leaq (%rcx,%r11,2),%r8 |
- shrq $63,%r11 |
- orq %r10,%r8 |
- movq 32(%rdi),%r10 |
- movq %r11,%r14 |
- mulq %rax |
- negq %r15 |
- movq 40(%rdi),%r11 |
- adcq %rax,%rbx |
- movq 0(%rsi,%rbp,1),%rax |
- movq %rbx,16(%rdi) |
- adcq %rdx,%r8 |
- leaq 16(%rbp),%rbp |
- movq %r8,24(%rdi) |
- sbbq %r15,%r15 |
- leaq 64(%rdi),%rdi |
- jmp .Lsqr4x_shift_n_add |
- |
-.align 32 |
-.Lsqr4x_shift_n_add: |
- leaq (%r14,%r10,2),%r12 |
- shrq $63,%r10 |
- leaq (%rcx,%r11,2),%r13 |
- shrq $63,%r11 |
- orq %r10,%r13 |
- movq -16(%rdi),%r10 |
- movq %r11,%r14 |
- mulq %rax |
- negq %r15 |
- movq -8(%rdi),%r11 |
- adcq %rax,%r12 |
- movq -8(%rsi,%rbp,1),%rax |
- movq %r12,-32(%rdi) |
- adcq %rdx,%r13 |
- |
- leaq (%r14,%r10,2),%rbx |
- movq %r13,-24(%rdi) |
- sbbq %r15,%r15 |
- shrq $63,%r10 |
- leaq (%rcx,%r11,2),%r8 |
- shrq $63,%r11 |
- orq %r10,%r8 |
- movq 0(%rdi),%r10 |
- movq %r11,%r14 |
- mulq %rax |
- negq %r15 |
- movq 8(%rdi),%r11 |
- adcq %rax,%rbx |
- movq 0(%rsi,%rbp,1),%rax |
- movq %rbx,-16(%rdi) |
- adcq %rdx,%r8 |
- |
- leaq (%r14,%r10,2),%r12 |
- movq %r8,-8(%rdi) |
- sbbq %r15,%r15 |
- shrq $63,%r10 |
- leaq (%rcx,%r11,2),%r13 |
- shrq $63,%r11 |
- orq %r10,%r13 |
- movq 16(%rdi),%r10 |
- movq %r11,%r14 |
- mulq %rax |
- negq %r15 |
- movq 24(%rdi),%r11 |
- adcq %rax,%r12 |
- movq 8(%rsi,%rbp,1),%rax |
- movq %r12,0(%rdi) |
- adcq %rdx,%r13 |
- |
- leaq (%r14,%r10,2),%rbx |
- movq %r13,8(%rdi) |
- sbbq %r15,%r15 |
- shrq $63,%r10 |
- leaq (%rcx,%r11,2),%r8 |
- shrq $63,%r11 |
- orq %r10,%r8 |
- movq 32(%rdi),%r10 |
- movq %r11,%r14 |
- mulq %rax |
- negq %r15 |
- movq 40(%rdi),%r11 |
- adcq %rax,%rbx |
- movq 16(%rsi,%rbp,1),%rax |
- movq %rbx,16(%rdi) |
- adcq %rdx,%r8 |
- movq %r8,24(%rdi) |
- sbbq %r15,%r15 |
- leaq 64(%rdi),%rdi |
- addq $32,%rbp |
- jnz .Lsqr4x_shift_n_add |
- |
- leaq (%r14,%r10,2),%r12 |
-.byte 0x67 |
- shrq $63,%r10 |
- leaq (%rcx,%r11,2),%r13 |
- shrq $63,%r11 |
- orq %r10,%r13 |
- movq -16(%rdi),%r10 |
- movq %r11,%r14 |
- mulq %rax |
- negq %r15 |
- movq -8(%rdi),%r11 |
- adcq %rax,%r12 |
- movq -8(%rsi),%rax |
- movq %r12,-32(%rdi) |
- adcq %rdx,%r13 |
- |
- leaq (%r14,%r10,2),%rbx |
- movq %r13,-24(%rdi) |
- sbbq %r15,%r15 |
- shrq $63,%r10 |
- leaq (%rcx,%r11,2),%r8 |
- shrq $63,%r11 |
- orq %r10,%r8 |
- mulq %rax |
- negq %r15 |
- adcq %rax,%rbx |
- adcq %rdx,%r8 |
- movq %rbx,-16(%rdi) |
- movq %r8,-8(%rdi) |
-.byte 102,72,15,126,213 |
-__bn_sqr8x_reduction: |
- xorq %rax,%rax |
- leaq (%r9,%rbp,1),%rcx |
- leaq 48+8(%rsp,%r9,2),%rdx |
- movq %rcx,0+8(%rsp) |
- leaq 48+8(%rsp,%r9,1),%rdi |
- movq %rdx,8+8(%rsp) |
- negq %r9 |
- jmp .L8x_reduction_loop |
- |
-.align 32 |
-.L8x_reduction_loop: |
- leaq (%rdi,%r9,1),%rdi |
-.byte 0x66 |
- movq 0(%rdi),%rbx |
- movq 8(%rdi),%r9 |
- movq 16(%rdi),%r10 |
- movq 24(%rdi),%r11 |
- movq 32(%rdi),%r12 |
- movq 40(%rdi),%r13 |
- movq 48(%rdi),%r14 |
- movq 56(%rdi),%r15 |
- movq %rax,(%rdx) |
- leaq 64(%rdi),%rdi |
- |
-.byte 0x67 |
- movq %rbx,%r8 |
- imulq 32+8(%rsp),%rbx |
- movq 0(%rbp),%rax |
- movl $8,%ecx |
- jmp .L8x_reduce |
- |
-.align 32 |
-.L8x_reduce: |
- mulq %rbx |
- movq 8(%rbp),%rax |
- negq %r8 |
- movq %rdx,%r8 |
- adcq $0,%r8 |
- |
- mulq %rbx |
- addq %rax,%r9 |
- movq 16(%rbp),%rax |
- adcq $0,%rdx |
- addq %r9,%r8 |
- movq %rbx,48-8+8(%rsp,%rcx,8) |
- movq %rdx,%r9 |
- adcq $0,%r9 |
- |
- mulq %rbx |
- addq %rax,%r10 |
- movq 24(%rbp),%rax |
- adcq $0,%rdx |
- addq %r10,%r9 |
- movq 32+8(%rsp),%rsi |
- movq %rdx,%r10 |
- adcq $0,%r10 |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq 32(%rbp),%rax |
- adcq $0,%rdx |
- imulq %r8,%rsi |
- addq %r11,%r10 |
- movq %rdx,%r11 |
- adcq $0,%r11 |
- |
- mulq %rbx |
- addq %rax,%r12 |
- movq 40(%rbp),%rax |
- adcq $0,%rdx |
- addq %r12,%r11 |
- movq %rdx,%r12 |
- adcq $0,%r12 |
- |
- mulq %rbx |
- addq %rax,%r13 |
- movq 48(%rbp),%rax |
- adcq $0,%rdx |
- addq %r13,%r12 |
- movq %rdx,%r13 |
- adcq $0,%r13 |
- |
- mulq %rbx |
- addq %rax,%r14 |
- movq 56(%rbp),%rax |
- adcq $0,%rdx |
- addq %r14,%r13 |
- movq %rdx,%r14 |
- adcq $0,%r14 |
- |
- mulq %rbx |
- movq %rsi,%rbx |
- addq %rax,%r15 |
- movq 0(%rbp),%rax |
- adcq $0,%rdx |
- addq %r15,%r14 |
- movq %rdx,%r15 |
- adcq $0,%r15 |
- |
- decl %ecx |
- jnz .L8x_reduce |
- |
- leaq 64(%rbp),%rbp |
- xorq %rax,%rax |
- movq 8+8(%rsp),%rdx |
- cmpq 0+8(%rsp),%rbp |
- jae .L8x_no_tail |
- |
-.byte 0x66 |
- addq 0(%rdi),%r8 |
- adcq 8(%rdi),%r9 |
- adcq 16(%rdi),%r10 |
- adcq 24(%rdi),%r11 |
- adcq 32(%rdi),%r12 |
- adcq 40(%rdi),%r13 |
- adcq 48(%rdi),%r14 |
- adcq 56(%rdi),%r15 |
- sbbq %rsi,%rsi |
- |
- movq 48+56+8(%rsp),%rbx |
- movl $8,%ecx |
- movq 0(%rbp),%rax |
- jmp .L8x_tail |
- |
-.align 32 |
-.L8x_tail: |
- mulq %rbx |
- addq %rax,%r8 |
- movq 8(%rbp),%rax |
- movq %r8,(%rdi) |
- movq %rdx,%r8 |
- adcq $0,%r8 |
- |
- mulq %rbx |
- addq %rax,%r9 |
- movq 16(%rbp),%rax |
- adcq $0,%rdx |
- addq %r9,%r8 |
- leaq 8(%rdi),%rdi |
- movq %rdx,%r9 |
- adcq $0,%r9 |
- |
- mulq %rbx |
- addq %rax,%r10 |
- movq 24(%rbp),%rax |
- adcq $0,%rdx |
- addq %r10,%r9 |
- movq %rdx,%r10 |
- adcq $0,%r10 |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq 32(%rbp),%rax |
- adcq $0,%rdx |
- addq %r11,%r10 |
- movq %rdx,%r11 |
- adcq $0,%r11 |
- |
- mulq %rbx |
- addq %rax,%r12 |
- movq 40(%rbp),%rax |
- adcq $0,%rdx |
- addq %r12,%r11 |
- movq %rdx,%r12 |
- adcq $0,%r12 |
- |
- mulq %rbx |
- addq %rax,%r13 |
- movq 48(%rbp),%rax |
- adcq $0,%rdx |
- addq %r13,%r12 |
- movq %rdx,%r13 |
- adcq $0,%r13 |
- |
- mulq %rbx |
- addq %rax,%r14 |
- movq 56(%rbp),%rax |
- adcq $0,%rdx |
- addq %r14,%r13 |
- movq %rdx,%r14 |
- adcq $0,%r14 |
- |
- mulq %rbx |
- movq 48-16+8(%rsp,%rcx,8),%rbx |
- addq %rax,%r15 |
- adcq $0,%rdx |
- addq %r15,%r14 |
- movq 0(%rbp),%rax |
- movq %rdx,%r15 |
- adcq $0,%r15 |
- |
- decl %ecx |
- jnz .L8x_tail |
- |
- leaq 64(%rbp),%rbp |
- movq 8+8(%rsp),%rdx |
- cmpq 0+8(%rsp),%rbp |
- jae .L8x_tail_done |
- |
- movq 48+56+8(%rsp),%rbx |
- negq %rsi |
- movq 0(%rbp),%rax |
- adcq 0(%rdi),%r8 |
- adcq 8(%rdi),%r9 |
- adcq 16(%rdi),%r10 |
- adcq 24(%rdi),%r11 |
- adcq 32(%rdi),%r12 |
- adcq 40(%rdi),%r13 |
- adcq 48(%rdi),%r14 |
- adcq 56(%rdi),%r15 |
- sbbq %rsi,%rsi |
- |
- movl $8,%ecx |
- jmp .L8x_tail |
- |
-.align 32 |
-.L8x_tail_done: |
- addq (%rdx),%r8 |
- adcq $0,%r9 |
- adcq $0,%r10 |
- adcq $0,%r11 |
- adcq $0,%r12 |
- adcq $0,%r13 |
- adcq $0,%r14 |
- adcq $0,%r15 |
- |
- |
- xorq %rax,%rax |
- |
- negq %rsi |
-.L8x_no_tail: |
- adcq 0(%rdi),%r8 |
- adcq 8(%rdi),%r9 |
- adcq 16(%rdi),%r10 |
- adcq 24(%rdi),%r11 |
- adcq 32(%rdi),%r12 |
- adcq 40(%rdi),%r13 |
- adcq 48(%rdi),%r14 |
- adcq 56(%rdi),%r15 |
- adcq $0,%rax |
- movq -8(%rbp),%rcx |
- xorq %rsi,%rsi |
- |
-.byte 102,72,15,126,213 |
- |
- movq %r8,0(%rdi) |
- movq %r9,8(%rdi) |
-.byte 102,73,15,126,217 |
- movq %r10,16(%rdi) |
- movq %r11,24(%rdi) |
- movq %r12,32(%rdi) |
- movq %r13,40(%rdi) |
- movq %r14,48(%rdi) |
- movq %r15,56(%rdi) |
- leaq 64(%rdi),%rdi |
- |
- cmpq %rdx,%rdi |
- jb .L8x_reduction_loop |
- .byte 0xf3,0xc3 |
-.size bn_sqr8x_internal,.-bn_sqr8x_internal |
-.type __bn_post4x_internal,@function |
-.align 32 |
-__bn_post4x_internal: |
- movq 0(%rbp),%r12 |
- leaq (%rdi,%r9,1),%rbx |
- movq %r9,%rcx |
-.byte 102,72,15,126,207 |
- negq %rax |
-.byte 102,72,15,126,206 |
- sarq $3+2,%rcx |
- decq %r12 |
- xorq %r10,%r10 |
- movq 8(%rbp),%r13 |
- movq 16(%rbp),%r14 |
- movq 24(%rbp),%r15 |
- jmp .Lsqr4x_sub_entry |
- |
-.align 16 |
-.Lsqr4x_sub: |
- movq 0(%rbp),%r12 |
- movq 8(%rbp),%r13 |
- movq 16(%rbp),%r14 |
- movq 24(%rbp),%r15 |
-.Lsqr4x_sub_entry: |
- leaq 32(%rbp),%rbp |
- notq %r12 |
- notq %r13 |
- notq %r14 |
- notq %r15 |
- andq %rax,%r12 |
- andq %rax,%r13 |
- andq %rax,%r14 |
- andq %rax,%r15 |
- |
- negq %r10 |
- adcq 0(%rbx),%r12 |
- adcq 8(%rbx),%r13 |
- adcq 16(%rbx),%r14 |
- adcq 24(%rbx),%r15 |
- movq %r12,0(%rdi) |
- leaq 32(%rbx),%rbx |
- movq %r13,8(%rdi) |
- sbbq %r10,%r10 |
- movq %r14,16(%rdi) |
- movq %r15,24(%rdi) |
- leaq 32(%rdi),%rdi |
- |
- incq %rcx |
- jnz .Lsqr4x_sub |
- |
- movq %r9,%r10 |
- negq %r9 |
- .byte 0xf3,0xc3 |
-.size __bn_post4x_internal,.-__bn_post4x_internal |
-.globl bn_from_montgomery |
-.hidden bn_from_montgomery |
-.type bn_from_montgomery,@function |
-.align 32 |
-bn_from_montgomery: |
- testl $7,%r9d |
- jz bn_from_mont8x |
- xorl %eax,%eax |
- .byte 0xf3,0xc3 |
-.size bn_from_montgomery,.-bn_from_montgomery |
- |
-.type bn_from_mont8x,@function |
-.align 32 |
-bn_from_mont8x: |
-.byte 0x67 |
- movq %rsp,%rax |
- pushq %rbx |
- pushq %rbp |
- pushq %r12 |
- pushq %r13 |
- pushq %r14 |
- pushq %r15 |
- |
- shll $3,%r9d |
- leaq (%r9,%r9,2),%r10 |
- negq %r9 |
- movq (%r8),%r8 |
- |
- |
- |
- |
- |
- |
- |
- |
- leaq -320(%rsp,%r9,2),%r11 |
- subq %rdi,%r11 |
- andq $4095,%r11 |
- cmpq %r11,%r10 |
- jb .Lfrom_sp_alt |
- subq %r11,%rsp |
- leaq -320(%rsp,%r9,2),%rsp |
- jmp .Lfrom_sp_done |
- |
-.align 32 |
-.Lfrom_sp_alt: |
- leaq 4096-320(,%r9,2),%r10 |
- leaq -320(%rsp,%r9,2),%rsp |
- subq %r10,%r11 |
- movq $0,%r10 |
- cmovcq %r10,%r11 |
- subq %r11,%rsp |
-.Lfrom_sp_done: |
- andq $-64,%rsp |
- movq %r9,%r10 |
- negq %r9 |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- |
- movq %r8,32(%rsp) |
- movq %rax,40(%rsp) |
-.Lfrom_body: |
- movq %r9,%r11 |
- leaq 48(%rsp),%rax |
- pxor %xmm0,%xmm0 |
- jmp .Lmul_by_1 |
- |
-.align 32 |
-.Lmul_by_1: |
- movdqu (%rsi),%xmm1 |
- movdqu 16(%rsi),%xmm2 |
- movdqu 32(%rsi),%xmm3 |
- movdqa %xmm0,(%rax,%r9,1) |
- movdqu 48(%rsi),%xmm4 |
- movdqa %xmm0,16(%rax,%r9,1) |
-.byte 0x48,0x8d,0xb6,0x40,0x00,0x00,0x00 |
- movdqa %xmm1,(%rax) |
- movdqa %xmm0,32(%rax,%r9,1) |
- movdqa %xmm2,16(%rax) |
- movdqa %xmm0,48(%rax,%r9,1) |
- movdqa %xmm3,32(%rax) |
- movdqa %xmm4,48(%rax) |
- leaq 64(%rax),%rax |
- subq $64,%r11 |
- jnz .Lmul_by_1 |
- |
-.byte 102,72,15,110,207 |
-.byte 102,72,15,110,209 |
-.byte 0x67 |
- movq %rcx,%rbp |
-.byte 102,73,15,110,218 |
- call __bn_sqr8x_reduction |
- call __bn_post4x_internal |
- |
- pxor %xmm0,%xmm0 |
- leaq 48(%rsp),%rax |
- movq 40(%rsp),%rsi |
- jmp .Lfrom_mont_zero |
- |
-.align 32 |
-.Lfrom_mont_zero: |
- movdqa %xmm0,0(%rax) |
- movdqa %xmm0,16(%rax) |
- movdqa %xmm0,32(%rax) |
- movdqa %xmm0,48(%rax) |
- leaq 64(%rax),%rax |
- subq $32,%r9 |
- jnz .Lfrom_mont_zero |
- |
- movq $1,%rax |
- movq -48(%rsi),%r15 |
- movq -40(%rsi),%r14 |
- movq -32(%rsi),%r13 |
- movq -24(%rsi),%r12 |
- movq -16(%rsi),%rbp |
- movq -8(%rsi),%rbx |
- leaq (%rsi),%rsp |
-.Lfrom_epilogue: |
- .byte 0xf3,0xc3 |
-.size bn_from_mont8x,.-bn_from_mont8x |
-.globl bn_scatter5 |
-.hidden bn_scatter5 |
-.type bn_scatter5,@function |
-.align 16 |
-bn_scatter5: |
- cmpl $0,%esi |
- jz .Lscatter_epilogue |
- leaq (%rdx,%rcx,8),%rdx |
-.Lscatter: |
- movq (%rdi),%rax |
- leaq 8(%rdi),%rdi |
- movq %rax,(%rdx) |
- leaq 256(%rdx),%rdx |
- subl $1,%esi |
- jnz .Lscatter |
-.Lscatter_epilogue: |
- .byte 0xf3,0xc3 |
-.size bn_scatter5,.-bn_scatter5 |
- |
-.globl bn_gather5 |
-.hidden bn_gather5 |
-.type bn_gather5,@function |
-.align 32 |
-bn_gather5: |
-.LSEH_begin_bn_gather5: |
- |
-.byte 0x4c,0x8d,0x14,0x24 |
-.byte 0x48,0x81,0xec,0x08,0x01,0x00,0x00 |
- leaq .Linc(%rip),%rax |
- andq $-16,%rsp |
- |
- movd %ecx,%xmm5 |
- movdqa 0(%rax),%xmm0 |
- movdqa 16(%rax),%xmm1 |
- leaq 128(%rdx),%r11 |
- leaq 128(%rsp),%rax |
- |
- pshufd $0,%xmm5,%xmm5 |
- movdqa %xmm1,%xmm4 |
- movdqa %xmm1,%xmm2 |
- paddd %xmm0,%xmm1 |
- pcmpeqd %xmm5,%xmm0 |
- movdqa %xmm4,%xmm3 |
- |
- paddd %xmm1,%xmm2 |
- pcmpeqd %xmm5,%xmm1 |
- movdqa %xmm0,-128(%rax) |
- movdqa %xmm4,%xmm0 |
- |
- paddd %xmm2,%xmm3 |
- pcmpeqd %xmm5,%xmm2 |
- movdqa %xmm1,-112(%rax) |
- movdqa %xmm4,%xmm1 |
- |
- paddd %xmm3,%xmm0 |
- pcmpeqd %xmm5,%xmm3 |
- movdqa %xmm2,-96(%rax) |
- movdqa %xmm4,%xmm2 |
- paddd %xmm0,%xmm1 |
- pcmpeqd %xmm5,%xmm0 |
- movdqa %xmm3,-80(%rax) |
- movdqa %xmm4,%xmm3 |
- |
- paddd %xmm1,%xmm2 |
- pcmpeqd %xmm5,%xmm1 |
- movdqa %xmm0,-64(%rax) |
- movdqa %xmm4,%xmm0 |
- |
- paddd %xmm2,%xmm3 |
- pcmpeqd %xmm5,%xmm2 |
- movdqa %xmm1,-48(%rax) |
- movdqa %xmm4,%xmm1 |
- |
- paddd %xmm3,%xmm0 |
- pcmpeqd %xmm5,%xmm3 |
- movdqa %xmm2,-32(%rax) |
- movdqa %xmm4,%xmm2 |
- paddd %xmm0,%xmm1 |
- pcmpeqd %xmm5,%xmm0 |
- movdqa %xmm3,-16(%rax) |
- movdqa %xmm4,%xmm3 |
- |
- paddd %xmm1,%xmm2 |
- pcmpeqd %xmm5,%xmm1 |
- movdqa %xmm0,0(%rax) |
- movdqa %xmm4,%xmm0 |
- |
- paddd %xmm2,%xmm3 |
- pcmpeqd %xmm5,%xmm2 |
- movdqa %xmm1,16(%rax) |
- movdqa %xmm4,%xmm1 |
- |
- paddd %xmm3,%xmm0 |
- pcmpeqd %xmm5,%xmm3 |
- movdqa %xmm2,32(%rax) |
- movdqa %xmm4,%xmm2 |
- paddd %xmm0,%xmm1 |
- pcmpeqd %xmm5,%xmm0 |
- movdqa %xmm3,48(%rax) |
- movdqa %xmm4,%xmm3 |
- |
- paddd %xmm1,%xmm2 |
- pcmpeqd %xmm5,%xmm1 |
- movdqa %xmm0,64(%rax) |
- movdqa %xmm4,%xmm0 |
- |
- paddd %xmm2,%xmm3 |
- pcmpeqd %xmm5,%xmm2 |
- movdqa %xmm1,80(%rax) |
- movdqa %xmm4,%xmm1 |
- |
- paddd %xmm3,%xmm0 |
- pcmpeqd %xmm5,%xmm3 |
- movdqa %xmm2,96(%rax) |
- movdqa %xmm4,%xmm2 |
- movdqa %xmm3,112(%rax) |
- jmp .Lgather |
- |
-.align 32 |
-.Lgather: |
- pxor %xmm4,%xmm4 |
- pxor %xmm5,%xmm5 |
- movdqa -128(%r11),%xmm0 |
- movdqa -112(%r11),%xmm1 |
- movdqa -96(%r11),%xmm2 |
- pand -128(%rax),%xmm0 |
- movdqa -80(%r11),%xmm3 |
- pand -112(%rax),%xmm1 |
- por %xmm0,%xmm4 |
- pand -96(%rax),%xmm2 |
- por %xmm1,%xmm5 |
- pand -80(%rax),%xmm3 |
- por %xmm2,%xmm4 |
- por %xmm3,%xmm5 |
- movdqa -64(%r11),%xmm0 |
- movdqa -48(%r11),%xmm1 |
- movdqa -32(%r11),%xmm2 |
- pand -64(%rax),%xmm0 |
- movdqa -16(%r11),%xmm3 |
- pand -48(%rax),%xmm1 |
- por %xmm0,%xmm4 |
- pand -32(%rax),%xmm2 |
- por %xmm1,%xmm5 |
- pand -16(%rax),%xmm3 |
- por %xmm2,%xmm4 |
- por %xmm3,%xmm5 |
- movdqa 0(%r11),%xmm0 |
- movdqa 16(%r11),%xmm1 |
- movdqa 32(%r11),%xmm2 |
- pand 0(%rax),%xmm0 |
- movdqa 48(%r11),%xmm3 |
- pand 16(%rax),%xmm1 |
- por %xmm0,%xmm4 |
- pand 32(%rax),%xmm2 |
- por %xmm1,%xmm5 |
- pand 48(%rax),%xmm3 |
- por %xmm2,%xmm4 |
- por %xmm3,%xmm5 |
- movdqa 64(%r11),%xmm0 |
- movdqa 80(%r11),%xmm1 |
- movdqa 96(%r11),%xmm2 |
- pand 64(%rax),%xmm0 |
- movdqa 112(%r11),%xmm3 |
- pand 80(%rax),%xmm1 |
- por %xmm0,%xmm4 |
- pand 96(%rax),%xmm2 |
- por %xmm1,%xmm5 |
- pand 112(%rax),%xmm3 |
- por %xmm2,%xmm4 |
- por %xmm3,%xmm5 |
- por %xmm5,%xmm4 |
- leaq 256(%r11),%r11 |
- pshufd $0x4e,%xmm4,%xmm0 |
- por %xmm4,%xmm0 |
- movq %xmm0,(%rdi) |
- leaq 8(%rdi),%rdi |
- subl $1,%esi |
- jnz .Lgather |
- |
- leaq (%r10),%rsp |
- .byte 0xf3,0xc3 |
-.LSEH_end_bn_gather5: |
-.size bn_gather5,.-bn_gather5 |
-.align 64 |
-.Linc: |
-.long 0,0, 1,1 |
-.long 2,2, 2,2 |
-.byte 77,111,110,116,103,111,109,101,114,121,32,77,117,108,116,105,112,108,105,99,97,116,105,111,110,32,119,105,116,104,32,115,99,97,116,116,101,114,47,103,97,116,104,101,114,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,103,62,0 |
-#endif |