Index: third_party/boringssl/linux-x86_64/crypto/bn/rsaz-x86_64.S |
diff --git a/third_party/boringssl/linux-x86_64/crypto/bn/rsaz-x86_64.S b/third_party/boringssl/linux-x86_64/crypto/bn/rsaz-x86_64.S |
deleted file mode 100644 |
index 21531d1c65099f55e3894c1875e7479a2d6d653c..0000000000000000000000000000000000000000 |
--- a/third_party/boringssl/linux-x86_64/crypto/bn/rsaz-x86_64.S |
+++ /dev/null |
@@ -1,1229 +0,0 @@ |
-#if defined(__x86_64__) |
-.text |
- |
-.extern OPENSSL_ia32cap_P |
-.hidden OPENSSL_ia32cap_P |
- |
-.globl rsaz_512_sqr |
-.hidden rsaz_512_sqr |
-.type rsaz_512_sqr,@function |
-.align 32 |
-rsaz_512_sqr: |
- pushq %rbx |
- pushq %rbp |
- pushq %r12 |
- pushq %r13 |
- pushq %r14 |
- pushq %r15 |
- |
- subq $128+24,%rsp |
-.Lsqr_body: |
- movq %rdx,%rbp |
- movq (%rsi),%rdx |
- movq 8(%rsi),%rax |
- movq %rcx,128(%rsp) |
- jmp .Loop_sqr |
- |
-.align 32 |
-.Loop_sqr: |
- movl %r8d,128+8(%rsp) |
- |
- movq %rdx,%rbx |
- mulq %rdx |
- movq %rax,%r8 |
- movq 16(%rsi),%rax |
- movq %rdx,%r9 |
- |
- mulq %rbx |
- addq %rax,%r9 |
- movq 24(%rsi),%rax |
- movq %rdx,%r10 |
- adcq $0,%r10 |
- |
- mulq %rbx |
- addq %rax,%r10 |
- movq 32(%rsi),%rax |
- movq %rdx,%r11 |
- adcq $0,%r11 |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq 40(%rsi),%rax |
- movq %rdx,%r12 |
- adcq $0,%r12 |
- |
- mulq %rbx |
- addq %rax,%r12 |
- movq 48(%rsi),%rax |
- movq %rdx,%r13 |
- adcq $0,%r13 |
- |
- mulq %rbx |
- addq %rax,%r13 |
- movq 56(%rsi),%rax |
- movq %rdx,%r14 |
- adcq $0,%r14 |
- |
- mulq %rbx |
- addq %rax,%r14 |
- movq %rbx,%rax |
- movq %rdx,%r15 |
- adcq $0,%r15 |
- |
- addq %r8,%r8 |
- movq %r9,%rcx |
- adcq %r9,%r9 |
- |
- mulq %rax |
- movq %rax,(%rsp) |
- addq %rdx,%r8 |
- adcq $0,%r9 |
- |
- movq %r8,8(%rsp) |
- shrq $63,%rcx |
- |
- |
- movq 8(%rsi),%r8 |
- movq 16(%rsi),%rax |
- mulq %r8 |
- addq %rax,%r10 |
- movq 24(%rsi),%rax |
- movq %rdx,%rbx |
- adcq $0,%rbx |
- |
- mulq %r8 |
- addq %rax,%r11 |
- movq 32(%rsi),%rax |
- adcq $0,%rdx |
- addq %rbx,%r11 |
- movq %rdx,%rbx |
- adcq $0,%rbx |
- |
- mulq %r8 |
- addq %rax,%r12 |
- movq 40(%rsi),%rax |
- adcq $0,%rdx |
- addq %rbx,%r12 |
- movq %rdx,%rbx |
- adcq $0,%rbx |
- |
- mulq %r8 |
- addq %rax,%r13 |
- movq 48(%rsi),%rax |
- adcq $0,%rdx |
- addq %rbx,%r13 |
- movq %rdx,%rbx |
- adcq $0,%rbx |
- |
- mulq %r8 |
- addq %rax,%r14 |
- movq 56(%rsi),%rax |
- adcq $0,%rdx |
- addq %rbx,%r14 |
- movq %rdx,%rbx |
- adcq $0,%rbx |
- |
- mulq %r8 |
- addq %rax,%r15 |
- movq %r8,%rax |
- adcq $0,%rdx |
- addq %rbx,%r15 |
- movq %rdx,%r8 |
- movq %r10,%rdx |
- adcq $0,%r8 |
- |
- addq %rdx,%rdx |
- leaq (%rcx,%r10,2),%r10 |
- movq %r11,%rbx |
- adcq %r11,%r11 |
- |
- mulq %rax |
- addq %rax,%r9 |
- adcq %rdx,%r10 |
- adcq $0,%r11 |
- |
- movq %r9,16(%rsp) |
- movq %r10,24(%rsp) |
- shrq $63,%rbx |
- |
- |
- movq 16(%rsi),%r9 |
- movq 24(%rsi),%rax |
- mulq %r9 |
- addq %rax,%r12 |
- movq 32(%rsi),%rax |
- movq %rdx,%rcx |
- adcq $0,%rcx |
- |
- mulq %r9 |
- addq %rax,%r13 |
- movq 40(%rsi),%rax |
- adcq $0,%rdx |
- addq %rcx,%r13 |
- movq %rdx,%rcx |
- adcq $0,%rcx |
- |
- mulq %r9 |
- addq %rax,%r14 |
- movq 48(%rsi),%rax |
- adcq $0,%rdx |
- addq %rcx,%r14 |
- movq %rdx,%rcx |
- adcq $0,%rcx |
- |
- mulq %r9 |
- movq %r12,%r10 |
- leaq (%rbx,%r12,2),%r12 |
- addq %rax,%r15 |
- movq 56(%rsi),%rax |
- adcq $0,%rdx |
- addq %rcx,%r15 |
- movq %rdx,%rcx |
- adcq $0,%rcx |
- |
- mulq %r9 |
- shrq $63,%r10 |
- addq %rax,%r8 |
- movq %r9,%rax |
- adcq $0,%rdx |
- addq %rcx,%r8 |
- movq %rdx,%r9 |
- adcq $0,%r9 |
- |
- movq %r13,%rcx |
- leaq (%r10,%r13,2),%r13 |
- |
- mulq %rax |
- addq %rax,%r11 |
- adcq %rdx,%r12 |
- adcq $0,%r13 |
- |
- movq %r11,32(%rsp) |
- movq %r12,40(%rsp) |
- shrq $63,%rcx |
- |
- |
- movq 24(%rsi),%r10 |
- movq 32(%rsi),%rax |
- mulq %r10 |
- addq %rax,%r14 |
- movq 40(%rsi),%rax |
- movq %rdx,%rbx |
- adcq $0,%rbx |
- |
- mulq %r10 |
- addq %rax,%r15 |
- movq 48(%rsi),%rax |
- adcq $0,%rdx |
- addq %rbx,%r15 |
- movq %rdx,%rbx |
- adcq $0,%rbx |
- |
- mulq %r10 |
- movq %r14,%r12 |
- leaq (%rcx,%r14,2),%r14 |
- addq %rax,%r8 |
- movq 56(%rsi),%rax |
- adcq $0,%rdx |
- addq %rbx,%r8 |
- movq %rdx,%rbx |
- adcq $0,%rbx |
- |
- mulq %r10 |
- shrq $63,%r12 |
- addq %rax,%r9 |
- movq %r10,%rax |
- adcq $0,%rdx |
- addq %rbx,%r9 |
- movq %rdx,%r10 |
- adcq $0,%r10 |
- |
- movq %r15,%rbx |
- leaq (%r12,%r15,2),%r15 |
- |
- mulq %rax |
- addq %rax,%r13 |
- adcq %rdx,%r14 |
- adcq $0,%r15 |
- |
- movq %r13,48(%rsp) |
- movq %r14,56(%rsp) |
- shrq $63,%rbx |
- |
- |
- movq 32(%rsi),%r11 |
- movq 40(%rsi),%rax |
- mulq %r11 |
- addq %rax,%r8 |
- movq 48(%rsi),%rax |
- movq %rdx,%rcx |
- adcq $0,%rcx |
- |
- mulq %r11 |
- addq %rax,%r9 |
- movq 56(%rsi),%rax |
- adcq $0,%rdx |
- movq %r8,%r12 |
- leaq (%rbx,%r8,2),%r8 |
- addq %rcx,%r9 |
- movq %rdx,%rcx |
- adcq $0,%rcx |
- |
- mulq %r11 |
- shrq $63,%r12 |
- addq %rax,%r10 |
- movq %r11,%rax |
- adcq $0,%rdx |
- addq %rcx,%r10 |
- movq %rdx,%r11 |
- adcq $0,%r11 |
- |
- movq %r9,%rcx |
- leaq (%r12,%r9,2),%r9 |
- |
- mulq %rax |
- addq %rax,%r15 |
- adcq %rdx,%r8 |
- adcq $0,%r9 |
- |
- movq %r15,64(%rsp) |
- movq %r8,72(%rsp) |
- shrq $63,%rcx |
- |
- |
- movq 40(%rsi),%r12 |
- movq 48(%rsi),%rax |
- mulq %r12 |
- addq %rax,%r10 |
- movq 56(%rsi),%rax |
- movq %rdx,%rbx |
- adcq $0,%rbx |
- |
- mulq %r12 |
- addq %rax,%r11 |
- movq %r12,%rax |
- movq %r10,%r15 |
- leaq (%rcx,%r10,2),%r10 |
- adcq $0,%rdx |
- shrq $63,%r15 |
- addq %rbx,%r11 |
- movq %rdx,%r12 |
- adcq $0,%r12 |
- |
- movq %r11,%rbx |
- leaq (%r15,%r11,2),%r11 |
- |
- mulq %rax |
- addq %rax,%r9 |
- adcq %rdx,%r10 |
- adcq $0,%r11 |
- |
- movq %r9,80(%rsp) |
- movq %r10,88(%rsp) |
- |
- |
- movq 48(%rsi),%r13 |
- movq 56(%rsi),%rax |
- mulq %r13 |
- addq %rax,%r12 |
- movq %r13,%rax |
- movq %rdx,%r13 |
- adcq $0,%r13 |
- |
- xorq %r14,%r14 |
- shlq $1,%rbx |
- adcq %r12,%r12 |
- adcq %r13,%r13 |
- adcq %r14,%r14 |
- |
- mulq %rax |
- addq %rax,%r11 |
- adcq %rdx,%r12 |
- adcq $0,%r13 |
- |
- movq %r11,96(%rsp) |
- movq %r12,104(%rsp) |
- |
- |
- movq 56(%rsi),%rax |
- mulq %rax |
- addq %rax,%r13 |
- adcq $0,%rdx |
- |
- addq %rdx,%r14 |
- |
- movq %r13,112(%rsp) |
- movq %r14,120(%rsp) |
- |
- movq (%rsp),%r8 |
- movq 8(%rsp),%r9 |
- movq 16(%rsp),%r10 |
- movq 24(%rsp),%r11 |
- movq 32(%rsp),%r12 |
- movq 40(%rsp),%r13 |
- movq 48(%rsp),%r14 |
- movq 56(%rsp),%r15 |
- |
- call __rsaz_512_reduce |
- |
- addq 64(%rsp),%r8 |
- adcq 72(%rsp),%r9 |
- adcq 80(%rsp),%r10 |
- adcq 88(%rsp),%r11 |
- adcq 96(%rsp),%r12 |
- adcq 104(%rsp),%r13 |
- adcq 112(%rsp),%r14 |
- adcq 120(%rsp),%r15 |
- sbbq %rcx,%rcx |
- |
- call __rsaz_512_subtract |
- |
- movq %r8,%rdx |
- movq %r9,%rax |
- movl 128+8(%rsp),%r8d |
- movq %rdi,%rsi |
- |
- decl %r8d |
- jnz .Loop_sqr |
- |
- leaq 128+24+48(%rsp),%rax |
- movq -48(%rax),%r15 |
- movq -40(%rax),%r14 |
- movq -32(%rax),%r13 |
- movq -24(%rax),%r12 |
- movq -16(%rax),%rbp |
- movq -8(%rax),%rbx |
- leaq (%rax),%rsp |
-.Lsqr_epilogue: |
- .byte 0xf3,0xc3 |
-.size rsaz_512_sqr,.-rsaz_512_sqr |
-.globl rsaz_512_mul |
-.hidden rsaz_512_mul |
-.type rsaz_512_mul,@function |
-.align 32 |
-rsaz_512_mul: |
- pushq %rbx |
- pushq %rbp |
- pushq %r12 |
- pushq %r13 |
- pushq %r14 |
- pushq %r15 |
- |
- subq $128+24,%rsp |
-.Lmul_body: |
-.byte 102,72,15,110,199 |
-.byte 102,72,15,110,201 |
- movq %r8,128(%rsp) |
- movq (%rdx),%rbx |
- movq %rdx,%rbp |
- call __rsaz_512_mul |
- |
-.byte 102,72,15,126,199 |
-.byte 102,72,15,126,205 |
- |
- movq (%rsp),%r8 |
- movq 8(%rsp),%r9 |
- movq 16(%rsp),%r10 |
- movq 24(%rsp),%r11 |
- movq 32(%rsp),%r12 |
- movq 40(%rsp),%r13 |
- movq 48(%rsp),%r14 |
- movq 56(%rsp),%r15 |
- |
- call __rsaz_512_reduce |
- addq 64(%rsp),%r8 |
- adcq 72(%rsp),%r9 |
- adcq 80(%rsp),%r10 |
- adcq 88(%rsp),%r11 |
- adcq 96(%rsp),%r12 |
- adcq 104(%rsp),%r13 |
- adcq 112(%rsp),%r14 |
- adcq 120(%rsp),%r15 |
- sbbq %rcx,%rcx |
- |
- call __rsaz_512_subtract |
- |
- leaq 128+24+48(%rsp),%rax |
- movq -48(%rax),%r15 |
- movq -40(%rax),%r14 |
- movq -32(%rax),%r13 |
- movq -24(%rax),%r12 |
- movq -16(%rax),%rbp |
- movq -8(%rax),%rbx |
- leaq (%rax),%rsp |
-.Lmul_epilogue: |
- .byte 0xf3,0xc3 |
-.size rsaz_512_mul,.-rsaz_512_mul |
-.globl rsaz_512_mul_gather4 |
-.hidden rsaz_512_mul_gather4 |
-.type rsaz_512_mul_gather4,@function |
-.align 32 |
-rsaz_512_mul_gather4: |
- pushq %rbx |
- pushq %rbp |
- pushq %r12 |
- pushq %r13 |
- pushq %r14 |
- pushq %r15 |
- |
- subq $152,%rsp |
-.Lmul_gather4_body: |
- movd %r9d,%xmm8 |
- movdqa .Linc+16(%rip),%xmm1 |
- movdqa .Linc(%rip),%xmm0 |
- |
- pshufd $0,%xmm8,%xmm8 |
- movdqa %xmm1,%xmm7 |
- movdqa %xmm1,%xmm2 |
- paddd %xmm0,%xmm1 |
- pcmpeqd %xmm8,%xmm0 |
- movdqa %xmm7,%xmm3 |
- paddd %xmm1,%xmm2 |
- pcmpeqd %xmm8,%xmm1 |
- movdqa %xmm7,%xmm4 |
- paddd %xmm2,%xmm3 |
- pcmpeqd %xmm8,%xmm2 |
- movdqa %xmm7,%xmm5 |
- paddd %xmm3,%xmm4 |
- pcmpeqd %xmm8,%xmm3 |
- movdqa %xmm7,%xmm6 |
- paddd %xmm4,%xmm5 |
- pcmpeqd %xmm8,%xmm4 |
- paddd %xmm5,%xmm6 |
- pcmpeqd %xmm8,%xmm5 |
- paddd %xmm6,%xmm7 |
- pcmpeqd %xmm8,%xmm6 |
- pcmpeqd %xmm8,%xmm7 |
- |
- movdqa 0(%rdx),%xmm8 |
- movdqa 16(%rdx),%xmm9 |
- movdqa 32(%rdx),%xmm10 |
- movdqa 48(%rdx),%xmm11 |
- pand %xmm0,%xmm8 |
- movdqa 64(%rdx),%xmm12 |
- pand %xmm1,%xmm9 |
- movdqa 80(%rdx),%xmm13 |
- pand %xmm2,%xmm10 |
- movdqa 96(%rdx),%xmm14 |
- pand %xmm3,%xmm11 |
- movdqa 112(%rdx),%xmm15 |
- leaq 128(%rdx),%rbp |
- pand %xmm4,%xmm12 |
- pand %xmm5,%xmm13 |
- pand %xmm6,%xmm14 |
- pand %xmm7,%xmm15 |
- por %xmm10,%xmm8 |
- por %xmm11,%xmm9 |
- por %xmm12,%xmm8 |
- por %xmm13,%xmm9 |
- por %xmm14,%xmm8 |
- por %xmm15,%xmm9 |
- |
- por %xmm9,%xmm8 |
- pshufd $0x4e,%xmm8,%xmm9 |
- por %xmm9,%xmm8 |
-.byte 102,76,15,126,195 |
- |
- movq %r8,128(%rsp) |
- movq %rdi,128+8(%rsp) |
- movq %rcx,128+16(%rsp) |
- |
- movq (%rsi),%rax |
- movq 8(%rsi),%rcx |
- mulq %rbx |
- movq %rax,(%rsp) |
- movq %rcx,%rax |
- movq %rdx,%r8 |
- |
- mulq %rbx |
- addq %rax,%r8 |
- movq 16(%rsi),%rax |
- movq %rdx,%r9 |
- adcq $0,%r9 |
- |
- mulq %rbx |
- addq %rax,%r9 |
- movq 24(%rsi),%rax |
- movq %rdx,%r10 |
- adcq $0,%r10 |
- |
- mulq %rbx |
- addq %rax,%r10 |
- movq 32(%rsi),%rax |
- movq %rdx,%r11 |
- adcq $0,%r11 |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq 40(%rsi),%rax |
- movq %rdx,%r12 |
- adcq $0,%r12 |
- |
- mulq %rbx |
- addq %rax,%r12 |
- movq 48(%rsi),%rax |
- movq %rdx,%r13 |
- adcq $0,%r13 |
- |
- mulq %rbx |
- addq %rax,%r13 |
- movq 56(%rsi),%rax |
- movq %rdx,%r14 |
- adcq $0,%r14 |
- |
- mulq %rbx |
- addq %rax,%r14 |
- movq (%rsi),%rax |
- movq %rdx,%r15 |
- adcq $0,%r15 |
- |
- leaq 8(%rsp),%rdi |
- movl $7,%ecx |
- jmp .Loop_mul_gather |
- |
-.align 32 |
-.Loop_mul_gather: |
- movdqa 0(%rbp),%xmm8 |
- movdqa 16(%rbp),%xmm9 |
- movdqa 32(%rbp),%xmm10 |
- movdqa 48(%rbp),%xmm11 |
- pand %xmm0,%xmm8 |
- movdqa 64(%rbp),%xmm12 |
- pand %xmm1,%xmm9 |
- movdqa 80(%rbp),%xmm13 |
- pand %xmm2,%xmm10 |
- movdqa 96(%rbp),%xmm14 |
- pand %xmm3,%xmm11 |
- movdqa 112(%rbp),%xmm15 |
- leaq 128(%rbp),%rbp |
- pand %xmm4,%xmm12 |
- pand %xmm5,%xmm13 |
- pand %xmm6,%xmm14 |
- pand %xmm7,%xmm15 |
- por %xmm10,%xmm8 |
- por %xmm11,%xmm9 |
- por %xmm12,%xmm8 |
- por %xmm13,%xmm9 |
- por %xmm14,%xmm8 |
- por %xmm15,%xmm9 |
- |
- por %xmm9,%xmm8 |
- pshufd $0x4e,%xmm8,%xmm9 |
- por %xmm9,%xmm8 |
-.byte 102,76,15,126,195 |
- |
- mulq %rbx |
- addq %rax,%r8 |
- movq 8(%rsi),%rax |
- movq %r8,(%rdi) |
- movq %rdx,%r8 |
- adcq $0,%r8 |
- |
- mulq %rbx |
- addq %rax,%r9 |
- movq 16(%rsi),%rax |
- adcq $0,%rdx |
- addq %r9,%r8 |
- movq %rdx,%r9 |
- adcq $0,%r9 |
- |
- mulq %rbx |
- addq %rax,%r10 |
- movq 24(%rsi),%rax |
- adcq $0,%rdx |
- addq %r10,%r9 |
- movq %rdx,%r10 |
- adcq $0,%r10 |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq 32(%rsi),%rax |
- adcq $0,%rdx |
- addq %r11,%r10 |
- movq %rdx,%r11 |
- adcq $0,%r11 |
- |
- mulq %rbx |
- addq %rax,%r12 |
- movq 40(%rsi),%rax |
- adcq $0,%rdx |
- addq %r12,%r11 |
- movq %rdx,%r12 |
- adcq $0,%r12 |
- |
- mulq %rbx |
- addq %rax,%r13 |
- movq 48(%rsi),%rax |
- adcq $0,%rdx |
- addq %r13,%r12 |
- movq %rdx,%r13 |
- adcq $0,%r13 |
- |
- mulq %rbx |
- addq %rax,%r14 |
- movq 56(%rsi),%rax |
- adcq $0,%rdx |
- addq %r14,%r13 |
- movq %rdx,%r14 |
- adcq $0,%r14 |
- |
- mulq %rbx |
- addq %rax,%r15 |
- movq (%rsi),%rax |
- adcq $0,%rdx |
- addq %r15,%r14 |
- movq %rdx,%r15 |
- adcq $0,%r15 |
- |
- leaq 8(%rdi),%rdi |
- |
- decl %ecx |
- jnz .Loop_mul_gather |
- |
- movq %r8,(%rdi) |
- movq %r9,8(%rdi) |
- movq %r10,16(%rdi) |
- movq %r11,24(%rdi) |
- movq %r12,32(%rdi) |
- movq %r13,40(%rdi) |
- movq %r14,48(%rdi) |
- movq %r15,56(%rdi) |
- |
- movq 128+8(%rsp),%rdi |
- movq 128+16(%rsp),%rbp |
- |
- movq (%rsp),%r8 |
- movq 8(%rsp),%r9 |
- movq 16(%rsp),%r10 |
- movq 24(%rsp),%r11 |
- movq 32(%rsp),%r12 |
- movq 40(%rsp),%r13 |
- movq 48(%rsp),%r14 |
- movq 56(%rsp),%r15 |
- |
- call __rsaz_512_reduce |
- addq 64(%rsp),%r8 |
- adcq 72(%rsp),%r9 |
- adcq 80(%rsp),%r10 |
- adcq 88(%rsp),%r11 |
- adcq 96(%rsp),%r12 |
- adcq 104(%rsp),%r13 |
- adcq 112(%rsp),%r14 |
- adcq 120(%rsp),%r15 |
- sbbq %rcx,%rcx |
- |
- call __rsaz_512_subtract |
- |
- leaq 128+24+48(%rsp),%rax |
- movq -48(%rax),%r15 |
- movq -40(%rax),%r14 |
- movq -32(%rax),%r13 |
- movq -24(%rax),%r12 |
- movq -16(%rax),%rbp |
- movq -8(%rax),%rbx |
- leaq (%rax),%rsp |
-.Lmul_gather4_epilogue: |
- .byte 0xf3,0xc3 |
-.size rsaz_512_mul_gather4,.-rsaz_512_mul_gather4 |
-.globl rsaz_512_mul_scatter4 |
-.hidden rsaz_512_mul_scatter4 |
-.type rsaz_512_mul_scatter4,@function |
-.align 32 |
-rsaz_512_mul_scatter4: |
- pushq %rbx |
- pushq %rbp |
- pushq %r12 |
- pushq %r13 |
- pushq %r14 |
- pushq %r15 |
- |
- movl %r9d,%r9d |
- subq $128+24,%rsp |
-.Lmul_scatter4_body: |
- leaq (%r8,%r9,8),%r8 |
-.byte 102,72,15,110,199 |
-.byte 102,72,15,110,202 |
-.byte 102,73,15,110,208 |
- movq %rcx,128(%rsp) |
- |
- movq %rdi,%rbp |
- movq (%rdi),%rbx |
- call __rsaz_512_mul |
- |
-.byte 102,72,15,126,199 |
-.byte 102,72,15,126,205 |
- |
- movq (%rsp),%r8 |
- movq 8(%rsp),%r9 |
- movq 16(%rsp),%r10 |
- movq 24(%rsp),%r11 |
- movq 32(%rsp),%r12 |
- movq 40(%rsp),%r13 |
- movq 48(%rsp),%r14 |
- movq 56(%rsp),%r15 |
- |
- call __rsaz_512_reduce |
- addq 64(%rsp),%r8 |
- adcq 72(%rsp),%r9 |
- adcq 80(%rsp),%r10 |
- adcq 88(%rsp),%r11 |
- adcq 96(%rsp),%r12 |
- adcq 104(%rsp),%r13 |
- adcq 112(%rsp),%r14 |
- adcq 120(%rsp),%r15 |
-.byte 102,72,15,126,214 |
- sbbq %rcx,%rcx |
- |
- call __rsaz_512_subtract |
- |
- movq %r8,0(%rsi) |
- movq %r9,128(%rsi) |
- movq %r10,256(%rsi) |
- movq %r11,384(%rsi) |
- movq %r12,512(%rsi) |
- movq %r13,640(%rsi) |
- movq %r14,768(%rsi) |
- movq %r15,896(%rsi) |
- |
- leaq 128+24+48(%rsp),%rax |
- movq -48(%rax),%r15 |
- movq -40(%rax),%r14 |
- movq -32(%rax),%r13 |
- movq -24(%rax),%r12 |
- movq -16(%rax),%rbp |
- movq -8(%rax),%rbx |
- leaq (%rax),%rsp |
-.Lmul_scatter4_epilogue: |
- .byte 0xf3,0xc3 |
-.size rsaz_512_mul_scatter4,.-rsaz_512_mul_scatter4 |
-.globl rsaz_512_mul_by_one |
-.hidden rsaz_512_mul_by_one |
-.type rsaz_512_mul_by_one,@function |
-.align 32 |
-rsaz_512_mul_by_one: |
- pushq %rbx |
- pushq %rbp |
- pushq %r12 |
- pushq %r13 |
- pushq %r14 |
- pushq %r15 |
- |
- subq $128+24,%rsp |
-.Lmul_by_one_body: |
- movq %rdx,%rbp |
- movq %rcx,128(%rsp) |
- |
- movq (%rsi),%r8 |
- pxor %xmm0,%xmm0 |
- movq 8(%rsi),%r9 |
- movq 16(%rsi),%r10 |
- movq 24(%rsi),%r11 |
- movq 32(%rsi),%r12 |
- movq 40(%rsi),%r13 |
- movq 48(%rsi),%r14 |
- movq 56(%rsi),%r15 |
- |
- movdqa %xmm0,(%rsp) |
- movdqa %xmm0,16(%rsp) |
- movdqa %xmm0,32(%rsp) |
- movdqa %xmm0,48(%rsp) |
- movdqa %xmm0,64(%rsp) |
- movdqa %xmm0,80(%rsp) |
- movdqa %xmm0,96(%rsp) |
- call __rsaz_512_reduce |
- movq %r8,(%rdi) |
- movq %r9,8(%rdi) |
- movq %r10,16(%rdi) |
- movq %r11,24(%rdi) |
- movq %r12,32(%rdi) |
- movq %r13,40(%rdi) |
- movq %r14,48(%rdi) |
- movq %r15,56(%rdi) |
- |
- leaq 128+24+48(%rsp),%rax |
- movq -48(%rax),%r15 |
- movq -40(%rax),%r14 |
- movq -32(%rax),%r13 |
- movq -24(%rax),%r12 |
- movq -16(%rax),%rbp |
- movq -8(%rax),%rbx |
- leaq (%rax),%rsp |
-.Lmul_by_one_epilogue: |
- .byte 0xf3,0xc3 |
-.size rsaz_512_mul_by_one,.-rsaz_512_mul_by_one |
-.type __rsaz_512_reduce,@function |
-.align 32 |
-__rsaz_512_reduce: |
- movq %r8,%rbx |
- imulq 128+8(%rsp),%rbx |
- movq 0(%rbp),%rax |
- movl $8,%ecx |
- jmp .Lreduction_loop |
- |
-.align 32 |
-.Lreduction_loop: |
- mulq %rbx |
- movq 8(%rbp),%rax |
- negq %r8 |
- movq %rdx,%r8 |
- adcq $0,%r8 |
- |
- mulq %rbx |
- addq %rax,%r9 |
- movq 16(%rbp),%rax |
- adcq $0,%rdx |
- addq %r9,%r8 |
- movq %rdx,%r9 |
- adcq $0,%r9 |
- |
- mulq %rbx |
- addq %rax,%r10 |
- movq 24(%rbp),%rax |
- adcq $0,%rdx |
- addq %r10,%r9 |
- movq %rdx,%r10 |
- adcq $0,%r10 |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq 32(%rbp),%rax |
- adcq $0,%rdx |
- addq %r11,%r10 |
- movq 128+8(%rsp),%rsi |
- |
- |
- adcq $0,%rdx |
- movq %rdx,%r11 |
- |
- mulq %rbx |
- addq %rax,%r12 |
- movq 40(%rbp),%rax |
- adcq $0,%rdx |
- imulq %r8,%rsi |
- addq %r12,%r11 |
- movq %rdx,%r12 |
- adcq $0,%r12 |
- |
- mulq %rbx |
- addq %rax,%r13 |
- movq 48(%rbp),%rax |
- adcq $0,%rdx |
- addq %r13,%r12 |
- movq %rdx,%r13 |
- adcq $0,%r13 |
- |
- mulq %rbx |
- addq %rax,%r14 |
- movq 56(%rbp),%rax |
- adcq $0,%rdx |
- addq %r14,%r13 |
- movq %rdx,%r14 |
- adcq $0,%r14 |
- |
- mulq %rbx |
- movq %rsi,%rbx |
- addq %rax,%r15 |
- movq 0(%rbp),%rax |
- adcq $0,%rdx |
- addq %r15,%r14 |
- movq %rdx,%r15 |
- adcq $0,%r15 |
- |
- decl %ecx |
- jne .Lreduction_loop |
- |
- .byte 0xf3,0xc3 |
-.size __rsaz_512_reduce,.-__rsaz_512_reduce |
-.type __rsaz_512_subtract,@function |
-.align 32 |
-__rsaz_512_subtract: |
- movq %r8,(%rdi) |
- movq %r9,8(%rdi) |
- movq %r10,16(%rdi) |
- movq %r11,24(%rdi) |
- movq %r12,32(%rdi) |
- movq %r13,40(%rdi) |
- movq %r14,48(%rdi) |
- movq %r15,56(%rdi) |
- |
- movq 0(%rbp),%r8 |
- movq 8(%rbp),%r9 |
- negq %r8 |
- notq %r9 |
- andq %rcx,%r8 |
- movq 16(%rbp),%r10 |
- andq %rcx,%r9 |
- notq %r10 |
- movq 24(%rbp),%r11 |
- andq %rcx,%r10 |
- notq %r11 |
- movq 32(%rbp),%r12 |
- andq %rcx,%r11 |
- notq %r12 |
- movq 40(%rbp),%r13 |
- andq %rcx,%r12 |
- notq %r13 |
- movq 48(%rbp),%r14 |
- andq %rcx,%r13 |
- notq %r14 |
- movq 56(%rbp),%r15 |
- andq %rcx,%r14 |
- notq %r15 |
- andq %rcx,%r15 |
- |
- addq (%rdi),%r8 |
- adcq 8(%rdi),%r9 |
- adcq 16(%rdi),%r10 |
- adcq 24(%rdi),%r11 |
- adcq 32(%rdi),%r12 |
- adcq 40(%rdi),%r13 |
- adcq 48(%rdi),%r14 |
- adcq 56(%rdi),%r15 |
- |
- movq %r8,(%rdi) |
- movq %r9,8(%rdi) |
- movq %r10,16(%rdi) |
- movq %r11,24(%rdi) |
- movq %r12,32(%rdi) |
- movq %r13,40(%rdi) |
- movq %r14,48(%rdi) |
- movq %r15,56(%rdi) |
- |
- .byte 0xf3,0xc3 |
-.size __rsaz_512_subtract,.-__rsaz_512_subtract |
-.type __rsaz_512_mul,@function |
-.align 32 |
-__rsaz_512_mul: |
- leaq 8(%rsp),%rdi |
- |
- movq (%rsi),%rax |
- mulq %rbx |
- movq %rax,(%rdi) |
- movq 8(%rsi),%rax |
- movq %rdx,%r8 |
- |
- mulq %rbx |
- addq %rax,%r8 |
- movq 16(%rsi),%rax |
- movq %rdx,%r9 |
- adcq $0,%r9 |
- |
- mulq %rbx |
- addq %rax,%r9 |
- movq 24(%rsi),%rax |
- movq %rdx,%r10 |
- adcq $0,%r10 |
- |
- mulq %rbx |
- addq %rax,%r10 |
- movq 32(%rsi),%rax |
- movq %rdx,%r11 |
- adcq $0,%r11 |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq 40(%rsi),%rax |
- movq %rdx,%r12 |
- adcq $0,%r12 |
- |
- mulq %rbx |
- addq %rax,%r12 |
- movq 48(%rsi),%rax |
- movq %rdx,%r13 |
- adcq $0,%r13 |
- |
- mulq %rbx |
- addq %rax,%r13 |
- movq 56(%rsi),%rax |
- movq %rdx,%r14 |
- adcq $0,%r14 |
- |
- mulq %rbx |
- addq %rax,%r14 |
- movq (%rsi),%rax |
- movq %rdx,%r15 |
- adcq $0,%r15 |
- |
- leaq 8(%rbp),%rbp |
- leaq 8(%rdi),%rdi |
- |
- movl $7,%ecx |
- jmp .Loop_mul |
- |
-.align 32 |
-.Loop_mul: |
- movq (%rbp),%rbx |
- mulq %rbx |
- addq %rax,%r8 |
- movq 8(%rsi),%rax |
- movq %r8,(%rdi) |
- movq %rdx,%r8 |
- adcq $0,%r8 |
- |
- mulq %rbx |
- addq %rax,%r9 |
- movq 16(%rsi),%rax |
- adcq $0,%rdx |
- addq %r9,%r8 |
- movq %rdx,%r9 |
- adcq $0,%r9 |
- |
- mulq %rbx |
- addq %rax,%r10 |
- movq 24(%rsi),%rax |
- adcq $0,%rdx |
- addq %r10,%r9 |
- movq %rdx,%r10 |
- adcq $0,%r10 |
- |
- mulq %rbx |
- addq %rax,%r11 |
- movq 32(%rsi),%rax |
- adcq $0,%rdx |
- addq %r11,%r10 |
- movq %rdx,%r11 |
- adcq $0,%r11 |
- |
- mulq %rbx |
- addq %rax,%r12 |
- movq 40(%rsi),%rax |
- adcq $0,%rdx |
- addq %r12,%r11 |
- movq %rdx,%r12 |
- adcq $0,%r12 |
- |
- mulq %rbx |
- addq %rax,%r13 |
- movq 48(%rsi),%rax |
- adcq $0,%rdx |
- addq %r13,%r12 |
- movq %rdx,%r13 |
- adcq $0,%r13 |
- |
- mulq %rbx |
- addq %rax,%r14 |
- movq 56(%rsi),%rax |
- adcq $0,%rdx |
- addq %r14,%r13 |
- movq %rdx,%r14 |
- leaq 8(%rbp),%rbp |
- adcq $0,%r14 |
- |
- mulq %rbx |
- addq %rax,%r15 |
- movq (%rsi),%rax |
- adcq $0,%rdx |
- addq %r15,%r14 |
- movq %rdx,%r15 |
- adcq $0,%r15 |
- |
- leaq 8(%rdi),%rdi |
- |
- decl %ecx |
- jnz .Loop_mul |
- |
- movq %r8,(%rdi) |
- movq %r9,8(%rdi) |
- movq %r10,16(%rdi) |
- movq %r11,24(%rdi) |
- movq %r12,32(%rdi) |
- movq %r13,40(%rdi) |
- movq %r14,48(%rdi) |
- movq %r15,56(%rdi) |
- |
- .byte 0xf3,0xc3 |
-.size __rsaz_512_mul,.-__rsaz_512_mul |
-.globl rsaz_512_scatter4 |
-.hidden rsaz_512_scatter4 |
-.type rsaz_512_scatter4,@function |
-.align 16 |
-rsaz_512_scatter4: |
- leaq (%rdi,%rdx,8),%rdi |
- movl $8,%r9d |
- jmp .Loop_scatter |
-.align 16 |
-.Loop_scatter: |
- movq (%rsi),%rax |
- leaq 8(%rsi),%rsi |
- movq %rax,(%rdi) |
- leaq 128(%rdi),%rdi |
- decl %r9d |
- jnz .Loop_scatter |
- .byte 0xf3,0xc3 |
-.size rsaz_512_scatter4,.-rsaz_512_scatter4 |
- |
-.globl rsaz_512_gather4 |
-.hidden rsaz_512_gather4 |
-.type rsaz_512_gather4,@function |
-.align 16 |
-rsaz_512_gather4: |
- movd %edx,%xmm8 |
- movdqa .Linc+16(%rip),%xmm1 |
- movdqa .Linc(%rip),%xmm0 |
- |
- pshufd $0,%xmm8,%xmm8 |
- movdqa %xmm1,%xmm7 |
- movdqa %xmm1,%xmm2 |
- paddd %xmm0,%xmm1 |
- pcmpeqd %xmm8,%xmm0 |
- movdqa %xmm7,%xmm3 |
- paddd %xmm1,%xmm2 |
- pcmpeqd %xmm8,%xmm1 |
- movdqa %xmm7,%xmm4 |
- paddd %xmm2,%xmm3 |
- pcmpeqd %xmm8,%xmm2 |
- movdqa %xmm7,%xmm5 |
- paddd %xmm3,%xmm4 |
- pcmpeqd %xmm8,%xmm3 |
- movdqa %xmm7,%xmm6 |
- paddd %xmm4,%xmm5 |
- pcmpeqd %xmm8,%xmm4 |
- paddd %xmm5,%xmm6 |
- pcmpeqd %xmm8,%xmm5 |
- paddd %xmm6,%xmm7 |
- pcmpeqd %xmm8,%xmm6 |
- pcmpeqd %xmm8,%xmm7 |
- movl $8,%r9d |
- jmp .Loop_gather |
-.align 16 |
-.Loop_gather: |
- movdqa 0(%rsi),%xmm8 |
- movdqa 16(%rsi),%xmm9 |
- movdqa 32(%rsi),%xmm10 |
- movdqa 48(%rsi),%xmm11 |
- pand %xmm0,%xmm8 |
- movdqa 64(%rsi),%xmm12 |
- pand %xmm1,%xmm9 |
- movdqa 80(%rsi),%xmm13 |
- pand %xmm2,%xmm10 |
- movdqa 96(%rsi),%xmm14 |
- pand %xmm3,%xmm11 |
- movdqa 112(%rsi),%xmm15 |
- leaq 128(%rsi),%rsi |
- pand %xmm4,%xmm12 |
- pand %xmm5,%xmm13 |
- pand %xmm6,%xmm14 |
- pand %xmm7,%xmm15 |
- por %xmm10,%xmm8 |
- por %xmm11,%xmm9 |
- por %xmm12,%xmm8 |
- por %xmm13,%xmm9 |
- por %xmm14,%xmm8 |
- por %xmm15,%xmm9 |
- |
- por %xmm9,%xmm8 |
- pshufd $0x4e,%xmm8,%xmm9 |
- por %xmm9,%xmm8 |
- movq %xmm8,(%rdi) |
- leaq 8(%rdi),%rdi |
- decl %r9d |
- jnz .Loop_gather |
- .byte 0xf3,0xc3 |
-.LSEH_end_rsaz_512_gather4: |
-.size rsaz_512_gather4,.-rsaz_512_gather4 |
- |
-.align 64 |
-.Linc: |
-.long 0,0, 1,1 |
-.long 2,2, 2,2 |
-#endif |