Index: third_party/boringssl/mac-x86/crypto/sha/sha512-586.S |
diff --git a/third_party/boringssl/mac-x86/crypto/sha/sha512-586.S b/third_party/boringssl/mac-x86/crypto/sha/sha512-586.S |
new file mode 100644 |
index 0000000000000000000000000000000000000000..3066100e6ff92929686ce6ce962c57c313cca399 |
--- /dev/null |
+++ b/third_party/boringssl/mac-x86/crypto/sha/sha512-586.S |
@@ -0,0 +1,2833 @@ |
+#if defined(__i386__) |
+.file "sha512-586.S" |
+.text |
+.globl _sha512_block_data_order |
+.private_extern _sha512_block_data_order |
+.align 4 |
+_sha512_block_data_order: |
+L_sha512_block_data_order_begin: |
+ pushl %ebp |
+ pushl %ebx |
+ pushl %esi |
+ pushl %edi |
+ movl 20(%esp),%esi |
+ movl 24(%esp),%edi |
+ movl 28(%esp),%eax |
+ movl %esp,%ebx |
+ call L000pic_point |
+L000pic_point: |
+ popl %ebp |
+ leal L001K512-L000pic_point(%ebp),%ebp |
+ subl $16,%esp |
+ andl $-64,%esp |
+ shll $7,%eax |
+ addl %edi,%eax |
+ movl %esi,(%esp) |
+ movl %edi,4(%esp) |
+ movl %eax,8(%esp) |
+ movl %ebx,12(%esp) |
+ movl L_OPENSSL_ia32cap_P$non_lazy_ptr-L001K512(%ebp),%edx |
+ movl (%edx),%ecx |
+ testl $67108864,%ecx |
+ jz L002loop_x86 |
+ movl 4(%edx),%edx |
+ movq (%esi),%mm0 |
+ andl $16777216,%ecx |
+ movq 8(%esi),%mm1 |
+ andl $512,%edx |
+ movq 16(%esi),%mm2 |
+ orl %edx,%ecx |
+ movq 24(%esi),%mm3 |
+ movq 32(%esi),%mm4 |
+ movq 40(%esi),%mm5 |
+ movq 48(%esi),%mm6 |
+ movq 56(%esi),%mm7 |
+ cmpl $16777728,%ecx |
+ je L003SSSE3 |
+ subl $80,%esp |
+ jmp L004loop_sse2 |
+.align 4,0x90 |
+L004loop_sse2: |
+ movq %mm1,8(%esp) |
+ movq %mm2,16(%esp) |
+ movq %mm3,24(%esp) |
+ movq %mm5,40(%esp) |
+ movq %mm6,48(%esp) |
+ pxor %mm1,%mm2 |
+ movq %mm7,56(%esp) |
+ movq %mm0,%mm3 |
+ movl (%edi),%eax |
+ movl 4(%edi),%ebx |
+ addl $8,%edi |
+ movl $15,%edx |
+ bswap %eax |
+ bswap %ebx |
+ jmp L00500_14_sse2 |
+.align 4,0x90 |
+L00500_14_sse2: |
+ movd %eax,%mm1 |
+ movl (%edi),%eax |
+ movd %ebx,%mm7 |
+ movl 4(%edi),%ebx |
+ addl $8,%edi |
+ bswap %eax |
+ bswap %ebx |
+ punpckldq %mm1,%mm7 |
+ movq %mm4,%mm1 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,32(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ movq %mm3,%mm0 |
+ movq %mm7,72(%esp) |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm0,(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 56(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ paddq (%ebp),%mm7 |
+ pxor %mm4,%mm3 |
+ movq 24(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm0,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm0,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 8(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ subl $8,%esp |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm0,%mm2 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ pxor %mm7,%mm6 |
+ movq 40(%esp),%mm5 |
+ paddq %mm2,%mm3 |
+ movq %mm0,%mm2 |
+ addl $8,%ebp |
+ paddq %mm6,%mm3 |
+ movq 48(%esp),%mm6 |
+ decl %edx |
+ jnz L00500_14_sse2 |
+ movd %eax,%mm1 |
+ movd %ebx,%mm7 |
+ punpckldq %mm1,%mm7 |
+ movq %mm4,%mm1 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,32(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ movq %mm3,%mm0 |
+ movq %mm7,72(%esp) |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm0,(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 56(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ paddq (%ebp),%mm7 |
+ pxor %mm4,%mm3 |
+ movq 24(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm0,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm0,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 8(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ subl $8,%esp |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm0,%mm2 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ pxor %mm7,%mm6 |
+ movq 192(%esp),%mm7 |
+ paddq %mm2,%mm3 |
+ movq %mm0,%mm2 |
+ addl $8,%ebp |
+ paddq %mm6,%mm3 |
+ pxor %mm0,%mm0 |
+ movl $32,%edx |
+ jmp L00616_79_sse2 |
+.align 4,0x90 |
+L00616_79_sse2: |
+ movq 88(%esp),%mm5 |
+ movq %mm7,%mm1 |
+ psrlq $1,%mm7 |
+ movq %mm5,%mm6 |
+ psrlq $6,%mm5 |
+ psllq $56,%mm1 |
+ paddq %mm3,%mm0 |
+ movq %mm7,%mm3 |
+ psrlq $6,%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $7,%mm1 |
+ pxor %mm7,%mm3 |
+ psrlq $1,%mm7 |
+ pxor %mm1,%mm3 |
+ movq %mm5,%mm1 |
+ psrlq $13,%mm5 |
+ pxor %mm3,%mm7 |
+ psllq $3,%mm6 |
+ pxor %mm5,%mm1 |
+ paddq 200(%esp),%mm7 |
+ pxor %mm6,%mm1 |
+ psrlq $42,%mm5 |
+ paddq 128(%esp),%mm7 |
+ pxor %mm5,%mm1 |
+ psllq $42,%mm6 |
+ movq 40(%esp),%mm5 |
+ pxor %mm6,%mm1 |
+ movq 48(%esp),%mm6 |
+ paddq %mm1,%mm7 |
+ movq %mm4,%mm1 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,32(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ movq %mm7,72(%esp) |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm0,(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 56(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ paddq (%ebp),%mm7 |
+ pxor %mm4,%mm3 |
+ movq 24(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm0,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm0,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 8(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ subl $8,%esp |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm0,%mm2 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ pxor %mm7,%mm6 |
+ movq 192(%esp),%mm7 |
+ paddq %mm6,%mm2 |
+ addl $8,%ebp |
+ movq 88(%esp),%mm5 |
+ movq %mm7,%mm1 |
+ psrlq $1,%mm7 |
+ movq %mm5,%mm6 |
+ psrlq $6,%mm5 |
+ psllq $56,%mm1 |
+ paddq %mm3,%mm2 |
+ movq %mm7,%mm3 |
+ psrlq $6,%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $7,%mm1 |
+ pxor %mm7,%mm3 |
+ psrlq $1,%mm7 |
+ pxor %mm1,%mm3 |
+ movq %mm5,%mm1 |
+ psrlq $13,%mm5 |
+ pxor %mm3,%mm7 |
+ psllq $3,%mm6 |
+ pxor %mm5,%mm1 |
+ paddq 200(%esp),%mm7 |
+ pxor %mm6,%mm1 |
+ psrlq $42,%mm5 |
+ paddq 128(%esp),%mm7 |
+ pxor %mm5,%mm1 |
+ psllq $42,%mm6 |
+ movq 40(%esp),%mm5 |
+ pxor %mm6,%mm1 |
+ movq 48(%esp),%mm6 |
+ paddq %mm1,%mm7 |
+ movq %mm4,%mm1 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,32(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ movq %mm7,72(%esp) |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm2,(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 56(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ paddq (%ebp),%mm7 |
+ pxor %mm4,%mm3 |
+ movq 24(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm2,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm2,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 8(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ subl $8,%esp |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm2,%mm0 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ pxor %mm7,%mm6 |
+ movq 192(%esp),%mm7 |
+ paddq %mm6,%mm0 |
+ addl $8,%ebp |
+ decl %edx |
+ jnz L00616_79_sse2 |
+ paddq %mm3,%mm0 |
+ movq 8(%esp),%mm1 |
+ movq 24(%esp),%mm3 |
+ movq 40(%esp),%mm5 |
+ movq 48(%esp),%mm6 |
+ movq 56(%esp),%mm7 |
+ pxor %mm1,%mm2 |
+ paddq (%esi),%mm0 |
+ paddq 8(%esi),%mm1 |
+ paddq 16(%esi),%mm2 |
+ paddq 24(%esi),%mm3 |
+ paddq 32(%esi),%mm4 |
+ paddq 40(%esi),%mm5 |
+ paddq 48(%esi),%mm6 |
+ paddq 56(%esi),%mm7 |
+ movl $640,%eax |
+ movq %mm0,(%esi) |
+ movq %mm1,8(%esi) |
+ movq %mm2,16(%esi) |
+ movq %mm3,24(%esi) |
+ movq %mm4,32(%esi) |
+ movq %mm5,40(%esi) |
+ movq %mm6,48(%esi) |
+ movq %mm7,56(%esi) |
+ leal (%esp,%eax,1),%esp |
+ subl %eax,%ebp |
+ cmpl 88(%esp),%edi |
+ jb L004loop_sse2 |
+ movl 92(%esp),%esp |
+ emms |
+ popl %edi |
+ popl %esi |
+ popl %ebx |
+ popl %ebp |
+ ret |
+.align 5,0x90 |
+L003SSSE3: |
+ leal -64(%esp),%edx |
+ subl $256,%esp |
+ movdqa 640(%ebp),%xmm1 |
+ movdqu (%edi),%xmm0 |
+.byte 102,15,56,0,193 |
+ movdqa (%ebp),%xmm3 |
+ movdqa %xmm1,%xmm2 |
+ movdqu 16(%edi),%xmm1 |
+ paddq %xmm0,%xmm3 |
+.byte 102,15,56,0,202 |
+ movdqa %xmm3,-128(%edx) |
+ movdqa 16(%ebp),%xmm4 |
+ movdqa %xmm2,%xmm3 |
+ movdqu 32(%edi),%xmm2 |
+ paddq %xmm1,%xmm4 |
+.byte 102,15,56,0,211 |
+ movdqa %xmm4,-112(%edx) |
+ movdqa 32(%ebp),%xmm5 |
+ movdqa %xmm3,%xmm4 |
+ movdqu 48(%edi),%xmm3 |
+ paddq %xmm2,%xmm5 |
+.byte 102,15,56,0,220 |
+ movdqa %xmm5,-96(%edx) |
+ movdqa 48(%ebp),%xmm6 |
+ movdqa %xmm4,%xmm5 |
+ movdqu 64(%edi),%xmm4 |
+ paddq %xmm3,%xmm6 |
+.byte 102,15,56,0,229 |
+ movdqa %xmm6,-80(%edx) |
+ movdqa 64(%ebp),%xmm7 |
+ movdqa %xmm5,%xmm6 |
+ movdqu 80(%edi),%xmm5 |
+ paddq %xmm4,%xmm7 |
+.byte 102,15,56,0,238 |
+ movdqa %xmm7,-64(%edx) |
+ movdqa %xmm0,(%edx) |
+ movdqa 80(%ebp),%xmm0 |
+ movdqa %xmm6,%xmm7 |
+ movdqu 96(%edi),%xmm6 |
+ paddq %xmm5,%xmm0 |
+.byte 102,15,56,0,247 |
+ movdqa %xmm0,-48(%edx) |
+ movdqa %xmm1,16(%edx) |
+ movdqa 96(%ebp),%xmm1 |
+ movdqa %xmm7,%xmm0 |
+ movdqu 112(%edi),%xmm7 |
+ paddq %xmm6,%xmm1 |
+.byte 102,15,56,0,248 |
+ movdqa %xmm1,-32(%edx) |
+ movdqa %xmm2,32(%edx) |
+ movdqa 112(%ebp),%xmm2 |
+ movdqa (%edx),%xmm0 |
+ paddq %xmm7,%xmm2 |
+ movdqa %xmm2,-16(%edx) |
+ nop |
+.align 5,0x90 |
+L007loop_ssse3: |
+ movdqa 16(%edx),%xmm2 |
+ movdqa %xmm3,48(%edx) |
+ leal 128(%ebp),%ebp |
+ movq %mm1,8(%esp) |
+ movl %edi,%ebx |
+ movq %mm2,16(%esp) |
+ leal 128(%edi),%edi |
+ movq %mm3,24(%esp) |
+ cmpl %eax,%edi |
+ movq %mm5,40(%esp) |
+ cmovbl %edi,%ebx |
+ movq %mm6,48(%esp) |
+ movl $4,%ecx |
+ pxor %mm1,%mm2 |
+ movq %mm7,56(%esp) |
+ pxor %mm3,%mm3 |
+ jmp L00800_47_ssse3 |
+.align 5,0x90 |
+L00800_47_ssse3: |
+ movdqa %xmm5,%xmm3 |
+ movdqa %xmm2,%xmm1 |
+.byte 102,15,58,15,208,8 |
+ movdqa %xmm4,(%edx) |
+.byte 102,15,58,15,220,8 |
+ movdqa %xmm2,%xmm4 |
+ psrlq $7,%xmm2 |
+ paddq %xmm3,%xmm0 |
+ movdqa %xmm4,%xmm3 |
+ psrlq $1,%xmm4 |
+ psllq $56,%xmm3 |
+ pxor %xmm4,%xmm2 |
+ psrlq $7,%xmm4 |
+ pxor %xmm3,%xmm2 |
+ psllq $7,%xmm3 |
+ pxor %xmm4,%xmm2 |
+ movdqa %xmm7,%xmm4 |
+ pxor %xmm3,%xmm2 |
+ movdqa %xmm7,%xmm3 |
+ psrlq $6,%xmm4 |
+ paddq %xmm2,%xmm0 |
+ movdqa %xmm7,%xmm2 |
+ psrlq $19,%xmm3 |
+ psllq $3,%xmm2 |
+ pxor %xmm3,%xmm4 |
+ psrlq $42,%xmm3 |
+ pxor %xmm2,%xmm4 |
+ psllq $42,%xmm2 |
+ pxor %xmm3,%xmm4 |
+ movdqa 32(%edx),%xmm3 |
+ pxor %xmm2,%xmm4 |
+ movdqa (%ebp),%xmm2 |
+ movq %mm4,%mm1 |
+ paddq %xmm4,%xmm0 |
+ movq -128(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,32(%esp) |
+ paddq %xmm0,%xmm2 |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm0 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm0,(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 56(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 24(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm0,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm0,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 8(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm0,%mm2 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ pxor %mm7,%mm6 |
+ movq 32(%esp),%mm5 |
+ paddq %mm6,%mm2 |
+ movq 40(%esp),%mm6 |
+ movq %mm4,%mm1 |
+ movq -120(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,24(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm2 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm2,56(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 48(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 16(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm2,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm2,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq (%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm2,%mm0 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ pxor %mm7,%mm6 |
+ movq 24(%esp),%mm5 |
+ paddq %mm6,%mm0 |
+ movq 32(%esp),%mm6 |
+ movdqa %xmm2,-128(%edx) |
+ movdqa %xmm6,%xmm4 |
+ movdqa %xmm3,%xmm2 |
+.byte 102,15,58,15,217,8 |
+ movdqa %xmm5,16(%edx) |
+.byte 102,15,58,15,229,8 |
+ movdqa %xmm3,%xmm5 |
+ psrlq $7,%xmm3 |
+ paddq %xmm4,%xmm1 |
+ movdqa %xmm5,%xmm4 |
+ psrlq $1,%xmm5 |
+ psllq $56,%xmm4 |
+ pxor %xmm5,%xmm3 |
+ psrlq $7,%xmm5 |
+ pxor %xmm4,%xmm3 |
+ psllq $7,%xmm4 |
+ pxor %xmm5,%xmm3 |
+ movdqa %xmm0,%xmm5 |
+ pxor %xmm4,%xmm3 |
+ movdqa %xmm0,%xmm4 |
+ psrlq $6,%xmm5 |
+ paddq %xmm3,%xmm1 |
+ movdqa %xmm0,%xmm3 |
+ psrlq $19,%xmm4 |
+ psllq $3,%xmm3 |
+ pxor %xmm4,%xmm5 |
+ psrlq $42,%xmm4 |
+ pxor %xmm3,%xmm5 |
+ psllq $42,%xmm3 |
+ pxor %xmm4,%xmm5 |
+ movdqa 48(%edx),%xmm4 |
+ pxor %xmm3,%xmm5 |
+ movdqa 16(%ebp),%xmm3 |
+ movq %mm4,%mm1 |
+ paddq %xmm5,%xmm1 |
+ movq -112(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,16(%esp) |
+ paddq %xmm1,%xmm3 |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm0 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm0,48(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 40(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 8(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm0,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm0,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 56(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm0,%mm2 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ pxor %mm7,%mm6 |
+ movq 16(%esp),%mm5 |
+ paddq %mm6,%mm2 |
+ movq 24(%esp),%mm6 |
+ movq %mm4,%mm1 |
+ movq -104(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,8(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm2 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm2,40(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 32(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq (%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm2,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm2,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 48(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm2,%mm0 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ pxor %mm7,%mm6 |
+ movq 8(%esp),%mm5 |
+ paddq %mm6,%mm0 |
+ movq 16(%esp),%mm6 |
+ movdqa %xmm3,-112(%edx) |
+ movdqa %xmm7,%xmm5 |
+ movdqa %xmm4,%xmm3 |
+.byte 102,15,58,15,226,8 |
+ movdqa %xmm6,32(%edx) |
+.byte 102,15,58,15,238,8 |
+ movdqa %xmm4,%xmm6 |
+ psrlq $7,%xmm4 |
+ paddq %xmm5,%xmm2 |
+ movdqa %xmm6,%xmm5 |
+ psrlq $1,%xmm6 |
+ psllq $56,%xmm5 |
+ pxor %xmm6,%xmm4 |
+ psrlq $7,%xmm6 |
+ pxor %xmm5,%xmm4 |
+ psllq $7,%xmm5 |
+ pxor %xmm6,%xmm4 |
+ movdqa %xmm1,%xmm6 |
+ pxor %xmm5,%xmm4 |
+ movdqa %xmm1,%xmm5 |
+ psrlq $6,%xmm6 |
+ paddq %xmm4,%xmm2 |
+ movdqa %xmm1,%xmm4 |
+ psrlq $19,%xmm5 |
+ psllq $3,%xmm4 |
+ pxor %xmm5,%xmm6 |
+ psrlq $42,%xmm5 |
+ pxor %xmm4,%xmm6 |
+ psllq $42,%xmm4 |
+ pxor %xmm5,%xmm6 |
+ movdqa (%edx),%xmm5 |
+ pxor %xmm4,%xmm6 |
+ movdqa 32(%ebp),%xmm4 |
+ movq %mm4,%mm1 |
+ paddq %xmm6,%xmm2 |
+ movq -96(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,(%esp) |
+ paddq %xmm2,%xmm4 |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm0 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm0,32(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 24(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 56(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm0,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm0,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 40(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm0,%mm2 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ pxor %mm7,%mm6 |
+ movq (%esp),%mm5 |
+ paddq %mm6,%mm2 |
+ movq 8(%esp),%mm6 |
+ movq %mm4,%mm1 |
+ movq -88(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,56(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm2 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm2,24(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 16(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 48(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm2,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm2,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 32(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm2,%mm0 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ pxor %mm7,%mm6 |
+ movq 56(%esp),%mm5 |
+ paddq %mm6,%mm0 |
+ movq (%esp),%mm6 |
+ movdqa %xmm4,-96(%edx) |
+ movdqa %xmm0,%xmm6 |
+ movdqa %xmm5,%xmm4 |
+.byte 102,15,58,15,235,8 |
+ movdqa %xmm7,48(%edx) |
+.byte 102,15,58,15,247,8 |
+ movdqa %xmm5,%xmm7 |
+ psrlq $7,%xmm5 |
+ paddq %xmm6,%xmm3 |
+ movdqa %xmm7,%xmm6 |
+ psrlq $1,%xmm7 |
+ psllq $56,%xmm6 |
+ pxor %xmm7,%xmm5 |
+ psrlq $7,%xmm7 |
+ pxor %xmm6,%xmm5 |
+ psllq $7,%xmm6 |
+ pxor %xmm7,%xmm5 |
+ movdqa %xmm2,%xmm7 |
+ pxor %xmm6,%xmm5 |
+ movdqa %xmm2,%xmm6 |
+ psrlq $6,%xmm7 |
+ paddq %xmm5,%xmm3 |
+ movdqa %xmm2,%xmm5 |
+ psrlq $19,%xmm6 |
+ psllq $3,%xmm5 |
+ pxor %xmm6,%xmm7 |
+ psrlq $42,%xmm6 |
+ pxor %xmm5,%xmm7 |
+ psllq $42,%xmm5 |
+ pxor %xmm6,%xmm7 |
+ movdqa 16(%edx),%xmm6 |
+ pxor %xmm5,%xmm7 |
+ movdqa 48(%ebp),%xmm5 |
+ movq %mm4,%mm1 |
+ paddq %xmm7,%xmm3 |
+ movq -80(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,48(%esp) |
+ paddq %xmm3,%xmm5 |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm0 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm0,16(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 8(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 40(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm0,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm0,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 24(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm0,%mm2 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ pxor %mm7,%mm6 |
+ movq 48(%esp),%mm5 |
+ paddq %mm6,%mm2 |
+ movq 56(%esp),%mm6 |
+ movq %mm4,%mm1 |
+ movq -72(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,40(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm2 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm2,8(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq (%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 32(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm2,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm2,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 16(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm2,%mm0 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ pxor %mm7,%mm6 |
+ movq 40(%esp),%mm5 |
+ paddq %mm6,%mm0 |
+ movq 48(%esp),%mm6 |
+ movdqa %xmm5,-80(%edx) |
+ movdqa %xmm1,%xmm7 |
+ movdqa %xmm6,%xmm5 |
+.byte 102,15,58,15,244,8 |
+ movdqa %xmm0,(%edx) |
+.byte 102,15,58,15,248,8 |
+ movdqa %xmm6,%xmm0 |
+ psrlq $7,%xmm6 |
+ paddq %xmm7,%xmm4 |
+ movdqa %xmm0,%xmm7 |
+ psrlq $1,%xmm0 |
+ psllq $56,%xmm7 |
+ pxor %xmm0,%xmm6 |
+ psrlq $7,%xmm0 |
+ pxor %xmm7,%xmm6 |
+ psllq $7,%xmm7 |
+ pxor %xmm0,%xmm6 |
+ movdqa %xmm3,%xmm0 |
+ pxor %xmm7,%xmm6 |
+ movdqa %xmm3,%xmm7 |
+ psrlq $6,%xmm0 |
+ paddq %xmm6,%xmm4 |
+ movdqa %xmm3,%xmm6 |
+ psrlq $19,%xmm7 |
+ psllq $3,%xmm6 |
+ pxor %xmm7,%xmm0 |
+ psrlq $42,%xmm7 |
+ pxor %xmm6,%xmm0 |
+ psllq $42,%xmm6 |
+ pxor %xmm7,%xmm0 |
+ movdqa 32(%edx),%xmm7 |
+ pxor %xmm6,%xmm0 |
+ movdqa 64(%ebp),%xmm6 |
+ movq %mm4,%mm1 |
+ paddq %xmm0,%xmm4 |
+ movq -64(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,32(%esp) |
+ paddq %xmm4,%xmm6 |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm0 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm0,(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 56(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 24(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm0,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm0,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 8(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm0,%mm2 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ pxor %mm7,%mm6 |
+ movq 32(%esp),%mm5 |
+ paddq %mm6,%mm2 |
+ movq 40(%esp),%mm6 |
+ movq %mm4,%mm1 |
+ movq -56(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,24(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm2 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm2,56(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 48(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 16(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm2,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm2,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq (%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm2,%mm0 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ pxor %mm7,%mm6 |
+ movq 24(%esp),%mm5 |
+ paddq %mm6,%mm0 |
+ movq 32(%esp),%mm6 |
+ movdqa %xmm6,-64(%edx) |
+ movdqa %xmm2,%xmm0 |
+ movdqa %xmm7,%xmm6 |
+.byte 102,15,58,15,253,8 |
+ movdqa %xmm1,16(%edx) |
+.byte 102,15,58,15,193,8 |
+ movdqa %xmm7,%xmm1 |
+ psrlq $7,%xmm7 |
+ paddq %xmm0,%xmm5 |
+ movdqa %xmm1,%xmm0 |
+ psrlq $1,%xmm1 |
+ psllq $56,%xmm0 |
+ pxor %xmm1,%xmm7 |
+ psrlq $7,%xmm1 |
+ pxor %xmm0,%xmm7 |
+ psllq $7,%xmm0 |
+ pxor %xmm1,%xmm7 |
+ movdqa %xmm4,%xmm1 |
+ pxor %xmm0,%xmm7 |
+ movdqa %xmm4,%xmm0 |
+ psrlq $6,%xmm1 |
+ paddq %xmm7,%xmm5 |
+ movdqa %xmm4,%xmm7 |
+ psrlq $19,%xmm0 |
+ psllq $3,%xmm7 |
+ pxor %xmm0,%xmm1 |
+ psrlq $42,%xmm0 |
+ pxor %xmm7,%xmm1 |
+ psllq $42,%xmm7 |
+ pxor %xmm0,%xmm1 |
+ movdqa 48(%edx),%xmm0 |
+ pxor %xmm7,%xmm1 |
+ movdqa 80(%ebp),%xmm7 |
+ movq %mm4,%mm1 |
+ paddq %xmm1,%xmm5 |
+ movq -48(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,16(%esp) |
+ paddq %xmm5,%xmm7 |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm0 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm0,48(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 40(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 8(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm0,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm0,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 56(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm0,%mm2 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ pxor %mm7,%mm6 |
+ movq 16(%esp),%mm5 |
+ paddq %mm6,%mm2 |
+ movq 24(%esp),%mm6 |
+ movq %mm4,%mm1 |
+ movq -40(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,8(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm2 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm2,40(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 32(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq (%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm2,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm2,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 48(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm2,%mm0 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ pxor %mm7,%mm6 |
+ movq 8(%esp),%mm5 |
+ paddq %mm6,%mm0 |
+ movq 16(%esp),%mm6 |
+ movdqa %xmm7,-48(%edx) |
+ movdqa %xmm3,%xmm1 |
+ movdqa %xmm0,%xmm7 |
+.byte 102,15,58,15,198,8 |
+ movdqa %xmm2,32(%edx) |
+.byte 102,15,58,15,202,8 |
+ movdqa %xmm0,%xmm2 |
+ psrlq $7,%xmm0 |
+ paddq %xmm1,%xmm6 |
+ movdqa %xmm2,%xmm1 |
+ psrlq $1,%xmm2 |
+ psllq $56,%xmm1 |
+ pxor %xmm2,%xmm0 |
+ psrlq $7,%xmm2 |
+ pxor %xmm1,%xmm0 |
+ psllq $7,%xmm1 |
+ pxor %xmm2,%xmm0 |
+ movdqa %xmm5,%xmm2 |
+ pxor %xmm1,%xmm0 |
+ movdqa %xmm5,%xmm1 |
+ psrlq $6,%xmm2 |
+ paddq %xmm0,%xmm6 |
+ movdqa %xmm5,%xmm0 |
+ psrlq $19,%xmm1 |
+ psllq $3,%xmm0 |
+ pxor %xmm1,%xmm2 |
+ psrlq $42,%xmm1 |
+ pxor %xmm0,%xmm2 |
+ psllq $42,%xmm0 |
+ pxor %xmm1,%xmm2 |
+ movdqa (%edx),%xmm1 |
+ pxor %xmm0,%xmm2 |
+ movdqa 96(%ebp),%xmm0 |
+ movq %mm4,%mm1 |
+ paddq %xmm2,%xmm6 |
+ movq -32(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,(%esp) |
+ paddq %xmm6,%xmm0 |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm0 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm0,32(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 24(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 56(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm0,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm0,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 40(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm0,%mm2 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ pxor %mm7,%mm6 |
+ movq (%esp),%mm5 |
+ paddq %mm6,%mm2 |
+ movq 8(%esp),%mm6 |
+ movq %mm4,%mm1 |
+ movq -24(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,56(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm2 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm2,24(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 16(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 48(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm2,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm2,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 32(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm2,%mm0 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ pxor %mm7,%mm6 |
+ movq 56(%esp),%mm5 |
+ paddq %mm6,%mm0 |
+ movq (%esp),%mm6 |
+ movdqa %xmm0,-32(%edx) |
+ movdqa %xmm4,%xmm2 |
+ movdqa %xmm1,%xmm0 |
+.byte 102,15,58,15,207,8 |
+ movdqa %xmm3,48(%edx) |
+.byte 102,15,58,15,211,8 |
+ movdqa %xmm1,%xmm3 |
+ psrlq $7,%xmm1 |
+ paddq %xmm2,%xmm7 |
+ movdqa %xmm3,%xmm2 |
+ psrlq $1,%xmm3 |
+ psllq $56,%xmm2 |
+ pxor %xmm3,%xmm1 |
+ psrlq $7,%xmm3 |
+ pxor %xmm2,%xmm1 |
+ psllq $7,%xmm2 |
+ pxor %xmm3,%xmm1 |
+ movdqa %xmm6,%xmm3 |
+ pxor %xmm2,%xmm1 |
+ movdqa %xmm6,%xmm2 |
+ psrlq $6,%xmm3 |
+ paddq %xmm1,%xmm7 |
+ movdqa %xmm6,%xmm1 |
+ psrlq $19,%xmm2 |
+ psllq $3,%xmm1 |
+ pxor %xmm2,%xmm3 |
+ psrlq $42,%xmm2 |
+ pxor %xmm1,%xmm3 |
+ psllq $42,%xmm1 |
+ pxor %xmm2,%xmm3 |
+ movdqa 16(%edx),%xmm2 |
+ pxor %xmm1,%xmm3 |
+ movdqa 112(%ebp),%xmm1 |
+ movq %mm4,%mm1 |
+ paddq %xmm3,%xmm7 |
+ movq -16(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,48(%esp) |
+ paddq %xmm7,%xmm1 |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm0 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm0,16(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 8(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 40(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm0,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm0,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 24(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm0,%mm2 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ pxor %mm7,%mm6 |
+ movq 48(%esp),%mm5 |
+ paddq %mm6,%mm2 |
+ movq 56(%esp),%mm6 |
+ movq %mm4,%mm1 |
+ movq -8(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,40(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm2 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm2,8(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq (%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 32(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm2,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm2,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 16(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm2,%mm0 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ pxor %mm7,%mm6 |
+ movq 40(%esp),%mm5 |
+ paddq %mm6,%mm0 |
+ movq 48(%esp),%mm6 |
+ movdqa %xmm1,-16(%edx) |
+ leal 128(%ebp),%ebp |
+ decl %ecx |
+ jnz L00800_47_ssse3 |
+ movdqa (%ebp),%xmm1 |
+ leal -640(%ebp),%ebp |
+ movdqu (%ebx),%xmm0 |
+.byte 102,15,56,0,193 |
+ movdqa (%ebp),%xmm3 |
+ movdqa %xmm1,%xmm2 |
+ movdqu 16(%ebx),%xmm1 |
+ paddq %xmm0,%xmm3 |
+.byte 102,15,56,0,202 |
+ movq %mm4,%mm1 |
+ movq -128(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,32(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm0 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm0,(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 56(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 24(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm0,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm0,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 8(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm0,%mm2 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ pxor %mm7,%mm6 |
+ movq 32(%esp),%mm5 |
+ paddq %mm6,%mm2 |
+ movq 40(%esp),%mm6 |
+ movq %mm4,%mm1 |
+ movq -120(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,24(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm2 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm2,56(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 48(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 16(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm2,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm2,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq (%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm2,%mm0 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ pxor %mm7,%mm6 |
+ movq 24(%esp),%mm5 |
+ paddq %mm6,%mm0 |
+ movq 32(%esp),%mm6 |
+ movdqa %xmm3,-128(%edx) |
+ movdqa 16(%ebp),%xmm4 |
+ movdqa %xmm2,%xmm3 |
+ movdqu 32(%ebx),%xmm2 |
+ paddq %xmm1,%xmm4 |
+.byte 102,15,56,0,211 |
+ movq %mm4,%mm1 |
+ movq -112(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,16(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm0 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm0,48(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 40(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 8(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm0,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm0,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 56(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm0,%mm2 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ pxor %mm7,%mm6 |
+ movq 16(%esp),%mm5 |
+ paddq %mm6,%mm2 |
+ movq 24(%esp),%mm6 |
+ movq %mm4,%mm1 |
+ movq -104(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,8(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm2 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm2,40(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 32(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq (%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm2,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm2,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 48(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm2,%mm0 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ pxor %mm7,%mm6 |
+ movq 8(%esp),%mm5 |
+ paddq %mm6,%mm0 |
+ movq 16(%esp),%mm6 |
+ movdqa %xmm4,-112(%edx) |
+ movdqa 32(%ebp),%xmm5 |
+ movdqa %xmm3,%xmm4 |
+ movdqu 48(%ebx),%xmm3 |
+ paddq %xmm2,%xmm5 |
+.byte 102,15,56,0,220 |
+ movq %mm4,%mm1 |
+ movq -96(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm0 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm0,32(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 24(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 56(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm0,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm0,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 40(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm0,%mm2 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ pxor %mm7,%mm6 |
+ movq (%esp),%mm5 |
+ paddq %mm6,%mm2 |
+ movq 8(%esp),%mm6 |
+ movq %mm4,%mm1 |
+ movq -88(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,56(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm2 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm2,24(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 16(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 48(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm2,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm2,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 32(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm2,%mm0 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ pxor %mm7,%mm6 |
+ movq 56(%esp),%mm5 |
+ paddq %mm6,%mm0 |
+ movq (%esp),%mm6 |
+ movdqa %xmm5,-96(%edx) |
+ movdqa 48(%ebp),%xmm6 |
+ movdqa %xmm4,%xmm5 |
+ movdqu 64(%ebx),%xmm4 |
+ paddq %xmm3,%xmm6 |
+.byte 102,15,56,0,229 |
+ movq %mm4,%mm1 |
+ movq -80(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,48(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm0 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm0,16(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 8(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 40(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm0,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm0,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 24(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm0,%mm2 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ pxor %mm7,%mm6 |
+ movq 48(%esp),%mm5 |
+ paddq %mm6,%mm2 |
+ movq 56(%esp),%mm6 |
+ movq %mm4,%mm1 |
+ movq -72(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,40(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm2 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm2,8(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq (%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 32(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm2,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm2,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 16(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm2,%mm0 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ pxor %mm7,%mm6 |
+ movq 40(%esp),%mm5 |
+ paddq %mm6,%mm0 |
+ movq 48(%esp),%mm6 |
+ movdqa %xmm6,-80(%edx) |
+ movdqa 64(%ebp),%xmm7 |
+ movdqa %xmm5,%xmm6 |
+ movdqu 80(%ebx),%xmm5 |
+ paddq %xmm4,%xmm7 |
+.byte 102,15,56,0,238 |
+ movq %mm4,%mm1 |
+ movq -64(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,32(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm0 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm0,(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 56(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 24(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm0,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm0,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 8(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm0,%mm2 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ pxor %mm7,%mm6 |
+ movq 32(%esp),%mm5 |
+ paddq %mm6,%mm2 |
+ movq 40(%esp),%mm6 |
+ movq %mm4,%mm1 |
+ movq -56(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,24(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm2 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm2,56(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 48(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 16(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm2,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm2,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq (%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm2,%mm0 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ pxor %mm7,%mm6 |
+ movq 24(%esp),%mm5 |
+ paddq %mm6,%mm0 |
+ movq 32(%esp),%mm6 |
+ movdqa %xmm7,-64(%edx) |
+ movdqa %xmm0,(%edx) |
+ movdqa 80(%ebp),%xmm0 |
+ movdqa %xmm6,%xmm7 |
+ movdqu 96(%ebx),%xmm6 |
+ paddq %xmm5,%xmm0 |
+.byte 102,15,56,0,247 |
+ movq %mm4,%mm1 |
+ movq -48(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,16(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm0 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm0,48(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 40(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 8(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm0,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm0,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 56(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm0,%mm2 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ pxor %mm7,%mm6 |
+ movq 16(%esp),%mm5 |
+ paddq %mm6,%mm2 |
+ movq 24(%esp),%mm6 |
+ movq %mm4,%mm1 |
+ movq -40(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,8(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm2 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm2,40(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 32(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq (%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm2,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm2,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 48(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm2,%mm0 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ pxor %mm7,%mm6 |
+ movq 8(%esp),%mm5 |
+ paddq %mm6,%mm0 |
+ movq 16(%esp),%mm6 |
+ movdqa %xmm0,-48(%edx) |
+ movdqa %xmm1,16(%edx) |
+ movdqa 96(%ebp),%xmm1 |
+ movdqa %xmm7,%xmm0 |
+ movdqu 112(%ebx),%xmm7 |
+ paddq %xmm6,%xmm1 |
+.byte 102,15,56,0,248 |
+ movq %mm4,%mm1 |
+ movq -32(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm0 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm0,32(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 24(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 56(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm0,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm0,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 40(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm0,%mm2 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ pxor %mm7,%mm6 |
+ movq (%esp),%mm5 |
+ paddq %mm6,%mm2 |
+ movq 8(%esp),%mm6 |
+ movq %mm4,%mm1 |
+ movq -24(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,56(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm2 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm2,24(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 16(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 48(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm2,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm2,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 32(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm2,%mm0 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ pxor %mm7,%mm6 |
+ movq 56(%esp),%mm5 |
+ paddq %mm6,%mm0 |
+ movq (%esp),%mm6 |
+ movdqa %xmm1,-32(%edx) |
+ movdqa %xmm2,32(%edx) |
+ movdqa 112(%ebp),%xmm2 |
+ movdqa (%edx),%xmm0 |
+ paddq %xmm7,%xmm2 |
+ movq %mm4,%mm1 |
+ movq -16(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,48(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm0 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm0,16(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq 8(%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 40(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm0,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm0,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 24(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm0,%mm2 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ pxor %mm7,%mm6 |
+ movq 48(%esp),%mm5 |
+ paddq %mm6,%mm2 |
+ movq 56(%esp),%mm6 |
+ movq %mm4,%mm1 |
+ movq -8(%edx),%mm7 |
+ pxor %mm6,%mm5 |
+ psrlq $14,%mm1 |
+ movq %mm4,40(%esp) |
+ pand %mm4,%mm5 |
+ psllq $23,%mm4 |
+ paddq %mm3,%mm2 |
+ movq %mm1,%mm3 |
+ psrlq $4,%mm1 |
+ pxor %mm6,%mm5 |
+ pxor %mm4,%mm3 |
+ psllq $23,%mm4 |
+ pxor %mm1,%mm3 |
+ movq %mm2,8(%esp) |
+ paddq %mm5,%mm7 |
+ pxor %mm4,%mm3 |
+ psrlq $23,%mm1 |
+ paddq (%esp),%mm7 |
+ pxor %mm1,%mm3 |
+ psllq $4,%mm4 |
+ pxor %mm4,%mm3 |
+ movq 32(%esp),%mm4 |
+ paddq %mm7,%mm3 |
+ movq %mm2,%mm5 |
+ psrlq $28,%mm5 |
+ paddq %mm3,%mm4 |
+ movq %mm2,%mm6 |
+ movq %mm5,%mm7 |
+ psllq $25,%mm6 |
+ movq 16(%esp),%mm1 |
+ psrlq $6,%mm5 |
+ pxor %mm6,%mm7 |
+ psllq $5,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm2 |
+ psrlq $5,%mm5 |
+ pxor %mm6,%mm7 |
+ pand %mm2,%mm0 |
+ psllq $6,%mm6 |
+ pxor %mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ pxor %mm7,%mm6 |
+ movq 40(%esp),%mm5 |
+ paddq %mm6,%mm0 |
+ movq 48(%esp),%mm6 |
+ movdqa %xmm2,-16(%edx) |
+ movq 8(%esp),%mm1 |
+ paddq %mm3,%mm0 |
+ movq 24(%esp),%mm3 |
+ movq 56(%esp),%mm7 |
+ pxor %mm1,%mm2 |
+ paddq (%esi),%mm0 |
+ paddq 8(%esi),%mm1 |
+ paddq 16(%esi),%mm2 |
+ paddq 24(%esi),%mm3 |
+ paddq 32(%esi),%mm4 |
+ paddq 40(%esi),%mm5 |
+ paddq 48(%esi),%mm6 |
+ paddq 56(%esi),%mm7 |
+ movq %mm0,(%esi) |
+ movq %mm1,8(%esi) |
+ movq %mm2,16(%esi) |
+ movq %mm3,24(%esi) |
+ movq %mm4,32(%esi) |
+ movq %mm5,40(%esi) |
+ movq %mm6,48(%esi) |
+ movq %mm7,56(%esi) |
+ cmpl %eax,%edi |
+ jb L007loop_ssse3 |
+ movl 76(%edx),%esp |
+ emms |
+ popl %edi |
+ popl %esi |
+ popl %ebx |
+ popl %ebp |
+ ret |
+.align 4,0x90 |
+L002loop_x86: |
+ movl (%edi),%eax |
+ movl 4(%edi),%ebx |
+ movl 8(%edi),%ecx |
+ movl 12(%edi),%edx |
+ bswap %eax |
+ bswap %ebx |
+ bswap %ecx |
+ bswap %edx |
+ pushl %eax |
+ pushl %ebx |
+ pushl %ecx |
+ pushl %edx |
+ movl 16(%edi),%eax |
+ movl 20(%edi),%ebx |
+ movl 24(%edi),%ecx |
+ movl 28(%edi),%edx |
+ bswap %eax |
+ bswap %ebx |
+ bswap %ecx |
+ bswap %edx |
+ pushl %eax |
+ pushl %ebx |
+ pushl %ecx |
+ pushl %edx |
+ movl 32(%edi),%eax |
+ movl 36(%edi),%ebx |
+ movl 40(%edi),%ecx |
+ movl 44(%edi),%edx |
+ bswap %eax |
+ bswap %ebx |
+ bswap %ecx |
+ bswap %edx |
+ pushl %eax |
+ pushl %ebx |
+ pushl %ecx |
+ pushl %edx |
+ movl 48(%edi),%eax |
+ movl 52(%edi),%ebx |
+ movl 56(%edi),%ecx |
+ movl 60(%edi),%edx |
+ bswap %eax |
+ bswap %ebx |
+ bswap %ecx |
+ bswap %edx |
+ pushl %eax |
+ pushl %ebx |
+ pushl %ecx |
+ pushl %edx |
+ movl 64(%edi),%eax |
+ movl 68(%edi),%ebx |
+ movl 72(%edi),%ecx |
+ movl 76(%edi),%edx |
+ bswap %eax |
+ bswap %ebx |
+ bswap %ecx |
+ bswap %edx |
+ pushl %eax |
+ pushl %ebx |
+ pushl %ecx |
+ pushl %edx |
+ movl 80(%edi),%eax |
+ movl 84(%edi),%ebx |
+ movl 88(%edi),%ecx |
+ movl 92(%edi),%edx |
+ bswap %eax |
+ bswap %ebx |
+ bswap %ecx |
+ bswap %edx |
+ pushl %eax |
+ pushl %ebx |
+ pushl %ecx |
+ pushl %edx |
+ movl 96(%edi),%eax |
+ movl 100(%edi),%ebx |
+ movl 104(%edi),%ecx |
+ movl 108(%edi),%edx |
+ bswap %eax |
+ bswap %ebx |
+ bswap %ecx |
+ bswap %edx |
+ pushl %eax |
+ pushl %ebx |
+ pushl %ecx |
+ pushl %edx |
+ movl 112(%edi),%eax |
+ movl 116(%edi),%ebx |
+ movl 120(%edi),%ecx |
+ movl 124(%edi),%edx |
+ bswap %eax |
+ bswap %ebx |
+ bswap %ecx |
+ bswap %edx |
+ pushl %eax |
+ pushl %ebx |
+ pushl %ecx |
+ pushl %edx |
+ addl $128,%edi |
+ subl $72,%esp |
+ movl %edi,204(%esp) |
+ leal 8(%esp),%edi |
+ movl $16,%ecx |
+.long 2784229001 |
+.align 4,0x90 |
+L00900_15_x86: |
+ movl 40(%esp),%ecx |
+ movl 44(%esp),%edx |
+ movl %ecx,%esi |
+ shrl $9,%ecx |
+ movl %edx,%edi |
+ shrl $9,%edx |
+ movl %ecx,%ebx |
+ shll $14,%esi |
+ movl %edx,%eax |
+ shll $14,%edi |
+ xorl %esi,%ebx |
+ shrl $5,%ecx |
+ xorl %edi,%eax |
+ shrl $5,%edx |
+ xorl %ecx,%eax |
+ shll $4,%esi |
+ xorl %edx,%ebx |
+ shll $4,%edi |
+ xorl %esi,%ebx |
+ shrl $4,%ecx |
+ xorl %edi,%eax |
+ shrl $4,%edx |
+ xorl %ecx,%eax |
+ shll $5,%esi |
+ xorl %edx,%ebx |
+ shll $5,%edi |
+ xorl %esi,%eax |
+ xorl %edi,%ebx |
+ movl 48(%esp),%ecx |
+ movl 52(%esp),%edx |
+ movl 56(%esp),%esi |
+ movl 60(%esp),%edi |
+ addl 64(%esp),%eax |
+ adcl 68(%esp),%ebx |
+ xorl %esi,%ecx |
+ xorl %edi,%edx |
+ andl 40(%esp),%ecx |
+ andl 44(%esp),%edx |
+ addl 192(%esp),%eax |
+ adcl 196(%esp),%ebx |
+ xorl %esi,%ecx |
+ xorl %edi,%edx |
+ movl (%ebp),%esi |
+ movl 4(%ebp),%edi |
+ addl %ecx,%eax |
+ adcl %edx,%ebx |
+ movl 32(%esp),%ecx |
+ movl 36(%esp),%edx |
+ addl %esi,%eax |
+ adcl %edi,%ebx |
+ movl %eax,(%esp) |
+ movl %ebx,4(%esp) |
+ addl %ecx,%eax |
+ adcl %edx,%ebx |
+ movl 8(%esp),%ecx |
+ movl 12(%esp),%edx |
+ movl %eax,32(%esp) |
+ movl %ebx,36(%esp) |
+ movl %ecx,%esi |
+ shrl $2,%ecx |
+ movl %edx,%edi |
+ shrl $2,%edx |
+ movl %ecx,%ebx |
+ shll $4,%esi |
+ movl %edx,%eax |
+ shll $4,%edi |
+ xorl %esi,%ebx |
+ shrl $5,%ecx |
+ xorl %edi,%eax |
+ shrl $5,%edx |
+ xorl %ecx,%ebx |
+ shll $21,%esi |
+ xorl %edx,%eax |
+ shll $21,%edi |
+ xorl %esi,%eax |
+ shrl $21,%ecx |
+ xorl %edi,%ebx |
+ shrl $21,%edx |
+ xorl %ecx,%eax |
+ shll $5,%esi |
+ xorl %edx,%ebx |
+ shll $5,%edi |
+ xorl %esi,%eax |
+ xorl %edi,%ebx |
+ movl 8(%esp),%ecx |
+ movl 12(%esp),%edx |
+ movl 16(%esp),%esi |
+ movl 20(%esp),%edi |
+ addl (%esp),%eax |
+ adcl 4(%esp),%ebx |
+ orl %esi,%ecx |
+ orl %edi,%edx |
+ andl 24(%esp),%ecx |
+ andl 28(%esp),%edx |
+ andl 8(%esp),%esi |
+ andl 12(%esp),%edi |
+ orl %esi,%ecx |
+ orl %edi,%edx |
+ addl %ecx,%eax |
+ adcl %edx,%ebx |
+ movl %eax,(%esp) |
+ movl %ebx,4(%esp) |
+ movb (%ebp),%dl |
+ subl $8,%esp |
+ leal 8(%ebp),%ebp |
+ cmpb $148,%dl |
+ jne L00900_15_x86 |
+.align 4,0x90 |
+L01016_79_x86: |
+ movl 312(%esp),%ecx |
+ movl 316(%esp),%edx |
+ movl %ecx,%esi |
+ shrl $1,%ecx |
+ movl %edx,%edi |
+ shrl $1,%edx |
+ movl %ecx,%eax |
+ shll $24,%esi |
+ movl %edx,%ebx |
+ shll $24,%edi |
+ xorl %esi,%ebx |
+ shrl $6,%ecx |
+ xorl %edi,%eax |
+ shrl $6,%edx |
+ xorl %ecx,%eax |
+ shll $7,%esi |
+ xorl %edx,%ebx |
+ shll $1,%edi |
+ xorl %esi,%ebx |
+ shrl $1,%ecx |
+ xorl %edi,%eax |
+ shrl $1,%edx |
+ xorl %ecx,%eax |
+ shll $6,%edi |
+ xorl %edx,%ebx |
+ xorl %edi,%eax |
+ movl %eax,(%esp) |
+ movl %ebx,4(%esp) |
+ movl 208(%esp),%ecx |
+ movl 212(%esp),%edx |
+ movl %ecx,%esi |
+ shrl $6,%ecx |
+ movl %edx,%edi |
+ shrl $6,%edx |
+ movl %ecx,%eax |
+ shll $3,%esi |
+ movl %edx,%ebx |
+ shll $3,%edi |
+ xorl %esi,%eax |
+ shrl $13,%ecx |
+ xorl %edi,%ebx |
+ shrl $13,%edx |
+ xorl %ecx,%eax |
+ shll $10,%esi |
+ xorl %edx,%ebx |
+ shll $10,%edi |
+ xorl %esi,%ebx |
+ shrl $10,%ecx |
+ xorl %edi,%eax |
+ shrl $10,%edx |
+ xorl %ecx,%ebx |
+ shll $13,%edi |
+ xorl %edx,%eax |
+ xorl %edi,%eax |
+ movl 320(%esp),%ecx |
+ movl 324(%esp),%edx |
+ addl (%esp),%eax |
+ adcl 4(%esp),%ebx |
+ movl 248(%esp),%esi |
+ movl 252(%esp),%edi |
+ addl %ecx,%eax |
+ adcl %edx,%ebx |
+ addl %esi,%eax |
+ adcl %edi,%ebx |
+ movl %eax,192(%esp) |
+ movl %ebx,196(%esp) |
+ movl 40(%esp),%ecx |
+ movl 44(%esp),%edx |
+ movl %ecx,%esi |
+ shrl $9,%ecx |
+ movl %edx,%edi |
+ shrl $9,%edx |
+ movl %ecx,%ebx |
+ shll $14,%esi |
+ movl %edx,%eax |
+ shll $14,%edi |
+ xorl %esi,%ebx |
+ shrl $5,%ecx |
+ xorl %edi,%eax |
+ shrl $5,%edx |
+ xorl %ecx,%eax |
+ shll $4,%esi |
+ xorl %edx,%ebx |
+ shll $4,%edi |
+ xorl %esi,%ebx |
+ shrl $4,%ecx |
+ xorl %edi,%eax |
+ shrl $4,%edx |
+ xorl %ecx,%eax |
+ shll $5,%esi |
+ xorl %edx,%ebx |
+ shll $5,%edi |
+ xorl %esi,%eax |
+ xorl %edi,%ebx |
+ movl 48(%esp),%ecx |
+ movl 52(%esp),%edx |
+ movl 56(%esp),%esi |
+ movl 60(%esp),%edi |
+ addl 64(%esp),%eax |
+ adcl 68(%esp),%ebx |
+ xorl %esi,%ecx |
+ xorl %edi,%edx |
+ andl 40(%esp),%ecx |
+ andl 44(%esp),%edx |
+ addl 192(%esp),%eax |
+ adcl 196(%esp),%ebx |
+ xorl %esi,%ecx |
+ xorl %edi,%edx |
+ movl (%ebp),%esi |
+ movl 4(%ebp),%edi |
+ addl %ecx,%eax |
+ adcl %edx,%ebx |
+ movl 32(%esp),%ecx |
+ movl 36(%esp),%edx |
+ addl %esi,%eax |
+ adcl %edi,%ebx |
+ movl %eax,(%esp) |
+ movl %ebx,4(%esp) |
+ addl %ecx,%eax |
+ adcl %edx,%ebx |
+ movl 8(%esp),%ecx |
+ movl 12(%esp),%edx |
+ movl %eax,32(%esp) |
+ movl %ebx,36(%esp) |
+ movl %ecx,%esi |
+ shrl $2,%ecx |
+ movl %edx,%edi |
+ shrl $2,%edx |
+ movl %ecx,%ebx |
+ shll $4,%esi |
+ movl %edx,%eax |
+ shll $4,%edi |
+ xorl %esi,%ebx |
+ shrl $5,%ecx |
+ xorl %edi,%eax |
+ shrl $5,%edx |
+ xorl %ecx,%ebx |
+ shll $21,%esi |
+ xorl %edx,%eax |
+ shll $21,%edi |
+ xorl %esi,%eax |
+ shrl $21,%ecx |
+ xorl %edi,%ebx |
+ shrl $21,%edx |
+ xorl %ecx,%eax |
+ shll $5,%esi |
+ xorl %edx,%ebx |
+ shll $5,%edi |
+ xorl %esi,%eax |
+ xorl %edi,%ebx |
+ movl 8(%esp),%ecx |
+ movl 12(%esp),%edx |
+ movl 16(%esp),%esi |
+ movl 20(%esp),%edi |
+ addl (%esp),%eax |
+ adcl 4(%esp),%ebx |
+ orl %esi,%ecx |
+ orl %edi,%edx |
+ andl 24(%esp),%ecx |
+ andl 28(%esp),%edx |
+ andl 8(%esp),%esi |
+ andl 12(%esp),%edi |
+ orl %esi,%ecx |
+ orl %edi,%edx |
+ addl %ecx,%eax |
+ adcl %edx,%ebx |
+ movl %eax,(%esp) |
+ movl %ebx,4(%esp) |
+ movb (%ebp),%dl |
+ subl $8,%esp |
+ leal 8(%ebp),%ebp |
+ cmpb $23,%dl |
+ jne L01016_79_x86 |
+ movl 840(%esp),%esi |
+ movl 844(%esp),%edi |
+ movl (%esi),%eax |
+ movl 4(%esi),%ebx |
+ movl 8(%esi),%ecx |
+ movl 12(%esi),%edx |
+ addl 8(%esp),%eax |
+ adcl 12(%esp),%ebx |
+ movl %eax,(%esi) |
+ movl %ebx,4(%esi) |
+ addl 16(%esp),%ecx |
+ adcl 20(%esp),%edx |
+ movl %ecx,8(%esi) |
+ movl %edx,12(%esi) |
+ movl 16(%esi),%eax |
+ movl 20(%esi),%ebx |
+ movl 24(%esi),%ecx |
+ movl 28(%esi),%edx |
+ addl 24(%esp),%eax |
+ adcl 28(%esp),%ebx |
+ movl %eax,16(%esi) |
+ movl %ebx,20(%esi) |
+ addl 32(%esp),%ecx |
+ adcl 36(%esp),%edx |
+ movl %ecx,24(%esi) |
+ movl %edx,28(%esi) |
+ movl 32(%esi),%eax |
+ movl 36(%esi),%ebx |
+ movl 40(%esi),%ecx |
+ movl 44(%esi),%edx |
+ addl 40(%esp),%eax |
+ adcl 44(%esp),%ebx |
+ movl %eax,32(%esi) |
+ movl %ebx,36(%esi) |
+ addl 48(%esp),%ecx |
+ adcl 52(%esp),%edx |
+ movl %ecx,40(%esi) |
+ movl %edx,44(%esi) |
+ movl 48(%esi),%eax |
+ movl 52(%esi),%ebx |
+ movl 56(%esi),%ecx |
+ movl 60(%esi),%edx |
+ addl 56(%esp),%eax |
+ adcl 60(%esp),%ebx |
+ movl %eax,48(%esi) |
+ movl %ebx,52(%esi) |
+ addl 64(%esp),%ecx |
+ adcl 68(%esp),%edx |
+ movl %ecx,56(%esi) |
+ movl %edx,60(%esi) |
+ addl $840,%esp |
+ subl $640,%ebp |
+ cmpl 8(%esp),%edi |
+ jb L002loop_x86 |
+ movl 12(%esp),%esp |
+ popl %edi |
+ popl %esi |
+ popl %ebx |
+ popl %ebp |
+ ret |
+.align 6,0x90 |
+L001K512: |
+.long 3609767458,1116352408 |
+.long 602891725,1899447441 |
+.long 3964484399,3049323471 |
+.long 2173295548,3921009573 |
+.long 4081628472,961987163 |
+.long 3053834265,1508970993 |
+.long 2937671579,2453635748 |
+.long 3664609560,2870763221 |
+.long 2734883394,3624381080 |
+.long 1164996542,310598401 |
+.long 1323610764,607225278 |
+.long 3590304994,1426881987 |
+.long 4068182383,1925078388 |
+.long 991336113,2162078206 |
+.long 633803317,2614888103 |
+.long 3479774868,3248222580 |
+.long 2666613458,3835390401 |
+.long 944711139,4022224774 |
+.long 2341262773,264347078 |
+.long 2007800933,604807628 |
+.long 1495990901,770255983 |
+.long 1856431235,1249150122 |
+.long 3175218132,1555081692 |
+.long 2198950837,1996064986 |
+.long 3999719339,2554220882 |
+.long 766784016,2821834349 |
+.long 2566594879,2952996808 |
+.long 3203337956,3210313671 |
+.long 1034457026,3336571891 |
+.long 2466948901,3584528711 |
+.long 3758326383,113926993 |
+.long 168717936,338241895 |
+.long 1188179964,666307205 |
+.long 1546045734,773529912 |
+.long 1522805485,1294757372 |
+.long 2643833823,1396182291 |
+.long 2343527390,1695183700 |
+.long 1014477480,1986661051 |
+.long 1206759142,2177026350 |
+.long 344077627,2456956037 |
+.long 1290863460,2730485921 |
+.long 3158454273,2820302411 |
+.long 3505952657,3259730800 |
+.long 106217008,3345764771 |
+.long 3606008344,3516065817 |
+.long 1432725776,3600352804 |
+.long 1467031594,4094571909 |
+.long 851169720,275423344 |
+.long 3100823752,430227734 |
+.long 1363258195,506948616 |
+.long 3750685593,659060556 |
+.long 3785050280,883997877 |
+.long 3318307427,958139571 |
+.long 3812723403,1322822218 |
+.long 2003034995,1537002063 |
+.long 3602036899,1747873779 |
+.long 1575990012,1955562222 |
+.long 1125592928,2024104815 |
+.long 2716904306,2227730452 |
+.long 442776044,2361852424 |
+.long 593698344,2428436474 |
+.long 3733110249,2756734187 |
+.long 2999351573,3204031479 |
+.long 3815920427,3329325298 |
+.long 3928383900,3391569614 |
+.long 566280711,3515267271 |
+.long 3454069534,3940187606 |
+.long 4000239992,4118630271 |
+.long 1914138554,116418474 |
+.long 2731055270,174292421 |
+.long 3203993006,289380356 |
+.long 320620315,460393269 |
+.long 587496836,685471733 |
+.long 1086792851,852142971 |
+.long 365543100,1017036298 |
+.long 2618297676,1126000580 |
+.long 3409855158,1288033470 |
+.long 4234509866,1501505948 |
+.long 987167468,1607167915 |
+.long 1246189591,1816402316 |
+.long 67438087,66051 |
+.long 202182159,134810123 |
+.byte 83,72,65,53,49,50,32,98,108,111,99,107,32,116,114,97 |
+.byte 110,115,102,111,114,109,32,102,111,114,32,120,56,54,44,32 |
+.byte 67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97 |
+.byte 112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,103 |
+.byte 62,0 |
+.section __IMPORT,__pointers,non_lazy_symbol_pointers |
+L_OPENSSL_ia32cap_P$non_lazy_ptr: |
+.indirect_symbol _OPENSSL_ia32cap_P |
+.long 0 |
+#endif |