Index: third_party/boringssl/linux-x86/crypto/sha/sha256-586.S |
diff --git a/third_party/boringssl/linux-x86/crypto/sha/sha256-586.S b/third_party/boringssl/linux-x86/crypto/sha/sha256-586.S |
index 196f7f9d58ee49235f2329d7e65f6366ea340ae0..38acbd8374c77f5ceed81e037f3e441c62175a44 100644 |
--- a/third_party/boringssl/linux-x86/crypto/sha/sha256-586.S |
+++ b/third_party/boringssl/linux-x86/crypto/sha/sha256-586.S |
@@ -40,12 +40,13 @@ sha256_block_data_order: |
orl %ebx,%ecx |
andl $1342177280,%ecx |
cmpl $1342177280,%ecx |
+ je .L004AVX |
testl $512,%ebx |
- jnz .L004SSSE3 |
+ jnz .L005SSSE3 |
.L003no_xmm: |
subl %edi,%eax |
cmpl $256,%eax |
- jae .L005unrolled |
+ jae .L006unrolled |
jmp .L002loop |
.align 16 |
.L002loop: |
@@ -117,7 +118,7 @@ sha256_block_data_order: |
movl %ecx,28(%esp) |
movl %edi,32(%esp) |
.align 16 |
-.L00600_15: |
+.L00700_15: |
movl %edx,%ecx |
movl 24(%esp),%esi |
rorl $14,%ecx |
@@ -155,11 +156,11 @@ sha256_block_data_order: |
addl $4,%ebp |
addl %ebx,%eax |
cmpl $3248222580,%esi |
- jne .L00600_15 |
+ jne .L00700_15 |
movl 156(%esp),%ecx |
- jmp .L00716_63 |
+ jmp .L00816_63 |
.align 16 |
-.L00716_63: |
+.L00816_63: |
movl %ecx,%ebx |
movl 104(%esp),%esi |
rorl $11,%ecx |
@@ -214,7 +215,7 @@ sha256_block_data_order: |
addl $4,%ebp |
addl %ebx,%eax |
cmpl $3329325298,%esi |
- jne .L00716_63 |
+ jne .L00816_63 |
movl 356(%esp),%esi |
movl 8(%esp),%ebx |
movl 16(%esp),%ecx |
@@ -258,7 +259,7 @@ sha256_block_data_order: |
.byte 112,112,114,111,64,111,112,101,110,115,115,108,46,111,114,103 |
.byte 62,0 |
.align 16 |
-.L005unrolled: |
+.L006unrolled: |
leal -96(%esp),%esp |
movl (%esi),%eax |
movl 4(%esi),%ebp |
@@ -275,9 +276,9 @@ sha256_block_data_order: |
movl %ebx,20(%esp) |
movl %ecx,24(%esp) |
movl %esi,28(%esp) |
- jmp .L008grand_loop |
+ jmp .L009grand_loop |
.align 16 |
-.L008grand_loop: |
+.L009grand_loop: |
movl (%edi),%ebx |
movl 4(%edi),%ecx |
bswap %ebx |
@@ -3157,7 +3158,7 @@ sha256_block_data_order: |
movl %ebx,24(%esp) |
movl %ecx,28(%esp) |
cmpl 104(%esp),%edi |
- jb .L008grand_loop |
+ jb .L009grand_loop |
movl 108(%esp),%esp |
popl %edi |
popl %esi |
@@ -3165,7 +3166,7 @@ sha256_block_data_order: |
popl %ebp |
ret |
.align 32 |
-.L004SSSE3: |
+.L005SSSE3: |
leal -96(%esp),%esp |
movl (%esi),%eax |
movl 4(%esi),%ebx |
@@ -3184,9 +3185,9 @@ sha256_block_data_order: |
movl %ecx,24(%esp) |
movl %esi,28(%esp) |
movdqa 256(%ebp),%xmm7 |
- jmp .L009grand_ssse3 |
+ jmp .L010grand_ssse3 |
.align 16 |
-.L009grand_ssse3: |
+.L010grand_ssse3: |
movdqu (%edi),%xmm0 |
movdqu 16(%edi),%xmm1 |
movdqu 32(%edi),%xmm2 |
@@ -3209,9 +3210,9 @@ sha256_block_data_order: |
paddd %xmm3,%xmm7 |
movdqa %xmm6,64(%esp) |
movdqa %xmm7,80(%esp) |
- jmp .L010ssse3_00_47 |
+ jmp .L011ssse3_00_47 |
.align 16 |
-.L010ssse3_00_47: |
+.L011ssse3_00_47: |
addl $64,%ebp |
movl %edx,%ecx |
movdqa %xmm1,%xmm4 |
@@ -3854,7 +3855,7 @@ sha256_block_data_order: |
addl %ecx,%eax |
movdqa %xmm6,80(%esp) |
cmpl $66051,64(%ebp) |
- jne .L010ssse3_00_47 |
+ jne .L011ssse3_00_47 |
movl %edx,%ecx |
rorl $14,%edx |
movl 20(%esp),%esi |
@@ -4368,12 +4369,1193 @@ sha256_block_data_order: |
movdqa 64(%ebp),%xmm7 |
subl $192,%ebp |
cmpl 104(%esp),%edi |
- jb .L009grand_ssse3 |
+ jb .L010grand_ssse3 |
movl 108(%esp),%esp |
popl %edi |
popl %esi |
popl %ebx |
popl %ebp |
ret |
+.align 32 |
+.L004AVX: |
+ leal -96(%esp),%esp |
+ vzeroall |
+ movl (%esi),%eax |
+ movl 4(%esi),%ebx |
+ movl 8(%esi),%ecx |
+ movl 12(%esi),%edi |
+ movl %ebx,4(%esp) |
+ xorl %ecx,%ebx |
+ movl %ecx,8(%esp) |
+ movl %edi,12(%esp) |
+ movl 16(%esi),%edx |
+ movl 20(%esi),%edi |
+ movl 24(%esi),%ecx |
+ movl 28(%esi),%esi |
+ movl %edi,20(%esp) |
+ movl 100(%esp),%edi |
+ movl %ecx,24(%esp) |
+ movl %esi,28(%esp) |
+ vmovdqa 256(%ebp),%xmm7 |
+ jmp .L012grand_avx |
+.align 32 |
+.L012grand_avx: |
+ vmovdqu (%edi),%xmm0 |
+ vmovdqu 16(%edi),%xmm1 |
+ vmovdqu 32(%edi),%xmm2 |
+ vmovdqu 48(%edi),%xmm3 |
+ addl $64,%edi |
+ vpshufb %xmm7,%xmm0,%xmm0 |
+ movl %edi,100(%esp) |
+ vpshufb %xmm7,%xmm1,%xmm1 |
+ vpshufb %xmm7,%xmm2,%xmm2 |
+ vpaddd (%ebp),%xmm0,%xmm4 |
+ vpshufb %xmm7,%xmm3,%xmm3 |
+ vpaddd 16(%ebp),%xmm1,%xmm5 |
+ vpaddd 32(%ebp),%xmm2,%xmm6 |
+ vpaddd 48(%ebp),%xmm3,%xmm7 |
+ vmovdqa %xmm4,32(%esp) |
+ vmovdqa %xmm5,48(%esp) |
+ vmovdqa %xmm6,64(%esp) |
+ vmovdqa %xmm7,80(%esp) |
+ jmp .L013avx_00_47 |
+.align 16 |
+.L013avx_00_47: |
+ addl $64,%ebp |
+ vpalignr $4,%xmm0,%xmm1,%xmm4 |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 20(%esp),%esi |
+ vpalignr $4,%xmm2,%xmm3,%xmm7 |
+ xorl %ecx,%edx |
+ movl 24(%esp),%edi |
+ xorl %edi,%esi |
+ vpsrld $7,%xmm4,%xmm6 |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,16(%esp) |
+ vpaddd %xmm7,%xmm0,%xmm0 |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ vpsrld $3,%xmm4,%xmm7 |
+ movl %eax,%ecx |
+ addl %edi,%edx |
+ movl 4(%esp),%edi |
+ vpslld $14,%xmm4,%xmm5 |
+ movl %eax,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %eax,(%esp) |
+ vpxor %xmm6,%xmm7,%xmm4 |
+ xorl %eax,%ecx |
+ xorl %edi,%eax |
+ addl 28(%esp),%edx |
+ vpshufd $250,%xmm3,%xmm7 |
+ shrdl $11,%ecx,%ecx |
+ andl %eax,%ebx |
+ xorl %esi,%ecx |
+ vpsrld $11,%xmm6,%xmm6 |
+ addl 32(%esp),%edx |
+ xorl %edi,%ebx |
+ shrdl $2,%ecx,%ecx |
+ vpxor %xmm5,%xmm4,%xmm4 |
+ addl %edx,%ebx |
+ addl 12(%esp),%edx |
+ addl %ecx,%ebx |
+ vpslld $11,%xmm5,%xmm5 |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 16(%esp),%esi |
+ vpxor %xmm6,%xmm4,%xmm4 |
+ xorl %ecx,%edx |
+ movl 20(%esp),%edi |
+ xorl %edi,%esi |
+ vpsrld $10,%xmm7,%xmm6 |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,12(%esp) |
+ vpxor %xmm5,%xmm4,%xmm4 |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ vpsrlq $17,%xmm7,%xmm5 |
+ movl %ebx,%ecx |
+ addl %edi,%edx |
+ movl (%esp),%edi |
+ vpaddd %xmm4,%xmm0,%xmm0 |
+ movl %ebx,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %ebx,28(%esp) |
+ vpxor %xmm5,%xmm6,%xmm6 |
+ xorl %ebx,%ecx |
+ xorl %edi,%ebx |
+ addl 24(%esp),%edx |
+ vpsrlq $19,%xmm7,%xmm7 |
+ shrdl $11,%ecx,%ecx |
+ andl %ebx,%eax |
+ xorl %esi,%ecx |
+ vpxor %xmm7,%xmm6,%xmm6 |
+ addl 36(%esp),%edx |
+ xorl %edi,%eax |
+ shrdl $2,%ecx,%ecx |
+ vpshufd $132,%xmm6,%xmm7 |
+ addl %edx,%eax |
+ addl 8(%esp),%edx |
+ addl %ecx,%eax |
+ vpsrldq $8,%xmm7,%xmm7 |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 12(%esp),%esi |
+ vpaddd %xmm7,%xmm0,%xmm0 |
+ xorl %ecx,%edx |
+ movl 16(%esp),%edi |
+ xorl %edi,%esi |
+ vpshufd $80,%xmm0,%xmm7 |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,8(%esp) |
+ vpsrld $10,%xmm7,%xmm6 |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ vpsrlq $17,%xmm7,%xmm5 |
+ movl %eax,%ecx |
+ addl %edi,%edx |
+ movl 28(%esp),%edi |
+ vpxor %xmm5,%xmm6,%xmm6 |
+ movl %eax,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %eax,24(%esp) |
+ vpsrlq $19,%xmm7,%xmm7 |
+ xorl %eax,%ecx |
+ xorl %edi,%eax |
+ addl 20(%esp),%edx |
+ vpxor %xmm7,%xmm6,%xmm6 |
+ shrdl $11,%ecx,%ecx |
+ andl %eax,%ebx |
+ xorl %esi,%ecx |
+ vpshufd $232,%xmm6,%xmm7 |
+ addl 40(%esp),%edx |
+ xorl %edi,%ebx |
+ shrdl $2,%ecx,%ecx |
+ vpslldq $8,%xmm7,%xmm7 |
+ addl %edx,%ebx |
+ addl 4(%esp),%edx |
+ addl %ecx,%ebx |
+ vpaddd %xmm7,%xmm0,%xmm0 |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 8(%esp),%esi |
+ vpaddd (%ebp),%xmm0,%xmm6 |
+ xorl %ecx,%edx |
+ movl 12(%esp),%edi |
+ xorl %edi,%esi |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,4(%esp) |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ movl %ebx,%ecx |
+ addl %edi,%edx |
+ movl 24(%esp),%edi |
+ movl %ebx,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %ebx,20(%esp) |
+ xorl %ebx,%ecx |
+ xorl %edi,%ebx |
+ addl 16(%esp),%edx |
+ shrdl $11,%ecx,%ecx |
+ andl %ebx,%eax |
+ xorl %esi,%ecx |
+ addl 44(%esp),%edx |
+ xorl %edi,%eax |
+ shrdl $2,%ecx,%ecx |
+ addl %edx,%eax |
+ addl (%esp),%edx |
+ addl %ecx,%eax |
+ vmovdqa %xmm6,32(%esp) |
+ vpalignr $4,%xmm1,%xmm2,%xmm4 |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 4(%esp),%esi |
+ vpalignr $4,%xmm3,%xmm0,%xmm7 |
+ xorl %ecx,%edx |
+ movl 8(%esp),%edi |
+ xorl %edi,%esi |
+ vpsrld $7,%xmm4,%xmm6 |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,(%esp) |
+ vpaddd %xmm7,%xmm1,%xmm1 |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ vpsrld $3,%xmm4,%xmm7 |
+ movl %eax,%ecx |
+ addl %edi,%edx |
+ movl 20(%esp),%edi |
+ vpslld $14,%xmm4,%xmm5 |
+ movl %eax,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %eax,16(%esp) |
+ vpxor %xmm6,%xmm7,%xmm4 |
+ xorl %eax,%ecx |
+ xorl %edi,%eax |
+ addl 12(%esp),%edx |
+ vpshufd $250,%xmm0,%xmm7 |
+ shrdl $11,%ecx,%ecx |
+ andl %eax,%ebx |
+ xorl %esi,%ecx |
+ vpsrld $11,%xmm6,%xmm6 |
+ addl 48(%esp),%edx |
+ xorl %edi,%ebx |
+ shrdl $2,%ecx,%ecx |
+ vpxor %xmm5,%xmm4,%xmm4 |
+ addl %edx,%ebx |
+ addl 28(%esp),%edx |
+ addl %ecx,%ebx |
+ vpslld $11,%xmm5,%xmm5 |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl (%esp),%esi |
+ vpxor %xmm6,%xmm4,%xmm4 |
+ xorl %ecx,%edx |
+ movl 4(%esp),%edi |
+ xorl %edi,%esi |
+ vpsrld $10,%xmm7,%xmm6 |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,28(%esp) |
+ vpxor %xmm5,%xmm4,%xmm4 |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ vpsrlq $17,%xmm7,%xmm5 |
+ movl %ebx,%ecx |
+ addl %edi,%edx |
+ movl 16(%esp),%edi |
+ vpaddd %xmm4,%xmm1,%xmm1 |
+ movl %ebx,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %ebx,12(%esp) |
+ vpxor %xmm5,%xmm6,%xmm6 |
+ xorl %ebx,%ecx |
+ xorl %edi,%ebx |
+ addl 8(%esp),%edx |
+ vpsrlq $19,%xmm7,%xmm7 |
+ shrdl $11,%ecx,%ecx |
+ andl %ebx,%eax |
+ xorl %esi,%ecx |
+ vpxor %xmm7,%xmm6,%xmm6 |
+ addl 52(%esp),%edx |
+ xorl %edi,%eax |
+ shrdl $2,%ecx,%ecx |
+ vpshufd $132,%xmm6,%xmm7 |
+ addl %edx,%eax |
+ addl 24(%esp),%edx |
+ addl %ecx,%eax |
+ vpsrldq $8,%xmm7,%xmm7 |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 28(%esp),%esi |
+ vpaddd %xmm7,%xmm1,%xmm1 |
+ xorl %ecx,%edx |
+ movl (%esp),%edi |
+ xorl %edi,%esi |
+ vpshufd $80,%xmm1,%xmm7 |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,24(%esp) |
+ vpsrld $10,%xmm7,%xmm6 |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ vpsrlq $17,%xmm7,%xmm5 |
+ movl %eax,%ecx |
+ addl %edi,%edx |
+ movl 12(%esp),%edi |
+ vpxor %xmm5,%xmm6,%xmm6 |
+ movl %eax,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %eax,8(%esp) |
+ vpsrlq $19,%xmm7,%xmm7 |
+ xorl %eax,%ecx |
+ xorl %edi,%eax |
+ addl 4(%esp),%edx |
+ vpxor %xmm7,%xmm6,%xmm6 |
+ shrdl $11,%ecx,%ecx |
+ andl %eax,%ebx |
+ xorl %esi,%ecx |
+ vpshufd $232,%xmm6,%xmm7 |
+ addl 56(%esp),%edx |
+ xorl %edi,%ebx |
+ shrdl $2,%ecx,%ecx |
+ vpslldq $8,%xmm7,%xmm7 |
+ addl %edx,%ebx |
+ addl 20(%esp),%edx |
+ addl %ecx,%ebx |
+ vpaddd %xmm7,%xmm1,%xmm1 |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 24(%esp),%esi |
+ vpaddd 16(%ebp),%xmm1,%xmm6 |
+ xorl %ecx,%edx |
+ movl 28(%esp),%edi |
+ xorl %edi,%esi |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,20(%esp) |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ movl %ebx,%ecx |
+ addl %edi,%edx |
+ movl 8(%esp),%edi |
+ movl %ebx,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %ebx,4(%esp) |
+ xorl %ebx,%ecx |
+ xorl %edi,%ebx |
+ addl (%esp),%edx |
+ shrdl $11,%ecx,%ecx |
+ andl %ebx,%eax |
+ xorl %esi,%ecx |
+ addl 60(%esp),%edx |
+ xorl %edi,%eax |
+ shrdl $2,%ecx,%ecx |
+ addl %edx,%eax |
+ addl 16(%esp),%edx |
+ addl %ecx,%eax |
+ vmovdqa %xmm6,48(%esp) |
+ vpalignr $4,%xmm2,%xmm3,%xmm4 |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 20(%esp),%esi |
+ vpalignr $4,%xmm0,%xmm1,%xmm7 |
+ xorl %ecx,%edx |
+ movl 24(%esp),%edi |
+ xorl %edi,%esi |
+ vpsrld $7,%xmm4,%xmm6 |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,16(%esp) |
+ vpaddd %xmm7,%xmm2,%xmm2 |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ vpsrld $3,%xmm4,%xmm7 |
+ movl %eax,%ecx |
+ addl %edi,%edx |
+ movl 4(%esp),%edi |
+ vpslld $14,%xmm4,%xmm5 |
+ movl %eax,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %eax,(%esp) |
+ vpxor %xmm6,%xmm7,%xmm4 |
+ xorl %eax,%ecx |
+ xorl %edi,%eax |
+ addl 28(%esp),%edx |
+ vpshufd $250,%xmm1,%xmm7 |
+ shrdl $11,%ecx,%ecx |
+ andl %eax,%ebx |
+ xorl %esi,%ecx |
+ vpsrld $11,%xmm6,%xmm6 |
+ addl 64(%esp),%edx |
+ xorl %edi,%ebx |
+ shrdl $2,%ecx,%ecx |
+ vpxor %xmm5,%xmm4,%xmm4 |
+ addl %edx,%ebx |
+ addl 12(%esp),%edx |
+ addl %ecx,%ebx |
+ vpslld $11,%xmm5,%xmm5 |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 16(%esp),%esi |
+ vpxor %xmm6,%xmm4,%xmm4 |
+ xorl %ecx,%edx |
+ movl 20(%esp),%edi |
+ xorl %edi,%esi |
+ vpsrld $10,%xmm7,%xmm6 |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,12(%esp) |
+ vpxor %xmm5,%xmm4,%xmm4 |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ vpsrlq $17,%xmm7,%xmm5 |
+ movl %ebx,%ecx |
+ addl %edi,%edx |
+ movl (%esp),%edi |
+ vpaddd %xmm4,%xmm2,%xmm2 |
+ movl %ebx,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %ebx,28(%esp) |
+ vpxor %xmm5,%xmm6,%xmm6 |
+ xorl %ebx,%ecx |
+ xorl %edi,%ebx |
+ addl 24(%esp),%edx |
+ vpsrlq $19,%xmm7,%xmm7 |
+ shrdl $11,%ecx,%ecx |
+ andl %ebx,%eax |
+ xorl %esi,%ecx |
+ vpxor %xmm7,%xmm6,%xmm6 |
+ addl 68(%esp),%edx |
+ xorl %edi,%eax |
+ shrdl $2,%ecx,%ecx |
+ vpshufd $132,%xmm6,%xmm7 |
+ addl %edx,%eax |
+ addl 8(%esp),%edx |
+ addl %ecx,%eax |
+ vpsrldq $8,%xmm7,%xmm7 |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 12(%esp),%esi |
+ vpaddd %xmm7,%xmm2,%xmm2 |
+ xorl %ecx,%edx |
+ movl 16(%esp),%edi |
+ xorl %edi,%esi |
+ vpshufd $80,%xmm2,%xmm7 |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,8(%esp) |
+ vpsrld $10,%xmm7,%xmm6 |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ vpsrlq $17,%xmm7,%xmm5 |
+ movl %eax,%ecx |
+ addl %edi,%edx |
+ movl 28(%esp),%edi |
+ vpxor %xmm5,%xmm6,%xmm6 |
+ movl %eax,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %eax,24(%esp) |
+ vpsrlq $19,%xmm7,%xmm7 |
+ xorl %eax,%ecx |
+ xorl %edi,%eax |
+ addl 20(%esp),%edx |
+ vpxor %xmm7,%xmm6,%xmm6 |
+ shrdl $11,%ecx,%ecx |
+ andl %eax,%ebx |
+ xorl %esi,%ecx |
+ vpshufd $232,%xmm6,%xmm7 |
+ addl 72(%esp),%edx |
+ xorl %edi,%ebx |
+ shrdl $2,%ecx,%ecx |
+ vpslldq $8,%xmm7,%xmm7 |
+ addl %edx,%ebx |
+ addl 4(%esp),%edx |
+ addl %ecx,%ebx |
+ vpaddd %xmm7,%xmm2,%xmm2 |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 8(%esp),%esi |
+ vpaddd 32(%ebp),%xmm2,%xmm6 |
+ xorl %ecx,%edx |
+ movl 12(%esp),%edi |
+ xorl %edi,%esi |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,4(%esp) |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ movl %ebx,%ecx |
+ addl %edi,%edx |
+ movl 24(%esp),%edi |
+ movl %ebx,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %ebx,20(%esp) |
+ xorl %ebx,%ecx |
+ xorl %edi,%ebx |
+ addl 16(%esp),%edx |
+ shrdl $11,%ecx,%ecx |
+ andl %ebx,%eax |
+ xorl %esi,%ecx |
+ addl 76(%esp),%edx |
+ xorl %edi,%eax |
+ shrdl $2,%ecx,%ecx |
+ addl %edx,%eax |
+ addl (%esp),%edx |
+ addl %ecx,%eax |
+ vmovdqa %xmm6,64(%esp) |
+ vpalignr $4,%xmm3,%xmm0,%xmm4 |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 4(%esp),%esi |
+ vpalignr $4,%xmm1,%xmm2,%xmm7 |
+ xorl %ecx,%edx |
+ movl 8(%esp),%edi |
+ xorl %edi,%esi |
+ vpsrld $7,%xmm4,%xmm6 |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,(%esp) |
+ vpaddd %xmm7,%xmm3,%xmm3 |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ vpsrld $3,%xmm4,%xmm7 |
+ movl %eax,%ecx |
+ addl %edi,%edx |
+ movl 20(%esp),%edi |
+ vpslld $14,%xmm4,%xmm5 |
+ movl %eax,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %eax,16(%esp) |
+ vpxor %xmm6,%xmm7,%xmm4 |
+ xorl %eax,%ecx |
+ xorl %edi,%eax |
+ addl 12(%esp),%edx |
+ vpshufd $250,%xmm2,%xmm7 |
+ shrdl $11,%ecx,%ecx |
+ andl %eax,%ebx |
+ xorl %esi,%ecx |
+ vpsrld $11,%xmm6,%xmm6 |
+ addl 80(%esp),%edx |
+ xorl %edi,%ebx |
+ shrdl $2,%ecx,%ecx |
+ vpxor %xmm5,%xmm4,%xmm4 |
+ addl %edx,%ebx |
+ addl 28(%esp),%edx |
+ addl %ecx,%ebx |
+ vpslld $11,%xmm5,%xmm5 |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl (%esp),%esi |
+ vpxor %xmm6,%xmm4,%xmm4 |
+ xorl %ecx,%edx |
+ movl 4(%esp),%edi |
+ xorl %edi,%esi |
+ vpsrld $10,%xmm7,%xmm6 |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,28(%esp) |
+ vpxor %xmm5,%xmm4,%xmm4 |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ vpsrlq $17,%xmm7,%xmm5 |
+ movl %ebx,%ecx |
+ addl %edi,%edx |
+ movl 16(%esp),%edi |
+ vpaddd %xmm4,%xmm3,%xmm3 |
+ movl %ebx,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %ebx,12(%esp) |
+ vpxor %xmm5,%xmm6,%xmm6 |
+ xorl %ebx,%ecx |
+ xorl %edi,%ebx |
+ addl 8(%esp),%edx |
+ vpsrlq $19,%xmm7,%xmm7 |
+ shrdl $11,%ecx,%ecx |
+ andl %ebx,%eax |
+ xorl %esi,%ecx |
+ vpxor %xmm7,%xmm6,%xmm6 |
+ addl 84(%esp),%edx |
+ xorl %edi,%eax |
+ shrdl $2,%ecx,%ecx |
+ vpshufd $132,%xmm6,%xmm7 |
+ addl %edx,%eax |
+ addl 24(%esp),%edx |
+ addl %ecx,%eax |
+ vpsrldq $8,%xmm7,%xmm7 |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 28(%esp),%esi |
+ vpaddd %xmm7,%xmm3,%xmm3 |
+ xorl %ecx,%edx |
+ movl (%esp),%edi |
+ xorl %edi,%esi |
+ vpshufd $80,%xmm3,%xmm7 |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,24(%esp) |
+ vpsrld $10,%xmm7,%xmm6 |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ vpsrlq $17,%xmm7,%xmm5 |
+ movl %eax,%ecx |
+ addl %edi,%edx |
+ movl 12(%esp),%edi |
+ vpxor %xmm5,%xmm6,%xmm6 |
+ movl %eax,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %eax,8(%esp) |
+ vpsrlq $19,%xmm7,%xmm7 |
+ xorl %eax,%ecx |
+ xorl %edi,%eax |
+ addl 4(%esp),%edx |
+ vpxor %xmm7,%xmm6,%xmm6 |
+ shrdl $11,%ecx,%ecx |
+ andl %eax,%ebx |
+ xorl %esi,%ecx |
+ vpshufd $232,%xmm6,%xmm7 |
+ addl 88(%esp),%edx |
+ xorl %edi,%ebx |
+ shrdl $2,%ecx,%ecx |
+ vpslldq $8,%xmm7,%xmm7 |
+ addl %edx,%ebx |
+ addl 20(%esp),%edx |
+ addl %ecx,%ebx |
+ vpaddd %xmm7,%xmm3,%xmm3 |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 24(%esp),%esi |
+ vpaddd 48(%ebp),%xmm3,%xmm6 |
+ xorl %ecx,%edx |
+ movl 28(%esp),%edi |
+ xorl %edi,%esi |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,20(%esp) |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ movl %ebx,%ecx |
+ addl %edi,%edx |
+ movl 8(%esp),%edi |
+ movl %ebx,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %ebx,4(%esp) |
+ xorl %ebx,%ecx |
+ xorl %edi,%ebx |
+ addl (%esp),%edx |
+ shrdl $11,%ecx,%ecx |
+ andl %ebx,%eax |
+ xorl %esi,%ecx |
+ addl 92(%esp),%edx |
+ xorl %edi,%eax |
+ shrdl $2,%ecx,%ecx |
+ addl %edx,%eax |
+ addl 16(%esp),%edx |
+ addl %ecx,%eax |
+ vmovdqa %xmm6,80(%esp) |
+ cmpl $66051,64(%ebp) |
+ jne .L013avx_00_47 |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 20(%esp),%esi |
+ xorl %ecx,%edx |
+ movl 24(%esp),%edi |
+ xorl %edi,%esi |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,16(%esp) |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ movl %eax,%ecx |
+ addl %edi,%edx |
+ movl 4(%esp),%edi |
+ movl %eax,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %eax,(%esp) |
+ xorl %eax,%ecx |
+ xorl %edi,%eax |
+ addl 28(%esp),%edx |
+ shrdl $11,%ecx,%ecx |
+ andl %eax,%ebx |
+ xorl %esi,%ecx |
+ addl 32(%esp),%edx |
+ xorl %edi,%ebx |
+ shrdl $2,%ecx,%ecx |
+ addl %edx,%ebx |
+ addl 12(%esp),%edx |
+ addl %ecx,%ebx |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 16(%esp),%esi |
+ xorl %ecx,%edx |
+ movl 20(%esp),%edi |
+ xorl %edi,%esi |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,12(%esp) |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ movl %ebx,%ecx |
+ addl %edi,%edx |
+ movl (%esp),%edi |
+ movl %ebx,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %ebx,28(%esp) |
+ xorl %ebx,%ecx |
+ xorl %edi,%ebx |
+ addl 24(%esp),%edx |
+ shrdl $11,%ecx,%ecx |
+ andl %ebx,%eax |
+ xorl %esi,%ecx |
+ addl 36(%esp),%edx |
+ xorl %edi,%eax |
+ shrdl $2,%ecx,%ecx |
+ addl %edx,%eax |
+ addl 8(%esp),%edx |
+ addl %ecx,%eax |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 12(%esp),%esi |
+ xorl %ecx,%edx |
+ movl 16(%esp),%edi |
+ xorl %edi,%esi |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,8(%esp) |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ movl %eax,%ecx |
+ addl %edi,%edx |
+ movl 28(%esp),%edi |
+ movl %eax,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %eax,24(%esp) |
+ xorl %eax,%ecx |
+ xorl %edi,%eax |
+ addl 20(%esp),%edx |
+ shrdl $11,%ecx,%ecx |
+ andl %eax,%ebx |
+ xorl %esi,%ecx |
+ addl 40(%esp),%edx |
+ xorl %edi,%ebx |
+ shrdl $2,%ecx,%ecx |
+ addl %edx,%ebx |
+ addl 4(%esp),%edx |
+ addl %ecx,%ebx |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 8(%esp),%esi |
+ xorl %ecx,%edx |
+ movl 12(%esp),%edi |
+ xorl %edi,%esi |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,4(%esp) |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ movl %ebx,%ecx |
+ addl %edi,%edx |
+ movl 24(%esp),%edi |
+ movl %ebx,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %ebx,20(%esp) |
+ xorl %ebx,%ecx |
+ xorl %edi,%ebx |
+ addl 16(%esp),%edx |
+ shrdl $11,%ecx,%ecx |
+ andl %ebx,%eax |
+ xorl %esi,%ecx |
+ addl 44(%esp),%edx |
+ xorl %edi,%eax |
+ shrdl $2,%ecx,%ecx |
+ addl %edx,%eax |
+ addl (%esp),%edx |
+ addl %ecx,%eax |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 4(%esp),%esi |
+ xorl %ecx,%edx |
+ movl 8(%esp),%edi |
+ xorl %edi,%esi |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,(%esp) |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ movl %eax,%ecx |
+ addl %edi,%edx |
+ movl 20(%esp),%edi |
+ movl %eax,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %eax,16(%esp) |
+ xorl %eax,%ecx |
+ xorl %edi,%eax |
+ addl 12(%esp),%edx |
+ shrdl $11,%ecx,%ecx |
+ andl %eax,%ebx |
+ xorl %esi,%ecx |
+ addl 48(%esp),%edx |
+ xorl %edi,%ebx |
+ shrdl $2,%ecx,%ecx |
+ addl %edx,%ebx |
+ addl 28(%esp),%edx |
+ addl %ecx,%ebx |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl (%esp),%esi |
+ xorl %ecx,%edx |
+ movl 4(%esp),%edi |
+ xorl %edi,%esi |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,28(%esp) |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ movl %ebx,%ecx |
+ addl %edi,%edx |
+ movl 16(%esp),%edi |
+ movl %ebx,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %ebx,12(%esp) |
+ xorl %ebx,%ecx |
+ xorl %edi,%ebx |
+ addl 8(%esp),%edx |
+ shrdl $11,%ecx,%ecx |
+ andl %ebx,%eax |
+ xorl %esi,%ecx |
+ addl 52(%esp),%edx |
+ xorl %edi,%eax |
+ shrdl $2,%ecx,%ecx |
+ addl %edx,%eax |
+ addl 24(%esp),%edx |
+ addl %ecx,%eax |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 28(%esp),%esi |
+ xorl %ecx,%edx |
+ movl (%esp),%edi |
+ xorl %edi,%esi |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,24(%esp) |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ movl %eax,%ecx |
+ addl %edi,%edx |
+ movl 12(%esp),%edi |
+ movl %eax,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %eax,8(%esp) |
+ xorl %eax,%ecx |
+ xorl %edi,%eax |
+ addl 4(%esp),%edx |
+ shrdl $11,%ecx,%ecx |
+ andl %eax,%ebx |
+ xorl %esi,%ecx |
+ addl 56(%esp),%edx |
+ xorl %edi,%ebx |
+ shrdl $2,%ecx,%ecx |
+ addl %edx,%ebx |
+ addl 20(%esp),%edx |
+ addl %ecx,%ebx |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 24(%esp),%esi |
+ xorl %ecx,%edx |
+ movl 28(%esp),%edi |
+ xorl %edi,%esi |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,20(%esp) |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ movl %ebx,%ecx |
+ addl %edi,%edx |
+ movl 8(%esp),%edi |
+ movl %ebx,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %ebx,4(%esp) |
+ xorl %ebx,%ecx |
+ xorl %edi,%ebx |
+ addl (%esp),%edx |
+ shrdl $11,%ecx,%ecx |
+ andl %ebx,%eax |
+ xorl %esi,%ecx |
+ addl 60(%esp),%edx |
+ xorl %edi,%eax |
+ shrdl $2,%ecx,%ecx |
+ addl %edx,%eax |
+ addl 16(%esp),%edx |
+ addl %ecx,%eax |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 20(%esp),%esi |
+ xorl %ecx,%edx |
+ movl 24(%esp),%edi |
+ xorl %edi,%esi |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,16(%esp) |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ movl %eax,%ecx |
+ addl %edi,%edx |
+ movl 4(%esp),%edi |
+ movl %eax,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %eax,(%esp) |
+ xorl %eax,%ecx |
+ xorl %edi,%eax |
+ addl 28(%esp),%edx |
+ shrdl $11,%ecx,%ecx |
+ andl %eax,%ebx |
+ xorl %esi,%ecx |
+ addl 64(%esp),%edx |
+ xorl %edi,%ebx |
+ shrdl $2,%ecx,%ecx |
+ addl %edx,%ebx |
+ addl 12(%esp),%edx |
+ addl %ecx,%ebx |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 16(%esp),%esi |
+ xorl %ecx,%edx |
+ movl 20(%esp),%edi |
+ xorl %edi,%esi |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,12(%esp) |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ movl %ebx,%ecx |
+ addl %edi,%edx |
+ movl (%esp),%edi |
+ movl %ebx,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %ebx,28(%esp) |
+ xorl %ebx,%ecx |
+ xorl %edi,%ebx |
+ addl 24(%esp),%edx |
+ shrdl $11,%ecx,%ecx |
+ andl %ebx,%eax |
+ xorl %esi,%ecx |
+ addl 68(%esp),%edx |
+ xorl %edi,%eax |
+ shrdl $2,%ecx,%ecx |
+ addl %edx,%eax |
+ addl 8(%esp),%edx |
+ addl %ecx,%eax |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 12(%esp),%esi |
+ xorl %ecx,%edx |
+ movl 16(%esp),%edi |
+ xorl %edi,%esi |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,8(%esp) |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ movl %eax,%ecx |
+ addl %edi,%edx |
+ movl 28(%esp),%edi |
+ movl %eax,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %eax,24(%esp) |
+ xorl %eax,%ecx |
+ xorl %edi,%eax |
+ addl 20(%esp),%edx |
+ shrdl $11,%ecx,%ecx |
+ andl %eax,%ebx |
+ xorl %esi,%ecx |
+ addl 72(%esp),%edx |
+ xorl %edi,%ebx |
+ shrdl $2,%ecx,%ecx |
+ addl %edx,%ebx |
+ addl 4(%esp),%edx |
+ addl %ecx,%ebx |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 8(%esp),%esi |
+ xorl %ecx,%edx |
+ movl 12(%esp),%edi |
+ xorl %edi,%esi |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,4(%esp) |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ movl %ebx,%ecx |
+ addl %edi,%edx |
+ movl 24(%esp),%edi |
+ movl %ebx,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %ebx,20(%esp) |
+ xorl %ebx,%ecx |
+ xorl %edi,%ebx |
+ addl 16(%esp),%edx |
+ shrdl $11,%ecx,%ecx |
+ andl %ebx,%eax |
+ xorl %esi,%ecx |
+ addl 76(%esp),%edx |
+ xorl %edi,%eax |
+ shrdl $2,%ecx,%ecx |
+ addl %edx,%eax |
+ addl (%esp),%edx |
+ addl %ecx,%eax |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 4(%esp),%esi |
+ xorl %ecx,%edx |
+ movl 8(%esp),%edi |
+ xorl %edi,%esi |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,(%esp) |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ movl %eax,%ecx |
+ addl %edi,%edx |
+ movl 20(%esp),%edi |
+ movl %eax,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %eax,16(%esp) |
+ xorl %eax,%ecx |
+ xorl %edi,%eax |
+ addl 12(%esp),%edx |
+ shrdl $11,%ecx,%ecx |
+ andl %eax,%ebx |
+ xorl %esi,%ecx |
+ addl 80(%esp),%edx |
+ xorl %edi,%ebx |
+ shrdl $2,%ecx,%ecx |
+ addl %edx,%ebx |
+ addl 28(%esp),%edx |
+ addl %ecx,%ebx |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl (%esp),%esi |
+ xorl %ecx,%edx |
+ movl 4(%esp),%edi |
+ xorl %edi,%esi |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,28(%esp) |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ movl %ebx,%ecx |
+ addl %edi,%edx |
+ movl 16(%esp),%edi |
+ movl %ebx,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %ebx,12(%esp) |
+ xorl %ebx,%ecx |
+ xorl %edi,%ebx |
+ addl 8(%esp),%edx |
+ shrdl $11,%ecx,%ecx |
+ andl %ebx,%eax |
+ xorl %esi,%ecx |
+ addl 84(%esp),%edx |
+ xorl %edi,%eax |
+ shrdl $2,%ecx,%ecx |
+ addl %edx,%eax |
+ addl 24(%esp),%edx |
+ addl %ecx,%eax |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 28(%esp),%esi |
+ xorl %ecx,%edx |
+ movl (%esp),%edi |
+ xorl %edi,%esi |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,24(%esp) |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ movl %eax,%ecx |
+ addl %edi,%edx |
+ movl 12(%esp),%edi |
+ movl %eax,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %eax,8(%esp) |
+ xorl %eax,%ecx |
+ xorl %edi,%eax |
+ addl 4(%esp),%edx |
+ shrdl $11,%ecx,%ecx |
+ andl %eax,%ebx |
+ xorl %esi,%ecx |
+ addl 88(%esp),%edx |
+ xorl %edi,%ebx |
+ shrdl $2,%ecx,%ecx |
+ addl %edx,%ebx |
+ addl 20(%esp),%edx |
+ addl %ecx,%ebx |
+ movl %edx,%ecx |
+ shrdl $14,%edx,%edx |
+ movl 24(%esp),%esi |
+ xorl %ecx,%edx |
+ movl 28(%esp),%edi |
+ xorl %edi,%esi |
+ shrdl $5,%edx,%edx |
+ andl %ecx,%esi |
+ movl %ecx,20(%esp) |
+ xorl %ecx,%edx |
+ xorl %esi,%edi |
+ shrdl $6,%edx,%edx |
+ movl %ebx,%ecx |
+ addl %edi,%edx |
+ movl 8(%esp),%edi |
+ movl %ebx,%esi |
+ shrdl $9,%ecx,%ecx |
+ movl %ebx,4(%esp) |
+ xorl %ebx,%ecx |
+ xorl %edi,%ebx |
+ addl (%esp),%edx |
+ shrdl $11,%ecx,%ecx |
+ andl %ebx,%eax |
+ xorl %esi,%ecx |
+ addl 92(%esp),%edx |
+ xorl %edi,%eax |
+ shrdl $2,%ecx,%ecx |
+ addl %edx,%eax |
+ addl 16(%esp),%edx |
+ addl %ecx,%eax |
+ movl 96(%esp),%esi |
+ xorl %edi,%ebx |
+ movl 12(%esp),%ecx |
+ addl (%esi),%eax |
+ addl 4(%esi),%ebx |
+ addl 8(%esi),%edi |
+ addl 12(%esi),%ecx |
+ movl %eax,(%esi) |
+ movl %ebx,4(%esi) |
+ movl %edi,8(%esi) |
+ movl %ecx,12(%esi) |
+ movl %ebx,4(%esp) |
+ xorl %edi,%ebx |
+ movl %edi,8(%esp) |
+ movl %ecx,12(%esp) |
+ movl 20(%esp),%edi |
+ movl 24(%esp),%ecx |
+ addl 16(%esi),%edx |
+ addl 20(%esi),%edi |
+ addl 24(%esi),%ecx |
+ movl %edx,16(%esi) |
+ movl %edi,20(%esi) |
+ movl %edi,20(%esp) |
+ movl 28(%esp),%edi |
+ movl %ecx,24(%esi) |
+ addl 28(%esi),%edi |
+ movl %ecx,24(%esp) |
+ movl %edi,28(%esi) |
+ movl %edi,28(%esp) |
+ movl 100(%esp),%edi |
+ vmovdqa 64(%ebp),%xmm7 |
+ subl $192,%ebp |
+ cmpl 104(%esp),%edi |
+ jb .L012grand_avx |
+ movl 108(%esp),%esp |
+ vzeroall |
+ popl %edi |
+ popl %esi |
+ popl %ebx |
+ popl %ebp |
+ ret |
.size sha256_block_data_order,.-.L_sha256_block_data_order_begin |
#endif |