| Index: third_party/boringssl/mac-x86/crypto/sha/sha256-586.S
|
| diff --git a/third_party/boringssl/mac-x86/crypto/sha/sha256-586.S b/third_party/boringssl/mac-x86/crypto/sha/sha256-586.S
|
| index 841854f7a9772d16ac1fcbec5c391930be066ef0..f0ba612fabb90014d18c5eb6a1d5cf12d2c9d952 100644
|
| --- a/third_party/boringssl/mac-x86/crypto/sha/sha256-586.S
|
| +++ b/third_party/boringssl/mac-x86/crypto/sha/sha256-586.S
|
| @@ -36,10 +36,11 @@ L000pic_point:
|
| jz L003no_xmm
|
| andl $1073741824,%ecx
|
| andl $268435968,%ebx
|
| + testl $536870912,%edx
|
| + jnz L004shaext
|
| orl %ebx,%ecx
|
| andl $1342177280,%ecx
|
| cmpl $1342177280,%ecx
|
| - je L004AVX
|
| testl $512,%ebx
|
| jnz L005SSSE3
|
| L003no_xmm:
|
| @@ -3165,6 +3166,204 @@ L009grand_loop:
|
| popl %ebp
|
| ret
|
| .align 5,0x90
|
| +L004shaext:
|
| + subl $32,%esp
|
| + movdqu (%esi),%xmm1
|
| + leal 128(%ebp),%ebp
|
| + movdqu 16(%esi),%xmm2
|
| + movdqa 128(%ebp),%xmm7
|
| + pshufd $27,%xmm1,%xmm0
|
| + pshufd $177,%xmm1,%xmm1
|
| + pshufd $27,%xmm2,%xmm2
|
| +.byte 102,15,58,15,202,8
|
| + punpcklqdq %xmm0,%xmm2
|
| + jmp L010loop_shaext
|
| +.align 4,0x90
|
| +L010loop_shaext:
|
| + movdqu (%edi),%xmm3
|
| + movdqu 16(%edi),%xmm4
|
| + movdqu 32(%edi),%xmm5
|
| +.byte 102,15,56,0,223
|
| + movdqu 48(%edi),%xmm6
|
| + movdqa %xmm2,16(%esp)
|
| + movdqa -128(%ebp),%xmm0
|
| + paddd %xmm3,%xmm0
|
| +.byte 102,15,56,0,231
|
| +.byte 15,56,203,209
|
| + pshufd $14,%xmm0,%xmm0
|
| + nop
|
| + movdqa %xmm1,(%esp)
|
| +.byte 15,56,203,202
|
| + movdqa -112(%ebp),%xmm0
|
| + paddd %xmm4,%xmm0
|
| +.byte 102,15,56,0,239
|
| +.byte 15,56,203,209
|
| + pshufd $14,%xmm0,%xmm0
|
| + leal 64(%edi),%edi
|
| +.byte 15,56,204,220
|
| +.byte 15,56,203,202
|
| + movdqa -96(%ebp),%xmm0
|
| + paddd %xmm5,%xmm0
|
| +.byte 102,15,56,0,247
|
| +.byte 15,56,203,209
|
| + pshufd $14,%xmm0,%xmm0
|
| + movdqa %xmm6,%xmm7
|
| +.byte 102,15,58,15,253,4
|
| + nop
|
| + paddd %xmm7,%xmm3
|
| +.byte 15,56,204,229
|
| +.byte 15,56,203,202
|
| + movdqa -80(%ebp),%xmm0
|
| + paddd %xmm6,%xmm0
|
| +.byte 15,56,205,222
|
| +.byte 15,56,203,209
|
| + pshufd $14,%xmm0,%xmm0
|
| + movdqa %xmm3,%xmm7
|
| +.byte 102,15,58,15,254,4
|
| + nop
|
| + paddd %xmm7,%xmm4
|
| +.byte 15,56,204,238
|
| +.byte 15,56,203,202
|
| + movdqa -64(%ebp),%xmm0
|
| + paddd %xmm3,%xmm0
|
| +.byte 15,56,205,227
|
| +.byte 15,56,203,209
|
| + pshufd $14,%xmm0,%xmm0
|
| + movdqa %xmm4,%xmm7
|
| +.byte 102,15,58,15,251,4
|
| + nop
|
| + paddd %xmm7,%xmm5
|
| +.byte 15,56,204,243
|
| +.byte 15,56,203,202
|
| + movdqa -48(%ebp),%xmm0
|
| + paddd %xmm4,%xmm0
|
| +.byte 15,56,205,236
|
| +.byte 15,56,203,209
|
| + pshufd $14,%xmm0,%xmm0
|
| + movdqa %xmm5,%xmm7
|
| +.byte 102,15,58,15,252,4
|
| + nop
|
| + paddd %xmm7,%xmm6
|
| +.byte 15,56,204,220
|
| +.byte 15,56,203,202
|
| + movdqa -32(%ebp),%xmm0
|
| + paddd %xmm5,%xmm0
|
| +.byte 15,56,205,245
|
| +.byte 15,56,203,209
|
| + pshufd $14,%xmm0,%xmm0
|
| + movdqa %xmm6,%xmm7
|
| +.byte 102,15,58,15,253,4
|
| + nop
|
| + paddd %xmm7,%xmm3
|
| +.byte 15,56,204,229
|
| +.byte 15,56,203,202
|
| + movdqa -16(%ebp),%xmm0
|
| + paddd %xmm6,%xmm0
|
| +.byte 15,56,205,222
|
| +.byte 15,56,203,209
|
| + pshufd $14,%xmm0,%xmm0
|
| + movdqa %xmm3,%xmm7
|
| +.byte 102,15,58,15,254,4
|
| + nop
|
| + paddd %xmm7,%xmm4
|
| +.byte 15,56,204,238
|
| +.byte 15,56,203,202
|
| + movdqa (%ebp),%xmm0
|
| + paddd %xmm3,%xmm0
|
| +.byte 15,56,205,227
|
| +.byte 15,56,203,209
|
| + pshufd $14,%xmm0,%xmm0
|
| + movdqa %xmm4,%xmm7
|
| +.byte 102,15,58,15,251,4
|
| + nop
|
| + paddd %xmm7,%xmm5
|
| +.byte 15,56,204,243
|
| +.byte 15,56,203,202
|
| + movdqa 16(%ebp),%xmm0
|
| + paddd %xmm4,%xmm0
|
| +.byte 15,56,205,236
|
| +.byte 15,56,203,209
|
| + pshufd $14,%xmm0,%xmm0
|
| + movdqa %xmm5,%xmm7
|
| +.byte 102,15,58,15,252,4
|
| + nop
|
| + paddd %xmm7,%xmm6
|
| +.byte 15,56,204,220
|
| +.byte 15,56,203,202
|
| + movdqa 32(%ebp),%xmm0
|
| + paddd %xmm5,%xmm0
|
| +.byte 15,56,205,245
|
| +.byte 15,56,203,209
|
| + pshufd $14,%xmm0,%xmm0
|
| + movdqa %xmm6,%xmm7
|
| +.byte 102,15,58,15,253,4
|
| + nop
|
| + paddd %xmm7,%xmm3
|
| +.byte 15,56,204,229
|
| +.byte 15,56,203,202
|
| + movdqa 48(%ebp),%xmm0
|
| + paddd %xmm6,%xmm0
|
| +.byte 15,56,205,222
|
| +.byte 15,56,203,209
|
| + pshufd $14,%xmm0,%xmm0
|
| + movdqa %xmm3,%xmm7
|
| +.byte 102,15,58,15,254,4
|
| + nop
|
| + paddd %xmm7,%xmm4
|
| +.byte 15,56,204,238
|
| +.byte 15,56,203,202
|
| + movdqa 64(%ebp),%xmm0
|
| + paddd %xmm3,%xmm0
|
| +.byte 15,56,205,227
|
| +.byte 15,56,203,209
|
| + pshufd $14,%xmm0,%xmm0
|
| + movdqa %xmm4,%xmm7
|
| +.byte 102,15,58,15,251,4
|
| + nop
|
| + paddd %xmm7,%xmm5
|
| +.byte 15,56,204,243
|
| +.byte 15,56,203,202
|
| + movdqa 80(%ebp),%xmm0
|
| + paddd %xmm4,%xmm0
|
| +.byte 15,56,205,236
|
| +.byte 15,56,203,209
|
| + pshufd $14,%xmm0,%xmm0
|
| + movdqa %xmm5,%xmm7
|
| +.byte 102,15,58,15,252,4
|
| +.byte 15,56,203,202
|
| + paddd %xmm7,%xmm6
|
| + movdqa 96(%ebp),%xmm0
|
| + paddd %xmm5,%xmm0
|
| +.byte 15,56,203,209
|
| + pshufd $14,%xmm0,%xmm0
|
| +.byte 15,56,205,245
|
| + movdqa 128(%ebp),%xmm7
|
| +.byte 15,56,203,202
|
| + movdqa 112(%ebp),%xmm0
|
| + paddd %xmm6,%xmm0
|
| + nop
|
| +.byte 15,56,203,209
|
| + pshufd $14,%xmm0,%xmm0
|
| + cmpl %edi,%eax
|
| + nop
|
| +.byte 15,56,203,202
|
| + paddd 16(%esp),%xmm2
|
| + paddd (%esp),%xmm1
|
| + jnz L010loop_shaext
|
| + pshufd $177,%xmm2,%xmm2
|
| + pshufd $27,%xmm1,%xmm7
|
| + pshufd $177,%xmm1,%xmm1
|
| + punpckhqdq %xmm2,%xmm1
|
| +.byte 102,15,58,15,215,8
|
| + movl 44(%esp),%esp
|
| + movdqu %xmm1,(%esi)
|
| + movdqu %xmm2,16(%esi)
|
| + popl %edi
|
| + popl %esi
|
| + popl %ebx
|
| + popl %ebp
|
| + ret
|
| +.align 5,0x90
|
| L005SSSE3:
|
| leal -96(%esp),%esp
|
| movl (%esi),%eax
|
| @@ -3184,9 +3383,9 @@ L005SSSE3:
|
| movl %ecx,24(%esp)
|
| movl %esi,28(%esp)
|
| movdqa 256(%ebp),%xmm7
|
| - jmp L010grand_ssse3
|
| + jmp L011grand_ssse3
|
| .align 4,0x90
|
| -L010grand_ssse3:
|
| +L011grand_ssse3:
|
| movdqu (%edi),%xmm0
|
| movdqu 16(%edi),%xmm1
|
| movdqu 32(%edi),%xmm2
|
| @@ -3209,9 +3408,9 @@ L010grand_ssse3:
|
| paddd %xmm3,%xmm7
|
| movdqa %xmm6,64(%esp)
|
| movdqa %xmm7,80(%esp)
|
| - jmp L011ssse3_00_47
|
| + jmp L012ssse3_00_47
|
| .align 4,0x90
|
| -L011ssse3_00_47:
|
| +L012ssse3_00_47:
|
| addl $64,%ebp
|
| movl %edx,%ecx
|
| movdqa %xmm1,%xmm4
|
| @@ -3854,7 +4053,7 @@ L011ssse3_00_47:
|
| addl %ecx,%eax
|
| movdqa %xmm6,80(%esp)
|
| cmpl $66051,64(%ebp)
|
| - jne L011ssse3_00_47
|
| + jne L012ssse3_00_47
|
| movl %edx,%ecx
|
| rorl $14,%edx
|
| movl 20(%esp),%esi
|
| @@ -4368,1189 +4567,8 @@ L011ssse3_00_47:
|
| movdqa 64(%ebp),%xmm7
|
| subl $192,%ebp
|
| cmpl 104(%esp),%edi
|
| - jb L010grand_ssse3
|
| - movl 108(%esp),%esp
|
| - popl %edi
|
| - popl %esi
|
| - popl %ebx
|
| - popl %ebp
|
| - ret
|
| -.align 5,0x90
|
| -L004AVX:
|
| - leal -96(%esp),%esp
|
| - vzeroall
|
| - movl (%esi),%eax
|
| - movl 4(%esi),%ebx
|
| - movl 8(%esi),%ecx
|
| - movl 12(%esi),%edi
|
| - movl %ebx,4(%esp)
|
| - xorl %ecx,%ebx
|
| - movl %ecx,8(%esp)
|
| - movl %edi,12(%esp)
|
| - movl 16(%esi),%edx
|
| - movl 20(%esi),%edi
|
| - movl 24(%esi),%ecx
|
| - movl 28(%esi),%esi
|
| - movl %edi,20(%esp)
|
| - movl 100(%esp),%edi
|
| - movl %ecx,24(%esp)
|
| - movl %esi,28(%esp)
|
| - vmovdqa 256(%ebp),%xmm7
|
| - jmp L012grand_avx
|
| -.align 5,0x90
|
| -L012grand_avx:
|
| - vmovdqu (%edi),%xmm0
|
| - vmovdqu 16(%edi),%xmm1
|
| - vmovdqu 32(%edi),%xmm2
|
| - vmovdqu 48(%edi),%xmm3
|
| - addl $64,%edi
|
| - vpshufb %xmm7,%xmm0,%xmm0
|
| - movl %edi,100(%esp)
|
| - vpshufb %xmm7,%xmm1,%xmm1
|
| - vpshufb %xmm7,%xmm2,%xmm2
|
| - vpaddd (%ebp),%xmm0,%xmm4
|
| - vpshufb %xmm7,%xmm3,%xmm3
|
| - vpaddd 16(%ebp),%xmm1,%xmm5
|
| - vpaddd 32(%ebp),%xmm2,%xmm6
|
| - vpaddd 48(%ebp),%xmm3,%xmm7
|
| - vmovdqa %xmm4,32(%esp)
|
| - vmovdqa %xmm5,48(%esp)
|
| - vmovdqa %xmm6,64(%esp)
|
| - vmovdqa %xmm7,80(%esp)
|
| - jmp L013avx_00_47
|
| -.align 4,0x90
|
| -L013avx_00_47:
|
| - addl $64,%ebp
|
| - vpalignr $4,%xmm0,%xmm1,%xmm4
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 20(%esp),%esi
|
| - vpalignr $4,%xmm2,%xmm3,%xmm7
|
| - xorl %ecx,%edx
|
| - movl 24(%esp),%edi
|
| - xorl %edi,%esi
|
| - vpsrld $7,%xmm4,%xmm6
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,16(%esp)
|
| - vpaddd %xmm7,%xmm0,%xmm0
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - vpsrld $3,%xmm4,%xmm7
|
| - movl %eax,%ecx
|
| - addl %edi,%edx
|
| - movl 4(%esp),%edi
|
| - vpslld $14,%xmm4,%xmm5
|
| - movl %eax,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %eax,(%esp)
|
| - vpxor %xmm6,%xmm7,%xmm4
|
| - xorl %eax,%ecx
|
| - xorl %edi,%eax
|
| - addl 28(%esp),%edx
|
| - vpshufd $250,%xmm3,%xmm7
|
| - shrdl $11,%ecx,%ecx
|
| - andl %eax,%ebx
|
| - xorl %esi,%ecx
|
| - vpsrld $11,%xmm6,%xmm6
|
| - addl 32(%esp),%edx
|
| - xorl %edi,%ebx
|
| - shrdl $2,%ecx,%ecx
|
| - vpxor %xmm5,%xmm4,%xmm4
|
| - addl %edx,%ebx
|
| - addl 12(%esp),%edx
|
| - addl %ecx,%ebx
|
| - vpslld $11,%xmm5,%xmm5
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 16(%esp),%esi
|
| - vpxor %xmm6,%xmm4,%xmm4
|
| - xorl %ecx,%edx
|
| - movl 20(%esp),%edi
|
| - xorl %edi,%esi
|
| - vpsrld $10,%xmm7,%xmm6
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,12(%esp)
|
| - vpxor %xmm5,%xmm4,%xmm4
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - vpsrlq $17,%xmm7,%xmm5
|
| - movl %ebx,%ecx
|
| - addl %edi,%edx
|
| - movl (%esp),%edi
|
| - vpaddd %xmm4,%xmm0,%xmm0
|
| - movl %ebx,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %ebx,28(%esp)
|
| - vpxor %xmm5,%xmm6,%xmm6
|
| - xorl %ebx,%ecx
|
| - xorl %edi,%ebx
|
| - addl 24(%esp),%edx
|
| - vpsrlq $19,%xmm7,%xmm7
|
| - shrdl $11,%ecx,%ecx
|
| - andl %ebx,%eax
|
| - xorl %esi,%ecx
|
| - vpxor %xmm7,%xmm6,%xmm6
|
| - addl 36(%esp),%edx
|
| - xorl %edi,%eax
|
| - shrdl $2,%ecx,%ecx
|
| - vpshufd $132,%xmm6,%xmm7
|
| - addl %edx,%eax
|
| - addl 8(%esp),%edx
|
| - addl %ecx,%eax
|
| - vpsrldq $8,%xmm7,%xmm7
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 12(%esp),%esi
|
| - vpaddd %xmm7,%xmm0,%xmm0
|
| - xorl %ecx,%edx
|
| - movl 16(%esp),%edi
|
| - xorl %edi,%esi
|
| - vpshufd $80,%xmm0,%xmm7
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,8(%esp)
|
| - vpsrld $10,%xmm7,%xmm6
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - vpsrlq $17,%xmm7,%xmm5
|
| - movl %eax,%ecx
|
| - addl %edi,%edx
|
| - movl 28(%esp),%edi
|
| - vpxor %xmm5,%xmm6,%xmm6
|
| - movl %eax,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %eax,24(%esp)
|
| - vpsrlq $19,%xmm7,%xmm7
|
| - xorl %eax,%ecx
|
| - xorl %edi,%eax
|
| - addl 20(%esp),%edx
|
| - vpxor %xmm7,%xmm6,%xmm6
|
| - shrdl $11,%ecx,%ecx
|
| - andl %eax,%ebx
|
| - xorl %esi,%ecx
|
| - vpshufd $232,%xmm6,%xmm7
|
| - addl 40(%esp),%edx
|
| - xorl %edi,%ebx
|
| - shrdl $2,%ecx,%ecx
|
| - vpslldq $8,%xmm7,%xmm7
|
| - addl %edx,%ebx
|
| - addl 4(%esp),%edx
|
| - addl %ecx,%ebx
|
| - vpaddd %xmm7,%xmm0,%xmm0
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 8(%esp),%esi
|
| - vpaddd (%ebp),%xmm0,%xmm6
|
| - xorl %ecx,%edx
|
| - movl 12(%esp),%edi
|
| - xorl %edi,%esi
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,4(%esp)
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - movl %ebx,%ecx
|
| - addl %edi,%edx
|
| - movl 24(%esp),%edi
|
| - movl %ebx,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %ebx,20(%esp)
|
| - xorl %ebx,%ecx
|
| - xorl %edi,%ebx
|
| - addl 16(%esp),%edx
|
| - shrdl $11,%ecx,%ecx
|
| - andl %ebx,%eax
|
| - xorl %esi,%ecx
|
| - addl 44(%esp),%edx
|
| - xorl %edi,%eax
|
| - shrdl $2,%ecx,%ecx
|
| - addl %edx,%eax
|
| - addl (%esp),%edx
|
| - addl %ecx,%eax
|
| - vmovdqa %xmm6,32(%esp)
|
| - vpalignr $4,%xmm1,%xmm2,%xmm4
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 4(%esp),%esi
|
| - vpalignr $4,%xmm3,%xmm0,%xmm7
|
| - xorl %ecx,%edx
|
| - movl 8(%esp),%edi
|
| - xorl %edi,%esi
|
| - vpsrld $7,%xmm4,%xmm6
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,(%esp)
|
| - vpaddd %xmm7,%xmm1,%xmm1
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - vpsrld $3,%xmm4,%xmm7
|
| - movl %eax,%ecx
|
| - addl %edi,%edx
|
| - movl 20(%esp),%edi
|
| - vpslld $14,%xmm4,%xmm5
|
| - movl %eax,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %eax,16(%esp)
|
| - vpxor %xmm6,%xmm7,%xmm4
|
| - xorl %eax,%ecx
|
| - xorl %edi,%eax
|
| - addl 12(%esp),%edx
|
| - vpshufd $250,%xmm0,%xmm7
|
| - shrdl $11,%ecx,%ecx
|
| - andl %eax,%ebx
|
| - xorl %esi,%ecx
|
| - vpsrld $11,%xmm6,%xmm6
|
| - addl 48(%esp),%edx
|
| - xorl %edi,%ebx
|
| - shrdl $2,%ecx,%ecx
|
| - vpxor %xmm5,%xmm4,%xmm4
|
| - addl %edx,%ebx
|
| - addl 28(%esp),%edx
|
| - addl %ecx,%ebx
|
| - vpslld $11,%xmm5,%xmm5
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl (%esp),%esi
|
| - vpxor %xmm6,%xmm4,%xmm4
|
| - xorl %ecx,%edx
|
| - movl 4(%esp),%edi
|
| - xorl %edi,%esi
|
| - vpsrld $10,%xmm7,%xmm6
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,28(%esp)
|
| - vpxor %xmm5,%xmm4,%xmm4
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - vpsrlq $17,%xmm7,%xmm5
|
| - movl %ebx,%ecx
|
| - addl %edi,%edx
|
| - movl 16(%esp),%edi
|
| - vpaddd %xmm4,%xmm1,%xmm1
|
| - movl %ebx,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %ebx,12(%esp)
|
| - vpxor %xmm5,%xmm6,%xmm6
|
| - xorl %ebx,%ecx
|
| - xorl %edi,%ebx
|
| - addl 8(%esp),%edx
|
| - vpsrlq $19,%xmm7,%xmm7
|
| - shrdl $11,%ecx,%ecx
|
| - andl %ebx,%eax
|
| - xorl %esi,%ecx
|
| - vpxor %xmm7,%xmm6,%xmm6
|
| - addl 52(%esp),%edx
|
| - xorl %edi,%eax
|
| - shrdl $2,%ecx,%ecx
|
| - vpshufd $132,%xmm6,%xmm7
|
| - addl %edx,%eax
|
| - addl 24(%esp),%edx
|
| - addl %ecx,%eax
|
| - vpsrldq $8,%xmm7,%xmm7
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 28(%esp),%esi
|
| - vpaddd %xmm7,%xmm1,%xmm1
|
| - xorl %ecx,%edx
|
| - movl (%esp),%edi
|
| - xorl %edi,%esi
|
| - vpshufd $80,%xmm1,%xmm7
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,24(%esp)
|
| - vpsrld $10,%xmm7,%xmm6
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - vpsrlq $17,%xmm7,%xmm5
|
| - movl %eax,%ecx
|
| - addl %edi,%edx
|
| - movl 12(%esp),%edi
|
| - vpxor %xmm5,%xmm6,%xmm6
|
| - movl %eax,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %eax,8(%esp)
|
| - vpsrlq $19,%xmm7,%xmm7
|
| - xorl %eax,%ecx
|
| - xorl %edi,%eax
|
| - addl 4(%esp),%edx
|
| - vpxor %xmm7,%xmm6,%xmm6
|
| - shrdl $11,%ecx,%ecx
|
| - andl %eax,%ebx
|
| - xorl %esi,%ecx
|
| - vpshufd $232,%xmm6,%xmm7
|
| - addl 56(%esp),%edx
|
| - xorl %edi,%ebx
|
| - shrdl $2,%ecx,%ecx
|
| - vpslldq $8,%xmm7,%xmm7
|
| - addl %edx,%ebx
|
| - addl 20(%esp),%edx
|
| - addl %ecx,%ebx
|
| - vpaddd %xmm7,%xmm1,%xmm1
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 24(%esp),%esi
|
| - vpaddd 16(%ebp),%xmm1,%xmm6
|
| - xorl %ecx,%edx
|
| - movl 28(%esp),%edi
|
| - xorl %edi,%esi
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,20(%esp)
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - movl %ebx,%ecx
|
| - addl %edi,%edx
|
| - movl 8(%esp),%edi
|
| - movl %ebx,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %ebx,4(%esp)
|
| - xorl %ebx,%ecx
|
| - xorl %edi,%ebx
|
| - addl (%esp),%edx
|
| - shrdl $11,%ecx,%ecx
|
| - andl %ebx,%eax
|
| - xorl %esi,%ecx
|
| - addl 60(%esp),%edx
|
| - xorl %edi,%eax
|
| - shrdl $2,%ecx,%ecx
|
| - addl %edx,%eax
|
| - addl 16(%esp),%edx
|
| - addl %ecx,%eax
|
| - vmovdqa %xmm6,48(%esp)
|
| - vpalignr $4,%xmm2,%xmm3,%xmm4
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 20(%esp),%esi
|
| - vpalignr $4,%xmm0,%xmm1,%xmm7
|
| - xorl %ecx,%edx
|
| - movl 24(%esp),%edi
|
| - xorl %edi,%esi
|
| - vpsrld $7,%xmm4,%xmm6
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,16(%esp)
|
| - vpaddd %xmm7,%xmm2,%xmm2
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - vpsrld $3,%xmm4,%xmm7
|
| - movl %eax,%ecx
|
| - addl %edi,%edx
|
| - movl 4(%esp),%edi
|
| - vpslld $14,%xmm4,%xmm5
|
| - movl %eax,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %eax,(%esp)
|
| - vpxor %xmm6,%xmm7,%xmm4
|
| - xorl %eax,%ecx
|
| - xorl %edi,%eax
|
| - addl 28(%esp),%edx
|
| - vpshufd $250,%xmm1,%xmm7
|
| - shrdl $11,%ecx,%ecx
|
| - andl %eax,%ebx
|
| - xorl %esi,%ecx
|
| - vpsrld $11,%xmm6,%xmm6
|
| - addl 64(%esp),%edx
|
| - xorl %edi,%ebx
|
| - shrdl $2,%ecx,%ecx
|
| - vpxor %xmm5,%xmm4,%xmm4
|
| - addl %edx,%ebx
|
| - addl 12(%esp),%edx
|
| - addl %ecx,%ebx
|
| - vpslld $11,%xmm5,%xmm5
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 16(%esp),%esi
|
| - vpxor %xmm6,%xmm4,%xmm4
|
| - xorl %ecx,%edx
|
| - movl 20(%esp),%edi
|
| - xorl %edi,%esi
|
| - vpsrld $10,%xmm7,%xmm6
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,12(%esp)
|
| - vpxor %xmm5,%xmm4,%xmm4
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - vpsrlq $17,%xmm7,%xmm5
|
| - movl %ebx,%ecx
|
| - addl %edi,%edx
|
| - movl (%esp),%edi
|
| - vpaddd %xmm4,%xmm2,%xmm2
|
| - movl %ebx,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %ebx,28(%esp)
|
| - vpxor %xmm5,%xmm6,%xmm6
|
| - xorl %ebx,%ecx
|
| - xorl %edi,%ebx
|
| - addl 24(%esp),%edx
|
| - vpsrlq $19,%xmm7,%xmm7
|
| - shrdl $11,%ecx,%ecx
|
| - andl %ebx,%eax
|
| - xorl %esi,%ecx
|
| - vpxor %xmm7,%xmm6,%xmm6
|
| - addl 68(%esp),%edx
|
| - xorl %edi,%eax
|
| - shrdl $2,%ecx,%ecx
|
| - vpshufd $132,%xmm6,%xmm7
|
| - addl %edx,%eax
|
| - addl 8(%esp),%edx
|
| - addl %ecx,%eax
|
| - vpsrldq $8,%xmm7,%xmm7
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 12(%esp),%esi
|
| - vpaddd %xmm7,%xmm2,%xmm2
|
| - xorl %ecx,%edx
|
| - movl 16(%esp),%edi
|
| - xorl %edi,%esi
|
| - vpshufd $80,%xmm2,%xmm7
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,8(%esp)
|
| - vpsrld $10,%xmm7,%xmm6
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - vpsrlq $17,%xmm7,%xmm5
|
| - movl %eax,%ecx
|
| - addl %edi,%edx
|
| - movl 28(%esp),%edi
|
| - vpxor %xmm5,%xmm6,%xmm6
|
| - movl %eax,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %eax,24(%esp)
|
| - vpsrlq $19,%xmm7,%xmm7
|
| - xorl %eax,%ecx
|
| - xorl %edi,%eax
|
| - addl 20(%esp),%edx
|
| - vpxor %xmm7,%xmm6,%xmm6
|
| - shrdl $11,%ecx,%ecx
|
| - andl %eax,%ebx
|
| - xorl %esi,%ecx
|
| - vpshufd $232,%xmm6,%xmm7
|
| - addl 72(%esp),%edx
|
| - xorl %edi,%ebx
|
| - shrdl $2,%ecx,%ecx
|
| - vpslldq $8,%xmm7,%xmm7
|
| - addl %edx,%ebx
|
| - addl 4(%esp),%edx
|
| - addl %ecx,%ebx
|
| - vpaddd %xmm7,%xmm2,%xmm2
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 8(%esp),%esi
|
| - vpaddd 32(%ebp),%xmm2,%xmm6
|
| - xorl %ecx,%edx
|
| - movl 12(%esp),%edi
|
| - xorl %edi,%esi
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,4(%esp)
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - movl %ebx,%ecx
|
| - addl %edi,%edx
|
| - movl 24(%esp),%edi
|
| - movl %ebx,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %ebx,20(%esp)
|
| - xorl %ebx,%ecx
|
| - xorl %edi,%ebx
|
| - addl 16(%esp),%edx
|
| - shrdl $11,%ecx,%ecx
|
| - andl %ebx,%eax
|
| - xorl %esi,%ecx
|
| - addl 76(%esp),%edx
|
| - xorl %edi,%eax
|
| - shrdl $2,%ecx,%ecx
|
| - addl %edx,%eax
|
| - addl (%esp),%edx
|
| - addl %ecx,%eax
|
| - vmovdqa %xmm6,64(%esp)
|
| - vpalignr $4,%xmm3,%xmm0,%xmm4
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 4(%esp),%esi
|
| - vpalignr $4,%xmm1,%xmm2,%xmm7
|
| - xorl %ecx,%edx
|
| - movl 8(%esp),%edi
|
| - xorl %edi,%esi
|
| - vpsrld $7,%xmm4,%xmm6
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,(%esp)
|
| - vpaddd %xmm7,%xmm3,%xmm3
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - vpsrld $3,%xmm4,%xmm7
|
| - movl %eax,%ecx
|
| - addl %edi,%edx
|
| - movl 20(%esp),%edi
|
| - vpslld $14,%xmm4,%xmm5
|
| - movl %eax,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %eax,16(%esp)
|
| - vpxor %xmm6,%xmm7,%xmm4
|
| - xorl %eax,%ecx
|
| - xorl %edi,%eax
|
| - addl 12(%esp),%edx
|
| - vpshufd $250,%xmm2,%xmm7
|
| - shrdl $11,%ecx,%ecx
|
| - andl %eax,%ebx
|
| - xorl %esi,%ecx
|
| - vpsrld $11,%xmm6,%xmm6
|
| - addl 80(%esp),%edx
|
| - xorl %edi,%ebx
|
| - shrdl $2,%ecx,%ecx
|
| - vpxor %xmm5,%xmm4,%xmm4
|
| - addl %edx,%ebx
|
| - addl 28(%esp),%edx
|
| - addl %ecx,%ebx
|
| - vpslld $11,%xmm5,%xmm5
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl (%esp),%esi
|
| - vpxor %xmm6,%xmm4,%xmm4
|
| - xorl %ecx,%edx
|
| - movl 4(%esp),%edi
|
| - xorl %edi,%esi
|
| - vpsrld $10,%xmm7,%xmm6
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,28(%esp)
|
| - vpxor %xmm5,%xmm4,%xmm4
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - vpsrlq $17,%xmm7,%xmm5
|
| - movl %ebx,%ecx
|
| - addl %edi,%edx
|
| - movl 16(%esp),%edi
|
| - vpaddd %xmm4,%xmm3,%xmm3
|
| - movl %ebx,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %ebx,12(%esp)
|
| - vpxor %xmm5,%xmm6,%xmm6
|
| - xorl %ebx,%ecx
|
| - xorl %edi,%ebx
|
| - addl 8(%esp),%edx
|
| - vpsrlq $19,%xmm7,%xmm7
|
| - shrdl $11,%ecx,%ecx
|
| - andl %ebx,%eax
|
| - xorl %esi,%ecx
|
| - vpxor %xmm7,%xmm6,%xmm6
|
| - addl 84(%esp),%edx
|
| - xorl %edi,%eax
|
| - shrdl $2,%ecx,%ecx
|
| - vpshufd $132,%xmm6,%xmm7
|
| - addl %edx,%eax
|
| - addl 24(%esp),%edx
|
| - addl %ecx,%eax
|
| - vpsrldq $8,%xmm7,%xmm7
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 28(%esp),%esi
|
| - vpaddd %xmm7,%xmm3,%xmm3
|
| - xorl %ecx,%edx
|
| - movl (%esp),%edi
|
| - xorl %edi,%esi
|
| - vpshufd $80,%xmm3,%xmm7
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,24(%esp)
|
| - vpsrld $10,%xmm7,%xmm6
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - vpsrlq $17,%xmm7,%xmm5
|
| - movl %eax,%ecx
|
| - addl %edi,%edx
|
| - movl 12(%esp),%edi
|
| - vpxor %xmm5,%xmm6,%xmm6
|
| - movl %eax,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %eax,8(%esp)
|
| - vpsrlq $19,%xmm7,%xmm7
|
| - xorl %eax,%ecx
|
| - xorl %edi,%eax
|
| - addl 4(%esp),%edx
|
| - vpxor %xmm7,%xmm6,%xmm6
|
| - shrdl $11,%ecx,%ecx
|
| - andl %eax,%ebx
|
| - xorl %esi,%ecx
|
| - vpshufd $232,%xmm6,%xmm7
|
| - addl 88(%esp),%edx
|
| - xorl %edi,%ebx
|
| - shrdl $2,%ecx,%ecx
|
| - vpslldq $8,%xmm7,%xmm7
|
| - addl %edx,%ebx
|
| - addl 20(%esp),%edx
|
| - addl %ecx,%ebx
|
| - vpaddd %xmm7,%xmm3,%xmm3
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 24(%esp),%esi
|
| - vpaddd 48(%ebp),%xmm3,%xmm6
|
| - xorl %ecx,%edx
|
| - movl 28(%esp),%edi
|
| - xorl %edi,%esi
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,20(%esp)
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - movl %ebx,%ecx
|
| - addl %edi,%edx
|
| - movl 8(%esp),%edi
|
| - movl %ebx,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %ebx,4(%esp)
|
| - xorl %ebx,%ecx
|
| - xorl %edi,%ebx
|
| - addl (%esp),%edx
|
| - shrdl $11,%ecx,%ecx
|
| - andl %ebx,%eax
|
| - xorl %esi,%ecx
|
| - addl 92(%esp),%edx
|
| - xorl %edi,%eax
|
| - shrdl $2,%ecx,%ecx
|
| - addl %edx,%eax
|
| - addl 16(%esp),%edx
|
| - addl %ecx,%eax
|
| - vmovdqa %xmm6,80(%esp)
|
| - cmpl $66051,64(%ebp)
|
| - jne L013avx_00_47
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 20(%esp),%esi
|
| - xorl %ecx,%edx
|
| - movl 24(%esp),%edi
|
| - xorl %edi,%esi
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,16(%esp)
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - movl %eax,%ecx
|
| - addl %edi,%edx
|
| - movl 4(%esp),%edi
|
| - movl %eax,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %eax,(%esp)
|
| - xorl %eax,%ecx
|
| - xorl %edi,%eax
|
| - addl 28(%esp),%edx
|
| - shrdl $11,%ecx,%ecx
|
| - andl %eax,%ebx
|
| - xorl %esi,%ecx
|
| - addl 32(%esp),%edx
|
| - xorl %edi,%ebx
|
| - shrdl $2,%ecx,%ecx
|
| - addl %edx,%ebx
|
| - addl 12(%esp),%edx
|
| - addl %ecx,%ebx
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 16(%esp),%esi
|
| - xorl %ecx,%edx
|
| - movl 20(%esp),%edi
|
| - xorl %edi,%esi
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,12(%esp)
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - movl %ebx,%ecx
|
| - addl %edi,%edx
|
| - movl (%esp),%edi
|
| - movl %ebx,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %ebx,28(%esp)
|
| - xorl %ebx,%ecx
|
| - xorl %edi,%ebx
|
| - addl 24(%esp),%edx
|
| - shrdl $11,%ecx,%ecx
|
| - andl %ebx,%eax
|
| - xorl %esi,%ecx
|
| - addl 36(%esp),%edx
|
| - xorl %edi,%eax
|
| - shrdl $2,%ecx,%ecx
|
| - addl %edx,%eax
|
| - addl 8(%esp),%edx
|
| - addl %ecx,%eax
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 12(%esp),%esi
|
| - xorl %ecx,%edx
|
| - movl 16(%esp),%edi
|
| - xorl %edi,%esi
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,8(%esp)
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - movl %eax,%ecx
|
| - addl %edi,%edx
|
| - movl 28(%esp),%edi
|
| - movl %eax,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %eax,24(%esp)
|
| - xorl %eax,%ecx
|
| - xorl %edi,%eax
|
| - addl 20(%esp),%edx
|
| - shrdl $11,%ecx,%ecx
|
| - andl %eax,%ebx
|
| - xorl %esi,%ecx
|
| - addl 40(%esp),%edx
|
| - xorl %edi,%ebx
|
| - shrdl $2,%ecx,%ecx
|
| - addl %edx,%ebx
|
| - addl 4(%esp),%edx
|
| - addl %ecx,%ebx
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 8(%esp),%esi
|
| - xorl %ecx,%edx
|
| - movl 12(%esp),%edi
|
| - xorl %edi,%esi
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,4(%esp)
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - movl %ebx,%ecx
|
| - addl %edi,%edx
|
| - movl 24(%esp),%edi
|
| - movl %ebx,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %ebx,20(%esp)
|
| - xorl %ebx,%ecx
|
| - xorl %edi,%ebx
|
| - addl 16(%esp),%edx
|
| - shrdl $11,%ecx,%ecx
|
| - andl %ebx,%eax
|
| - xorl %esi,%ecx
|
| - addl 44(%esp),%edx
|
| - xorl %edi,%eax
|
| - shrdl $2,%ecx,%ecx
|
| - addl %edx,%eax
|
| - addl (%esp),%edx
|
| - addl %ecx,%eax
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 4(%esp),%esi
|
| - xorl %ecx,%edx
|
| - movl 8(%esp),%edi
|
| - xorl %edi,%esi
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,(%esp)
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - movl %eax,%ecx
|
| - addl %edi,%edx
|
| - movl 20(%esp),%edi
|
| - movl %eax,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %eax,16(%esp)
|
| - xorl %eax,%ecx
|
| - xorl %edi,%eax
|
| - addl 12(%esp),%edx
|
| - shrdl $11,%ecx,%ecx
|
| - andl %eax,%ebx
|
| - xorl %esi,%ecx
|
| - addl 48(%esp),%edx
|
| - xorl %edi,%ebx
|
| - shrdl $2,%ecx,%ecx
|
| - addl %edx,%ebx
|
| - addl 28(%esp),%edx
|
| - addl %ecx,%ebx
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl (%esp),%esi
|
| - xorl %ecx,%edx
|
| - movl 4(%esp),%edi
|
| - xorl %edi,%esi
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,28(%esp)
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - movl %ebx,%ecx
|
| - addl %edi,%edx
|
| - movl 16(%esp),%edi
|
| - movl %ebx,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %ebx,12(%esp)
|
| - xorl %ebx,%ecx
|
| - xorl %edi,%ebx
|
| - addl 8(%esp),%edx
|
| - shrdl $11,%ecx,%ecx
|
| - andl %ebx,%eax
|
| - xorl %esi,%ecx
|
| - addl 52(%esp),%edx
|
| - xorl %edi,%eax
|
| - shrdl $2,%ecx,%ecx
|
| - addl %edx,%eax
|
| - addl 24(%esp),%edx
|
| - addl %ecx,%eax
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 28(%esp),%esi
|
| - xorl %ecx,%edx
|
| - movl (%esp),%edi
|
| - xorl %edi,%esi
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,24(%esp)
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - movl %eax,%ecx
|
| - addl %edi,%edx
|
| - movl 12(%esp),%edi
|
| - movl %eax,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %eax,8(%esp)
|
| - xorl %eax,%ecx
|
| - xorl %edi,%eax
|
| - addl 4(%esp),%edx
|
| - shrdl $11,%ecx,%ecx
|
| - andl %eax,%ebx
|
| - xorl %esi,%ecx
|
| - addl 56(%esp),%edx
|
| - xorl %edi,%ebx
|
| - shrdl $2,%ecx,%ecx
|
| - addl %edx,%ebx
|
| - addl 20(%esp),%edx
|
| - addl %ecx,%ebx
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 24(%esp),%esi
|
| - xorl %ecx,%edx
|
| - movl 28(%esp),%edi
|
| - xorl %edi,%esi
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,20(%esp)
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - movl %ebx,%ecx
|
| - addl %edi,%edx
|
| - movl 8(%esp),%edi
|
| - movl %ebx,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %ebx,4(%esp)
|
| - xorl %ebx,%ecx
|
| - xorl %edi,%ebx
|
| - addl (%esp),%edx
|
| - shrdl $11,%ecx,%ecx
|
| - andl %ebx,%eax
|
| - xorl %esi,%ecx
|
| - addl 60(%esp),%edx
|
| - xorl %edi,%eax
|
| - shrdl $2,%ecx,%ecx
|
| - addl %edx,%eax
|
| - addl 16(%esp),%edx
|
| - addl %ecx,%eax
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 20(%esp),%esi
|
| - xorl %ecx,%edx
|
| - movl 24(%esp),%edi
|
| - xorl %edi,%esi
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,16(%esp)
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - movl %eax,%ecx
|
| - addl %edi,%edx
|
| - movl 4(%esp),%edi
|
| - movl %eax,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %eax,(%esp)
|
| - xorl %eax,%ecx
|
| - xorl %edi,%eax
|
| - addl 28(%esp),%edx
|
| - shrdl $11,%ecx,%ecx
|
| - andl %eax,%ebx
|
| - xorl %esi,%ecx
|
| - addl 64(%esp),%edx
|
| - xorl %edi,%ebx
|
| - shrdl $2,%ecx,%ecx
|
| - addl %edx,%ebx
|
| - addl 12(%esp),%edx
|
| - addl %ecx,%ebx
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 16(%esp),%esi
|
| - xorl %ecx,%edx
|
| - movl 20(%esp),%edi
|
| - xorl %edi,%esi
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,12(%esp)
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - movl %ebx,%ecx
|
| - addl %edi,%edx
|
| - movl (%esp),%edi
|
| - movl %ebx,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %ebx,28(%esp)
|
| - xorl %ebx,%ecx
|
| - xorl %edi,%ebx
|
| - addl 24(%esp),%edx
|
| - shrdl $11,%ecx,%ecx
|
| - andl %ebx,%eax
|
| - xorl %esi,%ecx
|
| - addl 68(%esp),%edx
|
| - xorl %edi,%eax
|
| - shrdl $2,%ecx,%ecx
|
| - addl %edx,%eax
|
| - addl 8(%esp),%edx
|
| - addl %ecx,%eax
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 12(%esp),%esi
|
| - xorl %ecx,%edx
|
| - movl 16(%esp),%edi
|
| - xorl %edi,%esi
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,8(%esp)
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - movl %eax,%ecx
|
| - addl %edi,%edx
|
| - movl 28(%esp),%edi
|
| - movl %eax,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %eax,24(%esp)
|
| - xorl %eax,%ecx
|
| - xorl %edi,%eax
|
| - addl 20(%esp),%edx
|
| - shrdl $11,%ecx,%ecx
|
| - andl %eax,%ebx
|
| - xorl %esi,%ecx
|
| - addl 72(%esp),%edx
|
| - xorl %edi,%ebx
|
| - shrdl $2,%ecx,%ecx
|
| - addl %edx,%ebx
|
| - addl 4(%esp),%edx
|
| - addl %ecx,%ebx
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 8(%esp),%esi
|
| - xorl %ecx,%edx
|
| - movl 12(%esp),%edi
|
| - xorl %edi,%esi
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,4(%esp)
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - movl %ebx,%ecx
|
| - addl %edi,%edx
|
| - movl 24(%esp),%edi
|
| - movl %ebx,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %ebx,20(%esp)
|
| - xorl %ebx,%ecx
|
| - xorl %edi,%ebx
|
| - addl 16(%esp),%edx
|
| - shrdl $11,%ecx,%ecx
|
| - andl %ebx,%eax
|
| - xorl %esi,%ecx
|
| - addl 76(%esp),%edx
|
| - xorl %edi,%eax
|
| - shrdl $2,%ecx,%ecx
|
| - addl %edx,%eax
|
| - addl (%esp),%edx
|
| - addl %ecx,%eax
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 4(%esp),%esi
|
| - xorl %ecx,%edx
|
| - movl 8(%esp),%edi
|
| - xorl %edi,%esi
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,(%esp)
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - movl %eax,%ecx
|
| - addl %edi,%edx
|
| - movl 20(%esp),%edi
|
| - movl %eax,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %eax,16(%esp)
|
| - xorl %eax,%ecx
|
| - xorl %edi,%eax
|
| - addl 12(%esp),%edx
|
| - shrdl $11,%ecx,%ecx
|
| - andl %eax,%ebx
|
| - xorl %esi,%ecx
|
| - addl 80(%esp),%edx
|
| - xorl %edi,%ebx
|
| - shrdl $2,%ecx,%ecx
|
| - addl %edx,%ebx
|
| - addl 28(%esp),%edx
|
| - addl %ecx,%ebx
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl (%esp),%esi
|
| - xorl %ecx,%edx
|
| - movl 4(%esp),%edi
|
| - xorl %edi,%esi
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,28(%esp)
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - movl %ebx,%ecx
|
| - addl %edi,%edx
|
| - movl 16(%esp),%edi
|
| - movl %ebx,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %ebx,12(%esp)
|
| - xorl %ebx,%ecx
|
| - xorl %edi,%ebx
|
| - addl 8(%esp),%edx
|
| - shrdl $11,%ecx,%ecx
|
| - andl %ebx,%eax
|
| - xorl %esi,%ecx
|
| - addl 84(%esp),%edx
|
| - xorl %edi,%eax
|
| - shrdl $2,%ecx,%ecx
|
| - addl %edx,%eax
|
| - addl 24(%esp),%edx
|
| - addl %ecx,%eax
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 28(%esp),%esi
|
| - xorl %ecx,%edx
|
| - movl (%esp),%edi
|
| - xorl %edi,%esi
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,24(%esp)
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - movl %eax,%ecx
|
| - addl %edi,%edx
|
| - movl 12(%esp),%edi
|
| - movl %eax,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %eax,8(%esp)
|
| - xorl %eax,%ecx
|
| - xorl %edi,%eax
|
| - addl 4(%esp),%edx
|
| - shrdl $11,%ecx,%ecx
|
| - andl %eax,%ebx
|
| - xorl %esi,%ecx
|
| - addl 88(%esp),%edx
|
| - xorl %edi,%ebx
|
| - shrdl $2,%ecx,%ecx
|
| - addl %edx,%ebx
|
| - addl 20(%esp),%edx
|
| - addl %ecx,%ebx
|
| - movl %edx,%ecx
|
| - shrdl $14,%edx,%edx
|
| - movl 24(%esp),%esi
|
| - xorl %ecx,%edx
|
| - movl 28(%esp),%edi
|
| - xorl %edi,%esi
|
| - shrdl $5,%edx,%edx
|
| - andl %ecx,%esi
|
| - movl %ecx,20(%esp)
|
| - xorl %ecx,%edx
|
| - xorl %esi,%edi
|
| - shrdl $6,%edx,%edx
|
| - movl %ebx,%ecx
|
| - addl %edi,%edx
|
| - movl 8(%esp),%edi
|
| - movl %ebx,%esi
|
| - shrdl $9,%ecx,%ecx
|
| - movl %ebx,4(%esp)
|
| - xorl %ebx,%ecx
|
| - xorl %edi,%ebx
|
| - addl (%esp),%edx
|
| - shrdl $11,%ecx,%ecx
|
| - andl %ebx,%eax
|
| - xorl %esi,%ecx
|
| - addl 92(%esp),%edx
|
| - xorl %edi,%eax
|
| - shrdl $2,%ecx,%ecx
|
| - addl %edx,%eax
|
| - addl 16(%esp),%edx
|
| - addl %ecx,%eax
|
| - movl 96(%esp),%esi
|
| - xorl %edi,%ebx
|
| - movl 12(%esp),%ecx
|
| - addl (%esi),%eax
|
| - addl 4(%esi),%ebx
|
| - addl 8(%esi),%edi
|
| - addl 12(%esi),%ecx
|
| - movl %eax,(%esi)
|
| - movl %ebx,4(%esi)
|
| - movl %edi,8(%esi)
|
| - movl %ecx,12(%esi)
|
| - movl %ebx,4(%esp)
|
| - xorl %edi,%ebx
|
| - movl %edi,8(%esp)
|
| - movl %ecx,12(%esp)
|
| - movl 20(%esp),%edi
|
| - movl 24(%esp),%ecx
|
| - addl 16(%esi),%edx
|
| - addl 20(%esi),%edi
|
| - addl 24(%esi),%ecx
|
| - movl %edx,16(%esi)
|
| - movl %edi,20(%esi)
|
| - movl %edi,20(%esp)
|
| - movl 28(%esp),%edi
|
| - movl %ecx,24(%esi)
|
| - addl 28(%esi),%edi
|
| - movl %ecx,24(%esp)
|
| - movl %edi,28(%esi)
|
| - movl %edi,28(%esp)
|
| - movl 100(%esp),%edi
|
| - vmovdqa 64(%ebp),%xmm7
|
| - subl $192,%ebp
|
| - cmpl 104(%esp),%edi
|
| - jb L012grand_avx
|
| + jb L011grand_ssse3
|
| movl 108(%esp),%esp
|
| - vzeroall
|
| popl %edi
|
| popl %esi
|
| popl %ebx
|
|
|