| Index: third_party/boringssl/linux-x86/crypto/aes/aes-586.S
|
| diff --git a/third_party/boringssl/linux-x86/crypto/aes/aes-586.S b/third_party/boringssl/linux-x86/crypto/aes/aes-586.S
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..74282748ce0da92ead2f775480cd3409ac583f78
|
| --- /dev/null
|
| +++ b/third_party/boringssl/linux-x86/crypto/aes/aes-586.S
|
| @@ -0,0 +1,3257 @@
|
| +#if defined(__i386__)
|
| +.file "aes-586.S"
|
| +.text
|
| +.hidden _x86_AES_encrypt_compact
|
| +.type _x86_AES_encrypt_compact,@function
|
| +.align 16
|
| +_x86_AES_encrypt_compact:
|
| + movl %edi,20(%esp)
|
| + xorl (%edi),%eax
|
| + xorl 4(%edi),%ebx
|
| + xorl 8(%edi),%ecx
|
| + xorl 12(%edi),%edx
|
| + movl 240(%edi),%esi
|
| + leal -2(%esi,%esi,1),%esi
|
| + leal (%edi,%esi,8),%esi
|
| + movl %esi,24(%esp)
|
| + movl -128(%ebp),%edi
|
| + movl -96(%ebp),%esi
|
| + movl -64(%ebp),%edi
|
| + movl -32(%ebp),%esi
|
| + movl (%ebp),%edi
|
| + movl 32(%ebp),%esi
|
| + movl 64(%ebp),%edi
|
| + movl 96(%ebp),%esi
|
| +.align 16
|
| +.L000loop:
|
| + movl %eax,%esi
|
| + andl $255,%esi
|
| + movzbl -128(%ebp,%esi,1),%esi
|
| + movzbl %bh,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $8,%edi
|
| + xorl %edi,%esi
|
| + movl %ecx,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $16,%edi
|
| + xorl %edi,%esi
|
| + movl %edx,%edi
|
| + shrl $24,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $24,%edi
|
| + xorl %edi,%esi
|
| + movl %esi,4(%esp)
|
| +
|
| + movl %ebx,%esi
|
| + andl $255,%esi
|
| + shrl $16,%ebx
|
| + movzbl -128(%ebp,%esi,1),%esi
|
| + movzbl %ch,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $8,%edi
|
| + xorl %edi,%esi
|
| + movl %edx,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $16,%edi
|
| + xorl %edi,%esi
|
| + movl %eax,%edi
|
| + shrl $24,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $24,%edi
|
| + xorl %edi,%esi
|
| + movl %esi,8(%esp)
|
| +
|
| + movl %ecx,%esi
|
| + andl $255,%esi
|
| + shrl $24,%ecx
|
| + movzbl -128(%ebp,%esi,1),%esi
|
| + movzbl %dh,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $8,%edi
|
| + xorl %edi,%esi
|
| + movl %eax,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edx
|
| + andl $255,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $16,%edi
|
| + xorl %edi,%esi
|
| + movzbl %bh,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $24,%edi
|
| + xorl %edi,%esi
|
| +
|
| + andl $255,%edx
|
| + movzbl -128(%ebp,%edx,1),%edx
|
| + movzbl %ah,%eax
|
| + movzbl -128(%ebp,%eax,1),%eax
|
| + shll $8,%eax
|
| + xorl %eax,%edx
|
| + movl 4(%esp),%eax
|
| + andl $255,%ebx
|
| + movzbl -128(%ebp,%ebx,1),%ebx
|
| + shll $16,%ebx
|
| + xorl %ebx,%edx
|
| + movl 8(%esp),%ebx
|
| + movzbl -128(%ebp,%ecx,1),%ecx
|
| + shll $24,%ecx
|
| + xorl %ecx,%edx
|
| + movl %esi,%ecx
|
| +
|
| + movl $2155905152,%ebp
|
| + andl %ecx,%ebp
|
| + leal (%ecx,%ecx,1),%edi
|
| + movl %ebp,%esi
|
| + shrl $7,%ebp
|
| + andl $4278124286,%edi
|
| + subl %ebp,%esi
|
| + movl %ecx,%ebp
|
| + andl $454761243,%esi
|
| + rorl $16,%ebp
|
| + xorl %edi,%esi
|
| + movl %ecx,%edi
|
| + xorl %esi,%ecx
|
| + rorl $24,%edi
|
| + xorl %ebp,%esi
|
| + roll $24,%ecx
|
| + xorl %edi,%esi
|
| + movl $2155905152,%ebp
|
| + xorl %esi,%ecx
|
| + andl %edx,%ebp
|
| + leal (%edx,%edx,1),%edi
|
| + movl %ebp,%esi
|
| + shrl $7,%ebp
|
| + andl $4278124286,%edi
|
| + subl %ebp,%esi
|
| + movl %edx,%ebp
|
| + andl $454761243,%esi
|
| + rorl $16,%ebp
|
| + xorl %edi,%esi
|
| + movl %edx,%edi
|
| + xorl %esi,%edx
|
| + rorl $24,%edi
|
| + xorl %ebp,%esi
|
| + roll $24,%edx
|
| + xorl %edi,%esi
|
| + movl $2155905152,%ebp
|
| + xorl %esi,%edx
|
| + andl %eax,%ebp
|
| + leal (%eax,%eax,1),%edi
|
| + movl %ebp,%esi
|
| + shrl $7,%ebp
|
| + andl $4278124286,%edi
|
| + subl %ebp,%esi
|
| + movl %eax,%ebp
|
| + andl $454761243,%esi
|
| + rorl $16,%ebp
|
| + xorl %edi,%esi
|
| + movl %eax,%edi
|
| + xorl %esi,%eax
|
| + rorl $24,%edi
|
| + xorl %ebp,%esi
|
| + roll $24,%eax
|
| + xorl %edi,%esi
|
| + movl $2155905152,%ebp
|
| + xorl %esi,%eax
|
| + andl %ebx,%ebp
|
| + leal (%ebx,%ebx,1),%edi
|
| + movl %ebp,%esi
|
| + shrl $7,%ebp
|
| + andl $4278124286,%edi
|
| + subl %ebp,%esi
|
| + movl %ebx,%ebp
|
| + andl $454761243,%esi
|
| + rorl $16,%ebp
|
| + xorl %edi,%esi
|
| + movl %ebx,%edi
|
| + xorl %esi,%ebx
|
| + rorl $24,%edi
|
| + xorl %ebp,%esi
|
| + roll $24,%ebx
|
| + xorl %edi,%esi
|
| + xorl %esi,%ebx
|
| + movl 20(%esp),%edi
|
| + movl 28(%esp),%ebp
|
| + addl $16,%edi
|
| + xorl (%edi),%eax
|
| + xorl 4(%edi),%ebx
|
| + xorl 8(%edi),%ecx
|
| + xorl 12(%edi),%edx
|
| + cmpl 24(%esp),%edi
|
| + movl %edi,20(%esp)
|
| + jb .L000loop
|
| + movl %eax,%esi
|
| + andl $255,%esi
|
| + movzbl -128(%ebp,%esi,1),%esi
|
| + movzbl %bh,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $8,%edi
|
| + xorl %edi,%esi
|
| + movl %ecx,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $16,%edi
|
| + xorl %edi,%esi
|
| + movl %edx,%edi
|
| + shrl $24,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $24,%edi
|
| + xorl %edi,%esi
|
| + movl %esi,4(%esp)
|
| +
|
| + movl %ebx,%esi
|
| + andl $255,%esi
|
| + shrl $16,%ebx
|
| + movzbl -128(%ebp,%esi,1),%esi
|
| + movzbl %ch,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $8,%edi
|
| + xorl %edi,%esi
|
| + movl %edx,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $16,%edi
|
| + xorl %edi,%esi
|
| + movl %eax,%edi
|
| + shrl $24,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $24,%edi
|
| + xorl %edi,%esi
|
| + movl %esi,8(%esp)
|
| +
|
| + movl %ecx,%esi
|
| + andl $255,%esi
|
| + shrl $24,%ecx
|
| + movzbl -128(%ebp,%esi,1),%esi
|
| + movzbl %dh,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $8,%edi
|
| + xorl %edi,%esi
|
| + movl %eax,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edx
|
| + andl $255,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $16,%edi
|
| + xorl %edi,%esi
|
| + movzbl %bh,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $24,%edi
|
| + xorl %edi,%esi
|
| +
|
| + movl 20(%esp),%edi
|
| + andl $255,%edx
|
| + movzbl -128(%ebp,%edx,1),%edx
|
| + movzbl %ah,%eax
|
| + movzbl -128(%ebp,%eax,1),%eax
|
| + shll $8,%eax
|
| + xorl %eax,%edx
|
| + movl 4(%esp),%eax
|
| + andl $255,%ebx
|
| + movzbl -128(%ebp,%ebx,1),%ebx
|
| + shll $16,%ebx
|
| + xorl %ebx,%edx
|
| + movl 8(%esp),%ebx
|
| + movzbl -128(%ebp,%ecx,1),%ecx
|
| + shll $24,%ecx
|
| + xorl %ecx,%edx
|
| + movl %esi,%ecx
|
| +
|
| + xorl 16(%edi),%eax
|
| + xorl 20(%edi),%ebx
|
| + xorl 24(%edi),%ecx
|
| + xorl 28(%edi),%edx
|
| + ret
|
| +.size _x86_AES_encrypt_compact,.-_x86_AES_encrypt_compact
|
| +.hidden _sse_AES_encrypt_compact
|
| +.type _sse_AES_encrypt_compact,@function
|
| +.align 16
|
| +_sse_AES_encrypt_compact:
|
| + pxor (%edi),%mm0
|
| + pxor 8(%edi),%mm4
|
| + movl 240(%edi),%esi
|
| + leal -2(%esi,%esi,1),%esi
|
| + leal (%edi,%esi,8),%esi
|
| + movl %esi,24(%esp)
|
| + movl $454761243,%eax
|
| + movl %eax,8(%esp)
|
| + movl %eax,12(%esp)
|
| + movl -128(%ebp),%eax
|
| + movl -96(%ebp),%ebx
|
| + movl -64(%ebp),%ecx
|
| + movl -32(%ebp),%edx
|
| + movl (%ebp),%eax
|
| + movl 32(%ebp),%ebx
|
| + movl 64(%ebp),%ecx
|
| + movl 96(%ebp),%edx
|
| +.align 16
|
| +.L001loop:
|
| + pshufw $8,%mm0,%mm1
|
| + pshufw $13,%mm4,%mm5
|
| + movd %mm1,%eax
|
| + movd %mm5,%ebx
|
| + movl %edi,20(%esp)
|
| + movzbl %al,%esi
|
| + movzbl %ah,%edx
|
| + pshufw $13,%mm0,%mm2
|
| + movzbl -128(%ebp,%esi,1),%ecx
|
| + movzbl %bl,%edi
|
| + movzbl -128(%ebp,%edx,1),%edx
|
| + shrl $16,%eax
|
| + shll $8,%edx
|
| + movzbl -128(%ebp,%edi,1),%esi
|
| + movzbl %bh,%edi
|
| + shll $16,%esi
|
| + pshufw $8,%mm4,%mm6
|
| + orl %esi,%ecx
|
| + movzbl -128(%ebp,%edi,1),%esi
|
| + movzbl %ah,%edi
|
| + shll $24,%esi
|
| + shrl $16,%ebx
|
| + orl %esi,%edx
|
| + movzbl -128(%ebp,%edi,1),%esi
|
| + movzbl %bh,%edi
|
| + shll $8,%esi
|
| + orl %esi,%ecx
|
| + movzbl -128(%ebp,%edi,1),%esi
|
| + movzbl %al,%edi
|
| + shll $24,%esi
|
| + orl %esi,%ecx
|
| + movzbl -128(%ebp,%edi,1),%esi
|
| + movzbl %bl,%edi
|
| + movd %mm2,%eax
|
| + movd %ecx,%mm0
|
| + movzbl -128(%ebp,%edi,1),%ecx
|
| + movzbl %ah,%edi
|
| + shll $16,%ecx
|
| + movd %mm6,%ebx
|
| + orl %esi,%ecx
|
| + movzbl -128(%ebp,%edi,1),%esi
|
| + movzbl %bh,%edi
|
| + shll $24,%esi
|
| + orl %esi,%ecx
|
| + movzbl -128(%ebp,%edi,1),%esi
|
| + movzbl %bl,%edi
|
| + shll $8,%esi
|
| + shrl $16,%ebx
|
| + orl %esi,%ecx
|
| + movzbl -128(%ebp,%edi,1),%esi
|
| + movzbl %al,%edi
|
| + shrl $16,%eax
|
| + movd %ecx,%mm1
|
| + movzbl -128(%ebp,%edi,1),%ecx
|
| + movzbl %ah,%edi
|
| + shll $16,%ecx
|
| + andl $255,%eax
|
| + orl %esi,%ecx
|
| + punpckldq %mm1,%mm0
|
| + movzbl -128(%ebp,%edi,1),%esi
|
| + movzbl %bh,%edi
|
| + shll $24,%esi
|
| + andl $255,%ebx
|
| + movzbl -128(%ebp,%eax,1),%eax
|
| + orl %esi,%ecx
|
| + shll $16,%eax
|
| + movzbl -128(%ebp,%edi,1),%esi
|
| + orl %eax,%edx
|
| + shll $8,%esi
|
| + movzbl -128(%ebp,%ebx,1),%ebx
|
| + orl %esi,%ecx
|
| + orl %ebx,%edx
|
| + movl 20(%esp),%edi
|
| + movd %ecx,%mm4
|
| + movd %edx,%mm5
|
| + punpckldq %mm5,%mm4
|
| + addl $16,%edi
|
| + cmpl 24(%esp),%edi
|
| + ja .L002out
|
| + movq 8(%esp),%mm2
|
| + pxor %mm3,%mm3
|
| + pxor %mm7,%mm7
|
| + movq %mm0,%mm1
|
| + movq %mm4,%mm5
|
| + pcmpgtb %mm0,%mm3
|
| + pcmpgtb %mm4,%mm7
|
| + pand %mm2,%mm3
|
| + pand %mm2,%mm7
|
| + pshufw $177,%mm0,%mm2
|
| + pshufw $177,%mm4,%mm6
|
| + paddb %mm0,%mm0
|
| + paddb %mm4,%mm4
|
| + pxor %mm3,%mm0
|
| + pxor %mm7,%mm4
|
| + pshufw $177,%mm2,%mm3
|
| + pshufw $177,%mm6,%mm7
|
| + pxor %mm0,%mm1
|
| + pxor %mm4,%mm5
|
| + pxor %mm2,%mm0
|
| + pxor %mm6,%mm4
|
| + movq %mm3,%mm2
|
| + movq %mm7,%mm6
|
| + pslld $8,%mm3
|
| + pslld $8,%mm7
|
| + psrld $24,%mm2
|
| + psrld $24,%mm6
|
| + pxor %mm3,%mm0
|
| + pxor %mm7,%mm4
|
| + pxor %mm2,%mm0
|
| + pxor %mm6,%mm4
|
| + movq %mm1,%mm3
|
| + movq %mm5,%mm7
|
| + movq (%edi),%mm2
|
| + movq 8(%edi),%mm6
|
| + psrld $8,%mm1
|
| + psrld $8,%mm5
|
| + movl -128(%ebp),%eax
|
| + pslld $24,%mm3
|
| + pslld $24,%mm7
|
| + movl -64(%ebp),%ebx
|
| + pxor %mm1,%mm0
|
| + pxor %mm5,%mm4
|
| + movl (%ebp),%ecx
|
| + pxor %mm3,%mm0
|
| + pxor %mm7,%mm4
|
| + movl 64(%ebp),%edx
|
| + pxor %mm2,%mm0
|
| + pxor %mm6,%mm4
|
| + jmp .L001loop
|
| +.align 16
|
| +.L002out:
|
| + pxor (%edi),%mm0
|
| + pxor 8(%edi),%mm4
|
| + ret
|
| +.size _sse_AES_encrypt_compact,.-_sse_AES_encrypt_compact
|
| +.hidden _x86_AES_encrypt
|
| +.type _x86_AES_encrypt,@function
|
| +.align 16
|
| +_x86_AES_encrypt:
|
| + movl %edi,20(%esp)
|
| + xorl (%edi),%eax
|
| + xorl 4(%edi),%ebx
|
| + xorl 8(%edi),%ecx
|
| + xorl 12(%edi),%edx
|
| + movl 240(%edi),%esi
|
| + leal -2(%esi,%esi,1),%esi
|
| + leal (%edi,%esi,8),%esi
|
| + movl %esi,24(%esp)
|
| +.align 16
|
| +.L003loop:
|
| + movl %eax,%esi
|
| + andl $255,%esi
|
| + movl (%ebp,%esi,8),%esi
|
| + movzbl %bh,%edi
|
| + xorl 3(%ebp,%edi,8),%esi
|
| + movl %ecx,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edi
|
| + xorl 2(%ebp,%edi,8),%esi
|
| + movl %edx,%edi
|
| + shrl $24,%edi
|
| + xorl 1(%ebp,%edi,8),%esi
|
| + movl %esi,4(%esp)
|
| +
|
| + movl %ebx,%esi
|
| + andl $255,%esi
|
| + shrl $16,%ebx
|
| + movl (%ebp,%esi,8),%esi
|
| + movzbl %ch,%edi
|
| + xorl 3(%ebp,%edi,8),%esi
|
| + movl %edx,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edi
|
| + xorl 2(%ebp,%edi,8),%esi
|
| + movl %eax,%edi
|
| + shrl $24,%edi
|
| + xorl 1(%ebp,%edi,8),%esi
|
| + movl %esi,8(%esp)
|
| +
|
| + movl %ecx,%esi
|
| + andl $255,%esi
|
| + shrl $24,%ecx
|
| + movl (%ebp,%esi,8),%esi
|
| + movzbl %dh,%edi
|
| + xorl 3(%ebp,%edi,8),%esi
|
| + movl %eax,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edx
|
| + andl $255,%edi
|
| + xorl 2(%ebp,%edi,8),%esi
|
| + movzbl %bh,%edi
|
| + xorl 1(%ebp,%edi,8),%esi
|
| +
|
| + movl 20(%esp),%edi
|
| + movl (%ebp,%edx,8),%edx
|
| + movzbl %ah,%eax
|
| + xorl 3(%ebp,%eax,8),%edx
|
| + movl 4(%esp),%eax
|
| + andl $255,%ebx
|
| + xorl 2(%ebp,%ebx,8),%edx
|
| + movl 8(%esp),%ebx
|
| + xorl 1(%ebp,%ecx,8),%edx
|
| + movl %esi,%ecx
|
| +
|
| + addl $16,%edi
|
| + xorl (%edi),%eax
|
| + xorl 4(%edi),%ebx
|
| + xorl 8(%edi),%ecx
|
| + xorl 12(%edi),%edx
|
| + cmpl 24(%esp),%edi
|
| + movl %edi,20(%esp)
|
| + jb .L003loop
|
| + movl %eax,%esi
|
| + andl $255,%esi
|
| + movl 2(%ebp,%esi,8),%esi
|
| + andl $255,%esi
|
| + movzbl %bh,%edi
|
| + movl (%ebp,%edi,8),%edi
|
| + andl $65280,%edi
|
| + xorl %edi,%esi
|
| + movl %ecx,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edi
|
| + movl (%ebp,%edi,8),%edi
|
| + andl $16711680,%edi
|
| + xorl %edi,%esi
|
| + movl %edx,%edi
|
| + shrl $24,%edi
|
| + movl 2(%ebp,%edi,8),%edi
|
| + andl $4278190080,%edi
|
| + xorl %edi,%esi
|
| + movl %esi,4(%esp)
|
| + movl %ebx,%esi
|
| + andl $255,%esi
|
| + shrl $16,%ebx
|
| + movl 2(%ebp,%esi,8),%esi
|
| + andl $255,%esi
|
| + movzbl %ch,%edi
|
| + movl (%ebp,%edi,8),%edi
|
| + andl $65280,%edi
|
| + xorl %edi,%esi
|
| + movl %edx,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edi
|
| + movl (%ebp,%edi,8),%edi
|
| + andl $16711680,%edi
|
| + xorl %edi,%esi
|
| + movl %eax,%edi
|
| + shrl $24,%edi
|
| + movl 2(%ebp,%edi,8),%edi
|
| + andl $4278190080,%edi
|
| + xorl %edi,%esi
|
| + movl %esi,8(%esp)
|
| + movl %ecx,%esi
|
| + andl $255,%esi
|
| + shrl $24,%ecx
|
| + movl 2(%ebp,%esi,8),%esi
|
| + andl $255,%esi
|
| + movzbl %dh,%edi
|
| + movl (%ebp,%edi,8),%edi
|
| + andl $65280,%edi
|
| + xorl %edi,%esi
|
| + movl %eax,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edx
|
| + andl $255,%edi
|
| + movl (%ebp,%edi,8),%edi
|
| + andl $16711680,%edi
|
| + xorl %edi,%esi
|
| + movzbl %bh,%edi
|
| + movl 2(%ebp,%edi,8),%edi
|
| + andl $4278190080,%edi
|
| + xorl %edi,%esi
|
| + movl 20(%esp),%edi
|
| + andl $255,%edx
|
| + movl 2(%ebp,%edx,8),%edx
|
| + andl $255,%edx
|
| + movzbl %ah,%eax
|
| + movl (%ebp,%eax,8),%eax
|
| + andl $65280,%eax
|
| + xorl %eax,%edx
|
| + movl 4(%esp),%eax
|
| + andl $255,%ebx
|
| + movl (%ebp,%ebx,8),%ebx
|
| + andl $16711680,%ebx
|
| + xorl %ebx,%edx
|
| + movl 8(%esp),%ebx
|
| + movl 2(%ebp,%ecx,8),%ecx
|
| + andl $4278190080,%ecx
|
| + xorl %ecx,%edx
|
| + movl %esi,%ecx
|
| + addl $16,%edi
|
| + xorl (%edi),%eax
|
| + xorl 4(%edi),%ebx
|
| + xorl 8(%edi),%ecx
|
| + xorl 12(%edi),%edx
|
| + ret
|
| +.align 64
|
| +.LAES_Te:
|
| +.long 2774754246,2774754246
|
| +.long 2222750968,2222750968
|
| +.long 2574743534,2574743534
|
| +.long 2373680118,2373680118
|
| +.long 234025727,234025727
|
| +.long 3177933782,3177933782
|
| +.long 2976870366,2976870366
|
| +.long 1422247313,1422247313
|
| +.long 1345335392,1345335392
|
| +.long 50397442,50397442
|
| +.long 2842126286,2842126286
|
| +.long 2099981142,2099981142
|
| +.long 436141799,436141799
|
| +.long 1658312629,1658312629
|
| +.long 3870010189,3870010189
|
| +.long 2591454956,2591454956
|
| +.long 1170918031,1170918031
|
| +.long 2642575903,2642575903
|
| +.long 1086966153,1086966153
|
| +.long 2273148410,2273148410
|
| +.long 368769775,368769775
|
| +.long 3948501426,3948501426
|
| +.long 3376891790,3376891790
|
| +.long 200339707,200339707
|
| +.long 3970805057,3970805057
|
| +.long 1742001331,1742001331
|
| +.long 4255294047,4255294047
|
| +.long 3937382213,3937382213
|
| +.long 3214711843,3214711843
|
| +.long 4154762323,4154762323
|
| +.long 2524082916,2524082916
|
| +.long 1539358875,1539358875
|
| +.long 3266819957,3266819957
|
| +.long 486407649,486407649
|
| +.long 2928907069,2928907069
|
| +.long 1780885068,1780885068
|
| +.long 1513502316,1513502316
|
| +.long 1094664062,1094664062
|
| +.long 49805301,49805301
|
| +.long 1338821763,1338821763
|
| +.long 1546925160,1546925160
|
| +.long 4104496465,4104496465
|
| +.long 887481809,887481809
|
| +.long 150073849,150073849
|
| +.long 2473685474,2473685474
|
| +.long 1943591083,1943591083
|
| +.long 1395732834,1395732834
|
| +.long 1058346282,1058346282
|
| +.long 201589768,201589768
|
| +.long 1388824469,1388824469
|
| +.long 1696801606,1696801606
|
| +.long 1589887901,1589887901
|
| +.long 672667696,672667696
|
| +.long 2711000631,2711000631
|
| +.long 251987210,251987210
|
| +.long 3046808111,3046808111
|
| +.long 151455502,151455502
|
| +.long 907153956,907153956
|
| +.long 2608889883,2608889883
|
| +.long 1038279391,1038279391
|
| +.long 652995533,652995533
|
| +.long 1764173646,1764173646
|
| +.long 3451040383,3451040383
|
| +.long 2675275242,2675275242
|
| +.long 453576978,453576978
|
| +.long 2659418909,2659418909
|
| +.long 1949051992,1949051992
|
| +.long 773462580,773462580
|
| +.long 756751158,756751158
|
| +.long 2993581788,2993581788
|
| +.long 3998898868,3998898868
|
| +.long 4221608027,4221608027
|
| +.long 4132590244,4132590244
|
| +.long 1295727478,1295727478
|
| +.long 1641469623,1641469623
|
| +.long 3467883389,3467883389
|
| +.long 2066295122,2066295122
|
| +.long 1055122397,1055122397
|
| +.long 1898917726,1898917726
|
| +.long 2542044179,2542044179
|
| +.long 4115878822,4115878822
|
| +.long 1758581177,1758581177
|
| +.long 0,0
|
| +.long 753790401,753790401
|
| +.long 1612718144,1612718144
|
| +.long 536673507,536673507
|
| +.long 3367088505,3367088505
|
| +.long 3982187446,3982187446
|
| +.long 3194645204,3194645204
|
| +.long 1187761037,1187761037
|
| +.long 3653156455,3653156455
|
| +.long 1262041458,1262041458
|
| +.long 3729410708,3729410708
|
| +.long 3561770136,3561770136
|
| +.long 3898103984,3898103984
|
| +.long 1255133061,1255133061
|
| +.long 1808847035,1808847035
|
| +.long 720367557,720367557
|
| +.long 3853167183,3853167183
|
| +.long 385612781,385612781
|
| +.long 3309519750,3309519750
|
| +.long 3612167578,3612167578
|
| +.long 1429418854,1429418854
|
| +.long 2491778321,2491778321
|
| +.long 3477423498,3477423498
|
| +.long 284817897,284817897
|
| +.long 100794884,100794884
|
| +.long 2172616702,2172616702
|
| +.long 4031795360,4031795360
|
| +.long 1144798328,1144798328
|
| +.long 3131023141,3131023141
|
| +.long 3819481163,3819481163
|
| +.long 4082192802,4082192802
|
| +.long 4272137053,4272137053
|
| +.long 3225436288,3225436288
|
| +.long 2324664069,2324664069
|
| +.long 2912064063,2912064063
|
| +.long 3164445985,3164445985
|
| +.long 1211644016,1211644016
|
| +.long 83228145,83228145
|
| +.long 3753688163,3753688163
|
| +.long 3249976951,3249976951
|
| +.long 1977277103,1977277103
|
| +.long 1663115586,1663115586
|
| +.long 806359072,806359072
|
| +.long 452984805,452984805
|
| +.long 250868733,250868733
|
| +.long 1842533055,1842533055
|
| +.long 1288555905,1288555905
|
| +.long 336333848,336333848
|
| +.long 890442534,890442534
|
| +.long 804056259,804056259
|
| +.long 3781124030,3781124030
|
| +.long 2727843637,2727843637
|
| +.long 3427026056,3427026056
|
| +.long 957814574,957814574
|
| +.long 1472513171,1472513171
|
| +.long 4071073621,4071073621
|
| +.long 2189328124,2189328124
|
| +.long 1195195770,1195195770
|
| +.long 2892260552,2892260552
|
| +.long 3881655738,3881655738
|
| +.long 723065138,723065138
|
| +.long 2507371494,2507371494
|
| +.long 2690670784,2690670784
|
| +.long 2558624025,2558624025
|
| +.long 3511635870,3511635870
|
| +.long 2145180835,2145180835
|
| +.long 1713513028,1713513028
|
| +.long 2116692564,2116692564
|
| +.long 2878378043,2878378043
|
| +.long 2206763019,2206763019
|
| +.long 3393603212,3393603212
|
| +.long 703524551,703524551
|
| +.long 3552098411,3552098411
|
| +.long 1007948840,1007948840
|
| +.long 2044649127,2044649127
|
| +.long 3797835452,3797835452
|
| +.long 487262998,487262998
|
| +.long 1994120109,1994120109
|
| +.long 1004593371,1004593371
|
| +.long 1446130276,1446130276
|
| +.long 1312438900,1312438900
|
| +.long 503974420,503974420
|
| +.long 3679013266,3679013266
|
| +.long 168166924,168166924
|
| +.long 1814307912,1814307912
|
| +.long 3831258296,3831258296
|
| +.long 1573044895,1573044895
|
| +.long 1859376061,1859376061
|
| +.long 4021070915,4021070915
|
| +.long 2791465668,2791465668
|
| +.long 2828112185,2828112185
|
| +.long 2761266481,2761266481
|
| +.long 937747667,937747667
|
| +.long 2339994098,2339994098
|
| +.long 854058965,854058965
|
| +.long 1137232011,1137232011
|
| +.long 1496790894,1496790894
|
| +.long 3077402074,3077402074
|
| +.long 2358086913,2358086913
|
| +.long 1691735473,1691735473
|
| +.long 3528347292,3528347292
|
| +.long 3769215305,3769215305
|
| +.long 3027004632,3027004632
|
| +.long 4199962284,4199962284
|
| +.long 133494003,133494003
|
| +.long 636152527,636152527
|
| +.long 2942657994,2942657994
|
| +.long 2390391540,2390391540
|
| +.long 3920539207,3920539207
|
| +.long 403179536,403179536
|
| +.long 3585784431,3585784431
|
| +.long 2289596656,2289596656
|
| +.long 1864705354,1864705354
|
| +.long 1915629148,1915629148
|
| +.long 605822008,605822008
|
| +.long 4054230615,4054230615
|
| +.long 3350508659,3350508659
|
| +.long 1371981463,1371981463
|
| +.long 602466507,602466507
|
| +.long 2094914977,2094914977
|
| +.long 2624877800,2624877800
|
| +.long 555687742,555687742
|
| +.long 3712699286,3712699286
|
| +.long 3703422305,3703422305
|
| +.long 2257292045,2257292045
|
| +.long 2240449039,2240449039
|
| +.long 2423288032,2423288032
|
| +.long 1111375484,1111375484
|
| +.long 3300242801,3300242801
|
| +.long 2858837708,2858837708
|
| +.long 3628615824,3628615824
|
| +.long 84083462,84083462
|
| +.long 32962295,32962295
|
| +.long 302911004,302911004
|
| +.long 2741068226,2741068226
|
| +.long 1597322602,1597322602
|
| +.long 4183250862,4183250862
|
| +.long 3501832553,3501832553
|
| +.long 2441512471,2441512471
|
| +.long 1489093017,1489093017
|
| +.long 656219450,656219450
|
| +.long 3114180135,3114180135
|
| +.long 954327513,954327513
|
| +.long 335083755,335083755
|
| +.long 3013122091,3013122091
|
| +.long 856756514,856756514
|
| +.long 3144247762,3144247762
|
| +.long 1893325225,1893325225
|
| +.long 2307821063,2307821063
|
| +.long 2811532339,2811532339
|
| +.long 3063651117,3063651117
|
| +.long 572399164,572399164
|
| +.long 2458355477,2458355477
|
| +.long 552200649,552200649
|
| +.long 1238290055,1238290055
|
| +.long 4283782570,4283782570
|
| +.long 2015897680,2015897680
|
| +.long 2061492133,2061492133
|
| +.long 2408352771,2408352771
|
| +.long 4171342169,4171342169
|
| +.long 2156497161,2156497161
|
| +.long 386731290,386731290
|
| +.long 3669999461,3669999461
|
| +.long 837215959,837215959
|
| +.long 3326231172,3326231172
|
| +.long 3093850320,3093850320
|
| +.long 3275833730,3275833730
|
| +.long 2962856233,2962856233
|
| +.long 1999449434,1999449434
|
| +.long 286199582,286199582
|
| +.long 3417354363,3417354363
|
| +.long 4233385128,4233385128
|
| +.long 3602627437,3602627437
|
| +.long 974525996,974525996
|
| +.byte 99,124,119,123,242,107,111,197
|
| +.byte 48,1,103,43,254,215,171,118
|
| +.byte 202,130,201,125,250,89,71,240
|
| +.byte 173,212,162,175,156,164,114,192
|
| +.byte 183,253,147,38,54,63,247,204
|
| +.byte 52,165,229,241,113,216,49,21
|
| +.byte 4,199,35,195,24,150,5,154
|
| +.byte 7,18,128,226,235,39,178,117
|
| +.byte 9,131,44,26,27,110,90,160
|
| +.byte 82,59,214,179,41,227,47,132
|
| +.byte 83,209,0,237,32,252,177,91
|
| +.byte 106,203,190,57,74,76,88,207
|
| +.byte 208,239,170,251,67,77,51,133
|
| +.byte 69,249,2,127,80,60,159,168
|
| +.byte 81,163,64,143,146,157,56,245
|
| +.byte 188,182,218,33,16,255,243,210
|
| +.byte 205,12,19,236,95,151,68,23
|
| +.byte 196,167,126,61,100,93,25,115
|
| +.byte 96,129,79,220,34,42,144,136
|
| +.byte 70,238,184,20,222,94,11,219
|
| +.byte 224,50,58,10,73,6,36,92
|
| +.byte 194,211,172,98,145,149,228,121
|
| +.byte 231,200,55,109,141,213,78,169
|
| +.byte 108,86,244,234,101,122,174,8
|
| +.byte 186,120,37,46,28,166,180,198
|
| +.byte 232,221,116,31,75,189,139,138
|
| +.byte 112,62,181,102,72,3,246,14
|
| +.byte 97,53,87,185,134,193,29,158
|
| +.byte 225,248,152,17,105,217,142,148
|
| +.byte 155,30,135,233,206,85,40,223
|
| +.byte 140,161,137,13,191,230,66,104
|
| +.byte 65,153,45,15,176,84,187,22
|
| +.byte 99,124,119,123,242,107,111,197
|
| +.byte 48,1,103,43,254,215,171,118
|
| +.byte 202,130,201,125,250,89,71,240
|
| +.byte 173,212,162,175,156,164,114,192
|
| +.byte 183,253,147,38,54,63,247,204
|
| +.byte 52,165,229,241,113,216,49,21
|
| +.byte 4,199,35,195,24,150,5,154
|
| +.byte 7,18,128,226,235,39,178,117
|
| +.byte 9,131,44,26,27,110,90,160
|
| +.byte 82,59,214,179,41,227,47,132
|
| +.byte 83,209,0,237,32,252,177,91
|
| +.byte 106,203,190,57,74,76,88,207
|
| +.byte 208,239,170,251,67,77,51,133
|
| +.byte 69,249,2,127,80,60,159,168
|
| +.byte 81,163,64,143,146,157,56,245
|
| +.byte 188,182,218,33,16,255,243,210
|
| +.byte 205,12,19,236,95,151,68,23
|
| +.byte 196,167,126,61,100,93,25,115
|
| +.byte 96,129,79,220,34,42,144,136
|
| +.byte 70,238,184,20,222,94,11,219
|
| +.byte 224,50,58,10,73,6,36,92
|
| +.byte 194,211,172,98,145,149,228,121
|
| +.byte 231,200,55,109,141,213,78,169
|
| +.byte 108,86,244,234,101,122,174,8
|
| +.byte 186,120,37,46,28,166,180,198
|
| +.byte 232,221,116,31,75,189,139,138
|
| +.byte 112,62,181,102,72,3,246,14
|
| +.byte 97,53,87,185,134,193,29,158
|
| +.byte 225,248,152,17,105,217,142,148
|
| +.byte 155,30,135,233,206,85,40,223
|
| +.byte 140,161,137,13,191,230,66,104
|
| +.byte 65,153,45,15,176,84,187,22
|
| +.byte 99,124,119,123,242,107,111,197
|
| +.byte 48,1,103,43,254,215,171,118
|
| +.byte 202,130,201,125,250,89,71,240
|
| +.byte 173,212,162,175,156,164,114,192
|
| +.byte 183,253,147,38,54,63,247,204
|
| +.byte 52,165,229,241,113,216,49,21
|
| +.byte 4,199,35,195,24,150,5,154
|
| +.byte 7,18,128,226,235,39,178,117
|
| +.byte 9,131,44,26,27,110,90,160
|
| +.byte 82,59,214,179,41,227,47,132
|
| +.byte 83,209,0,237,32,252,177,91
|
| +.byte 106,203,190,57,74,76,88,207
|
| +.byte 208,239,170,251,67,77,51,133
|
| +.byte 69,249,2,127,80,60,159,168
|
| +.byte 81,163,64,143,146,157,56,245
|
| +.byte 188,182,218,33,16,255,243,210
|
| +.byte 205,12,19,236,95,151,68,23
|
| +.byte 196,167,126,61,100,93,25,115
|
| +.byte 96,129,79,220,34,42,144,136
|
| +.byte 70,238,184,20,222,94,11,219
|
| +.byte 224,50,58,10,73,6,36,92
|
| +.byte 194,211,172,98,145,149,228,121
|
| +.byte 231,200,55,109,141,213,78,169
|
| +.byte 108,86,244,234,101,122,174,8
|
| +.byte 186,120,37,46,28,166,180,198
|
| +.byte 232,221,116,31,75,189,139,138
|
| +.byte 112,62,181,102,72,3,246,14
|
| +.byte 97,53,87,185,134,193,29,158
|
| +.byte 225,248,152,17,105,217,142,148
|
| +.byte 155,30,135,233,206,85,40,223
|
| +.byte 140,161,137,13,191,230,66,104
|
| +.byte 65,153,45,15,176,84,187,22
|
| +.byte 99,124,119,123,242,107,111,197
|
| +.byte 48,1,103,43,254,215,171,118
|
| +.byte 202,130,201,125,250,89,71,240
|
| +.byte 173,212,162,175,156,164,114,192
|
| +.byte 183,253,147,38,54,63,247,204
|
| +.byte 52,165,229,241,113,216,49,21
|
| +.byte 4,199,35,195,24,150,5,154
|
| +.byte 7,18,128,226,235,39,178,117
|
| +.byte 9,131,44,26,27,110,90,160
|
| +.byte 82,59,214,179,41,227,47,132
|
| +.byte 83,209,0,237,32,252,177,91
|
| +.byte 106,203,190,57,74,76,88,207
|
| +.byte 208,239,170,251,67,77,51,133
|
| +.byte 69,249,2,127,80,60,159,168
|
| +.byte 81,163,64,143,146,157,56,245
|
| +.byte 188,182,218,33,16,255,243,210
|
| +.byte 205,12,19,236,95,151,68,23
|
| +.byte 196,167,126,61,100,93,25,115
|
| +.byte 96,129,79,220,34,42,144,136
|
| +.byte 70,238,184,20,222,94,11,219
|
| +.byte 224,50,58,10,73,6,36,92
|
| +.byte 194,211,172,98,145,149,228,121
|
| +.byte 231,200,55,109,141,213,78,169
|
| +.byte 108,86,244,234,101,122,174,8
|
| +.byte 186,120,37,46,28,166,180,198
|
| +.byte 232,221,116,31,75,189,139,138
|
| +.byte 112,62,181,102,72,3,246,14
|
| +.byte 97,53,87,185,134,193,29,158
|
| +.byte 225,248,152,17,105,217,142,148
|
| +.byte 155,30,135,233,206,85,40,223
|
| +.byte 140,161,137,13,191,230,66,104
|
| +.byte 65,153,45,15,176,84,187,22
|
| +.long 1,2,4,8
|
| +.long 16,32,64,128
|
| +.long 27,54,0,0
|
| +.long 0,0,0,0
|
| +.size _x86_AES_encrypt,.-_x86_AES_encrypt
|
| +.globl asm_AES_encrypt
|
| +.hidden asm_AES_encrypt
|
| +.type asm_AES_encrypt,@function
|
| +.align 16
|
| +asm_AES_encrypt:
|
| +.L_asm_AES_encrypt_begin:
|
| + pushl %ebp
|
| + pushl %ebx
|
| + pushl %esi
|
| + pushl %edi
|
| + movl 20(%esp),%esi
|
| + movl 28(%esp),%edi
|
| + movl %esp,%eax
|
| + subl $36,%esp
|
| + andl $-64,%esp
|
| + leal -127(%edi),%ebx
|
| + subl %esp,%ebx
|
| + negl %ebx
|
| + andl $960,%ebx
|
| + subl %ebx,%esp
|
| + addl $4,%esp
|
| + movl %eax,28(%esp)
|
| + call .L004pic_point
|
| +.L004pic_point:
|
| + popl %ebp
|
| + leal OPENSSL_ia32cap_P-.L004pic_point(%ebp),%eax
|
| + leal .LAES_Te-.L004pic_point(%ebp),%ebp
|
| + leal 764(%esp),%ebx
|
| + subl %ebp,%ebx
|
| + andl $768,%ebx
|
| + leal 2176(%ebp,%ebx,1),%ebp
|
| + btl $25,(%eax)
|
| + jnc .L005x86
|
| + movq (%esi),%mm0
|
| + movq 8(%esi),%mm4
|
| + call _sse_AES_encrypt_compact
|
| + movl 28(%esp),%esp
|
| + movl 24(%esp),%esi
|
| + movq %mm0,(%esi)
|
| + movq %mm4,8(%esi)
|
| + emms
|
| + popl %edi
|
| + popl %esi
|
| + popl %ebx
|
| + popl %ebp
|
| + ret
|
| +.align 16
|
| +.L005x86:
|
| + movl %ebp,24(%esp)
|
| + movl (%esi),%eax
|
| + movl 4(%esi),%ebx
|
| + movl 8(%esi),%ecx
|
| + movl 12(%esi),%edx
|
| + call _x86_AES_encrypt_compact
|
| + movl 28(%esp),%esp
|
| + movl 24(%esp),%esi
|
| + movl %eax,(%esi)
|
| + movl %ebx,4(%esi)
|
| + movl %ecx,8(%esi)
|
| + movl %edx,12(%esi)
|
| + popl %edi
|
| + popl %esi
|
| + popl %ebx
|
| + popl %ebp
|
| + ret
|
| +.size asm_AES_encrypt,.-.L_asm_AES_encrypt_begin
|
| +.hidden _x86_AES_decrypt_compact
|
| +.type _x86_AES_decrypt_compact,@function
|
| +.align 16
|
| +_x86_AES_decrypt_compact:
|
| + movl %edi,20(%esp)
|
| + xorl (%edi),%eax
|
| + xorl 4(%edi),%ebx
|
| + xorl 8(%edi),%ecx
|
| + xorl 12(%edi),%edx
|
| + movl 240(%edi),%esi
|
| + leal -2(%esi,%esi,1),%esi
|
| + leal (%edi,%esi,8),%esi
|
| + movl %esi,24(%esp)
|
| + movl -128(%ebp),%edi
|
| + movl -96(%ebp),%esi
|
| + movl -64(%ebp),%edi
|
| + movl -32(%ebp),%esi
|
| + movl (%ebp),%edi
|
| + movl 32(%ebp),%esi
|
| + movl 64(%ebp),%edi
|
| + movl 96(%ebp),%esi
|
| +.align 16
|
| +.L006loop:
|
| + movl %eax,%esi
|
| + andl $255,%esi
|
| + movzbl -128(%ebp,%esi,1),%esi
|
| + movzbl %dh,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $8,%edi
|
| + xorl %edi,%esi
|
| + movl %ecx,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $16,%edi
|
| + xorl %edi,%esi
|
| + movl %ebx,%edi
|
| + shrl $24,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $24,%edi
|
| + xorl %edi,%esi
|
| + movl %esi,4(%esp)
|
| + movl %ebx,%esi
|
| + andl $255,%esi
|
| + movzbl -128(%ebp,%esi,1),%esi
|
| + movzbl %ah,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $8,%edi
|
| + xorl %edi,%esi
|
| + movl %edx,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $16,%edi
|
| + xorl %edi,%esi
|
| + movl %ecx,%edi
|
| + shrl $24,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $24,%edi
|
| + xorl %edi,%esi
|
| + movl %esi,8(%esp)
|
| + movl %ecx,%esi
|
| + andl $255,%esi
|
| + movzbl -128(%ebp,%esi,1),%esi
|
| + movzbl %bh,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $8,%edi
|
| + xorl %edi,%esi
|
| + movl %eax,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $16,%edi
|
| + xorl %edi,%esi
|
| + movl %edx,%edi
|
| + shrl $24,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $24,%edi
|
| + xorl %edi,%esi
|
| + andl $255,%edx
|
| + movzbl -128(%ebp,%edx,1),%edx
|
| + movzbl %ch,%ecx
|
| + movzbl -128(%ebp,%ecx,1),%ecx
|
| + shll $8,%ecx
|
| + xorl %ecx,%edx
|
| + movl %esi,%ecx
|
| + shrl $16,%ebx
|
| + andl $255,%ebx
|
| + movzbl -128(%ebp,%ebx,1),%ebx
|
| + shll $16,%ebx
|
| + xorl %ebx,%edx
|
| + shrl $24,%eax
|
| + movzbl -128(%ebp,%eax,1),%eax
|
| + shll $24,%eax
|
| + xorl %eax,%edx
|
| + movl $2155905152,%edi
|
| + andl %ecx,%edi
|
| + movl %edi,%esi
|
| + shrl $7,%edi
|
| + leal (%ecx,%ecx,1),%eax
|
| + subl %edi,%esi
|
| + andl $4278124286,%eax
|
| + andl $454761243,%esi
|
| + xorl %esi,%eax
|
| + movl $2155905152,%edi
|
| + andl %eax,%edi
|
| + movl %edi,%esi
|
| + shrl $7,%edi
|
| + leal (%eax,%eax,1),%ebx
|
| + subl %edi,%esi
|
| + andl $4278124286,%ebx
|
| + andl $454761243,%esi
|
| + xorl %ecx,%eax
|
| + xorl %esi,%ebx
|
| + movl $2155905152,%edi
|
| + andl %ebx,%edi
|
| + movl %edi,%esi
|
| + shrl $7,%edi
|
| + leal (%ebx,%ebx,1),%ebp
|
| + subl %edi,%esi
|
| + andl $4278124286,%ebp
|
| + andl $454761243,%esi
|
| + xorl %ecx,%ebx
|
| + roll $8,%ecx
|
| + xorl %esi,%ebp
|
| + xorl %eax,%ecx
|
| + xorl %ebp,%eax
|
| + xorl %ebx,%ecx
|
| + xorl %ebp,%ebx
|
| + roll $24,%eax
|
| + xorl %ebp,%ecx
|
| + roll $16,%ebx
|
| + xorl %eax,%ecx
|
| + roll $8,%ebp
|
| + xorl %ebx,%ecx
|
| + movl 4(%esp),%eax
|
| + xorl %ebp,%ecx
|
| + movl %ecx,12(%esp)
|
| + movl $2155905152,%edi
|
| + andl %edx,%edi
|
| + movl %edi,%esi
|
| + shrl $7,%edi
|
| + leal (%edx,%edx,1),%ebx
|
| + subl %edi,%esi
|
| + andl $4278124286,%ebx
|
| + andl $454761243,%esi
|
| + xorl %esi,%ebx
|
| + movl $2155905152,%edi
|
| + andl %ebx,%edi
|
| + movl %edi,%esi
|
| + shrl $7,%edi
|
| + leal (%ebx,%ebx,1),%ecx
|
| + subl %edi,%esi
|
| + andl $4278124286,%ecx
|
| + andl $454761243,%esi
|
| + xorl %edx,%ebx
|
| + xorl %esi,%ecx
|
| + movl $2155905152,%edi
|
| + andl %ecx,%edi
|
| + movl %edi,%esi
|
| + shrl $7,%edi
|
| + leal (%ecx,%ecx,1),%ebp
|
| + subl %edi,%esi
|
| + andl $4278124286,%ebp
|
| + andl $454761243,%esi
|
| + xorl %edx,%ecx
|
| + roll $8,%edx
|
| + xorl %esi,%ebp
|
| + xorl %ebx,%edx
|
| + xorl %ebp,%ebx
|
| + xorl %ecx,%edx
|
| + xorl %ebp,%ecx
|
| + roll $24,%ebx
|
| + xorl %ebp,%edx
|
| + roll $16,%ecx
|
| + xorl %ebx,%edx
|
| + roll $8,%ebp
|
| + xorl %ecx,%edx
|
| + movl 8(%esp),%ebx
|
| + xorl %ebp,%edx
|
| + movl %edx,16(%esp)
|
| + movl $2155905152,%edi
|
| + andl %eax,%edi
|
| + movl %edi,%esi
|
| + shrl $7,%edi
|
| + leal (%eax,%eax,1),%ecx
|
| + subl %edi,%esi
|
| + andl $4278124286,%ecx
|
| + andl $454761243,%esi
|
| + xorl %esi,%ecx
|
| + movl $2155905152,%edi
|
| + andl %ecx,%edi
|
| + movl %edi,%esi
|
| + shrl $7,%edi
|
| + leal (%ecx,%ecx,1),%edx
|
| + subl %edi,%esi
|
| + andl $4278124286,%edx
|
| + andl $454761243,%esi
|
| + xorl %eax,%ecx
|
| + xorl %esi,%edx
|
| + movl $2155905152,%edi
|
| + andl %edx,%edi
|
| + movl %edi,%esi
|
| + shrl $7,%edi
|
| + leal (%edx,%edx,1),%ebp
|
| + subl %edi,%esi
|
| + andl $4278124286,%ebp
|
| + andl $454761243,%esi
|
| + xorl %eax,%edx
|
| + roll $8,%eax
|
| + xorl %esi,%ebp
|
| + xorl %ecx,%eax
|
| + xorl %ebp,%ecx
|
| + xorl %edx,%eax
|
| + xorl %ebp,%edx
|
| + roll $24,%ecx
|
| + xorl %ebp,%eax
|
| + roll $16,%edx
|
| + xorl %ecx,%eax
|
| + roll $8,%ebp
|
| + xorl %edx,%eax
|
| + xorl %ebp,%eax
|
| + movl $2155905152,%edi
|
| + andl %ebx,%edi
|
| + movl %edi,%esi
|
| + shrl $7,%edi
|
| + leal (%ebx,%ebx,1),%ecx
|
| + subl %edi,%esi
|
| + andl $4278124286,%ecx
|
| + andl $454761243,%esi
|
| + xorl %esi,%ecx
|
| + movl $2155905152,%edi
|
| + andl %ecx,%edi
|
| + movl %edi,%esi
|
| + shrl $7,%edi
|
| + leal (%ecx,%ecx,1),%edx
|
| + subl %edi,%esi
|
| + andl $4278124286,%edx
|
| + andl $454761243,%esi
|
| + xorl %ebx,%ecx
|
| + xorl %esi,%edx
|
| + movl $2155905152,%edi
|
| + andl %edx,%edi
|
| + movl %edi,%esi
|
| + shrl $7,%edi
|
| + leal (%edx,%edx,1),%ebp
|
| + subl %edi,%esi
|
| + andl $4278124286,%ebp
|
| + andl $454761243,%esi
|
| + xorl %ebx,%edx
|
| + roll $8,%ebx
|
| + xorl %esi,%ebp
|
| + xorl %ecx,%ebx
|
| + xorl %ebp,%ecx
|
| + xorl %edx,%ebx
|
| + xorl %ebp,%edx
|
| + roll $24,%ecx
|
| + xorl %ebp,%ebx
|
| + roll $16,%edx
|
| + xorl %ecx,%ebx
|
| + roll $8,%ebp
|
| + xorl %edx,%ebx
|
| + movl 12(%esp),%ecx
|
| + xorl %ebp,%ebx
|
| + movl 16(%esp),%edx
|
| + movl 20(%esp),%edi
|
| + movl 28(%esp),%ebp
|
| + addl $16,%edi
|
| + xorl (%edi),%eax
|
| + xorl 4(%edi),%ebx
|
| + xorl 8(%edi),%ecx
|
| + xorl 12(%edi),%edx
|
| + cmpl 24(%esp),%edi
|
| + movl %edi,20(%esp)
|
| + jb .L006loop
|
| + movl %eax,%esi
|
| + andl $255,%esi
|
| + movzbl -128(%ebp,%esi,1),%esi
|
| + movzbl %dh,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $8,%edi
|
| + xorl %edi,%esi
|
| + movl %ecx,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $16,%edi
|
| + xorl %edi,%esi
|
| + movl %ebx,%edi
|
| + shrl $24,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $24,%edi
|
| + xorl %edi,%esi
|
| + movl %esi,4(%esp)
|
| + movl %ebx,%esi
|
| + andl $255,%esi
|
| + movzbl -128(%ebp,%esi,1),%esi
|
| + movzbl %ah,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $8,%edi
|
| + xorl %edi,%esi
|
| + movl %edx,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $16,%edi
|
| + xorl %edi,%esi
|
| + movl %ecx,%edi
|
| + shrl $24,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $24,%edi
|
| + xorl %edi,%esi
|
| + movl %esi,8(%esp)
|
| + movl %ecx,%esi
|
| + andl $255,%esi
|
| + movzbl -128(%ebp,%esi,1),%esi
|
| + movzbl %bh,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $8,%edi
|
| + xorl %edi,%esi
|
| + movl %eax,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $16,%edi
|
| + xorl %edi,%esi
|
| + movl %edx,%edi
|
| + shrl $24,%edi
|
| + movzbl -128(%ebp,%edi,1),%edi
|
| + shll $24,%edi
|
| + xorl %edi,%esi
|
| + movl 20(%esp),%edi
|
| + andl $255,%edx
|
| + movzbl -128(%ebp,%edx,1),%edx
|
| + movzbl %ch,%ecx
|
| + movzbl -128(%ebp,%ecx,1),%ecx
|
| + shll $8,%ecx
|
| + xorl %ecx,%edx
|
| + movl %esi,%ecx
|
| + shrl $16,%ebx
|
| + andl $255,%ebx
|
| + movzbl -128(%ebp,%ebx,1),%ebx
|
| + shll $16,%ebx
|
| + xorl %ebx,%edx
|
| + movl 8(%esp),%ebx
|
| + shrl $24,%eax
|
| + movzbl -128(%ebp,%eax,1),%eax
|
| + shll $24,%eax
|
| + xorl %eax,%edx
|
| + movl 4(%esp),%eax
|
| + xorl 16(%edi),%eax
|
| + xorl 20(%edi),%ebx
|
| + xorl 24(%edi),%ecx
|
| + xorl 28(%edi),%edx
|
| + ret
|
| +.size _x86_AES_decrypt_compact,.-_x86_AES_decrypt_compact
|
| +.hidden _sse_AES_decrypt_compact
|
| +.type _sse_AES_decrypt_compact,@function
|
| +.align 16
|
| +_sse_AES_decrypt_compact:
|
| + pxor (%edi),%mm0
|
| + pxor 8(%edi),%mm4
|
| + movl 240(%edi),%esi
|
| + leal -2(%esi,%esi,1),%esi
|
| + leal (%edi,%esi,8),%esi
|
| + movl %esi,24(%esp)
|
| + movl $454761243,%eax
|
| + movl %eax,8(%esp)
|
| + movl %eax,12(%esp)
|
| + movl -128(%ebp),%eax
|
| + movl -96(%ebp),%ebx
|
| + movl -64(%ebp),%ecx
|
| + movl -32(%ebp),%edx
|
| + movl (%ebp),%eax
|
| + movl 32(%ebp),%ebx
|
| + movl 64(%ebp),%ecx
|
| + movl 96(%ebp),%edx
|
| +.align 16
|
| +.L007loop:
|
| + pshufw $12,%mm0,%mm1
|
| + pshufw $9,%mm4,%mm5
|
| + movd %mm1,%eax
|
| + movd %mm5,%ebx
|
| + movl %edi,20(%esp)
|
| + movzbl %al,%esi
|
| + movzbl %ah,%edx
|
| + pshufw $6,%mm0,%mm2
|
| + movzbl -128(%ebp,%esi,1),%ecx
|
| + movzbl %bl,%edi
|
| + movzbl -128(%ebp,%edx,1),%edx
|
| + shrl $16,%eax
|
| + shll $8,%edx
|
| + movzbl -128(%ebp,%edi,1),%esi
|
| + movzbl %bh,%edi
|
| + shll $16,%esi
|
| + pshufw $3,%mm4,%mm6
|
| + orl %esi,%ecx
|
| + movzbl -128(%ebp,%edi,1),%esi
|
| + movzbl %ah,%edi
|
| + shll $24,%esi
|
| + shrl $16,%ebx
|
| + orl %esi,%edx
|
| + movzbl -128(%ebp,%edi,1),%esi
|
| + movzbl %bh,%edi
|
| + shll $24,%esi
|
| + orl %esi,%ecx
|
| + movzbl -128(%ebp,%edi,1),%esi
|
| + movzbl %al,%edi
|
| + shll $8,%esi
|
| + movd %mm2,%eax
|
| + orl %esi,%ecx
|
| + movzbl -128(%ebp,%edi,1),%esi
|
| + movzbl %bl,%edi
|
| + shll $16,%esi
|
| + movd %mm6,%ebx
|
| + movd %ecx,%mm0
|
| + movzbl -128(%ebp,%edi,1),%ecx
|
| + movzbl %al,%edi
|
| + orl %esi,%ecx
|
| + movzbl -128(%ebp,%edi,1),%esi
|
| + movzbl %bl,%edi
|
| + orl %esi,%edx
|
| + movzbl -128(%ebp,%edi,1),%esi
|
| + movzbl %ah,%edi
|
| + shll $16,%esi
|
| + shrl $16,%eax
|
| + orl %esi,%edx
|
| + movzbl -128(%ebp,%edi,1),%esi
|
| + movzbl %bh,%edi
|
| + shrl $16,%ebx
|
| + shll $8,%esi
|
| + movd %edx,%mm1
|
| + movzbl -128(%ebp,%edi,1),%edx
|
| + movzbl %bh,%edi
|
| + shll $24,%edx
|
| + andl $255,%ebx
|
| + orl %esi,%edx
|
| + punpckldq %mm1,%mm0
|
| + movzbl -128(%ebp,%edi,1),%esi
|
| + movzbl %al,%edi
|
| + shll $8,%esi
|
| + movzbl %ah,%eax
|
| + movzbl -128(%ebp,%ebx,1),%ebx
|
| + orl %esi,%ecx
|
| + movzbl -128(%ebp,%edi,1),%esi
|
| + orl %ebx,%edx
|
| + shll $16,%esi
|
| + movzbl -128(%ebp,%eax,1),%eax
|
| + orl %esi,%edx
|
| + shll $24,%eax
|
| + orl %eax,%ecx
|
| + movl 20(%esp),%edi
|
| + movd %edx,%mm4
|
| + movd %ecx,%mm5
|
| + punpckldq %mm5,%mm4
|
| + addl $16,%edi
|
| + cmpl 24(%esp),%edi
|
| + ja .L008out
|
| + movq %mm0,%mm3
|
| + movq %mm4,%mm7
|
| + pshufw $228,%mm0,%mm2
|
| + pshufw $228,%mm4,%mm6
|
| + movq %mm0,%mm1
|
| + movq %mm4,%mm5
|
| + pshufw $177,%mm0,%mm0
|
| + pshufw $177,%mm4,%mm4
|
| + pslld $8,%mm2
|
| + pslld $8,%mm6
|
| + psrld $8,%mm3
|
| + psrld $8,%mm7
|
| + pxor %mm2,%mm0
|
| + pxor %mm6,%mm4
|
| + pxor %mm3,%mm0
|
| + pxor %mm7,%mm4
|
| + pslld $16,%mm2
|
| + pslld $16,%mm6
|
| + psrld $16,%mm3
|
| + psrld $16,%mm7
|
| + pxor %mm2,%mm0
|
| + pxor %mm6,%mm4
|
| + pxor %mm3,%mm0
|
| + pxor %mm7,%mm4
|
| + movq 8(%esp),%mm3
|
| + pxor %mm2,%mm2
|
| + pxor %mm6,%mm6
|
| + pcmpgtb %mm1,%mm2
|
| + pcmpgtb %mm5,%mm6
|
| + pand %mm3,%mm2
|
| + pand %mm3,%mm6
|
| + paddb %mm1,%mm1
|
| + paddb %mm5,%mm5
|
| + pxor %mm2,%mm1
|
| + pxor %mm6,%mm5
|
| + movq %mm1,%mm3
|
| + movq %mm5,%mm7
|
| + movq %mm1,%mm2
|
| + movq %mm5,%mm6
|
| + pxor %mm1,%mm0
|
| + pxor %mm5,%mm4
|
| + pslld $24,%mm3
|
| + pslld $24,%mm7
|
| + psrld $8,%mm2
|
| + psrld $8,%mm6
|
| + pxor %mm3,%mm0
|
| + pxor %mm7,%mm4
|
| + pxor %mm2,%mm0
|
| + pxor %mm6,%mm4
|
| + movq 8(%esp),%mm2
|
| + pxor %mm3,%mm3
|
| + pxor %mm7,%mm7
|
| + pcmpgtb %mm1,%mm3
|
| + pcmpgtb %mm5,%mm7
|
| + pand %mm2,%mm3
|
| + pand %mm2,%mm7
|
| + paddb %mm1,%mm1
|
| + paddb %mm5,%mm5
|
| + pxor %mm3,%mm1
|
| + pxor %mm7,%mm5
|
| + pshufw $177,%mm1,%mm3
|
| + pshufw $177,%mm5,%mm7
|
| + pxor %mm1,%mm0
|
| + pxor %mm5,%mm4
|
| + pxor %mm3,%mm0
|
| + pxor %mm7,%mm4
|
| + pxor %mm3,%mm3
|
| + pxor %mm7,%mm7
|
| + pcmpgtb %mm1,%mm3
|
| + pcmpgtb %mm5,%mm7
|
| + pand %mm2,%mm3
|
| + pand %mm2,%mm7
|
| + paddb %mm1,%mm1
|
| + paddb %mm5,%mm5
|
| + pxor %mm3,%mm1
|
| + pxor %mm7,%mm5
|
| + pxor %mm1,%mm0
|
| + pxor %mm5,%mm4
|
| + movq %mm1,%mm3
|
| + movq %mm5,%mm7
|
| + pshufw $177,%mm1,%mm2
|
| + pshufw $177,%mm5,%mm6
|
| + pxor %mm2,%mm0
|
| + pxor %mm6,%mm4
|
| + pslld $8,%mm1
|
| + pslld $8,%mm5
|
| + psrld $8,%mm3
|
| + psrld $8,%mm7
|
| + movq (%edi),%mm2
|
| + movq 8(%edi),%mm6
|
| + pxor %mm1,%mm0
|
| + pxor %mm5,%mm4
|
| + pxor %mm3,%mm0
|
| + pxor %mm7,%mm4
|
| + movl -128(%ebp),%eax
|
| + pslld $16,%mm1
|
| + pslld $16,%mm5
|
| + movl -64(%ebp),%ebx
|
| + psrld $16,%mm3
|
| + psrld $16,%mm7
|
| + movl (%ebp),%ecx
|
| + pxor %mm1,%mm0
|
| + pxor %mm5,%mm4
|
| + movl 64(%ebp),%edx
|
| + pxor %mm3,%mm0
|
| + pxor %mm7,%mm4
|
| + pxor %mm2,%mm0
|
| + pxor %mm6,%mm4
|
| + jmp .L007loop
|
| +.align 16
|
| +.L008out:
|
| + pxor (%edi),%mm0
|
| + pxor 8(%edi),%mm4
|
| + ret
|
| +.size _sse_AES_decrypt_compact,.-_sse_AES_decrypt_compact
|
| +.hidden _x86_AES_decrypt
|
| +.type _x86_AES_decrypt,@function
|
| +.align 16
|
| +_x86_AES_decrypt:
|
| + movl %edi,20(%esp)
|
| + xorl (%edi),%eax
|
| + xorl 4(%edi),%ebx
|
| + xorl 8(%edi),%ecx
|
| + xorl 12(%edi),%edx
|
| + movl 240(%edi),%esi
|
| + leal -2(%esi,%esi,1),%esi
|
| + leal (%edi,%esi,8),%esi
|
| + movl %esi,24(%esp)
|
| +.align 16
|
| +.L009loop:
|
| + movl %eax,%esi
|
| + andl $255,%esi
|
| + movl (%ebp,%esi,8),%esi
|
| + movzbl %dh,%edi
|
| + xorl 3(%ebp,%edi,8),%esi
|
| + movl %ecx,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edi
|
| + xorl 2(%ebp,%edi,8),%esi
|
| + movl %ebx,%edi
|
| + shrl $24,%edi
|
| + xorl 1(%ebp,%edi,8),%esi
|
| + movl %esi,4(%esp)
|
| +
|
| + movl %ebx,%esi
|
| + andl $255,%esi
|
| + movl (%ebp,%esi,8),%esi
|
| + movzbl %ah,%edi
|
| + xorl 3(%ebp,%edi,8),%esi
|
| + movl %edx,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edi
|
| + xorl 2(%ebp,%edi,8),%esi
|
| + movl %ecx,%edi
|
| + shrl $24,%edi
|
| + xorl 1(%ebp,%edi,8),%esi
|
| + movl %esi,8(%esp)
|
| +
|
| + movl %ecx,%esi
|
| + andl $255,%esi
|
| + movl (%ebp,%esi,8),%esi
|
| + movzbl %bh,%edi
|
| + xorl 3(%ebp,%edi,8),%esi
|
| + movl %eax,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edi
|
| + xorl 2(%ebp,%edi,8),%esi
|
| + movl %edx,%edi
|
| + shrl $24,%edi
|
| + xorl 1(%ebp,%edi,8),%esi
|
| +
|
| + movl 20(%esp),%edi
|
| + andl $255,%edx
|
| + movl (%ebp,%edx,8),%edx
|
| + movzbl %ch,%ecx
|
| + xorl 3(%ebp,%ecx,8),%edx
|
| + movl %esi,%ecx
|
| + shrl $16,%ebx
|
| + andl $255,%ebx
|
| + xorl 2(%ebp,%ebx,8),%edx
|
| + movl 8(%esp),%ebx
|
| + shrl $24,%eax
|
| + xorl 1(%ebp,%eax,8),%edx
|
| + movl 4(%esp),%eax
|
| +
|
| + addl $16,%edi
|
| + xorl (%edi),%eax
|
| + xorl 4(%edi),%ebx
|
| + xorl 8(%edi),%ecx
|
| + xorl 12(%edi),%edx
|
| + cmpl 24(%esp),%edi
|
| + movl %edi,20(%esp)
|
| + jb .L009loop
|
| + leal 2176(%ebp),%ebp
|
| + movl -128(%ebp),%edi
|
| + movl -96(%ebp),%esi
|
| + movl -64(%ebp),%edi
|
| + movl -32(%ebp),%esi
|
| + movl (%ebp),%edi
|
| + movl 32(%ebp),%esi
|
| + movl 64(%ebp),%edi
|
| + movl 96(%ebp),%esi
|
| + leal -128(%ebp),%ebp
|
| + movl %eax,%esi
|
| + andl $255,%esi
|
| + movzbl (%ebp,%esi,1),%esi
|
| + movzbl %dh,%edi
|
| + movzbl (%ebp,%edi,1),%edi
|
| + shll $8,%edi
|
| + xorl %edi,%esi
|
| + movl %ecx,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edi
|
| + movzbl (%ebp,%edi,1),%edi
|
| + shll $16,%edi
|
| + xorl %edi,%esi
|
| + movl %ebx,%edi
|
| + shrl $24,%edi
|
| + movzbl (%ebp,%edi,1),%edi
|
| + shll $24,%edi
|
| + xorl %edi,%esi
|
| + movl %esi,4(%esp)
|
| + movl %ebx,%esi
|
| + andl $255,%esi
|
| + movzbl (%ebp,%esi,1),%esi
|
| + movzbl %ah,%edi
|
| + movzbl (%ebp,%edi,1),%edi
|
| + shll $8,%edi
|
| + xorl %edi,%esi
|
| + movl %edx,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edi
|
| + movzbl (%ebp,%edi,1),%edi
|
| + shll $16,%edi
|
| + xorl %edi,%esi
|
| + movl %ecx,%edi
|
| + shrl $24,%edi
|
| + movzbl (%ebp,%edi,1),%edi
|
| + shll $24,%edi
|
| + xorl %edi,%esi
|
| + movl %esi,8(%esp)
|
| + movl %ecx,%esi
|
| + andl $255,%esi
|
| + movzbl (%ebp,%esi,1),%esi
|
| + movzbl %bh,%edi
|
| + movzbl (%ebp,%edi,1),%edi
|
| + shll $8,%edi
|
| + xorl %edi,%esi
|
| + movl %eax,%edi
|
| + shrl $16,%edi
|
| + andl $255,%edi
|
| + movzbl (%ebp,%edi,1),%edi
|
| + shll $16,%edi
|
| + xorl %edi,%esi
|
| + movl %edx,%edi
|
| + shrl $24,%edi
|
| + movzbl (%ebp,%edi,1),%edi
|
| + shll $24,%edi
|
| + xorl %edi,%esi
|
| + movl 20(%esp),%edi
|
| + andl $255,%edx
|
| + movzbl (%ebp,%edx,1),%edx
|
| + movzbl %ch,%ecx
|
| + movzbl (%ebp,%ecx,1),%ecx
|
| + shll $8,%ecx
|
| + xorl %ecx,%edx
|
| + movl %esi,%ecx
|
| + shrl $16,%ebx
|
| + andl $255,%ebx
|
| + movzbl (%ebp,%ebx,1),%ebx
|
| + shll $16,%ebx
|
| + xorl %ebx,%edx
|
| + movl 8(%esp),%ebx
|
| + shrl $24,%eax
|
| + movzbl (%ebp,%eax,1),%eax
|
| + shll $24,%eax
|
| + xorl %eax,%edx
|
| + movl 4(%esp),%eax
|
| + leal -2048(%ebp),%ebp
|
| + addl $16,%edi
|
| + xorl (%edi),%eax
|
| + xorl 4(%edi),%ebx
|
| + xorl 8(%edi),%ecx
|
| + xorl 12(%edi),%edx
|
| + ret
|
| +.align 64
|
| +.LAES_Td:
|
| +.long 1353184337,1353184337
|
| +.long 1399144830,1399144830
|
| +.long 3282310938,3282310938
|
| +.long 2522752826,2522752826
|
| +.long 3412831035,3412831035
|
| +.long 4047871263,4047871263
|
| +.long 2874735276,2874735276
|
| +.long 2466505547,2466505547
|
| +.long 1442459680,1442459680
|
| +.long 4134368941,4134368941
|
| +.long 2440481928,2440481928
|
| +.long 625738485,625738485
|
| +.long 4242007375,4242007375
|
| +.long 3620416197,3620416197
|
| +.long 2151953702,2151953702
|
| +.long 2409849525,2409849525
|
| +.long 1230680542,1230680542
|
| +.long 1729870373,1729870373
|
| +.long 2551114309,2551114309
|
| +.long 3787521629,3787521629
|
| +.long 41234371,41234371
|
| +.long 317738113,317738113
|
| +.long 2744600205,2744600205
|
| +.long 3338261355,3338261355
|
| +.long 3881799427,3881799427
|
| +.long 2510066197,2510066197
|
| +.long 3950669247,3950669247
|
| +.long 3663286933,3663286933
|
| +.long 763608788,763608788
|
| +.long 3542185048,3542185048
|
| +.long 694804553,694804553
|
| +.long 1154009486,1154009486
|
| +.long 1787413109,1787413109
|
| +.long 2021232372,2021232372
|
| +.long 1799248025,1799248025
|
| +.long 3715217703,3715217703
|
| +.long 3058688446,3058688446
|
| +.long 397248752,397248752
|
| +.long 1722556617,1722556617
|
| +.long 3023752829,3023752829
|
| +.long 407560035,407560035
|
| +.long 2184256229,2184256229
|
| +.long 1613975959,1613975959
|
| +.long 1165972322,1165972322
|
| +.long 3765920945,3765920945
|
| +.long 2226023355,2226023355
|
| +.long 480281086,480281086
|
| +.long 2485848313,2485848313
|
| +.long 1483229296,1483229296
|
| +.long 436028815,436028815
|
| +.long 2272059028,2272059028
|
| +.long 3086515026,3086515026
|
| +.long 601060267,601060267
|
| +.long 3791801202,3791801202
|
| +.long 1468997603,1468997603
|
| +.long 715871590,715871590
|
| +.long 120122290,120122290
|
| +.long 63092015,63092015
|
| +.long 2591802758,2591802758
|
| +.long 2768779219,2768779219
|
| +.long 4068943920,4068943920
|
| +.long 2997206819,2997206819
|
| +.long 3127509762,3127509762
|
| +.long 1552029421,1552029421
|
| +.long 723308426,723308426
|
| +.long 2461301159,2461301159
|
| +.long 4042393587,4042393587
|
| +.long 2715969870,2715969870
|
| +.long 3455375973,3455375973
|
| +.long 3586000134,3586000134
|
| +.long 526529745,526529745
|
| +.long 2331944644,2331944644
|
| +.long 2639474228,2639474228
|
| +.long 2689987490,2689987490
|
| +.long 853641733,853641733
|
| +.long 1978398372,1978398372
|
| +.long 971801355,971801355
|
| +.long 2867814464,2867814464
|
| +.long 111112542,111112542
|
| +.long 1360031421,1360031421
|
| +.long 4186579262,4186579262
|
| +.long 1023860118,1023860118
|
| +.long 2919579357,2919579357
|
| +.long 1186850381,1186850381
|
| +.long 3045938321,3045938321
|
| +.long 90031217,90031217
|
| +.long 1876166148,1876166148
|
| +.long 4279586912,4279586912
|
| +.long 620468249,620468249
|
| +.long 2548678102,2548678102
|
| +.long 3426959497,3426959497
|
| +.long 2006899047,2006899047
|
| +.long 3175278768,3175278768
|
| +.long 2290845959,2290845959
|
| +.long 945494503,945494503
|
| +.long 3689859193,3689859193
|
| +.long 1191869601,1191869601
|
| +.long 3910091388,3910091388
|
| +.long 3374220536,3374220536
|
| +.long 0,0
|
| +.long 2206629897,2206629897
|
| +.long 1223502642,1223502642
|
| +.long 2893025566,2893025566
|
| +.long 1316117100,1316117100
|
| +.long 4227796733,4227796733
|
| +.long 1446544655,1446544655
|
| +.long 517320253,517320253
|
| +.long 658058550,658058550
|
| +.long 1691946762,1691946762
|
| +.long 564550760,564550760
|
| +.long 3511966619,3511966619
|
| +.long 976107044,976107044
|
| +.long 2976320012,2976320012
|
| +.long 266819475,266819475
|
| +.long 3533106868,3533106868
|
| +.long 2660342555,2660342555
|
| +.long 1338359936,1338359936
|
| +.long 2720062561,2720062561
|
| +.long 1766553434,1766553434
|
| +.long 370807324,370807324
|
| +.long 179999714,179999714
|
| +.long 3844776128,3844776128
|
| +.long 1138762300,1138762300
|
| +.long 488053522,488053522
|
| +.long 185403662,185403662
|
| +.long 2915535858,2915535858
|
| +.long 3114841645,3114841645
|
| +.long 3366526484,3366526484
|
| +.long 2233069911,2233069911
|
| +.long 1275557295,1275557295
|
| +.long 3151862254,3151862254
|
| +.long 4250959779,4250959779
|
| +.long 2670068215,2670068215
|
| +.long 3170202204,3170202204
|
| +.long 3309004356,3309004356
|
| +.long 880737115,880737115
|
| +.long 1982415755,1982415755
|
| +.long 3703972811,3703972811
|
| +.long 1761406390,1761406390
|
| +.long 1676797112,1676797112
|
| +.long 3403428311,3403428311
|
| +.long 277177154,277177154
|
| +.long 1076008723,1076008723
|
| +.long 538035844,538035844
|
| +.long 2099530373,2099530373
|
| +.long 4164795346,4164795346
|
| +.long 288553390,288553390
|
| +.long 1839278535,1839278535
|
| +.long 1261411869,1261411869
|
| +.long 4080055004,4080055004
|
| +.long 3964831245,3964831245
|
| +.long 3504587127,3504587127
|
| +.long 1813426987,1813426987
|
| +.long 2579067049,2579067049
|
| +.long 4199060497,4199060497
|
| +.long 577038663,577038663
|
| +.long 3297574056,3297574056
|
| +.long 440397984,440397984
|
| +.long 3626794326,3626794326
|
| +.long 4019204898,4019204898
|
| +.long 3343796615,3343796615
|
| +.long 3251714265,3251714265
|
| +.long 4272081548,4272081548
|
| +.long 906744984,906744984
|
| +.long 3481400742,3481400742
|
| +.long 685669029,685669029
|
| +.long 646887386,646887386
|
| +.long 2764025151,2764025151
|
| +.long 3835509292,3835509292
|
| +.long 227702864,227702864
|
| +.long 2613862250,2613862250
|
| +.long 1648787028,1648787028
|
| +.long 3256061430,3256061430
|
| +.long 3904428176,3904428176
|
| +.long 1593260334,1593260334
|
| +.long 4121936770,4121936770
|
| +.long 3196083615,3196083615
|
| +.long 2090061929,2090061929
|
| +.long 2838353263,2838353263
|
| +.long 3004310991,3004310991
|
| +.long 999926984,999926984
|
| +.long 2809993232,2809993232
|
| +.long 1852021992,1852021992
|
| +.long 2075868123,2075868123
|
| +.long 158869197,158869197
|
| +.long 4095236462,4095236462
|
| +.long 28809964,28809964
|
| +.long 2828685187,2828685187
|
| +.long 1701746150,1701746150
|
| +.long 2129067946,2129067946
|
| +.long 147831841,147831841
|
| +.long 3873969647,3873969647
|
| +.long 3650873274,3650873274
|
| +.long 3459673930,3459673930
|
| +.long 3557400554,3557400554
|
| +.long 3598495785,3598495785
|
| +.long 2947720241,2947720241
|
| +.long 824393514,824393514
|
| +.long 815048134,815048134
|
| +.long 3227951669,3227951669
|
| +.long 935087732,935087732
|
| +.long 2798289660,2798289660
|
| +.long 2966458592,2966458592
|
| +.long 366520115,366520115
|
| +.long 1251476721,1251476721
|
| +.long 4158319681,4158319681
|
| +.long 240176511,240176511
|
| +.long 804688151,804688151
|
| +.long 2379631990,2379631990
|
| +.long 1303441219,1303441219
|
| +.long 1414376140,1414376140
|
| +.long 3741619940,3741619940
|
| +.long 3820343710,3820343710
|
| +.long 461924940,461924940
|
| +.long 3089050817,3089050817
|
| +.long 2136040774,2136040774
|
| +.long 82468509,82468509
|
| +.long 1563790337,1563790337
|
| +.long 1937016826,1937016826
|
| +.long 776014843,776014843
|
| +.long 1511876531,1511876531
|
| +.long 1389550482,1389550482
|
| +.long 861278441,861278441
|
| +.long 323475053,323475053
|
| +.long 2355222426,2355222426
|
| +.long 2047648055,2047648055
|
| +.long 2383738969,2383738969
|
| +.long 2302415851,2302415851
|
| +.long 3995576782,3995576782
|
| +.long 902390199,902390199
|
| +.long 3991215329,3991215329
|
| +.long 1018251130,1018251130
|
| +.long 1507840668,1507840668
|
| +.long 1064563285,1064563285
|
| +.long 2043548696,2043548696
|
| +.long 3208103795,3208103795
|
| +.long 3939366739,3939366739
|
| +.long 1537932639,1537932639
|
| +.long 342834655,342834655
|
| +.long 2262516856,2262516856
|
| +.long 2180231114,2180231114
|
| +.long 1053059257,1053059257
|
| +.long 741614648,741614648
|
| +.long 1598071746,1598071746
|
| +.long 1925389590,1925389590
|
| +.long 203809468,203809468
|
| +.long 2336832552,2336832552
|
| +.long 1100287487,1100287487
|
| +.long 1895934009,1895934009
|
| +.long 3736275976,3736275976
|
| +.long 2632234200,2632234200
|
| +.long 2428589668,2428589668
|
| +.long 1636092795,1636092795
|
| +.long 1890988757,1890988757
|
| +.long 1952214088,1952214088
|
| +.long 1113045200,1113045200
|
| +.byte 82,9,106,213,48,54,165,56
|
| +.byte 191,64,163,158,129,243,215,251
|
| +.byte 124,227,57,130,155,47,255,135
|
| +.byte 52,142,67,68,196,222,233,203
|
| +.byte 84,123,148,50,166,194,35,61
|
| +.byte 238,76,149,11,66,250,195,78
|
| +.byte 8,46,161,102,40,217,36,178
|
| +.byte 118,91,162,73,109,139,209,37
|
| +.byte 114,248,246,100,134,104,152,22
|
| +.byte 212,164,92,204,93,101,182,146
|
| +.byte 108,112,72,80,253,237,185,218
|
| +.byte 94,21,70,87,167,141,157,132
|
| +.byte 144,216,171,0,140,188,211,10
|
| +.byte 247,228,88,5,184,179,69,6
|
| +.byte 208,44,30,143,202,63,15,2
|
| +.byte 193,175,189,3,1,19,138,107
|
| +.byte 58,145,17,65,79,103,220,234
|
| +.byte 151,242,207,206,240,180,230,115
|
| +.byte 150,172,116,34,231,173,53,133
|
| +.byte 226,249,55,232,28,117,223,110
|
| +.byte 71,241,26,113,29,41,197,137
|
| +.byte 111,183,98,14,170,24,190,27
|
| +.byte 252,86,62,75,198,210,121,32
|
| +.byte 154,219,192,254,120,205,90,244
|
| +.byte 31,221,168,51,136,7,199,49
|
| +.byte 177,18,16,89,39,128,236,95
|
| +.byte 96,81,127,169,25,181,74,13
|
| +.byte 45,229,122,159,147,201,156,239
|
| +.byte 160,224,59,77,174,42,245,176
|
| +.byte 200,235,187,60,131,83,153,97
|
| +.byte 23,43,4,126,186,119,214,38
|
| +.byte 225,105,20,99,85,33,12,125
|
| +.byte 82,9,106,213,48,54,165,56
|
| +.byte 191,64,163,158,129,243,215,251
|
| +.byte 124,227,57,130,155,47,255,135
|
| +.byte 52,142,67,68,196,222,233,203
|
| +.byte 84,123,148,50,166,194,35,61
|
| +.byte 238,76,149,11,66,250,195,78
|
| +.byte 8,46,161,102,40,217,36,178
|
| +.byte 118,91,162,73,109,139,209,37
|
| +.byte 114,248,246,100,134,104,152,22
|
| +.byte 212,164,92,204,93,101,182,146
|
| +.byte 108,112,72,80,253,237,185,218
|
| +.byte 94,21,70,87,167,141,157,132
|
| +.byte 144,216,171,0,140,188,211,10
|
| +.byte 247,228,88,5,184,179,69,6
|
| +.byte 208,44,30,143,202,63,15,2
|
| +.byte 193,175,189,3,1,19,138,107
|
| +.byte 58,145,17,65,79,103,220,234
|
| +.byte 151,242,207,206,240,180,230,115
|
| +.byte 150,172,116,34,231,173,53,133
|
| +.byte 226,249,55,232,28,117,223,110
|
| +.byte 71,241,26,113,29,41,197,137
|
| +.byte 111,183,98,14,170,24,190,27
|
| +.byte 252,86,62,75,198,210,121,32
|
| +.byte 154,219,192,254,120,205,90,244
|
| +.byte 31,221,168,51,136,7,199,49
|
| +.byte 177,18,16,89,39,128,236,95
|
| +.byte 96,81,127,169,25,181,74,13
|
| +.byte 45,229,122,159,147,201,156,239
|
| +.byte 160,224,59,77,174,42,245,176
|
| +.byte 200,235,187,60,131,83,153,97
|
| +.byte 23,43,4,126,186,119,214,38
|
| +.byte 225,105,20,99,85,33,12,125
|
| +.byte 82,9,106,213,48,54,165,56
|
| +.byte 191,64,163,158,129,243,215,251
|
| +.byte 124,227,57,130,155,47,255,135
|
| +.byte 52,142,67,68,196,222,233,203
|
| +.byte 84,123,148,50,166,194,35,61
|
| +.byte 238,76,149,11,66,250,195,78
|
| +.byte 8,46,161,102,40,217,36,178
|
| +.byte 118,91,162,73,109,139,209,37
|
| +.byte 114,248,246,100,134,104,152,22
|
| +.byte 212,164,92,204,93,101,182,146
|
| +.byte 108,112,72,80,253,237,185,218
|
| +.byte 94,21,70,87,167,141,157,132
|
| +.byte 144,216,171,0,140,188,211,10
|
| +.byte 247,228,88,5,184,179,69,6
|
| +.byte 208,44,30,143,202,63,15,2
|
| +.byte 193,175,189,3,1,19,138,107
|
| +.byte 58,145,17,65,79,103,220,234
|
| +.byte 151,242,207,206,240,180,230,115
|
| +.byte 150,172,116,34,231,173,53,133
|
| +.byte 226,249,55,232,28,117,223,110
|
| +.byte 71,241,26,113,29,41,197,137
|
| +.byte 111,183,98,14,170,24,190,27
|
| +.byte 252,86,62,75,198,210,121,32
|
| +.byte 154,219,192,254,120,205,90,244
|
| +.byte 31,221,168,51,136,7,199,49
|
| +.byte 177,18,16,89,39,128,236,95
|
| +.byte 96,81,127,169,25,181,74,13
|
| +.byte 45,229,122,159,147,201,156,239
|
| +.byte 160,224,59,77,174,42,245,176
|
| +.byte 200,235,187,60,131,83,153,97
|
| +.byte 23,43,4,126,186,119,214,38
|
| +.byte 225,105,20,99,85,33,12,125
|
| +.byte 82,9,106,213,48,54,165,56
|
| +.byte 191,64,163,158,129,243,215,251
|
| +.byte 124,227,57,130,155,47,255,135
|
| +.byte 52,142,67,68,196,222,233,203
|
| +.byte 84,123,148,50,166,194,35,61
|
| +.byte 238,76,149,11,66,250,195,78
|
| +.byte 8,46,161,102,40,217,36,178
|
| +.byte 118,91,162,73,109,139,209,37
|
| +.byte 114,248,246,100,134,104,152,22
|
| +.byte 212,164,92,204,93,101,182,146
|
| +.byte 108,112,72,80,253,237,185,218
|
| +.byte 94,21,70,87,167,141,157,132
|
| +.byte 144,216,171,0,140,188,211,10
|
| +.byte 247,228,88,5,184,179,69,6
|
| +.byte 208,44,30,143,202,63,15,2
|
| +.byte 193,175,189,3,1,19,138,107
|
| +.byte 58,145,17,65,79,103,220,234
|
| +.byte 151,242,207,206,240,180,230,115
|
| +.byte 150,172,116,34,231,173,53,133
|
| +.byte 226,249,55,232,28,117,223,110
|
| +.byte 71,241,26,113,29,41,197,137
|
| +.byte 111,183,98,14,170,24,190,27
|
| +.byte 252,86,62,75,198,210,121,32
|
| +.byte 154,219,192,254,120,205,90,244
|
| +.byte 31,221,168,51,136,7,199,49
|
| +.byte 177,18,16,89,39,128,236,95
|
| +.byte 96,81,127,169,25,181,74,13
|
| +.byte 45,229,122,159,147,201,156,239
|
| +.byte 160,224,59,77,174,42,245,176
|
| +.byte 200,235,187,60,131,83,153,97
|
| +.byte 23,43,4,126,186,119,214,38
|
| +.byte 225,105,20,99,85,33,12,125
|
| +.size _x86_AES_decrypt,.-_x86_AES_decrypt
|
| +.globl asm_AES_decrypt
|
| +.hidden asm_AES_decrypt
|
| +.type asm_AES_decrypt,@function
|
| +.align 16
|
| +asm_AES_decrypt:
|
| +.L_asm_AES_decrypt_begin:
|
| + pushl %ebp
|
| + pushl %ebx
|
| + pushl %esi
|
| + pushl %edi
|
| + movl 20(%esp),%esi
|
| + movl 28(%esp),%edi
|
| + movl %esp,%eax
|
| + subl $36,%esp
|
| + andl $-64,%esp
|
| + leal -127(%edi),%ebx
|
| + subl %esp,%ebx
|
| + negl %ebx
|
| + andl $960,%ebx
|
| + subl %ebx,%esp
|
| + addl $4,%esp
|
| + movl %eax,28(%esp)
|
| + call .L010pic_point
|
| +.L010pic_point:
|
| + popl %ebp
|
| + leal OPENSSL_ia32cap_P-.L010pic_point(%ebp),%eax
|
| + leal .LAES_Td-.L010pic_point(%ebp),%ebp
|
| + leal 764(%esp),%ebx
|
| + subl %ebp,%ebx
|
| + andl $768,%ebx
|
| + leal 2176(%ebp,%ebx,1),%ebp
|
| + btl $25,(%eax)
|
| + jnc .L011x86
|
| + movq (%esi),%mm0
|
| + movq 8(%esi),%mm4
|
| + call _sse_AES_decrypt_compact
|
| + movl 28(%esp),%esp
|
| + movl 24(%esp),%esi
|
| + movq %mm0,(%esi)
|
| + movq %mm4,8(%esi)
|
| + emms
|
| + popl %edi
|
| + popl %esi
|
| + popl %ebx
|
| + popl %ebp
|
| + ret
|
| +.align 16
|
| +.L011x86:
|
| + movl %ebp,24(%esp)
|
| + movl (%esi),%eax
|
| + movl 4(%esi),%ebx
|
| + movl 8(%esi),%ecx
|
| + movl 12(%esi),%edx
|
| + call _x86_AES_decrypt_compact
|
| + movl 28(%esp),%esp
|
| + movl 24(%esp),%esi
|
| + movl %eax,(%esi)
|
| + movl %ebx,4(%esi)
|
| + movl %ecx,8(%esi)
|
| + movl %edx,12(%esi)
|
| + popl %edi
|
| + popl %esi
|
| + popl %ebx
|
| + popl %ebp
|
| + ret
|
| +.size asm_AES_decrypt,.-.L_asm_AES_decrypt_begin
|
| +.globl asm_AES_cbc_encrypt
|
| +.hidden asm_AES_cbc_encrypt
|
| +.type asm_AES_cbc_encrypt,@function
|
| +.align 16
|
| +asm_AES_cbc_encrypt:
|
| +.L_asm_AES_cbc_encrypt_begin:
|
| + pushl %ebp
|
| + pushl %ebx
|
| + pushl %esi
|
| + pushl %edi
|
| + movl 28(%esp),%ecx
|
| + cmpl $0,%ecx
|
| + je .L012drop_out
|
| + call .L013pic_point
|
| +.L013pic_point:
|
| + popl %ebp
|
| + leal OPENSSL_ia32cap_P-.L013pic_point(%ebp),%eax
|
| + cmpl $0,40(%esp)
|
| + leal .LAES_Te-.L013pic_point(%ebp),%ebp
|
| + jne .L014picked_te
|
| + leal .LAES_Td-.LAES_Te(%ebp),%ebp
|
| +.L014picked_te:
|
| + pushfl
|
| + cld
|
| + cmpl $512,%ecx
|
| + jb .L015slow_way
|
| + testl $15,%ecx
|
| + jnz .L015slow_way
|
| + btl $28,(%eax)
|
| + jc .L015slow_way
|
| + leal -324(%esp),%esi
|
| + andl $-64,%esi
|
| + movl %ebp,%eax
|
| + leal 2304(%ebp),%ebx
|
| + movl %esi,%edx
|
| + andl $4095,%eax
|
| + andl $4095,%ebx
|
| + andl $4095,%edx
|
| + cmpl %ebx,%edx
|
| + jb .L016tbl_break_out
|
| + subl %ebx,%edx
|
| + subl %edx,%esi
|
| + jmp .L017tbl_ok
|
| +.align 4
|
| +.L016tbl_break_out:
|
| + subl %eax,%edx
|
| + andl $4095,%edx
|
| + addl $384,%edx
|
| + subl %edx,%esi
|
| +.align 4
|
| +.L017tbl_ok:
|
| + leal 24(%esp),%edx
|
| + xchgl %esi,%esp
|
| + addl $4,%esp
|
| + movl %ebp,24(%esp)
|
| + movl %esi,28(%esp)
|
| + movl (%edx),%eax
|
| + movl 4(%edx),%ebx
|
| + movl 12(%edx),%edi
|
| + movl 16(%edx),%esi
|
| + movl 20(%edx),%edx
|
| + movl %eax,32(%esp)
|
| + movl %ebx,36(%esp)
|
| + movl %ecx,40(%esp)
|
| + movl %edi,44(%esp)
|
| + movl %esi,48(%esp)
|
| + movl $0,316(%esp)
|
| + movl %edi,%ebx
|
| + movl $61,%ecx
|
| + subl %ebp,%ebx
|
| + movl %edi,%esi
|
| + andl $4095,%ebx
|
| + leal 76(%esp),%edi
|
| + cmpl $2304,%ebx
|
| + jb .L018do_copy
|
| + cmpl $3852,%ebx
|
| + jb .L019skip_copy
|
| +.align 4
|
| +.L018do_copy:
|
| + movl %edi,44(%esp)
|
| +.long 2784229001
|
| +.L019skip_copy:
|
| + movl $16,%edi
|
| +.align 4
|
| +.L020prefetch_tbl:
|
| + movl (%ebp),%eax
|
| + movl 32(%ebp),%ebx
|
| + movl 64(%ebp),%ecx
|
| + movl 96(%ebp),%esi
|
| + leal 128(%ebp),%ebp
|
| + subl $1,%edi
|
| + jnz .L020prefetch_tbl
|
| + subl $2048,%ebp
|
| + movl 32(%esp),%esi
|
| + movl 48(%esp),%edi
|
| + cmpl $0,%edx
|
| + je .L021fast_decrypt
|
| + movl (%edi),%eax
|
| + movl 4(%edi),%ebx
|
| +.align 16
|
| +.L022fast_enc_loop:
|
| + movl 8(%edi),%ecx
|
| + movl 12(%edi),%edx
|
| + xorl (%esi),%eax
|
| + xorl 4(%esi),%ebx
|
| + xorl 8(%esi),%ecx
|
| + xorl 12(%esi),%edx
|
| + movl 44(%esp),%edi
|
| + call _x86_AES_encrypt
|
| + movl 32(%esp),%esi
|
| + movl 36(%esp),%edi
|
| + movl %eax,(%edi)
|
| + movl %ebx,4(%edi)
|
| + movl %ecx,8(%edi)
|
| + movl %edx,12(%edi)
|
| + leal 16(%esi),%esi
|
| + movl 40(%esp),%ecx
|
| + movl %esi,32(%esp)
|
| + leal 16(%edi),%edx
|
| + movl %edx,36(%esp)
|
| + subl $16,%ecx
|
| + movl %ecx,40(%esp)
|
| + jnz .L022fast_enc_loop
|
| + movl 48(%esp),%esi
|
| + movl 8(%edi),%ecx
|
| + movl 12(%edi),%edx
|
| + movl %eax,(%esi)
|
| + movl %ebx,4(%esi)
|
| + movl %ecx,8(%esi)
|
| + movl %edx,12(%esi)
|
| + cmpl $0,316(%esp)
|
| + movl 44(%esp),%edi
|
| + je .L023skip_ezero
|
| + movl $60,%ecx
|
| + xorl %eax,%eax
|
| +.align 4
|
| +.long 2884892297
|
| +.L023skip_ezero:
|
| + movl 28(%esp),%esp
|
| + popfl
|
| +.L012drop_out:
|
| + popl %edi
|
| + popl %esi
|
| + popl %ebx
|
| + popl %ebp
|
| + ret
|
| + pushfl
|
| +.align 16
|
| +.L021fast_decrypt:
|
| + cmpl 36(%esp),%esi
|
| + je .L024fast_dec_in_place
|
| + movl %edi,52(%esp)
|
| +.align 4
|
| +.align 16
|
| +.L025fast_dec_loop:
|
| + movl (%esi),%eax
|
| + movl 4(%esi),%ebx
|
| + movl 8(%esi),%ecx
|
| + movl 12(%esi),%edx
|
| + movl 44(%esp),%edi
|
| + call _x86_AES_decrypt
|
| + movl 52(%esp),%edi
|
| + movl 40(%esp),%esi
|
| + xorl (%edi),%eax
|
| + xorl 4(%edi),%ebx
|
| + xorl 8(%edi),%ecx
|
| + xorl 12(%edi),%edx
|
| + movl 36(%esp),%edi
|
| + movl 32(%esp),%esi
|
| + movl %eax,(%edi)
|
| + movl %ebx,4(%edi)
|
| + movl %ecx,8(%edi)
|
| + movl %edx,12(%edi)
|
| + movl 40(%esp),%ecx
|
| + movl %esi,52(%esp)
|
| + leal 16(%esi),%esi
|
| + movl %esi,32(%esp)
|
| + leal 16(%edi),%edi
|
| + movl %edi,36(%esp)
|
| + subl $16,%ecx
|
| + movl %ecx,40(%esp)
|
| + jnz .L025fast_dec_loop
|
| + movl 52(%esp),%edi
|
| + movl 48(%esp),%esi
|
| + movl (%edi),%eax
|
| + movl 4(%edi),%ebx
|
| + movl 8(%edi),%ecx
|
| + movl 12(%edi),%edx
|
| + movl %eax,(%esi)
|
| + movl %ebx,4(%esi)
|
| + movl %ecx,8(%esi)
|
| + movl %edx,12(%esi)
|
| + jmp .L026fast_dec_out
|
| +.align 16
|
| +.L024fast_dec_in_place:
|
| +.L027fast_dec_in_place_loop:
|
| + movl (%esi),%eax
|
| + movl 4(%esi),%ebx
|
| + movl 8(%esi),%ecx
|
| + movl 12(%esi),%edx
|
| + leal 60(%esp),%edi
|
| + movl %eax,(%edi)
|
| + movl %ebx,4(%edi)
|
| + movl %ecx,8(%edi)
|
| + movl %edx,12(%edi)
|
| + movl 44(%esp),%edi
|
| + call _x86_AES_decrypt
|
| + movl 48(%esp),%edi
|
| + movl 36(%esp),%esi
|
| + xorl (%edi),%eax
|
| + xorl 4(%edi),%ebx
|
| + xorl 8(%edi),%ecx
|
| + xorl 12(%edi),%edx
|
| + movl %eax,(%esi)
|
| + movl %ebx,4(%esi)
|
| + movl %ecx,8(%esi)
|
| + movl %edx,12(%esi)
|
| + leal 16(%esi),%esi
|
| + movl %esi,36(%esp)
|
| + leal 60(%esp),%esi
|
| + movl (%esi),%eax
|
| + movl 4(%esi),%ebx
|
| + movl 8(%esi),%ecx
|
| + movl 12(%esi),%edx
|
| + movl %eax,(%edi)
|
| + movl %ebx,4(%edi)
|
| + movl %ecx,8(%edi)
|
| + movl %edx,12(%edi)
|
| + movl 32(%esp),%esi
|
| + movl 40(%esp),%ecx
|
| + leal 16(%esi),%esi
|
| + movl %esi,32(%esp)
|
| + subl $16,%ecx
|
| + movl %ecx,40(%esp)
|
| + jnz .L027fast_dec_in_place_loop
|
| +.align 4
|
| +.L026fast_dec_out:
|
| + cmpl $0,316(%esp)
|
| + movl 44(%esp),%edi
|
| + je .L028skip_dzero
|
| + movl $60,%ecx
|
| + xorl %eax,%eax
|
| +.align 4
|
| +.long 2884892297
|
| +.L028skip_dzero:
|
| + movl 28(%esp),%esp
|
| + popfl
|
| + popl %edi
|
| + popl %esi
|
| + popl %ebx
|
| + popl %ebp
|
| + ret
|
| + pushfl
|
| +.align 16
|
| +.L015slow_way:
|
| + movl (%eax),%eax
|
| + movl 36(%esp),%edi
|
| + leal -80(%esp),%esi
|
| + andl $-64,%esi
|
| + leal -143(%edi),%ebx
|
| + subl %esi,%ebx
|
| + negl %ebx
|
| + andl $960,%ebx
|
| + subl %ebx,%esi
|
| + leal 768(%esi),%ebx
|
| + subl %ebp,%ebx
|
| + andl $768,%ebx
|
| + leal 2176(%ebp,%ebx,1),%ebp
|
| + leal 24(%esp),%edx
|
| + xchgl %esi,%esp
|
| + addl $4,%esp
|
| + movl %ebp,24(%esp)
|
| + movl %esi,28(%esp)
|
| + movl %eax,52(%esp)
|
| + movl (%edx),%eax
|
| + movl 4(%edx),%ebx
|
| + movl 16(%edx),%esi
|
| + movl 20(%edx),%edx
|
| + movl %eax,32(%esp)
|
| + movl %ebx,36(%esp)
|
| + movl %ecx,40(%esp)
|
| + movl %edi,44(%esp)
|
| + movl %esi,48(%esp)
|
| + movl %esi,%edi
|
| + movl %eax,%esi
|
| + cmpl $0,%edx
|
| + je .L029slow_decrypt
|
| + cmpl $16,%ecx
|
| + movl %ebx,%edx
|
| + jb .L030slow_enc_tail
|
| + btl $25,52(%esp)
|
| + jnc .L031slow_enc_x86
|
| + movq (%edi),%mm0
|
| + movq 8(%edi),%mm4
|
| +.align 16
|
| +.L032slow_enc_loop_sse:
|
| + pxor (%esi),%mm0
|
| + pxor 8(%esi),%mm4
|
| + movl 44(%esp),%edi
|
| + call _sse_AES_encrypt_compact
|
| + movl 32(%esp),%esi
|
| + movl 36(%esp),%edi
|
| + movl 40(%esp),%ecx
|
| + movq %mm0,(%edi)
|
| + movq %mm4,8(%edi)
|
| + leal 16(%esi),%esi
|
| + movl %esi,32(%esp)
|
| + leal 16(%edi),%edx
|
| + movl %edx,36(%esp)
|
| + subl $16,%ecx
|
| + cmpl $16,%ecx
|
| + movl %ecx,40(%esp)
|
| + jae .L032slow_enc_loop_sse
|
| + testl $15,%ecx
|
| + jnz .L030slow_enc_tail
|
| + movl 48(%esp),%esi
|
| + movq %mm0,(%esi)
|
| + movq %mm4,8(%esi)
|
| + emms
|
| + movl 28(%esp),%esp
|
| + popfl
|
| + popl %edi
|
| + popl %esi
|
| + popl %ebx
|
| + popl %ebp
|
| + ret
|
| + pushfl
|
| +.align 16
|
| +.L031slow_enc_x86:
|
| + movl (%edi),%eax
|
| + movl 4(%edi),%ebx
|
| +.align 4
|
| +.L033slow_enc_loop_x86:
|
| + movl 8(%edi),%ecx
|
| + movl 12(%edi),%edx
|
| + xorl (%esi),%eax
|
| + xorl 4(%esi),%ebx
|
| + xorl 8(%esi),%ecx
|
| + xorl 12(%esi),%edx
|
| + movl 44(%esp),%edi
|
| + call _x86_AES_encrypt_compact
|
| + movl 32(%esp),%esi
|
| + movl 36(%esp),%edi
|
| + movl %eax,(%edi)
|
| + movl %ebx,4(%edi)
|
| + movl %ecx,8(%edi)
|
| + movl %edx,12(%edi)
|
| + movl 40(%esp),%ecx
|
| + leal 16(%esi),%esi
|
| + movl %esi,32(%esp)
|
| + leal 16(%edi),%edx
|
| + movl %edx,36(%esp)
|
| + subl $16,%ecx
|
| + cmpl $16,%ecx
|
| + movl %ecx,40(%esp)
|
| + jae .L033slow_enc_loop_x86
|
| + testl $15,%ecx
|
| + jnz .L030slow_enc_tail
|
| + movl 48(%esp),%esi
|
| + movl 8(%edi),%ecx
|
| + movl 12(%edi),%edx
|
| + movl %eax,(%esi)
|
| + movl %ebx,4(%esi)
|
| + movl %ecx,8(%esi)
|
| + movl %edx,12(%esi)
|
| + movl 28(%esp),%esp
|
| + popfl
|
| + popl %edi
|
| + popl %esi
|
| + popl %ebx
|
| + popl %ebp
|
| + ret
|
| + pushfl
|
| +.align 16
|
| +.L030slow_enc_tail:
|
| + emms
|
| + movl %edx,%edi
|
| + movl $16,%ebx
|
| + subl %ecx,%ebx
|
| + cmpl %esi,%edi
|
| + je .L034enc_in_place
|
| +.align 4
|
| +.long 2767451785
|
| + jmp .L035enc_skip_in_place
|
| +.L034enc_in_place:
|
| + leal (%edi,%ecx,1),%edi
|
| +.L035enc_skip_in_place:
|
| + movl %ebx,%ecx
|
| + xorl %eax,%eax
|
| +.align 4
|
| +.long 2868115081
|
| + movl 48(%esp),%edi
|
| + movl %edx,%esi
|
| + movl (%edi),%eax
|
| + movl 4(%edi),%ebx
|
| + movl $16,40(%esp)
|
| + jmp .L033slow_enc_loop_x86
|
| +.align 16
|
| +.L029slow_decrypt:
|
| + btl $25,52(%esp)
|
| + jnc .L036slow_dec_loop_x86
|
| +.align 4
|
| +.L037slow_dec_loop_sse:
|
| + movq (%esi),%mm0
|
| + movq 8(%esi),%mm4
|
| + movl 44(%esp),%edi
|
| + call _sse_AES_decrypt_compact
|
| + movl 32(%esp),%esi
|
| + leal 60(%esp),%eax
|
| + movl 36(%esp),%ebx
|
| + movl 40(%esp),%ecx
|
| + movl 48(%esp),%edi
|
| + movq (%esi),%mm1
|
| + movq 8(%esi),%mm5
|
| + pxor (%edi),%mm0
|
| + pxor 8(%edi),%mm4
|
| + movq %mm1,(%edi)
|
| + movq %mm5,8(%edi)
|
| + subl $16,%ecx
|
| + jc .L038slow_dec_partial_sse
|
| + movq %mm0,(%ebx)
|
| + movq %mm4,8(%ebx)
|
| + leal 16(%ebx),%ebx
|
| + movl %ebx,36(%esp)
|
| + leal 16(%esi),%esi
|
| + movl %esi,32(%esp)
|
| + movl %ecx,40(%esp)
|
| + jnz .L037slow_dec_loop_sse
|
| + emms
|
| + movl 28(%esp),%esp
|
| + popfl
|
| + popl %edi
|
| + popl %esi
|
| + popl %ebx
|
| + popl %ebp
|
| + ret
|
| + pushfl
|
| +.align 16
|
| +.L038slow_dec_partial_sse:
|
| + movq %mm0,(%eax)
|
| + movq %mm4,8(%eax)
|
| + emms
|
| + addl $16,%ecx
|
| + movl %ebx,%edi
|
| + movl %eax,%esi
|
| +.align 4
|
| +.long 2767451785
|
| + movl 28(%esp),%esp
|
| + popfl
|
| + popl %edi
|
| + popl %esi
|
| + popl %ebx
|
| + popl %ebp
|
| + ret
|
| + pushfl
|
| +.align 16
|
| +.L036slow_dec_loop_x86:
|
| + movl (%esi),%eax
|
| + movl 4(%esi),%ebx
|
| + movl 8(%esi),%ecx
|
| + movl 12(%esi),%edx
|
| + leal 60(%esp),%edi
|
| + movl %eax,(%edi)
|
| + movl %ebx,4(%edi)
|
| + movl %ecx,8(%edi)
|
| + movl %edx,12(%edi)
|
| + movl 44(%esp),%edi
|
| + call _x86_AES_decrypt_compact
|
| + movl 48(%esp),%edi
|
| + movl 40(%esp),%esi
|
| + xorl (%edi),%eax
|
| + xorl 4(%edi),%ebx
|
| + xorl 8(%edi),%ecx
|
| + xorl 12(%edi),%edx
|
| + subl $16,%esi
|
| + jc .L039slow_dec_partial_x86
|
| + movl %esi,40(%esp)
|
| + movl 36(%esp),%esi
|
| + movl %eax,(%esi)
|
| + movl %ebx,4(%esi)
|
| + movl %ecx,8(%esi)
|
| + movl %edx,12(%esi)
|
| + leal 16(%esi),%esi
|
| + movl %esi,36(%esp)
|
| + leal 60(%esp),%esi
|
| + movl (%esi),%eax
|
| + movl 4(%esi),%ebx
|
| + movl 8(%esi),%ecx
|
| + movl 12(%esi),%edx
|
| + movl %eax,(%edi)
|
| + movl %ebx,4(%edi)
|
| + movl %ecx,8(%edi)
|
| + movl %edx,12(%edi)
|
| + movl 32(%esp),%esi
|
| + leal 16(%esi),%esi
|
| + movl %esi,32(%esp)
|
| + jnz .L036slow_dec_loop_x86
|
| + movl 28(%esp),%esp
|
| + popfl
|
| + popl %edi
|
| + popl %esi
|
| + popl %ebx
|
| + popl %ebp
|
| + ret
|
| + pushfl
|
| +.align 16
|
| +.L039slow_dec_partial_x86:
|
| + leal 60(%esp),%esi
|
| + movl %eax,(%esi)
|
| + movl %ebx,4(%esi)
|
| + movl %ecx,8(%esi)
|
| + movl %edx,12(%esi)
|
| + movl 32(%esp),%esi
|
| + movl (%esi),%eax
|
| + movl 4(%esi),%ebx
|
| + movl 8(%esi),%ecx
|
| + movl 12(%esi),%edx
|
| + movl %eax,(%edi)
|
| + movl %ebx,4(%edi)
|
| + movl %ecx,8(%edi)
|
| + movl %edx,12(%edi)
|
| + movl 40(%esp),%ecx
|
| + movl 36(%esp),%edi
|
| + leal 60(%esp),%esi
|
| +.align 4
|
| +.long 2767451785
|
| + movl 28(%esp),%esp
|
| + popfl
|
| + popl %edi
|
| + popl %esi
|
| + popl %ebx
|
| + popl %ebp
|
| + ret
|
| +.size asm_AES_cbc_encrypt,.-.L_asm_AES_cbc_encrypt_begin
|
| +.hidden _x86_AES_set_encrypt_key
|
| +.type _x86_AES_set_encrypt_key,@function
|
| +.align 16
|
| +_x86_AES_set_encrypt_key:
|
| + pushl %ebp
|
| + pushl %ebx
|
| + pushl %esi
|
| + pushl %edi
|
| + movl 24(%esp),%esi
|
| + movl 32(%esp),%edi
|
| + testl $-1,%esi
|
| + jz .L040badpointer
|
| + testl $-1,%edi
|
| + jz .L040badpointer
|
| + call .L041pic_point
|
| +.L041pic_point:
|
| + popl %ebp
|
| + leal .LAES_Te-.L041pic_point(%ebp),%ebp
|
| + leal 2176(%ebp),%ebp
|
| + movl -128(%ebp),%eax
|
| + movl -96(%ebp),%ebx
|
| + movl -64(%ebp),%ecx
|
| + movl -32(%ebp),%edx
|
| + movl (%ebp),%eax
|
| + movl 32(%ebp),%ebx
|
| + movl 64(%ebp),%ecx
|
| + movl 96(%ebp),%edx
|
| + movl 28(%esp),%ecx
|
| + cmpl $128,%ecx
|
| + je .L04210rounds
|
| + cmpl $192,%ecx
|
| + je .L04312rounds
|
| + cmpl $256,%ecx
|
| + je .L04414rounds
|
| + movl $-2,%eax
|
| + jmp .L045exit
|
| +.L04210rounds:
|
| + movl (%esi),%eax
|
| + movl 4(%esi),%ebx
|
| + movl 8(%esi),%ecx
|
| + movl 12(%esi),%edx
|
| + movl %eax,(%edi)
|
| + movl %ebx,4(%edi)
|
| + movl %ecx,8(%edi)
|
| + movl %edx,12(%edi)
|
| + xorl %ecx,%ecx
|
| + jmp .L04610shortcut
|
| +.align 4
|
| +.L04710loop:
|
| + movl (%edi),%eax
|
| + movl 12(%edi),%edx
|
| +.L04610shortcut:
|
| + movzbl %dl,%esi
|
| + movzbl -128(%ebp,%esi,1),%ebx
|
| + movzbl %dh,%esi
|
| + shll $24,%ebx
|
| + xorl %ebx,%eax
|
| + movzbl -128(%ebp,%esi,1),%ebx
|
| + shrl $16,%edx
|
| + movzbl %dl,%esi
|
| + xorl %ebx,%eax
|
| + movzbl -128(%ebp,%esi,1),%ebx
|
| + movzbl %dh,%esi
|
| + shll $8,%ebx
|
| + xorl %ebx,%eax
|
| + movzbl -128(%ebp,%esi,1),%ebx
|
| + shll $16,%ebx
|
| + xorl %ebx,%eax
|
| + xorl 896(%ebp,%ecx,4),%eax
|
| + movl %eax,16(%edi)
|
| + xorl 4(%edi),%eax
|
| + movl %eax,20(%edi)
|
| + xorl 8(%edi),%eax
|
| + movl %eax,24(%edi)
|
| + xorl 12(%edi),%eax
|
| + movl %eax,28(%edi)
|
| + incl %ecx
|
| + addl $16,%edi
|
| + cmpl $10,%ecx
|
| + jl .L04710loop
|
| + movl $10,80(%edi)
|
| + xorl %eax,%eax
|
| + jmp .L045exit
|
| +.L04312rounds:
|
| + movl (%esi),%eax
|
| + movl 4(%esi),%ebx
|
| + movl 8(%esi),%ecx
|
| + movl 12(%esi),%edx
|
| + movl %eax,(%edi)
|
| + movl %ebx,4(%edi)
|
| + movl %ecx,8(%edi)
|
| + movl %edx,12(%edi)
|
| + movl 16(%esi),%ecx
|
| + movl 20(%esi),%edx
|
| + movl %ecx,16(%edi)
|
| + movl %edx,20(%edi)
|
| + xorl %ecx,%ecx
|
| + jmp .L04812shortcut
|
| +.align 4
|
| +.L04912loop:
|
| + movl (%edi),%eax
|
| + movl 20(%edi),%edx
|
| +.L04812shortcut:
|
| + movzbl %dl,%esi
|
| + movzbl -128(%ebp,%esi,1),%ebx
|
| + movzbl %dh,%esi
|
| + shll $24,%ebx
|
| + xorl %ebx,%eax
|
| + movzbl -128(%ebp,%esi,1),%ebx
|
| + shrl $16,%edx
|
| + movzbl %dl,%esi
|
| + xorl %ebx,%eax
|
| + movzbl -128(%ebp,%esi,1),%ebx
|
| + movzbl %dh,%esi
|
| + shll $8,%ebx
|
| + xorl %ebx,%eax
|
| + movzbl -128(%ebp,%esi,1),%ebx
|
| + shll $16,%ebx
|
| + xorl %ebx,%eax
|
| + xorl 896(%ebp,%ecx,4),%eax
|
| + movl %eax,24(%edi)
|
| + xorl 4(%edi),%eax
|
| + movl %eax,28(%edi)
|
| + xorl 8(%edi),%eax
|
| + movl %eax,32(%edi)
|
| + xorl 12(%edi),%eax
|
| + movl %eax,36(%edi)
|
| + cmpl $7,%ecx
|
| + je .L05012break
|
| + incl %ecx
|
| + xorl 16(%edi),%eax
|
| + movl %eax,40(%edi)
|
| + xorl 20(%edi),%eax
|
| + movl %eax,44(%edi)
|
| + addl $24,%edi
|
| + jmp .L04912loop
|
| +.L05012break:
|
| + movl $12,72(%edi)
|
| + xorl %eax,%eax
|
| + jmp .L045exit
|
| +.L04414rounds:
|
| + movl (%esi),%eax
|
| + movl 4(%esi),%ebx
|
| + movl 8(%esi),%ecx
|
| + movl 12(%esi),%edx
|
| + movl %eax,(%edi)
|
| + movl %ebx,4(%edi)
|
| + movl %ecx,8(%edi)
|
| + movl %edx,12(%edi)
|
| + movl 16(%esi),%eax
|
| + movl 20(%esi),%ebx
|
| + movl 24(%esi),%ecx
|
| + movl 28(%esi),%edx
|
| + movl %eax,16(%edi)
|
| + movl %ebx,20(%edi)
|
| + movl %ecx,24(%edi)
|
| + movl %edx,28(%edi)
|
| + xorl %ecx,%ecx
|
| + jmp .L05114shortcut
|
| +.align 4
|
| +.L05214loop:
|
| + movl 28(%edi),%edx
|
| +.L05114shortcut:
|
| + movl (%edi),%eax
|
| + movzbl %dl,%esi
|
| + movzbl -128(%ebp,%esi,1),%ebx
|
| + movzbl %dh,%esi
|
| + shll $24,%ebx
|
| + xorl %ebx,%eax
|
| + movzbl -128(%ebp,%esi,1),%ebx
|
| + shrl $16,%edx
|
| + movzbl %dl,%esi
|
| + xorl %ebx,%eax
|
| + movzbl -128(%ebp,%esi,1),%ebx
|
| + movzbl %dh,%esi
|
| + shll $8,%ebx
|
| + xorl %ebx,%eax
|
| + movzbl -128(%ebp,%esi,1),%ebx
|
| + shll $16,%ebx
|
| + xorl %ebx,%eax
|
| + xorl 896(%ebp,%ecx,4),%eax
|
| + movl %eax,32(%edi)
|
| + xorl 4(%edi),%eax
|
| + movl %eax,36(%edi)
|
| + xorl 8(%edi),%eax
|
| + movl %eax,40(%edi)
|
| + xorl 12(%edi),%eax
|
| + movl %eax,44(%edi)
|
| + cmpl $6,%ecx
|
| + je .L05314break
|
| + incl %ecx
|
| + movl %eax,%edx
|
| + movl 16(%edi),%eax
|
| + movzbl %dl,%esi
|
| + movzbl -128(%ebp,%esi,1),%ebx
|
| + movzbl %dh,%esi
|
| + xorl %ebx,%eax
|
| + movzbl -128(%ebp,%esi,1),%ebx
|
| + shrl $16,%edx
|
| + shll $8,%ebx
|
| + movzbl %dl,%esi
|
| + xorl %ebx,%eax
|
| + movzbl -128(%ebp,%esi,1),%ebx
|
| + movzbl %dh,%esi
|
| + shll $16,%ebx
|
| + xorl %ebx,%eax
|
| + movzbl -128(%ebp,%esi,1),%ebx
|
| + shll $24,%ebx
|
| + xorl %ebx,%eax
|
| + movl %eax,48(%edi)
|
| + xorl 20(%edi),%eax
|
| + movl %eax,52(%edi)
|
| + xorl 24(%edi),%eax
|
| + movl %eax,56(%edi)
|
| + xorl 28(%edi),%eax
|
| + movl %eax,60(%edi)
|
| + addl $32,%edi
|
| + jmp .L05214loop
|
| +.L05314break:
|
| + movl $14,48(%edi)
|
| + xorl %eax,%eax
|
| + jmp .L045exit
|
| +.L040badpointer:
|
| + movl $-1,%eax
|
| +.L045exit:
|
| + popl %edi
|
| + popl %esi
|
| + popl %ebx
|
| + popl %ebp
|
| + ret
|
| +.size _x86_AES_set_encrypt_key,.-_x86_AES_set_encrypt_key
|
| +.globl asm_AES_set_encrypt_key
|
| +.hidden asm_AES_set_encrypt_key
|
| +.type asm_AES_set_encrypt_key,@function
|
| +.align 16
|
| +asm_AES_set_encrypt_key:
|
| +.L_asm_AES_set_encrypt_key_begin:
|
| + call _x86_AES_set_encrypt_key
|
| + ret
|
| +.size asm_AES_set_encrypt_key,.-.L_asm_AES_set_encrypt_key_begin
|
| +.globl asm_AES_set_decrypt_key
|
| +.hidden asm_AES_set_decrypt_key
|
| +.type asm_AES_set_decrypt_key,@function
|
| +.align 16
|
| +asm_AES_set_decrypt_key:
|
| +.L_asm_AES_set_decrypt_key_begin:
|
| + call _x86_AES_set_encrypt_key
|
| + cmpl $0,%eax
|
| + je .L054proceed
|
| + ret
|
| +.L054proceed:
|
| + pushl %ebp
|
| + pushl %ebx
|
| + pushl %esi
|
| + pushl %edi
|
| + movl 28(%esp),%esi
|
| + movl 240(%esi),%ecx
|
| + leal (,%ecx,4),%ecx
|
| + leal (%esi,%ecx,4),%edi
|
| +.align 4
|
| +.L055invert:
|
| + movl (%esi),%eax
|
| + movl 4(%esi),%ebx
|
| + movl (%edi),%ecx
|
| + movl 4(%edi),%edx
|
| + movl %eax,(%edi)
|
| + movl %ebx,4(%edi)
|
| + movl %ecx,(%esi)
|
| + movl %edx,4(%esi)
|
| + movl 8(%esi),%eax
|
| + movl 12(%esi),%ebx
|
| + movl 8(%edi),%ecx
|
| + movl 12(%edi),%edx
|
| + movl %eax,8(%edi)
|
| + movl %ebx,12(%edi)
|
| + movl %ecx,8(%esi)
|
| + movl %edx,12(%esi)
|
| + addl $16,%esi
|
| + subl $16,%edi
|
| + cmpl %edi,%esi
|
| + jne .L055invert
|
| + movl 28(%esp),%edi
|
| + movl 240(%edi),%esi
|
| + leal -2(%esi,%esi,1),%esi
|
| + leal (%edi,%esi,8),%esi
|
| + movl %esi,28(%esp)
|
| + movl 16(%edi),%eax
|
| +.align 4
|
| +.L056permute:
|
| + addl $16,%edi
|
| + movl $2155905152,%ebp
|
| + andl %eax,%ebp
|
| + leal (%eax,%eax,1),%ebx
|
| + movl %ebp,%esi
|
| + shrl $7,%ebp
|
| + subl %ebp,%esi
|
| + andl $4278124286,%ebx
|
| + andl $454761243,%esi
|
| + xorl %esi,%ebx
|
| + movl $2155905152,%ebp
|
| + andl %ebx,%ebp
|
| + leal (%ebx,%ebx,1),%ecx
|
| + movl %ebp,%esi
|
| + shrl $7,%ebp
|
| + subl %ebp,%esi
|
| + andl $4278124286,%ecx
|
| + andl $454761243,%esi
|
| + xorl %eax,%ebx
|
| + xorl %esi,%ecx
|
| + movl $2155905152,%ebp
|
| + andl %ecx,%ebp
|
| + leal (%ecx,%ecx,1),%edx
|
| + movl %ebp,%esi
|
| + shrl $7,%ebp
|
| + xorl %eax,%ecx
|
| + subl %ebp,%esi
|
| + andl $4278124286,%edx
|
| + andl $454761243,%esi
|
| + roll $8,%eax
|
| + xorl %esi,%edx
|
| + movl 4(%edi),%ebp
|
| + xorl %ebx,%eax
|
| + xorl %edx,%ebx
|
| + xorl %ecx,%eax
|
| + roll $24,%ebx
|
| + xorl %edx,%ecx
|
| + xorl %edx,%eax
|
| + roll $16,%ecx
|
| + xorl %ebx,%eax
|
| + roll $8,%edx
|
| + xorl %ecx,%eax
|
| + movl %ebp,%ebx
|
| + xorl %edx,%eax
|
| + movl %eax,(%edi)
|
| + movl $2155905152,%ebp
|
| + andl %ebx,%ebp
|
| + leal (%ebx,%ebx,1),%ecx
|
| + movl %ebp,%esi
|
| + shrl $7,%ebp
|
| + subl %ebp,%esi
|
| + andl $4278124286,%ecx
|
| + andl $454761243,%esi
|
| + xorl %esi,%ecx
|
| + movl $2155905152,%ebp
|
| + andl %ecx,%ebp
|
| + leal (%ecx,%ecx,1),%edx
|
| + movl %ebp,%esi
|
| + shrl $7,%ebp
|
| + subl %ebp,%esi
|
| + andl $4278124286,%edx
|
| + andl $454761243,%esi
|
| + xorl %ebx,%ecx
|
| + xorl %esi,%edx
|
| + movl $2155905152,%ebp
|
| + andl %edx,%ebp
|
| + leal (%edx,%edx,1),%eax
|
| + movl %ebp,%esi
|
| + shrl $7,%ebp
|
| + xorl %ebx,%edx
|
| + subl %ebp,%esi
|
| + andl $4278124286,%eax
|
| + andl $454761243,%esi
|
| + roll $8,%ebx
|
| + xorl %esi,%eax
|
| + movl 8(%edi),%ebp
|
| + xorl %ecx,%ebx
|
| + xorl %eax,%ecx
|
| + xorl %edx,%ebx
|
| + roll $24,%ecx
|
| + xorl %eax,%edx
|
| + xorl %eax,%ebx
|
| + roll $16,%edx
|
| + xorl %ecx,%ebx
|
| + roll $8,%eax
|
| + xorl %edx,%ebx
|
| + movl %ebp,%ecx
|
| + xorl %eax,%ebx
|
| + movl %ebx,4(%edi)
|
| + movl $2155905152,%ebp
|
| + andl %ecx,%ebp
|
| + leal (%ecx,%ecx,1),%edx
|
| + movl %ebp,%esi
|
| + shrl $7,%ebp
|
| + subl %ebp,%esi
|
| + andl $4278124286,%edx
|
| + andl $454761243,%esi
|
| + xorl %esi,%edx
|
| + movl $2155905152,%ebp
|
| + andl %edx,%ebp
|
| + leal (%edx,%edx,1),%eax
|
| + movl %ebp,%esi
|
| + shrl $7,%ebp
|
| + subl %ebp,%esi
|
| + andl $4278124286,%eax
|
| + andl $454761243,%esi
|
| + xorl %ecx,%edx
|
| + xorl %esi,%eax
|
| + movl $2155905152,%ebp
|
| + andl %eax,%ebp
|
| + leal (%eax,%eax,1),%ebx
|
| + movl %ebp,%esi
|
| + shrl $7,%ebp
|
| + xorl %ecx,%eax
|
| + subl %ebp,%esi
|
| + andl $4278124286,%ebx
|
| + andl $454761243,%esi
|
| + roll $8,%ecx
|
| + xorl %esi,%ebx
|
| + movl 12(%edi),%ebp
|
| + xorl %edx,%ecx
|
| + xorl %ebx,%edx
|
| + xorl %eax,%ecx
|
| + roll $24,%edx
|
| + xorl %ebx,%eax
|
| + xorl %ebx,%ecx
|
| + roll $16,%eax
|
| + xorl %edx,%ecx
|
| + roll $8,%ebx
|
| + xorl %eax,%ecx
|
| + movl %ebp,%edx
|
| + xorl %ebx,%ecx
|
| + movl %ecx,8(%edi)
|
| + movl $2155905152,%ebp
|
| + andl %edx,%ebp
|
| + leal (%edx,%edx,1),%eax
|
| + movl %ebp,%esi
|
| + shrl $7,%ebp
|
| + subl %ebp,%esi
|
| + andl $4278124286,%eax
|
| + andl $454761243,%esi
|
| + xorl %esi,%eax
|
| + movl $2155905152,%ebp
|
| + andl %eax,%ebp
|
| + leal (%eax,%eax,1),%ebx
|
| + movl %ebp,%esi
|
| + shrl $7,%ebp
|
| + subl %ebp,%esi
|
| + andl $4278124286,%ebx
|
| + andl $454761243,%esi
|
| + xorl %edx,%eax
|
| + xorl %esi,%ebx
|
| + movl $2155905152,%ebp
|
| + andl %ebx,%ebp
|
| + leal (%ebx,%ebx,1),%ecx
|
| + movl %ebp,%esi
|
| + shrl $7,%ebp
|
| + xorl %edx,%ebx
|
| + subl %ebp,%esi
|
| + andl $4278124286,%ecx
|
| + andl $454761243,%esi
|
| + roll $8,%edx
|
| + xorl %esi,%ecx
|
| + movl 16(%edi),%ebp
|
| + xorl %eax,%edx
|
| + xorl %ecx,%eax
|
| + xorl %ebx,%edx
|
| + roll $24,%eax
|
| + xorl %ecx,%ebx
|
| + xorl %ecx,%edx
|
| + roll $16,%ebx
|
| + xorl %eax,%edx
|
| + roll $8,%ecx
|
| + xorl %ebx,%edx
|
| + movl %ebp,%eax
|
| + xorl %ecx,%edx
|
| + movl %edx,12(%edi)
|
| + cmpl 28(%esp),%edi
|
| + jb .L056permute
|
| + xorl %eax,%eax
|
| + popl %edi
|
| + popl %esi
|
| + popl %ebx
|
| + popl %ebp
|
| + ret
|
| +.size asm_AES_set_decrypt_key,.-.L_asm_AES_set_decrypt_key_begin
|
| +.byte 65,69,83,32,102,111,114,32,120,56,54,44,32,67,82,89
|
| +.byte 80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114
|
| +.byte 111,64,111,112,101,110,115,115,108,46,111,114,103,62,0
|
| +#endif
|
|
|