Index: third_party/boringssl/linux-x86/crypto/aes/aes-586.S |
diff --git a/third_party/boringssl/linux-x86/crypto/aes/aes-586.S b/third_party/boringssl/linux-x86/crypto/aes/aes-586.S |
new file mode 100644 |
index 0000000000000000000000000000000000000000..74282748ce0da92ead2f775480cd3409ac583f78 |
--- /dev/null |
+++ b/third_party/boringssl/linux-x86/crypto/aes/aes-586.S |
@@ -0,0 +1,3257 @@ |
+#if defined(__i386__) |
+.file "aes-586.S" |
+.text |
+.hidden _x86_AES_encrypt_compact |
+.type _x86_AES_encrypt_compact,@function |
+.align 16 |
+_x86_AES_encrypt_compact: |
+ movl %edi,20(%esp) |
+ xorl (%edi),%eax |
+ xorl 4(%edi),%ebx |
+ xorl 8(%edi),%ecx |
+ xorl 12(%edi),%edx |
+ movl 240(%edi),%esi |
+ leal -2(%esi,%esi,1),%esi |
+ leal (%edi,%esi,8),%esi |
+ movl %esi,24(%esp) |
+ movl -128(%ebp),%edi |
+ movl -96(%ebp),%esi |
+ movl -64(%ebp),%edi |
+ movl -32(%ebp),%esi |
+ movl (%ebp),%edi |
+ movl 32(%ebp),%esi |
+ movl 64(%ebp),%edi |
+ movl 96(%ebp),%esi |
+.align 16 |
+.L000loop: |
+ movl %eax,%esi |
+ andl $255,%esi |
+ movzbl -128(%ebp,%esi,1),%esi |
+ movzbl %bh,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $8,%edi |
+ xorl %edi,%esi |
+ movl %ecx,%edi |
+ shrl $16,%edi |
+ andl $255,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $16,%edi |
+ xorl %edi,%esi |
+ movl %edx,%edi |
+ shrl $24,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $24,%edi |
+ xorl %edi,%esi |
+ movl %esi,4(%esp) |
+ |
+ movl %ebx,%esi |
+ andl $255,%esi |
+ shrl $16,%ebx |
+ movzbl -128(%ebp,%esi,1),%esi |
+ movzbl %ch,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $8,%edi |
+ xorl %edi,%esi |
+ movl %edx,%edi |
+ shrl $16,%edi |
+ andl $255,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $16,%edi |
+ xorl %edi,%esi |
+ movl %eax,%edi |
+ shrl $24,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $24,%edi |
+ xorl %edi,%esi |
+ movl %esi,8(%esp) |
+ |
+ movl %ecx,%esi |
+ andl $255,%esi |
+ shrl $24,%ecx |
+ movzbl -128(%ebp,%esi,1),%esi |
+ movzbl %dh,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $8,%edi |
+ xorl %edi,%esi |
+ movl %eax,%edi |
+ shrl $16,%edi |
+ andl $255,%edx |
+ andl $255,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $16,%edi |
+ xorl %edi,%esi |
+ movzbl %bh,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $24,%edi |
+ xorl %edi,%esi |
+ |
+ andl $255,%edx |
+ movzbl -128(%ebp,%edx,1),%edx |
+ movzbl %ah,%eax |
+ movzbl -128(%ebp,%eax,1),%eax |
+ shll $8,%eax |
+ xorl %eax,%edx |
+ movl 4(%esp),%eax |
+ andl $255,%ebx |
+ movzbl -128(%ebp,%ebx,1),%ebx |
+ shll $16,%ebx |
+ xorl %ebx,%edx |
+ movl 8(%esp),%ebx |
+ movzbl -128(%ebp,%ecx,1),%ecx |
+ shll $24,%ecx |
+ xorl %ecx,%edx |
+ movl %esi,%ecx |
+ |
+ movl $2155905152,%ebp |
+ andl %ecx,%ebp |
+ leal (%ecx,%ecx,1),%edi |
+ movl %ebp,%esi |
+ shrl $7,%ebp |
+ andl $4278124286,%edi |
+ subl %ebp,%esi |
+ movl %ecx,%ebp |
+ andl $454761243,%esi |
+ rorl $16,%ebp |
+ xorl %edi,%esi |
+ movl %ecx,%edi |
+ xorl %esi,%ecx |
+ rorl $24,%edi |
+ xorl %ebp,%esi |
+ roll $24,%ecx |
+ xorl %edi,%esi |
+ movl $2155905152,%ebp |
+ xorl %esi,%ecx |
+ andl %edx,%ebp |
+ leal (%edx,%edx,1),%edi |
+ movl %ebp,%esi |
+ shrl $7,%ebp |
+ andl $4278124286,%edi |
+ subl %ebp,%esi |
+ movl %edx,%ebp |
+ andl $454761243,%esi |
+ rorl $16,%ebp |
+ xorl %edi,%esi |
+ movl %edx,%edi |
+ xorl %esi,%edx |
+ rorl $24,%edi |
+ xorl %ebp,%esi |
+ roll $24,%edx |
+ xorl %edi,%esi |
+ movl $2155905152,%ebp |
+ xorl %esi,%edx |
+ andl %eax,%ebp |
+ leal (%eax,%eax,1),%edi |
+ movl %ebp,%esi |
+ shrl $7,%ebp |
+ andl $4278124286,%edi |
+ subl %ebp,%esi |
+ movl %eax,%ebp |
+ andl $454761243,%esi |
+ rorl $16,%ebp |
+ xorl %edi,%esi |
+ movl %eax,%edi |
+ xorl %esi,%eax |
+ rorl $24,%edi |
+ xorl %ebp,%esi |
+ roll $24,%eax |
+ xorl %edi,%esi |
+ movl $2155905152,%ebp |
+ xorl %esi,%eax |
+ andl %ebx,%ebp |
+ leal (%ebx,%ebx,1),%edi |
+ movl %ebp,%esi |
+ shrl $7,%ebp |
+ andl $4278124286,%edi |
+ subl %ebp,%esi |
+ movl %ebx,%ebp |
+ andl $454761243,%esi |
+ rorl $16,%ebp |
+ xorl %edi,%esi |
+ movl %ebx,%edi |
+ xorl %esi,%ebx |
+ rorl $24,%edi |
+ xorl %ebp,%esi |
+ roll $24,%ebx |
+ xorl %edi,%esi |
+ xorl %esi,%ebx |
+ movl 20(%esp),%edi |
+ movl 28(%esp),%ebp |
+ addl $16,%edi |
+ xorl (%edi),%eax |
+ xorl 4(%edi),%ebx |
+ xorl 8(%edi),%ecx |
+ xorl 12(%edi),%edx |
+ cmpl 24(%esp),%edi |
+ movl %edi,20(%esp) |
+ jb .L000loop |
+ movl %eax,%esi |
+ andl $255,%esi |
+ movzbl -128(%ebp,%esi,1),%esi |
+ movzbl %bh,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $8,%edi |
+ xorl %edi,%esi |
+ movl %ecx,%edi |
+ shrl $16,%edi |
+ andl $255,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $16,%edi |
+ xorl %edi,%esi |
+ movl %edx,%edi |
+ shrl $24,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $24,%edi |
+ xorl %edi,%esi |
+ movl %esi,4(%esp) |
+ |
+ movl %ebx,%esi |
+ andl $255,%esi |
+ shrl $16,%ebx |
+ movzbl -128(%ebp,%esi,1),%esi |
+ movzbl %ch,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $8,%edi |
+ xorl %edi,%esi |
+ movl %edx,%edi |
+ shrl $16,%edi |
+ andl $255,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $16,%edi |
+ xorl %edi,%esi |
+ movl %eax,%edi |
+ shrl $24,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $24,%edi |
+ xorl %edi,%esi |
+ movl %esi,8(%esp) |
+ |
+ movl %ecx,%esi |
+ andl $255,%esi |
+ shrl $24,%ecx |
+ movzbl -128(%ebp,%esi,1),%esi |
+ movzbl %dh,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $8,%edi |
+ xorl %edi,%esi |
+ movl %eax,%edi |
+ shrl $16,%edi |
+ andl $255,%edx |
+ andl $255,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $16,%edi |
+ xorl %edi,%esi |
+ movzbl %bh,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $24,%edi |
+ xorl %edi,%esi |
+ |
+ movl 20(%esp),%edi |
+ andl $255,%edx |
+ movzbl -128(%ebp,%edx,1),%edx |
+ movzbl %ah,%eax |
+ movzbl -128(%ebp,%eax,1),%eax |
+ shll $8,%eax |
+ xorl %eax,%edx |
+ movl 4(%esp),%eax |
+ andl $255,%ebx |
+ movzbl -128(%ebp,%ebx,1),%ebx |
+ shll $16,%ebx |
+ xorl %ebx,%edx |
+ movl 8(%esp),%ebx |
+ movzbl -128(%ebp,%ecx,1),%ecx |
+ shll $24,%ecx |
+ xorl %ecx,%edx |
+ movl %esi,%ecx |
+ |
+ xorl 16(%edi),%eax |
+ xorl 20(%edi),%ebx |
+ xorl 24(%edi),%ecx |
+ xorl 28(%edi),%edx |
+ ret |
+.size _x86_AES_encrypt_compact,.-_x86_AES_encrypt_compact |
+.hidden _sse_AES_encrypt_compact |
+.type _sse_AES_encrypt_compact,@function |
+.align 16 |
+_sse_AES_encrypt_compact: |
+ pxor (%edi),%mm0 |
+ pxor 8(%edi),%mm4 |
+ movl 240(%edi),%esi |
+ leal -2(%esi,%esi,1),%esi |
+ leal (%edi,%esi,8),%esi |
+ movl %esi,24(%esp) |
+ movl $454761243,%eax |
+ movl %eax,8(%esp) |
+ movl %eax,12(%esp) |
+ movl -128(%ebp),%eax |
+ movl -96(%ebp),%ebx |
+ movl -64(%ebp),%ecx |
+ movl -32(%ebp),%edx |
+ movl (%ebp),%eax |
+ movl 32(%ebp),%ebx |
+ movl 64(%ebp),%ecx |
+ movl 96(%ebp),%edx |
+.align 16 |
+.L001loop: |
+ pshufw $8,%mm0,%mm1 |
+ pshufw $13,%mm4,%mm5 |
+ movd %mm1,%eax |
+ movd %mm5,%ebx |
+ movl %edi,20(%esp) |
+ movzbl %al,%esi |
+ movzbl %ah,%edx |
+ pshufw $13,%mm0,%mm2 |
+ movzbl -128(%ebp,%esi,1),%ecx |
+ movzbl %bl,%edi |
+ movzbl -128(%ebp,%edx,1),%edx |
+ shrl $16,%eax |
+ shll $8,%edx |
+ movzbl -128(%ebp,%edi,1),%esi |
+ movzbl %bh,%edi |
+ shll $16,%esi |
+ pshufw $8,%mm4,%mm6 |
+ orl %esi,%ecx |
+ movzbl -128(%ebp,%edi,1),%esi |
+ movzbl %ah,%edi |
+ shll $24,%esi |
+ shrl $16,%ebx |
+ orl %esi,%edx |
+ movzbl -128(%ebp,%edi,1),%esi |
+ movzbl %bh,%edi |
+ shll $8,%esi |
+ orl %esi,%ecx |
+ movzbl -128(%ebp,%edi,1),%esi |
+ movzbl %al,%edi |
+ shll $24,%esi |
+ orl %esi,%ecx |
+ movzbl -128(%ebp,%edi,1),%esi |
+ movzbl %bl,%edi |
+ movd %mm2,%eax |
+ movd %ecx,%mm0 |
+ movzbl -128(%ebp,%edi,1),%ecx |
+ movzbl %ah,%edi |
+ shll $16,%ecx |
+ movd %mm6,%ebx |
+ orl %esi,%ecx |
+ movzbl -128(%ebp,%edi,1),%esi |
+ movzbl %bh,%edi |
+ shll $24,%esi |
+ orl %esi,%ecx |
+ movzbl -128(%ebp,%edi,1),%esi |
+ movzbl %bl,%edi |
+ shll $8,%esi |
+ shrl $16,%ebx |
+ orl %esi,%ecx |
+ movzbl -128(%ebp,%edi,1),%esi |
+ movzbl %al,%edi |
+ shrl $16,%eax |
+ movd %ecx,%mm1 |
+ movzbl -128(%ebp,%edi,1),%ecx |
+ movzbl %ah,%edi |
+ shll $16,%ecx |
+ andl $255,%eax |
+ orl %esi,%ecx |
+ punpckldq %mm1,%mm0 |
+ movzbl -128(%ebp,%edi,1),%esi |
+ movzbl %bh,%edi |
+ shll $24,%esi |
+ andl $255,%ebx |
+ movzbl -128(%ebp,%eax,1),%eax |
+ orl %esi,%ecx |
+ shll $16,%eax |
+ movzbl -128(%ebp,%edi,1),%esi |
+ orl %eax,%edx |
+ shll $8,%esi |
+ movzbl -128(%ebp,%ebx,1),%ebx |
+ orl %esi,%ecx |
+ orl %ebx,%edx |
+ movl 20(%esp),%edi |
+ movd %ecx,%mm4 |
+ movd %edx,%mm5 |
+ punpckldq %mm5,%mm4 |
+ addl $16,%edi |
+ cmpl 24(%esp),%edi |
+ ja .L002out |
+ movq 8(%esp),%mm2 |
+ pxor %mm3,%mm3 |
+ pxor %mm7,%mm7 |
+ movq %mm0,%mm1 |
+ movq %mm4,%mm5 |
+ pcmpgtb %mm0,%mm3 |
+ pcmpgtb %mm4,%mm7 |
+ pand %mm2,%mm3 |
+ pand %mm2,%mm7 |
+ pshufw $177,%mm0,%mm2 |
+ pshufw $177,%mm4,%mm6 |
+ paddb %mm0,%mm0 |
+ paddb %mm4,%mm4 |
+ pxor %mm3,%mm0 |
+ pxor %mm7,%mm4 |
+ pshufw $177,%mm2,%mm3 |
+ pshufw $177,%mm6,%mm7 |
+ pxor %mm0,%mm1 |
+ pxor %mm4,%mm5 |
+ pxor %mm2,%mm0 |
+ pxor %mm6,%mm4 |
+ movq %mm3,%mm2 |
+ movq %mm7,%mm6 |
+ pslld $8,%mm3 |
+ pslld $8,%mm7 |
+ psrld $24,%mm2 |
+ psrld $24,%mm6 |
+ pxor %mm3,%mm0 |
+ pxor %mm7,%mm4 |
+ pxor %mm2,%mm0 |
+ pxor %mm6,%mm4 |
+ movq %mm1,%mm3 |
+ movq %mm5,%mm7 |
+ movq (%edi),%mm2 |
+ movq 8(%edi),%mm6 |
+ psrld $8,%mm1 |
+ psrld $8,%mm5 |
+ movl -128(%ebp),%eax |
+ pslld $24,%mm3 |
+ pslld $24,%mm7 |
+ movl -64(%ebp),%ebx |
+ pxor %mm1,%mm0 |
+ pxor %mm5,%mm4 |
+ movl (%ebp),%ecx |
+ pxor %mm3,%mm0 |
+ pxor %mm7,%mm4 |
+ movl 64(%ebp),%edx |
+ pxor %mm2,%mm0 |
+ pxor %mm6,%mm4 |
+ jmp .L001loop |
+.align 16 |
+.L002out: |
+ pxor (%edi),%mm0 |
+ pxor 8(%edi),%mm4 |
+ ret |
+.size _sse_AES_encrypt_compact,.-_sse_AES_encrypt_compact |
+.hidden _x86_AES_encrypt |
+.type _x86_AES_encrypt,@function |
+.align 16 |
+_x86_AES_encrypt: |
+ movl %edi,20(%esp) |
+ xorl (%edi),%eax |
+ xorl 4(%edi),%ebx |
+ xorl 8(%edi),%ecx |
+ xorl 12(%edi),%edx |
+ movl 240(%edi),%esi |
+ leal -2(%esi,%esi,1),%esi |
+ leal (%edi,%esi,8),%esi |
+ movl %esi,24(%esp) |
+.align 16 |
+.L003loop: |
+ movl %eax,%esi |
+ andl $255,%esi |
+ movl (%ebp,%esi,8),%esi |
+ movzbl %bh,%edi |
+ xorl 3(%ebp,%edi,8),%esi |
+ movl %ecx,%edi |
+ shrl $16,%edi |
+ andl $255,%edi |
+ xorl 2(%ebp,%edi,8),%esi |
+ movl %edx,%edi |
+ shrl $24,%edi |
+ xorl 1(%ebp,%edi,8),%esi |
+ movl %esi,4(%esp) |
+ |
+ movl %ebx,%esi |
+ andl $255,%esi |
+ shrl $16,%ebx |
+ movl (%ebp,%esi,8),%esi |
+ movzbl %ch,%edi |
+ xorl 3(%ebp,%edi,8),%esi |
+ movl %edx,%edi |
+ shrl $16,%edi |
+ andl $255,%edi |
+ xorl 2(%ebp,%edi,8),%esi |
+ movl %eax,%edi |
+ shrl $24,%edi |
+ xorl 1(%ebp,%edi,8),%esi |
+ movl %esi,8(%esp) |
+ |
+ movl %ecx,%esi |
+ andl $255,%esi |
+ shrl $24,%ecx |
+ movl (%ebp,%esi,8),%esi |
+ movzbl %dh,%edi |
+ xorl 3(%ebp,%edi,8),%esi |
+ movl %eax,%edi |
+ shrl $16,%edi |
+ andl $255,%edx |
+ andl $255,%edi |
+ xorl 2(%ebp,%edi,8),%esi |
+ movzbl %bh,%edi |
+ xorl 1(%ebp,%edi,8),%esi |
+ |
+ movl 20(%esp),%edi |
+ movl (%ebp,%edx,8),%edx |
+ movzbl %ah,%eax |
+ xorl 3(%ebp,%eax,8),%edx |
+ movl 4(%esp),%eax |
+ andl $255,%ebx |
+ xorl 2(%ebp,%ebx,8),%edx |
+ movl 8(%esp),%ebx |
+ xorl 1(%ebp,%ecx,8),%edx |
+ movl %esi,%ecx |
+ |
+ addl $16,%edi |
+ xorl (%edi),%eax |
+ xorl 4(%edi),%ebx |
+ xorl 8(%edi),%ecx |
+ xorl 12(%edi),%edx |
+ cmpl 24(%esp),%edi |
+ movl %edi,20(%esp) |
+ jb .L003loop |
+ movl %eax,%esi |
+ andl $255,%esi |
+ movl 2(%ebp,%esi,8),%esi |
+ andl $255,%esi |
+ movzbl %bh,%edi |
+ movl (%ebp,%edi,8),%edi |
+ andl $65280,%edi |
+ xorl %edi,%esi |
+ movl %ecx,%edi |
+ shrl $16,%edi |
+ andl $255,%edi |
+ movl (%ebp,%edi,8),%edi |
+ andl $16711680,%edi |
+ xorl %edi,%esi |
+ movl %edx,%edi |
+ shrl $24,%edi |
+ movl 2(%ebp,%edi,8),%edi |
+ andl $4278190080,%edi |
+ xorl %edi,%esi |
+ movl %esi,4(%esp) |
+ movl %ebx,%esi |
+ andl $255,%esi |
+ shrl $16,%ebx |
+ movl 2(%ebp,%esi,8),%esi |
+ andl $255,%esi |
+ movzbl %ch,%edi |
+ movl (%ebp,%edi,8),%edi |
+ andl $65280,%edi |
+ xorl %edi,%esi |
+ movl %edx,%edi |
+ shrl $16,%edi |
+ andl $255,%edi |
+ movl (%ebp,%edi,8),%edi |
+ andl $16711680,%edi |
+ xorl %edi,%esi |
+ movl %eax,%edi |
+ shrl $24,%edi |
+ movl 2(%ebp,%edi,8),%edi |
+ andl $4278190080,%edi |
+ xorl %edi,%esi |
+ movl %esi,8(%esp) |
+ movl %ecx,%esi |
+ andl $255,%esi |
+ shrl $24,%ecx |
+ movl 2(%ebp,%esi,8),%esi |
+ andl $255,%esi |
+ movzbl %dh,%edi |
+ movl (%ebp,%edi,8),%edi |
+ andl $65280,%edi |
+ xorl %edi,%esi |
+ movl %eax,%edi |
+ shrl $16,%edi |
+ andl $255,%edx |
+ andl $255,%edi |
+ movl (%ebp,%edi,8),%edi |
+ andl $16711680,%edi |
+ xorl %edi,%esi |
+ movzbl %bh,%edi |
+ movl 2(%ebp,%edi,8),%edi |
+ andl $4278190080,%edi |
+ xorl %edi,%esi |
+ movl 20(%esp),%edi |
+ andl $255,%edx |
+ movl 2(%ebp,%edx,8),%edx |
+ andl $255,%edx |
+ movzbl %ah,%eax |
+ movl (%ebp,%eax,8),%eax |
+ andl $65280,%eax |
+ xorl %eax,%edx |
+ movl 4(%esp),%eax |
+ andl $255,%ebx |
+ movl (%ebp,%ebx,8),%ebx |
+ andl $16711680,%ebx |
+ xorl %ebx,%edx |
+ movl 8(%esp),%ebx |
+ movl 2(%ebp,%ecx,8),%ecx |
+ andl $4278190080,%ecx |
+ xorl %ecx,%edx |
+ movl %esi,%ecx |
+ addl $16,%edi |
+ xorl (%edi),%eax |
+ xorl 4(%edi),%ebx |
+ xorl 8(%edi),%ecx |
+ xorl 12(%edi),%edx |
+ ret |
+.align 64 |
+.LAES_Te: |
+.long 2774754246,2774754246 |
+.long 2222750968,2222750968 |
+.long 2574743534,2574743534 |
+.long 2373680118,2373680118 |
+.long 234025727,234025727 |
+.long 3177933782,3177933782 |
+.long 2976870366,2976870366 |
+.long 1422247313,1422247313 |
+.long 1345335392,1345335392 |
+.long 50397442,50397442 |
+.long 2842126286,2842126286 |
+.long 2099981142,2099981142 |
+.long 436141799,436141799 |
+.long 1658312629,1658312629 |
+.long 3870010189,3870010189 |
+.long 2591454956,2591454956 |
+.long 1170918031,1170918031 |
+.long 2642575903,2642575903 |
+.long 1086966153,1086966153 |
+.long 2273148410,2273148410 |
+.long 368769775,368769775 |
+.long 3948501426,3948501426 |
+.long 3376891790,3376891790 |
+.long 200339707,200339707 |
+.long 3970805057,3970805057 |
+.long 1742001331,1742001331 |
+.long 4255294047,4255294047 |
+.long 3937382213,3937382213 |
+.long 3214711843,3214711843 |
+.long 4154762323,4154762323 |
+.long 2524082916,2524082916 |
+.long 1539358875,1539358875 |
+.long 3266819957,3266819957 |
+.long 486407649,486407649 |
+.long 2928907069,2928907069 |
+.long 1780885068,1780885068 |
+.long 1513502316,1513502316 |
+.long 1094664062,1094664062 |
+.long 49805301,49805301 |
+.long 1338821763,1338821763 |
+.long 1546925160,1546925160 |
+.long 4104496465,4104496465 |
+.long 887481809,887481809 |
+.long 150073849,150073849 |
+.long 2473685474,2473685474 |
+.long 1943591083,1943591083 |
+.long 1395732834,1395732834 |
+.long 1058346282,1058346282 |
+.long 201589768,201589768 |
+.long 1388824469,1388824469 |
+.long 1696801606,1696801606 |
+.long 1589887901,1589887901 |
+.long 672667696,672667696 |
+.long 2711000631,2711000631 |
+.long 251987210,251987210 |
+.long 3046808111,3046808111 |
+.long 151455502,151455502 |
+.long 907153956,907153956 |
+.long 2608889883,2608889883 |
+.long 1038279391,1038279391 |
+.long 652995533,652995533 |
+.long 1764173646,1764173646 |
+.long 3451040383,3451040383 |
+.long 2675275242,2675275242 |
+.long 453576978,453576978 |
+.long 2659418909,2659418909 |
+.long 1949051992,1949051992 |
+.long 773462580,773462580 |
+.long 756751158,756751158 |
+.long 2993581788,2993581788 |
+.long 3998898868,3998898868 |
+.long 4221608027,4221608027 |
+.long 4132590244,4132590244 |
+.long 1295727478,1295727478 |
+.long 1641469623,1641469623 |
+.long 3467883389,3467883389 |
+.long 2066295122,2066295122 |
+.long 1055122397,1055122397 |
+.long 1898917726,1898917726 |
+.long 2542044179,2542044179 |
+.long 4115878822,4115878822 |
+.long 1758581177,1758581177 |
+.long 0,0 |
+.long 753790401,753790401 |
+.long 1612718144,1612718144 |
+.long 536673507,536673507 |
+.long 3367088505,3367088505 |
+.long 3982187446,3982187446 |
+.long 3194645204,3194645204 |
+.long 1187761037,1187761037 |
+.long 3653156455,3653156455 |
+.long 1262041458,1262041458 |
+.long 3729410708,3729410708 |
+.long 3561770136,3561770136 |
+.long 3898103984,3898103984 |
+.long 1255133061,1255133061 |
+.long 1808847035,1808847035 |
+.long 720367557,720367557 |
+.long 3853167183,3853167183 |
+.long 385612781,385612781 |
+.long 3309519750,3309519750 |
+.long 3612167578,3612167578 |
+.long 1429418854,1429418854 |
+.long 2491778321,2491778321 |
+.long 3477423498,3477423498 |
+.long 284817897,284817897 |
+.long 100794884,100794884 |
+.long 2172616702,2172616702 |
+.long 4031795360,4031795360 |
+.long 1144798328,1144798328 |
+.long 3131023141,3131023141 |
+.long 3819481163,3819481163 |
+.long 4082192802,4082192802 |
+.long 4272137053,4272137053 |
+.long 3225436288,3225436288 |
+.long 2324664069,2324664069 |
+.long 2912064063,2912064063 |
+.long 3164445985,3164445985 |
+.long 1211644016,1211644016 |
+.long 83228145,83228145 |
+.long 3753688163,3753688163 |
+.long 3249976951,3249976951 |
+.long 1977277103,1977277103 |
+.long 1663115586,1663115586 |
+.long 806359072,806359072 |
+.long 452984805,452984805 |
+.long 250868733,250868733 |
+.long 1842533055,1842533055 |
+.long 1288555905,1288555905 |
+.long 336333848,336333848 |
+.long 890442534,890442534 |
+.long 804056259,804056259 |
+.long 3781124030,3781124030 |
+.long 2727843637,2727843637 |
+.long 3427026056,3427026056 |
+.long 957814574,957814574 |
+.long 1472513171,1472513171 |
+.long 4071073621,4071073621 |
+.long 2189328124,2189328124 |
+.long 1195195770,1195195770 |
+.long 2892260552,2892260552 |
+.long 3881655738,3881655738 |
+.long 723065138,723065138 |
+.long 2507371494,2507371494 |
+.long 2690670784,2690670784 |
+.long 2558624025,2558624025 |
+.long 3511635870,3511635870 |
+.long 2145180835,2145180835 |
+.long 1713513028,1713513028 |
+.long 2116692564,2116692564 |
+.long 2878378043,2878378043 |
+.long 2206763019,2206763019 |
+.long 3393603212,3393603212 |
+.long 703524551,703524551 |
+.long 3552098411,3552098411 |
+.long 1007948840,1007948840 |
+.long 2044649127,2044649127 |
+.long 3797835452,3797835452 |
+.long 487262998,487262998 |
+.long 1994120109,1994120109 |
+.long 1004593371,1004593371 |
+.long 1446130276,1446130276 |
+.long 1312438900,1312438900 |
+.long 503974420,503974420 |
+.long 3679013266,3679013266 |
+.long 168166924,168166924 |
+.long 1814307912,1814307912 |
+.long 3831258296,3831258296 |
+.long 1573044895,1573044895 |
+.long 1859376061,1859376061 |
+.long 4021070915,4021070915 |
+.long 2791465668,2791465668 |
+.long 2828112185,2828112185 |
+.long 2761266481,2761266481 |
+.long 937747667,937747667 |
+.long 2339994098,2339994098 |
+.long 854058965,854058965 |
+.long 1137232011,1137232011 |
+.long 1496790894,1496790894 |
+.long 3077402074,3077402074 |
+.long 2358086913,2358086913 |
+.long 1691735473,1691735473 |
+.long 3528347292,3528347292 |
+.long 3769215305,3769215305 |
+.long 3027004632,3027004632 |
+.long 4199962284,4199962284 |
+.long 133494003,133494003 |
+.long 636152527,636152527 |
+.long 2942657994,2942657994 |
+.long 2390391540,2390391540 |
+.long 3920539207,3920539207 |
+.long 403179536,403179536 |
+.long 3585784431,3585784431 |
+.long 2289596656,2289596656 |
+.long 1864705354,1864705354 |
+.long 1915629148,1915629148 |
+.long 605822008,605822008 |
+.long 4054230615,4054230615 |
+.long 3350508659,3350508659 |
+.long 1371981463,1371981463 |
+.long 602466507,602466507 |
+.long 2094914977,2094914977 |
+.long 2624877800,2624877800 |
+.long 555687742,555687742 |
+.long 3712699286,3712699286 |
+.long 3703422305,3703422305 |
+.long 2257292045,2257292045 |
+.long 2240449039,2240449039 |
+.long 2423288032,2423288032 |
+.long 1111375484,1111375484 |
+.long 3300242801,3300242801 |
+.long 2858837708,2858837708 |
+.long 3628615824,3628615824 |
+.long 84083462,84083462 |
+.long 32962295,32962295 |
+.long 302911004,302911004 |
+.long 2741068226,2741068226 |
+.long 1597322602,1597322602 |
+.long 4183250862,4183250862 |
+.long 3501832553,3501832553 |
+.long 2441512471,2441512471 |
+.long 1489093017,1489093017 |
+.long 656219450,656219450 |
+.long 3114180135,3114180135 |
+.long 954327513,954327513 |
+.long 335083755,335083755 |
+.long 3013122091,3013122091 |
+.long 856756514,856756514 |
+.long 3144247762,3144247762 |
+.long 1893325225,1893325225 |
+.long 2307821063,2307821063 |
+.long 2811532339,2811532339 |
+.long 3063651117,3063651117 |
+.long 572399164,572399164 |
+.long 2458355477,2458355477 |
+.long 552200649,552200649 |
+.long 1238290055,1238290055 |
+.long 4283782570,4283782570 |
+.long 2015897680,2015897680 |
+.long 2061492133,2061492133 |
+.long 2408352771,2408352771 |
+.long 4171342169,4171342169 |
+.long 2156497161,2156497161 |
+.long 386731290,386731290 |
+.long 3669999461,3669999461 |
+.long 837215959,837215959 |
+.long 3326231172,3326231172 |
+.long 3093850320,3093850320 |
+.long 3275833730,3275833730 |
+.long 2962856233,2962856233 |
+.long 1999449434,1999449434 |
+.long 286199582,286199582 |
+.long 3417354363,3417354363 |
+.long 4233385128,4233385128 |
+.long 3602627437,3602627437 |
+.long 974525996,974525996 |
+.byte 99,124,119,123,242,107,111,197 |
+.byte 48,1,103,43,254,215,171,118 |
+.byte 202,130,201,125,250,89,71,240 |
+.byte 173,212,162,175,156,164,114,192 |
+.byte 183,253,147,38,54,63,247,204 |
+.byte 52,165,229,241,113,216,49,21 |
+.byte 4,199,35,195,24,150,5,154 |
+.byte 7,18,128,226,235,39,178,117 |
+.byte 9,131,44,26,27,110,90,160 |
+.byte 82,59,214,179,41,227,47,132 |
+.byte 83,209,0,237,32,252,177,91 |
+.byte 106,203,190,57,74,76,88,207 |
+.byte 208,239,170,251,67,77,51,133 |
+.byte 69,249,2,127,80,60,159,168 |
+.byte 81,163,64,143,146,157,56,245 |
+.byte 188,182,218,33,16,255,243,210 |
+.byte 205,12,19,236,95,151,68,23 |
+.byte 196,167,126,61,100,93,25,115 |
+.byte 96,129,79,220,34,42,144,136 |
+.byte 70,238,184,20,222,94,11,219 |
+.byte 224,50,58,10,73,6,36,92 |
+.byte 194,211,172,98,145,149,228,121 |
+.byte 231,200,55,109,141,213,78,169 |
+.byte 108,86,244,234,101,122,174,8 |
+.byte 186,120,37,46,28,166,180,198 |
+.byte 232,221,116,31,75,189,139,138 |
+.byte 112,62,181,102,72,3,246,14 |
+.byte 97,53,87,185,134,193,29,158 |
+.byte 225,248,152,17,105,217,142,148 |
+.byte 155,30,135,233,206,85,40,223 |
+.byte 140,161,137,13,191,230,66,104 |
+.byte 65,153,45,15,176,84,187,22 |
+.byte 99,124,119,123,242,107,111,197 |
+.byte 48,1,103,43,254,215,171,118 |
+.byte 202,130,201,125,250,89,71,240 |
+.byte 173,212,162,175,156,164,114,192 |
+.byte 183,253,147,38,54,63,247,204 |
+.byte 52,165,229,241,113,216,49,21 |
+.byte 4,199,35,195,24,150,5,154 |
+.byte 7,18,128,226,235,39,178,117 |
+.byte 9,131,44,26,27,110,90,160 |
+.byte 82,59,214,179,41,227,47,132 |
+.byte 83,209,0,237,32,252,177,91 |
+.byte 106,203,190,57,74,76,88,207 |
+.byte 208,239,170,251,67,77,51,133 |
+.byte 69,249,2,127,80,60,159,168 |
+.byte 81,163,64,143,146,157,56,245 |
+.byte 188,182,218,33,16,255,243,210 |
+.byte 205,12,19,236,95,151,68,23 |
+.byte 196,167,126,61,100,93,25,115 |
+.byte 96,129,79,220,34,42,144,136 |
+.byte 70,238,184,20,222,94,11,219 |
+.byte 224,50,58,10,73,6,36,92 |
+.byte 194,211,172,98,145,149,228,121 |
+.byte 231,200,55,109,141,213,78,169 |
+.byte 108,86,244,234,101,122,174,8 |
+.byte 186,120,37,46,28,166,180,198 |
+.byte 232,221,116,31,75,189,139,138 |
+.byte 112,62,181,102,72,3,246,14 |
+.byte 97,53,87,185,134,193,29,158 |
+.byte 225,248,152,17,105,217,142,148 |
+.byte 155,30,135,233,206,85,40,223 |
+.byte 140,161,137,13,191,230,66,104 |
+.byte 65,153,45,15,176,84,187,22 |
+.byte 99,124,119,123,242,107,111,197 |
+.byte 48,1,103,43,254,215,171,118 |
+.byte 202,130,201,125,250,89,71,240 |
+.byte 173,212,162,175,156,164,114,192 |
+.byte 183,253,147,38,54,63,247,204 |
+.byte 52,165,229,241,113,216,49,21 |
+.byte 4,199,35,195,24,150,5,154 |
+.byte 7,18,128,226,235,39,178,117 |
+.byte 9,131,44,26,27,110,90,160 |
+.byte 82,59,214,179,41,227,47,132 |
+.byte 83,209,0,237,32,252,177,91 |
+.byte 106,203,190,57,74,76,88,207 |
+.byte 208,239,170,251,67,77,51,133 |
+.byte 69,249,2,127,80,60,159,168 |
+.byte 81,163,64,143,146,157,56,245 |
+.byte 188,182,218,33,16,255,243,210 |
+.byte 205,12,19,236,95,151,68,23 |
+.byte 196,167,126,61,100,93,25,115 |
+.byte 96,129,79,220,34,42,144,136 |
+.byte 70,238,184,20,222,94,11,219 |
+.byte 224,50,58,10,73,6,36,92 |
+.byte 194,211,172,98,145,149,228,121 |
+.byte 231,200,55,109,141,213,78,169 |
+.byte 108,86,244,234,101,122,174,8 |
+.byte 186,120,37,46,28,166,180,198 |
+.byte 232,221,116,31,75,189,139,138 |
+.byte 112,62,181,102,72,3,246,14 |
+.byte 97,53,87,185,134,193,29,158 |
+.byte 225,248,152,17,105,217,142,148 |
+.byte 155,30,135,233,206,85,40,223 |
+.byte 140,161,137,13,191,230,66,104 |
+.byte 65,153,45,15,176,84,187,22 |
+.byte 99,124,119,123,242,107,111,197 |
+.byte 48,1,103,43,254,215,171,118 |
+.byte 202,130,201,125,250,89,71,240 |
+.byte 173,212,162,175,156,164,114,192 |
+.byte 183,253,147,38,54,63,247,204 |
+.byte 52,165,229,241,113,216,49,21 |
+.byte 4,199,35,195,24,150,5,154 |
+.byte 7,18,128,226,235,39,178,117 |
+.byte 9,131,44,26,27,110,90,160 |
+.byte 82,59,214,179,41,227,47,132 |
+.byte 83,209,0,237,32,252,177,91 |
+.byte 106,203,190,57,74,76,88,207 |
+.byte 208,239,170,251,67,77,51,133 |
+.byte 69,249,2,127,80,60,159,168 |
+.byte 81,163,64,143,146,157,56,245 |
+.byte 188,182,218,33,16,255,243,210 |
+.byte 205,12,19,236,95,151,68,23 |
+.byte 196,167,126,61,100,93,25,115 |
+.byte 96,129,79,220,34,42,144,136 |
+.byte 70,238,184,20,222,94,11,219 |
+.byte 224,50,58,10,73,6,36,92 |
+.byte 194,211,172,98,145,149,228,121 |
+.byte 231,200,55,109,141,213,78,169 |
+.byte 108,86,244,234,101,122,174,8 |
+.byte 186,120,37,46,28,166,180,198 |
+.byte 232,221,116,31,75,189,139,138 |
+.byte 112,62,181,102,72,3,246,14 |
+.byte 97,53,87,185,134,193,29,158 |
+.byte 225,248,152,17,105,217,142,148 |
+.byte 155,30,135,233,206,85,40,223 |
+.byte 140,161,137,13,191,230,66,104 |
+.byte 65,153,45,15,176,84,187,22 |
+.long 1,2,4,8 |
+.long 16,32,64,128 |
+.long 27,54,0,0 |
+.long 0,0,0,0 |
+.size _x86_AES_encrypt,.-_x86_AES_encrypt |
+.globl asm_AES_encrypt |
+.hidden asm_AES_encrypt |
+.type asm_AES_encrypt,@function |
+.align 16 |
+asm_AES_encrypt: |
+.L_asm_AES_encrypt_begin: |
+ pushl %ebp |
+ pushl %ebx |
+ pushl %esi |
+ pushl %edi |
+ movl 20(%esp),%esi |
+ movl 28(%esp),%edi |
+ movl %esp,%eax |
+ subl $36,%esp |
+ andl $-64,%esp |
+ leal -127(%edi),%ebx |
+ subl %esp,%ebx |
+ negl %ebx |
+ andl $960,%ebx |
+ subl %ebx,%esp |
+ addl $4,%esp |
+ movl %eax,28(%esp) |
+ call .L004pic_point |
+.L004pic_point: |
+ popl %ebp |
+ leal OPENSSL_ia32cap_P-.L004pic_point(%ebp),%eax |
+ leal .LAES_Te-.L004pic_point(%ebp),%ebp |
+ leal 764(%esp),%ebx |
+ subl %ebp,%ebx |
+ andl $768,%ebx |
+ leal 2176(%ebp,%ebx,1),%ebp |
+ btl $25,(%eax) |
+ jnc .L005x86 |
+ movq (%esi),%mm0 |
+ movq 8(%esi),%mm4 |
+ call _sse_AES_encrypt_compact |
+ movl 28(%esp),%esp |
+ movl 24(%esp),%esi |
+ movq %mm0,(%esi) |
+ movq %mm4,8(%esi) |
+ emms |
+ popl %edi |
+ popl %esi |
+ popl %ebx |
+ popl %ebp |
+ ret |
+.align 16 |
+.L005x86: |
+ movl %ebp,24(%esp) |
+ movl (%esi),%eax |
+ movl 4(%esi),%ebx |
+ movl 8(%esi),%ecx |
+ movl 12(%esi),%edx |
+ call _x86_AES_encrypt_compact |
+ movl 28(%esp),%esp |
+ movl 24(%esp),%esi |
+ movl %eax,(%esi) |
+ movl %ebx,4(%esi) |
+ movl %ecx,8(%esi) |
+ movl %edx,12(%esi) |
+ popl %edi |
+ popl %esi |
+ popl %ebx |
+ popl %ebp |
+ ret |
+.size asm_AES_encrypt,.-.L_asm_AES_encrypt_begin |
+.hidden _x86_AES_decrypt_compact |
+.type _x86_AES_decrypt_compact,@function |
+.align 16 |
+_x86_AES_decrypt_compact: |
+ movl %edi,20(%esp) |
+ xorl (%edi),%eax |
+ xorl 4(%edi),%ebx |
+ xorl 8(%edi),%ecx |
+ xorl 12(%edi),%edx |
+ movl 240(%edi),%esi |
+ leal -2(%esi,%esi,1),%esi |
+ leal (%edi,%esi,8),%esi |
+ movl %esi,24(%esp) |
+ movl -128(%ebp),%edi |
+ movl -96(%ebp),%esi |
+ movl -64(%ebp),%edi |
+ movl -32(%ebp),%esi |
+ movl (%ebp),%edi |
+ movl 32(%ebp),%esi |
+ movl 64(%ebp),%edi |
+ movl 96(%ebp),%esi |
+.align 16 |
+.L006loop: |
+ movl %eax,%esi |
+ andl $255,%esi |
+ movzbl -128(%ebp,%esi,1),%esi |
+ movzbl %dh,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $8,%edi |
+ xorl %edi,%esi |
+ movl %ecx,%edi |
+ shrl $16,%edi |
+ andl $255,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $16,%edi |
+ xorl %edi,%esi |
+ movl %ebx,%edi |
+ shrl $24,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $24,%edi |
+ xorl %edi,%esi |
+ movl %esi,4(%esp) |
+ movl %ebx,%esi |
+ andl $255,%esi |
+ movzbl -128(%ebp,%esi,1),%esi |
+ movzbl %ah,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $8,%edi |
+ xorl %edi,%esi |
+ movl %edx,%edi |
+ shrl $16,%edi |
+ andl $255,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $16,%edi |
+ xorl %edi,%esi |
+ movl %ecx,%edi |
+ shrl $24,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $24,%edi |
+ xorl %edi,%esi |
+ movl %esi,8(%esp) |
+ movl %ecx,%esi |
+ andl $255,%esi |
+ movzbl -128(%ebp,%esi,1),%esi |
+ movzbl %bh,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $8,%edi |
+ xorl %edi,%esi |
+ movl %eax,%edi |
+ shrl $16,%edi |
+ andl $255,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $16,%edi |
+ xorl %edi,%esi |
+ movl %edx,%edi |
+ shrl $24,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $24,%edi |
+ xorl %edi,%esi |
+ andl $255,%edx |
+ movzbl -128(%ebp,%edx,1),%edx |
+ movzbl %ch,%ecx |
+ movzbl -128(%ebp,%ecx,1),%ecx |
+ shll $8,%ecx |
+ xorl %ecx,%edx |
+ movl %esi,%ecx |
+ shrl $16,%ebx |
+ andl $255,%ebx |
+ movzbl -128(%ebp,%ebx,1),%ebx |
+ shll $16,%ebx |
+ xorl %ebx,%edx |
+ shrl $24,%eax |
+ movzbl -128(%ebp,%eax,1),%eax |
+ shll $24,%eax |
+ xorl %eax,%edx |
+ movl $2155905152,%edi |
+ andl %ecx,%edi |
+ movl %edi,%esi |
+ shrl $7,%edi |
+ leal (%ecx,%ecx,1),%eax |
+ subl %edi,%esi |
+ andl $4278124286,%eax |
+ andl $454761243,%esi |
+ xorl %esi,%eax |
+ movl $2155905152,%edi |
+ andl %eax,%edi |
+ movl %edi,%esi |
+ shrl $7,%edi |
+ leal (%eax,%eax,1),%ebx |
+ subl %edi,%esi |
+ andl $4278124286,%ebx |
+ andl $454761243,%esi |
+ xorl %ecx,%eax |
+ xorl %esi,%ebx |
+ movl $2155905152,%edi |
+ andl %ebx,%edi |
+ movl %edi,%esi |
+ shrl $7,%edi |
+ leal (%ebx,%ebx,1),%ebp |
+ subl %edi,%esi |
+ andl $4278124286,%ebp |
+ andl $454761243,%esi |
+ xorl %ecx,%ebx |
+ roll $8,%ecx |
+ xorl %esi,%ebp |
+ xorl %eax,%ecx |
+ xorl %ebp,%eax |
+ xorl %ebx,%ecx |
+ xorl %ebp,%ebx |
+ roll $24,%eax |
+ xorl %ebp,%ecx |
+ roll $16,%ebx |
+ xorl %eax,%ecx |
+ roll $8,%ebp |
+ xorl %ebx,%ecx |
+ movl 4(%esp),%eax |
+ xorl %ebp,%ecx |
+ movl %ecx,12(%esp) |
+ movl $2155905152,%edi |
+ andl %edx,%edi |
+ movl %edi,%esi |
+ shrl $7,%edi |
+ leal (%edx,%edx,1),%ebx |
+ subl %edi,%esi |
+ andl $4278124286,%ebx |
+ andl $454761243,%esi |
+ xorl %esi,%ebx |
+ movl $2155905152,%edi |
+ andl %ebx,%edi |
+ movl %edi,%esi |
+ shrl $7,%edi |
+ leal (%ebx,%ebx,1),%ecx |
+ subl %edi,%esi |
+ andl $4278124286,%ecx |
+ andl $454761243,%esi |
+ xorl %edx,%ebx |
+ xorl %esi,%ecx |
+ movl $2155905152,%edi |
+ andl %ecx,%edi |
+ movl %edi,%esi |
+ shrl $7,%edi |
+ leal (%ecx,%ecx,1),%ebp |
+ subl %edi,%esi |
+ andl $4278124286,%ebp |
+ andl $454761243,%esi |
+ xorl %edx,%ecx |
+ roll $8,%edx |
+ xorl %esi,%ebp |
+ xorl %ebx,%edx |
+ xorl %ebp,%ebx |
+ xorl %ecx,%edx |
+ xorl %ebp,%ecx |
+ roll $24,%ebx |
+ xorl %ebp,%edx |
+ roll $16,%ecx |
+ xorl %ebx,%edx |
+ roll $8,%ebp |
+ xorl %ecx,%edx |
+ movl 8(%esp),%ebx |
+ xorl %ebp,%edx |
+ movl %edx,16(%esp) |
+ movl $2155905152,%edi |
+ andl %eax,%edi |
+ movl %edi,%esi |
+ shrl $7,%edi |
+ leal (%eax,%eax,1),%ecx |
+ subl %edi,%esi |
+ andl $4278124286,%ecx |
+ andl $454761243,%esi |
+ xorl %esi,%ecx |
+ movl $2155905152,%edi |
+ andl %ecx,%edi |
+ movl %edi,%esi |
+ shrl $7,%edi |
+ leal (%ecx,%ecx,1),%edx |
+ subl %edi,%esi |
+ andl $4278124286,%edx |
+ andl $454761243,%esi |
+ xorl %eax,%ecx |
+ xorl %esi,%edx |
+ movl $2155905152,%edi |
+ andl %edx,%edi |
+ movl %edi,%esi |
+ shrl $7,%edi |
+ leal (%edx,%edx,1),%ebp |
+ subl %edi,%esi |
+ andl $4278124286,%ebp |
+ andl $454761243,%esi |
+ xorl %eax,%edx |
+ roll $8,%eax |
+ xorl %esi,%ebp |
+ xorl %ecx,%eax |
+ xorl %ebp,%ecx |
+ xorl %edx,%eax |
+ xorl %ebp,%edx |
+ roll $24,%ecx |
+ xorl %ebp,%eax |
+ roll $16,%edx |
+ xorl %ecx,%eax |
+ roll $8,%ebp |
+ xorl %edx,%eax |
+ xorl %ebp,%eax |
+ movl $2155905152,%edi |
+ andl %ebx,%edi |
+ movl %edi,%esi |
+ shrl $7,%edi |
+ leal (%ebx,%ebx,1),%ecx |
+ subl %edi,%esi |
+ andl $4278124286,%ecx |
+ andl $454761243,%esi |
+ xorl %esi,%ecx |
+ movl $2155905152,%edi |
+ andl %ecx,%edi |
+ movl %edi,%esi |
+ shrl $7,%edi |
+ leal (%ecx,%ecx,1),%edx |
+ subl %edi,%esi |
+ andl $4278124286,%edx |
+ andl $454761243,%esi |
+ xorl %ebx,%ecx |
+ xorl %esi,%edx |
+ movl $2155905152,%edi |
+ andl %edx,%edi |
+ movl %edi,%esi |
+ shrl $7,%edi |
+ leal (%edx,%edx,1),%ebp |
+ subl %edi,%esi |
+ andl $4278124286,%ebp |
+ andl $454761243,%esi |
+ xorl %ebx,%edx |
+ roll $8,%ebx |
+ xorl %esi,%ebp |
+ xorl %ecx,%ebx |
+ xorl %ebp,%ecx |
+ xorl %edx,%ebx |
+ xorl %ebp,%edx |
+ roll $24,%ecx |
+ xorl %ebp,%ebx |
+ roll $16,%edx |
+ xorl %ecx,%ebx |
+ roll $8,%ebp |
+ xorl %edx,%ebx |
+ movl 12(%esp),%ecx |
+ xorl %ebp,%ebx |
+ movl 16(%esp),%edx |
+ movl 20(%esp),%edi |
+ movl 28(%esp),%ebp |
+ addl $16,%edi |
+ xorl (%edi),%eax |
+ xorl 4(%edi),%ebx |
+ xorl 8(%edi),%ecx |
+ xorl 12(%edi),%edx |
+ cmpl 24(%esp),%edi |
+ movl %edi,20(%esp) |
+ jb .L006loop |
+ movl %eax,%esi |
+ andl $255,%esi |
+ movzbl -128(%ebp,%esi,1),%esi |
+ movzbl %dh,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $8,%edi |
+ xorl %edi,%esi |
+ movl %ecx,%edi |
+ shrl $16,%edi |
+ andl $255,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $16,%edi |
+ xorl %edi,%esi |
+ movl %ebx,%edi |
+ shrl $24,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $24,%edi |
+ xorl %edi,%esi |
+ movl %esi,4(%esp) |
+ movl %ebx,%esi |
+ andl $255,%esi |
+ movzbl -128(%ebp,%esi,1),%esi |
+ movzbl %ah,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $8,%edi |
+ xorl %edi,%esi |
+ movl %edx,%edi |
+ shrl $16,%edi |
+ andl $255,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $16,%edi |
+ xorl %edi,%esi |
+ movl %ecx,%edi |
+ shrl $24,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $24,%edi |
+ xorl %edi,%esi |
+ movl %esi,8(%esp) |
+ movl %ecx,%esi |
+ andl $255,%esi |
+ movzbl -128(%ebp,%esi,1),%esi |
+ movzbl %bh,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $8,%edi |
+ xorl %edi,%esi |
+ movl %eax,%edi |
+ shrl $16,%edi |
+ andl $255,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $16,%edi |
+ xorl %edi,%esi |
+ movl %edx,%edi |
+ shrl $24,%edi |
+ movzbl -128(%ebp,%edi,1),%edi |
+ shll $24,%edi |
+ xorl %edi,%esi |
+ movl 20(%esp),%edi |
+ andl $255,%edx |
+ movzbl -128(%ebp,%edx,1),%edx |
+ movzbl %ch,%ecx |
+ movzbl -128(%ebp,%ecx,1),%ecx |
+ shll $8,%ecx |
+ xorl %ecx,%edx |
+ movl %esi,%ecx |
+ shrl $16,%ebx |
+ andl $255,%ebx |
+ movzbl -128(%ebp,%ebx,1),%ebx |
+ shll $16,%ebx |
+ xorl %ebx,%edx |
+ movl 8(%esp),%ebx |
+ shrl $24,%eax |
+ movzbl -128(%ebp,%eax,1),%eax |
+ shll $24,%eax |
+ xorl %eax,%edx |
+ movl 4(%esp),%eax |
+ xorl 16(%edi),%eax |
+ xorl 20(%edi),%ebx |
+ xorl 24(%edi),%ecx |
+ xorl 28(%edi),%edx |
+ ret |
+.size _x86_AES_decrypt_compact,.-_x86_AES_decrypt_compact |
+.hidden _sse_AES_decrypt_compact |
+.type _sse_AES_decrypt_compact,@function |
+.align 16 |
+_sse_AES_decrypt_compact: |
+ pxor (%edi),%mm0 |
+ pxor 8(%edi),%mm4 |
+ movl 240(%edi),%esi |
+ leal -2(%esi,%esi,1),%esi |
+ leal (%edi,%esi,8),%esi |
+ movl %esi,24(%esp) |
+ movl $454761243,%eax |
+ movl %eax,8(%esp) |
+ movl %eax,12(%esp) |
+ movl -128(%ebp),%eax |
+ movl -96(%ebp),%ebx |
+ movl -64(%ebp),%ecx |
+ movl -32(%ebp),%edx |
+ movl (%ebp),%eax |
+ movl 32(%ebp),%ebx |
+ movl 64(%ebp),%ecx |
+ movl 96(%ebp),%edx |
+.align 16 |
+.L007loop: |
+ pshufw $12,%mm0,%mm1 |
+ pshufw $9,%mm4,%mm5 |
+ movd %mm1,%eax |
+ movd %mm5,%ebx |
+ movl %edi,20(%esp) |
+ movzbl %al,%esi |
+ movzbl %ah,%edx |
+ pshufw $6,%mm0,%mm2 |
+ movzbl -128(%ebp,%esi,1),%ecx |
+ movzbl %bl,%edi |
+ movzbl -128(%ebp,%edx,1),%edx |
+ shrl $16,%eax |
+ shll $8,%edx |
+ movzbl -128(%ebp,%edi,1),%esi |
+ movzbl %bh,%edi |
+ shll $16,%esi |
+ pshufw $3,%mm4,%mm6 |
+ orl %esi,%ecx |
+ movzbl -128(%ebp,%edi,1),%esi |
+ movzbl %ah,%edi |
+ shll $24,%esi |
+ shrl $16,%ebx |
+ orl %esi,%edx |
+ movzbl -128(%ebp,%edi,1),%esi |
+ movzbl %bh,%edi |
+ shll $24,%esi |
+ orl %esi,%ecx |
+ movzbl -128(%ebp,%edi,1),%esi |
+ movzbl %al,%edi |
+ shll $8,%esi |
+ movd %mm2,%eax |
+ orl %esi,%ecx |
+ movzbl -128(%ebp,%edi,1),%esi |
+ movzbl %bl,%edi |
+ shll $16,%esi |
+ movd %mm6,%ebx |
+ movd %ecx,%mm0 |
+ movzbl -128(%ebp,%edi,1),%ecx |
+ movzbl %al,%edi |
+ orl %esi,%ecx |
+ movzbl -128(%ebp,%edi,1),%esi |
+ movzbl %bl,%edi |
+ orl %esi,%edx |
+ movzbl -128(%ebp,%edi,1),%esi |
+ movzbl %ah,%edi |
+ shll $16,%esi |
+ shrl $16,%eax |
+ orl %esi,%edx |
+ movzbl -128(%ebp,%edi,1),%esi |
+ movzbl %bh,%edi |
+ shrl $16,%ebx |
+ shll $8,%esi |
+ movd %edx,%mm1 |
+ movzbl -128(%ebp,%edi,1),%edx |
+ movzbl %bh,%edi |
+ shll $24,%edx |
+ andl $255,%ebx |
+ orl %esi,%edx |
+ punpckldq %mm1,%mm0 |
+ movzbl -128(%ebp,%edi,1),%esi |
+ movzbl %al,%edi |
+ shll $8,%esi |
+ movzbl %ah,%eax |
+ movzbl -128(%ebp,%ebx,1),%ebx |
+ orl %esi,%ecx |
+ movzbl -128(%ebp,%edi,1),%esi |
+ orl %ebx,%edx |
+ shll $16,%esi |
+ movzbl -128(%ebp,%eax,1),%eax |
+ orl %esi,%edx |
+ shll $24,%eax |
+ orl %eax,%ecx |
+ movl 20(%esp),%edi |
+ movd %edx,%mm4 |
+ movd %ecx,%mm5 |
+ punpckldq %mm5,%mm4 |
+ addl $16,%edi |
+ cmpl 24(%esp),%edi |
+ ja .L008out |
+ movq %mm0,%mm3 |
+ movq %mm4,%mm7 |
+ pshufw $228,%mm0,%mm2 |
+ pshufw $228,%mm4,%mm6 |
+ movq %mm0,%mm1 |
+ movq %mm4,%mm5 |
+ pshufw $177,%mm0,%mm0 |
+ pshufw $177,%mm4,%mm4 |
+ pslld $8,%mm2 |
+ pslld $8,%mm6 |
+ psrld $8,%mm3 |
+ psrld $8,%mm7 |
+ pxor %mm2,%mm0 |
+ pxor %mm6,%mm4 |
+ pxor %mm3,%mm0 |
+ pxor %mm7,%mm4 |
+ pslld $16,%mm2 |
+ pslld $16,%mm6 |
+ psrld $16,%mm3 |
+ psrld $16,%mm7 |
+ pxor %mm2,%mm0 |
+ pxor %mm6,%mm4 |
+ pxor %mm3,%mm0 |
+ pxor %mm7,%mm4 |
+ movq 8(%esp),%mm3 |
+ pxor %mm2,%mm2 |
+ pxor %mm6,%mm6 |
+ pcmpgtb %mm1,%mm2 |
+ pcmpgtb %mm5,%mm6 |
+ pand %mm3,%mm2 |
+ pand %mm3,%mm6 |
+ paddb %mm1,%mm1 |
+ paddb %mm5,%mm5 |
+ pxor %mm2,%mm1 |
+ pxor %mm6,%mm5 |
+ movq %mm1,%mm3 |
+ movq %mm5,%mm7 |
+ movq %mm1,%mm2 |
+ movq %mm5,%mm6 |
+ pxor %mm1,%mm0 |
+ pxor %mm5,%mm4 |
+ pslld $24,%mm3 |
+ pslld $24,%mm7 |
+ psrld $8,%mm2 |
+ psrld $8,%mm6 |
+ pxor %mm3,%mm0 |
+ pxor %mm7,%mm4 |
+ pxor %mm2,%mm0 |
+ pxor %mm6,%mm4 |
+ movq 8(%esp),%mm2 |
+ pxor %mm3,%mm3 |
+ pxor %mm7,%mm7 |
+ pcmpgtb %mm1,%mm3 |
+ pcmpgtb %mm5,%mm7 |
+ pand %mm2,%mm3 |
+ pand %mm2,%mm7 |
+ paddb %mm1,%mm1 |
+ paddb %mm5,%mm5 |
+ pxor %mm3,%mm1 |
+ pxor %mm7,%mm5 |
+ pshufw $177,%mm1,%mm3 |
+ pshufw $177,%mm5,%mm7 |
+ pxor %mm1,%mm0 |
+ pxor %mm5,%mm4 |
+ pxor %mm3,%mm0 |
+ pxor %mm7,%mm4 |
+ pxor %mm3,%mm3 |
+ pxor %mm7,%mm7 |
+ pcmpgtb %mm1,%mm3 |
+ pcmpgtb %mm5,%mm7 |
+ pand %mm2,%mm3 |
+ pand %mm2,%mm7 |
+ paddb %mm1,%mm1 |
+ paddb %mm5,%mm5 |
+ pxor %mm3,%mm1 |
+ pxor %mm7,%mm5 |
+ pxor %mm1,%mm0 |
+ pxor %mm5,%mm4 |
+ movq %mm1,%mm3 |
+ movq %mm5,%mm7 |
+ pshufw $177,%mm1,%mm2 |
+ pshufw $177,%mm5,%mm6 |
+ pxor %mm2,%mm0 |
+ pxor %mm6,%mm4 |
+ pslld $8,%mm1 |
+ pslld $8,%mm5 |
+ psrld $8,%mm3 |
+ psrld $8,%mm7 |
+ movq (%edi),%mm2 |
+ movq 8(%edi),%mm6 |
+ pxor %mm1,%mm0 |
+ pxor %mm5,%mm4 |
+ pxor %mm3,%mm0 |
+ pxor %mm7,%mm4 |
+ movl -128(%ebp),%eax |
+ pslld $16,%mm1 |
+ pslld $16,%mm5 |
+ movl -64(%ebp),%ebx |
+ psrld $16,%mm3 |
+ psrld $16,%mm7 |
+ movl (%ebp),%ecx |
+ pxor %mm1,%mm0 |
+ pxor %mm5,%mm4 |
+ movl 64(%ebp),%edx |
+ pxor %mm3,%mm0 |
+ pxor %mm7,%mm4 |
+ pxor %mm2,%mm0 |
+ pxor %mm6,%mm4 |
+ jmp .L007loop |
+.align 16 |
+.L008out: |
+ pxor (%edi),%mm0 |
+ pxor 8(%edi),%mm4 |
+ ret |
+.size _sse_AES_decrypt_compact,.-_sse_AES_decrypt_compact |
+.hidden _x86_AES_decrypt |
+.type _x86_AES_decrypt,@function |
+.align 16 |
+_x86_AES_decrypt: |
+ movl %edi,20(%esp) |
+ xorl (%edi),%eax |
+ xorl 4(%edi),%ebx |
+ xorl 8(%edi),%ecx |
+ xorl 12(%edi),%edx |
+ movl 240(%edi),%esi |
+ leal -2(%esi,%esi,1),%esi |
+ leal (%edi,%esi,8),%esi |
+ movl %esi,24(%esp) |
+.align 16 |
+.L009loop: |
+ movl %eax,%esi |
+ andl $255,%esi |
+ movl (%ebp,%esi,8),%esi |
+ movzbl %dh,%edi |
+ xorl 3(%ebp,%edi,8),%esi |
+ movl %ecx,%edi |
+ shrl $16,%edi |
+ andl $255,%edi |
+ xorl 2(%ebp,%edi,8),%esi |
+ movl %ebx,%edi |
+ shrl $24,%edi |
+ xorl 1(%ebp,%edi,8),%esi |
+ movl %esi,4(%esp) |
+ |
+ movl %ebx,%esi |
+ andl $255,%esi |
+ movl (%ebp,%esi,8),%esi |
+ movzbl %ah,%edi |
+ xorl 3(%ebp,%edi,8),%esi |
+ movl %edx,%edi |
+ shrl $16,%edi |
+ andl $255,%edi |
+ xorl 2(%ebp,%edi,8),%esi |
+ movl %ecx,%edi |
+ shrl $24,%edi |
+ xorl 1(%ebp,%edi,8),%esi |
+ movl %esi,8(%esp) |
+ |
+ movl %ecx,%esi |
+ andl $255,%esi |
+ movl (%ebp,%esi,8),%esi |
+ movzbl %bh,%edi |
+ xorl 3(%ebp,%edi,8),%esi |
+ movl %eax,%edi |
+ shrl $16,%edi |
+ andl $255,%edi |
+ xorl 2(%ebp,%edi,8),%esi |
+ movl %edx,%edi |
+ shrl $24,%edi |
+ xorl 1(%ebp,%edi,8),%esi |
+ |
+ movl 20(%esp),%edi |
+ andl $255,%edx |
+ movl (%ebp,%edx,8),%edx |
+ movzbl %ch,%ecx |
+ xorl 3(%ebp,%ecx,8),%edx |
+ movl %esi,%ecx |
+ shrl $16,%ebx |
+ andl $255,%ebx |
+ xorl 2(%ebp,%ebx,8),%edx |
+ movl 8(%esp),%ebx |
+ shrl $24,%eax |
+ xorl 1(%ebp,%eax,8),%edx |
+ movl 4(%esp),%eax |
+ |
+ addl $16,%edi |
+ xorl (%edi),%eax |
+ xorl 4(%edi),%ebx |
+ xorl 8(%edi),%ecx |
+ xorl 12(%edi),%edx |
+ cmpl 24(%esp),%edi |
+ movl %edi,20(%esp) |
+ jb .L009loop |
+ leal 2176(%ebp),%ebp |
+ movl -128(%ebp),%edi |
+ movl -96(%ebp),%esi |
+ movl -64(%ebp),%edi |
+ movl -32(%ebp),%esi |
+ movl (%ebp),%edi |
+ movl 32(%ebp),%esi |
+ movl 64(%ebp),%edi |
+ movl 96(%ebp),%esi |
+ leal -128(%ebp),%ebp |
+ movl %eax,%esi |
+ andl $255,%esi |
+ movzbl (%ebp,%esi,1),%esi |
+ movzbl %dh,%edi |
+ movzbl (%ebp,%edi,1),%edi |
+ shll $8,%edi |
+ xorl %edi,%esi |
+ movl %ecx,%edi |
+ shrl $16,%edi |
+ andl $255,%edi |
+ movzbl (%ebp,%edi,1),%edi |
+ shll $16,%edi |
+ xorl %edi,%esi |
+ movl %ebx,%edi |
+ shrl $24,%edi |
+ movzbl (%ebp,%edi,1),%edi |
+ shll $24,%edi |
+ xorl %edi,%esi |
+ movl %esi,4(%esp) |
+ movl %ebx,%esi |
+ andl $255,%esi |
+ movzbl (%ebp,%esi,1),%esi |
+ movzbl %ah,%edi |
+ movzbl (%ebp,%edi,1),%edi |
+ shll $8,%edi |
+ xorl %edi,%esi |
+ movl %edx,%edi |
+ shrl $16,%edi |
+ andl $255,%edi |
+ movzbl (%ebp,%edi,1),%edi |
+ shll $16,%edi |
+ xorl %edi,%esi |
+ movl %ecx,%edi |
+ shrl $24,%edi |
+ movzbl (%ebp,%edi,1),%edi |
+ shll $24,%edi |
+ xorl %edi,%esi |
+ movl %esi,8(%esp) |
+ movl %ecx,%esi |
+ andl $255,%esi |
+ movzbl (%ebp,%esi,1),%esi |
+ movzbl %bh,%edi |
+ movzbl (%ebp,%edi,1),%edi |
+ shll $8,%edi |
+ xorl %edi,%esi |
+ movl %eax,%edi |
+ shrl $16,%edi |
+ andl $255,%edi |
+ movzbl (%ebp,%edi,1),%edi |
+ shll $16,%edi |
+ xorl %edi,%esi |
+ movl %edx,%edi |
+ shrl $24,%edi |
+ movzbl (%ebp,%edi,1),%edi |
+ shll $24,%edi |
+ xorl %edi,%esi |
+ movl 20(%esp),%edi |
+ andl $255,%edx |
+ movzbl (%ebp,%edx,1),%edx |
+ movzbl %ch,%ecx |
+ movzbl (%ebp,%ecx,1),%ecx |
+ shll $8,%ecx |
+ xorl %ecx,%edx |
+ movl %esi,%ecx |
+ shrl $16,%ebx |
+ andl $255,%ebx |
+ movzbl (%ebp,%ebx,1),%ebx |
+ shll $16,%ebx |
+ xorl %ebx,%edx |
+ movl 8(%esp),%ebx |
+ shrl $24,%eax |
+ movzbl (%ebp,%eax,1),%eax |
+ shll $24,%eax |
+ xorl %eax,%edx |
+ movl 4(%esp),%eax |
+ leal -2048(%ebp),%ebp |
+ addl $16,%edi |
+ xorl (%edi),%eax |
+ xorl 4(%edi),%ebx |
+ xorl 8(%edi),%ecx |
+ xorl 12(%edi),%edx |
+ ret |
+.align 64 |
+.LAES_Td: |
+.long 1353184337,1353184337 |
+.long 1399144830,1399144830 |
+.long 3282310938,3282310938 |
+.long 2522752826,2522752826 |
+.long 3412831035,3412831035 |
+.long 4047871263,4047871263 |
+.long 2874735276,2874735276 |
+.long 2466505547,2466505547 |
+.long 1442459680,1442459680 |
+.long 4134368941,4134368941 |
+.long 2440481928,2440481928 |
+.long 625738485,625738485 |
+.long 4242007375,4242007375 |
+.long 3620416197,3620416197 |
+.long 2151953702,2151953702 |
+.long 2409849525,2409849525 |
+.long 1230680542,1230680542 |
+.long 1729870373,1729870373 |
+.long 2551114309,2551114309 |
+.long 3787521629,3787521629 |
+.long 41234371,41234371 |
+.long 317738113,317738113 |
+.long 2744600205,2744600205 |
+.long 3338261355,3338261355 |
+.long 3881799427,3881799427 |
+.long 2510066197,2510066197 |
+.long 3950669247,3950669247 |
+.long 3663286933,3663286933 |
+.long 763608788,763608788 |
+.long 3542185048,3542185048 |
+.long 694804553,694804553 |
+.long 1154009486,1154009486 |
+.long 1787413109,1787413109 |
+.long 2021232372,2021232372 |
+.long 1799248025,1799248025 |
+.long 3715217703,3715217703 |
+.long 3058688446,3058688446 |
+.long 397248752,397248752 |
+.long 1722556617,1722556617 |
+.long 3023752829,3023752829 |
+.long 407560035,407560035 |
+.long 2184256229,2184256229 |
+.long 1613975959,1613975959 |
+.long 1165972322,1165972322 |
+.long 3765920945,3765920945 |
+.long 2226023355,2226023355 |
+.long 480281086,480281086 |
+.long 2485848313,2485848313 |
+.long 1483229296,1483229296 |
+.long 436028815,436028815 |
+.long 2272059028,2272059028 |
+.long 3086515026,3086515026 |
+.long 601060267,601060267 |
+.long 3791801202,3791801202 |
+.long 1468997603,1468997603 |
+.long 715871590,715871590 |
+.long 120122290,120122290 |
+.long 63092015,63092015 |
+.long 2591802758,2591802758 |
+.long 2768779219,2768779219 |
+.long 4068943920,4068943920 |
+.long 2997206819,2997206819 |
+.long 3127509762,3127509762 |
+.long 1552029421,1552029421 |
+.long 723308426,723308426 |
+.long 2461301159,2461301159 |
+.long 4042393587,4042393587 |
+.long 2715969870,2715969870 |
+.long 3455375973,3455375973 |
+.long 3586000134,3586000134 |
+.long 526529745,526529745 |
+.long 2331944644,2331944644 |
+.long 2639474228,2639474228 |
+.long 2689987490,2689987490 |
+.long 853641733,853641733 |
+.long 1978398372,1978398372 |
+.long 971801355,971801355 |
+.long 2867814464,2867814464 |
+.long 111112542,111112542 |
+.long 1360031421,1360031421 |
+.long 4186579262,4186579262 |
+.long 1023860118,1023860118 |
+.long 2919579357,2919579357 |
+.long 1186850381,1186850381 |
+.long 3045938321,3045938321 |
+.long 90031217,90031217 |
+.long 1876166148,1876166148 |
+.long 4279586912,4279586912 |
+.long 620468249,620468249 |
+.long 2548678102,2548678102 |
+.long 3426959497,3426959497 |
+.long 2006899047,2006899047 |
+.long 3175278768,3175278768 |
+.long 2290845959,2290845959 |
+.long 945494503,945494503 |
+.long 3689859193,3689859193 |
+.long 1191869601,1191869601 |
+.long 3910091388,3910091388 |
+.long 3374220536,3374220536 |
+.long 0,0 |
+.long 2206629897,2206629897 |
+.long 1223502642,1223502642 |
+.long 2893025566,2893025566 |
+.long 1316117100,1316117100 |
+.long 4227796733,4227796733 |
+.long 1446544655,1446544655 |
+.long 517320253,517320253 |
+.long 658058550,658058550 |
+.long 1691946762,1691946762 |
+.long 564550760,564550760 |
+.long 3511966619,3511966619 |
+.long 976107044,976107044 |
+.long 2976320012,2976320012 |
+.long 266819475,266819475 |
+.long 3533106868,3533106868 |
+.long 2660342555,2660342555 |
+.long 1338359936,1338359936 |
+.long 2720062561,2720062561 |
+.long 1766553434,1766553434 |
+.long 370807324,370807324 |
+.long 179999714,179999714 |
+.long 3844776128,3844776128 |
+.long 1138762300,1138762300 |
+.long 488053522,488053522 |
+.long 185403662,185403662 |
+.long 2915535858,2915535858 |
+.long 3114841645,3114841645 |
+.long 3366526484,3366526484 |
+.long 2233069911,2233069911 |
+.long 1275557295,1275557295 |
+.long 3151862254,3151862254 |
+.long 4250959779,4250959779 |
+.long 2670068215,2670068215 |
+.long 3170202204,3170202204 |
+.long 3309004356,3309004356 |
+.long 880737115,880737115 |
+.long 1982415755,1982415755 |
+.long 3703972811,3703972811 |
+.long 1761406390,1761406390 |
+.long 1676797112,1676797112 |
+.long 3403428311,3403428311 |
+.long 277177154,277177154 |
+.long 1076008723,1076008723 |
+.long 538035844,538035844 |
+.long 2099530373,2099530373 |
+.long 4164795346,4164795346 |
+.long 288553390,288553390 |
+.long 1839278535,1839278535 |
+.long 1261411869,1261411869 |
+.long 4080055004,4080055004 |
+.long 3964831245,3964831245 |
+.long 3504587127,3504587127 |
+.long 1813426987,1813426987 |
+.long 2579067049,2579067049 |
+.long 4199060497,4199060497 |
+.long 577038663,577038663 |
+.long 3297574056,3297574056 |
+.long 440397984,440397984 |
+.long 3626794326,3626794326 |
+.long 4019204898,4019204898 |
+.long 3343796615,3343796615 |
+.long 3251714265,3251714265 |
+.long 4272081548,4272081548 |
+.long 906744984,906744984 |
+.long 3481400742,3481400742 |
+.long 685669029,685669029 |
+.long 646887386,646887386 |
+.long 2764025151,2764025151 |
+.long 3835509292,3835509292 |
+.long 227702864,227702864 |
+.long 2613862250,2613862250 |
+.long 1648787028,1648787028 |
+.long 3256061430,3256061430 |
+.long 3904428176,3904428176 |
+.long 1593260334,1593260334 |
+.long 4121936770,4121936770 |
+.long 3196083615,3196083615 |
+.long 2090061929,2090061929 |
+.long 2838353263,2838353263 |
+.long 3004310991,3004310991 |
+.long 999926984,999926984 |
+.long 2809993232,2809993232 |
+.long 1852021992,1852021992 |
+.long 2075868123,2075868123 |
+.long 158869197,158869197 |
+.long 4095236462,4095236462 |
+.long 28809964,28809964 |
+.long 2828685187,2828685187 |
+.long 1701746150,1701746150 |
+.long 2129067946,2129067946 |
+.long 147831841,147831841 |
+.long 3873969647,3873969647 |
+.long 3650873274,3650873274 |
+.long 3459673930,3459673930 |
+.long 3557400554,3557400554 |
+.long 3598495785,3598495785 |
+.long 2947720241,2947720241 |
+.long 824393514,824393514 |
+.long 815048134,815048134 |
+.long 3227951669,3227951669 |
+.long 935087732,935087732 |
+.long 2798289660,2798289660 |
+.long 2966458592,2966458592 |
+.long 366520115,366520115 |
+.long 1251476721,1251476721 |
+.long 4158319681,4158319681 |
+.long 240176511,240176511 |
+.long 804688151,804688151 |
+.long 2379631990,2379631990 |
+.long 1303441219,1303441219 |
+.long 1414376140,1414376140 |
+.long 3741619940,3741619940 |
+.long 3820343710,3820343710 |
+.long 461924940,461924940 |
+.long 3089050817,3089050817 |
+.long 2136040774,2136040774 |
+.long 82468509,82468509 |
+.long 1563790337,1563790337 |
+.long 1937016826,1937016826 |
+.long 776014843,776014843 |
+.long 1511876531,1511876531 |
+.long 1389550482,1389550482 |
+.long 861278441,861278441 |
+.long 323475053,323475053 |
+.long 2355222426,2355222426 |
+.long 2047648055,2047648055 |
+.long 2383738969,2383738969 |
+.long 2302415851,2302415851 |
+.long 3995576782,3995576782 |
+.long 902390199,902390199 |
+.long 3991215329,3991215329 |
+.long 1018251130,1018251130 |
+.long 1507840668,1507840668 |
+.long 1064563285,1064563285 |
+.long 2043548696,2043548696 |
+.long 3208103795,3208103795 |
+.long 3939366739,3939366739 |
+.long 1537932639,1537932639 |
+.long 342834655,342834655 |
+.long 2262516856,2262516856 |
+.long 2180231114,2180231114 |
+.long 1053059257,1053059257 |
+.long 741614648,741614648 |
+.long 1598071746,1598071746 |
+.long 1925389590,1925389590 |
+.long 203809468,203809468 |
+.long 2336832552,2336832552 |
+.long 1100287487,1100287487 |
+.long 1895934009,1895934009 |
+.long 3736275976,3736275976 |
+.long 2632234200,2632234200 |
+.long 2428589668,2428589668 |
+.long 1636092795,1636092795 |
+.long 1890988757,1890988757 |
+.long 1952214088,1952214088 |
+.long 1113045200,1113045200 |
+.byte 82,9,106,213,48,54,165,56 |
+.byte 191,64,163,158,129,243,215,251 |
+.byte 124,227,57,130,155,47,255,135 |
+.byte 52,142,67,68,196,222,233,203 |
+.byte 84,123,148,50,166,194,35,61 |
+.byte 238,76,149,11,66,250,195,78 |
+.byte 8,46,161,102,40,217,36,178 |
+.byte 118,91,162,73,109,139,209,37 |
+.byte 114,248,246,100,134,104,152,22 |
+.byte 212,164,92,204,93,101,182,146 |
+.byte 108,112,72,80,253,237,185,218 |
+.byte 94,21,70,87,167,141,157,132 |
+.byte 144,216,171,0,140,188,211,10 |
+.byte 247,228,88,5,184,179,69,6 |
+.byte 208,44,30,143,202,63,15,2 |
+.byte 193,175,189,3,1,19,138,107 |
+.byte 58,145,17,65,79,103,220,234 |
+.byte 151,242,207,206,240,180,230,115 |
+.byte 150,172,116,34,231,173,53,133 |
+.byte 226,249,55,232,28,117,223,110 |
+.byte 71,241,26,113,29,41,197,137 |
+.byte 111,183,98,14,170,24,190,27 |
+.byte 252,86,62,75,198,210,121,32 |
+.byte 154,219,192,254,120,205,90,244 |
+.byte 31,221,168,51,136,7,199,49 |
+.byte 177,18,16,89,39,128,236,95 |
+.byte 96,81,127,169,25,181,74,13 |
+.byte 45,229,122,159,147,201,156,239 |
+.byte 160,224,59,77,174,42,245,176 |
+.byte 200,235,187,60,131,83,153,97 |
+.byte 23,43,4,126,186,119,214,38 |
+.byte 225,105,20,99,85,33,12,125 |
+.byte 82,9,106,213,48,54,165,56 |
+.byte 191,64,163,158,129,243,215,251 |
+.byte 124,227,57,130,155,47,255,135 |
+.byte 52,142,67,68,196,222,233,203 |
+.byte 84,123,148,50,166,194,35,61 |
+.byte 238,76,149,11,66,250,195,78 |
+.byte 8,46,161,102,40,217,36,178 |
+.byte 118,91,162,73,109,139,209,37 |
+.byte 114,248,246,100,134,104,152,22 |
+.byte 212,164,92,204,93,101,182,146 |
+.byte 108,112,72,80,253,237,185,218 |
+.byte 94,21,70,87,167,141,157,132 |
+.byte 144,216,171,0,140,188,211,10 |
+.byte 247,228,88,5,184,179,69,6 |
+.byte 208,44,30,143,202,63,15,2 |
+.byte 193,175,189,3,1,19,138,107 |
+.byte 58,145,17,65,79,103,220,234 |
+.byte 151,242,207,206,240,180,230,115 |
+.byte 150,172,116,34,231,173,53,133 |
+.byte 226,249,55,232,28,117,223,110 |
+.byte 71,241,26,113,29,41,197,137 |
+.byte 111,183,98,14,170,24,190,27 |
+.byte 252,86,62,75,198,210,121,32 |
+.byte 154,219,192,254,120,205,90,244 |
+.byte 31,221,168,51,136,7,199,49 |
+.byte 177,18,16,89,39,128,236,95 |
+.byte 96,81,127,169,25,181,74,13 |
+.byte 45,229,122,159,147,201,156,239 |
+.byte 160,224,59,77,174,42,245,176 |
+.byte 200,235,187,60,131,83,153,97 |
+.byte 23,43,4,126,186,119,214,38 |
+.byte 225,105,20,99,85,33,12,125 |
+.byte 82,9,106,213,48,54,165,56 |
+.byte 191,64,163,158,129,243,215,251 |
+.byte 124,227,57,130,155,47,255,135 |
+.byte 52,142,67,68,196,222,233,203 |
+.byte 84,123,148,50,166,194,35,61 |
+.byte 238,76,149,11,66,250,195,78 |
+.byte 8,46,161,102,40,217,36,178 |
+.byte 118,91,162,73,109,139,209,37 |
+.byte 114,248,246,100,134,104,152,22 |
+.byte 212,164,92,204,93,101,182,146 |
+.byte 108,112,72,80,253,237,185,218 |
+.byte 94,21,70,87,167,141,157,132 |
+.byte 144,216,171,0,140,188,211,10 |
+.byte 247,228,88,5,184,179,69,6 |
+.byte 208,44,30,143,202,63,15,2 |
+.byte 193,175,189,3,1,19,138,107 |
+.byte 58,145,17,65,79,103,220,234 |
+.byte 151,242,207,206,240,180,230,115 |
+.byte 150,172,116,34,231,173,53,133 |
+.byte 226,249,55,232,28,117,223,110 |
+.byte 71,241,26,113,29,41,197,137 |
+.byte 111,183,98,14,170,24,190,27 |
+.byte 252,86,62,75,198,210,121,32 |
+.byte 154,219,192,254,120,205,90,244 |
+.byte 31,221,168,51,136,7,199,49 |
+.byte 177,18,16,89,39,128,236,95 |
+.byte 96,81,127,169,25,181,74,13 |
+.byte 45,229,122,159,147,201,156,239 |
+.byte 160,224,59,77,174,42,245,176 |
+.byte 200,235,187,60,131,83,153,97 |
+.byte 23,43,4,126,186,119,214,38 |
+.byte 225,105,20,99,85,33,12,125 |
+.byte 82,9,106,213,48,54,165,56 |
+.byte 191,64,163,158,129,243,215,251 |
+.byte 124,227,57,130,155,47,255,135 |
+.byte 52,142,67,68,196,222,233,203 |
+.byte 84,123,148,50,166,194,35,61 |
+.byte 238,76,149,11,66,250,195,78 |
+.byte 8,46,161,102,40,217,36,178 |
+.byte 118,91,162,73,109,139,209,37 |
+.byte 114,248,246,100,134,104,152,22 |
+.byte 212,164,92,204,93,101,182,146 |
+.byte 108,112,72,80,253,237,185,218 |
+.byte 94,21,70,87,167,141,157,132 |
+.byte 144,216,171,0,140,188,211,10 |
+.byte 247,228,88,5,184,179,69,6 |
+.byte 208,44,30,143,202,63,15,2 |
+.byte 193,175,189,3,1,19,138,107 |
+.byte 58,145,17,65,79,103,220,234 |
+.byte 151,242,207,206,240,180,230,115 |
+.byte 150,172,116,34,231,173,53,133 |
+.byte 226,249,55,232,28,117,223,110 |
+.byte 71,241,26,113,29,41,197,137 |
+.byte 111,183,98,14,170,24,190,27 |
+.byte 252,86,62,75,198,210,121,32 |
+.byte 154,219,192,254,120,205,90,244 |
+.byte 31,221,168,51,136,7,199,49 |
+.byte 177,18,16,89,39,128,236,95 |
+.byte 96,81,127,169,25,181,74,13 |
+.byte 45,229,122,159,147,201,156,239 |
+.byte 160,224,59,77,174,42,245,176 |
+.byte 200,235,187,60,131,83,153,97 |
+.byte 23,43,4,126,186,119,214,38 |
+.byte 225,105,20,99,85,33,12,125 |
+.size _x86_AES_decrypt,.-_x86_AES_decrypt |
+.globl asm_AES_decrypt |
+.hidden asm_AES_decrypt |
+.type asm_AES_decrypt,@function |
+.align 16 |
+asm_AES_decrypt: |
+.L_asm_AES_decrypt_begin: |
+ pushl %ebp |
+ pushl %ebx |
+ pushl %esi |
+ pushl %edi |
+ movl 20(%esp),%esi |
+ movl 28(%esp),%edi |
+ movl %esp,%eax |
+ subl $36,%esp |
+ andl $-64,%esp |
+ leal -127(%edi),%ebx |
+ subl %esp,%ebx |
+ negl %ebx |
+ andl $960,%ebx |
+ subl %ebx,%esp |
+ addl $4,%esp |
+ movl %eax,28(%esp) |
+ call .L010pic_point |
+.L010pic_point: |
+ popl %ebp |
+ leal OPENSSL_ia32cap_P-.L010pic_point(%ebp),%eax |
+ leal .LAES_Td-.L010pic_point(%ebp),%ebp |
+ leal 764(%esp),%ebx |
+ subl %ebp,%ebx |
+ andl $768,%ebx |
+ leal 2176(%ebp,%ebx,1),%ebp |
+ btl $25,(%eax) |
+ jnc .L011x86 |
+ movq (%esi),%mm0 |
+ movq 8(%esi),%mm4 |
+ call _sse_AES_decrypt_compact |
+ movl 28(%esp),%esp |
+ movl 24(%esp),%esi |
+ movq %mm0,(%esi) |
+ movq %mm4,8(%esi) |
+ emms |
+ popl %edi |
+ popl %esi |
+ popl %ebx |
+ popl %ebp |
+ ret |
+.align 16 |
+.L011x86: |
+ movl %ebp,24(%esp) |
+ movl (%esi),%eax |
+ movl 4(%esi),%ebx |
+ movl 8(%esi),%ecx |
+ movl 12(%esi),%edx |
+ call _x86_AES_decrypt_compact |
+ movl 28(%esp),%esp |
+ movl 24(%esp),%esi |
+ movl %eax,(%esi) |
+ movl %ebx,4(%esi) |
+ movl %ecx,8(%esi) |
+ movl %edx,12(%esi) |
+ popl %edi |
+ popl %esi |
+ popl %ebx |
+ popl %ebp |
+ ret |
+.size asm_AES_decrypt,.-.L_asm_AES_decrypt_begin |
+.globl asm_AES_cbc_encrypt |
+.hidden asm_AES_cbc_encrypt |
+.type asm_AES_cbc_encrypt,@function |
+.align 16 |
+asm_AES_cbc_encrypt: |
+.L_asm_AES_cbc_encrypt_begin: |
+ pushl %ebp |
+ pushl %ebx |
+ pushl %esi |
+ pushl %edi |
+ movl 28(%esp),%ecx |
+ cmpl $0,%ecx |
+ je .L012drop_out |
+ call .L013pic_point |
+.L013pic_point: |
+ popl %ebp |
+ leal OPENSSL_ia32cap_P-.L013pic_point(%ebp),%eax |
+ cmpl $0,40(%esp) |
+ leal .LAES_Te-.L013pic_point(%ebp),%ebp |
+ jne .L014picked_te |
+ leal .LAES_Td-.LAES_Te(%ebp),%ebp |
+.L014picked_te: |
+ pushfl |
+ cld |
+ cmpl $512,%ecx |
+ jb .L015slow_way |
+ testl $15,%ecx |
+ jnz .L015slow_way |
+ btl $28,(%eax) |
+ jc .L015slow_way |
+ leal -324(%esp),%esi |
+ andl $-64,%esi |
+ movl %ebp,%eax |
+ leal 2304(%ebp),%ebx |
+ movl %esi,%edx |
+ andl $4095,%eax |
+ andl $4095,%ebx |
+ andl $4095,%edx |
+ cmpl %ebx,%edx |
+ jb .L016tbl_break_out |
+ subl %ebx,%edx |
+ subl %edx,%esi |
+ jmp .L017tbl_ok |
+.align 4 |
+.L016tbl_break_out: |
+ subl %eax,%edx |
+ andl $4095,%edx |
+ addl $384,%edx |
+ subl %edx,%esi |
+.align 4 |
+.L017tbl_ok: |
+ leal 24(%esp),%edx |
+ xchgl %esi,%esp |
+ addl $4,%esp |
+ movl %ebp,24(%esp) |
+ movl %esi,28(%esp) |
+ movl (%edx),%eax |
+ movl 4(%edx),%ebx |
+ movl 12(%edx),%edi |
+ movl 16(%edx),%esi |
+ movl 20(%edx),%edx |
+ movl %eax,32(%esp) |
+ movl %ebx,36(%esp) |
+ movl %ecx,40(%esp) |
+ movl %edi,44(%esp) |
+ movl %esi,48(%esp) |
+ movl $0,316(%esp) |
+ movl %edi,%ebx |
+ movl $61,%ecx |
+ subl %ebp,%ebx |
+ movl %edi,%esi |
+ andl $4095,%ebx |
+ leal 76(%esp),%edi |
+ cmpl $2304,%ebx |
+ jb .L018do_copy |
+ cmpl $3852,%ebx |
+ jb .L019skip_copy |
+.align 4 |
+.L018do_copy: |
+ movl %edi,44(%esp) |
+.long 2784229001 |
+.L019skip_copy: |
+ movl $16,%edi |
+.align 4 |
+.L020prefetch_tbl: |
+ movl (%ebp),%eax |
+ movl 32(%ebp),%ebx |
+ movl 64(%ebp),%ecx |
+ movl 96(%ebp),%esi |
+ leal 128(%ebp),%ebp |
+ subl $1,%edi |
+ jnz .L020prefetch_tbl |
+ subl $2048,%ebp |
+ movl 32(%esp),%esi |
+ movl 48(%esp),%edi |
+ cmpl $0,%edx |
+ je .L021fast_decrypt |
+ movl (%edi),%eax |
+ movl 4(%edi),%ebx |
+.align 16 |
+.L022fast_enc_loop: |
+ movl 8(%edi),%ecx |
+ movl 12(%edi),%edx |
+ xorl (%esi),%eax |
+ xorl 4(%esi),%ebx |
+ xorl 8(%esi),%ecx |
+ xorl 12(%esi),%edx |
+ movl 44(%esp),%edi |
+ call _x86_AES_encrypt |
+ movl 32(%esp),%esi |
+ movl 36(%esp),%edi |
+ movl %eax,(%edi) |
+ movl %ebx,4(%edi) |
+ movl %ecx,8(%edi) |
+ movl %edx,12(%edi) |
+ leal 16(%esi),%esi |
+ movl 40(%esp),%ecx |
+ movl %esi,32(%esp) |
+ leal 16(%edi),%edx |
+ movl %edx,36(%esp) |
+ subl $16,%ecx |
+ movl %ecx,40(%esp) |
+ jnz .L022fast_enc_loop |
+ movl 48(%esp),%esi |
+ movl 8(%edi),%ecx |
+ movl 12(%edi),%edx |
+ movl %eax,(%esi) |
+ movl %ebx,4(%esi) |
+ movl %ecx,8(%esi) |
+ movl %edx,12(%esi) |
+ cmpl $0,316(%esp) |
+ movl 44(%esp),%edi |
+ je .L023skip_ezero |
+ movl $60,%ecx |
+ xorl %eax,%eax |
+.align 4 |
+.long 2884892297 |
+.L023skip_ezero: |
+ movl 28(%esp),%esp |
+ popfl |
+.L012drop_out: |
+ popl %edi |
+ popl %esi |
+ popl %ebx |
+ popl %ebp |
+ ret |
+ pushfl |
+.align 16 |
+.L021fast_decrypt: |
+ cmpl 36(%esp),%esi |
+ je .L024fast_dec_in_place |
+ movl %edi,52(%esp) |
+.align 4 |
+.align 16 |
+.L025fast_dec_loop: |
+ movl (%esi),%eax |
+ movl 4(%esi),%ebx |
+ movl 8(%esi),%ecx |
+ movl 12(%esi),%edx |
+ movl 44(%esp),%edi |
+ call _x86_AES_decrypt |
+ movl 52(%esp),%edi |
+ movl 40(%esp),%esi |
+ xorl (%edi),%eax |
+ xorl 4(%edi),%ebx |
+ xorl 8(%edi),%ecx |
+ xorl 12(%edi),%edx |
+ movl 36(%esp),%edi |
+ movl 32(%esp),%esi |
+ movl %eax,(%edi) |
+ movl %ebx,4(%edi) |
+ movl %ecx,8(%edi) |
+ movl %edx,12(%edi) |
+ movl 40(%esp),%ecx |
+ movl %esi,52(%esp) |
+ leal 16(%esi),%esi |
+ movl %esi,32(%esp) |
+ leal 16(%edi),%edi |
+ movl %edi,36(%esp) |
+ subl $16,%ecx |
+ movl %ecx,40(%esp) |
+ jnz .L025fast_dec_loop |
+ movl 52(%esp),%edi |
+ movl 48(%esp),%esi |
+ movl (%edi),%eax |
+ movl 4(%edi),%ebx |
+ movl 8(%edi),%ecx |
+ movl 12(%edi),%edx |
+ movl %eax,(%esi) |
+ movl %ebx,4(%esi) |
+ movl %ecx,8(%esi) |
+ movl %edx,12(%esi) |
+ jmp .L026fast_dec_out |
+.align 16 |
+.L024fast_dec_in_place: |
+.L027fast_dec_in_place_loop: |
+ movl (%esi),%eax |
+ movl 4(%esi),%ebx |
+ movl 8(%esi),%ecx |
+ movl 12(%esi),%edx |
+ leal 60(%esp),%edi |
+ movl %eax,(%edi) |
+ movl %ebx,4(%edi) |
+ movl %ecx,8(%edi) |
+ movl %edx,12(%edi) |
+ movl 44(%esp),%edi |
+ call _x86_AES_decrypt |
+ movl 48(%esp),%edi |
+ movl 36(%esp),%esi |
+ xorl (%edi),%eax |
+ xorl 4(%edi),%ebx |
+ xorl 8(%edi),%ecx |
+ xorl 12(%edi),%edx |
+ movl %eax,(%esi) |
+ movl %ebx,4(%esi) |
+ movl %ecx,8(%esi) |
+ movl %edx,12(%esi) |
+ leal 16(%esi),%esi |
+ movl %esi,36(%esp) |
+ leal 60(%esp),%esi |
+ movl (%esi),%eax |
+ movl 4(%esi),%ebx |
+ movl 8(%esi),%ecx |
+ movl 12(%esi),%edx |
+ movl %eax,(%edi) |
+ movl %ebx,4(%edi) |
+ movl %ecx,8(%edi) |
+ movl %edx,12(%edi) |
+ movl 32(%esp),%esi |
+ movl 40(%esp),%ecx |
+ leal 16(%esi),%esi |
+ movl %esi,32(%esp) |
+ subl $16,%ecx |
+ movl %ecx,40(%esp) |
+ jnz .L027fast_dec_in_place_loop |
+.align 4 |
+.L026fast_dec_out: |
+ cmpl $0,316(%esp) |
+ movl 44(%esp),%edi |
+ je .L028skip_dzero |
+ movl $60,%ecx |
+ xorl %eax,%eax |
+.align 4 |
+.long 2884892297 |
+.L028skip_dzero: |
+ movl 28(%esp),%esp |
+ popfl |
+ popl %edi |
+ popl %esi |
+ popl %ebx |
+ popl %ebp |
+ ret |
+ pushfl |
+.align 16 |
+.L015slow_way: |
+ movl (%eax),%eax |
+ movl 36(%esp),%edi |
+ leal -80(%esp),%esi |
+ andl $-64,%esi |
+ leal -143(%edi),%ebx |
+ subl %esi,%ebx |
+ negl %ebx |
+ andl $960,%ebx |
+ subl %ebx,%esi |
+ leal 768(%esi),%ebx |
+ subl %ebp,%ebx |
+ andl $768,%ebx |
+ leal 2176(%ebp,%ebx,1),%ebp |
+ leal 24(%esp),%edx |
+ xchgl %esi,%esp |
+ addl $4,%esp |
+ movl %ebp,24(%esp) |
+ movl %esi,28(%esp) |
+ movl %eax,52(%esp) |
+ movl (%edx),%eax |
+ movl 4(%edx),%ebx |
+ movl 16(%edx),%esi |
+ movl 20(%edx),%edx |
+ movl %eax,32(%esp) |
+ movl %ebx,36(%esp) |
+ movl %ecx,40(%esp) |
+ movl %edi,44(%esp) |
+ movl %esi,48(%esp) |
+ movl %esi,%edi |
+ movl %eax,%esi |
+ cmpl $0,%edx |
+ je .L029slow_decrypt |
+ cmpl $16,%ecx |
+ movl %ebx,%edx |
+ jb .L030slow_enc_tail |
+ btl $25,52(%esp) |
+ jnc .L031slow_enc_x86 |
+ movq (%edi),%mm0 |
+ movq 8(%edi),%mm4 |
+.align 16 |
+.L032slow_enc_loop_sse: |
+ pxor (%esi),%mm0 |
+ pxor 8(%esi),%mm4 |
+ movl 44(%esp),%edi |
+ call _sse_AES_encrypt_compact |
+ movl 32(%esp),%esi |
+ movl 36(%esp),%edi |
+ movl 40(%esp),%ecx |
+ movq %mm0,(%edi) |
+ movq %mm4,8(%edi) |
+ leal 16(%esi),%esi |
+ movl %esi,32(%esp) |
+ leal 16(%edi),%edx |
+ movl %edx,36(%esp) |
+ subl $16,%ecx |
+ cmpl $16,%ecx |
+ movl %ecx,40(%esp) |
+ jae .L032slow_enc_loop_sse |
+ testl $15,%ecx |
+ jnz .L030slow_enc_tail |
+ movl 48(%esp),%esi |
+ movq %mm0,(%esi) |
+ movq %mm4,8(%esi) |
+ emms |
+ movl 28(%esp),%esp |
+ popfl |
+ popl %edi |
+ popl %esi |
+ popl %ebx |
+ popl %ebp |
+ ret |
+ pushfl |
+.align 16 |
+.L031slow_enc_x86: |
+ movl (%edi),%eax |
+ movl 4(%edi),%ebx |
+.align 4 |
+.L033slow_enc_loop_x86: |
+ movl 8(%edi),%ecx |
+ movl 12(%edi),%edx |
+ xorl (%esi),%eax |
+ xorl 4(%esi),%ebx |
+ xorl 8(%esi),%ecx |
+ xorl 12(%esi),%edx |
+ movl 44(%esp),%edi |
+ call _x86_AES_encrypt_compact |
+ movl 32(%esp),%esi |
+ movl 36(%esp),%edi |
+ movl %eax,(%edi) |
+ movl %ebx,4(%edi) |
+ movl %ecx,8(%edi) |
+ movl %edx,12(%edi) |
+ movl 40(%esp),%ecx |
+ leal 16(%esi),%esi |
+ movl %esi,32(%esp) |
+ leal 16(%edi),%edx |
+ movl %edx,36(%esp) |
+ subl $16,%ecx |
+ cmpl $16,%ecx |
+ movl %ecx,40(%esp) |
+ jae .L033slow_enc_loop_x86 |
+ testl $15,%ecx |
+ jnz .L030slow_enc_tail |
+ movl 48(%esp),%esi |
+ movl 8(%edi),%ecx |
+ movl 12(%edi),%edx |
+ movl %eax,(%esi) |
+ movl %ebx,4(%esi) |
+ movl %ecx,8(%esi) |
+ movl %edx,12(%esi) |
+ movl 28(%esp),%esp |
+ popfl |
+ popl %edi |
+ popl %esi |
+ popl %ebx |
+ popl %ebp |
+ ret |
+ pushfl |
+.align 16 |
+.L030slow_enc_tail: |
+ emms |
+ movl %edx,%edi |
+ movl $16,%ebx |
+ subl %ecx,%ebx |
+ cmpl %esi,%edi |
+ je .L034enc_in_place |
+.align 4 |
+.long 2767451785 |
+ jmp .L035enc_skip_in_place |
+.L034enc_in_place: |
+ leal (%edi,%ecx,1),%edi |
+.L035enc_skip_in_place: |
+ movl %ebx,%ecx |
+ xorl %eax,%eax |
+.align 4 |
+.long 2868115081 |
+ movl 48(%esp),%edi |
+ movl %edx,%esi |
+ movl (%edi),%eax |
+ movl 4(%edi),%ebx |
+ movl $16,40(%esp) |
+ jmp .L033slow_enc_loop_x86 |
+.align 16 |
+.L029slow_decrypt: |
+ btl $25,52(%esp) |
+ jnc .L036slow_dec_loop_x86 |
+.align 4 |
+.L037slow_dec_loop_sse: |
+ movq (%esi),%mm0 |
+ movq 8(%esi),%mm4 |
+ movl 44(%esp),%edi |
+ call _sse_AES_decrypt_compact |
+ movl 32(%esp),%esi |
+ leal 60(%esp),%eax |
+ movl 36(%esp),%ebx |
+ movl 40(%esp),%ecx |
+ movl 48(%esp),%edi |
+ movq (%esi),%mm1 |
+ movq 8(%esi),%mm5 |
+ pxor (%edi),%mm0 |
+ pxor 8(%edi),%mm4 |
+ movq %mm1,(%edi) |
+ movq %mm5,8(%edi) |
+ subl $16,%ecx |
+ jc .L038slow_dec_partial_sse |
+ movq %mm0,(%ebx) |
+ movq %mm4,8(%ebx) |
+ leal 16(%ebx),%ebx |
+ movl %ebx,36(%esp) |
+ leal 16(%esi),%esi |
+ movl %esi,32(%esp) |
+ movl %ecx,40(%esp) |
+ jnz .L037slow_dec_loop_sse |
+ emms |
+ movl 28(%esp),%esp |
+ popfl |
+ popl %edi |
+ popl %esi |
+ popl %ebx |
+ popl %ebp |
+ ret |
+ pushfl |
+.align 16 |
+.L038slow_dec_partial_sse: |
+ movq %mm0,(%eax) |
+ movq %mm4,8(%eax) |
+ emms |
+ addl $16,%ecx |
+ movl %ebx,%edi |
+ movl %eax,%esi |
+.align 4 |
+.long 2767451785 |
+ movl 28(%esp),%esp |
+ popfl |
+ popl %edi |
+ popl %esi |
+ popl %ebx |
+ popl %ebp |
+ ret |
+ pushfl |
+.align 16 |
+.L036slow_dec_loop_x86: |
+ movl (%esi),%eax |
+ movl 4(%esi),%ebx |
+ movl 8(%esi),%ecx |
+ movl 12(%esi),%edx |
+ leal 60(%esp),%edi |
+ movl %eax,(%edi) |
+ movl %ebx,4(%edi) |
+ movl %ecx,8(%edi) |
+ movl %edx,12(%edi) |
+ movl 44(%esp),%edi |
+ call _x86_AES_decrypt_compact |
+ movl 48(%esp),%edi |
+ movl 40(%esp),%esi |
+ xorl (%edi),%eax |
+ xorl 4(%edi),%ebx |
+ xorl 8(%edi),%ecx |
+ xorl 12(%edi),%edx |
+ subl $16,%esi |
+ jc .L039slow_dec_partial_x86 |
+ movl %esi,40(%esp) |
+ movl 36(%esp),%esi |
+ movl %eax,(%esi) |
+ movl %ebx,4(%esi) |
+ movl %ecx,8(%esi) |
+ movl %edx,12(%esi) |
+ leal 16(%esi),%esi |
+ movl %esi,36(%esp) |
+ leal 60(%esp),%esi |
+ movl (%esi),%eax |
+ movl 4(%esi),%ebx |
+ movl 8(%esi),%ecx |
+ movl 12(%esi),%edx |
+ movl %eax,(%edi) |
+ movl %ebx,4(%edi) |
+ movl %ecx,8(%edi) |
+ movl %edx,12(%edi) |
+ movl 32(%esp),%esi |
+ leal 16(%esi),%esi |
+ movl %esi,32(%esp) |
+ jnz .L036slow_dec_loop_x86 |
+ movl 28(%esp),%esp |
+ popfl |
+ popl %edi |
+ popl %esi |
+ popl %ebx |
+ popl %ebp |
+ ret |
+ pushfl |
+.align 16 |
+.L039slow_dec_partial_x86: |
+ leal 60(%esp),%esi |
+ movl %eax,(%esi) |
+ movl %ebx,4(%esi) |
+ movl %ecx,8(%esi) |
+ movl %edx,12(%esi) |
+ movl 32(%esp),%esi |
+ movl (%esi),%eax |
+ movl 4(%esi),%ebx |
+ movl 8(%esi),%ecx |
+ movl 12(%esi),%edx |
+ movl %eax,(%edi) |
+ movl %ebx,4(%edi) |
+ movl %ecx,8(%edi) |
+ movl %edx,12(%edi) |
+ movl 40(%esp),%ecx |
+ movl 36(%esp),%edi |
+ leal 60(%esp),%esi |
+.align 4 |
+.long 2767451785 |
+ movl 28(%esp),%esp |
+ popfl |
+ popl %edi |
+ popl %esi |
+ popl %ebx |
+ popl %ebp |
+ ret |
+.size asm_AES_cbc_encrypt,.-.L_asm_AES_cbc_encrypt_begin |
+.hidden _x86_AES_set_encrypt_key |
+.type _x86_AES_set_encrypt_key,@function |
+.align 16 |
+_x86_AES_set_encrypt_key: |
+ pushl %ebp |
+ pushl %ebx |
+ pushl %esi |
+ pushl %edi |
+ movl 24(%esp),%esi |
+ movl 32(%esp),%edi |
+ testl $-1,%esi |
+ jz .L040badpointer |
+ testl $-1,%edi |
+ jz .L040badpointer |
+ call .L041pic_point |
+.L041pic_point: |
+ popl %ebp |
+ leal .LAES_Te-.L041pic_point(%ebp),%ebp |
+ leal 2176(%ebp),%ebp |
+ movl -128(%ebp),%eax |
+ movl -96(%ebp),%ebx |
+ movl -64(%ebp),%ecx |
+ movl -32(%ebp),%edx |
+ movl (%ebp),%eax |
+ movl 32(%ebp),%ebx |
+ movl 64(%ebp),%ecx |
+ movl 96(%ebp),%edx |
+ movl 28(%esp),%ecx |
+ cmpl $128,%ecx |
+ je .L04210rounds |
+ cmpl $192,%ecx |
+ je .L04312rounds |
+ cmpl $256,%ecx |
+ je .L04414rounds |
+ movl $-2,%eax |
+ jmp .L045exit |
+.L04210rounds: |
+ movl (%esi),%eax |
+ movl 4(%esi),%ebx |
+ movl 8(%esi),%ecx |
+ movl 12(%esi),%edx |
+ movl %eax,(%edi) |
+ movl %ebx,4(%edi) |
+ movl %ecx,8(%edi) |
+ movl %edx,12(%edi) |
+ xorl %ecx,%ecx |
+ jmp .L04610shortcut |
+.align 4 |
+.L04710loop: |
+ movl (%edi),%eax |
+ movl 12(%edi),%edx |
+.L04610shortcut: |
+ movzbl %dl,%esi |
+ movzbl -128(%ebp,%esi,1),%ebx |
+ movzbl %dh,%esi |
+ shll $24,%ebx |
+ xorl %ebx,%eax |
+ movzbl -128(%ebp,%esi,1),%ebx |
+ shrl $16,%edx |
+ movzbl %dl,%esi |
+ xorl %ebx,%eax |
+ movzbl -128(%ebp,%esi,1),%ebx |
+ movzbl %dh,%esi |
+ shll $8,%ebx |
+ xorl %ebx,%eax |
+ movzbl -128(%ebp,%esi,1),%ebx |
+ shll $16,%ebx |
+ xorl %ebx,%eax |
+ xorl 896(%ebp,%ecx,4),%eax |
+ movl %eax,16(%edi) |
+ xorl 4(%edi),%eax |
+ movl %eax,20(%edi) |
+ xorl 8(%edi),%eax |
+ movl %eax,24(%edi) |
+ xorl 12(%edi),%eax |
+ movl %eax,28(%edi) |
+ incl %ecx |
+ addl $16,%edi |
+ cmpl $10,%ecx |
+ jl .L04710loop |
+ movl $10,80(%edi) |
+ xorl %eax,%eax |
+ jmp .L045exit |
+.L04312rounds: |
+ movl (%esi),%eax |
+ movl 4(%esi),%ebx |
+ movl 8(%esi),%ecx |
+ movl 12(%esi),%edx |
+ movl %eax,(%edi) |
+ movl %ebx,4(%edi) |
+ movl %ecx,8(%edi) |
+ movl %edx,12(%edi) |
+ movl 16(%esi),%ecx |
+ movl 20(%esi),%edx |
+ movl %ecx,16(%edi) |
+ movl %edx,20(%edi) |
+ xorl %ecx,%ecx |
+ jmp .L04812shortcut |
+.align 4 |
+.L04912loop: |
+ movl (%edi),%eax |
+ movl 20(%edi),%edx |
+.L04812shortcut: |
+ movzbl %dl,%esi |
+ movzbl -128(%ebp,%esi,1),%ebx |
+ movzbl %dh,%esi |
+ shll $24,%ebx |
+ xorl %ebx,%eax |
+ movzbl -128(%ebp,%esi,1),%ebx |
+ shrl $16,%edx |
+ movzbl %dl,%esi |
+ xorl %ebx,%eax |
+ movzbl -128(%ebp,%esi,1),%ebx |
+ movzbl %dh,%esi |
+ shll $8,%ebx |
+ xorl %ebx,%eax |
+ movzbl -128(%ebp,%esi,1),%ebx |
+ shll $16,%ebx |
+ xorl %ebx,%eax |
+ xorl 896(%ebp,%ecx,4),%eax |
+ movl %eax,24(%edi) |
+ xorl 4(%edi),%eax |
+ movl %eax,28(%edi) |
+ xorl 8(%edi),%eax |
+ movl %eax,32(%edi) |
+ xorl 12(%edi),%eax |
+ movl %eax,36(%edi) |
+ cmpl $7,%ecx |
+ je .L05012break |
+ incl %ecx |
+ xorl 16(%edi),%eax |
+ movl %eax,40(%edi) |
+ xorl 20(%edi),%eax |
+ movl %eax,44(%edi) |
+ addl $24,%edi |
+ jmp .L04912loop |
+.L05012break: |
+ movl $12,72(%edi) |
+ xorl %eax,%eax |
+ jmp .L045exit |
+.L04414rounds: |
+ movl (%esi),%eax |
+ movl 4(%esi),%ebx |
+ movl 8(%esi),%ecx |
+ movl 12(%esi),%edx |
+ movl %eax,(%edi) |
+ movl %ebx,4(%edi) |
+ movl %ecx,8(%edi) |
+ movl %edx,12(%edi) |
+ movl 16(%esi),%eax |
+ movl 20(%esi),%ebx |
+ movl 24(%esi),%ecx |
+ movl 28(%esi),%edx |
+ movl %eax,16(%edi) |
+ movl %ebx,20(%edi) |
+ movl %ecx,24(%edi) |
+ movl %edx,28(%edi) |
+ xorl %ecx,%ecx |
+ jmp .L05114shortcut |
+.align 4 |
+.L05214loop: |
+ movl 28(%edi),%edx |
+.L05114shortcut: |
+ movl (%edi),%eax |
+ movzbl %dl,%esi |
+ movzbl -128(%ebp,%esi,1),%ebx |
+ movzbl %dh,%esi |
+ shll $24,%ebx |
+ xorl %ebx,%eax |
+ movzbl -128(%ebp,%esi,1),%ebx |
+ shrl $16,%edx |
+ movzbl %dl,%esi |
+ xorl %ebx,%eax |
+ movzbl -128(%ebp,%esi,1),%ebx |
+ movzbl %dh,%esi |
+ shll $8,%ebx |
+ xorl %ebx,%eax |
+ movzbl -128(%ebp,%esi,1),%ebx |
+ shll $16,%ebx |
+ xorl %ebx,%eax |
+ xorl 896(%ebp,%ecx,4),%eax |
+ movl %eax,32(%edi) |
+ xorl 4(%edi),%eax |
+ movl %eax,36(%edi) |
+ xorl 8(%edi),%eax |
+ movl %eax,40(%edi) |
+ xorl 12(%edi),%eax |
+ movl %eax,44(%edi) |
+ cmpl $6,%ecx |
+ je .L05314break |
+ incl %ecx |
+ movl %eax,%edx |
+ movl 16(%edi),%eax |
+ movzbl %dl,%esi |
+ movzbl -128(%ebp,%esi,1),%ebx |
+ movzbl %dh,%esi |
+ xorl %ebx,%eax |
+ movzbl -128(%ebp,%esi,1),%ebx |
+ shrl $16,%edx |
+ shll $8,%ebx |
+ movzbl %dl,%esi |
+ xorl %ebx,%eax |
+ movzbl -128(%ebp,%esi,1),%ebx |
+ movzbl %dh,%esi |
+ shll $16,%ebx |
+ xorl %ebx,%eax |
+ movzbl -128(%ebp,%esi,1),%ebx |
+ shll $24,%ebx |
+ xorl %ebx,%eax |
+ movl %eax,48(%edi) |
+ xorl 20(%edi),%eax |
+ movl %eax,52(%edi) |
+ xorl 24(%edi),%eax |
+ movl %eax,56(%edi) |
+ xorl 28(%edi),%eax |
+ movl %eax,60(%edi) |
+ addl $32,%edi |
+ jmp .L05214loop |
+.L05314break: |
+ movl $14,48(%edi) |
+ xorl %eax,%eax |
+ jmp .L045exit |
+.L040badpointer: |
+ movl $-1,%eax |
+.L045exit: |
+ popl %edi |
+ popl %esi |
+ popl %ebx |
+ popl %ebp |
+ ret |
+.size _x86_AES_set_encrypt_key,.-_x86_AES_set_encrypt_key |
+.globl asm_AES_set_encrypt_key |
+.hidden asm_AES_set_encrypt_key |
+.type asm_AES_set_encrypt_key,@function |
+.align 16 |
+asm_AES_set_encrypt_key: |
+.L_asm_AES_set_encrypt_key_begin: |
+ call _x86_AES_set_encrypt_key |
+ ret |
+.size asm_AES_set_encrypt_key,.-.L_asm_AES_set_encrypt_key_begin |
+.globl asm_AES_set_decrypt_key |
+.hidden asm_AES_set_decrypt_key |
+.type asm_AES_set_decrypt_key,@function |
+.align 16 |
+asm_AES_set_decrypt_key: |
+.L_asm_AES_set_decrypt_key_begin: |
+ call _x86_AES_set_encrypt_key |
+ cmpl $0,%eax |
+ je .L054proceed |
+ ret |
+.L054proceed: |
+ pushl %ebp |
+ pushl %ebx |
+ pushl %esi |
+ pushl %edi |
+ movl 28(%esp),%esi |
+ movl 240(%esi),%ecx |
+ leal (,%ecx,4),%ecx |
+ leal (%esi,%ecx,4),%edi |
+.align 4 |
+.L055invert: |
+ movl (%esi),%eax |
+ movl 4(%esi),%ebx |
+ movl (%edi),%ecx |
+ movl 4(%edi),%edx |
+ movl %eax,(%edi) |
+ movl %ebx,4(%edi) |
+ movl %ecx,(%esi) |
+ movl %edx,4(%esi) |
+ movl 8(%esi),%eax |
+ movl 12(%esi),%ebx |
+ movl 8(%edi),%ecx |
+ movl 12(%edi),%edx |
+ movl %eax,8(%edi) |
+ movl %ebx,12(%edi) |
+ movl %ecx,8(%esi) |
+ movl %edx,12(%esi) |
+ addl $16,%esi |
+ subl $16,%edi |
+ cmpl %edi,%esi |
+ jne .L055invert |
+ movl 28(%esp),%edi |
+ movl 240(%edi),%esi |
+ leal -2(%esi,%esi,1),%esi |
+ leal (%edi,%esi,8),%esi |
+ movl %esi,28(%esp) |
+ movl 16(%edi),%eax |
+.align 4 |
+.L056permute: |
+ addl $16,%edi |
+ movl $2155905152,%ebp |
+ andl %eax,%ebp |
+ leal (%eax,%eax,1),%ebx |
+ movl %ebp,%esi |
+ shrl $7,%ebp |
+ subl %ebp,%esi |
+ andl $4278124286,%ebx |
+ andl $454761243,%esi |
+ xorl %esi,%ebx |
+ movl $2155905152,%ebp |
+ andl %ebx,%ebp |
+ leal (%ebx,%ebx,1),%ecx |
+ movl %ebp,%esi |
+ shrl $7,%ebp |
+ subl %ebp,%esi |
+ andl $4278124286,%ecx |
+ andl $454761243,%esi |
+ xorl %eax,%ebx |
+ xorl %esi,%ecx |
+ movl $2155905152,%ebp |
+ andl %ecx,%ebp |
+ leal (%ecx,%ecx,1),%edx |
+ movl %ebp,%esi |
+ shrl $7,%ebp |
+ xorl %eax,%ecx |
+ subl %ebp,%esi |
+ andl $4278124286,%edx |
+ andl $454761243,%esi |
+ roll $8,%eax |
+ xorl %esi,%edx |
+ movl 4(%edi),%ebp |
+ xorl %ebx,%eax |
+ xorl %edx,%ebx |
+ xorl %ecx,%eax |
+ roll $24,%ebx |
+ xorl %edx,%ecx |
+ xorl %edx,%eax |
+ roll $16,%ecx |
+ xorl %ebx,%eax |
+ roll $8,%edx |
+ xorl %ecx,%eax |
+ movl %ebp,%ebx |
+ xorl %edx,%eax |
+ movl %eax,(%edi) |
+ movl $2155905152,%ebp |
+ andl %ebx,%ebp |
+ leal (%ebx,%ebx,1),%ecx |
+ movl %ebp,%esi |
+ shrl $7,%ebp |
+ subl %ebp,%esi |
+ andl $4278124286,%ecx |
+ andl $454761243,%esi |
+ xorl %esi,%ecx |
+ movl $2155905152,%ebp |
+ andl %ecx,%ebp |
+ leal (%ecx,%ecx,1),%edx |
+ movl %ebp,%esi |
+ shrl $7,%ebp |
+ subl %ebp,%esi |
+ andl $4278124286,%edx |
+ andl $454761243,%esi |
+ xorl %ebx,%ecx |
+ xorl %esi,%edx |
+ movl $2155905152,%ebp |
+ andl %edx,%ebp |
+ leal (%edx,%edx,1),%eax |
+ movl %ebp,%esi |
+ shrl $7,%ebp |
+ xorl %ebx,%edx |
+ subl %ebp,%esi |
+ andl $4278124286,%eax |
+ andl $454761243,%esi |
+ roll $8,%ebx |
+ xorl %esi,%eax |
+ movl 8(%edi),%ebp |
+ xorl %ecx,%ebx |
+ xorl %eax,%ecx |
+ xorl %edx,%ebx |
+ roll $24,%ecx |
+ xorl %eax,%edx |
+ xorl %eax,%ebx |
+ roll $16,%edx |
+ xorl %ecx,%ebx |
+ roll $8,%eax |
+ xorl %edx,%ebx |
+ movl %ebp,%ecx |
+ xorl %eax,%ebx |
+ movl %ebx,4(%edi) |
+ movl $2155905152,%ebp |
+ andl %ecx,%ebp |
+ leal (%ecx,%ecx,1),%edx |
+ movl %ebp,%esi |
+ shrl $7,%ebp |
+ subl %ebp,%esi |
+ andl $4278124286,%edx |
+ andl $454761243,%esi |
+ xorl %esi,%edx |
+ movl $2155905152,%ebp |
+ andl %edx,%ebp |
+ leal (%edx,%edx,1),%eax |
+ movl %ebp,%esi |
+ shrl $7,%ebp |
+ subl %ebp,%esi |
+ andl $4278124286,%eax |
+ andl $454761243,%esi |
+ xorl %ecx,%edx |
+ xorl %esi,%eax |
+ movl $2155905152,%ebp |
+ andl %eax,%ebp |
+ leal (%eax,%eax,1),%ebx |
+ movl %ebp,%esi |
+ shrl $7,%ebp |
+ xorl %ecx,%eax |
+ subl %ebp,%esi |
+ andl $4278124286,%ebx |
+ andl $454761243,%esi |
+ roll $8,%ecx |
+ xorl %esi,%ebx |
+ movl 12(%edi),%ebp |
+ xorl %edx,%ecx |
+ xorl %ebx,%edx |
+ xorl %eax,%ecx |
+ roll $24,%edx |
+ xorl %ebx,%eax |
+ xorl %ebx,%ecx |
+ roll $16,%eax |
+ xorl %edx,%ecx |
+ roll $8,%ebx |
+ xorl %eax,%ecx |
+ movl %ebp,%edx |
+ xorl %ebx,%ecx |
+ movl %ecx,8(%edi) |
+ movl $2155905152,%ebp |
+ andl %edx,%ebp |
+ leal (%edx,%edx,1),%eax |
+ movl %ebp,%esi |
+ shrl $7,%ebp |
+ subl %ebp,%esi |
+ andl $4278124286,%eax |
+ andl $454761243,%esi |
+ xorl %esi,%eax |
+ movl $2155905152,%ebp |
+ andl %eax,%ebp |
+ leal (%eax,%eax,1),%ebx |
+ movl %ebp,%esi |
+ shrl $7,%ebp |
+ subl %ebp,%esi |
+ andl $4278124286,%ebx |
+ andl $454761243,%esi |
+ xorl %edx,%eax |
+ xorl %esi,%ebx |
+ movl $2155905152,%ebp |
+ andl %ebx,%ebp |
+ leal (%ebx,%ebx,1),%ecx |
+ movl %ebp,%esi |
+ shrl $7,%ebp |
+ xorl %edx,%ebx |
+ subl %ebp,%esi |
+ andl $4278124286,%ecx |
+ andl $454761243,%esi |
+ roll $8,%edx |
+ xorl %esi,%ecx |
+ movl 16(%edi),%ebp |
+ xorl %eax,%edx |
+ xorl %ecx,%eax |
+ xorl %ebx,%edx |
+ roll $24,%eax |
+ xorl %ecx,%ebx |
+ xorl %ecx,%edx |
+ roll $16,%ebx |
+ xorl %eax,%edx |
+ roll $8,%ecx |
+ xorl %ebx,%edx |
+ movl %ebp,%eax |
+ xorl %ecx,%edx |
+ movl %edx,12(%edi) |
+ cmpl 28(%esp),%edi |
+ jb .L056permute |
+ xorl %eax,%eax |
+ popl %edi |
+ popl %esi |
+ popl %ebx |
+ popl %ebp |
+ ret |
+.size asm_AES_set_decrypt_key,.-.L_asm_AES_set_decrypt_key_begin |
+.byte 65,69,83,32,102,111,114,32,120,56,54,44,32,67,82,89 |
+.byte 80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114 |
+.byte 111,64,111,112,101,110,115,115,108,46,111,114,103,62,0 |
+#endif |