Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(403)

Unified Diff: third_party/boringssl/win-x86_64/crypto/sha/sha256-x86_64.asm

Issue 1938433002: Revert BoringSSL Roll (Closed) Base URL: git@github.com:dart-lang/sdk.git@master
Patch Set: Created 4 years, 8 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
Index: third_party/boringssl/win-x86_64/crypto/sha/sha256-x86_64.asm
diff --git a/third_party/boringssl/win-x86_64/crypto/sha/sha256-x86_64.asm b/third_party/boringssl/win-x86_64/crypto/sha/sha256-x86_64.asm
index efaf9b55fccff2db693d37a0d8d653357c6cb87f..e6193c5b9d75453e8f81003c4b4296ff281455a9 100644
--- a/third_party/boringssl/win-x86_64/crypto/sha/sha256-x86_64.asm
+++ b/third_party/boringssl/win-x86_64/crypto/sha/sha256-x86_64.asm
@@ -23,11 +23,6 @@ $L$SEH_begin_sha256_block_data_order:
mov r9d,DWORD[r11]
mov r10d,DWORD[4+r11]
mov r11d,DWORD[8+r11]
- and r9d,1073741824
- and r10d,268435968
- or r10d,r9d
- cmp r10d,1342177792
- je NEAR $L$avx_shortcut
test r10d,512
jnz NEAR $L$ssse3_shortcut
push rbx
@@ -2882,1082 +2877,6 @@ $L$epilogue_ssse3:
mov rsi,QWORD[16+rsp]
DB 0F3h,0C3h ;repret
$L$SEH_end_sha256_block_data_order_ssse3:
-
-ALIGN 64
-sha256_block_data_order_avx:
- mov QWORD[8+rsp],rdi ;WIN64 prologue
- mov QWORD[16+rsp],rsi
- mov rax,rsp
-$L$SEH_begin_sha256_block_data_order_avx:
- mov rdi,rcx
- mov rsi,rdx
- mov rdx,r8
-
-
-$L$avx_shortcut:
- push rbx
- push rbp
- push r12
- push r13
- push r14
- push r15
- mov r11,rsp
- shl rdx,4
- sub rsp,160
- lea rdx,[rdx*4+rsi]
- and rsp,-64
- mov QWORD[((64+0))+rsp],rdi
- mov QWORD[((64+8))+rsp],rsi
- mov QWORD[((64+16))+rsp],rdx
- mov QWORD[((64+24))+rsp],r11
- movaps XMMWORD[(64+32)+rsp],xmm6
- movaps XMMWORD[(64+48)+rsp],xmm7
- movaps XMMWORD[(64+64)+rsp],xmm8
- movaps XMMWORD[(64+80)+rsp],xmm9
-$L$prologue_avx:
-
- vzeroupper
- mov eax,DWORD[rdi]
- mov ebx,DWORD[4+rdi]
- mov ecx,DWORD[8+rdi]
- mov edx,DWORD[12+rdi]
- mov r8d,DWORD[16+rdi]
- mov r9d,DWORD[20+rdi]
- mov r10d,DWORD[24+rdi]
- mov r11d,DWORD[28+rdi]
- vmovdqa xmm8,XMMWORD[((K256+512+32))]
- vmovdqa xmm9,XMMWORD[((K256+512+64))]
- jmp NEAR $L$loop_avx
-ALIGN 16
-$L$loop_avx:
- vmovdqa xmm7,XMMWORD[((K256+512))]
- vmovdqu xmm0,XMMWORD[rsi]
- vmovdqu xmm1,XMMWORD[16+rsi]
- vmovdqu xmm2,XMMWORD[32+rsi]
- vmovdqu xmm3,XMMWORD[48+rsi]
- vpshufb xmm0,xmm0,xmm7
- lea rbp,[K256]
- vpshufb xmm1,xmm1,xmm7
- vpshufb xmm2,xmm2,xmm7
- vpaddd xmm4,xmm0,XMMWORD[rbp]
- vpshufb xmm3,xmm3,xmm7
- vpaddd xmm5,xmm1,XMMWORD[32+rbp]
- vpaddd xmm6,xmm2,XMMWORD[64+rbp]
- vpaddd xmm7,xmm3,XMMWORD[96+rbp]
- vmovdqa XMMWORD[rsp],xmm4
- mov r14d,eax
- vmovdqa XMMWORD[16+rsp],xmm5
- mov edi,ebx
- vmovdqa XMMWORD[32+rsp],xmm6
- xor edi,ecx
- vmovdqa XMMWORD[48+rsp],xmm7
- mov r13d,r8d
- jmp NEAR $L$avx_00_47
-
-ALIGN 16
-$L$avx_00_47:
- sub rbp,-128
- vpalignr xmm4,xmm1,xmm0,4
- shrd r13d,r13d,14
- mov eax,r14d
- mov r12d,r9d
- vpalignr xmm7,xmm3,xmm2,4
- shrd r14d,r14d,9
- xor r13d,r8d
- xor r12d,r10d
- vpsrld xmm6,xmm4,7
- shrd r13d,r13d,5
- xor r14d,eax
- and r12d,r8d
- vpaddd xmm0,xmm0,xmm7
- xor r13d,r8d
- add r11d,DWORD[rsp]
- mov r15d,eax
- vpsrld xmm7,xmm4,3
- xor r12d,r10d
- shrd r14d,r14d,11
- xor r15d,ebx
- vpslld xmm5,xmm4,14
- add r11d,r12d
- shrd r13d,r13d,6
- and edi,r15d
- vpxor xmm4,xmm7,xmm6
- xor r14d,eax
- add r11d,r13d
- xor edi,ebx
- vpshufd xmm7,xmm3,250
- shrd r14d,r14d,2
- add edx,r11d
- add r11d,edi
- vpsrld xmm6,xmm6,11
- mov r13d,edx
- add r14d,r11d
- shrd r13d,r13d,14
- vpxor xmm4,xmm4,xmm5
- mov r11d,r14d
- mov r12d,r8d
- shrd r14d,r14d,9
- vpslld xmm5,xmm5,11
- xor r13d,edx
- xor r12d,r9d
- shrd r13d,r13d,5
- vpxor xmm4,xmm4,xmm6
- xor r14d,r11d
- and r12d,edx
- xor r13d,edx
- vpsrld xmm6,xmm7,10
- add r10d,DWORD[4+rsp]
- mov edi,r11d
- xor r12d,r9d
- vpxor xmm4,xmm4,xmm5
- shrd r14d,r14d,11
- xor edi,eax
- add r10d,r12d
- vpsrlq xmm7,xmm7,17
- shrd r13d,r13d,6
- and r15d,edi
- xor r14d,r11d
- vpaddd xmm0,xmm0,xmm4
- add r10d,r13d
- xor r15d,eax
- shrd r14d,r14d,2
- vpxor xmm6,xmm6,xmm7
- add ecx,r10d
- add r10d,r15d
- mov r13d,ecx
- vpsrlq xmm7,xmm7,2
- add r14d,r10d
- shrd r13d,r13d,14
- mov r10d,r14d
- vpxor xmm6,xmm6,xmm7
- mov r12d,edx
- shrd r14d,r14d,9
- xor r13d,ecx
- vpshufb xmm6,xmm6,xmm8
- xor r12d,r8d
- shrd r13d,r13d,5
- xor r14d,r10d
- vpaddd xmm0,xmm0,xmm6
- and r12d,ecx
- xor r13d,ecx
- add r9d,DWORD[8+rsp]
- vpshufd xmm7,xmm0,80
- mov r15d,r10d
- xor r12d,r8d
- shrd r14d,r14d,11
- vpsrld xmm6,xmm7,10
- xor r15d,r11d
- add r9d,r12d
- shrd r13d,r13d,6
- vpsrlq xmm7,xmm7,17
- and edi,r15d
- xor r14d,r10d
- add r9d,r13d
- vpxor xmm6,xmm6,xmm7
- xor edi,r11d
- shrd r14d,r14d,2
- add ebx,r9d
- vpsrlq xmm7,xmm7,2
- add r9d,edi
- mov r13d,ebx
- add r14d,r9d
- vpxor xmm6,xmm6,xmm7
- shrd r13d,r13d,14
- mov r9d,r14d
- mov r12d,ecx
- vpshufb xmm6,xmm6,xmm9
- shrd r14d,r14d,9
- xor r13d,ebx
- xor r12d,edx
- vpaddd xmm0,xmm0,xmm6
- shrd r13d,r13d,5
- xor r14d,r9d
- and r12d,ebx
- vpaddd xmm6,xmm0,XMMWORD[rbp]
- xor r13d,ebx
- add r8d,DWORD[12+rsp]
- mov edi,r9d
- xor r12d,edx
- shrd r14d,r14d,11
- xor edi,r10d
- add r8d,r12d
- shrd r13d,r13d,6
- and r15d,edi
- xor r14d,r9d
- add r8d,r13d
- xor r15d,r10d
- shrd r14d,r14d,2
- add eax,r8d
- add r8d,r15d
- mov r13d,eax
- add r14d,r8d
- vmovdqa XMMWORD[rsp],xmm6
- vpalignr xmm4,xmm2,xmm1,4
- shrd r13d,r13d,14
- mov r8d,r14d
- mov r12d,ebx
- vpalignr xmm7,xmm0,xmm3,4
- shrd r14d,r14d,9
- xor r13d,eax
- xor r12d,ecx
- vpsrld xmm6,xmm4,7
- shrd r13d,r13d,5
- xor r14d,r8d
- and r12d,eax
- vpaddd xmm1,xmm1,xmm7
- xor r13d,eax
- add edx,DWORD[16+rsp]
- mov r15d,r8d
- vpsrld xmm7,xmm4,3
- xor r12d,ecx
- shrd r14d,r14d,11
- xor r15d,r9d
- vpslld xmm5,xmm4,14
- add edx,r12d
- shrd r13d,r13d,6
- and edi,r15d
- vpxor xmm4,xmm7,xmm6
- xor r14d,r8d
- add edx,r13d
- xor edi,r9d
- vpshufd xmm7,xmm0,250
- shrd r14d,r14d,2
- add r11d,edx
- add edx,edi
- vpsrld xmm6,xmm6,11
- mov r13d,r11d
- add r14d,edx
- shrd r13d,r13d,14
- vpxor xmm4,xmm4,xmm5
- mov edx,r14d
- mov r12d,eax
- shrd r14d,r14d,9
- vpslld xmm5,xmm5,11
- xor r13d,r11d
- xor r12d,ebx
- shrd r13d,r13d,5
- vpxor xmm4,xmm4,xmm6
- xor r14d,edx
- and r12d,r11d
- xor r13d,r11d
- vpsrld xmm6,xmm7,10
- add ecx,DWORD[20+rsp]
- mov edi,edx
- xor r12d,ebx
- vpxor xmm4,xmm4,xmm5
- shrd r14d,r14d,11
- xor edi,r8d
- add ecx,r12d
- vpsrlq xmm7,xmm7,17
- shrd r13d,r13d,6
- and r15d,edi
- xor r14d,edx
- vpaddd xmm1,xmm1,xmm4
- add ecx,r13d
- xor r15d,r8d
- shrd r14d,r14d,2
- vpxor xmm6,xmm6,xmm7
- add r10d,ecx
- add ecx,r15d
- mov r13d,r10d
- vpsrlq xmm7,xmm7,2
- add r14d,ecx
- shrd r13d,r13d,14
- mov ecx,r14d
- vpxor xmm6,xmm6,xmm7
- mov r12d,r11d
- shrd r14d,r14d,9
- xor r13d,r10d
- vpshufb xmm6,xmm6,xmm8
- xor r12d,eax
- shrd r13d,r13d,5
- xor r14d,ecx
- vpaddd xmm1,xmm1,xmm6
- and r12d,r10d
- xor r13d,r10d
- add ebx,DWORD[24+rsp]
- vpshufd xmm7,xmm1,80
- mov r15d,ecx
- xor r12d,eax
- shrd r14d,r14d,11
- vpsrld xmm6,xmm7,10
- xor r15d,edx
- add ebx,r12d
- shrd r13d,r13d,6
- vpsrlq xmm7,xmm7,17
- and edi,r15d
- xor r14d,ecx
- add ebx,r13d
- vpxor xmm6,xmm6,xmm7
- xor edi,edx
- shrd r14d,r14d,2
- add r9d,ebx
- vpsrlq xmm7,xmm7,2
- add ebx,edi
- mov r13d,r9d
- add r14d,ebx
- vpxor xmm6,xmm6,xmm7
- shrd r13d,r13d,14
- mov ebx,r14d
- mov r12d,r10d
- vpshufb xmm6,xmm6,xmm9
- shrd r14d,r14d,9
- xor r13d,r9d
- xor r12d,r11d
- vpaddd xmm1,xmm1,xmm6
- shrd r13d,r13d,5
- xor r14d,ebx
- and r12d,r9d
- vpaddd xmm6,xmm1,XMMWORD[32+rbp]
- xor r13d,r9d
- add eax,DWORD[28+rsp]
- mov edi,ebx
- xor r12d,r11d
- shrd r14d,r14d,11
- xor edi,ecx
- add eax,r12d
- shrd r13d,r13d,6
- and r15d,edi
- xor r14d,ebx
- add eax,r13d
- xor r15d,ecx
- shrd r14d,r14d,2
- add r8d,eax
- add eax,r15d
- mov r13d,r8d
- add r14d,eax
- vmovdqa XMMWORD[16+rsp],xmm6
- vpalignr xmm4,xmm3,xmm2,4
- shrd r13d,r13d,14
- mov eax,r14d
- mov r12d,r9d
- vpalignr xmm7,xmm1,xmm0,4
- shrd r14d,r14d,9
- xor r13d,r8d
- xor r12d,r10d
- vpsrld xmm6,xmm4,7
- shrd r13d,r13d,5
- xor r14d,eax
- and r12d,r8d
- vpaddd xmm2,xmm2,xmm7
- xor r13d,r8d
- add r11d,DWORD[32+rsp]
- mov r15d,eax
- vpsrld xmm7,xmm4,3
- xor r12d,r10d
- shrd r14d,r14d,11
- xor r15d,ebx
- vpslld xmm5,xmm4,14
- add r11d,r12d
- shrd r13d,r13d,6
- and edi,r15d
- vpxor xmm4,xmm7,xmm6
- xor r14d,eax
- add r11d,r13d
- xor edi,ebx
- vpshufd xmm7,xmm1,250
- shrd r14d,r14d,2
- add edx,r11d
- add r11d,edi
- vpsrld xmm6,xmm6,11
- mov r13d,edx
- add r14d,r11d
- shrd r13d,r13d,14
- vpxor xmm4,xmm4,xmm5
- mov r11d,r14d
- mov r12d,r8d
- shrd r14d,r14d,9
- vpslld xmm5,xmm5,11
- xor r13d,edx
- xor r12d,r9d
- shrd r13d,r13d,5
- vpxor xmm4,xmm4,xmm6
- xor r14d,r11d
- and r12d,edx
- xor r13d,edx
- vpsrld xmm6,xmm7,10
- add r10d,DWORD[36+rsp]
- mov edi,r11d
- xor r12d,r9d
- vpxor xmm4,xmm4,xmm5
- shrd r14d,r14d,11
- xor edi,eax
- add r10d,r12d
- vpsrlq xmm7,xmm7,17
- shrd r13d,r13d,6
- and r15d,edi
- xor r14d,r11d
- vpaddd xmm2,xmm2,xmm4
- add r10d,r13d
- xor r15d,eax
- shrd r14d,r14d,2
- vpxor xmm6,xmm6,xmm7
- add ecx,r10d
- add r10d,r15d
- mov r13d,ecx
- vpsrlq xmm7,xmm7,2
- add r14d,r10d
- shrd r13d,r13d,14
- mov r10d,r14d
- vpxor xmm6,xmm6,xmm7
- mov r12d,edx
- shrd r14d,r14d,9
- xor r13d,ecx
- vpshufb xmm6,xmm6,xmm8
- xor r12d,r8d
- shrd r13d,r13d,5
- xor r14d,r10d
- vpaddd xmm2,xmm2,xmm6
- and r12d,ecx
- xor r13d,ecx
- add r9d,DWORD[40+rsp]
- vpshufd xmm7,xmm2,80
- mov r15d,r10d
- xor r12d,r8d
- shrd r14d,r14d,11
- vpsrld xmm6,xmm7,10
- xor r15d,r11d
- add r9d,r12d
- shrd r13d,r13d,6
- vpsrlq xmm7,xmm7,17
- and edi,r15d
- xor r14d,r10d
- add r9d,r13d
- vpxor xmm6,xmm6,xmm7
- xor edi,r11d
- shrd r14d,r14d,2
- add ebx,r9d
- vpsrlq xmm7,xmm7,2
- add r9d,edi
- mov r13d,ebx
- add r14d,r9d
- vpxor xmm6,xmm6,xmm7
- shrd r13d,r13d,14
- mov r9d,r14d
- mov r12d,ecx
- vpshufb xmm6,xmm6,xmm9
- shrd r14d,r14d,9
- xor r13d,ebx
- xor r12d,edx
- vpaddd xmm2,xmm2,xmm6
- shrd r13d,r13d,5
- xor r14d,r9d
- and r12d,ebx
- vpaddd xmm6,xmm2,XMMWORD[64+rbp]
- xor r13d,ebx
- add r8d,DWORD[44+rsp]
- mov edi,r9d
- xor r12d,edx
- shrd r14d,r14d,11
- xor edi,r10d
- add r8d,r12d
- shrd r13d,r13d,6
- and r15d,edi
- xor r14d,r9d
- add r8d,r13d
- xor r15d,r10d
- shrd r14d,r14d,2
- add eax,r8d
- add r8d,r15d
- mov r13d,eax
- add r14d,r8d
- vmovdqa XMMWORD[32+rsp],xmm6
- vpalignr xmm4,xmm0,xmm3,4
- shrd r13d,r13d,14
- mov r8d,r14d
- mov r12d,ebx
- vpalignr xmm7,xmm2,xmm1,4
- shrd r14d,r14d,9
- xor r13d,eax
- xor r12d,ecx
- vpsrld xmm6,xmm4,7
- shrd r13d,r13d,5
- xor r14d,r8d
- and r12d,eax
- vpaddd xmm3,xmm3,xmm7
- xor r13d,eax
- add edx,DWORD[48+rsp]
- mov r15d,r8d
- vpsrld xmm7,xmm4,3
- xor r12d,ecx
- shrd r14d,r14d,11
- xor r15d,r9d
- vpslld xmm5,xmm4,14
- add edx,r12d
- shrd r13d,r13d,6
- and edi,r15d
- vpxor xmm4,xmm7,xmm6
- xor r14d,r8d
- add edx,r13d
- xor edi,r9d
- vpshufd xmm7,xmm2,250
- shrd r14d,r14d,2
- add r11d,edx
- add edx,edi
- vpsrld xmm6,xmm6,11
- mov r13d,r11d
- add r14d,edx
- shrd r13d,r13d,14
- vpxor xmm4,xmm4,xmm5
- mov edx,r14d
- mov r12d,eax
- shrd r14d,r14d,9
- vpslld xmm5,xmm5,11
- xor r13d,r11d
- xor r12d,ebx
- shrd r13d,r13d,5
- vpxor xmm4,xmm4,xmm6
- xor r14d,edx
- and r12d,r11d
- xor r13d,r11d
- vpsrld xmm6,xmm7,10
- add ecx,DWORD[52+rsp]
- mov edi,edx
- xor r12d,ebx
- vpxor xmm4,xmm4,xmm5
- shrd r14d,r14d,11
- xor edi,r8d
- add ecx,r12d
- vpsrlq xmm7,xmm7,17
- shrd r13d,r13d,6
- and r15d,edi
- xor r14d,edx
- vpaddd xmm3,xmm3,xmm4
- add ecx,r13d
- xor r15d,r8d
- shrd r14d,r14d,2
- vpxor xmm6,xmm6,xmm7
- add r10d,ecx
- add ecx,r15d
- mov r13d,r10d
- vpsrlq xmm7,xmm7,2
- add r14d,ecx
- shrd r13d,r13d,14
- mov ecx,r14d
- vpxor xmm6,xmm6,xmm7
- mov r12d,r11d
- shrd r14d,r14d,9
- xor r13d,r10d
- vpshufb xmm6,xmm6,xmm8
- xor r12d,eax
- shrd r13d,r13d,5
- xor r14d,ecx
- vpaddd xmm3,xmm3,xmm6
- and r12d,r10d
- xor r13d,r10d
- add ebx,DWORD[56+rsp]
- vpshufd xmm7,xmm3,80
- mov r15d,ecx
- xor r12d,eax
- shrd r14d,r14d,11
- vpsrld xmm6,xmm7,10
- xor r15d,edx
- add ebx,r12d
- shrd r13d,r13d,6
- vpsrlq xmm7,xmm7,17
- and edi,r15d
- xor r14d,ecx
- add ebx,r13d
- vpxor xmm6,xmm6,xmm7
- xor edi,edx
- shrd r14d,r14d,2
- add r9d,ebx
- vpsrlq xmm7,xmm7,2
- add ebx,edi
- mov r13d,r9d
- add r14d,ebx
- vpxor xmm6,xmm6,xmm7
- shrd r13d,r13d,14
- mov ebx,r14d
- mov r12d,r10d
- vpshufb xmm6,xmm6,xmm9
- shrd r14d,r14d,9
- xor r13d,r9d
- xor r12d,r11d
- vpaddd xmm3,xmm3,xmm6
- shrd r13d,r13d,5
- xor r14d,ebx
- and r12d,r9d
- vpaddd xmm6,xmm3,XMMWORD[96+rbp]
- xor r13d,r9d
- add eax,DWORD[60+rsp]
- mov edi,ebx
- xor r12d,r11d
- shrd r14d,r14d,11
- xor edi,ecx
- add eax,r12d
- shrd r13d,r13d,6
- and r15d,edi
- xor r14d,ebx
- add eax,r13d
- xor r15d,ecx
- shrd r14d,r14d,2
- add r8d,eax
- add eax,r15d
- mov r13d,r8d
- add r14d,eax
- vmovdqa XMMWORD[48+rsp],xmm6
- cmp BYTE[131+rbp],0
- jne NEAR $L$avx_00_47
- shrd r13d,r13d,14
- mov eax,r14d
- mov r12d,r9d
- shrd r14d,r14d,9
- xor r13d,r8d
- xor r12d,r10d
- shrd r13d,r13d,5
- xor r14d,eax
- and r12d,r8d
- xor r13d,r8d
- add r11d,DWORD[rsp]
- mov r15d,eax
- xor r12d,r10d
- shrd r14d,r14d,11
- xor r15d,ebx
- add r11d,r12d
- shrd r13d,r13d,6
- and edi,r15d
- xor r14d,eax
- add r11d,r13d
- xor edi,ebx
- shrd r14d,r14d,2
- add edx,r11d
- add r11d,edi
- mov r13d,edx
- add r14d,r11d
- shrd r13d,r13d,14
- mov r11d,r14d
- mov r12d,r8d
- shrd r14d,r14d,9
- xor r13d,edx
- xor r12d,r9d
- shrd r13d,r13d,5
- xor r14d,r11d
- and r12d,edx
- xor r13d,edx
- add r10d,DWORD[4+rsp]
- mov edi,r11d
- xor r12d,r9d
- shrd r14d,r14d,11
- xor edi,eax
- add r10d,r12d
- shrd r13d,r13d,6
- and r15d,edi
- xor r14d,r11d
- add r10d,r13d
- xor r15d,eax
- shrd r14d,r14d,2
- add ecx,r10d
- add r10d,r15d
- mov r13d,ecx
- add r14d,r10d
- shrd r13d,r13d,14
- mov r10d,r14d
- mov r12d,edx
- shrd r14d,r14d,9
- xor r13d,ecx
- xor r12d,r8d
- shrd r13d,r13d,5
- xor r14d,r10d
- and r12d,ecx
- xor r13d,ecx
- add r9d,DWORD[8+rsp]
- mov r15d,r10d
- xor r12d,r8d
- shrd r14d,r14d,11
- xor r15d,r11d
- add r9d,r12d
- shrd r13d,r13d,6
- and edi,r15d
- xor r14d,r10d
- add r9d,r13d
- xor edi,r11d
- shrd r14d,r14d,2
- add ebx,r9d
- add r9d,edi
- mov r13d,ebx
- add r14d,r9d
- shrd r13d,r13d,14
- mov r9d,r14d
- mov r12d,ecx
- shrd r14d,r14d,9
- xor r13d,ebx
- xor r12d,edx
- shrd r13d,r13d,5
- xor r14d,r9d
- and r12d,ebx
- xor r13d,ebx
- add r8d,DWORD[12+rsp]
- mov edi,r9d
- xor r12d,edx
- shrd r14d,r14d,11
- xor edi,r10d
- add r8d,r12d
- shrd r13d,r13d,6
- and r15d,edi
- xor r14d,r9d
- add r8d,r13d
- xor r15d,r10d
- shrd r14d,r14d,2
- add eax,r8d
- add r8d,r15d
- mov r13d,eax
- add r14d,r8d
- shrd r13d,r13d,14
- mov r8d,r14d
- mov r12d,ebx
- shrd r14d,r14d,9
- xor r13d,eax
- xor r12d,ecx
- shrd r13d,r13d,5
- xor r14d,r8d
- and r12d,eax
- xor r13d,eax
- add edx,DWORD[16+rsp]
- mov r15d,r8d
- xor r12d,ecx
- shrd r14d,r14d,11
- xor r15d,r9d
- add edx,r12d
- shrd r13d,r13d,6
- and edi,r15d
- xor r14d,r8d
- add edx,r13d
- xor edi,r9d
- shrd r14d,r14d,2
- add r11d,edx
- add edx,edi
- mov r13d,r11d
- add r14d,edx
- shrd r13d,r13d,14
- mov edx,r14d
- mov r12d,eax
- shrd r14d,r14d,9
- xor r13d,r11d
- xor r12d,ebx
- shrd r13d,r13d,5
- xor r14d,edx
- and r12d,r11d
- xor r13d,r11d
- add ecx,DWORD[20+rsp]
- mov edi,edx
- xor r12d,ebx
- shrd r14d,r14d,11
- xor edi,r8d
- add ecx,r12d
- shrd r13d,r13d,6
- and r15d,edi
- xor r14d,edx
- add ecx,r13d
- xor r15d,r8d
- shrd r14d,r14d,2
- add r10d,ecx
- add ecx,r15d
- mov r13d,r10d
- add r14d,ecx
- shrd r13d,r13d,14
- mov ecx,r14d
- mov r12d,r11d
- shrd r14d,r14d,9
- xor r13d,r10d
- xor r12d,eax
- shrd r13d,r13d,5
- xor r14d,ecx
- and r12d,r10d
- xor r13d,r10d
- add ebx,DWORD[24+rsp]
- mov r15d,ecx
- xor r12d,eax
- shrd r14d,r14d,11
- xor r15d,edx
- add ebx,r12d
- shrd r13d,r13d,6
- and edi,r15d
- xor r14d,ecx
- add ebx,r13d
- xor edi,edx
- shrd r14d,r14d,2
- add r9d,ebx
- add ebx,edi
- mov r13d,r9d
- add r14d,ebx
- shrd r13d,r13d,14
- mov ebx,r14d
- mov r12d,r10d
- shrd r14d,r14d,9
- xor r13d,r9d
- xor r12d,r11d
- shrd r13d,r13d,5
- xor r14d,ebx
- and r12d,r9d
- xor r13d,r9d
- add eax,DWORD[28+rsp]
- mov edi,ebx
- xor r12d,r11d
- shrd r14d,r14d,11
- xor edi,ecx
- add eax,r12d
- shrd r13d,r13d,6
- and r15d,edi
- xor r14d,ebx
- add eax,r13d
- xor r15d,ecx
- shrd r14d,r14d,2
- add r8d,eax
- add eax,r15d
- mov r13d,r8d
- add r14d,eax
- shrd r13d,r13d,14
- mov eax,r14d
- mov r12d,r9d
- shrd r14d,r14d,9
- xor r13d,r8d
- xor r12d,r10d
- shrd r13d,r13d,5
- xor r14d,eax
- and r12d,r8d
- xor r13d,r8d
- add r11d,DWORD[32+rsp]
- mov r15d,eax
- xor r12d,r10d
- shrd r14d,r14d,11
- xor r15d,ebx
- add r11d,r12d
- shrd r13d,r13d,6
- and edi,r15d
- xor r14d,eax
- add r11d,r13d
- xor edi,ebx
- shrd r14d,r14d,2
- add edx,r11d
- add r11d,edi
- mov r13d,edx
- add r14d,r11d
- shrd r13d,r13d,14
- mov r11d,r14d
- mov r12d,r8d
- shrd r14d,r14d,9
- xor r13d,edx
- xor r12d,r9d
- shrd r13d,r13d,5
- xor r14d,r11d
- and r12d,edx
- xor r13d,edx
- add r10d,DWORD[36+rsp]
- mov edi,r11d
- xor r12d,r9d
- shrd r14d,r14d,11
- xor edi,eax
- add r10d,r12d
- shrd r13d,r13d,6
- and r15d,edi
- xor r14d,r11d
- add r10d,r13d
- xor r15d,eax
- shrd r14d,r14d,2
- add ecx,r10d
- add r10d,r15d
- mov r13d,ecx
- add r14d,r10d
- shrd r13d,r13d,14
- mov r10d,r14d
- mov r12d,edx
- shrd r14d,r14d,9
- xor r13d,ecx
- xor r12d,r8d
- shrd r13d,r13d,5
- xor r14d,r10d
- and r12d,ecx
- xor r13d,ecx
- add r9d,DWORD[40+rsp]
- mov r15d,r10d
- xor r12d,r8d
- shrd r14d,r14d,11
- xor r15d,r11d
- add r9d,r12d
- shrd r13d,r13d,6
- and edi,r15d
- xor r14d,r10d
- add r9d,r13d
- xor edi,r11d
- shrd r14d,r14d,2
- add ebx,r9d
- add r9d,edi
- mov r13d,ebx
- add r14d,r9d
- shrd r13d,r13d,14
- mov r9d,r14d
- mov r12d,ecx
- shrd r14d,r14d,9
- xor r13d,ebx
- xor r12d,edx
- shrd r13d,r13d,5
- xor r14d,r9d
- and r12d,ebx
- xor r13d,ebx
- add r8d,DWORD[44+rsp]
- mov edi,r9d
- xor r12d,edx
- shrd r14d,r14d,11
- xor edi,r10d
- add r8d,r12d
- shrd r13d,r13d,6
- and r15d,edi
- xor r14d,r9d
- add r8d,r13d
- xor r15d,r10d
- shrd r14d,r14d,2
- add eax,r8d
- add r8d,r15d
- mov r13d,eax
- add r14d,r8d
- shrd r13d,r13d,14
- mov r8d,r14d
- mov r12d,ebx
- shrd r14d,r14d,9
- xor r13d,eax
- xor r12d,ecx
- shrd r13d,r13d,5
- xor r14d,r8d
- and r12d,eax
- xor r13d,eax
- add edx,DWORD[48+rsp]
- mov r15d,r8d
- xor r12d,ecx
- shrd r14d,r14d,11
- xor r15d,r9d
- add edx,r12d
- shrd r13d,r13d,6
- and edi,r15d
- xor r14d,r8d
- add edx,r13d
- xor edi,r9d
- shrd r14d,r14d,2
- add r11d,edx
- add edx,edi
- mov r13d,r11d
- add r14d,edx
- shrd r13d,r13d,14
- mov edx,r14d
- mov r12d,eax
- shrd r14d,r14d,9
- xor r13d,r11d
- xor r12d,ebx
- shrd r13d,r13d,5
- xor r14d,edx
- and r12d,r11d
- xor r13d,r11d
- add ecx,DWORD[52+rsp]
- mov edi,edx
- xor r12d,ebx
- shrd r14d,r14d,11
- xor edi,r8d
- add ecx,r12d
- shrd r13d,r13d,6
- and r15d,edi
- xor r14d,edx
- add ecx,r13d
- xor r15d,r8d
- shrd r14d,r14d,2
- add r10d,ecx
- add ecx,r15d
- mov r13d,r10d
- add r14d,ecx
- shrd r13d,r13d,14
- mov ecx,r14d
- mov r12d,r11d
- shrd r14d,r14d,9
- xor r13d,r10d
- xor r12d,eax
- shrd r13d,r13d,5
- xor r14d,ecx
- and r12d,r10d
- xor r13d,r10d
- add ebx,DWORD[56+rsp]
- mov r15d,ecx
- xor r12d,eax
- shrd r14d,r14d,11
- xor r15d,edx
- add ebx,r12d
- shrd r13d,r13d,6
- and edi,r15d
- xor r14d,ecx
- add ebx,r13d
- xor edi,edx
- shrd r14d,r14d,2
- add r9d,ebx
- add ebx,edi
- mov r13d,r9d
- add r14d,ebx
- shrd r13d,r13d,14
- mov ebx,r14d
- mov r12d,r10d
- shrd r14d,r14d,9
- xor r13d,r9d
- xor r12d,r11d
- shrd r13d,r13d,5
- xor r14d,ebx
- and r12d,r9d
- xor r13d,r9d
- add eax,DWORD[60+rsp]
- mov edi,ebx
- xor r12d,r11d
- shrd r14d,r14d,11
- xor edi,ecx
- add eax,r12d
- shrd r13d,r13d,6
- and r15d,edi
- xor r14d,ebx
- add eax,r13d
- xor r15d,ecx
- shrd r14d,r14d,2
- add r8d,eax
- add eax,r15d
- mov r13d,r8d
- add r14d,eax
- mov rdi,QWORD[((64+0))+rsp]
- mov eax,r14d
-
- add eax,DWORD[rdi]
- lea rsi,[64+rsi]
- add ebx,DWORD[4+rdi]
- add ecx,DWORD[8+rdi]
- add edx,DWORD[12+rdi]
- add r8d,DWORD[16+rdi]
- add r9d,DWORD[20+rdi]
- add r10d,DWORD[24+rdi]
- add r11d,DWORD[28+rdi]
-
- cmp rsi,QWORD[((64+16))+rsp]
-
- mov DWORD[rdi],eax
- mov DWORD[4+rdi],ebx
- mov DWORD[8+rdi],ecx
- mov DWORD[12+rdi],edx
- mov DWORD[16+rdi],r8d
- mov DWORD[20+rdi],r9d
- mov DWORD[24+rdi],r10d
- mov DWORD[28+rdi],r11d
- jb NEAR $L$loop_avx
-
- mov rsi,QWORD[((64+24))+rsp]
- vzeroupper
- movaps xmm6,XMMWORD[((64+32))+rsp]
- movaps xmm7,XMMWORD[((64+48))+rsp]
- movaps xmm8,XMMWORD[((64+64))+rsp]
- movaps xmm9,XMMWORD[((64+80))+rsp]
- mov r15,QWORD[rsi]
- mov r14,QWORD[8+rsi]
- mov r13,QWORD[16+rsi]
- mov r12,QWORD[24+rsi]
- mov rbp,QWORD[32+rsi]
- mov rbx,QWORD[40+rsi]
- lea rsp,[48+rsi]
-$L$epilogue_avx:
- mov rdi,QWORD[8+rsp] ;WIN64 epilogue
- mov rsi,QWORD[16+rsp]
- DB 0F3h,0C3h ;repret
-$L$SEH_end_sha256_block_data_order_avx:
EXTERN __imp_RtlVirtualUnwind
ALIGN 16
@@ -4063,9 +2982,6 @@ ALIGN 4
DD $L$SEH_begin_sha256_block_data_order_ssse3 wrt ..imagebase
DD $L$SEH_end_sha256_block_data_order_ssse3 wrt ..imagebase
DD $L$SEH_info_sha256_block_data_order_ssse3 wrt ..imagebase
- DD $L$SEH_begin_sha256_block_data_order_avx wrt ..imagebase
- DD $L$SEH_end_sha256_block_data_order_avx wrt ..imagebase
- DD $L$SEH_info_sha256_block_data_order_avx wrt ..imagebase
section .xdata rdata align=8
ALIGN 8
$L$SEH_info_sha256_block_data_order:
@@ -4076,7 +2992,3 @@ $L$SEH_info_sha256_block_data_order_ssse3:
DB 9,0,0,0
DD se_handler wrt ..imagebase
DD $L$prologue_ssse3 wrt ..imagebase,$L$epilogue_ssse3 wrt ..imagebase
-$L$SEH_info_sha256_block_data_order_avx:
-DB 9,0,0,0
- DD se_handler wrt ..imagebase
- DD $L$prologue_avx wrt ..imagebase,$L$epilogue_avx wrt ..imagebase

Powered by Google App Engine
This is Rietveld 408576698