OLD | NEW |
1 #if defined(__x86_64__) | 1 #if defined(__x86_64__) |
2 .text | 2 .text |
3 .extern OPENSSL_ia32cap_P | 3 .extern OPENSSL_ia32cap_P |
4 .hidden OPENSSL_ia32cap_P | 4 .hidden OPENSSL_ia32cap_P |
5 .globl aesni_encrypt | 5 .globl aesni_encrypt |
6 .hidden aesni_encrypt | 6 .hidden aesni_encrypt |
7 .type aesni_encrypt,@function | 7 .type aesni_encrypt,@function |
8 .align 16 | 8 .align 16 |
9 aesni_encrypt: | 9 aesni_encrypt: |
10 movups (%rdi),%xmm2 | 10 movups (%rdi),%xmm2 |
11 movl 240(%rdx),%eax | 11 movl 240(%rdx),%eax |
12 movups (%rdx),%xmm0 | 12 movups (%rdx),%xmm0 |
13 movups 16(%rdx),%xmm1 | 13 movups 16(%rdx),%xmm1 |
14 leaq 32(%rdx),%rdx | 14 leaq 32(%rdx),%rdx |
15 xorps %xmm0,%xmm2 | 15 xorps %xmm0,%xmm2 |
16 .Loop_enc1_1: | 16 .Loop_enc1_1: |
17 .byte 102,15,56,220,209 | 17 .byte 102,15,56,220,209 |
18 decl %eax | 18 decl %eax |
19 movups (%rdx),%xmm1 | 19 movups (%rdx),%xmm1 |
20 leaq 16(%rdx),%rdx | 20 leaq 16(%rdx),%rdx |
21 jnz .Loop_enc1_1 | 21 jnz .Loop_enc1_1 |
22 .byte 102,15,56,221,209 | 22 .byte 102,15,56,221,209 |
23 pxor %xmm0,%xmm0 | |
24 pxor %xmm1,%xmm1 | |
25 movups %xmm2,(%rsi) | 23 movups %xmm2,(%rsi) |
26 pxor %xmm2,%xmm2 | |
27 .byte 0xf3,0xc3 | 24 .byte 0xf3,0xc3 |
28 .size aesni_encrypt,.-aesni_encrypt | 25 .size aesni_encrypt,.-aesni_encrypt |
29 | 26 |
30 .globl aesni_decrypt | 27 .globl aesni_decrypt |
31 .hidden aesni_decrypt | 28 .hidden aesni_decrypt |
32 .type aesni_decrypt,@function | 29 .type aesni_decrypt,@function |
33 .align 16 | 30 .align 16 |
34 aesni_decrypt: | 31 aesni_decrypt: |
35 movups (%rdi),%xmm2 | 32 movups (%rdi),%xmm2 |
36 movl 240(%rdx),%eax | 33 movl 240(%rdx),%eax |
37 movups (%rdx),%xmm0 | 34 movups (%rdx),%xmm0 |
38 movups 16(%rdx),%xmm1 | 35 movups 16(%rdx),%xmm1 |
39 leaq 32(%rdx),%rdx | 36 leaq 32(%rdx),%rdx |
40 xorps %xmm0,%xmm2 | 37 xorps %xmm0,%xmm2 |
41 .Loop_dec1_2: | 38 .Loop_dec1_2: |
42 .byte 102,15,56,222,209 | 39 .byte 102,15,56,222,209 |
43 decl %eax | 40 decl %eax |
44 movups (%rdx),%xmm1 | 41 movups (%rdx),%xmm1 |
45 leaq 16(%rdx),%rdx | 42 leaq 16(%rdx),%rdx |
46 jnz .Loop_dec1_2 | 43 jnz .Loop_dec1_2 |
47 .byte 102,15,56,223,209 | 44 .byte 102,15,56,223,209 |
48 pxor %xmm0,%xmm0 | |
49 pxor %xmm1,%xmm1 | |
50 movups %xmm2,(%rsi) | 45 movups %xmm2,(%rsi) |
51 pxor %xmm2,%xmm2 | |
52 .byte 0xf3,0xc3 | 46 .byte 0xf3,0xc3 |
53 .size aesni_decrypt, .-aesni_decrypt | 47 .size aesni_decrypt, .-aesni_decrypt |
54 .type _aesni_encrypt2,@function | 48 .type _aesni_encrypt2,@function |
55 .align 16 | 49 .align 16 |
56 _aesni_encrypt2: | 50 _aesni_encrypt2: |
57 movups (%rcx),%xmm0 | 51 movups (%rcx),%xmm0 |
58 shll $4,%eax | 52 shll $4,%eax |
59 movups 16(%rcx),%xmm1 | 53 movups 16(%rcx),%xmm1 |
60 xorps %xmm0,%xmm2 | 54 xorps %xmm0,%xmm2 |
61 xorps %xmm0,%xmm3 | 55 xorps %xmm0,%xmm3 |
(...skipping 205 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
267 pxor %xmm0,%xmm3 | 261 pxor %xmm0,%xmm3 |
268 pxor %xmm0,%xmm4 | 262 pxor %xmm0,%xmm4 |
269 .byte 102,15,56,220,209 | 263 .byte 102,15,56,220,209 |
270 leaq 32(%rcx,%rax,1),%rcx | 264 leaq 32(%rcx,%rax,1),%rcx |
271 negq %rax | 265 negq %rax |
272 .byte 102,15,56,220,217 | 266 .byte 102,15,56,220,217 |
273 pxor %xmm0,%xmm5 | 267 pxor %xmm0,%xmm5 |
274 pxor %xmm0,%xmm6 | 268 pxor %xmm0,%xmm6 |
275 .byte 102,15,56,220,225 | 269 .byte 102,15,56,220,225 |
276 pxor %xmm0,%xmm7 | 270 pxor %xmm0,%xmm7 |
277 movups (%rcx,%rax,1),%xmm0 | |
278 addq $16,%rax | 271 addq $16,%rax |
| 272 .byte 102,15,56,220,233 |
| 273 .byte 102,15,56,220,241 |
| 274 .byte 102,15,56,220,249 |
| 275 movups -16(%rcx,%rax,1),%xmm0 |
279 jmp .Lenc_loop6_enter | 276 jmp .Lenc_loop6_enter |
280 .align 16 | 277 .align 16 |
281 .Lenc_loop6: | 278 .Lenc_loop6: |
282 .byte 102,15,56,220,209 | 279 .byte 102,15,56,220,209 |
283 .byte 102,15,56,220,217 | 280 .byte 102,15,56,220,217 |
284 .byte 102,15,56,220,225 | 281 .byte 102,15,56,220,225 |
285 .Lenc_loop6_enter: | |
286 .byte 102,15,56,220,233 | 282 .byte 102,15,56,220,233 |
287 .byte 102,15,56,220,241 | 283 .byte 102,15,56,220,241 |
288 .byte 102,15,56,220,249 | 284 .byte 102,15,56,220,249 |
| 285 .Lenc_loop6_enter: |
289 movups (%rcx,%rax,1),%xmm1 | 286 movups (%rcx,%rax,1),%xmm1 |
290 addq $32,%rax | 287 addq $32,%rax |
291 .byte 102,15,56,220,208 | 288 .byte 102,15,56,220,208 |
292 .byte 102,15,56,220,216 | 289 .byte 102,15,56,220,216 |
293 .byte 102,15,56,220,224 | 290 .byte 102,15,56,220,224 |
294 .byte 102,15,56,220,232 | 291 .byte 102,15,56,220,232 |
295 .byte 102,15,56,220,240 | 292 .byte 102,15,56,220,240 |
296 .byte 102,15,56,220,248 | 293 .byte 102,15,56,220,248 |
297 movups -16(%rcx,%rax,1),%xmm0 | 294 movups -16(%rcx,%rax,1),%xmm0 |
298 jnz .Lenc_loop6 | 295 jnz .Lenc_loop6 |
(...skipping 22 matching lines...) Expand all Loading... |
321 pxor %xmm0,%xmm3 | 318 pxor %xmm0,%xmm3 |
322 pxor %xmm0,%xmm4 | 319 pxor %xmm0,%xmm4 |
323 .byte 102,15,56,222,209 | 320 .byte 102,15,56,222,209 |
324 leaq 32(%rcx,%rax,1),%rcx | 321 leaq 32(%rcx,%rax,1),%rcx |
325 negq %rax | 322 negq %rax |
326 .byte 102,15,56,222,217 | 323 .byte 102,15,56,222,217 |
327 pxor %xmm0,%xmm5 | 324 pxor %xmm0,%xmm5 |
328 pxor %xmm0,%xmm6 | 325 pxor %xmm0,%xmm6 |
329 .byte 102,15,56,222,225 | 326 .byte 102,15,56,222,225 |
330 pxor %xmm0,%xmm7 | 327 pxor %xmm0,%xmm7 |
331 movups (%rcx,%rax,1),%xmm0 | |
332 addq $16,%rax | 328 addq $16,%rax |
| 329 .byte 102,15,56,222,233 |
| 330 .byte 102,15,56,222,241 |
| 331 .byte 102,15,56,222,249 |
| 332 movups -16(%rcx,%rax,1),%xmm0 |
333 jmp .Ldec_loop6_enter | 333 jmp .Ldec_loop6_enter |
334 .align 16 | 334 .align 16 |
335 .Ldec_loop6: | 335 .Ldec_loop6: |
336 .byte 102,15,56,222,209 | 336 .byte 102,15,56,222,209 |
337 .byte 102,15,56,222,217 | 337 .byte 102,15,56,222,217 |
338 .byte 102,15,56,222,225 | 338 .byte 102,15,56,222,225 |
339 .Ldec_loop6_enter: | |
340 .byte 102,15,56,222,233 | 339 .byte 102,15,56,222,233 |
341 .byte 102,15,56,222,241 | 340 .byte 102,15,56,222,241 |
342 .byte 102,15,56,222,249 | 341 .byte 102,15,56,222,249 |
| 342 .Ldec_loop6_enter: |
343 movups (%rcx,%rax,1),%xmm1 | 343 movups (%rcx,%rax,1),%xmm1 |
344 addq $32,%rax | 344 addq $32,%rax |
345 .byte 102,15,56,222,208 | 345 .byte 102,15,56,222,208 |
346 .byte 102,15,56,222,216 | 346 .byte 102,15,56,222,216 |
347 .byte 102,15,56,222,224 | 347 .byte 102,15,56,222,224 |
348 .byte 102,15,56,222,232 | 348 .byte 102,15,56,222,232 |
349 .byte 102,15,56,222,240 | 349 .byte 102,15,56,222,240 |
350 .byte 102,15,56,222,248 | 350 .byte 102,15,56,222,248 |
351 movups -16(%rcx,%rax,1),%xmm0 | 351 movups -16(%rcx,%rax,1),%xmm0 |
352 jnz .Ldec_loop6 | 352 jnz .Ldec_loop6 |
(...skipping 19 matching lines...) Expand all Loading... |
372 shll $4,%eax | 372 shll $4,%eax |
373 movups 16(%rcx),%xmm1 | 373 movups 16(%rcx),%xmm1 |
374 xorps %xmm0,%xmm2 | 374 xorps %xmm0,%xmm2 |
375 xorps %xmm0,%xmm3 | 375 xorps %xmm0,%xmm3 |
376 pxor %xmm0,%xmm4 | 376 pxor %xmm0,%xmm4 |
377 pxor %xmm0,%xmm5 | 377 pxor %xmm0,%xmm5 |
378 pxor %xmm0,%xmm6 | 378 pxor %xmm0,%xmm6 |
379 leaq 32(%rcx,%rax,1),%rcx | 379 leaq 32(%rcx,%rax,1),%rcx |
380 negq %rax | 380 negq %rax |
381 .byte 102,15,56,220,209 | 381 .byte 102,15,56,220,209 |
| 382 addq $16,%rax |
382 pxor %xmm0,%xmm7 | 383 pxor %xmm0,%xmm7 |
| 384 .byte 102,15,56,220,217 |
383 pxor %xmm0,%xmm8 | 385 pxor %xmm0,%xmm8 |
384 .byte 102,15,56,220,217 | |
385 pxor %xmm0,%xmm9 | 386 pxor %xmm0,%xmm9 |
386 » movups» (%rcx,%rax,1),%xmm0 | 387 .byte» 102,15,56,220,225 |
387 » addq» $16,%rax | 388 .byte» 102,15,56,220,233 |
388 » jmp» .Lenc_loop8_inner | 389 .byte» 102,15,56,220,241 |
| 390 .byte» 102,15,56,220,249 |
| 391 .byte» 102,68,15,56,220,193 |
| 392 .byte» 102,68,15,56,220,201 |
| 393 » movups» -16(%rcx,%rax,1),%xmm0 |
| 394 » jmp» .Lenc_loop8_enter |
389 .align 16 | 395 .align 16 |
390 .Lenc_loop8: | 396 .Lenc_loop8: |
391 .byte 102,15,56,220,209 | 397 .byte 102,15,56,220,209 |
392 .byte 102,15,56,220,217 | 398 .byte 102,15,56,220,217 |
393 .Lenc_loop8_inner: | |
394 .byte 102,15,56,220,225 | 399 .byte 102,15,56,220,225 |
395 .byte 102,15,56,220,233 | 400 .byte 102,15,56,220,233 |
396 .byte 102,15,56,220,241 | 401 .byte 102,15,56,220,241 |
397 .byte 102,15,56,220,249 | 402 .byte 102,15,56,220,249 |
398 .byte 102,68,15,56,220,193 | 403 .byte 102,68,15,56,220,193 |
399 .byte 102,68,15,56,220,201 | 404 .byte 102,68,15,56,220,201 |
400 .Lenc_loop8_enter: | 405 .Lenc_loop8_enter: |
401 movups (%rcx,%rax,1),%xmm1 | 406 movups (%rcx,%rax,1),%xmm1 |
402 addq $32,%rax | 407 addq $32,%rax |
403 .byte 102,15,56,220,208 | 408 .byte 102,15,56,220,208 |
(...skipping 32 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
436 shll $4,%eax | 441 shll $4,%eax |
437 movups 16(%rcx),%xmm1 | 442 movups 16(%rcx),%xmm1 |
438 xorps %xmm0,%xmm2 | 443 xorps %xmm0,%xmm2 |
439 xorps %xmm0,%xmm3 | 444 xorps %xmm0,%xmm3 |
440 pxor %xmm0,%xmm4 | 445 pxor %xmm0,%xmm4 |
441 pxor %xmm0,%xmm5 | 446 pxor %xmm0,%xmm5 |
442 pxor %xmm0,%xmm6 | 447 pxor %xmm0,%xmm6 |
443 leaq 32(%rcx,%rax,1),%rcx | 448 leaq 32(%rcx,%rax,1),%rcx |
444 negq %rax | 449 negq %rax |
445 .byte 102,15,56,222,209 | 450 .byte 102,15,56,222,209 |
| 451 addq $16,%rax |
446 pxor %xmm0,%xmm7 | 452 pxor %xmm0,%xmm7 |
| 453 .byte 102,15,56,222,217 |
447 pxor %xmm0,%xmm8 | 454 pxor %xmm0,%xmm8 |
448 .byte 102,15,56,222,217 | |
449 pxor %xmm0,%xmm9 | 455 pxor %xmm0,%xmm9 |
450 » movups» (%rcx,%rax,1),%xmm0 | 456 .byte» 102,15,56,222,225 |
451 » addq» $16,%rax | 457 .byte» 102,15,56,222,233 |
452 » jmp» .Ldec_loop8_inner | 458 .byte» 102,15,56,222,241 |
| 459 .byte» 102,15,56,222,249 |
| 460 .byte» 102,68,15,56,222,193 |
| 461 .byte» 102,68,15,56,222,201 |
| 462 » movups» -16(%rcx,%rax,1),%xmm0 |
| 463 » jmp» .Ldec_loop8_enter |
453 .align 16 | 464 .align 16 |
454 .Ldec_loop8: | 465 .Ldec_loop8: |
455 .byte 102,15,56,222,209 | 466 .byte 102,15,56,222,209 |
456 .byte 102,15,56,222,217 | 467 .byte 102,15,56,222,217 |
457 .Ldec_loop8_inner: | |
458 .byte 102,15,56,222,225 | 468 .byte 102,15,56,222,225 |
459 .byte 102,15,56,222,233 | 469 .byte 102,15,56,222,233 |
460 .byte 102,15,56,222,241 | 470 .byte 102,15,56,222,241 |
461 .byte 102,15,56,222,249 | 471 .byte 102,15,56,222,249 |
462 .byte 102,68,15,56,222,193 | 472 .byte 102,68,15,56,222,193 |
463 .byte 102,68,15,56,222,201 | 473 .byte 102,68,15,56,222,201 |
464 .Ldec_loop8_enter: | 474 .Ldec_loop8_enter: |
465 movups (%rcx,%rax,1),%xmm1 | 475 movups (%rcx,%rax,1),%xmm1 |
466 addq $32,%rax | 476 addq $32,%rax |
467 .byte 102,15,56,222,208 | 477 .byte 102,15,56,222,208 |
(...skipping 107 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
575 cmpq $64,%rdx | 585 cmpq $64,%rdx |
576 jb .Lecb_enc_three | 586 jb .Lecb_enc_three |
577 movups 48(%rdi),%xmm5 | 587 movups 48(%rdi),%xmm5 |
578 je .Lecb_enc_four | 588 je .Lecb_enc_four |
579 movups 64(%rdi),%xmm6 | 589 movups 64(%rdi),%xmm6 |
580 cmpq $96,%rdx | 590 cmpq $96,%rdx |
581 jb .Lecb_enc_five | 591 jb .Lecb_enc_five |
582 movups 80(%rdi),%xmm7 | 592 movups 80(%rdi),%xmm7 |
583 je .Lecb_enc_six | 593 je .Lecb_enc_six |
584 movdqu 96(%rdi),%xmm8 | 594 movdqu 96(%rdi),%xmm8 |
585 xorps %xmm9,%xmm9 | |
586 call _aesni_encrypt8 | 595 call _aesni_encrypt8 |
587 movups %xmm2,(%rsi) | 596 movups %xmm2,(%rsi) |
588 movups %xmm3,16(%rsi) | 597 movups %xmm3,16(%rsi) |
589 movups %xmm4,32(%rsi) | 598 movups %xmm4,32(%rsi) |
590 movups %xmm5,48(%rsi) | 599 movups %xmm5,48(%rsi) |
591 movups %xmm6,64(%rsi) | 600 movups %xmm6,64(%rsi) |
592 movups %xmm7,80(%rsi) | 601 movups %xmm7,80(%rsi) |
593 movups %xmm8,96(%rsi) | 602 movups %xmm8,96(%rsi) |
594 jmp .Lecb_ret | 603 jmp .Lecb_ret |
595 .align 16 | 604 .align 16 |
(...skipping 93 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
689 leaq 128(%rdi),%rdi | 698 leaq 128(%rdi),%rdi |
690 .Lecb_dec_loop8_enter: | 699 .Lecb_dec_loop8_enter: |
691 | 700 |
692 call _aesni_decrypt8 | 701 call _aesni_decrypt8 |
693 | 702 |
694 movups (%r11),%xmm0 | 703 movups (%r11),%xmm0 |
695 subq $128,%rdx | 704 subq $128,%rdx |
696 jnc .Lecb_dec_loop8 | 705 jnc .Lecb_dec_loop8 |
697 | 706 |
698 movups %xmm2,(%rsi) | 707 movups %xmm2,(%rsi) |
699 pxor %xmm2,%xmm2 | |
700 movq %r11,%rcx | 708 movq %r11,%rcx |
701 movups %xmm3,16(%rsi) | 709 movups %xmm3,16(%rsi) |
702 pxor %xmm3,%xmm3 | |
703 movl %r10d,%eax | 710 movl %r10d,%eax |
704 movups %xmm4,32(%rsi) | 711 movups %xmm4,32(%rsi) |
705 pxor %xmm4,%xmm4 | |
706 movups %xmm5,48(%rsi) | 712 movups %xmm5,48(%rsi) |
707 pxor %xmm5,%xmm5 | |
708 movups %xmm6,64(%rsi) | 713 movups %xmm6,64(%rsi) |
709 pxor %xmm6,%xmm6 | |
710 movups %xmm7,80(%rsi) | 714 movups %xmm7,80(%rsi) |
711 pxor %xmm7,%xmm7 | |
712 movups %xmm8,96(%rsi) | 715 movups %xmm8,96(%rsi) |
713 pxor %xmm8,%xmm8 | |
714 movups %xmm9,112(%rsi) | 716 movups %xmm9,112(%rsi) |
715 pxor %xmm9,%xmm9 | |
716 leaq 128(%rsi),%rsi | 717 leaq 128(%rsi),%rsi |
717 addq $128,%rdx | 718 addq $128,%rdx |
718 jz .Lecb_ret | 719 jz .Lecb_ret |
719 | 720 |
720 .Lecb_dec_tail: | 721 .Lecb_dec_tail: |
721 movups (%rdi),%xmm2 | 722 movups (%rdi),%xmm2 |
722 cmpq $32,%rdx | 723 cmpq $32,%rdx |
723 jb .Lecb_dec_one | 724 jb .Lecb_dec_one |
724 movups 16(%rdi),%xmm3 | 725 movups 16(%rdi),%xmm3 |
725 je .Lecb_dec_two | 726 je .Lecb_dec_two |
726 movups 32(%rdi),%xmm4 | 727 movups 32(%rdi),%xmm4 |
727 cmpq $64,%rdx | 728 cmpq $64,%rdx |
728 jb .Lecb_dec_three | 729 jb .Lecb_dec_three |
729 movups 48(%rdi),%xmm5 | 730 movups 48(%rdi),%xmm5 |
730 je .Lecb_dec_four | 731 je .Lecb_dec_four |
731 movups 64(%rdi),%xmm6 | 732 movups 64(%rdi),%xmm6 |
732 cmpq $96,%rdx | 733 cmpq $96,%rdx |
733 jb .Lecb_dec_five | 734 jb .Lecb_dec_five |
734 movups 80(%rdi),%xmm7 | 735 movups 80(%rdi),%xmm7 |
735 je .Lecb_dec_six | 736 je .Lecb_dec_six |
736 movups 96(%rdi),%xmm8 | 737 movups 96(%rdi),%xmm8 |
737 movups (%rcx),%xmm0 | 738 movups (%rcx),%xmm0 |
738 xorps %xmm9,%xmm9 | |
739 call _aesni_decrypt8 | 739 call _aesni_decrypt8 |
740 movups %xmm2,(%rsi) | 740 movups %xmm2,(%rsi) |
741 pxor %xmm2,%xmm2 | |
742 movups %xmm3,16(%rsi) | 741 movups %xmm3,16(%rsi) |
743 pxor %xmm3,%xmm3 | |
744 movups %xmm4,32(%rsi) | 742 movups %xmm4,32(%rsi) |
745 pxor %xmm4,%xmm4 | |
746 movups %xmm5,48(%rsi) | 743 movups %xmm5,48(%rsi) |
747 pxor %xmm5,%xmm5 | |
748 movups %xmm6,64(%rsi) | 744 movups %xmm6,64(%rsi) |
749 pxor %xmm6,%xmm6 | |
750 movups %xmm7,80(%rsi) | 745 movups %xmm7,80(%rsi) |
751 pxor %xmm7,%xmm7 | |
752 movups %xmm8,96(%rsi) | 746 movups %xmm8,96(%rsi) |
753 pxor %xmm8,%xmm8 | |
754 pxor %xmm9,%xmm9 | |
755 jmp .Lecb_ret | 747 jmp .Lecb_ret |
756 .align 16 | 748 .align 16 |
757 .Lecb_dec_one: | 749 .Lecb_dec_one: |
758 movups (%rcx),%xmm0 | 750 movups (%rcx),%xmm0 |
759 movups 16(%rcx),%xmm1 | 751 movups 16(%rcx),%xmm1 |
760 leaq 32(%rcx),%rcx | 752 leaq 32(%rcx),%rcx |
761 xorps %xmm0,%xmm2 | 753 xorps %xmm0,%xmm2 |
762 .Loop_dec1_4: | 754 .Loop_dec1_4: |
763 .byte 102,15,56,222,209 | 755 .byte 102,15,56,222,209 |
764 decl %eax | 756 decl %eax |
765 movups (%rcx),%xmm1 | 757 movups (%rcx),%xmm1 |
766 leaq 16(%rcx),%rcx | 758 leaq 16(%rcx),%rcx |
767 jnz .Loop_dec1_4 | 759 jnz .Loop_dec1_4 |
768 .byte 102,15,56,223,209 | 760 .byte 102,15,56,223,209 |
769 movups %xmm2,(%rsi) | 761 movups %xmm2,(%rsi) |
770 pxor %xmm2,%xmm2 | |
771 jmp .Lecb_ret | 762 jmp .Lecb_ret |
772 .align 16 | 763 .align 16 |
773 .Lecb_dec_two: | 764 .Lecb_dec_two: |
774 call _aesni_decrypt2 | 765 call _aesni_decrypt2 |
775 movups %xmm2,(%rsi) | 766 movups %xmm2,(%rsi) |
776 pxor %xmm2,%xmm2 | |
777 movups %xmm3,16(%rsi) | 767 movups %xmm3,16(%rsi) |
778 pxor %xmm3,%xmm3 | |
779 jmp .Lecb_ret | 768 jmp .Lecb_ret |
780 .align 16 | 769 .align 16 |
781 .Lecb_dec_three: | 770 .Lecb_dec_three: |
782 call _aesni_decrypt3 | 771 call _aesni_decrypt3 |
783 movups %xmm2,(%rsi) | 772 movups %xmm2,(%rsi) |
784 pxor %xmm2,%xmm2 | |
785 movups %xmm3,16(%rsi) | 773 movups %xmm3,16(%rsi) |
786 pxor %xmm3,%xmm3 | |
787 movups %xmm4,32(%rsi) | 774 movups %xmm4,32(%rsi) |
788 pxor %xmm4,%xmm4 | |
789 jmp .Lecb_ret | 775 jmp .Lecb_ret |
790 .align 16 | 776 .align 16 |
791 .Lecb_dec_four: | 777 .Lecb_dec_four: |
792 call _aesni_decrypt4 | 778 call _aesni_decrypt4 |
793 movups %xmm2,(%rsi) | 779 movups %xmm2,(%rsi) |
794 pxor %xmm2,%xmm2 | |
795 movups %xmm3,16(%rsi) | 780 movups %xmm3,16(%rsi) |
796 pxor %xmm3,%xmm3 | |
797 movups %xmm4,32(%rsi) | 781 movups %xmm4,32(%rsi) |
798 pxor %xmm4,%xmm4 | |
799 movups %xmm5,48(%rsi) | 782 movups %xmm5,48(%rsi) |
800 pxor %xmm5,%xmm5 | |
801 jmp .Lecb_ret | 783 jmp .Lecb_ret |
802 .align 16 | 784 .align 16 |
803 .Lecb_dec_five: | 785 .Lecb_dec_five: |
804 xorps %xmm7,%xmm7 | 786 xorps %xmm7,%xmm7 |
805 call _aesni_decrypt6 | 787 call _aesni_decrypt6 |
806 movups %xmm2,(%rsi) | 788 movups %xmm2,(%rsi) |
807 pxor %xmm2,%xmm2 | |
808 movups %xmm3,16(%rsi) | 789 movups %xmm3,16(%rsi) |
809 pxor %xmm3,%xmm3 | |
810 movups %xmm4,32(%rsi) | 790 movups %xmm4,32(%rsi) |
811 pxor %xmm4,%xmm4 | |
812 movups %xmm5,48(%rsi) | 791 movups %xmm5,48(%rsi) |
813 pxor %xmm5,%xmm5 | |
814 movups %xmm6,64(%rsi) | 792 movups %xmm6,64(%rsi) |
815 pxor %xmm6,%xmm6 | |
816 pxor %xmm7,%xmm7 | |
817 jmp .Lecb_ret | 793 jmp .Lecb_ret |
818 .align 16 | 794 .align 16 |
819 .Lecb_dec_six: | 795 .Lecb_dec_six: |
820 call _aesni_decrypt6 | 796 call _aesni_decrypt6 |
821 movups %xmm2,(%rsi) | 797 movups %xmm2,(%rsi) |
822 pxor %xmm2,%xmm2 | |
823 movups %xmm3,16(%rsi) | 798 movups %xmm3,16(%rsi) |
824 pxor %xmm3,%xmm3 | |
825 movups %xmm4,32(%rsi) | 799 movups %xmm4,32(%rsi) |
826 pxor %xmm4,%xmm4 | |
827 movups %xmm5,48(%rsi) | 800 movups %xmm5,48(%rsi) |
828 pxor %xmm5,%xmm5 | |
829 movups %xmm6,64(%rsi) | 801 movups %xmm6,64(%rsi) |
830 pxor %xmm6,%xmm6 | |
831 movups %xmm7,80(%rsi) | 802 movups %xmm7,80(%rsi) |
832 pxor %xmm7,%xmm7 | |
833 | 803 |
834 .Lecb_ret: | 804 .Lecb_ret: |
835 xorps %xmm0,%xmm0 | |
836 pxor %xmm1,%xmm1 | |
837 .byte 0xf3,0xc3 | 805 .byte 0xf3,0xc3 |
838 .size aesni_ecb_encrypt,.-aesni_ecb_encrypt | 806 .size aesni_ecb_encrypt,.-aesni_ecb_encrypt |
839 .globl aesni_ccm64_encrypt_blocks | 807 .globl aesni_ccm64_encrypt_blocks |
840 .hidden aesni_ccm64_encrypt_blocks | 808 .hidden aesni_ccm64_encrypt_blocks |
841 .type aesni_ccm64_encrypt_blocks,@function | 809 .type aesni_ccm64_encrypt_blocks,@function |
842 .align 16 | 810 .align 16 |
843 aesni_ccm64_encrypt_blocks: | 811 aesni_ccm64_encrypt_blocks: |
844 movl 240(%rcx),%eax | 812 movl 240(%rcx),%eax |
845 movdqu (%r8),%xmm6 | 813 movdqu (%r8),%xmm6 |
846 movdqa .Lincrement64(%rip),%xmm9 | 814 movdqa .Lincrement64(%rip),%xmm9 |
(...skipping 37 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
884 .byte 102,15,56,221,216 | 852 .byte 102,15,56,221,216 |
885 | 853 |
886 leaq 16(%rdi),%rdi | 854 leaq 16(%rdi),%rdi |
887 xorps %xmm2,%xmm8 | 855 xorps %xmm2,%xmm8 |
888 movdqa %xmm6,%xmm2 | 856 movdqa %xmm6,%xmm2 |
889 movups %xmm8,(%rsi) | 857 movups %xmm8,(%rsi) |
890 .byte 102,15,56,0,215 | 858 .byte 102,15,56,0,215 |
891 leaq 16(%rsi),%rsi | 859 leaq 16(%rsi),%rsi |
892 jnz .Lccm64_enc_outer | 860 jnz .Lccm64_enc_outer |
893 | 861 |
894 pxor %xmm0,%xmm0 | |
895 pxor %xmm1,%xmm1 | |
896 pxor %xmm2,%xmm2 | |
897 movups %xmm3,(%r9) | 862 movups %xmm3,(%r9) |
898 pxor %xmm3,%xmm3 | |
899 pxor %xmm8,%xmm8 | |
900 pxor %xmm6,%xmm6 | |
901 .byte 0xf3,0xc3 | 863 .byte 0xf3,0xc3 |
902 .size aesni_ccm64_encrypt_blocks,.-aesni_ccm64_encrypt_blocks | 864 .size aesni_ccm64_encrypt_blocks,.-aesni_ccm64_encrypt_blocks |
903 .globl aesni_ccm64_decrypt_blocks | 865 .globl aesni_ccm64_decrypt_blocks |
904 .hidden aesni_ccm64_decrypt_blocks | 866 .hidden aesni_ccm64_decrypt_blocks |
905 .type aesni_ccm64_decrypt_blocks,@function | 867 .type aesni_ccm64_decrypt_blocks,@function |
906 .align 16 | 868 .align 16 |
907 aesni_ccm64_decrypt_blocks: | 869 aesni_ccm64_decrypt_blocks: |
908 movl 240(%rcx),%eax | 870 movl 240(%rcx),%eax |
909 movups (%r8),%xmm6 | 871 movups (%r8),%xmm6 |
910 movdqu (%r9),%xmm3 | 872 movdqu (%r9),%xmm3 |
(...skipping 71 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
982 xorps %xmm0,%xmm8 | 944 xorps %xmm0,%xmm8 |
983 leaq 32(%r11),%r11 | 945 leaq 32(%r11),%r11 |
984 xorps %xmm8,%xmm3 | 946 xorps %xmm8,%xmm3 |
985 .Loop_enc1_6: | 947 .Loop_enc1_6: |
986 .byte 102,15,56,220,217 | 948 .byte 102,15,56,220,217 |
987 decl %eax | 949 decl %eax |
988 movups (%r11),%xmm1 | 950 movups (%r11),%xmm1 |
989 leaq 16(%r11),%r11 | 951 leaq 16(%r11),%r11 |
990 jnz .Loop_enc1_6 | 952 jnz .Loop_enc1_6 |
991 .byte 102,15,56,221,217 | 953 .byte 102,15,56,221,217 |
992 pxor %xmm0,%xmm0 | |
993 pxor %xmm1,%xmm1 | |
994 pxor %xmm2,%xmm2 | |
995 movups %xmm3,(%r9) | 954 movups %xmm3,(%r9) |
996 pxor %xmm3,%xmm3 | |
997 pxor %xmm8,%xmm8 | |
998 pxor %xmm6,%xmm6 | |
999 .byte 0xf3,0xc3 | 955 .byte 0xf3,0xc3 |
1000 .size aesni_ccm64_decrypt_blocks,.-aesni_ccm64_decrypt_blocks | 956 .size aesni_ccm64_decrypt_blocks,.-aesni_ccm64_decrypt_blocks |
1001 .globl aesni_ctr32_encrypt_blocks | 957 .globl aesni_ctr32_encrypt_blocks |
1002 .hidden aesni_ctr32_encrypt_blocks | 958 .hidden aesni_ctr32_encrypt_blocks |
1003 .type aesni_ctr32_encrypt_blocks,@function | 959 .type aesni_ctr32_encrypt_blocks,@function |
1004 .align 16 | 960 .align 16 |
1005 aesni_ctr32_encrypt_blocks: | 961 aesni_ctr32_encrypt_blocks: |
1006 cmpq $1,%rdx | |
1007 jne .Lctr32_bulk | |
1008 | |
1009 | |
1010 | |
1011 movups (%r8),%xmm2 | |
1012 movups (%rdi),%xmm3 | |
1013 movl 240(%rcx),%edx | |
1014 movups (%rcx),%xmm0 | |
1015 movups 16(%rcx),%xmm1 | |
1016 leaq 32(%rcx),%rcx | |
1017 xorps %xmm0,%xmm2 | |
1018 .Loop_enc1_7: | |
1019 .byte 102,15,56,220,209 | |
1020 decl %edx | |
1021 movups (%rcx),%xmm1 | |
1022 leaq 16(%rcx),%rcx | |
1023 jnz .Loop_enc1_7 | |
1024 .byte 102,15,56,221,209 | |
1025 pxor %xmm0,%xmm0 | |
1026 pxor %xmm1,%xmm1 | |
1027 xorps %xmm3,%xmm2 | |
1028 pxor %xmm3,%xmm3 | |
1029 movups %xmm2,(%rsi) | |
1030 xorps %xmm2,%xmm2 | |
1031 jmp .Lctr32_epilogue | |
1032 | |
1033 .align 16 | |
1034 .Lctr32_bulk: | |
1035 leaq (%rsp),%rax | 962 leaq (%rsp),%rax |
1036 pushq %rbp | 963 pushq %rbp |
1037 subq $128,%rsp | 964 subq $128,%rsp |
1038 andq $-16,%rsp | 965 andq $-16,%rsp |
1039 leaq -8(%rax),%rbp | 966 leaq -8(%rax),%rbp |
1040 | 967 |
1041 | 968 » cmpq» $1,%rdx |
1042 | 969 » je» .Lctr32_one_shortcut |
1043 | 970 |
1044 movdqu (%r8),%xmm2 | 971 movdqu (%r8),%xmm2 |
1045 movdqu (%rcx),%xmm0 | 972 movdqu (%rcx),%xmm0 |
1046 movl 12(%r8),%r8d | 973 movl 12(%r8),%r8d |
1047 pxor %xmm0,%xmm2 | 974 pxor %xmm0,%xmm2 |
1048 movl 12(%rcx),%r11d | 975 movl 12(%rcx),%r11d |
1049 movdqa %xmm2,0(%rsp) | 976 movdqa %xmm2,0(%rsp) |
1050 bswapl %r8d | 977 bswapl %r8d |
1051 movdqa %xmm2,%xmm3 | 978 movdqa %xmm2,%xmm3 |
1052 movdqa %xmm2,%xmm4 | 979 movdqa %xmm2,%xmm4 |
(...skipping 370 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1423 leaq 128(%rsi),%rsi | 1350 leaq 128(%rsi),%rsi |
1424 | 1351 |
1425 subq $8,%rdx | 1352 subq $8,%rdx |
1426 jnc .Lctr32_loop8 | 1353 jnc .Lctr32_loop8 |
1427 | 1354 |
1428 addq $8,%rdx | 1355 addq $8,%rdx |
1429 jz .Lctr32_done | 1356 jz .Lctr32_done |
1430 leaq -128(%rcx),%rcx | 1357 leaq -128(%rcx),%rcx |
1431 | 1358 |
1432 .Lctr32_tail: | 1359 .Lctr32_tail: |
1433 | |
1434 | |
1435 leaq 16(%rcx),%rcx | 1360 leaq 16(%rcx),%rcx |
1436 cmpq $4,%rdx | 1361 cmpq $4,%rdx |
1437 jb .Lctr32_loop3 | 1362 jb .Lctr32_loop3 |
1438 je .Lctr32_loop4 | 1363 je .Lctr32_loop4 |
1439 | 1364 |
1440 | |
1441 shll $4,%eax | 1365 shll $4,%eax |
1442 movdqa 96(%rsp),%xmm8 | 1366 movdqa 96(%rsp),%xmm8 |
1443 pxor %xmm9,%xmm9 | 1367 pxor %xmm9,%xmm9 |
1444 | 1368 |
1445 movups 16(%rcx),%xmm0 | 1369 movups 16(%rcx),%xmm0 |
1446 .byte 102,15,56,220,209 | 1370 .byte 102,15,56,220,209 |
1447 .byte 102,15,56,220,217 | 1371 .byte 102,15,56,220,217 |
1448 leaq 32-16(%rcx,%rax,1),%rcx | 1372 leaq 32-16(%rcx,%rax,1),%rcx |
1449 negq %rax | 1373 negq %rax |
1450 .byte 102,15,56,220,225 | 1374 .byte 102,15,56,220,225 |
(...skipping 82 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1533 jb .Lctr32_done | 1457 jb .Lctr32_done |
1534 | 1458 |
1535 movups 16(%rdi),%xmm11 | 1459 movups 16(%rdi),%xmm11 |
1536 xorps %xmm11,%xmm3 | 1460 xorps %xmm11,%xmm3 |
1537 movups %xmm3,16(%rsi) | 1461 movups %xmm3,16(%rsi) |
1538 je .Lctr32_done | 1462 je .Lctr32_done |
1539 | 1463 |
1540 movups 32(%rdi),%xmm12 | 1464 movups 32(%rdi),%xmm12 |
1541 xorps %xmm12,%xmm4 | 1465 xorps %xmm12,%xmm4 |
1542 movups %xmm4,32(%rsi) | 1466 movups %xmm4,32(%rsi) |
| 1467 jmp .Lctr32_done |
1543 | 1468 |
| 1469 .align 16 |
| 1470 .Lctr32_one_shortcut: |
| 1471 movups (%r8),%xmm2 |
| 1472 movups (%rdi),%xmm10 |
| 1473 movl 240(%rcx),%eax |
| 1474 movups (%rcx),%xmm0 |
| 1475 movups 16(%rcx),%xmm1 |
| 1476 leaq 32(%rcx),%rcx |
| 1477 xorps %xmm0,%xmm2 |
| 1478 .Loop_enc1_7: |
| 1479 .byte 102,15,56,220,209 |
| 1480 decl %eax |
| 1481 movups (%rcx),%xmm1 |
| 1482 leaq 16(%rcx),%rcx |
| 1483 jnz .Loop_enc1_7 |
| 1484 .byte 102,15,56,221,209 |
| 1485 xorps %xmm10,%xmm2 |
| 1486 movups %xmm2,(%rsi) |
| 1487 jmp .Lctr32_done |
| 1488 |
| 1489 .align 16 |
1544 .Lctr32_done: | 1490 .Lctr32_done: |
1545 xorps %xmm0,%xmm0 | |
1546 xorl %r11d,%r11d | |
1547 pxor %xmm1,%xmm1 | |
1548 pxor %xmm2,%xmm2 | |
1549 pxor %xmm3,%xmm3 | |
1550 pxor %xmm4,%xmm4 | |
1551 pxor %xmm5,%xmm5 | |
1552 pxor %xmm6,%xmm6 | |
1553 pxor %xmm7,%xmm7 | |
1554 movaps %xmm0,0(%rsp) | |
1555 pxor %xmm8,%xmm8 | |
1556 movaps %xmm0,16(%rsp) | |
1557 pxor %xmm9,%xmm9 | |
1558 movaps %xmm0,32(%rsp) | |
1559 pxor %xmm10,%xmm10 | |
1560 movaps %xmm0,48(%rsp) | |
1561 pxor %xmm11,%xmm11 | |
1562 movaps %xmm0,64(%rsp) | |
1563 pxor %xmm12,%xmm12 | |
1564 movaps %xmm0,80(%rsp) | |
1565 pxor %xmm13,%xmm13 | |
1566 movaps %xmm0,96(%rsp) | |
1567 pxor %xmm14,%xmm14 | |
1568 movaps %xmm0,112(%rsp) | |
1569 pxor %xmm15,%xmm15 | |
1570 leaq (%rbp),%rsp | 1491 leaq (%rbp),%rsp |
1571 popq %rbp | 1492 popq %rbp |
1572 .Lctr32_epilogue: | 1493 .Lctr32_epilogue: |
1573 .byte 0xf3,0xc3 | 1494 .byte 0xf3,0xc3 |
1574 .size aesni_ctr32_encrypt_blocks,.-aesni_ctr32_encrypt_blocks | 1495 .size aesni_ctr32_encrypt_blocks,.-aesni_ctr32_encrypt_blocks |
1575 .globl aesni_xts_encrypt | 1496 .globl aesni_xts_encrypt |
1576 .hidden aesni_xts_encrypt | 1497 .hidden aesni_xts_encrypt |
1577 .type aesni_xts_encrypt,@function | 1498 .type aesni_xts_encrypt,@function |
1578 .align 16 | 1499 .align 16 |
1579 aesni_xts_encrypt: | 1500 aesni_xts_encrypt: |
(...skipping 251 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1831 movups %xmm7,-16(%rsi) | 1752 movups %xmm7,-16(%rsi) |
1832 subq $96,%rdx | 1753 subq $96,%rdx |
1833 jnc .Lxts_enc_grandloop | 1754 jnc .Lxts_enc_grandloop |
1834 | 1755 |
1835 movl $16+96,%eax | 1756 movl $16+96,%eax |
1836 subl %r10d,%eax | 1757 subl %r10d,%eax |
1837 movq %r11,%rcx | 1758 movq %r11,%rcx |
1838 shrl $4,%eax | 1759 shrl $4,%eax |
1839 | 1760 |
1840 .Lxts_enc_short: | 1761 .Lxts_enc_short: |
1841 | |
1842 movl %eax,%r10d | 1762 movl %eax,%r10d |
1843 pxor %xmm0,%xmm10 | 1763 pxor %xmm0,%xmm10 |
1844 addq $96,%rdx | 1764 addq $96,%rdx |
1845 jz .Lxts_enc_done | 1765 jz .Lxts_enc_done |
1846 | 1766 |
1847 pxor %xmm0,%xmm11 | 1767 pxor %xmm0,%xmm11 |
1848 cmpq $32,%rdx | 1768 cmpq $32,%rdx |
1849 jb .Lxts_enc_one | 1769 jb .Lxts_enc_one |
1850 pxor %xmm0,%xmm12 | 1770 pxor %xmm0,%xmm12 |
1851 je .Lxts_enc_two | 1771 je .Lxts_enc_two |
1852 | 1772 |
1853 pxor %xmm0,%xmm13 | 1773 pxor %xmm0,%xmm13 |
1854 cmpq $64,%rdx | 1774 cmpq $64,%rdx |
1855 jb .Lxts_enc_three | 1775 jb .Lxts_enc_three |
1856 pxor %xmm0,%xmm14 | 1776 pxor %xmm0,%xmm14 |
1857 je .Lxts_enc_four | 1777 je .Lxts_enc_four |
1858 | 1778 |
1859 movdqu (%rdi),%xmm2 | 1779 movdqu (%rdi),%xmm2 |
1860 movdqu 16(%rdi),%xmm3 | 1780 movdqu 16(%rdi),%xmm3 |
1861 movdqu 32(%rdi),%xmm4 | 1781 movdqu 32(%rdi),%xmm4 |
1862 pxor %xmm10,%xmm2 | 1782 pxor %xmm10,%xmm2 |
1863 movdqu 48(%rdi),%xmm5 | 1783 movdqu 48(%rdi),%xmm5 |
1864 pxor %xmm11,%xmm3 | 1784 pxor %xmm11,%xmm3 |
1865 movdqu 64(%rdi),%xmm6 | 1785 movdqu 64(%rdi),%xmm6 |
1866 leaq 80(%rdi),%rdi | 1786 leaq 80(%rdi),%rdi |
1867 pxor %xmm12,%xmm4 | 1787 pxor %xmm12,%xmm4 |
1868 pxor %xmm13,%xmm5 | 1788 pxor %xmm13,%xmm5 |
1869 pxor %xmm14,%xmm6 | 1789 pxor %xmm14,%xmm6 |
1870 pxor %xmm7,%xmm7 | |
1871 | 1790 |
1872 call _aesni_encrypt6 | 1791 call _aesni_encrypt6 |
1873 | 1792 |
1874 xorps %xmm10,%xmm2 | 1793 xorps %xmm10,%xmm2 |
1875 movdqa %xmm15,%xmm10 | 1794 movdqa %xmm15,%xmm10 |
1876 xorps %xmm11,%xmm3 | 1795 xorps %xmm11,%xmm3 |
1877 xorps %xmm12,%xmm4 | 1796 xorps %xmm12,%xmm4 |
1878 movdqu %xmm2,(%rsi) | 1797 movdqu %xmm2,(%rsi) |
1879 xorps %xmm13,%xmm5 | 1798 xorps %xmm13,%xmm5 |
1880 movdqu %xmm3,16(%rsi) | 1799 movdqu %xmm3,16(%rsi) |
(...skipping 122 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
2003 .byte 102,15,56,220,209 | 1922 .byte 102,15,56,220,209 |
2004 decl %eax | 1923 decl %eax |
2005 movups (%rcx),%xmm1 | 1924 movups (%rcx),%xmm1 |
2006 leaq 16(%rcx),%rcx | 1925 leaq 16(%rcx),%rcx |
2007 jnz .Loop_enc1_10 | 1926 jnz .Loop_enc1_10 |
2008 .byte 102,15,56,221,209 | 1927 .byte 102,15,56,221,209 |
2009 xorps %xmm10,%xmm2 | 1928 xorps %xmm10,%xmm2 |
2010 movups %xmm2,-16(%rsi) | 1929 movups %xmm2,-16(%rsi) |
2011 | 1930 |
2012 .Lxts_enc_ret: | 1931 .Lxts_enc_ret: |
2013 xorps %xmm0,%xmm0 | |
2014 pxor %xmm1,%xmm1 | |
2015 pxor %xmm2,%xmm2 | |
2016 pxor %xmm3,%xmm3 | |
2017 pxor %xmm4,%xmm4 | |
2018 pxor %xmm5,%xmm5 | |
2019 pxor %xmm6,%xmm6 | |
2020 pxor %xmm7,%xmm7 | |
2021 movaps %xmm0,0(%rsp) | |
2022 pxor %xmm8,%xmm8 | |
2023 movaps %xmm0,16(%rsp) | |
2024 pxor %xmm9,%xmm9 | |
2025 movaps %xmm0,32(%rsp) | |
2026 pxor %xmm10,%xmm10 | |
2027 movaps %xmm0,48(%rsp) | |
2028 pxor %xmm11,%xmm11 | |
2029 movaps %xmm0,64(%rsp) | |
2030 pxor %xmm12,%xmm12 | |
2031 movaps %xmm0,80(%rsp) | |
2032 pxor %xmm13,%xmm13 | |
2033 movaps %xmm0,96(%rsp) | |
2034 pxor %xmm14,%xmm14 | |
2035 pxor %xmm15,%xmm15 | |
2036 leaq (%rbp),%rsp | 1932 leaq (%rbp),%rsp |
2037 popq %rbp | 1933 popq %rbp |
2038 .Lxts_enc_epilogue: | 1934 .Lxts_enc_epilogue: |
2039 .byte 0xf3,0xc3 | 1935 .byte 0xf3,0xc3 |
2040 .size aesni_xts_encrypt,.-aesni_xts_encrypt | 1936 .size aesni_xts_encrypt,.-aesni_xts_encrypt |
2041 .globl aesni_xts_decrypt | 1937 .globl aesni_xts_decrypt |
2042 .hidden aesni_xts_decrypt | 1938 .hidden aesni_xts_decrypt |
2043 .type aesni_xts_decrypt,@function | 1939 .type aesni_xts_decrypt,@function |
2044 .align 16 | 1940 .align 16 |
2045 aesni_xts_decrypt: | 1941 aesni_xts_decrypt: |
(...skipping 257 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
2303 movups %xmm7,-16(%rsi) | 2199 movups %xmm7,-16(%rsi) |
2304 subq $96,%rdx | 2200 subq $96,%rdx |
2305 jnc .Lxts_dec_grandloop | 2201 jnc .Lxts_dec_grandloop |
2306 | 2202 |
2307 movl $16+96,%eax | 2203 movl $16+96,%eax |
2308 subl %r10d,%eax | 2204 subl %r10d,%eax |
2309 movq %r11,%rcx | 2205 movq %r11,%rcx |
2310 shrl $4,%eax | 2206 shrl $4,%eax |
2311 | 2207 |
2312 .Lxts_dec_short: | 2208 .Lxts_dec_short: |
2313 | |
2314 movl %eax,%r10d | 2209 movl %eax,%r10d |
2315 pxor %xmm0,%xmm10 | 2210 pxor %xmm0,%xmm10 |
2316 pxor %xmm0,%xmm11 | 2211 pxor %xmm0,%xmm11 |
2317 addq $96,%rdx | 2212 addq $96,%rdx |
2318 jz .Lxts_dec_done | 2213 jz .Lxts_dec_done |
2319 | 2214 |
2320 pxor %xmm0,%xmm12 | 2215 pxor %xmm0,%xmm12 |
2321 cmpq $32,%rdx | 2216 cmpq $32,%rdx |
2322 jb .Lxts_dec_one | 2217 jb .Lxts_dec_one |
2323 pxor %xmm0,%xmm13 | 2218 pxor %xmm0,%xmm13 |
(...skipping 182 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
2506 .byte 102,15,56,222,209 | 2401 .byte 102,15,56,222,209 |
2507 decl %eax | 2402 decl %eax |
2508 movups (%rcx),%xmm1 | 2403 movups (%rcx),%xmm1 |
2509 leaq 16(%rcx),%rcx | 2404 leaq 16(%rcx),%rcx |
2510 jnz .Loop_dec1_14 | 2405 jnz .Loop_dec1_14 |
2511 .byte 102,15,56,223,209 | 2406 .byte 102,15,56,223,209 |
2512 xorps %xmm10,%xmm2 | 2407 xorps %xmm10,%xmm2 |
2513 movups %xmm2,(%rsi) | 2408 movups %xmm2,(%rsi) |
2514 | 2409 |
2515 .Lxts_dec_ret: | 2410 .Lxts_dec_ret: |
2516 xorps %xmm0,%xmm0 | |
2517 pxor %xmm1,%xmm1 | |
2518 pxor %xmm2,%xmm2 | |
2519 pxor %xmm3,%xmm3 | |
2520 pxor %xmm4,%xmm4 | |
2521 pxor %xmm5,%xmm5 | |
2522 pxor %xmm6,%xmm6 | |
2523 pxor %xmm7,%xmm7 | |
2524 movaps %xmm0,0(%rsp) | |
2525 pxor %xmm8,%xmm8 | |
2526 movaps %xmm0,16(%rsp) | |
2527 pxor %xmm9,%xmm9 | |
2528 movaps %xmm0,32(%rsp) | |
2529 pxor %xmm10,%xmm10 | |
2530 movaps %xmm0,48(%rsp) | |
2531 pxor %xmm11,%xmm11 | |
2532 movaps %xmm0,64(%rsp) | |
2533 pxor %xmm12,%xmm12 | |
2534 movaps %xmm0,80(%rsp) | |
2535 pxor %xmm13,%xmm13 | |
2536 movaps %xmm0,96(%rsp) | |
2537 pxor %xmm14,%xmm14 | |
2538 pxor %xmm15,%xmm15 | |
2539 leaq (%rbp),%rsp | 2411 leaq (%rbp),%rsp |
2540 popq %rbp | 2412 popq %rbp |
2541 .Lxts_dec_epilogue: | 2413 .Lxts_dec_epilogue: |
2542 .byte 0xf3,0xc3 | 2414 .byte 0xf3,0xc3 |
2543 .size aesni_xts_decrypt,.-aesni_xts_decrypt | 2415 .size aesni_xts_decrypt,.-aesni_xts_decrypt |
2544 .globl aesni_cbc_encrypt | 2416 .globl aesni_cbc_encrypt |
2545 .hidden aesni_cbc_encrypt | 2417 .hidden aesni_cbc_encrypt |
2546 .type aesni_cbc_encrypt,@function | 2418 .type aesni_cbc_encrypt,@function |
2547 .align 16 | 2419 .align 16 |
2548 aesni_cbc_encrypt: | 2420 aesni_cbc_encrypt: |
(...skipping 29 matching lines...) Expand all Loading... |
2578 jnz .Loop_enc1_15 | 2450 jnz .Loop_enc1_15 |
2579 .byte 102,15,56,221,209 | 2451 .byte 102,15,56,221,209 |
2580 movl %r10d,%eax | 2452 movl %r10d,%eax |
2581 movq %r11,%rcx | 2453 movq %r11,%rcx |
2582 movups %xmm2,0(%rsi) | 2454 movups %xmm2,0(%rsi) |
2583 leaq 16(%rsi),%rsi | 2455 leaq 16(%rsi),%rsi |
2584 subq $16,%rdx | 2456 subq $16,%rdx |
2585 jnc .Lcbc_enc_loop | 2457 jnc .Lcbc_enc_loop |
2586 addq $16,%rdx | 2458 addq $16,%rdx |
2587 jnz .Lcbc_enc_tail | 2459 jnz .Lcbc_enc_tail |
2588 pxor %xmm0,%xmm0 | |
2589 pxor %xmm1,%xmm1 | |
2590 movups %xmm2,(%r8) | 2460 movups %xmm2,(%r8) |
2591 pxor %xmm2,%xmm2 | |
2592 pxor %xmm3,%xmm3 | |
2593 jmp .Lcbc_ret | 2461 jmp .Lcbc_ret |
2594 | 2462 |
2595 .Lcbc_enc_tail: | 2463 .Lcbc_enc_tail: |
2596 movq %rdx,%rcx | 2464 movq %rdx,%rcx |
2597 xchgq %rdi,%rsi | 2465 xchgq %rdi,%rsi |
2598 .long 0x9066A4F3 | 2466 .long 0x9066A4F3 |
2599 movl $16,%ecx | 2467 movl $16,%ecx |
2600 subq %rdx,%rcx | 2468 subq %rdx,%rcx |
2601 xorl %eax,%eax | 2469 xorl %eax,%eax |
2602 .long 0x9066AAF3 | 2470 .long 0x9066AAF3 |
2603 leaq -16(%rdi),%rdi | 2471 leaq -16(%rdi),%rdi |
2604 movl %r10d,%eax | 2472 movl %r10d,%eax |
2605 movq %rdi,%rsi | 2473 movq %rdi,%rsi |
2606 movq %r11,%rcx | 2474 movq %r11,%rcx |
2607 xorq %rdx,%rdx | 2475 xorq %rdx,%rdx |
2608 jmp .Lcbc_enc_loop | 2476 jmp .Lcbc_enc_loop |
2609 | 2477 |
2610 .align 16 | 2478 .align 16 |
2611 .Lcbc_decrypt: | 2479 .Lcbc_decrypt: |
2612 cmpq $16,%rdx | |
2613 jne .Lcbc_decrypt_bulk | |
2614 | |
2615 | |
2616 | |
2617 movdqu (%rdi),%xmm2 | |
2618 movdqu (%r8),%xmm3 | |
2619 movdqa %xmm2,%xmm4 | |
2620 movups (%rcx),%xmm0 | |
2621 movups 16(%rcx),%xmm1 | |
2622 leaq 32(%rcx),%rcx | |
2623 xorps %xmm0,%xmm2 | |
2624 .Loop_dec1_16: | |
2625 .byte 102,15,56,222,209 | |
2626 decl %r10d | |
2627 movups (%rcx),%xmm1 | |
2628 leaq 16(%rcx),%rcx | |
2629 jnz .Loop_dec1_16 | |
2630 .byte 102,15,56,223,209 | |
2631 pxor %xmm0,%xmm0 | |
2632 pxor %xmm1,%xmm1 | |
2633 movdqu %xmm4,(%r8) | |
2634 xorps %xmm3,%xmm2 | |
2635 pxor %xmm3,%xmm3 | |
2636 movups %xmm2,(%rsi) | |
2637 pxor %xmm2,%xmm2 | |
2638 jmp .Lcbc_ret | |
2639 .align 16 | |
2640 .Lcbc_decrypt_bulk: | |
2641 leaq (%rsp),%rax | 2480 leaq (%rsp),%rax |
2642 pushq %rbp | 2481 pushq %rbp |
2643 subq $16,%rsp | 2482 subq $16,%rsp |
2644 andq $-16,%rsp | 2483 andq $-16,%rsp |
2645 leaq -8(%rax),%rbp | 2484 leaq -8(%rax),%rbp |
2646 movups (%r8),%xmm10 | 2485 movups (%r8),%xmm10 |
2647 movl %r10d,%eax | 2486 movl %r10d,%eax |
2648 cmpq $80,%rdx | 2487 cmpq $80,%rdx |
2649 jbe .Lcbc_dec_tail | 2488 jbe .Lcbc_dec_tail |
2650 | 2489 |
(...skipping 216 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
2867 movdqa %xmm1,%xmm7 | 2706 movdqa %xmm1,%xmm7 |
2868 movups %xmm8,96(%rsi) | 2707 movups %xmm8,96(%rsi) |
2869 leaq 112(%rsi),%rsi | 2708 leaq 112(%rsi),%rsi |
2870 | 2709 |
2871 subq $128,%rdx | 2710 subq $128,%rdx |
2872 ja .Lcbc_dec_loop8 | 2711 ja .Lcbc_dec_loop8 |
2873 | 2712 |
2874 movaps %xmm9,%xmm2 | 2713 movaps %xmm9,%xmm2 |
2875 leaq -112(%rcx),%rcx | 2714 leaq -112(%rcx),%rcx |
2876 addq $112,%rdx | 2715 addq $112,%rdx |
2877 » jle» .Lcbc_dec_clear_tail_collected | 2716 » jle» .Lcbc_dec_tail_collected |
2878 movups %xmm9,(%rsi) | 2717 movups %xmm9,(%rsi) |
2879 leaq 16(%rsi),%rsi | 2718 leaq 16(%rsi),%rsi |
2880 cmpq $80,%rdx | 2719 cmpq $80,%rdx |
2881 jbe .Lcbc_dec_tail | 2720 jbe .Lcbc_dec_tail |
2882 | 2721 |
2883 movaps %xmm11,%xmm2 | 2722 movaps %xmm11,%xmm2 |
2884 .Lcbc_dec_six_or_seven: | 2723 .Lcbc_dec_six_or_seven: |
2885 cmpq $96,%rdx | 2724 cmpq $96,%rdx |
2886 ja .Lcbc_dec_seven | 2725 ja .Lcbc_dec_seven |
2887 | 2726 |
2888 movaps %xmm7,%xmm8 | 2727 movaps %xmm7,%xmm8 |
2889 call _aesni_decrypt6 | 2728 call _aesni_decrypt6 |
2890 pxor %xmm10,%xmm2 | 2729 pxor %xmm10,%xmm2 |
2891 movaps %xmm8,%xmm10 | 2730 movaps %xmm8,%xmm10 |
2892 pxor %xmm11,%xmm3 | 2731 pxor %xmm11,%xmm3 |
2893 movdqu %xmm2,(%rsi) | 2732 movdqu %xmm2,(%rsi) |
2894 pxor %xmm12,%xmm4 | 2733 pxor %xmm12,%xmm4 |
2895 movdqu %xmm3,16(%rsi) | 2734 movdqu %xmm3,16(%rsi) |
2896 pxor %xmm3,%xmm3 | |
2897 pxor %xmm13,%xmm5 | 2735 pxor %xmm13,%xmm5 |
2898 movdqu %xmm4,32(%rsi) | 2736 movdqu %xmm4,32(%rsi) |
2899 pxor %xmm4,%xmm4 | |
2900 pxor %xmm14,%xmm6 | 2737 pxor %xmm14,%xmm6 |
2901 movdqu %xmm5,48(%rsi) | 2738 movdqu %xmm5,48(%rsi) |
2902 pxor %xmm5,%xmm5 | |
2903 pxor %xmm15,%xmm7 | 2739 pxor %xmm15,%xmm7 |
2904 movdqu %xmm6,64(%rsi) | 2740 movdqu %xmm6,64(%rsi) |
2905 pxor %xmm6,%xmm6 | |
2906 leaq 80(%rsi),%rsi | 2741 leaq 80(%rsi),%rsi |
2907 movdqa %xmm7,%xmm2 | 2742 movdqa %xmm7,%xmm2 |
2908 pxor %xmm7,%xmm7 | |
2909 jmp .Lcbc_dec_tail_collected | 2743 jmp .Lcbc_dec_tail_collected |
2910 | 2744 |
2911 .align 16 | 2745 .align 16 |
2912 .Lcbc_dec_seven: | 2746 .Lcbc_dec_seven: |
2913 movups 96(%rdi),%xmm8 | 2747 movups 96(%rdi),%xmm8 |
2914 xorps %xmm9,%xmm9 | 2748 xorps %xmm9,%xmm9 |
2915 call _aesni_decrypt8 | 2749 call _aesni_decrypt8 |
2916 movups 80(%rdi),%xmm9 | 2750 movups 80(%rdi),%xmm9 |
2917 pxor %xmm10,%xmm2 | 2751 pxor %xmm10,%xmm2 |
2918 movups 96(%rdi),%xmm10 | 2752 movups 96(%rdi),%xmm10 |
2919 pxor %xmm11,%xmm3 | 2753 pxor %xmm11,%xmm3 |
2920 movdqu %xmm2,(%rsi) | 2754 movdqu %xmm2,(%rsi) |
2921 pxor %xmm12,%xmm4 | 2755 pxor %xmm12,%xmm4 |
2922 movdqu %xmm3,16(%rsi) | 2756 movdqu %xmm3,16(%rsi) |
2923 pxor %xmm3,%xmm3 | |
2924 pxor %xmm13,%xmm5 | 2757 pxor %xmm13,%xmm5 |
2925 movdqu %xmm4,32(%rsi) | 2758 movdqu %xmm4,32(%rsi) |
2926 pxor %xmm4,%xmm4 | |
2927 pxor %xmm14,%xmm6 | 2759 pxor %xmm14,%xmm6 |
2928 movdqu %xmm5,48(%rsi) | 2760 movdqu %xmm5,48(%rsi) |
2929 pxor %xmm5,%xmm5 | |
2930 pxor %xmm15,%xmm7 | 2761 pxor %xmm15,%xmm7 |
2931 movdqu %xmm6,64(%rsi) | 2762 movdqu %xmm6,64(%rsi) |
2932 pxor %xmm6,%xmm6 | |
2933 pxor %xmm9,%xmm8 | 2763 pxor %xmm9,%xmm8 |
2934 movdqu %xmm7,80(%rsi) | 2764 movdqu %xmm7,80(%rsi) |
2935 pxor %xmm7,%xmm7 | |
2936 leaq 96(%rsi),%rsi | 2765 leaq 96(%rsi),%rsi |
2937 movdqa %xmm8,%xmm2 | 2766 movdqa %xmm8,%xmm2 |
2938 pxor %xmm8,%xmm8 | |
2939 pxor %xmm9,%xmm9 | |
2940 jmp .Lcbc_dec_tail_collected | 2767 jmp .Lcbc_dec_tail_collected |
2941 | 2768 |
2942 .align 16 | 2769 .align 16 |
2943 .Lcbc_dec_loop6: | 2770 .Lcbc_dec_loop6: |
2944 movups %xmm7,(%rsi) | 2771 movups %xmm7,(%rsi) |
2945 leaq 16(%rsi),%rsi | 2772 leaq 16(%rsi),%rsi |
2946 movdqu 0(%rdi),%xmm2 | 2773 movdqu 0(%rdi),%xmm2 |
2947 movdqu 16(%rdi),%xmm3 | 2774 movdqu 16(%rdi),%xmm3 |
2948 movdqa %xmm2,%xmm11 | 2775 movdqa %xmm2,%xmm11 |
2949 movdqu 32(%rdi),%xmm4 | 2776 movdqu 32(%rdi),%xmm4 |
(...skipping 23 matching lines...) Expand all Loading... |
2973 movdqu %xmm5,48(%rsi) | 2800 movdqu %xmm5,48(%rsi) |
2974 pxor %xmm15,%xmm7 | 2801 pxor %xmm15,%xmm7 |
2975 movl %r10d,%eax | 2802 movl %r10d,%eax |
2976 movdqu %xmm6,64(%rsi) | 2803 movdqu %xmm6,64(%rsi) |
2977 leaq 80(%rsi),%rsi | 2804 leaq 80(%rsi),%rsi |
2978 subq $96,%rdx | 2805 subq $96,%rdx |
2979 ja .Lcbc_dec_loop6 | 2806 ja .Lcbc_dec_loop6 |
2980 | 2807 |
2981 movdqa %xmm7,%xmm2 | 2808 movdqa %xmm7,%xmm2 |
2982 addq $80,%rdx | 2809 addq $80,%rdx |
2983 » jle» .Lcbc_dec_clear_tail_collected | 2810 » jle» .Lcbc_dec_tail_collected |
2984 movups %xmm7,(%rsi) | 2811 movups %xmm7,(%rsi) |
2985 leaq 16(%rsi),%rsi | 2812 leaq 16(%rsi),%rsi |
2986 | 2813 |
2987 .Lcbc_dec_tail: | 2814 .Lcbc_dec_tail: |
2988 movups (%rdi),%xmm2 | 2815 movups (%rdi),%xmm2 |
2989 subq $16,%rdx | 2816 subq $16,%rdx |
2990 jbe .Lcbc_dec_one | 2817 jbe .Lcbc_dec_one |
2991 | 2818 |
2992 movups 16(%rdi),%xmm3 | 2819 movups 16(%rdi),%xmm3 |
2993 movaps %xmm2,%xmm11 | 2820 movaps %xmm2,%xmm11 |
(...skipping 14 matching lines...) Expand all Loading... |
3008 movaps %xmm5,%xmm14 | 2835 movaps %xmm5,%xmm14 |
3009 movaps %xmm6,%xmm15 | 2836 movaps %xmm6,%xmm15 |
3010 xorps %xmm7,%xmm7 | 2837 xorps %xmm7,%xmm7 |
3011 call _aesni_decrypt6 | 2838 call _aesni_decrypt6 |
3012 pxor %xmm10,%xmm2 | 2839 pxor %xmm10,%xmm2 |
3013 movaps %xmm15,%xmm10 | 2840 movaps %xmm15,%xmm10 |
3014 pxor %xmm11,%xmm3 | 2841 pxor %xmm11,%xmm3 |
3015 movdqu %xmm2,(%rsi) | 2842 movdqu %xmm2,(%rsi) |
3016 pxor %xmm12,%xmm4 | 2843 pxor %xmm12,%xmm4 |
3017 movdqu %xmm3,16(%rsi) | 2844 movdqu %xmm3,16(%rsi) |
3018 pxor %xmm3,%xmm3 | |
3019 pxor %xmm13,%xmm5 | 2845 pxor %xmm13,%xmm5 |
3020 movdqu %xmm4,32(%rsi) | 2846 movdqu %xmm4,32(%rsi) |
3021 pxor %xmm4,%xmm4 | |
3022 pxor %xmm14,%xmm6 | 2847 pxor %xmm14,%xmm6 |
3023 movdqu %xmm5,48(%rsi) | 2848 movdqu %xmm5,48(%rsi) |
3024 pxor %xmm5,%xmm5 | |
3025 leaq 64(%rsi),%rsi | 2849 leaq 64(%rsi),%rsi |
3026 movdqa %xmm6,%xmm2 | 2850 movdqa %xmm6,%xmm2 |
3027 pxor %xmm6,%xmm6 | |
3028 pxor %xmm7,%xmm7 | |
3029 subq $16,%rdx | 2851 subq $16,%rdx |
3030 jmp .Lcbc_dec_tail_collected | 2852 jmp .Lcbc_dec_tail_collected |
3031 | 2853 |
3032 .align 16 | 2854 .align 16 |
3033 .Lcbc_dec_one: | 2855 .Lcbc_dec_one: |
3034 movaps %xmm2,%xmm11 | 2856 movaps %xmm2,%xmm11 |
3035 movups (%rcx),%xmm0 | 2857 movups (%rcx),%xmm0 |
3036 movups 16(%rcx),%xmm1 | 2858 movups 16(%rcx),%xmm1 |
3037 leaq 32(%rcx),%rcx | 2859 leaq 32(%rcx),%rcx |
3038 xorps %xmm0,%xmm2 | 2860 xorps %xmm0,%xmm2 |
3039 .Loop_dec1_17: | 2861 .Loop_dec1_16: |
3040 .byte 102,15,56,222,209 | 2862 .byte 102,15,56,222,209 |
3041 decl %eax | 2863 decl %eax |
3042 movups (%rcx),%xmm1 | 2864 movups (%rcx),%xmm1 |
3043 leaq 16(%rcx),%rcx | 2865 leaq 16(%rcx),%rcx |
3044 » jnz» .Loop_dec1_17 | 2866 » jnz» .Loop_dec1_16 |
3045 .byte 102,15,56,223,209 | 2867 .byte 102,15,56,223,209 |
3046 xorps %xmm10,%xmm2 | 2868 xorps %xmm10,%xmm2 |
3047 movaps %xmm11,%xmm10 | 2869 movaps %xmm11,%xmm10 |
3048 jmp .Lcbc_dec_tail_collected | 2870 jmp .Lcbc_dec_tail_collected |
3049 .align 16 | 2871 .align 16 |
3050 .Lcbc_dec_two: | 2872 .Lcbc_dec_two: |
3051 movaps %xmm3,%xmm12 | 2873 movaps %xmm3,%xmm12 |
3052 call _aesni_decrypt2 | 2874 call _aesni_decrypt2 |
3053 pxor %xmm10,%xmm2 | 2875 pxor %xmm10,%xmm2 |
3054 movaps %xmm12,%xmm10 | 2876 movaps %xmm12,%xmm10 |
3055 pxor %xmm11,%xmm3 | 2877 pxor %xmm11,%xmm3 |
3056 movdqu %xmm2,(%rsi) | 2878 movdqu %xmm2,(%rsi) |
3057 movdqa %xmm3,%xmm2 | 2879 movdqa %xmm3,%xmm2 |
3058 pxor %xmm3,%xmm3 | |
3059 leaq 16(%rsi),%rsi | 2880 leaq 16(%rsi),%rsi |
3060 jmp .Lcbc_dec_tail_collected | 2881 jmp .Lcbc_dec_tail_collected |
3061 .align 16 | 2882 .align 16 |
3062 .Lcbc_dec_three: | 2883 .Lcbc_dec_three: |
3063 movaps %xmm4,%xmm13 | 2884 movaps %xmm4,%xmm13 |
3064 call _aesni_decrypt3 | 2885 call _aesni_decrypt3 |
3065 pxor %xmm10,%xmm2 | 2886 pxor %xmm10,%xmm2 |
3066 movaps %xmm13,%xmm10 | 2887 movaps %xmm13,%xmm10 |
3067 pxor %xmm11,%xmm3 | 2888 pxor %xmm11,%xmm3 |
3068 movdqu %xmm2,(%rsi) | 2889 movdqu %xmm2,(%rsi) |
3069 pxor %xmm12,%xmm4 | 2890 pxor %xmm12,%xmm4 |
3070 movdqu %xmm3,16(%rsi) | 2891 movdqu %xmm3,16(%rsi) |
3071 pxor %xmm3,%xmm3 | |
3072 movdqa %xmm4,%xmm2 | 2892 movdqa %xmm4,%xmm2 |
3073 pxor %xmm4,%xmm4 | |
3074 leaq 32(%rsi),%rsi | 2893 leaq 32(%rsi),%rsi |
3075 jmp .Lcbc_dec_tail_collected | 2894 jmp .Lcbc_dec_tail_collected |
3076 .align 16 | 2895 .align 16 |
3077 .Lcbc_dec_four: | 2896 .Lcbc_dec_four: |
3078 movaps %xmm5,%xmm14 | 2897 movaps %xmm5,%xmm14 |
3079 call _aesni_decrypt4 | 2898 call _aesni_decrypt4 |
3080 pxor %xmm10,%xmm2 | 2899 pxor %xmm10,%xmm2 |
3081 movaps %xmm14,%xmm10 | 2900 movaps %xmm14,%xmm10 |
3082 pxor %xmm11,%xmm3 | 2901 pxor %xmm11,%xmm3 |
3083 movdqu %xmm2,(%rsi) | 2902 movdqu %xmm2,(%rsi) |
3084 pxor %xmm12,%xmm4 | 2903 pxor %xmm12,%xmm4 |
3085 movdqu %xmm3,16(%rsi) | 2904 movdqu %xmm3,16(%rsi) |
3086 pxor %xmm3,%xmm3 | |
3087 pxor %xmm13,%xmm5 | 2905 pxor %xmm13,%xmm5 |
3088 movdqu %xmm4,32(%rsi) | 2906 movdqu %xmm4,32(%rsi) |
3089 pxor %xmm4,%xmm4 | |
3090 movdqa %xmm5,%xmm2 | 2907 movdqa %xmm5,%xmm2 |
3091 pxor %xmm5,%xmm5 | |
3092 leaq 48(%rsi),%rsi | 2908 leaq 48(%rsi),%rsi |
3093 jmp .Lcbc_dec_tail_collected | 2909 jmp .Lcbc_dec_tail_collected |
3094 | 2910 |
3095 .align 16 | 2911 .align 16 |
3096 .Lcbc_dec_clear_tail_collected: | |
3097 pxor %xmm3,%xmm3 | |
3098 pxor %xmm4,%xmm4 | |
3099 pxor %xmm5,%xmm5 | |
3100 pxor %xmm6,%xmm6 | |
3101 pxor %xmm7,%xmm7 | |
3102 pxor %xmm8,%xmm8 | |
3103 pxor %xmm9,%xmm9 | |
3104 .Lcbc_dec_tail_collected: | 2912 .Lcbc_dec_tail_collected: |
3105 movups %xmm10,(%r8) | 2913 movups %xmm10,(%r8) |
3106 andq $15,%rdx | 2914 andq $15,%rdx |
3107 jnz .Lcbc_dec_tail_partial | 2915 jnz .Lcbc_dec_tail_partial |
3108 movups %xmm2,(%rsi) | 2916 movups %xmm2,(%rsi) |
3109 pxor %xmm2,%xmm2 | |
3110 jmp .Lcbc_dec_ret | 2917 jmp .Lcbc_dec_ret |
3111 .align 16 | 2918 .align 16 |
3112 .Lcbc_dec_tail_partial: | 2919 .Lcbc_dec_tail_partial: |
3113 movaps %xmm2,(%rsp) | 2920 movaps %xmm2,(%rsp) |
3114 pxor %xmm2,%xmm2 | |
3115 movq $16,%rcx | 2921 movq $16,%rcx |
3116 movq %rsi,%rdi | 2922 movq %rsi,%rdi |
3117 subq %rdx,%rcx | 2923 subq %rdx,%rcx |
3118 leaq (%rsp),%rsi | 2924 leaq (%rsp),%rsi |
3119 .long 0x9066A4F3 | 2925 .long 0x9066A4F3 |
3120 movdqa %xmm2,(%rsp) | |
3121 | 2926 |
3122 .Lcbc_dec_ret: | 2927 .Lcbc_dec_ret: |
3123 xorps %xmm0,%xmm0 | |
3124 pxor %xmm1,%xmm1 | |
3125 leaq (%rbp),%rsp | 2928 leaq (%rbp),%rsp |
3126 popq %rbp | 2929 popq %rbp |
3127 .Lcbc_ret: | 2930 .Lcbc_ret: |
3128 .byte 0xf3,0xc3 | 2931 .byte 0xf3,0xc3 |
3129 .size aesni_cbc_encrypt,.-aesni_cbc_encrypt | 2932 .size aesni_cbc_encrypt,.-aesni_cbc_encrypt |
3130 .globl aesni_set_decrypt_key | 2933 .globl aesni_set_decrypt_key |
3131 .hidden aesni_set_decrypt_key | 2934 .hidden aesni_set_decrypt_key |
3132 .type aesni_set_decrypt_key,@function | 2935 .type aesni_set_decrypt_key,@function |
3133 .align 16 | 2936 .align 16 |
3134 aesni_set_decrypt_key: | 2937 aesni_set_decrypt_key: |
(...skipping 18 matching lines...) Expand all Loading... |
3153 .byte 102,15,56,219,201 | 2956 .byte 102,15,56,219,201 |
3154 leaq 16(%rdx),%rdx | 2957 leaq 16(%rdx),%rdx |
3155 leaq -16(%rdi),%rdi | 2958 leaq -16(%rdi),%rdi |
3156 movups %xmm0,16(%rdi) | 2959 movups %xmm0,16(%rdi) |
3157 movups %xmm1,-16(%rdx) | 2960 movups %xmm1,-16(%rdx) |
3158 cmpq %rdx,%rdi | 2961 cmpq %rdx,%rdi |
3159 ja .Ldec_key_inverse | 2962 ja .Ldec_key_inverse |
3160 | 2963 |
3161 movups (%rdx),%xmm0 | 2964 movups (%rdx),%xmm0 |
3162 .byte 102,15,56,219,192 | 2965 .byte 102,15,56,219,192 |
3163 pxor %xmm1,%xmm1 | |
3164 movups %xmm0,(%rdi) | 2966 movups %xmm0,(%rdi) |
3165 pxor %xmm0,%xmm0 | |
3166 .Ldec_key_ret: | 2967 .Ldec_key_ret: |
3167 addq $8,%rsp | 2968 addq $8,%rsp |
3168 .byte 0xf3,0xc3 | 2969 .byte 0xf3,0xc3 |
3169 .LSEH_end_set_decrypt_key: | 2970 .LSEH_end_set_decrypt_key: |
3170 .size aesni_set_decrypt_key,.-aesni_set_decrypt_key | 2971 .size aesni_set_decrypt_key,.-aesni_set_decrypt_key |
3171 .globl aesni_set_encrypt_key | 2972 .globl aesni_set_encrypt_key |
3172 .hidden aesni_set_encrypt_key | 2973 .hidden aesni_set_encrypt_key |
3173 .type aesni_set_encrypt_key,@function | 2974 .type aesni_set_encrypt_key,@function |
3174 .align 16 | 2975 .align 16 |
3175 aesni_set_encrypt_key: | 2976 aesni_set_encrypt_key: |
3176 __aesni_set_encrypt_key: | 2977 __aesni_set_encrypt_key: |
3177 .byte 0x48,0x83,0xEC,0x08 | 2978 .byte 0x48,0x83,0xEC,0x08 |
3178 movq $-1,%rax | 2979 movq $-1,%rax |
3179 testq %rdi,%rdi | 2980 testq %rdi,%rdi |
3180 jz .Lenc_key_ret | 2981 jz .Lenc_key_ret |
3181 testq %rdx,%rdx | 2982 testq %rdx,%rdx |
3182 jz .Lenc_key_ret | 2983 jz .Lenc_key_ret |
3183 | 2984 |
3184 movl $268437504,%r10d | |
3185 movups (%rdi),%xmm0 | 2985 movups (%rdi),%xmm0 |
3186 xorps %xmm4,%xmm4 | 2986 xorps %xmm4,%xmm4 |
3187 andl OPENSSL_ia32cap_P+4(%rip),%r10d | |
3188 leaq 16(%rdx),%rax | 2987 leaq 16(%rdx),%rax |
3189 cmpl $256,%esi | 2988 cmpl $256,%esi |
3190 je .L14rounds | 2989 je .L14rounds |
3191 cmpl $192,%esi | 2990 cmpl $192,%esi |
3192 je .L12rounds | 2991 je .L12rounds |
3193 cmpl $128,%esi | 2992 cmpl $128,%esi |
3194 jne .Lbad_keybits | 2993 jne .Lbad_keybits |
3195 | 2994 |
3196 .L10rounds: | 2995 .L10rounds: |
3197 movl $9,%esi | 2996 movl $9,%esi |
3198 cmpl $268435456,%r10d | |
3199 je .L10rounds_alt | |
3200 | |
3201 movups %xmm0,(%rdx) | 2997 movups %xmm0,(%rdx) |
3202 .byte 102,15,58,223,200,1 | 2998 .byte 102,15,58,223,200,1 |
3203 call .Lkey_expansion_128_cold | 2999 call .Lkey_expansion_128_cold |
3204 .byte 102,15,58,223,200,2 | 3000 .byte 102,15,58,223,200,2 |
3205 call .Lkey_expansion_128 | 3001 call .Lkey_expansion_128 |
3206 .byte 102,15,58,223,200,4 | 3002 .byte 102,15,58,223,200,4 |
3207 call .Lkey_expansion_128 | 3003 call .Lkey_expansion_128 |
3208 .byte 102,15,58,223,200,8 | 3004 .byte 102,15,58,223,200,8 |
3209 call .Lkey_expansion_128 | 3005 call .Lkey_expansion_128 |
3210 .byte 102,15,58,223,200,16 | 3006 .byte 102,15,58,223,200,16 |
3211 call .Lkey_expansion_128 | 3007 call .Lkey_expansion_128 |
3212 .byte 102,15,58,223,200,32 | 3008 .byte 102,15,58,223,200,32 |
3213 call .Lkey_expansion_128 | 3009 call .Lkey_expansion_128 |
3214 .byte 102,15,58,223,200,64 | 3010 .byte 102,15,58,223,200,64 |
3215 call .Lkey_expansion_128 | 3011 call .Lkey_expansion_128 |
3216 .byte 102,15,58,223,200,128 | 3012 .byte 102,15,58,223,200,128 |
3217 call .Lkey_expansion_128 | 3013 call .Lkey_expansion_128 |
3218 .byte 102,15,58,223,200,27 | 3014 .byte 102,15,58,223,200,27 |
3219 call .Lkey_expansion_128 | 3015 call .Lkey_expansion_128 |
3220 .byte 102,15,58,223,200,54 | 3016 .byte 102,15,58,223,200,54 |
3221 call .Lkey_expansion_128 | 3017 call .Lkey_expansion_128 |
3222 movups %xmm0,(%rax) | 3018 movups %xmm0,(%rax) |
3223 movl %esi,80(%rax) | 3019 movl %esi,80(%rax) |
3224 xorl %eax,%eax | 3020 xorl %eax,%eax |
3225 jmp .Lenc_key_ret | 3021 jmp .Lenc_key_ret |
3226 | 3022 |
3227 .align 16 | 3023 .align 16 |
3228 .L10rounds_alt: | |
3229 movdqa .Lkey_rotate(%rip),%xmm5 | |
3230 movl $8,%r10d | |
3231 movdqa .Lkey_rcon1(%rip),%xmm4 | |
3232 movdqa %xmm0,%xmm2 | |
3233 movdqu %xmm0,(%rdx) | |
3234 jmp .Loop_key128 | |
3235 | |
3236 .align 16 | |
3237 .Loop_key128: | |
3238 .byte 102,15,56,0,197 | |
3239 .byte 102,15,56,221,196 | |
3240 pslld $1,%xmm4 | |
3241 leaq 16(%rax),%rax | |
3242 | |
3243 movdqa %xmm2,%xmm3 | |
3244 pslldq $4,%xmm2 | |
3245 pxor %xmm2,%xmm3 | |
3246 pslldq $4,%xmm2 | |
3247 pxor %xmm2,%xmm3 | |
3248 pslldq $4,%xmm2 | |
3249 pxor %xmm3,%xmm2 | |
3250 | |
3251 pxor %xmm2,%xmm0 | |
3252 movdqu %xmm0,-16(%rax) | |
3253 movdqa %xmm0,%xmm2 | |
3254 | |
3255 decl %r10d | |
3256 jnz .Loop_key128 | |
3257 | |
3258 movdqa .Lkey_rcon1b(%rip),%xmm4 | |
3259 | |
3260 .byte 102,15,56,0,197 | |
3261 .byte 102,15,56,221,196 | |
3262 pslld $1,%xmm4 | |
3263 | |
3264 movdqa %xmm2,%xmm3 | |
3265 pslldq $4,%xmm2 | |
3266 pxor %xmm2,%xmm3 | |
3267 pslldq $4,%xmm2 | |
3268 pxor %xmm2,%xmm3 | |
3269 pslldq $4,%xmm2 | |
3270 pxor %xmm3,%xmm2 | |
3271 | |
3272 pxor %xmm2,%xmm0 | |
3273 movdqu %xmm0,(%rax) | |
3274 | |
3275 movdqa %xmm0,%xmm2 | |
3276 .byte 102,15,56,0,197 | |
3277 .byte 102,15,56,221,196 | |
3278 | |
3279 movdqa %xmm2,%xmm3 | |
3280 pslldq $4,%xmm2 | |
3281 pxor %xmm2,%xmm3 | |
3282 pslldq $4,%xmm2 | |
3283 pxor %xmm2,%xmm3 | |
3284 pslldq $4,%xmm2 | |
3285 pxor %xmm3,%xmm2 | |
3286 | |
3287 pxor %xmm2,%xmm0 | |
3288 movdqu %xmm0,16(%rax) | |
3289 | |
3290 movl %esi,96(%rax) | |
3291 xorl %eax,%eax | |
3292 jmp .Lenc_key_ret | |
3293 | |
3294 .align 16 | |
3295 .L12rounds: | 3024 .L12rounds: |
3296 movq 16(%rdi),%xmm2 | 3025 movq 16(%rdi),%xmm2 |
3297 movl $11,%esi | 3026 movl $11,%esi |
3298 cmpl $268435456,%r10d | |
3299 je .L12rounds_alt | |
3300 | |
3301 movups %xmm0,(%rdx) | 3027 movups %xmm0,(%rdx) |
3302 .byte 102,15,58,223,202,1 | 3028 .byte 102,15,58,223,202,1 |
3303 call .Lkey_expansion_192a_cold | 3029 call .Lkey_expansion_192a_cold |
3304 .byte 102,15,58,223,202,2 | 3030 .byte 102,15,58,223,202,2 |
3305 call .Lkey_expansion_192b | 3031 call .Lkey_expansion_192b |
3306 .byte 102,15,58,223,202,4 | 3032 .byte 102,15,58,223,202,4 |
3307 call .Lkey_expansion_192a | 3033 call .Lkey_expansion_192a |
3308 .byte 102,15,58,223,202,8 | 3034 .byte 102,15,58,223,202,8 |
3309 call .Lkey_expansion_192b | 3035 call .Lkey_expansion_192b |
3310 .byte 102,15,58,223,202,16 | 3036 .byte 102,15,58,223,202,16 |
3311 call .Lkey_expansion_192a | 3037 call .Lkey_expansion_192a |
3312 .byte 102,15,58,223,202,32 | 3038 .byte 102,15,58,223,202,32 |
3313 call .Lkey_expansion_192b | 3039 call .Lkey_expansion_192b |
3314 .byte 102,15,58,223,202,64 | 3040 .byte 102,15,58,223,202,64 |
3315 call .Lkey_expansion_192a | 3041 call .Lkey_expansion_192a |
3316 .byte 102,15,58,223,202,128 | 3042 .byte 102,15,58,223,202,128 |
3317 call .Lkey_expansion_192b | 3043 call .Lkey_expansion_192b |
3318 movups %xmm0,(%rax) | 3044 movups %xmm0,(%rax) |
3319 movl %esi,48(%rax) | 3045 movl %esi,48(%rax) |
3320 xorq %rax,%rax | 3046 xorq %rax,%rax |
3321 jmp .Lenc_key_ret | 3047 jmp .Lenc_key_ret |
3322 | 3048 |
3323 .align 16 | 3049 .align 16 |
3324 .L12rounds_alt: | |
3325 movdqa .Lkey_rotate192(%rip),%xmm5 | |
3326 movdqa .Lkey_rcon1(%rip),%xmm4 | |
3327 movl $8,%r10d | |
3328 movdqu %xmm0,(%rdx) | |
3329 jmp .Loop_key192 | |
3330 | |
3331 .align 16 | |
3332 .Loop_key192: | |
3333 movq %xmm2,0(%rax) | |
3334 movdqa %xmm2,%xmm1 | |
3335 .byte 102,15,56,0,213 | |
3336 .byte 102,15,56,221,212 | |
3337 pslld $1,%xmm4 | |
3338 leaq 24(%rax),%rax | |
3339 | |
3340 movdqa %xmm0,%xmm3 | |
3341 pslldq $4,%xmm0 | |
3342 pxor %xmm0,%xmm3 | |
3343 pslldq $4,%xmm0 | |
3344 pxor %xmm0,%xmm3 | |
3345 pslldq $4,%xmm0 | |
3346 pxor %xmm3,%xmm0 | |
3347 | |
3348 pshufd $255,%xmm0,%xmm3 | |
3349 pxor %xmm1,%xmm3 | |
3350 pslldq $4,%xmm1 | |
3351 pxor %xmm1,%xmm3 | |
3352 | |
3353 pxor %xmm2,%xmm0 | |
3354 pxor %xmm3,%xmm2 | |
3355 movdqu %xmm0,-16(%rax) | |
3356 | |
3357 decl %r10d | |
3358 jnz .Loop_key192 | |
3359 | |
3360 movl %esi,32(%rax) | |
3361 xorl %eax,%eax | |
3362 jmp .Lenc_key_ret | |
3363 | |
3364 .align 16 | |
3365 .L14rounds: | 3050 .L14rounds: |
3366 movups 16(%rdi),%xmm2 | 3051 movups 16(%rdi),%xmm2 |
3367 movl $13,%esi | 3052 movl $13,%esi |
3368 leaq 16(%rax),%rax | 3053 leaq 16(%rax),%rax |
3369 cmpl $268435456,%r10d | |
3370 je .L14rounds_alt | |
3371 | |
3372 movups %xmm0,(%rdx) | 3054 movups %xmm0,(%rdx) |
3373 movups %xmm2,16(%rdx) | 3055 movups %xmm2,16(%rdx) |
3374 .byte 102,15,58,223,202,1 | 3056 .byte 102,15,58,223,202,1 |
3375 call .Lkey_expansion_256a_cold | 3057 call .Lkey_expansion_256a_cold |
3376 .byte 102,15,58,223,200,1 | 3058 .byte 102,15,58,223,200,1 |
3377 call .Lkey_expansion_256b | 3059 call .Lkey_expansion_256b |
3378 .byte 102,15,58,223,202,2 | 3060 .byte 102,15,58,223,202,2 |
3379 call .Lkey_expansion_256a | 3061 call .Lkey_expansion_256a |
3380 .byte 102,15,58,223,200,2 | 3062 .byte 102,15,58,223,200,2 |
3381 call .Lkey_expansion_256b | 3063 call .Lkey_expansion_256b |
(...skipping 14 matching lines...) Expand all Loading... |
3396 .byte 102,15,58,223,200,32 | 3078 .byte 102,15,58,223,200,32 |
3397 call .Lkey_expansion_256b | 3079 call .Lkey_expansion_256b |
3398 .byte 102,15,58,223,202,64 | 3080 .byte 102,15,58,223,202,64 |
3399 call .Lkey_expansion_256a | 3081 call .Lkey_expansion_256a |
3400 movups %xmm0,(%rax) | 3082 movups %xmm0,(%rax) |
3401 movl %esi,16(%rax) | 3083 movl %esi,16(%rax) |
3402 xorq %rax,%rax | 3084 xorq %rax,%rax |
3403 jmp .Lenc_key_ret | 3085 jmp .Lenc_key_ret |
3404 | 3086 |
3405 .align 16 | 3087 .align 16 |
3406 .L14rounds_alt: | |
3407 movdqa .Lkey_rotate(%rip),%xmm5 | |
3408 movdqa .Lkey_rcon1(%rip),%xmm4 | |
3409 movl $7,%r10d | |
3410 movdqu %xmm0,0(%rdx) | |
3411 movdqa %xmm2,%xmm1 | |
3412 movdqu %xmm2,16(%rdx) | |
3413 jmp .Loop_key256 | |
3414 | |
3415 .align 16 | |
3416 .Loop_key256: | |
3417 .byte 102,15,56,0,213 | |
3418 .byte 102,15,56,221,212 | |
3419 | |
3420 movdqa %xmm0,%xmm3 | |
3421 pslldq $4,%xmm0 | |
3422 pxor %xmm0,%xmm3 | |
3423 pslldq $4,%xmm0 | |
3424 pxor %xmm0,%xmm3 | |
3425 pslldq $4,%xmm0 | |
3426 pxor %xmm3,%xmm0 | |
3427 pslld $1,%xmm4 | |
3428 | |
3429 pxor %xmm2,%xmm0 | |
3430 movdqu %xmm0,(%rax) | |
3431 | |
3432 decl %r10d | |
3433 jz .Ldone_key256 | |
3434 | |
3435 pshufd $255,%xmm0,%xmm2 | |
3436 pxor %xmm3,%xmm3 | |
3437 .byte 102,15,56,221,211 | |
3438 | |
3439 movdqa %xmm1,%xmm3 | |
3440 pslldq $4,%xmm1 | |
3441 pxor %xmm1,%xmm3 | |
3442 pslldq $4,%xmm1 | |
3443 pxor %xmm1,%xmm3 | |
3444 pslldq $4,%xmm1 | |
3445 pxor %xmm3,%xmm1 | |
3446 | |
3447 pxor %xmm1,%xmm2 | |
3448 movdqu %xmm2,16(%rax) | |
3449 leaq 32(%rax),%rax | |
3450 movdqa %xmm2,%xmm1 | |
3451 | |
3452 jmp .Loop_key256 | |
3453 | |
3454 .Ldone_key256: | |
3455 movl %esi,16(%rax) | |
3456 xorl %eax,%eax | |
3457 jmp .Lenc_key_ret | |
3458 | |
3459 .align 16 | |
3460 .Lbad_keybits: | 3088 .Lbad_keybits: |
3461 movq $-2,%rax | 3089 movq $-2,%rax |
3462 .Lenc_key_ret: | 3090 .Lenc_key_ret: |
3463 pxor %xmm0,%xmm0 | |
3464 pxor %xmm1,%xmm1 | |
3465 pxor %xmm2,%xmm2 | |
3466 pxor %xmm3,%xmm3 | |
3467 pxor %xmm4,%xmm4 | |
3468 pxor %xmm5,%xmm5 | |
3469 addq $8,%rsp | 3091 addq $8,%rsp |
3470 .byte 0xf3,0xc3 | 3092 .byte 0xf3,0xc3 |
3471 .LSEH_end_set_encrypt_key: | 3093 .LSEH_end_set_encrypt_key: |
3472 | 3094 |
3473 .align 16 | 3095 .align 16 |
3474 .Lkey_expansion_128: | 3096 .Lkey_expansion_128: |
3475 movups %xmm0,(%rax) | 3097 movups %xmm0,(%rax) |
3476 leaq 16(%rax),%rax | 3098 leaq 16(%rax),%rax |
3477 .Lkey_expansion_128_cold: | 3099 .Lkey_expansion_128_cold: |
3478 shufps $16,%xmm0,%xmm4 | 3100 shufps $16,%xmm0,%xmm4 |
(...skipping 65 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
3544 .Lbswap_mask: | 3166 .Lbswap_mask: |
3545 .byte 15,14,13,12,11,10,9,8,7,6,5,4,3,2,1,0 | 3167 .byte 15,14,13,12,11,10,9,8,7,6,5,4,3,2,1,0 |
3546 .Lincrement32: | 3168 .Lincrement32: |
3547 .long 6,6,6,0 | 3169 .long 6,6,6,0 |
3548 .Lincrement64: | 3170 .Lincrement64: |
3549 .long 1,0,0,0 | 3171 .long 1,0,0,0 |
3550 .Lxts_magic: | 3172 .Lxts_magic: |
3551 .long 0x87,0,1,0 | 3173 .long 0x87,0,1,0 |
3552 .Lincrement1: | 3174 .Lincrement1: |
3553 .byte 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1 | 3175 .byte 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1 |
3554 .Lkey_rotate: | |
3555 .long 0x0c0f0e0d,0x0c0f0e0d,0x0c0f0e0d,0x0c0f0e0d | |
3556 .Lkey_rotate192: | |
3557 .long 0x04070605,0x04070605,0x04070605,0x04070605 | |
3558 .Lkey_rcon1: | |
3559 .long 1,1,1,1 | |
3560 .Lkey_rcon1b: | |
3561 .long 0x1b,0x1b,0x1b,0x1b | |
3562 | 3176 |
3563 .byte 65,69,83,32,102,111,114,32,73,110,116,101,108,32,65,69,83,45,78,73,44,32
,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101
,110,115,115,108,46,111,114,103,62,0 | 3177 .byte 65,69,83,32,102,111,114,32,73,110,116,101,108,32,65,69,83,45,78,73,44,32
,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101
,110,115,115,108,46,111,114,103,62,0 |
3564 .align 64 | 3178 .align 64 |
3565 #endif | 3179 #endif |
OLD | NEW |