OLD | NEW |
1 #if defined(__x86_64__) | 1 #if defined(__x86_64__) |
2 .text | 2 .text |
3 .extern OPENSSL_ia32cap_P | 3 .extern OPENSSL_ia32cap_P |
4 .hidden OPENSSL_ia32cap_P | 4 .hidden OPENSSL_ia32cap_P |
5 .globl aesni_encrypt | 5 .globl aesni_encrypt |
6 .hidden aesni_encrypt | 6 .hidden aesni_encrypt |
7 .type aesni_encrypt,@function | 7 .type aesni_encrypt,@function |
8 .align 16 | 8 .align 16 |
9 aesni_encrypt: | 9 aesni_encrypt: |
10 movups (%rdi),%xmm2 | 10 movups (%rdi),%xmm2 |
(...skipping 490 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
501 andq $-16,%rdx | 501 andq $-16,%rdx |
502 jz .Lecb_ret | 502 jz .Lecb_ret |
503 | 503 |
504 movl 240(%rcx),%eax | 504 movl 240(%rcx),%eax |
505 movups (%rcx),%xmm0 | 505 movups (%rcx),%xmm0 |
506 movq %rcx,%r11 | 506 movq %rcx,%r11 |
507 movl %eax,%r10d | 507 movl %eax,%r10d |
508 testl %r8d,%r8d | 508 testl %r8d,%r8d |
509 jz .Lecb_decrypt | 509 jz .Lecb_decrypt |
510 | 510 |
511 » cmpq» $128,%rdx | 511 » cmpq» $0x80,%rdx |
512 jb .Lecb_enc_tail | 512 jb .Lecb_enc_tail |
513 | 513 |
514 movdqu (%rdi),%xmm2 | 514 movdqu (%rdi),%xmm2 |
515 movdqu 16(%rdi),%xmm3 | 515 movdqu 16(%rdi),%xmm3 |
516 movdqu 32(%rdi),%xmm4 | 516 movdqu 32(%rdi),%xmm4 |
517 movdqu 48(%rdi),%xmm5 | 517 movdqu 48(%rdi),%xmm5 |
518 movdqu 64(%rdi),%xmm6 | 518 movdqu 64(%rdi),%xmm6 |
519 movdqu 80(%rdi),%xmm7 | 519 movdqu 80(%rdi),%xmm7 |
520 movdqu 96(%rdi),%xmm8 | 520 movdqu 96(%rdi),%xmm8 |
521 movdqu 112(%rdi),%xmm9 | 521 movdqu 112(%rdi),%xmm9 |
522 leaq 128(%rdi),%rdi | 522 leaq 128(%rdi),%rdi |
523 » subq» $128,%rdx | 523 » subq» $0x80,%rdx |
524 jmp .Lecb_enc_loop8_enter | 524 jmp .Lecb_enc_loop8_enter |
525 .align 16 | 525 .align 16 |
526 .Lecb_enc_loop8: | 526 .Lecb_enc_loop8: |
527 movups %xmm2,(%rsi) | 527 movups %xmm2,(%rsi) |
528 movq %r11,%rcx | 528 movq %r11,%rcx |
529 movdqu (%rdi),%xmm2 | 529 movdqu (%rdi),%xmm2 |
530 movl %r10d,%eax | 530 movl %r10d,%eax |
531 movups %xmm3,16(%rsi) | 531 movups %xmm3,16(%rsi) |
532 movdqu 16(%rdi),%xmm3 | 532 movdqu 16(%rdi),%xmm3 |
533 movups %xmm4,32(%rsi) | 533 movups %xmm4,32(%rsi) |
534 movdqu 32(%rdi),%xmm4 | 534 movdqu 32(%rdi),%xmm4 |
535 movups %xmm5,48(%rsi) | 535 movups %xmm5,48(%rsi) |
536 movdqu 48(%rdi),%xmm5 | 536 movdqu 48(%rdi),%xmm5 |
537 movups %xmm6,64(%rsi) | 537 movups %xmm6,64(%rsi) |
538 movdqu 64(%rdi),%xmm6 | 538 movdqu 64(%rdi),%xmm6 |
539 movups %xmm7,80(%rsi) | 539 movups %xmm7,80(%rsi) |
540 movdqu 80(%rdi),%xmm7 | 540 movdqu 80(%rdi),%xmm7 |
541 movups %xmm8,96(%rsi) | 541 movups %xmm8,96(%rsi) |
542 movdqu 96(%rdi),%xmm8 | 542 movdqu 96(%rdi),%xmm8 |
543 movups %xmm9,112(%rsi) | 543 movups %xmm9,112(%rsi) |
544 leaq 128(%rsi),%rsi | 544 leaq 128(%rsi),%rsi |
545 movdqu 112(%rdi),%xmm9 | 545 movdqu 112(%rdi),%xmm9 |
546 leaq 128(%rdi),%rdi | 546 leaq 128(%rdi),%rdi |
547 .Lecb_enc_loop8_enter: | 547 .Lecb_enc_loop8_enter: |
548 | 548 |
549 call _aesni_encrypt8 | 549 call _aesni_encrypt8 |
550 | 550 |
551 » subq» $128,%rdx | 551 » subq» $0x80,%rdx |
552 jnc .Lecb_enc_loop8 | 552 jnc .Lecb_enc_loop8 |
553 | 553 |
554 movups %xmm2,(%rsi) | 554 movups %xmm2,(%rsi) |
555 movq %r11,%rcx | 555 movq %r11,%rcx |
556 movups %xmm3,16(%rsi) | 556 movups %xmm3,16(%rsi) |
557 movl %r10d,%eax | 557 movl %r10d,%eax |
558 movups %xmm4,32(%rsi) | 558 movups %xmm4,32(%rsi) |
559 movups %xmm5,48(%rsi) | 559 movups %xmm5,48(%rsi) |
560 movups %xmm6,64(%rsi) | 560 movups %xmm6,64(%rsi) |
561 movups %xmm7,80(%rsi) | 561 movups %xmm7,80(%rsi) |
562 movups %xmm8,96(%rsi) | 562 movups %xmm8,96(%rsi) |
563 movups %xmm9,112(%rsi) | 563 movups %xmm9,112(%rsi) |
564 leaq 128(%rsi),%rsi | 564 leaq 128(%rsi),%rsi |
565 » addq» $128,%rdx | 565 » addq» $0x80,%rdx |
566 jz .Lecb_ret | 566 jz .Lecb_ret |
567 | 567 |
568 .Lecb_enc_tail: | 568 .Lecb_enc_tail: |
569 movups (%rdi),%xmm2 | 569 movups (%rdi),%xmm2 |
570 » cmpq» $32,%rdx | 570 » cmpq» $0x20,%rdx |
571 jb .Lecb_enc_one | 571 jb .Lecb_enc_one |
572 movups 16(%rdi),%xmm3 | 572 movups 16(%rdi),%xmm3 |
573 je .Lecb_enc_two | 573 je .Lecb_enc_two |
574 movups 32(%rdi),%xmm4 | 574 movups 32(%rdi),%xmm4 |
575 » cmpq» $64,%rdx | 575 » cmpq» $0x40,%rdx |
576 jb .Lecb_enc_three | 576 jb .Lecb_enc_three |
577 movups 48(%rdi),%xmm5 | 577 movups 48(%rdi),%xmm5 |
578 je .Lecb_enc_four | 578 je .Lecb_enc_four |
579 movups 64(%rdi),%xmm6 | 579 movups 64(%rdi),%xmm6 |
580 » cmpq» $96,%rdx | 580 » cmpq» $0x60,%rdx |
581 jb .Lecb_enc_five | 581 jb .Lecb_enc_five |
582 movups 80(%rdi),%xmm7 | 582 movups 80(%rdi),%xmm7 |
583 je .Lecb_enc_six | 583 je .Lecb_enc_six |
584 movdqu 96(%rdi),%xmm8 | 584 movdqu 96(%rdi),%xmm8 |
585 xorps %xmm9,%xmm9 | 585 xorps %xmm9,%xmm9 |
586 call _aesni_encrypt8 | 586 call _aesni_encrypt8 |
587 movups %xmm2,(%rsi) | 587 movups %xmm2,(%rsi) |
588 movups %xmm3,16(%rsi) | 588 movups %xmm3,16(%rsi) |
589 movups %xmm4,32(%rsi) | 589 movups %xmm4,32(%rsi) |
590 movups %xmm5,48(%rsi) | 590 movups %xmm5,48(%rsi) |
(...skipping 53 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
644 movups %xmm2,(%rsi) | 644 movups %xmm2,(%rsi) |
645 movups %xmm3,16(%rsi) | 645 movups %xmm3,16(%rsi) |
646 movups %xmm4,32(%rsi) | 646 movups %xmm4,32(%rsi) |
647 movups %xmm5,48(%rsi) | 647 movups %xmm5,48(%rsi) |
648 movups %xmm6,64(%rsi) | 648 movups %xmm6,64(%rsi) |
649 movups %xmm7,80(%rsi) | 649 movups %xmm7,80(%rsi) |
650 jmp .Lecb_ret | 650 jmp .Lecb_ret |
651 | 651 |
652 .align 16 | 652 .align 16 |
653 .Lecb_decrypt: | 653 .Lecb_decrypt: |
654 » cmpq» $128,%rdx | 654 » cmpq» $0x80,%rdx |
655 jb .Lecb_dec_tail | 655 jb .Lecb_dec_tail |
656 | 656 |
657 movdqu (%rdi),%xmm2 | 657 movdqu (%rdi),%xmm2 |
658 movdqu 16(%rdi),%xmm3 | 658 movdqu 16(%rdi),%xmm3 |
659 movdqu 32(%rdi),%xmm4 | 659 movdqu 32(%rdi),%xmm4 |
660 movdqu 48(%rdi),%xmm5 | 660 movdqu 48(%rdi),%xmm5 |
661 movdqu 64(%rdi),%xmm6 | 661 movdqu 64(%rdi),%xmm6 |
662 movdqu 80(%rdi),%xmm7 | 662 movdqu 80(%rdi),%xmm7 |
663 movdqu 96(%rdi),%xmm8 | 663 movdqu 96(%rdi),%xmm8 |
664 movdqu 112(%rdi),%xmm9 | 664 movdqu 112(%rdi),%xmm9 |
665 leaq 128(%rdi),%rdi | 665 leaq 128(%rdi),%rdi |
666 » subq» $128,%rdx | 666 » subq» $0x80,%rdx |
667 jmp .Lecb_dec_loop8_enter | 667 jmp .Lecb_dec_loop8_enter |
668 .align 16 | 668 .align 16 |
669 .Lecb_dec_loop8: | 669 .Lecb_dec_loop8: |
670 movups %xmm2,(%rsi) | 670 movups %xmm2,(%rsi) |
671 movq %r11,%rcx | 671 movq %r11,%rcx |
672 movdqu (%rdi),%xmm2 | 672 movdqu (%rdi),%xmm2 |
673 movl %r10d,%eax | 673 movl %r10d,%eax |
674 movups %xmm3,16(%rsi) | 674 movups %xmm3,16(%rsi) |
675 movdqu 16(%rdi),%xmm3 | 675 movdqu 16(%rdi),%xmm3 |
676 movups %xmm4,32(%rsi) | 676 movups %xmm4,32(%rsi) |
677 movdqu 32(%rdi),%xmm4 | 677 movdqu 32(%rdi),%xmm4 |
678 movups %xmm5,48(%rsi) | 678 movups %xmm5,48(%rsi) |
679 movdqu 48(%rdi),%xmm5 | 679 movdqu 48(%rdi),%xmm5 |
680 movups %xmm6,64(%rsi) | 680 movups %xmm6,64(%rsi) |
681 movdqu 64(%rdi),%xmm6 | 681 movdqu 64(%rdi),%xmm6 |
682 movups %xmm7,80(%rsi) | 682 movups %xmm7,80(%rsi) |
683 movdqu 80(%rdi),%xmm7 | 683 movdqu 80(%rdi),%xmm7 |
684 movups %xmm8,96(%rsi) | 684 movups %xmm8,96(%rsi) |
685 movdqu 96(%rdi),%xmm8 | 685 movdqu 96(%rdi),%xmm8 |
686 movups %xmm9,112(%rsi) | 686 movups %xmm9,112(%rsi) |
687 leaq 128(%rsi),%rsi | 687 leaq 128(%rsi),%rsi |
688 movdqu 112(%rdi),%xmm9 | 688 movdqu 112(%rdi),%xmm9 |
689 leaq 128(%rdi),%rdi | 689 leaq 128(%rdi),%rdi |
690 .Lecb_dec_loop8_enter: | 690 .Lecb_dec_loop8_enter: |
691 | 691 |
692 call _aesni_decrypt8 | 692 call _aesni_decrypt8 |
693 | 693 |
694 movups (%r11),%xmm0 | 694 movups (%r11),%xmm0 |
695 » subq» $128,%rdx | 695 » subq» $0x80,%rdx |
696 jnc .Lecb_dec_loop8 | 696 jnc .Lecb_dec_loop8 |
697 | 697 |
698 movups %xmm2,(%rsi) | 698 movups %xmm2,(%rsi) |
699 pxor %xmm2,%xmm2 | 699 pxor %xmm2,%xmm2 |
700 movq %r11,%rcx | 700 movq %r11,%rcx |
701 movups %xmm3,16(%rsi) | 701 movups %xmm3,16(%rsi) |
702 pxor %xmm3,%xmm3 | 702 pxor %xmm3,%xmm3 |
703 movl %r10d,%eax | 703 movl %r10d,%eax |
704 movups %xmm4,32(%rsi) | 704 movups %xmm4,32(%rsi) |
705 pxor %xmm4,%xmm4 | 705 pxor %xmm4,%xmm4 |
706 movups %xmm5,48(%rsi) | 706 movups %xmm5,48(%rsi) |
707 pxor %xmm5,%xmm5 | 707 pxor %xmm5,%xmm5 |
708 movups %xmm6,64(%rsi) | 708 movups %xmm6,64(%rsi) |
709 pxor %xmm6,%xmm6 | 709 pxor %xmm6,%xmm6 |
710 movups %xmm7,80(%rsi) | 710 movups %xmm7,80(%rsi) |
711 pxor %xmm7,%xmm7 | 711 pxor %xmm7,%xmm7 |
712 movups %xmm8,96(%rsi) | 712 movups %xmm8,96(%rsi) |
713 pxor %xmm8,%xmm8 | 713 pxor %xmm8,%xmm8 |
714 movups %xmm9,112(%rsi) | 714 movups %xmm9,112(%rsi) |
715 pxor %xmm9,%xmm9 | 715 pxor %xmm9,%xmm9 |
716 leaq 128(%rsi),%rsi | 716 leaq 128(%rsi),%rsi |
717 » addq» $128,%rdx | 717 » addq» $0x80,%rdx |
718 jz .Lecb_ret | 718 jz .Lecb_ret |
719 | 719 |
720 .Lecb_dec_tail: | 720 .Lecb_dec_tail: |
721 movups (%rdi),%xmm2 | 721 movups (%rdi),%xmm2 |
722 » cmpq» $32,%rdx | 722 » cmpq» $0x20,%rdx |
723 jb .Lecb_dec_one | 723 jb .Lecb_dec_one |
724 movups 16(%rdi),%xmm3 | 724 movups 16(%rdi),%xmm3 |
725 je .Lecb_dec_two | 725 je .Lecb_dec_two |
726 movups 32(%rdi),%xmm4 | 726 movups 32(%rdi),%xmm4 |
727 » cmpq» $64,%rdx | 727 » cmpq» $0x40,%rdx |
728 jb .Lecb_dec_three | 728 jb .Lecb_dec_three |
729 movups 48(%rdi),%xmm5 | 729 movups 48(%rdi),%xmm5 |
730 je .Lecb_dec_four | 730 je .Lecb_dec_four |
731 movups 64(%rdi),%xmm6 | 731 movups 64(%rdi),%xmm6 |
732 » cmpq» $96,%rdx | 732 » cmpq» $0x60,%rdx |
733 jb .Lecb_dec_five | 733 jb .Lecb_dec_five |
734 movups 80(%rdi),%xmm7 | 734 movups 80(%rdi),%xmm7 |
735 je .Lecb_dec_six | 735 je .Lecb_dec_six |
736 movups 96(%rdi),%xmm8 | 736 movups 96(%rdi),%xmm8 |
737 movups (%rcx),%xmm0 | 737 movups (%rcx),%xmm0 |
738 xorps %xmm9,%xmm9 | 738 xorps %xmm9,%xmm9 |
739 call _aesni_decrypt8 | 739 call _aesni_decrypt8 |
740 movups %xmm2,(%rsi) | 740 movups %xmm2,(%rsi) |
741 pxor %xmm2,%xmm2 | 741 pxor %xmm2,%xmm2 |
742 movups %xmm3,16(%rsi) | 742 movups %xmm3,16(%rsi) |
(...skipping 857 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1600 movq %rcx,%r11 | 1600 movq %rcx,%r11 |
1601 movl %r10d,%eax | 1601 movl %r10d,%eax |
1602 shll $4,%r10d | 1602 shll $4,%r10d |
1603 movq %rdx,%r9 | 1603 movq %rdx,%r9 |
1604 andq $-16,%rdx | 1604 andq $-16,%rdx |
1605 | 1605 |
1606 movups 16(%rcx,%r10,1),%xmm1 | 1606 movups 16(%rcx,%r10,1),%xmm1 |
1607 | 1607 |
1608 movdqa .Lxts_magic(%rip),%xmm8 | 1608 movdqa .Lxts_magic(%rip),%xmm8 |
1609 movdqa %xmm2,%xmm15 | 1609 movdqa %xmm2,%xmm15 |
1610 » pshufd» $95,%xmm2,%xmm9 | 1610 » pshufd» $0x5f,%xmm2,%xmm9 |
1611 pxor %xmm0,%xmm1 | 1611 pxor %xmm0,%xmm1 |
1612 movdqa %xmm9,%xmm14 | 1612 movdqa %xmm9,%xmm14 |
1613 paddd %xmm9,%xmm9 | 1613 paddd %xmm9,%xmm9 |
1614 movdqa %xmm15,%xmm10 | 1614 movdqa %xmm15,%xmm10 |
1615 psrad $31,%xmm14 | 1615 psrad $31,%xmm14 |
1616 paddq %xmm15,%xmm15 | 1616 paddq %xmm15,%xmm15 |
1617 pand %xmm8,%xmm14 | 1617 pand %xmm8,%xmm14 |
1618 pxor %xmm0,%xmm10 | 1618 pxor %xmm0,%xmm10 |
1619 pxor %xmm14,%xmm15 | 1619 pxor %xmm14,%xmm15 |
1620 movdqa %xmm9,%xmm14 | 1620 movdqa %xmm9,%xmm14 |
(...skipping 78 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1699 pxor %xmm9,%xmm14 | 1699 pxor %xmm9,%xmm14 |
1700 movdqa %xmm12,32(%rsp) | 1700 movdqa %xmm12,32(%rsp) |
1701 .byte 102,15,56,220,224 | 1701 .byte 102,15,56,220,224 |
1702 .byte 102,15,56,220,232 | 1702 .byte 102,15,56,220,232 |
1703 pxor %xmm9,%xmm8 | 1703 pxor %xmm9,%xmm8 |
1704 movdqa %xmm14,64(%rsp) | 1704 movdqa %xmm14,64(%rsp) |
1705 .byte 102,15,56,220,240 | 1705 .byte 102,15,56,220,240 |
1706 .byte 102,15,56,220,248 | 1706 .byte 102,15,56,220,248 |
1707 movups 64(%r11),%xmm0 | 1707 movups 64(%r11),%xmm0 |
1708 movdqa %xmm8,80(%rsp) | 1708 movdqa %xmm8,80(%rsp) |
1709 » pshufd» $95,%xmm15,%xmm9 | 1709 » pshufd» $0x5f,%xmm15,%xmm9 |
1710 jmp .Lxts_enc_loop6 | 1710 jmp .Lxts_enc_loop6 |
1711 .align 32 | 1711 .align 32 |
1712 .Lxts_enc_loop6: | 1712 .Lxts_enc_loop6: |
1713 .byte 102,15,56,220,209 | 1713 .byte 102,15,56,220,209 |
1714 .byte 102,15,56,220,217 | 1714 .byte 102,15,56,220,217 |
1715 .byte 102,15,56,220,225 | 1715 .byte 102,15,56,220,225 |
1716 .byte 102,15,56,220,233 | 1716 .byte 102,15,56,220,233 |
1717 .byte 102,15,56,220,241 | 1717 .byte 102,15,56,220,241 |
1718 .byte 102,15,56,220,249 | 1718 .byte 102,15,56,220,249 |
1719 movups -64(%rcx,%rax,1),%xmm1 | 1719 movups -64(%rcx,%rax,1),%xmm1 |
(...skipping 118 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1838 shrl $4,%eax | 1838 shrl $4,%eax |
1839 | 1839 |
1840 .Lxts_enc_short: | 1840 .Lxts_enc_short: |
1841 | 1841 |
1842 movl %eax,%r10d | 1842 movl %eax,%r10d |
1843 pxor %xmm0,%xmm10 | 1843 pxor %xmm0,%xmm10 |
1844 addq $96,%rdx | 1844 addq $96,%rdx |
1845 jz .Lxts_enc_done | 1845 jz .Lxts_enc_done |
1846 | 1846 |
1847 pxor %xmm0,%xmm11 | 1847 pxor %xmm0,%xmm11 |
1848 » cmpq» $32,%rdx | 1848 » cmpq» $0x20,%rdx |
1849 jb .Lxts_enc_one | 1849 jb .Lxts_enc_one |
1850 pxor %xmm0,%xmm12 | 1850 pxor %xmm0,%xmm12 |
1851 je .Lxts_enc_two | 1851 je .Lxts_enc_two |
1852 | 1852 |
1853 pxor %xmm0,%xmm13 | 1853 pxor %xmm0,%xmm13 |
1854 » cmpq» $64,%rdx | 1854 » cmpq» $0x40,%rdx |
1855 jb .Lxts_enc_three | 1855 jb .Lxts_enc_three |
1856 pxor %xmm0,%xmm14 | 1856 pxor %xmm0,%xmm14 |
1857 je .Lxts_enc_four | 1857 je .Lxts_enc_four |
1858 | 1858 |
1859 movdqu (%rdi),%xmm2 | 1859 movdqu (%rdi),%xmm2 |
1860 movdqu 16(%rdi),%xmm3 | 1860 movdqu 16(%rdi),%xmm3 |
1861 movdqu 32(%rdi),%xmm4 | 1861 movdqu 32(%rdi),%xmm4 |
1862 pxor %xmm10,%xmm2 | 1862 pxor %xmm10,%xmm2 |
1863 movdqu 48(%rdi),%xmm5 | 1863 movdqu 48(%rdi),%xmm5 |
1864 pxor %xmm11,%xmm3 | 1864 pxor %xmm11,%xmm3 |
(...skipping 207 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
2072 movq %rcx,%r11 | 2072 movq %rcx,%r11 |
2073 movl %r10d,%eax | 2073 movl %r10d,%eax |
2074 shll $4,%r10d | 2074 shll $4,%r10d |
2075 movq %rdx,%r9 | 2075 movq %rdx,%r9 |
2076 andq $-16,%rdx | 2076 andq $-16,%rdx |
2077 | 2077 |
2078 movups 16(%rcx,%r10,1),%xmm1 | 2078 movups 16(%rcx,%r10,1),%xmm1 |
2079 | 2079 |
2080 movdqa .Lxts_magic(%rip),%xmm8 | 2080 movdqa .Lxts_magic(%rip),%xmm8 |
2081 movdqa %xmm2,%xmm15 | 2081 movdqa %xmm2,%xmm15 |
2082 » pshufd» $95,%xmm2,%xmm9 | 2082 » pshufd» $0x5f,%xmm2,%xmm9 |
2083 pxor %xmm0,%xmm1 | 2083 pxor %xmm0,%xmm1 |
2084 movdqa %xmm9,%xmm14 | 2084 movdqa %xmm9,%xmm14 |
2085 paddd %xmm9,%xmm9 | 2085 paddd %xmm9,%xmm9 |
2086 movdqa %xmm15,%xmm10 | 2086 movdqa %xmm15,%xmm10 |
2087 psrad $31,%xmm14 | 2087 psrad $31,%xmm14 |
2088 paddq %xmm15,%xmm15 | 2088 paddq %xmm15,%xmm15 |
2089 pand %xmm8,%xmm14 | 2089 pand %xmm8,%xmm14 |
2090 pxor %xmm0,%xmm10 | 2090 pxor %xmm0,%xmm10 |
2091 pxor %xmm14,%xmm15 | 2091 pxor %xmm14,%xmm15 |
2092 movdqa %xmm9,%xmm14 | 2092 movdqa %xmm9,%xmm14 |
(...skipping 78 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
2171 pxor %xmm9,%xmm14 | 2171 pxor %xmm9,%xmm14 |
2172 movdqa %xmm12,32(%rsp) | 2172 movdqa %xmm12,32(%rsp) |
2173 .byte 102,15,56,222,224 | 2173 .byte 102,15,56,222,224 |
2174 .byte 102,15,56,222,232 | 2174 .byte 102,15,56,222,232 |
2175 pxor %xmm9,%xmm8 | 2175 pxor %xmm9,%xmm8 |
2176 movdqa %xmm14,64(%rsp) | 2176 movdqa %xmm14,64(%rsp) |
2177 .byte 102,15,56,222,240 | 2177 .byte 102,15,56,222,240 |
2178 .byte 102,15,56,222,248 | 2178 .byte 102,15,56,222,248 |
2179 movups 64(%r11),%xmm0 | 2179 movups 64(%r11),%xmm0 |
2180 movdqa %xmm8,80(%rsp) | 2180 movdqa %xmm8,80(%rsp) |
2181 » pshufd» $95,%xmm15,%xmm9 | 2181 » pshufd» $0x5f,%xmm15,%xmm9 |
2182 jmp .Lxts_dec_loop6 | 2182 jmp .Lxts_dec_loop6 |
2183 .align 32 | 2183 .align 32 |
2184 .Lxts_dec_loop6: | 2184 .Lxts_dec_loop6: |
2185 .byte 102,15,56,222,209 | 2185 .byte 102,15,56,222,209 |
2186 .byte 102,15,56,222,217 | 2186 .byte 102,15,56,222,217 |
2187 .byte 102,15,56,222,225 | 2187 .byte 102,15,56,222,225 |
2188 .byte 102,15,56,222,233 | 2188 .byte 102,15,56,222,233 |
2189 .byte 102,15,56,222,241 | 2189 .byte 102,15,56,222,241 |
2190 .byte 102,15,56,222,249 | 2190 .byte 102,15,56,222,249 |
2191 movups -64(%rcx,%rax,1),%xmm1 | 2191 movups -64(%rcx,%rax,1),%xmm1 |
(...skipping 119 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
2311 | 2311 |
2312 .Lxts_dec_short: | 2312 .Lxts_dec_short: |
2313 | 2313 |
2314 movl %eax,%r10d | 2314 movl %eax,%r10d |
2315 pxor %xmm0,%xmm10 | 2315 pxor %xmm0,%xmm10 |
2316 pxor %xmm0,%xmm11 | 2316 pxor %xmm0,%xmm11 |
2317 addq $96,%rdx | 2317 addq $96,%rdx |
2318 jz .Lxts_dec_done | 2318 jz .Lxts_dec_done |
2319 | 2319 |
2320 pxor %xmm0,%xmm12 | 2320 pxor %xmm0,%xmm12 |
2321 » cmpq» $32,%rdx | 2321 » cmpq» $0x20,%rdx |
2322 jb .Lxts_dec_one | 2322 jb .Lxts_dec_one |
2323 pxor %xmm0,%xmm13 | 2323 pxor %xmm0,%xmm13 |
2324 je .Lxts_dec_two | 2324 je .Lxts_dec_two |
2325 | 2325 |
2326 pxor %xmm0,%xmm14 | 2326 pxor %xmm0,%xmm14 |
2327 » cmpq» $64,%rdx | 2327 » cmpq» $0x40,%rdx |
2328 jb .Lxts_dec_three | 2328 jb .Lxts_dec_three |
2329 je .Lxts_dec_four | 2329 je .Lxts_dec_four |
2330 | 2330 |
2331 movdqu (%rdi),%xmm2 | 2331 movdqu (%rdi),%xmm2 |
2332 movdqu 16(%rdi),%xmm3 | 2332 movdqu 16(%rdi),%xmm3 |
2333 movdqu 32(%rdi),%xmm4 | 2333 movdqu 32(%rdi),%xmm4 |
2334 pxor %xmm10,%xmm2 | 2334 pxor %xmm10,%xmm2 |
2335 movdqu 48(%rdi),%xmm5 | 2335 movdqu 48(%rdi),%xmm5 |
2336 pxor %xmm11,%xmm3 | 2336 pxor %xmm11,%xmm3 |
2337 movdqu 64(%rdi),%xmm6 | 2337 movdqu 64(%rdi),%xmm6 |
(...skipping 10 matching lines...) Expand all Loading... |
2348 movdqu %xmm2,(%rsi) | 2348 movdqu %xmm2,(%rsi) |
2349 xorps %xmm13,%xmm5 | 2349 xorps %xmm13,%xmm5 |
2350 movdqu %xmm3,16(%rsi) | 2350 movdqu %xmm3,16(%rsi) |
2351 xorps %xmm14,%xmm6 | 2351 xorps %xmm14,%xmm6 |
2352 movdqu %xmm4,32(%rsi) | 2352 movdqu %xmm4,32(%rsi) |
2353 pxor %xmm14,%xmm14 | 2353 pxor %xmm14,%xmm14 |
2354 movdqu %xmm5,48(%rsi) | 2354 movdqu %xmm5,48(%rsi) |
2355 pcmpgtd %xmm15,%xmm14 | 2355 pcmpgtd %xmm15,%xmm14 |
2356 movdqu %xmm6,64(%rsi) | 2356 movdqu %xmm6,64(%rsi) |
2357 leaq 80(%rsi),%rsi | 2357 leaq 80(%rsi),%rsi |
2358 » pshufd» $19,%xmm14,%xmm11 | 2358 » pshufd» $0x13,%xmm14,%xmm11 |
2359 andq $15,%r9 | 2359 andq $15,%r9 |
2360 jz .Lxts_dec_ret | 2360 jz .Lxts_dec_ret |
2361 | 2361 |
2362 movdqa %xmm15,%xmm10 | 2362 movdqa %xmm15,%xmm10 |
2363 paddq %xmm15,%xmm15 | 2363 paddq %xmm15,%xmm15 |
2364 pand %xmm8,%xmm11 | 2364 pand %xmm8,%xmm11 |
2365 pxor %xmm15,%xmm11 | 2365 pxor %xmm15,%xmm11 |
2366 jmp .Lxts_dec_done2 | 2366 jmp .Lxts_dec_done2 |
2367 | 2367 |
2368 .align 16 | 2368 .align 16 |
(...skipping 269 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
2638 jmp .Lcbc_ret | 2638 jmp .Lcbc_ret |
2639 .align 16 | 2639 .align 16 |
2640 .Lcbc_decrypt_bulk: | 2640 .Lcbc_decrypt_bulk: |
2641 leaq (%rsp),%rax | 2641 leaq (%rsp),%rax |
2642 pushq %rbp | 2642 pushq %rbp |
2643 subq $16,%rsp | 2643 subq $16,%rsp |
2644 andq $-16,%rsp | 2644 andq $-16,%rsp |
2645 leaq -8(%rax),%rbp | 2645 leaq -8(%rax),%rbp |
2646 movups (%r8),%xmm10 | 2646 movups (%r8),%xmm10 |
2647 movl %r10d,%eax | 2647 movl %r10d,%eax |
2648 » cmpq» $80,%rdx | 2648 » cmpq» $0x50,%rdx |
2649 jbe .Lcbc_dec_tail | 2649 jbe .Lcbc_dec_tail |
2650 | 2650 |
2651 movups (%rcx),%xmm0 | 2651 movups (%rcx),%xmm0 |
2652 movdqu 0(%rdi),%xmm2 | 2652 movdqu 0(%rdi),%xmm2 |
2653 movdqu 16(%rdi),%xmm3 | 2653 movdqu 16(%rdi),%xmm3 |
2654 movdqa %xmm2,%xmm11 | 2654 movdqa %xmm2,%xmm11 |
2655 movdqu 32(%rdi),%xmm4 | 2655 movdqu 32(%rdi),%xmm4 |
2656 movdqa %xmm3,%xmm12 | 2656 movdqa %xmm3,%xmm12 |
2657 movdqu 48(%rdi),%xmm5 | 2657 movdqu 48(%rdi),%xmm5 |
2658 movdqa %xmm4,%xmm13 | 2658 movdqa %xmm4,%xmm13 |
2659 movdqu 64(%rdi),%xmm6 | 2659 movdqu 64(%rdi),%xmm6 |
2660 movdqa %xmm5,%xmm14 | 2660 movdqa %xmm5,%xmm14 |
2661 movdqu 80(%rdi),%xmm7 | 2661 movdqu 80(%rdi),%xmm7 |
2662 movdqa %xmm6,%xmm15 | 2662 movdqa %xmm6,%xmm15 |
2663 movl OPENSSL_ia32cap_P+4(%rip),%r9d | 2663 movl OPENSSL_ia32cap_P+4(%rip),%r9d |
2664 » cmpq» $112,%rdx | 2664 » cmpq» $0x70,%rdx |
2665 jbe .Lcbc_dec_six_or_seven | 2665 jbe .Lcbc_dec_six_or_seven |
2666 | 2666 |
2667 andl $71303168,%r9d | 2667 andl $71303168,%r9d |
2668 » subq» $80,%rdx | 2668 » subq» $0x50,%rdx |
2669 cmpl $4194304,%r9d | 2669 cmpl $4194304,%r9d |
2670 je .Lcbc_dec_loop6_enter | 2670 je .Lcbc_dec_loop6_enter |
2671 » subq» $32,%rdx | 2671 » subq» $0x20,%rdx |
2672 leaq 112(%rcx),%rcx | 2672 leaq 112(%rcx),%rcx |
2673 jmp .Lcbc_dec_loop8_enter | 2673 jmp .Lcbc_dec_loop8_enter |
2674 .align 16 | 2674 .align 16 |
2675 .Lcbc_dec_loop8: | 2675 .Lcbc_dec_loop8: |
2676 movups %xmm9,(%rsi) | 2676 movups %xmm9,(%rsi) |
2677 leaq 16(%rsi),%rsi | 2677 leaq 16(%rsi),%rsi |
2678 .Lcbc_dec_loop8_enter: | 2678 .Lcbc_dec_loop8_enter: |
2679 movdqu 96(%rdi),%xmm8 | 2679 movdqu 96(%rdi),%xmm8 |
2680 pxor %xmm0,%xmm2 | 2680 pxor %xmm0,%xmm2 |
2681 movdqu 112(%rdi),%xmm9 | 2681 movdqu 112(%rdi),%xmm9 |
2682 pxor %xmm0,%xmm3 | 2682 pxor %xmm0,%xmm3 |
2683 movups 16-112(%rcx),%xmm1 | 2683 movups 16-112(%rcx),%xmm1 |
2684 pxor %xmm0,%xmm4 | 2684 pxor %xmm0,%xmm4 |
2685 xorq %r11,%r11 | 2685 xorq %r11,%r11 |
2686 » cmpq» $112,%rdx | 2686 » cmpq» $0x70,%rdx |
2687 pxor %xmm0,%xmm5 | 2687 pxor %xmm0,%xmm5 |
2688 pxor %xmm0,%xmm6 | 2688 pxor %xmm0,%xmm6 |
2689 pxor %xmm0,%xmm7 | 2689 pxor %xmm0,%xmm7 |
2690 pxor %xmm0,%xmm8 | 2690 pxor %xmm0,%xmm8 |
2691 | 2691 |
2692 .byte 102,15,56,222,209 | 2692 .byte 102,15,56,222,209 |
2693 pxor %xmm0,%xmm9 | 2693 pxor %xmm0,%xmm9 |
2694 movups 32-112(%rcx),%xmm0 | 2694 movups 32-112(%rcx),%xmm0 |
2695 .byte 102,15,56,222,217 | 2695 .byte 102,15,56,222,217 |
2696 .byte 102,15,56,222,225 | 2696 .byte 102,15,56,222,225 |
(...skipping 164 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
2861 movdqa %xmm13,%xmm4 | 2861 movdqa %xmm13,%xmm4 |
2862 movups %xmm5,48(%rsi) | 2862 movups %xmm5,48(%rsi) |
2863 movdqa %xmm14,%xmm5 | 2863 movdqa %xmm14,%xmm5 |
2864 movups %xmm6,64(%rsi) | 2864 movups %xmm6,64(%rsi) |
2865 movdqa %xmm15,%xmm6 | 2865 movdqa %xmm15,%xmm6 |
2866 movups %xmm7,80(%rsi) | 2866 movups %xmm7,80(%rsi) |
2867 movdqa %xmm1,%xmm7 | 2867 movdqa %xmm1,%xmm7 |
2868 movups %xmm8,96(%rsi) | 2868 movups %xmm8,96(%rsi) |
2869 leaq 112(%rsi),%rsi | 2869 leaq 112(%rsi),%rsi |
2870 | 2870 |
2871 » subq» $128,%rdx | 2871 » subq» $0x80,%rdx |
2872 ja .Lcbc_dec_loop8 | 2872 ja .Lcbc_dec_loop8 |
2873 | 2873 |
2874 movaps %xmm9,%xmm2 | 2874 movaps %xmm9,%xmm2 |
2875 leaq -112(%rcx),%rcx | 2875 leaq -112(%rcx),%rcx |
2876 » addq» $112,%rdx | 2876 » addq» $0x70,%rdx |
2877 jle .Lcbc_dec_clear_tail_collected | 2877 jle .Lcbc_dec_clear_tail_collected |
2878 movups %xmm9,(%rsi) | 2878 movups %xmm9,(%rsi) |
2879 leaq 16(%rsi),%rsi | 2879 leaq 16(%rsi),%rsi |
2880 » cmpq» $80,%rdx | 2880 » cmpq» $0x50,%rdx |
2881 jbe .Lcbc_dec_tail | 2881 jbe .Lcbc_dec_tail |
2882 | 2882 |
2883 movaps %xmm11,%xmm2 | 2883 movaps %xmm11,%xmm2 |
2884 .Lcbc_dec_six_or_seven: | 2884 .Lcbc_dec_six_or_seven: |
2885 » cmpq» $96,%rdx | 2885 » cmpq» $0x60,%rdx |
2886 ja .Lcbc_dec_seven | 2886 ja .Lcbc_dec_seven |
2887 | 2887 |
2888 movaps %xmm7,%xmm8 | 2888 movaps %xmm7,%xmm8 |
2889 call _aesni_decrypt6 | 2889 call _aesni_decrypt6 |
2890 pxor %xmm10,%xmm2 | 2890 pxor %xmm10,%xmm2 |
2891 movaps %xmm8,%xmm10 | 2891 movaps %xmm8,%xmm10 |
2892 pxor %xmm11,%xmm3 | 2892 pxor %xmm11,%xmm3 |
2893 movdqu %xmm2,(%rsi) | 2893 movdqu %xmm2,(%rsi) |
2894 pxor %xmm12,%xmm4 | 2894 pxor %xmm12,%xmm4 |
2895 movdqu %xmm3,16(%rsi) | 2895 movdqu %xmm3,16(%rsi) |
(...skipping 72 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
2968 movdqu %xmm3,16(%rsi) | 2968 movdqu %xmm3,16(%rsi) |
2969 pxor %xmm13,%xmm5 | 2969 pxor %xmm13,%xmm5 |
2970 movdqu %xmm4,32(%rsi) | 2970 movdqu %xmm4,32(%rsi) |
2971 pxor %xmm14,%xmm6 | 2971 pxor %xmm14,%xmm6 |
2972 movq %r11,%rcx | 2972 movq %r11,%rcx |
2973 movdqu %xmm5,48(%rsi) | 2973 movdqu %xmm5,48(%rsi) |
2974 pxor %xmm15,%xmm7 | 2974 pxor %xmm15,%xmm7 |
2975 movl %r10d,%eax | 2975 movl %r10d,%eax |
2976 movdqu %xmm6,64(%rsi) | 2976 movdqu %xmm6,64(%rsi) |
2977 leaq 80(%rsi),%rsi | 2977 leaq 80(%rsi),%rsi |
2978 » subq» $96,%rdx | 2978 » subq» $0x60,%rdx |
2979 ja .Lcbc_dec_loop6 | 2979 ja .Lcbc_dec_loop6 |
2980 | 2980 |
2981 movdqa %xmm7,%xmm2 | 2981 movdqa %xmm7,%xmm2 |
2982 » addq» $80,%rdx | 2982 » addq» $0x50,%rdx |
2983 jle .Lcbc_dec_clear_tail_collected | 2983 jle .Lcbc_dec_clear_tail_collected |
2984 movups %xmm7,(%rsi) | 2984 movups %xmm7,(%rsi) |
2985 leaq 16(%rsi),%rsi | 2985 leaq 16(%rsi),%rsi |
2986 | 2986 |
2987 .Lcbc_dec_tail: | 2987 .Lcbc_dec_tail: |
2988 movups (%rdi),%xmm2 | 2988 movups (%rdi),%xmm2 |
2989 » subq» $16,%rdx | 2989 » subq» $0x10,%rdx |
2990 jbe .Lcbc_dec_one | 2990 jbe .Lcbc_dec_one |
2991 | 2991 |
2992 movups 16(%rdi),%xmm3 | 2992 movups 16(%rdi),%xmm3 |
2993 movaps %xmm2,%xmm11 | 2993 movaps %xmm2,%xmm11 |
2994 » subq» $16,%rdx | 2994 » subq» $0x10,%rdx |
2995 jbe .Lcbc_dec_two | 2995 jbe .Lcbc_dec_two |
2996 | 2996 |
2997 movups 32(%rdi),%xmm4 | 2997 movups 32(%rdi),%xmm4 |
2998 movaps %xmm3,%xmm12 | 2998 movaps %xmm3,%xmm12 |
2999 » subq» $16,%rdx | 2999 » subq» $0x10,%rdx |
3000 jbe .Lcbc_dec_three | 3000 jbe .Lcbc_dec_three |
3001 | 3001 |
3002 movups 48(%rdi),%xmm5 | 3002 movups 48(%rdi),%xmm5 |
3003 movaps %xmm4,%xmm13 | 3003 movaps %xmm4,%xmm13 |
3004 » subq» $16,%rdx | 3004 » subq» $0x10,%rdx |
3005 jbe .Lcbc_dec_four | 3005 jbe .Lcbc_dec_four |
3006 | 3006 |
3007 movups 64(%rdi),%xmm6 | 3007 movups 64(%rdi),%xmm6 |
3008 movaps %xmm5,%xmm14 | 3008 movaps %xmm5,%xmm14 |
3009 movaps %xmm6,%xmm15 | 3009 movaps %xmm6,%xmm15 |
3010 xorps %xmm7,%xmm7 | 3010 xorps %xmm7,%xmm7 |
3011 call _aesni_decrypt6 | 3011 call _aesni_decrypt6 |
3012 pxor %xmm10,%xmm2 | 3012 pxor %xmm10,%xmm2 |
3013 movaps %xmm15,%xmm10 | 3013 movaps %xmm15,%xmm10 |
3014 pxor %xmm11,%xmm3 | 3014 pxor %xmm11,%xmm3 |
3015 movdqu %xmm2,(%rsi) | 3015 movdqu %xmm2,(%rsi) |
3016 pxor %xmm12,%xmm4 | 3016 pxor %xmm12,%xmm4 |
3017 movdqu %xmm3,16(%rsi) | 3017 movdqu %xmm3,16(%rsi) |
3018 pxor %xmm3,%xmm3 | 3018 pxor %xmm3,%xmm3 |
3019 pxor %xmm13,%xmm5 | 3019 pxor %xmm13,%xmm5 |
3020 movdqu %xmm4,32(%rsi) | 3020 movdqu %xmm4,32(%rsi) |
3021 pxor %xmm4,%xmm4 | 3021 pxor %xmm4,%xmm4 |
3022 pxor %xmm14,%xmm6 | 3022 pxor %xmm14,%xmm6 |
3023 movdqu %xmm5,48(%rsi) | 3023 movdqu %xmm5,48(%rsi) |
3024 pxor %xmm5,%xmm5 | 3024 pxor %xmm5,%xmm5 |
3025 leaq 64(%rsi),%rsi | 3025 leaq 64(%rsi),%rsi |
3026 movdqa %xmm6,%xmm2 | 3026 movdqa %xmm6,%xmm2 |
3027 pxor %xmm6,%xmm6 | 3027 pxor %xmm6,%xmm6 |
3028 pxor %xmm7,%xmm7 | 3028 pxor %xmm7,%xmm7 |
3029 » subq» $16,%rdx | 3029 » subq» $0x10,%rdx |
3030 jmp .Lcbc_dec_tail_collected | 3030 jmp .Lcbc_dec_tail_collected |
3031 | 3031 |
3032 .align 16 | 3032 .align 16 |
3033 .Lcbc_dec_one: | 3033 .Lcbc_dec_one: |
3034 movaps %xmm2,%xmm11 | 3034 movaps %xmm2,%xmm11 |
3035 movups (%rcx),%xmm0 | 3035 movups (%rcx),%xmm0 |
3036 movups 16(%rcx),%xmm1 | 3036 movups 16(%rcx),%xmm1 |
3037 leaq 32(%rcx),%rcx | 3037 leaq 32(%rcx),%rcx |
3038 xorps %xmm0,%xmm2 | 3038 xorps %xmm0,%xmm2 |
3039 .Loop_dec1_17: | 3039 .Loop_dec1_17: |
(...skipping 298 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
3338 leaq 24(%rax),%rax | 3338 leaq 24(%rax),%rax |
3339 | 3339 |
3340 movdqa %xmm0,%xmm3 | 3340 movdqa %xmm0,%xmm3 |
3341 pslldq $4,%xmm0 | 3341 pslldq $4,%xmm0 |
3342 pxor %xmm0,%xmm3 | 3342 pxor %xmm0,%xmm3 |
3343 pslldq $4,%xmm0 | 3343 pslldq $4,%xmm0 |
3344 pxor %xmm0,%xmm3 | 3344 pxor %xmm0,%xmm3 |
3345 pslldq $4,%xmm0 | 3345 pslldq $4,%xmm0 |
3346 pxor %xmm3,%xmm0 | 3346 pxor %xmm3,%xmm0 |
3347 | 3347 |
3348 » pshufd» $255,%xmm0,%xmm3 | 3348 » pshufd» $0xff,%xmm0,%xmm3 |
3349 pxor %xmm1,%xmm3 | 3349 pxor %xmm1,%xmm3 |
3350 pslldq $4,%xmm1 | 3350 pslldq $4,%xmm1 |
3351 pxor %xmm1,%xmm3 | 3351 pxor %xmm1,%xmm3 |
3352 | 3352 |
3353 pxor %xmm2,%xmm0 | 3353 pxor %xmm2,%xmm0 |
3354 pxor %xmm3,%xmm2 | 3354 pxor %xmm3,%xmm2 |
3355 movdqu %xmm0,-16(%rax) | 3355 movdqu %xmm0,-16(%rax) |
3356 | 3356 |
3357 decl %r10d | 3357 decl %r10d |
3358 jnz .Loop_key192 | 3358 jnz .Loop_key192 |
(...skipping 66 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
3425 pslldq $4,%xmm0 | 3425 pslldq $4,%xmm0 |
3426 pxor %xmm3,%xmm0 | 3426 pxor %xmm3,%xmm0 |
3427 pslld $1,%xmm4 | 3427 pslld $1,%xmm4 |
3428 | 3428 |
3429 pxor %xmm2,%xmm0 | 3429 pxor %xmm2,%xmm0 |
3430 movdqu %xmm0,(%rax) | 3430 movdqu %xmm0,(%rax) |
3431 | 3431 |
3432 decl %r10d | 3432 decl %r10d |
3433 jz .Ldone_key256 | 3433 jz .Ldone_key256 |
3434 | 3434 |
3435 » pshufd» $255,%xmm0,%xmm2 | 3435 » pshufd» $0xff,%xmm0,%xmm2 |
3436 pxor %xmm3,%xmm3 | 3436 pxor %xmm3,%xmm3 |
3437 .byte 102,15,56,221,211 | 3437 .byte 102,15,56,221,211 |
3438 | 3438 |
3439 movdqa %xmm1,%xmm3 | 3439 movdqa %xmm1,%xmm3 |
3440 pslldq $4,%xmm1 | 3440 pslldq $4,%xmm1 |
3441 pxor %xmm1,%xmm3 | 3441 pxor %xmm1,%xmm3 |
3442 pslldq $4,%xmm1 | 3442 pslldq $4,%xmm1 |
3443 pxor %xmm1,%xmm3 | 3443 pxor %xmm1,%xmm3 |
3444 pslldq $4,%xmm1 | 3444 pslldq $4,%xmm1 |
3445 pxor %xmm3,%xmm1 | 3445 pxor %xmm3,%xmm1 |
(...skipping 110 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
3556 .Lkey_rotate192: | 3556 .Lkey_rotate192: |
3557 .long 0x04070605,0x04070605,0x04070605,0x04070605 | 3557 .long 0x04070605,0x04070605,0x04070605,0x04070605 |
3558 .Lkey_rcon1: | 3558 .Lkey_rcon1: |
3559 .long 1,1,1,1 | 3559 .long 1,1,1,1 |
3560 .Lkey_rcon1b: | 3560 .Lkey_rcon1b: |
3561 .long 0x1b,0x1b,0x1b,0x1b | 3561 .long 0x1b,0x1b,0x1b,0x1b |
3562 | 3562 |
3563 .byte 65,69,83,32,102,111,114,32,73,110,116,101,108,32,65,69,83,45,78,73,44,32
,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101
,110,115,115,108,46,111,114,103,62,0 | 3563 .byte 65,69,83,32,102,111,114,32,73,110,116,101,108,32,65,69,83,45,78,73,44,32
,67,82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101
,110,115,115,108,46,111,114,103,62,0 |
3564 .align 64 | 3564 .align 64 |
3565 #endif | 3565 #endif |
OLD | NEW |