Lines Matching +full:64 +full:- +full:byte
1 /* Do not modify. This file is auto-generated from chacha-x86.pl. */
11 .byte 243,15,30,251
24 leal OPENSSL_ia32cap_P-.Lpic_point(%eax),%ebp
255 cmpl $64,%ebx
308 leal 64(%ebx),%ebx
317 leal 64(%eax),%eax
318 subl $64,%ecx
370 movb %al,-1(%ecx,%esi,1)
381 .size ChaCha20_ctr32,.-.L_ChaCha20_ctr32_begin
389 .byte 243,15,30,251
406 andl $-64,%esp
408 leal .Lssse3_data-.Lpic_point(%eax),%eax
437 .byte 102,15,56,0,222
446 .byte 102,15,56,0,223
459 .byte 102,15,56,0,222
468 .byte 102,15,56,0,223
484 cmpl $64,%ecx
494 leal 64(%esi),%esi
499 leal 64(%edi),%edi
500 subl $64,%ecx
516 movb %al,-1(%edi,%ebp,1)
526 .size ChaCha20_ssse3,.-.L_ChaCha20_ssse3_begin
527 .align 64
529 .byte 2,3,0,1,6,7,4,5,10,11,8,9,14,15,12,13
530 .byte 3,0,1,2,7,4,5,6,11,8,9,10,15,12,13,14
536 .long 0,-1,-1,-1
537 .align 64
538 .byte 67,104,97,67,104,97,50,48,32,102,111,114,32,120,56,54
539 .byte 44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32
540 .byte 60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111
541 .byte 114,103,62,0
549 .byte 243,15,30,251
565 andl $-64,%esp
567 leal .Lssse3_data-.Lpic_point(%eax),%eax
583 vpsubd 64(%eax),%xmm0,%xmm0
586 vmovdqa %xmm0,64(%ebp)
591 vmovdqa %xmm4,-64(%ebp)
592 vmovdqa %xmm5,-48(%ebp)
593 vmovdqa %xmm6,-32(%ebp)
594 vmovdqa %xmm7,-16(%ebp)
609 vmovdqa %xmm4,-128(%ebp)
610 vmovdqa %xmm5,-112(%ebp)
611 vmovdqa %xmm6,-96(%ebp)
612 vmovdqa %xmm7,-80(%ebp)
618 vmovdqa -112(%ebp),%xmm1
619 vmovdqa -96(%ebp),%xmm2
620 vmovdqa -80(%ebp),%xmm3
621 vmovdqa -48(%ebp),%xmm5
622 vmovdqa -32(%ebp),%xmm6
623 vmovdqa -16(%ebp),%xmm7
624 vmovdqa %xmm1,-112(%ebx)
625 vmovdqa %xmm2,-96(%ebx)
626 vmovdqa %xmm3,-80(%ebx)
627 vmovdqa %xmm5,-48(%ebx)
628 vmovdqa %xmm6,-32(%ebx)
629 vmovdqa %xmm7,-16(%ebx)
632 vmovdqa 64(%ebp),%xmm4
636 vpaddd 64(%eax),%xmm4,%xmm4
639 vmovdqa %xmm4,64(%ebx)
643 vmovdqa %xmm4,64(%ebp)
644 vmovdqa -128(%ebp),%xmm0
646 vmovdqa -64(%ebp),%xmm3
655 .byte 143,232,120,194,246,16
658 vmovdqa -112(%ebx),%xmm1
659 .byte 143,232,120,194,210,12
660 vmovdqa -48(%ebx),%xmm3
665 .byte 143,232,120,194,246,8
666 vmovdqa %xmm0,-128(%ebx)
668 vmovdqa %xmm6,64(%ebx)
671 .byte 143,232,120,194,210,7
673 .byte 143,232,120,194,255,16
674 vmovdqa %xmm2,-64(%ebx)
678 vmovdqa -96(%ebx),%xmm0
679 .byte 143,232,120,194,219,12
680 vmovdqa -32(%ebx),%xmm2
685 .byte 143,232,120,194,255,8
686 vmovdqa %xmm1,-112(%ebx)
691 .byte 143,232,120,194,219,7
693 .byte 143,232,120,194,246,16
694 vmovdqa %xmm3,-48(%ebx)
698 vmovdqa -80(%ebx),%xmm1
699 .byte 143,232,120,194,210,12
700 vmovdqa -16(%ebx),%xmm3
705 .byte 143,232,120,194,246,8
706 vmovdqa %xmm0,-96(%ebx)
711 .byte 143,232,120,194,210,7
712 .byte 143,232,120,194,255,16
713 vmovdqa %xmm2,-32(%ebx)
716 vmovdqa -128(%ebx),%xmm0
717 .byte 143,232,120,194,219,12
718 vmovdqa -48(%ebx),%xmm2
722 .byte 143,232,120,194,255,8
723 vmovdqa %xmm1,-80(%ebx)
727 .byte 143,232,120,194,219,7
728 .byte 143,232,120,194,246,16
729 vmovdqa %xmm3,-16(%ebx)
732 vmovdqa -112(%ebx),%xmm1
733 .byte 143,232,120,194,210,12
734 vmovdqa -32(%ebx),%xmm3
736 vmovdqa 64(%ebx),%xmm7
739 .byte 143,232,120,194,246,8
740 vmovdqa %xmm0,-128(%ebx)
745 .byte 143,232,120,194,210,7
747 .byte 143,232,120,194,255,16
748 vmovdqa %xmm2,-48(%ebx)
752 vmovdqa -96(%ebx),%xmm0
753 .byte 143,232,120,194,219,12
754 vmovdqa -16(%ebx),%xmm2
759 .byte 143,232,120,194,255,8
760 vmovdqa %xmm1,-112(%ebx)
762 vmovdqa %xmm7,64(%ebx)
765 .byte 143,232,120,194,219,7
767 .byte 143,232,120,194,246,16
768 vmovdqa %xmm3,-32(%ebx)
772 vmovdqa -80(%ebx),%xmm1
773 .byte 143,232,120,194,210,12
774 vmovdqa -64(%ebx),%xmm3
779 .byte 143,232,120,194,246,8
780 vmovdqa %xmm0,-96(%ebx)
785 .byte 143,232,120,194,210,7
786 .byte 143,232,120,194,255,16
787 vmovdqa %xmm2,-16(%ebx)
790 vmovdqa -128(%ebx),%xmm0
791 .byte 143,232,120,194,219,12
793 vmovdqa 64(%ebx),%xmm6
795 .byte 143,232,120,194,255,8
796 vmovdqa %xmm1,-80(%ebx)
800 .byte 143,232,120,194,219,7
803 vmovdqa %xmm3,-64(%ebx)
806 vmovdqa %xmm6,64(%ebx)
808 vmovdqa -112(%ebx),%xmm1
809 vmovdqa -96(%ebx),%xmm2
810 vmovdqa -80(%ebx),%xmm3
811 vpaddd -128(%ebp),%xmm0,%xmm0
812 vpaddd -112(%ebp),%xmm1,%xmm1
813 vpaddd -96(%ebp),%xmm2,%xmm2
814 vpaddd -80(%ebp),%xmm3,%xmm3
823 vpxor -128(%esi),%xmm1,%xmm4
824 vpxor -64(%esi),%xmm6,%xmm5
826 vpxor 64(%esi),%xmm3,%xmm7
828 vmovdqa -64(%ebx),%xmm0
829 vmovdqa -48(%ebx),%xmm1
830 vmovdqa -32(%ebx),%xmm2
831 vmovdqa -16(%ebx),%xmm3
832 vmovdqu %xmm4,-128(%edi)
833 vmovdqu %xmm5,-64(%edi)
835 vmovdqu %xmm7,64(%edi)
837 vpaddd -64(%ebp),%xmm0,%xmm0
838 vpaddd -48(%ebp),%xmm1,%xmm1
839 vpaddd -32(%ebp),%xmm2,%xmm2
840 vpaddd -16(%ebp),%xmm3,%xmm3
849 vpxor -128(%esi),%xmm1,%xmm4
850 vpxor -64(%esi),%xmm6,%xmm5
852 vpxor 64(%esi),%xmm3,%xmm7
858 vmovdqu %xmm4,-128(%edi)
859 vmovdqu %xmm5,-64(%edi)
861 vmovdqu %xmm7,64(%edi)
875 vpxor -128(%esi),%xmm1,%xmm4
876 vpxor -64(%esi),%xmm6,%xmm5
878 vpxor 64(%esi),%xmm3,%xmm7
880 vmovdqa 64(%ebx),%xmm0
884 vmovdqu %xmm4,-128(%edi)
885 vmovdqu %xmm5,-64(%edi)
887 vmovdqu %xmm7,64(%edi)
889 vpaddd 64(%ebp),%xmm0,%xmm0
901 vpxor -128(%esi),%xmm1,%xmm4
902 vpxor -64(%esi),%xmm6,%xmm5
904 vpxor 64(%esi),%xmm3,%xmm7
906 vmovdqu %xmm4,-128(%edi)
907 vmovdqu %xmm5,-64(%edi)
909 vmovdqu %xmm7,64(%edi)
916 leal -128(%esi),%esi
918 leal -128(%edi),%edi
919 vmovd 64(%ebp),%xmm2
951 .byte 143,232,120,194,219,16
954 .byte 143,232,120,194,201,12
957 .byte 143,232,120,194,219,8
960 .byte 143,232,120,194,201,7
966 .byte 143,232,120,194,219,16
969 .byte 143,232,120,194,201,12
972 .byte 143,232,120,194,219,8
975 .byte 143,232,120,194,201,7
985 cmpl $64,%ecx
991 leal 64(%esi),%esi
996 leal 64(%edi),%edi
997 subl $64,%ecx
1013 movb %al,-1(%edi,%ebp,1)
1024 .size ChaCha20_xop,.-.L_ChaCha20_xop_begin
1029 .long 1f - 0f
1030 .long 4f - 1f
1037 .long 3f - 2f
1052 .byte 243,15,30,251
1296 cmpl $64,%ebx
1349 leal 64(%ebx),%ebx
1358 leal 64(%eax),%eax
1359 subl $64,%ecx
1411 movb %al,-1(%ecx,%esi,1)
1422 .size ChaCha20_ctr32,.-.L_ChaCha20_ctr32_begin
1430 .byte 243,15,30,251
1447 andl $-64,%esp
1449 leal .Lssse3_data-.Lpic_point(%eax),%eax
1478 .byte 102,15,56,0,222
1487 .byte 102,15,56,0,223
1500 .byte 102,15,56,0,222
1509 .byte 102,15,56,0,223
1525 cmpl $64,%ecx
1535 leal 64(%esi),%esi
1540 leal 64(%edi),%edi
1541 subl $64,%ecx
1557 movb %al,-1(%edi,%ebp,1)
1567 .size ChaCha20_ssse3,.-.L_ChaCha20_ssse3_begin
1568 .align 64
1570 .byte 2,3,0,1,6,7,4,5,10,11,8,9,14,15,12,13
1571 .byte 3,0,1,2,7,4,5,6,11,8,9,10,15,12,13,14
1577 .long 0,-1,-1,-1
1578 .align 64
1579 .byte 67,104,97,67,104,97,50,48,32,102,111,114,32,120,56,54
1580 .byte 44,32,67,82,89,80,84,79,71,65,77,83,32,98,121,32
1581 .byte 60,97,112,112,114,111,64,111,112,101,110,115,115,108,46,111
1582 .byte 114,103,62,0
1590 .byte 243,15,30,251
1606 andl $-64,%esp
1608 leal .Lssse3_data-.Lpic_point(%eax),%eax
1624 vpsubd 64(%eax),%xmm0,%xmm0
1627 vmovdqa %xmm0,64(%ebp)
1632 vmovdqa %xmm4,-64(%ebp)
1633 vmovdqa %xmm5,-48(%ebp)
1634 vmovdqa %xmm6,-32(%ebp)
1635 vmovdqa %xmm7,-16(%ebp)
1650 vmovdqa %xmm4,-128(%ebp)
1651 vmovdqa %xmm5,-112(%ebp)
1652 vmovdqa %xmm6,-96(%ebp)
1653 vmovdqa %xmm7,-80(%ebp)
1659 vmovdqa -112(%ebp),%xmm1
1660 vmovdqa -96(%ebp),%xmm2
1661 vmovdqa -80(%ebp),%xmm3
1662 vmovdqa -48(%ebp),%xmm5
1663 vmovdqa -32(%ebp),%xmm6
1664 vmovdqa -16(%ebp),%xmm7
1665 vmovdqa %xmm1,-112(%ebx)
1666 vmovdqa %xmm2,-96(%ebx)
1667 vmovdqa %xmm3,-80(%ebx)
1668 vmovdqa %xmm5,-48(%ebx)
1669 vmovdqa %xmm6,-32(%ebx)
1670 vmovdqa %xmm7,-16(%ebx)
1673 vmovdqa 64(%ebp),%xmm4
1677 vpaddd 64(%eax),%xmm4,%xmm4
1680 vmovdqa %xmm4,64(%ebx)
1684 vmovdqa %xmm4,64(%ebp)
1685 vmovdqa -128(%ebp),%xmm0
1687 vmovdqa -64(%ebp),%xmm3
1696 .byte 143,232,120,194,246,16
1699 vmovdqa -112(%ebx),%xmm1
1700 .byte 143,232,120,194,210,12
1701 vmovdqa -48(%ebx),%xmm3
1706 .byte 143,232,120,194,246,8
1707 vmovdqa %xmm0,-128(%ebx)
1709 vmovdqa %xmm6,64(%ebx)
1712 .byte 143,232,120,194,210,7
1714 .byte 143,232,120,194,255,16
1715 vmovdqa %xmm2,-64(%ebx)
1719 vmovdqa -96(%ebx),%xmm0
1720 .byte 143,232,120,194,219,12
1721 vmovdqa -32(%ebx),%xmm2
1726 .byte 143,232,120,194,255,8
1727 vmovdqa %xmm1,-112(%ebx)
1732 .byte 143,232,120,194,219,7
1734 .byte 143,232,120,194,246,16
1735 vmovdqa %xmm3,-48(%ebx)
1739 vmovdqa -80(%ebx),%xmm1
1740 .byte 143,232,120,194,210,12
1741 vmovdqa -16(%ebx),%xmm3
1746 .byte 143,232,120,194,246,8
1747 vmovdqa %xmm0,-96(%ebx)
1752 .byte 143,232,120,194,210,7
1753 .byte 143,232,120,194,255,16
1754 vmovdqa %xmm2,-32(%ebx)
1757 vmovdqa -128(%ebx),%xmm0
1758 .byte 143,232,120,194,219,12
1759 vmovdqa -48(%ebx),%xmm2
1763 .byte 143,232,120,194,255,8
1764 vmovdqa %xmm1,-80(%ebx)
1768 .byte 143,232,120,194,219,7
1769 .byte 143,232,120,194,246,16
1770 vmovdqa %xmm3,-16(%ebx)
1773 vmovdqa -112(%ebx),%xmm1
1774 .byte 143,232,120,194,210,12
1775 vmovdqa -32(%ebx),%xmm3
1777 vmovdqa 64(%ebx),%xmm7
1780 .byte 143,232,120,194,246,8
1781 vmovdqa %xmm0,-128(%ebx)
1786 .byte 143,232,120,194,210,7
1788 .byte 143,232,120,194,255,16
1789 vmovdqa %xmm2,-48(%ebx)
1793 vmovdqa -96(%ebx),%xmm0
1794 .byte 143,232,120,194,219,12
1795 vmovdqa -16(%ebx),%xmm2
1800 .byte 143,232,120,194,255,8
1801 vmovdqa %xmm1,-112(%ebx)
1803 vmovdqa %xmm7,64(%ebx)
1806 .byte 143,232,120,194,219,7
1808 .byte 143,232,120,194,246,16
1809 vmovdqa %xmm3,-32(%ebx)
1813 vmovdqa -80(%ebx),%xmm1
1814 .byte 143,232,120,194,210,12
1815 vmovdqa -64(%ebx),%xmm3
1820 .byte 143,232,120,194,246,8
1821 vmovdqa %xmm0,-96(%ebx)
1826 .byte 143,232,120,194,210,7
1827 .byte 143,232,120,194,255,16
1828 vmovdqa %xmm2,-16(%ebx)
1831 vmovdqa -128(%ebx),%xmm0
1832 .byte 143,232,120,194,219,12
1834 vmovdqa 64(%ebx),%xmm6
1836 .byte 143,232,120,194,255,8
1837 vmovdqa %xmm1,-80(%ebx)
1841 .byte 143,232,120,194,219,7
1844 vmovdqa %xmm3,-64(%ebx)
1847 vmovdqa %xmm6,64(%ebx)
1849 vmovdqa -112(%ebx),%xmm1
1850 vmovdqa -96(%ebx),%xmm2
1851 vmovdqa -80(%ebx),%xmm3
1852 vpaddd -128(%ebp),%xmm0,%xmm0
1853 vpaddd -112(%ebp),%xmm1,%xmm1
1854 vpaddd -96(%ebp),%xmm2,%xmm2
1855 vpaddd -80(%ebp),%xmm3,%xmm3
1864 vpxor -128(%esi),%xmm1,%xmm4
1865 vpxor -64(%esi),%xmm6,%xmm5
1867 vpxor 64(%esi),%xmm3,%xmm7
1869 vmovdqa -64(%ebx),%xmm0
1870 vmovdqa -48(%ebx),%xmm1
1871 vmovdqa -32(%ebx),%xmm2
1872 vmovdqa -16(%ebx),%xmm3
1873 vmovdqu %xmm4,-128(%edi)
1874 vmovdqu %xmm5,-64(%edi)
1876 vmovdqu %xmm7,64(%edi)
1878 vpaddd -64(%ebp),%xmm0,%xmm0
1879 vpaddd -48(%ebp),%xmm1,%xmm1
1880 vpaddd -32(%ebp),%xmm2,%xmm2
1881 vpaddd -16(%ebp),%xmm3,%xmm3
1890 vpxor -128(%esi),%xmm1,%xmm4
1891 vpxor -64(%esi),%xmm6,%xmm5
1893 vpxor 64(%esi),%xmm3,%xmm7
1899 vmovdqu %xmm4,-128(%edi)
1900 vmovdqu %xmm5,-64(%edi)
1902 vmovdqu %xmm7,64(%edi)
1916 vpxor -128(%esi),%xmm1,%xmm4
1917 vpxor -64(%esi),%xmm6,%xmm5
1919 vpxor 64(%esi),%xmm3,%xmm7
1921 vmovdqa 64(%ebx),%xmm0
1925 vmovdqu %xmm4,-128(%edi)
1926 vmovdqu %xmm5,-64(%edi)
1928 vmovdqu %xmm7,64(%edi)
1930 vpaddd 64(%ebp),%xmm0,%xmm0
1942 vpxor -128(%esi),%xmm1,%xmm4
1943 vpxor -64(%esi),%xmm6,%xmm5
1945 vpxor 64(%esi),%xmm3,%xmm7
1947 vmovdqu %xmm4,-128(%edi)
1948 vmovdqu %xmm5,-64(%edi)
1950 vmovdqu %xmm7,64(%edi)
1957 leal -128(%esi),%esi
1959 leal -128(%edi),%edi
1960 vmovd 64(%ebp),%xmm2
1992 .byte 143,232,120,194,219,16
1995 .byte 143,232,120,194,201,12
1998 .byte 143,232,120,194,219,8
2001 .byte 143,232,120,194,201,7
2007 .byte 143,232,120,194,219,16
2010 .byte 143,232,120,194,201,12
2013 .byte 143,232,120,194,219,8
2016 .byte 143,232,120,194,201,7
2026 cmpl $64,%ecx
2032 leal 64(%esi),%esi
2037 leal 64(%edi),%edi
2038 subl $64,%ecx
2054 movb %al,-1(%edi,%ebp,1)
2065 .size ChaCha20_xop,.-.L_ChaCha20_xop_begin
2070 .long 1f - 0f
2071 .long 4f - 1f
2078 .long 3f - 2f