/freebsd/contrib/llvm-project/llvm/lib/Support/BLAKE3/ |
H A D | blake3_sse41_x86-64_windows_gnu.S | 27 movdqa xmmword ptr [rsp+0x190], xmm8 92 movdqu xmm8, xmmword ptr [r8+rdx-0x40] 96 movdqa xmm12, xmm8 97 punpckldq xmm8, xmm9 102 movdqa xmm9, xmm8 103 punpcklqdq xmm8, xmm10 108 movdqa xmmword ptr [rsp], xmm8 112 movdqu xmm8, xmmword ptr [r8+rdx-0x30] 116 movdqa xmm12, xmm8 117 punpckldq xmm8, xmm9 [all …]
|
H A D | blake3_sse41_x86-64_unix.S | 107 movdqu xmm8, xmmword ptr [r8+rdx-0x40] 111 movdqa xmm12, xmm8 112 punpckldq xmm8, xmm9 117 movdqa xmm9, xmm8 118 punpcklqdq xmm8, xmm10 123 movdqa xmmword ptr [rsp], xmm8 127 movdqu xmm8, xmmword ptr [r8+rdx-0x30] 131 movdqa xmm12, xmm8 132 punpckldq xmm8, xmm9 137 movdqa xmm9, xmm8 [all …]
|
H A D | blake3_sse2_x86-64_unix.S | 107 movdqu xmm8, xmmword ptr [r8+rdx-0x40] 111 movdqa xmm12, xmm8 112 punpckldq xmm8, xmm9 117 movdqa xmm9, xmm8 118 punpcklqdq xmm8, xmm10 123 movdqa xmmword ptr [rsp], xmm8 127 movdqu xmm8, xmmword ptr [r8+rdx-0x30] 131 movdqa xmm12, xmm8 132 punpckldq xmm8, xmm9 137 movdqa xmm9, xmm8 [all …]
|
H A D | blake3_sse41_x86-64_windows_msvc.asm | 26 movdqa xmmword ptr [rsp+190H], xmm8 91 movdqu xmm8, xmmword ptr [r8+rdx-40H] 95 movdqa xmm12, xmm8 96 punpckldq xmm8, xmm9 101 movdqa xmm9, xmm8 102 punpcklqdq xmm8, xmm10 107 movdqa xmmword ptr [rsp], xmm8 111 movdqu xmm8, xmmword ptr [r8+rdx-30H] 115 movdqa xmm12, xmm8 116 punpckldq xmm8, xmm9 [all …]
|
H A D | blake3_sse2_x86-64_windows_gnu.S | 27 movdqa xmmword ptr [rsp+0x190], xmm8 92 movdqu xmm8, xmmword ptr [r8+rdx-0x40] 96 movdqa xmm12, xmm8 97 punpckldq xmm8, xmm9 102 movdqa xmm9, xmm8 103 punpcklqdq xmm8, xmm10 108 movdqa xmmword ptr [rsp], xmm8 112 movdqu xmm8, xmmword ptr [r8+rdx-0x30] 116 movdqa xmm12, xmm8 117 punpckldq xmm8, xmm9 [all …]
|
H A D | blake3_sse2_x86-64_windows_msvc.asm | 26 movdqa xmmword ptr [rsp+190H], xmm8 91 movdqu xmm8, xmmword ptr [r8+rdx-40H] 95 movdqa xmm12, xmm8 96 punpckldq xmm8, xmm9 101 movdqa xmm9, xmm8 102 punpcklqdq xmm8, xmm10 107 movdqa xmmword ptr [rsp], xmm8 111 movdqu xmm8, xmmword ptr [r8+rdx-30H] 115 movdqa xmm12, xmm8 116 punpckldq xmm8, xmm9 [all …]
|
/freebsd/sys/contrib/openzfs/module/icp/asm-x86_64/blake3/ |
H A D | blake3_sse41.S | 99 movdqu xmm8, xmmword ptr [r8+rdx-0x40] 103 movdqa xmm12, xmm8 104 punpckldq xmm8, xmm9 109 movdqa xmm9, xmm8 110 punpcklqdq xmm8, xmm10 115 movdqa xmmword ptr [rsp], xmm8 119 movdqu xmm8, xmmword ptr [r8+rdx-0x30] 123 movdqa xmm12, xmm8 124 punpckldq xmm8, xmm9 129 movdqa xmm9, xmm8 [all …]
|
H A D | blake3_sse2.S | 99 movdqu xmm8, xmmword ptr [r8+rdx-0x40] 103 movdqa xmm12, xmm8 104 punpckldq xmm8, xmm9 109 movdqa xmm9, xmm8 110 punpcklqdq xmm8, xmm10 115 movdqa xmmword ptr [rsp], xmm8 119 movdqu xmm8, xmmword ptr [r8+rdx-0x30] 123 movdqa xmm12, xmm8 124 punpckldq xmm8, xmm9 129 movdqa xmm9, xmm8 [all …]
|
/freebsd/sys/crypto/openssl/amd64/ |
H A D | sha1-mb-x86_64.S | 97 movd -60(%r10),%xmm8 100 punpckldq %xmm8,%xmm1 101 movdqa %xmm10,%xmm8 106 pslld $5,%xmm8 119 por %xmm9,%xmm8 125 paddd %xmm8,%xmm14 127 movd -56(%r10),%xmm8 130 punpckldq %xmm8,%xmm2 131 movdqa %xmm14,%xmm8 136 pslld $5,%xmm8 [all …]
|
H A D | aesni-gcm-x86_64.S | 53 vpxor %xmm4,%xmm8,%xmm8 64 vpxor 16+8(%rsp),%xmm8,%xmm8 128 vpxor 112+8(%rsp),%xmm8,%xmm8 142 vpclmulqdq $0x10,%xmm3,%xmm8,%xmm5 145 vpclmulqdq $0x01,%xmm3,%xmm8,%xmm1 149 vpclmulqdq $0x00,%xmm3,%xmm8,%xmm2 152 vpclmulqdq $0x11,%xmm3,%xmm8,%xmm8 167 vpxor %xmm8,%xmm7,%xmm7 258 vpalignr $8,%xmm4,%xmm4,%xmm8 311 vpxor 16+8(%rsp),%xmm8,%xmm8 [all …]
|
H A D | bsaes-x86_64.S | 13 movdqa (%rax),%xmm8 16 pxor %xmm8,%xmm15 17 pxor %xmm8,%xmm0 18 pxor %xmm8,%xmm1 19 pxor %xmm8,%xmm2 22 pxor %xmm8,%xmm3 23 pxor %xmm8,%xmm4 26 pxor %xmm8,%xmm5 27 pxor %xmm8,%xmm6 34 movdqa 16(%r11),%xmm8 [all …]
|
H A D | ghash-x86_64.S | 976 movdqu 0(%rdx),%xmm8 981 pxor %xmm8,%xmm0 985 pshufd $78,%xmm0,%xmm8 986 pxor %xmm0,%xmm8 1013 xorps %xmm4,%xmm8 1017 pxor %xmm0,%xmm8 1019 pxor %xmm1,%xmm8 1021 movdqa %xmm8,%xmm9 1023 pslldq $8,%xmm8 1025 pxor %xmm8,%xmm0 [all …]
|
H A D | aesni-mb-x86_64.S | 109 movdqu (%r10),%xmm8 112 pxor %xmm8,%xmm4 238 movdqu (%r10,%rbx,1),%xmm8 242 pxor %xmm12,%xmm8 252 pxor %xmm8,%xmm4 373 movdqu 24(%rdi),%xmm8 521 pxor %xmm0,%xmm8 531 movdqu -16(%r10,%rbx,1),%xmm8 719 vmovdqu 104(%rdi),%xmm8 762 vpxor %xmm12,%xmm8,%xmm8 [all …]
|
H A D | aesni-sha1-x86_64.S | 106 pshufd $238,%xmm4,%xmm8 112 punpcklqdq %xmm5,%xmm8 119 pxor %xmm4,%xmm8 126 pxor %xmm12,%xmm8 134 movdqa %xmm8,%xmm3 138 movdqa %xmm8,%xmm12 141 paddd %xmm8,%xmm8 154 por %xmm12,%xmm8 161 pxor %xmm3,%xmm8 167 pxor %xmm13,%xmm8 [all …]
|
H A D | aesni-x86_64.S | 403 pxor %xmm0,%xmm8 469 pxor %xmm0,%xmm8 544 movdqu 96(%rdi),%xmm8 565 movups %xmm8,96(%rsi) 566 movdqu 96(%rdi),%xmm8 586 movups %xmm8,96(%rsi) 608 movdqu 96(%rdi),%xmm8 617 movups %xmm8,96(%rsi) 687 movdqu 96(%rdi),%xmm8 708 movups %xmm8,96(%rsi) [all …]
|
H A D | poly1305-x86_64.S | 687 vpsrldq $6,%xmm6,%xmm8 690 vpunpcklqdq %xmm8,%xmm7,%xmm8 695 vpsrlq $4,%xmm8,%xmm7 697 vpsrlq $30,%xmm8,%xmm8 699 vpand %xmm15,%xmm8,%xmm8 780 vpmuludq %xmm8,%xmm14,%xmm13 786 vpmuludq %xmm8,%xmm2,%xmm1 808 vpmuludq %xmm8,%xmm4,%xmm4 819 vpmuludq %xmm8,%xmm3,%xmm1 829 vpmuludq %xmm8,%xmm4,%xmm8 [all …]
|
H A D | rsaz-x86_64.S | 841 movd %r9d,%xmm8 845 pshufd $0,%xmm8,%xmm8 849 pcmpeqd %xmm8,%xmm0 852 pcmpeqd %xmm8,%xmm1 855 pcmpeqd %xmm8,%xmm2 858 pcmpeqd %xmm8,%xmm3 861 pcmpeqd %xmm8,%xmm4 863 pcmpeqd %xmm8,%xmm5 865 pcmpeqd %xmm8,%xmm6 866 pcmpeqd %xmm8,%xmm7 [all …]
|
H A D | sha256-mb-x86_64.S | 73 movdqu 0-128(%rdi),%xmm8 123 movdqa %xmm8,%xmm1 125 movdqa %xmm8,%xmm2 130 movdqa %xmm8,%xmm7 132 pxor %xmm8,%xmm3 194 movdqa %xmm8,%xmm4 210 movdqa %xmm8,%xmm14 344 paddd %xmm5,%xmm8 356 movdqa %xmm8,%xmm7 358 movdqa %xmm8,%xmm2 [all …]
|
/freebsd/sys/contrib/libsodium/src/libsodium/crypto_scalarmult/curve25519/sandy2x/ |
H A D | ladder_base.S | 48 vpxor %xmm8,%xmm8,%xmm8 146 vpxor %xmm13,%xmm8,%xmm15 148 vpxor %xmm15,%xmm8,%xmm8 222 vpaddq %xmm9,%xmm8,%xmm3 223 vpaddq subc2(%rip),%xmm8,%xmm8 224 vpsubq %xmm9,%xmm8,%xmm8 225 vpunpckhqdq %xmm3,%xmm8,%xmm9 226 vpunpcklqdq %xmm3,%xmm8,%xmm3 228 vpmuludq %xmm7,%xmm12,%xmm8 233 vpaddq %xmm7,%xmm8,%xmm8 [all …]
|
H A D | ladder.S | 53 vpxor %xmm8,%xmm8,%xmm8 195 vpxor %xmm13,%xmm8,%xmm15 197 vpxor %xmm15,%xmm8,%xmm8 271 vpaddq %xmm9,%xmm8,%xmm3 272 vpaddq subc2(%rip),%xmm8,%xmm8 273 vpsubq %xmm9,%xmm8,%xmm8 274 vpunpckhqdq %xmm3,%xmm8,%xmm9 275 vpunpcklqdq %xmm3,%xmm8,%xmm3 277 vpmuludq %xmm7,%xmm12,%xmm8 282 vpaddq %xmm7,%xmm8,%xmm8 [all …]
|
/freebsd/sys/contrib/openzfs/module/icp/asm-x86_64/modes/ |
H A D | ghash-x86_64.S | 307 vpunpckhqdq %xmm15,%xmm15,%xmm8 310 vpxor %xmm15,%xmm8,%xmm8 319 vpclmulqdq $0x10,%xmm7,%xmm8,%xmm5 326 vpunpckhqdq %xmm15,%xmm15,%xmm8 331 vpxor %xmm15,%xmm8,%xmm8 342 vpclmulqdq $0x10,%xmm7,%xmm8,%xmm5 354 vpunpckhqdq %xmm15,%xmm15,%xmm8 356 vpxor %xmm15,%xmm8,%xmm8 367 vpclmulqdq $0x10,%xmm7,%xmm8,%xmm5 391 vpunpckhqdq %xmm15,%xmm15,%xmm8 [all …]
|
H A D | aesni-gcm-x86_64.S | 115 vpxor %xmm4,%xmm8,%xmm8 126 vpxor 16+8(%rsp),%xmm8,%xmm8 190 vpxor 112+8(%rsp),%xmm8,%xmm8 204 vpclmulqdq $0x10,%xmm3,%xmm8,%xmm5 207 vpclmulqdq $0x01,%xmm3,%xmm8,%xmm1 211 vpclmulqdq $0x00,%xmm3,%xmm8,%xmm2 214 vpclmulqdq $0x11,%xmm3,%xmm8,%xmm8 229 vpxor %xmm8,%xmm7,%xmm7 321 vpalignr $8,%xmm4,%xmm4,%xmm8 374 vpxor 16+8(%rsp),%xmm8,%xmm8 [all …]
|
H A D | gcm_pclmulqdq.S | 185 movdqu %xmm6, %xmm8 189 psrld $31, %xmm8 191 pslldq $4, %xmm8 195 por %xmm8, %xmm6 204 movdqu %xmm3, %xmm8 207 pslld $30, %xmm8 // packed right shift shifting << 30 209 pxor %xmm8, %xmm7 // xor the shifted versions 211 movdqu %xmm7, %xmm8 213 psrldq $4, %xmm8 229 pxor %xmm8, %xmm2
|
/freebsd/crypto/openssl/crypto/bn/asm/ |
H A D | rsaz-x86_64.pl | 948 movaps %xmm8,0xc0(%rsp) 959 movd $pwr,%xmm8 963 pshufd \$0,%xmm8,%xmm8 # broadcast $power 973 pcmpeqd %xmm8,%xmm`$i` 980 pcmpeqd %xmm8,%xmm`$i` 984 pcmpeqd %xmm8,%xmm7 986 movdqa 16*0($bp),%xmm8 990 pand %xmm0,%xmm8 1003 por %xmm10,%xmm8 1005 por %xmm12,%xmm8 [all …]
|
/freebsd/crypto/openssl/crypto/aes/asm/ |
H A D | vpaes-x86_64.pl | 90 ## Preserves %xmm6 - %xmm8 so you get some local vectors 303 movdqa .Lk_rcon(%rip), %xmm8 # load rcon 504 ## Adds rcon from low byte of %xmm8, then rotates %xmm8 for 519 palignr \$15, %xmm8, %xmm1 520 palignr \$15, %xmm8, %xmm8 707 movaps %xmm8,0x30(%rsp) 730 movaps 0x30(%rsp),%xmm8 758 movaps %xmm8,0x30(%rsp) 786 movaps 0x30(%rsp),%xmm8 814 movaps %xmm8,0x30(%rsp) [all …]
|