| /linux/lib/crypto/x86/ |
| H A D | blake2s-core.S | 92 movd (DATA,%rax,4),%xmm4 99 punpckldq %xmm5,%xmm4 101 punpcklqdq %xmm6,%xmm4 102 paddd %xmm4,%xmm0 119 movd (DATA,%rax,4),%xmm4 121 punpckldq %xmm4,%xmm7 141 movd (DATA,%rax,4),%xmm4 145 punpckldq %xmm5,%xmm4 146 punpcklqdq %xmm4,%xmm6 160 movd (DATA,%rax,4),%xmm4 [all …]
|
| H A D | chacha-ssse3-x86_64.S | 38 movdqa ROT8(%rip),%xmm4 58 pshufb %xmm4,%xmm3 91 pshufb %xmm4,%xmm3 139 movdqu 0x00(%rdx),%xmm4 140 pxor %xmm4,%xmm0 253 pshufd $0x00,%xmm5,%xmm4 286 paddd %xmm4,%xmm0 311 pxor %xmm8,%xmm4 312 movdqa %xmm4,%xmm0 314 psrld $20,%xmm4 [all …]
|
| H A D | sha512-avx-asm.S | 172 vmovdqa W_t(idx), %xmm4 # XMM4 = W[t-2] 176 vpsrlq $61, %xmm4, %xmm0 # XMM0 = W[t-2]>>61 181 vpsrlq $19, %xmm4, %xmm1 # XMM1 = W[t-2]>>19 190 vpsrlq $6, %xmm4, %xmm2 # XMM2 = W[t-2]>>6 200 vpsllq $(64-61), %xmm4, %xmm3 # XMM3 = W[t-2]<<3 217 vpsllq $(64-19), %xmm4, %xmm4 # XMM4 = W[t-2]<<25 220 vpxor %xmm4, %xmm0, %xmm0 # XMM0 = W[t-2]>>61 ^ W[t-2]>>19 ^
|
| H A D | polyval-pclmul-avx.S | 95 vpclmulqdq $0x11, (16*\i)(KEY_POWERS), %xmm0, %xmm4 98 vpxor %xmm4, HI, HI
|
| H A D | sha512-ssse3-asm.S | 218 movdqa %xmm5, %xmm4 # XMM4 = W[t-15] 225 psllq $(64-1)-(64-8), %xmm4 # XMM4 = W[t-15] << 7 231 pxor %xmm5, %xmm4 # XMM4 = (W[t-15]<<7)^W[t-15] 237 psllq $(64-8), %xmm4 # XMM4 = ((W[t-15]<<7)^W[t-15])<<56 246 pxor %xmm4, %xmm3 # XMM3 = s0(W[t-15])
|
| H A D | sha1-ni-asm.S | 66 #define MSG1 %xmm4
|
| H A D | sha256-ni-asm.S | 68 #define MSG1 %xmm4 206 #define STATE1_B %xmm4
|
| H A D | chacha-avx512vl-x86_64.S | 385 vextracti128 $1,%ymm10,%xmm4 409 vmovdqa %xmm4,%xmm10
|
| H A D | sha1-ssse3-and-avx.S | 53 #define W12 %xmm4
|
| H A D | chacha-avx2-x86_64.S | 451 vextracti128 $1,%ymm10,%xmm4 475 vmovdqa %xmm4,%xmm10
|
| H A D | sha256-avx-asm.S | 80 X0 = %xmm4
|
| H A D | sha256-ssse3-asm.S | 74 X0 = %xmm4
|
| H A D | sha256-avx2-asm.S | 73 XWORD0 = %xmm4
|
| /linux/arch/x86/crypto/ |
| H A D | aria-aesni-avx-asm_64.S | 889 inpack16_post(%xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7, 893 %xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7, 895 aria_fe(%xmm1, %xmm0, %xmm3, %xmm2, %xmm4, %xmm5, %xmm6, %xmm7, 899 %xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7, 901 aria_fe(%xmm1, %xmm0, %xmm3, %xmm2, %xmm4, %xmm5, %xmm6, %xmm7, 905 %xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7, 907 aria_fe(%xmm1, %xmm0, %xmm3, %xmm2, %xmm4, %xmm5, %xmm6, %xmm7, 911 %xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7, 913 aria_fe(%xmm1, %xmm0, %xmm3, %xmm2, %xmm4, %xmm5, %xmm6, %xmm7, 917 %xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7, [all …]
|
| H A D | cast6-avx-x86_64-asm_64.S | 42 #define RA2 %xmm4
|
| H A D | twofish-avx-x86_64-asm_64.S | 42 #define RA2 %xmm4
|
| H A D | sm4-aesni-avx2-asm_64.S | 53 #define RTMP1x %xmm4
|
| H A D | sm4-aesni-avx-asm_64.S | 26 #define RTMP1 %xmm4
|
| H A D | aria-aesni-avx2-asm_64.S | 21 #define ymm4_x xmm4 1036 inc_le128(%xmm7, %xmm0, %xmm4);
|
| H A D | serpent-sse2-i586-asm_32.S | 31 #define RE %xmm4
|
| H A D | serpent-sse2-x86_64-asm_64.S | 26 #define RE1 %xmm4
|
| /linux/lib/crc/x86/ |
| H A D | crc-pclmul-template.S | 391 _fold_vec_final 16, %xmm0, %xmm1, CONSTS_XMM, BSWAP_MASK_XMM, %xmm4, %xmm5 427 %xmm0, %xmm0, unaligned_mem_tmp=%xmm4 433 movdqa %xmm0, %xmm4 436 movdqa %xmm4, %xmm0 444 _fold_vec %xmm0, %xmm1, CONSTS_XMM, %xmm4
|
| /linux/include/hyperv/ |
| H A D | hvhdk.h | 89 struct hv_u128 xmm4; member
|