Home
last modified time | relevance | path

Searched refs:xmm0 (Results 1 – 25 of 27) sorted by relevance

12

/linux/lib/crypto/x86/
H A Dchacha-ssse3-x86_64.S43 paddd %xmm1,%xmm0
44 pxor %xmm0,%xmm3
56 paddd %xmm1,%xmm0
57 pxor %xmm0,%xmm3
76 paddd %xmm1,%xmm0
77 pxor %xmm0,%xmm3
89 paddd %xmm1,%xmm0
90 pxor %xmm0,%xmm3
123 movdqu 0x00(%rdi),%xmm0
127 movdqa %xmm0,%xmm8
[all …]
H A Dblake2s-core.S69 movdqu (CTX),%xmm0 // Load h[0..3]
81 movdqa %xmm0,%xmm10 // Save h[0..3] and let v[0..3] = h[0..3]
102 paddd %xmm4,%xmm0
103 paddd %xmm1,%xmm0
104 pxor %xmm0,%xmm3
123 paddd %xmm5,%xmm0
124 paddd %xmm1,%xmm0
125 pxor %xmm0,%xmm3
133 pshufd $0x93,%xmm0,%xmm0
147 paddd %xmm6,%xmm0
[all …]
H A Dpolyval-pclmul-avx.S88 movups (16*\i)(MSG), %xmm0
90 pxor SUM, %xmm0
92 vpclmulqdq $0x01, (16*\i)(KEY_POWERS), %xmm0, %xmm2
93 vpclmulqdq $0x00, (16*\i)(KEY_POWERS), %xmm0, %xmm1
94 vpclmulqdq $0x10, (16*\i)(KEY_POWERS), %xmm0, %xmm3
95 vpclmulqdq $0x11, (16*\i)(KEY_POWERS), %xmm0, %xmm4
108 vpclmulqdq $0x01, %xmm0, %xmm1, MI
109 vpclmulqdq $0x10, %xmm0, %xmm1, %xmm2
110 vpclmulqdq $0x00, %xmm0, %xmm1, LO
111 vpclmulqdq $0x11, %xmm0, %xmm1, HI
[all …]
H A Dsha512-avx-asm.S176 vpsrlq $61, %xmm4, %xmm0 # XMM0 = W[t-2]>>61
184 vpxor %xmm1, %xmm0, %xmm0 # XMM0 = W[t-2]>>61 ^ W[t-2]>>19
220 vpxor %xmm4, %xmm0, %xmm0 # XMM0 = W[t-2]>>61 ^ W[t-2]>>19 ^
223 vpxor %xmm2, %xmm0, %xmm0 # XMM0 = s1(W[t-2])
227 vpaddq W_t(idx), %xmm0, %xmm0 # XMM0 = s1(W[t-2]) + W[t-16]
239 vpaddq %xmm6, %xmm0, %xmm0 # XMM0 = s1(W[t-2]) + W[t-16] + s0(W[t-15])
241 vpaddq %xmm1, %xmm0, %xmm0 # XMM0 = W[t] = s1(W[t-2]) + W[t-7] +
248 vmovdqa %xmm0, W_t(idx) # Store W[t]
249 vpaddq K_t(idx), %xmm0, %xmm0 # Compute W[t]+K[t]
250 vmovdqa %xmm0, WK_2(idx) # Store W[t]+K[t] for next rounds
[all …]
H A Dsha512-ssse3-asm.S174 movdqa %xmm2, %xmm0 # XMM0 = W[t-2]
185 psrlq $61-19, %xmm0 # XMM0 = W[t-2] >> 42
191 pxor %xmm2, %xmm0 # XMM0 = (W[t-2] >> 42) ^ W[t-2]
197 psrlq $(19-6), %xmm0 # XMM0 = ((W[t-2]>>42)^W[t-2])>>13
203 pxor %xmm2, %xmm0 # XMM0 = (((W[t-2]>>42)^W[t-2])>>13)^W[t-2]
209 psrlq $6, %xmm0 # XMM0 = ((((W[t-2]>>42)^W[t-2])>>13)^W[t-2])>>6
240 pxor %xmm1, %xmm0 # XMM0 = s1(W[t-2])
248 paddq %xmm3, %xmm0 # XMM0 = s1(W[t-2]) + s0(W[t-15])
251 paddq W_t(idx), %xmm0 # XMM0 = s1(W[t-2]) + s0(W[t-15]) + W[t-16]
253 paddq %xmm1, %xmm0 # XMM0 = s1(W[t-2]) + W[t-7] + s0(W[t-15]) + W[t-16]
[all …]
H A Dsha1-ni-asm.S62 #define ABCD %xmm0
H A Dsha256-ni-asm.S64 #define MSG %xmm0 /* sha256rnds2 implicit operand */
202 #define MSG %xmm0 // sha256rnds2 implicit operand
H A Dchacha-avx512vl-x86_64.S119 vextracti128 $1,%ymm7,%xmm0
143 vmovdqa %xmm0,%xmm7
331 vextracti128 $1,%ymm10,%xmm0
355 vmovdqa %xmm0,%xmm10
H A Dchacha-avx2-x86_64.S146 vextracti128 $1,%ymm7,%xmm0
170 vmovdqa %xmm0,%xmm7
397 vextracti128 $1,%ymm10,%xmm0
421 vmovdqa %xmm0,%xmm10
H A Dpoly1305-x86_64-cryptogams.pl3877 movdqu ($inp,$otp),%xmm0
3878 pxor ($otp),%xmm0
3879 movdqu %xmm0,($out,$otp)
3880 movdqa %xmm0,($otp)
3924 movdqu ($inp,$otp),%xmm0
3926 pxor %xmm0,%xmm1
3928 movdqa %xmm0,($otp)
H A Dsha1-avx2-asm.S125 #define W_TMP %xmm0
H A Dsha1-ssse3-and-avx.S47 #define W_TMP1 %xmm0
H A Dsha256-avx-asm.S85 XTMP0 = %xmm0
H A Dsha256-ssse3-asm.S79 XTMP0 = %xmm0
/linux/lib/crc/x86/
H A Dcrc-pclmul-template.S286 _cond_vex movd, CRC, %xmm0
288 _cond_vex movq, CRC, %xmm0
291 _cond_vex pslldq, $(128-\n)/8, %xmm0, %xmm0
310 _prepare_v0 16, %xmm0, %xmm1, BSWAP_MASK_XMM
391 _fold_vec_final 16, %xmm0, %xmm1, CONSTS_XMM, BSWAP_MASK_XMM, %xmm4, %xmm5
419 _cond_vex pshufb, %xmm3, %xmm0, %xmm1
427 %xmm0, %xmm0, unaligned_mem_tmp=%xmm4
433 movdqa %xmm0, %xmm4
434 movdqa %xmm3, %xmm0
436 movdqa %xmm4, %xmm0
[all …]
/linux/arch/x86/crypto/
H A Daes-xts-avx-x86_64.S311 _next_tweak TWEAK0, %xmm0, TWEAK1
312 _next_tweak TWEAK1, %xmm0, TWEAK2
313 _next_tweak TWEAK2, %xmm0, TWEAK3
319 _next_tweak TWEAK0_XMM, %xmm0, %xmm1
758 vmovdqu (SRC), %xmm0
759 _aes_crypt \enc, _XMM, TWEAK0_XMM, %xmm0, tmp=%xmm1
760 vmovdqu %xmm0, (DST)
761 _next_tweak TWEAK0_XMM, %xmm0, TWEAK0_XMM
780 vmovdqu (DST), %xmm0
785 _next_tweak TWEAK0_XMM, %xmm0, TWEAK1_XMM
[all …]
H A Daria-aesni-avx-asm_64.S889 inpack16_post(%xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7,
893 %xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7,
895 aria_fe(%xmm1, %xmm0, %xmm3, %xmm2, %xmm4, %xmm5, %xmm6, %xmm7,
899 %xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7,
901 aria_fe(%xmm1, %xmm0, %xmm3, %xmm2, %xmm4, %xmm5, %xmm6, %xmm7,
905 %xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7,
907 aria_fe(%xmm1, %xmm0, %xmm3, %xmm2, %xmm4, %xmm5, %xmm6, %xmm7,
911 %xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7,
913 aria_fe(%xmm1, %xmm0, %xmm3, %xmm2, %xmm4, %xmm5, %xmm6, %xmm7,
917 %xmm0, %xmm1, %xmm2, %xmm3, %xmm4, %xmm5, %xmm6, %xmm7,
[all …]
H A Dsm4-aesni-avx2-asm_64.S45 #define RX0x %xmm0
49 #define RNOTx %xmm0
H A Dsm4-aesni-avx-asm_64.S22 #define RX0 %xmm0
41 #define RNOT %xmm0
H A Dcast6-avx-x86_64-asm_64.S37 #define RA1 %xmm0
H A Dtwofish-avx-x86_64-asm_64.S37 #define RA1 %xmm0
H A Daes-ctr-avx-x86_64.S303 .set AESDATA0_XMM, %xmm0
H A Daria-aesni-avx2-asm_64.S17 #define ymm0_x xmm0
1036 inc_le128(%xmm7, %xmm0, %xmm4);
/linux/Documentation/trace/
H A Dtracepoint-analysis.rst321 12.40 : 34eee: 66 0f 7f 80 40 ff ff movdqa %xmm0,-0xc0(%eax)
323 12.40 : 34ef6: 66 0f 7f 80 50 ff ff movdqa %xmm0,-0xb0(%eax)
325 12.39 : 34efe: 66 0f 7f 80 60 ff ff movdqa %xmm0,-0xa0(%eax)
327 12.67 : 34f06: 66 0f 7f 80 70 ff ff movdqa %xmm0,-0x90(%eax)
329 12.58 : 34f0e: 66 0f 7f 40 80 movdqa %xmm0,-0x80(%eax)
330 12.31 : 34f13: 66 0f 7f 40 90 movdqa %xmm0,-0x70(%eax)
331 12.40 : 34f18: 66 0f 7f 40 a0 movdqa %xmm0,-0x60(%eax)
332 12.31 : 34f1d: 66 0f 7f 40 b0 movdqa %xmm0,-0x50(%eax)
/linux/include/hyperv/
H A Dhvhdk.h85 struct hv_u128 xmm0; member

12