/freebsd/contrib/bearssl/src/symcipher/ |
H A D | aes_x86ni.c | 62 expand_step192(__m128i *t1, __m128i *t2, __m128i *t3) in expand_step192() argument 75 t4 = _mm_slli_si128(*t3, 0x4); in expand_step192() 76 *t3 = _mm_xor_si128(*t3, t4); in expand_step192() 77 *t3 = _mm_xor_si128(*t3, *t2); in expand_step192() 98 expand_step256_2(__m128i *t1, __m128i *t3) in expand_step256_2() argument 104 t4 = _mm_slli_si128(*t3, 0x4); in expand_step256_2() 105 *t3 = _mm_xor_si128(*t3, t4); in expand_step256_2() 107 *t3 = _mm_xor_si128(*t3, t4); in expand_step256_2() 109 *t3 = _mm_xor_si128(*t3, t4); in expand_step256_2() 110 *t3 = _mm_xor_si128(*t3, t2); in expand_step256_2() [all …]
|
/freebsd/contrib/bearssl/src/ec/ |
H A D | ec_c25519_m64.c | 102 uint64_t t0, t1, t2, t3, cc; in f255_add() local 112 t3 = (uint64_t)z & MASK63; in f255_add() 129 d[3] = t3 + (uint64_t)(z >> 64); in f255_add() 133 uint64_t t0, t1, t2, t3, cc; in f255_add() 139 k = _addcarry_u64(k, a[3], b[3], &t3); in f255_add() 140 cc = (k << 1) + (t3 >> 63); in f255_add() 141 t3 &= MASK63; in f255_add() 154 (void)_addcarry_u64(k, t3, 0, &d[3]); in f255_add() 174 uint64_t t0, t1, t2, t3, t4, cc; in f255_sub() local 190 t3 = (uint64_t)z; in f255_sub() [all …]
|
H A D | ec_p256_m64.c | 263 uint64_t x, f, t0, t1, t2, t3, t4; in f256_montymul() local 289 t3 = (uint64_t)(z >> 64); in f256_montymul() 293 z = (unsigned __int128)t3 + (z >> 64) + (ff >> 64); in f256_montymul() 294 t3 = (uint64_t)z; in f256_montymul() 310 z = (unsigned __int128)b[3] * x + t3 + (z >> 64); in f256_montymul() 313 t3 = (uint64_t)z; in f256_montymul() 327 z = (unsigned __int128)t3 + (z >> 64) + (ff >> 64); in f256_montymul() 328 t3 = (uint64_t)z; in f256_montymul() 356 t3 = t3 - (uint64_t)(z >> 127) - t4 + (t4 << 32); in f256_montymul() 361 d[3] = t3; in f256_montymul() [all …]
|
/freebsd/sys/contrib/libsodium/src/libsodium/crypto_onetimeauth/poly1305/sse2/ |
H A D | poly1305_sse2.c | 207 xmmi T0, T1, T2, T3, T4, T5, T6, T7, T8; in poly1305_blocks() local 252 T3 = _mm_cvtsi32_si128(st->R[4]); in poly1305_blocks() 257 R24 = _mm_unpacklo_epi64(T1, T3); in poly1305_blocks() 324 T3 = H4; in poly1305_blocks() 325 T3 = _mm_mul_epu32(T3, T15); in poly1305_blocks() 346 T3 = _mm_add_epi64(T3, v31); in poly1305_blocks() 365 T3 = _mm_add_epi64(T3, v32); in poly1305_blocks() 381 T3 = _mm_add_epi64(T3, v33); in poly1305_blocks() 391 T3 = _mm_add_epi64(T3, v34); in poly1305_blocks() 413 T3 = _mm_add_epi64(T3, M8); in poly1305_blocks() [all …]
|
/freebsd/crypto/openssl/crypto/bn/asm/ |
H A D | ppc64-mont.pl | 142 $t3="r27"; 283 ld $t3,0($bp) ; bp[0] 286 mulld $t7,$a0,$t3 ; ap[0]*bp[0] 288 extrdi $t0,$t3,16,48 289 extrdi $t1,$t3,16,32 290 extrdi $t2,$t3,16,16 291 extrdi $t3,$t3,16,0 295 std $t3,`$FRAME+24`($sp) 311 lwz $t3,`8^$LITTLE_ENDIAN`($ap) 323 lwz $t3,4($bp) [all …]
|
H A D | ppc-mont.pl | 358 $t0,$t1,$t2,$t3, 459 $UMULL $t3,$a3,$bi 466 adde $acc3,$acc3,$t3 469 $UMULH $t3,$a3,$bi 478 adde $acc4,$acc4,$t3 # can't overflow 479 $UMULL $t3,$m3,$mi 496 adde $acc2,$acc3,$t3 497 $UMULH $t3,$m3,$mi 503 adde $acc3,$acc3,$t3 529 $UMULL $t3,$a3,$bi [all …]
|
H A D | armv8-mont.pl | 667 my ($t0,$t1,$t2,$t3)=map("x$_",(14..17)); 764 mul $t3,$a4,$a0 771 adcs $acc4,$acc4,$t3 772 umulh $t3,$a1,$a0 // hi(a[1..7]*a[0]) 781 adds $acc2,$acc2,$t3 // t[2]+lo(a[1]*a[0]) 782 umulh $t3,$a5,$a0 789 adcs $acc6,$acc6,$t3 790 mul $t3,$a3,$a1 798 adcs $acc4,$acc4,$t3 799 mul $t3,$a7,$a1 [all …]
|
/freebsd/crypto/openssl/crypto/sha/asm/ |
H A D | sha512-x86_64.pl | 781 my ($t0,$t1,$t2,$t3, $t4,$t5) = map("%xmm$_",(4..9)); 838 movdqa $TABLE+`$SZ*2*$rounds`(%rip),$t3 842 pshufb $t3,@X[0] 845 pshufb $t3,@X[1] 848 pshufb $t3,@X[2] 851 pshufb $t3,@X[3] 852 movdqa 0x60($Tbl),$t3 855 paddd @X[3],$t3 862 movdqa $t3,0x30(%rsp) 873 '&movdqa ($t3,@X[3])', [all …]
|
H A D | sha1-alpha.pl | 37 $t3="t12"; 60 sll @X[$i],8,$t3 67 zapnot $t3,0x44,$t3 75 or @X[$i],$t3,@X[$i] 77 bic $d,$b,$t3 81 or $t2,$t3,$t2 85 srl $b,32,$t3 90 or $t3,$b,$b 100 bic $d,$b,$t3 103 or $t2,$t3,$t2 [all …]
|
H A D | sha512-armv4.pl | 90 $t3="r12"; 118 ldr $t3,[sp,#$Hoff+4] @ h.hi 133 adc $Thi,$Thi,$t3 @ T += h 134 ldr $t3,[sp,#$Goff+4] @ g.hi 138 eor $t1,$t1,$t3 146 eor $t1,$t1,$t3 @ Ch(e,f,g) 147 ldr $t3,[$Ktbl,#$hi] @ K[i].hi 155 adc $Thi,$Thi,$t3 @ T += K[i] 161 ldr $t3,[sp,#$Coff+0] @ c.lo 188 and $Alo,$Alo,$t3 [all …]
|
H A D | sha256-armv4.pl | 67 $T1="r3"; $t3="r3"; 144 and $t3,$t3,$t2 @ (b^c)&=(a^b) 146 eor $t3,$t3,$b @ Maj(a,b,c) 148 @ add $h,$h,$t3 @ h+=Maj(a,b,c) 150 ($t2,$t3)=($t3,$t2); 259 eor $t3,$B,$C @ magic 269 ldreq $t3,[sp,#16*4] @ pull ctx 273 ldr $t0,[$t3,#0] 274 ldr $t1,[$t3,#4] 275 ldr $t2,[$t3,#8] [all …]
|
/freebsd/crypto/openssl/crypto/aes/asm/ |
H A D | aes-armv4.pl | 65 $t3="r6"; 220 ldrb $t3,[$rounds,#0] 225 orr $s0,$s0,$t3,lsl#24 227 ldrb $t3,[$rounds,#4] 232 orr $s1,$s1,$t3,lsl#24 234 ldrb $t3,[$rounds,#8] 239 orr $s2,$s2,$t3,lsl#24 241 ldrb $t3,[$rounds,#12] 244 orr $s3,$s3,$t3,lsl#24 274 mov $t3,$s0,lsr#8 [all …]
|
H A D | aes-parisc.pl | 65 ($t0,$t1,$t2,$t3) = ("%r5","%r6","%r7","%r8"); 204 ldw 12($key),$t3 211 xor $t3,$s3,$s3 213 ldw 28($key),$t3 277 xor $acc13,$t3,$t3 279 xor $acc14,$t3,$t3 281 xor $acc15,$t3,$t3 285 _srm $t3,0,$acc3 291 _srm $t3,8,$acc6 297 _srm $t3,16,$acc9 [all …]
|
/freebsd/crypto/openssl/crypto/ec/asm/ |
H A D | ecp_nistp521-ppc64.pl | 150 my ($t1, $t2, $t3, $t4) = ("v33", "v34", "v44", "v54"); 188 xxpermdi $t3,$in1[2],$in1[3],0b00 190 vmsumudm $out[3],$t3,$t4,$out[3] 195 vmsumudm $out[4],$t3,$t4,$out[4] 201 vmsumudm $out[5],$t3,$t4,$out[5] 206 vmsumudm $out[6],$t3,$t4,$out[6] 211 vmsumudm $out[7],$t3,$t4,$out[7] 216 vmsumudm $out[8],$t3,$t4,$out[8] 228 xxpermdi $t3,$in1[6],$in1[7],0b00 230 vmsumudm $out[7],$t3,$t4,$out[7] [all …]
|
H A D | ecp_nistz256-armv8.pl | 49 my ($rp,$ap,$bp,$bi,$a0,$a1,$a2,$a3,$t0,$t1,$t2,$t3,$poly1,$poly3, 231 ldp $t2,$t3,[$bp,#16] 279 mov $t3,$acc3 304 mov $t3,$acc3 315 mov $t3,$a3 385 umulh $t3,$a3,$bi 393 adc $acc4,xzr,$t3 423 sbc $t3,$acc0,$t1 430 adcs $acc3,$acc4,$t3 431 mul $t3,$a3,$bi // lo(a[3]*b[i]) [all …]
|
H A D | ecp_nistz256-ppc64.pl | 46 $acc4,$acc5,$a0,$a1,$a2,$a3,$t0,$t1,$t2,$t3) = 218 ld $t3, 24($bp) 293 mr $t3,$acc3 336 mr $t3,$acc3 349 ld $t3,88($sp) 452 mulhdu $t3,$a3,$bi 460 addze $acc4,$t3 491 subfe $t3,$t1,$acc0 495 adde $acc3,$acc4,$t3 501 mulld $t3,$a3,$bi # lo(a[3]*b[i]) [all …]
|
H A D | ecp_nistz256-sparcv9.pl | 101 my ($t0,$t1,$t2,$t3,$t4,$t5,$t6,$t7)=(map("%o$_",(0..5)),"%g4","%g5"); 180 # CPU is 64-bit. [Addition with 64-bit carry was introduced in T3, see 192 ld [$ap+12],$t3 200 mulx $t3,$bi,$t3 208 srlx $t3,32,@acc[4] 221 addccc @acc[3],$t3,@acc[3] 223 ld [$ap+12],$t3 285 mulx $t3,$bi,$t3 295 add @acc[3],$t3,$t3 298 srlx $t3,32,@acc[4] [all …]
|
/freebsd/crypto/openssl/crypto/md5/asm/ |
H A D | md5-sparcv9.pl | 43 ($t1,$t2,$t3,$saved_asi)=map("%l$_",(4..7)); 83 sll $a,$rot,$t3 86 add $b,$t3,$t3 88 add $t3,$a,$a 99 sll $a,$rot,$t3 102 add $b,$t3,$t3 104 add $t3,$a,$a 121 sll $a,$rot,$t3 124 add $b,$t3,$t3 126 add $t3,$a,$a [all …]
|
/freebsd/crypto/openssh/ |
H A D | rijndael.c | 765 u32 s0, s1, s2, s3, t0, t1, t2, t3; in rijndaelEncrypt() local 783 t3 = Te0[s3 >> 24] ^ Te1[(s0 >> 16) & 0xff] ^ Te2[(s1 >> 8) & 0xff] ^ Te3[s2 & 0xff] ^ rk[ 7]; in rijndaelEncrypt() 785 s0 = Te0[t0 >> 24] ^ Te1[(t1 >> 16) & 0xff] ^ Te2[(t2 >> 8) & 0xff] ^ Te3[t3 & 0xff] ^ rk[ 8]; in rijndaelEncrypt() 786 s1 = Te0[t1 >> 24] ^ Te1[(t2 >> 16) & 0xff] ^ Te2[(t3 >> 8) & 0xff] ^ Te3[t0 & 0xff] ^ rk[ 9]; in rijndaelEncrypt() 787 s2 = Te0[t2 >> 24] ^ Te1[(t3 >> 16) & 0xff] ^ Te2[(t0 >> 8) & 0xff] ^ Te3[t1 & 0xff] ^ rk[10]; in rijndaelEncrypt() 788 s3 = Te0[t3 >> 24] ^ Te1[(t0 >> 16) & 0xff] ^ Te2[(t1 >> 8) & 0xff] ^ Te3[t2 & 0xff] ^ rk[11]; in rijndaelEncrypt() 793 t3 = Te0[s3 >> 24] ^ Te1[(s0 >> 16) & 0xff] ^ Te2[(s1 >> 8) & 0xff] ^ Te3[s2 & 0xff] ^ rk[15]; in rijndaelEncrypt() 795 s0 = Te0[t0 >> 24] ^ Te1[(t1 >> 16) & 0xff] ^ Te2[(t2 >> 8) & 0xff] ^ Te3[t3 & 0xff] ^ rk[16]; in rijndaelEncrypt() 796 s1 = Te0[t1 >> 24] ^ Te1[(t2 >> 16) & 0xff] ^ Te2[(t3 >> 8) & 0xff] ^ Te3[t0 & 0xff] ^ rk[17]; in rijndaelEncrypt() 797 s2 = Te0[t2 >> 24] ^ Te1[(t3 >> 16) & 0xff] ^ Te2[(t0 >> 8) & 0xff] ^ Te3[t1 & 0xff] ^ rk[18]; in rijndaelEncrypt() [all …]
|
/freebsd/sys/crypto/rijndael/ |
H A D | rijndael-alg-fst.c | 862 u32 s0, s1, s2, s3, t0, t1, t2, t3; in rijndaelEncrypt() local 880 t3 = Te0[s3 >> 24] ^ Te1[(s0 >> 16) & 0xff] ^ Te2[(s1 >> 8) & 0xff] ^ Te3[s2 & 0xff] ^ rk[ 7]; in rijndaelEncrypt() 882 s0 = Te0[t0 >> 24] ^ Te1[(t1 >> 16) & 0xff] ^ Te2[(t2 >> 8) & 0xff] ^ Te3[t3 & 0xff] ^ rk[ 8]; in rijndaelEncrypt() 883 s1 = Te0[t1 >> 24] ^ Te1[(t2 >> 16) & 0xff] ^ Te2[(t3 >> 8) & 0xff] ^ Te3[t0 & 0xff] ^ rk[ 9]; in rijndaelEncrypt() 884 s2 = Te0[t2 >> 24] ^ Te1[(t3 >> 16) & 0xff] ^ Te2[(t0 >> 8) & 0xff] ^ Te3[t1 & 0xff] ^ rk[10]; in rijndaelEncrypt() 885 s3 = Te0[t3 >> 24] ^ Te1[(t0 >> 16) & 0xff] ^ Te2[(t1 >> 8) & 0xff] ^ Te3[t2 & 0xff] ^ rk[11]; in rijndaelEncrypt() 890 t3 = Te0[s3 >> 24] ^ Te1[(s0 >> 16) & 0xff] ^ Te2[(s1 >> 8) & 0xff] ^ Te3[s2 & 0xff] ^ rk[15]; in rijndaelEncrypt() 892 s0 = Te0[t0 >> 24] ^ Te1[(t1 >> 16) & 0xff] ^ Te2[(t2 >> 8) & 0xff] ^ Te3[t3 & 0xff] ^ rk[16]; in rijndaelEncrypt() 893 s1 = Te0[t1 >> 24] ^ Te1[(t2 >> 16) & 0xff] ^ Te2[(t3 >> 8) & 0xff] ^ Te3[t0 & 0xff] ^ rk[17]; in rijndaelEncrypt() 894 s2 = Te0[t2 >> 24] ^ Te1[(t3 >> 16) & 0xff] ^ Te2[(t0 >> 8) & 0xff] ^ Te3[t1 & 0xff] ^ rk[18]; in rijndaelEncrypt() [all …]
|
/freebsd/crypto/openssl/crypto/poly1305/asm/ |
H A D | poly1305-x86_64.pl | 344 my ($H0,$H1,$H2,$H3,$H4, $T0,$T1,$T2,$T3,$T4, $D0,$D1,$D2,$D3,$D4, $MASK) = 816 vpsrldq \$6,$T1,$T3 819 vpunpcklqdq $T3,$T2,$T3 # 2:3 824 vpsrlq \$4,$T3,$T2 826 vpsrlq \$30,$T3,$T3 828 vpand $MASK,$T3,$T3 # 3 909 vpmuludq $T3,$D4,$D3 # d3 = h3*r0 915 vpmuludq $T3,$H2,$H1 # h3*r1 937 vpmuludq $T3,$H4,$H4 # h3*s2 948 vpmuludq $T3,$H3,$H1 # h3*s3 [all …]
|
/freebsd/contrib/bearssl/src/hash/ |
H A D | ghash_pclmul.c | 274 __m128i t0, t1, t2, t3; in br_ghash_pclmul() local 285 t3 = pclmulqdq00(h1w, h2w); in br_ghash_pclmul() 287 _mm_xor_si128(t1, t3)); in br_ghash_pclmul() 290 t2 = _mm_xor_si128(t2, _mm_shuffle_epi32(t3, 0x0E)); in br_ghash_pclmul() 291 SL_256(t0, t1, t2, t3); in br_ghash_pclmul() 292 REDUCE_F128(t0, t1, t2, t3); in br_ghash_pclmul() 327 t3 = _mm_xor_si128( in br_ghash_pclmul() 341 t2 = _mm_xor_si128(t2, _mm_xor_si128(t1, t3)); in br_ghash_pclmul() 344 t2 = _mm_xor_si128(t2, _mm_shuffle_epi32(t3, 0x0E)); in br_ghash_pclmul() 345 SL_256(t0, t1, t2, t3); in br_ghash_pclmul() [all …]
|
/freebsd/contrib/bearssl/src/rsa/ |
H A D | rsa_i15_priv.c | 41 uint16_t *mp, *mq, *s1, *s2, *t1, *t2, *t3; in br_rsa_i15_private() local 125 t3 = mq + 4 * fwlen; in br_rsa_i15_private() 126 br_i15_encode(t3, xlen, t2); in br_rsa_i15_private() 133 wn = ((unsigned char *)t3)[u]; in br_rsa_i15_private() 188 * We need mq, s2 and t2. We use the t3 buffer as destination. in br_rsa_i15_private() 190 * reuse them for t3. Moreover, the first step of the computation in br_rsa_i15_private() 191 * is to copy s2 into t3, after which s2 is not needed. Right in br_rsa_i15_private() 193 * Therefore, we have ample room for t3 by simply using s2. in br_rsa_i15_private() 195 t3 = s2; in br_rsa_i15_private() 196 br_i15_mulacc(t3, mq, t2); in br_rsa_i15_private() [all …]
|
H A D | rsa_i31_priv.c | 41 uint32_t *mp, *mq, *s1, *s2, *t1, *t2, *t3; in br_rsa_i31_private() local 119 t3 = mq + 4 * fwlen; in br_rsa_i31_private() 120 br_i31_encode(t3, xlen, t2); in br_rsa_i31_private() 127 wn = ((unsigned char *)t3)[u]; in br_rsa_i31_private() 182 * We need mq, s2 and t2. We use the t3 buffer as destination. in br_rsa_i31_private() 184 * reuse them for t3. Moreover, the first step of the computation in br_rsa_i31_private() 185 * is to copy s2 into t3, after which s2 is not needed. Right in br_rsa_i31_private() 187 * Therefore, we have ample room for t3 by simply using s2. in br_rsa_i31_private() 189 t3 = s2; in br_rsa_i31_private() 190 br_i31_mulacc(t3, mq, t2); in br_rsa_i31_private() [all …]
|
/freebsd/contrib/llvm-project/llvm/lib/Support/BLAKE3/ |
H A D | blake3_sse2.c | 103 __m128i t0, t1, t2, t3, tt; in compress_pre() local 115 t3 = _mm_shuffle_ps2(m2, m3, _MM_SHUFFLE(3, 1, 3, 1)); // 15 13 11 9 in compress_pre() 116 t3 = _mm_shuffle_epi32(t3, _MM_SHUFFLE(2, 1, 0, 3)); // 13 11 9 15 in compress_pre() 117 g2(&rows[0], &rows[1], &rows[2], &rows[3], t3); in compress_pre() 122 m3 = t3; in compress_pre() 138 t3 = _mm_unpackhi_epi32(m1, m3); in compress_pre() 139 tt = _mm_unpacklo_epi32(m2, t3); in compress_pre() 140 t3 = _mm_shuffle_epi32(tt, _MM_SHUFFLE(0, 1, 3, 2)); in compress_pre() 141 g2(&rows[0], &rows[1], &rows[2], &rows[3], t3); in compress_pre() 146 m3 = t3; in compress_pre() [all …]
|