| /freebsd/contrib/bearssl/src/symcipher/ |
| H A D | aes_x86ni.c | 62 expand_step192(__m128i *t1, __m128i *t2, __m128i *t3) in expand_step192() argument 75 t4 = _mm_slli_si128(*t3, 0x4); in expand_step192() 76 *t3 = _mm_xor_si128(*t3, t4); in expand_step192() 77 *t3 = _mm_xor_si128(*t3, *t2); in expand_step192() 98 expand_step256_2(__m128i *t1, __m128i *t3) in expand_step256_2() argument 104 t4 = _mm_slli_si128(*t3, 0x4); in expand_step256_2() 105 *t3 = _mm_xor_si128(*t3, t4); in expand_step256_2() 107 *t3 = _mm_xor_si128(*t3, t4); in expand_step256_2() 109 *t3 = _mm_xor_si128(*t3, t4); in expand_step256_2() 110 *t3 = _mm_xor_si128(*t3, t2); in expand_step256_2() [all …]
|
| /freebsd/lib/libc/riscv/string/ |
| H A D | memcpy.S | 36 slli t3, t1, 3 37 sub t2, t2, t3 39 lb t3, 6(a1) 40 sb t3, 6(a3) 41 lb t3, 5(a1) 42 sb t3, 5(a3) 43 lb t3, 4(a1) 44 sb t3, 4(a3) 45 lb t3, 3(a1) 46 sb t3, 3(a3) [all …]
|
| H A D | memchr.S | 111 ld t3, 8(a0) 114 xor t3, t3, a1 117 not t4, t3 119 sub t3, t3, t0 123 and t3, t3, t4 130 mv a2, t3
|
| /freebsd/contrib/bearssl/src/ec/ |
| H A D | ec_c25519_m64.c | 102 uint64_t t0, t1, t2, t3, cc; in f255_add() local 112 t3 = (uint64_t)z & MASK63; in f255_add() 129 d[3] = t3 + (uint64_t)(z >> 64); in f255_add() 133 uint64_t t0, t1, t2, t3, cc; in f255_add() 139 k = _addcarry_u64(k, a[3], b[3], &t3); in f255_add() 140 cc = (k << 1) + (t3 >> 63); in f255_add() 141 t3 &= MASK63; in f255_add() 154 (void)_addcarry_u64(k, t3, 0, &d[3]); in f255_add() 174 uint64_t t0, t1, t2, t3, t4, cc; in f255_sub() local 190 t3 = (uint64_t)z; in f255_sub() [all …]
|
| H A D | ec_p256_m64.c | 263 uint64_t x, f, t0, t1, t2, t3, t4; in f256_montymul() local 289 t3 = (uint64_t)(z >> 64); in f256_montymul() 293 z = (unsigned __int128)t3 + (z >> 64) + (ff >> 64); in f256_montymul() 294 t3 = (uint64_t)z; in f256_montymul() 310 z = (unsigned __int128)b[3] * x + t3 + (z >> 64); in f256_montymul() 313 t3 = (uint64_t)z; in f256_montymul() 327 z = (unsigned __int128)t3 + (z >> 64) + (ff >> 64); in f256_montymul() 328 t3 = (uint64_t)z; in f256_montymul() 356 t3 = t3 - (uint64_t)(z >> 127) - t4 + (t4 << 32); in f256_montymul() 361 d[3] = t3; in f256_montymul() [all …]
|
| /freebsd/sys/contrib/libsodium/src/libsodium/crypto_onetimeauth/poly1305/sse2/ |
| H A D | poly1305_sse2.c | 207 xmmi T0, T1, T2, T3, T4, T5, T6, T7, T8; in poly1305_blocks() local 252 T3 = _mm_cvtsi32_si128(st->R[4]); in poly1305_blocks() 257 R24 = _mm_unpacklo_epi64(T1, T3); in poly1305_blocks() 324 T3 = H4; in poly1305_blocks() 325 T3 = _mm_mul_epu32(T3, T15); in poly1305_blocks() 346 T3 = _mm_add_epi64(T3, v31); in poly1305_blocks() 365 T3 = _mm_add_epi64(T3, v32); in poly1305_blocks() 381 T3 = _mm_add_epi64(T3, v33); in poly1305_blocks() 391 T3 = _mm_add_epi64(T3, v34); in poly1305_blocks() 413 T3 = _mm_add_epi64(T3, M8); in poly1305_blocks() [all …]
|
| /freebsd/crypto/openssl/crypto/aes/asm/ |
| H A D | aes-riscv32-zkn.pl | 193 my ($T0,$T1,$T2,$T3) = use_regs(13..16); 262 lw $T3,12($KEYP) 271 xor $Q3,$Q3,$T3 282 lw $T3,12($KEYP) 287 @{[aes32esmi4 $T3,$Q3,$Q0,$Q1,$Q2]} 288 # now T0~T3 hold the new state 297 @{[aes32esmi4 $Q0,$T0,$T1,$T2,$T3]} 298 @{[aes32esmi4 $Q1,$T1,$T2,$T3,$T0]} 299 @{[aes32esmi4 $Q2,$T2,$T3,$T0,$T1]} 300 @{[aes32esmi4 $Q3,$T3,$T0,$T1,$T2]} [all …]
|
| H A D | aes-armv4.pl | 65 $t3="r6"; 220 ldrb $t3,[$rounds,#0] 225 orr $s0,$s0,$t3,lsl#24 227 ldrb $t3,[$rounds,#4] 232 orr $s1,$s1,$t3,lsl#24 234 ldrb $t3,[$rounds,#8] 239 orr $s2,$s2,$t3,lsl#24 241 ldrb $t3,[$rounds,#12] 244 orr $s3,$s3,$t3,lsl#24 274 mov $t3,$s0,lsr#8 [all …]
|
| /freebsd/crypto/openssl/crypto/bn/asm/ |
| H A D | ppc64-mont.pl | 142 $t3="r27"; 283 ld $t3,0($bp) ; bp[0] 286 mulld $t7,$a0,$t3 ; ap[0]*bp[0] 288 extrdi $t0,$t3,16,48 289 extrdi $t1,$t3,16,32 290 extrdi $t2,$t3,16,16 291 extrdi $t3,$t3,16,0 295 std $t3,`$FRAME+24`($sp) 311 lwz $t3,`8^$LITTLE_ENDIAN`($ap) 323 lwz $t3,4($bp) [all …]
|
| H A D | ppc-mont.pl | 358 $t0,$t1,$t2,$t3, 459 $UMULL $t3,$a3,$bi 466 adde $acc3,$acc3,$t3 469 $UMULH $t3,$a3,$bi 478 adde $acc4,$acc4,$t3 # can't overflow 479 $UMULL $t3,$m3,$mi 496 adde $acc2,$acc3,$t3 497 $UMULH $t3,$m3,$mi 503 adde $acc3,$acc3,$t3 529 $UMULL $t3,$a3,$bi [all …]
|
| H A D | armv8-mont.pl | 669 my ($t0,$t1,$t2,$t3)=map("x$_",(14..17)); 766 mul $t3,$a4,$a0 773 adcs $acc4,$acc4,$t3 774 umulh $t3,$a1,$a0 // hi(a[1..7]*a[0]) 783 adds $acc2,$acc2,$t3 // t[2]+lo(a[1]*a[0]) 784 umulh $t3,$a5,$a0 791 adcs $acc6,$acc6,$t3 792 mul $t3,$a3,$a1 800 adcs $acc4,$acc4,$t3 801 mul $t3,$a7,$a1 [all …]
|
| /freebsd/crypto/openssl/crypto/sha/asm/ |
| H A D | sha512-x86_64.pl | 785 my ($t0,$t1,$t2,$t3, $t4,$t5) = map("%xmm$_",(4..9)); 842 movdqa $TABLE+`$SZ*2*$rounds`(%rip),$t3 846 pshufb $t3,@X[0] 849 pshufb $t3,@X[1] 852 pshufb $t3,@X[2] 855 pshufb $t3,@X[3] 856 movdqa 0x60($Tbl),$t3 859 paddd @X[3],$t3 866 movdqa $t3,0x30(%rsp) 877 '&movdqa ($t3,@X[3])', [all …]
|
| H A D | sha1-alpha.pl | 37 $t3="t12"; 60 sll @X[$i],8,$t3 67 zapnot $t3,0x44,$t3 75 or @X[$i],$t3,@X[$i] 77 bic $d,$b,$t3 81 or $t2,$t3,$t2 85 srl $b,32,$t3 90 or $t3,$b,$b 100 bic $d,$b,$t3 103 or $t2,$t3,$t2 [all …]
|
| H A D | sha512-armv4.pl | 90 $t3="r12"; 118 ldr $t3,[sp,#$Hoff+4] @ h.hi 133 adc $Thi,$Thi,$t3 @ T += h 134 ldr $t3,[sp,#$Goff+4] @ g.hi 138 eor $t1,$t1,$t3 146 eor $t1,$t1,$t3 @ Ch(e,f,g) 147 ldr $t3,[$Ktbl,#$hi] @ K[i].hi 155 adc $Thi,$Thi,$t3 @ T += K[i] 161 ldr $t3,[sp,#$Coff+0] @ c.lo 188 and $Alo,$Alo,$t3 [all …]
|
| H A D | sha256-armv4.pl | 67 $T1="r3"; $t3="r3"; 144 and $t3,$t3,$t2 @ (b^c)&=(a^b) 146 eor $t3,$t3,$b @ Maj(a,b,c) 148 @ add $h,$h,$t3 @ h+=Maj(a,b,c) 150 ($t2,$t3)=($t3,$t2); 259 eor $t3,$B,$C @ magic 269 ldreq $t3,[sp,#16*4] @ pull ctx 273 ldr $t0,[$t3,#0] 274 ldr $t1,[$t3,#4] 275 ldr $t2,[$t3,#8] [all …]
|
| /freebsd/crypto/openssl/crypto/ec/asm/ |
| H A D | ecp_sm2p256-armv8.pl | 25 my ($t0,$t1,$t2,$t3)=map("x$_",(3..6)); 54 mov $t3,$s3 60 sbcs $t3,$t3,$s7 67 csel $s3,$s3,$t3,cc 101 mov $t3,$s3 107 adcs $t3,$t3,$s7 114 csel $s3,$s3,$t3,eq 277 mov $t3,$s3 293 csel $s3,$s3,$t3,cs 308 mov $t3,$s3 [all …]
|
| H A D | ecp_nistp521-ppc64.pl | 150 my ($t1, $t2, $t3, $t4) = ("v33", "v34", "v44", "v54"); 188 xxpermdi $t3,$in1[2],$in1[3],0b00 190 vmsumudm $out[3],$t3,$t4,$out[3] 195 vmsumudm $out[4],$t3,$t4,$out[4] 201 vmsumudm $out[5],$t3,$t4,$out[5] 206 vmsumudm $out[6],$t3,$t4,$out[6] 211 vmsumudm $out[7],$t3,$t4,$out[7] 216 vmsumudm $out[8],$t3,$t4,$out[8] 228 xxpermdi $t3,$in1[6],$in1[7],0b00 230 vmsumudm $out[7],$t3,$t4,$out[7] [all …]
|
| H A D | ecp_nistz256-armv8.pl | 49 my ($rp,$ap,$bp,$bi,$a0,$a1,$a2,$a3,$t0,$t1,$t2,$t3,$poly1,$poly3, 244 ldp $t2,$t3,[$bp,#16] 298 mov $t3,$acc3 325 mov $t3,$acc3 336 mov $t3,$a3 410 umulh $t3,$a3,$bi 418 adc $acc4,xzr,$t3 448 sbc $t3,$acc0,$t1 455 adcs $acc3,$acc4,$t3 456 mul $t3,$a3,$bi // lo(a[3]*b[i]) [all …]
|
| H A D | ecp_nistz256-ppc64.pl | 46 $acc4,$acc5,$a0,$a1,$a2,$a3,$t0,$t1,$t2,$t3) = 218 ld $t3, 24($bp) 293 mr $t3,$acc3 336 mr $t3,$acc3 349 ld $t3,88($sp) 452 mulhdu $t3,$a3,$bi 460 addze $acc4,$t3 491 subfe $t3,$t1,$acc0 495 adde $acc3,$acc4,$t3 501 mulld $t3,$a3,$bi # lo(a[3]*b[i]) [all …]
|
| H A D | ecp_nistz256-sparcv9.pl | 101 my ($t0,$t1,$t2,$t3,$t4,$t5,$t6,$t7)=(map("%o$_",(0..5)),"%g4","%g5"); 180 # CPU is 64-bit. [Addition with 64-bit carry was introduced in T3, see 192 ld [$ap+12],$t3 200 mulx $t3,$bi,$t3 208 srlx $t3,32,@acc[4] 221 addccc @acc[3],$t3,@acc[3] 223 ld [$ap+12],$t3 285 mulx $t3,$bi,$t3 295 add @acc[3],$t3,$t3 298 srlx $t3,32,@acc[4] [all …]
|
| /freebsd/crypto/openssl/crypto/md5/asm/ |
| H A D | md5-sparcv9.pl | 43 ($t1,$t2,$t3,$saved_asi)=map("%l$_",(4..7)); 83 sll $a,$rot,$t3 86 add $b,$t3,$t3 88 add $t3,$a,$a 99 sll $a,$rot,$t3 102 add $b,$t3,$t3 104 add $t3,$a,$a 121 sll $a,$rot,$t3 124 add $b,$t3,$t3 126 add $t3,$a,$a [all …]
|
| /freebsd/crypto/openssh/ |
| H A D | rijndael.c | 765 u32 s0, s1, s2, s3, t0, t1, t2, t3; in rijndaelEncrypt() local 783 t3 = Te0[s3 >> 24] ^ Te1[(s0 >> 16) & 0xff] ^ Te2[(s1 >> 8) & 0xff] ^ Te3[s2 & 0xff] ^ rk[ 7]; in rijndaelEncrypt() 785 s0 = Te0[t0 >> 24] ^ Te1[(t1 >> 16) & 0xff] ^ Te2[(t2 >> 8) & 0xff] ^ Te3[t3 & 0xff] ^ rk[ 8]; in rijndaelEncrypt() 786 s1 = Te0[t1 >> 24] ^ Te1[(t2 >> 16) & 0xff] ^ Te2[(t3 >> 8) & 0xff] ^ Te3[t0 & 0xff] ^ rk[ 9]; in rijndaelEncrypt() 787 s2 = Te0[t2 >> 24] ^ Te1[(t3 >> 16) & 0xff] ^ Te2[(t0 >> 8) & 0xff] ^ Te3[t1 & 0xff] ^ rk[10]; in rijndaelEncrypt() 788 s3 = Te0[t3 >> 24] ^ Te1[(t0 >> 16) & 0xff] ^ Te2[(t1 >> 8) & 0xff] ^ Te3[t2 & 0xff] ^ rk[11]; in rijndaelEncrypt() 793 t3 = Te0[s3 >> 24] ^ Te1[(s0 >> 16) & 0xff] ^ Te2[(s1 >> 8) & 0xff] ^ Te3[s2 & 0xff] ^ rk[15]; in rijndaelEncrypt() 795 s0 = Te0[t0 >> 24] ^ Te1[(t1 >> 16) & 0xff] ^ Te2[(t2 >> 8) & 0xff] ^ Te3[t3 & 0xff] ^ rk[16]; in rijndaelEncrypt() 796 s1 = Te0[t1 >> 24] ^ Te1[(t2 >> 16) & 0xff] ^ Te2[(t3 >> 8) & 0xff] ^ Te3[t0 & 0xff] ^ rk[17]; in rijndaelEncrypt() 797 s2 = Te0[t2 >> 24] ^ Te1[(t3 >> 16) & 0xff] ^ Te2[(t0 >> 8) & 0xff] ^ Te3[t1 & 0xff] ^ rk[18]; in rijndaelEncrypt() [all …]
|
| /freebsd/sys/crypto/rijndael/ |
| H A D | rijndael-alg-fst.c | 862 u32 s0, s1, s2, s3, t0, t1, t2, t3; in rijndaelEncrypt() local 880 t3 = Te0[s3 >> 24] ^ Te1[(s0 >> 16) & 0xff] ^ Te2[(s1 >> 8) & 0xff] ^ Te3[s2 & 0xff] ^ rk[ 7]; in rijndaelEncrypt() 882 s0 = Te0[t0 >> 24] ^ Te1[(t1 >> 16) & 0xff] ^ Te2[(t2 >> 8) & 0xff] ^ Te3[t3 & 0xff] ^ rk[ 8]; in rijndaelEncrypt() 883 s1 = Te0[t1 >> 24] ^ Te1[(t2 >> 16) & 0xff] ^ Te2[(t3 >> 8) & 0xff] ^ Te3[t0 & 0xff] ^ rk[ 9]; in rijndaelEncrypt() 884 s2 = Te0[t2 >> 24] ^ Te1[(t3 >> 16) & 0xff] ^ Te2[(t0 >> 8) & 0xff] ^ Te3[t1 & 0xff] ^ rk[10]; in rijndaelEncrypt() 885 s3 = Te0[t3 >> 24] ^ Te1[(t0 >> 16) & 0xff] ^ Te2[(t1 >> 8) & 0xff] ^ Te3[t2 & 0xff] ^ rk[11]; in rijndaelEncrypt() 890 t3 = Te0[s3 >> 24] ^ Te1[(s0 >> 16) & 0xff] ^ Te2[(s1 >> 8) & 0xff] ^ Te3[s2 & 0xff] ^ rk[15]; in rijndaelEncrypt() 892 s0 = Te0[t0 >> 24] ^ Te1[(t1 >> 16) & 0xff] ^ Te2[(t2 >> 8) & 0xff] ^ Te3[t3 & 0xff] ^ rk[16]; in rijndaelEncrypt() 893 s1 = Te0[t1 >> 24] ^ Te1[(t2 >> 16) & 0xff] ^ Te2[(t3 >> 8) & 0xff] ^ Te3[t0 & 0xff] ^ rk[17]; in rijndaelEncrypt() 894 s2 = Te0[t2 >> 24] ^ Te1[(t3 >> 16) & 0xff] ^ Te2[(t0 >> 8) & 0xff] ^ Te3[t1 & 0xff] ^ rk[18]; in rijndaelEncrypt() [all …]
|
| /freebsd/crypto/openssl/crypto/poly1305/asm/ |
| H A D | poly1305-x86_64.pl | 346 my ($H0,$H1,$H2,$H3,$H4, $T0,$T1,$T2,$T3,$T4, $D0,$D1,$D2,$D3,$D4, $MASK) = 819 vpsrldq \$6,$T1,$T3 822 vpunpcklqdq $T3,$T2,$T3 # 2:3 827 vpsrlq \$4,$T3,$T2 829 vpsrlq \$30,$T3,$T3 831 vpand $MASK,$T3,$T3 # 3 912 vpmuludq $T3,$D4,$D3 # d3 = h3*r0 918 vpmuludq $T3,$H2,$H1 # h3*r1 940 vpmuludq $T3,$H4,$H4 # h3*s2 951 vpmuludq $T3,$H3,$H1 # h3*s3 [all …]
|
| /freebsd/contrib/bearssl/src/hash/ |
| H A D | ghash_pclmul.c | 274 __m128i t0, t1, t2, t3; in br_ghash_pclmul() local 285 t3 = pclmulqdq00(h1w, h2w); in br_ghash_pclmul() 287 _mm_xor_si128(t1, t3)); in br_ghash_pclmul() 290 t2 = _mm_xor_si128(t2, _mm_shuffle_epi32(t3, 0x0E)); in br_ghash_pclmul() 291 SL_256(t0, t1, t2, t3); in br_ghash_pclmul() 292 REDUCE_F128(t0, t1, t2, t3); in br_ghash_pclmul() 327 t3 = _mm_xor_si128( in br_ghash_pclmul() 341 t2 = _mm_xor_si128(t2, _mm_xor_si128(t1, t3)); in br_ghash_pclmul() 344 t2 = _mm_xor_si128(t2, _mm_shuffle_epi32(t3, 0x0E)); in br_ghash_pclmul() 345 SL_256(t0, t1, t2, t3); in br_ghash_pclmul() [all …]
|