/freebsd/crypto/openssl/crypto/aes/asm/ |
H A D | aes-armv4.pl | 64 $t2="r5"; 219 ldrb $t2,[$rounds,#1] 223 orr $s0,$s0,$t2,lsl#16 226 ldrb $t2,[$rounds,#5] 230 orr $s1,$s1,$t2,lsl#16 233 ldrb $t2,[$rounds,#9] 237 orr $s2,$s2,$t2,lsl#16 240 ldrb $t2,[$rounds,#13] 243 orr $s3,$s3,$t2,lsl#16 273 mov $t2,$s0,lsr#16 @ manner... [all …]
|
/freebsd/crypto/openssl/crypto/modes/asm/ |
H A D | ghashv8-armx.pl | 72 my ($t0,$t1,$t2,$xC2,$H,$Hhl,$H2)=map("q$_",(8..14)); 119 vshr.u64 $t2,$xC2,#63 121 vext.8 $t0,$t2,$xC2,#8 @ t0=0xc2....01 122 vshr.u64 $t2,$IN,#63 124 vand $t2,$t2,$t0 126 vext.8 $t2,$t2,$t2,#8 128 vorr $IN,$IN,$t2 @ H<<<=1 140 veor $t2,$Xl,$Xh 142 veor $Xm,$Xm,$t2 143 vpmull.p64 $t2,$Xl,$xC2 @ 1st phase [all …]
|
H A D | ghash-armv4.pl | 382 my ($t0,$t1,$t2,$t3)=map("q$_",(8..12)); 396 vext.8 $t2#lo, $a, $a, #3 @ A3 398 vmull.p8 $t2, $t2#lo, $b @ J = A3*B 408 veor $t2, $t2, $r @ N = I + J 411 veor $t2#lo, $t2#lo, $t2#hi @ t2 = (N) (P4 + P5) << 24 412 vand $t2#hi, $t2#hi, $k16 417 veor $t2#lo, $t2#lo, $t2#hi 420 vext.8 $t2, $t2, $t2, #13 422 veor $t2, $t2, $t3 424 veor $r, $r, $t2 [all …]
|
/freebsd/crypto/openssl/crypto/ec/asm/ |
H A D | ecp_nistp521-ppc64.pl | 150 my ($t1, $t2, $t3, $t4) = ("v33", "v34", "v44", "v54"); 179 xxpermdi $t2,$in2[1],$in2[0],0b00 180 vmsumudm $out[1],$t1,$t2,$vzero 182 xxpermdi $t2,$in2[2],$in2[1],0b00 183 vmsumudm $out[2],$t1,$t2,$vzero 186 xxpermdi $t2,$in2[3],$in2[2],0b00 187 vmsumudm $out[3],$t1,$t2,$vzero 192 xxpermdi $t2,$in2[4],$in2[3],0b00 193 vmsumudm $out[4],$t1,$t2,$vzero 198 xxpermdi $t2,$in2[5],$in2[4],0b00 [all …]
|
H A D | ecp_nistz256-armv8.pl | 49 my ($rp,$ap,$bp,$bi,$a0,$a1,$a2,$a3,$t0,$t1,$t2,$t3,$poly1,$poly3, 231 ldp $t2,$t3,[$bp,#16] 278 mov $t2,$acc2 303 mov $t2,$acc2 314 mov $t2,$a2 382 umulh $t2,$a2,$bi 392 adcs $acc3,$acc3,$t2 422 subs $t2,$acc0,$t0 // "*0xffff0001" 428 adcs $acc2,$acc3,$t2 // +=acc[0]*0xffff0001 429 mul $t2,$a2,$bi // lo(a[2]*b[i]) [all …]
|
H A D | ecp_nistz256-ppc64.pl | 46 $acc4,$acc5,$a0,$a1,$a2,$a3,$t0,$t1,$t2,$t3) = 216 ld $t2, 16($bp) 292 mr $t2,$acc2 334 mr $t2,$acc2 348 ld $t2,80($sp) 449 mulhdu $t2,$a2,$bi 459 adde $acc3,$acc3,$t2 490 subfc $t2,$t0,$acc0 # "*0xffff0001" 494 adde $acc2,$acc3,$t2 # +=acc[0]*0xffff0001 500 mulld $t2,$a2,$bi # lo(a[2]*b[i]) [all …]
|
H A D | ecp_nistz256-sparcv9.pl | 101 my ($t0,$t1,$t2,$t3,$t4,$t5,$t6,$t7)=(map("%o$_",(0..5)),"%g4","%g5"); 191 ld [$ap+8],$t2 199 mulx $t2,$bi,$t2 207 srlx $t2,32,@acc[3] 220 addccc @acc[2],$t2,@acc[2] 222 ld [$ap+8],$t2 284 mulx $t2,$bi,$t2 293 add @acc[2],$t2,$t2 296 srlx $t2,32,@acc[3] 310 addccc @acc[2],$t2,@acc[2] [all …]
|
/freebsd/crypto/openssl/crypto/sha/asm/ |
H A D | sha1-thumb.pl | 36 $t2="r2"; 53 lsl $t2,$a,#5 54 add $t2,$e 56 add $t2,$e @ E+=ROR(A,27) 57 add $t2,$t1 @ E+=X[i] 68 add $a,$t2,$t1 @ A=E+F_xx_xx(B,C,D) 140 mov $t2,sp 141 sub $t2,#16*4 @ [3] 155 cmp sp,$t2 159 sub $t2,#32*4 [all …]
|
H A D | sha512-armv4.pl | 89 $t2="r11"; 116 ldr $t2,[sp,#$Hoff+0] @ h.lo 131 adds $Tlo,$Tlo,$t2 132 ldr $t2,[sp,#$Goff+0] @ g.lo 136 eor $t0,$t0,$t2 144 eor $t0,$t0,$t2 145 ldr $t2,[$Ktbl,#$lo] @ K[i].lo 153 adds $Tlo,$Tlo,$t2 154 and $t0,$t2,#0xff 157 ldr $t2,[sp,#$Boff+0] @ b.lo [all …]
|
H A D | sha256-armv4.pl | 77 $t2="r12"; 95 add $a,$a,$t2 @ h+=Maj(a,b,c) from the past 102 add $a,$a,$t2 @ h+=Maj(a,b,c) from the past 103 ldrb $t2,[$inp,#2] 105 orr $t1,$t1,$t2,lsl#8 106 ldrb $t2,[$inp],#4 112 orr $t1,$t1,$t2,lsl#24 117 ldr $t2,[$Ktbl],#4 @ *K256++ 123 add $h,$h,$t2 @ h+=K256[i] 128 and $t2,$t2,#0xff [all …]
|
H A D | sha1-alpha.pl | 36 $t2="ra"; 58 srl @X[$i],8,$t2 63 zapnot $t2,0x22,$t2 66 or $t0,$t2,$t0 72 and $b,$c,$t2 81 or $t2,$t3,$t2 89 addl $t2,$e,$e 95 and $b,$c,$t2 103 or $t2,$t3,$t2 109 addl $t2,$e,$e [all …]
|
/freebsd/contrib/bc/scripts/ |
H A D | sqrt_frac_guess.bc | 53 t2 = (1/i) 56 l = length(t2$)/2 76 print "t2: ", t2, "\n" 92 t2 = (t2 >> l$) 96 t2 = (t2 >> l$) / 2 100 t2 = (t2 >> [all...] |
/freebsd/sys/contrib/libb2/ |
H A D | blake2s-load-sse41.h | 32 t2 = _mm_blend_epi16(t0, t1, 0xF0); \ 33 buf = _mm_shuffle_epi32(t2, _MM_SHUFFLE(2,1,0,3)); 38 t2 = _mm_blend_epi16(t0, t1, 0xF0); \ 39 buf = _mm_shuffle_epi32(t2, _MM_SHUFFLE(2,3,0,1)); 44 t2 = _mm_blend_epi16(m0, t1, 0xF0); \ 45 buf = _mm_shuffle_epi32(t2, _MM_SHUFFLE(2,3,0,1)); 50 t2 = _mm_blend_epi16(t0, t1, 0x0C); \ 51 buf = _mm_shuffle_epi32(t2, _MM_SHUFFLE(2,3,0,1)); 56 t2 = _mm_blend_epi16(t0, t1, 0x0F); \ 57 buf = _mm_shuffle_epi32(t2, _MM_SHUFFLE(3,1,0,2)); [all …]
|
/freebsd/crypto/openssl/crypto/md5/asm/ |
H A D | md5-sparcv9.pl | 43 ($t1,$t2,$t3,$saved_asi)=map("%l$_",(4..7)); 76 add $t2,$a,$a 80 sethi %hi(@K[$i+1]),$t2 82 or $t2,%lo(@K[$i+1]),$t2 84 add @X[$j],$t2,$t2 ! X[`$i+1`]+K[`$i+1`] 94 add $t2,$a,$a 96 sethi %hi(@K[$i+1]),$t2 98 or $t2,%lo(@K[$i+1]),$t2 100 add $tx,$t2,$t2 ! X[`2*$j+1`]+K[`$i+1`] 116 add $t2,$a,$a [all …]
|
/freebsd/crypto/openssh/ |
H A D | rijndael.c | 765 u32 s0, s1, s2, s3, t0, t1, t2, t3; in rijndaelEncrypt() local 782 t2 = Te0[s2 >> 24] ^ Te1[(s3 >> 16) & 0xff] ^ Te2[(s0 >> 8) & 0xff] ^ Te3[s1 & 0xff] ^ rk[ 6]; in rijndaelEncrypt() 785 s0 = Te0[t0 >> 24] ^ Te1[(t1 >> 16) & 0xff] ^ Te2[(t2 >> 8) & 0xff] ^ Te3[t3 & 0xff] ^ rk[ 8]; in rijndaelEncrypt() 786 s1 = Te0[t1 >> 24] ^ Te1[(t2 >> 16) & 0xff] ^ Te2[(t3 >> 8) & 0xff] ^ Te3[t0 & 0xff] ^ rk[ 9]; in rijndaelEncrypt() 787 s2 = Te0[t2 >> 24] ^ Te1[(t3 >> 16) & 0xff] ^ Te2[(t0 >> 8) & 0xff] ^ Te3[t1 & 0xff] ^ rk[10]; in rijndaelEncrypt() 788 s3 = Te0[t3 >> 24] ^ Te1[(t0 >> 16) & 0xff] ^ Te2[(t1 >> 8) & 0xff] ^ Te3[t2 & 0xff] ^ rk[11]; in rijndaelEncrypt() 792 t2 = Te0[s2 >> 24] ^ Te1[(s3 >> 16) & 0xff] ^ Te2[(s0 >> 8) & 0xff] ^ Te3[s1 & 0xff] ^ rk[14]; in rijndaelEncrypt() 795 s0 = Te0[t0 >> 24] ^ Te1[(t1 >> 16) & 0xff] ^ Te2[(t2 >> 8) & 0xff] ^ Te3[t3 & 0xff] ^ rk[16]; in rijndaelEncrypt() 796 s1 = Te0[t1 >> 24] ^ Te1[(t2 >> 16) & 0xff] ^ Te2[(t3 >> 8) & 0xff] ^ Te3[t0 & 0xff] ^ rk[17]; in rijndaelEncrypt() 797 s2 = Te0[t2 >> 24] ^ Te1[(t3 >> 16) & 0xff] ^ Te2[(t0 >> 8) & 0xff] ^ Te3[t1 & 0xff] ^ rk[18]; in rijndaelEncrypt() [all …]
|
/freebsd/sys/crypto/rijndael/ |
H A D | rijndael-alg-fst.c | 862 u32 s0, s1, s2, s3, t0, t1, t2, t3; in rijndaelEncrypt() local 879 t2 = Te0[s2 >> 24] ^ Te1[(s3 >> 16) & 0xff] ^ Te2[(s0 >> 8) & 0xff] ^ Te3[s1 & 0xff] ^ rk[ 6]; in rijndaelEncrypt() 882 s0 = Te0[t0 >> 24] ^ Te1[(t1 >> 16) & 0xff] ^ Te2[(t2 >> 8) & 0xff] ^ Te3[t3 & 0xff] ^ rk[ 8]; in rijndaelEncrypt() 883 s1 = Te0[t1 >> 24] ^ Te1[(t2 >> 16) & 0xff] ^ Te2[(t3 >> 8) & 0xff] ^ Te3[t0 & 0xff] ^ rk[ 9]; in rijndaelEncrypt() 884 s2 = Te0[t2 >> 24] ^ Te1[(t3 >> 16) & 0xff] ^ Te2[(t0 >> 8) & 0xff] ^ Te3[t1 & 0xff] ^ rk[10]; in rijndaelEncrypt() 885 s3 = Te0[t3 >> 24] ^ Te1[(t0 >> 16) & 0xff] ^ Te2[(t1 >> 8) & 0xff] ^ Te3[t2 & 0xff] ^ rk[11]; in rijndaelEncrypt() 889 t2 = Te0[s2 >> 24] ^ Te1[(s3 >> 16) & 0xff] ^ Te2[(s0 >> 8) & 0xff] ^ Te3[s1 & 0xff] ^ rk[14]; in rijndaelEncrypt() 892 s0 = Te0[t0 >> 24] ^ Te1[(t1 >> 16) & 0xff] ^ Te2[(t2 >> 8) & 0xff] ^ Te3[t3 & 0xff] ^ rk[16]; in rijndaelEncrypt() 893 s1 = Te0[t1 >> 24] ^ Te1[(t2 >> 16) & 0xff] ^ Te2[(t3 >> 8) & 0xff] ^ Te3[t0 & 0xff] ^ rk[17]; in rijndaelEncrypt() 894 s2 = Te0[t2 >> 24] ^ Te1[(t3 >> 16) & 0xff] ^ Te2[(t0 >> 8) & 0xff] ^ Te3[t1 & 0xff] ^ rk[18]; in rijndaelEncrypt() [all …]
|
/freebsd/crypto/openssl/crypto/bn/asm/ |
H A D | armv8-mont.pl | 669 my ($t0,$t1,$t2,$t3)=map("x$_",(14..17)); 765 mul $t2,$a3,$a0 771 adcs $acc3,$acc3,$t2 772 mul $t2,$a7,$a0 779 adcs $acc7,$acc7,$t2 780 umulh $t2,$a4,$a0 789 adcs $acc5,$acc5,$t2 790 mul $t2,$a2,$a1 // lo(a[2..7]*a[1]) (ii) 798 adds $acc3,$acc3,$t2 799 mul $t2,$a6,$a1 [all …]
|
H A D | ppc64-mont.pl | 141 $t2="r26"; 290 extrdi $t2,$t3,16,16 294 std $t2,`$FRAME+16`($sp) 310 lwz $t2,`12^$LITTLE_ENDIAN`($ap) ; load a[1] as 32-bit word pair 336 extrwi $t2,$t3,16,16 340 std $t2,`$FRAME+16`($sp) 345 mullw $t2,$t5,$n0 347 add $t1,$t1,$t2 361 lwz $t2,8($ap) ; load a[j..j+3] as 32-bit word pairs 379 std $t2,`$FRAME+80`($sp) [all …]
|
H A D | armv4-gf2m.pl | 253 my ($r,$t0,$t1,$t2,$t3)=map("q$_",(0..3,8..12)); 278 vext.8 $t2#lo, $a, $a, #3 @ A3 280 vmull.p8 $t2, $t2#lo, $b @ J = A3*B 290 veor $t2, $t2, $r @ N = I + J 293 veor $t2#lo, $t2#lo, $t2#hi @ t2 = (N) (P4 + P5) << 24 294 vand $t2#hi, $t2#hi, $k16 299 veor $t2#lo, $t2#lo, $t2#hi 302 vext.8 $t2, $t2, $t2, #13 304 veor $t2, $t2, $t3 306 veor $r, $r, $t2
|
H A D | ppc-mont.pl | 358 $t0,$t1,$t2,$t3, 457 $UMULL $t2,$a2,$bi 464 adde $acc2,$acc2,$t2 467 $UMULH $t2,$a2,$bi 476 adde $acc3,$acc3,$t2 477 $UMULL $t2,$m2,$mi 494 adde $acc1,$acc2,$t2 495 $UMULH $t2,$m2,$mi 502 adde $acc2,$acc2,$t2 527 $UMULL $t2,$a2,$bi [all …]
|
/freebsd/contrib/bearssl/src/ec/ |
H A D | ec_p256_m64.c | 263 uint64_t x, f, t0, t1, t2, t3, t4; in f256_montymul() local 288 t2 = (uint64_t)z; in f256_montymul() 291 z = (unsigned __int128)t2 + (uint64_t)ff; in f256_montymul() 292 t2 = (uint64_t)z; in f256_montymul() 308 z = (unsigned __int128)b[2] * x + t2 + (z >> 64); in f256_montymul() 311 t2 = (uint64_t)z; in f256_montymul() 325 z = (z >> 64) + (unsigned __int128)t2 + (uint64_t)ff; in f256_montymul() 326 t2 = (uint64_t)z; in f256_montymul() 354 z = (unsigned __int128)t2 - (z >> 127); in f256_montymul() 355 t2 = (uint64_t)z; in f256_montymul() [all …]
|
H A D | ec_c25519_m64.c | 102 uint64_t t0, t1, t2, t3, cc; in f255_add() local 110 t2 = (uint64_t)z; in f255_add() 127 z = (unsigned __int128)t2 + (z >> 64); in f255_add() 133 uint64_t t0, t1, t2, t3, cc; in f255_add() 138 k = _addcarry_u64(k, a[2], b[2], &t2); in f255_add() 153 k = _addcarry_u64(k, t2, 0, &d[2]); in f255_add() 174 uint64_t t0, t1, t2, t3, t4, cc; in f255_sub() local 186 t2 = (uint64_t)z; in f255_sub() 205 z = (unsigned __int128)t2 + (z >> 64); in f255_sub() 218 uint64_t t0, t1, t2, t3, t4; in f255_sub() [all …]
|
/freebsd/contrib/bearssl/src/symcipher/ |
H A D | aes_x86ni.c | 62 expand_step192(__m128i *t1, __m128i *t2, __m128i *t3) in expand_step192() argument 66 *t2 = _mm_shuffle_epi32(*t2, 0x55); in expand_step192() 73 *t1 = _mm_xor_si128(*t1, *t2); in expand_step192() 74 *t2 = _mm_shuffle_epi32(*t1, 0xFF); in expand_step192() 77 *t3 = _mm_xor_si128(*t3, *t2); in expand_step192() 82 expand_step256_1(__m128i *t1, __m128i *t2) in expand_step256_1() argument 86 *t2 = _mm_shuffle_epi32(*t2, 0xFF); in expand_step256_1() 93 *t1 = _mm_xor_si128(*t1, *t2); in expand_step256_1() 100 __m128i t2, t4; in expand_step256_2() local 103 t2 = _mm_shuffle_epi32(t4, 0xAA); in expand_step256_2() [all …]
|
/freebsd/contrib/bearssl/src/hash/ |
H A D | ghash_pclmul.c | 274 __m128i t0, t1, t2, t3; in br_ghash_pclmul() local 286 t2 = _mm_xor_si128(pclmulqdq00(h1x, h2x), in br_ghash_pclmul() 289 t1 = _mm_xor_si128(t1, _mm_shuffle_epi32(t2, 0x0E)); in br_ghash_pclmul() 290 t2 = _mm_xor_si128(t2, _mm_shuffle_epi32(t3, 0x0E)); in br_ghash_pclmul() 291 SL_256(t0, t1, t2, t3); in br_ghash_pclmul() 292 REDUCE_F128(t0, t1, t2, t3); in br_ghash_pclmul() 334 t2 = _mm_xor_si128( in br_ghash_pclmul() 341 t2 = _mm_xor_si128(t2, _mm_xor_si128(t1, t3)); in br_ghash_pclmul() 343 t1 = _mm_xor_si128(t1, _mm_shuffle_epi32(t2, 0x0E)); in br_ghash_pclmul() 344 t2 = _mm_xor_si128(t2, _mm_shuffle_epi32(t3, 0x0E)); in br_ghash_pclmul() [all …]
|
/freebsd/contrib/kyua/utils/config/ |
H A D | tree_test.cpp | 216 const config::tree t1, t2; in ATF_TEST_CASE_BODY() local 217 const config::tree combined = t1.combine(t2); in ATF_TEST_CASE_BODY() 238 config::tree t1, t2; in ATF_TEST_CASE_BODY() local 240 init_tree_for_combine_test(t2); in ATF_TEST_CASE_BODY() 246 const config::tree combined = t1.combine(t2); in ATF_TEST_CASE_BODY() 255 config::tree t1, t2; in ATF_TEST_CASE_BODY() local 257 init_tree_for_combine_test(t2); in ATF_TEST_CASE_BODY() 258 t2.set< config::int_node >("int-node", 3); in ATF_TEST_CASE_BODY() 259 t2.set< config::string_node >("string-node", "foo"); in ATF_TEST_CASE_BODY() 260 t2.set< config::int_node >("deeper.int.node", 15); in ATF_TEST_CASE_BODY() [all …]
|