| /freebsd/crypto/openssl/crypto/modes/asm/ |
| H A D | ghashv8-armx.pl | 72 my ($t0,$t1,$t2,$xC2,$H,$Hhl,$H2)=map("q$_",(8..14)); 119 vshr.u64 $t2,$xC2,#63 121 vext.8 $t0,$t2,$xC2,#8 @ t0=0xc2....01 122 vshr.u64 $t2,$IN,#63 124 vand $t2,$t2,$t0 126 vext.8 $t2,$t2,$t2,#8 128 vorr $IN,$IN,$t2 @ H<<<=1 140 veor $t2,$Xl,$Xh 142 veor $Xm,$Xm,$t2 143 vpmull.p64 $t2,$Xl,$xC2 @ 1st phase [all …]
|
| H A D | ghash-armv4.pl | 382 my ($t0,$t1,$t2,$t3)=map("q$_",(8..12)); 396 vext.8 $t2#lo, $a, $a, #3 @ A3 398 vmull.p8 $t2, $t2#lo, $b @ J = A3*B 408 veor $t2, $t2, $r @ N = I + J 411 veor $t2#lo, $t2#lo, $t2#hi @ t2 = (N) (P4 + P5) << 24 412 vand $t2#hi, $t2#hi, $k16 417 veor $t2#lo, $t2#lo, $t2#hi 420 vext.8 $t2, $t2, $t2, #13 422 veor $t2, $t2, $t3 424 veor $r, $r, $t2 [all …]
|
| /freebsd/crypto/openssl/crypto/aes/asm/ |
| H A D | aes-armv4.pl | 64 $t2="r5"; 219 ldrb $t2,[$rounds,#1] 223 orr $s0,$s0,$t2,lsl#16 226 ldrb $t2,[$rounds,#5] 230 orr $s1,$s1,$t2,lsl#16 233 ldrb $t2,[$rounds,#9] 237 orr $s2,$s2,$t2,lsl#16 240 ldrb $t2,[$rounds,#13] 243 orr $s3,$s3,$t2,lsl#16 273 mov $t2,$s0,lsr#16 @ manner... [all …]
|
| /freebsd/crypto/openssl/crypto/ec/asm/ |
| H A D | ecp_nistp521-ppc64.pl | 150 my ($t1, $t2, $t3, $t4) = ("v33", "v34", "v44", "v54"); 179 xxpermdi $t2,$in2[1],$in2[0],0b00 180 vmsumudm $out[1],$t1,$t2,$vzero 182 xxpermdi $t2,$in2[2],$in2[1],0b00 183 vmsumudm $out[2],$t1,$t2,$vzero 186 xxpermdi $t2,$in2[3],$in2[2],0b00 187 vmsumudm $out[3],$t1,$t2,$vzero 192 xxpermdi $t2,$in2[4],$in2[3],0b00 193 vmsumudm $out[4],$t1,$t2,$vzero 198 xxpermdi $t2,$in2[5],$in2[4],0b00 [all …]
|
| H A D | ecp_nistz256-armv8.pl | 49 my ($rp,$ap,$bp,$bi,$a0,$a1,$a2,$a3,$t0,$t1,$t2,$t3,$poly1,$poly3, 244 ldp $t2,$t3,[$bp,#16] 297 mov $t2,$acc2 324 mov $t2,$acc2 335 mov $t2,$a2 407 umulh $t2,$a2,$bi 417 adcs $acc3,$acc3,$t2 447 subs $t2,$acc0,$t0 // "*0xffff0001" 453 adcs $acc2,$acc3,$t2 // +=acc[0]*0xffff0001 454 mul $t2,$a2,$bi // lo(a[2]*b[i]) [all …]
|
| H A D | ecp_nistz256-ppc64.pl | 46 $acc4,$acc5,$a0,$a1,$a2,$a3,$t0,$t1,$t2,$t3) = 216 ld $t2, 16($bp) 292 mr $t2,$acc2 334 mr $t2,$acc2 348 ld $t2,80($sp) 449 mulhdu $t2,$a2,$bi 459 adde $acc3,$acc3,$t2 490 subfc $t2,$t0,$acc0 # "*0xffff0001" 494 adde $acc2,$acc3,$t2 # +=acc[0]*0xffff0001 500 mulld $t2,$a2,$bi # lo(a[2]*b[i]) [all …]
|
| H A D | ecp_sm2p256-armv8.pl | 25 my ($t0,$t1,$t2,$t3)=map("x$_",(3..6)); 53 mov $t2,$s2 59 sbcs $t2,$t2,$s6 66 csel $s2,$s2,$t2,cc 100 mov $t2,$s2 106 adcs $t2,$t2,$s6 113 csel $s2,$s2,$t2,eq 276 mov $t2,$s2 292 csel $s2,$s2,$t2,cs 307 mov $t2,$s2 [all …]
|
| H A D | ecp_nistz256-sparcv9.pl | 101 my ($t0,$t1,$t2,$t3,$t4,$t5,$t6,$t7)=(map("%o$_",(0..5)),"%g4","%g5"); 191 ld [$ap+8],$t2 199 mulx $t2,$bi,$t2 207 srlx $t2,32,@acc[3] 220 addccc @acc[2],$t2,@acc[2] 222 ld [$ap+8],$t2 284 mulx $t2,$bi,$t2 293 add @acc[2],$t2,$t2 296 srlx $t2,32,@acc[3] 310 addccc @acc[2],$t2,@acc[2] [all …]
|
| /freebsd/crypto/openssl/crypto/sha/asm/ |
| H A D | sha1-thumb.pl | 36 $t2="r2"; 53 lsl $t2,$a,#5 54 add $t2,$e 56 add $t2,$e @ E+=ROR(A,27) 57 add $t2,$t1 @ E+=X[i] 68 add $a,$t2,$t1 @ A=E+F_xx_xx(B,C,D) 140 mov $t2,sp 141 sub $t2,#16*4 @ [3] 155 cmp sp,$t2 159 sub $t2,#32*4 [all …]
|
| H A D | sha512-armv4.pl | 89 $t2="r11"; 116 ldr $t2,[sp,#$Hoff+0] @ h.lo 131 adds $Tlo,$Tlo,$t2 132 ldr $t2,[sp,#$Goff+0] @ g.lo 136 eor $t0,$t0,$t2 144 eor $t0,$t0,$t2 145 ldr $t2,[$Ktbl,#$lo] @ K[i].lo 153 adds $Tlo,$Tlo,$t2 154 and $t0,$t2,#0xff 157 ldr $t2,[sp,#$Boff+0] @ b.lo [all …]
|
| H A D | sha256-armv4.pl | 77 $t2="r12"; 95 add $a,$a,$t2 @ h+=Maj(a,b,c) from the past 102 add $a,$a,$t2 @ h+=Maj(a,b,c) from the past 103 ldrb $t2,[$inp,#2] 105 orr $t1,$t1,$t2,lsl#8 106 ldrb $t2,[$inp],#4 112 orr $t1,$t1,$t2,lsl#24 117 ldr $t2,[$Ktbl],#4 @ *K256++ 123 add $h,$h,$t2 @ h+=K256[i] 128 and $t2,$t2,#0xff [all …]
|
| H A D | sha1-alpha.pl | 36 $t2="ra"; 58 srl @X[$i],8,$t2 63 zapnot $t2,0x22,$t2 66 or $t0,$t2,$t0 72 and $b,$c,$t2 81 or $t2,$t3,$t2 89 addl $t2,$e,$e 95 and $b,$c,$t2 103 or $t2,$t3,$t2 109 addl $t2,$e,$e [all …]
|
| /freebsd/contrib/bc/scripts/ |
| H A D | sqrt_frac_guess.bc | 53 t2 = (1/i) 56 l = length(t2$)/2 76 print "t2: ", t2, "\n" 92 t2 = (t2 >> l$) 96 t2 = (t2 >> l$) / 2 100 t2 = (t2 >> l$) / 4 111 t2 = (t2 >> l$) * 2 115 t2 = (t2 >> l$) 120 good = (t1 < t2 && t2 < t3) 122 print t1, " < ", t2, " < ", t3, ": ", good, "\n\n"
|
| /freebsd/lib/libc/riscv/string/ |
| H A D | memchr.S | 39 sltu t2, a0, a3 41 neg t2, t2 43 and a3, a3, t2 45 slli t2, a0, 3 56 sll t2, t0, t2 58 xor t2, t2, t0 59 or a2, a2, t2 65 not t2, a2 67 and t2, t2, t1 69 and a2, a2, t2 [all …]
|
| H A D | memcpy.S | 35 la t2, .Lduff_start 37 sub t2, t2, t3 38 jr t2 105 ld t2, (t0) 112 sll t3, t2, a5 116 srl t1, t2, a4 128 slli t2, a2, 3 129 sub t1, t1, t2 131 lb t2, 6(a1) 132 sb t2, 6(a3) [all …]
|
| H A D | strnlen.S | 34 neg t2, a0 37 slli t2, t2, 3 40 srl t2, t0, t2 49 or a3, a3, t2 52 not t2, a3 54 and t2, t2, t1 55 and a3, a3, t2 62 sub t2, a4, a2 63 andi t2, t2, ~0b1111 64 add a5, a2, t2 [all …]
|
| H A D | strchrnul.S | 39 andi t2, a0, 0b111 45 beqz t2, .Lloop 50 slli t2, t2, 3 51 neg t2, t2 52 srl t2, t0, t2 55 or a3, a2, t2 58 or a2, a2, t2 62 not t2, a2 66 and a2, a2, t2 83 not t2, a2 [all …]
|
| /freebsd/sys/contrib/libb2/ |
| H A D | blake2s-load-sse41.h | 32 t2 = _mm_blend_epi16(t0, t1, 0xF0); \ 33 buf = _mm_shuffle_epi32(t2, _MM_SHUFFLE(2,1,0,3)); 38 t2 = _mm_blend_epi16(t0, t1, 0xF0); \ 39 buf = _mm_shuffle_epi32(t2, _MM_SHUFFLE(2,3,0,1)); 44 t2 = _mm_blend_epi16(m0, t1, 0xF0); \ 45 buf = _mm_shuffle_epi32(t2, _MM_SHUFFLE(2,3,0,1)); 50 t2 = _mm_blend_epi16(t0, t1, 0x0C); \ 51 buf = _mm_shuffle_epi32(t2, _MM_SHUFFLE(2,3,0,1)); 56 t2 = _mm_blend_epi16(t0, t1, 0x0F); \ 57 buf = _mm_shuffle_epi32(t2, _MM_SHUFFLE(3,1,0,2)); [all …]
|
| /freebsd/crypto/openssl/crypto/md5/asm/ |
| H A D | md5-sparcv9.pl | 43 ($t1,$t2,$t3,$saved_asi)=map("%l$_",(4..7)); 76 add $t2,$a,$a 80 sethi %hi(@K[$i+1]),$t2 82 or $t2,%lo(@K[$i+1]),$t2 84 add @X[$j],$t2,$t2 ! X[`$i+1`]+K[`$i+1`] 94 add $t2,$a,$a 96 sethi %hi(@K[$i+1]),$t2 98 or $t2,%lo(@K[$i+1]),$t2 100 add $tx,$t2,$t2 ! X[`2*$j+1`]+K[`$i+1`] 116 add $t2,$a,$a [all …]
|
| /freebsd/crypto/openssh/ |
| H A D | rijndael.c | 765 u32 s0, s1, s2, s3, t0, t1, t2, t3; in rijndaelEncrypt() local 782 t2 = Te0[s2 >> 24] ^ Te1[(s3 >> 16) & 0xff] ^ Te2[(s0 >> 8) & 0xff] ^ Te3[s1 & 0xff] ^ rk[ 6]; in rijndaelEncrypt() 785 s0 = Te0[t0 >> 24] ^ Te1[(t1 >> 16) & 0xff] ^ Te2[(t2 >> 8) & 0xff] ^ Te3[t3 & 0xff] ^ rk[ 8]; in rijndaelEncrypt() 786 s1 = Te0[t1 >> 24] ^ Te1[(t2 >> 16) & 0xff] ^ Te2[(t3 >> 8) & 0xff] ^ Te3[t0 & 0xff] ^ rk[ 9]; in rijndaelEncrypt() 787 s2 = Te0[t2 >> 24] ^ Te1[(t3 >> 16) & 0xff] ^ Te2[(t0 >> 8) & 0xff] ^ Te3[t1 & 0xff] ^ rk[10]; in rijndaelEncrypt() 788 s3 = Te0[t3 >> 24] ^ Te1[(t0 >> 16) & 0xff] ^ Te2[(t1 >> 8) & 0xff] ^ Te3[t2 & 0xff] ^ rk[11]; in rijndaelEncrypt() 792 t2 = Te0[s2 >> 24] ^ Te1[(s3 >> 16) & 0xff] ^ Te2[(s0 >> 8) & 0xff] ^ Te3[s1 & 0xff] ^ rk[14]; in rijndaelEncrypt() 795 s0 = Te0[t0 >> 24] ^ Te1[(t1 >> 16) & 0xff] ^ Te2[(t2 >> 8) & 0xff] ^ Te3[t3 & 0xff] ^ rk[16]; in rijndaelEncrypt() 796 s1 = Te0[t1 >> 24] ^ Te1[(t2 >> 16) & 0xff] ^ Te2[(t3 >> 8) & 0xff] ^ Te3[t0 & 0xff] ^ rk[17]; in rijndaelEncrypt() 797 s2 = Te0[t2 >> 24] ^ Te1[(t3 >> 16) & 0xff] ^ Te2[(t0 >> 8) & 0xff] ^ Te3[t1 & 0xff] ^ rk[18]; in rijndaelEncrypt() [all …]
|
| /freebsd/sys/crypto/rijndael/ |
| H A D | rijndael-alg-fst.c | 862 u32 s0, s1, s2, s3, t0, t1, t2, t3; in rijndaelEncrypt() local 879 t2 = Te0[s2 >> 24] ^ Te1[(s3 >> 16) & 0xff] ^ Te2[(s0 >> 8) & 0xff] ^ Te3[s1 & 0xff] ^ rk[ 6]; in rijndaelEncrypt() 882 s0 = Te0[t0 >> 24] ^ Te1[(t1 >> 16) & 0xff] ^ Te2[(t2 >> 8) & 0xff] ^ Te3[t3 & 0xff] ^ rk[ 8]; in rijndaelEncrypt() 883 s1 = Te0[t1 >> 24] ^ Te1[(t2 >> 16) & 0xff] ^ Te2[(t3 >> 8) & 0xff] ^ Te3[t0 & 0xff] ^ rk[ 9]; in rijndaelEncrypt() 884 s2 = Te0[t2 >> 24] ^ Te1[(t3 >> 16) & 0xff] ^ Te2[(t0 >> 8) & 0xff] ^ Te3[t1 & 0xff] ^ rk[10]; in rijndaelEncrypt() 885 s3 = Te0[t3 >> 24] ^ Te1[(t0 >> 16) & 0xff] ^ Te2[(t1 >> 8) & 0xff] ^ Te3[t2 & 0xff] ^ rk[11]; in rijndaelEncrypt() 889 t2 = Te0[s2 >> 24] ^ Te1[(s3 >> 16) & 0xff] ^ Te2[(s0 >> 8) & 0xff] ^ Te3[s1 & 0xff] ^ rk[14]; in rijndaelEncrypt() 892 s0 = Te0[t0 >> 24] ^ Te1[(t1 >> 16) & 0xff] ^ Te2[(t2 >> 8) & 0xff] ^ Te3[t3 & 0xff] ^ rk[16]; in rijndaelEncrypt() 893 s1 = Te0[t1 >> 24] ^ Te1[(t2 >> 16) & 0xff] ^ Te2[(t3 >> 8) & 0xff] ^ Te3[t0 & 0xff] ^ rk[17]; in rijndaelEncrypt() 894 s2 = Te0[t2 >> 24] ^ Te1[(t3 >> 16) & 0xff] ^ Te2[(t0 >> 8) & 0xff] ^ Te3[t1 & 0xff] ^ rk[18]; in rijndaelEncrypt() [all …]
|
| /freebsd/crypto/openssl/crypto/bn/asm/ |
| H A D | armv8-mont.pl | 669 my ($t0,$t1,$t2,$t3)=map("x$_",(14..17)); 765 mul $t2,$a3,$a0 771 adcs $acc3,$acc3,$t2 772 mul $t2,$a7,$a0 779 adcs $acc7,$acc7,$t2 780 umulh $t2,$a4,$a0 789 adcs $acc5,$acc5,$t2 790 mul $t2,$a2,$a1 // lo(a[2..7]*a[1]) (ii) 798 adds $acc3,$acc3,$t2 799 mul $t2,$a6,$a1 [all …]
|
| H A D | ppc64-mont.pl | 141 $t2="r26"; 290 extrdi $t2,$t3,16,16 294 std $t2,`$FRAME+16`($sp) 310 lwz $t2,`12^$LITTLE_ENDIAN`($ap) ; load a[1] as 32-bit word pair 336 extrwi $t2,$t3,16,16 340 std $t2,`$FRAME+16`($sp) 345 mullw $t2,$t5,$n0 347 add $t1,$t1,$t2 361 lwz $t2,8($ap) ; load a[j..j+3] as 32-bit word pairs 379 std $t2,`$FRAME+80`($sp) [all …]
|
| /freebsd/contrib/bearssl/src/ec/ |
| H A D | ec_p256_m64.c | 263 uint64_t x, f, t0, t1, t2, t3, t4; in f256_montymul() local 288 t2 = (uint64_t)z; in f256_montymul() 291 z = (unsigned __int128)t2 + (uint64_t)ff; in f256_montymul() 292 t2 = (uint64_t)z; in f256_montymul() 308 z = (unsigned __int128)b[2] * x + t2 + (z >> 64); in f256_montymul() 311 t2 = (uint64_t)z; in f256_montymul() 325 z = (z >> 64) + (unsigned __int128)t2 + (uint64_t)ff; in f256_montymul() 326 t2 = (uint64_t)z; in f256_montymul() 354 z = (unsigned __int128)t2 - (z >> 127); in f256_montymul() 355 t2 = (uint64_t)z; in f256_montymul() [all …]
|
| /freebsd/contrib/bearssl/src/symcipher/ |
| H A D | aes_x86ni.c | 62 expand_step192(__m128i *t1, __m128i *t2, __m128i *t3) in expand_step192() argument 66 *t2 = _mm_shuffle_epi32(*t2, 0x55); in expand_step192() 73 *t1 = _mm_xor_si128(*t1, *t2); in expand_step192() 74 *t2 = _mm_shuffle_epi32(*t1, 0xFF); in expand_step192() 77 *t3 = _mm_xor_si128(*t3, *t2); in expand_step192() 82 expand_step256_1(__m128i *t1, __m128i *t2) in expand_step256_1() argument 86 *t2 = _mm_shuffle_epi32(*t2, 0xFF); in expand_step256_1() 93 *t1 = _mm_xor_si128(*t1, *t2); in expand_step256_1() 100 __m128i t2, t4; in expand_step256_2() local 103 t2 = _mm_shuffle_epi32(t4, 0xAA); in expand_step256_2() [all …]
|