| /freebsd/crypto/openssl/crypto/aes/asm/ |
| H A D | aes-riscv64-zkn.pl | 120 my ($T0,$T1) = use_regs(13..14); 147 ld $T1,8($KEYP) 154 xor $Q1,$Q1,$T1 169 ld $T1,8($KEYP) 171 xor $Q1,$Q3,$T1 182 ld $T1,24($KEYP) 184 xor $Q1,$Q3,$T1 225 ld $T1,8($KEYP) 228 xor $Q1,$Q1,$T1 243 ld $T1,8($KEYP) [all …]
|
| H A D | aes-riscv32-zkn.pl | 193 my ($T0,$T1,$T2,$T3) = use_regs(13..16); 260 lw $T1,4($KEYP) 269 xor $Q1,$Q1,$T1 280 lw $T1,4($KEYP) 285 @{[aes32esmi4 $T1,$Q1,$Q2,$Q3,$Q0]} 297 @{[aes32esmi4 $Q0,$T0,$T1,$T2,$T3]} 298 @{[aes32esmi4 $Q1,$T1,$T2,$T3,$T0]} 299 @{[aes32esmi4 $Q2,$T2,$T3,$T0,$T1]} 300 @{[aes32esmi4 $Q3,$T3,$T0,$T1,$T2]} 310 lw $T1,4($KEYP) [all …]
|
| H A D | aes-riscv64.pl | 84 my ($T0,$T1,$T2,$T3) = use_regs(13..16); 267 [$T0,$T1,$T2,$T3], # Temporaries 279 [$T0,$T1,$T2,$T3], 291 [$T0,$T1,$T2,$T3], 301 [$T0,$T1,$T2,$T3], 303 [$T0,$T1,$T2,$T3], 321 xor $T1,$T1,$T5 340 xor $Q1,$T1,$T5 367 [$T0,$T1,$T2,$T3], 374 [$T0,$T1,$T2,$T3], [all …]
|
| /freebsd/crypto/openssl/crypto/poly1305/asm/ |
| H A D | poly1305-x86.pl | 392 my ($D0,$D1,$D2,$D3,$D4,$T0,$T1,$T2)=map("xmm$_",(0..7)); 430 &movdqa ($T1,$D1); 432 &pslld ($T1,2); 434 &paddd ($T1,$D1); # *5 436 &movdqa (&QWP(16*5,"esp"),$T1); 438 &movdqa ($T1,$D3); 440 &pslld ($T1,2); 442 &paddd ($T1,$D3); # *5 444 &movdqa (&QWP(16*7,"esp"),$T1); 447 &pshufd ($T1,$D0,0b01000100); [all …]
|
| H A D | poly1305-x86_64.pl | 346 my ($H0,$H1,$H2,$H3,$H4, $T0,$T1,$T2,$T3,$T4, $D0,$D1,$D2,$D3,$D4, $MASK) = 815 vmovdqu 16*3($inp),$T1 819 vpsrldq \$6,$T1,$T3 820 vpunpckhqdq $T1,$T0,$T4 # 4 821 vpunpcklqdq $T1,$T0,$T0 # 0:1 825 vpsrlq \$26,$T0,$T1 828 vpand $MASK,$T1,$T1 # 1 908 vpmuludq $T1,$D4,$D1 # d1 = h1*r0 923 vpmuludq $T1,$H2,$H1 # h1*r1 934 vpmuludq $T1,$H3,$H1 # h1*r2 [all …]
|
| /freebsd/crypto/openssl/crypto/modes/asm/ |
| H A D | ghash-x86_64.pl | 463 ($T1,$T2,$T3)=("%xmm3","%xmm4","%xmm5"); 471 pshufd \$0b01001110,$Xi,$T1 473 pxor $Xi,$T1 # 479 pshufd \$0b01001110,$Xi,$T1 480 pxor $Xi,$T1 # 486 pclmulqdq \$0x00,$HK,$T1 ####### 487 pxor $Xi,$T1 # 488 pxor $Xhi,$T1 # 490 movdqa $T1,$T2 # 491 psrldq \$8,$T1 [all …]
|
| H A D | aesni-gcm-x86_64.pl | 84 ($Ii,$T1,$T2,$Hkey, 102 vpaddb $T2,$T1,$inout1 107 vpxor $rndkey,$T1,$inout0 116 vpaddb $T2,$inout5,$T1 # next counter value 121 vmovdqu $T1,($ivp) # save next counter value 132 vpclmulqdq \$0x00,$Hkey,$Z3,$T1 145 vpxor $Z1,$T1,$Z0 148 vpclmulqdq \$0x10,$Hkey,$Ii,$T1 169 vpxor $T1,$Z2,$Z2 170 vpclmulqdq \$0x00,$Z1,$Ii,$T1 [all …]
|
| H A D | aes-gcm-riscv64-zvkb-zvkg-zvkned.pl | 80 my ($T0, $T1, $T2, $T3) = ("t0", "t1", "t2", "t3"); 419 addi $T1, $XIP, 32 421 @{[vle32_v $V31, $T1]} 424 li $T1, 1 427 slli $T1, $T1, 1 428 bltu $T1, $T0, 1b 476 srli $T1, $PADDING_LEN32, 2 483 @{[vsub_vx $V12, $V12, $T1, $CTR_MASK]} 512 addi $T1, $PADDING_LEN32, 4 513 @{[vsetvli "zero", $T1, "e32", "m4", "tu", "mu"]} [all …]
|
| H A D | ghash-x86.pl | 839 ($T1,$T2,$T3)=("xmm3","xmm4","xmm5"); 848 &pshufd ($T1,$Xi,0b01001110); 850 &pxor ($T1,$Xi); # 856 &pclmulqdq ($T1,$HK,0x00); ####### 857 &xorps ($T1,$Xi); # 858 &xorps ($T1,$Xhi); # 860 &movdqa ($T2,$T1); # 861 &psrldq ($T1,8); 863 &pxor ($Xhi,$T1); 875 &movdqa ($T1,$Xi); # [all …]
|
| /freebsd/contrib/llvm-project/libcxx/include/ |
| H A D | utility | 73 template <class T1, class T2> 76 typedef T1 first_type; 79 T1 first; 85 … explicit(see-below) pair(const T1& x, const T2& y); // constexpr in C++14 86 … template <class U = T1, class V = T2> explicit(see-below) pair(U&&, V&&); // constexpr in C++14 101 pair& operator=(pair&& p) noexcept(is_nothrow_move_assignable<T1>::value && 110 void swap(pair& p) noexcept(is_nothrow_swappable_v<T1> && 115 template<class T1, class T2, class U1, class U2, template<class> class TQual, template<class> class… 116 struct basic_common_reference<pair<T1, T2>, pair<U1, U2>, TQual, UQual>; // since C++23 118 template<class T1, class T2, class U1, class U2> [all …]
|
| /freebsd/crypto/openssl/crypto/seed/ |
| H A D | seed_local.h | 56 #define KEYSCHEDULE_UPDATE0(T0, T1, X1, X2, X3, X4, KC) \ argument 61 (T1) = ((X2) + (KC) - (X4)) & 0xffffffff 63 #define KEYSCHEDULE_UPDATE1(T0, T1, X1, X2, X3, X4, KC) \ argument 68 (T1) = ((X2) + (KC) - (X4)) & 0xffffffff 70 #define KEYUPDATE_TEMP(T0, T1, K) \ argument 72 (K)[1] = G_FUNC((T1)) 98 #define E_SEED(T0, T1, X1, X2, X3, X4, rbase) \ argument 100 (T1) = (X4) ^ (ks->data)[(rbase) + 1]; \ 101 (T1) ^= (T0); \ 102 (T1) = G_FUNC((T1)); \ [all …]
|
| /freebsd/crypto/openssl/crypto/sha/asm/ |
| H A D | sha512-sparcv9.pl | 118 $T1="%g2"; 148 ldx [$inp+64],$T1 158 srlx $T1,$tmp32,$T1 159 or $T1,@X[7],@X[7] 165 $code.="\tadd @X[$i/2],$h,$T1\n"; 167 $code.="\tsrlx @X[$i/2],32,$T1\n\tadd $h,$T1,$T1\n"; 196 add $h,$tmp2,$T1 215 add $h,$tmp2,$T1 230 $code.="\tadd $h,$T1,$T1\n"; 250 add $tmp2,$T1,$T1 [all …]
|
| H A D | sha256-c64xplus.pl | 36 ($E,$Ectx,$F,$Fctx,$G,$Gctx,$H,$Hctx,$T1,$S1,$s0,$t0e,$t1e,$t2e,$X1,$X15) 137 || ADD $K,$H,$T1 ; T1 = h + K256[i] 138 ADD $X0,$T1,$T1 ; T1 += X[i]; 145 || ADD $Ch,$T1,$T1 ; T1 += Ch(e,f,g) 146 ADD $S1,$T1,$T1 ; T1 += Sigma1(e) 154 || ADD $D,$T1,$E ; e = d + T1 158 || ADD $T1,$T2,$A ; a = T1 + T2 176 || ADD $K,$H,$T1 ; T1 = h + K256[i] 177 ADD $X0,$T1,$T1 ; T1 += X[i]; 184 || ADD $Ch,$T1,$T1 ; T1 += Ch(e,f,g) [all …]
|
| H A D | sha512-s390x.pl | 74 $T1="%r14"; 117 $LD $T1,`$i*$SZ`($inp) ### $i 126 $ST $T1,`$stdframe+$SZ*($i%16)`($sp) 128 algr $T1,$h # T1+=h 131 algr $T1,$t0 # T1+=Sigma1(e) 134 $ADD $T1,`$i*$SZ`($len,$tbl) # T1+=K[i] 136 algr $T1,$t2 # T1+=Ch(e,f,g) 144 algr $h,$T1 # h+=T1 146 algr $d,$T1 # d+=T1 155 $LD $T1,`$stdframe+$SZ*(($i+1)%16)`($sp) ### $i [all …]
|
| /freebsd/crypto/openssl/crypto/sha/ |
| H A D | sha256.c | 219 unsigned MD32_REG_T a, b, c, d, e, f, g, h, s0, s1, T1, T2; in sha256_block_data_order() local 237 T1 = X[i] = l; in sha256_block_data_order() 238 T1 += h + Sigma1(e) + Ch(e, f, g) + K256[i]; in sha256_block_data_order() 243 e = d + T1; in sha256_block_data_order() 247 a = T1 + T2; in sha256_block_data_order() 256 T1 = X[i & 0xf] += s0 + s1 + X[(i + 9) & 0xf]; in sha256_block_data_order() 257 T1 += h + Sigma1(e) + Ch(e, f, g) + K256[i]; in sha256_block_data_order() 262 e = d + T1; in sha256_block_data_order() 266 a = T1 + T2; in sha256_block_data_order() 284 T1 += h + Sigma1(e) + Ch(e, f, g) + K256[i]; \ [all …]
|
| H A D | sha512.c | 649 SHA_LONG64 a, b, c, d, e, f, g, h, s0, s1, T1, T2; in sha512_block_data_order() local 666 T1 = X[i] = W[i]; in sha512_block_data_order() 668 T1 = X[i] = PULL64(W[i]); in sha512_block_data_order() 670 T1 += h + Sigma1(e) + Ch(e, f, g) + K512[i]; in sha512_block_data_order() 675 e = d + T1; in sha512_block_data_order() 679 a = T1 + T2; in sha512_block_data_order() 688 T1 = X[i & 0xf] += s0 + s1 + X[(i + 9) & 0xf]; in sha512_block_data_order() 689 T1 += h + Sigma1(e) + Ch(e, f, g) + K512[i]; in sha512_block_data_order() 694 e = d + T1; in sha512_block_data_order() 698 a = T1 + T2; in sha512_block_data_order() [all …]
|
| /freebsd/sys/contrib/libsodium/src/libsodium/crypto_onetimeauth/poly1305/sse2/ |
| H A D | poly1305_sse2.c | 207 xmmi T0, T1, T2, T3, T4, T5, T6, T7, T8; in poly1305_blocks() local 240 T1 = _mm_loadu_si128((const xmmi *) (const void *) &st->H.hh[4]); in poly1305_blocks() 244 H2 = _mm_shuffle_epi32(T1, _MM_SHUFFLE(1, 1, 0, 0)); in poly1305_blocks() 245 H3 = _mm_shuffle_epi32(T1, _MM_SHUFFLE(3, 3, 2, 2)); in poly1305_blocks() 254 T1 = _mm_cvtsi32_si128(st->R2[4]); in poly1305_blocks() 257 R24 = _mm_unpacklo_epi64(T1, T3); in poly1305_blocks() 261 T1 = _mm_cvtsi32_si128(st->R[4]); in poly1305_blocks() 265 R24 = T1; in poly1305_blocks() 274 T1 = _mm_cvtsi32_si128(st->R2[4]); in poly1305_blocks() 279 R24 = _mm_shuffle_epi32(T1, _MM_SHUFFLE(0, 0, 0, 0)); in poly1305_blocks() [all …]
|
| /freebsd/crypto/openssl/crypto/chacha/asm/ |
| H A D | chacha-riscv64-v-zbb.pl | 88 my ( $T1 ) = ( "t4" ); 269 slli $T1, $VL, 6 270 bltu $T1, $LEN, 1f 273 addi $T1, $VL, -1 274 @{[vsetvli $VL, $T1, "e32", "m1", "ta", "ma"]} 361 addi $T1, $INPUT, 32 362 @{[vlsseg_nf_e32_v 8, $V24, $T1, $T0]} 372 li $T1, 0x3320646e 373 @{[vadd_vx $V1, $V1, $T1]} 374 add $STATE1, $STATE1, $T1 [all …]
|
| /freebsd/contrib/llvm-project/compiler-rt/lib/sanitizer_common/ |
| H A D | sanitizer_syscall_linux_x86_64.inc | 22 template <typename T1> 23 static uptr internal_syscall(u64 nr, T1 arg1) { 30 template <typename T1, typename T2> 31 static uptr internal_syscall(u64 nr, T1 arg1, T2 arg2) { 38 template <typename T1, typename T2, typename T3> 39 static uptr internal_syscall(u64 nr, T1 arg1, T2 arg2, T3 arg3) { 46 template <typename T1, typename T2, typename T3, typename T4> 47 static uptr internal_syscall(u64 nr, T1 arg1, T2 arg2, T3 arg3, T4 arg4) { 56 template <typename T1, typename T2, typename T3, typename T4, typename T5> 57 static uptr internal_syscall(u64 nr, T1 arg1, T2 arg2, T3 arg3, T4 arg4, [all …]
|
| /freebsd/contrib/ldns/ |
| H A D | sha2.c | 332 T1 = (h) + Sigma1_256(e) + Ch((e), (f), (g)) + \ 334 (d) += T1; \ 335 (h) = T1 + Sigma0_256(a) + Maj((a), (b), (c)); \ 342 T1 = (h) + Sigma1_256(e) + Ch((e), (f), (g)) + \ 344 (d) += T1; \ 345 (h) = T1 + Sigma0_256(a) + Maj((a), (b), (c)); \ 355 T1 = (h) + Sigma1_256(e) + Ch((e), (f), (g)) + K256[j] + \ 357 (d) += T1; \ 358 (h) = T1 + Sigma0_256(a) + Maj((a), (b), (c)); \ 364 sha2_word32 T1, *W256; in ldns_sha256_Transform() local [all …]
|
| /freebsd/crypto/openssl/crypto/bn/asm/ |
| H A D | rsaz-2k-avxifma.pl | 119 my ($T0,$T0h,$T1,$T1h,$T2,$tmp) = ("$zero", "$Bi", "$Yi", map("%ymm$_", (13..15))); 212 vpsrlq \$52, $_R1, $T1 222 vpermq \$3, $T1, $tmp 225 vpermq \$144, $T1, $T1 227 vblendpd \$1, $tmp, $T1, $T1 246 vpaddq $T1, $_R1, $_R1 254 vpcmpgtq .Lmask52x4(%rip), $_R1, $T1 259 vmovmskpd $T1, %r12d 266 vpcmpeqq .Lmask52x4(%rip), $_R1, $T1 271 vmovmskpd $T1, %ebx [all …]
|
| H A D | rsaz-3k-avxifma.pl | 104 my ($T0,$T0h,$T1,$T1h,$T2,$T2h,$T3,$T3h) = ("$zero", "$Bi", "$Yi", map("%ymm$_", (11..15))); 217 vpsrlq \$52, $_R1, $T1 245 vpermq \$3, $T1, $_R0 248 vpermq \$144, $T1, $T1 250 vblendpd \$1, $_R0, $T1, $T1 275 vpaddq $T1, $_R1, $_R1 291 vpcmpgtq .Lmask52x4(%rip),${_R1},$T1 293 vmovmskpd $T1, %r13d 325 vpcmpeqq .Lmask52x4(%rip),${_R1},$T1 327 vmovmskpd $T1, %r8d [all …]
|
| /freebsd/crypto/openssh/openbsd-compat/ |
| H A D | sha2.c | 370 T1 = (h) + Sigma1_256((e)) + Ch((e), (f), (g)) + K256[j] + W256[j]; \ 371 (d) += T1; \ 372 (h) = T1 + Sigma0_256((a)) + Maj((a), (b), (c)); \ 381 T1 = (h) + Sigma1_256((e)) + Ch((e), (f), (g)) + K256[j] + \ 383 (d) += T1; \ 384 (h) = T1 + Sigma0_256((a)) + Maj((a), (b), (c)); \ 392 u_int32_t T1, W256[16]; in SHA256Transform() local 441 a = b = c = d = e = f = g = h = T1 = 0; in SHA256Transform() 450 u_int32_t T1, T2, W256[16]; in SHA256Transform() local 468 T1 = h + Sigma1_256(e) + Ch(e, f, g) + K256[j] + W256[j]; in SHA256Transform() [all …]
|
| /freebsd/contrib/llvm-project/clang/lib/AST/ |
| H A D | ASTStructuralEquivalence.cpp | 99 QualType T1, QualType T2); 804 QualType T1, QualType T2) { in IsStructurallyEquivalent() argument 805 if (T1.isNull() || T2.isNull()) in IsStructurallyEquivalent() 806 return T1.isNull() && T2.isNull(); in IsStructurallyEquivalent() 808 QualType OrigT1 = T1; in IsStructurallyEquivalent() 814 T1 = Context.FromCtx.getCanonicalType(T1); in IsStructurallyEquivalent() 818 if (T1.getQualifiers() != T2.getQualifiers()) in IsStructurallyEquivalent() 821 Type::TypeClass TC = T1->getTypeClass(); in IsStructurallyEquivalent() 823 if (T1->getTypeClass() != T2->getTypeClass()) { in IsStructurallyEquivalent() 826 if (T1->getTypeClass() == Type::FunctionProto && in IsStructurallyEquivalent() [all …]
|
| /freebsd/crypto/openssl/crypto/aria/ |
| H A D | aria.c | 338 #define ARIA_ADD_ROUND_KEY(RK, T0, T1, T2, T3) \ argument 341 (T1) ^= (RK)->u[1]; \ 347 #define ARIA_SBOX_LAYER1_WITH_PRE_DIFF(T0, T1, T2, T3) … argument 350 …(T1) = S1[GET_U8_BE(T1, 0)] ^ S2[GET_U8_BE(T1, 1)] ^ X1[GET_U8_BE(T1, 2)] ^ X2[GET_U8_BE(T1, 3)]; \ 356 #define ARIA_SBOX_LAYER2_WITH_PRE_DIFF(T0, T1, T2, T3) … argument 359 …(T1) = X1[GET_U8_BE(T1, 0)] ^ X2[GET_U8_BE(T1, 1)] ^ S1[GET_U8_BE(T1, 2)] ^ S2[GET_U8_BE(T1, 3)]; \ 365 #define ARIA_DIFF_WORD(T0, T1, T2, T3) \ argument 367 (T1) ^= (T2); \ 369 (T0) ^= (T1); \ 371 (T3) ^= (T1); \ [all …]
|