Home
last modified time | relevance | path

Searched full:t4 (Results 1 – 25 of 296) sorted by relevance

12345678910>>...12

/freebsd/contrib/bearssl/src/symcipher/
H A Daes_x86ni.c64 __m128i t4; in expand_step192() local
67 t4 = _mm_slli_si128(*t1, 0x4); in expand_step192()
68 *t1 = _mm_xor_si128(*t1, t4); in expand_step192()
69 t4 = _mm_slli_si128(t4, 0x4); in expand_step192()
70 *t1 = _mm_xor_si128(*t1, t4); in expand_step192()
71 t4 = _mm_slli_si128(t4, 0x4); in expand_step192()
72 *t1 = _mm_xor_si128(*t1, t4); in expand_step192()
75 t4 = _mm_slli_si128(*t3, 0x4); in expand_step192()
76 *t3 = _mm_xor_si128(*t3, t4); in expand_step192()
84 __m128i t4; in expand_step256_1() local
[all …]
/freebsd/share/misc/
H A Dscsi_modes62 {Queue Algorithm Modifier} t4
85 {Reserved} *t4
90 {Reserved} *t4
91 {Initial Command Priority} t4
125 {Reserved} *t4
192 {Demand Retention Priority} t4
193 {Write Retention Priority} t4
206 {Reserved} *t4
226 {Reserved} *t4
227 {SPC} t4
[all …]
/freebsd/crypto/openssl/crypto/bn/asm/
H A Dppc64-mont.pl143 $t4="r28";
299 extrdi $t4,$t7,16,48
303 std $t4,`$FRAME+32`($sp)
312 lwz $t4,`4^$LITTLE_ENDIAN`($np) ; load n[0] as 32-bit word pair
327 mullw $t4,$a0,$t1 ; mulld ap[0]*bp[0]
343 mullw $t0,$t4,$n0 ; mulld tp[0]*n0
344 mulhwu $t1,$t4,$n0
346 mullw $t3,$t4,$n1
350 extrwi $t4,$t0,16,16
354 std $t4,`$FRAME+32`($sp) ; yes, std in 32-bit build
[all …]
/freebsd/sys/contrib/libsodium/src/libsodium/crypto_onetimeauth/poly1305/sse2/
H A Dpoly1305_sse2.c207 xmmi T0, T1, T2, T3, T4, T5, T6, T7, T8; in poly1305_blocks() local
255 T4 = _mm_unpacklo_epi32(T0, T2); in poly1305_blocks()
263 T4 = _mm_unpacklo_epi32(T0, T2); in poly1305_blocks()
267 R20 = _mm_shuffle_epi32(T4, _MM_SHUFFLE(1, 1, 0, 0)); in poly1305_blocks()
268 R21 = _mm_shuffle_epi32(T4, _MM_SHUFFLE(3, 3, 2, 2)); in poly1305_blocks()
338 T4 = H4; in poly1305_blocks()
339 T4 = _mm_mul_epu32(T4, T14); in poly1305_blocks()
368 T4 = _mm_add_epi64(T4, v41); in poly1305_blocks()
385 T4 = _mm_add_epi64(T4, v42); in poly1305_blocks()
393 T4 = _mm_add_epi64(T4, v43); in poly1305_blocks()
[all …]
/freebsd/contrib/bearssl/src/ec/
H A Dec_c25519_m64.c174 uint64_t t0, t1, t2, t3, t4, cc; in f255_sub() local
191 t4 = 1 + (uint64_t)(z >> 64); in f255_sub()
199 cc = (38 & -t4) + (19 & -(t3 >> 63)); in f255_sub()
218 uint64_t t0, t1, t2, t3, t4; in f255_sub()
225 (void)_subborrow_u64(k, 1, 0, &t4); in f255_sub()
231 (void)_subborrow_u64(k, t4, 0, &t4); in f255_sub()
239 t4 = (38 & -t4) + (19 & -(t3 >> 63)); in f255_sub()
241 k = _addcarry_u64(0, t0, t4, &d[0]); in f255_sub()
258 uint64_t t0, t1, t2, t3, t4, t5, t6, t7, th; in f255_mul() local
271 t4 = (uint64_t)(z >> 64); in f255_mul()
[all …]
H A Dec_p256_m64.c263 uint64_t x, f, t0, t1, t2, t3, t4; in f256_montymul() local
295 t4 = (uint64_t)(z >> 64); in f256_montymul()
312 z = t4 + (z >> 64); in f256_montymul()
314 t4 = (uint64_t)(z >> 64); in f256_montymul()
329 t4 += (uint64_t)(z >> 64); in f256_montymul()
345 * t4 can be only 0 or 1 in f256_montymul()
347 * We can therefore subtract p from t, conditionally on t4, to in f256_montymul()
350 z = (unsigned __int128)t0 + t4; in f256_montymul()
352 z = (unsigned __int128)t1 - (t4 << 32) + (z >> 64); in f256_montymul()
356 t3 = t3 - (uint64_t)(z >> 127) - t4 + (t4 << 32); in f256_montymul()
[all …]
H A Dec_p256_m62.c777 uint64_t t1[5], t2[5], t3[5], t4[5]; in p256_double() local
815 f256_montymul(t4, P->y, P->z); in p256_double()
816 f256_add(P->z, t4, t4); in p256_double()
825 f256_montysquare(t4, t3); in p256_double()
826 f256_add(t4, t4, t4); in p256_double()
827 f256_sub(P->y, P->y, t4); in p256_double()
879 uint64_t t1[5], t2[5], t3[5], t4[5], t5[5], t6[5], t7[5], tt; in p256_add() local
887 f256_montymul(t4, P2->z, t3); in p256_add()
888 f256_montymul(t3, P1->y, t4); in p256_add()
891 * Compute u2 = x2*z1^2 (in t2) and s2 = y2*z1^3 (in t4). in p256_add()
[all …]
H A Dec_prime_i31.c182 #define t4 9 macro
252 MMUL(t4, Py, Pz),
253 MSET(Pz, t4),
254 MADD(Pz, t4),
262 MMUL(t4, t3, t3),
263 MSUB(Py, t4),
264 MSUB(Py, t4),
325 MMUL(t4, P2z, t3),
326 MMUL(t3, P1y, t4),
329 * Compute u2 = x2*z1^2 (in t2) and s2 = y2*z1^3 (in t4).
[all …]
H A Dec_prime_i15.c183 #define t4 9 macro
253 MMUL(t4, Py, Pz),
254 MSET(Pz, t4),
255 MADD(Pz, t4),
263 MMUL(t4, t3, t3),
264 MSUB(Py, t4),
265 MSUB(Py, t4),
326 MMUL(t4, P2z, t3),
327 MMUL(t3, P1y, t4),
330 * Compute u2 = x2*z1^2 (in t2) and s2 = y2*z1^3 (in t4).
[all …]
H A Dec_p256_m31.c762 uint32_t t1[9], t2[9], t3[9], t4[9]; in p256_double() local
800 mul_f256(t4, Q->y, Q->z); in p256_double()
801 add_f256(Q->z, t4, t4); in p256_double()
809 square_f256(t4, t3); in p256_double()
810 add_f256(t4, t4, t4); in p256_double()
811 sub_f256(Q->y, Q->y, t4); in p256_double()
861 uint32_t t1[9], t2[9], t3[9], t4[9], t5[9], t6[9], t7[9]; in p256_add() local
870 mul_f256(t4, P2->z, t3); in p256_add()
871 mul_f256(t3, P1->y, t4); in p256_add()
874 * Compute u2 = x2*z1^2 (in t2) and s2 = y2*z1^3 (in t4). in p256_add()
[all …]
/freebsd/crypto/openssl/crypto/ec/asm/
H A Decp_sm2p256-armv8.pl26 my ($t4,$t5,$t6,$t7,$t8)=map("x$_",(15..17,19,20));
42 adc $t4,xzr,xzr
61 sbcs $t4,$t4,xzr
89 sbc $t4,xzr,xzr
108 tst $t4,$t4
272 adcs $t4,xzr,xzr
288 sbcs $t4,$t4,xzr
294 eor $t4,$t4,$t4
303 adcs $t4,xzr,xzr
319 sbcs $t4,$t4,xzr
[all …]
H A Decp_nistp521-ppc64.pl150 my ($t1, $t2, $t3, $t4) = ("v33", "v34", "v44", "v54");
189 xxpermdi $t4,$in2[1],$in2[0],0b00
190 vmsumudm $out[3],$t3,$t4,$out[3]
194 xxpermdi $t4,$in2[2],$in2[1],0b00
195 vmsumudm $out[4],$t3,$t4,$out[4]
200 xxpermdi $t4,$in2[3],$in2[2],0b00
201 vmsumudm $out[5],$t3,$t4,$out[5]
205 xxpermdi $t4,$in2[4],$in2[3],0b00
206 vmsumudm $out[6],$t3,$t4,$out[6]
210 xxpermdi $t4,$in2[5],$in2[4],0b00
[all …]
H A Decp_nistz256-x86_64.pl119 my ($t0,$t1,$t2,$t3,$t4)=("%rax","%rdx","%rcx","%r12","%r13");
136 xor $t4,$t4
147 adc \$0, $t4
155 sbb \$0, $t4
198 xor $t4, $t4
205 adc \$0, $t4
213 cmovz $a_ptr, $t4
227 shl \$63, $t4
229 or $t4, $a3
261 xor $t4, $t4
[all …]
/freebsd/crypto/openssl/crypto/aes/asm/
H A Daes-riscv32-zkn.pl474 my ($T0,$T1,$T2,$T3,$T4,$T5,$T6,$T7,$T8) = use_regs(13..17,28..31);
517 # use T4 to store rcon
518 li $T4,$rcon[$rnum]
522 xor $T0,$T0,$T4
523 # use T4 to store rotated T3
528 @{[rori $T4,$T3,8]}
532 srli $T4,$T3,8
534 or $T4,$T4,$T5
539 @{[fwdsbox4 $T0,$T4]}
566 lw $T4,16($UKEY)
[all …]
H A Daes-riscv64.pl85 my ($T4,$T5,$T6,$T7,$T8,$T9,$T10,$T11) = use_regs(17..24);
265 [$T4,$T5,$T6,$T7], # Destination registers
310 xor $T4,$T4,$T8
315 xor $T4,$T4,$T12
320 xor $T0,$T0,$T4
329 ld $T4,0($KEYP)
337 xor $Q0,$T0,$T4
339 srli $T5,$T4,32
365 [$T4,$T5,$T6,$T7],
412 xor $T4,$T4,$T0
[all …]
H A Daes-riscv64-zkn.pl282 my ($T0,$T1,$T2,$T3,$T4) = use_regs(6..8,13..14);
364 @{[aes64ks1i $T4,$T3,$rnum]}
365 @{[aes64ks2 $T0,$T4,$T0]}
375 @{[aes64ks1i $T4,$T1,0xA]}
376 @{[aes64ks2 $T2,$T4,$T2]}
537 @{[aes64im $T4,$T2]}
538 sd $T4,16($KEYP)
539 @{[aes64im $T4,$T3]}
540 sd $T4,24($KEYP)
544 @{[aes64ks1i $T4,$T3,$rnum]}
[all …]
H A Daes-mips.pl125 my ($t0,$t1,$t2,$t3,$t4,$t5,$t6,$t7,$t8,$t9,$t10,$t11) = map("\$$_",(12..23));
164 lwxs $t4,$i0($Tbl) # Te2[s2>>8]
187 rotr $t4,$t4,16
192 xor $t0,$t4
193 lwxs $t4,$i0($Tbl) # Te0[s0>>24]
215 xor $t0,$t4
247 ext $t4,$s2,8,8
254 $PTR_INS $i0,$t4,2,8
299 lw $t4,0($i0) # Te2[s2>>8]
321 lw $t4,0($i0) # Te2[s2>>8]
[all …]
/freebsd/crypto/openssl/crypto/poly1305/asm/
H A Dpoly1305-x86_64.pl346 my ($H0,$H1,$H2,$H3,$H4, $T0,$T1,$T2,$T3,$T4, $D0,$D1,$D2,$D3,$D4, $MASK) =
820 vpunpckhqdq $T1,$T0,$T4 # 4
824 vpsrlq \$40,$T4,$T4 # 4
832 vpor 32(%rcx),$T4,$T4 # padbit, yes, always
913 vpmuludq $T4,$D4,$D4 # d4 = h4*r0
916 vpmuludq 0x20(%rsp),$T4,$H0 # h4*s1
939 vpmuludq $T4,$H4,$H0 # h4*s2
950 vpmuludq $T4,$H3,$H0 # h4*s3
962 vpmuludq $T4,$H4,$T4 # h4*s4
965 vpaddq $T4,$D3,$D3 # d3 += h4*s4
[all …]
H A Dpoly1305-s390x.pl230 my ($T1, $T2, $T3, $T4) = map("%v$_",(24..27));
512 vlm ($T1,$T4,"0x00($inp)"); # load first input block
517 vperm ($I0,$T3,$T4,$bswaplo);
518 vperm ($I2,$T3,$T4,$bswapmi);
519 vperm ($T3,$T3,$T4,$bswaphi);
532 vperm ($T4,$T1,$T2,$bswaphi);
538 verimg ($I3,$T4,$mask26,50); # >>14
539 vesrlg ($T4,$T4,40);
540 vo ($I4,$I4,$T4);
571 vlm ($T1,$T4,"0x00($inp)"); # load next input block
[all …]
/freebsd/sys/riscv/riscv/
H A Dlocore.S154 li t4, (PTE_KERN)
155 or t4, t4, s8 /* t4 |= pte bits */
157 or t6, t4, s2
175 li t4, PTE_V
177 or t6, t4, t5
187 srli t4, s9, L2_SHIFT /* Div physmem base by 2 MiB */
189 add t3, t4, t2
193 slli t2, t4, PTE_PPN1_S /* << PTE_PPN1_S */
198 addi t4, t4, 1
199 bltu t4, t3, 1b
/freebsd/contrib/llvm-project/llvm/lib/Target/Hexagon/
H A DHexagonVectorLoopCarriedReuse.h23 // t4 = f(t3);
24 // t5 = g(t4);
35 // t4 = f(t3);
36 // t5 = g(t4);
52 // t4 = f(X1) <-- I2
53 // t5 = g(t4)
78 // Y2 = PHI<(Y0, LoopPreheader), (t4, Loop)>
83 // t4 = f(X1)
84 // t5 = g(t4)
/freebsd/sys/riscv/vmm/
H A Dvmm_switch.S70 la t4, .Lswitch_return
77 csrrw t4, stvec, t4
84 sd t4, (HYP_H_STVEC)(a0)
94 ld t4, (HYP_G_T + 4 * 8)(a0)
134 sd t4, (HYP_G_T + 4 * 8)(a0)
161 ld t4, (HYP_H_HSTATUS)(a0)
168 csrrw t4, hstatus, t4
174 sd t4, (HYP_G_HSTATUS)(a0)
/freebsd/crypto/openssl/crypto/sha/asm/
H A Dsha256-armv4.pl65 $inp="r1"; $t4="r1";
92 str $inp,[sp,#17*4] @ make room for $t4
109 str $inp,[sp,#17*4] @ make room for $t4
141 ldr $t4,[sp,#`($i+15)%16`*4] @ from future BODY_16_xx
158 @ ldr $t4,[sp,#`($i+14)%16`*4]
161 mov $t2,$t4,ror#$sigma1[0]
163 eor $t2,$t2,$t4,ror#$sigma1[1]
166 eor $t2,$t2,$t4,lsr#$sigma1[2] @ sigma1(X[i+14])
167 ldr $t4,[sp,#`($i+9)%16`*4]
173 add $t1,$t1,$t4 @ X[i]
[all …]
/freebsd/contrib/llvm-project/compiler-rt/lib/sanitizer_common/
H A Dsanitizer_syscall_linux_x86_64.inc46 template <typename T1, typename T2, typename T3, typename T4>
47 static uptr internal_syscall(u64 nr, T1 arg1, T2 arg2, T3 arg3, T4 arg4) {
56 template <typename T1, typename T2, typename T3, typename T4, typename T5>
57 static uptr internal_syscall(u64 nr, T1 arg1, T2 arg2, T3 arg3, T4 arg4,
68 template <typename T1, typename T2, typename T3, typename T4, typename T5,
70 static uptr internal_syscall(u64 nr, T1 arg1, T2 arg2, T3 arg3, T4 arg4,
/freebsd/lib/libc/riscv/string/
H A Dmemchr.S117 not t4, t3
121 and t4, t4, t1
123 and t3, t3, t4

12345678910>>...12