Home
last modified time | relevance | path

Searched +full:1 +full:v0 (Results 1 – 25 of 548) sorted by relevance

12345678910>>...22

/freebsd/sys/crypto/openssl/aarch64/
H A Dghashv8-armx.S17 dup v17.4s,v17.s[1]
22 shl v3.2d,v3.2d,#1
25 orr v3.16b,v3.16b,v18.16b //H<<<=1
31 pmull v0.1q,v20.1d,v20.1d
33 pmull2 v2.1q,v20.2d,v20.2d
34 pmull v1.1q,v16.1d,v16.1d
36 ext v17.16b,v0.16b,v2.16b,#8 //Karatsuba post-processing
37 eor v18.16b,v0.16b,v2.16b
40 pmull v18.1q,v0.1d,v19.1d //1st phase
42 ins v2.d[0],v1.d[1]
[all …]
H A Daesv8-armx.S22 mov x3,#-1
38 eor v0.16b,v0.16b,v0.16b
50 ext v5.16b,v0.16b,v3.16b,#12
52 aese v6.16b,v0.16b
53 subs w1,w1,#1
56 ext v5.16b,v0.16b,v5.16b,#12
58 ext v5.16b,v0.16b,v5.16b,#12
61 shl v1.16b,v1.16b,#1
68 ext v5.16b,v0.16b,v3.16b,#12
70 aese v6.16b,v0.16b
[all …]
H A Daes-gcm-armv8_64.S41 sub x5, x5, #1 //byte_len - 1
48 fmov d1, x10 //CTR block 1
51 add w12, w12, #1 //increment rev_ctr32
55 rev w9, w12 //CTR block 1
56 add w12, w12, #1 //CTR block 1
59 orr x9, x11, x9, lsl #32 //CTR block 1
60 …ld1 { v0.16b}, [x16] //special case vector load initial counter so we …
62 fmov v1.d[1], x9 //CTR block 1
67 add w12, w12, #1 //CTR block 2
69 fmov v2.d[1], x9 //CTR block 2
[all …]
H A Dvpaes-armv8.S137 ushr v0.16b, v7.16b, #4 // vpsrlb $4, %xmm0, %xmm0
140 tbl v2.16b, {v21.16b}, v0.16b // vpshufb %xmm0, %xmm3, %xmm2
141 eor v0.16b, v1.16b, v16.16b // vpxor %xmm5, %xmm1, %xmm0
142 eor v0.16b, v0.16b, v2.16b // vpxor %xmm2, %xmm0, %xmm0
151 tbl v0.16b, {v24.16b}, v3.16b // vpshufb %xmm3, %xmm12, %xmm0 # 0 = sb1t
154 eor v0.16b, v0.16b, v4.16b // vpxor %xmm4, %xmm0, %xmm0 # 0 = A
157 tbl v3.16b, {v0.16b}, v1.16b // vpshufb %xmm1, %xmm0, %xmm3 # 0 = B
159 tbl v0.16b, {v0.16b}, v4.16b // vpshufb %xmm4, %xmm0, %xmm0 # 3 = D
162 eor v0.16b, v0.16b, v3.16b // vpxor %xmm3, %xmm0, %xmm0 # 3 = 2A+B+D
163 and x11, x11, #~(1<<6) // and $0x30, %r11 # ... mod 4
[all …]
/freebsd/contrib/file/tests/
H A DMakefile.am6 android-vdex-1.result \
7 android-vdex-1.testfile \
40 gpkg-1-zst.result \
41 gpkg-1-zst.testfile \
80 keyman-1.result \
81 keyman-1.testfile \
141 zstd-dictionary-1.result \
147 zstd-v0.2-FF.result \
148 zstd-v0.2-FF.testfile \
149 zstd-v0.3-FF.result \
[all …]
/freebsd/crypto/openssl/crypto/aes/asm/
H A Dvpaes-ppc.pl62 $0 =~ m/(.*[\/\\])[^\/\\]+$/; $dir=$1;
246 vsrb v1, v0, v8 # vpsrlb \$4, %xmm0, %xmm0
247 vperm v0, $iptlo, $iptlo, v0 # vpshufb %xmm1, %xmm2, %xmm1
249 vxor v0, v0, v5 # vpxor %xmm5, %xmm1, %xmm0
250 vxor v0, v0, v1 # vpxor %xmm2, %xmm0, %xmm0
260 vperm v0, $sb1u, v7, v3 # vpshufb %xmm3, %xmm12, %xmm0 # 0 = sb1t
264 vxor v0, v0, v4 # vpxor %xmm4, %xmm0, %xmm0 # 0 = A
268 vperm v3, v0, v7, v1 # vpshufb %xmm1, %xmm0, %xmm3 # 0 = B
270 vperm v0, v0, v7, v4 # vpshufb %xmm4, %xmm0, %xmm0 # 3 = D
273 vxor v0, v0, v3 # vpxor %xmm3, %xmm0, %xmm0 # 3 = 2A+B+D
[all …]
H A Dvpaes-armv8.pl46 $0 =~ m/(.*[\/\\])[^\/\\]+$/; $dir=$1;
200 ushr v0.16b, v7.16b, #4 // vpsrlb \$4, %xmm0, %xmm0
203 tbl v2.16b, {$ipthi}, v0.16b // vpshufb %xmm0, %xmm3, %xmm2
204 eor v0.16b, v1.16b, v16.16b // vpxor %xmm5, %xmm1, %xmm0
205 eor v0.16b, v0.16b, v2.16b // vpxor %xmm2, %xmm0, %xmm0
214 tbl v0.16b, {$sb1u}, v3.16b // vpshufb %xmm3, %xmm12, %xmm0 # 0 = sb1t
217 eor v0.16b, v0.16b, v4.16b // vpxor %xmm4, %xmm0, %xmm0 # 0 = A
220 tbl v3.16b, {v0.16b}, v1.16b // vpshufb %xmm1, %xmm0, %xmm3 # 0 = B
222 tbl v0.16b, {v0.16b}, v4.16b // vpshufb %xmm4, %xmm0, %xmm0 # 3 = D
225 eor v0.16b, v0.16b, v3.16b // vpxor %xmm3, %xmm0, %xmm0 # 3 = 2A+B+D
[all …]
/freebsd/contrib/bearssl/src/rsa/
H A Drsa_i15_privexp.c33 * We want to invert e modulo phi = (p-1)(q-1). This first in br_rsa_i15_compute_privexp()
40 * values (phi/4, 1 and e) and calling moddiv, that requires in br_rsa_i15_compute_privexp()
52 * - We find small u, v such that u*e - v*r = 1 (using a in br_rsa_i15_compute_privexp()
65 uint32_t r, a, b, u0, v0, u1, v1, he, hr; in br_rsa_i15_compute_privexp() local
71 if (e < 3 || (e & 1) == 0) { in br_rsa_i15_compute_privexp()
85 || (pbuf[plen - 1] & 1) != 1) in br_rsa_i15_compute_privexp()
96 || (qbuf[qlen - 1] & 1) != 1) in br_rsa_i15_compute_privexp()
112 q = p + 1 + plen; in br_rsa_i15_compute_privexp()
117 * Compute phi = (p-1)*(q-1), then move it over p-1 and q-1 (that in br_rsa_i15_compute_privexp()
120 * p-1 and q-1, which is usually exact but may overshoot by one 1 in br_rsa_i15_compute_privexp()
[all …]
H A Drsa_i31_privexp.c33 * We want to invert e modulo phi = (p-1)(q-1). This first in br_rsa_i31_compute_privexp()
40 * values (phi/4, 1 and e) and calling moddiv, that requires in br_rsa_i31_compute_privexp()
52 * - We find small u, v such that u*e - v*r = 1 (using a in br_rsa_i31_compute_privexp()
65 uint32_t r, a, b, u0, v0, u1, v1, he, hr; in br_rsa_i31_compute_privexp() local
71 if (e < 3 || (e & 1) == 0) { in br_rsa_i31_compute_privexp()
85 || (pbuf[plen - 1] & 1) != 1) in br_rsa_i31_compute_privexp()
96 || (qbuf[qlen - 1] & 1) != 1) in br_rsa_i31_compute_privexp()
112 q = p + 1 + plen; in br_rsa_i31_compute_privexp()
117 * Compute phi = (p-1)*(q-1), then move it over p-1 and q-1 (that in br_rsa_i31_compute_privexp()
120 * p-1 and q-1, which is usually exact but may overshoot by one 1 in br_rsa_i31_compute_privexp()
[all …]
/freebsd/lib/libc/aarch64/string/
H A Dstrncmp.S21 subs x2, x2, #1
24 mov x13, #-1 // save constants for later
50 cmeq v5.16b, v0.16b, #0
66 tbl v0.16b, {v0.16b}, v4.16b
68 b 1f
72 1:
79 b 1f
86 1:
87 cmeq v2.16b, v0.16b, #0 // NUL byte present?
88 cmeq v4.16b, v0.16b, v4.16b // which bytes match?
[all …]
H A Dstrcmp.S21 mov x13, #-1
44 cmeq v5.16b, v0.16b, #0
60 tbl v0.16b, {v0.16b}, v4.16b
62 b 1f
66 1:
73 b 1f
80 1:
82 cmeq v2.16b, v0.16b, #0 // NUL byte present?
83 cmeq v4.16b, v0.16b, v4.16b // which bytes match?
102 cmeq v0.16b, v0.16b, v2.16b
[all …]
/freebsd/crypto/openssl/crypto/des/
H A Dcfb64ede.c29 register DES_LONG v0, v1; in DES_ede3_cfb64_encrypt() local
39 c2l(iv, v0); in DES_ede3_cfb64_encrypt()
42 ti[0] = v0; in DES_ede3_cfb64_encrypt()
43 ti[1] = v1; in DES_ede3_cfb64_encrypt()
45 v0 = ti[0]; in DES_ede3_cfb64_encrypt()
46 v1 = ti[1]; in DES_ede3_cfb64_encrypt()
49 l2c(v0, iv); in DES_ede3_cfb64_encrypt()
56 n = (n + 1) & 0x07; in DES_ede3_cfb64_encrypt()
61 c2l(iv, v0); in DES_ede3_cfb64_encrypt()
64 ti[0] = v0; in DES_ede3_cfb64_encrypt()
[all …]
H A Dcfb_enc.c27 * Until Aug 1 2003 this function did not correctly implement CFB-r, so it
34 register DES_LONG d0, d1, v0, v1; in DES_cfb_encrypt() local
55 c2l(iv, v0); in DES_cfb_encrypt()
60 ti[0] = v0; in DES_cfb_encrypt()
61 ti[1] = v1; in DES_cfb_encrypt()
66 d1 ^= ti[1]; in DES_cfb_encrypt()
74 v0 = v1; in DES_cfb_encrypt()
77 v0 = d0; in DES_cfb_encrypt()
82 l2c(v0, iv); in DES_cfb_encrypt()
87 sh[0] = v0, sh[1] = v1, sh[2] = d0, sh[3] = d1; in DES_cfb_encrypt()
[all …]
H A Dcfb64enc.c28 register DES_LONG v0, v1; in DES_cfb64_encrypt() local
38 c2l(iv, v0); in DES_cfb64_encrypt()
39 ti[0] = v0; in DES_cfb64_encrypt()
41 ti[1] = v1; in DES_cfb64_encrypt()
44 v0 = ti[0]; in DES_cfb64_encrypt()
45 l2c(v0, iv); in DES_cfb64_encrypt()
46 v0 = ti[1]; in DES_cfb64_encrypt()
47 l2c(v0, iv); in DES_cfb64_encrypt()
53 n = (n + 1) & 0x07; in DES_cfb64_encrypt()
58 c2l(iv, v0); in DES_cfb64_encrypt()
[all …]
H A Dofb_enc.c28 register DES_LONG d0, d1, vv0, vv1, v0, v1, n = (numbits + 7) / 8; in DES_ofb_encrypt() local
42 mask1 = (1L << (num - 32)) - 1; in DES_ofb_encrypt()
47 mask0 = (1L << num) - 1; in DES_ofb_encrypt()
52 c2l(iv, v0); in DES_ofb_encrypt()
54 ti[0] = v0; in DES_ofb_encrypt()
55 ti[1] = v1; in DES_ofb_encrypt()
57 ti[0] = v0; in DES_ofb_encrypt()
58 ti[1] = v1; in DES_ofb_encrypt()
61 vv1 = ti[1]; in DES_ofb_encrypt()
70 v0 = v1; in DES_ofb_encrypt()
[all …]
/freebsd/lib/libc/quad/
H A Dmuldi3.c14 * 1. Redistributions of source code must retain the above copyright
48 * v = 2^n v1 * v0
52 * uv = 2^2n u1 v1 + 2^n u1 v0 + 2^n v1 u0 + u0 v0
53 * = 2^2n u1 v1 + 2^n (u1 v0 + v1 u0) + u0 v0
56 * and add 2^n u0 v0 to the last term and subtract it from the middle.
60 * (2^n) (u1 v0 - u1 v1 + u0 v1 - u0 v0) +
61 * (2^n + 1) (u0 v0)
66 * (2^n) (u1 - u0) (v0 - v1) + [(u1-u0)... = mid]
67 * (2^n + 1) (u0 v0) [u0v0 = low]
69 * The terms (u1 v1), (u1 - u0) (v0 - v1), and (u0 v0) can all be done
[all …]
/freebsd/sys/libkern/arm/
H A Dmuldi3.c16 * 1. Redistributions of source code must retain the above copyright
50 * v = 2^n v1 * v0
54 * uv = 2^2n u1 v1 + 2^n u1 v0 + 2^n v1 u0 + u0 v0
55 * = 2^2n u1 v1 + 2^n (u1 v0 + v1 u0) + u0 v0
58 * and add 2^n u0 v0 to the last term and subtract it from the middle.
62 * (2^n) (u1 v0 - u1 v1 + u0 v1 - u0 v0) +
63 * (2^n + 1) (u0 v0)
68 * (2^n) (u1 - u0) (v0 - v1) + [(u1-u0)... = mid]
69 * (2^n + 1) (u0 v0) [u0v0 = low]
71 * The terms (u1 v1), (u1 - u0) (v0 - v1), and (u0 v0) can all be done
[all …]
/freebsd/contrib/libucl/klib/
H A Dkvec.h39 return 1;
55 #define kv_roundup32(x) (--(x), (x)|=(x)>>1, (x)|=(x)>>2, (x)|=(x)>>4, (x)|=(x)>>8, (x)|=(x)>>16, +…
77 size_t _ts = ((v).m > 1 ? (v).m * kv_grow_factor : 2); \
87 #define kv_copy_safe(type, v1, v0, el) do { \ argument
88 if ((v1).m < (v0).n) kv_resize_safe(type, v1, (v0).n, el); \
89 (v1).n = (v0).n; \
90 memcpy((v1).a, (v0).a, sizeof(type) * (v0).n); \
104 memmove((v).a + 1, (v).a, sizeof(type) * (v).n); \
109 #define kv_concat_safe(type, v1, v0, el) do { \ argument
110 if ((v1).m < (v0).n + (v1).n) \
[all …]
/freebsd/contrib/llvm-project/clang/lib/Headers/
H A Dvelintrin_approx.h12 static inline __vr _vel_approx_vfdivs_vvvl(__vr v0, __vr v1, int l) { in _vel_approx_vfdivs_vvvl() argument
19 v2 = _vel_vfmuls_vvvl(v0, v3, l); in _vel_approx_vfdivs_vvvl()
20 v4 = _vel_vfnmsbs_vvvvl(v0, v2, v1, l); in _vel_approx_vfdivs_vvvl()
22 v0 = _vel_vfnmsbs_vvvvl(v0, v2, v1, l); in _vel_approx_vfdivs_vvvl()
23 v0 = _vel_vfmads_vvvvl(v2, v3, v0, l); in _vel_approx_vfdivs_vvvl()
24 return v0; in _vel_approx_vfdivs_vvvl()
27 static inline __vr _vel_approx_pvfdiv_vvvl(__vr v0, __vr v1, int l) { in _vel_approx_pvfdiv_vvvl() argument
34 v2 = _vel_pvfmul_vvvl(v0, v3, l); in _vel_approx_pvfdiv_vvvl()
35 v4 = _vel_pvfnmsb_vvvvl(v0, v2, v1, l); in _vel_approx_pvfdiv_vvvl()
37 v0 = _vel_pvfnmsb_vvvvl(v0, v2, v1, l); in _vel_approx_pvfdiv_vvvl()
[all …]
/freebsd/sys/contrib/openzfs/module/icp/asm-aarch64/blake3/
H A Db3_aarch64_sse2.S81 eor v0.16b, v2.16b, v0.16b
109 mov v1.s[1], w5
114 and v0.8b, v1.8b, v0.8b
116 mov v3.d[1], v0.d[0]
119 uzp1 v1.4s, v0.4s, v6.4s
120 uzp2 v0.4s, v0
[all...]
H A Db3_aarch64_sse41.S27 * This is converted assembly: SSE4.1 -> ARMv8-A
81 eor v0.16b, v2.16b, v0.16b
102 .byte 1
116 .byte 1
142 mov v1.s[1], w5
147 and v0.8b, v1.8b, v0.8b
150 mov v1.d[1], v0
[all...]
/freebsd/contrib/llvm-project/llvm/lib/Target/RISCV/
H A DRISCVInstrInfoVVLPatterns.td24 def SDT_RISCVIntUnOp_VL : SDTypeProfile<1, 4, [SDTCisSameAs<0, 1>,
31 def SDT_RISCVIntBinOp_VL : SDTypeProfile<1, 5, [SDTCisSameAs<0, 1>,
40 def SDT_RISCVVNBinOp_RM_VL : SDTypeProfile<1, 6, [SDTCisVec<0>, SDTCisInt<0>,
42 SDTCisSameNumEltsAs<0, 1>,
43 SDTCisVec<1>,
44 SDTCisOpSmallerThanOp<2, 1>,
51 def SDT_RISCVFPUnOp_VL : SDTypeProfile<1, 3, [SDTCisSameAs<0, 1>,
56 def SDT_RISCVFPBinOp_VL : SDTypeProfile<1, 5, [SDTCisSameAs<0, 1>,
64 def SDT_RISCVCopySign_VL : SDTypeProfile<1, 5, [SDTCisSameAs<0, 1>,
73 SDTypeProfile<1, 3, [SDTCisVec<0>,
[all …]
/freebsd/contrib/bearssl/src/symcipher/
H A Dpoly1305_ctmulq.c129 r1 = r[1]; in poly1305_inner_big()
135 a1 = acc[1]; in poly1305_inner_big()
140 uint64_t v0, v1, v2; in poly1305_inner_big() local
143 v0 = br_dec64le(buf + 0); in poly1305_inner_big()
146 v1 = ((v0 >> 44) | (v1 << 20)) & MASK44; in poly1305_inner_big()
147 v0 &= MASK44; in poly1305_inner_big()
148 a0 += v0; in poly1305_inner_big()
150 a2 += v2 + ((uint64_t)1 << 40); in poly1305_inner_big()
158 v0 = br_dec64le(buf + 16); in poly1305_inner_big()
161 v1 = ((v0 >> 44) | (v1 << 20)) & MASK44; in poly1305_inner_big()
[all …]
/freebsd/crypto/openssl/crypto/bn/asm/
H A Dmips.pl99 ($zero,$at,$v0,$v1)=map("\$$_",(0..3));
143 move $v0,$zero
145 move $a0,$v0
161 $REG_S $t0,1*$SZREG($sp)
178 $ADDU $t1,$v0
179 sltu $v0,$t1,$v0 # All manuals say it "compares 32-bit
185 $ADDU $v0,$t0
189 $ADDU $v0,$at
193 $ADDU $t3,$v0
194 sltu $v0,$t3,$v0
[all …]
/freebsd/crypto/openssl/crypto/siphash/
H A Dsiphash.c36 (p)[1] = (uint8_t)((v) >> 8); \
45 (((uint64_t)((p)[0])) | ((uint64_t)((p)[1]) << 8) | \
52 v0 += v1; \
54 v1 ^= v0; \
55 v0 = ROTL(v0, 32); \
59 v0 += v3; \
61 v3 ^= v0; \
105 return 1; in SipHash_set_hash_size()
128 ctx->v0 = 0x736f6d6570736575ULL ^ k0; in SipHash_Init()
136 return 1; in SipHash_Init()
[all …]

12345678910>>...22