/freebsd/contrib/llvm-project/clang/lib/Headers/ |
H A D | velintrin_approx.h | 12 static inline __vr _vel_approx_vfdivs_vvvl(__vr v0, __vr v1, int l) { in _vel_approx_vfdivs_vvvl() argument 19 v2 = _vel_vfmuls_vvvl(v0, v3, l); in _vel_approx_vfdivs_vvvl() 20 v4 = _vel_vfnmsbs_vvvvl(v0, v2, v1, l); in _vel_approx_vfdivs_vvvl() 22 v0 = _vel_vfnmsbs_vvvvl(v0, v2, v1, l); in _vel_approx_vfdivs_vvvl() 23 v0 = _vel_vfmads_vvvvl(v2, v3, v0, l); in _vel_approx_vfdivs_vvvl() 24 return v0; in _vel_approx_vfdivs_vvvl() 27 static inline __vr _vel_approx_pvfdiv_vvvl(__vr v0, __vr v1, int l) { in _vel_approx_pvfdiv_vvvl() argument 34 v2 = _vel_pvfmul_vvvl(v0, v3, l); in _vel_approx_pvfdiv_vvvl() 35 v4 = _vel_pvfnmsb_vvvvl(v0, v2, v1, l); in _vel_approx_pvfdiv_vvvl() 37 v0 = _vel_pvfnmsb_vvvvl(v0, v2, v1, l); in _vel_approx_pvfdiv_vvvl() [all …]
|
/freebsd/sys/crypto/openssl/aarch64/ |
H A D | ghashv8-armx.S | 31 pmull v0.1q,v20.1d,v20.1d 36 ext v17.16b,v0.16b,v2.16b,#8 //Karatsuba post-processing 37 eor v18.16b,v0.16b,v2.16b 40 pmull v18.1q,v0.1d,v19.1d //1st phase 43 ins v1.d[1],v0.d[0] 44 eor v0.16b,v1.16b,v18.16b 46 ext v18.16b,v0.16b,v0.16b,#8 //2nd phase 47 pmull v0.1q,v0.1d,v19.1d 49 eor v22.16b,v0.16b,v18.16b 56 pmull v0.1q,v20.1d, v22.1d [all …]
|
H A D | aesv8-armx.S | 38 eor v0.16b,v0.16b,v0.16b 50 ext v5.16b,v0.16b,v3.16b,#12 52 aese v6.16b,v0.16b 56 ext v5.16b,v0.16b,v5.16b,#12 58 ext v5.16b,v0.16b,v5.16b,#12 68 ext v5.16b,v0.16b,v3.16b,#12 70 aese v6.16b,v0.16b 73 ext v5.16b,v0.16b,v5.16b,#12 75 ext v5.16b,v0.16b,v5.16b,#12 82 ext v5.16b,v0.16b,v3.16b,#12 [all …]
|
H A D | vpaes-armv8.S | 137 ushr v0.16b, v7.16b, #4 // vpsrlb $4, %xmm0, %xmm0 140 tbl v2.16b, {v21.16b}, v0.16b // vpshufb %xmm0, %xmm3, %xmm2 141 eor v0.16b, v1.16b, v16.16b // vpxor %xmm5, %xmm1, %xmm0 142 eor v0.16b, v0.16b, v2.16b // vpxor %xmm2, %xmm0, %xmm0 151 tbl v0.16b, {v24.16b}, v3.16b // vpshufb %xmm3, %xmm12, %xmm0 # 0 = sb1t 154 eor v0.16b, v0.16b, v4.16b // vpxor %xmm4, %xmm0, %xmm0 # 0 = A 157 tbl v3.16b, {v0.16b}, v1.16b // vpshufb %xmm1, %xmm0, %xmm3 # 0 = B 159 tbl v0.16b, {v0.16b}, v4.16b // vpshufb %xmm4, %xmm0, %xmm0 # 3 = D 162 eor v0.16b, v0.16b, v3.16b // vpxor %xmm3, %xmm0, %xmm0 # 3 = 2A+B+D 164 eor v0.16b, v0.16b, v4.16b // vpxor %xmm4, %xmm0, %xmm0 # 0 = 2A+3B+C+D [all …]
|
H A D | aes-gcm-armv8_64.S | 60 …ld1 { v0.16b}, [x16] //special case vector load initial counter so we … 93 aese v0.16b, v18.16b 94 aesmc v0.16b, v0.16b //AES block 0 - round 0 105 aese v0.16b, v19.16b 106 aesmc v0.16b, v0.16b //AES block 0 - round 1 117 aese v0.16b, v20.16b 118 aesmc v0.16b, v0.16b //AES block 0 - round 2 135 aese v0.16b, v21.16b 136 aesmc v0.16b, v0.16b //AES block 0 - round 3 159 aese v0.16b, v22.16b [all …]
|
/freebsd/sys/sys/ |
H A D | ktr.h | 143 #define KTR_EVENT1(m, egroup, ident, etype, edat, a0, v0) \ argument 144 CTR3(m, KTR_EFMT(egroup, ident, etype) a0, ident, edat, (v0)) 145 #define KTR_EVENT2(m, egroup, ident, etype, edat, a0, v0, a1, v1) \ argument 147 ident, edat, (v0), (v1)) 148 #define KTR_EVENT3(m, egroup, ident, etype, edat, a0, v0, a1, v1, a2, v2)\ argument 150 ident, edat, (v0), (v1), (v2)) 152 a0, v0, a1, v1, a2, v2, a3, v3) \ argument 154 ident, edat, (v0), (v1), (v2), (v3)) 161 #define KTR_STATE1(m, egroup, ident, state, a0, v0) \ argument 162 KTR_EVENT1(m, egroup, ident, "state:\"%s\"", state, a0, (v0)) [all …]
|
/freebsd/crypto/openssl/crypto/des/ |
H A D | cfb64ede.c | 29 register DES_LONG v0, v1; in DES_ede3_cfb64_encrypt() local 39 c2l(iv, v0); in DES_ede3_cfb64_encrypt() 42 ti[0] = v0; in DES_ede3_cfb64_encrypt() 45 v0 = ti[0]; in DES_ede3_cfb64_encrypt() 49 l2c(v0, iv); in DES_ede3_cfb64_encrypt() 61 c2l(iv, v0); in DES_ede3_cfb64_encrypt() 64 ti[0] = v0; in DES_ede3_cfb64_encrypt() 67 v0 = ti[0]; in DES_ede3_cfb64_encrypt() 71 l2c(v0, iv); in DES_ede3_cfb64_encrypt() 82 v0 = v1 = ti[0] = ti[1] = c = cc = 0; in DES_ede3_cfb64_encrypt() [all …]
|
H A D | cfb64enc.c | 28 register DES_LONG v0, v1; in DES_cfb64_encrypt() local 38 c2l(iv, v0); in DES_cfb64_encrypt() 39 ti[0] = v0; in DES_cfb64_encrypt() 44 v0 = ti[0]; in DES_cfb64_encrypt() 45 l2c(v0, iv); in DES_cfb64_encrypt() 46 v0 = ti[1]; in DES_cfb64_encrypt() 47 l2c(v0, iv); in DES_cfb64_encrypt() 58 c2l(iv, v0); in DES_cfb64_encrypt() 59 ti[0] = v0; in DES_cfb64_encrypt() 64 v0 = ti[0]; in DES_cfb64_encrypt() [all …]
|
H A D | cfb_enc.c | 34 register DES_LONG d0, d1, v0, v1; in DES_cfb_encrypt() local 55 c2l(iv, v0); in DES_cfb_encrypt() 60 ti[0] = v0; in DES_cfb_encrypt() 74 v0 = v1; in DES_cfb_encrypt() 77 v0 = d0; in DES_cfb_encrypt() 82 l2c(v0, iv); in DES_cfb_encrypt() 87 sh[0] = v0, sh[1] = v1, sh[2] = d0, sh[3] = d1; in DES_cfb_encrypt() 96 v0 = sh[0], v1 = sh[1]; in DES_cfb_encrypt() 99 c2l(iv, v0); in DES_cfb_encrypt() 107 ti[0] = v0; in DES_cfb_encrypt() [all …]
|
H A D | ofb_enc.c | 28 register DES_LONG d0, d1, vv0, vv1, v0, v1, n = (numbits + 7) / 8; in DES_ofb_encrypt() local 52 c2l(iv, v0); in DES_ofb_encrypt() 54 ti[0] = v0; in DES_ofb_encrypt() 57 ti[0] = v0; in DES_ofb_encrypt() 70 v0 = v1; in DES_ofb_encrypt() 73 v0 = vv0; in DES_ofb_encrypt() 76 v0 = ((v1 >> (num - 32)) | (vv0 << (64 - num))) & 0xffffffffL; in DES_ofb_encrypt() 80 v0 = ((v0 >> num) | (v1 << (32 - num))) & 0xffffffffL; in DES_ofb_encrypt() 85 l2c(v0, iv); in DES_ofb_encrypt() 87 v0 = v1 = d0 = d1 = ti[0] = ti[1] = vv0 = vv1 = 0; in DES_ofb_encrypt()
|
H A D | ofb64ede.c | 28 register DES_LONG v0, v1; in DES_ede3_ofb64_encrypt() local 38 c2l(iv, v0); in DES_ede3_ofb64_encrypt() 40 ti[0] = v0; in DES_ede3_ofb64_encrypt() 43 l2c(v0, dp); in DES_ede3_ofb64_encrypt() 50 v0 = ti[0]; in DES_ede3_ofb64_encrypt() 54 l2c(v0, dp); in DES_ede3_ofb64_encrypt() 63 l2c(v0, iv); in DES_ede3_ofb64_encrypt() 66 v0 = v1 = ti[0] = ti[1] = 0; in DES_ede3_ofb64_encrypt()
|
/freebsd/contrib/file/tests/ |
H A D | Makefile.am | 147 zstd-v0.2-FF.result \ 148 zstd-v0.2-FF.testfile \ 149 zstd-v0.3-FF.result \ 150 zstd-v0.3-FF.testfile \ 151 zstd-v0.4-FF.result \ 152 zstd-v0.4-FF.testfile \ 153 zstd-v0.5-FF.result \ 154 zstd-v0.5-FF.testfile \ 155 zstd-v0.6-FF.result \ 156 zstd-v0.6-FF.testfile \ [all …]
|
/freebsd/contrib/libucl/klib/ |
H A D | kvec.h | 87 #define kv_copy_safe(type, v1, v0, el) do { \ argument 88 if ((v1).m < (v0).n) kv_resize_safe(type, v1, (v0).n, el); \ 89 (v1).n = (v0).n; \ 90 memcpy((v1).a, (v0).a, sizeof(type) * (v0).n); \ 109 #define kv_concat_safe(type, v1, v0, el) do { \ argument 110 if ((v1).m < (v0).n + (v1).n) \ 111 kv_resize_safe(type, v1, (v0).n + (v1).n, el); \ 112 memcpy((v1).a + (v1).n, (v0).a, sizeof(type) * (v0).n); \ 113 (v1).n = (v0).n + (v1).n; \ 133 #define kv_copy(type, v1, v0) do { \ argument [all …]
|
/freebsd/lib/libc/aarch64/string/ |
H A D | strncmp.S | 50 cmeq v5.16b, v0.16b, #0 66 tbl v0.16b, {v0.16b}, v4.16b 87 cmeq v2.16b, v0.16b, #0 // NUL byte present? 88 cmeq v4.16b, v0.16b, v4.16b // which bytes match? 132 cmeq v5.16b, v0.16b, #0 151 tbl v0.16b, {v0.16b}, v4.16b 173 cmeq v2.16b, v0.16b, #0 // NUL byte present? 174 cmeq v4.16b, v0.16b, v4.16b // which bytes match? 201 cmeq v0.16b, v0.16b, v2.16b // Mismatch between chunks? 203 shrn v0.8b, v0.8h, #4 [all …]
|
H A D | timingsafe_bcmp.S | 76 eor v0.16b, v0.16b, v1.16b 78 orr v0.16b, v0.16b, v2.16b 79 umaxv s0, v0.4s // get a nonzero word if any 80 mov w0, v0.s[0] 86 eor v0.16b, v0.16b, v2.16b 88 orr v4.16b, v0.16b, v1.16b 94 eor v0.16b, v0.16b, v2.16b 96 orr v0.16b, v0.16b, v1.16b 97 orr v4.16b, v4.16b, v0.16b 106 eor v0.16b, v0.16b, v2.16b [all …]
|
H A D | strcmp.S | 44 cmeq v5.16b, v0.16b, #0 60 tbl v0.16b, {v0.16b}, v4.16b 82 cmeq v2.16b, v0.16b, #0 // NUL byte present? 83 cmeq v4.16b, v0.16b, v4.16b // which bytes match? 102 cmeq v0.16b, v0.16b, v2.16b 106 shrn v0.8b, v0.8h, #4 134 cmeq v0.16b, v0.16b, v2.16b // do the chunks match? 138 shrn v0.8b, v0.8h, #4 152 cmeq v0.16b, v0.16b, v2.16b 156 shrn v0.8b, v0.8h, #4 [all …]
|
/freebsd/contrib/bearssl/src/symcipher/ |
H A D | poly1305_ctmulq.c | 140 uint64_t v0, v1, v2; in poly1305_inner_big() local 143 v0 = br_dec64le(buf + 0); in poly1305_inner_big() 146 v1 = ((v0 >> 44) | (v1 << 20)) & MASK44; in poly1305_inner_big() 147 v0 &= MASK44; in poly1305_inner_big() 148 a0 += v0; in poly1305_inner_big() 158 v0 = br_dec64le(buf + 16); in poly1305_inner_big() 161 v1 = ((v0 >> 44) | (v1 << 20)) & MASK44; in poly1305_inner_big() 162 v0 &= MASK44; in poly1305_inner_big() 163 a0 += v0; in poly1305_inner_big() 173 v0 = br_dec64le(buf + 32); in poly1305_inner_big() [all …]
|
/freebsd/sys/contrib/openzfs/module/icp/asm-aarch64/blake3/ |
H A D | b3_aarch64_sse2.S | 81 eor v0.16b, v2.16b, v0.16b 114 and v0.8b, v1.8b, v0.8b 116 mov v3.d[1], v0.d[0] 119 uzp1 v1.4s, v0.4s, v6.4s 120 uzp2 v0.4s, v0.4s, v6.4s 125 add v2.4s, v2.4s, v0.4s 167 uzp2 v17.4s, v17.4s, v0 [all...] |
/freebsd/contrib/unbound/util/ |
H A D | siphash.c | 58 v0 += v1; \ 60 v1 ^= v0; \ 61 v0 = ROTL(v0, 32); \ 65 v0 += v3; \ 67 v3 ^= v0; \ 77 printf("(%3d) v0 %08x %08x\n", (int)inlen, (uint32_t)(v0 >> 32), \ 78 (uint32_t)v0); \ 93 uint64_t v0 = 0x736f6d6570736575ULL; in siphash() local 111 v0 ^= k0; in siphash() 124 v0 ^= m; in siphash() [all …]
|
/freebsd/crypto/openssl/crypto/siphash/ |
H A D | siphash.c | 52 v0 += v1; \ 54 v1 ^= v0; \ 55 v0 = ROTL(v0, 32); \ 59 v0 += v3; \ 61 v3 ^= v0; \ 128 ctx->v0 = 0x736f6d6570736575ULL ^ k0; in SipHash_Init() 145 uint64_t v0 = ctx->v0; in SipHash_Update() local 173 v0 ^= m; in SipHash_Update() 183 v0 ^= m; in SipHash_Update() 191 ctx->v0 = v0; in SipHash_Update() [all …]
|
/freebsd/lib/libc/quad/ |
H A D | muldi3.c | 106 #define v0 v.ul[L] in __muldi3() macro 128 prod.q = __lmulq(u0, v0); in __muldi3() 136 low.q = __lmulq(u0, v0); in __muldi3() 142 if (v0 >= v1) in __muldi3() 143 vdiff = v0 - v1; in __muldi3() 145 vdiff = v1 - v0, negmid ^= 1; in __muldi3() 161 #undef v0 in __muldi3() 184 u_long u1, u0, v1, v0, udiff, vdiff, high, mid, low; in __lmulq() local 192 v0 = LHALF(v); in __lmulq() 194 low = u0 * v0; in __lmulq() [all …]
|
/freebsd/sys/libkern/arm/ |
H A D | muldi3.c | 109 #define v0 v.ul[L] in __muldi3() macro 131 prod.q = __lmulq(u0, v0); in __muldi3() 139 low.q = __lmulq(u0, v0); in __muldi3() 145 if (v0 >= v1) in __muldi3() 146 vdiff = v0 - v1; in __muldi3() 148 vdiff = v1 - v0, negmid ^= 1; in __muldi3() 164 #undef v0 in __muldi3() 187 u_int u1, u0, v1, v0, udiff, vdiff, high, mid, low; in __lmulq() local 195 v0 = LHALF(v); in __lmulq() 197 low = u0 * v0; in __lmulq() [all …]
|
/freebsd/crypto/openssl/crypto/bn/asm/ |
H A D | mips.pl | 99 ($zero,$at,$v0,$v1)=map("\$$_",(0..3)); 143 move $v0,$zero 145 move $a0,$v0 178 $ADDU $t1,$v0 179 sltu $v0,$t1,$v0 # All manuals say it "compares 32-bit 185 $ADDU $v0,$t0 189 $ADDU $v0,$at 193 $ADDU $t3,$v0 194 sltu $v0,$t3,$v0 198 $ADDU $v0,$t2 [all …]
|
/freebsd/contrib/bearssl/src/rsa/ |
H A D | rsa_i15_privexp.c | 65 uint32_t r, a, b, u0, v0, u1, v1, he, hr; in br_rsa_i15_compute_privexp() local 243 v0 = 0; in br_rsa_i15_compute_privexp() 262 ctl = GT(v1, v0); in br_rsa_i15_compute_privexp() 265 v0 -= (v1 - (e & -ctl)) & -sab; in br_rsa_i15_compute_privexp() 268 ctl = GT(v0, v1); in br_rsa_i15_compute_privexp() 271 v1 -= (v0 - (e & -ctl)) & -sba; in br_rsa_i15_compute_privexp() 277 ctl = v0 & 1; in br_rsa_i15_compute_privexp() 280 v0 ^= (v0 ^ ((v0 >> 1) + (he & -ctl))) & -da; in br_rsa_i15_compute_privexp() 305 m[1] = v0 & 0x7FFF; in br_rsa_i15_compute_privexp() 306 m[2] = (v0 >> 15) & 0x7FFF; in br_rsa_i15_compute_privexp() [all …]
|
/freebsd/contrib/llvm-project/llvm/lib/Support/ |
H A D | SipHash.cpp | 36 v0 += v1; \ 38 v1 ^= v0; \ 39 v0 = ROTL(v0, 32); \ 43 v0 += v3; \ 45 v3 ^= v0; \ 70 uint64_t v0 = UINT64_C(0x736f6d6570736575); in siphash() local 84 v0 ^= k0; in siphash() 96 v0 ^= m; in siphash() 130 v0 ^= b; in siphash() 140 b = v0 ^ v1 ^ v2 ^ v3; in siphash() [all …]
|