/freebsd/sys/crypto/openssl/aarch64/ |
H A D | aes-gcm-armv8_64.S | 60 …ld1 { v0.16b}, [x16] //special case vector load initial counter so we … 93 aese v0.16b, v18.16b 94 aesmc v0.16b, v0.16b //AES block 0 - round 0 105 aese v0.16b, v19.16b 106 aesmc v0.16b, v0.16b //AES block 0 - round 1 117 aese v0.16b, v20.16b 118 aesmc v0.16b, v0.16b //AES block 0 - round 2 135 aese v0.16b, v21.16b 136 aesmc v0.16b, v0.16b //AES block 0 - round 3 159 aese v0.16b, v22.16b [all …]
|
H A D | sha1-armv8.S | 52 add w24,w24,w27 // e+=rot(a,5) 66 add w23,w23,w27 // e+=rot(a,5) 77 add w22,w22,w27 // e+=rot(a,5) 91 add w21,w21,w27 // e+=rot(a,5) 102 add w20,w20,w27 // e+=rot(a,5) 116 add w24,w24,w27 // e+=rot(a,5) 127 add w23,w23,w27 // e+=rot(a,5) 141 add w22,w22,w27 // e+=rot(a,5) 152 add w21,w21,w27 // e+=rot(a,5) 166 add w20,w20,w27 // e+=rot(a,5) [all …]
|
H A D | aesv8-armx.S | 7 .align 5 15 .align 5 38 eor v0.16b,v0.16b,v0.16b 50 ext v5.16b,v0.16b,v3.16b,#12 52 aese v6.16b,v0.16b 56 ext v5.16b,v0.16b,v5.16b,#12 58 ext v5.16b,v0.16b,v5.16b,#12 68 ext v5.16b,v0.16b,v3.16b,#12 70 aese v6.16b,v0.16b 73 ext v5.16b,v0.16b,v5.16b,#12 [all …]
|
H A D | arm64cpuid.S | 7 .align 5 32 aese v0.16b, v0.16b 48 sha256su0 v0.4s, v0.4s 56 pmull v0.1q, v0.1d, v0.1d 64 .long 0xcec08000 // sha512su0 v0.2d,v0.2d 78 .align 5
|
H A D | sha256-armv8.S | 226 rev w8,w8 // 5 1077 ld1 {v0.4s,v1.4s},[x0] 1088 orr v18.16b,v0.16b,v0.16b // offload 1093 orr v2.16b,v0.16b,v0.16b 1094 .inst 0x5e104020 //sha256h v0.16b,v1.16b,v16.4s 1100 orr v2.16b,v0.16b,v0.16b 1101 .inst 0x5e114020 //sha256h v0.16b,v1.16b,v17.4s 1107 orr v2.16b,v0.16b,v0.16b 1108 .inst 0x5e104020 //sha256h v0.16b,v1.16b,v16.4s 1114 orr v2.16b,v0.16b,v0.16b [all …]
|
H A D | vpaes-armv8.S | 137 ushr v0.16b, v7.16b, #4 // vpsrlb $4, %xmm0, %xmm0 140 tbl v2.16b, {v21.16b}, v0.16b // vpshufb %xmm0, %xmm3, %xmm2 141 eor v0.16b, v1.16b, v16.16b // vpxor %xmm5, %xmm1, %xmm0 142 eor v0.16b, v0.16b, v2.16b // vpxor %xmm2, %xmm0, %xmm0 151 tbl v0.16b, {v24.16b}, v3.16b // vpshufb %xmm3, %xmm12, %xmm0 # 0 = sb1t 154 eor v0.16b, v0.16b, v4.16b // vpxor %xmm4, %xmm0, %xmm0 # 0 = A 157 tbl v3.16b, {v0.16b}, v1.16b // vpshufb %xmm1, %xmm0, %xmm3 # 0 = B 159 tbl v0.16b, {v0.16b}, v4.16b // vpshufb %xmm4, %xmm0, %xmm0 # 3 = D 162 eor v0.16b, v0.16b, v3.16b // vpxor %xmm3, %xmm0, %xmm0 # 3 = 2A+B+D 164 eor v0.16b, v0.16b, v4.16b // vpxor %xmm4, %xmm0, %xmm0 # 0 = 2A+3B+C+D [all …]
|
H A D | sha512-armv8.S | 116 eor x17,x20,x20,ror#5 141 eor x17,x27,x27,ror#5 165 eor x17,x26,x26,ror#5 190 eor x17,x25,x25,ror#5 214 eor x17,x24,x24,ror#5 224 rev x8,x8 // 5 239 eor x17,x23,x23,ror#5 263 eor x17,x22,x22,ror#5 288 eor x17,x21,x21,ror#5 312 eor x17,x20,x20,ror#5 [all …]
|
H A D | armv8-mont.S | 11 .align 5 228 .align 5 267 ld1 {v0.4s,v1.4s},[x1],#32 280 umlal v6.2d,v28.2s,v0.s[0] 281 umlal v7.2d,v28.2s,v0.s[1] 282 umlal v8.2d,v28.2s,v0.s[2] 285 umlal v9.2d,v28.2s,v0.s[3] 311 umlal v7.2d,v28.2s,v0.s[0] 313 umlal v8.2d,v28.2s,v0.s[1] 314 umlal v9.2d,v28.2s,v0.s[2] [all …]
|
/freebsd/crypto/openssl/crypto/bn/asm/ |
H A D | mips.pl | 23 # The module is designed to work with either of the "new" MIPS ABI(5), 25 # IRIX 5.x not only because it doesn't support new ABIs but also 26 # because 5.x kernels put R4x00 CPU into 32-bit mode and all those 99 ($zero,$at,$v0,$v1)=map("\$$_",(0..3)); 137 .align 5 143 move $v0,$zero 145 move $a0,$v0 148 .align 5 157 $REG_S $ra,5*$SZREG($sp) 178 $ADDU $t1,$v0 [all …]
|
/freebsd/contrib/bearssl/src/rsa/ |
H A D | rsa_i31_privexp.c | 65 uint32_t r, a, b, u0, v0, u1, v1, he, hr; in br_rsa_i31_compute_privexp() local 84 if (plen < 5 || plen > (BR_MAX_RSA_FACTOR / 8) in br_rsa_i31_compute_privexp() 95 if (qlen < 5 || qlen > (BR_MAX_RSA_FACTOR / 8) in br_rsa_i31_compute_privexp() 111 plen = (p[0] + 31) >> 5; in br_rsa_i31_compute_privexp() 114 qlen = (q[0] + 31) >> 5; in br_rsa_i31_compute_privexp() 128 len = (phi[0] + 31) >> 5; in br_rsa_i31_compute_privexp() 132 len = (phi[0] + 31) >> 5; in br_rsa_i31_compute_privexp() 161 * u0, u1, v0 and v1. Initial values are: in br_rsa_i31_compute_privexp() 162 * a = e u0 = 1 v0 = 0 in br_rsa_i31_compute_privexp() 165 * a = u0*e - v0*r in br_rsa_i31_compute_privexp() [all …]
|
/freebsd/crypto/openssl/crypto/aes/asm/ |
H A D | vpaes-ppc.pl | 161 .align 5 177 my ($inp,$out,$key) = map("r$_",(3..5)); 236 .align 5 246 vsrb v1, v0, v8 # vpsrlb \$4, %xmm0, %xmm0 247 vperm v0, $iptlo, $iptlo, v0 # vpshufb %xmm1, %xmm2, %xmm1 249 vxor v0, v0, v5 # vpxor %xmm5, %xmm1, %xmm0 250 vxor v0, v0, v1 # vpxor %xmm2, %xmm0, %xmm0 260 vperm v0, $sb1u, v7, v3 # vpshufb %xmm3, %xmm12, %xmm0 # 0 = sb1t 264 vxor v0, v0, v4 # vpxor %xmm4, %xmm0, %xmm0 # 0 = A 268 vperm v3, v0, v7, v1 # vpshufb %xmm1, %xmm0, %xmm3 # 0 = B [all …]
|
H A D | vpaes-armv8.pl | 200 ushr v0.16b, v7.16b, #4 // vpsrlb \$4, %xmm0, %xmm0 203 tbl v2.16b, {$ipthi}, v0.16b // vpshufb %xmm0, %xmm3, %xmm2 204 eor v0.16b, v1.16b, v16.16b // vpxor %xmm5, %xmm1, %xmm0 205 eor v0.16b, v0.16b, v2.16b // vpxor %xmm2, %xmm0, %xmm0 214 tbl v0.16b, {$sb1u}, v3.16b // vpshufb %xmm3, %xmm12, %xmm0 # 0 = sb1t 217 eor v0.16b, v0.16b, v4.16b // vpxor %xmm4, %xmm0, %xmm0 # 0 = A 220 tbl v3.16b, {v0.16b}, v1.16b // vpshufb %xmm1, %xmm0, %xmm3 # 0 = B 222 tbl v0.16b, {v0.16b}, v4.16b // vpshufb %xmm4, %xmm0, %xmm0 # 3 = D 225 eor v0.16b, v0.16b, v3.16b // vpxor %xmm3, %xmm0, %xmm0 # 3 = 2A+B+D 227 eor v0.16b, v0.16b, v4.16b // vpxor %xmm4, %xmm0, %xmm0 # 0 = 2A+3B+C+D [all …]
|
/freebsd/contrib/bearssl/src/symcipher/ |
H A D | poly1305_ctmulq.c | 64 * We want to reduce that value modulo p = 2^130-5, so W^3 = 20 mod p, 133 u2 = r[5]; in poly1305_inner_big() 140 uint64_t v0, v1, v2; in poly1305_inner_big() local 143 v0 = br_dec64le(buf + 0); in poly1305_inner_big() 146 v1 = ((v0 >> 44) | (v1 << 20)) & MASK44; in poly1305_inner_big() 147 v0 &= MASK44; in poly1305_inner_big() 148 a0 += v0; in poly1305_inner_big() 158 v0 = br_dec64le(buf + 16); in poly1305_inner_big() 161 v1 = ((v0 >> 44) | (v1 << 20)) & MASK44; in poly1305_inner_big() 162 v0 &= MASK44; in poly1305_inner_big() [all …]
|
/freebsd/contrib/file/tests/ |
H A D | Makefile.am | 147 zstd-v0.2-FF.result \ 148 zstd-v0.2-FF.testfile \ 149 zstd-v0.3-FF.result \ 150 zstd-v0.3-FF.testfile \ 151 zstd-v0.4-FF.result \ 152 zstd-v0.4-FF.testfile \ 153 zstd-v0.5-FF.result \ 154 zstd-v0.5-FF.testfile \ 155 zstd-v0.6-FF.result \ 156 zstd-v0.6-FF.testfile \ [all …]
|
/freebsd/contrib/llvm-project/llvm/lib/Target/RISCV/ |
H A D | RISCVInstrInfoVVLPatterns.td | 31 def SDT_RISCVIntBinOp_VL : SDTypeProfile<1, 5, [SDTCisSameAs<0, 1>, 37 SDTCisVT<5, XLenVT>]>; 48 SDTCisVT<5, XLenVT>, 56 def SDT_RISCVFPBinOp_VL : SDTypeProfile<1, 5, [SDTCisSameAs<0, 1>, 62 SDTCisVT<5, XLenVT>]>; 64 def SDT_RISCVCopySign_VL : SDTypeProfile<1, 5, [SDTCisSameAs<0, 1>, 70 SDTCisVT<5, XLenVT>]>; 180 def SDT_RISCVVecFMA_VL : SDTypeProfile<1, 5, [SDTCisSameAs<0, 1>, 186 SDTCisVT<5, XLenVT>]>; 192 def SDT_RISCVWVecFMA_VL : SDTypeProfile<1, 5, [SDTCisVec<0>, SDTCisFP<0>, [all …]
|
/freebsd/contrib/unbound/util/ |
H A D | siphash.c | 53 ((uint64_t)((p)[4]) << 32) | ((uint64_t)((p)[5]) << 40) | \ 58 v0 += v1; \ 60 v1 ^= v0; \ 61 v0 = ROTL(v0, 32); \ 65 v0 += v3; \ 67 v3 ^= v0; \ 77 printf("(%3d) v0 %08x %08x\n", (int)inlen, (uint32_t)(v0 >> 32), \ 78 (uint32_t)v0); \ 93 uint64_t v0 = 0x736f6d6570736575ULL; in siphash() local 111 v0 ^= k0; in siphash() [all …]
|
/freebsd/crypto/openssl/crypto/siphash/ |
H A D | siphash.c | 47 ((uint64_t)((p)[4]) << 32) | ((uint64_t)((p)[5]) << 40) | \ 52 v0 += v1; \ 54 v1 ^= v0; \ 55 v0 = ROTL(v0, 32); \ 59 v0 += v3; \ 61 v3 ^= v0; \ 128 ctx->v0 = 0x736f6d6570736575ULL ^ k0; in SipHash_Init() 145 uint64_t v0 = ctx->v0; in SipHash_Update() local 173 v0 ^= m; in SipHash_Update() 183 v0 ^= m; in SipHash_Update() [all …]
|
/freebsd/sys/contrib/openzfs/module/icp/asm-aarch64/sha2/ |
H A D | sha256-armv8.S | 28 .word 5 208 rev w8,w8 // 5 1034 ld1 {v0.4s,v1.4s},[x0] 1045 orr v18.16b,v0.16b,v0.16b // offload 1050 orr v2.16b,v0.16b,v0.16b 1051 .inst 0x5e104020 //sha256h v0.16b,v1.16b,v16.4s 1057 orr v2.16b,v0.16b,v0 [all...] |
H A D | sha512-armv8.S | 28 .word 5 124 eor x17,x20,x20,ror#5 149 eor x17,x27,x27,ror#5 173 eor x17,x26,x26,ror#5 198 eor x17,x25,x25,ror#5 222 eor x17,x24,x24,ror#5 232 rev x8,x8 // 5 247 eor x17,x23,x23,ror#5 271 eor x17,x22,x22,ror#5 296 eor x17,x21,x21,ror#5 [all...] |
/freebsd/sys/contrib/libsodium/src/libsodium/crypto_shorthash/siphash24/ref/ |
H A D | shorthash_siphashx24_ref.c | 9 uint64_t v0 = 0x736f6d6570736575ULL; in crypto_shorthash_siphashx24() local 24 v0 ^= k0; in crypto_shorthash_siphashx24() 30 v0 ^= m; in crypto_shorthash_siphashx24() 37 b |= ((uint64_t) in[5]) << 40; in crypto_shorthash_siphashx24() 39 case 5: in crypto_shorthash_siphashx24() 60 v0 ^= b; in crypto_shorthash_siphashx24() 66 b = v0 ^ v1 ^ v2 ^ v3; in crypto_shorthash_siphashx24() 73 b = v0 ^ v1 ^ v2 ^ v3; in crypto_shorthash_siphashx24()
|
H A D | shorthash_siphash24_ref.c | 10 uint64_t v0 = 0x736f6d6570736575ULL; in crypto_shorthash_siphash24() local 25 v0 ^= k0; in crypto_shorthash_siphash24() 31 v0 ^= m; in crypto_shorthash_siphash24() 38 b |= ((uint64_t) in[5]) << 40; in crypto_shorthash_siphash24() 40 case 5: in crypto_shorthash_siphash24() 61 v0 ^= b; in crypto_shorthash_siphash24() 67 b = v0 ^ v1 ^ v2 ^ v3; in crypto_shorthash_siphash24()
|
/freebsd/crypto/openssl/crypto/ |
H A D | arm64cpuid.pl | 30 .align 5 55 aese v0.16b, v0.16b 71 sha256su0 v0.4s, v0.4s 79 pmull v0.1q, v0.1d, v0.1d 87 .long 0xcec08000 // sha512su0 v0.2d,v0.2d 101 .align 5
|
/freebsd/sys/contrib/zstd/ |
H A D | CHANGELOG | 169 perf: Improved decompression speed: x64 : +10% (clang) / +5% (gcc); ARM : from +15% to +50%, depend… 222 bug: Fix Buffer Overflow in legacy v0.3 decompression by @felixhandte (#1722) 337 build: Read Legacy format is limited to v0.5+ by default. Can be changed at compile time with macro… 439 v1.2.0 (May 5, 2017) 467 speed : improved decompression speed in streaming mode for single shot scenarios (+5%) 471 API : changed : dropped support of legacy versions <= v0.3 (can be changed by modifying ZSTD_LEGACY… 514 Improved : better memory usage when using streaming compression API, thanks to @Rogier-5 report 540 CLI supports legacy formats v0.4+ 544 v0.8.1 (Aug 18, 2016) 553 v0.8.0 (Aug 2, 2016) [all …]
|
/freebsd/lib/libc/aarch64/string/ |
H A D | timingsafe_bcmp.S | 19 cmp x2, #4 // at least 5 bytes to process? 76 eor v0.16b, v0.16b, v1.16b 78 orr v0.16b, v0.16b, v2.16b 79 umaxv s0, v0.4s // get a nonzero word if any 80 mov w0, v0.s[0] 86 eor v0.16b, v0.16b, v2.16b 88 orr v4.16b, v0.16b, v1.16b 94 eor v0.16b, v0.16b, v2.16b 96 orr v0.16b, v0.16b, v1.16b 97 orr v4.16b, v4.16b, v0.16b [all …]
|
/freebsd/contrib/llvm-project/llvm/lib/Support/ |
H A D | SipHash.cpp | 36 v0 += v1; \ 38 v1 ^= v0; \ 39 v0 = ROTL(v0, 32); \ 43 v0 += v3; \ 45 v3 ^= v0; \ 70 uint64_t v0 = UINT64_C(0x736f6d6570736575); in siphash() local 84 v0 ^= k0; in siphash() 96 v0 ^= m; in siphash() 104 b |= ((uint64_t)ni[5]) << 40; in siphash() 106 case 5: in siphash() [all …]
|