/freebsd/sys/crypto/openssl/aarch64/ |
H A D | aesv8-armx.S | 20 stp x29,x30,[sp,#-16]! 38 eor v0.16b,v0.16b,v0.16b 39 ld1 {v3.16b},[x0],#16 49 tbl v6.16b,{v3.16b},v2.16b 50 ext v5.16b,v0.16b,v3.16b,#12 51 st1 {v3.4s},[x2],#16 52 aese v6.16b,v0.16b 55 eor v3.16b,v3.16b,v5.16b 56 ext v5.16b,v0.16b,v5.16b,#12 57 eor v3.16b,v3.16b,v5.16b [all …]
|
H A D | ghashv8-armx.S | 13 movi v19.16b,#0xe1 15 ext v3.16b,v17.16b,v17.16b,#8 18 ext v16.16b,v18.16b,v19.16b,#8 //t0=0xc2....01 21 and v18.16b,v18.16b,v16.16b 23 ext v18.16b,v18.16b,v18.16b,#8 24 and v16.16b,v16.16b,v17.16b 25 orr v3.16b,v3.16b,v18.16b //H<<<=1 26 eor v20.16b,v3.16b,v16.16b //twisted H 27 st1 {v20.2d},[x0],#16 //store Htable[0] 30 ext v16.16b,v20.16b,v20.16b,#8 //Karatsuba pre-processing [all …]
|
H A D | aes-gcm-armv8_64.S | 16 stp x21, x22, [sp, #16] 33 ld1 {v11.16b}, [x3] 34 ext v11.16b, v11.16b, v11.16b, #8 35 rev64 v11.16b, v11.16b 39 ld1 {v18.4s}, [x8], #16 //load rk0 46 ext v15.16b, v15.16b, v15.16b, #8 53 ld1 {v19.4s}, [x8], #16 //load rk1 60 …ld1 { v0.16b}, [x16] //special case vector load initial counter so we … 73 ld1 {v20.4s}, [x8], #16 //load rk2 80 ext v14.16b, v14.16b, v14.16b, #8 [all …]
|
H A D | vpaes-armv8.S | 106 movi v17.16b, #0x0f 133 adr x11, .Lk_mc_forward+16 135 ld1 {v16.2d}, [x9], #16 // vmovdqu (%r9), %xmm5 # round0 key 136 and v1.16b, v7.16b, v17.16b // vpand %xmm9, %xmm0, %xmm1 137 ushr v0.16b, v7.16b, #4 // vpsrlb $4, %xmm0, %xmm0 138 tbl v1.16b, {v20.16b}, v1.16b // vpshufb %xmm1, %xmm2, %xmm1 139 // vmovdqa .Lk_ipt+16(%rip), %xmm3 # ipthi 140 tbl v2.16b, {v21.16b}, v0.16b // vpshufb %xmm0, %xmm3, %xmm2 141 eor v0.16b, v1.16b, v16.16b // vpxor %xmm5, %xmm1, %xmm0 142 eor v0.16b, v0.16b, v2.16b // vpxor %xmm2, %xmm0, %xmm0 [all …]
|
H A D | chacha-armv8.S | 41 stp x19,x20,[sp,#16] 50 ldp x26,x27,[x3,#16] 91 ror w17,w17,#16 92 ror w19,w19,#16 93 ror w20,w20,#16 94 ror w21,w21,#16 139 ror w21,w21,#16 140 ror w17,w17,#16 141 ror w19,w19,#16 142 ror w20,w20,#16 [all …]
|
H A D | keccak1600-armv8.S | 41 stp x28,x30,[sp,#16] // 32 bytes on top are mine 148 ldr x27,[sp,#16] 156 str x27,[sp,#16] 214 stp x19,x20,[sp,#16] 223 ldp x0,x1,[x0,#16*0] 224 ldp x2,x3,[x26,#16*1] 225 ldp x4,x5,[x26,#16*2] 226 ldp x6,x7,[x26,#16*3] 227 ldp x8,x9,[x26,#16*4] 228 ldp x10,x11,[x26,#16*5] [all …]
|
H A D | sha512-armv8.S | 82 stp x19,x20,[sp,#16] 452 str x9,[sp,#16] 540 ldr x9,[sp,#16] 572 str x13,[sp,#16] 664 ldr x13,[sp,#16] 696 str x1,[sp,#16] 788 ldr x1,[sp,#16] 820 str x5,[sp,#16] 912 ldr x5,[sp,#16] 944 str x9,[sp,#16] [all …]
|
/freebsd/sys/contrib/openzfs/module/icp/asm-aarch64/blake3/ |
H A D | b3_aarch64_sse41.S | 47 .word 16 68 .cfi_offset w19, -16 81 eor v0.16b, v2.16b, v0.16b 82 eor v1.16b, v3.16b, v1.16b 93 .section .rodata.cst16,"aM",@progbits,16 145 ldr q5, [x1, #16] [all...] |
H A D | b3_aarch64_sse2.S | 47 .word 16 68 .cfi_offset w19, -16 81 eor v0.16b, v2.16b, v0.16b 82 eor v1.16b, v3.16b, v1.16b 93 .section .rodata.cst16,"aM",@progbits,16 113 ldr q5, [x1, #16] [all...] |
/freebsd/tools/regression/ccd/layout/ |
H A D | ref.256k_128k_384k_128k_16_0 | 1 ccd3: 4 components (md90, md91, md92, md93), 1728 blocks interleaved at 16 blocks 6 0 16 16 7 16 16 16 8 32 16 16 9 48 16 16 10 64 32 16 11 80 32 16 12 96 32 16 13 112 32 16 14 128 48 16 [all …]
|
H A D | ref.256k_128k_384k_128k_16_2 | 1 ccd3: 4 components (md90, md91, md92, md93), 960 blocks interleaved at 16 blocks 6 0 16 257 9 16 16 16 10 32 16 16 11 48 16 16 12 64 32 16 13 80 32 16 14 96 32 16 15 112 32 16 16 128 48 16 [all …]
|
/freebsd/crypto/openssl/crypto/aes/asm/ |
H A D | vpaes-armv8.pl | 154 my ($invlo,$invhi,$iptlo,$ipthi,$sbou,$sbot) = map("v$_.16b",(18..23)); 155 my ($sb1u,$sb1t,$sb2u,$sb2t) = map("v$_.16b",(24..27)); 156 my ($sb9u,$sb9t,$sbdu,$sbdt,$sbbu,$sbbt,$sbeu,$sbet)=map("v$_.16b",(24..31)); 169 movi v17.16b, #0x0f 196 adr x11, .Lk_mc_forward+16 198 ld1 {v16.2d}, [x9], #16 // vmovdqu (%r9), %xmm5 # round0 key 199 and v1.16b, v7.16b, v17.16b // vpand %xmm9, %xmm0, %xmm1 200 ushr v0.16b, v7.16b, #4 // vpsrlb \$4, %xmm0, %xmm0 201 tbl v1.16b, {$iptlo}, v1.16b // vpshufb %xmm1, %xmm2, %xmm1 202 // vmovdqa .Lk_ipt+16(%rip), %xmm3 # ipthi [all …]
|
/freebsd/sys/contrib/openzfs/module/icp/asm-aarch64/sha2/ |
H A D | sha512-armv8.S | 27 .word 16 90 stp x19,x20,[sp,#16] 460 str x9,[sp,#16] 548 ldr x9,[sp,#16] 580 str x13,[sp,#16] 672 ldr x13,[sp,#16] 704 str x1,[sp,#16] 796 ldr x1,[sp,#16] 828 str x5,[sp,#16] 920 ldr x5,[sp,#16] [all...] |
/freebsd/sys/contrib/openzfs/module/zfs/ |
H A D | vdev_raidz_math_aarch64_neon_common.h | 132 extern const uint8_t gf_clmul_mod_lt[4*256][16]; 134 #define ELEM_SIZE 16 149 "eor " VR0(r) ".16b," VR0(r) ".16b,v21.16b\n" \ 150 "eor " VR1(r) ".16b," VR1(r) ".16b,v20.16b\n" \ 151 "eor " VR2(r) ".16b," VR2(r) ".16b,v19.16b\n" \ 152 "eor " VR3(r) ".16b," VR3(r) ".16b,v18.16b\n" \ 157 "eor " VR4(r) ".16b," VR4(r) ".16b,v21.16b\n" \ 158 "eor " VR5(r) ".16b," VR5(r) ".16b,v20.16b\n" \ 159 "eor " VR6(r) ".16b," VR6(r) ".16b,v19.16b\n" \ 160 "eor " VR7(r) ".16b," VR7(r) ".16b,v18.16b\n" \ [all …]
|
/freebsd/contrib/bearssl/src/symcipher/ |
H A D | aes_pwr8_cbcenc.c | 58 addi(%[cc], %[cc], 16) in cbcenc_128() 60 addi(%[cc], %[cc], 16) in cbcenc_128() 62 addi(%[cc], %[cc], 16) in cbcenc_128() 64 addi(%[cc], %[cc], 16) in cbcenc_128() 66 addi(%[cc], %[cc], 16) in cbcenc_128() 68 addi(%[cc], %[cc], 16) in cbcenc_128() 70 addi(%[cc], %[cc], 16) in cbcenc_128() 72 addi(%[cc], %[cc], 16) in cbcenc_128() 74 addi(%[cc], %[cc], 16) in cbcenc_128() 76 addi(%[cc], %[cc], 16) in cbcenc_128() [all …]
|
H A D | aes_pwr8_cbcdec.c | 52 cc1 = 16; in cbcdec_128() 61 addi(%[cc0], %[cc0], 16) in cbcdec_128() 63 addi(%[cc0], %[cc0], 16) in cbcdec_128() 65 addi(%[cc0], %[cc0], 16) in cbcdec_128() 67 addi(%[cc0], %[cc0], 16) in cbcdec_128() 69 addi(%[cc0], %[cc0], 16) in cbcdec_128() 71 addi(%[cc0], %[cc0], 16) in cbcdec_128() 73 addi(%[cc0], %[cc0], 16) in cbcdec_128() 75 addi(%[cc0], %[cc0], 16) in cbcdec_128() 77 addi(%[cc0], %[cc0], 16) in cbcdec_128() [all …]
|
/freebsd/contrib/cortex-strings/src/aarch64/ |
H A D | strchr.S | 92 movk wtmp2, #0x4010, lsl #16 93 dup vrepchr.16b, chrin 104 ld1 {vdata1.16b, vdata2.16b}, [src], #32 106 cmeq vhas_nul1.16b, vdata1.16b, #0 107 cmeq vhas_chr1.16b, vdata1.16b, vrepchr.16b 108 cmeq vhas_nul2.16b, vdata2.16b, #0 109 cmeq vhas_chr2.16b, vdata2.16b, vrepchr.16b 110 and vhas_nul1.16b, vhas_nul1.16b, vrepmask_0.16b 111 and vhas_nul2.16b, vhas_nul2.16b, vrepmask_0.16b 112 and vhas_chr1.16b, vhas_chr1.16b, vrepmask_c.16b [all …]
|
H A D | strchrnul.S | 81 movk wtmp2, #0x4010, lsl #16 82 dup vrepchr.16b, chrin 92 ld1 {vdata1.16b, vdata2.16b}, [src], #32 94 cmeq vhas_nul1.16b, vdata1.16b, #0 95 cmeq vhas_chr1.16b, vdata1.16b, vrepchr.16b 96 cmeq vhas_nul2.16b, vdata2.16b, #0 97 cmeq vhas_chr2.16b, vdata2.16b, vrepchr.16b 98 orr vhas_chr1.16b, vhas_chr1.16b, vhas_nul1.16b 99 orr vhas_chr2.16b, vhas_chr2.16b, vhas_nul2.16b 100 and vhas_chr1.16b, vhas_chr1.16b, vrepmask.16b [all …]
|
/freebsd/contrib/arm-optimized-routines/string/aarch64/ |
H A D | strrchr.S | 62 movk wtmp2, #0x4010, lsl #16 63 dup vrepchr.16b, chrin 75 ld1 {vdata1.16b, vdata2.16b}, [src], #32 77 cmeq vhas_nul1.16b, vdata1.16b, #0 78 cmeq vhas_chr1.16b, vdata1.16b, vrepchr.16b 79 cmeq vhas_nul2.16b, vdata2.16b, #0 80 cmeq vhas_chr2.16b, vdata2.16b, vrepchr.16b 81 and vhas_nul1.16b, vhas_nul1.16b, vrepmask_0.16b 82 and vhas_chr1.16b, vhas_chr1.16b, vrepmask_c.16b 83 and vhas_nul2.16b, vhas_nul2.16b, vrepmask_0.16b [all …]
|
H A D | strchr.S | 58 movk wtmp2, 0xc030, lsl 16 59 dup vrepchr.16b, chrin 70 ld1 {vdata1.16b, vdata2.16b}, [src], #32 72 cmeq vhas_nul1.16b, vdata1.16b, #0 73 cmeq vhas_chr1.16b, vdata1.16b, vrepchr.16b 74 cmeq vhas_nul2.16b, vdata2.16b, #0 75 cmeq vhas_chr2.16b, vdata2.16b, vrepchr.16b 76 bif vhas_nul1.16b, vhas_chr1.16b, vrepmask_0.16b 77 bif vhas_nul2.16b, vhas_chr2.16b, vrepmask_0.16b 78 and vend1.16b, vhas_nul1.16b, vrepmask_c.16b [all …]
|
H A D | strchrnul.S | 53 movk wtmp2, #0x4010, lsl #16 54 dup vrepchr.16b, chrin 64 ld1 {vdata1.16b, vdata2.16b}, [src], #32 66 cmeq vhas_chr1.16b, vdata1.16b, vrepchr.16b 67 cmeq vhas_chr2.16b, vdata2.16b, vrepchr.16b 68 cmhs vhas_nul1.16b, vhas_chr1.16b, vdata1.16b 69 cmhs vhas_nul2.16b, vhas_chr2.16b, vdata2.16b 70 and vhas_chr1.16b, vhas_nul1.16b, vrepmask.16b 71 and vhas_chr2.16b, vhas_nul2.16b, vrepmask.16b 73 addp vend1.16b, vhas_chr1.16b, vhas_chr2.16b // 256->128 [all …]
|
H A D | strrchr-mte.S | 38 For each 16-byte chunk we calculate a 64-bit syndrome value, with 46 dup vrepchr.16b, chrin 47 movi vrepmask.16b, 0x33 48 ld1 {vdata.16b}, [src] 49 cmeq vhas_nul.16b, vdata.16b, 0 50 cmeq vhas_chr.16b, vdata.16b, vrepchr.16b 51 bit vhas_nul.16b, vhas_chr.16b, vrepmask.16b 63 ldr q1, [src, 16] 64 cmeq vhas_chr.16b, vdata.16b, vrepchr.16b 65 cmhs vhas_nul.16b, vhas_chr.16b, vdata.16b [all …]
|
/freebsd/lib/libc/aarch64/string/ |
H A D | strncmp.S | 29 * into unmapped page. If so, we load 16 bytes from the nearest 33 add x3, x0, #16 // end of head 34 add x4, x1, #16 38 cmp x2,#16 50 cmeq v5.16b, v0.16b, #0 51 cmeq v6.16b, v1.16b, #0 66 tbl v0.16b, {v0.16b}, v4.16b 77 tbl v4.16b, {v1.16b}, v4.16b 87 cmeq v2.16b, v0.16b, #0 // NUL byte present? 88 cmeq v4.16b, v0.16b, v4.16b // which bytes match? [all …]
|
/freebsd/contrib/llvm-project/llvm/lib/Target/Mips/MCTargetDesc/ |
H A D | MipsAsmBackend.cpp | 81 // We now check if Value can be encoded as a 16-bit signed immediate. in adjustFixupValue() 82 if (!isInt<16>(Value)) { in adjustFixupValue() 110 // Get the 2nd 16-bits. Also add 1 if bit 15 is 1. in adjustFixupValue() 111 Value = ((Value + 0x8000) >> 16) & 0xffff; in adjustFixupValue() 115 // Get the 3rd 16-bits. in adjustFixupValue() 120 // Get the 4th 16-bits. in adjustFixupValue() 150 // We now check if Value can be encoded as a 16-bit signed immediate. in adjustFixupValue() 151 if (!isInt<16>(Value)) { in adjustFixupValue() 362 { "fixup_Mips_16", 0, 16, 0 }, in getFixupKindInfo() 366 { "fixup_Mips_HI16", 0, 16, 0 }, in getFixupKindInfo() [all …]
|
/freebsd/crypto/openssl/crypto/modes/ |
H A D | cts128.c | 29 const void *key, unsigned char ivec[16], in CRYPTO_cts128_encrypt_block() argument 34 if (len <= 16) in CRYPTO_cts128_encrypt_block() 37 if ((residue = len % 16) == 0) in CRYPTO_cts128_encrypt_block() 38 residue = 16; in CRYPTO_cts128_encrypt_block() 50 memcpy(out, out - 16, residue); in CRYPTO_cts128_encrypt_block() 51 memcpy(out - 16, ivec, 16); in CRYPTO_cts128_encrypt_block() 59 unsigned char ivec[16], in CRYPTO_nistcts128_encrypt_block() argument 64 if (len < 16) in CRYPTO_nistcts128_encrypt_block() 67 residue = len % 16; in CRYPTO_nistcts128_encrypt_block() 82 memcpy(out - 16 + residue, ivec, 16); in CRYPTO_nistcts128_encrypt_block() [all …]
|