/freebsd/stand/kboot/kboot/arch/powerpc64/ |
H A D | kerneltramp.S | 54 mfmsr %r10 55 andi. %r10, %r10, 1 /* test MSR_LE */ 66 lwz %r10, 0(%r8) 67 mtctr %r10 74 li %r10, 4 75 lwbrx %r3, %r8, %r10 76 li %r10, 8 77 lwbrx %r4, %r8, %r10 78 li %r10, 12 79 lwbrx %r5, %r8, %r10 [all …]
|
/freebsd/sys/crypto/openssl/arm/ |
H A D | sha1-armv4-large.S | 33 stmdb sp!,{r4,r5,r6,r7,r8,r9,r10,r11,r12,lr} 45 ldrb r10,[r1,#2] 50 orr r9,r9,r10,lsl#8 51 eor r10,r5,r6 @ F_xx_xx 58 eor r10,r5,r6 @ F_xx_xx 64 and r10,r4,r10,ror#2 66 eor r10,r10,r6,ror#2 @ F_00_19(B,C,D) 68 add r7,r7,r10 @ E+=F_00_19(B,C,D) 70 ldrb r10,[r1,#2] 75 orr r9,r9,r10,lsl#8 [all …]
|
H A D | keccak1600-armv4.S | 47 add r10,sp,#40 58 ldmia r10,{r10,r11,r12,r14} @ A[1][0..1] 60 eor r0,r0,r10 63 ldrd r10,r11,[sp,#56] 66 eor r4,r4,r10 69 ldrd r10,r11,[sp,#72] 72 eor r8,r8,r10 75 ldrd r10,r11,[sp,#88] 78 eor r2,r2,r10 81 ldrd r10,r11,[sp,#104] [all …]
|
H A D | chacha-armv4.S | 71 ldmia r3,{r4,r5,r6,r7,r8,r9,r10,r11} @ load key 73 stmdb sp!,{r4,r5,r6,r7,r8,r9,r10,r11} @ copy key 75 str r10,[sp,#4*(16+10)] @ off-load "rx" 88 ldr r10, [sp,#4*(13)] 100 mov r10,r10,ror#16 102 eor r10,r10,r1,ror#16 105 add r9,r9,r10 112 mov r10,r10,ror#24 114 eor r10,r10,r1,ror#24 117 add r9,r9,r10 [all …]
|
/freebsd/sys/crypto/openssl/amd64/ |
H A D | x86_64-mont5.S | 37 leaq -280(%rsp,%r9,8),%r10 39 andq $-1024,%r10 49 subq %r10,%r11 51 leaq (%r10,%r11,1),%rsp 53 cmpq %r10,%rsp 60 cmpq %r10,%rsp 64 leaq .Linc(%rip),%r10 70 movdqa 0(%r10),%xmm0 71 movdqa 16(%r10),%xmm1 72 leaq 24-112(%rsp,%r9,8),%r10 [all …]
|
H A D | x86_64-mont.S | 42 leaq -16(%rsp,%r9,8),%r10 44 andq $-1024,%r10 54 subq %r10,%r11 56 leaq (%r10,%r11,1),%rsp 58 cmpq %r10,%rsp 66 cmpq %r10,%rsp 83 movq %rax,%r10 86 imulq %r10,%rbp 90 addq %rax,%r10 104 movq %r10,%r11 [all …]
|
H A D | md5-x86_64.S | 55 leal -680876936(%rax,%r10,1),%eax 64 leal -389564586(%rdx,%r10,1),%edx 73 leal 606105819(%rcx,%r10,1),%ecx 82 leal -1044525330(%rbx,%r10,1),%ebx 91 leal -176418897(%rax,%r10,1),%eax 100 leal 1200080426(%rdx,%r10,1),%edx 109 leal -1473231341(%rcx,%r10,1),%ecx 118 leal -45705983(%rbx,%r10,1),%ebx 127 leal 1770035416(%rax,%r10,1),%eax 136 leal -1958414417(%rdx,%r10,1),%edx [all …]
|
H A D | ecp_nistz256-x86_64.S | 2420 movq 16(%rsi),%r10 2425 adcq %r10,%r10 2431 movq %r10,%rcx 2433 sbbq 16(%rsi),%r10 2441 cmovcq %rcx,%r10 2444 movq %r10,16(%rdi) 2475 movq 16(%rsi),%r10 2483 movq %r10,%rcx 2485 adcq 16(%rsi),%r10 2494 cmovzq %rcx,%r10 [all …]
|
H A D | sha512-x86_64.S | 55 movq 48(%rdi),%r10 73 xorq %r10,%r15 81 xorq %r10,%r15 121 addq %r10,%r12 134 movq %rax,%r10 140 xorq %r15,%r10 142 addq %r12,%r10 145 addq %r14,%r10 148 movq %r10,%r14 158 xorq %r10,%r14 [all …]
|
H A D | x86_64cpuid.S | 95 movzbq %cl,%r10 96 incq %r10 257 xorq %r10,%r10 262 movq (%rdi),%r10 265 xorq (%rsi),%r10 267 orq %r11,%r10 314 xorq %r10,%r10 326 movq %rdi,%r10 333 clflush (%r10) 335 addl %r9d,(%r10) [all …]
|
H A D | rsaz-avx2.S | 297 movq 8(%rsp),%r10 320 addq %rax,%r10 323 addq %r9,%r10 328 movq %r10,%rax 344 addq %rax,%r10 355 shrq $29,%r10 359 addq %r10,%r11 466 movq 8(%rsp),%r10 479 addq %rax,%r10 487 leaq (%r9,%r10,1),%rax [all …]
|
H A D | rsaz-x86_64.S | 57 movq %rdx,%r10 58 adcq $0,%r10 61 addq %rax,%r10 104 addq %rax,%r10 151 adcq %r10,%r10 160 adcq %rdx,%r10 164 movq %r10,24(%rsp) 213 movq 24(%rsi),%r10 224 mulq %r10 231 mulq %r10 [all …]
|
H A D | rc4-x86_64.S | 28 xorq %r10,%r10 39 subq %r10,%rbx 41 movl (%rdi,%r10,4),%eax 47 leaq 1(%r10),%rsi 54 movl %edx,(%rdi,%r10,4) 58 movl (%rdi,%r10,4),%eax 65 leaq 1(%r10),%rsi 74 movl %edx,0(%rdi,%r10,4) 82 movl %edx,4(%rdi,%r10,4) 90 movl %edx,8(%rdi,%r10,4) [all …]
|
H A D | keccak1600-x86_64.S | 19 movq -4(%rdi),%r10 31 xorq %r10,%rcx 62 xorq %rdx,%r10 66 rolq $43,%r10 70 orq %r10,%r9 80 xorq %r10,%r12 81 notq %r10 84 orq %r11,%r10 86 xorq %r13,%r10 87 movq %r10,-92(%rsi) [all …]
|
H A D | ghash-x86_64.S | 51 movq %r9,%r10 55 shlq $60,%r10 61 xorq %r10,%r8 67 movq %r9,%r10 70 shlq $60,%r10 75 xorq %r10,%r8 82 movq %r9,%r10 85 shlq $60,%r10 90 xorq %r10,%r8 94 movq %r9,%r10 [all …]
|
/freebsd/crypto/openssl/crypto/bn/asm/ |
H A D | x86_64-mont5.pl | 81 $lo0="%r10"; 129 lea -280(%rsp,$num,8),%r10 # future alloca(8*(num+2)+256+8) 131 and \$-1024,%r10 # minimize TLB usage 141 sub %r10,%r11 143 lea (%r10,%r11),%rsp 145 cmp %r10,%rsp 152 cmp %r10,%rsp 156 lea .Linc(%rip),%r10 167 movdqa 0(%r10),%xmm0 # 00000001000000010000000000000000 168 movdqa 16(%r10),%xmm1 # 00000002000000020000000200000002 [all …]
|
H A D | rsaz-x86_64.pl | 149 movq %rdx, %r10 150 adcq \$0, %r10 153 addq %rax, %r10 196 addq %rax, %r10 243 adcq %r10, %r10 252 adcq %rdx, %r10 256 movq %r10, 24(%rsp) 305 movq 24($inp), %r10 316 mulq %r10 323 mulq %r10 [all …]
|
H A D | x86_64-mont.pl | 92 $lo0="%r10"; 145 lea -16(%rsp,$num,8),%r10 # future alloca(8*(num+2)) 147 and \$-1024,%r10 # minimize TLB usage 157 sub %r10,%r11 159 lea (%r10,%r11),%rsp 161 cmp %r10,%rsp 169 cmp %r10,%rsp 367 my @A=("%r10","%r11"); 400 lea -32(%rsp,$num,8),%r10 # future alloca(8*(num+4)) 402 and \$-1024,%r10 # minimize TLB usage [all …]
|
/freebsd/lib/libc/amd64/string/ |
H A D | memset.S | 42 movabs $0x0101010101010101,%r10 43 imulq %r8,%r10 53 movq %r10,(%rdi) 54 movq %r10,8(%rdi) 55 movq %r10,16(%rdi) 56 movq %r10,24(%rdi) 63 movq %r10,-16(%rdi,%rcx) 64 movq %r10,-8(%rdi,%rcx) 71 movq %r10,(%rdi) 72 movq %r10,8(%rdi) [all …]
|
/freebsd/crypto/openssl/crypto/aes/asm/ |
H A D | vpaes-x86_64.pl | 89 ## Clobbers %xmm1-%xmm5, %r9, %r10, %r11, %rax 112 lea .Lk_mc_backward(%rip),%r10 125 movdqa -0x40(%r11,%r10), %xmm1 # .Lk_mc_forward[] 127 movdqa (%r11,%r10), %xmm4 # .Lk_mc_backward[] 168 movdqa -0x60(%r10), %xmm4 # 3 : sbou .Lk_sbo 169 movdqa -0x50(%r10), %xmm0 # 0 : sbot .Lk_sbo+16 173 movdqa 0x40(%r11,%r10), %xmm1 # .Lk_sr[] 202 lea .Lk_dsbd(%rip),%r10 209 add %r10, %r11 217 movdqa -0x20(%r10),%xmm4 # 4 : sb9u [all …]
|
H A D | vpaes-armv8.pl | 162 // Fills register %r10 -> .aes_consts (so you can -fPIC) 187 // Clobbers %xmm1-%xmm5, %r9, %r10, %r11, %rax 213 ld1 {v1.2d}, [x11], #16 // vmovdqa -0x40(%r11,%r10), %xmm1 # .Lk_mc_forward[] 219 ld1 {v4.2d}, [x10] // vmovdqa (%r11,%r10), %xmm4 # .Lk_mc_backward[] 249 // vmovdqa -0x60(%r10), %xmm4 # 3 : sbou .Lk_sbo 250 // vmovdqa -0x50(%r10), %xmm0 # 0 : sbot .Lk_sbo+16 252 ld1 {v1.2d}, [x10] // vmovdqa 0x40(%r11,%r10), %xmm1 # .Lk_sr[] 307 ld1 {v1.2d}, [x11], #16 // vmovdqa -0x40(%r11,%r10), %xmm1 # .Lk_mc_forward[] 318 ld1 {v4.2d}, [x10] // vmovdqa (%r11,%r10), %xmm4 # .Lk_mc_backward[] 367 // vmovdqa -0x60(%r10), %xmm4 # 3 : sbou .Lk_sbo [all …]
|
/freebsd/crypto/openssl/crypto/camellia/asm/ |
H A D | cmll-x86_64.pl | 527 shl \$32,%r10 # @S[2]|| 529 or %r11,%r10 # ||@S[3] 532 &_saveround (2,$out,-128,"%r8","%r10"); # KA<<<0 535 &_rotl128 ("%r8","%r10",15); 536 &_saveround (6,$out,-128,"%r8","%r10"); # KA<<<15 537 &_rotl128 ("%r8","%r10",15); # 15+15=30 538 &_saveround (8,$out,-128,"%r8","%r10"); # KA<<<30 541 &_rotl128 ("%r8","%r10",15); # 30+15=45 545 &_rotl128 ("%r8","%r10",15); # 45+15=60 546 &_saveround (14,$out,-128,"%r8","%r10"); # KA<<<60 [all …]
|
/freebsd/contrib/bionic-x86_64-string/ |
H A D | ssse3-strcmp-slm.S | 138 lea L(unaligned_table)(%rip), %r10 139 movslq (%r10, %r9,4), %r9 140 lea (%r10, %r9), %r10 141 jmp *%r10 /* jump to corresponding case */ 237 lea 1(%rdi), %r10 238 and $0xfff, %r10 /* offset into 4K page */ 239 sub $0x1000, %r10 /* subtract 4K pagesize */ 243 add $16, %r10 267 add $16, %r10 308 sub $0x1000, %r10 /* substract 4K from %r10 */ [all …]
|
/freebsd/sys/contrib/openzfs/module/icp/asm-x86_64/sha2/ |
H A D | sha512-x86_64.S | 148 movq 48(%rdi),%r10 164 xorq %r10,%r15 170 xorq %r10,%r15 201 addq %r10,%r12 211 movq %rax,%r10 215 xorq %r15,%r10 217 addq %r12,%r10 219 addq %r14,%r10 222 movq %r10,%r14 230 xorq %r10,%r14 [all …]
|
/freebsd/crypto/openssl/crypto/ |
H A D | x86_64cpuid.pl | 123 movzb %cl,%r10 # number of cores - 1 124 inc %r10 # number of cores 285 xor %r10,%r10 290 mov ($arg1),%r10 293 xor ($arg2),%r10 295 or %r11,%r10 345 xorq %r10,%r10 367 xorq %r10,%r10 374 my $out="%r10"; 485 ${rdop} %r10 [all …]
|