/freebsd/sys/contrib/openzfs/module/icp/asm-x86_64/sha2/ |
H A D | sha512-x86_64.S | 126 pushq %r13 127 .cfi_offset %r13,-40 157 movq %r8,%r13 160 rorq $23,%r13 162 xorq %r8,%r13 168 rorq $4,%r13 172 xorq %r8,%r13 178 rorq $14,%r13 182 addq %r13,%r12 189 movq %rdx,%r13 [all …]
|
/freebsd/sys/crypto/openssl/amd64/ |
H A D | sha512-x86_64.S | 32 pushq %r13 33 .cfi_offset %r13,-40 65 movq %r8,%r13 68 rorq $23,%r13 71 xorq %r8,%r13 79 rorq $4,%r13 84 xorq %r8,%r13 92 rorq $14,%r13 97 addq %r13,%r12 106 movq %rdx,%r13 [all …]
|
H A D | rsaz-avx512.S | 31 pushq %r13 33 .cfi_offset %r13,-40 60 movq 0(%r11),%r13 62 vpbroadcastq %r13,%ymm3 64 mulxq %r13,%r13,%r12 65 addq %r13,%r9 69 movq %r8,%r13 70 imulq %r9,%r13 71 andq %rax,%r13 73 vpbroadcastq %r13,%ymm4 [all …]
|
H A D | cmll-x86_64.S | 30 pushq %r13 32 .cfi_offset %r13,-32 42 movq %rcx,%r13 63 movl %r8d,0(%r13) 65 movl %r9d,4(%r13) 66 movl %r10d,8(%r13) 67 movl %r11d,12(%r13) 73 movq 16(%rsp),%r13 74 .cfi_restore %r13 321 pushq %r13 [all …]
|
H A D | keccak1600-x86_64.S | 43 movq %rcx,%r13 60 xorq %r13,%rbp 68 movq %r9,%r13 86 xorq %r13,%r10 89 andq %r8,%r13 91 xorq %r14,%r13 93 movq %r13,-68(%rsi) 111 movq %r8,%r13 119 andq %r13,%r9 137 xorq %r13,%r14 [all …]
|
H A D | x86_64-mont.S | 33 pushq %r13 34 .cfi_offset %r13,-40 93 movq %rdx,%r13 100 addq %rax,%r13 103 addq %r11,%r13 106 movq %r13,-16(%rsp,%r15,8) 107 movq %rdx,%r13 121 addq %rax,%r13 124 addq %r11,%r13 126 movq %r13,-16(%rsp,%r15,8) [all …]
|
H A D | ecp_nistz256-x86_64.S | 2411 pushq %r13 2413 .cfi_offset %r13,-24 2417 xorq %r13,%r13 2428 adcq $0,%r13 2436 sbbq $0,%r13 2447 movq 0(%rsp),%r13 2448 .cfi_restore %r13 2468 pushq %r13 2470 .cfi_offset %r13,-24 2481 xorq %r13,%r13 [all …]
|
H A D | bsaes-x86_64.S | 1095 pushq %r13 1097 .cfi_offset %r13,-40 1110 movq %rsi,%r13 1162 movdqu %xmm15,0(%r13) 1164 movdqu %xmm0,16(%r13) 1165 movdqu %xmm5,32(%r13) 1166 movdqu %xmm3,48(%r13) 1167 movdqu %xmm1,64(%r13) 1168 movdqu %xmm6,80(%r13) 1169 movdqu %xmm2,96(%r13) [all …]
|
H A D | x86_64-mont5.S | 28 pushq %r13 29 .cfi_offset %r13,-40 218 movq %rdx,%r13 225 addq %rax,%r13 228 addq %r11,%r13 231 movq %r13,-16(%rsp,%r15,8) 232 movq %rdx,%r13 247 addq %rax,%r13 249 addq %r11,%r13 251 movq %r13,-16(%rsp,%r9,8) [all …]
|
H A D | rsaz-x86_64.S | 20 pushq %r13 22 .cfi_offset %r13,-40 75 movq %rdx,%r13 76 adcq $0,%r13 79 addq %rax,%r13 126 addq %rax,%r13 129 addq %rbx,%r13 174 addq %rax,%r13 177 addq %rcx,%r13 256 addq %r13,%r13 [all …]
|
H A D | x25519-x86_64.S | 18 pushq %r13 20 .cfi_offset %r13,-40 34 movq 16(%rdx),%r13 51 mulq %r13 52 movq %r13,16(%rsp) 60 movq %rdx,%r13 82 adcq %rdx,%r13 105 adcq %rdx,%r13 128 adcq %rdx,%r13 150 adcq %rdx,%r13 [all …]
|
H A D | rsaz-avx2.S | 17 pushq %r13 18 .cfi_offset %r13,-40 26 movq %rdx,%r13 28 movq %r13,%r15 31 subq $-128,%r13 44 vmovdqu 0-128(%r13),%ymm0 46 vmovdqu 32-128(%r13),%ymm1 47 vmovdqu 64-128(%r13),%ymm2 48 vmovdqu 96-128(%r13),%ymm3 49 vmovdqu 128-128(%r13),%ymm4 [all …]
|
H A D | aes-x86_64.S | 218 movzbl (%r14,%r13,1),%r13d 350 pushq %r13 351 .cfi_offset %r13,-40 380 leaq (%r15,%r13,1),%rbp 405 movq -32(%rsi),%r13 406 .cfi_restore %r13 640 movzbl (%r14,%r13,1),%r13d 695 movq %rsi,%r13 698 andq %r11,%r13 700 movq %r13,%rdx [all …]
|
H A D | sha1-x86_64.S | 37 pushq %r13 38 .cfi_offset %r13,-40 69 leal 1518500249(%rdx,%r13,1),%r13d 134 leal 1518500249(%r14,%r13,1),%r13d 199 leal 1518500249(%rbp,%r13,1),%r13d 265 leal 1518500249(%rdx,%r13,1),%r13d 339 leal 1859775393(%r14,%r13,1),%r13d 409 leal 1859775393(%rbp,%r13,1),%r13d 479 leal 1859775393(%rdx,%r13,1),%r13d 549 leal 1859775393(%r14,%r13,1),%r13d [all …]
|
/freebsd/crypto/openssl/crypto/bn/asm/ |
H A D | rsaz-x86_64.pl | 111 push %r13 112 .cfi_push %r13 167 movq %rdx, %r13 168 adcq \$0, %r13 171 addq %rax, %r13 218 addq %rax, %r13 221 addq %rbx, %r13 266 addq %rax, %r13 269 addq %rcx, %r13 348 addq %r13, %r13 [all …]
|
H A D | rsaz-avx512.pl | 143 movq $_b_offset($b_ptr), %r13 # b[i] 145 vpbroadcastq %r13, $Bi # broadcast b[i] 147 mulx %r13, %r13, %r12 # a[0]*b[i] = (t0,t2) 148 addq %r13, $_acc # acc += t0 152 movq $_k0, %r13 153 imulq $_acc, %r13 # acc * k0 154 andq $mask52, %r13 # yi = (acc * k0) & mask52 156 vpbroadcastq %r13, $Yi # broadcast y[i] 158 mulx %r13, %r13, %r12 # yi * m[0] = (t0,t1) 159 addq %r13, $_acc # acc += t0 [all …]
|
H A D | x86_64-mont5.pl | 83 $hi1="%r13"; 120 push %r13 121 .cfi_push %r13 453 mov -32(%rsi),%r13 454 .cfi_restore %r13 470 my @N=("%r13","%rdi"); 493 push %r13 494 .cfi_push %r13 567 mov -32(%rsi),%r13 568 .cfi_restore %r13 [all …]
|
H A D | x86_64-mont.pl | 94 $hi1="%r13"; 136 push %r13 137 .cfi_push %r13 351 mov -32(%rsi),%r13 352 .cfi_restore %r13 368 my @N=("%r13","%rdi"); 391 push %r13 392 .cfi_push %r13 804 mov -32(%rsi),%r13 805 .cfi_restore %r13 [all …]
|
/freebsd/crypto/openssl/crypto/ec/asm/ |
H A D | x25519-x86_64.pl | 114 push %r13 115 .cfi_push %r13 127 mov 8*2(%rdx),%r13 144 mulq %r13 # f[0]*g[2] 145 mov %r13,8*2(%rsp) # offload g[2] 151 mov %rax,%r12 # %r12:%r13 = h3 153 mov %rdx,%r13 175 adc %rdx,%r13 198 adc %rdx,%r13 221 adc %rdx,%r13 [all …]
|
/freebsd/crypto/openssl/crypto/aes/asm/ |
H A D | aesni-sha256-x86_64.pl | 360 push %r13 361 .cfi_push %r13 402 lea $TABLE+`$SZ*2*$rounds+32`(%rip),%r13 # borrow $a0 418 vmovdqa 0x00(%r13,%r14,8),$mask14 419 vmovdqa 0x10(%r13,%r14,8),$mask12 420 vmovdqa 0x20(%r13,%r14,8),$mask10 586 mov $_out,%r13 # borrow $a0 593 vmovdqu $iv,(%r13,%r12) # write output 641 mov -32(%rsi),%r13 642 .cfi_restore %r13 [all …]
|
/freebsd/crypto/openssl/crypto/camellia/asm/ |
H A D | cmll-x86_64.pl | 66 $out="%r13"; 149 push %r13 150 .cfi_push %r13 189 mov 16(%rsp),%r13 190 .cfi_restore %r13 276 push %r13 277 .cfi_push %r13 316 mov 16(%rsp),%r13 317 .cfi_restore %r13 453 push %r13 [all …]
|
/freebsd/crypto/openssl/crypto/sha/asm/ |
H A D | keccak1600-x86_64.pl | 361 push %r13 362 .cfi_push %r13 399 pop %r13 400 .cfi_pop %r13 426 push %r13 427 .cfi_push %r13 493 pop %r13 494 .cfi_pop %r13 507 ($out,$len,$bsz) = ("%r12","%r13","%r14"); 517 push %r13 [all …]
|
/freebsd/contrib/llvm-project/compiler-rt/lib/builtins/hexagon/ |
H A D | udivdi3.S | 37 r13:12 = lsl(r5:4,r10) // shift divisor msb into same bit position as dividend msb 50 p0 = cmp.gtu(r13:12,r3:2) // set predicate reg if shifted divisor > current remainder 53 r7:6 = sub(r3:2, r13:12) // subtract shifted divisor from current remainder 62 r13:12 = lsr(r13:12, #1) // shift "shifted divisor" right by 1 for next iteration
|
H A D | udivmoddi4.S | 37 r13:12 = lsl(r5:4,r10) // shift divisor msb into same bit position as dividend msb 50 p0 = cmp.gtu(r13:12,r3:2) // set predicate reg if shifted divisor > current remainder 53 r7:6 = sub(r3:2, r13:12) // subtract shifted divisor from current remainder 62 r13:12 = lsr(r13:12, #1) // shift "shifted divisor" right by 1 for next iteration
|
H A D | umoddi3.S | 37 r13:12 = lsl(r5:4,r10) // shift divisor msb into same bit position as dividend msb 50 p0 = cmp.gtu(r13:12,r3:2) // set predicate reg if shifted divisor > current remainder 53 r7:6 = sub(r3:2, r13:12) // subtract shifted divisor from current remainder 62 r13:12 = lsr(r13:12, #1) // shift "shifted divisor" right by 1 for next iteration
|