| /linux/arch/x86/power/ |
| H A D | hibernate_asm_64.S | 34 movq mmu_cr4_features(%rip), %rax 35 movq %rax, %rdx 40 movq %rax, %cr4; # turn PGE back on 43 movq $saved_context, %rax 44 movq pt_regs_sp(%rax), %rsp 45 movq pt_regs_bp(%rax), %rbp 46 movq pt_regs_si(%rax), %rsi 47 movq pt_regs_di(%rax), %rdi 48 movq pt_regs_bx(%rax), %rbx 49 movq pt_regs_cx(%rax), %rcx [all …]
|
| /linux/arch/x86/hyperv/ |
| H A D | mshv_vtl_asm.S | 39 mov %_ASM_ARG1, %rax 42 mov MSHV_VTL_CPU_CONTEXT_rbx(%rax), %rbx 43 mov MSHV_VTL_CPU_CONTEXT_rbp(%rax), %rbp 44 mov MSHV_VTL_CPU_CONTEXT_rsi(%rax), %rsi 45 mov MSHV_VTL_CPU_CONTEXT_rdi(%rax), %rdi 46 mov MSHV_VTL_CPU_CONTEXT_r8(%rax), %r8 47 mov MSHV_VTL_CPU_CONTEXT_r9(%rax), %r9 48 mov MSHV_VTL_CPU_CONTEXT_r10(%rax), %r10 49 mov MSHV_VTL_CPU_CONTEXT_r11(%rax), %r11 50 mov MSHV_VTL_CPU_CONTEXT_r12(%rax), %r12 [all …]
|
| /linux/arch/x86/lib/ |
| H A D | bhi.S | 39 cmovne %rax, %rdi 56 cmovne %rax, %rdi 57 cmovne %rax, %rsi 67 cmovne %rax, %rdi 68 cmovne %rax, %rsi 69 cmovne %rax, %rdx 79 cmovne %rax, %rdi 80 cmovne %rax, %rsi 81 cmovne %rax, %rdx 82 cmovne %rax, %rcx [all …]
|
| H A D | copy_user_64.S | 60 2: movq (%rsi),%rax 61 3: movq %rax,(%rdi) 81 0: movq (%rsi),%rax 82 1: movq %rax,(%rdi) 88 leaq 8(%rdi),%rax 89 andq $-8,%rax 90 subq %rdi,%rax 93 addq %rax,%rdi 94 addq %rax,%rsi 95 subq %rax,%rcx [all …]
|
| H A D | csum-copy_64.S | 95 adcq %rbx, %rax 96 adcq %r8, %rax 97 adcq %r11, %rax 98 adcq %rdx, %rax 99 adcq %r10, %rax 100 adcq %r15, %rax 101 adcq %r14, %rax 102 adcq %r13, %rax 129 adcq %r9, %rax 143 adcq %rbx, %rax [all …]
|
| /linux/tools/arch/x86/lib/ |
| H A D | memset_64.S | 39 movq %r9,%rax 53 movabs $0x0101010101010101,%rax 54 imulq %rcx,%rax 69 movq %rax,(%rdi) 70 movq %rax,8(%rdi) 71 movq %rax,16(%rdi) 72 movq %rax,24(%rdi) 73 movq %rax,32(%rdi) 74 movq %rax,40(%rdi) 75 movq %rax,48(%rdi) [all …]
|
| /linux/arch/x86/platform/efi/ |
| H A D | efi_thunk_64.S | 37 movq %rsp, %rax 39 push %rax 45 movq 0x18(%rax), %rbp 46 movq 0x20(%rax), %rbx 47 movq 0x28(%rax), %rax 55 movq $__START_KERNEL_map, %rax 56 subq phys_base(%rip), %rax 60 subq %rax, %rbp 61 subq %rax, %rbx
|
| /linux/arch/x86/kernel/ |
| H A D | ftrace_64.S | 83 movq %rax, RAX(%rsp) 128 movq RAX(%rsp), %rax 185 movq RIP(%rsp), %rax 186 movq %rax, MCOUNT_REG_SIZE(%rsp) 250 movq EFLAGS(%rsp), %rax 251 movq %rax, MCOUNT_REG_SIZE(%rsp) 254 movq RIP(%rsp), %rax 255 movq %rax, MCOUNT_REG_SIZE+8(%rsp) 265 movq ORIG_RAX(%rsp), %rax 266 movq %rax, MCOUNT_REG_SIZE-8(%rsp) [all …]
|
| H A D | head_64.S | 78 leaq .Lon_kernel_cs(%rip), %rax 79 pushq %rax 118 addq %rcx, %rax 121 mov %rax, %rdi 136 movq %rax, %cr3 182 movq phys_base(%rip), %rax 183 addq $(init_top_pgt - __START_KERNEL_map), %rax 190 addq sme_me_mask(%rip), %rax 196 movq %rax, %cr3 305 movq trampoline_lock(%rip), %rax [all …]
|
| /linux/lib/crypto/x86/ |
| H A D | poly1305-x86_64-cryptogams.pl | 187 mov %rax,$d2 188 mov $r0,%rax 192 mov %rax,$h0 # future $h0 193 mov $r0,%rax 197 add %rax,$d2 198 mov $s1,%rax 203 add %rax,$h0 213 mov \$-4,%rax # mask value 216 and $d3,%rax # last reduction step 220 add $d3,%rax [all …]
|
| H A D | blake2s-core.S | 92 movd (DATA,%rax,4),%xmm4 94 movd (DATA,%rax,4),%xmm5 96 movd (DATA,%rax,4),%xmm6 98 movd (DATA,%rax,4),%xmm7 113 movd (DATA,%rax,4),%xmm5 115 movd (DATA,%rax,4),%xmm6 117 movd (DATA,%rax,4),%xmm7 119 movd (DATA,%rax,4),%xmm4 137 movd (DATA,%rax,4),%xmm6 139 movd (DATA,%rax,4),%xmm7 [all …]
|
| H A D | chacha-avx2-x86_64.S | 67 mov %rcx,%rax 142 cmp $0x10,%rax 149 cmp $0x20,%rax 156 cmp $0x30,%rax 163 cmp $0x40,%rax 171 cmp $0x50,%rax 177 cmp $0x60,%rax 183 cmp $0x70,%rax 189 cmp $0x80,%rax 200 mov %rax,%r9 [all …]
|
| H A D | chacha-ssse3-x86_64.S | 132 mov %rcx,%rax 137 cmp $0x10,%rax 145 cmp $0x20,%rax 153 cmp $0x30,%rax 161 cmp $0x40,%rax 173 mov %rax,%r9 176 and $~0x0f,%rax 184 lea (%rdx,%rax),%rsi 193 lea (%r11,%rax),%rdi 243 mov %rcx,%rax [all …]
|
| /linux/tools/testing/selftests/sgx/ |
| H A D | test_encl_bootstrap.S | 47 lea __encl_base(%rip), %rax 48 sub %rax, %rbx 49 lea encl_stack(%rip), %rax 50 add %rbx, %rax 55 lea -1(%rbx), %rax 57 xchg %rsp, %rax 58 push %rax 72 pop %rax 73 mov %rax, %rsp 76 mov $4, %rax
|
| /linux/arch/x86/crypto/ |
| H A D | aes-gcm-aesni-x86_64.S | 216 mov (\src, %rcx), %rax // Load last 8 bytes 219 shr %cl, %rax // Discard overlapping bytes 220 pinsrq $1, %rax, \dst 241 or \tmp64, %rax // Combine the two parts 243 movq %rax, \dst 254 pextrq $1, \src, %rax 257 ror %cl, %rax 258 mov %rax, (\dst, %rsi) // Store last LEN - 8 bytes 436 movdqa OFFSETOF_H_POWERS(KEY,%rax,2), TMP0 440 movdqu (DST,%rax,2), TMP1 [all …]
|
| H A D | aes-gcm-vaes-avx2.S | 243 lea 6*16(KEY,%rax,4), RNDKEYLAST_PTR 245 lea 16(KEY), %rax 247 vaesenc (%rax), H_CUR_XMM, H_CUR_XMM 248 add $16, %rax 249 cmp %rax, RNDKEYLAST_PTR 412 mov (\src, %rcx), %rax // Load last 8 bytes 415 shr %cl, %rax // Discard overlapping bytes 416 vpinsrq $1, %rax, \dst, \dst 437 or \tmp64, %rax // Combine the two parts 439 vmovq %rax, \dst [all …]
|
| H A D | aes-gcm-vaes-avx512.S | 304 lea 6*16(KEY,%rax,4), RNDKEYLAST_PTR 593 mov $-1, %rax 594 bzhi AADLEN64, %rax, %rax 595 kmovq %rax, %k1 837 lea 16(KEY), %rax 839 vbroadcasti32x4 (%rax), RNDKEY 841 add $16, %rax 842 cmp %rax, RNDKEYLAST_PTR 932 neg %rax 933 and $~15, %rax // -round_up(DATALEN, 16) [all …]
|
| H A D | aes-ctr-avx-x86_64.S | 124 mov (\src, %rcx), %rax // Load last 8 bytes 127 shr %cl, %rax // Discard overlapping bytes 128 vpinsrq $1, %rax, \dst, \dst 149 or \tmp64, %rax // Combine the two parts 151 vmovq %rax, \dst 162 vpextrq $1, \src, %rax 165 ror %cl, %rax 166 mov %rax, (\dst, \tmp64) // Store last LEN - 8 bytes 228 mov KEY, %rax 230 _vbroadcast128 (%rax), RNDKEY [all …]
|
| H A D | aria-aesni-avx-asm_64.S | 886 movq %rsi, %rax; 887 leaq 8 * 16(%rax), %r8; 891 %xmm15, %rax, %r8); 894 %rax, %r9, 0); 897 %xmm15, %rax, %r9, 1); 900 %rax, %r9, 2); 903 %xmm15, %rax, %r9, 3); 906 %rax, %r9, 4); 909 %xmm15, %rax, %r9, 5); 912 %rax, %r9, 6); [all …]
|
| /linux/arch/x86/entry/ |
| H A D | calling.h | 68 .macro PUSH_REGS rdx=%rdx rcx=%rcx rax=%rax save_ret=0 unwind_hint=1 81 pushq \rax /* pt_regs->ax */ 126 .macro PUSH_AND_CLEAR_REGS rdx=%rdx rcx=%rcx rax=%rax save_ret=0 clear_callee=1 unwind_hint=1 127 PUSH_REGS rdx=\rdx, rcx=\rcx, rax=\rax, save_ret=\save_ret unwind_hint=\unwind_hint 142 popq %rax 224 pushq %rax 225 SWITCH_TO_USER_CR3 scratch_reg=\scratch_reg scratch_reg2=%rax 226 popq %rax 313 or %rdx, %rax 314 mov %rax, \save_reg [all …]
|
| /linux/arch/x86/virt/vmx/tdx/ |
| H A D | tdxcall.S | 49 mov %rdi, %rax 123 pushq %rax /* save the TDCALL/SEAMCALL return code */ 124 movq 8(%rsp), %rax /* restore the structure pointer */ 125 movq %rsi, TDX_MODULE_rsi(%rax) /* save RSI */ 126 popq %rax /* restore the return code */ 195 mov $TDX_SEAMCALL_VMFAILINVALID, %rax 208 orq %rdi, %rax
|
| /linux/arch/x86/entry/vdso/ |
| H A D | vsgx.S | 107 mov %rbx, %rax 118 push %rax 121 push %rax 130 mov SGX_ENCLAVE_RUN_USER_HANDLER(%rax), %rax 132 call *%rax
|
| /linux/arch/x86/mm/ |
| H A D | mem_encrypt_boot.S | 37 movq %rcx, %rax /* Workarea stack page */ 38 leaq PAGE_SIZE(%rax), %rsp /* Set new stack pointer */ 39 addq $PAGE_SIZE, %rax /* Workarea encryption routine */ 47 movq %rax, %rdi /* Workarea encryption routine */ 57 movq %rax, %r8 /* Workarea encryption routine */ 61 call *%rax /* Call the encryption routine */
|
| /linux/arch/x86/purgatory/ |
| H A D | entry64.S | 32 leaq new_cs_exit(%rip), %rax 33 pushq %rax 38 movq rax(%rip), %rax 62 rax: .quad 0x0 label
|
| /linux/arch/x86/include/asm/ |
| H A D | div64.h | 106 static inline u64 mul_u64_add_u64_div_u64(u64 rax, u64 mul, u64 add, u64 div) in mul_u64_add_u64_div_u64() argument 110 asm ("mulq %[mul]" : "+a" (rax), "=d" (rdx) : [mul] "rm" (mul)); in mul_u64_add_u64_div_u64() 114 [lo] "+r" (rax), [hi] "+r" (rdx) : [add] "irm" (add)); in mul_u64_add_u64_div_u64() 116 asm ("divq %[div]" : "+a" (rax), "+d" (rdx) : [div] "rm" (div)); in mul_u64_add_u64_div_u64() 118 return rax; in mul_u64_add_u64_div_u64()
|