| /linux/arch/x86/include/asm/ |
| H A D | cmpxchg.h | 47 asm_inline volatile (lock #op "b %b0, %1" \ 52 asm_inline volatile (lock #op "w %w0, %1" \ 57 asm_inline volatile (lock #op "l %0, %1" \ 62 asm_inline volatile (lock #op "q %q0, %1" \ 94 asm_inline volatile(lock "cmpxchgb %2, %1" \ 103 asm_inline volatile(lock "cmpxchgw %2, %1" \ 112 asm_inline volatile(lock "cmpxchgl %2, %1" \ 121 asm_inline volatile(lock "cmpxchgq %2, %1" \ 168 asm_inline volatile(lock "cmpxchgb %[new], %[ptr]" \ 179 asm_inline volatile(lock "cmpxchgw %[new], %[ptr]" \ [all …]
|
| H A D | vmware.h | 111 asm_inline volatile (VMWARE_HYPERCALL in vmware_hypercall1() 136 asm_inline volatile (VMWARE_HYPERCALL in vmware_hypercall3() 161 asm_inline volatile (VMWARE_HYPERCALL in vmware_hypercall4() 187 asm_inline volatile (VMWARE_HYPERCALL in vmware_hypercall5() 215 asm_inline volatile (VMWARE_HYPERCALL in vmware_hypercall6() 243 asm_inline volatile (VMWARE_HYPERCALL in vmware_hypercall7() 277 asm_inline volatile ( in vmware_hypercall_hb_out() 305 asm_inline volatile ( in vmware_hypercall_hb_in()
|
| H A D | bitops.h | 55 asm_inline volatile(LOCK_PREFIX "orb %b1,%0" in arch_set_bit() 60 asm_inline volatile(LOCK_PREFIX __ASM_SIZE(bts) " %1,%0" in arch_set_bit() 75 asm_inline volatile(LOCK_PREFIX "andb %b1,%0" in arch_clear_bit() 79 asm_inline volatile(LOCK_PREFIX __ASM_SIZE(btr) " %1,%0" in arch_clear_bit() 101 asm_inline volatile(LOCK_PREFIX "xorb %2,%1" in arch_xor_unlock_is_negative_byte() 124 asm_inline volatile(LOCK_PREFIX "xorb %b1,%0" in arch_change_bit() 128 asm_inline volatile(LOCK_PREFIX __ASM_SIZE(btc) " %1,%0" in arch_change_bit()
|
| H A D | alternative.h | 247 asm_inline volatile(ALTERNATIVE(oldinstr, newinstr, ft_flags) : : : "memory") 250 …asm_inline volatile(ALTERNATIVE_2(oldinstr, newinstr1, ft_flags1, newinstr2, ft_flags2) ::: "memor… 261 asm_inline volatile(ALTERNATIVE(oldinstr, newinstr, ft_flags) \ 266 asm_inline volatile(ALTERNATIVE(oldinstr, newinstr, ft_flags) \ 280 asm_inline volatile(ALTERNATIVE("call %c[old]", "call %c[new]", ft_flags) \ 294 asm_inline volatile(ALTERNATIVE_2("call %c[old]", "call %c[new1]", ft_flags1, \
|
| H A D | cmpxchg_32.h | 22 asm_inline volatile(_lock "cmpxchg8b %[ptr]" \ 48 asm_inline volatile(_lock "cmpxchg8b %[ptr]" \ 93 asm_inline volatile( \ 123 asm_inline volatile( \
|
| H A D | runtime-const.h | 23 asm_inline("mov %1,%0\n1:\n" \ 37 asm_inline("shrl $12,%k0\n1:\n" \
|
| H A D | cmpxchg_64.h | 41 asm_inline volatile(_lock "cmpxchg16b %[ptr]" \ 68 asm_inline volatile(_lock "cmpxchg16b %[ptr]" \
|
| H A D | bug.h | 88 asm_inline volatile("1:\t" ins "\n" \ 160 asm_inline volatile("lea (2f)(%%rip), %[addr]\n1:\n" \
|
| H A D | gsseg.h | 23 asm_inline volatile("1: " LKGS_DI in native_lkgs()
|
| H A D | rmwcc.h | 12 asm_inline volatile (fullop \
|
| H A D | irq_stack.h | 87 asm_inline volatile( \
|
| H A D | percpu.h | 338 asm_inline qual ( \ 367 asm_inline qual ( \ 406 asm_inline qual ( \ 435 asm_inline qual ( \
|
| /linux/arch/riscv/include/asm/ |
| H A D | runtime-const.h | 18 asm_inline(".option push\n\t" \ 87 asm_inline(RISCV_RUNTIME_CONST_64_PREAMBLE \ 103 asm_inline(RISCV_RUNTIME_CONST_64_PREAMBLE \ 117 asm_inline(RISCV_RUNTIME_CONST_64_PREAMBLE \ 131 asm_inline(RISCV_RUNTIME_CONST_64_PREAMBLE \ 143 asm_inline(".option push\n\t" \
|
| /linux/samples/bpf/ |
| H A D | asm_goto_workaround.h | 22 #ifdef asm_inline 23 #undef asm_inline 24 #define asm_inline asm macro
|
| /linux/drivers/gpu/drm/vmwgfx/ |
| H A D | vmwgfx_msg_arm64.h | 62 asm_inline volatile ( in vmware_hypercall1() 87 asm_inline volatile ( in vmware_hypercall5() 113 asm_inline volatile ( in vmware_hypercall6() 143 asm_inline volatile ( in vmware_hypercall7() 173 asm_inline volatile ( in vmware_hypercall_hb()
|
| /linux/arch/s390/include/asm/ |
| H A D | alternative.h | 164 asm_inline volatile(ALTERNATIVE(oldinstr, altinstr, feature) : : : "memory") 167 asm_inline volatile(ALTERNATIVE_2(oldinstr, altinstr1, feature1, \ 172 asm_inline volatile (ALTERNATIVE(oldinstr, newinstr, feature) \ 177 asm_inline volatile(ALTERNATIVE(oldinstr, altinstr, feature) \
|
| H A D | runtime-const.h | 11 asm_inline( \ 27 asm_inline( \
|
| H A D | spinlock.h | 26 asm_inline( in spinlock_lockval() 99 asm_inline volatile( in arch_spin_unlock()
|
| /linux/drivers/s390/cio/ |
| H A D | ioasm.c | 25 asm_inline volatile( in __stsch() 55 asm_inline volatile( in __msch() 109 asm_inline volatile( in __ssch() 181 asm_inline volatile( in chsc()
|
| /linux/arch/s390/pci/ |
| H A D | pci_insn.c | 163 asm_inline volatile ( in ____pcilg() 232 asm_inline volatile ( in __pcilg_mio() 270 asm_inline volatile ( in __pcistg() 324 asm_inline volatile ( in __pcistg_mio() 359 asm_inline volatile ( in __pcistb() 413 asm_inline volatile ( in __pcistb_mio()
|
| /linux/arch/arm64/include/asm/ |
| H A D | runtime-const.h | 16 asm_inline("1:\t" \ 29 asm_inline("1:\t" \
|
| /linux/arch/s390/lib/ |
| H A D | spinlock.c | 97 asm_inline volatile( in arch_load_niai4() 110 asm_inline volatile( in arch_try_cmpxchg_niai8() 125 asm_inline volatile( in arch_try_cmpxchg_niai8()
|
| /linux/arch/s390/mm/ |
| H A D | pfault.c | 60 asm_inline volatile( in __pfault_init() 82 asm_inline volatile( in __pfault_fini()
|
| /linux/arch/s390/boot/ |
| H A D | physmem_info.c | 70 asm_inline volatile( in __diag260() 108 asm_inline volatile( in diag500_storage_limit() 129 asm_inline volatile( in tprot()
|
| /linux/include/linux/ |
| H A D | compiler_types.h | 603 #define asm_inline asm __inline macro 605 #define asm_inline asm macro
|