| /linux/arch/x86/include/asm/ |
| H A D | smap.h | 19 ALTERNATIVE "", "clac", X86_FEATURE_SMAP 22 ALTERNATIVE "", "stac", X86_FEATURE_SMAP 80 ALTERNATIVE(ANNOTATE_IGNORE_ALTERNATIVE "\n\t" in smap_save() 91 ALTERNATIVE(ANNOTATE_IGNORE_ALTERNATIVE "\n\t" in smap_restore() 99 ALTERNATIVE("", "clac", X86_FEATURE_SMAP) 101 ALTERNATIVE("", "stac", X86_FEATURE_SMAP) 104 ALTERNATIVE("", ANNOTATE_IGNORE_ALTERNATIVE "\n\t" "clac", X86_FEATURE_SMAP) 106 ALTERNATIVE("", ANNOTATE_IGNORE_ALTERNATIVE "\n\t" "stac", X86_FEATURE_SMAP)
|
| H A D | alternative.h | 217 #define ALTERNATIVE(oldinstr, newinstr, ft_flags) \ macro 223 ALTERNATIVE(ALTERNATIVE(oldinstr, newinstr1, ft_flags1), newinstr2, ft_flags2) 231 ALTERNATIVE(ALTERNATIVE_2(oldinstr, newinstr1, ft_flags1, newinstr2, ft_flags2), \ 247 asm_inline volatile(ALTERNATIVE(oldinstr, newinstr, ft_flags) : : : "memory") 261 asm_inline volatile(ALTERNATIVE(oldinstr, newinstr, ft_flags) \ 266 asm_inline volatile(ALTERNATIVE(oldinstr, newinstr, ft_flags) \ 280 asm_inline volatile(ALTERNATIVE("call %c[old]", "call %c[new]", ft_flags) \ 375 .macro ALTERNATIVE oldinstr, newinstr, ft_flags
|
| H A D | nospec-branch.h | 299 ALTERNATIVE "", \ 330 ALTERNATIVE "", __CLEAR_CPU_BUFFERS, X86_FEATURE_CLEAR_CPU_BUF 334 ALTERNATIVE "", "call clear_bhb_loop", X86_FEATURE_CLEAR_BHB_LOOP 338 ALTERNATIVE "", "call clear_bhb_loop", X86_FEATURE_CLEAR_BHB_VMEXIT 403 ALTERNATIVE("", \ 528 asm volatile(ALTERNATIVE("", "wrmsr", %c[feature]) in alternative_msr_write() 540 asm_inline volatile(ALTERNATIVE("", "call write_ibpb", X86_FEATURE_IBPB) in indirect_branch_prediction_barrier()
|
| H A D | cmpxchg_32.h | 94 ALTERNATIVE(_lock_loc \ 124 ALTERNATIVE(_lock_loc \
|
| /linux/arch/riscv/include/asm/ |
| H A D | errata_list.h | 18 ALTERNATIVE(__stringify(RISCV_PTR do_trap_insn_fault), \ 24 ALTERNATIVE(__stringify(RISCV_PTR do_page_fault), \ 31 asm(ALTERNATIVE("sfence.vma x0, %0", "sfence.vma", SIFIVE_VENDOR_ID, \ 36 asm(ALTERNATIVE("sfence.vma %0", "sfence.vma", SIFIVE_VENDOR_ID, \ 41 asm(ALTERNATIVE("sfence.vma %0, %1", "sfence.vma", SIFIVE_VENDOR_ID, \ 46 asm(ALTERNATIVE( \ 81 asm volatile(ALTERNATIVE( \ 102 asm volatile(ALTERNATIVE( \
|
| H A D | cpufeature-macros.h | 21 asm goto(ALTERNATIVE("j %l[l_no]", "nop", %[vendor], %[ext], 1) in __riscv_has_extension_likely() 35 asm goto(ALTERNATIVE("nop", "j %l[l_yes]", %[vendor], %[ext], 1) in __riscv_has_extension_unlikely()
|
| H A D | processor.h | 149 ALTERNATIVE(__nops(1), PREFETCH_R(x, 0), 0, \ 153 ALTERNATIVE(__nops(1), PREFETCH_W(x, 0), 0, \
|
| H A D | runtime-const.h | 108 ALTERNATIVE( \ 122 ALTERNATIVE( \
|
| /linux/arch/parisc/include/asm/ |
| H A D | cache.h | 53 ALTERNATIVE(ALT_COND_NO_SMP, INSN_PxTLB) \ 56 ALTERNATIVE(ALT_COND_NO_SMP, INSN_PxTLB) \ 57 ALTERNATIVE(ALT_COND_NO_SPLIT_TLB, INSN_NOP) \ 61 ALTERNATIVE(ALT_COND_NO_DCACHE, INSN_NOP) \ 62 ALTERNATIVE(ALT_COND_NO_IOC_FDC, INSN_NOP) \ 65 ALTERNATIVE(ALT_COND_NO_DCACHE, INSN_NOP) \ 66 ALTERNATIVE(ALT_COND_NO_IOC_FDC, INSN_NOP) :::"memory")
|
| H A D | alternative.h | 36 #define ALTERNATIVE(cond, replacement) "!0:" \ macro 47 #define ALTERNATIVE(from, to, cond, replacement)\ macro
|
| H A D | barrier.h | 12 ALTERNATIVE(ALT_COND_NO_SMP, INSN_NOP) \
|
| /linux/arch/x86/entry/ |
| H A D | calling.h | 168 ALTERNATIVE "", "SET_NOFLUSH_BIT \reg", X86_FEATURE_PCID 174 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 187 ALTERNATIVE "jmp .Lwrcr3_\@", "", X86_FEATURE_PCID 217 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 223 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 231 ALTERNATIVE "jmp .Ldone_\@", "", X86_FEATURE_PTI 250 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 260 ALTERNATIVE "jmp .Lwrcr3_\@", "", X86_FEATURE_PCID 307 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_KERNEL_IBRS 336 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_KERNEL_IBRS [all …]
|
| /linux/arch/s390/include/asm/ |
| H A D | alternative.h | 131 #define ALTERNATIVE(oldinstr, altinstr, feature) \ macro 164 asm_inline volatile(ALTERNATIVE(oldinstr, altinstr, feature) : : : "memory") 172 asm_inline volatile (ALTERNATIVE(oldinstr, newinstr, feature) \ 177 asm_inline volatile(ALTERNATIVE(oldinstr, altinstr, feature) \ 209 .macro ALTERNATIVE oldinstr, newinstr, feature
|
| H A D | spinlock.h | 27 ALTERNATIVE(" ly %[lockval],%[offzero](%%r0)\n", in spinlock_lockval() 100 ALTERNATIVE("nop", ".insn rre,0xb2fa0000,7,0", ALT_FACILITY(49)) /* NIAI 7 */ in arch_spin_unlock()
|
| /linux/arch/arm64/include/asm/ |
| H A D | arch_timer.h | 69 asm volatile(ALTERNATIVE("isb\n mrs %0, cntpct_el0", in arch_timer_read_cntpct_el0() 81 asm volatile(ALTERNATIVE("isb\n mrs %0, cntvct_el0", in arch_timer_read_cntvct_el0() 183 asm volatile(ALTERNATIVE("isb\n mrs %0, cntpct_el0", in __arch_counter_get_cntpct() 204 asm volatile(ALTERNATIVE("isb\n mrs %0, cntvct_el0", in __arch_counter_get_cntvct()
|
| H A D | io.h | 57 asm volatile(ALTERNATIVE("ldrb %w0, [%1]", in __raw_readb() 69 asm volatile(ALTERNATIVE("ldrh %w0, [%1]", in __raw_readw() 80 asm volatile(ALTERNATIVE("ldr %w0, [%1]", in __raw_readl() 91 asm volatile(ALTERNATIVE("ldr %0, [%1]", in __raw_readq()
|
| /linux/arch/s390/kernel/ |
| H A D | entry.S | 38 ALTERNATIVE "nop", ".insn s,0xb2010000,\address", ALT_FACILITY(193) 42 ALTERNATIVE "nop", ".insn s,0xb2000000,\address", ALT_FACILITY(193) 53 ALTERNATIVE "brcl 0,0", __stringify(mvc __PT_LAST_BREAK(8,\reg),__LC_LAST_BREAK(\lowcore)),\ 98 ALTERNATIVE "nop", ".insn rrf,0xb2e80000,0,0,12,0", ALT_SPEC(82) 102 ALTERNATIVE "nop", ".insn rrf,0xb2e80000,0,0,13,0", ALT_SPEC(82) 106 ALTERNATIVE "TSTMSK \tif_ptr,\tif_mask; jz .+8; .insn rrf,0xb2e80000,0,0,13,0", \ 112 ALTERNATIVE "jz .+8; .insn rrf,0xb2e80000,0,0,12,0", \ 168 ALTERNATIVE "nop", "lpp _LPP_OFFSET(%r13)", ALT_FACILITY(40)
|
| /linux/tools/include/asm/ |
| H A D | alternative.h | 7 .macro ALTERNATIVE oldinstr, newinstr, feature 15 #define ALTERNATIVE # macro
|
| /linux/arch/parisc/kernel/ |
| H A D | pacache.S | 106 ALTERNATIVE(88b, fitdone, ALT_COND_NO_SPLIT_TLB, INSN_NOP) 244 89: ALTERNATIVE(88b, 89b, ALT_COND_NO_ICACHE, INSN_NOP) 305 89: ALTERNATIVE(88b, 89b, ALT_COND_NO_DCACHE, INSN_NOP) 548 ALTERNATIVE(0b, 0b+4, ALT_COND_NO_SMP, INSN_PxTLB) 549 ALTERNATIVE(1b, 1b+4, ALT_COND_NO_SMP, INSN_PxTLB) 677 ALTERNATIVE(0b, 0b+4, ALT_COND_NO_SMP, INSN_PxTLB) 743 ALTERNATIVE(0b, 0b+4, ALT_COND_NO_SMP, INSN_PxTLB) 775 89: ALTERNATIVE(88b, 89b, ALT_COND_NO_DCACHE, INSN_NOP) 792 ALTERNATIVE(0b, 0b+4, ALT_COND_NO_SMP, INSN_PxTLB) 824 89: ALTERNATIVE(88b, 89b, ALT_COND_NO_DCACHE, INSN_NOP) [all …]
|
| /linux/arch/loongarch/include/asm/ |
| H A D | alternative.h | 71 #define ALTERNATIVE(oldinstr, newinstr, feature) \ macro 104 (asm volatile (ALTERNATIVE(oldinstr, newinstr, feature) : : : "memory"))
|
| H A D | alternative-asm.h | 29 .macro ALTERNATIVE oldinstr, newinstr, feature
|
| /linux/arch/s390/lib/ |
| H A D | spinlock.c | 98 ALTERNATIVE("nop", ".insn rre,0xb2fa0000,4,0", ALT_FACILITY(49)) /* NIAI 4 */ in arch_load_niai4() 111 ALTERNATIVE("nop", ".insn rre,0xb2fa0000,8,0", ALT_FACILITY(49)) /* NIAI 8 */ in arch_try_cmpxchg_niai8() 126 ALTERNATIVE("nop", ".insn rre,0xb2fa0000,8,0", ALT_FACILITY(49)) /* NIAI 8 */ in arch_try_cmpxchg_niai8()
|
| /linux/arch/x86/lib/ |
| H A D | retpoline.S | 369 ALTERNATIVE JMP_RETBLEED_UNTRAIN_RET, JMP_SRSO_UNTRAIN_RET, X86_FEATURE_SRSO 443 ALTERNATIVE __stringify(ANNOTATE_UNRET_SAFE; ret), \
|
| /linux/arch/riscv/lib/ |
| H A D | uaccess.S | 17 ALTERNATIVE("j fallback_scalar_usercopy", "nop", 0, RISCV_ISA_EXT_ZVE32X, CONFIG_RISCV_ISA_V) 45 …ALTERNATIVE("j fallback_scalar_usercopy_sum_enabled", "nop", 0, RISCV_ISA_EXT_ZVE32X, CONFIG_RISCV…
|
| /linux/arch/riscv/kernel/ |
| H A D | entry.S | 79 ALTERNATIVE("sfence.vma", "nop", 0, RISCV_ISA_EXT_SVVPTC, 1) 100 ALTERNATIVE("nops(4)", 113 ALTERNATIVE("nops(4)",
|