1 /* SPDX-License-Identifier: GPL-2.0 */ 2 #ifndef __KVM_X86_SVM_OPS_H 3 #define __KVM_X86_SVM_OPS_H 4 5 #include <linux/compiler_types.h> 6 7 #include <asm/kvm_host.h> 8 9 #define svm_asm(insn, clobber...) \ 10 do { \ 11 asm_volatile_goto("1: " __stringify(insn) "\n\t" \ 12 _ASM_EXTABLE(1b, %l[fault]) \ 13 ::: clobber : fault); \ 14 return; \ 15 fault: \ 16 kvm_spurious_fault(); \ 17 } while (0) 18 19 #define svm_asm1(insn, op1, clobber...) \ 20 do { \ 21 asm_volatile_goto("1: " __stringify(insn) " %0\n\t" \ 22 _ASM_EXTABLE(1b, %l[fault]) \ 23 :: op1 : clobber : fault); \ 24 return; \ 25 fault: \ 26 kvm_spurious_fault(); \ 27 } while (0) 28 29 #define svm_asm2(insn, op1, op2, clobber...) \ 30 do { \ 31 asm_volatile_goto("1: " __stringify(insn) " %1, %0\n\t" \ 32 _ASM_EXTABLE(1b, %l[fault]) \ 33 :: op1, op2 : clobber : fault); \ 34 return; \ 35 fault: \ 36 kvm_spurious_fault(); \ 37 } while (0) 38 39 static inline void clgi(void) 40 { 41 svm_asm(clgi); 42 } 43 44 static inline void stgi(void) 45 { 46 svm_asm(stgi); 47 } 48 49 static inline void invlpga(unsigned long addr, u32 asid) 50 { 51 svm_asm2(invlpga, "c"(asid), "a"(addr)); 52 } 53 54 /* 55 * Despite being a physical address, the portion of rAX that is consumed by 56 * VMSAVE, VMLOAD, etc... is still controlled by the effective address size, 57 * hence 'unsigned long' instead of 'hpa_t'. 58 */ 59 static inline void vmsave(unsigned long pa) 60 { 61 svm_asm1(vmsave, "a" (pa), "memory"); 62 } 63 64 static inline void vmload(unsigned long pa) 65 { 66 svm_asm1(vmload, "a" (pa), "memory"); 67 } 68 69 #endif /* __KVM_X86_SVM_OPS_H */ 70