1 /*- 2 * SPDX-License-Identifier: BSD-2-Clause 3 * 4 * Copyright (c) 2011 NetApp, Inc. 5 * All rights reserved. 6 * 7 * Redistribution and use in source and binary forms, with or without 8 * modification, are permitted provided that the following conditions 9 * are met: 10 * 1. Redistributions of source code must retain the above copyright 11 * notice, this list of conditions and the following disclaimer. 12 * 2. Redistributions in binary form must reproduce the above copyright 13 * notice, this list of conditions and the following disclaimer in the 14 * documentation and/or other materials provided with the distribution. 15 * 16 * THIS SOFTWARE IS PROVIDED BY NETAPP, INC ``AS IS'' AND 17 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 18 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 19 * ARE DISCLAIMED. IN NO EVENT SHALL NETAPP, INC OR CONTRIBUTORS BE LIABLE 20 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 21 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 22 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 23 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 24 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 25 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 26 * SUCH DAMAGE. 27 */ 28 29 #ifndef _VMX_CPUFUNC_H_ 30 #define _VMX_CPUFUNC_H_ 31 32 struct vmcs; 33 34 /* 35 * Section 5.2 "Conventions" from Intel Architecture Manual 2B. 36 * 37 * error 38 * VMsucceed 0 39 * VMFailInvalid 1 40 * VMFailValid 2 see also VMCS VM-Instruction Error Field 41 */ 42 #define VM_SUCCESS 0 43 #define VM_FAIL_INVALID 1 44 #define VM_FAIL_VALID 2 45 #define VMX_SET_ERROR_CODE \ 46 " jnc 1f;" \ 47 " mov $1, %[error];" /* CF: error = 1 */ \ 48 " jmp 3f;" \ 49 "1: jnz 2f;" \ 50 " mov $2, %[error];" /* ZF: error = 2 */ \ 51 " jmp 3f;" \ 52 "2: mov $0, %[error];" \ 53 "3:" 54 55 /* returns 0 on success and non-zero on failure */ 56 static __inline int 57 vmxon(char *region) 58 { 59 int error; 60 uint64_t addr; 61 62 addr = vtophys(region); 63 __asm __volatile("vmxon %[addr];" 64 VMX_SET_ERROR_CODE 65 : [error] "=r" (error) 66 : [addr] "m" (*(uint64_t *)&addr) 67 : "memory"); 68 69 return (error); 70 } 71 72 /* returns 0 on success and non-zero on failure */ 73 static __inline int 74 vmclear(struct vmcs *vmcs) 75 { 76 int error; 77 uint64_t addr; 78 79 addr = vtophys(vmcs); 80 __asm __volatile("vmclear %[addr];" 81 VMX_SET_ERROR_CODE 82 : [error] "=r" (error) 83 : [addr] "m" (*(uint64_t *)&addr) 84 : "memory"); 85 return (error); 86 } 87 88 static __inline void 89 vmxoff(void) 90 { 91 92 __asm __volatile("vmxoff"); 93 } 94 95 static __inline void 96 vmptrst(uint64_t *addr) 97 { 98 99 __asm __volatile("vmptrst %[addr]" :: [addr]"m" (*addr) : "memory"); 100 } 101 102 static __inline int 103 vmptrld(struct vmcs *vmcs) 104 { 105 int error; 106 uint64_t addr; 107 108 addr = vtophys(vmcs); 109 __asm __volatile("vmptrld %[addr];" 110 VMX_SET_ERROR_CODE 111 : [error] "=r" (error) 112 : [addr] "m" (*(uint64_t *)&addr) 113 : "memory"); 114 return (error); 115 } 116 117 static __inline int 118 vmwrite(uint64_t reg, uint64_t val) 119 { 120 int error; 121 122 __asm __volatile("vmwrite %[val], %[reg];" 123 VMX_SET_ERROR_CODE 124 : [error] "=r" (error) 125 : [val] "r" (val), [reg] "r" (reg) 126 : "memory"); 127 128 return (error); 129 } 130 131 static __inline int 132 vmread(uint64_t r, uint64_t *addr) 133 { 134 int error; 135 136 __asm __volatile("vmread %[r], %[addr];" 137 VMX_SET_ERROR_CODE 138 : [error] "=r" (error), [addr] "=m" (*addr) 139 : [r] "r" (r) 140 : "memory"); 141 return (error); 142 } 143 144 static void __inline 145 VMCLEAR(struct vmcs *vmcs) 146 { 147 int err; 148 149 err = vmclear(vmcs); 150 if (err != 0) 151 panic("%s: vmclear(%p) error %d", __func__, vmcs, err); 152 153 critical_exit(); 154 } 155 156 static void __inline 157 VMPTRLD(struct vmcs *vmcs) 158 { 159 int err; 160 161 critical_enter(); 162 163 err = vmptrld(vmcs); 164 if (err != 0) 165 panic("%s: vmptrld(%p) error %d", __func__, vmcs, err); 166 } 167 168 #define INVVPID_TYPE_ADDRESS 0UL 169 #define INVVPID_TYPE_SINGLE_CONTEXT 1UL 170 #define INVVPID_TYPE_ALL_CONTEXTS 2UL 171 172 struct invvpid_desc { 173 uint16_t vpid; 174 uint16_t _res1; 175 uint32_t _res2; 176 uint64_t linear_addr; 177 }; 178 CTASSERT(sizeof(struct invvpid_desc) == 16); 179 180 static void __inline 181 invvpid(uint64_t type, struct invvpid_desc desc) 182 { 183 int error; 184 185 __asm __volatile("invvpid %[desc], %[type];" 186 VMX_SET_ERROR_CODE 187 : [error] "=r" (error) 188 : [desc] "m" (desc), [type] "r" (type) 189 : "memory"); 190 191 if (error) 192 panic("invvpid error %d", error); 193 } 194 195 #define INVEPT_TYPE_SINGLE_CONTEXT 1UL 196 #define INVEPT_TYPE_ALL_CONTEXTS 2UL 197 struct invept_desc { 198 uint64_t eptp; 199 uint64_t _res; 200 }; 201 CTASSERT(sizeof(struct invept_desc) == 16); 202 203 static void __inline 204 invept(uint64_t type, struct invept_desc desc) 205 { 206 int error; 207 208 __asm __volatile("invept %[desc], %[type];" 209 VMX_SET_ERROR_CODE 210 : [error] "=r" (error) 211 : [desc] "m" (desc), [type] "r" (type) 212 : "memory"); 213 214 if (error) 215 panic("invept error %d", error); 216 } 217 #endif 218