1366f6083SPeter Grehan /*- 2c49761ddSPedro F. Giffuni * SPDX-License-Identifier: BSD-2-Clause-FreeBSD 3c49761ddSPedro F. Giffuni * 4366f6083SPeter Grehan * Copyright (c) 2011 NetApp, Inc. 5366f6083SPeter Grehan * All rights reserved. 6366f6083SPeter Grehan * 7366f6083SPeter Grehan * Redistribution and use in source and binary forms, with or without 8366f6083SPeter Grehan * modification, are permitted provided that the following conditions 9366f6083SPeter Grehan * are met: 10366f6083SPeter Grehan * 1. Redistributions of source code must retain the above copyright 11366f6083SPeter Grehan * notice, this list of conditions and the following disclaimer. 12366f6083SPeter Grehan * 2. Redistributions in binary form must reproduce the above copyright 13366f6083SPeter Grehan * notice, this list of conditions and the following disclaimer in the 14366f6083SPeter Grehan * documentation and/or other materials provided with the distribution. 15366f6083SPeter Grehan * 16366f6083SPeter Grehan * THIS SOFTWARE IS PROVIDED BY NETAPP, INC ``AS IS'' AND 17366f6083SPeter Grehan * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 18366f6083SPeter Grehan * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 19366f6083SPeter Grehan * ARE DISCLAIMED. IN NO EVENT SHALL NETAPP, INC OR CONTRIBUTORS BE LIABLE 20366f6083SPeter Grehan * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 21366f6083SPeter Grehan * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 22366f6083SPeter Grehan * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 23366f6083SPeter Grehan * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 24366f6083SPeter Grehan * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 25366f6083SPeter Grehan * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 26366f6083SPeter Grehan * SUCH DAMAGE. 27366f6083SPeter Grehan * 28366f6083SPeter Grehan * $FreeBSD$ 29366f6083SPeter Grehan */ 30366f6083SPeter Grehan 31366f6083SPeter Grehan #include <sys/cdefs.h> 32366f6083SPeter Grehan __FBSDID("$FreeBSD$"); 33366f6083SPeter Grehan 34483d953aSJohn Baldwin #include "opt_bhyve_snapshot.h" 35483d953aSJohn Baldwin 36366f6083SPeter Grehan #include <sys/param.h> 3738f1b189SPeter Grehan #include <sys/systm.h> 38366f6083SPeter Grehan #include <sys/kernel.h> 39366f6083SPeter Grehan #include <sys/module.h> 40366f6083SPeter Grehan #include <sys/sysctl.h> 41366f6083SPeter Grehan #include <sys/malloc.h> 42366f6083SPeter Grehan #include <sys/pcpu.h> 43366f6083SPeter Grehan #include <sys/lock.h> 44366f6083SPeter Grehan #include <sys/mutex.h> 45366f6083SPeter Grehan #include <sys/proc.h> 46318224bbSNeel Natu #include <sys/rwlock.h> 47366f6083SPeter Grehan #include <sys/sched.h> 48366f6083SPeter Grehan #include <sys/smp.h> 4967b69e76SJohn Baldwin #include <sys/sx.h> 50483d953aSJohn Baldwin #include <sys/vnode.h> 51366f6083SPeter Grehan 52366f6083SPeter Grehan #include <vm/vm.h> 533c48106aSKonstantin Belousov #include <vm/vm_param.h> 543c48106aSKonstantin Belousov #include <vm/vm_extern.h> 55318224bbSNeel Natu #include <vm/vm_object.h> 56318224bbSNeel Natu #include <vm/vm_page.h> 57318224bbSNeel Natu #include <vm/pmap.h> 58318224bbSNeel Natu #include <vm/vm_map.h> 59483d953aSJohn Baldwin #include <vm/vm_pager.h> 60483d953aSJohn Baldwin #include <vm/vm_kern.h> 61483d953aSJohn Baldwin #include <vm/vnode_pager.h> 62483d953aSJohn Baldwin #include <vm/swap_pager.h> 63483d953aSJohn Baldwin #include <vm/uma.h> 64366f6083SPeter Grehan 6563e62d39SJohn Baldwin #include <machine/cpu.h> 66366f6083SPeter Grehan #include <machine/pcb.h> 6775dd3366SNeel Natu #include <machine/smp.h> 68bd50262fSKonstantin Belousov #include <machine/md_var.h> 691c052192SNeel Natu #include <x86/psl.h> 7034a6b2d6SJohn Baldwin #include <x86/apicreg.h> 7115add60dSPeter Grehan #include <x86/ifunc.h> 72366f6083SPeter Grehan 73366f6083SPeter Grehan #include <machine/vmm.h> 74565bbb86SNeel Natu #include <machine/vmm_dev.h> 75e813a873SNeel Natu #include <machine/vmm_instruction_emul.h> 76483d953aSJohn Baldwin #include <machine/vmm_snapshot.h> 77565bbb86SNeel Natu 78d17b5104SNeel Natu #include "vmm_ioport.h" 79318224bbSNeel Natu #include "vmm_ktr.h" 80b01c2033SNeel Natu #include "vmm_host.h" 81366f6083SPeter Grehan #include "vmm_mem.h" 82366f6083SPeter Grehan #include "vmm_util.h" 83762fd208STycho Nightingale #include "vatpic.h" 84e883c9bbSTycho Nightingale #include "vatpit.h" 8508e3ff32SNeel Natu #include "vhpet.h" 86565bbb86SNeel Natu #include "vioapic.h" 87366f6083SPeter Grehan #include "vlapic.h" 88160ef77aSNeel Natu #include "vpmtmr.h" 890dafa5cdSNeel Natu #include "vrtc.h" 90366f6083SPeter Grehan #include "vmm_stat.h" 91f76fc5d4SNeel Natu #include "vmm_lapic.h" 92366f6083SPeter Grehan 93366f6083SPeter Grehan #include "io/ppt.h" 94366f6083SPeter Grehan #include "io/iommu.h" 95366f6083SPeter Grehan 96366f6083SPeter Grehan struct vlapic; 97366f6083SPeter Grehan 985fcf252fSNeel Natu /* 995fcf252fSNeel Natu * Initialization: 1005fcf252fSNeel Natu * (a) allocated when vcpu is created 1015fcf252fSNeel Natu * (i) initialized when vcpu is created and when it is reinitialized 1025fcf252fSNeel Natu * (o) initialized the first time the vcpu is created 1035fcf252fSNeel Natu * (x) initialized before use 1045fcf252fSNeel Natu */ 105366f6083SPeter Grehan struct vcpu { 1065fcf252fSNeel Natu struct mtx mtx; /* (o) protects 'state' and 'hostcpu' */ 1075fcf252fSNeel Natu enum vcpu_state state; /* (o) vcpu state */ 108950af9ffSJohn Baldwin int vcpuid; /* (o) */ 1095fcf252fSNeel Natu int hostcpu; /* (o) vcpu's host cpu */ 110248e6799SNeel Natu int reqidle; /* (i) request vcpu to idle */ 111950af9ffSJohn Baldwin struct vm *vm; /* (o) */ 1121aa51504SJohn Baldwin void *cookie; /* (i) cpu-specific data */ 1135fcf252fSNeel Natu struct vlapic *vlapic; /* (i) APIC device model */ 1145fcf252fSNeel Natu enum x2apic_state x2apic_state; /* (i) APIC mode */ 115091d4532SNeel Natu uint64_t exitintinfo; /* (i) events pending at VM exit */ 1165fcf252fSNeel Natu int nmi_pending; /* (i) NMI pending */ 1175fcf252fSNeel Natu int extint_pending; /* (i) INTR pending */ 1185fcf252fSNeel Natu int exception_pending; /* (i) exception pending */ 119c9c75df4SNeel Natu int exc_vector; /* (x) exception collateral */ 120c9c75df4SNeel Natu int exc_errcode_valid; 121c9c75df4SNeel Natu uint32_t exc_errcode; 1225fcf252fSNeel Natu struct savefpu *guestfpu; /* (a,i) guest fpu state */ 1235fcf252fSNeel Natu uint64_t guest_xcr0; /* (i) guest %xcr0 register */ 1245fcf252fSNeel Natu void *stats; /* (a,i) statistics */ 1255fcf252fSNeel Natu struct vm_exit exitinfo; /* (x) exit reason and collateral */ 126d087a399SNeel Natu uint64_t nextrip; /* (x) next instruction to execute */ 127483d953aSJohn Baldwin uint64_t tsc_offset; /* (o) TSC offsetting */ 128366f6083SPeter Grehan }; 129366f6083SPeter Grehan 130f76fc5d4SNeel Natu #define vcpu_lock_init(v) mtx_init(&((v)->mtx), "vcpu lock", 0, MTX_SPIN) 13108ebb360SJohn Baldwin #define vcpu_lock_destroy(v) mtx_destroy(&((v)->mtx)) 132f76fc5d4SNeel Natu #define vcpu_lock(v) mtx_lock_spin(&((v)->mtx)) 133f76fc5d4SNeel Natu #define vcpu_unlock(v) mtx_unlock_spin(&((v)->mtx)) 134318224bbSNeel Natu #define vcpu_assert_locked(v) mtx_assert(&((v)->mtx), MA_OWNED) 13575dd3366SNeel Natu 136318224bbSNeel Natu struct mem_seg { 1379b1aa8d6SNeel Natu size_t len; 1389b1aa8d6SNeel Natu bool sysmem; 1399b1aa8d6SNeel Natu struct vm_object *object; 1409b1aa8d6SNeel Natu }; 141e47fe318SCorvin Köhne #define VM_MAX_MEMSEGS 4 1429b1aa8d6SNeel Natu 1439b1aa8d6SNeel Natu struct mem_map { 144318224bbSNeel Natu vm_paddr_t gpa; 145318224bbSNeel Natu size_t len; 1469b1aa8d6SNeel Natu vm_ooffset_t segoff; 1479b1aa8d6SNeel Natu int segid; 1489b1aa8d6SNeel Natu int prot; 1499b1aa8d6SNeel Natu int flags; 150318224bbSNeel Natu }; 15100d3723fSConrad Meyer #define VM_MAX_MEMMAPS 8 152366f6083SPeter Grehan 153366f6083SPeter Grehan /* 1545fcf252fSNeel Natu * Initialization: 1555fcf252fSNeel Natu * (o) initialized the first time the VM is created 1565fcf252fSNeel Natu * (i) initialized when VM is created and when it is reinitialized 1575fcf252fSNeel Natu * (x) initialized before use 15867b69e76SJohn Baldwin * 15967b69e76SJohn Baldwin * Locking: 16067b69e76SJohn Baldwin * [m] mem_segs_lock 16167b69e76SJohn Baldwin * [r] rendezvous_mtx 16267b69e76SJohn Baldwin * [v] reads require one frozen vcpu, writes require freezing all vcpus 163366f6083SPeter Grehan */ 1645fcf252fSNeel Natu struct vm { 1655fcf252fSNeel Natu void *cookie; /* (i) cpu-specific data */ 1665fcf252fSNeel Natu void *iommu; /* (x) iommu-specific data */ 1675fcf252fSNeel Natu struct vhpet *vhpet; /* (i) virtual HPET */ 1685fcf252fSNeel Natu struct vioapic *vioapic; /* (i) virtual ioapic */ 1695fcf252fSNeel Natu struct vatpic *vatpic; /* (i) virtual atpic */ 1705fcf252fSNeel Natu struct vatpit *vatpit; /* (i) virtual atpit */ 171160ef77aSNeel Natu struct vpmtmr *vpmtmr; /* (i) virtual ACPI PM timer */ 1720dafa5cdSNeel Natu struct vrtc *vrtc; /* (o) virtual RTC */ 1735fcf252fSNeel Natu volatile cpuset_t active_cpus; /* (i) active vcpus */ 174fc276d92SJohn Baldwin volatile cpuset_t debug_cpus; /* (i) vcpus stopped for debug */ 175c0f35dbfSJohn Baldwin cpuset_t startup_cpus; /* (i) [r] waiting for startup */ 1765fcf252fSNeel Natu int suspend; /* (i) stop VM execution */ 177*98568a00SJohn Baldwin bool dying; /* (o) is dying */ 1785fcf252fSNeel Natu volatile cpuset_t suspended_cpus; /* (i) suspended vcpus */ 1795fcf252fSNeel Natu volatile cpuset_t halted_cpus; /* (x) cpus in a hard halt */ 18067b69e76SJohn Baldwin cpuset_t rendezvous_req_cpus; /* (x) [r] rendezvous requested */ 18167b69e76SJohn Baldwin cpuset_t rendezvous_done_cpus; /* (x) [r] rendezvous finished */ 18267b69e76SJohn Baldwin void *rendezvous_arg; /* (x) [r] rendezvous func/arg */ 1835b8a8cd1SNeel Natu vm_rendezvous_func_t rendezvous_func; 1845fcf252fSNeel Natu struct mtx rendezvous_mtx; /* (o) rendezvous lock */ 18567b69e76SJohn Baldwin struct mem_map mem_maps[VM_MAX_MEMMAPS]; /* (i) [m+v] guest address space */ 18667b69e76SJohn Baldwin struct mem_seg mem_segs[VM_MAX_MEMSEGS]; /* (o) [m+v] guest memory regions */ 1875fcf252fSNeel Natu struct vmspace *vmspace; /* (o) guest's address space */ 188df95cc76SKa Ho Ng char name[VM_MAX_NAMELEN+1]; /* (o) virtual machine name */ 189*98568a00SJohn Baldwin struct vcpu *vcpu[VM_MAXCPU]; /* (x) guest vcpus */ 19001d822d3SRodney W. Grimes /* The following describe the vm cpu topology */ 19101d822d3SRodney W. Grimes uint16_t sockets; /* (o) num of sockets */ 19201d822d3SRodney W. Grimes uint16_t cores; /* (o) num of cores/socket */ 19301d822d3SRodney W. Grimes uint16_t threads; /* (o) num of threads/core */ 19401d822d3SRodney W. Grimes uint16_t maxcpus; /* (o) max pluggable cpus */ 19567b69e76SJohn Baldwin struct sx mem_segs_lock; /* (o) */ 196*98568a00SJohn Baldwin struct sx vcpus_init_lock; /* (o) */ 197366f6083SPeter Grehan }; 198366f6083SPeter Grehan 199950af9ffSJohn Baldwin #define VMM_CTR0(vcpu, format) \ 200950af9ffSJohn Baldwin VCPU_CTR0((vcpu)->vm, (vcpu)->vcpuid, format) 201950af9ffSJohn Baldwin 202950af9ffSJohn Baldwin #define VMM_CTR1(vcpu, format, p1) \ 203950af9ffSJohn Baldwin VCPU_CTR1((vcpu)->vm, (vcpu)->vcpuid, format, p1) 204950af9ffSJohn Baldwin 205950af9ffSJohn Baldwin #define VMM_CTR2(vcpu, format, p1, p2) \ 206950af9ffSJohn Baldwin VCPU_CTR2((vcpu)->vm, (vcpu)->vcpuid, format, p1, p2) 207950af9ffSJohn Baldwin 208950af9ffSJohn Baldwin #define VMM_CTR3(vcpu, format, p1, p2, p3) \ 209950af9ffSJohn Baldwin VCPU_CTR3((vcpu)->vm, (vcpu)->vcpuid, format, p1, p2, p3) 210950af9ffSJohn Baldwin 211950af9ffSJohn Baldwin #define VMM_CTR4(vcpu, format, p1, p2, p3, p4) \ 212950af9ffSJohn Baldwin VCPU_CTR4((vcpu)->vm, (vcpu)->vcpuid, format, p1, p2, p3, p4) 213950af9ffSJohn Baldwin 214d5408b1dSNeel Natu static int vmm_initialized; 215d5408b1dSNeel Natu 21615add60dSPeter Grehan static void vmmops_panic(void); 217366f6083SPeter Grehan 21815add60dSPeter Grehan static void 21915add60dSPeter Grehan vmmops_panic(void) 22015add60dSPeter Grehan { 22115add60dSPeter Grehan panic("vmm_ops func called when !vmm_is_intel() && !vmm_is_svm()"); 22215add60dSPeter Grehan } 22315add60dSPeter Grehan 22415add60dSPeter Grehan #define DEFINE_VMMOPS_IFUNC(ret_type, opname, args) \ 22515add60dSPeter Grehan DEFINE_IFUNC(static, ret_type, vmmops_##opname, args) \ 22615add60dSPeter Grehan { \ 22715add60dSPeter Grehan if (vmm_is_intel()) \ 22815add60dSPeter Grehan return (vmm_ops_intel.opname); \ 22915add60dSPeter Grehan else if (vmm_is_svm()) \ 23015add60dSPeter Grehan return (vmm_ops_amd.opname); \ 23115add60dSPeter Grehan else \ 23215add60dSPeter Grehan return ((ret_type (*)args)vmmops_panic); \ 23315add60dSPeter Grehan } 23415add60dSPeter Grehan 23515add60dSPeter Grehan DEFINE_VMMOPS_IFUNC(int, modinit, (int ipinum)) 23615add60dSPeter Grehan DEFINE_VMMOPS_IFUNC(int, modcleanup, (void)) 23715add60dSPeter Grehan DEFINE_VMMOPS_IFUNC(void, modresume, (void)) 23815add60dSPeter Grehan DEFINE_VMMOPS_IFUNC(void *, init, (struct vm *vm, struct pmap *pmap)) 239869c8d19SJohn Baldwin DEFINE_VMMOPS_IFUNC(int, run, (void *vcpui, register_t rip, struct pmap *pmap, 240869c8d19SJohn Baldwin struct vm_eventinfo *info)) 24115add60dSPeter Grehan DEFINE_VMMOPS_IFUNC(void, cleanup, (void *vmi)) 242950af9ffSJohn Baldwin DEFINE_VMMOPS_IFUNC(void *, vcpu_init, (void *vmi, struct vcpu *vcpu, 243950af9ffSJohn Baldwin int vcpu_id)) 244869c8d19SJohn Baldwin DEFINE_VMMOPS_IFUNC(void, vcpu_cleanup, (void *vcpui)) 245869c8d19SJohn Baldwin DEFINE_VMMOPS_IFUNC(int, getreg, (void *vcpui, int num, uint64_t *retval)) 246869c8d19SJohn Baldwin DEFINE_VMMOPS_IFUNC(int, setreg, (void *vcpui, int num, uint64_t val)) 247869c8d19SJohn Baldwin DEFINE_VMMOPS_IFUNC(int, getdesc, (void *vcpui, int num, struct seg_desc *desc)) 248869c8d19SJohn Baldwin DEFINE_VMMOPS_IFUNC(int, setdesc, (void *vcpui, int num, struct seg_desc *desc)) 249869c8d19SJohn Baldwin DEFINE_VMMOPS_IFUNC(int, getcap, (void *vcpui, int num, int *retval)) 250869c8d19SJohn Baldwin DEFINE_VMMOPS_IFUNC(int, setcap, (void *vcpui, int num, int val)) 25115add60dSPeter Grehan DEFINE_VMMOPS_IFUNC(struct vmspace *, vmspace_alloc, (vm_offset_t min, 25215add60dSPeter Grehan vm_offset_t max)) 25315add60dSPeter Grehan DEFINE_VMMOPS_IFUNC(void, vmspace_free, (struct vmspace *vmspace)) 254869c8d19SJohn Baldwin DEFINE_VMMOPS_IFUNC(struct vlapic *, vlapic_init, (void *vcpui)) 255869c8d19SJohn Baldwin DEFINE_VMMOPS_IFUNC(void, vlapic_cleanup, (struct vlapic *vlapic)) 256483d953aSJohn Baldwin #ifdef BHYVE_SNAPSHOT 257869c8d19SJohn Baldwin DEFINE_VMMOPS_IFUNC(int, snapshot, (void *vmi, struct vm_snapshot_meta *meta)) 258869c8d19SJohn Baldwin DEFINE_VMMOPS_IFUNC(int, vcpu_snapshot, (void *vcpui, 259869c8d19SJohn Baldwin struct vm_snapshot_meta *meta)) 260869c8d19SJohn Baldwin DEFINE_VMMOPS_IFUNC(int, restore_tsc, (void *vcpui, uint64_t now)) 261483d953aSJohn Baldwin #endif 262366f6083SPeter Grehan 263014a52f3SNeel Natu #define fpu_start_emulating() load_cr0(rcr0() | CR0_TS) 264014a52f3SNeel Natu #define fpu_stop_emulating() clts() 265366f6083SPeter Grehan 2666ac73777STycho Nightingale SDT_PROVIDER_DEFINE(vmm); 2676ac73777STycho Nightingale 268366f6083SPeter Grehan static MALLOC_DEFINE(M_VM, "vm", "vm"); 269366f6083SPeter Grehan 270366f6083SPeter Grehan /* statistics */ 27161592433SNeel Natu static VMM_STAT(VCPU_TOTAL_RUNTIME, "vcpu total runtime"); 272366f6083SPeter Grehan 273b40598c5SPawel Biernacki SYSCTL_NODE(_hw, OID_AUTO, vmm, CTLFLAG_RW | CTLFLAG_MPSAFE, NULL, 274b40598c5SPawel Biernacki NULL); 275add611fdSNeel Natu 276055fc2cbSNeel Natu /* 277055fc2cbSNeel Natu * Halt the guest if all vcpus are executing a HLT instruction with 278055fc2cbSNeel Natu * interrupts disabled. 279055fc2cbSNeel Natu */ 280055fc2cbSNeel Natu static int halt_detection_enabled = 1; 281055fc2cbSNeel Natu SYSCTL_INT(_hw_vmm, OID_AUTO, halt_detection, CTLFLAG_RDTUN, 282055fc2cbSNeel Natu &halt_detection_enabled, 0, 283055fc2cbSNeel Natu "Halt VM if all vcpus execute HLT with interrupts disabled"); 284055fc2cbSNeel Natu 285978f3da1SAndriy Gapon static int vmm_ipinum; 286add611fdSNeel Natu SYSCTL_INT(_hw_vmm, OID_AUTO, ipinum, CTLFLAG_RD, &vmm_ipinum, 0, 287add611fdSNeel Natu "IPI vector used for vcpu notifications"); 288add611fdSNeel Natu 289b0538143SNeel Natu static int trace_guest_exceptions; 290b0538143SNeel Natu SYSCTL_INT(_hw_vmm, OID_AUTO, trace_guest_exceptions, CTLFLAG_RDTUN, 291b0538143SNeel Natu &trace_guest_exceptions, 0, 292b0538143SNeel Natu "Trap into hypervisor on all guest exceptions and reflect them back"); 293b0538143SNeel Natu 2943ba952e1SCorvin Köhne static int trap_wbinvd; 2953ba952e1SCorvin Köhne SYSCTL_INT(_hw_vmm, OID_AUTO, trap_wbinvd, CTLFLAG_RDTUN, &trap_wbinvd, 0, 2963ba952e1SCorvin Köhne "WBINVD triggers a VM-exit"); 2973ba952e1SCorvin Köhne 2989b1aa8d6SNeel Natu static void vm_free_memmap(struct vm *vm, int ident); 2999b1aa8d6SNeel Natu static bool sysmem_mapping(struct vm *vm, struct mem_map *mm); 300248e6799SNeel Natu static void vcpu_notify_event_locked(struct vcpu *vcpu, bool lapic_intr); 301248e6799SNeel Natu 302248e6799SNeel Natu #ifdef KTR 303248e6799SNeel Natu static const char * 304248e6799SNeel Natu vcpu_state2str(enum vcpu_state state) 305248e6799SNeel Natu { 306248e6799SNeel Natu 307248e6799SNeel Natu switch (state) { 308248e6799SNeel Natu case VCPU_IDLE: 309248e6799SNeel Natu return ("idle"); 310248e6799SNeel Natu case VCPU_FROZEN: 311248e6799SNeel Natu return ("frozen"); 312248e6799SNeel Natu case VCPU_RUNNING: 313248e6799SNeel Natu return ("running"); 314248e6799SNeel Natu case VCPU_SLEEPING: 315248e6799SNeel Natu return ("sleeping"); 316248e6799SNeel Natu default: 317248e6799SNeel Natu return ("unknown"); 318248e6799SNeel Natu } 319248e6799SNeel Natu } 320248e6799SNeel Natu #endif 321248e6799SNeel Natu 322366f6083SPeter Grehan static void 323*98568a00SJohn Baldwin vcpu_cleanup(struct vcpu *vcpu, bool destroy) 324366f6083SPeter Grehan { 325869c8d19SJohn Baldwin vmmops_vlapic_cleanup(vcpu->vlapic); 326869c8d19SJohn Baldwin vmmops_vcpu_cleanup(vcpu->cookie); 3271aa51504SJohn Baldwin vcpu->cookie = NULL; 3285fcf252fSNeel Natu if (destroy) { 329366f6083SPeter Grehan vmm_stat_free(vcpu->stats); 33038f1b189SPeter Grehan fpu_save_area_free(vcpu->guestfpu); 33108ebb360SJohn Baldwin vcpu_lock_destroy(vcpu); 332366f6083SPeter Grehan } 3335fcf252fSNeel Natu } 334366f6083SPeter Grehan 335*98568a00SJohn Baldwin static struct vcpu * 336*98568a00SJohn Baldwin vcpu_alloc(struct vm *vm, int vcpu_id) 337366f6083SPeter Grehan { 338366f6083SPeter Grehan struct vcpu *vcpu; 339366f6083SPeter Grehan 340a488c9c9SRodney W. Grimes KASSERT(vcpu_id >= 0 && vcpu_id < vm->maxcpus, 3415fcf252fSNeel Natu ("vcpu_init: invalid vcpu %d", vcpu_id)); 3425fcf252fSNeel Natu 343*98568a00SJohn Baldwin vcpu = malloc(sizeof(*vcpu), M_VM, M_WAITOK | M_ZERO); 34475dd3366SNeel Natu vcpu_lock_init(vcpu); 3455fcf252fSNeel Natu vcpu->state = VCPU_IDLE; 34675dd3366SNeel Natu vcpu->hostcpu = NOCPU; 347950af9ffSJohn Baldwin vcpu->vcpuid = vcpu_id; 348950af9ffSJohn Baldwin vcpu->vm = vm; 3495fcf252fSNeel Natu vcpu->guestfpu = fpu_save_area_alloc(); 3505fcf252fSNeel Natu vcpu->stats = vmm_stat_alloc(); 351483d953aSJohn Baldwin vcpu->tsc_offset = 0; 352*98568a00SJohn Baldwin return (vcpu); 3535fcf252fSNeel Natu } 3545fcf252fSNeel Natu 355*98568a00SJohn Baldwin static void 356*98568a00SJohn Baldwin vcpu_init(struct vcpu *vcpu) 357*98568a00SJohn Baldwin { 358*98568a00SJohn Baldwin vcpu->cookie = vmmops_vcpu_init(vcpu->vm->cookie, vcpu, vcpu->vcpuid); 359869c8d19SJohn Baldwin vcpu->vlapic = vmmops_vlapic_init(vcpu->cookie); 3603f0f4b15SJohn Baldwin vm_set_x2apic_state(vcpu, X2APIC_DISABLED); 361248e6799SNeel Natu vcpu->reqidle = 0; 362091d4532SNeel Natu vcpu->exitintinfo = 0; 3635fcf252fSNeel Natu vcpu->nmi_pending = 0; 3645fcf252fSNeel Natu vcpu->extint_pending = 0; 3655fcf252fSNeel Natu vcpu->exception_pending = 0; 366abb023fbSJohn Baldwin vcpu->guest_xcr0 = XFEATURE_ENABLED_X87; 36738f1b189SPeter Grehan fpu_save_area_reset(vcpu->guestfpu); 3685fcf252fSNeel Natu vmm_stat_init(vcpu->stats); 369366f6083SPeter Grehan } 370366f6083SPeter Grehan 371b0538143SNeel Natu int 37280cb5d84SJohn Baldwin vcpu_trace_exceptions(struct vcpu *vcpu) 373b0538143SNeel Natu { 374b0538143SNeel Natu 375b0538143SNeel Natu return (trace_guest_exceptions); 376b0538143SNeel Natu } 377b0538143SNeel Natu 3783ba952e1SCorvin Köhne int 37980cb5d84SJohn Baldwin vcpu_trap_wbinvd(struct vcpu *vcpu) 3803ba952e1SCorvin Köhne { 3813ba952e1SCorvin Köhne return (trap_wbinvd); 3823ba952e1SCorvin Köhne } 3833ba952e1SCorvin Köhne 38498ed632cSNeel Natu struct vm_exit * 38580cb5d84SJohn Baldwin vm_exitinfo(struct vcpu *vcpu) 38698ed632cSNeel Natu { 38798ed632cSNeel Natu return (&vcpu->exitinfo); 38898ed632cSNeel Natu } 38998ed632cSNeel Natu 390366f6083SPeter Grehan static int 391366f6083SPeter Grehan vmm_init(void) 392366f6083SPeter Grehan { 393366f6083SPeter Grehan int error; 394366f6083SPeter Grehan 39515add60dSPeter Grehan if (!vmm_is_hw_supported()) 39615add60dSPeter Grehan return (ENXIO); 39715add60dSPeter Grehan 398b01c2033SNeel Natu vmm_host_state_init(); 399add611fdSNeel Natu 400bd50262fSKonstantin Belousov vmm_ipinum = lapic_ipi_alloc(pti ? &IDTVEC(justreturn1_pti) : 401bd50262fSKonstantin Belousov &IDTVEC(justreturn)); 40218a2b08eSNeel Natu if (vmm_ipinum < 0) 403add611fdSNeel Natu vmm_ipinum = IPI_AST; 404366f6083SPeter Grehan 405366f6083SPeter Grehan error = vmm_mem_init(); 406366f6083SPeter Grehan if (error) 407366f6083SPeter Grehan return (error); 408366f6083SPeter Grehan 40915add60dSPeter Grehan vmm_resume_p = vmmops_modresume; 410366f6083SPeter Grehan 41115add60dSPeter Grehan return (vmmops_modinit(vmm_ipinum)); 412366f6083SPeter Grehan } 413366f6083SPeter Grehan 414366f6083SPeter Grehan static int 415366f6083SPeter Grehan vmm_handler(module_t mod, int what, void *arg) 416366f6083SPeter Grehan { 417366f6083SPeter Grehan int error; 418366f6083SPeter Grehan 419366f6083SPeter Grehan switch (what) { 420366f6083SPeter Grehan case MOD_LOAD: 42115add60dSPeter Grehan if (vmm_is_hw_supported()) { 422366f6083SPeter Grehan vmmdev_init(); 423366f6083SPeter Grehan error = vmm_init(); 424d5408b1dSNeel Natu if (error == 0) 425d5408b1dSNeel Natu vmm_initialized = 1; 42615add60dSPeter Grehan } else { 42715add60dSPeter Grehan error = ENXIO; 42815add60dSPeter Grehan } 429366f6083SPeter Grehan break; 430366f6083SPeter Grehan case MOD_UNLOAD: 43115add60dSPeter Grehan if (vmm_is_hw_supported()) { 432cdc5b9e7SNeel Natu error = vmmdev_cleanup(); 433cdc5b9e7SNeel Natu if (error == 0) { 43463e62d39SJohn Baldwin vmm_resume_p = NULL; 435366f6083SPeter Grehan iommu_cleanup(); 436add611fdSNeel Natu if (vmm_ipinum != IPI_AST) 43718a2b08eSNeel Natu lapic_ipi_free(vmm_ipinum); 43815add60dSPeter Grehan error = vmmops_modcleanup(); 43981ef6611SPeter Grehan /* 44081ef6611SPeter Grehan * Something bad happened - prevent new 44181ef6611SPeter Grehan * VMs from being created 44281ef6611SPeter Grehan */ 44381ef6611SPeter Grehan if (error) 444d5408b1dSNeel Natu vmm_initialized = 0; 44581ef6611SPeter Grehan } 44615add60dSPeter Grehan } else { 44715add60dSPeter Grehan error = 0; 44815add60dSPeter Grehan } 449366f6083SPeter Grehan break; 450366f6083SPeter Grehan default: 451366f6083SPeter Grehan error = 0; 452366f6083SPeter Grehan break; 453366f6083SPeter Grehan } 454366f6083SPeter Grehan return (error); 455366f6083SPeter Grehan } 456366f6083SPeter Grehan 457366f6083SPeter Grehan static moduledata_t vmm_kmod = { 458366f6083SPeter Grehan "vmm", 459366f6083SPeter Grehan vmm_handler, 460366f6083SPeter Grehan NULL 461366f6083SPeter Grehan }; 462366f6083SPeter Grehan 463366f6083SPeter Grehan /* 464e3f0800bSNeel Natu * vmm initialization has the following dependencies: 465e3f0800bSNeel Natu * 466e3f0800bSNeel Natu * - VT-x initialization requires smp_rendezvous() and therefore must happen 467e3f0800bSNeel Natu * after SMP is fully functional (after SI_SUB_SMP). 468366f6083SPeter Grehan */ 469e3f0800bSNeel Natu DECLARE_MODULE(vmm, vmm_kmod, SI_SUB_SMP + 1, SI_ORDER_ANY); 470366f6083SPeter Grehan MODULE_VERSION(vmm, 1); 471366f6083SPeter Grehan 4725fcf252fSNeel Natu static void 4735fcf252fSNeel Natu vm_init(struct vm *vm, bool create) 4745fcf252fSNeel Natu { 47515add60dSPeter Grehan vm->cookie = vmmops_init(vm, vmspace_pmap(vm->vmspace)); 4765fcf252fSNeel Natu vm->iommu = NULL; 4775fcf252fSNeel Natu vm->vioapic = vioapic_init(vm); 4785fcf252fSNeel Natu vm->vhpet = vhpet_init(vm); 4795fcf252fSNeel Natu vm->vatpic = vatpic_init(vm); 4805fcf252fSNeel Natu vm->vatpit = vatpit_init(vm); 481160ef77aSNeel Natu vm->vpmtmr = vpmtmr_init(vm); 4820dafa5cdSNeel Natu if (create) 4830dafa5cdSNeel Natu vm->vrtc = vrtc_init(vm); 4845fcf252fSNeel Natu 4855fcf252fSNeel Natu CPU_ZERO(&vm->active_cpus); 486fc276d92SJohn Baldwin CPU_ZERO(&vm->debug_cpus); 487c0f35dbfSJohn Baldwin CPU_ZERO(&vm->startup_cpus); 4885fcf252fSNeel Natu 4895fcf252fSNeel Natu vm->suspend = 0; 4905fcf252fSNeel Natu CPU_ZERO(&vm->suspended_cpus); 4915fcf252fSNeel Natu 492*98568a00SJohn Baldwin if (!create) { 493*98568a00SJohn Baldwin for (int i = 0; i < vm->maxcpus; i++) { 494*98568a00SJohn Baldwin if (vm->vcpu[i] != NULL) 495*98568a00SJohn Baldwin vcpu_init(vm->vcpu[i]); 496*98568a00SJohn Baldwin } 497*98568a00SJohn Baldwin } 498*98568a00SJohn Baldwin } 499*98568a00SJohn Baldwin 500*98568a00SJohn Baldwin void 501*98568a00SJohn Baldwin vm_disable_vcpu_creation(struct vm *vm) 502*98568a00SJohn Baldwin { 503*98568a00SJohn Baldwin sx_xlock(&vm->vcpus_init_lock); 504*98568a00SJohn Baldwin vm->dying = true; 505*98568a00SJohn Baldwin sx_xunlock(&vm->vcpus_init_lock); 506*98568a00SJohn Baldwin } 507*98568a00SJohn Baldwin 508*98568a00SJohn Baldwin struct vcpu * 509*98568a00SJohn Baldwin vm_alloc_vcpu(struct vm *vm, int vcpuid) 510*98568a00SJohn Baldwin { 511*98568a00SJohn Baldwin struct vcpu *vcpu; 512*98568a00SJohn Baldwin 513*98568a00SJohn Baldwin if (vcpuid < 0 || vcpuid >= vm_get_maxcpus(vm)) 514*98568a00SJohn Baldwin return (NULL); 515*98568a00SJohn Baldwin 516*98568a00SJohn Baldwin vcpu = atomic_load_ptr(&vm->vcpu[vcpuid]); 517*98568a00SJohn Baldwin if (__predict_true(vcpu != NULL)) 518*98568a00SJohn Baldwin return (vcpu); 519*98568a00SJohn Baldwin 520*98568a00SJohn Baldwin sx_xlock(&vm->vcpus_init_lock); 521*98568a00SJohn Baldwin vcpu = vm->vcpu[vcpuid]; 522*98568a00SJohn Baldwin if (vcpu == NULL && !vm->dying) { 523*98568a00SJohn Baldwin vcpu = vcpu_alloc(vm, vcpuid); 524*98568a00SJohn Baldwin vcpu_init(vcpu); 525*98568a00SJohn Baldwin 526*98568a00SJohn Baldwin /* 527*98568a00SJohn Baldwin * Ensure vCPU is fully created before updating pointer 528*98568a00SJohn Baldwin * to permit unlocked reads above. 529*98568a00SJohn Baldwin */ 530*98568a00SJohn Baldwin atomic_store_rel_ptr((uintptr_t *)&vm->vcpu[vcpuid], 531*98568a00SJohn Baldwin (uintptr_t)vcpu); 532*98568a00SJohn Baldwin } 533*98568a00SJohn Baldwin sx_xunlock(&vm->vcpus_init_lock); 534*98568a00SJohn Baldwin return (vcpu); 535*98568a00SJohn Baldwin } 536*98568a00SJohn Baldwin 537*98568a00SJohn Baldwin void 538*98568a00SJohn Baldwin vm_slock_vcpus(struct vm *vm) 539*98568a00SJohn Baldwin { 540*98568a00SJohn Baldwin sx_slock(&vm->vcpus_init_lock); 541*98568a00SJohn Baldwin } 542*98568a00SJohn Baldwin 543*98568a00SJohn Baldwin void 544*98568a00SJohn Baldwin vm_unlock_vcpus(struct vm *vm) 545*98568a00SJohn Baldwin { 546*98568a00SJohn Baldwin sx_unlock(&vm->vcpus_init_lock); 5475fcf252fSNeel Natu } 5485fcf252fSNeel Natu 54901d822d3SRodney W. Grimes /* 55001d822d3SRodney W. Grimes * The default CPU topology is a single thread per package. 55101d822d3SRodney W. Grimes */ 55201d822d3SRodney W. Grimes u_int cores_per_package = 1; 55301d822d3SRodney W. Grimes u_int threads_per_core = 1; 55401d822d3SRodney W. Grimes 555d5408b1dSNeel Natu int 556d5408b1dSNeel Natu vm_create(const char *name, struct vm **retvm) 557366f6083SPeter Grehan { 558366f6083SPeter Grehan struct vm *vm; 559318224bbSNeel Natu struct vmspace *vmspace; 560366f6083SPeter Grehan 561d5408b1dSNeel Natu /* 562d5408b1dSNeel Natu * If vmm.ko could not be successfully initialized then don't attempt 563d5408b1dSNeel Natu * to create the virtual machine. 564d5408b1dSNeel Natu */ 565d5408b1dSNeel Natu if (!vmm_initialized) 566d5408b1dSNeel Natu return (ENXIO); 567d5408b1dSNeel Natu 568df95cc76SKa Ho Ng if (name == NULL || strnlen(name, VM_MAX_NAMELEN + 1) == 569df95cc76SKa Ho Ng VM_MAX_NAMELEN + 1) 570d5408b1dSNeel Natu return (EINVAL); 571366f6083SPeter Grehan 5723c48106aSKonstantin Belousov vmspace = vmmops_vmspace_alloc(0, VM_MAXUSER_ADDRESS_LA48); 573318224bbSNeel Natu if (vmspace == NULL) 574318224bbSNeel Natu return (ENOMEM); 575318224bbSNeel Natu 576366f6083SPeter Grehan vm = malloc(sizeof(struct vm), M_VM, M_WAITOK | M_ZERO); 577366f6083SPeter Grehan strcpy(vm->name, name); 57888c4b8d1SNeel Natu vm->vmspace = vmspace; 5795b8a8cd1SNeel Natu mtx_init(&vm->rendezvous_mtx, "vm rendezvous lock", 0, MTX_DEF); 58067b69e76SJohn Baldwin sx_init(&vm->mem_segs_lock, "vm mem_segs"); 581*98568a00SJohn Baldwin sx_init(&vm->vcpus_init_lock, "vm vcpus"); 582366f6083SPeter Grehan 58301d822d3SRodney W. Grimes vm->sockets = 1; 58401d822d3SRodney W. Grimes vm->cores = cores_per_package; /* XXX backwards compatibility */ 58501d822d3SRodney W. Grimes vm->threads = threads_per_core; /* XXX backwards compatibility */ 586a488c9c9SRodney W. Grimes vm->maxcpus = VM_MAXCPU; /* XXX temp to keep code working */ 58701d822d3SRodney W. Grimes 5885fcf252fSNeel Natu vm_init(vm, true); 589366f6083SPeter Grehan 590d5408b1dSNeel Natu *retvm = vm; 591d5408b1dSNeel Natu return (0); 592366f6083SPeter Grehan } 593366f6083SPeter Grehan 59401d822d3SRodney W. Grimes void 59501d822d3SRodney W. Grimes vm_get_topology(struct vm *vm, uint16_t *sockets, uint16_t *cores, 59601d822d3SRodney W. Grimes uint16_t *threads, uint16_t *maxcpus) 59701d822d3SRodney W. Grimes { 59801d822d3SRodney W. Grimes *sockets = vm->sockets; 59901d822d3SRodney W. Grimes *cores = vm->cores; 60001d822d3SRodney W. Grimes *threads = vm->threads; 60101d822d3SRodney W. Grimes *maxcpus = vm->maxcpus; 60201d822d3SRodney W. Grimes } 60301d822d3SRodney W. Grimes 604a488c9c9SRodney W. Grimes uint16_t 605a488c9c9SRodney W. Grimes vm_get_maxcpus(struct vm *vm) 606a488c9c9SRodney W. Grimes { 607a488c9c9SRodney W. Grimes return (vm->maxcpus); 608a488c9c9SRodney W. Grimes } 609a488c9c9SRodney W. Grimes 61001d822d3SRodney W. Grimes int 61101d822d3SRodney W. Grimes vm_set_topology(struct vm *vm, uint16_t sockets, uint16_t cores, 612*98568a00SJohn Baldwin uint16_t threads, uint16_t maxcpus __unused) 61301d822d3SRodney W. Grimes { 614*98568a00SJohn Baldwin /* Ignore maxcpus. */ 615a488c9c9SRodney W. Grimes if ((sockets * cores * threads) > vm->maxcpus) 61601d822d3SRodney W. Grimes return (EINVAL); 61701d822d3SRodney W. Grimes vm->sockets = sockets; 61801d822d3SRodney W. Grimes vm->cores = cores; 61901d822d3SRodney W. Grimes vm->threads = threads; 62001d822d3SRodney W. Grimes return(0); 62101d822d3SRodney W. Grimes } 62201d822d3SRodney W. Grimes 623f7d51510SNeel Natu static void 6245fcf252fSNeel Natu vm_cleanup(struct vm *vm, bool destroy) 625366f6083SPeter Grehan { 6269b1aa8d6SNeel Natu struct mem_map *mm; 627366f6083SPeter Grehan int i; 628366f6083SPeter Grehan 629366f6083SPeter Grehan ppt_unassign_all(vm); 630366f6083SPeter Grehan 631318224bbSNeel Natu if (vm->iommu != NULL) 632318224bbSNeel Natu iommu_destroy_domain(vm->iommu); 633318224bbSNeel Natu 6340dafa5cdSNeel Natu if (destroy) 6350dafa5cdSNeel Natu vrtc_cleanup(vm->vrtc); 6360dafa5cdSNeel Natu else 6370dafa5cdSNeel Natu vrtc_reset(vm->vrtc); 638160ef77aSNeel Natu vpmtmr_cleanup(vm->vpmtmr); 639e883c9bbSTycho Nightingale vatpit_cleanup(vm->vatpit); 64008e3ff32SNeel Natu vhpet_cleanup(vm->vhpet); 641762fd208STycho Nightingale vatpic_cleanup(vm->vatpic); 64208e3ff32SNeel Natu vioapic_cleanup(vm->vioapic); 64308e3ff32SNeel Natu 644*98568a00SJohn Baldwin for (i = 0; i < vm->maxcpus; i++) { 645*98568a00SJohn Baldwin if (vm->vcpu[i] != NULL) 646*98568a00SJohn Baldwin vcpu_cleanup(vm->vcpu[i], destroy); 647*98568a00SJohn Baldwin } 6485fcf252fSNeel Natu 64915add60dSPeter Grehan vmmops_cleanup(vm->cookie); 6505fcf252fSNeel Natu 6519b1aa8d6SNeel Natu /* 6529b1aa8d6SNeel Natu * System memory is removed from the guest address space only when 6539b1aa8d6SNeel Natu * the VM is destroyed. This is because the mapping remains the same 6549b1aa8d6SNeel Natu * across VM reset. 6559b1aa8d6SNeel Natu * 6569b1aa8d6SNeel Natu * Device memory can be relocated by the guest (e.g. using PCI BARs) 6579b1aa8d6SNeel Natu * so those mappings are removed on a VM reset. 6589b1aa8d6SNeel Natu */ 6599b1aa8d6SNeel Natu for (i = 0; i < VM_MAX_MEMMAPS; i++) { 6609b1aa8d6SNeel Natu mm = &vm->mem_maps[i]; 6619b1aa8d6SNeel Natu if (destroy || !sysmem_mapping(vm, mm)) 6629b1aa8d6SNeel Natu vm_free_memmap(vm, i); 6639b1aa8d6SNeel Natu } 664f7d51510SNeel Natu 6659b1aa8d6SNeel Natu if (destroy) { 6669b1aa8d6SNeel Natu for (i = 0; i < VM_MAX_MEMSEGS; i++) 6679b1aa8d6SNeel Natu vm_free_memseg(vm, i); 668366f6083SPeter Grehan 66915add60dSPeter Grehan vmmops_vmspace_free(vm->vmspace); 6705fcf252fSNeel Natu vm->vmspace = NULL; 67108ebb360SJohn Baldwin 672*98568a00SJohn Baldwin sx_destroy(&vm->vcpus_init_lock); 67367b69e76SJohn Baldwin sx_destroy(&vm->mem_segs_lock); 67408ebb360SJohn Baldwin mtx_destroy(&vm->rendezvous_mtx); 6755fcf252fSNeel Natu } 6765fcf252fSNeel Natu } 677366f6083SPeter Grehan 6785fcf252fSNeel Natu void 6795fcf252fSNeel Natu vm_destroy(struct vm *vm) 6805fcf252fSNeel Natu { 6815fcf252fSNeel Natu vm_cleanup(vm, true); 682366f6083SPeter Grehan free(vm, M_VM); 683366f6083SPeter Grehan } 684366f6083SPeter Grehan 6855fcf252fSNeel Natu int 6865fcf252fSNeel Natu vm_reinit(struct vm *vm) 6875fcf252fSNeel Natu { 6885fcf252fSNeel Natu int error; 6895fcf252fSNeel Natu 6905fcf252fSNeel Natu /* 6915fcf252fSNeel Natu * A virtual machine can be reset only if all vcpus are suspended. 6925fcf252fSNeel Natu */ 6935fcf252fSNeel Natu if (CPU_CMP(&vm->suspended_cpus, &vm->active_cpus) == 0) { 6945fcf252fSNeel Natu vm_cleanup(vm, false); 6955fcf252fSNeel Natu vm_init(vm, false); 6965fcf252fSNeel Natu error = 0; 6975fcf252fSNeel Natu } else { 6985fcf252fSNeel Natu error = EBUSY; 6995fcf252fSNeel Natu } 7005fcf252fSNeel Natu 7015fcf252fSNeel Natu return (error); 7025fcf252fSNeel Natu } 7035fcf252fSNeel Natu 704366f6083SPeter Grehan const char * 705366f6083SPeter Grehan vm_name(struct vm *vm) 706366f6083SPeter Grehan { 707366f6083SPeter Grehan return (vm->name); 708366f6083SPeter Grehan } 709366f6083SPeter Grehan 71067b69e76SJohn Baldwin void 71167b69e76SJohn Baldwin vm_slock_memsegs(struct vm *vm) 71267b69e76SJohn Baldwin { 71367b69e76SJohn Baldwin sx_slock(&vm->mem_segs_lock); 71467b69e76SJohn Baldwin } 71567b69e76SJohn Baldwin 71667b69e76SJohn Baldwin void 71767b69e76SJohn Baldwin vm_xlock_memsegs(struct vm *vm) 71867b69e76SJohn Baldwin { 71967b69e76SJohn Baldwin sx_xlock(&vm->mem_segs_lock); 72067b69e76SJohn Baldwin } 72167b69e76SJohn Baldwin 72267b69e76SJohn Baldwin void 72367b69e76SJohn Baldwin vm_unlock_memsegs(struct vm *vm) 72467b69e76SJohn Baldwin { 72567b69e76SJohn Baldwin sx_unlock(&vm->mem_segs_lock); 72667b69e76SJohn Baldwin } 72767b69e76SJohn Baldwin 728366f6083SPeter Grehan int 729366f6083SPeter Grehan vm_map_mmio(struct vm *vm, vm_paddr_t gpa, size_t len, vm_paddr_t hpa) 730366f6083SPeter Grehan { 731318224bbSNeel Natu vm_object_t obj; 732366f6083SPeter Grehan 733318224bbSNeel Natu if ((obj = vmm_mmio_alloc(vm->vmspace, gpa, len, hpa)) == NULL) 734318224bbSNeel Natu return (ENOMEM); 735318224bbSNeel Natu else 736318224bbSNeel Natu return (0); 737366f6083SPeter Grehan } 738366f6083SPeter Grehan 739366f6083SPeter Grehan int 740366f6083SPeter Grehan vm_unmap_mmio(struct vm *vm, vm_paddr_t gpa, size_t len) 741366f6083SPeter Grehan { 742366f6083SPeter Grehan 743318224bbSNeel Natu vmm_mmio_free(vm->vmspace, gpa, len); 744318224bbSNeel Natu return (0); 745366f6083SPeter Grehan } 746366f6083SPeter Grehan 7479b1aa8d6SNeel Natu /* 7489b1aa8d6SNeel Natu * Return 'true' if 'gpa' is allocated in the guest address space. 7499b1aa8d6SNeel Natu * 7509b1aa8d6SNeel Natu * This function is called in the context of a running vcpu which acts as 7519b1aa8d6SNeel Natu * an implicit lock on 'vm->mem_maps[]'. 7529b1aa8d6SNeel Natu */ 7539b1aa8d6SNeel Natu bool 75480cb5d84SJohn Baldwin vm_mem_allocated(struct vcpu *vcpu, vm_paddr_t gpa) 755366f6083SPeter Grehan { 75680cb5d84SJohn Baldwin struct vm *vm = vcpu->vm; 7579b1aa8d6SNeel Natu struct mem_map *mm; 758341f19c9SNeel Natu int i; 759341f19c9SNeel Natu 7609b1aa8d6SNeel Natu #ifdef INVARIANTS 7619b1aa8d6SNeel Natu int hostcpu, state; 76280cb5d84SJohn Baldwin state = vcpu_get_state(vcpu, &hostcpu); 7639b1aa8d6SNeel Natu KASSERT(state == VCPU_RUNNING && hostcpu == curcpu, 7649b1aa8d6SNeel Natu ("%s: invalid vcpu state %d/%d", __func__, state, hostcpu)); 7659b1aa8d6SNeel Natu #endif 7669b1aa8d6SNeel Natu 7679b1aa8d6SNeel Natu for (i = 0; i < VM_MAX_MEMMAPS; i++) { 7689b1aa8d6SNeel Natu mm = &vm->mem_maps[i]; 7699b1aa8d6SNeel Natu if (mm->len != 0 && gpa >= mm->gpa && gpa < mm->gpa + mm->len) 7709b1aa8d6SNeel Natu return (true); /* 'gpa' is sysmem or devmem */ 771341f19c9SNeel Natu } 772341f19c9SNeel Natu 773318224bbSNeel Natu if (ppt_is_mmio(vm, gpa)) 7749b1aa8d6SNeel Natu return (true); /* 'gpa' is pci passthru mmio */ 775318224bbSNeel Natu 7769b1aa8d6SNeel Natu return (false); 777341f19c9SNeel Natu } 778341f19c9SNeel Natu 779341f19c9SNeel Natu int 7809b1aa8d6SNeel Natu vm_alloc_memseg(struct vm *vm, int ident, size_t len, bool sysmem) 781341f19c9SNeel Natu { 782318224bbSNeel Natu struct mem_seg *seg; 7839b1aa8d6SNeel Natu vm_object_t obj; 784366f6083SPeter Grehan 78567b69e76SJohn Baldwin sx_assert(&vm->mem_segs_lock, SX_XLOCKED); 78667b69e76SJohn Baldwin 7879b1aa8d6SNeel Natu if (ident < 0 || ident >= VM_MAX_MEMSEGS) 788341f19c9SNeel Natu return (EINVAL); 789341f19c9SNeel Natu 7909b1aa8d6SNeel Natu if (len == 0 || (len & PAGE_MASK)) 7919b1aa8d6SNeel Natu return (EINVAL); 792341f19c9SNeel Natu 7939b1aa8d6SNeel Natu seg = &vm->mem_segs[ident]; 7949b1aa8d6SNeel Natu if (seg->object != NULL) { 7959b1aa8d6SNeel Natu if (seg->len == len && seg->sysmem == sysmem) 7969b1aa8d6SNeel Natu return (EEXIST); 7979b1aa8d6SNeel Natu else 7989b1aa8d6SNeel Natu return (EINVAL); 799341f19c9SNeel Natu } 800341f19c9SNeel Natu 8016b389740SMark Johnston obj = vm_object_allocate(OBJT_SWAP, len >> PAGE_SHIFT); 8029b1aa8d6SNeel Natu if (obj == NULL) 803318224bbSNeel Natu return (ENOMEM); 804318224bbSNeel Natu 805318224bbSNeel Natu seg->len = len; 8069b1aa8d6SNeel Natu seg->object = obj; 8079b1aa8d6SNeel Natu seg->sysmem = sysmem; 808366f6083SPeter Grehan return (0); 809366f6083SPeter Grehan } 810366f6083SPeter Grehan 8119b1aa8d6SNeel Natu int 8129b1aa8d6SNeel Natu vm_get_memseg(struct vm *vm, int ident, size_t *len, bool *sysmem, 8139b1aa8d6SNeel Natu vm_object_t *objptr) 814477867a0SNeel Natu { 8159b1aa8d6SNeel Natu struct mem_seg *seg; 816477867a0SNeel Natu 81767b69e76SJohn Baldwin sx_assert(&vm->mem_segs_lock, SX_LOCKED); 81867b69e76SJohn Baldwin 8199b1aa8d6SNeel Natu if (ident < 0 || ident >= VM_MAX_MEMSEGS) 8209b1aa8d6SNeel Natu return (EINVAL); 8219b1aa8d6SNeel Natu 8229b1aa8d6SNeel Natu seg = &vm->mem_segs[ident]; 8239b1aa8d6SNeel Natu if (len) 8249b1aa8d6SNeel Natu *len = seg->len; 8259b1aa8d6SNeel Natu if (sysmem) 8269b1aa8d6SNeel Natu *sysmem = seg->sysmem; 8279b1aa8d6SNeel Natu if (objptr) 8289b1aa8d6SNeel Natu *objptr = seg->object; 8299b1aa8d6SNeel Natu return (0); 830477867a0SNeel Natu } 8319b1aa8d6SNeel Natu 8329b1aa8d6SNeel Natu void 8339b1aa8d6SNeel Natu vm_free_memseg(struct vm *vm, int ident) 8349b1aa8d6SNeel Natu { 8359b1aa8d6SNeel Natu struct mem_seg *seg; 8369b1aa8d6SNeel Natu 8379b1aa8d6SNeel Natu KASSERT(ident >= 0 && ident < VM_MAX_MEMSEGS, 8389b1aa8d6SNeel Natu ("%s: invalid memseg ident %d", __func__, ident)); 8399b1aa8d6SNeel Natu 8409b1aa8d6SNeel Natu seg = &vm->mem_segs[ident]; 8419b1aa8d6SNeel Natu if (seg->object != NULL) { 8429b1aa8d6SNeel Natu vm_object_deallocate(seg->object); 8439b1aa8d6SNeel Natu bzero(seg, sizeof(struct mem_seg)); 8449b1aa8d6SNeel Natu } 8459b1aa8d6SNeel Natu } 8469b1aa8d6SNeel Natu 8479b1aa8d6SNeel Natu int 8489b1aa8d6SNeel Natu vm_mmap_memseg(struct vm *vm, vm_paddr_t gpa, int segid, vm_ooffset_t first, 8499b1aa8d6SNeel Natu size_t len, int prot, int flags) 8509b1aa8d6SNeel Natu { 8519b1aa8d6SNeel Natu struct mem_seg *seg; 8529b1aa8d6SNeel Natu struct mem_map *m, *map; 8539b1aa8d6SNeel Natu vm_ooffset_t last; 8549b1aa8d6SNeel Natu int i, error; 8559b1aa8d6SNeel Natu 8569b1aa8d6SNeel Natu if (prot == 0 || (prot & ~(VM_PROT_ALL)) != 0) 8579b1aa8d6SNeel Natu return (EINVAL); 8589b1aa8d6SNeel Natu 8599b1aa8d6SNeel Natu if (flags & ~VM_MEMMAP_F_WIRED) 8609b1aa8d6SNeel Natu return (EINVAL); 8619b1aa8d6SNeel Natu 8629b1aa8d6SNeel Natu if (segid < 0 || segid >= VM_MAX_MEMSEGS) 8639b1aa8d6SNeel Natu return (EINVAL); 8649b1aa8d6SNeel Natu 8659b1aa8d6SNeel Natu seg = &vm->mem_segs[segid]; 8669b1aa8d6SNeel Natu if (seg->object == NULL) 8679b1aa8d6SNeel Natu return (EINVAL); 8689b1aa8d6SNeel Natu 8699b1aa8d6SNeel Natu last = first + len; 8709b1aa8d6SNeel Natu if (first < 0 || first >= last || last > seg->len) 8719b1aa8d6SNeel Natu return (EINVAL); 8729b1aa8d6SNeel Natu 8739b1aa8d6SNeel Natu if ((gpa | first | last) & PAGE_MASK) 8749b1aa8d6SNeel Natu return (EINVAL); 8759b1aa8d6SNeel Natu 8769b1aa8d6SNeel Natu map = NULL; 8779b1aa8d6SNeel Natu for (i = 0; i < VM_MAX_MEMMAPS; i++) { 8789b1aa8d6SNeel Natu m = &vm->mem_maps[i]; 8799b1aa8d6SNeel Natu if (m->len == 0) { 8809b1aa8d6SNeel Natu map = m; 8819b1aa8d6SNeel Natu break; 8829b1aa8d6SNeel Natu } 8839b1aa8d6SNeel Natu } 8849b1aa8d6SNeel Natu 8859b1aa8d6SNeel Natu if (map == NULL) 8869b1aa8d6SNeel Natu return (ENOSPC); 8879b1aa8d6SNeel Natu 8889b1aa8d6SNeel Natu error = vm_map_find(&vm->vmspace->vm_map, seg->object, first, &gpa, 8899b1aa8d6SNeel Natu len, 0, VMFS_NO_SPACE, prot, prot, 0); 8909b1aa8d6SNeel Natu if (error != KERN_SUCCESS) 8919b1aa8d6SNeel Natu return (EFAULT); 8929b1aa8d6SNeel Natu 8939b1aa8d6SNeel Natu vm_object_reference(seg->object); 8949b1aa8d6SNeel Natu 8959b1aa8d6SNeel Natu if (flags & VM_MEMMAP_F_WIRED) { 8969b1aa8d6SNeel Natu error = vm_map_wire(&vm->vmspace->vm_map, gpa, gpa + len, 8979b1aa8d6SNeel Natu VM_MAP_WIRE_USER | VM_MAP_WIRE_NOHOLES); 8989b1aa8d6SNeel Natu if (error != KERN_SUCCESS) { 8999b1aa8d6SNeel Natu vm_map_remove(&vm->vmspace->vm_map, gpa, gpa + len); 90054a3a114SMark Johnston return (error == KERN_RESOURCE_SHORTAGE ? ENOMEM : 90154a3a114SMark Johnston EFAULT); 9029b1aa8d6SNeel Natu } 9039b1aa8d6SNeel Natu } 9049b1aa8d6SNeel Natu 9059b1aa8d6SNeel Natu map->gpa = gpa; 9069b1aa8d6SNeel Natu map->len = len; 9079b1aa8d6SNeel Natu map->segoff = first; 9089b1aa8d6SNeel Natu map->segid = segid; 9099b1aa8d6SNeel Natu map->prot = prot; 9109b1aa8d6SNeel Natu map->flags = flags; 9119b1aa8d6SNeel Natu return (0); 9129b1aa8d6SNeel Natu } 9139b1aa8d6SNeel Natu 9149b1aa8d6SNeel Natu int 915f8a6ec2dSD Scott Phillips vm_munmap_memseg(struct vm *vm, vm_paddr_t gpa, size_t len) 916f8a6ec2dSD Scott Phillips { 917f8a6ec2dSD Scott Phillips struct mem_map *m; 918f8a6ec2dSD Scott Phillips int i; 919f8a6ec2dSD Scott Phillips 920f8a6ec2dSD Scott Phillips for (i = 0; i < VM_MAX_MEMMAPS; i++) { 921f8a6ec2dSD Scott Phillips m = &vm->mem_maps[i]; 922f8a6ec2dSD Scott Phillips if (m->gpa == gpa && m->len == len && 923f8a6ec2dSD Scott Phillips (m->flags & VM_MEMMAP_F_IOMMU) == 0) { 924f8a6ec2dSD Scott Phillips vm_free_memmap(vm, i); 925f8a6ec2dSD Scott Phillips return (0); 926f8a6ec2dSD Scott Phillips } 927f8a6ec2dSD Scott Phillips } 928f8a6ec2dSD Scott Phillips 929f8a6ec2dSD Scott Phillips return (EINVAL); 930f8a6ec2dSD Scott Phillips } 931f8a6ec2dSD Scott Phillips 932f8a6ec2dSD Scott Phillips int 9339b1aa8d6SNeel Natu vm_mmap_getnext(struct vm *vm, vm_paddr_t *gpa, int *segid, 9349b1aa8d6SNeel Natu vm_ooffset_t *segoff, size_t *len, int *prot, int *flags) 9359b1aa8d6SNeel Natu { 9369b1aa8d6SNeel Natu struct mem_map *mm, *mmnext; 9379b1aa8d6SNeel Natu int i; 9389b1aa8d6SNeel Natu 9399b1aa8d6SNeel Natu mmnext = NULL; 9409b1aa8d6SNeel Natu for (i = 0; i < VM_MAX_MEMMAPS; i++) { 9419b1aa8d6SNeel Natu mm = &vm->mem_maps[i]; 9429b1aa8d6SNeel Natu if (mm->len == 0 || mm->gpa < *gpa) 9439b1aa8d6SNeel Natu continue; 9449b1aa8d6SNeel Natu if (mmnext == NULL || mm->gpa < mmnext->gpa) 9459b1aa8d6SNeel Natu mmnext = mm; 9469b1aa8d6SNeel Natu } 9479b1aa8d6SNeel Natu 9489b1aa8d6SNeel Natu if (mmnext != NULL) { 9499b1aa8d6SNeel Natu *gpa = mmnext->gpa; 9509b1aa8d6SNeel Natu if (segid) 9519b1aa8d6SNeel Natu *segid = mmnext->segid; 9529b1aa8d6SNeel Natu if (segoff) 9539b1aa8d6SNeel Natu *segoff = mmnext->segoff; 9549b1aa8d6SNeel Natu if (len) 9559b1aa8d6SNeel Natu *len = mmnext->len; 9569b1aa8d6SNeel Natu if (prot) 9579b1aa8d6SNeel Natu *prot = mmnext->prot; 9589b1aa8d6SNeel Natu if (flags) 9599b1aa8d6SNeel Natu *flags = mmnext->flags; 9609b1aa8d6SNeel Natu return (0); 9619b1aa8d6SNeel Natu } else { 9629b1aa8d6SNeel Natu return (ENOENT); 9639b1aa8d6SNeel Natu } 964477867a0SNeel Natu } 965477867a0SNeel Natu 966318224bbSNeel Natu static void 9679b1aa8d6SNeel Natu vm_free_memmap(struct vm *vm, int ident) 968366f6083SPeter Grehan { 9699b1aa8d6SNeel Natu struct mem_map *mm; 97073505a10SRobert Wing int error __diagused; 9714db4fb2cSNeel Natu 9729b1aa8d6SNeel Natu mm = &vm->mem_maps[ident]; 9739b1aa8d6SNeel Natu if (mm->len) { 9749b1aa8d6SNeel Natu error = vm_map_remove(&vm->vmspace->vm_map, mm->gpa, 9759b1aa8d6SNeel Natu mm->gpa + mm->len); 9769b1aa8d6SNeel Natu KASSERT(error == KERN_SUCCESS, ("%s: vm_map_remove error %d", 9779b1aa8d6SNeel Natu __func__, error)); 9789b1aa8d6SNeel Natu bzero(mm, sizeof(struct mem_map)); 979318224bbSNeel Natu } 980318224bbSNeel Natu } 981318224bbSNeel Natu 9829b1aa8d6SNeel Natu static __inline bool 9839b1aa8d6SNeel Natu sysmem_mapping(struct vm *vm, struct mem_map *mm) 984318224bbSNeel Natu { 985318224bbSNeel Natu 9869b1aa8d6SNeel Natu if (mm->len != 0 && vm->mem_segs[mm->segid].sysmem) 9879b1aa8d6SNeel Natu return (true); 9889b1aa8d6SNeel Natu else 9899b1aa8d6SNeel Natu return (false); 990318224bbSNeel Natu } 991318224bbSNeel Natu 992147d12a7SAntoine Brodin vm_paddr_t 993147d12a7SAntoine Brodin vmm_sysmem_maxaddr(struct vm *vm) 9949b1aa8d6SNeel Natu { 9959b1aa8d6SNeel Natu struct mem_map *mm; 9969b1aa8d6SNeel Natu vm_paddr_t maxaddr; 9979b1aa8d6SNeel Natu int i; 998318224bbSNeel Natu 9999b1aa8d6SNeel Natu maxaddr = 0; 10009b1aa8d6SNeel Natu for (i = 0; i < VM_MAX_MEMMAPS; i++) { 10019b1aa8d6SNeel Natu mm = &vm->mem_maps[i]; 10029b1aa8d6SNeel Natu if (sysmem_mapping(vm, mm)) { 10039b1aa8d6SNeel Natu if (maxaddr < mm->gpa + mm->len) 10049b1aa8d6SNeel Natu maxaddr = mm->gpa + mm->len; 10059b1aa8d6SNeel Natu } 10069b1aa8d6SNeel Natu } 10079b1aa8d6SNeel Natu return (maxaddr); 1008318224bbSNeel Natu } 1009318224bbSNeel Natu 1010318224bbSNeel Natu static void 1011490d56c5SEd Maste vm_iommu_modify(struct vm *vm, bool map) 1012318224bbSNeel Natu { 1013318224bbSNeel Natu int i, sz; 1014318224bbSNeel Natu vm_paddr_t gpa, hpa; 10159b1aa8d6SNeel Natu struct mem_map *mm; 1016318224bbSNeel Natu void *vp, *cookie, *host_domain; 1017318224bbSNeel Natu 1018318224bbSNeel Natu sz = PAGE_SIZE; 1019318224bbSNeel Natu host_domain = iommu_host_domain(); 1020318224bbSNeel Natu 10219b1aa8d6SNeel Natu for (i = 0; i < VM_MAX_MEMMAPS; i++) { 10229b1aa8d6SNeel Natu mm = &vm->mem_maps[i]; 10239b1aa8d6SNeel Natu if (!sysmem_mapping(vm, mm)) 10249b1aa8d6SNeel Natu continue; 1025318224bbSNeel Natu 10269b1aa8d6SNeel Natu if (map) { 10279b1aa8d6SNeel Natu KASSERT((mm->flags & VM_MEMMAP_F_IOMMU) == 0, 10289b1aa8d6SNeel Natu ("iommu map found invalid memmap %#lx/%#lx/%#x", 10299b1aa8d6SNeel Natu mm->gpa, mm->len, mm->flags)); 10309b1aa8d6SNeel Natu if ((mm->flags & VM_MEMMAP_F_WIRED) == 0) 10319b1aa8d6SNeel Natu continue; 10329b1aa8d6SNeel Natu mm->flags |= VM_MEMMAP_F_IOMMU; 10339b1aa8d6SNeel Natu } else { 10349b1aa8d6SNeel Natu if ((mm->flags & VM_MEMMAP_F_IOMMU) == 0) 10359b1aa8d6SNeel Natu continue; 10369b1aa8d6SNeel Natu mm->flags &= ~VM_MEMMAP_F_IOMMU; 10379b1aa8d6SNeel Natu KASSERT((mm->flags & VM_MEMMAP_F_WIRED) != 0, 10389b1aa8d6SNeel Natu ("iommu unmap found invalid memmap %#lx/%#lx/%#x", 10399b1aa8d6SNeel Natu mm->gpa, mm->len, mm->flags)); 10409b1aa8d6SNeel Natu } 10419b1aa8d6SNeel Natu 10429b1aa8d6SNeel Natu gpa = mm->gpa; 10439b1aa8d6SNeel Natu while (gpa < mm->gpa + mm->len) { 104428b561adSJohn Baldwin vp = vm_gpa_hold_global(vm, gpa, PAGE_SIZE, 104528b561adSJohn Baldwin VM_PROT_WRITE, &cookie); 1046318224bbSNeel Natu KASSERT(vp != NULL, ("vm(%s) could not map gpa %#lx", 1047318224bbSNeel Natu vm_name(vm), gpa)); 1048318224bbSNeel Natu 1049318224bbSNeel Natu vm_gpa_release(cookie); 1050318224bbSNeel Natu 1051318224bbSNeel Natu hpa = DMAP_TO_PHYS((uintptr_t)vp); 1052318224bbSNeel Natu if (map) { 1053318224bbSNeel Natu iommu_create_mapping(vm->iommu, gpa, hpa, sz); 1054318224bbSNeel Natu } else { 1055318224bbSNeel Natu iommu_remove_mapping(vm->iommu, gpa, sz); 1056318224bbSNeel Natu } 1057318224bbSNeel Natu 1058318224bbSNeel Natu gpa += PAGE_SIZE; 1059318224bbSNeel Natu } 1060318224bbSNeel Natu } 1061318224bbSNeel Natu 1062318224bbSNeel Natu /* 1063318224bbSNeel Natu * Invalidate the cached translations associated with the domain 1064318224bbSNeel Natu * from which pages were removed. 1065318224bbSNeel Natu */ 1066318224bbSNeel Natu if (map) 1067318224bbSNeel Natu iommu_invalidate_tlb(host_domain); 1068318224bbSNeel Natu else 1069318224bbSNeel Natu iommu_invalidate_tlb(vm->iommu); 1070318224bbSNeel Natu } 1071318224bbSNeel Natu 1072490d56c5SEd Maste #define vm_iommu_unmap(vm) vm_iommu_modify((vm), false) 1073490d56c5SEd Maste #define vm_iommu_map(vm) vm_iommu_modify((vm), true) 1074318224bbSNeel Natu 1075318224bbSNeel Natu int 1076318224bbSNeel Natu vm_unassign_pptdev(struct vm *vm, int bus, int slot, int func) 1077318224bbSNeel Natu { 1078318224bbSNeel Natu int error; 1079318224bbSNeel Natu 1080318224bbSNeel Natu error = ppt_unassign_device(vm, bus, slot, func); 1081318224bbSNeel Natu if (error) 1082318224bbSNeel Natu return (error); 1083318224bbSNeel Natu 10849b1aa8d6SNeel Natu if (ppt_assigned_devices(vm) == 0) 1085318224bbSNeel Natu vm_iommu_unmap(vm); 10869b1aa8d6SNeel Natu 1087318224bbSNeel Natu return (0); 1088318224bbSNeel Natu } 1089318224bbSNeel Natu 1090318224bbSNeel Natu int 1091318224bbSNeel Natu vm_assign_pptdev(struct vm *vm, int bus, int slot, int func) 1092318224bbSNeel Natu { 1093318224bbSNeel Natu int error; 1094318224bbSNeel Natu vm_paddr_t maxaddr; 1095318224bbSNeel Natu 10969b1aa8d6SNeel Natu /* Set up the IOMMU to do the 'gpa' to 'hpa' translation */ 109751f45d01SNeel Natu if (ppt_assigned_devices(vm) == 0) { 1098318224bbSNeel Natu KASSERT(vm->iommu == NULL, 1099318224bbSNeel Natu ("vm_assign_pptdev: iommu must be NULL")); 1100147d12a7SAntoine Brodin maxaddr = vmm_sysmem_maxaddr(vm); 1101318224bbSNeel Natu vm->iommu = iommu_create_domain(maxaddr); 1102ffe1b10dSJohn Baldwin if (vm->iommu == NULL) 1103ffe1b10dSJohn Baldwin return (ENXIO); 1104318224bbSNeel Natu vm_iommu_map(vm); 1105318224bbSNeel Natu } 1106318224bbSNeel Natu 1107318224bbSNeel Natu error = ppt_assign_device(vm, bus, slot, func); 1108318224bbSNeel Natu return (error); 1109318224bbSNeel Natu } 1110318224bbSNeel Natu 111128b561adSJohn Baldwin static void * 111228b561adSJohn Baldwin _vm_gpa_hold(struct vm *vm, vm_paddr_t gpa, size_t len, int reqprot, 1113318224bbSNeel Natu void **cookie) 1114318224bbSNeel Natu { 11159b1aa8d6SNeel Natu int i, count, pageoff; 11169b1aa8d6SNeel Natu struct mem_map *mm; 1117318224bbSNeel Natu vm_page_t m; 111828b561adSJohn Baldwin 1119318224bbSNeel Natu pageoff = gpa & PAGE_MASK; 1120318224bbSNeel Natu if (len > PAGE_SIZE - pageoff) 1121318224bbSNeel Natu panic("vm_gpa_hold: invalid gpa/len: 0x%016lx/%lu", gpa, len); 1122318224bbSNeel Natu 11239b1aa8d6SNeel Natu count = 0; 11249b1aa8d6SNeel Natu for (i = 0; i < VM_MAX_MEMMAPS; i++) { 11259b1aa8d6SNeel Natu mm = &vm->mem_maps[i]; 112646567b4fSPeter Grehan if (gpa >= mm->gpa && gpa < mm->gpa + mm->len) { 1127318224bbSNeel Natu count = vm_fault_quick_hold_pages(&vm->vmspace->vm_map, 1128318224bbSNeel Natu trunc_page(gpa), PAGE_SIZE, reqprot, &m, 1); 11299b1aa8d6SNeel Natu break; 11309b1aa8d6SNeel Natu } 11319b1aa8d6SNeel Natu } 1132318224bbSNeel Natu 1133318224bbSNeel Natu if (count == 1) { 1134318224bbSNeel Natu *cookie = m; 1135318224bbSNeel Natu return ((void *)(PHYS_TO_DMAP(VM_PAGE_TO_PHYS(m)) + pageoff)); 1136318224bbSNeel Natu } else { 1137318224bbSNeel Natu *cookie = NULL; 1138318224bbSNeel Natu return (NULL); 1139318224bbSNeel Natu } 1140318224bbSNeel Natu } 1141318224bbSNeel Natu 114228b561adSJohn Baldwin void * 1143d3956e46SJohn Baldwin vm_gpa_hold(struct vcpu *vcpu, vm_paddr_t gpa, size_t len, int reqprot, 114428b561adSJohn Baldwin void **cookie) 114528b561adSJohn Baldwin { 114628b561adSJohn Baldwin #ifdef INVARIANTS 114728b561adSJohn Baldwin /* 114828b561adSJohn Baldwin * The current vcpu should be frozen to ensure 'vm_memmap[]' 114928b561adSJohn Baldwin * stability. 115028b561adSJohn Baldwin */ 1151d3956e46SJohn Baldwin int state = vcpu_get_state(vcpu, NULL); 115228b561adSJohn Baldwin KASSERT(state == VCPU_FROZEN, ("%s: invalid vcpu state %d", 115328b561adSJohn Baldwin __func__, state)); 115428b561adSJohn Baldwin #endif 1155d3956e46SJohn Baldwin return (_vm_gpa_hold(vcpu->vm, gpa, len, reqprot, cookie)); 115628b561adSJohn Baldwin } 115728b561adSJohn Baldwin 115828b561adSJohn Baldwin void * 115928b561adSJohn Baldwin vm_gpa_hold_global(struct vm *vm, vm_paddr_t gpa, size_t len, int reqprot, 116028b561adSJohn Baldwin void **cookie) 116128b561adSJohn Baldwin { 116267b69e76SJohn Baldwin sx_assert(&vm->mem_segs_lock, SX_LOCKED); 116328b561adSJohn Baldwin return (_vm_gpa_hold(vm, gpa, len, reqprot, cookie)); 116428b561adSJohn Baldwin } 116528b561adSJohn Baldwin 1166318224bbSNeel Natu void 1167318224bbSNeel Natu vm_gpa_release(void *cookie) 1168318224bbSNeel Natu { 1169318224bbSNeel Natu vm_page_t m = cookie; 1170318224bbSNeel Natu 1171eeacb3b0SMark Johnston vm_page_unwire(m, PQ_ACTIVE); 1172366f6083SPeter Grehan } 1173366f6083SPeter Grehan 1174366f6083SPeter Grehan int 1175d3956e46SJohn Baldwin vm_get_register(struct vcpu *vcpu, int reg, uint64_t *retval) 1176366f6083SPeter Grehan { 1177366f6083SPeter Grehan 1178366f6083SPeter Grehan if (reg >= VM_REG_LAST) 1179366f6083SPeter Grehan return (EINVAL); 1180366f6083SPeter Grehan 1181d3956e46SJohn Baldwin return (vmmops_getreg(vcpu->cookie, reg, retval)); 1182366f6083SPeter Grehan } 1183366f6083SPeter Grehan 1184366f6083SPeter Grehan int 1185d3956e46SJohn Baldwin vm_set_register(struct vcpu *vcpu, int reg, uint64_t val) 1186366f6083SPeter Grehan { 1187d087a399SNeel Natu int error; 1188366f6083SPeter Grehan 1189366f6083SPeter Grehan if (reg >= VM_REG_LAST) 1190366f6083SPeter Grehan return (EINVAL); 1191366f6083SPeter Grehan 1192869c8d19SJohn Baldwin error = vmmops_setreg(vcpu->cookie, reg, val); 1193d087a399SNeel Natu if (error || reg != VM_REG_GUEST_RIP) 1194d087a399SNeel Natu return (error); 1195d087a399SNeel Natu 1196d087a399SNeel Natu /* Set 'nextrip' to match the value of %rip */ 1197d3956e46SJohn Baldwin VMM_CTR1(vcpu, "Setting nextrip to %#lx", val); 1198d087a399SNeel Natu vcpu->nextrip = val; 1199d087a399SNeel Natu return (0); 1200366f6083SPeter Grehan } 1201366f6083SPeter Grehan 1202490d56c5SEd Maste static bool 1203366f6083SPeter Grehan is_descriptor_table(int reg) 1204366f6083SPeter Grehan { 1205366f6083SPeter Grehan 1206366f6083SPeter Grehan switch (reg) { 1207366f6083SPeter Grehan case VM_REG_GUEST_IDTR: 1208366f6083SPeter Grehan case VM_REG_GUEST_GDTR: 1209490d56c5SEd Maste return (true); 1210366f6083SPeter Grehan default: 1211490d56c5SEd Maste return (false); 1212366f6083SPeter Grehan } 1213366f6083SPeter Grehan } 1214366f6083SPeter Grehan 1215490d56c5SEd Maste static bool 1216366f6083SPeter Grehan is_segment_register(int reg) 1217366f6083SPeter Grehan { 1218366f6083SPeter Grehan 1219366f6083SPeter Grehan switch (reg) { 1220366f6083SPeter Grehan case VM_REG_GUEST_ES: 1221366f6083SPeter Grehan case VM_REG_GUEST_CS: 1222366f6083SPeter Grehan case VM_REG_GUEST_SS: 1223366f6083SPeter Grehan case VM_REG_GUEST_DS: 1224366f6083SPeter Grehan case VM_REG_GUEST_FS: 1225366f6083SPeter Grehan case VM_REG_GUEST_GS: 1226366f6083SPeter Grehan case VM_REG_GUEST_TR: 1227366f6083SPeter Grehan case VM_REG_GUEST_LDTR: 1228490d56c5SEd Maste return (true); 1229366f6083SPeter Grehan default: 1230490d56c5SEd Maste return (false); 1231366f6083SPeter Grehan } 1232366f6083SPeter Grehan } 1233366f6083SPeter Grehan 1234366f6083SPeter Grehan int 1235d3956e46SJohn Baldwin vm_get_seg_desc(struct vcpu *vcpu, int reg, struct seg_desc *desc) 1236366f6083SPeter Grehan { 1237366f6083SPeter Grehan 1238366f6083SPeter Grehan if (!is_segment_register(reg) && !is_descriptor_table(reg)) 1239366f6083SPeter Grehan return (EINVAL); 1240366f6083SPeter Grehan 1241d3956e46SJohn Baldwin return (vmmops_getdesc(vcpu->cookie, reg, desc)); 1242366f6083SPeter Grehan } 1243366f6083SPeter Grehan 1244366f6083SPeter Grehan int 12453f0f4b15SJohn Baldwin vm_set_seg_desc(struct vcpu *vcpu, int reg, struct seg_desc *desc) 1246366f6083SPeter Grehan { 1247366f6083SPeter Grehan 1248366f6083SPeter Grehan if (!is_segment_register(reg) && !is_descriptor_table(reg)) 1249366f6083SPeter Grehan return (EINVAL); 1250366f6083SPeter Grehan 12513f0f4b15SJohn Baldwin return (vmmops_setdesc(vcpu->cookie, reg, desc)); 1252366f6083SPeter Grehan } 1253366f6083SPeter Grehan 1254366f6083SPeter Grehan static void 1255366f6083SPeter Grehan restore_guest_fpustate(struct vcpu *vcpu) 1256366f6083SPeter Grehan { 1257366f6083SPeter Grehan 125838f1b189SPeter Grehan /* flush host state to the pcb */ 125938f1b189SPeter Grehan fpuexit(curthread); 1260bd8572e0SNeel Natu 1261bd8572e0SNeel Natu /* restore guest FPU state */ 1262366f6083SPeter Grehan fpu_stop_emulating(); 126338f1b189SPeter Grehan fpurestore(vcpu->guestfpu); 1264bd8572e0SNeel Natu 1265abb023fbSJohn Baldwin /* restore guest XCR0 if XSAVE is enabled in the host */ 1266abb023fbSJohn Baldwin if (rcr4() & CR4_XSAVE) 1267abb023fbSJohn Baldwin load_xcr(0, vcpu->guest_xcr0); 1268abb023fbSJohn Baldwin 1269bd8572e0SNeel Natu /* 1270bd8572e0SNeel Natu * The FPU is now "dirty" with the guest's state so turn on emulation 1271bd8572e0SNeel Natu * to trap any access to the FPU by the host. 1272bd8572e0SNeel Natu */ 1273bd8572e0SNeel Natu fpu_start_emulating(); 1274366f6083SPeter Grehan } 1275366f6083SPeter Grehan 1276366f6083SPeter Grehan static void 1277366f6083SPeter Grehan save_guest_fpustate(struct vcpu *vcpu) 1278366f6083SPeter Grehan { 1279366f6083SPeter Grehan 1280bd8572e0SNeel Natu if ((rcr0() & CR0_TS) == 0) 1281bd8572e0SNeel Natu panic("fpu emulation not enabled in host!"); 1282bd8572e0SNeel Natu 1283abb023fbSJohn Baldwin /* save guest XCR0 and restore host XCR0 */ 1284abb023fbSJohn Baldwin if (rcr4() & CR4_XSAVE) { 1285abb023fbSJohn Baldwin vcpu->guest_xcr0 = rxcr(0); 1286abb023fbSJohn Baldwin load_xcr(0, vmm_get_host_xcr0()); 1287abb023fbSJohn Baldwin } 1288abb023fbSJohn Baldwin 1289bd8572e0SNeel Natu /* save guest FPU state */ 1290bd8572e0SNeel Natu fpu_stop_emulating(); 129138f1b189SPeter Grehan fpusave(vcpu->guestfpu); 1292366f6083SPeter Grehan fpu_start_emulating(); 1293366f6083SPeter Grehan } 1294366f6083SPeter Grehan 129561592433SNeel Natu static VMM_STAT(VCPU_IDLE_TICKS, "number of ticks vcpu was idle"); 1296f76fc5d4SNeel Natu 1297318224bbSNeel Natu static int 12983f0f4b15SJohn Baldwin vcpu_set_state_locked(struct vcpu *vcpu, enum vcpu_state newstate, 1299f80330a8SNeel Natu bool from_idle) 1300366f6083SPeter Grehan { 1301318224bbSNeel Natu int error; 1302366f6083SPeter Grehan 1303318224bbSNeel Natu vcpu_assert_locked(vcpu); 1304366f6083SPeter Grehan 1305f76fc5d4SNeel Natu /* 1306f80330a8SNeel Natu * State transitions from the vmmdev_ioctl() must always begin from 1307f80330a8SNeel Natu * the VCPU_IDLE state. This guarantees that there is only a single 1308f80330a8SNeel Natu * ioctl() operating on a vcpu at any point. 1309f80330a8SNeel Natu */ 1310f80330a8SNeel Natu if (from_idle) { 1311248e6799SNeel Natu while (vcpu->state != VCPU_IDLE) { 1312248e6799SNeel Natu vcpu->reqidle = 1; 1313248e6799SNeel Natu vcpu_notify_event_locked(vcpu, false); 13143f0f4b15SJohn Baldwin VMM_CTR1(vcpu, "vcpu state change from %s to " 1315248e6799SNeel Natu "idle requested", vcpu_state2str(vcpu->state)); 1316f80330a8SNeel Natu msleep_spin(&vcpu->state, &vcpu->mtx, "vmstat", hz); 1317248e6799SNeel Natu } 1318f80330a8SNeel Natu } else { 1319f80330a8SNeel Natu KASSERT(vcpu->state != VCPU_IDLE, ("invalid transition from " 1320f80330a8SNeel Natu "vcpu idle state")); 1321f80330a8SNeel Natu } 1322f80330a8SNeel Natu 1323ef39d7e9SNeel Natu if (vcpu->state == VCPU_RUNNING) { 1324ef39d7e9SNeel Natu KASSERT(vcpu->hostcpu == curcpu, ("curcpu %d and hostcpu %d " 1325ef39d7e9SNeel Natu "mismatch for running vcpu", curcpu, vcpu->hostcpu)); 1326ef39d7e9SNeel Natu } else { 1327ef39d7e9SNeel Natu KASSERT(vcpu->hostcpu == NOCPU, ("Invalid hostcpu %d for a " 1328ef39d7e9SNeel Natu "vcpu that is not running", vcpu->hostcpu)); 1329ef39d7e9SNeel Natu } 1330ef39d7e9SNeel Natu 1331f80330a8SNeel Natu /* 1332318224bbSNeel Natu * The following state transitions are allowed: 1333318224bbSNeel Natu * IDLE -> FROZEN -> IDLE 1334318224bbSNeel Natu * FROZEN -> RUNNING -> FROZEN 1335318224bbSNeel Natu * FROZEN -> SLEEPING -> FROZEN 1336f76fc5d4SNeel Natu */ 1337318224bbSNeel Natu switch (vcpu->state) { 1338318224bbSNeel Natu case VCPU_IDLE: 1339318224bbSNeel Natu case VCPU_RUNNING: 1340318224bbSNeel Natu case VCPU_SLEEPING: 1341318224bbSNeel Natu error = (newstate != VCPU_FROZEN); 1342318224bbSNeel Natu break; 1343318224bbSNeel Natu case VCPU_FROZEN: 1344318224bbSNeel Natu error = (newstate == VCPU_FROZEN); 1345318224bbSNeel Natu break; 1346318224bbSNeel Natu default: 1347318224bbSNeel Natu error = 1; 1348318224bbSNeel Natu break; 1349318224bbSNeel Natu } 1350318224bbSNeel Natu 1351f80330a8SNeel Natu if (error) 1352f80330a8SNeel Natu return (EBUSY); 1353318224bbSNeel Natu 13543f0f4b15SJohn Baldwin VMM_CTR2(vcpu, "vcpu state changed from %s to %s", 1355248e6799SNeel Natu vcpu_state2str(vcpu->state), vcpu_state2str(newstate)); 1356248e6799SNeel Natu 1357f80330a8SNeel Natu vcpu->state = newstate; 1358ef39d7e9SNeel Natu if (newstate == VCPU_RUNNING) 1359ef39d7e9SNeel Natu vcpu->hostcpu = curcpu; 1360ef39d7e9SNeel Natu else 1361ef39d7e9SNeel Natu vcpu->hostcpu = NOCPU; 1362ef39d7e9SNeel Natu 1363f80330a8SNeel Natu if (newstate == VCPU_IDLE) 1364f80330a8SNeel Natu wakeup(&vcpu->state); 1365f80330a8SNeel Natu 1366f80330a8SNeel Natu return (0); 1367318224bbSNeel Natu } 1368318224bbSNeel Natu 1369318224bbSNeel Natu static void 13703f0f4b15SJohn Baldwin vcpu_require_state(struct vcpu *vcpu, enum vcpu_state newstate) 1371318224bbSNeel Natu { 1372318224bbSNeel Natu int error; 1373318224bbSNeel Natu 13743f0f4b15SJohn Baldwin if ((error = vcpu_set_state(vcpu, newstate, false)) != 0) 1375318224bbSNeel Natu panic("Error %d setting state to %d\n", error, newstate); 1376318224bbSNeel Natu } 1377318224bbSNeel Natu 1378318224bbSNeel Natu static void 13793f0f4b15SJohn Baldwin vcpu_require_state_locked(struct vcpu *vcpu, enum vcpu_state newstate) 1380318224bbSNeel Natu { 1381318224bbSNeel Natu int error; 1382318224bbSNeel Natu 13833f0f4b15SJohn Baldwin if ((error = vcpu_set_state_locked(vcpu, newstate, false)) != 0) 1384318224bbSNeel Natu panic("Error %d setting state to %d", error, newstate); 1385318224bbSNeel Natu } 1386318224bbSNeel Natu 1387b837daddSKonstantin Belousov static int 1388d8be3d52SJohn Baldwin vm_handle_rendezvous(struct vcpu *vcpu) 13895b8a8cd1SNeel Natu { 1390d8be3d52SJohn Baldwin struct vm *vm = vcpu->vm; 1391b837daddSKonstantin Belousov struct thread *td; 1392d8be3d52SJohn Baldwin int error, vcpuid; 13935b8a8cd1SNeel Natu 1394b837daddSKonstantin Belousov error = 0; 1395d8be3d52SJohn Baldwin vcpuid = vcpu->vcpuid; 1396b837daddSKonstantin Belousov td = curthread; 13975b8a8cd1SNeel Natu mtx_lock(&vm->rendezvous_mtx); 13985b8a8cd1SNeel Natu while (vm->rendezvous_func != NULL) { 139922d822c6SNeel Natu /* 'rendezvous_req_cpus' must be a subset of 'active_cpus' */ 1400e2650af1SStefan Eßer CPU_AND(&vm->rendezvous_req_cpus, &vm->rendezvous_req_cpus, &vm->active_cpus); 140122d822c6SNeel Natu 1402949f0f47SJohn Baldwin if (CPU_ISSET(vcpuid, &vm->rendezvous_req_cpus) && 140322d822c6SNeel Natu !CPU_ISSET(vcpuid, &vm->rendezvous_done_cpus)) { 1404d8be3d52SJohn Baldwin VMM_CTR0(vcpu, "Calling rendezvous func"); 1405d8be3d52SJohn Baldwin (*vm->rendezvous_func)(vcpu, vm->rendezvous_arg); 14065b8a8cd1SNeel Natu CPU_SET(vcpuid, &vm->rendezvous_done_cpus); 14075b8a8cd1SNeel Natu } 14085b8a8cd1SNeel Natu if (CPU_CMP(&vm->rendezvous_req_cpus, 14095b8a8cd1SNeel Natu &vm->rendezvous_done_cpus) == 0) { 1410d8be3d52SJohn Baldwin VMM_CTR0(vcpu, "Rendezvous completed"); 1411869dbab7SAndriy Gapon vm->rendezvous_func = NULL; 14125b8a8cd1SNeel Natu wakeup(&vm->rendezvous_func); 14135b8a8cd1SNeel Natu break; 14145b8a8cd1SNeel Natu } 1415d8be3d52SJohn Baldwin VMM_CTR0(vcpu, "Wait for rendezvous completion"); 14165b8a8cd1SNeel Natu mtx_sleep(&vm->rendezvous_func, &vm->rendezvous_mtx, 0, 1417b837daddSKonstantin Belousov "vmrndv", hz); 1418c6d31b83SKonstantin Belousov if (td_ast_pending(td, TDA_SUSPEND)) { 1419b837daddSKonstantin Belousov mtx_unlock(&vm->rendezvous_mtx); 1420b837daddSKonstantin Belousov error = thread_check_susp(td, true); 1421b837daddSKonstantin Belousov if (error != 0) 1422b837daddSKonstantin Belousov return (error); 1423b837daddSKonstantin Belousov mtx_lock(&vm->rendezvous_mtx); 1424b837daddSKonstantin Belousov } 14255b8a8cd1SNeel Natu } 14265b8a8cd1SNeel Natu mtx_unlock(&vm->rendezvous_mtx); 1427b837daddSKonstantin Belousov return (0); 14285b8a8cd1SNeel Natu } 14295b8a8cd1SNeel Natu 1430318224bbSNeel Natu /* 1431318224bbSNeel Natu * Emulate a guest 'hlt' by sleeping until the vcpu is ready to run. 1432318224bbSNeel Natu */ 1433318224bbSNeel Natu static int 14343f0f4b15SJohn Baldwin vm_handle_hlt(struct vcpu *vcpu, bool intr_disabled, bool *retu) 1435318224bbSNeel Natu { 14363f0f4b15SJohn Baldwin struct vm *vm = vcpu->vm; 1437c6a0cc2eSNeel Natu const char *wmesg; 1438b837daddSKonstantin Belousov struct thread *td; 14393f0f4b15SJohn Baldwin int error, t, vcpuid, vcpu_halted, vm_halted; 1440e50ce2aaSNeel Natu 14413f0f4b15SJohn Baldwin vcpuid = vcpu->vcpuid; 1442e50ce2aaSNeel Natu vcpu_halted = 0; 1443e50ce2aaSNeel Natu vm_halted = 0; 1444b837daddSKonstantin Belousov error = 0; 1445b837daddSKonstantin Belousov td = curthread; 1446318224bbSNeel Natu 14473f0f4b15SJohn Baldwin KASSERT(!CPU_ISSET(vcpuid, &vm->halted_cpus), ("vcpu already halted")); 14483f0f4b15SJohn Baldwin 1449f76fc5d4SNeel Natu vcpu_lock(vcpu); 1450c6a0cc2eSNeel Natu while (1) { 1451f76fc5d4SNeel Natu /* 1452f76fc5d4SNeel Natu * Do a final check for pending NMI or interrupts before 1453c6a0cc2eSNeel Natu * really putting this thread to sleep. Also check for 1454c6a0cc2eSNeel Natu * software events that would cause this vcpu to wakeup. 1455f76fc5d4SNeel Natu * 1456c6a0cc2eSNeel Natu * These interrupts/events could have happened after the 145715add60dSPeter Grehan * vcpu returned from vmmops_run() and before it acquired the 1458c6a0cc2eSNeel Natu * vcpu lock above. 1459f76fc5d4SNeel Natu */ 1460248e6799SNeel Natu if (vm->rendezvous_func != NULL || vm->suspend || vcpu->reqidle) 1461c6a0cc2eSNeel Natu break; 146280cb5d84SJohn Baldwin if (vm_nmi_pending(vcpu)) 1463c6a0cc2eSNeel Natu break; 1464c6a0cc2eSNeel Natu if (!intr_disabled) { 146580cb5d84SJohn Baldwin if (vm_extint_pending(vcpu) || 1466c6a0cc2eSNeel Natu vlapic_pending_intr(vcpu->vlapic, NULL)) { 1467c6a0cc2eSNeel Natu break; 1468c6a0cc2eSNeel Natu } 1469c6a0cc2eSNeel Natu } 1470c6a0cc2eSNeel Natu 1471f008d157SNeel Natu /* Don't go to sleep if the vcpu thread needs to yield */ 147280cb5d84SJohn Baldwin if (vcpu_should_yield(vcpu)) 1473f008d157SNeel Natu break; 1474f008d157SNeel Natu 147580cb5d84SJohn Baldwin if (vcpu_debugged(vcpu)) 1476fc276d92SJohn Baldwin break; 1477fc276d92SJohn Baldwin 1478e50ce2aaSNeel Natu /* 1479e50ce2aaSNeel Natu * Some Linux guests implement "halt" by having all vcpus 1480e50ce2aaSNeel Natu * execute HLT with interrupts disabled. 'halted_cpus' keeps 1481e50ce2aaSNeel Natu * track of the vcpus that have entered this state. When all 1482e50ce2aaSNeel Natu * vcpus enter the halted state the virtual machine is halted. 1483e50ce2aaSNeel Natu */ 1484e50ce2aaSNeel Natu if (intr_disabled) { 1485c6a0cc2eSNeel Natu wmesg = "vmhalt"; 14863f0f4b15SJohn Baldwin VMM_CTR0(vcpu, "Halted"); 1487055fc2cbSNeel Natu if (!vcpu_halted && halt_detection_enabled) { 1488e50ce2aaSNeel Natu vcpu_halted = 1; 1489e50ce2aaSNeel Natu CPU_SET_ATOMIC(vcpuid, &vm->halted_cpus); 1490e50ce2aaSNeel Natu } 1491e50ce2aaSNeel Natu if (CPU_CMP(&vm->halted_cpus, &vm->active_cpus) == 0) { 1492e50ce2aaSNeel Natu vm_halted = 1; 1493e50ce2aaSNeel Natu break; 1494e50ce2aaSNeel Natu } 1495e50ce2aaSNeel Natu } else { 1496e50ce2aaSNeel Natu wmesg = "vmidle"; 1497e50ce2aaSNeel Natu } 1498c6a0cc2eSNeel Natu 1499f76fc5d4SNeel Natu t = ticks; 15003f0f4b15SJohn Baldwin vcpu_require_state_locked(vcpu, VCPU_SLEEPING); 1501f008d157SNeel Natu /* 1502f008d157SNeel Natu * XXX msleep_spin() cannot be interrupted by signals so 1503f008d157SNeel Natu * wake up periodically to check pending signals. 1504f008d157SNeel Natu */ 1505f008d157SNeel Natu msleep_spin(vcpu, &vcpu->mtx, wmesg, hz); 15063f0f4b15SJohn Baldwin vcpu_require_state_locked(vcpu, VCPU_FROZEN); 15073dc3d32aSJohn Baldwin vmm_stat_incr(vcpu, VCPU_IDLE_TICKS, ticks - t); 1508c6d31b83SKonstantin Belousov if (td_ast_pending(td, TDA_SUSPEND)) { 1509b837daddSKonstantin Belousov vcpu_unlock(vcpu); 1510b837daddSKonstantin Belousov error = thread_check_susp(td, false); 15114d447b30SKonstantin Belousov if (error != 0) { 15124d447b30SKonstantin Belousov if (vcpu_halted) { 15134d447b30SKonstantin Belousov CPU_CLR_ATOMIC(vcpuid, 15144d447b30SKonstantin Belousov &vm->halted_cpus); 15154d447b30SKonstantin Belousov } 1516b837daddSKonstantin Belousov return (error); 15174d447b30SKonstantin Belousov } 1518b837daddSKonstantin Belousov vcpu_lock(vcpu); 1519b837daddSKonstantin Belousov } 1520f76fc5d4SNeel Natu } 1521e50ce2aaSNeel Natu 1522e50ce2aaSNeel Natu if (vcpu_halted) 1523e50ce2aaSNeel Natu CPU_CLR_ATOMIC(vcpuid, &vm->halted_cpus); 1524e50ce2aaSNeel Natu 1525f76fc5d4SNeel Natu vcpu_unlock(vcpu); 1526f76fc5d4SNeel Natu 1527e50ce2aaSNeel Natu if (vm_halted) 1528e50ce2aaSNeel Natu vm_suspend(vm, VM_SUSPEND_HALT); 1529e50ce2aaSNeel Natu 1530318224bbSNeel Natu return (0); 1531318224bbSNeel Natu } 1532318224bbSNeel Natu 1533318224bbSNeel Natu static int 15343f0f4b15SJohn Baldwin vm_handle_paging(struct vcpu *vcpu, bool *retu) 1535318224bbSNeel Natu { 15363f0f4b15SJohn Baldwin struct vm *vm = vcpu->vm; 1537318224bbSNeel Natu int rv, ftype; 1538318224bbSNeel Natu struct vm_map *map; 1539318224bbSNeel Natu struct vm_exit *vme; 1540318224bbSNeel Natu 1541318224bbSNeel Natu vme = &vcpu->exitinfo; 1542318224bbSNeel Natu 1543d087a399SNeel Natu KASSERT(vme->inst_length == 0, ("%s: invalid inst_length %d", 1544d087a399SNeel Natu __func__, vme->inst_length)); 1545d087a399SNeel Natu 1546318224bbSNeel Natu ftype = vme->u.paging.fault_type; 1547318224bbSNeel Natu KASSERT(ftype == VM_PROT_READ || 1548318224bbSNeel Natu ftype == VM_PROT_WRITE || ftype == VM_PROT_EXECUTE, 1549318224bbSNeel Natu ("vm_handle_paging: invalid fault_type %d", ftype)); 1550318224bbSNeel Natu 1551318224bbSNeel Natu if (ftype == VM_PROT_READ || ftype == VM_PROT_WRITE) { 1552318224bbSNeel Natu rv = pmap_emulate_accessed_dirty(vmspace_pmap(vm->vmspace), 1553318224bbSNeel Natu vme->u.paging.gpa, ftype); 15549d8d8e3eSNeel Natu if (rv == 0) { 15553f0f4b15SJohn Baldwin VMM_CTR2(vcpu, "%s bit emulation for gpa %#lx", 15569d8d8e3eSNeel Natu ftype == VM_PROT_READ ? "accessed" : "dirty", 15579d8d8e3eSNeel Natu vme->u.paging.gpa); 1558318224bbSNeel Natu goto done; 1559318224bbSNeel Natu } 15609d8d8e3eSNeel Natu } 1561318224bbSNeel Natu 1562318224bbSNeel Natu map = &vm->vmspace->vm_map; 1563df08823dSKonstantin Belousov rv = vm_fault(map, vme->u.paging.gpa, ftype, VM_FAULT_NORMAL, NULL); 1564318224bbSNeel Natu 15653f0f4b15SJohn Baldwin VMM_CTR3(vcpu, "vm_handle_paging rv = %d, gpa = %#lx, " 1566513c8d33SNeel Natu "ftype = %d", rv, vme->u.paging.gpa, ftype); 1567318224bbSNeel Natu 1568318224bbSNeel Natu if (rv != KERN_SUCCESS) 1569318224bbSNeel Natu return (EFAULT); 1570318224bbSNeel Natu done: 1571318224bbSNeel Natu return (0); 1572318224bbSNeel Natu } 1573318224bbSNeel Natu 1574318224bbSNeel Natu static int 15753f0f4b15SJohn Baldwin vm_handle_inst_emul(struct vcpu *vcpu, bool *retu) 1576318224bbSNeel Natu { 1577318224bbSNeel Natu struct vie *vie; 1578318224bbSNeel Natu struct vm_exit *vme; 1579e4f605eeSTycho Nightingale uint64_t gla, gpa, cs_base; 1580e813a873SNeel Natu struct vm_guest_paging *paging; 1581565bbb86SNeel Natu mem_region_read_t mread; 1582565bbb86SNeel Natu mem_region_write_t mwrite; 1583f7a9f178SNeel Natu enum vm_cpu_mode cpu_mode; 15841c73ea3eSNeel Natu int cs_d, error, fault; 1585318224bbSNeel Natu 1586318224bbSNeel Natu vme = &vcpu->exitinfo; 1587318224bbSNeel Natu 15881c73ea3eSNeel Natu KASSERT(vme->inst_length == 0, ("%s: invalid inst_length %d", 15891c73ea3eSNeel Natu __func__, vme->inst_length)); 15901c73ea3eSNeel Natu 1591318224bbSNeel Natu gla = vme->u.inst_emul.gla; 1592318224bbSNeel Natu gpa = vme->u.inst_emul.gpa; 1593e4f605eeSTycho Nightingale cs_base = vme->u.inst_emul.cs_base; 1594f7a9f178SNeel Natu cs_d = vme->u.inst_emul.cs_d; 1595318224bbSNeel Natu vie = &vme->u.inst_emul.vie; 1596e813a873SNeel Natu paging = &vme->u.inst_emul.paging; 1597f7a9f178SNeel Natu cpu_mode = paging->cpu_mode; 1598318224bbSNeel Natu 15993f0f4b15SJohn Baldwin VMM_CTR1(vcpu, "inst_emul fault accessing gpa %#lx", gpa); 16009d8d8e3eSNeel Natu 1601318224bbSNeel Natu /* Fetch, decode and emulate the faulting instruction */ 1602c2a875f9SNeel Natu if (vie->num_valid == 0) { 1603d3956e46SJohn Baldwin error = vmm_fetch_instruction(vcpu, paging, vme->rip + cs_base, 1604d3956e46SJohn Baldwin VIE_INST_SIZE, vie, &fault); 1605c2a875f9SNeel Natu } else { 1606c2a875f9SNeel Natu /* 1607c2a875f9SNeel Natu * The instruction bytes have already been copied into 'vie' 1608c2a875f9SNeel Natu */ 16099c4d5478SNeel Natu error = fault = 0; 1610c2a875f9SNeel Natu } 16119c4d5478SNeel Natu if (error || fault) 16129c4d5478SNeel Natu return (error); 1613318224bbSNeel Natu 1614d3956e46SJohn Baldwin if (vmm_decode_instruction(vcpu, gla, cpu_mode, cs_d, vie) != 0) { 16153f0f4b15SJohn Baldwin VMM_CTR1(vcpu, "Error decoding instruction at %#lx", 1616c07a0648SNeel Natu vme->rip + cs_base); 1617c07a0648SNeel Natu *retu = true; /* dump instruction bytes in userspace */ 1618c07a0648SNeel Natu return (0); 1619c07a0648SNeel Natu } 1620318224bbSNeel Natu 1621a0b78f09SPeter Grehan /* 16221c73ea3eSNeel Natu * Update 'nextrip' based on the length of the emulated instruction. 1623a0b78f09SPeter Grehan */ 1624a0b78f09SPeter Grehan vme->inst_length = vie->num_processed; 1625d087a399SNeel Natu vcpu->nextrip += vie->num_processed; 16263f0f4b15SJohn Baldwin VMM_CTR1(vcpu, "nextrip updated to %#lx after instruction decoding", 16273f0f4b15SJohn Baldwin vcpu->nextrip); 1628a0b78f09SPeter Grehan 162908e3ff32SNeel Natu /* return to userland unless this is an in-kernel emulated device */ 1630565bbb86SNeel Natu if (gpa >= DEFAULT_APIC_BASE && gpa < DEFAULT_APIC_BASE + PAGE_SIZE) { 1631565bbb86SNeel Natu mread = lapic_mmio_read; 1632565bbb86SNeel Natu mwrite = lapic_mmio_write; 1633565bbb86SNeel Natu } else if (gpa >= VIOAPIC_BASE && gpa < VIOAPIC_BASE + VIOAPIC_SIZE) { 1634565bbb86SNeel Natu mread = vioapic_mmio_read; 1635565bbb86SNeel Natu mwrite = vioapic_mmio_write; 163608e3ff32SNeel Natu } else if (gpa >= VHPET_BASE && gpa < VHPET_BASE + VHPET_SIZE) { 163708e3ff32SNeel Natu mread = vhpet_mmio_read; 163808e3ff32SNeel Natu mwrite = vhpet_mmio_write; 1639565bbb86SNeel Natu } else { 1640becd9849SNeel Natu *retu = true; 1641318224bbSNeel Natu return (0); 1642318224bbSNeel Natu } 1643318224bbSNeel Natu 1644d3956e46SJohn Baldwin error = vmm_emulate_instruction(vcpu, gpa, vie, paging, mread, mwrite, 1645d3956e46SJohn Baldwin retu); 1646318224bbSNeel Natu 1647318224bbSNeel Natu return (error); 1648318224bbSNeel Natu } 1649318224bbSNeel Natu 1650b15a09c0SNeel Natu static int 16513f0f4b15SJohn Baldwin vm_handle_suspend(struct vcpu *vcpu, bool *retu) 1652b15a09c0SNeel Natu { 16533f0f4b15SJohn Baldwin struct vm *vm = vcpu->vm; 1654b837daddSKonstantin Belousov int error, i; 1655b837daddSKonstantin Belousov struct thread *td; 1656b15a09c0SNeel Natu 1657b837daddSKonstantin Belousov error = 0; 1658b837daddSKonstantin Belousov td = curthread; 1659b15a09c0SNeel Natu 16603f0f4b15SJohn Baldwin CPU_SET_ATOMIC(vcpu->vcpuid, &vm->suspended_cpus); 1661b15a09c0SNeel Natu 1662b15a09c0SNeel Natu /* 1663b15a09c0SNeel Natu * Wait until all 'active_cpus' have suspended themselves. 1664b15a09c0SNeel Natu * 1665b15a09c0SNeel Natu * Since a VM may be suspended at any time including when one or 1666b15a09c0SNeel Natu * more vcpus are doing a rendezvous we need to call the rendezvous 1667b15a09c0SNeel Natu * handler while we are waiting to prevent a deadlock. 1668b15a09c0SNeel Natu */ 1669b15a09c0SNeel Natu vcpu_lock(vcpu); 1670b837daddSKonstantin Belousov while (error == 0) { 1671b15a09c0SNeel Natu if (CPU_CMP(&vm->suspended_cpus, &vm->active_cpus) == 0) { 16723f0f4b15SJohn Baldwin VMM_CTR0(vcpu, "All vcpus suspended"); 1673b15a09c0SNeel Natu break; 1674b15a09c0SNeel Natu } 1675b15a09c0SNeel Natu 1676b15a09c0SNeel Natu if (vm->rendezvous_func == NULL) { 16773f0f4b15SJohn Baldwin VMM_CTR0(vcpu, "Sleeping during suspend"); 16783f0f4b15SJohn Baldwin vcpu_require_state_locked(vcpu, VCPU_SLEEPING); 1679b15a09c0SNeel Natu msleep_spin(vcpu, &vcpu->mtx, "vmsusp", hz); 16803f0f4b15SJohn Baldwin vcpu_require_state_locked(vcpu, VCPU_FROZEN); 1681c6d31b83SKonstantin Belousov if (td_ast_pending(td, TDA_SUSPEND)) { 1682b837daddSKonstantin Belousov vcpu_unlock(vcpu); 1683b837daddSKonstantin Belousov error = thread_check_susp(td, false); 1684b837daddSKonstantin Belousov vcpu_lock(vcpu); 1685b837daddSKonstantin Belousov } 1686b15a09c0SNeel Natu } else { 16873f0f4b15SJohn Baldwin VMM_CTR0(vcpu, "Rendezvous during suspend"); 1688b15a09c0SNeel Natu vcpu_unlock(vcpu); 1689d8be3d52SJohn Baldwin error = vm_handle_rendezvous(vcpu); 1690b15a09c0SNeel Natu vcpu_lock(vcpu); 1691b15a09c0SNeel Natu } 1692b15a09c0SNeel Natu } 1693b15a09c0SNeel Natu vcpu_unlock(vcpu); 1694b15a09c0SNeel Natu 1695b15a09c0SNeel Natu /* 1696b15a09c0SNeel Natu * Wakeup the other sleeping vcpus and return to userspace. 1697b15a09c0SNeel Natu */ 1698a488c9c9SRodney W. Grimes for (i = 0; i < vm->maxcpus; i++) { 1699b15a09c0SNeel Natu if (CPU_ISSET(i, &vm->suspended_cpus)) { 17003f0f4b15SJohn Baldwin vcpu_notify_event(vm_vcpu(vm, i), false); 1701b15a09c0SNeel Natu } 1702b15a09c0SNeel Natu } 1703b15a09c0SNeel Natu 1704b15a09c0SNeel Natu *retu = true; 1705b837daddSKonstantin Belousov return (error); 1706b15a09c0SNeel Natu } 1707b15a09c0SNeel Natu 1708248e6799SNeel Natu static int 17093f0f4b15SJohn Baldwin vm_handle_reqidle(struct vcpu *vcpu, bool *retu) 1710248e6799SNeel Natu { 1711248e6799SNeel Natu vcpu_lock(vcpu); 1712248e6799SNeel Natu KASSERT(vcpu->reqidle, ("invalid vcpu reqidle %d", vcpu->reqidle)); 1713248e6799SNeel Natu vcpu->reqidle = 0; 1714248e6799SNeel Natu vcpu_unlock(vcpu); 1715248e6799SNeel Natu *retu = true; 1716248e6799SNeel Natu return (0); 1717248e6799SNeel Natu } 1718248e6799SNeel Natu 1719b15a09c0SNeel Natu int 1720f0fdcfe2SNeel Natu vm_suspend(struct vm *vm, enum vm_suspend_how how) 1721b15a09c0SNeel Natu { 1722f0fdcfe2SNeel Natu int i; 1723b15a09c0SNeel Natu 1724f0fdcfe2SNeel Natu if (how <= VM_SUSPEND_NONE || how >= VM_SUSPEND_LAST) 1725f0fdcfe2SNeel Natu return (EINVAL); 1726f0fdcfe2SNeel Natu 1727f0fdcfe2SNeel Natu if (atomic_cmpset_int(&vm->suspend, 0, how) == 0) { 1728f0fdcfe2SNeel Natu VM_CTR2(vm, "virtual machine already suspended %d/%d", 1729f0fdcfe2SNeel Natu vm->suspend, how); 1730b15a09c0SNeel Natu return (EALREADY); 1731b15a09c0SNeel Natu } 1732f0fdcfe2SNeel Natu 1733f0fdcfe2SNeel Natu VM_CTR1(vm, "virtual machine successfully suspended %d", how); 1734f0fdcfe2SNeel Natu 1735f0fdcfe2SNeel Natu /* 1736f0fdcfe2SNeel Natu * Notify all active vcpus that they are now suspended. 1737f0fdcfe2SNeel Natu */ 1738a488c9c9SRodney W. Grimes for (i = 0; i < vm->maxcpus; i++) { 1739f0fdcfe2SNeel Natu if (CPU_ISSET(i, &vm->active_cpus)) 17403f0f4b15SJohn Baldwin vcpu_notify_event(vm_vcpu(vm, i), false); 1741f0fdcfe2SNeel Natu } 1742f0fdcfe2SNeel Natu 1743f0fdcfe2SNeel Natu return (0); 1744f0fdcfe2SNeel Natu } 1745f0fdcfe2SNeel Natu 1746f0fdcfe2SNeel Natu void 174780cb5d84SJohn Baldwin vm_exit_suspended(struct vcpu *vcpu, uint64_t rip) 1748f0fdcfe2SNeel Natu { 174980cb5d84SJohn Baldwin struct vm *vm = vcpu->vm; 1750f0fdcfe2SNeel Natu struct vm_exit *vmexit; 1751f0fdcfe2SNeel Natu 1752f0fdcfe2SNeel Natu KASSERT(vm->suspend > VM_SUSPEND_NONE && vm->suspend < VM_SUSPEND_LAST, 1753f0fdcfe2SNeel Natu ("vm_exit_suspended: invalid suspend type %d", vm->suspend)); 1754f0fdcfe2SNeel Natu 175580cb5d84SJohn Baldwin vmexit = vm_exitinfo(vcpu); 1756f0fdcfe2SNeel Natu vmexit->rip = rip; 1757f0fdcfe2SNeel Natu vmexit->inst_length = 0; 1758f0fdcfe2SNeel Natu vmexit->exitcode = VM_EXITCODE_SUSPENDED; 1759f0fdcfe2SNeel Natu vmexit->u.suspended.how = vm->suspend; 1760b15a09c0SNeel Natu } 1761b15a09c0SNeel Natu 176240487465SNeel Natu void 176380cb5d84SJohn Baldwin vm_exit_debug(struct vcpu *vcpu, uint64_t rip) 1764fc276d92SJohn Baldwin { 1765fc276d92SJohn Baldwin struct vm_exit *vmexit; 1766fc276d92SJohn Baldwin 176780cb5d84SJohn Baldwin vmexit = vm_exitinfo(vcpu); 1768fc276d92SJohn Baldwin vmexit->rip = rip; 1769fc276d92SJohn Baldwin vmexit->inst_length = 0; 1770fc276d92SJohn Baldwin vmexit->exitcode = VM_EXITCODE_DEBUG; 1771fc276d92SJohn Baldwin } 1772fc276d92SJohn Baldwin 1773fc276d92SJohn Baldwin void 177480cb5d84SJohn Baldwin vm_exit_rendezvous(struct vcpu *vcpu, uint64_t rip) 177540487465SNeel Natu { 177640487465SNeel Natu struct vm_exit *vmexit; 177740487465SNeel Natu 177880cb5d84SJohn Baldwin KASSERT(vcpu->vm->rendezvous_func != NULL, 177980cb5d84SJohn Baldwin ("rendezvous not in progress")); 178040487465SNeel Natu 178180cb5d84SJohn Baldwin vmexit = vm_exitinfo(vcpu); 178240487465SNeel Natu vmexit->rip = rip; 178340487465SNeel Natu vmexit->inst_length = 0; 178440487465SNeel Natu vmexit->exitcode = VM_EXITCODE_RENDEZVOUS; 178580cb5d84SJohn Baldwin vmm_stat_incr(vcpu, VMEXIT_RENDEZVOUS, 1); 178640487465SNeel Natu } 178740487465SNeel Natu 178840487465SNeel Natu void 178980cb5d84SJohn Baldwin vm_exit_reqidle(struct vcpu *vcpu, uint64_t rip) 1790248e6799SNeel Natu { 1791248e6799SNeel Natu struct vm_exit *vmexit; 1792248e6799SNeel Natu 179380cb5d84SJohn Baldwin vmexit = vm_exitinfo(vcpu); 1794248e6799SNeel Natu vmexit->rip = rip; 1795248e6799SNeel Natu vmexit->inst_length = 0; 1796248e6799SNeel Natu vmexit->exitcode = VM_EXITCODE_REQIDLE; 179780cb5d84SJohn Baldwin vmm_stat_incr(vcpu, VMEXIT_REQIDLE, 1); 1798248e6799SNeel Natu } 1799248e6799SNeel Natu 1800248e6799SNeel Natu void 180180cb5d84SJohn Baldwin vm_exit_astpending(struct vcpu *vcpu, uint64_t rip) 180240487465SNeel Natu { 180340487465SNeel Natu struct vm_exit *vmexit; 180440487465SNeel Natu 180580cb5d84SJohn Baldwin vmexit = vm_exitinfo(vcpu); 180640487465SNeel Natu vmexit->rip = rip; 180740487465SNeel Natu vmexit->inst_length = 0; 180840487465SNeel Natu vmexit->exitcode = VM_EXITCODE_BOGUS; 180980cb5d84SJohn Baldwin vmm_stat_incr(vcpu, VMEXIT_ASTPENDING, 1); 181040487465SNeel Natu } 181140487465SNeel Natu 1812318224bbSNeel Natu int 18133f0f4b15SJohn Baldwin vm_run(struct vcpu *vcpu, struct vm_exit *vme_user) 1814318224bbSNeel Natu { 18153f0f4b15SJohn Baldwin struct vm *vm = vcpu->vm; 1816248e6799SNeel Natu struct vm_eventinfo evinfo; 1817318224bbSNeel Natu int error, vcpuid; 1818318224bbSNeel Natu struct pcb *pcb; 1819d087a399SNeel Natu uint64_t tscval; 1820318224bbSNeel Natu struct vm_exit *vme; 1821becd9849SNeel Natu bool retu, intr_disabled; 1822318224bbSNeel Natu pmap_t pmap; 1823318224bbSNeel Natu 18243f0f4b15SJohn Baldwin vcpuid = vcpu->vcpuid; 1825318224bbSNeel Natu 182695ebc360SNeel Natu if (!CPU_ISSET(vcpuid, &vm->active_cpus)) 182795ebc360SNeel Natu return (EINVAL); 182895ebc360SNeel Natu 182995ebc360SNeel Natu if (CPU_ISSET(vcpuid, &vm->suspended_cpus)) 183095ebc360SNeel Natu return (EINVAL); 183195ebc360SNeel Natu 1832318224bbSNeel Natu pmap = vmspace_pmap(vm->vmspace); 1833318224bbSNeel Natu vme = &vcpu->exitinfo; 1834248e6799SNeel Natu evinfo.rptr = &vm->rendezvous_func; 1835248e6799SNeel Natu evinfo.sptr = &vm->suspend; 1836248e6799SNeel Natu evinfo.iptr = &vcpu->reqidle; 1837318224bbSNeel Natu restart: 1838318224bbSNeel Natu critical_enter(); 1839318224bbSNeel Natu 1840318224bbSNeel Natu KASSERT(!CPU_ISSET(curcpu, &pmap->pm_active), 1841318224bbSNeel Natu ("vm_run: absurd pm_active")); 1842318224bbSNeel Natu 1843318224bbSNeel Natu tscval = rdtsc(); 1844318224bbSNeel Natu 1845318224bbSNeel Natu pcb = PCPU_GET(curpcb); 1846318224bbSNeel Natu set_pcb_flags(pcb, PCB_FULL_IRET); 1847318224bbSNeel Natu 1848318224bbSNeel Natu restore_guest_fpustate(vcpu); 1849318224bbSNeel Natu 18503f0f4b15SJohn Baldwin vcpu_require_state(vcpu, VCPU_RUNNING); 1851869c8d19SJohn Baldwin error = vmmops_run(vcpu->cookie, vcpu->nextrip, pmap, &evinfo); 18523f0f4b15SJohn Baldwin vcpu_require_state(vcpu, VCPU_FROZEN); 1853318224bbSNeel Natu 1854318224bbSNeel Natu save_guest_fpustate(vcpu); 1855318224bbSNeel Natu 18563dc3d32aSJohn Baldwin vmm_stat_incr(vcpu, VCPU_TOTAL_RUNTIME, rdtsc() - tscval); 1857318224bbSNeel Natu 1858318224bbSNeel Natu critical_exit(); 1859318224bbSNeel Natu 1860318224bbSNeel Natu if (error == 0) { 1861becd9849SNeel Natu retu = false; 1862d087a399SNeel Natu vcpu->nextrip = vme->rip + vme->inst_length; 1863318224bbSNeel Natu switch (vme->exitcode) { 1864248e6799SNeel Natu case VM_EXITCODE_REQIDLE: 18653f0f4b15SJohn Baldwin error = vm_handle_reqidle(vcpu, &retu); 1866248e6799SNeel Natu break; 1867b15a09c0SNeel Natu case VM_EXITCODE_SUSPENDED: 18683f0f4b15SJohn Baldwin error = vm_handle_suspend(vcpu, &retu); 1869b15a09c0SNeel Natu break; 187030b94db8SNeel Natu case VM_EXITCODE_IOAPIC_EOI: 1871e42c24d5SJohn Baldwin vioapic_process_eoi(vm, vme->u.ioapic_eoi.vector); 187230b94db8SNeel Natu break; 18735b8a8cd1SNeel Natu case VM_EXITCODE_RENDEZVOUS: 1874d8be3d52SJohn Baldwin error = vm_handle_rendezvous(vcpu); 18755b8a8cd1SNeel Natu break; 1876318224bbSNeel Natu case VM_EXITCODE_HLT: 1877becd9849SNeel Natu intr_disabled = ((vme->u.hlt.rflags & PSL_I) == 0); 18783f0f4b15SJohn Baldwin error = vm_handle_hlt(vcpu, intr_disabled, &retu); 1879318224bbSNeel Natu break; 1880318224bbSNeel Natu case VM_EXITCODE_PAGING: 18813f0f4b15SJohn Baldwin error = vm_handle_paging(vcpu, &retu); 1882318224bbSNeel Natu break; 1883318224bbSNeel Natu case VM_EXITCODE_INST_EMUL: 18843f0f4b15SJohn Baldwin error = vm_handle_inst_emul(vcpu, &retu); 1885318224bbSNeel Natu break; 1886d17b5104SNeel Natu case VM_EXITCODE_INOUT: 1887d17b5104SNeel Natu case VM_EXITCODE_INOUT_STR: 18883f0f4b15SJohn Baldwin error = vm_handle_inout(vcpu, vme, &retu); 1889d17b5104SNeel Natu break; 189065145c7fSNeel Natu case VM_EXITCODE_MONITOR: 189165145c7fSNeel Natu case VM_EXITCODE_MWAIT: 189227d26457SAndrew Turner case VM_EXITCODE_VMINSN: 1893d3956e46SJohn Baldwin vm_inject_ud(vcpu); 189465145c7fSNeel Natu break; 1895318224bbSNeel Natu default: 1896becd9849SNeel Natu retu = true; /* handled in userland */ 1897318224bbSNeel Natu break; 1898318224bbSNeel Natu } 1899318224bbSNeel Natu } 1900318224bbSNeel Natu 19010bda8d3eSCorvin Köhne /* 19020bda8d3eSCorvin Köhne * VM_EXITCODE_INST_EMUL could access the apic which could transform the 19030bda8d3eSCorvin Köhne * exit code into VM_EXITCODE_IPI. 19040bda8d3eSCorvin Köhne */ 19050bda8d3eSCorvin Köhne if (error == 0 && vme->exitcode == VM_EXITCODE_IPI) { 19060bda8d3eSCorvin Köhne retu = false; 1907d8be3d52SJohn Baldwin error = vm_handle_ipi(vcpu, vme, &retu); 19080bda8d3eSCorvin Köhne } 19090bda8d3eSCorvin Köhne 1910d087a399SNeel Natu if (error == 0 && retu == false) 1911f76fc5d4SNeel Natu goto restart; 1912f76fc5d4SNeel Natu 19133dc3d32aSJohn Baldwin vmm_stat_incr(vcpu, VMEXIT_USERSPACE, 1); 19143f0f4b15SJohn Baldwin VMM_CTR2(vcpu, "retu %d/%d", error, vme->exitcode); 1915248e6799SNeel Natu 1916318224bbSNeel Natu /* copy the exit information */ 19173f0f4b15SJohn Baldwin *vme_user = *vme; 1918366f6083SPeter Grehan return (error); 1919366f6083SPeter Grehan } 1920366f6083SPeter Grehan 1921366f6083SPeter Grehan int 1922d3956e46SJohn Baldwin vm_restart_instruction(struct vcpu *vcpu) 1923c9c75df4SNeel Natu { 1924d087a399SNeel Natu enum vcpu_state state; 1925d087a399SNeel Natu uint64_t rip; 192673505a10SRobert Wing int error __diagused; 1927c9c75df4SNeel Natu 1928d3956e46SJohn Baldwin state = vcpu_get_state(vcpu, NULL); 1929d087a399SNeel Natu if (state == VCPU_RUNNING) { 1930d087a399SNeel Natu /* 1931d087a399SNeel Natu * When a vcpu is "running" the next instruction is determined 1932d087a399SNeel Natu * by adding 'rip' and 'inst_length' in the vcpu's 'exitinfo'. 1933d087a399SNeel Natu * Thus setting 'inst_length' to zero will cause the current 1934d087a399SNeel Natu * instruction to be restarted. 1935d087a399SNeel Natu */ 1936c9c75df4SNeel Natu vcpu->exitinfo.inst_length = 0; 1937d3956e46SJohn Baldwin VMM_CTR1(vcpu, "restarting instruction at %#lx by " 1938d087a399SNeel Natu "setting inst_length to zero", vcpu->exitinfo.rip); 1939d087a399SNeel Natu } else if (state == VCPU_FROZEN) { 1940d087a399SNeel Natu /* 1941d087a399SNeel Natu * When a vcpu is "frozen" it is outside the critical section 194215add60dSPeter Grehan * around vmmops_run() and 'nextrip' points to the next 194315add60dSPeter Grehan * instruction. Thus instruction restart is achieved by setting 194415add60dSPeter Grehan * 'nextrip' to the vcpu's %rip. 1945d087a399SNeel Natu */ 1946d3956e46SJohn Baldwin error = vm_get_register(vcpu, VM_REG_GUEST_RIP, &rip); 1947d087a399SNeel Natu KASSERT(!error, ("%s: error %d getting rip", __func__, error)); 1948d3956e46SJohn Baldwin VMM_CTR2(vcpu, "restarting instruction by updating " 1949d087a399SNeel Natu "nextrip from %#lx to %#lx", vcpu->nextrip, rip); 1950d087a399SNeel Natu vcpu->nextrip = rip; 1951d087a399SNeel Natu } else { 1952d087a399SNeel Natu panic("%s: invalid state %d", __func__, state); 1953d087a399SNeel Natu } 1954c9c75df4SNeel Natu return (0); 1955c9c75df4SNeel Natu } 1956c9c75df4SNeel Natu 1957c9c75df4SNeel Natu int 195880cb5d84SJohn Baldwin vm_exit_intinfo(struct vcpu *vcpu, uint64_t info) 1959091d4532SNeel Natu { 1960091d4532SNeel Natu int type, vector; 1961091d4532SNeel Natu 1962091d4532SNeel Natu if (info & VM_INTINFO_VALID) { 1963091d4532SNeel Natu type = info & VM_INTINFO_TYPE; 1964091d4532SNeel Natu vector = info & 0xff; 1965091d4532SNeel Natu if (type == VM_INTINFO_NMI && vector != IDT_NMI) 1966091d4532SNeel Natu return (EINVAL); 1967091d4532SNeel Natu if (type == VM_INTINFO_HWEXCEPTION && vector >= 32) 1968091d4532SNeel Natu return (EINVAL); 1969091d4532SNeel Natu if (info & VM_INTINFO_RSVD) 1970091d4532SNeel Natu return (EINVAL); 1971091d4532SNeel Natu } else { 1972091d4532SNeel Natu info = 0; 1973091d4532SNeel Natu } 197480cb5d84SJohn Baldwin VMM_CTR2(vcpu, "%s: info1(%#lx)", __func__, info); 1975091d4532SNeel Natu vcpu->exitintinfo = info; 1976091d4532SNeel Natu return (0); 1977091d4532SNeel Natu } 1978091d4532SNeel Natu 1979091d4532SNeel Natu enum exc_class { 1980091d4532SNeel Natu EXC_BENIGN, 1981091d4532SNeel Natu EXC_CONTRIBUTORY, 1982091d4532SNeel Natu EXC_PAGEFAULT 1983091d4532SNeel Natu }; 1984091d4532SNeel Natu 1985091d4532SNeel Natu #define IDT_VE 20 /* Virtualization Exception (Intel specific) */ 1986091d4532SNeel Natu 1987091d4532SNeel Natu static enum exc_class 1988091d4532SNeel Natu exception_class(uint64_t info) 1989091d4532SNeel Natu { 1990091d4532SNeel Natu int type, vector; 1991091d4532SNeel Natu 1992091d4532SNeel Natu KASSERT(info & VM_INTINFO_VALID, ("intinfo must be valid: %#lx", info)); 1993091d4532SNeel Natu type = info & VM_INTINFO_TYPE; 1994091d4532SNeel Natu vector = info & 0xff; 1995091d4532SNeel Natu 1996091d4532SNeel Natu /* Table 6-4, "Interrupt and Exception Classes", Intel SDM, Vol 3 */ 1997091d4532SNeel Natu switch (type) { 1998091d4532SNeel Natu case VM_INTINFO_HWINTR: 1999091d4532SNeel Natu case VM_INTINFO_SWINTR: 2000091d4532SNeel Natu case VM_INTINFO_NMI: 2001091d4532SNeel Natu return (EXC_BENIGN); 2002091d4532SNeel Natu default: 2003091d4532SNeel Natu /* 2004091d4532SNeel Natu * Hardware exception. 2005091d4532SNeel Natu * 2006091d4532SNeel Natu * SVM and VT-x use identical type values to represent NMI, 2007091d4532SNeel Natu * hardware interrupt and software interrupt. 2008091d4532SNeel Natu * 2009091d4532SNeel Natu * SVM uses type '3' for all exceptions. VT-x uses type '3' 2010091d4532SNeel Natu * for exceptions except #BP and #OF. #BP and #OF use a type 2011091d4532SNeel Natu * value of '5' or '6'. Therefore we don't check for explicit 2012091d4532SNeel Natu * values of 'type' to classify 'intinfo' into a hardware 2013091d4532SNeel Natu * exception. 2014091d4532SNeel Natu */ 2015091d4532SNeel Natu break; 2016091d4532SNeel Natu } 2017091d4532SNeel Natu 2018091d4532SNeel Natu switch (vector) { 2019091d4532SNeel Natu case IDT_PF: 2020091d4532SNeel Natu case IDT_VE: 2021091d4532SNeel Natu return (EXC_PAGEFAULT); 2022091d4532SNeel Natu case IDT_DE: 2023091d4532SNeel Natu case IDT_TS: 2024091d4532SNeel Natu case IDT_NP: 2025091d4532SNeel Natu case IDT_SS: 2026091d4532SNeel Natu case IDT_GP: 2027091d4532SNeel Natu return (EXC_CONTRIBUTORY); 2028091d4532SNeel Natu default: 2029091d4532SNeel Natu return (EXC_BENIGN); 2030091d4532SNeel Natu } 2031091d4532SNeel Natu } 2032091d4532SNeel Natu 2033091d4532SNeel Natu static int 203480cb5d84SJohn Baldwin nested_fault(struct vcpu *vcpu, uint64_t info1, uint64_t info2, 2035091d4532SNeel Natu uint64_t *retinfo) 2036091d4532SNeel Natu { 2037091d4532SNeel Natu enum exc_class exc1, exc2; 2038091d4532SNeel Natu int type1, vector1; 2039091d4532SNeel Natu 2040091d4532SNeel Natu KASSERT(info1 & VM_INTINFO_VALID, ("info1 %#lx is not valid", info1)); 2041091d4532SNeel Natu KASSERT(info2 & VM_INTINFO_VALID, ("info2 %#lx is not valid", info2)); 2042091d4532SNeel Natu 2043091d4532SNeel Natu /* 2044091d4532SNeel Natu * If an exception occurs while attempting to call the double-fault 2045091d4532SNeel Natu * handler the processor enters shutdown mode (aka triple fault). 2046091d4532SNeel Natu */ 2047091d4532SNeel Natu type1 = info1 & VM_INTINFO_TYPE; 2048091d4532SNeel Natu vector1 = info1 & 0xff; 2049091d4532SNeel Natu if (type1 == VM_INTINFO_HWEXCEPTION && vector1 == IDT_DF) { 205080cb5d84SJohn Baldwin VMM_CTR2(vcpu, "triple fault: info1(%#lx), info2(%#lx)", 2051091d4532SNeel Natu info1, info2); 205280cb5d84SJohn Baldwin vm_suspend(vcpu->vm, VM_SUSPEND_TRIPLEFAULT); 2053091d4532SNeel Natu *retinfo = 0; 2054091d4532SNeel Natu return (0); 2055091d4532SNeel Natu } 2056091d4532SNeel Natu 2057091d4532SNeel Natu /* 2058091d4532SNeel Natu * Table 6-5 "Conditions for Generating a Double Fault", Intel SDM, Vol3 2059091d4532SNeel Natu */ 2060091d4532SNeel Natu exc1 = exception_class(info1); 2061091d4532SNeel Natu exc2 = exception_class(info2); 2062091d4532SNeel Natu if ((exc1 == EXC_CONTRIBUTORY && exc2 == EXC_CONTRIBUTORY) || 2063091d4532SNeel Natu (exc1 == EXC_PAGEFAULT && exc2 != EXC_BENIGN)) { 2064091d4532SNeel Natu /* Convert nested fault into a double fault. */ 2065091d4532SNeel Natu *retinfo = IDT_DF; 2066091d4532SNeel Natu *retinfo |= VM_INTINFO_VALID | VM_INTINFO_HWEXCEPTION; 2067091d4532SNeel Natu *retinfo |= VM_INTINFO_DEL_ERRCODE; 2068091d4532SNeel Natu } else { 2069091d4532SNeel Natu /* Handle exceptions serially */ 2070091d4532SNeel Natu *retinfo = info2; 2071091d4532SNeel Natu } 2072091d4532SNeel Natu return (1); 2073091d4532SNeel Natu } 2074091d4532SNeel Natu 2075091d4532SNeel Natu static uint64_t 2076091d4532SNeel Natu vcpu_exception_intinfo(struct vcpu *vcpu) 2077091d4532SNeel Natu { 2078091d4532SNeel Natu uint64_t info = 0; 2079091d4532SNeel Natu 2080091d4532SNeel Natu if (vcpu->exception_pending) { 2081c9c75df4SNeel Natu info = vcpu->exc_vector & 0xff; 2082091d4532SNeel Natu info |= VM_INTINFO_VALID | VM_INTINFO_HWEXCEPTION; 2083c9c75df4SNeel Natu if (vcpu->exc_errcode_valid) { 2084091d4532SNeel Natu info |= VM_INTINFO_DEL_ERRCODE; 2085c9c75df4SNeel Natu info |= (uint64_t)vcpu->exc_errcode << 32; 2086091d4532SNeel Natu } 2087091d4532SNeel Natu } 2088091d4532SNeel Natu return (info); 2089091d4532SNeel Natu } 2090091d4532SNeel Natu 2091091d4532SNeel Natu int 209280cb5d84SJohn Baldwin vm_entry_intinfo(struct vcpu *vcpu, uint64_t *retinfo) 2093091d4532SNeel Natu { 2094091d4532SNeel Natu uint64_t info1, info2; 2095091d4532SNeel Natu int valid; 2096091d4532SNeel Natu 2097091d4532SNeel Natu info1 = vcpu->exitintinfo; 2098091d4532SNeel Natu vcpu->exitintinfo = 0; 2099091d4532SNeel Natu 2100091d4532SNeel Natu info2 = 0; 2101091d4532SNeel Natu if (vcpu->exception_pending) { 2102091d4532SNeel Natu info2 = vcpu_exception_intinfo(vcpu); 2103091d4532SNeel Natu vcpu->exception_pending = 0; 210480cb5d84SJohn Baldwin VMM_CTR2(vcpu, "Exception %d delivered: %#lx", 2105c9c75df4SNeel Natu vcpu->exc_vector, info2); 2106091d4532SNeel Natu } 2107091d4532SNeel Natu 2108091d4532SNeel Natu if ((info1 & VM_INTINFO_VALID) && (info2 & VM_INTINFO_VALID)) { 210980cb5d84SJohn Baldwin valid = nested_fault(vcpu, info1, info2, retinfo); 2110091d4532SNeel Natu } else if (info1 & VM_INTINFO_VALID) { 2111091d4532SNeel Natu *retinfo = info1; 2112091d4532SNeel Natu valid = 1; 2113091d4532SNeel Natu } else if (info2 & VM_INTINFO_VALID) { 2114091d4532SNeel Natu *retinfo = info2; 2115091d4532SNeel Natu valid = 1; 2116091d4532SNeel Natu } else { 2117091d4532SNeel Natu valid = 0; 2118091d4532SNeel Natu } 2119091d4532SNeel Natu 2120091d4532SNeel Natu if (valid) { 2121d3956e46SJohn Baldwin VMM_CTR4(vcpu, "%s: info1(%#lx), info2(%#lx), " 2122091d4532SNeel Natu "retinfo(%#lx)", __func__, info1, info2, *retinfo); 2123091d4532SNeel Natu } 2124091d4532SNeel Natu 2125091d4532SNeel Natu return (valid); 2126091d4532SNeel Natu } 2127091d4532SNeel Natu 2128091d4532SNeel Natu int 21293f0f4b15SJohn Baldwin vm_get_intinfo(struct vcpu *vcpu, uint64_t *info1, uint64_t *info2) 2130091d4532SNeel Natu { 2131091d4532SNeel Natu *info1 = vcpu->exitintinfo; 2132091d4532SNeel Natu *info2 = vcpu_exception_intinfo(vcpu); 2133091d4532SNeel Natu return (0); 2134091d4532SNeel Natu } 2135091d4532SNeel Natu 2136091d4532SNeel Natu int 2137d3956e46SJohn Baldwin vm_inject_exception(struct vcpu *vcpu, int vector, int errcode_valid, 2138c9c75df4SNeel Natu uint32_t errcode, int restart_instruction) 2139366f6083SPeter Grehan { 214047b9935dSNeel Natu uint64_t regval; 214173505a10SRobert Wing int error __diagused; 2142dc506506SNeel Natu 2143c9c75df4SNeel Natu if (vector < 0 || vector >= 32) 2144366f6083SPeter Grehan return (EINVAL); 2145366f6083SPeter Grehan 2146091d4532SNeel Natu /* 2147091d4532SNeel Natu * A double fault exception should never be injected directly into 2148091d4532SNeel Natu * the guest. It is a derived exception that results from specific 2149091d4532SNeel Natu * combinations of nested faults. 2150091d4532SNeel Natu */ 2151c9c75df4SNeel Natu if (vector == IDT_DF) 2152091d4532SNeel Natu return (EINVAL); 2153091d4532SNeel Natu 2154dc506506SNeel Natu if (vcpu->exception_pending) { 2155d3956e46SJohn Baldwin VMM_CTR2(vcpu, "Unable to inject exception %d due to " 2156c9c75df4SNeel Natu "pending exception %d", vector, vcpu->exc_vector); 2157dc506506SNeel Natu return (EBUSY); 2158dc506506SNeel Natu } 2159dc506506SNeel Natu 216047b9935dSNeel Natu if (errcode_valid) { 216147b9935dSNeel Natu /* 216247b9935dSNeel Natu * Exceptions don't deliver an error code in real mode. 216347b9935dSNeel Natu */ 2164d3956e46SJohn Baldwin error = vm_get_register(vcpu, VM_REG_GUEST_CR0, ®val); 216547b9935dSNeel Natu KASSERT(!error, ("%s: error %d getting CR0", __func__, error)); 216647b9935dSNeel Natu if (!(regval & CR0_PE)) 216747b9935dSNeel Natu errcode_valid = 0; 216847b9935dSNeel Natu } 216947b9935dSNeel Natu 21702ce12423SNeel Natu /* 21712ce12423SNeel Natu * From section 26.6.1 "Interruptibility State" in Intel SDM: 21722ce12423SNeel Natu * 21732ce12423SNeel Natu * Event blocking by "STI" or "MOV SS" is cleared after guest executes 21742ce12423SNeel Natu * one instruction or incurs an exception. 21752ce12423SNeel Natu */ 2176d3956e46SJohn Baldwin error = vm_set_register(vcpu, VM_REG_GUEST_INTR_SHADOW, 0); 21772ce12423SNeel Natu KASSERT(error == 0, ("%s: error %d clearing interrupt shadow", 21782ce12423SNeel Natu __func__, error)); 21792ce12423SNeel Natu 2180c9c75df4SNeel Natu if (restart_instruction) 2181d3956e46SJohn Baldwin vm_restart_instruction(vcpu); 2182c9c75df4SNeel Natu 2183dc506506SNeel Natu vcpu->exception_pending = 1; 2184c9c75df4SNeel Natu vcpu->exc_vector = vector; 2185c9c75df4SNeel Natu vcpu->exc_errcode = errcode; 2186c9c75df4SNeel Natu vcpu->exc_errcode_valid = errcode_valid; 2187d3956e46SJohn Baldwin VMM_CTR1(vcpu, "Exception %d pending", vector); 2188dc506506SNeel Natu return (0); 2189dc506506SNeel Natu } 2190dc506506SNeel Natu 2191d37f2adbSNeel Natu void 2192d3956e46SJohn Baldwin vm_inject_fault(struct vcpu *vcpu, int vector, int errcode_valid, int errcode) 2193dc506506SNeel Natu { 219473505a10SRobert Wing int error __diagused, restart_instruction; 2195dc506506SNeel Natu 2196c9c75df4SNeel Natu restart_instruction = 1; 2197d37f2adbSNeel Natu 2198d3956e46SJohn Baldwin error = vm_inject_exception(vcpu, vector, errcode_valid, 2199c9c75df4SNeel Natu errcode, restart_instruction); 2200dc506506SNeel Natu KASSERT(error == 0, ("vm_inject_exception error %d", error)); 2201dc506506SNeel Natu } 2202dc506506SNeel Natu 2203dc506506SNeel Natu void 2204d3956e46SJohn Baldwin vm_inject_pf(struct vcpu *vcpu, int error_code, uint64_t cr2) 2205fd949af6SNeel Natu { 220673505a10SRobert Wing int error __diagused; 220737a723a5SNeel Natu 2208d3956e46SJohn Baldwin VMM_CTR2(vcpu, "Injecting page fault: error_code %#x, cr2 %#lx", 220937a723a5SNeel Natu error_code, cr2); 221037a723a5SNeel Natu 2211d3956e46SJohn Baldwin error = vm_set_register(vcpu, VM_REG_GUEST_CR2, cr2); 221237a723a5SNeel Natu KASSERT(error == 0, ("vm_set_register(cr2) error %d", error)); 2213fd949af6SNeel Natu 2214d3956e46SJohn Baldwin vm_inject_fault(vcpu, IDT_PF, 1, error_code); 2215366f6083SPeter Grehan } 2216366f6083SPeter Grehan 221761592433SNeel Natu static VMM_STAT(VCPU_NMI_COUNT, "number of NMIs delivered to vcpu"); 2218366f6083SPeter Grehan 2219f352ff0cSNeel Natu int 22203f0f4b15SJohn Baldwin vm_inject_nmi(struct vcpu *vcpu) 2221f352ff0cSNeel Natu { 2222f352ff0cSNeel Natu 2223f352ff0cSNeel Natu vcpu->nmi_pending = 1; 22243f0f4b15SJohn Baldwin vcpu_notify_event(vcpu, false); 2225f352ff0cSNeel Natu return (0); 2226f352ff0cSNeel Natu } 2227f352ff0cSNeel Natu 2228f352ff0cSNeel Natu int 222980cb5d84SJohn Baldwin vm_nmi_pending(struct vcpu *vcpu) 2230f352ff0cSNeel Natu { 2231f352ff0cSNeel Natu return (vcpu->nmi_pending); 2232f352ff0cSNeel Natu } 2233f352ff0cSNeel Natu 2234f352ff0cSNeel Natu void 223580cb5d84SJohn Baldwin vm_nmi_clear(struct vcpu *vcpu) 2236f352ff0cSNeel Natu { 2237f352ff0cSNeel Natu if (vcpu->nmi_pending == 0) 2238f352ff0cSNeel Natu panic("vm_nmi_clear: inconsistent nmi_pending state"); 2239f352ff0cSNeel Natu 2240f352ff0cSNeel Natu vcpu->nmi_pending = 0; 22413dc3d32aSJohn Baldwin vmm_stat_incr(vcpu, VCPU_NMI_COUNT, 1); 2242366f6083SPeter Grehan } 2243366f6083SPeter Grehan 22440775fbb4STycho Nightingale static VMM_STAT(VCPU_EXTINT_COUNT, "number of ExtINTs delivered to vcpu"); 22450775fbb4STycho Nightingale 22460775fbb4STycho Nightingale int 22473f0f4b15SJohn Baldwin vm_inject_extint(struct vcpu *vcpu) 22480775fbb4STycho Nightingale { 22490775fbb4STycho Nightingale 22500775fbb4STycho Nightingale vcpu->extint_pending = 1; 22513f0f4b15SJohn Baldwin vcpu_notify_event(vcpu, false); 22520775fbb4STycho Nightingale return (0); 22530775fbb4STycho Nightingale } 22540775fbb4STycho Nightingale 22550775fbb4STycho Nightingale int 225680cb5d84SJohn Baldwin vm_extint_pending(struct vcpu *vcpu) 22570775fbb4STycho Nightingale { 22580775fbb4STycho Nightingale return (vcpu->extint_pending); 22590775fbb4STycho Nightingale } 22600775fbb4STycho Nightingale 22610775fbb4STycho Nightingale void 226280cb5d84SJohn Baldwin vm_extint_clear(struct vcpu *vcpu) 22630775fbb4STycho Nightingale { 22640775fbb4STycho Nightingale if (vcpu->extint_pending == 0) 22650775fbb4STycho Nightingale panic("vm_extint_clear: inconsistent extint_pending state"); 22660775fbb4STycho Nightingale 22670775fbb4STycho Nightingale vcpu->extint_pending = 0; 22683dc3d32aSJohn Baldwin vmm_stat_incr(vcpu, VCPU_EXTINT_COUNT, 1); 22690775fbb4STycho Nightingale } 22700775fbb4STycho Nightingale 2271366f6083SPeter Grehan int 22723f0f4b15SJohn Baldwin vm_get_capability(struct vcpu *vcpu, int type, int *retval) 2273366f6083SPeter Grehan { 2274366f6083SPeter Grehan if (type < 0 || type >= VM_CAP_MAX) 2275366f6083SPeter Grehan return (EINVAL); 2276366f6083SPeter Grehan 22773f0f4b15SJohn Baldwin return (vmmops_getcap(vcpu->cookie, type, retval)); 2278366f6083SPeter Grehan } 2279366f6083SPeter Grehan 2280366f6083SPeter Grehan int 22813f0f4b15SJohn Baldwin vm_set_capability(struct vcpu *vcpu, int type, int val) 2282366f6083SPeter Grehan { 2283366f6083SPeter Grehan if (type < 0 || type >= VM_CAP_MAX) 2284366f6083SPeter Grehan return (EINVAL); 2285366f6083SPeter Grehan 22863f0f4b15SJohn Baldwin return (vmmops_setcap(vcpu->cookie, type, val)); 2287366f6083SPeter Grehan } 2288366f6083SPeter Grehan 2289950af9ffSJohn Baldwin struct vm * 2290950af9ffSJohn Baldwin vcpu_vm(struct vcpu *vcpu) 2291950af9ffSJohn Baldwin { 2292950af9ffSJohn Baldwin return (vcpu->vm); 2293950af9ffSJohn Baldwin } 2294950af9ffSJohn Baldwin 2295950af9ffSJohn Baldwin int 2296950af9ffSJohn Baldwin vcpu_vcpuid(struct vcpu *vcpu) 2297950af9ffSJohn Baldwin { 2298950af9ffSJohn Baldwin return (vcpu->vcpuid); 2299950af9ffSJohn Baldwin } 2300950af9ffSJohn Baldwin 2301950af9ffSJohn Baldwin struct vcpu * 2302950af9ffSJohn Baldwin vm_vcpu(struct vm *vm, int vcpuid) 2303950af9ffSJohn Baldwin { 2304*98568a00SJohn Baldwin return (vm->vcpu[vcpuid]); 2305950af9ffSJohn Baldwin } 2306950af9ffSJohn Baldwin 2307366f6083SPeter Grehan struct vlapic * 2308d3956e46SJohn Baldwin vm_lapic(struct vcpu *vcpu) 2309366f6083SPeter Grehan { 2310d3956e46SJohn Baldwin return (vcpu->vlapic); 2311366f6083SPeter Grehan } 2312366f6083SPeter Grehan 2313565bbb86SNeel Natu struct vioapic * 2314565bbb86SNeel Natu vm_ioapic(struct vm *vm) 2315565bbb86SNeel Natu { 2316565bbb86SNeel Natu 2317565bbb86SNeel Natu return (vm->vioapic); 2318565bbb86SNeel Natu } 2319565bbb86SNeel Natu 232008e3ff32SNeel Natu struct vhpet * 232108e3ff32SNeel Natu vm_hpet(struct vm *vm) 232208e3ff32SNeel Natu { 232308e3ff32SNeel Natu 232408e3ff32SNeel Natu return (vm->vhpet); 232508e3ff32SNeel Natu } 232608e3ff32SNeel Natu 2327490d56c5SEd Maste bool 2328366f6083SPeter Grehan vmm_is_pptdev(int bus, int slot, int func) 2329366f6083SPeter Grehan { 2330490d56c5SEd Maste int b, f, i, n, s; 2331366f6083SPeter Grehan char *val, *cp, *cp2; 2332490d56c5SEd Maste bool found; 2333366f6083SPeter Grehan 2334366f6083SPeter Grehan /* 233507044a96SNeel Natu * XXX 233607044a96SNeel Natu * The length of an environment variable is limited to 128 bytes which 233707044a96SNeel Natu * puts an upper limit on the number of passthru devices that may be 233807044a96SNeel Natu * specified using a single environment variable. 233907044a96SNeel Natu * 234007044a96SNeel Natu * Work around this by scanning multiple environment variable 234107044a96SNeel Natu * names instead of a single one - yuck! 2342366f6083SPeter Grehan */ 234307044a96SNeel Natu const char *names[] = { "pptdevs", "pptdevs2", "pptdevs3", NULL }; 234407044a96SNeel Natu 234507044a96SNeel Natu /* set pptdevs="1/2/3 4/5/6 7/8/9 10/11/12" */ 2346490d56c5SEd Maste found = false; 234707044a96SNeel Natu for (i = 0; names[i] != NULL && !found; i++) { 23482be111bfSDavide Italiano cp = val = kern_getenv(names[i]); 2349366f6083SPeter Grehan while (cp != NULL && *cp != '\0') { 2350366f6083SPeter Grehan if ((cp2 = strchr(cp, ' ')) != NULL) 2351366f6083SPeter Grehan *cp2 = '\0'; 2352366f6083SPeter Grehan 2353366f6083SPeter Grehan n = sscanf(cp, "%d/%d/%d", &b, &s, &f); 2354366f6083SPeter Grehan if (n == 3 && bus == b && slot == s && func == f) { 2355490d56c5SEd Maste found = true; 2356366f6083SPeter Grehan break; 2357366f6083SPeter Grehan } 2358366f6083SPeter Grehan 2359366f6083SPeter Grehan if (cp2 != NULL) 2360366f6083SPeter Grehan *cp2++ = ' '; 2361366f6083SPeter Grehan 2362366f6083SPeter Grehan cp = cp2; 2363366f6083SPeter Grehan } 2364366f6083SPeter Grehan freeenv(val); 236507044a96SNeel Natu } 2366366f6083SPeter Grehan return (found); 2367366f6083SPeter Grehan } 2368366f6083SPeter Grehan 2369366f6083SPeter Grehan void * 2370366f6083SPeter Grehan vm_iommu_domain(struct vm *vm) 2371366f6083SPeter Grehan { 2372366f6083SPeter Grehan 2373366f6083SPeter Grehan return (vm->iommu); 2374366f6083SPeter Grehan } 2375366f6083SPeter Grehan 237675dd3366SNeel Natu int 23773f0f4b15SJohn Baldwin vcpu_set_state(struct vcpu *vcpu, enum vcpu_state newstate, bool from_idle) 2378366f6083SPeter Grehan { 237975dd3366SNeel Natu int error; 2380366f6083SPeter Grehan 238175dd3366SNeel Natu vcpu_lock(vcpu); 23823f0f4b15SJohn Baldwin error = vcpu_set_state_locked(vcpu, newstate, from_idle); 238375dd3366SNeel Natu vcpu_unlock(vcpu); 238475dd3366SNeel Natu 238575dd3366SNeel Natu return (error); 238675dd3366SNeel Natu } 238775dd3366SNeel Natu 238875dd3366SNeel Natu enum vcpu_state 2389d3956e46SJohn Baldwin vcpu_get_state(struct vcpu *vcpu, int *hostcpu) 2390366f6083SPeter Grehan { 239175dd3366SNeel Natu enum vcpu_state state; 2392366f6083SPeter Grehan 239375dd3366SNeel Natu vcpu_lock(vcpu); 239475dd3366SNeel Natu state = vcpu->state; 2395d3c11f40SPeter Grehan if (hostcpu != NULL) 2396d3c11f40SPeter Grehan *hostcpu = vcpu->hostcpu; 239775dd3366SNeel Natu vcpu_unlock(vcpu); 2398366f6083SPeter Grehan 239975dd3366SNeel Natu return (state); 2400366f6083SPeter Grehan } 2401366f6083SPeter Grehan 240295ebc360SNeel Natu int 24033f0f4b15SJohn Baldwin vm_activate_cpu(struct vcpu *vcpu) 2404366f6083SPeter Grehan { 24053f0f4b15SJohn Baldwin struct vm *vm = vcpu->vm; 2406366f6083SPeter Grehan 24073f0f4b15SJohn Baldwin if (CPU_ISSET(vcpu->vcpuid, &vm->active_cpus)) 240895ebc360SNeel Natu return (EBUSY); 240922d822c6SNeel Natu 24103f0f4b15SJohn Baldwin VMM_CTR0(vcpu, "activated"); 24113f0f4b15SJohn Baldwin CPU_SET_ATOMIC(vcpu->vcpuid, &vm->active_cpus); 241295ebc360SNeel Natu return (0); 2413366f6083SPeter Grehan } 2414366f6083SPeter Grehan 2415fc276d92SJohn Baldwin int 24163f0f4b15SJohn Baldwin vm_suspend_cpu(struct vm *vm, struct vcpu *vcpu) 2417fc276d92SJohn Baldwin { 24183f0f4b15SJohn Baldwin if (vcpu == NULL) { 2419fc276d92SJohn Baldwin vm->debug_cpus = vm->active_cpus; 24203f0f4b15SJohn Baldwin for (int i = 0; i < vm->maxcpus; i++) { 2421fc276d92SJohn Baldwin if (CPU_ISSET(i, &vm->active_cpus)) 24223f0f4b15SJohn Baldwin vcpu_notify_event(vm_vcpu(vm, i), false); 2423fc276d92SJohn Baldwin } 2424fc276d92SJohn Baldwin } else { 24253f0f4b15SJohn Baldwin if (!CPU_ISSET(vcpu->vcpuid, &vm->active_cpus)) 2426fc276d92SJohn Baldwin return (EINVAL); 2427fc276d92SJohn Baldwin 24283f0f4b15SJohn Baldwin CPU_SET_ATOMIC(vcpu->vcpuid, &vm->debug_cpus); 24293f0f4b15SJohn Baldwin vcpu_notify_event(vcpu, false); 2430fc276d92SJohn Baldwin } 2431fc276d92SJohn Baldwin return (0); 2432fc276d92SJohn Baldwin } 2433fc276d92SJohn Baldwin 2434fc276d92SJohn Baldwin int 24353f0f4b15SJohn Baldwin vm_resume_cpu(struct vm *vm, struct vcpu *vcpu) 2436fc276d92SJohn Baldwin { 2437fc276d92SJohn Baldwin 24383f0f4b15SJohn Baldwin if (vcpu == NULL) { 2439fc276d92SJohn Baldwin CPU_ZERO(&vm->debug_cpus); 2440fc276d92SJohn Baldwin } else { 24413f0f4b15SJohn Baldwin if (!CPU_ISSET(vcpu->vcpuid, &vm->debug_cpus)) 2442fc276d92SJohn Baldwin return (EINVAL); 2443fc276d92SJohn Baldwin 24443f0f4b15SJohn Baldwin CPU_CLR_ATOMIC(vcpu->vcpuid, &vm->debug_cpus); 2445fc276d92SJohn Baldwin } 2446fc276d92SJohn Baldwin return (0); 2447fc276d92SJohn Baldwin } 2448fc276d92SJohn Baldwin 2449fc276d92SJohn Baldwin int 245080cb5d84SJohn Baldwin vcpu_debugged(struct vcpu *vcpu) 2451fc276d92SJohn Baldwin { 2452fc276d92SJohn Baldwin 245380cb5d84SJohn Baldwin return (CPU_ISSET(vcpu->vcpuid, &vcpu->vm->debug_cpus)); 2454fc276d92SJohn Baldwin } 2455fc276d92SJohn Baldwin 2456a5615c90SPeter Grehan cpuset_t 2457366f6083SPeter Grehan vm_active_cpus(struct vm *vm) 2458366f6083SPeter Grehan { 2459366f6083SPeter Grehan 2460366f6083SPeter Grehan return (vm->active_cpus); 2461366f6083SPeter Grehan } 2462366f6083SPeter Grehan 246395ebc360SNeel Natu cpuset_t 2464fc276d92SJohn Baldwin vm_debug_cpus(struct vm *vm) 2465fc276d92SJohn Baldwin { 2466fc276d92SJohn Baldwin 2467fc276d92SJohn Baldwin return (vm->debug_cpus); 2468fc276d92SJohn Baldwin } 2469fc276d92SJohn Baldwin 2470fc276d92SJohn Baldwin cpuset_t 247195ebc360SNeel Natu vm_suspended_cpus(struct vm *vm) 247295ebc360SNeel Natu { 247395ebc360SNeel Natu 247495ebc360SNeel Natu return (vm->suspended_cpus); 247595ebc360SNeel Natu } 247695ebc360SNeel Natu 2477c0f35dbfSJohn Baldwin /* 2478c0f35dbfSJohn Baldwin * Returns the subset of vCPUs in tostart that are awaiting startup. 2479c0f35dbfSJohn Baldwin * These vCPUs are also marked as no longer awaiting startup. 2480c0f35dbfSJohn Baldwin */ 2481c0f35dbfSJohn Baldwin cpuset_t 2482c0f35dbfSJohn Baldwin vm_start_cpus(struct vm *vm, const cpuset_t *tostart) 2483c0f35dbfSJohn Baldwin { 2484c0f35dbfSJohn Baldwin cpuset_t set; 2485c0f35dbfSJohn Baldwin 2486c0f35dbfSJohn Baldwin mtx_lock(&vm->rendezvous_mtx); 2487c0f35dbfSJohn Baldwin CPU_AND(&set, &vm->startup_cpus, tostart); 2488c0f35dbfSJohn Baldwin CPU_ANDNOT(&vm->startup_cpus, &vm->startup_cpus, &set); 2489c0f35dbfSJohn Baldwin mtx_unlock(&vm->rendezvous_mtx); 2490c0f35dbfSJohn Baldwin return (set); 2491c0f35dbfSJohn Baldwin } 2492c0f35dbfSJohn Baldwin 2493c0f35dbfSJohn Baldwin void 2494c0f35dbfSJohn Baldwin vm_await_start(struct vm *vm, const cpuset_t *waiting) 2495c0f35dbfSJohn Baldwin { 2496c0f35dbfSJohn Baldwin mtx_lock(&vm->rendezvous_mtx); 2497c0f35dbfSJohn Baldwin CPU_OR(&vm->startup_cpus, &vm->startup_cpus, waiting); 2498c0f35dbfSJohn Baldwin mtx_unlock(&vm->rendezvous_mtx); 2499c0f35dbfSJohn Baldwin } 2500c0f35dbfSJohn Baldwin 2501366f6083SPeter Grehan void * 25023dc3d32aSJohn Baldwin vcpu_stats(struct vcpu *vcpu) 2503366f6083SPeter Grehan { 2504366f6083SPeter Grehan 25053dc3d32aSJohn Baldwin return (vcpu->stats); 2506366f6083SPeter Grehan } 2507e9027382SNeel Natu 2508e9027382SNeel Natu int 25093f0f4b15SJohn Baldwin vm_get_x2apic_state(struct vcpu *vcpu, enum x2apic_state *state) 2510e9027382SNeel Natu { 25113f0f4b15SJohn Baldwin *state = vcpu->x2apic_state; 2512e9027382SNeel Natu 2513e9027382SNeel Natu return (0); 2514e9027382SNeel Natu } 2515e9027382SNeel Natu 2516e9027382SNeel Natu int 25173f0f4b15SJohn Baldwin vm_set_x2apic_state(struct vcpu *vcpu, enum x2apic_state state) 2518e9027382SNeel Natu { 25193f23d3caSNeel Natu if (state >= X2APIC_STATE_LAST) 2520e9027382SNeel Natu return (EINVAL); 2521e9027382SNeel Natu 2522d3956e46SJohn Baldwin vcpu->x2apic_state = state; 2523e9027382SNeel Natu 2524d3956e46SJohn Baldwin vlapic_set_x2apic_state(vcpu, state); 252573820fb0SNeel Natu 2526e9027382SNeel Natu return (0); 2527e9027382SNeel Natu } 252875dd3366SNeel Natu 252922821874SNeel Natu /* 253022821874SNeel Natu * This function is called to ensure that a vcpu "sees" a pending event 253122821874SNeel Natu * as soon as possible: 253222821874SNeel Natu * - If the vcpu thread is sleeping then it is woken up. 253322821874SNeel Natu * - If the vcpu is running on a different host_cpu then an IPI will be directed 253422821874SNeel Natu * to the host_cpu to cause the vcpu to trap into the hypervisor. 253522821874SNeel Natu */ 2536248e6799SNeel Natu static void 2537248e6799SNeel Natu vcpu_notify_event_locked(struct vcpu *vcpu, bool lapic_intr) 253875dd3366SNeel Natu { 253975dd3366SNeel Natu int hostcpu; 254075dd3366SNeel Natu 254175dd3366SNeel Natu hostcpu = vcpu->hostcpu; 2542ef39d7e9SNeel Natu if (vcpu->state == VCPU_RUNNING) { 2543ef39d7e9SNeel Natu KASSERT(hostcpu != NOCPU, ("vcpu running on invalid hostcpu")); 2544de5ea6b6SNeel Natu if (hostcpu != curcpu) { 2545ef39d7e9SNeel Natu if (lapic_intr) { 2546add611fdSNeel Natu vlapic_post_intr(vcpu->vlapic, hostcpu, 2547add611fdSNeel Natu vmm_ipinum); 2548ef39d7e9SNeel Natu } else { 254975dd3366SNeel Natu ipi_cpu(hostcpu, vmm_ipinum); 255075dd3366SNeel Natu } 2551ef39d7e9SNeel Natu } else { 2552ef39d7e9SNeel Natu /* 2553ef39d7e9SNeel Natu * If the 'vcpu' is running on 'curcpu' then it must 2554ef39d7e9SNeel Natu * be sending a notification to itself (e.g. SELF_IPI). 2555ef39d7e9SNeel Natu * The pending event will be picked up when the vcpu 2556ef39d7e9SNeel Natu * transitions back to guest context. 2557ef39d7e9SNeel Natu */ 2558ef39d7e9SNeel Natu } 2559ef39d7e9SNeel Natu } else { 2560ef39d7e9SNeel Natu KASSERT(hostcpu == NOCPU, ("vcpu state %d not consistent " 2561ef39d7e9SNeel Natu "with hostcpu %d", vcpu->state, hostcpu)); 2562366f6083SPeter Grehan if (vcpu->state == VCPU_SLEEPING) 2563366f6083SPeter Grehan wakeup_one(vcpu); 2564366f6083SPeter Grehan } 2565248e6799SNeel Natu } 2566248e6799SNeel Natu 2567248e6799SNeel Natu void 25683f0f4b15SJohn Baldwin vcpu_notify_event(struct vcpu *vcpu, bool lapic_intr) 2569248e6799SNeel Natu { 2570248e6799SNeel Natu vcpu_lock(vcpu); 2571248e6799SNeel Natu vcpu_notify_event_locked(vcpu, lapic_intr); 2572f76fc5d4SNeel Natu vcpu_unlock(vcpu); 2573f76fc5d4SNeel Natu } 2574318224bbSNeel Natu 2575318224bbSNeel Natu struct vmspace * 2576318224bbSNeel Natu vm_get_vmspace(struct vm *vm) 2577318224bbSNeel Natu { 2578318224bbSNeel Natu 2579318224bbSNeel Natu return (vm->vmspace); 2580318224bbSNeel Natu } 2581565bbb86SNeel Natu 2582565bbb86SNeel Natu int 2583565bbb86SNeel Natu vm_apicid2vcpuid(struct vm *vm, int apicid) 2584565bbb86SNeel Natu { 2585565bbb86SNeel Natu /* 2586565bbb86SNeel Natu * XXX apic id is assumed to be numerically identical to vcpu id 2587565bbb86SNeel Natu */ 2588565bbb86SNeel Natu return (apicid); 2589565bbb86SNeel Natu } 25905b8a8cd1SNeel Natu 2591b837daddSKonstantin Belousov int 2592d8be3d52SJohn Baldwin vm_smp_rendezvous(struct vcpu *vcpu, cpuset_t dest, 25935b8a8cd1SNeel Natu vm_rendezvous_func_t func, void *arg) 25945b8a8cd1SNeel Natu { 2595d8be3d52SJohn Baldwin struct vm *vm = vcpu->vm; 2596b837daddSKonstantin Belousov int error, i; 2597970955e4SNeel Natu 25985b8a8cd1SNeel Natu /* 25995b8a8cd1SNeel Natu * Enforce that this function is called without any locks 26005b8a8cd1SNeel Natu */ 26015b8a8cd1SNeel Natu WITNESS_WARN(WARN_PANIC, NULL, "vm_smp_rendezvous"); 26025b8a8cd1SNeel Natu 26035b8a8cd1SNeel Natu restart: 26045b8a8cd1SNeel Natu mtx_lock(&vm->rendezvous_mtx); 26055b8a8cd1SNeel Natu if (vm->rendezvous_func != NULL) { 26065b8a8cd1SNeel Natu /* 26075b8a8cd1SNeel Natu * If a rendezvous is already in progress then we need to 26083f0f4b15SJohn Baldwin * call the rendezvous handler in case this 'vcpu' is one 26095b8a8cd1SNeel Natu * of the targets of the rendezvous. 26105b8a8cd1SNeel Natu */ 2611d8be3d52SJohn Baldwin VMM_CTR0(vcpu, "Rendezvous already in progress"); 26125b8a8cd1SNeel Natu mtx_unlock(&vm->rendezvous_mtx); 2613d8be3d52SJohn Baldwin error = vm_handle_rendezvous(vcpu); 2614b837daddSKonstantin Belousov if (error != 0) 2615b837daddSKonstantin Belousov return (error); 26165b8a8cd1SNeel Natu goto restart; 26175b8a8cd1SNeel Natu } 26185b8a8cd1SNeel Natu KASSERT(vm->rendezvous_func == NULL, ("vm_smp_rendezvous: previous " 26195b8a8cd1SNeel Natu "rendezvous is still in progress")); 26205b8a8cd1SNeel Natu 2621d8be3d52SJohn Baldwin VMM_CTR0(vcpu, "Initiating rendezvous"); 26225b8a8cd1SNeel Natu vm->rendezvous_req_cpus = dest; 26235b8a8cd1SNeel Natu CPU_ZERO(&vm->rendezvous_done_cpus); 26245b8a8cd1SNeel Natu vm->rendezvous_arg = arg; 2625869dbab7SAndriy Gapon vm->rendezvous_func = func; 26265b8a8cd1SNeel Natu mtx_unlock(&vm->rendezvous_mtx); 26275b8a8cd1SNeel Natu 2628970955e4SNeel Natu /* 2629970955e4SNeel Natu * Wake up any sleeping vcpus and trigger a VM-exit in any running 2630970955e4SNeel Natu * vcpus so they handle the rendezvous as soon as possible. 2631970955e4SNeel Natu */ 2632a488c9c9SRodney W. Grimes for (i = 0; i < vm->maxcpus; i++) { 2633970955e4SNeel Natu if (CPU_ISSET(i, &dest)) 26343f0f4b15SJohn Baldwin vcpu_notify_event(vm_vcpu(vm, i), false); 2635970955e4SNeel Natu } 2636970955e4SNeel Natu 2637d8be3d52SJohn Baldwin return (vm_handle_rendezvous(vcpu)); 26385b8a8cd1SNeel Natu } 2639762fd208STycho Nightingale 2640762fd208STycho Nightingale struct vatpic * 2641762fd208STycho Nightingale vm_atpic(struct vm *vm) 2642762fd208STycho Nightingale { 2643762fd208STycho Nightingale return (vm->vatpic); 2644762fd208STycho Nightingale } 2645e883c9bbSTycho Nightingale 2646e883c9bbSTycho Nightingale struct vatpit * 2647e883c9bbSTycho Nightingale vm_atpit(struct vm *vm) 2648e883c9bbSTycho Nightingale { 2649e883c9bbSTycho Nightingale return (vm->vatpit); 2650e883c9bbSTycho Nightingale } 2651d17b5104SNeel Natu 2652160ef77aSNeel Natu struct vpmtmr * 2653160ef77aSNeel Natu vm_pmtmr(struct vm *vm) 2654160ef77aSNeel Natu { 2655160ef77aSNeel Natu 2656160ef77aSNeel Natu return (vm->vpmtmr); 2657160ef77aSNeel Natu } 2658160ef77aSNeel Natu 26590dafa5cdSNeel Natu struct vrtc * 26600dafa5cdSNeel Natu vm_rtc(struct vm *vm) 26610dafa5cdSNeel Natu { 26620dafa5cdSNeel Natu 26630dafa5cdSNeel Natu return (vm->vrtc); 26640dafa5cdSNeel Natu } 26650dafa5cdSNeel Natu 2666d17b5104SNeel Natu enum vm_reg_name 2667d17b5104SNeel Natu vm_segment_name(int seg) 2668d17b5104SNeel Natu { 2669d17b5104SNeel Natu static enum vm_reg_name seg_names[] = { 2670d17b5104SNeel Natu VM_REG_GUEST_ES, 2671d17b5104SNeel Natu VM_REG_GUEST_CS, 2672d17b5104SNeel Natu VM_REG_GUEST_SS, 2673d17b5104SNeel Natu VM_REG_GUEST_DS, 2674d17b5104SNeel Natu VM_REG_GUEST_FS, 2675d17b5104SNeel Natu VM_REG_GUEST_GS 2676d17b5104SNeel Natu }; 2677d17b5104SNeel Natu 2678d17b5104SNeel Natu KASSERT(seg >= 0 && seg < nitems(seg_names), 2679d17b5104SNeel Natu ("%s: invalid segment encoding %d", __func__, seg)); 2680d17b5104SNeel Natu return (seg_names[seg]); 2681d17b5104SNeel Natu } 2682cf1d80d8SPeter Grehan 2683d665d229SNeel Natu void 26842b4fe856SJohn Baldwin vm_copy_teardown(struct vm_copyinfo *copyinfo, int num_copyinfo) 2685d665d229SNeel Natu { 2686d665d229SNeel Natu int idx; 2687d665d229SNeel Natu 2688d665d229SNeel Natu for (idx = 0; idx < num_copyinfo; idx++) { 2689d665d229SNeel Natu if (copyinfo[idx].cookie != NULL) 2690d665d229SNeel Natu vm_gpa_release(copyinfo[idx].cookie); 2691d665d229SNeel Natu } 2692d665d229SNeel Natu bzero(copyinfo, num_copyinfo * sizeof(struct vm_copyinfo)); 2693d665d229SNeel Natu } 2694d665d229SNeel Natu 2695d665d229SNeel Natu int 2696d3956e46SJohn Baldwin vm_copy_setup(struct vcpu *vcpu, struct vm_guest_paging *paging, 2697d665d229SNeel Natu uint64_t gla, size_t len, int prot, struct vm_copyinfo *copyinfo, 26989c4d5478SNeel Natu int num_copyinfo, int *fault) 2699d665d229SNeel Natu { 2700d665d229SNeel Natu int error, idx, nused; 2701d665d229SNeel Natu size_t n, off, remaining; 2702d665d229SNeel Natu void *hva, *cookie; 2703d665d229SNeel Natu uint64_t gpa; 2704d665d229SNeel Natu 2705d665d229SNeel Natu bzero(copyinfo, sizeof(struct vm_copyinfo) * num_copyinfo); 2706d665d229SNeel Natu 2707d665d229SNeel Natu nused = 0; 2708d665d229SNeel Natu remaining = len; 2709d665d229SNeel Natu while (remaining > 0) { 2710d665d229SNeel Natu KASSERT(nused < num_copyinfo, ("insufficient vm_copyinfo")); 2711d3956e46SJohn Baldwin error = vm_gla2gpa(vcpu, paging, gla, prot, &gpa, fault); 27129c4d5478SNeel Natu if (error || *fault) 2713d665d229SNeel Natu return (error); 2714d665d229SNeel Natu off = gpa & PAGE_MASK; 2715d665d229SNeel Natu n = min(remaining, PAGE_SIZE - off); 2716d665d229SNeel Natu copyinfo[nused].gpa = gpa; 2717d665d229SNeel Natu copyinfo[nused].len = n; 2718d665d229SNeel Natu remaining -= n; 2719d665d229SNeel Natu gla += n; 2720d665d229SNeel Natu nused++; 2721d665d229SNeel Natu } 2722d665d229SNeel Natu 2723d665d229SNeel Natu for (idx = 0; idx < nused; idx++) { 2724d3956e46SJohn Baldwin hva = vm_gpa_hold(vcpu, copyinfo[idx].gpa, 27259b1aa8d6SNeel Natu copyinfo[idx].len, prot, &cookie); 2726d665d229SNeel Natu if (hva == NULL) 2727d665d229SNeel Natu break; 2728d665d229SNeel Natu copyinfo[idx].hva = hva; 2729d665d229SNeel Natu copyinfo[idx].cookie = cookie; 2730d665d229SNeel Natu } 2731d665d229SNeel Natu 2732d665d229SNeel Natu if (idx != nused) { 27332b4fe856SJohn Baldwin vm_copy_teardown(copyinfo, num_copyinfo); 27349c4d5478SNeel Natu return (EFAULT); 2735d665d229SNeel Natu } else { 27369c4d5478SNeel Natu *fault = 0; 2737d665d229SNeel Natu return (0); 2738d665d229SNeel Natu } 2739d665d229SNeel Natu } 2740d665d229SNeel Natu 2741d665d229SNeel Natu void 27422b4fe856SJohn Baldwin vm_copyin(struct vm_copyinfo *copyinfo, void *kaddr, size_t len) 2743d665d229SNeel Natu { 2744d665d229SNeel Natu char *dst; 2745d665d229SNeel Natu int idx; 2746d665d229SNeel Natu 2747d665d229SNeel Natu dst = kaddr; 2748d665d229SNeel Natu idx = 0; 2749d665d229SNeel Natu while (len > 0) { 2750d665d229SNeel Natu bcopy(copyinfo[idx].hva, dst, copyinfo[idx].len); 2751d665d229SNeel Natu len -= copyinfo[idx].len; 2752d665d229SNeel Natu dst += copyinfo[idx].len; 2753d665d229SNeel Natu idx++; 2754d665d229SNeel Natu } 2755d665d229SNeel Natu } 2756d665d229SNeel Natu 2757d665d229SNeel Natu void 27582b4fe856SJohn Baldwin vm_copyout(const void *kaddr, struct vm_copyinfo *copyinfo, size_t len) 2759d665d229SNeel Natu { 2760d665d229SNeel Natu const char *src; 2761d665d229SNeel Natu int idx; 2762d665d229SNeel Natu 2763d665d229SNeel Natu src = kaddr; 2764d665d229SNeel Natu idx = 0; 2765d665d229SNeel Natu while (len > 0) { 2766d665d229SNeel Natu bcopy(src, copyinfo[idx].hva, copyinfo[idx].len); 2767d665d229SNeel Natu len -= copyinfo[idx].len; 2768d665d229SNeel Natu src += copyinfo[idx].len; 2769d665d229SNeel Natu idx++; 2770d665d229SNeel Natu } 2771d665d229SNeel Natu } 2772cf1d80d8SPeter Grehan 2773cf1d80d8SPeter Grehan /* 2774cf1d80d8SPeter Grehan * Return the amount of in-use and wired memory for the VM. Since 2775cf1d80d8SPeter Grehan * these are global stats, only return the values with for vCPU 0 2776cf1d80d8SPeter Grehan */ 2777cf1d80d8SPeter Grehan VMM_STAT_DECLARE(VMM_MEM_RESIDENT); 2778cf1d80d8SPeter Grehan VMM_STAT_DECLARE(VMM_MEM_WIRED); 2779cf1d80d8SPeter Grehan 2780cf1d80d8SPeter Grehan static void 27813f0f4b15SJohn Baldwin vm_get_rescnt(struct vcpu *vcpu, struct vmm_stat_type *stat) 2782cf1d80d8SPeter Grehan { 2783cf1d80d8SPeter Grehan 27843f0f4b15SJohn Baldwin if (vcpu->vcpuid == 0) { 27853f0f4b15SJohn Baldwin vmm_stat_set(vcpu, VMM_MEM_RESIDENT, PAGE_SIZE * 27863f0f4b15SJohn Baldwin vmspace_resident_count(vcpu->vm->vmspace)); 2787cf1d80d8SPeter Grehan } 2788cf1d80d8SPeter Grehan } 2789cf1d80d8SPeter Grehan 2790cf1d80d8SPeter Grehan static void 27913f0f4b15SJohn Baldwin vm_get_wiredcnt(struct vcpu *vcpu, struct vmm_stat_type *stat) 2792cf1d80d8SPeter Grehan { 2793cf1d80d8SPeter Grehan 27943f0f4b15SJohn Baldwin if (vcpu->vcpuid == 0) { 27953f0f4b15SJohn Baldwin vmm_stat_set(vcpu, VMM_MEM_WIRED, PAGE_SIZE * 27963f0f4b15SJohn Baldwin pmap_wired_count(vmspace_pmap(vcpu->vm->vmspace))); 2797cf1d80d8SPeter Grehan } 2798cf1d80d8SPeter Grehan } 2799cf1d80d8SPeter Grehan 2800cf1d80d8SPeter Grehan VMM_STAT_FUNC(VMM_MEM_RESIDENT, "Resident memory", vm_get_rescnt); 2801cf1d80d8SPeter Grehan VMM_STAT_FUNC(VMM_MEM_WIRED, "Wired memory", vm_get_wiredcnt); 2802483d953aSJohn Baldwin 2803483d953aSJohn Baldwin #ifdef BHYVE_SNAPSHOT 2804483d953aSJohn Baldwin static int 2805483d953aSJohn Baldwin vm_snapshot_vcpus(struct vm *vm, struct vm_snapshot_meta *meta) 2806483d953aSJohn Baldwin { 2807a7db532eSJohn Baldwin uint64_t tsc, now; 2808483d953aSJohn Baldwin int ret; 2809483d953aSJohn Baldwin struct vcpu *vcpu; 281035abc6c2SJohn Baldwin uint16_t i, maxcpus; 2811483d953aSJohn Baldwin 2812a7db532eSJohn Baldwin now = rdtsc(); 281335abc6c2SJohn Baldwin maxcpus = vm_get_maxcpus(vm); 281435abc6c2SJohn Baldwin for (i = 0; i < maxcpus; i++) { 2815*98568a00SJohn Baldwin vcpu = vm->vcpu[i]; 2816*98568a00SJohn Baldwin if (vcpu == NULL) 2817*98568a00SJohn Baldwin continue; 2818483d953aSJohn Baldwin 2819483d953aSJohn Baldwin SNAPSHOT_VAR_OR_LEAVE(vcpu->x2apic_state, meta, ret, done); 2820483d953aSJohn Baldwin SNAPSHOT_VAR_OR_LEAVE(vcpu->exitintinfo, meta, ret, done); 2821483d953aSJohn Baldwin SNAPSHOT_VAR_OR_LEAVE(vcpu->exc_vector, meta, ret, done); 2822483d953aSJohn Baldwin SNAPSHOT_VAR_OR_LEAVE(vcpu->exc_errcode_valid, meta, ret, done); 2823483d953aSJohn Baldwin SNAPSHOT_VAR_OR_LEAVE(vcpu->exc_errcode, meta, ret, done); 2824483d953aSJohn Baldwin SNAPSHOT_VAR_OR_LEAVE(vcpu->guest_xcr0, meta, ret, done); 2825483d953aSJohn Baldwin SNAPSHOT_VAR_OR_LEAVE(vcpu->exitinfo, meta, ret, done); 2826483d953aSJohn Baldwin SNAPSHOT_VAR_OR_LEAVE(vcpu->nextrip, meta, ret, done); 2827a7db532eSJohn Baldwin 2828a7db532eSJohn Baldwin /* 2829a7db532eSJohn Baldwin * Save the absolute TSC value by adding now to tsc_offset. 2830483d953aSJohn Baldwin * 2831483d953aSJohn Baldwin * It will be turned turned back into an actual offset when the 2832483d953aSJohn Baldwin * TSC restore function is called 2833483d953aSJohn Baldwin */ 2834a7db532eSJohn Baldwin tsc = now + vcpu->tsc_offset; 2835a7db532eSJohn Baldwin SNAPSHOT_VAR_OR_LEAVE(tsc, meta, ret, done); 2836483d953aSJohn Baldwin } 2837483d953aSJohn Baldwin 2838483d953aSJohn Baldwin done: 2839483d953aSJohn Baldwin return (ret); 2840483d953aSJohn Baldwin } 2841483d953aSJohn Baldwin 2842483d953aSJohn Baldwin static int 2843483d953aSJohn Baldwin vm_snapshot_vm(struct vm *vm, struct vm_snapshot_meta *meta) 2844483d953aSJohn Baldwin { 2845483d953aSJohn Baldwin int ret; 2846483d953aSJohn Baldwin 2847483d953aSJohn Baldwin ret = vm_snapshot_vcpus(vm, meta); 2848a7db532eSJohn Baldwin if (ret != 0) 2849483d953aSJohn Baldwin goto done; 2850483d953aSJohn Baldwin 2851c0f35dbfSJohn Baldwin SNAPSHOT_VAR_OR_LEAVE(vm->startup_cpus, meta, ret, done); 2852483d953aSJohn Baldwin done: 2853483d953aSJohn Baldwin return (ret); 2854483d953aSJohn Baldwin } 2855483d953aSJohn Baldwin 2856483d953aSJohn Baldwin static int 28571aa51504SJohn Baldwin vm_snapshot_vcpu(struct vm *vm, struct vm_snapshot_meta *meta) 2858483d953aSJohn Baldwin { 285935abc6c2SJohn Baldwin int error; 28601aa51504SJohn Baldwin struct vcpu *vcpu; 286135abc6c2SJohn Baldwin uint16_t i, maxcpus; 2862483d953aSJohn Baldwin 2863483d953aSJohn Baldwin error = 0; 2864483d953aSJohn Baldwin 286535abc6c2SJohn Baldwin maxcpus = vm_get_maxcpus(vm); 286635abc6c2SJohn Baldwin for (i = 0; i < maxcpus; i++) { 2867*98568a00SJohn Baldwin vcpu = vm->vcpu[i]; 2868*98568a00SJohn Baldwin if (vcpu == NULL) 2869*98568a00SJohn Baldwin continue; 28701aa51504SJohn Baldwin 2871869c8d19SJohn Baldwin error = vmmops_vcpu_snapshot(vcpu->cookie, meta); 2872483d953aSJohn Baldwin if (error != 0) { 2873483d953aSJohn Baldwin printf("%s: failed to snapshot vmcs/vmcb data for " 2874483d953aSJohn Baldwin "vCPU: %d; error: %d\n", __func__, i, error); 2875483d953aSJohn Baldwin goto done; 2876483d953aSJohn Baldwin } 2877483d953aSJohn Baldwin } 2878483d953aSJohn Baldwin 2879483d953aSJohn Baldwin done: 2880483d953aSJohn Baldwin return (error); 2881483d953aSJohn Baldwin } 2882483d953aSJohn Baldwin 2883483d953aSJohn Baldwin /* 2884483d953aSJohn Baldwin * Save kernel-side structures to user-space for snapshotting. 2885483d953aSJohn Baldwin */ 2886483d953aSJohn Baldwin int 2887483d953aSJohn Baldwin vm_snapshot_req(struct vm *vm, struct vm_snapshot_meta *meta) 2888483d953aSJohn Baldwin { 2889483d953aSJohn Baldwin int ret = 0; 2890483d953aSJohn Baldwin 2891483d953aSJohn Baldwin switch (meta->dev_req) { 2892483d953aSJohn Baldwin case STRUCT_VMX: 289315add60dSPeter Grehan ret = vmmops_snapshot(vm->cookie, meta); 2894483d953aSJohn Baldwin break; 2895483d953aSJohn Baldwin case STRUCT_VMCX: 28961aa51504SJohn Baldwin ret = vm_snapshot_vcpu(vm, meta); 2897483d953aSJohn Baldwin break; 2898483d953aSJohn Baldwin case STRUCT_VM: 2899483d953aSJohn Baldwin ret = vm_snapshot_vm(vm, meta); 2900483d953aSJohn Baldwin break; 2901483d953aSJohn Baldwin case STRUCT_VIOAPIC: 2902483d953aSJohn Baldwin ret = vioapic_snapshot(vm_ioapic(vm), meta); 2903483d953aSJohn Baldwin break; 2904483d953aSJohn Baldwin case STRUCT_VLAPIC: 2905483d953aSJohn Baldwin ret = vlapic_snapshot(vm, meta); 2906483d953aSJohn Baldwin break; 2907483d953aSJohn Baldwin case STRUCT_VHPET: 2908483d953aSJohn Baldwin ret = vhpet_snapshot(vm_hpet(vm), meta); 2909483d953aSJohn Baldwin break; 2910483d953aSJohn Baldwin case STRUCT_VATPIC: 2911483d953aSJohn Baldwin ret = vatpic_snapshot(vm_atpic(vm), meta); 2912483d953aSJohn Baldwin break; 2913483d953aSJohn Baldwin case STRUCT_VATPIT: 2914483d953aSJohn Baldwin ret = vatpit_snapshot(vm_atpit(vm), meta); 2915483d953aSJohn Baldwin break; 2916483d953aSJohn Baldwin case STRUCT_VPMTMR: 2917483d953aSJohn Baldwin ret = vpmtmr_snapshot(vm_pmtmr(vm), meta); 2918483d953aSJohn Baldwin break; 2919483d953aSJohn Baldwin case STRUCT_VRTC: 2920483d953aSJohn Baldwin ret = vrtc_snapshot(vm_rtc(vm), meta); 2921483d953aSJohn Baldwin break; 2922483d953aSJohn Baldwin default: 2923483d953aSJohn Baldwin printf("%s: failed to find the requested type %#x\n", 2924483d953aSJohn Baldwin __func__, meta->dev_req); 2925483d953aSJohn Baldwin ret = (EINVAL); 2926483d953aSJohn Baldwin } 2927483d953aSJohn Baldwin return (ret); 2928483d953aSJohn Baldwin } 2929483d953aSJohn Baldwin 293080cb5d84SJohn Baldwin void 293180cb5d84SJohn Baldwin vm_set_tsc_offset(struct vcpu *vcpu, uint64_t offset) 2932483d953aSJohn Baldwin { 2933483d953aSJohn Baldwin vcpu->tsc_offset = offset; 2934483d953aSJohn Baldwin } 2935483d953aSJohn Baldwin 2936483d953aSJohn Baldwin int 2937483d953aSJohn Baldwin vm_restore_time(struct vm *vm) 2938483d953aSJohn Baldwin { 293935abc6c2SJohn Baldwin int error; 2940483d953aSJohn Baldwin uint64_t now; 2941483d953aSJohn Baldwin struct vcpu *vcpu; 294235abc6c2SJohn Baldwin uint16_t i, maxcpus; 2943483d953aSJohn Baldwin 2944483d953aSJohn Baldwin now = rdtsc(); 2945483d953aSJohn Baldwin 2946483d953aSJohn Baldwin error = vhpet_restore_time(vm_hpet(vm)); 2947483d953aSJohn Baldwin if (error) 2948483d953aSJohn Baldwin return (error); 2949483d953aSJohn Baldwin 295035abc6c2SJohn Baldwin maxcpus = vm_get_maxcpus(vm); 295135abc6c2SJohn Baldwin for (i = 0; i < maxcpus; i++) { 2952*98568a00SJohn Baldwin vcpu = vm->vcpu[i]; 2953*98568a00SJohn Baldwin if (vcpu == NULL) 2954*98568a00SJohn Baldwin continue; 2955483d953aSJohn Baldwin 2956869c8d19SJohn Baldwin error = vmmops_restore_tsc(vcpu->cookie, 29571aa51504SJohn Baldwin vcpu->tsc_offset - now); 2958483d953aSJohn Baldwin if (error) 2959483d953aSJohn Baldwin return (error); 2960483d953aSJohn Baldwin } 2961483d953aSJohn Baldwin 2962483d953aSJohn Baldwin return (0); 2963483d953aSJohn Baldwin } 2964483d953aSJohn Baldwin #endif 2965