Lines Matching refs:mmu
718 vmm_vcpu_mmu_mode(vmm_t *vmm, int vcpuid __unused, vmm_mmu_t *mmu) in vmm_vcpu_mmu_mode() argument
720 if ((mmu->vm_cr0 & CR0_PE) == 0) in vmm_vcpu_mmu_mode()
722 else if ((mmu->vm_cr4 & CR4_PAE) == 0) in vmm_vcpu_mmu_mode()
724 else if ((mmu->vm_efer & AMD_EFER_LME) == 0) in vmm_vcpu_mmu_mode()
733 vmm_mmu_t mmu = { 0 }; in vmm_vcpu_mode() local
736 vmm_mmu_regnum, (uint64_t *)&mmu) != 0) in vmm_vcpu_mode()
739 return (vmm_vcpu_mmu_mode(vmm, vcpuid, &mmu)); in vmm_vcpu_mode()
817 vmm_mmu_t mmu = { 0 }; in vmm_vtop() local
824 vmm_mmu_regnum, (uint64_t *)&mmu) != 0) in vmm_vtop()
827 if ((mmu.vm_cr0 & CR0_PG) == 0) { in vmm_vtop()
833 switch (vmm_vcpu_mmu_mode(vmm, vcpuid, &mmu)) { in vmm_vtop()
836 ret = vmm_pte2paddr(vmm, mmu.vm_cr3, B_TRUE, 2, vaddr, paddr); in vmm_vtop()
840 ret = vmm_pte2paddr(vmm, mmu.vm_cr3, B_FALSE, 3, vaddr, paddr); in vmm_vtop()
844 ret = vmm_pte2paddr(vmm, mmu.vm_cr3, B_FALSE, 4, vaddr, paddr); in vmm_vtop()