Searched refs:root_to_sp (Results 1 – 6 of 6) sorted by relevance
52 return root_to_sp(vcpu->arch.mmu->mirror_root_hpa); in tdp_mmu_get_root_for_fault()54 return root_to_sp(vcpu->arch.mmu->root.hpa); in tdp_mmu_get_root_for_fault()61 return root_to_sp(vcpu->arch.mmu->mirror_root_hpa); in tdp_mmu_get_root()63 return root_to_sp(vcpu->arch.mmu->root.hpa); in tdp_mmu_get_root()
276 static inline struct kvm_mmu_page *root_to_sp(hpa_t root) in root_to_sp() function295 struct kvm_mmu_page *root = root_to_sp(vcpu->arch.mmu->root.hpa); in kvm_vcpu_can_access_host_mmio()
143 struct kvm_mmu_page *root = root_to_sp(vcpu->arch.mmu->root.hpa); in kvm_track_host_mmio_mapping()
3788 sp = root_to_sp(*root_hpa); in mmu_free_root_page()3844 } else if (root_to_sp(mmu->root.hpa)) { in kvm_mmu_free_roots()3888 sp = root_to_sp(root_hpa); in kvm_mmu_free_guest_mode_roots()4264 sp = root_to_sp(root); in is_unsync_root()4298 sp = root_to_sp(root); in kvm_mmu_sync_roots()4762 struct kvm_mmu_page *sp = root_to_sp(vcpu->arch.mmu->root.hpa); in is_page_fault_stale()5115 sp = root_to_sp(root->hpa); in is_root_usable()5192 if (VALID_PAGE(mmu->root.hpa) && !root_to_sp(mmu->root.hpa)) in fast_pgd_switch()5239 struct kvm_mmu_page *sp = root_to_sp(vcpu->arch.mmu->root.hpa); in kvm_mmu_new_pgd()6115 sp = root_to_sp(root_hpa); in is_obsolete_root()
1942 struct kvm_mmu_page *root = root_to_sp(vcpu->arch.mmu->root.hpa); in kvm_tdp_mmu_get_walk()
3344 root = root_to_sp(root_hpa); in construct_eptp()3549 KVM_MMU_WARN_ON(root_to_sp(root_hpa) && in vmx_load_mmu_pgd()3550 root_level != root_to_sp(root_hpa)->role.level); in vmx_load_mmu_pgd()