Searched refs:hw_mmu (Results 1 – 9 of 9) sorted by relevance
62 if (mmu == vcpu->arch.hw_mmu || WARN_ON(mmu != host_s2_mmu)) in enter_vmid_context()65 cxt->mmu = vcpu->arch.hw_mmu; in enter_vmid_context()
528 hyp_vcpu->vcpu.arch.hw_mmu = &hyp_vm->kvm.arch.mmu; in init_pkvm_hyp_vcpu()746 .hw_mmu = &selftest_vm.kvm.arch.mmu,
1700 return BIT(selftest_vcpu->vcpu.arch.hw_mmu->pgt->ia_bits - 1); in selftest_ipa()
28 if (vcpu && mmu != vcpu->arch.hw_mmu) in enter_vmid_context()29 cxt->mmu = vcpu->arch.hw_mmu; in enter_vmid_context()
222 __load_stage2(vcpu->arch.hw_mmu, vcpu->arch.hw_mmu->arch); in kvm_vcpu_load_vhe()
787 if (!vcpu->arch.hw_mmu) in kvm_vcpu_load_hw_mmu()788 vcpu->arch.hw_mmu = &vcpu->kvm->arch.mmu; in kvm_vcpu_load_hw_mmu()790 if (!vcpu->arch.hw_mmu) { in kvm_vcpu_load_hw_mmu()792 vcpu->arch.hw_mmu = get_s2_mmu_nested(vcpu); in kvm_vcpu_load_hw_mmu()821 if (kvm_is_nested_s2_mmu(vcpu->kvm, vcpu->arch.hw_mmu)) in kvm_vcpu_put_hw_mmu()822 atomic_dec(&vcpu->arch.hw_mmu->refcnt); in kvm_vcpu_put_hw_mmu()824 vcpu->arch.hw_mmu = NULL; in kvm_vcpu_put_hw_mmu()1847 struct kvm_s2_mmu *mmu = vcpu->arch.hw_mmu; in check_nested_vcpu_requests()
1527 int min_pages = kvm_mmu_cache_min_pages(vcpu->arch.hw_mmu); in topup_mmu_memcache()1582 struct kvm_pgtable *pgt = s2fd->vcpu->arch.hw_mmu->pgt; in gmem_abort()1680 struct kvm_pgtable *pgt = vcpu->arch.hw_mmu->pgt; in pkvm_mem_abort()2000 pgt = s2fd->vcpu->arch.hw_mmu->pgt; in kvm_s2_fault_map()2120 mmu = vcpu->arch.hw_mmu; in handle_access_fault()2250 if (fault_ipa >= BIT_ULL(VTCR_EL2_IPA(vcpu->arch.hw_mmu->vtcr))) { in kvm_handle_guest_abort()2287 if (kvm_is_nested_s2_mmu(vcpu->kvm,vcpu->arch.hw_mmu) && in kvm_handle_guest_abort()2288 vcpu->arch.hw_mmu->nested_stage2_enabled) { in kvm_handle_guest_abort()2363 VM_BUG_ON(ipa >= kvm_phys_size(vcpu->arch.hw_mmu)); in kvm_handle_guest_abort()
548 vcpu->arch.hw_mmu = &vcpu->kvm->arch.mmu; in kvm_arch_vcpu_create()659 mmu = vcpu->arch.hw_mmu; in kvm_arch_vcpu_load()
878 struct kvm_s2_mmu *hw_mmu; member