Lines Matching refs:fault
303 int kvm_tdp_page_fault(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault);
343 struct kvm_page_fault *fault) in kvm_mmu_prepare_memory_fault_exit() argument
345 kvm_prepare_memory_fault_exit(vcpu, fault->gfn << PAGE_SHIFT, in kvm_mmu_prepare_memory_fault_exit()
346 PAGE_SIZE, fault->write, fault->exec, in kvm_mmu_prepare_memory_fault_exit()
347 fault->is_private); in kvm_mmu_prepare_memory_fault_exit()
354 struct kvm_page_fault fault = { in kvm_mmu_do_page_fault() local
382 fault.gfn = gpa_to_gfn(fault.addr) & ~kvm_gfn_direct_bits(vcpu->kvm); in kvm_mmu_do_page_fault()
383 fault.slot = kvm_vcpu_gfn_to_memslot(vcpu, fault.gfn); in kvm_mmu_do_page_fault()
390 if (IS_ENABLED(CONFIG_MITIGATION_RETPOLINE) && fault.is_tdp) in kvm_mmu_do_page_fault()
391 r = kvm_tdp_page_fault(vcpu, &fault); in kvm_mmu_do_page_fault()
393 r = vcpu->arch.mmu->page_fault(vcpu, &fault); in kvm_mmu_do_page_fault()
400 if (r == RET_PF_EMULATE && fault.is_private) { in kvm_mmu_do_page_fault()
402 kvm_mmu_prepare_memory_fault_exit(vcpu, &fault); in kvm_mmu_do_page_fault()
406 if (fault.write_fault_to_shadow_pgtable && emulation_type) in kvm_mmu_do_page_fault()
409 *level = fault.goal_level; in kvm_mmu_do_page_fault()
416 void kvm_mmu_hugepage_adjust(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault);
417 void disallowed_hugepage_adjust(struct kvm_page_fault *fault, u64 spte, int cur_level);