Searched refs:hmm_range (Results 1 – 10 of 10) sorted by relevance
177 struct hmm_range *hmm_range = &range->hmm_range; in amdgpu_hmm_range_get_pages() local185 hmm_range->notifier = notifier; in amdgpu_hmm_range_get_pages()186 hmm_range->default_flags = HMM_PFN_REQ_FAULT; in amdgpu_hmm_range_get_pages()188 hmm_range->default_flags |= HMM_PFN_REQ_WRITE; in amdgpu_hmm_range_get_pages()189 hmm_range->hmm_pfns = pfns; in amdgpu_hmm_range_get_pages()190 hmm_range->start = start; in amdgpu_hmm_range_get_pages()192 hmm_range->dev_private_owner = owner; in amdgpu_hmm_range_get_pages()195 hmm_range->end = min(hmm_range->start + MAX_WALK_BYTE, end); in amdgpu_hmm_range_get_pages()198 hmm_range->start, hmm_range->end); in amdgpu_hmm_range_get_pages()203 hmm_range->notifier_seq = mmu_interval_read_begin(notifier); in amdgpu_hmm_range_get_pages()[all …]
784 ttm->pages[i] = range ? hmm_pfn_to_page(range->hmm_range.hmm_pfns[i]) : NULL; in amdgpu_ttm_tt_set_user_pages()
696 struct hmm_range hmm_range = { in drm_gpusvm_check_pages() local715 hmm_range.notifier_seq = mmu_interval_read_begin(¬ifier->notifier); in drm_gpusvm_check_pages()716 hmm_range.hmm_pfns = pfns; in drm_gpusvm_check_pages()719 err = hmm_range_fault(&hmm_range); in drm_gpusvm_check_pages()724 hmm_range.notifier_seq = in drm_gpusvm_check_pages()769 struct hmm_range hmm_range = { in drm_gpusvm_scan_mm() local788 hmm_range.hmm_pfns = pfns; in drm_gpusvm_scan_mm()791 hmm_range.notifier_seq = mmu_interval_read_begin(notifier); in drm_gpusvm_scan_mm()795 err = hmm_range_fault(&hmm_range); in drm_gpusvm_scan_mm()800 hmm_range.notifier_seq = in drm_gpusvm_scan_mm()[all …]
34 struct hmm_range *range;51 struct hmm_range *range, unsigned long cpu_flags) in hmm_pfns_fill()100 struct hmm_range *range = hmm_vma_walk->range; in hmm_pte_need_fault()135 struct hmm_range *range = hmm_vma_walk->range; in hmm_range_need_fault()161 struct hmm_range *range = hmm_vma_walk->range; in hmm_vma_walk_hole()187 static inline unsigned long pmd_to_hmm_pfn_flags(struct hmm_range *range, in pmd_to_hmm_pfn_flags()202 struct hmm_range *range = hmm_vma_walk->range; in hmm_vma_handle_pmd()227 static inline unsigned long pte_to_hmm_pfn_flags(struct hmm_range *range, in pte_to_hmm_pfn_flags()240 struct hmm_range *range = hmm_vma_walk->range; in hmm_vma_handle_pte()340 struct hmm_range *range = hmm_vma_walk->range; in hmm_vma_handle_absent_pmd()[all …]
111 struct hmm_range { struct125 int hmm_range_fault(struct hmm_range *range); argument
29 struct hmm_range;
538 struct hmm_range *range, in nouveau_hmm_convert_pfn()662 struct hmm_range range = { in nouveau_range_fault()
405 struct hmm_range *range) in mshv_region_hmm_fault_and_lock()442 struct hmm_range range = { in mshv_region_range_fault()
333 struct hmm_range range = {}; in ib_umem_odp_map_dma_and_lock()
1821 range->hmm_range.hmm_pfns); in svm_range_validate_and_map()