Searched refs:uvma (Results 1 – 3 of 3) sorted by relevance
181 nouveau_uvma_vmm_put(struct nouveau_uvma *uvma) in nouveau_uvma_vmm_put() argument183 u64 addr = uvma->va.va.addr; in nouveau_uvma_vmm_put()184 u64 range = uvma->va.va.range; in nouveau_uvma_vmm_put()185 u8 page_shift = uvma->page_shift; in nouveau_uvma_vmm_put()187 return nouveau_uvmm_vmm_put(to_uvmm(uvma), addr, range, page_shift); in nouveau_uvma_vmm_put()191 nouveau_uvma_map(struct nouveau_uvma *uvma, in nouveau_uvma_map() argument194 u64 addr = uvma->va.va.addr; in nouveau_uvma_map()195 u64 offset = uvma->va.gem.offset; in nouveau_uvma_map()196 u64 range = uvma->va.va.range; in nouveau_uvma_map()197 u8 page_shift = uvma->page_shift; in nouveau_uvma_map()[all …]
1377 static bool xe_pt_userptr_inject_eagain(struct xe_userptr_vma *uvma) in xe_pt_userptr_inject_eagain() argument1379 u32 divisor = uvma->userptr.divisor ? uvma->userptr.divisor : 2; in xe_pt_userptr_inject_eagain()1383 uvma->userptr.divisor = divisor << 1; in xe_pt_userptr_inject_eagain()1392 static bool xe_pt_userptr_inject_eagain(struct xe_userptr_vma *uvma) in xe_pt_userptr_inject_eagain() argument1402 struct xe_userptr_vma *uvma; in vma_check_userptr() local1410 uvma = to_userptr_vma(vma); in vma_check_userptr()1411 if (xe_pt_userptr_inject_eagain(uvma)) in vma_check_userptr()1412 xe_vma_userptr_force_invalidate(uvma); in vma_check_userptr()1414 notifier_seq = uvma->userptr.pages.notifier_seq; in vma_check_userptr()1416 if (!mmu_interval_read_retry(&uvma->userptr.notifier, in vma_check_userptr()
1011 struct xe_userptr_vma *uvma = kzalloc_obj(*uvma); in xe_vma_create() local1013 if (!uvma) in xe_vma_create()1016 vma = &uvma->vma; in xe_vma_create()1059 struct xe_userptr_vma *uvma = to_userptr_vma(vma); in xe_vma_create() local1065 err = xe_userptr_setup(uvma, xe_vma_userptr(vma), size); in xe_vma_create()1088 struct xe_userptr_vma *uvma = to_userptr_vma(vma); in xe_vma_destroy_late() local1090 xe_userptr_remove(uvma); in xe_vma_destroy_late()