/linux/drivers/gpu/drm/xe/ |
H A D | xe_vm.h | 88 static inline struct xe_vm *gpuva_to_vm(struct drm_gpuva *gpuva) in gpuva_to_vm() argument 90 return gpuvm_to_vm(gpuva->vm); in gpuva_to_vm() 93 static inline struct xe_vma *gpuva_to_vma(struct drm_gpuva *gpuva) in gpuva_to_vma() argument 95 return container_of(gpuva, struct xe_vma, gpuva); in gpuva_to_vma() 109 return vma->gpuva.va.addr; in xe_vma_start() 114 return vma->gpuva.va.range; in xe_vma_size() 124 return vma->gpuva.gem.offset; in xe_vma_bo_offset() 129 return !vma->gpuva.gem.obj ? NULL : in xe_vma_bo() 130 container_of(vma->gpuva.gem.obj, struct xe_bo, ttm.base); in xe_vma_bo() 135 return container_of(vma->gpuva.vm, struct xe_vm, gpuvm); in xe_vma_vm() [all …]
|
H A D | xe_vm.c | 384 struct drm_gpuva *gpuva; in xe_gpuvm_validate() local 388 drm_gpuvm_bo_for_each_va(gpuva, vm_bo) in xe_gpuvm_validate() 389 list_move_tail(&gpuva_to_vma(gpuva)->combined_links.rebind, in xe_gpuvm_validate() 595 !(vma->gpuva.flags & XE_VMA_DESTROYED)) { in __vma_userptr_invalidate() 821 op->base.map.va.addr = vma->gpuva.va.addr; in xe_vm_populate_rebind() 822 op->base.map.va.range = vma->gpuva.va.range; in xe_vm_populate_rebind() 823 op->base.map.gem.obj = vma->gpuva.gem.obj; in xe_vm_populate_rebind() 824 op->base.map.gem.offset = vma->gpuva.gem.offset; in xe_vm_populate_rebind() 827 op->map.dumpable = vma->gpuva.flags & XE_VMA_DUMPABLE; in xe_vm_populate_rebind() 995 vma->gpuva.flags |= DRM_GPUVA_SPARSE; in xe_vma_create() [all …]
|
H A D | xe_hmm.c | 177 !(vma->gpuva.flags & XE_VMA_DESTROYED)) { in xe_hmm_userptr_unmap() 262 if (vma->gpuva.flags & XE_VMA_DESTROYED) in xe_hmm_userptr_populate_range()
|
H A D | xe_vm_types.h | 79 struct drm_gpuva gpuva; member
|
H A D | xe_pt.c | 205 XE_WARN_ON(!list_empty(&pt->bo->ttm.base.gpuva.list)); in xe_pt_destroy() 517 xe_walk->vma->gpuva.flags |= XE_VMA_PTE_64K; in xe_pt_stage_bind_entry() 531 xe_walk->vma->gpuva.flags |= (XE_VMA_PTE_4K << level); in xe_pt_stage_bind_entry() 573 xe_walk->vma->gpuva.flags |= XE_VMA_PTE_COMPACT; in xe_pt_stage_bind_entry() 647 if (vma->gpuva.flags & XE_VMA_ATOMIC_PTE_BIT) { in xe_pt_stage_bind()
|
H A D | xe_bo.c | 505 if (!list_empty(&bo->ttm.base.gpuva.list)) { in xe_bo_trigger_rebind() 515 struct drm_gpuva *gpuva; in xe_bo_trigger_rebind() local 542 drm_gpuvm_bo_for_each_va(gpuva, vm_bo) { in xe_bo_trigger_rebind() 543 struct xe_vma *vma = gpuva_to_vma(gpuva); in xe_bo_trigger_rebind() 1214 xe_assert(xe, list_empty(&ttm_bo->base.gpuva.list)); in xe_ttm_bo_destroy()
|
/linux/include/drm/ |
H A D | drm_gem.h | 409 } gpuva; member 589 if (!WARN((obj)->gpuva.lock_dep_map, \ 591 (obj)->gpuva.lock_dep_map = &(lock)->dep_map 593 lockdep_assert((obj)->gpuva.lock_dep_map ? \ 594 lock_is_held((obj)->gpuva.lock_dep_map) : \ 614 INIT_LIST_HEAD(&obj->gpuva.list); in drm_gem_gpuva_init() 626 list_for_each_entry(entry__, &(obj__)->gpuva.list, list.entry.gem) 640 list_for_each_entry_safe(entry__, next__, &(obj__)->gpuva.list, list.entry.gem)
|
H A D | drm_gpuvm.h | 679 struct list_head gpuva; member 772 list_for_each_entry(va__, &(vm_bo)->list.gpuva, gem.entry) 788 list_for_each_entry_safe(va__, next__, &(vm_bo)->list.gpuva, gem.entry)
|
/linux/drivers/gpu/drm/panfrost/ |
H A D | panfrost_perfcnt.c | 50 u64 gpuva; in panfrost_perfcnt_dump_locked() local 54 gpuva = pfdev->perfcnt->mapping->mmnode.start << PAGE_SHIFT; in panfrost_perfcnt_dump_locked() 55 gpu_write(pfdev, GPU_PERFCNT_BASE_LO, lower_32_bits(gpuva)); in panfrost_perfcnt_dump_locked() 56 gpu_write(pfdev, GPU_PERFCNT_BASE_HI, upper_32_bits(gpuva)); in panfrost_perfcnt_dump_locked()
|
/linux/drivers/gpu/drm/xe/display/ |
H A D | intel_fb_bo.c | 56 if (XE_IOCTL_DBG(xe, !list_empty(&bo->ttm.base.gpuva.list))) { in intel_fb_bo_framebuffer_init()
|
/linux/drivers/gpu/drm/ |
H A D | drm_gpuvm.c | 1488 INIT_LIST_HEAD(&vm_bo->list.gpuva); in drm_gpuvm_bo_create() 1619 list_add_tail(&vm_bo->list.entry.gem, &obj->gpuva.list); in drm_gpuvm_bo_obtain() 1655 list_add_tail(&__vm_bo->list.entry.gem, &obj->gpuva.list); in drm_gpuvm_bo_obtain_prealloc() 1843 list_add_tail(&va->gem.entry, &vm_bo->list.gpuva); in drm_gpuva_link()
|
/linux/drivers/gpu/drm/panthor/ |
H A D | panthor_mmu.c | 1404 struct drm_gpuva *gpuva; in panthor_vm_get_bo_for_va() local 1409 gpuva = drm_gpuva_find_first(&vm->base, va, 1); in panthor_vm_get_bo_for_va() 1410 vma = gpuva ? container_of(gpuva, struct panthor_vma, base) : NULL; in panthor_vm_get_bo_for_va()
|