Lines Matching defs:gpuva
332 struct drm_gpuva *gpuva;
336 drm_gpuvm_bo_for_each_va(gpuva, vm_bo)
337 list_move_tail(&gpuva_to_vma(gpuva)->combined_links.rebind,
733 op->base.map.va.addr = vma->gpuva.va.addr;
734 op->base.map.va.range = vma->gpuva.va.range;
735 op->base.map.gem.obj = vma->gpuva.gem.obj;
736 op->base.map.gem.offset = vma->gpuva.gem.offset;
739 op->map.vma_flags = vma->gpuva.flags & XE_VMA_CREATE_MASK;
1100 vma->gpuva.gem.obj = &bo->ttm.base;
1105 INIT_LIST_HEAD(&vma->gpuva.gem.entry);
1106 vma->gpuva.vm = &vm->gpuvm;
1107 vma->gpuva.va.addr = start;
1108 vma->gpuva.va.range = end - start + 1;
1109 vma->gpuva.flags = flags;
1115 vma->gpuva.flags |= XE_VMA_ATOMIC_PTE_BIT;
1123 vm_bo = drm_gpuvm_bo_obtain_locked(vma->gpuva.vm, &bo->ttm.base);
1131 vma->gpuva.gem.offset = bo_offset_or_userptr;
1132 drm_gpuva_link(&vma->gpuva, vm_bo);
1140 vma->gpuva.gem.offset = bo_offset_or_userptr;
1205 xe_assert(vm->xe, vma->gpuva.flags & XE_VMA_DESTROYED);
1210 drm_gpuva_unlink(&vma->gpuva);
1273 struct drm_gpuva *gpuva;
1282 gpuva = drm_gpuva_find_first(&vm->gpuvm, start, range);
1284 return gpuva ? gpuva_to_vma(gpuva) : NULL;
1295 err = drm_gpuva_insert(&vm->gpuvm, &vma->gpuva);
1308 drm_gpuva_remove(&vma->gpuva);
1852 struct drm_gpuva *gpuva, *next;
1889 drm_gpuvm_for_each_va_safe(gpuva, next, &vm->gpuvm) {
1890 vma = gpuva_to_vma(gpuva);
1894 vma->gpuva.flags |= XE_VMA_DESTROYED;
1908 vma->gpuva.flags |= XE_VMA_DESTROYED;
2146 struct drm_gpuva *gpuva;
2150 drm_gpuvm_for_each_va_range(gpuva, &vm->gpuvm, start, end)
2159 struct drm_gpuva *gpuva;
2164 drm_gpuvm_for_each_va_range(gpuva, &vm->gpuvm, start, end) {
2165 struct xe_vma *vma = gpuva_to_vma(gpuva);
2287 vma->gpuva.flags |= XE_VMA_DESTROYED;
2623 if (vma->gpuva.flags & XE_VMA_PTE_1G)
2625 else if (vma->gpuva.flags & (XE_VMA_PTE_2M | XE_VMA_PTE_COMPACT))
2627 else if (vma->gpuva.flags & XE_VMA_PTE_64K)
2629 else if (vma->gpuva.flags & XE_VMA_PTE_4K)
2639 vma->gpuva.flags |= XE_VMA_PTE_1G;
2642 vma->gpuva.flags |= XE_VMA_PTE_2M;
2645 vma->gpuva.flags |= XE_VMA_PTE_64K;
2648 vma->gpuva.flags |= XE_VMA_PTE_4K;
2943 vma->gpuva.flags &= ~XE_VMA_DESTROYED;
2964 vma->gpuva.flags &= ~XE_VMA_DESTROYED;
4371 struct drm_gpuva *gpuva;
4379 drm_gpuvm_for_each_va(gpuva, &vm->gpuvm) {
4380 struct xe_vma *vma = gpuva_to_vma(gpuva);
4381 struct xe_bo *bo = vma->gpuva.gem.obj ?
4382 gem_to_xe_bo(vma->gpuva.gem.obj) : NULL;
4421 struct drm_gpuva *gpuva;
4427 drm_gpuvm_for_each_va(gpuva, &vm->gpuvm) {
4428 if (gpuva->flags & XE_VMA_DUMPABLE)
4448 drm_gpuvm_for_each_va(gpuva, &vm->gpuvm) {
4449 struct xe_vma *vma = gpuva_to_vma(gpuva);
4450 struct xe_bo *bo = vma->gpuva.gem.obj ?
4451 gem_to_xe_bo(vma->gpuva.gem.obj) : NULL;
4453 if (!(gpuva->flags & XE_VMA_DUMPABLE))
4698 vma_flags = vma->gpuva.flags;
4704 vma_flags = vma->gpuva.flags;
4717 vma_flags = vma->gpuva.flags;