Lines Matching refs:svm_range
2342 struct xe_svm_range *svm_range; in vm_bind_ioctl_ops_create() local
2374 svm_range = xe_svm_range_find_or_insert(vm, addr, vma, &ctx); in vm_bind_ioctl_ops_create()
2376 if (PTR_ERR(svm_range) == -ENOENT) { in vm_bind_ioctl_ops_create()
2386 if (IS_ERR(svm_range)) { in vm_bind_ioctl_ops_create()
2387 err = PTR_ERR(svm_range); in vm_bind_ioctl_ops_create()
2391 if (xe_svm_range_validate(vm, svm_range, tile_mask, dpagemap)) { in vm_bind_ioctl_ops_create()
2392 xe_svm_range_debug(svm_range, "PREFETCH - RANGE IS VALID"); in vm_bind_ioctl_ops_create()
2397 &i, svm_range, xa_limit_32b, in vm_bind_ioctl_ops_create()
2405 xe_svm_range_debug(svm_range, "PREFETCH - RANGE CREATED"); in vm_bind_ioctl_ops_create()
2407 if (range_end > xe_svm_range_end(svm_range) && in vm_bind_ioctl_ops_create()
2408 xe_svm_range_end(svm_range) < xe_vma_end(vma)) { in vm_bind_ioctl_ops_create()
2409 addr = xe_svm_range_end(svm_range); in vm_bind_ioctl_ops_create()
2916 struct xe_svm_range *svm_range; in prefetch_ranges() local
2929 xa_for_each(&op->prefetch_range.range, i, svm_range) { in prefetch_ranges()
2931 xe_svm_range_migrate_to_smem(vm, svm_range); in prefetch_ranges()
2937 xe_svm_range_start(svm_range), xe_svm_range_end(svm_range)); in prefetch_ranges()
2940 if (xe_svm_range_needs_migrate_to_vram(svm_range, vma, dpagemap)) { in prefetch_ranges()
2941 err = xe_svm_alloc_vram(svm_range, &ctx, dpagemap); in prefetch_ranges()
2947 xe_svm_range_debug(svm_range, "PREFETCH - RANGE MIGRATED TO VRAM"); in prefetch_ranges()
2950 err = xe_svm_range_get_pages(vm, svm_range, &ctx); in prefetch_ranges()
2958 xe_svm_range_debug(svm_range, "PREFETCH - RANGE GET PAGES DONE"); in prefetch_ranges()