Lines Matching defs:sync

500 static int vm_update_pds(struct amdgpu_vm *vm, struct amdgpu_sync *sync)
510 return amdgpu_sync_fence(sync, vm->last_update, GFP_KERNEL);
1141 * all the reservation info in a context structure. Optionally a sync
1149 /* Pointer to sync object */
1150 struct amdgpu_sync *sync;
1175 ctx->sync = &mem->sync;
1214 ctx->sync = &mem->sync;
1249 * @wait: Optionally wait for a sync object representing pending VM updates
1262 ret = amdgpu_sync_wait(ctx->sync, intr);
1265 ctx->sync = NULL;
1271 struct amdgpu_sync *sync)
1290 (void)amdgpu_sync_fence(sync, bo_va->last_pt_update, GFP_KERNEL);
1297 struct amdgpu_sync *sync)
1314 return amdgpu_sync_fence(sync, bo_va->last_pt_update, GFP_KERNEL);
1319 struct amdgpu_sync *sync,
1337 ret = update_gpuvm_pte(mem, entry, sync);
1346 unmap_bo_from_gpuvm(mem, entry, sync);
1368 struct amdgpu_sync *sync)
1377 ret = amdgpu_sync_resv(NULL, sync, pd->tbo.base.resv,
1388 struct amdgpu_sync *sync)
1395 ret = vm_update_pds(peer_vm, sync);
1795 amdgpu_sync_create(&(*mem)->sync);
1883 amdgpu_sync_free(&(*mem)->sync);
1962 /* Free the sync object */
1963 amdgpu_sync_free(&mem->sync);
2086 ret = map_bo_to_gpuvm(mem, entry, ctx.sync,
2093 ret = vm_update_pds(avm, ctx.sync);
2184 ret = unmap_bo_from_gpuvm(mem, entry, ctx.sync);
2205 struct amdgpu_sync sync;
2208 amdgpu_sync_create(&sync);
2211 amdgpu_sync_clone(&mem->sync, &sync);
2214 ret = amdgpu_sync_wait(&sync, intr);
2215 amdgpu_sync_free(&sync);
2411 amdgpu_sync_create(&(*mem)->sync);
2660 struct amdgpu_sync sync;
2668 amdgpu_sync_create(&sync);
2728 ret = update_gpuvm_pte(mem, attachment, &sync);
2741 ret = process_update_pds(process_info, &sync);
2745 amdgpu_sync_wait(&sync, false);
2746 amdgpu_sync_free(&sync);
3059 pr_debug("Memory eviction: Failed to sync to PD BO moving fence. Try again\n");
3142 amdgpu_sync_create(&(*mem)->sync);
3178 amdgpu_sync_free(&(*mem)->sync);
3208 amdgpu_sync_free(&kgd_mem->sync);