Lines Matching refs:id_mgr

203 	struct amdgpu_vmid_mgr *id_mgr = &adev->vm_manager.id_mgr[vmhub];  in amdgpu_vmid_grab_idle()  local
212 fences = kmalloc_array(id_mgr->num_ids, sizeof(void *), GFP_KERNEL); in amdgpu_vmid_grab_idle()
218 list_for_each_entry((*idle), &id_mgr->ids_lru, list) { in amdgpu_vmid_grab_idle()
230 if (&(*idle)->list == &id_mgr->ids_lru) { in amdgpu_vmid_grab_idle()
278 struct amdgpu_vmid_mgr *id_mgr = &adev->vm_manager.id_mgr[vmhub]; in amdgpu_vmid_grab_reserved() local
284 *id = id_mgr->reserved; in amdgpu_vmid_grab_reserved()
355 struct amdgpu_vmid_mgr *id_mgr = &adev->vm_manager.id_mgr[vmhub]; in amdgpu_vmid_grab_used() local
363 list_for_each_entry_reverse((*id), &id_mgr->ids_lru, list) { in amdgpu_vmid_grab_used()
415 struct amdgpu_vmid_mgr *id_mgr = &adev->vm_manager.id_mgr[vmhub]; in amdgpu_vmid_grab() local
420 mutex_lock(&id_mgr->lock); in amdgpu_vmid_grab()
447 list_move_tail(&id->list, &id_mgr->ids_lru); in amdgpu_vmid_grab()
456 job->vmid = id - id_mgr->ids; in amdgpu_vmid_grab()
471 mutex_unlock(&id_mgr->lock); in amdgpu_vmid_grab()
495 struct amdgpu_vmid_mgr *id_mgr = &adev->vm_manager.id_mgr[vmhub]; in amdgpu_vmid_alloc_reserved() local
497 mutex_lock(&id_mgr->lock); in amdgpu_vmid_alloc_reserved()
499 ++id_mgr->reserved_use_count; in amdgpu_vmid_alloc_reserved()
500 if (!id_mgr->reserved) { in amdgpu_vmid_alloc_reserved()
503 id = list_first_entry(&id_mgr->ids_lru, struct amdgpu_vmid, in amdgpu_vmid_alloc_reserved()
507 id_mgr->reserved = id; in amdgpu_vmid_alloc_reserved()
510 mutex_unlock(&id_mgr->lock); in amdgpu_vmid_alloc_reserved()
517 struct amdgpu_vmid_mgr *id_mgr = &adev->vm_manager.id_mgr[vmhub]; in amdgpu_vmid_free_reserved() local
519 mutex_lock(&id_mgr->lock); in amdgpu_vmid_free_reserved()
520 if (!--id_mgr->reserved_use_count) { in amdgpu_vmid_free_reserved()
522 list_add(&id_mgr->reserved->list, &id_mgr->ids_lru); in amdgpu_vmid_free_reserved()
523 id_mgr->reserved = NULL; in amdgpu_vmid_free_reserved()
526 mutex_unlock(&id_mgr->lock); in amdgpu_vmid_free_reserved()
541 struct amdgpu_vmid_mgr *id_mgr = &adev->vm_manager.id_mgr[vmhub]; in amdgpu_vmid_reset() local
542 struct amdgpu_vmid *id = &id_mgr->ids[vmid]; in amdgpu_vmid_reset()
544 mutex_lock(&id_mgr->lock); in amdgpu_vmid_reset()
552 mutex_unlock(&id_mgr->lock); in amdgpu_vmid_reset()
567 struct amdgpu_vmid_mgr *id_mgr = in amdgpu_vmid_reset_all() local
568 &adev->vm_manager.id_mgr[i]; in amdgpu_vmid_reset_all()
570 for (j = 1; j < id_mgr->num_ids; ++j) in amdgpu_vmid_reset_all()
587 struct amdgpu_vmid_mgr *id_mgr = in amdgpu_vmid_mgr_init() local
588 &adev->vm_manager.id_mgr[i]; in amdgpu_vmid_mgr_init()
590 mutex_init(&id_mgr->lock); in amdgpu_vmid_mgr_init()
591 INIT_LIST_HEAD(&id_mgr->ids_lru); in amdgpu_vmid_mgr_init()
592 id_mgr->reserved_use_count = 0; in amdgpu_vmid_mgr_init()
595 id_mgr->num_ids = adev->vm_manager.first_kfd_vmid; in amdgpu_vmid_mgr_init()
598 for (j = 1; j < id_mgr->num_ids; ++j) { in amdgpu_vmid_mgr_init()
600 amdgpu_sync_create(&id_mgr->ids[j].active); in amdgpu_vmid_mgr_init()
601 list_add_tail(&id_mgr->ids[j].list, &id_mgr->ids_lru); in amdgpu_vmid_mgr_init()
623 struct amdgpu_vmid_mgr *id_mgr = in amdgpu_vmid_mgr_fini() local
624 &adev->vm_manager.id_mgr[i]; in amdgpu_vmid_mgr_fini()
626 mutex_destroy(&id_mgr->lock); in amdgpu_vmid_mgr_fini()
628 struct amdgpu_vmid *id = &id_mgr->ids[j]; in amdgpu_vmid_mgr_fini()