Home
last modified time | relevance | path

Searched refs:reset_domain (Results 1 – 23 of 23) sorted by relevance

/linux/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_reset.c265 struct amdgpu_reset_domain *reset_domain = container_of(ref, in amdgpu_reset_destroy_reset_domain() local
268 if (reset_domain->wq) in amdgpu_reset_destroy_reset_domain()
269 destroy_workqueue(reset_domain->wq); in amdgpu_reset_destroy_reset_domain()
271 kvfree(reset_domain); in amdgpu_reset_destroy_reset_domain()
277 struct amdgpu_reset_domain *reset_domain; in amdgpu_reset_create_reset_domain() local
279 reset_domain = kvzalloc_obj(struct amdgpu_reset_domain); in amdgpu_reset_create_reset_domain()
280 if (!reset_domain) { in amdgpu_reset_create_reset_domain()
285 reset_domain->type = type; in amdgpu_reset_create_reset_domain()
286 kref_init(&reset_domain->refcount); in amdgpu_reset_create_reset_domain()
288 reset_domain->wq = create_singlethread_workqueue(wq_name); in amdgpu_reset_create_reset_domain()
[all …]
H A Damdgpu_xgmi.c426 amdgpu_reset_put_reset_domain(hive->reset_domain); in amdgpu_xgmi_hive_release()
427 hive->reset_domain = NULL; in amdgpu_xgmi_hive_release()
725 if (adev->reset_domain->type != XGMI_HIVE) { in amdgpu_get_xgmi_hive()
726 hive->reset_domain = in amdgpu_get_xgmi_hive()
728 if (!hive->reset_domain) { in amdgpu_get_xgmi_hive()
736 amdgpu_reset_get_reset_domain(adev->reset_domain); in amdgpu_get_xgmi_hive()
737 hive->reset_domain = adev->reset_domain; in amdgpu_get_xgmi_hive()
1641 amdgpu_device_lock_reset_domain(tmp_adev->reset_domain); in amdgpu_xgmi_reset_on_init_work()
1652 amdgpu_device_unlock_reset_domain(tmp_adev->reset_domain); in amdgpu_xgmi_reset_on_init_work()
1665 amdgpu_reset_domain_schedule(hive->reset_domain, in amdgpu_xgmi_schedule_reset_on_init()
H A Damdgpu_ras_eeprom.c275 down_read(&adev->reset_domain->sem); in __write_table_header()
280 up_read(&adev->reset_domain->sem); in __write_table_header()
338 down_read(&adev->reset_domain->sem); in __write_table_ras_info()
343 up_read(&adev->reset_domain->sem); in __write_table_ras_info()
630 down_read(&adev->reset_domain->sem); in __amdgpu_ras_eeprom_write()
636 up_read(&adev->reset_domain->sem); in __amdgpu_ras_eeprom_write()
829 down_read(&adev->reset_domain->sem); in amdgpu_ras_eeprom_update_header()
834 up_read(&adev->reset_domain->sem); in amdgpu_ras_eeprom_update_header()
1020 down_read(&adev->reset_domain->sem); in __amdgpu_ras_eeprom_read()
1026 up_read(&adev->reset_domain->sem); in __amdgpu_ras_eeprom_read()
H A Damdgpu_userq.c88 amdgpu_reset_domain_schedule(adev->reset_domain, in amdgpu_userq_gpu_reset()
456 down_read(&adev->reset_domain->sem); in amdgpu_userq_cleanup()
468 up_read(&adev->reset_domain->sem); in amdgpu_userq_cleanup()
855 down_read(&adev->reset_domain->sem); in amdgpu_userq_create()
859 up_read(&adev->reset_domain->sem); in amdgpu_userq_create()
871 up_read(&adev->reset_domain->sem); in amdgpu_userq_create()
874 up_read(&adev->reset_domain->sem); in amdgpu_userq_create()
H A Damdgpu_amdkfd_arcturus.c320 if (!down_read_trylock(&adev->reset_domain->sem)) in set_barrier_auto_waitcnt()
338 up_read(&adev->reset_domain->sem); in set_barrier_auto_waitcnt()
H A Damdgpu_gart.c445 if (down_read_trylock(&adev->reset_domain->sem)) { in amdgpu_gart_invalidate_tlb()
447 up_read(&adev->reset_domain->sem); in amdgpu_gart_invalidate_tlb()
H A Damdgpu_device.c840 if (down_read_trylock(&adev->reset_domain->sem)) in amdgpu_device_skip_hw_access()
841 up_read(&adev->reset_domain->sem); in amdgpu_device_skip_hw_access()
843 lockdep_assert_held(&adev->reset_domain->sem); in amdgpu_device_skip_hw_access()
869 down_read_trylock(&adev->reset_domain->sem)) { in amdgpu_device_rreg()
871 up_read(&adev->reset_domain->sem); in amdgpu_device_rreg()
937 down_read_trylock(&adev->reset_domain->sem)) { in amdgpu_device_xcc_rreg()
939 up_read(&adev->reset_domain->sem); in amdgpu_device_xcc_rreg()
996 down_read_trylock(&adev->reset_domain->sem)) { in amdgpu_device_wreg()
998 up_read(&adev->reset_domain->sem); in amdgpu_device_wreg()
1068 down_read_trylock(&adev->reset_domain->sem)) { in amdgpu_device_xcc_wreg()
[all …]
H A Damdgpu_gmc.c723 if (!down_read_trylock(&adev->reset_domain->sem)) in amdgpu_gmc_flush_gpu_tlb()
736 up_read(&adev->reset_domain->sem); in amdgpu_gmc_flush_gpu_tlb()
784 if (!down_read_trylock(&adev->reset_domain->sem)) in amdgpu_gmc_flush_gpu_tlb_pasid()
845 !amdgpu_reset_pending(adev->reset_domain)) { in amdgpu_gmc_flush_gpu_tlb_pasid()
858 up_read(&adev->reset_domain->sem); in amdgpu_gmc_flush_gpu_tlb_pasid()
898 !amdgpu_reset_pending(adev->reset_domain)) { in amdgpu_gmc_fw_reg_write_reg_wait()
H A Damdgpu_virt.c660 if (amdgpu_reset_domain_schedule(adev->reset_domain, in amdgpu_virt_update_vf2pf_work_item()
1784 if (down_read_trylock(&adev->reset_domain->sem)) { in amdgpu_virt_req_ras_err_count()
1786 up_read(&adev->reset_domain->sem); in amdgpu_virt_req_ras_err_count()
1883 down_read_trylock(&adev->reset_domain->sem)) { in amdgpu_virt_req_ras_cper_dump()
1887 up_read(&adev->reset_domain->sem); in amdgpu_virt_req_ras_cper_dump()
2000 if (down_read_trylock(&adev->reset_domain->sem)) { in amdgpu_virt_send_remote_ras_cmd()
2003 up_read(&adev->reset_domain->sem); in amdgpu_virt_send_remote_ras_cmd()
H A Damdgpu_fence.c951 if (amdgpu_reset_domain_schedule(adev->reset_domain, &adev->reset_work)) in gpu_recover_get()
954 *val = atomic_read(&adev->reset_domain->reset_res); in gpu_recover_get()
H A Damdgpu_debugfs.c1653 r = down_write_killable(&adev->reset_domain->sem); in amdgpu_debugfs_test_ib_show()
1682 up_write(&adev->reset_domain->sem); in amdgpu_debugfs_test_ib_show()
1914 r = down_read_killable(&adev->reset_domain->sem); in amdgpu_debugfs_ib_preempt()
1955 up_read(&adev->reset_domain->sem); in amdgpu_debugfs_ib_preempt()
H A Dmxgpu_vi.c559 WARN_ONCE(!amdgpu_reset_domain_schedule(adev->reset_domain, in xgpu_vi_mailbox_rcv_irq()
H A Damdgpu_ras.c1525 if (!down_read_trylock(&adev->reset_domain->sem)) { in amdgpu_ras_query_error_status_with_event()
1534 up_read(&adev->reset_domain->sem); in amdgpu_ras_query_error_status_with_event()
3767 down_read(&adev->reset_domain->sem); in amdgpu_ras_page_retirement_thread()
3768 up_read(&adev->reset_domain->sem); in amdgpu_ras_page_retirement_thread()
5101 amdgpu_reset_domain_schedule(ras->adev->reset_domain, &ras->recovery_work); in amdgpu_ras_reset_gpu()
5106 amdgpu_reset_domain_schedule(ras->adev->reset_domain, &ras->recovery_work); in amdgpu_ras_reset_gpu()
H A Damdgpu_amdkfd.c320 (void)amdgpu_reset_domain_schedule(adev->reset_domain, &adev->kfd.reset_work); in amdgpu_amdkfd_gpu_reset()
H A Damdgpu_drv.c2666 amdgpu_device_lock_reset_domain(adev->reset_domain); in amdgpu_pmops_suspend_noirq()
2668 amdgpu_device_unlock_reset_domain(adev->reset_domain); in amdgpu_pmops_suspend_noirq()
H A Damdgpu_kms.c886 if (!down_read_trylock(&adev->reset_domain->sem)) in amdgpu_info_ioctl()
937 up_read(&adev->reset_domain->sem); in amdgpu_info_ioctl()
H A Damdgpu.h1188 struct amdgpu_reset_domain *reset_domain; member
/linux/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_device_queue_manager.c212 if (!down_read_trylock(&adev->reset_domain->sem)) in add_queue_mes()
252 up_read(&adev->reset_domain->sem); in add_queue_mes()
265 up_read(&adev->reset_domain->sem); in add_queue_mes()
285 if (!down_read_trylock(&adev->reset_domain->sem)) in remove_queue_mes()
296 up_read(&adev->reset_domain->sem); in remove_queue_mes()
367 if (!down_read_trylock(&adev->reset_domain->sem)) in suspend_all_queues_mes()
371 up_read(&adev->reset_domain->sem); in suspend_all_queues_mes()
387 if (!down_read_trylock(&adev->reset_domain->sem)) in resume_all_queues_mes()
391 up_read(&adev->reset_domain->sem); in resume_all_queues_mes()
2388 if (!down_read_trylock(&dqm->dev->adev->reset_domain->sem)) in unmap_queues_cpsch()
[all …]
H A Dkfd_kernel_queue.c174 …e->properties.type == KFD_QUEUE_TYPE_HIQ && down_read_trylock(&kq->dev->adev->reset_domain->sem)) { in kq_uninitialize()
181 up_read(&kq->dev->adev->reset_domain->sem); in kq_uninitialize()
H A Dkfd_process_queue_manager.c95 down_read_trylock(&dev->adev->reset_domain->sem)) { in kfd_process_dequeue_from_device()
99 up_read(&dev->adev->reset_domain->sem); in kfd_process_dequeue_from_device()
H A Dkfd_svm.c2385 if (!down_read_trylock(&pdd->dev->adev->reset_domain->sem)) in svm_range_drain_retry_fault()
2397 up_read(&pdd->dev->adev->reset_domain->sem); in svm_range_drain_retry_fault()
H A Dkfd_process.c1205 flush_workqueue(p->pdds[i]->dev->adev->reset_domain->wq); in kfd_process_wait_gpu_reset_complete()
/linux/drivers/gpu/drm/i915/gt/
H A Dintel_engine_cs.c398 u32 reset_domain; in get_reset_domain() local
432 reset_domain = engine_reset_domains[id]; in get_reset_domain()
443 reset_domain = engine_reset_domains[id]; in get_reset_domain()
446 return reset_domain; in get_reset_domain()
484 engine->reset_domain = get_reset_domain(GRAPHICS_VER(gt->i915), in intel_engine_setup()