| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_reset.c | 265 struct amdgpu_reset_domain *reset_domain = container_of(ref, in amdgpu_reset_destroy_reset_domain() local 268 if (reset_domain->wq) in amdgpu_reset_destroy_reset_domain() 269 destroy_workqueue(reset_domain->wq); in amdgpu_reset_destroy_reset_domain() 271 kvfree(reset_domain); in amdgpu_reset_destroy_reset_domain() 277 struct amdgpu_reset_domain *reset_domain; in amdgpu_reset_create_reset_domain() local 279 reset_domain = kvzalloc_obj(struct amdgpu_reset_domain); in amdgpu_reset_create_reset_domain() 280 if (!reset_domain) { in amdgpu_reset_create_reset_domain() 285 reset_domain->type = type; in amdgpu_reset_create_reset_domain() 286 kref_init(&reset_domain->refcount); in amdgpu_reset_create_reset_domain() 288 reset_domain->wq = create_singlethread_workqueue(wq_name); in amdgpu_reset_create_reset_domain() [all …]
|
| H A D | amdgpu_xgmi.c | 426 amdgpu_reset_put_reset_domain(hive->reset_domain); in amdgpu_xgmi_hive_release() 427 hive->reset_domain = NULL; in amdgpu_xgmi_hive_release() 725 if (adev->reset_domain->type != XGMI_HIVE) { in amdgpu_get_xgmi_hive() 726 hive->reset_domain = in amdgpu_get_xgmi_hive() 728 if (!hive->reset_domain) { in amdgpu_get_xgmi_hive() 736 amdgpu_reset_get_reset_domain(adev->reset_domain); in amdgpu_get_xgmi_hive() 737 hive->reset_domain = adev->reset_domain; in amdgpu_get_xgmi_hive() 1641 amdgpu_device_lock_reset_domain(tmp_adev->reset_domain); in amdgpu_xgmi_reset_on_init_work() 1652 amdgpu_device_unlock_reset_domain(tmp_adev->reset_domain); in amdgpu_xgmi_reset_on_init_work() 1665 amdgpu_reset_domain_schedule(hive->reset_domain, in amdgpu_xgmi_schedule_reset_on_init()
|
| H A D | amdgpu_ras_eeprom.c | 275 down_read(&adev->reset_domain->sem); in __write_table_header() 280 up_read(&adev->reset_domain->sem); in __write_table_header() 338 down_read(&adev->reset_domain->sem); in __write_table_ras_info() 343 up_read(&adev->reset_domain->sem); in __write_table_ras_info() 630 down_read(&adev->reset_domain->sem); in __amdgpu_ras_eeprom_write() 636 up_read(&adev->reset_domain->sem); in __amdgpu_ras_eeprom_write() 829 down_read(&adev->reset_domain->sem); in amdgpu_ras_eeprom_update_header() 834 up_read(&adev->reset_domain->sem); in amdgpu_ras_eeprom_update_header() 1020 down_read(&adev->reset_domain->sem); in __amdgpu_ras_eeprom_read() 1026 up_read(&adev->reset_domain->sem); in __amdgpu_ras_eeprom_read()
|
| H A D | amdgpu_userq.c | 88 amdgpu_reset_domain_schedule(adev->reset_domain, in amdgpu_userq_gpu_reset() 456 down_read(&adev->reset_domain->sem); in amdgpu_userq_cleanup() 468 up_read(&adev->reset_domain->sem); in amdgpu_userq_cleanup() 855 down_read(&adev->reset_domain->sem); in amdgpu_userq_create() 859 up_read(&adev->reset_domain->sem); in amdgpu_userq_create() 871 up_read(&adev->reset_domain->sem); in amdgpu_userq_create() 874 up_read(&adev->reset_domain->sem); in amdgpu_userq_create()
|
| H A D | amdgpu_amdkfd_arcturus.c | 320 if (!down_read_trylock(&adev->reset_domain->sem)) in set_barrier_auto_waitcnt() 338 up_read(&adev->reset_domain->sem); in set_barrier_auto_waitcnt()
|
| H A D | amdgpu_gart.c | 445 if (down_read_trylock(&adev->reset_domain->sem)) { in amdgpu_gart_invalidate_tlb() 447 up_read(&adev->reset_domain->sem); in amdgpu_gart_invalidate_tlb()
|
| H A D | amdgpu_device.c | 840 if (down_read_trylock(&adev->reset_domain->sem)) in amdgpu_device_skip_hw_access() 841 up_read(&adev->reset_domain->sem); in amdgpu_device_skip_hw_access() 843 lockdep_assert_held(&adev->reset_domain->sem); in amdgpu_device_skip_hw_access() 869 down_read_trylock(&adev->reset_domain->sem)) { in amdgpu_device_rreg() 871 up_read(&adev->reset_domain->sem); in amdgpu_device_rreg() 937 down_read_trylock(&adev->reset_domain->sem)) { in amdgpu_device_xcc_rreg() 939 up_read(&adev->reset_domain->sem); in amdgpu_device_xcc_rreg() 996 down_read_trylock(&adev->reset_domain->sem)) { in amdgpu_device_wreg() 998 up_read(&adev->reset_domain->sem); in amdgpu_device_wreg() 1068 down_read_trylock(&adev->reset_domain->sem)) { in amdgpu_device_xcc_wreg() [all …]
|
| H A D | amdgpu_gmc.c | 723 if (!down_read_trylock(&adev->reset_domain->sem)) in amdgpu_gmc_flush_gpu_tlb() 736 up_read(&adev->reset_domain->sem); in amdgpu_gmc_flush_gpu_tlb() 784 if (!down_read_trylock(&adev->reset_domain->sem)) in amdgpu_gmc_flush_gpu_tlb_pasid() 845 !amdgpu_reset_pending(adev->reset_domain)) { in amdgpu_gmc_flush_gpu_tlb_pasid() 858 up_read(&adev->reset_domain->sem); in amdgpu_gmc_flush_gpu_tlb_pasid() 898 !amdgpu_reset_pending(adev->reset_domain)) { in amdgpu_gmc_fw_reg_write_reg_wait()
|
| H A D | amdgpu_virt.c | 660 if (amdgpu_reset_domain_schedule(adev->reset_domain, in amdgpu_virt_update_vf2pf_work_item() 1784 if (down_read_trylock(&adev->reset_domain->sem)) { in amdgpu_virt_req_ras_err_count() 1786 up_read(&adev->reset_domain->sem); in amdgpu_virt_req_ras_err_count() 1883 down_read_trylock(&adev->reset_domain->sem)) { in amdgpu_virt_req_ras_cper_dump() 1887 up_read(&adev->reset_domain->sem); in amdgpu_virt_req_ras_cper_dump() 2000 if (down_read_trylock(&adev->reset_domain->sem)) { in amdgpu_virt_send_remote_ras_cmd() 2003 up_read(&adev->reset_domain->sem); in amdgpu_virt_send_remote_ras_cmd()
|
| H A D | amdgpu_fence.c | 951 if (amdgpu_reset_domain_schedule(adev->reset_domain, &adev->reset_work)) in gpu_recover_get() 954 *val = atomic_read(&adev->reset_domain->reset_res); in gpu_recover_get()
|
| H A D | amdgpu_debugfs.c | 1653 r = down_write_killable(&adev->reset_domain->sem); in amdgpu_debugfs_test_ib_show() 1682 up_write(&adev->reset_domain->sem); in amdgpu_debugfs_test_ib_show() 1914 r = down_read_killable(&adev->reset_domain->sem); in amdgpu_debugfs_ib_preempt() 1955 up_read(&adev->reset_domain->sem); in amdgpu_debugfs_ib_preempt()
|
| H A D | mxgpu_vi.c | 559 WARN_ONCE(!amdgpu_reset_domain_schedule(adev->reset_domain, in xgpu_vi_mailbox_rcv_irq()
|
| H A D | amdgpu_ras.c | 1525 if (!down_read_trylock(&adev->reset_domain->sem)) { in amdgpu_ras_query_error_status_with_event() 1534 up_read(&adev->reset_domain->sem); in amdgpu_ras_query_error_status_with_event() 3767 down_read(&adev->reset_domain->sem); in amdgpu_ras_page_retirement_thread() 3768 up_read(&adev->reset_domain->sem); in amdgpu_ras_page_retirement_thread() 5101 amdgpu_reset_domain_schedule(ras->adev->reset_domain, &ras->recovery_work); in amdgpu_ras_reset_gpu() 5106 amdgpu_reset_domain_schedule(ras->adev->reset_domain, &ras->recovery_work); in amdgpu_ras_reset_gpu()
|
| H A D | amdgpu_amdkfd.c | 320 (void)amdgpu_reset_domain_schedule(adev->reset_domain, &adev->kfd.reset_work); in amdgpu_amdkfd_gpu_reset()
|
| H A D | amdgpu_drv.c | 2666 amdgpu_device_lock_reset_domain(adev->reset_domain); in amdgpu_pmops_suspend_noirq() 2668 amdgpu_device_unlock_reset_domain(adev->reset_domain); in amdgpu_pmops_suspend_noirq()
|
| H A D | amdgpu_kms.c | 886 if (!down_read_trylock(&adev->reset_domain->sem)) in amdgpu_info_ioctl() 937 up_read(&adev->reset_domain->sem); in amdgpu_info_ioctl()
|
| H A D | amdgpu.h | 1188 struct amdgpu_reset_domain *reset_domain; member
|
| /linux/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_device_queue_manager.c | 212 if (!down_read_trylock(&adev->reset_domain->sem)) in add_queue_mes() 252 up_read(&adev->reset_domain->sem); in add_queue_mes() 265 up_read(&adev->reset_domain->sem); in add_queue_mes() 285 if (!down_read_trylock(&adev->reset_domain->sem)) in remove_queue_mes() 296 up_read(&adev->reset_domain->sem); in remove_queue_mes() 367 if (!down_read_trylock(&adev->reset_domain->sem)) in suspend_all_queues_mes() 371 up_read(&adev->reset_domain->sem); in suspend_all_queues_mes() 387 if (!down_read_trylock(&adev->reset_domain->sem)) in resume_all_queues_mes() 391 up_read(&adev->reset_domain->sem); in resume_all_queues_mes() 2388 if (!down_read_trylock(&dqm->dev->adev->reset_domain->sem)) in unmap_queues_cpsch() [all …]
|
| H A D | kfd_kernel_queue.c | 174 …e->properties.type == KFD_QUEUE_TYPE_HIQ && down_read_trylock(&kq->dev->adev->reset_domain->sem)) { in kq_uninitialize() 181 up_read(&kq->dev->adev->reset_domain->sem); in kq_uninitialize()
|
| H A D | kfd_process_queue_manager.c | 95 down_read_trylock(&dev->adev->reset_domain->sem)) { in kfd_process_dequeue_from_device() 99 up_read(&dev->adev->reset_domain->sem); in kfd_process_dequeue_from_device()
|
| H A D | kfd_svm.c | 2385 if (!down_read_trylock(&pdd->dev->adev->reset_domain->sem)) in svm_range_drain_retry_fault() 2397 up_read(&pdd->dev->adev->reset_domain->sem); in svm_range_drain_retry_fault()
|
| H A D | kfd_process.c | 1205 flush_workqueue(p->pdds[i]->dev->adev->reset_domain->wq); in kfd_process_wait_gpu_reset_complete()
|
| /linux/drivers/gpu/drm/i915/gt/ |
| H A D | intel_engine_cs.c | 398 u32 reset_domain; in get_reset_domain() local 432 reset_domain = engine_reset_domains[id]; in get_reset_domain() 443 reset_domain = engine_reset_domains[id]; in get_reset_domain() 446 return reset_domain; in get_reset_domain() 484 engine->reset_domain = get_reset_domain(GRAPHICS_VER(gt->i915), in intel_engine_setup()
|