/linux/drivers/gpu/drm/msm/adreno/ |
H A D | a5xx_preempt.c | 90 kthread_queue_work(gpu->worker, &gpu->recover_work); in a5xx_preempt_timer() 198 kthread_queue_work(gpu->worker, &gpu->recover_work); in a5xx_preempt_irq()
|
H A D | a6xx_gpu.c | 1512 kthread_queue_work(gpu->worker, &gpu->recover_work); in a6xx_fault_detect_irq() 1532 kthread_queue_work(gpu->worker, &gpu->recover_work); in a7xx_sw_fuse_violation_irq()
|
H A D | a5xx_gpu.c | 1258 kthread_queue_work(gpu->worker, &gpu->recover_work); in a5xx_fault_detect_irq()
|
H A D | a6xx_gmu.c | 33 kthread_queue_work(gpu->worker, &gpu->recover_work); in a6xx_gmu_fault()
|
/linux/drivers/gpu/drm/lima/ |
H A D | lima_sched.h | 85 struct work_struct recover_work; member
|
H A D | lima_sched.c | 497 container_of(work, struct lima_sched_pipe, recover_work); in lima_sched_recover_work() 522 INIT_WORK(&pipe->recover_work, lima_sched_recover_work); in lima_sched_pipe_init() 544 schedule_work(&pipe->recover_work); in lima_sched_pipe_task_done()
|
/linux/drivers/gpu/drm/tilcdc/ |
H A D | tilcdc_crtc.c | 55 struct work_struct recover_work; member 573 container_of(work, struct tilcdc_crtc, recover_work); in tilcdc_crtc_recover_work() 989 &tilcdc_crtc->recover_work); in tilcdc_crtc_irq() 1048 INIT_WORK(&tilcdc_crtc->recover_work, tilcdc_crtc_recover_work); in tilcdc_crtc_create()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en.h | 457 struct work_struct recover_work; member 545 struct work_struct recover_work; member 720 struct work_struct recover_work; member 1119 void mlx5e_tx_err_cqe_work(struct work_struct *recover_work);
|
H A D | en_main.c | 705 static void mlx5e_rq_err_cqe_work(struct work_struct *recover_work) in mlx5e_rq_err_cqe_work() argument 707 struct mlx5e_rq *rq = container_of(recover_work, struct mlx5e_rq, recover_work); in mlx5e_rq_err_cqe_work() 836 INIT_WORK(&rq->recover_work, mlx5e_rq_err_cqe_work); in mlx5e_alloc_rq() 1381 cancel_work_sync(&rq->recover_work); in mlx5e_close_rq() 1518 static void mlx5e_icosq_err_cqe_work(struct work_struct *recover_work) in mlx5e_icosq_err_cqe_work() argument 1520 struct mlx5e_icosq *sq = container_of(recover_work, struct mlx5e_icosq, in mlx5e_icosq_err_cqe_work() 1521 recover_work); in mlx5e_icosq_err_cqe_work() 1526 static void mlx5e_async_icosq_err_cqe_work(struct work_struct *recover_work) in mlx5e_async_icosq_err_cqe_work() argument 1528 struct mlx5e_icosq *sq = container_of(recover_work, struct mlx5e_icosq, in mlx5e_async_icosq_err_cqe_work() 1529 recover_work); in mlx5e_async_icosq_err_cqe_work() [all …]
|
H A D | en_tx.c | 870 queue_work(cq->workqueue, &sq->recover_work); in mlx5e_poll_tx_cq()
|
H A D | en_rx.c | 1039 queue_work(cq->workqueue, &sq->recover_work); in mlx5e_poll_ico_cq() 1806 queue_work(priv->wq, &rq->recover_work); in trigger_report()
|
/linux/drivers/gpu/drm/msm/ |
H A D | msm_gpu.c | 357 struct msm_gpu *gpu = container_of(work, struct msm_gpu, recover_work); in recover_worker() 536 kthread_queue_work(gpu->worker, &gpu->recover_work); in hangcheck_handler() 875 kthread_init_work(&gpu->recover_work, recover_worker); in msm_gpu_init()
|
H A D | msm_gpu.h | 266 struct kthread_work recover_work; member
|
/linux/fs/xfs/ |
H A D | xfs_extfree_item.c | 673 .recover_work = xfs_extent_free_recover_work, 686 .recover_work = xfs_extent_free_recover_work,
|
H A D | xfs_refcount_item.c | 536 .recover_work = xfs_refcount_recover_work,
|
H A D | xfs_rmap_item.c | 587 .recover_work = xfs_rmap_recover_work,
|
H A D | xfs_bmap_item.c | 601 .recover_work = xfs_bmap_recover_work,
|
H A D | xfs_attr_item.c | 948 .recover_work = xfs_attr_recover_work,
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | ptp.c | 341 INIT_WORK(&sq->recover_work, mlx5e_tx_err_cqe_work); in mlx5e_ptp_alloc_txqsq() 503 cancel_work_sync(&sq->recover_work); in mlx5e_ptp_close_txqsq()
|
/linux/fs/xfs/libxfs/ |
H A D | xfs_defer.c | 919 error = ops->recover_work(dfp, capture_list); in xfs_defer_finish_recovery()
|