Searched refs:nvmet_wq (Results 1 – 11 of 11) sorted by relevance
| /linux/drivers/nvme/target/ |
| H A D | io-cmd-file.c | 269 queue_work(nvmet_wq, &req->f.work); in nvmet_file_execute_flush() 329 queue_work(nvmet_wq, &req->f.work); in nvmet_file_execute_dsm() 359 queue_work(nvmet_wq, &req->f.work); in nvmet_file_execute_write_zeroes()
|
| H A D | fcloop.c | 374 queue_work(nvmet_wq, &rport->ls_work); in fcloop_h2t_ls_req() 411 queue_work(nvmet_wq, &rport->ls_work); in fcloop_h2t_xmt_ls_rsp() 470 queue_work(nvmet_wq, &tport->ls_work); in fcloop_t2h_ls_req() 527 queue_work(nvmet_wq, &tport->ls_work); in fcloop_t2h_xmt_ls_rsp() 576 queue_work(nvmet_wq, &tgt_rscn->work); in fcloop_tgt_discovery_evt() 793 queue_work(nvmet_wq, &tfcp_req->fcp_rcv_work); in fcloop_fcp_req() 977 queue_work(nvmet_wq, &tfcp_req->tio_done_work); in fcloop_fcp_req_release() 1036 WARN_ON(!queue_work(nvmet_wq, &tfcp_req->abort_rcv_work)); in fcloop_fcp_abort()
|
| H A D | core.c | 28 struct workqueue_struct *nvmet_wq; variable 29 EXPORT_SYMBOL_GPL(nvmet_wq); 406 queue_delayed_work(nvmet_wq, &ctrl->ka_work, ctrl->kato * HZ); in nvmet_keep_alive_timer() 424 queue_delayed_work(nvmet_wq, &ctrl->ka_work, ctrl->kato * HZ); in nvmet_start_keep_alive_timer() 1408 mod_delayed_work(nvmet_wq, &ctrl->ka_work, ctrl->kato * HZ); in nvmet_start_ctrl() 1775 queue_work(nvmet_wq, &ctrl->fatal_err_work); in nvmet_ctrl_fatal_error() 1956 nvmet_wq = alloc_workqueue("nvmet-wq", in nvmet_init() 1958 if (!nvmet_wq) in nvmet_init() 1987 destroy_workqueue(nvmet_wq); in nvmet_init() 2004 destroy_workqueue(nvmet_wq); in nvmet_exit()
|
| H A D | loop.c | 170 queue_work(nvmet_wq, &iod->work); in nvme_loop_queue_rq() 190 queue_work(nvmet_wq, &iod->work); in nvme_loop_submit_async_event()
|
| H A D | rdma.c | 1687 queue_work(nvmet_wq, &queue->release_work); in __nvmet_rdma_queue_disconnect() 1717 queue_work(nvmet_wq, &queue->release_work); in nvmet_rdma_queue_connect_fail() 1791 queue_delayed_work(nvmet_wq, &port->repair_work, 0); in nvmet_rdma_cm_handler() 1917 queue_delayed_work(nvmet_wq, &port->repair_work, 5 * HZ); in nvmet_rdma_repair_port_work() 2089 flush_workqueue(nvmet_wq); in nvmet_rdma_remove_one()
|
| H A D | tcp.c | 1388 queue_work(nvmet_wq, &queue->release_work); in nvmet_tcp_release_queue() 1887 queue_delayed_work(nvmet_wq, &queue->tls_handshake_tmo_work, in nvmet_tcp_tls_handshake() 2029 queue_work(nvmet_wq, &port->accept_work); in nvmet_tcp_listen_data_ready() 2248 flush_workqueue(nvmet_wq); in nvmet_tcp_exit() 2253 flush_workqueue(nvmet_wq); in nvmet_tcp_exit()
|
| H A D | fc.c | 371 queue_work(nvmet_wq, &lsop->put_work); in __nvmet_fc_finish_ls_req() 1084 if (!queue_work(nvmet_wq, &assoc->del_work)) in nvmet_fc_schedule_delete_assoc() 1649 flush_workqueue(nvmet_wq); in nvmet_fc_unregister_targetport() 2079 queue_work(nvmet_wq, &iod->work); in nvmet_fc_rcv_ls_req() 3011 flush_workqueue(nvmet_wq); in nvmet_fc_exit_module()
|
| H A D | passthru.c | 352 queue_work(nvmet_wq, &req->p.work); in nvmet_passthru_execute_cmd()
|
| H A D | nvmet.h | 503 extern struct workqueue_struct *nvmet_wq;
|
| H A D | pr.c | 689 queue_work(nvmet_wq, &req->r.abort_work); in nvmet_execute_pr_acquire()
|
| H A D | configfs.c | 1982 flush_workqueue(nvmet_wq); in nvmet_port_release()
|