Home
last modified time | relevance | path

Searched refs:io_req (Results 1 – 25 of 25) sorted by relevance

/linux/drivers/scsi/qedf/
H A Dqedf_io.c11 void qedf_cmd_timer_set(struct qedf_ctx *qedf, struct qedf_ioreq *io_req, in qedf_cmd_timer_set() argument
14 queue_delayed_work(qedf->timer_work_queue, &io_req->timeout_work, in qedf_cmd_timer_set()
21 struct qedf_ioreq *io_req = in qedf_cmd_timeout() local
26 fcport = io_req->fcport; in qedf_cmd_timeout()
27 if (io_req->fcport == NULL) { in qedf_cmd_timeout()
34 switch (io_req->cmd_type) { in qedf_cmd_timeout()
39 io_req->xid); in qedf_cmd_timeout()
44 io_req->xid); in qedf_cmd_timeout()
46 qedf_initiate_cleanup(io_req, true); in qedf_cmd_timeout()
47 complete(&io_req->abts_done); in qedf_cmd_timeout()
[all …]
H A Dqedf_els.c72 cb_arg->io_req = els_req; in qedf_initiate_els()
198 rrq_req = cb_arg->io_req; in qedf_rrq_compl()
430 els_req = cb_arg->io_req; in qedf_l2_els_compl()
552 srr_req = cb_arg->io_req; in qedf_srr_compl()
744 struct fcoe_cqe *cqe, struct qedf_ioreq *io_req) in qedf_process_seq_cleanup_compl() argument
749 cb_arg = io_req->cb_arg; in qedf_process_seq_cleanup_compl()
752 if (io_req->event == QEDF_IOREQ_EV_ELS_TMO || !cqe) { in qedf_process_seq_cleanup_compl()
754 "cqe is NULL or timeout event (0x%x)", io_req->event); in qedf_process_seq_cleanup_compl()
759 cancel_delayed_work_sync(&io_req->timeout_work); in qedf_process_seq_cleanup_compl()
761 rc = qedf_send_srr(io_req, cb_arg->offset, cb_arg->r_ctl); in qedf_process_seq_cleanup_compl()
[all …]
H A Dqedf_main.c728 struct qedf_ioreq *io_req; in qedf_eh_abort() local
751 io_req = qedf_priv(sc_cmd)->io_req; in qedf_eh_abort()
752 if (!io_req) { in qedf_eh_abort()
761 rval = kref_get_unless_zero(&io_req->refcount); /* ID: 005 */ in qedf_eh_abort()
766 if (!rval || io_req->sc_cmd != sc_cmd) { in qedf_eh_abort()
769 io_req->sc_cmd, sc_cmd, rdata->ids.port_id); in qedf_eh_abort()
775 refcount = kref_read(&io_req->refcount); in qedf_eh_abort()
778 io_req, io_req->xid, sc_cmd, sc_cmd->cmnd[0], in qedf_eh_abort()
791 io_req->xid, rdata->ids.port_id); in qedf_eh_abort()
792 while (io_req->sc_cmd && (wait_count != 0)) { in qedf_eh_abort()
[all …]
/linux/drivers/scsi/bnx2fc/
H A Dbnx2fc_io.c19 static int bnx2fc_split_bd(struct bnx2fc_cmd *io_req, u64 addr, int sg_len,
21 static int bnx2fc_map_sg(struct bnx2fc_cmd *io_req);
22 static int bnx2fc_build_bd_list_from_sg(struct bnx2fc_cmd *io_req);
23 static void bnx2fc_unmap_sg_list(struct bnx2fc_cmd *io_req);
24 static void bnx2fc_free_mp_resc(struct bnx2fc_cmd *io_req);
25 static void bnx2fc_parse_fcp_rsp(struct bnx2fc_cmd *io_req,
29 void bnx2fc_cmd_timer_set(struct bnx2fc_cmd *io_req, in bnx2fc_cmd_timer_set() argument
32 struct bnx2fc_interface *interface = io_req->port->priv; in bnx2fc_cmd_timer_set()
35 &io_req->timeout_work, in bnx2fc_cmd_timer_set()
37 kref_get(&io_req->refcount); in bnx2fc_cmd_timer_set()
[all …]
H A Dbnx2fc_tgt.c166 struct bnx2fc_cmd *io_req; in bnx2fc_flush_active_ios() local
176 list_for_each_entry_safe(io_req, tmp, &tgt->active_cmd_queue, link) { in bnx2fc_flush_active_ios()
178 list_del_init(&io_req->link); in bnx2fc_flush_active_ios()
179 io_req->on_active_queue = 0; in bnx2fc_flush_active_ios()
180 BNX2FC_IO_DBG(io_req, "cmd_queue cleanup\n"); in bnx2fc_flush_active_ios()
182 if (cancel_delayed_work(&io_req->timeout_work)) { in bnx2fc_flush_active_ios()
184 &io_req->req_flags)) { in bnx2fc_flush_active_ios()
186 BNX2FC_IO_DBG(io_req, "eh_abort for IO " in bnx2fc_flush_active_ios()
188 complete(&io_req->abts_done); in bnx2fc_flush_active_ios()
190 kref_put(&io_req->refcount, in bnx2fc_flush_active_ios()
[all …]
H A Dbnx2fc_debug.c17 void BNX2FC_IO_DBG(const struct bnx2fc_cmd *io_req, const char *fmt, ...) in BNX2FC_IO_DBG() argument
30 if (io_req && io_req->port && io_req->port->lport && in BNX2FC_IO_DBG()
31 io_req->port->lport->host) in BNX2FC_IO_DBG()
32 shost_printk(KERN_INFO, io_req->port->lport->host, in BNX2FC_IO_DBG()
34 io_req->xid, &vaf); in BNX2FC_IO_DBG()
H A Dbnx2fc_hwi.c635 struct bnx2fc_cmd *io_req = NULL; in bnx2fc_process_unsol_compl() local
710 io_req = (struct bnx2fc_cmd *)hba->cmd_mgr->cmds[xid]; in bnx2fc_process_unsol_compl()
711 if (!io_req) in bnx2fc_process_unsol_compl()
714 if (io_req->cmd_type != BNX2FC_SCSI_CMD) { in bnx2fc_process_unsol_compl()
720 &io_req->req_flags)) { in bnx2fc_process_unsol_compl()
721 BNX2FC_IO_DBG(io_req, "unsol_err: cleanup in " in bnx2fc_process_unsol_compl()
743 if (test_bit(BNX2FC_FLAG_ISSUE_ABTS, &io_req->req_flags)) { in bnx2fc_process_unsol_compl()
760 memcpy(&io_req->err_entry, err_entry, in bnx2fc_process_unsol_compl()
763 &io_req->req_flags)) { in bnx2fc_process_unsol_compl()
765 rc = bnx2fc_send_rec(io_req); in bnx2fc_process_unsol_compl()
[all …]
H A Dbnx2fc.h401 struct bnx2fc_cmd *io_req; member
468 struct bnx2fc_cmd *io_req; member
490 struct bnx2fc_cmd *io_req; member
529 int bnx2fc_initiate_cleanup(struct bnx2fc_cmd *io_req);
530 int bnx2fc_initiate_abts(struct bnx2fc_cmd *io_req);
531 void bnx2fc_cmd_timer_set(struct bnx2fc_cmd *io_req,
533 int bnx2fc_init_mp_req(struct bnx2fc_cmd *io_req);
534 void bnx2fc_init_cleanup_task(struct bnx2fc_cmd *io_req,
541 void bnx2fc_init_mp_task(struct bnx2fc_cmd *io_req,
543 void bnx2fc_init_task(struct bnx2fc_cmd *io_req,
[all …]
H A Dbnx2fc_debug.h41 void BNX2FC_IO_DBG(const struct bnx2fc_cmd *io_req, const char *fmt, ...);
H A Dbnx2fc_els.c35 rrq_req = cb_arg->io_req; in bnx2fc_rrq_compl()
138 els_req = cb_arg->io_req; in bnx2fc_l2_els_compl()
279 srr_req = cb_arg->io_req; in bnx2fc_srr_compl()
399 rec_req = cb_arg->io_req; in bnx2fc_rec_compl()
715 cb_arg->io_req = els_req; in bnx2fc_initiate_els()
/linux/drivers/scsi/fnic/
H A Dfnic_scsi.c112 struct fnic_io_req *io_req, in fnic_release_ioreq_buf() argument
115 if (io_req->sgl_list_pa) in fnic_release_ioreq_buf()
116 dma_unmap_single(&fnic->pdev->dev, io_req->sgl_list_pa, in fnic_release_ioreq_buf()
117 sizeof(io_req->sgl_list[0]) * io_req->sgl_cnt, in fnic_release_ioreq_buf()
121 if (io_req->sgl_cnt) in fnic_release_ioreq_buf()
122 mempool_free(io_req->sgl_list_alloc, in fnic_release_ioreq_buf()
123 fnic->io_sgl_pool[io_req->sgl_type]); in fnic_release_ioreq_buf()
124 if (io_req->sense_buf_pa) in fnic_release_ioreq_buf()
125 dma_unmap_single(&fnic->pdev->dev, io_req->sense_buf_pa, in fnic_release_ioreq_buf()
135 struct fnic_io_req *io_req = fnic_priv(sc)->io_req; in fnic_count_portid_ioreqs_iter() local
[all …]
H A Dfnic.h159 struct fnic_io_req *io_req; member
/linux/drivers/md/
H A Ddm-io.c464 static int dp_init(struct dm_io_request *io_req, struct dpages *dp, in dp_init()
472 switch (io_req->mem.type) { in dp_init()
474 list_dp_init(dp, io_req->mem.ptr.pl, io_req->mem.offset); in dp_init()
478 bio_dp_init(dp, io_req->mem.ptr.bio); in dp_init()
482 flush_kernel_vmap_range(io_req->mem.ptr.vma, size); in dp_init()
483 if ((io_req->bi_opf & REQ_OP_MASK) == REQ_OP_READ) { in dp_init()
484 dp->vma_invalidate_address = io_req->mem.ptr.vma; in dp_init()
487 vm_dp_init(dp, io_req->mem.ptr.vma); in dp_init()
491 km_dp_init(dp, io_req in dp_init()
463 dp_init(struct dm_io_request * io_req,struct dpages * dp,unsigned long size) dp_init() argument
500 dm_io(struct dm_io_request * io_req,unsigned int num_regions,struct dm_io_region * where,unsigned long * sync_error_bits,unsigned short ioprio) dm_io() argument
[all...]
H A Ddm-log.c243 struct dm_io_request io_req; member
301 lc->io_req.bi_opf = op; in rw_header()
303 return dm_io(&lc->io_req, 1, &lc->header_location, NULL, IOPRIO_DEFAULT); in rw_header()
314 lc->io_req.bi_opf = REQ_OP_WRITE | REQ_PREFLUSH; in flush_header()
316 return dm_io(&lc->io_req, 1, &null_location, NULL, IOPRIO_DEFAULT); in flush_header()
461 lc->io_req.mem.type = DM_IO_VMA; in create_log_context()
462 lc->io_req.notify.fn = NULL; in create_log_context()
463 lc->io_req.client = dm_io_client_create(); in create_log_context()
464 if (IS_ERR(lc->io_req.client)) { in create_log_context()
465 r = PTR_ERR(lc->io_req.client); in create_log_context()
[all …]
H A Ddm-integrity.c531 struct dm_io_request io_req; in sync_rw_sb() local
536 io_req.bi_opf = opf; in sync_rw_sb()
537 io_req.mem.type = DM_IO_KMEM; in sync_rw_sb()
538 io_req.mem.ptr.addr = ic->sb; in sync_rw_sb()
539 io_req.notify.fn = NULL; in sync_rw_sb()
540 io_req.client = ic->io; in sync_rw_sb()
554 r = dm_io(&io_req, 1, &io_loc, NULL, IOPRIO_DEFAULT); in sync_rw_sb()
1040 struct dm_io_request io_req; in rw_journal_sectors() local
1054 io_req.bi_opf = opf; in rw_journal_sectors()
1055 io_req.mem.type = DM_IO_PAGE_LIST; in rw_journal_sectors()
[all …]
H A Ddm-snap-persistent.c217 struct dm_io_request *io_req; member
226 req->result = dm_io(req->io_req, 1, req->where, NULL, IOPRIO_DEFAULT); in do_metadata()
240 struct dm_io_request io_req = { in chunk_io() local
250 return dm_io(&io_req, 1, &where, NULL, IOPRIO_DEFAULT); in chunk_io()
253 req.io_req = &io_req; in chunk_io()
H A Ddm-raid1.c267 struct dm_io_request io_req = { in mirror_flush() local
281 dm_io(&io_req, ms->nr_mirrors, io, &error_bits, IOPRIO_DEFAULT); in mirror_flush()
546 struct dm_io_request io_req = { in read_async_bio() local
557 BUG_ON(dm_io(&io_req, 1, &io, NULL, IOPRIO_DEFAULT)); in read_async_bio()
660 struct dm_io_request io_req = { in do_write() local
670 io_req.bi_opf = REQ_OP_DISCARD | op_flags; in do_write()
671 io_req.mem.type = DM_IO_KMEM; in do_write()
672 io_req.mem.ptr.addr = NULL; in do_write()
684 BUG_ON(dm_io(&io_req, ms->nr_mirrors, io, NULL, IOPRIO_DEFAULT)); in do_write()
H A Ddm-kcopyd.c558 struct dm_io_request io_req = { in run_io_job() local
581 r = dm_io(&io_req, 1, &job->source, NULL, IOPRIO_DEFAULT); in run_io_job()
583 r = dm_io(&io_req, job->num_dests, job->dests, NULL, IOPRIO_DEFAULT); in run_io_job()
/linux/samples/acrn/
H A Dvm-sample.c50 struct acrn_io_request *io_req; in main() local
109 io_req = &io_req_buf[vcpu_id]; in main()
110 if ((__sync_add_and_fetch(&io_req->processed, 0) == ACRN_IOREQ_STATE_PROCESSING) in main()
111 && (!io_req->kernel_handled)) in main()
112 if (io_req->type == ACRN_IOREQ_TYPE_PORTIO) { in main()
115 port = io_req->reqs.pio_request.address; in main()
116 bytes = io_req->reqs.pio_request.size; in main()
117 in = (io_req->reqs.pio_request.direction == ACRN_IOREQ_DIR_READ); in main()
/linux/drivers/staging/rtl8723bs/include/
H A Drtw_io.h52 struct io_req { struct
60 void (*_async_io_callback)(struct adapter *padater, struct io_req *pio_req, u8 *cnxt); argument
101 extern void sync_ioreq_enqueue(struct io_req *preq, struct io_queue *ioqueue);
105 extern uint free_ioreq(struct io_req *preq, struct io_queue *pio_queue);
106 extern struct io_req *alloc_ioreq(struct io_queue *pio_q);
136 void (*_async_io_callback)(struct adapter *padater, struct io_req *pio_req, u8 *cnxt), u8 *cnxt);
138 void (*_async_io_callback)(struct adapter *padater, struct io_req *pio_req, u8 *cnxt), u8 *cnxt);
140 void (*_async_io_callback)(struct adapter *padater, struct io_req *pio_req, u8 *cnxt), u8 *cnxt);
146 void (*_async_io_callback)(struct adapter *padater, struct io_req *pio_req, u8 *cnxt), u8 *cnxt);
148 void (*_async_io_callback)(struct adapter *padater, struct io_req *pio_req, u8 *cnxt), u8 *cnxt);
[all …]
/linux/drivers/scsi/csiostor/
H A Dcsio_lnode.c1435 struct csio_ioreq *io_req = NULL; in csio_ln_mgmt_wr_handler() local
1448 io_req = (struct csio_ioreq *) ((uintptr_t) wr_cmd->cookie); in csio_ln_mgmt_wr_handler()
1449 io_req->wr_status = csio_wr_status(wr_cmd); in csio_ln_mgmt_wr_handler()
1453 if (csio_mgmt_req_lookup(mgmtm, io_req) != 0) { in csio_ln_mgmt_wr_handler()
1456 io_req); in csio_ln_mgmt_wr_handler()
1465 list_del_init(&io_req->sm.sm_list); in csio_ln_mgmt_wr_handler()
1470 if (io_req->io_cbfn) in csio_ln_mgmt_wr_handler()
1471 io_req->io_cbfn(hw, io_req); in csio_ln_mgmt_wr_handler()
1680 csio_ln_prep_ecwr(struct csio_ioreq *io_req, uint32_t wr_len, in csio_ln_prep_ecwr() argument
1697 wr->cookie = io_req->fw_handle; in csio_ln_prep_ecwr()
[all …]
H A Dcsio_hw.c4087 csio_mgmt_req_lookup(struct csio_mgmtm *mgmtm, struct csio_ioreq *io_req) in csio_mgmt_req_lookup() argument
4093 if (io_req == (struct csio_ioreq *)tmp) in csio_mgmt_req_lookup()
4112 struct csio_ioreq *io_req; in csio_mgmt_tmo_handler() local
4119 io_req = (struct csio_ioreq *) tmp; in csio_mgmt_tmo_handler()
4120 io_req->tmo -= min_t(uint32_t, io_req->tmo, ECM_MIN_TMO); in csio_mgmt_tmo_handler()
4122 if (!io_req->tmo) { in csio_mgmt_tmo_handler()
4125 list_del_init(&io_req->sm.sm_list); in csio_mgmt_tmo_handler()
4126 if (io_req->io_cbfn) { in csio_mgmt_tmo_handler()
4128 io_req->wr_status = -ETIMEDOUT; in csio_mgmt_tmo_handler()
4129 io_req->io_cbfn(mgmtm->hw, io_req); in csio_mgmt_tmo_handler()
[all …]
/linux/include/linux/
H A Ddm-io.h82 int dm_io(struct dm_io_request *io_req, unsigned int num_regions,
/linux/fs/smb/client/
H A Dcifssmb.c2690 TRANSACT_IOCTL_REQ *io_req = NULL; in cifs_query_reparse_point() local
2720 (void **)&io_req, (void **)&io_rsp); in cifs_query_reparse_point()
2724 io_req->TotalParameterCount = 0; in cifs_query_reparse_point()
2725 io_req->TotalDataCount = 0; in cifs_query_reparse_point()
2726 io_req->MaxParameterCount = cpu_to_le32(2); in cifs_query_reparse_point()
2728 io_req->MaxDataCount = cpu_to_le32(CIFSMaxBufSize & 0xFFFFFF00); in cifs_query_reparse_point()
2729 io_req->MaxSetupCount = 4; in cifs_query_reparse_point()
2730 io_req->Reserved = 0; in cifs_query_reparse_point()
2731 io_req->ParameterOffset = 0; in cifs_query_reparse_point()
2732 io_req->DataCount = 0; in cifs_query_reparse_point()
[all …]
/linux/drivers/scsi/megaraid/
H A Dmegaraid_sas_fusion.c3892 struct MPI2_RAID_SCSI_IO_REQUEST *io_req; in build_mpt_mfi_pass_thru() local
3914 io_req = cmd->io_request; in build_mpt_mfi_pass_thru()
3918 (struct MPI25_IEEE_SGE_CHAIN64 *)&io_req->SGL; in build_mpt_mfi_pass_thru()
3924 (struct MPI25_IEEE_SGE_CHAIN64 *)&io_req->SGL.IeeeChain; in build_mpt_mfi_pass_thru()
3926 io_req->Function = MEGASAS_MPI2_FUNCTION_PASSTHRU_IO_REQUEST; in build_mpt_mfi_pass_thru()
3927 io_req->SGLOffset0 = offsetof(struct MPI2_RAID_SCSI_IO_REQUEST, in build_mpt_mfi_pass_thru()
3929 io_req->ChainOffset = fusion->chain_offset_mfi_pthru; in build_mpt_mfi_pass_thru()