| /linux/drivers/scsi/csiostor/ |
| H A D | csio_scsi.c | 178 if (req->nsge) in csio_scsi_fcp_cmnd() 302 ULPTX_NSGE_V(req->nsge)); in csio_scsi_init_ultptx_dsgl() 305 scsi_for_each_sg(scmnd, sgel, req->nsge, i) { in csio_scsi_init_ultptx_dsgl() 464 if (unlikely((req)->nsge > 1)) \ 466 (ALIGN(((req)->nsge - 1), 2) / 2)); \ 569 scsi_for_each_sg(scmnd, sgel, req->nsge, i) { in csio_setup_ddp() 583 if ((i != (req->nsge - 1)) && in csio_setup_ddp() 626 req->nsge = i; in csio_setup_ddp() 711 if (req->nsge) { in csio_scsis_uninit() 1720 if (req->nsge > 0) { in csio_scsi_err_handler() [all …]
|
| H A D | csio_wr.h | 245 uint32_t nsge; /* Number of SG elements */ member
|
| H A D | csio_isr.c | 235 ioreq->nsge); in csio_scsi_isr_handler()
|
| /linux/net/tls/ |
| H A D | tls_sw.c | 612 struct scatterlist *sge, *osge, *nsge; in tls_split_open_record() local 664 nsge = sk_msg_elem(msg_npl, j); in tls_split_open_record() 666 memcpy(nsge, &tmp, sizeof(*nsge)); in tls_split_open_record() 668 nsge = sk_msg_elem(msg_npl, j); in tls_split_open_record() 673 memcpy(nsge, osge, sizeof(*nsge)); in tls_split_open_record() 674 sg_unmark_end(nsge); in tls_split_open_record() 680 nsge = sk_msg_elem(msg_npl, j); in tls_split_open_record() 696 struct scatterlist *osge, *nsge; in tls_merge_open_record() local 704 nsge = sk_msg_elem(msg_npl, j); in tls_merge_open_record() 706 if (sg_page(osge) == sg_page(nsge) && in tls_merge_open_record() [all …]
|
| /linux/drivers/infiniband/hw/cxgb4/ |
| H A D | t4fw_ri_api.h | 122 __be16 nsge; member 137 __be16 nsge; member
|
| H A D | qp.c | 479 isglp->nsge = cpu_to_be16(num_sge); in build_isgl() 850 sglp->nsge = cpu_to_be16(1); in build_memreg()
|
| /linux/drivers/scsi/ |
| H A D | myrb.c | 1275 int nsge; in myrb_pthru_queuecommand() local 1281 nsge = scsi_dma_map(scmd); in myrb_pthru_queuecommand() 1282 if (nsge > 1) { in myrb_pthru_queuecommand() 1431 int nsge; in myrb_ldev_queuecommand() local 1557 nsge = scsi_dma_map(scmd); in myrb_ldev_queuecommand() 1558 if (nsge == 1) { in myrb_ldev_queuecommand() 1590 mbox->type5.sg_count = nsge; in myrb_ldev_queuecommand() 1592 scsi_for_each_sg(scmd, sgl, nsge, i) { in myrb_ldev_queuecommand()
|
| H A D | myrs.c | 1596 int nsge; in myrs_queuecommand() local 1718 nsge = scsi_dma_map(scmd); in myrs_queuecommand() 1719 if (nsge == 1) { in myrs_queuecommand() 1728 if (nsge > 2) { in myrs_queuecommand() 1752 hw_sge->ext.sge0_len = nsge; in myrs_queuecommand() 1757 scsi_for_each_sg(scmd, sgl, nsge, i) { in myrs_queuecommand()
|
| /linux/drivers/infiniband/hw/bnxt_re/ |
| H A D | qplib_fp.c | 973 u16 nsge; in bnxt_qplib_create_qp() local 1068 nsge = (qp->wqe_mode == BNXT_QPLIB_WQE_MODE_STATIC) ? in bnxt_qplib_create_qp() 1071 cpu_to_le16(((nsge & in bnxt_qplib_create_qp() 1788 u32 nsge, u32 *idx) in bnxt_qplib_put_sges() argument 1793 for (indx = 0; indx < nsge; indx++, (*idx)++) { in bnxt_qplib_put_sges() 1809 u16 nsge; in bnxt_qplib_required_slots() local 1812 nsge = wqe->num_sge; in bnxt_qplib_required_slots() 1814 bytes = sizeof(struct sq_send_hdr) + nsge * sizeof(struct sq_sge); in bnxt_qplib_required_slots()
|
| H A D | ib_verbs.c | 1068 static u16 bnxt_re_get_wqe_size(int ilsize, int nsge) in bnxt_re_get_wqe_size() argument 1072 wqe_size = bnxt_re_get_swqe_size(nsge); in bnxt_re_get_wqe_size()
|
| /linux/drivers/net/ethernet/pensando/ionic/ |
| H A D | ionic_txrx.c | 1392 dma_addr_t addr, u8 nsge, u16 len, in ionic_tx_tso_post() argument 1406 cmd = encode_txq_desc_cmd(IONIC_TXQ_DESC_OPCODE_TSO, flags, nsge, addr); in ionic_tx_tso_post()
|
| /linux/net/core/ |
| H A D | filter.c | 2791 struct scatterlist sge, nsge, nnsge, rsge = {0}, *psge; in BPF_CALL_4() local 2891 nsge = sk_msg_elem_cpy(msg, i); in BPF_CALL_4() 2900 sge = nsge; in BPF_CALL_4() 2903 nsge = nnsge; in BPF_CALL_4() 2906 nsge = sk_msg_elem_cpy(msg, i); in BPF_CALL_4() 3022 struct scatterlist *nsge, *sge = sk_msg_elem(msg, i); in BPF_CALL_4() local 3032 nsge = sk_msg_elem(msg, i); in BPF_CALL_4() 3034 sg_set_page(nsge, in BPF_CALL_4()
|