| /linux/drivers/misc/genwqe/ |
| H A D | card_utils.c | 291 int genwqe_alloc_sync_sgl(struct genwqe_dev *cd, struct genwqe_sgl *sgl, in genwqe_alloc_sync_sgl() argument 297 sgl->fpage_offs = offset_in_page((unsigned long)user_addr); in genwqe_alloc_sync_sgl() 298 sgl->fpage_size = min_t(size_t, PAGE_SIZE-sgl->fpage_offs, user_size); in genwqe_alloc_sync_sgl() 299 sgl->nr_pages = DIV_ROUND_UP(sgl->fpage_offs + user_size, PAGE_SIZE); in genwqe_alloc_sync_sgl() 300 sgl->lpage_size = (user_size - sgl->fpage_size) % PAGE_SIZE; in genwqe_alloc_sync_sgl() 303 __func__, user_addr, user_size, sgl->nr_pages, in genwqe_alloc_sync_sgl() 304 sgl->fpage_offs, sgl->fpage_size, sgl->lpage_size); in genwqe_alloc_sync_sgl() 306 sgl->user_addr = user_addr; in genwqe_alloc_sync_sgl() 307 sgl->user_size = user_size; in genwqe_alloc_sync_sgl() 308 sgl->write = write; in genwqe_alloc_sync_sgl() [all …]
|
| /linux/drivers/dma-buf/ |
| H A D | dma-buf-mapping.c | 9 static struct scatterlist *fill_sg_entry(struct scatterlist *sgl, size_t length, in fill_sg_entry() argument 26 sg_set_page(sgl, NULL, 0, 0); in fill_sg_entry() 27 sg_dma_address(sgl) = addr + (dma_addr_t)i * UINT_MAX; in fill_sg_entry() 28 sg_dma_len(sgl) = len; in fill_sg_entry() 29 sgl = sg_next(sgl); in fill_sg_entry() 32 return sgl; in fill_sg_entry() 100 struct scatterlist *sgl; in dma_buf_phys_vec_to_sgt() local 140 sgl = dma->sgt.sgl; in dma_buf_phys_vec_to_sgt() 165 sgl = fill_sg_entry(sgl, phys_vec[i].len, addr); in dma_buf_phys_vec_to_sgt() 174 sgl = fill_sg_entry(sgl, mapped_len, dma->state->addr); in dma_buf_phys_vec_to_sgt() [all …]
|
| /linux/crypto/ |
| H A D | af_alg.c | 540 sg_unmark_end(sgl_prev->sgt.sgl + sgl_prev->sgt.nents - 1); in af_alg_link_sg() 541 sg_chain(sgl_prev->sgt.sgl, sgl_prev->sgt.nents + 1, sgl_new->sgt.sgl); in af_alg_link_sg() 544 void af_alg_free_sg(struct af_alg_sgl *sgl) in af_alg_free_sg() argument 548 if (sgl->sgt.sgl) { in af_alg_free_sg() 549 if (sgl->need_unpin) in af_alg_free_sg() 550 for (i = 0; i < sgl->sgt.nents; i++) in af_alg_free_sg() 551 unpin_user_page(sg_page(&sgl->sgt.sgl[i])); in af_alg_free_sg() 552 if (sgl->sgt.sgl != sgl->sgl) in af_alg_free_sg() 553 kvfree(sgl->sgt.sgl); in af_alg_free_sg() 554 sgl->sgt.sgl = NULL; in af_alg_free_sg() [all …]
|
| H A D | algif_hash.c | 20 struct af_alg_sgl sgl; member 106 ctx->sgl.sgt.sgl = ctx->sgl.sgl; in hash_sendmsg() 107 ctx->sgl.sgt.nents = 0; in hash_sendmsg() 108 ctx->sgl.sgt.orig_nents = 0; in hash_sendmsg() 115 sg_init_table(ctx->sgl.sgl, npages); in hash_sendmsg() 117 ctx->sgl.need_unpin = iov_iter_extract_will_pin(&msg->msg_iter); in hash_sendmsg() 120 &ctx->sgl.sgt, npages, 0); in hash_sendmsg() 124 sg_mark_end(ctx->sgl.sgt.sgl + ctx->sgl.sgt.nents - 1); in hash_sendmsg() 132 ahash_request_set_crypt(&ctx->req, ctx->sgl.sgt.sgl, in hash_sendmsg() 160 af_alg_free_sg(&ctx->sgl); in hash_sendmsg() [all …]
|
| /linux/drivers/crypto/hisilicon/ |
| H A D | sgl.c | 37 struct hisi_acc_hw_sgl *sgl; member 92 block[i].sgl = dma_alloc_coherent(dev, block_size, in hisi_acc_create_sgl_pool() 95 if (!block[i].sgl) { in hisi_acc_create_sgl_pool() 104 block[i].sgl = dma_alloc_coherent(dev, remain_sgl * sgl_size, in hisi_acc_create_sgl_pool() 107 if (!block[i].sgl) { in hisi_acc_create_sgl_pool() 125 dma_free_coherent(dev, block_size, block[j].sgl, in hisi_acc_create_sgl_pool() 151 dma_free_coherent(dev, block[i].size, block[i].sgl, in hisi_acc_free_sgl_pool() 169 return (void *)block[block_index].sgl + pool->sgl_size * offset; in acc_get_sgl() 172 static void sg_map_to_hw_sg(struct scatterlist *sgl, in sg_map_to_hw_sg() argument 175 hw_sge->buf = sg_dma_address(sgl); in sg_map_to_hw_sg() [all …]
|
| /linux/mm/ |
| H A D | page_reporting.c | 106 struct scatterlist *sgl, unsigned int nents, bool reported) in page_reporting_drain() argument 108 struct scatterlist *sg = sgl; in page_reporting_drain() 137 sg_init_table(sgl, nents); in page_reporting_drain() 148 struct scatterlist *sgl, unsigned int *offset) in page_reporting_cycle() argument 208 sg_set_page(&sgl[*offset], page, page_len, 0); in page_reporting_cycle() 225 err = prdev->report(prdev, sgl, PAGE_REPORTING_CAPACITY); in page_reporting_cycle() 237 page_reporting_drain(prdev, sgl, PAGE_REPORTING_CAPACITY, !err); in page_reporting_cycle() 261 struct scatterlist *sgl, struct zone *zone) in page_reporting_process_zone() argument 286 sgl, &offset); in page_reporting_process_zone() 295 sgl = &sgl[offset]; in page_reporting_process_zone() [all …]
|
| /linux/lib/ |
| H A D | scatterlist.c | 82 int sg_nents_for_dma(struct scatterlist *sgl, unsigned int sglen, size_t len) in sg_nents_for_dma() argument 87 for_each_sg(sgl, sg, sglen, i) in sg_nents_for_dma() 107 struct scatterlist *sg_last(struct scatterlist *sgl, unsigned int nents) in sg_last() argument 112 for_each_sg(sgl, sg, nents, i) in sg_last() 130 void sg_init_table(struct scatterlist *sgl, unsigned int nents) in sg_init_table() argument 132 memset(sgl, 0, sizeof(*sgl) * nents); in sg_init_table() 133 sg_init_marker(sgl, nents); in sg_init_table() 202 struct scatterlist *sgl, *next; in __sg_free_table() local 205 if (unlikely(!table->sgl)) in __sg_free_table() 208 sgl = table->sgl; in __sg_free_table() [all …]
|
| H A D | kfifo.c | 309 static unsigned int setup_sgl_buf(struct __kfifo *fifo, struct scatterlist *sgl, in setup_sgl_buf() 318 sg_set_buf(sgl, buf, len); in setup_sgl_buf() 321 sg_dma_address(sgl) = dma + data_offset; in setup_sgl_buf() 322 sg_dma_len(sgl) = len; in setup_sgl_buf() 328 static unsigned int setup_sgl(struct __kfifo *fifo, struct scatterlist *sgl, in setup_sgl_buf() 344 n = setup_sgl_buf(fifo, sgl, off, nents, len_to_end, dma); in setup_sgl() 345 n += setup_sgl_buf(fifo, sgl + n, 0, nents - n, len - len_to_end, dma); in setup_sgl() 351 struct scatterlist *sgl, int nents, unsigned int len, in setup_sgl() 360 return setup_sgl(fifo, sgl, nents, len, fifo->in, dma); in __kfifo_dma_in_prepare() 365 struct scatterlist *sgl, in in __kfifo_dma_in_prepare() 295 setup_sgl_buf(struct scatterlist * sgl,void * buf,int nents,unsigned int len) setup_sgl_buf() argument 334 setup_sgl(struct __kfifo * fifo,struct scatterlist * sgl,int nents,unsigned int len,unsigned int off) setup_sgl() argument 357 __kfifo_dma_in_prepare(struct __kfifo * fifo,struct scatterlist * sgl,int nents,unsigned int len) __kfifo_dma_in_prepare() argument 370 __kfifo_dma_out_prepare(struct __kfifo * fifo,struct scatterlist * sgl,int nents,unsigned int len) __kfifo_dma_out_prepare() argument 549 __kfifo_dma_in_prepare_r(struct __kfifo * fifo,struct scatterlist * sgl,int nents,unsigned int len,size_t recsize) __kfifo_dma_in_prepare_r() argument 572 __kfifo_dma_out_prepare_r(struct __kfifo * fifo,struct scatterlist * sgl,int nents,unsigned int len,size_t recsize) __kfifo_dma_out_prepare_r() argument [all...] |
| /linux/drivers/gpu/drm/nouveau/nvkm/core/ |
| H A D | firmware.c | 121 case NVKM_FIRMWARE_IMG_DMA: return &fw->mem.sgl; in nvkm_firmware_mem_sgl() 122 case NVKM_FIRMWARE_IMG_SGT: return fw->mem.sgt.sgl; in nvkm_firmware_mem_sgl() 139 .sgl = nvkm_firmware_mem_sgl(memory), in nvkm_firmware_mem_map() 142 if (!map.sgl) in nvkm_firmware_mem_map() 151 struct scatterlist *sgl = nvkm_firmware_mem_sgl(memory); in nvkm_firmware_mem_size() local 153 return sgl ? sg_dma_len(sgl) : 0; in nvkm_firmware_mem_size() 208 dma_free_noncoherent(fw->device->dev, sg_dma_len(&fw->mem.sgl), in nvkm_firmware_dtor() 251 sg_init_one(&fw->mem.sgl, fw->img, len); in nvkm_firmware_ctor() 252 sg_dma_address(&fw->mem.sgl) = fw->phys; in nvkm_firmware_ctor() 253 sg_dma_len(&fw->mem.sgl) = len; in nvkm_firmware_ctor() [all …]
|
| /linux/drivers/target/iscsi/cxgbit/ |
| H A D | cxgbit_ddp.c | 133 struct scatterlist *sg = ttinfo->sgl; in cxgbit_ddp_set_map() 176 struct scatterlist *sgl = ttinfo->sgl; in cxgbit_ddp_reserve() local 178 unsigned int sg_offset = sgl->offset; in cxgbit_ddp_reserve() 188 if (cxgbit_ddp_sgl_check(sgl, sgcnt) < 0) in cxgbit_ddp_reserve() 191 ttinfo->nr_pages = (xferlen + sgl->offset + in cxgbit_ddp_reserve() 203 sgl->offset = 0; in cxgbit_ddp_reserve() 204 ret = dma_map_sg(&ppm->pdev->dev, sgl, sgcnt, DMA_FROM_DEVICE); in cxgbit_ddp_reserve() 205 sgl->offset = sg_offset; in cxgbit_ddp_reserve() 212 cxgbi_ppm_make_ppod_hdr(ppm, ttinfo->tag, csk->tid, sgl->offset, in cxgbit_ddp_reserve() 218 dma_unmap_sg(&ppm->pdev->dev, sgl, sgcnt, DMA_FROM_DEVICE); in cxgbit_ddp_reserve() [all …]
|
| /linux/tools/virtio/linux/ |
| H A D | scatterlist.h | 92 struct scatterlist *sgl) in sg_chain() argument 104 prv[prv_nents - 1].page_link = ((unsigned long) sgl | 0x01) & ~0x02; in sg_chain() 150 static inline void sg_init_table(struct scatterlist *sgl, unsigned int nents) in sg_init_table() argument 152 memset(sgl, 0, sizeof(*sgl) * nents); in sg_init_table() 153 sg_mark_end(&sgl[nents - 1]); in sg_init_table()
|
| /linux/drivers/spi/ |
| H A D | spi-bcm2835.c | 497 if (bs->tx_buf && !sg_is_last(&tfr->tx_sg.sgl[0])) in bcm2835_spi_transfer_prologue() 498 bs->tx_prologue = sg_dma_len(&tfr->tx_sg.sgl[0]) & 3; in bcm2835_spi_transfer_prologue() 500 if (bs->rx_buf && !sg_is_last(&tfr->rx_sg.sgl[0])) { in bcm2835_spi_transfer_prologue() 501 bs->rx_prologue = sg_dma_len(&tfr->rx_sg.sgl[0]) & 3; in bcm2835_spi_transfer_prologue() 504 if (!bs->tx_buf || sg_is_last(&tfr->tx_sg.sgl[0])) { in bcm2835_spi_transfer_prologue() 509 !(sg_dma_len(&tfr->tx_sg.sgl[0]) & ~3); in bcm2835_spi_transfer_prologue() 531 sg_dma_address(&tfr->rx_sg.sgl[0]), in bcm2835_spi_transfer_prologue() 534 sg_dma_address(&tfr->rx_sg.sgl[0]) += bs->rx_prologue; in bcm2835_spi_transfer_prologue() 535 sg_dma_len(&tfr->rx_sg.sgl[0]) -= bs->rx_prologue; in bcm2835_spi_transfer_prologue() 557 sg_dma_address(&tfr->tx_sg.sgl[0]) += bs->tx_prologue; in bcm2835_spi_transfer_prologue() [all …]
|
| /linux/drivers/scsi/lpfc/ |
| H A D | lpfc_scsi.c | 103 struct sli4_sge *sgl = (struct sli4_sge *)lpfc_cmd->dma_sgl; in lpfc_sli4_set_rsp_sgl_last() local 104 if (sgl) { in lpfc_sli4_set_rsp_sgl_last() 105 sgl += 1; in lpfc_sli4_set_rsp_sgl_last() 106 sgl->word2 = le32_to_cpu(sgl->word2); in lpfc_sli4_set_rsp_sgl_last() 107 bf_set(lpfc_sli4_sge_last, sgl, 1); in lpfc_sli4_set_rsp_sgl_last() 108 sgl->word2 = cpu_to_le32(sgl->word2); in lpfc_sli4_set_rsp_sgl_last() 608 struct sli4_sge_le *sgl; in lpfc_get_scsi_buf_s4() local 659 sgl = (struct sli4_sge_le *)lpfc_cmd->dma_sgl; in lpfc_get_scsi_buf_s4() 661 sgl->addr_hi = cpu_to_le32(putPaddrHigh(pdma_phys_fcp_cmd)); in lpfc_get_scsi_buf_s4() 662 sgl->addr_lo = cpu_to_le32(putPaddrLow(pdma_phys_fcp_cmd)); in lpfc_get_scsi_buf_s4() [all …]
|
| H A D | lpfc_nvme.c | 830 struct sli4_sge *sgl; in lpfc_nvme_adj_fcp_sgls() local 849 sgl = lpfc_ncmd->dma_sgl; in lpfc_nvme_adj_fcp_sgls() 850 sgl->sge_len = cpu_to_le32(nCmd->cmdlen); in lpfc_nvme_adj_fcp_sgls() 852 sgl->addr_hi = 0; in lpfc_nvme_adj_fcp_sgls() 853 sgl->addr_lo = 0; in lpfc_nvme_adj_fcp_sgls() 892 sgl->addr_hi = cpu_to_le32(putPaddrHigh(nCmd->cmddma)); in lpfc_nvme_adj_fcp_sgls() 893 sgl->addr_lo = cpu_to_le32(putPaddrLow(nCmd->cmddma)); in lpfc_nvme_adj_fcp_sgls() 898 wqe->generic.bde.addrHigh = sgl->addr_hi; in lpfc_nvme_adj_fcp_sgls() 899 wqe->generic.bde.addrLow = sgl->addr_lo; in lpfc_nvme_adj_fcp_sgls() 906 sgl++; in lpfc_nvme_adj_fcp_sgls() [all …]
|
| /linux/include/trace/events/ |
| H A D | dma.h | 181 for_each_sg(sgt->sgl, sg, sgt->orig_nents, i) 183 __entry->dma_addr = sg_dma_address(sgt->sgl); 261 for_each_sg(sgt->sgl, sg, sgt->orig_nents, i) 263 __entry->dma_addr = sg_dma_address(sgt->sgl); 281 TP_PROTO(struct device *dev, struct scatterlist *sgl, int nents, 283 TP_ARGS(dev, sgl, nents, ents, dir, attrs), 307 for_each_sg(sgl, sg, traced_nents, i) 309 for_each_sg(sgl, sg, traced_ents, i) { 338 TP_PROTO(struct device *dev, struct scatterlist *sgl, int nents, 340 TP_ARGS(dev, sgl, nents, err, dir, attrs), [all …]
|
| /linux/drivers/media/platform/ |
| H A D | m2m-deinterlace.c | 241 ctx->xt->sgl[0].size = s_width; in deinterlace_issue_dma() 242 ctx->xt->sgl[0].icg = s_width; in deinterlace_issue_dma() 248 ctx->xt->sgl[0].size = s_width; in deinterlace_issue_dma() 249 ctx->xt->sgl[0].icg = s_width; in deinterlace_issue_dma() 255 ctx->xt->sgl[0].size = s_width / 2; in deinterlace_issue_dma() 256 ctx->xt->sgl[0].icg = s_width / 2; in deinterlace_issue_dma() 262 ctx->xt->sgl[0].size = s_width / 2; in deinterlace_issue_dma() 263 ctx->xt->sgl[0].icg = s_width / 2; in deinterlace_issue_dma() 269 ctx->xt->sgl[0].size = s_width / 2; in deinterlace_issue_dma() 270 ctx->xt->sgl[0].icg = s_width / 2; in deinterlace_issue_dma() [all …]
|
| /linux/drivers/vfio/pci/pds/ |
| H A D | cmds.c | 206 struct pds_lm_sg_elem *sgl, *sge; in pds_vfio_dma_map_lm_file() local 225 sgl = kzalloc(sgl_size, GFP_KERNEL); in pds_vfio_dma_map_lm_file() 226 if (!sgl) { in pds_vfio_dma_map_lm_file() 232 sge = sgl; in pds_vfio_dma_map_lm_file() 240 sgl_addr = dma_map_single(dev, sgl, sgl_size, DMA_TO_DEVICE); in pds_vfio_dma_map_lm_file() 246 lm_file->sgl = sgl; in pds_vfio_dma_map_lm_file() 252 kfree(sgl); in pds_vfio_dma_map_lm_file() 267 if (lm_file->sgl) { in pds_vfio_dma_unmap_lm_file() 269 lm_file->num_sge * sizeof(*lm_file->sgl), in pds_vfio_dma_unmap_lm_file() 271 kfree(lm_file->sgl); in pds_vfio_dma_unmap_lm_file() [all …]
|
| /linux/drivers/xen/ |
| H A D | swiotlb-xen.c | 346 xen_swiotlb_unmap_sg(struct device *hwdev, struct scatterlist *sgl, int nelems, in xen_swiotlb_unmap_sg() argument 354 for_each_sg(sgl, sg, nelems, i) in xen_swiotlb_unmap_sg() 361 xen_swiotlb_map_sg(struct device *dev, struct scatterlist *sgl, int nelems, in xen_swiotlb_map_sg() argument 369 for_each_sg(sgl, sg, nelems, i) { in xen_swiotlb_map_sg() 379 xen_swiotlb_unmap_sg(dev, sgl, i, dir, attrs | DMA_ATTR_SKIP_CPU_SYNC); in xen_swiotlb_map_sg() 380 sg_dma_len(sgl) = 0; in xen_swiotlb_map_sg() 385 xen_swiotlb_sync_sg_for_cpu(struct device *dev, struct scatterlist *sgl, in xen_swiotlb_sync_sg_for_cpu() argument 391 for_each_sg(sgl, sg, nelems, i) { in xen_swiotlb_sync_sg_for_cpu() 398 xen_swiotlb_sync_sg_for_device(struct device *dev, struct scatterlist *sgl, in xen_swiotlb_sync_sg_for_device() argument 404 for_each_sg(sgl, sg, nelems, i) { in xen_swiotlb_sync_sg_for_device()
|
| /linux/drivers/dma/ |
| H A D | loongson1-apb-dma.c | 217 struct scatterlist *sgl, unsigned int sg_len, in ls1x_dma_prep_lli() argument 244 for_each_sg(sgl, sg, sg_len, i) { in ls1x_dma_prep_lli() 296 ls1x_dma_prep_slave_sg(struct dma_chan *dchan, struct scatterlist *sgl, in ls1x_dma_prep_slave_sg() argument 309 if (ls1x_dma_prep_lli(dchan, desc, sgl, sg_len, dir, false)) { in ls1x_dma_prep_slave_sg() 323 struct scatterlist *sgl; in ls1x_dma_prep_dma_cyclic() local 338 sgl = kmalloc_objs(*sgl, sg_len, GFP_NOWAIT); in ls1x_dma_prep_dma_cyclic() 339 if (!sgl) in ls1x_dma_prep_dma_cyclic() 342 sg_init_table(sgl, sg_len); in ls1x_dma_prep_dma_cyclic() 344 sg_set_page(&sgl[i], pfn_to_page(PFN_DOWN(buf_addr)), in ls1x_dma_prep_dma_cyclic() 346 sg_dma_address(&sgl[i]) = buf_addr; in ls1x_dma_prep_dma_cyclic() [all …]
|
| H A D | milbeaut-hdmac.c | 61 struct scatterlist *sgl; member 122 sg = &md->sgl[md->sg_cur]; in milbeaut_chan_start() 256 milbeaut_hdmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, in milbeaut_hdmac_prep_slave_sg() argument 272 md->sgl = kzalloc_objs(*sgl, sg_len, GFP_NOWAIT); in milbeaut_hdmac_prep_slave_sg() 273 if (!md->sgl) { in milbeaut_hdmac_prep_slave_sg() 279 md->sgl[i] = sgl[i]; in milbeaut_hdmac_prep_slave_sg() 350 sg = &md->sgl[md->sg_cur]; in milbeaut_hdmac_tx_status() 372 txstate->residue += sg_dma_len(&md->sgl[i]); in milbeaut_hdmac_tx_status() 398 kfree(md->sgl); in milbeaut_hdmac_desc_free()
|
| /linux/drivers/mtd/nand/ |
| H A D | qpic_common.c | 152 struct scatterlist *sgl; in qcom_prepare_bam_async_desc() local 164 sgl = &bam_txn->cmd_sgl[bam_txn->cmd_sgl_start]; in qcom_prepare_bam_async_desc() 170 sgl = &bam_txn->data_sgl[bam_txn->tx_sgl_start]; in qcom_prepare_bam_async_desc() 176 sgl = &bam_txn->data_sgl[bam_txn->rx_sgl_start]; in qcom_prepare_bam_async_desc() 183 sg_mark_end(sgl + sgl_cnt - 1); in qcom_prepare_bam_async_desc() 184 ret = dma_map_sg(nandc->dev, sgl, sgl_cnt, desc->dir); in qcom_prepare_bam_async_desc() 192 desc->bam_sgl = sgl; in qcom_prepare_bam_async_desc() 194 dma_desc = dmaengine_prep_slave_sg(chan, sgl, sgl_cnt, dir_eng, in qcom_prepare_bam_async_desc() 199 dma_unmap_sg(nandc->dev, sgl, sgl_cnt, desc->dir); in qcom_prepare_bam_async_desc() 364 struct scatterlist *sgl; in qcom_prep_adm_dma_desc() local [all …]
|
| /linux/drivers/net/ethernet/intel/ixgbe/ |
| H A D | ixgbe_fcoe.c | 29 ddp->sgl = NULL; in ixgbe_fcoe_clear_ddp() 110 if (ddp->sgl) in ixgbe_fcoe_ddp_put() 111 dma_unmap_sg(&adapter->pdev->dev, ddp->sgl, ddp->sgc, in ixgbe_fcoe_ddp_put() 134 struct scatterlist *sgl, unsigned int sgc, in ixgbe_fcoe_ddp_setup() argument 153 if (!netdev || !sgl) in ixgbe_fcoe_ddp_setup() 169 if (ddp->sgl) { in ixgbe_fcoe_ddp_setup() 171 xid, ddp->sgl, ddp->sgc); in ixgbe_fcoe_ddp_setup() 189 dmacount = dma_map_sg(&adapter->pdev->dev, sgl, sgc, DMA_FROM_DEVICE); in ixgbe_fcoe_ddp_setup() 202 ddp->sgl = sgl; in ixgbe_fcoe_ddp_setup() 206 for_each_sg(sgl, sg, dmacount, i) { in ixgbe_fcoe_ddp_setup() [all …]
|
| /linux/drivers/iio/buffer/ |
| H A D | industrialio-buffer-dmaengine.c | 66 struct scatterlist *sgl; in iio_dmaengine_buffer_submit_block() local 83 sgl = block->sg_table->sgl; in iio_dmaengine_buffer_submit_block() 84 nents = sg_nents_for_len(sgl, block->bytes_used); in iio_dmaengine_buffer_submit_block() 95 vecs[i].addr = sg_dma_address(sgl); in iio_dmaengine_buffer_submit_block() 96 vecs[i].len = min(sg_dma_len(sgl), len_total); in iio_dmaengine_buffer_submit_block() 99 sgl = sg_next(sgl); in iio_dmaengine_buffer_submit_block()
|
| /linux/drivers/scsi/esas2r/ |
| H A D | esas2r_io.c | 224 struct esas2r_mem_desc *sgl; in esas2r_build_sg_list_sge() local 231 sgl = esas2r_alloc_sgl(a); in esas2r_build_sg_list_sge() 233 if (unlikely(sgl == NULL)) in esas2r_build_sg_list_sge() 244 memcpy(sgl->virt_addr, sgc->sge.a64.last, sgelen); in esas2r_build_sg_list_sge() 248 (struct atto_vda_sge *)((u8 *)sgl->virt_addr + in esas2r_build_sg_list_sge() 253 (struct atto_vda_sge *)((u8 *)sgl->virt_addr in esas2r_build_sg_list_sge() 260 cpu_to_le64(sgl->phys_addr); in esas2r_build_sg_list_sge() 302 list_add(&sgl->next_desc, &rq->sg_table_head); in esas2r_build_sg_list_sge() 376 struct esas2r_mem_desc *sgl; in esas2r_build_prd_iblk() local 449 sgl = esas2r_alloc_sgl(a); in esas2r_build_prd_iblk() [all …]
|
| /linux/drivers/gpu/drm/i915/ |
| H A D | i915_scatterlist.h | 31 } __sgt_iter(struct scatterlist *sgl, bool dma) { in __sgt_iter() argument 32 struct sgt_iter s = { .sgp = sgl }; in __sgt_iter() 90 for ((__iter) = __sgt_iter((__sgt)->sgl, true); \ 112 for ((__iter) = __sgt_iter((__sgt)->sgl, false); \ 225 rsgt->table.sgl = NULL; in __i915_refct_sgt_init()
|