| /linux/lib/ |
| H A D | sg_split.c | 25 unsigned int sglen; in sg_calculate_split() local 36 sglen = mapped ? sg_dma_len(sg) : sg->length; in sg_calculate_split() 37 if (skip > sglen) { in sg_calculate_split() 38 skip -= sglen; in sg_calculate_split() 42 len = min_t(size_t, size, sglen - skip); in sg_calculate_split() 51 while (!size && (skip + len < sglen) && (--nb_splits > 0)) { in sg_calculate_split() 55 len = min_t(size_t, size, sglen - skip); in sg_calculate_split()
|
| /linux/drivers/usb/misc/ |
| H A D | usbtest.c | 34 __u32 sglen; member 51 __u32 sglen; member 1157 for (i = 1; i < ctx->param->sglen; i++) { in ctrl_complete() 1160 % ctx->param->sglen]; in ctrl_complete() 1208 if (param->sglen == 0 || param->iterations > UINT_MAX / param->sglen) in test_ctrl_queue() 1214 context.count = param->sglen * param->iterations; in test_ctrl_queue() 1221 * as with bulk/intr sglists, sglen is the queue depth; it also in test_ctrl_queue() 1222 * controls which subtests run (more tests than sglen) or rerun. in test_ctrl_queue() 1224 urb = kzalloc_objs(struct urb *, param->sglen); in test_ctrl_queue() 1227 for (i = 0; i < param->sglen; i++) { in test_ctrl_queue() [all …]
|
| /linux/drivers/misc/bcm-vk/ |
| H A D | bcm_vk_sg.c | 85 dma->sglen = (dma->nr_pages * sizeof(*sgdata)) + in bcm_vk_dma_alloc() 90 dma->sglen, in bcm_vk_dma_alloc() 160 "sgl 0x%llx handle 0x%llx, sglen: 0x%x sgsize: 0x%x\n", in bcm_vk_dma_alloc() 163 dma->sglen, in bcm_vk_dma_alloc() 232 dev_dbg(dev, "free sglist=%p sglen=0x%x\n", dma->sglist, dma->sglen); in bcm_vk_dma_free() 245 dma_free_coherent(dev, dma->sglen, dma->sglist, dma->handle); in bcm_vk_dma_free()
|
| H A D | bcm_vk_sg.h | 35 int sglen; /* Length (bytes) of sglist */ member
|
| /linux/drivers/dma/ |
| H A D | sa11x0-dma.c | 80 unsigned sglen; member 81 struct sa11x0_dma_sg sg[] __counted_by(sglen); 176 if (p->sg_load == txd->sglen) { in sa11x0_dma_start_sg() 227 if (++p->sg_done == txd->sglen) { in sa11x0_dma_complete() 464 for (i = 0; i < txd->sglen; i++) { in sa11x0_dma_tx_status() 480 for (; i < txd->sglen; i++) { in sa11x0_dma_tx_status() 523 struct dma_chan *chan, struct scatterlist *sg, unsigned int sglen, in sa11x0_dma_prep_slave_sg() argument 540 if (sglen == 0) in sa11x0_dma_prep_slave_sg() 543 for_each_sg(sg, sgent, sglen, i) { in sa11x0_dma_prep_slave_sg() 553 j = sg_nents_for_dma(sg, sglen, DMA_MAX_SIZE & ~DMA_ALIGN); in sa11x0_dma_prep_slave_sg() [all …]
|
| H A D | sprd-dma.c | 744 unsigned int sglen, int sg_index, in sprd_dma_fill_desc() argument 846 temp = sglen ? (sg_index + 1) % sglen : 0; in sprd_dma_fill_desc() 875 unsigned int sglen, int sg_index, in sprd_dma_fill_linklist_desc() argument 890 return sprd_dma_fill_desc(chan, hw, sglen, sg_index, src, dst, len, in sprd_dma_fill_linklist_desc() 951 unsigned int sglen, enum dma_transfer_direction dir, in sprd_dma_prep_slave_sg() argument 995 for_each_sg(sgl, sg, sglen, i) { in sprd_dma_prep_slave_sg() 1016 if (sglen < 2) in sprd_dma_prep_slave_sg() 1019 ret = sprd_dma_fill_linklist_desc(chan, sglen, i, src, dst, len, in sprd_dma_prep_slave_sg()
|
| H A D | moxart-dma.c | 123 unsigned int sglen; member 127 struct moxart_sg sg[] __counted_by(sglen); 307 d->sglen = sg_len; in moxart_prep_slave_sg() 447 for (size = i = completed_sgs; i < d->sglen; i++) in moxart_dma_desc_size() 535 if (++ch->sgidx < ch->desc->sglen) { in moxart_dma_interrupt()
|
| H A D | at_hdmac.c | 225 * @sglen: number of sg entries. 238 unsigned int sglen; member 248 struct atdma_sg sg[] __counted_by(sglen); 598 for (i = 0; i < desc->sglen; i++) { in atdma_desc_free() 722 for (i = 1; i < desc->sglen; i++) { in atc_get_llis_residue() 796 for (i = 0; i < desc->sglen; i++) in atc_handle_error() 935 desc->sglen = 1; in atc_prep_dma_interleaved() 998 desc->sglen = sg_len; in atc_prep_dma_memcpy() 1137 desc->sglen = 1; in atc_prep_dma_memset() 1197 desc->sglen = sg_len; in atc_prep_dma_memset_sg() [all …]
|
| H A D | k3dma.c | 531 struct dma_chan *chan, struct scatterlist *sgl, unsigned int sglen, in k3_dma_prep_slave_sg() argument 546 num = sg_nents_for_dma(sgl, sglen, DMA_MAX_SIZE); in k3_dma_prep_slave_sg() 553 for_each_sg(sgl, sg, sglen, i) { in k3_dma_prep_slave_sg()
|
| /linux/drivers/dma/ti/ |
| H A D | omap-dma.c | 126 unsigned sglen; member 127 struct omap_sg sg[] __counted_by(sglen); 256 for (i = 0; i < d->sglen; i++) { in omap_dma_desc_free() 621 } else if (d->using_ll || c->sgidx == d->sglen) { in omap_dma_callback() 796 for (size = i = 0; i < d->sglen; i++) in omap_dma_desc_size() 807 for (size = i = 0; i < d->sglen; i++) { in omap_dma_desc_size_pos() 960 struct dma_chan *chan, struct scatterlist *sgl, unsigned sglen, in omap_dma_prep_slave_sg() argument 1005 d = kzalloc_flex(*d, sg, sglen, GFP_ATOMIC); in omap_dma_prep_slave_sg() 1008 d->sglen = sglen; in omap_dma_prep_slave_sg() 1099 if (sglen >= 2) in omap_dma_prep_slave_sg() [all …]
|
| H A D | k3-udma.c | 226 unsigned int sglen; member 877 for (i = 0; i < uc->desc->sglen; i++) in udma_start_desc() 1049 d->desc_idx = (d->desc_idx + 1) % d->sglen; in udma_cyclic_packet_elapsed() 1228 d->tr_idx = (d->tr_idx + 1) % d->sglen; in udma_udma_irq_handler() 2787 d->sglen = tr_count; in udma_alloc_tr_desc() 2884 unsigned int sglen, enum dma_transfer_direction dir, in udma_prep_slave_sg_tr() argument 2898 for_each_sg(sgl, sgent, sglen, i) { in udma_prep_slave_sg_tr() 2911 d->sglen = sglen; in udma_prep_slave_sg_tr() 2919 for_each_sg(sgl, sgent, sglen, i) { in udma_prep_slave_sg_tr() 2968 unsigned int sglen, in udma_prep_slave_sg_triggered_tr() argument [all …]
|
| /linux/drivers/crypto/cavium/nitrox/ |
| H A D | nitrox_req.h | 595 unsigned int sglen; in create_multi_sg() local 597 for (; buflen && from_sg; buflen -= sglen) { in create_multi_sg() 598 sglen = from_sg->length; in create_multi_sg() 599 if (sglen > buflen) in create_multi_sg() 600 sglen = buflen; in create_multi_sg() 602 sg_set_buf(sg, sg_virt(from_sg), sglen); in create_multi_sg()
|
| /linux/tools/usb/ |
| H A D | testusb.c | 48 unsigned sglen; member 396 param.sglen = 32; in main() 418 if (parse_num(¶m.sglen, optarg)) in main() 456 "\t-g sglen default 32\n" in main()
|
| /linux/drivers/media/pci/cx23885/ |
| H A D | cx23885-alsa.c | 116 buf->sglen = dma_map_sg(&dev->pci->dev, buf->sglist, in cx23885_alsa_dma_map() 119 if (0 == buf->sglen) { in cx23885_alsa_dma_map() 130 if (!buf->sglen) in cx23885_alsa_dma_unmap() 134 buf->sglen = 0; in cx23885_alsa_dma_unmap()
|
| /linux/drivers/media/pci/cx25821/ |
| H A D | cx25821-alsa.c | 56 int sglen; member 179 buf->sglen = dma_map_sg(&dev->pci->dev, buf->sglist, in cx25821_alsa_dma_map() 182 if (0 == buf->sglen) { in cx25821_alsa_dma_map() 193 if (!buf->sglen) in cx25821_alsa_dma_unmap() 197 buf->sglen = 0; in cx25821_alsa_dma_unmap()
|
| /linux/drivers/media/pci/cx88/ |
| H A D | cx88-alsa.c | 49 int sglen; member 318 buf->sglen = dma_map_sg(&dev->pci->dev, buf->sglist, in cx88_alsa_dma_map() 321 if (buf->sglen == 0) { in cx88_alsa_dma_map() 332 if (!buf->sglen) in cx88_alsa_dma_unmap() 337 buf->sglen = 0; in cx88_alsa_dma_unmap()
|
| /linux/drivers/media/pci/cobalt/ |
| H A D | cobalt-omnitek.h | 35 struct scatterlist *scatter_list, bool to_pci, unsigned sglen,
|
| H A D | cobalt-omnitek.c | 150 struct scatterlist *scatter_list, bool to_pci, unsigned sglen, in descriptor_list_create() argument 200 if (sglen == 1) { in descriptor_list_create()
|
| /linux/drivers/scsi/megaraid/ |
| H A D | mega_common.h | 111 * @sglen : max sg elements supported 165 uint16_t sglen; member
|
| /linux/drivers/media/pci/saa7134/ |
| H A D | saa7134-alsa.c | 300 dma->sglen = dma_map_sg(&dev->pci->dev, dma->sglist, in saa7134_alsa_dma_map() 303 if (0 == dma->sglen) { in saa7134_alsa_dma_map() 314 if (!dma->sglen) in saa7134_alsa_dma_unmap() 318 dma->sglen = 0; in saa7134_alsa_dma_unmap() 716 dev->dmasound.sglist, dev->dmasound.sglen, 0); in snd_card_saa7134_hw_params()
|
| /linux/drivers/vdpa/mlx5/core/ |
| H A D | mr.c | 377 int sglen; in map_direct_mr() local 390 sglen = 1 << log_entity_size; in map_direct_mr() 403 for (; pa < paend; pa += sglen) { in map_direct_mr() 411 sg_set_page(sg, pg, sglen, 0); in map_direct_mr()
|
| /linux/drivers/dma/mediatek/ |
| H A D | mtk-uart-apdma.c | 339 * dmaengine_prep_slave_single will call the function. and sglen is 1. 344 unsigned int sglen, enum dma_transfer_direction dir, in mtk_uart_apdma_prep_slave_sg() argument 350 if (!is_slave_direction(dir) || sglen != 1) in mtk_uart_apdma_prep_slave_sg()
|
| /linux/drivers/media/common/saa7146/ |
| H A D | saa7146_core.c | 238 struct scatterlist *list, int sglen) in saa7146_pgtable_build_single() argument 245 if (WARN_ON(!sglen) || in saa7146_pgtable_build_single() 254 for_each_sg_dma_page(list, &dma_iter, sglen, 0) { in saa7146_pgtable_build_single()
|
| /linux/drivers/dma/lgm/ |
| H A D | lgm-dma.c | 1158 unsigned int sglen, enum dma_transfer_direction dir, in ldma_prep_slave_sg() argument 1174 return ldma_chan_desc_cfg(chan, sgl->dma_address, sglen); in ldma_prep_slave_sg() 1176 num = sg_nents_for_dma(sgl, sglen, DMA_MAX_SIZE); in ldma_prep_slave_sg() 1185 for_each_sg(sgl, sg, sglen, i) { in ldma_prep_slave_sg() 1194 switch (sglen) { in ldma_prep_slave_sg() 1209 } else if (num == (sglen - 1)) { in ldma_prep_slave_sg()
|
| /linux/drivers/crypto/inside-secure/ |
| H A D | safexcel_hash.c | 429 int sglen = sg_dma_len(sg); in safexcel_ahash_send_req() local 431 if (unlikely(sglen <= skip)) { in safexcel_ahash_send_req() 432 skip -= sglen; in safexcel_ahash_send_req() 437 if ((queued + skip) <= sglen) in safexcel_ahash_send_req() 438 sglen = queued; in safexcel_ahash_send_req() 440 sglen -= skip; in safexcel_ahash_send_req() 443 !(queued - sglen), in safexcel_ahash_send_req() 444 sg_dma_address(sg) + skip, sglen, in safexcel_ahash_send_req() 455 queued -= sglen; in safexcel_ahash_send_req()
|