Home
last modified time | relevance | path

Searched full:sglen (Results 1 – 25 of 36) sorted by relevance

12

/linux/lib/
H A Dsg_split.c25 unsigned int sglen; in sg_calculate_split() local
36 sglen = mapped ? sg_dma_len(sg) : sg->length; in sg_calculate_split()
37 if (skip > sglen) { in sg_calculate_split()
38 skip -= sglen; in sg_calculate_split()
42 len = min_t(size_t, size, sglen - skip); in sg_calculate_split()
51 while (!size && (skip + len < sglen) && (--nb_splits > 0)) { in sg_calculate_split()
55 len = min_t(size_t, size, sglen - skip); in sg_calculate_split()
/linux/drivers/usb/misc/
H A Dusbtest.c34 __u32 sglen; member
51 __u32 sglen; member
1157 for (i = 1; i < ctx->param->sglen; i++) { in ctrl_complete()
1160 % ctx->param->sglen]; in ctrl_complete()
1208 if (param->sglen == 0 || param->iterations > UINT_MAX / param->sglen) in test_ctrl_queue()
1214 context.count = param->sglen * param->iterations; in test_ctrl_queue()
1221 * as with bulk/intr sglists, sglen is the queue depth; it also in test_ctrl_queue()
1222 * controls which subtests run (more tests than sglen) or rerun. in test_ctrl_queue()
1224 urb = kzalloc_objs(struct urb *, param->sglen); in test_ctrl_queue()
1227 for (i = 0; i < param->sglen; i++) { in test_ctrl_queue()
[all …]
/linux/drivers/misc/bcm-vk/
H A Dbcm_vk_sg.c85 dma->sglen = (dma->nr_pages * sizeof(*sgdata)) + in bcm_vk_dma_alloc()
90 dma->sglen, in bcm_vk_dma_alloc()
160 "sgl 0x%llx handle 0x%llx, sglen: 0x%x sgsize: 0x%x\n", in bcm_vk_dma_alloc()
163 dma->sglen, in bcm_vk_dma_alloc()
232 dev_dbg(dev, "free sglist=%p sglen=0x%x\n", dma->sglist, dma->sglen); in bcm_vk_dma_free()
245 dma_free_coherent(dev, dma->sglen, dma->sglist, dma->handle); in bcm_vk_dma_free()
H A Dbcm_vk_sg.h35 int sglen; /* Length (bytes) of sglist */ member
/linux/drivers/dma/
H A Dsa11x0-dma.c80 unsigned sglen; member
81 struct sa11x0_dma_sg sg[] __counted_by(sglen);
176 if (p->sg_load == txd->sglen) { in sa11x0_dma_start_sg()
227 if (++p->sg_done == txd->sglen) { in sa11x0_dma_complete()
464 for (i = 0; i < txd->sglen; i++) { in sa11x0_dma_tx_status()
480 for (; i < txd->sglen; i++) { in sa11x0_dma_tx_status()
523 struct dma_chan *chan, struct scatterlist *sg, unsigned int sglen, in sa11x0_dma_prep_slave_sg() argument
540 if (sglen == 0) in sa11x0_dma_prep_slave_sg()
543 for_each_sg(sg, sgent, sglen, i) { in sa11x0_dma_prep_slave_sg()
553 j = sg_nents_for_dma(sg, sglen, DMA_MAX_SIZE & ~DMA_ALIGN); in sa11x0_dma_prep_slave_sg()
[all …]
H A Dsprd-dma.c744 unsigned int sglen, int sg_index, in sprd_dma_fill_desc() argument
846 temp = sglen ? (sg_index + 1) % sglen : 0; in sprd_dma_fill_desc()
875 unsigned int sglen, int sg_index, in sprd_dma_fill_linklist_desc() argument
890 return sprd_dma_fill_desc(chan, hw, sglen, sg_index, src, dst, len, in sprd_dma_fill_linklist_desc()
951 unsigned int sglen, enum dma_transfer_direction dir, in sprd_dma_prep_slave_sg() argument
995 for_each_sg(sgl, sg, sglen, i) { in sprd_dma_prep_slave_sg()
1016 if (sglen < 2) in sprd_dma_prep_slave_sg()
1019 ret = sprd_dma_fill_linklist_desc(chan, sglen, i, src, dst, len, in sprd_dma_prep_slave_sg()
H A Dmoxart-dma.c123 unsigned int sglen; member
127 struct moxart_sg sg[] __counted_by(sglen);
307 d->sglen = sg_len; in moxart_prep_slave_sg()
447 for (size = i = completed_sgs; i < d->sglen; i++) in moxart_dma_desc_size()
535 if (++ch->sgidx < ch->desc->sglen) { in moxart_dma_interrupt()
H A Dat_hdmac.c225 * @sglen: number of sg entries.
238 unsigned int sglen; member
248 struct atdma_sg sg[] __counted_by(sglen);
598 for (i = 0; i < desc->sglen; i++) { in atdma_desc_free()
722 for (i = 1; i < desc->sglen; i++) { in atc_get_llis_residue()
796 for (i = 0; i < desc->sglen; i++) in atc_handle_error()
935 desc->sglen = 1; in atc_prep_dma_interleaved()
998 desc->sglen = sg_len; in atc_prep_dma_memcpy()
1137 desc->sglen = 1; in atc_prep_dma_memset()
1197 desc->sglen = sg_len; in atc_prep_dma_memset_sg()
[all …]
H A Dk3dma.c531 struct dma_chan *chan, struct scatterlist *sgl, unsigned int sglen, in k3_dma_prep_slave_sg() argument
546 num = sg_nents_for_dma(sgl, sglen, DMA_MAX_SIZE); in k3_dma_prep_slave_sg()
553 for_each_sg(sgl, sg, sglen, i) { in k3_dma_prep_slave_sg()
/linux/drivers/dma/ti/
H A Domap-dma.c126 unsigned sglen; member
127 struct omap_sg sg[] __counted_by(sglen);
256 for (i = 0; i < d->sglen; i++) { in omap_dma_desc_free()
621 } else if (d->using_ll || c->sgidx == d->sglen) { in omap_dma_callback()
796 for (size = i = 0; i < d->sglen; i++) in omap_dma_desc_size()
807 for (size = i = 0; i < d->sglen; i++) { in omap_dma_desc_size_pos()
960 struct dma_chan *chan, struct scatterlist *sgl, unsigned sglen, in omap_dma_prep_slave_sg() argument
1005 d = kzalloc_flex(*d, sg, sglen, GFP_ATOMIC); in omap_dma_prep_slave_sg()
1008 d->sglen = sglen; in omap_dma_prep_slave_sg()
1099 if (sglen >= 2) in omap_dma_prep_slave_sg()
[all …]
H A Dk3-udma.c226 unsigned int sglen; member
877 for (i = 0; i < uc->desc->sglen; i++) in udma_start_desc()
1049 d->desc_idx = (d->desc_idx + 1) % d->sglen; in udma_cyclic_packet_elapsed()
1228 d->tr_idx = (d->tr_idx + 1) % d->sglen; in udma_udma_irq_handler()
2787 d->sglen = tr_count; in udma_alloc_tr_desc()
2884 unsigned int sglen, enum dma_transfer_direction dir, in udma_prep_slave_sg_tr() argument
2898 for_each_sg(sgl, sgent, sglen, i) { in udma_prep_slave_sg_tr()
2911 d->sglen = sglen; in udma_prep_slave_sg_tr()
2919 for_each_sg(sgl, sgent, sglen, i) { in udma_prep_slave_sg_tr()
2968 unsigned int sglen, in udma_prep_slave_sg_triggered_tr() argument
[all …]
/linux/drivers/crypto/cavium/nitrox/
H A Dnitrox_req.h595 unsigned int sglen; in create_multi_sg() local
597 for (; buflen && from_sg; buflen -= sglen) { in create_multi_sg()
598 sglen = from_sg->length; in create_multi_sg()
599 if (sglen > buflen) in create_multi_sg()
600 sglen = buflen; in create_multi_sg()
602 sg_set_buf(sg, sg_virt(from_sg), sglen); in create_multi_sg()
/linux/tools/usb/
H A Dtestusb.c48 unsigned sglen; member
396 param.sglen = 32; in main()
418 if (parse_num(&param.sglen, optarg)) in main()
456 "\t-g sglen default 32\n" in main()
/linux/drivers/media/pci/cx23885/
H A Dcx23885-alsa.c116 buf->sglen = dma_map_sg(&dev->pci->dev, buf->sglist, in cx23885_alsa_dma_map()
119 if (0 == buf->sglen) { in cx23885_alsa_dma_map()
130 if (!buf->sglen) in cx23885_alsa_dma_unmap()
134 buf->sglen = 0; in cx23885_alsa_dma_unmap()
/linux/drivers/media/pci/cx25821/
H A Dcx25821-alsa.c56 int sglen; member
179 buf->sglen = dma_map_sg(&dev->pci->dev, buf->sglist, in cx25821_alsa_dma_map()
182 if (0 == buf->sglen) { in cx25821_alsa_dma_map()
193 if (!buf->sglen) in cx25821_alsa_dma_unmap()
197 buf->sglen = 0; in cx25821_alsa_dma_unmap()
/linux/drivers/media/pci/cx88/
H A Dcx88-alsa.c49 int sglen; member
318 buf->sglen = dma_map_sg(&dev->pci->dev, buf->sglist, in cx88_alsa_dma_map()
321 if (buf->sglen == 0) { in cx88_alsa_dma_map()
332 if (!buf->sglen) in cx88_alsa_dma_unmap()
337 buf->sglen = 0; in cx88_alsa_dma_unmap()
/linux/drivers/media/pci/cobalt/
H A Dcobalt-omnitek.h35 struct scatterlist *scatter_list, bool to_pci, unsigned sglen,
H A Dcobalt-omnitek.c150 struct scatterlist *scatter_list, bool to_pci, unsigned sglen, in descriptor_list_create() argument
200 if (sglen == 1) { in descriptor_list_create()
/linux/drivers/scsi/megaraid/
H A Dmega_common.h111 * @sglen : max sg elements supported
165 uint16_t sglen; member
/linux/drivers/media/pci/saa7134/
H A Dsaa7134-alsa.c300 dma->sglen = dma_map_sg(&dev->pci->dev, dma->sglist, in saa7134_alsa_dma_map()
303 if (0 == dma->sglen) { in saa7134_alsa_dma_map()
314 if (!dma->sglen) in saa7134_alsa_dma_unmap()
318 dma->sglen = 0; in saa7134_alsa_dma_unmap()
716 dev->dmasound.sglist, dev->dmasound.sglen, 0); in snd_card_saa7134_hw_params()
/linux/drivers/vdpa/mlx5/core/
H A Dmr.c377 int sglen; in map_direct_mr() local
390 sglen = 1 << log_entity_size; in map_direct_mr()
403 for (; pa < paend; pa += sglen) { in map_direct_mr()
411 sg_set_page(sg, pg, sglen, 0); in map_direct_mr()
/linux/drivers/dma/mediatek/
H A Dmtk-uart-apdma.c339 * dmaengine_prep_slave_single will call the function. and sglen is 1.
344 unsigned int sglen, enum dma_transfer_direction dir, in mtk_uart_apdma_prep_slave_sg() argument
350 if (!is_slave_direction(dir) || sglen != 1) in mtk_uart_apdma_prep_slave_sg()
/linux/drivers/media/common/saa7146/
H A Dsaa7146_core.c238 struct scatterlist *list, int sglen) in saa7146_pgtable_build_single() argument
245 if (WARN_ON(!sglen) || in saa7146_pgtable_build_single()
254 for_each_sg_dma_page(list, &dma_iter, sglen, 0) { in saa7146_pgtable_build_single()
/linux/drivers/dma/lgm/
H A Dlgm-dma.c1158 unsigned int sglen, enum dma_transfer_direction dir, in ldma_prep_slave_sg() argument
1174 return ldma_chan_desc_cfg(chan, sgl->dma_address, sglen); in ldma_prep_slave_sg()
1176 num = sg_nents_for_dma(sgl, sglen, DMA_MAX_SIZE); in ldma_prep_slave_sg()
1185 for_each_sg(sgl, sg, sglen, i) { in ldma_prep_slave_sg()
1194 switch (sglen) { in ldma_prep_slave_sg()
1209 } else if (num == (sglen - 1)) { in ldma_prep_slave_sg()
/linux/drivers/crypto/inside-secure/
H A Dsafexcel_hash.c429 int sglen = sg_dma_len(sg); in safexcel_ahash_send_req() local
431 if (unlikely(sglen <= skip)) { in safexcel_ahash_send_req()
432 skip -= sglen; in safexcel_ahash_send_req()
437 if ((queued + skip) <= sglen) in safexcel_ahash_send_req()
438 sglen = queued; in safexcel_ahash_send_req()
440 sglen -= skip; in safexcel_ahash_send_req()
443 !(queued - sglen), in safexcel_ahash_send_req()
444 sg_dma_address(sg) + skip, sglen, in safexcel_ahash_send_req()
455 queued -= sglen; in safexcel_ahash_send_req()

12