Home
last modified time | relevance | path

Searched refs:rx_sg (Results 1 – 24 of 24) sorted by relevance

/linux/drivers/spi/
H A Dspi-dw-dma.c518 ret = dw_spi_dma_submit_rx(dws, xfer->rx_sg.sgl, in dw_spi_dma_transfer_all()
519 xfer->rx_sg.nents); in dw_spi_dma_transfer_all()
572 struct scatterlist *tx_sg = NULL, *rx_sg = NULL, tx_tmp, rx_tmp; in dw_spi_dma_transfer_one() local
590 rx_sg = !rx_sg ? &xfer->rx_sg.sgl[0] : sg_next(rx_sg); in dw_spi_dma_transfer_one()
591 sg_dma_address(&rx_tmp) = sg_dma_address(rx_sg); in dw_spi_dma_transfer_one()
592 rx_len = sg_dma_len(rx_sg); in dw_spi_dma_transfer_one()
643 nents = max(xfer->tx_sg.nents, xfer->rx_sg.nents); in dw_spi_dma_transfer()
H A Dspi-bcm2835.c500 if (bs->rx_buf && !sg_is_last(&tfr->rx_sg.sgl[0])) { in bcm2835_spi_transfer_prologue()
501 bs->rx_prologue = sg_dma_len(&tfr->rx_sg.sgl[0]) & 3; in bcm2835_spi_transfer_prologue()
531 sg_dma_address(&tfr->rx_sg.sgl[0]), in bcm2835_spi_transfer_prologue()
534 sg_dma_address(&tfr->rx_sg.sgl[0]) += bs->rx_prologue; in bcm2835_spi_transfer_prologue()
535 sg_dma_len(&tfr->rx_sg.sgl[0]) -= bs->rx_prologue; in bcm2835_spi_transfer_prologue()
582 sg_dma_address(&tfr->rx_sg.sgl[0]) -= bs->rx_prologue; in bcm2835_spi_undo_prologue()
583 sg_dma_len(&tfr->rx_sg.sgl[0]) += bs->rx_prologue; in bcm2835_spi_undo_prologue()
695 nents = tfr->rx_sg.nents; in bcm2835_spi_prepare_sg()
696 sgl = tfr->rx_sg.sgl; in bcm2835_spi_prepare_sg()
H A Dspi-amlogic-spisg.c333 if (xfer->rx_sg.nents && xfer->rx_sg.sgl) { in aml_spisg_setup_transfer()
334 ccsg_len = xfer->rx_sg.nents * sizeof(struct spisg_sg_link); in aml_spisg_setup_transfer()
341 aml_spisg_sg_xlate(&xfer->rx_sg, ccsg); in aml_spisg_setup_transfer()
356 &xfer->rx_sg, DMA_FROM_DEVICE); in aml_spisg_setup_transfer()
402 &xfer->rx_sg, DMA_FROM_DEVICE); in aml_spisg_cleanup_transfer()
H A Dspi-geni-qcom.c450 rx_desc = dmaengine_prep_slave_sg(mas->rx, xfer->rx_sg.sgl, xfer->rx_sg.nents, in setup_gsi_xfer()
842 if (!xfer->tx_sg.nents && !xfer->rx_sg.nents) in setup_se_xfer()
844 else if (xfer->tx_sg.nents > 1 || xfer->rx_sg.nents > 1) { in setup_se_xfer()
846 xfer->tx_sg.nents, xfer->rx_sg.nents); in setup_se_xfer()
866 geni_se_rx_init_dma(se, sg_dma_address(xfer->rx_sg.sgl), in setup_se_xfer()
867 sg_dma_len(xfer->rx_sg.sgl)); in setup_se_xfer()
H A Dspi-davinci.c685 t->rx_sg.sgl, t->rx_sg.nents, DMA_DEV_TO_MEM, in davinci_spi_bufs()
696 t->tx_sg.sgl = t->rx_sg.sgl; in davinci_spi_bufs()
697 t->tx_sg.nents = t->rx_sg.nents; in davinci_spi_bufs()
H A Dspi-pxa2xx-dma.c104 sgt = &xfer->rx_sg; in pxa2xx_spi_dma_prepare_one()
H A Dspi-at91-usart.c215 xfer->rx_sg.sgl, in at91_usart_spi_dma_transfer()
216 xfer->rx_sg.nents, in at91_usart_spi_dma_transfer()
H A Dspi-pic32-sqi.c276 sgl = xfer->rx_sg.sgl; in pic32_sqi_one_transfer()
277 nents = xfer->rx_sg.nents; in pic32_sqi_one_transfer()
H A Dspi-sun6i.c229 tfr->rx_sg.sgl, in sun6i_spi_prepare_dma()
230 tfr->rx_sg.nents, in sun6i_spi_prepare_dma()
H A Dspi-stm32.c1551 for_each_sg(xfer->rx_sg.sgl, spi_s, xfer->rx_sg.nents, i) in stm32_spi_prepare_rx_dma_mdma_chaining()
1562 spi_s = xfer->rx_sg.sgl; in stm32_spi_prepare_rx_dma_mdma_chaining()
1599 spi_s = xfer->rx_sg.sgl; in stm32_spi_prepare_rx_dma_mdma_chaining()
1668 rx_dma_desc = dmaengine_prep_slave_sg(spi->dma_rx, xfer->rx_sg.sgl, in stm32_spi_transfer_one_dma()
1669 xfer->rx_sg.nents, in stm32_spi_transfer_one_dma()
H A Dspi-omap2-mcspi.c509 ret = sg_split(xfer->rx_sg.sgl, xfer->rx_sg.nents, 0, nb_sizes, in omap2_mcspi_rx_dma()
H A Dspi-atmel.c793 xfer->rx_sg.sgl, in atmel_spi_next_xfer_dma_submit()
794 xfer->rx_sg.nents, in atmel_spi_next_xfer_dma_submit()
H A Dspi.c1260 ret = spi_map_buf_attrs(ctlr, rx_dev, &xfer->rx_sg, in __spi_map_msg()
1295 spi_unmap_buf_attrs(ctlr, rx_dev, &xfer->rx_sg, in __spi_unmap_msg()
1317 dma_sync_sgtable_for_device(rx_dev, &xfer->rx_sg, DMA_FROM_DEVICE); in spi_dma_sync_for_device()
1327 dma_sync_sgtable_for_cpu(rx_dev, &xfer->rx_sg, DMA_FROM_DEVICE); in spi_dma_sync_for_cpu()
H A Dspi-fsl-lpspi.c632 struct sg_table *tx = &transfer->tx_sg, *rx = &transfer->rx_sg; in fsl_lpspi_dma_transfer()
H A Dspi-pci1xxxx.c550 p->rx_sgl = xfer->rx_sg.sgl; in pci1xxxx_spi_transfer_with_dma()
H A Dspi-s3c64xx.c540 ret = s3c64xx_prepare_dma(&sdd->rx_dma, &xfer->rx_sg); in s3c64xx_enable_datapath()
H A Dspi-mt65xx.c720 mdata->rx_sgl = xfer->rx_sg.sgl; in mtk_spi_dma_transfer()
/linux/drivers/crypto/qce/
H A Ddma.c105 int qce_dma_prep_sgs(struct qce_dma_data *dma, struct scatterlist *rx_sg, in qce_dma_prep_sgs() argument
114 ret = qce_dma_prep_sg(rxchan, rx_sg, rx_nents, flags, DMA_MEM_TO_DEV, in qce_dma_prep_sgs()
/linux/drivers/infiniband/ulp/iser/
H A Diser_initiator.c229 struct ib_sge *rx_sg; in iser_alloc_rx_descriptors() local
258 rx_sg = &rx_desc->rx_sg; in iser_alloc_rx_descriptors()
259 rx_sg->addr = rx_desc->dma_addr; in iser_alloc_rx_descriptors()
260 rx_sg->length = ISER_RX_PAYLOAD_SIZE; in iser_alloc_rx_descriptors()
261 rx_sg->lkey = device->pd->local_dma_lkey; in iser_alloc_rx_descriptors()
H A Discsi_iser.h270 struct ib_sge rx_sg; member
H A Diser_verbs.c838 wr.sg_list = &rx_desc->rx_sg; in iser_post_recvm()
/linux/drivers/infiniband/ulp/isert/
H A Dib_isert.h91 struct ib_sge rx_sg; member
H A Dib_isert.c151 struct ib_sge *rx_sg; in isert_alloc_rx_descriptors() local
170 rx_sg = &rx_desc->rx_sg; in isert_alloc_rx_descriptors()
171 rx_sg->addr = rx_desc->dma_addr + isert_get_hdr_offset(rx_desc); in isert_alloc_rx_descriptors()
172 rx_sg->length = ISER_RX_PAYLOAD_SIZE; in isert_alloc_rx_descriptors()
173 rx_sg->lkey = device->pd->local_dma_lkey; in isert_alloc_rx_descriptors()
735 rx_wr->sg_list = &rx_desc->rx_sg; in isert_post_recvm()
766 rx_wr.sg_list = &rx_desc->rx_sg; in isert_post_recv()
/linux/include/linux/spi/
H A Dspi.h1117 struct sg_table rx_sg; member