Home
last modified time | relevance | path

Searched refs:dst_nents (Results 1 – 25 of 26) sorted by relevance

12

/linux/drivers/crypto/qce/
H A Dskcipher.c51 dma_unmap_sg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst); in qce_skcipher_done()
75 int dst_nents, src_nents, ret; in qce_skcipher_async_req_handle() local
87 rctx->dst_nents = sg_nents_for_len(req->dst, req->cryptlen); in qce_skcipher_async_req_handle()
89 rctx->dst_nents = rctx->src_nents; in qce_skcipher_async_req_handle()
94 if (rctx->dst_nents < 0) { in qce_skcipher_async_req_handle()
96 return -rctx->dst_nents; in qce_skcipher_async_req_handle()
99 rctx->dst_nents += 1; in qce_skcipher_async_req_handle()
104 ret = sg_alloc_table(&rctx->dst_tbl, rctx->dst_nents, gfp); in qce_skcipher_async_req_handle()
126 dst_nents = dma_map_sg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst); in qce_skcipher_async_req_handle()
127 if (!dst_nents) { in qce_skcipher_async_req_handle()
[all …]
H A Daead.c51 dma_unmap_sg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst); in qce_aead_done()
125 rctx->dst_nents = sg_nents_for_len(req->dst, totallen); in qce_aead_prepare_dst_buf()
126 if (rctx->dst_nents < 0) { in qce_aead_prepare_dst_buf()
131 rctx->dst_nents += 2; in qce_aead_prepare_dst_buf()
133 rctx->dst_nents += 1; in qce_aead_prepare_dst_buf()
137 ret = sg_alloc_table(&rctx->dst_tbl, rctx->dst_nents, gfp); in qce_aead_prepare_dst_buf()
175 rctx->dst_nents = sg_nents_for_len(rctx->dst_sg, totallen) + 1; in qce_aead_prepare_dst_buf()
289 rctx->dst_nents = rctx->src_nents + 1; in qce_aead_ccm_prepare_buf_assoclen()
291 rctx->dst_nents = rctx->src_nents; in qce_aead_ccm_prepare_buf_assoclen()
323 rctx->src_nents = rctx->dst_nents - 1; in qce_aead_prepare_buf()
[all …]
H A Dcipher.h39 int dst_nents; member
H A Daead.h31 int dst_nents; member
/linux/drivers/crypto/caam/
H A Dcaamalg_qi.c804 int dst_nents; member
826 int dst_nents; member
874 int dst_nents, dma_addr_t iv_dma, int ivsize, in caam_unmap() argument
881 if (dst_nents) in caam_unmap()
882 dma_unmap_sg(dev, dst, dst_nents, DMA_FROM_DEVICE); in caam_unmap()
900 caam_unmap(dev, req->src, req->dst, edesc->src_nents, edesc->dst_nents, in aead_unmap()
912 caam_unmap(dev, req->src, req->dst, edesc->src_nents, edesc->dst_nents, in skcipher_unmap()
951 int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0; in aead_edesc_alloc() local
1004 dst_nents = sg_nents_for_len(req->dst, dst_len); in aead_edesc_alloc()
1005 if (unlikely(dst_nents < 0)) { in aead_edesc_alloc()
[all …]
H A Dcaamalg_qi2.h110 int dst_nents; member
130 int dst_nents; member
H A Dcaamalg.c899 int dst_nents; member
925 int dst_nents; member
938 int dst_nents, in caam_unmap() argument
945 if (dst_nents) in caam_unmap()
946 dma_unmap_sg(dev, dst, dst_nents, DMA_FROM_DEVICE); in caam_unmap()
963 edesc->src_nents, edesc->dst_nents, 0, 0, in aead_unmap()
974 edesc->src_nents, edesc->dst_nents, in skcipher_unmap()
1054 edesc->dst_nents > 1 ? 100 : req->cryptlen, 1); in skcipher_crypt_done()
1315 int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0; in aead_edesc_alloc() local
1332 dst_nents = sg_nents_for_len(req->dst, dst_len); in aead_edesc_alloc()
[all …]
H A Dcaamalg_qi2.c152 int dst_nents, dma_addr_t iv_dma, int ivsize, in caam_unmap() argument
159 if (dst_nents) in caam_unmap()
160 dma_unmap_sg(dev, dst, dst_nents, DMA_FROM_DEVICE); in caam_unmap()
364 int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0; in aead_edesc_alloc() local
393 dst_nents = sg_nents_for_len(req->dst, dst_len); in aead_edesc_alloc()
394 if (unlikely(dst_nents < 0)) { in aead_edesc_alloc()
398 return ERR_PTR(dst_nents); in aead_edesc_alloc()
413 if (dst_nents) { in aead_edesc_alloc()
414 mapped_dst_nents = dma_map_sg(dev, req->dst, dst_nents, in aead_edesc_alloc()
478 caam_unmap(dev, req->src, req->dst, src_nents, dst_nents, 0, in aead_edesc_alloc()
[all …]
H A Dcaampkc.c54 dma_unmap_sg(dev, req->dst, edesc->dst_nents, DMA_FROM_DEVICE); in rsa_io_unmap()
262 int src_nents, dst_nents; in rsa_edesc_alloc() local
292 dst_nents = sg_nents_for_len(req->dst, req->dst_len); in rsa_edesc_alloc()
300 mapped_dst_nents = dma_map_sg(dev, req->dst, dst_nents, in rsa_edesc_alloc()
340 edesc->dst_nents = dst_nents; in rsa_edesc_alloc()
368 dma_unmap_sg(dev, req->dst, dst_nents, DMA_FROM_DEVICE); in rsa_edesc_alloc()
H A Dcaampkc.h133 int dst_nents; member
/linux/drivers/crypto/intel/keembay/
H A Dkeembay-ocs-aes-core.c86 int dst_nents; member
387 dma_unmap_sg(dev, req->dst, rctx->dst_nents, rctx->in_place ? in kmb_ocs_sk_dma_cleanup()
419 sg_swap_blocks(req->dst, rctx->dst_nents, in kmb_ocs_sk_prepare_inplace()
425 rctx->dst_nents, DMA_BIDIRECTIONAL); in kmb_ocs_sk_prepare_inplace()
475 rctx->dst_nents, DMA_FROM_DEVICE); in kmb_ocs_sk_prepare_notinplace()
523 rctx->dst_nents = sg_nents_for_len(req->dst, req->cryptlen); in kmb_ocs_sk_run()
524 if (rctx->dst_nents < 0) in kmb_ocs_sk_run()
560 sg_swap_blocks(req->dst, rctx->dst_nents, in kmb_ocs_sk_run()
694 dma_unmap_sg(dev, req->dst, rctx->dst_nents, rctx->in_place ? in kmb_ocs_aead_dma_cleanup()
783 rctx->dst_nents = sg_nents_for_len(req->dst, dst_size); in kmb_ocs_aead_dma_prepare()
[all …]
/linux/drivers/crypto/marvell/cesa/
H A Dcipher.c64 dma_unmap_sg(cesa_dev->dev, req->dst, creq->dst_nents, in mv_cesa_skcipher_dma_cleanup()
131 len = mv_cesa_sg_copy_from_sram(engine, req->dst, creq->dst_nents, in mv_cesa_skcipher_std_process()
332 ret = dma_map_sg(cesa_dev->dev, req->dst, creq->dst_nents, in mv_cesa_skcipher_dma_req_init()
396 dma_unmap_sg(cesa_dev->dev, req->dst, creq->dst_nents, in mv_cesa_skcipher_dma_req_init()
438 creq->dst_nents = sg_nents_for_len(req->dst, req->cryptlen); in mv_cesa_skcipher_req_init()
439 if (creq->dst_nents < 0) { in mv_cesa_skcipher_req_init()
441 return creq->dst_nents; in mv_cesa_skcipher_req_init()
H A Dcesa.h574 int dst_nents; member
/linux/drivers/crypto/
H A Dtalitos.c968 unsigned int dst_nents = edesc->dst_nents ? : 1; in talitos_sg_unmap() local
970 if (is_sec1 && dst && dst_nents > 1) { in talitos_sg_unmap()
973 sg_pcopy_from_buffer(dst, dst_nents, edesc->buf + offset, len, in talitos_sg_unmap()
980 if (dst && (dst_nents == 1 || !is_sec1)) in talitos_sg_unmap()
981 dma_unmap_sg(dev, dst, dst_nents, DMA_FROM_DEVICE); in talitos_sg_unmap()
1012 unsigned int dst_nents = edesc->dst_nents ? : 1; in ipsec_esp_unmap() local
1014 sg_pcopy_to_buffer(areq->dst, dst_nents, ctx->iv, ivsize, in ipsec_esp_unmap()
1265 sg_count = edesc->dst_nents ? : 1; in ipsec_esp()
1332 int src_nents, dst_nents, alloc_len, dma_len, src_len, dst_len; in talitos_edesc_alloc() local
1353 dst_nents = dst ? src_nents : 0; in talitos_edesc_alloc()
[all …]
H A Dsa2ul.c1088 int sg_nents, src_nents, dst_nents; in sa_run() local
1199 dst_nents = src_nents; in sa_run()
1202 dst_nents = sg_nents_for_len(req->dst, req->size); in sa_run()
1205 if (dst_nents == 1 && split_size <= req->dst->length) { in sa_run()
1207 dst_nents = 1; in sa_run()
1213 mapped_sg->sgt.orig_nents = dst_nents; in sa_run()
1223 mapped_sg->sgt.orig_nents = dst_nents; in sa_run()
1233 0, 1, &split_size, &dst, &dst_nents, in sa_run()
1236 dst_nents = mapped_sg->sgt.nents; in sa_run()
1244 rxd->tx_in = dmaengine_prep_slave_sg(dma_rx, dst, dst_nents, in sa_run()
H A Dtalitos.h62 int dst_nents; member
H A Dmxs-dcp.c319 int dst_nents = sg_nents(dst); in mxs_dcp_aes_block_crypt() local
381 sg_pcopy_from_buffer(dst, dst_nents, out_buf, in mxs_dcp_aes_block_crypt()
/linux/drivers/crypto/aspeed/
H A Daspeed-hace-crypto.c144 dma_unmap_sg(dev, req->dst, rctx->dst_nents, DMA_FROM_DEVICE); in aspeed_sk_transfer_sg()
164 nbytes = sg_copy_from_buffer(out_sg, rctx->dst_nents, in aspeed_sk_transfer()
174 "nb_out_sg", rctx->dst_nents, in aspeed_sk_transfer()
259 rctx->dst_nents, DMA_FROM_DEVICE); in aspeed_sk_start_sg()
348 dma_unmap_sg(hace_dev->dev, req->dst, rctx->dst_nents, in aspeed_sk_start_sg()
380 rctx->dst_nents = sg_nents(req->dst); in aspeed_hace_skcipher_trigger()
H A Daspeed-hace.h239 int dst_nents; member
/linux/drivers/crypto/virtio/
H A Dvirtio_crypto_skcipher_algs.c333 int src_nents, dst_nents; in __virtio_crypto_skcipher_do_req() local
349 dst_nents = sg_nents(req->dst); in __virtio_crypto_skcipher_do_req()
352 src_nents, dst_nents); in __virtio_crypto_skcipher_do_req()
355 sg_total = src_nents + dst_nents + 3; in __virtio_crypto_skcipher_do_req()
/linux/drivers/crypto/ccp/
H A Dccp-dmaengine.c354 unsigned int dst_nents, in ccp_create_desc() argument
375 if (!dst_nents || !src_nents) in ccp_create_desc()
406 dst_nents--; in ccp_create_desc()
407 if (!dst_nents) in ccp_create_desc()
/linux/crypto/
H A Dadiantum.c285 const unsigned int dst_nents = sg_nents(dst); in adiantum_finish() local
301 if (dst_nents == 1 && dst->offset + req->cryptlen <= PAGE_SIZE) { in adiantum_finish()
318 err = adiantum_hash_message(req, dst, dst_nents, &digest); in adiantum_finish()
/linux/drivers/crypto/starfive/
H A Djh7110-aes.c506 unsigned int len, src_nents, dst_nents; in starfive_aes_map_sg() local
517 dst_nents = src_nents; in starfive_aes_map_sg()
534 dst_nents = dma_map_sg(cryp->dev, dtsg, 1, DMA_FROM_DEVICE); in starfive_aes_map_sg()
535 if (dst_nents == 0) in starfive_aes_map_sg()
/linux/drivers/crypto/bcm/
H A Dcipher.h295 int dst_nents; /* Number of dst entries with data */ member
H A Dcipher.c161 rctx->dst_nents, chunksize); in spu_skcipher_rx_sg_create()
348 rctx->dst_nents = spu_sg_count(rctx->dst_sg, rctx->dst_skip, chunksize); in handle_skcipher_req()
433 rx_frag_num += rctx->dst_nents; in handle_skcipher_req()
1105 rctx->dst_nents, resp_len); in spu_aead_rx_sg_create()
1344 rctx->dst_nents = spu_sg_count(rctx->dst_sg, rctx->dst_skip, chunksize); in handle_aead_req()
1472 rx_frag_num += rctx->dst_nents; in handle_aead_req()
1491 rx_frag_num -= rctx->dst_nents; in handle_aead_req()
1737 rctx->dst_nents = 0; in skcipher_enqueue()
1925 rctx->dst_nents = 0; in ahash_enqueue()
2643 rctx->dst_nents = 0; in aead_enqueue()

12