| /linux/drivers/crypto/caam/ |
| H A D | caamhash.c | 544 int src_nents; member 557 if (edesc->src_nents) in ahash_unmap() 558 dma_unmap_sg(dev, req->src, edesc->src_nents, DMA_TO_DEVICE); in ahash_unmap() 832 int src_nents, mapped_nents, sec4_sg_bytes, sec4_sg_src_index; in ahash_update_ctx() local 854 src_nents = sg_nents_for_len(req->src, src_len); in ahash_update_ctx() 855 if (src_nents < 0) { in ahash_update_ctx() 857 return src_nents; in ahash_update_ctx() 860 if (src_nents) { in ahash_update_ctx() 861 mapped_nents = dma_map_sg(jrdev, req->src, src_nents, in ahash_update_ctx() 882 dma_unmap_sg(jrdev, req->src, src_nents, DMA_TO_DEVICE); in ahash_update_ctx() [all …]
|
| H A D | caamalg_qi2.c | 152 struct scatterlist *dst, int src_nents, in caam_unmap() argument 158 if (src_nents) in caam_unmap() 159 dma_unmap_sg(dev, src, src_nents, DMA_TO_DEVICE); in caam_unmap() 163 dma_unmap_sg(dev, src, src_nents, DMA_BIDIRECTIONAL); in caam_unmap() 365 int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0; in aead_edesc_alloc() local 386 src_nents = sg_nents_for_len(req->src, src_len); in aead_edesc_alloc() 387 if (unlikely(src_nents < 0)) { in aead_edesc_alloc() 391 return ERR_PTR(src_nents); in aead_edesc_alloc() 402 if (src_nents) { in aead_edesc_alloc() 403 mapped_src_nents = dma_map_sg(dev, req->src, src_nents, in aead_edesc_alloc() [all …]
|
| H A D | caamalg_qi.c | 803 int src_nents; member 825 int src_nents; member 873 struct scatterlist *dst, int src_nents, in caam_unmap() argument 879 if (src_nents) in caam_unmap() 880 dma_unmap_sg(dev, src, src_nents, DMA_TO_DEVICE); in caam_unmap() 884 dma_unmap_sg(dev, src, src_nents, DMA_BIDIRECTIONAL); in caam_unmap() 900 caam_unmap(dev, req->src, req->dst, edesc->src_nents, edesc->dst_nents, in aead_unmap() 912 caam_unmap(dev, req->src, req->dst, edesc->src_nents, edesc->dst_nents, in skcipher_unmap() 951 int src_nents, mapped_src_nents, dst_nents = 0, mapped_dst_nents = 0; in aead_edesc_alloc() local 977 src_nents = sg_nents_for_len(req->src, src_len); in aead_edesc_alloc() [all …]
|
| H A D | caamalg.c | 965 int src_nents; member 991 int src_nents; member 1004 struct scatterlist *dst, int src_nents, in caam_unmap() argument 1010 if (src_nents) in caam_unmap() 1011 dma_unmap_sg(dev, src, src_nents, DMA_TO_DEVICE); in caam_unmap() 1015 dma_unmap_sg(dev, src, src_nents, DMA_BIDIRECTIONAL); in caam_unmap() 1030 edesc->src_nents, edesc->dst_nents, 0, 0, in aead_unmap() 1041 edesc->src_nents, edesc->dst_nents, in skcipher_unmap() 1335 (int)edesc->src_nents > 1 ? 100 : req->cryptlen, req->cryptlen); in init_skcipher_job() 1339 edesc->src_nents > 1 ? 100 : req->cryptlen, 1); in init_skcipher_job() [all …]
|
| H A D | caampkc.h | 132 int src_nents; member
|
| /linux/drivers/crypto/qce/ |
| H A D | aead.c | 49 dma_unmap_sg(qce->dev, rctx->src_sg, rctx->src_nents, dir_src); in qce_aead_done() 239 rctx->src_nents = sg_nents_for_len(req->src, totallen) + 1; in qce_aead_ccm_prepare_buf_assoclen() 241 rctx->src_nents = sg_nents_for_len(req->src, totallen) + 2; in qce_aead_ccm_prepare_buf_assoclen() 244 ret = sg_alloc_table(&rctx->src_tbl, rctx->src_nents, gfp); in qce_aead_ccm_prepare_buf_assoclen() 279 rctx->src_nents = sg_nents_for_len(rctx->src_sg, totallen); in qce_aead_ccm_prepare_buf_assoclen() 289 rctx->dst_nents = rctx->src_nents + 1; in qce_aead_ccm_prepare_buf_assoclen() 291 rctx->dst_nents = rctx->src_nents; in qce_aead_ccm_prepare_buf_assoclen() 316 rctx->src_nents = sg_nents_for_len(req->src, totallen); in qce_aead_prepare_buf() 317 if (rctx->src_nents < 0) { in qce_aead_prepare_buf() 323 rctx->src_nents = rctx->dst_nents - 1; in qce_aead_prepare_buf() [all …]
|
| H A D | skcipher.c | 50 dma_unmap_sg(qce->dev, rctx->src_sg, rctx->src_nents, dir_src); in qce_skcipher_done() 75 int dst_nents, src_nents, ret; in qce_skcipher_async_req_handle() local 85 rctx->src_nents = sg_nents_for_len(req->src, req->cryptlen); in qce_skcipher_async_req_handle() 89 rctx->dst_nents = rctx->src_nents; in qce_skcipher_async_req_handle() 90 if (rctx->src_nents < 0) { in qce_skcipher_async_req_handle() 92 return rctx->src_nents; in qce_skcipher_async_req_handle() 133 src_nents = dma_map_sg(qce->dev, req->src, rctx->src_nents, dir_src); in qce_skcipher_async_req_handle() 134 if (!src_nents) { in qce_skcipher_async_req_handle() 141 src_nents = dst_nents - 1; in qce_skcipher_async_req_handle() 144 ret = qce_dma_prep_sgs(&qce->dma, rctx->src_sg, src_nents, in qce_skcipher_async_req_handle() [all …]
|
| H A D | sha.c | 53 dma_unmap_sg(qce->dev, req->src, rctx->src_nents, DMA_TO_DEVICE); in qce_ahash_done() 93 rctx->src_nents = sg_nents_for_len(req->src, req->nbytes); in qce_ahash_async_req_handle() 94 if (rctx->src_nents < 0) { in qce_ahash_async_req_handle() 96 return rctx->src_nents; in qce_ahash_async_req_handle() 99 ret = dma_map_sg(qce->dev, req->src, rctx->src_nents, DMA_TO_DEVICE); in qce_ahash_async_req_handle() 111 ret = qce_dma_prep_sgs(&qce->dma, req->src, rctx->src_nents, in qce_ahash_async_req_handle() 129 dma_unmap_sg(qce->dev, req->src, rctx->src_nents, DMA_TO_DEVICE); in qce_ahash_async_req_handle()
|
| H A D | cipher.h | 38 int src_nents; member
|
| H A D | sha.h | 50 int src_nents; member
|
| H A D | aead.h | 30 int src_nents; member
|
| /linux/drivers/crypto/inside-secure/eip93/ |
| H A D | eip93-common.c | 212 int src_nents, dst_nents; in check_valid_request() local 227 src_nents = sg_nents_for_len(src, totlen_src); in check_valid_request() 228 if (src_nents < 0) in check_valid_request() 229 return src_nents; in check_valid_request() 236 src_nents = max(src_nents, dst_nents); in check_valid_request() 237 dst_nents = src_nents; in check_valid_request() 238 if (unlikely((totlen_src || totlen_dst) && !src_nents)) in check_valid_request() 242 if (unlikely(totlen_src && !src_nents)) in check_valid_request() 250 if (dst_nents == 1 && src_nents == 1) { in check_valid_request() 281 src_nents = sg_nents_for_len(rctx->sg_src, totlen_src); in check_valid_request() [all …]
|
| /linux/drivers/crypto/virtio/ |
| H A D | virtio_crypto_skcipher_algs.c | 331 int src_nents, dst_nents; in __virtio_crypto_skcipher_do_req() local 341 src_nents = sg_nents_for_len(req->src, req->cryptlen); in __virtio_crypto_skcipher_do_req() 342 if (src_nents < 0) { in __virtio_crypto_skcipher_do_req() 344 return src_nents; in __virtio_crypto_skcipher_do_req() 350 src_nents, dst_nents); in __virtio_crypto_skcipher_do_req() 353 sg_total = src_nents + dst_nents + 3; in __virtio_crypto_skcipher_do_req() 433 for (sg = req->src; src_nents; sg = sg_next(sg), src_nents--) in __virtio_crypto_skcipher_do_req()
|
| /linux/drivers/crypto/ti/ |
| H A D | dthev2-aes.c | 273 int src_nents = sg_nents_for_len(src, len); in dthe_aes_run() local 311 src_mapped_nents = dma_map_sg(tx_dev, src, src_nents, src_dir); in dthe_aes_run() 318 dst_nents = src_nents; in dthe_aes_run() 324 dma_unmap_sg(tx_dev, src, src_nents, src_dir); in dthe_aes_run() 389 dma_unmap_sg(tx_dev, src, src_nents, src_dir); in dthe_aes_run()
|
| /linux/drivers/crypto/ccp/ |
| H A D | ccp-dmaengine.c | 356 unsigned int src_nents, in ccp_create_desc() argument 375 if (!dst_nents || !src_nents) in ccp_create_desc() 392 src_nents--; in ccp_create_desc() 393 if (!src_nents) in ccp_create_desc()
|
| /linux/drivers/crypto/starfive/ |
| H A D | jh7110-hash.c | 203 int ret, src_nents, i; in starfive_hash_one_request() local 230 src_nents = dma_map_sg(cryp->dev, tsg, 1, DMA_TO_DEVICE); in starfive_hash_one_request() 231 if (src_nents == 0) in starfive_hash_one_request()
|
| /linux/drivers/crypto/bcm/ |
| H A D | cipher.c | 224 rctx->src_nents, chunksize); in spu_skcipher_tx_sg_create() 346 rctx->src_nents = spu_sg_count(rctx->src_sg, rctx->src_skip, chunksize); in handle_skcipher_req() 444 tx_frag_num += rctx->src_nents; in handle_skcipher_req() 603 rctx->src_nents, new_data_len); in spu_ahash_tx_sg_create() 765 rctx->src_nents = spu_sg_count(rctx->src_sg, rctx->src_skip, in handle_ahash_req() 868 tx_frag_num += rctx->src_nents; in handle_ahash_req() 1212 rctx->src_nents, datalen); in spu_aead_tx_sg_create() 1338 rctx->src_nents = spu_sg_count(rctx->src_sg, rctx->src_skip, chunksize); in handle_aead_req() 1497 tx_frag_num += rctx->src_nents; in handle_aead_req() 1729 rctx->src_nents = 0; in skcipher_enqueue() [all …]
|
| H A D | cipher.h | 287 int src_nents; /* Number of src entries with data */ member
|
| /linux/drivers/crypto/ |
| H A D | sa2ul.c | 1088 int sg_nents, src_nents, dst_nents; in sa_run() local 1159 src_nents = 1; in sa_run() 1165 mapped_sg->sgt.orig_nents = src_nents; in sa_run() 1187 &split_size, &src, &src_nents, gfp_flags); in sa_run() 1189 src_nents = mapped_sg->sgt.nents; in sa_run() 1199 dst_nents = src_nents; in sa_run() 1261 src_nents, DMA_MEM_TO_DEV, in sa_run()
|
| H A D | talitos.h | 61 int src_nents; member
|