/linux/drivers/crypto/nx/ |
H A D | nx-aes-xcbc.c | 63 struct nx_sg *in_sg, *out_sg; in nx_xcbc_empty() local 81 in_sg = nx_build_sg_list(nx_ctx->in_sg, (u8 *) keys, &len, in nx_xcbc_empty() 93 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg); in nx_xcbc_empty() 108 in_sg = nx_build_sg_list(nx_ctx->in_sg, (u8 *) keys[1], &len, in nx_xcbc_empty() 121 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg); in nx_xcbc_empty() 172 struct nx_sg *in_sg; in nx_xcbc_update() local 195 in_sg = nx_ctx->in_sg; in nx_xcbc_update() 230 in_sg = nx_build_sg_list(nx_ctx->in_sg, in nx_xcbc_update() 241 in_sg = nx_build_sg_list(in_sg, in nx_xcbc_update() 251 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * in nx_xcbc_update() [all …]
|
H A D | nx-sha256.c | 109 struct nx_sg *in_sg = nx_ctx->in_sg; in nx_sha256_update() local 113 in_sg = nx_build_sg_list(in_sg, in nx_sha256_update() 122 used_sgs = in_sg - nx_ctx->in_sg; in nx_sha256_update() 137 in_sg = nx_build_sg_list(in_sg, (u8 *) data, in nx_sha256_update() 140 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg); in nx_sha256_update() 186 struct nx_sg *in_sg, *out_sg; in nx_sha256_final() local 215 in_sg = nx_build_sg_list(nx_ctx->in_sg, (u8 *) sctx->buf, in nx_sha256_final() 231 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg); in nx_sha256_final()
|
H A D | nx-sha512.c | 109 struct nx_sg *in_sg = nx_ctx->in_sg; in nx_sha512_update() local 113 in_sg = nx_build_sg_list(in_sg, in nx_sha512_update() 121 used_sgs = in_sg - nx_ctx->in_sg; in nx_sha512_update() 136 in_sg = nx_build_sg_list(in_sg, (u8 *) data, in nx_sha512_update() 139 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg); in nx_sha512_update() 189 struct nx_sg *in_sg, *out_sg; in nx_sha512_final() local 224 in_sg = nx_build_sg_list(nx_ctx->in_sg, sctx->buf, &len, in nx_sha512_final() 236 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg); in nx_sha512_final()
|
H A D | nx-aes-gcm.c | 107 struct nx_sg *nx_sg = nx_ctx->in_sg; in nx_gca() 137 nx_sg = nx_walk_and_build(nx_ctx->in_sg, max_sg_len, in nx_gca() 145 nx_ctx->op_aead.inlen = (nx_ctx->in_sg - nx_sg) in nx_gca() 204 nx_sg = nx_walk_and_build(nx_ctx->in_sg, max_sg_len, in gmac() 212 nx_ctx->op.inlen = (nx_ctx->in_sg - nx_sg) in gmac() 249 struct nx_sg *in_sg, *out_sg; in gcm_empty() local 269 in_sg = nx_build_sg_list(nx_ctx->in_sg, (u8 *) iv, in gcm_empty() 282 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg); in gcm_empty()
|
H A D | nx.c | 269 struct nx_sg *nx_insg = nx_ctx->in_sg; in nx_build_sg_lists() 294 nx_ctx->op.inlen = trim_sg_list(nx_ctx->in_sg, nx_insg, delta, nbytes); in nx_build_sg_lists() 314 nx_ctx->op.in = __pa(nx_ctx->in_sg); in nx_ctx_init() 322 nx_ctx->op_aead.in = __pa(nx_ctx->in_sg); in nx_ctx_init() 677 nx_ctx->in_sg = (struct nx_sg *)((u8 *)nx_ctx->csbcpb + NX_PAGE_SIZE); in nx_crypto_ctx_init() 678 nx_ctx->out_sg = (struct nx_sg *)((u8 *)nx_ctx->in_sg + NX_PAGE_SIZE); in nx_crypto_ctx_init() 753 nx_ctx->in_sg = NULL; in nx_crypto_ctx_exit()
|
H A D | nx-aes-ccm.c | 161 struct nx_sg *nx_insg = nx_ctx->in_sg; in generate_pat() 249 nx_ctx->op.inlen = (nx_ctx->in_sg - nx_insg) * in generate_pat() 282 nx_insg = nx_walk_and_build(nx_ctx->in_sg, in generate_pat() 296 nx_ctx->op_aead.inlen = (nx_ctx->in_sg - nx_insg) * in generate_pat()
|
H A D | nx.h | 126 struct nx_sg *in_sg; /* aligned pointer into kmem to an sg list */ member
|
/linux/lib/ |
H A D | sg_split.c | 80 struct scatterlist *in_sg, *out_sg; in sg_split_phys() local 84 in_sg = split->in_sg0; in sg_split_phys() 87 *out_sg = *in_sg; in sg_split_phys() 96 in_sg = sg_next(in_sg); in sg_split_phys() 106 struct scatterlist *in_sg, *out_sg; in sg_split_mapped() local 110 in_sg = split->in_sg0; in sg_split_mapped() 113 sg_dma_address(out_sg) = sg_dma_address(in_sg); in sg_split_mapped() 114 sg_dma_len(out_sg) = sg_dma_len(in_sg); in sg_split_mapped() 119 in_sg = sg_next(in_sg); in sg_split_mapped()
|
/linux/drivers/crypto/gemini/ |
H A D | sl3516-ce-cipher.c | 30 struct scatterlist *in_sg; in sl3516_ce_need_fallback() local 87 in_sg = areq->src; in sl3516_ce_need_fallback() 89 while (in_sg && out_sg) { in sl3516_ce_need_fallback() 90 if (in_sg->length != out_sg->length) { in sl3516_ce_need_fallback() 94 in_sg = sg_next(in_sg); in sl3516_ce_need_fallback() 97 if (in_sg || out_sg) in sl3516_ce_need_fallback()
|
/linux/drivers/crypto/allwinner/sun8i-ss/ |
H A D | sun8i-ss-cipher.c | 28 struct scatterlist *in_sg = areq->src; in sun8i_ss_need_fallback() local 76 in_sg = areq->src; in sun8i_ss_need_fallback() 78 while (in_sg && out_sg) { in sun8i_ss_need_fallback() 79 if (in_sg->length != out_sg->length) in sun8i_ss_need_fallback() 81 in_sg = sg_next(in_sg); in sun8i_ss_need_fallback() 84 if (in_sg || out_sg) in sun8i_ss_need_fallback()
|
/linux/drivers/crypto/allwinner/sun4i-ss/ |
H A D | sun4i-ss-hash.c | 198 struct scatterlist *in_sg = areq->src; in sun4i_hash() local 263 while (in_sg && i == 1) { in sun4i_hash() 264 if (in_sg->length % 4) in sun4i_hash() 266 in_sg = sg_next(in_sg); in sun4i_hash()
|
H A D | sun4i-ss-cipher.c | 182 struct scatterlist *in_sg = areq->src; in sun4i_ss_cipher_poll() local 225 while (in_sg && no_chunk == 1) { in sun4i_ss_cipher_poll() 226 if ((in_sg->length | in_sg->offset) & 3u) in sun4i_ss_cipher_poll() 228 in_sg = sg_next(in_sg); in sun4i_ss_cipher_poll()
|
/linux/drivers/crypto/ |
H A D | omap-aes-gcm.c | 36 dd->in_sg = NULL; in omap_aes_gcm_finish_req() 59 dma_unmap_sg(dd->dev, dd->in_sg, dd->in_sg_len, DMA_TO_DEVICE); in omap_aes_gcm_done_task() 148 dd->in_sg = dd->in_sgl; in omap_aes_gcm_copy_buffers() 171 dd->in_sg_len = sg_nents_for_len(dd->in_sg, alen + clen); in omap_aes_gcm_copy_buffers()
|
H A D | omap-aes.h | 181 struct scatterlist *in_sg; member
|
/linux/drivers/crypto/starfive/ |
H A D | jh7110-aes.c | 584 rctx->in_sg = req->src; in starfive_aes_do_one_req() 598 ret = starfive_aes_map_sg(cryp, rctx->in_sg, rctx->out_sg); in starfive_aes_do_one_req() 652 rctx->in_sg = scatterwalk_ffwd(_src, req->src, cryp->assoclen); in starfive_aes_aead_do_one_req() 654 rctx->out_sg = rctx->in_sg; in starfive_aes_aead_do_one_req() 681 sg_zero_buffer(rctx->in_sg, sg_nents(rctx->in_sg), in starfive_aes_aead_do_one_req() 682 sg_dma_len(rctx->in_sg) - cryp->total_in, in starfive_aes_aead_do_one_req() 710 ret = starfive_aes_map_sg(cryp, rctx->in_sg, rctx->out_sg); in starfive_aes_aead_do_one_req()
|
H A D | jh7110-hash.c | 229 for_each_sg(rctx->in_sg, tsg, rctx->in_sg_len, i) { in starfive_hash_one_request() 334 rctx->in_sg = req->src; in starfive_hash_digest() 337 rctx->in_sg_len = sg_nents_for_len(rctx->in_sg, rctx->total); in starfive_hash_digest()
|
H A D | jh7110-cryp.h | 217 struct scatterlist *in_sg; member
|
/linux/arch/um/drivers/ |
H A D | virt-pci.c | 77 struct scatterlist out_sg, extra_sg, in_sg; in um_pci_send_cmd() local 80 [1] = extra ? &extra_sg : &in_sg, in um_pci_send_cmd() 81 [2] = extra ? &in_sg : NULL, in um_pci_send_cmd() 132 sg_init_one(&in_sg, out, out_size); in um_pci_send_cmd()
|
/linux/drivers/crypto/aspeed/ |
H A D | aspeed-hace-crypto.c | 185 struct scatterlist *in_sg; in aspeed_sk_start() local 190 in_sg = req->src; in aspeed_sk_start() 192 nbytes = sg_copy_to_buffer(in_sg, rctx->src_nents, in aspeed_sk_start()
|
/linux/net/vmw_vsock/ |
H A D | virtio_transport.c | 101 int ret, in_sg = 0, out_sg = 0; in virtio_transport_send_skb() local 143 ret = virtqueue_add_sgs(vq, sgs, out_sg, in_sg, skb, gfp); in virtio_transport_send_skb()
|