/linux/drivers/crypto/amlogic/ |
H A D | amlogic-gxl-cipher.c | 33 if (sg_nents(src_sg) != sg_nents(dst_sg)) in meson_cipher_need_fallback() 37 if (sg_nents(src_sg) > MAXDESC - 3 || sg_nents(dst_sg) > MAXDESC - 3) in meson_cipher_need_fallback() 178 nr_sgs = dma_map_sg(mc->dev, areq->src, sg_nents(areq->src), in meson_cipher() 187 nr_sgs = dma_map_sg(mc->dev, areq->src, sg_nents(areq->src), in meson_cipher() 194 nr_sgd = dma_map_sg(mc->dev, areq->dst, sg_nents(areq->dst), in meson_cipher() 239 dma_unmap_sg(mc->dev, areq->src, sg_nents(areq->src), DMA_BIDIRECTIONAL); in meson_cipher() 241 dma_unmap_sg(mc->dev, areq->src, sg_nents(areq->src), DMA_TO_DEVICE); in meson_cipher() 242 dma_unmap_sg(mc->dev, areq->dst, sg_nents(areq->dst), DMA_FROM_DEVICE); in meson_cipher()
|
/linux/drivers/infiniband/sw/rdmavt/ |
H A D | trace_mr.h | 144 TP_PROTO(struct ib_mr *ibmr, int sg_nents, unsigned int *sg_offset), 145 TP_ARGS(ibmr, sg_nents, sg_offset), 152 __field(int, sg_nents) 161 __entry->sg_nents = sg_nents; 171 __entry->sg_nents,
|
H A D | mr.h | 34 int sg_nents, unsigned int *sg_offset);
|
H A D | mr.c | 589 int sg_nents, unsigned int *sg_offset) in rvt_map_mr_sg() argument 596 ret = ib_sg_to_pages(ibmr, sg, sg_nents, sg_offset, rvt_set_page); in rvt_map_mr_sg() 601 trace_rvt_map_mr_sg(ibmr, sg_nents, sg_offset); in rvt_map_mr_sg()
|
/linux/drivers/crypto/gemini/ |
H A D | sl3516-ce-cipher.c | 43 if (sg_nents(areq->src) > MAXDESC / 2) { in sl3516_ce_need_fallback() 48 if (sg_nents(areq->dst) > MAXDESC) { in sl3516_ce_need_fallback() 154 nr_sgs = dma_map_sg(ce->dev, areq->src, sg_nents(areq->src), in sl3516_ce_cipher() 163 nr_sgs = dma_map_sg(ce->dev, areq->src, sg_nents(areq->src), in sl3516_ce_cipher() 170 nr_sgd = dma_map_sg(ce->dev, areq->dst, sg_nents(areq->dst), in sl3516_ce_cipher() 251 dma_unmap_sg(ce->dev, areq->src, sg_nents(areq->src), in sl3516_ce_cipher() 254 dma_unmap_sg(ce->dev, areq->src, sg_nents(areq->src), in sl3516_ce_cipher() 256 dma_unmap_sg(ce->dev, areq->dst, sg_nents(areq->dst), in sl3516_ce_cipher()
|
/linux/drivers/crypto/intel/iaa/ |
H A D | iaa_crypto_main.c | 1115 dma_unmap_sg(dev, ctx->req->dst, sg_nents(ctx->req->dst), DMA_TO_DEVICE); in iaa_desc_complete() 1116 dma_unmap_sg(dev, ctx->req->src, sg_nents(ctx->req->src), DMA_FROM_DEVICE); in iaa_desc_complete() 1121 dma_unmap_sg(dev, ctx->req->dst, sg_nents(ctx->req->dst), DMA_FROM_DEVICE); in iaa_desc_complete() 1122 dma_unmap_sg(dev, ctx->req->src, sg_nents(ctx->req->src), DMA_TO_DEVICE); in iaa_desc_complete() 1255 dma_unmap_sg(dev, req->dst, sg_nents(req->dst), DMA_FROM_DEVICE); in iaa_remap_for_verify() 1256 dma_unmap_sg(dev, req->src, sg_nents(req->src), DMA_TO_DEVICE); in iaa_remap_for_verify() 1258 nr_sgs = dma_map_sg(dev, req->src, sg_nents(req->src), DMA_FROM_DEVICE); in iaa_remap_for_verify() 1271 nr_sgs = dma_map_sg(dev, req->dst, sg_nents(req->dst), DMA_TO_DEVICE); in iaa_remap_for_verify() 1277 dma_unmap_sg(dev, req->src, sg_nents(req->src), DMA_FROM_DEVICE); in iaa_remap_for_verify() 1556 nr_sgs = dma_map_sg(dev, req->src, sg_nents(req->src), DMA_TO_DEVICE); in iaa_comp_acompress() [all …]
|
/linux/drivers/crypto/cavium/nitrox/ |
H A D | nitrox_reqmgr.c | 61 dma_unmap_sg(dev, sr->in.sg, sg_nents(sr->in.sg), in softreq_unmap_sgbufs() 69 dma_unmap_sg(dev, sr->out.sg, sg_nents(sr->out.sg), in softreq_unmap_sgbufs() 165 nents = dma_map_sg(dev, req->src, sg_nents(req->src), in dma_map_inbufs() 182 dma_unmap_sg(dev, req->src, sg_nents(req->src), DMA_BIDIRECTIONAL); in dma_map_inbufs() 193 nents = dma_map_sg(dev, req->dst, sg_nents(req->dst), in dma_map_outbufs() 207 dma_unmap_sg(dev, req->dst, sg_nents(req->dst), DMA_BIDIRECTIONAL); in dma_map_outbufs()
|
H A D | nitrox_skcipher.c | 211 int nents = sg_nents(skreq->src) + 1; in alloc_src_sglist() 229 int nents = sg_nents(skreq->dst) + 3; in alloc_dst_sglist()
|
/linux/drivers/crypto/bcm/ |
H A D | util.c | 54 unsigned int nents = sg_nents(src); in sg_copy_part_to_buf() 73 unsigned int nents = sg_nents(dest); in sg_copy_part_from_buf() 96 int sg_nents = 0; in spu_sg_count() local 106 sg_nents++; in spu_sg_count() 111 return sg_nents; in spu_sg_count()
|
/linux/drivers/crypto/intel/qat/qat_common/ |
H A D | qat_bl.c | 62 int n = sg_nents(sgl); in __qat_bl_sgl_to_bufl() 130 int n_sglout = sg_nents(sglout); in __qat_bl_sgl_to_bufl() 200 n = sg_nents(sglout); in __qat_bl_sgl_to_bufl() 217 n = sg_nents(sgl); in __qat_bl_sgl_to_bufl() 279 n = sg_nents(sgl); in qat_bl_sgl_map()
|
/linux/drivers/crypto/hisilicon/ |
H A D | sgl.c | 233 sg_n = sg_nents(sgl); in hisi_acc_sg_buf_map_to_hw_sgl() 288 dma_unmap_sg(dev, sgl, sg_nents(sgl), DMA_BIDIRECTIONAL); in hisi_acc_sg_buf_unmap()
|
/linux/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_mr.c | 313 int pvrdma_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents, in pvrdma_map_mr_sg() argument 322 ret = ib_sg_to_pages(ibmr, sg, sg_nents, sg_offset, pvrdma_set_page); in pvrdma_map_mr_sg()
|
/linux/crypto/ |
H A D | scompress.c | 138 if (sg_nents(req->src) == 1 && !PageHighMem(sg_page(req->src))) { in scomp_acomp_comp_decomp() 146 if (req->dst && sg_nents(req->dst) == 1 && !PageHighMem(sg_page(req->dst))) in scomp_acomp_comp_decomp()
|
/linux/drivers/mailbox/ |
H A D | bcm-pdc-mailbox.c | 608 sg_nents(pdcs->src_sg[pdcs->txin]), DMA_TO_DEVICE); in pdc_receive_one() 623 dma_unmap_sg(dev, rx_ctx->dst_sg, sg_nents(rx_ctx->dst_sg), in pdc_receive_one() 717 num_desc = (u32)sg_nents(sg); in pdc_tx_list_sg_add() 884 num_desc = (u32)sg_nents(sg); in pdc_rx_list_sg_add() 1205 src_nent = sg_nents(mssg->spu.src); in pdc_send_data() 1212 dst_nent = sg_nents(mssg->spu.dst); in pdc_send_data()
|
H A D | bcm-flexrm-mailbox.c | 623 rc = dma_map_sg(dev, msg->spu.src, sg_nents(msg->spu.src), in flexrm_spu_dma_map() 628 rc = dma_map_sg(dev, msg->spu.dst, sg_nents(msg->spu.dst), in flexrm_spu_dma_map() 631 dma_unmap_sg(dev, msg->spu.src, sg_nents(msg->spu.src), in flexrm_spu_dma_map() 641 dma_unmap_sg(dev, msg->spu.dst, sg_nents(msg->spu.dst), in flexrm_spu_dma_unmap() 643 dma_unmap_sg(dev, msg->spu.src, sg_nents(msg->spu.src), in flexrm_spu_dma_unmap()
|
/linux/drivers/crypto/allwinner/sun4i-ss/ |
H A D | sun4i-ss-cipher.c | 80 sg_miter_start(&mi, areq->src, sg_nents(areq->src), in sun4i_ss_opti_poll() 108 sg_miter_start(&mo, areq->dst, sg_nents(areq->dst), in sun4i_ss_opti_poll() 272 sg_miter_start(&mi, areq->src, sg_nents(areq->src), in sun4i_ss_cipher_poll() 326 sg_miter_start(&mo, areq->dst, sg_nents(areq->dst), in sun4i_ss_cipher_poll()
|
H A D | sun4i-ss-hash.c | 220 copied = sg_pcopy_to_buffer(areq->src, sg_nents(areq->src), in sun4i_hash() 272 sg_miter_start(&mi, areq->src, sg_nents(areq->src), in sun4i_hash()
|
/linux/drivers/usb/storage/ |
H A D | protocol.c | 131 nents = sg_nents(sg); in usb_stor_access_xfer_buf()
|
/linux/drivers/infiniband/hw/ocrdma/ |
H A D | ocrdma_verbs.h | 104 int ocrdma_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents,
|
/linux/include/linux/ |
H A D | devcoredump.h | 34 for_each_sg(table, iter, sg_nents(table), i) { in _devcd_free_sgtable()
|
/linux/drivers/infiniband/hw/qedr/ |
H A D | verbs.h | 85 int sg_nents, unsigned int *sg_offset);
|
/linux/drivers/crypto/hisilicon/sec/ |
H A D | sec_algs.c | 460 sg_nents(sec_req_el->sgl_out), in sec_skcipher_alg_callback() 467 sg_nents(sec_req_el->sgl_in), in sec_skcipher_alg_callback() 730 sec_req->len_in = sg_nents(skreq->src); in sec_alg_skcipher_crypto() 744 sec_req->len_out = sg_nents(skreq->dst); in sec_alg_skcipher_crypto()
|
/linux/drivers/crypto/allwinner/sun8i-ss/ |
H A D | sun8i-ss-hash.c | 360 if (sg_nents(areq->src) > MAX_SG - 1) { in sun8i_ss_hash_need_fallback() 520 nr_sgs = dma_map_sg(ss->dev, areq->src, sg_nents(areq->src), DMA_TO_DEVICE); in sun8i_ss_hash_run() 686 dma_unmap_sg(ss->dev, areq->src, sg_nents(areq->src), in sun8i_ss_hash_run()
|
/linux/drivers/crypto/ |
H A D | omap-crypto.c | 20 int n = sg_nents(*sg); in omap_crypto_copy_sg_lists()
|
/linux/lib/ |
H A D | sg_split.c | 161 ret = sg_calculate_split(in, sg_nents(in), nb_splits, skip, split_sizes, in sg_split()
|