| /linux/crypto/ |
| H A D | jitterentropy-kcapi.c | 199 struct shash_desc *sdesc; member 208 if (rng->sdesc) { in jent_kcapi_cleanup() 209 shash_desc_zero(rng->sdesc); in jent_kcapi_cleanup() 210 kfree(rng->sdesc); in jent_kcapi_cleanup() 212 rng->sdesc = NULL; in jent_kcapi_cleanup() 228 struct shash_desc *sdesc; in jent_kcapi_init() local 242 sdesc = kmalloc(size, GFP_KERNEL); in jent_kcapi_init() 243 if (!sdesc) { in jent_kcapi_init() 248 sdesc->tfm = hash; in jent_kcapi_init() 249 crypto_shash_init(sdesc); in jent_kcapi_init() [all …]
|
| H A D | drbg.c | 1434 struct sdesc { struct 1440 struct sdesc *sdesc; in drbg_init_hash_kernel() local 1450 sdesc = kzalloc(sizeof(struct shash_desc) + crypto_shash_descsize(tfm), in drbg_init_hash_kernel() 1452 if (!sdesc) { in drbg_init_hash_kernel() 1457 sdesc->shash.tfm = tfm; in drbg_init_hash_kernel() 1458 drbg->priv_data = sdesc; in drbg_init_hash_kernel() 1465 struct sdesc *sdesc = drbg->priv_data; in drbg_fini_hash_kernel() local 1466 if (sdesc) { in drbg_fini_hash_kernel() 1467 crypto_free_shash(sdesc->shash.tfm); in drbg_fini_hash_kernel() 1468 kfree_sensitive(sdesc); in drbg_fini_hash_kernel() [all …]
|
| /linux/Documentation/crypto/ |
| H A D | api-samples.rst | 99 struct sdesc { 104 static struct sdesc *init_sdesc(struct crypto_shash *alg) 106 struct sdesc *sdesc; 110 sdesc = kmalloc(size, GFP_KERNEL); 111 if (!sdesc) 113 sdesc->shash.tfm = alg; 114 return sdesc; 121 struct sdesc *sdesc; 124 sdesc = init_sdesc(alg); 125 if (IS_ERR(sdesc)) { [all …]
|
| /linux/drivers/net/ethernet/chelsio/cxgb4vf/ |
| H A D | sge.c | 195 static inline dma_addr_t get_buf_addr(const struct rx_sw_desc *sdesc) in get_buf_addr() argument 197 return sdesc->dma_addr & ~(dma_addr_t)(RX_LARGE_BUF | RX_UNMAPPED_BUF); in get_buf_addr() 207 static inline bool is_buf_mapped(const struct rx_sw_desc *sdesc) in is_buf_mapped() argument 209 return !(sdesc->dma_addr & RX_UNMAPPED_BUF); in is_buf_mapped() 380 struct tx_sw_desc *sdesc; in free_tx_desc() local 386 sdesc = &tq->sdesc[cidx]; in free_tx_desc() 392 if (sdesc->skb) { in free_tx_desc() 394 unmap_sgl(dev, sdesc->skb, sdesc->sgl, tq); in free_tx_desc() 395 dev_consume_skb_any(sdesc->skb); in free_tx_desc() 396 sdesc->skb = NULL; in free_tx_desc() [all …]
|
| H A D | adapter.h | 143 struct rx_sw_desc *sdesc; /* address of SW RX descriptor ring */ member 249 struct tx_sw_desc *sdesc; /* address of SW TX descriptor ring */ member
|
| /linux/drivers/dma/sh/ |
| H A D | shdma-base.c | 56 struct shdma_desc *sdesc; in shdma_chan_xfer_ld_queue() local 63 list_for_each_entry(sdesc, &schan->ld_queue, node) in shdma_chan_xfer_ld_queue() 64 if (sdesc->mark == DESC_SUBMITTED) { in shdma_chan_xfer_ld_queue() 65 ops->start_xfer(schan, sdesc); in shdma_chan_xfer_ld_queue() 172 struct shdma_desc *sdesc; in shdma_get_desc() local 174 list_for_each_entry(sdesc, &schan->ld_free, node) in shdma_get_desc() 175 if (sdesc->mark != DESC_PREPARED) { in shdma_get_desc() 176 BUG_ON(sdesc->mark != DESC_IDLE); in shdma_get_desc() 177 list_del(&sdesc->node); in shdma_get_desc() 178 return sdesc; in shdma_get_desc() [all …]
|
| H A D | shdmac.c | 282 struct shdma_desc *sdesc) in sh_dmae_start_xfer() argument 286 struct sh_dmae_desc *sh_desc = container_of(sdesc, in sh_dmae_start_xfer() 289 sdesc->async_tx.cookie, sh_chan->shdma_chan.id, in sh_dmae_start_xfer() 386 struct shdma_desc *sdesc, in sh_dmae_desc_setup() argument 389 struct sh_dmae_desc *sh_desc = container_of(sdesc, in sh_dmae_desc_setup() 424 struct shdma_desc *sdesc) in sh_dmae_get_partial() argument 428 struct sh_dmae_desc *sh_desc = container_of(sdesc, in sh_dmae_get_partial() 462 struct shdma_desc *sdesc) in sh_dmae_desc_completed() argument 466 struct sh_dmae_desc *sh_desc = container_of(sdesc, in sh_dmae_desc_completed() 471 return (sdesc->direction == DMA_DEV_TO_MEM && in sh_dmae_desc_completed() [all …]
|
| /linux/drivers/dma/ |
| H A D | sprd-dma.c | 512 struct sprd_dma_desc *sdesc) in sprd_dma_set_chn_config() argument 514 struct sprd_dma_chn_hw *cfg = &sdesc->chn_hw; in sprd_dma_set_chn_config() 592 struct sprd_dma_desc *sdesc; in dma_irq_handle() local 605 sdesc = schan->cur_desc; in dma_irq_handle() 606 if (!sdesc) { in dma_irq_handle() 618 vchan_cyclic_callback(&sdesc->vd); in dma_irq_handle() 623 vchan_cookie_complete(&sdesc->vd); in dma_irq_handle() 676 struct sprd_dma_desc *sdesc = to_sprd_dma_desc(vd); in sprd_dma_tx_status() local 677 struct sprd_dma_chn_hw *hw = &sdesc->chn_hw; in sprd_dma_tx_status() 688 struct sprd_dma_desc *sdesc = schan->cur_desc; in sprd_dma_tx_status() local [all …]
|
| H A D | altera-msgdma.c | 535 struct msgdma_sw_desc *sdesc, *next; in msgdma_copy_desc_to_fifo() local 539 list_for_each_entry_safe(sdesc, next, &desc->tx_list, node) in msgdma_copy_desc_to_fifo() 540 msgdma_copy_one(mdev, sdesc); in msgdma_copy_desc_to_fifo()
|
| /linux/drivers/crypto/bcm/ |
| H A D | util.c | 188 struct sdesc { struct 218 struct sdesc *sdesc; in do_shash() local 228 sdesc = kmalloc(size, GFP_KERNEL); in do_shash() 229 if (!sdesc) { in do_shash() 233 sdesc->shash.tfm = hash; in do_shash() 243 rc = crypto_shash_init(&sdesc->shash); in do_shash() 248 rc = crypto_shash_update(&sdesc->shash, data1, data1_len); in do_shash() 254 rc = crypto_shash_update(&sdesc->shash, data2, data2_len); in do_shash() 260 rc = crypto_shash_final(&sdesc->shash, result); in do_shash() 266 kfree(sdesc); in do_shash()
|
| /linux/fs/smb/client/ |
| H A D | misc.c | 797 cifs_alloc_hash(const char *name, struct shash_desc **sdesc) in cifs_alloc_hash() argument 802 if (*sdesc) in cifs_alloc_hash() 809 *sdesc = NULL; in cifs_alloc_hash() 813 *sdesc = kmalloc(sizeof(struct shash_desc) + crypto_shash_descsize(alg), GFP_KERNEL); in cifs_alloc_hash() 814 if (*sdesc == NULL) { in cifs_alloc_hash() 820 (*sdesc)->tfm = alg; in cifs_alloc_hash() 831 cifs_free_hash(struct shash_desc **sdesc) in cifs_free_hash() argument 833 if (unlikely(!sdesc) || !*sdesc) in cifs_free_hash() 836 if ((*sdesc)->tfm) { in cifs_free_hash() 837 crypto_free_shash((*sdesc)->tfm); in cifs_free_hash() [all …]
|
| H A D | cifsproto.h | 353 int cifs_alloc_hash(const char *name, struct shash_desc **sdesc); 354 void cifs_free_hash(struct shash_desc **sdesc);
|
| /linux/drivers/dma/xilinx/ |
| H A D | zynqmp_dma.c | 316 struct zynqmp_dma_desc_ll *sdesc, in zynqmp_dma_config_sg_ll_desc() argument 320 struct zynqmp_dma_desc_ll *ddesc = sdesc + 1; in zynqmp_dma_config_sg_ll_desc() 322 sdesc->size = ddesc->size = len; in zynqmp_dma_config_sg_ll_desc() 323 sdesc->addr = src; in zynqmp_dma_config_sg_ll_desc() 326 sdesc->ctrl = ddesc->ctrl = ZYNQMP_DMA_DESC_CTRL_SIZE_256; in zynqmp_dma_config_sg_ll_desc() 328 sdesc->ctrl |= ZYNQMP_DMA_DESC_CTRL_COHRNT; in zynqmp_dma_config_sg_ll_desc() 334 ((uintptr_t)sdesc - (uintptr_t)chan->desc_pool_v); in zynqmp_dma_config_sg_ll_desc() 443 struct zynqmp_dma_desc_sw *sdesc) in zynqmp_dma_free_descriptor() argument 448 list_move_tail(&sdesc->node, &chan->free_list); in zynqmp_dma_free_descriptor() 449 list_for_each_entry_safe(child, next, &sdesc->tx_list, node) { in zynqmp_dma_free_descriptor()
|
| /linux/drivers/crypto/marvell/octeontx/ |
| H A D | otx_cptvf_algs.c | 561 ctx->sdesc = alloc_sdesc(ctx->hashalg); in cpt_aead_init() 562 if (!ctx->sdesc) { in cpt_aead_init() 624 kfree(ctx->sdesc); in otx_cpt_aead_exit() 691 struct otx_cpt_sdesc *sdesc; in alloc_sdesc() local 695 sdesc = kmalloc(size, GFP_KERNEL); in alloc_sdesc() 696 if (!sdesc) in alloc_sdesc() 699 sdesc->shash.tfm = alg; in alloc_sdesc() 701 return sdesc; in alloc_sdesc() 761 ret = crypto_shash_digest(&ctx->sdesc->shash, keys->authkey, in aead_hmac_init() 810 crypto_shash_init(&ctx->sdesc->shash); in aead_hmac_init() [all …]
|
| H A D | otx_cptvf_algs.h | 172 struct otx_cpt_sdesc *sdesc; member
|
| /linux/drivers/net/wireless/ath/ath10k/ |
| H A D | ce.c | 433 struct ce_desc *desc, sdesc; in _ath10k_ce_send_nolock() local 461 sdesc.addr = __cpu_to_le32(buffer); in _ath10k_ce_send_nolock() 462 sdesc.nbytes = __cpu_to_le16(nbytes); in _ath10k_ce_send_nolock() 463 sdesc.flags = __cpu_to_le16(desc_flags); in _ath10k_ce_send_nolock() 465 *desc = sdesc; in _ath10k_ce_send_nolock() 490 struct ce_desc_64 *desc, sdesc; in _ath10k_ce_send_nolock_64() local 528 addr = (__le32 *)&sdesc.addr; in _ath10k_ce_send_nolock_64() 538 sdesc.nbytes = __cpu_to_le16(nbytes); in _ath10k_ce_send_nolock_64() 539 sdesc.flags = __cpu_to_le16(desc_flags); in _ath10k_ce_send_nolock_64() 541 *desc = sdesc; in _ath10k_ce_send_nolock_64() [all …]
|
| /linux/drivers/net/ethernet/chelsio/cxgb3/ |
| H A D | sge.c | 237 struct tx_sw_desc *d = &q->sdesc[cidx]; in unmap_skb() 266 d = cidx + 1 == q->size ? q->sdesc : d + 1; in unmap_skb() 292 d = &q->sdesc[cidx]; in free_tx_desc() 305 d = q->sdesc; in free_tx_desc() 381 struct rx_sw_desc *d = &q->sdesc[cidx]; in free_rx_bufs() 503 struct rx_sw_desc *sd = &q->sdesc[q->pidx]; in refill_fl() 546 sd = q->sdesc; in refill_fl() 580 q->sdesc[q->pidx] = q->sdesc[idx]; in recycle_rx_buf() 683 kfree(q->fl[i].sdesc); in t3_free_qset() 695 if (q->txq[i].sdesc) { in t3_free_qset() [all …]
|
| H A D | adapter.h | 115 struct rx_sw_desc *sdesc; /* address of SW Rx descriptor ring */ member 176 struct tx_sw_desc *sdesc; /* address of SW Tx descriptor ring */ member
|
| /linux/drivers/net/ethernet/chelsio/cxgb4/ |
| H A D | sge.c | 319 d = &q->sdesc[cidx]; in free_tx_desc() 332 d = q->sdesc; in free_tx_desc() 441 struct rx_sw_desc *d = &q->sdesc[q->cidx]; in free_rx_bufs() 468 struct rx_sw_desc *d = &q->sdesc[q->cidx]; in unmap_rx_buf() 543 struct rx_sw_desc *sd = &q->sdesc[q->pidx]; in refill_fl() 584 sd = q->sdesc; in refill_fl() 613 sd = q->sdesc; in refill_fl() 1568 sgl_sdesc = &q->q.sdesc[last_desc]; in cxgb4_eth_xmit() 1876 sgl_sdesc = &txq->q.sdesc[last_desc]; in cxgb4_vf_eth_xmit() 3020 q->q.sdesc[last_desc].skb = skb; in service_ofldq() [all …]
|
| H A D | cxgb4_uld.c | 413 kfree(txq->q.sdesc); in free_sge_txq_uld()
|
| /linux/drivers/crypto/marvell/octeontx2/ |
| H A D | otx2_cptvf_algs.h | 165 struct otx2_cpt_sdesc *sdesc; member
|
| /linux/drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/ |
| H A D | chcr_ktls.c | 1137 sgl_sdesc = &q->q.sdesc[last_desc]; in chcr_ktls_xmit_wr_complete() 1322 sgl_sdesc = &q->q.sdesc[last_desc]; in chcr_ktls_xmit_wr_short() 1500 sgl_sdesc = &q->q.sdesc[last_desc]; in chcr_ktls_tx_plaintxt() 1600 sgl_sdesc = &q->q.sdesc[last_desc]; in chcr_ktls_tunnel_pkt()
|
| /linux/drivers/net/ethernet/chelsio/inline_crypto/ch_ipsec/ |
| H A D | chcr_ipsec.c | 751 sgl_sdesc = &q->q.sdesc[last_desc]; in ch_ipsec_xmit()
|