| /linux/drivers/crypto/axis/ |
| H A D | artpec6_crypto.c | 1090 struct artpec6_crypto_request_context *req_ctx = NULL; in artpec6_crypto_encrypt() local 1094 req_ctx = skcipher_request_ctx(req); in artpec6_crypto_encrypt() 1100 req_ctx->decrypt = 0; in artpec6_crypto_encrypt() 1115 ret = artpec6_crypto_common_init(&req_ctx->common, in artpec6_crypto_encrypt() 1124 artpec6_crypto_common_destroy(&req_ctx->common); in artpec6_crypto_encrypt() 1128 return artpec6_crypto_submit(&req_ctx->common); in artpec6_crypto_encrypt() 1136 struct artpec6_crypto_request_context *req_ctx = NULL; in artpec6_crypto_decrypt() local 1139 req_ctx = skcipher_request_ctx(req); in artpec6_crypto_decrypt() 1145 req_ctx->decrypt = 1; in artpec6_crypto_decrypt() 1161 ret = artpec6_crypto_common_init(&req_ctx->common, &req->base, in artpec6_crypto_decrypt() [all …]
|
| /linux/arch/x86/crypto/ |
| H A D | aria_aesni_avx2_glue.c | 88 struct aria_avx2_request_ctx *req_ctx = skcipher_request_ctx(req); in aria_avx2_ctr_encrypt() local 104 &req_ctx->keystream[0], in aria_avx2_ctr_encrypt() 115 &req_ctx->keystream[0], in aria_avx2_ctr_encrypt() 124 memcpy(&req_ctx->keystream[0], walk.iv, ARIA_BLOCK_SIZE); in aria_avx2_ctr_encrypt() 127 aria_encrypt(ctx, &req_ctx->keystream[0], in aria_avx2_ctr_encrypt() 128 &req_ctx->keystream[0]); in aria_avx2_ctr_encrypt() 130 crypto_xor_cpy(dst, src, &req_ctx->keystream[0], in aria_avx2_ctr_encrypt() 138 memcpy(&req_ctx->keystream[0], walk.iv, in aria_avx2_ctr_encrypt() 142 aria_encrypt(ctx, &req_ctx->keystream[0], in aria_avx2_ctr_encrypt() 143 &req_ctx->keystream[0]); in aria_avx2_ctr_encrypt() [all …]
|
| H A D | aria_aesni_avx_glue.c | 86 struct aria_avx_request_ctx *req_ctx = skcipher_request_ctx(req); in aria_avx_ctr_encrypt() local 102 &req_ctx->keystream[0], in aria_avx_ctr_encrypt() 111 memcpy(&req_ctx->keystream[0], walk.iv, ARIA_BLOCK_SIZE); in aria_avx_ctr_encrypt() 114 aria_encrypt(ctx, &req_ctx->keystream[0], in aria_avx_ctr_encrypt() 115 &req_ctx->keystream[0]); in aria_avx_ctr_encrypt() 117 crypto_xor_cpy(dst, src, &req_ctx->keystream[0], in aria_avx_ctr_encrypt() 125 memcpy(&req_ctx->keystream[0], walk.iv, in aria_avx_ctr_encrypt() 129 aria_encrypt(ctx, &req_ctx->keystream[0], in aria_avx_ctr_encrypt() 130 &req_ctx->keystream[0]); in aria_avx_ctr_encrypt() 132 crypto_xor_cpy(dst, src, &req_ctx->keystream[0], in aria_avx_ctr_encrypt()
|
| /linux/drivers/crypto/caam/ |
| H A D | caamalg_qi2.c | 356 struct caam_request *req_ctx = aead_request_ctx_dma(req); in aead_edesc_alloc() local 357 struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1]; in aead_edesc_alloc() 358 struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0]; in aead_edesc_alloc() 553 memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt)); in aead_edesc_alloc() 1116 struct caam_request *req_ctx = skcipher_request_ctx_dma(req); in skcipher_edesc_alloc() local 1117 struct dpaa2_fl_entry *in_fle = &req_ctx->fd_flt[1]; in skcipher_edesc_alloc() 1118 struct dpaa2_fl_entry *out_fle = &req_ctx->fd_flt[0]; in skcipher_edesc_alloc() 1242 memset(&req_ctx->fd_flt, 0, sizeof(req_ctx->fd_flt)); in skcipher_edesc_alloc() 1290 struct caam_request *req_ctx = to_caam_req(areq); in aead_encrypt_done() local 1291 struct aead_edesc *edesc = req_ctx->edesc; in aead_encrypt_done() [all …]
|
| /linux/drivers/crypto/ccree/ |
| H A D | cc_buffer_mgr.c | 343 struct cipher_req_ctx *req_ctx = (struct cipher_req_ctx *)ctx; in cc_unmap_cipher_request() local 345 if (req_ctx->gen_ctx.iv_dma_addr) { in cc_unmap_cipher_request() 347 &req_ctx->gen_ctx.iv_dma_addr, ivsize); in cc_unmap_cipher_request() 348 dma_unmap_single(dev, req_ctx->gen_ctx.iv_dma_addr, in cc_unmap_cipher_request() 352 if (req_ctx->dma_buf_type == CC_DMA_BUF_MLLI && in cc_unmap_cipher_request() 353 req_ctx->mlli_params.mlli_virt_addr) { in cc_unmap_cipher_request() 354 dma_pool_free(req_ctx->mlli_params.curr_pool, in cc_unmap_cipher_request() 355 req_ctx->mlli_params.mlli_virt_addr, in cc_unmap_cipher_request() 356 req_ctx->mlli_params.mlli_dma_addr); in cc_unmap_cipher_request() 360 dma_unmap_sg(dev, src, req_ctx->in_nents, DMA_TO_DEVICE); in cc_unmap_cipher_request() [all …]
|
| /linux/drivers/mtd/nand/ |
| H A D | ecc-sw-bch.c | 234 ret = nand_ecc_init_req_tweaking(&engine_conf->req_ctx, nand); in nand_ecc_sw_bch_init_ctx() 267 nand_ecc_cleanup_req_tweaking(&engine_conf->req_ctx); in nand_ecc_sw_bch_init_ctx() 283 nand_ecc_cleanup_req_tweaking(&engine_conf->req_ctx); in nand_ecc_sw_bch_cleanup_ctx() 312 nand_ecc_tweak_req(&engine_conf->req_ctx, req); in nand_ecc_sw_bch_prepare_io_req() 353 nand_ecc_restore_req(&engine_conf->req_ctx, req); in nand_ecc_sw_bch_finish_io_req() 382 nand_ecc_restore_req(&engine_conf->req_ctx, req); in nand_ecc_sw_bch_finish_io_req()
|
| H A D | ecc-realtek.c | 86 struct nand_ecc_req_tweak_ctx req_ctx; member 232 nand_ecc_tweak_req(&ctx->req_ctx, req); in rtl_ecc_prepare_io_req() 268 nand_ecc_restore_req(&ctx->req_ctx, req); in rtl_ecc_finish_io_req() 298 nand_ecc_restore_req(&ctx->req_ctx, req); in rtl_ecc_finish_io_req() 364 ret = nand_ecc_init_req_tweaking(&ctx->req_ctx, nand); in rtl_ecc_init_ctx() 380 nand_ecc_cleanup_req_tweaking(&ctx->req_ctx); in rtl_ecc_cleanup_ctx()
|
| H A D | ecc-sw-hamming.c | 503 ret = nand_ecc_init_req_tweaking(&engine_conf->req_ctx, nand); in nand_ecc_sw_hamming_init_ctx() 522 nand_ecc_cleanup_req_tweaking(&engine_conf->req_ctx); in nand_ecc_sw_hamming_init_ctx() 537 nand_ecc_cleanup_req_tweaking(&engine_conf->req_ctx); in nand_ecc_sw_hamming_cleanup_ctx() 566 nand_ecc_tweak_req(&engine_conf->req_ctx, req); in nand_ecc_sw_hamming_prepare_io_req() 607 nand_ecc_restore_req(&engine_conf->req_ctx, req); in nand_ecc_sw_hamming_finish_io_req() 636 nand_ecc_restore_req(&engine_conf->req_ctx, req); in nand_ecc_sw_hamming_finish_io_req()
|
| /linux/include/linux/mtd/ |
| H A D | nand-ecc-sw-bch.h | 26 struct nand_ecc_req_tweak_ctx req_ctx; member
|
| H A D | nand-ecc-sw-hamming.h | 25 struct nand_ecc_req_tweak_ctx req_ctx; member
|