| /linux/drivers/crypto/tegra/ |
| H A D | tegra-se-aes.c | 106 struct tegra_aes_reqctx *rctx = skcipher_request_ctx(req); in tegra_cbc_iv_copyback() local 111 if (rctx->encrypt) in tegra_cbc_iv_copyback() 112 memcpy(req->iv, rctx->datbuf.buf + offset, ctx->ivsize); in tegra_cbc_iv_copyback() 209 struct tegra_aes_reqctx *rctx) in tegra_aes_prep_cmd() argument 214 dma_addr_t addr = rctx->datbuf.addr; in tegra_aes_prep_cmd() 216 data_count = rctx->len / AES_BLOCK_SIZE; in tegra_aes_prep_cmd() 217 res_bits = (rctx->len % AES_BLOCK_SIZE) * 8; in tegra_aes_prep_cmd() 226 if (rctx->iv) { in tegra_aes_prep_cmd() 230 cpuvaddr[i++] = rctx->iv[j]; in tegra_aes_prep_cmd() 238 cpuvaddr[i++] = rctx->config; in tegra_aes_prep_cmd() [all …]
|
| /linux/drivers/crypto/qce/ |
| H A D | aead.c | 27 struct qce_aead_reqctx *rctx = aead_request_ctx_dma(req); in qce_aead_done() local 49 dma_unmap_sg(qce->dev, rctx->src_sg, rctx->src_nents, dir_src); in qce_aead_done() 51 dma_unmap_sg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst); in qce_aead_done() 53 if (IS_CCM(rctx->flags)) { in qce_aead_done() 55 sg_free_table(&rctx->src_tbl); in qce_aead_done() 57 sg_free_table(&rctx->dst_tbl); in qce_aead_done() 59 if (!(IS_DECRYPT(rctx->flags) && !diff_dst)) in qce_aead_done() 60 sg_free_table(&rctx->dst_tbl); in qce_aead_done() 63 sg_free_table(&rctx->dst_tbl); in qce_aead_done() 70 if (IS_ENCRYPT(rctx->flags)) { in qce_aead_done() [all …]
|
| H A D | sha.c | 41 struct qce_sha_reqctx *rctx = ahash_request_ctx_dma(req); in qce_ahash_done() local 53 dma_unmap_sg(qce->dev, req->src, rctx->src_nents, DMA_TO_DEVICE); in qce_ahash_done() 54 dma_unmap_sg(qce->dev, &rctx->result_sg, 1, DMA_FROM_DEVICE); in qce_ahash_done() 56 memcpy(rctx->digest, result->auth_iv, digestsize); in qce_ahash_done() 57 if (req->result && rctx->last_blk) in qce_ahash_done() 60 rctx->byte_count[0] = cpu_to_be32(result->auth_byte_count[0]); in qce_ahash_done() 61 rctx->byte_count[1] = cpu_to_be32(result->auth_byte_count[1]); in qce_ahash_done() 67 req->src = rctx->src_orig; in qce_ahash_done() 68 req->nbytes = rctx->nbytes_orig; in qce_ahash_done() 69 rctx->last_blk = false; in qce_ahash_done() [all …]
|
| H A D | skcipher.c | 31 struct qce_cipher_reqctx *rctx = skcipher_request_ctx(req); in qce_skcipher_done() local 50 dma_unmap_sg(qce->dev, rctx->src_sg, rctx->src_nents, dir_src); in qce_skcipher_done() 51 dma_unmap_sg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst); in qce_skcipher_done() 53 sg_free_table(&rctx->dst_tbl); in qce_skcipher_done() 59 memcpy(rctx->iv, result_buf->encr_cntr_iv, rctx->ivsize); in qce_skcipher_done() 67 struct qce_cipher_reqctx *rctx = skcipher_request_ctx(req); in qce_skcipher_async_req_handle() local 77 rctx->iv = req->iv; in qce_skcipher_async_req_handle() 78 rctx->ivsize = crypto_skcipher_ivsize(skcipher); in qce_skcipher_async_req_handle() 79 rctx->cryptlen = req->cryptlen; in qce_skcipher_async_req_handle() 85 rctx->src_nents = sg_nents_for_len(req->src, req->cryptlen); in qce_skcipher_async_req_handle() [all …]
|
| H A D | common.c | 151 struct qce_sha_reqctx *rctx = ahash_request_ctx_dma(req); in qce_setup_regs_ahash() local 162 if (!rctx->last_blk && req->nbytes % blocksize) in qce_setup_regs_ahash() 167 if (IS_CMAC(rctx->flags)) { in qce_setup_regs_ahash() 175 auth_cfg = qce_auth_cfg(rctx->flags, rctx->authklen, digestsize); in qce_setup_regs_ahash() 178 if (IS_SHA_HMAC(rctx->flags) || IS_CMAC(rctx->flags)) { in qce_setup_regs_ahash() 179 u32 authkey_words = rctx->authklen / sizeof(u32); in qce_setup_regs_ahash() 181 qce_cpu_to_be32p_array(mackey, rctx->authkey, rctx->authklen); in qce_setup_regs_ahash() 186 if (IS_CMAC(rctx->flags)) in qce_setup_regs_ahash() 189 if (rctx->first_blk) in qce_setup_regs_ahash() 190 memcpy(auth, rctx->digest, digestsize); in qce_setup_regs_ahash() [all …]
|
| /linux/drivers/crypto/ccp/ |
| H A D | ccp-crypto-aes-cmac.c | 28 struct ccp_aes_cmac_req_ctx *rctx = ahash_request_ctx_dma(req); in ccp_aes_cmac_complete() local 34 if (rctx->hash_rem) { in ccp_aes_cmac_complete() 36 unsigned int offset = rctx->nbytes - rctx->hash_rem; in ccp_aes_cmac_complete() 38 scatterwalk_map_and_copy(rctx->buf, rctx->src, in ccp_aes_cmac_complete() 39 offset, rctx->hash_rem, 0); in ccp_aes_cmac_complete() 40 rctx->buf_count = rctx->hash_rem; in ccp_aes_cmac_complete() 42 rctx->buf_count = 0; in ccp_aes_cmac_complete() 46 if (req->result && rctx->final) in ccp_aes_cmac_complete() 47 memcpy(req->result, rctx->iv, digest_size); in ccp_aes_cmac_complete() 50 sg_free_table(&rctx->data_sg); in ccp_aes_cmac_complete() [all …]
|
| H A D | ccp-crypto-sha.c | 31 struct ccp_sha_req_ctx *rctx = ahash_request_ctx_dma(req); in ccp_sha_complete() local 37 if (rctx->hash_rem) { in ccp_sha_complete() 39 unsigned int offset = rctx->nbytes - rctx->hash_rem; in ccp_sha_complete() 41 scatterwalk_map_and_copy(rctx->buf, rctx->src, in ccp_sha_complete() 42 offset, rctx->hash_rem, 0); in ccp_sha_complete() 43 rctx->buf_count = rctx->hash_rem; in ccp_sha_complete() 45 rctx->buf_count = 0; in ccp_sha_complete() 49 if (req->result && rctx->final) in ccp_sha_complete() 50 memcpy(req->result, rctx->ctx, digest_size); in ccp_sha_complete() 53 sg_free_table(&rctx->data_sg); in ccp_sha_complete() [all …]
|
| H A D | ccp-crypto-aes-xts.c | 65 struct ccp_aes_req_ctx *rctx = skcipher_request_ctx_dma(req); in ccp_aes_xts_complete() local 70 memcpy(req->iv, rctx->iv, AES_BLOCK_SIZE); in ccp_aes_xts_complete() 109 struct ccp_aes_req_ctx *rctx = skcipher_request_ctx_dma(req); in ccp_aes_xts_crypt() local 151 skcipher_request_set_tfm(&rctx->fallback_req, in ccp_aes_xts_crypt() 153 skcipher_request_set_callback(&rctx->fallback_req, in ccp_aes_xts_crypt() 157 skcipher_request_set_crypt(&rctx->fallback_req, req->src, in ccp_aes_xts_crypt() 159 ret = encrypt ? crypto_skcipher_encrypt(&rctx->fallback_req) : in ccp_aes_xts_crypt() 160 crypto_skcipher_decrypt(&rctx->fallback_req); in ccp_aes_xts_crypt() 164 memcpy(rctx->iv, req->iv, AES_BLOCK_SIZE); in ccp_aes_xts_crypt() 165 sg_init_one(&rctx->iv_sg, rctx->iv, AES_BLOCK_SIZE); in ccp_aes_xts_crypt() [all …]
|
| H A D | ccp-crypto-aes-galois.c | 80 struct ccp_aes_req_ctx *rctx = aead_request_ctx_dma(req); in ccp_aes_gcm_crypt() local 105 memcpy(rctx->iv, req->iv, GCM_AES_IV_SIZE); in ccp_aes_gcm_crypt() 107 rctx->iv[i + GCM_AES_IV_SIZE] = 0; in ccp_aes_gcm_crypt() 108 rctx->iv[AES_BLOCK_SIZE - 1] = 1; in ccp_aes_gcm_crypt() 111 iv_sg = &rctx->iv_sg; in ccp_aes_gcm_crypt() 113 sg_init_one(iv_sg, rctx->iv, iv_len); in ccp_aes_gcm_crypt() 116 memset(&rctx->cmd, 0, sizeof(rctx->cmd)); in ccp_aes_gcm_crypt() 117 INIT_LIST_HEAD(&rctx->cmd.entry); in ccp_aes_gcm_crypt() 118 rctx->cmd.engine = CCP_ENGINE_AES; in ccp_aes_gcm_crypt() 119 rctx->cmd.u.aes.authsize = crypto_aead_authsize(tfm); in ccp_aes_gcm_crypt() [all …]
|
| H A D | ccp-crypto-aes.c | 28 struct ccp_aes_req_ctx *rctx = skcipher_request_ctx_dma(req); in ccp_aes_complete() local 34 memcpy(req->iv, rctx->iv, AES_BLOCK_SIZE); in ccp_aes_complete() 71 struct ccp_aes_req_ctx *rctx = skcipher_request_ctx_dma(req); in ccp_aes_crypt() local 87 memcpy(rctx->iv, req->iv, AES_BLOCK_SIZE); in ccp_aes_crypt() 88 iv_sg = &rctx->iv_sg; in ccp_aes_crypt() 90 sg_init_one(iv_sg, rctx->iv, iv_len); in ccp_aes_crypt() 93 memset(&rctx->cmd, 0, sizeof(rctx->cmd)); in ccp_aes_crypt() 94 INIT_LIST_HEAD(&rctx->cmd.entry); in ccp_aes_crypt() 95 rctx->cmd.engine = CCP_ENGINE_AES; in ccp_aes_crypt() 96 rctx->cmd.u.aes.type = ctx->u.aes.type; in ccp_aes_crypt() [all …]
|
| H A D | ccp-crypto-des3.c | 27 struct ccp_des3_req_ctx *rctx = skcipher_request_ctx_dma(req); in ccp_des3_complete() local 33 memcpy(req->iv, rctx->iv, DES3_EDE_BLOCK_SIZE); in ccp_des3_complete() 66 struct ccp_des3_req_ctx *rctx = skcipher_request_ctx_dma(req); in ccp_des3_crypt() local 82 memcpy(rctx->iv, req->iv, DES3_EDE_BLOCK_SIZE); in ccp_des3_crypt() 83 iv_sg = &rctx->iv_sg; in ccp_des3_crypt() 85 sg_init_one(iv_sg, rctx->iv, iv_len); in ccp_des3_crypt() 88 memset(&rctx->cmd, 0, sizeof(rctx->cmd)); in ccp_des3_crypt() 89 INIT_LIST_HEAD(&rctx->cmd.entry); in ccp_des3_crypt() 90 rctx->cmd.engine = CCP_ENGINE_DES3; in ccp_des3_crypt() 91 rctx->cmd.u.des3.type = ctx->u.des3.type; in ccp_des3_crypt() [all …]
|
| H A D | ccp-crypto-rsa.c | 47 struct ccp_rsa_req_ctx *rctx = akcipher_request_ctx_dma(req); in ccp_rsa_complete() local 52 req->dst_len = rctx->cmd.u.rsa.key_size >> 3; in ccp_rsa_complete() 68 struct ccp_rsa_req_ctx *rctx = akcipher_request_ctx_dma(req); in ccp_rsa_crypt() local 71 memset(&rctx->cmd, 0, sizeof(rctx->cmd)); in ccp_rsa_crypt() 72 INIT_LIST_HEAD(&rctx->cmd.entry); in ccp_rsa_crypt() 73 rctx->cmd.engine = CCP_ENGINE_RSA; in ccp_rsa_crypt() 75 rctx->cmd.u.rsa.key_size = ctx->u.rsa.key_len; /* in bits */ in ccp_rsa_crypt() 77 rctx->cmd.u.rsa.exp = &ctx->u.rsa.e_sg; in ccp_rsa_crypt() 78 rctx->cmd.u.rsa.exp_len = ctx->u.rsa.e_len; in ccp_rsa_crypt() 80 rctx->cmd.u.rsa.exp = &ctx->u.rsa.d_sg; in ccp_rsa_crypt() [all …]
|
| /linux/drivers/crypto/inside-secure/eip93/ |
| H A D | eip93-common.c | 201 int check_valid_request(struct eip93_cipher_reqctx *rctx) in check_valid_request() argument 203 struct scatterlist *src = rctx->sg_src; in check_valid_request() 204 struct scatterlist *dst = rctx->sg_dst; in check_valid_request() 205 u32 textsize = rctx->textsize; in check_valid_request() 206 u32 authsize = rctx->authsize; in check_valid_request() 207 u32 blksize = rctx->blksize; in check_valid_request() 208 u32 totlen_src = rctx->assoclen + rctx->textsize; in check_valid_request() 209 u32 totlen_dst = rctx->assoclen + rctx->textsize; in check_valid_request() 215 if (!IS_CTR(rctx->flags)) { in check_valid_request() 221 if (IS_ENCRYPT(rctx->flags)) in check_valid_request() [all …]
|
| H A D | eip93-hash.c | 23 struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req); in eip93_hash_free_data_blocks() local 29 list_for_each_entry_safe(block, tmp, &rctx->blocks, list) { in eip93_hash_free_data_blocks() 34 if (!list_empty(&rctx->blocks)) in eip93_hash_free_data_blocks() 35 INIT_LIST_HEAD(&rctx->blocks); in eip93_hash_free_data_blocks() 37 if (rctx->finalize) in eip93_hash_free_data_blocks() 38 dma_unmap_single(eip93->dev, rctx->data_dma, in eip93_hash_free_data_blocks() 39 rctx->data_used, in eip93_hash_free_data_blocks() 45 struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req); in eip93_hash_free_sa_record() local 51 dma_unmap_single(eip93->dev, rctx->sa_record_hmac_base, in eip93_hash_free_sa_record() 52 sizeof(rctx->sa_record_hmac), DMA_TO_DEVICE); in eip93_hash_free_sa_record() [all …]
|
| H A D | eip93-cipher.c | 25 struct eip93_cipher_reqctx *rctx = skcipher_request_ctx(req); in eip93_skcipher_handle_result() local 27 eip93_unmap_dma(eip93, rctx, req->src, req->dst); in eip93_skcipher_handle_result() 28 eip93_handle_result(eip93, rctx, req->iv); in eip93_skcipher_handle_result() 36 struct eip93_cipher_reqctx *rctx = skcipher_request_ctx(req); in eip93_skcipher_send_req() local 39 err = check_valid_request(rctx); in eip93_skcipher_send_req() 46 return eip93_send_req(async, req->iv, rctx); in eip93_skcipher_send_req() 138 struct eip93_cipher_reqctx *rctx = skcipher_request_ctx(req); in eip93_skcipher_crypt() local 151 if (IS_ECB(rctx->flags) || IS_CBC(rctx->flags)) in eip93_skcipher_crypt() 162 rctx->assoclen = 0; in eip93_skcipher_crypt() 163 rctx->textsize = req->cryptlen; in eip93_skcipher_crypt() [all …]
|
| H A D | eip93-aead.c | 35 struct eip93_cipher_reqctx *rctx = aead_request_ctx(req); in eip93_aead_handle_result() local 37 eip93_unmap_dma(eip93, rctx, req->src, req->dst); in eip93_aead_handle_result() 38 eip93_handle_result(eip93, rctx, req->iv); in eip93_aead_handle_result() 46 struct eip93_cipher_reqctx *rctx = aead_request_ctx(req); in eip93_aead_send_req() local 49 err = check_valid_request(rctx); in eip93_aead_send_req() 55 return eip93_send_req(async, req->iv, rctx); in eip93_aead_send_req() 188 struct eip93_cipher_reqctx *rctx = aead_request_ctx(req); in eip93_aead_crypt() local 200 rctx->textsize = req->cryptlen; in eip93_aead_crypt() 201 rctx->blksize = ctx->blksize; in eip93_aead_crypt() 202 rctx->assoclen = req->assoclen; in eip93_aead_crypt() [all …]
|
| /linux/drivers/crypto/cavium/nitrox/ |
| H A D | nitrox_aead.c | 151 static int nitrox_set_creq(struct nitrox_aead_rctx *rctx) in nitrox_set_creq() argument 153 struct se_crypto_request *creq = &rctx->nkreq.creq; in nitrox_set_creq() 157 creq->flags = rctx->flags; in nitrox_set_creq() 158 creq->gfp = (rctx->flags & CRYPTO_TFM_REQ_MAY_SLEEP) ? GFP_KERNEL : in nitrox_set_creq() 163 creq->ctrl.s.arg = rctx->ctrl_arg; in nitrox_set_creq() 165 creq->gph.param0 = cpu_to_be16(rctx->cryptlen); in nitrox_set_creq() 166 creq->gph.param1 = cpu_to_be16(rctx->cryptlen + rctx->assoclen); in nitrox_set_creq() 167 creq->gph.param2 = cpu_to_be16(rctx->ivsize + rctx->assoclen); in nitrox_set_creq() 169 param3.auth_offset = rctx->ivsize; in nitrox_set_creq() 172 creq->ctx_handle = rctx->ctx_handle; in nitrox_set_creq() [all …]
|
| /linux/drivers/crypto/allwinner/sun8i-ss/ |
| H A D | sun8i-ss-hash.c | 146 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); in sun8i_ss_hash_init() local 150 memset(rctx, 0, sizeof(struct sun8i_ss_hash_reqctx)); in sun8i_ss_hash_init() 152 ahash_request_set_tfm(&rctx->fallback_req, tfmctx->fallback_tfm); in sun8i_ss_hash_init() 153 ahash_request_set_callback(&rctx->fallback_req, in sun8i_ss_hash_init() 157 return crypto_ahash_init(&rctx->fallback_req); in sun8i_ss_hash_init() 162 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); in sun8i_ss_hash_export() local 166 ahash_request_set_tfm(&rctx->fallback_req, tfmctx->fallback_tfm); in sun8i_ss_hash_export() 167 ahash_request_set_callback(&rctx->fallback_req, in sun8i_ss_hash_export() 171 return crypto_ahash_export(&rctx->fallback_req, out); in sun8i_ss_hash_export() 176 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); in sun8i_ss_hash_import() local [all …]
|
| /linux/drivers/crypto/bcm/ |
| H A D | cipher.c | 135 struct iproc_reqctx_s *rctx, in spu_skcipher_rx_sg_create() argument 141 struct iproc_ctx_s *ctx = rctx->ctx; in spu_skcipher_rx_sg_create() 144 mssg->spu.dst = kmalloc_objs(struct scatterlist, rx_frag_num, rctx->gfp); in spu_skcipher_rx_sg_create() 151 sg_set_buf(sg++, rctx->msg_buf.spu_resp_hdr, ctx->spu_resp_hdr_len); in spu_skcipher_rx_sg_create() 156 sg_set_buf(sg++, rctx->msg_buf.c.supdt_tweak, in spu_skcipher_rx_sg_create() 160 datalen = spu_msg_sg_add(&sg, &rctx->dst_sg, &rctx->dst_skip, in spu_skcipher_rx_sg_create() 161 rctx->dst_nents, chunksize); in spu_skcipher_rx_sg_create() 169 sg_set_buf(sg++, rctx->msg_buf.rx_stat_pad, stat_pad_len); in spu_skcipher_rx_sg_create() 171 memset(rctx->msg_buf.rx_stat, 0, SPU_RX_STATUS_LEN); in spu_skcipher_rx_sg_create() 172 sg_set_buf(sg, rctx->msg_buf.rx_stat, spu->spu_rx_status_len()); in spu_skcipher_rx_sg_create() [all …]
|
| /linux/drivers/crypto/rockchip/ |
| H A D | rk3288_crypto_skcipher.c | 72 struct rk_cipher_rctx *rctx = skcipher_request_ctx(areq); in rk_cipher_fallback() local 79 skcipher_request_set_tfm(&rctx->fallback_req, op->fallback_tfm); in rk_cipher_fallback() 80 skcipher_request_set_callback(&rctx->fallback_req, areq->base.flags, in rk_cipher_fallback() 82 skcipher_request_set_crypt(&rctx->fallback_req, areq->src, areq->dst, in rk_cipher_fallback() 84 if (rctx->mode & RK_CRYPTO_DEC) in rk_cipher_fallback() 85 err = crypto_skcipher_decrypt(&rctx->fallback_req); in rk_cipher_fallback() 87 err = crypto_skcipher_encrypt(&rctx->fallback_req); in rk_cipher_fallback() 93 struct rk_cipher_rctx *rctx = skcipher_request_ctx(req); in rk_cipher_handle_req() local 103 rctx->dev = rkc; in rk_cipher_handle_req() 157 struct rk_cipher_rctx *rctx = skcipher_request_ctx(req); in rk_aes_ecb_encrypt() local [all …]
|
| /linux/drivers/crypto/starfive/ |
| H A D | jh7110-hash.c | 66 struct starfive_cryp_request_ctx *rctx = ctx->rctx; in starfive_hash_hmac_key() local 74 rctx->csr.hash.hmac = 1; in starfive_hash_hmac_key() 75 rctx->csr.hash.key_flag = 1; in starfive_hash_hmac_key() 77 writel(rctx->csr.hash.v, cryp->base + STARFIVE_HASH_SHACSR); in starfive_hash_hmac_key() 167 struct starfive_cryp_request_ctx *rctx = ahash_request_ctx(req); in starfive_hash_copy_hash() local 175 mlen = rctx->digsize / sizeof(u32); in starfive_hash_copy_hash() 200 struct starfive_cryp_request_ctx *rctx = ctx->rctx; in starfive_hash_one_request() local 210 rctx->csr.hash.v = 0; in starfive_hash_one_request() 211 rctx->csr.hash.mode = ctx->hash_mode; in starfive_hash_one_request() 218 rctx->csr.hash.start = 1; in starfive_hash_one_request() [all …]
|
| /linux/drivers/crypto/gemini/ |
| H A D | sl3516-ce-cipher.c | 107 struct sl3516_ce_cipher_req_ctx *rctx = skcipher_request_ctx(areq); in sl3516_ce_cipher_fallback() local 115 skcipher_request_set_tfm(&rctx->fallback_req, op->fallback_tfm); in sl3516_ce_cipher_fallback() 116 skcipher_request_set_callback(&rctx->fallback_req, areq->base.flags, in sl3516_ce_cipher_fallback() 118 skcipher_request_set_crypt(&rctx->fallback_req, areq->src, areq->dst, in sl3516_ce_cipher_fallback() 120 if (rctx->op_dir == CE_DECRYPTION) in sl3516_ce_cipher_fallback() 121 err = crypto_skcipher_decrypt(&rctx->fallback_req); in sl3516_ce_cipher_fallback() 123 err = crypto_skcipher_encrypt(&rctx->fallback_req); in sl3516_ce_cipher_fallback() 132 struct sl3516_ce_cipher_req_ctx *rctx = skcipher_request_ctx(areq); in sl3516_ce_cipher() local 148 rctx->op_dir, areq->iv, crypto_skcipher_ivsize(tfm), in sl3516_ce_cipher() 185 rctx->t_src[i].addr = sg_dma_address(sg); in sl3516_ce_cipher() [all …]
|
| /linux/crypto/ |
| H A D | hctr2.c | 138 struct hctr2_request_ctx *rctx = skcipher_request_ctx(req); in hctr2_hash_tweak() local 139 struct polyval_ctx *poly_ctx = &rctx->u.poly_ctx; in hctr2_hash_tweak() 149 polyval_export_blkaligned(poly_ctx, &rctx->hashed_tweak); in hctr2_hash_tweak() 157 struct hctr2_request_ctx *rctx = skcipher_request_ctx(req); in hctr2_hash_message() local 158 struct polyval_ctx *poly_ctx = &rctx->u.poly_ctx; in hctr2_hash_message() 182 struct hctr2_request_ctx *rctx = skcipher_request_ctx(req); in hctr2_finish() local 183 struct polyval_ctx *poly_ctx = &rctx->u.poly_ctx; in hctr2_finish() 189 &rctx->hashed_tweak); in hctr2_finish() 190 hctr2_hash_message(req, rctx->bulk_part_dst, digest); in hctr2_finish() 191 crypto_xor(rctx->first_block, digest, BLOCKCIPHER_BLOCK_SIZE); in hctr2_finish() [all …]
|
| /linux/drivers/crypto/amlogic/ |
| H A D | amlogic-gxl-cipher.c | 62 struct meson_cipher_req_ctx *rctx = skcipher_request_ctx(areq); in meson_cipher_do_fallback() local 71 skcipher_request_set_tfm(&rctx->fallback_req, op->fallback_tfm); in meson_cipher_do_fallback() 72 skcipher_request_set_callback(&rctx->fallback_req, areq->base.flags, in meson_cipher_do_fallback() 74 skcipher_request_set_crypt(&rctx->fallback_req, areq->src, areq->dst, in meson_cipher_do_fallback() 77 if (rctx->op_dir == MESON_DECRYPT) in meson_cipher_do_fallback() 78 err = crypto_skcipher_decrypt(&rctx->fallback_req); in meson_cipher_do_fallback() 80 err = crypto_skcipher_encrypt(&rctx->fallback_req); in meson_cipher_do_fallback() 88 struct meson_cipher_req_ctx *rctx = skcipher_request_ctx(areq); in meson_cipher() local 92 int flow = rctx->flow; in meson_cipher() 109 rctx->op_dir, crypto_skcipher_ivsize(tfm), in meson_cipher() [all …]
|
| /linux/drivers/crypto/ |
| H A D | omap-sham.c | 727 struct omap_sham_reqctx *rctx) in omap_sham_align_sgs() argument 734 int offset = rctx->offset; in omap_sham_align_sgs() 735 int bufcnt = rctx->bufcnt; in omap_sham_align_sgs() 740 sg_init_table(rctx->sgl, 1); in omap_sham_align_sgs() 741 sg_set_buf(rctx->sgl, rctx->dd->xmit_buf, bufcnt); in omap_sham_align_sgs() 742 rctx->sg = rctx->sgl; in omap_sham_align_sgs() 743 rctx->sg_len = 1; in omap_sham_align_sgs() 824 return omap_sham_copy_sgs(rctx, sg, bs, new_len); in omap_sham_align_sgs() 826 return omap_sham_copy_sg_lists(rctx, sg, bs, new_len); in omap_sham_align_sgs() 828 rctx->total = new_len; in omap_sham_align_sgs() [all …]
|