Home
last modified time | relevance | path

Searched refs:rctx (Results 1 – 25 of 51) sorted by relevance

123

/linux/drivers/crypto/tegra/
H A Dtegra-se-aes.c106 struct tegra_aes_reqctx *rctx = skcipher_request_ctx(req); in tegra_cbc_iv_copyback() local
111 if (rctx->encrypt) in tegra_cbc_iv_copyback()
112 memcpy(req->iv, rctx->datbuf.buf + offset, ctx->ivsize); in tegra_cbc_iv_copyback()
209 struct tegra_aes_reqctx *rctx) in tegra_aes_prep_cmd() argument
214 dma_addr_t addr = rctx->datbuf.addr; in tegra_aes_prep_cmd()
216 data_count = rctx->len / AES_BLOCK_SIZE; in tegra_aes_prep_cmd()
217 res_bits = (rctx->len % AES_BLOCK_SIZE) * 8; in tegra_aes_prep_cmd()
226 if (rctx->iv) { in tegra_aes_prep_cmd()
230 cpuvaddr[i++] = rctx->iv[j]; in tegra_aes_prep_cmd()
238 cpuvaddr[i++] = rctx->config; in tegra_aes_prep_cmd()
[all …]
/linux/drivers/crypto/qce/
H A Daead.c27 struct qce_aead_reqctx *rctx = aead_request_ctx_dma(req); in qce_aead_done() local
49 dma_unmap_sg(qce->dev, rctx->src_sg, rctx->src_nents, dir_src); in qce_aead_done()
51 dma_unmap_sg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst); in qce_aead_done()
53 if (IS_CCM(rctx->flags)) { in qce_aead_done()
55 sg_free_table(&rctx->src_tbl); in qce_aead_done()
57 sg_free_table(&rctx->dst_tbl); in qce_aead_done()
59 if (!(IS_DECRYPT(rctx->flags) && !diff_dst)) in qce_aead_done()
60 sg_free_table(&rctx->dst_tbl); in qce_aead_done()
63 sg_free_table(&rctx->dst_tbl); in qce_aead_done()
70 if (IS_ENCRYPT(rctx->flags)) { in qce_aead_done()
[all …]
H A Dsha.c41 struct qce_sha_reqctx *rctx = ahash_request_ctx_dma(req); in qce_ahash_done() local
53 dma_unmap_sg(qce->dev, req->src, rctx->src_nents, DMA_TO_DEVICE); in qce_ahash_done()
54 dma_unmap_sg(qce->dev, &rctx->result_sg, 1, DMA_FROM_DEVICE); in qce_ahash_done()
56 memcpy(rctx->digest, result->auth_iv, digestsize); in qce_ahash_done()
57 if (req->result && rctx->last_blk) in qce_ahash_done()
60 rctx->byte_count[0] = cpu_to_be32(result->auth_byte_count[0]); in qce_ahash_done()
61 rctx->byte_count[1] = cpu_to_be32(result->auth_byte_count[1]); in qce_ahash_done()
67 req->src = rctx->src_orig; in qce_ahash_done()
68 req->nbytes = rctx->nbytes_orig; in qce_ahash_done()
69 rctx->last_blk = false; in qce_ahash_done()
[all …]
H A Dskcipher.c31 struct qce_cipher_reqctx *rctx = skcipher_request_ctx(req); in qce_skcipher_done() local
50 dma_unmap_sg(qce->dev, rctx->src_sg, rctx->src_nents, dir_src); in qce_skcipher_done()
51 dma_unmap_sg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst); in qce_skcipher_done()
53 sg_free_table(&rctx->dst_tbl); in qce_skcipher_done()
59 memcpy(rctx->iv, result_buf->encr_cntr_iv, rctx->ivsize); in qce_skcipher_done()
67 struct qce_cipher_reqctx *rctx = skcipher_request_ctx(req); in qce_skcipher_async_req_handle() local
77 rctx->iv = req->iv; in qce_skcipher_async_req_handle()
78 rctx->ivsize = crypto_skcipher_ivsize(skcipher); in qce_skcipher_async_req_handle()
79 rctx->cryptlen = req->cryptlen; in qce_skcipher_async_req_handle()
85 rctx->src_nents = sg_nents_for_len(req->src, req->cryptlen); in qce_skcipher_async_req_handle()
[all …]
H A Dcommon.c151 struct qce_sha_reqctx *rctx = ahash_request_ctx_dma(req); in qce_setup_regs_ahash() local
162 if (!rctx->last_blk && req->nbytes % blocksize) in qce_setup_regs_ahash()
167 if (IS_CMAC(rctx->flags)) { in qce_setup_regs_ahash()
175 auth_cfg = qce_auth_cfg(rctx->flags, rctx->authklen, digestsize); in qce_setup_regs_ahash()
178 if (IS_SHA_HMAC(rctx->flags) || IS_CMAC(rctx->flags)) { in qce_setup_regs_ahash()
179 u32 authkey_words = rctx->authklen / sizeof(u32); in qce_setup_regs_ahash()
181 qce_cpu_to_be32p_array(mackey, rctx->authkey, rctx->authklen); in qce_setup_regs_ahash()
186 if (IS_CMAC(rctx->flags)) in qce_setup_regs_ahash()
189 if (rctx->first_blk) in qce_setup_regs_ahash()
190 memcpy(auth, rctx->digest, digestsize); in qce_setup_regs_ahash()
[all …]
/linux/drivers/crypto/ccp/
H A Dccp-crypto-aes-cmac.c28 struct ccp_aes_cmac_req_ctx *rctx = ahash_request_ctx_dma(req); in ccp_aes_cmac_complete() local
34 if (rctx->hash_rem) { in ccp_aes_cmac_complete()
36 unsigned int offset = rctx->nbytes - rctx->hash_rem; in ccp_aes_cmac_complete()
38 scatterwalk_map_and_copy(rctx->buf, rctx->src, in ccp_aes_cmac_complete()
39 offset, rctx->hash_rem, 0); in ccp_aes_cmac_complete()
40 rctx->buf_count = rctx->hash_rem; in ccp_aes_cmac_complete()
42 rctx->buf_count = 0; in ccp_aes_cmac_complete()
46 if (req->result && rctx->final) in ccp_aes_cmac_complete()
47 memcpy(req->result, rctx->iv, digest_size); in ccp_aes_cmac_complete()
50 sg_free_table(&rctx->data_sg); in ccp_aes_cmac_complete()
[all …]
H A Dccp-crypto-sha.c31 struct ccp_sha_req_ctx *rctx = ahash_request_ctx_dma(req); in ccp_sha_complete() local
37 if (rctx->hash_rem) { in ccp_sha_complete()
39 unsigned int offset = rctx->nbytes - rctx->hash_rem; in ccp_sha_complete()
41 scatterwalk_map_and_copy(rctx->buf, rctx->src, in ccp_sha_complete()
42 offset, rctx->hash_rem, 0); in ccp_sha_complete()
43 rctx->buf_count = rctx->hash_rem; in ccp_sha_complete()
45 rctx->buf_count = 0; in ccp_sha_complete()
49 if (req->result && rctx->final) in ccp_sha_complete()
50 memcpy(req->result, rctx->ctx, digest_size); in ccp_sha_complete()
53 sg_free_table(&rctx->data_sg); in ccp_sha_complete()
[all …]
H A Dccp-crypto-aes-xts.c65 struct ccp_aes_req_ctx *rctx = skcipher_request_ctx_dma(req); in ccp_aes_xts_complete() local
70 memcpy(req->iv, rctx->iv, AES_BLOCK_SIZE); in ccp_aes_xts_complete()
109 struct ccp_aes_req_ctx *rctx = skcipher_request_ctx_dma(req); in ccp_aes_xts_crypt() local
151 skcipher_request_set_tfm(&rctx->fallback_req, in ccp_aes_xts_crypt()
153 skcipher_request_set_callback(&rctx->fallback_req, in ccp_aes_xts_crypt()
157 skcipher_request_set_crypt(&rctx->fallback_req, req->src, in ccp_aes_xts_crypt()
159 ret = encrypt ? crypto_skcipher_encrypt(&rctx->fallback_req) : in ccp_aes_xts_crypt()
160 crypto_skcipher_decrypt(&rctx->fallback_req); in ccp_aes_xts_crypt()
164 memcpy(rctx->iv, req->iv, AES_BLOCK_SIZE); in ccp_aes_xts_crypt()
165 sg_init_one(&rctx->iv_sg, rctx->iv, AES_BLOCK_SIZE); in ccp_aes_xts_crypt()
[all …]
H A Dccp-crypto-aes-galois.c80 struct ccp_aes_req_ctx *rctx = aead_request_ctx_dma(req); in ccp_aes_gcm_crypt() local
105 memcpy(rctx->iv, req->iv, GCM_AES_IV_SIZE); in ccp_aes_gcm_crypt()
107 rctx->iv[i + GCM_AES_IV_SIZE] = 0; in ccp_aes_gcm_crypt()
108 rctx->iv[AES_BLOCK_SIZE - 1] = 1; in ccp_aes_gcm_crypt()
111 iv_sg = &rctx->iv_sg; in ccp_aes_gcm_crypt()
113 sg_init_one(iv_sg, rctx->iv, iv_len); in ccp_aes_gcm_crypt()
116 memset(&rctx->cmd, 0, sizeof(rctx->cmd)); in ccp_aes_gcm_crypt()
117 INIT_LIST_HEAD(&rctx->cmd.entry); in ccp_aes_gcm_crypt()
118 rctx->cmd.engine = CCP_ENGINE_AES; in ccp_aes_gcm_crypt()
119 rctx->cmd.u.aes.authsize = crypto_aead_authsize(tfm); in ccp_aes_gcm_crypt()
[all …]
H A Dccp-crypto-aes.c28 struct ccp_aes_req_ctx *rctx = skcipher_request_ctx_dma(req); in ccp_aes_complete() local
34 memcpy(req->iv, rctx->iv, AES_BLOCK_SIZE); in ccp_aes_complete()
71 struct ccp_aes_req_ctx *rctx = skcipher_request_ctx_dma(req); in ccp_aes_crypt() local
87 memcpy(rctx->iv, req->iv, AES_BLOCK_SIZE); in ccp_aes_crypt()
88 iv_sg = &rctx->iv_sg; in ccp_aes_crypt()
90 sg_init_one(iv_sg, rctx->iv, iv_len); in ccp_aes_crypt()
93 memset(&rctx->cmd, 0, sizeof(rctx->cmd)); in ccp_aes_crypt()
94 INIT_LIST_HEAD(&rctx->cmd.entry); in ccp_aes_crypt()
95 rctx->cmd.engine = CCP_ENGINE_AES; in ccp_aes_crypt()
96 rctx->cmd.u.aes.type = ctx->u.aes.type; in ccp_aes_crypt()
[all …]
H A Dccp-crypto-des3.c27 struct ccp_des3_req_ctx *rctx = skcipher_request_ctx_dma(req); in ccp_des3_complete() local
33 memcpy(req->iv, rctx->iv, DES3_EDE_BLOCK_SIZE); in ccp_des3_complete()
66 struct ccp_des3_req_ctx *rctx = skcipher_request_ctx_dma(req); in ccp_des3_crypt() local
82 memcpy(rctx->iv, req->iv, DES3_EDE_BLOCK_SIZE); in ccp_des3_crypt()
83 iv_sg = &rctx->iv_sg; in ccp_des3_crypt()
85 sg_init_one(iv_sg, rctx->iv, iv_len); in ccp_des3_crypt()
88 memset(&rctx->cmd, 0, sizeof(rctx->cmd)); in ccp_des3_crypt()
89 INIT_LIST_HEAD(&rctx->cmd.entry); in ccp_des3_crypt()
90 rctx->cmd.engine = CCP_ENGINE_DES3; in ccp_des3_crypt()
91 rctx->cmd.u.des3.type = ctx->u.des3.type; in ccp_des3_crypt()
[all …]
H A Dccp-crypto-rsa.c47 struct ccp_rsa_req_ctx *rctx = akcipher_request_ctx_dma(req); in ccp_rsa_complete() local
52 req->dst_len = rctx->cmd.u.rsa.key_size >> 3; in ccp_rsa_complete()
68 struct ccp_rsa_req_ctx *rctx = akcipher_request_ctx_dma(req); in ccp_rsa_crypt() local
71 memset(&rctx->cmd, 0, sizeof(rctx->cmd)); in ccp_rsa_crypt()
72 INIT_LIST_HEAD(&rctx->cmd.entry); in ccp_rsa_crypt()
73 rctx->cmd.engine = CCP_ENGINE_RSA; in ccp_rsa_crypt()
75 rctx->cmd.u.rsa.key_size = ctx->u.rsa.key_len; /* in bits */ in ccp_rsa_crypt()
77 rctx->cmd.u.rsa.exp = &ctx->u.rsa.e_sg; in ccp_rsa_crypt()
78 rctx->cmd.u.rsa.exp_len = ctx->u.rsa.e_len; in ccp_rsa_crypt()
80 rctx->cmd.u.rsa.exp = &ctx->u.rsa.d_sg; in ccp_rsa_crypt()
[all …]
/linux/drivers/crypto/inside-secure/eip93/
H A Deip93-common.c201 int check_valid_request(struct eip93_cipher_reqctx *rctx) in check_valid_request() argument
203 struct scatterlist *src = rctx->sg_src; in check_valid_request()
204 struct scatterlist *dst = rctx->sg_dst; in check_valid_request()
205 u32 textsize = rctx->textsize; in check_valid_request()
206 u32 authsize = rctx->authsize; in check_valid_request()
207 u32 blksize = rctx->blksize; in check_valid_request()
208 u32 totlen_src = rctx->assoclen + rctx->textsize; in check_valid_request()
209 u32 totlen_dst = rctx->assoclen + rctx->textsize; in check_valid_request()
215 if (!IS_CTR(rctx->flags)) { in check_valid_request()
221 if (IS_ENCRYPT(rctx->flags)) in check_valid_request()
[all …]
H A Deip93-hash.c23 struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req); in eip93_hash_free_data_blocks() local
29 list_for_each_entry_safe(block, tmp, &rctx->blocks, list) { in eip93_hash_free_data_blocks()
34 if (!list_empty(&rctx->blocks)) in eip93_hash_free_data_blocks()
35 INIT_LIST_HEAD(&rctx->blocks); in eip93_hash_free_data_blocks()
37 if (rctx->finalize) in eip93_hash_free_data_blocks()
38 dma_unmap_single(eip93->dev, rctx->data_dma, in eip93_hash_free_data_blocks()
39 rctx->data_used, in eip93_hash_free_data_blocks()
45 struct eip93_hash_reqctx *rctx = ahash_request_ctx_dma(req); in eip93_hash_free_sa_record() local
51 dma_unmap_single(eip93->dev, rctx->sa_record_hmac_base, in eip93_hash_free_sa_record()
52 sizeof(rctx->sa_record_hmac), DMA_TO_DEVICE); in eip93_hash_free_sa_record()
[all …]
H A Deip93-cipher.c25 struct eip93_cipher_reqctx *rctx = skcipher_request_ctx(req); in eip93_skcipher_handle_result() local
27 eip93_unmap_dma(eip93, rctx, req->src, req->dst); in eip93_skcipher_handle_result()
28 eip93_handle_result(eip93, rctx, req->iv); in eip93_skcipher_handle_result()
36 struct eip93_cipher_reqctx *rctx = skcipher_request_ctx(req); in eip93_skcipher_send_req() local
39 err = check_valid_request(rctx); in eip93_skcipher_send_req()
46 return eip93_send_req(async, req->iv, rctx); in eip93_skcipher_send_req()
138 struct eip93_cipher_reqctx *rctx = skcipher_request_ctx(req); in eip93_skcipher_crypt() local
151 if (IS_ECB(rctx->flags) || IS_CBC(rctx->flags)) in eip93_skcipher_crypt()
162 rctx->assoclen = 0; in eip93_skcipher_crypt()
163 rctx->textsize = req->cryptlen; in eip93_skcipher_crypt()
[all …]
H A Deip93-aead.c35 struct eip93_cipher_reqctx *rctx = aead_request_ctx(req); in eip93_aead_handle_result() local
37 eip93_unmap_dma(eip93, rctx, req->src, req->dst); in eip93_aead_handle_result()
38 eip93_handle_result(eip93, rctx, req->iv); in eip93_aead_handle_result()
46 struct eip93_cipher_reqctx *rctx = aead_request_ctx(req); in eip93_aead_send_req() local
49 err = check_valid_request(rctx); in eip93_aead_send_req()
55 return eip93_send_req(async, req->iv, rctx); in eip93_aead_send_req()
188 struct eip93_cipher_reqctx *rctx = aead_request_ctx(req); in eip93_aead_crypt() local
200 rctx->textsize = req->cryptlen; in eip93_aead_crypt()
201 rctx->blksize = ctx->blksize; in eip93_aead_crypt()
202 rctx->assoclen = req->assoclen; in eip93_aead_crypt()
[all …]
/linux/drivers/crypto/cavium/nitrox/
H A Dnitrox_aead.c151 static int nitrox_set_creq(struct nitrox_aead_rctx *rctx) in nitrox_set_creq() argument
153 struct se_crypto_request *creq = &rctx->nkreq.creq; in nitrox_set_creq()
157 creq->flags = rctx->flags; in nitrox_set_creq()
158 creq->gfp = (rctx->flags & CRYPTO_TFM_REQ_MAY_SLEEP) ? GFP_KERNEL : in nitrox_set_creq()
163 creq->ctrl.s.arg = rctx->ctrl_arg; in nitrox_set_creq()
165 creq->gph.param0 = cpu_to_be16(rctx->cryptlen); in nitrox_set_creq()
166 creq->gph.param1 = cpu_to_be16(rctx->cryptlen + rctx->assoclen); in nitrox_set_creq()
167 creq->gph.param2 = cpu_to_be16(rctx->ivsize + rctx->assoclen); in nitrox_set_creq()
169 param3.auth_offset = rctx->ivsize; in nitrox_set_creq()
172 creq->ctx_handle = rctx->ctx_handle; in nitrox_set_creq()
[all …]
/linux/drivers/crypto/allwinner/sun8i-ss/
H A Dsun8i-ss-hash.c146 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); in sun8i_ss_hash_init() local
150 memset(rctx, 0, sizeof(struct sun8i_ss_hash_reqctx)); in sun8i_ss_hash_init()
152 ahash_request_set_tfm(&rctx->fallback_req, tfmctx->fallback_tfm); in sun8i_ss_hash_init()
153 ahash_request_set_callback(&rctx->fallback_req, in sun8i_ss_hash_init()
157 return crypto_ahash_init(&rctx->fallback_req); in sun8i_ss_hash_init()
162 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); in sun8i_ss_hash_export() local
166 ahash_request_set_tfm(&rctx->fallback_req, tfmctx->fallback_tfm); in sun8i_ss_hash_export()
167 ahash_request_set_callback(&rctx->fallback_req, in sun8i_ss_hash_export()
171 return crypto_ahash_export(&rctx->fallback_req, out); in sun8i_ss_hash_export()
176 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); in sun8i_ss_hash_import() local
[all …]
/linux/drivers/crypto/bcm/
H A Dcipher.c135 struct iproc_reqctx_s *rctx, in spu_skcipher_rx_sg_create() argument
141 struct iproc_ctx_s *ctx = rctx->ctx; in spu_skcipher_rx_sg_create()
144 mssg->spu.dst = kmalloc_objs(struct scatterlist, rx_frag_num, rctx->gfp); in spu_skcipher_rx_sg_create()
151 sg_set_buf(sg++, rctx->msg_buf.spu_resp_hdr, ctx->spu_resp_hdr_len); in spu_skcipher_rx_sg_create()
156 sg_set_buf(sg++, rctx->msg_buf.c.supdt_tweak, in spu_skcipher_rx_sg_create()
160 datalen = spu_msg_sg_add(&sg, &rctx->dst_sg, &rctx->dst_skip, in spu_skcipher_rx_sg_create()
161 rctx->dst_nents, chunksize); in spu_skcipher_rx_sg_create()
169 sg_set_buf(sg++, rctx->msg_buf.rx_stat_pad, stat_pad_len); in spu_skcipher_rx_sg_create()
171 memset(rctx->msg_buf.rx_stat, 0, SPU_RX_STATUS_LEN); in spu_skcipher_rx_sg_create()
172 sg_set_buf(sg, rctx->msg_buf.rx_stat, spu->spu_rx_status_len()); in spu_skcipher_rx_sg_create()
[all …]
/linux/drivers/crypto/rockchip/
H A Drk3288_crypto_skcipher.c72 struct rk_cipher_rctx *rctx = skcipher_request_ctx(areq); in rk_cipher_fallback() local
79 skcipher_request_set_tfm(&rctx->fallback_req, op->fallback_tfm); in rk_cipher_fallback()
80 skcipher_request_set_callback(&rctx->fallback_req, areq->base.flags, in rk_cipher_fallback()
82 skcipher_request_set_crypt(&rctx->fallback_req, areq->src, areq->dst, in rk_cipher_fallback()
84 if (rctx->mode & RK_CRYPTO_DEC) in rk_cipher_fallback()
85 err = crypto_skcipher_decrypt(&rctx->fallback_req); in rk_cipher_fallback()
87 err = crypto_skcipher_encrypt(&rctx->fallback_req); in rk_cipher_fallback()
93 struct rk_cipher_rctx *rctx = skcipher_request_ctx(req); in rk_cipher_handle_req() local
103 rctx->dev = rkc; in rk_cipher_handle_req()
157 struct rk_cipher_rctx *rctx = skcipher_request_ctx(req); in rk_aes_ecb_encrypt() local
[all …]
/linux/drivers/crypto/starfive/
H A Djh7110-hash.c66 struct starfive_cryp_request_ctx *rctx = ctx->rctx; in starfive_hash_hmac_key() local
74 rctx->csr.hash.hmac = 1; in starfive_hash_hmac_key()
75 rctx->csr.hash.key_flag = 1; in starfive_hash_hmac_key()
77 writel(rctx->csr.hash.v, cryp->base + STARFIVE_HASH_SHACSR); in starfive_hash_hmac_key()
167 struct starfive_cryp_request_ctx *rctx = ahash_request_ctx(req); in starfive_hash_copy_hash() local
175 mlen = rctx->digsize / sizeof(u32); in starfive_hash_copy_hash()
200 struct starfive_cryp_request_ctx *rctx = ctx->rctx; in starfive_hash_one_request() local
210 rctx->csr.hash.v = 0; in starfive_hash_one_request()
211 rctx->csr.hash.mode = ctx->hash_mode; in starfive_hash_one_request()
218 rctx->csr.hash.start = 1; in starfive_hash_one_request()
[all …]
/linux/drivers/crypto/gemini/
H A Dsl3516-ce-cipher.c107 struct sl3516_ce_cipher_req_ctx *rctx = skcipher_request_ctx(areq); in sl3516_ce_cipher_fallback() local
115 skcipher_request_set_tfm(&rctx->fallback_req, op->fallback_tfm); in sl3516_ce_cipher_fallback()
116 skcipher_request_set_callback(&rctx->fallback_req, areq->base.flags, in sl3516_ce_cipher_fallback()
118 skcipher_request_set_crypt(&rctx->fallback_req, areq->src, areq->dst, in sl3516_ce_cipher_fallback()
120 if (rctx->op_dir == CE_DECRYPTION) in sl3516_ce_cipher_fallback()
121 err = crypto_skcipher_decrypt(&rctx->fallback_req); in sl3516_ce_cipher_fallback()
123 err = crypto_skcipher_encrypt(&rctx->fallback_req); in sl3516_ce_cipher_fallback()
132 struct sl3516_ce_cipher_req_ctx *rctx = skcipher_request_ctx(areq); in sl3516_ce_cipher() local
148 rctx->op_dir, areq->iv, crypto_skcipher_ivsize(tfm), in sl3516_ce_cipher()
185 rctx->t_src[i].addr = sg_dma_address(sg); in sl3516_ce_cipher()
[all …]
/linux/crypto/
H A Dhctr2.c138 struct hctr2_request_ctx *rctx = skcipher_request_ctx(req); in hctr2_hash_tweak() local
139 struct polyval_ctx *poly_ctx = &rctx->u.poly_ctx; in hctr2_hash_tweak()
149 polyval_export_blkaligned(poly_ctx, &rctx->hashed_tweak); in hctr2_hash_tweak()
157 struct hctr2_request_ctx *rctx = skcipher_request_ctx(req); in hctr2_hash_message() local
158 struct polyval_ctx *poly_ctx = &rctx->u.poly_ctx; in hctr2_hash_message()
182 struct hctr2_request_ctx *rctx = skcipher_request_ctx(req); in hctr2_finish() local
183 struct polyval_ctx *poly_ctx = &rctx->u.poly_ctx; in hctr2_finish()
189 &rctx->hashed_tweak); in hctr2_finish()
190 hctr2_hash_message(req, rctx->bulk_part_dst, digest); in hctr2_finish()
191 crypto_xor(rctx->first_block, digest, BLOCKCIPHER_BLOCK_SIZE); in hctr2_finish()
[all …]
/linux/drivers/crypto/amlogic/
H A Damlogic-gxl-cipher.c62 struct meson_cipher_req_ctx *rctx = skcipher_request_ctx(areq); in meson_cipher_do_fallback() local
71 skcipher_request_set_tfm(&rctx->fallback_req, op->fallback_tfm); in meson_cipher_do_fallback()
72 skcipher_request_set_callback(&rctx->fallback_req, areq->base.flags, in meson_cipher_do_fallback()
74 skcipher_request_set_crypt(&rctx->fallback_req, areq->src, areq->dst, in meson_cipher_do_fallback()
77 if (rctx->op_dir == MESON_DECRYPT) in meson_cipher_do_fallback()
78 err = crypto_skcipher_decrypt(&rctx->fallback_req); in meson_cipher_do_fallback()
80 err = crypto_skcipher_encrypt(&rctx->fallback_req); in meson_cipher_do_fallback()
88 struct meson_cipher_req_ctx *rctx = skcipher_request_ctx(areq); in meson_cipher() local
92 int flow = rctx->flow; in meson_cipher()
109 rctx->op_dir, crypto_skcipher_ivsize(tfm), in meson_cipher()
[all …]
/linux/drivers/crypto/
H A Domap-sham.c727 struct omap_sham_reqctx *rctx) in omap_sham_align_sgs() argument
734 int offset = rctx->offset; in omap_sham_align_sgs()
735 int bufcnt = rctx->bufcnt; in omap_sham_align_sgs()
740 sg_init_table(rctx->sgl, 1); in omap_sham_align_sgs()
741 sg_set_buf(rctx->sgl, rctx->dd->xmit_buf, bufcnt); in omap_sham_align_sgs()
742 rctx->sg = rctx->sgl; in omap_sham_align_sgs()
743 rctx->sg_len = 1; in omap_sham_align_sgs()
824 return omap_sham_copy_sgs(rctx, sg, bs, new_len); in omap_sham_align_sgs()
826 return omap_sham_copy_sg_lists(rctx, sg, bs, new_len); in omap_sham_align_sgs()
828 rctx->total = new_len; in omap_sham_align_sgs()
[all …]

123