xref: /linux/drivers/crypto/gemini/sl3516-ce-cipher.c (revision eed4edda910fe34dfae8c6bfbcf57f4593a54295)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * sl3516-ce-cipher.c - hardware cryptographic offloader for Storlink SL3516 SoC
4  *
5  * Copyright (C) 2021 Corentin LABBE <clabbe@baylibre.com>
6  *
7  * This file adds support for AES cipher with 128,192,256 bits keysize in
8  * ECB mode.
9  */
10 
11 #include <crypto/engine.h>
12 #include <crypto/internal/skcipher.h>
13 #include <crypto/scatterwalk.h>
14 #include <linux/dma-mapping.h>
15 #include <linux/delay.h>
16 #include <linux/err.h>
17 #include <linux/io.h>
18 #include <linux/kernel.h>
19 #include <linux/pm_runtime.h>
20 #include <linux/slab.h>
21 #include <linux/string.h>
22 #include "sl3516-ce.h"
23 
24 /* sl3516_ce_need_fallback - check if a request can be handled by the CE */
25 static bool sl3516_ce_need_fallback(struct skcipher_request *areq)
26 {
27 	struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq);
28 	struct sl3516_ce_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
29 	struct sl3516_ce_dev *ce = op->ce;
30 	struct scatterlist *in_sg;
31 	struct scatterlist *out_sg;
32 	struct scatterlist *sg;
33 
34 	if (areq->cryptlen == 0 || areq->cryptlen % 16) {
35 		ce->fallback_mod16++;
36 		return true;
37 	}
38 
39 	/*
40 	 * check if we have enough descriptors for TX
41 	 * Note: TX need one control desc for each SG
42 	 */
43 	if (sg_nents(areq->src) > MAXDESC / 2) {
44 		ce->fallback_sg_count_tx++;
45 		return true;
46 	}
47 	/* check if we have enough descriptors for RX */
48 	if (sg_nents(areq->dst) > MAXDESC) {
49 		ce->fallback_sg_count_rx++;
50 		return true;
51 	}
52 
53 	sg = areq->src;
54 	while (sg) {
55 		if ((sg->length % 16) != 0) {
56 			ce->fallback_mod16++;
57 			return true;
58 		}
59 		if ((sg_dma_len(sg) % 16) != 0) {
60 			ce->fallback_mod16++;
61 			return true;
62 		}
63 		if (!IS_ALIGNED(sg->offset, 16)) {
64 			ce->fallback_align16++;
65 			return true;
66 		}
67 		sg = sg_next(sg);
68 	}
69 	sg = areq->dst;
70 	while (sg) {
71 		if ((sg->length % 16) != 0) {
72 			ce->fallback_mod16++;
73 			return true;
74 		}
75 		if ((sg_dma_len(sg) % 16) != 0) {
76 			ce->fallback_mod16++;
77 			return true;
78 		}
79 		if (!IS_ALIGNED(sg->offset, 16)) {
80 			ce->fallback_align16++;
81 			return true;
82 		}
83 		sg = sg_next(sg);
84 	}
85 
86 	/* need same numbers of SG (with same length) for source and destination */
87 	in_sg = areq->src;
88 	out_sg = areq->dst;
89 	while (in_sg && out_sg) {
90 		if (in_sg->length != out_sg->length) {
91 			ce->fallback_not_same_len++;
92 			return true;
93 		}
94 		in_sg = sg_next(in_sg);
95 		out_sg = sg_next(out_sg);
96 	}
97 	if (in_sg || out_sg)
98 		return true;
99 
100 	return false;
101 }
102 
103 static int sl3516_ce_cipher_fallback(struct skcipher_request *areq)
104 {
105 	struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq);
106 	struct sl3516_ce_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
107 	struct sl3516_ce_cipher_req_ctx *rctx = skcipher_request_ctx(areq);
108 	struct skcipher_alg *alg = crypto_skcipher_alg(tfm);
109 	struct sl3516_ce_alg_template *algt;
110 	int err;
111 
112 	algt = container_of(alg, struct sl3516_ce_alg_template, alg.skcipher.base);
113 	algt->stat_fb++;
114 
115 	skcipher_request_set_tfm(&rctx->fallback_req, op->fallback_tfm);
116 	skcipher_request_set_callback(&rctx->fallback_req, areq->base.flags,
117 				      areq->base.complete, areq->base.data);
118 	skcipher_request_set_crypt(&rctx->fallback_req, areq->src, areq->dst,
119 				   areq->cryptlen, areq->iv);
120 	if (rctx->op_dir == CE_DECRYPTION)
121 		err = crypto_skcipher_decrypt(&rctx->fallback_req);
122 	else
123 		err = crypto_skcipher_encrypt(&rctx->fallback_req);
124 	return err;
125 }
126 
127 static int sl3516_ce_cipher(struct skcipher_request *areq)
128 {
129 	struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq);
130 	struct sl3516_ce_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
131 	struct sl3516_ce_dev *ce = op->ce;
132 	struct sl3516_ce_cipher_req_ctx *rctx = skcipher_request_ctx(areq);
133 	struct skcipher_alg *alg = crypto_skcipher_alg(tfm);
134 	struct sl3516_ce_alg_template *algt;
135 	struct scatterlist *sg;
136 	unsigned int todo, len;
137 	struct pkt_control_ecb *ecb;
138 	int nr_sgs = 0;
139 	int nr_sgd = 0;
140 	int err = 0;
141 	int i;
142 
143 	algt = container_of(alg, struct sl3516_ce_alg_template, alg.skcipher.base);
144 
145 	dev_dbg(ce->dev, "%s %s %u %x IV(%p %u) key=%u\n", __func__,
146 		crypto_tfm_alg_name(areq->base.tfm),
147 		areq->cryptlen,
148 		rctx->op_dir, areq->iv, crypto_skcipher_ivsize(tfm),
149 		op->keylen);
150 
151 	algt->stat_req++;
152 
153 	if (areq->src == areq->dst) {
154 		nr_sgs = dma_map_sg(ce->dev, areq->src, sg_nents(areq->src),
155 				    DMA_BIDIRECTIONAL);
156 		if (nr_sgs <= 0 || nr_sgs > MAXDESC / 2) {
157 			dev_err(ce->dev, "Invalid sg number %d\n", nr_sgs);
158 			err = -EINVAL;
159 			goto theend;
160 		}
161 		nr_sgd = nr_sgs;
162 	} else {
163 		nr_sgs = dma_map_sg(ce->dev, areq->src, sg_nents(areq->src),
164 				    DMA_TO_DEVICE);
165 		if (nr_sgs <= 0 || nr_sgs > MAXDESC / 2) {
166 			dev_err(ce->dev, "Invalid sg number %d\n", nr_sgs);
167 			err = -EINVAL;
168 			goto theend;
169 		}
170 		nr_sgd = dma_map_sg(ce->dev, areq->dst, sg_nents(areq->dst),
171 				    DMA_FROM_DEVICE);
172 		if (nr_sgd <= 0 || nr_sgd > MAXDESC) {
173 			dev_err(ce->dev, "Invalid sg number %d\n", nr_sgd);
174 			err = -EINVAL;
175 			goto theend_sgs;
176 		}
177 	}
178 
179 	len = areq->cryptlen;
180 	i = 0;
181 	sg = areq->src;
182 	while (i < nr_sgs && sg && len) {
183 		if (sg_dma_len(sg) == 0)
184 			goto sgs_next;
185 		rctx->t_src[i].addr = sg_dma_address(sg);
186 		todo = min(len, sg_dma_len(sg));
187 		rctx->t_src[i].len = todo;
188 		dev_dbg(ce->dev, "%s total=%u SGS(%d %u off=%d) todo=%u\n", __func__,
189 			areq->cryptlen, i, rctx->t_src[i].len, sg->offset, todo);
190 		len -= todo;
191 		i++;
192 sgs_next:
193 		sg = sg_next(sg);
194 	}
195 	if (len > 0) {
196 		dev_err(ce->dev, "remaining len %d/%u nr_sgs=%d\n", len, areq->cryptlen, nr_sgs);
197 		err = -EINVAL;
198 		goto theend_sgs;
199 	}
200 
201 	len = areq->cryptlen;
202 	i = 0;
203 	sg = areq->dst;
204 	while (i < nr_sgd && sg && len) {
205 		if (sg_dma_len(sg) == 0)
206 			goto sgd_next;
207 		rctx->t_dst[i].addr = sg_dma_address(sg);
208 		todo = min(len, sg_dma_len(sg));
209 		rctx->t_dst[i].len = todo;
210 		dev_dbg(ce->dev, "%s total=%u SGD(%d %u off=%d) todo=%u\n", __func__,
211 			areq->cryptlen, i, rctx->t_dst[i].len, sg->offset, todo);
212 		len -= todo;
213 		i++;
214 
215 sgd_next:
216 		sg = sg_next(sg);
217 	}
218 	if (len > 0) {
219 		dev_err(ce->dev, "remaining len %d\n", len);
220 		err = -EINVAL;
221 		goto theend_sgs;
222 	}
223 
224 	switch (algt->mode) {
225 	case ECB_AES:
226 		rctx->pctrllen = sizeof(struct pkt_control_ecb);
227 		ecb = (struct pkt_control_ecb *)ce->pctrl;
228 
229 		rctx->tqflag = TQ0_TYPE_CTRL;
230 		rctx->tqflag |= TQ1_CIPHER;
231 		ecb->control.op_mode = rctx->op_dir;
232 		ecb->control.cipher_algorithm = ECB_AES;
233 		ecb->cipher.header_len = 0;
234 		ecb->cipher.algorithm_len = areq->cryptlen;
235 		cpu_to_be32_array((__be32 *)ecb->key, (u32 *)op->key, op->keylen / 4);
236 		rctx->h = &ecb->cipher;
237 
238 		rctx->tqflag |= TQ4_KEY0;
239 		rctx->tqflag |= TQ5_KEY4;
240 		rctx->tqflag |= TQ6_KEY6;
241 		ecb->control.aesnk = op->keylen / 4;
242 		break;
243 	}
244 
245 	rctx->nr_sgs = nr_sgs;
246 	rctx->nr_sgd = nr_sgd;
247 	err = sl3516_ce_run_task(ce, rctx, crypto_tfm_alg_name(areq->base.tfm));
248 
249 theend_sgs:
250 	if (areq->src == areq->dst) {
251 		dma_unmap_sg(ce->dev, areq->src, sg_nents(areq->src),
252 			     DMA_BIDIRECTIONAL);
253 	} else {
254 		dma_unmap_sg(ce->dev, areq->src, sg_nents(areq->src),
255 			     DMA_TO_DEVICE);
256 		dma_unmap_sg(ce->dev, areq->dst, sg_nents(areq->dst),
257 			     DMA_FROM_DEVICE);
258 	}
259 
260 theend:
261 
262 	return err;
263 }
264 
265 int sl3516_ce_handle_cipher_request(struct crypto_engine *engine, void *areq)
266 {
267 	int err;
268 	struct skcipher_request *breq = container_of(areq, struct skcipher_request, base);
269 
270 	err = sl3516_ce_cipher(breq);
271 	local_bh_disable();
272 	crypto_finalize_skcipher_request(engine, breq, err);
273 	local_bh_enable();
274 
275 	return 0;
276 }
277 
278 int sl3516_ce_skdecrypt(struct skcipher_request *areq)
279 {
280 	struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq);
281 	struct sl3516_ce_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
282 	struct sl3516_ce_cipher_req_ctx *rctx = skcipher_request_ctx(areq);
283 	struct crypto_engine *engine;
284 
285 	memset(rctx, 0, sizeof(struct sl3516_ce_cipher_req_ctx));
286 	rctx->op_dir = CE_DECRYPTION;
287 
288 	if (sl3516_ce_need_fallback(areq))
289 		return sl3516_ce_cipher_fallback(areq);
290 
291 	engine = op->ce->engine;
292 
293 	return crypto_transfer_skcipher_request_to_engine(engine, areq);
294 }
295 
296 int sl3516_ce_skencrypt(struct skcipher_request *areq)
297 {
298 	struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq);
299 	struct sl3516_ce_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
300 	struct sl3516_ce_cipher_req_ctx *rctx = skcipher_request_ctx(areq);
301 	struct crypto_engine *engine;
302 
303 	memset(rctx, 0, sizeof(struct sl3516_ce_cipher_req_ctx));
304 	rctx->op_dir = CE_ENCRYPTION;
305 
306 	if (sl3516_ce_need_fallback(areq))
307 		return sl3516_ce_cipher_fallback(areq);
308 
309 	engine = op->ce->engine;
310 
311 	return crypto_transfer_skcipher_request_to_engine(engine, areq);
312 }
313 
314 int sl3516_ce_cipher_init(struct crypto_tfm *tfm)
315 {
316 	struct sl3516_ce_cipher_tfm_ctx *op = crypto_tfm_ctx(tfm);
317 	struct sl3516_ce_alg_template *algt;
318 	const char *name = crypto_tfm_alg_name(tfm);
319 	struct crypto_skcipher *sktfm = __crypto_skcipher_cast(tfm);
320 	struct skcipher_alg *alg = crypto_skcipher_alg(sktfm);
321 	int err;
322 
323 	memset(op, 0, sizeof(struct sl3516_ce_cipher_tfm_ctx));
324 
325 	algt = container_of(alg, struct sl3516_ce_alg_template, alg.skcipher.base);
326 	op->ce = algt->ce;
327 
328 	op->fallback_tfm = crypto_alloc_skcipher(name, 0, CRYPTO_ALG_NEED_FALLBACK);
329 	if (IS_ERR(op->fallback_tfm)) {
330 		dev_err(op->ce->dev, "ERROR: Cannot allocate fallback for %s %ld\n",
331 			name, PTR_ERR(op->fallback_tfm));
332 		return PTR_ERR(op->fallback_tfm);
333 	}
334 
335 	crypto_skcipher_set_reqsize(sktfm, sizeof(struct sl3516_ce_cipher_req_ctx) +
336 				    crypto_skcipher_reqsize(op->fallback_tfm));
337 
338 	dev_info(op->ce->dev, "Fallback for %s is %s\n",
339 		 crypto_tfm_alg_driver_name(&sktfm->base),
340 		 crypto_tfm_alg_driver_name(crypto_skcipher_tfm(op->fallback_tfm)));
341 
342 	err = pm_runtime_get_sync(op->ce->dev);
343 	if (err < 0)
344 		goto error_pm;
345 
346 	return 0;
347 error_pm:
348 	pm_runtime_put_noidle(op->ce->dev);
349 	crypto_free_skcipher(op->fallback_tfm);
350 	return err;
351 }
352 
353 void sl3516_ce_cipher_exit(struct crypto_tfm *tfm)
354 {
355 	struct sl3516_ce_cipher_tfm_ctx *op = crypto_tfm_ctx(tfm);
356 
357 	kfree_sensitive(op->key);
358 	crypto_free_skcipher(op->fallback_tfm);
359 	pm_runtime_put_sync_suspend(op->ce->dev);
360 }
361 
362 int sl3516_ce_aes_setkey(struct crypto_skcipher *tfm, const u8 *key,
363 			 unsigned int keylen)
364 {
365 	struct sl3516_ce_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
366 	struct sl3516_ce_dev *ce = op->ce;
367 
368 	switch (keylen) {
369 	case 128 / 8:
370 		break;
371 	case 192 / 8:
372 		break;
373 	case 256 / 8:
374 		break;
375 	default:
376 		dev_dbg(ce->dev, "ERROR: Invalid keylen %u\n", keylen);
377 		return -EINVAL;
378 	}
379 	kfree_sensitive(op->key);
380 	op->keylen = keylen;
381 	op->key = kmemdup(key, keylen, GFP_KERNEL | GFP_DMA);
382 	if (!op->key)
383 		return -ENOMEM;
384 
385 	crypto_skcipher_clear_flags(op->fallback_tfm, CRYPTO_TFM_REQ_MASK);
386 	crypto_skcipher_set_flags(op->fallback_tfm, tfm->base.crt_flags & CRYPTO_TFM_REQ_MASK);
387 
388 	return crypto_skcipher_setkey(op->fallback_tfm, key, keylen);
389 }
390