xref: /linux/drivers/crypto/chelsio/chcr_algo.c (revision 906fd46a65383cd639e5eec72a047efc33045d86)
1 /*
2  * This file is part of the Chelsio T6 Crypto driver for Linux.
3  *
4  * Copyright (c) 2003-2016 Chelsio Communications, Inc. All rights reserved.
5  *
6  * This software is available to you under a choice of one of two
7  * licenses.  You may choose to be licensed under the terms of the GNU
8  * General Public License (GPL) Version 2, available from the file
9  * COPYING in the main directory of this source tree, or the
10  * OpenIB.org BSD license below:
11  *
12  *     Redistribution and use in source and binary forms, with or
13  *     without modification, are permitted provided that the following
14  *     conditions are met:
15  *
16  *      - Redistributions of source code must retain the above
17  *        copyright notice, this list of conditions and the following
18  *        disclaimer.
19  *
20  *      - Redistributions in binary form must reproduce the above
21  *        copyright notice, this list of conditions and the following
22  *        disclaimer in the documentation and/or other materials
23  *        provided with the distribution.
24  *
25  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
26  * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
27  * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
28  * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
29  * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
30  * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
31  * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
32  * SOFTWARE.
33  *
34  * Written and Maintained by:
35  *	Manoj Malviya (manojmalviya@chelsio.com)
36  *	Atul Gupta (atul.gupta@chelsio.com)
37  *	Jitendra Lulla (jlulla@chelsio.com)
38  *	Yeshaswi M R Gowda (yeshaswi@chelsio.com)
39  *	Harsh Jain (harsh@chelsio.com)
40  */
41 
42 #define pr_fmt(fmt) "chcr:" fmt
43 
44 #include <linux/kernel.h>
45 #include <linux/module.h>
46 #include <linux/crypto.h>
47 #include <linux/skbuff.h>
48 #include <linux/rtnetlink.h>
49 #include <linux/highmem.h>
50 #include <linux/scatterlist.h>
51 
52 #include <crypto/aes.h>
53 #include <crypto/algapi.h>
54 #include <crypto/hash.h>
55 #include <crypto/gcm.h>
56 #include <crypto/sha1.h>
57 #include <crypto/sha2.h>
58 #include <crypto/authenc.h>
59 #include <crypto/ctr.h>
60 #include <crypto/gf128mul.h>
61 #include <crypto/internal/aead.h>
62 #include <crypto/null.h>
63 #include <crypto/internal/skcipher.h>
64 #include <crypto/aead.h>
65 #include <crypto/scatterwalk.h>
66 #include <crypto/internal/hash.h>
67 
68 #include "t4fw_api.h"
69 #include "t4_msg.h"
70 #include "chcr_core.h"
71 #include "chcr_algo.h"
72 #include "chcr_crypto.h"
73 
74 #define IV AES_BLOCK_SIZE
75 
76 static unsigned int sgl_ent_len[] = {
77 	0, 0, 16, 24, 40, 48, 64, 72, 88,
78 	96, 112, 120, 136, 144, 160, 168, 184,
79 	192, 208, 216, 232, 240, 256, 264, 280,
80 	288, 304, 312, 328, 336, 352, 360, 376
81 };
82 
83 static unsigned int dsgl_ent_len[] = {
84 	0, 32, 32, 48, 48, 64, 64, 80, 80,
85 	112, 112, 128, 128, 144, 144, 160, 160,
86 	192, 192, 208, 208, 224, 224, 240, 240,
87 	272, 272, 288, 288, 304, 304, 320, 320
88 };
89 
90 static u32 round_constant[11] = {
91 	0x01000000, 0x02000000, 0x04000000, 0x08000000,
92 	0x10000000, 0x20000000, 0x40000000, 0x80000000,
93 	0x1B000000, 0x36000000, 0x6C000000
94 };
95 
96 static int chcr_handle_cipher_resp(struct skcipher_request *req,
97 				   unsigned char *input, int err);
98 
99 static inline  struct chcr_aead_ctx *AEAD_CTX(struct chcr_context *ctx)
100 {
101 	return &ctx->crypto_ctx->aeadctx;
102 }
103 
104 static inline struct ablk_ctx *ABLK_CTX(struct chcr_context *ctx)
105 {
106 	return &ctx->crypto_ctx->ablkctx;
107 }
108 
109 static inline struct hmac_ctx *HMAC_CTX(struct chcr_context *ctx)
110 {
111 	return &ctx->crypto_ctx->hmacctx;
112 }
113 
114 static inline struct chcr_gcm_ctx *GCM_CTX(struct chcr_aead_ctx *gctx)
115 {
116 	return gctx->ctx->gcm;
117 }
118 
119 static inline struct chcr_authenc_ctx *AUTHENC_CTX(struct chcr_aead_ctx *gctx)
120 {
121 	return gctx->ctx->authenc;
122 }
123 
124 static inline struct uld_ctx *ULD_CTX(struct chcr_context *ctx)
125 {
126 	return container_of(ctx->dev, struct uld_ctx, dev);
127 }
128 
129 static inline void chcr_init_hctx_per_wr(struct chcr_ahash_req_ctx *reqctx)
130 {
131 	memset(&reqctx->hctx_wr, 0, sizeof(struct chcr_hctx_per_wr));
132 }
133 
134 static int sg_nents_xlen(struct scatterlist *sg, unsigned int reqlen,
135 			 unsigned int entlen,
136 			 unsigned int skip)
137 {
138 	int nents = 0;
139 	unsigned int less;
140 	unsigned int skip_len = 0;
141 
142 	while (sg && skip) {
143 		if (sg_dma_len(sg) <= skip) {
144 			skip -= sg_dma_len(sg);
145 			skip_len = 0;
146 			sg = sg_next(sg);
147 		} else {
148 			skip_len = skip;
149 			skip = 0;
150 		}
151 	}
152 
153 	while (sg && reqlen) {
154 		less = min(reqlen, sg_dma_len(sg) - skip_len);
155 		nents += DIV_ROUND_UP(less, entlen);
156 		reqlen -= less;
157 		skip_len = 0;
158 		sg = sg_next(sg);
159 	}
160 	return nents;
161 }
162 
163 static inline int get_aead_subtype(struct crypto_aead *aead)
164 {
165 	struct aead_alg *alg = crypto_aead_alg(aead);
166 	struct chcr_alg_template *chcr_crypto_alg =
167 		container_of(alg, struct chcr_alg_template, alg.aead);
168 	return chcr_crypto_alg->type & CRYPTO_ALG_SUB_TYPE_MASK;
169 }
170 
171 void chcr_verify_tag(struct aead_request *req, u8 *input, int *err)
172 {
173 	u8 temp[SHA512_DIGEST_SIZE];
174 	struct crypto_aead *tfm = crypto_aead_reqtfm(req);
175 	int authsize = crypto_aead_authsize(tfm);
176 	struct cpl_fw6_pld *fw6_pld;
177 	int cmp = 0;
178 
179 	fw6_pld = (struct cpl_fw6_pld *)input;
180 	if ((get_aead_subtype(tfm) == CRYPTO_ALG_SUB_TYPE_AEAD_RFC4106) ||
181 	    (get_aead_subtype(tfm) == CRYPTO_ALG_SUB_TYPE_AEAD_GCM)) {
182 		cmp = crypto_memneq(&fw6_pld->data[2], (fw6_pld + 1), authsize);
183 	} else {
184 
185 		sg_pcopy_to_buffer(req->src, sg_nents(req->src), temp,
186 				authsize, req->assoclen +
187 				req->cryptlen - authsize);
188 		cmp = crypto_memneq(temp, (fw6_pld + 1), authsize);
189 	}
190 	if (cmp)
191 		*err = -EBADMSG;
192 	else
193 		*err = 0;
194 }
195 
196 static int chcr_inc_wrcount(struct chcr_dev *dev)
197 {
198 	if (dev->state == CHCR_DETACH)
199 		return 1;
200 	atomic_inc(&dev->inflight);
201 	return 0;
202 }
203 
204 static inline void chcr_dec_wrcount(struct chcr_dev *dev)
205 {
206 	atomic_dec(&dev->inflight);
207 }
208 
209 static inline int chcr_handle_aead_resp(struct aead_request *req,
210 					 unsigned char *input,
211 					 int err)
212 {
213 	struct chcr_aead_reqctx *reqctx = aead_request_ctx_dma(req);
214 	struct crypto_aead *tfm = crypto_aead_reqtfm(req);
215 	struct chcr_dev *dev = a_ctx(tfm)->dev;
216 
217 	chcr_aead_common_exit(req);
218 	if (reqctx->verify == VERIFY_SW) {
219 		chcr_verify_tag(req, input, &err);
220 		reqctx->verify = VERIFY_HW;
221 	}
222 	chcr_dec_wrcount(dev);
223 	aead_request_complete(req, err);
224 
225 	return err;
226 }
227 
228 static void get_aes_decrypt_key(unsigned char *dec_key,
229 				       const unsigned char *key,
230 				       unsigned int keylength)
231 {
232 	u32 temp;
233 	u32 w_ring[MAX_NK];
234 	int i, j, k;
235 	u8  nr, nk;
236 
237 	switch (keylength) {
238 	case AES_KEYLENGTH_128BIT:
239 		nk = KEYLENGTH_4BYTES;
240 		nr = NUMBER_OF_ROUNDS_10;
241 		break;
242 	case AES_KEYLENGTH_192BIT:
243 		nk = KEYLENGTH_6BYTES;
244 		nr = NUMBER_OF_ROUNDS_12;
245 		break;
246 	case AES_KEYLENGTH_256BIT:
247 		nk = KEYLENGTH_8BYTES;
248 		nr = NUMBER_OF_ROUNDS_14;
249 		break;
250 	default:
251 		return;
252 	}
253 	for (i = 0; i < nk; i++)
254 		w_ring[i] = get_unaligned_be32(&key[i * 4]);
255 
256 	i = 0;
257 	temp = w_ring[nk - 1];
258 	while (i + nk < (nr + 1) * 4) {
259 		if (!(i % nk)) {
260 			/* RotWord(temp) */
261 			temp = (temp << 8) | (temp >> 24);
262 			temp = aes_ks_subword(temp);
263 			temp ^= round_constant[i / nk];
264 		} else if (nk == 8 && (i % 4 == 0)) {
265 			temp = aes_ks_subword(temp);
266 		}
267 		w_ring[i % nk] ^= temp;
268 		temp = w_ring[i % nk];
269 		i++;
270 	}
271 	i--;
272 	for (k = 0, j = i % nk; k < nk; k++) {
273 		put_unaligned_be32(w_ring[j], &dec_key[k * 4]);
274 		j--;
275 		if (j < 0)
276 			j += nk;
277 	}
278 }
279 
280 static struct crypto_shash *chcr_alloc_shash(unsigned int ds)
281 {
282 	struct crypto_shash *base_hash = ERR_PTR(-EINVAL);
283 
284 	switch (ds) {
285 	case SHA1_DIGEST_SIZE:
286 		base_hash = crypto_alloc_shash("sha1", 0, 0);
287 		break;
288 	case SHA224_DIGEST_SIZE:
289 		base_hash = crypto_alloc_shash("sha224", 0, 0);
290 		break;
291 	case SHA256_DIGEST_SIZE:
292 		base_hash = crypto_alloc_shash("sha256", 0, 0);
293 		break;
294 	case SHA384_DIGEST_SIZE:
295 		base_hash = crypto_alloc_shash("sha384", 0, 0);
296 		break;
297 	case SHA512_DIGEST_SIZE:
298 		base_hash = crypto_alloc_shash("sha512", 0, 0);
299 		break;
300 	}
301 
302 	return base_hash;
303 }
304 
305 static int chcr_compute_partial_hash(struct shash_desc *desc,
306 				     char *iopad, char *result_hash,
307 				     int digest_size)
308 {
309 	struct sha1_state sha1_st;
310 	struct sha256_state sha256_st;
311 	struct sha512_state sha512_st;
312 	int error;
313 
314 	if (digest_size == SHA1_DIGEST_SIZE) {
315 		error = crypto_shash_init(desc) ?:
316 			crypto_shash_update(desc, iopad, SHA1_BLOCK_SIZE) ?:
317 			crypto_shash_export(desc, (void *)&sha1_st);
318 		memcpy(result_hash, sha1_st.state, SHA1_DIGEST_SIZE);
319 	} else if (digest_size == SHA224_DIGEST_SIZE) {
320 		error = crypto_shash_init(desc) ?:
321 			crypto_shash_update(desc, iopad, SHA256_BLOCK_SIZE) ?:
322 			crypto_shash_export(desc, (void *)&sha256_st);
323 		memcpy(result_hash, sha256_st.state, SHA256_DIGEST_SIZE);
324 
325 	} else if (digest_size == SHA256_DIGEST_SIZE) {
326 		error = crypto_shash_init(desc) ?:
327 			crypto_shash_update(desc, iopad, SHA256_BLOCK_SIZE) ?:
328 			crypto_shash_export(desc, (void *)&sha256_st);
329 		memcpy(result_hash, sha256_st.state, SHA256_DIGEST_SIZE);
330 
331 	} else if (digest_size == SHA384_DIGEST_SIZE) {
332 		error = crypto_shash_init(desc) ?:
333 			crypto_shash_update(desc, iopad, SHA512_BLOCK_SIZE) ?:
334 			crypto_shash_export(desc, (void *)&sha512_st);
335 		memcpy(result_hash, sha512_st.state, SHA512_DIGEST_SIZE);
336 
337 	} else if (digest_size == SHA512_DIGEST_SIZE) {
338 		error = crypto_shash_init(desc) ?:
339 			crypto_shash_update(desc, iopad, SHA512_BLOCK_SIZE) ?:
340 			crypto_shash_export(desc, (void *)&sha512_st);
341 		memcpy(result_hash, sha512_st.state, SHA512_DIGEST_SIZE);
342 	} else {
343 		error = -EINVAL;
344 		pr_err("Unknown digest size %d\n", digest_size);
345 	}
346 	return error;
347 }
348 
349 static void chcr_change_order(char *buf, int ds)
350 {
351 	int i;
352 
353 	if (ds == SHA512_DIGEST_SIZE) {
354 		for (i = 0; i < (ds / sizeof(u64)); i++)
355 			*((__be64 *)buf + i) =
356 				cpu_to_be64(*((u64 *)buf + i));
357 	} else {
358 		for (i = 0; i < (ds / sizeof(u32)); i++)
359 			*((__be32 *)buf + i) =
360 				cpu_to_be32(*((u32 *)buf + i));
361 	}
362 }
363 
364 static inline int is_hmac(struct crypto_tfm *tfm)
365 {
366 	struct crypto_alg *alg = tfm->__crt_alg;
367 	struct chcr_alg_template *chcr_crypto_alg =
368 		container_of(__crypto_ahash_alg(alg), struct chcr_alg_template,
369 			     alg.hash);
370 	if (chcr_crypto_alg->type == CRYPTO_ALG_TYPE_HMAC)
371 		return 1;
372 	return 0;
373 }
374 
375 static inline void dsgl_walk_init(struct dsgl_walk *walk,
376 				   struct cpl_rx_phys_dsgl *dsgl)
377 {
378 	walk->dsgl = dsgl;
379 	walk->nents = 0;
380 	walk->to = (struct phys_sge_pairs *)(dsgl + 1);
381 }
382 
383 static inline void dsgl_walk_end(struct dsgl_walk *walk, unsigned short qid,
384 				 int pci_chan_id)
385 {
386 	struct cpl_rx_phys_dsgl *phys_cpl;
387 
388 	phys_cpl = walk->dsgl;
389 
390 	phys_cpl->op_to_tid = htonl(CPL_RX_PHYS_DSGL_OPCODE_V(CPL_RX_PHYS_DSGL)
391 				    | CPL_RX_PHYS_DSGL_ISRDMA_V(0));
392 	phys_cpl->pcirlxorder_to_noofsgentr =
393 		htonl(CPL_RX_PHYS_DSGL_PCIRLXORDER_V(0) |
394 		      CPL_RX_PHYS_DSGL_PCINOSNOOP_V(0) |
395 		      CPL_RX_PHYS_DSGL_PCITPHNTENB_V(0) |
396 		      CPL_RX_PHYS_DSGL_PCITPHNT_V(0) |
397 		      CPL_RX_PHYS_DSGL_DCAID_V(0) |
398 		      CPL_RX_PHYS_DSGL_NOOFSGENTR_V(walk->nents));
399 	phys_cpl->rss_hdr_int.opcode = CPL_RX_PHYS_ADDR;
400 	phys_cpl->rss_hdr_int.qid = htons(qid);
401 	phys_cpl->rss_hdr_int.hash_val = 0;
402 	phys_cpl->rss_hdr_int.channel = pci_chan_id;
403 }
404 
405 static inline void dsgl_walk_add_page(struct dsgl_walk *walk,
406 					size_t size,
407 					dma_addr_t addr)
408 {
409 	int j;
410 
411 	if (!size)
412 		return;
413 	j = walk->nents;
414 	walk->to->len[j % 8] = htons(size);
415 	walk->to->addr[j % 8] = cpu_to_be64(addr);
416 	j++;
417 	if ((j % 8) == 0)
418 		walk->to++;
419 	walk->nents = j;
420 }
421 
422 static void  dsgl_walk_add_sg(struct dsgl_walk *walk,
423 			   struct scatterlist *sg,
424 			      unsigned int slen,
425 			      unsigned int skip)
426 {
427 	int skip_len = 0;
428 	unsigned int left_size = slen, len = 0;
429 	unsigned int j = walk->nents;
430 	int offset, ent_len;
431 
432 	if (!slen)
433 		return;
434 	while (sg && skip) {
435 		if (sg_dma_len(sg) <= skip) {
436 			skip -= sg_dma_len(sg);
437 			skip_len = 0;
438 			sg = sg_next(sg);
439 		} else {
440 			skip_len = skip;
441 			skip = 0;
442 		}
443 	}
444 
445 	while (left_size && sg) {
446 		len = min_t(u32, left_size, sg_dma_len(sg) - skip_len);
447 		offset = 0;
448 		while (len) {
449 			ent_len =  min_t(u32, len, CHCR_DST_SG_SIZE);
450 			walk->to->len[j % 8] = htons(ent_len);
451 			walk->to->addr[j % 8] = cpu_to_be64(sg_dma_address(sg) +
452 						      offset + skip_len);
453 			offset += ent_len;
454 			len -= ent_len;
455 			j++;
456 			if ((j % 8) == 0)
457 				walk->to++;
458 		}
459 		walk->last_sg = sg;
460 		walk->last_sg_len = min_t(u32, left_size, sg_dma_len(sg) -
461 					  skip_len) + skip_len;
462 		left_size -= min_t(u32, left_size, sg_dma_len(sg) - skip_len);
463 		skip_len = 0;
464 		sg = sg_next(sg);
465 	}
466 	walk->nents = j;
467 }
468 
469 static inline void ulptx_walk_init(struct ulptx_walk *walk,
470 				   struct ulptx_sgl *ulp)
471 {
472 	walk->sgl = ulp;
473 	walk->nents = 0;
474 	walk->pair_idx = 0;
475 	walk->pair = ulp->sge;
476 	walk->last_sg = NULL;
477 	walk->last_sg_len = 0;
478 }
479 
480 static inline void ulptx_walk_end(struct ulptx_walk *walk)
481 {
482 	walk->sgl->cmd_nsge = htonl(ULPTX_CMD_V(ULP_TX_SC_DSGL) |
483 			      ULPTX_NSGE_V(walk->nents));
484 }
485 
486 
487 static inline void ulptx_walk_add_page(struct ulptx_walk *walk,
488 					size_t size,
489 					dma_addr_t addr)
490 {
491 	if (!size)
492 		return;
493 
494 	if (walk->nents == 0) {
495 		walk->sgl->len0 = cpu_to_be32(size);
496 		walk->sgl->addr0 = cpu_to_be64(addr);
497 	} else {
498 		walk->pair->addr[walk->pair_idx] = cpu_to_be64(addr);
499 		walk->pair->len[walk->pair_idx] = cpu_to_be32(size);
500 		walk->pair_idx = !walk->pair_idx;
501 		if (!walk->pair_idx)
502 			walk->pair++;
503 	}
504 	walk->nents++;
505 }
506 
507 static void  ulptx_walk_add_sg(struct ulptx_walk *walk,
508 					struct scatterlist *sg,
509 			       unsigned int len,
510 			       unsigned int skip)
511 {
512 	int small;
513 	int skip_len = 0;
514 	unsigned int sgmin;
515 
516 	if (!len)
517 		return;
518 	while (sg && skip) {
519 		if (sg_dma_len(sg) <= skip) {
520 			skip -= sg_dma_len(sg);
521 			skip_len = 0;
522 			sg = sg_next(sg);
523 		} else {
524 			skip_len = skip;
525 			skip = 0;
526 		}
527 	}
528 	WARN(!sg, "SG should not be null here\n");
529 	if (sg && (walk->nents == 0)) {
530 		small = min_t(unsigned int, sg_dma_len(sg) - skip_len, len);
531 		sgmin = min_t(unsigned int, small, CHCR_SRC_SG_SIZE);
532 		walk->sgl->len0 = cpu_to_be32(sgmin);
533 		walk->sgl->addr0 = cpu_to_be64(sg_dma_address(sg) + skip_len);
534 		walk->nents++;
535 		len -= sgmin;
536 		walk->last_sg = sg;
537 		walk->last_sg_len = sgmin + skip_len;
538 		skip_len += sgmin;
539 		if (sg_dma_len(sg) == skip_len) {
540 			sg = sg_next(sg);
541 			skip_len = 0;
542 		}
543 	}
544 
545 	while (sg && len) {
546 		small = min(sg_dma_len(sg) - skip_len, len);
547 		sgmin = min_t(unsigned int, small, CHCR_SRC_SG_SIZE);
548 		walk->pair->len[walk->pair_idx] = cpu_to_be32(sgmin);
549 		walk->pair->addr[walk->pair_idx] =
550 			cpu_to_be64(sg_dma_address(sg) + skip_len);
551 		walk->pair_idx = !walk->pair_idx;
552 		walk->nents++;
553 		if (!walk->pair_idx)
554 			walk->pair++;
555 		len -= sgmin;
556 		skip_len += sgmin;
557 		walk->last_sg = sg;
558 		walk->last_sg_len = skip_len;
559 		if (sg_dma_len(sg) == skip_len) {
560 			sg = sg_next(sg);
561 			skip_len = 0;
562 		}
563 	}
564 }
565 
566 static inline int get_cryptoalg_subtype(struct crypto_skcipher *tfm)
567 {
568 	struct skcipher_alg *alg = crypto_skcipher_alg(tfm);
569 	struct chcr_alg_template *chcr_crypto_alg =
570 		container_of(alg, struct chcr_alg_template, alg.skcipher);
571 
572 	return chcr_crypto_alg->type & CRYPTO_ALG_SUB_TYPE_MASK;
573 }
574 
575 static int cxgb4_is_crypto_q_full(struct net_device *dev, unsigned int idx)
576 {
577 	struct adapter *adap = netdev2adap(dev);
578 	struct sge_uld_txq_info *txq_info =
579 		adap->sge.uld_txq_info[CXGB4_TX_CRYPTO];
580 	struct sge_uld_txq *txq;
581 	int ret = 0;
582 
583 	local_bh_disable();
584 	txq = &txq_info->uldtxq[idx];
585 	spin_lock(&txq->sendq.lock);
586 	if (txq->full)
587 		ret = -1;
588 	spin_unlock(&txq->sendq.lock);
589 	local_bh_enable();
590 	return ret;
591 }
592 
593 static int generate_copy_rrkey(struct ablk_ctx *ablkctx,
594 			       struct _key_ctx *key_ctx)
595 {
596 	if (ablkctx->ciph_mode == CHCR_SCMD_CIPHER_MODE_AES_CBC) {
597 		memcpy(key_ctx->key, ablkctx->rrkey, ablkctx->enckey_len);
598 	} else {
599 		memcpy(key_ctx->key,
600 		       ablkctx->key + (ablkctx->enckey_len >> 1),
601 		       ablkctx->enckey_len >> 1);
602 		memcpy(key_ctx->key + (ablkctx->enckey_len >> 1),
603 		       ablkctx->rrkey, ablkctx->enckey_len >> 1);
604 	}
605 	return 0;
606 }
607 
608 static int chcr_hash_ent_in_wr(struct scatterlist *src,
609 			     unsigned int minsg,
610 			     unsigned int space,
611 			     unsigned int srcskip)
612 {
613 	int srclen = 0;
614 	int srcsg = minsg;
615 	int soffset = 0, sless;
616 
617 	if (sg_dma_len(src) == srcskip) {
618 		src = sg_next(src);
619 		srcskip = 0;
620 	}
621 	while (src && space > (sgl_ent_len[srcsg + 1])) {
622 		sless = min_t(unsigned int, sg_dma_len(src) - soffset -	srcskip,
623 							CHCR_SRC_SG_SIZE);
624 		srclen += sless;
625 		soffset += sless;
626 		srcsg++;
627 		if (sg_dma_len(src) == (soffset + srcskip)) {
628 			src = sg_next(src);
629 			soffset = 0;
630 			srcskip = 0;
631 		}
632 	}
633 	return srclen;
634 }
635 
636 static int chcr_sg_ent_in_wr(struct scatterlist *src,
637 			     struct scatterlist *dst,
638 			     unsigned int minsg,
639 			     unsigned int space,
640 			     unsigned int srcskip,
641 			     unsigned int dstskip)
642 {
643 	int srclen = 0, dstlen = 0;
644 	int srcsg = minsg, dstsg = minsg;
645 	int offset = 0, soffset = 0, less, sless = 0;
646 
647 	if (sg_dma_len(src) == srcskip) {
648 		src = sg_next(src);
649 		srcskip = 0;
650 	}
651 	if (sg_dma_len(dst) == dstskip) {
652 		dst = sg_next(dst);
653 		dstskip = 0;
654 	}
655 
656 	while (src && dst &&
657 	       space > (sgl_ent_len[srcsg + 1] + dsgl_ent_len[dstsg])) {
658 		sless = min_t(unsigned int, sg_dma_len(src) - srcskip - soffset,
659 				CHCR_SRC_SG_SIZE);
660 		srclen += sless;
661 		srcsg++;
662 		offset = 0;
663 		while (dst && ((dstsg + 1) <= MAX_DSGL_ENT) &&
664 		       space > (sgl_ent_len[srcsg] + dsgl_ent_len[dstsg + 1])) {
665 			if (srclen <= dstlen)
666 				break;
667 			less = min_t(unsigned int, sg_dma_len(dst) - offset -
668 				     dstskip, CHCR_DST_SG_SIZE);
669 			dstlen += less;
670 			offset += less;
671 			if ((offset + dstskip) == sg_dma_len(dst)) {
672 				dst = sg_next(dst);
673 				offset = 0;
674 			}
675 			dstsg++;
676 			dstskip = 0;
677 		}
678 		soffset += sless;
679 		if ((soffset + srcskip) == sg_dma_len(src)) {
680 			src = sg_next(src);
681 			srcskip = 0;
682 			soffset = 0;
683 		}
684 
685 	}
686 	return min(srclen, dstlen);
687 }
688 
689 static int chcr_cipher_fallback(struct crypto_skcipher *cipher,
690 				struct skcipher_request *req,
691 				u8 *iv,
692 				unsigned short op_type)
693 {
694 	struct chcr_skcipher_req_ctx *reqctx = skcipher_request_ctx(req);
695 	int err;
696 
697 	skcipher_request_set_tfm(&reqctx->fallback_req, cipher);
698 	skcipher_request_set_callback(&reqctx->fallback_req, req->base.flags,
699 				      req->base.complete, req->base.data);
700 	skcipher_request_set_crypt(&reqctx->fallback_req, req->src, req->dst,
701 				   req->cryptlen, iv);
702 
703 	err = op_type ? crypto_skcipher_decrypt(&reqctx->fallback_req) :
704 			crypto_skcipher_encrypt(&reqctx->fallback_req);
705 
706 	return err;
707 
708 }
709 
710 static inline int get_qidxs(struct crypto_async_request *req,
711 			    unsigned int *txqidx, unsigned int *rxqidx)
712 {
713 	struct crypto_tfm *tfm = req->tfm;
714 	int ret = 0;
715 
716 	switch (tfm->__crt_alg->cra_flags & CRYPTO_ALG_TYPE_MASK) {
717 	case CRYPTO_ALG_TYPE_AEAD:
718 	{
719 		struct aead_request *aead_req =
720 			container_of(req, struct aead_request, base);
721 		struct chcr_aead_reqctx *reqctx = aead_request_ctx_dma(aead_req);
722 		*txqidx = reqctx->txqidx;
723 		*rxqidx = reqctx->rxqidx;
724 		break;
725 	}
726 	case CRYPTO_ALG_TYPE_SKCIPHER:
727 	{
728 		struct skcipher_request *sk_req =
729 			container_of(req, struct skcipher_request, base);
730 		struct chcr_skcipher_req_ctx *reqctx =
731 			skcipher_request_ctx(sk_req);
732 		*txqidx = reqctx->txqidx;
733 		*rxqidx = reqctx->rxqidx;
734 		break;
735 	}
736 	case CRYPTO_ALG_TYPE_AHASH:
737 	{
738 		struct ahash_request *ahash_req =
739 			container_of(req, struct ahash_request, base);
740 		struct chcr_ahash_req_ctx *reqctx =
741 			ahash_request_ctx(ahash_req);
742 		*txqidx = reqctx->txqidx;
743 		*rxqidx = reqctx->rxqidx;
744 		break;
745 	}
746 	default:
747 		ret = -EINVAL;
748 		/* should never get here */
749 		BUG();
750 		break;
751 	}
752 	return ret;
753 }
754 
755 static inline void create_wreq(struct chcr_context *ctx,
756 			       struct chcr_wr *chcr_req,
757 			       struct crypto_async_request *req,
758 			       unsigned int imm,
759 			       int hash_sz,
760 			       unsigned int len16,
761 			       unsigned int sc_len,
762 			       unsigned int lcb)
763 {
764 	struct uld_ctx *u_ctx = ULD_CTX(ctx);
765 	unsigned int tx_channel_id, rx_channel_id;
766 	unsigned int txqidx = 0, rxqidx = 0;
767 	unsigned int qid, fid, portno;
768 
769 	get_qidxs(req, &txqidx, &rxqidx);
770 	qid = u_ctx->lldi.rxq_ids[rxqidx];
771 	fid = u_ctx->lldi.rxq_ids[0];
772 	portno = rxqidx / ctx->rxq_perchan;
773 	tx_channel_id = txqidx / ctx->txq_perchan;
774 	rx_channel_id = cxgb4_port_e2cchan(u_ctx->lldi.ports[portno]);
775 
776 
777 	chcr_req->wreq.op_to_cctx_size = FILL_WR_OP_CCTX_SIZE;
778 	chcr_req->wreq.pld_size_hash_size =
779 		htonl(FW_CRYPTO_LOOKASIDE_WR_HASH_SIZE_V(hash_sz));
780 	chcr_req->wreq.len16_pkd =
781 		htonl(FW_CRYPTO_LOOKASIDE_WR_LEN16_V(DIV_ROUND_UP(len16, 16)));
782 	chcr_req->wreq.cookie = cpu_to_be64((uintptr_t)req);
783 	chcr_req->wreq.rx_chid_to_rx_q_id = FILL_WR_RX_Q_ID(rx_channel_id, qid,
784 							    !!lcb, txqidx);
785 
786 	chcr_req->ulptx.cmd_dest = FILL_ULPTX_CMD_DEST(tx_channel_id, fid);
787 	chcr_req->ulptx.len = htonl((DIV_ROUND_UP(len16, 16) -
788 				((sizeof(chcr_req->wreq)) >> 4)));
789 	chcr_req->sc_imm.cmd_more = FILL_CMD_MORE(!imm);
790 	chcr_req->sc_imm.len = cpu_to_be32(sizeof(struct cpl_tx_sec_pdu) +
791 					   sizeof(chcr_req->key_ctx) + sc_len);
792 }
793 
794 /**
795  *	create_cipher_wr - form the WR for cipher operations
796  *	@wrparam: Container for create_cipher_wr()'s parameters
797  */
798 static struct sk_buff *create_cipher_wr(struct cipher_wr_param *wrparam)
799 {
800 	struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(wrparam->req);
801 	struct chcr_context *ctx = c_ctx(tfm);
802 	struct uld_ctx *u_ctx = ULD_CTX(ctx);
803 	struct ablk_ctx *ablkctx = ABLK_CTX(ctx);
804 	struct sk_buff *skb = NULL;
805 	struct chcr_wr *chcr_req;
806 	struct cpl_rx_phys_dsgl *phys_cpl;
807 	struct ulptx_sgl *ulptx;
808 	struct chcr_skcipher_req_ctx *reqctx =
809 		skcipher_request_ctx(wrparam->req);
810 	unsigned int temp = 0, transhdr_len, dst_size;
811 	int error;
812 	int nents;
813 	unsigned int kctx_len;
814 	gfp_t flags = wrparam->req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP ?
815 			GFP_KERNEL : GFP_ATOMIC;
816 	struct adapter *adap = padap(ctx->dev);
817 	unsigned int rx_channel_id = reqctx->rxqidx / ctx->rxq_perchan;
818 
819 	rx_channel_id = cxgb4_port_e2cchan(u_ctx->lldi.ports[rx_channel_id]);
820 	nents = sg_nents_xlen(reqctx->dstsg,  wrparam->bytes, CHCR_DST_SG_SIZE,
821 			      reqctx->dst_ofst);
822 	dst_size = get_space_for_phys_dsgl(nents);
823 	kctx_len = roundup(ablkctx->enckey_len, 16);
824 	transhdr_len = CIPHER_TRANSHDR_SIZE(kctx_len, dst_size);
825 	nents = sg_nents_xlen(reqctx->srcsg, wrparam->bytes,
826 				  CHCR_SRC_SG_SIZE, reqctx->src_ofst);
827 	temp = reqctx->imm ? roundup(wrparam->bytes, 16) :
828 				     (sgl_len(nents) * 8);
829 	transhdr_len += temp;
830 	transhdr_len = roundup(transhdr_len, 16);
831 	skb = alloc_skb(SGE_MAX_WR_LEN, flags);
832 	if (!skb) {
833 		error = -ENOMEM;
834 		goto err;
835 	}
836 	chcr_req = __skb_put_zero(skb, transhdr_len);
837 	chcr_req->sec_cpl.op_ivinsrtofst =
838 			FILL_SEC_CPL_OP_IVINSR(rx_channel_id, 2, 1);
839 
840 	chcr_req->sec_cpl.pldlen = htonl(IV + wrparam->bytes);
841 	chcr_req->sec_cpl.aadstart_cipherstop_hi =
842 			FILL_SEC_CPL_CIPHERSTOP_HI(0, 0, IV + 1, 0);
843 
844 	chcr_req->sec_cpl.cipherstop_lo_authinsert =
845 			FILL_SEC_CPL_AUTHINSERT(0, 0, 0, 0);
846 	chcr_req->sec_cpl.seqno_numivs = FILL_SEC_CPL_SCMD0_SEQNO(reqctx->op, 0,
847 							 ablkctx->ciph_mode,
848 							 0, 0, IV >> 1);
849 	chcr_req->sec_cpl.ivgen_hdrlen = FILL_SEC_CPL_IVGEN_HDRLEN(0, 0, 0,
850 							  0, 1, dst_size);
851 
852 	chcr_req->key_ctx.ctx_hdr = ablkctx->key_ctx_hdr;
853 	if ((reqctx->op == CHCR_DECRYPT_OP) &&
854 	    (!(get_cryptoalg_subtype(tfm) ==
855 	       CRYPTO_ALG_SUB_TYPE_CTR)) &&
856 	    (!(get_cryptoalg_subtype(tfm) ==
857 	       CRYPTO_ALG_SUB_TYPE_CTR_RFC3686))) {
858 		generate_copy_rrkey(ablkctx, &chcr_req->key_ctx);
859 	} else {
860 		if ((ablkctx->ciph_mode == CHCR_SCMD_CIPHER_MODE_AES_CBC) ||
861 		    (ablkctx->ciph_mode == CHCR_SCMD_CIPHER_MODE_AES_CTR)) {
862 			memcpy(chcr_req->key_ctx.key, ablkctx->key,
863 			       ablkctx->enckey_len);
864 		} else {
865 			memcpy(chcr_req->key_ctx.key, ablkctx->key +
866 			       (ablkctx->enckey_len >> 1),
867 			       ablkctx->enckey_len >> 1);
868 			memcpy(chcr_req->key_ctx.key +
869 			       (ablkctx->enckey_len >> 1),
870 			       ablkctx->key,
871 			       ablkctx->enckey_len >> 1);
872 		}
873 	}
874 	phys_cpl = (struct cpl_rx_phys_dsgl *)((u8 *)(chcr_req + 1) + kctx_len);
875 	ulptx = (struct ulptx_sgl *)((u8 *)(phys_cpl + 1) + dst_size);
876 	chcr_add_cipher_src_ent(wrparam->req, ulptx, wrparam);
877 	chcr_add_cipher_dst_ent(wrparam->req, phys_cpl, wrparam, wrparam->qid);
878 
879 	atomic_inc(&adap->chcr_stats.cipher_rqst);
880 	temp = sizeof(struct cpl_rx_phys_dsgl) + dst_size + kctx_len + IV
881 		+ (reqctx->imm ? (wrparam->bytes) : 0);
882 	create_wreq(c_ctx(tfm), chcr_req, &(wrparam->req->base), reqctx->imm, 0,
883 		    transhdr_len, temp,
884 			ablkctx->ciph_mode == CHCR_SCMD_CIPHER_MODE_AES_CBC);
885 	reqctx->skb = skb;
886 
887 	if (reqctx->op && (ablkctx->ciph_mode ==
888 			   CHCR_SCMD_CIPHER_MODE_AES_CBC))
889 		sg_pcopy_to_buffer(wrparam->req->src,
890 			sg_nents(wrparam->req->src), wrparam->req->iv, 16,
891 			reqctx->processed + wrparam->bytes - AES_BLOCK_SIZE);
892 
893 	return skb;
894 err:
895 	return ERR_PTR(error);
896 }
897 
898 static inline int chcr_keyctx_ck_size(unsigned int keylen)
899 {
900 	int ck_size = 0;
901 
902 	if (keylen == AES_KEYSIZE_128)
903 		ck_size = CHCR_KEYCTX_CIPHER_KEY_SIZE_128;
904 	else if (keylen == AES_KEYSIZE_192)
905 		ck_size = CHCR_KEYCTX_CIPHER_KEY_SIZE_192;
906 	else if (keylen == AES_KEYSIZE_256)
907 		ck_size = CHCR_KEYCTX_CIPHER_KEY_SIZE_256;
908 	else
909 		ck_size = 0;
910 
911 	return ck_size;
912 }
913 static int chcr_cipher_fallback_setkey(struct crypto_skcipher *cipher,
914 				       const u8 *key,
915 				       unsigned int keylen)
916 {
917 	struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(cipher));
918 
919 	crypto_skcipher_clear_flags(ablkctx->sw_cipher,
920 				CRYPTO_TFM_REQ_MASK);
921 	crypto_skcipher_set_flags(ablkctx->sw_cipher,
922 				cipher->base.crt_flags & CRYPTO_TFM_REQ_MASK);
923 	return crypto_skcipher_setkey(ablkctx->sw_cipher, key, keylen);
924 }
925 
926 static int chcr_aes_cbc_setkey(struct crypto_skcipher *cipher,
927 			       const u8 *key,
928 			       unsigned int keylen)
929 {
930 	struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(cipher));
931 	unsigned int ck_size, context_size;
932 	u16 alignment = 0;
933 	int err;
934 
935 	err = chcr_cipher_fallback_setkey(cipher, key, keylen);
936 	if (err)
937 		goto badkey_err;
938 
939 	ck_size = chcr_keyctx_ck_size(keylen);
940 	alignment = ck_size == CHCR_KEYCTX_CIPHER_KEY_SIZE_192 ? 8 : 0;
941 	memcpy(ablkctx->key, key, keylen);
942 	ablkctx->enckey_len = keylen;
943 	get_aes_decrypt_key(ablkctx->rrkey, ablkctx->key, keylen << 3);
944 	context_size = (KEY_CONTEXT_HDR_SALT_AND_PAD +
945 			keylen + alignment) >> 4;
946 
947 	ablkctx->key_ctx_hdr = FILL_KEY_CTX_HDR(ck_size, CHCR_KEYCTX_NO_KEY,
948 						0, 0, context_size);
949 	ablkctx->ciph_mode = CHCR_SCMD_CIPHER_MODE_AES_CBC;
950 	return 0;
951 badkey_err:
952 	ablkctx->enckey_len = 0;
953 
954 	return err;
955 }
956 
957 static int chcr_aes_ctr_setkey(struct crypto_skcipher *cipher,
958 				   const u8 *key,
959 				   unsigned int keylen)
960 {
961 	struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(cipher));
962 	unsigned int ck_size, context_size;
963 	u16 alignment = 0;
964 	int err;
965 
966 	err = chcr_cipher_fallback_setkey(cipher, key, keylen);
967 	if (err)
968 		goto badkey_err;
969 	ck_size = chcr_keyctx_ck_size(keylen);
970 	alignment = (ck_size == CHCR_KEYCTX_CIPHER_KEY_SIZE_192) ? 8 : 0;
971 	memcpy(ablkctx->key, key, keylen);
972 	ablkctx->enckey_len = keylen;
973 	context_size = (KEY_CONTEXT_HDR_SALT_AND_PAD +
974 			keylen + alignment) >> 4;
975 
976 	ablkctx->key_ctx_hdr = FILL_KEY_CTX_HDR(ck_size, CHCR_KEYCTX_NO_KEY,
977 						0, 0, context_size);
978 	ablkctx->ciph_mode = CHCR_SCMD_CIPHER_MODE_AES_CTR;
979 
980 	return 0;
981 badkey_err:
982 	ablkctx->enckey_len = 0;
983 
984 	return err;
985 }
986 
987 static int chcr_aes_rfc3686_setkey(struct crypto_skcipher *cipher,
988 				   const u8 *key,
989 				   unsigned int keylen)
990 {
991 	struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(cipher));
992 	unsigned int ck_size, context_size;
993 	u16 alignment = 0;
994 	int err;
995 
996 	if (keylen < CTR_RFC3686_NONCE_SIZE)
997 		return -EINVAL;
998 	memcpy(ablkctx->nonce, key + (keylen - CTR_RFC3686_NONCE_SIZE),
999 	       CTR_RFC3686_NONCE_SIZE);
1000 
1001 	keylen -= CTR_RFC3686_NONCE_SIZE;
1002 	err = chcr_cipher_fallback_setkey(cipher, key, keylen);
1003 	if (err)
1004 		goto badkey_err;
1005 
1006 	ck_size = chcr_keyctx_ck_size(keylen);
1007 	alignment = (ck_size == CHCR_KEYCTX_CIPHER_KEY_SIZE_192) ? 8 : 0;
1008 	memcpy(ablkctx->key, key, keylen);
1009 	ablkctx->enckey_len = keylen;
1010 	context_size = (KEY_CONTEXT_HDR_SALT_AND_PAD +
1011 			keylen + alignment) >> 4;
1012 
1013 	ablkctx->key_ctx_hdr = FILL_KEY_CTX_HDR(ck_size, CHCR_KEYCTX_NO_KEY,
1014 						0, 0, context_size);
1015 	ablkctx->ciph_mode = CHCR_SCMD_CIPHER_MODE_AES_CTR;
1016 
1017 	return 0;
1018 badkey_err:
1019 	ablkctx->enckey_len = 0;
1020 
1021 	return err;
1022 }
1023 static void ctr_add_iv(u8 *dstiv, u8 *srciv, u32 add)
1024 {
1025 	unsigned int size = AES_BLOCK_SIZE;
1026 	__be32 *b = (__be32 *)(dstiv + size);
1027 	u32 c, prev;
1028 
1029 	memcpy(dstiv, srciv, AES_BLOCK_SIZE);
1030 	for (; size >= 4; size -= 4) {
1031 		prev = be32_to_cpu(*--b);
1032 		c = prev + add;
1033 		*b = cpu_to_be32(c);
1034 		if (prev < c)
1035 			break;
1036 		add = 1;
1037 	}
1038 
1039 }
1040 
1041 static unsigned int adjust_ctr_overflow(u8 *iv, u32 bytes)
1042 {
1043 	__be32 *b = (__be32 *)(iv + AES_BLOCK_SIZE);
1044 	u64 c;
1045 	u32 temp = be32_to_cpu(*--b);
1046 
1047 	temp = ~temp;
1048 	c = (u64)temp +  1; // No of block can processed without overflow
1049 	if ((bytes / AES_BLOCK_SIZE) >= c)
1050 		bytes = c * AES_BLOCK_SIZE;
1051 	return bytes;
1052 }
1053 
1054 static int chcr_update_tweak(struct skcipher_request *req, u8 *iv,
1055 			     u32 isfinal)
1056 {
1057 	struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
1058 	struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(tfm));
1059 	struct chcr_skcipher_req_ctx *reqctx = skcipher_request_ctx(req);
1060 	struct crypto_aes_ctx aes;
1061 	int ret, i;
1062 	u8 *key;
1063 	unsigned int keylen;
1064 	int round = reqctx->last_req_len / AES_BLOCK_SIZE;
1065 	int round8 = round / 8;
1066 
1067 	memcpy(iv, reqctx->iv, AES_BLOCK_SIZE);
1068 
1069 	keylen = ablkctx->enckey_len / 2;
1070 	key = ablkctx->key + keylen;
1071 	/* For a 192 bit key remove the padded zeroes which was
1072 	 * added in chcr_xts_setkey
1073 	 */
1074 	if (KEY_CONTEXT_CK_SIZE_G(ntohl(ablkctx->key_ctx_hdr))
1075 			== CHCR_KEYCTX_CIPHER_KEY_SIZE_192)
1076 		ret = aes_expandkey(&aes, key, keylen - 8);
1077 	else
1078 		ret = aes_expandkey(&aes, key, keylen);
1079 	if (ret)
1080 		return ret;
1081 	aes_encrypt(&aes, iv, iv);
1082 	for (i = 0; i < round8; i++)
1083 		gf128mul_x8_ble((le128 *)iv, (le128 *)iv);
1084 
1085 	for (i = 0; i < (round % 8); i++)
1086 		gf128mul_x_ble((le128 *)iv, (le128 *)iv);
1087 
1088 	if (!isfinal)
1089 		aes_decrypt(&aes, iv, iv);
1090 
1091 	memzero_explicit(&aes, sizeof(aes));
1092 	return 0;
1093 }
1094 
1095 static int chcr_update_cipher_iv(struct skcipher_request *req,
1096 				   struct cpl_fw6_pld *fw6_pld, u8 *iv)
1097 {
1098 	struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
1099 	struct chcr_skcipher_req_ctx *reqctx = skcipher_request_ctx(req);
1100 	int subtype = get_cryptoalg_subtype(tfm);
1101 	int ret = 0;
1102 
1103 	if (subtype == CRYPTO_ALG_SUB_TYPE_CTR)
1104 		ctr_add_iv(iv, req->iv, (reqctx->processed /
1105 			   AES_BLOCK_SIZE));
1106 	else if (subtype == CRYPTO_ALG_SUB_TYPE_CTR_RFC3686)
1107 		*(__be32 *)(reqctx->iv + CTR_RFC3686_NONCE_SIZE +
1108 			CTR_RFC3686_IV_SIZE) = cpu_to_be32((reqctx->processed /
1109 						AES_BLOCK_SIZE) + 1);
1110 	else if (subtype == CRYPTO_ALG_SUB_TYPE_XTS)
1111 		ret = chcr_update_tweak(req, iv, 0);
1112 	else if (subtype == CRYPTO_ALG_SUB_TYPE_CBC) {
1113 		if (reqctx->op)
1114 			/*Updated before sending last WR*/
1115 			memcpy(iv, req->iv, AES_BLOCK_SIZE);
1116 		else
1117 			memcpy(iv, &fw6_pld->data[2], AES_BLOCK_SIZE);
1118 	}
1119 
1120 	return ret;
1121 
1122 }
1123 
1124 /* We need separate function for final iv because in rfc3686  Initial counter
1125  * starts from 1 and buffer size of iv is 8 byte only which remains constant
1126  * for subsequent update requests
1127  */
1128 
1129 static int chcr_final_cipher_iv(struct skcipher_request *req,
1130 				   struct cpl_fw6_pld *fw6_pld, u8 *iv)
1131 {
1132 	struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
1133 	struct chcr_skcipher_req_ctx *reqctx = skcipher_request_ctx(req);
1134 	int subtype = get_cryptoalg_subtype(tfm);
1135 	int ret = 0;
1136 
1137 	if (subtype == CRYPTO_ALG_SUB_TYPE_CTR)
1138 		ctr_add_iv(iv, req->iv, DIV_ROUND_UP(reqctx->processed,
1139 						       AES_BLOCK_SIZE));
1140 	else if (subtype == CRYPTO_ALG_SUB_TYPE_XTS) {
1141 		if (!reqctx->partial_req)
1142 			memcpy(iv, reqctx->iv, AES_BLOCK_SIZE);
1143 		else
1144 			ret = chcr_update_tweak(req, iv, 1);
1145 	}
1146 	else if (subtype == CRYPTO_ALG_SUB_TYPE_CBC) {
1147 		/*Already updated for Decrypt*/
1148 		if (!reqctx->op)
1149 			memcpy(iv, &fw6_pld->data[2], AES_BLOCK_SIZE);
1150 
1151 	}
1152 	return ret;
1153 
1154 }
1155 
1156 static int chcr_handle_cipher_resp(struct skcipher_request *req,
1157 				   unsigned char *input, int err)
1158 {
1159 	struct chcr_skcipher_req_ctx *reqctx = skcipher_request_ctx(req);
1160 	struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
1161 	struct cpl_fw6_pld *fw6_pld = (struct cpl_fw6_pld *)input;
1162 	struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(tfm));
1163 	struct uld_ctx *u_ctx = ULD_CTX(c_ctx(tfm));
1164 	struct chcr_dev *dev = c_ctx(tfm)->dev;
1165 	struct chcr_context *ctx = c_ctx(tfm);
1166 	struct adapter *adap = padap(ctx->dev);
1167 	struct cipher_wr_param wrparam;
1168 	struct sk_buff *skb;
1169 	int bytes;
1170 
1171 	if (err)
1172 		goto unmap;
1173 	if (req->cryptlen == reqctx->processed) {
1174 		chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev,
1175 				      req);
1176 		err = chcr_final_cipher_iv(req, fw6_pld, req->iv);
1177 		goto complete;
1178 	}
1179 
1180 	if (!reqctx->imm) {
1181 		bytes = chcr_sg_ent_in_wr(reqctx->srcsg, reqctx->dstsg, 0,
1182 					  CIP_SPACE_LEFT(ablkctx->enckey_len),
1183 					  reqctx->src_ofst, reqctx->dst_ofst);
1184 		if ((bytes + reqctx->processed) >= req->cryptlen)
1185 			bytes  = req->cryptlen - reqctx->processed;
1186 		else
1187 			bytes = rounddown(bytes, 16);
1188 	} else {
1189 		/*CTR mode counter overfloa*/
1190 		bytes  = req->cryptlen - reqctx->processed;
1191 	}
1192 	err = chcr_update_cipher_iv(req, fw6_pld, reqctx->iv);
1193 	if (err)
1194 		goto unmap;
1195 
1196 	if (unlikely(bytes == 0)) {
1197 		chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev,
1198 				      req);
1199 		memcpy(req->iv, reqctx->init_iv, IV);
1200 		atomic_inc(&adap->chcr_stats.fallback);
1201 		err = chcr_cipher_fallback(ablkctx->sw_cipher, req, req->iv,
1202 					   reqctx->op);
1203 		goto complete;
1204 	}
1205 
1206 	if (get_cryptoalg_subtype(tfm) ==
1207 	    CRYPTO_ALG_SUB_TYPE_CTR)
1208 		bytes = adjust_ctr_overflow(reqctx->iv, bytes);
1209 	wrparam.qid = u_ctx->lldi.rxq_ids[reqctx->rxqidx];
1210 	wrparam.req = req;
1211 	wrparam.bytes = bytes;
1212 	skb = create_cipher_wr(&wrparam);
1213 	if (IS_ERR(skb)) {
1214 		pr_err("%s : Failed to form WR. No memory\n", __func__);
1215 		err = PTR_ERR(skb);
1216 		goto unmap;
1217 	}
1218 	skb->dev = u_ctx->lldi.ports[0];
1219 	set_wr_txq(skb, CPL_PRIORITY_DATA, reqctx->txqidx);
1220 	chcr_send_wr(skb);
1221 	reqctx->last_req_len = bytes;
1222 	reqctx->processed += bytes;
1223 	if (get_cryptoalg_subtype(tfm) ==
1224 		CRYPTO_ALG_SUB_TYPE_CBC && req->base.flags ==
1225 			CRYPTO_TFM_REQ_MAY_SLEEP ) {
1226 		complete(&ctx->cbc_aes_aio_done);
1227 	}
1228 	return 0;
1229 unmap:
1230 	chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, req);
1231 complete:
1232 	if (get_cryptoalg_subtype(tfm) ==
1233 		CRYPTO_ALG_SUB_TYPE_CBC && req->base.flags ==
1234 			CRYPTO_TFM_REQ_MAY_SLEEP ) {
1235 		complete(&ctx->cbc_aes_aio_done);
1236 	}
1237 	chcr_dec_wrcount(dev);
1238 	skcipher_request_complete(req, err);
1239 	return err;
1240 }
1241 
1242 static int process_cipher(struct skcipher_request *req,
1243 				  unsigned short qid,
1244 				  struct sk_buff **skb,
1245 				  unsigned short op_type)
1246 {
1247 	struct chcr_skcipher_req_ctx *reqctx = skcipher_request_ctx(req);
1248 	struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
1249 	unsigned int ivsize = crypto_skcipher_ivsize(tfm);
1250 	struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(tfm));
1251 	struct adapter *adap = padap(c_ctx(tfm)->dev);
1252 	struct	cipher_wr_param wrparam;
1253 	int bytes, err = -EINVAL;
1254 	int subtype;
1255 
1256 	reqctx->processed = 0;
1257 	reqctx->partial_req = 0;
1258 	if (!req->iv)
1259 		goto error;
1260 	subtype = get_cryptoalg_subtype(tfm);
1261 	if ((ablkctx->enckey_len == 0) || (ivsize > AES_BLOCK_SIZE) ||
1262 	    (req->cryptlen == 0) ||
1263 	    (req->cryptlen % crypto_skcipher_blocksize(tfm))) {
1264 		if (req->cryptlen == 0 && subtype != CRYPTO_ALG_SUB_TYPE_XTS)
1265 			goto fallback;
1266 		else if (req->cryptlen % crypto_skcipher_blocksize(tfm) &&
1267 			 subtype == CRYPTO_ALG_SUB_TYPE_XTS)
1268 			goto fallback;
1269 		pr_err("AES: Invalid value of Key Len %d nbytes %d IV Len %d\n",
1270 		       ablkctx->enckey_len, req->cryptlen, ivsize);
1271 		goto error;
1272 	}
1273 
1274 	err = chcr_cipher_dma_map(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, req);
1275 	if (err)
1276 		goto error;
1277 	if (req->cryptlen < (SGE_MAX_WR_LEN - (sizeof(struct chcr_wr) +
1278 					    AES_MIN_KEY_SIZE +
1279 					    sizeof(struct cpl_rx_phys_dsgl) +
1280 					/*Min dsgl size*/
1281 					    32))) {
1282 		/* Can be sent as Imm*/
1283 		unsigned int dnents = 0, transhdr_len, phys_dsgl, kctx_len;
1284 
1285 		dnents = sg_nents_xlen(req->dst, req->cryptlen,
1286 				       CHCR_DST_SG_SIZE, 0);
1287 		phys_dsgl = get_space_for_phys_dsgl(dnents);
1288 		kctx_len = roundup(ablkctx->enckey_len, 16);
1289 		transhdr_len = CIPHER_TRANSHDR_SIZE(kctx_len, phys_dsgl);
1290 		reqctx->imm = (transhdr_len + IV + req->cryptlen) <=
1291 			SGE_MAX_WR_LEN;
1292 		bytes = IV + req->cryptlen;
1293 
1294 	} else {
1295 		reqctx->imm = 0;
1296 	}
1297 
1298 	if (!reqctx->imm) {
1299 		bytes = chcr_sg_ent_in_wr(req->src, req->dst, 0,
1300 					  CIP_SPACE_LEFT(ablkctx->enckey_len),
1301 					  0, 0);
1302 		if ((bytes + reqctx->processed) >= req->cryptlen)
1303 			bytes  = req->cryptlen - reqctx->processed;
1304 		else
1305 			bytes = rounddown(bytes, 16);
1306 	} else {
1307 		bytes = req->cryptlen;
1308 	}
1309 	if (subtype == CRYPTO_ALG_SUB_TYPE_CTR) {
1310 		bytes = adjust_ctr_overflow(req->iv, bytes);
1311 	}
1312 	if (subtype == CRYPTO_ALG_SUB_TYPE_CTR_RFC3686) {
1313 		memcpy(reqctx->iv, ablkctx->nonce, CTR_RFC3686_NONCE_SIZE);
1314 		memcpy(reqctx->iv + CTR_RFC3686_NONCE_SIZE, req->iv,
1315 				CTR_RFC3686_IV_SIZE);
1316 
1317 		/* initialize counter portion of counter block */
1318 		*(__be32 *)(reqctx->iv + CTR_RFC3686_NONCE_SIZE +
1319 			CTR_RFC3686_IV_SIZE) = cpu_to_be32(1);
1320 		memcpy(reqctx->init_iv, reqctx->iv, IV);
1321 
1322 	} else {
1323 
1324 		memcpy(reqctx->iv, req->iv, IV);
1325 		memcpy(reqctx->init_iv, req->iv, IV);
1326 	}
1327 	if (unlikely(bytes == 0)) {
1328 		chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev,
1329 				      req);
1330 fallback:       atomic_inc(&adap->chcr_stats.fallback);
1331 		err = chcr_cipher_fallback(ablkctx->sw_cipher, req,
1332 					   subtype ==
1333 					   CRYPTO_ALG_SUB_TYPE_CTR_RFC3686 ?
1334 					   reqctx->iv : req->iv,
1335 					   op_type);
1336 		goto error;
1337 	}
1338 	reqctx->op = op_type;
1339 	reqctx->srcsg = req->src;
1340 	reqctx->dstsg = req->dst;
1341 	reqctx->src_ofst = 0;
1342 	reqctx->dst_ofst = 0;
1343 	wrparam.qid = qid;
1344 	wrparam.req = req;
1345 	wrparam.bytes = bytes;
1346 	*skb = create_cipher_wr(&wrparam);
1347 	if (IS_ERR(*skb)) {
1348 		err = PTR_ERR(*skb);
1349 		goto unmap;
1350 	}
1351 	reqctx->processed = bytes;
1352 	reqctx->last_req_len = bytes;
1353 	reqctx->partial_req = !!(req->cryptlen - reqctx->processed);
1354 
1355 	return 0;
1356 unmap:
1357 	chcr_cipher_dma_unmap(&ULD_CTX(c_ctx(tfm))->lldi.pdev->dev, req);
1358 error:
1359 	return err;
1360 }
1361 
1362 static int chcr_aes_encrypt(struct skcipher_request *req)
1363 {
1364 	struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
1365 	struct chcr_skcipher_req_ctx *reqctx = skcipher_request_ctx(req);
1366 	struct chcr_dev *dev = c_ctx(tfm)->dev;
1367 	struct sk_buff *skb = NULL;
1368 	int err;
1369 	struct uld_ctx *u_ctx = ULD_CTX(c_ctx(tfm));
1370 	struct chcr_context *ctx = c_ctx(tfm);
1371 	unsigned int cpu;
1372 
1373 	cpu = get_cpu();
1374 	reqctx->txqidx = cpu % ctx->ntxq;
1375 	reqctx->rxqidx = cpu % ctx->nrxq;
1376 	put_cpu();
1377 
1378 	err = chcr_inc_wrcount(dev);
1379 	if (err)
1380 		return -ENXIO;
1381 	if (unlikely(cxgb4_is_crypto_q_full(u_ctx->lldi.ports[0],
1382 						reqctx->txqidx) &&
1383 		(!(req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG)))) {
1384 			err = -ENOSPC;
1385 			goto error;
1386 	}
1387 
1388 	err = process_cipher(req, u_ctx->lldi.rxq_ids[reqctx->rxqidx],
1389 			     &skb, CHCR_ENCRYPT_OP);
1390 	if (err || !skb)
1391 		return  err;
1392 	skb->dev = u_ctx->lldi.ports[0];
1393 	set_wr_txq(skb, CPL_PRIORITY_DATA, reqctx->txqidx);
1394 	chcr_send_wr(skb);
1395 	if (get_cryptoalg_subtype(tfm) ==
1396 		CRYPTO_ALG_SUB_TYPE_CBC && req->base.flags ==
1397 			CRYPTO_TFM_REQ_MAY_SLEEP ) {
1398 			reqctx->partial_req = 1;
1399 			wait_for_completion(&ctx->cbc_aes_aio_done);
1400         }
1401 	return -EINPROGRESS;
1402 error:
1403 	chcr_dec_wrcount(dev);
1404 	return err;
1405 }
1406 
1407 static int chcr_aes_decrypt(struct skcipher_request *req)
1408 {
1409 	struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
1410 	struct chcr_skcipher_req_ctx *reqctx = skcipher_request_ctx(req);
1411 	struct uld_ctx *u_ctx = ULD_CTX(c_ctx(tfm));
1412 	struct chcr_dev *dev = c_ctx(tfm)->dev;
1413 	struct sk_buff *skb = NULL;
1414 	int err;
1415 	struct chcr_context *ctx = c_ctx(tfm);
1416 	unsigned int cpu;
1417 
1418 	cpu = get_cpu();
1419 	reqctx->txqidx = cpu % ctx->ntxq;
1420 	reqctx->rxqidx = cpu % ctx->nrxq;
1421 	put_cpu();
1422 
1423 	err = chcr_inc_wrcount(dev);
1424 	if (err)
1425 		return -ENXIO;
1426 
1427 	if (unlikely(cxgb4_is_crypto_q_full(u_ctx->lldi.ports[0],
1428 						reqctx->txqidx) &&
1429 		(!(req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG))))
1430 			return -ENOSPC;
1431 	err = process_cipher(req, u_ctx->lldi.rxq_ids[reqctx->rxqidx],
1432 			     &skb, CHCR_DECRYPT_OP);
1433 	if (err || !skb)
1434 		return err;
1435 	skb->dev = u_ctx->lldi.ports[0];
1436 	set_wr_txq(skb, CPL_PRIORITY_DATA, reqctx->txqidx);
1437 	chcr_send_wr(skb);
1438 	return -EINPROGRESS;
1439 }
1440 static int chcr_device_init(struct chcr_context *ctx)
1441 {
1442 	struct uld_ctx *u_ctx = NULL;
1443 	int txq_perchan, ntxq;
1444 	int err = 0, rxq_perchan;
1445 
1446 	if (!ctx->dev) {
1447 		u_ctx = assign_chcr_device();
1448 		if (!u_ctx) {
1449 			err = -ENXIO;
1450 			pr_err("chcr device assignment fails\n");
1451 			goto out;
1452 		}
1453 		ctx->dev = &u_ctx->dev;
1454 		ntxq = u_ctx->lldi.ntxq;
1455 		rxq_perchan = u_ctx->lldi.nrxq / u_ctx->lldi.nchan;
1456 		txq_perchan = ntxq / u_ctx->lldi.nchan;
1457 		ctx->ntxq = ntxq;
1458 		ctx->nrxq = u_ctx->lldi.nrxq;
1459 		ctx->rxq_perchan = rxq_perchan;
1460 		ctx->txq_perchan = txq_perchan;
1461 	}
1462 out:
1463 	return err;
1464 }
1465 
1466 static int chcr_init_tfm(struct crypto_skcipher *tfm)
1467 {
1468 	struct skcipher_alg *alg = crypto_skcipher_alg(tfm);
1469 	struct chcr_context *ctx = crypto_skcipher_ctx(tfm);
1470 	struct ablk_ctx *ablkctx = ABLK_CTX(ctx);
1471 
1472 	ablkctx->sw_cipher = crypto_alloc_skcipher(alg->base.cra_name, 0,
1473 				CRYPTO_ALG_NEED_FALLBACK);
1474 	if (IS_ERR(ablkctx->sw_cipher)) {
1475 		pr_err("failed to allocate fallback for %s\n", alg->base.cra_name);
1476 		return PTR_ERR(ablkctx->sw_cipher);
1477 	}
1478 	init_completion(&ctx->cbc_aes_aio_done);
1479 	crypto_skcipher_set_reqsize(tfm, sizeof(struct chcr_skcipher_req_ctx) +
1480 					 crypto_skcipher_reqsize(ablkctx->sw_cipher));
1481 
1482 	return chcr_device_init(ctx);
1483 }
1484 
1485 static int chcr_rfc3686_init(struct crypto_skcipher *tfm)
1486 {
1487 	struct skcipher_alg *alg = crypto_skcipher_alg(tfm);
1488 	struct chcr_context *ctx = crypto_skcipher_ctx(tfm);
1489 	struct ablk_ctx *ablkctx = ABLK_CTX(ctx);
1490 
1491 	/*RFC3686 initialises IV counter value to 1, rfc3686(ctr(aes))
1492 	 * cannot be used as fallback in chcr_handle_cipher_response
1493 	 */
1494 	ablkctx->sw_cipher = crypto_alloc_skcipher("ctr(aes)", 0,
1495 				CRYPTO_ALG_NEED_FALLBACK);
1496 	if (IS_ERR(ablkctx->sw_cipher)) {
1497 		pr_err("failed to allocate fallback for %s\n", alg->base.cra_name);
1498 		return PTR_ERR(ablkctx->sw_cipher);
1499 	}
1500 	crypto_skcipher_set_reqsize(tfm, sizeof(struct chcr_skcipher_req_ctx) +
1501 				    crypto_skcipher_reqsize(ablkctx->sw_cipher));
1502 	return chcr_device_init(ctx);
1503 }
1504 
1505 
1506 static void chcr_exit_tfm(struct crypto_skcipher *tfm)
1507 {
1508 	struct chcr_context *ctx = crypto_skcipher_ctx(tfm);
1509 	struct ablk_ctx *ablkctx = ABLK_CTX(ctx);
1510 
1511 	crypto_free_skcipher(ablkctx->sw_cipher);
1512 }
1513 
1514 static int get_alg_config(struct algo_param *params,
1515 			  unsigned int auth_size)
1516 {
1517 	switch (auth_size) {
1518 	case SHA1_DIGEST_SIZE:
1519 		params->mk_size = CHCR_KEYCTX_MAC_KEY_SIZE_160;
1520 		params->auth_mode = CHCR_SCMD_AUTH_MODE_SHA1;
1521 		params->result_size = SHA1_DIGEST_SIZE;
1522 		break;
1523 	case SHA224_DIGEST_SIZE:
1524 		params->mk_size = CHCR_KEYCTX_MAC_KEY_SIZE_256;
1525 		params->auth_mode = CHCR_SCMD_AUTH_MODE_SHA224;
1526 		params->result_size = SHA256_DIGEST_SIZE;
1527 		break;
1528 	case SHA256_DIGEST_SIZE:
1529 		params->mk_size = CHCR_KEYCTX_MAC_KEY_SIZE_256;
1530 		params->auth_mode = CHCR_SCMD_AUTH_MODE_SHA256;
1531 		params->result_size = SHA256_DIGEST_SIZE;
1532 		break;
1533 	case SHA384_DIGEST_SIZE:
1534 		params->mk_size = CHCR_KEYCTX_MAC_KEY_SIZE_512;
1535 		params->auth_mode = CHCR_SCMD_AUTH_MODE_SHA512_384;
1536 		params->result_size = SHA512_DIGEST_SIZE;
1537 		break;
1538 	case SHA512_DIGEST_SIZE:
1539 		params->mk_size = CHCR_KEYCTX_MAC_KEY_SIZE_512;
1540 		params->auth_mode = CHCR_SCMD_AUTH_MODE_SHA512_512;
1541 		params->result_size = SHA512_DIGEST_SIZE;
1542 		break;
1543 	default:
1544 		pr_err("ERROR, unsupported digest size\n");
1545 		return -EINVAL;
1546 	}
1547 	return 0;
1548 }
1549 
1550 static inline void chcr_free_shash(struct crypto_shash *base_hash)
1551 {
1552 		crypto_free_shash(base_hash);
1553 }
1554 
1555 /**
1556  *	create_hash_wr - Create hash work request
1557  *	@req: Cipher req base
1558  *	@param: Container for create_hash_wr()'s parameters
1559  */
1560 static struct sk_buff *create_hash_wr(struct ahash_request *req,
1561 				      struct hash_wr_param *param)
1562 {
1563 	struct chcr_ahash_req_ctx *req_ctx = ahash_request_ctx(req);
1564 	struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
1565 	struct chcr_context *ctx = h_ctx(tfm);
1566 	struct hmac_ctx *hmacctx = HMAC_CTX(ctx);
1567 	struct sk_buff *skb = NULL;
1568 	struct uld_ctx *u_ctx = ULD_CTX(ctx);
1569 	struct chcr_wr *chcr_req;
1570 	struct ulptx_sgl *ulptx;
1571 	unsigned int nents = 0, transhdr_len;
1572 	unsigned int temp = 0;
1573 	gfp_t flags = req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP ? GFP_KERNEL :
1574 		GFP_ATOMIC;
1575 	struct adapter *adap = padap(h_ctx(tfm)->dev);
1576 	int error = 0;
1577 	unsigned int rx_channel_id = req_ctx->rxqidx / ctx->rxq_perchan;
1578 
1579 	rx_channel_id = cxgb4_port_e2cchan(u_ctx->lldi.ports[rx_channel_id]);
1580 	transhdr_len = HASH_TRANSHDR_SIZE(param->kctx_len);
1581 	req_ctx->hctx_wr.imm = (transhdr_len + param->bfr_len +
1582 				param->sg_len) <= SGE_MAX_WR_LEN;
1583 	nents = sg_nents_xlen(req_ctx->hctx_wr.srcsg, param->sg_len,
1584 		      CHCR_SRC_SG_SIZE, req_ctx->hctx_wr.src_ofst);
1585 	nents += param->bfr_len ? 1 : 0;
1586 	transhdr_len += req_ctx->hctx_wr.imm ? roundup(param->bfr_len +
1587 				param->sg_len, 16) : (sgl_len(nents) * 8);
1588 	transhdr_len = roundup(transhdr_len, 16);
1589 
1590 	skb = alloc_skb(transhdr_len, flags);
1591 	if (!skb)
1592 		return ERR_PTR(-ENOMEM);
1593 	chcr_req = __skb_put_zero(skb, transhdr_len);
1594 
1595 	chcr_req->sec_cpl.op_ivinsrtofst =
1596 		FILL_SEC_CPL_OP_IVINSR(rx_channel_id, 2, 0);
1597 
1598 	chcr_req->sec_cpl.pldlen = htonl(param->bfr_len + param->sg_len);
1599 
1600 	chcr_req->sec_cpl.aadstart_cipherstop_hi =
1601 		FILL_SEC_CPL_CIPHERSTOP_HI(0, 0, 0, 0);
1602 	chcr_req->sec_cpl.cipherstop_lo_authinsert =
1603 		FILL_SEC_CPL_AUTHINSERT(0, 1, 0, 0);
1604 	chcr_req->sec_cpl.seqno_numivs =
1605 		FILL_SEC_CPL_SCMD0_SEQNO(0, 0, 0, param->alg_prm.auth_mode,
1606 					 param->opad_needed, 0);
1607 
1608 	chcr_req->sec_cpl.ivgen_hdrlen =
1609 		FILL_SEC_CPL_IVGEN_HDRLEN(param->last, param->more, 0, 1, 0, 0);
1610 
1611 	memcpy(chcr_req->key_ctx.key, req_ctx->partial_hash,
1612 	       param->alg_prm.result_size);
1613 
1614 	if (param->opad_needed)
1615 		memcpy(chcr_req->key_ctx.key +
1616 		       ((param->alg_prm.result_size <= 32) ? 32 :
1617 			CHCR_HASH_MAX_DIGEST_SIZE),
1618 		       hmacctx->opad, param->alg_prm.result_size);
1619 
1620 	chcr_req->key_ctx.ctx_hdr = FILL_KEY_CTX_HDR(CHCR_KEYCTX_NO_KEY,
1621 					    param->alg_prm.mk_size, 0,
1622 					    param->opad_needed,
1623 					    ((param->kctx_len +
1624 					     sizeof(chcr_req->key_ctx)) >> 4));
1625 	chcr_req->sec_cpl.scmd1 = cpu_to_be64((u64)param->scmd1);
1626 	ulptx = (struct ulptx_sgl *)((u8 *)(chcr_req + 1) + param->kctx_len +
1627 				     DUMMY_BYTES);
1628 	if (param->bfr_len != 0) {
1629 		req_ctx->hctx_wr.dma_addr =
1630 			dma_map_single(&u_ctx->lldi.pdev->dev, req_ctx->reqbfr,
1631 				       param->bfr_len, DMA_TO_DEVICE);
1632 		if (dma_mapping_error(&u_ctx->lldi.pdev->dev,
1633 				       req_ctx->hctx_wr. dma_addr)) {
1634 			error = -ENOMEM;
1635 			goto err;
1636 		}
1637 		req_ctx->hctx_wr.dma_len = param->bfr_len;
1638 	} else {
1639 		req_ctx->hctx_wr.dma_addr = 0;
1640 	}
1641 	chcr_add_hash_src_ent(req, ulptx, param);
1642 	/* Request upto max wr size */
1643 	temp = param->kctx_len + DUMMY_BYTES + (req_ctx->hctx_wr.imm ?
1644 				(param->sg_len + param->bfr_len) : 0);
1645 	atomic_inc(&adap->chcr_stats.digest_rqst);
1646 	create_wreq(h_ctx(tfm), chcr_req, &req->base, req_ctx->hctx_wr.imm,
1647 		    param->hash_size, transhdr_len,
1648 		    temp,  0);
1649 	req_ctx->hctx_wr.skb = skb;
1650 	return skb;
1651 err:
1652 	kfree_skb(skb);
1653 	return  ERR_PTR(error);
1654 }
1655 
1656 static int chcr_ahash_update(struct ahash_request *req)
1657 {
1658 	struct chcr_ahash_req_ctx *req_ctx = ahash_request_ctx(req);
1659 	struct crypto_ahash *rtfm = crypto_ahash_reqtfm(req);
1660 	struct uld_ctx *u_ctx = ULD_CTX(h_ctx(rtfm));
1661 	struct chcr_context *ctx = h_ctx(rtfm);
1662 	struct chcr_dev *dev = h_ctx(rtfm)->dev;
1663 	struct sk_buff *skb;
1664 	u8 remainder = 0, bs;
1665 	unsigned int nbytes = req->nbytes;
1666 	struct hash_wr_param params;
1667 	int error;
1668 	unsigned int cpu;
1669 
1670 	cpu = get_cpu();
1671 	req_ctx->txqidx = cpu % ctx->ntxq;
1672 	req_ctx->rxqidx = cpu % ctx->nrxq;
1673 	put_cpu();
1674 
1675 	bs = crypto_tfm_alg_blocksize(crypto_ahash_tfm(rtfm));
1676 
1677 	if (nbytes + req_ctx->reqlen >= bs) {
1678 		remainder = (nbytes + req_ctx->reqlen) % bs;
1679 		nbytes = nbytes + req_ctx->reqlen - remainder;
1680 	} else {
1681 		sg_pcopy_to_buffer(req->src, sg_nents(req->src), req_ctx->reqbfr
1682 				   + req_ctx->reqlen, nbytes, 0);
1683 		req_ctx->reqlen += nbytes;
1684 		return 0;
1685 	}
1686 	error = chcr_inc_wrcount(dev);
1687 	if (error)
1688 		return -ENXIO;
1689 	/* Detach state for CHCR means lldi or padap is freed. Increasing
1690 	 * inflight count for dev guarantees that lldi and padap is valid
1691 	 */
1692 	if (unlikely(cxgb4_is_crypto_q_full(u_ctx->lldi.ports[0],
1693 						req_ctx->txqidx) &&
1694 		(!(req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG)))) {
1695 			error = -ENOSPC;
1696 			goto err;
1697 	}
1698 
1699 	chcr_init_hctx_per_wr(req_ctx);
1700 	error = chcr_hash_dma_map(&u_ctx->lldi.pdev->dev, req);
1701 	if (error) {
1702 		error = -ENOMEM;
1703 		goto err;
1704 	}
1705 	get_alg_config(&params.alg_prm, crypto_ahash_digestsize(rtfm));
1706 	params.kctx_len = roundup(params.alg_prm.result_size, 16);
1707 	params.sg_len = chcr_hash_ent_in_wr(req->src, !!req_ctx->reqlen,
1708 				     HASH_SPACE_LEFT(params.kctx_len), 0);
1709 	if (params.sg_len > req->nbytes)
1710 		params.sg_len = req->nbytes;
1711 	params.sg_len = rounddown(params.sg_len + req_ctx->reqlen, bs) -
1712 			req_ctx->reqlen;
1713 	params.opad_needed = 0;
1714 	params.more = 1;
1715 	params.last = 0;
1716 	params.bfr_len = req_ctx->reqlen;
1717 	params.scmd1 = 0;
1718 	req_ctx->hctx_wr.srcsg = req->src;
1719 
1720 	params.hash_size = params.alg_prm.result_size;
1721 	req_ctx->data_len += params.sg_len + params.bfr_len;
1722 	skb = create_hash_wr(req, &params);
1723 	if (IS_ERR(skb)) {
1724 		error = PTR_ERR(skb);
1725 		goto unmap;
1726 	}
1727 
1728 	req_ctx->hctx_wr.processed += params.sg_len;
1729 	if (remainder) {
1730 		/* Swap buffers */
1731 		swap(req_ctx->reqbfr, req_ctx->skbfr);
1732 		sg_pcopy_to_buffer(req->src, sg_nents(req->src),
1733 				   req_ctx->reqbfr, remainder, req->nbytes -
1734 				   remainder);
1735 	}
1736 	req_ctx->reqlen = remainder;
1737 	skb->dev = u_ctx->lldi.ports[0];
1738 	set_wr_txq(skb, CPL_PRIORITY_DATA, req_ctx->txqidx);
1739 	chcr_send_wr(skb);
1740 	return -EINPROGRESS;
1741 unmap:
1742 	chcr_hash_dma_unmap(&u_ctx->lldi.pdev->dev, req);
1743 err:
1744 	chcr_dec_wrcount(dev);
1745 	return error;
1746 }
1747 
1748 static void create_last_hash_block(char *bfr_ptr, unsigned int bs, u64 scmd1)
1749 {
1750 	memset(bfr_ptr, 0, bs);
1751 	*bfr_ptr = 0x80;
1752 	if (bs == 64)
1753 		*(__be64 *)(bfr_ptr + 56) = cpu_to_be64(scmd1  << 3);
1754 	else
1755 		*(__be64 *)(bfr_ptr + 120) =  cpu_to_be64(scmd1  << 3);
1756 }
1757 
1758 static int chcr_ahash_final(struct ahash_request *req)
1759 {
1760 	struct chcr_ahash_req_ctx *req_ctx = ahash_request_ctx(req);
1761 	struct crypto_ahash *rtfm = crypto_ahash_reqtfm(req);
1762 	struct chcr_dev *dev = h_ctx(rtfm)->dev;
1763 	struct hash_wr_param params;
1764 	struct sk_buff *skb;
1765 	struct uld_ctx *u_ctx = ULD_CTX(h_ctx(rtfm));
1766 	struct chcr_context *ctx = h_ctx(rtfm);
1767 	u8 bs = crypto_tfm_alg_blocksize(crypto_ahash_tfm(rtfm));
1768 	int error;
1769 	unsigned int cpu;
1770 
1771 	cpu = get_cpu();
1772 	req_ctx->txqidx = cpu % ctx->ntxq;
1773 	req_ctx->rxqidx = cpu % ctx->nrxq;
1774 	put_cpu();
1775 
1776 	error = chcr_inc_wrcount(dev);
1777 	if (error)
1778 		return -ENXIO;
1779 
1780 	chcr_init_hctx_per_wr(req_ctx);
1781 	if (is_hmac(crypto_ahash_tfm(rtfm)))
1782 		params.opad_needed = 1;
1783 	else
1784 		params.opad_needed = 0;
1785 	params.sg_len = 0;
1786 	req_ctx->hctx_wr.isfinal = 1;
1787 	get_alg_config(&params.alg_prm, crypto_ahash_digestsize(rtfm));
1788 	params.kctx_len = roundup(params.alg_prm.result_size, 16);
1789 	if (is_hmac(crypto_ahash_tfm(rtfm))) {
1790 		params.opad_needed = 1;
1791 		params.kctx_len *= 2;
1792 	} else {
1793 		params.opad_needed = 0;
1794 	}
1795 
1796 	req_ctx->hctx_wr.result = 1;
1797 	params.bfr_len = req_ctx->reqlen;
1798 	req_ctx->data_len += params.bfr_len + params.sg_len;
1799 	req_ctx->hctx_wr.srcsg = req->src;
1800 	if (req_ctx->reqlen == 0) {
1801 		create_last_hash_block(req_ctx->reqbfr, bs, req_ctx->data_len);
1802 		params.last = 0;
1803 		params.more = 1;
1804 		params.scmd1 = 0;
1805 		params.bfr_len = bs;
1806 
1807 	} else {
1808 		params.scmd1 = req_ctx->data_len;
1809 		params.last = 1;
1810 		params.more = 0;
1811 	}
1812 	params.hash_size = crypto_ahash_digestsize(rtfm);
1813 	skb = create_hash_wr(req, &params);
1814 	if (IS_ERR(skb)) {
1815 		error = PTR_ERR(skb);
1816 		goto err;
1817 	}
1818 	req_ctx->reqlen = 0;
1819 	skb->dev = u_ctx->lldi.ports[0];
1820 	set_wr_txq(skb, CPL_PRIORITY_DATA, req_ctx->txqidx);
1821 	chcr_send_wr(skb);
1822 	return -EINPROGRESS;
1823 err:
1824 	chcr_dec_wrcount(dev);
1825 	return error;
1826 }
1827 
1828 static int chcr_ahash_finup(struct ahash_request *req)
1829 {
1830 	struct chcr_ahash_req_ctx *req_ctx = ahash_request_ctx(req);
1831 	struct crypto_ahash *rtfm = crypto_ahash_reqtfm(req);
1832 	struct chcr_dev *dev = h_ctx(rtfm)->dev;
1833 	struct uld_ctx *u_ctx = ULD_CTX(h_ctx(rtfm));
1834 	struct chcr_context *ctx = h_ctx(rtfm);
1835 	struct sk_buff *skb;
1836 	struct hash_wr_param params;
1837 	u8  bs;
1838 	int error;
1839 	unsigned int cpu;
1840 
1841 	cpu = get_cpu();
1842 	req_ctx->txqidx = cpu % ctx->ntxq;
1843 	req_ctx->rxqidx = cpu % ctx->nrxq;
1844 	put_cpu();
1845 
1846 	bs = crypto_tfm_alg_blocksize(crypto_ahash_tfm(rtfm));
1847 	error = chcr_inc_wrcount(dev);
1848 	if (error)
1849 		return -ENXIO;
1850 
1851 	if (unlikely(cxgb4_is_crypto_q_full(u_ctx->lldi.ports[0],
1852 						req_ctx->txqidx) &&
1853 		(!(req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG)))) {
1854 			error = -ENOSPC;
1855 			goto err;
1856 	}
1857 	chcr_init_hctx_per_wr(req_ctx);
1858 	error = chcr_hash_dma_map(&u_ctx->lldi.pdev->dev, req);
1859 	if (error) {
1860 		error = -ENOMEM;
1861 		goto err;
1862 	}
1863 
1864 	get_alg_config(&params.alg_prm, crypto_ahash_digestsize(rtfm));
1865 	params.kctx_len = roundup(params.alg_prm.result_size, 16);
1866 	if (is_hmac(crypto_ahash_tfm(rtfm))) {
1867 		params.kctx_len *= 2;
1868 		params.opad_needed = 1;
1869 	} else {
1870 		params.opad_needed = 0;
1871 	}
1872 
1873 	params.sg_len = chcr_hash_ent_in_wr(req->src, !!req_ctx->reqlen,
1874 				    HASH_SPACE_LEFT(params.kctx_len), 0);
1875 	if (params.sg_len < req->nbytes) {
1876 		if (is_hmac(crypto_ahash_tfm(rtfm))) {
1877 			params.kctx_len /= 2;
1878 			params.opad_needed = 0;
1879 		}
1880 		params.last = 0;
1881 		params.more = 1;
1882 		params.sg_len = rounddown(params.sg_len + req_ctx->reqlen, bs)
1883 					- req_ctx->reqlen;
1884 		params.hash_size = params.alg_prm.result_size;
1885 		params.scmd1 = 0;
1886 	} else {
1887 		params.last = 1;
1888 		params.more = 0;
1889 		params.sg_len = req->nbytes;
1890 		params.hash_size = crypto_ahash_digestsize(rtfm);
1891 		params.scmd1 = req_ctx->data_len + req_ctx->reqlen +
1892 				params.sg_len;
1893 	}
1894 	params.bfr_len = req_ctx->reqlen;
1895 	req_ctx->data_len += params.bfr_len + params.sg_len;
1896 	req_ctx->hctx_wr.result = 1;
1897 	req_ctx->hctx_wr.srcsg = req->src;
1898 	if ((req_ctx->reqlen + req->nbytes) == 0) {
1899 		create_last_hash_block(req_ctx->reqbfr, bs, req_ctx->data_len);
1900 		params.last = 0;
1901 		params.more = 1;
1902 		params.scmd1 = 0;
1903 		params.bfr_len = bs;
1904 	}
1905 	skb = create_hash_wr(req, &params);
1906 	if (IS_ERR(skb)) {
1907 		error = PTR_ERR(skb);
1908 		goto unmap;
1909 	}
1910 	req_ctx->reqlen = 0;
1911 	req_ctx->hctx_wr.processed += params.sg_len;
1912 	skb->dev = u_ctx->lldi.ports[0];
1913 	set_wr_txq(skb, CPL_PRIORITY_DATA, req_ctx->txqidx);
1914 	chcr_send_wr(skb);
1915 	return -EINPROGRESS;
1916 unmap:
1917 	chcr_hash_dma_unmap(&u_ctx->lldi.pdev->dev, req);
1918 err:
1919 	chcr_dec_wrcount(dev);
1920 	return error;
1921 }
1922 
1923 static int chcr_hmac_init(struct ahash_request *areq);
1924 static int chcr_sha_init(struct ahash_request *areq);
1925 
1926 static int chcr_ahash_digest(struct ahash_request *req)
1927 {
1928 	struct chcr_ahash_req_ctx *req_ctx = ahash_request_ctx(req);
1929 	struct crypto_ahash *rtfm = crypto_ahash_reqtfm(req);
1930 	struct chcr_dev *dev = h_ctx(rtfm)->dev;
1931 	struct uld_ctx *u_ctx = ULD_CTX(h_ctx(rtfm));
1932 	struct chcr_context *ctx = h_ctx(rtfm);
1933 	struct sk_buff *skb;
1934 	struct hash_wr_param params;
1935 	u8  bs;
1936 	int error;
1937 	unsigned int cpu;
1938 
1939 	cpu = get_cpu();
1940 	req_ctx->txqidx = cpu % ctx->ntxq;
1941 	req_ctx->rxqidx = cpu % ctx->nrxq;
1942 	put_cpu();
1943 
1944 	if (is_hmac(crypto_ahash_tfm(rtfm)))
1945 		chcr_hmac_init(req);
1946 	else
1947 		chcr_sha_init(req);
1948 
1949 	bs = crypto_tfm_alg_blocksize(crypto_ahash_tfm(rtfm));
1950 	error = chcr_inc_wrcount(dev);
1951 	if (error)
1952 		return -ENXIO;
1953 
1954 	if (unlikely(cxgb4_is_crypto_q_full(u_ctx->lldi.ports[0],
1955 						req_ctx->txqidx) &&
1956 		(!(req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG)))) {
1957 			error = -ENOSPC;
1958 			goto err;
1959 	}
1960 
1961 	chcr_init_hctx_per_wr(req_ctx);
1962 	error = chcr_hash_dma_map(&u_ctx->lldi.pdev->dev, req);
1963 	if (error) {
1964 		error = -ENOMEM;
1965 		goto err;
1966 	}
1967 
1968 	get_alg_config(&params.alg_prm, crypto_ahash_digestsize(rtfm));
1969 	params.kctx_len = roundup(params.alg_prm.result_size, 16);
1970 	if (is_hmac(crypto_ahash_tfm(rtfm))) {
1971 		params.kctx_len *= 2;
1972 		params.opad_needed = 1;
1973 	} else {
1974 		params.opad_needed = 0;
1975 	}
1976 	params.sg_len = chcr_hash_ent_in_wr(req->src, !!req_ctx->reqlen,
1977 				HASH_SPACE_LEFT(params.kctx_len), 0);
1978 	if (params.sg_len < req->nbytes) {
1979 		if (is_hmac(crypto_ahash_tfm(rtfm))) {
1980 			params.kctx_len /= 2;
1981 			params.opad_needed = 0;
1982 		}
1983 		params.last = 0;
1984 		params.more = 1;
1985 		params.scmd1 = 0;
1986 		params.sg_len = rounddown(params.sg_len, bs);
1987 		params.hash_size = params.alg_prm.result_size;
1988 	} else {
1989 		params.sg_len = req->nbytes;
1990 		params.hash_size = crypto_ahash_digestsize(rtfm);
1991 		params.last = 1;
1992 		params.more = 0;
1993 		params.scmd1 = req->nbytes + req_ctx->data_len;
1994 
1995 	}
1996 	params.bfr_len = 0;
1997 	req_ctx->hctx_wr.result = 1;
1998 	req_ctx->hctx_wr.srcsg = req->src;
1999 	req_ctx->data_len += params.bfr_len + params.sg_len;
2000 
2001 	if (req->nbytes == 0) {
2002 		create_last_hash_block(req_ctx->reqbfr, bs, req_ctx->data_len);
2003 		params.more = 1;
2004 		params.bfr_len = bs;
2005 	}
2006 
2007 	skb = create_hash_wr(req, &params);
2008 	if (IS_ERR(skb)) {
2009 		error = PTR_ERR(skb);
2010 		goto unmap;
2011 	}
2012 	req_ctx->hctx_wr.processed += params.sg_len;
2013 	skb->dev = u_ctx->lldi.ports[0];
2014 	set_wr_txq(skb, CPL_PRIORITY_DATA, req_ctx->txqidx);
2015 	chcr_send_wr(skb);
2016 	return -EINPROGRESS;
2017 unmap:
2018 	chcr_hash_dma_unmap(&u_ctx->lldi.pdev->dev, req);
2019 err:
2020 	chcr_dec_wrcount(dev);
2021 	return error;
2022 }
2023 
2024 static int chcr_ahash_continue(struct ahash_request *req)
2025 {
2026 	struct chcr_ahash_req_ctx *reqctx = ahash_request_ctx(req);
2027 	struct chcr_hctx_per_wr *hctx_wr = &reqctx->hctx_wr;
2028 	struct crypto_ahash *rtfm = crypto_ahash_reqtfm(req);
2029 	struct chcr_context *ctx = h_ctx(rtfm);
2030 	struct uld_ctx *u_ctx = ULD_CTX(ctx);
2031 	struct sk_buff *skb;
2032 	struct hash_wr_param params;
2033 	u8  bs;
2034 	int error;
2035 	unsigned int cpu;
2036 
2037 	cpu = get_cpu();
2038 	reqctx->txqidx = cpu % ctx->ntxq;
2039 	reqctx->rxqidx = cpu % ctx->nrxq;
2040 	put_cpu();
2041 
2042 	bs = crypto_tfm_alg_blocksize(crypto_ahash_tfm(rtfm));
2043 	get_alg_config(&params.alg_prm, crypto_ahash_digestsize(rtfm));
2044 	params.kctx_len = roundup(params.alg_prm.result_size, 16);
2045 	if (is_hmac(crypto_ahash_tfm(rtfm))) {
2046 		params.kctx_len *= 2;
2047 		params.opad_needed = 1;
2048 	} else {
2049 		params.opad_needed = 0;
2050 	}
2051 	params.sg_len = chcr_hash_ent_in_wr(hctx_wr->srcsg, 0,
2052 					    HASH_SPACE_LEFT(params.kctx_len),
2053 					    hctx_wr->src_ofst);
2054 	if ((params.sg_len + hctx_wr->processed) > req->nbytes)
2055 		params.sg_len = req->nbytes - hctx_wr->processed;
2056 	if (!hctx_wr->result ||
2057 	    ((params.sg_len + hctx_wr->processed) < req->nbytes)) {
2058 		if (is_hmac(crypto_ahash_tfm(rtfm))) {
2059 			params.kctx_len /= 2;
2060 			params.opad_needed = 0;
2061 		}
2062 		params.last = 0;
2063 		params.more = 1;
2064 		params.sg_len = rounddown(params.sg_len, bs);
2065 		params.hash_size = params.alg_prm.result_size;
2066 		params.scmd1 = 0;
2067 	} else {
2068 		params.last = 1;
2069 		params.more = 0;
2070 		params.hash_size = crypto_ahash_digestsize(rtfm);
2071 		params.scmd1 = reqctx->data_len + params.sg_len;
2072 	}
2073 	params.bfr_len = 0;
2074 	reqctx->data_len += params.sg_len;
2075 	skb = create_hash_wr(req, &params);
2076 	if (IS_ERR(skb)) {
2077 		error = PTR_ERR(skb);
2078 		goto err;
2079 	}
2080 	hctx_wr->processed += params.sg_len;
2081 	skb->dev = u_ctx->lldi.ports[0];
2082 	set_wr_txq(skb, CPL_PRIORITY_DATA, reqctx->txqidx);
2083 	chcr_send_wr(skb);
2084 	return 0;
2085 err:
2086 	return error;
2087 }
2088 
2089 static inline void chcr_handle_ahash_resp(struct ahash_request *req,
2090 					  unsigned char *input,
2091 					  int err)
2092 {
2093 	struct chcr_ahash_req_ctx *reqctx = ahash_request_ctx(req);
2094 	struct chcr_hctx_per_wr *hctx_wr = &reqctx->hctx_wr;
2095 	int digestsize, updated_digestsize;
2096 	struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
2097 	struct uld_ctx *u_ctx = ULD_CTX(h_ctx(tfm));
2098 	struct chcr_dev *dev = h_ctx(tfm)->dev;
2099 
2100 	if (input == NULL)
2101 		goto out;
2102 	digestsize = crypto_ahash_digestsize(crypto_ahash_reqtfm(req));
2103 	updated_digestsize = digestsize;
2104 	if (digestsize == SHA224_DIGEST_SIZE)
2105 		updated_digestsize = SHA256_DIGEST_SIZE;
2106 	else if (digestsize == SHA384_DIGEST_SIZE)
2107 		updated_digestsize = SHA512_DIGEST_SIZE;
2108 
2109 	if (hctx_wr->dma_addr) {
2110 		dma_unmap_single(&u_ctx->lldi.pdev->dev, hctx_wr->dma_addr,
2111 				 hctx_wr->dma_len, DMA_TO_DEVICE);
2112 		hctx_wr->dma_addr = 0;
2113 	}
2114 	if (hctx_wr->isfinal || ((hctx_wr->processed + reqctx->reqlen) ==
2115 				 req->nbytes)) {
2116 		if (hctx_wr->result == 1) {
2117 			hctx_wr->result = 0;
2118 			memcpy(req->result, input + sizeof(struct cpl_fw6_pld),
2119 			       digestsize);
2120 		} else {
2121 			memcpy(reqctx->partial_hash,
2122 			       input + sizeof(struct cpl_fw6_pld),
2123 			       updated_digestsize);
2124 
2125 		}
2126 		goto unmap;
2127 	}
2128 	memcpy(reqctx->partial_hash, input + sizeof(struct cpl_fw6_pld),
2129 	       updated_digestsize);
2130 
2131 	err = chcr_ahash_continue(req);
2132 	if (err)
2133 		goto unmap;
2134 	return;
2135 unmap:
2136 	if (hctx_wr->is_sg_map)
2137 		chcr_hash_dma_unmap(&u_ctx->lldi.pdev->dev, req);
2138 
2139 
2140 out:
2141 	chcr_dec_wrcount(dev);
2142 	ahash_request_complete(req, err);
2143 }
2144 
2145 /*
2146  *	chcr_handle_resp - Unmap the DMA buffers associated with the request
2147  *	@req: crypto request
2148  */
2149 int chcr_handle_resp(struct crypto_async_request *req, unsigned char *input,
2150 			 int err)
2151 {
2152 	struct crypto_tfm *tfm = req->tfm;
2153 	struct chcr_context *ctx = crypto_tfm_ctx(tfm);
2154 	struct adapter *adap = padap(ctx->dev);
2155 
2156 	switch (tfm->__crt_alg->cra_flags & CRYPTO_ALG_TYPE_MASK) {
2157 	case CRYPTO_ALG_TYPE_AEAD:
2158 		err = chcr_handle_aead_resp(aead_request_cast(req), input, err);
2159 		break;
2160 
2161 	case CRYPTO_ALG_TYPE_SKCIPHER:
2162 		 chcr_handle_cipher_resp(skcipher_request_cast(req),
2163 					       input, err);
2164 		break;
2165 	case CRYPTO_ALG_TYPE_AHASH:
2166 		chcr_handle_ahash_resp(ahash_request_cast(req), input, err);
2167 		}
2168 	atomic_inc(&adap->chcr_stats.complete);
2169 	return err;
2170 }
2171 static int chcr_ahash_export(struct ahash_request *areq, void *out)
2172 {
2173 	struct chcr_ahash_req_ctx *req_ctx = ahash_request_ctx(areq);
2174 	struct chcr_ahash_req_ctx *state = out;
2175 
2176 	state->reqlen = req_ctx->reqlen;
2177 	state->data_len = req_ctx->data_len;
2178 	memcpy(state->bfr1, req_ctx->reqbfr, req_ctx->reqlen);
2179 	memcpy(state->partial_hash, req_ctx->partial_hash,
2180 	       CHCR_HASH_MAX_DIGEST_SIZE);
2181 	chcr_init_hctx_per_wr(state);
2182 	return 0;
2183 }
2184 
2185 static int chcr_ahash_import(struct ahash_request *areq, const void *in)
2186 {
2187 	struct chcr_ahash_req_ctx *req_ctx = ahash_request_ctx(areq);
2188 	struct chcr_ahash_req_ctx *state = (struct chcr_ahash_req_ctx *)in;
2189 
2190 	req_ctx->reqlen = state->reqlen;
2191 	req_ctx->data_len = state->data_len;
2192 	req_ctx->reqbfr = req_ctx->bfr1;
2193 	req_ctx->skbfr = req_ctx->bfr2;
2194 	memcpy(req_ctx->bfr1, state->bfr1, CHCR_HASH_MAX_BLOCK_SIZE_128);
2195 	memcpy(req_ctx->partial_hash, state->partial_hash,
2196 	       CHCR_HASH_MAX_DIGEST_SIZE);
2197 	chcr_init_hctx_per_wr(req_ctx);
2198 	return 0;
2199 }
2200 
2201 static int chcr_ahash_setkey(struct crypto_ahash *tfm, const u8 *key,
2202 			     unsigned int keylen)
2203 {
2204 	struct hmac_ctx *hmacctx = HMAC_CTX(h_ctx(tfm));
2205 	unsigned int digestsize = crypto_ahash_digestsize(tfm);
2206 	unsigned int bs = crypto_tfm_alg_blocksize(crypto_ahash_tfm(tfm));
2207 	unsigned int i, err = 0, updated_digestsize;
2208 
2209 	SHASH_DESC_ON_STACK(shash, hmacctx->base_hash);
2210 
2211 	/* use the key to calculate the ipad and opad. ipad will sent with the
2212 	 * first request's data. opad will be sent with the final hash result
2213 	 * ipad in hmacctx->ipad and opad in hmacctx->opad location
2214 	 */
2215 	shash->tfm = hmacctx->base_hash;
2216 	if (keylen > bs) {
2217 		err = crypto_shash_digest(shash, key, keylen,
2218 					  hmacctx->ipad);
2219 		if (err)
2220 			goto out;
2221 		keylen = digestsize;
2222 	} else {
2223 		memcpy(hmacctx->ipad, key, keylen);
2224 	}
2225 	memset(hmacctx->ipad + keylen, 0, bs - keylen);
2226 	unsafe_memcpy(hmacctx->opad, hmacctx->ipad, bs,
2227 		      "fortified memcpy causes -Wrestrict warning");
2228 
2229 	for (i = 0; i < bs / sizeof(int); i++) {
2230 		*((unsigned int *)(&hmacctx->ipad) + i) ^= IPAD_DATA;
2231 		*((unsigned int *)(&hmacctx->opad) + i) ^= OPAD_DATA;
2232 	}
2233 
2234 	updated_digestsize = digestsize;
2235 	if (digestsize == SHA224_DIGEST_SIZE)
2236 		updated_digestsize = SHA256_DIGEST_SIZE;
2237 	else if (digestsize == SHA384_DIGEST_SIZE)
2238 		updated_digestsize = SHA512_DIGEST_SIZE;
2239 	err = chcr_compute_partial_hash(shash, hmacctx->ipad,
2240 					hmacctx->ipad, digestsize);
2241 	if (err)
2242 		goto out;
2243 	chcr_change_order(hmacctx->ipad, updated_digestsize);
2244 
2245 	err = chcr_compute_partial_hash(shash, hmacctx->opad,
2246 					hmacctx->opad, digestsize);
2247 	if (err)
2248 		goto out;
2249 	chcr_change_order(hmacctx->opad, updated_digestsize);
2250 out:
2251 	return err;
2252 }
2253 
2254 static int chcr_aes_xts_setkey(struct crypto_skcipher *cipher, const u8 *key,
2255 			       unsigned int key_len)
2256 {
2257 	struct ablk_ctx *ablkctx = ABLK_CTX(c_ctx(cipher));
2258 	unsigned short context_size = 0;
2259 	int err;
2260 
2261 	err = chcr_cipher_fallback_setkey(cipher, key, key_len);
2262 	if (err)
2263 		goto badkey_err;
2264 
2265 	memcpy(ablkctx->key, key, key_len);
2266 	ablkctx->enckey_len = key_len;
2267 	get_aes_decrypt_key(ablkctx->rrkey, ablkctx->key, key_len << 2);
2268 	context_size = (KEY_CONTEXT_HDR_SALT_AND_PAD + key_len) >> 4;
2269 	/* Both keys for xts must be aligned to 16 byte boundary
2270 	 * by padding with zeros. So for 24 byte keys padding 8 zeroes.
2271 	 */
2272 	if (key_len == 48) {
2273 		context_size = (KEY_CONTEXT_HDR_SALT_AND_PAD + key_len
2274 				+ 16) >> 4;
2275 		memmove(ablkctx->key + 32, ablkctx->key + 24, 24);
2276 		memset(ablkctx->key + 24, 0, 8);
2277 		memset(ablkctx->key + 56, 0, 8);
2278 		ablkctx->enckey_len = 64;
2279 		ablkctx->key_ctx_hdr =
2280 			FILL_KEY_CTX_HDR(CHCR_KEYCTX_CIPHER_KEY_SIZE_192,
2281 					 CHCR_KEYCTX_NO_KEY, 1,
2282 					 0, context_size);
2283 	} else {
2284 		ablkctx->key_ctx_hdr =
2285 		FILL_KEY_CTX_HDR((key_len == AES_KEYSIZE_256) ?
2286 				 CHCR_KEYCTX_CIPHER_KEY_SIZE_128 :
2287 				 CHCR_KEYCTX_CIPHER_KEY_SIZE_256,
2288 				 CHCR_KEYCTX_NO_KEY, 1,
2289 				 0, context_size);
2290 	}
2291 	ablkctx->ciph_mode = CHCR_SCMD_CIPHER_MODE_AES_XTS;
2292 	return 0;
2293 badkey_err:
2294 	ablkctx->enckey_len = 0;
2295 
2296 	return err;
2297 }
2298 
2299 static int chcr_sha_init(struct ahash_request *areq)
2300 {
2301 	struct chcr_ahash_req_ctx *req_ctx = ahash_request_ctx(areq);
2302 	struct crypto_ahash *tfm = crypto_ahash_reqtfm(areq);
2303 	int digestsize =  crypto_ahash_digestsize(tfm);
2304 
2305 	req_ctx->data_len = 0;
2306 	req_ctx->reqlen = 0;
2307 	req_ctx->reqbfr = req_ctx->bfr1;
2308 	req_ctx->skbfr = req_ctx->bfr2;
2309 	copy_hash_init_values(req_ctx->partial_hash, digestsize);
2310 
2311 	return 0;
2312 }
2313 
2314 static int chcr_sha_cra_init(struct crypto_tfm *tfm)
2315 {
2316 	crypto_ahash_set_reqsize(__crypto_ahash_cast(tfm),
2317 				 sizeof(struct chcr_ahash_req_ctx));
2318 	return chcr_device_init(crypto_tfm_ctx(tfm));
2319 }
2320 
2321 static int chcr_hmac_init(struct ahash_request *areq)
2322 {
2323 	struct chcr_ahash_req_ctx *req_ctx = ahash_request_ctx(areq);
2324 	struct crypto_ahash *rtfm = crypto_ahash_reqtfm(areq);
2325 	struct hmac_ctx *hmacctx = HMAC_CTX(h_ctx(rtfm));
2326 	unsigned int digestsize = crypto_ahash_digestsize(rtfm);
2327 	unsigned int bs = crypto_tfm_alg_blocksize(crypto_ahash_tfm(rtfm));
2328 
2329 	chcr_sha_init(areq);
2330 	req_ctx->data_len = bs;
2331 	if (is_hmac(crypto_ahash_tfm(rtfm))) {
2332 		if (digestsize == SHA224_DIGEST_SIZE)
2333 			memcpy(req_ctx->partial_hash, hmacctx->ipad,
2334 			       SHA256_DIGEST_SIZE);
2335 		else if (digestsize == SHA384_DIGEST_SIZE)
2336 			memcpy(req_ctx->partial_hash, hmacctx->ipad,
2337 			       SHA512_DIGEST_SIZE);
2338 		else
2339 			memcpy(req_ctx->partial_hash, hmacctx->ipad,
2340 			       digestsize);
2341 	}
2342 	return 0;
2343 }
2344 
2345 static int chcr_hmac_cra_init(struct crypto_tfm *tfm)
2346 {
2347 	struct chcr_context *ctx = crypto_tfm_ctx(tfm);
2348 	struct hmac_ctx *hmacctx = HMAC_CTX(ctx);
2349 	unsigned int digestsize =
2350 		crypto_ahash_digestsize(__crypto_ahash_cast(tfm));
2351 
2352 	crypto_ahash_set_reqsize(__crypto_ahash_cast(tfm),
2353 				 sizeof(struct chcr_ahash_req_ctx));
2354 	hmacctx->base_hash = chcr_alloc_shash(digestsize);
2355 	if (IS_ERR(hmacctx->base_hash))
2356 		return PTR_ERR(hmacctx->base_hash);
2357 	return chcr_device_init(crypto_tfm_ctx(tfm));
2358 }
2359 
2360 static void chcr_hmac_cra_exit(struct crypto_tfm *tfm)
2361 {
2362 	struct chcr_context *ctx = crypto_tfm_ctx(tfm);
2363 	struct hmac_ctx *hmacctx = HMAC_CTX(ctx);
2364 
2365 	if (hmacctx->base_hash) {
2366 		chcr_free_shash(hmacctx->base_hash);
2367 		hmacctx->base_hash = NULL;
2368 	}
2369 }
2370 
2371 inline void chcr_aead_common_exit(struct aead_request *req)
2372 {
2373 	struct chcr_aead_reqctx *reqctx = aead_request_ctx_dma(req);
2374 	struct crypto_aead *tfm = crypto_aead_reqtfm(req);
2375 	struct uld_ctx *u_ctx = ULD_CTX(a_ctx(tfm));
2376 
2377 	chcr_aead_dma_unmap(&u_ctx->lldi.pdev->dev, req, reqctx->op);
2378 }
2379 
2380 static int chcr_aead_common_init(struct aead_request *req)
2381 {
2382 	struct crypto_aead *tfm = crypto_aead_reqtfm(req);
2383 	struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
2384 	struct chcr_aead_reqctx *reqctx = aead_request_ctx_dma(req);
2385 	unsigned int authsize = crypto_aead_authsize(tfm);
2386 	int error = -EINVAL;
2387 
2388 	/* validate key size */
2389 	if (aeadctx->enckey_len == 0)
2390 		goto err;
2391 	if (reqctx->op && req->cryptlen < authsize)
2392 		goto err;
2393 	if (reqctx->b0_len)
2394 		reqctx->scratch_pad = reqctx->iv + IV;
2395 	else
2396 		reqctx->scratch_pad = NULL;
2397 
2398 	error = chcr_aead_dma_map(&ULD_CTX(a_ctx(tfm))->lldi.pdev->dev, req,
2399 				  reqctx->op);
2400 	if (error) {
2401 		error = -ENOMEM;
2402 		goto err;
2403 	}
2404 
2405 	return 0;
2406 err:
2407 	return error;
2408 }
2409 
2410 static int chcr_aead_need_fallback(struct aead_request *req, int dst_nents,
2411 				   int aadmax, int wrlen,
2412 				   unsigned short op_type)
2413 {
2414 	unsigned int authsize = crypto_aead_authsize(crypto_aead_reqtfm(req));
2415 
2416 	if (((req->cryptlen - (op_type ? authsize : 0)) == 0) ||
2417 	    dst_nents > MAX_DSGL_ENT ||
2418 	    (req->assoclen > aadmax) ||
2419 	    (wrlen > SGE_MAX_WR_LEN))
2420 		return 1;
2421 	return 0;
2422 }
2423 
2424 static int chcr_aead_fallback(struct aead_request *req, unsigned short op_type)
2425 {
2426 	struct crypto_aead *tfm = crypto_aead_reqtfm(req);
2427 	struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
2428 	struct aead_request *subreq = aead_request_ctx_dma(req);
2429 
2430 	aead_request_set_tfm(subreq, aeadctx->sw_cipher);
2431 	aead_request_set_callback(subreq, req->base.flags,
2432 				  req->base.complete, req->base.data);
2433 	aead_request_set_crypt(subreq, req->src, req->dst, req->cryptlen,
2434 				 req->iv);
2435 	aead_request_set_ad(subreq, req->assoclen);
2436 	return op_type ? crypto_aead_decrypt(subreq) :
2437 		crypto_aead_encrypt(subreq);
2438 }
2439 
2440 static struct sk_buff *create_authenc_wr(struct aead_request *req,
2441 					 unsigned short qid,
2442 					 int size)
2443 {
2444 	struct crypto_aead *tfm = crypto_aead_reqtfm(req);
2445 	struct chcr_context *ctx = a_ctx(tfm);
2446 	struct uld_ctx *u_ctx = ULD_CTX(ctx);
2447 	struct chcr_aead_ctx *aeadctx = AEAD_CTX(ctx);
2448 	struct chcr_authenc_ctx *actx = AUTHENC_CTX(aeadctx);
2449 	struct chcr_aead_reqctx *reqctx = aead_request_ctx_dma(req);
2450 	struct sk_buff *skb = NULL;
2451 	struct chcr_wr *chcr_req;
2452 	struct cpl_rx_phys_dsgl *phys_cpl;
2453 	struct ulptx_sgl *ulptx;
2454 	unsigned int transhdr_len;
2455 	unsigned int dst_size = 0, temp, subtype = get_aead_subtype(tfm);
2456 	unsigned int   kctx_len = 0, dnents, snents;
2457 	unsigned int  authsize = crypto_aead_authsize(tfm);
2458 	int error = -EINVAL;
2459 	u8 *ivptr;
2460 	int null = 0;
2461 	gfp_t flags = req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP ? GFP_KERNEL :
2462 		GFP_ATOMIC;
2463 	struct adapter *adap = padap(ctx->dev);
2464 	unsigned int rx_channel_id = reqctx->rxqidx / ctx->rxq_perchan;
2465 
2466 	rx_channel_id = cxgb4_port_e2cchan(u_ctx->lldi.ports[rx_channel_id]);
2467 	if (req->cryptlen == 0)
2468 		return NULL;
2469 
2470 	reqctx->b0_len = 0;
2471 	error = chcr_aead_common_init(req);
2472 	if (error)
2473 		return ERR_PTR(error);
2474 
2475 	if (subtype == CRYPTO_ALG_SUB_TYPE_CBC_NULL ||
2476 		subtype == CRYPTO_ALG_SUB_TYPE_CTR_NULL) {
2477 		null = 1;
2478 	}
2479 	dnents = sg_nents_xlen(req->dst, req->assoclen + req->cryptlen +
2480 		(reqctx->op ? -authsize : authsize), CHCR_DST_SG_SIZE, 0);
2481 	dnents += MIN_AUTH_SG; // For IV
2482 	snents = sg_nents_xlen(req->src, req->assoclen + req->cryptlen,
2483 			       CHCR_SRC_SG_SIZE, 0);
2484 	dst_size = get_space_for_phys_dsgl(dnents);
2485 	kctx_len = (KEY_CONTEXT_CTX_LEN_G(ntohl(aeadctx->key_ctx_hdr)) << 4)
2486 		- sizeof(chcr_req->key_ctx);
2487 	transhdr_len = CIPHER_TRANSHDR_SIZE(kctx_len, dst_size);
2488 	reqctx->imm = (transhdr_len + req->assoclen + req->cryptlen) <
2489 			SGE_MAX_WR_LEN;
2490 	temp = reqctx->imm ? roundup(req->assoclen + req->cryptlen, 16)
2491 			: (sgl_len(snents) * 8);
2492 	transhdr_len += temp;
2493 	transhdr_len = roundup(transhdr_len, 16);
2494 
2495 	if (chcr_aead_need_fallback(req, dnents, T6_MAX_AAD_SIZE,
2496 				    transhdr_len, reqctx->op)) {
2497 		atomic_inc(&adap->chcr_stats.fallback);
2498 		chcr_aead_common_exit(req);
2499 		return ERR_PTR(chcr_aead_fallback(req, reqctx->op));
2500 	}
2501 	skb = alloc_skb(transhdr_len, flags);
2502 	if (!skb) {
2503 		error = -ENOMEM;
2504 		goto err;
2505 	}
2506 
2507 	chcr_req = __skb_put_zero(skb, transhdr_len);
2508 
2509 	temp  = (reqctx->op == CHCR_ENCRYPT_OP) ? 0 : authsize;
2510 
2511 	/*
2512 	 * Input order	is AAD,IV and Payload. where IV should be included as
2513 	 * the part of authdata. All other fields should be filled according
2514 	 * to the hardware spec
2515 	 */
2516 	chcr_req->sec_cpl.op_ivinsrtofst =
2517 				FILL_SEC_CPL_OP_IVINSR(rx_channel_id, 2, 1);
2518 	chcr_req->sec_cpl.pldlen = htonl(req->assoclen + IV + req->cryptlen);
2519 	chcr_req->sec_cpl.aadstart_cipherstop_hi = FILL_SEC_CPL_CIPHERSTOP_HI(
2520 					null ? 0 : 1 + IV,
2521 					null ? 0 : IV + req->assoclen,
2522 					req->assoclen + IV + 1,
2523 					(temp & 0x1F0) >> 4);
2524 	chcr_req->sec_cpl.cipherstop_lo_authinsert = FILL_SEC_CPL_AUTHINSERT(
2525 					temp & 0xF,
2526 					null ? 0 : req->assoclen + IV + 1,
2527 					temp, temp);
2528 	if (subtype == CRYPTO_ALG_SUB_TYPE_CTR_NULL ||
2529 	    subtype == CRYPTO_ALG_SUB_TYPE_CTR_SHA)
2530 		temp = CHCR_SCMD_CIPHER_MODE_AES_CTR;
2531 	else
2532 		temp = CHCR_SCMD_CIPHER_MODE_AES_CBC;
2533 	chcr_req->sec_cpl.seqno_numivs = FILL_SEC_CPL_SCMD0_SEQNO(reqctx->op,
2534 					(reqctx->op == CHCR_ENCRYPT_OP) ? 1 : 0,
2535 					temp,
2536 					actx->auth_mode, aeadctx->hmac_ctrl,
2537 					IV >> 1);
2538 	chcr_req->sec_cpl.ivgen_hdrlen =  FILL_SEC_CPL_IVGEN_HDRLEN(0, 0, 1,
2539 					 0, 0, dst_size);
2540 
2541 	chcr_req->key_ctx.ctx_hdr = aeadctx->key_ctx_hdr;
2542 	if (reqctx->op == CHCR_ENCRYPT_OP ||
2543 		subtype == CRYPTO_ALG_SUB_TYPE_CTR_SHA ||
2544 		subtype == CRYPTO_ALG_SUB_TYPE_CTR_NULL)
2545 		memcpy(chcr_req->key_ctx.key, aeadctx->key,
2546 		       aeadctx->enckey_len);
2547 	else
2548 		memcpy(chcr_req->key_ctx.key, actx->dec_rrkey,
2549 		       aeadctx->enckey_len);
2550 
2551 	memcpy(chcr_req->key_ctx.key + roundup(aeadctx->enckey_len, 16),
2552 	       actx->h_iopad, kctx_len - roundup(aeadctx->enckey_len, 16));
2553 	phys_cpl = (struct cpl_rx_phys_dsgl *)((u8 *)(chcr_req + 1) + kctx_len);
2554 	ivptr = (u8 *)(phys_cpl + 1) + dst_size;
2555 	ulptx = (struct ulptx_sgl *)(ivptr + IV);
2556 	if (subtype == CRYPTO_ALG_SUB_TYPE_CTR_SHA ||
2557 	    subtype == CRYPTO_ALG_SUB_TYPE_CTR_NULL) {
2558 		memcpy(ivptr, aeadctx->nonce, CTR_RFC3686_NONCE_SIZE);
2559 		memcpy(ivptr + CTR_RFC3686_NONCE_SIZE, req->iv,
2560 				CTR_RFC3686_IV_SIZE);
2561 		*(__be32 *)(ivptr + CTR_RFC3686_NONCE_SIZE +
2562 			CTR_RFC3686_IV_SIZE) = cpu_to_be32(1);
2563 	} else {
2564 		memcpy(ivptr, req->iv, IV);
2565 	}
2566 	chcr_add_aead_dst_ent(req, phys_cpl, qid);
2567 	chcr_add_aead_src_ent(req, ulptx);
2568 	atomic_inc(&adap->chcr_stats.cipher_rqst);
2569 	temp = sizeof(struct cpl_rx_phys_dsgl) + dst_size + IV +
2570 		kctx_len + (reqctx->imm ? (req->assoclen + req->cryptlen) : 0);
2571 	create_wreq(a_ctx(tfm), chcr_req, &req->base, reqctx->imm, size,
2572 		   transhdr_len, temp, 0);
2573 	reqctx->skb = skb;
2574 
2575 	return skb;
2576 err:
2577 	chcr_aead_common_exit(req);
2578 
2579 	return ERR_PTR(error);
2580 }
2581 
2582 int chcr_aead_dma_map(struct device *dev,
2583 		      struct aead_request *req,
2584 		      unsigned short op_type)
2585 {
2586 	int error;
2587 	struct chcr_aead_reqctx *reqctx = aead_request_ctx_dma(req);
2588 	struct crypto_aead *tfm = crypto_aead_reqtfm(req);
2589 	unsigned int authsize = crypto_aead_authsize(tfm);
2590 	int src_len, dst_len;
2591 
2592 	/* calculate and handle src and dst sg length separately
2593 	 * for inplace and out-of place operations
2594 	 */
2595 	if (req->src == req->dst) {
2596 		src_len = req->assoclen + req->cryptlen + (op_type ?
2597 							0 : authsize);
2598 		dst_len = src_len;
2599 	} else {
2600 		src_len = req->assoclen + req->cryptlen;
2601 		dst_len = req->assoclen + req->cryptlen + (op_type ?
2602 							-authsize : authsize);
2603 	}
2604 
2605 	if (!req->cryptlen || !src_len || !dst_len)
2606 		return 0;
2607 	reqctx->iv_dma = dma_map_single(dev, reqctx->iv, (IV + reqctx->b0_len),
2608 					DMA_BIDIRECTIONAL);
2609 	if (dma_mapping_error(dev, reqctx->iv_dma))
2610 		return -ENOMEM;
2611 	if (reqctx->b0_len)
2612 		reqctx->b0_dma = reqctx->iv_dma + IV;
2613 	else
2614 		reqctx->b0_dma = 0;
2615 	if (req->src == req->dst) {
2616 		error = dma_map_sg(dev, req->src,
2617 				sg_nents_for_len(req->src, src_len),
2618 					DMA_BIDIRECTIONAL);
2619 		if (!error)
2620 			goto err;
2621 	} else {
2622 		error = dma_map_sg(dev, req->src,
2623 				   sg_nents_for_len(req->src, src_len),
2624 				   DMA_TO_DEVICE);
2625 		if (!error)
2626 			goto err;
2627 		error = dma_map_sg(dev, req->dst,
2628 				   sg_nents_for_len(req->dst, dst_len),
2629 				   DMA_FROM_DEVICE);
2630 		if (!error) {
2631 			dma_unmap_sg(dev, req->src,
2632 				     sg_nents_for_len(req->src, src_len),
2633 				     DMA_TO_DEVICE);
2634 			goto err;
2635 		}
2636 	}
2637 
2638 	return 0;
2639 err:
2640 	dma_unmap_single(dev, reqctx->iv_dma, IV, DMA_BIDIRECTIONAL);
2641 	return -ENOMEM;
2642 }
2643 
2644 void chcr_aead_dma_unmap(struct device *dev,
2645 			 struct aead_request *req,
2646 			 unsigned short op_type)
2647 {
2648 	struct chcr_aead_reqctx *reqctx = aead_request_ctx_dma(req);
2649 	struct crypto_aead *tfm = crypto_aead_reqtfm(req);
2650 	unsigned int authsize = crypto_aead_authsize(tfm);
2651 	int src_len, dst_len;
2652 
2653 	/* calculate and handle src and dst sg length separately
2654 	 * for inplace and out-of place operations
2655 	 */
2656 	if (req->src == req->dst) {
2657 		src_len = req->assoclen + req->cryptlen + (op_type ?
2658 							0 : authsize);
2659 		dst_len = src_len;
2660 	} else {
2661 		src_len = req->assoclen + req->cryptlen;
2662 		dst_len = req->assoclen + req->cryptlen + (op_type ?
2663 						-authsize : authsize);
2664 	}
2665 
2666 	if (!req->cryptlen || !src_len || !dst_len)
2667 		return;
2668 
2669 	dma_unmap_single(dev, reqctx->iv_dma, (IV + reqctx->b0_len),
2670 					DMA_BIDIRECTIONAL);
2671 	if (req->src == req->dst) {
2672 		dma_unmap_sg(dev, req->src,
2673 			     sg_nents_for_len(req->src, src_len),
2674 			     DMA_BIDIRECTIONAL);
2675 	} else {
2676 		dma_unmap_sg(dev, req->src,
2677 			     sg_nents_for_len(req->src, src_len),
2678 			     DMA_TO_DEVICE);
2679 		dma_unmap_sg(dev, req->dst,
2680 			     sg_nents_for_len(req->dst, dst_len),
2681 			     DMA_FROM_DEVICE);
2682 	}
2683 }
2684 
2685 void chcr_add_aead_src_ent(struct aead_request *req,
2686 			   struct ulptx_sgl *ulptx)
2687 {
2688 	struct ulptx_walk ulp_walk;
2689 	struct chcr_aead_reqctx *reqctx = aead_request_ctx_dma(req);
2690 
2691 	if (reqctx->imm) {
2692 		u8 *buf = (u8 *)ulptx;
2693 
2694 		if (reqctx->b0_len) {
2695 			memcpy(buf, reqctx->scratch_pad, reqctx->b0_len);
2696 			buf += reqctx->b0_len;
2697 		}
2698 		sg_pcopy_to_buffer(req->src, sg_nents(req->src),
2699 				   buf, req->cryptlen + req->assoclen, 0);
2700 	} else {
2701 		ulptx_walk_init(&ulp_walk, ulptx);
2702 		if (reqctx->b0_len)
2703 			ulptx_walk_add_page(&ulp_walk, reqctx->b0_len,
2704 					    reqctx->b0_dma);
2705 		ulptx_walk_add_sg(&ulp_walk, req->src, req->cryptlen +
2706 				  req->assoclen,  0);
2707 		ulptx_walk_end(&ulp_walk);
2708 	}
2709 }
2710 
2711 void chcr_add_aead_dst_ent(struct aead_request *req,
2712 			   struct cpl_rx_phys_dsgl *phys_cpl,
2713 			   unsigned short qid)
2714 {
2715 	struct chcr_aead_reqctx *reqctx = aead_request_ctx_dma(req);
2716 	struct crypto_aead *tfm = crypto_aead_reqtfm(req);
2717 	struct dsgl_walk dsgl_walk;
2718 	unsigned int authsize = crypto_aead_authsize(tfm);
2719 	struct chcr_context *ctx = a_ctx(tfm);
2720 	struct uld_ctx *u_ctx = ULD_CTX(ctx);
2721 	u32 temp;
2722 	unsigned int rx_channel_id = reqctx->rxqidx / ctx->rxq_perchan;
2723 
2724 	rx_channel_id = cxgb4_port_e2cchan(u_ctx->lldi.ports[rx_channel_id]);
2725 	dsgl_walk_init(&dsgl_walk, phys_cpl);
2726 	dsgl_walk_add_page(&dsgl_walk, IV + reqctx->b0_len, reqctx->iv_dma);
2727 	temp = req->assoclen + req->cryptlen +
2728 		(reqctx->op ? -authsize : authsize);
2729 	dsgl_walk_add_sg(&dsgl_walk, req->dst, temp, 0);
2730 	dsgl_walk_end(&dsgl_walk, qid, rx_channel_id);
2731 }
2732 
2733 void chcr_add_cipher_src_ent(struct skcipher_request *req,
2734 			     void *ulptx,
2735 			     struct  cipher_wr_param *wrparam)
2736 {
2737 	struct ulptx_walk ulp_walk;
2738 	struct chcr_skcipher_req_ctx *reqctx = skcipher_request_ctx(req);
2739 	u8 *buf = ulptx;
2740 
2741 	memcpy(buf, reqctx->iv, IV);
2742 	buf += IV;
2743 	if (reqctx->imm) {
2744 		sg_pcopy_to_buffer(req->src, sg_nents(req->src),
2745 				   buf, wrparam->bytes, reqctx->processed);
2746 	} else {
2747 		ulptx_walk_init(&ulp_walk, (struct ulptx_sgl *)buf);
2748 		ulptx_walk_add_sg(&ulp_walk, reqctx->srcsg, wrparam->bytes,
2749 				  reqctx->src_ofst);
2750 		reqctx->srcsg = ulp_walk.last_sg;
2751 		reqctx->src_ofst = ulp_walk.last_sg_len;
2752 		ulptx_walk_end(&ulp_walk);
2753 	}
2754 }
2755 
2756 void chcr_add_cipher_dst_ent(struct skcipher_request *req,
2757 			     struct cpl_rx_phys_dsgl *phys_cpl,
2758 			     struct  cipher_wr_param *wrparam,
2759 			     unsigned short qid)
2760 {
2761 	struct chcr_skcipher_req_ctx *reqctx = skcipher_request_ctx(req);
2762 	struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(wrparam->req);
2763 	struct chcr_context *ctx = c_ctx(tfm);
2764 	struct uld_ctx *u_ctx = ULD_CTX(ctx);
2765 	struct dsgl_walk dsgl_walk;
2766 	unsigned int rx_channel_id = reqctx->rxqidx / ctx->rxq_perchan;
2767 
2768 	rx_channel_id = cxgb4_port_e2cchan(u_ctx->lldi.ports[rx_channel_id]);
2769 	dsgl_walk_init(&dsgl_walk, phys_cpl);
2770 	dsgl_walk_add_sg(&dsgl_walk, reqctx->dstsg, wrparam->bytes,
2771 			 reqctx->dst_ofst);
2772 	reqctx->dstsg = dsgl_walk.last_sg;
2773 	reqctx->dst_ofst = dsgl_walk.last_sg_len;
2774 	dsgl_walk_end(&dsgl_walk, qid, rx_channel_id);
2775 }
2776 
2777 void chcr_add_hash_src_ent(struct ahash_request *req,
2778 			   struct ulptx_sgl *ulptx,
2779 			   struct hash_wr_param *param)
2780 {
2781 	struct ulptx_walk ulp_walk;
2782 	struct chcr_ahash_req_ctx *reqctx = ahash_request_ctx(req);
2783 
2784 	if (reqctx->hctx_wr.imm) {
2785 		u8 *buf = (u8 *)ulptx;
2786 
2787 		if (param->bfr_len) {
2788 			memcpy(buf, reqctx->reqbfr, param->bfr_len);
2789 			buf += param->bfr_len;
2790 		}
2791 
2792 		sg_pcopy_to_buffer(reqctx->hctx_wr.srcsg,
2793 				   sg_nents(reqctx->hctx_wr.srcsg), buf,
2794 				   param->sg_len, 0);
2795 	} else {
2796 		ulptx_walk_init(&ulp_walk, ulptx);
2797 		if (param->bfr_len)
2798 			ulptx_walk_add_page(&ulp_walk, param->bfr_len,
2799 					    reqctx->hctx_wr.dma_addr);
2800 		ulptx_walk_add_sg(&ulp_walk, reqctx->hctx_wr.srcsg,
2801 				  param->sg_len, reqctx->hctx_wr.src_ofst);
2802 		reqctx->hctx_wr.srcsg = ulp_walk.last_sg;
2803 		reqctx->hctx_wr.src_ofst = ulp_walk.last_sg_len;
2804 		ulptx_walk_end(&ulp_walk);
2805 	}
2806 }
2807 
2808 int chcr_hash_dma_map(struct device *dev,
2809 		      struct ahash_request *req)
2810 {
2811 	struct chcr_ahash_req_ctx *req_ctx = ahash_request_ctx(req);
2812 	int error = 0;
2813 
2814 	if (!req->nbytes)
2815 		return 0;
2816 	error = dma_map_sg(dev, req->src, sg_nents(req->src),
2817 			   DMA_TO_DEVICE);
2818 	if (!error)
2819 		return -ENOMEM;
2820 	req_ctx->hctx_wr.is_sg_map = 1;
2821 	return 0;
2822 }
2823 
2824 void chcr_hash_dma_unmap(struct device *dev,
2825 			 struct ahash_request *req)
2826 {
2827 	struct chcr_ahash_req_ctx *req_ctx = ahash_request_ctx(req);
2828 
2829 	if (!req->nbytes)
2830 		return;
2831 
2832 	dma_unmap_sg(dev, req->src, sg_nents(req->src),
2833 			   DMA_TO_DEVICE);
2834 	req_ctx->hctx_wr.is_sg_map = 0;
2835 
2836 }
2837 
2838 int chcr_cipher_dma_map(struct device *dev,
2839 			struct skcipher_request *req)
2840 {
2841 	int error;
2842 
2843 	if (req->src == req->dst) {
2844 		error = dma_map_sg(dev, req->src, sg_nents(req->src),
2845 				   DMA_BIDIRECTIONAL);
2846 		if (!error)
2847 			goto err;
2848 	} else {
2849 		error = dma_map_sg(dev, req->src, sg_nents(req->src),
2850 				   DMA_TO_DEVICE);
2851 		if (!error)
2852 			goto err;
2853 		error = dma_map_sg(dev, req->dst, sg_nents(req->dst),
2854 				   DMA_FROM_DEVICE);
2855 		if (!error) {
2856 			dma_unmap_sg(dev, req->src, sg_nents(req->src),
2857 				   DMA_TO_DEVICE);
2858 			goto err;
2859 		}
2860 	}
2861 
2862 	return 0;
2863 err:
2864 	return -ENOMEM;
2865 }
2866 
2867 void chcr_cipher_dma_unmap(struct device *dev,
2868 			   struct skcipher_request *req)
2869 {
2870 	if (req->src == req->dst) {
2871 		dma_unmap_sg(dev, req->src, sg_nents(req->src),
2872 				   DMA_BIDIRECTIONAL);
2873 	} else {
2874 		dma_unmap_sg(dev, req->src, sg_nents(req->src),
2875 				   DMA_TO_DEVICE);
2876 		dma_unmap_sg(dev, req->dst, sg_nents(req->dst),
2877 				   DMA_FROM_DEVICE);
2878 	}
2879 }
2880 
2881 static int set_msg_len(u8 *block, unsigned int msglen, int csize)
2882 {
2883 	__be32 data;
2884 
2885 	memset(block, 0, csize);
2886 	block += csize;
2887 
2888 	if (csize >= 4)
2889 		csize = 4;
2890 	else if (msglen > (unsigned int)(1 << (8 * csize)))
2891 		return -EOVERFLOW;
2892 
2893 	data = cpu_to_be32(msglen);
2894 	memcpy(block - csize, (u8 *)&data + 4 - csize, csize);
2895 
2896 	return 0;
2897 }
2898 
2899 static int generate_b0(struct aead_request *req, u8 *ivptr,
2900 			unsigned short op_type)
2901 {
2902 	unsigned int l, lp, m;
2903 	int rc;
2904 	struct crypto_aead *aead = crypto_aead_reqtfm(req);
2905 	struct chcr_aead_reqctx *reqctx = aead_request_ctx_dma(req);
2906 	u8 *b0 = reqctx->scratch_pad;
2907 
2908 	m = crypto_aead_authsize(aead);
2909 
2910 	memcpy(b0, ivptr, 16);
2911 
2912 	lp = b0[0];
2913 	l = lp + 1;
2914 
2915 	/* set m, bits 3-5 */
2916 	*b0 |= (8 * ((m - 2) / 2));
2917 
2918 	/* set adata, bit 6, if associated data is used */
2919 	if (req->assoclen)
2920 		*b0 |= 64;
2921 	rc = set_msg_len(b0 + 16 - l,
2922 			 (op_type == CHCR_DECRYPT_OP) ?
2923 			 req->cryptlen - m : req->cryptlen, l);
2924 
2925 	return rc;
2926 }
2927 
2928 static inline int crypto_ccm_check_iv(const u8 *iv)
2929 {
2930 	/* 2 <= L <= 8, so 1 <= L' <= 7. */
2931 	if (iv[0] < 1 || iv[0] > 7)
2932 		return -EINVAL;
2933 
2934 	return 0;
2935 }
2936 
2937 static int ccm_format_packet(struct aead_request *req,
2938 			     u8 *ivptr,
2939 			     unsigned int sub_type,
2940 			     unsigned short op_type,
2941 			     unsigned int assoclen)
2942 {
2943 	struct chcr_aead_reqctx *reqctx = aead_request_ctx_dma(req);
2944 	struct crypto_aead *tfm = crypto_aead_reqtfm(req);
2945 	struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
2946 	int rc = 0;
2947 
2948 	if (sub_type == CRYPTO_ALG_SUB_TYPE_AEAD_RFC4309) {
2949 		ivptr[0] = 3;
2950 		memcpy(ivptr + 1, &aeadctx->salt[0], 3);
2951 		memcpy(ivptr + 4, req->iv, 8);
2952 		memset(ivptr + 12, 0, 4);
2953 	} else {
2954 		memcpy(ivptr, req->iv, 16);
2955 	}
2956 	if (assoclen)
2957 		put_unaligned_be16(assoclen, &reqctx->scratch_pad[16]);
2958 
2959 	rc = generate_b0(req, ivptr, op_type);
2960 	/* zero the ctr value */
2961 	memset(ivptr + 15 - ivptr[0], 0, ivptr[0] + 1);
2962 	return rc;
2963 }
2964 
2965 static void fill_sec_cpl_for_aead(struct cpl_tx_sec_pdu *sec_cpl,
2966 				  unsigned int dst_size,
2967 				  struct aead_request *req,
2968 				  unsigned short op_type)
2969 {
2970 	struct crypto_aead *tfm = crypto_aead_reqtfm(req);
2971 	struct chcr_context *ctx = a_ctx(tfm);
2972 	struct uld_ctx *u_ctx = ULD_CTX(ctx);
2973 	struct chcr_aead_ctx *aeadctx = AEAD_CTX(ctx);
2974 	struct chcr_aead_reqctx *reqctx = aead_request_ctx_dma(req);
2975 	unsigned int cipher_mode = CHCR_SCMD_CIPHER_MODE_AES_CCM;
2976 	unsigned int mac_mode = CHCR_SCMD_AUTH_MODE_CBCMAC;
2977 	unsigned int rx_channel_id = reqctx->rxqidx / ctx->rxq_perchan;
2978 	unsigned int ccm_xtra;
2979 	unsigned int tag_offset = 0, auth_offset = 0;
2980 	unsigned int assoclen;
2981 
2982 	rx_channel_id = cxgb4_port_e2cchan(u_ctx->lldi.ports[rx_channel_id]);
2983 
2984 	if (get_aead_subtype(tfm) == CRYPTO_ALG_SUB_TYPE_AEAD_RFC4309)
2985 		assoclen = req->assoclen - 8;
2986 	else
2987 		assoclen = req->assoclen;
2988 	ccm_xtra = CCM_B0_SIZE +
2989 		((assoclen) ? CCM_AAD_FIELD_SIZE : 0);
2990 
2991 	auth_offset = req->cryptlen ?
2992 		(req->assoclen + IV + 1 + ccm_xtra) : 0;
2993 	if (op_type == CHCR_DECRYPT_OP) {
2994 		if (crypto_aead_authsize(tfm) != req->cryptlen)
2995 			tag_offset = crypto_aead_authsize(tfm);
2996 		else
2997 			auth_offset = 0;
2998 	}
2999 
3000 	sec_cpl->op_ivinsrtofst = FILL_SEC_CPL_OP_IVINSR(rx_channel_id, 2, 1);
3001 	sec_cpl->pldlen =
3002 		htonl(req->assoclen + IV + req->cryptlen + ccm_xtra);
3003 	/* For CCM there wil be b0 always. So AAD start will be 1 always */
3004 	sec_cpl->aadstart_cipherstop_hi = FILL_SEC_CPL_CIPHERSTOP_HI(
3005 				1 + IV,	IV + assoclen + ccm_xtra,
3006 				req->assoclen + IV + 1 + ccm_xtra, 0);
3007 
3008 	sec_cpl->cipherstop_lo_authinsert = FILL_SEC_CPL_AUTHINSERT(0,
3009 					auth_offset, tag_offset,
3010 					(op_type == CHCR_ENCRYPT_OP) ? 0 :
3011 					crypto_aead_authsize(tfm));
3012 	sec_cpl->seqno_numivs =  FILL_SEC_CPL_SCMD0_SEQNO(op_type,
3013 					(op_type == CHCR_ENCRYPT_OP) ? 0 : 1,
3014 					cipher_mode, mac_mode,
3015 					aeadctx->hmac_ctrl, IV >> 1);
3016 
3017 	sec_cpl->ivgen_hdrlen = FILL_SEC_CPL_IVGEN_HDRLEN(0, 0, 1, 0,
3018 					0, dst_size);
3019 }
3020 
3021 static int aead_ccm_validate_input(unsigned short op_type,
3022 				   struct aead_request *req,
3023 				   struct chcr_aead_ctx *aeadctx,
3024 				   unsigned int sub_type)
3025 {
3026 	if (sub_type != CRYPTO_ALG_SUB_TYPE_AEAD_RFC4309) {
3027 		if (crypto_ccm_check_iv(req->iv)) {
3028 			pr_err("CCM: IV check fails\n");
3029 			return -EINVAL;
3030 		}
3031 	} else {
3032 		if (req->assoclen != 16 && req->assoclen != 20) {
3033 			pr_err("RFC4309: Invalid AAD length %d\n",
3034 			       req->assoclen);
3035 			return -EINVAL;
3036 		}
3037 	}
3038 	return 0;
3039 }
3040 
3041 static struct sk_buff *create_aead_ccm_wr(struct aead_request *req,
3042 					  unsigned short qid,
3043 					  int size)
3044 {
3045 	struct crypto_aead *tfm = crypto_aead_reqtfm(req);
3046 	struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
3047 	struct chcr_aead_reqctx *reqctx = aead_request_ctx_dma(req);
3048 	struct sk_buff *skb = NULL;
3049 	struct chcr_wr *chcr_req;
3050 	struct cpl_rx_phys_dsgl *phys_cpl;
3051 	struct ulptx_sgl *ulptx;
3052 	unsigned int transhdr_len;
3053 	unsigned int dst_size = 0, kctx_len, dnents, temp, snents;
3054 	unsigned int sub_type, assoclen = req->assoclen;
3055 	unsigned int authsize = crypto_aead_authsize(tfm);
3056 	int error = -EINVAL;
3057 	u8 *ivptr;
3058 	gfp_t flags = req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP ? GFP_KERNEL :
3059 		GFP_ATOMIC;
3060 	struct adapter *adap = padap(a_ctx(tfm)->dev);
3061 
3062 	sub_type = get_aead_subtype(tfm);
3063 	if (sub_type == CRYPTO_ALG_SUB_TYPE_AEAD_RFC4309)
3064 		assoclen -= 8;
3065 	reqctx->b0_len = CCM_B0_SIZE + (assoclen ? CCM_AAD_FIELD_SIZE : 0);
3066 	error = chcr_aead_common_init(req);
3067 	if (error)
3068 		return ERR_PTR(error);
3069 
3070 	error = aead_ccm_validate_input(reqctx->op, req, aeadctx, sub_type);
3071 	if (error)
3072 		goto err;
3073 	dnents = sg_nents_xlen(req->dst, req->assoclen + req->cryptlen
3074 			+ (reqctx->op ? -authsize : authsize),
3075 			CHCR_DST_SG_SIZE, 0);
3076 	dnents += MIN_CCM_SG; // For IV and B0
3077 	dst_size = get_space_for_phys_dsgl(dnents);
3078 	snents = sg_nents_xlen(req->src, req->assoclen + req->cryptlen,
3079 			       CHCR_SRC_SG_SIZE, 0);
3080 	snents += MIN_CCM_SG; //For B0
3081 	kctx_len = roundup(aeadctx->enckey_len, 16) * 2;
3082 	transhdr_len = CIPHER_TRANSHDR_SIZE(kctx_len, dst_size);
3083 	reqctx->imm = (transhdr_len + req->assoclen + req->cryptlen +
3084 		       reqctx->b0_len) <= SGE_MAX_WR_LEN;
3085 	temp = reqctx->imm ? roundup(req->assoclen + req->cryptlen +
3086 				     reqctx->b0_len, 16) :
3087 		(sgl_len(snents) *  8);
3088 	transhdr_len += temp;
3089 	transhdr_len = roundup(transhdr_len, 16);
3090 
3091 	if (chcr_aead_need_fallback(req, dnents, T6_MAX_AAD_SIZE -
3092 				reqctx->b0_len, transhdr_len, reqctx->op)) {
3093 		atomic_inc(&adap->chcr_stats.fallback);
3094 		chcr_aead_common_exit(req);
3095 		return ERR_PTR(chcr_aead_fallback(req, reqctx->op));
3096 	}
3097 	skb = alloc_skb(transhdr_len,  flags);
3098 
3099 	if (!skb) {
3100 		error = -ENOMEM;
3101 		goto err;
3102 	}
3103 
3104 	chcr_req = __skb_put_zero(skb, transhdr_len);
3105 
3106 	fill_sec_cpl_for_aead(&chcr_req->sec_cpl, dst_size, req, reqctx->op);
3107 
3108 	chcr_req->key_ctx.ctx_hdr = aeadctx->key_ctx_hdr;
3109 	memcpy(chcr_req->key_ctx.key, aeadctx->key, aeadctx->enckey_len);
3110 	memcpy(chcr_req->key_ctx.key + roundup(aeadctx->enckey_len, 16),
3111 			aeadctx->key, aeadctx->enckey_len);
3112 
3113 	phys_cpl = (struct cpl_rx_phys_dsgl *)((u8 *)(chcr_req + 1) + kctx_len);
3114 	ivptr = (u8 *)(phys_cpl + 1) + dst_size;
3115 	ulptx = (struct ulptx_sgl *)(ivptr + IV);
3116 	error = ccm_format_packet(req, ivptr, sub_type, reqctx->op, assoclen);
3117 	if (error)
3118 		goto dstmap_fail;
3119 	chcr_add_aead_dst_ent(req, phys_cpl, qid);
3120 	chcr_add_aead_src_ent(req, ulptx);
3121 
3122 	atomic_inc(&adap->chcr_stats.aead_rqst);
3123 	temp = sizeof(struct cpl_rx_phys_dsgl) + dst_size + IV +
3124 		kctx_len + (reqctx->imm ? (req->assoclen + req->cryptlen +
3125 		reqctx->b0_len) : 0);
3126 	create_wreq(a_ctx(tfm), chcr_req, &req->base, reqctx->imm, 0,
3127 		    transhdr_len, temp, 0);
3128 	reqctx->skb = skb;
3129 
3130 	return skb;
3131 dstmap_fail:
3132 	kfree_skb(skb);
3133 err:
3134 	chcr_aead_common_exit(req);
3135 	return ERR_PTR(error);
3136 }
3137 
3138 static struct sk_buff *create_gcm_wr(struct aead_request *req,
3139 				     unsigned short qid,
3140 				     int size)
3141 {
3142 	struct crypto_aead *tfm = crypto_aead_reqtfm(req);
3143 	struct chcr_context *ctx = a_ctx(tfm);
3144 	struct uld_ctx *u_ctx = ULD_CTX(ctx);
3145 	struct chcr_aead_ctx *aeadctx = AEAD_CTX(ctx);
3146 	struct chcr_aead_reqctx *reqctx = aead_request_ctx_dma(req);
3147 	struct sk_buff *skb = NULL;
3148 	struct chcr_wr *chcr_req;
3149 	struct cpl_rx_phys_dsgl *phys_cpl;
3150 	struct ulptx_sgl *ulptx;
3151 	unsigned int transhdr_len, dnents = 0, snents;
3152 	unsigned int dst_size = 0, temp = 0, kctx_len, assoclen = req->assoclen;
3153 	unsigned int authsize = crypto_aead_authsize(tfm);
3154 	int error = -EINVAL;
3155 	u8 *ivptr;
3156 	gfp_t flags = req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP ? GFP_KERNEL :
3157 		GFP_ATOMIC;
3158 	struct adapter *adap = padap(ctx->dev);
3159 	unsigned int rx_channel_id = reqctx->rxqidx / ctx->rxq_perchan;
3160 
3161 	rx_channel_id = cxgb4_port_e2cchan(u_ctx->lldi.ports[rx_channel_id]);
3162 	if (get_aead_subtype(tfm) == CRYPTO_ALG_SUB_TYPE_AEAD_RFC4106)
3163 		assoclen = req->assoclen - 8;
3164 
3165 	reqctx->b0_len = 0;
3166 	error = chcr_aead_common_init(req);
3167 	if (error)
3168 		return ERR_PTR(error);
3169 	dnents = sg_nents_xlen(req->dst, req->assoclen + req->cryptlen +
3170 				(reqctx->op ? -authsize : authsize),
3171 				CHCR_DST_SG_SIZE, 0);
3172 	snents = sg_nents_xlen(req->src, req->assoclen + req->cryptlen,
3173 			       CHCR_SRC_SG_SIZE, 0);
3174 	dnents += MIN_GCM_SG; // For IV
3175 	dst_size = get_space_for_phys_dsgl(dnents);
3176 	kctx_len = roundup(aeadctx->enckey_len, 16) + AEAD_H_SIZE;
3177 	transhdr_len = CIPHER_TRANSHDR_SIZE(kctx_len, dst_size);
3178 	reqctx->imm = (transhdr_len + req->assoclen + req->cryptlen) <=
3179 			SGE_MAX_WR_LEN;
3180 	temp = reqctx->imm ? roundup(req->assoclen + req->cryptlen, 16) :
3181 		(sgl_len(snents) * 8);
3182 	transhdr_len += temp;
3183 	transhdr_len = roundup(transhdr_len, 16);
3184 	if (chcr_aead_need_fallback(req, dnents, T6_MAX_AAD_SIZE,
3185 			    transhdr_len, reqctx->op)) {
3186 
3187 		atomic_inc(&adap->chcr_stats.fallback);
3188 		chcr_aead_common_exit(req);
3189 		return ERR_PTR(chcr_aead_fallback(req, reqctx->op));
3190 	}
3191 	skb = alloc_skb(transhdr_len, flags);
3192 	if (!skb) {
3193 		error = -ENOMEM;
3194 		goto err;
3195 	}
3196 
3197 	chcr_req = __skb_put_zero(skb, transhdr_len);
3198 
3199 	//Offset of tag from end
3200 	temp = (reqctx->op == CHCR_ENCRYPT_OP) ? 0 : authsize;
3201 	chcr_req->sec_cpl.op_ivinsrtofst = FILL_SEC_CPL_OP_IVINSR(
3202 						rx_channel_id, 2, 1);
3203 	chcr_req->sec_cpl.pldlen =
3204 		htonl(req->assoclen + IV + req->cryptlen);
3205 	chcr_req->sec_cpl.aadstart_cipherstop_hi = FILL_SEC_CPL_CIPHERSTOP_HI(
3206 					assoclen ? 1 + IV : 0,
3207 					assoclen ? IV + assoclen : 0,
3208 					req->assoclen + IV + 1, 0);
3209 	chcr_req->sec_cpl.cipherstop_lo_authinsert =
3210 			FILL_SEC_CPL_AUTHINSERT(0, req->assoclen + IV + 1,
3211 						temp, temp);
3212 	chcr_req->sec_cpl.seqno_numivs =
3213 			FILL_SEC_CPL_SCMD0_SEQNO(reqctx->op, (reqctx->op ==
3214 					CHCR_ENCRYPT_OP) ? 1 : 0,
3215 					CHCR_SCMD_CIPHER_MODE_AES_GCM,
3216 					CHCR_SCMD_AUTH_MODE_GHASH,
3217 					aeadctx->hmac_ctrl, IV >> 1);
3218 	chcr_req->sec_cpl.ivgen_hdrlen =  FILL_SEC_CPL_IVGEN_HDRLEN(0, 0, 1,
3219 					0, 0, dst_size);
3220 	chcr_req->key_ctx.ctx_hdr = aeadctx->key_ctx_hdr;
3221 	memcpy(chcr_req->key_ctx.key, aeadctx->key, aeadctx->enckey_len);
3222 	memcpy(chcr_req->key_ctx.key + roundup(aeadctx->enckey_len, 16),
3223 	       GCM_CTX(aeadctx)->ghash_h, AEAD_H_SIZE);
3224 
3225 	phys_cpl = (struct cpl_rx_phys_dsgl *)((u8 *)(chcr_req + 1) + kctx_len);
3226 	ivptr = (u8 *)(phys_cpl + 1) + dst_size;
3227 	/* prepare a 16 byte iv */
3228 	/* S   A   L  T |  IV | 0x00000001 */
3229 	if (get_aead_subtype(tfm) ==
3230 	    CRYPTO_ALG_SUB_TYPE_AEAD_RFC4106) {
3231 		memcpy(ivptr, aeadctx->salt, 4);
3232 		memcpy(ivptr + 4, req->iv, GCM_RFC4106_IV_SIZE);
3233 	} else {
3234 		memcpy(ivptr, req->iv, GCM_AES_IV_SIZE);
3235 	}
3236 	put_unaligned_be32(0x01, &ivptr[12]);
3237 	ulptx = (struct ulptx_sgl *)(ivptr + 16);
3238 
3239 	chcr_add_aead_dst_ent(req, phys_cpl, qid);
3240 	chcr_add_aead_src_ent(req, ulptx);
3241 	atomic_inc(&adap->chcr_stats.aead_rqst);
3242 	temp = sizeof(struct cpl_rx_phys_dsgl) + dst_size + IV +
3243 		kctx_len + (reqctx->imm ? (req->assoclen + req->cryptlen) : 0);
3244 	create_wreq(a_ctx(tfm), chcr_req, &req->base, reqctx->imm, size,
3245 		    transhdr_len, temp, reqctx->verify);
3246 	reqctx->skb = skb;
3247 	return skb;
3248 
3249 err:
3250 	chcr_aead_common_exit(req);
3251 	return ERR_PTR(error);
3252 }
3253 
3254 
3255 
3256 static int chcr_aead_cra_init(struct crypto_aead *tfm)
3257 {
3258 	struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
3259 	struct aead_alg *alg = crypto_aead_alg(tfm);
3260 
3261 	aeadctx->sw_cipher = crypto_alloc_aead(alg->base.cra_name, 0,
3262 					       CRYPTO_ALG_NEED_FALLBACK |
3263 					       CRYPTO_ALG_ASYNC);
3264 	if  (IS_ERR(aeadctx->sw_cipher))
3265 		return PTR_ERR(aeadctx->sw_cipher);
3266 	crypto_aead_set_reqsize_dma(
3267 		tfm, max(sizeof(struct chcr_aead_reqctx),
3268 			 sizeof(struct aead_request) +
3269 			 crypto_aead_reqsize(aeadctx->sw_cipher)));
3270 	return chcr_device_init(a_ctx(tfm));
3271 }
3272 
3273 static void chcr_aead_cra_exit(struct crypto_aead *tfm)
3274 {
3275 	struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
3276 
3277 	crypto_free_aead(aeadctx->sw_cipher);
3278 }
3279 
3280 static int chcr_authenc_null_setauthsize(struct crypto_aead *tfm,
3281 					unsigned int authsize)
3282 {
3283 	struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
3284 
3285 	aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_NOP;
3286 	aeadctx->mayverify = VERIFY_HW;
3287 	return crypto_aead_setauthsize(aeadctx->sw_cipher, authsize);
3288 }
3289 static int chcr_authenc_setauthsize(struct crypto_aead *tfm,
3290 				    unsigned int authsize)
3291 {
3292 	struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
3293 	u32 maxauth = crypto_aead_maxauthsize(tfm);
3294 
3295 	/*SHA1 authsize in ipsec is 12 instead of 10 i.e maxauthsize / 2 is not
3296 	 * true for sha1. authsize == 12 condition should be before
3297 	 * authsize == (maxauth >> 1)
3298 	 */
3299 	if (authsize == ICV_4) {
3300 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_PL1;
3301 		aeadctx->mayverify = VERIFY_HW;
3302 	} else if (authsize == ICV_6) {
3303 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_PL2;
3304 		aeadctx->mayverify = VERIFY_HW;
3305 	} else if (authsize == ICV_10) {
3306 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_TRUNC_RFC4366;
3307 		aeadctx->mayverify = VERIFY_HW;
3308 	} else if (authsize == ICV_12) {
3309 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_IPSEC_96BIT;
3310 		aeadctx->mayverify = VERIFY_HW;
3311 	} else if (authsize == ICV_14) {
3312 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_PL3;
3313 		aeadctx->mayverify = VERIFY_HW;
3314 	} else if (authsize == (maxauth >> 1)) {
3315 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_DIV2;
3316 		aeadctx->mayverify = VERIFY_HW;
3317 	} else if (authsize == maxauth) {
3318 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_NO_TRUNC;
3319 		aeadctx->mayverify = VERIFY_HW;
3320 	} else {
3321 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_NO_TRUNC;
3322 		aeadctx->mayverify = VERIFY_SW;
3323 	}
3324 	return crypto_aead_setauthsize(aeadctx->sw_cipher, authsize);
3325 }
3326 
3327 
3328 static int chcr_gcm_setauthsize(struct crypto_aead *tfm, unsigned int authsize)
3329 {
3330 	struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
3331 
3332 	switch (authsize) {
3333 	case ICV_4:
3334 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_PL1;
3335 		aeadctx->mayverify = VERIFY_HW;
3336 		break;
3337 	case ICV_8:
3338 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_DIV2;
3339 		aeadctx->mayverify = VERIFY_HW;
3340 		break;
3341 	case ICV_12:
3342 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_IPSEC_96BIT;
3343 		aeadctx->mayverify = VERIFY_HW;
3344 		break;
3345 	case ICV_14:
3346 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_PL3;
3347 		aeadctx->mayverify = VERIFY_HW;
3348 		break;
3349 	case ICV_16:
3350 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_NO_TRUNC;
3351 		aeadctx->mayverify = VERIFY_HW;
3352 		break;
3353 	case ICV_13:
3354 	case ICV_15:
3355 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_NO_TRUNC;
3356 		aeadctx->mayverify = VERIFY_SW;
3357 		break;
3358 	default:
3359 		return -EINVAL;
3360 	}
3361 	return crypto_aead_setauthsize(aeadctx->sw_cipher, authsize);
3362 }
3363 
3364 static int chcr_4106_4309_setauthsize(struct crypto_aead *tfm,
3365 					  unsigned int authsize)
3366 {
3367 	struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
3368 
3369 	switch (authsize) {
3370 	case ICV_8:
3371 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_DIV2;
3372 		aeadctx->mayverify = VERIFY_HW;
3373 		break;
3374 	case ICV_12:
3375 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_IPSEC_96BIT;
3376 		aeadctx->mayverify = VERIFY_HW;
3377 		break;
3378 	case ICV_16:
3379 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_NO_TRUNC;
3380 		aeadctx->mayverify = VERIFY_HW;
3381 		break;
3382 	default:
3383 		return -EINVAL;
3384 	}
3385 	return crypto_aead_setauthsize(aeadctx->sw_cipher, authsize);
3386 }
3387 
3388 static int chcr_ccm_setauthsize(struct crypto_aead *tfm,
3389 				unsigned int authsize)
3390 {
3391 	struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(tfm));
3392 
3393 	switch (authsize) {
3394 	case ICV_4:
3395 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_PL1;
3396 		aeadctx->mayverify = VERIFY_HW;
3397 		break;
3398 	case ICV_6:
3399 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_PL2;
3400 		aeadctx->mayverify = VERIFY_HW;
3401 		break;
3402 	case ICV_8:
3403 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_DIV2;
3404 		aeadctx->mayverify = VERIFY_HW;
3405 		break;
3406 	case ICV_10:
3407 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_TRUNC_RFC4366;
3408 		aeadctx->mayverify = VERIFY_HW;
3409 		break;
3410 	case ICV_12:
3411 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_IPSEC_96BIT;
3412 		aeadctx->mayverify = VERIFY_HW;
3413 		break;
3414 	case ICV_14:
3415 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_PL3;
3416 		aeadctx->mayverify = VERIFY_HW;
3417 		break;
3418 	case ICV_16:
3419 		aeadctx->hmac_ctrl = CHCR_SCMD_HMAC_CTRL_NO_TRUNC;
3420 		aeadctx->mayverify = VERIFY_HW;
3421 		break;
3422 	default:
3423 		return -EINVAL;
3424 	}
3425 	return crypto_aead_setauthsize(aeadctx->sw_cipher, authsize);
3426 }
3427 
3428 static int chcr_ccm_common_setkey(struct crypto_aead *aead,
3429 				const u8 *key,
3430 				unsigned int keylen)
3431 {
3432 	struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(aead));
3433 	unsigned char ck_size, mk_size;
3434 	int key_ctx_size = 0;
3435 
3436 	key_ctx_size = sizeof(struct _key_ctx) + roundup(keylen, 16) * 2;
3437 	if (keylen == AES_KEYSIZE_128) {
3438 		ck_size = CHCR_KEYCTX_CIPHER_KEY_SIZE_128;
3439 		mk_size = CHCR_KEYCTX_MAC_KEY_SIZE_128;
3440 	} else if (keylen == AES_KEYSIZE_192) {
3441 		ck_size = CHCR_KEYCTX_CIPHER_KEY_SIZE_192;
3442 		mk_size = CHCR_KEYCTX_MAC_KEY_SIZE_192;
3443 	} else if (keylen == AES_KEYSIZE_256) {
3444 		ck_size = CHCR_KEYCTX_CIPHER_KEY_SIZE_256;
3445 		mk_size = CHCR_KEYCTX_MAC_KEY_SIZE_256;
3446 	} else {
3447 		aeadctx->enckey_len = 0;
3448 		return	-EINVAL;
3449 	}
3450 	aeadctx->key_ctx_hdr = FILL_KEY_CTX_HDR(ck_size, mk_size, 0, 0,
3451 						key_ctx_size >> 4);
3452 	memcpy(aeadctx->key, key, keylen);
3453 	aeadctx->enckey_len = keylen;
3454 
3455 	return 0;
3456 }
3457 
3458 static int chcr_aead_ccm_setkey(struct crypto_aead *aead,
3459 				const u8 *key,
3460 				unsigned int keylen)
3461 {
3462 	struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(aead));
3463 	int error;
3464 
3465 	crypto_aead_clear_flags(aeadctx->sw_cipher, CRYPTO_TFM_REQ_MASK);
3466 	crypto_aead_set_flags(aeadctx->sw_cipher, crypto_aead_get_flags(aead) &
3467 			      CRYPTO_TFM_REQ_MASK);
3468 	error = crypto_aead_setkey(aeadctx->sw_cipher, key, keylen);
3469 	if (error)
3470 		return error;
3471 	return chcr_ccm_common_setkey(aead, key, keylen);
3472 }
3473 
3474 static int chcr_aead_rfc4309_setkey(struct crypto_aead *aead, const u8 *key,
3475 				    unsigned int keylen)
3476 {
3477 	struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(aead));
3478 	int error;
3479 
3480 	if (keylen < 3) {
3481 		aeadctx->enckey_len = 0;
3482 		return	-EINVAL;
3483 	}
3484 	crypto_aead_clear_flags(aeadctx->sw_cipher, CRYPTO_TFM_REQ_MASK);
3485 	crypto_aead_set_flags(aeadctx->sw_cipher, crypto_aead_get_flags(aead) &
3486 			      CRYPTO_TFM_REQ_MASK);
3487 	error = crypto_aead_setkey(aeadctx->sw_cipher, key, keylen);
3488 	if (error)
3489 		return error;
3490 	keylen -= 3;
3491 	memcpy(aeadctx->salt, key + keylen, 3);
3492 	return chcr_ccm_common_setkey(aead, key, keylen);
3493 }
3494 
3495 static int chcr_gcm_setkey(struct crypto_aead *aead, const u8 *key,
3496 			   unsigned int keylen)
3497 {
3498 	struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(aead));
3499 	struct chcr_gcm_ctx *gctx = GCM_CTX(aeadctx);
3500 	unsigned int ck_size;
3501 	int ret = 0, key_ctx_size = 0;
3502 	struct crypto_aes_ctx aes;
3503 
3504 	aeadctx->enckey_len = 0;
3505 	crypto_aead_clear_flags(aeadctx->sw_cipher, CRYPTO_TFM_REQ_MASK);
3506 	crypto_aead_set_flags(aeadctx->sw_cipher, crypto_aead_get_flags(aead)
3507 			      & CRYPTO_TFM_REQ_MASK);
3508 	ret = crypto_aead_setkey(aeadctx->sw_cipher, key, keylen);
3509 	if (ret)
3510 		goto out;
3511 
3512 	if (get_aead_subtype(aead) == CRYPTO_ALG_SUB_TYPE_AEAD_RFC4106 &&
3513 	    keylen > 3) {
3514 		keylen -= 4;  /* nonce/salt is present in the last 4 bytes */
3515 		memcpy(aeadctx->salt, key + keylen, 4);
3516 	}
3517 	if (keylen == AES_KEYSIZE_128) {
3518 		ck_size = CHCR_KEYCTX_CIPHER_KEY_SIZE_128;
3519 	} else if (keylen == AES_KEYSIZE_192) {
3520 		ck_size = CHCR_KEYCTX_CIPHER_KEY_SIZE_192;
3521 	} else if (keylen == AES_KEYSIZE_256) {
3522 		ck_size = CHCR_KEYCTX_CIPHER_KEY_SIZE_256;
3523 	} else {
3524 		pr_err("GCM: Invalid key length %d\n", keylen);
3525 		ret = -EINVAL;
3526 		goto out;
3527 	}
3528 
3529 	memcpy(aeadctx->key, key, keylen);
3530 	aeadctx->enckey_len = keylen;
3531 	key_ctx_size = sizeof(struct _key_ctx) + roundup(keylen, 16) +
3532 		AEAD_H_SIZE;
3533 	aeadctx->key_ctx_hdr = FILL_KEY_CTX_HDR(ck_size,
3534 						CHCR_KEYCTX_MAC_KEY_SIZE_128,
3535 						0, 0,
3536 						key_ctx_size >> 4);
3537 	/* Calculate the H = CIPH(K, 0 repeated 16 times).
3538 	 * It will go in key context
3539 	 */
3540 	ret = aes_expandkey(&aes, key, keylen);
3541 	if (ret) {
3542 		aeadctx->enckey_len = 0;
3543 		goto out;
3544 	}
3545 	memset(gctx->ghash_h, 0, AEAD_H_SIZE);
3546 	aes_encrypt(&aes, gctx->ghash_h, gctx->ghash_h);
3547 	memzero_explicit(&aes, sizeof(aes));
3548 
3549 out:
3550 	return ret;
3551 }
3552 
3553 static int chcr_authenc_setkey(struct crypto_aead *authenc, const u8 *key,
3554 				   unsigned int keylen)
3555 {
3556 	struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(authenc));
3557 	struct chcr_authenc_ctx *actx = AUTHENC_CTX(aeadctx);
3558 	/* it contains auth and cipher key both*/
3559 	struct crypto_authenc_keys keys;
3560 	unsigned int bs, subtype;
3561 	unsigned int max_authsize = crypto_aead_alg(authenc)->maxauthsize;
3562 	int err = 0, i, key_ctx_len = 0;
3563 	unsigned char ck_size = 0;
3564 	unsigned char pad[CHCR_HASH_MAX_BLOCK_SIZE_128] = { 0 };
3565 	struct crypto_shash *base_hash = ERR_PTR(-EINVAL);
3566 	struct algo_param param;
3567 	int align;
3568 	u8 *o_ptr = NULL;
3569 
3570 	crypto_aead_clear_flags(aeadctx->sw_cipher, CRYPTO_TFM_REQ_MASK);
3571 	crypto_aead_set_flags(aeadctx->sw_cipher, crypto_aead_get_flags(authenc)
3572 			      & CRYPTO_TFM_REQ_MASK);
3573 	err = crypto_aead_setkey(aeadctx->sw_cipher, key, keylen);
3574 	if (err)
3575 		goto out;
3576 
3577 	if (crypto_authenc_extractkeys(&keys, key, keylen) != 0)
3578 		goto out;
3579 
3580 	if (get_alg_config(&param, max_authsize)) {
3581 		pr_err("Unsupported digest size\n");
3582 		goto out;
3583 	}
3584 	subtype = get_aead_subtype(authenc);
3585 	if (subtype == CRYPTO_ALG_SUB_TYPE_CTR_SHA ||
3586 		subtype == CRYPTO_ALG_SUB_TYPE_CTR_NULL) {
3587 		if (keys.enckeylen < CTR_RFC3686_NONCE_SIZE)
3588 			goto out;
3589 		memcpy(aeadctx->nonce, keys.enckey + (keys.enckeylen
3590 		- CTR_RFC3686_NONCE_SIZE), CTR_RFC3686_NONCE_SIZE);
3591 		keys.enckeylen -= CTR_RFC3686_NONCE_SIZE;
3592 	}
3593 	if (keys.enckeylen == AES_KEYSIZE_128) {
3594 		ck_size = CHCR_KEYCTX_CIPHER_KEY_SIZE_128;
3595 	} else if (keys.enckeylen == AES_KEYSIZE_192) {
3596 		ck_size = CHCR_KEYCTX_CIPHER_KEY_SIZE_192;
3597 	} else if (keys.enckeylen == AES_KEYSIZE_256) {
3598 		ck_size = CHCR_KEYCTX_CIPHER_KEY_SIZE_256;
3599 	} else {
3600 		pr_err("Unsupported cipher key\n");
3601 		goto out;
3602 	}
3603 
3604 	/* Copy only encryption key. We use authkey to generate h(ipad) and
3605 	 * h(opad) so authkey is not needed again. authkeylen size have the
3606 	 * size of the hash digest size.
3607 	 */
3608 	memcpy(aeadctx->key, keys.enckey, keys.enckeylen);
3609 	aeadctx->enckey_len = keys.enckeylen;
3610 	if (subtype == CRYPTO_ALG_SUB_TYPE_CBC_SHA ||
3611 		subtype == CRYPTO_ALG_SUB_TYPE_CBC_NULL) {
3612 
3613 		get_aes_decrypt_key(actx->dec_rrkey, aeadctx->key,
3614 			    aeadctx->enckey_len << 3);
3615 	}
3616 	base_hash  = chcr_alloc_shash(max_authsize);
3617 	if (IS_ERR(base_hash)) {
3618 		pr_err("Base driver cannot be loaded\n");
3619 		goto out;
3620 	}
3621 	{
3622 		SHASH_DESC_ON_STACK(shash, base_hash);
3623 
3624 		shash->tfm = base_hash;
3625 		bs = crypto_shash_blocksize(base_hash);
3626 		align = KEYCTX_ALIGN_PAD(max_authsize);
3627 		o_ptr =  actx->h_iopad + param.result_size + align;
3628 
3629 		if (keys.authkeylen > bs) {
3630 			err = crypto_shash_digest(shash, keys.authkey,
3631 						  keys.authkeylen,
3632 						  o_ptr);
3633 			if (err) {
3634 				pr_err("Base driver cannot be loaded\n");
3635 				goto out;
3636 			}
3637 			keys.authkeylen = max_authsize;
3638 		} else
3639 			memcpy(o_ptr, keys.authkey, keys.authkeylen);
3640 
3641 		/* Compute the ipad-digest*/
3642 		memset(pad + keys.authkeylen, 0, bs - keys.authkeylen);
3643 		memcpy(pad, o_ptr, keys.authkeylen);
3644 		for (i = 0; i < bs >> 2; i++)
3645 			*((unsigned int *)pad + i) ^= IPAD_DATA;
3646 
3647 		if (chcr_compute_partial_hash(shash, pad, actx->h_iopad,
3648 					      max_authsize))
3649 			goto out;
3650 		/* Compute the opad-digest */
3651 		memset(pad + keys.authkeylen, 0, bs - keys.authkeylen);
3652 		memcpy(pad, o_ptr, keys.authkeylen);
3653 		for (i = 0; i < bs >> 2; i++)
3654 			*((unsigned int *)pad + i) ^= OPAD_DATA;
3655 
3656 		if (chcr_compute_partial_hash(shash, pad, o_ptr, max_authsize))
3657 			goto out;
3658 
3659 		/* convert the ipad and opad digest to network order */
3660 		chcr_change_order(actx->h_iopad, param.result_size);
3661 		chcr_change_order(o_ptr, param.result_size);
3662 		key_ctx_len = sizeof(struct _key_ctx) +
3663 			roundup(keys.enckeylen, 16) +
3664 			(param.result_size + align) * 2;
3665 		aeadctx->key_ctx_hdr = FILL_KEY_CTX_HDR(ck_size, param.mk_size,
3666 						0, 1, key_ctx_len >> 4);
3667 		actx->auth_mode = param.auth_mode;
3668 		chcr_free_shash(base_hash);
3669 
3670 		memzero_explicit(&keys, sizeof(keys));
3671 		return 0;
3672 	}
3673 out:
3674 	aeadctx->enckey_len = 0;
3675 	memzero_explicit(&keys, sizeof(keys));
3676 	if (!IS_ERR(base_hash))
3677 		chcr_free_shash(base_hash);
3678 	return -EINVAL;
3679 }
3680 
3681 static int chcr_aead_digest_null_setkey(struct crypto_aead *authenc,
3682 					const u8 *key, unsigned int keylen)
3683 {
3684 	struct chcr_aead_ctx *aeadctx = AEAD_CTX(a_ctx(authenc));
3685 	struct chcr_authenc_ctx *actx = AUTHENC_CTX(aeadctx);
3686 	struct crypto_authenc_keys keys;
3687 	int err;
3688 	/* it contains auth and cipher key both*/
3689 	unsigned int subtype;
3690 	int key_ctx_len = 0;
3691 	unsigned char ck_size = 0;
3692 
3693 	crypto_aead_clear_flags(aeadctx->sw_cipher, CRYPTO_TFM_REQ_MASK);
3694 	crypto_aead_set_flags(aeadctx->sw_cipher, crypto_aead_get_flags(authenc)
3695 			      & CRYPTO_TFM_REQ_MASK);
3696 	err = crypto_aead_setkey(aeadctx->sw_cipher, key, keylen);
3697 	if (err)
3698 		goto out;
3699 
3700 	if (crypto_authenc_extractkeys(&keys, key, keylen) != 0)
3701 		goto out;
3702 
3703 	subtype = get_aead_subtype(authenc);
3704 	if (subtype == CRYPTO_ALG_SUB_TYPE_CTR_SHA ||
3705 	    subtype == CRYPTO_ALG_SUB_TYPE_CTR_NULL) {
3706 		if (keys.enckeylen < CTR_RFC3686_NONCE_SIZE)
3707 			goto out;
3708 		memcpy(aeadctx->nonce, keys.enckey + (keys.enckeylen
3709 			- CTR_RFC3686_NONCE_SIZE), CTR_RFC3686_NONCE_SIZE);
3710 		keys.enckeylen -= CTR_RFC3686_NONCE_SIZE;
3711 	}
3712 	if (keys.enckeylen == AES_KEYSIZE_128) {
3713 		ck_size = CHCR_KEYCTX_CIPHER_KEY_SIZE_128;
3714 	} else if (keys.enckeylen == AES_KEYSIZE_192) {
3715 		ck_size = CHCR_KEYCTX_CIPHER_KEY_SIZE_192;
3716 	} else if (keys.enckeylen == AES_KEYSIZE_256) {
3717 		ck_size = CHCR_KEYCTX_CIPHER_KEY_SIZE_256;
3718 	} else {
3719 		pr_err("Unsupported cipher key %d\n", keys.enckeylen);
3720 		goto out;
3721 	}
3722 	memcpy(aeadctx->key, keys.enckey, keys.enckeylen);
3723 	aeadctx->enckey_len = keys.enckeylen;
3724 	if (subtype == CRYPTO_ALG_SUB_TYPE_CBC_SHA ||
3725 	    subtype == CRYPTO_ALG_SUB_TYPE_CBC_NULL) {
3726 		get_aes_decrypt_key(actx->dec_rrkey, aeadctx->key,
3727 				aeadctx->enckey_len << 3);
3728 	}
3729 	key_ctx_len =  sizeof(struct _key_ctx) + roundup(keys.enckeylen, 16);
3730 
3731 	aeadctx->key_ctx_hdr = FILL_KEY_CTX_HDR(ck_size, CHCR_KEYCTX_NO_KEY, 0,
3732 						0, key_ctx_len >> 4);
3733 	actx->auth_mode = CHCR_SCMD_AUTH_MODE_NOP;
3734 	memzero_explicit(&keys, sizeof(keys));
3735 	return 0;
3736 out:
3737 	aeadctx->enckey_len = 0;
3738 	memzero_explicit(&keys, sizeof(keys));
3739 	return -EINVAL;
3740 }
3741 
3742 static int chcr_aead_op(struct aead_request *req,
3743 			int size,
3744 			create_wr_t create_wr_fn)
3745 {
3746 	struct crypto_aead *tfm = crypto_aead_reqtfm(req);
3747 	struct chcr_aead_reqctx *reqctx = aead_request_ctx_dma(req);
3748 	struct chcr_context *ctx = a_ctx(tfm);
3749 	struct uld_ctx *u_ctx = ULD_CTX(ctx);
3750 	struct sk_buff *skb;
3751 	struct chcr_dev *cdev;
3752 
3753 	cdev = a_ctx(tfm)->dev;
3754 	if (!cdev) {
3755 		pr_err("%s : No crypto device.\n", __func__);
3756 		return -ENXIO;
3757 	}
3758 
3759 	if (chcr_inc_wrcount(cdev)) {
3760 	/* Detach state for CHCR means lldi or padap is freed.
3761 	 * We cannot increment fallback here.
3762 	 */
3763 		return chcr_aead_fallback(req, reqctx->op);
3764 	}
3765 
3766 	if (cxgb4_is_crypto_q_full(u_ctx->lldi.ports[0],
3767 					reqctx->txqidx) &&
3768 		(!(req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG))) {
3769 			chcr_dec_wrcount(cdev);
3770 			return -ENOSPC;
3771 	}
3772 
3773 	if (get_aead_subtype(tfm) == CRYPTO_ALG_SUB_TYPE_AEAD_RFC4106 &&
3774 	    crypto_ipsec_check_assoclen(req->assoclen) != 0) {
3775 		pr_err("RFC4106: Invalid value of assoclen %d\n",
3776 		       req->assoclen);
3777 		return -EINVAL;
3778 	}
3779 
3780 	/* Form a WR from req */
3781 	skb = create_wr_fn(req, u_ctx->lldi.rxq_ids[reqctx->rxqidx], size);
3782 
3783 	if (IS_ERR_OR_NULL(skb)) {
3784 		chcr_dec_wrcount(cdev);
3785 		return PTR_ERR_OR_ZERO(skb);
3786 	}
3787 
3788 	skb->dev = u_ctx->lldi.ports[0];
3789 	set_wr_txq(skb, CPL_PRIORITY_DATA, reqctx->txqidx);
3790 	chcr_send_wr(skb);
3791 	return -EINPROGRESS;
3792 }
3793 
3794 static int chcr_aead_encrypt(struct aead_request *req)
3795 {
3796 	struct crypto_aead *tfm = crypto_aead_reqtfm(req);
3797 	struct chcr_aead_reqctx *reqctx = aead_request_ctx_dma(req);
3798 	struct chcr_context *ctx = a_ctx(tfm);
3799 	unsigned int cpu;
3800 
3801 	cpu = get_cpu();
3802 	reqctx->txqidx = cpu % ctx->ntxq;
3803 	reqctx->rxqidx = cpu % ctx->nrxq;
3804 	put_cpu();
3805 
3806 	reqctx->verify = VERIFY_HW;
3807 	reqctx->op = CHCR_ENCRYPT_OP;
3808 
3809 	switch (get_aead_subtype(tfm)) {
3810 	case CRYPTO_ALG_SUB_TYPE_CTR_SHA:
3811 	case CRYPTO_ALG_SUB_TYPE_CBC_SHA:
3812 	case CRYPTO_ALG_SUB_TYPE_CBC_NULL:
3813 	case CRYPTO_ALG_SUB_TYPE_CTR_NULL:
3814 		return chcr_aead_op(req, 0, create_authenc_wr);
3815 	case CRYPTO_ALG_SUB_TYPE_AEAD_CCM:
3816 	case CRYPTO_ALG_SUB_TYPE_AEAD_RFC4309:
3817 		return chcr_aead_op(req, 0, create_aead_ccm_wr);
3818 	default:
3819 		return chcr_aead_op(req, 0, create_gcm_wr);
3820 	}
3821 }
3822 
3823 static int chcr_aead_decrypt(struct aead_request *req)
3824 {
3825 	struct crypto_aead *tfm = crypto_aead_reqtfm(req);
3826 	struct chcr_context *ctx = a_ctx(tfm);
3827 	struct chcr_aead_ctx *aeadctx = AEAD_CTX(ctx);
3828 	struct chcr_aead_reqctx *reqctx = aead_request_ctx_dma(req);
3829 	int size;
3830 	unsigned int cpu;
3831 
3832 	cpu = get_cpu();
3833 	reqctx->txqidx = cpu % ctx->ntxq;
3834 	reqctx->rxqidx = cpu % ctx->nrxq;
3835 	put_cpu();
3836 
3837 	if (aeadctx->mayverify == VERIFY_SW) {
3838 		size = crypto_aead_maxauthsize(tfm);
3839 		reqctx->verify = VERIFY_SW;
3840 	} else {
3841 		size = 0;
3842 		reqctx->verify = VERIFY_HW;
3843 	}
3844 	reqctx->op = CHCR_DECRYPT_OP;
3845 	switch (get_aead_subtype(tfm)) {
3846 	case CRYPTO_ALG_SUB_TYPE_CBC_SHA:
3847 	case CRYPTO_ALG_SUB_TYPE_CTR_SHA:
3848 	case CRYPTO_ALG_SUB_TYPE_CBC_NULL:
3849 	case CRYPTO_ALG_SUB_TYPE_CTR_NULL:
3850 		return chcr_aead_op(req, size, create_authenc_wr);
3851 	case CRYPTO_ALG_SUB_TYPE_AEAD_CCM:
3852 	case CRYPTO_ALG_SUB_TYPE_AEAD_RFC4309:
3853 		return chcr_aead_op(req, size, create_aead_ccm_wr);
3854 	default:
3855 		return chcr_aead_op(req, size, create_gcm_wr);
3856 	}
3857 }
3858 
3859 static struct chcr_alg_template driver_algs[] = {
3860 	/* AES-CBC */
3861 	{
3862 		.type = CRYPTO_ALG_TYPE_SKCIPHER | CRYPTO_ALG_SUB_TYPE_CBC,
3863 		.is_registered = 0,
3864 		.alg.skcipher = {
3865 			.base.cra_name		= "cbc(aes)",
3866 			.base.cra_driver_name	= "cbc-aes-chcr",
3867 			.base.cra_blocksize	= AES_BLOCK_SIZE,
3868 
3869 			.init			= chcr_init_tfm,
3870 			.exit			= chcr_exit_tfm,
3871 			.min_keysize		= AES_MIN_KEY_SIZE,
3872 			.max_keysize		= AES_MAX_KEY_SIZE,
3873 			.ivsize			= AES_BLOCK_SIZE,
3874 			.setkey			= chcr_aes_cbc_setkey,
3875 			.encrypt		= chcr_aes_encrypt,
3876 			.decrypt		= chcr_aes_decrypt,
3877 			}
3878 	},
3879 	{
3880 		.type = CRYPTO_ALG_TYPE_SKCIPHER | CRYPTO_ALG_SUB_TYPE_XTS,
3881 		.is_registered = 0,
3882 		.alg.skcipher = {
3883 			.base.cra_name		= "xts(aes)",
3884 			.base.cra_driver_name	= "xts-aes-chcr",
3885 			.base.cra_blocksize	= AES_BLOCK_SIZE,
3886 
3887 			.init			= chcr_init_tfm,
3888 			.exit			= chcr_exit_tfm,
3889 			.min_keysize		= 2 * AES_MIN_KEY_SIZE,
3890 			.max_keysize		= 2 * AES_MAX_KEY_SIZE,
3891 			.ivsize			= AES_BLOCK_SIZE,
3892 			.setkey			= chcr_aes_xts_setkey,
3893 			.encrypt		= chcr_aes_encrypt,
3894 			.decrypt		= chcr_aes_decrypt,
3895 			}
3896 	},
3897 	{
3898 		.type = CRYPTO_ALG_TYPE_SKCIPHER | CRYPTO_ALG_SUB_TYPE_CTR,
3899 		.is_registered = 0,
3900 		.alg.skcipher = {
3901 			.base.cra_name		= "ctr(aes)",
3902 			.base.cra_driver_name	= "ctr-aes-chcr",
3903 			.base.cra_blocksize	= 1,
3904 
3905 			.init			= chcr_init_tfm,
3906 			.exit			= chcr_exit_tfm,
3907 			.min_keysize		= AES_MIN_KEY_SIZE,
3908 			.max_keysize		= AES_MAX_KEY_SIZE,
3909 			.ivsize			= AES_BLOCK_SIZE,
3910 			.setkey			= chcr_aes_ctr_setkey,
3911 			.encrypt		= chcr_aes_encrypt,
3912 			.decrypt		= chcr_aes_decrypt,
3913 		}
3914 	},
3915 	{
3916 		.type = CRYPTO_ALG_TYPE_SKCIPHER |
3917 			CRYPTO_ALG_SUB_TYPE_CTR_RFC3686,
3918 		.is_registered = 0,
3919 		.alg.skcipher = {
3920 			.base.cra_name		= "rfc3686(ctr(aes))",
3921 			.base.cra_driver_name	= "rfc3686-ctr-aes-chcr",
3922 			.base.cra_blocksize	= 1,
3923 
3924 			.init			= chcr_rfc3686_init,
3925 			.exit			= chcr_exit_tfm,
3926 			.min_keysize		= AES_MIN_KEY_SIZE + CTR_RFC3686_NONCE_SIZE,
3927 			.max_keysize		= AES_MAX_KEY_SIZE + CTR_RFC3686_NONCE_SIZE,
3928 			.ivsize			= CTR_RFC3686_IV_SIZE,
3929 			.setkey			= chcr_aes_rfc3686_setkey,
3930 			.encrypt		= chcr_aes_encrypt,
3931 			.decrypt		= chcr_aes_decrypt,
3932 		}
3933 	},
3934 	/* SHA */
3935 	{
3936 		.type = CRYPTO_ALG_TYPE_AHASH,
3937 		.is_registered = 0,
3938 		.alg.hash = {
3939 			.halg.digestsize = SHA1_DIGEST_SIZE,
3940 			.halg.base = {
3941 				.cra_name = "sha1",
3942 				.cra_driver_name = "sha1-chcr",
3943 				.cra_blocksize = SHA1_BLOCK_SIZE,
3944 			}
3945 		}
3946 	},
3947 	{
3948 		.type = CRYPTO_ALG_TYPE_AHASH,
3949 		.is_registered = 0,
3950 		.alg.hash = {
3951 			.halg.digestsize = SHA256_DIGEST_SIZE,
3952 			.halg.base = {
3953 				.cra_name = "sha256",
3954 				.cra_driver_name = "sha256-chcr",
3955 				.cra_blocksize = SHA256_BLOCK_SIZE,
3956 			}
3957 		}
3958 	},
3959 	{
3960 		.type = CRYPTO_ALG_TYPE_AHASH,
3961 		.is_registered = 0,
3962 		.alg.hash = {
3963 			.halg.digestsize = SHA224_DIGEST_SIZE,
3964 			.halg.base = {
3965 				.cra_name = "sha224",
3966 				.cra_driver_name = "sha224-chcr",
3967 				.cra_blocksize = SHA224_BLOCK_SIZE,
3968 			}
3969 		}
3970 	},
3971 	{
3972 		.type = CRYPTO_ALG_TYPE_AHASH,
3973 		.is_registered = 0,
3974 		.alg.hash = {
3975 			.halg.digestsize = SHA384_DIGEST_SIZE,
3976 			.halg.base = {
3977 				.cra_name = "sha384",
3978 				.cra_driver_name = "sha384-chcr",
3979 				.cra_blocksize = SHA384_BLOCK_SIZE,
3980 			}
3981 		}
3982 	},
3983 	{
3984 		.type = CRYPTO_ALG_TYPE_AHASH,
3985 		.is_registered = 0,
3986 		.alg.hash = {
3987 			.halg.digestsize = SHA512_DIGEST_SIZE,
3988 			.halg.base = {
3989 				.cra_name = "sha512",
3990 				.cra_driver_name = "sha512-chcr",
3991 				.cra_blocksize = SHA512_BLOCK_SIZE,
3992 			}
3993 		}
3994 	},
3995 	/* HMAC */
3996 	{
3997 		.type = CRYPTO_ALG_TYPE_HMAC,
3998 		.is_registered = 0,
3999 		.alg.hash = {
4000 			.halg.digestsize = SHA1_DIGEST_SIZE,
4001 			.halg.base = {
4002 				.cra_name = "hmac(sha1)",
4003 				.cra_driver_name = "hmac-sha1-chcr",
4004 				.cra_blocksize = SHA1_BLOCK_SIZE,
4005 			}
4006 		}
4007 	},
4008 	{
4009 		.type = CRYPTO_ALG_TYPE_HMAC,
4010 		.is_registered = 0,
4011 		.alg.hash = {
4012 			.halg.digestsize = SHA224_DIGEST_SIZE,
4013 			.halg.base = {
4014 				.cra_name = "hmac(sha224)",
4015 				.cra_driver_name = "hmac-sha224-chcr",
4016 				.cra_blocksize = SHA224_BLOCK_SIZE,
4017 			}
4018 		}
4019 	},
4020 	{
4021 		.type = CRYPTO_ALG_TYPE_HMAC,
4022 		.is_registered = 0,
4023 		.alg.hash = {
4024 			.halg.digestsize = SHA256_DIGEST_SIZE,
4025 			.halg.base = {
4026 				.cra_name = "hmac(sha256)",
4027 				.cra_driver_name = "hmac-sha256-chcr",
4028 				.cra_blocksize = SHA256_BLOCK_SIZE,
4029 			}
4030 		}
4031 	},
4032 	{
4033 		.type = CRYPTO_ALG_TYPE_HMAC,
4034 		.is_registered = 0,
4035 		.alg.hash = {
4036 			.halg.digestsize = SHA384_DIGEST_SIZE,
4037 			.halg.base = {
4038 				.cra_name = "hmac(sha384)",
4039 				.cra_driver_name = "hmac-sha384-chcr",
4040 				.cra_blocksize = SHA384_BLOCK_SIZE,
4041 			}
4042 		}
4043 	},
4044 	{
4045 		.type = CRYPTO_ALG_TYPE_HMAC,
4046 		.is_registered = 0,
4047 		.alg.hash = {
4048 			.halg.digestsize = SHA512_DIGEST_SIZE,
4049 			.halg.base = {
4050 				.cra_name = "hmac(sha512)",
4051 				.cra_driver_name = "hmac-sha512-chcr",
4052 				.cra_blocksize = SHA512_BLOCK_SIZE,
4053 			}
4054 		}
4055 	},
4056 	/* Add AEAD Algorithms */
4057 	{
4058 		.type = CRYPTO_ALG_TYPE_AEAD | CRYPTO_ALG_SUB_TYPE_AEAD_GCM,
4059 		.is_registered = 0,
4060 		.alg.aead = {
4061 			.base = {
4062 				.cra_name = "gcm(aes)",
4063 				.cra_driver_name = "gcm-aes-chcr",
4064 				.cra_blocksize	= 1,
4065 				.cra_priority = CHCR_AEAD_PRIORITY,
4066 				.cra_ctxsize =	sizeof(struct chcr_context) +
4067 						sizeof(struct chcr_aead_ctx) +
4068 						sizeof(struct chcr_gcm_ctx),
4069 			},
4070 			.ivsize = GCM_AES_IV_SIZE,
4071 			.maxauthsize = GHASH_DIGEST_SIZE,
4072 			.setkey = chcr_gcm_setkey,
4073 			.setauthsize = chcr_gcm_setauthsize,
4074 		}
4075 	},
4076 	{
4077 		.type = CRYPTO_ALG_TYPE_AEAD | CRYPTO_ALG_SUB_TYPE_AEAD_RFC4106,
4078 		.is_registered = 0,
4079 		.alg.aead = {
4080 			.base = {
4081 				.cra_name = "rfc4106(gcm(aes))",
4082 				.cra_driver_name = "rfc4106-gcm-aes-chcr",
4083 				.cra_blocksize	 = 1,
4084 				.cra_priority = CHCR_AEAD_PRIORITY + 1,
4085 				.cra_ctxsize =	sizeof(struct chcr_context) +
4086 						sizeof(struct chcr_aead_ctx) +
4087 						sizeof(struct chcr_gcm_ctx),
4088 
4089 			},
4090 			.ivsize = GCM_RFC4106_IV_SIZE,
4091 			.maxauthsize	= GHASH_DIGEST_SIZE,
4092 			.setkey = chcr_gcm_setkey,
4093 			.setauthsize	= chcr_4106_4309_setauthsize,
4094 		}
4095 	},
4096 	{
4097 		.type = CRYPTO_ALG_TYPE_AEAD | CRYPTO_ALG_SUB_TYPE_AEAD_CCM,
4098 		.is_registered = 0,
4099 		.alg.aead = {
4100 			.base = {
4101 				.cra_name = "ccm(aes)",
4102 				.cra_driver_name = "ccm-aes-chcr",
4103 				.cra_blocksize	 = 1,
4104 				.cra_priority = CHCR_AEAD_PRIORITY,
4105 				.cra_ctxsize =	sizeof(struct chcr_context) +
4106 						sizeof(struct chcr_aead_ctx),
4107 
4108 			},
4109 			.ivsize = AES_BLOCK_SIZE,
4110 			.maxauthsize	= GHASH_DIGEST_SIZE,
4111 			.setkey = chcr_aead_ccm_setkey,
4112 			.setauthsize	= chcr_ccm_setauthsize,
4113 		}
4114 	},
4115 	{
4116 		.type = CRYPTO_ALG_TYPE_AEAD | CRYPTO_ALG_SUB_TYPE_AEAD_RFC4309,
4117 		.is_registered = 0,
4118 		.alg.aead = {
4119 			.base = {
4120 				.cra_name = "rfc4309(ccm(aes))",
4121 				.cra_driver_name = "rfc4309-ccm-aes-chcr",
4122 				.cra_blocksize	 = 1,
4123 				.cra_priority = CHCR_AEAD_PRIORITY + 1,
4124 				.cra_ctxsize =	sizeof(struct chcr_context) +
4125 						sizeof(struct chcr_aead_ctx),
4126 
4127 			},
4128 			.ivsize = 8,
4129 			.maxauthsize	= GHASH_DIGEST_SIZE,
4130 			.setkey = chcr_aead_rfc4309_setkey,
4131 			.setauthsize = chcr_4106_4309_setauthsize,
4132 		}
4133 	},
4134 	{
4135 		.type = CRYPTO_ALG_TYPE_AEAD | CRYPTO_ALG_SUB_TYPE_CBC_SHA,
4136 		.is_registered = 0,
4137 		.alg.aead = {
4138 			.base = {
4139 				.cra_name = "authenc(hmac(sha1),cbc(aes))",
4140 				.cra_driver_name =
4141 					"authenc-hmac-sha1-cbc-aes-chcr",
4142 				.cra_blocksize	 = AES_BLOCK_SIZE,
4143 				.cra_priority = CHCR_AEAD_PRIORITY,
4144 				.cra_ctxsize =	sizeof(struct chcr_context) +
4145 						sizeof(struct chcr_aead_ctx) +
4146 						sizeof(struct chcr_authenc_ctx),
4147 
4148 			},
4149 			.ivsize = AES_BLOCK_SIZE,
4150 			.maxauthsize = SHA1_DIGEST_SIZE,
4151 			.setkey = chcr_authenc_setkey,
4152 			.setauthsize = chcr_authenc_setauthsize,
4153 		}
4154 	},
4155 	{
4156 		.type = CRYPTO_ALG_TYPE_AEAD | CRYPTO_ALG_SUB_TYPE_CBC_SHA,
4157 		.is_registered = 0,
4158 		.alg.aead = {
4159 			.base = {
4160 
4161 				.cra_name = "authenc(hmac(sha256),cbc(aes))",
4162 				.cra_driver_name =
4163 					"authenc-hmac-sha256-cbc-aes-chcr",
4164 				.cra_blocksize	 = AES_BLOCK_SIZE,
4165 				.cra_priority = CHCR_AEAD_PRIORITY,
4166 				.cra_ctxsize =	sizeof(struct chcr_context) +
4167 						sizeof(struct chcr_aead_ctx) +
4168 						sizeof(struct chcr_authenc_ctx),
4169 
4170 			},
4171 			.ivsize = AES_BLOCK_SIZE,
4172 			.maxauthsize	= SHA256_DIGEST_SIZE,
4173 			.setkey = chcr_authenc_setkey,
4174 			.setauthsize = chcr_authenc_setauthsize,
4175 		}
4176 	},
4177 	{
4178 		.type = CRYPTO_ALG_TYPE_AEAD | CRYPTO_ALG_SUB_TYPE_CBC_SHA,
4179 		.is_registered = 0,
4180 		.alg.aead = {
4181 			.base = {
4182 				.cra_name = "authenc(hmac(sha224),cbc(aes))",
4183 				.cra_driver_name =
4184 					"authenc-hmac-sha224-cbc-aes-chcr",
4185 				.cra_blocksize	 = AES_BLOCK_SIZE,
4186 				.cra_priority = CHCR_AEAD_PRIORITY,
4187 				.cra_ctxsize =	sizeof(struct chcr_context) +
4188 						sizeof(struct chcr_aead_ctx) +
4189 						sizeof(struct chcr_authenc_ctx),
4190 			},
4191 			.ivsize = AES_BLOCK_SIZE,
4192 			.maxauthsize = SHA224_DIGEST_SIZE,
4193 			.setkey = chcr_authenc_setkey,
4194 			.setauthsize = chcr_authenc_setauthsize,
4195 		}
4196 	},
4197 	{
4198 		.type = CRYPTO_ALG_TYPE_AEAD | CRYPTO_ALG_SUB_TYPE_CBC_SHA,
4199 		.is_registered = 0,
4200 		.alg.aead = {
4201 			.base = {
4202 				.cra_name = "authenc(hmac(sha384),cbc(aes))",
4203 				.cra_driver_name =
4204 					"authenc-hmac-sha384-cbc-aes-chcr",
4205 				.cra_blocksize	 = AES_BLOCK_SIZE,
4206 				.cra_priority = CHCR_AEAD_PRIORITY,
4207 				.cra_ctxsize =	sizeof(struct chcr_context) +
4208 						sizeof(struct chcr_aead_ctx) +
4209 						sizeof(struct chcr_authenc_ctx),
4210 
4211 			},
4212 			.ivsize = AES_BLOCK_SIZE,
4213 			.maxauthsize = SHA384_DIGEST_SIZE,
4214 			.setkey = chcr_authenc_setkey,
4215 			.setauthsize = chcr_authenc_setauthsize,
4216 		}
4217 	},
4218 	{
4219 		.type = CRYPTO_ALG_TYPE_AEAD | CRYPTO_ALG_SUB_TYPE_CBC_SHA,
4220 		.is_registered = 0,
4221 		.alg.aead = {
4222 			.base = {
4223 				.cra_name = "authenc(hmac(sha512),cbc(aes))",
4224 				.cra_driver_name =
4225 					"authenc-hmac-sha512-cbc-aes-chcr",
4226 				.cra_blocksize	 = AES_BLOCK_SIZE,
4227 				.cra_priority = CHCR_AEAD_PRIORITY,
4228 				.cra_ctxsize =	sizeof(struct chcr_context) +
4229 						sizeof(struct chcr_aead_ctx) +
4230 						sizeof(struct chcr_authenc_ctx),
4231 
4232 			},
4233 			.ivsize = AES_BLOCK_SIZE,
4234 			.maxauthsize = SHA512_DIGEST_SIZE,
4235 			.setkey = chcr_authenc_setkey,
4236 			.setauthsize = chcr_authenc_setauthsize,
4237 		}
4238 	},
4239 	{
4240 		.type = CRYPTO_ALG_TYPE_AEAD | CRYPTO_ALG_SUB_TYPE_CBC_NULL,
4241 		.is_registered = 0,
4242 		.alg.aead = {
4243 			.base = {
4244 				.cra_name = "authenc(digest_null,cbc(aes))",
4245 				.cra_driver_name =
4246 					"authenc-digest_null-cbc-aes-chcr",
4247 				.cra_blocksize	 = AES_BLOCK_SIZE,
4248 				.cra_priority = CHCR_AEAD_PRIORITY,
4249 				.cra_ctxsize =	sizeof(struct chcr_context) +
4250 						sizeof(struct chcr_aead_ctx) +
4251 						sizeof(struct chcr_authenc_ctx),
4252 
4253 			},
4254 			.ivsize  = AES_BLOCK_SIZE,
4255 			.maxauthsize = 0,
4256 			.setkey  = chcr_aead_digest_null_setkey,
4257 			.setauthsize = chcr_authenc_null_setauthsize,
4258 		}
4259 	},
4260 	{
4261 		.type = CRYPTO_ALG_TYPE_AEAD | CRYPTO_ALG_SUB_TYPE_CTR_SHA,
4262 		.is_registered = 0,
4263 		.alg.aead = {
4264 			.base = {
4265 				.cra_name = "authenc(hmac(sha1),rfc3686(ctr(aes)))",
4266 				.cra_driver_name =
4267 				"authenc-hmac-sha1-rfc3686-ctr-aes-chcr",
4268 				.cra_blocksize	 = 1,
4269 				.cra_priority = CHCR_AEAD_PRIORITY,
4270 				.cra_ctxsize =	sizeof(struct chcr_context) +
4271 						sizeof(struct chcr_aead_ctx) +
4272 						sizeof(struct chcr_authenc_ctx),
4273 
4274 			},
4275 			.ivsize = CTR_RFC3686_IV_SIZE,
4276 			.maxauthsize = SHA1_DIGEST_SIZE,
4277 			.setkey = chcr_authenc_setkey,
4278 			.setauthsize = chcr_authenc_setauthsize,
4279 		}
4280 	},
4281 	{
4282 		.type = CRYPTO_ALG_TYPE_AEAD | CRYPTO_ALG_SUB_TYPE_CTR_SHA,
4283 		.is_registered = 0,
4284 		.alg.aead = {
4285 			.base = {
4286 
4287 				.cra_name = "authenc(hmac(sha256),rfc3686(ctr(aes)))",
4288 				.cra_driver_name =
4289 				"authenc-hmac-sha256-rfc3686-ctr-aes-chcr",
4290 				.cra_blocksize	 = 1,
4291 				.cra_priority = CHCR_AEAD_PRIORITY,
4292 				.cra_ctxsize =	sizeof(struct chcr_context) +
4293 						sizeof(struct chcr_aead_ctx) +
4294 						sizeof(struct chcr_authenc_ctx),
4295 
4296 			},
4297 			.ivsize = CTR_RFC3686_IV_SIZE,
4298 			.maxauthsize	= SHA256_DIGEST_SIZE,
4299 			.setkey = chcr_authenc_setkey,
4300 			.setauthsize = chcr_authenc_setauthsize,
4301 		}
4302 	},
4303 	{
4304 		.type = CRYPTO_ALG_TYPE_AEAD | CRYPTO_ALG_SUB_TYPE_CTR_SHA,
4305 		.is_registered = 0,
4306 		.alg.aead = {
4307 			.base = {
4308 				.cra_name = "authenc(hmac(sha224),rfc3686(ctr(aes)))",
4309 				.cra_driver_name =
4310 				"authenc-hmac-sha224-rfc3686-ctr-aes-chcr",
4311 				.cra_blocksize	 = 1,
4312 				.cra_priority = CHCR_AEAD_PRIORITY,
4313 				.cra_ctxsize =	sizeof(struct chcr_context) +
4314 						sizeof(struct chcr_aead_ctx) +
4315 						sizeof(struct chcr_authenc_ctx),
4316 			},
4317 			.ivsize = CTR_RFC3686_IV_SIZE,
4318 			.maxauthsize = SHA224_DIGEST_SIZE,
4319 			.setkey = chcr_authenc_setkey,
4320 			.setauthsize = chcr_authenc_setauthsize,
4321 		}
4322 	},
4323 	{
4324 		.type = CRYPTO_ALG_TYPE_AEAD | CRYPTO_ALG_SUB_TYPE_CTR_SHA,
4325 		.is_registered = 0,
4326 		.alg.aead = {
4327 			.base = {
4328 				.cra_name = "authenc(hmac(sha384),rfc3686(ctr(aes)))",
4329 				.cra_driver_name =
4330 				"authenc-hmac-sha384-rfc3686-ctr-aes-chcr",
4331 				.cra_blocksize	 = 1,
4332 				.cra_priority = CHCR_AEAD_PRIORITY,
4333 				.cra_ctxsize =	sizeof(struct chcr_context) +
4334 						sizeof(struct chcr_aead_ctx) +
4335 						sizeof(struct chcr_authenc_ctx),
4336 
4337 			},
4338 			.ivsize = CTR_RFC3686_IV_SIZE,
4339 			.maxauthsize = SHA384_DIGEST_SIZE,
4340 			.setkey = chcr_authenc_setkey,
4341 			.setauthsize = chcr_authenc_setauthsize,
4342 		}
4343 	},
4344 	{
4345 		.type = CRYPTO_ALG_TYPE_AEAD | CRYPTO_ALG_SUB_TYPE_CTR_SHA,
4346 		.is_registered = 0,
4347 		.alg.aead = {
4348 			.base = {
4349 				.cra_name = "authenc(hmac(sha512),rfc3686(ctr(aes)))",
4350 				.cra_driver_name =
4351 				"authenc-hmac-sha512-rfc3686-ctr-aes-chcr",
4352 				.cra_blocksize	 = 1,
4353 				.cra_priority = CHCR_AEAD_PRIORITY,
4354 				.cra_ctxsize =	sizeof(struct chcr_context) +
4355 						sizeof(struct chcr_aead_ctx) +
4356 						sizeof(struct chcr_authenc_ctx),
4357 
4358 			},
4359 			.ivsize = CTR_RFC3686_IV_SIZE,
4360 			.maxauthsize = SHA512_DIGEST_SIZE,
4361 			.setkey = chcr_authenc_setkey,
4362 			.setauthsize = chcr_authenc_setauthsize,
4363 		}
4364 	},
4365 	{
4366 		.type = CRYPTO_ALG_TYPE_AEAD | CRYPTO_ALG_SUB_TYPE_CTR_NULL,
4367 		.is_registered = 0,
4368 		.alg.aead = {
4369 			.base = {
4370 				.cra_name = "authenc(digest_null,rfc3686(ctr(aes)))",
4371 				.cra_driver_name =
4372 				"authenc-digest_null-rfc3686-ctr-aes-chcr",
4373 				.cra_blocksize	 = 1,
4374 				.cra_priority = CHCR_AEAD_PRIORITY,
4375 				.cra_ctxsize =	sizeof(struct chcr_context) +
4376 						sizeof(struct chcr_aead_ctx) +
4377 						sizeof(struct chcr_authenc_ctx),
4378 
4379 			},
4380 			.ivsize  = CTR_RFC3686_IV_SIZE,
4381 			.maxauthsize = 0,
4382 			.setkey  = chcr_aead_digest_null_setkey,
4383 			.setauthsize = chcr_authenc_null_setauthsize,
4384 		}
4385 	},
4386 };
4387 
4388 /*
4389  *	chcr_unregister_alg - Deregister crypto algorithms with
4390  *	kernel framework.
4391  */
4392 static int chcr_unregister_alg(void)
4393 {
4394 	int i;
4395 
4396 	for (i = 0; i < ARRAY_SIZE(driver_algs); i++) {
4397 		switch (driver_algs[i].type & CRYPTO_ALG_TYPE_MASK) {
4398 		case CRYPTO_ALG_TYPE_SKCIPHER:
4399 			if (driver_algs[i].is_registered && refcount_read(
4400 			    &driver_algs[i].alg.skcipher.base.cra_refcnt)
4401 			    == 1) {
4402 				crypto_unregister_skcipher(
4403 						&driver_algs[i].alg.skcipher);
4404 				driver_algs[i].is_registered = 0;
4405 			}
4406 			break;
4407 		case CRYPTO_ALG_TYPE_AEAD:
4408 			if (driver_algs[i].is_registered && refcount_read(
4409 			    &driver_algs[i].alg.aead.base.cra_refcnt) == 1) {
4410 				crypto_unregister_aead(
4411 						&driver_algs[i].alg.aead);
4412 				driver_algs[i].is_registered = 0;
4413 			}
4414 			break;
4415 		case CRYPTO_ALG_TYPE_AHASH:
4416 			if (driver_algs[i].is_registered && refcount_read(
4417 			    &driver_algs[i].alg.hash.halg.base.cra_refcnt)
4418 			    == 1) {
4419 				crypto_unregister_ahash(
4420 						&driver_algs[i].alg.hash);
4421 				driver_algs[i].is_registered = 0;
4422 			}
4423 			break;
4424 		}
4425 	}
4426 	return 0;
4427 }
4428 
4429 #define SZ_AHASH_CTX sizeof(struct chcr_context)
4430 #define SZ_AHASH_H_CTX (sizeof(struct chcr_context) + sizeof(struct hmac_ctx))
4431 #define SZ_AHASH_REQ_CTX sizeof(struct chcr_ahash_req_ctx)
4432 
4433 /*
4434  *	chcr_register_alg - Register crypto algorithms with kernel framework.
4435  */
4436 static int chcr_register_alg(void)
4437 {
4438 	struct crypto_alg ai;
4439 	struct ahash_alg *a_hash;
4440 	int err = 0, i;
4441 	char *name = NULL;
4442 
4443 	for (i = 0; i < ARRAY_SIZE(driver_algs); i++) {
4444 		if (driver_algs[i].is_registered)
4445 			continue;
4446 		switch (driver_algs[i].type & CRYPTO_ALG_TYPE_MASK) {
4447 		case CRYPTO_ALG_TYPE_SKCIPHER:
4448 			driver_algs[i].alg.skcipher.base.cra_priority =
4449 				CHCR_CRA_PRIORITY;
4450 			driver_algs[i].alg.skcipher.base.cra_module = THIS_MODULE;
4451 			driver_algs[i].alg.skcipher.base.cra_flags =
4452 				CRYPTO_ALG_TYPE_SKCIPHER | CRYPTO_ALG_ASYNC |
4453 				CRYPTO_ALG_ALLOCATES_MEMORY |
4454 				CRYPTO_ALG_NEED_FALLBACK;
4455 			driver_algs[i].alg.skcipher.base.cra_ctxsize =
4456 				sizeof(struct chcr_context) +
4457 				sizeof(struct ablk_ctx);
4458 			driver_algs[i].alg.skcipher.base.cra_alignmask = 0;
4459 
4460 			err = crypto_register_skcipher(&driver_algs[i].alg.skcipher);
4461 			name = driver_algs[i].alg.skcipher.base.cra_driver_name;
4462 			break;
4463 		case CRYPTO_ALG_TYPE_AEAD:
4464 			driver_algs[i].alg.aead.base.cra_flags =
4465 				CRYPTO_ALG_ASYNC | CRYPTO_ALG_NEED_FALLBACK |
4466 				CRYPTO_ALG_ALLOCATES_MEMORY;
4467 			driver_algs[i].alg.aead.encrypt = chcr_aead_encrypt;
4468 			driver_algs[i].alg.aead.decrypt = chcr_aead_decrypt;
4469 			driver_algs[i].alg.aead.init = chcr_aead_cra_init;
4470 			driver_algs[i].alg.aead.exit = chcr_aead_cra_exit;
4471 			driver_algs[i].alg.aead.base.cra_module = THIS_MODULE;
4472 			err = crypto_register_aead(&driver_algs[i].alg.aead);
4473 			name = driver_algs[i].alg.aead.base.cra_driver_name;
4474 			break;
4475 		case CRYPTO_ALG_TYPE_AHASH:
4476 			a_hash = &driver_algs[i].alg.hash;
4477 			a_hash->update = chcr_ahash_update;
4478 			a_hash->final = chcr_ahash_final;
4479 			a_hash->finup = chcr_ahash_finup;
4480 			a_hash->digest = chcr_ahash_digest;
4481 			a_hash->export = chcr_ahash_export;
4482 			a_hash->import = chcr_ahash_import;
4483 			a_hash->halg.statesize = SZ_AHASH_REQ_CTX;
4484 			a_hash->halg.base.cra_priority = CHCR_CRA_PRIORITY;
4485 			a_hash->halg.base.cra_module = THIS_MODULE;
4486 			a_hash->halg.base.cra_flags =
4487 				CRYPTO_ALG_ASYNC | CRYPTO_ALG_ALLOCATES_MEMORY;
4488 			a_hash->halg.base.cra_alignmask = 0;
4489 			a_hash->halg.base.cra_exit = NULL;
4490 
4491 			if (driver_algs[i].type == CRYPTO_ALG_TYPE_HMAC) {
4492 				a_hash->halg.base.cra_init = chcr_hmac_cra_init;
4493 				a_hash->halg.base.cra_exit = chcr_hmac_cra_exit;
4494 				a_hash->init = chcr_hmac_init;
4495 				a_hash->setkey = chcr_ahash_setkey;
4496 				a_hash->halg.base.cra_ctxsize = SZ_AHASH_H_CTX;
4497 			} else {
4498 				a_hash->init = chcr_sha_init;
4499 				a_hash->halg.base.cra_ctxsize = SZ_AHASH_CTX;
4500 				a_hash->halg.base.cra_init = chcr_sha_cra_init;
4501 			}
4502 			err = crypto_register_ahash(&driver_algs[i].alg.hash);
4503 			ai = driver_algs[i].alg.hash.halg.base;
4504 			name = ai.cra_driver_name;
4505 			break;
4506 		}
4507 		if (err) {
4508 			pr_err("%s : Algorithm registration failed\n", name);
4509 			goto register_err;
4510 		} else {
4511 			driver_algs[i].is_registered = 1;
4512 		}
4513 	}
4514 	return 0;
4515 
4516 register_err:
4517 	chcr_unregister_alg();
4518 	return err;
4519 }
4520 
4521 /*
4522  *	start_crypto - Register the crypto algorithms.
4523  *	This should called once when the first device comesup. After this
4524  *	kernel will start calling driver APIs for crypto operations.
4525  */
4526 int start_crypto(void)
4527 {
4528 	return chcr_register_alg();
4529 }
4530 
4531 /*
4532  *	stop_crypto - Deregister all the crypto algorithms with kernel.
4533  *	This should be called once when the last device goes down. After this
4534  *	kernel will not call the driver API for crypto operations.
4535  */
4536 int stop_crypto(void)
4537 {
4538 	chcr_unregister_alg();
4539 	return 0;
4540 }
4541