xref: /linux/arch/arm64/crypto/aes-ce-ccm-core.S (revision 9e56ff53b4115875667760445b028357848b4748)
1/* SPDX-License-Identifier: GPL-2.0-only */
2/*
3 * aesce-ccm-core.S - AES-CCM transform for ARMv8 with Crypto Extensions
4 *
5 * Copyright (C) 2013 - 2017 Linaro Ltd <ard.biesheuvel@linaro.org>
6 */
7
8#include <linux/linkage.h>
9#include <asm/assembler.h>
10
11	.text
12	.arch	armv8-a+crypto
13
14	/*
15	 * u32 ce_aes_ccm_auth_data(u8 mac[], u8 const in[], u32 abytes,
16	 *			    u32 macp, u8 const rk[], u32 rounds);
17	 */
18SYM_FUNC_START(ce_aes_ccm_auth_data)
19	ld1	{v0.16b}, [x0]			/* load mac */
20	cbz	w3, 1f
21	sub	w3, w3, #16
22	eor	v1.16b, v1.16b, v1.16b
230:	ldrb	w7, [x1], #1			/* get 1 byte of input */
24	subs	w2, w2, #1
25	add	w3, w3, #1
26	ins	v1.b[0], w7
27	ext	v1.16b, v1.16b, v1.16b, #1	/* rotate in the input bytes */
28	beq	8f				/* out of input? */
29	cbnz	w3, 0b
30	eor	v0.16b, v0.16b, v1.16b
311:	ld1	{v3.4s}, [x4]			/* load first round key */
32	prfm	pldl1strm, [x1]
33	cmp	w5, #12				/* which key size? */
34	add	x6, x4, #16
35	sub	w7, w5, #2			/* modified # of rounds */
36	bmi	2f
37	bne	5f
38	mov	v5.16b, v3.16b
39	b	4f
402:	mov	v4.16b, v3.16b
41	ld1	{v5.4s}, [x6], #16		/* load 2nd round key */
423:	aese	v0.16b, v4.16b
43	aesmc	v0.16b, v0.16b
444:	ld1	{v3.4s}, [x6], #16		/* load next round key */
45	aese	v0.16b, v5.16b
46	aesmc	v0.16b, v0.16b
475:	ld1	{v4.4s}, [x6], #16		/* load next round key */
48	subs	w7, w7, #3
49	aese	v0.16b, v3.16b
50	aesmc	v0.16b, v0.16b
51	ld1	{v5.4s}, [x6], #16		/* load next round key */
52	bpl	3b
53	aese	v0.16b, v4.16b
54	subs	w2, w2, #16			/* last data? */
55	eor	v0.16b, v0.16b, v5.16b		/* final round */
56	bmi	6f
57	ld1	{v1.16b}, [x1], #16		/* load next input block */
58	eor	v0.16b, v0.16b, v1.16b		/* xor with mac */
59	bne	1b
606:	st1	{v0.16b}, [x0]			/* store mac */
61	beq	10f
62	adds	w2, w2, #16
63	beq	10f
64	mov	w3, w2
657:	ldrb	w7, [x1], #1
66	umov	w6, v0.b[0]
67	eor	w6, w6, w7
68	strb	w6, [x0], #1
69	subs	w2, w2, #1
70	beq	10f
71	ext	v0.16b, v0.16b, v0.16b, #1	/* rotate out the mac bytes */
72	b	7b
738:	cbz	w3, 91f
74	mov	w7, w3
75	add	w3, w3, #16
769:	ext	v1.16b, v1.16b, v1.16b, #1
77	adds	w7, w7, #1
78	bne	9b
7991:	eor	v0.16b, v0.16b, v1.16b
80	st1	{v0.16b}, [x0]
8110:	mov	w0, w3
82	ret
83SYM_FUNC_END(ce_aes_ccm_auth_data)
84
85	/*
86	 * void ce_aes_ccm_final(u8 mac[], u8 const ctr[], u8 const rk[],
87	 * 			 u32 rounds);
88	 */
89SYM_FUNC_START(ce_aes_ccm_final)
90	ld1	{v3.4s}, [x2], #16		/* load first round key */
91	ld1	{v0.16b}, [x0]			/* load mac */
92	cmp	w3, #12				/* which key size? */
93	sub	w3, w3, #2			/* modified # of rounds */
94	ld1	{v1.16b}, [x1]			/* load 1st ctriv */
95	bmi	0f
96	bne	3f
97	mov	v5.16b, v3.16b
98	b	2f
990:	mov	v4.16b, v3.16b
1001:	ld1	{v5.4s}, [x2], #16		/* load next round key */
101	aese	v0.16b, v4.16b
102	aesmc	v0.16b, v0.16b
103	aese	v1.16b, v4.16b
104	aesmc	v1.16b, v1.16b
1052:	ld1	{v3.4s}, [x2], #16		/* load next round key */
106	aese	v0.16b, v5.16b
107	aesmc	v0.16b, v0.16b
108	aese	v1.16b, v5.16b
109	aesmc	v1.16b, v1.16b
1103:	ld1	{v4.4s}, [x2], #16		/* load next round key */
111	subs	w3, w3, #3
112	aese	v0.16b, v3.16b
113	aesmc	v0.16b, v0.16b
114	aese	v1.16b, v3.16b
115	aesmc	v1.16b, v1.16b
116	bpl	1b
117	aese	v0.16b, v4.16b
118	aese	v1.16b, v4.16b
119	/* final round key cancels out */
120	eor	v0.16b, v0.16b, v1.16b		/* en-/decrypt the mac */
121	st1	{v0.16b}, [x0]			/* store result */
122	ret
123SYM_FUNC_END(ce_aes_ccm_final)
124
125	.macro	aes_ccm_do_crypt,enc
126	cbz	x2, 5f
127	ldr	x8, [x6, #8]			/* load lower ctr */
128	ld1	{v0.16b}, [x5]			/* load mac */
129CPU_LE(	rev	x8, x8			)	/* keep swabbed ctr in reg */
1300:	/* outer loop */
131	ld1	{v1.8b}, [x6]			/* load upper ctr */
132	prfm	pldl1strm, [x1]
133	add	x8, x8, #1
134	rev	x9, x8
135	cmp	w4, #12				/* which key size? */
136	sub	w7, w4, #2			/* get modified # of rounds */
137	ins	v1.d[1], x9			/* no carry in lower ctr */
138	ld1	{v3.4s}, [x3]			/* load first round key */
139	add	x10, x3, #16
140	bmi	1f
141	bne	4f
142	mov	v5.16b, v3.16b
143	b	3f
1441:	mov	v4.16b, v3.16b
145	ld1	{v5.4s}, [x10], #16		/* load 2nd round key */
1462:	/* inner loop: 3 rounds, 2x interleaved */
147	aese	v0.16b, v4.16b
148	aesmc	v0.16b, v0.16b
149	aese	v1.16b, v4.16b
150	aesmc	v1.16b, v1.16b
1513:	ld1	{v3.4s}, [x10], #16		/* load next round key */
152	aese	v0.16b, v5.16b
153	aesmc	v0.16b, v0.16b
154	aese	v1.16b, v5.16b
155	aesmc	v1.16b, v1.16b
1564:	ld1	{v4.4s}, [x10], #16		/* load next round key */
157	subs	w7, w7, #3
158	aese	v0.16b, v3.16b
159	aesmc	v0.16b, v0.16b
160	aese	v1.16b, v3.16b
161	aesmc	v1.16b, v1.16b
162	ld1	{v5.4s}, [x10], #16		/* load next round key */
163	bpl	2b
164	aese	v0.16b, v4.16b
165	aese	v1.16b, v4.16b
166	subs	w2, w2, #16
167	bmi	6f				/* partial block? */
168	ld1	{v2.16b}, [x1], #16		/* load next input block */
169	.if	\enc == 1
170	eor	v2.16b, v2.16b, v5.16b		/* final round enc+mac */
171	eor	v1.16b, v1.16b, v2.16b		/* xor with crypted ctr */
172	.else
173	eor	v2.16b, v2.16b, v1.16b		/* xor with crypted ctr */
174	eor	v1.16b, v2.16b, v5.16b		/* final round enc */
175	.endif
176	eor	v0.16b, v0.16b, v2.16b		/* xor mac with pt ^ rk[last] */
177	st1	{v1.16b}, [x0], #16		/* write output block */
178	bne	0b
179CPU_LE(	rev	x8, x8			)
180	st1	{v0.16b}, [x5]			/* store mac */
181	str	x8, [x6, #8]			/* store lsb end of ctr (BE) */
1825:	ret
183
1846:	eor	v0.16b, v0.16b, v5.16b		/* final round mac */
185	eor	v1.16b, v1.16b, v5.16b		/* final round enc */
186	st1	{v0.16b}, [x5]			/* store mac */
187	add	w2, w2, #16			/* process partial tail block */
1887:	ldrb	w9, [x1], #1			/* get 1 byte of input */
189	umov	w6, v1.b[0]			/* get top crypted ctr byte */
190	umov	w7, v0.b[0]			/* get top mac byte */
191	.if	\enc == 1
192	eor	w7, w7, w9
193	eor	w9, w9, w6
194	.else
195	eor	w9, w9, w6
196	eor	w7, w7, w9
197	.endif
198	strb	w9, [x0], #1			/* store out byte */
199	strb	w7, [x5], #1			/* store mac byte */
200	subs	w2, w2, #1
201	beq	5b
202	ext	v0.16b, v0.16b, v0.16b, #1	/* shift out mac byte */
203	ext	v1.16b, v1.16b, v1.16b, #1	/* shift out ctr byte */
204	b	7b
205	.endm
206
207	/*
208	 * void ce_aes_ccm_encrypt(u8 out[], u8 const in[], u32 cbytes,
209	 * 			   u8 const rk[], u32 rounds, u8 mac[],
210	 * 			   u8 ctr[]);
211	 * void ce_aes_ccm_decrypt(u8 out[], u8 const in[], u32 cbytes,
212	 * 			   u8 const rk[], u32 rounds, u8 mac[],
213	 * 			   u8 ctr[]);
214	 */
215SYM_FUNC_START(ce_aes_ccm_encrypt)
216	aes_ccm_do_crypt	1
217SYM_FUNC_END(ce_aes_ccm_encrypt)
218
219SYM_FUNC_START(ce_aes_ccm_decrypt)
220	aes_ccm_do_crypt	0
221SYM_FUNC_END(ce_aes_ccm_decrypt)
222