xref: /linux/drivers/crypto/ccp/ccp-dev.c (revision c0c914eca7f251c70facc37dfebeaf176601918d)
1 /*
2  * AMD Cryptographic Coprocessor (CCP) driver
3  *
4  * Copyright (C) 2013 Advanced Micro Devices, Inc.
5  *
6  * Author: Tom Lendacky <thomas.lendacky@amd.com>
7  *
8  * This program is free software; you can redistribute it and/or modify
9  * it under the terms of the GNU General Public License version 2 as
10  * published by the Free Software Foundation.
11  */
12 
13 #include <linux/module.h>
14 #include <linux/kernel.h>
15 #include <linux/kthread.h>
16 #include <linux/sched.h>
17 #include <linux/interrupt.h>
18 #include <linux/spinlock.h>
19 #include <linux/mutex.h>
20 #include <linux/delay.h>
21 #include <linux/hw_random.h>
22 #include <linux/cpu.h>
23 #ifdef CONFIG_X86
24 #include <asm/cpu_device_id.h>
25 #endif
26 #include <linux/ccp.h>
27 
28 #include "ccp-dev.h"
29 
30 MODULE_AUTHOR("Tom Lendacky <thomas.lendacky@amd.com>");
31 MODULE_LICENSE("GPL");
32 MODULE_VERSION("1.0.0");
33 MODULE_DESCRIPTION("AMD Cryptographic Coprocessor driver");
34 
35 struct ccp_tasklet_data {
36 	struct completion completion;
37 	struct ccp_cmd *cmd;
38 };
39 
40 static struct ccp_device *ccp_dev;
41 static inline struct ccp_device *ccp_get_device(void)
42 {
43 	return ccp_dev;
44 }
45 
46 static inline void ccp_add_device(struct ccp_device *ccp)
47 {
48 	ccp_dev = ccp;
49 }
50 
51 static inline void ccp_del_device(struct ccp_device *ccp)
52 {
53 	ccp_dev = NULL;
54 }
55 
56 /**
57  * ccp_present - check if a CCP device is present
58  *
59  * Returns zero if a CCP device is present, -ENODEV otherwise.
60  */
61 int ccp_present(void)
62 {
63 	if (ccp_get_device())
64 		return 0;
65 
66 	return -ENODEV;
67 }
68 EXPORT_SYMBOL_GPL(ccp_present);
69 
70 /**
71  * ccp_enqueue_cmd - queue an operation for processing by the CCP
72  *
73  * @cmd: ccp_cmd struct to be processed
74  *
75  * Queue a cmd to be processed by the CCP. If queueing the cmd
76  * would exceed the defined length of the cmd queue the cmd will
77  * only be queued if the CCP_CMD_MAY_BACKLOG flag is set and will
78  * result in a return code of -EBUSY.
79  *
80  * The callback routine specified in the ccp_cmd struct will be
81  * called to notify the caller of completion (if the cmd was not
82  * backlogged) or advancement out of the backlog. If the cmd has
83  * advanced out of the backlog the "err" value of the callback
84  * will be -EINPROGRESS. Any other "err" value during callback is
85  * the result of the operation.
86  *
87  * The cmd has been successfully queued if:
88  *   the return code is -EINPROGRESS or
89  *   the return code is -EBUSY and CCP_CMD_MAY_BACKLOG flag is set
90  */
91 int ccp_enqueue_cmd(struct ccp_cmd *cmd)
92 {
93 	struct ccp_device *ccp = ccp_get_device();
94 	unsigned long flags;
95 	unsigned int i;
96 	int ret;
97 
98 	if (!ccp)
99 		return -ENODEV;
100 
101 	/* Caller must supply a callback routine */
102 	if (!cmd->callback)
103 		return -EINVAL;
104 
105 	cmd->ccp = ccp;
106 
107 	spin_lock_irqsave(&ccp->cmd_lock, flags);
108 
109 	i = ccp->cmd_q_count;
110 
111 	if (ccp->cmd_count >= MAX_CMD_QLEN) {
112 		ret = -EBUSY;
113 		if (cmd->flags & CCP_CMD_MAY_BACKLOG)
114 			list_add_tail(&cmd->entry, &ccp->backlog);
115 	} else {
116 		ret = -EINPROGRESS;
117 		ccp->cmd_count++;
118 		list_add_tail(&cmd->entry, &ccp->cmd);
119 
120 		/* Find an idle queue */
121 		if (!ccp->suspending) {
122 			for (i = 0; i < ccp->cmd_q_count; i++) {
123 				if (ccp->cmd_q[i].active)
124 					continue;
125 
126 				break;
127 			}
128 		}
129 	}
130 
131 	spin_unlock_irqrestore(&ccp->cmd_lock, flags);
132 
133 	/* If we found an idle queue, wake it up */
134 	if (i < ccp->cmd_q_count)
135 		wake_up_process(ccp->cmd_q[i].kthread);
136 
137 	return ret;
138 }
139 EXPORT_SYMBOL_GPL(ccp_enqueue_cmd);
140 
141 static void ccp_do_cmd_backlog(struct work_struct *work)
142 {
143 	struct ccp_cmd *cmd = container_of(work, struct ccp_cmd, work);
144 	struct ccp_device *ccp = cmd->ccp;
145 	unsigned long flags;
146 	unsigned int i;
147 
148 	cmd->callback(cmd->data, -EINPROGRESS);
149 
150 	spin_lock_irqsave(&ccp->cmd_lock, flags);
151 
152 	ccp->cmd_count++;
153 	list_add_tail(&cmd->entry, &ccp->cmd);
154 
155 	/* Find an idle queue */
156 	for (i = 0; i < ccp->cmd_q_count; i++) {
157 		if (ccp->cmd_q[i].active)
158 			continue;
159 
160 		break;
161 	}
162 
163 	spin_unlock_irqrestore(&ccp->cmd_lock, flags);
164 
165 	/* If we found an idle queue, wake it up */
166 	if (i < ccp->cmd_q_count)
167 		wake_up_process(ccp->cmd_q[i].kthread);
168 }
169 
170 static struct ccp_cmd *ccp_dequeue_cmd(struct ccp_cmd_queue *cmd_q)
171 {
172 	struct ccp_device *ccp = cmd_q->ccp;
173 	struct ccp_cmd *cmd = NULL;
174 	struct ccp_cmd *backlog = NULL;
175 	unsigned long flags;
176 
177 	spin_lock_irqsave(&ccp->cmd_lock, flags);
178 
179 	cmd_q->active = 0;
180 
181 	if (ccp->suspending) {
182 		cmd_q->suspended = 1;
183 
184 		spin_unlock_irqrestore(&ccp->cmd_lock, flags);
185 		wake_up_interruptible(&ccp->suspend_queue);
186 
187 		return NULL;
188 	}
189 
190 	if (ccp->cmd_count) {
191 		cmd_q->active = 1;
192 
193 		cmd = list_first_entry(&ccp->cmd, struct ccp_cmd, entry);
194 		list_del(&cmd->entry);
195 
196 		ccp->cmd_count--;
197 	}
198 
199 	if (!list_empty(&ccp->backlog)) {
200 		backlog = list_first_entry(&ccp->backlog, struct ccp_cmd,
201 					   entry);
202 		list_del(&backlog->entry);
203 	}
204 
205 	spin_unlock_irqrestore(&ccp->cmd_lock, flags);
206 
207 	if (backlog) {
208 		INIT_WORK(&backlog->work, ccp_do_cmd_backlog);
209 		schedule_work(&backlog->work);
210 	}
211 
212 	return cmd;
213 }
214 
215 static void ccp_do_cmd_complete(unsigned long data)
216 {
217 	struct ccp_tasklet_data *tdata = (struct ccp_tasklet_data *)data;
218 	struct ccp_cmd *cmd = tdata->cmd;
219 
220 	cmd->callback(cmd->data, cmd->ret);
221 	complete(&tdata->completion);
222 }
223 
224 static int ccp_cmd_queue_thread(void *data)
225 {
226 	struct ccp_cmd_queue *cmd_q = (struct ccp_cmd_queue *)data;
227 	struct ccp_cmd *cmd;
228 	struct ccp_tasklet_data tdata;
229 	struct tasklet_struct tasklet;
230 
231 	tasklet_init(&tasklet, ccp_do_cmd_complete, (unsigned long)&tdata);
232 
233 	set_current_state(TASK_INTERRUPTIBLE);
234 	while (!kthread_should_stop()) {
235 		schedule();
236 
237 		set_current_state(TASK_INTERRUPTIBLE);
238 
239 		cmd = ccp_dequeue_cmd(cmd_q);
240 		if (!cmd)
241 			continue;
242 
243 		__set_current_state(TASK_RUNNING);
244 
245 		/* Execute the command */
246 		cmd->ret = ccp_run_cmd(cmd_q, cmd);
247 
248 		/* Schedule the completion callback */
249 		tdata.cmd = cmd;
250 		init_completion(&tdata.completion);
251 		tasklet_schedule(&tasklet);
252 		wait_for_completion(&tdata.completion);
253 	}
254 
255 	__set_current_state(TASK_RUNNING);
256 
257 	return 0;
258 }
259 
260 static int ccp_trng_read(struct hwrng *rng, void *data, size_t max, bool wait)
261 {
262 	struct ccp_device *ccp = container_of(rng, struct ccp_device, hwrng);
263 	u32 trng_value;
264 	int len = min_t(int, sizeof(trng_value), max);
265 
266 	/*
267 	 * Locking is provided by the caller so we can update device
268 	 * hwrng-related fields safely
269 	 */
270 	trng_value = ioread32(ccp->io_regs + TRNG_OUT_REG);
271 	if (!trng_value) {
272 		/* Zero is returned if not data is available or if a
273 		 * bad-entropy error is present. Assume an error if
274 		 * we exceed TRNG_RETRIES reads of zero.
275 		 */
276 		if (ccp->hwrng_retries++ > TRNG_RETRIES)
277 			return -EIO;
278 
279 		return 0;
280 	}
281 
282 	/* Reset the counter and save the rng value */
283 	ccp->hwrng_retries = 0;
284 	memcpy(data, &trng_value, len);
285 
286 	return len;
287 }
288 
289 /**
290  * ccp_alloc_struct - allocate and initialize the ccp_device struct
291  *
292  * @dev: device struct of the CCP
293  */
294 struct ccp_device *ccp_alloc_struct(struct device *dev)
295 {
296 	struct ccp_device *ccp;
297 
298 	ccp = devm_kzalloc(dev, sizeof(*ccp), GFP_KERNEL);
299 	if (!ccp)
300 		return NULL;
301 	ccp->dev = dev;
302 
303 	INIT_LIST_HEAD(&ccp->cmd);
304 	INIT_LIST_HEAD(&ccp->backlog);
305 
306 	spin_lock_init(&ccp->cmd_lock);
307 	mutex_init(&ccp->req_mutex);
308 	mutex_init(&ccp->ksb_mutex);
309 	ccp->ksb_count = KSB_COUNT;
310 	ccp->ksb_start = 0;
311 
312 	return ccp;
313 }
314 
315 /**
316  * ccp_init - initialize the CCP device
317  *
318  * @ccp: ccp_device struct
319  */
320 int ccp_init(struct ccp_device *ccp)
321 {
322 	struct device *dev = ccp->dev;
323 	struct ccp_cmd_queue *cmd_q;
324 	struct dma_pool *dma_pool;
325 	char dma_pool_name[MAX_DMAPOOL_NAME_LEN];
326 	unsigned int qmr, qim, i;
327 	int ret;
328 
329 	/* Find available queues */
330 	qim = 0;
331 	qmr = ioread32(ccp->io_regs + Q_MASK_REG);
332 	for (i = 0; i < MAX_HW_QUEUES; i++) {
333 		if (!(qmr & (1 << i)))
334 			continue;
335 
336 		/* Allocate a dma pool for this queue */
337 		snprintf(dma_pool_name, sizeof(dma_pool_name), "ccp_q%d", i);
338 		dma_pool = dma_pool_create(dma_pool_name, dev,
339 					   CCP_DMAPOOL_MAX_SIZE,
340 					   CCP_DMAPOOL_ALIGN, 0);
341 		if (!dma_pool) {
342 			dev_err(dev, "unable to allocate dma pool\n");
343 			ret = -ENOMEM;
344 			goto e_pool;
345 		}
346 
347 		cmd_q = &ccp->cmd_q[ccp->cmd_q_count];
348 		ccp->cmd_q_count++;
349 
350 		cmd_q->ccp = ccp;
351 		cmd_q->id = i;
352 		cmd_q->dma_pool = dma_pool;
353 
354 		/* Reserve 2 KSB regions for the queue */
355 		cmd_q->ksb_key = KSB_START + ccp->ksb_start++;
356 		cmd_q->ksb_ctx = KSB_START + ccp->ksb_start++;
357 		ccp->ksb_count -= 2;
358 
359 		/* Preset some register values and masks that are queue
360 		 * number dependent
361 		 */
362 		cmd_q->reg_status = ccp->io_regs + CMD_Q_STATUS_BASE +
363 				    (CMD_Q_STATUS_INCR * i);
364 		cmd_q->reg_int_status = ccp->io_regs + CMD_Q_INT_STATUS_BASE +
365 					(CMD_Q_STATUS_INCR * i);
366 		cmd_q->int_ok = 1 << (i * 2);
367 		cmd_q->int_err = 1 << ((i * 2) + 1);
368 
369 		cmd_q->free_slots = CMD_Q_DEPTH(ioread32(cmd_q->reg_status));
370 
371 		init_waitqueue_head(&cmd_q->int_queue);
372 
373 		/* Build queue interrupt mask (two interrupts per queue) */
374 		qim |= cmd_q->int_ok | cmd_q->int_err;
375 
376 #ifdef CONFIG_ARM64
377 		/* For arm64 set the recommended queue cache settings */
378 		iowrite32(ccp->axcache, ccp->io_regs + CMD_Q_CACHE_BASE +
379 			  (CMD_Q_CACHE_INC * i));
380 #endif
381 
382 		dev_dbg(dev, "queue #%u available\n", i);
383 	}
384 	if (ccp->cmd_q_count == 0) {
385 		dev_notice(dev, "no command queues available\n");
386 		ret = -EIO;
387 		goto e_pool;
388 	}
389 	dev_notice(dev, "%u command queues available\n", ccp->cmd_q_count);
390 
391 	/* Disable and clear interrupts until ready */
392 	iowrite32(0x00, ccp->io_regs + IRQ_MASK_REG);
393 	for (i = 0; i < ccp->cmd_q_count; i++) {
394 		cmd_q = &ccp->cmd_q[i];
395 
396 		ioread32(cmd_q->reg_int_status);
397 		ioread32(cmd_q->reg_status);
398 	}
399 	iowrite32(qim, ccp->io_regs + IRQ_STATUS_REG);
400 
401 	/* Request an irq */
402 	ret = ccp->get_irq(ccp);
403 	if (ret) {
404 		dev_err(dev, "unable to allocate an IRQ\n");
405 		goto e_pool;
406 	}
407 
408 	/* Initialize the queues used to wait for KSB space and suspend */
409 	init_waitqueue_head(&ccp->ksb_queue);
410 	init_waitqueue_head(&ccp->suspend_queue);
411 
412 	/* Create a kthread for each queue */
413 	for (i = 0; i < ccp->cmd_q_count; i++) {
414 		struct task_struct *kthread;
415 
416 		cmd_q = &ccp->cmd_q[i];
417 
418 		kthread = kthread_create(ccp_cmd_queue_thread, cmd_q,
419 					 "ccp-q%u", cmd_q->id);
420 		if (IS_ERR(kthread)) {
421 			dev_err(dev, "error creating queue thread (%ld)\n",
422 				PTR_ERR(kthread));
423 			ret = PTR_ERR(kthread);
424 			goto e_kthread;
425 		}
426 
427 		cmd_q->kthread = kthread;
428 		wake_up_process(kthread);
429 	}
430 
431 	/* Register the RNG */
432 	ccp->hwrng.name = "ccp-rng";
433 	ccp->hwrng.read = ccp_trng_read;
434 	ret = hwrng_register(&ccp->hwrng);
435 	if (ret) {
436 		dev_err(dev, "error registering hwrng (%d)\n", ret);
437 		goto e_kthread;
438 	}
439 
440 	/* Make the device struct available before enabling interrupts */
441 	ccp_add_device(ccp);
442 
443 	/* Enable interrupts */
444 	iowrite32(qim, ccp->io_regs + IRQ_MASK_REG);
445 
446 	return 0;
447 
448 e_kthread:
449 	for (i = 0; i < ccp->cmd_q_count; i++)
450 		if (ccp->cmd_q[i].kthread)
451 			kthread_stop(ccp->cmd_q[i].kthread);
452 
453 	ccp->free_irq(ccp);
454 
455 e_pool:
456 	for (i = 0; i < ccp->cmd_q_count; i++)
457 		dma_pool_destroy(ccp->cmd_q[i].dma_pool);
458 
459 	return ret;
460 }
461 
462 /**
463  * ccp_destroy - tear down the CCP device
464  *
465  * @ccp: ccp_device struct
466  */
467 void ccp_destroy(struct ccp_device *ccp)
468 {
469 	struct ccp_cmd_queue *cmd_q;
470 	struct ccp_cmd *cmd;
471 	unsigned int qim, i;
472 
473 	/* Remove general access to the device struct */
474 	ccp_del_device(ccp);
475 
476 	/* Unregister the RNG */
477 	hwrng_unregister(&ccp->hwrng);
478 
479 	/* Stop the queue kthreads */
480 	for (i = 0; i < ccp->cmd_q_count; i++)
481 		if (ccp->cmd_q[i].kthread)
482 			kthread_stop(ccp->cmd_q[i].kthread);
483 
484 	/* Build queue interrupt mask (two interrupt masks per queue) */
485 	qim = 0;
486 	for (i = 0; i < ccp->cmd_q_count; i++) {
487 		cmd_q = &ccp->cmd_q[i];
488 		qim |= cmd_q->int_ok | cmd_q->int_err;
489 	}
490 
491 	/* Disable and clear interrupts */
492 	iowrite32(0x00, ccp->io_regs + IRQ_MASK_REG);
493 	for (i = 0; i < ccp->cmd_q_count; i++) {
494 		cmd_q = &ccp->cmd_q[i];
495 
496 		ioread32(cmd_q->reg_int_status);
497 		ioread32(cmd_q->reg_status);
498 	}
499 	iowrite32(qim, ccp->io_regs + IRQ_STATUS_REG);
500 
501 	ccp->free_irq(ccp);
502 
503 	for (i = 0; i < ccp->cmd_q_count; i++)
504 		dma_pool_destroy(ccp->cmd_q[i].dma_pool);
505 
506 	/* Flush the cmd and backlog queue */
507 	while (!list_empty(&ccp->cmd)) {
508 		/* Invoke the callback directly with an error code */
509 		cmd = list_first_entry(&ccp->cmd, struct ccp_cmd, entry);
510 		list_del(&cmd->entry);
511 		cmd->callback(cmd->data, -ENODEV);
512 	}
513 	while (!list_empty(&ccp->backlog)) {
514 		/* Invoke the callback directly with an error code */
515 		cmd = list_first_entry(&ccp->backlog, struct ccp_cmd, entry);
516 		list_del(&cmd->entry);
517 		cmd->callback(cmd->data, -ENODEV);
518 	}
519 }
520 
521 /**
522  * ccp_irq_handler - handle interrupts generated by the CCP device
523  *
524  * @irq: the irq associated with the interrupt
525  * @data: the data value supplied when the irq was created
526  */
527 irqreturn_t ccp_irq_handler(int irq, void *data)
528 {
529 	struct device *dev = data;
530 	struct ccp_device *ccp = dev_get_drvdata(dev);
531 	struct ccp_cmd_queue *cmd_q;
532 	u32 q_int, status;
533 	unsigned int i;
534 
535 	status = ioread32(ccp->io_regs + IRQ_STATUS_REG);
536 
537 	for (i = 0; i < ccp->cmd_q_count; i++) {
538 		cmd_q = &ccp->cmd_q[i];
539 
540 		q_int = status & (cmd_q->int_ok | cmd_q->int_err);
541 		if (q_int) {
542 			cmd_q->int_status = status;
543 			cmd_q->q_status = ioread32(cmd_q->reg_status);
544 			cmd_q->q_int_status = ioread32(cmd_q->reg_int_status);
545 
546 			/* On error, only save the first error value */
547 			if ((q_int & cmd_q->int_err) && !cmd_q->cmd_error)
548 				cmd_q->cmd_error = CMD_Q_ERROR(cmd_q->q_status);
549 
550 			cmd_q->int_rcvd = 1;
551 
552 			/* Acknowledge the interrupt and wake the kthread */
553 			iowrite32(q_int, ccp->io_regs + IRQ_STATUS_REG);
554 			wake_up_interruptible(&cmd_q->int_queue);
555 		}
556 	}
557 
558 	return IRQ_HANDLED;
559 }
560 
561 #ifdef CONFIG_PM
562 bool ccp_queues_suspended(struct ccp_device *ccp)
563 {
564 	unsigned int suspended = 0;
565 	unsigned long flags;
566 	unsigned int i;
567 
568 	spin_lock_irqsave(&ccp->cmd_lock, flags);
569 
570 	for (i = 0; i < ccp->cmd_q_count; i++)
571 		if (ccp->cmd_q[i].suspended)
572 			suspended++;
573 
574 	spin_unlock_irqrestore(&ccp->cmd_lock, flags);
575 
576 	return ccp->cmd_q_count == suspended;
577 }
578 #endif
579 
580 #ifdef CONFIG_X86
581 static const struct x86_cpu_id ccp_support[] = {
582 	{ X86_VENDOR_AMD, 22, },
583 	{ },
584 };
585 #endif
586 
587 static int __init ccp_mod_init(void)
588 {
589 #ifdef CONFIG_X86
590 	struct cpuinfo_x86 *cpuinfo = &boot_cpu_data;
591 	int ret;
592 
593 	if (!x86_match_cpu(ccp_support))
594 		return -ENODEV;
595 
596 	switch (cpuinfo->x86) {
597 	case 22:
598 		if ((cpuinfo->x86_model < 48) || (cpuinfo->x86_model > 63))
599 			return -ENODEV;
600 
601 		ret = ccp_pci_init();
602 		if (ret)
603 			return ret;
604 
605 		/* Don't leave the driver loaded if init failed */
606 		if (!ccp_get_device()) {
607 			ccp_pci_exit();
608 			return -ENODEV;
609 		}
610 
611 		return 0;
612 
613 		break;
614 	}
615 #endif
616 
617 #ifdef CONFIG_ARM64
618 	int ret;
619 
620 	ret = ccp_platform_init();
621 	if (ret)
622 		return ret;
623 
624 	/* Don't leave the driver loaded if init failed */
625 	if (!ccp_get_device()) {
626 		ccp_platform_exit();
627 		return -ENODEV;
628 	}
629 
630 	return 0;
631 #endif
632 
633 	return -ENODEV;
634 }
635 
636 static void __exit ccp_mod_exit(void)
637 {
638 #ifdef CONFIG_X86
639 	struct cpuinfo_x86 *cpuinfo = &boot_cpu_data;
640 
641 	switch (cpuinfo->x86) {
642 	case 22:
643 		ccp_pci_exit();
644 		break;
645 	}
646 #endif
647 
648 #ifdef CONFIG_ARM64
649 	ccp_platform_exit();
650 #endif
651 }
652 
653 module_init(ccp_mod_init);
654 module_exit(ccp_mod_exit);
655