1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * Cryptographic API. 4 * 5 * Support for ATMEL DES/TDES HW acceleration. 6 * 7 * Copyright (c) 2012 Eukréa Electromatique - ATMEL 8 * Author: Nicolas Royer <nicolas@eukrea.com> 9 * 10 * Some ideas are from omap-aes.c drivers. 11 */ 12 13 14 #include <linux/kernel.h> 15 #include <linux/module.h> 16 #include <linux/slab.h> 17 #include <linux/err.h> 18 #include <linux/clk.h> 19 #include <linux/io.h> 20 #include <linux/hw_random.h> 21 #include <linux/platform_device.h> 22 23 #include <linux/device.h> 24 #include <linux/dmaengine.h> 25 #include <linux/init.h> 26 #include <linux/errno.h> 27 #include <linux/interrupt.h> 28 #include <linux/irq.h> 29 #include <linux/scatterlist.h> 30 #include <linux/dma-mapping.h> 31 #include <linux/mod_devicetable.h> 32 #include <linux/delay.h> 33 #include <linux/crypto.h> 34 #include <crypto/scatterwalk.h> 35 #include <crypto/algapi.h> 36 #include <crypto/internal/des.h> 37 #include <crypto/internal/skcipher.h> 38 #include "atmel-tdes-regs.h" 39 40 #define ATMEL_TDES_PRIORITY 300 41 42 /* TDES flags */ 43 /* Reserve bits [17:16], [13:12], [2:0] for AES Mode Register */ 44 #define TDES_FLAGS_ENCRYPT TDES_MR_CYPHER_ENC 45 #define TDES_FLAGS_OPMODE_MASK (TDES_MR_OPMOD_MASK | TDES_MR_CFBS_MASK) 46 #define TDES_FLAGS_ECB TDES_MR_OPMOD_ECB 47 #define TDES_FLAGS_CBC TDES_MR_OPMOD_CBC 48 49 #define TDES_FLAGS_MODE_MASK (TDES_FLAGS_OPMODE_MASK | TDES_FLAGS_ENCRYPT) 50 51 #define TDES_FLAGS_INIT BIT(3) 52 #define TDES_FLAGS_FAST BIT(4) 53 #define TDES_FLAGS_BUSY BIT(5) 54 #define TDES_FLAGS_DMA BIT(6) 55 56 #define ATMEL_TDES_QUEUE_LENGTH 50 57 58 struct atmel_tdes_caps { 59 bool has_dma; 60 }; 61 62 struct atmel_tdes_dev; 63 64 struct atmel_tdes_ctx { 65 struct atmel_tdes_dev *dd; 66 67 int keylen; 68 u32 key[DES3_EDE_KEY_SIZE / sizeof(u32)]; 69 unsigned long flags; 70 71 u16 block_size; 72 }; 73 74 struct atmel_tdes_reqctx { 75 unsigned long mode; 76 u8 lastc[DES_BLOCK_SIZE]; 77 }; 78 79 struct atmel_tdes_dma { 80 struct dma_chan *chan; 81 struct dma_slave_config dma_conf; 82 }; 83 84 struct atmel_tdes_dev { 85 struct list_head list; 86 unsigned long phys_base; 87 void __iomem *io_base; 88 89 struct atmel_tdes_ctx *ctx; 90 struct device *dev; 91 struct clk *iclk; 92 int irq; 93 94 unsigned long flags; 95 96 spinlock_t lock; 97 struct crypto_queue queue; 98 99 struct tasklet_struct done_task; 100 struct tasklet_struct queue_task; 101 102 struct skcipher_request *req; 103 size_t total; 104 105 struct scatterlist *in_sg; 106 unsigned int nb_in_sg; 107 size_t in_offset; 108 struct scatterlist *out_sg; 109 unsigned int nb_out_sg; 110 size_t out_offset; 111 112 size_t buflen; 113 size_t dma_size; 114 115 void *buf_in; 116 int dma_in; 117 dma_addr_t dma_addr_in; 118 struct atmel_tdes_dma dma_lch_in; 119 120 void *buf_out; 121 int dma_out; 122 dma_addr_t dma_addr_out; 123 struct atmel_tdes_dma dma_lch_out; 124 125 struct atmel_tdes_caps caps; 126 127 u32 hw_version; 128 }; 129 130 struct atmel_tdes_drv { 131 struct list_head dev_list; 132 spinlock_t lock; 133 }; 134 135 static struct atmel_tdes_drv atmel_tdes = { 136 .dev_list = LIST_HEAD_INIT(atmel_tdes.dev_list), 137 .lock = __SPIN_LOCK_UNLOCKED(atmel_tdes.lock), 138 }; 139 140 static int atmel_tdes_sg_copy(struct scatterlist **sg, size_t *offset, 141 void *buf, size_t buflen, size_t total, int out) 142 { 143 size_t count, off = 0; 144 145 while (buflen && total) { 146 count = min((*sg)->length - *offset, total); 147 count = min(count, buflen); 148 149 if (!count) 150 return off; 151 152 scatterwalk_map_and_copy(buf + off, *sg, *offset, count, out); 153 154 off += count; 155 buflen -= count; 156 *offset += count; 157 total -= count; 158 159 if (*offset == (*sg)->length) { 160 *sg = sg_next(*sg); 161 if (*sg) 162 *offset = 0; 163 else 164 total = 0; 165 } 166 } 167 168 return off; 169 } 170 171 static inline u32 atmel_tdes_read(struct atmel_tdes_dev *dd, u32 offset) 172 { 173 return readl_relaxed(dd->io_base + offset); 174 } 175 176 static inline void atmel_tdes_write(struct atmel_tdes_dev *dd, 177 u32 offset, u32 value) 178 { 179 writel_relaxed(value, dd->io_base + offset); 180 } 181 182 static void atmel_tdes_write_n(struct atmel_tdes_dev *dd, u32 offset, 183 const u32 *value, int count) 184 { 185 for (; count--; value++, offset += 4) 186 atmel_tdes_write(dd, offset, *value); 187 } 188 189 static struct atmel_tdes_dev *atmel_tdes_dev_alloc(void) 190 { 191 struct atmel_tdes_dev *tdes_dd; 192 193 spin_lock_bh(&atmel_tdes.lock); 194 /* One TDES IP per SoC. */ 195 tdes_dd = list_first_entry_or_null(&atmel_tdes.dev_list, 196 struct atmel_tdes_dev, list); 197 spin_unlock_bh(&atmel_tdes.lock); 198 return tdes_dd; 199 } 200 201 static int atmel_tdes_hw_init(struct atmel_tdes_dev *dd) 202 { 203 int err; 204 205 err = clk_prepare_enable(dd->iclk); 206 if (err) 207 return err; 208 209 if (!(dd->flags & TDES_FLAGS_INIT)) { 210 atmel_tdes_write(dd, TDES_CR, TDES_CR_SWRST); 211 dd->flags |= TDES_FLAGS_INIT; 212 } 213 214 return 0; 215 } 216 217 static inline unsigned int atmel_tdes_get_version(struct atmel_tdes_dev *dd) 218 { 219 return atmel_tdes_read(dd, TDES_HW_VERSION) & 0x00000fff; 220 } 221 222 static int atmel_tdes_hw_version_init(struct atmel_tdes_dev *dd) 223 { 224 int err; 225 226 err = atmel_tdes_hw_init(dd); 227 if (err) 228 return err; 229 230 dd->hw_version = atmel_tdes_get_version(dd); 231 232 dev_info(dd->dev, 233 "version: 0x%x\n", dd->hw_version); 234 235 clk_disable_unprepare(dd->iclk); 236 237 return 0; 238 } 239 240 static void atmel_tdes_dma_callback(void *data) 241 { 242 struct atmel_tdes_dev *dd = data; 243 244 /* dma_lch_out - completed */ 245 tasklet_schedule(&dd->done_task); 246 } 247 248 static int atmel_tdes_write_ctrl(struct atmel_tdes_dev *dd) 249 { 250 int err; 251 u32 valmr = TDES_MR_SMOD_PDC; 252 253 err = atmel_tdes_hw_init(dd); 254 255 if (err) 256 return err; 257 258 if (!dd->caps.has_dma) 259 atmel_tdes_write(dd, TDES_PTCR, 260 TDES_PTCR_TXTDIS | TDES_PTCR_RXTDIS); 261 262 /* MR register must be set before IV registers */ 263 if (dd->ctx->keylen > (DES_KEY_SIZE << 1)) { 264 valmr |= TDES_MR_KEYMOD_3KEY; 265 valmr |= TDES_MR_TDESMOD_TDES; 266 } else if (dd->ctx->keylen > DES_KEY_SIZE) { 267 valmr |= TDES_MR_KEYMOD_2KEY; 268 valmr |= TDES_MR_TDESMOD_TDES; 269 } else { 270 valmr |= TDES_MR_TDESMOD_DES; 271 } 272 273 valmr |= dd->flags & TDES_FLAGS_MODE_MASK; 274 275 atmel_tdes_write(dd, TDES_MR, valmr); 276 277 atmel_tdes_write_n(dd, TDES_KEY1W1R, dd->ctx->key, 278 dd->ctx->keylen >> 2); 279 280 if (dd->req->iv && (valmr & TDES_MR_OPMOD_MASK) != TDES_MR_OPMOD_ECB) 281 atmel_tdes_write_n(dd, TDES_IV1R, (void *)dd->req->iv, 2); 282 283 return 0; 284 } 285 286 static int atmel_tdes_crypt_pdc_stop(struct atmel_tdes_dev *dd) 287 { 288 int err = 0; 289 size_t count; 290 291 atmel_tdes_write(dd, TDES_PTCR, TDES_PTCR_TXTDIS|TDES_PTCR_RXTDIS); 292 293 if (dd->flags & TDES_FLAGS_FAST) { 294 dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_FROM_DEVICE); 295 dma_unmap_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE); 296 } else { 297 dma_sync_single_for_device(dd->dev, dd->dma_addr_out, 298 dd->dma_size, DMA_FROM_DEVICE); 299 300 /* copy data */ 301 count = atmel_tdes_sg_copy(&dd->out_sg, &dd->out_offset, 302 dd->buf_out, dd->buflen, dd->dma_size, 1); 303 if (count != dd->dma_size) { 304 err = -EINVAL; 305 dev_dbg(dd->dev, "not all data converted: %zu\n", count); 306 } 307 } 308 309 return err; 310 } 311 312 static int atmel_tdes_buff_init(struct atmel_tdes_dev *dd) 313 { 314 int err = -ENOMEM; 315 316 dd->buf_in = (void *)__get_free_pages(GFP_KERNEL, 0); 317 dd->buf_out = (void *)__get_free_pages(GFP_KERNEL, 0); 318 dd->buflen = PAGE_SIZE; 319 dd->buflen &= ~(DES_BLOCK_SIZE - 1); 320 321 if (!dd->buf_in || !dd->buf_out) { 322 dev_dbg(dd->dev, "unable to alloc pages.\n"); 323 goto err_alloc; 324 } 325 326 /* MAP here */ 327 dd->dma_addr_in = dma_map_single(dd->dev, dd->buf_in, 328 dd->buflen, DMA_TO_DEVICE); 329 err = dma_mapping_error(dd->dev, dd->dma_addr_in); 330 if (err) { 331 dev_dbg(dd->dev, "dma %zd bytes error\n", dd->buflen); 332 goto err_map_in; 333 } 334 335 dd->dma_addr_out = dma_map_single(dd->dev, dd->buf_out, 336 dd->buflen, DMA_FROM_DEVICE); 337 err = dma_mapping_error(dd->dev, dd->dma_addr_out); 338 if (err) { 339 dev_dbg(dd->dev, "dma %zd bytes error\n", dd->buflen); 340 goto err_map_out; 341 } 342 343 return 0; 344 345 err_map_out: 346 dma_unmap_single(dd->dev, dd->dma_addr_in, dd->buflen, 347 DMA_TO_DEVICE); 348 err_map_in: 349 err_alloc: 350 free_page((unsigned long)dd->buf_out); 351 free_page((unsigned long)dd->buf_in); 352 return err; 353 } 354 355 static void atmel_tdes_buff_cleanup(struct atmel_tdes_dev *dd) 356 { 357 dma_unmap_single(dd->dev, dd->dma_addr_out, dd->buflen, 358 DMA_FROM_DEVICE); 359 dma_unmap_single(dd->dev, dd->dma_addr_in, dd->buflen, 360 DMA_TO_DEVICE); 361 free_page((unsigned long)dd->buf_out); 362 free_page((unsigned long)dd->buf_in); 363 } 364 365 static int atmel_tdes_crypt_pdc(struct atmel_tdes_dev *dd, 366 dma_addr_t dma_addr_in, 367 dma_addr_t dma_addr_out, int length) 368 { 369 int len32; 370 371 dd->dma_size = length; 372 373 if (!(dd->flags & TDES_FLAGS_FAST)) { 374 dma_sync_single_for_device(dd->dev, dma_addr_in, length, 375 DMA_TO_DEVICE); 376 } 377 378 len32 = DIV_ROUND_UP(length, sizeof(u32)); 379 380 atmel_tdes_write(dd, TDES_PTCR, TDES_PTCR_TXTDIS|TDES_PTCR_RXTDIS); 381 atmel_tdes_write(dd, TDES_TPR, dma_addr_in); 382 atmel_tdes_write(dd, TDES_TCR, len32); 383 atmel_tdes_write(dd, TDES_RPR, dma_addr_out); 384 atmel_tdes_write(dd, TDES_RCR, len32); 385 386 /* Enable Interrupt */ 387 atmel_tdes_write(dd, TDES_IER, TDES_INT_ENDRX); 388 389 /* Start DMA transfer */ 390 atmel_tdes_write(dd, TDES_PTCR, TDES_PTCR_TXTEN | TDES_PTCR_RXTEN); 391 392 return 0; 393 } 394 395 static int atmel_tdes_crypt_dma(struct atmel_tdes_dev *dd, 396 dma_addr_t dma_addr_in, 397 dma_addr_t dma_addr_out, int length) 398 { 399 struct scatterlist sg[2]; 400 struct dma_async_tx_descriptor *in_desc, *out_desc; 401 enum dma_slave_buswidth addr_width; 402 403 dd->dma_size = length; 404 405 if (!(dd->flags & TDES_FLAGS_FAST)) { 406 dma_sync_single_for_device(dd->dev, dma_addr_in, length, 407 DMA_TO_DEVICE); 408 } 409 410 addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES; 411 412 dd->dma_lch_in.dma_conf.dst_addr_width = addr_width; 413 dd->dma_lch_out.dma_conf.src_addr_width = addr_width; 414 415 dmaengine_slave_config(dd->dma_lch_in.chan, &dd->dma_lch_in.dma_conf); 416 dmaengine_slave_config(dd->dma_lch_out.chan, &dd->dma_lch_out.dma_conf); 417 418 dd->flags |= TDES_FLAGS_DMA; 419 420 sg_init_table(&sg[0], 1); 421 sg_dma_address(&sg[0]) = dma_addr_in; 422 sg_dma_len(&sg[0]) = length; 423 424 sg_init_table(&sg[1], 1); 425 sg_dma_address(&sg[1]) = dma_addr_out; 426 sg_dma_len(&sg[1]) = length; 427 428 in_desc = dmaengine_prep_slave_sg(dd->dma_lch_in.chan, &sg[0], 429 1, DMA_MEM_TO_DEV, 430 DMA_PREP_INTERRUPT | DMA_CTRL_ACK); 431 if (!in_desc) 432 return -EINVAL; 433 434 out_desc = dmaengine_prep_slave_sg(dd->dma_lch_out.chan, &sg[1], 435 1, DMA_DEV_TO_MEM, 436 DMA_PREP_INTERRUPT | DMA_CTRL_ACK); 437 if (!out_desc) 438 return -EINVAL; 439 440 out_desc->callback = atmel_tdes_dma_callback; 441 out_desc->callback_param = dd; 442 443 dmaengine_submit(out_desc); 444 dma_async_issue_pending(dd->dma_lch_out.chan); 445 446 dmaengine_submit(in_desc); 447 dma_async_issue_pending(dd->dma_lch_in.chan); 448 449 return 0; 450 } 451 452 static int atmel_tdes_crypt_start(struct atmel_tdes_dev *dd) 453 { 454 int err, fast = 0, in, out; 455 size_t count; 456 dma_addr_t addr_in, addr_out; 457 458 if ((!dd->in_offset) && (!dd->out_offset)) { 459 /* check for alignment */ 460 in = IS_ALIGNED((u32)dd->in_sg->offset, sizeof(u32)) && 461 IS_ALIGNED(dd->in_sg->length, dd->ctx->block_size); 462 out = IS_ALIGNED((u32)dd->out_sg->offset, sizeof(u32)) && 463 IS_ALIGNED(dd->out_sg->length, dd->ctx->block_size); 464 fast = in && out; 465 466 if (sg_dma_len(dd->in_sg) != sg_dma_len(dd->out_sg)) 467 fast = 0; 468 } 469 470 471 if (fast) { 472 count = min_t(size_t, dd->total, sg_dma_len(dd->in_sg)); 473 count = min_t(size_t, count, sg_dma_len(dd->out_sg)); 474 475 err = dma_map_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE); 476 if (!err) { 477 dev_dbg(dd->dev, "dma_map_sg() error\n"); 478 return -EINVAL; 479 } 480 481 err = dma_map_sg(dd->dev, dd->out_sg, 1, 482 DMA_FROM_DEVICE); 483 if (!err) { 484 dev_dbg(dd->dev, "dma_map_sg() error\n"); 485 dma_unmap_sg(dd->dev, dd->in_sg, 1, 486 DMA_TO_DEVICE); 487 return -EINVAL; 488 } 489 490 addr_in = sg_dma_address(dd->in_sg); 491 addr_out = sg_dma_address(dd->out_sg); 492 493 dd->flags |= TDES_FLAGS_FAST; 494 495 } else { 496 /* use cache buffers */ 497 count = atmel_tdes_sg_copy(&dd->in_sg, &dd->in_offset, 498 dd->buf_in, dd->buflen, dd->total, 0); 499 500 addr_in = dd->dma_addr_in; 501 addr_out = dd->dma_addr_out; 502 503 dd->flags &= ~TDES_FLAGS_FAST; 504 } 505 506 dd->total -= count; 507 508 if (dd->caps.has_dma) 509 err = atmel_tdes_crypt_dma(dd, addr_in, addr_out, count); 510 else 511 err = atmel_tdes_crypt_pdc(dd, addr_in, addr_out, count); 512 513 if (err && (dd->flags & TDES_FLAGS_FAST)) { 514 dma_unmap_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE); 515 dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_TO_DEVICE); 516 } 517 518 return err; 519 } 520 521 static void 522 atmel_tdes_set_iv_as_last_ciphertext_block(struct atmel_tdes_dev *dd) 523 { 524 struct skcipher_request *req = dd->req; 525 struct atmel_tdes_reqctx *rctx = skcipher_request_ctx(req); 526 struct crypto_skcipher *skcipher = crypto_skcipher_reqtfm(req); 527 unsigned int ivsize = crypto_skcipher_ivsize(skcipher); 528 529 if (req->cryptlen < ivsize) 530 return; 531 532 if (rctx->mode & TDES_FLAGS_ENCRYPT) 533 scatterwalk_map_and_copy(req->iv, req->dst, 534 req->cryptlen - ivsize, ivsize, 0); 535 else 536 memcpy(req->iv, rctx->lastc, ivsize); 537 538 } 539 540 static void atmel_tdes_finish_req(struct atmel_tdes_dev *dd, int err) 541 { 542 struct skcipher_request *req = dd->req; 543 struct atmel_tdes_reqctx *rctx = skcipher_request_ctx(req); 544 545 clk_disable_unprepare(dd->iclk); 546 547 dd->flags &= ~TDES_FLAGS_BUSY; 548 549 if (!err && (rctx->mode & TDES_FLAGS_OPMODE_MASK) != TDES_FLAGS_ECB) 550 atmel_tdes_set_iv_as_last_ciphertext_block(dd); 551 552 skcipher_request_complete(req, err); 553 } 554 555 static int atmel_tdes_handle_queue(struct atmel_tdes_dev *dd, 556 struct skcipher_request *req) 557 { 558 struct crypto_async_request *async_req, *backlog; 559 struct atmel_tdes_ctx *ctx; 560 struct atmel_tdes_reqctx *rctx; 561 unsigned long flags; 562 int err, ret = 0; 563 564 spin_lock_irqsave(&dd->lock, flags); 565 if (req) 566 ret = crypto_enqueue_request(&dd->queue, &req->base); 567 if (dd->flags & TDES_FLAGS_BUSY) { 568 spin_unlock_irqrestore(&dd->lock, flags); 569 return ret; 570 } 571 backlog = crypto_get_backlog(&dd->queue); 572 async_req = crypto_dequeue_request(&dd->queue); 573 if (async_req) 574 dd->flags |= TDES_FLAGS_BUSY; 575 spin_unlock_irqrestore(&dd->lock, flags); 576 577 if (!async_req) 578 return ret; 579 580 if (backlog) 581 crypto_request_complete(backlog, -EINPROGRESS); 582 583 req = skcipher_request_cast(async_req); 584 585 /* assign new request to device */ 586 dd->req = req; 587 dd->total = req->cryptlen; 588 dd->in_offset = 0; 589 dd->in_sg = req->src; 590 dd->out_offset = 0; 591 dd->out_sg = req->dst; 592 593 rctx = skcipher_request_ctx(req); 594 ctx = crypto_skcipher_ctx(crypto_skcipher_reqtfm(req)); 595 rctx->mode &= TDES_FLAGS_MODE_MASK; 596 dd->flags = (dd->flags & ~TDES_FLAGS_MODE_MASK) | rctx->mode; 597 dd->ctx = ctx; 598 599 err = atmel_tdes_write_ctrl(dd); 600 if (!err) 601 err = atmel_tdes_crypt_start(dd); 602 if (err) { 603 /* des_task will not finish it, so do it here */ 604 atmel_tdes_finish_req(dd, err); 605 tasklet_schedule(&dd->queue_task); 606 } 607 608 return ret; 609 } 610 611 static int atmel_tdes_crypt_dma_stop(struct atmel_tdes_dev *dd) 612 { 613 int err = -EINVAL; 614 size_t count; 615 616 if (dd->flags & TDES_FLAGS_DMA) { 617 err = 0; 618 if (dd->flags & TDES_FLAGS_FAST) { 619 dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_FROM_DEVICE); 620 dma_unmap_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE); 621 } else { 622 dma_sync_single_for_device(dd->dev, dd->dma_addr_out, 623 dd->dma_size, DMA_FROM_DEVICE); 624 625 /* copy data */ 626 count = atmel_tdes_sg_copy(&dd->out_sg, &dd->out_offset, 627 dd->buf_out, dd->buflen, dd->dma_size, 1); 628 if (count != dd->dma_size) { 629 err = -EINVAL; 630 dev_dbg(dd->dev, "not all data converted: %zu\n", count); 631 } 632 } 633 } 634 return err; 635 } 636 637 static int atmel_tdes_crypt(struct skcipher_request *req, unsigned long mode) 638 { 639 struct crypto_skcipher *skcipher = crypto_skcipher_reqtfm(req); 640 struct atmel_tdes_ctx *ctx = crypto_skcipher_ctx(skcipher); 641 struct atmel_tdes_reqctx *rctx = skcipher_request_ctx(req); 642 struct device *dev = ctx->dd->dev; 643 644 if (!req->cryptlen) 645 return 0; 646 647 if (!IS_ALIGNED(req->cryptlen, DES_BLOCK_SIZE)) { 648 dev_dbg(dev, "request size is not exact amount of DES blocks\n"); 649 return -EINVAL; 650 } 651 ctx->block_size = DES_BLOCK_SIZE; 652 653 rctx->mode = mode; 654 655 if ((mode & TDES_FLAGS_OPMODE_MASK) != TDES_FLAGS_ECB && 656 !(mode & TDES_FLAGS_ENCRYPT)) { 657 unsigned int ivsize = crypto_skcipher_ivsize(skcipher); 658 659 if (req->cryptlen >= ivsize) 660 scatterwalk_map_and_copy(rctx->lastc, req->src, 661 req->cryptlen - ivsize, 662 ivsize, 0); 663 } 664 665 return atmel_tdes_handle_queue(ctx->dd, req); 666 } 667 668 static int atmel_tdes_dma_init(struct atmel_tdes_dev *dd) 669 { 670 int ret; 671 672 /* Try to grab 2 DMA channels */ 673 dd->dma_lch_in.chan = dma_request_chan(dd->dev, "tx"); 674 if (IS_ERR(dd->dma_lch_in.chan)) { 675 ret = PTR_ERR(dd->dma_lch_in.chan); 676 goto err_dma_in; 677 } 678 679 dd->dma_lch_in.dma_conf.dst_addr = dd->phys_base + 680 TDES_IDATA1R; 681 dd->dma_lch_in.dma_conf.src_maxburst = 1; 682 dd->dma_lch_in.dma_conf.src_addr_width = 683 DMA_SLAVE_BUSWIDTH_4_BYTES; 684 dd->dma_lch_in.dma_conf.dst_maxburst = 1; 685 dd->dma_lch_in.dma_conf.dst_addr_width = 686 DMA_SLAVE_BUSWIDTH_4_BYTES; 687 dd->dma_lch_in.dma_conf.device_fc = false; 688 689 dd->dma_lch_out.chan = dma_request_chan(dd->dev, "rx"); 690 if (IS_ERR(dd->dma_lch_out.chan)) { 691 ret = PTR_ERR(dd->dma_lch_out.chan); 692 goto err_dma_out; 693 } 694 695 dd->dma_lch_out.dma_conf.src_addr = dd->phys_base + 696 TDES_ODATA1R; 697 dd->dma_lch_out.dma_conf.src_maxburst = 1; 698 dd->dma_lch_out.dma_conf.src_addr_width = 699 DMA_SLAVE_BUSWIDTH_4_BYTES; 700 dd->dma_lch_out.dma_conf.dst_maxburst = 1; 701 dd->dma_lch_out.dma_conf.dst_addr_width = 702 DMA_SLAVE_BUSWIDTH_4_BYTES; 703 dd->dma_lch_out.dma_conf.device_fc = false; 704 705 return 0; 706 707 err_dma_out: 708 dma_release_channel(dd->dma_lch_in.chan); 709 err_dma_in: 710 dev_err(dd->dev, "no DMA channel available\n"); 711 return ret; 712 } 713 714 static void atmel_tdes_dma_cleanup(struct atmel_tdes_dev *dd) 715 { 716 dma_release_channel(dd->dma_lch_in.chan); 717 dma_release_channel(dd->dma_lch_out.chan); 718 } 719 720 static int atmel_des_setkey(struct crypto_skcipher *tfm, const u8 *key, 721 unsigned int keylen) 722 { 723 struct atmel_tdes_ctx *ctx = crypto_skcipher_ctx(tfm); 724 int err; 725 726 err = verify_skcipher_des_key(tfm, key); 727 if (err) 728 return err; 729 730 memcpy(ctx->key, key, keylen); 731 ctx->keylen = keylen; 732 733 return 0; 734 } 735 736 static int atmel_tdes_setkey(struct crypto_skcipher *tfm, const u8 *key, 737 unsigned int keylen) 738 { 739 struct atmel_tdes_ctx *ctx = crypto_skcipher_ctx(tfm); 740 int err; 741 742 err = verify_skcipher_des3_key(tfm, key); 743 if (err) 744 return err; 745 746 memcpy(ctx->key, key, keylen); 747 ctx->keylen = keylen; 748 749 return 0; 750 } 751 752 static int atmel_tdes_ecb_encrypt(struct skcipher_request *req) 753 { 754 return atmel_tdes_crypt(req, TDES_FLAGS_ECB | TDES_FLAGS_ENCRYPT); 755 } 756 757 static int atmel_tdes_ecb_decrypt(struct skcipher_request *req) 758 { 759 return atmel_tdes_crypt(req, TDES_FLAGS_ECB); 760 } 761 762 static int atmel_tdes_cbc_encrypt(struct skcipher_request *req) 763 { 764 return atmel_tdes_crypt(req, TDES_FLAGS_CBC | TDES_FLAGS_ENCRYPT); 765 } 766 767 static int atmel_tdes_cbc_decrypt(struct skcipher_request *req) 768 { 769 return atmel_tdes_crypt(req, TDES_FLAGS_CBC); 770 } 771 772 static int atmel_tdes_init_tfm(struct crypto_skcipher *tfm) 773 { 774 struct atmel_tdes_ctx *ctx = crypto_skcipher_ctx(tfm); 775 776 ctx->dd = atmel_tdes_dev_alloc(); 777 if (!ctx->dd) 778 return -ENODEV; 779 780 crypto_skcipher_set_reqsize(tfm, sizeof(struct atmel_tdes_reqctx)); 781 782 return 0; 783 } 784 785 static void atmel_tdes_skcipher_alg_init(struct skcipher_alg *alg) 786 { 787 alg->base.cra_priority = ATMEL_TDES_PRIORITY; 788 alg->base.cra_flags = CRYPTO_ALG_ASYNC; 789 alg->base.cra_ctxsize = sizeof(struct atmel_tdes_ctx); 790 alg->base.cra_module = THIS_MODULE; 791 792 alg->init = atmel_tdes_init_tfm; 793 } 794 795 static struct skcipher_alg tdes_algs[] = { 796 { 797 .base.cra_name = "ecb(des)", 798 .base.cra_driver_name = "atmel-ecb-des", 799 .base.cra_blocksize = DES_BLOCK_SIZE, 800 .base.cra_alignmask = 0x7, 801 802 .min_keysize = DES_KEY_SIZE, 803 .max_keysize = DES_KEY_SIZE, 804 .setkey = atmel_des_setkey, 805 .encrypt = atmel_tdes_ecb_encrypt, 806 .decrypt = atmel_tdes_ecb_decrypt, 807 }, 808 { 809 .base.cra_name = "cbc(des)", 810 .base.cra_driver_name = "atmel-cbc-des", 811 .base.cra_blocksize = DES_BLOCK_SIZE, 812 .base.cra_alignmask = 0x7, 813 814 .min_keysize = DES_KEY_SIZE, 815 .max_keysize = DES_KEY_SIZE, 816 .ivsize = DES_BLOCK_SIZE, 817 .setkey = atmel_des_setkey, 818 .encrypt = atmel_tdes_cbc_encrypt, 819 .decrypt = atmel_tdes_cbc_decrypt, 820 }, 821 { 822 .base.cra_name = "ecb(des3_ede)", 823 .base.cra_driver_name = "atmel-ecb-tdes", 824 .base.cra_blocksize = DES_BLOCK_SIZE, 825 .base.cra_alignmask = 0x7, 826 827 .min_keysize = DES3_EDE_KEY_SIZE, 828 .max_keysize = DES3_EDE_KEY_SIZE, 829 .setkey = atmel_tdes_setkey, 830 .encrypt = atmel_tdes_ecb_encrypt, 831 .decrypt = atmel_tdes_ecb_decrypt, 832 }, 833 { 834 .base.cra_name = "cbc(des3_ede)", 835 .base.cra_driver_name = "atmel-cbc-tdes", 836 .base.cra_blocksize = DES_BLOCK_SIZE, 837 .base.cra_alignmask = 0x7, 838 839 .min_keysize = DES3_EDE_KEY_SIZE, 840 .max_keysize = DES3_EDE_KEY_SIZE, 841 .setkey = atmel_tdes_setkey, 842 .encrypt = atmel_tdes_cbc_encrypt, 843 .decrypt = atmel_tdes_cbc_decrypt, 844 .ivsize = DES_BLOCK_SIZE, 845 }, 846 }; 847 848 static void atmel_tdes_queue_task(unsigned long data) 849 { 850 struct atmel_tdes_dev *dd = (struct atmel_tdes_dev *)data; 851 852 atmel_tdes_handle_queue(dd, NULL); 853 } 854 855 static void atmel_tdes_done_task(unsigned long data) 856 { 857 struct atmel_tdes_dev *dd = (struct atmel_tdes_dev *) data; 858 int err; 859 860 if (!(dd->flags & TDES_FLAGS_DMA)) 861 err = atmel_tdes_crypt_pdc_stop(dd); 862 else 863 err = atmel_tdes_crypt_dma_stop(dd); 864 865 if (dd->total && !err) { 866 if (dd->flags & TDES_FLAGS_FAST) { 867 dd->in_sg = sg_next(dd->in_sg); 868 dd->out_sg = sg_next(dd->out_sg); 869 if (!dd->in_sg || !dd->out_sg) 870 err = -EINVAL; 871 } 872 if (!err) 873 err = atmel_tdes_crypt_start(dd); 874 if (!err) 875 return; /* DMA started. Not fininishing. */ 876 } 877 878 atmel_tdes_finish_req(dd, err); 879 atmel_tdes_handle_queue(dd, NULL); 880 } 881 882 static irqreturn_t atmel_tdes_irq(int irq, void *dev_id) 883 { 884 struct atmel_tdes_dev *tdes_dd = dev_id; 885 u32 reg; 886 887 reg = atmel_tdes_read(tdes_dd, TDES_ISR); 888 if (reg & atmel_tdes_read(tdes_dd, TDES_IMR)) { 889 atmel_tdes_write(tdes_dd, TDES_IDR, reg); 890 if (TDES_FLAGS_BUSY & tdes_dd->flags) 891 tasklet_schedule(&tdes_dd->done_task); 892 else 893 dev_warn(tdes_dd->dev, "TDES interrupt when no active requests.\n"); 894 return IRQ_HANDLED; 895 } 896 897 return IRQ_NONE; 898 } 899 900 static void atmel_tdes_unregister_algs(struct atmel_tdes_dev *dd) 901 { 902 int i; 903 904 for (i = 0; i < ARRAY_SIZE(tdes_algs); i++) 905 crypto_unregister_skcipher(&tdes_algs[i]); 906 } 907 908 static int atmel_tdes_register_algs(struct atmel_tdes_dev *dd) 909 { 910 int err, i, j; 911 912 for (i = 0; i < ARRAY_SIZE(tdes_algs); i++) { 913 atmel_tdes_skcipher_alg_init(&tdes_algs[i]); 914 915 err = crypto_register_skcipher(&tdes_algs[i]); 916 if (err) 917 goto err_tdes_algs; 918 } 919 920 return 0; 921 922 err_tdes_algs: 923 for (j = 0; j < i; j++) 924 crypto_unregister_skcipher(&tdes_algs[j]); 925 926 return err; 927 } 928 929 static void atmel_tdes_get_cap(struct atmel_tdes_dev *dd) 930 { 931 932 dd->caps.has_dma = 0; 933 934 /* keep only major version number */ 935 switch (dd->hw_version & 0xf00) { 936 case 0x800: 937 case 0x700: 938 dd->caps.has_dma = 1; 939 break; 940 case 0x600: 941 break; 942 default: 943 dev_warn(dd->dev, 944 "Unmanaged tdes version, set minimum capabilities\n"); 945 break; 946 } 947 } 948 949 static const struct of_device_id atmel_tdes_dt_ids[] = { 950 { .compatible = "atmel,at91sam9g46-tdes" }, 951 { /* sentinel */ } 952 }; 953 MODULE_DEVICE_TABLE(of, atmel_tdes_dt_ids); 954 955 static int atmel_tdes_probe(struct platform_device *pdev) 956 { 957 struct atmel_tdes_dev *tdes_dd; 958 struct device *dev = &pdev->dev; 959 struct resource *tdes_res; 960 int err; 961 962 tdes_dd = devm_kmalloc(&pdev->dev, sizeof(*tdes_dd), GFP_KERNEL); 963 if (!tdes_dd) 964 return -ENOMEM; 965 966 tdes_dd->dev = dev; 967 968 platform_set_drvdata(pdev, tdes_dd); 969 970 INIT_LIST_HEAD(&tdes_dd->list); 971 spin_lock_init(&tdes_dd->lock); 972 973 tasklet_init(&tdes_dd->done_task, atmel_tdes_done_task, 974 (unsigned long)tdes_dd); 975 tasklet_init(&tdes_dd->queue_task, atmel_tdes_queue_task, 976 (unsigned long)tdes_dd); 977 978 crypto_init_queue(&tdes_dd->queue, ATMEL_TDES_QUEUE_LENGTH); 979 980 tdes_dd->io_base = devm_platform_get_and_ioremap_resource(pdev, 0, &tdes_res); 981 if (IS_ERR(tdes_dd->io_base)) { 982 err = PTR_ERR(tdes_dd->io_base); 983 goto err_tasklet_kill; 984 } 985 tdes_dd->phys_base = tdes_res->start; 986 987 /* Get the IRQ */ 988 tdes_dd->irq = platform_get_irq(pdev, 0); 989 if (tdes_dd->irq < 0) { 990 err = tdes_dd->irq; 991 goto err_tasklet_kill; 992 } 993 994 err = devm_request_irq(&pdev->dev, tdes_dd->irq, atmel_tdes_irq, 995 IRQF_SHARED, "atmel-tdes", tdes_dd); 996 if (err) { 997 dev_err(dev, "unable to request tdes irq.\n"); 998 goto err_tasklet_kill; 999 } 1000 1001 /* Initializing the clock */ 1002 tdes_dd->iclk = devm_clk_get(&pdev->dev, "tdes_clk"); 1003 if (IS_ERR(tdes_dd->iclk)) { 1004 dev_err(dev, "clock initialization failed.\n"); 1005 err = PTR_ERR(tdes_dd->iclk); 1006 goto err_tasklet_kill; 1007 } 1008 1009 err = atmel_tdes_hw_version_init(tdes_dd); 1010 if (err) 1011 goto err_tasklet_kill; 1012 1013 atmel_tdes_get_cap(tdes_dd); 1014 1015 err = atmel_tdes_buff_init(tdes_dd); 1016 if (err) 1017 goto err_tasklet_kill; 1018 1019 if (tdes_dd->caps.has_dma) { 1020 err = atmel_tdes_dma_init(tdes_dd); 1021 if (err) 1022 goto err_buff_cleanup; 1023 1024 dev_info(dev, "using %s, %s for DMA transfers\n", 1025 dma_chan_name(tdes_dd->dma_lch_in.chan), 1026 dma_chan_name(tdes_dd->dma_lch_out.chan)); 1027 } 1028 1029 spin_lock(&atmel_tdes.lock); 1030 list_add_tail(&tdes_dd->list, &atmel_tdes.dev_list); 1031 spin_unlock(&atmel_tdes.lock); 1032 1033 err = atmel_tdes_register_algs(tdes_dd); 1034 if (err) 1035 goto err_algs; 1036 1037 dev_info(dev, "Atmel DES/TDES\n"); 1038 1039 return 0; 1040 1041 err_algs: 1042 spin_lock(&atmel_tdes.lock); 1043 list_del(&tdes_dd->list); 1044 spin_unlock(&atmel_tdes.lock); 1045 if (tdes_dd->caps.has_dma) 1046 atmel_tdes_dma_cleanup(tdes_dd); 1047 err_buff_cleanup: 1048 atmel_tdes_buff_cleanup(tdes_dd); 1049 err_tasklet_kill: 1050 tasklet_kill(&tdes_dd->done_task); 1051 tasklet_kill(&tdes_dd->queue_task); 1052 1053 return err; 1054 } 1055 1056 static void atmel_tdes_remove(struct platform_device *pdev) 1057 { 1058 struct atmel_tdes_dev *tdes_dd = platform_get_drvdata(pdev); 1059 1060 spin_lock(&atmel_tdes.lock); 1061 list_del(&tdes_dd->list); 1062 spin_unlock(&atmel_tdes.lock); 1063 1064 atmel_tdes_unregister_algs(tdes_dd); 1065 1066 tasklet_kill(&tdes_dd->done_task); 1067 tasklet_kill(&tdes_dd->queue_task); 1068 1069 if (tdes_dd->caps.has_dma) 1070 atmel_tdes_dma_cleanup(tdes_dd); 1071 1072 atmel_tdes_buff_cleanup(tdes_dd); 1073 } 1074 1075 static struct platform_driver atmel_tdes_driver = { 1076 .probe = atmel_tdes_probe, 1077 .remove_new = atmel_tdes_remove, 1078 .driver = { 1079 .name = "atmel_tdes", 1080 .of_match_table = atmel_tdes_dt_ids, 1081 }, 1082 }; 1083 1084 module_platform_driver(atmel_tdes_driver); 1085 1086 MODULE_DESCRIPTION("Atmel DES/TDES hw acceleration support."); 1087 MODULE_LICENSE("GPL v2"); 1088 MODULE_AUTHOR("Nicolas Royer - Eukréa Electromatique"); 1089