1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * QMC driver 4 * 5 * Copyright 2022 CS GROUP France 6 * 7 * Author: Herve Codina <herve.codina@bootlin.com> 8 */ 9 10 #include <soc/fsl/qe/qmc.h> 11 #include <linux/dma-mapping.h> 12 #include <linux/hdlc.h> 13 #include <linux/interrupt.h> 14 #include <linux/io.h> 15 #include <linux/module.h> 16 #include <linux/of.h> 17 #include <linux/of_platform.h> 18 #include <linux/platform_device.h> 19 #include <linux/slab.h> 20 #include <soc/fsl/cpm.h> 21 #include <sysdev/fsl_soc.h> 22 #include "tsa.h" 23 24 /* SCC general mode register high (32 bits) */ 25 #define SCC_GSMRL 0x00 26 #define SCC_GSMRL_ENR (1 << 5) 27 #define SCC_GSMRL_ENT (1 << 4) 28 #define SCC_GSMRL_MODE_QMC (0x0A << 0) 29 30 /* SCC general mode register low (32 bits) */ 31 #define SCC_GSMRH 0x04 32 #define SCC_GSMRH_CTSS (1 << 7) 33 #define SCC_GSMRH_CDS (1 << 8) 34 #define SCC_GSMRH_CTSP (1 << 9) 35 #define SCC_GSMRH_CDP (1 << 10) 36 37 /* SCC event register (16 bits) */ 38 #define SCC_SCCE 0x10 39 #define SCC_SCCE_IQOV (1 << 3) 40 #define SCC_SCCE_GINT (1 << 2) 41 #define SCC_SCCE_GUN (1 << 1) 42 #define SCC_SCCE_GOV (1 << 0) 43 44 /* SCC mask register (16 bits) */ 45 #define SCC_SCCM 0x14 46 /* Multichannel base pointer (32 bits) */ 47 #define QMC_GBL_MCBASE 0x00 48 /* Multichannel controller state (16 bits) */ 49 #define QMC_GBL_QMCSTATE 0x04 50 /* Maximum receive buffer length (16 bits) */ 51 #define QMC_GBL_MRBLR 0x06 52 /* Tx time-slot assignment table pointer (16 bits) */ 53 #define QMC_GBL_TX_S_PTR 0x08 54 /* Rx pointer (16 bits) */ 55 #define QMC_GBL_RXPTR 0x0A 56 /* Global receive frame threshold (16 bits) */ 57 #define QMC_GBL_GRFTHR 0x0C 58 /* Global receive frame count (16 bits) */ 59 #define QMC_GBL_GRFCNT 0x0E 60 /* Multichannel interrupt base address (32 bits) */ 61 #define QMC_GBL_INTBASE 0x10 62 /* Multichannel interrupt pointer (32 bits) */ 63 #define QMC_GBL_INTPTR 0x14 64 /* Rx time-slot assignment table pointer (16 bits) */ 65 #define QMC_GBL_RX_S_PTR 0x18 66 /* Tx pointer (16 bits) */ 67 #define QMC_GBL_TXPTR 0x1A 68 /* CRC constant (32 bits) */ 69 #define QMC_GBL_C_MASK32 0x1C 70 /* Time slot assignment table Rx (32 x 16 bits) */ 71 #define QMC_GBL_TSATRX 0x20 72 /* Time slot assignment table Tx (32 x 16 bits) */ 73 #define QMC_GBL_TSATTX 0x60 74 /* CRC constant (16 bits) */ 75 #define QMC_GBL_C_MASK16 0xA0 76 77 /* TSA entry (16bit entry in TSATRX and TSATTX) */ 78 #define QMC_TSA_VALID (1 << 15) 79 #define QMC_TSA_WRAP (1 << 14) 80 #define QMC_TSA_MASK (0x303F) 81 #define QMC_TSA_CHANNEL(x) ((x) << 6) 82 83 /* Tx buffer descriptor base address (16 bits, offset from MCBASE) */ 84 #define QMC_SPE_TBASE 0x00 85 86 /* Channel mode register (16 bits) */ 87 #define QMC_SPE_CHAMR 0x02 88 #define QMC_SPE_CHAMR_MODE_HDLC (1 << 15) 89 #define QMC_SPE_CHAMR_MODE_TRANSP ((0 << 15) | (1 << 13)) 90 #define QMC_SPE_CHAMR_ENT (1 << 12) 91 #define QMC_SPE_CHAMR_POL (1 << 8) 92 #define QMC_SPE_CHAMR_HDLC_IDLM (1 << 13) 93 #define QMC_SPE_CHAMR_HDLC_CRC (1 << 7) 94 #define QMC_SPE_CHAMR_HDLC_NOF (0x0f << 0) 95 #define QMC_SPE_CHAMR_TRANSP_RD (1 << 14) 96 #define QMC_SPE_CHAMR_TRANSP_SYNC (1 << 10) 97 98 /* Tx internal state (32 bits) */ 99 #define QMC_SPE_TSTATE 0x04 100 /* Tx buffer descriptor pointer (16 bits) */ 101 #define QMC_SPE_TBPTR 0x0C 102 /* Zero-insertion state (32 bits) */ 103 #define QMC_SPE_ZISTATE 0x14 104 /* Channel’s interrupt mask flags (16 bits) */ 105 #define QMC_SPE_INTMSK 0x1C 106 /* Rx buffer descriptor base address (16 bits, offset from MCBASE) */ 107 #define QMC_SPE_RBASE 0x20 108 /* HDLC: Maximum frame length register (16 bits) */ 109 #define QMC_SPE_MFLR 0x22 110 /* TRANSPARENT: Transparent maximum receive length (16 bits) */ 111 #define QMC_SPE_TMRBLR 0x22 112 /* Rx internal state (32 bits) */ 113 #define QMC_SPE_RSTATE 0x24 114 /* Rx buffer descriptor pointer (16 bits) */ 115 #define QMC_SPE_RBPTR 0x2C 116 /* Packs 4 bytes to 1 long word before writing to buffer (32 bits) */ 117 #define QMC_SPE_RPACK 0x30 118 /* Zero deletion state (32 bits) */ 119 #define QMC_SPE_ZDSTATE 0x34 120 121 /* Transparent synchronization (16 bits) */ 122 #define QMC_SPE_TRNSYNC 0x3C 123 #define QMC_SPE_TRNSYNC_RX(x) ((x) << 8) 124 #define QMC_SPE_TRNSYNC_TX(x) ((x) << 0) 125 126 /* Interrupt related registers bits */ 127 #define QMC_INT_V (1 << 15) 128 #define QMC_INT_W (1 << 14) 129 #define QMC_INT_NID (1 << 13) 130 #define QMC_INT_IDL (1 << 12) 131 #define QMC_INT_GET_CHANNEL(x) (((x) & 0x0FC0) >> 6) 132 #define QMC_INT_MRF (1 << 5) 133 #define QMC_INT_UN (1 << 4) 134 #define QMC_INT_RXF (1 << 3) 135 #define QMC_INT_BSY (1 << 2) 136 #define QMC_INT_TXB (1 << 1) 137 #define QMC_INT_RXB (1 << 0) 138 139 /* BD related registers bits */ 140 #define QMC_BD_RX_E (1 << 15) 141 #define QMC_BD_RX_W (1 << 13) 142 #define QMC_BD_RX_I (1 << 12) 143 #define QMC_BD_RX_L (1 << 11) 144 #define QMC_BD_RX_F (1 << 10) 145 #define QMC_BD_RX_CM (1 << 9) 146 #define QMC_BD_RX_UB (1 << 7) 147 #define QMC_BD_RX_LG (1 << 5) 148 #define QMC_BD_RX_NO (1 << 4) 149 #define QMC_BD_RX_AB (1 << 3) 150 #define QMC_BD_RX_CR (1 << 2) 151 152 #define QMC_BD_TX_R (1 << 15) 153 #define QMC_BD_TX_W (1 << 13) 154 #define QMC_BD_TX_I (1 << 12) 155 #define QMC_BD_TX_L (1 << 11) 156 #define QMC_BD_TX_TC (1 << 10) 157 #define QMC_BD_TX_CM (1 << 9) 158 #define QMC_BD_TX_UB (1 << 7) 159 #define QMC_BD_TX_PAD (0x0f << 0) 160 161 /* Numbers of BDs and interrupt items */ 162 #define QMC_NB_TXBDS 8 163 #define QMC_NB_RXBDS 8 164 #define QMC_NB_INTS 128 165 166 struct qmc_xfer_desc { 167 union { 168 void (*tx_complete)(void *context); 169 void (*rx_complete)(void *context, size_t length); 170 }; 171 void *context; 172 }; 173 174 struct qmc_chan { 175 struct list_head list; 176 unsigned int id; 177 struct qmc *qmc; 178 void *__iomem s_param; 179 enum qmc_mode mode; 180 u64 tx_ts_mask; 181 u64 rx_ts_mask; 182 bool is_reverse_data; 183 184 spinlock_t tx_lock; 185 cbd_t __iomem *txbds; 186 cbd_t __iomem *txbd_free; 187 cbd_t __iomem *txbd_done; 188 struct qmc_xfer_desc tx_desc[QMC_NB_TXBDS]; 189 u64 nb_tx_underrun; 190 bool is_tx_stopped; 191 192 spinlock_t rx_lock; 193 cbd_t __iomem *rxbds; 194 cbd_t __iomem *rxbd_free; 195 cbd_t __iomem *rxbd_done; 196 struct qmc_xfer_desc rx_desc[QMC_NB_RXBDS]; 197 u64 nb_rx_busy; 198 int rx_pending; 199 bool is_rx_halted; 200 bool is_rx_stopped; 201 }; 202 203 struct qmc { 204 struct device *dev; 205 struct tsa_serial *tsa_serial; 206 void *__iomem scc_regs; 207 void *__iomem scc_pram; 208 void *__iomem dpram; 209 u16 scc_pram_offset; 210 cbd_t __iomem *bd_table; 211 dma_addr_t bd_dma_addr; 212 size_t bd_size; 213 u16 __iomem *int_table; 214 u16 __iomem *int_curr; 215 dma_addr_t int_dma_addr; 216 size_t int_size; 217 struct list_head chan_head; 218 struct qmc_chan *chans[64]; 219 }; 220 221 static inline void qmc_write16(void *__iomem addr, u16 val) 222 { 223 iowrite16be(val, addr); 224 } 225 226 static inline u16 qmc_read16(void *__iomem addr) 227 { 228 return ioread16be(addr); 229 } 230 231 static inline void qmc_setbits16(void *__iomem addr, u16 set) 232 { 233 qmc_write16(addr, qmc_read16(addr) | set); 234 } 235 236 static inline void qmc_clrbits16(void *__iomem addr, u16 clr) 237 { 238 qmc_write16(addr, qmc_read16(addr) & ~clr); 239 } 240 241 static inline void qmc_write32(void *__iomem addr, u32 val) 242 { 243 iowrite32be(val, addr); 244 } 245 246 static inline u32 qmc_read32(void *__iomem addr) 247 { 248 return ioread32be(addr); 249 } 250 251 static inline void qmc_setbits32(void *__iomem addr, u32 set) 252 { 253 qmc_write32(addr, qmc_read32(addr) | set); 254 } 255 256 257 int qmc_chan_get_info(struct qmc_chan *chan, struct qmc_chan_info *info) 258 { 259 struct tsa_serial_info tsa_info; 260 int ret; 261 262 /* Retrieve info from the TSA related serial */ 263 ret = tsa_serial_get_info(chan->qmc->tsa_serial, &tsa_info); 264 if (ret) 265 return ret; 266 267 info->mode = chan->mode; 268 info->rx_fs_rate = tsa_info.rx_fs_rate; 269 info->rx_bit_rate = tsa_info.rx_bit_rate; 270 info->nb_tx_ts = hweight64(chan->tx_ts_mask); 271 info->tx_fs_rate = tsa_info.tx_fs_rate; 272 info->tx_bit_rate = tsa_info.tx_bit_rate; 273 info->nb_rx_ts = hweight64(chan->rx_ts_mask); 274 275 return 0; 276 } 277 EXPORT_SYMBOL(qmc_chan_get_info); 278 279 int qmc_chan_set_param(struct qmc_chan *chan, const struct qmc_chan_param *param) 280 { 281 if (param->mode != chan->mode) 282 return -EINVAL; 283 284 switch (param->mode) { 285 case QMC_HDLC: 286 if ((param->hdlc.max_rx_buf_size % 4) || 287 (param->hdlc.max_rx_buf_size < 8)) 288 return -EINVAL; 289 290 qmc_write16(chan->qmc->scc_pram + QMC_GBL_MRBLR, 291 param->hdlc.max_rx_buf_size - 8); 292 qmc_write16(chan->s_param + QMC_SPE_MFLR, 293 param->hdlc.max_rx_frame_size); 294 if (param->hdlc.is_crc32) { 295 qmc_setbits16(chan->s_param + QMC_SPE_CHAMR, 296 QMC_SPE_CHAMR_HDLC_CRC); 297 } else { 298 qmc_clrbits16(chan->s_param + QMC_SPE_CHAMR, 299 QMC_SPE_CHAMR_HDLC_CRC); 300 } 301 break; 302 303 case QMC_TRANSPARENT: 304 qmc_write16(chan->s_param + QMC_SPE_TMRBLR, 305 param->transp.max_rx_buf_size); 306 break; 307 308 default: 309 return -EINVAL; 310 } 311 312 return 0; 313 } 314 EXPORT_SYMBOL(qmc_chan_set_param); 315 316 int qmc_chan_write_submit(struct qmc_chan *chan, dma_addr_t addr, size_t length, 317 void (*complete)(void *context), void *context) 318 { 319 struct qmc_xfer_desc *xfer_desc; 320 unsigned long flags; 321 cbd_t *__iomem bd; 322 u16 ctrl; 323 int ret; 324 325 /* 326 * R bit UB bit 327 * 0 0 : The BD is free 328 * 1 1 : The BD is in used, waiting for transfer 329 * 0 1 : The BD is in used, waiting for completion 330 * 1 0 : Should not append 331 */ 332 333 spin_lock_irqsave(&chan->tx_lock, flags); 334 bd = chan->txbd_free; 335 336 ctrl = qmc_read16(&bd->cbd_sc); 337 if (ctrl & (QMC_BD_TX_R | QMC_BD_TX_UB)) { 338 /* We are full ... */ 339 ret = -EBUSY; 340 goto end; 341 } 342 343 qmc_write16(&bd->cbd_datlen, length); 344 qmc_write32(&bd->cbd_bufaddr, addr); 345 346 xfer_desc = &chan->tx_desc[bd - chan->txbds]; 347 xfer_desc->tx_complete = complete; 348 xfer_desc->context = context; 349 350 /* Activate the descriptor */ 351 ctrl |= (QMC_BD_TX_R | QMC_BD_TX_UB); 352 wmb(); /* Be sure to flush the descriptor before control update */ 353 qmc_write16(&bd->cbd_sc, ctrl); 354 355 if (!chan->is_tx_stopped) 356 qmc_setbits16(chan->s_param + QMC_SPE_CHAMR, QMC_SPE_CHAMR_POL); 357 358 if (ctrl & QMC_BD_TX_W) 359 chan->txbd_free = chan->txbds; 360 else 361 chan->txbd_free++; 362 363 ret = 0; 364 365 end: 366 spin_unlock_irqrestore(&chan->tx_lock, flags); 367 return ret; 368 } 369 EXPORT_SYMBOL(qmc_chan_write_submit); 370 371 static void qmc_chan_write_done(struct qmc_chan *chan) 372 { 373 struct qmc_xfer_desc *xfer_desc; 374 void (*complete)(void *context); 375 unsigned long flags; 376 void *context; 377 cbd_t *__iomem bd; 378 u16 ctrl; 379 380 /* 381 * R bit UB bit 382 * 0 0 : The BD is free 383 * 1 1 : The BD is in used, waiting for transfer 384 * 0 1 : The BD is in used, waiting for completion 385 * 1 0 : Should not append 386 */ 387 388 spin_lock_irqsave(&chan->tx_lock, flags); 389 bd = chan->txbd_done; 390 391 ctrl = qmc_read16(&bd->cbd_sc); 392 while (!(ctrl & QMC_BD_TX_R)) { 393 if (!(ctrl & QMC_BD_TX_UB)) 394 goto end; 395 396 xfer_desc = &chan->tx_desc[bd - chan->txbds]; 397 complete = xfer_desc->tx_complete; 398 context = xfer_desc->context; 399 xfer_desc->tx_complete = NULL; 400 xfer_desc->context = NULL; 401 402 qmc_write16(&bd->cbd_sc, ctrl & ~QMC_BD_TX_UB); 403 404 if (ctrl & QMC_BD_TX_W) 405 chan->txbd_done = chan->txbds; 406 else 407 chan->txbd_done++; 408 409 if (complete) { 410 spin_unlock_irqrestore(&chan->tx_lock, flags); 411 complete(context); 412 spin_lock_irqsave(&chan->tx_lock, flags); 413 } 414 415 bd = chan->txbd_done; 416 ctrl = qmc_read16(&bd->cbd_sc); 417 } 418 419 end: 420 spin_unlock_irqrestore(&chan->tx_lock, flags); 421 } 422 423 int qmc_chan_read_submit(struct qmc_chan *chan, dma_addr_t addr, size_t length, 424 void (*complete)(void *context, size_t length), void *context) 425 { 426 struct qmc_xfer_desc *xfer_desc; 427 unsigned long flags; 428 cbd_t *__iomem bd; 429 u16 ctrl; 430 int ret; 431 432 /* 433 * E bit UB bit 434 * 0 0 : The BD is free 435 * 1 1 : The BD is in used, waiting for transfer 436 * 0 1 : The BD is in used, waiting for completion 437 * 1 0 : Should not append 438 */ 439 440 spin_lock_irqsave(&chan->rx_lock, flags); 441 bd = chan->rxbd_free; 442 443 ctrl = qmc_read16(&bd->cbd_sc); 444 if (ctrl & (QMC_BD_RX_E | QMC_BD_RX_UB)) { 445 /* We are full ... */ 446 ret = -EBUSY; 447 goto end; 448 } 449 450 qmc_write16(&bd->cbd_datlen, 0); /* data length is updated by the QMC */ 451 qmc_write32(&bd->cbd_bufaddr, addr); 452 453 xfer_desc = &chan->rx_desc[bd - chan->rxbds]; 454 xfer_desc->rx_complete = complete; 455 xfer_desc->context = context; 456 457 /* Activate the descriptor */ 458 ctrl |= (QMC_BD_RX_E | QMC_BD_RX_UB); 459 wmb(); /* Be sure to flush data before descriptor activation */ 460 qmc_write16(&bd->cbd_sc, ctrl); 461 462 /* Restart receiver if needed */ 463 if (chan->is_rx_halted && !chan->is_rx_stopped) { 464 /* Restart receiver */ 465 if (chan->mode == QMC_TRANSPARENT) 466 qmc_write32(chan->s_param + QMC_SPE_ZDSTATE, 0x18000080); 467 else 468 qmc_write32(chan->s_param + QMC_SPE_ZDSTATE, 0x00000080); 469 qmc_write32(chan->s_param + QMC_SPE_RSTATE, 0x31000000); 470 chan->is_rx_halted = false; 471 } 472 chan->rx_pending++; 473 474 if (ctrl & QMC_BD_RX_W) 475 chan->rxbd_free = chan->rxbds; 476 else 477 chan->rxbd_free++; 478 479 ret = 0; 480 end: 481 spin_unlock_irqrestore(&chan->rx_lock, flags); 482 return ret; 483 } 484 EXPORT_SYMBOL(qmc_chan_read_submit); 485 486 static void qmc_chan_read_done(struct qmc_chan *chan) 487 { 488 void (*complete)(void *context, size_t size); 489 struct qmc_xfer_desc *xfer_desc; 490 unsigned long flags; 491 cbd_t *__iomem bd; 492 void *context; 493 u16 datalen; 494 u16 ctrl; 495 496 /* 497 * E bit UB bit 498 * 0 0 : The BD is free 499 * 1 1 : The BD is in used, waiting for transfer 500 * 0 1 : The BD is in used, waiting for completion 501 * 1 0 : Should not append 502 */ 503 504 spin_lock_irqsave(&chan->rx_lock, flags); 505 bd = chan->rxbd_done; 506 507 ctrl = qmc_read16(&bd->cbd_sc); 508 while (!(ctrl & QMC_BD_RX_E)) { 509 if (!(ctrl & QMC_BD_RX_UB)) 510 goto end; 511 512 xfer_desc = &chan->rx_desc[bd - chan->rxbds]; 513 complete = xfer_desc->rx_complete; 514 context = xfer_desc->context; 515 xfer_desc->rx_complete = NULL; 516 xfer_desc->context = NULL; 517 518 datalen = qmc_read16(&bd->cbd_datlen); 519 qmc_write16(&bd->cbd_sc, ctrl & ~QMC_BD_RX_UB); 520 521 if (ctrl & QMC_BD_RX_W) 522 chan->rxbd_done = chan->rxbds; 523 else 524 chan->rxbd_done++; 525 526 chan->rx_pending--; 527 528 if (complete) { 529 spin_unlock_irqrestore(&chan->rx_lock, flags); 530 complete(context, datalen); 531 spin_lock_irqsave(&chan->rx_lock, flags); 532 } 533 534 bd = chan->rxbd_done; 535 ctrl = qmc_read16(&bd->cbd_sc); 536 } 537 538 end: 539 spin_unlock_irqrestore(&chan->rx_lock, flags); 540 } 541 542 static int qmc_chan_command(struct qmc_chan *chan, u8 qmc_opcode) 543 { 544 return cpm_command(chan->id << 2, (qmc_opcode << 4) | 0x0E); 545 } 546 547 static int qmc_chan_stop_rx(struct qmc_chan *chan) 548 { 549 unsigned long flags; 550 int ret; 551 552 spin_lock_irqsave(&chan->rx_lock, flags); 553 554 /* Send STOP RECEIVE command */ 555 ret = qmc_chan_command(chan, 0x0); 556 if (ret) { 557 dev_err(chan->qmc->dev, "chan %u: Send STOP RECEIVE failed (%d)\n", 558 chan->id, ret); 559 goto end; 560 } 561 562 chan->is_rx_stopped = true; 563 564 end: 565 spin_unlock_irqrestore(&chan->rx_lock, flags); 566 return ret; 567 } 568 569 static int qmc_chan_stop_tx(struct qmc_chan *chan) 570 { 571 unsigned long flags; 572 int ret; 573 574 spin_lock_irqsave(&chan->tx_lock, flags); 575 576 /* Send STOP TRANSMIT command */ 577 ret = qmc_chan_command(chan, 0x1); 578 if (ret) { 579 dev_err(chan->qmc->dev, "chan %u: Send STOP TRANSMIT failed (%d)\n", 580 chan->id, ret); 581 goto end; 582 } 583 584 chan->is_tx_stopped = true; 585 586 end: 587 spin_unlock_irqrestore(&chan->tx_lock, flags); 588 return ret; 589 } 590 591 int qmc_chan_stop(struct qmc_chan *chan, int direction) 592 { 593 int ret; 594 595 if (direction & QMC_CHAN_READ) { 596 ret = qmc_chan_stop_rx(chan); 597 if (ret) 598 return ret; 599 } 600 601 if (direction & QMC_CHAN_WRITE) { 602 ret = qmc_chan_stop_tx(chan); 603 if (ret) 604 return ret; 605 } 606 607 return 0; 608 } 609 EXPORT_SYMBOL(qmc_chan_stop); 610 611 static void qmc_chan_start_rx(struct qmc_chan *chan) 612 { 613 unsigned long flags; 614 615 spin_lock_irqsave(&chan->rx_lock, flags); 616 617 /* Restart the receiver */ 618 if (chan->mode == QMC_TRANSPARENT) 619 qmc_write32(chan->s_param + QMC_SPE_ZDSTATE, 0x18000080); 620 else 621 qmc_write32(chan->s_param + QMC_SPE_ZDSTATE, 0x00000080); 622 qmc_write32(chan->s_param + QMC_SPE_RSTATE, 0x31000000); 623 chan->is_rx_halted = false; 624 625 chan->is_rx_stopped = false; 626 627 spin_unlock_irqrestore(&chan->rx_lock, flags); 628 } 629 630 static void qmc_chan_start_tx(struct qmc_chan *chan) 631 { 632 unsigned long flags; 633 634 spin_lock_irqsave(&chan->tx_lock, flags); 635 636 /* 637 * Enable channel transmitter as it could be disabled if 638 * qmc_chan_reset() was called. 639 */ 640 qmc_setbits16(chan->s_param + QMC_SPE_CHAMR, QMC_SPE_CHAMR_ENT); 641 642 /* Set the POL bit in the channel mode register */ 643 qmc_setbits16(chan->s_param + QMC_SPE_CHAMR, QMC_SPE_CHAMR_POL); 644 645 chan->is_tx_stopped = false; 646 647 spin_unlock_irqrestore(&chan->tx_lock, flags); 648 } 649 650 int qmc_chan_start(struct qmc_chan *chan, int direction) 651 { 652 if (direction & QMC_CHAN_READ) 653 qmc_chan_start_rx(chan); 654 655 if (direction & QMC_CHAN_WRITE) 656 qmc_chan_start_tx(chan); 657 658 return 0; 659 } 660 EXPORT_SYMBOL(qmc_chan_start); 661 662 static void qmc_chan_reset_rx(struct qmc_chan *chan) 663 { 664 struct qmc_xfer_desc *xfer_desc; 665 unsigned long flags; 666 cbd_t *__iomem bd; 667 u16 ctrl; 668 669 spin_lock_irqsave(&chan->rx_lock, flags); 670 bd = chan->rxbds; 671 do { 672 ctrl = qmc_read16(&bd->cbd_sc); 673 qmc_write16(&bd->cbd_sc, ctrl & ~(QMC_BD_RX_UB | QMC_BD_RX_E)); 674 675 xfer_desc = &chan->rx_desc[bd - chan->rxbds]; 676 xfer_desc->rx_complete = NULL; 677 xfer_desc->context = NULL; 678 679 bd++; 680 } while (!(ctrl & QMC_BD_RX_W)); 681 682 chan->rxbd_free = chan->rxbds; 683 chan->rxbd_done = chan->rxbds; 684 qmc_write16(chan->s_param + QMC_SPE_RBPTR, 685 qmc_read16(chan->s_param + QMC_SPE_RBASE)); 686 687 chan->rx_pending = 0; 688 chan->is_rx_stopped = false; 689 690 spin_unlock_irqrestore(&chan->rx_lock, flags); 691 } 692 693 static void qmc_chan_reset_tx(struct qmc_chan *chan) 694 { 695 struct qmc_xfer_desc *xfer_desc; 696 unsigned long flags; 697 cbd_t *__iomem bd; 698 u16 ctrl; 699 700 spin_lock_irqsave(&chan->tx_lock, flags); 701 702 /* Disable transmitter. It will be re-enable on qmc_chan_start() */ 703 qmc_clrbits16(chan->s_param + QMC_SPE_CHAMR, QMC_SPE_CHAMR_ENT); 704 705 bd = chan->txbds; 706 do { 707 ctrl = qmc_read16(&bd->cbd_sc); 708 qmc_write16(&bd->cbd_sc, ctrl & ~(QMC_BD_TX_UB | QMC_BD_TX_R)); 709 710 xfer_desc = &chan->tx_desc[bd - chan->txbds]; 711 xfer_desc->tx_complete = NULL; 712 xfer_desc->context = NULL; 713 714 bd++; 715 } while (!(ctrl & QMC_BD_TX_W)); 716 717 chan->txbd_free = chan->txbds; 718 chan->txbd_done = chan->txbds; 719 qmc_write16(chan->s_param + QMC_SPE_TBPTR, 720 qmc_read16(chan->s_param + QMC_SPE_TBASE)); 721 722 /* Reset TSTATE and ZISTATE to their initial value */ 723 qmc_write32(chan->s_param + QMC_SPE_TSTATE, 0x30000000); 724 qmc_write32(chan->s_param + QMC_SPE_ZISTATE, 0x00000100); 725 726 spin_unlock_irqrestore(&chan->tx_lock, flags); 727 } 728 729 int qmc_chan_reset(struct qmc_chan *chan, int direction) 730 { 731 if (direction & QMC_CHAN_READ) 732 qmc_chan_reset_rx(chan); 733 734 if (direction & QMC_CHAN_WRITE) 735 qmc_chan_reset_tx(chan); 736 737 return 0; 738 } 739 EXPORT_SYMBOL(qmc_chan_reset); 740 741 static int qmc_check_chans(struct qmc *qmc) 742 { 743 struct tsa_serial_info info; 744 bool is_one_table = false; 745 struct qmc_chan *chan; 746 u64 tx_ts_mask = 0; 747 u64 rx_ts_mask = 0; 748 u64 tx_ts_assigned_mask; 749 u64 rx_ts_assigned_mask; 750 int ret; 751 752 /* Retrieve info from the TSA related serial */ 753 ret = tsa_serial_get_info(qmc->tsa_serial, &info); 754 if (ret) 755 return ret; 756 757 if ((info.nb_tx_ts > 64) || (info.nb_rx_ts > 64)) { 758 dev_err(qmc->dev, "Number of TSA Tx/Rx TS assigned not supported\n"); 759 return -EINVAL; 760 } 761 762 /* 763 * If more than 32 TS are assigned to this serial, one common table is 764 * used for Tx and Rx and so masks must be equal for all channels. 765 */ 766 if ((info.nb_tx_ts > 32) || (info.nb_rx_ts > 32)) { 767 if (info.nb_tx_ts != info.nb_rx_ts) { 768 dev_err(qmc->dev, "Number of TSA Tx/Rx TS assigned are not equal\n"); 769 return -EINVAL; 770 } 771 is_one_table = true; 772 } 773 774 tx_ts_assigned_mask = info.nb_tx_ts == 64 ? U64_MAX : (((u64)1) << info.nb_tx_ts) - 1; 775 rx_ts_assigned_mask = info.nb_rx_ts == 64 ? U64_MAX : (((u64)1) << info.nb_rx_ts) - 1; 776 777 list_for_each_entry(chan, &qmc->chan_head, list) { 778 if (chan->tx_ts_mask > tx_ts_assigned_mask) { 779 dev_err(qmc->dev, "chan %u uses TSA unassigned Tx TS\n", chan->id); 780 return -EINVAL; 781 } 782 if (tx_ts_mask & chan->tx_ts_mask) { 783 dev_err(qmc->dev, "chan %u uses an already used Tx TS\n", chan->id); 784 return -EINVAL; 785 } 786 787 if (chan->rx_ts_mask > rx_ts_assigned_mask) { 788 dev_err(qmc->dev, "chan %u uses TSA unassigned Rx TS\n", chan->id); 789 return -EINVAL; 790 } 791 if (rx_ts_mask & chan->rx_ts_mask) { 792 dev_err(qmc->dev, "chan %u uses an already used Rx TS\n", chan->id); 793 return -EINVAL; 794 } 795 796 if (is_one_table && (chan->tx_ts_mask != chan->rx_ts_mask)) { 797 dev_err(qmc->dev, "chan %u uses different Rx and Tx TS\n", chan->id); 798 return -EINVAL; 799 } 800 801 tx_ts_mask |= chan->tx_ts_mask; 802 rx_ts_mask |= chan->rx_ts_mask; 803 } 804 805 return 0; 806 } 807 808 static unsigned int qmc_nb_chans(struct qmc *qmc) 809 { 810 unsigned int count = 0; 811 struct qmc_chan *chan; 812 813 list_for_each_entry(chan, &qmc->chan_head, list) 814 count++; 815 816 return count; 817 } 818 819 static int qmc_of_parse_chans(struct qmc *qmc, struct device_node *np) 820 { 821 struct device_node *chan_np; 822 struct qmc_chan *chan; 823 const char *mode; 824 u32 chan_id; 825 u64 ts_mask; 826 int ret; 827 828 for_each_available_child_of_node(np, chan_np) { 829 ret = of_property_read_u32(chan_np, "reg", &chan_id); 830 if (ret) { 831 dev_err(qmc->dev, "%pOF: failed to read reg\n", chan_np); 832 of_node_put(chan_np); 833 return ret; 834 } 835 if (chan_id > 63) { 836 dev_err(qmc->dev, "%pOF: Invalid chan_id\n", chan_np); 837 of_node_put(chan_np); 838 return -EINVAL; 839 } 840 841 chan = devm_kzalloc(qmc->dev, sizeof(*chan), GFP_KERNEL); 842 if (!chan) { 843 of_node_put(chan_np); 844 return -ENOMEM; 845 } 846 847 chan->id = chan_id; 848 spin_lock_init(&chan->rx_lock); 849 spin_lock_init(&chan->tx_lock); 850 851 ret = of_property_read_u64(chan_np, "fsl,tx-ts-mask", &ts_mask); 852 if (ret) { 853 dev_err(qmc->dev, "%pOF: failed to read fsl,tx-ts-mask\n", 854 chan_np); 855 of_node_put(chan_np); 856 return ret; 857 } 858 chan->tx_ts_mask = ts_mask; 859 860 ret = of_property_read_u64(chan_np, "fsl,rx-ts-mask", &ts_mask); 861 if (ret) { 862 dev_err(qmc->dev, "%pOF: failed to read fsl,rx-ts-mask\n", 863 chan_np); 864 of_node_put(chan_np); 865 return ret; 866 } 867 chan->rx_ts_mask = ts_mask; 868 869 mode = "transparent"; 870 ret = of_property_read_string(chan_np, "fsl,operational-mode", &mode); 871 if (ret && ret != -EINVAL) { 872 dev_err(qmc->dev, "%pOF: failed to read fsl,operational-mode\n", 873 chan_np); 874 of_node_put(chan_np); 875 return ret; 876 } 877 if (!strcmp(mode, "transparent")) { 878 chan->mode = QMC_TRANSPARENT; 879 } else if (!strcmp(mode, "hdlc")) { 880 chan->mode = QMC_HDLC; 881 } else { 882 dev_err(qmc->dev, "%pOF: Invalid fsl,operational-mode (%s)\n", 883 chan_np, mode); 884 of_node_put(chan_np); 885 return -EINVAL; 886 } 887 888 chan->is_reverse_data = of_property_read_bool(chan_np, 889 "fsl,reverse-data"); 890 891 list_add_tail(&chan->list, &qmc->chan_head); 892 qmc->chans[chan->id] = chan; 893 } 894 895 return qmc_check_chans(qmc); 896 } 897 898 static int qmc_setup_tsa_64rxtx(struct qmc *qmc, const struct tsa_serial_info *info) 899 { 900 struct qmc_chan *chan; 901 unsigned int i; 902 u16 val; 903 904 /* 905 * Use a common Tx/Rx 64 entries table. 906 * Everything was previously checked, Tx and Rx related stuffs are 907 * identical -> Used Rx related stuff to build the table 908 */ 909 910 /* Invalidate all entries */ 911 for (i = 0; i < 64; i++) 912 qmc_write16(qmc->scc_pram + QMC_GBL_TSATRX + (i * 2), 0x0000); 913 914 /* Set entries based on Rx stuff*/ 915 list_for_each_entry(chan, &qmc->chan_head, list) { 916 for (i = 0; i < info->nb_rx_ts; i++) { 917 if (!(chan->rx_ts_mask & (((u64)1) << i))) 918 continue; 919 920 val = QMC_TSA_VALID | QMC_TSA_MASK | 921 QMC_TSA_CHANNEL(chan->id); 922 qmc_write16(qmc->scc_pram + QMC_GBL_TSATRX + (i * 2), val); 923 } 924 } 925 926 /* Set Wrap bit on last entry */ 927 qmc_setbits16(qmc->scc_pram + QMC_GBL_TSATRX + ((info->nb_rx_ts - 1) * 2), 928 QMC_TSA_WRAP); 929 930 /* Init pointers to the table */ 931 val = qmc->scc_pram_offset + QMC_GBL_TSATRX; 932 qmc_write16(qmc->scc_pram + QMC_GBL_RX_S_PTR, val); 933 qmc_write16(qmc->scc_pram + QMC_GBL_RXPTR, val); 934 qmc_write16(qmc->scc_pram + QMC_GBL_TX_S_PTR, val); 935 qmc_write16(qmc->scc_pram + QMC_GBL_TXPTR, val); 936 937 return 0; 938 } 939 940 static int qmc_setup_tsa_32rx_32tx(struct qmc *qmc, const struct tsa_serial_info *info) 941 { 942 struct qmc_chan *chan; 943 unsigned int i; 944 u16 val; 945 946 /* 947 * Use a Tx 32 entries table and a Rx 32 entries table. 948 * Everything was previously checked. 949 */ 950 951 /* Invalidate all entries */ 952 for (i = 0; i < 32; i++) { 953 qmc_write16(qmc->scc_pram + QMC_GBL_TSATRX + (i * 2), 0x0000); 954 qmc_write16(qmc->scc_pram + QMC_GBL_TSATTX + (i * 2), 0x0000); 955 } 956 957 /* Set entries based on Rx and Tx stuff*/ 958 list_for_each_entry(chan, &qmc->chan_head, list) { 959 /* Rx part */ 960 for (i = 0; i < info->nb_rx_ts; i++) { 961 if (!(chan->rx_ts_mask & (((u64)1) << i))) 962 continue; 963 964 val = QMC_TSA_VALID | QMC_TSA_MASK | 965 QMC_TSA_CHANNEL(chan->id); 966 qmc_write16(qmc->scc_pram + QMC_GBL_TSATRX + (i * 2), val); 967 } 968 /* Tx part */ 969 for (i = 0; i < info->nb_tx_ts; i++) { 970 if (!(chan->tx_ts_mask & (((u64)1) << i))) 971 continue; 972 973 val = QMC_TSA_VALID | QMC_TSA_MASK | 974 QMC_TSA_CHANNEL(chan->id); 975 qmc_write16(qmc->scc_pram + QMC_GBL_TSATTX + (i * 2), val); 976 } 977 } 978 979 /* Set Wrap bit on last entries */ 980 qmc_setbits16(qmc->scc_pram + QMC_GBL_TSATRX + ((info->nb_rx_ts - 1) * 2), 981 QMC_TSA_WRAP); 982 qmc_setbits16(qmc->scc_pram + QMC_GBL_TSATTX + ((info->nb_tx_ts - 1) * 2), 983 QMC_TSA_WRAP); 984 985 /* Init Rx pointers ...*/ 986 val = qmc->scc_pram_offset + QMC_GBL_TSATRX; 987 qmc_write16(qmc->scc_pram + QMC_GBL_RX_S_PTR, val); 988 qmc_write16(qmc->scc_pram + QMC_GBL_RXPTR, val); 989 990 /* ... and Tx pointers */ 991 val = qmc->scc_pram_offset + QMC_GBL_TSATTX; 992 qmc_write16(qmc->scc_pram + QMC_GBL_TX_S_PTR, val); 993 qmc_write16(qmc->scc_pram + QMC_GBL_TXPTR, val); 994 995 return 0; 996 } 997 998 static int qmc_setup_tsa(struct qmc *qmc) 999 { 1000 struct tsa_serial_info info; 1001 int ret; 1002 1003 /* Retrieve info from the TSA related serial */ 1004 ret = tsa_serial_get_info(qmc->tsa_serial, &info); 1005 if (ret) 1006 return ret; 1007 1008 /* 1009 * Setup one common 64 entries table or two 32 entries (one for Tx and 1010 * one for Tx) according to assigned TS numbers. 1011 */ 1012 return ((info.nb_tx_ts > 32) || (info.nb_rx_ts > 32)) ? 1013 qmc_setup_tsa_64rxtx(qmc, &info) : 1014 qmc_setup_tsa_32rx_32tx(qmc, &info); 1015 } 1016 1017 static int qmc_setup_chan_trnsync(struct qmc *qmc, struct qmc_chan *chan) 1018 { 1019 struct tsa_serial_info info; 1020 u16 first_rx, last_tx; 1021 u16 trnsync; 1022 int ret; 1023 1024 /* Retrieve info from the TSA related serial */ 1025 ret = tsa_serial_get_info(chan->qmc->tsa_serial, &info); 1026 if (ret) 1027 return ret; 1028 1029 /* Find the first Rx TS allocated to the channel */ 1030 first_rx = chan->rx_ts_mask ? __ffs64(chan->rx_ts_mask) + 1 : 0; 1031 1032 /* Find the last Tx TS allocated to the channel */ 1033 last_tx = fls64(chan->tx_ts_mask); 1034 1035 trnsync = 0; 1036 if (info.nb_rx_ts) 1037 trnsync |= QMC_SPE_TRNSYNC_RX((first_rx % info.nb_rx_ts) * 2); 1038 if (info.nb_tx_ts) 1039 trnsync |= QMC_SPE_TRNSYNC_TX((last_tx % info.nb_tx_ts) * 2); 1040 1041 qmc_write16(chan->s_param + QMC_SPE_TRNSYNC, trnsync); 1042 1043 dev_dbg(qmc->dev, "chan %u: trnsync=0x%04x, rx %u/%u 0x%llx, tx %u/%u 0x%llx\n", 1044 chan->id, trnsync, 1045 first_rx, info.nb_rx_ts, chan->rx_ts_mask, 1046 last_tx, info.nb_tx_ts, chan->tx_ts_mask); 1047 1048 return 0; 1049 } 1050 1051 static int qmc_setup_chan(struct qmc *qmc, struct qmc_chan *chan) 1052 { 1053 unsigned int i; 1054 cbd_t __iomem *bd; 1055 int ret; 1056 u16 val; 1057 1058 chan->qmc = qmc; 1059 1060 /* Set channel specific parameter base address */ 1061 chan->s_param = qmc->dpram + (chan->id * 64); 1062 /* 16 bd per channel (8 rx and 8 tx) */ 1063 chan->txbds = qmc->bd_table + (chan->id * (QMC_NB_TXBDS + QMC_NB_RXBDS)); 1064 chan->rxbds = qmc->bd_table + (chan->id * (QMC_NB_TXBDS + QMC_NB_RXBDS)) + QMC_NB_TXBDS; 1065 1066 chan->txbd_free = chan->txbds; 1067 chan->txbd_done = chan->txbds; 1068 chan->rxbd_free = chan->rxbds; 1069 chan->rxbd_done = chan->rxbds; 1070 1071 /* TBASE and TBPTR*/ 1072 val = chan->id * (QMC_NB_TXBDS + QMC_NB_RXBDS) * sizeof(cbd_t); 1073 qmc_write16(chan->s_param + QMC_SPE_TBASE, val); 1074 qmc_write16(chan->s_param + QMC_SPE_TBPTR, val); 1075 1076 /* RBASE and RBPTR*/ 1077 val = ((chan->id * (QMC_NB_TXBDS + QMC_NB_RXBDS)) + QMC_NB_TXBDS) * sizeof(cbd_t); 1078 qmc_write16(chan->s_param + QMC_SPE_RBASE, val); 1079 qmc_write16(chan->s_param + QMC_SPE_RBPTR, val); 1080 qmc_write32(chan->s_param + QMC_SPE_TSTATE, 0x30000000); 1081 qmc_write32(chan->s_param + QMC_SPE_RSTATE, 0x31000000); 1082 qmc_write32(chan->s_param + QMC_SPE_ZISTATE, 0x00000100); 1083 if (chan->mode == QMC_TRANSPARENT) { 1084 qmc_write32(chan->s_param + QMC_SPE_ZDSTATE, 0x18000080); 1085 qmc_write16(chan->s_param + QMC_SPE_TMRBLR, 60); 1086 val = QMC_SPE_CHAMR_MODE_TRANSP | QMC_SPE_CHAMR_TRANSP_SYNC; 1087 if (chan->is_reverse_data) 1088 val |= QMC_SPE_CHAMR_TRANSP_RD; 1089 qmc_write16(chan->s_param + QMC_SPE_CHAMR, val); 1090 ret = qmc_setup_chan_trnsync(qmc, chan); 1091 if (ret) 1092 return ret; 1093 } else { 1094 qmc_write32(chan->s_param + QMC_SPE_ZDSTATE, 0x00000080); 1095 qmc_write16(chan->s_param + QMC_SPE_MFLR, 60); 1096 qmc_write16(chan->s_param + QMC_SPE_CHAMR, 1097 QMC_SPE_CHAMR_MODE_HDLC | QMC_SPE_CHAMR_HDLC_IDLM); 1098 } 1099 1100 /* Do not enable interrupts now. They will be enabled later */ 1101 qmc_write16(chan->s_param + QMC_SPE_INTMSK, 0x0000); 1102 1103 /* Init Rx BDs and set Wrap bit on last descriptor */ 1104 BUILD_BUG_ON(QMC_NB_RXBDS == 0); 1105 val = QMC_BD_RX_I; 1106 for (i = 0; i < QMC_NB_RXBDS; i++) { 1107 bd = chan->rxbds + i; 1108 qmc_write16(&bd->cbd_sc, val); 1109 } 1110 bd = chan->rxbds + QMC_NB_RXBDS - 1; 1111 qmc_write16(&bd->cbd_sc, val | QMC_BD_RX_W); 1112 1113 /* Init Tx BDs and set Wrap bit on last descriptor */ 1114 BUILD_BUG_ON(QMC_NB_TXBDS == 0); 1115 val = QMC_BD_TX_I; 1116 if (chan->mode == QMC_HDLC) 1117 val |= QMC_BD_TX_L | QMC_BD_TX_TC; 1118 for (i = 0; i < QMC_NB_TXBDS; i++) { 1119 bd = chan->txbds + i; 1120 qmc_write16(&bd->cbd_sc, val); 1121 } 1122 bd = chan->txbds + QMC_NB_TXBDS - 1; 1123 qmc_write16(&bd->cbd_sc, val | QMC_BD_TX_W); 1124 1125 return 0; 1126 } 1127 1128 static int qmc_setup_chans(struct qmc *qmc) 1129 { 1130 struct qmc_chan *chan; 1131 int ret; 1132 1133 list_for_each_entry(chan, &qmc->chan_head, list) { 1134 ret = qmc_setup_chan(qmc, chan); 1135 if (ret) 1136 return ret; 1137 } 1138 1139 return 0; 1140 } 1141 1142 static int qmc_finalize_chans(struct qmc *qmc) 1143 { 1144 struct qmc_chan *chan; 1145 int ret; 1146 1147 list_for_each_entry(chan, &qmc->chan_head, list) { 1148 /* Unmask channel interrupts */ 1149 if (chan->mode == QMC_HDLC) { 1150 qmc_write16(chan->s_param + QMC_SPE_INTMSK, 1151 QMC_INT_NID | QMC_INT_IDL | QMC_INT_MRF | 1152 QMC_INT_UN | QMC_INT_RXF | QMC_INT_BSY | 1153 QMC_INT_TXB | QMC_INT_RXB); 1154 } else { 1155 qmc_write16(chan->s_param + QMC_SPE_INTMSK, 1156 QMC_INT_UN | QMC_INT_BSY | 1157 QMC_INT_TXB | QMC_INT_RXB); 1158 } 1159 1160 /* Forced stop the channel */ 1161 ret = qmc_chan_stop(chan, QMC_CHAN_ALL); 1162 if (ret) 1163 return ret; 1164 } 1165 1166 return 0; 1167 } 1168 1169 static int qmc_setup_ints(struct qmc *qmc) 1170 { 1171 unsigned int i; 1172 u16 __iomem *last; 1173 1174 /* Raz all entries */ 1175 for (i = 0; i < (qmc->int_size / sizeof(u16)); i++) 1176 qmc_write16(qmc->int_table + i, 0x0000); 1177 1178 /* Set Wrap bit on last entry */ 1179 if (qmc->int_size >= sizeof(u16)) { 1180 last = qmc->int_table + (qmc->int_size / sizeof(u16)) - 1; 1181 qmc_write16(last, QMC_INT_W); 1182 } 1183 1184 return 0; 1185 } 1186 1187 static void qmc_irq_gint(struct qmc *qmc) 1188 { 1189 struct qmc_chan *chan; 1190 unsigned int chan_id; 1191 unsigned long flags; 1192 u16 int_entry; 1193 1194 int_entry = qmc_read16(qmc->int_curr); 1195 while (int_entry & QMC_INT_V) { 1196 /* Clear all but the Wrap bit */ 1197 qmc_write16(qmc->int_curr, int_entry & QMC_INT_W); 1198 1199 chan_id = QMC_INT_GET_CHANNEL(int_entry); 1200 chan = qmc->chans[chan_id]; 1201 if (!chan) { 1202 dev_err(qmc->dev, "interrupt on invalid chan %u\n", chan_id); 1203 goto int_next; 1204 } 1205 1206 if (int_entry & QMC_INT_TXB) 1207 qmc_chan_write_done(chan); 1208 1209 if (int_entry & QMC_INT_UN) { 1210 dev_info(qmc->dev, "intr chan %u, 0x%04x (UN)\n", chan_id, 1211 int_entry); 1212 chan->nb_tx_underrun++; 1213 } 1214 1215 if (int_entry & QMC_INT_BSY) { 1216 dev_info(qmc->dev, "intr chan %u, 0x%04x (BSY)\n", chan_id, 1217 int_entry); 1218 chan->nb_rx_busy++; 1219 /* Restart the receiver if needed */ 1220 spin_lock_irqsave(&chan->rx_lock, flags); 1221 if (chan->rx_pending && !chan->is_rx_stopped) { 1222 if (chan->mode == QMC_TRANSPARENT) 1223 qmc_write32(chan->s_param + QMC_SPE_ZDSTATE, 0x18000080); 1224 else 1225 qmc_write32(chan->s_param + QMC_SPE_ZDSTATE, 0x00000080); 1226 qmc_write32(chan->s_param + QMC_SPE_RSTATE, 0x31000000); 1227 chan->is_rx_halted = false; 1228 } else { 1229 chan->is_rx_halted = true; 1230 } 1231 spin_unlock_irqrestore(&chan->rx_lock, flags); 1232 } 1233 1234 if (int_entry & QMC_INT_RXB) 1235 qmc_chan_read_done(chan); 1236 1237 int_next: 1238 if (int_entry & QMC_INT_W) 1239 qmc->int_curr = qmc->int_table; 1240 else 1241 qmc->int_curr++; 1242 int_entry = qmc_read16(qmc->int_curr); 1243 } 1244 } 1245 1246 static irqreturn_t qmc_irq_handler(int irq, void *priv) 1247 { 1248 struct qmc *qmc = (struct qmc *)priv; 1249 u16 scce; 1250 1251 scce = qmc_read16(qmc->scc_regs + SCC_SCCE); 1252 qmc_write16(qmc->scc_regs + SCC_SCCE, scce); 1253 1254 if (unlikely(scce & SCC_SCCE_IQOV)) 1255 dev_info(qmc->dev, "IRQ queue overflow\n"); 1256 1257 if (unlikely(scce & SCC_SCCE_GUN)) 1258 dev_err(qmc->dev, "Global transmitter underrun\n"); 1259 1260 if (unlikely(scce & SCC_SCCE_GOV)) 1261 dev_err(qmc->dev, "Global receiver overrun\n"); 1262 1263 /* normal interrupt */ 1264 if (likely(scce & SCC_SCCE_GINT)) 1265 qmc_irq_gint(qmc); 1266 1267 return IRQ_HANDLED; 1268 } 1269 1270 static int qmc_probe(struct platform_device *pdev) 1271 { 1272 struct device_node *np = pdev->dev.of_node; 1273 unsigned int nb_chans; 1274 struct resource *res; 1275 struct qmc *qmc; 1276 int irq; 1277 int ret; 1278 1279 qmc = devm_kzalloc(&pdev->dev, sizeof(*qmc), GFP_KERNEL); 1280 if (!qmc) 1281 return -ENOMEM; 1282 1283 qmc->dev = &pdev->dev; 1284 INIT_LIST_HEAD(&qmc->chan_head); 1285 1286 qmc->scc_regs = devm_platform_ioremap_resource_byname(pdev, "scc_regs"); 1287 if (IS_ERR(qmc->scc_regs)) 1288 return PTR_ERR(qmc->scc_regs); 1289 1290 1291 res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "scc_pram"); 1292 if (!res) 1293 return -EINVAL; 1294 qmc->scc_pram_offset = res->start - get_immrbase(); 1295 qmc->scc_pram = devm_ioremap_resource(qmc->dev, res); 1296 if (IS_ERR(qmc->scc_pram)) 1297 return PTR_ERR(qmc->scc_pram); 1298 1299 qmc->dpram = devm_platform_ioremap_resource_byname(pdev, "dpram"); 1300 if (IS_ERR(qmc->dpram)) 1301 return PTR_ERR(qmc->dpram); 1302 1303 qmc->tsa_serial = devm_tsa_serial_get_byphandle(qmc->dev, np, "fsl,tsa-serial"); 1304 if (IS_ERR(qmc->tsa_serial)) { 1305 return dev_err_probe(qmc->dev, PTR_ERR(qmc->tsa_serial), 1306 "Failed to get TSA serial\n"); 1307 } 1308 1309 /* Connect the serial (SCC) to TSA */ 1310 ret = tsa_serial_connect(qmc->tsa_serial); 1311 if (ret) { 1312 dev_err(qmc->dev, "Failed to connect TSA serial\n"); 1313 return ret; 1314 } 1315 1316 /* Parse channels informationss */ 1317 ret = qmc_of_parse_chans(qmc, np); 1318 if (ret) 1319 goto err_tsa_serial_disconnect; 1320 1321 nb_chans = qmc_nb_chans(qmc); 1322 1323 /* Init GMSR_H and GMSR_L registers */ 1324 qmc_write32(qmc->scc_regs + SCC_GSMRH, 1325 SCC_GSMRH_CDS | SCC_GSMRH_CTSS | SCC_GSMRH_CDP | SCC_GSMRH_CTSP); 1326 1327 /* enable QMC mode */ 1328 qmc_write32(qmc->scc_regs + SCC_GSMRL, SCC_GSMRL_MODE_QMC); 1329 1330 /* 1331 * Allocate the buffer descriptor table 1332 * 8 rx and 8 tx descriptors per channel 1333 */ 1334 qmc->bd_size = (nb_chans * (QMC_NB_TXBDS + QMC_NB_RXBDS)) * sizeof(cbd_t); 1335 qmc->bd_table = dmam_alloc_coherent(qmc->dev, qmc->bd_size, 1336 &qmc->bd_dma_addr, GFP_KERNEL); 1337 if (!qmc->bd_table) { 1338 dev_err(qmc->dev, "Failed to allocate bd table\n"); 1339 ret = -ENOMEM; 1340 goto err_tsa_serial_disconnect; 1341 } 1342 memset(qmc->bd_table, 0, qmc->bd_size); 1343 1344 qmc_write32(qmc->scc_pram + QMC_GBL_MCBASE, qmc->bd_dma_addr); 1345 1346 /* Allocate the interrupt table */ 1347 qmc->int_size = QMC_NB_INTS * sizeof(u16); 1348 qmc->int_table = dmam_alloc_coherent(qmc->dev, qmc->int_size, 1349 &qmc->int_dma_addr, GFP_KERNEL); 1350 if (!qmc->int_table) { 1351 dev_err(qmc->dev, "Failed to allocate interrupt table\n"); 1352 ret = -ENOMEM; 1353 goto err_tsa_serial_disconnect; 1354 } 1355 memset(qmc->int_table, 0, qmc->int_size); 1356 1357 qmc->int_curr = qmc->int_table; 1358 qmc_write32(qmc->scc_pram + QMC_GBL_INTBASE, qmc->int_dma_addr); 1359 qmc_write32(qmc->scc_pram + QMC_GBL_INTPTR, qmc->int_dma_addr); 1360 1361 /* Set MRBLR (valid for HDLC only) max MRU + max CRC */ 1362 qmc_write16(qmc->scc_pram + QMC_GBL_MRBLR, HDLC_MAX_MRU + 4); 1363 1364 qmc_write16(qmc->scc_pram + QMC_GBL_GRFTHR, 1); 1365 qmc_write16(qmc->scc_pram + QMC_GBL_GRFCNT, 1); 1366 1367 qmc_write32(qmc->scc_pram + QMC_GBL_C_MASK32, 0xDEBB20E3); 1368 qmc_write16(qmc->scc_pram + QMC_GBL_C_MASK16, 0xF0B8); 1369 1370 ret = qmc_setup_tsa(qmc); 1371 if (ret) 1372 goto err_tsa_serial_disconnect; 1373 1374 qmc_write16(qmc->scc_pram + QMC_GBL_QMCSTATE, 0x8000); 1375 1376 ret = qmc_setup_chans(qmc); 1377 if (ret) 1378 goto err_tsa_serial_disconnect; 1379 1380 /* Init interrupts table */ 1381 ret = qmc_setup_ints(qmc); 1382 if (ret) 1383 goto err_tsa_serial_disconnect; 1384 1385 /* Disable and clear interrupts, set the irq handler */ 1386 qmc_write16(qmc->scc_regs + SCC_SCCM, 0x0000); 1387 qmc_write16(qmc->scc_regs + SCC_SCCE, 0x000F); 1388 irq = platform_get_irq(pdev, 0); 1389 if (irq < 0) 1390 goto err_tsa_serial_disconnect; 1391 ret = devm_request_irq(qmc->dev, irq, qmc_irq_handler, 0, "qmc", qmc); 1392 if (ret < 0) 1393 goto err_tsa_serial_disconnect; 1394 1395 /* Enable interrupts */ 1396 qmc_write16(qmc->scc_regs + SCC_SCCM, 1397 SCC_SCCE_IQOV | SCC_SCCE_GINT | SCC_SCCE_GUN | SCC_SCCE_GOV); 1398 1399 ret = qmc_finalize_chans(qmc); 1400 if (ret < 0) 1401 goto err_disable_intr; 1402 1403 /* Enable transmiter and receiver */ 1404 qmc_setbits32(qmc->scc_regs + SCC_GSMRL, SCC_GSMRL_ENR | SCC_GSMRL_ENT); 1405 1406 platform_set_drvdata(pdev, qmc); 1407 1408 return 0; 1409 1410 err_disable_intr: 1411 qmc_write16(qmc->scc_regs + SCC_SCCM, 0); 1412 1413 err_tsa_serial_disconnect: 1414 tsa_serial_disconnect(qmc->tsa_serial); 1415 return ret; 1416 } 1417 1418 static void qmc_remove(struct platform_device *pdev) 1419 { 1420 struct qmc *qmc = platform_get_drvdata(pdev); 1421 1422 /* Disable transmiter and receiver */ 1423 qmc_setbits32(qmc->scc_regs + SCC_GSMRL, 0); 1424 1425 /* Disable interrupts */ 1426 qmc_write16(qmc->scc_regs + SCC_SCCM, 0); 1427 1428 /* Disconnect the serial from TSA */ 1429 tsa_serial_disconnect(qmc->tsa_serial); 1430 } 1431 1432 static const struct of_device_id qmc_id_table[] = { 1433 { .compatible = "fsl,cpm1-scc-qmc" }, 1434 {} /* sentinel */ 1435 }; 1436 MODULE_DEVICE_TABLE(of, qmc_id_table); 1437 1438 static struct platform_driver qmc_driver = { 1439 .driver = { 1440 .name = "fsl-qmc", 1441 .of_match_table = of_match_ptr(qmc_id_table), 1442 }, 1443 .probe = qmc_probe, 1444 .remove_new = qmc_remove, 1445 }; 1446 module_platform_driver(qmc_driver); 1447 1448 struct qmc_chan *qmc_chan_get_byphandle(struct device_node *np, const char *phandle_name) 1449 { 1450 struct of_phandle_args out_args; 1451 struct platform_device *pdev; 1452 struct qmc_chan *qmc_chan; 1453 struct qmc *qmc; 1454 int ret; 1455 1456 ret = of_parse_phandle_with_fixed_args(np, phandle_name, 1, 0, 1457 &out_args); 1458 if (ret < 0) 1459 return ERR_PTR(ret); 1460 1461 if (!of_match_node(qmc_driver.driver.of_match_table, out_args.np)) { 1462 of_node_put(out_args.np); 1463 return ERR_PTR(-EINVAL); 1464 } 1465 1466 pdev = of_find_device_by_node(out_args.np); 1467 of_node_put(out_args.np); 1468 if (!pdev) 1469 return ERR_PTR(-ENODEV); 1470 1471 qmc = platform_get_drvdata(pdev); 1472 if (!qmc) { 1473 platform_device_put(pdev); 1474 return ERR_PTR(-EPROBE_DEFER); 1475 } 1476 1477 if (out_args.args_count != 1) { 1478 platform_device_put(pdev); 1479 return ERR_PTR(-EINVAL); 1480 } 1481 1482 if (out_args.args[0] >= ARRAY_SIZE(qmc->chans)) { 1483 platform_device_put(pdev); 1484 return ERR_PTR(-EINVAL); 1485 } 1486 1487 qmc_chan = qmc->chans[out_args.args[0]]; 1488 if (!qmc_chan) { 1489 platform_device_put(pdev); 1490 return ERR_PTR(-ENOENT); 1491 } 1492 1493 return qmc_chan; 1494 } 1495 EXPORT_SYMBOL(qmc_chan_get_byphandle); 1496 1497 void qmc_chan_put(struct qmc_chan *chan) 1498 { 1499 put_device(chan->qmc->dev); 1500 } 1501 EXPORT_SYMBOL(qmc_chan_put); 1502 1503 static void devm_qmc_chan_release(struct device *dev, void *res) 1504 { 1505 struct qmc_chan **qmc_chan = res; 1506 1507 qmc_chan_put(*qmc_chan); 1508 } 1509 1510 struct qmc_chan *devm_qmc_chan_get_byphandle(struct device *dev, 1511 struct device_node *np, 1512 const char *phandle_name) 1513 { 1514 struct qmc_chan *qmc_chan; 1515 struct qmc_chan **dr; 1516 1517 dr = devres_alloc(devm_qmc_chan_release, sizeof(*dr), GFP_KERNEL); 1518 if (!dr) 1519 return ERR_PTR(-ENOMEM); 1520 1521 qmc_chan = qmc_chan_get_byphandle(np, phandle_name); 1522 if (!IS_ERR(qmc_chan)) { 1523 *dr = qmc_chan; 1524 devres_add(dev, dr); 1525 } else { 1526 devres_free(dr); 1527 } 1528 1529 return qmc_chan; 1530 } 1531 EXPORT_SYMBOL(devm_qmc_chan_get_byphandle); 1532 1533 MODULE_AUTHOR("Herve Codina <herve.codina@bootlin.com>"); 1534 MODULE_DESCRIPTION("CPM QMC driver"); 1535 MODULE_LICENSE("GPL"); 1536