1 /* 2 * Renesas USB driver 3 * 4 * Copyright (C) 2011 Renesas Solutions Corp. 5 * Kuninori Morimoto <kuninori.morimoto.gx@renesas.com> 6 * 7 * This program is distributed in the hope that it will be useful, 8 * but WITHOUT ANY WARRANTY; without even the implied warranty of 9 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 10 * GNU General Public License for more details. 11 * 12 * You should have received a copy of the GNU General Public License 13 * along with this program; if not, write to the Free Software 14 * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA 15 * 16 */ 17 #include <linux/delay.h> 18 #include <linux/io.h> 19 #include <linux/scatterlist.h> 20 #include "common.h" 21 #include "pipe.h" 22 23 #define usbhsf_get_cfifo(p) (&((p)->fifo_info.cfifo)) 24 #define usbhsf_get_d0fifo(p) (&((p)->fifo_info.d0fifo)) 25 #define usbhsf_get_d1fifo(p) (&((p)->fifo_info.d1fifo)) 26 #define usbhsf_is_cfifo(p, f) (usbhsf_get_cfifo(p) == f) 27 28 #define usbhsf_fifo_is_busy(f) ((f)->pipe) /* see usbhs_pipe_select_fifo */ 29 30 /* 31 * packet initialize 32 */ 33 void usbhs_pkt_init(struct usbhs_pkt *pkt) 34 { 35 INIT_LIST_HEAD(&pkt->node); 36 } 37 38 /* 39 * packet control function 40 */ 41 static int usbhsf_null_handle(struct usbhs_pkt *pkt, int *is_done) 42 { 43 struct usbhs_priv *priv = usbhs_pipe_to_priv(pkt->pipe); 44 struct device *dev = usbhs_priv_to_dev(priv); 45 46 dev_err(dev, "null handler\n"); 47 48 return -EINVAL; 49 } 50 51 static struct usbhs_pkt_handle usbhsf_null_handler = { 52 .prepare = usbhsf_null_handle, 53 .try_run = usbhsf_null_handle, 54 }; 55 56 void usbhs_pkt_push(struct usbhs_pipe *pipe, struct usbhs_pkt *pkt, 57 void (*done)(struct usbhs_priv *priv, 58 struct usbhs_pkt *pkt), 59 void *buf, int len, int zero, int sequence) 60 { 61 struct usbhs_priv *priv = usbhs_pipe_to_priv(pipe); 62 struct device *dev = usbhs_priv_to_dev(priv); 63 unsigned long flags; 64 65 if (!done) { 66 dev_err(dev, "no done function\n"); 67 return; 68 } 69 70 /******************** spin lock ********************/ 71 usbhs_lock(priv, flags); 72 73 if (!pipe->handler) { 74 dev_err(dev, "no handler function\n"); 75 pipe->handler = &usbhsf_null_handler; 76 } 77 78 list_move_tail(&pkt->node, &pipe->list); 79 80 /* 81 * each pkt must hold own handler. 82 * because handler might be changed by its situation. 83 * dma handler -> pio handler. 84 */ 85 pkt->pipe = pipe; 86 pkt->buf = buf; 87 pkt->handler = pipe->handler; 88 pkt->length = len; 89 pkt->zero = zero; 90 pkt->actual = 0; 91 pkt->done = done; 92 pkt->sequence = sequence; 93 94 usbhs_unlock(priv, flags); 95 /******************** spin unlock ******************/ 96 } 97 98 static void __usbhsf_pkt_del(struct usbhs_pkt *pkt) 99 { 100 list_del_init(&pkt->node); 101 } 102 103 static struct usbhs_pkt *__usbhsf_pkt_get(struct usbhs_pipe *pipe) 104 { 105 if (list_empty(&pipe->list)) 106 return NULL; 107 108 return list_first_entry(&pipe->list, struct usbhs_pkt, node); 109 } 110 111 struct usbhs_pkt *usbhs_pkt_pop(struct usbhs_pipe *pipe, struct usbhs_pkt *pkt) 112 { 113 struct usbhs_priv *priv = usbhs_pipe_to_priv(pipe); 114 unsigned long flags; 115 116 /******************** spin lock ********************/ 117 usbhs_lock(priv, flags); 118 119 if (!pkt) 120 pkt = __usbhsf_pkt_get(pipe); 121 122 if (pkt) 123 __usbhsf_pkt_del(pkt); 124 125 usbhs_unlock(priv, flags); 126 /******************** spin unlock ******************/ 127 128 return pkt; 129 } 130 131 enum { 132 USBHSF_PKT_PREPARE, 133 USBHSF_PKT_TRY_RUN, 134 USBHSF_PKT_DMA_DONE, 135 }; 136 137 static int usbhsf_pkt_handler(struct usbhs_pipe *pipe, int type) 138 { 139 struct usbhs_priv *priv = usbhs_pipe_to_priv(pipe); 140 struct usbhs_pkt *pkt; 141 struct device *dev = usbhs_priv_to_dev(priv); 142 int (*func)(struct usbhs_pkt *pkt, int *is_done); 143 unsigned long flags; 144 int ret = 0; 145 int is_done = 0; 146 147 /******************** spin lock ********************/ 148 usbhs_lock(priv, flags); 149 150 pkt = __usbhsf_pkt_get(pipe); 151 if (!pkt) 152 goto __usbhs_pkt_handler_end; 153 154 switch (type) { 155 case USBHSF_PKT_PREPARE: 156 func = pkt->handler->prepare; 157 break; 158 case USBHSF_PKT_TRY_RUN: 159 func = pkt->handler->try_run; 160 break; 161 case USBHSF_PKT_DMA_DONE: 162 func = pkt->handler->dma_done; 163 break; 164 default: 165 dev_err(dev, "unknown pkt handler\n"); 166 goto __usbhs_pkt_handler_end; 167 } 168 169 ret = func(pkt, &is_done); 170 171 if (is_done) 172 __usbhsf_pkt_del(pkt); 173 174 __usbhs_pkt_handler_end: 175 usbhs_unlock(priv, flags); 176 /******************** spin unlock ******************/ 177 178 if (is_done) { 179 pkt->done(priv, pkt); 180 usbhs_pkt_start(pipe); 181 } 182 183 return ret; 184 } 185 186 void usbhs_pkt_start(struct usbhs_pipe *pipe) 187 { 188 usbhsf_pkt_handler(pipe, USBHSF_PKT_PREPARE); 189 } 190 191 /* 192 * irq enable/disable function 193 */ 194 #define usbhsf_irq_empty_ctrl(p, e) usbhsf_irq_callback_ctrl(p, irq_bempsts, e) 195 #define usbhsf_irq_ready_ctrl(p, e) usbhsf_irq_callback_ctrl(p, irq_brdysts, e) 196 #define usbhsf_irq_callback_ctrl(pipe, status, enable) \ 197 ({ \ 198 struct usbhs_priv *priv = usbhs_pipe_to_priv(pipe); \ 199 struct usbhs_mod *mod = usbhs_mod_get_current(priv); \ 200 u16 status = (1 << usbhs_pipe_number(pipe)); \ 201 if (!mod) \ 202 return; \ 203 if (enable) \ 204 mod->status |= status; \ 205 else \ 206 mod->status &= ~status; \ 207 usbhs_irq_callback_update(priv, mod); \ 208 }) 209 210 static void usbhsf_tx_irq_ctrl(struct usbhs_pipe *pipe, int enable) 211 { 212 /* 213 * And DCP pipe can NOT use "ready interrupt" for "send" 214 * it should use "empty" interrupt. 215 * see 216 * "Operation" - "Interrupt Function" - "BRDY Interrupt" 217 * 218 * on the other hand, normal pipe can use "ready interrupt" for "send" 219 * even though it is single/double buffer 220 */ 221 if (usbhs_pipe_is_dcp(pipe)) 222 usbhsf_irq_empty_ctrl(pipe, enable); 223 else 224 usbhsf_irq_ready_ctrl(pipe, enable); 225 } 226 227 static void usbhsf_rx_irq_ctrl(struct usbhs_pipe *pipe, int enable) 228 { 229 usbhsf_irq_ready_ctrl(pipe, enable); 230 } 231 232 /* 233 * FIFO ctrl 234 */ 235 static void usbhsf_send_terminator(struct usbhs_pipe *pipe, 236 struct usbhs_fifo *fifo) 237 { 238 struct usbhs_priv *priv = usbhs_pipe_to_priv(pipe); 239 240 usbhs_bset(priv, fifo->ctr, BVAL, BVAL); 241 } 242 243 static int usbhsf_fifo_barrier(struct usbhs_priv *priv, 244 struct usbhs_fifo *fifo) 245 { 246 int timeout = 1024; 247 248 do { 249 /* The FIFO port is accessible */ 250 if (usbhs_read(priv, fifo->ctr) & FRDY) 251 return 0; 252 253 udelay(10); 254 } while (timeout--); 255 256 return -EBUSY; 257 } 258 259 static void usbhsf_fifo_clear(struct usbhs_pipe *pipe, 260 struct usbhs_fifo *fifo) 261 { 262 struct usbhs_priv *priv = usbhs_pipe_to_priv(pipe); 263 264 if (!usbhs_pipe_is_dcp(pipe)) 265 usbhsf_fifo_barrier(priv, fifo); 266 267 usbhs_write(priv, fifo->ctr, BCLR); 268 } 269 270 static int usbhsf_fifo_rcv_len(struct usbhs_priv *priv, 271 struct usbhs_fifo *fifo) 272 { 273 return usbhs_read(priv, fifo->ctr) & DTLN_MASK; 274 } 275 276 static void usbhsf_fifo_unselect(struct usbhs_pipe *pipe, 277 struct usbhs_fifo *fifo) 278 { 279 struct usbhs_priv *priv = usbhs_pipe_to_priv(pipe); 280 281 usbhs_pipe_select_fifo(pipe, NULL); 282 usbhs_write(priv, fifo->sel, 0); 283 } 284 285 static int usbhsf_fifo_select(struct usbhs_pipe *pipe, 286 struct usbhs_fifo *fifo, 287 int write) 288 { 289 struct usbhs_priv *priv = usbhs_pipe_to_priv(pipe); 290 struct device *dev = usbhs_priv_to_dev(priv); 291 int timeout = 1024; 292 u16 mask = ((1 << 5) | 0xF); /* mask of ISEL | CURPIPE */ 293 u16 base = usbhs_pipe_number(pipe); /* CURPIPE */ 294 295 if (usbhs_pipe_is_busy(pipe) || 296 usbhsf_fifo_is_busy(fifo)) 297 return -EBUSY; 298 299 if (usbhs_pipe_is_dcp(pipe)) { 300 base |= (1 == write) << 5; /* ISEL */ 301 302 if (usbhs_mod_is_host(priv)) 303 usbhs_dcp_dir_for_host(pipe, write); 304 } 305 306 /* "base" will be used below */ 307 if (usbhs_get_dparam(priv, has_sudmac) && !usbhsf_is_cfifo(priv, fifo)) 308 usbhs_write(priv, fifo->sel, base); 309 else 310 usbhs_write(priv, fifo->sel, base | MBW_32); 311 312 /* check ISEL and CURPIPE value */ 313 while (timeout--) { 314 if (base == (mask & usbhs_read(priv, fifo->sel))) { 315 usbhs_pipe_select_fifo(pipe, fifo); 316 return 0; 317 } 318 udelay(10); 319 } 320 321 dev_err(dev, "fifo select error\n"); 322 323 return -EIO; 324 } 325 326 /* 327 * DCP status stage 328 */ 329 static int usbhs_dcp_dir_switch_to_write(struct usbhs_pkt *pkt, int *is_done) 330 { 331 struct usbhs_pipe *pipe = pkt->pipe; 332 struct usbhs_priv *priv = usbhs_pipe_to_priv(pipe); 333 struct usbhs_fifo *fifo = usbhsf_get_cfifo(priv); /* CFIFO */ 334 struct device *dev = usbhs_priv_to_dev(priv); 335 int ret; 336 337 usbhs_pipe_disable(pipe); 338 339 ret = usbhsf_fifo_select(pipe, fifo, 1); 340 if (ret < 0) { 341 dev_err(dev, "%s() faile\n", __func__); 342 return ret; 343 } 344 345 usbhs_pipe_sequence_data1(pipe); /* DATA1 */ 346 347 usbhsf_fifo_clear(pipe, fifo); 348 usbhsf_send_terminator(pipe, fifo); 349 350 usbhsf_fifo_unselect(pipe, fifo); 351 352 usbhsf_tx_irq_ctrl(pipe, 1); 353 usbhs_pipe_enable(pipe); 354 355 return ret; 356 } 357 358 static int usbhs_dcp_dir_switch_to_read(struct usbhs_pkt *pkt, int *is_done) 359 { 360 struct usbhs_pipe *pipe = pkt->pipe; 361 struct usbhs_priv *priv = usbhs_pipe_to_priv(pipe); 362 struct usbhs_fifo *fifo = usbhsf_get_cfifo(priv); /* CFIFO */ 363 struct device *dev = usbhs_priv_to_dev(priv); 364 int ret; 365 366 usbhs_pipe_disable(pipe); 367 368 ret = usbhsf_fifo_select(pipe, fifo, 0); 369 if (ret < 0) { 370 dev_err(dev, "%s() fail\n", __func__); 371 return ret; 372 } 373 374 usbhs_pipe_sequence_data1(pipe); /* DATA1 */ 375 usbhsf_fifo_clear(pipe, fifo); 376 377 usbhsf_fifo_unselect(pipe, fifo); 378 379 usbhsf_rx_irq_ctrl(pipe, 1); 380 usbhs_pipe_enable(pipe); 381 382 return ret; 383 384 } 385 386 static int usbhs_dcp_dir_switch_done(struct usbhs_pkt *pkt, int *is_done) 387 { 388 struct usbhs_pipe *pipe = pkt->pipe; 389 390 if (pkt->handler == &usbhs_dcp_status_stage_in_handler) 391 usbhsf_tx_irq_ctrl(pipe, 0); 392 else 393 usbhsf_rx_irq_ctrl(pipe, 0); 394 395 pkt->actual = pkt->length; 396 *is_done = 1; 397 398 return 0; 399 } 400 401 struct usbhs_pkt_handle usbhs_dcp_status_stage_in_handler = { 402 .prepare = usbhs_dcp_dir_switch_to_write, 403 .try_run = usbhs_dcp_dir_switch_done, 404 }; 405 406 struct usbhs_pkt_handle usbhs_dcp_status_stage_out_handler = { 407 .prepare = usbhs_dcp_dir_switch_to_read, 408 .try_run = usbhs_dcp_dir_switch_done, 409 }; 410 411 /* 412 * DCP data stage (push) 413 */ 414 static int usbhsf_dcp_data_stage_try_push(struct usbhs_pkt *pkt, int *is_done) 415 { 416 struct usbhs_pipe *pipe = pkt->pipe; 417 418 usbhs_pipe_sequence_data1(pipe); /* DATA1 */ 419 420 /* 421 * change handler to PIO push 422 */ 423 pkt->handler = &usbhs_fifo_pio_push_handler; 424 425 return pkt->handler->prepare(pkt, is_done); 426 } 427 428 struct usbhs_pkt_handle usbhs_dcp_data_stage_out_handler = { 429 .prepare = usbhsf_dcp_data_stage_try_push, 430 }; 431 432 /* 433 * DCP data stage (pop) 434 */ 435 static int usbhsf_dcp_data_stage_prepare_pop(struct usbhs_pkt *pkt, 436 int *is_done) 437 { 438 struct usbhs_pipe *pipe = pkt->pipe; 439 struct usbhs_priv *priv = usbhs_pipe_to_priv(pipe); 440 struct usbhs_fifo *fifo = usbhsf_get_cfifo(priv); 441 442 if (usbhs_pipe_is_busy(pipe)) 443 return 0; 444 445 /* 446 * prepare pop for DCP should 447 * - change DCP direction, 448 * - clear fifo 449 * - DATA1 450 */ 451 usbhs_pipe_disable(pipe); 452 453 usbhs_pipe_sequence_data1(pipe); /* DATA1 */ 454 455 usbhsf_fifo_select(pipe, fifo, 0); 456 usbhsf_fifo_clear(pipe, fifo); 457 usbhsf_fifo_unselect(pipe, fifo); 458 459 /* 460 * change handler to PIO pop 461 */ 462 pkt->handler = &usbhs_fifo_pio_pop_handler; 463 464 return pkt->handler->prepare(pkt, is_done); 465 } 466 467 struct usbhs_pkt_handle usbhs_dcp_data_stage_in_handler = { 468 .prepare = usbhsf_dcp_data_stage_prepare_pop, 469 }; 470 471 /* 472 * PIO push handler 473 */ 474 static int usbhsf_pio_try_push(struct usbhs_pkt *pkt, int *is_done) 475 { 476 struct usbhs_pipe *pipe = pkt->pipe; 477 struct usbhs_priv *priv = usbhs_pipe_to_priv(pipe); 478 struct device *dev = usbhs_priv_to_dev(priv); 479 struct usbhs_fifo *fifo = usbhsf_get_cfifo(priv); /* CFIFO */ 480 void __iomem *addr = priv->base + fifo->port; 481 u8 *buf; 482 int maxp = usbhs_pipe_get_maxpacket(pipe); 483 int total_len; 484 int i, ret, len; 485 int is_short; 486 487 usbhs_pipe_data_sequence(pipe, pkt->sequence); 488 pkt->sequence = -1; /* -1 sequence will be ignored */ 489 490 usbhs_pipe_set_trans_count_if_bulk(pipe, pkt->length); 491 492 ret = usbhsf_fifo_select(pipe, fifo, 1); 493 if (ret < 0) 494 return 0; 495 496 ret = usbhs_pipe_is_accessible(pipe); 497 if (ret < 0) { 498 /* inaccessible pipe is not an error */ 499 ret = 0; 500 goto usbhs_fifo_write_busy; 501 } 502 503 ret = usbhsf_fifo_barrier(priv, fifo); 504 if (ret < 0) 505 goto usbhs_fifo_write_busy; 506 507 buf = pkt->buf + pkt->actual; 508 len = pkt->length - pkt->actual; 509 len = min(len, maxp); 510 total_len = len; 511 is_short = total_len < maxp; 512 513 /* 514 * FIXME 515 * 516 * 32-bit access only 517 */ 518 if (len >= 4 && !((unsigned long)buf & 0x03)) { 519 iowrite32_rep(addr, buf, len / 4); 520 len %= 4; 521 buf += total_len - len; 522 } 523 524 /* the rest operation */ 525 for (i = 0; i < len; i++) 526 iowrite8(buf[i], addr + (0x03 - (i & 0x03))); 527 528 /* 529 * variable update 530 */ 531 pkt->actual += total_len; 532 533 if (pkt->actual < pkt->length) 534 *is_done = 0; /* there are remainder data */ 535 else if (is_short) 536 *is_done = 1; /* short packet */ 537 else 538 *is_done = !pkt->zero; /* send zero packet ? */ 539 540 /* 541 * pipe/irq handling 542 */ 543 if (is_short) 544 usbhsf_send_terminator(pipe, fifo); 545 546 usbhsf_tx_irq_ctrl(pipe, !*is_done); 547 usbhs_pipe_enable(pipe); 548 549 dev_dbg(dev, " send %d (%d/ %d/ %d/ %d)\n", 550 usbhs_pipe_number(pipe), 551 pkt->length, pkt->actual, *is_done, pkt->zero); 552 553 /* 554 * Transmission end 555 */ 556 if (*is_done) { 557 if (usbhs_pipe_is_dcp(pipe)) 558 usbhs_dcp_control_transfer_done(pipe); 559 } 560 561 usbhsf_fifo_unselect(pipe, fifo); 562 563 return 0; 564 565 usbhs_fifo_write_busy: 566 usbhsf_fifo_unselect(pipe, fifo); 567 568 /* 569 * pipe is busy. 570 * retry in interrupt 571 */ 572 usbhsf_tx_irq_ctrl(pipe, 1); 573 574 return ret; 575 } 576 577 struct usbhs_pkt_handle usbhs_fifo_pio_push_handler = { 578 .prepare = usbhsf_pio_try_push, 579 .try_run = usbhsf_pio_try_push, 580 }; 581 582 /* 583 * PIO pop handler 584 */ 585 static int usbhsf_prepare_pop(struct usbhs_pkt *pkt, int *is_done) 586 { 587 struct usbhs_pipe *pipe = pkt->pipe; 588 589 if (usbhs_pipe_is_busy(pipe)) 590 return 0; 591 592 /* 593 * pipe enable to prepare packet receive 594 */ 595 usbhs_pipe_data_sequence(pipe, pkt->sequence); 596 pkt->sequence = -1; /* -1 sequence will be ignored */ 597 598 usbhs_pipe_set_trans_count_if_bulk(pipe, pkt->length); 599 usbhs_pipe_enable(pipe); 600 usbhsf_rx_irq_ctrl(pipe, 1); 601 602 return 0; 603 } 604 605 static int usbhsf_pio_try_pop(struct usbhs_pkt *pkt, int *is_done) 606 { 607 struct usbhs_pipe *pipe = pkt->pipe; 608 struct usbhs_priv *priv = usbhs_pipe_to_priv(pipe); 609 struct device *dev = usbhs_priv_to_dev(priv); 610 struct usbhs_fifo *fifo = usbhsf_get_cfifo(priv); /* CFIFO */ 611 void __iomem *addr = priv->base + fifo->port; 612 u8 *buf; 613 u32 data = 0; 614 int maxp = usbhs_pipe_get_maxpacket(pipe); 615 int rcv_len, len; 616 int i, ret; 617 int total_len = 0; 618 619 ret = usbhsf_fifo_select(pipe, fifo, 0); 620 if (ret < 0) 621 return 0; 622 623 ret = usbhsf_fifo_barrier(priv, fifo); 624 if (ret < 0) 625 goto usbhs_fifo_read_busy; 626 627 rcv_len = usbhsf_fifo_rcv_len(priv, fifo); 628 629 buf = pkt->buf + pkt->actual; 630 len = pkt->length - pkt->actual; 631 len = min(len, rcv_len); 632 total_len = len; 633 634 /* 635 * update actual length first here to decide disable pipe. 636 * if this pipe keeps BUF status and all data were popped, 637 * then, next interrupt/token will be issued again 638 */ 639 pkt->actual += total_len; 640 641 if ((pkt->actual == pkt->length) || /* receive all data */ 642 (total_len < maxp)) { /* short packet */ 643 *is_done = 1; 644 usbhsf_rx_irq_ctrl(pipe, 0); 645 usbhs_pipe_disable(pipe); /* disable pipe first */ 646 } 647 648 /* 649 * Buffer clear if Zero-Length packet 650 * 651 * see 652 * "Operation" - "FIFO Buffer Memory" - "FIFO Port Function" 653 */ 654 if (0 == rcv_len) { 655 pkt->zero = 1; 656 usbhsf_fifo_clear(pipe, fifo); 657 goto usbhs_fifo_read_end; 658 } 659 660 /* 661 * FIXME 662 * 663 * 32-bit access only 664 */ 665 if (len >= 4 && !((unsigned long)buf & 0x03)) { 666 ioread32_rep(addr, buf, len / 4); 667 len %= 4; 668 buf += total_len - len; 669 } 670 671 /* the rest operation */ 672 for (i = 0; i < len; i++) { 673 if (!(i & 0x03)) 674 data = ioread32(addr); 675 676 buf[i] = (data >> ((i & 0x03) * 8)) & 0xff; 677 } 678 679 usbhs_fifo_read_end: 680 dev_dbg(dev, " recv %d (%d/ %d/ %d/ %d)\n", 681 usbhs_pipe_number(pipe), 682 pkt->length, pkt->actual, *is_done, pkt->zero); 683 684 /* 685 * Transmission end 686 */ 687 if (*is_done) { 688 if (usbhs_pipe_is_dcp(pipe)) 689 usbhs_dcp_control_transfer_done(pipe); 690 } 691 692 usbhs_fifo_read_busy: 693 usbhsf_fifo_unselect(pipe, fifo); 694 695 return ret; 696 } 697 698 struct usbhs_pkt_handle usbhs_fifo_pio_pop_handler = { 699 .prepare = usbhsf_prepare_pop, 700 .try_run = usbhsf_pio_try_pop, 701 }; 702 703 /* 704 * DCP ctrol statge handler 705 */ 706 static int usbhsf_ctrl_stage_end(struct usbhs_pkt *pkt, int *is_done) 707 { 708 usbhs_dcp_control_transfer_done(pkt->pipe); 709 710 *is_done = 1; 711 712 return 0; 713 } 714 715 struct usbhs_pkt_handle usbhs_ctrl_stage_end_handler = { 716 .prepare = usbhsf_ctrl_stage_end, 717 .try_run = usbhsf_ctrl_stage_end, 718 }; 719 720 /* 721 * DMA fifo functions 722 */ 723 static struct dma_chan *usbhsf_dma_chan_get(struct usbhs_fifo *fifo, 724 struct usbhs_pkt *pkt) 725 { 726 if (&usbhs_fifo_dma_push_handler == pkt->handler) 727 return fifo->tx_chan; 728 729 if (&usbhs_fifo_dma_pop_handler == pkt->handler) 730 return fifo->rx_chan; 731 732 return NULL; 733 } 734 735 static struct usbhs_fifo *usbhsf_get_dma_fifo(struct usbhs_priv *priv, 736 struct usbhs_pkt *pkt) 737 { 738 struct usbhs_fifo *fifo; 739 740 /* DMA :: D0FIFO */ 741 fifo = usbhsf_get_d0fifo(priv); 742 if (usbhsf_dma_chan_get(fifo, pkt) && 743 !usbhsf_fifo_is_busy(fifo)) 744 return fifo; 745 746 /* DMA :: D1FIFO */ 747 fifo = usbhsf_get_d1fifo(priv); 748 if (usbhsf_dma_chan_get(fifo, pkt) && 749 !usbhsf_fifo_is_busy(fifo)) 750 return fifo; 751 752 return NULL; 753 } 754 755 #define usbhsf_dma_start(p, f) __usbhsf_dma_ctrl(p, f, DREQE) 756 #define usbhsf_dma_stop(p, f) __usbhsf_dma_ctrl(p, f, 0) 757 static void __usbhsf_dma_ctrl(struct usbhs_pipe *pipe, 758 struct usbhs_fifo *fifo, 759 u16 dreqe) 760 { 761 struct usbhs_priv *priv = usbhs_pipe_to_priv(pipe); 762 763 usbhs_bset(priv, fifo->sel, DREQE, dreqe); 764 } 765 766 #define usbhsf_dma_map(p) __usbhsf_dma_map_ctrl(p, 1) 767 #define usbhsf_dma_unmap(p) __usbhsf_dma_map_ctrl(p, 0) 768 static int __usbhsf_dma_map_ctrl(struct usbhs_pkt *pkt, int map) 769 { 770 struct usbhs_pipe *pipe = pkt->pipe; 771 struct usbhs_priv *priv = usbhs_pipe_to_priv(pipe); 772 struct usbhs_pipe_info *info = usbhs_priv_to_pipeinfo(priv); 773 774 return info->dma_map_ctrl(pkt, map); 775 } 776 777 static void usbhsf_dma_complete(void *arg); 778 static void xfer_work(struct work_struct *work) 779 { 780 struct usbhs_pkt *pkt = container_of(work, struct usbhs_pkt, work); 781 struct usbhs_pipe *pipe = pkt->pipe; 782 struct usbhs_fifo *fifo = usbhs_pipe_to_fifo(pipe); 783 struct usbhs_priv *priv = usbhs_pipe_to_priv(pipe); 784 struct dma_async_tx_descriptor *desc; 785 struct dma_chan *chan = usbhsf_dma_chan_get(fifo, pkt); 786 struct device *dev = usbhs_priv_to_dev(priv); 787 enum dma_transfer_direction dir; 788 789 dir = usbhs_pipe_is_dir_in(pipe) ? DMA_DEV_TO_MEM : DMA_MEM_TO_DEV; 790 791 desc = dmaengine_prep_slave_single(chan, pkt->dma + pkt->actual, 792 pkt->trans, dir, 793 DMA_PREP_INTERRUPT | DMA_CTRL_ACK); 794 if (!desc) 795 return; 796 797 desc->callback = usbhsf_dma_complete; 798 desc->callback_param = pipe; 799 800 if (dmaengine_submit(desc) < 0) { 801 dev_err(dev, "Failed to submit dma descriptor\n"); 802 return; 803 } 804 805 dev_dbg(dev, " %s %d (%d/ %d)\n", 806 fifo->name, usbhs_pipe_number(pipe), pkt->length, pkt->zero); 807 808 usbhs_pipe_set_trans_count_if_bulk(pipe, pkt->trans); 809 usbhs_pipe_enable(pipe); 810 usbhsf_dma_start(pipe, fifo); 811 dma_async_issue_pending(chan); 812 } 813 814 /* 815 * DMA push handler 816 */ 817 static int usbhsf_dma_prepare_push(struct usbhs_pkt *pkt, int *is_done) 818 { 819 struct usbhs_pipe *pipe = pkt->pipe; 820 struct usbhs_priv *priv = usbhs_pipe_to_priv(pipe); 821 struct usbhs_fifo *fifo; 822 int len = pkt->length - pkt->actual; 823 int ret; 824 825 if (usbhs_pipe_is_busy(pipe)) 826 return 0; 827 828 /* use PIO if packet is less than pio_dma_border or pipe is DCP */ 829 if ((len < usbhs_get_dparam(priv, pio_dma_border)) || 830 usbhs_pipe_is_dcp(pipe)) 831 goto usbhsf_pio_prepare_push; 832 833 if (len & 0x7) /* 8byte alignment */ 834 goto usbhsf_pio_prepare_push; 835 836 if ((uintptr_t)(pkt->buf + pkt->actual) & 0x7) /* 8byte alignment */ 837 goto usbhsf_pio_prepare_push; 838 839 /* get enable DMA fifo */ 840 fifo = usbhsf_get_dma_fifo(priv, pkt); 841 if (!fifo) 842 goto usbhsf_pio_prepare_push; 843 844 if (usbhsf_dma_map(pkt) < 0) 845 goto usbhsf_pio_prepare_push; 846 847 ret = usbhsf_fifo_select(pipe, fifo, 0); 848 if (ret < 0) 849 goto usbhsf_pio_prepare_push_unmap; 850 851 pkt->trans = len; 852 853 INIT_WORK(&pkt->work, xfer_work); 854 schedule_work(&pkt->work); 855 856 return 0; 857 858 usbhsf_pio_prepare_push_unmap: 859 usbhsf_dma_unmap(pkt); 860 usbhsf_pio_prepare_push: 861 /* 862 * change handler to PIO 863 */ 864 pkt->handler = &usbhs_fifo_pio_push_handler; 865 866 return pkt->handler->prepare(pkt, is_done); 867 } 868 869 static int usbhsf_dma_push_done(struct usbhs_pkt *pkt, int *is_done) 870 { 871 struct usbhs_pipe *pipe = pkt->pipe; 872 873 pkt->actual = pkt->trans; 874 875 *is_done = !pkt->zero; /* send zero packet ? */ 876 877 usbhsf_dma_stop(pipe, pipe->fifo); 878 usbhsf_dma_unmap(pkt); 879 usbhsf_fifo_unselect(pipe, pipe->fifo); 880 881 return 0; 882 } 883 884 struct usbhs_pkt_handle usbhs_fifo_dma_push_handler = { 885 .prepare = usbhsf_dma_prepare_push, 886 .dma_done = usbhsf_dma_push_done, 887 }; 888 889 /* 890 * DMA pop handler 891 */ 892 static int usbhsf_dma_try_pop(struct usbhs_pkt *pkt, int *is_done) 893 { 894 struct usbhs_pipe *pipe = pkt->pipe; 895 struct usbhs_priv *priv = usbhs_pipe_to_priv(pipe); 896 struct usbhs_fifo *fifo; 897 int len, ret; 898 899 if (usbhs_pipe_is_busy(pipe)) 900 return 0; 901 902 if (usbhs_pipe_is_dcp(pipe)) 903 goto usbhsf_pio_prepare_pop; 904 905 /* get enable DMA fifo */ 906 fifo = usbhsf_get_dma_fifo(priv, pkt); 907 if (!fifo) 908 goto usbhsf_pio_prepare_pop; 909 910 if ((uintptr_t)(pkt->buf + pkt->actual) & 0x7) /* 8byte alignment */ 911 goto usbhsf_pio_prepare_pop; 912 913 ret = usbhsf_fifo_select(pipe, fifo, 0); 914 if (ret < 0) 915 goto usbhsf_pio_prepare_pop; 916 917 /* use PIO if packet is less than pio_dma_border */ 918 len = usbhsf_fifo_rcv_len(priv, fifo); 919 len = min(pkt->length - pkt->actual, len); 920 if (len & 0x7) /* 8byte alignment */ 921 goto usbhsf_pio_prepare_pop_unselect; 922 923 if (len < usbhs_get_dparam(priv, pio_dma_border)) 924 goto usbhsf_pio_prepare_pop_unselect; 925 926 ret = usbhsf_fifo_barrier(priv, fifo); 927 if (ret < 0) 928 goto usbhsf_pio_prepare_pop_unselect; 929 930 if (usbhsf_dma_map(pkt) < 0) 931 goto usbhsf_pio_prepare_pop_unselect; 932 933 /* DMA */ 934 935 /* 936 * usbhs_fifo_dma_pop_handler :: prepare 937 * enabled irq to come here. 938 * but it is no longer needed for DMA. disable it. 939 */ 940 usbhsf_rx_irq_ctrl(pipe, 0); 941 942 pkt->trans = len; 943 944 INIT_WORK(&pkt->work, xfer_work); 945 schedule_work(&pkt->work); 946 947 return 0; 948 949 usbhsf_pio_prepare_pop_unselect: 950 usbhsf_fifo_unselect(pipe, fifo); 951 usbhsf_pio_prepare_pop: 952 953 /* 954 * change handler to PIO 955 */ 956 pkt->handler = &usbhs_fifo_pio_pop_handler; 957 958 return pkt->handler->try_run(pkt, is_done); 959 } 960 961 static int usbhsf_dma_pop_done(struct usbhs_pkt *pkt, int *is_done) 962 { 963 struct usbhs_pipe *pipe = pkt->pipe; 964 int maxp = usbhs_pipe_get_maxpacket(pipe); 965 966 usbhsf_dma_stop(pipe, pipe->fifo); 967 usbhsf_dma_unmap(pkt); 968 usbhsf_fifo_unselect(pipe, pipe->fifo); 969 970 pkt->actual += pkt->trans; 971 972 if ((pkt->actual == pkt->length) || /* receive all data */ 973 (pkt->trans < maxp)) { /* short packet */ 974 *is_done = 1; 975 } else { 976 /* re-enable */ 977 usbhsf_prepare_pop(pkt, is_done); 978 } 979 980 return 0; 981 } 982 983 struct usbhs_pkt_handle usbhs_fifo_dma_pop_handler = { 984 .prepare = usbhsf_prepare_pop, 985 .try_run = usbhsf_dma_try_pop, 986 .dma_done = usbhsf_dma_pop_done 987 }; 988 989 /* 990 * DMA setting 991 */ 992 static bool usbhsf_dma_filter(struct dma_chan *chan, void *param) 993 { 994 struct sh_dmae_slave *slave = param; 995 996 /* 997 * FIXME 998 * 999 * usbhs doesn't recognize id = 0 as valid DMA 1000 */ 1001 if (0 == slave->shdma_slave.slave_id) 1002 return false; 1003 1004 chan->private = slave; 1005 1006 return true; 1007 } 1008 1009 static void usbhsf_dma_quit(struct usbhs_priv *priv, struct usbhs_fifo *fifo) 1010 { 1011 if (fifo->tx_chan) 1012 dma_release_channel(fifo->tx_chan); 1013 if (fifo->rx_chan) 1014 dma_release_channel(fifo->rx_chan); 1015 1016 fifo->tx_chan = NULL; 1017 fifo->rx_chan = NULL; 1018 } 1019 1020 static void usbhsf_dma_init(struct usbhs_priv *priv, 1021 struct usbhs_fifo *fifo) 1022 { 1023 struct device *dev = usbhs_priv_to_dev(priv); 1024 dma_cap_mask_t mask; 1025 1026 dma_cap_zero(mask); 1027 dma_cap_set(DMA_SLAVE, mask); 1028 fifo->tx_chan = dma_request_channel(mask, usbhsf_dma_filter, 1029 &fifo->tx_slave); 1030 1031 dma_cap_zero(mask); 1032 dma_cap_set(DMA_SLAVE, mask); 1033 fifo->rx_chan = dma_request_channel(mask, usbhsf_dma_filter, 1034 &fifo->rx_slave); 1035 1036 if (fifo->tx_chan || fifo->rx_chan) 1037 dev_dbg(dev, "enable DMAEngine (%s%s%s)\n", 1038 fifo->name, 1039 fifo->tx_chan ? "[TX]" : " ", 1040 fifo->rx_chan ? "[RX]" : " "); 1041 } 1042 1043 /* 1044 * irq functions 1045 */ 1046 static int usbhsf_irq_empty(struct usbhs_priv *priv, 1047 struct usbhs_irq_state *irq_state) 1048 { 1049 struct usbhs_pipe *pipe; 1050 struct device *dev = usbhs_priv_to_dev(priv); 1051 int i, ret; 1052 1053 if (!irq_state->bempsts) { 1054 dev_err(dev, "debug %s !!\n", __func__); 1055 return -EIO; 1056 } 1057 1058 dev_dbg(dev, "irq empty [0x%04x]\n", irq_state->bempsts); 1059 1060 /* 1061 * search interrupted "pipe" 1062 * not "uep". 1063 */ 1064 usbhs_for_each_pipe_with_dcp(pipe, priv, i) { 1065 if (!(irq_state->bempsts & (1 << i))) 1066 continue; 1067 1068 ret = usbhsf_pkt_handler(pipe, USBHSF_PKT_TRY_RUN); 1069 if (ret < 0) 1070 dev_err(dev, "irq_empty run_error %d : %d\n", i, ret); 1071 } 1072 1073 return 0; 1074 } 1075 1076 static int usbhsf_irq_ready(struct usbhs_priv *priv, 1077 struct usbhs_irq_state *irq_state) 1078 { 1079 struct usbhs_pipe *pipe; 1080 struct device *dev = usbhs_priv_to_dev(priv); 1081 int i, ret; 1082 1083 if (!irq_state->brdysts) { 1084 dev_err(dev, "debug %s !!\n", __func__); 1085 return -EIO; 1086 } 1087 1088 dev_dbg(dev, "irq ready [0x%04x]\n", irq_state->brdysts); 1089 1090 /* 1091 * search interrupted "pipe" 1092 * not "uep". 1093 */ 1094 usbhs_for_each_pipe_with_dcp(pipe, priv, i) { 1095 if (!(irq_state->brdysts & (1 << i))) 1096 continue; 1097 1098 ret = usbhsf_pkt_handler(pipe, USBHSF_PKT_TRY_RUN); 1099 if (ret < 0) 1100 dev_err(dev, "irq_ready run_error %d : %d\n", i, ret); 1101 } 1102 1103 return 0; 1104 } 1105 1106 static void usbhsf_dma_complete(void *arg) 1107 { 1108 struct usbhs_pipe *pipe = arg; 1109 struct usbhs_priv *priv = usbhs_pipe_to_priv(pipe); 1110 struct device *dev = usbhs_priv_to_dev(priv); 1111 int ret; 1112 1113 ret = usbhsf_pkt_handler(pipe, USBHSF_PKT_DMA_DONE); 1114 if (ret < 0) 1115 dev_err(dev, "dma_complete run_error %d : %d\n", 1116 usbhs_pipe_number(pipe), ret); 1117 } 1118 1119 /* 1120 * fifo init 1121 */ 1122 void usbhs_fifo_init(struct usbhs_priv *priv) 1123 { 1124 struct usbhs_mod *mod = usbhs_mod_get_current(priv); 1125 struct usbhs_fifo *cfifo = usbhsf_get_cfifo(priv); 1126 struct usbhs_fifo *d0fifo = usbhsf_get_d0fifo(priv); 1127 struct usbhs_fifo *d1fifo = usbhsf_get_d1fifo(priv); 1128 1129 mod->irq_empty = usbhsf_irq_empty; 1130 mod->irq_ready = usbhsf_irq_ready; 1131 mod->irq_bempsts = 0; 1132 mod->irq_brdysts = 0; 1133 1134 cfifo->pipe = NULL; 1135 d0fifo->pipe = NULL; 1136 d1fifo->pipe = NULL; 1137 } 1138 1139 void usbhs_fifo_quit(struct usbhs_priv *priv) 1140 { 1141 struct usbhs_mod *mod = usbhs_mod_get_current(priv); 1142 1143 mod->irq_empty = NULL; 1144 mod->irq_ready = NULL; 1145 mod->irq_bempsts = 0; 1146 mod->irq_brdysts = 0; 1147 } 1148 1149 int usbhs_fifo_probe(struct usbhs_priv *priv) 1150 { 1151 struct usbhs_fifo *fifo; 1152 1153 /* CFIFO */ 1154 fifo = usbhsf_get_cfifo(priv); 1155 fifo->name = "CFIFO"; 1156 fifo->port = CFIFO; 1157 fifo->sel = CFIFOSEL; 1158 fifo->ctr = CFIFOCTR; 1159 1160 /* D0FIFO */ 1161 fifo = usbhsf_get_d0fifo(priv); 1162 fifo->name = "D0FIFO"; 1163 fifo->port = D0FIFO; 1164 fifo->sel = D0FIFOSEL; 1165 fifo->ctr = D0FIFOCTR; 1166 fifo->tx_slave.shdma_slave.slave_id = usbhs_get_dparam(priv, d0_tx_id); 1167 fifo->rx_slave.shdma_slave.slave_id = usbhs_get_dparam(priv, d0_rx_id); 1168 usbhsf_dma_init(priv, fifo); 1169 1170 /* D1FIFO */ 1171 fifo = usbhsf_get_d1fifo(priv); 1172 fifo->name = "D1FIFO"; 1173 fifo->port = D1FIFO; 1174 fifo->sel = D1FIFOSEL; 1175 fifo->ctr = D1FIFOCTR; 1176 fifo->tx_slave.shdma_slave.slave_id = usbhs_get_dparam(priv, d1_tx_id); 1177 fifo->rx_slave.shdma_slave.slave_id = usbhs_get_dparam(priv, d1_rx_id); 1178 usbhsf_dma_init(priv, fifo); 1179 1180 return 0; 1181 } 1182 1183 void usbhs_fifo_remove(struct usbhs_priv *priv) 1184 { 1185 usbhsf_dma_quit(priv, usbhsf_get_d0fifo(priv)); 1186 usbhsf_dma_quit(priv, usbhsf_get_d1fifo(priv)); 1187 } 1188