1 // SPDX-License-Identifier: GPL-2.0-or-later 2 /* 3 * 4 * Linux MegaRAID device driver 5 * 6 * Copyright (c) 2003-2004 LSI Logic Corporation. 7 * 8 * FILE : megaraid_mm.c 9 * Version : v2.20.2.7 (Jul 16 2006) 10 * 11 * Common management module 12 */ 13 #include <linux/sched.h> 14 #include <linux/slab.h> 15 #include <linux/mutex.h> 16 #include "megaraid_mm.h" 17 18 19 // Entry points for char node driver 20 static DEFINE_MUTEX(mraid_mm_mutex); 21 static int mraid_mm_open(struct inode *, struct file *); 22 static long mraid_mm_unlocked_ioctl(struct file *, uint, unsigned long); 23 24 25 // routines to convert to and from the old the format 26 static int mimd_to_kioc(mimd_t __user *, mraid_mmadp_t *, uioc_t *); 27 static int kioc_to_mimd(uioc_t *, mimd_t __user *); 28 29 30 // Helper functions 31 static int handle_drvrcmd(void __user *, uint8_t, int *); 32 static int lld_ioctl(mraid_mmadp_t *, uioc_t *); 33 static void ioctl_done(uioc_t *); 34 static void lld_timedout(struct timer_list *); 35 static void hinfo_to_cinfo(mraid_hba_info_t *, mcontroller_t *); 36 static mraid_mmadp_t *mraid_mm_get_adapter(mimd_t __user *, int *); 37 static uioc_t *mraid_mm_alloc_kioc(mraid_mmadp_t *); 38 static void mraid_mm_dealloc_kioc(mraid_mmadp_t *, uioc_t *); 39 static int mraid_mm_attach_buf(mraid_mmadp_t *, uioc_t *, int); 40 static int mraid_mm_setup_dma_pools(mraid_mmadp_t *); 41 static void mraid_mm_free_adp_resources(mraid_mmadp_t *); 42 static void mraid_mm_teardown_dma_pools(mraid_mmadp_t *); 43 44 MODULE_AUTHOR("LSI Logic Corporation"); 45 MODULE_DESCRIPTION("LSI Logic Management Module"); 46 MODULE_LICENSE("GPL"); 47 MODULE_VERSION(LSI_COMMON_MOD_VERSION); 48 49 static int dbglevel = CL_ANN; 50 module_param_named(dlevel, dbglevel, int, 0); 51 MODULE_PARM_DESC(dlevel, "Debug level (default=0)"); 52 53 EXPORT_SYMBOL(mraid_mm_register_adp); 54 EXPORT_SYMBOL(mraid_mm_unregister_adp); 55 EXPORT_SYMBOL(mraid_mm_adapter_app_handle); 56 57 static uint32_t drvr_ver = 0x02200207; 58 59 static int adapters_count_g; 60 static struct list_head adapters_list_g; 61 62 static wait_queue_head_t wait_q; 63 64 static const struct file_operations lsi_fops = { 65 .open = mraid_mm_open, 66 .unlocked_ioctl = mraid_mm_unlocked_ioctl, 67 .compat_ioctl = compat_ptr_ioctl, 68 .owner = THIS_MODULE, 69 .llseek = noop_llseek, 70 }; 71 72 static struct miscdevice megaraid_mm_dev = { 73 .minor = MISC_DYNAMIC_MINOR, 74 .name = "megadev0", 75 .fops = &lsi_fops, 76 }; 77 78 /** 79 * mraid_mm_open - open routine for char node interface 80 * @inode : unused 81 * @filep : unused 82 * 83 * Allow ioctl operations by apps only if they have superuser privilege. 84 */ 85 static int 86 mraid_mm_open(struct inode *inode, struct file *filep) 87 { 88 /* 89 * Only allow superuser to access private ioctl interface 90 */ 91 if (!capable(CAP_SYS_ADMIN)) return (-EACCES); 92 93 return 0; 94 } 95 96 /** 97 * mraid_mm_ioctl - module entry-point for ioctls 98 * @filep : file operations pointer (ignored) 99 * @cmd : ioctl command 100 * @arg : user ioctl packet 101 */ 102 static int 103 mraid_mm_ioctl(struct file *filep, unsigned int cmd, unsigned long arg) 104 { 105 uioc_t *kioc; 106 char signature[EXT_IOCTL_SIGN_SZ] = {0}; 107 int rval; 108 mraid_mmadp_t *adp; 109 uint8_t old_ioctl; 110 int drvrcmd_rval; 111 void __user *argp = (void __user *)arg; 112 113 /* 114 * Make sure only USCSICMD are issued through this interface. 115 * MIMD application would still fire different command. 116 */ 117 118 if ((_IOC_TYPE(cmd) != MEGAIOC_MAGIC) && (cmd != USCSICMD)) { 119 return (-EINVAL); 120 } 121 122 /* 123 * Look for signature to see if this is the new or old ioctl format. 124 */ 125 if (copy_from_user(signature, argp, EXT_IOCTL_SIGN_SZ)) { 126 con_log(CL_ANN, (KERN_WARNING 127 "megaraid cmm: copy from usr addr failed\n")); 128 return (-EFAULT); 129 } 130 131 if (memcmp(signature, EXT_IOCTL_SIGN, EXT_IOCTL_SIGN_SZ) == 0) 132 old_ioctl = 0; 133 else 134 old_ioctl = 1; 135 136 /* 137 * At present, we don't support the new ioctl packet 138 */ 139 if (!old_ioctl ) 140 return (-EINVAL); 141 142 /* 143 * If it is a driver ioctl (as opposed to fw ioctls), then we can 144 * handle the command locally. rval > 0 means it is not a drvr cmd 145 */ 146 rval = handle_drvrcmd(argp, old_ioctl, &drvrcmd_rval); 147 148 if (rval < 0) 149 return rval; 150 else if (rval == 0) 151 return drvrcmd_rval; 152 153 rval = 0; 154 if ((adp = mraid_mm_get_adapter(argp, &rval)) == NULL) { 155 return rval; 156 } 157 158 /* 159 * Check if adapter can accept ioctl. We may have marked it offline 160 * if any previous kioc had timedout on this controller. 161 */ 162 if (!adp->quiescent) { 163 con_log(CL_ANN, (KERN_WARNING 164 "megaraid cmm: controller cannot accept cmds due to " 165 "earlier errors\n" )); 166 return -EFAULT; 167 } 168 169 /* 170 * The following call will block till a kioc is available 171 * or return NULL if the list head is empty for the pointer 172 * of type mraid_mmapt passed to mraid_mm_alloc_kioc 173 */ 174 kioc = mraid_mm_alloc_kioc(adp); 175 if (!kioc) 176 return -ENXIO; 177 178 /* 179 * User sent the old mimd_t ioctl packet. Convert it to uioc_t. 180 */ 181 if ((rval = mimd_to_kioc(argp, adp, kioc))) { 182 mraid_mm_dealloc_kioc(adp, kioc); 183 return rval; 184 } 185 186 kioc->done = ioctl_done; 187 188 /* 189 * Issue the IOCTL to the low level driver. After the IOCTL completes 190 * release the kioc if and only if it was _not_ timedout. If it was 191 * timedout, that means that resources are still with low level driver. 192 */ 193 if ((rval = lld_ioctl(adp, kioc))) { 194 195 if (!kioc->timedout) 196 mraid_mm_dealloc_kioc(adp, kioc); 197 198 return rval; 199 } 200 201 /* 202 * Convert the kioc back to user space 203 */ 204 rval = kioc_to_mimd(kioc, argp); 205 206 /* 207 * Return the kioc to free pool 208 */ 209 mraid_mm_dealloc_kioc(adp, kioc); 210 211 return rval; 212 } 213 214 static long 215 mraid_mm_unlocked_ioctl(struct file *filep, unsigned int cmd, 216 unsigned long arg) 217 { 218 int err; 219 220 mutex_lock(&mraid_mm_mutex); 221 err = mraid_mm_ioctl(filep, cmd, arg); 222 mutex_unlock(&mraid_mm_mutex); 223 224 return err; 225 } 226 227 /** 228 * mraid_mm_get_adapter - Returns corresponding adapters for the mimd packet 229 * @umimd : User space mimd_t ioctl packet 230 * @rval : returned success/error status 231 * 232 * The function return value is a pointer to the located @adapter. 233 */ 234 static mraid_mmadp_t * 235 mraid_mm_get_adapter(mimd_t __user *umimd, int *rval) 236 { 237 mraid_mmadp_t *adapter; 238 mimd_t mimd; 239 uint32_t adapno; 240 int iterator; 241 bool is_found; 242 243 if (copy_from_user(&mimd, umimd, sizeof(mimd_t))) { 244 *rval = -EFAULT; 245 return NULL; 246 } 247 248 adapno = GETADAP(mimd.ui.fcs.adapno); 249 250 if (adapno >= adapters_count_g) { 251 *rval = -ENODEV; 252 return NULL; 253 } 254 255 adapter = NULL; 256 iterator = 0; 257 is_found = false; 258 259 list_for_each_entry(adapter, &adapters_list_g, list) { 260 if (iterator++ == adapno) { 261 is_found = true; 262 break; 263 } 264 } 265 266 if (!is_found) { 267 *rval = -ENODEV; 268 return NULL; 269 } 270 271 return adapter; 272 } 273 274 /** 275 * handle_drvrcmd - Checks if the opcode is a driver cmd and if it is, handles it. 276 * @arg : packet sent by the user app 277 * @old_ioctl : mimd if 1; uioc otherwise 278 * @rval : pointer for command's returned value (not function status) 279 */ 280 static int 281 handle_drvrcmd(void __user *arg, uint8_t old_ioctl, int *rval) 282 { 283 mimd_t __user *umimd; 284 mimd_t kmimd; 285 uint8_t opcode; 286 uint8_t subopcode; 287 288 if (old_ioctl) 289 goto old_packet; 290 else 291 goto new_packet; 292 293 new_packet: 294 return (-ENOTSUPP); 295 296 old_packet: 297 *rval = 0; 298 umimd = arg; 299 300 if (copy_from_user(&kmimd, umimd, sizeof(mimd_t))) 301 return (-EFAULT); 302 303 opcode = kmimd.ui.fcs.opcode; 304 subopcode = kmimd.ui.fcs.subopcode; 305 306 /* 307 * If the opcode is 0x82 and the subopcode is either GET_DRVRVER or 308 * GET_NUMADP, then we can handle. Otherwise we should return 1 to 309 * indicate that we cannot handle this. 310 */ 311 if (opcode != 0x82) 312 return 1; 313 314 switch (subopcode) { 315 316 case MEGAIOC_QDRVRVER: 317 318 if (copy_to_user(kmimd.data, &drvr_ver, sizeof(uint32_t))) 319 return (-EFAULT); 320 321 return 0; 322 323 case MEGAIOC_QNADAP: 324 325 *rval = adapters_count_g; 326 327 if (copy_to_user(kmimd.data, &adapters_count_g, 328 sizeof(uint32_t))) 329 return (-EFAULT); 330 331 return 0; 332 333 default: 334 /* cannot handle */ 335 return 1; 336 } 337 338 return 0; 339 } 340 341 342 /** 343 * mimd_to_kioc - Converter from old to new ioctl format 344 * @umimd : user space old MIMD IOCTL 345 * @adp : adapter softstate 346 * @kioc : kernel space new format IOCTL 347 * 348 * Routine to convert MIMD interface IOCTL to new interface IOCTL packet. The 349 * new packet is in kernel space so that driver can perform operations on it 350 * freely. 351 */ 352 353 static int 354 mimd_to_kioc(mimd_t __user *umimd, mraid_mmadp_t *adp, uioc_t *kioc) 355 { 356 mbox64_t *mbox64; 357 mbox_t *mbox; 358 mraid_passthru_t *pthru32; 359 uint32_t adapno; 360 uint8_t opcode; 361 uint8_t subopcode; 362 mimd_t mimd; 363 364 if (copy_from_user(&mimd, umimd, sizeof(mimd_t))) 365 return (-EFAULT); 366 367 /* 368 * Applications are not allowed to send extd pthru 369 */ 370 if ((mimd.mbox[0] == MBOXCMD_PASSTHRU64) || 371 (mimd.mbox[0] == MBOXCMD_EXTPTHRU)) 372 return (-EINVAL); 373 374 opcode = mimd.ui.fcs.opcode; 375 subopcode = mimd.ui.fcs.subopcode; 376 adapno = GETADAP(mimd.ui.fcs.adapno); 377 378 if (adapno >= adapters_count_g) 379 return (-ENODEV); 380 381 kioc->adapno = adapno; 382 kioc->mb_type = MBOX_LEGACY; 383 kioc->app_type = APPTYPE_MIMD; 384 385 switch (opcode) { 386 387 case 0x82: 388 389 if (subopcode == MEGAIOC_QADAPINFO) { 390 391 kioc->opcode = GET_ADAP_INFO; 392 kioc->data_dir = UIOC_RD; 393 kioc->xferlen = sizeof(mraid_hba_info_t); 394 395 if (mraid_mm_attach_buf(adp, kioc, kioc->xferlen)) 396 return (-ENOMEM); 397 } 398 else { 399 con_log(CL_ANN, (KERN_WARNING 400 "megaraid cmm: Invalid subop\n")); 401 return (-EINVAL); 402 } 403 404 break; 405 406 case 0x81: 407 408 kioc->opcode = MBOX_CMD; 409 kioc->xferlen = mimd.ui.fcs.length; 410 kioc->user_data_len = kioc->xferlen; 411 kioc->user_data = mimd.ui.fcs.buffer; 412 413 if (mraid_mm_attach_buf(adp, kioc, kioc->xferlen)) 414 return (-ENOMEM); 415 416 if (mimd.outlen) kioc->data_dir = UIOC_RD; 417 if (mimd.inlen) kioc->data_dir |= UIOC_WR; 418 419 break; 420 421 case 0x80: 422 423 kioc->opcode = MBOX_CMD; 424 kioc->xferlen = (mimd.outlen > mimd.inlen) ? 425 mimd.outlen : mimd.inlen; 426 kioc->user_data_len = kioc->xferlen; 427 kioc->user_data = mimd.data; 428 429 if (mraid_mm_attach_buf(adp, kioc, kioc->xferlen)) 430 return (-ENOMEM); 431 432 if (mimd.outlen) kioc->data_dir = UIOC_RD; 433 if (mimd.inlen) kioc->data_dir |= UIOC_WR; 434 435 break; 436 437 default: 438 return (-EINVAL); 439 } 440 441 /* 442 * If driver command, nothing else to do 443 */ 444 if (opcode == 0x82) 445 return 0; 446 447 /* 448 * This is a mailbox cmd; copy the mailbox from mimd 449 */ 450 mbox64 = (mbox64_t *)((unsigned long)kioc->cmdbuf); 451 mbox = &mbox64->mbox32; 452 memcpy(mbox, mimd.mbox, 14); 453 454 if (mbox->cmd != MBOXCMD_PASSTHRU) { // regular DCMD 455 456 mbox->xferaddr = (uint32_t)kioc->buf_paddr; 457 458 if (kioc->data_dir & UIOC_WR) { 459 if (copy_from_user(kioc->buf_vaddr, kioc->user_data, 460 kioc->xferlen)) { 461 return (-EFAULT); 462 } 463 } 464 465 return 0; 466 } 467 468 /* 469 * This is a regular 32-bit pthru cmd; mbox points to pthru struct. 470 * Just like in above case, the beginning for memblk is treated as 471 * a mailbox. The passthru will begin at next 1K boundary. And the 472 * data will start 1K after that. 473 */ 474 pthru32 = kioc->pthru32; 475 kioc->user_pthru = &umimd->pthru; 476 mbox->xferaddr = (uint32_t)kioc->pthru32_h; 477 478 if (copy_from_user(pthru32, kioc->user_pthru, 479 sizeof(mraid_passthru_t))) { 480 return (-EFAULT); 481 } 482 483 pthru32->dataxferaddr = kioc->buf_paddr; 484 if (kioc->data_dir & UIOC_WR) { 485 if (pthru32->dataxferlen > kioc->xferlen) 486 return -EINVAL; 487 if (copy_from_user(kioc->buf_vaddr, kioc->user_data, 488 pthru32->dataxferlen)) { 489 return (-EFAULT); 490 } 491 } 492 493 return 0; 494 } 495 496 /** 497 * mraid_mm_attach_buf - Attach a free dma buffer for required size 498 * @adp : Adapter softstate 499 * @kioc : kioc that the buffer needs to be attached to 500 * @xferlen : required length for buffer 501 * 502 * First we search for a pool with smallest buffer that is >= @xferlen. If 503 * that pool has no free buffer, we will try for the next bigger size. If none 504 * is available, we will try to allocate the smallest buffer that is >= 505 * @xferlen and attach it the pool. 506 */ 507 static int 508 mraid_mm_attach_buf(mraid_mmadp_t *adp, uioc_t *kioc, int xferlen) 509 { 510 mm_dmapool_t *pool; 511 int right_pool = -1; 512 unsigned long flags; 513 int i; 514 515 kioc->pool_index = -1; 516 kioc->buf_vaddr = NULL; 517 kioc->buf_paddr = 0; 518 kioc->free_buf = 0; 519 520 /* 521 * We need xferlen amount of memory. See if we can get it from our 522 * dma pools. If we don't get exact size, we will try bigger buffer 523 */ 524 525 for (i = 0; i < MAX_DMA_POOLS; i++) { 526 527 pool = &adp->dma_pool_list[i]; 528 529 if (xferlen > pool->buf_size) 530 continue; 531 532 if (right_pool == -1) 533 right_pool = i; 534 535 spin_lock_irqsave(&pool->lock, flags); 536 537 if (!pool->in_use) { 538 539 pool->in_use = 1; 540 kioc->pool_index = i; 541 kioc->buf_vaddr = pool->vaddr; 542 kioc->buf_paddr = pool->paddr; 543 544 spin_unlock_irqrestore(&pool->lock, flags); 545 return 0; 546 } 547 else { 548 spin_unlock_irqrestore(&pool->lock, flags); 549 continue; 550 } 551 } 552 553 /* 554 * If xferlen doesn't match any of our pools, return error 555 */ 556 if (right_pool == -1) 557 return -EINVAL; 558 559 /* 560 * We did not get any buffer from the preallocated pool. Let us try 561 * to allocate one new buffer. NOTE: This is a blocking call. 562 */ 563 pool = &adp->dma_pool_list[right_pool]; 564 565 spin_lock_irqsave(&pool->lock, flags); 566 567 kioc->pool_index = right_pool; 568 kioc->free_buf = 1; 569 kioc->buf_vaddr = dma_pool_alloc(pool->handle, GFP_ATOMIC, 570 &kioc->buf_paddr); 571 spin_unlock_irqrestore(&pool->lock, flags); 572 573 if (!kioc->buf_vaddr) 574 return -ENOMEM; 575 576 return 0; 577 } 578 579 /** 580 * mraid_mm_alloc_kioc - Returns a uioc_t from free list 581 * @adp : Adapter softstate for this module 582 * 583 * The kioc_semaphore is initialized with number of kioc nodes in the 584 * free kioc pool. If the kioc pool is empty, this function blocks till 585 * a kioc becomes free. 586 */ 587 static uioc_t * 588 mraid_mm_alloc_kioc(mraid_mmadp_t *adp) 589 { 590 uioc_t *kioc; 591 struct list_head* head; 592 unsigned long flags; 593 594 down(&adp->kioc_semaphore); 595 596 spin_lock_irqsave(&adp->kioc_pool_lock, flags); 597 598 head = &adp->kioc_pool; 599 600 if (list_empty(head)) { 601 up(&adp->kioc_semaphore); 602 spin_unlock_irqrestore(&adp->kioc_pool_lock, flags); 603 604 con_log(CL_ANN, ("megaraid cmm: kioc list empty!\n")); 605 return NULL; 606 } 607 608 kioc = list_entry(head->next, uioc_t, list); 609 list_del_init(&kioc->list); 610 611 spin_unlock_irqrestore(&adp->kioc_pool_lock, flags); 612 613 memset((caddr_t)(unsigned long)kioc->cmdbuf, 0, sizeof(mbox64_t)); 614 memset((caddr_t) kioc->pthru32, 0, sizeof(mraid_passthru_t)); 615 616 kioc->buf_vaddr = NULL; 617 kioc->buf_paddr = 0; 618 kioc->pool_index =-1; 619 kioc->free_buf = 0; 620 kioc->user_data = NULL; 621 kioc->user_data_len = 0; 622 kioc->user_pthru = NULL; 623 kioc->timedout = 0; 624 625 return kioc; 626 } 627 628 /** 629 * mraid_mm_dealloc_kioc - Return kioc to free pool 630 * @adp : Adapter softstate 631 * @kioc : uioc_t node to be returned to free pool 632 */ 633 static void 634 mraid_mm_dealloc_kioc(mraid_mmadp_t *adp, uioc_t *kioc) 635 { 636 mm_dmapool_t *pool; 637 unsigned long flags; 638 639 if (kioc->pool_index != -1) { 640 pool = &adp->dma_pool_list[kioc->pool_index]; 641 642 /* This routine may be called in non-isr context also */ 643 spin_lock_irqsave(&pool->lock, flags); 644 645 /* 646 * While attaching the dma buffer, if we didn't get the 647 * required buffer from the pool, we would have allocated 648 * it at the run time and set the free_buf flag. We must 649 * free that buffer. Otherwise, just mark that the buffer is 650 * not in use 651 */ 652 if (kioc->free_buf == 1) 653 dma_pool_free(pool->handle, kioc->buf_vaddr, 654 kioc->buf_paddr); 655 else 656 pool->in_use = 0; 657 658 spin_unlock_irqrestore(&pool->lock, flags); 659 } 660 661 /* Return the kioc to the free pool */ 662 spin_lock_irqsave(&adp->kioc_pool_lock, flags); 663 list_add(&kioc->list, &adp->kioc_pool); 664 spin_unlock_irqrestore(&adp->kioc_pool_lock, flags); 665 666 /* increment the free kioc count */ 667 up(&adp->kioc_semaphore); 668 669 return; 670 } 671 672 /** 673 * lld_ioctl - Routine to issue ioctl to low level drvr 674 * @adp : The adapter handle 675 * @kioc : The ioctl packet with kernel addresses 676 */ 677 static int 678 lld_ioctl(mraid_mmadp_t *adp, uioc_t *kioc) 679 { 680 int rval; 681 struct uioc_timeout timeout = { }; 682 683 kioc->status = -ENODATA; 684 rval = adp->issue_uioc(adp->drvr_data, kioc, IOCTL_ISSUE); 685 686 if (rval) return rval; 687 688 /* 689 * Start the timer 690 */ 691 if (adp->timeout > 0) { 692 timeout.uioc = kioc; 693 timer_setup_on_stack(&timeout.timer, lld_timedout, 0); 694 695 timeout.timer.expires = jiffies + adp->timeout * HZ; 696 697 add_timer(&timeout.timer); 698 } 699 700 /* 701 * Wait till the low level driver completes the ioctl. After this 702 * call, the ioctl either completed successfully or timedout. 703 */ 704 wait_event(wait_q, (kioc->status != -ENODATA)); 705 if (timeout.timer.function) { 706 del_timer_sync(&timeout.timer); 707 destroy_timer_on_stack(&timeout.timer); 708 } 709 710 /* 711 * If the command had timedout, we mark the controller offline 712 * before returning 713 */ 714 if (kioc->timedout) { 715 adp->quiescent = 0; 716 } 717 718 return kioc->status; 719 } 720 721 722 /** 723 * ioctl_done - callback from the low level driver 724 * @kioc : completed ioctl packet 725 */ 726 static void 727 ioctl_done(uioc_t *kioc) 728 { 729 uint32_t adapno; 730 int iterator; 731 mraid_mmadp_t* adapter; 732 bool is_found; 733 734 /* 735 * When the kioc returns from driver, make sure it still doesn't 736 * have ENODATA in status. Otherwise, driver will hang on wait_event 737 * forever 738 */ 739 if (kioc->status == -ENODATA) { 740 con_log(CL_ANN, (KERN_WARNING 741 "megaraid cmm: lld didn't change status!\n")); 742 743 kioc->status = -EINVAL; 744 } 745 746 /* 747 * Check if this kioc was timedout before. If so, nobody is waiting 748 * on this kioc. We don't have to wake up anybody. Instead, we just 749 * have to free the kioc 750 */ 751 if (kioc->timedout) { 752 iterator = 0; 753 adapter = NULL; 754 adapno = kioc->adapno; 755 is_found = false; 756 757 con_log(CL_ANN, ( KERN_WARNING "megaraid cmm: completed " 758 "ioctl that was timedout before\n")); 759 760 list_for_each_entry(adapter, &adapters_list_g, list) { 761 if (iterator++ == adapno) { 762 is_found = true; 763 break; 764 } 765 } 766 767 kioc->timedout = 0; 768 769 if (is_found) 770 mraid_mm_dealloc_kioc( adapter, kioc ); 771 772 } 773 else { 774 wake_up(&wait_q); 775 } 776 } 777 778 779 /** 780 * lld_timedout - callback from the expired timer 781 * @t : timer that timed out 782 */ 783 static void 784 lld_timedout(struct timer_list *t) 785 { 786 struct uioc_timeout *timeout = from_timer(timeout, t, timer); 787 uioc_t *kioc = timeout->uioc; 788 789 kioc->status = -ETIME; 790 kioc->timedout = 1; 791 792 con_log(CL_ANN, (KERN_WARNING "megaraid cmm: ioctl timed out\n")); 793 794 wake_up(&wait_q); 795 } 796 797 798 /** 799 * kioc_to_mimd - Converter from new back to old format 800 * @kioc : Kernel space IOCTL packet (successfully issued) 801 * @mimd : User space MIMD packet 802 */ 803 static int 804 kioc_to_mimd(uioc_t *kioc, mimd_t __user *mimd) 805 { 806 mimd_t kmimd; 807 uint8_t opcode; 808 uint8_t subopcode; 809 810 mbox64_t *mbox64; 811 mraid_passthru_t __user *upthru32; 812 mraid_passthru_t *kpthru32; 813 mcontroller_t cinfo; 814 mraid_hba_info_t *hinfo; 815 816 817 if (copy_from_user(&kmimd, mimd, sizeof(mimd_t))) 818 return (-EFAULT); 819 820 opcode = kmimd.ui.fcs.opcode; 821 subopcode = kmimd.ui.fcs.subopcode; 822 823 if (opcode == 0x82) { 824 switch (subopcode) { 825 826 case MEGAIOC_QADAPINFO: 827 828 hinfo = (mraid_hba_info_t *)(unsigned long) 829 kioc->buf_vaddr; 830 831 hinfo_to_cinfo(hinfo, &cinfo); 832 833 if (copy_to_user(kmimd.data, &cinfo, sizeof(cinfo))) 834 return (-EFAULT); 835 836 return 0; 837 838 default: 839 return (-EINVAL); 840 } 841 842 return 0; 843 } 844 845 mbox64 = (mbox64_t *)(unsigned long)kioc->cmdbuf; 846 847 if (kioc->user_pthru) { 848 849 upthru32 = kioc->user_pthru; 850 kpthru32 = kioc->pthru32; 851 852 if (copy_to_user(&upthru32->scsistatus, 853 &kpthru32->scsistatus, 854 sizeof(uint8_t))) { 855 return (-EFAULT); 856 } 857 } 858 859 if (kioc->user_data) { 860 if (copy_to_user(kioc->user_data, kioc->buf_vaddr, 861 kioc->user_data_len)) { 862 return (-EFAULT); 863 } 864 } 865 866 if (copy_to_user(&mimd->mbox[17], 867 &mbox64->mbox32.status, sizeof(uint8_t))) { 868 return (-EFAULT); 869 } 870 871 return 0; 872 } 873 874 875 /** 876 * hinfo_to_cinfo - Convert new format hba info into old format 877 * @hinfo : New format, more comprehensive adapter info 878 * @cinfo : Old format adapter info to support mimd_t apps 879 */ 880 static void 881 hinfo_to_cinfo(mraid_hba_info_t *hinfo, mcontroller_t *cinfo) 882 { 883 if (!hinfo || !cinfo) 884 return; 885 886 cinfo->base = hinfo->baseport; 887 cinfo->irq = hinfo->irq; 888 cinfo->numldrv = hinfo->num_ldrv; 889 cinfo->pcibus = hinfo->pci_bus; 890 cinfo->pcidev = hinfo->pci_slot; 891 cinfo->pcifun = PCI_FUNC(hinfo->pci_dev_fn); 892 cinfo->pciid = hinfo->pci_device_id; 893 cinfo->pcivendor = hinfo->pci_vendor_id; 894 cinfo->pcislot = hinfo->pci_slot; 895 cinfo->uid = hinfo->unique_id; 896 } 897 898 899 /** 900 * mraid_mm_register_adp - Registration routine for low level drivers 901 * @lld_adp : Adapter object 902 */ 903 int 904 mraid_mm_register_adp(mraid_mmadp_t *lld_adp) 905 { 906 mraid_mmadp_t *adapter; 907 mbox64_t *mbox_list; 908 uioc_t *kioc; 909 uint32_t rval; 910 int i; 911 912 913 if (lld_adp->drvr_type != DRVRTYPE_MBOX) 914 return (-EINVAL); 915 916 adapter = kzalloc(sizeof(mraid_mmadp_t), GFP_KERNEL); 917 918 if (!adapter) 919 return -ENOMEM; 920 921 922 adapter->unique_id = lld_adp->unique_id; 923 adapter->drvr_type = lld_adp->drvr_type; 924 adapter->drvr_data = lld_adp->drvr_data; 925 adapter->pdev = lld_adp->pdev; 926 adapter->issue_uioc = lld_adp->issue_uioc; 927 adapter->timeout = lld_adp->timeout; 928 adapter->max_kioc = lld_adp->max_kioc; 929 adapter->quiescent = 1; 930 931 /* 932 * Allocate single blocks of memory for all required kiocs, 933 * mailboxes and passthru structures. 934 */ 935 adapter->kioc_list = kmalloc_array(lld_adp->max_kioc, 936 sizeof(uioc_t), 937 GFP_KERNEL); 938 adapter->mbox_list = kmalloc_array(lld_adp->max_kioc, 939 sizeof(mbox64_t), 940 GFP_KERNEL); 941 adapter->pthru_dma_pool = dma_pool_create("megaraid mm pthru pool", 942 &adapter->pdev->dev, 943 sizeof(mraid_passthru_t), 944 16, 0); 945 946 if (!adapter->kioc_list || !adapter->mbox_list || 947 !adapter->pthru_dma_pool) { 948 949 con_log(CL_ANN, (KERN_WARNING 950 "megaraid cmm: out of memory, %s %d\n", __func__, 951 __LINE__)); 952 953 rval = (-ENOMEM); 954 955 goto memalloc_error; 956 } 957 958 /* 959 * Slice kioc_list and make a kioc_pool with the individiual kiocs 960 */ 961 INIT_LIST_HEAD(&adapter->kioc_pool); 962 spin_lock_init(&adapter->kioc_pool_lock); 963 sema_init(&adapter->kioc_semaphore, lld_adp->max_kioc); 964 965 mbox_list = (mbox64_t *)adapter->mbox_list; 966 967 for (i = 0; i < lld_adp->max_kioc; i++) { 968 969 kioc = adapter->kioc_list + i; 970 kioc->cmdbuf = (uint64_t)(unsigned long)(mbox_list + i); 971 kioc->pthru32 = dma_pool_alloc(adapter->pthru_dma_pool, 972 GFP_KERNEL, &kioc->pthru32_h); 973 974 if (!kioc->pthru32) { 975 976 con_log(CL_ANN, (KERN_WARNING 977 "megaraid cmm: out of memory, %s %d\n", 978 __func__, __LINE__)); 979 980 rval = (-ENOMEM); 981 982 goto pthru_dma_pool_error; 983 } 984 985 list_add_tail(&kioc->list, &adapter->kioc_pool); 986 } 987 988 // Setup the dma pools for data buffers 989 if ((rval = mraid_mm_setup_dma_pools(adapter)) != 0) { 990 goto dma_pool_error; 991 } 992 993 list_add_tail(&adapter->list, &adapters_list_g); 994 995 adapters_count_g++; 996 997 return 0; 998 999 dma_pool_error: 1000 /* Do nothing */ 1001 1002 pthru_dma_pool_error: 1003 1004 for (i = 0; i < lld_adp->max_kioc; i++) { 1005 kioc = adapter->kioc_list + i; 1006 if (kioc->pthru32) { 1007 dma_pool_free(adapter->pthru_dma_pool, kioc->pthru32, 1008 kioc->pthru32_h); 1009 } 1010 } 1011 1012 memalloc_error: 1013 1014 kfree(adapter->kioc_list); 1015 kfree(adapter->mbox_list); 1016 1017 dma_pool_destroy(adapter->pthru_dma_pool); 1018 1019 kfree(adapter); 1020 1021 return rval; 1022 } 1023 1024 1025 /** 1026 * mraid_mm_adapter_app_handle - return the application handle for this adapter 1027 * @unique_id : adapter unique identifier 1028 * 1029 * For the given driver data, locate the adapter in our global list and 1030 * return the corresponding handle, which is also used by applications to 1031 * uniquely identify an adapter. 1032 * 1033 * Return adapter handle if found in the list. 1034 * Return 0 if adapter could not be located, should never happen though. 1035 */ 1036 uint32_t 1037 mraid_mm_adapter_app_handle(uint32_t unique_id) 1038 { 1039 mraid_mmadp_t *adapter; 1040 mraid_mmadp_t *tmp; 1041 int index = 0; 1042 1043 list_for_each_entry_safe(adapter, tmp, &adapters_list_g, list) { 1044 1045 if (adapter->unique_id == unique_id) { 1046 1047 return MKADAP(index); 1048 } 1049 1050 index++; 1051 } 1052 1053 return 0; 1054 } 1055 1056 1057 /** 1058 * mraid_mm_setup_dma_pools - Set up dma buffer pools per adapter 1059 * @adp : Adapter softstate 1060 * 1061 * We maintain a pool of dma buffers per each adapter. Each pool has one 1062 * buffer. E.g, we may have 5 dma pools - one each for 4k, 8k ... 64k buffers. 1063 * We have just one 4k buffer in 4k pool, one 8k buffer in 8k pool etc. We 1064 * dont' want to waste too much memory by allocating more buffers per each 1065 * pool. 1066 */ 1067 static int 1068 mraid_mm_setup_dma_pools(mraid_mmadp_t *adp) 1069 { 1070 mm_dmapool_t *pool; 1071 int bufsize; 1072 int i; 1073 1074 /* 1075 * Create MAX_DMA_POOLS number of pools 1076 */ 1077 bufsize = MRAID_MM_INIT_BUFF_SIZE; 1078 1079 for (i = 0; i < MAX_DMA_POOLS; i++){ 1080 1081 pool = &adp->dma_pool_list[i]; 1082 1083 pool->buf_size = bufsize; 1084 spin_lock_init(&pool->lock); 1085 1086 pool->handle = dma_pool_create("megaraid mm data buffer", 1087 &adp->pdev->dev, bufsize, 1088 16, 0); 1089 1090 if (!pool->handle) { 1091 goto dma_pool_setup_error; 1092 } 1093 1094 pool->vaddr = dma_pool_alloc(pool->handle, GFP_KERNEL, 1095 &pool->paddr); 1096 1097 if (!pool->vaddr) 1098 goto dma_pool_setup_error; 1099 1100 bufsize = bufsize * 2; 1101 } 1102 1103 return 0; 1104 1105 dma_pool_setup_error: 1106 1107 mraid_mm_teardown_dma_pools(adp); 1108 return (-ENOMEM); 1109 } 1110 1111 1112 /** 1113 * mraid_mm_unregister_adp - Unregister routine for low level drivers 1114 * @unique_id : UID of the adpater 1115 * 1116 * Assumes no outstanding ioctls to llds. 1117 */ 1118 int 1119 mraid_mm_unregister_adp(uint32_t unique_id) 1120 { 1121 mraid_mmadp_t *adapter; 1122 mraid_mmadp_t *tmp; 1123 1124 list_for_each_entry_safe(adapter, tmp, &adapters_list_g, list) { 1125 1126 1127 if (adapter->unique_id == unique_id) { 1128 1129 adapters_count_g--; 1130 1131 list_del_init(&adapter->list); 1132 1133 mraid_mm_free_adp_resources(adapter); 1134 1135 kfree(adapter); 1136 1137 con_log(CL_ANN, ( 1138 "megaraid cmm: Unregistered one adapter:%#x\n", 1139 unique_id)); 1140 1141 return 0; 1142 } 1143 } 1144 1145 return (-ENODEV); 1146 } 1147 1148 /** 1149 * mraid_mm_free_adp_resources - Free adapter softstate 1150 * @adp : Adapter softstate 1151 */ 1152 static void 1153 mraid_mm_free_adp_resources(mraid_mmadp_t *adp) 1154 { 1155 uioc_t *kioc; 1156 int i; 1157 1158 mraid_mm_teardown_dma_pools(adp); 1159 1160 for (i = 0; i < adp->max_kioc; i++) { 1161 1162 kioc = adp->kioc_list + i; 1163 1164 dma_pool_free(adp->pthru_dma_pool, kioc->pthru32, 1165 kioc->pthru32_h); 1166 } 1167 1168 kfree(adp->kioc_list); 1169 kfree(adp->mbox_list); 1170 1171 dma_pool_destroy(adp->pthru_dma_pool); 1172 1173 1174 return; 1175 } 1176 1177 1178 /** 1179 * mraid_mm_teardown_dma_pools - Free all per adapter dma buffers 1180 * @adp : Adapter softstate 1181 */ 1182 static void 1183 mraid_mm_teardown_dma_pools(mraid_mmadp_t *adp) 1184 { 1185 int i; 1186 mm_dmapool_t *pool; 1187 1188 for (i = 0; i < MAX_DMA_POOLS; i++) { 1189 1190 pool = &adp->dma_pool_list[i]; 1191 1192 if (pool->handle) { 1193 1194 if (pool->vaddr) 1195 dma_pool_free(pool->handle, pool->vaddr, 1196 pool->paddr); 1197 1198 dma_pool_destroy(pool->handle); 1199 pool->handle = NULL; 1200 } 1201 } 1202 1203 return; 1204 } 1205 1206 /** 1207 * mraid_mm_init - Module entry point 1208 */ 1209 static int __init 1210 mraid_mm_init(void) 1211 { 1212 int err; 1213 1214 // Announce the driver version 1215 con_log(CL_ANN, (KERN_INFO "megaraid cmm: %s %s\n", 1216 LSI_COMMON_MOD_VERSION, LSI_COMMON_MOD_EXT_VERSION)); 1217 1218 err = misc_register(&megaraid_mm_dev); 1219 if (err < 0) { 1220 con_log(CL_ANN, ("megaraid cmm: cannot register misc device\n")); 1221 return err; 1222 } 1223 1224 init_waitqueue_head(&wait_q); 1225 1226 INIT_LIST_HEAD(&adapters_list_g); 1227 1228 return 0; 1229 } 1230 1231 1232 /** 1233 * mraid_mm_exit - Module exit point 1234 */ 1235 static void __exit 1236 mraid_mm_exit(void) 1237 { 1238 con_log(CL_DLEVEL1 , ("exiting common mod\n")); 1239 1240 misc_deregister(&megaraid_mm_dev); 1241 } 1242 1243 module_init(mraid_mm_init); 1244 module_exit(mraid_mm_exit); 1245 1246 /* vi: set ts=8 sw=8 tw=78: */ 1247