1 /* $Id: osm_bsd.c,v 1.36 2010/05/11 03:12:11 lcn Exp $ */ 2 /*- 3 * SPDX-License-Identifier: BSD-2-Clause-FreeBSD 4 * 5 * HighPoint RAID Driver for FreeBSD 6 * Copyright (C) 2005-2011 HighPoint Technologies, Inc. 7 * All rights reserved. 8 * 9 * Redistribution and use in source and binary forms, with or without 10 * modification, are permitted provided that the following conditions 11 * are met: 12 * 1. Redistributions of source code must retain the above copyright 13 * notice, this list of conditions and the following disclaimer. 14 * 2. Redistributions in binary form must reproduce the above copyright 15 * notice, this list of conditions and the following disclaimer in the 16 * documentation and/or other materials provided with the distribution. 17 * 18 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 19 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 20 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 21 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 22 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 23 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 24 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 25 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 26 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 27 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 28 * SUCH DAMAGE. 29 * 30 * $FreeBSD$ 31 */ 32 #include <dev/hptnr/hptnr_config.h> 33 #include <dev/hptnr/os_bsd.h> 34 #include <dev/hptnr/hptintf.h> 35 int msi = 0; 36 int debug_flag = 0; 37 static HIM *hpt_match(device_t dev, int scan) 38 { 39 PCI_ID pci_id; 40 HIM *him; 41 int i; 42 43 for (him = him_list; him; him = him->next) { 44 for (i=0; him->get_supported_device_id(i, &pci_id); i++) { 45 if (scan && him->get_controller_count) 46 him->get_controller_count(&pci_id,0,0); 47 if ((pci_get_vendor(dev) == pci_id.vid) && 48 (pci_get_device(dev) == pci_id.did)){ 49 return (him); 50 } 51 } 52 } 53 54 return (NULL); 55 } 56 57 static int hpt_probe(device_t dev) 58 { 59 HIM *him; 60 61 him = hpt_match(dev, 0); 62 if (him != NULL) { 63 KdPrint(("hpt_probe: adapter at PCI %d:%d:%d, IRQ %d", 64 pci_get_bus(dev), pci_get_slot(dev), pci_get_function(dev), pci_get_irq(dev) 65 )); 66 device_set_desc(dev, him->name); 67 return (BUS_PROBE_DEFAULT); 68 } 69 70 return (ENXIO); 71 } 72 73 static int hpt_attach(device_t dev) 74 { 75 PHBA hba = (PHBA)device_get_softc(dev); 76 HIM *him; 77 PCI_ID pci_id; 78 HPT_UINT size; 79 PVBUS vbus; 80 PVBUS_EXT vbus_ext; 81 82 KdPrint(("hpt_attach(%d/%d/%d)", pci_get_bus(dev), pci_get_slot(dev), pci_get_function(dev))); 83 84 him = hpt_match(dev, 1); 85 hba->ext_type = EXT_TYPE_HBA; 86 hba->ldm_adapter.him = him; 87 88 pci_enable_busmaster(dev); 89 90 pci_id.vid = pci_get_vendor(dev); 91 pci_id.did = pci_get_device(dev); 92 pci_id.rev = pci_get_revid(dev); 93 pci_id.subsys = (HPT_U32)(pci_get_subdevice(dev)) << 16 | pci_get_subvendor(dev); 94 95 size = him->get_adapter_size(&pci_id); 96 hba->ldm_adapter.him_handle = malloc(size, M_DEVBUF, M_WAITOK); 97 98 hba->pcidev = dev; 99 hba->pciaddr.tree = 0; 100 hba->pciaddr.bus = pci_get_bus(dev); 101 hba->pciaddr.device = pci_get_slot(dev); 102 hba->pciaddr.function = pci_get_function(dev); 103 104 if (!him->create_adapter(&pci_id, hba->pciaddr, hba->ldm_adapter.him_handle, hba)) { 105 free(hba->ldm_adapter.him_handle, M_DEVBUF); 106 return ENXIO; 107 } 108 109 os_printk("adapter at PCI %d:%d:%d, IRQ %d", 110 hba->pciaddr.bus, hba->pciaddr.device, hba->pciaddr.function, pci_get_irq(dev)); 111 112 if (!ldm_register_adapter(&hba->ldm_adapter)) { 113 size = ldm_get_vbus_size(); 114 vbus_ext = malloc(sizeof(VBUS_EXT) + size, M_DEVBUF, M_WAITOK | 115 M_ZERO); 116 vbus_ext->ext_type = EXT_TYPE_VBUS; 117 ldm_create_vbus((PVBUS)vbus_ext->vbus, vbus_ext); 118 ldm_register_adapter(&hba->ldm_adapter); 119 } 120 121 ldm_for_each_vbus(vbus, vbus_ext) { 122 if (hba->ldm_adapter.vbus==vbus) { 123 hba->vbus_ext = vbus_ext; 124 hba->next = vbus_ext->hba_list; 125 vbus_ext->hba_list = hba; 126 break; 127 } 128 } 129 return 0; 130 } 131 132 /* 133 * Maybe we'd better to use the bus_dmamem_alloc to alloc DMA memory, 134 * but there are some problems currently (alignment, etc). 135 */ 136 static __inline void *__get_free_pages(int order) 137 { 138 /* don't use low memory - other devices may get starved */ 139 return contigmalloc(PAGE_SIZE<<order, 140 M_DEVBUF, M_WAITOK, BUS_SPACE_MAXADDR_24BIT, BUS_SPACE_MAXADDR, PAGE_SIZE, 0); 141 } 142 143 static __inline void free_pages(void *p, int order) 144 { 145 contigfree(p, PAGE_SIZE<<order, M_DEVBUF); 146 } 147 148 static int hpt_alloc_mem(PVBUS_EXT vbus_ext) 149 { 150 PHBA hba; 151 struct freelist *f; 152 HPT_UINT i; 153 void **p; 154 155 for (hba = vbus_ext->hba_list; hba; hba = hba->next) 156 hba->ldm_adapter.him->get_meminfo(hba->ldm_adapter.him_handle); 157 158 ldm_get_mem_info((PVBUS)vbus_ext->vbus, 0); 159 160 for (f=vbus_ext->freelist_head; f; f=f->next) { 161 KdPrint(("%s: %d*%d=%d bytes", 162 f->tag, f->count, f->size, f->count*f->size)); 163 for (i=0; i<f->count; i++) { 164 p = (void **)malloc(f->size, M_DEVBUF, M_WAITOK); 165 if (!p) return (ENXIO); 166 *p = f->head; 167 f->head = p; 168 } 169 } 170 171 for (f=vbus_ext->freelist_dma_head; f; f=f->next) { 172 int order, size, j; 173 174 HPT_ASSERT((f->size & (f->alignment-1))==0); 175 176 for (order=0, size=PAGE_SIZE; size<f->size; order++, size<<=1) 177 ; 178 179 KdPrint(("%s: %d*%d=%d bytes, order %d", 180 f->tag, f->count, f->size, f->count*f->size, order)); 181 HPT_ASSERT(f->alignment<=PAGE_SIZE); 182 183 for (i=0; i<f->count;) { 184 p = (void **)__get_free_pages(order); 185 if (!p) return -1; 186 for (j = size/f->size; j && i<f->count; i++,j--) { 187 *p = f->head; 188 *(BUS_ADDRESS *)(p+1) = (BUS_ADDRESS)vtophys(p); 189 f->head = p; 190 p = (void **)((unsigned long)p + f->size); 191 } 192 } 193 } 194 195 HPT_ASSERT(PAGE_SIZE==DMAPOOL_PAGE_SIZE); 196 197 for (i=0; i<os_max_cache_pages; i++) { 198 p = (void **)__get_free_pages(0); 199 if (!p) return -1; 200 HPT_ASSERT(((HPT_UPTR)p & (DMAPOOL_PAGE_SIZE-1))==0); 201 dmapool_put_page((PVBUS)vbus_ext->vbus, p, (BUS_ADDRESS)vtophys(p)); 202 } 203 204 return 0; 205 } 206 207 static void hpt_free_mem(PVBUS_EXT vbus_ext) 208 { 209 struct freelist *f; 210 void *p; 211 int i; 212 BUS_ADDRESS bus; 213 214 for (f=vbus_ext->freelist_head; f; f=f->next) { 215 #if DBG 216 if (f->count!=f->reserved_count) { 217 KdPrint(("memory leak for freelist %s (%d/%d)", f->tag, f->count, f->reserved_count)); 218 } 219 #endif 220 while ((p=freelist_get(f))) 221 free(p, M_DEVBUF); 222 } 223 224 for (i=0; i<os_max_cache_pages; i++) { 225 p = dmapool_get_page((PVBUS)vbus_ext->vbus, &bus); 226 HPT_ASSERT(p); 227 free_pages(p, 0); 228 } 229 230 for (f=vbus_ext->freelist_dma_head; f; f=f->next) { 231 int order, size; 232 #if DBG 233 if (f->count!=f->reserved_count) { 234 KdPrint(("memory leak for dma freelist %s (%d/%d)", f->tag, f->count, f->reserved_count)); 235 } 236 #endif 237 for (order=0, size=PAGE_SIZE; size<f->size; order++, size<<=1) ; 238 239 while ((p=freelist_get_dma(f, &bus))) { 240 if (order) 241 free_pages(p, order); 242 else { 243 /* can't free immediately since other blocks in this page may still be in the list */ 244 if (((HPT_UPTR)p & (PAGE_SIZE-1))==0) 245 dmapool_put_page((PVBUS)vbus_ext->vbus, p, bus); 246 } 247 } 248 } 249 250 while ((p = dmapool_get_page((PVBUS)vbus_ext->vbus, &bus))) 251 free_pages(p, 0); 252 } 253 254 static int hpt_init_vbus(PVBUS_EXT vbus_ext) 255 { 256 PHBA hba; 257 258 for (hba = vbus_ext->hba_list; hba; hba = hba->next) 259 if (!hba->ldm_adapter.him->initialize(hba->ldm_adapter.him_handle)) { 260 KdPrint(("fail to initialize %p", hba)); 261 return -1; 262 } 263 264 ldm_initialize_vbus((PVBUS)vbus_ext->vbus, &vbus_ext->hba_list->ldm_adapter); 265 return 0; 266 } 267 268 static void hpt_flush_done(PCOMMAND pCmd) 269 { 270 PVDEV vd = pCmd->target; 271 272 if (mIsArray(vd->type) && vd->u.array.transform && vd!=vd->u.array.transform->target) { 273 vd = vd->u.array.transform->target; 274 HPT_ASSERT(vd); 275 pCmd->target = vd; 276 pCmd->Result = RETURN_PENDING; 277 vdev_queue_cmd(pCmd); 278 return; 279 } 280 281 *(int *)pCmd->priv = 1; 282 wakeup(pCmd); 283 } 284 285 /* 286 * flush a vdev (without retry). 287 */ 288 static int hpt_flush_vdev(PVBUS_EXT vbus_ext, PVDEV vd) 289 { 290 PCOMMAND pCmd; 291 int result = 0, done; 292 HPT_UINT count; 293 294 KdPrint(("flusing dev %p", vd)); 295 296 hpt_assert_vbus_locked(vbus_ext); 297 298 if (mIsArray(vd->type) && vd->u.array.transform) 299 count = max(vd->u.array.transform->source->cmds_per_request, 300 vd->u.array.transform->target->cmds_per_request); 301 else 302 count = vd->cmds_per_request; 303 304 pCmd = ldm_alloc_cmds(vd->vbus, count); 305 306 if (!pCmd) { 307 return -1; 308 } 309 310 pCmd->type = CMD_TYPE_FLUSH; 311 pCmd->flags.hard_flush = 1; 312 pCmd->target = vd; 313 pCmd->done = hpt_flush_done; 314 done = 0; 315 pCmd->priv = &done; 316 317 ldm_queue_cmd(pCmd); 318 319 if (!done) { 320 while (hpt_sleep(vbus_ext, pCmd, PPAUSE, "hptfls", HPT_OSM_TIMEOUT)) { 321 ldm_reset_vbus(vd->vbus); 322 } 323 } 324 325 KdPrint(("flush result %d", pCmd->Result)); 326 327 if (pCmd->Result!=RETURN_SUCCESS) 328 result = -1; 329 330 ldm_free_cmds(pCmd); 331 332 return result; 333 } 334 335 static void hpt_stop_tasks(PVBUS_EXT vbus_ext); 336 static void hpt_shutdown_vbus(PVBUS_EXT vbus_ext, int howto) 337 { 338 PVBUS vbus = (PVBUS)vbus_ext->vbus; 339 PHBA hba; 340 int i; 341 342 KdPrint(("hpt_shutdown_vbus")); 343 344 /* stop all ctl tasks and disable the worker taskqueue */ 345 hpt_stop_tasks(vbus_ext); 346 hpt_lock_vbus(vbus_ext); 347 vbus_ext->worker.ta_context = 0; 348 349 /* flush devices */ 350 for (i=0; i<osm_max_targets; i++) { 351 PVDEV vd = ldm_find_target(vbus, i); 352 if (vd) { 353 /* retry once */ 354 if (hpt_flush_vdev(vbus_ext, vd)) 355 hpt_flush_vdev(vbus_ext, vd); 356 } 357 } 358 359 ldm_shutdown(vbus); 360 hpt_unlock_vbus(vbus_ext); 361 362 ldm_release_vbus(vbus); 363 364 for (hba=vbus_ext->hba_list; hba; hba=hba->next) 365 bus_teardown_intr(hba->pcidev, hba->irq_res, hba->irq_handle); 366 367 hpt_free_mem(vbus_ext); 368 369 while ((hba=vbus_ext->hba_list)) { 370 vbus_ext->hba_list = hba->next; 371 free(hba->ldm_adapter.him_handle, M_DEVBUF); 372 } 373 374 callout_drain(&vbus_ext->timer); 375 mtx_destroy(&vbus_ext->lock); 376 free(vbus_ext, M_DEVBUF); 377 KdPrint(("hpt_shutdown_vbus done")); 378 } 379 380 static void __hpt_do_tasks(PVBUS_EXT vbus_ext) 381 { 382 OSM_TASK *tasks; 383 384 tasks = vbus_ext->tasks; 385 vbus_ext->tasks = 0; 386 387 while (tasks) { 388 OSM_TASK *t = tasks; 389 tasks = t->next; 390 t->next = 0; 391 t->func(vbus_ext->vbus, t->data); 392 } 393 } 394 395 static void hpt_do_tasks(PVBUS_EXT vbus_ext, int pending) 396 { 397 if(vbus_ext){ 398 hpt_lock_vbus(vbus_ext); 399 __hpt_do_tasks(vbus_ext); 400 hpt_unlock_vbus(vbus_ext); 401 } 402 } 403 404 static void hpt_action(struct cam_sim *sim, union ccb *ccb); 405 static void hpt_poll(struct cam_sim *sim); 406 static void hpt_async(void * callback_arg, u_int32_t code, struct cam_path * path, void * arg); 407 static void hpt_pci_intr(void *arg); 408 409 static __inline POS_CMDEXT cmdext_get(PVBUS_EXT vbus_ext) 410 { 411 POS_CMDEXT p = vbus_ext->cmdext_list; 412 if (p) 413 vbus_ext->cmdext_list = p->next; 414 return p; 415 } 416 417 static __inline void cmdext_put(POS_CMDEXT p) 418 { 419 p->next = p->vbus_ext->cmdext_list; 420 p->vbus_ext->cmdext_list = p; 421 } 422 423 static void hpt_timeout(void *arg) 424 { 425 PCOMMAND pCmd = (PCOMMAND)arg; 426 POS_CMDEXT ext = (POS_CMDEXT)pCmd->priv; 427 428 KdPrint(("pCmd %p timeout", pCmd)); 429 430 ldm_reset_vbus((PVBUS)ext->vbus_ext->vbus); 431 } 432 433 static void os_cmddone(PCOMMAND pCmd) 434 { 435 POS_CMDEXT ext = (POS_CMDEXT)pCmd->priv; 436 union ccb *ccb = ext->ccb; 437 HPT_U8 *cdb; 438 439 if (ccb->ccb_h.flags & CAM_CDB_POINTER) 440 cdb = ccb->csio.cdb_io.cdb_ptr; 441 else 442 cdb = ccb->csio.cdb_io.cdb_bytes; 443 444 KdPrint(("os_cmddone(%p, %d)", pCmd, pCmd->Result)); 445 446 callout_stop(&ext->timeout); 447 switch(cdb[0]) { 448 case 0x85: /*ATA_16*/ 449 case 0xA1: /*ATA_12*/ 450 { 451 PassthroughCmd *passthru = &pCmd->uCmd.Passthrough; 452 HPT_U8 *sense_buffer = (HPT_U8 *)&ccb->csio.sense_data; 453 memset(&ccb->csio.sense_data, 0,sizeof(ccb->csio.sense_data)); 454 455 sense_buffer[0] = 0x72; /* Response Code */ 456 sense_buffer[7] = 14; /* Additional Sense Length */ 457 458 sense_buffer[8] = 0x9; /* ATA Return Descriptor */ 459 sense_buffer[9] = 0xc; /* Additional Descriptor Length */ 460 sense_buffer[11] = (HPT_U8)passthru->bFeaturesReg; /* Error */ 461 sense_buffer[13] = (HPT_U8)passthru->bSectorCountReg; /* Sector Count (7:0) */ 462 sense_buffer[15] = (HPT_U8)passthru->bLbaLowReg; /* LBA Low (7:0) */ 463 sense_buffer[17] = (HPT_U8)passthru->bLbaMidReg; /* LBA Mid (7:0) */ 464 sense_buffer[19] = (HPT_U8)passthru->bLbaHighReg; /* LBA High (7:0) */ 465 466 if ((cdb[0] == 0x85) && (cdb[1] & 0x1)) 467 { 468 sense_buffer[10] = 1; 469 sense_buffer[12] = (HPT_U8)(passthru->bSectorCountReg >> 8); /* Sector Count (15:8) */ 470 sense_buffer[14] = (HPT_U8)(passthru->bLbaLowReg >> 8); /* LBA Low (15:8) */ 471 sense_buffer[16] = (HPT_U8)(passthru->bLbaMidReg >> 8); /* LBA Mid (15:8) */ 472 sense_buffer[18] = (HPT_U8)(passthru->bLbaHighReg >> 8); /* LBA High (15:8) */ 473 } 474 475 sense_buffer[20] = (HPT_U8)passthru->bDriveHeadReg; /* Device */ 476 sense_buffer[21] = (HPT_U8)passthru->bCommandReg; /* Status */ 477 KdPrint(("sts 0x%x err 0x%x low 0x%x mid 0x%x hig 0x%x dh 0x%x sc 0x%x", 478 passthru->bCommandReg, 479 passthru->bFeaturesReg, 480 passthru->bLbaLowReg, 481 passthru->bLbaMidReg, 482 passthru->bLbaHighReg, 483 passthru->bDriveHeadReg, 484 passthru->bSectorCountReg)); 485 KdPrint(("result:0x%x,bFeaturesReg:0x%04x,bSectorCountReg:0x%04x,LBA:0x%04x%04x%04x ", 486 pCmd->Result,passthru->bFeaturesReg,passthru->bSectorCountReg, 487 passthru->bLbaHighReg,passthru->bLbaMidReg,passthru->bLbaLowReg)); 488 } 489 default: 490 break; 491 } 492 493 switch(pCmd->Result) { 494 case RETURN_SUCCESS: 495 ccb->ccb_h.status = CAM_REQ_CMP; 496 break; 497 case RETURN_BAD_DEVICE: 498 ccb->ccb_h.status = CAM_DEV_NOT_THERE; 499 break; 500 case RETURN_DEVICE_BUSY: 501 ccb->ccb_h.status = CAM_BUSY; 502 break; 503 case RETURN_INVALID_REQUEST: 504 ccb->ccb_h.status = CAM_REQ_INVALID; 505 break; 506 case RETURN_SELECTION_TIMEOUT: 507 ccb->ccb_h.status = CAM_SEL_TIMEOUT; 508 break; 509 case RETURN_RETRY: 510 ccb->ccb_h.status = CAM_BUSY; 511 break; 512 default: 513 ccb->ccb_h.status = CAM_SCSI_STATUS_ERROR; 514 break; 515 } 516 517 if (pCmd->flags.data_in) { 518 bus_dmamap_sync(ext->vbus_ext->io_dmat, ext->dma_map, BUS_DMASYNC_POSTREAD); 519 } 520 else if (pCmd->flags.data_out) { 521 bus_dmamap_sync(ext->vbus_ext->io_dmat, ext->dma_map, BUS_DMASYNC_POSTWRITE); 522 } 523 524 bus_dmamap_unload(ext->vbus_ext->io_dmat, ext->dma_map); 525 526 cmdext_put(ext); 527 ldm_free_cmds(pCmd); 528 xpt_done(ccb); 529 } 530 531 static int os_buildsgl(PCOMMAND pCmd, PSG pSg, int logical) 532 { 533 /* since we have provided physical sg, nobody will ask us to build physical sg */ 534 HPT_ASSERT(0); 535 return FALSE; 536 } 537 538 static void hpt_io_dmamap_callback(void *arg, bus_dma_segment_t *segs, int nsegs, int error) 539 { 540 PCOMMAND pCmd = (PCOMMAND)arg; 541 POS_CMDEXT ext = (POS_CMDEXT)pCmd->priv; 542 PSG psg = pCmd->psg; 543 int idx; 544 545 HPT_ASSERT(pCmd->flags.physical_sg); 546 547 if (error) 548 panic("busdma error"); 549 550 HPT_ASSERT(nsegs<=os_max_sg_descriptors); 551 552 if (nsegs != 0) { 553 for (idx = 0; idx < nsegs; idx++, psg++) { 554 psg->addr.bus = segs[idx].ds_addr; 555 psg->size = segs[idx].ds_len; 556 psg->eot = 0; 557 } 558 psg[-1].eot = 1; 559 560 if (pCmd->flags.data_in) { 561 bus_dmamap_sync(ext->vbus_ext->io_dmat, ext->dma_map, 562 BUS_DMASYNC_PREREAD); 563 } 564 else if (pCmd->flags.data_out) { 565 bus_dmamap_sync(ext->vbus_ext->io_dmat, ext->dma_map, 566 BUS_DMASYNC_PREWRITE); 567 } 568 } 569 570 callout_reset(&ext->timeout, HPT_OSM_TIMEOUT, hpt_timeout, pCmd); 571 ldm_queue_cmd(pCmd); 572 } 573 574 static void hpt_scsi_io(PVBUS_EXT vbus_ext, union ccb *ccb) 575 { 576 PVBUS vbus = (PVBUS)vbus_ext->vbus; 577 PVDEV vd; 578 PCOMMAND pCmd; 579 POS_CMDEXT ext; 580 HPT_U8 *cdb; 581 582 if (ccb->ccb_h.flags & CAM_CDB_POINTER) 583 cdb = ccb->csio.cdb_io.cdb_ptr; 584 else 585 cdb = ccb->csio.cdb_io.cdb_bytes; 586 587 KdPrint(("hpt_scsi_io: ccb %x id %d lun %d cdb %x-%x-%x", 588 ccb, 589 ccb->ccb_h.target_id, ccb->ccb_h.target_lun, 590 *(HPT_U32 *)&cdb[0], *(HPT_U32 *)&cdb[4], *(HPT_U32 *)&cdb[8] 591 )); 592 593 /* ccb->ccb_h.path_id is not our bus id - don't check it */ 594 if (ccb->ccb_h.target_lun != 0 || 595 ccb->ccb_h.target_id >= osm_max_targets || 596 (ccb->ccb_h.flags & CAM_CDB_PHYS)) 597 { 598 ccb->ccb_h.status = CAM_TID_INVALID; 599 xpt_done(ccb); 600 return; 601 } 602 603 vd = ldm_find_target(vbus, ccb->ccb_h.target_id); 604 605 if (!vd) { 606 ccb->ccb_h.status = CAM_SEL_TIMEOUT; 607 xpt_done(ccb); 608 return; 609 } 610 611 switch (cdb[0]) { 612 case TEST_UNIT_READY: 613 case START_STOP_UNIT: 614 case SYNCHRONIZE_CACHE: 615 ccb->ccb_h.status = CAM_REQ_CMP; 616 break; 617 618 case 0x85: /*ATA_16*/ 619 case 0xA1: /*ATA_12*/ 620 { 621 int error; 622 HPT_U8 prot; 623 PassthroughCmd *passthru; 624 625 if (mIsArray(vd->type)) { 626 ccb->ccb_h.status = CAM_REQ_INVALID; 627 break; 628 } 629 630 HPT_ASSERT(vd->type == VD_RAW && vd->u.raw.legacy_disk); 631 632 prot = (cdb[1] & 0x1e) >> 1; 633 634 635 if (prot < 3 || prot > 5) 636 { 637 ccb->ccb_h.status = CAM_REQ_INVALID; 638 break; 639 } 640 641 pCmd = ldm_alloc_cmds(vbus, vd->cmds_per_request); 642 if (!pCmd) { 643 HPT_ASSERT(0); 644 ccb->ccb_h.status = CAM_BUSY; 645 break; 646 } 647 648 passthru = &pCmd->uCmd.Passthrough; 649 if (cdb[0] == 0x85/*ATA_16*/) { 650 if (cdb[1] & 0x1) { 651 passthru->bFeaturesReg = 652 ((HPT_U16)cdb[3] << 8) 653 | cdb[4]; 654 passthru->bSectorCountReg = 655 ((HPT_U16)cdb[5] << 8) | 656 cdb[6]; 657 passthru->bLbaLowReg = 658 ((HPT_U16)cdb[7] << 8) | 659 cdb[8]; 660 passthru->bLbaMidReg = 661 ((HPT_U16)cdb[9] << 8) | 662 cdb[10]; 663 passthru->bLbaHighReg = 664 ((HPT_U16)cdb[11] << 8) | 665 cdb[12]; 666 } else { 667 passthru->bFeaturesReg = cdb[4]; 668 passthru->bSectorCountReg = cdb[6]; 669 passthru->bLbaLowReg = cdb[8]; 670 passthru->bLbaMidReg = cdb[10]; 671 passthru->bLbaHighReg = cdb[12]; 672 } 673 passthru->bDriveHeadReg = cdb[13]; 674 passthru->bCommandReg = cdb[14]; 675 676 } else { /*ATA_12*/ 677 678 passthru->bFeaturesReg = cdb[3]; 679 passthru->bSectorCountReg = cdb[4]; 680 passthru->bLbaLowReg = cdb[5]; 681 passthru->bLbaMidReg = cdb[6]; 682 passthru->bLbaHighReg = cdb[7]; 683 passthru->bDriveHeadReg = cdb[8]; 684 passthru->bCommandReg = cdb[9]; 685 } 686 687 if (cdb[1] & 0xe0) { 688 689 690 if (!(passthru->bCommandReg == ATA_CMD_READ_MULTI || 691 passthru->bCommandReg == ATA_CMD_READ_MULTI_EXT || 692 passthru->bCommandReg == ATA_CMD_WRITE_MULTI || 693 passthru->bCommandReg == ATA_CMD_WRITE_MULTI_EXT || 694 passthru->bCommandReg == ATA_CMD_WRITE_MULTI_FUA_EXT) 695 ) { 696 goto error; 697 } 698 } 699 700 701 if (passthru->bFeaturesReg == ATA_SET_FEATURES_XFER && 702 passthru->bCommandReg == ATA_CMD_SET_FEATURES) { 703 goto error; 704 } 705 706 707 passthru->nSectors = ccb->csio.dxfer_len/ATA_SECTOR_SIZE; 708 switch (prot) { 709 default: /*None data*/ 710 break; 711 case 4: /*PIO data in, T_DIR=1 match check*/ 712 if ((cdb[2] & 3) && 713 (cdb[2] & 0x8) == 0) 714 { 715 OsPrint(("PIO data in, T_DIR=1 match check")); 716 goto error; 717 } 718 pCmd->flags.data_in = 1; 719 break; 720 case 5: /*PIO data out, T_DIR=0 match check*/ 721 if ((cdb[2] & 3) && 722 (cdb[2] & 0x8)) 723 { 724 OsPrint(("PIO data out, T_DIR=0 match check")); 725 goto error; 726 } 727 728 pCmd->flags.data_out = 1; 729 break; 730 } 731 pCmd->type = CMD_TYPE_PASSTHROUGH; 732 pCmd->priv = ext = cmdext_get(vbus_ext); 733 HPT_ASSERT(ext); 734 ext->ccb = ccb; 735 pCmd->target = vd; 736 pCmd->done = os_cmddone; 737 pCmd->buildsgl = os_buildsgl; 738 pCmd->psg = ext->psg; 739 740 if(!ccb->csio.dxfer_len) 741 { 742 ldm_queue_cmd(pCmd); 743 return; 744 } 745 pCmd->flags.physical_sg = 1; 746 error = bus_dmamap_load_ccb(vbus_ext->io_dmat, 747 ext->dma_map, ccb, 748 hpt_io_dmamap_callback, pCmd, 749 BUS_DMA_WAITOK 750 ); 751 KdPrint(("bus_dmamap_load return %d", error)); 752 if (error && error!=EINPROGRESS) { 753 os_printk("bus_dmamap_load error %d", error); 754 cmdext_put(ext); 755 ldm_free_cmds(pCmd); 756 ccb->ccb_h.status = CAM_REQ_CMP_ERR; 757 xpt_done(ccb); 758 } 759 return; 760 error: 761 ldm_free_cmds(pCmd); 762 ccb->ccb_h.status = CAM_REQ_INVALID; 763 break; 764 } 765 766 case INQUIRY: 767 { 768 PINQUIRYDATA inquiryData; 769 HIM_DEVICE_CONFIG devconf; 770 HPT_U8 *rbuf; 771 772 memset(ccb->csio.data_ptr, 0, ccb->csio.dxfer_len); 773 inquiryData = (PINQUIRYDATA)ccb->csio.data_ptr; 774 775 if (cdb[1] & 1) { 776 rbuf = (HPT_U8 *)inquiryData; 777 switch(cdb[2]) { 778 case 0: 779 rbuf[0] = 0; 780 rbuf[1] = 0; 781 rbuf[2] = 0; 782 rbuf[3] = 3; 783 rbuf[4] = 0; 784 rbuf[5] = 0x80; 785 rbuf[6] = 0x83; 786 ccb->ccb_h.status = CAM_REQ_CMP; 787 break; 788 case 0x80: { 789 rbuf[0] = 0; 790 rbuf[1] = 0x80; 791 rbuf[2] = 0; 792 if (vd->type == VD_RAW) { 793 rbuf[3] = 20; 794 vd->u.raw.him->get_device_config(vd->u.raw.phy_dev,&devconf); 795 memcpy(&rbuf[4], devconf.pIdentifyData->SerialNumber, 20); 796 ldm_ide_fixstring(&rbuf[4], 20); 797 } else { 798 rbuf[3] = 1; 799 rbuf[4] = 0x20; 800 } 801 ccb->ccb_h.status = CAM_REQ_CMP; 802 break; 803 } 804 case 0x83: 805 rbuf[0] = 0; 806 rbuf[1] = 0x83; 807 rbuf[2] = 0; 808 rbuf[3] = 12; 809 rbuf[4] = 1; 810 rbuf[5] = 2; 811 rbuf[6] = 0; 812 rbuf[7] = 8; 813 rbuf[8] = 0; 814 rbuf[9] = 0x19; 815 rbuf[10] = 0x3C; 816 rbuf[11] = 0; 817 rbuf[12] = 0; 818 rbuf[13] = 0; 819 rbuf[14] = 0; 820 rbuf[15] = 0; 821 ccb->ccb_h.status = CAM_REQ_CMP; 822 break; 823 default: 824 ccb->ccb_h.status = CAM_REQ_INVALID; 825 break; 826 } 827 828 break; 829 } 830 else if (cdb[2]) { 831 ccb->ccb_h.status = CAM_REQ_INVALID; 832 break; 833 } 834 835 inquiryData->DeviceType = 0; /*DIRECT_ACCESS_DEVICE*/ 836 inquiryData->Versions = 5; /*SPC-3*/ 837 inquiryData->ResponseDataFormat = 2; 838 inquiryData->AdditionalLength = 0x5b; 839 inquiryData->CommandQueue = 1; 840 841 if (ccb->csio.dxfer_len > 63) { 842 rbuf = (HPT_U8 *)inquiryData; 843 rbuf[58] = 0x60; 844 rbuf[59] = 0x3; 845 846 rbuf[64] = 0x3; 847 rbuf[66] = 0x3; 848 rbuf[67] = 0x20; 849 850 } 851 852 if (vd->type == VD_RAW) { 853 vd->u.raw.him->get_device_config(vd->u.raw.phy_dev,&devconf); 854 855 if ((devconf.pIdentifyData->GeneralConfiguration & 0x80)) 856 inquiryData->RemovableMedia = 1; 857 858 859 memcpy(&inquiryData->VendorId, "ATA ", 8); 860 memcpy(&inquiryData->ProductId, devconf.pIdentifyData->ModelNumber, 16); 861 ldm_ide_fixstring((HPT_U8 *)&inquiryData->ProductId, 16); 862 memcpy(&inquiryData->ProductRevisionLevel, devconf.pIdentifyData->FirmwareRevision, 4); 863 ldm_ide_fixstring((HPT_U8 *)&inquiryData->ProductRevisionLevel, 4); 864 if (inquiryData->ProductRevisionLevel[0] == 0 || inquiryData->ProductRevisionLevel[0] == ' ') 865 memcpy(&inquiryData->ProductRevisionLevel, "n/a ", 4); 866 } else { 867 memcpy(&inquiryData->VendorId, "HPT ", 8); 868 snprintf((char *)&inquiryData->ProductId, 16, "DISK_%d_%d ", 869 os_get_vbus_seq(vbus_ext), vd->target_id); 870 inquiryData->ProductId[15] = ' '; 871 memcpy(&inquiryData->ProductRevisionLevel, "4.00", 4); 872 } 873 874 ccb->ccb_h.status = CAM_REQ_CMP; 875 break; 876 } 877 case READ_CAPACITY: 878 { 879 HPT_U8 *rbuf = ccb->csio.data_ptr; 880 HPT_U32 cap; 881 HPT_U8 sector_size_shift = 0; 882 HPT_U64 new_cap; 883 HPT_U32 sector_size = 0; 884 885 if (mIsArray(vd->type)) 886 sector_size_shift = vd->u.array.sector_size_shift; 887 else{ 888 if(vd->type == VD_RAW){ 889 sector_size = vd->u.raw.logical_sector_size; 890 } 891 892 switch (sector_size) { 893 case 0x1000: 894 KdPrint(("set 4k setctor size in READ_CAPACITY")); 895 sector_size_shift = 3; 896 break; 897 default: 898 break; 899 } 900 } 901 new_cap = vd->capacity >> sector_size_shift; 902 903 if (new_cap > 0xfffffffful) 904 cap = 0xffffffff; 905 else 906 cap = new_cap - 1; 907 908 rbuf[0] = (HPT_U8)(cap>>24); 909 rbuf[1] = (HPT_U8)(cap>>16); 910 rbuf[2] = (HPT_U8)(cap>>8); 911 rbuf[3] = (HPT_U8)cap; 912 rbuf[4] = 0; 913 rbuf[5] = 0; 914 rbuf[6] = 2 << sector_size_shift; 915 rbuf[7] = 0; 916 917 ccb->ccb_h.status = CAM_REQ_CMP; 918 break; 919 } 920 921 case REPORT_LUNS: 922 { 923 HPT_U8 *rbuf = ccb->csio.data_ptr; 924 memset(rbuf, 0, 16); 925 rbuf[3] = 8; 926 ccb->ccb_h.status = CAM_REQ_CMP; 927 break; 928 } 929 case SERVICE_ACTION_IN: 930 { 931 HPT_U8 *rbuf = ccb->csio.data_ptr; 932 HPT_U64 cap = 0; 933 HPT_U8 sector_size_shift = 0; 934 HPT_U32 sector_size = 0; 935 936 if(mIsArray(vd->type)) 937 sector_size_shift = vd->u.array.sector_size_shift; 938 else{ 939 if(vd->type == VD_RAW){ 940 sector_size = vd->u.raw.logical_sector_size; 941 } 942 943 switch (sector_size) { 944 case 0x1000: 945 KdPrint(("set 4k setctor size in SERVICE_ACTION_IN")); 946 sector_size_shift = 3; 947 break; 948 default: 949 break; 950 } 951 } 952 cap = (vd->capacity >> sector_size_shift) - 1; 953 954 rbuf[0] = (HPT_U8)(cap>>56); 955 rbuf[1] = (HPT_U8)(cap>>48); 956 rbuf[2] = (HPT_U8)(cap>>40); 957 rbuf[3] = (HPT_U8)(cap>>32); 958 rbuf[4] = (HPT_U8)(cap>>24); 959 rbuf[5] = (HPT_U8)(cap>>16); 960 rbuf[6] = (HPT_U8)(cap>>8); 961 rbuf[7] = (HPT_U8)cap; 962 rbuf[8] = 0; 963 rbuf[9] = 0; 964 rbuf[10] = 2 << sector_size_shift; 965 rbuf[11] = 0; 966 967 if(!mIsArray(vd->type)){ 968 rbuf[13] = vd->u.raw.logicalsectors_per_physicalsector; 969 rbuf[14] = (HPT_U8)((vd->u.raw.lowest_aligned >> 8) & 0x3f); 970 rbuf[15] = (HPT_U8)(vd->u.raw.lowest_aligned); 971 } 972 973 ccb->ccb_h.status = CAM_REQ_CMP; 974 break; 975 } 976 977 case READ_6: 978 case READ_10: 979 case READ_16: 980 case WRITE_6: 981 case WRITE_10: 982 case WRITE_16: 983 case 0x13: 984 case 0x2f: 985 case 0x8f: /* VERIFY_16 */ 986 { 987 int error; 988 HPT_U8 sector_size_shift = 0; 989 HPT_U32 sector_size = 0; 990 pCmd = ldm_alloc_cmds(vbus, vd->cmds_per_request); 991 if(!pCmd){ 992 KdPrint(("Failed to allocate command!")); 993 ccb->ccb_h.status = CAM_BUSY; 994 break; 995 } 996 997 switch (cdb[0]) { 998 case READ_6: 999 case WRITE_6: 1000 case 0x13: 1001 pCmd->uCmd.Ide.Lba = ((HPT_U32)cdb[1] << 16) | ((HPT_U32)cdb[2] << 8) | (HPT_U32)cdb[3]; 1002 pCmd->uCmd.Ide.nSectors = (HPT_U16) cdb[4]; 1003 break; 1004 case READ_16: 1005 case WRITE_16: 1006 case 0x8f: /* VERIFY_16 */ 1007 { 1008 HPT_U64 block = 1009 ((HPT_U64)cdb[2]<<56) | 1010 ((HPT_U64)cdb[3]<<48) | 1011 ((HPT_U64)cdb[4]<<40) | 1012 ((HPT_U64)cdb[5]<<32) | 1013 ((HPT_U64)cdb[6]<<24) | 1014 ((HPT_U64)cdb[7]<<16) | 1015 ((HPT_U64)cdb[8]<<8) | 1016 ((HPT_U64)cdb[9]); 1017 pCmd->uCmd.Ide.Lba = block; 1018 pCmd->uCmd.Ide.nSectors = (HPT_U16)cdb[13] | ((HPT_U16)cdb[12]<<8); 1019 break; 1020 } 1021 1022 default: 1023 pCmd->uCmd.Ide.Lba = (HPT_U32)cdb[5] | ((HPT_U32)cdb[4] << 8) | ((HPT_U32)cdb[3] << 16) | ((HPT_U32)cdb[2] << 24); 1024 pCmd->uCmd.Ide.nSectors = (HPT_U16) cdb[8] | ((HPT_U16)cdb[7]<<8); 1025 break; 1026 } 1027 1028 if(mIsArray(vd->type)) { 1029 sector_size_shift = vd->u.array.sector_size_shift; 1030 } 1031 else{ 1032 if(vd->type == VD_RAW){ 1033 sector_size = vd->u.raw.logical_sector_size; 1034 } 1035 1036 switch (sector_size) { 1037 case 0x1000: 1038 KdPrint(("<8>resize sector size from 4k to 512")); 1039 sector_size_shift = 3; 1040 break; 1041 default: 1042 break; 1043 } 1044 } 1045 pCmd->uCmd.Ide.Lba <<= sector_size_shift; 1046 pCmd->uCmd.Ide.nSectors <<= sector_size_shift; 1047 1048 1049 switch (cdb[0]) { 1050 case READ_6: 1051 case READ_10: 1052 case READ_16: 1053 pCmd->flags.data_in = 1; 1054 break; 1055 case WRITE_6: 1056 case WRITE_10: 1057 case WRITE_16: 1058 pCmd->flags.data_out = 1; 1059 break; 1060 } 1061 pCmd->priv = ext = cmdext_get(vbus_ext); 1062 HPT_ASSERT(ext); 1063 ext->ccb = ccb; 1064 pCmd->target = vd; 1065 pCmd->done = os_cmddone; 1066 pCmd->buildsgl = os_buildsgl; 1067 pCmd->psg = ext->psg; 1068 pCmd->flags.physical_sg = 1; 1069 error = bus_dmamap_load_ccb(vbus_ext->io_dmat, 1070 ext->dma_map, ccb, 1071 hpt_io_dmamap_callback, pCmd, 1072 BUS_DMA_WAITOK 1073 ); 1074 KdPrint(("bus_dmamap_load return %d", error)); 1075 if (error && error!=EINPROGRESS) { 1076 os_printk("bus_dmamap_load error %d", error); 1077 cmdext_put(ext); 1078 ldm_free_cmds(pCmd); 1079 ccb->ccb_h.status = CAM_REQ_CMP_ERR; 1080 xpt_done(ccb); 1081 } 1082 return; 1083 } 1084 1085 default: 1086 ccb->ccb_h.status = CAM_REQ_INVALID; 1087 break; 1088 } 1089 1090 xpt_done(ccb); 1091 return; 1092 } 1093 1094 static void hpt_action(struct cam_sim *sim, union ccb *ccb) 1095 { 1096 PVBUS_EXT vbus_ext = (PVBUS_EXT)cam_sim_softc(sim); 1097 1098 KdPrint(("hpt_action(fn=%d, id=%d)", ccb->ccb_h.func_code, ccb->ccb_h.target_id)); 1099 1100 hpt_assert_vbus_locked(vbus_ext); 1101 switch (ccb->ccb_h.func_code) { 1102 1103 case XPT_SCSI_IO: 1104 hpt_scsi_io(vbus_ext, ccb); 1105 return; 1106 1107 case XPT_RESET_BUS: 1108 ldm_reset_vbus((PVBUS)vbus_ext->vbus); 1109 break; 1110 1111 case XPT_GET_TRAN_SETTINGS: 1112 case XPT_SET_TRAN_SETTINGS: 1113 ccb->ccb_h.status = CAM_FUNC_NOTAVAIL; 1114 break; 1115 1116 case XPT_CALC_GEOMETRY: 1117 ccb->ccg.heads = 255; 1118 ccb->ccg.secs_per_track = 63; 1119 ccb->ccg.cylinders = ccb->ccg.volume_size / (ccb->ccg.heads * ccb->ccg.secs_per_track); 1120 ccb->ccb_h.status = CAM_REQ_CMP; 1121 break; 1122 1123 case XPT_PATH_INQ: 1124 { 1125 struct ccb_pathinq *cpi = &ccb->cpi; 1126 1127 cpi->version_num = 1; 1128 cpi->hba_inquiry = PI_SDTR_ABLE; 1129 cpi->target_sprt = 0; 1130 cpi->hba_misc = PIM_NOBUSRESET; 1131 cpi->hba_eng_cnt = 0; 1132 cpi->max_target = osm_max_targets; 1133 cpi->max_lun = 0; 1134 cpi->unit_number = cam_sim_unit(sim); 1135 cpi->bus_id = cam_sim_bus(sim); 1136 cpi->initiator_id = osm_max_targets; 1137 cpi->base_transfer_speed = 3300; 1138 1139 strlcpy(cpi->sim_vid, "FreeBSD", SIM_IDLEN); 1140 strlcpy(cpi->hba_vid, "HPT ", HBA_IDLEN); 1141 strlcpy(cpi->dev_name, cam_sim_name(sim), DEV_IDLEN); 1142 cpi->transport = XPORT_SPI; 1143 cpi->transport_version = 2; 1144 cpi->protocol = PROTO_SCSI; 1145 cpi->protocol_version = SCSI_REV_2; 1146 cpi->ccb_h.status = CAM_REQ_CMP; 1147 break; 1148 } 1149 1150 default: 1151 ccb->ccb_h.status = CAM_REQ_INVALID; 1152 break; 1153 } 1154 1155 xpt_done(ccb); 1156 return; 1157 } 1158 1159 static void hpt_pci_intr(void *arg) 1160 { 1161 PVBUS_EXT vbus_ext = (PVBUS_EXT)arg; 1162 hpt_lock_vbus(vbus_ext); 1163 ldm_intr((PVBUS)vbus_ext->vbus); 1164 hpt_unlock_vbus(vbus_ext); 1165 } 1166 1167 static void hpt_poll(struct cam_sim *sim) 1168 { 1169 PVBUS_EXT vbus_ext = cam_sim_softc(sim); 1170 hpt_assert_vbus_locked(vbus_ext); 1171 ldm_intr((PVBUS)vbus_ext->vbus); 1172 } 1173 1174 static void hpt_async(void * callback_arg, u_int32_t code, struct cam_path * path, void * arg) 1175 { 1176 KdPrint(("hpt_async")); 1177 } 1178 1179 static int hpt_shutdown(device_t dev) 1180 { 1181 KdPrint(("hpt_shutdown(dev=%p)", dev)); 1182 return 0; 1183 } 1184 1185 static int hpt_detach(device_t dev) 1186 { 1187 /* we don't allow the driver to be unloaded. */ 1188 return EBUSY; 1189 } 1190 1191 static void hpt_ioctl_done(struct _IOCTL_ARG *arg) 1192 { 1193 arg->ioctl_cmnd = 0; 1194 wakeup(arg); 1195 } 1196 1197 static void __hpt_do_ioctl(PVBUS_EXT vbus_ext, IOCTL_ARG *ioctl_args) 1198 { 1199 ioctl_args->result = -1; 1200 ioctl_args->done = hpt_ioctl_done; 1201 ioctl_args->ioctl_cmnd = (void *)1; 1202 1203 hpt_lock_vbus(vbus_ext); 1204 ldm_ioctl((PVBUS)vbus_ext->vbus, ioctl_args); 1205 1206 while (ioctl_args->ioctl_cmnd) { 1207 if (hpt_sleep(vbus_ext, ioctl_args, PPAUSE, "hptctl", HPT_OSM_TIMEOUT)==0) 1208 break; 1209 ldm_reset_vbus((PVBUS)vbus_ext->vbus); 1210 __hpt_do_tasks(vbus_ext); 1211 } 1212 1213 /* KdPrint(("ioctl %x result %d", ioctl_args->dwIoControlCode, ioctl_args->result)); */ 1214 1215 hpt_unlock_vbus(vbus_ext); 1216 } 1217 1218 static void hpt_do_ioctl(IOCTL_ARG *ioctl_args) 1219 { 1220 PVBUS vbus; 1221 PVBUS_EXT vbus_ext; 1222 1223 ldm_for_each_vbus(vbus, vbus_ext) { 1224 __hpt_do_ioctl(vbus_ext, ioctl_args); 1225 if (ioctl_args->result!=HPT_IOCTL_RESULT_WRONG_VBUS) 1226 return; 1227 } 1228 } 1229 1230 #define HPT_DO_IOCTL(code, inbuf, insize, outbuf, outsize) ({\ 1231 IOCTL_ARG arg;\ 1232 arg.dwIoControlCode = code;\ 1233 arg.lpInBuffer = inbuf;\ 1234 arg.lpOutBuffer = outbuf;\ 1235 arg.nInBufferSize = insize;\ 1236 arg.nOutBufferSize = outsize;\ 1237 arg.lpBytesReturned = 0;\ 1238 hpt_do_ioctl(&arg);\ 1239 arg.result;\ 1240 }) 1241 1242 #define DEVICEID_VALID(id) ((id) && ((HPT_U32)(id)!=0xffffffff)) 1243 1244 static int hpt_get_logical_devices(DEVICEID * pIds, int nMaxCount) 1245 { 1246 int i; 1247 HPT_U32 count = nMaxCount-1; 1248 1249 if (HPT_DO_IOCTL(HPT_IOCTL_GET_LOGICAL_DEVICES, 1250 &count, sizeof(HPT_U32), pIds, sizeof(DEVICEID)*nMaxCount)) 1251 return -1; 1252 1253 nMaxCount = (int)pIds[0]; 1254 for (i=0; i<nMaxCount; i++) pIds[i] = pIds[i+1]; 1255 return nMaxCount; 1256 } 1257 1258 static int hpt_get_device_info_v3(DEVICEID id, PLOGICAL_DEVICE_INFO_V3 pInfo) 1259 { 1260 return HPT_DO_IOCTL(HPT_IOCTL_GET_DEVICE_INFO_V3, 1261 &id, sizeof(DEVICEID), pInfo, sizeof(LOGICAL_DEVICE_INFO_V3)); 1262 } 1263 1264 /* not belong to this file logically, but we want to use ioctl interface */ 1265 static int __hpt_stop_tasks(PVBUS_EXT vbus_ext, DEVICEID id) 1266 { 1267 LOGICAL_DEVICE_INFO_V3 devinfo; 1268 int i, result; 1269 DEVICEID param[2] = { id, 0 }; 1270 1271 if (hpt_get_device_info_v3(id, &devinfo)) 1272 return -1; 1273 1274 if (devinfo.Type!=LDT_ARRAY) 1275 return -1; 1276 1277 if (devinfo.u.array.Flags & ARRAY_FLAG_REBUILDING) 1278 param[1] = AS_REBUILD_ABORT; 1279 else if (devinfo.u.array.Flags & ARRAY_FLAG_VERIFYING) 1280 param[1] = AS_VERIFY_ABORT; 1281 else if (devinfo.u.array.Flags & ARRAY_FLAG_INITIALIZING) 1282 param[1] = AS_INITIALIZE_ABORT; 1283 else if (devinfo.u.array.Flags & ARRAY_FLAG_TRANSFORMING) 1284 param[1] = AS_TRANSFORM_ABORT; 1285 else 1286 return -1; 1287 1288 KdPrint(("SET_ARRAY_STATE(%x, %d)", param[0], param[1])); 1289 result = HPT_DO_IOCTL(HPT_IOCTL_SET_ARRAY_STATE, 1290 param, sizeof(param), 0, 0); 1291 1292 for (i=0; i<devinfo.u.array.nDisk; i++) 1293 if (DEVICEID_VALID(devinfo.u.array.Members[i])) 1294 __hpt_stop_tasks(vbus_ext, devinfo.u.array.Members[i]); 1295 1296 return result; 1297 } 1298 1299 static void hpt_stop_tasks(PVBUS_EXT vbus_ext) 1300 { 1301 DEVICEID ids[32]; 1302 int i, count; 1303 1304 count = hpt_get_logical_devices((DEVICEID *)&ids, sizeof(ids)/sizeof(ids[0])); 1305 1306 for (i=0; i<count; i++) 1307 __hpt_stop_tasks(vbus_ext, ids[i]); 1308 } 1309 1310 static d_open_t hpt_open; 1311 static d_close_t hpt_close; 1312 static d_ioctl_t hpt_ioctl; 1313 static int hpt_rescan_bus(void); 1314 1315 static struct cdevsw hpt_cdevsw = { 1316 .d_open = hpt_open, 1317 .d_close = hpt_close, 1318 .d_ioctl = hpt_ioctl, 1319 .d_name = driver_name, 1320 .d_version = D_VERSION, 1321 }; 1322 1323 static struct intr_config_hook hpt_ich; 1324 1325 /* 1326 * hpt_final_init will be called after all hpt_attach. 1327 */ 1328 static void hpt_final_init(void *dummy) 1329 { 1330 int i,unit_number=0; 1331 PVBUS_EXT vbus_ext; 1332 PVBUS vbus; 1333 PHBA hba; 1334 1335 /* Clear the config hook */ 1336 config_intrhook_disestablish(&hpt_ich); 1337 1338 /* allocate memory */ 1339 i = 0; 1340 ldm_for_each_vbus(vbus, vbus_ext) { 1341 if (hpt_alloc_mem(vbus_ext)) { 1342 os_printk("out of memory"); 1343 return; 1344 } 1345 i++; 1346 } 1347 1348 if (!i) { 1349 if (bootverbose) 1350 os_printk("no controller detected."); 1351 return; 1352 } 1353 1354 /* initializing hardware */ 1355 ldm_for_each_vbus(vbus, vbus_ext) { 1356 /* make timer available here */ 1357 mtx_init(&vbus_ext->lock, "hptsleeplock", NULL, MTX_DEF); 1358 callout_init_mtx(&vbus_ext->timer, &vbus_ext->lock, 0); 1359 if (hpt_init_vbus(vbus_ext)) { 1360 os_printk("fail to initialize hardware"); 1361 break; /* FIXME */ 1362 } 1363 } 1364 1365 /* register CAM interface */ 1366 ldm_for_each_vbus(vbus, vbus_ext) { 1367 struct cam_devq *devq; 1368 struct ccb_setasync ccb; 1369 1370 if (bus_dma_tag_create(NULL,/* parent */ 1371 4, /* alignment */ 1372 BUS_SPACE_MAXADDR_32BIT+1, /* boundary */ 1373 BUS_SPACE_MAXADDR, /* lowaddr */ 1374 BUS_SPACE_MAXADDR, /* highaddr */ 1375 NULL, NULL, /* filter, filterarg */ 1376 PAGE_SIZE * (os_max_sg_descriptors-1), /* maxsize */ 1377 os_max_sg_descriptors, /* nsegments */ 1378 0x10000, /* maxsegsize */ 1379 BUS_DMA_WAITOK, /* flags */ 1380 busdma_lock_mutex, /* lockfunc */ 1381 &vbus_ext->lock, /* lockfuncarg */ 1382 &vbus_ext->io_dmat /* tag */)) 1383 { 1384 return ; 1385 } 1386 1387 for (i=0; i<os_max_queue_comm; i++) { 1388 POS_CMDEXT ext = (POS_CMDEXT)malloc(sizeof(OS_CMDEXT), M_DEVBUF, M_WAITOK); 1389 if (!ext) { 1390 os_printk("Can't alloc cmdext(%d)", i); 1391 return ; 1392 } 1393 ext->vbus_ext = vbus_ext; 1394 ext->next = vbus_ext->cmdext_list; 1395 vbus_ext->cmdext_list = ext; 1396 1397 if (bus_dmamap_create(vbus_ext->io_dmat, 0, &ext->dma_map)) { 1398 os_printk("Can't create dma map(%d)", i); 1399 return ; 1400 } 1401 callout_init_mtx(&ext->timeout, &vbus_ext->lock, 0); 1402 } 1403 1404 if ((devq = cam_simq_alloc(os_max_queue_comm)) == NULL) { 1405 os_printk("cam_simq_alloc failed"); 1406 return ; 1407 } 1408 1409 hpt_lock_vbus(vbus_ext); 1410 vbus_ext->sim = cam_sim_alloc(hpt_action, hpt_poll, driver_name, 1411 vbus_ext, unit_number, &vbus_ext->lock, 1412 os_max_queue_comm, /*tagged*/8, devq); 1413 unit_number++; 1414 if (!vbus_ext->sim) { 1415 os_printk("cam_sim_alloc failed"); 1416 cam_simq_free(devq); 1417 hpt_unlock_vbus(vbus_ext); 1418 return ; 1419 } 1420 1421 if (xpt_bus_register(vbus_ext->sim, NULL, 0) != CAM_SUCCESS) { 1422 os_printk("xpt_bus_register failed"); 1423 cam_sim_free(vbus_ext->sim, /*free devq*/ TRUE); 1424 vbus_ext->sim = NULL; 1425 return ; 1426 } 1427 1428 if (xpt_create_path(&vbus_ext->path, /*periph */ NULL, 1429 cam_sim_path(vbus_ext->sim), CAM_TARGET_WILDCARD, 1430 CAM_LUN_WILDCARD) != CAM_REQ_CMP) 1431 { 1432 os_printk("xpt_create_path failed"); 1433 xpt_bus_deregister(cam_sim_path(vbus_ext->sim)); 1434 cam_sim_free(vbus_ext->sim, /*free_devq*/TRUE); 1435 hpt_unlock_vbus(vbus_ext); 1436 vbus_ext->sim = NULL; 1437 return ; 1438 } 1439 hpt_unlock_vbus(vbus_ext); 1440 1441 xpt_setup_ccb(&ccb.ccb_h, vbus_ext->path, /*priority*/5); 1442 ccb.ccb_h.func_code = XPT_SASYNC_CB; 1443 ccb.event_enable = AC_LOST_DEVICE; 1444 ccb.callback = hpt_async; 1445 ccb.callback_arg = vbus_ext; 1446 xpt_action((union ccb *)&ccb); 1447 1448 for (hba = vbus_ext->hba_list; hba; hba = hba->next) { 1449 int rid = 0; 1450 if ((hba->irq_res = bus_alloc_resource_any(hba->pcidev, 1451 SYS_RES_IRQ, &rid, RF_SHAREABLE | RF_ACTIVE)) == NULL) 1452 { 1453 os_printk("can't allocate interrupt"); 1454 return ; 1455 } 1456 1457 if (bus_setup_intr(hba->pcidev, hba->irq_res, INTR_TYPE_CAM | INTR_MPSAFE, 1458 NULL, hpt_pci_intr, vbus_ext, &hba->irq_handle)) 1459 { 1460 os_printk("can't set up interrupt"); 1461 return ; 1462 } 1463 hba->ldm_adapter.him->intr_control(hba->ldm_adapter.him_handle, HPT_TRUE); 1464 1465 } 1466 1467 vbus_ext->shutdown_eh = EVENTHANDLER_REGISTER(shutdown_final, 1468 hpt_shutdown_vbus, vbus_ext, SHUTDOWN_PRI_DEFAULT); 1469 if (!vbus_ext->shutdown_eh) 1470 os_printk("Shutdown event registration failed"); 1471 } 1472 1473 ldm_for_each_vbus(vbus, vbus_ext) { 1474 TASK_INIT(&vbus_ext->worker, 0, (task_fn_t *)hpt_do_tasks, vbus_ext); 1475 if (vbus_ext->tasks) 1476 TASK_ENQUEUE(&vbus_ext->worker); 1477 } 1478 1479 make_dev(&hpt_cdevsw, DRIVER_MINOR, UID_ROOT, GID_OPERATOR, 1480 S_IRUSR | S_IWUSR, "%s", driver_name); 1481 } 1482 1483 #if defined(KLD_MODULE) 1484 1485 typedef struct driverlink *driverlink_t; 1486 struct driverlink { 1487 kobj_class_t driver; 1488 TAILQ_ENTRY(driverlink) link; /* list of drivers in devclass */ 1489 }; 1490 1491 typedef TAILQ_HEAD(driver_list, driverlink) driver_list_t; 1492 1493 struct devclass { 1494 TAILQ_ENTRY(devclass) link; 1495 devclass_t parent; /* parent in devclass hierarchy */ 1496 driver_list_t drivers; /* bus devclasses store drivers for bus */ 1497 char *name; 1498 device_t *devices; /* array of devices indexed by unit */ 1499 int maxunit; /* size of devices array */ 1500 }; 1501 1502 static void override_kernel_driver(void) 1503 { 1504 driverlink_t dl, dlfirst; 1505 driver_t *tmpdriver; 1506 devclass_t dc = devclass_find("pci"); 1507 1508 if (dc){ 1509 dlfirst = TAILQ_FIRST(&dc->drivers); 1510 for (dl = dlfirst; dl; dl = TAILQ_NEXT(dl, link)) { 1511 if(strcmp(dl->driver->name, driver_name) == 0) { 1512 tmpdriver=dl->driver; 1513 dl->driver=dlfirst->driver; 1514 dlfirst->driver=tmpdriver; 1515 break; 1516 } 1517 } 1518 } 1519 } 1520 1521 #else 1522 #define override_kernel_driver() 1523 #endif 1524 1525 static void hpt_init(void *dummy) 1526 { 1527 if (bootverbose) 1528 os_printk("%s %s", driver_name_long, driver_ver); 1529 1530 override_kernel_driver(); 1531 init_config(); 1532 1533 hpt_ich.ich_func = hpt_final_init; 1534 hpt_ich.ich_arg = NULL; 1535 if (config_intrhook_establish(&hpt_ich) != 0) { 1536 printf("%s: cannot establish configuration hook\n", 1537 driver_name_long); 1538 } 1539 1540 } 1541 SYSINIT(hptinit, SI_SUB_CONFIGURE, SI_ORDER_FIRST, hpt_init, NULL); 1542 1543 /* 1544 * CAM driver interface 1545 */ 1546 static device_method_t driver_methods[] = { 1547 /* Device interface */ 1548 DEVMETHOD(device_probe, hpt_probe), 1549 DEVMETHOD(device_attach, hpt_attach), 1550 DEVMETHOD(device_detach, hpt_detach), 1551 DEVMETHOD(device_shutdown, hpt_shutdown), 1552 { 0, 0 } 1553 }; 1554 1555 static driver_t hpt_pci_driver = { 1556 driver_name, 1557 driver_methods, 1558 sizeof(HBA) 1559 }; 1560 1561 static devclass_t hpt_devclass; 1562 1563 #ifndef TARGETNAME 1564 #error "no TARGETNAME found" 1565 #endif 1566 1567 /* use this to make TARGETNAME be expanded */ 1568 #define __DRIVER_MODULE(p1, p2, p3, p4, p5, p6) DRIVER_MODULE(p1, p2, p3, p4, p5, p6) 1569 #define __MODULE_VERSION(p1, p2) MODULE_VERSION(p1, p2) 1570 #define __MODULE_DEPEND(p1, p2, p3, p4, p5) MODULE_DEPEND(p1, p2, p3, p4, p5) 1571 __DRIVER_MODULE(TARGETNAME, pci, hpt_pci_driver, hpt_devclass, 0, 0); 1572 __MODULE_VERSION(TARGETNAME, 1); 1573 __MODULE_DEPEND(TARGETNAME, cam, 1, 1, 1); 1574 1575 static int hpt_open(struct cdev *dev, int flags, int devtype, struct thread *td) 1576 { 1577 return 0; 1578 } 1579 1580 static int hpt_close(struct cdev *dev, int flags, int devtype, struct thread *td) 1581 { 1582 return 0; 1583 } 1584 1585 static int hpt_ioctl(struct cdev *dev, u_long cmd, caddr_t data, int fflag, struct thread *td) 1586 { 1587 PHPT_IOCTL_PARAM piop=(PHPT_IOCTL_PARAM)data; 1588 IOCTL_ARG ioctl_args; 1589 HPT_U32 bytesReturned = 0; 1590 1591 switch (cmd){ 1592 case HPT_DO_IOCONTROL: 1593 { 1594 if (piop->Magic == HPT_IOCTL_MAGIC || piop->Magic == HPT_IOCTL_MAGIC32) { 1595 KdPrint(("ioctl=%x in=%p len=%d out=%p len=%d\n", 1596 piop->dwIoControlCode, 1597 piop->lpInBuffer, 1598 piop->nInBufferSize, 1599 piop->lpOutBuffer, 1600 piop->nOutBufferSize)); 1601 1602 memset(&ioctl_args, 0, sizeof(ioctl_args)); 1603 1604 ioctl_args.dwIoControlCode = piop->dwIoControlCode; 1605 ioctl_args.nInBufferSize = piop->nInBufferSize; 1606 ioctl_args.nOutBufferSize = piop->nOutBufferSize; 1607 ioctl_args.lpBytesReturned = &bytesReturned; 1608 1609 if (ioctl_args.nInBufferSize) { 1610 ioctl_args.lpInBuffer = malloc(ioctl_args.nInBufferSize, M_DEVBUF, M_WAITOK); 1611 if (!ioctl_args.lpInBuffer) 1612 goto invalid; 1613 if (copyin((void*)piop->lpInBuffer, 1614 ioctl_args.lpInBuffer, piop->nInBufferSize)) 1615 goto invalid; 1616 } 1617 1618 if (ioctl_args.nOutBufferSize) { 1619 ioctl_args.lpOutBuffer = malloc(ioctl_args.nOutBufferSize, M_DEVBUF, M_WAITOK | M_ZERO); 1620 if (!ioctl_args.lpOutBuffer) 1621 goto invalid; 1622 } 1623 1624 hpt_do_ioctl(&ioctl_args); 1625 1626 if (ioctl_args.result==HPT_IOCTL_RESULT_OK) { 1627 if (piop->nOutBufferSize) { 1628 if (copyout(ioctl_args.lpOutBuffer, 1629 (void*)piop->lpOutBuffer, piop->nOutBufferSize)) 1630 goto invalid; 1631 } 1632 if (piop->lpBytesReturned) { 1633 if (copyout(&bytesReturned, 1634 (void*)piop->lpBytesReturned, sizeof(HPT_U32))) 1635 goto invalid; 1636 } 1637 if (ioctl_args.lpInBuffer) free(ioctl_args.lpInBuffer, M_DEVBUF); 1638 if (ioctl_args.lpOutBuffer) free(ioctl_args.lpOutBuffer, M_DEVBUF); 1639 return 0; 1640 } 1641 invalid: 1642 if (ioctl_args.lpInBuffer) free(ioctl_args.lpInBuffer, M_DEVBUF); 1643 if (ioctl_args.lpOutBuffer) free(ioctl_args.lpOutBuffer, M_DEVBUF); 1644 return EFAULT; 1645 } 1646 return EFAULT; 1647 } 1648 1649 case HPT_SCAN_BUS: 1650 { 1651 return hpt_rescan_bus(); 1652 } 1653 default: 1654 KdPrint(("invalid command!")); 1655 return EFAULT; 1656 } 1657 1658 } 1659 1660 static int hpt_rescan_bus(void) 1661 { 1662 union ccb *ccb; 1663 PVBUS vbus; 1664 PVBUS_EXT vbus_ext; 1665 1666 ldm_for_each_vbus(vbus, vbus_ext) { 1667 if ((ccb = xpt_alloc_ccb()) == NULL) 1668 { 1669 return(ENOMEM); 1670 } 1671 if (xpt_create_path(&ccb->ccb_h.path, NULL, cam_sim_path(vbus_ext->sim), 1672 CAM_TARGET_WILDCARD, CAM_LUN_WILDCARD) != CAM_REQ_CMP) 1673 { 1674 xpt_free_ccb(ccb); 1675 return(EIO); 1676 } 1677 xpt_rescan(ccb); 1678 } 1679 return(0); 1680 } 1681