1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 22 /* 23 * Copyright 2008 Sun Microsystems, Inc. All rights reserved. 24 * Use is subject to license terms. 25 */ 26 27 28 /* 29 * Direct Attached disk driver for SPARC machines. 30 */ 31 32 /* 33 * Includes, Declarations and Local Data 34 */ 35 #include <sys/dada/dada.h> 36 #include <sys/dkbad.h> 37 #include <sys/dklabel.h> 38 #include <sys/dkio.h> 39 #include <sys/cdio.h> 40 #include <sys/vtoc.h> 41 #include <sys/dada/targets/daddef.h> 42 #include <sys/dada/targets/dadpriv.h> 43 #include <sys/file.h> 44 #include <sys/stat.h> 45 #include <sys/kstat.h> 46 #include <sys/vtrace.h> 47 #include <sys/aio_req.h> 48 #include <sys/note.h> 49 #include <sys/cmlb.h> 50 51 /* 52 * Global Error Levels for Error Reporting 53 */ 54 int dcd_error_level = DCD_ERR_RETRYABLE; 55 /* 56 * Local Static Data 57 */ 58 59 static int dcd_io_time = DCD_IO_TIME; 60 static int dcd_retry_count = DCD_RETRY_COUNT; 61 #ifndef lint 62 static int dcd_report_pfa = 1; 63 #endif 64 static int dcd_rot_delay = 4; 65 static int dcd_poll_busycnt = DCD_POLL_TIMEOUT; 66 67 /* 68 * Local Function Prototypes 69 */ 70 71 static int dcdopen(dev_t *dev_p, int flag, int otyp, cred_t *cred_p); 72 static int dcdclose(dev_t dev, int flag, int otyp, cred_t *cred_p); 73 static int dcdstrategy(struct buf *bp); 74 static int dcddump(dev_t dev, caddr_t addr, daddr_t blkno, int nblk); 75 static int dcdioctl(dev_t, int, intptr_t, int, cred_t *, int *); 76 static int dcdread(dev_t dev, struct uio *uio, cred_t *cred_p); 77 static int dcdwrite(dev_t dev, struct uio *uio, cred_t *cred_p); 78 static int dcd_prop_op(dev_t, dev_info_t *, ddi_prop_op_t, int, 79 char *, caddr_t, int *); 80 static int dcdaread(dev_t dev, struct aio_req *aio, cred_t *cred_p); 81 static int dcdawrite(dev_t dev, struct aio_req *aio, cred_t *cred_p); 82 83 84 static void dcd_free_softstate(struct dcd_disk *un, dev_info_t *devi); 85 static int dcd_doattach(dev_info_t *devi, int (*f)()); 86 static int dcd_validate_geometry(struct dcd_disk *un); 87 static ddi_devid_t dcd_get_devid(struct dcd_disk *un); 88 static ddi_devid_t dcd_create_devid(struct dcd_disk *un); 89 static int dcd_make_devid_from_serial(struct dcd_disk *un); 90 static void dcd_validate_model_serial(char *str, int *retlen, int totallen); 91 static int dcd_read_deviceid(struct dcd_disk *un); 92 static int dcd_write_deviceid(struct dcd_disk *un); 93 static int dcd_poll(struct dcd_pkt *pkt); 94 static char *dcd_rname(int reason); 95 static void dcd_flush_cache(struct dcd_disk *un); 96 97 static int dcd_compute_dk_capacity(struct dcd_device *devp, 98 diskaddr_t *capacity); 99 static int dcd_send_lb_rw_cmd(dev_info_t *devinfo, void *bufaddr, 100 diskaddr_t start_block, size_t reqlength, uchar_t cmd); 101 102 static void dcdmin(struct buf *bp); 103 104 static int dcdioctl_cmd(dev_t, struct udcd_cmd *, 105 enum uio_seg, enum uio_seg); 106 107 static void dcdstart(struct dcd_disk *un); 108 static void dcddone_and_mutex_exit(struct dcd_disk *un, struct buf *bp); 109 static void make_dcd_cmd(struct dcd_disk *un, struct buf *bp, int (*f)()); 110 static void dcdudcdmin(struct buf *bp); 111 112 static int dcdrunout(caddr_t); 113 static int dcd_check_wp(dev_t dev); 114 static int dcd_unit_ready(dev_t dev); 115 static void dcd_handle_tran_busy(struct buf *bp, struct diskhd *dp, 116 struct dcd_disk *un); 117 static void dcdintr(struct dcd_pkt *pkt); 118 static int dcd_handle_incomplete(struct dcd_disk *un, struct buf *bp); 119 static void dcd_offline(struct dcd_disk *un, int bechatty); 120 static int dcd_ready_and_valid(dev_t dev, struct dcd_disk *un); 121 static void dcd_reset_disk(struct dcd_disk *un, struct dcd_pkt *pkt); 122 static void dcd_translate(struct dadkio_status32 *statp, struct udcd_cmd *cmdp); 123 static int dcdflushdone(struct buf *bp); 124 125 /* Function prototypes for cmlb */ 126 127 static int dcd_lb_rdwr(dev_info_t *devi, uchar_t cmd, void *bufaddr, 128 diskaddr_t start_block, size_t reqlength, void *tg_cookie); 129 130 static int dcd_lb_getphygeom(dev_info_t *devi, cmlb_geom_t *phygeomp); 131 static int dcd_lb_getinfo(dev_info_t *devi, int cmd, void *arg, 132 void *tg_cookie); 133 134 135 static cmlb_tg_ops_t dcd_lb_ops = { 136 TG_DK_OPS_VERSION_1, 137 dcd_lb_rdwr, 138 dcd_lb_getinfo 139 }; 140 141 /* 142 * Error and Logging Functions 143 */ 144 #ifndef lint 145 static void clean_print(dev_info_t *dev, char *label, uint_t level, 146 char *title, char *data, int len); 147 static void dcdrestart(void *arg); 148 #endif /* lint */ 149 150 static int dcd_check_error(struct dcd_disk *un, struct buf *bp); 151 152 /* 153 * Error statistics create/update functions 154 */ 155 static int dcd_create_errstats(struct dcd_disk *, int); 156 157 158 159 /*PRINTFLIKE4*/ 160 extern void dcd_log(dev_info_t *, char *, uint_t, const char *, ...) 161 __KPRINTFLIKE(4); 162 extern void makecommand(struct dcd_pkt *, int, uchar_t, uint32_t, 163 uchar_t, uint32_t, uchar_t, uchar_t); 164 165 166 /* 167 * Configuration Routines 168 */ 169 static int dcdinfo(dev_info_t *dip, ddi_info_cmd_t infocmd, void *arg, 170 void **result); 171 static int dcdprobe(dev_info_t *devi); 172 static int dcdattach(dev_info_t *devi, ddi_attach_cmd_t cmd); 173 static int dcddetach(dev_info_t *devi, ddi_detach_cmd_t cmd); 174 static int dcdreset(dev_info_t *dip, ddi_reset_cmd_t cmd); 175 static int dcd_dr_detach(dev_info_t *devi); 176 static int dcdpower(dev_info_t *devi, int component, int level); 177 178 static void *dcd_state; 179 static int dcd_max_instance; 180 static char *dcd_label = "dad"; 181 182 static char *diskokay = "disk okay\n"; 183 184 #if DEBUG || lint 185 #define DCDDEBUG 186 #endif 187 188 int dcd_test_flag = 0; 189 /* 190 * Debugging macros 191 */ 192 #ifdef DCDDEBUG 193 static int dcddebug = 0; 194 #define DEBUGGING (dcddebug > 1) 195 #define DAD_DEBUG if (dcddebug == 1) dcd_log 196 #define DAD_DEBUG2 if (dcddebug > 1) dcd_log 197 #else /* DCDDEBUG */ 198 #define dcddebug (0) 199 #define DEBUGGING (0) 200 #define DAD_DEBUG if (0) dcd_log 201 #define DAD_DEBUG2 if (0) dcd_log 202 #endif 203 204 /* 205 * we use pkt_private area for storing bp and retry_count 206 * XXX: Really is this usefull. 207 */ 208 struct dcd_pkt_private { 209 struct buf *dcdpp_bp; 210 short dcdpp_retry_count; 211 short dcdpp_victim_retry_count; 212 }; 213 214 215 _NOTE(SCHEME_PROTECTS_DATA("Unique per pkt", dcd_pkt_private buf)) 216 217 #define PP_LEN (sizeof (struct dcd_pkt_private)) 218 219 #define PKT_SET_BP(pkt, bp) \ 220 ((struct dcd_pkt_private *)pkt->pkt_private)->dcdpp_bp = bp 221 #define PKT_GET_BP(pkt) \ 222 (((struct dcd_pkt_private *)pkt->pkt_private)->dcdpp_bp) 223 224 225 #define PKT_SET_RETRY_CNT(pkt, n) \ 226 ((struct dcd_pkt_private *)pkt->pkt_private)->dcdpp_retry_count = n 227 228 #define PKT_GET_RETRY_CNT(pkt) \ 229 (((struct dcd_pkt_private *)pkt->pkt_private)->dcdpp_retry_count) 230 231 #define PKT_INCR_RETRY_CNT(pkt, n) \ 232 ((struct dcd_pkt_private *)pkt->pkt_private)->dcdpp_retry_count += n 233 234 #define PKT_SET_VICTIM_RETRY_CNT(pkt, n) \ 235 ((struct dcd_pkt_private *)pkt->pkt_private)->dcdpp_victim_retry_count \ 236 = n 237 238 #define PKT_GET_VICTIM_RETRY_CNT(pkt) \ 239 (((struct dcd_pkt_private *)pkt->pkt_private)->dcdpp_victim_retry_count) 240 #define PKT_INCR_VICTIM_RETRY_CNT(pkt, n) \ 241 ((struct dcd_pkt_private *)pkt->pkt_private)->dcdpp_victim_retry_count \ 242 += n 243 244 #define DISK_NOT_READY_RETRY_COUNT (dcd_retry_count / 2) 245 246 247 /* 248 * Urk! 249 */ 250 #define SET_BP_ERROR(bp, err) \ 251 bioerror(bp, err); 252 253 #define IOSP KSTAT_IO_PTR(un->un_stats) 254 #define IO_PARTITION_STATS un->un_pstats[DCDPART(bp->b_edev)] 255 #define IOSP_PARTITION KSTAT_IO_PTR(IO_PARTITION_STATS) 256 257 #define DCD_DO_KSTATS(un, kstat_function, bp) \ 258 ASSERT(mutex_owned(DCD_MUTEX)); \ 259 if (bp != un->un_sbufp) { \ 260 if (un->un_stats) { \ 261 kstat_function(IOSP); \ 262 } \ 263 if (IO_PARTITION_STATS) { \ 264 kstat_function(IOSP_PARTITION); \ 265 } \ 266 } 267 268 #define DCD_DO_ERRSTATS(un, x) \ 269 if (un->un_errstats) { \ 270 struct dcd_errstats *dtp; \ 271 dtp = (struct dcd_errstats *)un->un_errstats->ks_data; \ 272 dtp->x.value.ui32++; \ 273 } 274 275 #define GET_SOFT_STATE(dev) \ 276 struct dcd_disk *un; \ 277 int instance, part; \ 278 minor_t minor = getminor(dev); \ 279 \ 280 part = minor & DCDPART_MASK; \ 281 instance = minor >> DCDUNIT_SHIFT; \ 282 if ((un = ddi_get_soft_state(dcd_state, instance)) == NULL) \ 283 return (ENXIO); 284 285 #define LOGICAL_BLOCK_ALIGN(blkno, blknoshift) \ 286 (((blkno) & ((1 << (blknoshift)) - 1)) == 0) 287 288 /* 289 * After the following number of sectors, the cylinder number spills over 290 * 0xFFFF if sectors = 63 and heads = 16. 291 */ 292 #define NUM_SECTORS_32G 0x3EFFC10 293 294 /* 295 * Configuration Data 296 */ 297 298 /* 299 * Device driver ops vector 300 */ 301 302 static struct cb_ops dcd_cb_ops = { 303 dcdopen, /* open */ 304 dcdclose, /* close */ 305 dcdstrategy, /* strategy */ 306 nodev, /* print */ 307 dcddump, /* dump */ 308 dcdread, /* read */ 309 dcdwrite, /* write */ 310 dcdioctl, /* ioctl */ 311 nodev, /* devmap */ 312 nodev, /* mmap */ 313 nodev, /* segmap */ 314 nochpoll, /* poll */ 315 dcd_prop_op, /* cb_prop_op */ 316 0, /* streamtab */ 317 D_64BIT | D_MP | D_NEW, /* Driver compatibility flag */ 318 CB_REV, /* cb_rev */ 319 dcdaread, /* async I/O read entry point */ 320 dcdawrite /* async I/O write entry point */ 321 }; 322 323 static struct dev_ops dcd_ops = { 324 DEVO_REV, /* devo_rev, */ 325 0, /* refcnt */ 326 dcdinfo, /* info */ 327 nulldev, /* identify */ 328 dcdprobe, /* probe */ 329 dcdattach, /* attach */ 330 dcddetach, /* detach */ 331 dcdreset, /* reset */ 332 &dcd_cb_ops, /* driver operations */ 333 (struct bus_ops *)0, /* bus operations */ 334 dcdpower, /* power */ 335 ddi_quiesce_not_supported, /* devo_quiesce */ 336 }; 337 338 339 /* 340 * This is the loadable module wrapper. 341 */ 342 #include <sys/modctl.h> 343 344 static struct modldrv modldrv = { 345 &mod_driverops, /* Type of module. This one is a driver */ 346 "DAD Disk Driver", /* Name of the module. */ 347 &dcd_ops, /* driver ops */ 348 }; 349 350 351 352 static struct modlinkage modlinkage = { 353 MODREV_1, &modldrv, NULL 354 }; 355 356 /* 357 * the dcd_attach_mutex only protects dcd_max_instance in multi-threaded 358 * attach situations 359 */ 360 static kmutex_t dcd_attach_mutex; 361 362 int 363 _init(void) 364 { 365 int e; 366 367 if ((e = ddi_soft_state_init(&dcd_state, sizeof (struct dcd_disk), 368 DCD_MAXUNIT)) != 0) 369 return (e); 370 371 mutex_init(&dcd_attach_mutex, NULL, MUTEX_DRIVER, NULL); 372 e = mod_install(&modlinkage); 373 if (e != 0) { 374 mutex_destroy(&dcd_attach_mutex); 375 ddi_soft_state_fini(&dcd_state); 376 return (e); 377 } 378 379 return (e); 380 } 381 382 int 383 _fini(void) 384 { 385 int e; 386 387 if ((e = mod_remove(&modlinkage)) != 0) 388 return (e); 389 390 ddi_soft_state_fini(&dcd_state); 391 mutex_destroy(&dcd_attach_mutex); 392 393 return (e); 394 } 395 396 int 397 _info(struct modinfo *modinfop) 398 { 399 400 return (mod_info(&modlinkage, modinfop)); 401 } 402 403 static int 404 dcdprobe(dev_info_t *devi) 405 { 406 struct dcd_device *devp; 407 int rval = DDI_PROBE_PARTIAL; 408 int instance; 409 410 devp = ddi_get_driver_private(devi); 411 instance = ddi_get_instance(devi); 412 413 /* 414 * Keep a count of how many disks (ie. highest instance no) we have 415 * XXX currently not used but maybe useful later again 416 */ 417 mutex_enter(&dcd_attach_mutex); 418 if (instance > dcd_max_instance) 419 dcd_max_instance = instance; 420 mutex_exit(&dcd_attach_mutex); 421 422 DAD_DEBUG2(devp->dcd_dev, dcd_label, DCD_DEBUG, "dcdprobe:\n"); 423 424 if (ddi_get_soft_state(dcd_state, instance) != NULL) 425 return (DDI_PROBE_PARTIAL); 426 427 /* 428 * Turn around and call utility probe routine 429 * to see whether we actually have a disk at 430 */ 431 432 DAD_DEBUG2(devp->dcd_dev, dcd_label, DCD_DEBUG, 433 "dcdprobe: %x\n", dcd_probe(devp, NULL_FUNC)); 434 435 switch (dcd_probe(devp, NULL_FUNC)) { 436 default: 437 case DCDPROBE_NORESP: 438 case DCDPROBE_NONCCS: 439 case DCDPROBE_NOMEM: 440 case DCDPROBE_FAILURE: 441 case DCDPROBE_BUSY: 442 break; 443 444 case DCDPROBE_EXISTS: 445 /* 446 * Check whether it is a ATA device and then 447 * return SUCCESS. 448 */ 449 DAD_DEBUG2(devp->dcd_dev, dcd_label, DCD_DEBUG, 450 "config %x\n", devp->dcd_ident->dcd_config); 451 if ((devp->dcd_ident->dcd_config & ATAPI_DEVICE) == 0) { 452 if (devp->dcd_ident->dcd_config & ATANON_REMOVABLE) { 453 rval = DDI_PROBE_SUCCESS; 454 } else 455 rval = DDI_PROBE_FAILURE; 456 } else { 457 rval = DDI_PROBE_FAILURE; 458 } 459 break; 460 } 461 dcd_unprobe(devp); 462 463 DAD_DEBUG2(devp->dcd_dev, dcd_label, DCD_DEBUG, 464 "dcdprobe returns %x\n", rval); 465 466 return (rval); 467 } 468 469 470 /*ARGSUSED*/ 471 static int 472 dcdattach(dev_info_t *devi, ddi_attach_cmd_t cmd) 473 { 474 int instance, rval; 475 struct dcd_device *devp; 476 struct dcd_disk *un; 477 struct diskhd *dp; 478 char *pm_comp[] = 479 { "NAME=ide-disk", "0=standby", "1=idle", "2=active" }; 480 481 /* CONSTCOND */ 482 ASSERT(NO_COMPETING_THREADS); 483 484 485 devp = ddi_get_driver_private(devi); 486 instance = ddi_get_instance(devi); 487 DAD_DEBUG2(devp->dcd_dev, dcd_label, DCD_DEBUG, "Attach Started\n"); 488 489 switch (cmd) { 490 case DDI_ATTACH: 491 break; 492 493 case DDI_RESUME: 494 if (!(un = ddi_get_soft_state(dcd_state, instance))) 495 return (DDI_FAILURE); 496 mutex_enter(DCD_MUTEX); 497 Restore_state(un); 498 /* 499 * Restore the state which was saved to give the 500 * the right state in un_last_state 501 */ 502 un->un_last_state = un->un_save_state; 503 un->un_throttle = 2; 504 cv_broadcast(&un->un_suspend_cv); 505 /* 506 * Raise the power level of the device to active. 507 */ 508 mutex_exit(DCD_MUTEX); 509 (void) pm_raise_power(DCD_DEVINFO, 0, DCD_DEVICE_ACTIVE); 510 mutex_enter(DCD_MUTEX); 511 512 /* 513 * start unit - if this is a low-activity device 514 * commands in queue will have to wait until new 515 * commands come in, which may take awhile. 516 * Also, we specifically don't check un_ncmds 517 * because we know that there really are no 518 * commands in progress after the unit was suspended 519 * and we could have reached the throttle level, been 520 * suspended, and have no new commands coming in for 521 * awhile. Highly unlikely, but so is the low- 522 * activity disk scenario. 523 */ 524 dp = &un->un_utab; 525 if (dp->b_actf && (dp->b_forw == NULL)) { 526 dcdstart(un); 527 } 528 529 mutex_exit(DCD_MUTEX); 530 return (DDI_SUCCESS); 531 532 default: 533 return (DDI_FAILURE); 534 } 535 536 if (dcd_doattach(devi, SLEEP_FUNC) == DDI_FAILURE) { 537 return (DDI_FAILURE); 538 } 539 540 if (!(un = (struct dcd_disk *) 541 ddi_get_soft_state(dcd_state, instance))) { 542 return (DDI_FAILURE); 543 } 544 devp->dcd_private = (ataopaque_t)un; 545 546 /* 547 * Add a zero-length attribute to tell the world we support 548 * kernel ioctls (for layered drivers) 549 */ 550 (void) ddi_prop_create(DDI_DEV_T_NONE, devi, DDI_PROP_CANSLEEP, 551 DDI_KERNEL_IOCTL, NULL, 0); 552 553 /* 554 * Since the dad device does not have the 'reg' property, 555 * cpr will not call its DDI_SUSPEND/DDI_RESUME entries. 556 * The following code is to tell cpr that this device 557 * does need to be suspended and resumed. 558 */ 559 (void) ddi_prop_update_string(DDI_DEV_T_NONE, devi, 560 "pm-hardware-state", (caddr_t)"needs-suspend-resume"); 561 562 /* 563 * Initialize power management bookkeeping; 564 * Create components - In IDE case there are 3 levels and one 565 * component. The levels being - active, idle, standby. 566 */ 567 568 rval = ddi_prop_update_string_array(DDI_DEV_T_NONE, 569 devi, "pm-components", pm_comp, 4); 570 if (rval == DDI_PROP_SUCCESS) { 571 /* 572 * Ignore the return value of pm_raise_power 573 * Even if we check the return values and 574 * remove the property created above, PM 575 * framework will not honour the change after 576 * first call to pm_raise_power. Hence, the 577 * removal of that property does not help if 578 * pm_raise_power fails. 579 */ 580 (void) pm_raise_power(DCD_DEVINFO, 0, DCD_DEVICE_ACTIVE); 581 } 582 583 ddi_report_dev(devi); 584 585 cmlb_alloc_handle(&un->un_dklbhandle); 586 587 if (cmlb_attach(devi, 588 &dcd_lb_ops, 589 0, 590 0, 591 0, 592 DDI_NT_BLOCK_CHAN, 593 CMLB_FAKE_GEOM_LABEL_IOCTLS_VTOC8, 594 un->un_dklbhandle, 595 0) != 0) { 596 cmlb_free_handle(&un->un_dklbhandle); 597 dcd_free_softstate(un, devi); 598 return (DDI_FAILURE); 599 } 600 601 mutex_enter(DCD_MUTEX); 602 (void) dcd_validate_geometry(un); 603 604 /* Get devid; create a devid ONLY IF could not get ID */ 605 if (dcd_get_devid(un) == NULL) { 606 /* Create the fab'd devid */ 607 (void) dcd_create_devid(un); 608 } 609 mutex_exit(DCD_MUTEX); 610 611 return (DDI_SUCCESS); 612 } 613 614 static void 615 dcd_free_softstate(struct dcd_disk *un, dev_info_t *devi) 616 { 617 struct dcd_device *devp; 618 int instance = ddi_get_instance(devi); 619 620 devp = ddi_get_driver_private(devi); 621 622 if (un) { 623 sema_destroy(&un->un_semoclose); 624 cv_destroy(&un->un_sbuf_cv); 625 cv_destroy(&un->un_state_cv); 626 cv_destroy(&un->un_disk_busy_cv); 627 cv_destroy(&un->un_suspend_cv); 628 629 /* 630 * Deallocate command packet resources. 631 */ 632 if (un->un_sbufp) 633 freerbuf(un->un_sbufp); 634 if (un->un_dp) { 635 kmem_free((caddr_t)un->un_dp, sizeof (*un->un_dp)); 636 } 637 /* 638 * Unregister the devid and free devid resources allocated 639 */ 640 ddi_devid_unregister(DCD_DEVINFO); 641 if (un->un_devid) { 642 ddi_devid_free(un->un_devid); 643 un->un_devid = NULL; 644 } 645 646 /* 647 * Delete kstats. Kstats for non CD devices are deleted 648 * in dcdclose. 649 */ 650 if (un->un_stats) { 651 kstat_delete(un->un_stats); 652 } 653 654 } 655 656 /* 657 * Cleanup scsi_device resources. 658 */ 659 ddi_soft_state_free(dcd_state, instance); 660 devp->dcd_private = (ataopaque_t)0; 661 /* unprobe scsi device */ 662 dcd_unprobe(devp); 663 664 /* Remove properties created during attach */ 665 ddi_prop_remove_all(devi); 666 } 667 668 static int 669 dcddetach(dev_info_t *devi, ddi_detach_cmd_t cmd) 670 { 671 int instance; 672 struct dcd_disk *un; 673 clock_t wait_cmds_complete; 674 instance = ddi_get_instance(devi); 675 676 if (!(un = ddi_get_soft_state(dcd_state, instance))) 677 return (DDI_FAILURE); 678 679 switch (cmd) { 680 case DDI_DETACH: 681 return (dcd_dr_detach(devi)); 682 683 case DDI_SUSPEND: 684 mutex_enter(DCD_MUTEX); 685 if (un->un_state == DCD_STATE_SUSPENDED) { 686 mutex_exit(DCD_MUTEX); 687 return (DDI_SUCCESS); 688 } 689 un->un_throttle = 0; 690 /* 691 * Save the last state first 692 */ 693 un->un_save_state = un->un_last_state; 694 695 New_state(un, DCD_STATE_SUSPENDED); 696 697 /* 698 * wait till current operation completed. If we are 699 * in the resource wait state (with an intr outstanding) 700 * then we need to wait till the intr completes and 701 * starts the next cmd. We wait for 702 * DCD_WAIT_CMDS_COMPLETE seconds before failing the 703 * DDI_SUSPEND. 704 */ 705 wait_cmds_complete = ddi_get_lbolt(); 706 wait_cmds_complete += 707 DCD_WAIT_CMDS_COMPLETE * drv_usectohz(1000000); 708 709 while (un->un_ncmds) { 710 if (cv_timedwait(&un->un_disk_busy_cv, 711 DCD_MUTEX, wait_cmds_complete) == -1) { 712 /* 713 * commands Didn't finish in the 714 * specified time, fail the DDI_SUSPEND. 715 */ 716 DAD_DEBUG2(DCD_DEVINFO, dcd_label, 717 DCD_DEBUG, "dcddetach: SUSPEND " 718 "failed due to outstanding cmds\n"); 719 Restore_state(un); 720 mutex_exit(DCD_MUTEX); 721 return (DDI_FAILURE); 722 } 723 } 724 mutex_exit(DCD_MUTEX); 725 return (DDI_SUCCESS); 726 } 727 return (DDI_FAILURE); 728 } 729 730 /* 731 * The reset entry point gets invoked at the system shutdown time or through 732 * CPR code at system suspend. 733 * Will be flushing the cache and expect this to be last I/O operation to the 734 * disk before system reset/power off. 735 */ 736 /*ARGSUSED*/ 737 static int 738 dcdreset(dev_info_t *dip, ddi_reset_cmd_t cmd) 739 { 740 struct dcd_disk *un; 741 int instance; 742 743 instance = ddi_get_instance(dip); 744 745 if (!(un = ddi_get_soft_state(dcd_state, instance))) 746 return (DDI_FAILURE); 747 748 dcd_flush_cache(un); 749 750 return (DDI_SUCCESS); 751 } 752 753 754 static int 755 dcd_dr_detach(dev_info_t *devi) 756 { 757 struct dcd_device *devp; 758 struct dcd_disk *un; 759 760 /* 761 * Get scsi_device structure for this instance. 762 */ 763 if ((devp = ddi_get_driver_private(devi)) == NULL) 764 return (DDI_FAILURE); 765 766 /* 767 * Get dcd_disk structure containing target 'private' information 768 */ 769 un = (struct dcd_disk *)devp->dcd_private; 770 771 /* 772 * Verify there are NO outstanding commands issued to this device. 773 * ie, un_ncmds == 0. 774 * It's possible to have outstanding commands through the physio 775 * code path, even though everything's closed. 776 */ 777 #ifndef lint 778 _NOTE(COMPETING_THREADS_NOW); 779 #endif 780 mutex_enter(DCD_MUTEX); 781 if (un->un_ncmds) { 782 mutex_exit(DCD_MUTEX); 783 _NOTE(NO_COMPETING_THREADS_NOW); 784 return (DDI_FAILURE); 785 } 786 787 mutex_exit(DCD_MUTEX); 788 789 cmlb_detach(un->un_dklbhandle, 0); 790 cmlb_free_handle(&un->un_dklbhandle); 791 792 793 /* 794 * Lower the power state of the device 795 * i.e. the minimum power consumption state - sleep. 796 */ 797 (void) pm_lower_power(DCD_DEVINFO, 0, DCD_DEVICE_STANDBY); 798 799 _NOTE(NO_COMPETING_THREADS_NOW); 800 801 /* 802 * at this point there are no competing threads anymore 803 * release active MT locks and all device resources. 804 */ 805 dcd_free_softstate(un, devi); 806 807 return (DDI_SUCCESS); 808 } 809 810 static int 811 dcdpower(dev_info_t *devi, int component, int level) 812 { 813 struct dcd_pkt *pkt; 814 struct dcd_disk *un; 815 int instance; 816 uchar_t cmd; 817 818 819 instance = ddi_get_instance(devi); 820 821 if (!(un = ddi_get_soft_state(dcd_state, instance)) || 822 (DCD_DEVICE_STANDBY > level) || (level > DCD_DEVICE_ACTIVE) || 823 component != 0) { 824 return (DDI_FAILURE); 825 } 826 827 mutex_enter(DCD_MUTEX); 828 /* 829 * if there are active commands for the device or device will be 830 * active soon. At the same time there is request to lower power 831 * return failure. 832 */ 833 if ((un->un_ncmds) && (level != DCD_DEVICE_ACTIVE)) { 834 mutex_exit(DCD_MUTEX); 835 return (DDI_FAILURE); 836 } 837 838 if ((un->un_state == DCD_STATE_OFFLINE) || 839 (un->un_state == DCD_STATE_FATAL)) { 840 mutex_exit(DCD_MUTEX); 841 return (DDI_FAILURE); 842 } 843 844 if (level == DCD_DEVICE_ACTIVE) { 845 /* 846 * No need to fire any command, just set the state structure 847 * to indicate previous state and set the level to active 848 */ 849 un->un_power_level = DCD_DEVICE_ACTIVE; 850 if (un->un_state == DCD_STATE_PM_SUSPENDED) 851 Restore_state(un); 852 mutex_exit(DCD_MUTEX); 853 } else { 854 pkt = dcd_init_pkt(ROUTE, (struct dcd_pkt *)NULL, 855 NULL, (uint32_t)sizeof (struct dcd_cmd), 2, PP_LEN, 856 PKT_CONSISTENT, NULL_FUNC, NULL); 857 858 if (pkt == (struct dcd_pkt *)NULL) { 859 mutex_exit(DCD_MUTEX); 860 return (DDI_FAILURE); 861 } 862 863 switch (level) { 864 case DCD_DEVICE_IDLE: 865 cmd = ATA_IDLE_IMMEDIATE; 866 break; 867 868 case DCD_DEVICE_STANDBY: 869 cmd = ATA_STANDBY_IMMEDIATE; 870 break; 871 } 872 873 makecommand(pkt, 0, cmd, 0, 0, 0, NO_DATA_XFER, 0); 874 mutex_exit(DCD_MUTEX); 875 /* 876 * Issue the appropriate command 877 */ 878 if ((dcd_poll(pkt)) || (SCBP_C(pkt) != STATUS_GOOD)) { 879 dcd_destroy_pkt(pkt); 880 return (DDI_FAILURE); 881 } 882 dcd_destroy_pkt(pkt); 883 mutex_enter(DCD_MUTEX); 884 if (un->un_state != DCD_STATE_PM_SUSPENDED) 885 New_state(un, DCD_STATE_PM_SUSPENDED); 886 un->un_power_level = level; 887 mutex_exit(DCD_MUTEX); 888 } 889 890 return (DDI_SUCCESS); 891 } 892 893 static int 894 dcd_doattach(dev_info_t *devi, int (*canwait)()) 895 { 896 struct dcd_device *devp; 897 struct dcd_disk *un = (struct dcd_disk *)0; 898 int instance; 899 int km_flags = (canwait != NULL_FUNC)? KM_SLEEP : KM_NOSLEEP; 900 int rval; 901 char *prop_template = "target%x-dcd-options"; 902 int options; 903 char prop_str[32]; 904 int target; 905 diskaddr_t capacity; 906 907 devp = ddi_get_driver_private(devi); 908 909 /* 910 * Call the routine scsi_probe to do some of the dirty work. 911 * If the INQUIRY command succeeds, the field dcd_inq in the 912 * device structure will be filled in. The dcd_sense structure 913 * will also be allocated. 914 */ 915 916 switch (dcd_probe(devp, canwait)) { 917 default: 918 return (DDI_FAILURE); 919 920 case DCDPROBE_EXISTS: 921 if ((devp->dcd_ident->dcd_config & ATAPI_DEVICE) == 0) { 922 if (devp->dcd_ident->dcd_config & ATANON_REMOVABLE) { 923 rval = DDI_SUCCESS; 924 } else { 925 rval = DDI_FAILURE; 926 goto error; 927 } 928 } else { 929 rval = DDI_FAILURE; 930 goto error; 931 } 932 } 933 934 935 instance = ddi_get_instance(devp->dcd_dev); 936 937 if (ddi_soft_state_zalloc(dcd_state, instance) != DDI_SUCCESS) { 938 rval = DDI_FAILURE; 939 goto error; 940 } 941 942 un = ddi_get_soft_state(dcd_state, instance); 943 944 un->un_sbufp = getrbuf(km_flags); 945 if (un->un_sbufp == (struct buf *)NULL) { 946 rval = DDI_FAILURE; 947 goto error; 948 } 949 950 951 un->un_dcd = devp; 952 un->un_power_level = -1; 953 un->un_tgattribute.media_is_writable = 1; 954 955 sema_init(&un->un_semoclose, 1, NULL, SEMA_DRIVER, NULL); 956 cv_init(&un->un_sbuf_cv, NULL, CV_DRIVER, NULL); 957 cv_init(&un->un_state_cv, NULL, CV_DRIVER, NULL); 958 /* Initialize power management conditional variable */ 959 cv_init(&un->un_disk_busy_cv, NULL, CV_DRIVER, NULL); 960 cv_init(&un->un_suspend_cv, NULL, CV_DRIVER, NULL); 961 962 if (un->un_dp == 0) { 963 /* 964 * Assume CCS drive, assume parity, but call 965 * it a CDROM if it is a RODIRECT device. 966 */ 967 un->un_dp = (struct dcd_drivetype *) 968 kmem_zalloc(sizeof (struct dcd_drivetype), km_flags); 969 if (!un->un_dp) { 970 rval = DDI_FAILURE; 971 goto error; 972 } 973 if ((devp->dcd_ident->dcd_config & ATAPI_DEVICE) == 0) { 974 if (devp->dcd_ident->dcd_config & ATANON_REMOVABLE) { 975 un->un_dp->ctype = CTYPE_DISK; 976 } 977 } else { 978 rval = DDI_FAILURE; 979 goto error; 980 } 981 un->un_dp->name = "CCS"; 982 un->un_dp->options = 0; 983 } 984 985 /* 986 * Allow I/O requests at un_secsize offset in multiple of un_secsize. 987 */ 988 un->un_secsize = DEV_BSIZE; 989 990 /* 991 * If the device is not a removable media device, make sure that 992 * that the device is ready, by issuing the another identify but 993 * not needed. Get the capacity from identify data and store here. 994 */ 995 if (dcd_compute_dk_capacity(devp, &capacity) == 0) { 996 un->un_diskcapacity = capacity; 997 un->un_lbasize = DEV_BSIZE; 998 } 999 1000 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, "Geometry Data\n"); 1001 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, "cyls %x, heads %x", 1002 devp->dcd_ident->dcd_fixcyls, 1003 devp->dcd_ident->dcd_heads); 1004 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, "sectors %x,", 1005 devp->dcd_ident->dcd_sectors); 1006 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, "capacity %llx\n", 1007 capacity); 1008 1009 DAD_DEBUG(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1010 "dcdprobe: drive selected\n"); 1011 1012 /* 1013 * Check for the property target<n>-dcd-options to find the option 1014 * set by the HBA driver for this target so that we can set the 1015 * Unit structure variable so that we can send commands accordingly. 1016 */ 1017 target = devp->dcd_address->a_target; 1018 (void) sprintf(prop_str, prop_template, target); 1019 options = ddi_prop_get_int(DDI_DEV_T_ANY, devi, DDI_PROP_NOTPROM, 1020 prop_str, -1); 1021 if (options < 0) { 1022 DAD_DEBUG(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1023 "No per target properties"); 1024 } else { 1025 if ((options & DCD_DMA_MODE) == DCD_DMA_MODE) { 1026 un->un_dp->options |= DMA_SUPPORTTED; 1027 un->un_dp->dma_mode = (options >> 3) & 0x03; 1028 DAD_DEBUG(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1029 "mode %x\n", un->un_dp->dma_mode); 1030 } else { 1031 un->un_dp->options &= ~DMA_SUPPORTTED; 1032 un->un_dp->pio_mode = options & 0x7; 1033 if (options & DCD_BLOCK_MODE) 1034 un->un_dp->options |= BLOCK_MODE; 1035 DAD_DEBUG(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1036 "mode %x\n", un->un_dp->pio_mode); 1037 } 1038 DAD_DEBUG(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1039 "options %x,", un->un_dp->options); 1040 } 1041 1042 un->un_throttle = 2; 1043 /* 1044 * set default max_xfer_size - This should depend on whether the 1045 * Block mode is supported by the device or not. 1046 */ 1047 un->un_max_xfer_size = MAX_ATA_XFER_SIZE; 1048 1049 /* 1050 * Set write cache enable softstate 1051 * 1052 * WCE is only supported in ATAPI-4 or higher; for 1053 * lower rev devices, must assume write cache is 1054 * enabled. 1055 */ 1056 mutex_enter(DCD_MUTEX); 1057 un->un_write_cache_enabled = (devp->dcd_ident->dcd_majvers == 0xffff) || 1058 ((devp->dcd_ident->dcd_majvers & IDENTIFY_80_ATAPI_4) == 0) || 1059 (devp->dcd_ident->dcd_features85 & IDENTIFY_85_WCE) != 0; 1060 mutex_exit(DCD_MUTEX); 1061 1062 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1063 "dcd_doattach returns good\n"); 1064 1065 return (rval); 1066 1067 error: 1068 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, "dcd_doattach failed\n"); 1069 dcd_free_softstate(un, devi); 1070 return (rval); 1071 } 1072 1073 #ifdef NOTNEEDED 1074 /* 1075 * This routine is used to set the block mode of operation by issuing the 1076 * Set Block mode ata command with the maximum block mode possible 1077 */ 1078 dcd_set_multiple(struct dcd_disk *un) 1079 { 1080 int status; 1081 struct udcd_cmd ucmd; 1082 struct dcd_cmd cdb; 1083 dev_t dev; 1084 1085 1086 /* Zero all the required structure */ 1087 (void) bzero((caddr_t)&ucmd, sizeof (ucmd)); 1088 1089 (void) bzero((caddr_t)&cdb, sizeof (struct dcd_cmd)); 1090 1091 cdb.cmd = ATA_SET_MULTIPLE; 1092 /* 1093 * Here we should pass what needs to go into sector count REGISTER. 1094 * Eventhough this field indicates the number of bytes to read we 1095 * need to specify the block factor in terms of bytes so that it 1096 * will be programmed by the HBA driver into the sector count register. 1097 */ 1098 cdb.size = un->un_lbasize * un->un_dp->block_factor; 1099 1100 cdb.sector_num.lba_num = 0; 1101 cdb.address_mode = ADD_LBA_MODE; 1102 cdb.direction = NO_DATA_XFER; 1103 1104 ucmd.udcd_flags = 0; 1105 ucmd.udcd_cmd = &cdb; 1106 ucmd.udcd_bufaddr = NULL; 1107 ucmd.udcd_buflen = 0; 1108 ucmd.udcd_flags |= UDCD_SILENT; 1109 1110 dev = makedevice(ddi_driver_major(DCD_DEVINFO), 1111 ddi_get_instance(DCD_DEVINFO) << DCDUNIT_SHIFT); 1112 1113 1114 status = dcdioctl_cmd(dev, &ucmd, UIO_SYSSPACE, UIO_SYSSPACE); 1115 1116 return (status); 1117 } 1118 /* 1119 * The following routine is used only for setting the transfer mode 1120 * and it is not designed for transferring any other features subcommand. 1121 */ 1122 dcd_set_features(struct dcd_disk *un, uchar_t mode) 1123 { 1124 int status; 1125 struct udcd_cmd ucmd; 1126 struct dcd_cmd cdb; 1127 dev_t dev; 1128 1129 1130 /* Zero all the required structure */ 1131 (void) bzero((caddr_t)&ucmd, sizeof (ucmd)); 1132 1133 (void) bzero((caddr_t)&cdb, sizeof (struct dcd_cmd)); 1134 1135 cdb.cmd = ATA_SET_FEATURES; 1136 /* 1137 * Here we need to pass what needs to go into the sector count register 1138 * But in the case of SET FEATURES command the value taken in the 1139 * sector count register depends what type of subcommand is 1140 * passed in the features register. Since we have defined the size to 1141 * be the size in bytes in this context it does not indicate bytes 1142 * instead it indicates the mode to be programmed. 1143 */ 1144 cdb.size = un->un_lbasize * mode; 1145 1146 cdb.sector_num.lba_num = 0; 1147 cdb.address_mode = ADD_LBA_MODE; 1148 cdb.direction = NO_DATA_XFER; 1149 cdb.features = ATA_FEATURE_SET_MODE; 1150 DAD_DEBUG(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1151 "size %x, features %x, cmd %x\n", 1152 cdb.size, cdb.features, cdb.cmd); 1153 1154 ucmd.udcd_flags = 0; 1155 ucmd.udcd_cmd = &cdb; 1156 ucmd.udcd_bufaddr = NULL; 1157 ucmd.udcd_buflen = 0; 1158 ucmd.udcd_flags |= UDCD_SILENT; 1159 1160 dev = makedevice(ddi_driver_major(DCD_DEVINFO), 1161 ddi_get_instance(DCD_DEVINFO) << DCDUNIT_SHIFT); 1162 1163 status = dcdioctl_cmd(dev, &ucmd, UIO_SYSSPACE, UIO_SYSSPACE); 1164 1165 return (status); 1166 } 1167 #endif 1168 1169 /* 1170 * Validate the geometry for this disk, e.g., 1171 * see whether it has a valid label. 1172 */ 1173 static int 1174 dcd_validate_geometry(struct dcd_disk *un) 1175 { 1176 int secsize = 0; 1177 struct dcd_device *devp; 1178 int secdiv; 1179 int rval; 1180 1181 ASSERT(mutex_owned(DCD_MUTEX)); 1182 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1183 "dcd_validate_geometry: started \n"); 1184 1185 if (un->un_lbasize < 0) { 1186 return (DCD_BAD_LABEL); 1187 } 1188 1189 if (un->un_state == DCD_STATE_PM_SUSPENDED) { 1190 mutex_exit(DCD_MUTEX); 1191 if (pm_raise_power(DCD_DEVINFO, 0, DCD_DEVICE_ACTIVE) != 1192 DDI_SUCCESS) { 1193 mutex_enter(DCD_MUTEX); 1194 return (DCD_BAD_LABEL); 1195 } 1196 mutex_enter(DCD_MUTEX); 1197 } 1198 1199 secsize = un->un_secsize; 1200 1201 /* 1202 * take a log base 2 of sector size (sorry) 1203 */ 1204 for (secdiv = 0; secsize = secsize >> 1; secdiv++) 1205 ; 1206 un->un_secdiv = secdiv; 1207 1208 /* 1209 * Only DIRECT ACCESS devices will have Sun labels. 1210 * CD's supposedly have a Sun label, too 1211 */ 1212 1213 devp = un->un_dcd; 1214 1215 if (((devp->dcd_ident->dcd_config & ATAPI_DEVICE) == 0) && 1216 (devp->dcd_ident->dcd_config & ATANON_REMOVABLE)) { 1217 mutex_exit(DCD_MUTEX); 1218 rval = cmlb_validate(un->un_dklbhandle, 0, 0); 1219 mutex_enter(DCD_MUTEX); 1220 if (rval == ENOMEM) 1221 return (DCD_NO_MEM_FOR_LABEL); 1222 else if (rval != 0) 1223 return (DCD_BAD_LABEL); 1224 } else { 1225 /* it should never get here. */ 1226 return (DCD_BAD_LABEL); 1227 } 1228 1229 /* 1230 * take a log base 2 of logical block size 1231 */ 1232 secsize = un->un_lbasize; 1233 for (secdiv = 0; secsize = secsize >> 1; secdiv++) 1234 ; 1235 un->un_lbadiv = secdiv; 1236 1237 /* 1238 * take a log base 2 of the multiple of DEV_BSIZE blocks that 1239 * make up one logical block 1240 */ 1241 secsize = un->un_lbasize >> DEV_BSHIFT; 1242 for (secdiv = 0; secsize = secsize >> 1; secdiv++) 1243 ; 1244 un->un_blknoshift = secdiv; 1245 return (0); 1246 } 1247 1248 /* 1249 * Unix Entry Points 1250 */ 1251 1252 /* ARGSUSED3 */ 1253 static int 1254 dcdopen(dev_t *dev_p, int flag, int otyp, cred_t *cred_p) 1255 { 1256 dev_t dev = *dev_p; 1257 int rval = EIO; 1258 int partmask; 1259 int nodelay = (flag & (FNDELAY | FNONBLOCK)); 1260 int i; 1261 char kstatname[KSTAT_STRLEN]; 1262 diskaddr_t lblocks; 1263 char *partname; 1264 1265 GET_SOFT_STATE(dev); 1266 1267 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1268 "Inside Open flag %x, otyp %x\n", flag, otyp); 1269 1270 if (otyp >= OTYPCNT) { 1271 return (EINVAL); 1272 } 1273 1274 partmask = 1 << part; 1275 1276 /* 1277 * We use a semaphore here in order to serialize 1278 * open and close requests on the device. 1279 */ 1280 sema_p(&un->un_semoclose); 1281 1282 mutex_enter(DCD_MUTEX); 1283 1284 if ((un->un_state & DCD_STATE_FATAL) == DCD_STATE_FATAL) { 1285 rval = ENXIO; 1286 goto done; 1287 } 1288 1289 while (un->un_state == DCD_STATE_SUSPENDED) { 1290 cv_wait(&un->un_suspend_cv, DCD_MUTEX); 1291 } 1292 1293 if ((un->un_state == DCD_STATE_PM_SUSPENDED) && (!nodelay)) { 1294 mutex_exit(DCD_MUTEX); 1295 if (pm_raise_power(DCD_DEVINFO, 0, DCD_DEVICE_ACTIVE) 1296 != DDI_SUCCESS) { 1297 mutex_enter(DCD_MUTEX); 1298 rval = EIO; 1299 goto done; 1300 } 1301 mutex_enter(DCD_MUTEX); 1302 } 1303 1304 /* 1305 * set make_dcd_cmd() flags and stat_size here since these 1306 * are unlikely to change 1307 */ 1308 un->un_cmd_flags = 0; 1309 1310 un->un_cmd_stat_size = 2; 1311 1312 DAD_DEBUG(DCD_DEVINFO, dcd_label, DCD_DEBUG, "dcdopen un=0x%p\n", 1313 (void *)un); 1314 /* 1315 * check for previous exclusive open 1316 */ 1317 DAD_DEBUG(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1318 "exclopen=%x, flag=%x, regopen=%x\n", 1319 un->un_exclopen, flag, un->un_ocmap.regopen[otyp]); 1320 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1321 "Exclusive open flag %x, partmask %x\n", 1322 un->un_exclopen, partmask); 1323 1324 if (un->un_exclopen & (partmask)) { 1325 failed_exclusive: 1326 DAD_DEBUG(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1327 "exclusive open fails\n"); 1328 rval = EBUSY; 1329 goto done; 1330 } 1331 1332 if (flag & FEXCL) { 1333 int i; 1334 if (un->un_ocmap.lyropen[part]) { 1335 goto failed_exclusive; 1336 } 1337 for (i = 0; i < (OTYPCNT - 1); i++) { 1338 if (un->un_ocmap.regopen[i] & (partmask)) { 1339 goto failed_exclusive; 1340 } 1341 } 1342 } 1343 if (flag & FWRITE) { 1344 mutex_exit(DCD_MUTEX); 1345 if (dcd_check_wp(dev)) { 1346 sema_v(&un->un_semoclose); 1347 return (EROFS); 1348 } 1349 mutex_enter(DCD_MUTEX); 1350 } 1351 1352 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1353 "Check Write Protect handled\n"); 1354 1355 if (!nodelay) { 1356 mutex_exit(DCD_MUTEX); 1357 if ((rval = dcd_ready_and_valid(dev, un)) != 0) { 1358 rval = EIO; 1359 } 1360 (void) pm_idle_component(DCD_DEVINFO, 0); 1361 /* 1362 * Fail if device is not ready or if the number of disk 1363 * blocks is zero or negative for non CD devices. 1364 */ 1365 if (rval || cmlb_partinfo(un->un_dklbhandle, 1366 part, &lblocks, NULL, &partname, NULL, 0) || 1367 lblocks <= 0) { 1368 rval = EIO; 1369 mutex_enter(DCD_MUTEX); 1370 goto done; 1371 } 1372 mutex_enter(DCD_MUTEX); 1373 } 1374 1375 if (otyp == OTYP_LYR) { 1376 un->un_ocmap.lyropen[part]++; 1377 } else { 1378 un->un_ocmap.regopen[otyp] |= partmask; 1379 } 1380 1381 /* 1382 * set up open and exclusive open flags 1383 */ 1384 if (flag & FEXCL) { 1385 un->un_exclopen |= (partmask); 1386 } 1387 1388 1389 DAD_DEBUG(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1390 "open of part %d type %d\n", 1391 part, otyp); 1392 1393 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1394 "Kstats getting updated\n"); 1395 /* 1396 * only create kstats for disks, CD kstats created in dcdattach 1397 */ 1398 _NOTE(NO_COMPETING_THREADS_NOW); 1399 mutex_exit(DCD_MUTEX); 1400 if (un->un_stats == (kstat_t *)0) { 1401 un->un_stats = kstat_create("dad", instance, 1402 NULL, "disk", KSTAT_TYPE_IO, 1, 1403 KSTAT_FLAG_PERSISTENT); 1404 if (un->un_stats) { 1405 un->un_stats->ks_lock = DCD_MUTEX; 1406 kstat_install(un->un_stats); 1407 } 1408 1409 /* 1410 * set up partition statistics for each partition 1411 * with number of blocks > 0 1412 */ 1413 if (!nodelay) { 1414 for (i = 0; i < NDKMAP; i++) { 1415 if ((un->un_pstats[i] == (kstat_t *)0) && 1416 (cmlb_partinfo(un->un_dklbhandle, 1417 i, &lblocks, NULL, &partname, 1418 NULL, 0) == 0) && lblocks > 0) { 1419 (void) sprintf(kstatname, "dad%d,%s", 1420 instance, partname); 1421 un->un_pstats[i] = kstat_create("dad", 1422 instance, 1423 kstatname, 1424 "partition", 1425 KSTAT_TYPE_IO, 1426 1, 1427 KSTAT_FLAG_PERSISTENT); 1428 if (un->un_pstats[i]) { 1429 un->un_pstats[i]->ks_lock = 1430 DCD_MUTEX; 1431 kstat_install(un->un_pstats[i]); 1432 } 1433 } 1434 } 1435 } 1436 /* 1437 * set up error kstats 1438 */ 1439 (void) dcd_create_errstats(un, instance); 1440 } 1441 #ifndef lint 1442 _NOTE(COMPETING_THREADS_NOW); 1443 #endif 1444 1445 sema_v(&un->un_semoclose); 1446 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, "Open success\n"); 1447 return (0); 1448 1449 done: 1450 mutex_exit(DCD_MUTEX); 1451 sema_v(&un->un_semoclose); 1452 return (rval); 1453 1454 } 1455 1456 /* 1457 * Test if disk is ready and has a valid geometry. 1458 */ 1459 static int 1460 dcd_ready_and_valid(dev_t dev, struct dcd_disk *un) 1461 { 1462 int rval = 1; 1463 int g_error = 0; 1464 1465 mutex_enter(DCD_MUTEX); 1466 /* 1467 * cmds outstanding 1468 */ 1469 if (un->un_ncmds == 0) { 1470 (void) dcd_unit_ready(dev); 1471 } 1472 1473 /* 1474 * If device is not yet ready here, inform it is offline 1475 */ 1476 if (un->un_state == DCD_STATE_NORMAL) { 1477 rval = dcd_unit_ready(dev); 1478 if (rval != 0 && rval != EACCES) { 1479 dcd_offline(un, 1); 1480 goto done; 1481 } 1482 } 1483 1484 if (un->un_format_in_progress == 0) { 1485 g_error = dcd_validate_geometry(un); 1486 } 1487 1488 /* 1489 * check if geometry was valid. We don't check the validity of 1490 * geometry for CDROMS. 1491 */ 1492 1493 if (g_error == DCD_BAD_LABEL) { 1494 rval = 1; 1495 goto done; 1496 } 1497 1498 1499 /* 1500 * the state has changed; inform the media watch routines 1501 */ 1502 un->un_mediastate = DKIO_INSERTED; 1503 cv_broadcast(&un->un_state_cv); 1504 rval = 0; 1505 1506 done: 1507 mutex_exit(DCD_MUTEX); 1508 return (rval); 1509 } 1510 1511 1512 /*ARGSUSED*/ 1513 static int 1514 dcdclose(dev_t dev, int flag, int otyp, cred_t *cred_p) 1515 { 1516 uchar_t *cp; 1517 int i; 1518 1519 GET_SOFT_STATE(dev); 1520 1521 1522 if (otyp >= OTYPCNT) 1523 return (ENXIO); 1524 1525 DAD_DEBUG(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1526 "close of part %d type %d\n", 1527 part, otyp); 1528 sema_p(&un->un_semoclose); 1529 1530 mutex_enter(DCD_MUTEX); 1531 1532 if (un->un_exclopen & (1<<part)) { 1533 un->un_exclopen &= ~(1<<part); 1534 } 1535 1536 if (otyp == OTYP_LYR) { 1537 un->un_ocmap.lyropen[part] -= 1; 1538 } else { 1539 un->un_ocmap.regopen[otyp] &= ~(1<<part); 1540 } 1541 1542 cp = &un->un_ocmap.chkd[0]; 1543 while (cp < &un->un_ocmap.chkd[OCSIZE]) { 1544 if (*cp != (uchar_t)0) { 1545 break; 1546 } 1547 cp++; 1548 } 1549 1550 if (cp == &un->un_ocmap.chkd[OCSIZE]) { 1551 DAD_DEBUG(DCD_DEVINFO, dcd_label, DCD_DEBUG, "last close\n"); 1552 if (un->un_state == DCD_STATE_OFFLINE) { 1553 dcd_offline(un, 1); 1554 } 1555 1556 mutex_exit(DCD_MUTEX); 1557 (void) cmlb_close(un->un_dklbhandle, 0); 1558 1559 _NOTE(NO_COMPETING_THREADS_NOW); 1560 if (un->un_stats) { 1561 kstat_delete(un->un_stats); 1562 un->un_stats = 0; 1563 } 1564 for (i = 0; i < NDKMAP; i++) { 1565 if (un->un_pstats[i]) { 1566 kstat_delete(un->un_pstats[i]); 1567 un->un_pstats[i] = (kstat_t *)0; 1568 } 1569 } 1570 1571 if (un->un_errstats) { 1572 kstat_delete(un->un_errstats); 1573 un->un_errstats = (kstat_t *)0; 1574 } 1575 mutex_enter(DCD_MUTEX); 1576 1577 #ifndef lint 1578 _NOTE(COMPETING_THREADS_NOW); 1579 #endif 1580 } 1581 1582 mutex_exit(DCD_MUTEX); 1583 sema_v(&un->un_semoclose); 1584 return (0); 1585 } 1586 1587 static void 1588 dcd_offline(struct dcd_disk *un, int bechatty) 1589 { 1590 if (bechatty) 1591 dcd_log(DCD_DEVINFO, dcd_label, CE_WARN, "offline\n"); 1592 1593 mutex_exit(DCD_MUTEX); 1594 cmlb_invalidate(un->un_dklbhandle, 0); 1595 mutex_enter(DCD_MUTEX); 1596 } 1597 1598 /* 1599 * Given the device number return the devinfo pointer 1600 * from the scsi_device structure. 1601 */ 1602 /*ARGSUSED*/ 1603 static int 1604 dcdinfo(dev_info_t *dip, ddi_info_cmd_t infocmd, void *arg, void **result) 1605 { 1606 dev_t dev; 1607 struct dcd_disk *un; 1608 int instance, error; 1609 1610 1611 switch (infocmd) { 1612 case DDI_INFO_DEVT2DEVINFO: 1613 dev = (dev_t)arg; 1614 instance = DCDUNIT(dev); 1615 if ((un = ddi_get_soft_state(dcd_state, instance)) == NULL) 1616 return (DDI_FAILURE); 1617 *result = (void *) DCD_DEVINFO; 1618 error = DDI_SUCCESS; 1619 break; 1620 case DDI_INFO_DEVT2INSTANCE: 1621 dev = (dev_t)arg; 1622 instance = DCDUNIT(dev); 1623 *result = (void *)(uintptr_t)instance; 1624 error = DDI_SUCCESS; 1625 break; 1626 default: 1627 error = DDI_FAILURE; 1628 } 1629 return (error); 1630 } 1631 1632 /* 1633 * property operation routine. return the number of blocks for the partition 1634 * in question or forward the request to the propery facilities. 1635 */ 1636 static int 1637 dcd_prop_op(dev_t dev, dev_info_t *dip, ddi_prop_op_t prop_op, int mod_flags, 1638 char *name, caddr_t valuep, int *lengthp) 1639 { 1640 struct dcd_disk *un; 1641 1642 if ((un = ddi_get_soft_state(dcd_state, ddi_get_instance(dip))) == NULL) 1643 return (ddi_prop_op(dev, dip, prop_op, mod_flags, 1644 name, valuep, lengthp)); 1645 1646 return (cmlb_prop_op(un->un_dklbhandle, 1647 dev, dip, prop_op, mod_flags, name, valuep, lengthp, 1648 DCDPART(dev), NULL)); 1649 } 1650 1651 /* 1652 * These routines perform raw i/o operations. 1653 */ 1654 /*ARGSUSED*/ 1655 void 1656 dcduscsimin(struct buf *bp) 1657 { 1658 1659 } 1660 1661 1662 static void 1663 dcdmin(struct buf *bp) 1664 { 1665 struct dcd_disk *un; 1666 int instance; 1667 minor_t minor = getminor(bp->b_edev); 1668 instance = minor >> DCDUNIT_SHIFT; 1669 un = ddi_get_soft_state(dcd_state, instance); 1670 1671 if (bp->b_bcount > un->un_max_xfer_size) 1672 bp->b_bcount = un->un_max_xfer_size; 1673 } 1674 1675 1676 /* ARGSUSED2 */ 1677 static int 1678 dcdread(dev_t dev, struct uio *uio, cred_t *cred_p) 1679 { 1680 int secmask; 1681 GET_SOFT_STATE(dev); 1682 #ifdef lint 1683 part = part; 1684 #endif /* lint */ 1685 secmask = un->un_secsize - 1; 1686 1687 if (uio->uio_loffset & ((offset_t)(secmask))) { 1688 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1689 "file offset not modulo %d\n", 1690 un->un_secsize); 1691 return (EINVAL); 1692 } else if (uio->uio_iov->iov_len & (secmask)) { 1693 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1694 "transfer length not modulo %d\n", un->un_secsize); 1695 return (EINVAL); 1696 } 1697 return (physio(dcdstrategy, (struct buf *)0, dev, B_READ, dcdmin, uio)); 1698 } 1699 1700 /* ARGSUSED2 */ 1701 static int 1702 dcdaread(dev_t dev, struct aio_req *aio, cred_t *cred_p) 1703 { 1704 int secmask; 1705 struct uio *uio = aio->aio_uio; 1706 GET_SOFT_STATE(dev); 1707 #ifdef lint 1708 part = part; 1709 #endif /* lint */ 1710 secmask = un->un_secsize - 1; 1711 1712 if (uio->uio_loffset & ((offset_t)(secmask))) { 1713 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1714 "file offset not modulo %d\n", 1715 un->un_secsize); 1716 return (EINVAL); 1717 } else if (uio->uio_iov->iov_len & (secmask)) { 1718 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1719 "transfer length not modulo %d\n", un->un_secsize); 1720 return (EINVAL); 1721 } 1722 return (aphysio(dcdstrategy, anocancel, dev, B_READ, dcdmin, aio)); 1723 } 1724 1725 /* ARGSUSED2 */ 1726 static int 1727 dcdwrite(dev_t dev, struct uio *uio, cred_t *cred_p) 1728 { 1729 int secmask; 1730 GET_SOFT_STATE(dev); 1731 #ifdef lint 1732 part = part; 1733 #endif /* lint */ 1734 secmask = un->un_secsize - 1; 1735 1736 if (uio->uio_loffset & ((offset_t)(secmask))) { 1737 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1738 "file offset not modulo %d\n", 1739 un->un_secsize); 1740 return (EINVAL); 1741 } else if (uio->uio_iov->iov_len & (secmask)) { 1742 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1743 "transfer length not modulo %d\n", un->un_secsize); 1744 return (EINVAL); 1745 } 1746 return (physio(dcdstrategy, (struct buf *)0, dev, B_WRITE, dcdmin, 1747 uio)); 1748 } 1749 1750 /* ARGSUSED2 */ 1751 static int 1752 dcdawrite(dev_t dev, struct aio_req *aio, cred_t *cred_p) 1753 { 1754 int secmask; 1755 struct uio *uio = aio->aio_uio; 1756 GET_SOFT_STATE(dev); 1757 #ifdef lint 1758 part = part; 1759 #endif /* lint */ 1760 secmask = un->un_secsize - 1; 1761 1762 if (uio->uio_loffset & ((offset_t)(secmask))) { 1763 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1764 "file offset not modulo %d\n", 1765 un->un_secsize); 1766 return (EINVAL); 1767 } else if (uio->uio_iov->iov_len & (secmask)) { 1768 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1769 "transfer length not modulo %d\n", un->un_secsize); 1770 return (EINVAL); 1771 } 1772 return (aphysio(dcdstrategy, anocancel, dev, B_WRITE, dcdmin, aio)); 1773 } 1774 1775 /* 1776 * strategy routine 1777 */ 1778 static int 1779 dcdstrategy(struct buf *bp) 1780 { 1781 struct dcd_disk *un; 1782 struct diskhd *dp; 1783 int i; 1784 minor_t minor = getminor(bp->b_edev); 1785 diskaddr_t p_lblksrt; 1786 diskaddr_t lblocks; 1787 diskaddr_t bn; 1788 1789 if ((un = ddi_get_soft_state(dcd_state, 1790 minor >> DCDUNIT_SHIFT)) == NULL || 1791 un->un_state == DCD_STATE_DUMPING || 1792 ((un->un_state & DCD_STATE_FATAL) == DCD_STATE_FATAL)) { 1793 SET_BP_ERROR(bp, ((un) ? ENXIO : EIO)); 1794 error: 1795 bp->b_resid = bp->b_bcount; 1796 biodone(bp); 1797 return (0); 1798 } 1799 1800 /* 1801 * If the request size (buf->b_bcount)is greater than the size 1802 * (un->un_max_xfer_size) supported by the target driver fail 1803 * the request with EINVAL error code. 1804 * 1805 * We are not supposed to receive requests exceeding 1806 * un->un_max_xfer_size size because the caller is expected to 1807 * check what is the maximum size that is supported by this 1808 * driver either through ioctl or dcdmin routine(which is private 1809 * to this driver). 1810 * But we have seen cases (like meta driver(md))where dcdstrategy 1811 * called with more than supported size and cause data corruption. 1812 */ 1813 1814 if (bp->b_bcount > un->un_max_xfer_size) { 1815 SET_BP_ERROR(bp, EINVAL); 1816 goto error; 1817 } 1818 1819 TRACE_2(TR_FAC_DADA, TR_DCDSTRATEGY_START, 1820 "dcdstrategy_start: bp 0x%p un 0x%p", bp, un); 1821 1822 /* 1823 * Commands may sneak in while we released the mutex in 1824 * DDI_SUSPEND, we should block new commands. 1825 */ 1826 mutex_enter(DCD_MUTEX); 1827 while (un->un_state == DCD_STATE_SUSPENDED) { 1828 cv_wait(&un->un_suspend_cv, DCD_MUTEX); 1829 } 1830 1831 if (un->un_state == DCD_STATE_PM_SUSPENDED) { 1832 mutex_exit(DCD_MUTEX); 1833 (void) pm_idle_component(DCD_DEVINFO, 0); 1834 if (pm_raise_power(DCD_DEVINFO, 0, 1835 DCD_DEVICE_ACTIVE) != DDI_SUCCESS) { 1836 SET_BP_ERROR(bp, EIO); 1837 goto error; 1838 } 1839 mutex_enter(DCD_MUTEX); 1840 } 1841 mutex_exit(DCD_MUTEX); 1842 1843 /* 1844 * Map-in the buffer in case starting address is not word aligned. 1845 */ 1846 1847 if (((uintptr_t)bp->b_un.b_addr) & 0x1) 1848 bp_mapin(bp); 1849 1850 bp->b_flags &= ~(B_DONE|B_ERROR); 1851 bp->b_resid = 0; 1852 bp->av_forw = 0; 1853 1854 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1855 "bp->b_bcount %lx\n", bp->b_bcount); 1856 1857 if (bp != un->un_sbufp) { 1858 validated: if (cmlb_partinfo(un->un_dklbhandle, 1859 minor & DCDPART_MASK, 1860 &lblocks, 1861 &p_lblksrt, 1862 NULL, 1863 NULL, 1864 0) == 0) { 1865 1866 bn = dkblock(bp); 1867 1868 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1869 "dkblock(bp) is %llu\n", bn); 1870 1871 i = 0; 1872 if (bn < 0) { 1873 i = -1; 1874 } else if (bn >= lblocks) { 1875 /* 1876 * For proper comparison, file system block 1877 * number has to be scaled to actual CD 1878 * transfer size. 1879 * Since all the CDROM operations 1880 * that have Sun Labels are in the correct 1881 * block size this will work for CD's. This 1882 * will have to change when we have different 1883 * sector sizes. 1884 * 1885 * if bn == lblocks, 1886 * Not an error, resid == count 1887 */ 1888 if (bn > lblocks) { 1889 i = -1; 1890 } else { 1891 i = 1; 1892 } 1893 } else if (bp->b_bcount & (un->un_secsize-1)) { 1894 /* 1895 * This should really be: 1896 * 1897 * ... if (bp->b_bcount & (un->un_lbasize-1)) 1898 * 1899 */ 1900 i = -1; 1901 } else { 1902 if (!bp->b_bcount) { 1903 printf("Waring : Zero read or Write\n"); 1904 goto error; 1905 } 1906 /* 1907 * sort by absolute block number. 1908 */ 1909 bp->b_resid = bn; 1910 bp->b_resid += p_lblksrt; 1911 /* 1912 * zero out av_back - this will be a signal 1913 * to dcdstart to go and fetch the resources 1914 */ 1915 bp->av_back = NO_PKT_ALLOCATED; 1916 } 1917 1918 /* 1919 * Check to see whether or not we are done 1920 * (with or without errors). 1921 */ 1922 1923 if (i != 0) { 1924 if (i < 0) { 1925 bp->b_flags |= B_ERROR; 1926 } 1927 goto error; 1928 } 1929 } else { 1930 /* 1931 * opened in NDELAY/NONBLOCK mode? 1932 * Check if disk is ready and has a valid geometry 1933 */ 1934 if (dcd_ready_and_valid(bp->b_edev, un) == 0) { 1935 goto validated; 1936 } else { 1937 dcd_log(DCD_DEVINFO, dcd_label, CE_WARN, 1938 "i/o to invalid geometry\n"); 1939 SET_BP_ERROR(bp, EIO); 1940 goto error; 1941 } 1942 } 1943 } else if (BP_HAS_NO_PKT(bp)) { 1944 struct udcd_cmd *tscmdp; 1945 struct dcd_cmd *tcmdp; 1946 /* 1947 * This indicates that it is a special buffer 1948 * This could be a udcd-cmd and hence call bp_mapin just 1949 * in case that it could be a PIO command issued. 1950 */ 1951 tscmdp = (struct udcd_cmd *)bp->b_forw; 1952 tcmdp = tscmdp->udcd_cmd; 1953 if ((tcmdp->cmd != ATA_READ_DMA) && (tcmdp->cmd != 0xc9) && 1954 (tcmdp->cmd != ATA_WRITE_DMA) && (tcmdp->cmd != 0xcb) && 1955 (tcmdp->cmd != IDENTIFY_DMA) && 1956 (tcmdp->cmd != ATA_FLUSH_CACHE)) { 1957 bp_mapin(bp); 1958 } 1959 } 1960 1961 /* 1962 * We are doing it a bit non-standard. That is, the 1963 * head of the b_actf chain is *not* the active command- 1964 * it is just the head of the wait queue. The reason 1965 * we do this is that the head of the b_actf chain is 1966 * guaranteed to not be moved by disksort(), so that 1967 * our restart command (pointed to by 1968 * b_forw) and the head of the wait queue (b_actf) can 1969 * have resources granted without it getting lost in 1970 * the queue at some later point (where we would have 1971 * to go and look for it). 1972 */ 1973 mutex_enter(DCD_MUTEX); 1974 1975 DCD_DO_KSTATS(un, kstat_waitq_enter, bp); 1976 1977 dp = &un->un_utab; 1978 1979 if (dp->b_actf == NULL) { 1980 dp->b_actf = bp; 1981 dp->b_actl = bp; 1982 } else if ((un->un_state == DCD_STATE_SUSPENDED) && 1983 bp == un->un_sbufp) { 1984 bp->b_actf = dp->b_actf; 1985 dp->b_actf = bp; 1986 } else { 1987 TRACE_3(TR_FAC_DADA, TR_DCDSTRATEGY_DISKSORT_START, 1988 "dcdstrategy_disksort_start: dp 0x%p bp 0x%p un 0x%p", 1989 dp, bp, un); 1990 disksort(dp, bp); 1991 TRACE_0(TR_FAC_DADA, TR_DCDSTRATEGY_DISKSORT_END, 1992 "dcdstrategy_disksort_end"); 1993 } 1994 1995 DAD_DEBUG(DCD_DEVINFO, dcd_label, DCD_DEBUG, 1996 "ncmd %x , throttle %x, forw 0x%p\n", 1997 un->un_ncmds, un->un_throttle, (void *)dp->b_forw); 1998 ASSERT(un->un_ncmds >= 0); 1999 ASSERT(un->un_throttle >= 0); 2000 if ((un->un_ncmds < un->un_throttle) && (dp->b_forw == NULL)) { 2001 dcdstart(un); 2002 } else if (BP_HAS_NO_PKT(dp->b_actf)) { 2003 struct buf *cmd_bp; 2004 2005 cmd_bp = dp->b_actf; 2006 cmd_bp->av_back = ALLOCATING_PKT; 2007 mutex_exit(DCD_MUTEX); 2008 /* 2009 * try and map this one 2010 */ 2011 TRACE_0(TR_FAC_DADA, TR_DCDSTRATEGY_SMALL_WINDOW_START, 2012 "dcdstrategy_small_window_call (begin)"); 2013 2014 make_dcd_cmd(un, cmd_bp, NULL_FUNC); 2015 2016 TRACE_0(TR_FAC_DADA, TR_DCDSTRATEGY_SMALL_WINDOW_END, 2017 "dcdstrategy_small_window_call (end)"); 2018 2019 /* 2020 * there is a small window where the active cmd 2021 * completes before make_dcd_cmd returns. 2022 * consequently, this cmd never gets started so 2023 * we start it from here 2024 */ 2025 mutex_enter(DCD_MUTEX); 2026 if ((un->un_ncmds < un->un_throttle) && 2027 (dp->b_forw == NULL)) { 2028 dcdstart(un); 2029 } 2030 } 2031 mutex_exit(DCD_MUTEX); 2032 2033 done: 2034 TRACE_0(TR_FAC_DADA, TR_DCDSTRATEGY_END, "dcdstrategy_end"); 2035 return (0); 2036 } 2037 2038 2039 /* 2040 * Unit start and Completion 2041 * NOTE: we assume that the caller has at least checked for: 2042 * (un->un_ncmds < un->un_throttle) 2043 * if not, there is no real harm done, dcd_transport() will 2044 * return BUSY 2045 */ 2046 static void 2047 dcdstart(struct dcd_disk *un) 2048 { 2049 int status, sort_key; 2050 struct buf *bp; 2051 struct diskhd *dp; 2052 uchar_t state = un->un_last_state; 2053 2054 TRACE_1(TR_FAC_DADA, TR_DCDSTART_START, "dcdstart_start: un 0x%p", un); 2055 2056 retry: 2057 ASSERT(mutex_owned(DCD_MUTEX)); 2058 2059 dp = &un->un_utab; 2060 if (((bp = dp->b_actf) == NULL) || (bp->av_back == ALLOCATING_PKT) || 2061 (dp->b_forw != NULL)) { 2062 TRACE_0(TR_FAC_DADA, TR_DCDSTART_NO_WORK_END, 2063 "dcdstart_end (no work)"); 2064 return; 2065 } 2066 2067 /* 2068 * remove from active queue 2069 */ 2070 dp->b_actf = bp->b_actf; 2071 bp->b_actf = 0; 2072 2073 /* 2074 * increment ncmds before calling dcd_transport because dcdintr 2075 * may be called before we return from dcd_transport! 2076 */ 2077 un->un_ncmds++; 2078 2079 /* 2080 * If measuring stats, mark exit from wait queue and 2081 * entrance into run 'queue' if and only if we are 2082 * going to actually start a command. 2083 * Normally the bp already has a packet at this point 2084 */ 2085 DCD_DO_KSTATS(un, kstat_waitq_to_runq, bp); 2086 2087 mutex_exit(DCD_MUTEX); 2088 2089 if (BP_HAS_NO_PKT(bp)) { 2090 make_dcd_cmd(un, bp, dcdrunout); 2091 if (BP_HAS_NO_PKT(bp) && !(bp->b_flags & B_ERROR)) { 2092 mutex_enter(DCD_MUTEX); 2093 DCD_DO_KSTATS(un, kstat_runq_back_to_waitq, bp); 2094 2095 bp->b_actf = dp->b_actf; 2096 dp->b_actf = bp; 2097 New_state(un, DCD_STATE_RWAIT); 2098 un->un_ncmds--; 2099 TRACE_0(TR_FAC_DADA, TR_DCDSTART_NO_RESOURCES_END, 2100 "dcdstart_end (No Resources)"); 2101 goto done; 2102 2103 } else if (bp->b_flags & B_ERROR) { 2104 mutex_enter(DCD_MUTEX); 2105 DCD_DO_KSTATS(un, kstat_runq_exit, bp); 2106 2107 un->un_ncmds--; 2108 bp->b_resid = bp->b_bcount; 2109 if (bp->b_error == 0) { 2110 SET_BP_ERROR(bp, EIO); 2111 } 2112 2113 /* 2114 * restore old state 2115 */ 2116 un->un_state = un->un_last_state; 2117 un->un_last_state = state; 2118 2119 mutex_exit(DCD_MUTEX); 2120 2121 biodone(bp); 2122 mutex_enter(DCD_MUTEX); 2123 if (un->un_state == DCD_STATE_SUSPENDED) { 2124 cv_broadcast(&un->un_disk_busy_cv); 2125 } 2126 2127 if ((un->un_ncmds < un->un_throttle) && 2128 (dp->b_forw == NULL)) { 2129 goto retry; 2130 } else { 2131 goto done; 2132 } 2133 } 2134 } 2135 2136 /* 2137 * Restore resid from the packet, b_resid had been the 2138 * disksort key. 2139 */ 2140 sort_key = bp->b_resid; 2141 bp->b_resid = BP_PKT(bp)->pkt_resid; 2142 BP_PKT(bp)->pkt_resid = 0; 2143 2144 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 2145 "bp->b_resid %lx, pkt_resid %lx\n", 2146 bp->b_resid, BP_PKT(bp)->pkt_resid); 2147 2148 /* 2149 * We used to check whether or not to try and link commands here. 2150 * Since we have found that there is no performance improvement 2151 * for linked commands, this has not made much sense. 2152 */ 2153 if ((status = dcd_transport((struct dcd_pkt *)BP_PKT(bp))) 2154 != TRAN_ACCEPT) { 2155 mutex_enter(DCD_MUTEX); 2156 un->un_ncmds--; 2157 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 2158 "transport returned %x\n", status); 2159 if (status == TRAN_BUSY) { 2160 DCD_DO_ERRSTATS(un, dcd_transerrs); 2161 DCD_DO_KSTATS(un, kstat_runq_back_to_waitq, bp); 2162 dcd_handle_tran_busy(bp, dp, un); 2163 if (un->un_ncmds > 0) { 2164 bp->b_resid = sort_key; 2165 } 2166 } else { 2167 DCD_DO_KSTATS(un, kstat_runq_exit, bp); 2168 mutex_exit(DCD_MUTEX); 2169 2170 dcd_log(DCD_DEVINFO, dcd_label, CE_WARN, 2171 "transport rejected (%d)\n", 2172 status); 2173 SET_BP_ERROR(bp, EIO); 2174 bp->b_resid = bp->b_bcount; 2175 if (bp != un->un_sbufp) { 2176 dcd_destroy_pkt(BP_PKT(bp)); 2177 } 2178 biodone(bp); 2179 2180 mutex_enter(DCD_MUTEX); 2181 if (un->un_state == DCD_STATE_SUSPENDED) { 2182 cv_broadcast(&un->un_disk_busy_cv); 2183 } 2184 if ((un->un_ncmds < un->un_throttle) && 2185 (dp->b_forw == NULL)) { 2186 goto retry; 2187 } 2188 } 2189 } else { 2190 mutex_enter(DCD_MUTEX); 2191 2192 if (dp->b_actf && BP_HAS_NO_PKT(dp->b_actf)) { 2193 struct buf *cmd_bp; 2194 2195 cmd_bp = dp->b_actf; 2196 cmd_bp->av_back = ALLOCATING_PKT; 2197 mutex_exit(DCD_MUTEX); 2198 /* 2199 * try and map this one 2200 */ 2201 TRACE_0(TR_FAC_DADA, TR_DCASTART_SMALL_WINDOW_START, 2202 "dcdstart_small_window_start"); 2203 2204 make_dcd_cmd(un, cmd_bp, NULL_FUNC); 2205 2206 TRACE_0(TR_FAC_DADA, TR_DCDSTART_SMALL_WINDOW_END, 2207 "dcdstart_small_window_end"); 2208 /* 2209 * there is a small window where the active cmd 2210 * completes before make_dcd_cmd returns. 2211 * consequently, this cmd never gets started so 2212 * we start it from here 2213 */ 2214 mutex_enter(DCD_MUTEX); 2215 if ((un->un_ncmds < un->un_throttle) && 2216 (dp->b_forw == NULL)) { 2217 goto retry; 2218 } 2219 } 2220 } 2221 2222 done: 2223 ASSERT(mutex_owned(DCD_MUTEX)); 2224 TRACE_0(TR_FAC_DADA, TR_DCDSTART_END, "dcdstart_end"); 2225 } 2226 2227 /* 2228 * make_dcd_cmd: create a pkt 2229 */ 2230 static void 2231 make_dcd_cmd(struct dcd_disk *un, struct buf *bp, int (*func)()) 2232 { 2233 auto int count, com, direction; 2234 struct dcd_pkt *pkt; 2235 int flags, tval; 2236 2237 _NOTE(DATA_READABLE_WITHOUT_LOCK(dcd_disk::un_dp)) 2238 TRACE_3(TR_FAC_DADA, TR_MAKE_DCD_CMD_START, 2239 "make_dcd_cmd_start: un 0x%p bp 0x%p un 0x%p", un, bp, un); 2240 2241 2242 flags = un->un_cmd_flags; 2243 2244 if (bp != un->un_sbufp) { 2245 int partition = DCDPART(bp->b_edev); 2246 diskaddr_t p_lblksrt; 2247 diskaddr_t lblocks; 2248 long secnt; 2249 uint32_t blkno; 2250 int dkl_nblk, delta; 2251 long resid; 2252 2253 if (cmlb_partinfo(un->un_dklbhandle, 2254 partition, 2255 &lblocks, 2256 &p_lblksrt, 2257 NULL, 2258 NULL, 2259 0) != NULL) { 2260 lblocks = 0; 2261 p_lblksrt = 0; 2262 } 2263 2264 dkl_nblk = (int)lblocks; 2265 2266 /* 2267 * Make sure we don't run off the end of a partition. 2268 * 2269 * Put this test here so that we can adjust b_count 2270 * to accurately reflect the actual amount we are 2271 * goint to transfer. 2272 */ 2273 2274 /* 2275 * First, compute partition-relative block number 2276 */ 2277 blkno = dkblock(bp); 2278 secnt = (bp->b_bcount + (un->un_secsize - 1)) >> un->un_secdiv; 2279 count = MIN(secnt, dkl_nblk - blkno); 2280 if (count != secnt) { 2281 /* 2282 * We have an overrun 2283 */ 2284 resid = (secnt - count) << un->un_secdiv; 2285 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 2286 "overrun by %ld sectors\n", 2287 secnt - count); 2288 bp->b_bcount -= resid; 2289 } else { 2290 resid = 0; 2291 } 2292 2293 /* 2294 * Adjust block number to absolute 2295 */ 2296 delta = (int)p_lblksrt; 2297 blkno += delta; 2298 2299 mutex_enter(DCD_MUTEX); 2300 /* 2301 * This is for devices having block size different from 2302 * from DEV_BSIZE (e.g. 2K CDROMs). 2303 */ 2304 if (un->un_lbasize != un->un_secsize) { 2305 blkno >>= un->un_blknoshift; 2306 count >>= un->un_blknoshift; 2307 } 2308 mutex_exit(DCD_MUTEX); 2309 2310 TRACE_0(TR_FAC_DADA, TR_MAKE_DCD_CMD_INIT_PKT_START, 2311 "make_dcd_cmd_init_pkt_call (begin)"); 2312 pkt = dcd_init_pkt(ROUTE, NULL, bp, 2313 (uint32_t)sizeof (struct dcd_cmd), 2314 un->un_cmd_stat_size, PP_LEN, PKT_CONSISTENT, 2315 func, (caddr_t)un); 2316 TRACE_1(TR_FAC_DADA, TR_MAKE_DCD_CMD_INIT_PKT_END, 2317 "make_dcd_cmd_init_pkt_call (end): pkt 0x%p", pkt); 2318 if (!pkt) { 2319 bp->b_bcount += resid; 2320 bp->av_back = NO_PKT_ALLOCATED; 2321 TRACE_0(TR_FAC_DADA, 2322 TR_MAKE_DCD_CMD_NO_PKT_ALLOCATED1_END, 2323 "make_dcd_cmd_end (NO_PKT_ALLOCATED1)"); 2324 return; 2325 } 2326 if (bp->b_flags & B_READ) { 2327 if ((un->un_dp->options & DMA_SUPPORTTED) == 2328 DMA_SUPPORTTED) { 2329 com = ATA_READ_DMA; 2330 } else { 2331 if (un->un_dp->options & BLOCK_MODE) 2332 com = ATA_READ_MULTIPLE; 2333 else 2334 com = ATA_READ; 2335 } 2336 direction = DATA_READ; 2337 } else { 2338 if ((un->un_dp->options & DMA_SUPPORTTED) == 2339 DMA_SUPPORTTED) { 2340 com = ATA_WRITE_DMA; 2341 } else { 2342 if (un->un_dp->options & BLOCK_MODE) 2343 com = ATA_WRITE_MULTIPLE; 2344 else 2345 com = ATA_WRITE; 2346 } 2347 direction = DATA_WRITE; 2348 } 2349 2350 /* 2351 * Save the resid in the packet, temporarily until 2352 * we transport the command. 2353 */ 2354 pkt->pkt_resid = resid; 2355 2356 makecommand(pkt, flags, com, blkno, ADD_LBA_MODE, 2357 bp->b_bcount, direction, 0); 2358 tval = dcd_io_time; 2359 } else { 2360 2361 struct udcd_cmd *scmd = (struct udcd_cmd *)bp->b_forw; 2362 2363 /* 2364 * set options 2365 */ 2366 if ((scmd->udcd_flags & UDCD_SILENT) && !(DEBUGGING)) { 2367 flags |= FLAG_SILENT; 2368 } 2369 if (scmd->udcd_flags & UDCD_DIAGNOSE) 2370 flags |= FLAG_DIAGNOSE; 2371 2372 if (scmd->udcd_flags & UDCD_NOINTR) 2373 flags |= FLAG_NOINTR; 2374 2375 pkt = dcd_init_pkt(ROUTE, (struct dcd_pkt *)NULL, 2376 (bp->b_bcount)? bp: NULL, 2377 (uint32_t)sizeof (struct dcd_cmd), 2378 2, PP_LEN, PKT_CONSISTENT, func, (caddr_t)un); 2379 2380 if (!pkt) { 2381 bp->av_back = NO_PKT_ALLOCATED; 2382 return; 2383 } 2384 2385 makecommand(pkt, 0, scmd->udcd_cmd->cmd, 2386 scmd->udcd_cmd->sector_num.lba_num, 2387 scmd->udcd_cmd->address_mode, 2388 scmd->udcd_cmd->size, 2389 scmd->udcd_cmd->direction, scmd->udcd_cmd->features); 2390 2391 pkt->pkt_flags = flags; 2392 if (scmd->udcd_timeout == 0) 2393 tval = dcd_io_time; 2394 else 2395 tval = scmd->udcd_timeout; 2396 /* UDAD interface should be decided. */ 2397 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 2398 "udcd interface\n"); 2399 } 2400 2401 pkt->pkt_comp = dcdintr; 2402 pkt->pkt_time = tval; 2403 PKT_SET_BP(pkt, bp); 2404 bp->av_back = (struct buf *)pkt; 2405 2406 TRACE_0(TR_FAC_DADA, TR_MAKE_DCD_CMD_END, "make_dcd_cmd_end"); 2407 } 2408 2409 /* 2410 * Command completion processing 2411 */ 2412 static void 2413 dcdintr(struct dcd_pkt *pkt) 2414 { 2415 struct dcd_disk *un; 2416 struct buf *bp; 2417 int action; 2418 int status; 2419 2420 bp = PKT_GET_BP(pkt); 2421 un = ddi_get_soft_state(dcd_state, DCDUNIT(bp->b_edev)); 2422 2423 TRACE_1(TR_FAC_DADA, TR_DCDINTR_START, "dcdintr_start: un 0x%p", un); 2424 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, "dcdintr\n"); 2425 2426 mutex_enter(DCD_MUTEX); 2427 un->un_ncmds--; 2428 DCD_DO_KSTATS(un, kstat_runq_exit, bp); 2429 ASSERT(un->un_ncmds >= 0); 2430 2431 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 2432 "reason %x and Status %x\n", pkt->pkt_reason, SCBP_C(pkt)); 2433 2434 /* 2435 * do most common case first 2436 */ 2437 if ((pkt->pkt_reason == CMD_CMPLT) && (SCBP_C(pkt) == 0)) { 2438 int com = GETATACMD((struct dcd_cmd *)pkt->pkt_cdbp); 2439 2440 if (un->un_state == DCD_STATE_OFFLINE) { 2441 un->un_state = un->un_last_state; 2442 dcd_log(DCD_DEVINFO, dcd_label, CE_NOTE, 2443 (const char *) diskokay); 2444 } 2445 /* 2446 * If the command is a read or a write, and we have 2447 * a non-zero pkt_resid, that is an error. We should 2448 * attempt to retry the operation if possible. 2449 */ 2450 action = COMMAND_DONE; 2451 if (pkt->pkt_resid && (com == ATA_READ || com == ATA_WRITE)) { 2452 DCD_DO_ERRSTATS(un, dcd_harderrs); 2453 if ((int)PKT_GET_RETRY_CNT(pkt) < dcd_retry_count) { 2454 PKT_INCR_RETRY_CNT(pkt, 1); 2455 action = QUE_COMMAND; 2456 } else { 2457 /* 2458 * if we have exhausted retries 2459 * a command with a residual is in error in 2460 * this case. 2461 */ 2462 action = COMMAND_DONE_ERROR; 2463 } 2464 dcd_log(DCD_DEVINFO, dcd_label, 2465 CE_WARN, "incomplete %s- %s\n", 2466 (bp->b_flags & B_READ)? "read" : "write", 2467 (action == QUE_COMMAND)? "retrying" : 2468 "giving up"); 2469 } 2470 2471 /* 2472 * pkt_resid will reflect, at this point, a residual 2473 * of how many bytes left to be transferred there were 2474 * from the actual scsi command. Add this to b_resid i.e 2475 * the amount this driver could not see to transfer, 2476 * to get the total number of bytes not transfered. 2477 */ 2478 if (action != QUE_COMMAND) { 2479 bp->b_resid += pkt->pkt_resid; 2480 } 2481 2482 } else if (pkt->pkt_reason != CMD_CMPLT) { 2483 action = dcd_handle_incomplete(un, bp); 2484 } 2485 2486 /* 2487 * If we are in the middle of syncing or dumping, we have got 2488 * here because dcd_transport has called us explictly after 2489 * completing the command in a polled mode. We don't want to 2490 * have a recursive call into dcd_transport again. 2491 */ 2492 if (ddi_in_panic() && (action == QUE_COMMAND)) { 2493 action = COMMAND_DONE_ERROR; 2494 } 2495 2496 /* 2497 * save pkt reason; consecutive failures are not reported unless 2498 * fatal 2499 * do not reset last_pkt_reason when the cmd was retried and 2500 * succeeded because 2501 * there maybe more commands comming back with last_pkt_reason 2502 */ 2503 if ((un->un_last_pkt_reason != pkt->pkt_reason) && 2504 ((pkt->pkt_reason != CMD_CMPLT) || 2505 (PKT_GET_RETRY_CNT(pkt) == 0))) { 2506 un->un_last_pkt_reason = pkt->pkt_reason; 2507 } 2508 2509 switch (action) { 2510 case COMMAND_DONE_ERROR: 2511 error: 2512 if (bp->b_resid == 0) { 2513 bp->b_resid = bp->b_bcount; 2514 } 2515 if (bp->b_error == 0) { 2516 struct dcd_cmd *cdbp = (struct dcd_cmd *)pkt->pkt_cdbp; 2517 if (cdbp->cmd == ATA_FLUSH_CACHE && 2518 (pkt->pkt_scbp[0] & STATUS_ATA_ERR) && 2519 (pkt->pkt_scbp[1] & ERR_ABORT)) { 2520 SET_BP_ERROR(bp, ENOTSUP); 2521 un->un_flush_not_supported = 1; 2522 } else { 2523 SET_BP_ERROR(bp, EIO); 2524 } 2525 } 2526 bp->b_flags |= B_ERROR; 2527 /*FALLTHROUGH*/ 2528 case COMMAND_DONE: 2529 dcddone_and_mutex_exit(un, bp); 2530 2531 TRACE_0(TR_FAC_DADA, TR_DCDINTR_COMMAND_DONE_END, 2532 "dcdintr_end (COMMAND_DONE)"); 2533 return; 2534 2535 case QUE_COMMAND: 2536 if (un->un_ncmds >= un->un_throttle) { 2537 struct diskhd *dp = &un->un_utab; 2538 2539 bp->b_actf = dp->b_actf; 2540 dp->b_actf = bp; 2541 2542 DCD_DO_KSTATS(un, kstat_waitq_enter, bp); 2543 2544 mutex_exit(DCD_MUTEX); 2545 goto exit; 2546 } 2547 2548 un->un_ncmds++; 2549 /* reset the pkt reason again */ 2550 pkt->pkt_reason = 0; 2551 DCD_DO_KSTATS(un, kstat_runq_enter, bp); 2552 mutex_exit(DCD_MUTEX); 2553 if ((status = dcd_transport(BP_PKT(bp))) != TRAN_ACCEPT) { 2554 struct diskhd *dp = &un->un_utab; 2555 2556 mutex_enter(DCD_MUTEX); 2557 un->un_ncmds--; 2558 if (status == TRAN_BUSY) { 2559 DCD_DO_KSTATS(un, kstat_runq_back_to_waitq, bp); 2560 dcd_handle_tran_busy(bp, dp, un); 2561 mutex_exit(DCD_MUTEX); 2562 goto exit; 2563 } 2564 DCD_DO_ERRSTATS(un, dcd_transerrs); 2565 DCD_DO_KSTATS(un, kstat_runq_exit, bp); 2566 2567 dcd_log(DCD_DEVINFO, dcd_label, CE_WARN, 2568 "requeue of command fails (%x)\n", status); 2569 SET_BP_ERROR(bp, EIO); 2570 bp->b_resid = bp->b_bcount; 2571 2572 dcddone_and_mutex_exit(un, bp); 2573 goto exit; 2574 } 2575 break; 2576 2577 case JUST_RETURN: 2578 default: 2579 DCD_DO_KSTATS(un, kstat_waitq_enter, bp); 2580 mutex_exit(DCD_MUTEX); 2581 break; 2582 } 2583 2584 exit: 2585 TRACE_0(TR_FAC_DADA, TR_DCDINTR_END, "dcdintr_end"); 2586 } 2587 2588 2589 /* 2590 * Done with a command. 2591 */ 2592 static void 2593 dcddone_and_mutex_exit(struct dcd_disk *un, register struct buf *bp) 2594 { 2595 struct diskhd *dp; 2596 2597 TRACE_1(TR_FAC_DADA, TR_DCDONE_START, "dcddone_start: un 0x%p", un); 2598 2599 _NOTE(LOCK_RELEASED_AS_SIDE_EFFECT(&un->un_dcd->dcd_mutex)); 2600 2601 dp = &un->un_utab; 2602 if (bp == dp->b_forw) { 2603 dp->b_forw = NULL; 2604 } 2605 2606 if (un->un_stats) { 2607 ulong_t n_done = bp->b_bcount - bp->b_resid; 2608 if (bp->b_flags & B_READ) { 2609 IOSP->reads++; 2610 IOSP->nread += n_done; 2611 } else { 2612 IOSP->writes++; 2613 IOSP->nwritten += n_done; 2614 } 2615 } 2616 if (IO_PARTITION_STATS) { 2617 ulong_t n_done = bp->b_bcount - bp->b_resid; 2618 if (bp->b_flags & B_READ) { 2619 IOSP_PARTITION->reads++; 2620 IOSP_PARTITION->nread += n_done; 2621 } else { 2622 IOSP_PARTITION->writes++; 2623 IOSP_PARTITION->nwritten += n_done; 2624 } 2625 } 2626 2627 /* 2628 * Start the next one before releasing resources on this one 2629 */ 2630 if (un->un_state == DCD_STATE_SUSPENDED) { 2631 cv_broadcast(&un->un_disk_busy_cv); 2632 } else if (dp->b_actf && (un->un_ncmds < un->un_throttle) && 2633 (dp->b_forw == NULL && un->un_state != DCD_STATE_SUSPENDED)) { 2634 dcdstart(un); 2635 } 2636 2637 mutex_exit(DCD_MUTEX); 2638 2639 if (bp != un->un_sbufp) { 2640 dcd_destroy_pkt(BP_PKT(bp)); 2641 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 2642 "regular done: resid %ld\n", bp->b_resid); 2643 } else { 2644 ASSERT(un->un_sbuf_busy); 2645 } 2646 TRACE_0(TR_FAC_DADA, TR_DCDDONE_BIODONE_CALL, "dcddone_biodone_call"); 2647 2648 biodone(bp); 2649 2650 (void) pm_idle_component(DCD_DEVINFO, 0); 2651 2652 TRACE_0(TR_FAC_DADA, TR_DCDDONE_END, "dcddone end"); 2653 } 2654 2655 2656 /* 2657 * reset the disk unless the transport layer has already 2658 * cleared the problem 2659 */ 2660 #define C1 (STAT_ATA_BUS_RESET|STAT_ATA_DEV_RESET|STAT_ATA_ABORTED) 2661 static void 2662 dcd_reset_disk(struct dcd_disk *un, struct dcd_pkt *pkt) 2663 { 2664 2665 if ((pkt->pkt_statistics & C1) == 0) { 2666 mutex_exit(DCD_MUTEX); 2667 if (!dcd_reset(ROUTE, RESET_ALL)) { 2668 DAD_DEBUG(DCD_DEVINFO, dcd_label, DCD_DEBUG, 2669 "Reset failed"); 2670 } 2671 mutex_enter(DCD_MUTEX); 2672 } 2673 } 2674 2675 static int 2676 dcd_handle_incomplete(struct dcd_disk *un, struct buf *bp) 2677 { 2678 static char *fail = "ATA transport failed: reason '%s': %s\n"; 2679 static char *notresp = "disk not responding to selection\n"; 2680 int rval = COMMAND_DONE_ERROR; 2681 int action = COMMAND_SOFT_ERROR; 2682 struct dcd_pkt *pkt = BP_PKT(bp); 2683 int be_chatty = (un->un_state != DCD_STATE_SUSPENDED) && 2684 (bp != un->un_sbufp || !(pkt->pkt_flags & FLAG_SILENT)); 2685 2686 ASSERT(mutex_owned(DCD_MUTEX)); 2687 2688 switch (pkt->pkt_reason) { 2689 2690 case CMD_TIMEOUT: 2691 /* 2692 * This Indicates the already the HBA would have reset 2693 * so Just indicate to retry the command 2694 */ 2695 break; 2696 2697 case CMD_INCOMPLETE: 2698 action = dcd_check_error(un, bp); 2699 DCD_DO_ERRSTATS(un, dcd_transerrs); 2700 if (action == COMMAND_HARD_ERROR) { 2701 (void) dcd_reset_disk(un, pkt); 2702 } 2703 break; 2704 2705 case CMD_FATAL: 2706 /* 2707 * Something drastic has gone wrong 2708 */ 2709 break; 2710 case CMD_DMA_DERR: 2711 case CMD_DATA_OVR: 2712 /* FALLTHROUGH */ 2713 2714 default: 2715 /* 2716 * the target may still be running the command, 2717 * so we should try and reset that target. 2718 */ 2719 DCD_DO_ERRSTATS(un, dcd_transerrs); 2720 if ((pkt->pkt_reason != CMD_RESET) && 2721 (pkt->pkt_reason != CMD_ABORTED)) { 2722 (void) dcd_reset_disk(un, pkt); 2723 } 2724 break; 2725 } 2726 2727 /* 2728 * If pkt_reason is CMD_RESET/ABORTED, chances are that this pkt got 2729 * reset/aborted because another disk on this bus caused it. 2730 * The disk that caused it, should get CMD_TIMEOUT with pkt_statistics 2731 * of STAT_TIMEOUT/STAT_DEV_RESET 2732 */ 2733 if ((pkt->pkt_reason == CMD_RESET) ||(pkt->pkt_reason == CMD_ABORTED)) { 2734 /* To be written : XXX */ 2735 DAD_DEBUG(DCD_DEVINFO, dcd_label, DCD_DEBUG, 2736 "Command aborted\n"); 2737 } 2738 2739 if (bp == un->un_sbufp && (pkt->pkt_flags & FLAG_DIAGNOSE)) { 2740 rval = COMMAND_DONE_ERROR; 2741 } else { 2742 if ((rval == COMMAND_DONE_ERROR) && 2743 (action == COMMAND_SOFT_ERROR) && 2744 ((int)PKT_GET_RETRY_CNT(pkt) < dcd_retry_count)) { 2745 PKT_INCR_RETRY_CNT(pkt, 1); 2746 rval = QUE_COMMAND; 2747 } 2748 } 2749 2750 if (pkt->pkt_reason == CMD_INCOMPLETE && rval == COMMAND_DONE_ERROR) { 2751 /* 2752 * Looks like someone turned off this shoebox. 2753 */ 2754 if (un->un_state != DCD_STATE_OFFLINE) { 2755 dcd_log(DCD_DEVINFO, dcd_label, CE_WARN, 2756 (const char *) notresp); 2757 New_state(un, DCD_STATE_OFFLINE); 2758 } 2759 } else if (pkt->pkt_reason == CMD_FATAL) { 2760 /* 2761 * Suppressing the following message for the time being 2762 * dcd_log(DCD_DEVINFO, dcd_label, CE_WARN, 2763 * (const char *) notresp); 2764 */ 2765 PKT_INCR_RETRY_CNT(pkt, 6); 2766 rval = COMMAND_DONE_ERROR; 2767 New_state(un, DCD_STATE_FATAL); 2768 } else if (be_chatty) { 2769 int in_panic = ddi_in_panic(); 2770 if (!in_panic || (rval == COMMAND_DONE_ERROR)) { 2771 if (((pkt->pkt_reason != un->un_last_pkt_reason) && 2772 (pkt->pkt_reason != CMD_RESET)) || 2773 (rval == COMMAND_DONE_ERROR) || 2774 (dcd_error_level == DCD_ERR_ALL)) { 2775 dcd_log(DCD_DEVINFO, dcd_label, CE_WARN, 2776 fail, dcd_rname(pkt->pkt_reason), 2777 (rval == COMMAND_DONE_ERROR) ? 2778 "giving up": "retrying command"); 2779 DAD_DEBUG(DCD_DEVINFO, dcd_label, DCD_DEBUG, 2780 "retrycount=%x\n", 2781 PKT_GET_RETRY_CNT(pkt)); 2782 } 2783 } 2784 } 2785 error: 2786 return (rval); 2787 } 2788 2789 static int 2790 dcd_check_error(struct dcd_disk *un, struct buf *bp) 2791 { 2792 struct diskhd *dp = &un->un_utab; 2793 struct dcd_pkt *pkt = BP_PKT(bp); 2794 int rval = 0; 2795 unsigned char status; 2796 unsigned char error; 2797 2798 TRACE_0(TR_FAC_DADA, TR_DCD_CHECK_ERROR_START, "dcd_check_error_start"); 2799 ASSERT(mutex_owned(DCD_MUTEX)); 2800 2801 DAD_DEBUG(DCD_DEVINFO, dcd_label, DCD_DEBUG, 2802 "Pkt: 0x%p dp: 0x%p\n", (void *)pkt, (void *)dp); 2803 2804 /* 2805 * Here we need to check status first and then if error is indicated 2806 * Then the error register. 2807 */ 2808 2809 status = (pkt->pkt_scbp)[0]; 2810 if ((status & STATUS_ATA_DWF) == STATUS_ATA_DWF) { 2811 /* 2812 * There has been a Device Fault - reason for such error 2813 * is vendor specific 2814 * Action to be taken is - Indicate error and reset device. 2815 */ 2816 2817 dcd_log(DCD_DEVINFO, dcd_label, CE_WARN, "Device Fault\n"); 2818 rval = COMMAND_HARD_ERROR; 2819 } else if ((status & STATUS_ATA_CORR) == STATUS_ATA_CORR) { 2820 2821 /* 2822 * The sector read or written is marginal and hence ECC 2823 * Correction has been applied. Indicate to repair 2824 * Here we need to probably re-assign based on the badblock 2825 * mapping. 2826 */ 2827 2828 dcd_log(DCD_DEVINFO, dcd_label, CE_WARN, 2829 "Soft Error on block %x\n", 2830 ((struct dcd_cmd *)pkt->pkt_cdbp)->sector_num.lba_num); 2831 rval = COMMAND_SOFT_ERROR; 2832 } else if ((status & STATUS_ATA_ERR) == STATUS_ATA_ERR) { 2833 error = pkt->pkt_scbp[1]; 2834 2835 dcd_log(DCD_DEVINFO, dcd_label, CE_WARN, 2836 "Command:0x%x,Error:0x%x,Status:0x%x\n", 2837 GETATACMD((struct dcd_cmd *)pkt->pkt_cdbp), 2838 error, status); 2839 if ((error & ERR_AMNF) == ERR_AMNF) { 2840 /* Address make not found */ 2841 dcd_log(DCD_DEVINFO, dcd_label, CE_WARN, 2842 "Address Mark Not Found"); 2843 } else if ((error & ERR_TKONF) == ERR_TKONF) { 2844 /* Track 0 Not found */ 2845 dcd_log(DCD_DEVINFO, dcd_label, CE_WARN, 2846 "Track 0 Not found \n"); 2847 } else if ((error & ERR_IDNF) == ERR_IDNF) { 2848 dcd_log(DCD_DEVINFO, dcd_label, CE_WARN, 2849 " ID not found \n"); 2850 } else if ((error & ERR_UNC) == ERR_UNC) { 2851 dcd_log(DCD_DEVINFO, dcd_label, CE_WARN, 2852 "Uncorrectable data Error: Block %x\n", 2853 ((struct dcd_cmd *)pkt->pkt_cdbp)-> 2854 sector_num.lba_num); 2855 } else if ((error & ERR_BBK) == ERR_BBK) { 2856 dcd_log(DCD_DEVINFO, dcd_label, CE_WARN, 2857 "Bad block detected: Block %x\n", 2858 ((struct dcd_cmd *)pkt->pkt_cdbp)-> 2859 sector_num.lba_num); 2860 } else if ((error & ERR_ABORT) == ERR_ABORT) { 2861 /* Aborted Command */ 2862 dcd_log(DCD_DEVINFO, dcd_label, CE_WARN, 2863 " Aborted Command \n"); 2864 } 2865 /* 2866 * Return the soft error so that the command 2867 * will be retried. 2868 */ 2869 rval = COMMAND_SOFT_ERROR; 2870 } 2871 2872 TRACE_0(TR_FAC_DADA, TR_DCD_CHECK_ERROR_END, "dcd_check_error_end"); 2873 return (rval); 2874 } 2875 2876 2877 /* 2878 * System Crash Dump routine 2879 */ 2880 2881 #define NDUMP_RETRIES 5 2882 2883 static int 2884 dcddump(dev_t dev, caddr_t addr, daddr_t blkno, int nblk) 2885 { 2886 struct dcd_pkt *pkt; 2887 int i; 2888 struct buf local, *bp; 2889 int err; 2890 unsigned char com; 2891 diskaddr_t p_lblksrt; 2892 diskaddr_t lblocks; 2893 2894 GET_SOFT_STATE(dev); 2895 #ifdef lint 2896 part = part; 2897 #endif /* lint */ 2898 2899 _NOTE(NOW_INVISIBLE_TO_OTHER_THREADS(*un)) 2900 2901 if ((un->un_state & DCD_STATE_FATAL) == DCD_STATE_FATAL) 2902 return (ENXIO); 2903 2904 if (cmlb_partinfo(un->un_dklbhandle, DCDPART(dev), 2905 &lblocks, &p_lblksrt, NULL, NULL, 0)) 2906 return (ENXIO); 2907 2908 if (blkno+nblk > lblocks) { 2909 return (EINVAL); 2910 } 2911 2912 2913 if ((un->un_state == DCD_STATE_SUSPENDED) || 2914 (un->un_state == DCD_STATE_PM_SUSPENDED)) { 2915 if (pm_raise_power(DCD_DEVINFO, 0, 2916 DCD_DEVICE_ACTIVE) != DDI_SUCCESS) { 2917 return (EIO); 2918 } 2919 } 2920 2921 /* 2922 * When cpr calls dcddump, we know that dad is in a 2923 * a good state, so no bus reset is required 2924 */ 2925 un->un_throttle = 0; 2926 2927 if ((un->un_state != DCD_STATE_SUSPENDED) && 2928 (un->un_state != DCD_STATE_DUMPING)) { 2929 2930 New_state(un, DCD_STATE_DUMPING); 2931 2932 /* 2933 * Reset the bus. I'd like to not have to do this, 2934 * but this is the safest thing to do... 2935 */ 2936 2937 if (dcd_reset(ROUTE, RESET_ALL) == 0) { 2938 return (EIO); 2939 } 2940 2941 } 2942 2943 blkno += p_lblksrt; 2944 2945 /* 2946 * It should be safe to call the allocator here without 2947 * worrying about being locked for DVMA mapping because 2948 * the address we're passed is already a DVMA mapping 2949 * 2950 * We are also not going to worry about semaphore ownership 2951 * in the dump buffer. Dumping is single threaded at present. 2952 */ 2953 2954 bp = &local; 2955 bzero((caddr_t)bp, sizeof (*bp)); 2956 bp->b_flags = B_BUSY; 2957 bp->b_un.b_addr = addr; 2958 bp->b_bcount = nblk << DEV_BSHIFT; 2959 bp->b_resid = 0; 2960 2961 for (i = 0; i < NDUMP_RETRIES; i++) { 2962 bp->b_flags &= ~B_ERROR; 2963 if ((pkt = dcd_init_pkt(ROUTE, NULL, bp, 2964 (uint32_t)sizeof (struct dcd_cmd), 2, PP_LEN, 2965 PKT_CONSISTENT, NULL_FUNC, NULL)) != NULL) { 2966 break; 2967 } 2968 if (i == 0) { 2969 if (bp->b_flags & B_ERROR) { 2970 dcd_log(DCD_DEVINFO, dcd_label, CE_WARN, 2971 "no resources for dumping; " 2972 "error code: 0x%x, retrying", 2973 geterror(bp)); 2974 } else { 2975 dcd_log(DCD_DEVINFO, dcd_label, CE_WARN, 2976 "no resources for dumping; retrying"); 2977 } 2978 } else if (i != (NDUMP_RETRIES - 1)) { 2979 if (bp->b_flags & B_ERROR) { 2980 dcd_log(DCD_DEVINFO, dcd_label, CE_CONT, "no " 2981 "resources for dumping; error code: 0x%x, " 2982 "retrying\n", geterror(bp)); 2983 } 2984 } else { 2985 if (bp->b_flags & B_ERROR) { 2986 dcd_log(DCD_DEVINFO, dcd_label, CE_CONT, 2987 "no resources for dumping; " 2988 "error code: 0x%x, retries failed, " 2989 "giving up.\n", geterror(bp)); 2990 } else { 2991 dcd_log(DCD_DEVINFO, dcd_label, CE_CONT, 2992 "no resources for dumping; " 2993 "retries failed, giving up.\n"); 2994 } 2995 return (EIO); 2996 } 2997 delay(10); 2998 } 2999 if ((un->un_dp->options & DMA_SUPPORTTED) == DMA_SUPPORTTED) { 3000 com = ATA_WRITE_DMA; 3001 } else { 3002 if (un->un_dp->options & BLOCK_MODE) 3003 com = ATA_WRITE_MULTIPLE; 3004 else 3005 com = ATA_WRITE; 3006 } 3007 3008 makecommand(pkt, 0, com, blkno, ADD_LBA_MODE, 3009 (int)nblk*un->un_secsize, DATA_WRITE, 0); 3010 3011 for (err = EIO, i = 0; i < NDUMP_RETRIES && err == EIO; i++) { 3012 3013 if (dcd_poll(pkt) == 0) { 3014 switch (SCBP_C(pkt)) { 3015 case STATUS_GOOD: 3016 if (pkt->pkt_resid == 0) { 3017 err = 0; 3018 } 3019 break; 3020 case STATUS_ATA_BUSY: 3021 (void) dcd_reset(ROUTE, RESET_TARGET); 3022 break; 3023 default: 3024 mutex_enter(DCD_MUTEX); 3025 (void) dcd_reset_disk(un, pkt); 3026 mutex_exit(DCD_MUTEX); 3027 break; 3028 } 3029 } else if (i > NDUMP_RETRIES/2) { 3030 (void) dcd_reset(ROUTE, RESET_ALL); 3031 } 3032 3033 } 3034 dcd_destroy_pkt(pkt); 3035 return (err); 3036 } 3037 3038 /* 3039 * This routine implements the ioctl calls. It is called 3040 * from the device switch at normal priority. 3041 */ 3042 /* ARGSUSED3 */ 3043 static int 3044 dcdioctl(dev_t dev, int cmd, intptr_t arg, int flag, 3045 cred_t *cred_p, int *rval_p) 3046 { 3047 auto int32_t data[512 / (sizeof (int32_t))]; 3048 struct dk_cinfo *info; 3049 struct dk_minfo media_info; 3050 struct udcd_cmd *scmd; 3051 int i, err; 3052 enum uio_seg uioseg = 0; 3053 enum dkio_state state = 0; 3054 #ifdef _MULTI_DATAMODEL 3055 struct dadkio_rwcmd rwcmd; 3056 #endif 3057 struct dadkio_rwcmd32 rwcmd32; 3058 struct dcd_cmd dcdcmd; 3059 3060 GET_SOFT_STATE(dev); 3061 #ifdef lint 3062 part = part; 3063 state = state; 3064 uioseg = uioseg; 3065 #endif /* lint */ 3066 3067 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 3068 "dcd_ioctl : cmd %x, arg %lx\n", cmd, arg); 3069 3070 bzero((caddr_t)data, sizeof (data)); 3071 3072 switch (cmd) { 3073 3074 #ifdef DCDDEBUG 3075 /* 3076 * Following ioctl are for testing RESET/ABORTS 3077 */ 3078 #define DKIOCRESET (DKIOC|14) 3079 #define DKIOCABORT (DKIOC|15) 3080 3081 case DKIOCRESET: 3082 if (ddi_copyin((caddr_t)arg, (caddr_t)data, 4, flag)) 3083 return (EFAULT); 3084 DAD_DEBUG(DCD_DEVINFO, dcd_label, DCD_DEBUG, 3085 "DKIOCRESET: data = 0x%x\n", data[0]); 3086 if (dcd_reset(ROUTE, data[0])) { 3087 return (0); 3088 } else { 3089 return (EIO); 3090 } 3091 case DKIOCABORT: 3092 DAD_DEBUG(DCD_DEVINFO, dcd_label, DCD_DEBUG, 3093 "DKIOCABORT:\n"); 3094 if (dcd_abort(ROUTE, (struct dcd_pkt *)0)) { 3095 return (0); 3096 } else { 3097 return (EIO); 3098 } 3099 #endif 3100 3101 case DKIOCINFO: 3102 /* 3103 * Controller Information 3104 */ 3105 info = (struct dk_cinfo *)data; 3106 3107 mutex_enter(DCD_MUTEX); 3108 switch (un->un_dp->ctype) { 3109 default: 3110 info->dki_ctype = DKC_DIRECT; 3111 break; 3112 } 3113 mutex_exit(DCD_MUTEX); 3114 info->dki_cnum = ddi_get_instance(ddi_get_parent(DCD_DEVINFO)); 3115 (void) strcpy(info->dki_cname, 3116 ddi_get_name(ddi_get_parent(DCD_DEVINFO))); 3117 /* 3118 * Unit Information 3119 */ 3120 info->dki_unit = ddi_get_instance(DCD_DEVINFO); 3121 info->dki_slave = (Tgt(DCD_DCD_DEVP)<<3); 3122 (void) strcpy(info->dki_dname, ddi_driver_name(DCD_DEVINFO)); 3123 info->dki_flags = DKI_FMTVOL; 3124 info->dki_partition = DCDPART(dev); 3125 3126 /* 3127 * Max Transfer size of this device in blocks 3128 */ 3129 info->dki_maxtransfer = un->un_max_xfer_size / DEV_BSIZE; 3130 3131 /* 3132 * We can't get from here to there yet 3133 */ 3134 info->dki_addr = 0; 3135 info->dki_space = 0; 3136 info->dki_prio = 0; 3137 info->dki_vec = 0; 3138 3139 i = sizeof (struct dk_cinfo); 3140 if (ddi_copyout((caddr_t)data, (caddr_t)arg, i, flag)) 3141 return (EFAULT); 3142 else 3143 return (0); 3144 3145 case DKIOCGMEDIAINFO: 3146 /* 3147 * As dad target driver is used for IDE disks only 3148 * Can keep the return value hardcoded to FIXED_DISK 3149 */ 3150 media_info.dki_media_type = DK_FIXED_DISK; 3151 3152 mutex_enter(DCD_MUTEX); 3153 media_info.dki_lbsize = un->un_lbasize; 3154 media_info.dki_capacity = un->un_diskcapacity; 3155 mutex_exit(DCD_MUTEX); 3156 3157 if (ddi_copyout(&media_info, (caddr_t)arg, 3158 sizeof (struct dk_minfo), flag)) 3159 return (EFAULT); 3160 else 3161 return (0); 3162 3163 case DKIOCGGEOM: 3164 case DKIOCGVTOC: 3165 case DKIOCGETEFI: 3166 3167 mutex_enter(DCD_MUTEX); 3168 if (un->un_ncmds == 0) { 3169 if ((err = dcd_unit_ready(dev)) != 0) { 3170 mutex_exit(DCD_MUTEX); 3171 return (err); 3172 } 3173 } 3174 3175 mutex_exit(DCD_MUTEX); 3176 err = cmlb_ioctl(un->un_dklbhandle, dev, cmd, 3177 arg, flag, cred_p, rval_p, 0); 3178 return (err); 3179 3180 case DKIOCGAPART: 3181 case DKIOCSAPART: 3182 case DKIOCSGEOM: 3183 case DKIOCSVTOC: 3184 case DKIOCSETEFI: 3185 case DKIOCPARTITION: 3186 case DKIOCPARTINFO: 3187 case DKIOCGMBOOT: 3188 case DKIOCSMBOOT: 3189 3190 err = cmlb_ioctl(un->un_dklbhandle, dev, cmd, 3191 arg, flag, cred_p, rval_p, 0); 3192 return (err); 3193 3194 case DIOCTL_RWCMD: 3195 if (drv_priv(cred_p) != 0) { 3196 return (EPERM); 3197 } 3198 3199 #ifdef _MULTI_DATAMODEL 3200 switch (ddi_model_convert_from(flag & FMODELS)) { 3201 case DDI_MODEL_NONE: 3202 if (ddi_copyin((caddr_t)arg, (caddr_t)&rwcmd, 3203 sizeof (struct dadkio_rwcmd), flag)) { 3204 return (EFAULT); 3205 } 3206 rwcmd32.cmd = rwcmd.cmd; 3207 rwcmd32.flags = rwcmd.flags; 3208 rwcmd32.blkaddr = rwcmd.blkaddr; 3209 rwcmd32.buflen = rwcmd.buflen; 3210 rwcmd32.bufaddr = (caddr32_t)(uintptr_t)rwcmd.bufaddr; 3211 break; 3212 case DDI_MODEL_ILP32: 3213 if (ddi_copyin((caddr_t)arg, (caddr_t)&rwcmd32, 3214 sizeof (struct dadkio_rwcmd32), flag)) { 3215 return (EFAULT); 3216 } 3217 break; 3218 } 3219 #else 3220 if (ddi_copyin((caddr_t)arg, (caddr_t)&rwcmd32, 3221 sizeof (struct dadkio_rwcmd32), flag)) { 3222 return (EFAULT); 3223 } 3224 #endif 3225 mutex_enter(DCD_MUTEX); 3226 3227 uioseg = UIO_SYSSPACE; 3228 scmd = (struct udcd_cmd *)data; 3229 scmd->udcd_cmd = &dcdcmd; 3230 /* 3231 * Convert the dadkio_rwcmd structure to udcd_cmd so that 3232 * it can take the normal path to get the io done 3233 */ 3234 if (rwcmd32.cmd == DADKIO_RWCMD_READ) { 3235 if ((un->un_dp->options & DMA_SUPPORTTED) == 3236 DMA_SUPPORTTED) 3237 scmd->udcd_cmd->cmd = ATA_READ_DMA; 3238 else 3239 scmd->udcd_cmd->cmd = ATA_READ; 3240 scmd->udcd_cmd->address_mode = ADD_LBA_MODE; 3241 scmd->udcd_cmd->direction = DATA_READ; 3242 scmd->udcd_flags |= UDCD_READ|UDCD_SILENT; 3243 } else if (rwcmd32.cmd == DADKIO_RWCMD_WRITE) { 3244 if ((un->un_dp->options & DMA_SUPPORTTED) == 3245 DMA_SUPPORTTED) 3246 scmd->udcd_cmd->cmd = ATA_WRITE_DMA; 3247 else 3248 scmd->udcd_cmd->cmd = ATA_WRITE; 3249 scmd->udcd_cmd->direction = DATA_WRITE; 3250 scmd->udcd_flags |= UDCD_WRITE|UDCD_SILENT; 3251 } else { 3252 mutex_exit(DCD_MUTEX); 3253 return (EINVAL); 3254 } 3255 3256 scmd->udcd_cmd->address_mode = ADD_LBA_MODE; 3257 scmd->udcd_cmd->features = 0; 3258 scmd->udcd_cmd->size = rwcmd32.buflen; 3259 scmd->udcd_cmd->sector_num.lba_num = rwcmd32.blkaddr; 3260 scmd->udcd_bufaddr = (caddr_t)(uintptr_t)rwcmd32.bufaddr; 3261 scmd->udcd_buflen = rwcmd32.buflen; 3262 scmd->udcd_timeout = (ushort_t)dcd_io_time; 3263 scmd->udcd_resid = 0ULL; 3264 scmd->udcd_status = 0; 3265 scmd->udcd_error_reg = 0; 3266 scmd->udcd_status_reg = 0; 3267 3268 mutex_exit(DCD_MUTEX); 3269 3270 i = dcdioctl_cmd(dev, scmd, UIO_SYSSPACE, UIO_USERSPACE); 3271 mutex_enter(DCD_MUTEX); 3272 /* 3273 * After return convert the status from scmd to 3274 * dadkio_status 3275 */ 3276 (void) dcd_translate(&(rwcmd32.status), scmd); 3277 rwcmd32.status.resid = scmd->udcd_resid; 3278 mutex_exit(DCD_MUTEX); 3279 3280 #ifdef _MULTI_DATAMODEL 3281 switch (ddi_model_convert_from(flag & FMODELS)) { 3282 case DDI_MODEL_NONE: { 3283 int counter; 3284 rwcmd.status.status = rwcmd32.status.status; 3285 rwcmd.status.resid = rwcmd32.status.resid; 3286 rwcmd.status.failed_blk_is_valid = 3287 rwcmd32.status.failed_blk_is_valid; 3288 rwcmd.status.failed_blk = rwcmd32.status.failed_blk; 3289 rwcmd.status.fru_code_is_valid = 3290 rwcmd32.status.fru_code_is_valid; 3291 rwcmd.status.fru_code = rwcmd32.status.fru_code; 3292 for (counter = 0; 3293 counter < DADKIO_ERROR_INFO_LEN; counter++) 3294 rwcmd.status.add_error_info[counter] = 3295 rwcmd32.status.add_error_info[counter]; 3296 } 3297 /* Copy out the result back to the user program */ 3298 if (ddi_copyout((caddr_t)&rwcmd, (caddr_t)arg, 3299 sizeof (struct dadkio_rwcmd), flag)) { 3300 if (i != 0) { 3301 i = EFAULT; 3302 } 3303 } 3304 break; 3305 case DDI_MODEL_ILP32: 3306 /* Copy out the result back to the user program */ 3307 if (ddi_copyout((caddr_t)&rwcmd32, (caddr_t)arg, 3308 sizeof (struct dadkio_rwcmd32), flag)) { 3309 if (i != 0) { 3310 i = EFAULT; 3311 } 3312 } 3313 break; 3314 } 3315 #else 3316 /* Copy out the result back to the user program */ 3317 if (ddi_copyout((caddr_t)&rwcmd32, (caddr_t)arg, 3318 sizeof (struct dadkio_rwcmd32), flag)) { 3319 if (i != 0) 3320 i = EFAULT; 3321 } 3322 #endif 3323 return (i); 3324 3325 case UDCDCMD: { 3326 #ifdef _MULTI_DATAMODEL 3327 /* 3328 * For use when a 32 bit app makes a call into a 3329 * 64 bit ioctl 3330 */ 3331 struct udcd_cmd32 udcd_cmd_32_for_64; 3332 struct udcd_cmd32 *ucmd32 = &udcd_cmd_32_for_64; 3333 model_t model; 3334 #endif /* _MULTI_DATAMODEL */ 3335 3336 if (drv_priv(cred_p) != 0) { 3337 return (EPERM); 3338 } 3339 3340 scmd = (struct udcd_cmd *)data; 3341 3342 #ifdef _MULTI_DATAMODEL 3343 switch (model = ddi_model_convert_from(flag & FMODELS)) { 3344 case DDI_MODEL_ILP32: 3345 if (ddi_copyin((caddr_t)arg, ucmd32, 3346 sizeof (struct udcd_cmd32), flag)) { 3347 return (EFAULT); 3348 } 3349 /* 3350 * Convert the ILP32 uscsi data from the 3351 * application to LP64 for internal use. 3352 */ 3353 udcd_cmd32toudcd_cmd(ucmd32, scmd); 3354 break; 3355 case DDI_MODEL_NONE: 3356 if (ddi_copyin((caddr_t)arg, scmd, sizeof (*scmd), 3357 flag)) { 3358 return (EFAULT); 3359 } 3360 break; 3361 } 3362 #else /* ! _MULTI_DATAMODEL */ 3363 if (ddi_copyin((caddr_t)arg, (caddr_t)scmd, 3364 sizeof (*scmd), flag)) { 3365 return (EFAULT); 3366 } 3367 #endif /* ! _MULTI_DATAMODEL */ 3368 3369 scmd->udcd_flags &= ~UDCD_NOINTR; 3370 uioseg = (flag & FKIOCTL)? UIO_SYSSPACE: UIO_USERSPACE; 3371 3372 i = dcdioctl_cmd(dev, scmd, uioseg, uioseg); 3373 #ifdef _MULTI_DATAMODEL 3374 switch (model) { 3375 case DDI_MODEL_ILP32: 3376 /* 3377 * Convert back to ILP32 before copyout to the 3378 * application 3379 */ 3380 udcd_cmdtoudcd_cmd32(scmd, ucmd32); 3381 if (ddi_copyout(ucmd32, (caddr_t)arg, 3382 sizeof (*ucmd32), flag)) { 3383 if (i != 0) 3384 i = EFAULT; 3385 } 3386 break; 3387 case DDI_MODEL_NONE: 3388 if (ddi_copyout(scmd, (caddr_t)arg, sizeof (*scmd), 3389 flag)) { 3390 if (i != 0) 3391 i = EFAULT; 3392 } 3393 break; 3394 } 3395 #else /* ! _MULTI_DATAMODE */ 3396 if (ddi_copyout((caddr_t)scmd, (caddr_t)arg, 3397 sizeof (*scmd), flag)) { 3398 if (i != 0) 3399 i = EFAULT; 3400 } 3401 #endif 3402 return (i); 3403 } 3404 case DKIOCFLUSHWRITECACHE: { 3405 struct dk_callback *dkc = (struct dk_callback *)arg; 3406 struct dcd_pkt *pkt; 3407 struct buf *bp; 3408 int is_sync = 1; 3409 3410 mutex_enter(DCD_MUTEX); 3411 if (un->un_flush_not_supported || 3412 ! un->un_write_cache_enabled) { 3413 i = un->un_flush_not_supported ? ENOTSUP : 0; 3414 mutex_exit(DCD_MUTEX); 3415 /* 3416 * If a callback was requested: a callback will 3417 * always be done if the caller saw the 3418 * DKIOCFLUSHWRITECACHE ioctl return 0, and 3419 * never done if the caller saw the ioctl return 3420 * an error. 3421 */ 3422 if ((flag & FKIOCTL) && dkc != NULL && 3423 dkc->dkc_callback != NULL) { 3424 (*dkc->dkc_callback)(dkc->dkc_cookie, i); 3425 /* 3426 * Did callback and reported error. 3427 * Since we did a callback, ioctl 3428 * should return 0. 3429 */ 3430 i = 0; 3431 } 3432 return (i); 3433 } 3434 3435 /* 3436 * Get the special buffer 3437 */ 3438 while (un->un_sbuf_busy) { 3439 cv_wait(&un->un_sbuf_cv, DCD_MUTEX); 3440 } 3441 un->un_sbuf_busy = 1; 3442 bp = un->un_sbufp; 3443 mutex_exit(DCD_MUTEX); 3444 3445 pkt = dcd_init_pkt(ROUTE, (struct dcd_pkt *)NULL, 3446 NULL, (uint32_t)sizeof (struct dcd_cmd), 3447 2, PP_LEN, PKT_CONSISTENT, SLEEP_FUNC, (caddr_t)un); 3448 ASSERT(pkt != NULL); 3449 3450 makecommand(pkt, un->un_cmd_flags | FLAG_SILENT, 3451 ATA_FLUSH_CACHE, 0, ADD_LBA_MODE, 0, NO_DATA_XFER, 0); 3452 3453 pkt->pkt_comp = dcdintr; 3454 pkt->pkt_time = DCD_FLUSH_TIME; 3455 PKT_SET_BP(pkt, bp); 3456 3457 bp->av_back = (struct buf *)pkt; 3458 bp->b_forw = NULL; 3459 bp->b_flags = B_BUSY; 3460 bp->b_error = 0; 3461 bp->b_edev = dev; 3462 bp->b_dev = cmpdev(dev); 3463 bp->b_bcount = 0; 3464 bp->b_blkno = 0; 3465 bp->b_un.b_addr = 0; 3466 bp->b_iodone = NULL; 3467 bp->b_list = NULL; 3468 3469 if ((flag & FKIOCTL) && dkc != NULL && 3470 dkc->dkc_callback != NULL) { 3471 struct dk_callback *dkc2 = (struct dk_callback *) 3472 kmem_zalloc(sizeof (*dkc2), KM_SLEEP); 3473 bcopy(dkc, dkc2, sizeof (*dkc2)); 3474 3475 bp->b_list = (struct buf *)dkc2; 3476 bp->b_iodone = dcdflushdone; 3477 is_sync = 0; 3478 } 3479 3480 (void) dcdstrategy(bp); 3481 3482 i = 0; 3483 if (is_sync) { 3484 i = biowait(bp); 3485 (void) dcdflushdone(bp); 3486 } 3487 3488 return (i); 3489 } 3490 default: 3491 break; 3492 } 3493 return (ENOTTY); 3494 } 3495 3496 3497 static int 3498 dcdflushdone(struct buf *bp) 3499 { 3500 struct dcd_disk *un = ddi_get_soft_state(dcd_state, 3501 DCDUNIT(bp->b_edev)); 3502 struct dcd_pkt *pkt = BP_PKT(bp); 3503 struct dk_callback *dkc = (struct dk_callback *)bp->b_list; 3504 3505 ASSERT(un != NULL); 3506 ASSERT(bp == un->un_sbufp); 3507 ASSERT(pkt != NULL); 3508 3509 dcd_destroy_pkt(pkt); 3510 bp->av_back = NO_PKT_ALLOCATED; 3511 3512 if (dkc != NULL) { 3513 ASSERT(bp->b_iodone != NULL); 3514 (*dkc->dkc_callback)(dkc->dkc_cookie, geterror(bp)); 3515 kmem_free(dkc, sizeof (*dkc)); 3516 bp->b_iodone = NULL; 3517 bp->b_list = NULL; 3518 } 3519 3520 /* 3521 * Tell anybody who cares that the buffer is now free 3522 */ 3523 mutex_enter(DCD_MUTEX); 3524 un->un_sbuf_busy = 0; 3525 cv_signal(&un->un_sbuf_cv); 3526 mutex_exit(DCD_MUTEX); 3527 return (0); 3528 } 3529 3530 /* 3531 * dcdrunout: 3532 * the callback function for resource allocation 3533 * 3534 * XXX it would be preferable that dcdrunout() scans the whole 3535 * list for possible candidates for dcdstart(); this avoids 3536 * that a bp at the head of the list whose request cannot be 3537 * satisfied is retried again and again 3538 */ 3539 /*ARGSUSED*/ 3540 static int 3541 dcdrunout(caddr_t arg) 3542 { 3543 int serviced; 3544 struct dcd_disk *un; 3545 struct diskhd *dp; 3546 3547 TRACE_1(TR_FAC_DADA, TR_DCDRUNOUT_START, "dcdrunout_start: arg 0x%p", 3548 arg); 3549 serviced = 1; 3550 3551 un = (struct dcd_disk *)arg; 3552 dp = &un->un_utab; 3553 3554 /* 3555 * We now support passing a structure to the callback 3556 * routine. 3557 */ 3558 ASSERT(un != NULL); 3559 mutex_enter(DCD_MUTEX); 3560 if ((un->un_ncmds < un->un_throttle) && (dp->b_forw == NULL)) { 3561 dcdstart(un); 3562 } 3563 if (un->un_state == DCD_STATE_RWAIT) { 3564 serviced = 0; 3565 } 3566 mutex_exit(DCD_MUTEX); 3567 TRACE_1(TR_FAC_DADA, TR_DCDRUNOUT_END, 3568 "dcdrunout_end: serviced %d", serviced); 3569 return (serviced); 3570 } 3571 3572 3573 /* 3574 * This routine called to see whether unit is (still) there. Must not 3575 * be called when un->un_sbufp is in use, and must not be called with 3576 * an unattached disk. Soft state of disk is restored to what it was 3577 * upon entry- up to caller to set the correct state. 3578 * 3579 * We enter with the disk mutex held. 3580 */ 3581 3582 /* ARGSUSED0 */ 3583 static int 3584 dcd_unit_ready(dev_t dev) 3585 { 3586 #ifndef lint 3587 auto struct udcd_cmd dcmd, *com = &dcmd; 3588 auto struct dcd_cmd cmdblk; 3589 #endif 3590 int error; 3591 #ifndef lint 3592 GET_SOFT_STATE(dev); 3593 #endif 3594 3595 /* 3596 * Now that we protect the special buffer with 3597 * a mutex, we could probably do a mutex_tryenter 3598 * on it here and return failure if it were held... 3599 */ 3600 3601 error = 0; 3602 return (error); 3603 } 3604 3605 /* ARGSUSED0 */ 3606 int 3607 dcdioctl_cmd(dev_t devp, struct udcd_cmd *in, enum uio_seg cdbspace, 3608 enum uio_seg dataspace) 3609 { 3610 3611 struct buf *bp; 3612 struct udcd_cmd *scmd; 3613 struct dcd_pkt *pkt; 3614 int err, rw; 3615 caddr_t cdb; 3616 int flags = 0; 3617 3618 GET_SOFT_STATE(devp); 3619 3620 #ifdef lint 3621 part = part; 3622 #endif 3623 3624 /* 3625 * Is this a request to reset the bus? 3626 * if so, we need to do reseting. 3627 */ 3628 3629 if (in->udcd_flags & UDCD_RESET) { 3630 int flag = RESET_TARGET; 3631 err = dcd_reset(ROUTE, flag) ? 0: EIO; 3632 return (err); 3633 } 3634 3635 scmd = in; 3636 3637 3638 /* Do some sanity checks */ 3639 if (scmd->udcd_buflen <= 0) { 3640 if (scmd->udcd_flags & (UDCD_READ | UDCD_WRITE)) { 3641 return (EINVAL); 3642 } else { 3643 scmd->udcd_buflen = 0; 3644 } 3645 } 3646 3647 /* Make a copy of the dcd_cmd passed */ 3648 cdb = kmem_zalloc(sizeof (struct dcd_cmd), KM_SLEEP); 3649 if (cdbspace == UIO_SYSSPACE) { 3650 flags |= FKIOCTL; 3651 } 3652 3653 if (ddi_copyin((void *)scmd->udcd_cmd, cdb, sizeof (struct dcd_cmd), 3654 flags)) { 3655 kmem_free(cdb, sizeof (struct dcd_cmd)); 3656 return (EFAULT); 3657 } 3658 scmd = (struct udcd_cmd *)kmem_alloc(sizeof (*scmd), KM_SLEEP); 3659 bcopy((caddr_t)in, (caddr_t)scmd, sizeof (*scmd)); 3660 scmd->udcd_cmd = (struct dcd_cmd *)cdb; 3661 rw = (scmd->udcd_flags & UDCD_READ) ? B_READ: B_WRITE; 3662 3663 3664 /* 3665 * Get the special buffer 3666 */ 3667 3668 mutex_enter(DCD_MUTEX); 3669 while (un->un_sbuf_busy) { 3670 if (cv_wait_sig(&un->un_sbuf_cv, DCD_MUTEX) == 0) { 3671 kmem_free(scmd->udcd_cmd, sizeof (struct dcd_cmd)); 3672 kmem_free((caddr_t)scmd, sizeof (*scmd)); 3673 mutex_exit(DCD_MUTEX); 3674 return (EINTR); 3675 } 3676 } 3677 3678 un->un_sbuf_busy = 1; 3679 bp = un->un_sbufp; 3680 mutex_exit(DCD_MUTEX); 3681 3682 3683 /* 3684 * If we are going to do actual I/O, let physio do all the 3685 * things 3686 */ 3687 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 3688 "dcdioctl_cmd : buflen %x\n", scmd->udcd_buflen); 3689 3690 if (scmd->udcd_buflen) { 3691 auto struct iovec aiov; 3692 auto struct uio auio; 3693 struct uio *uio = &auio; 3694 3695 bzero((caddr_t)&auio, sizeof (struct uio)); 3696 bzero((caddr_t)&aiov, sizeof (struct iovec)); 3697 3698 aiov.iov_base = scmd->udcd_bufaddr; 3699 aiov.iov_len = scmd->udcd_buflen; 3700 3701 uio->uio_iov = &aiov; 3702 uio->uio_iovcnt = 1; 3703 uio->uio_resid = scmd->udcd_buflen; 3704 uio->uio_segflg = dataspace; 3705 3706 /* 3707 * Let physio do the rest... 3708 */ 3709 bp->av_back = NO_PKT_ALLOCATED; 3710 bp->b_forw = (struct buf *)scmd; 3711 err = physio(dcdstrategy, bp, devp, rw, dcdudcdmin, uio); 3712 } else { 3713 /* 3714 * We have to mimic what physio would do here. 3715 */ 3716 bp->av_back = NO_PKT_ALLOCATED; 3717 bp->b_forw = (struct buf *)scmd; 3718 bp->b_flags = B_BUSY | rw; 3719 bp->b_edev = devp; 3720 bp->b_dev = cmpdev(devp); 3721 bp->b_bcount = bp->b_blkno = 0; 3722 (void) dcdstrategy(bp); 3723 err = biowait(bp); 3724 } 3725 3726 done: 3727 if ((pkt = BP_PKT(bp)) != NULL) { 3728 bp->av_back = NO_PKT_ALLOCATED; 3729 /* we need to update the completion status of udcd command */ 3730 in->udcd_resid = bp->b_resid; 3731 in->udcd_status_reg = SCBP_C(pkt); 3732 /* XXX: we need to give error_reg also */ 3733 dcd_destroy_pkt(pkt); 3734 } 3735 /* 3736 * Tell anybody who cares that the buffer is now free 3737 */ 3738 mutex_enter(DCD_MUTEX); 3739 un->un_sbuf_busy = 0; 3740 cv_signal(&un->un_sbuf_cv); 3741 mutex_exit(DCD_MUTEX); 3742 3743 kmem_free(scmd->udcd_cmd, sizeof (struct dcd_cmd)); 3744 kmem_free((caddr_t)scmd, sizeof (*scmd)); 3745 return (err); 3746 } 3747 3748 static void 3749 dcdudcdmin(struct buf *bp) 3750 { 3751 3752 #ifdef lint 3753 bp = bp; 3754 #endif 3755 3756 } 3757 3758 /* 3759 * restart a cmd from timeout() context 3760 * 3761 * the cmd is expected to be in un_utab.b_forw. If this pointer is non-zero 3762 * a restart timeout request has been issued and no new timeouts should 3763 * be requested. b_forw is reset when the cmd eventually completes in 3764 * dcddone_and_mutex_exit() 3765 */ 3766 void 3767 dcdrestart(void *arg) 3768 { 3769 struct dcd_disk *un = (struct dcd_disk *)arg; 3770 struct buf *bp; 3771 int status; 3772 3773 DAD_DEBUG(DCD_DEVINFO, dcd_label, DCD_DEBUG, "dcdrestart\n"); 3774 3775 mutex_enter(DCD_MUTEX); 3776 bp = un->un_utab.b_forw; 3777 if (bp) { 3778 un->un_ncmds++; 3779 DCD_DO_KSTATS(un, kstat_waitq_to_runq, bp); 3780 } 3781 3782 3783 if (bp) { 3784 struct dcd_pkt *pkt = BP_PKT(bp); 3785 3786 mutex_exit(DCD_MUTEX); 3787 3788 pkt->pkt_flags = 0; 3789 3790 if ((status = dcd_transport(pkt)) != TRAN_ACCEPT) { 3791 mutex_enter(DCD_MUTEX); 3792 DCD_DO_KSTATS(un, kstat_runq_back_to_waitq, bp); 3793 un->un_ncmds--; 3794 if (status == TRAN_BUSY) { 3795 /* XXX : To be checked */ 3796 /* 3797 * if (un->un_throttle > 1) { 3798 * ASSERT(un->un_ncmds >= 0); 3799 * un->un_throttle = un->un_ncmds; 3800 * } 3801 */ 3802 un->un_reissued_timeid = 3803 timeout(dcdrestart, (caddr_t)un, 3804 DCD_BSY_TIMEOUT/500); 3805 mutex_exit(DCD_MUTEX); 3806 return; 3807 } 3808 DCD_DO_ERRSTATS(un, dcd_transerrs); 3809 dcd_log(DCD_DEVINFO, dcd_label, CE_WARN, 3810 "dcdrestart transport failed (%x)\n", status); 3811 bp->b_resid = bp->b_bcount; 3812 SET_BP_ERROR(bp, EIO); 3813 3814 DCD_DO_KSTATS(un, kstat_waitq_exit, bp); 3815 un->un_reissued_timeid = 0L; 3816 dcddone_and_mutex_exit(un, bp); 3817 return; 3818 } 3819 mutex_enter(DCD_MUTEX); 3820 } 3821 un->un_reissued_timeid = 0L; 3822 mutex_exit(DCD_MUTEX); 3823 DAD_DEBUG(DCD_DEVINFO, dcd_label, DCD_DEBUG, "dcdrestart done\n"); 3824 } 3825 3826 /* 3827 * This routine gets called to reset the throttle to its saved 3828 * value wheneven we lower the throttle. 3829 */ 3830 void 3831 dcd_reset_throttle(caddr_t arg) 3832 { 3833 struct dcd_disk *un = (struct dcd_disk *)arg; 3834 struct diskhd *dp; 3835 3836 mutex_enter(DCD_MUTEX); 3837 dp = &un->un_utab; 3838 3839 /* 3840 * start any commands that didn't start while throttling. 3841 */ 3842 if (dp->b_actf && (un->un_ncmds < un->un_throttle) && 3843 (dp->b_forw == NULL)) { 3844 dcdstart(un); 3845 } 3846 mutex_exit(DCD_MUTEX); 3847 } 3848 3849 3850 /* 3851 * This routine handles the case when a TRAN_BUSY is 3852 * returned by HBA. 3853 * 3854 * If there are some commands already in the transport, the 3855 * bp can be put back on queue and it will 3856 * be retried when the queue is emptied after command 3857 * completes. But if there is no command in the tranport 3858 * and it still return busy, we have to retry the command 3859 * after some time like 10ms. 3860 */ 3861 /* ARGSUSED0 */ 3862 static void 3863 dcd_handle_tran_busy(struct buf *bp, struct diskhd *dp, struct dcd_disk *un) 3864 { 3865 ASSERT(mutex_owned(DCD_MUTEX)); 3866 3867 3868 if (dp->b_forw == NULL || dp->b_forw == bp) { 3869 dp->b_forw = bp; 3870 } else if (dp->b_forw != bp) { 3871 bp->b_actf = dp->b_actf; 3872 dp->b_actf = bp; 3873 3874 } 3875 if (!un->un_reissued_timeid) { 3876 un->un_reissued_timeid = 3877 timeout(dcdrestart, (caddr_t)un, DCD_BSY_TIMEOUT/500); 3878 } 3879 } 3880 3881 static int 3882 dcd_write_deviceid(struct dcd_disk *un) 3883 { 3884 3885 int status; 3886 diskaddr_t blk; 3887 struct udcd_cmd ucmd; 3888 struct dcd_cmd cdb; 3889 struct dk_devid *dkdevid; 3890 uint_t *ip, chksum; 3891 int i; 3892 dev_t dev; 3893 3894 mutex_exit(DCD_MUTEX); 3895 if (cmlb_get_devid_block(un->un_dklbhandle, &blk, 0)) { 3896 mutex_enter(DCD_MUTEX); 3897 return (EINVAL); 3898 } 3899 mutex_enter(DCD_MUTEX); 3900 3901 /* Allocate the buffer */ 3902 dkdevid = kmem_zalloc(un->un_secsize, KM_SLEEP); 3903 3904 /* Fill in the revision */ 3905 dkdevid->dkd_rev_hi = DK_DEVID_REV_MSB; 3906 dkdevid->dkd_rev_lo = DK_DEVID_REV_LSB; 3907 3908 /* Copy in the device id */ 3909 bcopy(un->un_devid, &dkdevid->dkd_devid, 3910 ddi_devid_sizeof(un->un_devid)); 3911 3912 /* Calculate the chksum */ 3913 chksum = 0; 3914 ip = (uint_t *)dkdevid; 3915 for (i = 0; i < ((un->un_secsize - sizeof (int))/sizeof (int)); i++) 3916 chksum ^= ip[i]; 3917 3918 /* Fill in the checksum */ 3919 DKD_FORMCHKSUM(chksum, dkdevid); 3920 3921 (void) bzero((caddr_t)&ucmd, sizeof (ucmd)); 3922 (void) bzero((caddr_t)&cdb, sizeof (struct dcd_cmd)); 3923 3924 if ((un->un_dp->options & DMA_SUPPORTTED) == DMA_SUPPORTTED) { 3925 cdb.cmd = ATA_WRITE_DMA; 3926 } else { 3927 if (un->un_dp->options & BLOCK_MODE) 3928 cdb.cmd = ATA_WRITE_MULTIPLE; 3929 else 3930 cdb.cmd = ATA_WRITE; 3931 } 3932 cdb.size = un->un_secsize; 3933 cdb.sector_num.lba_num = blk; 3934 cdb.address_mode = ADD_LBA_MODE; 3935 cdb.direction = DATA_WRITE; 3936 3937 ucmd.udcd_flags = UDCD_WRITE; 3938 ucmd.udcd_cmd = &cdb; 3939 ucmd.udcd_bufaddr = (caddr_t)dkdevid; 3940 ucmd.udcd_buflen = un->un_secsize; 3941 ucmd.udcd_flags |= UDCD_SILENT; 3942 dev = makedevice(ddi_driver_major(DCD_DEVINFO), 3943 ddi_get_instance(DCD_DEVINFO) << DCDUNIT_SHIFT); 3944 mutex_exit(DCD_MUTEX); 3945 status = dcdioctl_cmd(dev, &ucmd, UIO_SYSSPACE, UIO_SYSSPACE); 3946 mutex_enter(DCD_MUTEX); 3947 3948 kmem_free(dkdevid, un->un_secsize); 3949 return (status); 3950 } 3951 3952 static int 3953 dcd_read_deviceid(struct dcd_disk *un) 3954 { 3955 int status; 3956 diskaddr_t blk; 3957 struct udcd_cmd ucmd; 3958 struct dcd_cmd cdb; 3959 struct dk_devid *dkdevid; 3960 uint_t *ip; 3961 int chksum; 3962 int i, sz; 3963 dev_t dev; 3964 3965 mutex_exit(DCD_MUTEX); 3966 if (cmlb_get_devid_block(un->un_dklbhandle, &blk, 0)) { 3967 mutex_enter(DCD_MUTEX); 3968 return (EINVAL); 3969 } 3970 mutex_enter(DCD_MUTEX); 3971 3972 dkdevid = kmem_alloc(un->un_secsize, KM_SLEEP); 3973 3974 (void) bzero((caddr_t)&ucmd, sizeof (ucmd)); 3975 (void) bzero((caddr_t)&cdb, sizeof (cdb)); 3976 3977 if ((un->un_dp->options & DMA_SUPPORTTED) == DMA_SUPPORTTED) { 3978 cdb.cmd = ATA_READ_DMA; 3979 } else { 3980 if (un->un_dp->options & BLOCK_MODE) 3981 cdb.cmd = ATA_READ_MULTIPLE; 3982 else 3983 cdb.cmd = ATA_READ; 3984 } 3985 cdb.size = un->un_secsize; 3986 cdb.sector_num.lba_num = blk; 3987 cdb.address_mode = ADD_LBA_MODE; 3988 cdb.direction = DATA_READ; 3989 3990 ucmd.udcd_flags = UDCD_READ; 3991 ucmd.udcd_cmd = &cdb; 3992 ucmd.udcd_bufaddr = (caddr_t)dkdevid; 3993 ucmd.udcd_buflen = un->un_secsize; 3994 ucmd.udcd_flags |= UDCD_SILENT; 3995 dev = makedevice(ddi_driver_major(DCD_DEVINFO), 3996 ddi_get_instance(DCD_DEVINFO) << DCDUNIT_SHIFT); 3997 mutex_exit(DCD_MUTEX); 3998 status = dcdioctl_cmd(dev, &ucmd, UIO_SYSSPACE, UIO_SYSSPACE); 3999 mutex_enter(DCD_MUTEX); 4000 4001 if (status != 0) { 4002 kmem_free((caddr_t)dkdevid, un->un_secsize); 4003 return (status); 4004 } 4005 4006 /* Validate the revision */ 4007 4008 if ((dkdevid->dkd_rev_hi != DK_DEVID_REV_MSB) || 4009 (dkdevid->dkd_rev_lo != DK_DEVID_REV_LSB)) { 4010 kmem_free((caddr_t)dkdevid, un->un_secsize); 4011 return (EINVAL); 4012 } 4013 4014 /* Calculate the checksum */ 4015 chksum = 0; 4016 ip = (uint_t *)dkdevid; 4017 for (i = 0; i < ((un->un_secsize - sizeof (int))/sizeof (int)); i++) 4018 chksum ^= ip[i]; 4019 4020 /* Compare the checksums */ 4021 4022 if (DKD_GETCHKSUM(dkdevid) != chksum) { 4023 kmem_free((caddr_t)dkdevid, un->un_secsize); 4024 return (EINVAL); 4025 } 4026 4027 /* VAlidate the device id */ 4028 if (ddi_devid_valid((ddi_devid_t)&dkdevid->dkd_devid) != DDI_SUCCESS) { 4029 kmem_free((caddr_t)dkdevid, un->un_secsize); 4030 return (EINVAL); 4031 } 4032 4033 /* return a copy of the device id */ 4034 sz = ddi_devid_sizeof((ddi_devid_t)&dkdevid->dkd_devid); 4035 un->un_devid = (ddi_devid_t)kmem_alloc(sz, KM_SLEEP); 4036 bcopy(&dkdevid->dkd_devid, un->un_devid, sz); 4037 kmem_free((caddr_t)dkdevid, un->un_secsize); 4038 4039 return (0); 4040 } 4041 4042 /* 4043 * Return the device id for the device. 4044 * 1. If the device ID exists then just return it - nothing to do in that case. 4045 * 2. Build one from the drives model number and serial number. 4046 * 3. If there is a problem in building it from serial/model #, then try 4047 * to read it from the acyl region of the disk. 4048 * Note: If this function is unable to return a valid ID then the calling 4049 * point will invoke the routine to create a fabricated ID ans stor it on the 4050 * acyl region of the disk. 4051 */ 4052 static ddi_devid_t 4053 dcd_get_devid(struct dcd_disk *un) 4054 { 4055 int rc; 4056 4057 /* If already registered, return that value */ 4058 if (un->un_devid != NULL) 4059 return (un->un_devid); 4060 4061 /* Build a devid from model and serial number, if present */ 4062 rc = dcd_make_devid_from_serial(un); 4063 4064 if (rc != DDI_SUCCESS) { 4065 /* Read the devid from the disk. */ 4066 if (dcd_read_deviceid(un)) 4067 return (NULL); 4068 } 4069 4070 (void) ddi_devid_register(DCD_DEVINFO, un->un_devid); 4071 return (un->un_devid); 4072 } 4073 4074 4075 static ddi_devid_t 4076 dcd_create_devid(struct dcd_disk *un) 4077 { 4078 if (ddi_devid_init(DCD_DEVINFO, DEVID_FAB, 0, NULL, (ddi_devid_t *) 4079 &un->un_devid) == DDI_FAILURE) 4080 return (NULL); 4081 4082 if (dcd_write_deviceid(un)) { 4083 ddi_devid_free(un->un_devid); 4084 un->un_devid = NULL; 4085 return (NULL); 4086 } 4087 4088 (void) ddi_devid_register(DCD_DEVINFO, un->un_devid); 4089 return (un->un_devid); 4090 } 4091 4092 /* 4093 * Build a devid from the model and serial number, if present 4094 * Return DDI_SUCCESS or DDI_FAILURE. 4095 */ 4096 static int 4097 dcd_make_devid_from_serial(struct dcd_disk *un) 4098 { 4099 int rc = DDI_SUCCESS; 4100 char *hwid; 4101 char *model; 4102 int model_len; 4103 char *serno; 4104 int serno_len; 4105 int total_len; 4106 4107 /* initialize the model and serial number information */ 4108 model = un->un_dcd->dcd_ident->dcd_model; 4109 model_len = DCD_MODEL_NUMBER_LENGTH; 4110 serno = un->un_dcd->dcd_ident->dcd_drvser; 4111 serno_len = DCD_SERIAL_NUMBER_LENGTH; 4112 4113 /* Verify the model and serial number */ 4114 dcd_validate_model_serial(model, &model_len, model_len); 4115 if (model_len == 0) { 4116 rc = DDI_FAILURE; 4117 goto out; 4118 } 4119 dcd_validate_model_serial(serno, &serno_len, serno_len); 4120 if (serno_len == 0) { 4121 rc = DDI_FAILURE; 4122 goto out; 4123 } 4124 4125 /* 4126 * The device ID will be concatenation of the model number, 4127 * the '=' separator, the serial number. Allocate 4128 * the string and concatenate the components. 4129 */ 4130 total_len = model_len + 1 + serno_len; 4131 hwid = kmem_alloc(total_len, KM_SLEEP); 4132 bcopy((caddr_t)model, (caddr_t)hwid, model_len); 4133 bcopy((caddr_t)"=", (caddr_t)&hwid[model_len], 1); 4134 bcopy((caddr_t)serno, (caddr_t)&hwid[model_len + 1], serno_len); 4135 4136 /* Initialize the device ID, trailing NULL not included */ 4137 rc = ddi_devid_init(DCD_DEVINFO, DEVID_ATA_SERIAL, total_len, 4138 hwid, (ddi_devid_t *)&un->un_devid); 4139 4140 /* Free the allocated string */ 4141 kmem_free(hwid, total_len); 4142 4143 out: return (rc); 4144 } 4145 4146 /* 4147 * Test for a valid model or serial number. Assume that a valid representation 4148 * contains at least one character that is neither a space, 0 digit, or NULL. 4149 * Trim trailing blanks and NULLS from returned length. 4150 */ 4151 static void 4152 dcd_validate_model_serial(char *str, int *retlen, int totallen) 4153 { 4154 char ch; 4155 boolean_t ret = B_FALSE; 4156 int i; 4157 int tb; 4158 4159 for (i = 0, tb = 0; i < totallen; i++) { 4160 ch = *str++; 4161 if ((ch != ' ') && (ch != '\0') && (ch != '0')) 4162 ret = B_TRUE; 4163 if ((ch == ' ') || (ch == '\0')) 4164 tb++; 4165 else 4166 tb = 0; 4167 } 4168 4169 if (ret == B_TRUE) { 4170 /* Atleast one non 0 or blank character. */ 4171 *retlen = totallen - tb; 4172 } else { 4173 *retlen = 0; 4174 } 4175 } 4176 4177 #ifndef lint 4178 void 4179 clean_print(dev_info_t *dev, char *label, uint_t level, 4180 char *title, char *data, int len) 4181 { 4182 int i; 4183 char buf[256]; 4184 4185 (void) sprintf(buf, "%s:", title); 4186 for (i = 0; i < len; i++) { 4187 (void) sprintf(&buf[strlen(buf)], "0x%x ", (data[i] & 0xff)); 4188 } 4189 (void) sprintf(&buf[strlen(buf)], "\n"); 4190 4191 dcd_log(dev, label, level, "%s", buf); 4192 } 4193 #endif /* Not lint */ 4194 4195 #ifndef lint 4196 /* 4197 * Print a piece of inquiry data- cleaned up for non-printable characters 4198 * and stopping at the first space character after the beginning of the 4199 * passed string; 4200 */ 4201 4202 void 4203 inq_fill(char *p, int l, char *s) 4204 { 4205 unsigned i = 0; 4206 char c; 4207 4208 while (i++ < l) { 4209 if ((c = *p++) < ' ' || c >= 0177) { 4210 c = '*'; 4211 } else if (i != 1 && c == ' ') { 4212 break; 4213 } 4214 *s++ = c; 4215 } 4216 *s++ = 0; 4217 } 4218 #endif /* Not lint */ 4219 4220 char * 4221 dcd_sname(uchar_t status) 4222 { 4223 switch (status & STATUS_ATA_MASK) { 4224 case STATUS_GOOD: 4225 return ("good status"); 4226 4227 case STATUS_ATA_BUSY: 4228 return ("busy"); 4229 4230 default: 4231 return ("<unknown status>"); 4232 } 4233 } 4234 4235 /* ARGSUSED0 */ 4236 char * 4237 dcd_rname(int reason) 4238 { 4239 static char *rnames[] = { 4240 "cmplt", 4241 "incomplete", 4242 "dma_derr", 4243 "tran_err", 4244 "reset", 4245 "aborted", 4246 "timeout", 4247 "data_ovr", 4248 }; 4249 if (reason > CMD_DATA_OVR) { 4250 return ("<unknown reason>"); 4251 } else { 4252 return (rnames[reason]); 4253 } 4254 } 4255 4256 4257 4258 /* ARGSUSED0 */ 4259 int 4260 dcd_check_wp(dev_t dev) 4261 { 4262 4263 return (0); 4264 } 4265 4266 /* 4267 * Create device error kstats 4268 */ 4269 static int 4270 dcd_create_errstats(struct dcd_disk *un, int instance) 4271 { 4272 4273 char kstatname[KSTAT_STRLEN]; 4274 4275 if (un->un_errstats == (kstat_t *)0) { 4276 (void) sprintf(kstatname, "dad%d,error", instance); 4277 un->un_errstats = kstat_create("daderror", instance, kstatname, 4278 "device_error", KSTAT_TYPE_NAMED, 4279 sizeof (struct dcd_errstats)/ sizeof (kstat_named_t), 4280 KSTAT_FLAG_PERSISTENT); 4281 4282 if (un->un_errstats) { 4283 struct dcd_errstats *dtp; 4284 4285 dtp = (struct dcd_errstats *)un->un_errstats->ks_data; 4286 kstat_named_init(&dtp->dcd_softerrs, "Soft Errors", 4287 KSTAT_DATA_UINT32); 4288 kstat_named_init(&dtp->dcd_harderrs, "Hard Errors", 4289 KSTAT_DATA_UINT32); 4290 kstat_named_init(&dtp->dcd_transerrs, 4291 "Transport Errors", KSTAT_DATA_UINT32); 4292 kstat_named_init(&dtp->dcd_model, "Model", 4293 KSTAT_DATA_CHAR); 4294 kstat_named_init(&dtp->dcd_revision, "Revision", 4295 KSTAT_DATA_CHAR); 4296 kstat_named_init(&dtp->dcd_serial, "Serial No", 4297 KSTAT_DATA_CHAR); 4298 kstat_named_init(&dtp->dcd_capacity, "Size", 4299 KSTAT_DATA_ULONGLONG); 4300 kstat_named_init(&dtp->dcd_rq_media_err, "Media Error", 4301 KSTAT_DATA_UINT32); 4302 kstat_named_init(&dtp->dcd_rq_ntrdy_err, 4303 "Device Not Ready", KSTAT_DATA_UINT32); 4304 kstat_named_init(&dtp->dcd_rq_nodev_err, " No Device", 4305 KSTAT_DATA_UINT32); 4306 kstat_named_init(&dtp->dcd_rq_recov_err, "Recoverable", 4307 KSTAT_DATA_UINT32); 4308 kstat_named_init(&dtp->dcd_rq_illrq_err, 4309 "Illegal Request", KSTAT_DATA_UINT32); 4310 4311 un->un_errstats->ks_private = un; 4312 un->un_errstats->ks_update = nulldev; 4313 kstat_install(un->un_errstats); 4314 4315 (void) strncpy(&dtp->dcd_model.value.c[0], 4316 un->un_dcd->dcd_ident->dcd_model, 16); 4317 (void) strncpy(&dtp->dcd_serial.value.c[0], 4318 un->un_dcd->dcd_ident->dcd_drvser, 16); 4319 (void) strncpy(&dtp->dcd_revision.value.c[0], 4320 un->un_dcd->dcd_ident->dcd_fw, 8); 4321 dtp->dcd_capacity.value.ui64 = 4322 (uint64_t)((uint64_t)un->un_diskcapacity * 4323 (uint64_t)un->un_lbasize); 4324 } 4325 } 4326 return (0); 4327 } 4328 4329 4330 /* 4331 * This has been moved from DADA layer as this does not do anything other than 4332 * retrying the command when it is busy or it does not complete 4333 */ 4334 int 4335 dcd_poll(struct dcd_pkt *pkt) 4336 { 4337 int busy_count, rval = -1, savef; 4338 clock_t savet; 4339 void (*savec)(); 4340 4341 4342 /* 4343 * Save old flags 4344 */ 4345 savef = pkt->pkt_flags; 4346 savec = pkt->pkt_comp; 4347 savet = pkt->pkt_time; 4348 4349 pkt->pkt_flags |= FLAG_NOINTR; 4350 4351 4352 /* 4353 * Set the Pkt_comp to NULL 4354 */ 4355 4356 pkt->pkt_comp = 0; 4357 4358 /* 4359 * Set the Pkt time for the polled command 4360 */ 4361 if (pkt->pkt_time == 0) { 4362 pkt->pkt_time = DCD_POLL_TIMEOUT; 4363 } 4364 4365 4366 /* Now transport the command */ 4367 for (busy_count = 0; busy_count < dcd_poll_busycnt; busy_count++) { 4368 if ((rval = dcd_transport(pkt)) == TRAN_ACCEPT) { 4369 if (pkt->pkt_reason == CMD_INCOMPLETE && 4370 pkt->pkt_state == 0) { 4371 delay(100); 4372 } else if (pkt->pkt_reason == CMD_CMPLT) { 4373 rval = 0; 4374 break; 4375 } 4376 } 4377 if (rval == TRAN_BUSY) { 4378 delay(100); 4379 continue; 4380 } 4381 } 4382 4383 pkt->pkt_flags = savef; 4384 pkt->pkt_comp = savec; 4385 pkt->pkt_time = savet; 4386 return (rval); 4387 } 4388 4389 4390 void 4391 dcd_translate(struct dadkio_status32 *statp, struct udcd_cmd *cmdp) 4392 { 4393 if (cmdp->udcd_status_reg & STATUS_ATA_BUSY) 4394 statp->status = DADKIO_STAT_NOT_READY; 4395 else if (cmdp->udcd_status_reg & STATUS_ATA_DWF) 4396 statp->status = DADKIO_STAT_HARDWARE_ERROR; 4397 else if (cmdp->udcd_status_reg & STATUS_ATA_CORR) 4398 statp->status = DADKIO_STAT_SOFT_ERROR; 4399 else if (cmdp->udcd_status_reg & STATUS_ATA_ERR) { 4400 /* 4401 * The error register is valid only when BSY and DRQ not set 4402 * Assumed that HBA has checked this before it gives the data 4403 */ 4404 if (cmdp->udcd_error_reg & ERR_AMNF) 4405 statp->status = DADKIO_STAT_NOT_FORMATTED; 4406 else if (cmdp->udcd_error_reg & ERR_TKONF) 4407 statp->status = DADKIO_STAT_NOT_FORMATTED; 4408 else if (cmdp->udcd_error_reg & ERR_ABORT) 4409 statp->status = DADKIO_STAT_ILLEGAL_REQUEST; 4410 else if (cmdp->udcd_error_reg & ERR_IDNF) 4411 statp->status = DADKIO_STAT_NOT_FORMATTED; 4412 else if (cmdp->udcd_error_reg & ERR_UNC) 4413 statp->status = DADKIO_STAT_BUS_ERROR; 4414 else if (cmdp->udcd_error_reg & ERR_BBK) 4415 statp->status = DADKIO_STAT_MEDIUM_ERROR; 4416 } else 4417 statp->status = DADKIO_STAT_NO_ERROR; 4418 } 4419 4420 static void 4421 dcd_flush_cache(struct dcd_disk *un) 4422 { 4423 struct dcd_pkt *pkt; 4424 int retry_count; 4425 4426 4427 if ((pkt = dcd_init_pkt(ROUTE, NULL, NULL, 4428 (uint32_t)sizeof (struct dcd_cmd), 2, PP_LEN, 4429 PKT_CONSISTENT, NULL_FUNC, NULL)) == NULL) { 4430 return; 4431 } 4432 4433 makecommand(pkt, 0, ATA_FLUSH_CACHE, 0, ADD_LBA_MODE, 0, 4434 NO_DATA_XFER, 0); 4435 4436 /* 4437 * Send the command. There are chances it might fail on some 4438 * disks since it is not a mandatory command as per ata-4. Try 4439 * 3 times if it fails. The retry count has been randomly selected. 4440 * There is a need for retry since as per the spec FLUSH CACHE can fail 4441 * as a result of unrecoverable error encountered during execution 4442 * of writing data and subsequent command should continue flushing 4443 * cache. 4444 */ 4445 for (retry_count = 0; retry_count < 3; retry_count++) { 4446 /* 4447 * Set the packet fields. 4448 */ 4449 pkt->pkt_comp = 0; 4450 pkt->pkt_time = DCD_POLL_TIMEOUT; 4451 pkt->pkt_flags |= FLAG_FORCENOINTR; 4452 pkt->pkt_flags |= FLAG_NOINTR; 4453 if (dcd_transport(pkt) == TRAN_ACCEPT) { 4454 if (pkt->pkt_reason == CMD_CMPLT) { 4455 break; 4456 } 4457 } 4458 /* 4459 * Note the wait time value of 100ms is same as in the 4460 * dcd_poll routine. 4461 */ 4462 drv_usecwait(1000000); 4463 } 4464 (void) dcd_destroy_pkt(pkt); 4465 } 4466 4467 static int 4468 dcd_send_lb_rw_cmd(dev_info_t *devi, void *bufaddr, 4469 diskaddr_t start_block, size_t reqlength, uchar_t cmd) 4470 { 4471 struct dcd_pkt *pkt; 4472 struct buf *bp; 4473 diskaddr_t real_addr = start_block; 4474 size_t buffer_size = reqlength; 4475 uchar_t command, tmp; 4476 int i, rval = 0; 4477 struct dcd_disk *un; 4478 4479 un = ddi_get_soft_state(dcd_state, ddi_get_instance(devi)); 4480 if (un == NULL) 4481 return (ENXIO); 4482 4483 bp = dcd_alloc_consistent_buf(ROUTE, (struct buf *)NULL, 4484 buffer_size, B_READ, NULL_FUNC, NULL); 4485 if (!bp) { 4486 dcd_log(DCD_DEVINFO, dcd_label, CE_WARN, 4487 "no bp for disk label\n"); 4488 return (ENOMEM); 4489 } 4490 4491 pkt = dcd_init_pkt(ROUTE, (struct dcd_pkt *)NULL, 4492 bp, (uint32_t)sizeof (struct dcd_cmd), 2, PP_LEN, 4493 PKT_CONSISTENT, NULL_FUNC, NULL); 4494 4495 if (!pkt) { 4496 dcd_log(DCD_DEVINFO, dcd_label, CE_WARN, 4497 "no memory for disk label\n"); 4498 dcd_free_consistent_buf(bp); 4499 return (ENOMEM); 4500 } 4501 4502 if (cmd == TG_READ) { 4503 bzero(bp->b_un.b_addr, buffer_size); 4504 tmp = DATA_READ; 4505 } else { 4506 bcopy((caddr_t)bufaddr, bp->b_un.b_addr, buffer_size); 4507 tmp = DATA_WRITE; 4508 } 4509 4510 mutex_enter(DCD_MUTEX); 4511 if ((un->un_dp->options & DMA_SUPPORTTED) == DMA_SUPPORTTED) { 4512 if (cmd == TG_READ) { 4513 command = ATA_READ_DMA; 4514 } else { 4515 command = ATA_WRITE_DMA; 4516 } 4517 } else { 4518 if (cmd == TG_READ) { 4519 if (un->un_dp->options & BLOCK_MODE) 4520 command = ATA_READ_MULTIPLE; 4521 else 4522 command = ATA_READ; 4523 } else { 4524 if (un->un_dp->options & BLOCK_MODE) 4525 command = ATA_READ_MULTIPLE; 4526 else 4527 command = ATA_WRITE; 4528 } 4529 } 4530 mutex_exit(DCD_MUTEX); 4531 (void) makecommand(pkt, 0, command, real_addr, ADD_LBA_MODE, 4532 buffer_size, tmp, 0); 4533 4534 for (i = 0; i < 3; i++) { 4535 if (dcd_poll(pkt) || SCBP_C(pkt) != STATUS_GOOD || 4536 (pkt->pkt_state & STATE_XFERRED_DATA) == 0 || 4537 (pkt->pkt_resid != 0)) { 4538 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 4539 "Status %x, state %x, resid %lx\n", 4540 SCBP_C(pkt), pkt->pkt_state, pkt->pkt_resid); 4541 rval = EIO; 4542 } else { 4543 break; 4544 } 4545 } 4546 4547 if (rval != 0) { 4548 dcd_destroy_pkt(pkt); 4549 dcd_free_consistent_buf(bp); 4550 return (EIO); 4551 } 4552 4553 if (cmd == TG_READ) { 4554 bcopy(bp->b_un.b_addr, bufaddr, reqlength); 4555 rval = 0; 4556 } 4557 4558 dcd_destroy_pkt(pkt); 4559 dcd_free_consistent_buf(bp); 4560 return (rval); 4561 } 4562 4563 static int dcd_compute_dk_capacity(struct dcd_device *devp, 4564 diskaddr_t *capacity) 4565 { 4566 diskaddr_t cap; 4567 diskaddr_t no_of_lbasec; 4568 4569 cap = devp->dcd_ident->dcd_fixcyls * 4570 devp->dcd_ident->dcd_heads * 4571 devp->dcd_ident->dcd_sectors; 4572 no_of_lbasec = devp->dcd_ident->dcd_addrsec[1]; 4573 no_of_lbasec = no_of_lbasec << 16; 4574 no_of_lbasec = no_of_lbasec | devp->dcd_ident->dcd_addrsec[0]; 4575 4576 if (no_of_lbasec > cap) { 4577 cap = no_of_lbasec; 4578 } 4579 4580 if (cap != ((uint32_t)-1)) 4581 *capacity = cap; 4582 else 4583 return (EINVAL); 4584 return (0); 4585 } 4586 4587 /*ARGSUSED5*/ 4588 static int 4589 dcd_lb_rdwr(dev_info_t *devi, uchar_t cmd, void *bufaddr, 4590 diskaddr_t start_block, size_t reqlength, void *tg_cookie) 4591 { 4592 if (cmd != TG_READ && cmd != TG_WRITE) 4593 return (EINVAL); 4594 4595 return (dcd_send_lb_rw_cmd(devi, bufaddr, start_block, 4596 reqlength, cmd)); 4597 } 4598 4599 static int 4600 dcd_lb_getphygeom(dev_info_t *devi, cmlb_geom_t *phygeomp) 4601 { 4602 struct dcd_device *devp; 4603 uint32_t no_of_lbasec, capacity, calculated_cylinders; 4604 4605 devp = ddi_get_driver_private(devi); 4606 4607 if ((devp->dcd_ident->dcd_config & ATAPI_DEVICE) == 0) { 4608 if (devp->dcd_ident->dcd_config & ATANON_REMOVABLE) { 4609 phygeomp->g_ncyl = devp->dcd_ident->dcd_fixcyls - 2; 4610 phygeomp->g_acyl = 2; 4611 phygeomp->g_nhead = devp->dcd_ident->dcd_heads; 4612 phygeomp->g_nsect = devp->dcd_ident->dcd_sectors; 4613 4614 no_of_lbasec = devp->dcd_ident->dcd_addrsec[1]; 4615 no_of_lbasec = no_of_lbasec << 16; 4616 no_of_lbasec = no_of_lbasec | 4617 devp->dcd_ident->dcd_addrsec[0]; 4618 capacity = devp->dcd_ident->dcd_fixcyls * 4619 devp->dcd_ident->dcd_heads * 4620 devp->dcd_ident->dcd_sectors; 4621 if (no_of_lbasec > capacity) { 4622 capacity = no_of_lbasec; 4623 if (capacity > NUM_SECTORS_32G) { 4624 /* 4625 * if the capacity is greater than 32G, 4626 * then 255 is the sectors per track. 4627 * This should be good until 128G disk 4628 * capacity, which is the current ATA-4 4629 * limitation. 4630 */ 4631 phygeomp->g_nsect = 255; 4632 } 4633 4634 /* 4635 * If the disk capacity is >= 128GB then no. of 4636 * addressable sectors will be set to 0xfffffff 4637 * in the IDENTIFY info. In that case set the 4638 * no. of pcyl to the Max. 16bit value. 4639 */ 4640 4641 calculated_cylinders = (capacity) / 4642 (phygeomp->g_nhead * phygeomp->g_nsect); 4643 if (calculated_cylinders >= USHRT_MAX) { 4644 phygeomp->g_ncyl = USHRT_MAX - 2; 4645 } else { 4646 phygeomp->g_ncyl = 4647 calculated_cylinders - 2; 4648 } 4649 } 4650 4651 phygeomp->g_capacity = capacity; 4652 phygeomp->g_intrlv = 0; 4653 phygeomp->g_rpm = 5400; 4654 phygeomp->g_secsize = devp->dcd_ident->dcd_secsiz; 4655 4656 return (0); 4657 } else 4658 return (ENOTSUP); 4659 } else { 4660 return (EINVAL); 4661 } 4662 } 4663 4664 4665 /*ARGSUSED3*/ 4666 static int 4667 dcd_lb_getinfo(dev_info_t *devi, int cmd, void *arg, void *tg_cookie) 4668 { 4669 struct dcd_disk *un; 4670 4671 un = ddi_get_soft_state(dcd_state, ddi_get_instance(devi)); 4672 4673 if (un == NULL) 4674 return (ENXIO); 4675 4676 switch (cmd) { 4677 case TG_GETPHYGEOM: 4678 return (dcd_lb_getphygeom(devi, (cmlb_geom_t *)arg)); 4679 4680 case TG_GETVIRTGEOM: 4681 return (-1); 4682 4683 case TG_GETCAPACITY: 4684 case TG_GETBLOCKSIZE: 4685 mutex_enter(DCD_MUTEX); 4686 if (un->un_diskcapacity <= 0) { 4687 mutex_exit(DCD_MUTEX); 4688 dcd_log(DCD_DEVINFO, dcd_label, CE_WARN, 4689 "invalid disk capacity\n"); 4690 return (EIO); 4691 } 4692 if (cmd == TG_GETCAPACITY) 4693 *(diskaddr_t *)arg = un->un_diskcapacity; 4694 else 4695 *(uint32_t *)arg = DEV_BSIZE; 4696 4697 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, "capacity %x\n", 4698 un->un_diskcapacity); 4699 mutex_exit(DCD_MUTEX); 4700 return (0); 4701 4702 case TG_GETATTR: 4703 mutex_enter(DCD_MUTEX); 4704 *(tg_attribute_t *)arg = un->un_tgattribute; 4705 DAD_DEBUG2(DCD_DEVINFO, dcd_label, DCD_DEBUG, 4706 "media_is_writable %x\n", 4707 un->un_tgattribute.media_is_writable); 4708 mutex_exit(DCD_MUTEX); 4709 return (0); 4710 default: 4711 return (ENOTTY); 4712 } 4713 } 4714